{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,23]],"date-time":"2026-03-23T23:09:52Z","timestamp":1774307392866,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":69,"publisher":"ACM","license":[{"start":{"date-parts":[[2019,6,26]],"date-time":"2019-06-26T00:00:00Z","timestamp":1561507200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2019,6,26]]},"DOI":"10.1145\/3330345.3331057","type":"proceedings-article","created":{"date-parts":[[2019,6,18]],"date-time":"2019-06-18T12:14:30Z","timestamp":1560860070000},"page":"46-57","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":77,"title":["Accelerating reduction and scan using tensor core units"],"prefix":"10.1145","author":[{"given":"Abdul","family":"Dakkak","sequence":"first","affiliation":[{"name":"University of Illinois Urbana-Champaign"}]},{"given":"Cheng","family":"Li","sequence":"additional","affiliation":[{"name":"University of Illinois Urbana-Champaign"}]},{"given":"Jinjun","family":"Xiong","sequence":"additional","affiliation":[{"name":"IBM T. J. Watson Research Center"}]},{"given":"Isaac","family":"Gelado","sequence":"additional","affiliation":[{"name":"NVIDIA Corporation"}]},{"given":"Wen-mei","family":"Hwu","sequence":"additional","affiliation":[{"name":"University of Illinois Urbana-Champaign"}]}],"member":"320","published-online":{"date-parts":[[2019,6,26]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"Journal of Physics: Conference Series number 1.","author":"Agullo Emmanuel","year":"2037","unstructured":"Emmanuel Agullo , Jim Demmel , Jack Dongarra , Bilel Hadri , Jakub Kurzak , Julien Langou , Hatem Ltaief , Piotr Luszczek , and Stanimire Tomov . 2009. Numerical linear algebra on emerging architectures: the plasma and magma projects . In Journal of Physics: Conference Series number 1. Vol. 180 . IOP Publishing , 01 2037 . Emmanuel Agullo, Jim Demmel, Jack Dongarra, Bilel Hadri, Jakub Kurzak, Julien Langou, Hatem Ltaief, Piotr Luszczek, and Stanimire Tomov. 2009. Numerical linear algebra on emerging architectures: the plasma and magma projects. In Journal of Physics: Conference Series number 1. Vol. 180. IOP Publishing, 012037."},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.1145\/1543135.1542481"},{"key":"e_1_3_2_1_3_1","unstructured":"Apple. 2019 (accessed January 14 2019). A11 Bionic. https:\/\/www.apple.com\/iphone-x.  Apple. 2019 (accessed January 14 2019). A11 Bionic. https:\/\/www.apple.com\/iphone-x."},{"key":"e_1_3_2_1_4_1","unstructured":"Arm. 2019 (accessed January 14 2019). Arm Machine Learning Processor. https:\/\/developer.arm.com\/products\/processors\/machine-learning\/arm-ml-processor.  Arm. 2019 (accessed January 14 2019). Arm Machine Learning Processor. https:\/\/developer.arm.com\/products\/processors\/machine-learning\/arm-ml-processor."},{"key":"e_1_3_2_1_5_1","volume-title":"GPU computing gems Jade edition","author":"Bell Nathan","unstructured":"Nathan Bell and Jared Hoberock . 2011. Thrust: a productivity-oriented library for cuda . In GPU computing gems Jade edition . Elsevier , 359--371. Nathan Bell and Jared Hoberock. 2011. Thrust: a productivity-oriented library for cuda. In GPU computing gems Jade edition. Elsevier, 359--371."},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1109\/12.42122"},{"key":"e_1_3_2_1_7_1","unstructured":"Guy E Blelloch Michael A Heroux and Marco Zagha. 1993. Segmented operations for sparse matrix computation on vector multiprocessors. Tech. rep. Carnegie-Mellon Univ Pittsburgh PA School of Computer Science.   Guy E Blelloch Michael A Heroux and Marco Zagha. 1993. Segmented operations for sparse matrix computation on vector multiprocessors. Tech. rep. Carnegie-Mellon Univ Pittsburgh PA School of Computer Science."},{"key":"e_1_3_2_1_8_1","unstructured":"Caffe2. 2019 (accessed January 14 2019). Caffe2. https:\/\/caffe2.ai.  Caffe2. 2019 (accessed January 14 2019). Caffe2. https:\/\/caffe2.ai."},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.5555\/1958016.1958031"},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.5555\/3195638.3195653"},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1145\/2790060.2790063"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1109\/99.660313"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1109\/CGO.2019.8661187"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1145\/1375527.1375559"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1109\/TCAD.2016.2584062"},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1145\/2975991.2975996"},{"key":"e_1_3_2_1_17_1","volume-title":"Multireduce and multiscan on modern gpus. Department of Computer Science","author":"Eilers Marco","unstructured":"Marco Eilers . 2014. Multireduce and multiscan on modern gpus. Department of Computer Science , University of Copenhagen. Master's thesis. Marco Eilers. 2014. Multireduce and multiscan on modern gpus. Department of Computer Science, University of Copenhagen. Master's thesis."},{"key":"e_1_3_2_1_18_1","unstructured":"Google. 2019 (accessed January 14 2019). Edge TPU. https:\/\/cloud.google.com\/edge-tpu.  Google. 2019 (accessed January 14 2019). Edge TPU. https:\/\/cloud.google.com\/edge-tpu."},{"key":"e_1_3_2_1_19_1","unstructured":"Google. 2019 (accessed January 14 2019). Google Cloud TPU. https:\/\/cloud.google.com\/tpu.  Google. 2019 (accessed January 14 2019). Google Cloud TPU. https:\/\/cloud.google.com\/tpu."},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1145\/3148226.3148237"},{"key":"e_1_3_2_1_21_1","unstructured":"Mark Harris Shubhabrata Sengupta and John D Owens. 2007. Parallel prefix sum (scan) with cuda. GPU gems 3 39 851--876.  Mark Harris Shubhabrata Sengupta and John D Owens. 2007. Parallel prefix sum (scan) with cuda. GPU gems 3 39 851--876."},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1006\/jcom.1998.0476"},{"key":"e_1_3_2_1_23_1","unstructured":"Sergey Ioffe and Christian Szegedy. 2015. Batch normalization: accelerating deep network training by reducing internal covariate shift. arXiv preprint arXiv:1502.03167.  Sergey Ioffe and Christian Szegedy. 2015. Batch normalization: accelerating deep network training by reducing internal covariate shift. arXiv preprint arXiv:1502.03167."},{"key":"e_1_3_2_1_24_1","unstructured":"X. Jia et al. 2018. Highly Scalable Deep Learning Training System with Mixed-Precision: Training ImageNet in Four Minutes. ArXiv e-prints (July 2018). arXiv: 1807.11205.  X. Jia et al. 2018. Highly Scalable Deep Learning Training System with Mixed-Precision: Training ImageNet in Four Minutes. ArXiv e-prints (July 2018). arXiv: 1807.11205."},{"key":"e_1_3_2_1_25_1","unstructured":"Yangqing Jia Evan Shelhamer Jeff Donahue Sergey Karayev Jonathan Long Ross Girshick Sergio Guadarrama and Trevor Darrell. 2014. Caffe: convolutional architecture for fast feature embedding. arXiv preprint arXiv:1408.5093.  Yangqing Jia Evan Shelhamer Jeff Donahue Sergey Karayev Jonathan Long Ross Girshick Sergio Guadarrama and Trevor Darrell. 2014. Caffe: convolutional architecture for fast feature embedding. arXiv preprint arXiv:1408.5093."},{"key":"e_1_3_2_1_26_1","unstructured":"Zhe Jia Marco Maggioni Benjamin Staiger and Daniele P Scarpazza. 2018. Dissecting the nvidia volta gpu architecture via microbenchmarking. arXiv preprint arXiv:1804.06826.  Zhe Jia Marco Maggioni Benjamin Staiger and Daniele P Scarpazza. 2018. Dissecting the nvidia volta gpu architecture via microbenchmarking. arXiv preprint arXiv:1804.06826."},{"key":"e_1_3_2_1_27_1","volume-title":"Computer Architecture (ISCA), 2017 ACM\/IEEE 44th Annual International Symposium on. IEEE, 1--12","author":"Norman","unstructured":"Norman P Jouppi et al. 2017. In-datacenter performance analysis of a tensor processing unit . In Computer Architecture (ISCA), 2017 ACM\/IEEE 44th Annual International Symposium on. IEEE, 1--12 . Norman P Jouppi et al. 2017. In-datacenter performance analysis of a tensor processing unit. In Computer Architecture (ISCA), 2017 ACM\/IEEE 44th Annual International Symposium on. IEEE, 1--12."},{"key":"e_1_3_2_1_28_1","unstructured":"Julien Demouth. 2019 (accessed January 14 2019). Kepler Shuffle: Tips and Tricks. http:\/\/on-demand.gputechconf.com\/gtc\/2013\/presentations\/S3174-Kepler-Shuffle-Tips-Tricks.pdf.  Julien Demouth. 2019 (accessed January 14 2019). Kepler Shuffle: Tips and Tricks. http:\/\/on-demand.gputechconf.com\/gtc\/2013\/presentations\/S3174-Kepler-Shuffle-Tips-Tricks.pdf."},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICPP.2011.41"},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1137\/07070111X"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1145\/3122948.3122952"},{"key":"e_1_3_2_1_32_1","unstructured":"Lawrence Livermore National Laboratory. 2019 (accessed January 14 2019). Sierra Supercomputer. https:\/\/computation.llnl.gov\/computers\/sierra.  Lawrence Livermore National Laboratory. 2019 (accessed January 14 2019). Sierra Supercomputer. https:\/\/computation.llnl.gov\/computers\/sierra."},{"key":"e_1_3_2_1_33_1","unstructured":"Chris Leary and Todd Wang. 2017. Xla: tensorflow compiled. TensorFlow Dev Summit.  Chris Leary and Todd Wang. 2017. Xla: tensorflow compiled. TensorFlow Dev Summit."},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.1145\/2908080.2908089"},{"key":"e_1_3_2_1_35_1","volume-title":"Erwin Laure, Ivy Bo Peng, and Jeffrey S Vetter.","author":"Markidis Stefano","year":"2018","unstructured":"Stefano Markidis , Steven Wei Der Chien , Erwin Laure, Ivy Bo Peng, and Jeffrey S Vetter. 2018 . Nvidia tensor core programmability, performance & precision. arXiv preprint arXiv:1803.04014. Stefano Markidis, Steven Wei Der Chien, Erwin Laure, Ivy Bo Peng, and Jeffrey S Vetter. 2018. Nvidia tensor core programmability, performance & precision. arXiv preprint arXiv:1803.04014."},{"key":"e_1_3_2_1_36_1","volume-title":"Structured parallel programming: patterns for efficient computation","author":"McCool Michael D","unstructured":"Michael D McCool , Arch D Robison , and James Reinders . 2012. Structured parallel programming: patterns for efficient computation . Elsevier . Michael D McCool, Arch D Robison, and James Reinders. 2012. Structured parallel programming: patterns for efficient computation. Elsevier."},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1145\/2544174.2500595"},{"key":"e_1_3_2_1_38_1","unstructured":"D Merrill. 2018. CUB v1.8.0: CUDA Unbound a library of warp-wide block-wide and device-wide GPU parallel primitives. NVIDIA Research.  D Merrill. 2018. CUB v1.8.0: CUDA Unbound a library of warp-wide block-wide and device-wide GPU parallel primitives. NVIDIA Research."},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1145\/1854273.1854344"},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.1142\/S0129626415500073"},{"key":"e_1_3_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.7717\/peerj-cs.127"},{"key":"e_1_3_2_1_43_1","unstructured":"MXNet. 2019 (accessed January 14 2019). MXNet. https:\/\/mxnet.apache.org.  MXNet. 2019 (accessed January 14 2019). MXNet. https:\/\/mxnet.apache.org."},{"key":"e_1_3_2_1_44_1","unstructured":"Wilt Nicholas. 2013. The cuda handbook: a comprehensive guide to gpu programming. (2013).  Wilt Nicholas. 2013. The cuda handbook: a comprehensive guide to gpu programming. (2013)."},{"key":"e_1_3_2_1_45_1","unstructured":"NVIDIA. 2019 (accessed January 14 2019). Cooperative Groups: Flexible CUDA Thread Programming. https:\/\/devblogs.nvidia.com\/cooperative-groups\/.  NVIDIA. 2019 (accessed January 14 2019). Cooperative Groups: Flexible CUDA Thread Programming. https:\/\/devblogs.nvidia.com\/cooperative-groups\/."},{"key":"e_1_3_2_1_46_1","unstructured":"NVIDIA. 2019 (accessed January 14 2019). cuBLAS. https:\/\/developer.nvidia.com\/cublas.  NVIDIA. 2019 (accessed January 14 2019). cuBLAS. https:\/\/developer.nvidia.com\/cublas."},{"key":"e_1_3_2_1_47_1","unstructured":"NVIDIA. 2019 (accessed January 14 2019). CUDA C Programming Guide. https:\/\/docs.nvidia.com\/cuda\/cuda-c-programming-guide\/index.html.  NVIDIA. 2019 (accessed January 14 2019). CUDA C Programming Guide. https:\/\/docs.nvidia.com\/cuda\/cuda-c-programming-guide\/index.html."},{"key":"e_1_3_2_1_48_1","unstructured":"NVIDIA. 2019 (accessed January 14 2019). cuDNN. https:\/\/developer.nvidia.com\/cudnn.  NVIDIA. 2019 (accessed January 14 2019). cuDNN. https:\/\/developer.nvidia.com\/cudnn."},{"key":"e_1_3_2_1_49_1","unstructured":"NVIDIA. 2019 (accessed January 14 2019). CUTLASS. https:\/\/devblogs.nvidia.com\/cutlass-linear-algebra-cuda.  NVIDIA. 2019 (accessed January 14 2019). CUTLASS. https:\/\/devblogs.nvidia.com\/cutlass-linear-algebra-cuda."},{"key":"e_1_3_2_1_50_1","unstructured":"NVIDIA. 2019 (accessed January 14 2019). Mixed Precision Training. https:\/\/docs.nvidia.com\/deeplearning\/sdk\/mixed-precision-training.  NVIDIA. 2019 (accessed January 14 2019). Mixed Precision Training. https:\/\/docs.nvidia.com\/deeplearning\/sdk\/mixed-precision-training."},{"key":"e_1_3_2_1_51_1","unstructured":"NVIDIA. 2019 (accessed January 14 2019). NVPROF. https:\/\/docs.nvidia.com\/cuda\/profiler-users-guide\/index.html#system-profiling.  NVIDIA. 2019 (accessed January 14 2019). NVPROF. https:\/\/docs.nvidia.com\/cuda\/profiler-users-guide\/index.html#system-profiling."},{"key":"e_1_3_2_1_52_1","unstructured":"NVIDIA. 2019 (accessed January 14 2019). NVVM IR Specification 1.5. https:\/\/docs.nvidia.com\/cuda\/nvvm-ir-spec\/index.html.  NVIDIA. 2019 (accessed January 14 2019). NVVM IR Specification 1.5. https:\/\/docs.nvidia.com\/cuda\/nvvm-ir-spec\/index.html."},{"key":"e_1_3_2_1_53_1","unstructured":"NVIDIA. 2019 (accessed January 14 2019). Parallel Thread Execution ISA Version 6.2. https:\/\/docs.nvidia.com\/cuda\/parallel-thread-execution\/index.html.  NVIDIA. 2019 (accessed January 14 2019). Parallel Thread Execution ISA Version 6.2. https:\/\/docs.nvidia.com\/cuda\/parallel-thread-execution\/index.html."},{"key":"e_1_3_2_1_54_1","unstructured":"NVIDIA. 2019 (accessed January 14 2019). Programming Tensor Cores in CUDA 9. https:\/\/devblogs.nvidia.com\/programming-tensor-cores-cuda-9.  NVIDIA. 2019 (accessed January 14 2019). Programming Tensor Cores in CUDA 9. https:\/\/devblogs.nvidia.com\/programming-tensor-cores-cuda-9."},{"key":"e_1_3_2_1_55_1","unstructured":"NVIDIA. 2019 (accessed January 14 2019). Tensor Cores. https:\/\/www.nvidia.com\/en-us\/data-center\/tensorcore.  NVIDIA. 2019 (accessed January 14 2019). Tensor Cores. https:\/\/www.nvidia.com\/en-us\/data-center\/tensorcore."},{"key":"e_1_3_2_1_56_1","unstructured":"NVIDIA. 2019 (accessed January 14 2019). TensorRT. https:\/\/developer.nvidia.com\/tensorrt.  NVIDIA. 2019 (accessed January 14 2019). TensorRT. https:\/\/developer.nvidia.com\/tensorrt."},{"key":"e_1_3_2_1_57_1","unstructured":"Oak Ridge National Laboratory. 2019 (accessed January 14 2019). Summit Supercomputer. https:\/\/www.olcf.ornl.gov\/summit.  Oak Ridge National Laboratory. 2019 (accessed January 14 2019). Summit Supercomputer. https:\/\/www.olcf.ornl.gov\/summit."},{"key":"e_1_3_2_1_58_1","volume-title":"Spring","author":"P\u00e4rssinen Paavo","year":"2018","unstructured":"Paavo P\u00e4rssinen . {n. d.} Modern mobile graphics processors. Science: Internet, Data and Things (CS-E4000) , Spring 2018 , 211. Paavo P\u00e4rssinen. {n. d.} Modern mobile graphics processors. Science: Internet, Data and Things (CS-E4000), Spring 2018, 211."},{"key":"e_1_3_2_1_59_1","unstructured":"PyTorch. 2019 (accessed January 14 2019). PyTorch. https:\/\/pytorch.org.  PyTorch. 2019 (accessed January 14 2019). PyTorch. https:\/\/pytorch.org."},{"key":"e_1_3_2_1_60_1","unstructured":"Stephan Rabanser Oleksandr Shchur and Stephan G\u00fcnnemann. 2017. Introduction to tensor decompositions and their applications in machine learning. arXiv preprint arXiv:1711.10781.  Stephan Rabanser Oleksandr Shchur and Stephan G\u00fcnnemann. 2017. Introduction to tensor decompositions and their applications in machine learning. arXiv preprint arXiv:1711.10781."},{"key":"e_1_3_2_1_61_1","doi-asserted-by":"publisher","DOI":"10.1145\/3150211"},{"key":"e_1_3_2_1_62_1","volume-title":"Aamodt","author":"Raihan Md Aamir","year":"2018","unstructured":"Md Aamir Raihan , Negar Goli , and Tor M . Aamodt . 2018 . Modeling deep learning accelerator enabled gpus. CoRR , abs\/1811.08309. arXiv: 1811.08309. http:\/\/arxiv.org\/abs\/1811.08309. Md Aamir Raihan, Negar Goli, and Tor M. Aamodt. 2018. Modeling deep learning accelerator enabled gpus. CoRR, abs\/1811.08309. arXiv: 1811.08309. http:\/\/arxiv.org\/abs\/1811.08309."},{"key":"e_1_3_2_1_63_1","doi-asserted-by":"publisher","DOI":"10.2200\/S00783ED1V01Y201706CAC041"},{"key":"e_1_3_2_1_65_1","volume-title":"Workshop on edge computing using new commodity architectures, 26--27","author":"Sengupta Shubhabrata","year":"2006","unstructured":"Shubhabrata Sengupta , Aaron E Lefohn , and John D Owens . 2006 . A work-efficient step-efficient prefix sum algorithm . In Workshop on edge computing using new commodity architectures, 26--27 . Shubhabrata Sengupta, Aaron E Lefohn, and John D Owens. 2006. A work-efficient step-efficient prefix sum algorithm. In Workshop on edge computing using new commodity architectures, 26--27."},{"key":"e_1_3_2_1_66_1","doi-asserted-by":"publisher","DOI":"10.5555\/3049832.3049841"},{"key":"e_1_3_2_1_67_1","unstructured":"TensorFlow. 2019 (accessed January 14 2019). TensorFlow. https:\/\/www.tensorflow.org.  TensorFlow. 2019 (accessed January 14 2019). TensorFlow. https:\/\/www.tensorflow.org."},{"key":"e_1_3_2_1_68_1","doi-asserted-by":"crossref","unstructured":"Charles Van Loan. 1992. A survey of matrix computations. Handbooks in operations research and management science 3 247--321.  Charles Van Loan. 1992. A survey of matrix computations. Handbooks in operations research and management science 3 247--321.","DOI":"10.1016\/S0927-0507(05)80203-8"},{"key":"e_1_3_2_1_69_1","unstructured":"WikiChip. 2019 (accessed January 14 2019). Cascade Lake-Microarchitectures - Intel. https:\/\/en.wikichip.org\/wiki\/intel\/microarchitectures\/cascade_lake.  WikiChip. 2019 (accessed January 14 2019). Cascade Lake-Microarchitectures - Intel. https:\/\/en.wikichip.org\/wiki\/intel\/microarchitectures\/cascade_lake."},{"key":"e_1_3_2_1_70_1","doi-asserted-by":"publisher","DOI":"10.1145\/2517327.2442539"},{"key":"e_1_3_2_1_71_1","unstructured":"Yuhao Zhu Matthew Mattina and Paul Whatmough. 2018. Mobile machine learning hardware at arm: a systems-on-chip (soc) perspective. arXiv preprint arXiv:1801.06274.  Yuhao Zhu Matthew Mattina and Paul Whatmough. 2018. Mobile machine learning hardware at arm: a systems-on-chip (soc) perspective. arXiv preprint arXiv:1801.06274."}],"event":{"name":"ICS '19: 2019 International Conference on Supercomputing","location":"Phoenix Arizona","acronym":"ICS '19","sponsor":["SIGARCH ACM Special Interest Group on Computer Architecture"]},"container-title":["Proceedings of the ACM International Conference on Supercomputing"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3330345.3331057","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3330345.3331057","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T23:53:26Z","timestamp":1750204406000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3330345.3331057"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019,6,26]]},"references-count":69,"alternative-id":["10.1145\/3330345.3331057","10.1145\/3330345"],"URL":"https:\/\/doi.org\/10.1145\/3330345.3331057","relation":{},"subject":[],"published":{"date-parts":[[2019,6,26]]},"assertion":[{"value":"2019-06-26","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}