{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,11]],"date-time":"2026-03-11T16:30:59Z","timestamp":1773246659085,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":64,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,10,28]],"date-time":"2024-10-28T00:00:00Z","timestamp":1730073600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"DOI":"10.13039\/https:\/\/doi.org\/10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["U20B2063, 62220106008, 62106038"],"award-info":[{"award-number":["U20B2063, 62220106008, 62106038"]}],"id":[{"id":"10.13039\/https:\/\/doi.org\/10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/https:\/\/doi.org\/10.13039\/100012542","name":"Sichuan Province Science and Technology Support Program","doi-asserted-by":"publisher","award":["2024NSFTD0034 and 2023YFG0259"],"award-info":[{"award-number":["2024NSFTD0034 and 2023YFG0259"]}],"id":[{"id":"10.13039\/https:\/\/doi.org\/10.13039\/100012542","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,10,28]]},"DOI":"10.1145\/3664647.3681186","type":"proceedings-article","created":{"date-parts":[[2024,10,26]],"date-time":"2024-10-26T06:59:49Z","timestamp":1729925989000},"page":"8441-8450","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":16,"title":["Q-SNNs: Quantized Spiking Neural Networks"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0002-7753-2948","authenticated-orcid":false,"given":"Wenjie","family":"Wei","sequence":"first","affiliation":[{"name":"The University of Electronic Science and Technology of China, Chengdu, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0003-6080-0173","authenticated-orcid":false,"given":"Yu","family":"Liang","sequence":"additional","affiliation":[{"name":"The University of Electronic Science and Technology of China, Chengdu, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1927-9366","authenticated-orcid":false,"given":"Ammar","family":"Belatreche","sequence":"additional","affiliation":[{"name":"Northumbria University, Newcastle, United Kingdom"}]},{"ORCID":"https:\/\/orcid.org\/0009-0009-0524-6281","authenticated-orcid":false,"given":"Yichen","family":"Xiao","sequence":"additional","affiliation":[{"name":"The University of Electronic Science and Technology of China, Chengdu, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0002-7851-3810","authenticated-orcid":false,"given":"Honglin","family":"Cao","sequence":"additional","affiliation":[{"name":"The University of Electronic Science and Technology of China, Chengdu, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0007-2346-6101","authenticated-orcid":false,"given":"Zhenbang","family":"Ren","sequence":"additional","affiliation":[{"name":"The University of Electronic Science and Technology of China, Chengdu, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0255-4923","authenticated-orcid":false,"given":"Guoqing","family":"Wang","sequence":"additional","affiliation":[{"name":"The University of Electronic Science and Technology of China, Chengdu, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2345-0974","authenticated-orcid":false,"given":"Malu","family":"Zhang","sequence":"additional","affiliation":[{"name":"University of Electronic Science and Technology of China, Chengdu, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6297-5722","authenticated-orcid":false,"given":"Yang","family":"Yang","sequence":"additional","affiliation":[{"name":"University of Electronic Science and Technology of China, Chengdu, China"}]}],"member":"320","published-online":{"date-parts":[[2024,10,28]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"Truenorth: Design and tool flow of a 65 mw 1 million neuron programmable neurosynaptic chip","author":"Akopyan Filipp","year":"2015","unstructured":"Filipp Akopyan, Jun Sawada, Andrew Cassidy, Rodrigo Alvarez-Icaza, John Arthur, Paul Merolla, Nabil Imam, Yutaka Nakamura, Pallab Datta, Gi-Joon Nam, et al. 2015. Truenorth: Design and tool flow of a 65 mw 1 million neuron programmable neurosynaptic chip. IEEE transactions on computer-aided design of integrated circuits and systems, Vol. 34, 10 (2015), 1537--1557."},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.781"},{"key":"e_1_3_2_1_3_1","volume-title":"Spikingbert: Distilling bert to train spiking language models using implicit differentiation. arXiv preprint arXiv:2308.10873","author":"Bal Malyaban","year":"2023","unstructured":"Malyaban Bal and Abhronil Sengupta. 2023. Spikingbert: Distilling bert to train spiking language models using implicit differentiation. arXiv preprint arXiv:2308.10873 (2023)."},{"key":"e_1_3_2_1_4_1","first-page":"24975","article-title":"Differentiable hierarchical and surrogate gradient search for spiking neural networks","volume":"35","author":"Che Kaiwei","year":"2022","unstructured":"Kaiwei Che, Luziwei Leng, Kaixuan Zhang, Jianguo Zhang, Qinghu Meng, Jie Cheng, Qinghai Guo, and Jianxing Liao. 2022. Differentiable hierarchical and surrogate gradient search for spiking neural networks. Advances in Neural Information Processing Systems, Vol. 35 (2022), 24975--24990.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_5_1","volume-title":"Pruning of deep spiking neural networks through gradient rewiring. arXiv preprint arXiv:2105.04916","author":"Chen Yanqi","year":"2021","unstructured":"Yanqi Chen, Zhaofei Yu, Wei Fang, Tiejun Huang, and Yonghong Tian. 2021. Pruning of deep spiking neural networks through gradient rewiring. arXiv preprint arXiv:2105.04916 (2021)."},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1109\/IJCNN52387.2021.9534111"},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1109\/MM.2018.112130359"},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"e_1_3_2_1_9_1","volume-title":"Comprehensive snn compression using admm optimization and activity regularization","author":"Deng Lei","year":"2021","unstructured":"Lei Deng, Yujie Wu, Yifan Hu, Ling Liang, Guoqi Li, Xing Hu, Yufei Ding, Peng Li, and Yuan Xie. 2021. Comprehensive snn compression using admm optimization and activity regularization. IEEE transactions on neural networks and learning systems, Vol. 34, 6 (2021), 2791--2805."},{"key":"e_1_3_2_1_10_1","volume-title":"Temporal efficient training of spiking neural network via gradient re-weighting. arXiv preprint arXiv:2202.11946","author":"Deng Shikuang","year":"2022","unstructured":"Shikuang Deng, Yuhang Li, Shanghang Zhang, and Shi Gu. 2022. Temporal efficient training of spiking neural network via gradient re-weighting. arXiv preprint arXiv:2202.11946 (2022)."},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1109\/TAI.2024.3374268"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.3389\/fnins.2023.1209795"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2023.109639"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.00042"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","unstructured":"A. L. Hodgkin and A. F. Huxley. [n. d.]. A quantitative description of membrane current and its application to conduction and excitation in nerve. The Journal of Physiology ( [n. d.]) 500--544. https:\/\/doi.org\/10.1113\/jphysiol.1952.sp004764","DOI":"10.1113\/jphysiol.1952.sp004764"},{"key":"e_1_3_2_1_16_1","volume-title":"BiLLM: Pushing the Limit of Post-Training Quantization for LLMs. arXiv preprint arXiv:2402.04291","author":"Huang Wei","year":"2024","unstructured":"Wei Huang, Yangdong Liu, Haotong Qin, Ying Li, Shiming Zhang, Xianglong Liu, Michele Magno, and Xiaojuan Qi. 2024. BiLLM: Pushing the Limit of Post-Training Quantization for LLMs. arXiv preprint arXiv:2402.04291 (2024)."},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1109\/TNN.2003.820440"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1109\/DSLW51110.2021.9523415"},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11063-021-10680-x"},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i07.6787"},{"key":"e_1_3_2_1_21_1","unstructured":"Alex Krizhevsky Geoffrey Hinton et al. 2009. Learning multiple layers of features from tiny images. (2009)."},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICPR48806.2021.9412147"},{"key":"e_1_3_2_1_23_1","volume-title":"Cifar10-dvs: an event-stream dataset for object classification. Frontiers in neuroscience","author":"Li Hongmin","year":"2017","unstructured":"Hongmin Li, Hanchao Liu, Xiangyang Ji, Guoqi Li, and Luping Shi. 2017. Cifar10-dvs: an event-stream dataset for object classification. Frontiers in neuroscience, Vol. 11 (2017), 244131."},{"key":"e_1_3_2_1_24_1","volume-title":"Additive powers-of-two quantization: An efficient non-uniform discretization for neural networks. arXiv preprint arXiv:1909.13144","author":"Li Yuhang","year":"2019","unstructured":"Yuhang Li, Xin Dong, and Wei Wang. 2019. Additive powers-of-two quantization: An efficient non-uniform discretization for neural networks. arXiv preprint arXiv:1909.13144 (2019)."},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1109\/TAI.2024.3352533"},{"key":"e_1_3_2_1_26_1","volume-title":"Proceedings, Part VII. Springer, 631--649","author":"Li Yuhang","year":"2022","unstructured":"Yuhang Li, Youngeun Kim, Hyoungseob Park, Tamar Geller, and Priyadarshini Panda. 2022. Neuromorphic data augmentation for training spiking neural networks. In Computer Vision--ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23--27, 2022, Proceedings, Part VII. Springer, 631--649."},{"key":"e_1_3_2_1_27_1","volume-title":"Towards Efficient Deep Spiking Neural Networks Construction with Spiking Activity based Pruning. arXiv preprint arXiv:2406.01072","author":"Li Yaxin","year":"2024","unstructured":"Yaxin Li, Qi Xu, Jiangrong Shen, Hongming Xu, Long Chen, and Gang Pan. 2024. Towards Efficient Deep Spiking Neural Networks Construction with Spiking Activity based Pruning. arXiv preprint arXiv:2406.01072 (2024)."},{"key":"e_1_3_2_1_28_1","volume-title":"LitE-SNN: Designing Lightweight and Efficient Spiking Neural Network through Spatial-Temporal Compressive Network Search and Joint Optimization. arXiv preprint arXiv:2401.14652","author":"Liu Qianhui","year":"2024","unstructured":"Qianhui Liu, Jiaqi Yan, Malu Zhang, Gang Pan, and Haizhou Li. 2024. LitE-SNN: Designing Lightweight and Efficient Spiking Neural Network through Spatial-Temporal Compressive Network Search and Joint Optimization. arXiv preprint arXiv:2401.14652 (2024)."},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.3389\/fnins.2020.00535"},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1109\/TASLP.2021.3100684"},{"key":"e_1_3_2_1_31_1","volume-title":"Nature","volume":"572","author":"Pei Jing","year":"2019","unstructured":"Jing Pei, Lei Deng, Sen Song, Mingguo Zhao, Youhui Zhang, Shuang Wu, Guanrui Wang, Zhe Zou, Zhenzhi Wu, Wei He, et al. 2019. Towards artificial general intelligence with hybrid Tianjic chip architecture. Nature, Vol. 572, 7767 (2019), 106--111."},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.3389\/fnins.2023.1225871"},{"key":"e_1_3_2_1_33_1","volume-title":"SpikeNAS: A Fast Memory-Aware Neural Architecture Search Framework for Spiking Neural Network Systems. arXiv preprint arXiv:2402.11322","author":"Wicaksana Putra Rachmad Vidya","year":"2024","unstructured":"Rachmad Vidya Wicaksana Putra and Muhammad Shafique. 2024. SpikeNAS: A Fast Memory-Aware Neural Architecture Search Framework for Spiking Neural Network Systems. arXiv preprint arXiv:2402.11322 (2024)."},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2021.06.070"},{"key":"e_1_3_2_1_35_1","volume-title":"Bibert: Accurate fully binarized bert. arXiv preprint arXiv:2203.06390","author":"Qin Haotong","year":"2022","unstructured":"Haotong Qin, Yifu Ding, Mingyuan Zhang, Qinghua Yan, Aishan Liu, Qingqing Dang, Ziwei Liu, and Xianglong Liu. 2022. Bibert: Accurate fully binarized bert. arXiv preprint arXiv:2203.06390 (2022)."},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2020.107281"},{"key":"e_1_3_2_1_37_1","volume-title":"Accurate LoRA-Finetuning Quantization of LLMs via Information Retention. arXiv preprint arXiv:2402.05445","author":"Qin Haotong","year":"2024","unstructured":"Haotong Qin, Xudong Ma, Xingyu Zheng, Xiaoyang Li, Yang Zhang, Shouda Liu, Jie Luo, Xianglong Liu, and Michele Magno. 2024. Accurate LoRA-Finetuning Quantization of LLMs via Information Retention. arXiv preprint arXiv:2402.05445 (2024)."},{"key":"e_1_3_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46493-0_32"},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCC.2019.00020"},{"key":"e_1_3_2_1_40_1","doi-asserted-by":"publisher","DOI":"10.3389\/fnins.2017.00682"},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i1.25079"},{"key":"e_1_3_2_1_42_1","volume-title":"Towards Energy Efficient Spiking Neural Networks: An Unstructured Pruning Framework. In The Twelfth International Conference on Learning Representations.","author":"Shi Xinyu","year":"2023","unstructured":"Xinyu Shi, Jianhao Ding, Zecheng Hao, and Zhaofei Yu. 2023. Towards Energy Efficient Spiking Neural Networks: An Unstructured Pruning Framework. In The Twelfth International Conference on Learning Representations."},{"key":"e_1_3_2_1_43_1","volume-title":"Optimizing the energy consumption of spiking neural networks for neuromorphic applications. Frontiers in neuroscience","author":"Sorbaro Martino","year":"2020","unstructured":"Martino Sorbaro, Qian Liu, Massimo Bortone, and Sadique Sheik. 2020. Optimizing the energy consumption of spiking neural networks for neuromorphic applications. Frontiers in neuroscience, Vol. 14 (2020), 516916."},{"key":"e_1_3_2_1_44_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00603"},{"key":"e_1_3_2_1_45_1","doi-asserted-by":"publisher","DOI":"10.1109\/COOLCHIPS52128.2021.9410323"},{"key":"e_1_3_2_1_46_1","doi-asserted-by":"publisher","DOI":"10.1109\/TCDS.2020.2971655"},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00968"},{"key":"e_1_3_2_1_48_1","volume-title":"Event-Driven Learning for Spiking Neural Networks. arXiv preprint arXiv:2403.00270","author":"Wei Wenjie","year":"2024","unstructured":"Wenjie Wei, Malu Zhang, Jilin Zhang, Ammar Belatreche, Jibin Wu, Zijing Xu, Xuerui Qiu, Hong Chen, Yang Yang, and Haizhou Li. 2024. Event-Driven Learning for Spiking Neural Networks. arXiv preprint arXiv:2403.00270 (2024)."},{"key":"e_1_3_2_1_49_1","volume-title":"A spiking neural network framework for robust sound classification. Frontiers in neuroscience","author":"Wu Jibin","year":"2018","unstructured":"Jibin Wu, Yansong Chua, Malu Zhang, Haizhou Li, and Kay Chen Tan. 2018. A spiking neural network framework for robust sound classification. Frontiers in neuroscience, Vol. 12 (2018), 836."},{"key":"e_1_3_2_1_50_1","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2021.3114196"},{"key":"e_1_3_2_1_51_1","volume-title":"Deep spiking neural networks for large vocabulary automatic speech recognition. Frontiers in neuroscience","author":"Wu Jibin","year":"2020","unstructured":"Jibin Wu, Emre Yilmaz, Malu Zhang, Haizhou Li, and Kay Chen Tan. 2020. Deep spiking neural networks for large vocabulary automatic speech recognition. Frontiers in neuroscience, Vol. 14 (2020), 199."},{"key":"e_1_3_2_1_52_1","volume-title":"Spatio-temporal backpropagation for training high-performance spiking neural networks. Frontiers in neuroscience","author":"Wu Yujie","year":"2018","unstructured":"Yujie Wu, Lei Deng, Guoqi Li, and Luping Shi. 2018. Spatio-temporal backpropagation for training high-performance spiking neural networks. Frontiers in neuroscience, Vol. 12 (2018), 323875."},{"key":"e_1_3_2_1_53_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v33i01.33011311"},{"key":"e_1_3_2_1_54_1","unstructured":"Qi Xu Jie Deng Jiangrong Shen Biwu Chen Huajin Tang and Gang Pan. [n. d.]. Hybrid Spiking Vision Transformer for Event-Based Object Detection. Available at SSRN 4790563 ( [n. d.])."},{"key":"e_1_3_2_1_55_1","volume-title":"Biologically inspired structure learning with reverse knowledge distillation for spiking neural networks. arXiv preprint arXiv:2304.09500","author":"Xu Qi","year":"2023","unstructured":"Qi Xu, Yaxin Li, Xuanye Fang, Jiangrong Shen, Jian K Liu, Huajin Tang, and Gang Pan. 2023. Biologically inspired structure learning with reverse knowledge distillation for spiking neural networks. arXiv preprint arXiv:2304.09500 (2023)."},{"key":"e_1_3_2_1_56_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00762"},{"key":"e_1_3_2_1_57_1","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2023.3241201"},{"key":"e_1_3_2_1_58_1","volume-title":"Workload-balanced pruning for sparse spiking neural networks. arXiv preprint arXiv:2302.06746","author":"Yin Ruokai","year":"2023","unstructured":"Ruokai Yin, Youngeun Kim, Yuhang Li, Abhishek Moitra, Nitin Satpute, Anna Hambitzer, and Priyadarshini Panda. 2023. Workload-balanced pruning for sparse spiking neural networks. arXiv preprint arXiv:2302.06746 (2023)."},{"key":"e_1_3_2_1_59_1","unstructured":"Ruokai Yin Yuhang Li Abhishek Moitra and Priyadarshini Panda. [n. d.]. MINT: Multiplier-less INTeger Quantization for Energy Efficient Spiking Neural Networks. ( [n. d.])."},{"key":"e_1_3_2_1_60_1","doi-asserted-by":"publisher","DOI":"10.1109\/JETCAS.2023.3328911"},{"key":"e_1_3_2_1_61_1","volume-title":"A highly effective and robust membrane potential-driven supervised learning method for spiking neurons","author":"Zhang Malu","year":"2018","unstructured":"Malu Zhang, Hong Qu, Ammar Belatreche, Yi Chen, and Zhang Yi. 2018. A highly effective and robust membrane potential-driven supervised learning method for spiking neurons. IEEE transactions on neural networks and learning systems, Vol. 30, 1 (2018), 123--137."},{"key":"e_1_3_2_1_62_1","volume-title":"Hong Qu, Yansong Chua, Trevor E Carlson, et al.","author":"Zhang Malu","year":"2021","unstructured":"Malu Zhang, Jiadong Wang, Jibin Wu, Ammar Belatreche, Burin Amornpaisannon, Zhixuan Zhang, Venkata Pavan Kumar Miriyala, Hong Qu, Yansong Chua, Trevor E Carlson, et al. 2021. Rectified linear postsynaptic potential function for backpropagation in deep spiking neural networks. IEEE transactions on neural networks and learning systems, Vol. 33, 5 (2021), 1947--1958."},{"key":"e_1_3_2_1_63_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i12.17329"},{"key":"e_1_3_2_1_64_1","volume-title":"Spikegpt: Generative pre-trained language model with spiking neural networks. arXiv preprint arXiv:2302.13939","author":"Zhu Rui-Jie","year":"2023","unstructured":"Rui-Jie Zhu, Qihang Zhao, Guoqi Li, and Jason K Eshraghian. 2023. Spikegpt: Generative pre-trained language model with spiking neural networks. arXiv preprint arXiv:2302.13939 (2023)."}],"event":{"name":"MM '24: The 32nd ACM International Conference on Multimedia","location":"Melbourne VIC Australia","acronym":"MM '24","sponsor":["SIGMM ACM Special Interest Group on Multimedia"]},"container-title":["Proceedings of the 32nd ACM International Conference on Multimedia"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3664647.3681186","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3664647.3681186","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T01:18:02Z","timestamp":1750295882000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3664647.3681186"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,28]]},"references-count":64,"alternative-id":["10.1145\/3664647.3681186","10.1145\/3664647"],"URL":"https:\/\/doi.org\/10.1145\/3664647.3681186","relation":{},"subject":[],"published":{"date-parts":[[2024,10,28]]},"assertion":[{"value":"2024-10-28","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}