{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,15]],"date-time":"2026-03-15T15:31:55Z","timestamp":1773588715487,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":61,"publisher":"ACM","funder":[{"name":"National Natural Science Foundation of China","award":["62325205"],"award-info":[{"award-number":["62325205"]}]},{"name":"National Natural Science Foundation of China","award":["62502197"],"award-info":[{"award-number":["62502197"]}]},{"name":"National Natural Science Foundation of China","award":["U25B2035"],"award-info":[{"award-number":["U25B2035"]}]},{"name":"Basic Research Program of Jiangsu","award":["BK20251206"],"award-info":[{"award-number":["BK20251206"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2026,3,22]]},"DOI":"10.1145\/3779212.3790163","type":"proceedings-article","created":{"date-parts":[[2026,3,10]],"date-time":"2026-03-10T13:55:26Z","timestamp":1773150926000},"page":"764-782","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["Fine-grained and Non-intrusive LLM Training Monitoring via Microsecond-level Traffic Measurement"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0002-9036-4563","authenticated-orcid":false,"given":"Yibo","family":"Xiao","sequence":"first","affiliation":[{"name":"State Key Laboratory of Novel Software Technology, Nanjing University, Nanjing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2663-4146","authenticated-orcid":false,"given":"Hao","family":"Zheng","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Novel Software Technology, Nanjing University, Nanjing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9358-8808","authenticated-orcid":false,"given":"Haifeng","family":"Sun","sequence":"additional","affiliation":[{"name":"National University of Singapore, Singapore, Singapore"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5394-8450","authenticated-orcid":false,"given":"Qingkai","family":"Meng","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Novel Software Technology, Nanjing University, Nanjing, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0002-3543-1233","authenticated-orcid":false,"given":"Jiong","family":"Duan","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Novel Software Technology, Nanjing University, Nanjing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1487-2419","authenticated-orcid":false,"given":"Xiaohe","family":"Hu","sequence":"additional","affiliation":[{"name":"Infrawaves, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1565-9997","authenticated-orcid":false,"given":"Rong","family":"Gu","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Novel Software Technology, Nanjing University, Nanjing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6934-1685","authenticated-orcid":false,"given":"Guihai","family":"Chen","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Novel Software Technology, Nanjing University, Nanjing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2710-7628","authenticated-orcid":false,"given":"Chen","family":"Tian","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Novel Software Technology, Nanjing University, Nanjing, China"}]}],"member":"320","published-online":{"date-parts":[[2026,3,22]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"Diogo Almeida, Janko Altenschmidt, Sam Altman, Shyamal Anadkat, et al.","author":"Achiam Josh","year":"2023","unstructured":"Josh Achiam, Steven Adler, Sandhini Agarwal, Lama Ahmad, Ilge Akkaya, Florencia Leoni Aleman, Diogo Almeida, Janko Altenschmidt, Sam Altman, Shyamal Anadkat, et al., 2023. Gpt-4 technical report. arXiv preprint arXiv:2303.08774 (2023)."},{"key":"e_1_3_2_1_2_1","unstructured":"Aliyun. 2025. Aliyun Web Services. https:\/\/www.aliyun.com."},{"key":"e_1_3_2_1_3_1","unstructured":"Amazon. 2025. Amazon Web Services. https:\/\/aws.amazon.com."},{"key":"e_1_3_2_1_4_1","unstructured":"AMD. 2025. AMD Pensando. https:\/\/www.amd.com\/en\/solutions\/data-center\/networking.html."},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1145\/3123878.3131979"},{"key":"e_1_3_2_1_6_1","unstructured":"Azure. 2025. Microsoft Azure. https:\/\/azure.microsoft.com."},{"key":"e_1_3_2_1_7_1","unstructured":"Jinze Bai Shuai Bai Yunfei Chu Zeyu Cui Kai Dang Xiaodong Deng Yang Fan Wenbin Ge Yu Han Fei Huang et al. 2023. Qwen technical report. arXiv preprint arXiv:2309.16609 (2023)."},{"key":"e_1_3_2_1_8_1","volume-title":"INSERT: In-Network Stateful End-to-End RDMA Telemetry. In IEEE INFOCOM 2024-IEEE Conference on Computer Communications. IEEE, 1061-1070","author":"Chang Hyunseok","year":"2024","unstructured":"Hyunseok Chang, Walid A Hanafy, Sarit Mukherjee, and Limin Wang. 2024. INSERT: In-Network Stateful End-to-End RDMA Telemetry. In IEEE INFOCOM 2024-IEEE Conference on Computer Communications. IEEE, 1061-1070."},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICNP61940.2024.10858560"},{"key":"e_1_3_2_1_10_1","volume-title":"Deepseekmoe: Towards ultimate expert specialization in mixture-of-experts language models. arXiv preprint arXiv:2401.06066","author":"Dai Damai","year":"2024","unstructured":"Damai Dai, Chengqi Deng, Chenggang Zhao, RX Xu, Huazuo Gao, Deli Chen, Jiashi Li, Wangding Zeng, Xingkai Yu, Yu Wu, et al., 2024. Deepseekmoe: Towards ultimate expert specialization in mixture-of-experts language models. arXiv preprint arXiv:2401.06066 (2024)."},{"key":"e_1_3_2_1_11_1","volume-title":"Minder: Faulty Machine Detection for Large-scale Distributed Model Training. In 22nd USENIX Symposium on Networked Systems Design and Implementation (NSDI 25)","author":"Deng Yangtao","year":"2025","unstructured":"Yangtao Deng, Xiang Shi, Zhuo Jiang, Xingjian Zhang, Lei Zhang, Zhang Zhang, Bo Li, Zuquan Song, Hang Zhu, Gaohong Liu, et al., 2025. Minder: Faulty Machine Detection for Large-scale Distributed Model Training. In 22nd USENIX Symposium on Networked Systems Design and Implementation (NSDI 25)."},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1145\/3651890.3672256"},{"key":"e_1_3_2_1_13_1","volume-title":"22nd USENIX Symposium on Networked Systems Design and Implementation (NSDI 25)","author":"Dong Jianbo","year":"2025","unstructured":"Jianbo Dong, Kun Qian, Pengcheng Zhang, Zhilong Zheng, Liang Chen, Fei Feng, Yikai Zhu, Gang Lu, Zhihui Ren, Xue Li, et al., 2025. Evolution of Aegis: Fault Diagnosis for AI Model Training Cloud Service in Production (Experience Track). In 22nd USENIX Symposium on Networked Systems Design and Implementation (NSDI 25)."},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1109\/CCGrid57682.2023.00039"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1145\/3517745.3561430"},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1145\/2785956.2787496"},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1145\/3387514.3405877"},{"key":"e_1_3_2_1_18_1","first-page":"331","volume-title":"19th USENIX Symposium on Operating Systems Design and Implementation (OSDI 25)","author":"Huang Songlin","year":"2025","unstructured":"Songlin Huang and Chenshu Wu. 2025. Neutrino: Fine-grained Kernel Profiling via Programmable Probing. In 19th USENIX Symposium on Operating Systems Design and Implementation (OSDI 25). 331-355."},{"key":"e_1_3_2_1_19_1","volume-title":"Gpipe: Efficient training of giant neural networks using pipeline parallelism. Advances in neural information processing systems","author":"Huang Yanping","year":"2019","unstructured":"Yanping Huang, Youlong Cheng, Ankur Bapna, Orhan Firat, Dehao Chen, Mia Chen, HyoukJoong Lee, Jiquan Ngiam, Quoc V Le, Yonghui Wu, et al., 2019. Gpipe: Efficient training of giant neural networks using pipeline parallelism. Advances in neural information processing systems, Vol. 32 (2019)."},{"key":"e_1_3_2_1_20_1","volume-title":"PAT: a new algorithm for all-gather and reduce-scatter operations at scale. arXiv preprint arXiv:2506.20252","author":"Jeaugey Sylvain","year":"2025","unstructured":"Sylvain Jeaugey. 2025. PAT: a new algorithm for all-gather and reduce-scatter operations at scale. arXiv preprint arXiv:2506.20252 (2025)."},{"key":"e_1_3_2_1_21_1","unstructured":"Albert Q. Jiang Alexandre Sablayrolles Antoine Roux Arthur Mensch Blanche Savary Chris Bamford Devendra Singh Chaplot Diego de las Casas Emma Bou Hanna Florian Bressand Gianna Lengyel Guillaume Bour Guillaume Lample L\u00e9lio Renard Lavaud Lucile Saulnier Marie-Anne Lachaux Pierre Stock Sandeep Subramanian Sophia Yang Szymon Antoniak Teven Le Scao Th\u00e9ophile Gervet Thibaut Lavril Thomas Wang Timoth\u00e9e Lacroix and William El Sayed. 2024b. Mixtral of Experts. arXiv:2401.04088 [cs.LG] https:\/\/arxiv.org\/abs\/2401.04088"},{"key":"e_1_3_2_1_22_1","first-page":"745","volume-title":"21st USENIX Symposium on Networked Systems Design and Implementation (NSDI 24)","author":"Jiang Ziheng","year":"2024","unstructured":"Ziheng Jiang, Haibin Lin, Yinmin Zhong, Qi Huang, Yangrui Chen, Zhi Zhang, Yanghua Peng, Xiang Li, Cong Xie, Shibiao Nong, et al., 2024a. : Scaling large language model training to more than 10,000 . In 21st USENIX Symposium on Networked Systems Design and Implementation (NSDI 24). 745-760."},{"key":"e_1_3_2_1_23_1","first-page":"31","volume-title":"20th USENIX Symposium on Networked Systems Design and Implementation (NSDI 23)","author":"Kong Xinhao","year":"2023","unstructured":"Xinhao Kong, Jingrong Chen, Wei Bai, Yechen Xu, Mahmoud Elhaddad, Shachar Raindel, Jitendra Padhye, Alvin R Lebeck, and Danyang Zhuo. 2023. Understanding microarchitecture resources for performance isolation. In 20th USENIX Symposium on Networked Systems Design and Implementation (NSDI 23). 31-48."},{"key":"e_1_3_2_1_24_1","first-page":"287","volume-title":"19th USENIX Symposium on Networked Systems Design and Implementation (NSDI 22)","author":"Kong Xinhao","year":"2022","unstructured":"Xinhao Kong, Yibo Zhu, Huaping Zhou, Zhuo Jiang, Jianxi Ye, Chuanxiong Guo, and Danyang Zhuo. 2022. Collie: Finding performance anomalies in subsystems. In 19th USENIX Symposium on Networked Systems Design and Implementation (NSDI 22). 287-305."},{"key":"e_1_3_2_1_25_1","volume-title":"Gshard: Scaling giant models with conditional computation and automatic sharding. arXiv preprint arXiv:2006.16668","author":"Lepikhin Dmitry","year":"2020","unstructured":"Dmitry Lepikhin, HyoukJoong Lee, Yuanzhong Xu, Dehao Chen, Orhan Firat, Yanping Huang, Maxim Krikun, Noam Shazeer, and Zhifeng Chen. 2020. Gshard: Scaling giant models with conditional computation and automatic sharding. arXiv preprint arXiv:2006.16668 (2020)."},{"key":"e_1_3_2_1_26_1","volume-title":"Yan Zhuang, Fei Feng, Lingbo Tang, Zheng Cao, Ming Zhang, Frank Kelly, Mohammad Alizadeh, et al.","author":"Li Yuliang","year":"2019","unstructured":"Yuliang Li, Rui Miao, Hongqiang Harry Liu, Yan Zhuang, Fei Feng, Lingbo Tang, Zheng Cao, Ming Zhang, Frank Kelly, Mohammad Alizadeh, et al., 2019. HPCC: High precision congestion control. In Proceedings of the ACM special interest group on data communication. 44-58."},{"key":"e_1_3_2_1_27_1","unstructured":"Aixin Liu Bei Feng Bing Xue Bingxuan Wang Bochao Wu Chengda Lu Chenggang Zhao Chengqi Deng Chenyu Zhang Chong Ruan et al. 2024a. Deepseek-v3 technical report. arXiv preprint arXiv:2412.19437 (2024)."},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1145\/3651890.3672264"},{"key":"e_1_3_2_1_29_1","first-page":"15","volume-title":"20th USENIX Symposium on Networked Systems Design and Implementation (NSDI 23)","author":"Liu Kefei","year":"2023","unstructured":"Kefei Liu, Zhuo Jiang, Jiao Zhang, Haoran Wei, Xiaolong Zhong, Lizhuang Tan, Tian Pan, and Tao Huang. 2023. Hostping: Diagnosing intra-host network bottlenecks in servers. In 20th USENIX Symposium on Networked Systems Design and Implementation (NSDI 23). 15-29."},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1145\/3718958.3750521"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1109\/26.103043"},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1145\/2934872.2934879"},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.1145\/3341301.3359646"},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.1145\/3458817.3476209"},{"key":"e_1_3_2_1_35_1","unstructured":"NVIDIA. [n.d.]. NCCL GitHub Issue #578: Ring AllReduce performance discrepancy. https:\/\/github.com\/NVIDIA\/nccl\/issues\/578."},{"key":"e_1_3_2_1_36_1","unstructured":"Nvidia. 2019. NCCL Tree. https:\/\/developer.nvidia.com\/blog\/massively-scale-deep-learning-training-nccl-2-4\/."},{"key":"e_1_3_2_1_37_1","unstructured":"Nvidia. 2025a. Data-Path Accelerator Subsystem. https:\/\/docs.nvidia.com\/doca\/sdk\/dpasubsystem\/index.html."},{"key":"e_1_3_2_1_38_1","unstructured":"Nvidia. 2025b. DOCA Documentation v2.9.0. https:\/\/docs.nvidia.com\/doca\/archive\/2-9-0\/index.html."},{"key":"e_1_3_2_1_39_1","unstructured":"Nvidia. 2025c. DOCA PCC. https:\/\/docs.nvidia.com\/doca\/sdk\/docapcc\/index.html."},{"key":"e_1_3_2_1_40_1","unstructured":"Nvidia. 2025d. DPA L2 Reflector. https:\/\/docs.nvidia.com\/doca\/archive\/doca-v2-5-0\/nvidiadocadpal2reflectorapplicationguide\/index.html."},{"key":"e_1_3_2_1_41_1","unstructured":"Nvidia. 2025 e. NCCL. https:\/\/github.com\/nvidia\/nccl."},{"key":"e_1_3_2_1_42_1","volume-title":"2025 f","unstructured":"Nvidia. 2025 f. NCCL Group calls. https:\/\/docs.nvidia.com\/deeplearning\/nccl\/user-guide\/docs\/usage\/groups.html."},{"key":"e_1_3_2_1_43_1","unstructured":"Nvidia. 2025 g. NCCL Point-to-point communication. https:\/\/docs.nvidia.com\/deeplearning\/nccl\/user-guide\/docs\/usage\/p2p.html."},{"key":"e_1_3_2_1_44_1","unstructured":"Nvidia. 2025 h. NCCL Profiler Plugin. https:\/\/github.com\/NVIDIA\/nccl\/blob\/master\/ext-profiler\/README.md."},{"key":"e_1_3_2_1_45_1","unstructured":"Pytorch. 2025. Pytorch NCCL backend. https:\/\/github.com\/pytorch\/pytorch\/blob\/viable\/strict\/1766582867\/torch\/csrc\/cuda\/nccl.cpp."},{"key":"e_1_3_2_1_46_1","doi-asserted-by":"publisher","DOI":"10.1145\/3651890.3672265"},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.1109\/SC41405.2020.00024"},{"key":"e_1_3_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.1145\/3394486.3406703"},{"key":"e_1_3_2_1_49_1","volume-title":"Megatron-lm: Training multi-billion parameter language models using model parallelism. arXiv preprint arXiv:1909.08053","author":"Shoeybi Mohammad","year":"2019","unstructured":"Mohammad Shoeybi, Mostofa Patwary, Raul Puri, Patrick LeGresley, Jared Casper, and Bryan Catanzaro. 2019. Megatron-lm: Training multi-billion parameter language models using model parallelism. arXiv preprint arXiv:1909.08053 (2019)."},{"key":"e_1_3_2_1_50_1","doi-asserted-by":"publisher","DOI":"10.1145\/3603269.3604847"},{"key":"e_1_3_2_1_51_1","unstructured":"Kimi Team Angang Du Bofei Gao Bowei Xing Changjiu Jiang Cheng Chen Cheng Li Chenjun Xiao Chenzhuang Du Chonghua Liao et al. 2025. Kimi k1. 5: Scaling reinforcement learning with llms. arXiv preprint arXiv:2501.12599 (2025)."},{"key":"e_1_3_2_1_52_1","volume-title":"Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971","author":"Touvron Hugo","year":"2023","unstructured":"Hugo Touvron, Thibaut Lavril, Gautier Izacard, Xavier Martinet, Marie-Anne Lachaux, Timoth\u00e9e Lacroix, Baptiste Rozi\u00e8re, Naman Goyal, Eric Hambro, Faisal Azhar, et al., 2023. Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971 (2023)."},{"key":"e_1_3_2_1_53_1","volume-title":"22nd USENIX Symposium on Networked Systems Design and Implementation (NSDI 25)","author":"Wang Xizheng","unstructured":"Xizheng Wang, Qingxu Li, Yichi Xu, Gang Lu, Dan Li, Li Chen, Heyang Zhou, Linkang Zheng, Sen Zhang, Yikai Zhu, Yang Liu, Pengcheng Zhang, Kun Qian, Kunling He, Jiaqi Gao, Ennan Zhai, Dennis Cai, and Binzhang Fu. 2025. SimAI: Unifying Architecture Design and Performance Tuning for Large-Scale Large Language Model Training with Scalability and Precision. In 22nd USENIX Symposium on Networked Systems Design and Implementation (NSDI 25). USENIX Association, Philadelphia, PA, 541-558. https:\/\/www.usenix.org\/conference\/nsdi25\/presentation\/wang-xizheng-simai"},{"key":"e_1_3_2_1_54_1","first-page":"731","volume-title":"2025 USENIX Annual Technical Conference (USENIX ATC 25)","author":"Wu Tianyuan","year":"2025","unstructured":"Tianyuan Wu, Wei Wang, Yinghao Yu, Siran Yang, Wenchao Wu, Qinkai Duan, Guodong Yang, Jiamang Wang, Lin Qu, and Liping Zhang. 2025. : Hunting in Training at Scale. In 2025 USENIX Annual Technical Conference (USENIX ATC 25). 731-747."},{"key":"e_1_3_2_1_55_1","first-page":"835","volume-title":"2024 USENIX Annual Technical Conference (USENIX ATC 24)","author":"Xiong Yifan","year":"2024","unstructured":"Yifan Xiong, Yuting Jiang, Ziyue Yang, Lei Qu, Guoshuai Zhao, Shuguang Liu, Dong Zhong, Boris Pinzur, Jie Zhang, Yang Wang, et al., 2024. : Improving Cloud Infrastructure Reliability with Proactive Validation. In 2024 USENIX Annual Technical Conference (USENIX ATC 24). 835-850."},{"key":"e_1_3_2_1_56_1","doi-asserted-by":"publisher","DOI":"10.1145\/3230543.3230544"},{"key":"e_1_3_2_1_57_1","volume-title":"Holmes: Localizing Irregularities in LLM Training with Mega-scale GPU Clusters. In 22nd USENIX Symposium on Networked Systems Design and Implementation (NSDI 25)","author":"Yao Zhiyi","year":"2025","unstructured":"Zhiyi Yao, Pengbo Hu, CongCong Miao, Xuya Jia, Zuning Liang, Yuedong Xu, Chunzhia He, Hao Lu, Mingzhuo Chen, Xiang Li, Zekun He, Yachen Wang, Xianneng Zou, and Junchen Jiang. 2025. Holmes: Localizing Irregularities in LLM Training with Mega-scale GPU Clusters. In 22nd USENIX Symposium on Networked Systems Design and Implementation (NSDI 25)."},{"key":"e_1_3_2_1_58_1","doi-asserted-by":"publisher","DOI":"10.1145\/3603269.3604837"},{"key":"e_1_3_2_1_59_1","unstructured":"Chenggang Zhao Shangyan Zhou Liyue Zhang Chengqi Deng Zhean Xu Yuxuan Liu Kuai Yu Jiashi Li and Liang Zhao. 2025. DeepEP: an efficient expert-parallel communication library. https:\/\/github.com\/deepseek-ai\/DeepEP."},{"key":"e_1_3_2_1_60_1","doi-asserted-by":"publisher","DOI":"10.1145\/3651890.3672236"},{"key":"e_1_3_2_1_61_1","first-page":"193","volume-title":"18th USENIX Symposium on Operating Systems Design and Implementation (OSDI 24)","author":"Zhong Yinmin","year":"2024","unstructured":"Yinmin Zhong, Shengyu Liu, Junda Chen, Jianbo Hu, Yibo Zhu, Xuanzhe Liu, Xin Jin, and Hao Zhang. 2024. : Disaggregating prefill and decoding for goodput-optimized large language model serving. In 18th USENIX Symposium on Operating Systems Design and Implementation (OSDI 24). 193-210."}],"event":{"name":"ASPLOS '26: 31st ACM International Conference on Architectural Support for Programming Languages and Operating Systems","location":"Pittsburgh PA USA","sponsor":["SIGOPS ACM Special Interest Group on Operating Systems","SIGPLAN ACM Special Interest Group on Programming Languages","SIGARCH ACM Special Interest Group on Computer Architecture","SIGBED ACM Special Interest Group on Embedded Systems"]},"container-title":["Proceedings of the 31st ACM International Conference on Architectural Support for Programming Languages and Operating Systems, Volume 2"],"original-title":[],"deposited":{"date-parts":[[2026,3,15]],"date-time":"2026-03-15T14:08:38Z","timestamp":1773583718000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3779212.3790163"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,3,22]]},"references-count":61,"alternative-id":["10.1145\/3779212.3790163","10.1145\/3779212"],"URL":"https:\/\/doi.org\/10.1145\/3779212.3790163","relation":{},"subject":[],"published":{"date-parts":[[2026,3,22]]},"assertion":[{"value":"2026-03-22","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}