{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T05:05:56Z","timestamp":1750309556045,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":59,"publisher":"ACM","license":[{"start":{"date-parts":[[2025,4,22]],"date-time":"2025-04-22T00:00:00Z","timestamp":1745280000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"Young Elite Scientist Sponsorship Program by CAST (2023QNRC001)","award":["YESS20230367"],"award-info":[{"award-number":["YESS20230367"]}]},{"DOI":"10.13039\/https:\/\/doi.org\/10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62306330, 62106275, 62025208, 62421002"],"award-info":[{"award-number":["62306330, 62106275, 62025208, 62421002"]}],"id":[{"id":"10.13039\/https:\/\/doi.org\/10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,4,22]]},"DOI":"10.1145\/3696410.3714880","type":"proceedings-article","created":{"date-parts":[[2025,4,22]],"date-time":"2025-04-22T22:47:11Z","timestamp":1745362031000},"page":"2378-2390","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["Scenario-independent Uncertainty Estimation for LLM-based Question Answering via Factor Analysis"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-2224-6197","authenticated-orcid":false,"given":"Zhihua","family":"Wen","sequence":"first","affiliation":[{"name":"National University of Defense Technology, Changsha, Hunan, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0006-1242-0661","authenticated-orcid":false,"given":"Zhizhao","family":"Liu","sequence":"additional","affiliation":[{"name":"National University of Defense Technology, Changsha, Hunan, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8906-5198","authenticated-orcid":false,"given":"Zhiliang","family":"Tian","sequence":"additional","affiliation":[{"name":"National University of Defense Technology, Changsha, Hunan, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0004-2165-8247","authenticated-orcid":false,"given":"Shilong","family":"Pan","sequence":"additional","affiliation":[{"name":"National University of Defense Technology, Changsha, Hunan, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4819-373X","authenticated-orcid":false,"given":"Zhen","family":"Huang","sequence":"additional","affiliation":[{"name":"National University of Defense Technology, Changsha, Hunan, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9743-2034","authenticated-orcid":false,"given":"Dongsheng","family":"Li","sequence":"additional","affiliation":[{"name":"National University of Defense Technology, Changsha, Hunan, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7111-1849","authenticated-orcid":false,"given":"Minlie","family":"Huang","sequence":"additional","affiliation":[{"name":"Tsinghua University, Beijing, China"}]}],"member":"320","published-online":{"date-parts":[[2025,4,22]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1098\/rspa.1937.0109"},{"key":"e_1_3_2_1_4_1","volume-title":"PCANet: A Simple Deep Learning Baseline for Image Classification? IEEE transactions on image processing","author":"Chan Tsung-Han","year":"2015","unstructured":"Tsung-Han Chan, Kui Jia, Shenghua Gao, Jiwen Lu, Zinan Zeng, and Yi Ma. 2015. PCANet: A Simple Deep Learning Baseline for Image Classification? IEEE transactions on image processing, Vol. 24, 12 (2015), 5017--5032."},{"key":"e_1_3_2_1_5_1","volume-title":"INSIDE: LLMs' Internal States Retain the Power of Hallucination Detection. In ICLR. https:\/\/openreview.net\/forum?id=Zj12nzlQbz","author":"Chen Chao","year":"2024","unstructured":"Chao Chen, Kai Liu, Ze Chen, Yi Gu, Yue Wu, Mingyuan Tao, Zhihang Fu, and Jieping Ye. 2024b. INSIDE: LLMs' Internal States Retain the Power of Hallucination Detection. In ICLR. https:\/\/openreview.net\/forum?id=Zj12nzlQbz"},{"key":"e_1_3_2_1_6_1","volume-title":"The Twelfth International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=ccxD4mtkTU","author":"Chen Canyu","year":"2024","unstructured":"Canyu Chen and Kai Shu. 2024. Can LLM-Generated Misinformation Be Detected?. In The Twelfth International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=ccxD4mtkTU"},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","unstructured":"Jifan Chen Grace Kim Aniruddh Sriram Greg Durrett and Eunsol Choi. 2024a. Complex Claim Verification with Evidence Retrieved in the Wild. In NAACL Kevin Duh Helena Gomez and Steven Bethard (Eds.). Association for Computational Linguistics Mexico City Mexico. hrefhttps:\/\/doi.org\/10.18653\/v1\/2024.naacl-long.196doi:nolinkurl10.18653\/v1\/2024.naacl-long.196","DOI":"10.18653\/v1"},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1145\/3583780.3614905"},{"key":"e_1_3_2_1_9_1","volume-title":"Hongzhi Zhang, Fuzheng Zhang, Di Zhang, Kun Gai, and Ji-Rong Wen.","author":"Cheng Xiaoxue","year":"2024","unstructured":"Xiaoxue Cheng, Junyi Li, Wayne Xin Zhao, Hongzhi Zhang, Fuzheng Zhang, Di Zhang, Kun Gai, and Ji-Rong Wen. 2024. Small Agent Can Also Rock! Empowering Small Language Models as Hallucination Detector. https:\/\/arxiv.org\/abs\/2406.11277"},{"key":"e_1_3_2_1_10_1","unstructured":"I-Chun Chern Steffi Chern Shiqi Chen Weizhe Yuan Kehua Feng Chunting Zhou Junxian He Graham Neubig Pengfei Liu et al. 2023. FacTool: Factuality Detection in Generative AI--A Tool Augmented Framework for Multi-Task and Multi-Domain Scenarios. arXiv preprint arXiv:2307.13528 (2023)."},{"key":"e_1_3_2_1_11_1","volume-title":"Lookback Lens: Detecting and Mitigating Contextual Hallucinations in Large Language Models Using Only Attention Maps. arXiv preprint arXiv:2407.07071","author":"Chuang Yung-Sung","year":"2024","unstructured":"Yung-Sung Chuang, Linlu Qiu, Cheng-Yu Hsieh, Ranjay Krishna, Yoon Kim, and James Glass. 2024. Lookback Lens: Detecting and Mitigating Contextual Hallucinations in Large Language Models Using Only Attention Maps. arXiv preprint arXiv:2407.07071 (2024)."},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_1_13_1","unstructured":"Longchao Da Tiejin Chen Lu Cheng and Hua Wei. 2024. LLM Uncertainty Quantification through Directional Entailment Graph and Claim Level Response Augmentation. arxiv: 2407.00994 [cs.CL] https:\/\/arxiv.org\/abs\/2407.00994"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"crossref","unstructured":"Jinhao Duan Hao Cheng Shiqi Wang Alex Zavalny Chenan Wang Renjing Xu Bhavya Kailkhura and Kaidi Xu. 2024. Shifting Attention to Relevance: Towards the Predictive Uncertainty Quantification of Free-Form Large Language Models. In ACL. 5050--5063.","DOI":"10.18653\/v1\/2024.acl-long.276"},{"key":"e_1_3_2_1_15_1","unstructured":"Abhimanyu Dubey Abhinav Jauhri Abhinav Pandey and Abhishek Kadian et al. 2024. The Llama 3 Herd of Models. arxiv: 2407.21783 [cs.AI] https:\/\/arxiv.org\/abs\/2407.21783"},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1162\/tacl_a_00330"},{"key":"e_1_3_2_1_17_1","volume-title":"Zero-Shot Text Classification with Self-Training. In Conference on Empirical Methods in Natural Language Processing.","author":"Gera Ariel","year":"2022","unstructured":"Ariel Gera, Alon Halfon, Eyal Shnarch, Yotam Perlitz, Liat Ein-Dor, and Noam Slonim. 2022. Zero-Shot Text Classification with Self-Training. In Conference on Empirical Methods in Natural Language Processing."},{"key":"e_1_3_2_1_18_1","volume-title":"Yujiu Yang, Nan Duan, and Weizhu Chen.","author":"Gou Zhibin","year":"2024","unstructured":"Zhibin Gou, Zhihong Shao, Yeyun Gong, yelong shen, Yujiu Yang, Nan Duan, and Weizhu Chen. 2024. CRITIC: Large Language Models Can Self-Correct with Tool-Interactive Critiquing. In ICLR. https:\/\/openreview.net\/forum?id=Sx038qxjek"},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1007\/BF02289162"},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1148\/radiology.143.1.7063747"},{"key":"e_1_3_2_1_21_1","volume-title":"DeBERTa: Decoding-Enhanced BERT with Disentangled Attention. In 2021 International Conference on Learning Representations. Under review.","author":"He Pengcheng","year":"2021","unstructured":"Pengcheng He, Xiaodong Liu, Jianfeng Gao, and Wei Chen. 2021. DeBERTa: Decoding-Enhanced BERT with Disentangled Attention. In 2021 International Conference on Learning Representations. Under review."},{"key":"e_1_3_2_1_22_1","unstructured":"Lei Huang Weijiang Yu Weitao Ma Weihong Zhong Zhangyin Feng Haotian Wang Qianglong Chen Weihua Peng Xiaocheng Feng Bing Qin and Ting Liu. 2023. A Survey on Hallucination in Large Language Models: Principles Taxonomy Challenges and Open Questions. arxiv: 2311.05232 [cs.CL]"},{"key":"e_1_3_2_1_23_1","volume-title":"Retrieving supporting evidence for llms generated answers. arXiv preprint arXiv:2306.13781","author":"Huo Siqing","year":"2023","unstructured":"Siqing Huo, Negar Arabzadeh, and Charles LA Clarke. 2023. Retrieving supporting evidence for llms generated answers. arXiv preprint arXiv:2306.13781 (2023)."},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1121\/1.2016299"},{"key":"e_1_3_2_1_25_1","unstructured":"Ziwei Ji Delong Chen Etsuko Ishii Samuel Cahyawijaya Yejin Bang Bryan Wilie and Pascale Fung. 2024a. LLM Internal States Reveal Hallucination Risk Faced With a Query. arxiv: 2407.03282 [cs.CL]"},{"key":"e_1_3_2_1_26_1","volume-title":"ANAH: Analytical Annotation of Hallucinations in Large Language Models. In Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Lun-Wei Ku","author":"Ji Ziwei","year":"2024","unstructured":"Ziwei Ji, Yuzhe Gu, Wenwei Zhang, Chengqi Lyu, Dahua Lin, and Kai Chen. 2024b. ANAH: Analytical Annotation of Hallucinations in Large Language Models. In Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Lun-Wei Ku, Andre Martins, and Vivek Srikumar (Eds.). Association for Computational Linguistics, Bangkok, Thailand. https:\/\/aclanthology.org\/2024.acl-long.442"},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P17-1147"},{"key":"e_1_3_2_1_29_1","unstructured":"Saurav Kadavath Tom Conerly Amanda Askell Tom Henighan Dawn Drain Ethan Perez Nicholas Schiefer Zac Hatfield-Dodds Nova DasSarma Eli Tran-Johnson et al. 2022a. Language models (mostly) know what they know. arXiv preprint arXiv:2207.05221 (2022)."},{"key":"e_1_3_2_1_30_1","unstructured":"Saurav Kadavath Tom Conerly Amanda Askell Tom Henighan Dawn Drain Ethan Perez Nicholas Schiefer Zac Hatfield-Dodds Nova DasSarma Eli Tran-Johnson Scott Johnston Sheer El-Showk Andy Jones Nelson Elhage Tristan Hume Anna Chen Yuntao Bai Sam Bowman Stanislav Fort Deep Ganguli Danny Hernandez Josh Jacobson Jackson Kernion Shauna Kravec Liane Lovitt Kamal Ndousse Catherine Olsson Sam Ringer Dario Amodei Tom Brown Jack Clark Nicholas Joseph Ben Mann Sam McCandlish Chris Olah and Jared Kaplan. 2022b. Language Models (Mostly) Know What They Know. arxiv: 2207.05221 [cs.CL] https:\/\/arxiv.org\/abs\/2207.05221"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1177\/001316447403400115"},{"key":"e_1_3_2_1_32_1","unstructured":"Lorenz Kuhn Yarin Gal and Sebastian Farquhar. 2023. Semantic Uncertainty: Linguistic Invariances for Uncertainty Estimation in Natural Language Generation. In The Eleventh International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=VD-AYtP0dve"},{"key":"e_1_3_2_1_33_1","volume-title":"Principal component analysis. Nature methods","author":"Lever Jake","year":"2017","unstructured":"Jake Lever, Martin Krzywinski, and Naomi Altman. 2017. Principal component analysis. Nature methods, Vol. 14, 7 (2017)."},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_1_35_1","volume-title":"Generating with Confidence: Uncertainty Quantification for Black-box Large Language Models. Transactions on Machine Learning Research","author":"Lin Zhen","year":"2024","unstructured":"Zhen Lin, Shubhendu Trivedi, and Jimeng Sun. 2024. Generating with Confidence: Uncertainty Quantification for Black-box Large Language Models. Transactions on Machine Learning Research (2024). https:\/\/openreview.net\/forum?id=DWkJCSxKU5"},{"key":"e_1_3_2_1_36_1","volume-title":"Yu Pan and Guanting Chen","author":"Linyu Liu Xiaocheng Li","year":"2024","unstructured":"Xiaocheng Li Linyu Liu, Yu Pan and Guanting Chen. 2024. Uncertainty Estimation and Quantification for LLMs: A Simple Supervised Approach. arxiv: 2404.15993 [cs.LG] https:\/\/arxiv.org\/abs\/2404.15993"},{"key":"e_1_3_2_1_37_1","volume-title":"Proceedings of 5-th Berkeley Symposium on Mathematical Statistics and Probability\/University of California Press.","author":"Macqueen J","year":"1967","unstructured":"J Macqueen. 1967. Some methods for classification and analysis of multivariate observations. In Proceedings of 5-th Berkeley Symposium on Mathematical Statistics and Probability\/University of California Press."},{"key":"e_1_3_2_1_38_1","unstructured":"Andrey Malinin and Mark Gales. 2021. Uncertainty Estimation in Autoregressive Structured Prediction. In ICLR. https:\/\/openreview.net\/forum?id=jN5y-zb5Q7m"},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_1_40_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_1_41_1","unstructured":"OpenAI. 2024. Hello GPT-4o. https:\/\/openai.com\/index\/hello-gpt-4o\/ Accessed: 2024--10-07."},{"key":"e_1_3_2_1_42_1","unstructured":"Benjamin Plaut Khanh Nguyen and Tu Trinh. 2024. Softmax Probabilities (Mostly) Predict Large Language Model Correctness on Multiple-Choice Q&A. arxiv: 2402.13213 [cs.CL] https:\/\/arxiv.org\/abs\/2402.13213"},{"key":"e_1_3_2_1_43_1","doi-asserted-by":"publisher","DOI":"10.1016\/0377-0427(87)90125--7nolinkurl10.1016"},{"key":"e_1_3_2_1_44_1","volume-title":"Hermina Petric Maretic, and Juba Nait Saada","author":"Sansford Hannah","year":"2024","unstructured":"Hannah Sansford, Nicholas Richardson, Hermina Petric Maretic, and Juba Nait Saada. 2024. Grapheval: A knowledge-graph based llm hallucination evaluation framework. arXiv preprint arXiv:2407.10793 (2024)."},{"key":"e_1_3_2_1_45_1","volume-title":"Simple Entity-Centric Questions Challenge Dense Retrievers. In 2021 Conference on Empirical Methods in Natural Language Processing, EMNLP","author":"Sciavolino Christopher","year":"2021","unstructured":"Christopher Sciavolino, Zexuan Zhong, Jinhyuk Lee, and Danqi Chen. 2021. Simple Entity-Centric Questions Challenge Dense Retrievers. In 2021 Conference on Empirical Methods in Natural Language Processing, EMNLP 2021. Association for Computational Linguistics (ACL), 6138--6148."},{"key":"e_1_3_2_1_46_1","volume-title":"Jannik Kossen and Yarin Gal","author":"Sebastian Farquhar Lorenz Kuhn","year":"2024","unstructured":"Lorenz Kuhn Sebastian Farquhar, Jannik Kossen and Yarin Gal. 2024. Detecting hallucinations in large language models using semantic entropy. In Nature. Nature. https:\/\/www.nature.com\/articles\/s41586-024-07421-0#citeas"},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.2307\/1412107"},{"key":"e_1_3_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.findings-acl.854"},{"key":"e_1_3_2_1_49_1","unstructured":"Xiaoxi Sun Jinpeng Li Yan Zhong Dongyan Zhao and Rui Yan. 2024. Towards Detecting LLMs Hallucination via Markov Chain-based Multi-agent Debate Framework. arxiv: 2406.03075 [cs.CL] https:\/\/arxiv.org\/abs\/2406.03075"},{"key":"e_1_3_2_1_50_1","volume-title":"Conference on Neural Information Processing Systems.","author":"Tenenbaum JB","year":"1998","unstructured":"JB Tenenbaum. 1998. Mapping a Manifold of Perceptual Observations.. In Conference on Neural Information Processing Systems."},{"key":"e_1_3_2_1_51_1","doi-asserted-by":"publisher","DOI":"10.1007\/BF02288916"},{"key":"e_1_3_2_1_52_1","volume-title":"Unconditional Truthfulness: Learning Conditional Dependency for Uncertainty Quantification of Large Language Models. arxiv: 2408.10692 [cs.CL] https:\/\/arxiv.org\/abs\/2408.10692","author":"Vazhentsev Artem","year":"2024","unstructured":"Artem Vazhentsev, Ekaterina Fadeeva, Rui Xing, Alexander Panchenko, Preslav Nakov, Timothy Baldwin, Maxim Panov, and Artem Shelmanov. 2024. Unconditional Truthfulness: Learning Conditional Dependency for Uncertainty Quantification of Large Language Models. arxiv: 2408.10692 [cs.CL] https:\/\/arxiv.org\/abs\/2408.10692"},{"key":"e_1_3_2_1_53_1","doi-asserted-by":"crossref","unstructured":"Fanqi Wan Xinting Huang Leyang Cui Xiaojun Quan Wei Bi and Shuming Shi. 2024. Knowledge Verification to Nip Hallucination in the Bud. https:\/\/arxiv.org\/abs\/2401.10768","DOI":"10.18653\/v1\/2024.emnlp-main.152"},{"key":"e_1_3_2_1_54_1","doi-asserted-by":"publisher","DOI":"10.1162\/tacl_a_00021"},{"key":"e_1_3_2_1_55_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.findings-emnlp.262"},{"key":"e_1_3_2_1_56_1","unstructured":"Miao Xiong Zhiyuan Hu Xinyang Lu Yifei Li Jie Fu Junxian He and Bryan Hooi. 2024. Can LLMs Express Their Uncertainty? An Empirical Evaluation of Confidence Elicitation in LLMs. arxiv: 2306.13063 [cs.CL] https:\/\/arxiv.org\/abs\/2306.13063"},{"key":"e_1_3_2_1_57_1","doi-asserted-by":"crossref","unstructured":"Dongxu Zhang Varun Gangal Barrett Lattimer and Yi Yang. 2024. Enhancing Hallucination Detection through Perturbation-Based Synthetic Data Generation in System Responses. In Findings of ACL Lun-Wei Ku Andre Martins and Vivek Srikumar (Eds.). Bangkok Thailand and virtual meeting. https:\/\/aclanthology.org\/2024.findings-acl.789","DOI":"10.18653\/v1\/2024.findings-acl.789"},{"key":"e_1_3_2_1_58_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_1_59_1","unstructured":"Xiao Zhang and Ji Wu. 2024. Dissecting learning and forgetting in language model finetuning. In ICLR. https:\/\/openreview.net\/forum?id=tmsqb6WpLz"},{"key":"e_1_3_2_1_60_1","doi-asserted-by":"publisher","unstructured":"Yukun Zhao Lingyong Yan Weiwei Sun Guoliang Xing Chong Meng Shuaiqiang Wang Zhicong Cheng Zhaochun Ren and Dawei Yin. 2024. Knowing What LLMs DO NOT Know: A Simple Yet Effective Self-Detection Method. In NAACL Kevin Duh Helena Gomez and Steven Bethard (Eds.). Association for Computational Linguistics Mexico City Mexico. hrefhttps:\/\/doi.org\/10.18653\/v1\/2024.naacl-long.390doi:nolinkurl10.18653\/v1\/2024.naacl-long.390","DOI":"10.18653\/v1"}],"event":{"name":"WWW '25: The ACM Web Conference 2025","sponsor":["SIGWEB ACM Special Interest Group on Hypertext, Hypermedia, and Web"],"location":"Sydney NSW Australia","acronym":"WWW '25"},"container-title":["Proceedings of the ACM on Web Conference 2025"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3696410.3714880","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3696410.3714880","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T01:18:53Z","timestamp":1750295933000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3696410.3714880"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,4,22]]},"references-count":59,"alternative-id":["10.1145\/3696410.3714880","10.1145\/3696410"],"URL":"https:\/\/doi.org\/10.1145\/3696410.3714880","relation":{},"subject":[],"published":{"date-parts":[[2025,4,22]]},"assertion":[{"value":"2025-04-22","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}