{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,10]],"date-time":"2025-12-10T04:09:17Z","timestamp":1765339757839,"version":"3.46.0"},"publisher-location":"New York, NY, USA","reference-count":52,"publisher":"ACM","funder":[{"name":"State Key Laboratory of Industrial Control Technology","award":["No. ICT2024A09"],"award-info":[{"award-number":["No. ICT2024A09"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,10,27]]},"DOI":"10.1145\/3746027.3755832","type":"proceedings-article","created":{"date-parts":[[2025,10,25]],"date-time":"2025-10-25T05:56:43Z","timestamp":1761371803000},"page":"5237-5246","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["CogDDN: A Cognitive Demand-Driven Navigation with Decision Optimization and Dual-Process Thinking"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0002-9397-8341","authenticated-orcid":false,"given":"Yuehao","family":"Huang","sequence":"first","affiliation":[{"name":"Zhejiang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7910-810X","authenticated-orcid":false,"given":"Liang","family":"Liu","sequence":"additional","affiliation":[{"name":"vivo AI Lab, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0001-5344-6079","authenticated-orcid":false,"given":"Shuangming","family":"Lei","sequence":"additional","affiliation":[{"name":"Zhejiang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8135-9012","authenticated-orcid":false,"given":"Yukai","family":"Ma","sequence":"additional","affiliation":[{"name":"Zhejiang University, Hang zhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0001-5450-1383","authenticated-orcid":false,"given":"Hao","family":"Su","sequence":"additional","affiliation":[{"name":"Zhejiang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3849-2736","authenticated-orcid":false,"given":"Jianbiao","family":"Mei","sequence":"additional","affiliation":[{"name":"Zhejiang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0004-6567-5815","authenticated-orcid":false,"given":"Pengxiang","family":"Zhao","sequence":"additional","affiliation":[{"name":"Zhejiang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0003-5918-3235","authenticated-orcid":false,"given":"Yaqing","family":"Gu","sequence":"additional","affiliation":[{"name":"Zhejiang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4822-8939","authenticated-orcid":false,"given":"Yong","family":"Liu","sequence":"additional","affiliation":[{"name":"Zhejiang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8545-9464","authenticated-orcid":false,"given":"Jiajun","family":"Lv","sequence":"additional","affiliation":[{"name":"Zhejiang University, Hangzhou, China"}]}],"member":"320","published-online":{"date-parts":[[2025,10,27]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"Alexey Dosovitskiy, Saurabh Gupta, Vladlen Koltun, Jana Kosecka, Jitendra Malik, Roozbeh Mottaghi, Manolis Savva, et al.","author":"Anderson Peter","year":"2018","unstructured":"Peter Anderson, Angel Chang, Devendra Singh Chaplot, Alexey Dosovitskiy, Saurabh Gupta, Vladlen Koltun, Jana Kosecka, Jitendra Malik, Roozbeh Mottaghi, Manolis Savva, et al., 2018. On evaluation of embodied navigation agents. arXiv preprint arXiv:1807.06757 (2018)."},{"key":"e_1_3_2_1_2_1","volume-title":"Yuzhen Liu and Ming Cao","author":"Hamidreza Kasaei Tingguang Li Lei Han","year":"2024","unstructured":"Lei Han Hamidreza Kasaei Tingguang Li Bangguo Yu, Yuzhen Liu and Ming Cao. 2024. VLN-Game: Vision-Language Equilibrium Search for Zero-Shot Semantic Navigation. arXiv:2411.11609 (2024)."},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA57147.2024.10610499"},{"key":"e_1_3_2_1_4_1","volume-title":"CL-CoTNav: Closed-Loop Hierarchical Chain-of-Thought for Zero-Shot Object-Goal Navigation with Vision-Language Models. arXiv preprint arXiv:2504.09000","author":"Cai Yuxin","year":"2025","unstructured":"Yuxin Cai, Xiangkun He, Maonan Wang, Hongliang Guo, Wei-Yun Yau, and Chen Lv. 2025. CL-CoTNav: Closed-Loop Hierarchical Chain-of-Thought for Zero-Shot Object-Goal Navigation with Vision-Language Models. arXiv preprint arXiv:2504.09000 (2025)."},{"key":"e_1_3_2_1_5_1","first-page":"4247","article-title":"Object goal navigation using goal-oriented semantic exploration","volume":"33","author":"Chaplot Devendra Singh","year":"2020","unstructured":"Devendra Singh Chaplot, Dhiraj Prakashchand Gandhi, Abhinav Gupta, and Russ R Salakhutdinov. 2020. Object goal navigation using goal-oriented semantic exploration. Advances in Neural Information Processing Systems, Vol. 33 (2020), 4247-4258.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01501"},{"key":"e_1_3_2_1_7_1","first-page":"5982","article-title":"ProcTHOR: Large-Scale Embodied AI Using Procedural Generation","volume":"35","author":"Deitke Matt","year":"2022","unstructured":"Matt Deitke, Eli VanderBilt, Alvaro Herrasti, Luca Weihs, Kiana Ehsani, Jordi Salvador, Winson Han, Eric Kolve, Aniruddha Kembhavi, and Roozbeh Mottaghi. 2022. ProcTHOR: Large-Scale Embodied AI Using Procedural Generation. Advances in Neural Information Processing Systems, Vol. 35 (2022), 5982-5994.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1177\/1745691612460685"},{"key":"e_1_3_2_1_9_1","volume-title":"Clip on wheels: Zero-shot object navigation as object localization and exploration. arXiv preprint arXiv:2203.10421","author":"Gadre Samir Yitzhak","year":"2022","unstructured":"Samir Yitzhak Gadre, Mitchell Wortsman, Gabriel Ilharco, Ludwig Schmidt, and Shuran Song. 2022. Clip on wheels: Zero-shot object navigation as object localization and exploration. arXiv preprint arXiv:2203.10421, Vol. 3, 4 (2022), 7."},{"key":"e_1_3_2_1_10_1","volume-title":"Himanshu Gaurav Singh, and Antonio Loquercio","author":"Goetting Dylan","year":"2024","unstructured":"Dylan Goetting, Himanshu Gaurav Singh, and Antonio Loquercio. 2024. End-to-End Navigation with Vision Language Models: Transforming Spatial Reasoning into Question-Answering. arXiv preprint arXiv:2411.05755 (2024)."},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1109\/TSSC.1968.300136"},{"key":"e_1_3_2_1_12_1","volume-title":"General Scene Adaptation for Vision-and-Language Navigation. arXiv preprint arXiv:2501.17403","author":"Hong Haodong","year":"2025","unstructured":"Haodong Hong, Yanyuan Qiao, Sen Wang, Jiajun Liu, and Qi Wu. 2025. General Scene Adaptation for Vision-and-Language Navigation. arXiv preprint arXiv:2501.17403 (2025)."},{"key":"e_1_3_2_1_13_1","volume-title":"Fast and slow thinking","author":"Kahneman Daniel","year":"2011","unstructured":"Daniel Kahneman. 2011. Fast and slow thinking. Allen Lane and Penguin Books, New York (2011)."},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.01041"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1109\/CASE49997.2022.9926534"},{"key":"e_1_3_2_1_16_1","unstructured":"Eric Kolve Roozbeh Mottaghi Winson Han Eli VanderBilt Luca Weihs Alvaro Herrasti Matt Deitke Kiana Ehsani Daniel Gordon Yuke Zhu et al. 2017. Ai2-thor: An interactive 3d environment for visual ai. arXiv preprint arXiv:1712.05474 (2017)."},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52733.2024.01567"},{"key":"e_1_3_2_1_18_1","volume-title":"Learnact: Few-shot mobile gui agent with a unified demonstration benchmark. arXiv preprint arXiv:2504.13805","author":"Liu Guangyi","year":"2025","unstructured":"Guangyi Liu, Pengxiang Zhao, Liang Liu, Zhiming Chen, Yuxiang Chai, Shuai Ren, Hao Wang, Shibo He, and Wenchao Meng. 2025a. Learnact: Few-shot mobile gui agent with a unified demonstration benchmark. arXiv preprint arXiv:2504.13805 (2025)."},{"key":"e_1_3_2_1_19_1","unstructured":"Guangyi Liu Pengxiang Zhao Liang Liu Yaxuan Guo Han Xiao Weifeng Lin Yuxiang Chai Yue Han Shuai Ren Hao Wang et al. 2025b. Llm-powered gui agents in phone automation: Surveying progress and prospects. arXiv preprint arXiv:2504.19838 (2025)."},{"key":"e_1_3_2_1_20_1","volume-title":"Instructnav: Zero-shot system for generic instruction navigation in unexplored environment. arXiv preprint arXiv:2406.04882","author":"Long Yuxing","year":"2024","unstructured":"Yuxing Long, Wenzhe Cai, Hongcheng Wang, Guanqi Zhan, and Hao Dong. 2024. Instructnav: Zero-shot system for generic instruction navigation in unexplored environment. arXiv preprint arXiv:2406.04882 (2024)."},{"key":"e_1_3_2_1_21_1","volume-title":"UI-R1: Enhancing Efficient Action Prediction of GUI Agents by Reinforcement Learning. arXiv preprint arXiv:2503.21620","author":"Lu Zhengxi","year":"2025","unstructured":"Zhengxi Lu, Yuxiang Chai, Yaxuan Guo, Xi Yin, Liang Liu, Hao Wang, Han Xiao, Shuai Ren, Guanjing Xiong, and Hongsheng Li. 2025. UI-R1: Enhancing Efficient Action Prediction of GUI Agents by Reinforcement Learning. arXiv preprint arXiv:2503.21620 (2025)."},{"key":"e_1_3_2_1_22_1","volume-title":"Licrocc: Teach radar for accurate semantic occupancy prediction using lidar and camera","author":"Ma Yukai","year":"2024","unstructured":"Yukai Ma, Jianbiao Mei, Xuemeng Yang, Licheng Wen, Weihua Xu, Jiangning Zhang, Xingxing Zuo, Botian Shi, and Yong Liu. 2024. Licrocc: Teach radar for accurate semantic occupancy prediction using lidar and camera. IEEE Robotics and Automation Letters (2024)."},{"key":"e_1_3_2_1_23_1","volume-title":"LeapVAD: A Leap in Autonomous Driving via Cognitive Perception and Dual-Process Thinking. arXiv preprint arXiv:2501.08168","author":"Ma Yukai","year":"2025","unstructured":"Yukai Ma, Tiantian Wei, Naiting Zhong, Jianbiao Mei, Tao Hu, Licheng Wen, Xuemeng Yang, Botian Shi, and Yong Liu. 2025. LeapVAD: A Leap in Autonomous Driving via Cognitive Perception and Dual-Process Thinking. arXiv preprint arXiv:2501.08168 (2025)."},{"key":"e_1_3_2_1_24_1","unstructured":"Jianbiao Mei Yukai Ma Xuemeng Yang Licheng Wen Xinyu Cai Xin Li Daocheng Fu Bo Zhang Pinlong Cai Min Dou et al. 2024. Continuously learning adapting and improving: A dual-process approach to autonomous driving. arXiv preprint arXiv:2405.15324 (2024)."},{"key":"e_1_3_2_1_25_1","volume-title":"VLM-Vac: Enhancing Smart Vacuums through VLM Knowledge Distillation and Language-Guided Experience Replay. ArXiv","author":"Mirjalili Reihaneh","year":"2024","unstructured":"Reihaneh Mirjalili, Michael Krawez, Florian Walter, and Wolfram Burgard. 2024. VLM-Vac: Enhancing Smart Vacuums through VLM Knowledge Distillation and Language-Guided Experience Replay. ArXiv, Vol. abs\/2409.14096 (2024). https:\/\/api.semanticscholar.org\/CorpusID:272826708"},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.01281"},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"crossref","unstructured":"Dujun Nie Xianda Guo Yiqun Duan Ruijun Zhang and Long Chen. 2025. WMNav: Integrating Vision-Language Models into World Models for Object Goal Navigation. https:\/\/api.semanticscholar.org\/CorpusID:276776282","DOI":"10.1109\/IROS60139.2025.11246684"},{"key":"e_1_3_2_1_28_1","volume-title":"Conference on Robot Learning. PMLR, 517-528","author":"Pal Anwesan","year":"2021","unstructured":"Anwesan Pal, Yiding Qiu, and Henrik Christensen. 2021. Learning hierarchical relationships for object-goal navigation. In Conference on Robot Learning. PMLR, 517-528."},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01832"},{"key":"e_1_3_2_1_30_1","volume-title":"6th Annual Conference on Robot Learning. https:\/\/openreview.net\/forum?id=UW5A3SweAH","author":"Shah Dhruv","year":"2022","unstructured":"Dhruv Shah, Blazej Osinski, Brian Ichter, and Sergey Levine. 2022. LM-Nav: Robotic Navigation with Large Pre-Trained Models of Language, Vision, and Action. In 6th Annual Conference on Robot Learning. https:\/\/openreview.net\/forum?id=UW5A3SweAH"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1109\/LRA.2024.3511409"},{"key":"e_1_3_2_1_32_1","first-page":"16353","article-title":"Find what you want: Learning demand-conditioned object attribute space for demand-driven navigation","volume":"36","author":"Wang Hongcheng","year":"2023","unstructured":"Hongcheng Wang, Andy Guan Hong Chen, Xiaoqi Li, Mingdong Wu, and Hao Dong. 2023. Find what you want: Learning demand-conditioned object attribute space for demand-driven navigation. Advances in Neural Information Processing Systems, Vol. 36 (2023), 16353-16366.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_33_1","volume-title":"MO-DDN: A Coarse-to-Fine Attribute-based Exploration Agent for Multi-object Demand-driven Navigation. In The Thirty-eighth Annual Conference on Neural Information Processing Systems.","author":"Wang Hongcheng","year":"2024","unstructured":"Hongcheng Wang, Peiqi Liu, Wenzhe Cai, Mingdong Wu, Zhengyu Qian, and Hao Dong. 2024c. MO-DDN: A Coarse-to-Fine Attribute-based Exploration Agent for Multi-object Demand-driven Navigation. In The Thirty-eighth Annual Conference on Neural Information Processing Systems."},{"key":"e_1_3_2_1_34_1","unstructured":"Peng Wang Shuai Bai Sinan Tan Shijie Wang Zhihao Fan Jinze Bai Keqin Chen Xuejing Liu Jialin Wang Wenbin Ge et al. 2024a. Qwen2-vl: Enhancing vision-language model's perception of the world at any resolution. arXiv preprint arXiv:2409.12191 (2024)."},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.jmsy.2024.04.020"},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.1145\/3664647.3681212"},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01270-0_3"},{"key":"e_1_3_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.1016\/0010-0277(74)90017-1"},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.emnlp-main.775"},{"key":"e_1_3_2_1_40_1","volume-title":"Denny Zhou, et al.","author":"Wei Jason","year":"2022","unstructured":"Jason Wei, Xuezhi Wang, Dale Schuurmans, Maarten Bosma, Fei Xia, Ed Chi, Quoc V Le, Denny Zhou, et al., 2022. Chain-of-thought prompting elicits reasoning in large language models. Advances in neural information processing systems, Vol. 35 (2022), 24824-24837."},{"key":"e_1_3_2_1_41_1","volume-title":"Voronav: Voronoi-based zero-shot object navigation with large language model. arXiv preprint arXiv:2401.02695","author":"Wu Pengying","year":"2024","unstructured":"Pengying Wu, Yao Mu, Bingxian Wu, Yi Hou, Ji Ma, Shanghang Zhang, and Chang Liu. 2024. Voronav: Voronoi-based zero-shot object navigation with large language model. arXiv preprint arXiv:2401.02695 (2024)."},{"key":"e_1_3_2_1_42_1","volume-title":"DFRot: Achieving Outlier-Free and Massive Activation-Free for Rotated LLMs with Refined Rotation. arXiv preprint arXiv:2412.00648","author":"Xiang Jingyang","year":"2024","unstructured":"Jingyang Xiang and Sai Qian Zhang. 2024. DFRot: Achieving Outlier-Free and Massive Activation-Free for Rotated LLMs with Refined Rotation. arXiv preprint arXiv:2412.00648 (2024)."},{"key":"e_1_3_2_1_43_1","volume-title":"Navigation with VLM framework: Go to Any Language. ArXiv","author":"Yin Zecheng","year":"2024","unstructured":"Zecheng Yin, \u2020 ChonghaoCheng, Yinghong Liao, Zhihao Yuan, Shuguang Cui, and Zhen Li. 2024. Navigation with VLM framework: Go to Any Language. ArXiv, Vol. abs\/2410.02787 (2024). https:\/\/api.semanticscholar.org\/CorpusID:273163041"},{"key":"e_1_3_2_1_44_1","volume-title":"Safe-vln: Collision avoidance for vision-and-language navigation of autonomous robots operating in continuous environments","author":"Yue Lu","year":"2024","unstructured":"Lu Yue, Dongliang Zhou, Liang Xie, Feitian Zhang, Ye Yan, and Erwei Yin. 2024. Safe-vln: Collision avoidance for vision-and-language navigation of autonomous robots operating in continuous environments. IEEE Robotics and Automation Letters (2024)."},{"key":"e_1_3_2_1_45_1","volume-title":"Mc-gpt: Empowering vision-and-language navigation with memory map and reasoning chains. arXiv preprint arXiv:2405.10620","author":"Zhan Zhaohuan","year":"2024","unstructured":"Zhaohuan Zhan, Lisha Yu, Sijie Yu, and Guang Tan. 2024. Mc-gpt: Empowering vision-and-language navigation with memory map and reasoning chains. arXiv preprint arXiv:2405.10620 (2024)."},{"key":"e_1_3_2_1_46_1","volume-title":"Navid: Video-based vlm plans the next step for vision-and-language navigation. arXiv preprint arXiv:2402.15852","author":"Zhang Jiazhao","year":"2024","unstructured":"Jiazhao Zhang, Kunyu Wang, Rongtao Xu, Gengze Zhou, Yicong Hong, Xiaomeng Fang, Qi Wu, Zhizheng Zhang, and He Wang. 2024b. Navid: Video-based vlm plans the next step for vision-and-language navigation. arXiv preprint arXiv:2402.15852 (2024)."},{"key":"e_1_3_2_1_47_1","volume-title":"MapNav: A Novel Memory Representation via Annotated Semantic Maps for VLM-based Vision-and-Language Navigation. ArXiv","author":"Zhang Lingfeng","year":"2025","unstructured":"Lingfeng Zhang, Xiaoshuai Hao, Qinwen Xu, Qiang Zhang, Xinyao Zhang, Pengwei Wang, Jing Zhang, Zhongyuan Wang, Shanghang Zhang, and Renjing Xu. 2025. MapNav: A Novel Memory Representation via Annotated Semantic Maps for VLM-based Vision-and-Language Navigation. ArXiv, Vol. abs\/2502.13451 (2025). https:\/\/api.semanticscholar.org\/CorpusID:276449634"},{"key":"e_1_3_2_1_48_1","volume-title":"Vision-and-Language Navigation Today and Tomorrow: A Survey in the Era of Foundation Models. ArXiv","author":"Zhang Yue","year":"2024","unstructured":"Yue Zhang, Ziqiao Ma, Jialu Li, Yanyuan Qiao, Zun Wang, Joyce Chai, Qi Wu, Mohit Bansal, and Parisa Kordjamshidi. 2024a. Vision-and-Language Navigation Today and Tomorrow: A Survey in the Era of Foundation Models. ArXiv, Vol. abs\/2407.07035 (2024). https:\/\/api.semanticscholar.org\/CorpusID:271064503"},{"key":"e_1_3_2_1_49_1","volume-title":"Qqq: Quality quattuor-bit quantization for large language models. arXiv preprint arXiv:2406.09904","author":"Zhang Ying","year":"2024","unstructured":"Ying Zhang, Peng Zhang, Mincong Huang, Jingyang Xiang, Yujie Wang, Chao Wang, Yineng Zhang, Lei Yu, Chuan Liu, and Wei Lin. 2024c. Qqq: Quality quattuor-bit quantization for large language models. arXiv preprint arXiv:2406.09904 (2024)."},{"key":"e_1_3_2_1_50_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA48891.2023.10161289"},{"key":"e_1_3_2_1_51_1","volume-title":"European Conference on Computer Vision. Springer, 260-278","author":"Zhou Gengze","year":"2024","unstructured":"Gengze Zhou, Yicong Hong, Zun Wang, Xin Eric Wang, and Qi Wu. 2024b. Navgpt-2: Unleashing navigational reasoning capability for large vision-language models. In European Conference on Computer Vision. Springer, 260-278."},{"key":"e_1_3_2_1_52_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v38i7.28597"}],"event":{"name":"MM '25: The 33rd ACM International Conference on Multimedia","sponsor":["SIGMM ACM Special Interest Group on Multimedia"],"location":"Dublin Ireland","acronym":"MM '25"},"container-title":["Proceedings of the 33rd ACM International Conference on Multimedia"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3746027.3755832","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,12,10]],"date-time":"2025-12-10T04:04:58Z","timestamp":1765339498000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3746027.3755832"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,10,27]]},"references-count":52,"alternative-id":["10.1145\/3746027.3755832","10.1145\/3746027"],"URL":"https:\/\/doi.org\/10.1145\/3746027.3755832","relation":{},"subject":[],"published":{"date-parts":[[2025,10,27]]},"assertion":[{"value":"2025-10-27","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}