{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,11]],"date-time":"2026-03-11T12:45:15Z","timestamp":1773233115409,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":74,"publisher":"ACM","funder":[{"name":"Beijing Natural Science Foundation","award":["4252054"],"award-info":[{"award-number":["4252054"]}]},{"name":"Youth Innovation Promotion Association CAS","award":["2022132"],"award-info":[{"award-number":["2022132"]}]},{"DOI":"10.13039\/501100005090","name":"Beijing Nova Program","doi-asserted-by":"publisher","award":["20230484276"],"award-info":[{"award-number":["20230484276"]}],"id":[{"id":"10.13039\/501100005090","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,10,27]]},"DOI":"10.1145\/3746027.3755145","type":"proceedings-article","created":{"date-parts":[[2025,10,25]],"date-time":"2025-10-25T07:30:51Z","timestamp":1761377451000},"page":"3664-3672","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":1,"title":["Vision Transformer with Sparse Scan Prior"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0003-4812-7625","authenticated-orcid":false,"given":"Yuguang","family":"Zhang","sequence":"first","affiliation":[{"name":"School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing, China, MAIS &amp; NLPR, Institute of Automation, Chinese Academy of Sciences, Beijing, China, and China Electronics Standardization Insitute, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6115-5503","authenticated-orcid":false,"given":"Qihang","family":"Fan","sequence":"additional","affiliation":[{"name":"MAIS &amp; NLPR, Institute of Automation, Chinese Academy of Sciences, Beijing, China and School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5866-2283","authenticated-orcid":false,"given":"Huaibo","family":"Huang","sequence":"additional","affiliation":[{"name":"MAIS &amp; NLPR, Institute of Automation, Chinese Academy of Sciences, Beijing, China and School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing, China"}]}],"member":"320","published-online":{"date-parts":[[2025,10,27]]},"reference":[{"key":"e_1_3_2_2_1_1","volume-title":"Bermano","author":"Arar Moab","year":"2022","unstructured":"Moab Arar, Ariel Shamir, and Amit H. Bermano. 2022. Learned Queries for Efficient Local Attention. In CVPR."},{"key":"e_1_3_2_2_2_1","unstructured":"Zhaowei Cai and Nuno Vasconcelos. 2018. Cascade R-CNN: Delving into High Quality Object Detection. In CVPR."},{"key":"e_1_3_2_2_3_1","doi-asserted-by":"crossref","unstructured":"Chun-Fu (Richard) Chen Quanfu Fan and Rameswar Panda. 2021. CrossViT: Cross-Attention Multi-Scale Vision Transformer for Image Classification. In ICCV.","DOI":"10.1109\/ICCV48922.2021.00041"},{"key":"e_1_3_2_2_4_1","unstructured":"Chun-Fu (Richard) Chen Rameswar Panda and Quanfu Fan. 2022. RegionViT: Regional-to-Local Attention for Vision Transformers. In ICLR."},{"key":"e_1_3_2_2_5_1","unstructured":"Kai Chen Jiaqi Wang Jiangmiao Pang et al. 2019. MMDetection: Open MMLab Detection Toolbox and Benchmark. arXiv preprint arXiv:1906.07155 (2019)."},{"key":"e_1_3_2_2_6_1","unstructured":"Xiangxiang Chu Zhi Tian Bo Zhang Xinlong Wang and Chunhua Shen. 2023. Conditional Positional Encodings for Vision Transformers. In ICLR."},{"key":"e_1_3_2_2_7_1","unstructured":"MMSegmentation Contributors. 2020. Mmsegmentation an open source semantic segmentation toolbox."},{"key":"e_1_3_2_2_8_1","volume-title":"Imagenet: A large-scale hierarchical image database. In CVPR.","author":"Deng Jia","year":"2009","unstructured":"Jia Deng, Wei Dong, Richard Socher, et al. 2009. Imagenet: A large-scale hierarchical image database. In CVPR."},{"key":"e_1_3_2_2_9_1","doi-asserted-by":"crossref","unstructured":"Mingyu Ding Bin Xiao Noel Codella et al. 2022. DaViT: Dual Attention Vision Transformers. In ECCV.","DOI":"10.1007\/978-3-031-20053-3_5"},{"key":"e_1_3_2_2_10_1","doi-asserted-by":"crossref","unstructured":"Xiaoyi Dong Jianmin Bao Dongdong Chen et al. 2022. CSWin Transformer: A General Vision Transformer Backbone with Cross-Shaped Windows. In CVPR.","DOI":"10.1109\/CVPR52688.2022.01181"},{"key":"e_1_3_2_2_11_1","volume-title":"Words: Transformers for Image Recognition at Scale. In ICLR.","author":"Dosovitskiy Alexey","year":"2021","unstructured":"Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, et al. 2021. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. In ICLR."},{"key":"e_1_3_2_2_12_1","volume-title":"Microsaccades uncover the orientation of covert attention. Vision research 43, 9","author":"Engbert Ralf","year":"2003","unstructured":"Ralf Engbert and Reinhold Kliegl. 2003. Microsaccades uncover the orientation of covert attention. Vision research 43, 9 (2003), 1035--1045."},{"key":"e_1_3_2_2_13_1","volume-title":"RMT: Retentive Networks Meet Vision Transformers. In CVPR.","author":"Fan Qihang","year":"2024","unstructured":"Qihang Fan, Huaibo Huang, Mingrui Chen, Hongmin Liu, and Ran He. 2024. RMT: Retentive Networks Meet Vision Transformers. In CVPR."},{"key":"e_1_3_2_2_14_1","unstructured":"Qihang Fan Huaibo Huang Jiyang Guan and Ran He. 2023. Rethinking Local Perception in Lightweight Vision Transformer. arXiv:2303.17803 [cs.CV]"},{"key":"e_1_3_2_2_15_1","unstructured":"Qihang Fan Huaibo Huang Xiaoqiang Zhou and Ran He. 2023. Lightweight Vision Transformer with Bidirectional Interaction. In NeurIPS."},{"key":"e_1_3_2_2_16_1","volume-title":"Self guided Transformer with Evolving Token Reallocation","year":"2023","unstructured":"SG-Former: Self guided Transformer with Evolving Token Reallocation. 2023. Sucheng Ren, Xingyi Yang, Songhua Liu, Xinchao Wang. In ICCV."},{"key":"e_1_3_2_2_17_1","volume-title":"CMT: Convolutional neural networks meet vision transformers. In CVPR.","author":"Guo Jianyuan","year":"2022","unstructured":"Jianyuan Guo, Kai Han, HanWu, Chang Xu, Yehui Tang, Chunjing Xu, and Yunhe Wang. 2022. CMT: Convolutional neural networks meet vision transformers. In CVPR."},{"key":"e_1_3_2_2_18_1","volume-title":"Visual Attention Network. arXiv preprint arXiv:2202.09741","author":"Guo Meng-Hao","year":"2022","unstructured":"Meng-Hao Guo, Cheng-Ze Lu, Zheng-Ning Liu, Ming-Ming Cheng, and Shi-Min Hu. 2022. Visual Attention Network. arXiv preprint arXiv:2202.09741 (2022)."},{"key":"e_1_3_2_2_19_1","unstructured":"Dongchen Han Ziyi Wang Zhuofan Xia Yizeng Han Yifan Pu Chunjiang Ge Jun Song Shiji Song Bo Zheng and Gao Huang. 2024. Demystify Mamba in Vision: A Linear Attention Perspective. In NeurIPS."},{"key":"e_1_3_2_2_20_1","unstructured":"Ali Hassani and Humphrey Shi. 2022. Dilated Neighborhood Attention Transformer. arXiv:2209.15001 [cs.CV] https:\/\/arxiv.org\/abs\/2209.15001"},{"key":"e_1_3_2_2_21_1","doi-asserted-by":"crossref","unstructured":"Ali Hassani Steven Walton Jiachen Li Shen Li and Humphrey Shi. 2023. Neighborhood Attention Transformer. In CVPR.","DOI":"10.1109\/CVPR52729.2023.00599"},{"key":"e_1_3_2_2_22_1","unstructured":"Ali Hatamizadeh Greg Heinrich Hongxu Yin Andrew Tao Jose M Alvarez Jan Kautz and Pavlo Molchanov. 2024. FasterViT: Fast Vision Transformers with Hierarchical Attention. In ICLR."},{"key":"e_1_3_2_2_23_1","unstructured":"Ali Hatamizadeh Hongxu Yin Greg Heinrich Jan Kautz and Pavlo Molchanov. 2023. Global context vision transformers. In ICML."},{"key":"e_1_3_2_2_24_1","unstructured":"Kaiming He Xinlei Chen Saining Xie Yanghao Li Piotr Doll\u00e1r and Ross Girshick. 2022. Masked Autoencoders Are Scalable Vision Learners. In CVPR."},{"key":"e_1_3_2_2_25_1","volume-title":"Girshick","author":"He Kaiming","year":"2017","unstructured":"Kaiming He, Georgia Gkioxari, Piotr Doll\u00e1r, and Ross B. Girshick. 2017. Mask R-CNN. In ICCV."},{"key":"e_1_3_2_2_26_1","doi-asserted-by":"crossref","unstructured":"Dan Hendrycks Steven Basart Norman Mu Saurav Kadavath FrankWang Evan Dorundo Rahul Desai Tyler Zhu Samyak Parajuli Mike Guo Dawn Song Jacob Steinhardt and Justin Gilmer. 2021. The Many Faces of Robustness: A Critical Analysis of Out-of-Distribution Generalization. In ICCV.","DOI":"10.1109\/ICCV48922.2021.00823"},{"key":"e_1_3_2_2_27_1","doi-asserted-by":"crossref","unstructured":"Dan Hendrycks Kevin Zhao Steven Basart Jacob Steinhardt and Dawn Song. 2021. Natural Adversarial Examples. In CVPR.","DOI":"10.1109\/CVPR46437.2021.01501"},{"key":"e_1_3_2_2_28_1","volume-title":"Conv2Former: A Simple Transformer-Style ConvNet for Visual Recognition. arXiv preprint arXiv:2211.11943","author":"Hou Qibin","year":"2022","unstructured":"Qibin Hou, Cheng-Ze Lu, Ming-Ming Cheng, and Jiashi Feng. 2022. Conv2Former: A Simple Transformer-Style ConvNet for Visual Recognition. arXiv preprint arXiv:2211.11943 (2022)."},{"key":"e_1_3_2_2_29_1","doi-asserted-by":"crossref","unstructured":"Gao Huang Yu Sun and Zhuang Liu. 2016. Deep networks with stochastic depth. In ECCV.","DOI":"10.1007\/978-3-319-46493-0_39"},{"key":"e_1_3_2_2_30_1","unstructured":"Huaibo Huang Xiaoqiang Zhou Jie Cao Ran He and Tieniu Tan. 2023. Vision Transformer with Super Token Sampling. In CVPR."},{"key":"e_1_3_2_2_31_1","volume-title":"Orthogonal Transformer: An Efficient Vision Transformer Backbone with Token Orthogonalization. In NeurIPS.","author":"Huang Huaibo","year":"2022","unstructured":"Huaibo Huang, Xiaoqiang Zhou, and Ran He. 2022. Orthogonal Transformer: An Efficient Vision Transformer Backbone with Token Orthogonalization. In NeurIPS."},{"key":"e_1_3_2_2_32_1","doi-asserted-by":"publisher","DOI":"10.1113\/jphysiol.1962.sp006837"},{"key":"e_1_3_2_2_33_1","unstructured":"Zi-Hang Jiang Qibin Hou Li Yuan Daquan Zhou Yujun Shi Xiaojie Jin Anran Wang and Jiashi Feng. 2021. All Tokens Matter: Token Labeling for Training Better Vision Transformers. In NeurIPS."},{"key":"e_1_3_2_2_34_1","volume-title":"Cordelia Schmid, and Minsu Cho.","author":"Kim Manjin","year":"2024","unstructured":"Manjin Kim, Paul Hongsuck Seo, Cordelia Schmid, and Minsu Cho. 2024. Learning Correlation Structures for Vision Transformers. In CVPR."},{"key":"e_1_3_2_2_35_1","doi-asserted-by":"crossref","unstructured":"Alexander Kirillov Ross Girshick Kaiming He and Piotr Doll\u00e1r. 2019. Panoptic feature pyramid networks. In CVPR.","DOI":"10.1109\/CVPR.2019.00656"},{"key":"e_1_3_2_2_36_1","unstructured":"Youngwan Lee Jonghee Kim Jeffrey Willette and Sung Ju Hwang. 2022. MPViT: Multi-Path Vision Transformer for Dense Prediction. In CVPR."},{"key":"e_1_3_2_2_37_1","unstructured":"Kunchang Li Yali Wang Peng Gao Guanglu Song Yu Liu Hongsheng Li and Yu Qiao. 2022. UniFormer: Unified Transformer for Efficient Spatiotemporal Representation Learning. arXiv:2201.04676 [cs.CV]"},{"key":"e_1_3_2_2_38_1","unstructured":"Yanghao Li Chao-YuanWu Haoqi Fan Karttikeya Mangalam Bo Xiong Jitendra Malik and Christoph Feichtenhofer. 2022. MViTv2: Improved multiscale vision transformers for classification and detection. In CVPR."},{"key":"e_1_3_2_2_39_1","volume-title":"Efficientformer: Vision transformers at mobilenet speed. In NeurIPS.","author":"Li Yanyu","year":"2022","unstructured":"Yanyu Li, Geng Yuan, Yang Wen, Ju Hu, Georgios Evangelidis, Sergey Tulyakov, Yanzhi Wang, and Jian Ren. 2022. Efficientformer: Vision transformers at mobilenet speed. In NeurIPS."},{"key":"e_1_3_2_2_40_1","unstructured":"Tsung-Yi Lin Priya Goyal Ross B. Girshick and Kaiming He andPiotr Doll\u00e1r. 2017. Focal loss for dense object detection. In ICCV."},{"key":"e_1_3_2_2_41_1","unstructured":"Tsung-Yi Lin Michael Maire Serge Belongie et al. 2014. Microsoft coco: Common objects in context. In ECCV."},{"key":"e_1_3_2_2_42_1","volume-title":"Scale-Aware Modulation Meet Transformer. In ICCV.","author":"Lin Weifeng","year":"2023","unstructured":"Weifeng Lin, Ziheng Wu, Jiayu Chen, Jun Huang, and Lianwen Jin. 2023. Scale-Aware Modulation Meet Transformer. In ICCV."},{"key":"e_1_3_2_2_43_1","volume-title":"Swin Transformer: Hierarchical Vision Transformer using Shifted Windows. In ICCV.","author":"Liu Ze","year":"2021","unstructured":"Ze Liu, Yutong Lin, Yue Cao, Han Hu, Yixuan Wei, Zheng Zhang, Stephen Lin, and Baining Guo. 2021. Swin Transformer: Hierarchical Vision Transformer using Shifted Windows. In ICCV."},{"key":"e_1_3_2_2_44_1","unstructured":"Zhuang Liu Hanzi Mao Chao-Yuan Wu et al. 2022. A convnet for the 2020s. In CVPR."},{"key":"e_1_3_2_2_45_1","unstructured":"Xiaofeng Mao Gege Qi Yuefeng Chen Xiaodan Li Ranjie Duan Shaokai Ye Yuan He and Hui Xue. 2022. Towards Robust Vision Transformer. In CVPR."},{"key":"e_1_3_2_2_46_1","doi-asserted-by":"publisher","DOI":"10.1167\/8.14.28"},{"key":"e_1_3_2_2_47_1","doi-asserted-by":"publisher","DOI":"10.1038\/nrn1348"},{"key":"e_1_3_2_2_48_1","doi-asserted-by":"publisher","DOI":"10.1038\/nrn2619"},{"key":"e_1_3_2_2_49_1","unstructured":"Junting Pan Adrian Bulat Fuwen Tan et al. 2022. EdgeViTs: Competing Lightweight CNNs on Mobile Devices with Vision Transformers. In ECCV."},{"key":"e_1_3_2_2_50_1","unstructured":"Zizheng Pan Jianfei Cai and Bohan Zhuang. 2022. Fast Vision Transformers with HiLo Attention. In NeurIPS."},{"key":"e_1_3_2_2_51_1","volume-title":"A compact field guide to the study of microsaccades: Challenges and functions. Vision research 118","author":"Poletti Martina","year":"2016","unstructured":"Martina Poletti and Michele Rucci. 2016. A compact field guide to the study of microsaccades: Challenges and functions. Vision research 118 (2016), 83--97."},{"key":"e_1_3_2_2_52_1","unstructured":"Yongming Rao Wenliang Zhao Benlin Liu Jiwen Lu Jie Zhou and Cho-Jui Hsieh. 2021. DynamicViT: Efficient Vision Transformers with Dynamic Token Sparsification. In NeurIPS."},{"key":"e_1_3_2_2_53_1","volume-title":"Miniature eye movements enhance fine spatial detail. Nature 447, 7146","author":"Rucci Michele","year":"2007","unstructured":"Michele Rucci, Ramona Iovin, Martina Poletti, and Francesca Santini. 2007. Miniature eye movements enhance fine spatial detail. Nature 447, 7146 (2007), 851--854."},{"key":"e_1_3_2_2_54_1","doi-asserted-by":"publisher","DOI":"10.1146\/annurev-vision-082114-035742"},{"key":"e_1_3_2_2_55_1","unstructured":"Chenyang Si Weihao Yu Pan Zhou Yichen Zhou XinchaoWang and Shuicheng YAN. 2022. Inception Transformer. In NeurIPS."},{"key":"e_1_3_2_2_56_1","unstructured":"Shitao Tang Jiahui Zhang Siyu Zhu et al. 2022. Quadtree attention for vision transformers. In ICLR."},{"key":"e_1_3_2_2_57_1","doi-asserted-by":"publisher","DOI":"10.1523\/JNEUROSCI.0741-05.2005"},{"key":"e_1_3_2_2_58_1","unstructured":"Zhengzhong Tu Hossein Talebi Han Zhang Feng Yang Peyman Milanfar Alan Bovik and Yinxiao Li. 2022. MaxViT: Multi-Axis Vision Transformer. In ECCV."},{"key":"e_1_3_2_2_59_1","unstructured":"Ashish Vaswani Noam Shazeer Niki Parmar et al. 2017. Attention is all you need. In NeurIPS."},{"key":"e_1_3_2_2_60_1","doi-asserted-by":"crossref","unstructured":"Wenhai Wang Jifeng Dai Zhe Chen Zhenhang Huang Zhiqi Li Xizhou Zhu Xiaowei Hu Tong Lu Lewei Lu Hongsheng Li et al. 2023. InternImage: Exploring Large-Scale Vision Foundation Models with Deformable Convolutions. In CVPR.","DOI":"10.1109\/CVPR52729.2023.01385"},{"key":"e_1_3_2_2_61_1","doi-asserted-by":"crossref","unstructured":"WenhaiWang Enze Xie Xiang Li Deng-Ping Fan Kaitao Song Ding Liang Tong Lu Ping Luo and Ling Shao. 2021. Pyramid Vision Transformer: A Versatile Backbone for Dense Prediction without Convolutions. In ICCV.","DOI":"10.1109\/ICCV48922.2021.00061"},{"key":"e_1_3_2_2_62_1","first-page":"1","article-title":"Pvtv2: Improved baselines with pyramid vision transformer","volume":"8","author":"Xie Enze","year":"2022","unstructured":"WenhaiWang, Enze Xie, Xiang Li, Deng-Ping Fan, Kaitao Song, Ding Liang, Tong Lu, Ping Luo, and Ling Shao. 2022. Pvtv2: Improved baselines with pyramid vision transformer. Computational Visual Media 8, 3 (2022), 1--10.","journal-title":"Computational Visual Media"},{"key":"e_1_3_2_2_63_1","unstructured":"Wenxiao Wang Lu Yao Long Chen Binbin Lin Deng Cai Xiaofei He and Wei Liu. 2022. CrossFormer: A Versatile Vision Transformer Hinging on Cross-scale Attention. In ICLR."},{"key":"e_1_3_2_2_64_1","volume-title":"Li Erran Li, and Gao Huang","author":"Xia Zhuofan","year":"2022","unstructured":"Zhuofan Xia, Xuran Pan, Shiji Song, Li Erran Li, and Gao Huang. 2022. Vision Transformer With Deformable Attention. In CVPR."},{"key":"e_1_3_2_2_65_1","doi-asserted-by":"crossref","unstructured":"Tete Xiao Yingcheng Liu Bolei Zhou Yuning Jiang and Jian Sun. 2018. Unified Perceptual Parsing for Scene Understanding. In ECCV.","DOI":"10.1007\/978-3-030-01228-1_26"},{"key":"e_1_3_2_2_66_1","volume-title":"MOAT: Alternating mobile convolution and attention brings strong vision models. In ICLR.","author":"Yang Chenglin","year":"2023","unstructured":"Chenglin Yang, Siyuan Qiao, Qihang Yu, et al. 2023. MOAT: Alternating mobile convolution and attention brings strong vision models. In ICLR."},{"key":"e_1_3_2_2_67_1","doi-asserted-by":"crossref","unstructured":"Chenglin Yang Yilin Wang Jianming Zhang et al. 2022. Lite Vision Transformer with Enhanced Self-Attention. In CVPR.","DOI":"10.1109\/CVPR52688.2022.01169"},{"key":"e_1_3_2_2_68_1","unstructured":"Jianwei Yang Chunyuan Li Pengchuan Zhang Xiyang Dai Bin Xiao Lu Yuan and Jianfeng Gao. 2021. Focal Self-Attention for Local-Global Interactions in Vision Transformers. In NeurIPS."},{"key":"e_1_3_2_2_69_1","doi-asserted-by":"crossref","unstructured":"Rui Yang Hailong Ma JieWu Yansong Tang Xuefeng Xiao Min Zheng and Xiu Li. 2022. ScalableViT: Rethinking the context-oriented generalization of vision transformer. In ECCV.","DOI":"10.1007\/978-3-031-20053-3_28"},{"key":"e_1_3_2_2_70_1","volume-title":"Dual Vision Transformer. TPAMI","author":"Yao Ting","year":"2023","unstructured":"Ting Yao, Yehao Li, Yingwei Pan, Yu Wang, Xiao-Ping Zhang, and Tao Mei. 2023. Dual Vision Transformer. TPAMI (2023)."},{"key":"e_1_3_2_2_71_1","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2023.3329173"},{"key":"e_1_3_2_2_72_1","doi-asserted-by":"crossref","unstructured":"Bolei Zhou Hang Zhao Xavier Puig et al. 2017. Scene parsing through ade20k dataset. In CVPR.","DOI":"10.1109\/CVPR.2017.544"},{"key":"e_1_3_2_2_73_1","volume-title":"Alvarez","author":"Zhou Daquan","year":"2022","unstructured":"Daquan Zhou, Zhiding Yu, Enze Xie, Chaowei Xiao, Anima Anandkumar, Jiashi Feng, and Jose M. Alvarez. 2022. Understanding The Robustness in Vision Transformers. In ICML."},{"key":"e_1_3_2_2_74_1","doi-asserted-by":"crossref","unstructured":"Lei Zhu Xinjiang Wang Zhanghan Ke Wayne Zhang and Rynson Lau. 2023. BiFormer: Vision Transformer with Bi-Level Routing Attention. In CVPR.","DOI":"10.1109\/CVPR52729.2023.00995"}],"event":{"name":"MM '25: The 33rd ACM International Conference on Multimedia","location":"Dublin Ireland","acronym":"MM '25","sponsor":["SIGMM ACM Special Interest Group on Multimedia"]},"container-title":["Proceedings of the 33rd ACM International Conference on Multimedia"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3746027.3755145","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,12,9]],"date-time":"2025-12-09T19:56:15Z","timestamp":1765310175000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3746027.3755145"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,10,27]]},"references-count":74,"alternative-id":["10.1145\/3746027.3755145","10.1145\/3746027"],"URL":"https:\/\/doi.org\/10.1145\/3746027.3755145","relation":{},"subject":[],"published":{"date-parts":[[2025,10,27]]},"assertion":[{"value":"2025-10-27","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}