{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T01:10:24Z","timestamp":1755825024433,"version":"3.44.0"},"publisher-location":"New York, NY, USA","reference-count":52,"publisher":"ACM","funder":[{"name":"Meituan"},{"DOI":"10.13039\/501100006374","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62476201;62222203"],"award-info":[{"award-number":["62476201;62222203"]}],"id":[{"id":"10.13039\/501100006374","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,6,30]]},"DOI":"10.1145\/3731715.3733293","type":"proceedings-article","created":{"date-parts":[[2025,6,25]],"date-time":"2025-06-25T18:31:04Z","timestamp":1750876264000},"page":"1045-1053","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["Composed Query-Based Event Retrieval in Video Corpus with Multimodal Episodic Perceptron"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0006-0981-5021","authenticated-orcid":false,"given":"Fan","family":"Ni","sequence":"first","affiliation":[{"name":"School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2209-651X","authenticated-orcid":false,"given":"Xun","family":"Jiang","sequence":"additional","affiliation":[{"name":"School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5711-7769","authenticated-orcid":false,"given":"Hao","family":"Yang","sequence":"additional","affiliation":[{"name":"Meituan, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6083-2656","authenticated-orcid":false,"given":"Chong","family":"Peng","sequence":"additional","affiliation":[{"name":"Meituan, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7122-7147","authenticated-orcid":false,"given":"Peng","family":"Yan","sequence":"additional","affiliation":[{"name":"Meituan, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9318-0084","authenticated-orcid":false,"given":"Zheng","family":"Wang","sequence":"additional","affiliation":[{"name":"School of Computer Science and Technology, Tongji University, Shanghai, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7303-3231","authenticated-orcid":false,"given":"Fumin","family":"Shen","sequence":"additional","affiliation":[{"name":"School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5685-3123","authenticated-orcid":false,"given":"Xing","family":"Xu","sequence":"additional","affiliation":[{"name":"School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu, China and School of Computer Science and Technology, Tongji University, Shanghai, China"}]}],"member":"320","published-online":{"date-parts":[[2025,6,30]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"crossref","unstructured":"Ranjay Krishna Kenji Hata Frederic Ren Li Fei-Fei and Juan Carlos Niebles. 2017. Dense-Captioning Events in Videos. In ICCV. 706--715.","DOI":"10.1109\/ICCV.2017.83"},{"key":"e_1_3_2_1_2_1","volume-title":"Tvr: A large-scale dataset for video-subtitle moment retrieval. In ECCV. 447--463.","author":"Lei Jie","year":"2020","unstructured":"Jie Lei, Licheng Yu, Tamara L Berg, and Mohit Bansal. 2020. Tvr: A large-scale dataset for video-subtitle moment retrieval. In ECCV. 447--463."},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"crossref","unstructured":"Lisa Anne Hendricks Oliver Wang Eli Shechtman Josef Sivic Trevor Darrell and Bryan Russell. 2017. Localizing Moments in Video with Natural Language.. In ICCV. 5804--5813.","DOI":"10.1109\/ICCV.2017.618"},{"key":"e_1_3_2_1_4_1","volume-title":"Tall: Temporal activity localization via language query. In ICCV. 5277--5285.","author":"Gao Jiyang","year":"2017","unstructured":"Jiyang Gao, Chen Sun, Zhenheng Yang, and Ram Nevatia. 2017. Tall: Temporal activity localization via language query. In ICCV. 5277--5285."},{"key":"e_1_3_2_1_5_1","volume-title":"Joey Tianyi Zhou, and Rick Siow Mong Goh","author":"Zhang Hao","year":"2021","unstructured":"Hao Zhang, Aixin Sun, Wei Jing, Guoshun Nan, Liangli Zhen, Joey Tianyi Zhou, and Rick Siow Mong Goh. 2021. Video corpus moment retrieval with contrastive learning. In SIGIR. 685--695."},{"key":"e_1_3_2_1_6_1","volume-title":"Eunseop Yoon, Dahyun Kim, Junyeong Kim, Hee Suk Yoon, and Chang D Yoo.","author":"Yoon Sunjae","year":"2022","unstructured":"Sunjae Yoon, Ji Woo Hong, Eunseop Yoon, Dahyun Kim, Junyeong Kim, Hee Suk Yoon, and Chang D Yoo. 2022. Selective query-guided debiasing for video corpus moment retrieval. In ECCV. 185--200."},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"crossref","unstructured":"Minjoon Jung SeongHo Choi JooChan Kim Jin-Hwa Kim and Byoung-Tak Zhang. 2022. Modal-specific Pseudo Query Generation for Video Corpus Moment Retrieval. In EMNLP. 7769--7781.","DOI":"10.18653\/v1\/2022.emnlp-main.530"},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"crossref","unstructured":"Zhiguo Chen Xun Jiang Xing Xu Zuo Cao Yijun Mo and Heng Tao Shen. 2023. Joint searching and grounding: Multi-granularity video content retrieval. In ACM MM. 975--983.","DOI":"10.1145\/3581783.3612349"},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1145\/3474085.3475515"},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"crossref","unstructured":"Shizhe Chen Yida Zhao Qin Jin and Qi Wu. 2020. Fine-grained video-text retrieval with hierarchical graph reasoning. In CVPR. 10638--10647.","DOI":"10.1109\/CVPR42600.2020.01065"},{"key":"e_1_3_2_1_11_1","volume-title":"X-clip: End-to-end multi-grained contrastive learning for video-text retrieval. In ACM MM. 638--647.","author":"Ma Yiwei","year":"2022","unstructured":"Yiwei Ma, Guohai Xu, Xiaoshuai Sun, Ming Yan, Ji Zhang, and Rongrong Ji. 2022. X-clip: End-to-end multi-grained contrastive learning for video-text retrieval. In ACM MM. 638--647."},{"key":"e_1_3_2_1_12_1","first-page":"1","article-title":"Cross-Modal Attention Preservation with Self-Contrastive Learning for Composed Query-Based Image Retrieval","volume":"20","author":"Li Shenshen","year":"2024","unstructured":"Shenshen Li, Xing Xu, Xun Jiang, Fumin Shen, Zhe Sun, and Andrzej Cichocki. 2024. Cross-Modal Attention Preservation with Self-Contrastive Learning for Composed Query-Based Image Retrieval. TOMM, Vol. 20, 6 (2024), 1--22.","journal-title":"TOMM"},{"key":"e_1_3_2_1_13_1","first-page":"5509","article-title":"Geometric Matching for Cross-Modal Retrieval","volume":"36","author":"Wang Zheng","year":"2025","unstructured":"Zheng Wang, Zhenwei Gao, Yang Yang, Guoqing Wang, Chengbo Jiao, and Heng Tao Shen. 2025. Geometric Matching for Cross-Modal Retrieval. IEEE TNNLS, Vol. 36, 3 (2025), 5509--5521.","journal-title":"IEEE TNNLS"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"crossref","unstructured":"Zheng Wang Zhenwei Gao Kangshuai Guo Yang Yang Xiaoming Wang and Heng Tao Shen. 2023. Multilateral Semantic Relations Modeling for Image Text Retrieval. In CVPR. 2830--2839.","DOI":"10.1109\/CVPR52729.2023.00277"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"crossref","unstructured":"Xun Jiang Xing Xu Jingran Zhang Fumin Shen Zuo Cao and Heng Tao Shen. 2022. Semi-supervised video paragraph grounding with contrastive encoder. In CVPR. 2466--2475.","DOI":"10.1109\/CVPR52688.2022.00250"},{"key":"e_1_3_2_1_16_1","first-page":"2226","article-title":"Semantics disentangling for cross-modal retrieval","volume":"33","author":"Wang Zheng","year":"2024","unstructured":"Zheng Wang, Xing Xu, Jiwei Wei, Ning Xie, Yang Yang, and Heng Tao Shen. 2024. Semantics disentangling for cross-modal retrieval. IEEE TIP, Vol. 33 (2024), 2226--2237.","journal-title":"IEEE TIP"},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"crossref","unstructured":"Jinpeng Wang Yixiao Ge Guanyu Cai Rui Yan Xudong Lin Ying Shan Xiaohu Qie and Mike Zheng Shou. 2022. Object-aware video-language pre-training for retrieval. In CVPR. 3313--3322.","DOI":"10.1109\/CVPR52688.2022.00331"},{"key":"e_1_3_2_1_18_1","volume-title":"MILES: Visual BERT Pre-training with Injected Language Semantics for Video-Text Retrieval. In ECCV. 691--708.","author":"Ge Yuying","year":"2022","unstructured":"Yuying Ge, Yixiao Ge, Xihui Liu, Jinpeng Wang, Jianping Wu, Ying Shan, Xiaohu Qie, and Ping Luo. 2022. MILES: Visual BERT Pre-training with Injected Language Semantics for Video-Text Retrieval. In ECCV. 691--708."},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"crossref","unstructured":"Jianfeng Dong Xianke Chen Minsong Zhang Xun Yang Shujie Chen Xirong Li and Xun Wang. 2022. Partially relevant video retrieval. In ACM MM. 246--257.","DOI":"10.1145\/3503161.3547976"},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"crossref","unstructured":"Jianfeng Dong Minsong Zhang Zheng Zhang Xianke Chen Daizong Liu Xiaoye Qu Xun Wang and Baolong Liu. 2023. Dual learning with dynamic knowledge distillation for partially relevant video retrieval. In ICCV. 11302--11312.","DOI":"10.1109\/ICCV51070.2023.01038"},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"crossref","unstructured":"Xun Jiang Zhiguo Chen Xing Xu Fumin Shen Zuo Cao and Xunliang Cai. 2023. Progressive Event Alignment Network for Partial Relevant Video Retrieval. In ICME. 1973--1978.","DOI":"10.1109\/ICME55011.2023.00338"},{"key":"e_1_3_2_1_22_1","volume-title":"GMMFormer v2: An Uncertainty-aware Framework for Partially Relevant Video Retrieval. arXiv preprint arXiv:2405.13824","author":"Wang Yuting","year":"2024","unstructured":"Yuting Wang, Jinpeng Wang, Bin Chen, Tao Dai, Ruisheng Luo, and Shu-Tao Xia. 2024. GMMFormer v2: An Uncertainty-aware Framework for Partially Relevant Video Retrieval. arXiv preprint arXiv:2405.13824 (2024)."},{"key":"e_1_3_2_1_23_1","first-page":"3921","article-title":"Siamese alignment network for weakly supervised video moment retrieval","volume":"25","author":"Wang Yunxiao","year":"2022","unstructured":"Yunxiao Wang, Meng Liu, Yinwei Wei, Zhiyong Cheng, Yinglong Wang, and Liqiang Nie. 2022. Siamese alignment network for weakly supervised video moment retrieval. TMM, Vol. 25 (2022), 3921--3933.","journal-title":"TMM"},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"crossref","unstructured":"Xun Jiang Zailei Zhou Xing Xu Yang Yang Guoqing Wang and Heng Tao Shen. 2023. Faster video moment retrieval with point-level supervision. In ACM MM. 1334--1342.","DOI":"10.1145\/3581783.3612394"},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2024.3396272"},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"crossref","unstructured":"Junyu Gao and Changsheng Xu. 2021. Fast video moment retrieval. In ICCV. 1523--1532.","DOI":"10.1109\/ICCV48922.2021.00155"},{"key":"e_1_3_2_1_27_1","volume-title":"Sdn: Semantic decoupling network for temporal language grounding. TNNLS","author":"Jiang Xun","year":"2022","unstructured":"Xun Jiang, Xing Xu, Jingran Zhang, Fumin Shen, Zuo Cao, and Heng Tao Shen. 2022. Sdn: Semantic decoupling network for temporal language grounding. TNNLS (2022)."},{"key":"e_1_3_2_1_28_1","volume-title":"VERIFIED: A Video Corpus Moment Retrieval Benchmark for Fine-Grained Video Understanding. arXiv preprint arXiv:2410.08593","author":"Chen Houlun","year":"2024","unstructured":"Houlun Chen, Xin Wang, Hong Chen, Zeyang Zhang, Wei Feng, Bin Huang, Jia Jia, and Wenwu Zhu. 2024. VERIFIED: A Video Corpus Moment Retrieval Benchmark for Fine-Grained Video Understanding. arXiv preprint arXiv:2410.08593 (2024)."},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v38i4.28101"},{"key":"e_1_3_2_1_30_1","unstructured":"Zheyuan Liu Cristian Rodriguez-Opazo Damien Teney and Stephen Gould. 2021. Image Retrieval on Real-Life Images With Pre-Trained Vision-and-Language Models. In ICCV. 2105--2114."},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"crossref","unstructured":"Hui Wu Yupeng Gao Xiaoxiao Guo Ziad Al-Halah Steven Rennie Kristen Grauman and Rogerio Feris. 2021. Fashion iq: A new dataset towards retrieving images by natural language feedback. In CVPR. 11307--11317.","DOI":"10.1109\/CVPR46437.2021.01115"},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"crossref","unstructured":"Shenshen Li. 2023. Dual-Path Semantic Construction Network for Composed Query-Based Image Retrieval. In ICMR. 636--639.","DOI":"10.1145\/3591106.3592245"},{"key":"e_1_3_2_1_33_1","volume-title":"Set of Diverse Queries with Uncertainty Regularization for Composed Image Retrieval. TCSVT","author":"Xu Yahui","year":"2024","unstructured":"Yahui Xu, Jiwei Wei, Yi Bin, Yang Yang, Zeyu Ma, and Heng Tao Shen. 2024. Set of Diverse Queries with Uncertainty Regularization for Composed Image Retrieval. TCSVT (2024), 10494--10506."},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"crossref","unstructured":"Alberto Baldrati Marco Bertini Tiberio Uricchio and Alberto Del Bimbo. 2022. Effective conditioned and composed image retrieval combining CLIP-based features. In CVPR. 21434--21442.","DOI":"10.1109\/CVPR52688.2022.02080"},{"key":"e_1_3_2_1_35_1","volume-title":"Multi-grained attention network with mutual exclusion for composed query-based image retrieval. TCSVT","author":"Li Shenshen","year":"2023","unstructured":"Shenshen Li, Xing Xu, Xun Jiang, Fumin Shen, Xin Liu, and Heng Tao Shen. 2023. Multi-grained attention network with mutual exclusion for composed query-based image retrieval. TCSVT (2023), 2959--2972."},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v38i7.28479"},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"crossref","unstructured":"Haokun Wen Xuemeng Song Xiaolin Chen Yinwei Wei Liqiang Nie and Tat-Seng Chua. 2024. Simple but Effective Raw-Data Level Multimodal Fusion for Composed Image Retrieval. In SIGIR. 229--239.","DOI":"10.1145\/3626772.3657727"},{"key":"e_1_3_2_1_38_1","volume-title":"Covr: Learning composed video retrieval from web video captions. In AAAI. 5270--5279.","author":"Ventura Lucas","year":"2024","unstructured":"Lucas Ventura, Antoine Yang, Cordelia Schmid, and G\u00fcl Varol. 2024a. Covr: Learning composed video retrieval from web video captions. In AAAI. 5270--5279."},{"key":"e_1_3_2_1_39_1","volume-title":"CoVR-2: Automatic Data Construction for Composed Video Retrieval. TPAMI","author":"Ventura Lucas","year":"2024","unstructured":"Lucas Ventura, Antoine Yang, Cordelia Schmid, and G\u00fcl Varol. 2024b. CoVR-2: Automatic Data Construction for Composed Video Retrieval. TPAMI (2024), 11409--11421."},{"key":"e_1_3_2_1_40_1","volume-title":"Salman Khan, Michael Felsberg, Mubarak Shah, and Fahad Shahbaz Khan.","author":"Thawakar Omkar","year":"2024","unstructured":"Omkar Thawakar, Muzammal Naseer, Rao Muhammad Anwer, Salman Khan, Michael Felsberg, Mubarak Shah, and Fahad Shahbaz Khan. 2024. Composed Video Retrieval via Enriched Context and Discriminative Embeddings. In CVPR. 26896--26906."},{"key":"e_1_3_2_1_41_1","volume-title":"EgoCVR: An Egocentric Benchmark for Fine-Grained Composed Video Retrieval. arXiv preprint arXiv:2407.16658","author":"Hummel Thomas","year":"2024","unstructured":"Thomas Hummel, Shyamgopal Karthik, Mariana-Iuliana Georgescu, and Zeynep Akata. 2024. EgoCVR: An Egocentric Benchmark for Fine-Grained Composed Video Retrieval. arXiv preprint arXiv:2407.16658 (2024)."},{"key":"e_1_3_2_1_42_1","volume-title":"Yan Xia, Yansong Tang, Daniel Cremers, Philip Torr, Volker Tresp, and Jindong Gu.","author":"Zhang Gengyuan","year":"2024","unstructured":"Gengyuan Zhang, Mang Ling Ada Fok, Yan Xia, Yansong Tang, Daniel Cremers, Philip Torr, Volker Tresp, and Jindong Gu. 2024. Localizing Events in Videos with Multimodal Queries. arXiv preprint arXiv:2406.10079 (2024)."},{"key":"e_1_3_2_1_43_1","volume-title":"NeurIPS (2020)","author":"Brown Tom","year":"2020","unstructured":"Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared D Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, et al. 2020. Language models are few-shot learners. NeurIPS (2020), 1877--1901."},{"key":"e_1_3_2_1_44_1","volume-title":"Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al.","author":"Radford Alec","year":"2021","unstructured":"Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. 2021. Learning transferable visual models from natural language supervision. In ICML. 8748--8763."},{"key":"e_1_3_2_1_45_1","volume-title":"Vl-bert: Pre-training of generic visual-linguistic representations. arXiv preprint arXiv:1908.08530","author":"Su Weijie","year":"2019","unstructured":"Weijie Su, Xizhou Zhu, Yue Cao, Bin Li, Lewei Lu, Furu Wei, and Jifeng Dai. 2019. Vl-bert: Pre-training of generic visual-linguistic representations. arXiv preprint arXiv:1908.08530 (2019)."},{"key":"e_1_3_2_1_46_1","volume-title":"Carl Vondrick, Kevin Murphy, and Cordelia Schmid.","author":"Sun Chen","year":"2019","unstructured":"Chen Sun, Austin Myers, Carl Vondrick, Kevin Murphy, and Cordelia Schmid. 2019. Videobert: A joint model for video and language representation learning. In ICCV. 7463--7472."},{"key":"e_1_3_2_1_47_1","volume-title":"Visualbert: A simple and performant baseline for vision and language. arXiv preprint arXiv:1908.03557","author":"Li Liunian Harold","year":"2019","unstructured":"Liunian Harold Li, Mark Yatskar, Da Yin, Cho-Jui Hsieh, and Kai-Wei Chang. 2019. Visualbert: A simple and performant baseline for vision and language. arXiv preprint arXiv:1908.03557 (2019)."},{"key":"e_1_3_2_1_48_1","volume-title":"Unicoder-vl: A universal encoder for vision and language by cross-modal pre-training. In AAAI. 11336--11344.","author":"Li Gen","year":"2020","unstructured":"Gen Li, Nan Duan, Yuejian Fang, Ming Gong, and Daxin Jiang. 2020. Unicoder-vl: A universal encoder for vision and language by cross-modal pre-training. In AAAI. 11336--11344."},{"key":"e_1_3_2_1_49_1","volume-title":"Faisal Ahmed, Zhe Gan, Yu Cheng, and Jingjing Liu.","author":"Chen Yen-Chun","year":"2020","unstructured":"Yen-Chun Chen, Linjie Li, Licheng Yu, Ahmed El Kholy, Faisal Ahmed, Zhe Gan, Yu Cheng, and Jingjing Liu. 2020. Uniter: Universal image-text representation learning. In ECCV. 104--120."},{"key":"e_1_3_2_1_50_1","volume-title":"Neural machine translation by jointly learning to align and translate. ICLR","author":"Bahdanau Dzmitry","year":"2015","unstructured":"Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Bengio. 2015. Neural machine translation by jointly learning to align and translate. ICLR (2015)."},{"key":"e_1_3_2_1_51_1","unstructured":"Kaiming He Haoqi Fan Yuxin Wu Saining Xie and Ross Girshick. 2020. Momentum contrast for unsupervised visual representation learning. In CVPR. 9726--9735."},{"key":"e_1_3_2_1_52_1","volume-title":"Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692","author":"Liu Yinhan","year":"2019","unstructured":"Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692 (2019)."}],"event":{"name":"ICMR '25: International Conference on Multimedia Retrieval","sponsor":["SIGMM ACM Special Interest Group on Multimedia"],"location":"Chicago IL USA","acronym":"ICMR '25"},"container-title":["Proceedings of the 2025 International Conference on Multimedia Retrieval"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3731715.3733293","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,21]],"date-time":"2025-08-21T04:02:32Z","timestamp":1755748952000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3731715.3733293"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,6,30]]},"references-count":52,"alternative-id":["10.1145\/3731715.3733293","10.1145\/3731715"],"URL":"https:\/\/doi.org\/10.1145\/3731715.3733293","relation":{},"subject":[],"published":{"date-parts":[[2025,6,30]]},"assertion":[{"value":"2025-06-30","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}