{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,19]],"date-time":"2025-12-19T10:08:55Z","timestamp":1766138935853,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":67,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,10,28]],"date-time":"2024-10-28T00:00:00Z","timestamp":1730073600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"DOI":"10.13039\/https:\/\/doi.org\/10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["U21A20390"],"award-info":[{"award-number":["U21A20390"]}],"id":[{"id":"10.13039\/https:\/\/doi.org\/10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Development Project of Jilin Province of China","award":["20240601039RC"],"award-info":[{"award-number":["20240601039RC"]}]},{"name":"Fundamental Research Funds for the Central University, JLU"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,10,28]]},"DOI":"10.1145\/3664647.3681219","type":"proceedings-article","created":{"date-parts":[[2024,10,26]],"date-time":"2024-10-26T06:59:33Z","timestamp":1729925973000},"page":"1148-1157","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":5,"title":["Caption-Aware Multimodal Relation Extraction with Mutual Information Maximization"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-5627-050X","authenticated-orcid":false,"given":"Zefan","family":"Zhang","sequence":"first","affiliation":[{"name":"College of Computer Science and Technology, Jilin University &amp; Key Laboratory of Symbolic Computation and Knowledge Engineering, Ministry of Education, Jilin University, Changchun, Jilin, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0008-0225-1600","authenticated-orcid":false,"given":"Weiqi","family":"Zhang","sequence":"additional","affiliation":[{"name":"College of Computer Science and Technology, Jilin University &amp; Key Laboratory of Symbolic Computation and Knowledge Engineering, Ministry of Education, Jilin University, Changchun, Jilin, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0006-5270-9814","authenticated-orcid":false,"given":"Yanhui","family":"Li","sequence":"additional","affiliation":[{"name":"College of Computer Science and Technology, Jilin University &amp; Key Laboratory of Symbolic Computation and Knowledge Engineering, Ministry of Education, Jilin University, Changchun, Jilin, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8060-4725","authenticated-orcid":false,"given":"Tian","family":"Bai","sequence":"additional","affiliation":[{"name":"College of Computer Science and Technology, Jilin University &amp; Key Laboratory of Symbolic Computation and Knowledge Engineering, Ministry of Education, Jilin University, Changchun, Jilin, China"}]}],"member":"320","published-online":{"date-parts":[[2024,10,28]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"Deep variational information bottleneck. arXiv preprint arXiv:1612.00410","author":"Alemi Alexander A","year":"2016","unstructured":"Alexander A Alemi, Ian Fischer, Joshua V Dillon, and Kevin Murphy. 2016. Deep variational information bottleneck. arXiv preprint arXiv:1612.00410 (2016)."},{"key":"e_1_3_2_1_2_1","volume-title":"Learning representations for neural network-based classification using the information bottleneck principle","author":"Amjad Rana Ali","year":"2019","unstructured":"Rana Ali Amjad and Bernhard C Geiger. 2019. Learning representations for neural network-based classification using the information bottleneck principle. IEEE transactions on pattern analysis and machine intelligence, Vol. 42, 9 (2019), 2225--2239."},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICDAR.2019.00061"},{"key":"e_1_3_2_1_4_1","unstructured":"Jinze Bai Shuai Bai Yunfei Chu Zeyu Cui Kai Dang Xiaodong Deng Yang Fan Wenbin Ge Yu Han Fei Huang et al. 2023. Qwen technical report. arXiv preprint arXiv:2309.16609 (2023)."},{"key":"e_1_3_2_1_5_1","volume-title":"International conference on machine learning. PMLR, 531--540","author":"Belghazi Mohamed Ishmael","year":"2018","unstructured":"Mohamed Ishmael Belghazi, Aristide Baratin, Sai Rajeshwar, Sherjil Ozair, Yoshua Bengio, Aaron Courville, and Devon Hjelm. 2018. Mutual information neural estimation. In International conference on machine learning. PMLR, 531--540."},{"key":"e_1_3_2_1_6_1","volume-title":"Good visual guidance makes a better extractor: Hierarchical visual prefix for multimodal entity and relation extraction. arXiv preprint arXiv:2205.03521","author":"Chen Xiang","year":"2022","unstructured":"Xiang Chen, Ningyu Zhang, Lei Li, Yunzhi Yao, Shumin Deng, Chuanqi Tan, Fei Huang, Luo Si, and Huajun Chen. 2022. Good visual guidance makes a better extractor: Hierarchical visual prefix for multimodal entity and relation extraction. arXiv preprint arXiv:2205.03521 (2022)."},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1145\/3485447.3511998"},{"key":"e_1_3_2_1_8_1","volume-title":"International conference on machine learning. PMLR, 1779--1788","author":"Cheng Pengyu","year":"2020","unstructured":"Pengyu Cheng, Weituo Hao, Shuyang Dai, Jiachang Liu, Zhe Gan, and Lawrence Carin. 2020. Club: A contrastive log-ratio upper bound of mutual information. In International conference on machine learning. PMLR, 1779--1788."},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1109\/TASLP.2023.3345146"},{"key":"e_1_3_2_1_10_1","volume-title":"Junqi Zhao, Weisheng Wang, Boyang Li, Pascale Fung, and Steven Hoi.","author":"Dai Wenliang","year":"2023","unstructured":"Wenliang Dai, Junnan Li, Dongxu Li, Anthony Meng Huat Tiong, Junqi Zhao, Weisheng Wang, Boyang Li, Pascale Fung, and Steven Hoi. 2023. InstructBLIP: Towards General-purpose Vision-Language Models with Instruction Tuning. arxiv: 2305.06500 [cs.CV] https:\/\/arxiv.org\/abs\/2305.06500"},{"key":"e_1_3_2_1_11_1","volume-title":"Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805","author":"Devlin Jacob","year":"2018","unstructured":"Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018)."},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i1.25127"},{"key":"e_1_3_2_1_13_1","volume-title":"Proceedings of the thirteenth international conference on artificial intelligence and statistics. JMLR Workshop and Conference Proceedings, 297--304","author":"Gutmann Michael","year":"2010","unstructured":"Michael Gutmann and Aapo Hyv\u00e4rinen. 2010. Noise-contrastive estimation: A new estimation principle for unnormalized statistical models. In Proceedings of the thirteenth international conference on artificial intelligence and statistics. JMLR Workshop and Conference Proceedings, 297--304."},{"key":"e_1_3_2_1_14_1","volume-title":"Improving multimodal fusion with hierarchical mutual information maximization for multimodal sentiment analysis. arXiv preprint arXiv:2109.00412","author":"Han Wei","year":"2021","unstructured":"Wei Han, Hui Chen, and Soujanya Poria. 2021. Improving multimodal fusion with hierarchical mutual information maximization for multimodal sentiment analysis. arXiv preprint arXiv:2109.00412 (2021)."},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00975"},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"e_1_3_2_1_17_1","volume-title":"2023 f. Large Multilingual Models Pivot Zero-Shot Multimodal Learning across Languages. arXiv preprint arXiv:2308.12038","author":"Hu Jinyi","year":"2023","unstructured":"Jinyi Hu, Yuan Yao, Chongyi Wang, Shan Wang, Yinxu Pan, Qianyu Chen, Tianyu Yu, Hanghao Wu, Yue Zhao, Haoye Zhang, Xu Han, Yankai Lin, Jiao Xue, Dahai Li, Zhiyuan Liu, and Maosong Sun. 2023 f. Large Multilingual Models Pivot Zero-Shot Multimodal Learning across Languages. arXiv preprint arXiv:2308.12038 (2023)."},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1145\/3581783.3611899"},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1145\/3539618.3592058"},{"key":"e_1_3_2_1_20_1","volume-title":"Multimodal relation extraction with cross-modal retrieval and synthesis. arXiv preprint arXiv:2305.16166","author":"Hu Xuming","year":"2023","unstructured":"Xuming Hu, Zhijiang Guo, Zhiyang Teng, Irwin King, and Philip S Yu. 2023. Multimodal relation extraction with cross-modal retrieval and synthesis. arXiv preprint arXiv:2305.16166 (2023)."},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1145\/3539618.3592072"},{"key":"e_1_3_2_1_22_1","volume-title":"Entity-to-text based data augmentation for various named entity recognition tasks. arXiv preprint arXiv:2210.10343","author":"Hu Xuming","year":"2022","unstructured":"Xuming Hu, Yong Jiang, Aiwei Liu, Zhongqiang Huang, Pengjun Xie, Fei Huang, Lijie Wen, and Philip S Yu. 2022. Entity-to-text based data augmentation for various named entity recognition tasks. arXiv preprint arXiv:2210.10343 (2022)."},{"key":"e_1_3_2_1_23_1","volume-title":"2023 e. Gda: Generative data augmentation techniques for relation extraction tasks. arXiv preprint arXiv:2305.16663","author":"Hu Xuming","year":"2023","unstructured":"Xuming Hu, Aiwei Liu, Zeqi Tan, Xin Zhang, Chenwei Zhang, Irwin King, and Philip S Yu. 2023 e. Gda: Generative data augmentation techniques for relation extraction tasks. arXiv preprint arXiv:2305.16663 (2023)."},{"key":"e_1_3_2_1_24_1","volume-title":"Hiure: Hierarchical exemplar contrastive learning for unsupervised relation extraction. arXiv preprint arXiv:2205.02225","author":"Hu Xuming","year":"2022","unstructured":"Xuming Hu, Shuliang Liu, Chenwei Zhang, Shuang Li, Lijie Wen, and Philip S Yu. 2022. Hiure: Hierarchical exemplar contrastive learning for unsupervised relation extraction. arXiv preprint arXiv:2205.02225 (2022)."},{"key":"e_1_3_2_1_25_1","volume-title":"Semi-supervised relation extraction via incremental meta self-training. arXiv preprint arXiv:2010.16410","author":"Hu Xuming","year":"2020","unstructured":"Xuming Hu, Chenwei Zhang, Fukun Ma, Chenyao Liu, Lijie Wen, and Philip S Yu. 2020. Semi-supervised relation extraction via incremental meta self-training. arXiv preprint arXiv:2010.16410 (2020)."},{"key":"e_1_3_2_1_26_1","volume-title":"SelfORE: Self-supervised relational feature learning for open relation extraction. arXiv preprint arXiv:2004.02438","author":"Hu Xuming","year":"2020","unstructured":"Xuming Hu, Chenwei Zhang, Yusong Xu, Lijie Wen, and Philip S Yu. 2020. SelfORE: Self-supervised relational feature learning for open relation extraction. arXiv preprint arXiv:2004.02438 (2020)."},{"key":"e_1_3_2_1_27_1","volume-title":"Gradient imitation reinforcement learning for low resource relation extraction. arXiv preprint arXiv:2109.06415","author":"Hu Xuming","year":"2021","unstructured":"Xuming Hu, Chenwei Zhang, Yawen Yang, Xiaohe Li, Li Lin, Lijie Wen, and Philip S Yu. 2021. Gradient imitation reinforcement learning for low resource relation extraction. arXiv preprint arXiv:2109.06415 (2021)."},{"key":"e_1_3_2_1_28_1","unstructured":"Bo Li Hao Zhang Kaichen Zhang Dong Guo Yuanhan Zhang Renrui Zhang Feng Li Ziwei Liu and Chunyuan Li. 2024. LLaVA-NeXT: What Else Influences Visual Instruction Tuning Beyond Data? https:\/\/llava-vl.github.io\/blog\/2024-05--25-llava-next-ablations\/"},{"key":"e_1_3_2_1_29_1","unstructured":"Junnan Li Dongxu Li Silvio Savarese and Steven Hoi. 2023. BLIP-2: Bootstrapping Language-Image Pre-training with Frozen Image Encoders and Large Language Models. arxiv: 2301.12597 [cs.CV] https:\/\/arxiv.org\/abs\/2301.12597"},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i13.26987"},{"key":"e_1_3_2_1_31_1","volume-title":"GCDT: A global context enhanced deep transition architecture for sequence labeling. arXiv preprint arXiv:1906.02437","author":"Liu Yijin","year":"2019","unstructured":"Yijin Liu, Fandong Meng, Jinchao Zhang, Jinan Xu, Yufeng Chen, and Jie Zhou. 2019. GCDT: A global context enhanced deep transition architecture for sequence labeling. arXiv preprint arXiv:1906.02437 (2019)."},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2022\/171"},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P18-1185"},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2023.3298356"},{"key":"e_1_3_2_1_35_1","volume-title":"Multimodal named entity recognition for short social media posts. arXiv preprint arXiv:1802.07862","author":"Moon Seungwhan","year":"2018","unstructured":"Seungwhan Moon, Leonardo Neves, and Vitor Carvalho. 2018. Multimodal named entity recognition for short social media posts. arXiv preprint arXiv:1802.07862 (2018)."},{"key":"e_1_3_2_1_36_1","volume-title":"Duc-Trong Le, and Quang-Thuy Ha.","author":"Thi Nguyen Cam-Van","year":"2023","unstructured":"Cam-Van Thi Nguyen, Ngoc-Hoa Thi Nguyen, Duc-Trong Le, and Quang-Thuy Ha. 2023. Self-MI: Efficient Multimodal Fusion via Self-Supervised Multi-Task Learning with Auxiliary Mutual Information Maximization. arXiv preprint arXiv:2311.03785 (2023)."},{"key":"e_1_3_2_1_37_1","volume-title":"Representation learning with contrastive predictive coding. arXiv preprint arXiv:1807.03748","author":"van den Oord Aaron","year":"2018","unstructured":"Aaron van den Oord, Yazhe Li, and Oriol Vinyals. 2018. Representation learning with contrastive predictive coding. arXiv preprint arXiv:1807.03748 (2018)."},{"key":"e_1_3_2_1_38_1","volume-title":"International Conference on Machine Learning. PMLR, 5171--5180","author":"Poole Ben","year":"2019","unstructured":"Ben Poole, Sherjil Ozair, Aaron Van Den Oord, Alex Alemi, and George Tucker. 2019. On variational bounds of mutual information. In International Conference on Machine Learning. PMLR, 5171--5180."},{"key":"e_1_3_2_1_39_1","volume-title":"ERICA: Improving entity and relation understanding for pre-trained language models via contrastive learning. arXiv preprint arXiv:2012.15022","author":"Qin Yujia","year":"2020","unstructured":"Yujia Qin, Yankai Lin, Ryuichi Takanobu, Zhiyuan Liu, Peng Li, Heng Ji, Minlie Huang, Maosong Sun, and Jie Zhou. 2020. ERICA: Improving entity and relation understanding for pre-trained language models via contrastive learning. arXiv preprint arXiv:2012.15022 (2020)."},{"key":"e_1_3_2_1_40_1","volume-title":"Complementary Information Mutual Learning for Multimodality Medical Image Segmentation. arXiv preprint arXiv:2401.02717","author":"Shen Chuyun","year":"2024","unstructured":"Chuyun Shen, Wenhao Li, Haoqing Chen, Xiaoling Wang, Fengping Zhu, Yuxin Li, Xiangfeng Wang, and Bo Jin. 2024. Complementary Information Mutual Learning for Multimodality Medical Image Segmentation. arXiv preprint arXiv:2401.02717 (2024)."},{"key":"e_1_3_2_1_41_1","volume-title":"Matching the blanks: Distributional similarity for relation learning. arXiv preprint arXiv:1906.03158","author":"Soares Livio Baldini","year":"2019","unstructured":"Livio Baldini Soares, Nicholas FitzGerald, Jeffrey Ling, and Tom Kwiatkowski. 2019. Matching the blanks: Distributional similarity for relation learning. arXiv preprint arXiv:1906.03158 (2019)."},{"key":"e_1_3_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i15.17633"},{"key":"e_1_3_2_1_43_1","doi-asserted-by":"publisher","DOI":"10.1145\/3581783.3612449"},{"volume-title":"Deep learning and the information bottleneck principle. In 2015 ieee information theory workshop (itw)","author":"Tishby Naftali","key":"e_1_3_2_1_44_1","unstructured":"Naftali Tishby and Noga Zaslavsky. 2015. Deep learning and the information bottleneck principle. In 2015 ieee information theory workshop (itw). IEEE, 1--5."},{"key":"e_1_3_2_1_45_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i05.6445"},{"key":"e_1_3_2_1_46_1","doi-asserted-by":"publisher","DOI":"10.1145\/3503161.3548324"},{"key":"e_1_3_2_1_47_1","unstructured":"Weihan Wang Qingsong Lv Wenmeng Yu Wenyi Hong Ji Qi Yan Wang Junhui Ji Zhuoyi Yang Lei Zhao Xixuan Song Jiazheng Xu Bin Xu Juanzi Li Yuxiao Dong Ming Ding and Jie Tang. 2023. CogVLM: Visual Expert for Pretrained Language Models. arxiv: 2311.03079 [cs.CV]"},{"key":"e_1_3_2_1_48_1","volume-title":"Named entity and relation extraction with multi-modal retrieval. arXiv preprint arXiv:2212.01612","author":"Wang Xinyu","year":"2022","unstructured":"Xinyu Wang, Jiong Cai, Yong Jiang, Pengjun Xie, Kewei Tu, and Wei Lu. 2022. Named entity and relation extraction with multi-modal retrieval. arXiv preprint arXiv:2212.01612 (2022)."},{"key":"e_1_3_2_1_49_1","volume-title":"Information screening whilst exploiting! multimodal relation extraction with feature denoising and multimodal topic modeling. arXiv preprint arXiv:2305.11719","author":"Wu Shengqiong","year":"2023","unstructured":"Shengqiong Wu, Hao Fei, Yixin Cao, Lidong Bing, and Tat-Seng Chua. 2023. Information screening whilst exploiting! multimodal relation extraction with feature denoising and multimodal topic modeling. arXiv preprint arXiv:2305.11719 (2023)."},{"key":"e_1_3_2_1_50_1","volume-title":"Do as I Do: Pose Guided Human Motion Copy","author":"Wu Sifan","year":"2024","unstructured":"Sifan Wu, Zhenguang Liu, Beibei Zhang, Roger Zimmermann, Zhongjie Ba, Xiaosong Zhang, and Kui Ren. 2024. Do as I Do: Pose Guided Human Motion Copy. IEEE Transactions on Dependable and Secure Computing (2024)."},{"key":"e_1_3_2_1_51_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00478"},{"key":"e_1_3_2_1_52_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00867"},{"key":"e_1_3_2_1_53_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP48485.2024.10447410"},{"key":"e_1_3_2_1_54_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52733.2024.01334"},{"volume-title":"Improving multimodal named entity recognition via entity span detection with unified multimodal transformer","author":"Yu Jianfei","key":"e_1_3_2_1_55_1","unstructured":"Jianfei Yu, Jing Jiang, Li Yang, and Rui Xia. 2020. Improving multimodal named entity recognition via entity span detection with unified multimodal transformer. Association for Computational Linguistics."},{"key":"e_1_3_2_1_56_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D15-1203"},{"key":"e_1_3_2_1_57_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i16.17687"},{"key":"e_1_3_2_1_58_1","volume-title":"Document-level relation extraction as semantic segmentation. arXiv preprint arXiv:2106.03618","author":"Zhang Ningyu","year":"2021","unstructured":"Ningyu Zhang, Xiang Chen, Xin Xie, Shumin Deng, Chuanqi Tan, Mosha Chen, Fei Huang, Luo Si, and Huajun Chen. 2021. Document-level relation extraction as semantic segmentation. arXiv preprint arXiv:2106.03618 (2021)."},{"key":"e_1_3_2_1_59_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-demos.1"},{"key":"e_1_3_2_1_60_1","doi-asserted-by":"publisher","DOI":"10.1145\/3447548.3467057"},{"key":"e_1_3_2_1_61_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v32i1.11962"},{"key":"e_1_3_2_1_62_1","doi-asserted-by":"publisher","DOI":"10.1145\/3591106.3592272"},{"key":"e_1_3_2_1_63_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ipm.2023.103264"},{"key":"e_1_3_2_1_64_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.acl-long.376"},{"key":"e_1_3_2_1_65_1","doi-asserted-by":"publisher","DOI":"10.1145\/3474085.3476968"},{"key":"e_1_3_2_1_66_1","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2023.3297842"},{"key":"e_1_3_2_1_67_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICME51207.2021.9428274"}],"event":{"name":"MM '24: The 32nd ACM International Conference on Multimedia","sponsor":["SIGMM ACM Special Interest Group on Multimedia"],"location":"Melbourne VIC Australia","acronym":"MM '24"},"container-title":["Proceedings of the 32nd ACM International Conference on Multimedia"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3664647.3681219","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3664647.3681219","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T01:18:02Z","timestamp":1750295882000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3664647.3681219"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,28]]},"references-count":67,"alternative-id":["10.1145\/3664647.3681219","10.1145\/3664647"],"URL":"https:\/\/doi.org\/10.1145\/3664647.3681219","relation":{},"subject":[],"published":{"date-parts":[[2024,10,28]]},"assertion":[{"value":"2024-10-28","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}