{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,21]],"date-time":"2025-08-21T18:50:09Z","timestamp":1755802209604,"version":"3.44.0"},"publisher-location":"New York, NY, USA","reference-count":85,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,5,30]],"date-time":"2024-05-30T00:00:00Z","timestamp":1717027200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,5,30]]},"DOI":"10.1145\/3652583.3658010","type":"proceedings-article","created":{"date-parts":[[2024,6,7]],"date-time":"2024-06-07T06:30:40Z","timestamp":1717741840000},"page":"785-794","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["Pattern4Ego: Learning Egocentric Video Representation Using Cross-video Activity Patterns"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0008-5767-7529","authenticated-orcid":false,"given":"Ruihai","family":"Wu","sequence":"first","affiliation":[{"name":"CFCS, School of CS, Peking University, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0007-8491-4211","authenticated-orcid":false,"given":"Yourong","family":"Zhang","sequence":"additional","affiliation":[{"name":"Peking University, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-8850-4397","authenticated-orcid":false,"given":"Yu","family":"Qi","sequence":"additional","affiliation":[{"name":"Peking University &amp; Northeastern University, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0006-9431-6461","authenticated-orcid":false,"given":"Andy Guanhong","family":"Chen","sequence":"additional","affiliation":[{"name":"Peking University, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7984-9909","authenticated-orcid":false,"given":"Hao","family":"Dong","sequence":"additional","affiliation":[{"name":"CFCS, School of CS, Peking University, Beijing, China"}]}],"member":"320","published-online":{"date-parts":[[2024,6,7]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"crossref","unstructured":"Mehmet Ali Arabaci Elif Surer and Alptekin Temizel. 2023. Egocentric Activity Recognition Using Two-Stage Decision Fusion. (2023).","DOI":"10.21203\/rs.3.rs-3024911\/v1"},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01261-8_7"},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00994"},{"key":"e_1_3_2_1_4_1","volume-title":"Kristof Van Laerhoven, and Hilde Kuehne","author":"Bock Marius","year":"2023","unstructured":"Marius Bock, Michael Moeller, Kristof Van Laerhoven, and Hilde Kuehne. 2023. WEAR: A Multimodal Dataset for Wearable and Egocentric Video Activity Recognition. arXiv preprint arXiv:2304.05088 (2023)."},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.15607\/RSS.2016.XII.034"},{"key":"e_1_3_2_1_6_1","volume-title":"Generalizing Hand Segmentation in Egocentric Videos With Uncertainty-Guided Model Adaptation. In IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR).","author":"Cai Minjie","year":"2020","unstructured":"Minjie Cai, Feng Lu, and Yoichi Sato. 2020. Generalizing Hand Segmentation in Egocentric Videos With Uncertainty-Guided Model Adaptation. In IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR)."},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.502"},{"key":"e_1_3_2_1_8_1","volume-title":"Caglar Gulcehre, Dzmitry Bahdanau, Fethi Bougares, Holger Schwenk, and Yoshua Bengio.","author":"Cho Kyunghyun","year":"2014","unstructured":"Kyunghyun Cho, Bart Van Merri\u00ebnboer, Caglar Gulcehre, Dzmitry Bahdanau, Fethi Bougares, Holger Schwenk, and Yoshua Bengio. 2014. Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv preprint arXiv:1406.1078 (2014)."},{"key":"e_1_3_2_1_9_1","volume-title":"Proceedings of the European Conference on Computer Vision (ECCV). 720--736","author":"Damen Dima","year":"2018","unstructured":"Dima Damen, Hazel Doughty, Giovanni Maria Farinella, Sanja Fidler, Antonino Furnari, Evangelos Kazakos, Davide Moltisanti, Jonathan Munro, Toby Perrett, Will Price, et al. 2018. Scaling egocentric vision: The epic-kitchens dataset. In Proceedings of the European Conference on Computer Vision (ECCV). 720--736."},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.cviu.2016.02.016"},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58536-5_13"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00190"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00630"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.jvcir.2017.10.004"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00635"},{"key":"e_1_3_2_1_16_1","volume-title":"Red: Reinforced encoder-decoder networks for action anticipation. arXiv preprint arXiv:1707.04818","author":"Gao Jiyang","year":"2017","unstructured":"Jiyang Gao, Zhenheng Yang, and Ram Nevatia. 2017. Red: Reinforced encoder-decoder networks for action anticipation. arXiv preprint arXiv:1707.04818 (2017)."},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.337"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00627"},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"crossref","unstructured":"Kristen Grauman Andrew Westbury Eugene Byrne Zachary Chavis Antonino Furnari Rohit Girdhar Jackson Hamburger Hao Jiang Miao Liu Xingyu Liu Miguel Martin Tushar Nagarajan Ilija Radosavovic Santhosh Kumar Ramakrishnan Fiona Ryan Jayant Sharma Michael Wray Mengmeng Xu Eric Zhongcong Xu Chen Zhao Siddhant Bansal Dhruv Batra Vincent Cartillier Sean Crane Tien Do Morrie Doulaty Akshay Erapalli Christoph Feichtenhofer Adriano Fragomeni Qichen Fu Christian Fuegen Abrham Gebreselasie Cristina Gonzalez James Hillis Xuhua Huang Yifei Huang Wenqi Jia Weslie Khoo Jachym Kolar Satwik Kottur Anurag Kumar Federico Landini Chao Li Yanghao Li Zhenqiang Li Karttikeya Mangalam Raghava Modhugu Jonathan Munro Tullie Murrell Takumi Nishiyasu Will Price Paola Ruiz Puentes Merey Ramazanova Leda Sari Kiran Somasundaram Audrey Southerland Yusuke Sugano Ruijie Tao Minh Vo Yuchen Wang Xindi Wu Takuma Yagi Yunyi Zhu Pablo Arbelaez David Crandall Dima Damen Giovanni Maria Farinella Bernard Ghanem Vamsi Krishna Ithapu C. V. Jawahar Hanbyul Joo Kris Kitani Haizhou Li Richard Newcombe Aude Oliva Hyun Soo Park James M. Rehg Yoichi Sato Jianbo Shi Mike Zheng Shou Antonio Torralba Lorenzo Torresani Mingfei Yan and Jitendra Malik. 2021. Ego4D: Around the World in 3 000 Hours of Egocentric Video. CoRR Vol. abs\/2110.07058 (2021). showeprint[arXiv]2110.07058 https:\/\/arxiv.org\/abs\/2110.07058","DOI":"10.1109\/CVPR52688.2022.01842"},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"crossref","unstructured":"Kristen Grauman Andrew Westbury Lorenzo Torresani Kris Kitani Jitendra Malik Triantafyllos Afouras Kumar Ashutosh Vijay Baiyya Siddhant Bansal Bikram Boote et al. 2023. Ego-exo4d: Understanding skilled human activity from first-and third-person perspectives. arXiv preprint arXiv:2311.18259 (2023).","DOI":"10.1109\/CVPR52733.2024.01834"},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00025"},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCVW.2019.00186"},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58580-8_19"},{"key":"e_1_3_2_1_24_1","volume-title":"Long short-term memory. Neural computation","author":"Hochreiter Sepp","year":"1997","unstructured":"Sepp Hochreiter and J\u00fcrgen Schmidhuber. 1997. Long short-term memory. Neural computation, Vol. 9, 8 (1997), 1735--1780."},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.464"},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.01426"},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01225-0_46"},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00034"},{"key":"e_1_3_2_1_29_1","volume-title":"Videograph: Recognizing minutes-long human activities in videos. arXiv preprint arXiv:1905.05143","author":"Hussein Noureldien","year":"2019","unstructured":"Noureldien Hussein, Efstratios Gavves, and Arnold WM Smeulders. 2019b. Videograph: Recognizing minutes-long human activities in videos. arXiv preprint arXiv:1905.05143 (2019)."},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.373"},{"key":"e_1_3_2_1_31_1","volume-title":"Egocentric Pose Estimation from Human Vision Span. arXiv preprint arXiv:2104.05167","author":"Jiang Hao","year":"2021","unstructured":"Hao Jiang and Vamsi Krishna Ithapu. 2021. Egocentric Pose Estimation from Human Vision Span. arXiv preprint arXiv:2104.05167 (2021)."},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00559"},{"key":"e_1_3_2_1_33_1","volume-title":"Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907","author":"Kipf Thomas N","year":"2016","unstructured":"Thomas N Kipf and Max Welling. 2016. Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907 (2016)."},{"key":"e_1_3_2_1_34_1","volume-title":"AI2-THOR: An Interactive 3D Environment for Visual AI. arXiv","author":"Kolve Eric","year":"2017","unstructured":"Eric Kolve, Roozbeh Mottaghi, Winson Han, Eli VanderBilt, Luca Weihs, Alvaro Herrasti, Daniel Gordon, Yuke Zhu, Abhinav Gupta, and Ali Farhadi. 2017. AI2-THOR: An Interactive 3D Environment for Visual AI. arXiv (2017)."},{"key":"e_1_3_2_1_35_1","unstructured":"Bolin Lai Miao Liu Fiona Ryan and James Rehg. 2022. In the Eye of Transformer: Global-Local Correlation for Egocentric Gaze Estimation. arXiv preprint arXiv:2208.04464 (2022)."},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-014-0794-5"},{"key":"e_1_3_2_1_37_1","volume-title":"Exploiting Multimodal Synthetic Data for Egocentric Human-Object Interaction Detection in an Industrial Scenario. arXiv preprint arXiv:2306.12152","author":"Leonardi Rosario","year":"2023","unstructured":"Rosario Leonardi, Francesco Ragusa, Antonino Furnari, and Giovanni Maria Farinella. 2023. Exploiting Multimodal Synthetic Data for Egocentric Human-Object Interaction Detection in an Industrial Scenario. arXiv preprint arXiv:2306.12152 (2023)."},{"key":"e_1_3_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00379"},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1109\/WACV51458.2022.00090"},{"key":"e_1_3_2_1_40_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2013.399"},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00099"},{"key":"e_1_3_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01228-1_38"},{"key":"e_1_3_2_1_43_1","volume-title":"Deep predictive coding networks for video prediction and unsupervised learning. arXiv preprint arXiv:1605.08104","author":"Lotter William","year":"2016","unstructured":"William Lotter, Gabriel Kreiman, and David Cox. 2016. Deep predictive coding networks for video prediction and unsupervised learning. arXiv preprint arXiv:1605.08104 (2016)."},{"key":"e_1_3_2_1_44_1","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2015.2487868"},{"key":"e_1_3_2_1_45_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2013.350"},{"key":"e_1_3_2_1_46_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.00229"},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00710"},{"key":"e_1_3_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00441"},{"key":"e_1_3_2_1_49_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00878"},{"key":"e_1_3_2_1_50_1","unstructured":"Tushar Nagarajan and Kristen Grauman. 2020. Learning Affordance Landscapes for Interaction Exploration in 3D Environments. In NeurIPS."},{"key":"e_1_3_2_1_51_1","doi-asserted-by":"crossref","unstructured":"Tushar Nagarajan Yanghao Li Christoph Feichtenhofer and Kristen Grauman. 2020. EGO-TOPO: Environment Affordances from Egocentric Video. In CVPR.","DOI":"10.1109\/CVPR42600.2020.00024"},{"key":"e_1_3_2_1_52_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00991"},{"key":"e_1_3_2_1_53_1","unstructured":"Chuanruo Ning Ruihai Wu Haoran Lu Kaichun Mo and Hao Dong. 2023. Where2Explore: Few-shot Affordance Learning for Unseen Novel Categories of Articulated Objects. In Advances in Neural Information Processing Systems (NeurIPS)."},{"key":"e_1_3_2_1_54_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.01249"},{"key":"e_1_3_2_1_55_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.508"},{"key":"e_1_3_2_1_56_1","volume-title":"Anticipation and next action forecasting in video: an end-to-end model with memory. arXiv preprint arXiv:1901.03728","author":"Pirri Fiora","year":"2019","unstructured":"Fiora Pirri, Lorenzo Mauro, Edoardo Alati, Valsamis Ntouskos, Mahdieh Izadpanahkakhk, and Elham Omrani. 2019. Anticipation and next action forecasting in video: an end-to-end model with memory. arXiv preprint arXiv:1901.03728 (2019)."},{"key":"e_1_3_2_1_57_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2012.6248010"},{"key":"e_1_3_2_1_58_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00481"},{"key":"e_1_3_2_1_59_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00098"},{"key":"e_1_3_2_1_60_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.399"},{"key":"e_1_3_2_1_61_1","doi-asserted-by":"publisher","DOI":"10.5121\/ijcses.2011.2408"},{"key":"e_1_3_2_1_62_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01249-6_19"},{"key":"e_1_3_2_1_63_1","volume-title":"Charades-ego: A large-scale dataset of paired third and first person videos. arXiv preprint arXiv:1804.09626","author":"Sigurdsson Gunnar A","year":"2018","unstructured":"Gunnar A Sigurdsson, Abhinav Gupta, Cordelia Schmid, Ali Farhadi, and Karteek Alahari. 2018. Charades-ego: A large-scale dataset of paired third and first person videos. arXiv preprint arXiv:1804.09626 (2018)."},{"key":"e_1_3_2_1_64_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58555-6_5"},{"key":"e_1_3_2_1_65_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00232"},{"key":"e_1_3_2_1_66_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00782"},{"key":"e_1_3_2_1_67_1","article-title":"Visualizing data using t-SNE","volume":"9","author":"der Maaten Laurens Van","year":"2008","unstructured":"Laurens Van der Maaten and Geoffrey Hinton. 2008. Visualizing data using t-SNE. Journal of machine learning research, Vol. 9, 11 (2008).","journal-title":"Journal of machine learning research"},{"key":"e_1_3_2_1_68_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00413"},{"key":"e_1_3_2_1_69_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58520-4_30"},{"key":"e_1_3_2_1_70_1","volume-title":"Estimating Egocentric 3D Human Pose in Global Space. arXiv preprint arXiv:2104.13454","author":"Wang Jian","year":"2021","unstructured":"Jian Wang, Lingjie Liu, Weipeng Xu, Kripasindhu Sarkar, and Christian Theobalt. 2021a. Estimating Egocentric 3D Human Pose in Global Space. arXiv preprint arXiv:2104.13454 (2021)."},{"key":"e_1_3_2_1_71_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.01252"},{"key":"e_1_3_2_1_72_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01228-1_25"},{"key":"e_1_3_2_1_73_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.01854"},{"key":"e_1_3_2_1_74_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00806"},{"key":"e_1_3_2_1_75_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-19818-2_6"},{"key":"e_1_3_2_1_76_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00999"},{"key":"e_1_3_2_1_77_1","unstructured":"Ruihai Wu Kai Cheng Yan Shen Chuanruo Ning Guanqi Zhan and Hao Dong. 2023 a. Learning Environment-Aware Affordance for 3D Articulated Object Manipulation under Occlusions. In Advances in Neural Information Processing Systems (NeurIPS)."},{"key":"e_1_3_2_1_78_1","volume-title":"Learning Foresightful Dense Visual Affordance for Deformable Object Manipulation. In IEEE International Conference on Computer Vision (ICCV).","author":"Wu Ruihai","year":"2023","unstructured":"Ruihai Wu, Chuanruo Ning, and Hao Dong. 2023 b. Learning Foresightful Dense Visual Affordance for Deformable Object Manipulation. In IEEE International Conference on Computer Vision (ICCV)."},{"key":"e_1_3_2_1_79_1","volume-title":"International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=iEx3PiooLy","author":"Wu Ruihai","year":"2022","unstructured":"Ruihai Wu, Yan Zhao, Kaichun Mo, Zizheng Guo, Yian Wang, Tianhao Wu, Qingnan Fan, Xuelin Chen, Leonidas Guibas, and Hao Dong. 2022. VAT-Mart: Learning Visual Action Trajectory Proposals for Manipulating 3D ARTiculated Objects. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=iEx3PiooLy"},{"key":"e_1_3_2_1_80_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.01001"},{"key":"e_1_3_2_1_81_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46475-6_12"},{"key":"e_1_3_2_1_82_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-19818-2_8"},{"key":"e_1_3_2_1_83_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.01021"},{"key":"e_1_3_2_1_84_1","volume-title":"DualAfford: Learning Collaborative Visual Affordance for Dual-gripper Object Manipulation. International Conference on Learning Representations (ICLR)","author":"Zhao Yan","year":"2023","unstructured":"Yan Zhao, Ruihai Wu, Zhehuan Chen, Yourong Zhang, Qingnan Fan, Kaichun Mo, and Hao Dong. 2023. DualAfford: Learning Collaborative Visual Affordance for Dual-gripper Object Manipulation. International Conference on Learning Representations (ICLR) (2023)."},{"key":"e_1_3_2_1_85_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.511"}],"event":{"name":"ICMR '24: International Conference on Multimedia Retrieval","sponsor":["SIGMM ACM Special Interest Group on Multimedia","SIGSOFT ACM Special Interest Group on Software Engineering"],"location":"Phuket Thailand","acronym":"ICMR '24"},"container-title":["Proceedings of the 2024 International Conference on Multimedia Retrieval"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3652583.3658010","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3652583.3658010","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,21]],"date-time":"2025-08-21T08:46:42Z","timestamp":1755766002000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3652583.3658010"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,5,30]]},"references-count":85,"alternative-id":["10.1145\/3652583.3658010","10.1145\/3652583"],"URL":"https:\/\/doi.org\/10.1145\/3652583.3658010","relation":{},"subject":[],"published":{"date-parts":[[2024,5,30]]},"assertion":[{"value":"2024-06-07","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}