{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,13]],"date-time":"2026-01-13T23:02:28Z","timestamp":1768345348699,"version":"3.49.0"},"reference-count":64,"publisher":"Association for Computing Machinery (ACM)","issue":"3","license":[{"start":{"date-parts":[[2025,3,10]],"date-time":"2025-03-10T00:00:00Z","timestamp":1741564800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"Exploratory Research Project of Zhejiang Lab","award":["2022PG0AN01"],"award-info":[{"award-number":["2022PG0AN01"]}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62106235, 62201538"],"award-info":[{"award-number":["62106235, 62201538"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100012166","name":"National Key R&D Program of China","doi-asserted-by":"crossref","award":["2022YFE0137800"],"award-info":[{"award-number":["2022YFE0137800"]}],"id":[{"id":"10.13039\/501100012166","id-type":"DOI","asserted-by":"crossref"}]},{"name":"Key R&D Program of Xinjiang, China","award":["2022B01006"],"award-info":[{"award-number":["2022B01006"]}]},{"DOI":"10.13039\/501100007129","name":"Natural Science Foundation of Shandong Province","doi-asserted-by":"crossref","award":["ZR2022QF006, ZR2024MF116"],"award-info":[{"award-number":["ZR2022QF006, ZR2024MF116"]}],"id":[{"id":"10.13039\/501100007129","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["ACM Trans. Multimedia Comput. Commun. Appl."],"published-print":{"date-parts":[[2025,3,31]]},"abstract":"<jats:p>The practical use of the Transformer-based methods for processing videos is constrained by the high computing complexity. Although previous approaches adopt the spatiotemporal decomposition of 3D attention to mitigate the issue, they suffer from the drawback of neglecting the majority of visual tokens. This article presents a novel mixed attention operation that subtly fuses the random, spatial, and temporal attention mechanisms. The proposed random attention stochastically samples video tokens in a simple yet effective way, complementing other attention methods. Furthermore, since the attention operation concentrates on learning long-distance relationships, we employ the channel shift operation to encode short-term temporal characteristics. Our model can provide more comprehensive motion representations thanks to the amalgamation of these techniques. Experimental results show that the proposed method produces competitive action recognition results with low computational overhead on both large-scale and small-scale public video datasets.<\/jats:p>","DOI":"10.1145\/3712594","type":"journal-article","created":{"date-parts":[[2025,1,17]],"date-time":"2025-01-17T16:33:41Z","timestamp":1737131621000},"page":"1-20","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":2,"title":["Mixed Attention and Channel Shift Transformer for Efficient Action Recognition"],"prefix":"10.1145","volume":"21","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-1543-5395","authenticated-orcid":false,"given":"Xiusheng","family":"Lu","sequence":"first","affiliation":[{"name":"Tsinghua University, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0695-1566","authenticated-orcid":false,"given":"Yanbin","family":"Hao","sequence":"additional","affiliation":[{"name":"University of Science and Technology of China, Hefei, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7546-9052","authenticated-orcid":false,"given":"Lechao","family":"Cheng","sequence":"additional","affiliation":[{"name":"Hefei University of Technology, Hefei, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5843-6411","authenticated-orcid":false,"given":"Sicheng","family":"Zhao","sequence":"additional","affiliation":[{"name":"Tsinghua University, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3066-1884","authenticated-orcid":false,"given":"Yutao","family":"Liu","sequence":"additional","affiliation":[{"name":"Ocean University of China, Qingdao, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2621-6048","authenticated-orcid":false,"given":"Mingli","family":"Song","sequence":"additional","affiliation":[{"name":"Zhejiang University, Hangzhou, China"}]}],"member":"320","published-online":{"date-parts":[[2025,3,10]]},"reference":[{"key":"e_1_3_1_2_2","first-page":"6836","volume-title":"Proceedings of the IEEE International Conference on Computer Vision","author":"Arnab Anurag","year":"2021","unstructured":"Anurag Arnab, Mostafa Dehghani, Georg Heigold, Chen Sun, Mario Lu\u010di\u0107, and Cordelia Schmid. 2021. ViViT: A video vision transformer. In Proceedings of the IEEE International Conference on Computer Vision, 6836\u20136846."},{"key":"e_1_3_1_3_2","unstructured":"Jimmy Lei Ba Jamie Ryan Kiros and Geoffrey E. Hinton. 2016. Layer normalization. arXiv:1607.06450. Retrieved from https:\/\/arxiv.org\/abs\/1607.06450"},{"key":"e_1_3_1_4_2","unstructured":"Iz Beltagy Matthew E. Peters and Arman Cohan. 2020. Longformer: The long-document transformer. arXiv:2004.05150. Retrieved from https:\/\/arxiv.org\/abs\/2004.05150"},{"key":"e_1_3_1_5_2","first-page":"813","volume-title":"Proceedings of the International Conference on Machine Learning","author":"Bertasius Gedas","year":"2021","unstructured":"Gedas Bertasius, Heng Wang, and Lorenzo Torresani. 2021. Is space-time attention all you need for video understanding? In Proceedings of the International Conference on Machine Learning, 813\u2013824."},{"key":"e_1_3_1_6_2","first-page":"19594","volume-title":"Proceedings of the Advances in Neural Information Processing Systems","author":"Bulat Adrian","year":"2021","unstructured":"Adrian Bulat, Juan Manuel Perez Rua, Swathikiran Sudhakaran, Martinez Brais, and Tzimiropoulos Georgios. 2021. Space-time mixing attention for video transformer. In Proceedings of the Advances in Neural Information Processing Systems, 19594\u201319607."},{"key":"e_1_3_1_7_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58452-8_13"},{"key":"e_1_3_1_8_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.502"},{"key":"e_1_3_1_9_2","first-page":"589","volume-title":"Proceedings of the IEEE International Conference on Computer Vision","author":"Chen Zhengsu","year":"2021","unstructured":"Zhengsu Chen, Lingxi Xie, Jianwei Niu, Xuefeng Liu, Longhui Wei, and Qi Tian. 2021. Visformer: The vision-friendly transformer. In Proceedings of the IEEE International Conference on Computer Vision, 589\u2013598."},{"key":"e_1_3_1_10_2","unstructured":"Jacob Devlin Ming-Wei Chang Kenton Lee and Kristina Toutanova. 2018. BERT: Pre-training of deep bidirectional transformers for language understanding. arXiv:1810.04805. Retrieved from https:\/\/arxiv.org\/abs\/1810.04805"},{"key":"e_1_3_1_11_2","unstructured":"Alexey Dosovitskiy Lucas Beyer Alexander Kolesnikov Dirk Weissenborn Xiaohua Zhai Thomas Unterthiner Mostafa Dehghani Matthias Minderer Georg Heigold Sylvain Gelly et al. 2020. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv:2010.11929. Retrieved from https:\/\/arxiv.org\/abs\/2010.11929"},{"key":"e_1_3_1_12_2","first-page":"6824","volume-title":"Proceedings of the IEEE International Conference on Computer Vision","author":"Fan Haoqi","year":"2021","unstructured":"Haoqi Fan, Bo Xiong, Karttikeya Mangalam, Yanghao Li, Zhicheng Yan, Jitendra Malik, and Christoph Feichtenhofer. 2021. Multiscale vision transformers. In Proceedings of the IEEE International Conference on Computer Vision, 6824\u20136835."},{"key":"e_1_3_1_13_2","first-page":"2261","volume-title":"Proceedings of the Advances in Neural Information Processing Systems","author":"Fan Quanfu","year":"2019","unstructured":"Quanfu Fan, Chun-Fu (Richard) Chen, Hilde Kuehne, Marco Pistoia, and David Cox. 2019. More is less: Learning efficient video representations by big-little network and depthwise temporal aggregation. In Proceedings of the Advances in Neural Information Processing Systems, 2261\u20132270."},{"key":"e_1_3_1_14_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00028"},{"key":"e_1_3_1_15_2","first-page":"6202","volume-title":"Proceedings of the IEEE International Conference on Computer Vision","author":"Feichtenhofer Christoph","year":"2019","unstructured":"Christoph Feichtenhofer, Haoqi Fan, Jitendra Malik, and Kaiming He. 2019. Slowfast networks for video recognition. In Proceedings of the IEEE International Conference on Computer Vision, 6202\u20136211."},{"key":"e_1_3_1_16_2","doi-asserted-by":"publisher","DOI":"10.1145\/3633781"},{"key":"e_1_3_1_17_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2024.111852"},{"key":"e_1_3_1_18_2","unstructured":"Priya Goyal Piotr Doll\u00e1r Ross Girshick Pieter Noordhuis Lukasz Wesolowski Aapo Kyrola Andrew Tulloch Yangqing Jia and Kaiming He. 2017. Accurate large minibatch SGD: Training ImageNet in 1 hour. arXiv:1706.02677. Retrieved from https:\/\/arxiv.org\/abs\/1706.02677"},{"key":"e_1_3_1_19_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.622"},{"key":"e_1_3_1_20_2","unstructured":"Qipeng Guo Xipeng Qiu Pengfei Liu Yunfan Shao Xiangyang Xue and Zheng Zhang. 2019. Star-transformer. arXiv:1902.09113. Retrieved from https:\/\/arxiv.org\/abs\/1902.09113"},{"key":"e_1_3_1_21_2","first-page":"928","volume-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition","author":"Hao Yanbin","year":"2022","unstructured":"Yanbin Hao, Hao Zhang, Chong-Wah Ngo, and Xiangnan He. 2022. Group contextualization for video recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 928\u2013938."},{"key":"e_1_3_1_22_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.procs.2017.09.092"},{"key":"e_1_3_1_23_2","first-page":"770","volume-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition","author":"He Kaiming","year":"2016","unstructured":"Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 770\u2013778."},{"key":"e_1_3_1_24_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2014.223"},{"key":"e_1_3_1_25_2","doi-asserted-by":"publisher","DOI":"10.1145\/3065386"},{"key":"e_1_3_1_26_2","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2023.3282631"},{"key":"e_1_3_1_27_2","first-page":"1092","volume-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition","author":"Li Xianhang","year":"2020","unstructured":"Xianhang Li, Yali Wang, Zhipeng Zhou, and Yu Qiao. 2020. Smallbignet: Integrating core and contextual views for video classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 1092\u20131101."},{"key":"e_1_3_1_28_2","first-page":"13557","volume-title":"Proceedings of the IEEE International Conference on Computer Vision","author":"Li Xinyu","year":"2021","unstructured":"Xinyu Li, Yanyi Zhang, Chunhui Liu, Bing Shuai, Yi Zhu, Biagio Brattoli, Hao Chen, Ivan Marsic, and Joseph Tighe. 2021. VidTr: Video transformer without convolutions. In Proceedings of the IEEE International Conference on Computer Vision, 13557\u201313567."},{"key":"e_1_3_1_29_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00099"},{"key":"e_1_3_1_30_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01228-1_38"},{"key":"e_1_3_1_31_2","unstructured":"Jingyun Liang Jiezhang Cao Yuchen Fan Kai Zhang Rakesh Ranjan Yawei Li Radu Timofte and Luc Van Gool. 2022. VRT: A video restoration transformer. arXiv:2201.12288. Retrieved from https:\/\/arxiv.org\/abs\/2201.12288"},{"key":"e_1_3_1_32_2","first-page":"13334","volume-title":"Proceedings of the International Conference on Machine Learning","author":"Lin Jing","year":"2022","unstructured":"Jing Lin, Yuanhao Cai, Xiaowan Hu, Haoqian Wang, Youliang Yan, Xueyi Zou, Henghui Ding, Yulun Zhang, Radu Timofte, and Luc Van Gool. 2022. Flow-guided sparse transformer for video deblurring. In Proceedings of the International Conference on Machine Learning, 13334\u201313343."},{"key":"e_1_3_1_33_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00718"},{"key":"e_1_3_1_34_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"e_1_3_1_35_2","first-page":"11669","volume-title":"Proceedings of the 34th AAAI Conference on Artificial Intelligence","author":"Liu Zhaoyang","year":"2020","unstructured":"Zhaoyang Liu, Donghao Luo, Yabiao Wang, Limin Wang, Ying Tai, Chengjie Wang, Jilin Li, Feiyue Huang, and Tong Lu. 2020. TEINet: Towards an efficient architecture for video recognition. In Proceedings of the 34th AAAI Conference on Artificial Intelligence, 11669\u201311676."},{"key":"e_1_3_1_36_2","first-page":"3202","volume-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition","author":"Liu Ze","year":"2022","unstructured":"Ze Liu, Jia Ning, Yue Cao, Yixuan Wei, Zheng Zhang, Stephen Lin, and Han Hu. 2022. Video swin transformer. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 3202\u20133211."},{"key":"e_1_3_1_37_2","first-page":"13708","volume-title":"Proceedings of the IEEE International Conference on Computer Vision","author":"Liu Zhaoyang","year":"2021","unstructured":"Zhaoyang Liu, Limin Wang, Wayne Wu, Chen Qian, and Tong Lu. 2021. TAM: Temporal adaptive module for video recognition. In Proceedings of the IEEE International Conference on Computer Vision, 13708\u201313718."},{"key":"e_1_3_1_38_2","first-page":"4898","volume-title":"Proceedings of the Advances in Neural Information Processing Systems","author":"Luo Wenjie","year":"2016","unstructured":"Wenjie Luo, Yujia Li, Raquel Urtasun, and Richard Zemel. 2016. Understanding the effective receptive field in deep convolutional neural networks. In Proceedings of the Advances in Neural Information Processing Systems, 4898\u20134906."},{"key":"e_1_3_1_39_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2021.108487"},{"key":"e_1_3_1_40_2","first-page":"3163","volume-title":"Proceedings of the IEEE International Conference on Computer Vision","author":"Neimark Daniel","year":"2021","unstructured":"Daniel Neimark, Omri Bar, Maya Zohar, and Dotan Asselmann. 2021. Video transformer network. In Proceedings of the IEEE International Conference on Computer Vision, 3163\u20133172."},{"key":"e_1_3_1_41_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.bspc.2018.04.014"},{"key":"e_1_3_1_42_2","doi-asserted-by":"publisher","DOI":"10.1109\/TCSVT.2019.2894161"},{"key":"e_1_3_1_43_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.590"},{"key":"e_1_3_1_44_2","doi-asserted-by":"publisher","DOI":"10.1145\/3485665"},{"key":"e_1_3_1_45_2","unstructured":"Khurram Soomro Amir Roshan Zamir and Mubarak Shah. 2012. UCF101: A dataset of 101 human actions classes from videos in the wild. arXiv:1212.0402. Retrieved from https:\/\/arxiv.org\/abs\/1212.0402"},{"key":"e_1_3_1_46_2","first-page":"7262","volume-title":"Proceedings of the IEEE International Conference on Computer Vision","author":"Strudel Robin","year":"2021","unstructured":"Robin Strudel, Ricardo Garcia, Ivan Laptev, and Cordelia Schmid. 2021. Segmenter: Transformer for semantic segmentation. In Proceedings of the IEEE International Conference on Computer Vision, 7262\u20137272."},{"key":"e_1_3_1_47_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.510"},{"key":"e_1_3_1_48_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00565"},{"key":"e_1_3_1_49_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00675"},{"issue":"11","key":"e_1_3_1_50_2","first-page":"2579","article-title":"Visualizing data using t-SNE","volume":"9","author":"Van der Maaten Laurens","year":"2008","unstructured":"Laurens Van der Maaten and Geoffrey Hinton. 2008. Visualizing data using t-SNE. Journal of Machine Learning Research 9, 11 (2008), 2579\u20132605.","journal-title":"Journal of Machine Learning Research"},{"key":"e_1_3_1_51_2","doi-asserted-by":"crossref","first-page":"124596","DOI":"10.1016\/j.eswa.2024.124596","article-title":"An efficient motion visual learning method for video action recognition","volume":"255","author":"Wang Bin","year":"2024","unstructured":"Bin Wang, Faliang Chang, Chunsheng Liu, Wenqian Wang, and Ruiyi Ma. 2024. An efficient motion visual learning method for video action recognition. Expert Systems with Applications 255 (2024), 124596.","journal-title":"Expert Systems with Applications"},{"key":"e_1_3_1_52_2","first-page":"352","volume-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition","author":"Wang Heng","year":"2020","unstructured":"Heng Wang, Du Tran, Lorenzo Torresani, and Matt Feiszli. 2020. Video modeling with correlation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 352\u2013361."},{"key":"e_1_3_1_53_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00155"},{"key":"e_1_3_1_54_2","first-page":"1895","volume-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition","author":"Wang Limin","year":"2021","unstructured":"Limin Wang, Zhan Tong, Bin Ji, and Gangshan Wu. 2021. TDN: Temporal difference networks for efficient action recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 1895\u20131904."},{"key":"e_1_3_1_55_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00813"},{"key":"e_1_3_1_56_2","first-page":"12249","volume-title":"Proceedings of the 34th AAAI Conference on Artificial Intelligence","author":"Wang Xiaohan","year":"2020","unstructured":"Xiaohan Wang, Yu Wu, Linchao Zhu, and Yi Yang. 2020. Symbiotic attention with privileged information for egocentric action recognition. In Proceedings of the 34th AAAI Conference on Artificial Intelligence, 12249\u201312256."},{"key":"e_1_3_1_57_2","first-page":"13214","volume-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition","author":"Wang Zhengwei","year":"2021","unstructured":"Zhengwei Wang, Qi She, and Aljosa Smolic. 2021. Action-net: Multipath excitation for action recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 13214\u201313223."},{"key":"e_1_3_1_58_2","first-page":"627","volume-title":"Proceedings of the European Conference on Computer Vision","author":"Xiang Wangmeng","year":"2022","unstructured":"Wangmeng Xiang, Chao Li, Biao Wang, Xihan Wei, Xian-Sheng Hua, and Lei Zhang. 2022. Spatiotemporal self-attention modeling with temporal patch shift for action recognition. In Proceedings of the European Conference on Computer Vision, 627\u2013644."},{"key":"e_1_3_1_59_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01267-0_19"},{"issue":"2","key":"e_1_3_1_60_2","first-page":"1","article-title":"Exploiting attention-consistency loss for spatial-temporal stream action recognition","volume":"18","author":"Xu Haotian","year":"2022","unstructured":"Haotian Xu, Xiaobo Jin, Qiufeng Wang, Amir Hussain, and Kaizhu Huang. 2022. Exploiting attention-consistency loss for spatial-temporal stream action recognition. ACM Transactions on Multimedia Computing, Communications, and Applications 18, 2 (2022), 1\u201315.","journal-title":"ACM Transactions on Multimedia Computing, Communications, and Applications"},{"key":"e_1_3_1_61_2","doi-asserted-by":"publisher","DOI":"10.1145\/3578518"},{"key":"e_1_3_1_62_2","first-page":"17283","volume-title":"Proceedings of the Advances in Neural Information Processing Systems","author":"Zaheer Manzil","year":"2020","unstructured":"Manzil Zaheer, Guru Guruganesh, Kumar Avinava Dubey, Joshua Ainslie, Chris Alberti, Santiago Ontanon, Philip Pham, Anirudh Ravula, Qifan Wang, Li Yang, et al. 2020. Big bird: Transformers for longer sequences. In Proceedings of the Advances in Neural Information Processing Systems, 17283\u201317297."},{"key":"e_1_3_1_63_2","doi-asserted-by":"crossref","first-page":"5773","DOI":"10.1145\/3503161.3547908","volume-title":"Proceedings of the ACM International Conference on Multimedia","author":"Zhang Hao","year":"2022","unstructured":"Hao Zhang, Lechao Cheng, Yanbin Hao, and Chong-wah Ngo. 2022. Long-term leap attention, short-term periodic shift for video classification. In Proceedings of the ACM International Conference on Multimedia, 5773\u20135782."},{"key":"e_1_3_1_64_2","first-page":"917","volume-title":"Proceedings of the ACM International Conference on Multimedia","author":"Zhang Hao","year":"2021","unstructured":"Hao Zhang, Yanbin Hao, and Chong-Wah Ngo. 2021. Token shift transformer for video classification. In Proceedings of the ACM International Conference on Multimedia, 917\u2013925."},{"key":"e_1_3_1_65_2","doi-asserted-by":"publisher","DOI":"10.1145\/3568312"}],"container-title":["ACM Transactions on Multimedia Computing, Communications, and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3712594","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3712594","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T01:18:37Z","timestamp":1750295917000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3712594"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,3,10]]},"references-count":64,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2025,3,31]]}},"alternative-id":["10.1145\/3712594"],"URL":"https:\/\/doi.org\/10.1145\/3712594","relation":{},"ISSN":["1551-6857","1551-6865"],"issn-type":[{"value":"1551-6857","type":"print"},{"value":"1551-6865","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,3,10]]},"assertion":[{"value":"2024-03-26","order":0,"name":"received","label":"Received","group":{"name":"publication_history","label":"Publication History"}},{"value":"2024-12-10","order":2,"name":"accepted","label":"Accepted","group":{"name":"publication_history","label":"Publication History"}},{"value":"2025-03-10","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}