{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,24]],"date-time":"2026-02-24T17:17:01Z","timestamp":1771953421766,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":64,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,10,26]],"date-time":"2023-10-26T00:00:00Z","timestamp":1698278400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"name":"Shenzhen Stable Supporting Program","award":["WDZC20200820200655001"],"award-info":[{"award-number":["WDZC20200820200655001"]}]},{"name":"Shenzhen Key Laboratory of Next Generation Interactive Media Innovative Technology","award":["ZDSYS20210623092001004"],"award-info":[{"award-number":["ZDSYS20210623092001004"]}]},{"name":"Shenzhen Science and Technology Project","award":["JCYJ20200109143041798"],"award-info":[{"award-number":["JCYJ20200109143041798"]}]},{"name":"National Key R&D Program of China","award":["2020AAA0108303"],"award-info":[{"award-number":["2020AAA0108303"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,10,26]]},"DOI":"10.1145\/3581783.3611795","type":"proceedings-article","created":{"date-parts":[[2023,10,27]],"date-time":"2023-10-27T07:27:30Z","timestamp":1698391650000},"page":"6695-6704","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":17,"title":["Ada-DQA: Adaptive Diverse Quality-aware Feature Acquisition for Video Quality Assessment"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-0908-6587","authenticated-orcid":false,"given":"Hongbo","family":"Liu","sequence":"first","affiliation":[{"name":"Tsinghua University, Shenzhen, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5006-1990","authenticated-orcid":false,"given":"Mingda","family":"Wu","sequence":"additional","affiliation":[{"name":"Kuaishou Technology, Peking, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3681-2196","authenticated-orcid":false,"given":"Kun","family":"Yuan","sequence":"additional","affiliation":[{"name":"Kuaishou Technology, Shenzhen, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8173-9290","authenticated-orcid":false,"given":"Ming","family":"Sun","sequence":"additional","affiliation":[{"name":"Kuaishou Technology, Peking, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1534-4549","authenticated-orcid":false,"given":"Yansong","family":"Tang","sequence":"additional","affiliation":[{"name":"Tsinghua University, Shenzhen, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3617-2184","authenticated-orcid":false,"given":"Chuanchuan","family":"Zheng","sequence":"additional","affiliation":[{"name":"Kuaishou Technology, Peking, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3368-9206","authenticated-orcid":false,"given":"Xing","family":"Wen","sequence":"additional","affiliation":[{"name":"Kuaishou Technology, Peking, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6906-6735","authenticated-orcid":false,"given":"Xiu","family":"Li","sequence":"additional","affiliation":[{"name":"Tsinghua University, Shenzhen, China"}]}],"member":"320","published-online":{"date-parts":[[2023,10,27]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"ViViT: A Video Vision Transformer","author":"Arnab Anurag","unstructured":"Anurag Arnab, Mostafa Dehghani, Georg Heigold, Chen Sun, Mario Lucic, and Cordelia Schmid. 2021. ViViT: A Video Vision Transformer. In ICCV. IEEE, 6816--6826."},{"key":"e_1_3_2_1_2_1","volume-title":"ICML","volume":"139","author":"Bertasius Gedas","year":"2021","unstructured":"Gedas Bertasius, Heng Wang, and Lorenzo Torresani. 2021. Is Space-Time Attention All You Need for Video Understanding?. In ICML, Vol. 139. PMLR, 813--824."},{"key":"e_1_3_2_1_3_1","volume-title":"A Short Note about Kinetics-600. CoRR","author":"Carreira Jo\u00e3o","year":"2018","unstructured":"Jo\u00e3o Carreira, Eric Noland, Andras Banki-Horvath, Chloe Hillier, and Andrew Zisserman. 2018. A Short Note about Kinetics-600. CoRR, Vol. abs\/1808.01340 (2018)."},{"key":"e_1_3_2_1_4_1","volume-title":"Deep Perceptual Preprocessing for Video Coding","author":"Chadha Aaron","unstructured":"Aaron Chadha and Yiannis Andreopoulos. 2021. Deep Perceptual Preprocessing for Video Coding. In CVPR. IEEE, 14852--14861."},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1109\/MNET.2010.5430141"},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"crossref","unstructured":"Pengfei Chen Leida Li Lei Ma Jinjian Wu and Guangming Shi. 2020b. RIRNet: Recurrent-In-Recurrent Network for Video Quality Assessment. In ACM MM. ACM 834--842.","DOI":"10.1145\/3394171.3413717"},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2021.3130536"},{"key":"e_1_3_2_1_8_1","volume-title":"Hinton","author":"Chen Ting","year":"2020","unstructured":"Ting Chen, Simon Kornblith, Mohammad Norouzi, and Geoffrey E. Hinton. 2020a. A Simple Framework for Contrastive Learning of Visual Representations. In ICML, Vol. 119. PMLR, 1597--1607."},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1109\/COMST.2014.2363139"},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1109\/TBC.2011.2104671"},{"key":"e_1_3_2_1_11_1","unstructured":"Cisco. 2021. Cisco annual internet report white paper. https:\/\/www.cisco.com\/c\/en\/us\/solutions\/collateral\/ executive-perspectives\/annual-internet-report\/white-paper-c11-741490.html."},{"key":"e_1_3_2_1_12_1","unstructured":"MMAction2 Contributors. 2020. OpenMMLab's Next Generation Video Understanding Toolbox and Benchmark. https:\/\/github.com\/open-mmlab\/mmaction2."},{"key":"e_1_3_2_1_13_1","volume-title":"ImageNet: A large-scale hierarchical image database","author":"Deng Jia","unstructured":"Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. 2009. ImageNet: A large-scale hierarchical image database. In CVPR. IEEE Computer Society, 248--255."},{"key":"e_1_3_2_1_14_1","volume-title":"Bovik","author":"Ebenezer Joshua Peter","year":"2020","unstructured":"Joshua Peter Ebenezer, Zaixi Shang, Yongjun Wu, Hai Wei, and Alan C. Bovik. 2020. No-Reference Video Quality Assessment Using Space-Time Chips. In MMSP. IEEE, 1--6."},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"crossref","unstructured":"Christoph Feichtenhofer Haoqi Fan Jitendra Malik and Kaiming He. 2019. SlowFast Networks for Video Recognition. In ICCV. 6201--6210.","DOI":"10.1109\/ICCV.2019.00630"},{"key":"e_1_3_2_1_16_1","volume-title":"Jensen-Shannon divergence and Hilbert space embedding","author":"Fuglede Bent","unstructured":"Bent Fuglede and Flemming Tops\u00f8e. 2004. Jensen-Shannon divergence and Hilbert space embedding. In ISIT. IEEE, 31."},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2021.3077642"},{"key":"e_1_3_2_1_18_1","volume-title":"No-Reference Image Quality Assessment with Reinforcement Recursive List-Wise Ranking","author":"Gu Jie","unstructured":"Jie Gu, Gaofeng Meng, Cheng Da, Shiming Xiang, and Chunhong Pan. 2019a. No-Reference Image Quality Assessment with Reinforcement Recursive List-Wise Ranking. In AAAI. AAAI Press, 8336--8343."},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2019.02.021"},{"key":"e_1_3_2_1_20_1","volume-title":"Girshick","author":"He Kaiming","year":"2020","unstructured":"Kaiming He, Haoqi Fan, Yuxin Wu, Saining Xie, and Ross B. Girshick. 2020. Momentum Contrast for Unsupervised Visual Representation Learning. In CVPR. Computer Vision Foundation \/ IEEE, 9726--9735."},{"key":"e_1_3_2_1_21_1","volume-title":"Distilling the Knowledge in a Neural Network. CoRR","author":"Hinton Geoffrey E.","year":"2015","unstructured":"Geoffrey E. Hinton, Oriol Vinyals, and Jeffrey Dean. 2015. Distilling the Knowledge in a Neural Network. CoRR, Vol. abs\/1503.02531 (2015)."},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2013.2291663"},{"key":"e_1_3_2_1_23_1","volume-title":"The Konstanz natural video database (KoNViD-1k)","author":"Hosu Vlad","unstructured":"Vlad Hosu, Franz Hahn, Mohsen Jenadeleh, Hanhe Lin, Hui Men, Tam\u00e1s Szir\u00e1nyi, Shujun Li, and Dietmar Saupe. 2017. The Konstanz natural video database (KoNViD-1k). In QoMEX. IEEE, 1--6."},{"key":"e_1_3_2_1_24_1","first-page":"115782","article-title":"No-reference video quality evaluation by a deep transfer CNN architecture","volume":"83","author":"Hou Rui","year":"2020","unstructured":"Rui Hou, YunHao Zhao, Yang Hu, and Huan Liu. 2020. No-reference video quality evaluation by a deep transfer CNN architecture. SPIC, Vol. 83 (2020), 115782.","journal-title":"SPIC"},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"crossref","unstructured":"Andrej Karpathy George Toderici Sanketh Shetty Thomas Leung Rahul Sukthankar and Li Fei-Fei. 2014. Large-scale Video Classification with Convolutional Neural Networks. In CVPR.","DOI":"10.1109\/CVPR.2014.223"},{"key":"e_1_3_2_1_26_1","volume-title":"Karen Simonyan, Brian Zhang, Chloe Hillier, Sudheendra Vijayanarasimhan, Fabio Viola, Tim Green, Trevor Back, Paul Natsev, Mustafa Suleyman, and Andrew Zisserman.","author":"Kay Will","year":"2017","unstructured":"Will Kay, Jo a o Carreira, Karen Simonyan, Brian Zhang, Chloe Hillier, Sudheendra Vijayanarasimhan, Fabio Viola, Tim Green, Trevor Back, Paul Natsev, Mustafa Suleyman, and Andrew Zisserman. 2017. The Kinetics Human Action Video Dataset. CoRR, Vol. abs\/1705.06950 (2017)."},{"key":"e_1_3_2_1_27_1","volume-title":"MUSIQ: Multi-scale Image Quality Transformer. (October","author":"Ke Junjie","year":"2021","unstructured":"Junjie Ke, Qifei Wang, Yilin Wang, Peyman Milanfar, and Feng Yang. 2021. MUSIQ: Multi-scale Image Quality Transformer. (October 2021), 5148--5157."},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2019.2923051"},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1145\/3394171.3413845"},{"key":"e_1_3_2_1_30_1","volume-title":"Blindly Assess Quality of In-the-Wild Videos via Quality-aware Pre-training and Motion Perception. CoRR","author":"Li Bowen","year":"2021","unstructured":"Bowen Li, Weixia Zhang, Meng Tian, Guangtao Zhai, and Xianpei Wang. 2021b. Blindly Assess Quality of In-the-Wild Videos via Quality-aware Pre-training and Motion Perception. CoRR, Vol. abs\/2108.08505 (2021)."},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"crossref","unstructured":"Dingquan Li Tingting Jiang and Ming Jiang. 2019. Quality Assessment of In-the-Wild Videos. In ACM Multimedia. ACM 2351--2359.","DOI":"10.1145\/3343031.3351028"},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-020-01408-w"},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.1145\/3503161.3547849"},{"key":"e_1_3_2_1_34_1","volume-title":"Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. In ICCV. 10012--10022.","author":"Liu Ze","year":"2021","unstructured":"Ze Liu, Yutong Lin, Yue Cao, Han Hu, Yixuan Wei, Zheng Zhang, Stephen Lin, and Baining Guo. 2021a. Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. In ICCV. 10012--10022."},{"key":"e_1_3_2_1_35_1","volume-title":"Video Swin Transformer. CoRR","author":"Liu Ze","year":"2021","unstructured":"Ze Liu, Jia Ning, Yue Cao, Yixuan Wei, Zheng Zhang, Stephen Lin, and Han Hu. 2021b. Video Swin Transformer. CoRR, Vol. abs\/2106.13230 (2021)."},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2012.2214050"},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2015.2502725"},{"key":"e_1_3_2_1_38_1","first-page":"209","article-title":"Making a \"Completely Blind\" Image Quality Analyzer","volume":"20","author":"Mittal Anish","year":"2013","unstructured":"Anish Mittal, Rajiv Soundararajan, and Alan C. Bovik. 2013. Making a \"Completely Blind\" Image Quality Analyzer. IEEE SPL, Vol. 20, 3 (2013), 209--212.","journal-title":"IEEE SPL"},{"key":"e_1_3_2_1_39_1","volume-title":"PyTorch: An Imperative Style","author":"Paszke Adam","unstructured":"Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, Alban Desmaison, Andreas K\u00f6pf, Edward Z. Yang, Zachary DeVito, Martin Raison, Alykhan Tejani, Sasank Chilamkurthy, Benoit Steiner, Lu Fang, Junjie Bai, and Soumith Chintala. 2019. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In NeurIPS. 8024--8035."},{"key":"e_1_3_2_1_40_1","volume-title":"Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, Gretchen Krueger, and Ilya Sutskever.","author":"Radford Alec","year":"2021","unstructured":"Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, Gretchen Krueger, and Ilya Sutskever. 2021. Learning Transferable Visual Models From Natural Language Supervision. In ICML. PMLR, 8748--8763."},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2012.2191563"},{"key":"e_1_3_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2014.2299154"},{"key":"e_1_3_2_1_43_1","doi-asserted-by":"publisher","DOI":"10.1109\/QoMEX.2014.6982289"},{"key":"e_1_3_2_1_44_1","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2018.2869673"},{"key":"e_1_3_2_1_45_1","volume-title":"Le","author":"Tan Mingxing","year":"2019","unstructured":"Mingxing Tan and Quoc V. Le. 2019. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. In ICML (Proceedings of Machine Learning Research, Vol. 97). PMLR, 6105--6114."},{"key":"e_1_3_2_1_46_1","volume-title":"Video Classification With Channel-Separated Convolutional Networks","author":"Tran Du","unstructured":"Du Tran, Heng Wang, Matt Feiszli, and Lorenzo Torresani. 2019. Video Classification With Channel-Separated Convolutional Networks. In ICCV. IEEE, 5551--5560."},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.1109\/PCS50896.2021.9477483"},{"key":"e_1_3_2_1_48_1","first-page":"4449","article-title":"UGC-VQA: Benchmarking Blind Video Quality Assessment for User Generated Content","volume":"30","author":"Tu Zhengzhong","year":"2021","unstructured":"Zhengzhong Tu, Yilin Wang, Neil Birkbeck, Balu Adsumilli, and Alan C. Bovik. 2021b. UGC-VQA: Benchmarking Blind Video Quality Assessment for User Generated Content. IEEE TIP, Vol. 30 (2021), 4449--4464.","journal-title":"IEEE TIP"},{"key":"e_1_3_2_1_49_1","first-page":"4449","article-title":"UGC-VQA: Benchmarking Blind Video Quality Assessment for User Generated Content","volume":"30","author":"Tu Zhengzhong","year":"2021","unstructured":"Zhengzhong Tu, Yilin Wang, Neil Birkbeck, Balu Adsumilli, and Alan C. Bovik. 2021c. UGC-VQA: Benchmarking Blind Video Quality Assessment for User Generated Content. IEEE TIP, Vol. 30 (2021), 4449--4464.","journal-title":"IEEE TIP"},{"key":"e_1_3_2_1_50_1","volume-title":"Bovik","author":"Tu Zhengzhong","year":"2021","unstructured":"Zhengzhong Tu, Xiangxu Yu, Yilin Wang, Neil Birkbeck, Balu Adsumilli, and Alan C. Bovik. 2021d. RAPIQUE: Rapid and Accurate Video Quality Prediction of User Generated Content. CoRR, Vol. abs\/2101.10955 (2021)."},{"key":"e_1_3_2_1_51_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11760-019-01510-8"},{"key":"e_1_3_2_1_52_1","volume-title":"YouTube UGC Dataset for Video Compression Research","author":"Wang Yilin","unstructured":"Yilin Wang, Sasi Inguva, and Balu Adsumilli. 2019. YouTube UGC Dataset for Video Compression Research. In MMSP. IEEE, 1--5."},{"key":"e_1_3_2_1_53_1","volume-title":"Neil Birkbeck, Balu Adsumilli, Peyman Milanfar, and Feng Yang.","author":"Wang Yilin","year":"2021","unstructured":"Yilin Wang, Junjie Ke, Hossein Talebi, Joong Gon Yim, Neil Birkbeck, Balu Adsumilli, Peyman Milanfar, and Feng Yang. 2021. Rich Features for Perceptual Quality Assessment of UGC Videos. In CVPR. 13435--13444."},{"key":"e_1_3_2_1_54_1","doi-asserted-by":"publisher","DOI":"10.1016\/S0923-5965(03)00076-6"},{"key":"e_1_3_2_1_55_1","volume-title":"ECCV (6)","author":"Wu Haoning","unstructured":"Haoning Wu, Chaofeng Chen, Jingwen Hou, Liang Liao, Annan Wang, Wenxiu Sun, Qiong Yan, and Weisi Lin. 2022a. FAST-VQA: Efficient End-to-End Video Quality Assessment with Fragment Sampling. In ECCV (6), Vol. 13666. 538--554."},{"key":"e_1_3_2_1_56_1","volume-title":"DisCoVQA: Temporal Distortion-Content Transformers for Video Quality Assessment. CoRR","author":"Wu Haoning","year":"2022","unstructured":"Haoning Wu, Chaofeng Chen, Liang Liao, Jingwen Hou, Qiong Yan, and Weisi Lin. 2022b. DisCoVQA: Temporal Distortion-Content Transformers for Video Quality Assessment. CoRR, Vol. abs\/2206.09853 (2022)."},{"key":"e_1_3_2_1_57_1","volume-title":"StarVQA: Space-Time Attention for Video Quality Assessment. CoRR","author":"Xing Fengchuang","year":"2021","unstructured":"Fengchuang Xing, Yuan-Gen Wang, Hanpin Wang, Leida Li, and Guopu Zhu. 2021. StarVQA: Space-Time Attention for Video Quality Assessment. CoRR, Vol. abs\/2108.09635 (2021)."},{"key":"e_1_3_2_1_58_1","doi-asserted-by":"crossref","unstructured":"Jiahua Xu Jing Li Xingguang Zhou Wei Zhou Baichao Wang and Zhibo Chen. 2021. Perceptual Quality Assessment of Internet Videos. In ACM Multimedia. ACM 1248--1257.","DOI":"10.1145\/3474085.3475486"},{"key":"e_1_3_2_1_59_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-54190-7_1"},{"key":"e_1_3_2_1_60_1","volume-title":"Patch-VQ: 'Patching Up' the Video Quality Problem","author":"Ying Zhenqiang","unstructured":"Zhenqiang Ying, Maniratnam Mandal, Deepti Ghadiyaram, and Alan Bovik. 2021. Patch-VQ: 'Patching Up' the Video Quality Problem. In CVPR. Computer Vision Foundation \/ IEEE, 14019--14029."},{"key":"e_1_3_2_1_61_1","doi-asserted-by":"crossref","unstructured":"Junyong You. 2021. Long Short-term Convolutional Transformer for No-Reference Video Quality Assessment. In ACM Multimedia. ACM 2112--2120.","DOI":"10.1145\/3474085.3475368"},{"key":"e_1_3_2_1_62_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICIP.2019.8803395"},{"key":"e_1_3_2_1_63_1","volume-title":"2023 b. Quality-Aware Pre-Trained Models for Blind Image Quality Assessment","author":"Zhao Kai","unstructured":"Kai Zhao, Kun Yuan, Ming Sun, Mading Li, and Xing Wen. 2023 b. Quality-Aware Pre-Trained Models for Blind Image Quality Assessment. In CVPR. IEEE Computer Society, 22302--22313."},{"key":"e_1_3_2_1_64_1","volume-title":"Frames and Clips Integration for Video Quality Assessment. In CVPR Workshops. IEEE Computer Society, 1302--1310","author":"Zhao Kai","year":"2023","unstructured":"Kai Zhao, Kun Yuan, Ming Sun, and Xing Wen. 2023 a. Zoom-VQA: Patches, Frames and Clips Integration for Video Quality Assessment. In CVPR Workshops. IEEE Computer Society, 1302--1310."}],"event":{"name":"MM '23: The 31st ACM International Conference on Multimedia","location":"Ottawa ON Canada","acronym":"MM '23","sponsor":["SIGMM ACM Special Interest Group on Multimedia"]},"container-title":["Proceedings of the 31st ACM International Conference on Multimedia"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3581783.3611795","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3581783.3611795","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T00:02:07Z","timestamp":1755820927000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3581783.3611795"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,10,26]]},"references-count":64,"alternative-id":["10.1145\/3581783.3611795","10.1145\/3581783"],"URL":"https:\/\/doi.org\/10.1145\/3581783.3611795","relation":{},"subject":[],"published":{"date-parts":[[2023,10,26]]},"assertion":[{"value":"2023-10-27","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}