{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,11]],"date-time":"2026-02-11T12:40:16Z","timestamp":1770813616219,"version":"3.50.1"},"reference-count":60,"publisher":"Association for Computing Machinery (ACM)","issue":"6","license":[{"start":{"date-parts":[[2023,7,12]],"date-time":"2023-07-12T00:00:00Z","timestamp":1689120000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"name":"National Key Research and Development Program of China","award":["2020AAA0106300"],"award-info":[{"award-number":["2020AAA0106300"]}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62250008, 62222209, 62102222, 61872215"],"award-info":[{"award-number":["62250008, 62222209, 62102222, 61872215"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["ACM Trans. Multimedia Comput. Commun. Appl."],"published-print":{"date-parts":[[2023,11,30]]},"abstract":"<jats:p>\n            <jats:bold>Temporal Sentence Grounding in Videos (TSGV)<\/jats:bold>\n            , which aims to ground a natural language sentence that indicates complex human activities in an untrimmed video, has drawn widespread attention over the past few years. However, recent studies have found that current benchmark datasets may have obvious moment annotation biases, enabling several simple baselines even without training to achieve\n            <jats:bold>state-of-the-art (SOTA)<\/jats:bold>\n            performance. In this paper, we take a closer look at existing evaluation protocols for TSGV, and find that both the prevailing dataset splits and evaluation metrics are the devils that lead to untrustworthy benchmarking. Therefore, we propose to re-organize the two widely-used datasets, making the ground-truth moment distributions different in the training and test splits, i.e.,\n            <jats:bold>out-of-distribution (OOD)<\/jats:bold>\n            test. Meanwhile, we introduce a new evaluation metric \u201cdR@\n            <jats:italic>n<\/jats:italic>\n            ,IoU=\n            <jats:italic>m<\/jats:italic>\n            \u201d that discounts the basic recall scores especially with small IoU thresholds, so as to alleviate the inflating evaluation caused by biased datasets with a large proportion of long ground-truth moments. New benchmarking results indicate that our proposed evaluation protocols can better monitor the research progress in TSGV. Furthermore, we propose a novel causality-based\n            <jats:bold>Multi-branch Deconfounding Debiasing (MDD)<\/jats:bold>\n            framework for unbiased moment prediction. Specifically, we design a multi-branch deconfounder to eliminate the effects caused by multiple confounders with causal intervention. In order to help the model better align the semantics between sentence queries and video moments, we enhance the representations during feature encoding. Specifically, for textual information, the query is parsed into several verb-centered phrases to obtain a more fine-grained textual feature. For visual information, the positional information has been decomposed from the moment features to enhance the representations of moments with diverse locations. Extensive experiments demonstrate that our proposed approach can achieve competitive results among existing SOTA approaches and outperform the base model with great gains.\n          <\/jats:p>","DOI":"10.1145\/3565573","type":"journal-article","created":{"date-parts":[[2022,10,7]],"date-time":"2022-10-07T13:19:40Z","timestamp":1665148780000},"page":"1-23","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":12,"title":["A Closer Look at Debiased Temporal Sentence Grounding in Videos: Dataset, Metric, and Approach"],"prefix":"10.1145","volume":"19","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-5382-6699","authenticated-orcid":false,"given":"Xiaohan","family":"Lan","sequence":"first","affiliation":[{"name":"Tsinghua Shenzhen International Graduate School, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8701-7689","authenticated-orcid":false,"given":"Yitian","family":"Yuan","sequence":"additional","affiliation":[{"name":"Meituan, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0351-2939","authenticated-orcid":false,"given":"Xin","family":"Wang","sequence":"additional","affiliation":[{"name":"Tsinghua University, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6148-9709","authenticated-orcid":false,"given":"Long","family":"Chen","sequence":"additional","affiliation":[{"name":"Columbia University, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5462-6178","authenticated-orcid":false,"given":"Zhi","family":"Wang","sequence":"additional","affiliation":[{"name":"Tsinghua Shenzhen International Graduate School, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7331-6132","authenticated-orcid":false,"given":"Lin","family":"Ma","sequence":"additional","affiliation":[{"name":"Meituan, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2236-9290","authenticated-orcid":false,"given":"Wenwu","family":"Zhu","sequence":"additional","affiliation":[{"name":"Tsinghua University, China"}]}],"member":"320","published-online":{"date-parts":[[2023,7,12]]},"reference":[{"key":"e_1_3_1_2_2","first-page":"4971","volume-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","author":"Agrawal Aishwarya","year":"2018","unstructured":"Aishwarya Agrawal, Dhruv Batra, Devi Parikh, and Aniruddha Kembhavi. 2018. Don\u2019t just assume; look and answer: Overcoming priors for visual question answering. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 4971\u20134980."},{"key":"e_1_3_1_3_2","first-page":"839","volume-title":"Proceedings of the International Conference on Neural Information Processing Systems","author":"Cad\u00e8ne R\u00e9mi","year":"2019","unstructured":"R\u00e9mi Cad\u00e8ne, Corentin Dancette, Hedi Ben-younes, Matthieu Cord, and Devi Parikh. 2019. RUBi: Reducing unimodal biases for visual question answering. In Proceedings of the International Conference on Neural Information Processing Systems. 839\u2013850."},{"key":"e_1_3_1_4_2","volume-title":"Proceedings of the Conference on Empirical Methods in Natural Language Processing","author":"Cao Meng","year":"2021","unstructured":"Meng Cao, Long Chen, Mike Zheng Shou, Can Zhang, and Yuexian Zou. 2021. On pursuit of designing multi-modal transformer for video grounding. In Proceedings of the Conference on Empirical Methods in Natural Language Processing."},{"key":"e_1_3_1_5_2","first-page":"4724","volume-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","author":"Carreira Jo\u00e3o","year":"2017","unstructured":"Jo\u00e3o Carreira and Andrew Zisserman. 2017. Quo vadis, action recognition? A new model and the kinetics dataset. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 4724\u20134733."},{"key":"e_1_3_1_6_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D18-1015"},{"key":"e_1_3_1_7_2","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i07.6627"},{"key":"e_1_3_1_8_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.01081"},{"key":"e_1_3_1_9_2","article-title":"Counterfactual samples synthesizing and training for robust visual question answering","volume":"2110","author":"Chen Long","year":"2021","unstructured":"Long Chen, Yuhang Zheng, Yulei Niu, Hanwang Zhang, and Jun Xiao. 2021. Counterfactual samples synthesizing and training for robust visual question answering. ArXiv preprint abs\/2110.01013 (2021).","journal-title":"ArXiv preprint"},{"key":"e_1_3_1_10_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1418"},{"key":"e_1_3_1_11_2","first-page":"3063","volume-title":"Proceedings of the International Conference on Neural Information Processing Systems","author":"Duan Xuguang","year":"2018","unstructured":"Xuguang Duan, Wen-bing Huang, Chuang Gan, Jingdong Wang, Wenwu Zhu, and Junzhou Huang. 2018. Weakly supervised dense event captioning in videos. In Proceedings of the International Conference on Neural Information Processing Systems. 3063\u20133073."},{"key":"e_1_3_1_12_2","first-page":"5277","volume-title":"Proceedings of the IEEE\/CVF International Conference on Computer Vision","author":"Gao Jiyang","year":"2017","unstructured":"Jiyang Gao, Chen Sun, Zhenheng Yang, and Ram Nevatia. 2017. TALL: Temporal activity localization via language query. In Proceedings of the IEEE\/CVF International Conference on Computer Vision. 5277\u20135285."},{"key":"e_1_3_1_13_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00155"},{"key":"e_1_3_1_14_2","first-page":"1481","volume-title":"Proceedings of the Conference on Empirical Methods in Natural Language Processing","author":"Gao Mingfei","year":"2019","unstructured":"Mingfei Gao, Larry Davis, Richard Socher, and Caiming Xiong. 2019. WSLLN:Weakly supervised natural language localization networks. In Proceedings of the Conference on Empirical Methods in Natural Language Processing. 1481\u20131487."},{"key":"e_1_3_1_15_2","first-page":"245","volume-title":"IEEE Winter Conference on Applications of Computer Vision (WACV)","author":"Ge Runzhou","year":"2019","unstructured":"Runzhou Ge, Jiyang Gao, Kan Chen, and Ram Nevatia. 2019. MAC: Mining activity concepts for language-based temporal localization. In IEEE Winter Conference on Applications of Computer Vision (WACV). 245\u2013253."},{"key":"e_1_3_1_16_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.63"},{"key":"e_1_3_1_17_2","first-page":"1","volume-title":"Proceedings of the Second Workshop on Shortcomings in Vision and Language","author":"Grand Gabriel","year":"2019","unstructured":"Gabriel Grand and Yonatan Belinkov. 2019. Adversarial regularization for visual question answering: Strengths, shortcomings, and side effects. In Proceedings of the Second Workshop on Shortcomings in Vision and Language. 1\u201313."},{"key":"e_1_3_1_18_2","volume-title":"The British Machine Vision Conference (BMVC)","author":"Hahn Meera","year":"2020","unstructured":"Meera Hahn, Asim Kadav, James M. Rehg, and Hans Peter Graf. 2020. Tripping through time: Efficient localization of activities in videos. In The British Machine Vision Conference (BMVC)."},{"key":"e_1_3_1_19_2","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v33i01.33018393"},{"key":"e_1_3_1_20_2","first-page":"961","volume-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","author":"Heilbron Fabian Caba","year":"2015","unstructured":"Fabian Caba Heilbron, Victor Escorcia, Bernard Ghanem, and Juan Carlos Niebles. 2015. ActivityNet: A Large-scale video benchmark for human activity understanding. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 961\u2013970."},{"key":"e_1_3_1_21_2","first-page":"5804","volume-title":"Proceedings of the IEEE\/CVF International Conference on Computer Vision","author":"Hendricks Lisa Anne","year":"2017","unstructured":"Lisa Anne Hendricks, Oliver Wang, Eli Shechtman, Josef Sivic, Trevor Darrell, and Bryan C. Russell. 2017. Localizing moments in video with natural language. In Proceedings of the IEEE\/CVF International Conference on Computer Vision. 5804\u20135813."},{"key":"e_1_3_1_22_2","doi-asserted-by":"publisher","DOI":"10.1162\/neco.1997.9.8.1735"},{"key":"e_1_3_1_23_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00711"},{"key":"e_1_3_1_24_2","doi-asserted-by":"publisher","DOI":"10.1145\/3323873.3325019"},{"key":"e_1_3_1_25_2","volume-title":"International Conference on Learning Representations, ICLR","author":"Kingma Diederik P.","year":"2015","unstructured":"Diederik P. Kingma and Jimmy Ba. 2015. Adam: A method for stochastic optimization. In International Conference on Learning Representations, ICLR."},{"key":"e_1_3_1_26_2","first-page":"706","volume-title":"Proceedings of the IEEE\/CVF International Conference on Computer Vision","author":"Krishna Ranjay","year":"2017","unstructured":"Ranjay Krishna, Kenji Hata, Frederic Ren, Li Fei-Fei, and Juan Carlos Niebles. 2017. Dense-captioning events in videos. In Proceedings of the IEEE\/CVF International Conference on Computer Vision. 706\u2013715."},{"key":"e_1_3_1_27_2","doi-asserted-by":"publisher","DOI":"10.1145\/3209978.3210003"},{"key":"e_1_3_1_28_2","doi-asserted-by":"publisher","DOI":"10.1145\/3240508.3240549"},{"key":"e_1_3_1_29_2","first-page":"5144","volume-title":"Proceedings of the Conference on Empirical Methods in Natural Language Processing","author":"Lu Chujie","year":"2019","unstructured":"Chujie Lu, Long Chen, Chilie Tan, Xiaolin Li, and Jun Xiao. 2019. DEBUG: A dense bottom-up grounding approach for natural language video localization. In Proceedings of the Conference on Empirical Methods in Natural Language Processing. 5144\u20135153."},{"key":"e_1_3_1_30_2","first-page":"11592","volume-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","author":"Mithun Niluthpol Chowdhury","year":"2019","unstructured":"Niluthpol Chowdhury Mithun, Sujoy Paul, and Amit K. Roy-Chowdhury. 2019. Weakly supervised video moment retrieval from text queries. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 11592\u201311601."},{"key":"e_1_3_1_31_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00279"},{"key":"e_1_3_1_32_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01251"},{"key":"e_1_3_1_33_2","volume-title":"The British Machine Vision Conference (BMVC)","author":"Otani Mayu","year":"2020","unstructured":"Mayu Otani, Yuta Nakashima, Esa Rahtu, and Janne Heikkil\u00e4. 2020. Uncovering hidden challenges in query-based video moment retrieval. In The British Machine Vision Conference (BMVC)."},{"key":"e_1_3_1_34_2","doi-asserted-by":"crossref","first-page":"44","DOI":"10.1631\/FITEE.1601787","article-title":"Cross-media analysis and reasoning: Advances and directions","author":"Peng Yu-xin","year":"2017","unstructured":"Yu-xin Peng, Wen-wu Zhu, Yao Zhao, Chang-sheng Xu, Qing-ming Huang, Han-qing Lu, Qing-hua Zheng, Tie-jun Huang, and Wen Gao. 2017. Cross-media analysis and reasoning: Advances and directions. Frontiers of Information Technology & Electronic Engineering (2017), 44\u201357.","journal-title":"Frontiers of Information Technology & Electronic Engineering"},{"key":"e_1_3_1_35_2","doi-asserted-by":"publisher","DOI":"10.3115\/v1\/D14-1162"},{"key":"e_1_3_1_36_2","first-page":"1548","volume-title":"Proceedings of the International Conference on Neural Information Processing Systems","author":"Ramakrishnan Sainandan","year":"2018","unstructured":"Sainandan Ramakrishnan, Aishwarya Agrawal, and Stefan Lee. 2018. Overcoming language priors in visual question answering with adversarial regularization. In Proceedings of the International Conference on Neural Information Processing Systems. 1548\u20131558."},{"key":"e_1_3_1_37_2","doi-asserted-by":"publisher","DOI":"10.1162\/tacl_a_00207"},{"key":"e_1_3_1_38_2","first-page":"510","volume-title":"Proceedings of the European Conference on Computer Vision (ECCV)","author":"Sigurdsson Gunnar A.","year":"2016","unstructured":"Gunnar A. Sigurdsson, G\u00fcl Varol, Xiaolong Wang, Ali Farhadi, Ivan Laptev, and Abhinav Gupta. 2016. Hollywood in homes: Crowdsourcing data collection for activity understanding. In Proceedings of the European Conference on Computer Vision (ECCV). 510\u2013526."},{"key":"e_1_3_1_39_2","volume-title":"Pacific Rim Conference on Multimedia","author":"Song Xiaomeng","year":"2018","unstructured":"Xiaomeng Song and Yahong Han. 2018. VAL: Visual-attention action localizer. In Pacific Rim Conference on Multimedia."},{"key":"e_1_3_1_40_2","article-title":"Weakly-supervised multi-level attentional reconstruction network for grounding textual queries in videos","volume":"2003","author":"Song Yijun","year":"2020","unstructured":"Yijun Song, Jingwen Wang, Lin Ma, Zhou Yu, and Jun Yu. 2020. Weakly-supervised multi-level attentional reconstruction network for grounding textual queries in videos. ArXiv preprint abs\/2003.07048 (2020).","journal-title":"ArXiv preprint"},{"key":"e_1_3_1_41_2","volume-title":"arXiv","author":"Tan Reuben","year":"2019","unstructured":"Reuben Tan, Huijuan Xu, Kate Saenko, and Bryan A. Plummer. 2019. wMAN: Weakly-supervised moment alignment network for text-based video segment retrieval. In arXiv."},{"key":"e_1_3_1_42_2","first-page":"3713","volume-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","author":"Tang Kaihua","year":"2020","unstructured":"Kaihua Tang, Yulei Niu, Jianqiang Huang, Jiaxin Shi, and Hanwang Zhang. 2020. Unbiased scene graph generation from biased training. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 3713\u20133722."},{"key":"e_1_3_1_43_2","first-page":"4489","volume-title":"Proceedings of the IEEE\/CVF International Conference on Computer Vision","author":"Tran Du","year":"2015","unstructured":"Du Tran, Lubomir D. Bourdev, Rob Fergus, Lorenzo Torresani, and Manohar Paluri. 2015. Learning spatiotemporal features with 3D convolutional networks. In Proceedings of the IEEE\/CVF International Conference on Computer Vision. 4489\u20134497."},{"key":"e_1_3_1_44_2","first-page":"5998","volume-title":"Proceedings of the International Conference on Neural Information Processing Systems","author":"Vaswani Ashish","year":"2017","unstructured":"Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. In Proceedings of the International Conference on Neural Information Processing Systems. 5998\u20136008."},{"key":"e_1_3_1_45_2","first-page":"334","volume-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","author":"Wang Weining","year":"2019","unstructured":"Weining Wang, Yan Huang, and Liang Wang. 2019. Language-driven temporal activity localization: A semantic matching reinforcement learning model. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 334\u2013343."},{"key":"e_1_3_1_46_2","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i07.6924"},{"key":"e_1_3_1_47_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICME51207.2021.9428369"},{"key":"e_1_3_1_48_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.emnlp-main.327"},{"key":"e_1_3_1_49_2","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i4.16406"},{"key":"e_1_3_1_50_2","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v33i01.33019062"},{"key":"e_1_3_1_51_2","doi-asserted-by":"publisher","DOI":"10.1145\/3404835.3462823"},{"key":"e_1_3_1_52_2","article-title":"Deconfounded image captioning: A causal retrospect","author":"Yang Xu","year":"2021","unstructured":"Xu Yang, Hanwang Zhang, and Jianfei Cai. 2021. Deconfounded image captioning: A causal retrospect. IEEE Transactions on Pattern Analysis and Machine Intelligence (2021).","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence"},{"key":"e_1_3_1_53_2","doi-asserted-by":"publisher","DOI":"10.1145\/3475723.3484247"},{"key":"e_1_3_1_54_2","first-page":"534","volume-title":"Proceedings of the International Conference on Neural Information Processing Systems","author":"Yuan Yitian","year":"2019","unstructured":"Yitian Yuan, Lin Ma, Jingwen Wang, Wei Liu, and Wenwu Zhu. 2019. Semantic conditioned dynamic modulation for temporal sentence grounding in videos. In Proceedings of the International Conference on Neural Information Processing Systems. 534\u2013544."},{"key":"e_1_3_1_55_2","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v33i01.33019159"},{"key":"e_1_3_1_56_2","first-page":"10284","volume-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","author":"Zeng Runhao","year":"2020","unstructured":"Runhao Zeng, Haoming Xu, Wenbing Huang, Peihao Chen, Mingkui Tan, and Chuang Gan. 2020. Dense regression network for video grounding. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 10284\u201310293."},{"key":"e_1_3_1_57_2","first-page":"1247","volume-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","author":"Zhang Da","year":"2019","unstructured":"Da Zhang, Xiyang Dai, Xin Wang, Yuan-Fang Wang, and Larry S. Davis. 2019. MAN: Moment alignment network for natural language moment retrieval via iterative graph adjustment. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 1247\u20131257."},{"key":"e_1_3_1_58_2","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i07.6984"},{"key":"e_1_3_1_59_2","doi-asserted-by":"publisher","DOI":"10.1145\/3331184.3331235"},{"key":"e_1_3_1_60_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00834"},{"key":"e_1_3_1_61_2","article-title":"Multimedia big data computing","author":"Zhu Wenwu","year":"2015","unstructured":"Wenwu Zhu, Peng Cui, Zhi Wang, and Gang Hua. 2015. Multimedia big data computing. IEEE Multimedia (2015).","journal-title":"IEEE Multimedia"}],"container-title":["ACM Transactions on Multimedia Computing, Communications, and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3565573","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3565573","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T17:48:51Z","timestamp":1750182531000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3565573"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,7,12]]},"references-count":60,"journal-issue":{"issue":"6","published-print":{"date-parts":[[2023,11,30]]}},"alternative-id":["10.1145\/3565573"],"URL":"https:\/\/doi.org\/10.1145\/3565573","relation":{},"ISSN":["1551-6857","1551-6865"],"issn-type":[{"value":"1551-6857","type":"print"},{"value":"1551-6865","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,7,12]]},"assertion":[{"value":"2022-02-23","order":0,"name":"received","label":"Received","group":{"name":"publication_history","label":"Publication History"}},{"value":"2022-08-03","order":1,"name":"accepted","label":"Accepted","group":{"name":"publication_history","label":"Publication History"}},{"value":"2023-07-12","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}