{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,4]],"date-time":"2026-03-04T03:16:13Z","timestamp":1772594173435,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":74,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,7,19]],"date-time":"2023-07-19T00:00:00Z","timestamp":1689724800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,7,19]]},"DOI":"10.1145\/3571884.3603754","type":"proceedings-article","created":{"date-parts":[[2023,7,17]],"date-time":"2023-07-17T18:04:11Z","timestamp":1689617051000},"page":"1-6","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":33,"title":["Deceptive AI Ecosystems: The Case of ChatGPT"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-1755-0976","authenticated-orcid":false,"given":"Xiao","family":"Zhan","sequence":"first","affiliation":[{"name":"Informatics, King's College London, United Kingdom"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2303-1531","authenticated-orcid":false,"given":"Yifan","family":"Xu","sequence":"additional","affiliation":[{"name":"Computer Science, University of Machester, United Kingdom"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3999-528X","authenticated-orcid":false,"given":"Stefan","family":"Sarkadi","sequence":"additional","affiliation":[{"name":"Informatics, King's College London, United Kingdom"}]}],"member":"320","published-online":{"date-parts":[[2023,7,19]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1109\/TTS.2022.3179756"},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.mlwa.2020.100006"},{"key":"e_1_3_2_1_3_1","volume-title":"AIAI 2020, Neos Marmaras, Greece, June 5\u20137, 2020, Proceedings, Part II 16","author":"Adamopoulou Eleni","year":"2020","unstructured":"Eleni Adamopoulou and Lefteris Moussiades. 2020. An overview of chatbot technology. In Artificial Intelligence Applications and Innovations: 16th IFIP WG 12.5 International Conference, AIAI 2020, Neos Marmaras, Greece, June 5\u20137, 2020, Proceedings, Part II 16. Springer, 373\u2013383."},{"key":"e_1_3_2_1_4_1","unstructured":"HLEG AI. 2019. High-level expert group on artificial intelligence. 6\u00a0pages."},{"key":"e_1_3_2_1_5_1","first-page":"23716","article-title":"Flamingo: a visual language model for few-shot learning","volume":"35","author":"Alayrac Jean-Baptiste","year":"2022","unstructured":"Jean-Baptiste Alayrac, Jeff Donahue, Pauline Luc, Antoine Miech, Iain Barr, Yana Hasson, Karel Lenc, Arthur Mensch, Katherine Millican, Malcolm Reynolds, 2022. Flamingo: a visual language model for few-shot learning. Advances in Neural Information Processing Systems 35 (2022), 23716\u201323736.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_6_1","first-page":"27","article-title":"A Current Regime of Uncertainty: Improving Assessments of Liability for Damages Caused by Artificial Intelligence","volume":"21","author":"Alshanteer Mousa","year":"2019","unstructured":"Mousa Alshanteer. 2019. A Current Regime of Uncertainty: Improving Assessments of Liability for Damages Caused by Artificial Intelligence. NCJL & Tech. 21 (2019), 27.","journal-title":"NCJL & Tech."},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1038\/scientificamerican1155-31"},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijinfomgt.2021.102433"},{"key":"e_1_3_2_1_9_1","volume-title":"Is ChatGPT Leading Generative AI? What is Beyond Expectations?What is beyond expectations","author":"Ayd\u0131n \u00d6mer","year":"2023","unstructured":"\u00d6mer Ayd\u0131n and Enis Karaarslan. 2023. Is ChatGPT Leading Generative AI? What is Beyond Expectations?What is beyond expectations (2023)."},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"crossref","unstructured":"Amos Azaria. 2022. ChatGPT Usage and Limitations. (2022).","DOI":"10.31219\/osf.io\/5ue7n"},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"crossref","unstructured":"Nathan Ballantyne. 2019. Epistemic trespassing. (2019).","DOI":"10.1093\/oso\/9780190847289.003.0008"},{"key":"e_1_3_2_1_12_1","volume-title":"Psychological Aspects of AI. An Introduction to Ethics in Robotics and AI","author":"Bartneck Christoph","year":"2021","unstructured":"Christoph Bartneck, Christoph L\u00fctge, Alan Wagner, Sean Welsh, Christoph Bartneck, Christoph L\u00fctge, Alan Wagner, and Sean Welsh. 2021. Psychological Aspects of AI. An Introduction to Ethics in Robotics and AI (2021), 55\u201360."},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445922"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.2196\/11510"},{"key":"e_1_3_2_1_15_1","volume-title":"A categorical archive of ChatGPT failures. arXiv preprint arXiv:2302.03494","author":"Borji Ali","year":"2023","unstructured":"Ali Borji. 2023. A categorical archive of ChatGPT failures. arXiv preprint arXiv:2302.03494 (2023)."},{"key":"e_1_3_2_1_16_1","unstructured":"Siobhan Calafiore. 2023. Beware ChatGPT\u2019s deceptive medical information researchers warn. https:\/\/thelimbic.com\/beware-chatgpts-deceptive-medical-information-researchers-warn\/."},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.3390\/info13010041"},{"key":"e_1_3_2_1_18_1","first-page":"53","article-title":"Creating ethical chatbots","volume":"101","author":"Castelluccio Michael","year":"2019","unstructured":"Michael Castelluccio. 2019. Creating ethical chatbots. Strategic Finance 101, 6 (2019), 53\u201355.","journal-title":"Strategic Finance"},{"key":"e_1_3_2_1_19_1","volume-title":"Cognitive illusions and the lying machine: a blueprint for sophistic mendacity. Ph.\u00a0D. Dissertation","author":"Clark H","unstructured":"Micah\u00a0H Clark. 2010. Cognitive illusions and the lying machine: a blueprint for sophistic mendacity. Ph.\u00a0D. Dissertation. Rensselaer Polytechnic Institute."},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1145\/3342775.3342786"},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.54097\/fcis.v2i2.4465"},{"key":"e_1_3_2_1_22_1","volume-title":"Segment Anything Model (SAM) for Digital Pathology: Assess Zero-shot Segmentation on Whole Slide Imaging. arXiv preprint arXiv:2304.04155","author":"Deng Ruining","year":"2023","unstructured":"Ruining Deng, Can Cui, Quan Liu, Tianyuan Yao, Lucas\u00a0W Remedios, Shunxing Bao, Bennett\u00a0A Landman, Lee\u00a0E Wheless, Lori\u00a0A Coburn, Keith\u00a0T Wilson, 2023. Segment Anything Model (SAM) for Digital Pathology: Assess Zero-shot Segmentation on Whole Slide Imaging. arXiv preprint arXiv:2304.04155 (2023)."},{"key":"e_1_3_2_1_23_1","unstructured":"Hillemann Dennis and Zimprich Stephan. 2023. ChatGPT - legal challenges legal opportunities. https:\/\/www.fieldfisher.com\/en\/insights\/chatgpt-legal-challenges-legal-opportunities."},{"key":"e_1_3_2_1_24_1","volume-title":"What\u2019s wrong with epistemic trespassing?Philosophical Studies 179, 1","author":"DiPaolo Joshua","year":"2022","unstructured":"Joshua DiPaolo. 2022. What\u2019s wrong with epistemic trespassing?Philosophical Studies 179, 1 (2022), 223\u2013243."},{"key":"e_1_3_2_1_25_1","unstructured":"Mariama\u00a0Corca Djalo\u00a0Djalo. 2023. KAI: An AI-powered Chatbot To Support Therapy. B.S. thesis. Universitat Polit\u00e8cnica de Catalunya."},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1145\/3543829.3544525"},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"crossref","unstructured":"L Festinger and JM Carlsmith. 1959. Cognitive consequences of forced compliance. e Journal of Abnormal and Social Psychology 58 (2) 203-210.","DOI":"10.1037\/h0041593"},{"key":"e_1_3_2_1_28_1","unstructured":"Alex Friedland. 2023. Stopping dangerous AI \u2014 two public letters frame the debate Biden discusses AI risks Bloomberg trains an LLM on its own data and White House emerging tech announcements. https:\/\/cset.georgetown.edu\/newsletter\/april-6-2023\/"},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10676-004-2895-2"},{"key":"e_1_3_2_1_30_1","volume-title":"How Close is ChatGPT to Human Experts? Comparison Corpus, Evaluation, and Detection. arXiv preprint arXiv:2301.07597","author":"Guo Biyang","year":"2023","unstructured":"Biyang Guo, Xin Zhang, Ziyuan Wang, Minqi Jiang, Jinran Nie, Yuxuan Ding, Jianwei Yue, and Yupeng Wu. 2023. How Close is ChatGPT to Human Experts? Comparison Corpus, Evaluation, and Detection. arXiv preprint arXiv:2301.07597 (2023)."},{"key":"e_1_3_2_1_31_1","volume-title":"Language is not all you need: Aligning perception with language models. arXiv preprint arXiv:2302.14045","author":"Huang Shaohan","year":"2023","unstructured":"Shaohan Huang, Li Dong, Wenhui Wang, Yaru Hao, Saksham Singhal, Shuming Ma, Tengchao Lv, Lei Cui, Owais\u00a0Khan Mohammed, Qiang Liu, 2023. Language is not all you need: Aligning perception with language models. arXiv preprint arXiv:2302.14045 (2023)."},{"key":"e_1_3_2_1_32_1","unstructured":"IBM. 2018. IBM Watson. https:\/\/www.ibm.com\/watson."},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.1145\/3543829.3543836"},{"key":"e_1_3_2_1_34_1","volume-title":"When is it permissible for artificial intelligence to lie? A trust-based approach. arXiv preprint arXiv:2103.05434","author":"Kim Tae\u00a0Wan","year":"2021","unstructured":"Tae\u00a0Wan Kim, Kyusong Lee, Zhaoqi Cheng, Yanhan Tang, John Hooker, 2021. When is it permissible for artificial intelligence to lie? A trust-based approach. arXiv preprint arXiv:2103.05434 (2021)."},{"key":"e_1_3_2_1_35_1","volume-title":"Anthropomorphism of computers: Is it mindful or mindless?Computers in Human Behavior 28, 1","author":"Kim Youjeong","year":"2012","unstructured":"Youjeong Kim and S\u00a0Shyam Sundar. 2012. Anthropomorphism of computers: Is it mindful or mindless?Computers in Human Behavior 28, 1 (2012), 241\u2013250."},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.1145\/3405755.3406124"},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1145\/3543829.3543835"},{"key":"e_1_3_2_1_38_1","volume-title":"What Makes Good In-Context Examples for GPT-3 ?arXiv preprint arXiv:2101.06804","author":"Liu Jiachang","year":"2021","unstructured":"Jiachang Liu, Dinghan Shen, Yizhe Zhang, Bill Dolan, Lawrence Carin, and Weizhu Chen. 2021. What Makes Good In-Context Examples for GPT-3 ?arXiv preprint arXiv:2101.06804 (2021)."},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1145\/3613250"},{"key":"e_1_3_2_1_40_1","doi-asserted-by":"crossref","unstructured":"Peta Masters and Sebastian Sardina. 2017. Deceptive Path-Planning.. In IJCAI. 4368\u20134375.","DOI":"10.24963\/ijcai.2017\/610"},{"key":"e_1_3_2_1_41_1","volume-title":"The radicalization risks of GPT-3 and advanced neural language models. arXiv preprint arXiv:2009.06807","author":"McGuffie Kris","year":"2020","unstructured":"Kris McGuffie and Alex Newhouse. 2020. The radicalization risks of GPT-3 and advanced neural language models. arXiv preprint arXiv:2009.06807 (2020)."},{"key":"e_1_3_2_1_42_1","volume-title":"Retrieved","author":"Milmo Dan","year":"2023","unstructured":"Dan Milmo. 2023. ChatGPT reaches 100 million users two months after launch. Retrieved Feb 24, 2023 from https:\/\/www.theguardian.com\/technology\/2023\/feb\/02\/chatgpt-100-million-users-open-ai-fastest-growing-app"},{"key":"e_1_3_2_1_43_1","volume-title":"Smartphone-based conversational agents and responses to questions about mental health, interpersonal violence, and physical health. JAMA internal medicine 176, 5","author":"Miner S","year":"2016","unstructured":"Adam\u00a0S Miner, Arnold Milstein, Stephen Schueller, Roshini Hegde, Christina Mangurian, and Eleni Linos. 2016. Smartphone-based conversational agents and responses to questions about mental health, interpersonal violence, and physical health. JAMA internal medicine 176, 5 (2016), 619\u2013625."},{"key":"e_1_3_2_1_44_1","volume-title":"Deceitful media: Artificial intelligence and social life after the Turing test","author":"Natale Simone","unstructured":"Simone Natale. 2021. Deceitful media: Artificial intelligence and social life after the Turing test. Oxford University Press, USA."},{"key":"e_1_3_2_1_45_1","doi-asserted-by":"publisher","DOI":"10.1002\/pra2.487"},{"key":"e_1_3_2_1_46_1","unstructured":"Susan\u00a0A. Nolan. 2023. Learning to Lie: The Perils of ChatGPT. https:\/\/www.psychologytoday.com\/intl\/blog\/misinformation-desk\/202303\/learning-to-lie-the-perils-of-chatgpt."},{"key":"e_1_3_2_1_47_1","unstructured":"Future of Life\u00a0Institute. 2023. Pause Giant AI Experiments: An Open Letter. https:\/\/futureoflife.org\/open-letter\/pause-giant-ai-experiments\/"},{"key":"e_1_3_2_1_48_1","unstructured":"OpenAI. 2022. ChatGPT. https:\/\/openai.com\/blog\/chatgpt"},{"key":"e_1_3_2_1_49_1","unstructured":"OpenAI. 2023. GPT-4. https:\/\/openai.com\/product\/gpt-4"},{"key":"e_1_3_2_1_50_1","volume-title":"Training language models to follow instructions with human feedback. arXiv preprint arXiv:2203.02155","author":"Ouyang Long","year":"2022","unstructured":"Long Ouyang, Jeff Wu, Xu Jiang, Diogo Almeida, Carroll\u00a0L Wainwright, Pamela Mishkin, Chong Zhang, Sandhini Agarwal, Katarina Slama, Alex Ray, 2022. Training language models to follow instructions with human feedback. arXiv preprint arXiv:2203.02155 (2022)."},{"key":"e_1_3_2_1_51_1","volume-title":"Proceedings of the 20th International TRUST Workshop @ IJCAI\/AAMAS\/ECAI\/ICML. CEUR Workshop Proceedings","author":"Panisson R.","year":"2018","unstructured":"Alison\u00a0R. Panisson, Stefan Sarkadi, Peter McBurney, Simon Parsons, and Rafael\u00a0H. Bordini. 2018. Lies, Bullshit, and Deception in Agent-Oriented Programming Languages. In Proceedings of the 20th International TRUST Workshop @ IJCAI\/AAMAS\/ECAI\/ICML. CEUR Workshop Proceedings, Stockholm, Sweden, 50\u201361."},{"key":"e_1_3_2_1_52_1","unstructured":"Katyanna Quach. 2020. Researchers made an OpenAI GPT-3 medical chatbot as an experiment. It told a mock patient to kill themselves. https:\/\/www.theregister.com\/2020\/10\/28\/gpt3_medical_chatbot_experiment\/."},{"key":"e_1_3_2_1_53_1","volume-title":"Machine behaviour. Nature 568, 7753","author":"Rahwan Iyad","year":"2019","unstructured":"Iyad Rahwan, Manuel Cebrian, Nick Obradovich, Josh Bongard, Jean-Fran\u00e7ois Bonnefon, Cynthia Breazeal, Jacob\u00a0W Crandall, Nicholas\u00a0A Christakis, Iain\u00a0D Couzin, Matthew\u00a0O Jackson, 2019. Machine behaviour. Nature 568, 7753 (2019), 477."},{"key":"e_1_3_2_1_54_1","volume-title":"Object hallucination in image captioning. arXiv preprint arXiv:1809.02156","author":"Rohrbach Anna","year":"2018","unstructured":"Anna Rohrbach, Lisa\u00a0Anne Hendricks, Kaylee Burns, Trevor Darrell, and Kate Saenko. 2018. Object hallucination in image captioning. arXiv preprint arXiv:1809.02156 (2018)."},{"key":"e_1_3_2_1_55_1","volume-title":"ChatGPT: Bullshit spewer or the end of traditional assessments in higher education?Journal of Applied Learning and Teaching 6, 1","author":"Rudolph J\u00fcrgen","year":"2023","unstructured":"J\u00fcrgen Rudolph, Samson Tan, and Shannon Tan. 2023. ChatGPT: Bullshit spewer or the end of traditional assessments in higher education?Journal of Applied Learning and Teaching 6, 1 (2023)."},{"key":"e_1_3_2_1_56_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-1-4842-5741-8"},{"key":"e_1_3_2_1_57_1","volume-title":"Proc. of the 22nd International Conference on Autonomous Agents and Multiagent Systems (AAMAS","author":"Sarkadi Stefan","year":"2023","unstructured":"Stefan Sarkadi, Peidong Mei, and Edmond Awad. 2023. Should my agent lie for me? A study on attitudes of US-based participants towards deceptive AI in selected future-of-work scenarios. In Proc. of the 22nd International Conference on Autonomous Agents and Multiagent Systems (AAMAS 2023). International Foundation for Autonomous Agents and Multiagent Systems (IFAAMAS)."},{"key":"e_1_3_2_1_58_1","doi-asserted-by":"publisher","DOI":"10.3233\/AIC-190615"},{"key":"e_1_3_2_1_59_1","doi-asserted-by":"publisher","DOI":"10.1098\/rsos.201032"},{"key":"e_1_3_2_1_60_1","doi-asserted-by":"publisher","DOI":"10.1145\/3543829.3544521"},{"key":"e_1_3_2_1_61_1","doi-asserted-by":"publisher","DOI":"10.1145\/3543829.3544513"},{"key":"e_1_3_2_1_62_1","doi-asserted-by":"publisher","DOI":"10.1145\/3479515"},{"key":"e_1_3_2_1_63_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.chbr.2021.100092"},{"key":"e_1_3_2_1_64_1","doi-asserted-by":"publisher","DOI":"10.1631\/FITEE.1700826"},{"key":"e_1_3_2_1_65_1","volume-title":"ChatGPT: The End of Online Exam Integrity?arXiv preprint arXiv:2212.09292","author":"Susnjak Teo","year":"2022","unstructured":"Teo Susnjak. 2022. ChatGPT: The End of Online Exam Integrity?arXiv preprint arXiv:2212.09292 (2022)."},{"key":"e_1_3_2_1_66_1","doi-asserted-by":"publisher","DOI":"10.1145\/3405755.3406123"},{"key":"e_1_3_2_1_67_1","unstructured":"Hsu Tiffany and Thompson Stuart\u00a0A.2023. Disinformation Researchers Raise Alarms About A.I. Chatbots. https:\/\/www.nytimes.com\/2023\/02\/08\/technology\/ai-chatbots-disinformation.html."},{"key":"e_1_3_2_1_68_1","volume-title":"LLaMA: Open and Efficient Foundation Language Models. arXiv:2302.13971","author":"Touvron Hugo","year":"2023","unstructured":"Hugo Touvron, Thibaut Lavril, Gautier Izacard, Xavier Martinet, Marie-Anne Lachaux, Timoth\u00e9e Lacroix, Baptiste Rozi\u00e8re, Naman Goyal, Eric Hambro, Faisal Azhar, Aurelien Rodriguez, Armand Joulin, Edouard Grave, and Guillaume Lample. 2023. LLaMA: Open and Efficient Foundation Language Models. arXiv:2302.13971 (2023)."},{"key":"e_1_3_2_1_69_1","unstructured":"Salil Tripathi. 2023. We asked ChatGPT about its impact on human rights and business. Here\u2019s what it told us. https:\/\/www.ihrb.org\/focus-areas\/information-communication-technology\/we-asked-chatgpt-about-its-impact-on-human-rights-on-business-heres-what-it-told-us."},{"key":"e_1_3_2_1_70_1","unstructured":"Ido Vock. 2022. ChatGPT proves that AI still has a racism problem. https:\/\/www.newstatesman.com\/quickfire\/2022\/12\/chatgpt-shows-ai-racism-problem."},{"key":"e_1_3_2_1_71_1","volume-title":"Ethical and social risks of harm from language models. arXiv preprint arXiv:2112.04359","author":"Weidinger Laura","year":"2021","unstructured":"Laura Weidinger, John Mellor, Maribeth Rauh, Conor Griffin, Jonathan Uesato, Po-Sen Huang, Myra Cheng, Mia Glaese, Borja Balle, Atoosa Kasirzadeh, 2021. Ethical and social risks of harm from language models. arXiv preprint arXiv:2112.04359 (2021)."},{"key":"e_1_3_2_1_72_1","doi-asserted-by":"publisher","DOI":"10.1145\/365153.365168"},{"key":"e_1_3_2_1_73_1","volume-title":"On hallucination and predictive uncertainty in conditional language generation. arXiv preprint arXiv:2103.15025","author":"Xiao Yijun","year":"2021","unstructured":"Yijun Xiao and William\u00a0Yang Wang. 2021. On hallucination and predictive uncertainty in conditional language generation. arXiv preprint arXiv:2103.15025 (2021)."},{"key":"e_1_3_2_1_74_1","doi-asserted-by":"publisher","DOI":"10.1007\/s12369-014-0267-6"}],"event":{"name":"CUI '23: ACM conference on Conversational User Interfaces","location":"Eindhoven Netherlands","acronym":"CUI '23","sponsor":["SIGCHI ACM Special Interest Group on Computer-Human Interaction"]},"container-title":["Proceedings of the 5th International Conference on Conversational User Interfaces"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3571884.3603754","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3571884.3603754","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T17:51:13Z","timestamp":1750182673000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3571884.3603754"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,7,19]]},"references-count":74,"alternative-id":["10.1145\/3571884.3603754","10.1145\/3571884"],"URL":"https:\/\/doi.org\/10.1145\/3571884.3603754","relation":{},"subject":[],"published":{"date-parts":[[2023,7,19]]},"assertion":[{"value":"2023-07-19","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}