{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T17:09:41Z","timestamp":1776100181340,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":74,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,10,17]],"date-time":"2024-10-17T00:00:00Z","timestamp":1729123200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"National Science Foundation","award":["CCF-1909963, CNS-2120350, and III-2311598"],"award-info":[{"award-number":["CCF-1909963, CNS-2120350, and III-2311598"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,10,17]]},"DOI":"10.1145\/3723178.3723224","type":"proceedings-article","created":{"date-parts":[[2025,6,6]],"date-time":"2025-06-06T07:16:47Z","timestamp":1749194207000},"page":"346-354","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":11,"title":["Programming with AI: Evaluating ChatGPT, Gemini, AlphaCode, and GitHub Copilot for Programmers"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-9986-210X","authenticated-orcid":false,"given":"Md Kamrul","family":"Siam","sequence":"first","affiliation":[{"name":"Computer Science, New York Institute of Technology, New York, New York, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-9510-6696","authenticated-orcid":false,"given":"Huanying","family":"Gu","sequence":"additional","affiliation":[{"name":"Computer Science, New York Institute of Technology, New York, New York, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3968-9699","authenticated-orcid":false,"given":"Jerry Q.","family":"Cheng","sequence":"additional","affiliation":[{"name":"Computer Science, New York Institute of Technology, New York, New York, USA"}]}],"member":"320","published-online":{"date-parts":[[2025,6,6]]},"reference":[{"key":"e_1_3_3_2_2_2","unstructured":"[n.d.]. About GitHub Copilot Individual. https:\/\/docs.github.com\/en\/copilot\/copilot-individual\/about-github-copilot-individual"},{"key":"e_1_3_3_2_3_2","unstructured":"[n.d.]. Build with the Gemini API. https:\/\/ai.google.dev\/"},{"key":"e_1_3_3_2_4_2","unstructured":"[n.d.]. ChatGPT vs. Microsoft Copilot vs. Gemini: Which is the best AI chatbot?https:\/\/www.zdnet.com\/article\/chatgpt-vs-microsoft-copilot-vs-gemini-which-is-the-best-ai-chatbot\/"},{"key":"e_1_3_3_2_5_2","unstructured":"[n.d.]. Chord. https:\/\/chord.pub\/article\/39449\/how-to-use-alphacode"},{"key":"e_1_3_3_2_6_2","unstructured":"[n.d.]. Code and debug with Bard. https:\/\/blog.google\/technology\/ai\/code-with-bard\/"},{"key":"e_1_3_3_2_7_2","unstructured":"[n.d.]. Competitive programming with AlphaCode. https:\/\/deepmind.google\/discover\/blog\/competitive-programming-with-alphacode\/"},{"key":"e_1_3_3_2_8_2","unstructured":"[n.d.]. Gemini - Google DeepMind. https:\/\/deepmind.google\/technologies\/gemini\/"},{"key":"e_1_3_3_2_9_2","unstructured":"[n.d.]. Gemini: All About This Zodiac Sign\u2019s Personality Traits Compatibility and More. https:\/\/astrostyle.com\/astrology\/zodiac-signs\/gemini\/"},{"key":"e_1_3_3_2_10_2","unstructured":"[n.d.]. Getting started with GitHub Copilot. https:\/\/docs.github.com\/en\/copilot\/using-github-copilot\/getting-started-with-github-copilot"},{"key":"e_1_3_3_2_11_2","unstructured":"[n.d.]. GitHub Copilot vs. ChatGPT: Which is Better for Coding in 2024? ([n. d.])."},{"key":"e_1_3_3_2_12_2","unstructured":"[n.d.]. GPT-3.5 model architecture. https:\/\/iq.opengenus.org\/gpt-3-5-model\/"},{"key":"e_1_3_3_2_13_2","unstructured":"[n.d.]. Models comparison: OpenAI documentation. https:\/\/platform.openai.com\/docs\/models\/overview"},{"key":"e_1_3_3_2_14_2","unstructured":"[n.d.]. PLM ChatGPT and Large Language Model Thoughts. https:\/\/beyondplm.com\/2023\/01\/28\/plm-chatgpt-and-large-language-model-thoughts\/"},{"key":"e_1_3_3_2_15_2","unstructured":"[n.d.]. The purpose benefits and downsides of GitHub Copilot | Proxify.io. https:\/\/proxify.io\/articles\/what-is-github-copilot"},{"key":"e_1_3_3_2_16_2","unstructured":"[n.d.]. Quickstart for GitHub Copilot. https:\/\/docs.github.com\/en\/copilot\/quickstart"},{"key":"e_1_3_3_2_17_2","unstructured":"[n.d.]. Safeguarding Data Integrity and Privacy in the Age of LLMs | Sentra Blog. https:\/\/www.sentra.io\/blog\/safeguarding-data-integrity-and-privacy-in-the-age-of-ai-powered-large-language-models-llms"},{"key":"e_1_3_3_2_18_2","unstructured":"[n.d.]. The transformer architecture | Python. https:\/\/campus.datacamp.com\/courses\/introduction-to-llms-in-python\/the-large-language-models-llms-landscape?ex=7"},{"key":"e_1_3_3_2_19_2","unstructured":"[n.d.]. Understanding Transformers & the Architecture of LLMs. https:\/\/www.mlq.ai\/llm-transformer-architecture\/"},{"key":"e_1_3_3_2_20_2","unstructured":"[n.d.]. What Are Large Language Models (LLMs)? | IBM. https:\/\/www.ibm.com\/topics\/large-language-models"},{"key":"e_1_3_3_2_21_2","unstructured":"[n.d.]. What is ChatGPT and why does it matter? Here\u2019s what you need to know. https:\/\/www.zdnet.com\/article\/what-is-chatgpt-and-why-does-it-matter-heres-everything-you-need-to-know\/"},{"key":"e_1_3_3_2_22_2","unstructured":"2024. https:\/\/blog.google\/technology\/ai\/google-gemini-next-generation-model-february-2024\/"},{"key":"e_1_3_3_2_23_2","doi-asserted-by":"publisher","unstructured":"Syeda\u00a0Nahida Akter et\u00a0al. 2023. An In-depth Look at Gemini\u2019s Language Abilities. (2023). 10.48550\/ARXIV.2312.11444","DOI":"10.48550\/ARXIV.2312.11444"},{"key":"e_1_3_3_2_24_2","unstructured":"K.\u00a0C.\u00a0Sabreena Basheer. [n.d.]. Unleashing the Power of DeepMind\u2019s AlphaCode: Revolutionizing Code Writing. https:\/\/www.analyticsvidhya.com\/blog\/2023\/12\/unleashing-the-power-of-deepminds-alphacode-revolutionizing-code-writing\/"},{"key":"e_1_3_3_2_25_2","volume-title":"Advances in Neural Information Processing Systems","author":"Bengio Yoshua","year":"2000","unstructured":"Yoshua Bengio, R\u00e9jean Ducharme, and Pascal Vincent. 2000. A Neural Probabilistic Language Model. In Advances in Neural Information Processing Systems , T.\u00a0Leen, T.\u00a0Dietterich, and V.\u00a0Tresp (Eds.), Vol.\u00a013. MIT Press. https:\/\/proceedings.neurips.cc\/paper_files\/paper\/2000\/file\/728f206c2a01bf572b5940d7d9a8fa4c-Paper.pdf"},{"key":"e_1_3_3_2_26_2","doi-asserted-by":"publisher","unstructured":"Davide Castelvecchi. 2022. Are ChatGPT and AlphaCode going to replace programmers? Nature (Dec. 2022). 10.1038\/d41586-022-04383-z","DOI":"10.1038\/d41586-022-04383-z"},{"key":"e_1_3_3_2_27_2","unstructured":"Sam\u00a0McKay CFA. [n.d.]. How to Use Chat GPT: A Simple Guide for Beginners | Master Data Skills + AI. https:\/\/blog.enterprisedna.co\/how-to-use-chat-gpt\/"},{"key":"e_1_3_3_2_28_2","doi-asserted-by":"publisher","unstructured":"Abel Chandra Laura T\u00fcnnermann Tommy L\u00f6fstedt and Regina Gratz. [n.d.]. Transformer-based deep learning for predicting protein properties in the life sciences. 12 ([n. d.]) e82819. 10.7554\/eLife.82819","DOI":"10.7554\/eLife.82819"},{"key":"e_1_3_3_2_29_2","unstructured":"Jeffrey Dastin and Jeffrey Dastin. [n.d.]. Google rebrands Bard chatbot as Gemini rolls out paid subscription. ([n. d.]). https:\/\/www.reuters.com\/technology\/google-rebrands-bard-chatbot-gemini-rolls-out-paid-subscription-2024-02-08\/"},{"key":"e_1_3_3_2_30_2","unstructured":"Jeffrey Dastin and Jeffrey Dastin. [n.d.]. Google unveils ChatGPT rival Bard AI search plans in battle with Microsoft. ([n. d.]). https:\/\/www.reuters.com\/technology\/google-opens-bard-chatbot-test-users-plans-more-ai-search-2023-02-06\/"},{"key":"e_1_3_3_2_31_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/N19-1423"},{"key":"e_1_3_3_2_32_2","unstructured":"Victor Dibia. [n.d.]. AlphaCode: Competition-Level Code Generation with Transformer Based Architectures | Paper Review. https:\/\/victordibia.com"},{"key":"e_1_3_3_2_33_2","unstructured":"Thomas Dohmke. [n.d.]. GitHub Copilot is generally available to all developers. https:\/\/github.blog\/2022-06-21-github-copilot-is-generally-available-to-all-developers\/"},{"key":"e_1_3_3_2_34_2","doi-asserted-by":"publisher","unstructured":"Yogesh\u00a0K. Dwivedi et\u00a0al. [n.d.]. Opinion Paper: \u201cSo what if ChatGPT wrote it?\u201d Multidisciplinary perspectives on opportunities challenges and implications of generative conversational AI for research practice and policy. 71 ([n. d.]) 102642. 10.1016\/j.ijinfomgt.2023.102642","DOI":"10.1016\/j.ijinfomgt.2023.102642"},{"key":"e_1_3_3_2_35_2","doi-asserted-by":"publisher","unstructured":"Xiao Fang Shangkun Che Minjia Mao Hongzhe Zhang Ming Zhao and Xiaohang Zhao. 2024. Bias of AI-generated content: an examination of news produced by large language models. Scientific Reports 14 1 (March 2024) 5224. 10.1038\/s41598-024-55686-2","DOI":"10.1038\/s41598-024-55686-2"},{"key":"e_1_3_3_2_36_2","unstructured":"Nat Friedman. [n.d.]. Introducing GitHub Copilot: your AI pair programmer. https:\/\/github.blog\/2021-06-29-introducing-github-copilot-ai-pair-programmer\/"},{"key":"e_1_3_3_2_37_2","unstructured":"Bill Gates. [n.d.]. The Age of AI has begun. https:\/\/www.gatesnotes.com\/The-Age-of-AI-Has-Begun"},{"key":"e_1_3_3_2_38_2","unstructured":"Dave Gershgorn. [n.d.]. GitHub and OpenAI launch a new AI tool that generates its own code. https:\/\/www.theverge.com\/2021\/6\/29\/22555777\/github-openai-ai-tool-autocomplete-code"},{"key":"e_1_3_3_2_39_2","unstructured":"Kristi Hines. [n.d.]. History Of ChatGPT: A Timeline Of The Meteoric Rise Of Generative AI Chatbots. https:\/\/www.searchenginejournal.com\/history-of-chatgpt-timeline\/488370\/"},{"key":"e_1_3_3_2_40_2","doi-asserted-by":"publisher","unstructured":"Xu Huajie. [n.d.]. Github Copilot - A Groundbreaking Code Autocomplete Tool. ([n. d.]). 10.13140\/RG.2.2.29962.24002","DOI":"10.13140\/RG.2.2.29962.24002"},{"key":"e_1_3_3_2_41_2","unstructured":"Senol Isci. [n.d.]. Comprehensive Guide on Evaluation of Response Generation and Retrieval in LLMs. https:\/\/medium.com\/@senol.isci\/comprehensive-guide-on-evaluation-of-response-generation-and-retrieval-with-llms-0cbc2adb3ae6"},{"key":"e_1_3_3_2_42_2","volume-title":"Statistical methods for speech recognition","author":"Jelinek Frederick","unstructured":"Frederick Jelinek. [n.d.]. Statistical methods for speech recognition. MIT Press."},{"key":"e_1_3_3_2_43_2","doi-asserted-by":"publisher","unstructured":"Ahmed\u00a0Naseer Kaftan Majid\u00a0Kadhum Hussain and Farah\u00a0Hasson Naser. [n.d.]. Response accuracy of ChatGPT 3.5 Copilot and Gemini in interpreting biochemical laboratory data a pilot study. 14 1 ([n. d.]) 8233. 10.1038\/s41598-024-58964-1","DOI":"10.1038\/s41598-024-58964-1"},{"key":"e_1_3_3_2_44_2","doi-asserted-by":"publisher","unstructured":"Antonia Karamolegkou Jiaang Li Li Zhou and Anders S\u00f8gaard. [n.d.]. Copyright Violations and Large Language Models. 10.48550\/arXiv.2310.13771 arxiv:https:\/\/arXiv.org\/abs\/2310.13771 [cs]","DOI":"10.48550\/arXiv.2310.13771"},{"key":"e_1_3_3_2_45_2","unstructured":"Tahsin\u00a0Alamgir Kheya Mohamed\u00a0Reda Bouadjenek and Sunil Aryal. [n.d.]. The Pursuit of Fairness in Artificial Intelligence Models: A Survey. arxiv:https:\/\/arXiv.org\/abs\/2403.17333 [cs]http:\/\/arxiv.org\/abs\/2403.17333"},{"key":"e_1_3_3_2_46_2","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2011-720"},{"key":"e_1_3_3_2_47_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-1-4842-9994-4_5"},{"key":"e_1_3_3_2_48_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-53085-2_20"},{"key":"e_1_3_3_2_49_2","doi-asserted-by":"publisher","unstructured":"Yujia Li et\u00a0al. 2022. Competition-level code generation with AlphaCode. Science 378 6624 (Dec. 2022) 1092\u20131097. 10.1126\/science.abq1158","DOI":"10.1126\/science.abq1158"},{"key":"e_1_3_3_2_50_2","doi-asserted-by":"publisher","unstructured":"Yujia Li David Choi Junyoung Chung Nate Kushman Julian Schrittwieser R\u00e9mi Leblond Tom Eccles James Keeling Felix Gimeno Agustin Dal\u00a0Lago Thomas Hubert Peter Choy Cyprien De\u00a0Masson\u00a0d\u2019Autume Igor Babuschkin Xinyun Chen Po-Sen Huang Johannes Welbl Sven Gowal Alexey Cherepanov James Molloy Daniel\u00a0J. Mankowitz Esme Sutherland\u00a0Robson Pushmeet Kohli Nando De\u00a0Freitas Koray Kavukcuoglu and Oriol Vinyals. [n.d.]. Competition-level code generation with AlphaCode. 378 6624 ([n. d.]) 1092\u20131097. 10.1126\/science.abq1158","DOI":"10.1126\/science.abq1158"},{"key":"e_1_3_3_2_51_2","doi-asserted-by":"publisher","unstructured":"Louis Lippens. 2024. Computer says \u2018no\u2019: Exploring systemic bias in ChatGPT using an audit approach. Computers in Human Behavior: Artificial Humans 2 1 (Jan. 2024) 100054. 10.1016\/j.chbah.2024.100054","DOI":"10.1016\/j.chbah.2024.100054"},{"key":"e_1_3_3_2_52_2","doi-asserted-by":"publisher","unstructured":"Yue Liu et\u00a0al. [n.d.]. Refining ChatGPT-Generated Code: Characterizing and Mitigating Code Quality Issues. 33 5 ([n. d.]) 1\u201326. 10.1145\/3643674","DOI":"10.1145\/3643674"},{"key":"e_1_3_3_2_53_2","doi-asserted-by":"publisher","unstructured":"Wei Ma Shangqing Liu Zhihao Lin Wenhan Wang Qiang Hu Ye Liu Cen Zhang Liming Nie Li Li and Yang Liu. [n.d.]. LMs: Understanding Code Syntax and Semantics for Code Analysis. ([n. d.]). 10.48550\/ARXIV.2305.12138","DOI":"10.48550\/ARXIV.2305.12138"},{"key":"e_1_3_3_2_54_2","unstructured":"Dan Milmo. [n.d.]. ChatGPT reaches 100 million users two months after launch. ([n. d.]). https:\/\/www.theguardian.com\/technology\/2023\/feb\/02\/chatgpt-100-million-users-open-ai-fastest-growing-app"},{"key":"e_1_3_3_2_55_2","unstructured":"Ethan Mollick. [n.d.]. ChatGPT Is a Tipping Point for AI. ([n. d.]). https:\/\/hbr.org\/2022\/12\/chatgpt-is-a-tipping-point-for-ai"},{"key":"e_1_3_3_2_56_2","doi-asserted-by":"publisher","DOI":"10.1145\/3524842.3528470"},{"key":"e_1_3_3_2_57_2","unstructured":"OpenAI. [n.d.]. Code generated by ChatGPT. https:\/\/chat.openai.com Generated by ChatGPT."},{"key":"e_1_3_3_2_58_2","doi-asserted-by":"publisher","unstructured":"OpenAI Josh Achiam et\u00a0al. 2023. GPT-4 Technical Report. (2023). 10.48550\/ARXIV.2303.08774","DOI":"10.48550\/ARXIV.2303.08774"},{"key":"e_1_3_3_2_59_2","doi-asserted-by":"publisher","unstructured":"Debalina\u00a0Ghosh Paul Hong Zhu and Ian Bayley. 2024. Benchmarks and Metrics for Evaluations of Code Generation: A Critical Review. (2024). 10.48550\/ARXIV.2406.12655","DOI":"10.48550\/ARXIV.2406.12655"},{"key":"e_1_3_3_2_60_2","unstructured":"Leone Perdig\u00e3o. [n.d.]. ChatGPT: a deep dive. https:\/\/leoneperdigao.medium.com\/chatgpt-a-deep-dive-1feade9c4d77"},{"key":"e_1_3_3_2_61_2","unstructured":"Cameron R.\u00a0Wolfe Ph.D. [n.d.]. Decoder-Only Transformers: The Workhorse of Generative LLMs. https:\/\/cameronrwolfe.substack.com\/p\/decoder-only-transformers-the-workhorse"},{"key":"e_1_3_3_2_62_2","unstructured":"Narasimhan K. Salimans T. & Sutskever\u00a0I. Radford A.[n.d.]. Improving language understanding by generative pre-training. ([n. d.]). https:\/\/www.mikecaptain.com\/resources\/pdf\/GPT-1.pdf"},{"key":"e_1_3_3_2_63_2","unstructured":"Tal Ridnik. [n.d.]. State-of-the-art Code Generation with AlphaCodium - From Prompt Engineering to Flow Engineering. https:\/\/www.codium.ai\/blog\/alphacodium-state-of-the-art-code-generation-for-code-contests\/"},{"key":"e_1_3_3_2_64_2","doi-asserted-by":"publisher","unstructured":"Tal Ridnik Dedy Kredo and Itamar Friedman. [n.d.]. Code Generation with AlphaCodium: From Prompt Engineering to Flow Engineering. 10.48550\/arXiv.2401.08500 arxiv:https:\/\/arXiv.org\/abs\/2401.08500 [cs]","DOI":"10.48550\/arXiv.2401.08500"},{"key":"e_1_3_3_2_65_2","doi-asserted-by":"publisher","unstructured":"Konstantinos\u00a0I. Roumeliotis and Nikolaos\u00a0D. Tselikas. [n.d.]. ChatGPT and Open-AI Models: A Preliminary Review. 15 6 ([n. d.]) 192. 10.3390\/fi15060192","DOI":"10.3390\/fi15060192"},{"key":"e_1_3_3_2_66_2","doi-asserted-by":"publisher","unstructured":"Ilja Siro\u0161 Dave Singel\u00e9e and Bart Preneel. 2024. GitHub Copilot: the perfect Code compLeeter? (2024). 10.48550\/ARXIV.2406.11326","DOI":"10.48550\/ARXIV.2406.11326"},{"key":"e_1_3_3_2_67_2","doi-asserted-by":"publisher","unstructured":"Gemini Team Rohan Anil et\u00a0al. 2023. Gemini: A Family of Highly Capable Multimodal Models. (2023). 10.48550\/ARXIV.2312.11805","DOI":"10.48550\/ARXIV.2312.11805"},{"key":"e_1_3_3_2_68_2","unstructured":"James Vincent. [n.d.]. DeepMind says its new AI coding engine is as good as an average human programmer. https:\/\/www.theverge.com\/2022\/2\/2\/22914085\/alphacode-ai-coding-program-automatic-deepmind-codeforce"},{"key":"e_1_3_3_2_69_2","unstructured":"Yuqing Wang and Yun Zhao. [n.d.]. Gemini in Reasoning: Unveiling Commonsense in Multimodal Large Language Models. arxiv:https:\/\/arXiv.org\/abs\/2312.17661 [cs]http:\/\/arxiv.org\/abs\/2312.17661"},{"key":"e_1_3_3_2_70_2","unstructured":"Kyle Wiggers. 2023. Google unveils AlphaCode 2 powered by Gemini. https:\/\/techcrunch.com\/2023\/12\/06\/deepmind-unveils-alphacode-2-powered-by-gemini\/"},{"key":"e_1_3_3_2_71_2","unstructured":"Tong Xiao and Jingbo Zhu. [n.d.]. Introduction to Transformers: an NLP Perspective. arxiv:https:\/\/arXiv.org\/abs\/2311.17633 [cs]http:\/\/arxiv.org\/abs\/2311.17633"},{"key":"e_1_3_3_2_72_2","doi-asserted-by":"publisher","unstructured":"Lixiang Yan Lele Sha Linxuan Zhao Yuheng Li Roberto Martinez\u2010Maldonado Guanliang Chen Xinyu Li Yueqiao Jin and Dragan Ga\u0161evi\u0107. [n.d.]. Practical and ethical challenges of large language models in education: A systematic scoping review. 55 1 ([n. d.]) 90\u2013112. 10.1111\/bjet.13370","DOI":"10.1111\/bjet.13370"},{"key":"e_1_3_3_2_73_2","doi-asserted-by":"publisher","unstructured":"Abdelrahman Zayed Gon\u00e7alo Mordido Samira Shabanian Ioana Baldini and Sarath Chandar. 2024. Fairness-Aware Structured Pruning in Transformers. Proceedings of the AAAI Conference on Artificial Intelligence 38 20 (March 2024) 22484\u201322492. 10.1609\/aaai.v38i20.30256","DOI":"10.1609\/aaai.v38i20.30256"},{"key":"e_1_3_3_2_74_2","doi-asserted-by":"publisher","unstructured":"Shudan Zhang et\u00a0al. 2024. NaturalCodeBench: Examining Coding Performance Mismatch on HumanEval and Natural User Prompts. (2024). 10.48550\/ARXIV.2405.04520","DOI":"10.48550\/ARXIV.2405.04520"},{"key":"e_1_3_3_2_75_2","unstructured":"Shuyin Zhao. [n.d.]. GitHub Copilot now has a better AI model and new capabilities. https:\/\/github.blog\/2023-02-14-github-copilot-now-has-a-better-ai-model-and-new-capabilities\/"}],"event":{"name":"ICCA 2024: 3rd International Conference on Computing Advancements","location":"Dhaka Bangladesh","acronym":"ICCA 2024"},"container-title":["Proceedings of the 3rd International Conference on Computing Advancements"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3723178.3723224","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3723178.3723224","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T01:56:47Z","timestamp":1750298207000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3723178.3723224"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,17]]},"references-count":74,"alternative-id":["10.1145\/3723178.3723224","10.1145\/3723178"],"URL":"https:\/\/doi.org\/10.1145\/3723178.3723224","relation":{},"subject":[],"published":{"date-parts":[[2024,10,17]]},"assertion":[{"value":"2025-06-06","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}