{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,21]],"date-time":"2026-04-21T14:50:51Z","timestamp":1776783051574,"version":"3.51.2"},"publisher-location":"New York, NY, USA","reference-count":49,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,6,3]],"date-time":"2024-06-03T00:00:00Z","timestamp":1717372800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"DOI":"10.13039\/100000001","name":"National Science Foundation","doi-asserted-by":"publisher","award":["2047120"],"award-info":[{"award-number":["2047120"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,6,3]]},"DOI":"10.1145\/3625549.3658689","type":"proceedings-article","created":{"date-parts":[[2024,8,30]],"date-time":"2024-08-30T15:55:29Z","timestamp":1725033329000},"page":"281-294","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":25,"title":["Can Large Language Models Write Parallel Code?"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-3538-6164","authenticated-orcid":false,"given":"Daniel","family":"Nichols","sequence":"first","affiliation":[{"name":"University of Maryland, College Park, United States of America"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6704-0520","authenticated-orcid":false,"given":"Joshua H.","family":"Davis","sequence":"additional","affiliation":[{"name":"University of Maryland, College Park, United States of America"}]},{"ORCID":"https:\/\/orcid.org\/0009-0004-2952-4198","authenticated-orcid":false,"given":"Zhaojun","family":"Xie","sequence":"additional","affiliation":[{"name":"University of Maryland, College Park, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7016-1617","authenticated-orcid":false,"given":"Arjun","family":"Rajaram","sequence":"additional","affiliation":[{"name":"University of Maryland, College Park, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3069-3701","authenticated-orcid":false,"given":"Abhinav","family":"Bhatele","sequence":"additional","affiliation":[{"name":"University of Maryland, College Park, United States of America"}]}],"member":"320","published-online":{"date-parts":[[2024,8,30]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"2023. Big Code Models Leaderboard - a Hugging Face Space by bigcode. https:\/\/huggingface.co\/spaces\/bigcode\/bigcode-models-leaderboard"},{"key":"e_1_3_2_1_2_1","unstructured":"2023. HIP Documentation. https:\/\/rocm.docs.amd.com\/projects\/HIP\/en\/latest\/"},{"key":"e_1_3_2_1_3_1","unstructured":"2023. Zero-Shot Replication Framework. https:\/\/github.com\/emrgnt-cmplxty\/zero-shot-replication."},{"key":"e_1_3_2_1_4_1","volume-title":"A Transformer-based Approach for Source Code Summarization. ArXiv abs\/2005.00653","author":"Ahmad Wasi Uddin","year":"2020","unstructured":"Wasi Uddin Ahmad, Saikat Chakraborty, Baishakhi Ray, and Kai-Wei Chang. 2020. A Transformer-based Approach for Source Code Summarization. ArXiv abs\/2005.00653 (2020)."},{"key":"e_1_3_2_1_5_1","volume-title":"Learning code summarization from a small and local dataset. ArXiv abs\/2206.00804","author":"Ahmed Toufique","year":"2022","unstructured":"Toufique Ahmed and Prem Devanbu. 2022. Learning code summarization from a small and local dataset. ArXiv abs\/2206.00804 (2022)."},{"key":"e_1_3_2_1_6_1","volume-title":"Niklas Muennighoff, Mayank Mishra, Alex Gu, Manan Dey, et al.","author":"Allal Loubna Ben","year":"2023","unstructured":"Loubna Ben Allal, Raymond Li, Denis Kocetkov, Chenghao Mou, Christopher Akiki, Carlos Munoz Ferrandis, Niklas Muennighoff, Mayank Mishra, Alex Gu, Manan Dey, et al. 2023. SantaCoder: don't reach for the stars! arXiv preprint arXiv:2301.03988 (2023)."},{"key":"e_1_3_2_1_7_1","volume-title":"Program Synthesis with Large Language Models. CoRR abs\/2108.07732","author":"Austin Jacob","year":"2021","unstructured":"Jacob Austin, Augustus Odena, Maxwell I. Nye, Maarten Bosma, Henryk Michalewski, David Dohan, Ellen Jiang, Carrie J. Cai, Michael Terry, Quoc V. Le, and Charles Sutton. 2021. Program Synthesis with Large Language Models. CoRR abs\/2108.07732 (2021). arXiv:2108.07732 https:\/\/arxiv.org\/abs\/2108.07732"},{"key":"e_1_3_2_1_8_1","volume-title":"Brown et al","author":"Tom","year":"2020","unstructured":"Tom B. Brown et al. 2020. Language Models are Few-Shot Learners. CoRR abs\/2005.14165 (2020). arXiv:2005.14165 https:\/\/arxiv.org\/abs\/2005.14165"},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1109\/TSE.2023.3267446"},{"key":"e_1_3_2_1_10_1","volume-title":"Pei hung Lin, and Chuanhua Liao","author":"Chen Le","year":"2023","unstructured":"Le Chen, Xianzhong Ding, Murali Emani, Tristan Vanderbruggen, Pei hung Lin, and Chuanhua Liao. 2023. Data Race Detection Using Large Language Models. arXiv:2308.07505 [cs.LG]"},{"key":"e_1_3_2_1_11_1","volume-title":"LM4HPC: Towards Effective Language Model Application in High-Performance Computing","author":"Chen Le","unstructured":"Le Chen, Pei-Hung Lin, Tristan Vanderbruggen, Chunhua Liao, Murali Emani, and Bronis de Supinski. 2023. LM4HPC: Towards Effective Language Model Application in High-Performance Computing. In OpenMP: Advanced Task-Based, Device and Compiler Programming, Simon McIntosh-Smith, Michael Klemm, Bronis R. de Supinski, Tom Deakin, and Jannis Klinkenberg (Eds.). Springer Nature Switzerland, Cham, 18--33."},{"key":"e_1_3_2_1_12_1","unstructured":"Mark Chen and et al. 2021. Evaluating Large Language Models Trained on Code. arXiv:arXiv:2107.03374"},{"key":"e_1_3_2_1_13_1","unstructured":"Mark Chen Jerry Tworek Heewoo Jun Qiming Yuan Henrique Ponde de Oliveira Pinto Jared Kaplan Harri Edwards Yuri Burda Nicholas Joseph Greg Brockman Alex Ray Raul Puri Gretchen Krueger Michael Petrov Heidy Khlaaf Girish Sastry Pamela Mishkin Brooke Chan Scott Gray Nick Ryder Mikhail Pavlov Alethea Power Lukasz Kaiser Mohammad Bavarian Clemens Winter Philippe Tillet Felipe Petroski Such Dave Cummings Matthias Plappert Fotios Chantzis Elizabeth Barnes Ariel Herbert-Voss William Hebgen Guss Alex Nichol Alex Paino Nikolas Tezak Jie Tang Igor Babuschkin Suchir Balaji Shantanu Jain William Saunders Christopher Hesse Andrew N. Carr Jan Leike Josh Achiam Vedant Misra Evan Morikawa Alec Radford Matthew Knight Miles Brundage Mira Murati Katie Mayer Peter Welinder Bob McGrew Dario Amodei Sam McCandlish Ilya Sutskever and Wojciech Zaremba. 2021. Evaluating Large Language Models Trained on Code. arXiv:arXiv:2107.03374"},{"key":"e_1_3_2_1_14_1","volume-title":"Training Verifiers to Solve Math Word Problems. arXiv preprint arXiv:2110.14168","author":"Cobbe Karl","year":"2021","unstructured":"Karl Cobbe, Vineet Kosaraju, Mohammad Bavarian, Mark Chen, Heewoo Jun, Lukasz Kaiser, Matthias Plappert, Jerry Tworek, Jacob Hilton, Reiichiro Nakano, Christopher Hesse, and John Schulman. 2021. Training Verifiers to Solve Math Word Problems. arXiv preprint arXiv:2110.14168 (2021)."},{"key":"e_1_3_2_1_15_1","unstructured":"Xueying Du Mingwei Liu Kaixin Wang Hanlin Wang Junwei Liu Yixuan Chen Jiayi Feng Chaofeng Sha Xin Peng and Yiling Lou. 2023. ClassEval: A Manually-Crafted Benchmark for Evaluating LLMs on Class-level Code Generation. arXiv:2308.01861 [cs.CL]"},{"key":"e_1_3_2_1_16_1","volume-title":"The Pile: An 800GB Dataset of Diverse Text for Language Modeling. CoRR abs\/2101.00027","author":"Gao Leo","year":"2021","unstructured":"Leo Gao, Stella Biderman, Sid Black, Laurence Golding, Travis Hoppe, Charles Foster, Jason Phang, Horace He, Anish Thite, Noa Nabeshima, Shawn Presser, and Connor Leahy. 2021. The Pile: An 800GB Dataset of Diverse Text for Language Modeling. CoRR abs\/2101.00027 (2021). arXiv:2101.00027 https:\/\/arxiv.org\/abs\/2101.00027"},{"key":"e_1_3_2_1_17_1","volume-title":"PAL: Program-aided Language Models. arXiv preprint arXiv:2211.10435","author":"Gao Luyu","year":"2022","unstructured":"Luyu Gao, Aman Madaan, Shuyan Zhou, Uri Alon, Pengfei Liu, Yiming Yang, Jamie Callan, and Graham Neubig. 2022. PAL: Program-aided Language Models. arXiv preprint arXiv:2211.10435 (2022)."},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1145\/3540250.3549096"},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1145\/3605731.3605886"},{"key":"e_1_3_2_1_20_1","volume-title":"Assemble Foundation Models for Automatic Code Summarization. 2022 IEEE International Conference on Software Analysis, Evolution and Reengineering (SANER)","author":"Gu Jian","year":"2022","unstructured":"Jian Gu, Pasquale Salza, and Harald C. Gall. 2022. Assemble Foundation Models for Automatic Code Summarization. 2022 IEEE International Conference on Software Analysis, Evolution and Reengineering (SANER) (2022), 935--946."},{"key":"e_1_3_2_1_21_1","volume-title":"Semantic Similarity Metrics for Evaluating Source Code Summarization. 2022 IEEE\/ACM 30th International Conference on Program Comprehension (ICPC)","author":"Haque Sakib","year":"2022","unstructured":"Sakib Haque, Zachary Eberhart, Aakash Bansal, and Collin McMillan. 2022. Semantic Similarity Metrics for Evaluating Source Code Summarization. 2022 IEEE\/ACM 30th International Conference on Program Comprehension (ICPC) (2022), 36--47."},{"key":"e_1_3_2_1_22_1","volume-title":"The Curious Case of Neural Text Degeneration. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=rygGQyrFvH","author":"Holtzman Ari","year":"2020","unstructured":"Ari Holtzman, Jan Buys, Li Du, Maxwell Forbes, and Yejin Choi. 2020. The Curious Case of Neural Text Degeneration. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=rygGQyrFvH"},{"key":"e_1_3_2_1_23_1","volume-title":"LoRA: Low-Rank Adaptation of Large Language Models. CoRR abs\/2106.09685","author":"Hu Edward J.","year":"2021","unstructured":"Edward J. Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, and Weizhu Chen. 2021. LoRA: Low-Rank Adaptation of Large Language Models. CoRR abs\/2106.09685 (2021). arXiv:2106.09685 https:\/\/arxiv.org\/abs\/2106.09685"},{"key":"e_1_3_2_1_24_1","unstructured":"Tal Kadosh Niranjan Hasabnis Vy A. Vo Nadav Schneider Neva Krien Abdul Wasay Nesreen Ahmed Ted Willke Guy Tamir Yuval Pinter Timothy Mattson and Gal Oren. 2023. Scope is all you need: Transforming LLMs for HPC Code. arXiv:2308.09440 [cs.CL]"},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1109\/MSR59073.2023.00037"},{"key":"e_1_3_2_1_26_1","volume-title":"2022 IEEE\/ACM 44th International Conference on Software Engineering (ICSE)","author":"Kharkar Anant","year":"2022","unstructured":"Anant Kharkar, Roshanak Zilouchian Moghaddam, Matthew Jin, Xiaoyu Liu, Xin Shi, Colin B. Clement, and Neel Sundaresan. 2022. Learning to Reduce False Positives in Analytic Bug Detectors. 2022 IEEE\/ACM 44th International Conference on Software Engineering (ICSE) (2022), 1307--1316."},{"key":"e_1_3_2_1_27_1","volume-title":"Jia Li, Chenghao Mou, Carlos Mu\u00f1oz Ferrandis, Yacine Jernite, Margaret Mitchell, Sean Hughes, Thomas Wolf, Dzmitry Bahdanau, Leandro von Werra, and Harm de Vries.","author":"Kocetkov Denis","year":"2022","unstructured":"Denis Kocetkov, Raymond Li, Loubna Ben Allal, Jia Li, Chenghao Mou, Carlos Mu\u00f1oz Ferrandis, Yacine Jernite, Margaret Mitchell, Sean Hughes, Thomas Wolf, Dzmitry Bahdanau, Leandro von Werra, and Harm de Vries. 2022. The Stack: 3 TB of permissively licensed source code. Preprint (2022)."},{"key":"e_1_3_2_1_28_1","volume-title":"Daniel Fried, Sida Wang, and Tao Yu.","author":"Lai Yuhang","year":"2022","unstructured":"Yuhang Lai, Chengxi Li, Yiming Wang, Tianyi Zhang, Ruiqi Zhong, Luke Zettlemoyer, Scott Wen tau Yih, Daniel Fried, Sida Wang, and Tao Yu. 2022. DS-1000: A Natural and Reliable Benchmark for Data Science Code Generation. arXiv:2211.11501 [cs.SE]"},{"key":"e_1_3_2_1_29_1","unstructured":"Raymond Li Loubna Ben Allal Yangtian Zi Niklas Muennighoff Denis Kocetkov Chenghao Mou Marc Marone Christopher Akiki Jia Li Jenny Chim Qian Liu Evgenii Zheltonozhskii Terry Yue Zhuo Thomas Wang Olivier Dehaene Mishig Davaadorj Joel Lamy-Poirier Jo\u00e3o Monteiro Oleh Shliazhko Nicolas Gontier Nicholas Meade Armel Zebaze Ming-Ho Yee Logesh Kumar Umapathi Jian Zhu Benjamin Lipkin Muhtasham Oblokulov Zhiruo Wang Rudra Murthy Jason Stillerman Siva Sankalp Patel Dmitry Abulkhanov Marco Zocca Manan Dey Zhihan Zhang Nour Fahmy Urvashi Bhattacharyya Wenhao Yu Swayam Singh Sasha Luccioni Paulo Villegas Maxim Kunakov Fedor Zhdanov Manuel Romero Tony Lee Nadav Timor Jennifer Ding Claire Schlesinger Hailey Schoelkopf Jan Ebert Tri Dao Mayank Mishra Alex Gu Jennifer Robinson Carolyn Jane Anderson Brendan Dolan-Gavitt Danish Contractor Siva Reddy Daniel Fried Dzmitry Bahdanau Yacine Jernite Carlos Mu\u00f1oz Ferrandis Sean Hughes Thomas Wolf Arjun Guha Leandro von Werra and Harm de Vries. 2023. StarCoder: may the source be with you! (2023). arXiv:2305.06161 [cs.CL]"},{"key":"e_1_3_2_1_30_1","unstructured":"Mingjie Liu Nathaniel Pinckney Brucek Khailany and Haoxing Ren. 2023. VerilogEval: Evaluating Large Language Models for Verilog Code Generation. arXiv:2309.07544 [cs.LG]"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"crossref","unstructured":"Christian Munley Aaron Jarmusch and Sunita Chandrasekaran. 2023. LLM4VV: Developing LLM-Driven Testsuite for Compiler Validation. arXiv:2310.04963 [cs.AI]","DOI":"10.1016\/j.future.2024.05.034"},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"crossref","unstructured":"Daniel Nichols Aniruddha Marathe Harshitha Menon Todd Gamblin and Abhinav Bhatele. 2023. Modeling Parallel Programs using Large Language Models. arXiv:2306.17281 [cs.DC]","DOI":"10.23919\/ISC.2024.10528929"},{"key":"e_1_3_2_1_33_1","volume-title":"Fitzek","author":"P\u00e9ter Vingelmann NVIDIA","year":"2020","unstructured":"NVIDIA, P\u00e9ter Vingelmann, and Frank H.P. Fitzek. 2020. CUDA, release: 10.2.89. https:\/\/developer.nvidia.com\/cuda-toolkit"},{"key":"e_1_3_2_1_35_1","unstructured":"OpenAI. 2023. OpenAI API. https:\/\/platform.openai.com\/docs\/api-reference\/"},{"key":"e_1_3_2_1_36_1","unstructured":"OpenAI. 2023. OpenAI Python API library. https:\/\/github.com\/openai\/openaipython"},{"key":"e_1_3_2_1_37_1","volume-title":"OpenMP Application Program Interface. Version 4.0","year":"2013","unstructured":"OpenMP4 2013. OpenMP Application Program Interface. Version 4.0. July 2013."},{"key":"e_1_3_2_1_38_1","volume-title":"PyTorch: An Imperative Style","author":"Paszke Adam","year":"1912","unstructured":"Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, Alban Desmaison, Andreas K\u00f6pf, Edward Yang, Zach DeVito, Martin Raison, Alykhan Tejani, Sasank Chilamkurthy, Benoit Steiner, Lu Fang, Junjie Bai, and Soumith Chintala. 2019. PyTorch: An Imperative Style, High-Performance Deep Learning Library. arXiv:1912.01703 [cs.LG]"},{"key":"e_1_3_2_1_39_1","unstructured":"Phind. 2023. Phind-CodeLlama-34B-v2. https:\/\/huggingface.co\/Phind\/Phind-CodeLlama-34B-v2"},{"key":"e_1_3_2_1_40_1","volume-title":"Can we learn from developer mistakes? Learning to localize and repair real bugs from real bug fixes. ArXiv abs\/2207.00301","author":"Richter Cedric","year":"2022","unstructured":"Cedric Richter and Heike Wehrheim. 2022. Can we learn from developer mistakes? Learning to localize and repair real bugs from real bug fixes. ArXiv abs\/2207.00301 (2022)."},{"key":"e_1_3_2_1_41_1","unstructured":"Baptiste Rozi\u00e8re Jonas Gehring Fabian Gloeckle Sten Sootla Itai Gat Xiaoqing Ellen Tan Yossi Adi Jingyu Liu Tal Remez J\u00e9r\u00e9my Rapin Artyom Kozhevnikov Ivan Evtimov Joanna Bitton Manish Bhatt Cristian Canton Ferrer Aaron Grattafiori Wenhan Xiong Alexandre D\u00e9fossez Jade Copet Faisal Azhar Hugo Touvron Louis Martin Nicolas Usunier Thomas Scialom and Gabriel Synnaeve. 2023. Code Llama: Open Foundation Models for Code. arXiv:2308.12950 [cs.CL]"},{"key":"e_1_3_2_1_42_1","unstructured":"M. Snir. 1998. MPI-the Complete Reference: The MPI core. Mass. https:\/\/books.google.com\/books?id=x79puJ2YkroC"},{"key":"e_1_3_2_1_43_1","doi-asserted-by":"crossref","unstructured":"Xiangru Tang Bill Qian Rick Gao Jiakang Chen Xinyun Chen and Mark Gerstein. 2023. BioCoder: A Benchmark for Bioinformatics Code Generation with Contextual Pragmatic Knowledge. arXiv:2308.16458 [cs.LG]","DOI":"10.1093\/bioinformatics\/btae230"},{"key":"e_1_3_2_1_44_1","unstructured":"Hugo Touvron et al. 2023. Llama 2: Open Foundation and Fine-Tuned Chat Models. arXiv:2307.09288 [cs.CL]"},{"key":"e_1_3_2_1_45_1","doi-asserted-by":"publisher","DOI":"10.1109\/TPDS.2021.3097283"},{"key":"e_1_3_2_1_46_1","volume-title":"William F. Godoy, Keita Teranishi, Prasanna Balaprakash, and Jeffrey S. Vetter.","author":"Valero-Lara Pedro","year":"2023","unstructured":"Pedro Valero-Lara, Alexis Huante, Mustafa Al Lail, William F. Godoy, Keita Teranishi, Prasanna Balaprakash, and Jeffrey S. Vetter. 2023. Comparing Llama-2 and GPT-3 LLMs for HPC kernels generation. arXiv:2309.07103 [cs.SE]"},{"key":"e_1_3_2_1_47_1","volume-title":"CoRR abs\/1706.03762","author":"Vaswani Ashish","year":"2017","unstructured":"Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, and Illia Polosukhin. 2017. Attention Is All You Need. CoRR abs\/1706.03762 (2017). arXiv:1706.03762 http:\/\/arxiv.org\/abs\/1706.03762"},{"key":"e_1_3_2_1_48_1","volume-title":"Sylvain Gugger, Mariama Drame, Quentin Lhoest, and Alexander M. Rush.","author":"Wolf Thomas","year":"2020","unstructured":"Thomas Wolf, Lysandre Debut, Victor Sanh, Julien Chaumond, Clement Delangue, Anthony Moi, Perric Cistac, Clara Ma, Yacine Jernite, Julien Plu, Canwen Xu, Teven Le Scao, Sylvain Gugger, Mariama Drame, Quentin Lhoest, and Alexander M. Rush. 2020. Transformers: State-of-the-Art Natural Language Processing. Association for Computational Linguistics, 38--45. https:\/\/www.aclweb.org\/anthology\/2020.emnlp-demos.6"},{"key":"e_1_3_2_1_49_1","doi-asserted-by":"publisher","DOI":"10.5281\/zenodo.6363556"},{"key":"e_1_3_2_1_50_1","volume-title":"CoderEval: A Benchmark of Pragmatic Code Generation with Generative Pre-trained Models. arXiv preprint arXiv:2302.00288","author":"Yu Hao","year":"2023","unstructured":"Hao Yu, Bo Shen, Dezhi Ran, Jiaxin Zhang, Qi Zhang, Yuchi Ma, Guangtai Liang, Ying Li, Tao Xie, and Qianxiang Wang. 2023. CoderEval: A Benchmark of Pragmatic Code Generation with Generative Pre-trained Models. arXiv preprint arXiv:2302.00288 (2023)."}],"event":{"name":"HPDC '24: 33rd International Symposium on High-Performance Parallel and Distributed Computing","location":"Pisa Italy","acronym":"HPDC '24","sponsor":["SIGARCH ACM Special Interest Group on Computer Architecture","SIGHPC ACM Special Interest Group on High Performance Computing, Special Interest Group on High Performance Computing"]},"container-title":["Proceedings of the 33rd International Symposium on High-Performance Parallel and Distributed Computing"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3625549.3658689","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3625549.3658689","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T22:50:38Z","timestamp":1750287038000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3625549.3658689"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,6,3]]},"references-count":49,"alternative-id":["10.1145\/3625549.3658689","10.1145\/3625549"],"URL":"https:\/\/doi.org\/10.1145\/3625549.3658689","relation":{},"subject":[],"published":{"date-parts":[[2024,6,3]]},"assertion":[{"value":"2024-08-30","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}