{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,20]],"date-time":"2026-02-20T19:05:05Z","timestamp":1771614305356,"version":"3.50.1"},"reference-count":85,"publisher":"Association for Computing Machinery (ACM)","issue":"1","funder":[{"name":"Ministry of Education, Singapore under its Academic Research Fund Tier 3","award":["MOET32020-0004"],"award-info":[{"award-number":["MOET32020-0004"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["ACM Trans. Softw. Eng. Methodol."],"published-print":{"date-parts":[[2026,1,31]]},"abstract":"<jats:p>\n                    The extensive availability of publicly accessible source code and the advances in language models, coupled with increasing computational resources, have led to a remarkable rise of large language models for code (LLM4Code). These models do not exist in isolation but rather depend on and interact with each other, forming a complex ecosystem that is worth studying. It motivates us to introduce a pioneering analysis of the\n                    <jats:italic toggle=\"yes\">LLM4Code ecosystem<\/jats:italic>\n                    . Utilizing Hugging Face [\n                    <jats:inline-formula content-type=\"math\/tex\">\n                      <jats:inline-graphic xmlns:xlink=\"http:\/\/www.w3.org\/1999\/xlink\" xlink:href=\"tosem-2024-0656-inline1.jpg\"\/>\n                    <\/jats:inline-formula>\n                    ]\u2014the premier hub for transformer-based models\u2014as our primary source, we manually curate a list of datasets and models focused on software engineering tasks. We first identify key datasets, models, and users in the ecosystem and quantify their contributions and importance. We then examine each model\u2019s documentation to trace its base model and understand the process for deriving new models. We categorize LLM4Code model reuse into nine categories, with the top three being\n                    <jats:italic toggle=\"yes\">fine-tuning<\/jats:italic>\n                    ,\n                    <jats:italic toggle=\"yes\">architecture sharing<\/jats:italic>\n                    , and\n                    <jats:italic toggle=\"yes\">quantization<\/jats:italic>\n                    . Additionally, we examine documentation and licensing practices, revealing that LLM4Code documentation is less detailed than that of general AI repositories on GitHub. The license usage pattern is also different from other software repositories, and we further analyze potential license incompatibility issues. To analyze the rapidly growing LLM4Code, we explore the potential of using LLMs to assist in constructing and analyzing the ecosystem. Advanced LLMs from OpenAI identify LLM4Code with 98% accuracy, infer base models with 87% accuracy, and predict reuse types with 89% accuracy. We employ LLMs to expand the ecosystem and find that conclusions from the manually curated dataset align with those from the automatically created one. Based on our findings, we discuss the implications and suggestions to facilitate the healthy growth of LLM4Code.\n                  <\/jats:p>","DOI":"10.1145\/3731753","type":"journal-article","created":{"date-parts":[[2025,4,25]],"date-time":"2025-04-25T13:23:23Z","timestamp":1745587403000},"page":"1-30","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":9,"title":["Ecosystem of Large Language Models for Code"],"prefix":"10.1145","volume":"35","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-5938-1918","authenticated-orcid":false,"given":"Zhou","family":"Yang","sequence":"first","affiliation":[{"name":"Singapore Management University, Singapore, Singapore"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0799-5018","authenticated-orcid":false,"given":"Jieke","family":"Shi","sequence":"additional","affiliation":[{"name":"Singapore Management University, Singapore, Singapore"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4346-5276","authenticated-orcid":false,"given":"Premkumar","family":"Devanbu","sequence":"additional","affiliation":[{"name":"UC Davis, Davis, California, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4367-7201","authenticated-orcid":false,"given":"David","family":"Lo","sequence":"additional","affiliation":[{"name":"Singapore Management University, Singapore, Singapore"}]}],"member":"320","published-online":{"date-parts":[[2025,12,11]]},"reference":[{"key":"e_1_3_2_2_2","unstructured":"The Verge. 2025. OpenAI Has Evidence That Its Models Helped Train China\u2019s DeepSeek. Retrieved February 13 2025 from https:\/\/www.theverge.com\/news\/601195\/openai-evidence-deepseek-distillation-ai-data"},{"key":"e_1_3_2_3_2","unstructured":"The Linux Foundation. 2019. Fulfilling Open Source License Obligations: Can Checklists Help? Retrieved from https:\/\/events19.linuxfoundation.org\/wp-content\/uploads\/2018\/07\/OSLS-2019-Fulfilling-Open-Source-license-obligations-Can-checklists-help.pdf"},{"key":"e_1_3_2_4_2","doi-asserted-by":"publisher","DOI":"10.1145\/2597073.2597120"},{"key":"e_1_3_2_5_2","unstructured":"Toufique Ahmed Premkumar Devanbu Christoph Treude and Michael Pradel. 2024. Can LLMs replace manual annotation of software engineering artifacts? arXiv:2408.05534. Retrieved from https:\/\/arxiv.org\/abs\/2408.05534"},{"key":"e_1_3_2_6_2","doi-asserted-by":"publisher","DOI":"10.1109\/SANER50967.2021.00048"},{"key":"e_1_3_2_7_2","unstructured":"Loubna Ben Allal Raymond Li Denis Kocetkov Chenghao Mou Christopher Akiki Carlos Munoz Ferrandis Niklas Muennighoff Mayank Mishra Alex Gu Manan Dey et al. 2023. SantaCoder: Don\u2019t reach for the stars! arXiv:2301.03988. Retrieved from https:\/\/arxiv.org\/abs\/2301.03988"},{"key":"e_1_3_2_8_2","doi-asserted-by":"publisher","DOI":"10.1145\/3359591.3359735"},{"key":"e_1_3_2_9_2","doi-asserted-by":"publisher","DOI":"10.1126\/science.286.5439.509"},{"key":"e_1_3_2_10_2","doi-asserted-by":"publisher","DOI":"10.1145\/2568225.2568233"},{"key":"e_1_3_2_11_2","unstructured":"Amanda Bertsch Maor Ivgi Uri Alon Jonathan Berant Matthew R. Gormley and Graham Neubig. 2024. In-context learning with long-context models: An in-depth exploration. arXiv:2405.00200. Retrieved from https:\/\/arxiv.org\/abs\/2405.00200"},{"key":"e_1_3_2_12_2","unstructured":"Rishi Bommasani Drew A. Hudson Ehsan Adeli Russ Altman Simran Arora Sydney von Arx Michael S. Bernstein Jeannette Bohg Antoine Bosselut Emma Brunskill et al. 2021. On the opportunities and risks of foundation models. arXiv:2108.07258. Retrieved from https:\/\/arxiv.org\/abs\/2108.07258"},{"key":"e_1_3_2_13_2","doi-asserted-by":"crossref","unstructured":"Rishi Bommasani Dilara Soylu Thomas I. Liao Kathleen A. Creel and Percy Liang. 2025. Ecosystem graphs: The social footprint of foundation models. In Proceedings of the 2024 AAAI\/ACM Conference on AI Ethics and Society. AAAI Press 196\u2013209.","DOI":"10.1609\/aies.v7i1.31629"},{"key":"e_1_3_2_14_2","doi-asserted-by":"publisher","DOI":"10.1145\/1718918.1718973"},{"key":"e_1_3_2_15_2","first-page":"1877","volume-title":"Advances in Neural Information Processing Systems","author":"Brown Tom","year":"2020","unstructured":"Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared D. Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, et al. 2020. Language models are few-shot learners. In Advances in Neural Information Processing Systems. H. Larochelle, M. Ranzato, R. Hadsell, M.F. Balcan, and H. Lin (Eds.), Vol. 33, Curran Associates, Inc., 1877\u20131901."},{"key":"e_1_3_2_16_2","doi-asserted-by":"publisher","DOI":"10.1145\/3643664.3648204"},{"key":"e_1_3_2_17_2","first-page":"607","volume-title":"Proceedings of the 2024 IEEE\/ACM 21st International Conference on Mining Software Repositories (MSR)","author":"Castano J.","year":"2024","unstructured":"J. Castano, S. Martinez-Fernandez, X. Franch, and J. Bogner. 2024. Analyzing the evolution and maintenance of ML models on hugging face. In Proceedings of the 2024 IEEE\/ACM 21st International Conference on Mining Software Repositories (MSR). IEEE Computer Society, Los Alamitos, CA, 607\u2013618. DOI: https:\/\/doi.ieeecomputersociety.org\/."},{"key":"e_1_3_2_18_2","doi-asserted-by":"publisher","DOI":"10.1109\/ESEM56168.2023.10304801"},{"key":"e_1_3_2_19_2","doi-asserted-by":"publisher","DOI":"10.1109\/MSR.2017.67"},{"key":"e_1_3_2_20_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.socnet.2012.12.003"},{"key":"e_1_3_2_21_2","doi-asserted-by":"publisher","DOI":"10.1007\/s10515-024-00453-w"},{"key":"e_1_3_2_22_2","doi-asserted-by":"publisher","DOI":"10.1109\/JVA60410.2023.00015"},{"key":"e_1_3_2_23_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/N19-1423"},{"key":"e_1_3_2_24_2","doi-asserted-by":"publisher","DOI":"10.1145\/3661167.3661215"},{"key":"e_1_3_2_25_2","doi-asserted-by":"publisher","DOI":"10.1109\/SANER56733.2023.00095"},{"issue":"1","key":"e_1_3_2_26_2","first-page":"1","article-title":"What makes a popular academic AI repository","volume":"26","author":"Fan Yuanrui","year":"2021","unstructured":"Yuanrui Fan, Xin Xia, David Lo, Ahmed E. Hassan, and Shanping Li. 2021. What makes a popular academic AI repository? Empirical Software Engineering 26, 1 (2021), 1\u201335.","journal-title":"Empirical Software Engineering"},{"key":"e_1_3_2_27_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.findings-emnlp.139"},{"key":"e_1_3_2_28_2","doi-asserted-by":"publisher","DOI":"10.1037\/h0031619"},{"key":"e_1_3_2_29_2","doi-asserted-by":"publisher","DOI":"10.1007\/s10664-023-10377-w"},{"key":"e_1_3_2_30_2","doi-asserted-by":"publisher","unstructured":"Xinyi Hou Yanjie Zhao Yue Liu Zhou Yang Kailong Wang Li Li Xiapu Luo David Lo John Grundy and Haoyu Wang. 2024. Large language models for software engineering: A systematic literature review. ACM Transactions on Software Engineering and Methodology 33 8 (Sept. 2024) 1\u201379. DOI: 10.1145\/3695988","DOI":"10.1145\/3695988"},{"key":"e_1_3_2_31_2","unstructured":"Edward J. Hu Yelong Shen Phillip Wallis Zeyuan Allen-Zhu Yuanzhi Li Shean Wang Lu Wang and Weizhu Chen. 2021. LoRA: Low-rank adaptation of large language models. arXiv:2106.09685. Retrieved from https:\/\/arxiv.org\/abs\/2106.09685"},{"key":"e_1_3_2_32_2","unstructured":"Hamel Husain Ho-Hsiang Wu Tiferet Gazit Miltiadis Allamanis and Marc Brockschmidt. 2019. CodeSearchNet challenge: Evaluating the state of semantic code search. arXiv:1909.09436. Retrieved from https:\/\/arxiv.org\/abs\/1909.09436"},{"key":"e_1_3_2_33_2","doi-asserted-by":"publisher","DOI":"10.1145\/2372251.2372257"},{"key":"e_1_3_2_34_2","unstructured":"Wenxin Jiang Chingwo Cheung Mingyu Kim Heesoo Kim George K. Thiruvathukal and James C. Davis. 2024. Naming practices of pre-trained models in Hugging Face. arXiv:2310.01642. Retrieved from https:\/\/arxiv.org\/abs\/2310.01642"},{"key":"e_1_3_2_35_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICSE48619.2023.00206"},{"key":"e_1_3_2_36_2","doi-asserted-by":"crossref","unstructured":"Wenxin Jiang Nicholas Synovic Purvish Jajal Taylor R. Schorlemmer Arav Tewari Bhavesh Pareek George K. Thiruvathukal and James C. Davis. 2023. PTMTorrent: A dataset for mining open-source pre-trained model packages. arXiv:2303.08934. Retrieved from https:\/\/arxiv.org\/abs\/2303.08934","DOI":"10.1109\/MSR59073.2023.00021"},{"key":"e_1_3_2_37_2","doi-asserted-by":"publisher","DOI":"10.1145\/3560835.3564547"},{"key":"e_1_3_2_38_2","first-page":"431","volume-title":"Proceedings of the 2024 IEEE\/ACM 21st International Conference on Mining Software Repositories (MSR)","author":"Jiang Wenxin","year":"2024","unstructured":"Wenxin Jiang, Jerin Yasmin, Jason Jones, Nicholas Synovic, Jiashen Kuo, Nathaniel Bielanski, Yuan Tian, George K. Thiruvathukal, and James C. Davis. 2024. PeaTMOSS: A dataset and initial analysis of pre-trained models in open-source software. In Proceedings of the 2024 IEEE\/ACM 21st International Conference on Mining Software Repositories (MSR), 431\u2013443."},{"key":"e_1_3_2_39_2","doi-asserted-by":"crossref","unstructured":"Jason Jones Wenxin Jiang Nicholas Synovic George K. Thiruvathukal and James C. Davis. 2024. What do we know about Hugging Face? A systematic literature review and quantitative validation of qualitative claims. arXiv:2406.08205. Retrieved from https:\/\/arxiv.org\/abs\/2406.08205","DOI":"10.1145\/3674805.3686665"},{"key":"e_1_3_2_40_2","doi-asserted-by":"publisher","DOI":"10.1109\/ISI58743.2023.10297271"},{"key":"e_1_3_2_41_2","doi-asserted-by":"publisher","DOI":"10.2307\/2529310"},{"key":"e_1_3_2_42_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.acl-long.577"},{"key":"e_1_3_2_43_2","unstructured":"Raymond Li Loubna Ben Allal Yangtian Zi Niklas Muennighoff Denis Kocetkov Chenghao Mou Marc Marone Christopher Akiki Jia Li Jenny Chim et al. 2023. StarCoder: May the source be with you! Transactions on Machine Learning Research (2023). Reproducibility Certification. Retrieved from https:\/\/openreview.net\/forum?id=KoFOg41haE"},{"key":"e_1_3_2_44_2","doi-asserted-by":"publisher","DOI":"10.1145\/3510003.3510142"},{"key":"e_1_3_2_45_2","unstructured":"Yinhan Liu Myle Ott Naman Goyal Jingfei Du Mandar Joshi Danqi Chen Omer Levy Mike Lewis Luke Zettlemoyer and Veselin Stoyanov. 2019. RoBERTa: A robustly optimized BERT pretraining approach. arXiv:1907.11692. Retrieved from http:\/\/arxiv.org\/abs\/1907.11692"},{"key":"e_1_3_2_46_2","unstructured":"Anton Lozhkov Raymond Li Loubna Ben Allal Federico Cassano Joel Lamy-Poirier Nouamane Tazi Ao Tang Dmytro Pykhtar Jiawei Liu Yuxiang Wei et al. 2024. StarCoder 2 and The Stack v2: The next generation. arXiv:2402.19173. Retrieved from https:\/\/arxiv.org\/abs\/2402.19173"},{"key":"e_1_3_2_47_2","unstructured":"Shuai Lu Daya Guo Shuo Ren Junjie Huang Alexey Svyatkovskiy Ambrosio Blanco Colin B. Clement Dawn Drain Daxin Jiang Duyu Tang et al. 2021. CodeXGLUE: A machine learning benchmark dataset for code understanding and generation. arXiv:2102.04664. Retrieved from https:\/\/arxiv.org\/abs\/2102.04664"},{"key":"e_1_3_2_48_2","first-page":"100","volume-title":"Proceedings of the 2020 IEEE\/ACM 42nd International Conference on Software Engineering (ICSE)","author":"Ma Wanwangying","year":"2020","unstructured":"Wanwangying Ma, Lin Chen, Xiangyu Zhang, Yang Feng, Zhaogui Xu, Zhifei Chen, Yuming Zhou, and Baowen Xu. 2020. Impact analysis of cross-project bugs on software ecosystems. In Proceedings of the 2020 IEEE\/ACM 42nd International Conference on Software Engineering (ICSE), 100\u2013111."},{"key":"e_1_3_2_49_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-08129-3_6"},{"key":"e_1_3_2_50_2","first-page":"1386","volume-title":"Proceedings of the 2021 IEEE\/ACM 43rd International Conference on Software Engineering (ICSE)","author":"Massacci Fabio","year":"2021","unstructured":"Fabio Massacci and Ivan Pashchenko. 2021. Technical leverage in a software ecosystem: Development opportunities and security risks. In Proceedings of the 2021 IEEE\/ACM 43rd International Conference on Software Engineering (ICSE). IEEE, 1386\u20131397."},{"key":"e_1_3_2_51_2","doi-asserted-by":"publisher","DOI":"10.11613\/BM.2012.031"},{"key":"e_1_3_2_52_2","doi-asserted-by":"publisher","DOI":"10.5555\/949280"},{"key":"e_1_3_2_53_2","doi-asserted-by":"publisher","DOI":"10.1109\/SANER56733.2023.00028"},{"key":"e_1_3_2_54_2","volume-title":"Proceedings of the 11th International Conference on Learning Representations","author":"Nijkamp Erik","year":"2023","unstructured":"Erik Nijkamp, Bo Pang, Hiroaki Hayashi, Lifu Tu, Huan Wang, Yingbo Zhou, Silvio Savarese, and Caiming Xiong. 2023. CodeGen: An open large language model for code with multi-turn program synthesis. In Proceedings of the 11th International Conference on Learning Representations."},{"key":"e_1_3_2_55_2","doi-asserted-by":"publisher","DOI":"10.1145\/3643916.3644412"},{"key":"e_1_3_2_56_2","doi-asserted-by":"publisher","DOI":"10.1145\/3379597.3387510"},{"key":"e_1_3_2_57_2","doi-asserted-by":"publisher","DOI":"10.1007\/s10664-018-9660-3"},{"issue":"8","key":"e_1_3_2_58_2","first-page":"9","article-title":"Language models are unsupervised multitask learners","volume":"1","author":"Radford Alec","year":"2019","unstructured":"Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, and Ilya Sutskever. 2019. Language models are unsupervised multitask learners. OpenAI Blog 1, 8 (2019), 9. Retrieved from https:\/\/cdn.openai.com\/better-language-models\/language_models_are_unsupervised_multitask_learners.pdf","journal-title":"OpenAI Blog"},{"key":"e_1_3_2_59_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICPR56361.2022.9956690"},{"key":"e_1_3_2_60_2","doi-asserted-by":"publisher","DOI":"10.1145\/3639475.3640097"},{"key":"e_1_3_2_61_2","doi-asserted-by":"publisher","unstructured":"Jieke Shi Zhou Yang and David Lo. 2024. Efficient and green large language models for software engineering: Vision and the road ahead. ACM Transactions on Software Engineering and Methodology (Dec. 2024). Just Accepted. DOI: 10.1145\/3708525","DOI":"10.1145\/3708525"},{"key":"e_1_3_2_62_2","doi-asserted-by":"publisher","DOI":"10.1145\/3551349.3556964"},{"key":"e_1_3_2_63_2","doi-asserted-by":"publisher","DOI":"10.1145\/3540250.3549145"},{"key":"e_1_3_2_64_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.acl-long.540"},{"key":"e_1_3_2_65_2","doi-asserted-by":"publisher","DOI":"10.1109\/SANER60148.2024.00059"},{"key":"e_1_3_2_66_2","doi-asserted-by":"publisher","DOI":"10.1109\/ASE56229.2023.00149"},{"key":"e_1_3_2_67_2","doi-asserted-by":"publisher","DOI":"10.1145\/3180155.3180209"},{"key":"e_1_3_2_68_2","first-page":"6000","article-title":"Attention is all you need","author":"Vaswani Ashish","year":"2017","unstructured":"Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, \u0141ukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. In Proceedings of the 31st International Conference on Neural Information Processing Systems. Curran Associates Inc., Red Hook, NY, 6000\u20136010.","journal-title":"Proceedings of the 31st International Conference on Neural Information Processing Systems"},{"key":"e_1_3_2_69_2","doi-asserted-by":"publisher","DOI":"10.1145\/3540250.3549153"},{"key":"e_1_3_2_70_2","unstructured":"Shenao Wang Yanjie Zhao Xinyi Hou and Haoyu Wang. 2024. Large language model supply chain: A research agenda. arXiv:2404.12736. Retrieved from https:\/\/arxiv.org\/abs\/2404.12736"},{"key":"e_1_3_2_71_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.emnlp-main.685"},{"key":"e_1_3_2_72_2","doi-asserted-by":"publisher","DOI":"10.1145\/3377811.3380426"},{"key":"e_1_3_2_73_2","doi-asserted-by":"publisher","DOI":"10.1017\/CBO9780511815478"},{"key":"e_1_3_2_74_2","doi-asserted-by":"publisher","DOI":"10.1145\/3611643.3616302"},{"key":"e_1_3_2_75_2","unstructured":"Martin Weyssow Xin Zhou Kisub Kim David Lo and Houari Sahraoui. 2024. Exploring parameter-efficient fine-tuning techniques for code generation with large language models. arXiv:2308.10462. Retrieved from https:\/\/arxiv.org\/abs\/2308.10462"},{"key":"e_1_3_2_76_2","doi-asserted-by":"publisher","DOI":"10.1145\/2601248.2601268"},{"key":"e_1_3_2_77_2","doi-asserted-by":"publisher","DOI":"10.1007\/s10664-024-10486-0"},{"key":"e_1_3_2_78_2","unstructured":"Zhou Yang and David Lo. 2024. Hotfixing large language models for code. arXiv:2408.05727. Retrieved from https:\/\/arxiv.org\/abs\/2408.05727"},{"key":"e_1_3_2_79_2","doi-asserted-by":"publisher","DOI":"10.1145\/3510003.3510146"},{"key":"e_1_3_2_80_2","doi-asserted-by":"crossref","unstructured":"Zhou Yang Zhensu Sun Terry Zhuo Yue Premkumar Devanbu and David Lo. 2024. Robustness security privacy explainability efficiency and usability of large language models for code. arXiv:2403.07506. Retrieved from https:\/\/arxiv.org\/abs\/2403.07506","DOI":"10.1145\/3731753"},{"key":"e_1_3_2_81_2","doi-asserted-by":"publisher","DOI":"10.1109\/MSR59073.2023.00024"},{"key":"e_1_3_2_82_2","doi-asserted-by":"publisher","DOI":"10.1109\/TSE.2024.3361661"},{"key":"e_1_3_2_83_2","doi-asserted-by":"publisher","DOI":"10.1109\/TSE.2024.3482719"},{"key":"e_1_3_2_84_2","doi-asserted-by":"publisher","DOI":"10.1145\/3428230"},{"key":"e_1_3_2_85_2","doi-asserted-by":"publisher","DOI":"10.1145\/3446905"},{"key":"e_1_3_2_86_2","volume-title":"Advances in Neural Information Processing Systems","author":"Zhu Ligeng","year":"2019","unstructured":"Ligeng Zhu, Zhijian Liu, and Song Han. 2019. Deep Leakage from Gradients. In Advances in Neural Information Processing Systems. H. Wallach, H. Larochelle, A. Beygelzimer, F. d\u2019Alch\u00e9-Buc, E. Fox, and R. Garnett (Eds.), Vol. 32, Curran Associates, Inc. Retrieved from https:\/\/proceedings.neurips.cc\/paper_files\/paper\/2019\/file\/60a6c4002cc7b29142def8871531281a-Paper.pdf"}],"container-title":["ACM Transactions on Software Engineering and Methodology"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3731753","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,12,11]],"date-time":"2025-12-11T15:56:20Z","timestamp":1765468580000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3731753"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,12,11]]},"references-count":85,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2026,1,31]]}},"alternative-id":["10.1145\/3731753"],"URL":"https:\/\/doi.org\/10.1145\/3731753","relation":{},"ISSN":["1049-331X","1557-7392"],"issn-type":[{"value":"1049-331X","type":"print"},{"value":"1557-7392","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,12,11]]},"assertion":[{"value":"2024-09-28","order":0,"name":"received","label":"Received","group":{"name":"publication_history","label":"Publication History"}},{"value":"2025-04-03","order":2,"name":"accepted","label":"Accepted","group":{"name":"publication_history","label":"Publication History"}},{"value":"2025-12-11","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}