{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,30]],"date-time":"2026-03-30T20:56:56Z","timestamp":1774904216269,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":71,"publisher":"ACM","content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,6,17]]},"DOI":"10.1145\/3756681.3756999","type":"proceedings-article","created":{"date-parts":[[2025,12,24]],"date-time":"2025-12-24T08:30:04Z","timestamp":1766565004000},"page":"1024-1034","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":1,"title":["LAMeD: LLM-generated Annotations for Memory Leak Detection"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-1577-8347","authenticated-orcid":false,"given":"Ekaterina","family":"Shemetova","sequence":"first","affiliation":[{"name":"St. Petersburg University, St. Petersburg, Russian Federation"}]},{"ORCID":"https:\/\/orcid.org\/0009-0002-3690-4733","authenticated-orcid":false,"given":"Ivan","family":"Smirnov","sequence":"additional","affiliation":[{"name":"ITMO University, St. Petersburg, Russian Federation"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6456-3329","authenticated-orcid":false,"given":"Anton","family":"Alekseev","sequence":"additional","affiliation":[{"name":"St. Petersburg Department of Steklov Institute of Mathematics, St. Petersburg, Russian Federation; St. Petersburg University, St. Petersburg, Russian Federation and KSTU n. a. I. Razzakov, Bishkek, Kirghizstan"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6778-225X","authenticated-orcid":false,"given":"Ilya","family":"Shenbin","sequence":"additional","affiliation":[{"name":"St. Petersburg Department of Steklov Institute of Mathematics, St. Petersburg, Russian Federation"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5701-1785","authenticated-orcid":false,"given":"Alexey","family":"Rukhovich","sequence":"additional","affiliation":[{"name":"AI Foundation and Algorithm Lab, Moscow, Russian Federation"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7787-2251","authenticated-orcid":false,"given":"Sergey","family":"Nikolenko","sequence":"additional","affiliation":[{"name":"St. Petersburg Department of Steklov Institute of Mathematics, St. Petersburg, Russian Federation"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8991-9264","authenticated-orcid":false,"given":"Vadim","family":"Lomshakov","sequence":"additional","affiliation":[{"name":"St. Petersburg Department of Steklov Institute of Mathematics, St. Petersburg, Russian Federation"}]},{"ORCID":"https:\/\/orcid.org\/0009-0003-0299-5849","authenticated-orcid":false,"given":"Irina","family":"Piontkovskaya","sequence":"additional","affiliation":[{"name":"AI Foundation and Algorithm Lab, Moscow, Russian Federation"}]}],"member":"320","published-online":{"date-parts":[[2025,12,24]]},"reference":[{"key":"e_1_3_3_2_2_2","unstructured":"Josh Achiam Steven Adler Sandhini Agarwal Lama Ahmad Ilge Akkaya Florencia\u00a0Leoni Aleman Diogo Almeida Janko Altenschmidt Sam Altman Shyamal Anadkat et\u00a0al. 2023. Gpt-4 technical report. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2303.08774 (2023)."},{"key":"e_1_3_3_2_3_2","unstructured":"Mistral AI. 2024. Codestral-22B-v0.1: A 22-Billion-Parameter Model for Code Generation. https:\/\/huggingface.co\/mistralai\/Codestral-22B-v0.1 Accessed: 2025-03-23."},{"key":"e_1_3_3_2_4_2","doi-asserted-by":"crossref","unstructured":"Meysam Alizadeh Ma\u00ebl Kubli Zeynab Samei Shirin Dehghani Mohammadmasiha Zahedivafa Juan\u00a0Diego Bermeo Maria Korobeynikova and Fabrizio Gilardi. 2024. Open-Source LLMs for Text Annotation: A Practical Guide for Model Setting and Fine-Tuning. arxiv:https:\/\/arXiv.org\/abs\/2307.02179\u00a0[cs.CL] https:\/\/arxiv.org\/abs\/2307.02179","DOI":"10.1007\/s42001-024-00345-9"},{"key":"e_1_3_3_2_5_2","unstructured":"Abhinav Anand Shweta Verma Krishna Narasimhan and Mira Mezini. 2024. A Critical Study of What Code-LLMs (Do Not) Learn. arxiv:https:\/\/arXiv.org\/abs\/2406.11930\u00a0[cs.SE] https:\/\/arxiv.org\/abs\/2406.11930"},{"key":"e_1_3_3_2_6_2","doi-asserted-by":"crossref","unstructured":"Hayk Aslanyan Hovhannes Movsisyan Hripsime Hovhannisyan Zhora Gevorgyan Ruslan Mkoyan Arutyun Avetisyan and Sevak Sargsyan. 2024. Combining Static Analysis with Directed Symbolic Execution for Scalable and Accurate Memory Leak Detection. IEEE Access (2024).","DOI":"10.1109\/ACCESS.2024.3409838"},{"key":"e_1_3_3_2_7_2","unstructured":"Sam Blackshear and PW O\u2019Hearn. 2017. Open-sourcing RacerD: fast static race detection at scale. code.facebook.com blog post (2017)."},{"key":"e_1_3_3_2_8_2","unstructured":"Tom\u00a0B Brown. 2020. Language models are few-shot learners. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2005.14165 (2020)."},{"key":"e_1_3_3_2_9_2","doi-asserted-by":"publisher","DOI":"10.1109\/CGO.2011.5764689"},{"key":"e_1_3_3_2_10_2","first-page":"209","volume-title":"OSDI","author":"Cadar Cristian","year":"2008","unstructured":"Cristian Cadar, Daniel Dunbar, Dawson\u00a0R Engler, et\u00a0al. 2008. Klee: unassisted and automatic generation of high-coverage tests for complex systems programs.. In OSDI , Vol.\u00a08. 209\u2013224."},{"key":"e_1_3_3_2_11_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-20398-5_33"},{"key":"e_1_3_3_2_12_2","doi-asserted-by":"publisher","DOI":"10.1145\/1480881.1480917"},{"key":"e_1_3_3_2_13_2","unstructured":"Cristiano Calcagno Dino Distefano and Peter O\u2019Hearn. 2015. Open-sourcing Facebook Infer: Identify bugs before you ship. code. facebook. com blog post 11 (2015)."},{"key":"e_1_3_3_2_14_2","doi-asserted-by":"publisher","unstructured":"Cristiano Calcagno Dino Distefano Peter\u00a0W. O\u2019Hearn and Hongseok Yang. 2011. Compositional Shape Analysis by Means of Bi-Abduction. J. ACM 58 6 Article 26 (dec 2011) 66\u00a0pages. 10.1145\/2049697.2049700","DOI":"10.1145\/2049697.2049700"},{"key":"e_1_3_3_2_15_2","doi-asserted-by":"publisher","DOI":"10.1145\/3652588.3663317"},{"key":"e_1_3_3_2_16_2","doi-asserted-by":"publisher","DOI":"10.1145\/3607199.3607242"},{"key":"e_1_3_3_2_17_2","doi-asserted-by":"publisher","DOI":"10.1145\/3639478.3647633"},{"key":"e_1_3_3_2_18_2","volume-title":"Clang Static Analyzer","author":"Team Clang","year":"2024","unstructured":"Clang Team. 2024. Clang Static Analyzer. LLVM Project. https:\/\/clang-analyzer.llvm.org\/ Accessed: 2024-08-23."},{"key":"e_1_3_3_2_19_2","doi-asserted-by":"publisher","DOI":"10.1145\/260028.260173"},{"key":"e_1_3_3_2_20_2","first-page":"4171","volume-title":"Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers)","author":"Devlin Jacob","year":"2019","unstructured":"Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers). 4171\u20134186."},{"key":"e_1_3_3_2_21_2","unstructured":"Yangruibo Ding Yanjun Fu Omniyyah Ibrahim Chawin Sitawarin Xinyun Chen Basel Alomair David Wagner Baishakhi Ray and Yizheng Chen. 2024. Vulnerability detection with code language models: How far are we? arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2403.18624 (2024)."},{"key":"e_1_3_3_2_22_2","doi-asserted-by":"crossref","unstructured":"Dino Distefano Manuel F\u00e4hndrich Francesco Logozzo and Peter\u00a0W O\u2019Hearn. 2019. Scaling static analyses at Facebook. Commun. ACM 62 8 (2019) 62\u201370.","DOI":"10.1145\/3338112"},{"key":"e_1_3_3_2_23_2","doi-asserted-by":"publisher","DOI":"10.1145\/2663716.2663755"},{"key":"e_1_3_3_2_24_2","doi-asserted-by":"publisher","DOI":"10.14722\/ndss.2021.24416"},{"key":"e_1_3_3_2_25_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICSE.2019.00025"},{"key":"e_1_3_3_2_26_2","unstructured":"Chongzhou Fang Ning Miao Shaurya Srivastav Jialin Liu Ruoyu Zhang Ruijie Fang Asmita Ryan Tsang Najmeh Nazari Han Wang and Houman Homayoun. 2024. Large Language Models for Code Analysis: Do LLMs Really Do Their Job? arxiv:https:\/\/arXiv.org\/abs\/2310.12357\u00a0[cs.SE] https:\/\/arxiv.org\/abs\/2310.12357"},{"key":"e_1_3_3_2_27_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.findings-emnlp.139"},{"key":"e_1_3_3_2_28_2","unstructured":"Zeyu Gao Hao Wang Yuchen Zhou Wenyu Zhu and Chao Zhang. 2023. How far have we gone in vulnerability detection using large language models. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2311.12420 (2023)."},{"key":"e_1_3_3_2_29_2","doi-asserted-by":"crossref","unstructured":"Alexander\u00a0Yurievich Gerasimov Alexey\u00a0Alexeyevich Kanakhin Petr\u00a0Alekseevich Privalov Andrey\u00a0Alexandrovich Zhukov and Evgenii\u00a0Arkadievich Kaminskii. 2022. Case study: Source code static analysis for performance issues detection. Proceedings of the Institute for System Programming of the RAS (Proceedings of ISP RAS) 34 4 (2022) 7\u201320.","DOI":"10.15514\/ISPRAS-2022-34(4)-1"},{"key":"e_1_3_3_2_30_2","doi-asserted-by":"publisher","unstructured":"Seyed\u00a0Mohammad Ghaffarian and Hamid\u00a0Reza Shahriari. 2017. Software Vulnerability Analysis and Discovery Using Machine-Learning and Data-Mining Techniques: A Survey. ACM Comput. Surv. 50 4 Article 56 (aug 2017) 36\u00a0pages. 10.1145\/3092566","DOI":"10.1145\/3092566"},{"key":"e_1_3_3_2_31_2","unstructured":"Daya Guo Dejian Yang Haowei Zhang Junxiao Song Ruoyu Zhang Runxin Xu Qihao Zhu Shirong Ma Peiyi Wang Xiao Bi et\u00a0al. 2025. Deepseek-r1: Incentivizing reasoning capability in llms via reinforcement learning. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2501.12948 (2025)."},{"key":"e_1_3_3_2_32_2","unstructured":"Daya Guo Qihao Zhu Dejian Yang Zhenda Xie Kai Dong Wentao Zhang Guanting Chen Xiao Bi Yu Wu YK Li et\u00a0al. 2024. DeepSeek-Coder: When the Large Language Model Meets Programming\u2013The Rise of Code Intelligence. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2401.14196 (2024)."},{"key":"e_1_3_3_2_33_2","doi-asserted-by":"publisher","unstructured":"Yu Hao Weiteng Chen Ziqiao Zhou and Weidong Cui. 2023. E&V: Prompting Large Language Models to Perform Static Analysis by Pseudo-code Execution and Verification. CoRR abs\/2312.08477 (2023). 10.48550\/ARXIV.2312.08477 arXiv:https:\/\/arXiv.org\/abs\/2312.08477","DOI":"10.48550\/ARXIV.2312.08477"},{"key":"e_1_3_3_2_34_2","doi-asserted-by":"publisher","DOI":"10.5555\/2190025.2190075"},{"key":"e_1_3_3_2_35_2","doi-asserted-by":"publisher","unstructured":"Junda He Christoph Treude and David Lo. 2025. LLM-Based Multi-Agent Systems for Software Engineering: Literature Review Vision and the Road Ahead. ACM Trans. Softw. Eng. Methodol. (Jan. 2025). 10.1145\/3712003Just Accepted.","DOI":"10.1145\/3712003"},{"key":"e_1_3_3_2_36_2","doi-asserted-by":"crossref","unstructured":"Soneya\u00a0Binta Hossain Nan Jiang Qiang Zhou Xiaopeng Li Wen-Hao Chiang Yingjun Lyu Hoan Nguyen and Omer Tripp. 2024. A deep dive into large language models for automated bug localization and repair. Proceedings of the ACM on Software Engineering 1 FSE (2024) 1471\u20131493.","DOI":"10.1145\/3660773"},{"key":"e_1_3_3_2_37_2","volume-title":"An awesome and curated list of best code-LLM for research","author":"Hui Binyuan","year":"2023","unstructured":"Binyuan Hui. 2023. An awesome and curated list of best code-LLM for research."},{"key":"e_1_3_3_2_38_2","unstructured":"Binyuan Hui Jian Yang Zeyu Cui Jiaxi Yang Dayiheng Liu Lei Zhang Tianyu Liu Jiajun Zhang Bowen Yu Keming Lu et\u00a0al. 2024. Qwen2. 5-coder technical report. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2409.12186 (2024)."},{"key":"e_1_3_3_2_39_2","volume-title":"Joern: The Bug Hunter\u2019s Workbench","year":"2024","unstructured":"joern.io. 2024. Joern: The Bug Hunter\u2019s Workbench. https:\/\/github.com\/joernio\/joern"},{"key":"e_1_3_3_2_40_2","doi-asserted-by":"publisher","DOI":"10.1145\/1375634.1375653"},{"key":"e_1_3_3_2_41_2","doi-asserted-by":"crossref","unstructured":"Sungmin Kang Gabin An and Shin Yoo. 2024. A quantitative and qualitative evaluation of LLM-based explainable fault localization. Proceedings of the ACM on Software Engineering 1 FSE (2024) 1424\u20131446.","DOI":"10.1145\/3660771"},{"key":"e_1_3_3_2_42_2","doi-asserted-by":"publisher","DOI":"10.1145\/3510003.3510153"},{"key":"e_1_3_3_2_43_2","unstructured":"Hannah Kim Kushan Mitra Rafael\u00a0Li Chen Sajjadur Rahman and Dan Zhang. 2024. MEGAnno+: A Human-LLM Collaborative Annotation System. arxiv:https:\/\/arXiv.org\/abs\/2402.18050\u00a0[cs.CL] https:\/\/arxiv.org\/abs\/2402.18050"},{"key":"e_1_3_3_2_44_2","unstructured":"Harrison Lee Samrat Phatale Hassan Mansoor Thomas Mesnard Johan Ferret Kellie Lu Colton Bishop Ethan Hall Victor Carbune Abhinav Rastogi et\u00a0al. 2023. Rlaif: Scaling reinforcement learning from human feedback with ai feedback. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2309.00267 (2023)."},{"key":"e_1_3_3_2_45_2","doi-asserted-by":"crossref","unstructured":"Haonan Li Yu Hao Yizhuo Zhai and Zhiyun Qian. 2024. Enhancing Static Analysis for Practical Bug Detection: An LLM-Integrated Approach. Proceedings of the ACM on Programming Languages 8 OOPSLA1 (2024) 474\u2013499.","DOI":"10.1145\/3649828"},{"key":"e_1_3_3_2_46_2","doi-asserted-by":"publisher","DOI":"10.1145\/3368089.3417923"},{"key":"e_1_3_3_2_47_2","doi-asserted-by":"crossref","unstructured":"Ye Liu Yue Xue Daoyuan Wu Yuqiang Sun Yi Li Miaolei Shi and Yang Liu. 2024. PropertyGPT: LLM-driven Formal Verification of Smart Contracts through Retrieval-Augmented Property Generation. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2405.02580 (2024).","DOI":"10.14722\/ndss.2025.241357"},{"key":"e_1_3_3_2_48_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-05237-8_66"},{"key":"e_1_3_3_2_49_2","doi-asserted-by":"publisher","unstructured":"Mohammad\u00a0Mahdi Mohajer Reem Aleithan Nima\u00a0Shiri Harzevili Moshi Wei Alvine\u00a0Boaye Belle Hung\u00a0Viet Pham and Song Wang. 2023. SkipAnalyzer: An Embodied Agent for Code Analysis with Large Language Models. CoRR abs\/2310.18532 (2023). 10.48550\/ARXIV.2310.18532 arXiv:https:\/\/arXiv.org\/abs\/2310.18532","DOI":"10.48550\/ARXIV.2310.18532"},{"key":"e_1_3_3_2_50_2","unstructured":"L\u00e1szl\u00f3 Nagy. 2025. Bear: A tool that generates a compilation database for clang tooling. https:\/\/github.com\/rizsotto\/Bear Accessed: 2025-03-23."},{"key":"e_1_3_3_2_51_2","doi-asserted-by":"crossref","unstructured":"Arbi\u00a0Haza Nasution and Aytug Onan. 2024. ChatGPT Label: Comparing the Quality of Human-Generated and LLM-Generated Annotations in Low-resource Language NLP Tasks. IEEE Access (2024).","DOI":"10.1109\/ACCESS.2024.3402809"},{"key":"e_1_3_3_2_52_2","doi-asserted-by":"crossref","unstructured":"Nicholas Nethercote and Julian Seward. 2007. Valgrind: a framework for heavyweight dynamic binary instrumentation. ACM Sigplan notices 42 6 (2007) 89\u2013100.","DOI":"10.1145\/1273442.1250746"},{"key":"e_1_3_3_2_53_2","doi-asserted-by":"crossref","unstructured":"Peter O\u2019Hearn. 2019. Separation logic. Commun. ACM 62 2 (2019) 86\u201395.","DOI":"10.1145\/3211968"},{"key":"e_1_3_3_2_54_2","doi-asserted-by":"publisher","DOI":"10.1145\/2872427.2883029"},{"key":"e_1_3_3_2_55_2","unstructured":"Alec Radford Jeffrey Wu Rewon Child David Luan Dario Amodei Ilya Sutskever et\u00a0al. 2019. Language models are unsupervised multitask learners. OpenAI blog 1 8 (2019) 9."},{"key":"e_1_3_3_2_56_2","doi-asserted-by":"publisher","DOI":"10.1109\/LICS.2002.1029817"},{"key":"e_1_3_3_2_57_2","unstructured":"Baptiste Roziere Jonas Gehring Fabian Gloeckle Sten Sootla Itai Gat Xiaoqing\u00a0Ellen Tan Yossi Adi Jingyu Liu Romain Sauvestre Tal Remez et\u00a0al. 2023. Code llama: Open foundation models for code. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2308.12950 (2023)."},{"key":"e_1_3_3_2_58_2","first-page":"309","volume-title":"2012 USENIX annual technical conference (USENIX ATC 12)","author":"Serebryany Konstantin","year":"2012","unstructured":"Konstantin Serebryany, Derek Bruening, Alexander Potapenko, and Dmitriy Vyukov. 2012. { AddressSanitizer} : A fast address sanity checker. In 2012 USENIX annual technical conference (USENIX ATC 12). 309\u2013318."},{"key":"e_1_3_3_2_59_2","doi-asserted-by":"publisher","DOI":"10.1145\/2854038.2854043"},{"key":"e_1_3_3_2_60_2","doi-asserted-by":"publisher","DOI":"10.1145\/2892208.2892235"},{"key":"e_1_3_3_2_61_2","doi-asserted-by":"publisher","DOI":"10.1145\/2338965.2336784"},{"key":"e_1_3_3_2_62_2","unstructured":"Yuqiang Sun Daoyuan Wu Yue Xue Han Liu Wei Ma Lyuye Zhang Miaolei Shi and Yang Liu. 2024. LLM4Vuln: A Unified Evaluation Framework for Decoupling and Enhancing LLMs\u2019 Vulnerability Reasoning. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2401.16185 (2024)."},{"key":"e_1_3_3_2_63_2","doi-asserted-by":"publisher","DOI":"10.5555\/3295222.3295349"},{"key":"e_1_3_3_2_64_2","doi-asserted-by":"crossref","unstructured":"Nalin Wadhwa Jui Pradhan Atharv Sonwane Surya\u00a0Prakash Sahu Nagarajan Natarajan Aditya Kanade Suresh Parthasarathy and Sriram Rajamani. 2024. CORE: Resolving Code Quality Issues using LLMs. Proceedings of the ACM on Software Engineering 1 FSE (2024) 789\u2013811.","DOI":"10.1145\/3643762"},{"key":"e_1_3_3_2_65_2","doi-asserted-by":"crossref","unstructured":"Chong Wang Jianan Liu Xin Peng Yang Liu and Yiling Lou. 2025. Boosting Static Resource Leak Detection via LLM-based Resource-Oriented Intention Inference. Proceedings of the 47th International Conference on Software Engineering (2025).","DOI":"10.1109\/ICSE55347.2025.00131"},{"key":"e_1_3_3_2_66_2","unstructured":"Jin Wang Zishan Huang Hengli Liu Nianyi Yang and Yinhao Xiao. 2023. Defecthunter: A novel llm-driven boosted-conformer-based code vulnerability detection mechanism. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2309.15324 (2023)."},{"key":"e_1_3_3_2_67_2","unstructured":"Jiwei Yan Jinhao Huang Chunrong Fang Jun Yan and Jian Zhang. 2024. Better Debugging: Combining Static Analysis and LLMs for Explainable Crashing Fault Localization. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2408.12070 (2024)."},{"key":"e_1_3_3_2_68_2","doi-asserted-by":"publisher","DOI":"10.1145\/2581122.2544154"},{"key":"e_1_3_3_2_69_2","unstructured":"Kwangkeun Yi. 2017. Inferbo: Infer-based buffer overrun analyzer. Meta Research. Retrieved Feburary 6 (2017) 2023. https:\/\/research.facebook.com\/blog\/2017\/2\/inferbo-infer-based-buffer-overrun-analyzer\/"},{"key":"e_1_3_3_2_70_2","unstructured":"Jie Zhang Haoyu Bu Hui Wen Yu Chen Lun Li and Hongsong Zhu. 2024. When llms meet cybersecurity: A systematic literature review. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2405.03644 (2024)."},{"key":"e_1_3_3_2_71_2","unstructured":"Ziyin Zhang Chaoyu Chen Bingchang Liu Cong Liao Zi Gong Hang Yu Jianguo Li and Rui Wang. 2024. Unifying the Perspectives of NLP and Software Engineering: A Survey on Language Models for Code. arxiv:https:\/\/arXiv.org\/abs\/2311.07989\u00a0[cs.CL] https:\/\/arxiv.org\/abs\/2311.07989"},{"key":"e_1_3_3_2_72_2","unstructured":"Xin Zhou Sicong Cao Xiaobing Sun and David Lo. 2024. Large language model for vulnerability detection and repair: Literature review and the road ahead. ACM Transactions on Software Engineering and Methodology (2024)."}],"event":{"name":"EASE '25: Evaluation and Assessment in Software Engineering","location":"Istanbul Turkiye","acronym":"EASE '25"},"container-title":["Proceedings of the 29th International Conference on Evaluation and Assessment in Software Engineering"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3756681.3756999","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,12,24]],"date-time":"2025-12-24T08:39:44Z","timestamp":1766565584000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3756681.3756999"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,6,17]]},"references-count":71,"alternative-id":["10.1145\/3756681.3756999","10.1145\/3756681"],"URL":"https:\/\/doi.org\/10.1145\/3756681.3756999","relation":{},"subject":[],"published":{"date-parts":[[2025,6,17]]},"assertion":[{"value":"2025-12-24","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}