{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,30]],"date-time":"2026-01-30T06:40:33Z","timestamp":1769755233728,"version":"3.49.0"},"publisher-location":"New York, NY, USA","reference-count":51,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,9,11]],"date-time":"2024-09-11T00:00:00Z","timestamp":1726012800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"name":"National Research Foundation, under its Investigatorship Grant","award":["NRF-NRFI08-2022-0002"],"award-info":[{"award-number":["NRF-NRFI08-2022-0002"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,9,11]]},"DOI":"10.1145\/3650212.3680347","type":"proceedings-article","created":{"date-parts":[[2024,9,11]],"date-time":"2024-09-11T11:44:25Z","timestamp":1726055065000},"page":"1124-1136","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":8,"title":["AI Coders Are among Us: Rethinking Programming Language Grammar towards Efficient Code Generation"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-5393-7858","authenticated-orcid":false,"given":"Zhensu","family":"Sun","sequence":"first","affiliation":[{"name":"Singapore Management University, Singapore, Singapore"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3728-9541","authenticated-orcid":false,"given":"Xiaoning","family":"Du","sequence":"additional","affiliation":[{"name":"Monash University, Melbourne, Australia"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5938-1918","authenticated-orcid":false,"given":"Zhou","family":"Yang","sequence":"additional","affiliation":[{"name":"Singapore Management University, Singapore, Singapore"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2990-1614","authenticated-orcid":false,"given":"Li","family":"Li","sequence":"additional","affiliation":[{"name":"Beihang University, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4367-7201","authenticated-orcid":false,"given":"David","family":"Lo","sequence":"additional","affiliation":[{"name":"Singapore Management University, Singapore, Singapore"}]}],"member":"320","published-online":{"date-parts":[[2024,9,11]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"DeepSeek AI. 2023. DeepSeek Coder: Let the Code Write Itself. https:\/\/github.com\/deepseek-ai\/DeepSeek-Coder"},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.48550\/ARXIV.2301.03988"},{"key":"e_1_3_2_1_4_1","volume-title":"International Conference on Machine Learning. 2397\u20132430","author":"Biderman Stella","year":"2023","unstructured":"Stella Biderman, Hailey Schoelkopf, Quentin Gregory Anthony, Herbie Bradley, Kyle O\u2019Brien, Eric Hallahan, Mohammad Aflah Khan, Shivanshu Purohit, USVSN Sai Prashanth, and Edward Raff. 2023. Pythia: A suite for analyzing large language models across training and scaling. In International Conference on Machine Learning. 2397\u20132430."},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1109\/ASE.2019.00014"},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1145\/3377816.3381720"},{"key":"e_1_3_2_1_7_1","unstructured":"Harrison Chase. 2022. LangChain. https:\/\/github.com\/langchain-ai\/langchain"},{"key":"e_1_3_2_1_8_1","unstructured":"Mark Chen Jerry Tworek Heewoo Jun Qiming Yuan Henrique Ponde Jared Kaplan Harrison Edwards Yura Burda Nicholas Joseph Greg Brockman Alex Ray Raul Puri Gretchen Krueger Michael Petrov Heidy Khlaaf Girish Sastry Pamela Mishkin Brooke Chan Scott Gray Nick Ryder Mikhail Pavlov Alethea Power Lukasz Kaiser Mohammad Bavarian Clemens Winter Philippe Tillet Felipe Petroski Such David W. Cummings Matthias Plappert Fotios Chantzis Elizabeth Barnes Ariel Herbert-Voss William H. Guss Alex Nichol Igor Babuschkin S. Arun Balaji Shantanu Jain Andrew Carr Jan Leike Joshua Achiam Vedant Misra Evan Morikawa Alec Radford Matthew M. Knight Miles Brundage Mira Murati Katie Mayer Peter Welinder Bob McGrew Dario Amodei Sam McCandlish Ilya Sutskever and Wojciech Zaremba. 2021. Evaluating Large Language Models Trained on Code. ArXiv abs\/2107.03374 (2021) https:\/\/api.semanticscholar.org\/CorpusID:235755472"},{"key":"e_1_3_2_1_9_1","volume-title":"Conference on Empirical Methods in Natural Language Processing. https:\/\/api.semanticscholar.org\/CorpusID:252780909","author":"Chen Nuo","year":"2022","unstructured":"Nuo Chen, Qiushi Sun, Renyu Zhu, Xiang Li, Xuesong Lu, and Ming Gao. 2022. CAT-probing: A Metric-based Approach to Interpret How Pre-trained Models for Programming Language Attend Code Structure. In Conference on Empirical Methods in Natural Language Processing. https:\/\/api.semanticscholar.org\/CorpusID:252780909"},{"key":"e_1_3_2_1_10_1","volume-title":"CodeBPE: Investigating Subtokenization Options for Large Language Model Pretraining on Source Code. ArXiv, abs\/2308.00683","author":"Chirkova Nadezhda","year":"2023","unstructured":"Nadezhda Chirkova and Sergey Troshin. 2023. CodeBPE: Investigating Subtokenization Options for Large Language Model Pretraining on Source Code. ArXiv, abs\/2308.00683 (2023), https:\/\/api.semanticscholar.org\/CorpusID:252600018"},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.suscom.2023.100857"},{"key":"e_1_3_2_1_12_1","volume-title":"CodeBERT: A Pre-Trained Model for Programming and Natural Languages. ArXiv, abs\/2002.08155","author":"Feng Zhangyin","year":"2020","unstructured":"Zhangyin Feng, Daya Guo, Duyu Tang, Nan Duan, Xiaocheng Feng, Ming Gong, Linjun Shou, Bing Qin, Ting Liu, Daxin Jiang, and Ming Zhou. 2020. CodeBERT: A Pre-Trained Model for Programming and Natural Languages. ArXiv, abs\/2002.08155 (2020), https:\/\/api.semanticscholar.org\/CorpusID:211171605"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1145\/321312.321318"},{"key":"e_1_3_2_1_14_1","unstructured":"Google. 2023. ChatGPT. https:\/\/bard.google.com\/"},{"key":"e_1_3_2_1_15_1","volume-title":"Jacobs","author":"Grune Dick","year":"2007","unstructured":"Dick Grune and Ceriel J. H. Jacobs. 2007. Parsing Techniques - A Practical Guide. In Monographs in Computer Science. https:\/\/api.semanticscholar.org\/CorpusID:33077869"},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"crossref","unstructured":"Xinyi Hou Yanjie Zhao Yue Liu Zhou Yang Kailong Wang Li Li Xiapu Luo David Lo John Grundy and Haoyu Wang. 2023. Large Language Models for Software Engineering: A Systematic Literature Review. arxiv:2308.10620.","DOI":"10.1145\/3695988"},{"key":"e_1_3_2_1_17_1","unstructured":"[n. d.]."},{"key":"e_1_3_2_1_18_1","volume-title":"Program Translation via Code Distillation. ArXiv, abs\/2310.11476","author":"Huang Yufan","year":"2023","unstructured":"Yufan Huang, Mengnan Qi, Yongqiang Yao, Maoquan Wang, Bin Gu, Colin B. Clement, and Neel Sundaresan. 2023. Program Translation via Code Distillation. ArXiv, abs\/2310.11476 (2023), https:\/\/api.semanticscholar.org\/CorpusID:264289043"},{"key":"e_1_3_2_1_19_1","volume-title":"Scaling Laws for Neural Language Models. ArXiv, abs\/2001.08361","author":"Kaplan Jared","year":"2020","unstructured":"Jared Kaplan, Sam McCandlish, T. J. Henighan, Tom B. Brown, Benjamin Chess, Rewon Child, Scott Gray, Alec Radford, Jeff Wu, and Dario Amodei. 2020. Scaling Laws for Neural Language Models. ArXiv, abs\/2001.08361 (2020), https:\/\/api.semanticscholar.org\/CorpusID:210861095"},{"key":"e_1_3_2_1_20_1","volume-title":"2020 IEEE\/ACM 42nd International Conference on Software Engineering (ICSE), 1073\u20131085","author":"Karampatsis Rafael-Michael","year":"2020","unstructured":"Rafael-Michael Karampatsis, Hlib Babii, Romain Robbes, Charles Sutton, and Andrea Janes. 2020. Big Code != Big Vocabulary: Open-Vocabulary Models for Source Code. 2020 IEEE\/ACM 42nd International Conference on Software Engineering (ICSE), 1073\u20131085. https:\/\/api.semanticscholar.org\/CorpusID:211161525"},{"key":"e_1_3_2_1_21_1","volume-title":"Jia Li, Chenghao Mou, Carlos Mu\u00f1oz Ferrandis, Yacine Jernite, Margaret Mitchell, Sean Hughes, Thomas Wolf, Dzmitry Bahdanau, Leandro von Werra, and Harm de Vries.","author":"Kocetkov Denis","year":"2022","unstructured":"Denis Kocetkov, Raymond Li, Loubna Ben Allal, Jia Li, Chenghao Mou, Carlos Mu\u00f1oz Ferrandis, Yacine Jernite, Margaret Mitchell, Sean Hughes, Thomas Wolf, Dzmitry Bahdanau, Leandro von Werra, and Harm de Vries. 2022. The Stack: 3 TB of permissively licensed source code. Preprint."},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"crossref","unstructured":"Bernard Lang. 1974. Deterministic Techniques for Efficient Non-Deterministic Parsers. In International Colloquium on Automata Languages and Programming. https:\/\/api.semanticscholar.org\/CorpusID:27069587","DOI":"10.1007\/3-540-06841-4_65"},{"key":"e_1_3_2_1_23_1","unstructured":"Raymond Li Loubna Ben Allal Yangtian Zi Niklas Muennighoff Denis Kocetkov Chenghao Mou Marc Marone Christopher Akiki Jia Li Jenny Chim Qian Liu Evgenii Zheltonozhskii Terry Yue Zhuo Thomas Wang Olivier Dehaene Mishig Davaadorj Joel Lamy-Poirier Jo\u00e3o Monteiro Oleh Shliazhko Nicolas Gontier Nicholas Meade Armel Zebaze Ming-Ho Yee Logesh Kumar Umapathi Jian Zhu Benjamin Lipkin Muhtasham Oblokulov Zhiruo Wang Rudra Murthy Jason Stillerman Siva Sankalp Patel Dmitry Abulkhanov Marco Zocca Manan Dey Zhihan Zhang Nourhan Fahmy Urvashi Bhattacharyya W. Yu Swayam Singh Sasha Luccioni Paulo Villegas Maxim Kunakov Fedor Zhdanov Manuel Romero Tony Lee Nadav Timor Jennifer Ding Claire Schlesinger Hailey Schoelkopf Jana Ebert Tri Dao Mayank Mishra Alexander Gu Jennifer Robinson Carolyn Jane Anderson Brendan Dolan-Gavitt Danish Contractor Siva Reddy Daniel Fried Dzmitry Bahdanau Yacine Jernite Carlos Mu\u00f1oz Ferrandis Sean M. Hughes Thomas Wolf Arjun Guha Leandro von Werra and Harm de Vries. 2023. StarCoder: may the source be with you!. ArXiv abs\/2305.06161 (2023) https:\/\/api.semanticscholar.org\/CorpusID:258588247"},{"key":"e_1_3_2_1_24_1","volume-title":"WizardCoder: Empowering Code Large Language Models with Evol-Instruct. ArXiv, abs\/2306.08568","author":"Luo Ziyang","year":"2023","unstructured":"Ziyang Luo, Can Xu, Pu Zhao, Qingfeng Sun, Xiubo Geng, Wenxiang Hu, Chongyang Tao, Jing Ma, Qingwei Lin, and Daxin Jiang. 2023. WizardCoder: Empowering Code Large Language Models with Evol-Instruct. ArXiv, abs\/2306.08568 (2023), https:\/\/api.semanticscholar.org\/CorpusID:259164815"},{"key":"e_1_3_2_1_25_1","volume-title":"Is Self-Attention Powerful to Learn Code Syntax and Semantics? ArXiv, abs\/2212.10017","author":"Ma Wei","year":"2022","unstructured":"Wei Ma, Mengjie Zhao, Xiaofei Xie, Qiang Hu, Shangqing Liu, Jiexin Zhang, Wenhan Wang, and Yang Liu. 2022. Is Self-Attention Powerful to Learn Code Syntax and Semantics? ArXiv, abs\/2212.10017 (2022), https:\/\/api.semanticscholar.org\/CorpusID:254877330"},{"key":"e_1_3_2_1_26_1","unstructured":"Anthony Moi and Nicolas Patry. 2023. HuggingFace\u2019s Tokenizers. https:\/\/github.com\/huggingface\/tokenizers"},{"key":"e_1_3_2_1_27_1","volume-title":"CodeGen: An Open Large Language Model for Code with Multi-Turn Program Synthesis. In International Conference on Learning Representations. https:\/\/api.semanticscholar.org\/CorpusID:252668917","author":"Nijkamp Erik","year":"2022","unstructured":"Erik Nijkamp, Bo Pang, Hiroaki Hayashi, Lifu Tu, Haiquan Wang, Yingbo Zhou, Silvio Savarese, and Caiming Xiong. 2022. CodeGen: An Open Large Language Model for Code with Multi-Turn Program Synthesis. In International Conference on Learning Representations. https:\/\/api.semanticscholar.org\/CorpusID:252668917"},{"key":"e_1_3_2_1_28_1","unstructured":"OpenAI. 2023. ChatGPT. https:\/\/chat.openai.com\/"},{"key":"e_1_3_2_1_29_1","unstructured":"OpenAI. 2023. GPT-3.5. https:\/\/platform.openai.com\/docs\/models\/gpt-3-5"},{"key":"e_1_3_2_1_31_1","unstructured":"Tim Peters. 2023. PEP 20 \u2013 The Zen of Python. https:\/\/peps.python.org\/pep-0020\/"},{"key":"e_1_3_2_1_32_1","unstructured":"Python. 2023. Full Grammar specification. https:\/\/docs.python.org\/3\/reference\/grammar.html"},{"key":"e_1_3_2_1_33_1","volume-title":"Proceedings of the 29th ACM Joint Meeting on European Software Engineering Conference and Symposium on the Foundations of Software Engineering, https:\/\/api.semanticscholar.org\/CorpusID:235359051","author":"Islam Rabin Md Rafiqul","year":"2021","unstructured":"Md Rafiqul Islam Rabin, Vincent J. Hellendoorn, and Mohammad Amin Alipour. 2021. Understanding neural code intelligence through program simplification. Proceedings of the 29th ACM Joint Meeting on European Software Engineering Conference and Symposium on the Foundations of Software Engineering, https:\/\/api.semanticscholar.org\/CorpusID:235359051"},{"key":"e_1_3_2_1_34_1","volume-title":"Proceedings of the 6th ACM SIGPLAN International Symposium on Machine Programming, https:\/\/api.semanticscholar.org\/CorpusID:249191662","author":"Islam Rabin Md Rafiqul","year":"2022","unstructured":"Md Rafiqul Islam Rabin, Aftab Hussain, and Mohammad Amin Alipour. 2022. Syntax-guided program reduction for understanding neural code intelligence models. Proceedings of the 6th ACM SIGPLAN International Symposium on Machine Programming, https:\/\/api.semanticscholar.org\/CorpusID:249191662"},{"key":"e_1_3_2_1_35_1","unstructured":"Alec Radford Jeff Wu Rewon Child David Luan Dario Amodei and Ilya Sutskever. 2019. Language Models are Unsupervised Multitask Learners. https:\/\/api.semanticscholar.org\/CorpusID:160025533"},{"key":"e_1_3_2_1_36_1","unstructured":"Repilt. 2023. ReplitLM. https:\/\/github.com\/replit\/replitLM"},{"key":"e_1_3_2_1_37_1","volume-title":"Aaron Grattafiori, Wenhan Xiong, Alexandre D\u2019efossez, Jade Copet, Faisal Azhar, Hugo Touvron, Louis Martin, Nicolas Usunier, Thomas Scialom, and Gabriel Synnaeve.","author":"Rozi\u00e8re Baptiste","year":"2023","unstructured":"Baptiste Rozi\u00e8re, Jonas Gehring, Fabian Gloeckle, Sten Sootla, Itai Gat, Xiaoqing Tan, Yossi Adi, Jingyu Liu, Tal Remez, J\u00e9r\u00e9my Rapin, Artyom Kozhevnikov, I. Evtimov, Joanna Bitton, Manish P Bhatt, Cristian Cant\u00f3n Ferrer, Aaron Grattafiori, Wenhan Xiong, Alexandre D\u2019efossez, Jade Copet, Faisal Azhar, Hugo Touvron, Louis Martin, Nicolas Usunier, Thomas Scialom, and Gabriel Synnaeve. 2023. Code Llama: Open Foundation Models for Code. ArXiv, abs\/2308.12950 (2023), https:\/\/api.semanticscholar.org\/CorpusID:261100919"},{"key":"e_1_3_2_1_38_1","volume-title":"Neural Machine Translation of Rare Words with Subword Units. ArXiv, abs\/1508.07909","author":"Sennrich Rico","year":"2015","unstructured":"Rico Sennrich, Barry Haddow, and Alexandra Birch. 2015. Neural Machine Translation of Rare Words with Subword Units. ArXiv, abs\/1508.07909 (2015), https:\/\/api.semanticscholar.org\/CorpusID:1114678"},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1145\/3609437.3609438"},{"key":"e_1_3_2_1_40_1","unstructured":"Significant Gravitas. [n. d.]. AutoGPT. https:\/\/github.com\/Significant-Gravitas\/AutoGPT"},{"key":"e_1_3_2_1_41_1","unstructured":"Hugo Touvron Louis Martin Kevin R. Stone Peter Albert Amjad Almahairi Yasmine Babaei Nikolay Bashlykov Soumya Batra Prajjwal Bhargava Shruti Bhosale Daniel M. Bikel Lukas Blecher Cristian Cant\u00f3n Ferrer Moya Chen Guillem Cucurull David Esiobu Jude Fernandes Jeremy Fu Wenyin Fu Brian Fuller Cynthia Gao Vedanuj Goswami Naman Goyal Anthony S. Hartshorn Saghar Hosseini Rui Hou Hakan Inan Marcin Kardas Viktor Kerkez Madian Khabsa Isabel M. Kloumann A. V. Korenev Punit Singh Koura Marie-Anne Lachaux Thibaut Lavril Jenya Lee Diana Liskovich Yinghai Lu Yuning Mao Xavier Martinet Todor Mihaylov Pushkar Mishra Igor Molybog Yixin Nie Andrew Poulton Jeremy Reizenstein Rashi Rungta Kalyan Saladi Alan Schelten Ruan Silva Eric Michael Smith R. Subramanian Xia Tan Binh Tang Ross Taylor Adina Williams Jian Xiang Kuan Puxin Xu Zhengxu Yan Iliyan Zarov Yuchen Zhang Angela Fan Melanie Kambadur Sharan Narang Aurelien Rodriguez Robert Stojnic Sergey Edunov and Thomas Scialom. 2023. Llama 2: Open Foundation and Fine-Tuned Chat Models. ArXiv abs\/2307.09288 (2023) https:\/\/api.semanticscholar.org\/CorpusID:259950998"},{"key":"e_1_3_2_1_42_1","unstructured":"tree sitter. 2023. tree-sitter\/tree-sitter: An incremental parsing system for programming tools. https:\/\/github.com\/tree-sitter\/tree-sitter"},{"key":"e_1_3_2_1_43_1","volume-title":"Probing Pretrained Models of Source Codes. ArXiv, abs\/2202.08975","author":"Troshin Sergey","year":"2022","unstructured":"Sergey Troshin and Nadezhda Chirkova. 2022. Probing Pretrained Models of Source Codes. ArXiv, abs\/2202.08975 (2022), https:\/\/api.semanticscholar.org\/CorpusID:246996634"},{"key":"e_1_3_2_1_44_1","unstructured":"Guido van Rossum Barry Warsaw and Alyssa Coghlan. 2023. PEP 8 \u2013 Style Guide for Python Code. https:\/\/peps.python.org\/pep-0008\/"},{"key":"e_1_3_2_1_45_1","volume-title":"2022 IEEE\/ACM 44th International Conference on Software Engineering (ICSE), 2377\u20132388","author":"Wan Yao","year":"2022","unstructured":"Yao Wan, Wei Zhao, Hongyu Zhang, Yulei Sui, Guandong Xu, and Hairong Jin. 2022. What Do They Capture? - A Structural Analysis of Pre-Trained Language Models for Source Code. 2022 IEEE\/ACM 44th International Conference on Software Engineering (ICSE), 2377\u20132388. https:\/\/api.semanticscholar.org\/CorpusID:246823289"},{"key":"e_1_3_2_1_46_1","volume-title":"Nghi D. Q. Bui, Junnan Li, and Steven C. H. Hoi.","author":"Wang Yue","year":"2023","unstructured":"Yue Wang, Hung Le, Akhilesh Deepak Gotmare, Nghi D. Q. Bui, Junnan Li, and Steven C. H. Hoi. 2023. CodeT5+: Open Code Large Language Models for Code Understanding and Generation. ArXiv, abs\/2305.07922 (2023), https:\/\/api.semanticscholar.org\/CorpusID:258685677"},{"key":"e_1_3_2_1_47_1","volume-title":"Hoi","author":"Wang Yue","year":"2021","unstructured":"Yue Wang, Weishi Wang, Shafiq R. Joty, and Steven C. H. Hoi. 2021. CodeT5: Identifier-aware Unified Pre-trained Encoder-Decoder Models for Code Understanding and Generation. ArXiv, abs\/2109.00859 (2021), https:\/\/api.semanticscholar.org\/CorpusID:237386541"},{"key":"e_1_3_2_1_48_1","volume-title":"Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations. Association for Computational Linguistics, Online. 38\u201345","author":"Wolf Thomas","year":"2020","unstructured":"Thomas Wolf, Lysandre Debut, Victor Sanh, Julien Chaumond, Clement Delangue, Anthony Moi, Pierric Cistac, Tim Rault, R\u00e9mi Louf, Morgan Funtowicz, Joe Davison, Sam Shleifer, Patrick von Platen, Clara Ma, Yacine Jernite, Julien Plu, Canwen Xu, Teven Le Scao, Sylvain Gugger, Mariama Drame, Quentin Lhoest, and Alexander M. Rush. 2020. Transformers: State-of-the-Art Natural Language Processing. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations. Association for Computational Linguistics, Online. 38\u201345. https:\/\/www.aclweb.org\/anthology\/2020.emnlp-demos.6"},{"key":"e_1_3_2_1_49_1","volume-title":"Prem Devanbu, and David Lo.","author":"Yang Zhou","year":"2024","unstructured":"Zhou Yang, Zhensu Sun, Terry Yue Zhuo, Prem Devanbu, and David Lo. 2024. Robustness, Security, Privacy, Explainability, Efficiency, and Usability of Large Language Models for Code. ArXiv, abs\/2403.07506 (2024), https:\/\/api.semanticscholar.org\/CorpusID:268364103"},{"key":"e_1_3_2_1_50_1","doi-asserted-by":"publisher","DOI":"10.1145\/3533767.3534390"},{"key":"e_1_3_2_1_51_1","doi-asserted-by":"publisher","unstructured":"Peiyuan Zhang Guangtao Zeng Tianduo Wang and Wei Lu. 2024. TinyLlama: An Open-Source Small Language Model. https:\/\/doi.org\/10.48550\/ARXIV.2401.02385 arxiv:2401.02385. 10.48550\/ARXIV.2401.02385","DOI":"10.48550\/ARXIV.2401.02385"},{"key":"e_1_3_2_1_52_1","doi-asserted-by":"publisher","DOI":"10.1145\/3540250.3549094"},{"key":"e_1_3_2_1_53_1","volume-title":"Proceedings of the 29th ACM Joint Meeting on European Software Engineering Conference and Symposium on the Foundations of Software Engineering, https:\/\/api.semanticscholar.org\/CorpusID:227227733","author":"Zheng Yunhui","year":"2020","unstructured":"Yunhui Zheng, Sahil Suneja, Yufan Zhuang, Alessandro Morari, and Jim Laredo. 2020. Probing model signal-awareness via prediction-preserving input minimization. Proceedings of the 29th ACM Joint Meeting on European Software Engineering Conference and Symposium on the Foundations of Software Engineering, https:\/\/api.semanticscholar.org\/CorpusID:227227733"}],"event":{"name":"ISSTA '24: 33rd ACM SIGSOFT International Symposium on Software Testing and Analysis","location":"Vienna Austria","acronym":"ISSTA '24","sponsor":["SIGSOFT ACM Special Interest Group on Software Engineering","AITO"]},"container-title":["Proceedings of the 33rd ACM SIGSOFT International Symposium on Software Testing and Analysis"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3650212.3680347","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3650212.3680347","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T22:50:07Z","timestamp":1750287007000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3650212.3680347"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,9,11]]},"references-count":51,"alternative-id":["10.1145\/3650212.3680347","10.1145\/3650212"],"URL":"https:\/\/doi.org\/10.1145\/3650212.3680347","relation":{},"subject":[],"published":{"date-parts":[[2024,9,11]]},"assertion":[{"value":"2024-09-11","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}