{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T17:09:47Z","timestamp":1776100187148,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":68,"publisher":"ACM","content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,6,23]]},"DOI":"10.1145\/3696630.3728541","type":"proceedings-article","created":{"date-parts":[[2025,7,28]],"date-time":"2025-07-28T19:08:09Z","timestamp":1753729689000},"page":"150-161","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":5,"title":["Natural Language Outlines for Code: Literate Programming in the LLM Era"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-7140-7869","authenticated-orcid":false,"given":"Kensen","family":"Shi","sequence":"first","affiliation":[{"name":"Google, Mountain View, California, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4558-2847","authenticated-orcid":false,"given":"Deniz","family":"Alt\u0131nb\u00fcken","sequence":"additional","affiliation":[{"name":"Google, Mountain View, California, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0000-4812-7199","authenticated-orcid":false,"given":"Saswat","family":"Anand","sequence":"additional","affiliation":[{"name":"Google, Mountain View, California, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5808-8015","authenticated-orcid":false,"given":"Mihai","family":"Christodorescu","sequence":"additional","affiliation":[{"name":"Google, Mountain View, California, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-3745-8888","authenticated-orcid":false,"given":"Katja","family":"Gr\u00fcnwedel","sequence":"additional","affiliation":[{"name":"Google, Munich, Germany"}]},{"ORCID":"https:\/\/orcid.org\/0009-0004-6834-0498","authenticated-orcid":false,"given":"Alexa","family":"Koenings","sequence":"additional","affiliation":[{"name":"Google, Munich, Germany"}]},{"ORCID":"https:\/\/orcid.org\/0009-0003-9741-0830","authenticated-orcid":false,"given":"Sai","family":"Naidu","sequence":"additional","affiliation":[{"name":"Google, New York, New York, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-7606-1656","authenticated-orcid":false,"given":"Anurag","family":"Pathak","sequence":"additional","affiliation":[{"name":"Google, Munich, Germany"}]},{"ORCID":"https:\/\/orcid.org\/0009-0007-3929-8843","authenticated-orcid":false,"given":"Marc","family":"Rasi","sequence":"additional","affiliation":[{"name":"Google, Mountain View, California, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0001-1771-104X","authenticated-orcid":false,"given":"Fredde","family":"Ribeiro","sequence":"additional","affiliation":[{"name":"Google, Munich, Germany"}]},{"ORCID":"https:\/\/orcid.org\/0009-0000-3885-3429","authenticated-orcid":false,"given":"Brandon","family":"Ruffin","sequence":"additional","affiliation":[{"name":"Google, San Francisco, California, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0008-6784-4757","authenticated-orcid":false,"given":"Siddhant","family":"Sanyam","sequence":"additional","affiliation":[{"name":"Google, New York, New York, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0001-5710-2302","authenticated-orcid":false,"given":"Maxim","family":"Tabachnyk","sequence":"additional","affiliation":[{"name":"Google, Munich, Germany"}]},{"ORCID":"https:\/\/orcid.org\/0009-0000-6652-0867","authenticated-orcid":false,"given":"Sara","family":"Toth","sequence":"additional","affiliation":[{"name":"Google, Munich, Germany"}]},{"ORCID":"https:\/\/orcid.org\/0009-0000-4761-6644","authenticated-orcid":false,"given":"Roy","family":"Tu","sequence":"additional","affiliation":[{"name":"Google, Kirkland, Washington, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0006-6726-837X","authenticated-orcid":false,"given":"Tobias","family":"Welp","sequence":"additional","affiliation":[{"name":"Google, Munich, Germany"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2739-1032","authenticated-orcid":false,"given":"Pengcheng","family":"Yin","sequence":"additional","affiliation":[{"name":"Google, Mountain View, California, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0009-2708-691X","authenticated-orcid":false,"given":"Manzil","family":"Zaheer","sequence":"additional","affiliation":[{"name":"Google, Mountain View, California, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2546-9000","authenticated-orcid":false,"given":"Satish","family":"Chandra","sequence":"additional","affiliation":[{"name":"Google, Sunnyvale, California, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0041-3820","authenticated-orcid":false,"given":"Charles","family":"Sutton","sequence":"additional","affiliation":[{"name":"Google, Mountain View, California, USA"}]}],"member":"320","published-online":{"date-parts":[[2025,7,28]]},"reference":[{"key":"e_1_3_2_2_1_1","unstructured":"[n. d.]. Android App Security and Obfuscation | DexGuard. https:\/\/www.guardsquare.com\/dexguard."},{"key":"e_1_3_2_2_2_1","unstructured":"[n. d.]. Code Search. https:\/\/developers.google.com\/code-search."},{"key":"e_1_3_2_2_3_1","unstructured":"[n. d.]. GitHub Copilot - Your AI pair programmer. https:\/\/github.com\/features\/copilot."},{"key":"e_1_3_2_2_4_1","unstructured":"[n. d.]. Java Obfuscator and Android App Optimizer | ProGuard. https:\/\/www.guardsquare.com\/proguard."},{"key":"e_1_3_2_2_5_1","unstructured":"2024. Google Python Style Guide. https:\/\/google.github.io\/styleguide\/pyguide.html."},{"key":"e_1_3_2_2_6_1","volume-title":"Three Ways to Prime Students for Learning","author":"Addison Stephen","unstructured":"Stephen Addison. 2022. Three Ways to Prime Students for Learning. Faculty Focus. https:\/\/www.facultyfocus.com\/articles\/course-design-ideas\/three-ways-to-prime-students-for-learning\/"},{"key":"e_1_3_2_2_7_1","volume-title":"A Transformer-based Approach for Source Code Summarization","author":"Ahmad Wasi","unstructured":"Wasi Ahmad, Saikat Chakraborty, Baishakhi Ray, and Kai-Wei Chang. 2020. A Transformer-based Approach for Source Code Summarization. In Association for Computational Linguistics (ACL)."},{"key":"e_1_3_2_2_8_1","volume-title":"International Conference on Automated Software Engineering (ASE).","author":"Ahmed Toufique","year":"2022","unstructured":"Toufique Ahmed and Premkumar Devanbu. 2022. Few-shot training LLMs for project-specific code-summarization. In International Conference on Automated Software Engineering (ASE)."},{"key":"e_1_3_2_2_9_1","volume-title":"International Conference on Software Engineering (ICSE).","author":"Ahmed Toufique","year":"2024","unstructured":"Toufique Ahmed, Kunal Suresh Pai, Premkumar Devanbu, and Earl Barr. 2024. Automatic semantic augmentation of language model prompts (for code summarization). In International Conference on Software Engineering (ICSE)."},{"key":"e_1_3_2_2_10_1","volume-title":"Unsupervised Evaluation of Code LLMs with Round-Trip Correctness. In International Conference on Machine Learning (ICML).","author":"Allamanis Miltiadis","year":"2024","unstructured":"Miltiadis Allamanis, Sheena Panthaplackel, and Pengcheng Yin. 2024. Unsupervised Evaluation of Code LLMs with Round-Trip Correctness. In International Conference on Machine Learning (ICML)."},{"key":"e_1_3_2_2_11_1","unstructured":"Shengnan An Bo Zhou Zeqi Lin Qiang Fu Bei Chen Nanning Zheng Weizhu Chen and Jian-Guang Lou. 2023. Skill-Based Few-Shot Selection for In-Context Learning. In Empirical Methods in Natural Language Processing (EMNLP)."},{"key":"e_1_3_2_2_12_1","volume-title":"Spellburst: A node-based interface for exploratory creative coding with natural language prompts. In User Interface Software and Technology (UIST).","author":"Angert Tyler","year":"2023","unstructured":"Tyler Angert, Miroslav Suzara, Jenny Han, Christopher Pondoc, and Hariharan Subramonyam. 2023. Spellburst: A node-based interface for exploratory creative coding with natural language prompts. In User Interface Software and Technology (UIST)."},{"key":"e_1_3_2_2_13_1","volume-title":"International Conference on Software Engineering (ICSE).","author":"Bacchelli Alberto","year":"2013","unstructured":"Alberto Bacchelli and Christian Bird. 2013. Expectations, outcomes, and challenges of modern code review. In International Conference on Software Engineering (ICSE)."},{"key":"e_1_3_2_2_14_1","volume-title":"Assessing the Impact of File Ordering Strategies on Code Review Process. In International Conference on Evaluation and Assessment in Software Engineering (EASE).","author":"Bagirov Farid","year":"2023","unstructured":"Farid Bagirov, Pouria Derakhshanfar, Alexey Kalina, Elena Kartysheva, and Vladimir Kovalenko. 2023. Assessing the Impact of File Ordering Strategies on Code Review Process. In International Conference on Evaluation and Assessment in Software Engineering (EASE)."},{"key":"e_1_3_2_2_15_1","unstructured":"Yuntao Bai Andy Jones Kamal Ndousse Amanda Askell Anna Chen Nova DasSarma Dawn Drain Stanislav Fort Deep Ganguli Tom Henighan et al. 2022. Training a helpful and harmless assistant with reinforcement learning from human feedback. arXiv preprint arXiv:2204.05862 (2022)."},{"key":"e_1_3_2_2_16_1","volume-title":"Efficient training of language models to fill in the middle. arXiv preprint arXiv:2207.14255","author":"Bavarian Mohammad","year":"2022","unstructured":"Mohammad Bavarian, Heewoo Jun, Nikolas Tezak, John Schulman, Christine McLeavey, Jerry Tworek, and Mark Chen. 2022. Efficient training of language models to fill in the middle. arXiv preprint arXiv:2207.14255 (2022)."},{"key":"e_1_3_2_2_17_1","article-title":"Why my code summarization model does not work: Code comment improvement with category prediction","volume":"30","author":"Chen Qiuyuan","year":"2021","unstructured":"Qiuyuan Chen, Xin Xia, Han Hu, David Lo, and Shanping Li. 2021. Why my code summarization model does not work: Code comment improvement with category prediction. Transactions on Software Engineering and Methodology (TOSEM) 30, 2 (2021).","journal-title":"Transactions on Software Engineering and Methodology (TOSEM)"},{"key":"e_1_3_2_2_18_1","article-title":"Prompt Sapper: a LLM-empowered production tool for building AI chains","volume":"33","author":"Cheng Yu","year":"2024","unstructured":"Yu Cheng, Jieshan Chen, Qing Huang, Zhenchang Xing, Xiwei Xu, and Qinghua Lu. 2024. Prompt Sapper: a LLM-empowered production tool for building AI chains. Transactions on Software Engineering and Methodology (TOSEM) 33, 5 (2024).","journal-title":"Transactions on Software Engineering and Methodology (TOSEM)"},{"key":"e_1_3_2_2_19_1","unstructured":"Paul F Christiano Jan Leike Tom Brown Miljan Martic Shane Legg and Dario Amodei. 2017. Deep reinforcement learning from human preferences. In Advances in Neural Information Processing Systems (NeurIPS)."},{"key":"e_1_3_2_2_20_1","volume-title":"Introduction to Algorithms","author":"Cormen Thomas H.","unstructured":"Thomas H. Cormen, Charles E. Leiserson, Ronald L. Rivest, and Clifford Stein. 2009. Introduction to Algorithms, 3rd Edition. MIT Press.","edition":"3"},{"key":"e_1_3_2_2_21_1","volume-title":"Enabling Language Models to Fill in the Blanks","author":"Donahue Chris","unstructured":"Chris Donahue, Mina Lee, and Percy Liang. 2020. Enabling Language Models to Fill in the Blanks. In Association for Computational Linguistics (ACL)."},{"key":"e_1_3_2_2_22_1","volume-title":"International Conference on AI-Powered Software (AIware.","author":"Dvivedi Shubhang Shekhar","year":"2024","unstructured":"Shubhang Shekhar Dvivedi, Vyshnav Vijay, Sai Leela Rahul Pujari, Shoumik Lodh, and Dhruv Kumar. 2024. A comparative analysis of large language models for code documentation generation. In International Conference on AI-Powered Software (AIware."},{"key":"e_1_3_2_2_23_1","doi-asserted-by":"crossref","unstructured":"Ernst N.A. and Robillard M.P. 2023. A study of documentation for software architecture. Empirical Software Engineering 28 122 (2023).","DOI":"10.1007\/s10664-023-10347-2"},{"key":"e_1_3_2_2_24_1","volume-title":"International Conference on Software Engineering: Future of Software Engineering (ICSE-FoSE).","author":"Fan Angela","year":"2023","unstructured":"Angela Fan, Beliz Gokkaya, Mark Harman, Mitya Lyubarskiy, Shubho Sengupta, Shin Yoo, and Jie M Zhang. 2023. Large language models for software engineering: Survey and open problems. In International Conference on Software Engineering: Future of Software Engineering (ICSE-FoSE)."},{"key":"e_1_3_2_2_25_1","doi-asserted-by":"crossref","first-page":"1095","DOI":"10.1109\/TSE.2017.2664836","article-title":"Autofolding for source code summarization","volume":"43","author":"Fowkes Jaroslav","year":"2017","unstructured":"Jaroslav Fowkes, Pankajan Chanthirasegaran, Razvan Ranca, Miltiadis Allamanis, Mirella Lapata, and Charles Sutton. 2017. Autofolding for source code summarization. Transactions on Software Engineering (TSE) 43, 12 (2017), 1095\u20131109.","journal-title":"Transactions on Software Engineering (TSE)"},{"key":"e_1_3_2_2_26_1","volume-title":"Making Pre-trained Language Models Better Few-shot Learners. In Association for Computational Linguistics and International Joint Conference on Natural Language Processing (ACL-IJCNLP).","author":"Gao Tianyu","year":"2021","unstructured":"Tianyu Gao, Adam Fisch, and Danqi Chen. 2021. Making Pre-trained Language Models Better Few-shot Learners. In Association for Computational Linguistics and International Joint Conference on Natural Language Processing (ACL-IJCNLP)."},{"key":"e_1_3_2_2_27_1","volume-title":"Retrieval-augmented generation for large language models: A survey. arXiv preprint arXiv:2312.10997","author":"Gao Yunfan","year":"2023","unstructured":"Yunfan Gao, Yun Xiong, Xinyu Gao, Kangxiang Jia, Jinliu Pan, Yuxi Bi, Yi Dai, Jiawei Sun, and Haofen Wang. 2023. Retrieval-augmented generation for large language models: A survey. arXiv preprint arXiv:2312.10997 (2023)."},{"key":"e_1_3_2_2_28_1","volume-title":"Gemini: A Family of Highly Capable Multimodal Models. arXiv preprint arXiv:2312.11805","author":"Team Gemini","year":"2023","unstructured":"Gemini Team. 2023. Gemini: A Family of Highly Capable Multimodal Models. arXiv preprint arXiv:2312.11805 (2023)."},{"key":"e_1_3_2_2_29_1","volume-title":"Unlocking multimodal understanding across millions of tokens of context. arXiv preprint arXiv:2403.05530","author":"Team Gemini","year":"2024","unstructured":"Gemini Team. 2024. Gemini 1.5: Unlocking multimodal understanding across millions of tokens of context. arXiv preprint arXiv:2403.05530 (2024)."},{"key":"e_1_3_2_2_30_1","volume-title":"International Conference on Software Engineering (ICSE).","author":"Geng Mingyang","year":"2024","unstructured":"Mingyang Geng, Shangwen Wang, Dezun Dong, Haotian Wang, Ge Li, Zhi Jin, Xiaoguang Mao, and Xiangke Liao. 2024. Large language models are few-shot summarizers: Multi-intent comment generation via in-context learning. In International Conference on Software Engineering (ICSE)."},{"key":"e_1_3_2_2_31_1","unstructured":"David Goodger and Guido van Rossum. 2001. PEP 257: Docstring Conventions. https:\/\/peps.python.org\/pep-0257\/."},{"key":"e_1_3_2_2_32_1","volume-title":"DeepSeek-Coder: When the Large Language Model Meets Programming - The Rise of Code Intelligence. arXiv preprint arXiv:2401.14196","author":"Guo Daya","year":"2024","unstructured":"Daya Guo, Qihao Zhu, Dejian Yang, Zhenda Xie, Kai Dong, Wentao Zhang, Guanting Chen, Xiao Bi, Y. Wu, Y. K. Li, Fuli Luo, Yingfei Xiong, and Wenfeng Liang. 2024. DeepSeek-Coder: When the Large Language Model Meets Programming - The Rise of Code Intelligence. arXiv preprint arXiv:2401.14196 (2024)."},{"key":"e_1_3_2_2_33_1","volume-title":"Large language models for software engineering: A systematic literature review. arXiv preprint arXiv:2308.10620","author":"Hou Xinyi","year":"2023","unstructured":"Xinyi Hou, Yanjie Zhao, Yue Liu, Zhou Yang, Kailong Wang, Li Li, Xiapu Luo, David Lo, John Grundy, and Haoyu Wang. 2023. Large language models for software engineering: A systematic literature review. arXiv preprint arXiv:2308.10620 (2023)."},{"key":"e_1_3_2_2_34_1","volume-title":"International Conference on Software Engineering (ICSE).","author":"Hu Xing","year":"2022","unstructured":"Xing Hu, Xin Xia, David Lo, Zhiyuan Wan, Qiuyuan Chen, and Thomas Zimmermann. 2022. Practitioners' expectations on automated code comment generation. In International Conference on Software Engineering (ICSE)."},{"key":"e_1_3_2_2_35_1","volume-title":"A comprehensive survey on process-oriented automatic text summarization with exploration of LLM-based methods. arXiv preprint arXiv:2403.02901","author":"Jin Hanlei","year":"2024","unstructured":"Hanlei Jin, Yang Zhang, Dan Meng, Jun Wang, and Jinghua Tan. 2024. A comprehensive survey on process-oriented automatic text summarization with exploration of LLM-based methods. arXiv preprint arXiv:2403.02901 (2024)."},{"key":"e_1_3_2_2_36_1","volume-title":"Extended Abstracts of the CHI Conference on Human Factors in Computing Systems.","author":"Kahng Minsuk","year":"2024","unstructured":"Minsuk Kahng, Ian Tenney, Mahima Pushkarna, Michael Xieyang Liu, James Wexler, Emily Reif, Krystal Kallarackal, Minsuk Chang, Michael Terry, and Lucas Dixon. 2024. LLM Comparator: Visual analytics for side-by-side evaluation of large language models. In Extended Abstracts of the CHI Conference on Human Factors in Computing Systems."},{"key":"e_1_3_2_2_37_1","volume-title":"A survey of reinforcement learning from human feedback. arXiv preprint arXiv:2312.14925","author":"Kaufmann Timo","year":"2023","unstructured":"Timo Kaufmann, Paul Weng, Viktor Bengs, and Eyke H\u00fcllermeier. 2023. A survey of reinforcement learning from human feedback. arXiv preprint arXiv:2312.14925 (2023)."},{"key":"e_1_3_2_2_38_1","doi-asserted-by":"crossref","first-page":"97","DOI":"10.1093\/comjnl\/27.2.97","article-title":"Literate programming","volume":"27","author":"Knuth Donald","year":"1984","unstructured":"Donald Knuth. 1984. Literate programming. Comput. J. 27, 2 (1984), 97\u2013111.","journal-title":"Comput. J."},{"key":"e_1_3_2_2_39_1","volume-title":"Conference on Language Modeling (COLM).","author":"Koo Terry","year":"2024","unstructured":"Terry Koo, Frederick Liu, and Luheng He. 2024. Automata-based constraints for language model decoding. In Conference on Language Modeling (COLM)."},{"key":"e_1_3_2_2_40_1","volume-title":"The WAC Clearinghouse","author":"LeCourt Donna","unstructured":"Donna LeCourt, Kate Kiefer, Luann Barnes, Mike Palmquist, and Tom Siller. 2024. Abstracts. The WAC Clearinghouse, Colorado State University. https:\/\/wac.colostate.edu\/repository\/writing\/guides\/abstracts\/"},{"key":"e_1_3_2_2_41_1","unstructured":"Patrick Lewis Ethan Perez Aleksandra Piktus Fabio Petroni Vladimir Karpukhin Naman Goyal Heinrich K\u00fcttler Mike Lewis Wen-tau Yih Tim Rockt\u00e4schel Sebastian Riedel and Douwe Kiela. 2020. Retrieval-augmented generation for knowledge-intensive NLP tasks. In Advances in Neural Information Processing Systems (NeurIPS)."},{"key":"e_1_3_2_2_42_1","volume-title":"Enhancing LLM-Based Coding Tools through Native Integration of IDE-Derived Static Context. arXiv preprint arXiv:2402.03630","author":"Li Yichen","year":"2024","unstructured":"Yichen Li, Yun Peng, Yintong Huo, and Michael R Lyu. 2024. Enhancing LLM-Based Coding Tools through Native Integration of IDE-Derived Static Context. arXiv preprint arXiv:2402.03630 (2024)."},{"key":"e_1_3_2_2_43_1","volume-title":"Conference on Human Factors in Computing Systems (CHI).","author":"Liu Michael Xieyang","year":"2023","unstructured":"Michael Xieyang Liu, Advait Sarkar, Carina Negreanu, Benjamin Zorn, Jack Williams, Neil Toronto, and Andrew D Gordon. 2023. \"What it wants me to say\": Bridging the abstraction gap between end-user programmers and code-generating large language models. In Conference on Human Factors in Computing Systems (CHI)."},{"key":"e_1_3_2_2_44_1","volume-title":"Technical Symposium on Computer Science Education (SIGCSE TS).","author":"MacNeil Stephen","year":"2023","unstructured":"Stephen MacNeil, Andrew Tran, Arto Hellas, Joanne Kim, Sami Sarsa, Paul Denny, Seth Bernstein, and Juho Leinonen. 2023. Experiences from using code explanations generated by large language models in a web software development e-book. In Technical Symposium on Computer Science Education (SIGCSE TS)."},{"key":"e_1_3_2_2_45_1","volume-title":"International Conference on Evaluation and Assessment in Software Engineering (EASE).","author":"Manfredi Gilda","year":"2023","unstructured":"Gilda Manfredi, Ugo Erra, and Gabriele Gilio. 2023. A mixed reality approach for innovative pair programming education with a conversational AI virtual avatar. In International Conference on Evaluation and Assessment in Software Engineering (EASE)."},{"key":"e_1_3_2_2_46_1","volume-title":"International Conference on Program Comprehension (ICPC).","author":"Minelli Roberto","year":"2015","unstructured":"Roberto Minelli, Andrea Mocci, and Michele Lanza. 2015. I know what you did last summer - an investigation of how developers spend their time. In International Conference on Program Comprehension (ICPC)."},{"key":"e_1_3_2_2_47_1","volume-title":"International Conference on Software Engineering (ICSE).","author":"Mu Fangwen","year":"2023","unstructured":"Fangwen Mu, Xiao Chen, Lin Shi, Song Wang, and Qing Wang. 2023. Developer-intent driven code comment generation. In International Conference on Software Engineering (ICSE)."},{"key":"e_1_3_2_2_48_1","volume-title":"International Conference on Software Engineering (ICSE).","author":"Nam Daye","year":"2024","unstructured":"Daye Nam, Andrew Macvean, Vincent Hellendoorn, Bogdan Vasilescu, and Brad Myers. 2024. Using an LLM to help with code understanding. In International Conference on Software Engineering (ICSE)."},{"key":"e_1_3_2_2_49_1","volume-title":"International Conference on Software Engineering (ICSE).","author":"Nashid Noor","year":"2023","unstructured":"Noor Nashid, Mifta Sintaha, and Ali Mesbah. 2023. Retrieval-based prompt selection for code-related few-shot learning. In International Conference on Software Engineering (ICSE)."},{"key":"e_1_3_2_2_50_1","volume-title":"GPT-4 technical report. arXiv preprint arXiv:2303.08774","author":"AI.","year":"2023","unstructured":"OpenAI. 2023. GPT-4 technical report. arXiv preprint arXiv:2303.08774 (2023)."},{"key":"e_1_3_2_2_51_1","volume-title":"Junyi Jessy Li, and Raymond Mooney","author":"Panthaplackel Sheena","year":"2020","unstructured":"Sheena Panthaplackel, Pengyu Nie, Milos Gligoric, Junyi Jessy Li, and Raymond Mooney. 2020. Learning to Update Natural Language Comments Based on Code Changes. In Association for Computational Linguistics (ACL)."},{"key":"e_1_3_2_2_52_1","volume-title":"Classifying Code Comments in Java Open-Source Software Systems. In International Conference on Mining Software Repositories (MSR).","author":"Pascarella Luca","year":"2017","unstructured":"Luca Pascarella and Alberto Bacchelli. 2017. Classifying Code Comments in Java Open-Source Software Systems. In International Conference on Mining Software Repositories (MSR)."},{"key":"e_1_3_2_2_53_1","volume-title":"Together We Go Further: LLMs and IDE Static Analysis for Extract Method Refactoring. arXiv preprint arXiv:2401.15298","author":"Pomian Dorin","year":"2024","unstructured":"Dorin Pomian, Abhiram Bellur, Malinda Dilhara, Zarina Kurbatova, Egor Bogomolov, Timofey Bryksin, and Danny Dig. 2024. Together We Go Further: LLMs and IDE Static Analysis for Extract Method Refactoring. arXiv preprint arXiv:2401.15298 (2024)."},{"key":"e_1_3_2_2_54_1","volume-title":"Code Health: To Comment or Not to Comment? https:\/\/testing.googleblog.com\/2017\/07\/code-health-to-comment-or-not-to-comment.html.","author":"Reuveni Dori","year":"2017","unstructured":"Dori Reuveni and Kevin Bourrillion. 2017. Code Health: To Comment or Not to Comment? https:\/\/testing.googleblog.com\/2017\/07\/code-health-to-comment-or-not-to-comment.html."},{"key":"e_1_3_2_2_55_1","volume-title":"International Conference on Intelligent User Interfaces (IUI).","author":"Ross Steven I","year":"2023","unstructured":"Steven I Ross, Fernando Martinez, Stephanie Houde, Michael Muller, and Justin D Weisz. 2023. The programmer's assistant: Conversational interaction with a large language model for software development. In International Conference on Intelligent User Interfaces (IUI)."},{"key":"e_1_3_2_2_56_1","volume-title":"International Conference on Software Engineering: Software Engineering in Practice (ICSE SEIP).","author":"Sadowski Caitlin","year":"2018","unstructured":"Caitlin Sadowski, Emma S\u00f6derberg, Luke Church, Michal Sipko, and Alberto Bacchelli. 2018. Modern code review: a case study at Google. In International Conference on Software Engineering: Software Engineering in Practice (ICSE SEIP)."},{"key":"e_1_3_2_2_57_1","doi-asserted-by":"crossref","unstructured":"Caitlin Sadowski Kathryn T Stolee and Sebastian Elbaum. 2015. How developers search for code: a case study. In Foundations of Software Engineering (FSE).","DOI":"10.1145\/2786805.2786855"},{"key":"e_1_3_2_2_58_1","volume-title":"International Conference on Automated Software Engineering (ASE).","author":"Scalabrino Simone","year":"2017","unstructured":"Simone Scalabrino, Gabriele Bavota, Christopher Vendome, Mario Linares-V\u00e1squez, Denys Poshyvanyk, and Rocco Oliveto. 2017. Automatically assessing code understandability: How far are we?. In International Conference on Automated Software Engineering (ASE)."},{"key":"e_1_3_2_2_59_1","volume-title":"In-IDE Human-AI Experience in the Era of Large Language Models","author":"Sergeyuk Agnia","year":"2024","unstructured":"Agnia Sergeyuk, Sergey Titov, and Maliheh Izadi. 2024. In-IDE Human-AI Experience in the Era of Large Language Models; A Literature Review. arXiv preprint arXiv:2401.10739 (2024)."},{"key":"e_1_3_2_2_60_1","volume-title":"Source Code Summarization in the Era of Large Language Models. arXiv preprint arXiv:2407.07959","author":"Sun Weisong","year":"2024","unstructured":"Weisong Sun, Yun Miao, Yuekang Li, Hongyu Zhang, Chunrong Fang, Yi Liu, Gelei Deng, Yang Liu, and Zhenyu Chen. 2024. Source Code Summarization in the Era of Large Language Models. arXiv preprint arXiv:2407.07959 (2024)."},{"key":"e_1_3_2_2_61_1","volume-title":"Man Fai Wong, and Ching Nam Hang","author":"Tan Chee Wei","year":"2023","unstructured":"Chee Wei Tan, Shangxin Guo, Man Fai Wong, and Ching Nam Hang. 2023. Copilot for Xcode: Exploring AI-Assisted Programming by Prompting Cloud-based Large Language Models. arXiv preprint arXiv:2307.14349 (2023)."},{"key":"e_1_3_2_2_62_1","volume-title":"International Conference on Learning Representations (ICLR).","author":"Tay Yi","year":"2023","unstructured":"Yi Tay, Mostafa Dehghani, Vinh Q. Tran, Xavier Garcia, Jason Wei, Xuezhi Wang, Hyung Won Chung, Dara Bahri, Tal Schuster, Steven Zheng, Denny Zhou, Neil Houlsby, and Donald Metzler. 2023. UL2: Unifying Language Learning Paradigms. In International Conference on Learning Representations (ICLR)."},{"key":"e_1_3_2_2_63_1","volume-title":"Denny Zhou, et al.","author":"Wei Jason","year":"2022","unstructured":"Jason Wei, Xuezhi Wang, Dale Schuurmans, Maarten Bosma, Fei Xia, Ed Chi, Quoc V Le, Denny Zhou, et al. 2022. Chain-of-thought prompting elicits reasoning in large language models. In Advances in Neural Information Processing Systems (NeurIPS)."},{"key":"e_1_3_2_2_64_1","volume-title":"International Conference on Program Comprehension (ICPC).","author":"Wen Fengcai","year":"2019","unstructured":"Fengcai Wen, Csaba Nagy, Gabriele Bavota, and Michele Lanza. 2019. A large-scale empirical study on code-comment inconsistencies. In International Conference on Program Comprehension (ICPC)."},{"key":"e_1_3_2_2_65_1","volume-title":"Efficient guided generation for large language models. arXiv preprint arXiv:2307.09702","author":"Willard Brandon T","year":"2023","unstructured":"Brandon T Willard and R\u00e9mi Louf. 2023. Efficient guided generation for large language models. arXiv preprint arXiv:2307.09702 (2023)."},{"key":"e_1_3_2_2_66_1","doi-asserted-by":"crossref","first-page":"471","DOI":"10.3390\/sym14030471","article-title":"A survey of automatic source code summarization","volume":"14","author":"Zhang Chunyan","year":"2022","unstructured":"Chunyan Zhang, Junchao Wang, Qinglei Zhou, Ting Xu, Ke Tang, Hairen Gui, and Fudong Liu. 2022. A survey of automatic source code summarization. Symmetry 14, 3 (2022), 471.","journal-title":"Symmetry"},{"key":"e_1_3_2_2_67_1","volume-title":"Unifying the perspectives of NLP and software engineering: A survey on language models for code. arXiv preprint arXiv:2311.07989","author":"Zhang Ziyin","year":"2023","unstructured":"Ziyin Zhang, Chaoyu Chen, Bingchang Liu, Cong Liao, Zi Gong, Hang Yu, Jianguo Li, and Rui Wang. 2023. Unifying the perspectives of NLP and software engineering: A survey on language models for code. arXiv preprint arXiv:2311.07989 (2023)."},{"key":"e_1_3_2_2_68_1","volume-title":"Automatic code summarization: A systematic literature review. arXiv preprint arXiv:1909.04352","author":"Zhu Yuxiang","year":"2019","unstructured":"Yuxiang Zhu and Minxue Pan. 2019. Automatic code summarization: A systematic literature review. arXiv preprint arXiv:1909.04352 (2019)."}],"event":{"name":"FSE Companion '25: 33rd ACM International Conference on the Foundations of Software Engineering","location":"Clarion Hotel Trondheim Trondheim Norway","acronym":"FSE Companion '25","sponsor":["SIGSOFT ACM Special Interest Group on Software Engineering"]},"container-title":["Proceedings of the 33rd ACM International Conference on the Foundations of Software Engineering"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3696630.3728541","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,7,28]],"date-time":"2025-07-28T19:16:01Z","timestamp":1753730161000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3696630.3728541"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,6,23]]},"references-count":68,"alternative-id":["10.1145\/3696630.3728541","10.1145\/3696630"],"URL":"https:\/\/doi.org\/10.1145\/3696630.3728541","relation":{},"subject":[],"published":{"date-parts":[[2025,6,23]]},"assertion":[{"value":"2025-07-28","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}