{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,11]],"date-time":"2026-04-11T02:12:04Z","timestamp":1775873524853,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":59,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,11,30]],"date-time":"2023-11-30T00:00:00Z","timestamp":1701302400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/100000001","name":"NSF (National Science Foundation)","doi-asserted-by":"publisher","award":["CCF-2107291, IIS-2107524, IIS-2145479, CCF-2217696, CCF-2313027"],"award-info":[{"award-number":["CCF-2107291, IIS-2107524, IIS-2145479, CCF-2217696, CCF-2313027"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,11,30]]},"DOI":"10.1145\/3611643.3616350","type":"proceedings-article","created":{"date-parts":[[2023,11,30]],"date-time":"2023-11-30T23:14:38Z","timestamp":1701386078000},"page":"695-707","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":30,"title":["Multilingual Code Co-evolution using Large Language Models"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-8211-3321","authenticated-orcid":false,"given":"Jiyang","family":"Zhang","sequence":"first","affiliation":[{"name":"University of Texas at Austin, Austin, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1529-3216","authenticated-orcid":false,"given":"Pengyu","family":"Nie","sequence":"additional","affiliation":[{"name":"University of Texas at Austin, Austin, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2550-5262","authenticated-orcid":false,"given":"Junyi Jessy","family":"Li","sequence":"additional","affiliation":[{"name":"University of Texas at Austin, Austin, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5894-7649","authenticated-orcid":false,"given":"Milos","family":"Gligoric","sequence":"additional","affiliation":[{"name":"University of Texas at Austin, Austin, USA"}]}],"member":"320","published-online":{"date-parts":[[2023,11,30]]},"reference":[{"key":"e_1_3_2_2_1_1","volume-title":"Unified Pre-training for Program Understanding and Generation. In Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. 2655\u20132668","author":"Ahmad Wasi","year":"2021","unstructured":"Wasi Ahmad, Saikat Chakraborty, Baishakhi Ray, and Kai-Wei Chang. 2021. Unified Pre-training for Program Understanding and Generation. In Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. 2655\u20132668."},{"key":"e_1_3_2_2_2_1","doi-asserted-by":"crossref","unstructured":"Toufique Ahmed and Premkumar Devanbu. 2022. Few-Shot Training LLMs for Project-Specific Code-Summarization. In Automated Software Engineering. 1\u20135.","DOI":"10.1145\/3551349.3559555"},{"key":"e_1_3_2_2_3_1","unstructured":"Christian Mauceri Alexandre FAU. 2013. Java2csharp. http:\/\/sourceforge.net\/projects\/j2cstranslator\/"},{"key":"e_1_3_2_2_4_1","unstructured":"Jacob Austin Augustus Odena Maxwell Nye Maarten Bosma Henryk Michalewski David Dohan Ellen Jiang Carrie Cai Michael Terry and Quoc Le. 2021. Program synthesis with large language models. arXiv preprint arXiv:2108.07732."},{"key":"e_1_3_2_2_5_1","volume-title":"Longformer: The long-document transformer. arXiv preprint arXiv:2004.05150.","author":"Beltagy Iz","year":"2020","unstructured":"Iz Beltagy, Matthew E Peters, and Arman Cohan. 2020. Longformer: The long-document transformer. arXiv preprint arXiv:2004.05150."},{"key":"e_1_3_2_2_6_1","volume-title":"An Empirical Investigation of Statistical Significance in NLP. In Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning. 995\u20131005","author":"Berg-Kirkpatrick Taylor","year":"2012","unstructured":"Taylor Berg-Kirkpatrick, David Burkett, and Dan Klein. 2012. An Empirical Investigation of Statistical Significance in NLP. In Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning. 995\u20131005."},{"key":"e_1_3_2_2_7_1","volume-title":"Unlimiformer: Long-range transformers with unlimited length input. arXiv preprint arXiv:2305.01625.","author":"Bertsch Amanda","year":"2023","unstructured":"Amanda Bertsch, Uri Alon, Graham Neubig, and Matthew R Gormley. 2023. Unlimiformer: Long-range transformers with unlimited length input. arXiv preprint arXiv:2305.01625."},{"key":"e_1_3_2_2_8_1","volume-title":"Language models are few-shot learners. Advances in neural information processing systems, 33","author":"Brown Tom","year":"2020","unstructured":"Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared D Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, and Amanda Askell. 2020. Language models are few-shot learners. Advances in neural information processing systems, 33 (2020), 1877\u20131901."},{"key":"e_1_3_2_2_9_1","volume-title":"Hierarchical Learning of Cross-Language Mappings Through Distributed Vector Representations for Code. In International Conference on Software Engineering, NIER. 33\u201336","author":"Bui Nghi DQ","year":"2018","unstructured":"Nghi DQ Bui and Lingxiao Jiang. 2018. Hierarchical Learning of Cross-Language Mappings Through Distributed Vector Representations for Code. In International Conference on Software Engineering, NIER. 33\u201336."},{"key":"e_1_3_2_2_10_1","first-page":"1385","article-title":"Codit: Code Editing with Tree-based Neural Models","author":"Chakraborty Saikat","year":"2020","unstructured":"Saikat Chakraborty, Yangruibo Ding, Miltiadis Allamanis, and Baishakhi Ray. 2020. Codit: Code Editing with Tree-based Neural Models. Transactions on Software Engineering, 1385\u20131399.","journal-title":"Transactions on Software Engineering"},{"key":"e_1_3_2_2_11_1","doi-asserted-by":"crossref","unstructured":"Saikat Chakraborty and Baishakhi Ray. 2021. On Multi-Modal Learning of Editing Source Code. In Automated Software Engineering. 443\u2013455.","DOI":"10.1109\/ASE51524.2021.9678559"},{"key":"e_1_3_2_2_12_1","volume-title":"Jared Kaplan, Harri Edwards, Yuri Burda, Nicholas Joseph, and Greg Brockman.","author":"Chen Mark","year":"2021","unstructured":"Mark Chen, Jerry Tworek, Heewoo Jun, Qiming Yuan, Henrique Ponde de Oliveira Pinto, Jared Kaplan, Harri Edwards, Yuri Burda, Nicholas Joseph, and Greg Brockman. 2021. Evaluating Large Language Models Trained on Code. arXiv preprint arXiv:2107.03374."},{"key":"e_1_3_2_2_13_1","unstructured":"Xinyun Chen Chang Liu and Dawn Song. 2018. Tree-to-Tree Neural Networks for Program Translation. In Advances in Neural Information Processing Systems. 31."},{"key":"e_1_3_2_2_14_1","first-page":"16344","article-title":"Flashattention: Fast and memory-efficient exact attention with io-awareness","volume":"35","author":"Dao Tri","year":"2022","unstructured":"Tri Dao, Dan Fu, Stefano Ermon, Atri Rudra, and Christopher R\u00e9. 2022. Flashattention: Fast and memory-efficient exact attention with io-awareness. Advances in Neural Information Processing Systems, 35 (2022), 16344\u201316359.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_2_15_1","doi-asserted-by":"crossref","unstructured":"Yangruibo Ding Baishakhi Ray Premkumar Devanbu and Vincent J Hellendoorn. 2020. Patching as Translation: the Data and the Metaphor. In Automated Software Engineering. 275\u2013286.","DOI":"10.1145\/3324884.3416587"},{"key":"e_1_3_2_2_16_1","unstructured":"Elliot Chance et al.. 2021. A tool for transpiling C to Go. https:\/\/github.com\/elliotchance\/c2go"},{"key":"e_1_3_2_2_17_1","unstructured":"Python Software Foundation. 2023. difflib \u2014 Helpers for computing deltas. https:\/\/docs.python.org\/3\/library\/difflib.html"},{"key":"e_1_3_2_2_18_1","unstructured":"Galois and Immunant. 2023. C2Rust. https:\/\/github.com\/immunant\/c2rust"},{"key":"e_1_3_2_2_19_1","doi-asserted-by":"publisher","DOI":"10.1145\/3468264.3468553"},{"key":"e_1_3_2_2_20_1","unstructured":"Google. 2023. Google Cloud. https:\/\/cloud.google.com\/"},{"key":"e_1_3_2_2_21_1","volume-title":"UniXcoder: Unified Cross-Modal Pre-training for Code Representation. In Annual Meeting of the Association for Computational Linguistics. 7212\u20137225","author":"Guo Daya","year":"2022","unstructured":"Daya Guo, Shuai Lu, Nan Duan, Yanlin Wang, Ming Zhou, and Jian Yin. 2022. UniXcoder: Unified Cross-Modal Pre-training for Code Representation. In Annual Meeting of the Association for Computational Linguistics. 7212\u20137225."},{"key":"e_1_3_2_2_22_1","doi-asserted-by":"publisher","DOI":"10.1145\/2491411.2491461"},{"key":"e_1_3_2_2_23_1","unstructured":"MongoDB Inc.. 2023. MongoDB. https:\/\/www.mongodb.com\/"},{"key":"e_1_3_2_2_24_1","doi-asserted-by":"crossref","unstructured":"Paul Jaccard. 1912. The Distribution of the Flora in the Alpine Zone. New phytologist 37\u201350.","DOI":"10.1111\/j.1469-8137.1912.tb05611.x"},{"key":"e_1_3_2_2_25_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.acl-long.597"},{"key":"e_1_3_2_2_26_1","unstructured":"Junaed Younus Khan and Gias Uddin. 2022. Automatic Code Documentation Generation Using GPT-3. In Automated Software Engineering. 1\u20136."},{"key":"e_1_3_2_2_27_1","unstructured":"Marie-Anne Lachaux Baptiste Roziere Lowik Chanussot and Guillaume Lample. 2020. Unsupervised Translation of Programming Languages. In Advances in Neural Information Processing Systems. 20601\u201320611."},{"key":"e_1_3_2_2_28_1","doi-asserted-by":"publisher","DOI":"10.1145\/3540250.3549081"},{"key":"e_1_3_2_2_29_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICPC52881.2021.00013"},{"key":"e_1_3_2_2_30_1","doi-asserted-by":"publisher","DOI":"10.1109\/TSE.2022.3185458"},{"key":"e_1_3_2_2_31_1","volume-title":"Roberta: A Robustly Optimized Bert Pretraining Approach. arXiv preprint arXiv:1907.11692.","author":"Liu Yinhan","year":"2019","unstructured":"Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. Roberta: A Robustly Optimized Bert Pretraining Approach. arXiv preprint arXiv:1907.11692."},{"key":"e_1_3_2_2_32_1","article-title":"Just-in-time obsolete comment detection and update","author":"Liu Zhongxin","year":"2021","unstructured":"Zhongxin Liu, Xin Xia, David Lo, Meng Yan, and Shanping Li. 2021. Just-in-time obsolete comment detection and update. IEEE Transactions on Software Engineering.","journal-title":"IEEE Transactions on Software Engineering."},{"key":"e_1_3_2_2_33_1","unstructured":"Shuai Lu Daya Guo Shuo Ren Junjie Huang Alexey Svyatkovskiy Ambrosio Blanco Colin Clement Dawn Drain Daxin Jiang and Duyu Tang. 2021. CodeXGLUE: A Machine Learning Benchmark Dataset for Code Understanding and Generation. arXiv preprint arXiv:2102.04664."},{"key":"e_1_3_2_2_34_1","doi-asserted-by":"publisher","DOI":"10.1145\/3527315"},{"key":"e_1_3_2_2_35_1","volume-title":"Ground Truth for Grammatical Error Correction Metrics. In Annual Meeting of the Association for Computational Linguistics and International Joint Conference on Natural Language Processing. 588\u2013593","author":"Napoles Courtney","year":"2015","unstructured":"Courtney Napoles, Keisuke Sakaguchi, Matt Post, and Joel Tetreault. 2015. Ground Truth for Grammatical Error Correction Metrics. In Annual Meeting of the Association for Computational Linguistics and International Joint Conference on Natural Language Processing. 588\u2013593."},{"key":"e_1_3_2_2_36_1","volume-title":"Tung Thanh Nguyen, and Tien N Nguyen","author":"Nguyen Anh Tuan","year":"2015","unstructured":"Anh Tuan Nguyen, Tung Thanh Nguyen, and Tien N Nguyen. 2015. Divide-and-Conquer Approach for Multi-Phase Statistical Migration for Source Code. In Automated Software Engineering. 585\u2013596."},{"key":"e_1_3_2_2_37_1","volume-title":"Machine Learning for Executable Code in Software Testing and Verification. Ph. D. Dissertation","author":"Nie Pengyu","unstructured":"Pengyu Nie. 2023. Machine Learning for Executable Code in Software Testing and Verification. Ph. D. Dissertation. The University of Texas at Austin."},{"key":"e_1_3_2_2_38_1","volume-title":"Impact of Evaluation Methodologies on Code Summarization. In Annual Meeting of the Association for Computational Linguistics. 4936\u20134960","author":"Nie Pengyu","year":"2022","unstructured":"Pengyu Nie, Jiyang Zhang, Junyi Jessy Li, Raymond J. Mooney, and Milos Gligoric. 2022. Impact of Evaluation Methodologies on Code Summarization. In Annual Meeting of the Association for Computational Linguistics. 4936\u20134960."},{"key":"e_1_3_2_2_40_1","unstructured":"OpenAI. 2023. Introducing ChatGPT. https:\/\/openai.com\/blog\/chatgpt"},{"key":"e_1_3_2_2_41_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.168"},{"key":"e_1_3_2_2_42_1","volume-title":"Annual Meeting of the Association for Computational Linguistics. 311\u2013318","author":"Papineni Kishore","year":"2002","unstructured":"Kishore Papineni, Salim Roukos, Todd Ward, and Wei-Jing Zhu. 2002. BLEU: a Method for Automatic Evaluation of Machine Translation. In Annual Meeting of the Association for Computational Linguistics. 311\u2013318."},{"key":"e_1_3_2_2_43_1","doi-asserted-by":"publisher","DOI":"10.1002\/spe.4380250705"},{"key":"e_1_3_2_2_44_1","doi-asserted-by":"publisher","DOI":"10.1145\/2660193.2660228"},{"key":"e_1_3_2_2_45_1","unstructured":"Shuo Ren Daya Guo Shuai Lu Long Zhou Shujie Liu Duyu Tang Neel Sundaresan Ming Zhou Ambrosio Blanco and Shuai Ma. 2020. CodeBLEU: a Method for Automatic Evaluation of Code Synthesis. arXiv preprint arXiv:2009.10297."},{"key":"e_1_3_2_2_46_1","unstructured":"Baptiste Roziere Jie M Zhang Francois Charton Mark Harman Gabriel Synnaeve and Guillaume Lample. 2021. Leveraging Automated Unit Tests for Unsupervised Code Translation. arXiv preprint arXiv:2110.06773."},{"key":"e_1_3_2_2_47_1","unstructured":"Apache Software. 2022. Apache Lucene. https:\/\/lucene.apache.org\/"},{"key":"e_1_3_2_2_48_1","doi-asserted-by":"crossref","unstructured":"Felix Stahlberg and Shankar Kumar. 2020. Seq2Edits: Sequence Transduction Using Span-level Edit Operations. In Empirical Methods in Natural Language Processing. 5147\u20135159.","DOI":"10.18653\/v1\/2020.emnlp-main.418"},{"key":"e_1_3_2_2_49_1","unstructured":"Sindhu Tipirneni Ming Zhu and Chandan K Reddy. 2022. StructCoder: Structure-Aware Transformer for Code Generation. arXiv preprint arXiv:2206.05239."},{"key":"e_1_3_2_2_50_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-21952-8_4"},{"key":"e_1_3_2_2_51_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICSE.2019.00021"},{"key":"e_1_3_2_2_52_1","doi-asserted-by":"publisher","DOI":"10.1145\/3510003.3510621"},{"key":"e_1_3_2_2_53_1","volume-title":"\u0141 ukasz Kaiser, and Illia Polosukhin","author":"Vaswani Ashish","year":"2017","unstructured":"Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, \u0141 ukasz Kaiser, and Illia Polosukhin. 2017. Attention is All You Need. In Advances in Neural Information Processing Systems. 5998\u20136008."},{"key":"e_1_3_2_2_54_1","volume-title":"Nghi D.Q. Bui, Junnan Li, and Steven C. H. Hoi.","author":"Wang Yue","year":"2023","unstructured":"Yue Wang, Hung Le, Akhilesh Deepak Gotmare, Nghi D.Q. Bui, Junnan Li, and Steven C. H. Hoi. 2023. CodeT5+: Open Code Large Language Models for Code Understanding and Generation. arXiv preprint."},{"key":"e_1_3_2_2_55_1","doi-asserted-by":"crossref","unstructured":"Yue Wang Weishi Wang Shafiq Joty and Steven CH Hoi. 2021. CodeT5: Identifier-aware Unified Pre-trained Encoder-Decoder Models for Code Understanding and Generation. In Empirical Methods in Natural Language Processing. 8696\u20138708.","DOI":"10.18653\/v1\/2021.emnlp-main.685"},{"key":"e_1_3_2_2_56_1","doi-asserted-by":"publisher","DOI":"10.1162\/tacl_a_00107"},{"key":"e_1_3_2_2_57_1","volume-title":"International Conference on Learning Representations.","author":"Yao Ziyu","year":"2021","unstructured":"Ziyu Yao, Frank F. Xu, Pengcheng Yin, Huan Sun, and Graham Neubig. 2021. Learning Structural Edits via Incremental Tree Transformations. In International Conference on Learning Representations."},{"key":"e_1_3_2_2_58_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICSE-SEIP58684.2023.00020"},{"key":"e_1_3_2_2_59_1","volume-title":"Junyi Jessy Li, and Milos Gligoric","author":"Zhang Jiyang","year":"2022","unstructured":"Jiyang Zhang, Sheena Panthaplackel, Pengyu Nie, Junyi Jessy Li, and Milos Gligoric. 2022. CoditT5: Pretraining for Source Code and Natural Language Editing. In Automated Software Engineering. 1\u201312."},{"key":"e_1_3_2_2_60_1","volume-title":"Multilingual Code Snippets Training for Program Translation. In AAAI Conference on Artificial Intelligence. 11783\u201311790","author":"Zhu Ming","year":"2022","unstructured":"Ming Zhu, Karthik Suresh, and Chandan K Reddy. 2022. Multilingual Code Snippets Training for Program Translation. In AAAI Conference on Artificial Intelligence. 11783\u201311790."}],"event":{"name":"ESEC\/FSE '23: 31st ACM Joint European Software Engineering Conference and Symposium on the Foundations of Software Engineering","location":"San Francisco CA USA","acronym":"ESEC\/FSE '23","sponsor":["SIGSOFT ACM Special Interest Group on Software Engineering"]},"container-title":["Proceedings of the 31st ACM Joint European Software Engineering Conference and Symposium on the Foundations of Software Engineering"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3611643.3616350","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3611643.3616350","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T16:36:11Z","timestamp":1750178171000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3611643.3616350"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,11,30]]},"references-count":59,"alternative-id":["10.1145\/3611643.3616350","10.1145\/3611643"],"URL":"https:\/\/doi.org\/10.1145\/3611643.3616350","relation":{},"subject":[],"published":{"date-parts":[[2023,11,30]]},"assertion":[{"value":"2023-11-30","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}