{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,28]],"date-time":"2026-03-28T20:58:53Z","timestamp":1774731533328,"version":"3.50.1"},"reference-count":115,"publisher":"Association for Computing Machinery (ACM)","issue":"3","license":[{"start":{"date-parts":[[2024,3,15]],"date-time":"2024-03-15T00:00:00Z","timestamp":1710460800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["ACM Trans. Softw. Eng. Methodol."],"published-print":{"date-parts":[[2024,3,31]]},"abstract":"<jats:p>Automated Program Repair (APR) is defined as the process of fixing a bug\/defect in the source code, by an automated tool. APR tools have recently experienced promising results by leveraging state-of-the-art Neural Language Processing (NLP) techniques. APR tools such as TFix and CodeXGLUE that combine text-to-text transformers with software-specific techniques are outperforming alternatives, these days. However, in most APR studies, the train and test sets are chosen from the same set of projects (i.e., when APR fixes a bug in the test set from project A, the model has already seen example fixed bugs from project A in the training set). In the real world, however, APR models are meant to be generalizable to new and different projects. Therefore, there is a potential threat that reported APR models with high effectiveness perform poorly when the characteristics of the new project or its bugs are different than the training set\u2019s (\u201cDomain Shift\u201d).<\/jats:p>\n          <jats:p>\n            In this study, we first define the problem of domain shift in automated program repair. Next, we measure the potential damage of domain shift on two recent APR models (TFix and CodeXGLUE). Based on this observation, we then propose a domain adaptation framework that can adapt an APR model for a given target project. We conduct an empirical study with three domain adaptation methods\n            <jats:italic>FullFineTuning<\/jats:italic>\n            ,\n            <jats:italic>TuningWithLightWeightAdapterLayers<\/jats:italic>\n            , and\n            <jats:italic>CurriculumLearning<\/jats:italic>\n            and two APR models on 2,672 bugs from 12 projects.\n          <\/jats:p>\n          <jats:p>The results show that our proposed framework on average can improve the effectiveness of TFix by 13.05% and CodeXGLUE by 48.78%, in terms of \u201cExact Match\u201d. Through experiments, we also show that the framework provides high efficiency and reliability (in terms of \u201cExposure Bias\u201d). Using synthetic data to domain adapt TFix and CodeXGLUE on the projects with no data (Zero-shot learning), also results in an average improvement of 5.76% and 17.62% for TFix and CodeXGLUE, respectively.<\/jats:p>","DOI":"10.1145\/3631972","type":"journal-article","created":{"date-parts":[[2023,11,21]],"date-time":"2023-11-21T08:26:14Z","timestamp":1700555174000},"page":"1-43","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":14,"title":["Improving Automated Program Repair with Domain Adaptation"],"prefix":"10.1145","volume":"33","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-7095-166X","authenticated-orcid":false,"given":"Armin","family":"Zirak","sequence":"first","affiliation":[{"name":"University of Calgary, Alberta, Canada"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0204-9812","authenticated-orcid":false,"given":"Hadi","family":"Hemmati","sequence":"additional","affiliation":[{"name":"University of Calgary, Alberta, Canada"}]}],"member":"320","published-online":{"date-parts":[[2024,3,15]]},"reference":[{"key":"e_1_3_2_2_2","unstructured":"2022. ESLint Find and Fix Problems in Your JavaScript code. https:\/\/eslint.org\/"},{"key":"e_1_3_2_3_2","unstructured":"2022. Lightrun a Developer-Native Observability Platform. https:\/\/lightrun.com\/"},{"key":"e_1_3_2_4_2","volume-title":"Conference on Machine Translation","author":"Abdul-Rauf Sadaf","year":"2020","unstructured":"Sadaf Abdul-Rauf, Jos\u00e9 Carlos Rosales, Minh Quang Pham, and Fran\u00e7ois Yvon. 2020. Limsi@ wmt 2020. In Conference on Machine Translation."},{"key":"e_1_3_2_5_2","article-title":"Deep learning using rectified linear units (relu)","author":"Agarap Abien Fred","year":"2018","unstructured":"Abien Fred Agarap. 2018. Deep learning using rectified linear units (relu). arXiv preprint arXiv:1803.08375 (2018).","journal-title":"arXiv preprint arXiv:1803.08375"},{"key":"e_1_3_2_6_2","doi-asserted-by":"publisher","DOI":"10.1145\/3183377.3183383"},{"key":"e_1_3_2_7_2","article-title":"Gradient based sample selection for online continual learning","volume":"32","author":"Aljundi Rahaf","year":"2019","unstructured":"Rahaf Aljundi, Min Lin, Baptiste Goujaud, and Yoshua Bengio. 2019. Gradient based sample selection for online continual learning. Advances in Neural Information Processing Systems 32 (2019).","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_8_2","article-title":"Learning to represent programs with graphs","author":"Allamanis Miltiadis","year":"2017","unstructured":"Miltiadis Allamanis, Marc Brockschmidt, and Mahmoud Khademi. 2017. Learning to represent programs with graphs. arXiv preprint arXiv:1711.00740 (2017).","journal-title":"arXiv preprint arXiv:1711.00740"},{"key":"e_1_3_2_9_2","first-page":"27865","article-title":"Self-supervised bug detection and repair","volume":"34","author":"Allamanis Miltiadis","year":"2021","unstructured":"Miltiadis Allamanis, Henry Jackson-Flux, and Marc Brockschmidt. 2021. Self-supervised bug detection and repair. Advances in Neural Information Processing Systems 34 (2021), 27865\u201327876.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_10_2","article-title":"Concrete problems in AI safety","author":"Amodei Dario","year":"2016","unstructured":"Dario Amodei, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Man\u00e9. 2016. Concrete problems in AI safety. arXiv preprint arXiv:1606.06565 (2016).","journal-title":"arXiv preprint arXiv:1606.06565"},{"issue":"2019","key":"e_1_3_2_11_2","first-page":"1","article-title":"Getafix: Learning to fix bugs automatically","volume":"3","author":"Bader Johannes","year":"2019","unstructured":"Johannes Bader, Andrew Scott, Michael Pradel, and Satish Chandra. 2019. Getafix: Learning to fix bugs automatically. In ACM on Programming Languages 3(OOPSLA2019), 1\u201327.","journal-title":"ACM on Programming Languages"},{"key":"e_1_3_2_12_2","article-title":"Simple, scalable adaptation for neural machine translation","author":"Bapna Ankur","year":"2019","unstructured":"Ankur Bapna, Naveen Arivazhagan, and Orhan Firat. 2019. Simple, scalable adaptation for neural machine translation. arXiv preprint arXiv:1909.08478 (2019).","journal-title":"arXiv preprint arXiv:1909.08478"},{"key":"e_1_3_2_13_2","article-title":"Non-parametric adaptation for neural machine translation","author":"Bapna Ankur","year":"2019","unstructured":"Ankur Bapna and Orhan Firat. 2019. Non-parametric adaptation for neural machine translation. arXiv preprint arXiv:1903.00058 (2019).","journal-title":"arXiv preprint arXiv:1903.00058"},{"key":"e_1_3_2_14_2","first-page":"56","volume-title":"Fifth Conference on Machine Translation (EMNLP 2020) November 19-20, 2020, online: Proceedings of the conference","author":"Barrault Lo\u00efc","year":"2020","unstructured":"Lo\u00efc Barrault, Magdalena Marta Biesialska, Marta Ruiz Costa-Juss\u00e0, Fethi Bougares, and Olivier Galibert. 2020. Findings of the first shared task on lifelong learning machine translation. In Fifth Conference on Machine Translation (EMNLP 2020) November 19-20, 2020, online: Proceedings of the conference. Association for Computational Linguistics, 56\u201364."},{"key":"e_1_3_2_15_2","first-page":"780","volume-title":"International Conference on Machine Learning","author":"Berabi Berkay","year":"2021","unstructured":"Berkay Berabi, Jingxuan He, Veselin Raychev, and Martin Vechev. 2021. Tfix: Learning to fix coding errors with a text-to-text transformer. In International Conference on Machine Learning. PMLR, 780\u2013791."},{"key":"e_1_3_2_16_2","doi-asserted-by":"publisher","DOI":"10.1145\/3180155.3180219"},{"key":"e_1_3_2_17_2","first-page":"1877","article-title":"Language models are few-shot learners","volume":"33","author":"Brown Tom","year":"2020","unstructured":"Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared D. Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, et\u00a0al. 2020. Language models are few-shot learners. Advances in Neural Information Processing Systems 33 (2020), 1877\u20131901.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_18_2","first-page":"249","volume-title":"11th Conference of the European Chapter of the Association for Computational Linguistics","author":"Callison-Burch Chris","year":"2006","unstructured":"Chris Callison-Burch, Miles Osborne, and Philipp Koehn. 2006. Re-evaluating the role of BLEU in machine translation research. In 11th Conference of the European Chapter of the Association for Computational Linguistics. 249\u2013256."},{"key":"e_1_3_2_19_2","article-title":"Cross-domain deep code search with few-shot meta learning","author":"Chai Yitian","year":"2022","unstructured":"Yitian Chai, Hongyu Zhang, Beijun Shen, and Xiaodong Gu. 2022. Cross-domain deep code search with few-shot meta learning. arXiv preprint arXiv:2201.00150 (2022).","journal-title":"arXiv preprint arXiv:2201.00150"},{"key":"e_1_3_2_20_2","doi-asserted-by":"publisher","DOI":"10.1145\/3524610.3527917"},{"key":"e_1_3_2_21_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.iwslt-1.14"},{"key":"e_1_3_2_22_2","first-page":"23089","article-title":"PLUR: A unifying, graph-based view of program learning, understanding, and repair","volume":"34","author":"Chen Zimin","year":"2021","unstructured":"Zimin Chen, Vincent J. Hellendoorn, Pascal Lamblin, Petros Maniatis, Pierre-Antoine Manzagol, Daniel Tarlow, and Subhodeep Moitra. 2021. PLUR: A unifying, graph-based view of program learning, understanding, and repair. Advances in Neural Information Processing Systems 34 (2021), 23089\u201323101.","journal-title":"Advances in Neural Information Processing Systems"},{"issue":"9","key":"e_1_3_2_23_2","first-page":"1943","article-title":"Sequencer: Sequence-to-sequence learning for end-to-end program repair","volume":"47","author":"Chen Zimin","year":"2019","unstructured":"Zimin Chen, Steve Kommrusch, Michele Tufano, Louis-No\u00ebl Pouchet, Denys Poshyvanyk, and Martin Monperrus. 2019. Sequencer: Sequence-to-sequence learning for end-to-end program repair. IEEE Transactions on Software Engineering 47, 9 (2019), 1943\u20131959.","journal-title":"IEEE Transactions on Software Engineering"},{"key":"e_1_3_2_24_2","first-page":"385","volume-title":"55th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)","author":"Chu Chenhui","year":"2017","unstructured":"Chenhui Chu, Raj Dabre, and Sadao Kurohashi. 2017. An empirical comparison of domain adaptation methods for neural machine translation. In 55th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers). 385\u2013391."},{"key":"e_1_3_2_25_2","volume-title":"Practical Nonparametric Statistics","author":"Conover William Jay","year":"1999","unstructured":"William Jay Conover. 1999. Practical Nonparametric Statistics. Vol. 350. Wiley."},{"key":"e_1_3_2_26_2","article-title":"Data Preparation for software vulnerability prediction: A systematic literature review","author":"Croft Roland","year":"2022","unstructured":"Roland Croft, Yongzheng Xie, and Muhammad Ali Babar. 2022. Data Preparation for software vulnerability prediction: A systematic literature review. IEEE Transactions on Software Engineering (2022).","journal-title":"IEEE Transactions on Software Engineering"},{"key":"e_1_3_2_27_2","doi-asserted-by":"publisher","DOI":"10.1007\/BF02289138"},{"key":"e_1_3_2_28_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.364"},{"key":"e_1_3_2_29_2","first-page":"1","volume-title":"5th International Workshop on Machine Learning Techniques for Software Quality Evolution","author":"Stefano Manuel De","year":"2021","unstructured":"Manuel De Stefano, Fabiano Pecorelli, Fabio Palomba, and Andrea De Lucia. 2021. Comparing within-and cross-project machine learning algorithms for code smell detection. In 5th International Workshop on Machine Learning Techniques for Software Quality Evolution. 1\u20136."},{"key":"e_1_3_2_30_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICSTW55395.2022.00039"},{"key":"e_1_3_2_31_2","article-title":"BERTt: Pre-training of deep bidirectional transformers for language understanding","author":"Devlin Jacob","year":"2018","unstructured":"Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. BERTt: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018).","journal-title":"arXiv preprint arXiv:1810.04805"},{"key":"e_1_3_2_32_2","volume-title":"International Conference on Learning Representations (ICLR)","author":"Dinella Elizabeth","year":"2020","unstructured":"Elizabeth Dinella, Hanjun Dai, Ziyang Li, Mayur Naik, Le Song, and Ke Wang. 2020. Hoppity: Learning graph transformations to detect and fix bugs in programs. In International Conference on Learning Representations (ICLR)."},{"key":"e_1_3_2_33_2","article-title":"Understanding back-translation at scale","author":"Edunov Sergey","year":"2018","unstructured":"Sergey Edunov, Myle Ott, Michael Auli, and David Grangier. 2018. Understanding back-translation at scale. arXiv preprint arXiv:1808.09381 (2018).","journal-title":"arXiv preprint arXiv:1808.09381"},{"key":"e_1_3_2_34_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-71704-9_65"},{"key":"e_1_3_2_35_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/W17-4713"},{"key":"e_1_3_2_36_2","article-title":"Codebert: A pre-trained model for programming and natural languages","author":"Feng Zhangyin","year":"2020","unstructured":"Zhangyin Feng, Daya Guo, Duyu Tang, Nan Duan, Xiaocheng Feng, Ming Gong, Linjun Shou, Bing Qin, Ting Liu, Daxin Jiang, et\u00a0al. 2020. Codebert: A pre-trained model for programming and natural languages. arXiv preprint arXiv:2002.08155 (2020).","journal-title":"arXiv preprint arXiv:2002.08155"},{"key":"e_1_3_2_37_2","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2020.3025954"},{"key":"e_1_3_2_38_2","doi-asserted-by":"publisher","DOI":"10.1109\/TSE.2017.2755013"},{"key":"e_1_3_2_39_2","article-title":"Graphcodebert: Pre-training code representations with data flow","author":"Guo Daya","year":"2020","unstructured":"Daya Guo, Shuo Ren, Shuai Lu, Zhangyin Feng, Duyu Tang, Shujie Liu, Long Zhou, Nan Duan, Alexey Svyatkovskiy, Shengyu Fu, et\u00a0al. 2020. Graphcodebert: Pre-training code representations with data flow. arXiv preprint arXiv:2009.08366 (2020).","journal-title":"arXiv preprint arXiv:2009.08366"},{"key":"e_1_3_2_40_2","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v31i1.10742"},{"key":"e_1_3_2_41_2","article-title":"On distribution shift in learning-based bug detectors","author":"He Jingxuan","year":"2022","unstructured":"Jingxuan He, Luca Beurer-Kellner, and Martin Vechev. 2022. On distribution shift in learning-based bug detectors. arXiv preprint arXiv:2204.10049 (2022).","journal-title":"arXiv preprint arXiv:2204.10049"},{"key":"e_1_3_2_42_2","volume-title":"International Conference on Learning Representations","author":"Hellendoorn Vincent J.","year":"2019","unstructured":"Vincent J. Hellendoorn, Charles Sutton, Rishabh Singh, Petros Maniatis, and David Bieber. 2019. Global relational models of source code. In International Conference on Learning Representations."},{"key":"e_1_3_2_43_2","article-title":"Codesearchnet challenge: Evaluating the state of semantic code search","author":"Husain Hamel","year":"2019","unstructured":"Hamel Husain, Ho-Hsiang Wu, Tiferet Gazit, Miltiadis Allamanis, and Marc Brockschmidt. 2019. Codesearchnet challenge: Evaluating the state of semantic code search. arXiv preprint arXiv:1909.09436 (2019).","journal-title":"arXiv preprint arXiv:1909.09436"},{"key":"e_1_3_2_44_2","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i01.5341"},{"key":"e_1_3_2_45_2","article-title":"Impact of code language models on automated program repair","author":"Jiang Nan","year":"2023","unstructured":"Nan Jiang, Kevin Liu, Thibaud Lutellier, and Lin Tan. 2023. Impact of code language models on automated program repair. arXiv preprint arXiv:2302.05020 (2023).","journal-title":"arXiv preprint arXiv:2302.05020"},{"key":"e_1_3_2_46_2","article-title":"KNOD: Domain knowledge distilled tree decoder for automated program repair","author":"Jiang Nan","year":"2023","unstructured":"Nan Jiang, Thibaud Lutellier, Yiling Lou, Lin Tan, Dan Goldwasser, and Xiangyu Zhang. 2023. KNOD: Domain knowledge distilled tree decoder for automated program repair. arXiv preprint arXiv:2302.01857 (2023).","journal-title":"arXiv preprint arXiv:2302.01857"},{"key":"e_1_3_2_47_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICSE43902.2021.00107"},{"key":"e_1_3_2_48_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2021.114637"},{"key":"e_1_3_2_49_2","article-title":"A simple baseline to semi-supervised domain adaptation for machine translation","author":"Jin Di","year":"2020","unstructured":"Di Jin, Zhijing Jin, Joey Tianyi Zhou, and Peter Szolovits. 2020. A simple baseline to semi-supervised domain adaptation for machine translation. arXiv preprint arXiv:2001.08140 (2020).","journal-title":"arXiv preprint arXiv:2001.08140"},{"key":"e_1_3_2_50_2","first-page":"5110","volume-title":"International Conference on Machine Learning","author":"Kanade Aditya","year":"2020","unstructured":"Aditya Kanade, Petros Maniatis, Gogul Balakrishnan, and Kensen Shi. 2020. Learning and evaluating contextual embedding of source code. In International Conference on Machine Learning. PMLR, 5110\u20135121."},{"key":"e_1_3_2_51_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.privatenlp-1.6"},{"key":"e_1_3_2_52_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.282"},{"key":"e_1_3_2_53_2","article-title":"Six challenges for neural machine translation","author":"Koehn Philipp","year":"2017","unstructured":"Philipp Koehn and Rebecca Knowles. 2017. Six challenges for neural machine translation. arXiv preprint arXiv:1706.03872 (2017).","journal-title":"arXiv preprint arXiv:1706.03872"},{"key":"e_1_3_2_54_2","first-page":"284","volume-title":"6th Workshop on Statistical Machine Translation","author":"Lambert Patrik","year":"2011","unstructured":"Patrik Lambert, Holger Schwenk, Christophe Servan, and Sadaf Abdul-Rauf. 2011. Investigations on translation model adaptation using monolingual data. In 6th Workshop on Statistical Machine Translation. 284\u2013293."},{"key":"e_1_3_2_55_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICSE.2012.6227211"},{"key":"e_1_3_2_56_2","doi-asserted-by":"publisher","DOI":"10.1109\/TSE.2011.104"},{"key":"e_1_3_2_57_2","doi-asserted-by":"publisher","DOI":"10.1038\/nature14539"},{"key":"e_1_3_2_58_2","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i05.6339"},{"key":"e_1_3_2_59_2","article-title":"One sentence one model for neural machine translation","author":"Li Xiaoqing","year":"2016","unstructured":"Xiaoqing Li, Jiajun Zhang, and Chengqing Zong. 2016. One sentence one model for neural machine translation. arXiv preprint arXiv:1609.06490 (2016).","journal-title":"arXiv preprint arXiv:1609.06490"},{"key":"e_1_3_2_60_2","doi-asserted-by":"publisher","DOI":"10.1145\/3377811.3380345"},{"key":"e_1_3_2_61_2","article-title":"DEAR: A novel deep learning-based approach for automated program repair","author":"Li Yi","year":"2022","unstructured":"Yi Li, Shaohua Wang, and Tien N. Nguyen. 2022. DEAR: A novel deep learning-based approach for automated program repair. arXiv preprint arXiv:2205.01859 (2022).","journal-title":"arXiv preprint arXiv:2205.01859"},{"key":"e_1_3_2_62_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.infsof.2018.04.001"},{"key":"e_1_3_2_63_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.infsof.2018.11.005"},{"key":"e_1_3_2_64_2","article-title":"CD-VulD: Cross-domain vulnerability discovery based on deep domain adaptation","author":"Liu Shigang","year":"2020","unstructured":"Shigang Liu, Guanjun Lin, Lizhen Qu, Jun Zhang, Olivier De Vel, Paul Montague, and Yang Xiang. 2020. CD-VulD: Cross-domain vulnerability discovery based on deep domain adaptation. IEEE Transactions on Dependable and Secure Computing (2020).","journal-title":"IEEE Transactions on Dependable and Secure Computing"},{"key":"e_1_3_2_65_2","doi-asserted-by":"crossref","first-page":"727","DOI":"10.1145\/3106237.3106253","volume-title":"2017 11th Joint Meeting on Foundations of Software Engineering","author":"Long Fan","year":"2017","unstructured":"Fan Long, Peter Amidon, and Martin Rinard. 2017. Automatic inference of code transforms for patch generation. In 2017 11th Joint Meeting on Foundations of Software Engineering. 727\u2013739."},{"key":"e_1_3_2_66_2","first-page":"298","volume-title":"43rd Annual ACM SIGPLAN-SIGACT Symposium on Principles of Programming Languages","author":"Long Fan","year":"2016","unstructured":"Fan Long and Martin Rinard. 2016. Automatic patch generation by learning correct code. In 43rd Annual ACM SIGPLAN-SIGACT Symposium on Principles of Programming Languages. 298\u2013312."},{"key":"e_1_3_2_67_2","article-title":"CodexGLUE: A machine learning benchmark dataset for code understanding and generation","author":"Lu Shuai","year":"2021","unstructured":"Shuai Lu, Daya Guo, Shuo Ren, Junjie Huang, Alexey Svyatkovskiy, Ambrosio Blanco, Colin Clement, Dawn Drain, Daxin Jiang, Duyu Tang, et\u00a0al. 2021. CodexGLUE: A machine learning benchmark dataset for code understanding and generation. arXiv preprint arXiv:2102.04664 (2021).","journal-title":"arXiv preprint arXiv:2102.04664"},{"key":"e_1_3_2_68_2","volume-title":"12th International Workshop on Spoken Language Translation: Evaluation Campaign","author":"Luong Minh-Thang","year":"2015","unstructured":"Minh-Thang Luong and Christopher D. Manning. 2015. Stanford neural machine translation systems for spoken language domains. In 12th International Workshop on Spoken Language Translation: Evaluation Campaign."},{"key":"e_1_3_2_69_2","first-page":"101","volume-title":"29th ACM SIGSOFT International Symposium on Software Testing and Analysis","author":"Lutellier Thibaud","year":"2020","unstructured":"Thibaud Lutellier, Hung Viet Pham, Lawrence Pang, Yitong Li, Moshi Wei, and Lin Tan. 2020. Coconut: Combining context-aware neural translation models using ensemble for program repair. In 29th ACM SIGSOFT International Symposium on Software Testing and Analysis. 101\u2013114."},{"key":"e_1_3_2_70_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.infsof.2011.09.007"},{"key":"e_1_3_2_71_2","first-page":"441","volume-title":"25th International Symposium on Software Testing and Analysis","author":"Martinez Matias","year":"2016","unstructured":"Matias Martinez and Martin Monperrus. 2016. Astor: A program repair library for java. In 25th International Symposium on Software Testing and Analysis. 441\u2013444."},{"key":"e_1_3_2_72_2","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i05.6369"},{"key":"e_1_3_2_73_2","doi-asserted-by":"publisher","DOI":"10.1145\/3105906"},{"key":"e_1_3_2_74_2","doi-asserted-by":"publisher","DOI":"10.5555\/2486788.2486839"},{"key":"e_1_3_2_75_2","doi-asserted-by":"publisher","DOI":"10.1109\/IJCNN.2019.8851923"},{"key":"e_1_3_2_76_2","doi-asserted-by":"publisher","DOI":"10.1007\/s11390-017-1785-0"},{"key":"e_1_3_2_77_2","first-page":"311","volume-title":"40th Annual Meeting of the Association for Computational Linguistics","author":"Papineni Kishore","year":"2002","unstructured":"Kishore Papineni, Salim Roukos, Todd Ward, and Wei-Jing Zhu. 2002. BLEU: A method for automatic evaluation of machine translation. In 40th Annual Meeting of the Association for Computational Linguistics. 311\u2013318."},{"key":"e_1_3_2_78_2","doi-asserted-by":"publisher","DOI":"10.1109\/MSP.2014.2347059"},{"key":"e_1_3_2_79_2","doi-asserted-by":"publisher","DOI":"10.1145\/3468264.3468623"},{"key":"e_1_3_2_80_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01252-6_47"},{"key":"e_1_3_2_81_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.361"},{"key":"e_1_3_2_82_2","article-title":"Selecting artificially-generated sentences for fine-tuning neural machine translation","author":"Poncelas Alberto","year":"2019","unstructured":"Alberto Poncelas and Andy Way. 2019. Selecting artificially-generated sentences for fine-tuning neural machine translation. arXiv preprint arXiv:1909.12016 (2019).","journal-title":"arXiv preprint arXiv:1909.12016"},{"key":"e_1_3_2_83_2","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3276517","article-title":"Deepbugs: A learning approach to name-based bug detection","volume":"2","author":"Pradel Michael","year":"2018","unstructured":"Michael Pradel and Koushik Sen. 2018. Deepbugs: A learning approach to name-based bug detection. In ACM on Programming Languages 2, (OOPSLA), (2018), 1\u201325.","journal-title":"ACM on Programming Languages"},{"key":"e_1_3_2_84_2","doi-asserted-by":"publisher","DOI":"10.1145\/2568225.2568254"},{"issue":"140","key":"e_1_3_2_85_2","first-page":"1","article-title":"Exploring the limits of transfer learning with a unified text-to-text transformer.","volume":"21","author":"Raffel Colin","year":"2020","unstructured":"Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, Peter J. Liu, et\u00a0al. 2020. Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 21, 140 (2020), 1\u201367.","journal-title":"J. Mach. Learn. Res."},{"key":"e_1_3_2_86_2","volume-title":"Advances in Domain Adaptation Theory","author":"Redko Ievgen","year":"2019","unstructured":"Ievgen Redko, Emilie Morvant, Amaury Habrard, Marc Sebban, and Younes Bennani. 2019. Advances in Domain Adaptation Theory. Elsevier."},{"key":"e_1_3_2_87_2","article-title":"CodeBLEU: A method for automatic evaluation of code synthesis","author":"Ren Shuo","year":"2020","unstructured":"Shuo Ren, Daya Guo, Shuai Lu, Long Zhou, Shujie Liu, Duyu Tang, Neel Sundaresan, Ming Zhou, Ambrosio Blanco, and Shuai Ma. 2020. CodeBLEU: A method for automatic evaluation of code synthesis. arXiv preprint arXiv:2009.10297 (2020).","journal-title":"arXiv preprint arXiv:2009.10297"},{"key":"e_1_3_2_88_2","article-title":"On the effectiveness of transfer learning for code search","author":"Salza Pasquale","year":"2021","unstructured":"Pasquale Salza, Christoph Schwizer, Jian Gu, and Harald C. Gall. 2021. On the effectiveness of transfer learning for code search. arXiv preprint arXiv:2108.05890 (2021).","journal-title":"arXiv preprint arXiv:2108.05890"},{"key":"e_1_3_2_89_2","article-title":"Domain adaptation and multi-domain adaptation for neural machine translation: A survey","author":"Saunders Danielle","year":"2021","unstructured":"Danielle Saunders. 2021. Domain adaptation and multi-domain adaptation for neural machine translation: A survey. arXiv preprint arXiv:2104.06951 (2021).","journal-title":"arXiv preprint arXiv:2104.06951"},{"key":"e_1_3_2_90_2","article-title":"Addressing exposure bias with document minimum risk training: Cambridge at the WMT20 biomedical translation task","author":"Saunders Danielle","year":"2020","unstructured":"Danielle Saunders and Bill Byrne. 2020. Addressing exposure bias with document minimum risk training: Cambridge at the WMT20 biomedical translation task. arXiv preprint arXiv:2010.05333 (2020).","journal-title":"arXiv preprint arXiv:2010.05333"},{"key":"e_1_3_2_91_2","volume-title":"5th International Workshop on Spoken Language Translation: Papers","author":"Schwenk Holger","year":"2008","unstructured":"Holger Schwenk. 2008. Investigations on large-scale lightly-supervised training for statistical machine translation.. In 5th International Workshop on Spoken Language Translation: Papers."},{"issue":"4","key":"e_1_3_2_92_2","first-page":"35","article-title":"Modern information retrieval: A brief overview","volume":"24","author":"Singhal Amit","year":"2001","unstructured":"Amit Singhal et\u00a0al. 2001. Modern information retrieval: A brief overview. IEEE Data Eng. Bull. 24, 4 (2001), 35\u201343.","journal-title":"IEEE Data Eng. Bull."},{"key":"e_1_3_2_93_2","doi-asserted-by":"publisher","DOI":"10.1109\/SANER.2018.8330203"},{"key":"e_1_3_2_94_2","article-title":"Code-switching for enhancing NMT with pre-specified translation","author":"Song Kai","year":"2019","unstructured":"Kai Song, Yue Zhang, Heng Yu, Weihua Luo, Kun Wang, and Min Zhang. 2019. Code-switching for enhancing NMT with pre-specified translation. arXiv preprint arXiv:1904.09107 (2019).","journal-title":"arXiv preprint arXiv:1904.09107"},{"key":"e_1_3_2_95_2","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v30i1.10306"},{"key":"e_1_3_2_96_2","doi-asserted-by":"publisher","DOI":"10.1145\/3238147.3240732"},{"key":"e_1_3_2_97_2","doi-asserted-by":"publisher","DOI":"10.1109\/ICSME.2019.00046"},{"key":"e_1_3_2_98_2","doi-asserted-by":"publisher","DOI":"10.1145\/3340544"},{"key":"e_1_3_2_99_2","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.316"},{"key":"e_1_3_2_100_2","article-title":"Dynamic data selection for neural machine translation","author":"Wees Marlies Van Der","year":"2017","unstructured":"Marlies Van Der Wees, Arianna Bisazza, and Christof Monz. 2017. Dynamic data selection for neural machine translation. arXiv preprint arXiv:1708.00712 (2017).","journal-title":"arXiv preprint arXiv:1708.00712"},{"key":"e_1_3_2_101_2","article-title":"Neural program repair by jointly learning to localize and repair","author":"Vasic Marko","year":"2019","unstructured":"Marko Vasic, Aditya Kanade, Petros Maniatis, David Bieber, and Rishabh Singh. 2019. Neural program repair by jointly learning to localize and repair. arXiv preprint arXiv:1904.01720 (2019).","journal-title":"arXiv preprint arXiv:1904.01720"},{"key":"e_1_3_2_102_2","article-title":"Attention is all you need","volume":"30","author":"Vaswani Ashish","year":"2017","unstructured":"Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, \u0141ukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. Advances in Neural Information Processing systems 30 (2017).","journal-title":"Advances in Neural Information Processing systems"},{"key":"e_1_3_2_103_2","article-title":"On exposure bias, hallucination and domain shift in neural machine translation","author":"Wang Chaojun","year":"2020","unstructured":"Chaojun Wang and Rico Sennrich. 2020. On exposure bias, hallucination and domain shift in neural machine translation. arXiv preprint arXiv:2005.03642 (2020).","journal-title":"arXiv preprint arXiv:2005.03642"},{"key":"e_1_3_2_104_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58589-1_20"},{"key":"e_1_3_2_105_2","article-title":"Dynamic neural program embedding for program repair","author":"Wang Ke","year":"2017","unstructured":"Ke Wang, Rishabh Singh, and Zhendong Su. 2017. Dynamic neural program embedding for program repair. arXiv preprint arXiv:1711.07163 (2017).","journal-title":"arXiv preprint arXiv:1711.07163"},{"key":"e_1_3_2_106_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2018.05.083"},{"key":"e_1_3_2_107_2","article-title":"Denoising neural machine translation training with trusted data and online data selection","author":"Wang Wei","year":"2018","unstructured":"Wei Wang, Taro Watanabe, Macduff Hughes, Tetsuji Nakagawa, and Ciprian Chelba. 2018. Denoising neural machine translation training with trusted data and online data selection. arXiv preprint arXiv:1809.00068 (2018).","journal-title":"arXiv preprint arXiv:1809.00068"},{"key":"e_1_3_2_108_2","doi-asserted-by":"publisher","DOI":"10.1109\/SANER.2019.8668043"},{"key":"e_1_3_2_109_2","doi-asserted-by":"publisher","DOI":"10.1145\/3540250.3549101"},{"key":"e_1_3_2_110_2","first-page":"10799","volume-title":"International Conference on Machine Learning","author":"Yasunaga Michihiro","year":"2020","unstructured":"Michihiro Yasunaga and Percy Liang. 2020. Graph-based, self-supervised program repair from diagnostic feedback. In International Conference on Machine Learning. PMLR, 10799\u201310808."},{"key":"e_1_3_2_111_2","first-page":"11941","volume-title":"International Conference on Machine Learning","author":"Yasunaga Michihiro","year":"2021","unstructured":"Michihiro Yasunaga and Percy Liang. 2021. Break-it-fix-it: Unsupervised learning for program repair. In International Conference on Machine Learning. PMLR, 11941\u201311952."},{"key":"e_1_3_2_112_2","article-title":"SelfAPR: Self-supervised program repair with test execution diagnostics","author":"Ye He","year":"2022","unstructured":"He Ye, Matias Martinez, Xiapu Luo, Tao Zhang, and Martin Monperrus. 2022. SelfAPR: Self-supervised program repair with test execution diagnostics. arXiv preprint arXiv:2203.12755 (2022).","journal-title":"arXiv preprint arXiv:2203.12755"},{"key":"e_1_3_2_113_2","doi-asserted-by":"publisher","DOI":"10.1145\/3510003.3510222"},{"key":"e_1_3_2_114_2","article-title":"Improving massively multilingual neural machine translation and zero-shot translation","author":"Zhang Biao","year":"2020","unstructured":"Biao Zhang, Philip Williams, Ivan Titov, and Rico Sennrich. 2020. Improving massively multilingual neural machine translation and zero-shot translation. arXiv preprint arXiv:2004.11867 (2020).","journal-title":"arXiv preprint arXiv:2004.11867"},{"key":"e_1_3_2_115_2","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i05.6513"},{"key":"e_1_3_2_116_2","doi-asserted-by":"publisher","DOI":"10.1145\/3468264.3468544"}],"container-title":["ACM Transactions on Software Engineering and Methodology"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3631972","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3631972","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T22:51:02Z","timestamp":1750287062000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3631972"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,3,15]]},"references-count":115,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2024,3,31]]}},"alternative-id":["10.1145\/3631972"],"URL":"https:\/\/doi.org\/10.1145\/3631972","relation":{},"ISSN":["1049-331X","1557-7392"],"issn-type":[{"value":"1049-331X","type":"print"},{"value":"1557-7392","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,3,15]]},"assertion":[{"value":"2022-12-21","order":0,"name":"received","label":"Received","group":{"name":"publication_history","label":"Publication History"}},{"value":"2023-10-19","order":1,"name":"accepted","label":"Accepted","group":{"name":"publication_history","label":"Publication History"}},{"value":"2024-03-15","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}