{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T15:59:11Z","timestamp":1776095951913,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":41,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,4,20]],"date-time":"2024-04-20T00:00:00Z","timestamp":1713571200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,4,20]]},"DOI":"10.1145\/3643795.3648388","type":"proceedings-article","created":{"date-parts":[[2024,9,10]],"date-time":"2024-09-10T13:46:19Z","timestamp":1725975979000},"page":"46-53","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":7,"title":["Translation of Low-Resource COBOL to Logically Correct and Readable Java leveraging High-Resource Java Refinement"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-8842-6895","authenticated-orcid":false,"given":"Shubham","family":"Gandhi","sequence":"first","affiliation":[{"name":"TCS Research, New Delhi, India"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2775-3497","authenticated-orcid":false,"given":"Manasi","family":"Patwardhan","sequence":"additional","affiliation":[{"name":"TCS Research, New Delhi, India"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8519-661X","authenticated-orcid":false,"given":"Jyotsana","family":"Khatri","sequence":"additional","affiliation":[{"name":"TCS Research, New Delhi, India"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9834-3308","authenticated-orcid":false,"given":"Lovekesh","family":"Vig","sequence":"additional","affiliation":[{"name":"TCS Research, New Delhi, India"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9162-4825","authenticated-orcid":false,"given":"Raveendra Kumar","family":"Medicherla","sequence":"additional","affiliation":[{"name":"TCS Research, New Delhi, India"}]}],"member":"320","published-online":{"date-parts":[[2024,9,10]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"Karan Aggarwal et al. 2015. Using machine translation for converting python 2 to python 3 code. Technical Report. PeerJ PrePrints."},{"key":"e_1_3_2_1_2_1","volume-title":"Back-translate: Unsupervised Translation of Programming Languages. arXiv:2205.11116 [cs.CL]","author":"Ahmad Wasi Uddin","year":"2023","unstructured":"Wasi Uddin Ahmad et al. 2023. Summarize and Generate to Back-translate: Unsupervised Translation of Programming Languages. arXiv:2205.11116 [cs.CL]"},{"key":"e_1_3_2_1_3_1","volume-title":"Proceedings of the 37th IEEE\/ACM International Conference on Automated Software Engineering. 1--5.","author":"Madi Naser Al","year":"2022","unstructured":"Naser Al Madi. 2022. How readable is model-generated code? examining readability and visual inspection of github copilot. In Proceedings of the 37th IEEE\/ACM International Conference on Automated Software Engineering. 1--5."},{"key":"e_1_3_2_1_4_1","volume-title":"Multi-lingual Evaluation of Code Generation Models. In The Eleventh International Conference on Learning Representations.","author":"Ben","unstructured":"Ben Athiwaratkun et al. 2022. Multi-lingual Evaluation of Code Generation Models. In The Eleventh International Conference on Learning Representations."},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1109\/TSE.2009.70"},{"key":"e_1_3_2_1_6_1","unstructured":"Angelica Chen et al. 2023. Improving Code Generation by Training with Natural Language Feedback. arXiv:2303.16749 [cs.SE]"},{"key":"e_1_3_2_1_7_1","unstructured":"Mark Chen et al. 2021. Evaluating Large Language Models Trained on Code. arXiv e-prints (2021) arXiv-2107."},{"key":"e_1_3_2_1_8_1","unstructured":"Xinyun Chen et al. 2018. Tree-to-tree neural networks for program translation. Advances in neural information processing systems 31 (2018)."},{"key":"e_1_3_2_1_9_1","unstructured":"Aakanksha Chowdhery et al. 2022. PaLM: Scaling Language Modeling with Pathways. arXiv:2204.02311 [cs.CL]"},{"key":"e_1_3_2_1_10_1","unstructured":"Jonathan Dorn. 2012. A general software readability model. (2012)."},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"crossref","unstructured":"Jessica Ficler and Yoav Goldberg. 2017. Controlling Linguistic Style Aspects in Neural Language Generation. arXiv:1707.02633 [cs.CL]","DOI":"10.18653\/v1\/W17-4912"},{"key":"e_1_3_2_1_12_1","unstructured":"Jie Huang et al. 2023. Large Language Models Cannot Self-Correct Reasoning Yet. (2023). arXiv:2310.01798 [cs.CL]"},{"key":"e_1_3_2_1_13_1","unstructured":"Xue Jiang et al. 2023. Self-planning Code Generation with Large Language Model. arXiv e-prints (2023) arXiv-2303."},{"key":"e_1_3_2_1_14_1","first-page":"14967","article-title":"DOBF: A deobfuscation pre-training objective for programming languages","volume":"34","author":"Lachaux Marie-Anne","year":"2021","unstructured":"Marie-Anne Lachaux et al. 2021. DOBF: A deobfuscation pre-training objective for programming languages. Advances in Neural Information Processing Systems 34 (2021), 14967--14979.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_15_1","volume-title":"Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing. 5039--5049","author":"Guillaume","unstructured":"Guillaume Lample et al. 2018. Phrase-Based & Neural Unsupervised Machine Translation. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing. 5039--5049."},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"crossref","unstructured":"Jia Li et al. 2023. Structured Chain-of-Thought Prompting for Code Generation. arXiv:2305.06599 [cs.SE]","DOI":"10.1145\/3690635"},{"key":"e_1_3_2_1_17_1","unstructured":"Raymond Li et al. 2023. StarCoder: may the source be with you! arXiv:2305.06161 [cs.CL]"},{"key":"e_1_3_2_1_18_1","unstructured":"Xiping Liu and Zhao Tan. 2023. Divide and Prompt: Chain of Thought Prompting for Text-to-SQL. arXiv:2304.11556 [cs.CL]"},{"key":"e_1_3_2_1_19_1","unstructured":"Ziyang Luo et al. 2023. WizardCoder: Empowering Code Large Language Models with Evol-Instruct. arXiv:2306.08568 [cs.CL]"},{"key":"e_1_3_2_1_20_1","volume-title":"Self-refine: Iterative refinement with self-feedback. arXiv preprint arXiv:2303.17651","author":"Aman Madaan","year":"2023","unstructured":"Aman Madaan et al. 2023. Self-refine: Iterative refinement with self-feedback. arXiv preprint arXiv:2303.17651 (2023)."},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"crossref","unstructured":"Aniketh Malyala et al. 2023. On ML-Based Program Translation: Perils and Promises. arXiv preprint arXiv:2302.10812 (2023).","DOI":"10.1109\/ICSE-NIER58687.2023.00017"},{"key":"e_1_3_2_1_22_1","unstructured":"Erik Nijkamp et al. 2023. CodeGen: An Open Large Language Model for Code with Multi-Turn Program Synthesis. arXiv:2203.13474 [cs.LG]"},{"key":"e_1_3_2_1_24_1","unstructured":"Rangeet Pan et al. 2023. Understanding the Effectiveness of Large Language Models in Code Translation. arXiv e-prints (2023) arXiv-2308."},{"key":"e_1_3_2_1_25_1","volume-title":"Proceedings of the 8th working conference on mining software repositories. 73--82","author":"Daryl","unstructured":"Daryl Posnett et al. 2011. A simpler model of software readability. In Proceedings of the 8th working conference on mining software repositories. 73--82."},{"key":"e_1_3_2_1_26_1","volume-title":"Thirty-fifth Conference on Neural Information Processing Systems Datasets and Benchmarks Track (Round 2).","author":"Ruchir","unstructured":"Ruchir Puri et al. 2021. CodeNet: A Large-Scale AI for Code Dataset for Learning a Diversity of Coding Tasks. In Thirty-fifth Conference on Neural Information Processing Systems Datasets and Benchmarks Track (Round 2)."},{"key":"e_1_3_2_1_27_1","first-page":"20601","article-title":"Unsupervised translation of programming languages","volume":"33","author":"Baptiste Roziere","year":"2020","unstructured":"Baptiste Roziere et al. 2020. Unsupervised translation of programming languages. Advances in Neural Information Processing Systems 33 (2020), 20601--20611.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_28_1","unstructured":"Baptiste Roziere et al. 2021. Leveraging automated unit tests for unsupervised code translation. arXiv preprint arXiv:2110.06773 (2021)."},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1002\/smr.1958"},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1109\/TSE.2019.2901468"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1109\/CSMR.2013.32"},{"key":"e_1_3_2_1_32_1","unstructured":"Hugo Touvron et al. 2023. Llama 2: Open Foundation and Fine-Tuned Chat Models. arXiv:2307.09288 [cs.CL]"},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.5120\/13734-1532"},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.blackboxnlp-1.31"},{"key":"e_1_3_2_1_35_1","unstructured":"Junjie Wang et al. 2023. Software Testing with Large Language Model: Survey Landscape and Vision. arXiv:2307.07221 [cs.SE]"},{"key":"e_1_3_2_1_36_1","unstructured":"Xuezhi Wang et al. 2023. Self-Consistency Improves Chain of Thought Reasoning in Language Models. (2023). arXiv:2203.11171 [cs.CL]"},{"key":"e_1_3_2_1_37_1","unstructured":"Jason Wei et al. 2023. Chain-of-Thought Prompting Elicits Reasoning in Large Language Models. arXiv:2201.11903 [cs.CL]"},{"key":"e_1_3_2_1_38_1","unstructured":"Chunqiu Steven Xia et al. 2022. Practical Program Repair in the Era of Large Pre-trained Language Models. arXiv:2210.14179 [cs.SE]"},{"key":"e_1_3_2_1_39_1","unstructured":"Chunqiu Steven Xia and Lingming Zhang. 2023. Conversational Automated Program Repair. arXiv:2301.13246 [cs.SE]"},{"key":"e_1_3_2_1_40_1","unstructured":"Michihiro Yasunaga and Percy Liang. 2020. Graph-based Self-Supervised Program Repair from Diagnostic Feedback. arXiv:2005.10636 [cs.SE]"},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"crossref","unstructured":"Qinkai Zheng et al. 2023. CodeGeeX: A Pre-Trained Model for Code Generation with Multilingual Evaluations on HumanEval-X. arXiv:2303.17568 [cs.LG]","DOI":"10.1145\/3580305.3599790"},{"key":"e_1_3_2_1_42_1","unstructured":"Yuqi Zhu et al. 2023. Improving Code Generation by Dynamic Temperature Sampling. arXiv preprint arXiv:2309.02772 (2023)."}],"event":{"name":"LLM4Code '24: 1st International Workshop on Large Language Models for Code","location":"Lisbon Portugal","acronym":"LLM4Code '24","sponsor":["SIGSOFT ACM Special Interest Group on Software Engineering","IEEE CS","Faculty of Engineering of University of Porto"]},"container-title":["Proceedings of the 1st International Workshop on Large Language Models for Code"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3643795.3648388","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3643795.3648388","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T00:57:45Z","timestamp":1750294665000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3643795.3648388"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,4,20]]},"references-count":41,"alternative-id":["10.1145\/3643795.3648388","10.1145\/3643795"],"URL":"https:\/\/doi.org\/10.1145\/3643795.3648388","relation":{},"subject":[],"published":{"date-parts":[[2024,4,20]]},"assertion":[{"value":"2024-09-10","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}