{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,27]],"date-time":"2025-10-27T12:57:33Z","timestamp":1761569853940,"version":"build-2065373602"},"publisher-location":"New York, NY, USA","reference-count":45,"publisher":"ACM","funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62272473"],"award-info":[{"award-number":["62272473"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["U2441238"],"award-info":[{"award-number":["U2441238"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62202474"],"award-info":[{"award-number":["62202474"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"the Science and Technology Innovation Program of Hunan Province","doi-asserted-by":"publisher","award":["2023RC1001"],"award-info":[{"award-number":["2023RC1001"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,6,20]]},"DOI":"10.1145\/3755881.3755884","type":"proceedings-article","created":{"date-parts":[[2025,10,27]],"date-time":"2025-10-27T11:46:17Z","timestamp":1761565577000},"page":"437-448","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["MetaCoder: Generating Code from Multiple Perspectives"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0005-9264-5272","authenticated-orcid":false,"given":"Xin","family":"Chen","sequence":"first","affiliation":[{"name":"National University of Defense Technology, Changsha, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7076-4239","authenticated-orcid":false,"given":"Zhijie","family":"Jiang","sequence":"additional","affiliation":[{"name":"National University of Defense Technology, Changsha, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5903-5302","authenticated-orcid":false,"given":"Yong","family":"Guo","sequence":"additional","affiliation":[{"name":"National University of Defense Technology, Changsha, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2533-4547","authenticated-orcid":false,"given":"Zhouyang","family":"Jia","sequence":"additional","affiliation":[{"name":"National University of Defense Technology, Changsha, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0002-6363-859X","authenticated-orcid":false,"given":"Si","family":"Zheng","sequence":"additional","affiliation":[{"name":"National University of Defense Technology, Changsha, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0006-6042-7489","authenticated-orcid":false,"given":"Yuanliang","family":"Zhang","sequence":"additional","affiliation":[{"name":"National University of Defense Technology, Changsha, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0798-974X","authenticated-orcid":false,"given":"Shanshan","family":"Li","sequence":"additional","affiliation":[{"name":"National University of Defense Technology, Changsha, China"}]}],"member":"320","published-online":{"date-parts":[[2025,10,27]]},"reference":[{"key":"e_1_3_3_1_2_2","unstructured":"2025. ChatGPT. https:\/\/chatgpt.com\/ Accessed: 2025-2-26."},{"key":"e_1_3_3_1_3_2","unstructured":"2025. Claude 3.5 Sonnet. https:\/\/www.anthropic.com\/news\/claude-3-5-sonnet Accessed: 2025-2-26."},{"key":"e_1_3_3_1_4_2","unstructured":"2025. Copilot. https:\/\/copilot.microsoft.com\/ Accessed: 2025-2-26."},{"key":"e_1_3_3_1_5_2","unstructured":"2025. Cursor. https:\/\/www.cursor.com\/ Accessed: 2025-2-26."},{"key":"e_1_3_3_1_6_2","unstructured":"2025. HackerRank. https:\/\/www.hackerrank.com\/ Accessed: 2025-2-26."},{"key":"e_1_3_3_1_7_2","unstructured":"2025. LeetCode. https:\/\/leetcode.com\/ Accessed: 2025-2-26."},{"key":"e_1_3_3_1_8_2","unstructured":"Marah Abdin Jyoti Aneja Hany Awadalla Ahmed Awadallah Ammar\u00a0Ahmad Awan Nguyen Bach Amit Bahree Arash Bakhtiari Jianmin Bao Harkirat Behl et\u00a0al. 2024. Phi-3 technical report: A highly capable language model locally on your phone. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2404.14219 (2024)."},{"key":"e_1_3_3_1_9_2","unstructured":"Josh Achiam Steven Adler Sandhini Agarwal Lama Ahmad Ilge Akkaya Florencia\u00a0Leoni Aleman Diogo Almeida Janko Altenschmidt Sam Altman Shyamal Anadkat et\u00a0al. 2023. Gpt-4 technical report. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2303.08774 (2023)."},{"key":"e_1_3_3_1_10_2","unstructured":"Ben Athiwaratkun Sanjay\u00a0Krishna Gouda Zijian Wang Xiaopeng Li Yuchen Tian Ming Tan Wasi\u00a0Uddin Ahmad Shiqi Wang Qing Sun Mingyue Shang et\u00a0al. 2022. Multi-lingual evaluation of code generation models. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2210.14868 (2022)."},{"key":"e_1_3_3_1_11_2","doi-asserted-by":"crossref","unstructured":"Ramakrishna Bairi Atharv Sonwane Aditya Kanade Arun Iyer Suresh Parthasarathy Sriram Rajamani B Ashok and Shashank Shet. 2024. Codeplan: Repository-level coding using llms and planning. Proceedings of the ACM on Software Engineering 1 FSE (2024) 675\u2013698.","DOI":"10.1145\/3643757"},{"key":"e_1_3_3_1_12_2","doi-asserted-by":"crossref","unstructured":"Federico Cassano John Gouwar Francesca Lucchetti Claire Schlesinger Anders Freeman Carolyn\u00a0Jane Anderson Molly\u00a0Q Feldman Michael Greenberg Abhinav Jangda and Arjun Guha. 2024. Knowledge transfer from high-resource to low-resource programming languages for code llms. Proceedings of the ACM on Programming Languages 8 OOPSLA2 (2024) 677\u2013708.","DOI":"10.1145\/3689735"},{"key":"e_1_3_3_1_13_2","unstructured":"Angelica Chen J\u00e9r\u00e9my Scheurer Tomasz Korbak Jon\u00a0Ander Campos Jun\u00a0Shern Chan Samuel\u00a0R Bowman Kyunghyun Cho and Ethan Perez. 2023. Improving code generation by training with natural language feedback. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2303.16749 (2023)."},{"key":"e_1_3_3_1_14_2","unstructured":"Bei Chen Fengji Zhang Anh Nguyen Daoguang Zan Zeqi Lin Jian-Guang Lou and Weizhu Chen. 2022. Codet: Code generation with generated tests. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2207.10397 (2022)."},{"key":"e_1_3_3_1_15_2","unstructured":"Jiawei Chen Wentao Chen Jing Su Jingjing Xu Hongyu Lin Mengjie Ren Yaojie Lu Xianpei Han and Le Sun. 2024. The Rise and Down of Babel Tower: Investigating the Evolution Process of Multilingual Code Large Language Model. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2412.07298 (2024)."},{"key":"e_1_3_3_1_16_2","unstructured":"Mark Chen Jerry Tworek Heewoo Jun Qiming Yuan Henrique Ponde De\u00a0Oliveira Pinto Jared Kaplan Harri Edwards Yuri Burda Nicholas Joseph Greg Brockman et\u00a0al. 2021. Evaluating large language models trained on code. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2107.03374 (2021)."},{"key":"e_1_3_3_1_17_2","doi-asserted-by":"crossref","unstructured":"Simin Chen Zexin Li Wei Yang and Cong Liu. 2024. DeciX: Explain Deep Learning Based Code Generation Applications. Proceedings of the ACM on Software Engineering 1 FSE (2024) 2424\u20132446.","DOI":"10.1145\/3660814"},{"key":"e_1_3_3_1_18_2","unstructured":"Xinyun Chen Maxwell Lin Nathanael Sch\u00e4rli and Denny Zhou. 2023. Teaching large language models to self-debug. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2304.05128 (2023)."},{"key":"e_1_3_3_1_19_2","doi-asserted-by":"crossref","unstructured":"Yihong Dong Xue Jiang Zhi Jin and Ge Li. 2024. Self-collaboration code generation via chatgpt. ACM Transactions on Software Engineering and Methodology 33 7 (2024) 1\u201338.","DOI":"10.1145\/3672459"},{"key":"e_1_3_3_1_20_2","unstructured":"Abhimanyu Dubey Abhinav Jauhri Abhinav Pandey Abhishek Kadian Ahmad Al-Dahle Aiesha Letman Akhil Mathur Alan Schelten Amy Yang Angela Fan et\u00a0al. 2024. The llama 3 herd of models. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2407.21783 (2024)."},{"key":"e_1_3_3_1_21_2","doi-asserted-by":"crossref","unstructured":"Zhangyin Feng Daya Guo Duyu Tang Nan Duan Xiaocheng Feng Ming Gong Linjun Shou Bing Qin Ting Liu Daxin Jiang et\u00a0al. 2020. Codebert: A pre-trained model for programming and natural languages. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2002.08155 (2020).","DOI":"10.18653\/v1\/2020.findings-emnlp.139"},{"key":"e_1_3_3_1_22_2","unstructured":"Daniel Fried Armen Aghajanyan Jessy Lin Sida Wang Eric Wallace Freda Shi Ruiqi Zhong Wen-tau Yih Luke Zettlemoyer and Mike Lewis. 2022. Incoder: A generative model for code infilling and synthesis. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2204.05999 (2022)."},{"key":"e_1_3_3_1_23_2","unstructured":"Daya Guo Dejian Yang Haowei Zhang Junxiao Song Ruoyu Zhang Runxin Xu Qihao Zhu Shirong Ma Peiyi Wang Xiao Bi et\u00a0al. 2025. Deepseek-r1: Incentivizing reasoning capability in llms via reinforcement learning. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2501.12948 (2025)."},{"key":"e_1_3_3_1_24_2","unstructured":"Takeshi Kojima Shixiang\u00a0Shane Gu Machel Reid Yutaka Matsuo and Yusuke Iwasawa. 2022. Large language models are zero-shot reasoners. Advances in neural information processing systems 35 (2022) 22199\u201322213."},{"key":"e_1_3_3_1_25_2","unstructured":"Hung Le Hailin Chen Amrita Saha Akash Gokul Doyen Sahoo and Shafiq Joty. 2023. Codechain: Towards modular code generation through chain of self-revisions with representative sub-modules. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2310.08992 (2023)."},{"key":"e_1_3_3_1_26_2","doi-asserted-by":"crossref","unstructured":"Bolun Li Zhihong Sun Tao Huang Hongyu Zhang Yao Wan Ge Li Zhi Jin and Chen Lyu. 2024. Ircoco: Immediate rewards-guided deep reinforcement learning for code completion. Proceedings of the ACM on Software Engineering 1 FSE (2024) 182\u2013203.","DOI":"10.1145\/3643735"},{"key":"e_1_3_3_1_27_2","doi-asserted-by":"crossref","unstructured":"Jia Li Ge Li Yongmin Li and Zhi Jin. 2025. Structured chain-of-thought prompting for code generation. ACM Transactions on Software Engineering and Methodology 34 2 (2025) 1\u201323.","DOI":"10.1145\/3690635"},{"key":"e_1_3_3_1_28_2","unstructured":"Tian Liang Zhiwei He Wenxiang Jiao Xing Wang Yan Wang Rui Wang Yujiu Yang Shuming Shi and Zhaopeng Tu. 2023. Encouraging divergent thinking in large language models through multi-agent debate. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2305.19118 (2023)."},{"key":"e_1_3_3_1_29_2","doi-asserted-by":"publisher","unstructured":"Yingwei Ma Yue Liu Yue Yu Yuanliang Zhang Yu Jiang Changjian Wang and Shanshan Li. 2023. At which training stage does code data help llms reasoning? arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2309.16298 (2023). 10.48550\/arXiv.2309.16298","DOI":"10.48550\/arXiv.2309.16298"},{"key":"e_1_3_3_1_30_2","unstructured":"Aman Madaan Niket Tandon Prakhar Gupta Skyler Hallinan Luyu Gao Sarah Wiegreffe Uri Alon Nouha Dziri Shrimai Prabhumoye Yiming Yang et\u00a0al. 2023. Self-refine: Iterative refinement with self-feedback. Advances in Neural Information Processing Systems 36 (2023) 46534\u201346594."},{"key":"e_1_3_3_1_31_2","unstructured":"Fangwen Mu Lin Shi Song Wang Zhuohao Yu Binquan Zhang Chenxue Wang Shichao Liu and Qing Wang. 2023. Clarifygpt: Empowering llm-based code generation with intention clarification. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2310.10996 (2023)."},{"key":"e_1_3_3_1_32_2","volume-title":"NeurIPS 2023 Workshop on Instruction Tuning and Instruction Following","author":"Muennighoff Niklas","year":"2023","unstructured":"Niklas Muennighoff, Qian Liu, Armel Zebaze, Qinkai Zheng, Binyuan Hui, Terry\u00a0Yue Zhuo, Swayam Singh, Xiangru Tang, Leandro Von\u00a0Werra, and Shayne Longpre. 2023. Octopack: Instruction tuning code large language models. In NeurIPS 2023 Workshop on Instruction Tuning and Instruction Following."},{"key":"e_1_3_3_1_33_2","unstructured":"Erik Nijkamp Bo Pang Hiroaki Hayashi Lifu Tu Huan Wang Yingbo Zhou Silvio Savarese and Caiming Xiong. 2022. Codegen: An open large language model for code with multi-turn program synthesis. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2203.13474 (2022)."},{"key":"e_1_3_3_1_34_2","unstructured":"Theo\u00a0X Olausson Jeevana\u00a0Priya Inala Chenglong Wang Jianfeng Gao and Armando Solar-Lezama. 2023. Is self-repair a silver bullet for code generation? arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2306.09896 (2023)."},{"key":"e_1_3_3_1_35_2","first-page":"311","volume-title":"Proceedings of the 40th annual meeting of the Association for Computational Linguistics","author":"Papineni Kishore","year":"2002","unstructured":"Kishore Papineni, Salim Roukos, Todd Ward, and Wei-Jing Zhu. 2002. Bleu: a method for automatic evaluation of machine translation. In Proceedings of the 40th annual meeting of the Association for Computational Linguistics. 311\u2013318."},{"key":"e_1_3_3_1_36_2","unstructured":"Chen Qian Xin Cong Cheng Yang Weize Chen Yusheng Su Juyuan Xu Zhiyuan Liu and Maosong Sun. 2023. Communicative agents for software development. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2307.07924 6 3 (2023)."},{"key":"e_1_3_3_1_37_2","unstructured":"Noah Shinn Federico Cassano Ashwin Gopinath Karthik Narasimhan and Shunyu Yao. 2023. Reflexion: Language agents with verbal reinforcement learning. Advances in Neural Information Processing Systems 36 (2023) 8634\u20138652."},{"key":"e_1_3_3_1_38_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.acl-long.100"},{"key":"e_1_3_3_1_39_2","unstructured":"Gemma Team Morgane Riviere Shreya Pathak Pier\u00a0Giuseppe Sessa Cassidy Hardin Surya Bhupatiraju L\u00e9onard Hussenot Thomas Mesnard Bobak Shahriari Alexandre Ram\u00e9 et\u00a0al. 2024. Gemma 2: Improving open language models at a practical size. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2408.00118 (2024)."},{"key":"e_1_3_3_1_40_2","unstructured":"Hanbin Wang Zhenghao Liu Shuo Wang Ganqu Cui Ning Ding Zhiyuan Liu and Ge Yu. 2023. Intervenor: Prompting the coding ability of large language models with the interactive chain of repair. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2311.09868 (2023)."},{"key":"e_1_3_3_1_41_2","unstructured":"Jason Wei Xuezhi Wang Dale Schuurmans Maarten Bosma Fei Xia Ed Chi Quoc\u00a0V Le Denny Zhou et\u00a0al. 2022. Chain-of-thought prompting elicits reasoning in large language models. Advances in neural information processing systems 35 (2022) 24824\u201324837."},{"key":"e_1_3_3_1_42_2","unstructured":"An Yang Baosong Yang Beichen Zhang Binyuan Hui Bo Zheng Bowen Yu Chengyuan Li Dayiheng Liu Fei Huang Haoran Wei et\u00a0al. 2024. Qwen2. 5 technical report. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2412.15115 (2024)."},{"key":"e_1_3_3_1_43_2","doi-asserted-by":"crossref","unstructured":"Zhen Yang Fang Liu Zhongxing Yu Jacky\u00a0Wai Keung Jia Li Shuo Liu Yifan Hong Xiaoxue Ma Zhi Jin and Ge Li. 2024. Exploring and unleashing the power of large language models in automated code translation. Proceedings of the ACM on Software Engineering 1 FSE (2024) 1585\u20131608.","DOI":"10.1145\/3660778"},{"key":"e_1_3_3_1_44_2","doi-asserted-by":"crossref","unstructured":"Kechi Zhang Zhuo Li Jia Li Ge Li and Zhi Jin. 2023. Self-edit: Fault-aware code editor for code generation. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2305.04087 (2023).","DOI":"10.18653\/v1\/2023.acl-long.45"},{"key":"e_1_3_3_1_45_2","doi-asserted-by":"publisher","DOI":"10.1145\/3580305.3599790"},{"key":"e_1_3_3_1_46_2","unstructured":"Qihao Zhu Daya Guo Zhihong Shao Dejian Yang Peiyi Wang Runxin Xu Y Wu Yukun Li Huazuo Gao Shirong Ma et\u00a0al. 2024. Deepseek-coder-v2: Breaking the barrier of closed-source models in code intelligence. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2406.11931 (2024)."}],"event":{"name":"Internetware 2025: the 16th International Conference on Internetware","location":"Trondheim Norway","acronym":"Internetware 2025","sponsor":["SIGSOFT ACM Special Interest Group on Artificial Intelligence"]},"container-title":["Proceedings of the 16th International Conference on Internetware"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3755881.3755884","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,27]],"date-time":"2025-10-27T11:46:41Z","timestamp":1761565601000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3755881.3755884"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,6,20]]},"references-count":45,"alternative-id":["10.1145\/3755881.3755884","10.1145\/3755881"],"URL":"https:\/\/doi.org\/10.1145\/3755881.3755884","relation":{},"subject":[],"published":{"date-parts":[[2025,6,20]]},"assertion":[{"value":"2025-10-27","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}