{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,1]],"date-time":"2026-03-01T10:13:12Z","timestamp":1772359992870,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":51,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,11,30]],"date-time":"2023-11-30T00:00:00Z","timestamp":1701302400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,11,30]]},"DOI":"10.1145\/3611643.3616302","type":"proceedings-article","created":{"date-parts":[[2023,11,30]],"date-time":"2023-11-30T23:14:38Z","timestamp":1701386078000},"page":"224-236","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":21,"title":["Towards Greener Yet Powerful Code Generation via Quantization: An Empirical Study"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-2749-7419","authenticated-orcid":false,"given":"Xiaokai","family":"Wei","sequence":"first","affiliation":[{"name":"AWS AI Labs, New York City, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4743-6461","authenticated-orcid":false,"given":"Sujan Kumar","family":"Gonugondla","sequence":"additional","affiliation":[{"name":"AWS AI Labs, New York City, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6338-1432","authenticated-orcid":false,"given":"Shiqi","family":"Wang","sequence":"additional","affiliation":[{"name":"AWS AI Labs, New York City, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8171-9583","authenticated-orcid":false,"given":"Wasi","family":"Ahmad","sequence":"additional","affiliation":[{"name":"AWS AI Labs, New York City, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3406-5235","authenticated-orcid":false,"given":"Baishakhi","family":"Ray","sequence":"additional","affiliation":[{"name":"AWS AI Labs, New York City, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7189-6903","authenticated-orcid":false,"given":"Haifeng","family":"Qian","sequence":"additional","affiliation":[{"name":"AWS AI Labs, New York City, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4916-1131","authenticated-orcid":false,"given":"Xiaopeng","family":"Li","sequence":"additional","affiliation":[{"name":"AWS AI Labs, New York City, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0000-3961-8439","authenticated-orcid":false,"given":"Varun","family":"Kumar","sequence":"additional","affiliation":[{"name":"AWS AI Labs, New York City, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4368-5092","authenticated-orcid":false,"given":"Zijian","family":"Wang","sequence":"additional","affiliation":[{"name":"AWS AI Labs, New York City, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0002-4952-721X","authenticated-orcid":false,"given":"Yuchen","family":"Tian","sequence":"additional","affiliation":[{"name":"AWS AI Labs, Santa Clara, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0002-0411-178X","authenticated-orcid":false,"given":"Qing","family":"Sun","sequence":"additional","affiliation":[{"name":"AWS AI Labs, Santa Clara, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2009-496X","authenticated-orcid":false,"given":"Ben","family":"Athiwaratkun","sequence":"additional","affiliation":[{"name":"AWS AI Labs, New York City, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0000-6523-3516","authenticated-orcid":false,"given":"Mingyue","family":"Shang","sequence":"additional","affiliation":[{"name":"AWS AI Labs, New York City, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6568-017X","authenticated-orcid":false,"given":"Murali Krishna","family":"Ramanathan","sequence":"additional","affiliation":[{"name":"AWS AI Labs, New York City, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0038-5081","authenticated-orcid":false,"given":"Parminder","family":"Bhatia","sequence":"additional","affiliation":[{"name":"AWS AI Labs, New York City, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0006-4028-4935","authenticated-orcid":false,"given":"Bing","family":"Xiang","sequence":"additional","affiliation":[{"name":"AWS AI Labs, New York City, USA"}]}],"member":"320","published-online":{"date-parts":[[2023,11,30]]},"reference":[{"key":"e_1_3_2_2_1_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_2_2_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_2_3_1","unstructured":"Jacob Austin Augustus Odena Maxwell Nye Maarten Bosma Henryk Michalewski David Dohan Ellen Jiang Carrie Cai Michael Terry Quoc Le and Charles Sutton. 2021. Program Synthesis with Large Language Models. arXiv e-prints Article arXiv:2108.07732 Aug. arXiv:2108.07732 pages. arxiv:2108.07732."},{"key":"e_1_3_2_2_4_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_2_5_1","volume-title":"Jared Kaplan, Harri Edwards, Yuri Burda, Nicholas Joseph, and Greg Brockman.","author":"Chen Mark","year":"2021","unstructured":"Mark Chen, Jerry Tworek, Heewoo Jun, Qiming Yuan, Henrique Ponde de Oliveira Pinto, Jared Kaplan, Harri Edwards, Yuri Burda, Nicholas Joseph, and Greg Brockman. 2021. Evaluating large language models trained on code. arXiv preprint arXiv:2107.03374."},{"key":"e_1_3_2_2_6_1","unstructured":"Mark Chen Jerry Tworek Heewoo Jun Qiming Yuan Henrique Ponde de Oliveira Pinto Jared Kaplan Harri Edwards Yuri Burda Nicholas Joseph Greg Brockman Alex Ray Raul Puri Gretchen Krueger Michael Petrov Heidy Khlaaf Girish Sastry Pamela Mishkin Brooke Chan Scott Gray Nick Ryder Mikhail Pavlov Alethea Power Lukasz Kaiser Mohammad Bavarian Clemens Winter Philippe Tillet Felipe Petroski Such Dave Cummings Matthias Plappert Fotios Chantzis Elizabeth Barnes Ariel Herbert-Voss William Hebgen Guss Alex Nichol Alex Paino Nikolas Tezak Jie Tang Igor Babuschkin Suchir Balaji Shantanu Jain William Saunders Christopher Hesse Andrew N. Carr Jan Leike Josh Achiam Vedant Misra Evan Morikawa Alec Radford Matthew Knight Miles Brundage Mira Murati Katie Mayer Peter Welinder Bob McGrew Dario Amodei Sam McCandlish Ilya Sutskever and Wojciech Zaremba. 2021. Evaluating Large Language Models Trained on Code. arXiv e-prints Article arXiv:2107.03374 July arXiv:2107.03374 pages. arxiv:2107.03374."},{"key":"e_1_3_2_2_7_1","unstructured":"Aakanksha Chowdhery Sharan Narang Jacob Devlin Maarten Bosma Gaurav Mishra Adam Roberts Paul Barham Hyung Won Chung Charles Sutton Sebastian Gehrmann Parker Schuh Kensen Shi Sasha Tsvyashchenko Joshua Maynez Abhishek Rao Parker Barnes Yi Tay Noam Shazeer Vinodkumar Prabhakaran Emily Reif Nan Du Ben Hutchinson Reiner Pope James Bradbury Jacob Austin Michael Isard Guy Gur-Ari Pengcheng Yin Toju Duke Anselm Levskaya Sanjay Ghemawat Sunipa Dev Henryk Michalewski Xavier Garcia Vedant Misra Kevin Robinson Liam Fedus Denny Zhou Daphne Ippolito David Luan Hyeontaek Lim Barret Zoph Alexander Spiridonov Ryan Sepassi David Dohan Shivani Agrawal Mark Omernick Andrew M. Dai Thanumalayan Sankaranarayana Pillai Marie Pellat Aitor Lewkowycz Erica Moreira Rewon Child Oleksandr Polozov Katherine Lee Zongwei Zhou Xuezhi Wang Brennan Saeta Mark Diaz Orhan Firat Michele Catasta Jason Wei Kathy Meier-Hellstern Douglas Eck Jeff Dean Slav Petrov and Noah Fiedel. 2022. PaLM: Scaling Language Modeling with Pathways. arXiv e-prints Article arXiv:2204.02311 April arXiv:2204.02311 pages. arxiv:2204.02311."},{"key":"e_1_3_2_2_8_1","volume-title":"Nl-augmenter: A framework for task-sensitive natural language augmentation. arXiv preprint arXiv:2112.02721.","author":"Dhole Kaustubh D","year":"2021","unstructured":"Kaustubh D Dhole, Varun Gangal, Sebastian Gehrmann, Aadesh Gupta, Zhenhao Li, Saad Mahamood, Abinaya Mahendiran, Simon Mille, Ashish Srivastava, and Samson Tan. 2021. Nl-augmenter: A framework for task-sensitive natural language augmentation. arXiv preprint arXiv:2112.02721."},{"key":"e_1_3_2_2_9_1","volume-title":"Hotflip: White-box adversarial examples for text classification. arXiv preprint arXiv:1712.06751.","author":"Ebrahimi Javid","year":"2017","unstructured":"Javid Ebrahimi, Anyi Rao, Daniel Lowd, and Dejing Dou. 2017. Hotflip: White-box adversarial examples for text classification. arXiv preprint arXiv:1712.06751."},{"key":"e_1_3_2_2_10_1","volume-title":"Wilcoxon-Mann-Whitney or t-test? On assumptions for hypothesis tests and multiple interpretations of decision rules. Statistics surveys, 4","author":"Fay Michael P","year":"2010","unstructured":"Michael P Fay and Michael A Proschan. 2010. Wilcoxon-Mann-Whitney or t-test? On assumptions for hypothesis tests and multiple interpretations of decision rules. Statistics surveys, 4 (2010), 1."},{"key":"e_1_3_2_2_11_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_2_12_1","unstructured":"Daniel Fried Armen Aghajanyan Jessy Lin Sida Wang Eric Wallace Freda Shi Ruiqi Zhong Wen-tau Yih Luke Zettlemoyer and Mike Lewis. 2022. InCoder: A Generative Model for Code Infilling and Synthesis. arXiv e-prints Article arXiv:2204.05999 April arXiv:2204.05999 pages. arxiv:2204.05999."},{"key":"e_1_3_2_2_13_1","doi-asserted-by":"publisher","DOI":"10.1109\/SPW.2018.00016"},{"key":"e_1_3_2_2_14_1","volume-title":"https:\/\/copilot.github.com Last accessed","author":"GitHub AI","year":"2022","unstructured":"OpenAI GitHub. 2022. GitHub Copilot. https:\/\/copilot.github.com Last accessed 31 August 2022"},{"key":"e_1_3_2_2_15_1","doi-asserted-by":"crossref","unstructured":"Karan Goel Nazneen Rajani Jesse Vig Samson Tan Jason Wu Stephan Zheng Caiming Xiong Mohit Bansal and Christopher R\u00e9. 2021. Robustness gym: Unifying the nlp evaluation landscape. arXiv preprint arXiv:2101.04840.","DOI":"10.18653\/v1\/2021.naacl-demos.6"},{"key":"e_1_3_2_2_16_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_2_17_1","article-title":"Towards the Systematic Reporting of the Energy and Carbon Footprints of Machine Learning","volume":"21","author":"Henderson Peter","year":"2020","unstructured":"Peter Henderson, Jieru Hu, Joshua Romoff, Emma Brunskill, Dan Jurafsky, and Joelle Pineau. 2020. Towards the Systematic Reporting of the Energy and Carbon Footprints of Machine Learning. J. Mach. Learn. Res., 21, 1 (2020), Article 248, jan, 43 pages. issn:1532-4435","journal-title":"J. Mach. Learn. Res."},{"key":"e_1_3_2_2_18_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_2_19_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i05.6311"},{"key":"e_1_3_2_2_20_1","volume-title":"Spoc: Search-based pseudocode to code. Advances in Neural Information Processing Systems, 32","author":"Kulal Sumith","year":"2019","unstructured":"Sumith Kulal, Panupong Pasupat, Kartik Chandra, Mina Lee, Oded Padon, Alex Aiken, and Percy S Liang. 2019. Spoc: Search-based pseudocode to code. Advances in Neural Information Processing Systems, 32 (2019)."},{"key":"e_1_3_2_2_21_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_2_22_1","volume-title":"Silvio Savarese, and Steven C. H. Hoi.","author":"Le Hung","year":"2022","unstructured":"Hung Le, Yue Wang, Akhilesh Deepak Gotmare, Silvio Savarese, and Steven C. H. Hoi. 2022. CodeRL: Mastering Code Generation through Pretrained Models and Deep Reinforcement Learning. arXiv e-prints, Article arXiv:2207.01780, July, arXiv:2207.01780 pages. arxiv:2207.01780."},{"key":"e_1_3_2_2_23_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_2_24_1","unstructured":"Zhenhao Li and Lucia Specia. 2019. Improving neural machine translation robustness via data augmentation: Beyond back translation. arXiv preprint arXiv:1910.03009."},{"key":"e_1_3_2_2_25_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.acl-short.22"},{"key":"e_1_3_2_2_26_1","doi-asserted-by":"publisher","DOI":"10.3115\/1220355.1220427"},{"key":"e_1_3_2_2_27_1","doi-asserted-by":"publisher","DOI":"10.1145\/3238147.3238202"},{"key":"e_1_3_2_2_28_1","doi-asserted-by":"publisher","DOI":"10.1145\/219717.219748"},{"key":"e_1_3_2_2_29_1","volume-title":"Jake Grigsby, Di Jin, and Yanjun Qi.","author":"Morris John X","year":"2020","unstructured":"John X Morris, Eli Lifland, Jin Yong Yoo, Jake Grigsby, Di Jin, and Yanjun Qi. 2020. Textattack: A framework for adversarial attacks, data augmentation, and adversarial training in nlp. arXiv preprint arXiv:2005.05909."},{"key":"e_1_3_2_2_30_1","unstructured":"Erik Nijkamp Bo Pang Hiroaki Hayashi Lifu Tu Huan Wang Yingbo Zhou Silvio Savarese and Caiming Xiong. 2022. A Conversational Paradigm for Program Synthesis. arXiv e-prints Article arXiv:2203.13474 March arXiv:2203.13474 pages. arxiv:2203.13474."},{"key":"e_1_3_2_2_31_1","unstructured":"David Patterson Joseph Gonzalez Quoc Le Chen Liang Lluis-Miquel Munguia Daniel Rothchild David So Maud Texier and Jeff Dean. 2021. Carbon Emissions and Large Neural Network Training. arXiv e-prints Article arXiv:2104.10350 April arXiv:2104.10350 pages. arxiv:2104.10350."},{"key":"e_1_3_2_2_32_1","volume-title":"Language models are unsupervised multitask learners. OpenAI blog, 1, 8","author":"Radford Alec","year":"2019","unstructured":"Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, and Ilya Sutskever. 2019. Language models are unsupervised multitask learners. OpenAI blog, 1, 8 (2019), 9."},{"key":"e_1_3_2_2_33_1","first-page":"1","article-title":"Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer","volume":"21","author":"Raffel Colin","year":"2020","unstructured":"Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, and Peter J. Liu. 2020. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. Journal of Machine Learning Research, 21, 140 (2020), 1\u201367. http:\/\/jmlr.org\/papers\/v21\/20-074.html","journal-title":"Journal of Machine Learning Research"},{"key":"e_1_3_2_2_34_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_2_35_1","first-page":"20601","article-title":"Unsupervised translation of programming languages","volume":"33","author":"Roziere Baptiste","year":"2020","unstructured":"Baptiste Roziere, Marie-Anne Lachaux, Lowik Chanussot, and Guillaume Lample. 2020. Unsupervised translation of programming languages. Advances in Neural Information Processing Systems, 33 (2020), 20601\u201320611.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_2_36_1","doi-asserted-by":"publisher","DOI":"10.5555\/3305890.3305992"},{"key":"e_1_3_2_2_37_1","doi-asserted-by":"publisher","DOI":"10.1145\/3381831"},{"key":"e_1_3_2_2_38_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-6504"},{"key":"e_1_3_2_2_39_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_2_40_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_2_41_1","doi-asserted-by":"publisher","DOI":"10.1145\/3180155.3180220"},{"key":"e_1_3_2_2_42_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_2_43_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_2_44_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_2_45_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_2_46_1","unstructured":"Guangxuan Xiao Ji Lin Mickael Seznec Hao Wu Julien Demouth and Song Han. 2023. SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models. In Proceedings of the 40th International Conference on Machine Learning Andreas Krause Emma Brunskill Kyunghyun Cho Barbara Engelhardt Sivan Sabato and Jonathan Scarlett (Eds.) (Proceedings of Machine Learning Research Vol. 202). 38087\u201338099. https:\/\/proceedings.mlr.press\/v202\/xiao23c.html"},{"key":"e_1_3_2_2_47_1","volume-title":"Minjia Zhang, Xiaoxia Wu, Conglong Li, and Yuxiong He.","author":"Yao Zhewei","year":"2022","unstructured":"Zhewei Yao, Reza Yazdani Aminabadi, Minjia Zhang, Xiaoxia Wu, Conglong Li, and Yuxiong He. 2022. ZeroQuant: Efficient and Affordable Post-Training Quantization for Large-Scale Transformers. In Advances in Neural Information Processing Systems, S. Koyejo, S. Mohamed, A. Agarwal, D. Belgrave, K. Cho, and A. Oh (Eds.). 35, 27168\u201327183. https:\/\/proceedings.neurips.cc\/paper_files\/paper\/2022\/file\/adf7fa39d65e2983d724ff7da57f00ac-Paper-Conference.pdf"},{"key":"e_1_3_2_2_48_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_2_49_1","doi-asserted-by":"publisher","DOI":"10.1109\/TSE.2019.2962027"},{"key":"e_1_3_2_2_50_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_2_51_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.findings-acl.315"}],"event":{"name":"ESEC\/FSE '23: 31st ACM Joint European Software Engineering Conference and Symposium on the Foundations of Software Engineering","location":"San Francisco CA USA","acronym":"ESEC\/FSE '23","sponsor":["SIGSOFT ACM Special Interest Group on Software Engineering"]},"container-title":["Proceedings of the 31st ACM Joint European Software Engineering Conference and Symposium on the Foundations of Software Engineering"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3611643.3616302","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3611643.3616302","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T16:36:04Z","timestamp":1750178164000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3611643.3616302"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,11,30]]},"references-count":51,"alternative-id":["10.1145\/3611643.3616302","10.1145\/3611643"],"URL":"https:\/\/doi.org\/10.1145\/3611643.3616302","relation":{},"subject":[],"published":{"date-parts":[[2023,11,30]]},"assertion":[{"value":"2023-11-30","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}