{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,4]],"date-time":"2026-05-04T10:20:12Z","timestamp":1777890012401,"version":"3.51.4"},"publisher-location":"New York, NY, USA","reference-count":33,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,4,20]],"date-time":"2024-04-20T00:00:00Z","timestamp":1713571200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,4,20]]},"DOI":"10.1145\/3643795.3648394","type":"proceedings-article","created":{"date-parts":[[2024,9,10]],"date-time":"2024-09-10T13:46:19Z","timestamp":1725975979000},"page":"30-37","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":16,"title":["Learn to Code Sustainably: An Empirical Study on Green Code Generation"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0009-5300-2572","authenticated-orcid":false,"given":"Tina","family":"Vartziotis","sequence":"first","affiliation":[{"name":"TWT Science and Innovation, Stuttgart, Germany"},{"name":"NTUA, Athens, Greece"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3431-8450","authenticated-orcid":false,"given":"Ippolyti","family":"Dellatolas","sequence":"additional","affiliation":[{"name":"MIT, Cambridge, Massachusetts, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0562-5136","authenticated-orcid":false,"given":"George","family":"Dasoulas","sequence":"additional","affiliation":[{"name":"Harvard University, Cambridge, Massachusetts, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0007-9472-7022","authenticated-orcid":false,"given":"Maximilian","family":"Schmidt","sequence":"additional","affiliation":[{"name":"TWT Science and Innovation, Stuttgart, Germany"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3737-6215","authenticated-orcid":false,"given":"Florian","family":"Schneider","sequence":"additional","affiliation":[{"name":"TWT Science and Innovation, Stuttgart, Germany"}]},{"ORCID":"https:\/\/orcid.org\/0009-0000-6476-6978","authenticated-orcid":false,"given":"Tim","family":"Hoffmann","sequence":"additional","affiliation":[{"name":"Mercedes-Benz, Stuttgart, Germany"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2709-752X","authenticated-orcid":false,"given":"Sotirios","family":"Kotsopoulos","sequence":"additional","affiliation":[{"name":"National Technical University of Athens, Massachusetts Institute of Technology, Stuttgart, Germany"}]},{"ORCID":"https:\/\/orcid.org\/0009-0000-2214-9445","authenticated-orcid":false,"given":"Michael","family":"Keckeisen","sequence":"additional","affiliation":[{"name":"TWT Science and Innovation, Stuttgart, Germany"}]}],"member":"320","published-online":{"date-parts":[[2024,9,10]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1145\/3630614.3630619"},{"key":"e_1_3_2_1_2_1","unstructured":"Amodei and Hernandez. 2018. . https:\/\/openai.com\/research\/ai-and-compute"},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","unstructured":"Mahmoud Assran Joshua Romoff Nicolas Ballas Joelle Pineau and Michael Rabbat. 2019. Gossip-based Actor-Learner Architectures for Deep Reinforcement Learning. (2019). 10.48550\/ARXIV.1906.04585","DOI":"10.48550\/ARXIV.1906.04585"},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","unstructured":"Alfredo Canziani Adam Paszke and Eugenio Culurciello. 2016. An Analysis of Deep Neural Network Models for Practical Applications. 10.48550\/ARXIV.1605.07678","DOI":"10.48550\/ARXIV.1605.07678"},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.5555\/3495724.3497383"},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1145\/1840845.1840883"},{"key":"e_1_3_2_1_7_1","unstructured":"Arnaldo Carvalho de Melo and Red Hat. 2010. The New Linux' perf' Tools. https:\/\/api.semanticscholar.org\/CorpusID:10296207"},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.suscom.2023.100857"},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.suscom.2023.100857"},{"key":"e_1_3_2_1_10_1","volume-title":"execute, assess: Program synthesis with a repl. Advances in Neural Information Processing Systems 32","author":"Ellis Kevin","year":"2019","unstructured":"Kevin Ellis, Maxwell Nye, Yewen Pu, Felix Sosa, Josh Tenenbaum, and Armando Solar-Lezama. 2019. Write, execute, assess: Program synthesis with a repl. Advances in Neural Information Processing Systems 32 (2019)."},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1145\/1836089.1836091"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1561\/2500000010"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","unstructured":"Peter Henderson Jieru Hu Joshua Romoff Emma Brunskill Dan Jurafsky and Joelle Pineau. 2020. Towards the Systematic Reporting of the Energy and Carbon Footprints of Machine Learning. 10.48550\/ARXIV.2002.05651","DOI":"10.48550\/ARXIV.2002.05651"},{"key":"e_1_3_2_1_14_1","unstructured":"Ralph Hintemann and Simon Hinterholzer. 2020. Energy consumption of data centers worldwide How will the Internet become green?"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1145\/3385412.3386027"},{"key":"e_1_3_2_1_16_1","volume-title":"World Energy Outlook","author":"IEA.","year":"2022","unstructured":"IEA. 2022. World Energy Outlook 2022. Technical Report. IEA, Paris. https:\/\/www.iea.org\/reports\/world-energy-outlook-2022"},{"key":"e_1_3_2_1_17_1","volume-title":"Tracking Clean Energy Progress","author":"IEA.","year":"2023","unstructured":"IEA. 2023. Tracking Clean Energy Progress 2023. Technical Report. IEA, Paris. https:\/\/www.iea.org\/reports\/tracking-clean-energy-progress-2023"},{"key":"e_1_3_2_1_18_1","first-page":"2811","article-title":"Data Centres Metered Electricity Consumption 2022","author":"Statistics Office Ireland Central","year":"2023","unstructured":"Ireland Central Statistics Office. 2023. Data Centres Metered Electricity Consumption 2022. ISSN: 2811-5422.","journal-title":"ISSN"},{"key":"e_1_3_2_1_19_1","volume-title":"Garnett (Eds.)","volume":"31","author":"Jeon Yunho","year":"2018","unstructured":"Yunho Jeon and Junmo Kim. 2018. Constructing Fast Network through Deconstruction of Convolution. In Advances in Neural Information Processing Systems, S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, and R. Garnett (Eds.), Vol. 31. Curran Associates, Inc. https:\/\/proceedings.neurips.cc\/paper_files\/paper\/2018\/file\/9719a00ed0c5709d80dfef33795dcef3-Paper.pdf"},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","unstructured":"Alexandre Lacoste Alexandra Luccioni Victor Schmidt and Thomas Dandres. 2019. Quantifying the Carbon Emissions of Machine Learning. 10.48550\/ARXIV.1910.09700","DOI":"10.48550\/ARXIV.1910.09700"},{"key":"e_1_3_2_1_21_1","volume-title":"Proceedings of the Thirty-second Conference on Neural Information Processing Systems (NeurIPS-18)","author":"Lazic Nevena","year":"2018","unstructured":"Nevena Lazic, Tyler Lu, Craig Boutilier, MK Ryu, Eehern Jay Wong, Binz Roy, and Greg Imwalle. 2018. Data Center Cooling using Model-predictive Control. In Proceedings of the Thirty-second Conference on Neural Information Processing Systems (NeurIPS-18). Montreal, QC, 3818--3827. https:\/\/papers.nips.cc\/paper\/7638-data-center-cooling-using-model-predictive-control"},{"key":"e_1_3_2_1_22_1","unstructured":"LeetCode. 2010. The World's Leading Online Programming Learning Platform --- leetcode.com. https:\/\/leetcode.com\/. [Accessed 04-12-2023]."},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01264-9_8"},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.infsof.2021.106560"},{"key":"e_1_3_2_1_25_1","volume-title":"Recalibrating global data center energy-use estimates. Science 367, 6481 (02","author":"Masanet Eric","year":"2020","unstructured":"Eric Masanet, Arman Shehabi, Nuoa Lei, and Sarah Smith. 2020. Recalibrating global data center energy-use estimates. Science 367, 6481 (02 2020), 984--986."},{"key":"e_1_3_2_1_26_1","volume-title":"Carbon Emissions and Large Neural Network Training. arXiv preprint arxiv:2104.10350","year":"2021","unstructured":"Patterson, Gonzalez, Le, Liang, Munguia, Rothchild, So, Texier, and Dean. 2021. Carbon Emissions and Large Neural Network Training. arXiv preprint arxiv:2104.10350 (2021)."},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1109\/ijcnn.2018.8489312"},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","unstructured":"Mark Sandler Andrew Howard Menglong Zhu Andrey Zhmoginov and Liang-Chieh Chen. 2018. MobileNetV2: Inverted Residuals and Linear Bottlenecks. (2018). 10.48550\/ARXIV.1801.04381","DOI":"10.48550\/ARXIV.1801.04381"},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1145\/3381831"},{"key":"e_1_3_2_1_30_1","unstructured":"David Richard So Quoc V. Le and Chen Liang. 2019. The Evolved Transformer. https:\/\/arxiv.org\/pdf\/1901.11117.pdf"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P19-1355"},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.48550\/ARXIV.1905.11946"},{"key":"e_1_3_2_1_33_1","first-page":"795","article-title":"Sustainable ai: Environmental implications, challenges and opportunities","volume":"4","author":"Wu Carole-Jean","year":"2022","unstructured":"Carole-Jean Wu, Ramya Raghavendra, Udit Gupta, Bilge Acun, Newsha Ardalani, Kiwan Maeng, Gloria Chang, Fiona Aga, Jinshi Huang, Charles Bai, et al. 2022. Sustainable ai: Environmental implications, challenges and opportunities. Proceedings of Machine Learning and Systems 4 (2022), 795--813.","journal-title":"Proceedings of Machine Learning and Systems"}],"event":{"name":"LLM4Code '24: 1st International Workshop on Large Language Models for Code","location":"Lisbon Portugal","acronym":"LLM4Code '24","sponsor":["SIGSOFT ACM Special Interest Group on Software Engineering","IEEE CS","Faculty of Engineering of University of Porto"]},"container-title":["Proceedings of the 1st International Workshop on Large Language Models for Code"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3643795.3648394","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3643795.3648394","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T00:57:45Z","timestamp":1750294665000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3643795.3648394"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,4,20]]},"references-count":33,"alternative-id":["10.1145\/3643795.3648394","10.1145\/3643795"],"URL":"https:\/\/doi.org\/10.1145\/3643795.3648394","relation":{},"subject":[],"published":{"date-parts":[[2024,4,20]]},"assertion":[{"value":"2024-09-10","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}