{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T10:50:53Z","timestamp":1777632653296,"version":"3.51.4"},"publisher-location":"New York, NY, USA","reference-count":60,"publisher":"ACM","license":[{"start":{"date-parts":[[2025,2,25]],"date-time":"2025-02-25T00:00:00Z","timestamp":1740441600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,2,25]]},"DOI":"10.1145\/3708493.3712691","type":"proceedings-article","created":{"date-parts":[[2025,2,25]],"date-time":"2025-02-25T17:02:04Z","timestamp":1740502924000},"page":"141-153","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":24,"title":["LLM Compiler: Foundation Language Models for Compiler Optimization"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-9382-4302","authenticated-orcid":false,"given":"Chris","family":"Cummins","sequence":"first","affiliation":[{"name":"Meta, Menlo Park, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0120-895X","authenticated-orcid":false,"given":"Volker","family":"Seeker","sequence":"additional","affiliation":[{"name":"Meta, Menlo Park, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3336-0726","authenticated-orcid":false,"given":"Dejan","family":"Grubisic","sequence":"additional","affiliation":[{"name":"Meta, Menlo Park, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9014-4379","authenticated-orcid":false,"given":"Baptiste","family":"Roziere","sequence":"additional","affiliation":[{"name":"Meta, Paris, France"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-4115-3964","authenticated-orcid":false,"given":"Jonas","family":"Gehring","sequence":"additional","affiliation":[{"name":"Meta, Paris, France"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1715-3356","authenticated-orcid":false,"given":"Gabriel","family":"Synnaeve","sequence":"additional","affiliation":[{"name":"Meta, Paris, France"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0664-4176","authenticated-orcid":false,"given":"Hugh","family":"Leather","sequence":"additional","affiliation":[{"name":"Meta, Menlo Park, USA"}]}],"member":"320","published-online":{"date-parts":[[2025,2,25]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","unstructured":"F. Agakov E. Bonilla J. Cavazos B. Franke G. Fursin M.F.P. O\u2019Boyle J. Thomson M. Toussaint and C.K.I. Williams. 2006. Using Machine Learning to Focus Iterative Optimization. In CGO. https:\/\/doi.org\/10.1109\/CGO.2006.37 10.1109\/CGO.2006.37","DOI":"10.1109\/CGO.2006.37"},{"key":"e_1_3_2_1_2_1","volume-title":"Niklas Muennighoff, Mayank Mishra, Alex Gu, Manan Dey, Logesh Kumar Umapathi, Carolyn Jane Anderson, Yangtian Zi, Joel Lamy Poirier, Hailey Schoelkopf, Sergey Troshin, Dmitry Abulkhanov, Manuel Romero","author":"Allal Loubna Ben","year":"2023","unstructured":"Loubna Ben Allal, Raymond Li, Denis Kocetkov, Chenghao Mou, Christopher Akiki, Carlos Munoz Ferrandis, Niklas Muennighoff, Mayank Mishra, Alex Gu, Manan Dey, Logesh Kumar Umapathi, Carolyn Jane Anderson, Yangtian Zi, Joel Lamy Poirier, Hailey Schoelkopf, Sergey Troshin, Dmitry Abulkhanov, Manuel Romero, Michael Lappert, Francesco De Toni, Bernardo Garc\u00eda del R\u00edo, Qian Liu, Shamik Bose, Urvashi Bhattacharyya, Terry Yue Zhuo, Ian Yu, Paulo Villegas, Marco Zocca, Sourab Mangrulkar, David Lansky, Huu Nguyen, Danish Contractor, Luis Villa, Jia Li, Dzmitry Bahdanau, Yacine Jernite, Sean Hughes, Daniel Fried, Arjun Guha, Harm de Vries, and Leandro von Werra. 2023. SantaCoder: Don\u2019t Reach for the Stars!. arXiv:2301.03988."},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","unstructured":"Jordi Armengol-Estap\u00e9 Jackson Woodruff Chris Cummins and Michael FP O\u2019Boyle. 2024. SLaDe: A Portable Small Language Model Decompiler for Optimized Assembler. In CGO. https:\/\/doi.org\/10.1109\/CGO57630.2024.10444788 10.1109\/CGO57630.2024.10444788","DOI":"10.1109\/CGO57630.2024.10444788"},{"key":"e_1_3_2_1_4_1","volume-title":"Bryan Chan, and Yaoqing Gao.","author":"Ashouri Amir H","year":"2022","unstructured":"Amir H Ashouri, Mostafa Elhoushi, Yuzhe Hua, Xiang Wang, Muhammad Asif Manzoor, Bryan Chan, and Yaoqing Gao. 2022. MLGOPerf: An ML Guided Inliner to Optimize Performance. arXiv:2207.08389."},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"crossref","unstructured":"Ying Cao Ruigang Liang Kai Chen and Peiwei Hu. 2022. Boosting Neural Networks to Decompile Optimized Binaries. In ACSAC.","DOI":"10.1145\/3564625.3567998"},{"key":"e_1_3_2_1_6_1","unstructured":"Mark Chen Jerry Tworek Heewoo Jun Qiming Yuan Henrique Ponde de Oliveira Pinto Jared Kaplan Harri Edwards Yuri Burda Nicholas Joseph Greg Brockman Alex Ray Raul Puri Gretchen Krueger Michael Petrov Heidy Khlaaf Girish Sastry Pamela Mishkin Brooke Chan Scott Gray Nick Ryder Mikhail Pavlov Alethea Power Lukasz Kaiser Mohammad Bavarian Clemens Winter Philippe Tillet Felipe Petroski Such Dave Cummings Matthias Plappert Fotios Chantzis Elizabeth Barnes Ariel Herbert-Voss William Hebgen Guss Alex Nichol Alex Paino Nikolas Tezak Jie Tang Igor Babuschkin Suchir Balaji Shantanu Jain William Saunders Christopher Hesse Andrew N. Carr Jan Leike Josh Achiam Vedant Misra Evan Morikawa Alec Radford Matthew Knight Miles Brundage Mira Murati Katie Mayer Peter Welinder Bob McGrew Dario Amodei Sam McCandlish Ilya Sutskever and Wojciech Zaremba. 2021. Evaluating Large Language Models Trained on Code. arXiv:2107.03374."},{"key":"e_1_3_2_1_7_1","unstructured":"Chris Cummins Zacharias Fisches Tal Ben-Nun Torsten Hoefler Michael O\u2019Boyle and Hugh Leather. 2021. ProGraML: A Graph-based Program Representation for Data Flow Analysis and Compiler Optimizations. In ICML."},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","unstructured":"Chris Cummins Pavlos Petoumenos Zheng Wang and Hugh Leather. 2017. End-to-End Deep Learning of Optimization Heuristics. In PACT. https:\/\/doi.org\/10.1109\/PACT.2017.24 10.1109\/PACT.2017.24","DOI":"10.1109\/PACT.2017.24"},{"key":"e_1_3_2_1_9_1","unstructured":"Chris Cummins Volker Seeker Dejan Grubisic Mostafa Elhoushi Youwei Liang Baptiste Roziere Jonas Gehring Fabian Gloeckle Kim Hazelwood Gabriel Synnaeve and Hugh Leather. 2023. Large language models for compiler optimization. arXiv preprint arXiv:2309.07062."},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","unstructured":"Chris Cummins Bram Wasti Jiadong Guo Brandon Cui Jason Ansel Sahir Gomez Somya Jain Jia Liu Olivier Teytaud Benoit Steiner Yuandong Tian and Hugh Leather. 2022. CompilerGym: Robust Performant Compiler Optimization Environments for AI Research. In CGO. https:\/\/doi.org\/10.1109\/CGO53902.2022.9741258 10.1109\/CGO53902.2022.9741258","DOI":"10.1109\/CGO53902.2022.9741258"},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1145\/3597926.3598067"},{"key":"e_1_3_2_1_12_1","unstructured":"Tim Dettmers Artidoro Pagnoni Ari Holtzman and Luke Zettlemoyer. 2023. QLoRA: Efficient Finetuning of Quantized LLMs. arXiv preprint arXiv:2305.14314."},{"key":"e_1_3_2_1_13_1","unstructured":"Jiayu Ding Shuming Ma Li Dong Xingxing Zhang Shaohan Huang Wenhui Wang and Furu Wei. 2023. LongNet: Scaling Transformers to 1 000 000 000 Tokens. arXiv:2307.02486."},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"crossref","unstructured":"Zhangyin Feng Daya Guo Duyu Tang Nan Duan Xiaocheng Feng Ming Gong Linjun Shou Bing Qin Ting Liu Daxin Jiang and Ming Zhou. 2020. CodeBERT: A Pre-trained Model for Programming and Natural Languages. arXiv:2002.08155.","DOI":"10.18653\/v1\/2020.findings-emnlp.139"},{"key":"e_1_3_2_1_15_1","unstructured":"Daniel Fried Armen Aghajanyan Jessy Lin Sida Wang Eric Wallace Freda Shi Ruiqi Zhong Wen-tau Yih Luke Zettlemoyer and Mike Lewis. 2023. InCoder: A Generative Model for Code Infilling and Synthesis. arXiv:2204.05999."},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","unstructured":"G. G. Fursin M. F. P. O\u2019Boyle and P. M. W. Knijnenburg. 2005. Evaluating Iterative Compilation. In LCPC. https:\/\/doi.org\/10.1007\/11596110_24 10.1007\/11596110_24","DOI":"10.1007\/11596110_24"},{"key":"e_1_3_2_1_17_1","article-title":"A New Algorithm for Data Compression","volume":"12","author":"Gage Philip","year":"1994","unstructured":"Philip Gage. 1994. A New Algorithm for Data Compression. C Users Journal, 12, 2 (1994).","journal-title":"C Users Journal"},{"key":"e_1_3_2_1_18_1","unstructured":"Shannon K Gallagher William E Klieber and David Svoboda. 2022. LLVM Intermediate Representation for Code Weakness Identification."},{"key":"e_1_3_2_1_19_1","volume-title":"Mircea Trofin, and Johannes Doerfert.","author":"Grossman Aiden","year":"2024","unstructured":"Aiden Grossman, Ludger Paehler, Konstantinos Parasyris, Tal Ben-Nun, Jacob Hegna, William Moses, Jose M Monsalve Diaz, Mircea Trofin, and Johannes Doerfert. 2024. ComPile: A Large IR Dataset from Production Sources. arXiv:2309.15432."},{"key":"e_1_3_2_1_20_1","unstructured":"Dejan Grubisic Chris Cummins Volker Seeker and Hugh Leather. 2024. Compiler generated feedback for Large Language Models. arXiv:2403.14714."},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"crossref","unstructured":"Dejan Grubisic Chris Cummins Volker Seeker and Hugh Leather. 2024. Priority Sampling of Large Language Models for Compilers. arXiv:2402.18734.","DOI":"10.1145\/3642970.3655831"},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","unstructured":"Yi Gui Yao Wan Hongyu Zhang Huifang Huang Yulei Sui Guandong Xu Zhiyuan Shao and Hai Jin. 2022. Cross-language binary-source code matching with intermediate representations. In SANER. https:\/\/doi.org\/10.1109\/SANER53432.2022.00077 10.1109\/SANER53432.2022.00077","DOI":"10.1109\/SANER53432.2022.00077"},{"key":"e_1_3_2_1_23_1","unstructured":"Suriya Gunasekar Yi Zhang Jyoti Aneja Caio C\u00e9sar Teodoro Mendes Allie Del Giorno Sivakanth Gopi Mojan Javaheripi Piero Kauffmann Gustavo de Rosa Olli Saarikivi Adil Salim Shital Shah Harkirat Singh Behl Xin Wang S\u00e9bastien Bubeck Ronen Eldan Adam Tauman Kalai Yin Tat Lee and Yuanzhi Li. 2023. Textbooks Are All You Need. arXiv:2306.11644."},{"key":"e_1_3_2_1_24_1","unstructured":"Daya Guo Qihao Zhu Dejian Yang Zhenda Xie Kai Dong Wentao Zhang Guanting Chen Xiao Bi Y. Wu Y. K. Li Fuli Luo Yingfei Xiong and Wenfeng Liang. 2024. DeepSeek-Coder: When the Large Language Model Meets Programming \u2013 The Rise of Code Intelligence. arXiv:2401.14196."},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1109\/WWC.2001.990739"},{"key":"e_1_3_2_1_26_1","unstructured":"Ameer Haj-Ali Qijing Huang William Moses John Xiang John Wawrzynek Krste Asanovic and Ion Stoica. 2020. AutoPhase: Juggling HLS Phase Orderings in Random Forests with Deep Reinforcement Learning. In MLSys."},{"key":"e_1_3_2_1_27_1","unstructured":"Xinyi Hou Yanjie Zhao Yue Liu Zhou Yang Kailong Wang Li Li Xiapu Luo David Lo John Grundy and Haoyu Wang. 2023. Large Language Models for Software Engineering: A Systematic Literature Review. arXiv:2308.10620."},{"key":"e_1_3_2_1_28_1","volume-title":"LoRA: Low-Rank Adaptation of Large Language Models. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=nZeVKeeFYf9","author":"Hu Edward J","year":"2022","unstructured":"Edward J Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, and Weizhu Chen. 2022. LoRA: Low-Rank Adaptation of Large Language Models. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=nZeVKeeFYf9"},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","unstructured":"Tarindu Jayatilaka Hideto Ueno Giorgis Georgakoudis EunJung Park and Johannes Doerfert. 2021. Towards Compile-Time-Reducing Compiler Optimization Selection via Machine Learning. In ICPP. https:\/\/doi.org\/10.1145\/3458744.3473355 10.1145\/3458744.3473355","DOI":"10.1145\/3458744.3473355"},{"key":"e_1_3_2_1_30_1","unstructured":"Juyong Jiang Fan Wang Jiasi Shen Sungju Kim and Sunghun Kim. 2024. A Survey on Large Language Models for Code Generation. arXiv:2406.00515."},{"key":"e_1_3_2_1_31_1","volume-title":"Jia Li, Chenghao Mou, Carlos Mu\u00f1oz Ferrandis, Yacine Jernite, Margaret Mitchell, Sean Hughes, Thomas Wolf, Dzmitry Bahdanau, Leandro von Werra, and Harm de Vries.","author":"Kocetkov Denis","year":"2022","unstructured":"Denis Kocetkov, Raymond Li, Loubna Ben Allal, Jia Li, Chenghao Mou, Carlos Mu\u00f1oz Ferrandis, Yacine Jernite, Margaret Mitchell, Sean Hughes, Thomas Wolf, Dzmitry Bahdanau, Leandro von Werra, and Harm de Vries. 2022. The Stack: 3TB of Permissively Licensed Source Code. arXiv:2211.15533."},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1109\/CGO.2004.1281665"},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"publisher","unstructured":"Hugh Leather and Chris Cummins. 2020. Machine Learning in Compilers: Past Present and Future. In FDL. https:\/\/doi.org\/10.1109\/FDL50818.2020.9232934 10.1109\/FDL50818.2020.9232934","DOI":"10.1109\/FDL50818.2020.9232934"},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","unstructured":"Yujia Li David Choi Junyoung Chung Nate Kushman Julian Schrittwieser R\u00e9mi Leblond Tom Eccles James Keeling Felix Gimeno Agustin Dal Lago Thomas Hubert Peter Choy Cyprien de Masson d\u2019Autume Igor Babuschkin Xinyun Chen Po-Sen Huang Johannes Welbl Sven Gowal Alexey Cherepanov James Molloy Daniel J. Mankowitz Esme Sutherland Robson Pushmeet Kohli Nando de Freitas Koray Kavukcuoglu and Oriol Vinyals. 2022. Competition-Level Code Generation with AlphaCode. Science 378 6624 (2022) https:\/\/doi.org\/10.1126\/science.abq1158 10.1126\/science.abq1158","DOI":"10.1126\/science.abq1158"},{"key":"e_1_3_2_1_35_1","unstructured":"Youwei Liang Kevin Stone Ali Shameli Chris Cummins Mostafa Elhoushi Jiadong Guo Benoit Steiner Xiaomeng Yang Pengtao Xie Hugh Leather and Yuandong Tian. 2023. Learning Compiler Pass Orders using Coreset and Normalized Value Prediction. In ICML."},{"key":"e_1_3_2_1_36_1","unstructured":"Ilya Loshchilov and Frank Hutter. 2017. Decoupled Weight Decay Regularization. arXiv:1711.05101."},{"key":"e_1_3_2_1_37_1","unstructured":"Anton Lozhkov Raymond Li Loubna Ben Allal Federico Cassano Joel Lamy-Poirier Nouamane Tazi Ao Tang Dmytro Pykhtar Jiawei Liu Yuxiang Wei Tianyang Liu Max Tian Denis Kocetkov Arthur Zucker Younes Belkada Zijian Wang Qian Liu Dmitry Abulkhanov Indraneil Paul Zhuang Li Wen-Ding Li Megan Risdal Jia Li Jian Zhu Terry Yue Zhuo Evgenii Zheltonozhskii Nii Osae Osae Dade Wenhao Yu Lucas Krau\u00df Naman Jain Yixuan Su Xuanli He Manan Dey Edoardo Abati Yekun Chai Niklas Muennighoff Xiangru Tang Muhtasham Oblokulov Christopher Akiki Marc Marone Chenghao Mou Mayank Mishra Alex Gu Binyuan Hui Tri Dao Armel Zebaze Olivier Dehaene Nicolas Patry Canwen Xu Julian McAuley Han Hu Torsten Scholak Sebastien Paquet Jennifer Robinson Carolyn Jane Anderson Nicolas Chapados Mostofa Patwary Nima Tajbakhsh Yacine Jernite Carlos Mu\u00f1oz Ferrandis Lingming Zhang Sean Hughes Thomas Wolf Arjun Guha Leandro von Werra and Harm de Vries. 2024. StarCoder 2 and The Stack v2: The Next Generation. arXiv:2402.19173."},{"key":"e_1_3_2_1_38_1","unstructured":"Aman Madaan Alexander Shypula Uri Alon Milad Hashemi Parthasarathy Ranganathan Yiming Yang Graham Neubig and Amir Yazdanbakhsh. 2023. Learning Performance-Improving Code Edits. arXiv:2302.07867."},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","unstructured":"William F. Ogilvie Pavlos Petoumenos Zheng Wang and Hugh Leather. 2017. Minimizing the Cost of Iterative Compilation with Active Learning. In CGO. https:\/\/doi.org\/10.1109\/CGO.2017.7863744 10.1109\/CGO.2017.7863744","DOI":"10.1109\/CGO.2017.7863744"},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.3115\/1073083.1073135"},{"key":"e_1_3_2_1_42_1","unstructured":"Indraneil Paul Goran Glava\u0161 and Iryna Gurevych. 2024. IRCoder: Intermediate Representations Make Language Models Robust Multilingual Code Generators. arXiv:2403.03894."},{"key":"e_1_3_2_1_43_1","doi-asserted-by":"publisher","DOI":"10.1109\/PACT52795.2021.00008"},{"key":"e_1_3_2_1_44_1","unstructured":"LLVM PM. 2021. Using the New Pass Manager \u2014 LLVM 17.0.6 documentation. https:\/\/llvm.org\/docs\/NewPassManager.html"},{"key":"e_1_3_2_1_45_1","volume-title":"A graph-based model for build optimization sequences: A study of optimization sequence length impacts on code size and speedup. COLA, 74","author":"Queiroz Nilton Luiz","year":"2023","unstructured":"Nilton Luiz Queiroz Jr and Anderson Faustino da Silva. 2023. A graph-based model for build optimization sequences: A study of optimization sequence length impacts on code size and speedup. COLA, 74 (2023)."},{"key":"e_1_3_2_1_46_1","unstructured":"Baptiste Rozi\u00e8re Jonas Gehring Fabian Gloeckle Sten Sootla Itai Gat Xiaoqing Ellen Tan Yossi Adi Jingyu Liu Tal Remez J\u00e9r\u00e9my Rapin Artyom Kozhevnikov Ivan Evtimov Joanna Bitton Manish Bhatt Cristian Canton Ferrer Aaron Grattafiori Wenhan Xiong Alexandre D\u00e9fossez Jade Copet Faisal Azhar Hugo Touvron Louis Martin Nicolas Usunier Thomas Scialom and Gabriel Synnaeve. 2023. Code Llama: Open Foundation Models for Code. arXiv:2308.12950."},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"crossref","unstructured":"Eric Schulte Jason Ruchti Matt Noonan David Ciarletta and Alexey Loginov. 2018. Evolving Exact Decompilation. In BAR.","DOI":"10.14722\/bar.2018.23008"},{"key":"e_1_3_2_1_48_1","unstructured":"Max Sch\u00e4fer Sarah Nadi Aryaz Eghbali and Frank Tip. 2023. Adaptive Test Generation Using a Large Language Model. arXiv:2302.06527."},{"key":"e_1_3_2_1_49_1","doi-asserted-by":"publisher","unstructured":"Volker Seeker Chris Cummins Murray Cole Bj\u00f6rn Franke Kim Hazelwood and Hugh Leather. 2024. Revealing Compiler Heuristics Through Automated Discovery and Optimization. In CGO. https:\/\/doi.org\/10.1109\/CGO57630.2024.10444847 10.1109\/CGO57630.2024.10444847","DOI":"10.1109\/CGO57630.2024.10444847"},{"key":"e_1_3_2_1_50_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2023.127063"},{"key":"e_1_3_2_1_51_1","unstructured":"Marc Szafraniec Baptiste Roziere Francois Charton Hugh Leather Patrick Labatut and Gabriel Synnaeve. 2022. Code Translation with Compiler Representations. arXiv:2207.03578."},{"key":"e_1_3_2_1_52_1","volume-title":"Lahiri","author":"Taneja Jubi","year":"2024","unstructured":"Jubi Taneja, Avery Laird, Cong Yan, Madan Musuvathi, and Shuvendu K. Lahiri. 2024. LLM-Vectorizer: LLM-based Verified Loop Vectorizer. arXiv:2406.04693."},{"key":"e_1_3_2_1_53_1","volume-title":"Llama: Open and efficient foundation language models. arXiv:2302.13971.","author":"Touvron Hugo","year":"2023","unstructured":"Hugo Touvron, Thibaut Lavril, Gautier Izacard, Xavier Martinet, Marie-Anne Lachaux, Timoth\u00e9e Lacroix, Baptiste Rozi\u00e8re, Naman Goyal, Eric Hambro, Faisal Azhar, Aurelien Rodriguez, Armand Joulin, Edouard Grave, and Guillaume Lample. 2023. Llama: Open and efficient foundation language models. arXiv:2302.13971."},{"key":"e_1_3_2_1_54_1","unstructured":"Hugo Touvron Louis Martin Kevin Stone Peter Albert Amjad Almahairi Yasmine Babaei Nikolay Bashlykov Soumya Batra Prajjwal Bhargava Shruti Bhosale Dan Bikel Lukas Blecher Cristian Canton Ferrer Moya Chen Guillem Cucurull David Esiobu Jude Fernandes Jeremy Fu Wenyin Fu Brian Fuller Cynthia Gao Vedanuj Goswami Naman Goyal Anthony Hartshorn Saghar Hosseini Rui Hou Hakan Inan Marcin Kardas Viktor Kerkez Madian Khabsa Isabel Kloumann Artem Korenev Punit Singh Koura Marie-Anne Lachaux Thibaut Lavril Jenya Lee Diana Liskovich Yinghai Lu Yuning Mao Xavier Martinet Todor Mihaylov Pushkar Mishra Igor Molybog Yixin Nie Andrew Poulton Jeremy Reizenstein Rashi Rungta Kalyan Saladi Alan Schelten Ruan Silva Eric Michael Smith Ranjan Subramanian Xiaoqing Ellen Tan Binh Tang Ross Taylor Adina Williams Jian Xiang Kuan Puxin Xu Zheng Yan Iliyan Zarov Yuchen Zhang Angela Fan Melanie Kambadur Sharan Narang Aurelien Rodriguez Robert Stojnic Sergey Edunov and Thomas Scialom. 2023. Llama 2: Open Foundation and Fine-Tuned Chat Models. arXiv:2307.09288."},{"key":"e_1_3_2_1_55_1","unstructured":"Mircea Trofin Yundi Qian Eugene Brevdo Zinan Lin Krzysztof Choromanski and David Li. 2021. MLGO: a Machine Learning Guided Compiler Optimizations Framework. arXiv:2101.04808."},{"key":"e_1_3_2_1_56_1","volume-title":"Nghi DQ Bui, Junnan Li, and Steven CH Hoi.","author":"Wang Yue","year":"2023","unstructured":"Yue Wang, Hung Le, Akhilesh Deepak Gotmare, Nghi DQ Bui, Junnan Li, and Steven CH Hoi. 2023. Codet5+: Open code large language models for code understanding and generation. arXiv preprint arXiv:2305.07922."},{"key":"e_1_3_2_1_57_1","unstructured":"Zheng Wang and Michael O\u2019Boyle. 2018. Machine Learning in Compiler Optimisation. arXiv:1805.03441."},{"key":"e_1_3_2_1_58_1","volume-title":"Magicoder: Empowering Code Generation with OSS-Instruct. arXiv:2312.02120.","author":"Wei Yuxiang","year":"2024","unstructured":"Yuxiang Wei, Zhe Wang, Jiawei Liu, Yifeng Ding, and Lingming Zhang. 2024. Magicoder: Empowering Code Generation with OSS-Instruct. arXiv:2312.02120."},{"key":"e_1_3_2_1_59_1","volume-title":"Michael Pradel, and Lingming Zhang.","author":"Xia Chunqiu Steven","year":"2023","unstructured":"Chunqiu Steven Xia, Matteo Paltenghi, Jia Le Tian, Michael Pradel, and Lingming Zhang. 2023. Universal Fuzzing via Large Language Models. arXiv:2308.04748."},{"key":"e_1_3_2_1_60_1","doi-asserted-by":"publisher","unstructured":"Chunqiu Steven Xia Yuxiang Wei and Lingming Zhang. 2023. Automated Program Repair in the Era of Large Pre-Trained Language Models. In ICSE. https:\/\/doi.org\/10.1109\/ICSE48619.2023.00129 10.1109\/ICSE48619.2023.00129","DOI":"10.1109\/ICSE48619.2023.00129"},{"key":"e_1_3_2_1_61_1","doi-asserted-by":"crossref","unstructured":"Qinkai Zheng Xiao Xia Xu Zou Yuxiao Dong Shan Wang Yufei Xue Zihan Wang Lei Shen Andi Wang Yang Li Teng Su Zhilin Yang and Jie Tang. 2023. CodeGeeX: A Pre-Trained Model for Code Generation with Multilingual Evaluations on HumanEval-X. arXiv:2303.17568.","DOI":"10.1145\/3580305.3599790"}],"event":{"name":"CC '25: 34th ACM SIGPLAN International Conference on Compiler Construction","location":"Las Vegas NV USA","acronym":"CC '25","sponsor":["SIGPLAN SIGPLAN Programming Languages"]},"container-title":["Proceedings of the 34th ACM SIGPLAN International Conference on Compiler Construction"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3708493.3712691","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3708493.3712691","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T01:09:54Z","timestamp":1750295394000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3708493.3712691"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,2,25]]},"references-count":60,"alternative-id":["10.1145\/3708493.3712691","10.1145\/3708493"],"URL":"https:\/\/doi.org\/10.1145\/3708493.3712691","relation":{},"subject":[],"published":{"date-parts":[[2025,2,25]]},"assertion":[{"value":"2025-02-25","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}