{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,15]],"date-time":"2025-11-15T10:35:15Z","timestamp":1763202915993,"version":"3.44.0"},"publisher-location":"New York, NY, USA","reference-count":59,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,10,13]],"date-time":"2024-10-13T00:00:00Z","timestamp":1728777600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100006374","name":"National Science Foundation","doi-asserted-by":"publisher","award":["2211982"],"award-info":[{"award-number":["2211982"]}],"id":[{"id":"10.13039\/501100006374","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,10,14]]},"DOI":"10.1145\/3656019.3676895","type":"proceedings-article","created":{"date-parts":[[2024,10,11]],"date-time":"2024-10-11T10:34:08Z","timestamp":1728642848000},"page":"156-167","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":3,"title":["MIREncoder: Multi-modal IR-based Pretrained Embeddings for Performance Optimizations"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0007-0947-1182","authenticated-orcid":false,"given":"Akash","family":"Dutta","sequence":"first","affiliation":[{"name":"Iowa State University, United States of America"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8672-5317","authenticated-orcid":false,"given":"Ali","family":"Jannesari","sequence":"additional","affiliation":[{"name":"Iowa State University, United States of America"}]}],"member":"320","published-online":{"date-parts":[[2024,10,13]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1007\/s00607-022-01081-6"},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.1145\/3212695"},{"key":"e_1_3_2_1_3_1","volume-title":"Learning to represent programs with graphs. arXiv preprint arXiv:1711.00740","author":"Allamanis Miltiadis","year":"2017","unstructured":"Miltiadis Allamanis, Marc Brockschmidt, and Mahmoud Khademi. 2017. Learning to represent programs with graphs. arXiv preprint arXiv:1711.00740 (2017)."},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1145\/3296979.3192412"},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1145\/3290353"},{"key":"e_1_3_2_1_6_1","unstructured":"AMD. [n. d.]. AMD OpenCL accelerated parallel processing SDK. https:\/\/developer.amd.com\/amd-accelerated-parallel-processing-app-sdk\/."},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1145\/2628071.2628092"},{"key":"e_1_3_2_1_8_1","unstructured":"E Barszcz J Barton L Dagum P Frederickson T Lasinski R Schreiber V Venkatakrishnan S Weeratunga D Bailey D Browning 1991. The nas parallel benchmarks. In The International Journal of Supercomputer Applications. Citeseer."},{"key":"e_1_3_2_1_9_1","volume-title":"Neural code comprehension: A learnable representation of code semantics. Advances in neural information processing systems 31","author":"Ben-Nun Tal","year":"2018","unstructured":"Tal Ben-Nun, Alice\u00a0Shoshana Jakobovits, and Torsten Hoefler. 2018. Neural code comprehension: A learnable representation of code semantics. Advances in neural information processing systems 31 (2018)."},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICAPAI49758.2021.9462050"},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1145\/3377555.3377894"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-540-79561-2_2"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1109\/HiPC.2011.6152736"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1109\/IISWC.2009.5306797"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1109\/IISWC.2010.5650274"},{"key":"e_1_3_2_1_16_1","volume-title":"Proceedings of Machine Learning and Systems 5","author":"Chen Le","year":"2023","unstructured":"Le Chen, Quazi\u00a0Ishtiaque Mahmud, Hung Phan, Nesreen Ahmed, and Ali Jannesari. 2023. Learning to parallelize with openMP by augmented heterogeneous AST representation. Proceedings of Machine Learning and Systems 5 (2023)."},{"key":"e_1_3_2_1_17_1","unstructured":"NVIDIA Corporation. [n. d.]. CUDA. http:\/\/developer.nvidia.com\/object\/cuda.html."},{"key":"e_1_3_2_1_18_1","volume-title":"International Conference on Machine Learning. PMLR, 2244\u20132253","author":"Cummins Chris","year":"2021","unstructured":"Chris Cummins, Zacharias\u00a0V Fisches, Tal Ben-Nun, Torsten Hoefler, Michael\u00a0FP O\u2019Boyle, and Hugh Leather. 2021. Programl: A graph-based program representation for data flow analysis and compiler optimizations. In International Conference on Machine Learning. PMLR, 2244\u20132253."},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1109\/PACT.2017.24"},{"key":"e_1_3_2_1_20_1","volume-title":"Large language models for compiler optimization. arXiv preprint arXiv:2309.07062","author":"Cummins Chris","year":"2023","unstructured":"Chris Cummins, Volker Seeker, Dejan Grubisic, Mostafa Elhoushi, Youwei Liang, Baptiste Roziere, Jonas Gehring, Fabian Gloeckle, Kim Hazelwood, Gabriel Synnaeve, 2023. Large language models for compiler optimization. arXiv preprint arXiv:2309.07062 (2023)."},{"key":"e_1_3_2_1_21_1","volume-title":"A deep tree-based model for software defect prediction. arXiv preprint arXiv:1802.00921","author":"Dam Hoa\u00a0Khanh","year":"2018","unstructured":"Hoa\u00a0Khanh Dam, Trang Pham, Shien\u00a0Wee Ng, Truyen Tran, John Grundy, Aditya Ghose, Taeksu Kim, and Chul-Joo Kim. 2018. A deep tree-based model for software defect prediction. arXiv preprint arXiv:1802.00921 (2018)."},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1145\/1735688.1735702"},{"key":"e_1_3_2_1_23_1","volume-title":"Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805","author":"Devlin Jacob","year":"2018","unstructured":"Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018)."},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1145\/3588195.3592984"},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1109\/AI4S56813.2022.00010"},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1109\/IPDPS54959.2023.00060"},{"key":"e_1_3_2_1_27_1","volume-title":"Codebert: A pre-trained model for programming and natural languages. arXiv preprint arXiv:2002.08155","author":"Feng Zhangyin","year":"2020","unstructured":"Zhangyin Feng, Daya Guo, Duyu Tang, Nan Duan, Xiaocheng Feng, Ming Gong, Linjun Shou, Bing Qin, Ting Liu, Daxin Jiang, 2020. Codebert: A pre-trained model for programming and natural languages. arXiv preprint arXiv:2002.08155 (2020)."},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1109\/CGO.2013.6494993"},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1145\/3368826.3377928"},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1145\/2807426.2807429"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1109\/HPEC58863.2023.10363459"},{"volume-title":"Technical Report.","author":"Karlin Ian","key":"e_1_3_2_1_32_1","unstructured":"Ian Karlin, Jeff Keasler, and J\u00a0Robert Neely. 2013. Lulesh 2.0 updates and changes. Technical Report. Lawrence Livermore National Lab.(LLNL), Livermore, CA (United States)."},{"key":"e_1_3_2_1_33_1","volume-title":"International conference on machine learning. PMLR, 3835\u20133845","author":"Li Yujia","year":"2019","unstructured":"Yujia Li, Chenjie Gu, Thomas Dullien, Oriol Vinyals, and Pushmeet Kohli. 2019. Graph matching networks for learning the similarity of graph structured objects. In International conference on machine learning. PMLR, 3835\u20133845."},{"key":"e_1_3_2_1_34_1","volume-title":"Vilbert: Pretraining task-agnostic visiolinguistic representations for vision-and-language tasks. Advances in neural information processing systems 32","author":"Lu Jiasen","year":"2019","unstructured":"Jiasen Lu, Dhruv Batra, Devi Parikh, and Stefan Lee. 2019. Vilbert: Pretraining task-agnostic visiolinguistic representations for vision-and-language tasks. Advances in neural information processing systems 32 (2019)."},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1145\/2628071.2628087"},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.1145\/2503210.2503268"},{"key":"e_1_3_2_1_37_1","unstructured":"Jiquan Ngiam Aditya Khosla Mingyu Kim Juhan Nam Honglak Lee and Andrew\u00a0Y Ng. 2011. Multimodal deep learning. In ICML."},{"key":"e_1_3_2_1_38_1","volume-title":"Polybench: The polyhedral benchmark suite. URL: http:\/\/www. cs. ucla. edu\/pouchet\/software\/polybench 437","author":"Louis-No\u00ebl","year":"2012","unstructured":"Louis-No\u00ebl Pouchet 2012. Polybench: The polyhedral benchmark suite. URL: http:\/\/www. cs. ucla. edu\/pouchet\/software\/polybench 437 (2012), 1\u20131."},{"key":"e_1_3_2_1_39_1","unstructured":"Ruchir Puri David Kung Geert Janssen Wei Zhang Giacomo Domeniconi Vladmir Zolotov Julian Dolby Jie Chen Mihir Choudhury Lindsey Decker Veronika Thost Luca Buratti Saurabh Pujar Shyam Ramji Ulrich Finkler Susan Malaika and Frederick Reiss. 2021. CodeNet: A Large-Scale AI for Code Dataset for Learning a Diversity of Coding Tasks."},{"key":"e_1_3_2_1_40_1","volume-title":"Rigel: A Framework for OpenMP PerformanceTuning. In 2019 IEEE 21st International Conference on High Performance Computing and Communications","author":"Rameshka Piyumi","year":"2019","unstructured":"Piyumi Rameshka, Pasindu Senanayake, Thulana Kannangara, Praveen Seneviratne, Sanath Jayasena, Tharindu Rusira, and Mary Hall. 2019. Rigel: A Framework for OpenMP PerformanceTuning. In 2019 IEEE 21st International Conference on High Performance Computing and Communications; IEEE 17th International Conference on Smart City; IEEE 5th International Conference on Data Science and Systems (HPCC\/SmartCity\/DSS). IEEE, 2093\u20132102."},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.1145\/2775051.2677009"},{"key":"e_1_3_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1145\/3453483.3454109"},{"key":"e_1_3_2_1_43_1","doi-asserted-by":"publisher","DOI":"10.1145\/3392717.3392765"},{"key":"e_1_3_2_1_44_1","doi-asserted-by":"publisher","DOI":"10.1109\/IISWC.2011.6114174"},{"key":"e_1_3_2_1_45_1","doi-asserted-by":"publisher","DOI":"10.1145\/3194242"},{"key":"e_1_3_2_1_46_1","volume-title":"Proceedings of Machine Learning and Systems 3","author":"Steiner Benoit","year":"2021","unstructured":"Benoit Steiner, Chris Cummins, Horace He, and Hugh Leather. 2021. Value learning for throughput optimization of deep learning workloads. Proceedings of Machine Learning and Systems 3 (2021)."},{"key":"e_1_3_2_1_47_1","first-page":"27","article-title":"Parboil: A revised benchmark suite for scientific and commercial throughput computing","volume":"127","author":"Stratton A","year":"2012","unstructured":"John\u00a0A Stratton, Christopher Rodrigues, I-Jui Sung, Nady Obeid, Li-Wen Chang, Nasser Anssari, Geng\u00a0Daniel Liu, and Wen-mei\u00a0W Hwu. 2012. Parboil: A revised benchmark suite for scientific and commercial throughput computing. Center for Reliable and High-Performance Computing 127 (2012), 27.","journal-title":"Center for Reliable and High-Performance Computing"},{"key":"e_1_3_2_1_48_1","volume-title":"Recent Advances and Trends in Multimodal Deep Learning: A Review. arXiv preprint arXiv:2105.11087","author":"Summaira Jabeen","year":"2021","unstructured":"Jabeen Summaira, Xi Li, Amin\u00a0Muhammad Shoib, Songyuan Li, and Jabbar Abdul. 2021. Recent Advances and Trends in Multimodal Deep Learning: A Review. arXiv preprint arXiv:2105.11087 (2021)."},{"key":"e_1_3_2_1_49_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00756"},{"key":"e_1_3_2_1_50_1","doi-asserted-by":"publisher","DOI":"10.1109\/SC.2002.10062"},{"key":"e_1_3_2_1_51_1","volume-title":"Perfograph: A numerical aware program graph representation for performance optimization and program analysis. Advances in Neural Information Processing Systems 36","author":"TehraniJamsaz Ali","year":"2024","unstructured":"Ali TehraniJamsaz, Quazi\u00a0Ishtiaque Mahmud, Le Chen, Nesreen\u00a0K Ahmed, and Ali Jannesari. 2024. Perfograph: A numerical aware program graph representation for performance optimization and program analysis. Advances in Neural Information Processing Systems 36 (2024)."},{"key":"e_1_3_2_1_52_1","doi-asserted-by":"publisher","DOI":"10.1109\/IPDPS53621.2022.00120"},{"key":"e_1_3_2_1_53_1","doi-asserted-by":"publisher","DOI":"10.1145\/3418463"},{"key":"e_1_3_2_1_54_1","doi-asserted-by":"publisher","DOI":"10.1109\/SC.2008.5214359"},{"key":"e_1_3_2_1_55_1","doi-asserted-by":"publisher","DOI":"10.1145\/2633046"},{"key":"e_1_3_2_1_56_1","volume-title":"Artemis: Automatic Runtime Tuning of Parallel Execution Parameters Using Machine Learning. In International Conference on High Performance Computing. Springer, 453\u2013472","author":"Wood Chad","year":"2021","unstructured":"Chad Wood, Giorgis Georgakoudis, David Beckingsale, David Poliakoff, Alfredo Gimenez, Kevin Huck, Allen Malony, and Todd Gamblin. 2021. Artemis: Automatic Runtime Tuning of Parallel Execution Parameters Using Machine Learning. In International Conference on High Performance Computing. Springer, 453\u2013472."},{"key":"e_1_3_2_1_57_1","volume-title":"ytopt: Autotuning scientific applications for energy efficiency at large scales. arXiv preprint arXiv:2303.16245","author":"Wu Xingfu","year":"2023","unstructured":"Xingfu Wu, Prasanna Balaprakash, Michael Kruse, Jaehoon Koo, Brice Videau, Paul Hovland, Valerie Taylor, Brad Geltz, Siddhartha Jana, and Mary Hall. 2023. ytopt: Autotuning scientific applications for energy efficiency at large scales. arXiv preprint arXiv:2303.16245 (2023)."},{"key":"e_1_3_2_1_58_1","doi-asserted-by":"publisher","DOI":"10.1002\/cpe.6683"},{"key":"e_1_3_2_1_59_1","volume-title":"Google\u2019s neural machine translation system: Bridging the gap between human and machine translation. arXiv preprint arXiv:1609.08144","author":"Wu Yonghui","year":"2016","unstructured":"Yonghui Wu, Mike Schuster, Zhifeng Chen, Quoc\u00a0V Le, Mohammad Norouzi, Wolfgang Macherey, Maxim Krikun, Yuan Cao, Qin Gao, Klaus Macherey, 2016. Google\u2019s neural machine translation system: Bridging the gap between human and machine translation. arXiv preprint arXiv:1609.08144 (2016)."}],"event":{"name":"PACT '24: International Conference on Parallel Architectures and Compilation Techniques","sponsor":["SIGARCH ACM Special Interest Group on Computer Architecture"],"location":"Long Beach CA USA","acronym":"PACT '24"},"container-title":["Proceedings of the 2024 International Conference on Parallel Architectures and Compilation Techniques"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3656019.3676895","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3656019.3676895","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,22]],"date-time":"2025-08-22T19:56:25Z","timestamp":1755892585000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3656019.3676895"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,13]]},"references-count":59,"alternative-id":["10.1145\/3656019.3676895","10.1145\/3656019"],"URL":"https:\/\/doi.org\/10.1145\/3656019.3676895","relation":{},"subject":[],"published":{"date-parts":[[2024,10,13]]},"assertion":[{"value":"2024-10-13","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}