{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,21]],"date-time":"2026-01-21T13:54:08Z","timestamp":1769003648632,"version":"3.49.0"},"reference-count":43,"publisher":"IEEE","funder":[{"DOI":"10.13039\/100000001","name":"National Science Foundation","doi-asserted-by":"publisher","award":["2047120"],"award-info":[{"award-number":["2047120"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024,5]]},"DOI":"10.23919\/isc.2024.10528929","type":"proceedings-article","created":{"date-parts":[[2024,5,10]],"date-time":"2024-05-10T17:22:23Z","timestamp":1715361743000},"page":"1-12","source":"Crossref","is-referenced-by-count":36,"title":["HPC-Coder: Modeling Parallel Programs using Large Language Models"],"prefix":"10.23919","author":[{"given":"Daniel","family":"Nichols","sequence":"first","affiliation":[{"name":"University of Maryland,Department of Computer Science,College Park,MD,USA"}]},{"given":"Aniruddha","family":"Marathe","sequence":"additional","affiliation":[{"name":"Lawrence Livermore National Laboratory,Center for Applied Scientific Computing,Livermore,CA,USA"}]},{"given":"Harshitha","family":"Menon","sequence":"additional","affiliation":[{"name":"Lawrence Livermore National Laboratory,Center for Applied Scientific Computing,Livermore,CA,USA"}]},{"given":"Todd","family":"Gamblin","sequence":"additional","affiliation":[{"name":"Lawrence Livermore National Laboratory,Livermore Computing,Livermore,CA,USA"}]},{"given":"Abhinav","family":"Bhatele","sequence":"additional","affiliation":[{"name":"University of Maryland,Department of Computer Science,College Park,MD,USA"}]}],"member":"263","reference":[{"key":"ref1","author":"Zhao","year":"2023","journal-title":"A survey of large language models"},{"key":"ref2","author":"Chen","year":"2021","journal-title":"Evaluating large language models trained on code"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1108\/ws.2000.07949fab.004"},{"key":"ref4","author":"Roziere","year":"2023","journal-title":"Code llama: Open foundation models for code"},{"issue":"13","key":"ref5","article-title":"Android mobile malware detection using machine learning: A systematic review","volume-title":"Electronics","volume":"10","author":"Senanayake","year":"2021"},{"key":"ref6","volume-title":"M14code","year":"2022"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/SANER53432.2022.00112"},{"key":"ref8","article-title":"Learning code summarization from a small and local dataset","volume":"abs\/2206.00804","author":"Ahmed","year":"2022","journal-title":"ArXiv"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1145\/nnnnnnn.nnnnnnn"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.449"},{"key":"ref11","article-title":"Can we learn from developer mistakes? learning to localize and repair real bugs from real bug fixes","volume":"abs\/2207.00301","author":"Richter","year":"2022","journal-title":"ArXiv"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1145\/3510003.3510153"},{"key":"ref13","article-title":"Attention is all you need","volume-title":"CoRR","volume":"abs\/1706.03762","author":"Vaswani","year":"2017"},{"key":"ref14","doi-asserted-by":"crossref","DOI":"10.1145\/3520312.3534862","volume-title":"A Systematic Evaluation of Large Language Models of Code","author":"Xu","year":"2022"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1145\/3359591.3359735"},{"key":"ref16","author":"Radford","year":"2019","journal-title":"Language models are unsupervised multitask learners"},{"key":"ref17","article-title":"KRIPKE-a massively parallel transport mini-app","author":"Kunen","year":"2015","journal-title":"Lawrence Livermore National Laboratory (LLNL), Livermore, CA, Tech. Rep"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1137\/120864672"},{"key":"ref19","doi-asserted-by":"crossref","DOI":"10.1126\/science.abq1158","volume-title":"Competition-level code generation with alphacode","author":"Li","year":"2022"},{"key":"ref20","volume-title":"Openwebtext corpus","author":"Gokaslan","year":"2019"},{"key":"ref21","article-title":"GPT-Neo: Large Scale Autoregressive Language Modeling with Mesh-Tensorflow","author":"Black","year":"2021","journal-title":"If you use this software, please cite it using these metadata"},{"key":"ref22","article-title":"The pile: An 800gb dataset of diverse text for language modeling","volume-title":"CoRR","volume":"abs\/2101.00027","author":"Gao","year":"2021"},{"key":"ref23","article-title":"Language models are few-shot learners","volume-title":"CoRR","volume":"abs\/2005.14165","author":"Brown","year":"2020"},{"key":"ref24","article-title":"OpenAI","year":"2023","journal-title":"Gpt-4 technical report"},{"key":"ref25","first-page":"38","article-title":"Transformers: State-of-the-Art Natural Language Processing","author":"Wolf","year":"2020","journal-title":"Association for Computational Linguistics"},{"key":"ref26","volume-title":"Deepspeed: Extreme-scale model training for everyone"},{"key":"ref27","article-title":"Zero-offload: Democratizing billion-scale model training","volume-title":"CoRR","volume":"abs\/2101.06840","author":"Ren","year":"2021"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1145\/3320060"},{"key":"ref29","author":"Nichols","year":"2022","journal-title":"A survey and empirical evaluation of parallel deep learning frameworks"},{"key":"ref30","article-title":"Fixing weight decay regularization in adam","volume-title":"CoRR","volume":"abs\/1711.05101","author":"Loshchilov","year":"2017"},{"key":"ref31","author":"Chen","year":"2021","journal-title":"Evaluating large language models trained on code"},{"key":"ref32","first-page":"4171","article-title":"BERT: Pre-training of deep bidirectional transformers for language understanding","volume-title":"Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers)","author":"Devlin"},{"key":"ref33","article-title":"Roberta: A robustly optimized BERT pretraining approach","volume-title":"CoRR","volume":"abs\/1907.11692","author":"Liu","year":"2019"},{"key":"ref34","article-title":"Magicoder: Source code is all you need","author":"Wei","year":"2023","journal-title":"arXiv preprint"},{"key":"ref35","article-title":"Piloting copilot and codex: Hot temperature, cold prompts, or black magic?","volume":"abs\/2210.14699","author":"Doderlein","year":"2022","journal-title":"ArXiv"},{"key":"ref36","article-title":"Grounded copilot: How programmers interact with code-generating models","volume":"abs\/2206.15000","author":"Barke","year":"2022","journal-title":"ArXiv"},{"key":"ref37","article-title":"What is it like to program with artificial intelligence?","volume":"abs\/2208.06213","author":"Sarkar","year":"2022","journal-title":"ArXiv"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1145\/3624062.3624088"},{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1016\/j.future.2024.05.034"},{"key":"ref40","volume-title":"code2vec: Learning distributed representations of code","author":"Alon","year":"2018"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1145\/3418463"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1145\/3540250.3549096"},{"key":"ref43","volume-title":"Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension","author":"Lewis","year":"2019"}],"event":{"name":"ISC High Performance 2024 Research Paper Proceedings (39th International Conference)","location":"Hamburg, Germany","start":{"date-parts":[[2024,5,12]]},"end":{"date-parts":[[2024,5,16]]}},"container-title":["ISC High Performance 2024 Research Paper Proceedings (39th International Conference)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/10528919\/10528920\/10528929.pdf?arnumber=10528929","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,27]],"date-time":"2024-11-27T00:16:20Z","timestamp":1732666580000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10528929\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,5]]},"references-count":43,"URL":"https:\/\/doi.org\/10.23919\/isc.2024.10528929","relation":{},"subject":[],"published":{"date-parts":[[2024,5]]}}}