{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,5]],"date-time":"2026-02-05T06:38:22Z","timestamp":1770273502961,"version":"3.49.0"},"publisher-location":"New York, NY, USA","reference-count":35,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,11,12]],"date-time":"2023-11-12T00:00:00Z","timestamp":1699747200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,11,12]]},"DOI":"10.1145\/3624062.3624088","type":"proceedings-article","created":{"date-parts":[[2023,11,10]],"date-time":"2023-11-10T13:53:39Z","timestamp":1699624419000},"page":"215-223","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":15,"title":["Data Race Detection Using Large Language Models"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-3847-4108","authenticated-orcid":false,"given":"Le","family":"Chen","sequence":"first","affiliation":[{"name":"Iowa State University, United States of America and Lawrence Livermore National Laboratory, United States of America"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6114-2801","authenticated-orcid":false,"given":"Xianzhong","family":"Ding","sequence":"additional","affiliation":[{"name":"University of California, Merced, United States of America"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6279-0007","authenticated-orcid":false,"given":"Murali","family":"Emani","sequence":"additional","affiliation":[{"name":"Argonne National Laboratory (ANL), USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0609-3925","authenticated-orcid":false,"given":"Tristan","family":"Vanderbruggen","sequence":"additional","affiliation":[{"name":"Lawrence Livermore National Laboratory, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4977-814X","authenticated-orcid":false,"given":"Pei-Hung","family":"Lin","sequence":"additional","affiliation":[{"name":"Lawrence Livermore National Laboratory, United States of America"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6477-0547","authenticated-orcid":false,"given":"Chunhua","family":"Liao","sequence":"additional","affiliation":[{"name":"Lawrence Livermore National Laboratory, United States of America"}]}],"member":"320","published-online":{"date-parts":[[2023,11,12]]},"reference":[{"key":"e_1_3_2_2_1_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-21487-5_4"},{"key":"e_1_3_2_2_2_1","volume-title":"Language models are few-shot learners. Advances in neural information processing systems 33","author":"Brown Tom","year":"2020","unstructured":"Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared\u00a0D Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, 2020. Language models are few-shot learners. Advances in neural information processing systems 33 (2020), 1877\u20131901."},{"key":"e_1_3_2_2_3_1","volume-title":"LM4HPC: Towards Effective Language Model Application in High-Performance Computing. arXiv preprint arXiv:2306.14979","author":"Chen Le","year":"2023","unstructured":"Le Chen, Pei-Hung Lin, Tristan Vanderbruggen, Chunhua Liao, Murali Emani, and Bronis de Supinski. 2023. LM4HPC: Towards Effective Language Model Application in High-Performance Computing. arXiv preprint arXiv:2306.14979 (2023)."},{"key":"e_1_3_2_2_4_1","volume-title":"Multi-View Learning for Parallelism Discovery of Sequential Programs. In 2022 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPSW). IEEE, 295\u2013303","author":"Chen Le","year":"2022","unstructured":"Le Chen, Quazi\u00a0Ishtiaque Mahmud, and Ali Jannesari. 2022. Multi-View Learning for Parallelism Discovery of Sequential Programs. In 2022 IEEE International Parallel and Distributed Processing Symposium Workshops (IPDPSW). IEEE, 295\u2013303."},{"key":"e_1_3_2_2_5_1","volume-title":"Proceedings of Machine Learning and Systems 5","author":"Chen Le","year":"2023","unstructured":"Le Chen, Quazi\u00a0Ishtiaque Mahmud, Hung Phan, Nesreen Ahmed, and Ali Jannesari. 2023. Learning to Parallelize with OpenMP by Augmented Heterogeneous AST Representation. Proceedings of Machine Learning and Systems 5 (2023)."},{"key":"e_1_3_2_2_6_1","volume-title":"Jared Kaplan, Harri Edwards, Yuri Burda","author":"Chen Mark","year":"2021","unstructured":"Mark Chen, Jerry Tworek, Heewoo Jun, Qiming Yuan, Henrique Ponde de\u00a0Oliveira Pinto, Jared Kaplan, Harri Edwards, Yuri Burda, Nicholas Joseph, Greg Brockman, 2021. Evaluating large language models trained on code. arXiv preprint arXiv:2107.03374 (2021)."},{"key":"e_1_3_2_2_7_1","volume-title":"Qlora: Efficient finetuning of quantized llms. arXiv preprint arXiv:2305.14314","author":"Dettmers Tim","year":"2023","unstructured":"Tim Dettmers, Artidoro Pagnoni, Ari Holtzman, and Luke Zettlemoyer. 2023. Qlora: Efficient finetuning of quantized llms. arXiv preprint arXiv:2305.14314 (2023)."},{"key":"e_1_3_2_2_8_1","volume-title":"Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805","author":"Devlin Jacob","year":"2018","unstructured":"Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018)."},{"key":"e_1_3_2_2_9_1","volume-title":"HPC-GPT: Integrating Large Language Model for High-Performance Computing. In Workshops of The International Conference on High Performance Computing, Network, Storage, and Analysis (SC-W","author":"Ding Xianzhong","year":"2023","unstructured":"Xianzhong Ding, Le Chen, Murali Emani, Chunhua Liao, Pei-Hung Lin, Tristan Vanderbruggen, Zhen Xie, Alberto\u00a0E. Cerpa, and Wan Du. 2023. HPC-GPT: Integrating Large Language Model for High-Performance Computing. In Workshops of The International Conference on High Performance Computing, Network, Storage, and Analysis (SC-W 2023)."},{"key":"e_1_3_2_2_10_1","doi-asserted-by":"publisher","DOI":"10.1109\/IPSN54338.2022.00011"},{"key":"e_1_3_2_2_11_1","doi-asserted-by":"publisher","DOI":"10.1145\/3360322.3360857"},{"key":"e_1_3_2_2_12_1","doi-asserted-by":"publisher","DOI":"10.1145\/3408308.3427986"},{"key":"e_1_3_2_2_13_1","volume-title":"Codebert: A pre-trained model for programming and natural languages. arXiv preprint arXiv:2002.08155","author":"Feng Zhangyin","year":"2020","unstructured":"Zhangyin Feng, Daya Guo, Duyu Tang, Nan Duan, Xiaocheng Feng, Ming Gong, Linjun Shou, Bing Qin, Ting Liu, Daxin Jiang, 2020. Codebert: A pre-trained model for programming and natural languages. arXiv preprint arXiv:2002.08155 (2020)."},{"key":"e_1_3_2_2_14_1","volume-title":"Finding Reusable Machine Learning Components to Build Programming Language Processing Pipelines. arXiv preprint arXiv:2208.05596","author":"Flynn Patrick","year":"2022","unstructured":"Patrick Flynn, Tristan Vanderbruggen, Chunhua Liao, Pei-Hung Lin, Murali Emani, and Xipeng Shen. 2022. Finding Reusable Machine Learning Components to Build Programming Language Processing Pipelines. arXiv preprint arXiv:2208.05596 (2022)."},{"key":"e_1_3_2_2_15_1","unstructured":"Intel. [n. d.]. Inspector. https:\/\/www.intel.com\/content\/www\/us\/en\/developer\/tools\/oneapi\/inspector.html"},{"key":"e_1_3_2_2_16_1","volume-title":"Large language models effectively leverage document-level context for literary translation, but critical errors persist. arXiv preprint arXiv:2304.03245","author":"Karpinska Marzena","year":"2023","unstructured":"Marzena Karpinska and Mohit Iyyer. 2023. Large language models effectively leverage document-level context for literary translation, but critical errors persist. arXiv preprint arXiv:2304.03245 (2023)."},{"key":"e_1_3_2_2_17_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.lindif.2023.102274"},{"key":"e_1_3_2_2_18_1","volume-title":"Creating a Dataset Supporting Translation Between OpenMP Fortran and C++ Code. arXiv preprint arXiv:2307.07686","author":"Lei Bin","year":"2023","unstructured":"Bin Lei, Caiwen Ding, Le Chen, Pei-Hung Lin, and Chunhua Liao. 2023. Creating a Dataset Supporting Translation Between OpenMP Fortran and C++ Code. arXiv preprint arXiv:2307.07686 (2023)."},{"key":"e_1_3_2_2_19_1","volume-title":"Boosting Logical Reasoning in Large Language Models through a New Framework: The Graph of Thought. arXiv preprint arXiv:2308.08614","author":"Lei Bin","year":"2023","unstructured":"Bin Lei, Chunhua Liao, Caiwen Ding, 2023. Boosting Logical Reasoning in Large Language Models through a New Framework: The Graph of Thought. arXiv preprint arXiv:2308.08614 (2023)."},{"key":"e_1_3_2_2_20_1","volume-title":"StarCoder: may the source be with you!arXiv preprint arXiv:2305.06161","author":"Li Raymond","year":"2023","unstructured":"Raymond Li, Loubna\u00a0Ben Allal, Yangtian Zi, Niklas Muennighoff, Denis Kocetkov, Chenghao Mou, Marc Marone, Christopher Akiki, Jia Li, Jenny Chim, 2023. StarCoder: may the source be with you!arXiv preprint arXiv:2305.06161 (2023)."},{"key":"e_1_3_2_2_21_1","doi-asserted-by":"publisher","DOI":"10.1145\/3126908.3126958"},{"key":"e_1_3_2_2_22_1","unstructured":"LLVM. 2023. THREADSANITIZER. https:\/\/clang.llvm.org\/docs\/ThreadSanitizer.html"},{"key":"e_1_3_2_2_23_1","unstructured":"OpenAI. 2023. GPT-4 Technical Report. ArXiv abs\/2303.08774 (2023)."},{"key":"e_1_3_2_2_24_1","doi-asserted-by":"publisher","DOI":"10.1145\/1133255.1134019"},{"key":"e_1_3_2_2_25_1","volume-title":"PanGu-Coder2: Boosting Large Language Models for Code with Ranking Feedback. arXiv preprint arXiv:2307.14936","author":"Shen Bo","year":"2023","unstructured":"Bo Shen, Jiaxin Zhang, Taihong Chen, Daoguang Zan, Bing Geng, An Fu, Muhan Zeng, Ailun Yu, Jichuan Ji, Jingyang Zhao, 2023. PanGu-Coder2: Boosting Large Language Models for Code with Ranking Feedback. arXiv preprint arXiv:2307.14936 (2023)."},{"key":"e_1_3_2_2_26_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICSTW52544.2021.00046"},{"key":"e_1_3_2_2_27_1","volume-title":"Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288","author":"Touvron Hugo","year":"2023","unstructured":"Hugo Touvron, Louis Martin, Kevin Stone, Peter Albert, Amjad Almahairi, Yasmine Babaei, Nikolay Bashlykov, Soumya Batra, Prajjwal Bhargava, Shruti Bhosale, 2023. Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288 (2023)."},{"key":"e_1_3_2_2_28_1","doi-asserted-by":"publisher","DOI":"10.1145\/1287624.1287654"},{"key":"e_1_3_2_2_29_1","volume-title":"Codet5: Identifier-aware unified pre-trained encoder-decoder models for code understanding and generation. arXiv preprint arXiv:2109.00859","author":"Wang Yue","year":"2021","unstructured":"Yue Wang, Weishi Wang, Shafiq Joty, and Steven\u00a0CH Hoi. 2021. Codet5: Identifier-aware unified pre-trained encoder-decoder models for code understanding and generation. arXiv preprint arXiv:2109.00859 (2021)."},{"key":"e_1_3_2_2_30_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.acl-long.411"},{"key":"e_1_3_2_2_31_1","volume-title":"Sentiment Analysis in the Era of Large Language Models: A Reality Check. arXiv preprint arXiv:2305.15005","author":"Zhang Wenxuan","year":"2023","unstructured":"Wenxuan Zhang, Yue Deng, Bing Liu, Sinno\u00a0Jialin Pan, and Lidong Bing. 2023. Sentiment Analysis in the Era of Large Language Models: A Reality Check. arXiv preprint arXiv:2305.15005 (2023)."},{"key":"e_1_3_2_2_32_1","volume-title":"Automatic chain of thought prompting in large language models. arXiv preprint arXiv:2210.03493","author":"Zhang Zhuosheng","year":"2022","unstructured":"Zhuosheng Zhang, Aston Zhang, Mu Li, and Alex Smola. 2022. Automatic chain of thought prompting in large language models. arXiv preprint arXiv:2210.03493 (2022)."},{"key":"e_1_3_2_2_33_1","volume-title":"A survey of large language models. arXiv preprint arXiv:2303.18223","author":"Zhao Wayne\u00a0Xin","year":"2023","unstructured":"Wayne\u00a0Xin Zhao, Kun Zhou, Junyi Li, Tianyi Tang, Xiaolei Wang, Yupeng Hou, Yingqian Min, Beichen Zhang, Junjie Zhang, Zican Dong, 2023. A survey of large language models. arXiv preprint arXiv:2303.18223 (2023)."},{"key":"e_1_3_2_2_34_1","volume-title":"Large language models are human-level prompt engineers. arXiv preprint arXiv:2211.01910","author":"Zhou Yongchao","year":"2022","unstructured":"Yongchao Zhou, Andrei\u00a0Ioan Muresanu, Ziwen Han, Keiran Paster, Silviu Pitis, Harris Chan, and Jimmy Ba. 2022. Large language models are human-level prompt engineers. arXiv preprint arXiv:2211.01910 (2022)."},{"key":"e_1_3_2_2_35_1","volume-title":"Fine-tuning language models from human preferences. arXiv preprint arXiv:1909.08593","author":"Ziegler M","year":"2019","unstructured":"Daniel\u00a0M Ziegler, Nisan Stiennon, Jeffrey Wu, Tom\u00a0B Brown, Alec Radford, Dario Amodei, Paul Christiano, and Geoffrey Irving. 2019. Fine-tuning language models from human preferences. arXiv preprint arXiv:1909.08593 (2019)."}],"event":{"name":"SC-W 2023: Workshops of The International Conference on High Performance Computing, Network, Storage, and Analysis","location":"Denver CO USA","acronym":"SC-W 2023"},"container-title":["Proceedings of the SC '23 Workshops of the International Conference on High Performance Computing, Network, Storage, and Analysis"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3624062.3624088","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3624062.3624088","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,21]],"date-time":"2025-08-21T03:00:59Z","timestamp":1755745259000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3624062.3624088"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,11,12]]},"references-count":35,"alternative-id":["10.1145\/3624062.3624088","10.1145\/3624062"],"URL":"https:\/\/doi.org\/10.1145\/3624062.3624088","relation":{},"subject":[],"published":{"date-parts":[[2023,11,12]]},"assertion":[{"value":"2023-11-12","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}