{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,20]],"date-time":"2026-02-20T18:32:29Z","timestamp":1771612349647,"version":"3.50.1"},"reference-count":33,"publisher":"IEEE","license":[{"start":{"date-parts":[[2023,11,9]],"date-time":"2023-11-09T00:00:00Z","timestamp":1699488000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2023,11,9]],"date-time":"2023-11-09T00:00:00Z","timestamp":1699488000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023,11,9]]},"DOI":"10.1109\/icast57874.2023.10359304","type":"proceedings-article","created":{"date-parts":[[2023,12,21]],"date-time":"2023-12-21T14:23:24Z","timestamp":1703168604000},"page":"192-197","source":"Crossref","is-referenced-by-count":11,"title":["Using Large Language Models for Bug Localization and Fixing"],"prefix":"10.1109","author":[{"given":"Tung Do","family":"Viet","sequence":"first","affiliation":[{"name":"The University of Aizu,Fukushima,Japan"}]},{"given":"Konstantin","family":"Markov","sequence":"additional","affiliation":[{"name":"The University of Aizu,Fukushima,Japan"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1145\/3106237.3106253"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1145\/3510418"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1145\/3296979.3192387"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/ASE.2019.00044"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1145\/3183377.3183383"},{"key":"ref6","article-title":"Automatic program repair with openai\u2019s codex: Evaluating quixbugs","author":"Prenner","year":"2021","journal-title":"ArXiv"},{"key":"ref7","article-title":"Galactica: A large language model for science","author":"Taylor","year":"2022","journal-title":"ArXiv"},{"key":"ref8","article-title":"Stanford alpaca: An instruction-following llama model","author":"Taori","year":"2023"},{"key":"ref9","article-title":"Gpt-3.5: Language model","volume-title":"OpenAI","year":"2021"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1145\/3135932.3135941"},{"key":"ref11","article-title":"Socratic models: Composing zero-shot multimodal reasoning with language","author":"Zeng","year":"2022","journal-title":"ArXiv"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/TAIC.PART.2007.13"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/ICSE.2017.62"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1145\/581396.581397"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1002\/stvr.1509"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/TR.2011.2172031"},{"key":"ref17","article-title":"D&c: A divide-and-conquer approach to ir-based bug localization","author":"Koyuncu","year":"2019"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1145\/2931037.2948705"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/ICSE.2013.6606623"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1145\/2884781.2884807"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/ICSE.2009.5070536"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1145\/3293882.3330577"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1145\/3395363.3397369"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/tse.2019.2940179"},{"key":"ref25","first-page":"1877","article-title":"Language models are few-shot learners","volume":"33","author":"Brown","year":"2020","journal-title":"Advances in neural information processing systems"},{"key":"ref26","article-title":"Training verifiers to solve math word problems","author":"Cobbe","year":"2021"},{"key":"ref27","article-title":"Chain of thought prompting elicits reasoning in large language models","author":"Wei","year":"2022"},{"key":"ref28","article-title":"Large language models are zero-shot reasoners","author":"Kojima","year":"2022"},{"key":"ref29","article-title":"Multitask prompted training enables zero-shot task generalization","author":"Sanh","year":"2021"},{"key":"ref30","article-title":"Neural program repair with program dependence analysis and effective filter mechanism","author":"Zhang","year":"2023","journal-title":"ArXiv"},{"key":"ref31","article-title":"Deep learning for bug-localization in student programs","author":"Gupta","year":"2019"},{"key":"ref32","article-title":"Llama: Open and efficient foundation language models","author":"Touvron","year":"2023"},{"key":"ref33","article-title":"Improving automatically generated code from codex via automated program repair","author":"Fan","year":"2022","journal-title":"ArXiv"}],"event":{"name":"2023 12th International Conference on Awareness Science and Technology (iCAST)","location":"Taichung, Taiwan","start":{"date-parts":[[2023,11,9]]},"end":{"date-parts":[[2023,11,11]]}},"container-title":["2023 12th International Conference on Awareness Science and Technology (iCAST)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/10359246\/10359247\/10359304.pdf?arnumber=10359304","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,1,12]],"date-time":"2024-01-12T18:41:21Z","timestamp":1705084881000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10359304\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,11,9]]},"references-count":33,"URL":"https:\/\/doi.org\/10.1109\/icast57874.2023.10359304","relation":{},"subject":[],"published":{"date-parts":[[2023,11,9]]}}}