{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,18]],"date-time":"2025-09-18T10:16:49Z","timestamp":1758190609031,"version":"3.44.0"},"reference-count":19,"publisher":"IEEE","license":[{"start":{"date-parts":[[2025,5,29]],"date-time":"2025-05-29T00:00:00Z","timestamp":1748476800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,5,29]],"date-time":"2025-05-29T00:00:00Z","timestamp":1748476800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025,5,29]]},"DOI":"10.1109\/sera65747.2025.11154541","type":"proceedings-article","created":{"date-parts":[[2025,9,16]],"date-time":"2025-09-16T17:31:58Z","timestamp":1758043918000},"page":"158-162","source":"Crossref","is-referenced-by-count":0,"title":["SYNC: SYnergistic aNnotation Collaboration between Humans and LLMs for Enhanced Model Training"],"prefix":"10.1109","author":[{"given":"Tammy","family":"Le","sequence":"first","affiliation":[{"name":"University of Nebraska at Omaha,Computer Science,Nebraska,USA"}]},{"given":"Will","family":"Taylor","sequence":"additional","affiliation":[{"name":"University of Nebraska at Omaha,Computer Science,Nebraska,USA"}]},{"given":"Shradha","family":"Maharjan","sequence":"additional","affiliation":[{"name":"University of Nebraska at Omaha,Computer Science,Nebraska,USA"}]},{"given":"Meng","family":"Xia","sequence":"additional","affiliation":[{"name":"University of Nebraska at Omaha,Computer Science,Nebraska,USA"}]},{"given":"Myoungkyu","family":"Song","sequence":"additional","affiliation":[{"name":"University of Nebraska at Omaha,Computer Science,Nebraska,USA"}]}],"member":"263","reference":[{"key":"ref1","article-title":"Bert: Pre-training of deep bidirectional transformers for language understanding","author":"Devlin","year":"2018","journal-title":"arXiv preprint arXiv:1810.04805"},{"issue":"8","key":"ref2","first-page":"9","article-title":"Language models are unsupervised multitask learners","volume":"1","author":"Radford","year":"2019","journal-title":"OpenAI blog"},{"issue":"140","key":"ref3","first-page":"1","article-title":"Exploring the limits of transfer learning with a unified text-to-text transformer","volume":"21","author":"Raffel","year":"2020","journal-title":"Journal of machine learning research"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/ICSE-FoSE59343.2023.00008"},{"volume-title":"A survey of large language models for code: Evolution, benchmarking, and future trends","year":"2023","author":"Zheng","key":"ref5"},{"volume-title":"Graphcodebert: Pre-training code representations with data flow","year":"2020","author":"Guo","key":"ref6"},{"key":"ref7","first-page":"5110","article-title":"Learning and evaluating contextual embedding of source code","volume-title":"International conference on machine learning","author":"Kanade"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.findings-emnlp.139"},{"key":"ref9","article-title":"Practical program repair in the era of large pre-trained language models","author":"Xia","year":"2022","journal-title":"arXiv preprint arXiv:2210.14179"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1145\/3324884.3416591"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1145\/3368089.3417058"},{"key":"ref12","first-page":"54","article-title":"Treebert: A tree-based pre-trained model for programming language","author":"Jiang","year":"2021","journal-title":"Uncertainty in Artificial Intelligence"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/ICSE43902.2021.00041"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1145\/3510003.3510096"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.naacl-industry.15"},{"key":"ref16","article-title":"Can chatgpt reproduce human-generated labels? a study of social computing tasks","author":"Zhu","year":"2023","journal-title":"arXiv preprint arXiv:2304.10145"},{"key":"ref17","doi-asserted-by":"crossref","DOI":"10.18653\/v1\/D19-1410","article-title":"Sentence-bert: Sentence embeddings using siamese bert-networks","volume-title":"Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing","author":"Reimers"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2022.acl-long.499"},{"key":"ref19","article-title":"Codesearchnet challenge: Evaluating the state of semantic code search","author":"Husain","year":"2019","journal-title":"arXiv preprint arXiv:1909.09436"}],"event":{"name":"2025 IEEE\/ACIS 23rd International Conference on Software Engineering Research, Management and Applications (SERA)","start":{"date-parts":[[2025,5,29]]},"location":"Las Vegas, NV, USA","end":{"date-parts":[[2025,5,31]]}},"container-title":["2025 IEEE\/ACIS 23rd International Conference on Software Engineering Research, Management and Applications (SERA)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/11154458\/11154479\/11154541.pdf?arnumber=11154541","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,17]],"date-time":"2025-09-17T05:34:58Z","timestamp":1758087298000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11154541\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,5,29]]},"references-count":19,"URL":"https:\/\/doi.org\/10.1109\/sera65747.2025.11154541","relation":{},"subject":[],"published":{"date-parts":[[2025,5,29]]}}}