{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,19]],"date-time":"2025-11-19T19:27:10Z","timestamp":1763580430845,"version":"3.40.3"},"publisher-location":"Cham","reference-count":28,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031360206"},{"type":"electronic","value":"9783031360213"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-36021-3_15","type":"book-chapter","created":{"date-parts":[[2023,6,30]],"date-time":"2023-06-30T09:06:16Z","timestamp":1688115976000},"page":"171-179","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["Fine-Tuning Large Language Models for\u00a0Answering Programming Questions with\u00a0Code Snippets"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-8991-9264","authenticated-orcid":false,"given":"Vadim","family":"Lomshakov","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8828-4615","authenticated-orcid":false,"given":"Sergey","family":"Kovalchuk","sequence":"additional","affiliation":[]},{"given":"Maxim","family":"Omelchenko","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7787-2251","authenticated-orcid":false,"given":"Sergey","family":"Nikolenko","sequence":"additional","affiliation":[]},{"given":"Artem","family":"Aliev","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,6,26]]},"reference":[{"key":"15_CR1","doi-asserted-by":"publisher","unstructured":"Beau, N., Crabb\u00e9, B.: The impact of lexical and grammatical processing on generating code from natural language. In: Findings of the Association for Computational Linguistics: ACL 2022, pp. 2204\u20132214. Association for Computational Linguistics, Dublin, Ireland (2022). https:\/\/doi.org\/10.18653\/v1\/2022.findings-acl.173","DOI":"10.18653\/v1\/2022.findings-acl.173"},{"issue":"3","key":"15_CR2","doi-asserted-by":"publisher","first-page":"2258","DOI":"10.1007\/s10664-019-09758-x","volume":"25","author":"S Beyer","year":"2019","unstructured":"Beyer, S., Macho, C., Di Penta, M., Pinzger, M.: What kind of questions do developers ask on Stack Overflow? A comparison of automated approaches to classify posts into question categories. Empir. Softw. Eng. 25(3), 2258\u20132301 (2019). https:\/\/doi.org\/10.1007\/s10664-019-09758-x","journal-title":"Empir. Softw. Eng."},{"key":"15_CR3","doi-asserted-by":"publisher","unstructured":"Black, S., Gao, L., Wang, P., Leahy, C., Biderman, S.: GPT-Neo: large Scale autoregressive language modeling with mesh-tensorflow (2021). https:\/\/doi.org\/10.5281\/zenodo.5297715","DOI":"10.5281\/zenodo.5297715"},{"key":"15_CR4","doi-asserted-by":"publisher","unstructured":"Brown, T.B. et al.: Language models are few-shot learners (2020). https:\/\/doi.org\/10.48550\/ARXIV.2005.14165","DOI":"10.48550\/ARXIV.2005.14165"},{"key":"15_CR5","unstructured":"Chen, M. et al.: Evaluating large language models trained on code. CoRR abs\/2107.03374 (2021), https:\/\/arxiv.org\/abs\/2107.03374"},{"key":"15_CR6","doi-asserted-by":"publisher","unstructured":"Christopoulou, F. et al.: PanGu-Coder: program synthesis with function-level language modeling (2022). https:\/\/doi.org\/10.48550\/ARXIV.2207.11280","DOI":"10.48550\/ARXIV.2207.11280"},{"key":"15_CR7","doi-asserted-by":"crossref","unstructured":"Ding, N. et al.: Openprompt: an open-source framework for prompt-learning. arXiv preprint arXiv:2111.01998 (2021)","DOI":"10.18653\/v1\/2022.acl-demo.10"},{"key":"15_CR8","doi-asserted-by":"publisher","unstructured":"Evtikhiev, M., Bogomolov, E., Sokolov, Y., Bryksin, T.: Out of the bleu: how should we assess quality of the code generation models? (2022). https:\/\/doi.org\/10.48550\/ARXIV.2208.03133","DOI":"10.48550\/ARXIV.2208.03133"},{"key":"15_CR9","unstructured":"Gao, L., et al.: The pile: An 800 GB dataset of diverse text for language modeling. arXiv preprint arXiv:2101.00027 (2020)"},{"key":"15_CR10","doi-asserted-by":"publisher","unstructured":"Hall, P., Hart, J.D.: Bootstrap test for difference between means in nonparametric regression. J. Am. Statist. Assoc. 85(412), 1039\u20131049 (1990). https:\/\/doi.org\/10.1080\/01621459.1990.10474974","DOI":"10.1080\/01621459.1990.10474974"},{"key":"15_CR11","doi-asserted-by":"publisher","unstructured":"Hendrycks, D., Basart, S., Kadavath, S., Mazeika, M., Arora, A., Guo, E., Burns, C., Puranik, S., He, H., Song, D., Steinhardt, J.: Measuring coding challenge competence with apps (2021). https:\/\/doi.org\/10.48550\/ARXIV.2105.09938","DOI":"10.48550\/ARXIV.2105.09938"},{"key":"15_CR12","unstructured":"Kovalchuk, S.V., Lomshakov, V., Aliev, A.: Human perceiving behavior modeling in evaluation of code generation models. In: Proceedings of the 2nd Workshop on Natural Language Generation, Evaluation, and Metrics (GEM), pp. 287\u2013294. ACL, Abu Dhabi, UAE (2022). https:\/\/aclanthology.org\/2022.gem-1.24"},{"key":"15_CR13","doi-asserted-by":"publisher","unstructured":"Lee, N., Li, B.Z., Wang, S., Yih, W.T., Ma, H., Khabsa, M.: Language models as fact checkers? (2020). https:\/\/doi.org\/10.48550\/ARXIV.2006.04102","DOI":"10.48550\/ARXIV.2006.04102"},{"key":"15_CR14","doi-asserted-by":"publisher","unstructured":"Li, Y. et al.: Competition-level code generation with alphacode (2022). https:\/\/doi.org\/10.48550\/ARXIV.2203.07814","DOI":"10.48550\/ARXIV.2203.07814"},{"key":"15_CR15","unstructured":"Lin, C.Y.: ROUGE: a package for automatic evaluation of summaries. In: Text Summarization Branches Out, pp. 74\u201381. Association for Computational Linguistics, Barcelona, Spain (2004). https:\/\/aclanthology.org\/W04-1013"},{"key":"15_CR16","doi-asserted-by":"publisher","unstructured":"Nijkamp, E. et al.: CodeGen: an open large language model for code with multi-turn program synthesis (2022). https:\/\/doi.org\/10.48550\/ARXIV.2203.13474","DOI":"10.48550\/ARXIV.2203.13474"},{"key":"15_CR17","unstructured":"Norouzi, S., Cao, Y.: Semantic parsing with less prior and more monolingual data. CoRR abs\/2101.00259 (2021). https:\/\/arxiv.org\/abs\/2101.00259"},{"key":"15_CR18","doi-asserted-by":"publisher","unstructured":"Petroni, F. et al.: Language models as knowledge bases? (2019). https:\/\/doi.org\/10.48550\/ARXIV.1909.01066","DOI":"10.48550\/ARXIV.1909.01066"},{"key":"15_CR19","doi-asserted-by":"publisher","unstructured":"Ren, S. et al.: CodeBLEU: a method for automatic evaluation of code synthesis (2020). https:\/\/doi.org\/10.48550\/ARXIV.2009.10297","DOI":"10.48550\/ARXIV.2009.10297"},{"key":"15_CR20","doi-asserted-by":"publisher","unstructured":"Roberts, A., Raffel, C., Shazeer, N.: How much knowledge can you pack into the parameters of a language model? (2020). https:\/\/doi.org\/10.48550\/ARXIV.2002.08910","DOI":"10.48550\/ARXIV.2002.08910"},{"issue":"1","key":"15_CR21","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1186\/s44147-022-00159-4","volume":"69","author":"AS Soliman","year":"2022","unstructured":"Soliman, A.S., Hadhoud, M.M., Shaheen, S.I.: MarianCG: a code generation transformer model inspired by machine translation. J. Eng. Appl. Sci. 69(1), 1\u201323 (2022)","journal-title":"J. Eng. Appl. Sci."},{"key":"15_CR22","unstructured":"Song, K., Tan, X., Qin, T., Lu, J., Liu, T.Y.: MPNet: masked and permuted pre-training for language understanding. In: Proceedings of the 34th International Conference on Neural Information Processing Systems. NIPS2020, Curran Associates Inc., Red Hook, NY, USA (2020)"},{"key":"15_CR23","doi-asserted-by":"crossref","unstructured":"Tran, N., Tran, H., Nguyen, S., Nguyen, H., Nguyen, T.: Does BLEU score work for code migration? In: 2019 IEEE\/ACM 27th International Conference on Program Comprehension (ICPC), pp. 165\u2013176 (2019)","DOI":"10.1109\/ICPC.2019.00034"},{"key":"15_CR24","doi-asserted-by":"publisher","unstructured":"Xu, F.F., Alon, U., Neubig, G., Hellendoorn, V.J.: A systematic evaluation of large language models of code (2022). https:\/\/doi.org\/10.48550\/ARXIV.2202.13169","DOI":"10.48550\/ARXIV.2202.13169"},{"key":"15_CR25","doi-asserted-by":"publisher","unstructured":"Ye, Q. et al.: Studying strategically: learning to mask for closed-book QA (2020). https:\/\/doi.org\/10.48550\/ARXIV.2012.15856","DOI":"10.48550\/ARXIV.2012.15856"},{"key":"15_CR26","doi-asserted-by":"crossref","unstructured":"Yin, P., Deng, B., Chen, E., Vasilescu, B., Neubig, G.: Learning to mine aligned code and natural language pairs from stack overflow. In: 2018 IEEE\/ACM 15th Intl. Conf. on Mining Software Repositories (MSR), pp. 476\u2013486 (2018)","DOI":"10.1145\/3196398.3196408"},{"key":"15_CR27","doi-asserted-by":"publisher","unstructured":"Yin, P., Neubig, G.: TRANX: A transition-based neural abstract syntax parser for semantic parsing and code generation. In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, pp. 7\u201312. ACL, Brussels, Belgium (2018). https:\/\/doi.org\/10.18653\/v1\/D18-2002","DOI":"10.18653\/v1\/D18-2002"},{"key":"15_CR28","doi-asserted-by":"publisher","unstructured":"Zhang, T., Kishore, V., Wu, F., Weinberger, K.Q., Artzi, Y.: Bertscore: Evaluating text generation with BERT (2019). https:\/\/doi.org\/10.48550\/ARXIV.1904.09675","DOI":"10.48550\/ARXIV.1904.09675"}],"container-title":["Lecture Notes in Computer Science","Computational Science \u2013 ICCS 2023"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-36021-3_15","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,6,30]],"date-time":"2023-06-30T09:08:08Z","timestamp":1688116088000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-36021-3_15"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031360206","9783031360213"],"references-count":28,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-36021-3_15","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"26 June 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICCS","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Computational Science","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Prague","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Czech Republic","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"3 July 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"5 July 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"iccs-computsci2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/www.iccs-meeting.org\/iccs2023\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"530","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"188","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"94","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"35% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2,8","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3,2","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}