{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,8]],"date-time":"2026-03-08T00:49:21Z","timestamp":1772930961690,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":39,"publisher":"ACM","content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,6,20]]},"DOI":"10.1145\/3755881.3755908","type":"proceedings-article","created":{"date-parts":[[2025,10,27]],"date-time":"2025-10-27T11:46:17Z","timestamp":1761565577000},"page":"426-436","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":1,"title":["Enhancement Report Approval Prediction: A Comparative Study of Large Language Models"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0000-2772-7240","authenticated-orcid":false,"given":"Haosheng","family":"Zuo","sequence":"first","affiliation":[{"name":"State Key Laboratory for Novel Software and Technology, Nanjing University, Nanjing University, Nanjing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4123-4554","authenticated-orcid":false,"given":"Feifei","family":"Niu","sequence":"additional","affiliation":[{"name":"University of Ottawa, Ottawa, Canada"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9270-5072","authenticated-orcid":false,"given":"Chuanyi","family":"Li","sequence":"additional","affiliation":[{"name":"State Key Laboratory for Novel Software and Technology, Nanjing University, Nanjing University, Nanjing, China"}]}],"member":"320","published-online":{"date-parts":[[2025,10,27]]},"reference":[{"key":"e_1_3_3_1_2_2","unstructured":"Josh Achiam Steven Adler Sandhini Agarwal Lama Ahmad Ilge Akkaya Florencia\u00a0Leoni Aleman Diogo Almeida Janko Altenschmidt Sam Altman Shyamal Anadkat et\u00a0al. 2023. Gpt-4 technical report. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2303.08774 (2023)."},{"key":"e_1_3_3_1_3_2","doi-asserted-by":"publisher","DOI":"10.1109\/CCWC51732.2021.9375841"},{"key":"e_1_3_3_1_4_2","doi-asserted-by":"publisher","DOI":"10.1088\/1742-6596\/1828\/1\/012106"},{"key":"e_1_3_3_1_5_2","unstructured":"Tom Brown Benjamin Mann Nick Ryder Melanie Subbiah Jared\u00a0D Kaplan Prafulla Dhariwal Arvind Neelakantan Pranav Shyam Girish Sastry Amanda Askell et\u00a0al. 2020. Language models are few-shot learners. Advances in neural information processing systems 33 (2020) 1877\u20131901."},{"key":"e_1_3_3_1_6_2","doi-asserted-by":"crossref","unstructured":"Jun Cheng Mazhar Sadiq Olga\u00a0A Kalugina Sadeem\u00a0Ahmad Nafees and Qasim Umer. 2021. Convolutional Neural Network Based Approval Prediction of Enhancement Reports. IEEE Access 9 (2021) 122412\u2013122424.","DOI":"10.1109\/ACCESS.2021.3108624"},{"key":"e_1_3_3_1_7_2","unstructured":"Kevin Clark Minh-Thang Luong Quoc\u00a0V Le and Christopher\u00a0D Manning. 2020. Electra: Pre-training text encoders as discriminators rather than generators. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2003.10555 (2020)."},{"key":"e_1_3_3_1_8_2","first-page":"4171","volume-title":"Proceedings of the 2019 conference of the North American chapter of the association for computational linguistics: human language technologies, volume 1 (long and short papers)","author":"Devlin Jacob","year":"2019","unstructured":"Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. Bert: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 conference of the North American chapter of the association for computational linguistics: human language technologies, volume 1 (long and short papers). 4171\u20134186."},{"key":"e_1_3_3_1_9_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-64849-7_18"},{"key":"e_1_3_3_1_10_2","unstructured":"Aaron Grattafiori Abhimanyu Dubey Abhinav Jauhri Abhinav Pandey Abhishek Kadian Ahmad Al-Dahle Aiesha Letman Akhil Mathur Alan Schelten Alex Vaughan et\u00a0al. 2024. The llama 3 herd of models. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2407.21783 (2024)."},{"key":"e_1_3_3_1_11_2","unstructured":"Standish Group et\u00a0al. 2015. CHAOS report 2015. The Standish Group International Inc (2015) 1\u201313."},{"key":"e_1_3_3_1_12_2","unstructured":"Pengcheng He Jianfeng Gao and Weizhu Chen. 2021. Debertav3: Improving deberta using electra-style pre-training with gradient-disentangled embedding sharing. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2111.09543 (2021)."},{"key":"e_1_3_3_1_13_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-39225-7_7"},{"key":"e_1_3_3_1_14_2","doi-asserted-by":"crossref","unstructured":"Sepp Hochreiter and J\u00fcrgen Schmidhuber. 1997. Long short-term memory. Neural computation 9 8 (1997) 1735\u20131780.","DOI":"10.1162\/neco.1997.9.8.1735"},{"key":"e_1_3_3_1_15_2","unstructured":"Edward\u00a0J Hu Yelong Shen Phillip Wallis Zeyuan Allen-Zhu Yuanzhi Li Shean Wang Lu Wang Weizhu Chen et\u00a0al. 2022. Lora: Low-rank adaptation of large language models. ICLR 1 2 (2022) 3."},{"key":"e_1_3_3_1_16_2","doi-asserted-by":"crossref","unstructured":"Shalinka Jayatilleke Richard Lai and Karl Reed. 2018. A method of requirements change analysis. Requirements Engineering 23 4 (2018) 493\u2013508.","DOI":"10.1007\/s00766-017-0277-7"},{"key":"e_1_3_3_1_17_2","unstructured":"Jared Kaplan Sam McCandlish Tom Henighan Tom\u00a0B Brown Benjamin Chess Rewon Child Scott Gray Alec Radford Jeffrey Wu and Dario Amodei. 2020. Scaling laws for neural language models. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2001.08361 (2020)."},{"key":"e_1_3_3_1_18_2","doi-asserted-by":"crossref","unstructured":"Chuanyi Li Liguo Huang Jidong Ge Bin Luo and Vincent Ng. 2018. Automatically classifying user requests in crowdsourcing requirements engineering. Journal of Systems and Software 138 (2018) 108\u2013123.","DOI":"10.1016\/j.jss.2017.12.028"},{"key":"e_1_3_3_1_19_2","unstructured":"Aixin Liu Bei Feng Bing Xue Bingxuan Wang Bochao Wu Chengda Lu Chenggang Zhao Chengqi Deng Chenyu Zhang Chong Ruan et\u00a0al. 2024. Deepseek-v3 technical report. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2412.19437 (2024)."},{"key":"e_1_3_3_1_20_2","unstructured":"Yinhan Liu Myle Ott Naman Goyal Jingfei Du Mandar Joshi Danqi Chen Omer Levy Mike Lewis Luke Zettlemoyer and Veselin Stoyanov. 2019. Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/1907.11692 (2019)."},{"key":"e_1_3_3_1_21_2","doi-asserted-by":"publisher","DOI":"10.1109\/RE.2017.41"},{"key":"e_1_3_3_1_22_2","doi-asserted-by":"publisher","DOI":"10.1109\/IBCAST51254.2021.9393180"},{"key":"e_1_3_3_1_23_2","doi-asserted-by":"publisher","DOI":"10.1145\/3545258.3545265"},{"key":"e_1_3_3_1_24_2","unstructured":"Feifei Niu Chuanyi Li Heng Chen Ji-Dong Ge Bin Luo and Alexander Egyed. 2023. Utilizing Creator Profiles for Predicting Valuable User Enhancement Reports. Authorea Preprints (2023)."},{"key":"e_1_3_3_1_25_2","unstructured":"Feifei Niu Chuanyi Li Jidong Ge and B Luo. 2023. Survey on User Feature Requests Analysis and Processing. Ruan Jian Xue Bao\/Journal of Software (2023) 3605\u20133636. http:\/\/www.jos.org.cn\/1000-9825\/6558.htm (in Chinese)."},{"key":"e_1_3_3_1_26_2","doi-asserted-by":"crossref","unstructured":"Zeeshan\u00a0Ahmed Nizamani Hui Liu David\u00a0Matthew Chen and Zhendong Niu. 2018. Automatic approval prediction for software enhancement requests. Automated Software Engineering 25 2 (2018) 347\u2013381.","DOI":"10.1007\/s10515-017-0229-y"},{"key":"e_1_3_3_1_27_2","unstructured":"OpenAI. 2024. GPT-3.5 Turbo. https:\/\/platform.openai.com\/docs\/models\/gpt-3.5-turbo Accessed: 2025-04-16."},{"key":"e_1_3_3_1_28_2","unstructured":"OpenAI. 2024. GPT-4o-mini. https:\/\/platform.openai.com\/docs\/models\/gpt-4o-mini Accessed: 2025-04-16."},{"key":"e_1_3_3_1_29_2","unstructured":"OpenAI. 2024. GPT-4o Mini: Advancing Cost-Efficient Intelligence. https:\/\/openai.com\/index\/gpt-4o-mini-advancing-cost-efficient-intelligence\/ Accessed: 2025-04-16."},{"key":"e_1_3_3_1_30_2","unstructured":"Alec Radford Karthik Narasimhan Tim Salimans Ilya Sutskever et\u00a0al. 2018. Improving language understanding by generative pre-training. (2018)."},{"key":"e_1_3_3_1_31_2","unstructured":"Alec Radford Jeffrey Wu Rewon Child David Luan Dario Amodei Ilya Sutskever et\u00a0al. 2019. Language models are unsupervised multitask learners. OpenAI blog 1 8 (2019) 9."},{"key":"e_1_3_3_1_32_2","doi-asserted-by":"publisher","DOI":"10.1109\/ASE.2017.8115656"},{"key":"e_1_3_3_1_33_2","unstructured":"Hugo Touvron Thibaut Lavril Gautier Izacard Xavier Martinet Marie-Anne Lachaux Timoth\u00e9e Lacroix Baptiste Rozi\u00e8re Naman Goyal Eric Hambro Faisal Azhar et\u00a0al. 2023. Llama: Open and efficient foundation language models. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2302.13971 (2023)."},{"key":"e_1_3_3_1_34_2","unstructured":"Hugo Touvron Louis Martin Kevin Stone Peter Albert Amjad Almahairi Yasmine Babaei Nikolay Bashlykov Soumya Batra Prajjwal Bhargava Shruti Bhosale et\u00a0al. 2023. Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2307.09288 (2023)."},{"key":"e_1_3_3_1_35_2","doi-asserted-by":"crossref","unstructured":"Qasim Umer Hui Liu and Yasir Sultan. 2019. Sentiment based approval prediction for enhancement reports. Journal of Systems and Software 155 (2019) 57\u201369.","DOI":"10.1016\/j.jss.2019.05.026"},{"key":"e_1_3_3_1_36_2","unstructured":"Ashish Vaswani Noam Shazeer Niki Parmar Jakob Uszkoreit Llion Jones Aidan\u00a0N Gomez \u0141ukasz Kaiser and Illia Polosukhin. 2017. Attention is all you need. Advances in neural information processing systems 30 (2017)."},{"key":"e_1_3_3_1_37_2","doi-asserted-by":"crossref","unstructured":"Congying Xu Xiaobing Sun Bin Li Xintong Lu and Hongjing Guo. 2018. MULAPI: Improving API method recommendation with API usage location. Journal of Systems and Software 142 (2018) 195\u2013205.","DOI":"10.1016\/j.jss.2018.04.060"},{"key":"e_1_3_3_1_38_2","unstructured":"Zhilin Yang Zihang Dai Yiming Yang Jaime Carbonell Russ\u00a0R Salakhutdinov and Quoc\u00a0V Le. 2019. Xlnet: Generalized autoregressive pretraining for language understanding. Advances in neural information processing systems 32 (2019)."},{"key":"e_1_3_3_1_39_2","doi-asserted-by":"crossref","unstructured":"Tao Zhang Jiachi Chen Xian Zhan Xiapu Luo David Lo and He Jiang. 2019. Where2Change: Change request localization for app reviews. IEEE Transactions on Software Engineering 47 11 (2019) 2590\u20132616.","DOI":"10.1109\/TSE.2019.2956941"},{"key":"e_1_3_3_1_40_2","unstructured":"Yaowei Zheng Richong Zhang Junhao Zhang Yanhan Ye Zheyan Luo Zhangchi Feng and Yongqiang Ma. 2024. Llamafactory: Unified efficient fine-tuning of 100+ language models. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2403.13372 (2024)."}],"event":{"name":"Internetware 2025: the 16th International Conference on Internetware","location":"Trondheim Norway","acronym":"Internetware 2025","sponsor":["SIGSOFT ACM Special Interest Group on Artificial Intelligence"]},"container-title":["Proceedings of the 16th International Conference on Internetware"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3755881.3755908","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,27]],"date-time":"2025-10-27T11:52:31Z","timestamp":1761565951000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3755881.3755908"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,6,20]]},"references-count":39,"alternative-id":["10.1145\/3755881.3755908","10.1145\/3755881"],"URL":"https:\/\/doi.org\/10.1145\/3755881.3755908","relation":{},"subject":[],"published":{"date-parts":[[2025,6,20]]},"assertion":[{"value":"2025-10-27","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}