{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T19:59:30Z","timestamp":1776110370090,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":68,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,5,11]],"date-time":"2024-05-11T00:00:00Z","timestamp":1715385600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"DOI":"10.13039\/https:\/\/doi.org\/10.13039\/501100010256","name":"Guangzhou Municipal Science and Technology Project","doi-asserted-by":"publisher","award":["No. 2023A03J0011"],"award-info":[{"award-number":["No. 2023A03J0011"]}],"id":[{"id":"10.13039\/https:\/\/doi.org\/10.13039\/501100010256","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Guangzhou Science and Technology Program City-University Joint Funding Project","award":["No. 2023A03J0001"],"award-info":[{"award-number":["No. 2023A03J0001"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,5,11]]},"DOI":"10.1145\/3613904.3641917","type":"proceedings-article","created":{"date-parts":[[2024,5,11]],"date-time":"2024-05-11T08:39:12Z","timestamp":1715416752000},"page":"1-18","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":18,"title":["Evaluating Large Language Models on Academic Literature Understanding and Review: An Empirical Study among Early-stage Scholars"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-0743-0121","authenticated-orcid":false,"given":"Jiyao","family":"Wang","sequence":"first","affiliation":[{"name":"Robotics and Autonomous Systems Thrust, The Hong Kong University of Science and Technology (Guangzhou), China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0007-0681-5162","authenticated-orcid":false,"given":"Haolong","family":"Hu","sequence":"additional","affiliation":[{"name":"Intelligent Transportation Thrust, The Hong Kong University of Science and Technology (Guangzhou), China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0008-5914-3643","authenticated-orcid":false,"given":"Zuyuan","family":"Wang","sequence":"additional","affiliation":[{"name":"Robotics and Autonomous Systems Thrust, The Hong Kong University of Science and Technology (Guangzhou), China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5169-1804","authenticated-orcid":false,"given":"Song","family":"Yan","sequence":"additional","affiliation":[{"name":"Robotics and Autonomous Systems Thrust, The Hong Kong University of Science and Technology (Guangzhou), China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8234-6512","authenticated-orcid":false,"given":"Youyu","family":"Sheng","sequence":"additional","affiliation":[{"name":"Robotics and Autonomous Systems Thrust, The Hong Kong University of Science and Technology (Guangzhou), China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4359-4083","authenticated-orcid":false,"given":"Dengbo","family":"He","sequence":"additional","affiliation":[{"name":"Intelligent Transportation Thrust, The Hong Kong University of Science and Technology (Guangzhou), China"}]}],"member":"320","published-online":{"date-parts":[[2024,5,11]]},"reference":[{"key":"e_1_3_3_3_1_1","volume-title":"Large language models (LLM) and ChatGPT: what will the impact on nuclear medicine be?European Journal of Nuclear Medicine and Molecular Imaging 50, 6","author":"Alberts L","year":"2023","unstructured":"Ian\u00a0L Alberts, Lorenzo Mercolli, Thomas Pyka, George Prenosil, Kuangyu Shi, Axel Rominger, and Ali Afshar-Oromieh. 2023. Large language models (LLM) and ChatGPT: what will the impact on nuclear medicine be?European Journal of Nuclear Medicine and Molecular Imaging 50, 6 (2023), 1549\u20131552."},{"key":"e_1_3_3_3_2_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ajmo.2023.100036"},{"key":"e_1_3_3_3_3_1","volume-title":"OpenAI ChatGPT generated literature review: Digital twin in healthcare. Available at SSRN 4308687","author":"Ayd\u0131n \u00d6mer","year":"2022","unstructured":"\u00d6mer Ayd\u0131n and Enis Karaarslan. 2022. OpenAI ChatGPT generated literature review: Digital twin in healthcare. Available at SSRN 4308687 (2022)."},{"key":"e_1_3_3_3_4_1","volume-title":"Promptsource: An integrated development environment and repository for natural language prompts. arXiv preprint arXiv:2202.01279","author":"Bach H","year":"2022","unstructured":"Stephen\u00a0H Bach, Victor Sanh, Zheng-Xin Yong, Albert Webson, Colin Raffel, Nihal\u00a0V Nayak, Abheesht Sharma, Taewoon Kim, M\u00a0Saiful Bari, Thibault Fevry, 2022. Promptsource: An integrated development environment and repository for natural language prompts. arXiv preprint arXiv:2202.01279 (2022)."},{"key":"e_1_3_3_3_5_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445922"},{"key":"e_1_3_3_3_6_1","volume-title":"On the opportunities and risks of foundation models. arXiv preprint arXiv:2108.07258","author":"Bommasani Rishi","year":"2021","unstructured":"Rishi Bommasani, Drew\u00a0A Hudson, Ehsan Adeli, Russ Altman, Simran Arora, Sydney von Arx, Michael\u00a0S Bernstein, Jeannette Bohg, Antoine Bosselut, Emma Brunskill, 2021. On the opportunities and risks of foundation models. arXiv preprint arXiv:2108.07258 (2021)."},{"key":"e_1_3_3_3_7_1","doi-asserted-by":"publisher","DOI":"10.1145\/3236669"},{"key":"e_1_3_3_3_8_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-49065-2_27"},{"key":"e_1_3_3_3_9_1","doi-asserted-by":"publisher","DOI":"10.1111\/j.1467-8535.2008.00914.x"},{"key":"e_1_3_3_3_10_1","doi-asserted-by":"publisher","DOI":"10.1145\/2858036.2858498"},{"key":"e_1_3_3_3_11_1","doi-asserted-by":"publisher","DOI":"10.1089\/dst.2013.0013"},{"key":"e_1_3_3_3_12_1","volume-title":"CHI Conference on Human Factors in Computing Systems Extended Abstracts. 1\u20134.","author":"Joon\u00a0Young Chung John","year":"2022","unstructured":"John Joon\u00a0Young Chung, Wooseok Kim, Kang\u00a0Min Yoo, Hwaran Lee, Eytan Adar, and Minsuk Chang. 2022. TaleBrush: visual sketching of story generation with pretrained language models. In CHI Conference on Human Factors in Computing Systems Extended Abstracts. 1\u20134."},{"key":"e_1_3_3_3_13_1","volume-title":"Perceived usefulness, perceived ease of use, and user acceptance of information technology. MIS Quarterly","author":"Davis D","year":"1989","unstructured":"Fred\u00a0D Davis. 1989. Perceived usefulness, perceived ease of use, and user acceptance of information technology. MIS Quarterly (1989), 319\u2013340."},{"key":"e_1_3_3_3_14_1","doi-asserted-by":"publisher","DOI":"10.5114\/biolsport.2023.125623"},{"key":"e_1_3_3_3_15_1","volume-title":"Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805","author":"Devlin Jacob","year":"2018","unstructured":"Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018)."},{"key":"e_1_3_3_3_16_1","doi-asserted-by":"publisher","DOI":"10.1038\/d41586-023-00288-7"},{"key":"e_1_3_3_3_17_1","volume-title":"Unified language model pre-training for natural language understanding and generation. Advances in Neural Information Processing Systems 32","author":"Dong Li","year":"2019","unstructured":"Li Dong, Nan Yang, Wenhui Wang, Furu Wei, Xiaodong Liu, Yu Wang, Jianfeng Gao, Ming Zhou, and Hsiao-Wuen Hon. 2019. Unified language model pre-training for natural language understanding and generation. Advances in Neural Information Processing Systems 32 (2019)."},{"key":"e_1_3_3_3_18_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.frl.2023.103662"},{"key":"e_1_3_3_3_19_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-82199-9_12"},{"key":"e_1_3_3_3_20_1","doi-asserted-by":"publisher","DOI":"10.3390\/info10040143"},{"key":"e_1_3_3_3_21_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.aorn.2007.07.001"},{"key":"e_1_3_3_3_22_1","doi-asserted-by":"publisher","DOI":"10.1101\/2022.12.23.521610"},{"key":"e_1_3_3_3_23_1","doi-asserted-by":"publisher","DOI":"10.1101\/2022.12.23.521610"},{"key":"e_1_3_3_3_24_1","volume-title":"Realtoxicityprompts: Evaluating neural toxic degeneration in language models. arXiv preprint arXiv:2009.11462","author":"Gehman Samuel","year":"2020","unstructured":"Samuel Gehman, Suchin Gururangan, Maarten Sap, Yejin Choi, and Noah\u00a0A Smith. 2020. Realtoxicityprompts: Evaluating neural toxic degeneration in language models. arXiv preprint arXiv:2009.11462 (2020)."},{"key":"e_1_3_3_3_25_1","doi-asserted-by":"publisher","DOI":"10.1145\/3532106.3533533"},{"key":"e_1_3_3_3_26_1","doi-asserted-by":"publisher","DOI":"10.1145\/3544548.3580782"},{"key":"e_1_3_3_3_27_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11019-023-10136-0"},{"key":"e_1_3_3_3_28_1","doi-asserted-by":"publisher","DOI":"10.1021\/acsenergylett.2c02828"},{"key":"e_1_3_3_3_29_1","doi-asserted-by":"crossref","unstructured":"Mohanad Halaweh. 2023. ChatGPT in education: Strategies for responsible implementation. (2023).","DOI":"10.30935\/cedtech\/13036"},{"key":"e_1_3_3_3_30_1","doi-asserted-by":"publisher","DOI":"10.3389\/ffutr.2022.772910"},{"key":"e_1_3_3_3_31_1","volume-title":"Prior familiarization with takeover requests affects drivers","author":"Hergeth Sebastian","year":"2017","unstructured":"Sebastian Hergeth, Lutz Lorenz, and Josef\u00a0F Krems. 2017. Prior familiarization with takeover requests affects drivers\u2019 takeover performance and automation trust. Human factors 59, 3 (2017), 457\u2013470."},{"key":"e_1_3_3_3_32_1","doi-asserted-by":"publisher","DOI":"10.1177\/0018720814547570"},{"key":"e_1_3_3_3_33_1","volume-title":"Could AI help you to write your next paper?Nature 611, 7934","author":"Hutson Matthew","year":"2022","unstructured":"Matthew Hutson. 2022. Could AI help you to write your next paper?Nature 611, 7934 (2022), 192\u2013193."},{"key":"e_1_3_3_3_34_1","volume-title":"How can we know what language models know?Transactions of the Association for Computational Linguistics 8","author":"Jiang Zhengbao","year":"2020","unstructured":"Zhengbao Jiang, Frank\u00a0F Xu, Jun Araki, and Graham Neubig. 2020. How can we know what language models know?Transactions of the Association for Computational Linguistics 8 (2020), 423\u2013438."},{"key":"e_1_3_3_3_35_1","volume-title":"Using ChatGPT and Other Large Language Model","author":"Jungherr Andreas","year":"2023","unstructured":"Andreas Jungherr. 2023. Using ChatGPT and Other Large Language Model (LLM) Applications for Academic Paper Assignments. (2023)."},{"key":"e_1_3_3_3_36_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.lindif.2023.102274"},{"key":"e_1_3_3_3_37_1","doi-asserted-by":"publisher","DOI":"10.1145\/2858036.2858402"},{"key":"e_1_3_3_3_38_1","doi-asserted-by":"publisher","DOI":"10.2307\/3001469"},{"key":"e_1_3_3_3_39_1","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pdig.0000198"},{"key":"e_1_3_3_3_40_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2021.05.103"},{"key":"e_1_3_3_3_41_1","doi-asserted-by":"publisher","DOI":"10.1518\/hfes.46.1.50.30392"},{"key":"e_1_3_3_3_42_1","volume-title":"Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. arXiv preprint arXiv:1910.13461","author":"Lewis Mike","year":"2019","unstructured":"Mike Lewis, Yinhan Liu, Naman Goyal, Marjan Ghazvininejad, Abdelrahman Mohamed, Omer Levy, Ves Stoyanov, and Luke Zettlemoyer. 2019. Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. arXiv preprint arXiv:1910.13461 (2019)."},{"key":"e_1_3_3_3_43_1","doi-asserted-by":"publisher","DOI":"10.1016\/S2589-7500(23)00019-5"},{"key":"e_1_3_3_3_44_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ergon.2012.09.001"},{"key":"e_1_3_3_3_45_1","first-page":"1","article-title":"Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing","volume":"55","author":"Liu Pengfei","year":"2023","unstructured":"Pengfei Liu, Weizhe Yuan, Jinlan Fu, Zhengbao Jiang, Hiroaki Hayashi, and Graham Neubig. 2023. Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing. Comput. Surveys 55, 9 (2023), 1\u201335.","journal-title":"Comput. Surveys"},{"key":"e_1_3_3_3_46_1","volume-title":"Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 2: Short Papers). 182\u2013189","author":"Luccioni Alexandra","year":"2021","unstructured":"Alexandra Luccioni and Joseph Viviano. 2021. What\u2019s in the box? an analysis of undesirable content in the Common Crawl corpus. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 2: Short Papers). 182\u2013189."},{"key":"e_1_3_3_3_47_1","doi-asserted-by":"publisher","DOI":"10.1080\/00313831.2011.568674"},{"key":"e_1_3_3_3_48_1","volume-title":"Exploring the role of artificial intelligence in enhancing academic performance: A case study of ChatGPT. Available at SSRN","author":"M\u00a0Alshater Muneer","year":"2022","unstructured":"Muneer M\u00a0Alshater. 2022. Exploring the role of artificial intelligence in enhancing academic performance: A case study of ChatGPT. Available at SSRN (2022)."},{"key":"e_1_3_3_3_49_1","doi-asserted-by":"publisher","DOI":"10.1177\/1071181321651063"},{"key":"e_1_3_3_3_50_1","first-page":"47","article-title":"Strategies in study time allocation: Why is study time sometimes not effective?Journal of Experimental Psychology","volume":"122","author":"Mazzoni Giuliana","year":"1993","unstructured":"Giuliana Mazzoni and Cesare Cornoldi. 1993. Strategies in study time allocation: Why is study time sometimes not effective?Journal of Experimental Psychology: General 122, 1 (1993), 47.","journal-title":"General"},{"key":"e_1_3_3_3_51_1","doi-asserted-by":"publisher","DOI":"10.1186\/s13040-023-00339-9"},{"key":"e_1_3_3_3_52_1","volume-title":"Perspectives on the Potential for Generative AI in their Fields. arXiv preprint arXiv:2304.01420","author":"Morris Meredith\u00a0Ringel","year":"2023","unstructured":"Meredith\u00a0Ringel Morris. 2023. Scientists\u2019 Perspectives on the Potential for Generative AI in their Fields. arXiv preprint arXiv:2304.01420 (2023)."},{"key":"e_1_3_3_3_53_1","doi-asserted-by":"publisher","DOI":"10.1145\/3491101.3503719"},{"key":"e_1_3_3_3_54_1","doi-asserted-by":"publisher","DOI":"10.1518\/001872097778543886"},{"key":"e_1_3_3_3_55_1","volume-title":"Red teaming language models with language models. arXiv preprint arXiv:2202.03286","author":"Perez Ethan","year":"2022","unstructured":"Ethan Perez, Saffron Huang, Francis Song, Trevor Cai, Roman Ring, John Aslanides, Amelia Glaese, Nat McAleese, and Geoffrey Irving. 2022. Red teaming language models with language models. arXiv preprint arXiv:2202.03286 (2022)."},{"key":"e_1_3_3_3_56_1","doi-asserted-by":"publisher","DOI":"10.1145\/3544548.3580907"},{"key":"e_1_3_3_3_57_1","doi-asserted-by":"publisher","DOI":"10.52631\/jemds.v3i1.175"},{"key":"e_1_3_3_3_58_1","volume-title":"Fast and frugal heuristics. Encyclopedia of Social Psychology","author":"Reimer Torsten","year":"2007","unstructured":"Torsten Reimer and J\u00f6rg Rieskamp. 2007. Fast and frugal heuristics. Encyclopedia of Social Psychology (2007), 346\u2013348."},{"key":"e_1_3_3_3_59_1","doi-asserted-by":"publisher","DOI":"10.58496\/MJCSC\/2023\/006"},{"key":"e_1_3_3_3_60_1","unstructured":"Anuschka Schmitt Thiemo Wambsganss and Andreas Janson. 2022. Designing for conversational system trustworthiness: the impact of model transparency on trust and task performance. (2022)."},{"key":"e_1_3_3_3_61_1","unstructured":"Horrock Stevens. 2019. What Human Factors Isn\u2019t: 1. Common Sense. https:\/\/humanisticsystems.com\/2019\/07\/10\/what-human-factors-isnt-1-common-sense\/"},{"key":"e_1_3_3_3_62_1","volume-title":"Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971","author":"Touvron Hugo","year":"2023","unstructured":"Hugo Touvron, Thibaut Lavril, Gautier Izacard, Xavier Martinet, Marie-Anne Lachaux, Timoth\u00e9e Lacroix, Baptiste Rozi\u00e8re, Naman Goyal, Eric Hambro, Faisal Azhar, 2023. Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971 (2023)."},{"key":"e_1_3_3_3_63_1","volume-title":"Attention is all you need. Advances in Neural Information Processing Systems 30","author":"Vaswani Ashish","year":"2017","unstructured":"Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan\u00a0N Gomez, \u0141ukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. Advances in Neural Information Processing Systems 30 (2017)."},{"key":"e_1_3_3_3_64_1","volume-title":"Engineering psychology and human performance","author":"Wickens D","unstructured":"Christopher\u00a0D Wickens, William\u00a0S Helton, Justin\u00a0G Hollands, and Simon Banbury. 2021. Engineering psychology and human performance. Routledge."},{"key":"e_1_3_3_3_65_1","doi-asserted-by":"publisher","DOI":"10.1037\/e578042012-008"},{"key":"e_1_3_3_3_66_1","volume-title":"Xlnet: Generalized autoregressive pretraining for language understanding. Advances in Neural Information Processing Systems 32","author":"Yang Zhilin","year":"2019","unstructured":"Zhilin Yang, Zihang Dai, Yiming Yang, Jaime Carbonell, Russ\u00a0R Salakhutdinov, and Quoc\u00a0V Le. 2019. Xlnet: Generalized autoregressive pretraining for language understanding. Advances in Neural Information Processing Systems 32 (2019)."},{"key":"e_1_3_3_3_67_1","doi-asserted-by":"publisher","DOI":"10.1109\/IICSPI.2018.8690387"},{"key":"e_1_3_3_3_68_1","doi-asserted-by":"publisher","DOI":"10.1145\/3474085.3475505"}],"event":{"name":"CHI '24: CHI Conference on Human Factors in Computing Systems","location":"Honolulu HI USA","acronym":"CHI '24","sponsor":["SIGCHI ACM Special Interest Group on Computer-Human Interaction","SIGACCESS ACM Special Interest Group on Accessible Computing"]},"container-title":["Proceedings of the CHI Conference on Human Factors in Computing Systems"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3613904.3641917","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3613904.3641917","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T23:57:28Z","timestamp":1750291048000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3613904.3641917"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,5,11]]},"references-count":68,"alternative-id":["10.1145\/3613904.3641917","10.1145\/3613904"],"URL":"https:\/\/doi.org\/10.1145\/3613904.3641917","relation":{},"subject":[],"published":{"date-parts":[[2024,5,11]]},"assertion":[{"value":"2024-05-11","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}