{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,14]],"date-time":"2026-03-14T12:09:09Z","timestamp":1773490149331,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":46,"publisher":"ACM","funder":[{"name":"Notre Dame-IBM Technology Ethics Lab","award":["11199"],"award-info":[{"award-number":["11199"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2026,3,23]]},"DOI":"10.1145\/3742414.3794723","type":"proceedings-article","created":{"date-parts":[[2026,3,9]],"date-time":"2026-03-09T11:03:52Z","timestamp":1773054232000},"page":"61-65","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["From Verification Burden to Trusted Collaboration: Design Goals for LLM-Assisted Literature Reviews"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-1936-9863","authenticated-orcid":false,"given":"Brenda","family":"Nogueira","sequence":"first","affiliation":[{"name":"Lucy Institute for Data and Society, University of Notre Dame, Notre Dame, Indiana, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4699-5026","authenticated-orcid":false,"given":"Werner","family":"Geyer","sequence":"additional","affiliation":[{"name":"IBM Research, Cambridge, Massachusetts, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4964-6059","authenticated-orcid":false,"given":"Andrew A.","family":"Anderson","sequence":"additional","affiliation":[{"name":"IBM Research, Almaden, California, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7902-7625","authenticated-orcid":false,"given":"Toby Jia-Jun","family":"Li","sequence":"additional","affiliation":[{"name":"Department of Computer Science and Engineering, University of Notre Dame, Notre Dame, Indiana, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0008-0125-9352","authenticated-orcid":false,"given":"Dongwhi","family":"Kim","sequence":"additional","affiliation":[{"name":"University of Notre Dame, Notre Dame, Indiana, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4322-1076","authenticated-orcid":false,"given":"Nuno","family":"Moniz","sequence":"additional","affiliation":[{"name":"Lucy Family Institute for Data &amp; Society, University of Notre Dame, Notre Dame, Indiana, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3932-5956","authenticated-orcid":false,"given":"Nitesh V","family":"Chawla","sequence":"additional","affiliation":[{"name":"Department of Computer Science and Engineering, University of Notre Dame, South Bend, Indiana, USA"}]}],"member":"320","published-online":{"date-parts":[[2026,3,22]]},"reference":[{"key":"e_1_3_3_1_2_2","doi-asserted-by":"publisher","unstructured":"Shubham Agarwal Gaurav Sahu Abhay Puri et\u00a0al. 2024. LLMs for Literature Review: Are we there yet? arxiv:https:\/\/arXiv.org\/abs\/2412.15249\u00a0[cs.IR] 10.48550\/arxiv.2412.15249arXiv preprint.","DOI":"10.48550\/arxiv.2412.15249"},{"key":"e_1_3_3_1_3_2","doi-asserted-by":"crossref","unstructured":"Norin Arshed and Mike Danson. 2015. The literature review. Research methods for business and management: a guide to writing your dissertation (2015) 31\u201349.","DOI":"10.23912\/978-1-910158-51-7-2790"},{"key":"e_1_3_3_1_4_2","doi-asserted-by":"crossref","unstructured":"Christopher\u00a0A Bail. 2024. Can Generative AI improve social science? Proceedings of the National Academy of Sciences 121 21 (2024) e2314021121.","DOI":"10.1073\/pnas.2314021121"},{"key":"e_1_3_3_1_5_2","doi-asserted-by":"crossref","unstructured":"Francisco Bolanos Angelo Salatino Francesco Osborne and Enrico Motta. 2024. Artificial intelligence for literature reviews: Opportunities and challenges. Artificial Intelligence Review 57 10 (2024) 259.","DOI":"10.1007\/s10462-024-10902-3"},{"key":"e_1_3_3_1_6_2","doi-asserted-by":"crossref","unstructured":"Virginia Braun and Victoria Clarke. 2006. Using thematic analysis in psychology. Qualitative research in psychology 3 2 (2006) 77\u2013101.","DOI":"10.1191\/1478088706qp063oa"},{"key":"e_1_3_3_1_7_2","doi-asserted-by":"publisher","DOI":"10.1145\/3477495.3532065"},{"key":"e_1_3_3_1_8_2","doi-asserted-by":"crossref","unstructured":"Xiuying Chen Hind Alamro Mingzhe Li Shen Gao Xiangliang Zhang Dongyan Zhao and Rui Yan. 2021. Capturing relations between scientific papers: An abstractive model for related work section generation. Association for Computational Linguistics (ACL).","DOI":"10.18653\/v1\/2021.acl-long.473"},{"key":"e_1_3_3_1_9_2","unstructured":"Nicholas Crispino Kyle Montgomery Fankun Zeng Dawn Song and Chenguang Wang. 2023. Agent instructs large language models to be general zero-shot reasoners. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2310.03710 (2023)."},{"key":"e_1_3_3_1_10_2","unstructured":"Yunfan Gao Yun Xiong Xinyu Gao Kangxiang Jia Jinliu Pan Yuxi Bi Yixin Dai Jiawei Sun Haofen Wang and Haofen Wang. 2023. Retrieval-augmented generation for large language models: A survey. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2312.10997 2 1 (2023)."},{"key":"e_1_3_3_1_11_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.acl-long.116"},{"key":"e_1_3_3_1_12_2","doi-asserted-by":"publisher","unstructured":"Julia Gehrmann Lars Quakulinski and Oya Beyan. 2024. Large Language Models for Literature Reviews - an Exemplary Comparison of LLM-based Approaches with Manual Methods. ieee:1085244710.1109\/fllm63129.2024.10852447Proceedings paper.","DOI":"10.1109\/fllm63129.2024.10852447"},{"key":"e_1_3_3_1_13_2","unstructured":"Mourad Gridach Ameer Shaban Naci Kocyigit and et al.2025. Agentic Artificial Intelligence for Scientific Discovery. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2503.08979 (2025). https:\/\/arxiv.org\/abs\/2503.08979"},{"key":"e_1_3_3_1_14_2","doi-asserted-by":"crossref","unstructured":"Dritjon Gruda. 2024. Three ways ChatGPT helps me in my academic writing. Nature 10 (2024).","DOI":"10.1038\/d41586-024-01042-3"},{"key":"e_1_3_3_1_15_2","first-page":"427","volume-title":"Coling 2010: Posters","author":"Hoang Cong Duy\u00a0Vu","year":"2010","unstructured":"Cong Duy\u00a0Vu Hoang and Min-Yen Kan. 2010. Towards automated related work summarization. In Coling 2010: Posters. 427\u2013435."},{"key":"e_1_3_3_1_16_2","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-56060-6_24"},{"key":"e_1_3_3_1_17_2","unstructured":"Jingshan Huang and Ming Tan. 2023. The role of ChatGPT in scientific communication: writing better scientific review articles. American journal of cancer research 13 4 (2023) 1148."},{"key":"e_1_3_3_1_18_2","doi-asserted-by":"crossref","unstructured":"Lei Huang Weijiang Yu Weitao Ma Weihong Zhong Zhangyin Feng Haotian Wang Qianglong Chen Weihua Peng Xiaocheng Feng Bing Qin et\u00a0al. 2025. A survey on hallucination in large language models: Principles taxonomy challenges and open questions. ACM Transactions on Information Systems 43 2 (2025) 1\u201355.","DOI":"10.1145\/3703155"},{"key":"e_1_3_3_1_19_2","unstructured":"Yuzhong Huang Baolin Peng Tianyi Shen Yutong Zhang and Jianfeng Gao. 2024. A Survey on Agentic Large Language Models: Architectures Applications and Open Problems. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2407.01502 (2024). https:\/\/arxiv.org\/abs\/2407.01502"},{"key":"e_1_3_3_1_20_2","unstructured":"Prabha Kannan. 2024. How much research is being written by large language models?https:\/\/hai.stanford.edu\/news\/how-much-research-being-written-large-language-models"},{"key":"e_1_3_3_1_21_2","unstructured":"Patrick Lewis Ethan Perez Aleksandra Piktus Fabio Petroni Vladimir Karpukhin Naman Goyal Heinrich K\u00fcttler Mike Lewis Wen-tau Yih Tim Rockt\u00e4schel et\u00a0al. 2020. Retrieval-augmented generation for knowledge-intensive nlp tasks. Advances in neural information processing systems 33 (2020) 9459\u20139474."},{"key":"e_1_3_3_1_22_2","unstructured":"Junyi Li Jie Chen Ruiyang Ren Xiaoxue Cheng Wayne\u00a0Xin Zhao Jian-Yun Nie and Ji-Rong Wen. 2024. The dawn after the dark: An empirical study on factuality hallucination in large language models. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2401.03205 (2024)."},{"key":"e_1_3_3_1_23_2","unstructured":"Junlong Li Jinyuan Wang Zhuosheng Zhang and Hai Zhao. 2022. Self-prompting large language models for zero-shot open-domain QA. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2212.08635 (2022)."},{"key":"e_1_3_3_1_24_2","unstructured":"Yutong Li Lu Chen Aiwei Liu et\u00a0al. 2024. ChatCite: LLM Agent with Human Workflow Guidance for Comparative Literature Summary. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2403.02574 (2024). arxiv:https:\/\/arXiv.org\/abs\/2403.02574\u00a0[cs.CL] https:\/\/arxiv.org\/abs\/2403.02574"},{"key":"e_1_3_3_1_25_2","unstructured":"Weixin Liang Yaohui Zhang Zhengxuan Wu Haley Lepp Wenlong Ji Xuandong Zhao Hancheng Cao Sheng Liu Siyu He Zhi Huang et\u00a0al. 2024. Mapping the increasing use of LLMs in scientific papers. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2404.01268 (2024)."},{"key":"e_1_3_3_1_26_2","unstructured":"Zhehui Liao Maria Antoniak Inyoung Cheong Evie Yu-Yen Cheng Ai-Heng Lee Kyle Lo Joseph\u00a0Chee Chang and Amy\u00a0X Zhang. 2024. LLMs as Research Tools: A Large Scale Survey of Researchers\u2019 Usage and Perceptions. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2411.05025 (2024)."},{"key":"e_1_3_3_1_27_2","unstructured":"Adian Liusie Potsawee Manakul and Mark\u00a0JF Gales. 2023. LLM comparative assessment: Zero-shot NLG evaluation through pairwise comparisons using large language models. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2307.07889 (2023)."},{"key":"e_1_3_3_1_28_2","unstructured":"Lawrence\u00a0A Machi and Brenda\u00a0T McEvoy. 2009. The literature review: Six steps to success. (2009)."},{"key":"e_1_3_3_1_29_2","doi-asserted-by":"publisher","unstructured":"Xiangbin Meng Xiangyu Yan Kuo Zhang et\u00a0al. 2024. The Application of Large Language Models in Medicine: A Scoping Review. iScience (March 2024). 10.1016\/j.isci.2024.109713","DOI":"10.1016\/j.isci.2024.109713"},{"key":"e_1_3_3_1_30_2","doi-asserted-by":"publisher","unstructured":"Tanisha Mishra Edward Sutanto Rini Rossanti et\u00a0al. 2024. Use of large language models as artificial intelligence tools in academic research and publishing among global clinical researchers. Dental Science Reports (December 2024). 10.1038\/s41598-024-81370-6Study reports 54.2% outlook; check journal details.","DOI":"10.1038\/s41598-024-81370-6"},{"key":"e_1_3_3_1_31_2","doi-asserted-by":"publisher","unstructured":"Rock\u00a0Yuren Pang Hope Schroeder Kendal Smith et\u00a0al. 2025. Understanding the LLM-ification of CHI: Unpacking the Impact of LLMs at CHI through a Systematic Literature Review. arxiv:https:\/\/arXiv.org\/abs\/2501.12557\u00a0[cs.HC] 10.48550\/arxiv.2501.12557arXiv preprint.","DOI":"10.48550\/arxiv.2501.12557"},{"key":"e_1_3_3_1_32_2","unstructured":"Joon\u00a0Sung Park Joseph\u00a0C. O\u2019Brien Carrie\u00a0J. Cai Meredith\u00a0Ringel Morris Percy Liang and Michael\u00a0S. Bernstein. 2023. Generative Agents: Interactive Simulacra of Human Behavior. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2304.03442 (2023). https:\/\/arxiv.org\/abs\/2304.03442"},{"key":"e_1_3_3_1_33_2","unstructured":"Ronak Pradeep Sahel Sharifymoghaddam and Jimmy Lin. 2023. RankZephyr: Effective and Robust Zero-Shot Listwise Reranking is a Breeze! arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2312.02724 (2023)."},{"key":"e_1_3_3_1_34_2","unstructured":"Quthor. 2024. Understanding the necessity of AI in academic writing. https:\/\/quickcreator.io\/blog\/necessity-of-ai-in-academic-writing\/"},{"key":"e_1_3_3_1_35_2","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v39i27.35084"},{"key":"e_1_3_3_1_36_2","doi-asserted-by":"publisher","DOI":"10.1109\/miucc62295.2024.10783597"},{"key":"e_1_3_3_1_37_2","unstructured":"Fobo Shi Peijun Qing Dong Yang Nan Wang Youbo Lei Haonan Lu Xiaodong Lin and Duantengchuan Li. 2023. Prompt space optimizing few-shot reasoning success with large language models. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2306.03799 (2023)."},{"key":"e_1_3_3_1_38_2","unstructured":"Noah Shinn Zachary Labash Ashwin Gopinath Aman Madaan Percy Liu Yisen Chen Luke Zettlemoyer et\u00a0al. 2023. Reflexion: Language Agents with Verbal Reinforcement Learning. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2303.11366 (2023). https:\/\/arxiv.org\/abs\/2303.11366"},{"key":"e_1_3_3_1_39_2","unstructured":"Chenglei Si Diyi Yang and Tatsunori Hashimoto. 2024. Can llms generate novel research ideas. A large-scale human study with 100 (2024)."},{"key":"e_1_3_3_1_40_2","doi-asserted-by":"crossref","unstructured":"Weiwei Sun Lingyong Yan Xinyu Ma Shuaiqiang Wang Pengjie Ren Zhumin Chen Dawei Yin and Zhaochun Ren. 2023. Is ChatGPT good at search? investigating large language models as re-ranking agents. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2304.09542 (2023).","DOI":"10.18653\/v1\/2023.emnlp-main.923"},{"key":"e_1_3_3_1_41_2","unstructured":"Xuemei Tang Xufeng Duan and Zhenguang\u00a0G. Cai. 2024. Are LLMs Good Literature Review Writers? Evaluating the Literature Review Writing Ability of Large Language Models. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2412.13612 (2024). arxiv:https:\/\/arXiv.org\/abs\/2412.13612\u00a0[cs.CL] https:\/\/arxiv.org\/abs\/2412.13612"},{"key":"e_1_3_3_1_42_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2024.acl-long.18"},{"key":"e_1_3_3_1_43_2","doi-asserted-by":"crossref","unstructured":"Jane Webster and Richard\u00a0T Watson. 2002. Analyzing the past to prepare for the future: Writing a literature review. MIS quarterly (2002) xiii\u2013xxiii.","DOI":"10.2307\/4132319"},{"key":"e_1_3_3_1_44_2","unstructured":"Jason Wei Maarten Bosma Vincent\u00a0Y Zhao Kelvin Guu Adams\u00a0Wei Yu Brian Lester Nan Du Andrew\u00a0M Dai and Quoc\u00a0V Le. 2021. Finetuned language models are zero-shot learners. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2109.01652 (2021)."},{"key":"e_1_3_3_1_45_2","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.550"},{"key":"e_1_3_3_1_46_2","doi-asserted-by":"crossref","unstructured":"Shunyu Yao Dian Yu Jeffrey Zhao Izhak Shafran Tom Griffiths Yuan Cao and Karthik Narasimhan. 2023. Tree of thoughts: Deliberate problem solving with large language models. Advances in neural information processing systems 36 (2023) 11809\u201311822.","DOI":"10.52202\/075280-0517"},{"key":"e_1_3_3_1_47_2","doi-asserted-by":"publisher","unstructured":"Hye\u00a0Sun Yun Iain\u00a0J. Marshall Thomas\u00a0A. Trikalinos et\u00a0al. 2023. Appraising the Potential Uses and Harms of LLMs for Medical Systematic Reviews. arxiv:https:\/\/arXiv.org\/abs\/2305.11828\u00a0[cs.CY] 10.48550\/arXiv.2305.11828arXiv preprint.","DOI":"10.48550\/arXiv.2305.11828"}],"event":{"name":"IUI '26: 31st International Conference on Intelligent User Interfaces","location":"Paphos Cyprus","acronym":"IUI '26 Companion","sponsor":["SIGCHI ACM Special Interest Group on Computer-Human Interaction","SIGAI ACM Special Interest Group on Artificial Intelligence"]},"container-title":["Companion Proceedings of the 31st International Conference on Intelligent User Interfaces"],"original-title":[],"deposited":{"date-parts":[[2026,3,14]],"date-time":"2026-03-14T11:04:46Z","timestamp":1773486286000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3742414.3794723"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,3,22]]},"references-count":46,"alternative-id":["10.1145\/3742414.3794723","10.1145\/3742414"],"URL":"https:\/\/doi.org\/10.1145\/3742414.3794723","relation":{},"subject":[],"published":{"date-parts":[[2026,3,22]]},"assertion":[{"value":"2026-03-22","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}