{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T23:16:46Z","timestamp":1776122206678,"version":"3.50.1"},"reference-count":132,"publisher":"Association for Computing Machinery (ACM)","issue":"2","license":[{"start":{"date-parts":[[2025,5,2]],"date-time":"2025-05-02T00:00:00Z","timestamp":1746144000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["Proc. ACM Hum.-Comput. Interact."],"published-print":{"date-parts":[[2025,5,2]]},"abstract":"<jats:p>Large language models are increasingly applied in real-world scenarios, including research and education. These models, however, come with well-known ethical issues, which may manifest in unexpected ways in human-computer interaction research due to the extensive engagement with human subjects. This paper reports on research practices related to LLM use, drawing on 16 semi-structured interviews and a survey with 50 HCI researchers. We discuss the ways in which LLMs are already being utilized throughout the entire HCI research pipeline, from ideation to system development and paper writing. While researchers described nuanced understandings of ethical issues, they were rarely or only partially able to identify and address those ethical concerns in their own projects. This lack of action and reliance on workarounds was explained through the perceived lack of control and distributed responsibility in the LLM supply chain, the conditional nature of engaging with ethics, and competing priorities. Finally, we reflect on the implications of our findings and present opportunities to shape emerging norms of engaging with large language models in HCI research.<\/jats:p>","DOI":"10.1145\/3711000","type":"journal-article","created":{"date-parts":[[2025,5,20]],"date-time":"2025-05-20T11:36:19Z","timestamp":1747740979000},"page":"1-26","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":23,"title":["'I'm Categorizing LLM as a Productivity Tool': Examining Ethics of LLM Use in HCI Research Practices"],"prefix":"10.1145","volume":"9","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-0152-4311","authenticated-orcid":false,"given":"Shivani","family":"Kapania","sequence":"first","affiliation":[{"name":"Carnegie Mellon University, Pittsburgh, PA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6195-2510","authenticated-orcid":false,"given":"Ruiyi","family":"Wang","sequence":"additional","affiliation":[{"name":"Carnegie Mellon University, Pittsburgh, PA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7902-7625","authenticated-orcid":false,"given":"Toby Jia-Jun","family":"Li","sequence":"additional","affiliation":[{"name":"University of Notre Dame, Notre Dame, IN, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0877-5727","authenticated-orcid":false,"given":"Tianshi","family":"Li","sequence":"additional","affiliation":[{"name":"Northeastern University, Boston, MA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5364-3718","authenticated-orcid":false,"given":"Hong","family":"Shen","sequence":"additional","affiliation":[{"name":"Carnegie Mellon University, Pittsburgh, PA, USA"}]}],"member":"320","published-online":{"date-parts":[[2025,5,2]]},"reference":[{"key":"e_1_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1145\/3290605.3300233"},{"key":"e_1_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.1145\/3137107"},{"key":"e_1_2_1_3_1","unstructured":"Christopher Bail. 2023. Can generative artificial intelligence improve social science."},{"key":"e_1_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445922"},{"key":"e_1_2_1_5_1","unstructured":"ACM Publications Board. 2021. ACM Publications Policy on Research Involving Human Participants and Subjects. https:\/\/www.acm.org\/publications\/policies\/research-involving-human-participants-and-subjects. (Accessed on 01\/16\/2024)."},{"key":"e_1_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1038\/d41586-"},{"key":"e_1_2_1_7_1","volume-title":"Using thematic analysis in psychology. Qualitative research in psychology 3, 2","author":"Braun Virginia","year":"2006","unstructured":"Virginia Braun and Victoria Clarke. 2006. Using thematic analysis in psychology. Qualitative research in psychology 3, 2 (2006), 77--101."},{"key":"e_1_2_1_8_1","volume-title":"Reflecting on reflexive thematic analysis. Qualitative research in sport, exercise and health 11, 4","author":"Braun Virginia","year":"2019","unstructured":"Virginia Braun and Victoria Clarke. 2019. Reflecting on reflexive thematic analysis. Qualitative research in sport, exercise and health 11, 4 (2019), 589--597."},{"key":"e_1_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1145\/2858036.2858313"},{"key":"e_1_2_1_10_1","unstructured":"Ryan Browne. 2023. OpenAI CEO says a bug allowed some ChatGPT to see others' chat titles. https:\/\/www.cnbc.com\/ 2023\/03\/23\/openai-ceo-says-a-bug-allowed-some-chatgpt-to-see-others-chat-titles.html. (Accessed on 01\/15\/2024)."},{"key":"e_1_2_1_11_1","doi-asserted-by":"crossref","unstructured":"Amy Bruckman. 2014. Research ethics and HCI. 449--468 pages.","DOI":"10.1007\/978-1-4939-0378-8_18"},{"key":"e_1_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1145\/3022198.3022199"},{"key":"e_1_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1525\/jer.2011.6.2.71"},{"key":"e_1_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1145\/1713066.1713069"},{"key":"e_1_2_1_15_1","volume-title":"Buchanan and Michael Zimmer","author":"Elizabeth","year":"2023","unstructured":"Elizabeth A. Buchanan and Michael Zimmer. 2023. Internet Research Ethics. https:\/\/plato.stanford.edu\/archives\/ win2023\/entries\/ethics-internet-research\/."},{"key":"e_1_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1145\/3544549.3582749"},{"key":"e_1_2_1_17_1","unstructured":"Nicholas Carlini Florian Tramer Eric Wallace Matthew Jagielski Ariel Herbert-Voss Katherine Lee Adam Roberts Tom Brown Dawn Song Ulfar Erlingsson Alina Oprea and Colin Raffel. 2021. Extracting Training Data from Large Language Models. arXiv:2012.07805 [cs.CR]"},{"key":"e_1_2_1_18_1","doi-asserted-by":"publisher","unstructured":"Stephen Cave Claire Craig Kanta Dihal Sarah Dillon Jessica Montgomery Beth Singler and Lindsay Taylor. 2018. Portrayals and perceptions of AI and why they matter. doi:10.17863\/CAM.34502","DOI":"10.17863\/CAM.34502"},{"key":"e_1_2_1_19_1","volume-title":"ACL 2023 Policy on AI Writing Assistance - ACL 2023. https:\/\/2023","author":"ACL","year":"2023","unstructured":"ACL 2023 Program Chairs. 2023. ACL 2023 Policy on AI Writing Assistance - ACL 2023. https:\/\/2023.aclweb.org\/blog\/ ACL-2023-policy\/. (Accessed on 01\/11\/2024)."},{"key":"e_1_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1145\/1979742.1979574"},{"key":"e_1_2_1_21_1","unstructured":"Canyu Chen and Kai Shu. 2023. Can LLM-Generated Misinformation Be Detected? arXiv:2309.13788 [cs.CL]"},{"key":"e_1_2_1_22_1","unstructured":"Canyu Chen and Kai Shu. 2023. Combating Misinformation in the Age of LLMs: Opportunities and Challenges. arXiv:2311.05656 [cs.CY]"},{"key":"e_1_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.emnlp-main.669"},{"key":"e_1_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.emnlp-main.669"},{"key":"e_1_2_1_25_1","unstructured":"Robert Chew John Bollenbacher Michael Wenger Jessica Speer and Annice Kim. 2023. LLM-Assisted Content Analysis: Using Large Language Models to Support Deductive Coding. arXiv:2306.14924 [cs.CL]"},{"key":"e_1_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10676-018--9490--4"},{"key":"e_1_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1145\/3593013.3594073"},{"key":"e_1_2_1_28_1","first-page":"688","article-title":"Using large language models in psychology","volume":"2","author":"Demszky Dorottya","year":"2023","unstructured":"Dorottya Demszky, Diyi Yang, David S Yeager, Christopher J Bryan, Margarett Clapper, Susannah Chandhok, Johannes C Eichstaedt, Cameron Hecht, Jeremy Jamieson, Meghann Johnson, et al. 2023. Using large language models in psychology. Nature Reviews Psychology 2, 11 (2023), 688--701.","journal-title":"Nature Reviews Psychology"},{"key":"e_1_2_1_29_1","volume-title":"Vidhisha Balachandran, Sachin Kumar, Yulia Tsvetkov, M. R. Leiser, and Saif Mohammad.","author":"Derczynski Leon","year":"2023","unstructured":"Leon Derczynski, Hannah Rose Kirk, Vidhisha Balachandran, Sachin Kumar, Yulia Tsvetkov, M. R. Leiser, and Saif Mohammad. 2023. Assessing Language Model Deployment with Risk Cards. arXiv:2303.18190 [cs.CL]"},{"key":"e_1_2_1_30_1","doi-asserted-by":"crossref","unstructured":"Ameet Deshpande Tanmay Rajpurohit Karthik Narasimhan and Ashwin Kalyan. 2023. Anthropomorphization of AI: Opportunities and Risks. arXiv:2305.14784 [cs.AI]","DOI":"10.18653\/v1\/2023.nllp-1.1"},{"key":"e_1_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1145\/3544548.3581347"},{"key":"e_1_2_1_32_1","unstructured":"Elicit. 2024. Elicit Help Center. https:\/\/support.elicit.com\/en\/categories\/146369-about-elicit. (Accessed on 07\/16\/2024)."},{"key":"e_1_2_1_33_1","doi-asserted-by":"crossref","unstructured":"Xiao Fang Shangkun Che Minjia Mao Hongzhe Zhang Ming Zhao and Xiaohang Zhao. 2023. Bias of AI-Generated Content: An Examination of News Produced by Large Language Models. arXiv:2309.09825 [cs.AI]","DOI":"10.21203\/rs.3.rs-3499674\/v1"},{"key":"e_1_2_1_34_1","doi-asserted-by":"crossref","unstructured":"Yunhe Feng Sreecharan Vanam Manasa Cherukupally Weijian Zheng Meikang Qiu and Haihua Chen. 2023. Investigating Code Generation Performance of ChatGPT with Crowdsourcing Social Data. 876--885 pages. doi:10. 1109\/COMPSAC57700.2023.00117","DOI":"10.1109\/COMPSAC57700.2023.00117"},{"key":"e_1_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1145\/3462204.3483283"},{"key":"e_1_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.1145\/3491101.3516400"},{"key":"e_1_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1145\/3170427.3186321"},{"key":"e_1_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.1177\/2056305118763366"},{"key":"e_1_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1145\/3633070"},{"key":"e_1_2_1_40_1","unstructured":"Association for Computing Machinery. 2023. ACM Policy on Authorship. https:\/\/www.acm.org\/publications\/policies\/ new-acm-policy-on-authorship. (Accessed on 01\/11\/2024)."},{"key":"e_1_2_1_41_1","unstructured":"Office for Human Research Protections. 1998. Informed Consent Checklist (1998) | HHS.gov. https:\/\/www.hhs.gov\/ohrp\/regulations-and-policy\/guidance\/checklists\/index.html. (Accessed on 01\/15\/2024)."},{"key":"e_1_2_1_42_1","unstructured":"ACM Code 2018 Task Force. 2018. Code of Ethics. https:\/\/www.acm.org\/code-of-ethics. (Accessed on 01\/16\/2024)."},{"key":"e_1_2_1_43_1","doi-asserted-by":"publisher","DOI":"10.1145\/3027063.3051135"},{"key":"e_1_2_1_44_1","unstructured":"Iason Gabriel Arianna Manzini Geoff Keeling Lisa Anne Hendricks Verena Rieser Hasan Iqbal Nenad Toma\u0161ev Ira Ktena Zachary Kenton Mikel Rodriguez Seliem El-Sayed Sasha Brown Canfer Akbulut Andrew Trask Edward Hughes A. Stevie Bergman Renee Shelby Nahema Marchal Conor Griffin Juan Mateos-Garcia Laura Weidinger Winnie Street Benjamin Lange Alex Ingerman Alison Lentz Reed Enger Andrew Barakat Victoria Krakovna John Oliver Siy Zeb Kurth-Nelson Amanda McCroskery Vijay Bolina Harry Law Murray Shanahan Lize Alberts Borja Balle Sarah de Haas Yetunde Ibitoye Allan Dafoe Beth Goldberg S\u00e9bastien Krier Alexander Reese Sims Witherspoon Will Hawkins Maribeth Rauh Don Wallace Matija Franklin Josh A. Goldstein Joel Lehman Michael Klenk Shannon Vallor Courtney Biles Meredith Ringel Morris Helen King Blaise Ag\u00fcera y Arcas William Isaac and James Manyika. 2024. The Ethics of Advanced AI Assistants. arXiv:2404.16244 [cs.CY] https:\/\/arxiv.org\/abs\/2404.16244"},{"key":"e_1_2_1_45_1","doi-asserted-by":"crossref","unstructured":"Chen Gao Xiaochong Lan Zhihong Lu Jinzhu Mao Jinghua Piao Huandong Wang Depeng Jin and Yong Li. 2023. S3: Social-network Simulation System with Large Language Model-Empowered Agents. arXiv:2307.14984 [cs.SI]","DOI":"10.2139\/ssrn.4607026"},{"key":"e_1_2_1_46_1","volume-title":"Toby Jia-Jun Li, and Simon Tangi Perrault","author":"Gao Jie","year":"2023","unstructured":"Jie Gao, Yuchen Guo, Gionnieve Lim, Tianqin Zhang, Zheng Zhang, Toby Jia-Jun Li, and Simon Tangi Perrault. 2023. CollabCoder: A GPT-Powered Workflow for Collaborative Qualitative Analysis. arXiv:2304.07366 [cs.HC]"},{"key":"e_1_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.1145\/3544548.3581352"},{"key":"e_1_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.findings-emnlp.301"},{"key":"e_1_2_1_49_1","volume-title":"Ulrich","author":"Girotra Karan","year":"2023","unstructured":"Karan Girotra, Lennart Meincke, Christian Terwiesch, and Karl T. Ulrich. 2023. Ideas are dimes a dozen: Large language models for Idea generation in innovation. https:\/\/papers.ssrn.com\/sol3\/papers.cfm?abstract_id=4526071#"},{"key":"e_1_2_1_50_1","doi-asserted-by":"publisher","DOI":"10.1145\/3544548.3580688"},{"key":"e_1_2_1_51_1","doi-asserted-by":"publisher","unstructured":"Jennifer A. Hamilton. 2011. On the Ethics of Unusable Data. 73--88 pages. https:\/\/doi.org\/10.7591\/9780801463594-005","DOI":"10.7591\/9780801463594-005"},{"key":"e_1_2_1_52_1","volume-title":"Nathan Lambert, Yejin Choi, and Nouha Dziri.","author":"Han Seungju","year":"2024","unstructured":"Seungju Han, Kavel Rao, Allyson Ettinger, Liwei Jiang, Bill Yuchen Lin, Nathan Lambert, Yejin Choi, and Nouha Dziri. 2024. WildGuard: Open One-Stop Moderation Tools for Safety Risks, Jailbreaks, and Refusals of LLMs. arXiv:2406.18495 [cs.CL] https:\/\/arxiv.org\/abs\/2406.18495"},{"key":"e_1_2_1_53_1","doi-asserted-by":"publisher","unstructured":"Transparent Statistics in Human--Computer Interaction Working Group. 2019. Transparent Statistics Guidelines. doi:10.5281\/zenodo.1186169 (Available at https:\/\/transparentstats.github.io\/guidelines).","DOI":"10.5281\/zenodo.1186169"},{"key":"e_1_2_1_54_1","volume-title":"New in Dovetail: Search API, new editor","author":"Dovetail Inc. 2024.","year":"2024","unstructured":"Dovetail Inc. 2024. New in Dovetail: Search API, new editor, improved features. https:\/\/dovetail.com\/blog\/evolutiondovetail-magic\/. (Accessed on 07\/16\/2024)."},{"key":"e_1_2_1_55_1","volume-title":"Using mixed-methods sequential explanatory design: From theory to practice. Field methods 18, 1","author":"Ivankova Nataliya V","year":"2006","unstructured":"Nataliya V Ivankova, John W Creswell, and Sheldon L Stick. 2006. Using mixed-methods sequential explanatory design: From theory to practice. Field methods 18, 1 (2006), 3--20."},{"key":"e_1_2_1_56_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.nlp.2023.100020"},{"key":"e_1_2_1_57_1","unstructured":"Jenni. 2024. Jenni AI. https:\/\/jenni.ai\/. (Accessed on 07\/16\/2024)."},{"key":"e_1_2_1_58_1","unstructured":"Jigsaw and Google's Counter Abuse Technology team. 2017. Perspective API. https:\/\/perspectiveapi.com\/. (Accessed on 01\/16\/2024)."},{"key":"e_1_2_1_59_1","doi-asserted-by":"publisher","DOI":"10.1145\/3491102.3517533"},{"key":"e_1_2_1_60_1","doi-asserted-by":"publisher","DOI":"10.1145\/3579592"},{"key":"e_1_2_1_61_1","doi-asserted-by":"crossref","unstructured":"Enkelejda Kasneci Kathrin Se\u00dfler Stefan K\u00fcchemann Maria Bannert Daryna Dementieva Frank Fischer Urs Gasser Georg Groh Stephan G\u00fcnnemann Eyke H\u00fcllermeier et al. 2023. ChatGPT for good? On opportunities and challenges of large language models for education. Learning and individual differences 103 (2023) 102274.","DOI":"10.1016\/j.lindif.2023.102274"},{"key":"e_1_2_1_62_1","doi-asserted-by":"publisher","DOI":"10.1177\/14687941211014366"},{"key":"e_1_2_1_63_1","volume-title":"Faisal Ladhak, Frieda Rong, Rose E. Wang, Minae Kwon, Joon Sung Park, Hancheng Cao, Tony Lee, Rishi Bommasani, Michael Bernstein, and Percy Liang.","author":"Lee Mina","year":"2024","unstructured":"Mina Lee, Megha Srivastava, Amelia Hardy, John Thickstun, Esin Durmus, Ashwin Paranjape, Ines Gerard-Ursin, Xiang Lisa Li, Faisal Ladhak, Frieda Rong, Rose E. Wang, Minae Kwon, Joon Sung Park, Hancheng Cao, Tony Lee, Rishi Bommasani, Michael Bernstein, and Percy Liang. 2024. Evaluating Human-Language Model Interaction. arXiv:2212.09746 [cs.CL] https:\/\/arxiv.org\/abs\/2212.09746"},{"key":"e_1_2_1_64_1","volume-title":"Ashley Ricker Gyllen, and Mutlu Cukurova","author":"Leiker Daniel","year":"2023","unstructured":"Daniel Leiker, Sara Finnigan, Ashley Ricker Gyllen, and Mutlu Cukurova. 2023. Prototyping the use of Large Language Models (LLMs) for adult learning content creation at scale. arXiv:2306.01815 [cs.CY]"},{"key":"e_1_2_1_65_1","unstructured":"Yuan Li Yixuan Zhang and Lichao Sun. 2023. MetaAgents: Simulating Interactions of Human Behaviors for LLM-based Task-oriented Coordination via Collaborative Generative Agents. arXiv:2310.06500 [cs.AI]"},{"key":"e_1_2_1_66_1","doi-asserted-by":"publisher","DOI":"10.1145\/3443686"},{"key":"e_1_2_1_67_1","doi-asserted-by":"crossref","unstructured":"Q. Vera Liao and JenniferWortman Vaughan. 2023. AI Transparency in the Age of LLMs: A Human-Centered Research Roadmap. arXiv:2306.01941 [cs.HC]","DOI":"10.1162\/99608f92.8036d03b"},{"key":"e_1_2_1_68_1","doi-asserted-by":"crossref","unstructured":"Yiren Liu Si Chen Haocong Cheng Mengxia Yu Xiao Ran Andrew Mo Yiliu Tang and Yun Huang. 2023. How AI Processing Delays Foster Creativity: Exploring Research Question Co-Creation with an LLM-based Agent. arXiv:2310.06155 [cs.HC]","DOI":"10.1145\/3613904.3642698"},{"key":"e_1_2_1_69_1","doi-asserted-by":"publisher","DOI":"10.1145\/3491101.3519809"},{"key":"e_1_2_1_70_1","doi-asserted-by":"crossref","unstructured":"Nils Lukas Ahmed Salem Robert Sim Shruti Tople Lukas Wutschitz and Santiago Zanella-B\u00e9guelin. 2023. Analyzing Leakage of Personally Identifiable Information in Language Models. arXiv:2302.00539 [cs.LG]","DOI":"10.1109\/SP46215.2023.10179300"},{"key":"e_1_2_1_71_1","unstructured":"Kevin Ma Daniele Grandi Christopher McComb and Kosa Goucher-Lambert. 2023. Conceptual Design Generation Using Large Language Models. arXiv:2306.01779 [cs.CL]"},{"key":"e_1_2_1_72_1","doi-asserted-by":"publisher","DOI":"10.1145\/223904.223922"},{"key":"e_1_2_1_73_1","doi-asserted-by":"publisher","DOI":"10.1145\/3359174"},{"key":"e_1_2_1_74_1","doi-asserted-by":"publisher","DOI":"10.1145\/3544548.3580940"},{"key":"e_1_2_1_75_1","unstructured":"Niloofar Mireshghallah Hyunwoo Kim Xuhui Zhou Yulia Tsvetkov Maarten Sap Reza Shokri and Yejin Choi. 2023. Can LLMs Keep a Secret? Testing Privacy Implications of Language Models via Contextual Integrity Theory. arXiv:2310.17884 [cs.AI]"},{"key":"e_1_2_1_76_1","volume-title":"Griffin","author":"Mozes Maximilian","year":"2023","unstructured":"Maximilian Mozes, Xuanli He, Bennett Kleinberg, and Lewis D. Griffin. 2023. Use of LLMs for Illicit Purposes: Threats, Prevention Measures, and Vulnerabilities. arXiv:2308.12833 [cs.CL]"},{"key":"e_1_2_1_77_1","doi-asserted-by":"publisher","DOI":"10.1145\/3290607.3311742"},{"key":"e_1_2_1_78_1","doi-asserted-by":"publisher","DOI":"10.1145\/2702123.2702481"},{"key":"e_1_2_1_79_1","doi-asserted-by":"publisher","DOI":"10.1007\/s43681-023-00289--2"},{"key":"e_1_2_1_80_1","doi-asserted-by":"publisher","DOI":"10.1145\/3581641.3584067"},{"key":"e_1_2_1_81_1","unstructured":"US Department of Health and Human Services. 2018. 2018 Requirements (2018 Common Rule) | HHS.gov. https:\/\/www. hhs.gov\/ohrp\/regulations-and-policy\/regulations\/45-cfr-46\/revised-common-rule-regulatory-text\/index.html. (Accessed on 01\/15\/2024)."},{"key":"e_1_2_1_82_1","unstructured":"US Department of Health and Human Services. 2018. Policy for the Protection of Human Subjects. https:\/\/www.hhs.gov\/ohrp\/regulations-and-policy\/regulations\/45-cfr-46\/index.html. (Accessed on 01\/15\/2024)."},{"key":"e_1_2_1_83_1","unstructured":"National Institute of Standards and Technology. 2023. Artificial Intelligence Risk Management Framework (AI RMF 1.0)."},{"key":"e_1_2_1_84_1","unstructured":"OpenAI. 2023. GPT-4 Technical Report. arXiv:2303.08774 [cs.CL]"},{"key":"e_1_2_1_85_1","unstructured":"Jonas Oppenlaender and Joonas H\u00e4m\u00e4l\u00e4inen. 2023. Mapping the Challenges of HCI: An Application and Evaluation of ChatGPT and GPT-4 for Mining Insights at Scale. arXiv:2306.05036 [cs.HC]"},{"key":"e_1_2_1_86_1","unstructured":"Google PAIR. 2021. People AI Guidebook. https:\/\/pair.withgoogle.com\/guidebook. (Accessed on 01\/15\/2024)."},{"key":"e_1_2_1_87_1","unstructured":"Google PAIR. 2022. People AI Research Guidebook - Case Studies. https:\/\/pair.withgoogle.com\/guidebook\/casestudies. (Accessed on 01\/15\/2024)."},{"key":"e_1_2_1_88_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10488-013-0528-y"},{"key":"e_1_2_1_89_1","unstructured":"Yikang Pan Liangming Pan Wenhu Chen Preslav Nakov Min-Yen Kan and William Yang Wang. 2023. On the Risk of Misinformation Pollution with Large Language Models. arXiv:2305.13661 [cs.CL]"},{"key":"e_1_2_1_90_1","doi-asserted-by":"crossref","unstructured":"Stefano De Paoli. 2023. Can Large Language Models emulate an inductive Thematic Analysis of semi-structured interviews? An exploration and provocation on the limits of the approach and the model. arXiv:2305.13014 [cs.CL]","DOI":"10.1177\/08944393231220483"},{"key":"e_1_2_1_91_1","doi-asserted-by":"publisher","DOI":"10.1145\/3586183.3606763"},{"key":"e_1_2_1_92_1","volume-title":"Percy Liang, and Michael S.","author":"Park Joon Sung","year":"2022","unstructured":"Joon Sung Park, Lindsay Popowski, Carrie J. Cai, Meredith Ringel Morris, Percy Liang, and Michael S. Bernstein. 2022. Social Simulacra: Creating Populated Prototypes for Social Computing Systems. arXiv:2208.04024 [cs.HC]"},{"key":"e_1_2_1_93_1","doi-asserted-by":"publisher","DOI":"10.1145\/3544549.3585628"},{"key":"e_1_2_1_94_1","doi-asserted-by":"publisher","DOI":"10.1007\/s43681-020-00030-3"},{"key":"e_1_2_1_95_1","doi-asserted-by":"publisher","unstructured":"Lumpapun Punchoojit and Nuttanont Hongwarittorrn. 2015. Research ethics in human-computer interaction: A review of ethical concerns in the past five years. 180--185 pages. doi:10.1109\/NICS.2015.7302187","DOI":"10.1109\/NICS.2015.7302187"},{"key":"e_1_2_1_96_1","doi-asserted-by":"publisher","DOI":"10.1186\/s13643-023-02243-z"},{"key":"e_1_2_1_97_1","volume-title":"Anh Nguyen Duc, and Pekka Abrahamsson","author":"Rasheed Zeeshan","year":"2024","unstructured":"Zeeshan Rasheed, Muhammad Waseem, Aakash Ahmad, Kai-Kristian Kemell, Wang Xiaofeng, Anh Nguyen Duc, and Pekka Abrahamsson. 2024. Can Large Language Models Serve as Data Analysts? A Multi-Agent Assisted Approach for Qualitative Data Analysis. arXiv:2402.01386 [cs.SE] https:\/\/arxiv.org\/abs\/2402.01386"},{"key":"e_1_2_1_98_1","doi-asserted-by":"publisher","DOI":"10.1145\/3544548.3580848"},{"key":"e_1_2_1_99_1","doi-asserted-by":"publisher","DOI":"10.1145\/3341164"},{"key":"e_1_2_1_100_1","volume-title":"PEER: Empowering Writing with Large Language Models","author":"Se\u00dfler Kathrin","year":"2023","unstructured":"Kathrin Se\u00dfler, Tao Xiang, Lukas Bogenrieder, and Enkelejda Kasneci. 2023. PEER: Empowering Writing with Large Language Models. In Responsive and Sustainable Educational Futures, Olga Viberg, Ioana Jivet, Pedro J. Mu\u00f1oz-Merino, Maria Perifanou, and Tina Papathoma (Eds.). Springer Nature Switzerland, Cham, 755--761."},{"key":"e_1_2_1_101_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445971"},{"key":"e_1_2_1_102_1","doi-asserted-by":"publisher","DOI":"10.1145\/3584931.3606955"},{"key":"e_1_2_1_103_1","doi-asserted-by":"crossref","unstructured":"Irina Shklovski and Janet Vertesi. 2012. ''Un-Googling'': Research Technologies Communities at Risk and the Ethics of User Studies in HCI. 4 pages.","DOI":"10.14236\/ewic\/HCI2012.75"},{"key":"e_1_2_1_104_1","doi-asserted-by":"publisher","DOI":"10.1145\/3419764"},{"key":"e_1_2_1_105_1","doi-asserted-by":"publisher","DOI":"10.1145\/3545945.3569855"},{"key":"e_1_2_1_106_1","volume-title":"Carolyn Ashurst, and Adrian Weller.","author":"Smith Victoria","year":"2023","unstructured":"Victoria Smith, Ali Shahin Shamsabadi, Carolyn Ashurst, and Adrian Weller. 2023. Identifying and Mitigating Privacy Risks Stemming from Language Models: A Survey. arXiv:2310.01424 [cs.CL]"},{"key":"e_1_2_1_107_1","doi-asserted-by":"publisher","DOI":"10.1145\/3633200"},{"key":"e_1_2_1_108_1","unstructured":"David Soergel Adam Saunders and Andrew McCallum. 2013. Open Scholarship and Peer Review: a Time for Experimentation. https:\/\/api.semanticscholar.org\/CorpusID:14548845"},{"key":"e_1_2_1_109_1","volume-title":"Toby Jia-Jun Li, and Haijun Xia","author":"Suh Sangho","year":"2023","unstructured":"Sangho Suh, Meng Chen, Bryan Min, Toby Jia-Jun Li, and Haijun Xia. 2023. Structured Generation and Exploration of Design Space with Large Language Models for Human-AI Co-Creation. arXiv:2310.12953 [cs.HC]"},{"key":"e_1_2_1_110_1","doi-asserted-by":"publisher","DOI":"10.1145\/3637371"},{"key":"e_1_2_1_111_1","volume-title":"Timothy R McIntosh, and Surangika Ranathunga.","author":"Susnjak Teo","year":"2024","unstructured":"Teo Susnjak, Peter Hwang, Napoleon H Reyes, Andre LC Barczak, Timothy R McIntosh, and Surangika Ranathunga. 2024. Automating research synthesis with domain-specific large language model fine-tuning."},{"key":"e_1_2_1_112_1","doi-asserted-by":"publisher","DOI":"10.1177\/16094069241231168"},{"key":"e_1_2_1_113_1","volume-title":"Kabilan Elangovan, Laura Gutierrez, Ting Fang Tan, and Daniel Shu Wei Ting.","author":"Thirunavukarasu Arun James","year":"2023","unstructured":"Arun James Thirunavukarasu, Darren Shu Jeng Ting, Kabilan Elangovan, Laura Gutierrez, Ting Fang Tan, and Daniel Shu Wei Ting. 2023. Large language models in medicine. Nature medicine 29, 8 (2023), 1930--1940."},{"key":"e_1_2_1_114_1","doi-asserted-by":"publisher","DOI":"10.1145\/3563657.3596014"},{"key":"e_1_2_1_115_1","doi-asserted-by":"publisher","DOI":"10.1145\/2818048.2820078"},{"key":"e_1_2_1_116_1","unstructured":"Jane Wakefield. 2016. Microsoft chatbot is taught to swear on Twitter - BBC News. https:\/\/www.bbc.com\/news\/ technology-35890188. (Accessed on 07\/16\/2024)."},{"key":"e_1_2_1_117_1","doi-asserted-by":"publisher","DOI":"10.1145\/3661145"},{"key":"e_1_2_1_118_1","doi-asserted-by":"crossref","unstructured":"Jing Wei Sungdong Kim Hyunhoon Jung and Young-Ho Kim. 2023. Leveraging Large Language Models to Power Chatbots for Collecting User Self-Reported Data. arXiv:2301.05843 [cs.HC]","DOI":"10.1145\/3637364"},{"key":"e_1_2_1_119_1","volume-title":"William Isaac, Sean Legassick, Geoffrey Irving, and Iason Gabriel.","author":"Weidinger Laura","year":"2021","unstructured":"Laura Weidinger, John Mellor, Maribeth Rauh, Conor Griffin, Jonathan Uesato, Po-Sen Huang, Myra Cheng, Mia Glaese, Borja Balle, Atoosa Kasirzadeh, Zac Kenton, Sasha Brown, Will Hawkins, Tom Stepleton, Courtney Biles, Abeba Birhane, Julia Haas, Laura Rimell, Lisa Anne Hendricks, William Isaac, Sean Legassick, Geoffrey Irving, and Iason Gabriel. 2021. Ethical and social risks of harm from Language Models. arXiv:2112.04359 [cs.CL]"},{"key":"e_1_2_1_120_1","doi-asserted-by":"publisher","DOI":"10.1145\/3531146.3533088"},{"key":"e_1_2_1_121_1","doi-asserted-by":"publisher","DOI":"10.1177\/20539517231177620"},{"key":"e_1_2_1_122_1","unstructured":"Tongshuang Wu Ellen Jiang Aaron Donsbach Jeff Gray Alejandra Molina Michael Terry and Carrie J Cai. 2022. PromptChainer: Chaining Large Language Model Prompts through Visual Programming. arXiv:2203.06566 [cs.HC]"},{"key":"e_1_2_1_123_1","doi-asserted-by":"publisher","DOI":"10.1145\/3581754.3584136"},{"key":"e_1_2_1_124_1","doi-asserted-by":"publisher","DOI":"10.1145\/3581754.3584136"},{"key":"e_1_2_1_125_1","unstructured":"Chen Xu Wenjie Wang Yuxin Li Liang Pang Jun Xu and Tat-Seng Chua. 2023. Do LLMs Implicitly Exhibit User Discrimination in Recommendation? An Empirical Study. arXiv:2311.07054 [cs.IR]"},{"key":"e_1_2_1_126_1","unstructured":"Jingfeng Yang Hongye Jin Ruixiang Tang Xiaotian Han Qizhang Feng Haoming Jiang Bing Yin and Xia Hu. 2023. Harnessing the Power of LLMs in Practice: A Survey on ChatGPT and Beyond. arXiv:2304.13712 [cs.CL]"},{"key":"e_1_2_1_127_1","volume-title":"Andre Manoel, Lukas Wutschitz, Sergey Yekhanin, and Huishuai Zhang.","author":"Yu Da","year":"2021","unstructured":"Da Yu, Saurabh Naik, Arturs Backurs, Sivakanth Gopi, Huseyin A. Inan, Gautam Kamath, Janardhan Kulkarni, Yin Tat Lee, Andre Manoel, Lukas Wutschitz, Sergey Yekhanin, and Huishuai Zhang. 2021. Differentially Private Fine-tuning of Language Models. arXiv:2110.06500 https:\/\/arxiv.org\/abs\/2110.06500"},{"key":"e_1_2_1_128_1","doi-asserted-by":"crossref","unstructured":"Xuan Zhang and Wei Gao. 2023. Towards LLM-based Fact Verification on News Claims with a Hierarchical Step-by-Step Prompting Method. arXiv:2310.00305 [cs.CL]","DOI":"10.18653\/v1\/2023.ijcnlp-main.64"},{"key":"e_1_2_1_129_1","volume-title":"Wei Bi, Freda Shi, and Shuming Shi.","author":"Zhang Yue","year":"2023","unstructured":"Yue Zhang, Yafu Li, Leyang Cui, Deng Cai, Lemao Liu, Tingchen Fu, Xinting Huang, Enbo Zhao, Yu Zhang, Yulong Chen, Longyue Wang, Anh Tuan Luu, Wei Bi, Freda Shi, and Shuming Shi. 2023. Siren's Song in the AI Ocean: A Survey on Hallucination in Large Language Models. arXiv:2309.01219 [cs.CL]"},{"key":"e_1_2_1_130_1","doi-asserted-by":"publisher","DOI":"10.1145\/3586183.3606800"},{"key":"e_1_2_1_131_1","doi-asserted-by":"publisher","DOI":"10.1145\/3613904.3642385"},{"key":"e_1_2_1_132_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10676-010-9227-5"}],"container-title":["Proceedings of the ACM on Human-Computer Interaction"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3711000","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3711000","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,21]],"date-time":"2025-08-21T09:34:03Z","timestamp":1755768843000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3711000"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,5,2]]},"references-count":132,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2025,5,2]]}},"alternative-id":["10.1145\/3711000"],"URL":"https:\/\/doi.org\/10.1145\/3711000","relation":{},"ISSN":["2573-0142"],"issn-type":[{"value":"2573-0142","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,5,2]]},"assertion":[{"value":"2025-05-02","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}