{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,11]],"date-time":"2026-05-11T14:34:49Z","timestamp":1778510089200,"version":"3.51.4"},"publisher-location":"New York, NY, USA","reference-count":81,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,5,11]],"date-time":"2024-05-11T00:00:00Z","timestamp":1715385600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-sa\/4.0\/"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,5,11]]},"DOI":"10.1145\/3613904.3642385","type":"proceedings-article","created":{"date-parts":[[2024,5,11]],"date-time":"2024-05-11T08:37:41Z","timestamp":1715416661000},"page":"1-26","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":83,"title":["\u201cIt's a Fair Game\u201d, or Is It? Examining How Users Navigate Disclosure Risks and Benefits When Using LLM-Based Conversational Agents"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-6794-0054","authenticated-orcid":false,"given":"Zhiping","family":"Zhang","sequence":"first","affiliation":[{"name":"Northeastern University,, Khoury College of Computer Sciences, United States"}]},{"ORCID":"https:\/\/orcid.org\/0009-0008-5685-4618","authenticated-orcid":false,"given":"Michelle","family":"Jia","sequence":"additional","affiliation":[{"name":"Carnegie Mellon University, United States"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8063-1034","authenticated-orcid":false,"given":"Hao-Ping (Hank)","family":"Lee","sequence":"additional","affiliation":[{"name":"Human-Computer Interaction Institute, Carnegie Mellon University, United States"}]},{"ORCID":"https:\/\/orcid.org\/0009-0004-8329-4610","authenticated-orcid":false,"given":"Bingsheng","family":"Yao","sequence":"additional","affiliation":[{"name":"Computer Science, Rensselaer Polytechnic Institute, United States"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9073-8054","authenticated-orcid":false,"given":"Sauvik","family":"Das","sequence":"additional","affiliation":[{"name":"Human-Computer Interaction Institute, Carnegie Mellon University, United States"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3238-2109","authenticated-orcid":false,"given":"Ada","family":"Lerner","sequence":"additional","affiliation":[{"name":"Khoury College of Computer Sciences, Northeastern University, United States"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9371-9441","authenticated-orcid":false,"given":"Dakuo","family":"Wang","sequence":"additional","affiliation":[{"name":"Northeastern University, United States"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0877-5727","authenticated-orcid":false,"given":"Tianshi","family":"Li","sequence":"additional","affiliation":[{"name":"Khoury College of Computer Sciences, Northeastern University, United States"}]}],"member":"320","published-online":{"date-parts":[[2024,5,11]]},"reference":[{"key":"e_1_3_3_3_1_1","doi-asserted-by":"publisher","DOI":"10.1145\/3366485"},{"key":"e_1_3_3_3_2_1","doi-asserted-by":"publisher","DOI":"10.1080\/1463922X.2022.2061080"},{"key":"e_1_3_3_3_3_1","doi-asserted-by":"publisher","DOI":"10.1109\/EuroSPW51379.2020.00038"},{"key":"e_1_3_3_3_4_1","doi-asserted-by":"publisher","DOI":"10.1609\/hcomp.v7i1.5285"},{"key":"e_1_3_3_3_5_1","volume-title":"Contextual design. interactions 6, 1","author":"Beyer Hugh","year":"1999","unstructured":"Hugh Beyer and Karen Holtzblatt. 1999. Contextual design. interactions 6, 1 (1999), 32\u201342."},{"key":"e_1_3_3_3_6_1","volume-title":"Eighteenth Symposium on Usable Privacy and Security (SOUPS","author":"Bieringer Lukas","year":"2022","unstructured":"Lukas Bieringer, Kathrin Grosse, Michael Backes, Battista Biggio, and Katharina Krombholz. 2022. Industrial practitioners\u2019 mental models of adversarial machine learning. In Eighteenth Symposium on Usable Privacy and Security (SOUPS 2022). 97\u2013116."},{"key":"e_1_3_3_3_7_1","doi-asserted-by":"publisher","DOI":"10.1145\/3531146.3534642"},{"key":"e_1_3_3_3_8_1","volume-title":"Quantifying memorization across neural language models. arXiv preprint arXiv:2202.07646","author":"Carlini Nicholas","year":"2022","unstructured":"Nicholas Carlini, Daphne Ippolito, Matthew Jagielski, Katherine Lee, Florian Tramer, and Chiyuan Zhang. 2022. Quantifying memorization across neural language models. arXiv preprint arXiv:2202.07646 (2022)."},{"key":"e_1_3_3_3_9_1","volume-title":"USENIX Security Symposium, Vol.\u00a06.","author":"Carlini Nicholas","year":"2021","unstructured":"Nicholas Carlini, Florian Tramer, Eric Wallace, Matthew Jagielski, Ariel Herbert-Voss, Katherine Lee, Adam Roberts, Tom\u00a0B Brown, Dawn Song, Ulfar Erlingsson, 2021. Extracting Training Data from Large Language Models.. In USENIX Security Symposium, Vol.\u00a06."},{"key":"e_1_3_3_3_10_1","volume-title":"Vicuna: An open-source chatbot impressing gpt-4 with 90%* chatgpt quality. See https:\/\/vicuna. lmsys. org (accessed","author":"Chiang Wei-Lin","year":"2023","unstructured":"Wei-Lin Chiang, Zhuohan Li, Zi Lin, Ying Sheng, Zhanghao Wu, Hao Zhang, Lianmin Zheng, Siyuan Zhuang, Yonghao Zhuang, Joseph\u00a0E Gonzalez, 2023. Vicuna: An open-source chatbot impressing gpt-4 with 90%* chatgpt quality. See https:\/\/vicuna. lmsys. org (accessed 14 April 2023) (2023)."},{"key":"e_1_3_3_3_11_1","doi-asserted-by":"publisher","DOI":"10.56553\/popets-2023-0027"},{"key":"e_1_3_3_3_12_1","doi-asserted-by":"publisher","DOI":"10.1177\/0149206305279602"},{"key":"e_1_3_3_3_13_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP43922.2022.9746975"},{"key":"e_1_3_3_3_14_1","volume-title":"AMCIS 2007 proceedings","author":"Dwyer Catherine","year":"2007","unstructured":"Catherine Dwyer, Starr Hiltz, and Katia Passerini. 2007. Trust and privacy concern within social networking sites: A comparison of Facebook and MySpace. AMCIS 2007 proceedings (2007), 339."},{"key":"e_1_3_3_3_15_1","volume-title":"Security And Privacy In Computing And Communications\/12th IEEE International Conference On Big Data Science And Engineering (TrustCom\/BigDataSE)","author":"El\u00a0Haddad Ghada","unstructured":"Ghada El\u00a0Haddad, Esma Aimeur, and Hicham Hage. 2018. Understanding trust, privacy and financial fears in online payment. In 2018 17th IEEE International Conference On Trust, Security And Privacy In Computing And Communications\/12th IEEE International Conference On Big Data Science And Engineering (TrustCom\/BigDataSE). IEEE, 28\u201336."},{"key":"e_1_3_3_3_16_1","unstructured":"Sheryl Estrada. 2023. A startup CFO used ChatGPT to build an FP&A tool\u2014here\u2019s how it went. https:\/\/fortune.com\/2023\/03\/01\/startup-cfo-chatgpt-finance-tool\/ Accessed: 09\/11\/2023."},{"key":"e_1_3_3_3_17_1","unstructured":"Pedro Ferreira. 2023. Can ChatGPT Improve Technical Analysis and Trading Techniques?https:\/\/www.financemagnates.com\/trending\/can-chatgpt-improve-technical-analysis-and-trading-techniques\/ Accessed: 09\/11\/2023."},{"key":"e_1_3_3_3_18_1","volume-title":"ChatGPT scored 72study shows. https:\/\/www.healthcareitnews.com\/news\/chatgpt-scored-72-clinical-decision-accuracy-mgb-study-shows Accessed: 09\/11\/2023","author":"Fox Andrea","unstructured":"Andrea Fox. 2023. ChatGPT scored 72study shows. https:\/\/www.healthcareitnews.com\/news\/chatgpt-scored-72-clinical-decision-accuracy-mgb-study-shows Accessed: 09\/11\/2023."},{"key":"e_1_3_3_3_19_1","volume-title":"Thirteenth Symposium on Usable Privacy and Security (SOUPS","author":"Gallagher Kevin","year":"2017","unstructured":"Kevin Gallagher, Sameer Patil, and Nasir Memon. 2017. New Me: Understanding Expert and { Non-Expert} Perceptions and Usage of the Tor Anonymity Network. In Thirteenth Symposium on Usable Privacy and Security (SOUPS 2017). 385\u2013398."},{"key":"e_1_3_3_3_20_1","volume-title":"Koala: A dialogue model for academic research. Blog post, April 1","author":"Geng Xinyang","year":"2023","unstructured":"Xinyang Geng, Arnav Gudibande, Hao Liu, Eric Wallace, Pieter Abbeel, Sergey Levine, and Dawn Song. 2023. Koala: A dialogue model for academic research. Blog post, April 1 (2023)."},{"key":"e_1_3_3_3_21_1","unstructured":"Thomas Germain. 2023. A Mental Health App Tested ChatGPT on Its Users. The Founder Said Backlash Was Just a Misunderstanding.https:\/\/gizmodo.com\/mental-health-therapy-app-ai-koko-chatgpt-rob-morris-1849965534\/ Accessed: 09\/11\/2023."},{"key":"e_1_3_3_3_22_1","doi-asserted-by":"publisher","DOI":"10.1177\/0093650210377091"},{"key":"e_1_3_3_3_23_1","doi-asserted-by":"publisher","DOI":"10.1002\/9781444316568.wiem02057"},{"key":"e_1_3_3_3_24_1","volume-title":"\u201ccontextual integrity","author":"Grodzinsky Frances","year":"2010","unstructured":"Frances Grodzinsky and Herman\u00a0T Tavani. 2010. Applying the \u201ccontextual integrity\u201d model of privacy to personal blogs in the blogoshere. (2010)."},{"key":"e_1_3_3_3_25_1","volume-title":"The False Promise of Imitating Proprietary LLMs. arXiv preprint arXiv:2305.15717","author":"Gudibande Arnav","year":"2023","unstructured":"Arnav Gudibande, Eric Wallace, Charlie Snell, Xinyang Geng, Hao Liu, Pieter Abbeel, Sergey Levine, and Dawn Song. 2023. The False Promise of Imitating Proprietary LLMs. arXiv preprint arXiv:2305.15717 (2023)."},{"key":"e_1_3_3_3_26_1","volume-title":"How many interviews are enough? An experiment with data saturation and variability. Field methods 18, 1","author":"Guest Greg","year":"2006","unstructured":"Greg Guest, Arwen Bunce, and Laura Johnson. 2006. How many interviews are enough? An experiment with data saturation and variability. Field methods 18, 1 (2006), 59\u201382."},{"key":"e_1_3_3_3_27_1","first-page":"21","article-title":"What can I really do?","volume":"10","author":"Hargittai Eszter","year":"2016","unstructured":"Eszter Hargittai and Alice Marwick. 2016. \u201cWhat can I really do?\u201d Explaining the privacy paradox with online apathy. International journal of communication 10 (2016), 21.","journal-title":"International journal of communication"},{"key":"e_1_3_3_3_28_1","doi-asserted-by":"publisher","DOI":"10.1145\/3290605.3300645"},{"key":"e_1_3_3_3_29_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-39540-7_3"},{"key":"e_1_3_3_3_30_1","volume-title":"Knowledge unlearning for mitigating privacy risks in language models. arXiv preprint arXiv:2210.01504","author":"Jang Joel","year":"2022","unstructured":"Joel Jang, Dongkeun Yoon, Sohee Yang, Sungmin Cha, Moontae Lee, Lajanugen Logeswaran, and Minjoon Seo. 2022. Knowledge unlearning for mitigating privacy risks in language models. arXiv preprint arXiv:2210.01504 (2022)."},{"key":"e_1_3_3_3_31_1","volume-title":"Towards Better Instruction Following Language Models for Chinese: Investigating the Impact of Training Data and Evaluation. arXiv preprint arXiv:2304.07854","author":"Ji Yunjie","year":"2023","unstructured":"Yunjie Ji, Yan Gong, Yong Deng, Yiping Peng, Qiang Niu, Baochang Ma, and Xiangang Li. 2023. Towards Better Instruction Following Language Models for Chinese: Investigating the Impact of Training Data and Evaluation. arXiv preprint arXiv:2304.07854 (2023)."},{"key":"e_1_3_3_3_32_1","doi-asserted-by":"publisher","DOI":"10.1145\/2157689.2157696"},{"key":"e_1_3_3_3_33_1","volume-title":"International Conference on Machine Learning. PMLR, 10697\u201310707","author":"Kandpal Nikhil","year":"2022","unstructured":"Nikhil Kandpal, Eric Wallace, and Colin Raffel. 2022. Deduplicating training data mitigates privacy risks in language models. In International Conference on Machine Learning. PMLR, 10697\u201310707."},{"key":"e_1_3_3_3_34_1","volume-title":"User Mental Models of the Internet and Implications for Privacy and Security. In Eleventh symposium on usable privacy and security (SOUPS","author":"Kang Ruogu","year":"2015","unstructured":"Ruogu Kang, Laura Dabbish, Nathaniel Fruchter, and Sara Kiesler. 2015. { \u201cMy} Data Just Goes { Everywhere:\u201d} User Mental Models of the Internet and Implications for Privacy and Security. In Eleventh symposium on usable privacy and security (SOUPS 2015). 39\u201352."},{"key":"e_1_3_3_3_35_1","doi-asserted-by":"publisher","DOI":"10.1145\/3313831.3376219"},{"key":"e_1_3_3_3_36_1","volume-title":"ProPILE: Probing Privacy Leakage in Large Language Models. arXiv preprint arXiv:2307.01881","author":"Kim Siwon","year":"2023","unstructured":"Siwon Kim, Sangdoo Yun, Hwaran Lee, Martin Gubri, Sungroh Yoon, and Seong\u00a0Joon Oh. 2023. ProPILE: Probing Privacy Leakage in Large Language Models. arXiv preprint arXiv:2307.01881 (2023)."},{"key":"e_1_3_3_3_37_1","volume-title":"Anthropomorphism of computers: Is it mindful or mindless?Computers in Human Behavior 28, 1","author":"Kim Youjeong","year":"2012","unstructured":"Youjeong Kim and S\u00a0Shyam Sundar. 2012. Anthropomorphism of computers: Is it mindful or mindless?Computers in Human Behavior 28, 1 (2012), 241\u2013250."},{"key":"e_1_3_3_3_38_1","unstructured":"Daniel Kimmel. 2023. ChatGPT Therapy Is Good But It Misses What Makes Us Human. https:\/\/www.columbiapsychiatry.org\/news\/chatgpt-therapy-is-good-but-it-misses-what-makes-us-human. Accessed: 09\/11\/2023."},{"key":"e_1_3_3_3_39_1","doi-asserted-by":"publisher","DOI":"10.1109\/MITP.2023.3275489"},{"key":"e_1_3_3_3_40_1","volume-title":"Google","author":"Leonard Andrew","year":"2023","unstructured":"Andrew Leonard. 2023. \u2018Dr. Google\u2019 meets its match: Dr. ChatGPT. https:\/\/www.latimes.com\/science\/story\/2023-09-08\/dr-google-meets-its-match-dr-chatgpt Accessed: 09\/11\/2023."},{"key":"e_1_3_3_3_41_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_3_3_42_1","volume-title":"Large language models can be strong differentially private learners. arXiv preprint arXiv:2110.05679","author":"Li Xuechen","year":"2021","unstructured":"Xuechen Li, Florian Tramer, Percy Liang, and Tatsunori Hashimoto. 2021. Large language models can be strong differentially private learners. arXiv preprint arXiv:2110.05679 (2021)."},{"key":"e_1_3_3_3_43_1","volume-title":"Privacy protection and self-disclosure across societies: A study of global Twitter users. new media & society 19, 9","author":"Liang Hai","year":"2017","unstructured":"Hai Liang, Fei Shen, and King-wa Fu. 2017. Privacy protection and self-disclosure across societies: A study of global Twitter users. new media & society 19, 9 (2017), 1476\u20131497."},{"key":"e_1_3_3_3_44_1","volume-title":"AI Transparency in the Age of LLMs: A Human-Centered Research Roadmap. arXiv preprint arXiv:2306.01941","author":"Liao Q\u00a0Vera","year":"2023","unstructured":"Q\u00a0Vera Liao and Jennifer\u00a0Wortman Vaughan. 2023. AI Transparency in the Age of LLMs: A Human-Centered Research Roadmap. arXiv preprint arXiv:2306.01941 (2023)."},{"key":"e_1_3_3_3_45_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_3_3_46_1","volume-title":"Sixteenth Symposium on Usable Privacy and Security (SOUPS","author":"Mai Alexandra","year":"2020","unstructured":"Alexandra Mai, Katharina Pfeffer, Matthias Gusenbauer, Edgar Weippl, and Katharina Krombholz. 2020. User mental models of cryptocurrency systems-a grounded theory approach. In Sixteenth Symposium on Usable Privacy and Security (SOUPS 2020). 341\u2013358."},{"key":"e_1_3_3_3_47_1","volume-title":"Differentially private decoding in large language models. arXiv preprint arXiv:2205.13621","author":"Majmudar Jimit","year":"2022","unstructured":"Jimit Majmudar, Christophe Dupuy, Charith Peris, Sami Smaili, Rahul Gupta, and Richard Zemel. 2022. Differentially private decoding in large language models. arXiv preprint arXiv:2205.13621 (2022)."},{"key":"e_1_3_3_3_48_1","doi-asserted-by":"publisher","DOI":"10.1145\/3359174"},{"key":"e_1_3_3_3_49_1","volume-title":"Understanding human impressions of artificial intelligence. Preprint]. PsyArXiv. https:\/\/doi. org\/10.31234\/osf. io\/5ursp","author":"McKee KR","year":"2021","unstructured":"KR McKee, X Bai, and S Fiske. 2021. Understanding human impressions of artificial intelligence. Preprint]. PsyArXiv. https:\/\/doi. org\/10.31234\/osf. io\/5ursp (2021)."},{"key":"e_1_3_3_3_50_1","volume-title":"EmbodiedGPT: Vision-Language Pre-Training via Embodied Chain of Thought. arXiv preprint arXiv:2305.15021","author":"Mu Yao","year":"2023","unstructured":"Yao Mu, Qinglong Zhang, Mengkang Hu, Wenhai Wang, Mingyu Ding, Jun Jin, Bin Wang, Jifeng Dai, Yu Qiao, and Ping Luo. 2023. EmbodiedGPT: Vision-Language Pre-Training via Embodied Chain of Thought. arXiv preprint arXiv:2305.15021 (2023)."},{"key":"e_1_3_3_3_51_1","doi-asserted-by":"publisher","DOI":"10.1145\/3319502.3374839"},{"key":"e_1_3_3_3_52_1","first-page":"119","article-title":"Privacy as contextual integrity","volume":"79","author":"Nissenbaum Helen","year":"2004","unstructured":"Helen Nissenbaum. 2004. Privacy as contextual integrity. Wash. L. Rev. 79 (2004), 119.","journal-title":"Wash. L. Rev."},{"key":"e_1_3_3_3_53_1","volume-title":"Privacy in context: Technology, policy, and the integrity of social life","author":"Nissenbaum Helen","unstructured":"Helen Nissenbaum. 2020. Privacy in context: Technology, policy, and the integrity of social life. Stanford University Press."},{"key":"e_1_3_3_3_54_1","volume-title":"Mental models","author":"Norman A","unstructured":"Donald\u00a0A Norman. 2014. Some observations on mental models. In Mental models. Psychology Press, 15\u201322."},{"key":"e_1_3_3_3_55_1","first-page":"27730","article-title":"Training language models to follow instructions with human feedback","volume":"35","author":"Ouyang Long","year":"2022","unstructured":"Long Ouyang, Jeffrey Wu, Xu Jiang, Diogo Almeida, Carroll Wainwright, Pamela Mishkin, Chong Zhang, Sandhini Agarwal, Katarina Slama, Alex Ray, 2022. Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems 35 (2022), 27730\u201327744.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_3_3_56_1","volume-title":"Several categories of Large Language Models (LLMs): A Short Survey. arXiv preprint arXiv:2307.10188","author":"Pahune Saurabh","year":"2023","unstructured":"Saurabh Pahune and Manoj Chandrasekharan. 2023. Several categories of Large Language Models (LLMs): A Short Survey. arXiv preprint arXiv:2307.10188 (2023)."},{"key":"e_1_3_3_3_57_1","doi-asserted-by":"publisher","DOI":"10.1109\/SP40000.2020.00095"},{"key":"e_1_3_3_3_58_1","doi-asserted-by":"publisher","unstructured":"Charith Peris Christophe Dupuy Jimit Majmudar Rahil Parikh Sami Smaili Richard Zemel and Rahul Gupta. 2023. Privacy in the Time of Language Models. https:\/\/doi.org\/10.1145\/3539597.3575792","DOI":"10.1145\/3539597.3575792"},{"key":"e_1_3_3_3_59_1","doi-asserted-by":"publisher","unstructured":"Karen Renaud Melanie Volkamer and Arne Renkema-Padmos. 2014. Why Doesn\u2019t Jane Protect Her Privacy?Springer International Publishing 244\u2013262. https:\/\/doi.org\/10.1007\/978-3-319-08506-7_13","DOI":"10.1007\/978-3-319-08506-7_13"},{"key":"e_1_3_3_3_60_1","volume-title":"CHI 2019 Workshop: Where is the Human? Bridging the Gap Between AI and HCI. Association for Computing Machinery, Inc.","author":"Rutjes Heleen","year":"2019","unstructured":"Heleen Rutjes, Martijn Willemsen, and Wijnand IJsselsteijn. 2019. Considerations on explainable AI and users\u2019 mental models. In CHI 2019 Workshop: Where is the Human? Bridging the Gap Between AI and HCI. Association for Computing Machinery, Inc."},{"key":"e_1_3_3_3_61_1","volume-title":"The coding manual for qualitative researchers","author":"Salda\u00f1a Johnny","unstructured":"Johnny Salda\u00f1a. 2015. The coding manual for qualitative researchers. Sage."},{"key":"e_1_3_3_3_62_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICPHM49022.2020.9187036"},{"key":"e_1_3_3_3_63_1","doi-asserted-by":"publisher","DOI":"10.1609\/aimag.v32i4.2383"},{"key":"e_1_3_3_3_64_1","doi-asserted-by":"publisher","DOI":"10.2139\/ssrn.3536265"},{"key":"e_1_3_3_3_65_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.chb.2010.10.017"},{"key":"e_1_3_3_3_66_1","unstructured":"Mikhail Taver. 2023. ChatGPT is Coming to Finance So Let\u2019s Talk About the Risks and Rewards. https:\/\/www.unite.ai\/chatgpt-is-coming-to-finance-so-lets-talk-about-the-risks-and-rewards\/. Accessed: 09\/11\/2023."},{"key":"e_1_3_3_3_67_1","volume-title":"24th USENIX Security Symposium (USENIX Security 15)","author":"Ur Blase","year":"2015","unstructured":"Blase Ur, Sean\u00a0M Segreti, Lujo Bauer, Nicolas Christin, Lorrie\u00a0Faith Cranor, Saranga Komanduri, Darya Kurilova, Michelle\u00a0L Mazurek, William Melicher, and Richard Shay. 2015. Measuring { Real-World} Accuracies and Biases in Modeling Password Guessability. In 24th USENIX Security Symposium (USENIX Security 15). 463\u2013481."},{"key":"e_1_3_3_3_68_1","volume-title":"Privacy as trust: Information privacy for an information age","author":"Waldman Ari\u00a0Ezra","unstructured":"Ari\u00a0Ezra Waldman. 2018. Privacy as trust: Information privacy for an information age. Cambridge University Press."},{"key":"e_1_3_3_3_69_1","volume-title":"Industry unbound: The inside story of privacy, data, and corporate power","author":"Waldman Ari\u00a0Ezra","unstructured":"Ari\u00a0Ezra Waldman. 2021. Industry unbound: The inside story of privacy, data, and corporate power. Cambridge University Press."},{"key":"e_1_3_3_3_70_1","doi-asserted-by":"publisher","DOI":"10.1145\/2078827.2078841"},{"key":"e_1_3_3_3_71_1","volume-title":"Ethical and social risks of harm from language models. arXiv preprint arXiv:2112.04359","author":"Weidinger Laura","year":"2021","unstructured":"Laura Weidinger, John Mellor, Maribeth Rauh, Conor Griffin, Jonathan Uesato, Po-Sen Huang, Myra Cheng, Mia Glaese, Borja Balle, Atoosa Kasirzadeh, 2021. Ethical and social risks of harm from language models. arXiv preprint arXiv:2112.04359 (2021)."},{"key":"e_1_3_3_3_72_1","doi-asserted-by":"publisher","DOI":"10.1109\/access.2023.3282111"},{"key":"e_1_3_3_3_73_1","volume-title":"Differentially private fine-tuning of language models. arXiv preprint arXiv:2110.06500","author":"Yu Da","year":"2021","unstructured":"Da Yu, Saurabh Naik, Arturs Backurs, Sivakanth Gopi, Huseyin\u00a0A Inan, Gautam Kamath, Janardhan Kulkarni, Yin\u00a0Tat Lee, Andre Manoel, Lukas Wutschitz, 2021. Differentially private fine-tuning of language models. arXiv preprint arXiv:2110.06500 (2021)."},{"key":"e_1_3_3_3_74_1","volume-title":"End user security and privacy concerns with smart homes. In thirteenth symposium on usable privacy and security (SOUPS","author":"Zeng Eric","year":"2017","unstructured":"Eric Zeng, Shrirang Mare, and Franziska Roesner. 2017. End user security and privacy concerns with smart homes. In thirteenth symposium on usable privacy and security (SOUPS 2017). 65\u201380."},{"key":"e_1_3_3_3_75_1","volume-title":"Counterfactual memorization in neural language models. arXiv preprint arXiv:2112.12938","author":"Zhang Chiyuan","year":"2021","unstructured":"Chiyuan Zhang, Daphne Ippolito, Katherine Lee, Matthew Jagielski, Florian Tram\u00e8r, and Nicholas Carlini. 2021. Counterfactual memorization in neural language models. arXiv preprint arXiv:2112.12938 (2021)."},{"key":"e_1_3_3_3_76_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_3_3_77_1","volume-title":"A survey of large language models. arXiv preprint arXiv:2303.18223","author":"Zhao Wayne\u00a0Xin","year":"2023","unstructured":"Wayne\u00a0Xin Zhao, Kun Zhou, Junyi Li, Tianyi Tang, Xiaolei Wang, Yupeng Hou, Yingqian Min, Beichen Zhang, Junjie Zhang, Zican Dong, 2023. A survey of large language models. arXiv preprint arXiv:2303.18223 (2023)."},{"key":"e_1_3_3_3_78_1","volume-title":"Judging LLM-as-a-judge with MT-Bench and Chatbot Arena. arXiv preprint arXiv:2306.05685","author":"Zheng Lianmin","year":"2023","unstructured":"Lianmin Zheng, Wei-Lin Chiang, Ying Sheng, Siyuan Zhuang, Zhanghao Wu, Yonghao Zhuang, Zi Lin, Zhuohan Li, Dacheng Li, Eric Xing, 2023. Judging LLM-as-a-judge with MT-Bench and Chatbot Arena. arXiv preprint arXiv:2306.05685 (2023)."},{"key":"e_1_3_3_3_79_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.chb.2014.12.012"},{"key":"e_1_3_3_3_80_1","doi-asserted-by":"publisher","DOI":"10.1007\/s12369-014-0267-6"},{"key":"e_1_3_3_3_81_1","volume-title":"Social Theory Re-Wired","author":"Zuboff Shoshana","unstructured":"Shoshana Zuboff. 2023. The age of surveillance capitalism. In Social Theory Re-Wired. Routledge, 203\u2013213."}],"event":{"name":"CHI '24: CHI Conference on Human Factors in Computing Systems","location":"Honolulu HI USA","acronym":"CHI '24","sponsor":["SIGCHI ACM Special Interest Group on Computer-Human Interaction","SIGACCESS ACM Special Interest Group on Accessible Computing"]},"container-title":["Proceedings of the CHI Conference on Human Factors in Computing Systems"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3613904.3642385","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3613904.3642385","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T23:56:47Z","timestamp":1750291007000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3613904.3642385"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,5,11]]},"references-count":81,"alternative-id":["10.1145\/3613904.3642385","10.1145\/3613904"],"URL":"https:\/\/doi.org\/10.1145\/3613904.3642385","relation":{},"subject":[],"published":{"date-parts":[[2024,5,11]]},"assertion":[{"value":"2024-05-11","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}