{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,10]],"date-time":"2026-03-10T12:12:51Z","timestamp":1773144771584,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":80,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,8,8]],"date-time":"2023-08-08T00:00:00Z","timestamp":1691452800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,8,8]]},"DOI":"10.1145\/3600211.3604685","type":"proceedings-article","created":{"date-parts":[[2023,8,29]],"date-time":"2023-08-29T18:41:37Z","timestamp":1693334497000},"page":"70-83","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":6,"title":["Beyond the ML Model: Applying Safety Engineering Frameworks to Text-to-Image Development"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-5281-2428","authenticated-orcid":false,"given":"Shalaleh","family":"Rismani","sequence":"first","affiliation":[{"name":"McGill University, Canada"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4720-3844","authenticated-orcid":false,"given":"Renee","family":"Shelby","sequence":"additional","affiliation":[{"name":"Google Research, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9816-7348","authenticated-orcid":false,"given":"Andrew","family":"Smart","sequence":"additional","affiliation":[{"name":"Google Research, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0008-0407-5946","authenticated-orcid":false,"given":"Renelito","family":"Delos Santos","sequence":"additional","affiliation":[{"name":"Google Research, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9387-6284","authenticated-orcid":false,"given":"AJung","family":"Moon","sequence":"additional","affiliation":[{"name":"McGill University, Canada"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9638-4664","authenticated-orcid":false,"given":"Negar","family":"Rostamzadeh","sequence":"additional","affiliation":[{"name":"Google Research, Canada"}]}],"member":"320","published-online":{"date-parts":[[2023,8,29]]},"reference":[{"key":"e_1_3_2_2_2_1","unstructured":"Stability AI. 2022. Stable Diffusion. https:\/\/stablediffusionweb.com\/. Accessed: 2023-1-27."},{"key":"e_1_3_2_2_3_1","doi-asserted-by":"publisher","DOI":"10.1016\/S0140-6736(23)00216-7"},{"key":"e_1_3_2_2_4_1","unstructured":"Pesala Bandara. 2022. New Tool Allows Users to See Bias in AI Image Generators. https:\/\/petapixel.com\/2022\/11\/09\/new-tool-allows-users-to-see-bias-in-ai-image-generators\/. Accessed: 2023-1-28."},{"key":"e_1_3_2_2_5_1","first-page":"2","article-title":"Algorithmic injustice: a relational ethics approach","volume":"2","author":"Birhane Abeba","year":"2021","unstructured":"Abeba Birhane. 2021. Algorithmic injustice: a relational ethics approach. Patterns (N Y) 2, 2 (Feb. 2021), 100205.","journal-title":"Patterns (N Y)"},{"key":"e_1_3_2_2_6_1","volume-title":"Multimodal datasets: misogyny, pornography, and malignant stereotypes. (Oct","author":"Birhane Abeba","year":"2021","unstructured":"Abeba Birhane, Vinay\u00a0Uday Prabhu, and Emmanuel Kahembwe. 2021. Multimodal datasets: misogyny, pornography, and malignant stereotypes. (Oct. 2021). arxiv:2110.01963\u00a0[cs.CY]"},{"key":"e_1_3_2_2_7_1","volume-title":"Responsible Language Technologies: Foreseeing and Mitigating Harms. In Extended Abstracts of the 2022 CHI Conference on Human Factors in Computing Systems","author":"Blodgett Su\u00a0Lin","year":"2022","unstructured":"Su\u00a0Lin Blodgett, Q\u00a0Vera Liao, Alexandra Olteanu, Rada Mihalcea, Michael Muller, Morgan\u00a0Klaus Scheuerman, Chenhao Tan, and Qian Yang. 2022. Responsible Language Technologies: Foreseeing and Mitigating Harms. In Extended Abstracts of the 2022 CHI Conference on Human Factors in Computing Systems (New Orleans, LA, USA) (CHI EA \u201922, Article 152). Association for Computing Machinery, New York, NY, USA, 1\u20133."},{"key":"e_1_3_2_2_8_1","volume-title":"On the Opportunities and Risks of Foundation Models. (Aug","author":"Bommasani Rishi","year":"2021","unstructured":"Rishi Bommasani, Drew\u00a0A Hudson, Ehsan Adeli, Russ Altman, Simran Arora, Sydney von Arx, Michael\u00a0S Bernstein, Jeannette Bohg, Antoine Bosselut, Emma Brunskill, Erik Brynjolfsson, Shyamal Buch, Dallas Card, Rodrigo Castellon, Niladri Chatterji, Annie Chen, Kathleen Creel, Jared\u00a0Quincy Davis, Dora Demszky, Chris Donahue, Moussa Doumbouya, Esin Durmus, Stefano Ermon, John Etchemendy, Kawin Ethayarajh, Li Fei-Fei, Chelsea Finn, Trevor Gale, Lauren Gillespie, Karan Goel, Noah Goodman, Shelby Grossman, Neel Guha, Tatsunori Hashimoto, Peter Henderson, John Hewitt, Daniel\u00a0E Ho, Jenny Hong, Kyle Hsu, Jing Huang, Thomas Icard, Saahil Jain, Dan Jurafsky, Pratyusha Kalluri, Siddharth Karamcheti, Geoff Keeling, Fereshte Khani, Omar Khattab, Pang\u00a0Wei Koh, Mark Krass, Ranjay Krishna, Rohith Kuditipudi, Ananya Kumar, Faisal Ladhak, Mina Lee, Tony Lee, Jure Leskovec, Isabelle Levent, Xiang\u00a0Lisa Li, Xuechen Li, Tengyu Ma, Ali Malik, Christopher\u00a0D Manning, Suvir Mirchandani, Eric Mitchell, Zanele Munyikwa, Suraj Nair, Avanika Narayan, Deepak Narayanan, Ben Newman, Allen Nie, Juan\u00a0Carlos Niebles, Hamed Nilforoshan, Julian Nyarko, Giray Ogut, Laurel Orr, Isabel Papadimitriou, Joon\u00a0Sung Park, Chris Piech, Eva Portelance, Christopher Potts, Aditi Raghunathan, Rob Reich, Hongyu Ren, Frieda Rong, Yusuf Roohani, Camilo Ruiz, Jack Ryan, Christopher R\u00e9, Dorsa Sadigh, Shiori Sagawa, Keshav Santhanam, Andy Shih, Krishnan Srinivasan, Alex Tamkin, Rohan Taori, Armin\u00a0W Thomas, Florian Tram\u00e8r, Rose\u00a0E Wang, William Wang, Bohan Wu, Jiajun Wu, Yuhuai Wu, Sang\u00a0Michael Xie, Michihiro Yasunaga, Jiaxuan You, Matei Zaharia, Michael Zhang, Tianyi Zhang, Xikun Zhang, Yuhui Zhang, Lucia Zheng, Kaitlyn Zhou, and Percy Liang. 2021. On the Opportunities and Risks of Foundation Models. (Aug. 2021). arxiv:2108.07258\u00a0[cs.LG]"},{"key":"e_1_3_2_2_9_1","volume-title":"Explainable and Transparent AI and Multi-Agent Systems","author":"Carli Rachele","unstructured":"Rachele Carli, Amro Najjar, and Davide Calvaresi. 2022. Risk and Exposure of XAI in Persuasion and Argumentation: The case of Manipulation. In Explainable and Transparent AI and Multi-Agent Systems. Springer International Publishing, 204\u2013220."},{"key":"e_1_3_2_2_10_1","volume-title":"Effective FMEAs: achieving safe, reliable, and economical products and processes using failure mode and effects analysis","author":"Carlson Carl","unstructured":"Carl Carlson. 2012. Effective FMEAs: achieving safe, reliable, and economical products and processes using failure mode and effects analysis. Wiley, Hoboken, N.J."},{"key":"e_1_3_2_2_11_1","first-page":"2s","article-title":"Art by Computing Machinery: Is Machine Art Acceptable in the Artworld?ACM Trans","volume":"15","author":"Ch\u2019ng Eugene","year":"2019","unstructured":"Eugene Ch\u2019ng. 2019. Art by Computing Machinery: Is Machine Art Acceptable in the Artworld?ACM Trans. Multimedia Comput. Commun. Appl. 15, 2s (July 2019), 1\u201317.","journal-title":"Multimedia Comput. Commun. Appl."},{"key":"e_1_3_2_2_12_1","volume-title":"DallEval: DALL-Eval: Probing the Reasoning Skills and Social Biases of Text-to-Image Generative Transformers. (Nov","author":"Cho Jaemin","year":"2022","unstructured":"Jaemin Cho, Abhay Zala, and Mohit Bansal. 2022. DallEval: DALL-Eval: Probing the Reasoning Skills and Social Biases of Text-to-Image Generative Transformers. (Nov. 2022)."},{"key":"e_1_3_2_2_13_1","unstructured":"Laurie Clarke. 2022. When AI can make art \u2013 what does it mean for creativity?https:\/\/www.theguardian.com\/technology\/2022\/nov\/12\/when-ai-can-make-art-what-does-it-mean-for-creativity-dall-e-midjourney. Accessed: 2022-11-12."},{"key":"e_1_3_2_2_15_1","volume-title":"How Generative AI Is Changing Creative Work. Harvard Business Review (Nov","author":"Davenport H","year":"2022","unstructured":"Thomas\u00a0H Davenport and Nitin Mittal. 2022. How Generative AI Is Changing Creative Work. Harvard Business Review (Nov. 2022)."},{"key":"e_1_3_2_2_16_1","volume-title":"Foundations of Safety Science: A Century of Understanding Accidents and Disasters","author":"Dekker Sidney","unstructured":"Sidney Dekker. 2019. Foundations of Safety Science: A Century of Understanding Accidents and Disasters. Routledge."},{"key":"e_1_3_2_2_17_1","unstructured":"Emily Denton and Burcu\u00a0Karagol Ayan. 2022. FIT400M Data Card. https:\/\/github.com\/google-research\/parti\/blob\/main\/data_cards\/fit400m_data_card.pdf. Accessed: 2023-3-10."},{"key":"e_1_3_2_2_18_1","doi-asserted-by":"publisher","DOI":"10.1177\/20539517211035955"},{"key":"e_1_3_2_2_19_1","unstructured":"Stable Diffusion. 2022. Stable Diffusion Online. https:\/\/stablediffusionweb.com\/. Accessed: 2023-1-30."},{"key":"e_1_3_2_2_20_1","volume-title":"System Safety and Artificial Intelligence. (Feb","author":"Dobbe Roel I\u00a0J","year":"2022","unstructured":"Roel I\u00a0J Dobbe. 2022. System Safety and Artificial Intelligence. (Feb. 2022). arxiv:2202.09292\u00a0[eess.SY]"},{"key":"e_1_3_2_2_21_1","unstructured":"Benj Edwards. 2022. Have AI image generators assimilated your art? New tool lets you check. https:\/\/arstechnica.com\/information-technology\/2022\/09\/have-ai-image-generators-assimilated-your-art-new-tool-lets-you-check\/. Accessed: 2022-09-15."},{"key":"e_1_3_2_2_22_1","doi-asserted-by":"publisher","DOI":"10.1145\/3411764.3445188"},{"key":"e_1_3_2_2_23_1","doi-asserted-by":"publisher","DOI":"10.1145\/3514094.3534137"},{"key":"e_1_3_2_2_24_1","volume-title":"Value Sensitive Design","author":"Friedman Batya","unstructured":"Batya Friedman and David\u00a0G Hendry. 2019. Value Sensitive Design. MIT Press."},{"key":"e_1_3_2_2_25_1","volume-title":"Visual Conceptual Blending with Large-scale Language and Vision Models. (June","author":"Ge Songwei","year":"2021","unstructured":"Songwei Ge and Devi Parikh. 2021. Visual Conceptual Blending with Large-scale Language and Vision Models. (June 2021). arxiv:2106.14127\u00a0[cs.CL]"},{"key":"e_1_3_2_2_26_1","doi-asserted-by":"publisher","DOI":"10.1145\/3458723"},{"key":"e_1_3_2_2_27_1","volume-title":"How do we audit generative algorithms?","author":"Gero Katy\u00a0Ilonka","year":"2022","unstructured":"Katy\u00a0Ilonka Gero. 2022. How do we audit generative algorithms? (2022)."},{"key":"e_1_3_2_2_28_1","volume-title":"An Interaction Framework for Studying Co-Creative AI. (March","author":"Guzdial Matthew","year":"2019","unstructured":"Matthew Guzdial and Mark Riedl. 2019. An Interaction Framework for Studying Co-Creative AI. (March 2019). arxiv:1903.09709\u00a0[cs.HC]"},{"key":"e_1_3_2_2_29_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3372826"},{"key":"e_1_3_2_2_30_1","doi-asserted-by":"publisher","DOI":"10.1145\/3347092"},{"key":"e_1_3_2_2_31_1","volume-title":"Alex Hanna, Emily Denton, Christina Greer, Oddur Kjartansson, Parker Barnes, and Margaret Mitchell.","author":"Hutchinson Ben","year":"2020","unstructured":"Ben Hutchinson, Andrew Smart, Alex Hanna, Emily Denton, Christina Greer, Oddur Kjartansson, Parker Barnes, and Margaret Mitchell. 2020. Towards Accountability for Machine Learning Datasets: Practices from Software Engineering and Infrastructure. (Oct. 2020). arxiv:2010.13561\u00a0[cs.LG]"},{"key":"e_1_3_2_2_33_1","unstructured":"Takuto Ishimatsu Nancy\u00a0G Leveson John Thomas Masa Katahira Yuko Miyamoto and Haruka Nakao. 2010. Modeling and hazard analysis using STPA. (2010)."},{"key":"e_1_3_2_2_34_1","doi-asserted-by":"publisher","DOI":"10.2514\/1.A32449"},{"key":"e_1_3_2_2_35_1","doi-asserted-by":"publisher","DOI":"10.5267\/j.msl.2015.5.006"},{"key":"e_1_3_2_2_36_1","unstructured":"johannezz. 2021. The Promptist Manifesto. https:\/\/deeplearn.art\/the-promptist-manifesto\/. Accessed: 2023-1-27."},{"key":"e_1_3_2_2_37_1","volume-title":"DALL-E 2 Creates Incredible Images\u2014and Biased Ones You Don\u2019t See. Wired (May","author":"Johnson Khari","year":"2022","unstructured":"Khari Johnson. 2022. DALL-E 2 Creates Incredible Images\u2014and Biased Ones You Don\u2019t See. Wired (May 2022)."},{"key":"e_1_3_2_2_39_1","volume-title":"A Hazard Analysis Framework for Code Synthesis Large Language Models. (July","author":"Khlaaf Heidy","year":"2022","unstructured":"Heidy Khlaaf, Pamela Mishkin, Joshua Achiam, Gretchen Krueger, and Miles Brundage. 2022. A Hazard Analysis Framework for Code Synthesis Large Language Models. (July 2022). arxiv:2207.14157\u00a0[cs.SE]"},{"key":"e_1_3_2_2_40_1","doi-asserted-by":"publisher","DOI":"10.1145\/3411764.3445261"},{"key":"e_1_3_2_2_41_1","unstructured":"Nancy Leveson and John Thomas. 2018. STPA_Handbook."},{"key":"e_1_3_2_2_42_1","volume-title":"Engineering a safer world: Systems thinking applied to safety","author":"Leveson G","unstructured":"Nancy\u00a0G Leveson. 2016. Engineering a safer world: Systems thinking applied to safety. The MIT Press."},{"key":"e_1_3_2_2_43_1","volume-title":"FMEA-AI: AI fairness impact assessment using failure mode and effects analysis. AI and Ethics (March","author":"Li Jamy","year":"2022","unstructured":"Jamy Li and Mark Chignell. 2022. FMEA-AI: AI fairness impact assessment using failure mode and effects analysis. AI and Ethics (March 2022)."},{"key":"e_1_3_2_2_44_1","doi-asserted-by":"publisher","DOI":"10.1145\/3313831.3376590"},{"key":"e_1_3_2_2_45_1","volume-title":"Jen-Jen Yang, Chun-Nen Huang, and Yu-Hsuan Lu.","author":"Lo Huai-Wei","year":"2021","unstructured":"Huai-Wei Lo, James J\u00a0H Liou, Jen-Jen Yang, Chun-Nen Huang, and Yu-Hsuan Lu. 2021. An Extended FMEA Model for Exploring the Potential Failure Modes: A Case Study of a Steam Turbine for a Nuclear Power Plant. Hindawi (2021)."},{"key":"e_1_3_2_2_46_1","doi-asserted-by":"crossref","unstructured":"Michael Madaio Lisa Egede Hariharan Subramonyam Jennifer\u00a0Wortman Vaughan and Hanna Wallach. 2022. Assessing the Fairness of AI Systems: AI Practitioners\u2019 Processes Challenges and Needs for Support. 26\u00a0pages.","DOI":"10.1145\/3512899"},{"key":"e_1_3_2_2_47_1","doi-asserted-by":"publisher","DOI":"10.1145\/3313831.3376445"},{"key":"e_1_3_2_2_48_1","volume-title":"Putting AI Ethics into Practice: The Hourglass Model of Organizational AI Governance. (June","author":"M\u00e4ntym\u00e4ki Matti","year":"2022","unstructured":"Matti M\u00e4ntym\u00e4ki, Matti Minkkinen, Teemu Birkstedt, and Mika Viljanen. 2022. Putting AI Ethics into Practice: The Hourglass Model of Organizational AI Governance. (June 2022). arxiv:2206.00335\u00a0[cs.AI]"},{"key":"e_1_3_2_2_49_1","doi-asserted-by":"publisher","unstructured":"Nikolas Martelaro Carol\u00a0J. Smith and Tamara Zilovic. 2022. Exploring Opportunities in Usable Hazard Analysis Processes for AI Engineering. https:\/\/doi.org\/10.48550\/ARXIV.2203.15628","DOI":"10.48550\/ARXIV.2203.15628"},{"key":"e_1_3_2_2_50_1","volume-title":"Andrew Smart, and William\u00a0S Isaac","author":"Martin Donald","year":"2020","unstructured":"Donald Martin, Jr, Vinodkumar Prabhakaran, Jill Kuhlberg, Andrew Smart, and William\u00a0S Isaac. 2020. Participatory Problem Formulation for Fairer Machine Learning Through Community Based System Dynamics. (May 2020). arxiv:2005.07572\u00a0[cs.CY]"},{"key":"e_1_3_2_2_51_1","doi-asserted-by":"publisher","DOI":"10.1145\/3287560.3287596"},{"key":"e_1_3_2_2_52_1","doi-asserted-by":"publisher","DOI":"10.1007\/s13347-020-00405-8"},{"key":"e_1_3_2_2_54_1","volume-title":"Ivan Zhang, Carol Chen, Adrien Morisot, and Nicholas Frosst.","author":"Ngo Helen","year":"2021","unstructured":"Helen Ngo, Cooper Raterink, Jo\u00e3o G\u00a0M Ara\u00fajo, Ivan Zhang, Carol Chen, Adrien Morisot, and Nicholas Frosst. 2021. Mitigating harm in language models with conditional-likelihood filtration. (Aug. 2021). arxiv:2108.07790\u00a0[cs.CL]"},{"key":"e_1_3_2_2_55_1","volume-title":"Purposeful sampling for qualitative data collection and analysis in mixed method implementation research. Administration and policy in mental health and mental health services research 42","author":"Palinkas A","year":"2015","unstructured":"Lawrence\u00a0A Palinkas, Sarah\u00a0M Horwitz, Carla\u00a0A Green, Jennifer\u00a0P Wisdom, Naihua Duan, and Kimberly Hoagwood. 2015. Purposeful sampling for qualitative data collection and analysis in mixed method implementation research. Administration and policy in mental health and mental health services research 42 (2015), 533\u2013544."},{"key":"e_1_3_2_2_56_1","unstructured":"Charlie Parker Sam Scott and Alistair Geddes. 2019. Snowball sampling."},{"key":"e_1_3_2_2_57_1","volume-title":"Initial Images: Using Image Prompts to Improve Subject Representation in Multimodal AI Generated Art. In Creativity and Cognition (Venice, Italy) (C&C \u201922)","author":"Qiao Han","year":"2022","unstructured":"Han Qiao, Vivian Liu, and Lydia Chilton. 2022. Initial Images: Using Image Prompts to Improve Subject Representation in Multimodal AI Generated Art. In Creativity and Cognition (Venice, Italy) (C&C \u201922). Association for Computing Machinery, New York, NY, USA, 15\u201328."},{"key":"e_1_3_2_2_58_1","volume-title":"Learning Transferable Visual Models From Natural Language Supervision. 139","author":"Radford Alec","year":"2021","unstructured":"Alec Radford, Jong\u00a0Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, Gretchen Krueger, and Ilya Sutskever. 2021. Learning Transferable Visual Models From Natural Language Supervision. 139 (2021), 8748\u20138763."},{"key":"e_1_3_2_2_59_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3372873"},{"key":"e_1_3_2_2_60_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3372873"},{"key":"e_1_3_2_2_61_1","doi-asserted-by":"publisher","DOI":"10.1145\/3449081"},{"key":"e_1_3_2_2_62_1","doi-asserted-by":"publisher","DOI":"10.1145\/3449081"},{"key":"e_1_3_2_2_63_1","unstructured":"Aditya Ramesh Mikhail Pavlov Gabriel Goh and Scott Gray. 2022. DALL-E 2. https:\/\/openai.com\/dall-e-2\/. Accessed: 2023-1-27."},{"key":"e_1_3_2_2_64_1","volume-title":"Red-Teaming the Stable Diffusion Safety Filter. (Oct","author":"Rando Javier","year":"2022","unstructured":"Javier Rando, Daniel Paleka, David Lindner, Lennart Heim, and Florian Tram\u00e8r. 2022. Red-Teaming the Stable Diffusion Safety Filter. (Oct. 2022). arxiv:2210.04610\u00a0[cs.AI]"},{"key":"e_1_3_2_2_65_1","doi-asserted-by":"publisher","DOI":"10.1177\/1609406919862424"},{"key":"e_1_3_2_2_66_1","doi-asserted-by":"publisher","DOI":"10.1109\/ETHICS53270.2021.9632769"},{"key":"e_1_3_2_2_67_1","volume-title":"Edgar Jatho, Joshua Kroll, Ajung Moon, and Negar Rostamzadeh.","author":"Rismani Shalaleh","year":"2022","unstructured":"Shalaleh Rismani, Renee Shelby, Andrew Smart, Edgar Jatho, Joshua Kroll, Ajung Moon, and Negar Rostamzadeh. 2022. From plane crashes to algorithmic harm: applicability of safety engineering frameworks for responsible ML. (Oct. 2022). arxiv:2210.03535\u00a0[cs.HC]"},{"key":"e_1_3_2_2_68_1","volume-title":"Burcu\u00a0Karagol Ayan, S Sara\u00a0Mahdavi, Rapha\u00a0Gontijo Lopes, Tim Salimans, Jonathan Ho, David\u00a0J Fleet, and Mohammad Norouzi.","author":"Saharia Chitwan","year":"2022","unstructured":"Chitwan Saharia, William Chan, Saurabh Saxena, Lala Li, Jay Whang, Emily Denton, Seyed Kamyar\u00a0Seyed Ghasemipour, Burcu\u00a0Karagol Ayan, S Sara\u00a0Mahdavi, Rapha\u00a0Gontijo Lopes, Tim Salimans, Jonathan Ho, David\u00a0J Fleet, and Mohammad Norouzi. 2022. Photorealistic Text-to-Image Diffusion Models with Deep Language Understanding. (May 2022). arxiv:2205.11487\u00a0[cs.CV]"},{"key":"e_1_3_2_2_69_1","unstructured":"Rob Salkowitz. 2022. AI Is Coming For Commercial Art Jobs. Can It Be Stopped?https:\/\/www.forbes.com\/sites\/robsalkowitz\/2022\/09\/16\/ai-is-coming-for-commercial-art-jobs-can-it-be-stopped\/?sh=7956f23a54b0. Accessed: 2022-09-16."},{"key":"e_1_3_2_2_70_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445896"},{"key":"e_1_3_2_2_71_1","doi-asserted-by":"publisher","DOI":"10.1145\/3411764.3445518"},{"key":"e_1_3_2_2_72_1","volume-title":"Schmidt and Sebastian Schmieg","author":"A.","year":"2022","unstructured":"Florian\u00a0A. Schmidt and Sebastian Schmieg. 2022. Prompt Battle. https:\/\/promptbattle.com\/. Accessed: 2022-10-22."},{"key":"e_1_3_2_2_73_1","volume-title":"Emilio Garcia, and Gurleen Virk.","author":"Shelby Renee","year":"2022","unstructured":"Renee Shelby, Shalaleh Rismani, Kathryn Henne, Ajung Moon, Negar Rostamzadeh, Paul Nicholas, N\u2019mah Yilla, Jess Gallegos, Andrew Smart, Emilio Garcia, and Gurleen Virk. 2022. Identifying Sociotechnical Harms of Algorithmic Systems: Scoping a Taxonomy for Harm Reduction. (Oct. 2022). arxiv:2210.05791\u00a0[cs.HC]"},{"key":"e_1_3_2_2_74_1","volume-title":"Inseok Jang, and Jinkyun Park.","author":"Shin Sung-Min","year":"2021","unstructured":"Sung-Min Shin, Sang\u00a0Hun Lee, Seung K\u00a0I Shin, Inseok Jang, and Jinkyun Park. 2021. STPA-Based Hazard and Importance Analysis on NPP Safety I&C Systems Focusing on Human\u2013System Interactions. Reliab. Eng. Syst. Saf. 213 (Sept. 2021), 107698."},{"key":"e_1_3_2_2_75_1","volume-title":"The Biased Artist: Exploiting Cultural Biases via Homoglyphs in Text-Guided Image Generation Models. (Sept","author":"Struppek Lukas","year":"2022","unstructured":"Lukas Struppek, Dominik Hintersdorf, and Kristian Kersting. 2022. The Biased Artist: Exploiting Cultural Biases via Homoglyphs in Text-Guided Image Generation Models. (Sept. 2022). arxiv:2209.08891\u00a0[cs.CV]"},{"key":"e_1_3_2_2_76_1","volume-title":"Toward Human-Centered Responsible Artificial Intelligence: A Review of CHI Research and Industry Toolkits. (Feb","author":"Tahaei Mohammad","year":"2023","unstructured":"Mohammad Tahaei, Marios Constantinides, and Daniele Quercia. 2023. Toward Human-Centered Responsible Artificial Intelligence: A Review of CHI Research and Industry Toolkits. (Feb. 2023). arxiv:2302.05284\u00a0[cs.HC]"},{"key":"e_1_3_2_2_77_1","volume-title":"Manifestations of Xenophobia in AI Systems. (Dec","author":"Tomasev Nenad","year":"2022","unstructured":"Nenad Tomasev, Jonathan\u00a0Leader Maynard, and Iason Gabriel. 2022. Manifestations of Xenophobia in AI Systems. (Dec. 2022). arxiv:2212.07877\u00a0[cs.CY]"},{"key":"e_1_3_2_2_78_1","volume-title":"Measuring Representational Harms in Image Captioning. In 2022 ACM Conference on Fairness, Accountability, and Transparency","author":"Wang Angelina","year":"2022","unstructured":"Angelina Wang, Solon Barocas, Kristen Laird, and Hanna Wallach. 2022. Measuring Representational Harms in Image Captioning. In 2022 ACM Conference on Fairness, Accountability, and Transparency (Seoul, Republic of Korea) (FAccT \u201922). Association for Computing Machinery, New York, NY, USA, 324\u2013335."},{"key":"e_1_3_2_2_79_1","doi-asserted-by":"publisher","DOI":"10.1145\/3512977"},{"key":"e_1_3_2_2_80_1","volume-title":"RePrompt: Automatic Prompt Editing to Refine AI-Generative Art Towards Precise Expressions. (Feb","author":"Wang Yunlong","year":"2023","unstructured":"Yunlong Wang, Shuyuan Shen, and Brian\u00a0Y Lim. 2023. RePrompt: Automatic Prompt Editing to Refine AI-Generative Art Towards Precise Expressions. (Feb. 2023). arxiv:2302.09466\u00a0[cs.HC]"},{"key":"e_1_3_2_2_81_1","doi-asserted-by":"publisher","DOI":"10.1145\/3531146.3533088"},{"key":"e_1_3_2_2_82_1","volume-title":"Notions of Responsibility. (Sept.","author":"Widder David\u00a0Gray","year":"2022","unstructured":"David\u00a0Gray Widder and Dawn Nafus. 2022. Dislocated Accountabilities in the AI Supply Chain: Modularity and Developers\u2019 Notions of Responsibility. (Sept. 2022). arxiv:2209.09780\u00a0[cs.CY]"},{"key":"e_1_3_2_2_83_1","doi-asserted-by":"publisher","DOI":"10.1145\/3448016.3457566"},{"key":"e_1_3_2_2_84_1","volume-title":"Scaling Autoregressive Models for Content-Rich Text-to-Image Generation. (June","author":"Yu Jiahui","year":"2022","unstructured":"Jiahui Yu, Yuanzhong Xu, Jing\u00a0Yu Koh, Thang Luong, Gunjan Baid, Zirui Wang, Vijay Vasudevan, Alexander Ku, Yinfei Yang, Burcu\u00a0Karagol Ayan, Ben Hutchinson, Wei Han, Zarana Parekh, Xin Li, Han Zhang, Jason Baldridge, and Yonghui Wu. 2022. Scaling Autoregressive Models for Content-Rich Text-to-Image Generation. (June 2022). arxiv:2206.10789\u00a0[cs.CV]"},{"key":"e_1_3_2_2_85_1","unstructured":"Zack Zwiezen. 2022. Rick And Morty Creator Used Controversial AI Art Voice Acting In New Shooter. https:\/\/kotaku.com\/high-on-life-justin-roiland-ai-art-rick-morty-1849900835. Accessed: 2023-1-27."}],"event":{"name":"AIES '23: AAAI\/ACM Conference on AI, Ethics, and Society","location":"Montr\u00e9al QC Canada","acronym":"AIES '23","sponsor":["SIGAI ACM Special Interest Group on Artificial Intelligence"]},"container-title":["Proceedings of the 2023 AAAI\/ACM Conference on AI, Ethics, and Society"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3600211.3604685","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3600211.3604685","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T16:37:39Z","timestamp":1750178259000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3600211.3604685"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,8,8]]},"references-count":80,"alternative-id":["10.1145\/3600211.3604685","10.1145\/3600211"],"URL":"https:\/\/doi.org\/10.1145\/3600211.3604685","relation":{},"subject":[],"published":{"date-parts":[[2023,8,8]]},"assertion":[{"value":"2023-08-29","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}