{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,28]],"date-time":"2026-02-28T01:55:54Z","timestamp":1772243754221,"version":"3.50.1"},"reference-count":144,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2024,7,24]],"date-time":"2024-07-24T00:00:00Z","timestamp":1721779200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,7,24]],"date-time":"2024-07-24T00:00:00Z","timestamp":1721779200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["AI Ethics"],"published-print":{"date-parts":[[2025,4]]},"DOI":"10.1007\/s43681-024-00517-3","type":"journal-article","created":{"date-parts":[[2024,7,24]],"date-time":"2024-07-24T13:08:52Z","timestamp":1721826532000},"page":"1769-1786","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":12,"title":["Situating the social issues of image generation models in the model life cycle: a sociotechnical approach"],"prefix":"10.1007","volume":"5","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-4698-1479","authenticated-orcid":false,"given":"Amelia","family":"Katirai","sequence":"first","affiliation":[]},{"given":"Noa","family":"Garcia","sequence":"additional","affiliation":[]},{"given":"Kazuki","family":"Ide","sequence":"additional","affiliation":[]},{"given":"Yuta","family":"Nakashima","sequence":"additional","affiliation":[]},{"given":"Atsuo","family":"Kishimoto","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,7,24]]},"reference":[{"key":"517_CR1","unstructured":"Ramesh, A., Dhariwal, P., Nichol, A., Chu, C., Chen, M.: Hierarchical text-conditional image generation with CLIP latents. (2022). ArXiv Prepr ArXiv220406125"},{"key":"517_CR2","unstructured":"Dayma, B., Patil, S., Cuenca, P., Saifullah, K., Abraham, T., Le Khac, P., et al.: Dall\u00b7 e mini. HuggingFacecom Httpshuggingface Cospacesdalleminidalle-Mini Accessed Sep 29. 2021; (2022)"},{"key":"517_CR3","doi-asserted-by":"crossref","unstructured":"Rombach, R., Blattmann, A., Lorenz, D., Esser, P., Ommer, B.: High-resolution image synthesis with latent diffusion models. In: CVPR. pp. 10684\u201395. (2022)","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"517_CR4","unstructured":"Hong, W., Ding, M., Zheng, W., Liu, X., Tang, J., CogVideo: Large-scale pretraining for text-to-video generation via transformers. In: ICLR (2023)"},{"key":"517_CR5","unstructured":"Singer, U., Polyak, A., Hayes, T., Yin, X., An, J., Zhang, S., et al.: Make-A-Video: Text-to-video generation without text-video data. In: ICLR (2022)"},{"key":"517_CR6","unstructured":"Ho, J., Chan, W., Saharia, C., Whang, J., Gao, R., Gritsenko, A., et al.: Imagen video: High definition video generation with diffusion models. ArXiv Prepr ArXiv221002303; (2022)"},{"key":"517_CR7","doi-asserted-by":"crossref","unstructured":"Dehouche, N., Dehouche, K.: What\u2019s in a text-to-image prompt? The potential of stable diffusion in visual arts education. Heliyon; (2023)","DOI":"10.1016\/j.heliyon.2023.e16757"},{"issue":"11","key":"517_CR8","doi-asserted-by":"publisher","first-page":"139","DOI":"10.1145\/3422622","volume":"63","author":"I Goodfellow","year":"2020","unstructured":"Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., et al.: Generative adversarial networks. Commun. ACM. 63(11), 139\u2013144 (2020)","journal-title":"Commun. ACM"},{"key":"517_CR9","unstructured":"Radford, A., Kim, J.W., Hallacy, C., Ramesh, A., Goh, G., Agarwal, S., et al.: Learning transferable visual models from natural language supervision. In: ICML (2021)"},{"key":"517_CR10","unstructured":"Bendel, O.: Image synthesis from an ethical perspective. AI Soc.;1\u201310. (2023)"},{"key":"517_CR11","doi-asserted-by":"crossref","unstructured":"Bird, C., Ungless, E., Kasirzadeh, A.: Typology of risks of generative text-to-image models. In: Proceedings of the 2023 AAAI\/ACM Conference on AI, Ethics, and Society. pp. 396\u2013410. (2023)","DOI":"10.1145\/3600211.3604722"},{"key":"517_CR12","doi-asserted-by":"publisher","unstructured":"Weidinger, L., Uesato, J., Rauh, M., Griffin, C., Huang, P.S., Mellor, J., et al.: Taxonomy of Risks posed by Language Models. In: 2022 ACM Conference on Fairness, Accountability, and Transparency [Internet]. Seoul Republic of Korea: ACM; 2022 [cited 2023 Mar 20]. pp. 214\u201329. https:\/\/doi.org\/10.1145\/3531146.3533088","DOI":"10.1145\/3531146.3533088"},{"key":"517_CR13","first-page":"237802312199958","volume":"7","author":"K Joyce","year":"2021","unstructured":"Joyce, K., Smith-Doerr, L., Alegria, S., Bell, S., Cruz, T., Hoffman, S.G., et al.: Toward a sociology of Artificial Intelligence: A call for Research on inequalities and Structural Change. Socius Sociol. Res. Dyn. World. 7, 237802312199958 (2021)","journal-title":"Socius Sociol. Res. Dyn. World"},{"key":"517_CR14","unstructured":"Weidinger, L., Rauh, M., Marchal, N., Manzini, A., Hendricks, L.A., Mateos-Garcia, J., et al.: Sociotechnical Safety Evaluation of Generative AI Systems [Internet]. arXiv; [cited 2024 Mar 27]. (2023). http:\/\/arxiv.org\/abs\/2310.11986"},{"issue":"4","key":"517_CR15","doi-asserted-by":"publisher","first-page":"799","DOI":"10.1109\/TAI.2022.3194503","volume":"4","author":"C Huang","year":"2022","unstructured":"Huang, C., Zhang, Z., Mao, B., Yao, X.: An overview of artificial intelligence ethics. IEEE Trans. Artif. Intell. 4(4), 799\u2013819 (2022)","journal-title":"IEEE Trans. Artif. Intell."},{"issue":"3","key":"517_CR16","doi-asserted-by":"publisher","first-page":"213","DOI":"10.1007\/s43681-021-00043-6","volume":"1","author":"A van Wynsberghe","year":"2021","unstructured":"van Wynsberghe, A., Sustainable, A.I.: AI for sustainability and the sustainability of AI. AI Ethics. 1(3), 213\u2013218 (2021)","journal-title":"AI Ethics"},{"issue":"4","key":"517_CR17","doi-asserted-by":"publisher","first-page":"1105","DOI":"10.1007\/s00146-021-01301-1","volume":"36","author":"K Crawford","year":"2021","unstructured":"Crawford, K., Paglen, T., Excavating, A.I.: The politics of images in machine learning training sets. AI Soc. 36(4), 1105\u20131116 (2021)","journal-title":"AI Soc."},{"key":"517_CR18","doi-asserted-by":"publisher","DOI":"10.2307\/j.ctv1ghv45t","volume-title":"Atlas of AI: Power, Politics, and the Planetary Costs of Artificial Intelligence","author":"K Crawford","year":"2021","unstructured":"Crawford, K.: Atlas of AI: Power, Politics, and the Planetary Costs of Artificial Intelligence. Yale University Press, New Haven (2021)"},{"key":"517_CR19","unstructured":"Benjamin, R.: Race after Technology: Abolitionist Tools for the new Jim code. Oxford University Press (2020)"},{"key":"517_CR20","doi-asserted-by":"crossref","unstructured":"Gebru, T.: Race and Gender. In: Dubber MD, Pasquale F, Das S, editors. The Oxford Handbook of Ethics of AI [Internet]. 2020 [cited 2022 Nov 4]. pp. 252\u201369. https:\/\/academic.oup.com\/edited-volume\/34287\/chapter-abstract\/290662826?redirectedFrom=fulltext","DOI":"10.1093\/oxfordhb\/9780190067397.013.16"},{"key":"517_CR21","unstructured":"Anderson, J., Laracy, F.J.R., Marlowe, T.A., Multi-Disciplinary: Anal. Catholic Social Teach. Implications Eng. Technol.;18(6). (2020)"},{"key":"517_CR22","doi-asserted-by":"crossref","unstructured":"McStay, A., Emotional, A.I., Ethics, Spice, J.: Contributing community, wholeness, sincerity, and heart. Philos. Technol. 34(4), 1781\u20131802 (2021)","DOI":"10.1007\/s13347-021-00487-y"},{"key":"517_CR23","doi-asserted-by":"publisher","DOI":"10.7551\/mitpress\/12549.001.0001","volume-title":"AI Ethics","author":"M Coeckelbergh","year":"2020","unstructured":"Coeckelbergh, M.: AI Ethics. The MIT Press, Cambridge (2020)"},{"issue":"4","key":"517_CR24","first-page":"42","volume":"21","author":"VD KIROVA","year":"2023","unstructured":"KIROVA, V.D., Ku, C., Laracy, J., Marlowe, T.: The Ethics of Artificial Intelligence in the era of generative AI. J. Syst. Cybern Inf. 21(4), 42\u201350 (2023)","journal-title":"J. Syst. Cybern Inf."},{"key":"517_CR25","unstructured":"Mansimov, E., Parisotto, E., Ba, J.L., Salakhutdinov, R.: Generating images from captions with attention. In: ICML. (2016)"},{"key":"517_CR26","unstructured":"Ramesh, A., Pavlov, M., Goh, G., Gray, S., Voss, C., Radford, A., et al.: Zero-shot text-to-image Generation, pp. 8821\u20138831. In: ICML. PMLR (2021)"},{"key":"517_CR27","unstructured":"Ordonez, V., Kulkarni, G., Berg, T.: Im2text: Describing images using 1 million captioned photographs. Adv. Neural Inf. Process. Syst.;24. (2011)"},{"key":"517_CR28","doi-asserted-by":"publisher","first-page":"67","DOI":"10.1162\/tacl_a_00166","volume":"2","author":"P Young","year":"2014","unstructured":"Young, P., Lai, A., Hodosh, M., Hockenmaier, J.: From image descriptions to visual denotations: New similarity metrics for semantic inference over event descriptions. Trans. Assoc. Comput. Linguist. 2, 67\u201378 (2014)","journal-title":"Trans. Assoc. Comput. Linguist"},{"key":"517_CR29","unstructured":"Chen, X., Fang, H., Lin, T.Y., Vedantam, R., Gupta, S., Doll\u00e1r, P., et al.: Microsoft COCO captions: Data collection and evaluation server. (2015). ArXiv Prepr ArXiv150400325"},{"key":"517_CR30","doi-asserted-by":"publisher","first-page":"32","DOI":"10.1007\/s11263-016-0981-7","volume":"123","author":"R Krishna","year":"2017","unstructured":"Krishna, R., Zhu, Y., Groth, O., Johnson, J., Hata, K., Kravitz, J., et al.: Visual genome: Connecting language and vision using crowdsourced dense image annotations. Int. J. Comput. Vis. 123, 32\u201373 (2017)","journal-title":"Int. J. Comput. Vis."},{"key":"517_CR31","doi-asserted-by":"crossref","unstructured":"Sharma, P., Ding, N., Goodman, S., Soricut, R., Conceptual Captions: A Cleaned, Hypernymed, Image Alt-text Dataset For Automatic Image Captioning. In: Proceedings of ACL. (2018)","DOI":"10.18653\/v1\/P18-1238"},{"key":"517_CR32","doi-asserted-by":"crossref","unstructured":"Agrawal, H., Desai, K., Wang, Y., Chen, X., Jain, R., Johnson, M., et al.: Nocaps: Novel object captioning at scale. In: Proceedings of the IEEE\/CVF international conference on computer vision. pp. 8948\u201357. (2019)","DOI":"10.1109\/ICCV.2019.00904"},{"key":"517_CR33","doi-asserted-by":"crossref","unstructured":"Sidorov, O., Hu, R., Rohrbach, M., Singh, A.: Textcaps: a dataset for image captioning with reading comprehension. In: Computer Vision\u2013ECCV.: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part II 16. Springer; 2020. pp. 742\u201358. (2020)","DOI":"10.1007\/978-3-030-58536-5_44"},{"key":"517_CR34","doi-asserted-by":"crossref","unstructured":"Changpinyo, S., Sharma, P., Ding, N., Soricut, R.: Conceptual 12m: Pushing web-scale image-text pre-training to recognize long-tail visual concepts. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. pp. 3558\u201368. (2021)","DOI":"10.1109\/CVPR46437.2021.00356"},{"key":"517_CR35","doi-asserted-by":"crossref","unstructured":"Srinivasan, K., Raman, K., Chen, J., Bendersky, M., Najork, M., Wit: Wikipedia-based image text dataset for multimodal multilingual machine learning. In: Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval. pp. 2443\u20139. (2021)","DOI":"10.1145\/3404835.3463257"},{"key":"517_CR36","unstructured":"Desai, K., Kaul, G., Aysola, Z., Johnson, J., RedCaps: Web-curated image-text data created by the people, for the people. ArXiv Prepr ArXiv211111431; (2021)"},{"key":"517_CR37","unstructured":"Schuhmann, C., Kaczmarczyk, R., Komatsuzaki, A., Katta, A., Vencu, R., Beaumont, R., et al.: NeurIPS Workshop Datacentric AI. J\u00fclich Supercomputing Center (2021). LAION-400\u00a0M: Open Dataset of CLIP-Filtered 400 Million Image-Text Pairs"},{"key":"517_CR38","unstructured":"Schuhmann, C., Beaumont, R., Vencu, R., Gordon, C., Wightman, R., Cherti, M., et al.: LAION-5B: An open large-scale dataset for training next generation image-text models. ArXiv Prepr ArXiv221008402; (2022)"},{"key":"517_CR39","doi-asserted-by":"crossref","unstructured":"Nilsback, M.E., Zisserman, A.: Automated flower classification over a large number of classes. In: Indian Conference on Computer Vision, Graphics & Image Processing. IEEE; pp. 722\u20139. (2008)","DOI":"10.1109\/ICVGIP.2008.47"},{"key":"517_CR40","unstructured":"Wah, C., Branson, S., Welinder, P., Perona, P., Belongie, S.: The Caltech-UCSD Birds-200-2011 Dataset. California Institute of Technology (2011). Report No.: CNS-TR-2011-001"},{"key":"517_CR41","doi-asserted-by":"crossref","unstructured":"Lin, T.Y., Maire, M., Belongie, S., Hays, J., Perona, P., Ramanan, D., et al.: Microsoft COCO: Common objects in context. In: ECCV. Springer; pp. 740\u201355. (2014)","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"517_CR42","unstructured":"Reed, S., Akata, Z., Yan, X., Logeswaran, L., Schiele, B., Lee, H.: Generative Adversarial text to Image Synthesis, pp. 1060\u20131069. In: ICML. PMLR (2016)"},{"key":"517_CR43","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., et al.: Attention is all you need. In: NeurIPS (2017)"},{"key":"517_CR44","unstructured":"Kingma, D.P., Welling, M.: Auto-encoding variational bayes. ICLR; (2014)"},{"key":"517_CR45","unstructured":"Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., et al.: Generative adversarial nets. In: NeurIPS (2014)"},{"key":"517_CR46","unstructured":"Sohl-Dickstein, J., Weiss, E., Maheswaranathan, N., Ganguli, S.: Deep unsupervised learning using nonequilibrium thermodynamics. In: ICML p. 2256\u20132265. (2015)"},{"key":"517_CR47","unstructured":"Song, Y., Ermon, S.: Generative modeling by estimating gradients of the data distribution. In: NeurIPS (2019)"},{"key":"517_CR48","unstructured":"Ho, J., Jain, A., Abbeel, P.: Denoising diffusion probabilistic models. In: NeurIPS p. 6840\u20136851. (2020)"},{"key":"517_CR49","unstructured":"Xu, J., Liu, X., Wu, Y., Tong, Y., Li, Q., Ding, M., et al.: ImageReward: Learning and evaluating human preferences for text-to-image generation. (2023). ArXiv Prepr ArXiv230405977"},{"key":"517_CR50","unstructured":"Hugging Face: Stable Diffusion text-to-image fine-tuning [Internet]. [cited 2024 May 27]. https:\/\/huggingface.co\/docs\/diffusers\/v0.13.0\/en\/training\/text2image"},{"key":"517_CR51","unstructured":"Zhang, C., Zhang, C., Zheng, S., Qiao, Y., Li, C., Zhang, M., et al.: A complete survey on generative AI (AIGC): Is ChatGPT from GPT-4 to GPT-5 all you need? (2023). ArXiv Prepr ArXiv230311717"},{"key":"517_CR52","doi-asserted-by":"crossref","unstructured":"Ko, H.K., Park, G., Jeon, H., Jo, J., Kim, J., Seo, J.: Large-scale text-to-image generation models for visual artists\u2019 creative works. In: IUI. pp. 919\u201333. (2023)","DOI":"10.1145\/3581641.3584078"},{"key":"517_CR53","doi-asserted-by":"crossref","unstructured":"Bird, C., Ungless, E.L., Kasirzadeh, A.: Typology of Risks of Generative Text-to-Image Models. (2023)","DOI":"10.1145\/3600211.3604722"},{"key":"517_CR54","doi-asserted-by":"publisher","unstructured":"Bender, E.M., Gebru, T., McMillan-Major, A., Shmitchell, S.: On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? \u5217. In: Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency [Internet]. Virtual Event Canada: ACM; 2021 [cited 2023 Mar 31]. pp. 610\u201323. https:\/\/doi.org\/10.1145\/3442188.3445922","DOI":"10.1145\/3442188.3445922"},{"issue":"2","key":"517_CR55","doi-asserted-by":"publisher","first-page":"77","DOI":"10.1191\/1478088706qp063oa","volume":"3","author":"V Braun","year":"2006","unstructured":"Braun, V., Clarke, V.: Using thematic analysis in psychology. Qual. Res. Psychol. 3(2), 77\u2013101 (2006)","journal-title":"Qual. Res. Psychol."},{"issue":"11","key":"517_CR56","doi-asserted-by":"publisher","first-page":"100336","DOI":"10.1016\/j.patter.2021.100336","volume":"2","author":"A Paullada","year":"2021","unstructured":"Paullada, A., Raji, I.D., Bender, E.M., Denton, E., Hanna, A.: Data and its (dis)contents: A survey of dataset development and use in machine learning research. Patterns. 2(11), 100336 (2021)","journal-title":"Patterns"},{"issue":"12","key":"517_CR57","doi-asserted-by":"publisher","first-page":"86","DOI":"10.1145\/3458723","volume":"64","author":"T Gebru","year":"2021","unstructured":"Gebru, T., Morgenstern, J., Vecchione, B., Vaughan, J.W., Wallach, H., Iii, H.D., et al.: Datasheets for datasets. Commun. ACM. 64(12), 86\u201392 (2021)","journal-title":"Commun. ACM"},{"key":"517_CR58","unstructured":"Wu, Y., Yu, N., Li, Z., Backes, M., Zhang, Y.: Membership Inference Attacks Against Text-to-image Generation Models. ArXiv Prepr ArXiv221000968.; (2022)"},{"key":"517_CR59","doi-asserted-by":"crossref","unstructured":"Schramowski, P., Brack, M., Deiseroth, B., Kersting, K.: Safe latent diffusion: Mitigating Inappropriate Degeneration in Diffusion models. (2023). ArXiv Prepr ArXiv221105105","DOI":"10.1109\/CVPR52729.2023.02157"},{"key":"517_CR60","doi-asserted-by":"crossref","unstructured":"Monea, A.: The Digital Closet. MIT Press (2022)","DOI":"10.7551\/mitpress\/12551.001.0001"},{"key":"517_CR61","doi-asserted-by":"crossref","unstructured":"Ungless, E.L., Ross, B., Lauscher, A.: Stereotypes and Smut: The (Mis) representation of Non-cisgender Identities by Text-to-Image Models. ArXiv Prepr ArXiv230517072.; (2023)","DOI":"10.18653\/v1\/2023.findings-acl.502"},{"key":"517_CR62","unstructured":"Nichol, A.Q., Dhariwal, P., Ramesh, A., Shyam, P., Mishkin, P., Mcgrew, B., et al.: GLIDE: Towards Photorealistic Image Generation and Editing with Text-Guided Diffusion Models, pp. 16784\u201316804. In: ICML. PMLR (2022)"},{"key":"517_CR63","unstructured":"Gray, M.L., Suri, S.: Ghost work: How to stop Silicon Valley from Building a new Global Underclass. Eamon Dolan Books (2019)"},{"key":"517_CR64","unstructured":"Perrigo, B.: Exclusive: OpenAI used Kenyan workers on Less Than 2 per hour to make ChatGPT less toxic. Vice (2023)"},{"key":"517_CR65","unstructured":"Vyas, N., Kakade, S., Barak, B.: Provable copyright protection for generative models. ArXiv Prepr ArXiv230210870; (2023)"},{"key":"517_CR66","unstructured":"Ghosh, A., Fossas, G.: Can There be Art Without an Artist? In: NeurIPS Workshop: Human Evaluation of Generative Models. (2022)"},{"key":"517_CR67","doi-asserted-by":"crossref","unstructured":"Zhang, E., Wang, K., Xu, X., Wang, Z., Shi, H.: Forget-Me-Not: Learning to forget in text-to-image Diffusion models. (2023). ArXiv Prepr ArXiv230317591","DOI":"10.1109\/CVPRW63382.2024.00182"},{"key":"517_CR68","unstructured":"Nichol, A.: DALL-E 2 pre-training mitigations. OpenAI Blog (2022)"},{"key":"517_CR69","unstructured":"Vincent, J.: Getty Images sues AI art generator Stable Diffusion in the US for copyright infringement. The Verge. The Verge. (2023)"},{"key":"517_CR70","unstructured":"Cetinic, E.: The Myth of Culturally Agnostic AI Models. ArXiv Prepr ArXiv221115271.; (2022)"},{"key":"#cr-split#-517_CR71.1","unstructured":"Weatherbed, J.: How to keep your art out of AI generators. The Verge [Internet]. Feb 8 [cited 2024 May 27]"},{"key":"#cr-split#-517_CR71.2","unstructured":"(2024). https:\/\/www.theverge.com\/24063327\/ai-art-protect-images-copyright-generators"},{"key":"517_CR72","doi-asserted-by":"crossref","unstructured":"Shan, S., Ding, W., Passananti, J., Wu, S., Zheng, H., Zhao, B.Y., Nightshade: Prompt-Specific Poisoning Attacks on Text-to-Image Generative Models. In: 2024 IEEE Symposium on Security and Privacy (SP). IEEE Computer Society; pp. 212\u2013212. (2024)","DOI":"10.1109\/SP54263.2024.00207"},{"key":"517_CR73","unstructured":"United States Copyright Office: Copyrightable Authorship: What Can Be Registered. (2021)"},{"key":"517_CR74","doi-asserted-by":"crossref","unstructured":"Chesterman, S.: AI-generated content is taking over the world. But who owns it? NUS Law Work Pap No 2023002.; (2023)","DOI":"10.2139\/ssrn.4321596"},{"key":"517_CR75","doi-asserted-by":"crossref","unstructured":"Fenwick, M., Jurcys, P.: Originality and the future of copyright in an age of generative AI. Available SSRN 4354449.; (2023)","DOI":"10.1016\/j.clsr.2023.105892"},{"key":"517_CR76","unstructured":"Guadamuz, A.: Do androids dream of electric copyright? Comparative analysis of originality in artificial intelligence generated works. Intellect. Prop. Q.; (2017)"},{"key":"517_CR77","unstructured":"Shen, X., Qu, Y., Backes, M., Zhang, Y.: Prompt Stealing Attacks Against Text-to-Image Generation Models. ArXiv Prepr ArXiv230209923.; (2023)"},{"key":"517_CR78","unstructured":"Gervais, D.J.: The machine as author. Iowa Rev.;105:2053. (2019)"},{"key":"517_CR79","doi-asserted-by":"crossref","unstructured":"Epstein, Z., Levine, S., Rand, D.G., Rahwan, I.: Who gets credit for AI-Generated art? iScience.;23(9):101515. (2020)","DOI":"10.1016\/j.isci.2020.101515"},{"key":"517_CR80","unstructured":"Luccioni, A.S., Akiki, C., Mitchell, M., Jernite, Y.: Stable Bias: Analyzing Societal representations in Diffusion models. (2023). ArXiv Prepr ArXiv230311408"},{"key":"517_CR81","doi-asserted-by":"crossref","unstructured":"Cheong, M., Abedin, E., Ferreira, M., Reimann, R.W., Chalson, S., Robinson, P., et al.: Investigating gender and racial biases in DALL-E Mini images. PNAS; (2023)","DOI":"10.1145\/3649883"},{"key":"517_CR82","doi-asserted-by":"crossref","unstructured":"Bianchi, F., Kalluri, P., Durmus, E., Ladhak, F., Cheng, M., Nozza, D., et al.: Easily accessible text-to-image generation amplifies demographic stereotypes at large scale. (2022). ArXiv Prepr ArXiv221103759","DOI":"10.1145\/3593013.3594095"},{"key":"517_CR83","unstructured":"Fraser, K.C., Kiritchenko, S., Nejadgholi, I.: A Friendly Face: Do text-to-Image systems Rely on stereotypes when the Input is Under-Specified? (2023). ArXiv Prepr ArXiv230207159"},{"key":"517_CR84","doi-asserted-by":"publisher","unstructured":"Steed, R., Caliskan, A.: Image Representations Learned With Unsupervised Pre-Training Contain Human-like Biases. In: Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency [Internet]. Virtual Event Canada: ACM; 2021 [cited 2023 Apr 7]. pp. 701\u201313. https:\/\/doi.org\/10.1145\/3442188.3445932","DOI":"10.1145\/3442188.3445932"},{"key":"517_CR85","doi-asserted-by":"crossref","unstructured":"Struppek, L., Hintersdorf, D., Kersting, K.: The biased artist: Exploiting cultural biases via homoglyphs in text-guided image generation models. ArXiv Prepr ArXiv220908891; (2022)","DOI":"10.1613\/jair.1.15388"},{"key":"517_CR86","unstructured":"Seshadri, P., Singh, S., Elazar, Y.: The Bias amplification Paradox in text-to-image generation. (2023). ArXiv Prepr ArXiv230800755"},{"key":"517_CR87","doi-asserted-by":"crossref","unstructured":"Cho, J., Zala, A., Bansal, M.: DALL-Eval: Probing the reasoning skills and social biases of text-to-image generative transformers. ArXiv Prepr ArXiv220204053; (2022)","DOI":"10.1109\/ICCV51070.2023.00283"},{"key":"517_CR88","doi-asserted-by":"crossref","unstructured":"Bansal, H., Yin, D., Monajatipoor, M., Chang, K.W.: How well can Text-to-Image Generative Models understand Ethical Natural Language Interventions? ArXiv Prepr ArXiv221015230.; (2022)","DOI":"10.18653\/v1\/2022.emnlp-main.88"},{"key":"517_CR89","doi-asserted-by":"publisher","first-page":"e43110","DOI":"10.2196\/43110","volume":"25","author":"LC Adams","year":"2023","unstructured":"Adams, L.C., Busch, F., Truhn, D., Makowski, M.R., Aerts, H.J., Bressem, K.K.: What does DALL-E 2 Know about Radiology? J. Med. Internet Res. 25, e43110 (2023)","journal-title":"J. Med. Internet Res."},{"key":"517_CR90","unstructured":"Fraser, K.C., Kiritchenko, S., Nejadgholi, I.: Diversity is not a one-way street: Pilot study on ethical interventions for racial bias in text-to-image systems. ICCV Accept; (2023)"},{"key":"517_CR91","unstructured":"Munn, L., Magee, L., Arora, V., Unmaking, A.I., Imagemaking: A methodological toolkit for critical investigation. (2023). ArXiv Prepr ArXiv230709753"},{"key":"517_CR92","doi-asserted-by":"crossref","unstructured":"Hutchinson, B., Baldridge, J., Prabhakaran, V.: Underspecification in Scene Description-to-Depiction Tasks. In: Proceedings of the 2nd Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 12th International Joint Conference on Natural Language Processing (Volume 1: Long Papers) [Internet]. Online only: Association for Computational Linguistics; pp. 1172\u201384. (2022). https:\/\/aclanthology.org\/2022.aacl-main.86","DOI":"10.18653\/v1\/2022.aacl-main.86"},{"key":"517_CR93","unstructured":"Offert, F., Phan, T.A., Sign That, Spells: DALL-E 2, Invisual images and the racial politics of Feature Space. (2022). ArXiv Prepr ArXiv221106323"},{"key":"517_CR94","unstructured":"Andrews, J.T., Zhao, D., Thong, W., Modas, A., Papakyriakopoulos, O., Nagpal, S., et al.: Ethical considerations for collecting human-centric image datasets. ArXiv Prepr ArXiv230203629; (2023)"},{"key":"517_CR95","doi-asserted-by":"crossref","unstructured":"Wolfe, R., Caliskan, A.: American\u2009=\u2009=\u2009white in multimodal language-and-image AI. In: AIES. pp. 800\u201312. (2022)","DOI":"10.1145\/3514094.3534136"},{"key":"517_CR96","doi-asserted-by":"crossref","unstructured":"Hirota, Y., Nakashima, Y., Garcia, N.: Quantifying societal bias amplification in image captioning. In: CVPR p. 13450\u201313459. (2022)","DOI":"10.1109\/CVPR52688.2022.01309"},{"key":"517_CR97","doi-asserted-by":"publisher","DOI":"10.7312\/katz19490","volume-title":"Artificial Whiteness: Politics and Ideology in Artificial Intelligence","author":"Y Katz","year":"2020","unstructured":"Katz, Y.: Artificial Whiteness: Politics and Ideology in Artificial Intelligence. Columbia University, New York (2020)"},{"issue":"4","key":"517_CR98","doi-asserted-by":"publisher","first-page":"685","DOI":"10.1007\/s13347-020-00415-6","volume":"33","author":"S Cave","year":"2020","unstructured":"Cave, S., Dihal, K.: The whiteness of AI. Philos. Technol. 33(4), 685\u2013703 (2020)","journal-title":"Philos. Technol."},{"key":"517_CR99","unstructured":"Carlini, N., Hayes, J., Nasr, M., Jagielski, M., Sehwag, V., Tramer, F., et al.: Extracting training data from diffusion models. ArXiv Prepr ArXiv230113188; (2023)"},{"key":"517_CR100","doi-asserted-by":"crossref","unstructured":"Struppek, L., Hintersdorf, D., Kersting, K.: Rickrolling the artist: Injecting Invisible backdoors into text-guided image generation models. (2022). ArXiv Prepr ArXiv221102408","DOI":"10.1109\/ICCV51070.2023.00423"},{"key":"517_CR101","doi-asserted-by":"crossref","unstructured":"Van Le, T., Phung, H., Nguyen, T.H., Dao, Q., Tran, N., Tran, A.: Anti-DreamBooth: Protecting users from personalized text-to-image synthesis. ArXiv Prepr ArXiv230315433; (2023)","DOI":"10.1109\/ICCV51070.2023.00202"},{"issue":"3","key":"517_CR102","doi-asserted-by":"publisher","first-page":"50","DOI":"10.1145\/3313129","volume":"25","author":"L Stark","year":"2019","unstructured":"Stark, L.: Facial recognition is the plutonium of AI. XRDS Crossroads ACM Mag Stud. 25(3), 50\u201355 (2019)","journal-title":"XRDS Crossroads ACM Mag Stud."},{"key":"517_CR103","unstructured":"Chen, C., Fu, J., Lyu, L.: A pathway towards responsible AI generated content. (2023). ArXiv Prepr ArXiv230301325"},{"key":"517_CR104","unstructured":"OpenAI: GPT-4 System Card. (2023)"},{"key":"517_CR105","doi-asserted-by":"crossref","unstructured":"Sha, Z., Li, Z., Yu, N., Zhang, Y.: DE-FAKE: Detection and attribution of fake images generated by text-to-image generation models. (2023). ArXiv Prepr ArXiv221006998","DOI":"10.1145\/3576915.3616588"},{"key":"517_CR106","doi-asserted-by":"crossref","unstructured":"Ricker, J., Damm, S., Holz, T., Fischer, A.: Towards the detection of Diffusion Model deepfakes. (2023). ArXiv Prepr ArXiv221014571","DOI":"10.5220\/0012422000003660"},{"key":"517_CR107","first-page":"147","volume":"98","author":"R Chesney","year":"2019","unstructured":"Chesney, R., Citron, D.: Deepfakes and the new disinformation war: The coming age of post-truth geopolitics. Foreign Aff. 98, 147 (2019)","journal-title":"Foreign Aff"},{"key":"517_CR108","unstructured":"Oppenlaender, J., Visuri, A., Paananen, V., Linder, R., Silvennoinen, J.: Text-to-image generation: Perceptions and realities. (2023). ArXiv Prepr ArXiv230313530"},{"issue":"8","key":"517_CR109","doi-asserted-by":"publisher","first-page":"e2120481119","DOI":"10.1073\/pnas.2120481119","volume":"119","author":"SJ Nightingale","year":"2022","unstructured":"Nightingale, S.J., Farid, H.: AI-synthesized faces are indistinguishable from real faces and more trustworthy. Proc. Natl. Acad. Sci. 119(8), e2120481119 (2022)","journal-title":"Proc. Natl. Acad. Sci."},{"key":"517_CR110","unstructured":"Birhane, A., Prabhu, V.U., Kahembwe, E.: Multimodal datasets: misogyny, pornography, and malignant stereotypes [Internet]. arXiv; 2021 [cited 2023 Apr 10]. http:\/\/arxiv.org\/abs\/2110.01963"},{"key":"517_CR111","unstructured":"Internet Watch Foundation. How AI is being abused to create child sexual abuse imagery [Internet]: (2023). https:\/\/www.iwf.org.uk\/about-us\/why-we-exist\/our-research\/how-ai-is-being-abused-to-create-child-sexual-abuse-imagery\/"},{"issue":"5","key":"517_CR112","doi-asserted-by":"publisher","first-page":"910","DOI":"10.1109\/JSTSP.2020.3002101","volume":"14","author":"L Verdoliva","year":"2020","unstructured":"Verdoliva, L.: Media forensics and deepfakes: An overview. IEEE J. Sel. Top. Signal. Process. 14(5), 910\u2013932 (2020)","journal-title":"IEEE J. Sel. Top. Signal. Process."},{"key":"517_CR113","unstructured":"Agarwal, S., Farid, H., Gu, Y., He, M., Nagano, K., Li, H.: Protecting World Leaders Against Deep Fakes. In: CVPR Workshops. p. 38. (2019)"},{"key":"517_CR114","doi-asserted-by":"crossref","unstructured":"Qu, Y., Shen, X., He, X., Backes, M., Zannettou, S., Zhang, Y.: Unsafe diffusion: On the generation of unsafe images and hateful memes from text-to-image models. ArXiv Prepr ArXiv230513873; (2023)","DOI":"10.1145\/3576915.3616679"},{"key":"517_CR115","doi-asserted-by":"crossref","unstructured":"Srinivasan, R., Uchino, K.: Biases in generative art: A causal look from the lens of art history. In: FAccT p. 41\u201351. (2021)","DOI":"10.1145\/3442188.3445869"},{"key":"517_CR116","unstructured":"Shan, S., Cryan, J., Wenger, E., Zheng, H., Hanocka, R., Zhao, B.Y.: GLAZE: Protecting artists from style mimicry by text-to-image models. ArXiv Prepr ArXiv230204222; (2023)"},{"key":"517_CR117","doi-asserted-by":"crossref","unstructured":"Vimpari, V., Kultima, A., H\u00e4m\u00e4l\u00e4inen, P., Guckelsberger, C.: An adapt-or-die type of Situation: Perception, Adoption, and Use of Text-To-Image-generation AI by Game Industry professionals. (2023). ArXiv Prepr ArXiv230212601","DOI":"10.1145\/3611025"},{"key":"517_CR118","doi-asserted-by":"publisher","unstructured":"Oppenlaender, J.: The Creativity of Text-to-Image Generation. In: Proceedings of the 25th International Academic Mindtrek Conference [Internet]. Tampere Finland: ACM; 2022 [cited 2023 Apr 26]. pp. 192\u2013202. https:\/\/doi.org\/10.1145\/3569219.3569352","DOI":"10.1145\/3569219.3569352"},{"key":"517_CR119","unstructured":"Colton, S., Charnley, J., Pease, A.: Computational creativity theory: The FACE and IDEA descriptive models. In: Proc. Int. Conf. Comp. Creativity (2011)"},{"key":"517_CR120","unstructured":"Jordanous, A.: Evaluating computational creativity: a standardised procedure for evaluating creative systems and its application [PhD Thesis]. University of Sussex; (2013)"},{"key":"517_CR121","doi-asserted-by":"publisher","unstructured":"Ragot, M., Martin, N., Cojean, S.: AI-generated vs. Human Artworks. A Perception Bias Towards Artificial Intelligence? In: Extended Abstracts of the 2020 CHI Conference on Human Factors in Computing Systems [Internet]. Honolulu HI USA: ACM; 2020 [cited 2023 Apr 26]. pp. 1\u201310. https:\/\/doi.org\/10.1145\/3334480.3382892","DOI":"10.1145\/3334480.3382892"},{"key":"517_CR122","unstructured":"Seymour, T.: The camera never lies? Creator of AI image rejects prestigious photo award. The Art Newspaper [Internet]. 2023 [cited 2023 May 4]. Available from: The camera never lies? Creator of AI image rejects prestigious photo award"},{"key":"517_CR123","volume-title":"The Uninhabitable Earth","author":"D Wallace-Wells","year":"2019","unstructured":"Wallace-Wells, D.: The Uninhabitable Earth. Tim Duggan Books, New York (2019)"},{"issue":"37","key":"517_CR124","doi-asserted-by":"publisher","first-page":"eadh2458","DOI":"10.1126\/sciadv.adh2458","volume":"9","author":"K Richardson","year":"2023","unstructured":"Richardson, K., Steffen, W., Lucht, W., Bendtsen, J., Cornell, S.E., Donges, J.F., et al.: Earth beyond six of nine planetary boundaries. Sci. Adv. 9(37), eadh2458 (2023)","journal-title":"Sci. Adv."},{"issue":"8","key":"517_CR125","doi-asserted-by":"publisher","first-page":"423","DOI":"10.1038\/s42256-020-0219-9","volume":"2","author":"P Dhar","year":"2020","unstructured":"Dhar, P.: The carbon impact of artificial intelligence. Nat. Mach. Intell. 2(8), 423\u2013425 (2020)","journal-title":"Nat. Mach. Intell."},{"key":"517_CR126","volume-title":"Is AI good for the Planet?","author":"B Brevini","year":"2021","unstructured":"Brevini, B.: Is AI good for the Planet? Polity, Cambridge (2021)"},{"issue":"9","key":"517_CR127","doi-asserted-by":"publisher","first-page":"3464","DOI":"10.1021\/acs.est.3c01106","volume":"57","author":"MC Rillig","year":"2023","unstructured":"Rillig, M.C., \u00c5gerstrand, M., Bi, M., Gould, K.A., Sauerland, U.: Risks and benefits of large language models for the environment. Environ. Sci. Technol. 57(9), 3464\u20133466 (2023)","journal-title":"Environ. Sci. Technol."},{"key":"517_CR128","doi-asserted-by":"crossref","unstructured":"Luccioni, A.S., Jernite, Y., Strubell, E.: Power hungry processing: Watts driving the cost of ai deployment? ArXiv Prepr ArXiv231116863.; (2023)","DOI":"10.1145\/3630106.3658542"},{"key":"517_CR129","unstructured":"Amodei, D., Hernandez, D.: AI and compute. OpenAI Blog. (2018)"},{"key":"517_CR130","unstructured":"Lacoste, A., Luccioni, A., Schmidt, V., Dandres, T.: Quantifying the Carbon Emissions of Machine Learning. In: NeurIPS Workshop: Tackling Climate Change with Machine Learning. (2019)"},{"key":"517_CR131","unstructured":"Schwartz, R., Dodge, J., Smith, N.A., Etzioni, O., Green, A.I.: (2019). ArXiv Prepr ArXiv190710597"},{"key":"517_CR132","unstructured":"Li, P., Yang, J., Islam, M.A., Ren, S., Making, A.I.: Less thirsty: Uncovering and addressing the Secret Water Footprint of AI models. (2023). ArXiv Prepr ArXiv230403271"},{"issue":"12","key":"517_CR133","doi-asserted-by":"publisher","first-page":"e905","DOI":"10.1016\/S2542-5196(21)00263-1","volume":"5","author":"SM Parvez","year":"2021","unstructured":"Parvez, S.M., Jahan, F., Brune, M.N., Gorman, J.F., Rahman, M.J., Carpenter, D., et al.: Health consequences of exposure to e-waste: An updated systematic review. Lancet Planet. Health. 5(12), e905\u2013e920 (2021)","journal-title":"Lancet Planet. Health"},{"key":"517_CR134","unstructured":"Pitron, G.: The Rare Metals War: The dark side of Clean Energy and Digital Technologies. Scribe (2020)"},{"key":"517_CR135","unstructured":"The Economist: So long iPhone, Generative AI needs a new device. The Economist [Internet]. https:\/\/www.economist.com \u203a business \u203a 2023\/10\/05"},{"key":"517_CR136","unstructured":"ISO\/IEC 29134:2023 Information technology \u2014 Security techniques \u2014 Guidelines for privacy impact assessment [Internet]: (2023). https:\/\/www.iso.org\/standard\/86012.html"},{"key":"517_CR137","unstructured":"OpenAI: March 20 ChatGPT outage: Here\u2019s what happened. (2023)"},{"issue":"2","key":"517_CR138","doi-asserted-by":"publisher","first-page":"127","DOI":"10.1093\/scan\/nsn048","volume":"4","author":"MP Ewbank","year":"2009","unstructured":"Ewbank, M.P., Barnard, P.J., Croucher, C.J., Ramponi, C., Calder, A.J.: The amygdala response to images with impact. Soc. Cogn. Affect. Neurosci. 4(2), 127\u2013133 (2009)","journal-title":"Soc. Cogn. Affect. Neurosci."},{"key":"517_CR139","unstructured":"Birhane, A., Prabhu, V.U., Kahembwe, E.: Multimodal datasets: Misogyny, pornography, and malignant stereotypes. ArXiv Prepr ArXiv211001963; (2021)"},{"key":"517_CR140","doi-asserted-by":"crossref","unstructured":"Garcia, N., Hirota, Y., Wu, Y., Nakashima, Y.: Uncurated Image-Text Datasets: Shedding Light on Demographic Bias. (2023)","DOI":"10.1109\/CVPR52729.2023.00672"},{"key":"517_CR141","doi-asserted-by":"crossref","unstructured":"Bourtoule, L., Chandrasekaran, V., Choquette-Choo, C.A., Jia, H., Travers, A., Zhang, B., et al.: Machine unlearning. In: 2021 IEEE Symposium on Security and Privacy (SP). IEEE; pp. 141\u201359. (2021)","DOI":"10.1109\/SP40001.2021.00019"},{"key":"517_CR142","doi-asserted-by":"crossref","unstructured":"Koch, K., Soll, M.: No Matter How You Slice It: Machine Unlearning with SISA Comes at the Expense of Minority Classes. In: 2023 IEEE Conference on Secure and Trustworthy Machine Learning (SaTML). pp. 622\u201337. (2023)","DOI":"10.1109\/SaTML54575.2023.00047"},{"key":"517_CR143","doi-asserted-by":"crossref","unstructured":"van Maanen, G.A.I., Ethics: Ethics washing, and the need to Politicize Data Ethics. Digit. Soc. 1(2), 9 (2022)","DOI":"10.1007\/s44206-022-00013-3"}],"container-title":["AI and Ethics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s43681-024-00517-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s43681-024-00517-3\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s43681-024-00517-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,5,7]],"date-time":"2025-05-07T06:03:08Z","timestamp":1746597788000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s43681-024-00517-3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,7,24]]},"references-count":144,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2025,4]]}},"alternative-id":["517"],"URL":"https:\/\/doi.org\/10.1007\/s43681-024-00517-3","relation":{},"ISSN":["2730-5953","2730-5961"],"issn-type":[{"value":"2730-5953","type":"print"},{"value":"2730-5961","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,7,24]]},"assertion":[{"value":"5 April 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"27 June 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"24 July 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors have no competing interests to declare that are relevant to the content of this article.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}]}}