{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,7]],"date-time":"2026-05-07T23:45:12Z","timestamp":1778197512473,"version":"3.51.4"},"reference-count":53,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2025,1,9]],"date-time":"2025-01-09T00:00:00Z","timestamp":1736380800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2025,1,9]],"date-time":"2025-01-09T00:00:00Z","timestamp":1736380800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100001381","name":"National Research Foundation Singapore","doi-asserted-by":"publisher","award":["AISG3-GV-2023-012"],"award-info":[{"award-number":["AISG3-GV-2023-012"]}],"id":[{"id":"10.13039\/501100001381","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100013373","name":"NIHR Oxford Biomedical Research Centre","doi-asserted-by":"publisher","award":["NIHR203316"],"award-info":[{"award-number":["NIHR203316"]}],"id":[{"id":"10.13039\/501100013373","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100010269","name":"Wellcome Trust","doi-asserted-by":"publisher","award":["203132\/Z\/16\/Z"],"award-info":[{"award-number":["203132\/Z\/16\/Z"]}],"id":[{"id":"10.13039\/100010269","id-type":"DOI","asserted-by":"publisher"}]},{"name":"International Collaborative Bioscience Innovation & Law Programme","award":["NNF23SA0087056"],"award-info":[{"award-number":["NNF23SA0087056"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["AI Ethics"],"published-print":{"date-parts":[[2025,2]]},"abstract":"<jats:title>Abstract<\/jats:title>\n          <jats:p>As artificial intelligence (AI) systems begin to take on social roles traditionally filled by humans, it will be crucial to understand how this affects people\u2019s cooperative expectations. In the case of human\u2013human dyads, different relationships are governed by different norms: For example, how two strangers\u2014versus two friends or colleagues\u2014should interact when faced with a similar coordination problem often differs. How will the rise of \u2018social\u2019 artificial intelligence (and ultimately, superintelligent AI) complicate people\u2019s expectations about the cooperative norms that should govern different types of relationships, whether human\u2013human or human\u2013AI? Do people expect AI to adhere to the same cooperative dynamics as humans when in a given social role? Conversely, will they begin to expect humans in certain types of relationships to act more like AI? Here, we consider how people\u2019s cooperative expectations may pull apart between human\u2013human and human\u2013AI relationships, detailing an empirical proposal for mapping these distinctions across relationship types. We see the data resulting from our proposal as relevant for understanding people\u2019s relationship\u2013specific cooperative expectations in an age of social AI, which may also forecast potential resistance towards AI systems occupying certain social roles. Finally, these data can form the basis for ethical evaluations: What relationship\u2013specific cooperative norms we should adopt for human\u2013AI interactions, or reinforce through responsible AI design, depends partly on empirical facts about what norms people find intuitive for such interactions (along with the costs and benefits of maintaining these). Toward the end of the paper, we discuss how these relational norms may change over time and consider the implications of this for the proposed research program.<\/jats:p>","DOI":"10.1007\/s43681-024-00631-2","type":"journal-article","created":{"date-parts":[[2025,1,9]],"date-time":"2025-01-09T10:51:24Z","timestamp":1736419884000},"page":"71-80","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["The need for an empirical research program regarding human\u2013AI relational norms"],"prefix":"10.1007","volume":"5","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-5944-0209","authenticated-orcid":false,"given":"Madeline G.","family":"Reinecke","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0867-6630","authenticated-orcid":false,"given":"Andreas","family":"Kappes","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1867-2097","authenticated-orcid":false,"given":"Sebastian","family":"Porsdam Mann","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1691-6403","authenticated-orcid":false,"given":"Julian","family":"Savulescu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9691-2888","authenticated-orcid":false,"given":"Brian D.","family":"Earp","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,1,9]]},"reference":[{"issue":"5","key":"631_CR1","doi-asserted-by":"publisher","first-page":"1157","DOI":"10.1037\/pspp0000470","volume":"125","author":"M Atari","year":"2023","unstructured":"Atari, M., Haidt, J., Graham, J., Koleva, S., Stevens, S.T., Dehghani, M.: Morality beyond the WEIRD: how the nomological network of morality varies across cultures. J. Pers. Soc. Psychol. 125(5), 1157\u20131188 (2023). https:\/\/doi.org\/10.1037\/pspp0000470","journal-title":"J. Pers. Soc. Psychol."},{"key":"631_CR2","doi-asserted-by":"crossref","unstructured":"Atari, M., Xue, M.J., Park, P.S., Blasi, D., Henrich, J.: Which humans? (2023). https:\/\/osf.io\/5b26t","DOI":"10.31234\/osf.io\/5b26t"},{"key":"631_CR3","doi-asserted-by":"publisher","unstructured":"Bender, E.M., Gebru, T., McMillan-Major, A., Shmitchell, S.: On the dangers of stochastic parrots: can language models be too big? In: Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency, pp. 610\u2013623 (2021). https:\/\/doi.org\/10.1145\/3442188.3445922","DOI":"10.1145\/3442188.3445922"},{"key":"631_CR4","unstructured":"Biden, J.R.: Executive order on the safe, secure, and trustworthy development and use of artificial intelligence (2023). https:\/\/www.whitehouse.gov\/briefing-room\/presidential-actions\/2023\/10\/30\/executive-order-on-the-safe-secure-and-trustworthy-development-and-use-of-artificial-intelligence\/"},{"issue":"1","key":"631_CR5","doi-asserted-by":"publisher","first-page":"4","DOI":"10.1037\/xge0001250","volume":"152","author":"YE Bigman","year":"2023","unstructured":"Bigman, Y.E., Wilson, D., Arnestad, M.N., Waytz, A., Gray, K.: Algorithmic discrimination causes less moral outrage than human discrimination. J. Exp. Psychol. Gen. 152(1), 4\u201327 (2023). https:\/\/doi.org\/10.1037\/xge0001250","journal-title":"J. Exp. Psychol. Gen."},{"issue":"2","key":"631_CR6","doi-asserted-by":"publisher","first-page":"26","DOI":"10.1111\/j.1467-9736.2011.00701.x","volume":"99","author":"P Bloom","year":"2011","unstructured":"Bloom, P.: Family, community, trolley problems, and the crisis in moral psychology. Yale Rev. 99(2), 26\u201343 (2011). https:\/\/doi.org\/10.1111\/j.1467-9736.2011.00701.x","journal-title":"Yale Rev."},{"key":"631_CR7","volume-title":"Superintelligence: Paths, Dangers, Strategies","author":"N Bostrom","year":"2014","unstructured":"Bostrom, N.: Superintelligence: Paths, Dangers, Strategies. Oxford University Press, Oxford (2014)"},{"key":"631_CR8","doi-asserted-by":"publisher","unstructured":"Bubeck, S., Chandrasekaran, V., Eldan, R., Gehrke, J., Horvitz, E., Kamar, E., Lee, P., Lee, Y.T., Li, Y., Lundberg, S., Nori, H., Palangi, H., Ribeiro, M.T., Zhang, Y.: Sparks of artificial general intelligence: early experiments with GPT-4 (2023). https:\/\/doi.org\/10.48550\/arXiv.2303.12712","DOI":"10.48550\/arXiv.2303.12712"},{"issue":"2","key":"631_CR9","doi-asserted-by":"publisher","first-page":"187","DOI":"10.1037\/0033-2909.126.2.187","volume":"126","author":"DB Bugental","year":"2000","unstructured":"Bugental, D.B.: Acquisition of the algorithms of social life: a domain-based approach. Psychol. Bull. 126(2), 187\u2013219 (2000). https:\/\/doi.org\/10.1037\/0033-2909.126.2.187","journal-title":"Psychol. Bull."},{"key":"631_CR10","doi-asserted-by":"publisher","unstructured":"Chalmers, D.J.: Could a large language model be conscious? (2023). https:\/\/doi.org\/10.48550\/arXiv.2303.07103","DOI":"10.48550\/arXiv.2303.07103"},{"issue":"6","key":"631_CR11","doi-asserted-by":"publisher","first-page":"684","DOI":"10.1177\/0146167293196003","volume":"19","author":"MS Clark","year":"1993","unstructured":"Clark, M.S., Mills, J.: The difference between communal and exchange relationships: what it is and is not. Pers. Soc. Psychol. Bull. 19(6), 684\u2013691 (1993). https:\/\/doi.org\/10.1177\/0146167293196003","journal-title":"Pers. Soc. Psychol. Bull."},{"key":"631_CR12","doi-asserted-by":"publisher","unstructured":"Coeckelbergh, M.: Technology games\/gender games. from Wittgenstein\u2019s toolbox and language games to gendered robots and biased artificial intelligence. In: Feminist Philosophy of Technology, pp. 27\u201338. Springer (2020). https:\/\/doi.org\/10.1007\/978-3-476-04967-4_2","DOI":"10.1007\/978-3-476-04967-4_2"},{"issue":"1","key":"631_CR13","doi-asserted-by":"publisher","first-page":"niae013","DOI":"10.1093\/nc\/niae013","volume":"2024","author":"C Colombatto","year":"2024","unstructured":"Colombatto, C., Fleming, S.M.: Folk psychological attributions of consciousness to large language models. Neurosci. Conscious. 2024(1), niae013 (2024). https:\/\/doi.org\/10.1093\/nc\/niae013","journal-title":"Neurosci. Conscious."},{"key":"631_CR14","doi-asserted-by":"publisher","first-page":"106","DOI":"10.1016\/j.jrp.2018.10.008","volume":"78","author":"OS Curry","year":"2019","unstructured":"Curry, O.S., Jones Chesters, M., Van Lissa, C.J.: Mapping morality with a compass: testing the theory of \u2018morality-as-cooperation\u2019 with a new questionnaire. J. Res. Pers. 78, 106\u2013124 (2019). https:\/\/doi.org\/10.1016\/j.jrp.2018.10.008","journal-title":"J. Res. Pers."},{"issue":"4","key":"631_CR15","doi-asserted-by":"publisher","first-page":"629","DOI":"10.1007\/s13347-018-0317-3","volume":"31","author":"J Danaher","year":"2018","unstructured":"Danaher, J.: Toward an ethics of AI assistants: an initial framework. Philos. Technol. 31(4), 629\u2013653 (2018). https:\/\/doi.org\/10.1007\/s13347-018-0317-3","journal-title":"Philos. Technol."},{"issue":"1","key":"631_CR16","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1080\/09515089.2014.989967","volume":"29","author":"J Demaree-Cotton","year":"2016","unstructured":"Demaree-Cotton, J.: Do framing effects make moral intuitions unreliable? Philos. Psychol. 29(1), 1\u201322 (2016). https:\/\/doi.org\/10.1080\/09515089.2014.989967","journal-title":"Philos. Psychol."},{"issue":"4","key":"631_CR17","doi-asserted-by":"publisher","first-page":"720","DOI":"10.1177\/01634437221119021","volume":"45","author":"I Depounti","year":"2023","unstructured":"Depounti, I., Saukko, P., Natale, S.: Ideal technologies, ideal women: AI and gender imaginaries in Redditors\u2019 discussions on the Replika bot girlfriend. Media Cult. Soc. 45(4), 720\u2013736 (2023). https:\/\/doi.org\/10.1177\/01634437221119021","journal-title":"Media Cult. Soc."},{"key":"631_CR18","doi-asserted-by":"crossref","unstructured":"Dong, M., Conway, J., Bonnefon, J.-F., Shariff, A., Rahwan, I.: A psychological model predicts fears about artificial intelligence across 20 countries and 6 domains of application (2023). https:\/\/osf.io\/preprints\/psyarxiv\/pjvqt","DOI":"10.31234\/osf.io\/pjvqt"},{"key":"631_CR19","unstructured":"Earp, B.D., McLoughlin, K.L., Calcott, R., Caraccio, M., Monrad, J.T., Owen, A.M., Crockett, M.J., Clark, M.S.: How social relationships shape praise and blame: Strengthening and extending the relational norms model (2025). https:\/\/www.researchgate.net\/publication\/387959265"},{"issue":"3","key":"631_CR20","doi-asserted-by":"publisher","first-page":"91","DOI":"10.1007\/s11017-021-09546-z","volume":"42","author":"BD Earp","year":"2021","unstructured":"Earp, B.D., Lewis, J., Dranseika, V., Hannikainen, I.R.: Experimental philosophical bioethics and normative inference. Theor. Med. Bioeth. 42(3), 91\u2013111 (2021). https:\/\/doi.org\/10.1007\/s11017-021-09546-z","journal-title":"Theor. Med. Bioeth."},{"issue":"1","key":"631_CR21","doi-asserted-by":"publisher","first-page":"5776","DOI":"10.1038\/s41467-021-26067-4","volume":"12","author":"BD Earp","year":"2021","unstructured":"Earp, B.D., McLoughlin, K.L., Monrad, J.T., Clark, M.S., Crockett, M.J.: How social relationships shape moral wrongness judgments. Nat. Commun. 12(1), 5776 (2021). https:\/\/doi.org\/10.1038\/s41467-021-26067-4","journal-title":"Nat. Commun."},{"key":"631_CR22","unstructured":"European Parliament. EU AI Act: First Regulation on Artificial Intelligence (2024). https:\/\/artificialintelligenceact.eu\/the-act\/"},{"issue":"4","key":"631_CR23","doi-asserted-by":"publisher","first-page":"995","DOI":"10.1007\/s00146-020-00942-y","volume":"35","author":"E Firt","year":"2020","unstructured":"Firt, E.: The missing G. AI Soc. 35(4), 995\u20131007 (2020). https:\/\/doi.org\/10.1007\/s00146-020-00942-y","journal-title":"AI Soc."},{"key":"631_CR24","doi-asserted-by":"publisher","unstructured":"Gabriel, I., Manzini, A., Keeling, G., Hendricks, L.A., Rieser, V., Iqbal, H., Toma\u0161ev, N., Ktena, I., Kenton, Z., Rodriguez, M., El-Sayed, S., Brown, S., Akbulut, C., Trask, A., Hughes, E., Bergman, A.S., Shelby, R., Marchal, N., Griffin, C., Isaac, W., Manyika, J.: The ethics of advanced AI assistants (2024). https:\/\/doi.org\/10.48550\/arXiv.2404.16244","DOI":"10.48550\/arXiv.2404.16244"},{"key":"631_CR25","doi-asserted-by":"publisher","unstructured":"Gallagher, S., Gelman, B., Taoufiq, S., V\u00f6r\u00f6s, T., Lee, Y., Kyadige, A., Bergeron, S.: Phishing and social engineering in the age of LLMs. In: Kucharavy, A., Plancherel, O., Mulder, V., Mermoud, A., Lenders, V. (eds.) Large Language Models in Cybersecurity: Threats, Exposure and Mitigation, pp. 81\u201386. Springer Nature, Switzerland (2024). https:\/\/doi.org\/10.1007\/978-3-031-54827-7_8","DOI":"10.1007\/978-3-031-54827-7_8"},{"issue":"7","key":"631_CR26","doi-asserted-by":"publisher","first-page":"e16649","DOI":"10.2196\/16649","volume":"22","author":"S Gao","year":"2020","unstructured":"Gao, S., He, L., Chen, Y., Li, D., Lai, K.: Public perception of artificial intelligence in medical care: content analysis of social media. J. Med. Internet Res. 22(7), e16649 (2020). https:\/\/doi.org\/10.2196\/16649","journal-title":"J. Med. Internet Res."},{"key":"631_CR27","volume-title":"The Hidden Pattern: A Patternist Philosophy of Mind","author":"B Goertzel","year":"2006","unstructured":"Goertzel, B.: The Hidden Pattern: A Patternist Philosophy of Mind. BrownWalker Press, Irvine (2006)"},{"key":"631_CR28","doi-asserted-by":"publisher","DOI":"10.1016\/S0065-2458(08)60418-0","volume-title":"Speculations Concerning the First Ultraintelligent Machine","author":"IJ Good","year":"1966","unstructured":"Good, I.J.: Speculations Concerning the First Ultraintelligent Machine. Elsevier, Amsterdam (1966)"},{"issue":"5812","key":"631_CR29","doi-asserted-by":"publisher","first-page":"619","DOI":"10.1126\/science.1134475","volume":"315","author":"HM Gray","year":"2007","unstructured":"Gray, H.M., Gray, K., Wegner, D.M.: Dimensions of mind perception. Science 315(5812), 619 (2007). https:\/\/doi.org\/10.1126\/science.1134475","journal-title":"Science"},{"key":"631_CR30","unstructured":"Gruetzemacher, R., Whittlestone, J.: Defining and unpacking transformative AI (2019). arXiv:1912.00747v1"},{"key":"631_CR31","unstructured":"Gumusel, E., Zhou, K. Z., Sanfilippo, M.R.: User privacy harms and risks in conversational AI: a proposed framework (2024). arXiv:2402.09716"},{"issue":"2","key":"631_CR32","doi-asserted-by":"publisher","first-page":"201","DOI":"10.1002\/(SICI)1099-0992(199603)26:2<201::AID-EJSP745>3.0.CO;2-J","volume":"26","author":"J Haidt","year":"1996","unstructured":"Haidt, J., Baron, J.: Social roles and the moral judgement of acts and omissions. Eur. J. Soc. Psychol. 26(2), 201\u2013218 (1996)","journal-title":"Eur. J. Soc. Psychol."},{"issue":"3","key":"631_CR33","doi-asserted-by":"publisher","first-page":"237","DOI":"10.2307\/3033836","volume":"44","author":"VL Hamilton","year":"1981","unstructured":"Hamilton, V.L., Sanders, J.: The effect of roles and deeds on responsibility judgments: the normative structure of wrongdoing. Soc. Psychol. Q. 44(3), 237\u2013254 (1981). https:\/\/doi.org\/10.2307\/3033836","journal-title":"Soc. Psychol. Q."},{"issue":"2","key":"631_CR34","doi-asserted-by":"publisher","first-page":"216","DOI":"10.1177\/1745691619885840","volume":"15","author":"N Hester","year":"2020","unstructured":"Hester, N., Gray, K.: The moral psychology of raceless, genderless strangers. Perspect. Psychol. Sci. 15(2), 216\u2013230 (2020). https:\/\/doi.org\/10.1177\/1745691619885840","journal-title":"Perspect. Psychol. Sci."},{"issue":"9","key":"631_CR35","doi-asserted-by":"publisher","first-page":"389","DOI":"10.1038\/s42256-019-0088-2","volume":"1","author":"A Jobin","year":"2019","unstructured":"Jobin, A., Ienca, M., Vayena, E.: The global landscape of AI ethics guidelines. Nat. Mach. Intell. 1(9), 389\u2013399 (2019). https:\/\/doi.org\/10.1038\/s42256-019-0088-2","journal-title":"Nat. Mach. Intell."},{"key":"631_CR36","doi-asserted-by":"publisher","first-page":"111","DOI":"10.1007\/978-3-031-13631-3_7","volume-title":"Synthetic Friends","author":"H Kempt","year":"2022","unstructured":"Kempt, H.: Synthetic friends. In: Synthetic Friends, pp. 111\u2013161. Springer International Publishing, Berlin (2022)"},{"key":"631_CR37","doi-asserted-by":"publisher","unstructured":"Khan, A.A., Badshah, S., Liang, P., Khan, B., Waseem, M., Niazi, M., Akbar, M.A.: Ethics of AI: a systematic literature review of principles and challenges (2021). https:\/\/doi.org\/10.48550\/arXiv.2109.07906","DOI":"10.48550\/arXiv.2109.07906"},{"issue":"2","key":"631_CR38","doi-asserted-by":"publisher","first-page":"32","DOI":"10.1007\/s10676-023-09703-z","volume":"25","author":"JJ Koplin","year":"2023","unstructured":"Koplin, J.J.: Dual-use implications of AI text generation. Ethics Inf. Technol. 25(2), 32 (2023). https:\/\/doi.org\/10.1007\/s10676-023-09703-z","journal-title":"Ethics Inf. Technol."},{"key":"631_CR39","doi-asserted-by":"publisher","first-page":"102783","DOI":"10.1016\/j.ijinfomgt.2024.102783","volume":"77","author":"V Kumar","year":"2024","unstructured":"Kumar, V., Ashraf, A.R., Nadeem, W.: AI-powered marketing: what, where, and how? Int. J. Inf. Manag. 77, 102783 (2024). https:\/\/doi.org\/10.1016\/j.ijinfomgt.2024.102783","journal-title":"Int. J. Inf. Manag."},{"issue":"1","key":"631_CR40","doi-asserted-by":"publisher","first-page":"205395171875668","DOI":"10.1177\/2053951718756684","volume":"5","author":"MK Lee","year":"2018","unstructured":"Lee, M.K.: Understanding perception of algorithmic decisions: fairness, trust, and emotion in response to algorithmic management. Big Data Soc. 5(1), 2053951718756684 (2018). https:\/\/doi.org\/10.1177\/2053951718756684","journal-title":"Big Data Soc."},{"key":"631_CR41","unstructured":"Legg, S., Hutter, M.: A formal measure of machine intelligence (2006). arXiv:cs\/0605024"},{"issue":"1","key":"631_CR42","doi-asserted-by":"publisher","first-page":"3108","DOI":"10.1038\/s41467-023-38592-5","volume":"14","author":"K Makovi","year":"2023","unstructured":"Makovi, K., Sargsyan, A., Li, W., Bonnefon, J.-F., Rahwan, T.: Trust within human-machine collectives depends on the perceived consensus about cooperative norms. Nat. Commun. 14(1), 3108 (2023). https:\/\/doi.org\/10.1038\/s41467-023-38592-5","journal-title":"Nat. Commun."},{"key":"631_CR43","doi-asserted-by":"crossref","unstructured":"Mallick, R., Flathmann, C., Lancaster, C., Hauptman, A., McNeese, N., Freeman, G.: The pursuit of happiness: the power and influence of AI teammate emotion in human-AI teamwork (2023). https:\/\/www.tandfonline.com\/doi\/abs\/10.1080\/0144929X.2023.2277909","DOI":"10.1080\/0144929X.2023.2277909"},{"key":"631_CR44","first-page":"109","volume":"30","author":"K Maria","year":"2022","unstructured":"Maria, K., Drigas, A., Skianis, C.: Chatbots as cognitive, educational, advisory & coaching systems education. Tech. Soc. Sci. J. 30, 109\u2013126 (2022)","journal-title":"Tech. Soc. Sci. J."},{"key":"631_CR45","volume-title":"Theoretical Frameworks for Personal Relationships","author":"J Mills","year":"1994","unstructured":"Mills, J., Clark, M.S.: Communal and exchange relationships: controversies and research. In: Theoretical Frameworks for Personal Relationships. Psychology Press, Hove (1994)"},{"key":"631_CR46","doi-asserted-by":"publisher","first-page":"555","DOI":"10.1007\/978-3-319-26485-1_33","volume-title":"Fundamental Issues of Artificial Intelligence","author":"VC M\u00fcller","year":"2016","unstructured":"M\u00fcller, V.C., Bostrom, N.: Future progress in artificial intelligence: a survey of expert opinion. In: M\u00fcller, V.C. (ed.) Fundamental Issues of Artificial Intelligence, pp. 555\u2013572. Springer International Publishing, Berlin (2016). https:\/\/doi.org\/10.1007\/978-3-319-26485-1_33"},{"key":"631_CR47","doi-asserted-by":"publisher","first-page":"107372","DOI":"10.1016\/j.chb.2022.107372","volume":"136","author":"JVT Pauketat","year":"2022","unstructured":"Pauketat, J.V.T., Anthis, J.R.: Predicting the moral consideration of artificial intelligences. Comput. Hum. Behav. 136, 107372 (2022). https:\/\/doi.org\/10.1016\/j.chb.2022.107372","journal-title":"Comput. Hum. Behav."},{"issue":"5","key":"631_CR48","doi-asserted-by":"publisher","first-page":"472","DOI":"10.1038\/s42256-023-00653-1","volume":"5","author":"S Porsdam Mann","year":"2023","unstructured":"Porsdam Mann, S., Earp, B.D., Nyholm, S., Danaher, J., M\u00f8ller, N., Bowman-Smart, H., Hatherley, J., Koplin, J., Plozza, M., Rodger, D., Treit, P.V., Renard, G., McMillan, J., Savulescu, J.: Generative AI entails a credit-blame asymmetry. Nat. Mach. Intell. 5(5), 472\u2013475 (2023). https:\/\/doi.org\/10.1038\/s42256-023-00653-1","journal-title":"Nat. Mach. Intell."},{"issue":"1","key":"631_CR49","doi-asserted-by":"publisher","first-page":"57","DOI":"10.1037\/a0021867","volume":"118","author":"TS Rai","year":"2011","unstructured":"Rai, T.S., Fiske, A.P.: Moral psychology is relationship regulation: moral motives for unity, hierarchy, equality, and proportionality. Psychol. Rev. 118(1), 57\u201375 (2011). https:\/\/doi.org\/10.1037\/a0021867","journal-title":"Psychol. Rev."},{"issue":"7","key":"631_CR50","doi-asserted-by":"publisher","first-page":"652","DOI":"10.1111\/bioe.12869","volume":"35","author":"J Savulescu","year":"2021","unstructured":"Savulescu, J., Gyngell, C., Kahane, G.: Collective reflective equilibrium in practice (CREP) and controversial novel technologies. Bioethics 35(7), 652\u2013663 (2021). https:\/\/doi.org\/10.1111\/bioe.12869","journal-title":"Bioethics"},{"key":"631_CR51","unstructured":"UNESCO. Recommendation on the ethics of artificial intelligence (2021). https:\/\/www.unesco.org\/en\/articles\/recommendation-ethics-artificial-intelligence"},{"key":"631_CR52","doi-asserted-by":"publisher","unstructured":"Weidinger, L., Uesato, J., Rauh, M., Griffin, C., Huang, P.-S., Mellor, J., Glaese, A., Cheng, M., Balle, B., Kasirzadeh, A., Biles, C., Brown, S., Kenton, Z., Hawkins, W., Stepleton, T., Birhane, A., Hendricks, L.A., Rimell, L., Isaac, W., Irving, G., Gabriel, I.: Taxonomy of risks posed by language models. In: Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency, pp. 214\u2013229 (2022). https:\/\/doi.org\/10.1145\/3531146.3533088","DOI":"10.1145\/3531146.3533088"},{"key":"631_CR53","doi-asserted-by":"publisher","first-page":"1272","DOI":"10.1038\/s42256-024-00922-7","volume":"6","author":"S Porsdam Mann","year":"2024","unstructured":"Porsdam Mann, S., Vazirani, A.A., Aboy, M. et al.: Guidelines for ethical use and acknowledgement of large language models in academic writing. Nat. Mach. Intell. 6, 1272\u20131274 (2024). https:\/\/doi.org\/10.1038\/s42256-024-00922-7","journal-title":"Nat. Mach. Intell."}],"updated-by":[{"DOI":"10.1007\/s43681-025-00659-y","type":"correction","label":"Correction","source":"publisher","updated":{"date-parts":[[2025,2,18]],"date-time":"2025-02-18T00:00:00Z","timestamp":1739836800000}}],"container-title":["AI and Ethics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s43681-024-00631-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s43681-024-00631-2\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s43681-024-00631-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,3,24]],"date-time":"2025-03-24T07:03:32Z","timestamp":1742799812000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s43681-024-00631-2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,1,9]]},"references-count":53,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2025,2]]}},"alternative-id":["631"],"URL":"https:\/\/doi.org\/10.1007\/s43681-024-00631-2","relation":{"correction":[{"id-type":"doi","id":"10.1007\/s43681-025-00659-y","asserted-by":"object"}]},"ISSN":["2730-5953","2730-5961"],"issn-type":[{"value":"2730-5953","type":"print"},{"value":"2730-5961","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,1,9]]},"assertion":[{"value":"10 July 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"16 November 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"9 January 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"17 January 2025","order":4,"name":"change_date","label":"Change Date","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"Update","order":5,"name":"change_type","label":"Change Type","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"The original online version of this article was revised: In this article, on the first page, where it says \u201c\u2026characteristically or under certain conditions [21],\u201d (three lines up from the bottom of the right-hand column), the reference \u2018[21]\u2019 should have read \u2018[19]\u2019. In this article, on the third page (Section 1), about halfway down the page, where it says \u201cFor example, in Earp [19],\u201d the term \u201cEarp\u201d should be deleted. In this article, in the references section, the reference to my PhD dissertation (Earp, B. D., Relational morality \u2014 etc.) was incorrect and should have been replaced with the following: Earp, B. D., McLoughlin, K. L., Calcott, R., Caraccio, M., Monrad, J. T., Owen, A. M., Crockett, M. J., & Clark, M. S. (2025). How social relationships shape praise and blame: Strengthening and extending the relational norms model. Retrieved from https:\/\/www.researchgate.net\/publication\/387959265. The original article has been corrected.","order":6,"name":"change_details","label":"Change Details","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"18 February 2025","order":7,"name":"change_date","label":"Change Date","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"Correction","order":8,"name":"change_type","label":"Change Type","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"A Correction to this paper has been published:","order":9,"name":"change_details","label":"Change Details","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"https:\/\/doi.org\/10.1007\/s43681-025-00659-y","URL":"https:\/\/doi.org\/10.1007\/s43681-025-00659-y","order":10,"name":"change_details","label":"Change Details","group":{"name":"ArticleHistory","label":"Article History"}}]}}