{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,30]],"date-time":"2025-10-30T09:18:39Z","timestamp":1761815919786,"version":"build-2065373602"},"publisher-location":"New York, NY, USA","reference-count":59,"publisher":"ACM","funder":[{"DOI":"10.13039\/501100000275","name":"Leverhulme Trust","doi-asserted-by":"publisher","award":["ECF-2021-429"],"award-info":[{"award-number":["ECF-2021-429"]}],"id":[{"id":"10.13039\/501100000275","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,11,5]]},"DOI":"10.1145\/3757887.3763018","type":"proceedings-article","created":{"date-parts":[[2025,10,29]],"date-time":"2025-10-29T07:42:58Z","timestamp":1761723778000},"page":"151-169","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["From Incidents to Insights: Patterns of Responsibility following AI Harms"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-3129-6022","authenticated-orcid":false,"given":"Isabel","family":"Richards","sequence":"first","affiliation":[{"name":"University of Cambridge, Cambridge, United Kingdom"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7072-2701","authenticated-orcid":false,"given":"Claire","family":"Benn","sequence":"additional","affiliation":[{"name":"University of Cambridge, Cambridge, United Kingdom"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9640-8139","authenticated-orcid":false,"given":"Miri","family":"Zilka","sequence":"additional","affiliation":[{"name":"University of Cambridge, Cambridge, United Kingdom"}]}],"member":"320","published-online":{"date-parts":[[2025,11,4]]},"reference":[{"volume-title":"AI Incident Database - About","year":"2024","key":"e_1_3_3_2_2_2","unstructured":"2024. AI Incident Database - About. https:\/\/incidentdatabase.ai\/about"},{"key":"e_1_3_3_2_3_2","doi-asserted-by":"crossref","unstructured":"Mike Ananny and Kate Crawford. 2018. Seeing without knowing: Limitations of the transparency ideal and its application to algorithmic accountability. new media & society 20 3 (2018) 973\u2013989.","DOI":"10.1177\/1461444816676645"},{"key":"e_1_3_3_2_4_2","unstructured":"Daniel Atherton. 2024. AI Incident Information Sharing Resources. https:\/\/github.com\/jphall663\/awesome-machine-learning-interpretability\/blob\/master \/README.md#ai-incident-information-sharing-resources"},{"key":"e_1_3_3_2_5_2","doi-asserted-by":"crossref","unstructured":"Jascha Bareis and Christian Katzenbach. 2022. Talking AI into being: The narratives and imaginaries of national AI strategies and their performative politics. Science Technology & Human Values 47 5 (2022) 855\u2013881.","DOI":"10.1177\/01622439211030007"},{"key":"e_1_3_3_2_6_2","doi-asserted-by":"crossref","unstructured":"Solon Barocas Sophie Hood and Malte Ziewitz. 2013. Governing algorithms: A provocation piece. Available at SSRN 2245322 (2013).","DOI":"10.2139\/ssrn.2245322"},{"key":"e_1_3_3_2_7_2","doi-asserted-by":"crossref","unstructured":"Mads Borup Nik Brown Kornelia Konrad and Harro Van\u00a0Lente. 2006. The sociology of expectations in science and technology. Technology analysis & strategic management 18 3-4 (2006) 285\u2013298.","DOI":"10.1080\/09537320600777002"},{"key":"e_1_3_3_2_8_2","doi-asserted-by":"crossref","unstructured":"Jenna Burrell. 2016. How the machine \u2018thinks\u2019: Understanding opacity in machine learning algorithms. Big data & society 3 1 (2016) 2053951715622512.","DOI":"10.1177\/2053951715622512"},{"key":"e_1_3_3_2_9_2","doi-asserted-by":"publisher","DOI":"10.1145\/3617694.3623249"},{"key":"e_1_3_3_2_10_2","unstructured":"Christopher Carey. 2023. Google rolls out AI to optimise traffic lights and cut emissions. Cities Today (October 2023). https:\/\/cities-today.com\/google-rolls-out-ai-to-optimise-traffic-lights-and-cut-emissions\/"},{"key":"e_1_3_3_2_11_2","doi-asserted-by":"crossref","unstructured":"D\u00a0Justin Coates. 2016. The epistemic norm of blame. Ethical Theory and Moral Practice 19 2 (2016) 457\u2013473.","DOI":"10.1007\/s10677-015-9639-8"},{"key":"e_1_3_3_2_12_2","doi-asserted-by":"crossref","unstructured":"Gerald\u00a0A Cohen. 2006. Casting the First Stone: Who Can and Who Can\u2019t Condemn the Terrorists? 1. Royal Institute of Philosophy Supplements 58 (2006) 113\u2013136.","DOI":"10.1017\/S1358246106058061"},{"key":"e_1_3_3_2_13_2","doi-asserted-by":"publisher","DOI":"10.1145\/3531146.3533150"},{"key":"e_1_3_3_2_14_2","doi-asserted-by":"crossref","unstructured":"Guillaume Dandurand Fenwick McKelvey and Jonathan Roberge. 2023. Freezing out: Legacy media\u2019s shaping of AI as a cold controversy. Big Data & Society 10 2 (2023) 20539517231219242.","DOI":"10.1177\/20539517231219242"},{"key":"e_1_3_3_2_15_2","doi-asserted-by":"crossref","unstructured":"Adrienne De\u00a0Ruiter. 2021. The distinct wrong of deepfakes. Philosophy & Technology 34 4 (2021) 1311\u20131332.","DOI":"10.1007\/s13347-021-00459-2"},{"key":"e_1_3_3_2_16_2","doi-asserted-by":"publisher","DOI":"10.1109\/STC55697.2022.00010"},{"key":"e_1_3_3_2_17_2","volume-title":"Reasoning practically","author":"Dworkin Gerald","year":"2000","unstructured":"Gerald Dworkin. 2000. Reasoning practically. Oxford University Press, Chapter Morally Speaking."},{"key":"e_1_3_3_2_18_2","unstructured":"Madeleine\u00a0Clare Elish. 2019. Moral crumple zones: Cautionary tales in human-robot interaction (pre-print). Engaging Science Technology and Society (pre-print) (2019)."},{"key":"e_1_3_3_2_19_2","doi-asserted-by":"publisher","DOI":"10.1145\/3617694.3623223"},{"key":"e_1_3_3_2_20_2","unstructured":"Sarah Fortinsky. 2023. Sports illustrated responds to accusations it published AI-generated content. https:\/\/thehill.com\/policy\/technology\/4330197-sports-illustrated-responds-to-accusations-it-published-ai-generated-content\/. [Accessed 08-01-2025]."},{"key":"e_1_3_3_2_21_2","unstructured":"Adrian Groza and Anca Marginean. 2023. Brave new world: Artificial Intelligence in teaching and learning. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2310.06856 (2023)."},{"key":"e_1_3_3_2_22_2","doi-asserted-by":"publisher","DOI":"10.1515\/9781503618220"},{"key":"e_1_3_3_2_23_2","doi-asserted-by":"crossref","unstructured":"Mia Hoffmann and Heather Frase. 2023. Adding structure to AI harm. https:\/\/cset.georgetown.edu\/publication\/adding-structure-to-ai-harm\/","DOI":"10.51593\/20230022"},{"key":"e_1_3_3_2_24_2","doi-asserted-by":"crossref","unstructured":"Julian Jacobs. 2024. The artificial intelligence shock and socio-political polarization. Technological Forecasting and Social Change 199 (2024) 123006.","DOI":"10.1016\/j.techfore.2023.123006"},{"key":"e_1_3_3_2_25_2","doi-asserted-by":"publisher","DOI":"10.1109\/ISSREW55968.2022.00094"},{"key":"e_1_3_3_2_26_2","unstructured":"Khoa Lam. 2023. ChatGPT Incidents and Issues. https:\/\/incidentdatabase.ai\/blog\/chatgpt-incidents-and-issues\/."},{"key":"e_1_3_3_2_27_2","unstructured":"Nora\u00a0Freya Lindemann. 2022. The ethical permissibility of chatting with the dead: Towards a normative framework for \u2018Deathbots\u2019. Publications of the Institute of Cognitive Science 1 (2022)."},{"key":"e_1_3_3_2_28_2","unstructured":"Ben Lovejoy. 2023. TikTok deepfakes: MrBeast Tom Hanks Gayle King; call for ban. https:\/\/9to5mac.com\/2023\/10\/04\/tiktok-deepfakes\/. [Accessed 08-01-2025]."},{"key":"e_1_3_3_2_29_2","doi-asserted-by":"crossref","unstructured":"Giampiero Lupo. 2023. Risky artificial intelligence: The role of incidents in the path to AI regulation. Law Technology and Humans 5 1 (2023) 133\u2013152.","DOI":"10.5204\/lthj.2682"},{"key":"e_1_3_3_2_30_2","doi-asserted-by":"crossref","unstructured":"Carl Macrae. 2016. The problem with incident reporting. BMJ quality & safety 25 2 (2016) 71\u201375.","DOI":"10.1136\/bmjqs-2015-004732"},{"key":"e_1_3_3_2_31_2","unstructured":"Carl Macrae. 2022. Understanding Autonomous Vehicle Risk: A Case Study Analysis. Safety Science (2022)."},{"key":"e_1_3_3_2_32_2","doi-asserted-by":"crossref","unstructured":"Noortje Marres. 2015. Why map issues? On controversy analysis as a digital method. Science Technology & Human Values 40 5 (2015) 655\u2013686.","DOI":"10.1177\/0162243915574602"},{"key":"e_1_3_3_2_33_2","first-page":"228","volume-title":"MEDIA","author":"Marres Noortje","year":"2021","unstructured":"Noortje Marres. 2021. No issues without media: The changing politics of public controversy in digital societies. In MEDIA. Intellect, 228\u2013243."},{"key":"e_1_3_3_2_34_2","doi-asserted-by":"crossref","unstructured":"Noortje Marres and David Moats. 2015. Mapping controversies with social media: The case for symmetry. Social Media+ Society 1 2 (2015) 2056305115604176.","DOI":"10.1177\/2056305115604176"},{"key":"e_1_3_3_2_35_2","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v35i17.17817"},{"key":"e_1_3_3_2_36_2","doi-asserted-by":"publisher","DOI":"10.1093\/acprof:oso\/9780199740031.001.0001"},{"key":"e_1_3_3_2_37_2","unstructured":"L McWilliams. 2024. New EY research finds AI investment is surging with senior leaders seeing more positive ROI as hype continues to become reality. https:\/\/www.ey.com\/en_us\/newsroom\/2024\/07\/new-ey-research-finds-ai-investment-is-surging-with-senior-leaders-seeing-more-positive-roi-as-hype-continues-to-become-reality"},{"key":"e_1_3_3_2_38_2","doi-asserted-by":"crossref","unstructured":"Helen Nissenbaum. 1996. Accountability in a computerized society. Science and engineering ethics 2 1 (1996) 25\u201342.","DOI":"10.1007\/BF02639315"},{"key":"e_1_3_3_2_39_2","doi-asserted-by":"publisher","DOI":"10.4159\/harvard.9780674061194"},{"key":"e_1_3_3_2_40_2","unstructured":"Nikiforos Pittaras and Sean McGregor. 2022. A taxonomic system for failure cause analysis of open source AI incidents. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2211.07280 (2022)."},{"key":"e_1_3_3_2_41_2","doi-asserted-by":"publisher","DOI":"10.1145\/3514094.3534181"},{"key":"e_1_3_3_2_42_2","doi-asserted-by":"crossref","unstructured":"Rowena Rodrigues Anais Resseguier and Nicole Santiago. 2023. When artificial intelligence fails: The emerging role of incident databases. Pub. Governance Admin. & Fin. L. Rev. 8 (2023) 17.","DOI":"10.53116\/pgaflr.7030"},{"key":"e_1_3_3_2_43_2","doi-asserted-by":"crossref","unstructured":"Gideon Rosen. 2004. Skepticism about moral responsibility. Philosophical perspectives 18 (2004) 295\u2013313.","DOI":"10.1111\/j.1520-8583.2004.00030.x"},{"key":"e_1_3_3_2_44_2","doi-asserted-by":"publisher","DOI":"10.2307\/j.ctv134vmrn"},{"key":"e_1_3_3_2_45_2","doi-asserted-by":"publisher","DOI":"10.4159\/9780674043145"},{"key":"e_1_3_3_2_46_2","doi-asserted-by":"crossref","unstructured":"Tim Schr\u00f6der and Michael Schulz. 2022. Monitoring machine learning models: a categorization of challenges and methods. Data Science and Management 5 3 (2022) 105\u2013116.","DOI":"10.1016\/j.dsm.2022.07.004"},{"key":"e_1_3_3_2_47_2","unstructured":"Janet Schwartz. 2022. Introducing AI Incident Responses. https:\/\/incidentdatabase.ai\/blog\/introducing-ai-incident-responses\/."},{"key":"e_1_3_3_2_48_2","doi-asserted-by":"crossref","unstructured":"Kris Shrishak. 2023. How to deal with an AI near-miss: Look to the skies. Bulletin of the Atomic Scientists 79 3 (2023) 166\u2013169.","DOI":"10.1080\/00963402.2023.2199580"},{"key":"e_1_3_3_2_49_2","doi-asserted-by":"crossref","unstructured":"Jack Stilgoe. 2018. Machine learning social learning and the governance of self-driving cars. Social studies of science 48 1 (2018) 25\u201356.","DOI":"10.1177\/0306312717741687"},{"key":"e_1_3_3_2_50_2","doi-asserted-by":"crossref","unstructured":"Jack Stilgoe. 2020. Who\u2019s driving innovation. New Technologies and the Collaborative State. Cham Switzerland: Palgrave Macmillan (2020).","DOI":"10.1007\/978-3-030-32320-2"},{"key":"e_1_3_3_2_51_2","unstructured":"Peter\u00a0F Strawson et\u00a0al. 2003. Freedom and resentment. Free will 2 (2003) 72\u201393."},{"key":"e_1_3_3_2_52_2","doi-asserted-by":"publisher","DOI":"10.1093\/0198248490.001.0001"},{"key":"e_1_3_3_2_53_2","unstructured":"Lionel\u00a0Nganyewou Tidjon and Foutse Khomh. 2022. Threat assessment in machine learning based systems. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2207.00091 (2022)."},{"key":"e_1_3_3_2_54_2","unstructured":"The\u00a0Straits Times. 2023. News anchors targeted by deepfake scammers on Facebook. https:\/\/www.straitstimes.com\/world\/news-anchors-targeted-by-deepfake-scammers-on-facebook. [Accessed 08-01-2025]."},{"key":"e_1_3_3_2_55_2","doi-asserted-by":"publisher","DOI":"10.1145\/3600211.3604700"},{"key":"e_1_3_3_2_56_2","unstructured":"Mengyi Wei and Zhixuan Zhou. 2022. AI ethics issues in real world: Evidence from AI incident database. arXiv preprint arXiv:https:\/\/arXiv.org\/abs\/2206.07635 (2022)."},{"key":"e_1_3_3_2_57_2","doi-asserted-by":"crossref","unstructured":"Roger Wertheimer. 1998. Constraining condemning. Ethics 108 3 (1998) 489\u2013501.","DOI":"10.1086\/233823"},{"key":"e_1_3_3_2_58_2","doi-asserted-by":"crossref","unstructured":"Tammy Worth. 2024. Are robots the solution to the crisis in older-person care? Nature (2024).","DOI":"10.1038\/d41586-024-01184-4"},{"key":"e_1_3_3_2_59_2","unstructured":"Xinhui Zhan Heshan Sun and Shaila\u00a0M Miranda. 2023. How does AI fail us? A typological theorization of AI failures. ICIS 2023 Proceedings 25 (2023) 1\u201317."},{"key":"e_1_3_3_2_60_2","first-page":"131","volume-title":"Moral Responsibility and the Problem of Many Hands","author":"Zwart Sjoerd\u00a0D","year":"2015","unstructured":"Sjoerd\u00a0D Zwart. 2015. Responsibility and the problem of many hands in networks. In Moral Responsibility and the Problem of Many Hands. Routledge, 131\u2013166."}],"event":{"name":"EAAMO '25: Equity and Access in Algorithms, Mechanisms, and Optimization","sponsor":["SIGecom Special Interest Group on Economics and Computation","SIGAI ACM Special Interest Group on Artificial Intelligence"],"location":"Pittsburgh USA","acronym":"EAAMO '25"},"container-title":["Proceedings of the 5th ACM Conference on Equity and Access in Algorithms, Mechanisms, and Optimization"],"original-title":[],"deposited":{"date-parts":[[2025,10,30]],"date-time":"2025-10-30T09:14:12Z","timestamp":1761815652000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3757887.3763018"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,11,4]]},"references-count":59,"alternative-id":["10.1145\/3757887.3763018","10.1145\/3757887"],"URL":"https:\/\/doi.org\/10.1145\/3757887.3763018","relation":{},"subject":[],"published":{"date-parts":[[2025,11,4]]},"assertion":[{"value":"2025-11-04","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}