{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,21]],"date-time":"2026-04-21T04:11:57Z","timestamp":1776744717494,"version":"3.51.2"},"publisher-location":"New York, NY, USA","reference-count":144,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,6,12]],"date-time":"2023-06-12T00:00:00Z","timestamp":1686528000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"European Union Horizon 2020","award":["Marie Sklodowska-Curie grant agreement No 860621"],"award-info":[{"award-number":["Marie Sklodowska-Curie grant agreement No 860621"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,6,12]]},"DOI":"10.1145\/3593013.3594074","type":"proceedings-article","created":{"date-parts":[[2023,6,12]],"date-time":"2023-06-12T14:40:46Z","timestamp":1686580846000},"page":"1198-1212","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":46,"title":["Explainability in AI Policies: A Critical Review of Communications, Reports, Regulations, and Standards in the EU, US, and UK"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-4733-9760","authenticated-orcid":false,"given":"Luca","family":"Nannini","sequence":"first","affiliation":[{"name":"Minsait - Indra Sistemas, Spain and CiTIUS (Centro Singular de Investigaci\u00f3n en Tecnolox\u00edas Intelixentes), Universidade de Santiago de Compostela, Spain"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2725-5305","authenticated-orcid":false,"given":"Agathe","family":"Balayn","sequence":"additional","affiliation":[{"name":"Delft University of Technology, Netherlands"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6614-0888","authenticated-orcid":false,"given":"Adam Leon","family":"Smith","sequence":"additional","affiliation":[{"name":"Dragonfly, Spain"}]}],"member":"320","published-online":{"date-parts":[[2023,6,12]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"European Commission Directorate-General for Internal Market Industry Entrepreneurship and SMEs. 2022. Draft standardisation request to the European Standardisation Organisations in support of safe and trustworthy artificial intelligence. https:\/\/ec.europa.eu\/docsroom\/documents\/52376"},{"key":"e_1_3_2_1_2_1","volume-title":"National Artificial Intelligence Initiative Act of","author":"116th Congress of the United States of America. 2020.","year":"2020","unstructured":"116th Congress of the United States of America. 2020. National Artificial Intelligence Initiative Act of 2020. https:\/\/www.congress.gov\/116\/crpt\/hrpt617\/CRPT-116hrpt617.pdf#page=1210"},{"key":"e_1_3_2_1_3_1","volume-title":"Algorithmic Accountability Act of","author":"117th Congress of the United States of America. 2022.","year":"2022","unstructured":"117th Congress of the United States of America. 2022. Algorithmic Accountability Act of 2022, H.R.6580. https:\/\/www.congress.gov\/bill\/117th-congress\/house-bill\/6580\/text"},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1145\/3173574.3174156"},{"key":"e_1_3_2_1_5_1","volume-title":"Sanity Checks for Saliency Maps. In Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018","author":"Adebayo Julius","year":"2018","unstructured":"Julius Adebayo, Justin Gilmer, Michael Muelly, Ian J. Goodfellow, Moritz Hardt, and Been Kim. 2018. Sanity Checks for Saliency Maps. In Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, NeurIPS 2018, December 3-8, 2018, Montr\u00e9al, Canada, Samy Bengio, Hanna M. Wallach, Hugo Larochelle, Kristen Grauman, Nicol\u00f2 Cesa-Bianchi, and Roman Garnett (Eds.). 9525\u20139536. https:\/\/proceedings.neurips.cc\/paper\/2018\/hash\/294a8ed24b1ad22ec2e7efea049b8737-Abstract.html"},{"key":"e_1_3_2_1_6_1","volume-title":"The Tenth International Conference on Learning Representations, ICLR 2022","author":"Adebayo Julius","year":"2022","unstructured":"Julius Adebayo, Michael Muelly, Harold Abelson, and Been Kim. 2022. Post hoc Explanations may be Ineffective for Detecting Unknown Spurious Correlation. In The Tenth International Conference on Learning Representations, ICLR 2022, Virtual Event, April 25-29, 2022. OpenReview.net. https:\/\/openreview.net\/forum?id=xNOVfCCvDpM"},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","unstructured":"Mhairi Aitken David Leslie Florian Ostmann Jacob Pratt Helen Margetts and Cosmina Dorobantu. 2022. Common Regulatory Capacity for AI. (2022). https:\/\/doi.org\/10.5281\/zenodo.6838946","DOI":"10.5281\/zenodo.6838946"},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1145\/3377325.3377519"},{"key":"e_1_3_2_1_9_1","unstructured":"XAI Explainable Artificial Intelligence IEEE Computer Society (IEEE C\/AISC\/XAI) Artificial Intelligence Standards Committee. 2020. IEEE P2894 - Guide for an Architectural Framework for Explainable Artificial Intelligence. https:\/\/standards.ieee.org\/ieee\/2894\/10284\/."},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1007\/s43681-021-00084-x"},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1145\/3531146.3533179"},{"key":"e_1_3_2_1_12_1","volume-title":"Beyond Debiasing: Regulating AI and its inequalities. Technical Report. https:\/\/edri.org\/wp-content\/uploads\/2021\/09\/EDRi_Beyond-Debiasing-Report_Online.pdf","author":"Balayn Agathe","year":"2021","unstructured":"Agathe Balayn and Seda G\u00fcrses. 2021. Beyond Debiasing: Regulating AI and its inequalities. Technical Report. https:\/\/edri.org\/wp-content\/uploads\/2021\/09\/EDRi_Beyond-Debiasing-Report_Online.pdf"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1145\/3491102.3517474"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2019.12.012"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1145\/3531146.3533090"},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.3389\/fdata.2021.688969"},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3375624"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3372860"},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","unstructured":"Rishi Bommasani Drew A. Hudson Ehsan Adeli Russ Altman Simran Arora Sydney von Arx Michael S. Bernstein Jeannette Bohg Antoine Bosselut Emma Brunskill Erik Brynjolfsson Shyamal Buch Dallas Card Rodrigo Castellon Niladri Chatterji Annie Chen Kathleen Creel Jared Quincy Davis Dora Demszky Chris Donahue Moussa Doumbouya Esin Durmus Stefano Ermon John Etchemendy Kawin Ethayarajh Li Fei-Fei Chelsea Finn Trevor Gale Lauren Gillespie Karan Goel Noah Goodman Shelby Grossman Neel Guha Tatsunori Hashimoto Peter Henderson John Hewitt Daniel E. Ho Jenny Hong Kyle Hsu Jing Huang Thomas Icard Saahil Jain Dan Jurafsky Pratyusha Kalluri Siddharth Karamcheti Geoff Keeling Fereshte Khani Omar Khattab Pang Wei Koh Mark Krass Ranjay Krishna Rohith Kuditipudi Ananya Kumar Faisal Ladhak Mina Lee Tony Lee Jure Leskovec Isabelle Levent Xiang Lisa Li Xuechen Li Tengyu Ma Ali Malik Christopher D. Manning Suvir Mirchandani Eric Mitchell Zanele Munyikwa Suraj Nair Avanika Narayan Deepak Narayanan Ben Newman Allen Nie Juan Carlos Niebles Hamed Nilforoshan Julian Nyarko Giray Ogut Laurel Orr Isabel Papadimitriou Joon Sung Park Chris Piech Eva Portelance Christopher Potts Aditi Raghunathan Rob Reich Hongyu Ren Frieda Rong Yusuf Roohani Camilo Ruiz Jack Ryan Christopher R\u00e9 Dorsa Sadigh Shiori Sagawa Keshav Santhanam Andy Shih Krishnan Srinivasan Alex Tamkin Rohan Taori Armin W. Thomas Florian Tram\u00e8r Rose E. Wang William Wang Bohan Wu Jiajun Wu Yuhuai Wu Sang Michael Xie Michihiro Yasunaga Jiaxuan You Matei Zaharia Michael Zhang Tianyi Zhang Xikun Zhang Yuhui Zhang Lucia Zheng Kaitlyn Zhou and Percy Liang. 2021. On the Opportunities and Risks of Foundation Models. https:\/\/doi.org\/10.48550\/ARXIV.2108.07258","DOI":"10.48550\/ARXIV.2108.07258"},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1145\/3531146.3533153"},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.2304.00612"},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1017\/err.2020.10"},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.5445\/IR"},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1613\/jair.1.12228"},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2022.118888"},{"key":"e_1_3_2_1_27_1","unstructured":"CEN-CENELEC. 2020. Focus Group Report - Road Map on Artificial Intelligence (AI). https:\/\/www.standict.eu\/node\/4854 Accessed on: 2023-01-30."},{"key":"e_1_3_2_1_28_1","unstructured":"CEN-CENELEC. 2021. Joint Technical Committee 21 (JTC 21) \u2018Artificial Intelligence\u2019. https:\/\/www.cencenelec.eu\/areas-of-work\/cen-cenelec-topics\/artificial-intelligence\/ Accessed on: 2023-01-30."},{"key":"e_1_3_2_1_29_1","unstructured":"Central Digital & Data Office & Centre for Data Ethics & Innovation. 2021. Algorithmic Transparency Recording Standard - Guidance. https:\/\/www.gov.uk\/government\/publications\/algorithmic-transparency-template"},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","unstructured":"Supriyo Chakraborty Richard Tomsett Ramya Raghavendra Daniel Harborne Moustafa Alzantot Federico Cerutti Mani Srivastava Alun Preece Simon Julier Raghuveer M. Rao Troy D. Kelley Dave Braines Murat Sensoy Christopher J. Willis and Prudhvi Gurram. 2017. Interpretability of deep learning models: A survey of results. In 2017 IEEE SmartWorld Ubiquitous Intelligence & Computing Advanced & Trusted Computed Scalable Computing & Communications Cloud & Big Data Computing Internet of People and Smart City Innovation (SmartWorld\/SCALCOM\/UIC\/ATC\/CBDCom\/IOP\/SCI). IEEE San Francisco CA 1\u20136. https:\/\/doi.org\/10.1109\/UIC-ATC.2017.8397411","DOI":"10.1109\/UIC-ATC.2017.8397411"},{"key":"e_1_3_2_1_31_1","volume-title":"L-Shapley and C-Shapley: Efficient Model Interpretation for Structured Data. In 7th International Conference on Learning Representations, ICLR 2019","author":"Chen Jianbo","year":"2019","unstructured":"Jianbo Chen, Le Song, Martin J. Wainwright, and Michael I. Jordan. 2019. L-Shapley and C-Shapley: Efficient Model Interpretation for Structured Data. In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019. OpenReview.net. https:\/\/openreview.net\/forum?id=S1E3Ko09F7"},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1145\/3397481.3450644"},{"key":"e_1_3_2_1_33_1","unstructured":"European Commission. 2018. Communication Artificial Intelligence for Europe: Shaping Europe\u2019s Digital Future. https:\/\/digital-strategy.ec.europa.eu\/en\/library\/communication-artificial-intelligence-europe"},{"key":"e_1_3_2_1_34_1","volume-title":"White paper on artificial intelligence: A European approach to excellence and trust","author":"European Commission","year":"2020","unstructured":"European Commission. 2020. White paper on artificial intelligence: A European approach to excellence and trust. Brussels, 1st edn. European Commission, Brussels (2020). https:\/\/commission.europa.eu\/publications\/white-paper-artificial-intelligence-european-approach-excellence-and-trust_en","edition":"1"},{"key":"e_1_3_2_1_35_1","unstructured":"European Commission. 2021. Communication on Fostering a European approach to Artificial Intelligence. https:\/\/digital-strategy.ec.europa.eu\/en\/library\/communication-fostering-european-approach-artificial-intelligence"},{"key":"e_1_3_2_1_36_1","unstructured":"European Commission. 2021. Coordinated Plan on Artificial Intelligence 2021 Review. https:\/\/digital-strategy.ec.europa.eu\/en\/library\/coordinated-plan-artificial-intelligence-2021-review"},{"key":"e_1_3_2_1_37_1","volume-title":"Rolling Plan for ICT Standardisation","author":"European Commission","year":"2022","unstructured":"European Commission. 2022. Rolling Plan for ICT Standardisation 2022. https:\/\/joinup.ec.europa.eu\/collection\/rolling-plan-ict-standardisation\/rolling-plan-2022"},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1002\/widm.1391"},{"key":"e_1_3_2_1_40_1","volume-title":"Rodica Tirtea, and Stefan Schiffner.","author":"Danezis George","year":"2015","unstructured":"George Danezis, Josep Domingo-Ferrer, Marit Hansen, Jaap-Henk Hoepman, Daniel Le M\u00e9tayer, Rodica Tirtea, and Stefan Schiffner. 2015. Privacy and Data Protection by Design - from policy to engineering. CoRR abs\/1501.03726 (2015). arXiv:1501.03726http:\/\/arxiv.org\/abs\/1501.03726"},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.aacl-main.46"},{"key":"e_1_3_2_1_42_1","volume-title":"Opportunities and Challenges in Explainable Artificial Intelligence (XAI): A Survey. CoRR abs\/2006.11371","author":"Das Arun","year":"2020","unstructured":"Arun Das and Paul Rad. 2020. Opportunities and Challenges in Explainable Artificial Intelligence (XAI): A Survey. CoRR abs\/2006.11371 (2020). arXiv:2006.11371https:\/\/arxiv.org\/abs\/2006.11371"},{"key":"e_1_3_2_1_43_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.giq.2021.101666"},{"key":"e_1_3_2_1_44_1","doi-asserted-by":"publisher","DOI":"10.1145\/3461778.3462131"},{"key":"e_1_3_2_1_45_1","doi-asserted-by":"publisher","DOI":"10.53292\/208f5901.ff492fb3"},{"key":"e_1_3_2_1_46_1","doi-asserted-by":"publisher","DOI":"10.1109\/MSP.2018.2701152"},{"key":"e_1_3_2_1_47_1","volume-title":"Riedl","author":"Ehsan Upol","year":"2021","unstructured":"Upol Ehsan, Samir Passi, Q. Vera Liao, Larry Chan, I-Hsiang Lee, Michael J. Muller, and Mark O. Riedl. 2021. The Who in Explainable AI: How AI Background Shapes Perceptions of AI Explanations. CoRR abs\/2107.13509 (2021). arXiv:2107.13509https:\/\/arxiv.org\/abs\/2107.13509"},{"key":"e_1_3_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-60117-1_33"},{"key":"e_1_3_2_1_49_1","doi-asserted-by":"publisher","DOI":"10.1080\/03637751.2017.1375130"},{"key":"e_1_3_2_1_50_1","unstructured":"European Telecommunications Standards Institute (ETSI). 2019. ETSI Securing Artificial Intelligence (SAI) Committee. https:\/\/www.etsi.org\/technologies\/securing-artificial-intelligence"},{"key":"e_1_3_2_1_51_1","unstructured":"European Commission. 2016. Regulation (EU) 2016\/679 of the European Parliament and of the Council of 27 April 2016 on the protection of natural persons with regard to the processing of personal data and on the free movement of such data and repealing Directive 95\/46\/EC (General Data Protection Regulation) (Text with EEA relevance). https:\/\/eur-lex.europa.eu\/eli\/reg\/2016\/679\/oj"},{"key":"e_1_3_2_1_52_1","unstructured":"Experiential Networked Intelligence (ENI) European Telecommunications Standards Institute (ETSI). 2021. ETSI GS ENI 005 V2.1.1 - Experiential Networked Intelligence (ENI) System Architecture. https:\/\/www.etsi.org\/deliver\/etsi_gs\/ENI\/001_099\/005\/02.01.01_60\/gs_ENI005v020101p.pdf"},{"key":"e_1_3_2_1_53_1","unstructured":"Securing Artificial Intelligence (SAI) European Telecommunications Standards Institute (ETSI). 2023. DGR\/SAI-007\u2019 Work Item: Explicability and transparency of AI processing. https:\/\/portal.etsi.org\/webapp\/WorkProgram\/Report_WorkItem.asp?WKI_ID=63078"},{"key":"e_1_3_2_1_54_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-981-13-2874-9_4"},{"key":"e_1_3_2_1_55_1","doi-asserted-by":"publisher","DOI":"10.1007\/s13347-019-00354-x"},{"key":"e_1_3_2_1_56_1","volume-title":"Artificial Intelligence and Democratic Values Index -","author":"Center for AI and Digital Policy (CAIDP). 2022.","year":"2022","unstructured":"Center for AI and Digital Policy (CAIDP). 2022. Artificial Intelligence and Democratic Values Index - February, 2022. https:\/\/caidp.org\/reports\/aidv-2021\/"},{"key":"e_1_3_2_1_57_1","unstructured":"Department for Digital Culture and Media & Sport of the United Kingdom. 2019. National Data Strategy. https:\/\/www.gov.uk\/guidance\/national-data-strategy"},{"key":"e_1_3_2_1_58_1","volume-title":"Department for Business, and Energy & Industrial Strategy of the United Kingdom.","author":"Department for Digital, Culture","year":"2019","unstructured":"Department for Digital, Culture, Media & Sport, Department for Business, and Energy & Industrial Strategy of the United Kingdom. 2019. AI Sector Deal - Policy paper. https:\/\/www.gov.uk\/government\/publications\/artificial-intelligence-sector-deal\/ai-sector-deal"},{"key":"e_1_3_2_1_59_1","volume-title":"Department for Business, Energy & Industrial Strategy, and Office for Artificial Intelligence of the United Kingdom.","author":"Department for Digital, Culture","year":"2022","unstructured":"Department for Digital, Culture, Media & Sport, Department for Business, Energy & Industrial Strategy, and Office for Artificial Intelligence of the United Kingdom. 2022. Establishing a pro-innovation approach to regulating AI - Policy paper presented to UK Parliament. https:\/\/www.gov.uk\/government\/publications\/establishing-a-pro-innovation-approach-to-regulating-ai\/establishing-a-pro-innovation-approach-to-regulating-ai-policy-statement#fnref:35"},{"key":"e_1_3_2_1_60_1","volume-title":"Department for Business, Energy & Industrial Strategy, and Office for Artificial Intelligence of the United Kingdom.","author":"Department for Digital, Culture","year":"2022","unstructured":"Department for Digital, Culture, Media & Sport, Department for Business, Energy & Industrial Strategy, and Office for Artificial Intelligence of the United Kingdom. 2022. National AI Strategy - AI Action Plan. https:\/\/www.gov.uk\/government\/publications\/national-ai-strategy-ai-action-plan\/national-ai-strategy-ai-action-plan"},{"key":"e_1_3_2_1_61_1","unstructured":"Stanford Institute for Human-Centered Artificial Intelligence (HAI). 2022. The 2022 AI Index Report - Measuring trends in Artificial Intelligence. https:\/\/aiindex.stanford.edu\/report\/"},{"key":"e_1_3_2_1_62_1","unstructured":"Standard for XAI eXplainable AI Working Group IEEE Computational Intelligence Society\/ Standards Committee (IEEE CIS\/SC\/XAI WG). 2024. IEEE CIS\/SC\/XAI WG P2976 - Standard for XAI \u2013 eXplainable Artificial Intelligence - for Achieving Clarity and Interoperability of AI Systems Design. https:\/\/standards.ieee.org\/ieee\/2976\/10522\/."},{"key":"e_1_3_2_1_63_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v33i01.33013681"},{"key":"e_1_3_2_1_64_1","doi-asserted-by":"publisher","DOI":"10.23919\/JSC.2021.0018"},{"key":"e_1_3_2_1_65_1","first-page":"25","article-title":"Engineering privacy by design","volume":"14","author":"G\u00fcrses Seda","year":"2011","unstructured":"Seda G\u00fcrses, Carmela Troncoso, and Claudia Diaz. 2011. Engineering privacy by design. Computers, Privacy & Data Protection 14, 3 (2011), 25. https:\/\/software.imdea.org\/ carmela.troncoso\/papers\/Gurses-CPDP11.pdf","journal-title":"Computers, Privacy & Data Protection"},{"key":"e_1_3_2_1_66_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-04083-2_17"},{"key":"e_1_3_2_1_67_1","doi-asserted-by":"publisher","DOI":"10.1145\/3491102.3517650"},{"key":"e_1_3_2_1_68_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.491"},{"key":"e_1_3_2_1_69_1","doi-asserted-by":"publisher","DOI":"10.1145\/3284751.3284760"},{"key":"e_1_3_2_1_70_1","doi-asserted-by":"publisher","DOI":"10.1007\/s43681-020-00008-1"},{"key":"e_1_3_2_1_71_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-99740-7_1"},{"key":"e_1_3_2_1_72_1","doi-asserted-by":"publisher","DOI":"10.1007\/s00146-021-01267-0"},{"key":"e_1_3_2_1_73_1","unstructured":"Intelligent Transporation Systems IEEE Vehicular Technology Society (IEEE VT\/ITS). 2022. IEEE 7001-2021 - Standard for Transparency of Autonomous Systems. https:\/\/standards.ieee.org\/ieee\/7001\/6929\/."},{"key":"e_1_3_2_1_74_1","unstructured":"Information Commissioner\u2019s Office (ICO) of the United Kingdom and The Alan Turing Institute. 2019. Project ExplAIn - Interim Report. https:\/\/ico.org.uk\/media\/about-the-ico\/documents\/2615039\/project-explain-20190603.pdf"},{"key":"e_1_3_2_1_75_1","unstructured":"Information Commissioner\u2019s Office (ICO) of the United Kingdom and The Alan Turing Institute. 2020. Explaining decisions made with AI. https:\/\/ico.org.uk\/for-organisations\/guide-to-data-protection\/key-dp-themes\/explaining-decisions-made-with-ai\/"},{"key":"e_1_3_2_1_76_1","unstructured":"JTC 1 \/SC 42 Artificial Intelligence International Standards Association (ISO). 2017. ISO\/IEC JTC 1\/SC 42 - Artificial intelligence Committee. https:\/\/www.iso.org\/committee\/6794475.html."},{"key":"e_1_3_2_1_77_1","unstructured":"Securing Artificial Intelligence (SAI) International Standards Association (ISO). 2023. ISO\/IEC AWI TS 6254 -Information technology \u2014 Artificial intelligence \u2014 Objectives and approaches for explainability of ML models and AI systems. https:\/\/www.iso.org\/standard\/82148.html."},{"key":"e_1_3_2_1_78_1","unstructured":"International Standards Association (ISO). 2020. ISO\/IEC TR 24028:2020 - Information technology \u2014 Artificial intelligence \u2014 Overview of trustworthiness in artificial intelligence. https:\/\/www.iso.org\/standard\/77608.html."},{"key":"e_1_3_2_1_79_1","unstructured":"International Standards Association (ISO). 2025. ISO\/IEC AWI 12792 - Information technology \u2014 Artificial intelligence \u2014 Transparency taxonomy of AI systems. https:\/\/www.iso.org\/standard\/84111.html."},{"key":"e_1_3_2_1_80_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445941"},{"key":"e_1_3_2_1_81_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v36i11.21452"},{"key":"e_1_3_2_1_82_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11023-017-9417-6"},{"key":"e_1_3_2_1_83_1","unstructured":"Kate Kaye. 2022. This Senate bill would force companies to audit AI used for housing and loans. https:\/\/www.protocol.com\/enterprise\/revised-algorithmic-accountability-bill-ai"},{"key":"e_1_3_2_1_84_1","doi-asserted-by":"publisher","DOI":"10.1145\/3461702.3462605"},{"key":"e_1_3_2_1_85_1","doi-asserted-by":"publisher","DOI":"10.1145\/3375627.3375835"},{"key":"e_1_3_2_1_86_1","volume-title":"The Disagreement Problem in Explainable Machine Learning: A Practitioner\u2019s Perspective. CoRR abs\/2202.01602","author":"Krishna Satyapriya","year":"2022","unstructured":"Satyapriya Krishna, Tessa Han, Alex Gu, Javin Pombra, Shahin Jabbari, Steven Wu, and Himabindu Lakkaraju. 2022. The Disagreement Problem in Explainable Machine Learning: A Practitioner\u2019s Perspective. CoRR abs\/2202.01602 (2022). arXiv:2202.01602https:\/\/arxiv.org\/abs\/2202.01602"},{"key":"e_1_3_2_1_87_1","volume-title":"An Evaluation of the Human-Interpretability of Explanation. CoRR abs\/1902.00006","author":"Lage Isaac","year":"2019","unstructured":"Isaac Lage, Emily Chen, Jeffrey He, Menaka Narayanan, Been Kim, Sam Gershman, and Finale Doshi-Velez. 2019. An Evaluation of the Human-Interpretability of Explanation. CoRR abs\/1902.00006 (2019). arXiv:1902.00006http:\/\/arxiv.org\/abs\/1902.00006"},{"key":"e_1_3_2_1_88_1","doi-asserted-by":"publisher","DOI":"10.1609\/hcomp.v7i1.5280"},{"key":"e_1_3_2_1_89_1","doi-asserted-by":"publisher","DOI":"10.1117\/12.2520149"},{"key":"e_1_3_2_1_90_1","doi-asserted-by":"publisher","DOI":"10.1109\/MSP.2022.3153277"},{"key":"e_1_3_2_1_91_1","doi-asserted-by":"publisher","DOI":"10.1145\/3313831.3376590"},{"key":"e_1_3_2_1_92_1","doi-asserted-by":"publisher","DOI":"10.1609\/hcomp.v10i1.21995"},{"key":"e_1_3_2_1_93_1","doi-asserted-by":"publisher","DOI":"10.1145\/3313831.3376445"},{"key":"e_1_3_2_1_94_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.clsr.2019.05.002"},{"key":"e_1_3_2_1_95_1","doi-asserted-by":"publisher","DOI":"10.1038\/s42256-019-0114-4"},{"key":"e_1_3_2_1_96_1","doi-asserted-by":"publisher","DOI":"10.1007\/s13347-020-00405-8"},{"key":"e_1_3_2_1_97_1","doi-asserted-by":"publisher","DOI":"10.1145\/3397481.3450689"},{"key":"e_1_3_2_1_98_1","doi-asserted-by":"publisher","DOI":"10.1145\/3387166"},{"key":"e_1_3_2_1_99_1","doi-asserted-by":"publisher","DOI":"10.1007\/s00146-021-01308-8"},{"key":"e_1_3_2_1_100_1","unstructured":"Markus Mueck Raymond Forbes Scott Cadzow Suno Wood and European Telecommunications Standards Institute (ETSI) Gazis Evangelos. 2022. ETSI Activities in the field of Artificial Intelligence - Preparing the implementation of the European AI Act. https:\/\/www.etsi.org\/images\/files\/ETSIWhitePapers\/ETSI-WP52-ETSI-activities-in-the-field-of-AI.pdf. (2022)."},{"key":"e_1_3_2_1_101_1","doi-asserted-by":"publisher","DOI":"10.1145\/3359221"},{"key":"e_1_3_2_1_102_1","unstructured":"IMCO-LIBE Committees of European Parliament. 2022-04-20. Draft Report on the proposal for a regulation of the European Parliament and of the Council on harmonised rules on Artificial Intelligence (Artificial Intelligence Act) and amending certain Union Legislative Acts."},{"key":"e_1_3_2_1_103_1","unstructured":"JURI Committee of European Parliament. 2022-09-12. Opinion of the Committee on Legal Affairs for the Committee on the Internal Market and Consumer Protection and the Committee on Civil Liberties Justice and Home Affairs on the proposal for a regulation of the European Parliament and of the Council laying down harmonised rules on artificial intelligence (Artificial Intelligence Act) and amending certain Union Legislative Acts."},{"key":"e_1_3_2_1_104_1","unstructured":"Office of Science and Technology Policy of the White House of United States of America. 2022. Blueprint for an AI Bill of Rights: Making Automated Systems Work for the American People. https:\/\/www.whitehouse.gov\/ostp\/ai-bill-of-rights\/"},{"key":"e_1_3_2_1_105_1","volume-title":"Artificial Intelligence: AI Fundamental Research - Explainability. https:\/\/www.nist.gov\/artificial-intelligence\/ai-fundamental-research-explainability","author":"National Institute of Standards and Technology (NIST) of United States of America.","year":"2021","unstructured":"National Institute of Standards and Technology (NIST) of United States of America. 2021. Artificial Intelligence: AI Fundamental Research - Explainability. https:\/\/www.nist.gov\/artificial-intelligence\/ai-fundamental-research-explainability"},{"key":"e_1_3_2_1_106_1","volume-title":"Leadership in AI: A Plan for Federal Engagement in Developing Technical Standards and Related Tools - Prepared in response to Executive Order 13859 Submitted on","author":"National Institute of Standards and Technology (NIST) of United States of America. 2021. U.S.","year":"2019","unstructured":"National Institute of Standards and Technology (NIST) of United States of America. 2021. U.S. Leadership in AI: A Plan for Federal Engagement in Developing Technical Standards and Related Tools - Prepared in response to Executive Order 13859 Submitted on August 9, 2019. https:\/\/www.nist.gov\/news-events\/news\/2019\/08\/plan-outlines-priorities-federal-agency-engagement-ai-standards-development"},{"key":"e_1_3_2_1_107_1","doi-asserted-by":"publisher","unstructured":"National Institute of Standards and Technology (NIST) of United States of America. 2023. Risk Management Framework v1.0. https:\/\/doi.org\/10.6028\/NIST.AI.100-1","DOI":"10.6028\/NIST.AI.100-1"},{"key":"e_1_3_2_1_108_1","volume-title":"Maintaining American Leadership in Artificial Intelligence - Executive Order 13859 of","author":"Executive Office of the President of United States of America. 2019.","year":"2019","unstructured":"Executive Office of the President of United States of America. 2019. Maintaining American Leadership in Artificial Intelligence - Executive Order 13859 of February 11, 2019. https:\/\/www.federalregister.gov\/d\/2019-02544"},{"key":"e_1_3_2_1_109_1","unstructured":"Executive Office of the President of United States of America Office of Management and Budget. 2020. Guidance for Regulation of Artificial Intelligence Applications. https:\/\/www.whitehouse.gov\/wp-content\/uploads\/2020\/11\/M-21-06.pdf"},{"key":"e_1_3_2_1_110_1","unstructured":"Information Commissioner\u2019s Office (ICO) of the United Kingdom. 2020. Guidance on AI and data protection. https:\/\/ico.org.uk\/for-organisations\/guide-to-data-protection\/key-dp-themes\/guidance-on-artificial-intelligence-and-data-protection\/"},{"key":"e_1_3_2_1_111_1","unstructured":"Information Commissioner\u2019s Office (ICO) of the United Kingdom. 2020. Guidance on the AI auditing framework: Draft guidance for consultation. https:\/\/ico.org.uk\/media\/2617219\/guidance-on-the-ai-auditing-framework-draft-for-consultation.pdf"},{"key":"e_1_3_2_1_112_1","volume-title":"Data Protection Act","author":"Parliament of the United Kingdom. 2018.","year":"2018","unstructured":"Parliament of the United Kingdom. 2018. Data Protection Act 2018. https:\/\/www.legislation.gov.uk\/ukpga\/2018\/12\/contents\/enacted"},{"key":"e_1_3_2_1_113_1","unstructured":"Federal Trade Commission of the United States of America. 2022. Trade Regulation Rule on Commercial Surveillance and Data Security (Billing Code: 6750-01-P). https:\/\/www.federalregister.gov\/documents\/2022\/08\/22\/2022-17752\/trade-regulation-rule-on-commercial-surveillance-and-data-security"},{"key":"e_1_3_2_1_114_1","unstructured":"Defense Advanced Research Projects Agency of United States of America. 2016-08-050. Federal Contract Opportunity for Explainable Artificial Intelligence (XAI) DARPA-BAA-16-53. https:\/\/www.darpa.mil\/attachments\/DARPA-BAA-16-53.pdf"},{"key":"e_1_3_2_1_115_1","unstructured":"Cabinet Office & Central Digital & Data Office & Office for Artificial Intelligence. 2021. Ethics Transparency and Accountability Framework for Automated Decision-Making - Guidance. https:\/\/www.gov.uk\/government\/publications\/ethics-transparency-and-accountability-framework-for-automated-decision-making\/ethics-transparency-and-accountability-framework-for-automated-decision-making"},{"key":"e_1_3_2_1_116_1","unstructured":"High-Level Expert Group on Artificial Intelligence. 2019. Ethics Guidelines for Trustworthy AI. https:\/\/digital-strategy.ec.europa.eu\/en\/library\/ethics-guidelines-trustworthy-ai"},{"key":"e_1_3_2_1_117_1","unstructured":"High-Level Expert Group on Artificial Intelligence. 2020. The Assessment List for Trustworthy Artificial Intelligence (ALTAI). https:\/\/ec.europa.eu\/futurium\/en\/ai-alliance-consultation"},{"key":"e_1_3_2_1_118_1","unstructured":"National Security Commission on Artificial Intelligence of United States of America. 2021. Final Report. https:\/\/www.nscai.gov\/2021-final-report\/"},{"key":"e_1_3_2_1_119_1","unstructured":"European Parliament and Council. 2020-12-15. Proposal for a regulation of the European Parliament and of the Council on a Single Market For Digital Services (Digital Services Act) and amending Directive 2000\/31\/EC."},{"key":"e_1_3_2_1_120_1","unstructured":"European Parliament and Council. 2021-04-21. Proposal for a Regulation of the European Parliament and of the Council laying down harmonised rules on Artificial Intelligence (Artificial Intelligence Act) and amending certain Union legislative acts."},{"key":"e_1_3_2_1_121_1","unstructured":"European Parliament and Council. 2022-09-28. Proposal for a Directive of the European Parliament and of the Council on adapting non-contractual civil liability rules to artificial intelligence (AI Liability Directive). https:\/\/commission.europa.eu\/business-economy-euro\/doing-business-eu\/contract-rules\/digital-contracts\/liability-rules-artificial-intelligence_en"},{"key":"e_1_3_2_1_122_1","unstructured":"European Parliament and Council. 2022-11-25. Proposal for a Regulation of the European Parliament and of the Council laying down harmonised rules on artificial intelligence (Artificial Intelligence Act) and amending certain Union legislative acts - General Approach."},{"key":"e_1_3_2_1_123_1","doi-asserted-by":"publisher","DOI":"10.6028\/NIST.IR.8312"},{"key":"e_1_3_2_1_124_1","doi-asserted-by":"publisher","DOI":"10.1145\/3531146.3533200"},{"key":"e_1_3_2_1_125_1","doi-asserted-by":"publisher","DOI":"10.1145\/3449081"},{"key":"e_1_3_2_1_126_1","unstructured":"Marco Tulio Ribeiro Sameer Singh and Carlos Guestrin. 2016. Model-Agnostic Interpretability of Machine Learning. http:\/\/arxiv.org\/abs\/1606.05386 arXiv:1606.05386 [cs stat]."},{"key":"e_1_3_2_1_127_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11023-019-09509-3"},{"key":"e_1_3_2_1_128_1","doi-asserted-by":"publisher","DOI":"10.1038\/s42256-019-0048-x"},{"key":"e_1_3_2_1_129_1","volume-title":"Artificial Intelligence: A Modern Approach","author":"Russell Stuart","year":"2020","unstructured":"Stuart Russell and Peter Norvig. 2020. Artificial Intelligence: A Modern Approach (4th Edition). Pearson. http:\/\/aima.cs.berkeley.edu\/","edition":"4"},{"key":"e_1_3_2_1_130_1","doi-asserted-by":"publisher","DOI":"10.1145\/3411764.3445518"},{"key":"e_1_3_2_1_131_1","doi-asserted-by":"publisher","unstructured":"Wojciech Samek and Klaus-Robert M\u00fcller. 2019. Towards Explainable Artificial Intelligence. In Explainable AI: Interpreting Explaining and Visualizing Deep Learning Wojciech Samek Gr\u00e9goire Montavon Andrea Vedaldi Lars Kai Hansen and Klaus-Robert M\u00fcller (Eds.). Lecture Notes in Computer Science Vol. 11700. Springer 5\u201322. https:\/\/doi.org\/10.1007\/978-3-030-28954-6_1","DOI":"10.1007\/978-3-030-28954-6_1"},{"key":"e_1_3_2_1_132_1","doi-asserted-by":"publisher","DOI":"10.1145\/3470918"},{"key":"e_1_3_2_1_133_1","doi-asserted-by":"publisher","DOI":"10.1145\/3498366.3505816"},{"key":"e_1_3_2_1_134_1","volume-title":"Proceedings of the 37th International Conference on Machine Learning, ICML 2020","volume":"9057","author":"Sixt Leon","year":"2020","unstructured":"Leon Sixt, Maximilian Granz, and Tim Landgraf. 2020. When Explanations Lie: Why Many Modified BP Attributions Fail. In Proceedings of the 37th International Conference on Machine Learning, ICML 2020, 13-18 July 2020, Virtual Event(Proceedings of Machine Learning Research, Vol. 119). PMLR, 9046\u20139057. http:\/\/proceedings.mlr.press\/v119\/sixt20a.html"},{"key":"e_1_3_2_1_135_1","doi-asserted-by":"publisher","DOI":"10.1145\/3375627.3375830"},{"key":"e_1_3_2_1_136_1","doi-asserted-by":"publisher","DOI":"10.14763\/2021.3.1574"},{"key":"e_1_3_2_1_137_1","unstructured":"Algorithmic Bias Working Group IEEE Computer Society (IEEE C\/S2ESC\/ALGB-WG) Software & Systems Engineering Standards Committee. 2017. IEEE P7003 - Algorithmic Bias Considerations. https:\/\/standards.ieee.org\/ieee\/7003\/6980\/."},{"key":"e_1_3_2_1_138_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3372870"},{"key":"e_1_3_2_1_139_1","doi-asserted-by":"publisher","DOI":"10.3390\/j5010010"},{"key":"e_1_3_2_1_140_1","doi-asserted-by":"publisher","DOI":"10.1145\/3397481.3450662"},{"key":"e_1_3_2_1_141_1","doi-asserted-by":"publisher","DOI":"10.1093\/idpl"},{"key":"e_1_3_2_1_142_1","doi-asserted-by":"publisher","DOI":"10.1515\/9789048550180-016"},{"key":"e_1_3_2_1_143_1","volume-title":"BIM: Towards quantitative evaluation of interpretability methods with ground truth. arXiv preprint arXiv:1907.09701","author":"Yang Mengjiao","year":"2019","unstructured":"Mengjiao Yang and Been Kim. 2019. BIM: Towards quantitative evaluation of interpretability methods with ground truth. arXiv preprint arXiv:1907.09701 (2019)."},{"key":"e_1_3_2_1_144_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-022-01657-x"},{"key":"e_1_3_2_1_145_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00920"},{"key":"e_1_3_2_1_146_1","doi-asserted-by":"publisher","DOI":"10.1561\/1500000066"}],"event":{"name":"FAccT '23: the 2023 ACM Conference on Fairness, Accountability, and Transparency","location":"Chicago IL USA","acronym":"FAccT '23"},"container-title":["2023 ACM Conference on Fairness Accountability and Transparency"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3593013.3594074","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3593013.3594074","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T16:37:18Z","timestamp":1750178238000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3593013.3594074"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,6,12]]},"references-count":144,"alternative-id":["10.1145\/3593013.3594074","10.1145\/3593013"],"URL":"https:\/\/doi.org\/10.1145\/3593013.3594074","relation":{},"subject":[],"published":{"date-parts":[[2023,6,12]]},"assertion":[{"value":"2023-06-12","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}