{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T17:52:23Z","timestamp":1775065943895,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":120,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,6,12]],"date-time":"2023-06-12T00:00:00Z","timestamp":1686528000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,6,12]]},"DOI":"10.1145\/3593013.3594069","type":"proceedings-article","created":{"date-parts":[[2023,6,12]],"date-time":"2023-06-12T14:40:46Z","timestamp":1686580846000},"page":"1139-1150","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":105,"title":["The role of explainable AI in the context of the AI Act"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-6552-787X","authenticated-orcid":false,"given":"Cecilia","family":"Panigutti","sequence":"first","affiliation":[{"name":"European Commission, Joint Research Centre (JRC), Ispra, Italy"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1987-5707","authenticated-orcid":false,"given":"Ronan","family":"Hamon","sequence":"additional","affiliation":[{"name":"European Commission, Joint Research Centre (JRC), Ispra, Italy"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9811-9397","authenticated-orcid":false,"given":"Isabelle","family":"Hupont","sequence":"additional","affiliation":[{"name":"European Commission, Joint Research Centre (JRC), Sevilla, Spain"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2433-7110","authenticated-orcid":false,"given":"David","family":"Fernandez Llorca","sequence":"additional","affiliation":[{"name":"European Commission, Joint Research Centre (JRC), Sevilla, Spain"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7999-3241","authenticated-orcid":false,"given":"Delia","family":"Fano Yela","sequence":"additional","affiliation":[{"name":"European Commission, Joint Research Centre (JRC), Sevilla, Spain"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0452-6865","authenticated-orcid":false,"given":"Henrik","family":"Junklewitz","sequence":"additional","affiliation":[{"name":"European Commission, Joint Research Centre (JRC), Ispra, Italy"}]},{"ORCID":"https:\/\/orcid.org\/0009-0004-6162-5714","authenticated-orcid":false,"given":"Salvatore","family":"Scalzo","sequence":"additional","affiliation":[{"name":"European Commission, Belgium"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-5075-3523","authenticated-orcid":false,"given":"Gabriele","family":"Mazzini","sequence":"additional","affiliation":[{"name":"European Commission, Belgium"}]},{"ORCID":"https:\/\/orcid.org\/0009-0003-9916-8806","authenticated-orcid":false,"given":"Ignacio","family":"Sanchez","sequence":"additional","affiliation":[{"name":"European Commission, Joint Research Centre (JRC), Ispra, Italy"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9669-3459","authenticated-orcid":false,"given":"Josep","family":"Soler Garrido","sequence":"additional","affiliation":[{"name":"European Commission, Joint Research Centre (JRC), Sevilla, Spain"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4983-3989","authenticated-orcid":false,"given":"Emilia","family":"Gomez","sequence":"additional","affiliation":[{"name":"European Commission, Joint Research Centre (JRC), Sevilla, Spain"}]}],"member":"320","published-online":{"date-parts":[[2023,6,12]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"2016. Regulation (EU) 2016\/679 of the European Parliament and of the Council of 27 April 2016 on the Protection of Natural Persons with Regard to the Processing of Personal Data and on the Free Movement of Such Data and Repealing Directive 95\/46\/EC (General Data Protection Regulation)."},{"key":"e_1_3_2_1_2_1","unstructured":"2022. Draft standardisation request to the European Standardisation Organisations in support of safe and trustworthy artificial intelligence. https:\/\/ec.europa.eu\/docsroom\/documents\/52376"},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1145\/3173574.3174156"},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.5555\/3291125.3309612"},{"key":"e_1_3_2_1_5_1","volume-title":"Peeking inside the black-box: a survey on explainable artificial intelligence (XAI)","author":"Adadi Amina","year":"2018","unstructured":"Amina Adadi and Mohammed Berrada. 2018. Peeking inside the black-box: a survey on explainable artificial intelligence (XAI). IEEE access 6 (2018), 52138\u201352160."},{"key":"e_1_3_2_1_6_1","volume-title":"Sanity checks for saliency maps. Advances in neural information processing systems 31","author":"Adebayo Julius","year":"2018","unstructured":"Julius Adebayo, Justin Gilmer, Michael Muelly, Ian Goodfellow, Moritz Hardt, and Been Kim. 2018. Sanity checks for saliency maps. Advances in neural information processing systems 31 (2018)."},{"key":"e_1_3_2_1_7_1","first-page":"14822","article-title":"Characterizing the risk of fairwashing","volume":"34","author":"A\u00efvodji Ulrich","year":"2021","unstructured":"Ulrich A\u00efvodji, Hiromi Arai, S\u00e9bastien Gambs, and Satoshi Hara. 2021. Characterizing the risk of fairwashing. Advances in Neural Information Processing Systems 34 (2021), 14822\u201314834.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_8_1","volume-title":"On the robustness of interpretability methods. arXiv preprint arXiv:1806.08049","author":"Alvarez-Melis David","year":"2018","unstructured":"David Alvarez-Melis and Tommi S Jaakkola. 2018. On the robustness of interpretability methods. arXiv preprint arXiv:1806.08049 (2018)."},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ipm.2022.103225"},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.7717\/peerj-cs.479"},{"key":"e_1_3_2_1_11_1","volume-title":"International Conference on Machine Learning. PMLR, 314\u2013323","author":"Anders Christopher","year":"2020","unstructured":"Christopher Anders, Plamen Pasliev, Ann-Kathrin Dombrowski, Klaus-Robert M\u00fcller, and Pan Kessel. 2020. Fairwashing explanations with off-manifold detergent. In International Conference on Machine Learning. PMLR, 314\u2013323."},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1111\/rssb.12377"},{"key":"e_1_3_2_1_13_1","volume-title":"Assessing the (un) trustworthiness of saliency maps for localizing abnormalities in medical imaging. arXiv preprint arXiv:2008.02766","author":"Arun Nishanth","year":"2020","unstructured":"Nishanth Arun, Nathan Gaw, Praveer Singh, Ken Chang, Mehak Aggarwal, Bryan Chen, Katharina Hoebel, Sharut Gupta, Jay Patel, Mishka Gidwani, 2020. Assessing the (un) trustworthiness of saliency maps for localizing abnormalities in medical imaging. arXiv preprint arXiv:2008.02766 (2020)."},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1109\/FG.2018.00019"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1073\/pnas.1907375117"},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3375624"},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1145\/3531146.3533153"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1145\/3377325.3377498"},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1145\/3449287"},{"key":"e_1_3_2_1_20_1","volume-title":"Conference on fairness, accountability and transparency. PMLR, 77\u201391","author":"Buolamwini Joy","year":"2018","unstructured":"Joy Buolamwini and Timnit Gebru. 2018. Gender shades: Intersectional accuracy disparities in commercial gender classification. In Conference on fairness, accountability and transparency. PMLR, 77\u201391."},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1145\/2783258.2788613"},{"key":"e_1_3_2_1_22_1","volume-title":"This looks like that: deep learning for interpretable image recognition. Advances in neural information processing systems 32","author":"Chen Chaofan","year":"2019","unstructured":"Chaofan Chen, Oscar Li, Daniel Tao, Alina Barnett, Cynthia Rudin, and Jonathan K Su. 2019. This looks like that: deep learning for interpretable image recognition. Advances in neural information processing systems 32 (2019)."},{"key":"e_1_3_2_1_23_1","unstructured":"European Commission. 2020. White Paper: On Artificial Intelligence - A European Approach to Excellence and Trust. https:\/\/commission.europa.eu\/publications\/white-paper-artificial-intelligence-european-approach-excellence-and-trust_en"},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1002\/widm.1391"},{"key":"e_1_3_2_1_25_1","volume-title":"Congress. 2022","author":"US","year":"2022","unstructured":"US Congress. 2022. Algorithmic Accountability Act of 2022. https:\/\/www.congress.gov\/bill\/117th-congress\/senate-bill\/3572\/text"},{"key":"e_1_3_2_1_26_1","volume-title":"Translation: Internet Information Service Algorithmic Recommendation Management Provisions \u2013 Effective","author":"Creemers Rogier","year":"2022","unstructured":"Rogier Creemers, Graham Webster, and Helen Toner. 2022. Translation: Internet Information Service Algorithmic Recommendation Management Provisions \u2013 Effective March 1, 2022. https:\/\/digichina.stanford.edu\/work\/translation-internet-information-service-algorithmic-recommendation-management-provisions-effective-march-1-2022\/"},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00482"},{"key":"e_1_3_2_1_28_1","unstructured":"Botty Dimanov Umang Bhatt Mateja Jamnik and Adrian Weller. 2020. You Shouldn\u2019t Trust Me: Learning Models Which Conceal Unfairness From Multiple Explanation Methods.. In SafeAI@ AAAI."},{"key":"e_1_3_2_1_29_1","volume-title":"Explanations can be manipulated and geometry is to blame. Advances in Neural Information Processing Systems 32","author":"Dombrowski Ann-Kathrin","year":"2019","unstructured":"Ann-Kathrin Dombrowski, Maximillian Alber, Christopher Anders, Marcel Ackermann, Klaus-Robert M\u00fcller, and Pan Kessel. 2019. Explanations can be manipulated and geometry is to blame. Advances in Neural Information Processing Systems 32 (2019)."},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2021.108194"},{"key":"e_1_3_2_1_31_1","volume-title":"Towards a rigorous science of interpretable machine learning. arXiv preprint arXiv:1702.08608","author":"Doshi-Velez Finale","year":"2017","unstructured":"Finale Doshi-Velez and Been Kim. 2017. Towards a rigorous science of interpretable machine learning. arXiv preprint arXiv:1702.08608 (2017)."},{"key":"e_1_3_2_1_32_1","volume-title":"The who in explainable ai: How ai background shapes perceptions of ai explanations. arXiv preprint arXiv:2107.13509","author":"Ehsan Upol","year":"2021","unstructured":"Upol Ehsan, Samir Passi, Q Vera Liao, Larry Chan, I Lee, Michael Muller, Mark O Riedl, 2021. The who in explainable ai: How ai background shapes perceptions of ai explanations. arXiv preprint arXiv:2107.13509 (2021)."},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"crossref","unstructured":"Malin Eiband Daniel Buschek Alexander Kremer and Heinrich Hussmann. 2019. The impact of placebic explanations on trust in intelligent systems. In Extended abstracts of the 2019 CHI conference on human factors in computing systems. 1\u20136.","DOI":"10.1145\/3290607.3312787"},{"key":"e_1_3_2_1_34_1","volume-title":"Alzheimer\u2019s Disease Neuroimaging Initiative (ADNI","author":"Eitel Fabian","year":"2019","unstructured":"Fabian Eitel, Kerstin Ritter, Alzheimer\u2019s Disease Neuroimaging Initiative (ADNI, 2019. Testing the robustness of attribution methods for convolutional neural networks in MRI-based Alzheimer\u2019s disease classification. In Interpretability of Machine Intelligence in Medical Image Computing and Multimodal Learning for Clinical Decision Support. Springer, 3\u201311."},{"key":"e_1_3_2_1_35_1","volume-title":"What the face reveals: Basic and applied studies of spontaneous expression using the Facial Action Coding System (FACS)","author":"Ekman Rosenberg","unstructured":"Rosenberg Ekman. 1997. What the face reveals: Basic and applied studies of spontaneous expression using the Facial Action Coding System (FACS). Oxford University Press, USA."},{"key":"e_1_3_2_1_39_1","unstructured":"European Commission. 2021. Proposal for a Regulation laying down harmonised rules on Artificial Intelligence and amending certain union legislative acts. https:\/\/eur-lex.europa.eu\/legal-content\/EN\/TXT\/?uri=celex%3A52021PC0206"},{"key":"e_1_3_2_1_40_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11948-020-00276-4"},{"key":"e_1_3_2_1_41_1","first-page":"376","article-title":"Reasoned A(I)Administration: Explanation Requirements in EU Law and the Automation of Public Administration","volume":"47","author":"Fink M.","year":"2022","unstructured":"M. Fink and M. Finck. 2022. Reasoned A(I)Administration: Explanation Requirements in EU Law and the Automation of Public Administration. European Law Review 47, 3 (2022), 376\u2013392.","journal-title":"European Law Review"},{"key":"e_1_3_2_1_42_1","volume-title":"Isaac S Kohane, and Andrew L Beam.","author":"Finlayson Samuel G","year":"2018","unstructured":"Samuel G Finlayson, Hyung Won Chung, Isaac S Kohane, and Andrew L Beam. 2018. Adversarial attacks against medical deep learning systems. arXiv preprint arXiv:1804.05296 (2018)."},{"key":"e_1_3_2_1_43_1","first-page":"1","article-title":"All Models are Wrong, but Many are Useful: Learning a Variable\u2019s Importance by Studying an Entire Class of Prediction Models Simultaneously.J","volume":"20","author":"Fisher Aaron","year":"2019","unstructured":"Aaron Fisher, Cynthia Rudin, and Francesca Dominici. 2019. All Models are Wrong, but Many are Useful: Learning a Variable\u2019s Importance by Studying an Entire Class of Prediction Models Simultaneously.J. Mach. Learn. Res. 20, 177 (2019), 1\u201381.","journal-title":"Mach. Learn. Res."},{"key":"e_1_3_2_1_44_1","unstructured":"G20. 2019. G20 Ministerial Statement on Trade and Digital Economy. https:\/\/wp.oecd.ai\/app\/uploads\/2021\/06\/G20-AI-Principles.pdf"},{"key":"e_1_3_2_1_45_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v33i01.33013681"},{"key":"e_1_3_2_1_46_1","doi-asserted-by":"publisher","DOI":"10.1136\/amiajnl-2011-000089"},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.artint.2020.103428"},{"key":"e_1_3_2_1_48_1","volume-title":"A survey of methods for explaining black box models. ACM computing surveys (CSUR) 51, 5","author":"Guidotti Riccardo","year":"2018","unstructured":"Riccardo Guidotti, Anna Monreale, Salvatore Ruggieri, Franco Turini, Fosca Giannotti, and Dino Pedreschi. 2018. A survey of methods for explaining black box models. ACM computing surveys (CSUR) 51, 5 (2018), 1\u201342."},{"key":"e_1_3_2_1_49_1","volume-title":"Explainable artificial intelligence (xai)","author":"Gunning David","year":"2017","unstructured":"David Gunning. 2017. Explainable artificial intelligence (xai). Defense Advanced Research Projects Agency (DARPA), nd Web 2 (2017)."},{"key":"e_1_3_2_1_50_1","doi-asserted-by":"publisher","DOI":"10.1109\/MCI.2021.3129960"},{"key":"e_1_3_2_1_51_1","first-page":"3650","article-title":"The out-of-distribution problem in explainability and search methods for feature importance explanations","volume":"34","author":"Hase Peter","year":"2021","unstructured":"Peter Hase, Harry Xie, and Mohit Bansal. 2021. The out-of-distribution problem in explainability and search methods for feature importance explanations. Advances in Neural Information Processing Systems 34 (2021), 3650\u20133666.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_52_1","volume-title":"Benchmarking neural network robustness to common corruptions and perturbations. arXiv preprint arXiv:1903.12261","author":"Hendrycks Dan","year":"2019","unstructured":"Dan Hendrycks and Thomas Dietterich. 2019. Benchmarking neural network robustness to common corruptions and perturbations. arXiv preprint arXiv:1903.12261 (2019)."},{"key":"e_1_3_2_1_53_1","volume-title":"Metrics for explainable AI: Challenges and prospects. arXiv preprint arXiv:1812.04608","author":"Hoffman Robert R","year":"2018","unstructured":"Robert R Hoffman, Shane T Mueller, Gary Klein, and Jordan Litman. 2018. Metrics for explainable AI: Challenges and prospects. arXiv preprint arXiv:1812.04608 (2018)."},{"key":"e_1_3_2_1_54_1","doi-asserted-by":"publisher","DOI":"10.1145\/3290605.3300809"},{"key":"e_1_3_2_1_55_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-04083-2_2"},{"key":"e_1_3_2_1_56_1","unstructured":"The White House. 2022. Blueprint for an AI Bill of Rights. https:\/\/www.whitehouse.gov\/ostp\/ai-bill-of-rights\/"},{"key":"e_1_3_2_1_57_1","volume-title":"Optimal sparse decision trees. Advances in Neural Information Processing Systems 32","author":"Hu Xiyang","year":"2019","unstructured":"Xiyang Hu, Cynthia Rudin, and Margo Seltzer. 2019. Optimal sparse decision trees. Advances in Neural Information Processing Systems 32 (2019)."},{"key":"e_1_3_2_1_58_1","volume-title":"The Landscape of Facial Processing Applications in the Context of the European AI Act and the Development of Trustworthy Systems. Nature Scientific Reports","author":"Hupont Isabelle","year":"2022","unstructured":"Isabelle Hupont, Song\u00fcl Tolan, Hatice Gunes, and Emilia G\u00f3mez. 2022. The Landscape of Facial Processing Applications in the Context of the European AI Act and the Development of Trustworthy Systems. Nature Scientific Reports (2022)."},{"key":"e_1_3_2_1_59_1","volume-title":"Attention is not explanation. arXiv preprint arXiv:1902.10186","author":"Jain Sarthak","year":"2019","unstructured":"Sarthak Jain and Byron C Wallace. 2019. Attention is not explanation. arXiv preprint arXiv:1902.10186 (2019)."},{"key":"e_1_3_2_1_60_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10115-019-01368-9"},{"key":"e_1_3_2_1_61_1","doi-asserted-by":"publisher","DOI":"10.15779\/Z38TD9N83H"},{"key":"e_1_3_2_1_62_1","doi-asserted-by":"publisher","DOI":"10.1145\/3313831.3376219"},{"key":"e_1_3_2_1_63_1","volume-title":"Explainable AI: Interpreting, Explaining and Visualizing Deep Learning","author":"Kindermans Pieter-Jan","unstructured":"Pieter-Jan Kindermans, Sara Hooker, Julius Adebayo, Maximilian Alber, Kristof T Sch\u00fctt, Sven D\u00e4hne, Dumitru Erhan, and Been Kim. 2019. The (un) reliability of saliency methods. In Explainable AI: Interpreting, Explaining and Visualizing Deep Learning. Springer, 267\u2013280."},{"key":"e_1_3_2_1_64_1","doi-asserted-by":"crossref","unstructured":"Anastasiya Kiseleva. 2021. Making AI\u2019s Transparency Transparent: notes on the EU Proposal for the AI Act. https:\/\/europeanlawblog.eu\/2021\/07\/29\/making-ais-transparency-transparent-notes-on-the-eu-proposal-for-the-ai-act\/","DOI":"10.21428\/9885764c.0440ae5f"},{"key":"e_1_3_2_1_65_1","doi-asserted-by":"publisher","DOI":"10.1038\/s42256-020-0216-z"},{"key":"e_1_3_2_1_66_1","volume-title":"Trust in automation: Designing for appropriate reliance. Human factors 46, 1","author":"Lee John D","year":"2004","unstructured":"John D Lee and Katrina A See. 2004. Trust in automation: Designing for appropriate reliance. Human factors 46, 1 (2004), 50\u201380."},{"key":"e_1_3_2_1_67_1","doi-asserted-by":"publisher","DOI":"10.1214\/15-AOAS848"},{"key":"e_1_3_2_1_68_1","doi-asserted-by":"publisher","DOI":"10.1145\/3411764.3445522"},{"key":"e_1_3_2_1_69_1","doi-asserted-by":"publisher","DOI":"10.1145\/3531146.3534628"},{"key":"e_1_3_2_1_70_1","volume-title":"The mythos of model interpretability: In machine learning, the concept of interpretability is both important and slippery.Queue 16, 3","author":"Lipton Zachary C","year":"2018","unstructured":"Zachary C Lipton. 2018. The mythos of model interpretability: In machine learning, the concept of interpretability is both important and slippery.Queue 16, 3 (2018), 31\u201357."},{"key":"e_1_3_2_1_71_1","volume-title":"Proceedings of machine learning research 151","author":"Liu Jiachang","year":"2022","unstructured":"Jiachang Liu, Chudi Zhong, Margo Seltzer, and Cynthia Rudin. 2022. Fast Sparse Classification for Generalized Linear and Additive Models. Proceedings of machine learning research 151 (2022), 9304."},{"key":"e_1_3_2_1_72_1","doi-asserted-by":"publisher","DOI":"10.3390\/app12199423"},{"key":"e_1_3_2_1_73_1","volume-title":"A unified approach to interpreting model predictions. Advances in neural information processing systems 30","author":"Lundberg Scott M","year":"2017","unstructured":"Scott M Lundberg and Su-In Lee. 2017. A unified approach to interpreting model predictions. Advances in neural information processing systems 30 (2017)."},{"key":"e_1_3_2_1_74_1","doi-asserted-by":"publisher","DOI":"10.1093\/idpl\/ipx019"},{"key":"e_1_3_2_1_75_1","doi-asserted-by":"publisher","unstructured":"Gabriele Mazzini and Salvatore Scalzo. 2022. The Proposal for the Artificial Intelligence Act: Considerations around Some Key Concepts. La via europea per l\u2019Intelligenza artificiale. Atti del Convegno del Progetto Dottorale di Alta Formazione in Scienze Giuridiche-Ca\u2019Foscari Venezia 25-26 novembre 2021 (2022). https:\/\/doi.org\/10.2139\/ssrn.4098809","DOI":"10.2139\/ssrn.4098809"},{"key":"e_1_3_2_1_76_1","volume-title":"Explanation in artificial intelligence: Insights from the social sciences. Artificial intelligence 267","author":"Miller Tim","year":"2019","unstructured":"Tim Miller. 2019. Explanation in artificial intelligence: Insights from the social sciences. Artificial intelligence 267 (2019), 1\u201338."},{"key":"e_1_3_2_1_77_1","volume-title":"Explainable AI: Beware of inmates running the asylum or: How I learnt to stop worrying and love the social and behavioural sciences. arXiv preprint arXiv:1712.00547","author":"Miller Tim","year":"2017","unstructured":"Tim Miller, Piers Howe, and Liz Sonenberg. 2017. Explainable AI: Beware of inmates running the asylum or: How I learnt to stop worrying and love the social and behavioural sciences. arXiv preprint arXiv:1712.00547 (2017)."},{"key":"e_1_3_2_1_78_1","volume-title":"Taxonomy of Machine Learning Safety: A Survey and Primer. ACM Computing Surveys (CSUR)","author":"Mohseni Sina","year":"2021","unstructured":"Sina Mohseni, Haotao Wang, Chaowei Xiao, Zhiding Yu, Zhangyang Wang, and Jay Yadawa. 2021. Taxonomy of Machine Learning Safety: A Survey and Primer. ACM Computing Surveys (CSUR) (2021)."},{"key":"e_1_3_2_1_79_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-65965-3_28"},{"key":"e_1_3_2_1_81_1","volume-title":"Adversarial Robustness of MR Image Reconstruction Under Realistic Perturbations. In International Workshop on Machine Learning for Medical Image Reconstruction. Springer, 24\u201333","author":"Morshuis Jan Nikolas","year":"2022","unstructured":"Jan Nikolas Morshuis, Sergios Gatidis, Matthias Hein, and Christian F Baumgartner. 2022. Adversarial Robustness of MR Image Reconstruction Under Realistic Perturbations. In International Workshop on Machine Learning for Medical Image Reconstruction. Springer, 24\u201333."},{"key":"e_1_3_2_1_82_1","doi-asserted-by":"publisher","DOI":"10.1073\/pnas.1900654116"},{"key":"e_1_3_2_1_83_1","volume-title":"From anecdotal evidence to quantitative evaluation methods: A systematic review on evaluating explainable ai. arXiv preprint arXiv:2201.08164","author":"Nauta Meike","year":"2022","unstructured":"Meike Nauta, Jan Trienes, Shreyasi Pathak, Elisa Nguyen, Michelle Peters, Yasmin Schmitt, J\u00f6rg Schl\u00f6tterer, Maurice van Keulen, and Christin Seifert. 2022. From anecdotal evidence to quantitative evaluation methods: A systematic review on evaluating explainable ai. arXiv preprint arXiv:2201.08164 (2022)."},{"key":"e_1_3_2_1_84_1","doi-asserted-by":"publisher","DOI":"10.1109\/MSP.2022.3142719"},{"key":"e_1_3_2_1_85_1","doi-asserted-by":"publisher","DOI":"10.1609\/hcomp.v8i1.7469"},{"key":"e_1_3_2_1_86_1","volume-title":"Dissecting racial bias in an algorithm used to manage the health of populations. Science 366, 6464","author":"Obermeyer Ziad","year":"2019","unstructured":"Ziad Obermeyer, Brian Powers, Christine Vogeli, and Sendhil Mullainathan. 2019. Dissecting racial bias in an algorithm used to manage the health of populations. Science 366, 6464 (2019), 447\u2013453."},{"key":"e_1_3_2_1_87_1","unstructured":"OECD. 2019. Recommendation of the Council on Artificial Intelligence OECD\/LEGAL\/0449."},{"key":"e_1_3_2_1_88_1","unstructured":"High Level Expert Group on Artificial Intelligence. 2019. Ethics Guidelines for Trustworthy AI."},{"key":"e_1_3_2_1_89_1","volume-title":"Co-design of human-centered, explainable AI for clinical decision support. ACM Transactions on Interactive Intelligent Systems","author":"Panigutti Cecilia","year":"2023","unstructured":"Cecilia Panigutti, Andrea Beretta, Daniele Fadda, Fosca Giannotti, Dino Pedreschi, Alan Perotti, and Salvatore Rinzivillo. 2023. Co-design of human-centered, explainable AI for clinical decision support. ACM Transactions on Interactive Intelligent Systems (2023)."},{"key":"e_1_3_2_1_90_1","doi-asserted-by":"publisher","DOI":"10.1145\/3491102.3502104"},{"key":"e_1_3_2_1_91_1","doi-asserted-by":"publisher","DOI":"10.1145\/3495013"},{"key":"e_1_3_2_1_92_1","volume-title":"Learning to deceive with attention-based explanations. arXiv preprint arXiv:1909.07913","author":"Pruthi Danish","year":"2019","unstructured":"Danish Pruthi, Mansi Gupta, Bhuwan Dhingra, Graham Neubig, and Zachary C Lipton. 2019. Learning to deceive with attention-based explanations. arXiv preprint arXiv:1909.07913 (2019)."},{"key":"e_1_3_2_1_93_1","volume-title":"Jing Liu, Yueyuan Zheng, Hilary Hei Ting Ngai, Janet Hsiao, and Lei Chen.","author":"Qiu Luyu","year":"2021","unstructured":"Luyu Qiu, Yi Yang, Caleb Chen Cao, Jing Liu, Yueyuan Zheng, Hilary Hei Ting Ngai, Janet Hsiao, and Lei Chen. 2021. Resisting out-of-distribution data problem in perturbation of xai. arXiv preprint arXiv:2107.14000 (2021)."},{"key":"e_1_3_2_1_94_1","doi-asserted-by":"publisher","DOI":"10.1145\/3485447.3512254"},{"key":"e_1_3_2_1_95_1","doi-asserted-by":"publisher","DOI":"10.1145\/2939672.2939778"},{"key":"e_1_3_2_1_96_1","doi-asserted-by":"publisher","DOI":"10.1214\/21-SS133"},{"key":"e_1_3_2_1_97_1","doi-asserted-by":"publisher","DOI":"10.3390\/app12199545"},{"key":"e_1_3_2_1_98_1","volume-title":"Explaining Deep Neural Networks: A Survey on the Global Interpretation Methods. Neurocomputing","author":"Saleem Rabia","year":"2022","unstructured":"Rabia Saleem, Bo Yuan, Fatih Kurugollu, Ashiq Anjum, and Lu Liu. 2022. Explaining Deep Neural Networks: A Survey on the Global Interpretation Methods. Neurocomputing (2022)."},{"key":"e_1_3_2_1_99_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-57321-8_24"},{"key":"e_1_3_2_1_100_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.74"},{"key":"e_1_3_2_1_101_1","volume-title":"International conference on machine learning. PMLR, 3145\u20133153","author":"Shrikumar Avanti","year":"2017","unstructured":"Avanti Shrikumar, Peyton Greenside, and Anshul Kundaje. 2017. Learning important features through propagating activation differences. In International conference on machine learning. PMLR, 3145\u20133153."},{"key":"e_1_3_2_1_102_1","first-page":"62","article-title":"Counterfactual explanations can be manipulated","volume":"34","author":"Slack Dylan","year":"2021","unstructured":"Dylan Slack, Anna Hilgard, Himabindu Lakkaraju, and Sameer Singh. 2021. Counterfactual explanations can be manipulated. Advances in Neural Information Processing Systems 34 (2021), 62\u201375.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_103_1","first-page":"9391","article-title":"Reliable post hoc explanations: Modeling uncertainty in explainability","volume":"34","author":"Slack Dylan","year":"2021","unstructured":"Dylan Slack, Anna Hilgard, Sameer Singh, and Himabindu Lakkaraju. 2021. Reliable post hoc explanations: Modeling uncertainty in explainability. Advances in Neural Information Processing Systems 34 (2021), 9391\u20139404.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_104_1","doi-asserted-by":"publisher","DOI":"10.1145\/3375627.3375830"},{"key":"e_1_3_2_1_105_1","doi-asserted-by":"publisher","DOI":"10.3390\/j5010010"},{"key":"e_1_3_2_1_106_1","doi-asserted-by":"publisher","DOI":"10.1145\/3531146.3534639"},{"key":"e_1_3_2_1_107_1","volume-title":"International conference on machine learning. PMLR, 3319\u20133328","author":"Sundararajan Mukund","year":"2017","unstructured":"Mukund Sundararajan, Ankur Taly, and Qiqi Yan. 2017. Axiomatic attribution for deep networks. In International conference on machine learning. PMLR, 3319\u20133328."},{"key":"e_1_3_2_1_108_1","doi-asserted-by":"publisher","DOI":"10.1145\/3411764.3445088"},{"key":"e_1_3_2_1_109_1","volume-title":"Assessing social and intersectional biases in contextualized word representations. Advances in Neural Information Processing Systems 32","author":"Tan Yi Chern","year":"2019","unstructured":"Yi Chern Tan and L Elisa Celis. 2019. Assessing social and intersectional biases in contextualized word representations. Advances in Neural Information Processing Systems 32 (2019)."},{"key":"e_1_3_2_1_110_1","doi-asserted-by":"publisher","DOI":"10.3390\/a15100367"},{"key":"e_1_3_2_1_111_1","unstructured":"Media UK Secretary of State for Digital Culture and Sport. 2022. AI Regulation Policy Paper. https:\/\/www.gov.uk\/government\/publications\/establishing-a-pro-innovation-approach-to-regulating-ai\/establishing-a-pro-innovation-approach-to-regulating-ai-policy-statement"},{"key":"e_1_3_2_1_112_1","unstructured":"UNESCO. 2022. Recommendation on the Ethics of Artificial Intelligence. https:\/\/unesdoc.unesco.org\/ark:\/48223\/pf0000381137"},{"key":"e_1_3_2_1_113_1","doi-asserted-by":"publisher","DOI":"10.2760\/069178"},{"key":"e_1_3_2_1_114_1","volume-title":"Attention is all you need. Advances in neural information processing systems 30","author":"Vaswani Ashish","year":"2017","unstructured":"Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, \u0141ukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. Advances in neural information processing systems 30 (2017)."},{"key":"e_1_3_2_1_115_1","doi-asserted-by":"publisher","DOI":"10.1080\/01605682.2020.1865846"},{"key":"e_1_3_2_1_116_1","doi-asserted-by":"publisher","DOI":"10.1093\/idpl\/ipx005"},{"key":"e_1_3_2_1_117_1","doi-asserted-by":"publisher","DOI":"10.1145\/3397481.3450650"},{"key":"e_1_3_2_1_118_1","unstructured":"David S Watson Limor Gultchin Ankur Taly and Luciano Floridi. 2021. Local explanations via necessity and sufficiency: Unifying theory and practice. In Uncertainty in Artificial Intelligence. PMLR 1382\u20131392."},{"key":"e_1_3_2_1_119_1","doi-asserted-by":"publisher","DOI":"10.1145\/3544548.3581197"},{"key":"e_1_3_2_1_120_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-10590-1_53"},{"key":"e_1_3_2_1_121_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3372852"},{"key":"e_1_3_2_1_122_1","volume-title":"arXiv preprint arXiv:1904.12991","author":"Zhang Yujia","year":"2019","unstructured":"Yujia Zhang, Kuangyan Song, Yiming Sun, Sarah Tan, and Madeleine Udell. 2019. \" Why Should You Trust My Explanation?\" Understanding Uncertainty in LIME Explanations. arXiv preprint arXiv:1904.12991 (2019)."},{"key":"e_1_3_2_1_123_1","doi-asserted-by":"publisher","DOI":"10.3390\/electronics10050593"},{"key":"e_1_3_2_1_124_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v36i9.21196"}],"event":{"name":"FAccT '23: the 2023 ACM Conference on Fairness, Accountability, and Transparency","location":"Chicago IL USA","acronym":"FAccT '23"},"container-title":["2023 ACM Conference on Fairness Accountability and Transparency"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3593013.3594069","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3593013.3594069","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T16:37:18Z","timestamp":1750178238000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3593013.3594069"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,6,12]]},"references-count":120,"alternative-id":["10.1145\/3593013.3594069","10.1145\/3593013"],"URL":"https:\/\/doi.org\/10.1145\/3593013.3594069","relation":{},"subject":[],"published":{"date-parts":[[2023,6,12]]},"assertion":[{"value":"2023-06-12","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}