{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,14]],"date-time":"2026-04-14T00:41:49Z","timestamp":1776127309811,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":49,"publisher":"ACM","license":[{"start":{"date-parts":[[2023,6,12]],"date-time":"2023-06-12T00:00:00Z","timestamp":1686528000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"DOI":"10.13039\/100000001","name":"NSF (National Science Foundation)","doi-asserted-by":"publisher","award":["IIS-1901168, IIS-2008139"],"award-info":[{"award-number":["IIS-1901168, IIS-2008139"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]},{"name":"GFSD (Graduate Fellowships for STEM Diversity)"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2023,6,12]]},"DOI":"10.1145\/3593013.3593972","type":"proceedings-article","created":{"date-parts":[[2023,6,12]],"date-time":"2023-06-12T14:40:46Z","timestamp":1686580846000},"page":"12-21","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":9,"title":["How to Explain and Justify Almost Any Decision: Potential Pitfalls for Accountability in AI Decision-Making"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-1205-3970","authenticated-orcid":false,"given":"Joyce","family":"Zhou","sequence":"first","affiliation":[{"name":"Cornell University, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3654-3683","authenticated-orcid":false,"given":"Thorsten","family":"Joachims","sequence":"additional","affiliation":[{"name":"Cornell University, USA"}]}],"member":"320","published-online":{"date-parts":[[2023,6,12]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"Adrien Bennetot, Siham Tabik, Alberto Barbado, Salvador Garc\u00eda, Sergio Gil-L\u00f3pez, Daniel Molina, Richard Benjamins, Raja Chatila, and Francisco Herrera.","author":"Arrieta Alejandro Barredo","year":"2019","unstructured":"Alejandro Barredo Arrieta , Natalia D\u00edaz-Rodr\u00edguez , Javier Del Ser , Adrien Bennetot, Siham Tabik, Alberto Barbado, Salvador Garc\u00eda, Sergio Gil-L\u00f3pez, Daniel Molina, Richard Benjamins, Raja Chatila, and Francisco Herrera. 2019 . Explainable Artificial Intelligence (XAI): Concepts, Taxonomies , Opportunities and Challenges toward Responsible AI. ( 2019). https:\/\/doi.org\/10.48550\/ARXIV.1910.10045 arxiv:1910.10045 [cs.AI] 10.48550\/ARXIV.1910.10045 Alejandro Barredo Arrieta, Natalia D\u00edaz-Rodr\u00edguez, Javier Del Ser, Adrien Bennetot, Siham Tabik, Alberto Barbado, Salvador Garc\u00eda, Sergio Gil-L\u00f3pez, Daniel Molina, Richard Benjamins, Raja Chatila, and Francisco Herrera. 2019. Explainable Artificial Intelligence (XAI): Concepts, Taxonomies, Opportunities and Challenges toward Responsible AI. (2019). https:\/\/doi.org\/10.48550\/ARXIV.1910.10045 arxiv:1910.10045 [cs.AI]"},{"key":"#cr-split#-e_1_3_2_1_2_1.1","unstructured":"Ulrich A\u00efvodji Hiromi Arai Olivier Fortineau S\u00e9bastien Gambs Satoshi Hara and Alain Tapp. 2019. Fairwashing: the risk of rationalization. (2019). https:\/\/doi.org\/10.48550\/ARXIV.1901.09749 arxiv:1901.09749 [cs.LG] 10.48550\/ARXIV.1901.09749"},{"key":"#cr-split#-e_1_3_2_1_2_1.2","unstructured":"Ulrich A\u00efvodji Hiromi Arai Olivier Fortineau S\u00e9bastien Gambs Satoshi Hara and Alain Tapp. 2019. Fairwashing: the risk of rationalization. (2019). https:\/\/doi.org\/10.48550\/ARXIV.1901.09749 arxiv:1901.09749 [cs.LG]"},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1145\/3411764.3445717"},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3372830"},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2020\/417"},{"key":"#cr-split#-e_1_3_2_1_6_1.1","unstructured":"Miles Brundage Shahar Avin Jasmine Wang Haydn Belfield Gretchen Krueger Gillian Hadfield Heidy Khlaaf Jingying Yang Helen Toner Ruth Fong Tegan Maharaj Pang Wei Koh Sara Hooker Jade Leung Andrew Trask Emma Bluemke Jonathan Lebensold Cullen O'Keefe Mark Koren Th\u00e9o Ryffel JB Rubinovitz Tamay Besiroglu Federica Carugati Jack Clark Peter Eckersley Sarah de Haas Maritza Johnson Ben Laurie Alex Ingerman Igor Krawczuk Amanda Askell Rosario Cammarota Andrew Lohn David Krueger Charlotte Stix Peter Henderson Logan Graham Carina Prunkl Bianca Martin Elizabeth Seger Noa Zilberman Se\u00e1n \u00d3 h\u00c9igeartaigh Frens Kroeger Girish Sastry Rebecca Kagan Adrian Weller Brian Tse Elizabeth Barnes Allan Dafoe Paul Scharre Ariel Herbert-Voss Martijn Rasser Shagun Sodhani Carrick Flynn Thomas Krendl Gilbert Lisa Dyer Saif Khan Yoshua Bengio and Markus Anderljung. 2020. Toward Trustworthy AI Development: Mechanisms for Supporting Verifiable Claims. (2020). https:\/\/doi.org\/10.48550\/ARXIV.2004.07213 arxiv:2004.07213 [cs.CY] 10.48550\/ARXIV.2004.07213"},{"key":"#cr-split#-e_1_3_2_1_6_1.2","unstructured":"Miles Brundage Shahar Avin Jasmine Wang Haydn Belfield Gretchen Krueger Gillian Hadfield Heidy Khlaaf Jingying Yang Helen Toner Ruth Fong Tegan Maharaj Pang Wei Koh Sara Hooker Jade Leung Andrew Trask Emma Bluemke Jonathan Lebensold Cullen O'Keefe Mark Koren Th\u00e9o Ryffel JB Rubinovitz Tamay Besiroglu Federica Carugati Jack Clark Peter Eckersley Sarah de Haas Maritza Johnson Ben Laurie Alex Ingerman Igor Krawczuk Amanda Askell Rosario Cammarota Andrew Lohn David Krueger Charlotte Stix Peter Henderson Logan Graham Carina Prunkl Bianca Martin Elizabeth Seger Noa Zilberman Se\u00e1n \u00d3 h\u00c9igeartaigh Frens Kroeger Girish Sastry Rebecca Kagan Adrian Weller Brian Tse Elizabeth Barnes Allan Dafoe Paul Scharre Ariel Herbert-Voss Martijn Rasser Shagun Sodhani Carrick Flynn Thomas Krendl Gilbert Lisa Dyer Saif Khan Yoshua Bengio and Markus Anderljung. 2020. Toward Trustworthy AI Development: Mechanisms for Supporting Verifiable Claims. (2020). https:\/\/doi.org\/10.48550\/ARXIV.2004.07213 arxiv:2004.07213 [cs.CY]"},{"key":"e_1_3_2_1_7_1","volume-title":"Proc. ACM Hum.-Comput. Interact. 5, CSCW1 (April","author":"Bu\u00e7inca Zana","year":"2021","unstructured":"Zana Bu\u00e7inca , Maja Barbara Malaya , and Krzysztof Z. Gajos . 2021. To Trust or to Think: Cognitive Forcing Functions Can Reduce Overreliance on AI in AI-Assisted Decision-Making . Proc. ACM Hum.-Comput. Interact. 5, CSCW1 (April 2021 ). https:\/\/doi.org\/10.1145\/3449287 10.1145\/3449287 Zana Bu\u00e7inca, Maja Barbara Malaya, and Krzysztof Z. Gajos. 2021. To Trust or to Think: Cognitive Forcing Functions Can Reduce Overreliance on AI in AI-Assisted Decision-Making. Proc. ACM Hum.-Comput. Interact. 5, CSCW1 (April 2021). https:\/\/doi.org\/10.1145\/3449287"},{"key":"e_1_3_2_1_8_1","volume-title":"Algorithmic Decision Making and the Cost of Fairness(KDD \u201917)","author":"Corbett-Davies Sam","unstructured":"Sam Corbett-Davies , Emma Pierson , Avi Feller , Sharad Goel , and Aziz Huq . 2017. Algorithmic Decision Making and the Cost of Fairness(KDD \u201917) . Association for Computing Machinery , New York, NY, USA , 797\u2013806. https:\/\/doi.org\/10.1145\/3097983.3098095 10.1145\/3097983.3098095 Sam Corbett-Davies, Emma Pierson, Avi Feller, Sharad Goel, and Aziz Huq. 2017. Algorithmic Decision Making and the Cost of Fairness(KDD \u201917). Association for Computing Machinery, New York, NY, USA, 797\u2013806. https:\/\/doi.org\/10.1145\/3097983.3098095"},{"key":"#cr-split#-e_1_3_2_1_9_1.1","unstructured":"Sanjoy Dasgupta Nave Frost and Michal Moshkovitz. 2022. Framework for Evaluating Faithfulness of Local Explanations. (2022). https:\/\/doi.org\/10.48550\/ARXIV.2202.00734 arxiv:2202.00734 [cs.LG] 10.48550\/ARXIV.2202.00734"},{"key":"#cr-split#-e_1_3_2_1_9_1.2","unstructured":"Sanjoy Dasgupta Nave Frost and Michal Moshkovitz. 2022. Framework for Evaluating Faithfulness of Local Explanations. (2022). https:\/\/doi.org\/10.48550\/ARXIV.2202.00734 arxiv:2202.00734 [cs.LG]"},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1145\/3301275.3302310"},{"key":"#cr-split#-e_1_3_2_1_11_1.1","unstructured":"Finale Doshi-Velez and Been Kim. 2017. Towards A Rigorous Science of Interpretable Machine Learning. (2017). https:\/\/doi.org\/10.48550\/ARXIV.1702.08608 arxiv:1702.08608 [cs.AI] 10.48550\/ARXIV.1702.08608"},{"key":"#cr-split#-e_1_3_2_1_11_1.2","unstructured":"Finale Doshi-Velez and Been Kim. 2017. Towards A Rigorous Science of Interpretable Machine Learning. (2017). https:\/\/doi.org\/10.48550\/ARXIV.1702.08608 arxiv:1702.08608 [cs.AI]"},{"key":"#cr-split#-e_1_3_2_1_12_1.1","unstructured":"Finale Doshi-Velez Mason Kortz Ryan Budish Chris Bavitz Sam Gershman David O'Brien Kate Scott Stuart Schieber James Waldo David Weinberger Adrian Weller and Alexandra Wood. 2017. Accountability of AI Under the Law: The Role of Explanation. (2017). https:\/\/doi.org\/10.48550\/ARXIV.1711.01134 arxiv:1711.01134 [cs.AI] 10.48550\/ARXIV.1711.01134"},{"key":"#cr-split#-e_1_3_2_1_12_1.2","doi-asserted-by":"crossref","unstructured":"Finale Doshi-Velez Mason Kortz Ryan Budish Chris Bavitz Sam Gershman David O'Brien Kate Scott Stuart Schieber James Waldo David Weinberger Adrian Weller and Alexandra Wood. 2017. Accountability of AI Under the Law: The Role of Explanation. (2017). https:\/\/doi.org\/10.48550\/ARXIV.1711.01134 arxiv:1711.01134 [cs.AI]","DOI":"10.2139\/ssrn.3064761"},{"key":"e_1_3_2_1_13_1","first-page":"18","article-title":"Slave to the Algorithm? Why a \u2019Right to an Explanation\u2019 Is Probably Not the Remedy You Are Looking For","volume":"16","author":"Edwards Lilian","year":"2017","unstructured":"Lilian Edwards and Michael Veale . 2017 . Slave to the Algorithm? Why a \u2019Right to an Explanation\u2019 Is Probably Not the Remedy You Are Looking For . Duke Law & Technology Review 16 (2017), 18 \u2013 84 . https:\/\/doi.org\/10.2139\/ssrn.2972855 10.2139\/ssrn.2972855 Lilian Edwards and Michael Veale. 2017. Slave to the Algorithm? Why a \u2019Right to an Explanation\u2019 Is Probably Not the Remedy You Are Looking For. Duke Law & Technology Review 16 (2017), 18\u201384. https:\/\/doi.org\/10.2139\/ssrn.2972855","journal-title":"Duke Law & Technology Review"},{"key":"e_1_3_2_1_14_1","volume-title":"Beam","author":"Ghassemi Marzyeh","year":"2021","unstructured":"Marzyeh Ghassemi , Luke Oakden-Rayner , and Andrew L . Beam . 2021 . The false hope of current approaches to explainable artificial intelligence in health care.The Lancet. Digital health 3, 11 (Nov. 2021), e745\u2013e750. https:\/\/doi.org\/10.1016\/S2589-7500(21)00208-9 10.1016\/S2589-7500(21)00208-9 Marzyeh Ghassemi, Luke Oakden-Rayner, and Andrew L. Beam. 2021. The false hope of current approaches to explainable artificial intelligence in health care.The Lancet. Digital health 3, 11 (Nov. 2021), e745\u2013e750. https:\/\/doi.org\/10.1016\/S2589-7500(21)00208-9"},{"key":"e_1_3_2_1_15_1","unstructured":"Gillian K. Hadfield. 2021. Explanation and justification: AI decision-making law and the rights of citizens. https:\/\/srinstitute.utoronto.ca\/news\/hadfield-justifiable-ai  Gillian K. Hadfield. 2021. Explanation and justification: AI decision-making law and the rights of citizens. https:\/\/srinstitute.utoronto.ca\/news\/hadfield-justifiable-ai"},{"key":"e_1_3_2_1_16_1","volume-title":"Findings of the Association for Computational Linguistics: ACL-IJCNLP","author":"Gonz\u00e1lez Ana Valeria","year":"2021","unstructured":"Ana Valeria Gonz\u00e1lez , Gagan Bansal , Angela Fan , Yashar Mehdad , Robin Jia , and Srinivasan Iyer . 2021. Do Explanations Help Users Detect Errors in Open-Domain QA? An Evaluation of Spoken vs. Visual Explanations . In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021 . Association for Computational Linguistics , Online , 1103\u20131116. https:\/\/doi.org\/10.18653\/v1\/2021.findings-acl.95 10.18653\/v1 Ana Valeria Gonz\u00e1lez, Gagan Bansal, Angela Fan, Yashar Mehdad, Robin Jia, and Srinivasan Iyer. 2021. Do Explanations Help Users Detect Errors in Open-Domain QA? An Evaluation of Spoken vs. Visual Explanations. In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021. Association for Computational Linguistics, Online, 1103\u20131116. https:\/\/doi.org\/10.18653\/v1\/2021.findings-acl.95"},{"key":"e_1_3_2_1_17_1","volume-title":"Proc. ACM Hum.-Comput. Interact. 5, CSCW2 (Oct.","author":"Tsung-Yu Hou Yoyo","year":"2021","unstructured":"Yoyo Tsung-Yu Hou and Malte F. Jung . 2021. Who is the Expert? Reconciling Algorithm Aversion and Algorithm Appreciation in AI-Supported Decision Making . Proc. ACM Hum.-Comput. Interact. 5, CSCW2 (Oct. 2021 ). https:\/\/doi.org\/10.1145\/3479864 10.1145\/3479864 Yoyo Tsung-Yu Hou and Malte F. Jung. 2021. Who is the Expert? Reconciling Algorithm Aversion and Algorithm Appreciation in AI-Supported Decision Making. Proc. ACM Hum.-Comput. Interact. 5, CSCW2 (Oct. 2021). https:\/\/doi.org\/10.1145\/3479864"},{"key":"e_1_3_2_1_18_1","unstructured":"Julia Angwin Jeff Larson Surya Mattu and Lauren Kirchner. 2016. Machine Bias: There\u2019s software used across the country to predict future criminals. And it\u2019s biased against blacks.https:\/\/www.propublica.org\/article\/machine-bias-risk-assessments-in-criminal-sentencing  Julia Angwin Jeff Larson Surya Mattu and Lauren Kirchner. 2016. Machine Bias: There\u2019s software used across the country to predict future criminals. And it\u2019s biased against blacks.https:\/\/www.propublica.org\/article\/machine-bias-risk-assessments-in-criminal-sentencing"},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1109\/VLHCC.2013.6645235"},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1145\/3306618.3314229"},{"key":"e_1_3_2_1_21_1","volume-title":"EDDA: Explanation-driven Data Augmentation to Improve Explanation Faithfulness.","author":"Li Ruiwen","year":"2021","unstructured":"Ruiwen Li , Zhibo Zhang , Jiani Li , Chiheb Trabelsi , Scott Sanner , Jongseong Jang , Yeonjeong Jeong , and Dongsub Shim . 2021 . EDDA: Explanation-driven Data Augmentation to Improve Explanation Faithfulness. (2021). https:\/\/doi.org\/10.48550\/ARXIV.2105.14162 arxiv:2105.14162 [cs.AI] 10.48550\/ARXIV.2105.14162 Ruiwen Li, Zhibo Zhang, Jiani Li, Chiheb Trabelsi, Scott Sanner, Jongseong Jang, Yeonjeong Jeong, and Dongsub Shim. 2021. EDDA: Explanation-driven Data Augmentation to Improve Explanation Faithfulness. (2021). https:\/\/doi.org\/10.48550\/ARXIV.2105.14162 arxiv:2105.14162 [cs.AI]"},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1145\/3313831.3376590"},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1145\/1518701.1519023"},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1145\/3236386.3241340"},{"key":"e_1_3_2_1_25_1","volume-title":"Proc. ACM Hum.-Comput. Interact. 5, CSCW2 (Oct.","author":"Liu Han","year":"2021","unstructured":"Han Liu , Vivian Lai , and Chenhao Tan . 2021 . Understanding the Effect of Out-of-Distribution Examples and Interactive Explanations on Human-AI Decision Making . Proc. ACM Hum.-Comput. Interact. 5, CSCW2 (Oct. 2021). https:\/\/doi.org\/10.1145\/3479552 10.1145\/3479552 Han Liu, Vivian Lai, and Chenhao Tan. 2021. Understanding the Effect of Out-of-Distribution Examples and Interactive Explanations on Human-AI Decision Making. Proc. ACM Hum.-Comput. Interact. 5, CSCW2 (Oct. 2021). https:\/\/doi.org\/10.1145\/3479552"},{"key":"e_1_3_2_1_26_1","first-page":"19","volume-title":"Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics","author":"Liu Hui","year":"2019","unstructured":"Hui Liu , Qingyu Yin , and William Yang Wang . 2019 . Towards Explainable NLP: A Generative Explanation Framework for Text Classification . In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics , Florence, Italy, 5570\u20135581. https:\/\/doi.org\/10. 18653\/v1\/P 19 - 1560 10.18653\/v1 Hui Liu, Qingyu Yin, and William Yang Wang. 2019. Towards Explainable NLP: A Generative Explanation Framework for Text Classification. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, Florence, Italy, 5570\u20135581. https:\/\/doi.org\/10.18653\/v1\/P19-1560"},{"key":"#cr-split#-e_1_3_2_1_27_1.1","unstructured":"Shixia Liu Xiting Wang Mengchen Liu and Jun Zhu. 2017. Towards Better Analysis of Machine Learning Models: A Visual Analytics Perspective. (2017). https:\/\/doi.org\/10.48550\/ARXIV.1702.01226 arxiv:1702.01226 [cs.LG] 10.48550\/ARXIV.1702.01226"},{"key":"#cr-split#-e_1_3_2_1_27_1.2","unstructured":"Shixia Liu Xiting Wang Mengchen Liu and Jun Zhu. 2017. Towards Better Analysis of Machine Learning Models: A Visual Analytics Perspective. (2017). https:\/\/doi.org\/10.48550\/ARXIV.1702.01226 arxiv:1702.01226 [cs.LG]"},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1145\/3457607"},{"key":"#cr-split#-e_1_3_2_1_29_1.1","unstructured":"Tim Miller. 2017. Explanation in Artificial Intelligence: Insights from the Social Sciences. (2017). https:\/\/doi.org\/10.48550\/ARXIV.1706.07269 arxiv:1706.07269 [cs.AI] 10.48550\/ARXIV.1706.07269"},{"key":"#cr-split#-e_1_3_2_1_29_1.2","unstructured":"Tim Miller. 2017. Explanation in Artificial Intelligence: Insights from the Social Sciences. (2017). https:\/\/doi.org\/10.48550\/ARXIV.1706.07269 arxiv:1706.07269 [cs.AI]"},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1145\/3387166"},{"key":"#cr-split#-e_1_3_2_1_31_1.1","unstructured":"Menaka Narayanan Emily Chen Jeffrey He Been Kim Sam Gershman and Finale Doshi-Velez. 2018. How do Humans Understand Explanations from Machine Learning Systems? An Evaluation of the Human-Interpretability of Explanation. (2018). https:\/\/doi.org\/10.48550\/ARXIV.1802.00682 arxiv:1802.00682 [cs.AI] 10.48550\/ARXIV.1802.00682"},{"key":"#cr-split#-e_1_3_2_1_31_1.2","unstructured":"Menaka Narayanan Emily Chen Jeffrey He Been Kim Sam Gershman and Finale Doshi-Velez. 2018. How do Humans Understand Explanations from Machine Learning Systems? An Evaluation of the Human-Interpretability of Explanation. (2018). https:\/\/doi.org\/10.48550\/ARXIV.1802.00682 arxiv:1802.00682 [cs.AI]"},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1145\/2939672.2939778"},{"key":"#cr-split#-e_1_3_2_1_33_1.1","unstructured":"Cynthia Rudin. 2019. Stop Explaining Black Box Machine Learning Models for High Stakes Decisions and Use Interpretable Models Instead. https:\/\/doi.org\/10.48550\/arXiv.1811.10154 arXiv:1811.10154 [cs stat]. 10.48550\/arXiv.1811.10154"},{"key":"#cr-split#-e_1_3_2_1_33_1.2","doi-asserted-by":"crossref","unstructured":"Cynthia Rudin. 2019. Stop Explaining Black Box Machine Learning Models for High Stakes Decisions and Use Interpretable Models Instead. https:\/\/doi.org\/10.48550\/arXiv.1811.10154 arXiv:1811.10154 [cs stat].","DOI":"10.1038\/s42256-019-0048-x"},{"key":"e_1_3_2_1_34_1","first-page":"3","article-title":"The Intuitive Appeal of Explainable Machines","volume":"87","author":"Selbst Andrew","year":"2018","unstructured":"Andrew Selbst and Solon Barocas . 2018 . The Intuitive Appeal of Explainable Machines . Fordham Law Review 87 , 3 (Jan. 2018), 1085. https:\/\/ir.lawnet.fordham.edu\/flr\/vol87\/iss3\/11 Andrew Selbst and Solon Barocas. 2018. The Intuitive Appeal of Explainable Machines. Fordham Law Review 87, 3 (Jan. 2018), 1085. https:\/\/ir.lawnet.fordham.edu\/flr\/vol87\/iss3\/11","journal-title":"Fordham Law Review"},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1093\/idpl\/ipx022"},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1145\/3375627.3375830"},{"key":"e_1_3_2_1_38_1","first-page":"47","article-title":"Counterfactual Explanations Without Opening the Black Box: Automated Decisions and the GDPR","volume":"31","author":"Wachter Sandra","year":"2017","unstructured":"Sandra Wachter , Brent Daniel Mittelstadt , and Chris Russell . 2017 . Counterfactual Explanations Without Opening the Black Box: Automated Decisions and the GDPR . Harvard Journal of Law & Technology 31 , 2 (2017), 47 . https:\/\/doi.org\/10.2139\/ssrn.3063289 10.2139\/ssrn.3063289 Sandra Wachter, Brent Daniel Mittelstadt, and Chris Russell. 2017. Counterfactual Explanations Without Opening the Black Box: Automated Decisions and the GDPR. Harvard Journal of Law & Technology 31, 2 (2017), 47. https:\/\/doi.org\/10.2139\/ssrn.3063289","journal-title":"Harvard Journal of Law & Technology"},{"key":"#cr-split#-e_1_3_2_1_39_1.1","unstructured":"Fan Yang Mengnan Du and Xia Hu. 2019. Evaluating Explanation Without Ground Truth in Interpretable Machine Learning. (2019). https:\/\/doi.org\/10.48550\/ARXIV.1907.06831 arxiv:1907.06831 [cs.AI] 10.48550\/ARXIV.1907.06831"},{"key":"#cr-split#-e_1_3_2_1_39_1.2","unstructured":"Fan Yang Mengnan Du and Xia Hu. 2019. Evaluating Explanation Without Ground Truth in Interpretable Machine Learning. (2019). https:\/\/doi.org\/10.48550\/ARXIV.1907.06831 arxiv:1907.06831 [cs.AI]"},{"key":"e_1_3_2_1_40_1","volume-title":"Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers)","author":"Zhang Wei","year":"1865","unstructured":"Wei Zhang , Ziming Huang , Yada Zhu , Guangnan Ye , Xiaodong Cui , and Fan Zhang . 2021. On Sample Based Explanation Methods for NLP: Faithfulness, Efficiency and Semantic Evaluation . In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers) . Association for Computational Linguistics , Online , 5399\u20135411. https:\/\/doi.org\/10. 1865 3\/v1\/2021.acl-long.419 10.18653\/v1 Wei Zhang, Ziming Huang, Yada Zhu, Guangnan Ye, Xiaodong Cui, and Fan Zhang. 2021. On Sample Based Explanation Methods for NLP: Faithfulness, Efficiency and Semantic Evaluation. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers). Association for Computational Linguistics, Online, 5399\u20135411. https:\/\/doi.org\/10.18653\/v1\/2021.acl-long.419"}],"event":{"name":"FAccT '23: the 2023 ACM Conference on Fairness, Accountability, and Transparency","location":"Chicago IL USA","acronym":"FAccT '23"},"container-title":["2023 ACM Conference on Fairness, Accountability, and Transparency"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3593013.3593972","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3593013.3593972","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T16:48:02Z","timestamp":1750178882000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3593013.3593972"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,6,12]]},"references-count":49,"alternative-id":["10.1145\/3593013.3593972","10.1145\/3593013"],"URL":"https:\/\/doi.org\/10.1145\/3593013.3593972","relation":{},"subject":[],"published":{"date-parts":[[2023,6,12]]},"assertion":[{"value":"2023-06-12","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}