{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,9]],"date-time":"2026-04-09T01:21:12Z","timestamp":1775697672426,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":64,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,6,20]],"date-time":"2022-06-20T00:00:00Z","timestamp":1655683200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2022,6,21]]},"DOI":"10.1145\/3531146.3533153","type":"proceedings-article","created":{"date-parts":[[2022,6,20]],"date-time":"2022-06-20T14:27:10Z","timestamp":1655735230000},"page":"891-905","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":56,"title":["Post-Hoc Explanations Fail to Achieve their Purpose in Adversarial Contexts"],"prefix":"10.1145","author":[{"given":"Sebastian","family":"Bordt","sequence":"first","affiliation":[{"name":"Department of Computer Science, University of T\u00fcbingen, Germany"}]},{"given":"Mich\u00e8le","family":"Finck","sequence":"additional","affiliation":[{"name":"Law Faculty, University of T\u00fcbingen, Germany"}]},{"given":"Eric","family":"Raidl","sequence":"additional","affiliation":[{"name":"Ethics and Philosophy Lab, University of T\u00fcbingen, Germany"}]},{"given":"Ulrike","family":"von Luxburg","sequence":"additional","affiliation":[{"name":"Department of Computer Science, University of T\u00fcbingen, Germany"}]}],"member":"320","published-online":{"date-parts":[[2022,6,20]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"The Nature of Explanation","author":"Achinstein P.","unstructured":"P. Achinstein. 1983. The Nature of Explanation. Oxford University Press, New York."},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2018.2870052"},{"key":"e_1_3_2_1_3_1","unstructured":"J. Adebayo J. Gilmer M. Muelly I. Goodfellow M. Hardt and B. Kim. 2018. Sanity checks for saliency maps. In Neural Information Processing Systems (NeurIPS)."},{"key":"e_1_3_2_1_4_1","unstructured":"A.Karimi G. Barthe B. Sch\u00f6lkopf and I. Valera. 2021. A survey of algorithmic recourse: definitions formulations solutions and prospects. arxiv:2010.04050"},{"key":"e_1_3_2_1_5_1","volume-title":"International Conference on Machine Learning (ICML).","author":"Anders C.","unstructured":"C. Anders, P. Pasliev, A.\u00a0K. Dombrowski, K.\u00a0R. M\u00fcller, and P. Kessel. 2020. Fairwashing explanations with off-manifold detergent. In International Conference on Machine Learning (ICML)."},{"key":"e_1_3_2_1_6_1","unstructured":"S. Barocas M. Hardt and A. Narayanan. 2019. Fairness and Machine Learning. fairmlbook.org. http:\/\/www.fairmlbook.org."},{"key":"e_1_3_2_1_7_1","volume-title":"ACM Conference on Fairness, Accountability, and Transparency.","author":"Barocas S.","unstructured":"S. Barocas, A. Selbst, and M. Raghavan. 2020. The hidden assumptions behind counterfactual explanations and principal reasons. In ACM Conference on Fairness, Accountability, and Transparency."},{"key":"e_1_3_2_1_8_1","volume-title":"Scientific Explanation: A Study of the Function of Theory, Probability and Law in Science","author":"Braithwaite B.","year":"1953","unstructured":"R.\u00a0B. Braithwaite. 1953. Scientific Explanation: A Study of the Function of Theory, Probability and Law in Science. Cambridge University Press, Cambridge."},{"key":"e_1_3_2_1_9_1","unstructured":"O. Camburu E. Giunchiglia J. Foerster T. Lukasiewicz and P. Blunsom. 2019. Can I trust the explainer? Verifying post-hoc explanatory methods. arXiv:1910.02065 (2019)."},{"key":"e_1_3_2_1_10_1","volume-title":"IEEE 29th International Requirements Engineering Conference (RE).","author":"Chazette L.","unstructured":"L. Chazette, W. Brunotte, and T. Speith. 2021. Exploring explainability: A definition, a model, and a knowledge catalogue. In IEEE 29th International Requirements Engineering Conference (RE)."},{"key":"e_1_3_2_1_11_1","volume-title":"White Paper on Artificial Intelligence-A European approach to excellence and trust. Com (2020) 65 Final","author":"European Commission","year":"2020","unstructured":"European Commission. 2020. White Paper on Artificial Intelligence-A European approach to excellence and trust. Com (2020) 65 Final (2020)."},{"key":"e_1_3_2_1_12_1","first-page":"1","article-title":"Explaining by removing: A unified framework for model explanation","volume":"22","author":"Covert I.","year":"2021","unstructured":"I. Covert, S. Lundberg, and S.I. Lee. 2021. Explaining by removing: A unified framework for model explanation. Journal of Machine Learning Research (JMLR) 22, 209 (2021), 1\u201390.","journal-title":"Journal of Machine Learning Research (JMLR)"},{"key":"e_1_3_2_1_13_1","volume-title":"Retiring Adult: New Datasets for Fair Machine Learning. In Neural Information Processing Systems (NeurIPS).","author":"Ding F.","year":"2021","unstructured":"F. Ding, M. Hardt, J. Miller, and L. Schmidt. 2021. Retiring Adult: New Datasets for Fair Machine Learning. In Neural Information Processing Systems (NeurIPS)."},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"crossref","unstructured":"L. Edwards and M. Veale. 2017. Slave to the algorithm: Why a right to an explanation is probably not the remedy you are looking for. Duke Law and Technology Review 16 (2017).","DOI":"10.31228\/osf.io\/97upg"},{"key":"e_1_3_2_1_15_1","volume-title":"Explaining the Explainer: A First Theoretical Analysis of LIME. In Conference on Artificial Intelligence and Statistics (AISTATS).","author":"Garreau D.","year":"2020","unstructured":"D. Garreau and U. von\u00a0Luxburg. 2020. Explaining the Explainer: A First Theoretical Analysis of LIME. In Conference on Artificial Intelligence and Statistics (AISTATS)."},{"key":"e_1_3_2_1_16_1","unstructured":"S. Ghalebikesabi L. Ter-Minassian K. DiazOrdaz and C.\u00a0C. Holmes. 2021. On locality of local explanation models. In Advances in Neural Information Processing Systems (NeurIPS)."},{"key":"e_1_3_2_1_17_1","volume-title":"Aspects of Scientific Explanation and Other Essays in the Philosophy of Science","author":"Hempel C.","unstructured":"C. Hempel. 1965. Aspects of Scientific Explanation and Other Essays in the Philosophy of Science. Free Press, New York."},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1515\/til-2019-0004"},{"key":"e_1_3_2_1_19_1","volume-title":"ACM conference on Fairness, Accountability, and Transparency.","author":"Jacobs Z.","unstructured":"A.\u00a0Z. Jacobs and H. Wallach. 2021. Measurement and fairness. In ACM conference on Fairness, Accountability, and Transparency."},{"key":"e_1_3_2_1_20_1","volume-title":"International Conference on Artificial Intelligence and Statistics (AISTATS).","author":"Janzing D.","unstructured":"D. Janzing, L. Minorics, and P. Bl\u00f6baum. 2020. Feature relevance quantification in explainable AI: A causal problem. In International Conference on Artificial Intelligence and Statistics (AISTATS)."},{"key":"e_1_3_2_1_21_1","unstructured":"M. Kaminski and J. Urban. 2021. The Right to Contest AI. Columbia Law Review (2021)."},{"key":"e_1_3_2_1_22_1","volume-title":"IEEE 29th International Requirements Engineering Conference Workshops (REW).","author":"K\u00e4stner L.","unstructured":"L. K\u00e4stner, M. Langer, V. Lazar, A. Schom\u00e4cker, T. Speith, and S. Sterz. 2021. On the Relation of Trust and Explainability: Why to Engineer for Trustworthiness. In IEEE 29th International Requirements Engineering Conference Workshops (REW)."},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1093\/jla\/laz001"},{"key":"e_1_3_2_1_24_1","volume-title":"AAAI\/ACM Conference on AI, Ethics, and Society.","author":"Kommiya\u00a0Mothilal R.","unstructured":"R. Kommiya\u00a0Mothilal, D. Mahajan, C. Tan, and A. Sharma. 2021. Towards unifying feature attribution and counterfactual explanations: Different means to the same end. In AAAI\/ACM Conference on AI, Ethics, and Society."},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"crossref","unstructured":"S. Krishna T. Han A. Gu J. Pombra S. Jabbari S. Wu and H. Lakkaraju. 2022. The Disagreement Problem in Explainable Machine Learning: A Practitioner\u2019s Perspective. arXiv preprint arXiv:2202.01602(2022).","DOI":"10.21203\/rs.3.rs-2963888\/v1"},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"crossref","unstructured":"M. Langer D. Oster T. Speith H. Hermanns L. K\u00e4stner E. Schmidt A. Sesing and K. Baum. 2021. What do we want from Explainable Artificial Intelligence (XAI)? \u2013 A stakeholder perspective on XAI and a conceptual model guiding interdisciplinary XAI research. Artificial Intelligence 296 (2021).","DOI":"10.1016\/j.artint.2021.103473"},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"crossref","unstructured":"E. Lee D. Braines Mi. Stiffler A. Hudler and D. Harborne. 2019. Developing the sensitivity of LIME for better machine learning explanation. In Artificial Intelligence and Machine Learning for Multi-Domain Operations Applications.","DOI":"10.1117\/12.2520149"},{"key":"e_1_3_2_1_28_1","unstructured":"D. Lewis. 1973. Counterfactuals. Blackwell."},{"key":"e_1_3_2_1_29_1","unstructured":"Q.\u00a0V. Liao and K.\u00a0R. Varshney. 2021. Human-Centered Explainable AI (XAI): From Algorithms to User Experiences. arXiv preprint arXiv:2110.10790(2021)."},{"key":"e_1_3_2_1_30_1","unstructured":"S. Lundberg and S. Lee. 2017. A unified approach to interpreting model predictions. In Neural Information Processing Systems (NeurIPS)."},{"key":"e_1_3_2_1_31_1","volume-title":"From local explanations to global understanding with explainable AI for trees. Nature machine intelligence 2, 1","author":"Lundberg M.","year":"2020","unstructured":"S.\u00a0M. Lundberg, G. Erion, H. Chen, A. DeGrave, J.\u00a0M. Prutkin, B. Nair, R. Katz, J. Himmelfarb, N. Bansal, and S.\u00a0I. Lee. 2020. From local explanations to global understanding with explainable AI for trees. Nature machine intelligence 2, 1 (2020), 56\u201367."},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"crossref","unstructured":"G. Malgieri and G. Comand\u00e9. 2017. Why a Right to Legibility of Automated Decision-Making Exists in the General Data Protection Regulation. International Data Privacy Law 7 4 (11 2017) 243\u2013265.","DOI":"10.1093\/idpl\/ipx019"},{"key":"e_1_3_2_1_33_1","unstructured":"C. Molnar. 2020. Interpretable machine learning. Lulu.com."},{"key":"e_1_3_2_1_34_1","volume-title":"ACM Conference on Fairness, Accountability, and Transparency.","author":"Mothilal R.","unstructured":"R. Mothilal, A. Sharma, and C. Tan. 2020. Explaining machine learning classifiers through diverse counterfactual explanations. In ACM Conference on Fairness, Accountability, and Transparency."},{"key":"e_1_3_2_1_35_1","unstructured":"High-Level Expert\u00a0Group on AI. 2019. Ethics Guidelines for Trustworthy AI."},{"key":"e_1_3_2_1_36_1","unstructured":"Working Party. 2016. Guidelines on Automated individual decision-making and Profiling for the purposes of RegulationGuidelines on Automated individual decision-making and Profiling for the purposes of Regulation 2016\/679."},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"crossref","unstructured":"A. Paullada I. Raji E. Bender E.and\u00a0Denton and A. Hanna. 2021. Data and its (dis) contents: A survey of dataset development and use in machine learning research. Patterns 2 11 (2021).","DOI":"10.1016\/j.patter.2021.100336"},{"key":"e_1_3_2_1_38_1","volume-title":"Causality: Models, Reasoning and Inference","author":"Pearl J.","year":"2000","unstructured":"J. Pearl. 2000. Causality: Models, Reasoning and Inference. Cambridge University Press, Cambridge."},{"key":"e_1_3_2_1_39_1","volume-title":"The Logic of Scientific Discovery","author":"Popper K.","unstructured":"K. Popper. 1959. The Logic of Scientific Discovery. Hutchinson, London."},{"key":"e_1_3_2_1_40_1","doi-asserted-by":"crossref","unstructured":"A. Reutlinger and J. Saatsi. 2018. Explanation Beyond Causation; Philosophical Perspectives on Non-Causal Explanations. Oxford University Press Oxford.","DOI":"10.1093\/oso\/9780198777946.001.0001"},{"key":"e_1_3_2_1_41_1","volume-title":"22nd ACM SIGKDD international conference on knowledge discovery and data mining.","author":"Ribeiro T.","unstructured":"M.\u00a0T. Ribeiro, S. Singh, and C. Guestrin. 2016. Why should i trust you? Explaining the predictions of any classifier. In 22nd ACM SIGKDD international conference on knowledge discovery and data mining."},{"key":"e_1_3_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1038\/s42256-019-0048-x"},{"key":"e_1_3_2_1_43_1","volume-title":"Statistical Explanation and Statistical Relevance","author":"Salmon W.","unstructured":"W. Salmon. 1971. Statistical Explanation and Statistical Relevance. University of Pittsburgh Press, Pittsburgh, PA."},{"key":"e_1_3_2_1_44_1","volume-title":"Minnesota Studies in the Philosophy of Science, Vol.\u00a013","author":"Salmon W.","unstructured":"W. Salmon. 1989. Four Decades of Scientific Explanation. In Scientific Explanation, Kitcher and Salmon (Eds.). Minnesota Studies in the Philosophy of Science, Vol.\u00a013. University of Minnesota Press, 3\u2013219."},{"key":"e_1_3_2_1_45_1","volume-title":"ACM Conference on Fairness, Accountability, and Transparency.","author":"Selbst A.","unstructured":"A. Selbst and J. Powles. 2018. Meaningful Information and the Right to Explanation. In ACM Conference on Fairness, Accountability, and Transparency."},{"key":"e_1_3_2_1_46_1","unstructured":"D. Slack A. Hilgard S. Singh and H. Lakkaraju. 2021. Reliable post hoc explanations: Modeling uncertainty in explainability. In Neural Information Processing Systems (NeurIPS)."},{"key":"e_1_3_2_1_47_1","volume-title":"AAAI\/ACM Conference on AI, Ethics, and Society.","author":"Slack D.","unstructured":"D. Slack, S. Hilgard, E. Jia, S. Singh, and H. Lakkaraju. 2020. Fooling lime and shap: Adversarial attacks on post hoc explanation methods. In AAAI\/ACM Conference on AI, Ethics, and Society."},{"key":"e_1_3_2_1_48_1","unstructured":"D. Slack S. Hilgard H. Lakkaraju and S. Singh. 2021. Counterfactual Explanations Can Be Manipulated. arXiv:2106.02666 (2021)."},{"key":"e_1_3_2_1_49_1","doi-asserted-by":"crossref","unstructured":"P. Spirtes C. Glymour and R. Scheines. 1993. Causation Prediction and Search. Springer Berlin.","DOI":"10.1007\/978-1-4612-2748-9"},{"key":"e_1_3_2_1_50_1","doi-asserted-by":"publisher","DOI":"10.1007\/BF00258078"},{"key":"e_1_3_2_1_51_1","volume-title":"International Conference on Machine Learning (ICML).","author":"Sundararajan M.","unstructured":"M. Sundararajan and A. Najmi. 2020. The many Shapley values for model explanation. In International Conference on Machine Learning (ICML)."},{"key":"e_1_3_2_1_52_1","volume-title":"ICML Workshop on Human Interpretability in Machine Learning.","author":"Tomsett R.","unstructured":"R. Tomsett, D. Braines, D. Harborne, A. Preece, and S. Chakraborty. 2018. Interpretable to Whom? A Role-based Model for Analyzing Interpretable Machine Learning Systems. In ICML Workshop on Human Interpretability in Machine Learning."},{"key":"e_1_3_2_1_53_1","doi-asserted-by":"publisher","DOI":"10.1038\/s41591-020-0942-0"},{"key":"e_1_3_2_1_54_1","doi-asserted-by":"publisher","DOI":"10.9785\/cri-2021-220402"},{"key":"e_1_3_2_1_55_1","volume-title":"The Philosophical Basis of Algorithmic Recourse. In ACM Conference on Fairness, Accountability, and Transparency.","author":"Venkatasubramanian S.","unstructured":"S. Venkatasubramanian and M. Alfano. 2020. The Philosophical Basis of Algorithmic Recourse. In ACM Conference on Fairness, Accountability, and Transparency."},{"key":"e_1_3_2_1_56_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2021.05.009"},{"key":"e_1_3_2_1_57_1","first-page":"1607","article-title":"Transparency and the Black Box Problem","volume":"34","author":"von Eschenbach J.","year":"2021","unstructured":"W.\u00a0J. von Eschenbach. 2021. Transparency and the Black Box Problem: Why We Do Not Trust AI. Philos. Technol. 34(2021), 1607\u20131622.","journal-title":"Why We Do Not Trust AI. Philos. Technol."},{"key":"e_1_3_2_1_58_1","volume-title":"R. Williamson, and I. Guyon.","author":"U.","year":"2012","unstructured":"U. von\u00a0Luxburg, R. Williamson, and I. Guyon. 2012. Clustering: Science or Art?JMLR Workshop and Conference Proceedings (Workshop on Unsupervised Learning and Transfer Learning)(2012), 65\u00a0\u2013\u00a079."},{"key":"e_1_3_2_1_59_1","doi-asserted-by":"crossref","unstructured":"S. Wachter B. Mittelstadt and L. Floridi. 2017. Why a Right to Explanation of Automated Decision-Making Does Not Exist in the General Data Protection Regulation. International Data Privacy Law 7 2 (06 2017) 76\u201399.","DOI":"10.1093\/idpl\/ipx005"},{"key":"e_1_3_2_1_60_1","first-page":"841","article-title":"Counterfactual explanations without opening the black box: Automated decisions and the GDPR","volume":"31","author":"Wachter S.","year":"2017","unstructured":"S. Wachter, B. Mittelstadt, and C. Russell. 2017. Counterfactual explanations without opening the black box: Automated decisions and the GDPR. Harv. JL & Tech. 31(2017), 841.","journal-title":"Harv. JL & Tech."},{"key":"e_1_3_2_1_61_1","volume-title":"Making Things Happen: A Theory of Causal Explanation","author":"Woodward J.","unstructured":"J. Woodward. 2003. Making Things Happen: A Theory of Causal Explanation. Oxford University Press."},{"key":"e_1_3_2_1_62_1","unstructured":"J. Woodward and L. Ross. 2003. Scientific Explanation. The Stanford Encyclopedia of Philosophy (Summer Edition 2021) (2003). https:\/\/plato.stanford.edu\/archives\/sum2021\/entries\/scientific-explanation\/"},{"key":"e_1_3_2_1_63_1","unstructured":"C. Zednik and H. Boelsen. forthcoming. Scientific Exploration and Explainable Artificial Intelligence. Minds and Machines(forthcoming)."},{"key":"e_1_3_2_1_64_1","unstructured":"Y. Zhang K. Song Y. Sun S. Tan and M. Udell. 2019. Why Should You Trust My Explanation? Understanding Uncertainty in LIME Explanations. arXiv preprint arXiv:1904.12991(2019)."}],"event":{"name":"FAccT '22: 2022 ACM Conference on Fairness, Accountability, and Transparency","location":"Seoul Republic of Korea","acronym":"FAccT '22","sponsor":["ACM Association for Computing Machinery"]},"container-title":["2022 ACM Conference on Fairness Accountability and Transparency"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3531146.3533153","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3531146.3533153","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T19:02:08Z","timestamp":1750186928000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3531146.3533153"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,6,20]]},"references-count":64,"alternative-id":["10.1145\/3531146.3533153","10.1145\/3531146"],"URL":"https:\/\/doi.org\/10.1145\/3531146.3533153","relation":{},"subject":[],"published":{"date-parts":[[2022,6,20]]},"assertion":[{"value":"2022-06-20","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}