{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,8]],"date-time":"2026-04-08T16:23:51Z","timestamp":1775665431872,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":62,"publisher":"ACM","license":[{"start":{"date-parts":[[2021,7,21]],"date-time":"2021-07-21T00:00:00Z","timestamp":1626825600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2021,7,21]]},"DOI":"10.1145\/3461702.3462597","type":"proceedings-article","created":{"date-parts":[[2021,7,31]],"date-time":"2021-07-31T01:21:32Z","timestamp":1627694492000},"page":"652-663","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":60,"title":["Towards Unifying Feature Attribution and Counterfactual Explanations: Different Means to the Same End"],"prefix":"10.1145","author":[{"given":"Ramaravind","family":"Kommiya Mothilal","sequence":"first","affiliation":[{"name":"Microsoft Research India, Bangalore, India"}]},{"given":"Divyat","family":"Mahajan","sequence":"additional","affiliation":[{"name":"Microsoft Research India, Bangalore, India"}]},{"given":"Chenhao","family":"Tan","sequence":"additional","affiliation":[{"name":"University of Chicago, Chicago, IL, USA"}]},{"given":"Amit","family":"Sharma","sequence":"additional","affiliation":[{"name":"Microsoft Research India, Bangalore, India"}]}],"member":"320","published-online":{"date-parts":[[2021,7,30]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"[n.d.]. ([n. d.]).  [n.d.]. ([n. d.])."},{"key":"e_1_3_2_1_2_1","unstructured":"Accessed 2019. UCI Machine Learning Repository. German credit dataset. https:\/\/archive.ics.uci.edu\/ml\/support\/statlog+(german+credit+data)  Accessed 2019. UCI Machine Learning Repository. German credit dataset. https:\/\/archive.ics.uci.edu\/ml\/support\/statlog+(german+credit+data)"},{"key":"e_1_3_2_1_3_1","volume-title":"Explaining individual predictions when features are dependent: More accurate approximations to Shapley values. Artificial Intelligence","author":"Aas Kjersti","year":"2021","unstructured":"Kjersti Aas , Martin Jullum , and Anders L\u00f8land . 2021. Explaining individual predictions when features are dependent: More accurate approximations to Shapley values. Artificial Intelligence ( 2021 ), 103502. Kjersti Aas, Martin Jullum, and Anders L\u00f8land. 2021. Explaining individual predictions when features are dependent: More accurate approximations to Shapley values. Artificial Intelligence (2021), 103502."},{"key":"e_1_3_2_1_4_1","volume-title":"On the robustness of interpretability methods. arXiv preprint arXiv:1806.08049","author":"Alvarez-Melis David","year":"2018","unstructured":"David Alvarez-Melis and Tommi S Jaakkola . 2018. On the robustness of interpretability methods. arXiv preprint arXiv:1806.08049 ( 2018 ). David Alvarez-Melis and Tommi S Jaakkola. 2018. On the robustness of interpretability methods. arXiv preprint arXiv:1806.08049 (2018)."},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1111\/acem.12249"},{"key":"e_1_3_2_1_6_1","volume-title":"Pin-Yu Chen, Amit Dhurandhar, Michael Hind, Samuel C Hoffman, Stephanie Houde, Q Vera Liao, Ronny Luss, Aleksandra Mojsilovi\u0107, et al.","author":"Arya Vijay","year":"2019","unstructured":"Vijay Arya , Rachel KE Bellamy , Pin-Yu Chen, Amit Dhurandhar, Michael Hind, Samuel C Hoffman, Stephanie Houde, Q Vera Liao, Ronny Luss, Aleksandra Mojsilovi\u0107, et al. 2019 . One explanation does not fit all: A toolkit and taxonomy of ai explainability techniques. arXiv preprint arXiv:1909.03012 (2019). Vijay Arya, Rachel KE Bellamy, Pin-Yu Chen, Amit Dhurandhar, Michael Hind, Samuel C Hoffman, Stephanie Houde, Q Vera Liao, Ronny Luss, Aleksandra Mojsilovi\u0107, et al. 2019. One explanation does not fit all: A toolkit and taxonomy of ai explainability techniques. arXiv preprint arXiv:1909.03012 (2019)."},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3372830"},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.747"},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1145\/2783258.2788613"},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58112-1_31"},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58112-1_31"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"crossref","unstructured":"Thomas Desautels Jacob Calvert Jana Hoffman Melissa Jay Yaniv Kerem Lisa Shieh David Shimabukuro Uli Chettipally Mitchell D Feldman Chris Barton etal 2016. Prediction of sepsis in the intensive care unit with minimal electronic health record data: a machine learning approach. JMIR medical informatics Vol. 4 3 (2016) e28.  Thomas Desautels Jacob Calvert Jana Hoffman Melissa Jay Yaniv Kerem Lisa Shieh David Shimabukuro Uli Chettipally Mitchell D Feldman Chris Barton et al. 2016. Prediction of sepsis in the intensive care unit with minimal electronic health record data: a machine learning approach. JMIR medical informatics Vol. 4 3 (2016) e28.","DOI":"10.2196\/medinform.5909"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.408"},{"key":"e_1_3_2_1_14_1","unstructured":"Amit Dhurandhar Pin-Yu Chen Ronny Luss Chun-Chen Tu Paishun Ting Karthikeyan Shanmugam and Payel Das. 2018. Explanations based on the missing: Towards contrastive explanations with pertinent negatives. In Advances in Neural Information Processing Systems. 592--603.  Amit Dhurandhar Pin-Yu Chen Ronny Luss Chun-Chen Tu Paishun Ting Karthikeyan Shanmugam and Payel Das. 2018. Explanations based on the missing: Towards contrastive explanations with pertinent negatives. In Advances in Neural Information Processing Systems. 592--603."},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.jemermed.2016.02.026"},{"key":"e_1_3_2_1_16_1","volume-title":"Explaining Black-Box Algorithms Using Probabilistic Contrastive Counterfactuals. arXiv preprint arXiv:2103.11972","author":"Galhotra Sainyam","year":"2021","unstructured":"Sainyam Galhotra , Romila Pradhan , and Babak Salimi . 2021. Explaining Black-Box Algorithms Using Probabilistic Contrastive Counterfactuals. arXiv preprint arXiv:2103.11972 ( 2021 ). Sainyam Galhotra, Romila Pradhan, and Babak Salimi. 2021. Explaining Black-Box Algorithms Using Probabilistic Contrastive Counterfactuals. arXiv preprint arXiv:2103.11972 (2021)."},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0172049"},{"key":"e_1_3_2_1_18_1","volume-title":"Actual causality","author":"Halpern Joseph Y","unstructured":"Joseph Y Halpern . 2016. Actual causality . MIT Press . Joseph Y Halpern. 2016. Actual causality .MIT Press."},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0201016"},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0174708"},{"key":"e_1_3_2_1_21_1","volume-title":"International Conference on Artificial Intelligence and Statistics. PMLR, 895--905","author":"Karimi Amir-Hossein","year":"2020","unstructured":"Amir-Hossein Karimi , Gilles Barthe , Borja Balle , and Isabel Valera . 2020 . Model-agnostic counterfactual explanations for consequential decisions . In International Conference on Artificial Intelligence and Statistics. PMLR, 895--905 . Amir-Hossein Karimi, Gilles Barthe, Borja Balle, and Isabel Valera. 2020. Model-agnostic counterfactual explanations for consequential decisions. In International Conference on Artificial Intelligence and Statistics. PMLR, 895--905."},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445899"},{"key":"e_1_3_2_1_23_1","volume-title":"Giovanni Vacanti, and Alexandru Coca.","author":"Klaise Janis","year":"2019","unstructured":"Janis Klaise , Arnaud Van Looveren , Giovanni Vacanti, and Alexandru Coca. 2019 . Alibi : Algorithms for monitoring and explaining machine learning models. https:\/\/github.com\/SeldonIO\/alibi Janis Klaise, Arnaud Van Looveren, Giovanni Vacanti, and Alexandru Coca. 2019. Alibi: Algorithms for monitoring and explaining machine learning models. https:\/\/github.com\/SeldonIO\/alibi"},{"key":"e_1_3_2_1_24_1","unstructured":"Ronny Kohavi and Barry Becker. 1996. UCI Machine Learning Repository. https:\/\/archive.ics.uci.edu\/ml\/datasets\/adult  Ronny Kohavi and Barry Becker. 1996. UCI Machine Learning Repository. https:\/\/archive.ics.uci.edu\/ml\/datasets\/adult"},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1561\/9781601986290"},{"key":"e_1_3_2_1_26_1","volume-title":"International Conference on Machine Learning. PMLR, 5491--5500","author":"Kumar I Elizabeth","year":"2020","unstructured":"I Elizabeth Kumar , Suresh Venkatasubramanian , Carlos Scheidegger , and Sorelle Friedler . 2020 . Problems with Shapley-value-based explanations as feature importance measures . In International Conference on Machine Learning. PMLR, 5491--5500 . I Elizabeth Kumar, Suresh Venkatasubramanian, Carlos Scheidegger, and Sorelle Friedler. 2020. Problems with Shapley-value-based explanations as feature importance measures. In International Conference on Machine Learning. PMLR, 5491--5500."},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1046"},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1145\/3287560.3287590"},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1145\/2939672.2939874"},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.annemergmed.2017.08.005"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1145\/3313831.3376590"},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1017\/S1358246100005130"},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.1145\/3236386.3241340"},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.1145\/2339530.2339556"},{"key":"e_1_3_2_1_35_1","unstructured":"Scott M Lundberg and Su-In Lee. 2017. A unified approach to interpreting model predictions. In Advances in neural information processing systems. 4765--4774.  Scott M Lundberg and Su-In Lee. 2017. A unified approach to interpreting model predictions. In Advances in neural information processing systems. 4765--4774."},{"key":"e_1_3_2_1_36_1","volume-title":"Preserving causal constraints in counterfactual explanations for machine learning classifiers. arXiv preprint arXiv:1912.03277","author":"Mahajan Divyat","year":"2019","unstructured":"Divyat Mahajan , Chenhao Tan , and Amit Sharma . 2019. Preserving causal constraints in counterfactual explanations for machine learning classifiers. arXiv preprint arXiv:1912.03277 ( 2019 ). Divyat Mahajan, Chenhao Tan, and Amit Sharma. 2019. Preserving causal constraints in counterfactual explanations for machine learning classifiers. arXiv preprint arXiv:1912.03277 (2019)."},{"key":"e_1_3_2_1_37_1","volume-title":"Explanation in artificial intelligence: Insights from the social sciences. Artificial Intelligence","author":"Miller Tim","year":"2018","unstructured":"Tim Miller . 2018. Explanation in artificial intelligence: Insights from the social sciences. Artificial Intelligence ( 2018 ). Tim Miller. 2018. Explanation in artificial intelligence: Insights from the social sciences. Artificial Intelligence (2018)."},{"key":"e_1_3_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.1136\/heartjnl-2011-301246"},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3372850"},{"key":"e_1_3_2_1_40_1","volume-title":"Predicting the future-big data, machine learning, and clinical medicine. The New England journal of medicine","author":"Obermeyer Ziad","year":"2016","unstructured":"Ziad Obermeyer and Ezekiel J Emanuel . 2016. Predicting the future-big data, machine learning, and clinical medicine. The New England journal of medicine , Vol. 375 , 13 ( 2016 ), 1216. Ziad Obermeyer and Ezekiel J Emanuel. 2016. Predicting the future-big data, machine learning, and clinical medicine. The New England journal of medicine, Vol. 375, 13 (2016), 1216."},{"key":"e_1_3_2_1_41_1","volume-title":"Causal inference in statistics: A primer","author":"Pearl Judea","unstructured":"Judea Pearl , Madelyn Glymour , and Nicholas P Jewell . 2016. Causal inference in statistics: A primer . John Wiley & Sons . Judea Pearl, Madelyn Glymour, and Nicholas P Jewell. 2016. Causal inference in statistics: A primer. John Wiley & Sons."},{"key":"e_1_3_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1145\/3375627.3375850"},{"key":"e_1_3_2_1_43_1","volume-title":"Can I Still Trust You?: Understanding the Impact of Distribution Shifts on Algorithmic Recourses. arXiv preprint arXiv:2012.11788","author":"Rawal Kaivalya","year":"2020","unstructured":"Kaivalya Rawal , Ece Kamar , and Himabindu Lakkaraju . 2020. Can I Still Trust You?: Understanding the Impact of Distribution Shifts on Algorithmic Recourses. arXiv preprint arXiv:2012.11788 ( 2020 ). Kaivalya Rawal, Ece Kamar, and Himabindu Lakkaraju. 2020. Can I Still Trust You?: Understanding the Impact of Distribution Shifts on Algorithmic Recourses. arXiv preprint arXiv:2012.11788 (2020)."},{"key":"e_1_3_2_1_44_1","doi-asserted-by":"publisher","DOI":"10.1145\/2939672.2939778"},{"key":"e_1_3_2_1_45_1","doi-asserted-by":"publisher","DOI":"10.1198\/016214504000001880"},{"key":"e_1_3_2_1_46_1","doi-asserted-by":"publisher","DOI":"10.1038\/s42256-019-0048-x"},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.1145\/3287560.3287569"},{"key":"e_1_3_2_1_48_1","volume-title":"GeCo: Quality Counterfactual Explanations in Real Time. arXiv preprint arXiv:2101.01292","author":"Schleich Maximilian","year":"2021","unstructured":"Maximilian Schleich , Zixuan Geng , Yihong Zhang , and Dan Suciu . 2021. GeCo: Quality Counterfactual Explanations in Real Time. arXiv preprint arXiv:2101.01292 ( 2021 ). Maximilian Schleich, Zixuan Geng, Yihong Zhang, and Dan Suciu. 2021. GeCo: Quality Counterfactual Explanations in Real Time. arXiv preprint arXiv:2101.01292 (2021)."},{"key":"e_1_3_2_1_49_1","volume-title":"Certifai: Counterfactual explanations for robustness, transparency, interpretability, and fairness of artificial intelligence models. arXiv preprint arXiv:1905.07857","author":"Sharma Shubham","year":"2019","unstructured":"Shubham Sharma , Jette Henderson , and Joydeep Ghosh . 2019 . Certifai: Counterfactual explanations for robustness, transparency, interpretability, and fairness of artificial intelligence models. arXiv preprint arXiv:1905.07857 (2019). Shubham Sharma, Jette Henderson, and Joydeep Ghosh. 2019. Certifai: Counterfactual explanations for robustness, transparency, interpretability, and fairness of artificial intelligence models. arXiv preprint arXiv:1905.07857 (2019)."},{"key":"e_1_3_2_1_50_1","doi-asserted-by":"publisher","DOI":"10.1145\/3375627.3375812"},{"key":"e_1_3_2_1_51_1","volume-title":"International Conference on Machine Learning. PMLR, 3145--3153","author":"Shrikumar Avanti","year":"2017","unstructured":"Avanti Shrikumar , Peyton Greenside , and Anshul Kundaje . 2017 . Learning important features through propagating activation differences . In International Conference on Machine Learning. PMLR, 3145--3153 . Avanti Shrikumar, Peyton Greenside, and Anshul Kundaje. 2017. Learning important features through propagating activation differences. In International Conference on Machine Learning. PMLR, 3145--3153."},{"key":"e_1_3_2_1_52_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3372870"},{"key":"e_1_3_2_1_53_1","volume-title":"International Conference on Machine Learning. PMLR, 9269--9278","author":"Sundararajan Mukund","year":"2020","unstructured":"Mukund Sundararajan and Amir Najmi . 2020 . The many Shapley values for model explanation . In International Conference on Machine Learning. PMLR, 9269--9278 . Mukund Sundararajan and Amir Najmi. 2020. The many Shapley values for model explanation. In International Conference on Machine Learning. PMLR, 9269--9278."},{"key":"e_1_3_2_1_54_1","volume-title":"International Conference on Machine Learning. PMLR, 3319--3328","author":"Sundararajan Mukund","year":"2017","unstructured":"Mukund Sundararajan , Ankur Taly , and Qiqi Yan . 2017 . Axiomatic attribution for deep networks . In International Conference on Machine Learning. PMLR, 3319--3328 . Mukund Sundararajan, Ankur Taly, and Qiqi Yan. 2017. Axiomatic attribution for deep networks. In International Conference on Machine Learning. PMLR, 3319--3328."},{"key":"e_1_3_2_1_55_1","volume-title":"Detecting bias in black-box models using transparent model distillation. arXiv preprint arXiv:1710.06169","author":"Tan Sarah","year":"2017","unstructured":"Sarah Tan , Rich Caruana , Giles Hooker , and Yin Lou . 2017. Detecting bias in black-box models using transparent model distillation. arXiv preprint arXiv:1710.06169 ( 2017 ). Sarah Tan, Rich Caruana, Giles Hooker, and Yin Lou. 2017. Detecting bias in black-box models using transparent model distillation. arXiv preprint arXiv:1710.06169 (2017)."},{"key":"e_1_3_2_1_56_1","doi-asserted-by":"publisher","DOI":"10.1145\/3287560.3287566"},{"key":"e_1_3_2_1_57_1","volume-title":"Counterfactual Explanations for Machine Learning: A Review. arXiv preprint arXiv:2010.10596","author":"Verma Sahil","year":"2020","unstructured":"Sahil Verma , John Dickerson , and Keegan Hines . 2020. Counterfactual Explanations for Machine Learning: A Review. arXiv preprint arXiv:2010.10596 ( 2020 ). Sahil Verma, John Dickerson, and Keegan Hines. 2020. Counterfactual Explanations for Machine Learning: A Review. arXiv preprint arXiv:2010.10596 (2020)."},{"key":"e_1_3_2_1_58_1","first-page":"841","article-title":"Counterfactual explanations without opening the black box: Automated decisions and the GDPR","volume":"31","author":"Wachter Sandra","year":"2017","unstructured":"Sandra Wachter , Brent Mittelstadt , and Chris Russell . 2017 . Counterfactual explanations without opening the black box: Automated decisions and the GDPR . Harv. JL & Tech. , Vol. 31 (2017), 841 . Sandra Wachter, Brent Mittelstadt, and Chris Russell. 2017. Counterfactual explanations without opening the black box: Automated decisions and the GDPR. Harv. JL & Tech., Vol. 31 (2017), 841.","journal-title":"Harv. JL & Tech."},{"key":"e_1_3_2_1_59_1","doi-asserted-by":"publisher","DOI":"10.1215\/00318108-2005-001"},{"key":"e_1_3_2_1_60_1","volume-title":"David I Inouye, and Pradeep Ravikumar.","author":"Yeh Chih-Kuan","year":"2019","unstructured":"Chih-Kuan Yeh , Cheng-Yu Hsieh , Arun Sai Suggala , David I Inouye, and Pradeep Ravikumar. 2019 . On the (in) fidelity and sensitivity for explanations. arXiv preprint arXiv:1901.09392 (2019). Chih-Kuan Yeh, Cheng-Yu Hsieh, Arun Sai Suggala, David I Inouye, and Pradeep Ravikumar. 2019. On the (in) fidelity and sensitivity for explanations. arXiv preprint arXiv:1901.09392 (2019)."},{"key":"e_1_3_2_1_61_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1420"},{"key":"e_1_3_2_1_62_1","volume-title":"arXiv preprint arXiv:1904.12991","author":"Zhang Yujia","year":"2019","unstructured":"Yujia Zhang , Kuangyan Song , Yiming Sun , Sarah Tan , and Madeleine Udell . 2019. \" Why Should You Trust My Explanation?\" Understanding Uncertainty in LIME Explanations . arXiv preprint arXiv:1904.12991 ( 2019 ). Yujia Zhang, Kuangyan Song, Yiming Sun, Sarah Tan, and Madeleine Udell. 2019. \" Why Should You Trust My Explanation?\" Understanding Uncertainty in LIME Explanations. arXiv preprint arXiv:1904.12991 (2019)."}],"event":{"name":"AIES '21: AAAI\/ACM Conference on AI, Ethics, and Society","location":"Virtual Event USA","acronym":"AIES '21","sponsor":["SIGAI ACM Special Interest Group on Artificial Intelligence","AAAI"]},"container-title":["Proceedings of the 2021 AAAI\/ACM Conference on AI, Ethics, and Society"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3461702.3462597","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3461702.3462597","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T20:17:06Z","timestamp":1750191426000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3461702.3462597"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,7,21]]},"references-count":62,"alternative-id":["10.1145\/3461702.3462597","10.1145\/3461702"],"URL":"https:\/\/doi.org\/10.1145\/3461702.3462597","relation":{},"subject":[],"published":{"date-parts":[[2021,7,21]]},"assertion":[{"value":"2021-07-30","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}