{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,27]],"date-time":"2026-02-27T11:47:04Z","timestamp":1772192824913,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":71,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,6,20]],"date-time":"2022-06-20T00:00:00Z","timestamp":1655683200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"DOI":"10.13039\/100000001","name":"National Science Foundation","doi-asserted-by":"publisher","award":["1916505, 1934464, 1928614, 2129076"],"award-info":[{"award-number":["1916505, 1934464, 1928614, 2129076"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2022,6,21]]},"DOI":"10.1145\/3531146.3533090","type":"proceedings-article","created":{"date-parts":[[2022,6,20]],"date-time":"2022-06-20T14:27:10Z","timestamp":1655735230000},"page":"248-266","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":71,"title":["It\u2019s Just Not That Simple: An Empirical Study of the Accuracy-Explainability Trade-off in Machine Learning for Public Policy"],"prefix":"10.1145","author":[{"given":"Andrew","family":"Bell","sequence":"first","affiliation":[{"name":"New York University, USA"}]},{"given":"Ian","family":"Solano-Kamaiko","sequence":"additional","affiliation":[{"name":"New York University, USA"}]},{"given":"Oded","family":"Nov","sequence":"additional","affiliation":[{"name":"New York University, USA"}]},{"given":"Julia","family":"Stoyanovich","sequence":"additional","affiliation":[{"name":"New York University, USA"}]}],"member":"320","published-online":{"date-parts":[[2022,6,20]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1145\/3313831.3376615"},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.1145\/2723576.2723619"},{"key":"e_1_3_2_1_3_1","first-page":"3","article-title":"Instance-based prediction of heart-disease presence with the Cleveland database","volume":"3","author":"Aha D","year":"1988","unstructured":"D Aha and Dennis Kibler. 1988. Instance-based prediction of heart-disease presence with the Cleveland database. University of California 3, 1 (1988), 3\u20132.","journal-title":"University of California"},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.3233\/978-1-60750-754-3-11"},{"key":"e_1_3_2_1_5_1","unstructured":"Kasun Amarasinghe Kit Rodolfa Hemank Lamba and Rayid Ghani. 2020. Explainable machine learning for public policy: Use cases gaps and research directions. arXiv preprint arXiv:2010.14374(2020)."},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","unstructured":"Ryan\u00a0S Baker and Aaron Hawn. 2021. Algorithmic Bias in Education. https:\/\/doi.org\/10.35542\/osf.io\/pbmvz","DOI":"10.35542\/osf.io"},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","unstructured":"Robert Bartlett Adair Morse Richard Stanton and Nancy Wallace. 2021. Consumer-lending discrimination in the FinTech Era. Journal of Financial Economics(2021). https:\/\/doi.org\/10.1016\/j.jfineco.2021.05.047","DOI":"10.1016\/j.jfineco.2021.05.047"},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-20055-8_4"},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICHI.2019.8904616"},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"crossref","unstructured":"Umang Bhatt Alice Xiang Shubham Sharma Adrian Weller Ankur Taly Yunhan Jia Joydeep Ghosh Ruchir Puri Jos\u00e9 M.\u00a0F. Moura and Peter Eckersley. 2020. Explainable Machine Learning in Deployment. arxiv:1909.06342\u00a0[cs.LG]","DOI":"10.1145\/3351095.3375624"},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1145\/3377325.3377498"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1145\/2939672.2939698"},{"key":"e_1_3_2_1_13_1","unstructured":"Alexandra Chouldechova. 2016. Fair prediction with disparate impact: A study of bias in recidivism prediction instruments. arxiv:1610.07524\u00a0[stat.AP]"},{"key":"e_1_3_2_1_14_1","volume-title":"Federal Deposit\u00a0Insurance Corporation","author":"FDIC","year":"1968","unstructured":"FDIC: Federal Deposit\u00a0Insurance Corporation. 1968. Civil Rights Act of 1968. (1968). https:\/\/www.fdic.gov\/regulations\/laws\/rules\/6000-1400.html"},{"key":"e_1_3_2_1_15_1","unstructured":"P. Cortez and A.\u00a0M.\u00a0G. Silva. 2008. Using data mining to predict secondary school student performance."},{"key":"e_1_3_2_1_16_1","unstructured":"Ian Covert Scott Lundberg and Su-In Lee. 2020. Understanding global feature contributions with additive importance measures. arXiv preprint arXiv:2004.00668(2020)."},{"key":"e_1_3_2_1_17_1","unstructured":"Ian Covert Scott\u00a0M. Lundberg and Su-In Lee. 2020. Understanding Global Feature Contributions Through Additive Importance Measures. CoRR abs\/2004.00668(2020). arXiv:2004.00668https:\/\/arxiv.org\/abs\/2004.00668"},{"key":"e_1_3_2_1_18_1","unstructured":"Arun Das and Paul Rad. 2020. Opportunities and Challenges in Explainable Artificial Intelligence (XAI): A Survey. arxiv:2006.11371\u00a0[cs.CV]"},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1109\/SP.2016.42"},{"key":"e_1_3_2_1_20_1","volume-title":"NeurIPS Workshop on AI for Social Good.","author":"de Troya Inigo\u00a0Martinez","year":"2018","unstructured":"Inigo\u00a0Martinez de Troya, Ruqian Chen, Laura\u00a0O Moraes, Pranjal Bajaj, Jordan Kupersmith, Rayid Ghani, Nuno\u00a0B Br\u00e1s, and Leid Zejnilovic. 2018. Predicting, explaining, and understanding risk of long-term unemployment. In NeurIPS Workshop on AI for Social Good."},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1093\/jamia\/ocaa094"},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1093\/jamia"},{"key":"e_1_3_2_1_23_1","volume-title":"Proceedings of the ACM on Human-Computer Interaction 4, CSCW2(2020)","author":"Dove Graham","year":"2020","unstructured":"Graham Dove, Martina Balestra, Devin Mann, and Oded Nov. 2020. Good for the Many or Best for the Few? A Dilemma in the Design of Algorithmic Advice. Proceedings of the ACM on Human-Computer Interaction 4, CSCW2(2020), 1\u201322."},{"key":"e_1_3_2_1_24_1","unstructured":"Gintare\u00a0Karolina Dziugaite Shai Ben-David and Daniel\u00a0M Roy. 2020. Enforcing Interpretability and its Statistical Impacts: Trade-offs between Accuracy and Interpretability. arXiv preprint arXiv:2010.13764(2020)."},{"key":"e_1_3_2_1_25_1","volume-title":"Automating Inequality: How High-Tech Tools Profile, Police, and Punish the Poor","author":"Eubanks Virginia","year":"2018","unstructured":"Virginia Eubanks. 2018. Automating Inequality: How High-Tech Tools Profile, Police, and Punish the Poor. St. Martin\u2019s Press, Inc., USA."},{"key":"e_1_3_2_1_26_1","volume-title":"Predictably unequal? the effects of machine learning on credit markets. The Effects of Machine Learning on Credit Markets (October 1","author":"Fuster Andreas","year":"2020","unstructured":"Andreas Fuster, Paul Goldsmith-Pinkham, Tarun Ramadorai, and Ansgar Walther. 2020. Predictably unequal? the effects of machine learning on credit markets. The Effects of Machine Learning on Credit Markets (October 1, 2020) (2020)."},{"key":"e_1_3_2_1_27_1","volume-title":"Can predictive algorithms assist decision-making in social work with children and families?Child abuse review 28, 2","author":"Gillingham Philip","year":"2019","unstructured":"Philip Gillingham. 2019. Can predictive algorithms assist decision-making in social work with children and families?Child abuse review 28, 2 (2019), 114\u2013126."},{"key":"e_1_3_2_1_28_1","volume-title":"A framework for considering comprehensibility in modeling. Big data 4, 2","author":"Gleicher Michael","year":"2016","unstructured":"Michael Gleicher. 2016. A framework for considering comprehensibility in modeling. Big data 4, 2 (2016), 75\u201388."},{"key":"e_1_3_2_1_29_1","doi-asserted-by":"publisher","DOI":"10.1609\/aimag.v38i3.2741"},{"key":"e_1_3_2_1_30_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.clsr.2022.105681"},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1145\/3236009"},{"key":"e_1_3_2_1_32_1","volume-title":"XAI\u2014Explainable artificial intelligence. Science Robotics 4, 37","author":"Gunning David","year":"2019","unstructured":"David Gunning, Mark Stefik, Jaesik Choi, Timothy Miller, Simone Stumpf, and Guang-Zhong Yang. 2019. XAI\u2014Explainable artificial intelligence. Science Robotics 4, 37 (2019)."},{"key":"e_1_3_2_1_33_1","unstructured":"Satoshi Hara and Kohei Hayashi. 2016. Making Tree Ensembles Interpretable. arxiv:1606.05390\u00a0[stat.ML]"},{"key":"e_1_3_2_1_34_1","unstructured":"Kenneth Holstein and Shayan Doroudi. 2021. Equity and Artificial Intelligence in Education: Will \u201dAIEd\u201d Amplify or Alleviate Inequities in Education?arxiv:2104.12920\u00a0[cs.HC]"},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"crossref","unstructured":"Andreas Holzinger Andr\u00e9 Carrington and Heimo M\u00fcller. 2020. Measuring the quality of explanations: the system causability scale (SCS). KI-K\u00fcnstliche Intelligenz(2020) 1\u20136.","DOI":"10.1007\/s13218-020-00636-z"},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.1145\/3392878"},{"key":"e_1_3_2_1_37_1","unstructured":"Qian|Rangwala Hu. 2020. Towards Fair Educational Data Mining: A Case Study on Detecting At-Risk Students.https:\/\/eric.ed.gov\/?id=ED608050"},{"key":"e_1_3_2_1_38_1","doi-asserted-by":"crossref","unstructured":"Johan Huysmans Bart Baesens and Jan Vanthienen. 2006. Using rule extraction to improve the comprehensibility of predictive models. (2006).","DOI":"10.2139\/ssrn.961358"},{"key":"e_1_3_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1145\/3442188.3445941"},{"key":"e_1_3_2_1_40_1","unstructured":"Jeff\u00a0Larson Julia\u00a0Angwin. 2016. Machine Bias. https:\/\/www.propublica.org\/article\/machine-bias-risk-assessments-in-criminal-sentencing"},{"key":"e_1_3_2_1_41_1","volume-title":"Detecting racial bias in algorithms and machine learning. Journal of Information, Communication and Ethics in Society","author":"Lee Nicol\u00a0Turner","year":"2018","unstructured":"Nicol\u00a0Turner Lee. 2018. Detecting racial bias in algorithms and machine learning. Journal of Information, Communication and Ethics in Society (2018)."},{"key":"e_1_3_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1145\/3313831.3376590"},{"key":"e_1_3_2_1_43_1","doi-asserted-by":"publisher","DOI":"10.1145\/1620545.1620576"},{"key":"e_1_3_2_1_44_1","doi-asserted-by":"publisher","DOI":"10.1145\/1518701.1519023"},{"key":"e_1_3_2_1_45_1","volume-title":"Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017","author":"M.","year":"2017","unstructured":"Scott\u00a0M. Lundberg and Su-In Lee. 2017. A Unified Approach to Interpreting Model Predictions. In Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, December 4-9, 2017, Long Beach, CA, USA, Isabelle Guyon, Ulrike von Luxburg, Samy Bengio, Hanna\u00a0M. Wallach, Rob Fergus, S.\u00a0V.\u00a0N. Vishwanathan, and Roman Garnett (Eds.). 4765\u20134774. https:\/\/proceedings.neurips.cc\/paper\/2017\/hash\/8a20a8621978632d76c43dfd28b67767-Abstract.html"},{"key":"e_1_3_2_1_46_1","unstructured":"Ricards Marcinkevics and Julia\u00a0E. Vogt. 2020. Interpretability and Explainability: A Machine Learning Zoo Mini-tour. CoRR abs\/2012.01805(2020). arxiv:2012.01805https:\/\/arxiv.org\/abs\/2012.01805"},{"key":"e_1_3_2_1_47_1","volume-title":"Explanation in artificial intelligence: Insights from the social sciences. Artificial intelligence 267","author":"Miller Tim","year":"2019","unstructured":"Tim Miller. 2019. Explanation in artificial intelligence: Insights from the social sciences. Artificial intelligence 267 (2019), 1\u201338."},{"key":"e_1_3_2_1_48_1","unstructured":"Christoph Molnar. 2020. Interpretable machine learning. Lulu. com."},{"key":"e_1_3_2_1_49_1","unstructured":"Menaka Narayanan Emily Chen Jeffrey He Been Kim Sam Gershman and Finale Doshi-Velez. 2018. How do humans understand explanations from machine learning systems? an evaluation of the human-interpretability of explanation. arXiv preprint arXiv:1802.00682(2018)."},{"key":"e_1_3_2_1_50_1","doi-asserted-by":"publisher","DOI":"10.1145\/3417518"},{"key":"e_1_3_2_1_51_1","volume-title":"Weapons of Math Destruction: How Big Data Increases Inequality and Threatens Democracy","author":"O\u2019Neil Cathy","unstructured":"Cathy O\u2019Neil. 2016. Weapons of Math Destruction: How Big Data Increases Inequality and Threatens Democracy. Crown Publishing Group, USA."},{"key":"e_1_3_2_1_52_1","doi-asserted-by":"publisher","DOI":"10.1002\/isaf.1422"},{"key":"e_1_3_2_1_53_1","doi-asserted-by":"publisher","DOI":"10.1017\/cls.2017.6"},{"key":"e_1_3_2_1_54_1","doi-asserted-by":"publisher","DOI":"10.1145\/2939672.2939778"},{"key":"e_1_3_2_1_55_1","doi-asserted-by":"crossref","unstructured":"Marco\u00a0Tulio Ribeiro Sameer Singh and Carlos Guestrin. 2016. \u201dWhy Should I Trust You?\u201d: Explaining the Predictions of Any Classifier. arxiv:1602.04938\u00a0[cs.LG]","DOI":"10.18653\/v1\/N16-3020"},{"key":"e_1_3_2_1_56_1","doi-asserted-by":"crossref","unstructured":"Kit\u00a0T Rodolfa Hemank Lamba and Rayid Ghani. 2020. Empirical observation of negligible fairness-accuracy trade-offs in machine learning for public policy. arXiv preprint arXiv:2012.02972(2020).","DOI":"10.1038\/s42256-021-00396-x"},{"key":"e_1_3_2_1_57_1","volume-title":"IL. Univerity of Chicago, Harris School of Public Policy Working Paper","author":"Ross Robert","year":"2017","unstructured":"Robert Ross. 2017. The impact of property tax appeals on vertical equity in Cook County, IL. Univerity of Chicago, Harris School of Public Policy Working Paper (2017)."},{"key":"e_1_3_2_1_58_1","doi-asserted-by":"crossref","unstructured":"Cynthia Rudin. 2019. Stop Explaining Black Box Machine Learning Models for High Stakes Decisions and Use Interpretable Models Instead. arxiv:1811.10154\u00a0[stat.ML]","DOI":"10.1038\/s42256-019-0048-x"},{"key":"e_1_3_2_1_59_1","unstructured":"Piotr Sapiezynski Valentin Kassarnig and Christo Wilson. 2017. Academic performance prediction in a gender-imbalanced environment."},{"key":"e_1_3_2_1_60_1","doi-asserted-by":"crossref","unstructured":"Dylan Slack Sophie Hilgard Emily Jia Sameer Singh and Himabindu Lakkaraju. 2020. Fooling LIME and SHAP: Adversarial Attacks on Post hoc Explanation Methods. arxiv:1911.02508\u00a0[cs.LG]","DOI":"10.1145\/3375627.3375830"},{"key":"e_1_3_2_1_61_1","doi-asserted-by":"publisher","DOI":"10.1145\/3375627.3375830"},{"key":"e_1_3_2_1_62_1","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0144439"},{"key":"e_1_3_2_1_63_1","doi-asserted-by":"publisher","DOI":"10.1038\/s42256-020-0171-8"},{"key":"e_1_3_2_1_64_1","doi-asserted-by":"publisher","DOI":"10.1145\/2641190.2641198"},{"key":"e_1_3_2_1_65_1","doi-asserted-by":"publisher","DOI":"10.1002\/poi3.198"},{"key":"e_1_3_2_1_66_1","doi-asserted-by":"publisher","DOI":"10.1017\/dap.2020.23"},{"key":"e_1_3_2_1_67_1","volume-title":"IUI Workshops.","author":"Yang Yiwei","year":"2019","unstructured":"Yiwei Yang, Eser Kandogan, Yunyao Li, Prithviraj Sen, and Walter\u00a0S Lasecki. 2019. A study on interaction in human-in-the-loop machine learning for text analytics. In IUI Workshops."},{"key":"e_1_3_2_1_68_1","doi-asserted-by":"publisher","DOI":"10.1109\/VIS49827.2021.9623303"},{"key":"e_1_3_2_1_69_1","first-page":"15424","article-title":"Machine Learning Informed Decision-Making with Interpreted Model\u2019s Outputs: A Field Intervention. In Academy of Management Proceedings, Vol.\u00a02021. Academy of Management Briarcliff Manor","volume":"10510","author":"Zejnilovic Leid","year":"2021","unstructured":"Leid Zejnilovic, Susana Lavado, Carlos Soares, \u00cd\u00f1igo Mart\u00ednez De Rituerto De\u00a0Troya, Andrew Bell, and Rayid Ghani. 2021. Machine Learning Informed Decision-Making with Interpreted Model\u2019s Outputs: A Field Intervention. In Academy of Management Proceedings, Vol.\u00a02021. Academy of Management Briarcliff Manor, NY 10510, 15424.","journal-title":"NY"},{"key":"e_1_3_2_1_70_1","doi-asserted-by":"publisher","DOI":"10.1525\/gp.2020.12908"},{"key":"e_1_3_2_1_71_1","volume-title":"Why Should You Trust My Explanation?","author":"Zhang Yujia","year":"1904","unstructured":"Yujia Zhang, Kuangyan Song, Yiming Sun, Sarah Tan, and Madeleine Udell. 2019. \u201d Why Should You Trust My Explanation?\u201d Understanding Uncertainty in LIME Explanations. arXiv preprint arXiv:1904.12991(2019)."}],"event":{"name":"FAccT '22: 2022 ACM Conference on Fairness, Accountability, and Transparency","location":"Seoul Republic of Korea","acronym":"FAccT '22","sponsor":["ACM Association for Computing Machinery"]},"container-title":["2022 ACM Conference on Fairness Accountability and Transparency"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3531146.3533090","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3531146.3533090","content-type":"application\/pdf","content-version":"vor","intended-application":"syndication"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3531146.3533090","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T19:02:10Z","timestamp":1750186930000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3531146.3533090"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,6,20]]},"references-count":71,"alternative-id":["10.1145\/3531146.3533090","10.1145\/3531146"],"URL":"https:\/\/doi.org\/10.1145\/3531146.3533090","relation":{},"subject":[],"published":{"date-parts":[[2022,6,20]]},"assertion":[{"value":"2022-06-20","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}