{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,14]],"date-time":"2026-04-14T06:16:21Z","timestamp":1776147381157,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":81,"publisher":"ACM","license":[{"start":{"date-parts":[[2021,3,1]],"date-time":"2021-03-01T00:00:00Z","timestamp":1614556800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"DOI":"10.13039\/501100000781","name":"European Research Council","doi-asserted-by":"publisher","award":["802774"],"award-info":[{"award-number":["802774"]}],"id":[{"id":"10.13039\/501100000781","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2021,3,3]]},"DOI":"10.1145\/3442188.3445923","type":"proceedings-article","created":{"date-parts":[[2021,3,3]],"date-time":"2021-03-03T01:26:24Z","timestamp":1614734784000},"page":"624-635","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":420,"title":["Formalizing Trust in Artificial Intelligence"],"prefix":"10.1145","author":[{"given":"Alon","family":"Jacovi","sequence":"first","affiliation":[{"name":"Bar Ilan University"}]},{"given":"Ana","family":"Marasovi\u0107","sequence":"additional","affiliation":[{"name":"Allen Institute for Artificial Intelligence, University of Washington"}]},{"given":"Tim","family":"Miller","sequence":"additional","affiliation":[{"name":"School of Computing and Information Systems, The University of Melbourne"}]},{"given":"Yoav","family":"Goldberg","sequence":"additional","affiliation":[{"name":"Bar Ilan University, Allen Institute for Artificial Intelligence"}]}],"member":"320","published-online":{"date-parts":[[2021,3]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"David Alvarez Melis and Tommi Jaakkola. 2018. Towards Robust Interpretability with Self-Explaining Neural Networks. In Advances in Neural Information Processing Systems 31 S. Bengio H. Wallach H. Larochelle K. Grauman N. Cesa-Bianchi and R. Garnett (Eds.). 7775--7784. http:\/\/papers.nips.cc\/paper\/8003-towards-robust-interpretability-with-self-explaining-neural-networks.pdf"},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.1147\/JRD.2019.2942288"},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_1_4_1","volume-title":"Trust and antitrust. ethics 96, 2","author":"Baier Annette","year":"1986","unstructured":"Annette Baier. 1986. Trust and antitrust. ethics 96, 2 (1986), 231--260."},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1162\/tacl_a_00041"},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.3390\/electronics8080832"},{"key":"e_1_3_2_1_9_1","volume-title":"An Interpretable Model with Globally Consistent Explanations for Credit Risk. CoRR abs\/1811.12615","author":"Chen Chaofan","year":"2018","unstructured":"Chaofan Chen, Kangcheng Lin, Cynthia Rudin, Yaron Shaposhnik, Sijia Wang, and Tong Wang. 2018. An Interpretable Model with Globally Consistent Explanations for Credit Risk. CoRR abs\/1811.12615 (2018). arXiv:1811.12615 http:\/\/arxiv.org\/abs\/1811.12615"},{"key":"e_1_3_2_1_10_1","volume-title":"Glass","author":"Dalvi F.","year":"2019","unstructured":"F. Dalvi, Nadir Durrani, Hassan Sajjad, Yonatan Belinkov, A. Bau, and James R. Glass. 2019. What Is One Grain of Sand in the Desert? Analyzing Individual Neurons in Deep NLP Models. In AAAI."},{"key":"e_1_3_2_1_11_1","unstructured":"Arun Das and Paul Rad. 2020. Opportunities and Challenges in Explainable Artificial Intelligence (XAI): A Survey. arXiv:2006.11371 [cs.CV]"},{"key":"e_1_3_2_1_12_1","volume-title":"Manning","author":"de Vries Harm","year":"2020","unstructured":"Harm de Vries, Dzmitry Bahdanau, and Christopher D. Manning. 2020. Towards Ecologically Valid Research on Language User Interfaces. CoRR abs\/2007.14435 (2020). arXiv:2007.14435 https:\/\/arxiv.org\/abs\/2007.14435"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_1_15_1","volume-title":"Towards a rigorous science of interpretable machine learning. arXiv preprint arXiv:1702.08608","author":"Doshi-Velez Finale","year":"2017","unstructured":"Finale Doshi-Velez and Been Kim. 2017. Towards a rigorous science of interpretable machine learning. arXiv preprint arXiv:1702.08608 (2017)."},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_1_17_1","volume-title":"Evaluating NLP Models via Contrast Sets. CoRR abs\/2004.02709","author":"Gardner Matt","year":"2020","unstructured":"Matt Gardner, Yoav Artzi, Victoria Basmova, Jonathan Berant, Ben Bogin, Sihao Chen, Pradeep Dasigi, Dheeru Dua, Yanai Elazar, Ananth Gottumukkala, Nitish Gupta, Hanna Hajishirzi, Gabriel Ilharco, Daniel Khashabi, Kevin Lin, Jiangming Liu, Nelson F. Liu, Phoebe Mulcaire, Qiang Ning, Sameer Singh, Noah A. Smith, Sanjay Subramanian, Reut Tsarfaty, Eric Wallace, Ally Zhang, and Ben Zhou. 2020. Evaluating NLP Models via Contrast Sets. CoRR abs\/2004.02709 (2020). arXiv:2004.02709 https:\/\/arxiv.org\/abs\/2004.02709"},{"key":"e_1_3_2_1_18_1","volume-title":"Proceedings of the 5th Workshop on Fairness, Accountability, and Transparency in Machine Learning.","author":"Gebru Timnit","year":"2018","unstructured":"Timnit Gebru, Jamie Morgenstern, Briana Vecchione, Jennifer Wortman Vaughan, Hanna M. Wallach, Hal Daum\u00e9, and Kate Crawford. 2018. Datasheets for Datasets. In Proceedings of the 5th Workshop on Fairness, Accountability, and Transparency in Machine Learning."},{"key":"e_1_3_2_1_19_1","volume-title":"ClinicalVis: Supporting Clinical Task-Focused Design Evaluation. CoRR abs\/1810.05798","author":"Ghassemi Marzyeh","year":"2018","unstructured":"Marzyeh Ghassemi, Mahima Pushkarna, James Wexler, Jesse Johnson, and Paul Varghese. 2018. ClinicalVis: Supporting Clinical Task-Focused Design Evaluation. CoRR abs\/1810.05798 (2018). arXiv:1810.05798 http:\/\/arxiv.org\/abs\/1810.05798"},{"key":"e_1_3_2_1_20_1","volume-title":"Towards Automatic Concept-based Explanations. In Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019","author":"Ghorbani Amirata","year":"2019","unstructured":"Amirata Ghorbani, James Wexler, James Y. Zou, and Been Kim. 2019. Towards Automatic Concept-based Explanations. In Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, NeurIPS 2019, 8-14 December 2019, Vancouver, BC, Canada, Hanna M. Wallach, Hugo Larochelle, Alina Beygelzimer, Florence d'Alch\u00e9-Buc, Emily B. Fox, and Roman Garnett (Eds.). 9273--9282. http:\/\/papers.nips.cc\/paper\/9126-towards-automatic-concept-based-explanations"},{"key":"e_1_3_2_1_21_1","volume-title":"3rd International Conference on Learning Representations, ICLR","author":"Goodfellow Ian J.","year":"2015","unstructured":"Ian J. Goodfellow, Jonathon Shlens, and Christian Szegedy. 2015. Explaining and Harnessing Adversarial Examples. In 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings, Yoshua Bengio and Yann LeCun (Eds.). http:\/\/arxiv.org\/abs\/1412.6572"},{"key":"e_1_3_2_1_22_1","volume-title":"Counterfactual Visual Explanations (Proceedings of Machine Learning Research","author":"Goyal Yash","unstructured":"Yash Goyal, Ziyan Wu, Jan Ernst, Dhruv Batra, Devi Parikh, and Stefan Lee. 2019. Counterfactual Visual Explanations (Proceedings of Machine Learning Research, Vol. 97), Kamalika Chaudhuri and Ruslan Salakhutdinov (Eds.). PMLR, Long Beach, California, USA, 2376--2384. http:\/\/proceedings.mlr.press\/v97\/goyal19a.html"},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1145\/3236009"},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_1_25_1","volume-title":"The Stanford Encyclopedia of Philosophy (fall 2018 ed.), Edward N","author":"Hansson Sven Ove","unstructured":"Sven Ove Hansson. 2018. Risk. In The Stanford Encyclopedia of Philosophy (fall 2018 ed.), Edward N. Zalta (Ed.). Metaphysics Research Lab, Stanford University."},{"key":"e_1_3_2_1_26_1","volume-title":"Evaluating Explainable AI: Which Algorithmic Explanations Help Users Predict Model Behavior? CoRR abs\/2005.01831","author":"Hase Peter","year":"2020","unstructured":"Peter Hase and Mohit Bansal. 2020. Evaluating Explainable AI: Which Algorithmic Explanations Help Users Predict Model Behavior? CoRR abs\/2005.01831 (2020). arXiv:2005.01831 https:\/\/arxiv.org\/abs\/2005.01831"},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1111\/nous.12000"},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"crossref","unstructured":"Lisa Anne Hendricks Ronghang Hu Trevor Darrell and Zeynep Akata. 2018. Grounding Visual Explanations. In ECCV. https:\/\/arxiv.org\/abs\/1807.09685","DOI":"10.1007\/978-3-030-01216-8_17"},{"key":"e_1_3_2_1_29_1","volume-title":"A taxonomy of emergent trusting in the human-machine relationship. Cognitive systems engineering: The future for a changing world","author":"Hoffman Robert R","year":"2017","unstructured":"Robert R Hoffman. 2017. A taxonomy of emergent trusting in the human-machine relationship. Cognitive systems engineering: The future for a changing world (2017), 137--163."},{"key":"e_1_3_2_1_30_1","volume-title":"99th Seminar","author":"Hofstede Gert Jan","year":"2006","unstructured":"Gert Jan Hofstede. 2006. Intrinsic and Enforceable Trust: A Research Agenda. European Association of Agricultural Economists, 99th Seminar, February 8-10, 2006, Bonn, Germany (01 2006)."},{"key":"e_1_3_2_1_31_1","volume-title":"Aligning Faithful Interpretations with their Social Attribution. CoRR abs\/2006.01067","author":"Jacovi Alon","year":"2020","unstructured":"Alon Jacovi and Yoav Goldberg. 2020. Aligning Faithful Interpretations with their Social Attribution. CoRR abs\/2006.01067 (2020). arXiv:2006.01067 https:\/\/arxiv.org\/abs\/2006.01067"},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_1_33_1","volume-title":"Interpreting Interpretability: Understanding Data Scientists' Use of Interpretability Tools for Machine Learning.","author":"Kaur Harmanpreet","year":"2019","unstructured":"Harmanpreet Kaur, Harsha Nori, Samuel Jenkins, Rich Caruana, Hanna M. Wallach, and Jennifer Wortman Vaughan. 2019. Interpreting Interpretability: Understanding Data Scientists' Use of Interpretability Tools for Machine Learning."},{"key":"e_1_3_2_1_34_1","volume-title":"Learning The Difference That Makes A Difference With Counterfactually-Augmented Data. In 8th International Conference on Learning Representations, ICLR 2020","author":"Kaushik Divyansh","year":"2020","unstructured":"Divyansh Kaushik, Eduard H. Hovy, and Zachary Chase Lipton. 2020. Learning The Difference That Makes A Difference With Counterfactually-Augmented Data. In 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26-30, 2020. OpenReview.net."},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_1_36_1","unstructured":"Been Kim Martin Wattenberg Justin Gilmer Carrie Cai James Wexler Fernanda Viegas and Rory Sayres. 2017. Interpretability Beyond Feature Attribution: Quantitative Testing with Concept Activation Vectors (TCAV). arXiv:1711.11279 [stat.ML]"},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.1145\/3328526.3329621"},{"key":"e_1_3_2_1_39_1","volume-title":"International Convention Centre","author":"Koh Pang Wei","year":"2017","unstructured":"Pang Wei Koh and Percy Liang. 2017. Understanding Black-box Predictions via Influence Functions (Proceedings of Machine Learning Research, Vol. 70), Doina Precup and Yee Whye Teh (Eds.). PMLR, International Convention Centre, Sydney, Australia, 1885--1894. http:\/\/proceedings.mlr.press\/v70\/koh17a.html"},{"key":"e_1_3_2_1_40_1","volume-title":"DEXAHAI","author":"Kopecka Hana","year":"2020","unstructured":"Hana Kopecka and Jose M Such. 2020. Explainable AI for Cultural Minds. https:\/\/sites.google.com\/view\/dexahai-at-ecai2020\/home Workshop on Dialogue, Explanation and Argumentation for Human-Agent Interaction, DEXAHAI; Conference date: 07-09-2020."},{"key":"e_1_3_2_1_41_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1145\/3290605.3300717"},{"key":"e_1_3_2_1_43_1","volume-title":"The parable of Google Flu: traps in big data analysis. Science 343, 6176","author":"Lazer David","year":"2014","unstructured":"David Lazer, Ryan Kennedy, Gary King, and Alessandro Vespignani. 2014. The parable of Google Flu: traps in big data analysis. Science 343, 6176 (2014), 1203--1205."},{"key":"e_1_3_2_1_44_1","volume-title":"Trust in automation: Designing for appropriate reliance. Human factors 46, 1","author":"Lee John D","year":"2004","unstructured":"John D Lee and Katrina A See. 2004. Trust in automation: Designing for appropriate reliance. Human factors 46, 1 (2004), 50--80."},{"key":"e_1_3_2_1_45_1","volume-title":"Trust as a social reality. Social forces 63, 4","author":"David Lewis J","year":"1985","unstructured":"J David Lewis and Andrew Weigert. 1985. Trust as a social reality. Social forces 63, 4 (1985), 967--985."},{"key":"e_1_3_2_1_46_1","volume-title":"Decision making and change in human affairs","author":"Lichtenstein Sarah","unstructured":"Sarah Lichtenstein, Baruch Fischhoff, and Lawrence D Phillips. 1977. Calibration of probabilities: The state of the art. In Decision making and change in human affairs. Springer, 275--324."},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.1145\/3233231"},{"key":"e_1_3_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_1_49_1","volume-title":"Advances in Neural Information Processing Systems 32, H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alch\u00e9-Buc","author":"Lubars Brian","unstructured":"Brian Lubars and Chenhao Tan. 2019. Ask not what AI can do, but what AI should do: Towards a framework of task delegability. In Advances in Neural Information Processing Systems 32, H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alch\u00e9-Buc, E. Fox, and R. Garnett (Eds.). Curran Associates, Inc., 57--67."},{"key":"e_1_3_2_1_50_1","doi-asserted-by":"publisher","DOI":"10.1145\/3313831.3376445"},{"key":"e_1_3_2_1_51_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_1_52_1","doi-asserted-by":"publisher","DOI":"10.2307\/258792"},{"key":"e_1_3_2_1_53_1","volume-title":"The Stanford Encyclopedia of Philosophy (fall 2015 ed.), Edward N","author":"McLeod Carolyn","unstructured":"Carolyn McLeod. 2015. Trust. In The Stanford Encyclopedia of Philosophy (fall 2015 ed.), Edward N. Zalta (Ed.). Metaphysics Research Lab, Stanford University."},{"key":"e_1_3_2_1_54_1","volume-title":"Contrastive Explanation: A Structural-Model Approach. CoRR abs\/1811.03163","author":"Miller Tim","year":"2018","unstructured":"Tim Miller. 2018. Contrastive Explanation: A Structural-Model Approach. CoRR abs\/1811.03163 (2018). arXiv:1811.03163 http:\/\/arxiv.org\/abs\/1811.03163"},{"key":"e_1_3_2_1_55_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.artint.2018.07.007"},{"key":"e_1_3_2_1_56_1","volume-title":"Trust in Modern Societies: The Search for the Bases of Social Order","author":"Misztal B.","unstructured":"B. Misztal. 1996. Trust in Modern Societies: The Search for the Bases of Social Order. Wiley. https:\/\/books.google.co.il\/books?id=q3R1QgAACAAJ"},{"key":"e_1_3_2_1_57_1","doi-asserted-by":"publisher","DOI":"10.1145\/3287560.3287596"},{"key":"e_1_3_2_1_58_1","doi-asserted-by":"publisher","DOI":"10.1518\/001872097778543886"},{"key":"e_1_3_2_1_59_1","unstructured":"Joelle Pineau. 2020. The Machine Learning Reproducibility Checklist. https:\/\/www.cs.mcgill.ca\/~jpineau\/ReproducibilityChecklist.pdf."},{"key":"e_1_3_2_1_60_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_1_61_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_1_62_1","volume-title":"IUI Workshops.","author":"Mireia","unstructured":"Mireia Ribera and \u00c0gata Lapedriza. 2019. Can we do better explanations? A proposal of user-centered explainable AI. In IUI Workshops."},{"key":"e_1_3_2_1_63_1","unstructured":"D. Schlangen. 2020. Targeting the Benchmark: On Methodology in Current Natural Language Processing Research. (2020). https:\/\/arxiv.org\/abs\/2007.04792 arXiv:2007.04792."},{"key":"e_1_3_2_1_64_1","volume-title":"Quantifying Interpretability and Trust in Machine Learning Systems. CoRR abs\/1901.08558","author":"Schmidt Philipp","year":"2019","unstructured":"Philipp Schmidt and Felix Bie\u00dfmann. 2019. Quantifying Interpretability and Trust in Machine Learning Systems. CoRR abs\/1901.08558 (2019). arXiv:1901.08558 http:\/\/arxiv.org\/abs\/1901.08558"},{"key":"e_1_3_2_1_65_1","volume-title":"2nd International Conference on Learning Representations ICLR, Workshop Track Proceedings. https:\/\/arxiv.org\/abs\/1312.6034","author":"Simonyan K.","unstructured":"K. Simonyan, A. Vedaldi, and Andrew Zisserman. 2014. Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps. In 2nd International Conference on Learning Representations ICLR, Workshop Track Proceedings. https:\/\/arxiv.org\/abs\/1312.6034"},{"key":"e_1_3_2_1_66_1","doi-asserted-by":"publisher","DOI":"10.1145\/3313831.3376624"},{"key":"e_1_3_2_1_67_1","volume-title":"We Need to Talk About Random Splits. CoRR abs\/2005.00636","author":"S\u00f8gaard Anders","year":"2020","unstructured":"Anders S\u00f8gaard, Sebastian Ebert, Jasmijn Bastings, and Katja Filippova. 2020. We Need to Talk About Random Splits. CoRR abs\/2005.00636 (2020). arXiv:2005.00636 https:\/\/arxiv.org\/abs\/2005.00636"},{"key":"e_1_3_2_1_68_1","doi-asserted-by":"publisher","DOI":"10.2307\/3857326"},{"key":"e_1_3_2_1_69_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_1_70_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P19-1159"},{"key":"e_1_3_2_1_71_1","doi-asserted-by":"publisher","DOI":"10.1002\/tht3.259"},{"key":"e_1_3_2_1_72_1","doi-asserted-by":"publisher","DOI":"10.1007\/s40926-019-00107-y"},{"key":"e_1_3_2_1_73_1","volume-title":"oLMpics - On what Language Model Pre-training Captures. CoRR abs\/1912.13283","author":"Talmor Alon","year":"2019","unstructured":"Alon Talmor, Yanai Elazar, Yoav Goldberg, and Jonathan Berant. 2019. oLMpics - On what Language Model Pre-training Captures. CoRR abs\/1912.13283 (2019). arXiv:1912.13283 http:\/\/arxiv.org\/abs\/1912.13283"},{"key":"e_1_3_2_1_74_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1"},{"key":"e_1_3_2_1_75_1","doi-asserted-by":"publisher","DOI":"10.1007\/s12525-020-00441-4"},{"key":"e_1_3_2_1_76_1","volume-title":"A Survey on Explainable Artificial Intelligence (XAI): Towards Medical XAI. CoRR abs\/1907.07374","author":"Tjoa Erico","year":"2019","unstructured":"Erico Tjoa and Cuntai Guan. 2019. A Survey on Explainable Artificial Intelligence (XAI): Towards Medical XAI. CoRR abs\/1907.07374 (2019). arXiv:1907.07374 http:\/\/arxiv.org\/abs\/1907.07374"},{"key":"e_1_3_2_1_77_1","doi-asserted-by":"publisher","DOI":"10.1145\/3351095.3372834"},{"key":"e_1_3_2_1_78_1","doi-asserted-by":"publisher","DOI":"10.1145\/3294052.3322192"},{"key":"e_1_3_2_1_79_1","unstructured":"Sarah Myers West Meredith Whittaker and Kate Crawford. 2019. Discriminating systems: Gender race and power in AI. (2019). https:\/\/ainowinstitute.org\/discriminatingsystems.pdf"},{"key":"e_1_3_2_1_80_1","doi-asserted-by":"publisher","DOI":"10.1007\/s11406-009-9218-0"},{"key":"e_1_3_2_1_81_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-32236-6_51"}],"event":{"name":"FAccT '21: 2021 ACM Conference on Fairness, Accountability, and Transparency","location":"Virtual Event Canada","acronym":"FAccT '21","sponsor":["ACM Association for Computing Machinery"]},"container-title":["Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3442188.3445923","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3442188.3445923","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T20:48:57Z","timestamp":1750193337000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3442188.3445923"}},"subtitle":["Prerequisites, Causes and Goals of Human Trust in AI"],"short-title":[],"issued":{"date-parts":[[2021,3]]},"references-count":81,"alternative-id":["10.1145\/3442188.3445923","10.1145\/3442188"],"URL":"https:\/\/doi.org\/10.1145\/3442188.3445923","relation":{},"subject":[],"published":{"date-parts":[[2021,3]]},"assertion":[{"value":"2021-03-01","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}