{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T15:09:11Z","timestamp":1776092951867,"version":"3.50.1"},"reference-count":44,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2022,1,24]],"date-time":"2022-01-24T00:00:00Z","timestamp":1642982400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,24]],"date-time":"2022-01-24T00:00:00Z","timestamp":1642982400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Ethics Inf Technol"],"published-print":{"date-parts":[[2022,3]]},"DOI":"10.1007\/s10676-022-09623-4","type":"journal-article","created":{"date-parts":[[2022,1,31]],"date-time":"2022-01-31T15:02:58Z","timestamp":1643641378000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":42,"title":["Fairness, explainability and in-between: understanding the impact of different explanation methods on non-expert users\u2019 perceptions of fairness toward an algorithmic system"],"prefix":"10.1007","volume":"24","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-2091-2966","authenticated-orcid":false,"given":"Avital","family":"Shulner-Tal","sequence":"first","affiliation":[]},{"given":"Tsvi","family":"Kuflik","sequence":"additional","affiliation":[]},{"given":"Doron","family":"Kliger","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,1,24]]},"reference":[{"key":"9623_CR1","doi-asserted-by":"crossref","unstructured":"Abdollahi, B., & Nasraoui, O. (2018). Transparency in fair machine learning: The case of explainable recommender systems. Human and Machine Learning (pp. 21\u201335). Springer.","DOI":"10.1007\/978-3-319-90403-0_2"},{"key":"9623_CR2","doi-asserted-by":"publisher","first-page":"52138","DOI":"10.1109\/ACCESS.2018.2870052","volume":"6","author":"A Adadi","year":"2018","unstructured":"Adadi, A., & Berrada, M. (2018). Peeking inside the black-box: A survey on Explainable Artificial Intelligence (XAI). IEEE Access, 6, 52138\u201352160.","journal-title":"IEEE Access"},{"key":"9623_CR3","doi-asserted-by":"crossref","unstructured":"Arrieta, A. B., D\u00edaz-Rodr\u00edguez, N., Del Ser, J., Bennetot, A., Tabik, S., Barbado, A., ... Chatila, R. (2020). Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI. Information Fusion, 58, 82-115.","DOI":"10.1016\/j.inffus.2019.12.012"},{"key":"9623_CR4","unstructured":"Arya, V., Bellamy, R. K., Chen, P. Y., Dhurandhar, A., Hind, M., Hoffman, S. C., ... Mourad, S. (2019). One explanation does not fit all: A toolkit and taxonomy of AI explainability techniques. arXiv preprint arXiv:1909.03012"},{"key":"9623_CR5","unstructured":"Barocas, S., Hardt, M., & Narayanan, A. (2018). Fairness and Machine Learning. fairmlbook. org. Retrieved from http:\/\/www.fairmlbook.org"},{"key":"9623_CR6","doi-asserted-by":"crossref","unstructured":"Binns, R., Van Kleek, M., Veale, M., Lyngs, U., Zhao, J., & Shadbolt, N. (2018, April). \u2018It\u2019s reducing a human being to a percentage\u2019 perceptions of justice in algorithmic decisions. In Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems (pp. 1\u201314).","DOI":"10.1145\/3173574.3173951"},{"key":"9623_CR7","unstructured":"Craven, M. W. (1996). Extracting comprehensible models from trained neural networks. University of Wisconsin-Madison Department of Computer Sciences."},{"key":"9623_CR8","doi-asserted-by":"crossref","unstructured":"Dodge, J., Liao, Q. V., Zhang, Y., Bellamy, R. K., & Dugan, C. (2019, March). Explaining models: an empirical study of how explanations impact fairness judgment. In Proceedings of the 24th International Conference on Intelligent User Interfaces (pp. 275\u2013285).","DOI":"10.1145\/3301275.3302310"},{"key":"9623_CR9","doi-asserted-by":"crossref","unstructured":"Do\u0161ilovi\u0107, F. K., Br\u010di\u0107, M., & Hlupi\u0107, N. (2018). Explainable artificial intelligence: A survey. 2018 41st International convention on information and communication technology, electronics and microelectronics (MIPRO) (pp. 0210\u20130215). IEEE.","DOI":"10.23919\/MIPRO.2018.8400040"},{"key":"9623_CR10","unstructured":"Eiband, M., Schneider, H., & Buschek, D. (2018). Normative vs. Pragmatic: Two perspectives on the design of explanations in intelligent systems. In: IUI workshops on explainable smart systems (EXSS)"},{"key":"9623_CR11","doi-asserted-by":"crossref","unstructured":"Eiband, M., Schneider, H., Bilandzic, M., Fazekas-Con, J., Haug, M., & Hussmann, H. (2018). Bringing transparency design into practice. 23rd International conference on intelligent user interfaces (pp. 211\u2013223). ACM.","DOI":"10.1145\/3172944.3172961"},{"key":"9623_CR12","unstructured":"Felfernig, A., Gula, B. (2006). Consumer behavior in the interaction with knowledge-based recommender applications. In: ECAI 2006 workshop on recommender systems, pp. 37\u201341"},{"key":"9623_CR13","doi-asserted-by":"crossref","unstructured":"Fernandez, A., Herrera, F., Cordon, O., del Jesus, M. J., & Marcelloni, F. (2019). Evolutionary fuzzy systems for explainable artificial intelligence: Why, when, what for, and where to? IEEE Computational Intelligence Magazine, 14(1), 69\u201381.","DOI":"10.1109\/MCI.2018.2881645"},{"issue":"4","key":"9623_CR14","doi-asserted-by":"publisher","first-page":"367","DOI":"10.1016\/j.ijhcs.2013.12.007","volume":"72","author":"F Gedikli","year":"2014","unstructured":"Gedikli, F., Jannach, D., & Ge, M. (2014). How should I explain? A comparison of different explanation types for recommender systems. International Journal of Human-Computer Studies, 72(4), 367\u2013382.","journal-title":"International Journal of Human-Computer Studies"},{"issue":"2","key":"9623_CR15","doi-asserted-by":"publisher","first-page":"75","DOI":"10.1089\/big.2016.0007","volume":"4","author":"M Gleicher","year":"2016","unstructured":"Gleicher, M. (2016). A framework for considering comprehensibility in modeling. Big Data, 4(2), 75\u201388.","journal-title":"Big Data"},{"issue":"3","key":"9623_CR16","doi-asserted-by":"publisher","first-page":"50","DOI":"10.1609\/aimag.v38i3.2741","volume":"38","author":"B Goodman","year":"2017","unstructured":"Goodman, B., & Flaxman, S. (2017). European Union regulations on algorithmic decision-making and a \u201cright to explanation.\u201d AI Magazine, 38(3), 50\u201357.","journal-title":"AI Magazine"},{"key":"9623_CR17","unstructured":"Green, B. (2018). \u201cFair\u201d risk assessments: A precarious approach for criminal justice reform. In: 5th Workshop on fairness, accountability, and transparency in machine learning."},{"key":"9623_CR18","unstructured":"Griffin, R. W., Phillips, J., & Gully, S. M. (2017). Organizational behavior: Managing people and organizations."},{"issue":"5","key":"9623_CR19","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3236009","volume":"51","author":"R Guidotti","year":"2018","unstructured":"Guidotti, R., Monreale, A., Ruggieri, S., Turini, F., Giannotti, F., & Pedreschi, D. (2018). A survey of methods for explaining black box models. ACM Computing Surveys (CSUR), 51(5), 1\u201342.","journal-title":"ACM Computing Surveys (CSUR)"},{"key":"9623_CR20","unstructured":"Gunning, D. (2017). Explainable artificial intelligence (xai). Defense Advanced Research Projects Agency (DARPA), nd Web, 2, 2."},{"issue":"2","key":"9623_CR21","doi-asserted-by":"publisher","first-page":"44","DOI":"10.1609\/aimag.v40i2.2850","volume":"40","author":"D Gunning","year":"2019","unstructured":"Gunning, D., & Aha, D. W. (2019). DARPA\u2019s explainable artificial intelligence program. AI Magazine, 40(2), 44\u201358.","journal-title":"AI Magazine"},{"key":"9623_CR22","doi-asserted-by":"crossref","unstructured":"Holzinger, A., Carrington, A., & M\u00fcller, H. (2020). Measuring the quality of explanations: The system causability scale (SCS) (pp. 1\u20136). KI-K\u00fcnstliche Intelligenz.","DOI":"10.1007\/s13218-020-00636-z"},{"issue":"4","key":"9623_CR23","first-page":"e1312","volume":"9","author":"A Holzinger","year":"2019","unstructured":"Holzinger, A., Langs, G., Denk, H., Zatloukal, K., & M\u00fcller, H. (2019). Causability and explainability of artificial intelligence in medicine. Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery, 9(4), e1312.","journal-title":"Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery"},{"key":"9623_CR24","doi-asserted-by":"crossref","unstructured":"Jesus, S., Bel\u00e9m, C., Balayan, V., Bento, J., Saleiro, P., Bizarro, P., & Gama, J. (2021). How can I choose an explainer? An Application-grounded Evaluation of Post-hoc Explanations. arXiv preprint arXiv:2101.08758","DOI":"10.1145\/3442188.3445941"},{"key":"9623_CR25","unstructured":"Kilbertus, N., Gasc\u00f3n, A., Kusner, M. J., Veale, M., Gummadi, K. P., & Weller, A. (2018). Blind justice: Fairness with encrypted sensitive attributes. arXiv preprint arXiv:1806.03281"},{"key":"9623_CR26","unstructured":"Kim, B., Glassman, E., Johnson, B., & Shah, J. (2015). iBCM: Interactive Bayesian case model empowering humans via intuitive interaction."},{"key":"9623_CR27","unstructured":"Lipton, Z. C. (2016). The mythos of model interpretability. arXiv preprint arXiv:1606.03490"},{"key":"9623_CR28","doi-asserted-by":"publisher","first-page":"154096","DOI":"10.1109\/ACCESS.2019.2949286","volume":"7","author":"O Loyola-Gonzalez","year":"2019","unstructured":"Loyola-Gonzalez, O. (2019). Black-box vs. white-box: Understanding their advantages and weaknesses from a practical point of view. IEEE Access, 7, 154096\u2013154113.","journal-title":"IEEE Access"},{"key":"9623_CR29","unstructured":"Lundberg, S. M., & Lee, S. I. (2017). A unified approach to interpreting model predictions. In: Advances in neural information processing systems (pp. 4765\u20134774)."},{"key":"9623_CR30","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.dsp.2017.10.011","volume":"73","author":"G Montavon","year":"2018","unstructured":"Montavon, G., Samek, W., & M\u00fcller, K. R. (2018). Methods for interpreting and understanding deep neural networks. Digital Signal Processing, 73, 1\u201315.","journal-title":"Digital Signal Processing"},{"issue":"3\u20135","key":"9623_CR31","doi-asserted-by":"publisher","first-page":"393","DOI":"10.1007\/s11257-017-9195-0","volume":"27","author":"I Nunes","year":"2017","unstructured":"Nunes, I., & Jannach, D. (2017). A systematic review and taxonomy of explanations in decision support and recommender systems. User Modeling and User-Adapted Interaction, 27(3\u20135), 393\u2013444.","journal-title":"User Modeling and User-Adapted Interaction"},{"issue":"1","key":"9623_CR32","doi-asserted-by":"publisher","first-page":"137","DOI":"10.1007\/s11747-019-00710-5","volume":"48","author":"A Rai","year":"2020","unstructured":"Rai, A. (2020). Explainable AI: From black box to glass box. Journal of the Academy of Marketing Science, 48(1), 137\u2013141.","journal-title":"Journal of the Academy of Marketing Science"},{"key":"9623_CR33","doi-asserted-by":"crossref","unstructured":"Ribeiro, M. T., Singh, S., & Guestrin, C. (2016). Why should I trust you?: Explaining the predictions of any classifier. Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining (pp. 1135\u20131144). ACM.","DOI":"10.1145\/2939672.2939778"},{"issue":"5","key":"9623_CR34","doi-asserted-by":"publisher","first-page":"206","DOI":"10.1038\/s42256-019-0048-x","volume":"1","author":"C Rudin","year":"2019","unstructured":"Rudin, C. (2019). Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nature Machine Intelligence, 1(5), 206\u2013215.","journal-title":"Nature Machine Intelligence"},{"key":"9623_CR35","unstructured":"Samek, W., Wiegand, T., & M\u00fcller, K. R. (2017). Explainable artificial intelligence: Understanding, visualizing and interpreting deep learning models. arXiv preprint arXiv:1708.08296"},{"key":"9623_CR36","unstructured":"Singh, C., Murdoch, W. J., & Yu, B. (2018). Hierarchical interpretations for neural network predictions. arXiv preprint arXiv:1806.05337"},{"key":"9623_CR37","doi-asserted-by":"crossref","unstructured":"Sinha, R., Swearingen, K. (2002). The role of transparency in recommender systems. In: Conference on Human Factors in Computing Systems, pp. 830\u2013831","DOI":"10.1145\/506443.506619"},{"key":"9623_CR38","doi-asserted-by":"publisher","unstructured":"Tal, A. S., Batsuren, K., Bogina, V., Giunchiglia, F., Hartman, A., Loizou, S. K., Kuflik, T. & Otterbacher, J. (2019) \u201cEnd to End\u201d towards a framework for reducing biases and promoting transparency of algorithmic systems. In: 2019 14th International Workshop on Semantic and Social Media Adaptation and Personalization (SMAP), Larnaca, Cyprus, , pp. 1-6. https:\/\/doi.org\/10.1109\/SMAP.2019.8864914","DOI":"10.1109\/SMAP.2019.8864914"},{"issue":"3","key":"9623_CR39","doi-asserted-by":"publisher","first-page":"230","DOI":"10.1080\/09540091.2017.1310182","volume":"29","author":"A Theodorou","year":"2017","unstructured":"Theodorou, A., Wortham, R. H., & Bryson, J. J. (2017). Designing and implementing transparency for real time inspection of autonomous robots. Connection Science, 29(3), 230\u2013241.","journal-title":"Connection Science"},{"key":"9623_CR40","doi-asserted-by":"crossref","unstructured":"Tintarev, N., & Masthoff, J. (2011). Designing and evaluating explanations for recommender systems. Recommender systems handbook (pp. 479\u2013510). Springer.","DOI":"10.1007\/978-0-387-85820-3_15"},{"issue":"CSCW","key":"9623_CR41","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3359130","volume":"3","author":"N Van Berkel","year":"2019","unstructured":"Van Berkel, N., Goncalves, J., Hettiachchi, D., Wijenayake, S., Kelly, R. M., & Kostakos, V. (2019). Crowdsourcing perceptions of fair predictors for machine learning: A recidivism case study. Proceedings of the ACM on Human-Computer Interaction, 3(CSCW), 1\u201321.","journal-title":"Proceedings of the ACM on Human-Computer Interaction"},{"issue":"4","key":"9623_CR42","doi-asserted-by":"publisher","first-page":"217","DOI":"10.2753\/MIS0742-1222230410","volume":"23","author":"W Wang","year":"2007","unstructured":"Wang, W., & Benbasat, I. (2007). Recommendation agents for electronic commerce: Effects of explanation facilities on trusting beliefs. Journal of Management Information Systems, 23(4), 217\u2013246.","journal-title":"Journal of Management Information Systems"},{"key":"9623_CR43","unstructured":"Wortham, R. H., Theodorou, A., & Bryson, J. J. (2016, June). What does the robot think? Transparency as a fundamental design requirement for intelligent systems. In: Ijcai-2016 ethics for artificial intelligence workshop."},{"key":"9623_CR44","unstructured":"Zhang, J. M., Harman, M., Ma, L., & Liu, Y. (2019). Machine learning testing: Survey, landscapes and horizons. arXiv preprint arXiv:1906.10742"}],"container-title":["Ethics and Information Technology"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10676-022-09623-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10676-022-09623-4\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10676-022-09623-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,3,22]],"date-time":"2022-03-22T19:15:33Z","timestamp":1647976533000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10676-022-09623-4"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,1,24]]},"references-count":44,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2022,3]]}},"alternative-id":["9623"],"URL":"https:\/\/doi.org\/10.1007\/s10676-022-09623-4","relation":{},"ISSN":["1388-1957","1572-8439"],"issn-type":[{"value":"1388-1957","type":"print"},{"value":"1572-8439","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,1,24]]},"assertion":[{"value":"5 January 2022","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"24 January 2022","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors have no conflict of interest to declare that are relevant to the content of this article.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"The experiments conducted in this study were approved by the Committee for Ethical Research and the Protection of Human Participants, University of Haifa, Israel (Approval 350\/19).","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}},{"value":"The following consent to take part in academic research was presented to the participants: \u201cThis research is conducted by researchers from the Departments of Information Systems and Economics at the University of Haifa, which deals with the transparency and fairness of algorithmic systems. We request your participation in this online study. It should be emphasized that the answers to the questionnaires will be kept confidential and used only for research purposes. No personal or identifying information is requested or kept. Your participation in this study is voluntary. If you decide at any time that you do not wish to participate, you may do so without penalty. This research is approved by Committee for Ethical Research and the Protection of Human Participants, University of Haifa: (350\/19) Thank you in advance for your cooperation.\u201d","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent to participate"}},{"value":"This work has not been published before; it is not under consideration for publication anywhere else; its publication has been approved by all co-authors.","order":5,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent for publication"}}],"article-number":"2"}}