{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,4]],"date-time":"2026-03-04T03:38:12Z","timestamp":1772595492598,"version":"3.50.1"},"publisher-location":"Cham","reference-count":28,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031503955","type":"print"},{"value":"9783031503962","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2024,1,21]],"date-time":"2024-01-21T00:00:00Z","timestamp":1705795200000},"content-version":"vor","delay-in-days":20,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:p>In explainable artificial intelligence (XAI) research, the predominant focus has been on interpreting models for experts and practitioners. Model agnostic and local explanation approaches are deemed interpretable and sufficient in many applications. However, in domains like healthcare, where end users are patients without AI or domain expertise, there is an urgent need for model explanations that are more comprehensible and instil trust in the model\u2019s operations. We hypothesise that generating model explanations that are narrative, patient-specific and <jats:italic>global<\/jats:italic> (holistic of the model) would enable better understandability and enable decision-making. We test this using a decision tree model to generate both local and global explanations for patients identified as having a high risk of coronary heart disease. These explanations are presented to non-expert users. We find a strong individual preference for a specific type of explanation. The majority of participants prefer global explanations, while a smaller group prefers local explanations. A task based evaluation of mental models of these participants provide valuable feedback to enhance narrative global explanations. This, in turn, guides the design of health informatics systems that are both trustworthy and actionable.<\/jats:p>","DOI":"10.1007\/978-3-031-50396-2_3","type":"book-chapter","created":{"date-parts":[[2024,1,20]],"date-time":"2024-01-20T14:02:08Z","timestamp":1705759328000},"page":"43-65","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":10,"title":["Evaluation of\u00a0Human-Understandability of\u00a0Global Model Explanations Using Decision Tree"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-4460-1671","authenticated-orcid":false,"given":"Adarsa","family":"Sivaprasad","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7548-9504","authenticated-orcid":false,"given":"Ehud","family":"Reiter","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1663-1627","authenticated-orcid":false,"given":"Nava","family":"Tintarev","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4854-9014","authenticated-orcid":false,"given":"Nir","family":"Oren","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,1,21]]},"reference":[{"key":"3_CR1","doi-asserted-by":"publisher","first-page":"52138","DOI":"10.1109\/ACCESS.2018.2870052","volume":"6","author":"A Adadi","year":"2018","unstructured":"Adadi, A., Berrada, M.: Peeking inside the black-box: a survey on explainable artificial intelligence (XAI). IEEE Access 6, 52138\u201352160 (2018). https:\/\/doi.org\/10.1109\/ACCESS.2018.2870052","journal-title":"IEEE Access"},{"issue":"7","key":"3_CR2","doi-asserted-by":"publisher","first-page":"1039","DOI":"10.1007\/s10994-017-5633-9","volume":"106","author":"D Bertsimas","year":"2017","unstructured":"Bertsimas, D., Dunn, J.: Optimal classification trees. Mach. Learn. 106(7), 1039\u20131082 (2017). https:\/\/doi.org\/10.1007\/s10994-017-5633-9","journal-title":"Mach. Learn."},{"key":"3_CR3","unstructured":"Biran, O., Cotton, C.V.: Explanation and justification in machine learning: a survey. In: IJCAI-17 Workshop on Explainable AI (XAI), vol. 8 (2017)"},{"key":"3_CR4","doi-asserted-by":"publisher","unstructured":"Blanquero, R., Carrizosa, E., Molero-R\u00edo, C., Morales, D.R.: Optimal randomized classification trees. Comput. Oper. Res. 132, 105281 (2021). https:\/\/doi.org\/10.1016\/j.cor.2021.105281","DOI":"10.1016\/j.cor.2021.105281"},{"key":"3_CR5","unstructured":"Breiman, L., Friedman, J.H., Olshen, R.A., Stone, C.J.: Classification and regression trees (1984)"},{"key":"3_CR6","doi-asserted-by":"publisher","unstructured":"Glik, D.C.: Risk communication for public health emergencies. Annu. Rev. Publ. Health 28(1), 33\u201354 (2007). https:\/\/doi.org\/10.1146\/annurev.publhealth.28.021406.144123, pMID: 17222081","DOI":"10.1146\/annurev.publhealth.28.021406.144123"},{"key":"3_CR7","doi-asserted-by":"publisher","unstructured":"Hippisley-Cox, J., Coupland, C., Brindle, P.: Development and validation of qrisk3 risk prediction algorithms to estimate future risk of cardiovascular disease: prospective cohort study. BMJ 357 (2017). https:\/\/doi.org\/10.1136\/bmj.j2099","DOI":"10.1136\/bmj.j2099"},{"key":"3_CR8","unstructured":"Hu, X., Rudin, C., Seltzer, M.: Optimal sparse decision trees. In: Wallach, H., Larochelle, H., Beygelzimer, A., d\u2019Alch\u00e9-Buc, F., Fox, E., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 32. Curran Associates, Inc. (2019)"},{"issue":"1","key":"3_CR9","doi-asserted-by":"publisher","first-page":"15","DOI":"10.1016\/0020-0190(76)90095-8","volume":"5","author":"L Hyafil","year":"1976","unstructured":"Hyafil, L., Rivest, R.L.: Constructing optimal binary decision trees is np-complete. Inf. Process. Lett. 5(1), 15\u201317 (1976)","journal-title":"Inf. Process. Lett."},{"key":"3_CR10","series-title":"Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence)","doi-asserted-by":"publisher","first-page":"224","DOI":"10.1007\/978-3-540-27819-1_16","volume-title":"Learning Theory","author":"AR Klivans","year":"2004","unstructured":"Klivans, A.R., Servedio, R.A.: Toward attribute efficient learning of decision lists and parities. In: Shawe-Taylor, J., Singer, Y. (eds.) COLT 2004. LNCS (LNAI), vol. 3120, pp. 224\u2013238. Springer, Heidelberg (2004). https:\/\/doi.org\/10.1007\/978-3-540-27819-1_16"},{"issue":"7","key":"3_CR11","doi-asserted-by":"publisher","first-page":"747","DOI":"10.1111\/j.1467-842X.1998.tb01487.x","volume":"22","author":"MW Knuiman","year":"1998","unstructured":"Knuiman, M.W., Vu, H.T., Bartholomew, H.C.: Multivariate risk estimation for coronary heart disease: the Busselton health study. Aust. N. Z. J. Publ. Health 22(7), 747\u2013753 (1998)","journal-title":"Aust. N. Z. J. Publ. Health"},{"key":"3_CR12","doi-asserted-by":"publisher","unstructured":"Lakkaraju, H., Bach, S.H., Leskovec, J.: Interpretable decision sets: a joint framework for description and prediction. In: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. KDD \u201916, pp. 1675\u20131684. Association for Computing Machinery, New York, NY, USA (2016). https:\/\/doi.org\/10.1145\/2939672.2939874","DOI":"10.1145\/2939672.2939874"},{"key":"3_CR13","doi-asserted-by":"publisher","unstructured":"Letham, B., Rudin, C., McCormick, T., Madigan, D.: Interpretable classifiers using rules and Bayesian analysis: building a better stroke prediction model. Ann. Appl. Stat. 9, 1350\u20131371 (2015). https:\/\/doi.org\/10.1214\/15-AOAS848","DOI":"10.1214\/15-AOAS848"},{"key":"3_CR14","unstructured":"Lin, J., Zhong, C., Hu, D., Rudin, C., Seltzer, M.: Generalized and scalable optimal sparse decision trees. In: III, H.D., Singh, A. (eds.) Proceedings of the 37th International Conference on Machine Learning, vol. 119, pp. 6150\u20136160. PMLR, 13\u201318 July 2020"},{"key":"3_CR15","unstructured":"Lundberg, S.M., Lee, S.I.: A unified approach to interpreting model predictions. In: Proceedings of the 31st International Conference on Neural Information Processing Systems. NIPS\u201917, pp. 4768\u20134777. Curran Associates Inc., Red Hook, NY, USA (2017)"},{"key":"3_CR16","doi-asserted-by":"publisher","DOI":"10.1016\/j.jbi.2020.103655","volume":"113","author":"AF Markus","year":"2021","unstructured":"Markus, A.F., Kors, J.A., Rijnbeek, P.R.: The role of explainability in creating trustworthy artificial intelligence for health care: a comprehensive survey of the terminology, design choices, and evaluation strategies. J. Biomed. Inform. 113, 103655 (2021). https:\/\/doi.org\/10.1016\/j.jbi.2020.103655","journal-title":"J. Biomed. Inform."},{"key":"3_CR17","doi-asserted-by":"publisher","DOI":"10.1016\/j.csl.2023.101483","volume":"81","author":"S Maruf","year":"2023","unstructured":"Maruf, S., Zukerman, I., Reiter, E., Haffari, G.: Influence of context on users\u2019 views about explanations for decision-tree predictions. Comput. Speech Lang. 81, 101483 (2023). https:\/\/doi.org\/10.1016\/j.csl.2023.101483","journal-title":"Comput. Speech Lang."},{"key":"3_CR18","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.artint.2018.07.007","volume":"267","author":"T Miller","year":"2019","unstructured":"Miller, T.: Explanation in artificial intelligence: insights from the social sciences. Artif. Intell. 267, 1\u201338 (2019). https:\/\/doi.org\/10.1016\/j.artint.2018.07.007","journal-title":"Artif. Intell."},{"issue":"2","key":"3_CR19","doi-asserted-by":"publisher","first-page":"197","DOI":"10.1016\/j.actpsy.2007.12.008","volume":"128","author":"S Moreno-R\u00edos","year":"2008","unstructured":"Moreno-R\u00edos, S., Garc\u00eda-Madruga, J.A., Byrne, R.M.: Inferences from semifactual \u2018even if\u2019 conditionals. Acta Physiol. (OXF) 128(2), 197\u2013209 (2008). https:\/\/doi.org\/10.1016\/j.actpsy.2007.12.008","journal-title":"Acta Physiol. (OXF)"},{"key":"3_CR20","unstructured":"Narayanan, M., Chen, E., He, J., Kim, B., Gershman, S., Doshi-Velez, F.: How do humans understand explanations from machine learning systems? An evaluation of the human-interpretability of explanation. CoRR abs\/1802.00682 (2018)"},{"key":"3_CR21","volume-title":"The Book of Why: The New Science of Cause and Effect","author":"J Pearl","year":"2018","unstructured":"Pearl, J., Mackenzie, D.: The Book of Why: The New Science of Cause and Effect, 1st edn. Basic Books Inc., New York (2018)","edition":"1"},{"key":"3_CR22","unstructured":"Quinlan, J.R.: C4.5: Programs for Machine Learning. Morgan Kaufmann Publishers Inc., San Francisco, CA, USA (1993)"},{"key":"3_CR23","doi-asserted-by":"publisher","unstructured":"Reiter, E.: Natural language generation challenges for explainable AI. In: Proceedings of the 1st Workshop on Interactive Natural Language Technology for Explainable Artif. Intell. (NL4XAI 2019), pp. 3\u20137. Association for Computational Linguistics (2019). https:\/\/doi.org\/10.18653\/v1\/W19-8402","DOI":"10.18653\/v1\/W19-8402"},{"key":"3_CR24","doi-asserted-by":"publisher","unstructured":"Ribeiro, M., Singh, S., Guestrin, C.: \u201cwhy should I trust you?\u201d: Explaining the predictions of any classifier. In: Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Demonstrations, pp. 97\u2013101. Association for Computational Linguistics, San Diego, California (2016). https:\/\/doi.org\/10.18653\/v1\/N16-3020","DOI":"10.18653\/v1\/N16-3020"},{"key":"3_CR25","doi-asserted-by":"crossref","unstructured":"Sendak, M.P., Gao, M., Brajer, N., Balu, S.: Presenting machine learning model information to clinical end users with model facts labels. NPJ Digit. Med. 3 (2020)","DOI":"10.1038\/s41746-020-0253-3"},{"key":"3_CR26","doi-asserted-by":"publisher","unstructured":"Spiegelhalter, D.: Risk and uncertainty communication. Annu. Rev. Stat. Appl. 4(1), 31\u201360 (2017). https:\/\/doi.org\/10.1146\/annurev-statistics-010814-020148","DOI":"10.1146\/annurev-statistics-010814-020148"},{"key":"3_CR27","doi-asserted-by":"crossref","unstructured":"Verwer, S., Zhang, Y.: Learning optimal classification trees using a binary linear program formulation. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33, no. 01, pp. 1625\u20131632 (2019)","DOI":"10.1609\/aaai.v33i01.33011624"},{"key":"3_CR28","doi-asserted-by":"crossref","unstructured":"Wachter, S., Mittelstadt, B.D., Russell, C.: Counterfactual explanations without opening the black box: automated decisions and the GDPR. CoRR abs\/1711.00399 (2017)","DOI":"10.2139\/ssrn.3063289"}],"container-title":["Communications in Computer and Information Science","Artificial Intelligence. ECAI 2023 International Workshops"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-50396-2_3","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,1,20]],"date-time":"2024-01-20T14:09:29Z","timestamp":1705759769000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-50396-2_3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9783031503955","9783031503962"],"references-count":28,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-50396-2_3","relation":{},"ISSN":["1865-0929","1865-0937"],"issn-type":[{"value":"1865-0929","type":"print"},{"value":"1865-0937","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"21 January 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECAI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Artificial Intelligence","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Krak\u00f3w","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Poland","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"30 September 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"26","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ecai2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/ecai2023.eu\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"EasyChair and SMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"134","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"69","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"51% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2.86","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2.15","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Presented figures are relevant for the ECAI 2023 workshops","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}