{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,25]],"date-time":"2025-03-25T14:16:09Z","timestamp":1742912169296,"version":"3.40.3"},"publisher-location":"Cham","reference-count":44,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031606052"},{"type":"electronic","value":"9783031606069"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-3-031-60606-9_18","type":"book-chapter","created":{"date-parts":[[2024,6,1]],"date-time":"2024-06-01T01:06:47Z","timestamp":1717204007000},"page":"316-333","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Towards a\u00a0Framework for\u00a0Interdisciplinary Studies in\u00a0Explainable Artificial Intelligence"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0009-1985-1642","authenticated-orcid":false,"given":"Paula","family":"Ziethmann","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0004-3827-9809","authenticated-orcid":false,"given":"Fabian","family":"Stieler","sequence":"additional","affiliation":[]},{"given":"Raphael","family":"Pfrommer","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6957-9549","authenticated-orcid":false,"given":"Kerstin","family":"Schl\u00f6gl-Flierl","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7931-1105","authenticated-orcid":false,"given":"Bernhard","family":"Bauer","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,6,1]]},"reference":[{"key":"18_CR1","unstructured":"AI Act - Shaping Europe\u2019s digital future. https:\/\/digital-strategy.ec.europa.eu\/en\/policies\/regulatory-framework-ai"},{"key":"18_CR2","unstructured":"Proposal for a Regulation of the European parliament and of the council laying down harmonised rules on artificial intelligence (artificial intelligence act) and amending certain union legislative acts. https:\/\/artificialintelligenceact.eu\/wp-content\/uploads\/2024\/01\/AI-Act-FullText.pdf"},{"key":"18_CR3","doi-asserted-by":"publisher","first-page":"52138","DOI":"10.1109\/ACCESS.2018.2870052","volume":"6","author":"A Adadi","year":"2018","unstructured":"Adadi, A., Berrada, M.: Peeking inside the black-box: a survey on Explainable Artificial Intelligence (XAI). IEEE Access 6, 52138\u201352160 (2018). https:\/\/doi.org\/10.1109\/ACCESS.2018.2870052","journal-title":"IEEE Access"},{"key":"18_CR4","unstructured":"Arrieta, A.B., et al.: Explainable Artificial Intelligence (XAI): Concepts, Taxonomies, Opportunities and Challenges toward Responsible AI (2019). arXiv:1910.10045"},{"key":"18_CR5","unstructured":"Arya, V., et al.: One explanation does not fit all: a toolkit and taxonomy of AI Explainability Techniques (2019). arXiv:1909.03012"},{"issue":"7","key":"18_CR6","doi-asserted-by":"publisher","DOI":"10.1371\/journal.pone.0130140","volume":"10","author":"S Bach","year":"2015","unstructured":"Bach, S., Binder, A., Montavon, G., Klauschen, F., M\u00fcller, K.R., Samek, W.: On pixel-wise explanations for non-linear classifier decisions by layer-wise relevance propagation. PLoS ONE 10(7), e0130140 (2015). https:\/\/doi.org\/10.1371\/journal.pone.0130140","journal-title":"PLoS ONE"},{"key":"18_CR7","doi-asserted-by":"publisher","unstructured":"Bienefeld, N., et al.: Solving the explainable AI conundrum by bridging clinicians\u2019 needs and developers\u2019 goals. NPJ Digital Medicine 6(1), 94 (2023). https:\/\/doi.org\/10.1038\/s41746-023-00837-4, https:\/\/www.nature.com\/articles\/s41746-023-00837-4","DOI":"10.1038\/s41746-023-00837-4"},{"key":"18_CR8","unstructured":"Braun, M., Breuer, S.: \u201cEmbedded Ethics and Social Sciences\u201d in HRI Research: Scenarios and Subjectivities (2022)"},{"issue":"8","key":"18_CR9","doi-asserted-by":"publisher","first-page":"832","DOI":"10.3390\/electronics8080832","volume":"8","author":"DV Carvalho","year":"2019","unstructured":"Carvalho, D.V., Pereira, E.M., Cardoso, J.S.: Machine learning interpretability: a survey on methods and metrics. Electronics 8(8), 832 (2019)","journal-title":"Electronics"},{"key":"18_CR10","unstructured":"Chromik, M., Schuessler, M.: A taxonomy for human subject evaluation of black-box explanations in XAI. In: ExSS-ATEC@IUI. Cagliari, Italy (2020)"},{"key":"18_CR11","doi-asserted-by":"publisher","unstructured":"Dhanorkar, S., Wolf, C.T., Qian, K., Xu, A., Popa, L., Li, Y.: Who needs to know what, when?: broadening the Explainable AI (XAI) Design Space by Looking at Explanations Across the AI Lifecycle. In: Designing Interactive Systems Conference 2021, pp. 1591\u20131602. ACM, Virtual Event USA (2021). https:\/\/doi.org\/10.1145\/3461778.3462131","DOI":"10.1145\/3461778.3462131"},{"key":"18_CR12","unstructured":"Doshi-Velez, F., Kim, B.: Towards a rigorous science of interpretable machine learning (2017). arXiv:1702.08608"},{"key":"18_CR13","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"56","DOI":"10.1007\/978-3-030-49760-6_4","volume-title":"Design, User Experience, and Usability. Design for Contemporary Interactive Environments","author":"JJ Ferreira","year":"2020","unstructured":"Ferreira, J.J., Monteiro, M.S.: What are people doing about XAI user experience? A survey on ai explainability research and practice. In: Marcus, A., Rosenzweig, E. (eds.) HCII 2020. LNCS, vol. 12201, pp. 56\u201373. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-49760-6_4"},{"issue":"2","key":"18_CR14","doi-asserted-by":"publisher","first-page":"215","DOI":"10.1007\/s13347-021-00460-9","volume":"34","author":"L Floridi","year":"2021","unstructured":"Floridi, L.: The European legislation on AI: a brief analysis of its philosophical approach. Philos. Technol. 34(2), 215\u2013222 (2021). https:\/\/doi.org\/10.1007\/s13347-021-00460-9","journal-title":"Philos. Technol."},{"issue":"2","key":"18_CR15","doi-asserted-by":"publisher","first-page":"75","DOI":"10.1089\/big.2016.0007","volume":"4","author":"M Gleicher","year":"2016","unstructured":"Gleicher, M.: A framework for considering comprehensibility in modeling. Big Data 4(2), 75\u201388 (2016). https:\/\/doi.org\/10.1089\/big.2016.0007","journal-title":"Big Data"},{"key":"18_CR16","doi-asserted-by":"publisher","unstructured":"Hohman, F., Head, A., Caruana, R., DeLine, R., Drucker, S.M.: Gamut: a Design Probe to Understand How Data Scientists Understand Machine Learning Models. In: Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems. ACM (2019). https:\/\/doi.org\/10.1145\/3290605.3300809","DOI":"10.1145\/3290605.3300809"},{"key":"18_CR17","doi-asserted-by":"publisher","unstructured":"Hong, S.R., Hullman, J., Bertini, E.: Human factors in model interpretability: industry practices, challenges, and needs. In: Proceedings of the ACM on Human-Computer Interaction 4(CSCW1) (2020). https:\/\/doi.org\/10.1145\/3392878, arXiv:2004.11440","DOI":"10.1145\/3392878"},{"key":"18_CR18","doi-asserted-by":"publisher","unstructured":"Johs, A.J., Agosto, D.E., Weber, R.O.: Qualitative investigation in explainable artificial intelligence: a bit more insight from social science (2020). https:\/\/doi.org\/10.22541\/au.163284810.09140868\/v1.","DOI":"10.22541\/au.163284810.09140868\/v1."},{"key":"18_CR19","doi-asserted-by":"publisher","unstructured":"Johs, A.J., Agosto, D.E., Weber, R.O.: Explainable artificial intelligence and social science: further insights for qualitative investigation. Appl. AI Lett. 3(1) (2022). https:\/\/doi.org\/10.1002\/ail2.64","DOI":"10.1002\/ail2.64"},{"key":"18_CR20","doi-asserted-by":"publisher","unstructured":"Kaur, H., Nori, H., Jenkins, S., Caruana, R., Wallach, H., Wortman\u00a0Vaughan, J.: Interpreting interpretability: understanding data scientists\u2019 use of interpretability tools for machine learning. In: Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems. ACM, Honolulu HI USA (2020). https:\/\/doi.org\/10.1145\/3313831.3376219","DOI":"10.1145\/3313831.3376219"},{"key":"18_CR21","doi-asserted-by":"publisher","unstructured":"Kitzinger, J.: Qualitative research: introducing focus groups. BMJ 311(7000), 299\u2013302 (1995). https:\/\/doi.org\/10.1136\/bmj.311.7000.299, https:\/\/www.bmj.com\/content\/311\/7000\/299","DOI":"10.1136\/bmj.311.7000.299"},{"key":"18_CR22","doi-asserted-by":"publisher","unstructured":"Langer, M., et al.: What do we want from explainable artificial intelligence (XAI)? \u2013 a stakeholder perspective on xai and a conceptual model guiding interdisciplinary XAI research. Artif. Intell. 296 (2021). https:\/\/doi.org\/10.1016\/j.artint.2021.103473, arXiv:2102.07817","DOI":"10.1016\/j.artint.2021.103473"},{"key":"18_CR23","volume-title":"A Unified Approach to Interpreting Model Predictions","author":"SM Lundberg","year":"2017","unstructured":"Lundberg, S.M., Lee, S.I.: A Unified Approach to Interpreting Model Predictions. Long Beach, CA, USA (2017)"},{"key":"18_CR24","series-title":"Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence)","doi-asserted-by":"publisher","first-page":"129","DOI":"10.1007\/978-3-030-51924-7_8","volume-title":"Explainable, Transparent Autonomous Agents and Multi-Agent Systems","author":"A Malhi","year":"2020","unstructured":"Malhi, A., Knapic, S., Fr\u00e4mling, K.: explainable agents for less bias in human-agent decision making. In: Calvaresi, D., Najjar, A., Winikoff, M., Fr\u00e4mling, K. (eds.) EXTRAAMAS 2020. LNCS (LNAI), vol. 12175, pp. 129\u2013146. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-51924-7_8"},{"key":"18_CR25","doi-asserted-by":"crossref","unstructured":"Mayring, P.: Qualitative inhaltsanalyse. In: Mey, G., Mruck, K. (eds.) Handbuch Qualitative Forschung in der Psychologie. VS Verlag f\u00fcr Sozialwissenschaften (2010)","DOI":"10.1007\/978-3-531-92052-8_42"},{"key":"18_CR26","doi-asserted-by":"crossref","unstructured":"Mayring, P.: Qualitative content analysis: theoretical foundation, basic procedures and software solution. Klagenfurt (2014)","DOI":"10.1007\/978-94-017-9181-6_13"},{"key":"18_CR27","doi-asserted-by":"publisher","unstructured":"Mayring, P., Fenzel, T.: Qualitative Inhaltsanalyse. In: Baur, N., Blasius, J. (eds) Handbuch Methoden der empirischen Sozialforschung, pp. 691\u2013706. Springer Fachmedien, Wiesbaden (2022). https:\/\/doi.org\/10.1007\/978-3-658-37985-8_43","DOI":"10.1007\/978-3-658-37985-8_43"},{"key":"18_CR28","doi-asserted-by":"publisher","unstructured":"Mercado, J.E., Rupp, M.A., Chen, J.Y.C., Barnes, M.J., Barber, D., Procci, K.: Intelligent agent transparency in human\u2013agent teaming for Multi-UxV management. Hum. Factors: J. Hum. Factors Ergonomics Soc. 58(3), 401\u2013415 (2016). https:\/\/doi.org\/10.1177\/0018720815621206","DOI":"10.1177\/0018720815621206"},{"key":"18_CR29","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.artint.2018.07.007","volume":"267","author":"T Miller","year":"2019","unstructured":"Miller, T.: Explanation in artificial intelligence: insights from the social sciences. Artif. Intell. 267, 1\u201338 (2019). https:\/\/doi.org\/10.1016\/j.artint.2018.07.007","journal-title":"Artif. Intell."},{"key":"18_CR30","unstructured":"Molnar, C.: Interpretable Machine Learning - A Guide for Making Black Box Models Explainable. 2nd edn. (2022). https:\/\/christophm.github.io\/interpretable-ml-book"},{"key":"18_CR31","unstructured":"Mueller, S.T., Veinott, E.S., Hoffman, R.R., Klein, G., Alam, L., Mamun, T. et al.: Principles of explanation in human-AI systems (2021)"},{"key":"18_CR32","doi-asserted-by":"publisher","unstructured":"Nauta, M., et al.: From anecdotal evidence to quantitative evaluation methods: a systematic review on evaluating explainable AI. ACM Comput. Surv. 55(13s) (2023). https:\/\/doi.org\/10.1145\/3583558","DOI":"10.1145\/3583558"},{"key":"18_CR33","doi-asserted-by":"publisher","DOI":"10.2139\/ssrn.4447964","author":"C Novelli","year":"2023","unstructured":"Novelli, C., Casolari, F., Rotolo, A., Taddeo, M., Floridi, L.: Taking AI risks seriously: a proposal for the AI act. SSRN Electron. J. (2023). https:\/\/doi.org\/10.2139\/ssrn.4447964","journal-title":"SSRN Electron. J."},{"key":"18_CR34","doi-asserted-by":"publisher","unstructured":"Ribeiro, M.T., Singh, S., Guestrin, C.: \u201cWhy Should I Trust You?\u201d: explaining the predictions of any classifier. In: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 1135\u20131144. ACM, San Francisco California USA (2016). https:\/\/doi.org\/10.1145\/2939672.2939778","DOI":"10.1145\/2939672.2939778"},{"issue":"4","key":"18_CR35","doi-asserted-by":"publisher","first-page":"905","DOI":"10.1046\/j.1365-2648.1999.00966.x","volume":"29","author":"N Robinson","year":"1999","unstructured":"Robinson, N.: The use of focus group methodology - with selected examples from sexual health research. J. Adv. Nurs. 29(4), 905\u2013913 (1999). https:\/\/doi.org\/10.1046\/j.1365-2648.1999.00966.x","journal-title":"J. Adv. Nurs."},{"key":"18_CR36","doi-asserted-by":"crossref","unstructured":"Rudin, C., Chen, C., Chen, Z., Huang, H., Semenova, L., Zhong, C.: Interpretable machine learning: fundamental principles and 10 grand challenges (Jul 2021). http:\/\/arxiv.org\/abs\/2103.11251, arXiv:2103.11251","DOI":"10.1214\/21-SS133"},{"key":"18_CR37","doi-asserted-by":"publisher","unstructured":"Saeed, W., Omlin, C.: Explainable AI (XAI): a systematic meta-survey of current challenges and future opportunities. Knowl.-Based Syst. 263 (2023). https:\/\/doi.org\/10.1016\/j.knosys.2023.110273","DOI":"10.1016\/j.knosys.2023.110273"},{"issue":"3\u20134","key":"18_CR38","doi-asserted-by":"publisher","first-page":"303","DOI":"10.1007\/s13218-022-00786-2","volume":"36","author":"U Schmid","year":"2022","unstructured":"Schmid, U., Wrede, B.: What is missing in XAI so far?: an interdisciplinary perspective. KI - K\u00fcnstliche Intelligenz 36(3\u20134), 303\u2013315 (2022). https:\/\/doi.org\/10.1007\/s13218-022-00786-2","journal-title":"KI - K\u00fcnstliche Intelligenz"},{"key":"18_CR39","doi-asserted-by":"publisher","unstructured":"Schwalbe, G., Finzel, B.: A comprehensive taxonomy for explainable artificial intelligence: a systematic survey of surveys on methods and concepts. Data Min. Knowl. Discov. (2023). https:\/\/doi.org\/10.1007\/s10618-022-00867-8","DOI":"10.1007\/s10618-022-00867-8"},{"key":"18_CR40","doi-asserted-by":"publisher","unstructured":"Selvaraju, R.R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., Batra, D.: Grad-CAM: visual explanations from deep networks via gradient-based localization. In: 2017 IEEE International Conference on Computer Vision (ICCV), pp. 618\u2013626. IEEE, Venice (2017). https:\/\/doi.org\/10.1109\/ICCV.2017.74","DOI":"10.1109\/ICCV.2017.74"},{"key":"18_CR41","doi-asserted-by":"publisher","unstructured":"Siontis, K.C., Noseworthy, P.A., Attia, Z.I., Friedman, P.A.: Artificial intelligence-enhanced electrocardiography in cardiovascular disease management. Nat. Rev. Cardiol. 18(7), 465\u2013478 (2021). https:\/\/doi.org\/10.1038\/s41569-020-00503-2, https:\/\/www.nature.com\/articles\/s41569-020-00503-2","DOI":"10.1038\/s41569-020-00503-2"},{"key":"18_CR42","doi-asserted-by":"publisher","unstructured":"Stieler, F., et al.: LIFEDATA - A framework for traceable active learning projects. In: 2023 IEEE 31st International Requirements Engineering Conference Workshops (REW), pp. 465\u2013474. IEEE, Hannover, Germany (2023). https:\/\/doi.org\/10.1109\/REW57809.2023.00088","DOI":"10.1109\/REW57809.2023.00088"},{"key":"18_CR43","doi-asserted-by":"publisher","unstructured":"Vilone, G., Longo, L.: Notions of explainability and evaluation approaches for explainable artificial intelligence. Inf. Fusion 76, 89\u2013106 (2021). https:\/\/doi.org\/10.1016\/j.inffus.2021.05.009, https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S1566253521001093","DOI":"10.1016\/j.inffus.2021.05.009"},{"issue":"4","key":"18_CR44","doi-asserted-by":"publisher","first-page":"2159","DOI":"10.1007\/s12525-022-00608-1","volume":"32","author":"J Zacharias","year":"2022","unstructured":"Zacharias, J., Von Zahn, M., Chen, J., Hinz, O.: Designing a feature selection method based on explainable artificial intelligence. Electron. Mark. 32(4), 2159\u20132184 (2022). https:\/\/doi.org\/10.1007\/s12525-022-00608-1","journal-title":"Electron. Mark."}],"container-title":["Lecture Notes in Computer Science","Artificial Intelligence in HCI"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-60606-9_18","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,6,1]],"date-time":"2024-06-01T01:08:15Z","timestamp":1717204095000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-60606-9_18"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9783031606052","9783031606069"],"references-count":44,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-60606-9_18","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"1 June 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"HCII","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Human-Computer Interaction","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Washington DC","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"USA","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 June 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 July 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"26","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"hcii2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/2024.hci.international\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}