{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T00:38:03Z","timestamp":1760056683002,"version":"build-2065373602"},"publisher-location":"Cham","reference-count":28,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783032013989","type":"print"},{"value":"9783032013996","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T00:00:00Z","timestamp":1760054400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T00:00:00Z","timestamp":1760054400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-3-032-01399-6_13","type":"book-chapter","created":{"date-parts":[[2025,10,9]],"date-time":"2025-10-09T12:51:03Z","timestamp":1760014263000},"page":"229-241","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Designing for Effective Human-XAI Interaction: User Experience Research Plays and Cards"],"prefix":"10.1007","author":[{"given":"Mohammad","family":"Naiseh","sequence":"first","affiliation":[]},{"given":"Huseyin","family":"Dogan","sequence":"additional","affiliation":[]},{"given":"Stephen","family":"Giff","sequence":"additional","affiliation":[]},{"given":"Avleen","family":"Malhi","sequence":"additional","affiliation":[]},{"given":"Nan","family":"Jiang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,10,10]]},"reference":[{"key":"13_CR1","doi-asserted-by":"crossref","unstructured":"Dogan, H., Giff, S., Barsoum, R.: User experience research: point of view playbook. In:\u00a0Extended Abstracts of the CHI Conference on Human Factors in Computing Systems, pp. 1\u20137 (2024)","DOI":"10.1145\/3613905.3637136"},{"key":"13_CR2","doi-asserted-by":"publisher","unstructured":"Naiseh, M.: Social eXplainable AI (Social XAI): towards expanding the social benefits of XAI. In: Montag, C., Ali, R. (eds.) The Impact of Artificial Intelligence on Societies. Studies in Neuroscience, Psychology and Behavioral Economics. Springer, Cham (2024) https:\/\/doi.org\/10.1007\/978-3-031-70355-3_13","DOI":"10.1007\/978-3-031-70355-3_13"},{"key":"13_CR3","doi-asserted-by":"publisher","DOI":"10.1016\/j.ijhcs.2022.102941","volume":"169","author":"M Naiseh","year":"2023","unstructured":"Naiseh, M., Al-Thani, D., Jiang, N., Ali, R.: How the different explanation classes impact trust calibration: the case of clinical decision support systems. Int. J. Hum Comput Stud. 169, 102941 (2023)","journal-title":"Int. J. Hum Comput Stud."},{"key":"13_CR4","doi-asserted-by":"crossref","unstructured":"Naiseh, M., Bentley, C., Ramchurn, S.D.: Trustworthy autonomous systems (TAS): engaging TAS experts in curriculum design. In: 2022 IEEE global engineering education conference (EDUCON), pp. 901\u2013905. IEEE (2022)","DOI":"10.1109\/EDUCON52537.2022.9766663"},{"issue":"10","key":"13_CR5","doi-asserted-by":"publisher","first-page":"28","DOI":"10.1109\/MC.2021.3076131","volume":"54","author":"M Naiseh","year":"2021","unstructured":"Naiseh, M., Cemiloglu, D., Al Thani, D., Jiang, N., Ali, R.: Explainable recommendations and calibrated trust: two systematic user errors. Computer 54(10), 28\u201337 (2021)","journal-title":"Computer"},{"key":"13_CR6","doi-asserted-by":"publisher","unstructured":"Xu, F., Uszkoreit, H., Du, Y., Fan, W., Zhao, D., Zhu, J.: Explainable AI: a brief survey on history, research areas, approaches and challenges. In: Tang, J., Kan, MY., Zhao, D., Li, S., Zan, H. (eds.) Natural Language Processing and Chinese Computing. NLPCC 2019. LNCS, vol. 11839. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-32236-6_51","DOI":"10.1007\/978-3-030-32236-6_51"},{"key":"13_CR7","doi-asserted-by":"crossref","unstructured":"Naiseh, M., Al-Mansoori, R.S., Al-Thani, D., Jiang, N., Ali, R.: Nudging through friction: an approach for calibrating trust in explainable AI. In:\u00a02021 8th International Conference on Behavioral and Social Computing (BESC), pp. 1\u20135. IEEE (2021)","DOI":"10.1109\/BESC53957.2021.9635271"},{"key":"13_CR8","doi-asserted-by":"crossref","unstructured":"Cabitza, F., Fregosi, C., Campagner, A., Natali, C.: Explanations considered harmful: the impact of misleading explanations on accuracy in hybrid human-AI decision making. In:\u00a0World conference on explainable artificial intelligence, pp. 255\u2013269. Springer Nature Switzerland, Cham (2024)","DOI":"10.1007\/978-3-031-63803-9_14"},{"issue":"5","key":"13_CR9","doi-asserted-by":"publisher","first-page":"1857","DOI":"10.1007\/s11280-021-00916-0","volume":"24","author":"M Naiseh","year":"2021","unstructured":"Naiseh, M., Al-Thani, D., Jiang, N., Ali, R.: Explainable recommendation: when design meets trust calibration. World Wide Web 24(5), 1857\u20131884 (2021)","journal-title":"World Wide Web"},{"key":"13_CR10","doi-asserted-by":"crossref","unstructured":"Simkute, A., Surana, A., Luger, E., Evans, M., Jones, R.: XAI for learning: Narrowing down the digital divide between \u201cnew\u201d and \u201cold\u201d experts. In:\u00a0Adjunct Proceedings of the 2022 Nordic Human-Computer Interaction Conference, pp. 1\u20136 (2022)","DOI":"10.1145\/3547522.3547678"},{"key":"13_CR11","doi-asserted-by":"publisher","unstructured":"Naiseh, M., Jiang, N., Ma, J., Ali, R.: Personalising explainable recommendations: literature and conceptualisation. In: Rocha, \u00c1., Adeli, H., Reis, L., Costanzo, S., Orovic, I., Moreira, F. (eds.) Trends and Innovations in Information Systems and Technologies. WorldCIST 2020. AISC, vol. 1160. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-45691-7_49","DOI":"10.1007\/978-3-030-45691-7_49"},{"key":"13_CR12","doi-asserted-by":"crossref","unstructured":"Sneirson, M., Chai, J., Howley, I.: A learning approach for increasing AI literacy via XAI in informal settings. In: International Conference on Artificial Intelligence in Education, pp. 336\u2013343. Springer Nature Switzerland, Cham (2024)","DOI":"10.1007\/978-3-031-64299-9_29"},{"key":"13_CR13","doi-asserted-by":"publisher","unstructured":"Naiseh, M., Soorati, M.D., Ramchurn, S.: Outlining the\u00a0design space of\u00a0eXplainable Swarm (xSwarm): experts\u2019 perspective. In: Bourgeois, J., et al. (eds.) Distributed Autonomous Robotic Systems. DARS 2022. SPAR, vol. 28. Springer, Cham (2024). https:\/\/doi.org\/10.1007\/978-3-031-51497-5_3","DOI":"10.1007\/978-3-031-51497-5_3"},{"key":"13_CR14","unstructured":"UXR POV PlayBook. UXR POV Playbook (n.d.). https:\/\/www.uxrpovplaybook.com\/"},{"key":"13_CR15","doi-asserted-by":"crossref","unstructured":"Severes, B., Carreira, C., Vieira, A.B., Gomes, E., Apar\u00edcio, J.T., Pereira, I.: The human side of XAI: bridging the gap between AI and non-expert audiences. In: Proceedings of the 41st ACM International Conference on Design of Communication, pp. 126\u2013132 (2023)","DOI":"10.1145\/3615335.3623062"},{"key":"13_CR16","unstructured":"Williams, O.: Towards human-centred explainable AI: a systematic literature review.\u00a0Master\u2019s Thesis (2021)"},{"issue":"4","key":"13_CR17","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3419764","volume":"10","author":"B Shneiderman","year":"2020","unstructured":"Shneiderman, B.: Bridging the gap between ethics and practice: guidelines for reliable, safe, and trustworthy human-centered AI systems. ACM Trans. Interact. Intell. Syst. 10(4), 1\u201331 (2020)","journal-title":"ACM Trans. Interact. Intell. Syst."},{"key":"13_CR18","doi-asserted-by":"publisher","unstructured":"Natali, C., Famiglini, L., Campagner, A., La Maida, G.A., Gallazzi, E., Cabitza, F.: Color shadows 2: Assessing the impact of xai on diagnostic decision-making. In: Longo, L. (ed.) Explainable Artificial Intelligence. xAI 2023. CCIS, vol. 1901. Springer, Cham (2023). https:\/\/doi.org\/10.1007\/978-3-031-44064-9_33","DOI":"10.1007\/978-3-031-44064-9_33"},{"issue":"9","key":"13_CR19","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3561048","volume":"55","author":"R Dwivedi","year":"2023","unstructured":"Dwivedi, R., et al.: Explainable AI (XAI): core ideas, techniques, and solutions. ACM Comput. Surv. 55(9), 1\u201333 (2023)","journal-title":"ACM Comput. Surv."},{"key":"13_CR20","doi-asserted-by":"crossref","unstructured":"Eiband, M., Schneider, H., Bilandzic, M., Fazekas-Con, J., Haug, M., Hussmann, H.: Bringing transparency design into practice. In:\u00a0Proceedings of the 23rd International Conference on Intelligent User Interfaces, pp. 211\u2013223 (2018)","DOI":"10.1145\/3172944.3172961"},{"key":"13_CR21","unstructured":"Liao, Q.V., Varshney, K.R.: Human-centered explainable ai (xai): from algorithms to user experiences. arXiv preprint arXiv:2110.10790 (2021)"},{"key":"13_CR22","doi-asserted-by":"crossref","unstructured":"Schmitt, V., et al.: Evaluating Human-Centered AI Explanations: introduction of an XAI evaluation framework for fact-checking. In:\u00a0Proceedings of the 3rd ACM International Workshop on Multimedia AI against Disinformation, pp. 91\u2013100 (2024)","DOI":"10.1145\/3643491.3660283"},{"key":"13_CR23","unstructured":"Bombassei De Bona, F., Dominici, G., Miller, T., Langheinrich, M., Gjoreski, M.: Evaluating explanations through LLMs: beyond traditional user studies.\u00a0arXiv e-prints, pp. arXiv-2410 (2024)"},{"key":"13_CR24","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.artint.2018.07.007","volume":"267","author":"T Miller","year":"2019","unstructured":"Miller, T.: Explanation in artificial intelligence: insights from the social sciences. Artif. Intell. 267, 1\u201338 (2019)","journal-title":"Artif. Intell."},{"key":"13_CR25","doi-asserted-by":"publisher","DOI":"10.1016\/j.jrt.2024.100076","volume":"17","author":"M Naiseh","year":"2024","unstructured":"Naiseh, M., Simkute, A., Zieni, B., Jiang, N., Ali, R.: C-XAI: a conceptual framework for designing XAI tools that support trust calibration. J. Respon. Technol. 17, 100076 (2024)","journal-title":"J. Respon. Technol."},{"key":"13_CR26","doi-asserted-by":"crossref","unstructured":"Liao, Q.V., Gruen, D., Miller, S.: Questioning the AI: informing design practices for explainable AI user experiences. In:\u00a0Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems, pp. 1\u201315 (2020)","DOI":"10.1145\/3313831.3376590"},{"key":"13_CR27","doi-asserted-by":"publisher","unstructured":"Ferreira, J.J., Monteiro, M.S.: What are people doing about XAI user experience? A survey on AI explainability research and practice. In: Marcus, A., Rosenzweig, E. (eds.) Design, User Experience, and Usability. Design for Contemporary Interactive Environments. HCII 2020. LNCS, vol. 12201. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-49760-6_4","DOI":"10.1007\/978-3-030-49760-6_4"},{"key":"13_CR28","doi-asserted-by":"crossref","unstructured":"Naiseh, M., et al.: Attitudes towards AI: the interplay of self-efficacy, well-being, and competency.\u00a0J. Technol. Behav. Sci. 1\u201314 (2025)","DOI":"10.1007\/s41347-025-00486-2"}],"container-title":["Lecture Notes in Computer Science","Explainable, Trustworthy, and Responsible AI and Multi-Agent Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-032-01399-6_13","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,9]],"date-time":"2025-10-09T12:51:17Z","timestamp":1760014277000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-032-01399-6_13"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,10,10]]},"ISBN":["9783032013989","9783032013996"],"references-count":28,"URL":"https:\/\/doi.org\/10.1007\/978-3-032-01399-6_13","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,10,10]]},"assertion":[{"value":"10 October 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"EXTRAAMAS","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Workshop on Explainable, Trustworthy, and Responsible AI and Multi-Agent Systems","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Detroit, MI","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"USA","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"19 May 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"20 May 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"7","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"extraamas2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/extraamas.ehealth.hevs.ch\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}