{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,25]],"date-time":"2026-03-25T05:55:28Z","timestamp":1774418128392,"version":"3.50.1"},"publisher-location":"Cham","reference-count":101,"publisher":"Springer International Publishing","isbn-type":[{"value":"9783030497590","type":"print"},{"value":"9783030497606","type":"electronic"}],"license":[{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020]]},"DOI":"10.1007\/978-3-030-49760-6_4","type":"book-chapter","created":{"date-parts":[[2020,7,9]],"date-time":"2020-07-09T23:07:54Z","timestamp":1594336074000},"page":"56-73","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":57,"title":["What Are People Doing About XAI User Experience? A Survey on AI Explainability Research and Practice"],"prefix":"10.1007","author":[{"given":"Juliana J.","family":"Ferreira","sequence":"first","affiliation":[]},{"given":"Mateus S.","family":"Monteiro","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2020,7,10]]},"reference":[{"key":"4_CR1","unstructured":"Apicella, A., Isgro, F., Prevete, R., Tamburrini, G., Vietri, A.: Sparse dictionaries for the explanation of classification systems. In: PIE, p. 009 (2015)"},{"key":"4_CR2","doi-asserted-by":"crossref","unstructured":"Barria-Pineda, J., Brusilovsky, P.: Making educational recommendations transparent through a fine-grained open learner model. In: IUI Workshops (2019)","DOI":"10.1145\/3314183.3323463"},{"key":"4_CR3","doi-asserted-by":"crossref","unstructured":"Belle, V.: Logic meets probability: towards explainable AI systems for uncertain worlds. In: IJCAI, pp. 5116\u20135120 (2017)","DOI":"10.24963\/ijcai.2017\/733"},{"key":"4_CR4","doi-asserted-by":"crossref","unstructured":"Benjamin, J.J., M\u00fcller-Birn, C.: Materializing interpretability: exploring meaning in algorithmic systems. In: Companion Publication of the 2019 on Designing Interactive Systems Conference 2019 Companion, pp. 123\u2013127. ACM (2019)","DOI":"10.1145\/3301019.3323900"},{"key":"4_CR5","doi-asserted-by":"crossref","unstructured":"Bhatia, A., Garg, V., Haves, P., Pudi, V.: Explainable clustering using hyper-rectangles for building energy simulation data. In: IOP Conference Series: Earth and Environmental Science, vol. 238, p. 012068. IOP Publishing (2019)","DOI":"10.1088\/1755-1315\/238\/1\/012068"},{"key":"4_CR6","doi-asserted-by":"crossref","unstructured":"Browne, J.T.: Wizard of OZ prototyping for machine learning experiences. In: Extended Abstracts of the 2019 CHI Conference on Human Factors in Computing Systems, p. LBW2621. ACM (2019)","DOI":"10.1145\/3290607.3312877"},{"key":"4_CR7","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"27","DOI":"10.1007\/978-3-030-29726-8_3","volume-title":"Machine Learning and Knowledge Extraction","author":"F Cabitza","year":"2019","unstructured":"Cabitza, F., Campagner, A., Ciucci, D.: New frontiers in explainable AI: understanding the GI to interpret the GO. In: Holzinger, A., Kieseberg, P., Tjoa, A.M., Weippl, E. (eds.) CD-MAKE 2019. LNCS, vol. 11713, pp. 27\u201347. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-29726-8_3"},{"key":"4_CR8","doi-asserted-by":"crossref","unstructured":"Cai, C.J., Jongejan, J., Holbrook, J.: The effects of example-based explanations in a machine learning interface. In: Proceedings of the 24th International Conference on Intelligent User Interfaces, pp. 258\u2013262. ACM (2019)","DOI":"10.1145\/3301275.3302289"},{"key":"4_CR9","unstructured":"Chander, A., Srinivasan, R., Chelian, S., Wang, J., Uchino, K.: Working with beliefs: AI transparency in the enterprise. In: IUI Workshops (2018)"},{"key":"4_CR10","doi-asserted-by":"crossref","unstructured":"Charleer, S., Guti\u00e9rrez, F., Verbert, K.: Supporting job mediator and job seeker through an actionable dashboard. In: Proceedings of the 24th International Conference on Intelligent User Interfaces, pp. 121\u2013131 (2019)","DOI":"10.1145\/3301275.3302312"},{"key":"4_CR11","doi-asserted-by":"crossref","unstructured":"Chen, L., Wang, F.: Explaining recommendations based on feature sentiments in product reviews. In: Proceedings of the 22nd International Conference on Intelligent User Interfaces, pp. 17\u201328. ACM (2017)","DOI":"10.1145\/3025171.3025173"},{"key":"4_CR12","doi-asserted-by":"crossref","unstructured":"Cheng, H.F., et al.: Explaining decision-making algorithms through UI: strategies to help non-expert stakeholders. In: Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems, p. 559. ACM (2019)","DOI":"10.1145\/3290605.3300789"},{"key":"4_CR13","unstructured":"Chromik, M., Eiband, M., V\u00f6lkel, S.T., Buschek, D.: Dark patterns of explainability, transparency, and user control for intelligent systems. In: IUI Workshops (2019)"},{"key":"4_CR14","doi-asserted-by":"crossref","unstructured":"Clewley, N., Dodd, L., Smy, V., Witheridge, A., Louvieris, P.: Eliciting expert knowledge to inform training design. In: Proceedings of the 31st European Conference on Cognitive Ergonomics, pp. 138\u2013143 (2019)","DOI":"10.1145\/3335082.3335091"},{"key":"4_CR15","doi-asserted-by":"crossref","unstructured":"Datta, A., Sen, S., Zick, Y.: Algorithmic transparency via quantitative input influence: theory and experiments with learning systems. In: 2016 IEEE Symposium on Security and Privacy (SP), pp. 598\u2013617. IEEE (2016)","DOI":"10.1109\/SP.2016.42"},{"key":"4_CR16","unstructured":"Di Castro, F., Bertini, E.: Surrogate decision tree visualization interpreting and visualizing black-box classification models with surrogate decision tree. In: CEUR Workshop Proceedings, vol. 2327 (2019)"},{"key":"4_CR17","doi-asserted-by":"crossref","unstructured":"Dimitrova, R., Majumdar, R., Prabhu, V.S.: Causality analysis for concurrent reactive systems. arXiv preprint arXiv:1901.00589 (2019)","DOI":"10.4204\/EPTCS.286.3"},{"key":"4_CR18","doi-asserted-by":"publisher","first-page":"1561","DOI":"10.1016\/j.procs.2018.08.129","volume":"126","author":"L Ding","year":"2018","unstructured":"Ding, L.: Human knowledge in constructing AI systems-neural logic networks approach towards an explainable AI. Procedia Comput. Sci. 126, 1561\u20131570 (2018)","journal-title":"Procedia Comput. Sci."},{"key":"4_CR19","doi-asserted-by":"crossref","unstructured":"Dodge, J., Liao, Q.V., Zhang, Y., Bellamy, R.K., Dugan, C.: Explaining models: an empirical study of how explanations impact fairness judgment. In: Proceedings of the 24th International Conference on Intelligent User Interfaces, pp. 275\u2013285. ACM (2019)","DOI":"10.1145\/3301275.3302310"},{"key":"4_CR20","unstructured":"Dodge, J., Penney, S., Anderson, A., Burnett, M.M.: What should be in an XAI explanation? what IFT reveals. In: IUI Workshops (2018)"},{"key":"4_CR21","doi-asserted-by":"crossref","unstructured":"Dominguez, V., Messina, P., Donoso-Guzm\u00e1n, I., Parra, D.: The effect of explanations and algorithmic accuracy on visual recommender systems of artistic images. In: Proceedings of the 24th International Conference on Intelligent User Interfaces, pp. 408\u2013416. ACM (2019)","DOI":"10.1145\/3301275.3302274"},{"key":"4_CR22","doi-asserted-by":"crossref","unstructured":"Ehsan, U., Tambwekar, P., Chan, L., Harrison, B., Riedl, M.O.: Automated rationale generation: a technique for explainable AI and its effects on human perceptions. In: Proceedings of the 24th International Conference on Intelligent User Interfaces, pp. 263\u2013274. ACM (2019)","DOI":"10.1145\/3301275.3302316"},{"key":"4_CR23","doi-asserted-by":"crossref","unstructured":"Eiband, M., Buschek, D., Kremer, A., Hussmann, H.: The impact of placebic explanations on trust in intelligent systems. In: Extended Abstracts of the 2019 CHI Conference on Human Factors in Computing Systems, p. LBW0243. ACM (2019)","DOI":"10.1145\/3290607.3312787"},{"key":"4_CR24","unstructured":"Eiband, M., Schneider, H., Buschek, D.: Normative vs. pragmatic: two perspectives on the design of explanations in intelligent systems. In: IUI Workshops (2018)"},{"key":"4_CR25","series-title":"Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence)","doi-asserted-by":"publisher","first-page":"78","DOI":"10.1007\/978-3-030-01081-2_6","volume-title":"Case-Based Reasoning Research and Development","author":"V Eisenstadt","year":"2018","unstructured":"Eisenstadt, V., Espinoza-Stapelfeld, C., Mikyas, A., Althoff, K.-D.: Explainable distributed case-based support systems: patterns for enhancement and validation of design recommendations. In: Cox, M.T., Funk, P., Begum, S. (eds.) ICCBR 2018. LNCS (LNAI), vol. 11156, pp. 78\u201394. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01081-2_6"},{"key":"4_CR26","series-title":"Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence)","doi-asserted-by":"publisher","first-page":"49","DOI":"10.1007\/978-3-030-29249-2_4","volume-title":"Case-Based Reasoning Research and Development","author":"V Eisenstadt","year":"2019","unstructured":"Eisenstadt, V., Langenhan, C., Althoff, K.-D.: FLEA-CBR \u2013 a flexible alternative to the classic 4R cycle of case-based reasoning. In: Bach, K., Marling, C. (eds.) ICCBR 2019. LNCS (LNAI), vol. 11680, pp. 49\u201363. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-29249-2_4"},{"key":"4_CR27","unstructured":"Eljasik-Swoboda, T., Engel, F., Hemmje, M.: Using topic specific features for argument stance recognition"},{"key":"4_CR28","doi-asserted-by":"crossref","unstructured":"Escalante, H.J., et al.: Design of an explainable machine learning challenge for video interviews. In: 2017 International Joint Conference on Neural Networks (IJCNN), pp. 3688\u20133695. IEEE (2017)","DOI":"10.1109\/IJCNN.2017.7966320"},{"key":"4_CR29","doi-asserted-by":"crossref","unstructured":"Finkbeiner, B., Kleinberg, S.: Proceedings 3rd workshop on formal reasoning about causation, responsibility, and explanations in science and technology. arXiv preprint arXiv:1901.00073 (2019)","DOI":"10.4204\/EPTCS.286.0"},{"key":"4_CR30","doi-asserted-by":"publisher","first-page":"30","DOI":"10.1016\/j.cag.2018.09.018","volume":"77","author":"R Garcia","year":"2018","unstructured":"Garcia, R., Telea, A.C., da Silva, B.C., T\u00f8rresen, J., Comba, J.L.D.: A task-and-technique centered survey on visual analytics for deep learning model engineering. Comput. Graph. 77, 30\u201349 (2018)","journal-title":"Comput. Graph."},{"key":"4_CR31","unstructured":"Gervasio, M.T., Myers, K.L., Yeh, E., Adkins, B.: Explanation to avert surprise. In: IUI Workshops, vol. 2068 (2018)"},{"key":"4_CR32","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"295","DOI":"10.1007\/978-3-319-99740-7_21","volume-title":"Machine Learning and Knowledge Extraction","author":"R Goebel","year":"2018","unstructured":"Goebel, R., et al.: Explainable AI: the new 42? In: Holzinger, A., Kieseberg, P., Tjoa, A.M., Weippl, E. (eds.) CD-MAKE 2018. LNCS, vol. 11015, pp. 295\u2013303. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-319-99740-7_21"},{"key":"4_CR33","doi-asserted-by":"publisher","first-page":"26","DOI":"10.1016\/j.eswa.2016.11.017","volume":"71","author":"MB Gorza\u0142czany","year":"2017","unstructured":"Gorza\u0142czany, M.B., Rudzi\u0144ski, F.: Interpretable and accurate medical data classification-a multi-objective genetic-fuzzy optimization approach. Expert Syst. Appl. 71, 26\u201339 (2017)","journal-title":"Expert Syst. Appl."},{"key":"4_CR34","series-title":"Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence)","doi-asserted-by":"publisher","first-page":"255","DOI":"10.1007\/978-3-319-91470-1_22","volume-title":"Augmented Cognition: Intelligent Technologies","author":"SS Grigsby","year":"2018","unstructured":"Grigsby, S.S.: Artificial intelligence for advanced human-machine symbiosis. In: Schmorrow, D.D., Fidopiastis, C.M. (eds.) AC 2018. LNCS (LNAI), vol. 10915, pp. 255\u2013266. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-319-91470-1_22"},{"key":"4_CR35","unstructured":"Guo, K., Pratt, D., MacDonald III, A., Schrater, P.: Labeling images by interpretation from natural viewing. In: IUI Workshops (2018)"},{"key":"4_CR36","unstructured":"Guzdial, M., Reno, J., Chen, J., Smith, G., Riedl, M.: Explainable PCGML via game design patterns. arXiv preprint arXiv:1809.09419 (2018)"},{"key":"4_CR37","unstructured":"Hamidi-Haines, M., Qi, Z., Fern, A., Li, F., Tadepalli, P.: Interactive naming for explaining deep neural networks: a formative study. arXiv preprint arXiv:1812.07150 (2018)"},{"key":"4_CR38","unstructured":"Hepenstal, S., Kodagoda, N., Zhang, L., Paudyal, P., Wong, B.W.: Algorithmic transparency of conversational agents. In: IUI Workshops (2019)"},{"key":"4_CR39","doi-asserted-by":"crossref","unstructured":"Hohman, F., Head, A., Caruana, R., DeLine, R., Drucker, S.M.: Gamut: a design probe to understand how data scientists understand machine learning models. In: Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems, p. 579. ACM (2019)","DOI":"10.1145\/3290605.3300809"},{"issue":"8","key":"4_CR40","doi-asserted-by":"publisher","first-page":"2674","DOI":"10.1109\/TVCG.2018.2843369","volume":"25","author":"FM Hohman","year":"2018","unstructured":"Hohman, F.M., Kahng, M., Pienta, R., Chau, D.H.: Visual analytics in deep learning: an interrogative survey for the next frontiers. IEEE Trans. Vis. Comput. Graph. 25(8), 2674\u20132693 (2018)","journal-title":"IEEE Trans. Vis. Comput. Graph."},{"key":"4_CR41","first-page":"1","volume":"34","author":"K Ishii","year":"2017","unstructured":"Ishii, K.: Comparative legal study on privacy and personal data protection for robots equipped with artificial intelligence: looking at functional and technological aspects. AI Soc. 34, 1\u201325 (2017)","journal-title":"AI Soc."},{"key":"4_CR42","doi-asserted-by":"crossref","unstructured":"Jain, A., Keller, J., Popescu, M.: Explainable AI for dataset comparison. In: 2019 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE), pp. 1\u20137. IEEE (2019)","DOI":"10.1109\/FUZZ-IEEE.2019.8858911"},{"key":"4_CR43","series-title":"Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence)","doi-asserted-by":"publisher","first-page":"77","DOI":"10.1007\/978-3-030-30391-4_5","volume-title":"Explainable, Transparent Autonomous Agents and Multi-Agent Systems","author":"SF Jentzsch","year":"2019","unstructured":"Jentzsch, S.F., H\u00f6hn, S., Hochgeschwender, N.: Conversational interfaces for explainable AI: a human-centred approach. In: Calvaresi, D., Najjar, A., Schumacher, M., Fr\u00e4mling, K. (eds.) EXTRAAMAS 2019. LNCS (LNAI), vol. 11763, pp. 77\u201392. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-30391-4_5"},{"key":"4_CR44","series-title":"Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence)","doi-asserted-by":"publisher","first-page":"59","DOI":"10.1007\/978-3-030-30391-4_4","volume-title":"Explainable, Transparent Autonomous Agents and Multi-Agent Systems","author":"T Kampik","year":"2019","unstructured":"Kampik, T., Nieves, J.C., Lindgren, H.: Explaining sympathetic actions of rational agents. In: Calvaresi, D., Najjar, A., Schumacher, M., Fr\u00e4mling, K. (eds.) EXTRAAMAS 2019. LNCS (LNAI), vol. 11763, pp. 59\u201376. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-30391-4_4"},{"key":"4_CR45","doi-asserted-by":"crossref","unstructured":"Kizilcec, R.F.: How much information?: Effects of transparency on trust in an algorithmic interface. In: Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems, pp. 2390\u20132395. ACM (2016)","DOI":"10.1145\/2858036.2858402"},{"key":"4_CR46","doi-asserted-by":"crossref","unstructured":"Krebs, L.M., et al.: Tell me what you know: GDPR implications on designing transparency and accountability for news recommender systems. In: Extended Abstracts of the 2019 CHI Conference on Human Factors in Computing Systems, p. LBW2610. ACM (2019)","DOI":"10.1145\/3290607.3312808"},{"key":"4_CR47","unstructured":"Krishnan, J., Coronado, P., Reed, T.: SEVA: a systems engineer\u2019s virtual assistant. In: AAAI Spring Symposium: Combining Machine Learning with Knowledge Engineering (2019)"},{"issue":"1","key":"4_CR48","doi-asserted-by":"publisher","first-page":"299","DOI":"10.1109\/TVCG.2018.2865027","volume":"25","author":"BC Kwon","year":"2018","unstructured":"Kwon, B.C., et al.: RetainVis: visual analytics with interpretable and interactive recurrent neural networks on electronic medical records. IEEE Trans. Vis. Comput. Graph. 25(1), 299\u2013309 (2018)","journal-title":"IEEE Trans. Vis. Comput. Graph."},{"key":"4_CR49","unstructured":"Lee, O.J., Jung, J.J.: Explainable movie recommendation systems by using story-based similarity. In: IUI Workshops (2018)"},{"key":"4_CR50","unstructured":"Lim, B.Y., Wang, D., Loh, T.P., Ngiam, K.Y.: Interpreting intelligibility under uncertain data imputation. In: IUI Workshops (2018)"},{"key":"4_CR51","unstructured":"Lim, B.Y., Yang, Q., Abdul, A.M., Wang, D.: Why these explanations? selecting intelligibility types for explanation goals. In: IUI Workshops (2019)"},{"key":"4_CR52","doi-asserted-by":"crossref","unstructured":"Loi, D., Wolf, C.T., Blomberg, J.L., Arar, R., Brereton, M.: Co-designing AI futures: Integrating AI ethics, social computing, and design. In: A Companion Publication of the 2019 on Designing Interactive Systems Conference 2019 Companion, pp. 381\u2013384. ACM (2019)","DOI":"10.1145\/3301019.3320000"},{"key":"4_CR53","doi-asserted-by":"publisher","first-page":"109","DOI":"10.1016\/j.ins.2019.05.016","volume":"496","author":"L Magdalena","year":"2019","unstructured":"Magdalena, L.: Semantic interpretability in hierarchical fuzzy systems: creating semantically decouplable hierarchies. Inf. Sci. 496, 109\u2013123 (2019)","journal-title":"Inf. Sci."},{"key":"4_CR54","series-title":"Advances in Intelligent Systems and Computing","doi-asserted-by":"publisher","first-page":"939","DOI":"10.1007\/978-3-030-22871-2_67","volume-title":"Intelligent Computing","author":"S Meacham","year":"2019","unstructured":"Meacham, S., Isaac, G., Nauck, D., Virginas, B.: Towards explainable AI: design and development for explanation of machine learning predictions for a patient readmittance medical application. In: Arai, K., Bhatia, R., Kapoor, S. (eds.) CompCom 2019. AISC, vol. 997, pp. 939\u2013955. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-22871-2_67"},{"key":"4_CR55","doi-asserted-by":"crossref","unstructured":"Millecamp, M., Htun, N.N., Conati, C., Verbert, K.: To explain or not to explain: the effects of personal characteristics when explaining music recommendations. In: IUI, pp. 397\u2013407 (2019)","DOI":"10.1145\/3301275.3302313"},{"issue":"1","key":"4_CR56","doi-asserted-by":"publisher","first-page":"342","DOI":"10.1109\/TVCG.2018.2864812","volume":"25","author":"Y Ming","year":"2018","unstructured":"Ming, Y., Qu, H., Bertini, E.: RuleMatrix: visualizing and understanding classifiers with rules. IEEE Trans. Vis. Comput. Graph. 25(1), 342\u2013352 (2018)","journal-title":"IEEE Trans. Vis. Comput. Graph."},{"key":"4_CR57","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.dsp.2017.10.011","volume":"73","author":"G Montavon","year":"2018","unstructured":"Montavon, G., Samek, W., M\u00fcller, K.R.: Methods for interpreting and understanding deep neural networks. Digit. Signal Proc. 73, 1\u201315 (2018)","journal-title":"Digit. Signal Proc."},{"key":"4_CR58","doi-asserted-by":"publisher","first-page":"56","DOI":"10.1016\/j.eswa.2019.03.054","volume":"129","author":"JLZ Montenegro","year":"2019","unstructured":"Montenegro, J.L.Z., da Costa, C.A., Righi, R.D.R.: Survey of conversational agents in health. Expert Syst. Appl. 129, 56\u201367 (2019). https:\/\/doi.org\/10.1016\/j.eswa.2019.03.054. http:\/\/www.sciencedirect.com\/science\/article\/pii\/S0957417419302283","journal-title":"Expert Syst. Appl."},{"key":"4_CR59","doi-asserted-by":"crossref","unstructured":"Nassar, M., Salah, K., ur Rehman, M.H., Svetinovic, D.: Blockchain for explainable and trustworthy artificial intelligence. Wiley Interdisc. Rev.: Data Min. Knowl. Discovery 10(1), e1340 (2020)","DOI":"10.1002\/widm.1340"},{"key":"4_CR60","series-title":"Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence)","doi-asserted-by":"publisher","first-page":"204","DOI":"10.1007\/978-3-319-91122-9_18","volume-title":"Engineering Psychology and Cognitive Ergonomics","author":"MA Neerincx","year":"2018","unstructured":"Neerincx, M.A., van der Waa, J., Kaptein, F., van Diggelen, J.: Using perceptual and cognitive explanations for enhanced human-agent team performance. In: Harris, D. (ed.) EPCE 2018. LNCS (LNAI), vol. 10906, pp. 204\u2013214. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-319-91122-9_18"},{"key":"4_CR61","doi-asserted-by":"crossref","unstructured":"Nguyen, A.T., et al.: Believe it or not: designing a human-AI partnership for mixed-initiative fact-checking. In: The 31st Annual ACM Symposium on User Interface Software and Technology, pp. 189\u2013199. ACM (2018)","DOI":"10.1145\/3242587.3242666"},{"key":"4_CR62","doi-asserted-by":"crossref","unstructured":"Nguyen, A.T., Lease, M., Wallace, B.C.: Explainable modeling of annotations in crowdsourcing. In: IUI, pp. 575\u2013579 (2019)","DOI":"10.1145\/3301275.3302276"},{"key":"4_CR63","unstructured":"Nguyen, A.T., Lease, M., Wallace, B.C.: Mash: software tools for developing interactive and transparent machine learning systems. In: IUI Workshops (2019)"},{"issue":"3\u20135","key":"4_CR64","doi-asserted-by":"publisher","first-page":"393","DOI":"10.1007\/s11257-017-9195-0","volume":"27","author":"I Nunes","year":"2017","unstructured":"Nunes, I., Jannach, D.: A systematic review and taxonomy of explanations in decision support and recommender systems. User Model. User-Adap. Inter. 27(3\u20135), 393\u2013444 (2017)","journal-title":"User Model. User-Adap. Inter."},{"key":"4_CR65","doi-asserted-by":"crossref","unstructured":"Olszewska, J.I.: Designing transparent and autonomous intelligent vision systems. In: Proceedings of the International Conference on Agents and Artificial Intelligence (ICAART), pp. 850\u2013856 (2019)","DOI":"10.5220\/0007585208500856"},{"key":"4_CR66","doi-asserted-by":"crossref","unstructured":"van Oosterhout, A.: Understanding the benefits and drawbacks of shape change in contrast or addition to other modalities. In: Companion Publication of the 2019 on Designing Interactive Systems Conference 2019 Companion, pp. 113\u2013116. ACM (2019)","DOI":"10.1145\/3301019.3324875"},{"key":"4_CR67","unstructured":"van Otterlo, M., Atzmueller, M.: On requirements and design criteria for explainability in legal AI (2018)"},{"key":"4_CR68","unstructured":"Paudyal, P., Lee, J., Kamzin, A., Soudki, M., Banerjee, A., Gupta, S.K.: Learn2sign: explainable AI for sign language learning. In: IUI Workshops (2019)"},{"key":"4_CR69","doi-asserted-by":"crossref","unstructured":"Petersen, K., Feldt, R., Mujtaba, S., Mattsson, M.: Systematic mapping studies in software engineering. In: Ease, vol. 8, pp. 68\u201377 (2008)","DOI":"10.14236\/ewic\/EASE2008.8"},{"key":"4_CR70","unstructured":"Ribera, M., Lapedriza, \u00c0.: Can we do better explanations? A proposal of user-centered explainable AI. In: IUI Workshops (2019)"},{"key":"4_CR71","unstructured":"Rotsidis, A., Theodorou, A., Wortham, R.H.: Robots that make sense: transparent intelligence through augmented reality. In: IUI Workshops (2019)"},{"key":"4_CR72","doi-asserted-by":"publisher","unstructured":"Santos, T.I., Abel, A.: Using feature visualisation for explaining deep learning models in visual speech. In: 2019 IEEE 4th International Conference on Big Data Analytics (ICBDA), pp. 231\u2013235, March 2019. https:\/\/doi.org\/10.1109\/ICBDA.2019.8713256","DOI":"10.1109\/ICBDA.2019.8713256"},{"key":"4_CR73","doi-asserted-by":"crossref","unstructured":"Schmidmaier, M., Han, Z., Weber, T., Liu, Y., Hu\u00dfmann, H.: Real-time personalization in adaptive ides (2019)","DOI":"10.1145\/3314183.3324975"},{"key":"4_CR74","doi-asserted-by":"crossref","unstructured":"Schuessler, M., Wei\u00df, P.: Minimalistic explanations: capturing the essence of decisions. arXiv preprint arXiv:1905.02994 (2019)","DOI":"10.1145\/3290607.3312823"},{"key":"4_CR75","doi-asserted-by":"crossref","unstructured":"Sellam, T., Lin, K., Huang, I., Yang, M., Vondrick, C., Wu, E.: DeepBase: deep inspection of neural networks. In: Proceedings of the 2019 International Conference on Management of Data, pp. 1117\u20131134 (2019)","DOI":"10.1145\/3299869.3300073"},{"key":"4_CR76","doi-asserted-by":"crossref","unstructured":"Singh, M., Martins, L.M., Joanis, P., Mago, V.K.: Building a cardiovascular disease predictive model using structural equation model & fuzzy cognitive map. In: 2016 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE), pp. 1377\u20131382. IEEE (2016)","DOI":"10.1109\/FUZZ-IEEE.2016.7737850"},{"key":"4_CR77","unstructured":"Sliwinski, J., Strobel, M., Zick, Y.: An axiomatic approach to linear explanations in data classification. In: IUI Workshops (2018)"},{"key":"4_CR78","unstructured":"Smith, A., Nolan, J.: The problem of explanations without user feedback. In: IUI Workshops (2018)"},{"key":"4_CR79","unstructured":"Smith-Renner, A., Rua, R., Colony, M.: Towards an explainable threat detection tool. In: IUI Workshops (2019)"},{"key":"4_CR80","doi-asserted-by":"crossref","unstructured":"Sokol, K., Flach, P.A.: Conversational explanations of machine learning predictions through class-contrastive counterfactual statements. In: IJCAI, pp. 5785\u20135786 (2018)","DOI":"10.24963\/ijcai.2018\/836"},{"key":"4_CR81","doi-asserted-by":"crossref","unstructured":"Springer, A., Whittaker, S.: Progressive disclosure: designing for effective transparency. arXiv preprint arXiv:1811.02164 (2018)","DOI":"10.1145\/3301275.3302322"},{"key":"4_CR82","unstructured":"Stumpf, S.: Horses for courses: making the case for persuasive engagement in smart systems. In: Joint Proceedings of the ACM IUI 2019 Workshops, vol. 2327. CEUR (2019)"},{"key":"4_CR83","unstructured":"Stumpf, S., Skrebe, S., Aymer, G., Hobson, J.: Explaining smart heating systems to discourage fiddling with optimized behavior. In: CEUR Workshop Proceedings, vol. 2068 (2018)"},{"key":"4_CR84","unstructured":"Sundararajan, M., Xu, J., Taly, A., Sayres, R., Najmi, A.: Exploring principled visualizations for deep network attributions. In: IUI Workshops (2019)"},{"issue":"3","key":"4_CR85","doi-asserted-by":"publisher","first-page":"230","DOI":"10.1080\/09540091.2017.1310182","volume":"29","author":"A Theodorou","year":"2017","unstructured":"Theodorou, A., Wortham, R.H., Bryson, J.J.: Designing and implementing transparency for real time inspection of autonomous robots. Connect. Sci. 29(3), 230\u2013241 (2017)","journal-title":"Connect. Sci."},{"key":"4_CR86","doi-asserted-by":"crossref","unstructured":"Tsai, C.H., Brusilovsky, P.: Explaining social recommendations to casual users: design principles and opportunities. In: Proceedings of the 23rd International Conference on Intelligent User Interfaces Companion, p. 59. ACM (2018)","DOI":"10.1145\/3180308.3180368"},{"key":"4_CR87","unstructured":"Tsai, C.H., Brusilovsky, P.: Designing explanation interfaces for transparency and beyond. In: IUI Workshops (2019)"},{"key":"4_CR88","unstructured":"Vellido, A.: The importance of interpretability and visualization in machine learning for applications in medicine and health care. Neural Comput. Appl. 1\u201315 (2019)"},{"key":"4_CR89","doi-asserted-by":"crossref","unstructured":"Vijay, A., Umadevi, K.: Secured AI guided architecture for D2D systems of massive MIMO deployed in 5G networks. In: 2019 3rd International Conference on Trends in Electronics and Informatics (ICOEI), pp. 468\u2013472. IEEE (2019)","DOI":"10.1109\/ICOEI.2019.8862712"},{"key":"4_CR90","unstructured":"Vorm, E.S., Miller, A.D.: Assessing the value of transparency in recommender systems: an end-user perspective (2018)"},{"key":"4_CR91","doi-asserted-by":"crossref","unstructured":"Wang, D., Yang, Q., Abdul, A., Lim, B.Y.: Designing theory-driven user-centric explainable AI. In: Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems, p. 601. ACM (2019)","DOI":"10.1145\/3290605.3300831"},{"key":"4_CR92","doi-asserted-by":"crossref","unstructured":"Wang, Q., et al.: ATMSeer: increasing transparency and controllability in automated machine learning. In: Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems, p. 681. ACM (2019)","DOI":"10.1145\/3290605.3300911"},{"key":"4_CR93","doi-asserted-by":"crossref","unstructured":"Wang, X., Chen, Y., Yang, J., Wu, L., Wu, Z., Xie, X.: A reinforcement learning framework for explainable recommendation. In: 2018 IEEE International Conference on Data Mining (ICDM), pp. 587\u2013596. IEEE (2018)","DOI":"10.1109\/ICDM.2018.00074"},{"key":"4_CR94","doi-asserted-by":"crossref","unstructured":"Wolf, C.T., Blomberg, J.: Explainability in context: lessons from an intelligent system in the it services domain. In: IUI Workshops (2019)","DOI":"10.1145\/3301275.3302317"},{"key":"4_CR95","unstructured":"Xie, Y., Gao, G., Chen, X.: Outlining the design space of explainable intelligent systems for medical diagnosis. arXiv preprint arXiv:1902.06019 (2019)"},{"key":"4_CR96","doi-asserted-by":"crossref","unstructured":"Yang, Q., Banovic, N., Zimmerman, J.: Mapping machine learning advances from HCI research to reveal starting places for design innovation. In: Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems, p. 130. ACM (2018)","DOI":"10.1145\/3173574.3173704"},{"key":"4_CR97","doi-asserted-by":"crossref","unstructured":"Yeganejou, M., Dick, S.: Improved deep fuzzy clustering for accurate and interpretable classifiers. In: 2019 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE), pp. 1\u20137. IEEE (2019)","DOI":"10.1109\/FUZZ-IEEE.2019.8858809"},{"key":"4_CR98","unstructured":"Zhao, R., Benbasat, I., Cavusoglu, H.: Transparency in advice-giving systems: a framework and a research model for transparency provision. In: IUI Workshops (2019)"},{"key":"4_CR99","doi-asserted-by":"crossref","unstructured":"Zheng, X.l., Zhu, M.Y., Li, Q.B., Chen, C.C., Tan, Y.C.: FinBrain: when finance meets AI 2.0. Front. Inf. Technol. Electron. Eng. 20(7), 914\u2013924 (2019)","DOI":"10.1631\/FITEE.1700822"},{"key":"4_CR100","doi-asserted-by":"crossref","unstructured":"Zhou, J., et al.: Effects of influence on user trust in predictive decision making. In: Extended Abstracts of the 2019 CHI Conference on Human Factors in Computing Systems, pp. 1\u20136 (2019)","DOI":"10.1145\/3290607.3312962"},{"key":"4_CR101","doi-asserted-by":"crossref","unstructured":"Zhu, J., Liapis, A., Risi, S., Bidarra, R., Youngblood, G.M.: Explainable AI for designers: a human-centered perspective on mixed-initiative co-creation. In: 2018 IEEE Conference on Computational Intelligence and Games (CIG), pp. 1\u20138. IEEE (2018)","DOI":"10.1109\/CIG.2018.8490433"}],"container-title":["Lecture Notes in Computer Science","Design, User Experience, and Usability. Design for Contemporary Interactive Environments"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-49760-6_4","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,7,10]],"date-time":"2024-07-10T00:23:36Z","timestamp":1720571016000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-49760-6_4"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020]]},"ISBN":["9783030497590","9783030497606"],"references-count":101,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-49760-6_4","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020]]},"assertion":[{"value":"10 July 2020","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"HCII","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Human-Computer Interaction","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Copenhagen","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Denmark","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2020","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"19 July 2020","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"24 July 2020","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"22","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"hcii2020","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/2020.hci.international\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"This content has been made available to all.","name":"free","label":"Free to read"}]}}