{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,2]],"date-time":"2026-04-02T08:32:52Z","timestamp":1775118772973,"version":"3.50.1"},"publisher-location":"Cham","reference-count":91,"publisher":"Springer International Publishing","isbn-type":[{"value":"9783031040825","type":"print"},{"value":"9783031040832","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2022,4,17]],"date-time":"2022-04-17T00:00:00Z","timestamp":1650153600000},"content-version":"vor","delay-in-days":106,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:p>Explainable Artificial Intelligence (xAI) is an established field with a vibrant community that has developed a variety of very successful approaches to explain and interpret predictions of complex machine learning models such as deep neural networks. In this article, we briefly introduce a few selected methods and discuss them in a short, clear and concise way. The goal of this article is to give beginners, especially application engineers and data scientists, a quick overview of the state of the art in this current topic. The following 17 methods are covered in this chapter: LIME, Anchors, GraphLIME, LRP, DTD, PDA, TCAV, XGNN, SHAP, ASV, Break-Down, Shapley Flow, Textual Explanations of Visual Models, Integrated Gradients, Causal Models, Meaningful Perturbations, and X-NeSyL.<\/jats:p>","DOI":"10.1007\/978-3-031-04083-2_2","type":"book-chapter","created":{"date-parts":[[2022,4,16]],"date-time":"2022-04-16T17:03:23Z","timestamp":1650128603000},"page":"13-38","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":265,"title":["Explainable AI Methods - A Brief Overview"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-6786-5194","authenticated-orcid":false,"given":"Andreas","family":"Holzinger","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1085-8428","authenticated-orcid":false,"given":"Anna","family":"Saranti","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2331-868X","authenticated-orcid":false,"given":"Christoph","family":"Molnar","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8423-1823","authenticated-orcid":false,"given":"Przemyslaw","family":"Biecek","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6283-3265","authenticated-orcid":false,"given":"Wojciech","family":"Samek","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,4,17]]},"reference":[{"key":"2_CR1","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"101","DOI":"10.1007\/978-3-030-69544-6_7","volume-title":"Computer Vision \u2013 ACCV 2020","author":"C Agarwal","year":"2021","unstructured":"Agarwal, C., Nguyen, A.: Explaining image classifiers by removing input features using generative models. In: Ishikawa, H., Liu, C.-L., Pajdla, T., Shi, J. (eds.) ACCV 2020. LNCS, vol. 12627, pp. 101\u2013118. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-69544-6_7"},{"issue":"93","key":"2_CR2","first-page":"1","volume":"20","author":"M Alber","year":"2019","unstructured":"Alber, M., et al.: iNNvestigate neural networks! J. Mach. Learn. Res. (JMLR) 20(93), 1\u20138 (2019)","journal-title":"J. Mach. Learn. Res. (JMLR)"},{"key":"2_CR3","unstructured":"Ali, A., Schnake, T., Eberle, O., Montavon, G., M\u00fcller, K.R., Wolf, L.: XAI for transformers: better explanations through conservative propagation. arXiv preprint arXiv:2202.07304 (2022)"},{"key":"2_CR4","unstructured":"Anders, C.J., Neumann, D., Samek, W., M\u00fcller, K.R., Lapuschkin, S.: Software for dataset-wide XAI: from local explanations to global insights with Zennit, CoRelAy, and ViRelAy. arXiv preprint arXiv:2106.13200 (2021)"},{"key":"2_CR5","doi-asserted-by":"publisher","first-page":"261","DOI":"10.1016\/j.inffus.2021.07.015","volume":"77","author":"CJ Anders","year":"2022","unstructured":"Anders, C.J., Weber, L., Neumann, D., Samek, W., M\u00fcller, K.R., Lapuschkin, S.: Finding and removing clever HANs: using explanation methods to debug and improve deep models. Inf. Fusion 77, 261\u2013295 (2022)","journal-title":"Inf. Fusion"},{"key":"2_CR6","series-title":"Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence)","doi-asserted-by":"publisher","first-page":"211","DOI":"10.1007\/978-3-030-28954-6_11","volume-title":"Explainable AI: Interpreting, Explaining and Visualizing Deep Learning","author":"L Arras","year":"2019","unstructured":"Arras, L., et al.: Explaining and interpreting LSTMs. In: Samek, W., Montavon, G., Vedaldi, A., Hansen, L.K., M\u00fcller, K.-R. (eds.) Explainable AI: Interpreting, Explaining and Visualizing Deep Learning. LNCS (LNAI), vol. 11700, pp. 211\u2013238. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-28954-6_11"},{"key":"2_CR7","doi-asserted-by":"crossref","unstructured":"Arras, L., Montavon, G., M\u00fcller, K.R., Samek, W.: Explaining recurrent neural network predictions in sentiment analysis. In: Proceedings of the EMNLP 2017 Workshop on Computational Approaches to Subjectivity, Sentiment & Social Media Analysis (WASSA), pp. 159\u2013168. Association for Computational Linguistics (2017)","DOI":"10.18653\/v1\/W17-5221"},{"key":"2_CR8","doi-asserted-by":"publisher","first-page":"14","DOI":"10.1016\/j.inffus.2021.11.008","volume":"81","author":"L Arras","year":"2022","unstructured":"Arras, L., Osman, A., Samek, W.: CLEVR-XAI: a benchmark dataset for the ground truth evaluation of neural network explanations. Inf. Fusion 81, 14\u201340 (2022)","journal-title":"Inf. Fusion"},{"key":"2_CR9","doi-asserted-by":"publisher","first-page":"60588","DOI":"10.1109\/ACCESS.2021.3071274","volume":"9","author":"NA Asif","year":"2021","unstructured":"Asif, N.A., et al.: Graph neural network: a comprehensive review on Non-Euclidean space. IEEE Access 9, 60588\u201360606 (2021)","journal-title":"IEEE Access"},{"key":"2_CR10","doi-asserted-by":"crossref","unstructured":"Bach, S., Binder, A., Montavon, G., Klauschen, F., M\u00fcller, K.R., Samek, W.: On pixel-wise explanations for non-linear classifier decisions by layer-wise relevance propagation. PLoS ONE 10(7), e0130140 (2015)","DOI":"10.1371\/journal.pone.0130140"},{"key":"2_CR11","first-page":"1803","volume":"11","author":"D Baehrens","year":"2010","unstructured":"Baehrens, D., Schroeter, T., Harmeling, S., Kawanabe, M., Hansen, K., M\u00fcller, K.R.: How to explain individual classification decisions. J. Mach. Learn. Res. 11, 1803\u20131831 (2010)","journal-title":"J. Mach. Learn. Res."},{"key":"2_CR12","unstructured":"Balduzzi, D., Frean, M., Leary, L., Lewis, J., Ma, K.W.D., McWilliams, B.: The shattered gradients problem: if ResNets are the answer, then what is the question? In: International Conference on Machine Learning, pp. 342\u2013350. PMLR (2017)"},{"key":"2_CR13","unstructured":"Banerjee, S., Lavie, A.: Meteor: an automatic metric for MT evaluation with improved correlation with human judgments. In: Proceedings of the ACL Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and\/or Summarization, pp. 65\u201372 (2005)"},{"key":"2_CR14","doi-asserted-by":"crossref","unstructured":"Becking, D., Dreyer, M., Samek, W., M\u00fcller, K., Lapuschkin, S.: Ecq$$^{\\text{x}}$$: explainability-driven quantization for low-bit and sparse DNNs. In: Holzinger, A., et al. (eds.) xxAI 2020. LNAI, vol. 13200, pp. 271\u2013296. Springer, Cham (2022)","DOI":"10.1007\/978-3-031-04083-2_14"},{"issue":"7","key":"2_CR15","doi-asserted-by":"publisher","first-page":"58","DOI":"10.1145\/3448250","volume":"64","author":"Y Bengio","year":"2021","unstructured":"Bengio, Y., Lecun, Y., Hinton, G.: Deep learning for AI. Commun. ACM 64(7), 58\u201365 (2021)","journal-title":"Commun. ACM"},{"key":"2_CR16","unstructured":"Biecek, P.: DALEX: explainers for complex predictive models in R. J. Mach. Learn. Res. 19(84), 1\u20135 (2018). http:\/\/jmlr.org\/papers\/v19\/18-416.html"},{"key":"2_CR17","doi-asserted-by":"crossref","unstructured":"Biecek, P., Burzykowski, T.: Explanatory Model Analysis. Chapman and Hall\/CRC, New York (2021). https:\/\/pbiecek.github.io\/ema\/","DOI":"10.1201\/9780429027192"},{"key":"2_CR18","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"63","DOI":"10.1007\/978-3-319-44781-0_8","volume-title":"Artificial Neural Networks and Machine Learning \u2013 ICANN 2016","author":"A Binder","year":"2016","unstructured":"Binder, A., Montavon, G., Lapuschkin, S., M\u00fcller, K.-R., Samek, W.: Layer-wise relevance propagation for neural networks with local renormalization layers. In: Villa, A.E.P., Masulli, P., Pons Rivero, A.J. (eds.) ICANN 2016. LNCS, vol. 9887, pp. 63\u201371. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-44781-0_8"},{"key":"2_CR19","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"656","DOI":"10.1007\/978-3-030-32251-9_72","volume-title":"Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2019","author":"JR Clough","year":"2019","unstructured":"Clough, J.R., Oksuz, I., Puyol-Ant\u00f3n, E., Ruijsink, B., King, A.P., Schnabel, J.A.: Global and local interpretability for cardiac MRI classification. In: Shen, D., et al. (eds.) MICCAI 2019. LNCS, vol. 11767, pp. 656\u2013664. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-32251-9_72"},{"key":"2_CR20","doi-asserted-by":"crossref","unstructured":"D\u00edaz-Rodr\u00edguez, N., et al.: Explainable neural-symbolic learning (X-NeSyL) methodology to fuse deep learning representations with expert knowledge graphs: the MonuMAI cultural heritage use case. arXiv preprint arXiv:2104.11914 (2021)","DOI":"10.1016\/j.inffus.2021.09.022"},{"key":"2_CR21","doi-asserted-by":"crossref","unstructured":"Fong, R.C., Vedaldi, A.: Interpretable explanations of black boxes by meaningful perturbation. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 3429\u20133437 (2017)","DOI":"10.1109\/ICCV.2017.371"},{"key":"2_CR22","unstructured":"Frye, C., de Mijolla, D., Cowton, L., Stanley, M., Feige, I.: Shapley-based explainability on the data manifold. arXiv preprint arXiv:2006.01272 (2020)"},{"key":"2_CR23","unstructured":"Frye, C., Rowat, C., Feige, I.: Asymmetric shapley values: incorporating causal knowledge into model-agnostic explainability. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 1229\u20131239 (2020)"},{"key":"2_CR24","doi-asserted-by":"publisher","DOI":"10.1007\/978-1-4471-0211-3","volume-title":"Neural-Symbolic Learning Systems: Foundations and Applications","author":"AS d\u2019Avila Garcez","year":"2012","unstructured":"d\u2019Avila Garcez, A.S., Broda, K.B., Gabbay, D.M.: Neural-Symbolic Learning Systems: Foundations and Applications. Springer, Heidelberg (2012). https:\/\/doi.org\/10.1007\/978-1-4471-0211-3"},{"key":"2_CR25","volume-title":"Deep Learning","author":"I Goodfellow","year":"2016","unstructured":"Goodfellow, I., Bengio, Y., Courville, A.: Deep Learning. MIT Press, Cambridge (2016)"},{"key":"2_CR26","unstructured":"Gosiewska, A., Biecek, P.: iBreakDown: Uncertainty of Model Explanations for Non-additive Predictive Models. arXiv preprint arXiv:1903.11420 (2019)"},{"key":"2_CR27","unstructured":"Graesser, L., Keng, W.L.: Foundations of Deep Reinforcement Learning: Theory and Practice in Python. Addison-Wesley Professional (2019)"},{"issue":"2","key":"2_CR28","first-page":"44","volume":"40","author":"D Gunning","year":"2019","unstructured":"Gunning, D., Aha, D.W.: Darpa\u2019s explainable artificial intelligence program. AI Mag. 40(2), 44\u201358 (2019)","journal-title":"AI Mag."},{"key":"2_CR29","doi-asserted-by":"crossref","unstructured":"Halpern, J.Y., Pearl, J.: Causes and explanations: a structural-model approach. Part II: Explanations. Br. J. Philos. Sci. 56(4), 889\u2013911 (2005)","DOI":"10.1093\/bjps\/axi148"},{"key":"2_CR30","unstructured":"Hedstr\u00f6m, A., et al.: Quantus: an explainable AI toolkit for responsible evaluation of neural network explanations. arXiv preprint arXiv:2202.06861 (2022)"},{"key":"2_CR31","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1007\/978-3-319-46493-0_1","volume-title":"Computer Vision \u2013 ECCV 2016","author":"LA Hendricks","year":"2016","unstructured":"Hendricks, L.A., Akata, Z., Rohrbach, M., Donahue, J., Schiele, B., Darrell, T.: Generating visual explanations. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9908, pp. 3\u201319. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46493-0_1"},{"issue":"6","key":"2_CR32","doi-asserted-by":"publisher","first-page":"750","DOI":"10.1007\/s10458-019-09421-1","volume":"33","author":"P Hernandez-Leal","year":"2019","unstructured":"Hernandez-Leal, P., Kartal, B., Taylor, M.E.: A survey and critique of multiagent deep reinforcement learning. Auton. Agent. Multi-Agent Syst. 33(6), 750\u2013797 (2019). https:\/\/doi.org\/10.1007\/s10458-019-09421-1","journal-title":"Auton. Agent. Multi-Agent Syst."},{"key":"2_CR33","unstructured":"Hoffman, R.R., Mueller, S.T., Klein, G., Litman, J.: Metrics for explainable AI: challenges and prospects. arXiv preprint arXiv:1812.04608 (2018)"},{"key":"2_CR34","doi-asserted-by":"crossref","unstructured":"Holzinger, A., Carrington, A., Mueller, H.: Measuring the quality of explanations: the system causability scale (SCS). Comparing human and machine explanations. KI - K\u00fcnstliche Intelligenz (German Journal of Artificial intelligence), Special Issue on Interactive Machine Learning, Edited by Kristian Kersting, TU Darmstadt 34(2), 193\u2013198 (2020)","DOI":"10.1007\/s13218-020-00636-z"},{"issue":"7","key":"2_CR35","doi-asserted-by":"publisher","first-page":"28","DOI":"10.1016\/j.inffus.2021.01.008","volume":"71","author":"A Holzinger","year":"2021","unstructured":"Holzinger, A., Malle, B., Saranti, A., Pfeifer, B.: Towards multi-modal causability with graph neural networks enabling information fusion for explainable AI. Inf. Fusion 71(7), 28\u201337 (2021)","journal-title":"Inf. Fusion"},{"issue":"10","key":"2_CR36","doi-asserted-by":"publisher","first-page":"78","DOI":"10.1109\/MC.2021.3092610","volume":"54","author":"A Holzinger","year":"2021","unstructured":"Holzinger, A., Mueller, H.: Toward human-AI interfaces to support explainability and causability in medical AI. IEEE Comput. 54(10), 78\u201386 (2021)","journal-title":"IEEE Comput."},{"issue":"7","key":"2_CR37","doi-asserted-by":"publisher","first-page":"2401","DOI":"10.1007\/s10489-018-1361-5","volume":"49","author":"A Holzinger","year":"2018","unstructured":"Holzinger, A., et al.: Interactive machine learning: experimental evidence for the human in the algorithmic loop. Appl. Intell. 49(7), 2401\u20132414 (2018). https:\/\/doi.org\/10.1007\/s10489-018-1361-5","journal-title":"Appl. Intell."},{"key":"2_CR38","doi-asserted-by":"crossref","unstructured":"Huang, Q., Yamada, M., Tian, Y., Singh, D., Yin, D., Chang, Y.: GraphLIME: local interpretable model explanations for graph neural networks. arXiv preprint arXiv:2001.06216v1 (2020)","DOI":"10.1109\/TKDE.2022.3187455"},{"key":"2_CR39","unstructured":"Kauffmann, J., Esders, M., Montavon, G., Samek, W., M\u00fcller, K.R.: From clustering to cluster explanations via neural networks. arXiv preprint arXiv:1906.07633 (2019)"},{"key":"2_CR40","doi-asserted-by":"crossref","unstructured":"Kauffmann, J., M\u00fcller, K.R., Montavon, G.: Towards explaining anomalies: a deep Taylor decomposition of one-class models. Pattern Recogn. 101, 107198 (2020)","DOI":"10.1016\/j.patcog.2020.107198"},{"key":"2_CR41","unstructured":"Kim, B., et al.: Interpretability beyond feature attribution: quantitative testing with concept activation vectors (TCAV). In: International Conference on Machine Learning, pp. 2668\u20132677. PMLR (2018)"},{"key":"2_CR42","doi-asserted-by":"crossref","unstructured":"Kohlbrenner, M., Bauer, A., Nakajima, S., Binder, A., Samek, W., Lapuschkin, S.: Towards best practice in explaining neural network decisions with LRP. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1\u20137. IEEE (2020)","DOI":"10.1109\/IJCNN48605.2020.9206975"},{"key":"2_CR43","doi-asserted-by":"crossref","unstructured":"Kole, S., Bruna, J., Kutyniok, G., Levie, R., Nguyen, D.A.: A rate-distortion framework for explaining neural network decisions. In: Holzinger, A., et al. (eds.) xxAI 2020. LNAI, vol. 13200, pp. 91\u2013115. Springer, Cham (2022)","DOI":"10.1007\/978-3-031-04083-2_6"},{"key":"2_CR44","unstructured":"Koller, D., Friedman, N.: Probabilistic Graphical Models: Principles and Techniques. MIT Press (2009)"},{"issue":"1","key":"2_CR45","doi-asserted-by":"publisher","first-page":"32","DOI":"10.1007\/s11263-016-0981-7","volume":"123","author":"R Krishna","year":"2017","unstructured":"Krishna, R., et al.: Visual genome: connecting language and vision using crowdsourced dense image annotations. Int. J. Comput. Vis. 123(1), 32\u201373 (2017)","journal-title":"Int. J. Comput. Vis."},{"key":"2_CR46","unstructured":"Lapuschkin, S., Binder, A., M\u00fcller, K.R., Samek, W.: Understanding and comparing deep neural networks for age and gender classification. In: Proceedings of the IEEE International Conference on Computer Vision Workshops (ICCVW), pp. 1629\u20131638 (2017)"},{"key":"2_CR47","doi-asserted-by":"publisher","first-page":"1096","DOI":"10.1038\/s41467-019-08987-4","volume":"10","author":"S Lapuschkin","year":"2019","unstructured":"Lapuschkin, S., W\u00e4ldchen, S., Binder, A., Montavon, G., Samek, W., M\u00fcller, K.R.: Unmasking clever HANs predictors and assessing what machines really learn. Nat. Commun. 10, 1096 (2019)","journal-title":"Nat. Commun."},{"key":"2_CR48","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"740","DOI":"10.1007\/978-3-319-10602-1_48","volume-title":"Computer Vision \u2013 ECCV 2014","author":"T-Y Lin","year":"2014","unstructured":"Lin, T.-Y., et al.: Microsoft COCO: common objects in context. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol. 8693, pp. 740\u2013755. Springer, Cham (2014). https:\/\/doi.org\/10.1007\/978-3-319-10602-1_48"},{"issue":"1","key":"2_CR49","doi-asserted-by":"publisher","first-page":"56","DOI":"10.1038\/s42256-019-0138-9","volume":"2","author":"SM Lundberg","year":"2020","unstructured":"Lundberg, S.M., et al.: From local explanations to global understanding with explainable AI for trees. Nat. Mach. Intell. 2(1), 56\u201367 (2020)","journal-title":"Nat. Mach. Intell."},{"key":"2_CR50","unstructured":"Lundberg, S.M., Erion, G.G., Lee, S.I.: Consistent individualized feature attribution for tree ensembles. arXiv preprint arXiv:1802.03888 (2018)"},{"key":"2_CR51","unstructured":"Lundberg, S.M., Lee, S.I.: A unified approach to interpreting model predictions. In: Advances in Neural Information Processing Systems, vol. 30, pp. 4765\u20134774 (2017)"},{"key":"2_CR52","unstructured":"Luo, D., et al.: Parameterized explainer for graph neural network. In: Advances in Neural Information Processing Systems, vol. 33, pp. 19620\u201319631 (2020)"},{"key":"2_CR53","doi-asserted-by":"crossref","unstructured":"Madumal, P., Miller, T., Sonenberg, L., Vetere, F.: Explainable reinforcement learning through a causal lens. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, pp. 2493\u20132500 (2020)","DOI":"10.1609\/aaai.v34i03.5631"},{"key":"2_CR54","doi-asserted-by":"crossref","unstructured":"Mamalakis, A., Ebert-Uphoff, I., Barnes, E.: Explainable artificial intelligence in meteorology and climate science: Model fine-tuning, calibrating trust and learning new science. In: Holzinger, A., et al. (eds.) xxAI 2020. LNAI, vol. 13200, pp. 315\u2013339. Springer, Cham (2022)","DOI":"10.1007\/978-3-031-04083-2_16"},{"key":"2_CR55","unstructured":"Mao, J., Gan, C., Kohli, P., Tenenbaum, J.B., Wu, J.: The neuro-symbolic concept learner: interpreting scenes, words, and sentences from natural supervision. arXiv preprint arXiv:1904.12584 (2019)"},{"key":"2_CR56","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1038\/s42256-019-0114-4","volume":"1","author":"B Mittelstadt","year":"2019","unstructured":"Mittelstadt, B.: Principles alone cannot guarantee ethical AI. Nat. Mach. Intell. 1, 1\u20137 (2019)","journal-title":"Nat. Mach. Intell."},{"key":"2_CR57","unstructured":"Molnar, C., et al.: Pitfalls to avoid when interpreting machine learning models. arXiv preprint arXiv:2007.04131 (2020)"},{"key":"2_CR58","series-title":"Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence)","doi-asserted-by":"publisher","first-page":"193","DOI":"10.1007\/978-3-030-28954-6_10","volume-title":"Explainable AI: Interpreting, Explaining and Visualizing Deep Learning","author":"G Montavon","year":"2019","unstructured":"Montavon, G., Binder, A., Lapuschkin, S., Samek, W., M\u00fcller, K.-R.: Layer-wise relevance propagation: an overview. In: Samek, W., Montavon, G., Vedaldi, A., Hansen, L.K., M\u00fcller, K.-R. (eds.) Explainable AI: Interpreting, Explaining and Visualizing Deep Learning. LNCS (LNAI), vol. 11700, pp. 193\u2013209. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-28954-6_10"},{"key":"2_CR59","doi-asserted-by":"crossref","unstructured":"Montavon, G., Kauffmann, J., Samek, W., M\u00fcller, K.R.: Explaining the predictions of unsupervised learning models. In: Holzinger, A., et al. (eds.) xxAI 2020. LNAI, vol. 13200, pp. 117\u2013138. Springer, Cham (2022)","DOI":"10.1007\/978-3-031-04083-2_7"},{"key":"2_CR60","doi-asserted-by":"publisher","first-page":"211","DOI":"10.1016\/j.patcog.2016.11.008","volume":"65","author":"G Montavon","year":"2017","unstructured":"Montavon, G., Lapuschkin, S., Binder, A., Samek, W., M\u00fcller, K.R.: Explaining nonlinear classification decisions with deep Taylor decomposition. Pattern Recogn. 65, 211\u2013222 (2017)","journal-title":"Pattern Recogn."},{"key":"2_CR61","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.dsp.2017.10.011","volume":"73","author":"G Montavon","year":"2018","unstructured":"Montavon, G., Samek, W., M\u00fcller, K.R.: Methods for interpreting and understanding deep neural networks. Digit. Signal Process. 73, 1\u201315 (2018)","journal-title":"Digit. Signal Process."},{"key":"2_CR62","doi-asserted-by":"crossref","unstructured":"Morch, N.J., et al.: Visualization of neural networks using saliency maps. In: Proceedings of ICNN 1995-International Conference on Neural Networks, vol. 4, pp. 2085\u20132090 (1995)","DOI":"10.1109\/ICNN.1995.488997"},{"key":"2_CR63","doi-asserted-by":"crossref","unstructured":"O\u2019Sullivan, S., et al.: Legal, regulatory, and ethical frameworks for development of standards in artificial intelligence (AI) and autonomous robotic surgery. Int. J. Med. Robot. Comput. Assisted Surg. 15(1), e1968 (2019)","DOI":"10.1002\/rcs.1968"},{"key":"2_CR64","first-page":"13","volume-title":"Possible Minds: 25 Ways of Looking at AI","author":"J Pearl","year":"2019","unstructured":"Pearl, J.: The limitations of opaque learning machines. In: Brockman, J. (ed.) Possible Minds: 25 Ways of Looking at AI, pp. 13\u201319. Penguin, New York (2019)"},{"issue":"3","key":"2_CR65","doi-asserted-by":"publisher","first-page":"54","DOI":"10.1145\/3241036","volume":"62","author":"J Pearl","year":"2019","unstructured":"Pearl, J.: The seven tools of causal inference, with reflections on machine learning. Commun. ACM 62(3), 54\u201360 (2019)","journal-title":"Commun. ACM"},{"key":"2_CR66","volume-title":"The Book of Why","author":"J Pearl","year":"2018","unstructured":"Pearl, J., Mackenzie, D.: The Book of Why. Basic Books, New York (2018)"},{"key":"2_CR67","doi-asserted-by":"crossref","unstructured":"Ribeiro, M.T., Singh, S., Guestrin, C.: Why should I trust you?: explaining the predictions of any classifier. In: 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD 2016), pp. 1135\u20131144. ACM (2016)","DOI":"10.1145\/2939672.2939778"},{"key":"2_CR68","doi-asserted-by":"crossref","unstructured":"Ribeiro, M.T., Singh, S., Guestrin, C.: Anchors: high-precision model-agnostic explanations. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 32, no. 1 (2018)","DOI":"10.1609\/aaai.v32i1.11491"},{"issue":"5","key":"2_CR69","doi-asserted-by":"publisher","first-page":"589","DOI":"10.1109\/TKDE.2007.190734","volume":"20","author":"M Robnik-\u0160ikonja","year":"2008","unstructured":"Robnik-\u0160ikonja, M., Kononenko, I.: Explaining classifications for individual instances. IEEE Trans. Knowl. Data Eng. 20(5), 589\u2013600 (2008)","journal-title":"IEEE Trans. Knowl. Data Eng."},{"issue":"3","key":"2_CR70","doi-asserted-by":"publisher","first-page":"247","DOI":"10.1109\/JPROC.2021.3060483","volume":"109","author":"W Samek","year":"2021","unstructured":"Samek, W., Montavon, G., Lapuschkin, S., Anders, C.J., M\u00fcller, K.R.: Explaining deep neural networks and beyond: a review of methods and applications. Proc. IEEE 109(3), 247\u2013278 (2021)","journal-title":"Proc. IEEE"},{"key":"2_CR71","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"250","DOI":"10.1007\/978-3-030-29726-8_16","volume-title":"Machine Learning and Knowledge Extraction","author":"A Saranti","year":"2019","unstructured":"Saranti, A., Taraghi, B., Ebner, M., Holzinger, A.: Insights into learning competence through probabilistic graphical models. In: Holzinger, A., Kieseberg, P., Tjoa, A.M., Weippl, E. (eds.) CD-MAKE 2019. LNCS, vol. 11713, pp. 250\u2013271. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-29726-8_16"},{"key":"2_CR72","unstructured":"Schnake, T., et al.: XAI for graphs: explaining graph neural network predictions by identifying relevant walks. arXiv preprint arXiv:2006.03589 (2020)"},{"key":"2_CR73","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"209","DOI":"10.1007\/978-3-030-57321-8_12","volume-title":"Machine Learning and Knowledge Extraction","author":"D Schneeberger","year":"2020","unstructured":"Schneeberger, D., St\u00f6ger, K., Holzinger, A.: The European legal framework for medical AI. In: Holzinger, A., Kieseberg, P., Tjoa, A.M., Weippl, E. (eds.) CD-MAKE 2020. LNCS, vol. 12279, pp. 209\u2013226. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-57321-8_12"},{"key":"2_CR74","unstructured":"Schoelkopf, B.: Causality for machine learning. arXiv preprint arXiv:1911.10500 (2019)"},{"key":"2_CR75","unstructured":"Simonyan, K., Vedaldi, A., Zisserman, A.: Deep inside convolutional networks: visualising image classification models and saliency maps. arXiv preprint arXiv:1312.6034 (2013)"},{"key":"2_CR76","doi-asserted-by":"crossref","unstructured":"Stoeger, K., Schneeberger, D., Kieseberg, P., Holzinger, A.: Legal aspects of data cleansing in medical AI. Comput. Law Secur. Rev. 42, 105587 (2021)","DOI":"10.1016\/j.clsr.2021.105587"},{"key":"2_CR77","doi-asserted-by":"crossref","unstructured":"Student: The probable error of a mean. Biometrika, pp. 1\u201325 (1908)","DOI":"10.2307\/2331554"},{"key":"2_CR78","doi-asserted-by":"publisher","first-page":"141","DOI":"10.1016\/j.jneumeth.2016.10.008","volume":"274","author":"I Sturm","year":"2016","unstructured":"Sturm, I., Lapuschkin, S., Samek, W., M\u00fcller, K.R.: Interpretable deep neural networks for single-trial EEG classification. J. Neurosci. Methods 274, 141\u2013145 (2016)","journal-title":"J. Neurosci. Methods"},{"key":"2_CR79","doi-asserted-by":"publisher","first-page":"233","DOI":"10.1016\/j.inffus.2021.07.008","volume":"77","author":"J Sun","year":"2022","unstructured":"Sun, J., Lapuschkin, S., Samek, W., Binder, A.: Explain and improve: LRP-inference fine tuning for image captioning models. Inf. Fusion 77, 233\u2013246 (2022)","journal-title":"Inf. Fusion"},{"key":"2_CR80","unstructured":"Sundararajan, M., Taly, A., Yan, Q.: Axiomatic attribution for deep networks. In: Proceedings of the 34th International Conference on Machine Learning. Proceedings of Machine Learning Research, vol. 70, pp. 3319\u20133328. PMLR, 06\u201311 August 2017"},{"key":"2_CR81","volume-title":"Reinforcement Learning: An Introduction","author":"RS Sutton","year":"2018","unstructured":"Sutton, R.S., Barto, A.G.: Reinforcement Learning: An Introduction. MIT Press, Cambridge (2018)"},{"key":"2_CR82","doi-asserted-by":"crossref","unstructured":"Vedantam, R., Lawrence Zitnick, C., Parikh, D.: Cider: consensus-based image description evaluation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4566\u20134575 (2015)","DOI":"10.1109\/CVPR.2015.7299087"},{"key":"2_CR83","unstructured":"Vu, M., Thai, M.T.: PGM-explainer: probabilistic graphical model explanations for graph neural networks. In: Advances in Neural Information Processing Systems, vol. 33, pp. 12225\u201312235 (2020)"},{"key":"2_CR84","unstructured":"Wang, J., Wiens, J., Lundberg, S.: Shapley flow: a graph-based approach to interpreting model predictions. In: 24th International Conference on Artificial Intelligence and Statistics (AISTATS). Proceedings of Machine Learning Research, vol. 130, pp. 721\u2013729. PMLR (2021)"},{"issue":"3","key":"2_CR85","first-page":"229","volume":"8","author":"RJ Williams","year":"1992","unstructured":"Williams, R.J.: Simple statistical gradient-following algorithms for connectionist reinforcement learning. Mach. Learn. 8(3), 229\u2013256 (1992)","journal-title":"Mach. Learn."},{"key":"2_CR86","doi-asserted-by":"crossref","unstructured":"Xu, J., Xue, M., Picek, S.: Explainability-based backdoor attacks against graph neural networks. In: Proceedings of the 3rd ACM Workshop on Wireless Security and Machine Learning, pp. 31\u201336 (2021)","DOI":"10.1145\/3468218.3469046"},{"key":"2_CR87","doi-asserted-by":"crossref","unstructured":"Yeom, S.K., et al.: Pruning by explaining: a novel criterion for deep neural network pruning. Pattern Recogn. 115, 107899 (2021)","DOI":"10.1016\/j.patcog.2021.107899"},{"key":"2_CR88","doi-asserted-by":"crossref","unstructured":"Yuan, H., Tang, J., Hu, X., Ji, S.: XGNN: towards model-level explanations of graph neural networks. In: Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pp. 430\u2013438 (2020)","DOI":"10.1145\/3394486.3403085"},{"key":"2_CR89","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"818","DOI":"10.1007\/978-3-319-10590-1_53","volume-title":"Computer Vision \u2013 ECCV 2014","author":"MD Zeiler","year":"2014","unstructured":"Zeiler, M.D., Fergus, R.: Visualizing and understanding convolutional networks. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol. 8689, pp. 818\u2013833. Springer, Cham (2014). https:\/\/doi.org\/10.1007\/978-3-319-10590-1_53"},{"key":"2_CR90","unstructured":"Zhang, A., Lipton, Z.C., Li, M., Smola, A.J.: Dive into deep learning. Release 0.17.0, Open Source (2021)"},{"key":"2_CR91","unstructured":"Zintgraf, L.M., Cohen, T.S., Adel, T., Welling, M.: Visualizing deep neural network decisions: prediction difference analysis. arXiv preprint arXiv:1702.04595 (2017)"}],"container-title":["Lecture Notes in Computer Science","xxAI - Beyond Explainable AI"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-04083-2_2","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,9,22]],"date-time":"2024-09-22T10:07:10Z","timestamp":1726999630000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-04083-2_2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031040825","9783031040832"],"references-count":91,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-04083-2_2","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"17 April 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"xxAI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Workshop on Extending Explainable AI Beyond Deep Models and Classifiers","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Vienna","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Austria","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2020","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 July 2020","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 July 2020","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"xxai2020","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/human-centered.ai\/xxai-icml-2020\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}