{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T13:45:31Z","timestamp":1776087931256,"version":"3.50.1"},"publisher-location":"Cham","reference-count":97,"publisher":"Springer International Publishing","isbn-type":[{"value":"9783030289539","type":"print"},{"value":"9783030289546","type":"electronic"}],"license":[{"start":{"date-parts":[[2019,1,1]],"date-time":"2019-01-01T00:00:00Z","timestamp":1546300800000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2019]]},"DOI":"10.1007\/978-3-030-28954-6_1","type":"book-chapter","created":{"date-parts":[[2019,9,9]],"date-time":"2019-09-09T23:08:50Z","timestamp":1568070530000},"page":"5-22","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":396,"title":["Towards Explainable Artificial Intelligence"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-6283-3265","authenticated-orcid":false,"given":"Wojciech","family":"Samek","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3861-7685","authenticated-orcid":false,"given":"Klaus-Robert","family":"M\u00fcller","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2019,9,10]]},"reference":[{"issue":"93","key":"1_CR1","first-page":"1","volume":"20","author":"M Alber","year":"2019","unstructured":"Alber, M., et al.: iNNvestigate neural networks!. J. Mach. Learn. Res. 20(93), 1\u20138 (2019)","journal-title":"J. Mach. Learn. Res."},{"key":"1_CR2","doi-asserted-by":"crossref","first-page":"169","DOI":"10.1007\/978-3-030-28954-6_9","volume-title":"Explainable AI: Interpreting, Explaining and Visualizing Deep Learning","author":"Marco Ancona","year":"2019","unstructured":"Ancona, M., Ceolini, E., \u00d6ztireli, C., Gross, M.: Gradient-based attribution methods. In: Samek, W., Montavon, G., Vedaldi, A., Hansen, L.K., M\u00fcller, K.-R. (eds.) Explainable AI. LNCS, vol. 11700, pp. 169\u2013191. Springer, Cham (2019)"},{"issue":"2","key":"1_CR3","doi-asserted-by":"crossref","first-page":"8","DOI":"10.1145\/2089125.2089128","volume":"44","author":"P Antunes","year":"2012","unstructured":"Antunes, P., Herskovic, V., Ochoa, S.F., Pino, J.A.: Structuring dimensions for collaborative systems evaluation. ACM Comput. Surv. (CSUR) 44(2), 8 (2012)","journal-title":"ACM Comput. Surv. (CSUR)"},{"key":"1_CR4","unstructured":"Arjona-Medina, J.A., Gillhofer, M., Widrich, M., Unterthiner, T., Hochreiter, S.: RUDDER: return decomposition for delayed rewards. arXiv preprint arXiv:1806.07857 (2018)"},{"key":"1_CR5","doi-asserted-by":"crossref","first-page":"211","DOI":"10.1007\/978-3-030-28954-6_11","volume-title":"Explainable AI: Interpreting, Explaining and Visualizing Deep Learning","author":"Leila Arras","year":"2019","unstructured":"Arras, L., et al.: Explaining and interpreting LSTMs. In: Samek, W., Montavon, G., Vedaldi, A., Hansen, L.K., M\u00fcller, K.-R. (eds.) Explainable AI. LNCS, vol. 11700, pp. 211\u2013238. Springer, Cham (2019)"},{"issue":"8","key":"1_CR6","doi-asserted-by":"crossref","first-page":"e0181142","DOI":"10.1371\/journal.pone.0181142","volume":"12","author":"L Arras","year":"2017","unstructured":"Arras, L., Horn, F., Montavon, G., M\u00fcller, K.R., Samek, W.: What is relevant in a text document?: An interpretable machine learning approach. PLoS ONE 12(8), e0181142 (2017)","journal-title":"PLoS ONE"},{"key":"1_CR7","doi-asserted-by":"crossref","unstructured":"Arras, L., Montavon, G., M\u00fcller, K.R., Samek, W.: Explaining recurrent neural network predictions in sentiment analysis. In: EMNLP 2017 Workshop on Computational Approaches to Subjectivity, Sentiment & Social Media Analysis (WASSA), pp. 159\u2013168 (2017)","DOI":"10.18653\/v1\/W17-5221"},{"key":"1_CR8","doi-asserted-by":"crossref","unstructured":"Arras, L., Osman, A., M\u00fcller, K.R., Samek, W.: Evaluating recurrent neural network explanations. In: ACL 2019 Workshop on BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP (2019)","DOI":"10.18653\/v1\/W19-4813"},{"issue":"7","key":"1_CR9","doi-asserted-by":"crossref","first-page":"e0130140","DOI":"10.1371\/journal.pone.0130140","volume":"10","author":"S Bach","year":"2015","unstructured":"Bach, S., Binder, A., Montavon, G., Klauschen, F., M\u00fcller, K.R., Samek, W.: On pixel-wise explanations for non-linear classifier decisions by layer-wise relevance propagation. PLoS ONE 10(7), e0130140 (2015)","journal-title":"PLoS ONE"},{"key":"1_CR10","first-page":"1803","volume":"11","author":"D Baehrens","year":"2010","unstructured":"Baehrens, D., Schroeter, T., Harmeling, S., Kawanabe, M., Hansen, K., M\u00fcller, K.R.: How to explain individual classification decisions. J. Mach. Learn. Res. 11, 1803\u20131831 (2010)","journal-title":"J. Mach. Learn. Res."},{"key":"1_CR11","unstructured":"Bahdanau, D., Cho, K., Bengio, Y.: Neural machine translation by jointly learning to align and translate. In: International Conference on Learning Representations (ICLR) (2015)"},{"key":"1_CR12","doi-asserted-by":"crossref","unstructured":"Bau, D., Zhou, B., Khosla, A., Oliva, A., Torralba, A.: Network dissection: quantifying interpretability of deep visual representations. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6541\u20136549 (2017)","DOI":"10.1109\/CVPR.2017.354"},{"key":"1_CR13","series-title":"Lecture Notes in Electrical Engineering","doi-asserted-by":"publisher","first-page":"913","DOI":"10.1007\/978-981-10-0557-2_87","volume-title":"Information Science and Applications (ICISA) 2016","author":"A Binder","year":"2016","unstructured":"Binder, A., Bach, S., Montavon, G., M\u00fcller, K.-R., Samek, W.: Layer-wise relevance propagation for deep neural network architectures. Information Science and Applications (ICISA) 2016. LNEE, vol. 376, pp. 913\u2013922. Springer, Singapore (2016). https:\/\/doi.org\/10.1007\/978-981-10-0557-2_87"},{"key":"1_CR14","unstructured":"Binder, A., et al.: Towards computational fluorescence microscopy: machine learning-based integrated prediction of morphological and molecular tumor profiles. arXiv preprint arXiv:1805.11178 (2018)"},{"issue":"1","key":"1_CR15","doi-asserted-by":"crossref","first-page":"3887","DOI":"10.1038\/s41467-018-06169-2","volume":"9","author":"S Chmiela","year":"2018","unstructured":"Chmiela, S., Sauceda, H.E., M\u00fcller, K.R., Tkatchenko, A.: Towards exact molecular dynamics simulations with machine-learned force fields. Nat. Commun. 9(1), 3887 (2018)","journal-title":"Nat. Commun."},{"key":"1_CR16","doi-asserted-by":"crossref","unstructured":"Cire\u015fan, D., Meier, U., Masci, J., Schmidhuber, J.: A committee of neural networks for traffic sign classification. In: International Joint Conference on Neural Networks (IJCNN), pp. 1918\u20131921 (2011)","DOI":"10.1109\/IJCNN.2011.6033458"},{"key":"1_CR17","doi-asserted-by":"crossref","unstructured":"Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Fei-Fei, L.: Imagenet: a large-scale hierarchical image database. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 248\u2013255 (2009)","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"1_CR18","unstructured":"Doshi-Velez, F., Kim, B.: Towards a rigorous science of interpretable machine learning. arXiv preprint arXiv:1702.08608 (2017)"},{"key":"1_CR19","doi-asserted-by":"crossref","unstructured":"Doshi-Velez, F., et al.: Accountability of AI under the law: the role of explanation. arXiv preprint arXiv:1711.01134 (2017)","DOI":"10.2139\/ssrn.3064761"},{"key":"1_CR20","doi-asserted-by":"crossref","unstructured":"Eitel, F., et al.: Uncovering convolutional neural network decisions for diagnosing multiple sclerosis on conventional MRI using layer-wise relevance propagation. arXiv preprint arXiv:1904.08771 (2019)","DOI":"10.1016\/j.nicl.2019.102003"},{"key":"1_CR21","unstructured":"European Commission\u2019s High-Level Expert Group: Draft ethics guidelines for trustworthy AI. European Commission (2019)"},{"issue":"1","key":"1_CR22","doi-asserted-by":"crossref","first-page":"98","DOI":"10.1007\/s11263-014-0733-5","volume":"111","author":"M Everingham","year":"2015","unstructured":"Everingham, M., Eslami, S.A., Van Gool, L., Williams, C.K., Winn, J., Zisserman, A.: The PASCAL visual object classes challenge: a retrospective. Int. J. Comput. Vision 111(1), 98\u2013136 (2015)","journal-title":"Int. J. Comput. Vision"},{"issue":"2","key":"1_CR23","doi-asserted-by":"crossref","first-page":"303","DOI":"10.1007\/s11263-009-0275-4","volume":"88","author":"M Everingham","year":"2010","unstructured":"Everingham, M., Van Gool, L., Williams, C.K., Winn, J., Zisserman, A.: The pascal visual object classes (VOC) challenge. Int. J. Comput. Vision 88(2), 303\u2013338 (2010)","journal-title":"Int. J. Comput. Vision"},{"key":"1_CR24","unstructured":"Eykholt, K., et al.: Robust physical-world attacks on deep learning models. arXiv preprint arXiv:1707.08945 (2017)"},{"key":"1_CR25","doi-asserted-by":"crossref","unstructured":"Fong, R.C., Vedaldi, A.: Interpretable explanations of black boxes by meaningful perturbation. In: IEEE International Conference on Computer Vision (CVPR), pp. 3429\u20133437 (2017)","DOI":"10.1109\/ICCV.2017.371"},{"key":"1_CR26","doi-asserted-by":"crossref","first-page":"149","DOI":"10.1007\/978-3-030-28954-6_8","volume-title":"Explainable AI: Interpreting, Explaining and Visualizing Deep Learning","author":"Ruth Fong","year":"2019","unstructured":"Fong, R., Vedaldi, A.: Explanations for attributing deep neural network predictions. In: Samek, W., Montavon, G., Vedaldi, A., Hansen, L.K., M\u00fcller, K.-R. (eds.) Explainable AI. LNCS, vol. 11700, pp. 149\u2013167. Springer, Cham (2019)"},{"issue":"3","key":"1_CR27","doi-asserted-by":"crossref","first-page":"50","DOI":"10.1609\/aimag.v38i3.2741","volume":"38","author":"B Goodman","year":"2017","unstructured":"Goodman, B., Flaxman, S.: European union regulations on algorithmic decision-making and a \u201cright to explanation\u201d. AI Mag. 38(3), 50\u201357 (2017)","journal-title":"AI Mag."},{"key":"1_CR28","doi-asserted-by":"crossref","unstructured":"Hajian, S., Bonchi, F., Castillo, C.: Algorithmic bias: from discrimination discovery to fairness-aware data mining. In: 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 2125\u20132126 (2016)","DOI":"10.1145\/2939672.2945386"},{"key":"1_CR29","unstructured":"Han, S., Pool, J., Tran, J., Dally, W.: Learning both weights and connections for efficient neural network. In: Advances in Neural Information Processing Systems (NIPS), pp. 1135\u20131143 (2015)"},{"key":"1_CR30","doi-asserted-by":"crossref","DOI":"10.4324\/9781410605481","volume-title":"Human Communication Theory and Research: Concepts, Contexts, and Challenges","author":"RL Heath","year":"2013","unstructured":"Heath, R.L., Bryant, J.: Human Communication Theory and Research: Concepts, Contexts, and Challenges. Routledge, New York (2013)"},{"key":"1_CR31","doi-asserted-by":"crossref","first-page":"285","DOI":"10.1007\/978-3-030-28954-6_15","volume-title":"Explainable AI: Interpreting, Explaining and Visualizing Deep Learning","author":"Markus Hofmarcher","year":"2019","unstructured":"Hofmarcher, M., Unterthiner, T., Arjona-Medina, J., Klambauer, G., Hochreiter, S., Nessler, B.: Visual scene understanding for autonomous driving using semantic segmentation. In: Samek, W., Montavon, G., Vedaldi, A., Hansen, L.K., M\u00fcller, K.-R. (eds.) Explainable AI. LNCS, vol. 11700, pp. 285\u2013296. Springer, Cham (2019)"},{"key":"1_CR32","doi-asserted-by":"crossref","first-page":"e1312","DOI":"10.1002\/widm.1312","volume":"9","author":"A Holzinger","year":"2019","unstructured":"Holzinger, A., Langs, G., Denk, H., Zatloukal, K., M\u00fcller, H.: Causability and explainabilty of artificial intelligence in medicine. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 9, e1312 (2019)","journal-title":"Wiley Interdiscip. Rev. Data Min. Knowl. Discov."},{"key":"1_CR33","doi-asserted-by":"crossref","first-page":"2391","DOI":"10.1038\/s41598-019-38748-8","volume":"9","author":"F Horst","year":"2019","unstructured":"Horst, F., Lapuschkin, S., Samek, W., M\u00fcller, K.R., Sch\u00f6llhorn, W.I.: Explaining the unique nature of individual gait patterns with deep learning. Sci. Rep. 9, 2391 (2019)","journal-title":"Sci. Rep."},{"key":"1_CR34","doi-asserted-by":"crossref","unstructured":"Karpathy, A., Toderici, G., Shetty, S., Leung, T., Sukthankar, R., Fei-Fei, L.: Large-scale video classification with convolutional neural networks. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1725\u20131732 (2014)","DOI":"10.1109\/CVPR.2014.223"},{"key":"1_CR35","unstructured":"Kauffmann, J., M\u00fcller, K.R., Montavon, G.: Towards explaining anomalies: a deep Taylor decomposition of one-class models. arXiv preprint arXiv:1805.06230 (2018)"},{"key":"1_CR36","unstructured":"Kauffmann, J., Esders, M., Montavon, G., Samek, W., M\u00fcller, K.R.: From clustering to cluster explanations via neural networks. arXiv preprint arXiv:1906.07633 (2019)"},{"key":"1_CR37","unstructured":"Khanna, R., Kim, B., Ghosh, J., Koyejo, O.: Interpreting black box predictions using fisher kernels. arXiv preprint arXiv:1810.10118 (2018)"},{"key":"1_CR38","unstructured":"Kim, B., et al.: Interpretability beyond feature attribution: quantitative testing with concept activation vectors (TCAV). In: International Conference on Machine Learning (ICML), pp. 2673\u20132682 (2018)"},{"key":"1_CR39","unstructured":"Kindermans, P.J., et al.: Learning how to explain neural networks: patternnet and patternattribution. In: International Conference on Learning Representations (ICLR) (2018)"},{"issue":"2","key":"1_CR40","doi-asserted-by":"crossref","first-page":"151","DOI":"10.1016\/j.semcancer.2018.07.001","volume":"52","author":"F Klauschen","year":"2018","unstructured":"Klauschen, F., et al.: Scoring of tumor-infiltrating lymphocytes: from visual estimation to machine learning. Semin. Cancer Biol. 52(2), 151\u2013157 (2018)","journal-title":"Semin. Cancer Biol."},{"key":"1_CR41","unstructured":"Koh, P.W., Liang, P.: Understanding black-box predictions via influence functions. In: International Conference on Machine Learning (ICML), pp. 1885\u20131894 (2017)"},{"issue":"10","key":"1_CR42","doi-asserted-by":"crossref","first-page":"3863","DOI":"10.1073\/pnas.0600244103","volume":"103","author":"N Kriegeskorte","year":"2006","unstructured":"Kriegeskorte, N., Goebel, R., Bandettini, P.: Information-based functional brain mapping. Proc. Nat. Acad. Sci. 103(10), 3863\u20133868 (2006)","journal-title":"Proc. Nat. Acad. Sci."},{"key":"1_CR43","unstructured":"Lage, I., et al.: An evaluation of the human-interpretability of explanation. arXiv preprint arXiv:1902.00006 (2019)"},{"key":"1_CR44","doi-asserted-by":"crossref","unstructured":"Lapuschkin, S., Binder, A., Montavon, G., M\u00fcller, K.R., Samek, W.: Analyzing classifiers: fisher vectors and deep neural networks. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2912\u20132920 (2016)","DOI":"10.1109\/CVPR.2016.318"},{"key":"1_CR45","unstructured":"Lapuschkin, S.: Opening the machine learning black box with layer-wise relevance propagation. Ph.D. thesis, Technische Universit\u00e4t Berlin (2019)"},{"key":"1_CR46","doi-asserted-by":"crossref","first-page":"1096","DOI":"10.1038\/s41467-019-08987-4","volume":"10","author":"S Lapuschkin","year":"2019","unstructured":"Lapuschkin, S., W\u00e4ldchen, S., Binder, A., Montavon, G., Samek, W., M\u00fcller, K.R.: Unmasking clever hans predictors and assessing what machines really learn. Nat. Commun. 10, 1096 (2019)","journal-title":"Nat. Commun."},{"issue":"7553","key":"1_CR47","doi-asserted-by":"crossref","first-page":"436","DOI":"10.1038\/nature14539","volume":"521","author":"Y LeCun","year":"2015","unstructured":"LeCun, Y., Bengio, Y., Hinton, G.: Deep learning. Nature 521(7553), 436\u2013444 (2015)","journal-title":"Nature"},{"key":"1_CR48","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"9","DOI":"10.1007\/978-3-642-35289-8_3","volume-title":"Neural Networks: Tricks of the Trade","author":"YA LeCun","year":"2012","unstructured":"LeCun, Y.A., Bottou, L., Orr, G.B., M\u00fcller, K.-R.: Efficient backprop. In: Montavon, G., Orr, G.B., M\u00fcller, K.-R. (eds.) Neural Networks: Tricks of the Trade. LNCS, vol. 7700, pp. 9\u201348. Springer, Heidelberg (2012). https:\/\/doi.org\/10.1007\/978-3-642-35289-8_3"},{"issue":"2","key":"1_CR49","doi-asserted-by":"crossref","first-page":"387","DOI":"10.1016\/j.neuroimage.2010.11.004","volume":"56","author":"S Lemm","year":"2011","unstructured":"Lemm, S., Blankertz, B., Dickhaus, T., M\u00fcller, K.R.: Introduction to machine learning for brain imaging. Neuroimage 56(2), 387\u2013399 (2011)","journal-title":"Neuroimage"},{"key":"1_CR50","unstructured":"Li, J., Monroe, W., Jurafsky, D.: Understanding neural networks through representation erasure. arXiv preprint arXiv:1612.08220 (2016)"},{"issue":"6","key":"1_CR51","doi-asserted-by":"crossref","first-page":"321","DOI":"10.1038\/nrg3920","volume":"16","author":"MW Libbrecht","year":"2015","unstructured":"Libbrecht, M.W., Noble, W.S.: Machine learning applications in genetics and genomics. Nat. Rev. Genet. 16(6), 321 (2015)","journal-title":"Nat. Rev. Genet."},{"issue":"2","key":"1_CR52","doi-asserted-by":"crossref","first-page":"39","DOI":"10.1109\/MM.2008.31","volume":"28","author":"E Lindholm","year":"2008","unstructured":"Lindholm, E., Nickolls, J., Oberman, S., Montrym, J.: NVIDIA tesla: a unified graphics and computing architecture. IEEE Micro 28(2), 39\u201355 (2008)","journal-title":"IEEE Micro"},{"key":"1_CR53","doi-asserted-by":"crossref","unstructured":"Lu, C., Tang, X.: Surpassing human-level face verification performance on LFW with GaussianFace. In: 29th AAAI Conference on Artificial Intelligence, pp. 3811\u20133819 (2015)","DOI":"10.1609\/aaai.v29i1.9797"},{"key":"1_CR54","unstructured":"Lundberg, S.M., Lee, S.I.: A unified approach to interpreting model predictions. In: Advances in Neural Information Processing Systems (NIPS), pp. 4765\u20134774 (2017)"},{"key":"1_CR55","unstructured":"Madry, A., Makelov, A., Schmidt, L., Tsipras, D., Vladu, A.: Towards deep learning models resistant to adversarial attacks. In: International Conference on Learning Representations (ICLR) (2018)"},{"issue":"7540","key":"1_CR56","doi-asserted-by":"crossref","first-page":"529","DOI":"10.1038\/nature14236","volume":"518","author":"V Mnih","year":"2015","unstructured":"Mnih, V., et al.: Human-level control through deep reinforcement learning. Nature 518(7540), 529\u2013533 (2015)","journal-title":"Nature"},{"key":"1_CR57","doi-asserted-by":"crossref","first-page":"253","DOI":"10.1007\/978-3-030-28954-6_13","volume-title":"Explainable AI: Interpreting, Explaining and Visualizing Deep Learning","author":"Gr\u00e9goire Montavon","year":"2019","unstructured":"Montavon, G.: Gradient-based vs. propagation-based explanations: an axiomatic comparison. In: Samek, W., Montavon, G., Vedaldi, A., Hansen, L.K., M\u00fcller, K.-R. (eds.) Explainable AI. LNCS, vol. 11700, pp. 253\u2013265. Springer, Cham (2019)"},{"key":"1_CR58","doi-asserted-by":"crossref","first-page":"193","DOI":"10.1007\/978-3-030-28954-6_10","volume-title":"Explainable AI: Interpreting, Explaining and Visualizing Deep Learning","author":"Gr\u00e9goire Montavon","year":"2019","unstructured":"Montavon, G., Binder, A., Lapuschkin, S., Samek, W., M\u00fcller, K.-R.: Layer-wise relevance propagation: an overview. In: Samek, W., Montavon, G., Vedaldi, A., Hansen, L.K., M\u00fcller, K.-R. (eds.) Explainable AI. LNCS, vol. 11700, pp. 193\u2013209. Springer, Cham (2019)"},{"key":"1_CR59","doi-asserted-by":"crossref","first-page":"211","DOI":"10.1016\/j.patcog.2016.11.008","volume":"65","author":"G Montavon","year":"2017","unstructured":"Montavon, G., Lapuschkin, S., Binder, A., Samek, W., M\u00fcller, K.R.: Explaining nonlinear classification decisions with deep Taylor decomposition. Pattern Recogn. 65, 211\u2013222 (2017)","journal-title":"Pattern Recogn."},{"key":"1_CR60","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1016\/j.dsp.2017.10.011","volume":"73","author":"G Montavon","year":"2018","unstructured":"Montavon, G., Samek, W., M\u00fcller, K.R.: Methods for interpreting and understanding deep neural networks. Digit. Signal Process. 73, 1\u201315 (2018)","journal-title":"Digit. Signal Process."},{"issue":"6337","key":"1_CR61","doi-asserted-by":"crossref","first-page":"508","DOI":"10.1126\/science.aam6960","volume":"356","author":"M Morav\u010d\u00edk","year":"2017","unstructured":"Morav\u010d\u00edk, M., et al.: Deepstack: expert-level artificial intelligence in heads-up no-limit poker. Science 356(6337), 508\u2013513 (2017)","journal-title":"Science"},{"key":"1_CR62","unstructured":"Morch, N., et al.: Visualization of neural networks using saliency maps. In: International Conference on Neural Networks (ICNN), vol. 4, pp. 2085\u20132090 (1995)"},{"key":"1_CR63","unstructured":"Mordvintsev, A., Olah, C., Tyka, M.: Inceptionism: going deeper into neural networks (2015)"},{"key":"1_CR64","unstructured":"Nguyen, A., Dosovitskiy, A., Yosinski, J., Brox, T., Clune, J.: Synthesizing the preferred inputs for neurons in neural networks via deep generator networks. In: Advances in Neural Information Processing Systems (NIPS), pp. 3387\u20133395 (2016)"},{"key":"1_CR65","doi-asserted-by":"crossref","first-page":"55","DOI":"10.1007\/978-3-030-28954-6_4","volume-title":"Explainable AI: Interpreting, Explaining and Visualizing Deep Learning","author":"Anh Nguyen","year":"2019","unstructured":"Nguyen, A., Yosinski, J., Clune, J.: Understanding neural networks via feature visualization: a survey. In: Samek, W., Montavon, G., Vedaldi, A., Hansen, L.K., M\u00fcller, K.-R. (eds.) Explainable AI. LNCS, vol. 11700, pp. 55\u201376. Springer, Cham (2019)"},{"key":"1_CR66","doi-asserted-by":"crossref","unstructured":"Nguyen, D.: Comparing automatic and human evaluation of local explanations for text classification. In: Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT), pp. 1069\u20131078 (2018)","DOI":"10.18653\/v1\/N18-1097"},{"issue":"2","key":"1_CR67","doi-asserted-by":"crossref","first-page":"244","DOI":"10.1007\/s40846-017-0297-2","volume":"38","author":"A Phinyomark","year":"2018","unstructured":"Phinyomark, A., Petri, G., Ib\u00e1\u00f1ez-Marcelo, E., Osis, S.T., Ferber, R.: Analysis of big data in gait biomechanics: current trends and future directions. J. Med. Biol. Eng. 38(2), 244\u2013260 (2018)","journal-title":"J. Med. Biol. Eng."},{"key":"1_CR68","doi-asserted-by":"crossref","first-page":"2810","DOI":"10.1038\/srep02810","volume":"3","author":"G Pilania","year":"2013","unstructured":"Pilania, G., Wang, C., Jiang, X., Rajasekaran, S., Ramprasad, R.: Accelerating materials property predictions using machine learning. Sci. Rep. 3, 2810 (2013)","journal-title":"Sci. Rep."},{"key":"1_CR69","doi-asserted-by":"crossref","unstructured":"Poerner, N., Roth, B., Sch\u00fctze, H.: Evaluating neural network explanation methods using hybrid documents and morphosyntactic agreement. In: 56th Annual Meeting of the Association for Computational Linguistics (ACL), pp. 340\u2013350 (2018)","DOI":"10.18653\/v1\/P18-1032"},{"key":"1_CR70","doi-asserted-by":"crossref","first-page":"331","DOI":"10.1007\/978-3-030-28954-6_18","volume-title":"Explainable AI: Interpreting, Explaining and Visualizing Deep Learning","author":"Kristina Preuer","year":"2019","unstructured":"Preuer, K., Klambauer, G., Rippmann, F., Hochreiter, S., Unterthiner, T.: Interpretable deep learning in drug discovery. In: Samek, W., Montavon, G., Vedaldi, A., Hansen, L.K., M\u00fcller, K.-R. (eds.) Explainable AI. LNCS, vol. 11700, pp. 331\u2013345. Springer, Cham (2019)"},{"key":"1_CR71","doi-asserted-by":"crossref","unstructured":"Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: unified, real-time object detection. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 779\u2013788 (2016)","DOI":"10.1109\/CVPR.2016.91"},{"key":"1_CR72","doi-asserted-by":"crossref","unstructured":"Reyes, E., et al.: Enhanced rotational invariant convolutional neural network for supernovae detection. In: International Joint Conference on Neural Networks (IJCNN), pp. 1\u20138 (2018)","DOI":"10.1109\/IJCNN.2018.8489627"},{"key":"1_CR73","doi-asserted-by":"crossref","unstructured":"Ribeiro, M.T., Singh, S., Guestrin, C.: Why should I trust you?: explaining the predictions of any classifier. In: ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 1135\u20131144 (2016)","DOI":"10.18653\/v1\/N16-3020"},{"key":"1_CR74","doi-asserted-by":"crossref","unstructured":"Ross, A.S., Hughes, M.C., Doshi-Velez, F.: Right for the right reasons: training differentiable models by constraining their explanations. In: 26th International Joint Conferences on Artificial Intelligence (IJCAI), pp. 2662\u20132670 (2017)","DOI":"10.24963\/ijcai.2017\/371"},{"issue":"11","key":"1_CR75","doi-asserted-by":"crossref","first-page":"2660","DOI":"10.1109\/TNNLS.2016.2599820","volume":"28","author":"W Samek","year":"2017","unstructured":"Samek, W., Binder, A., Montavon, G., Lapuschkin, S., M\u00fcller, K.R.: Evaluating the visualization of what a deep neural network has learned. IEEE Trans. Neural Netw. Learn. Syst. 28(11), 2660\u20132673 (2017)","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"issue":"1","key":"1_CR76","first-page":"39","volume":"1","author":"W Samek","year":"2018","unstructured":"Samek, W., Wiegand, T., M\u00fcller, K.R.: Explainable artificial intelligence: understanding, visualizing and interpreting deep learning models. ITU J. ICT Discov. 1(1), 39\u201348 (2018). Special Issue 1 - The Impact of Artificial Intelligence (AI) on Communication Networks and Services","journal-title":"ITU J. ICT Discov."},{"issue":"3","key":"1_CR77","doi-asserted-by":"crossref","first-page":"222","DOI":"10.1007\/s11263-013-0636-x","volume":"105","author":"J S\u00e1nchez","year":"2013","unstructured":"S\u00e1nchez, J., Perronnin, F., Mensink, T., Verbeek, J.J.: Image classification with the fisher vector: theory and practice. Int. J. Comput. Vision 105(3), 222\u2013245 (2013)","journal-title":"Int. J. Comput. Vision"},{"key":"1_CR78","doi-asserted-by":"crossref","first-page":"13890","DOI":"10.1038\/ncomms13890","volume":"8","author":"KT Sch\u00fctt","year":"2017","unstructured":"Sch\u00fctt, K.T., Arbabzadah, F., Chmiela, S., M\u00fcller, K.R., Tkatchenko, A.: Quantum-chemical insights from deep tensor neural networks. Nat. Commun. 8, 13890 (2017)","journal-title":"Nat. Commun."},{"key":"1_CR79","doi-asserted-by":"crossref","unstructured":"Selvaraju, R.R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., Batra, D.: Grad-CAM: visual explanations from deep networks via gradient-based localization. In: IEEE International Conference on Computer Vision (CVPR), pp. 618\u2013626 (2017)","DOI":"10.1109\/ICCV.2017.74"},{"issue":"28","key":"1_CR80","first-page":"307","volume":"2","author":"LS Shapley","year":"1953","unstructured":"Shapley, L.S.: A value for n-person games. Contrib. Theory Games 2(28), 307\u2013317 (1953)","journal-title":"Contrib. Theory Games"},{"key":"1_CR81","unstructured":"Shrikumar, A., Greenside, P., Kundaje, A.: Learning important features through propagating activation differences. arXiv preprint arXiv:1704.02685 (2017)"},{"issue":"7587","key":"1_CR82","doi-asserted-by":"crossref","first-page":"484","DOI":"10.1038\/nature16961","volume":"529","author":"D Silver","year":"2016","unstructured":"Silver, D., et al.: Mastering the game of Go with deep neural networks and tree search. Nature 529(7587), 484\u2013489 (2016)","journal-title":"Nature"},{"issue":"7676","key":"1_CR83","doi-asserted-by":"crossref","first-page":"354","DOI":"10.1038\/nature24270","volume":"550","author":"D Silver","year":"2017","unstructured":"Silver, D., et al.: Mastering the game of Go without human knowledge. Nature 550(7676), 354\u2013359 (2017)","journal-title":"Nature"},{"key":"1_CR84","unstructured":"Simonyan, K., Vedaldi, A., Zisserman, A.: Deep inside convolutional networks: visualising image classification models and saliency maps. In: ICLR Workshop (2014)"},{"key":"1_CR85","unstructured":"Smilkov, D., Thorat, N., Kim, B., Vi\u00e9gas, F., Wattenberg, M.: SmoothGrad: removing noise by adding noise. arXiv preprint arXiv:1706.03825 (2017)"},{"key":"1_CR86","unstructured":"Springenberg, J.T., Dosovitskiy, A., Brox, T., Riedmiller, M.: Striving for simplicity: the all convolutional net. In: ICLR Workshop (2015)"},{"key":"1_CR87","doi-asserted-by":"crossref","first-page":"141","DOI":"10.1016\/j.jneumeth.2016.10.008","volume":"274","author":"I Sturm","year":"2016","unstructured":"Sturm, I., Lapuschkin, S., Samek, W., M\u00fcller, K.R.: Interpretable deep neural networks for single-trial EEG classification. J. Neurosci. Methods 274, 141\u2013145 (2016)","journal-title":"J. Neurosci. Methods"},{"key":"1_CR88","unstructured":"Sundararajan, M., Taly, A., Yan, Q.: Axiomatic attribution for deep networks. In: International Conference on Machine Learning (ICML), pp. 3319\u20133328 (2017)"},{"key":"1_CR89","doi-asserted-by":"crossref","unstructured":"Thomas, A.W., Heekeren, H.R., M\u00fcller, K.R., Samek, W.: Analyzing neuroimaging data through recurrent deep learning models. arXiv preprint arXiv:1810.09945 (2018)","DOI":"10.3389\/fnins.2019.01321"},{"key":"1_CR90","unstructured":"Van Den Oord, A., et al.: Wavenet: a generative model for raw audio. arXiv preprint arXiv:1609.03499 (2016)"},{"key":"1_CR91","unstructured":"Weller, A.: Transparency: motivations and challenges. In: Samek, W., Montavon, G., Vedaldi, A., Hansen, L.K., M\u00fcller, K.-R. (eds.) Explainable AI. LNCS, vol. 11700, pp. 23\u201340. Springer, Cham (2019)"},{"issue":"8","key":"1_CR92","doi-asserted-by":"crossref","first-page":"080602","DOI":"10.1103\/PhysRevLett.122.080602","volume":"122","author":"D Wu","year":"2019","unstructured":"Wu, D., Wang, L., Zhang, P.: Solving statistical mechanics using variational autoregressive networks. Phys. Rev. Lett. 122(8), 080602 (2019)","journal-title":"Phys. Rev. Lett."},{"key":"1_CR93","unstructured":"Yosinski, J., Clune, J., Nguyen, A., Fuchs, T., Lipson, H.: Understanding neural networks through deep visualization. arXiv preprint arXiv:1506.06579 (2015)"},{"key":"1_CR94","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"818","DOI":"10.1007\/978-3-319-10590-1_53","volume-title":"Computer Vision \u2013 ECCV 2014","author":"MD Zeiler","year":"2014","unstructured":"Zeiler, M.D., Fergus, R.: Visualizing and understanding convolutional networks. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol. 8689, pp. 818\u2013833. Springer, Cham (2014). https:\/\/doi.org\/10.1007\/978-3-319-10590-1_53"},{"key":"1_CR95","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"543","DOI":"10.1007\/978-3-319-46493-0_33","volume-title":"Computer Vision \u2013 ECCV 2016","author":"J Zhang","year":"2016","unstructured":"Zhang, J., Lin, Z., Brandt, J., Shen, X., Sclaroff, S.: Top-down neural attention by excitation backprop. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9908, pp. 543\u2013559. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46493-0_33"},{"key":"1_CR96","doi-asserted-by":"crossref","first-page":"243","DOI":"10.1007\/978-3-030-28954-6_12","volume-title":"Explainable AI: Interpreting, Explaining and Visualizing Deep Learning","author":"Bolei Zhou","year":"2019","unstructured":"Zhou, B., Bau, D., Oliva, A., Torralba, A.: Comparing the interpretability of deep networks via network dissection. In: Samek, W., Montavon, G., Vedaldi, A., Hansen, L.K., M\u00fcller, K.-R. (eds.) Explainable AI. LNCS, vol. 11700, pp. 243\u2013252. Springer, Cham (2019)"},{"key":"1_CR97","unstructured":"Zintgraf, L.M., Cohen, T.S., Adel, T., Welling, M.: Visualizing deep neural network decisions: prediction difference analysis. In: International Conference on Learning Representations (ICLR) (2017)"}],"container-title":["Lecture Notes in Computer Science","Explainable AI: Interpreting, Explaining and Visualizing Deep Learning"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-28954-6_1","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,9,28]],"date-time":"2022-09-28T00:10:52Z","timestamp":1664323852000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/978-3-030-28954-6_1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019]]},"ISBN":["9783030289539","9783030289546"],"references-count":97,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-28954-6_1","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2019]]},"assertion":[{"value":"10 September 2019","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}}]}}