{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,19]],"date-time":"2026-03-19T14:53:05Z","timestamp":1773931985068,"version":"3.50.1"},"reference-count":54,"publisher":"Springer Science and Business Media LLC","issue":"5","license":[{"start":{"date-parts":[[2022,4,13]],"date-time":"2022-04-13T00:00:00Z","timestamp":1649808000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2022,4,13]],"date-time":"2022-04-13T00:00:00Z","timestamp":1649808000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100001659","name":"Deutsche Forschungsgemeinschaft","doi-asserted-by":"publisher","award":["EXC 2046\/1"],"award-info":[{"award-number":["EXC 2046\/1"]}],"id":[{"id":"10.13039\/501100001659","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001659","name":"Deutsche Forschungsgemeinschaft","doi-asserted-by":"publisher","award":["project-ID: 390685689"],"award-info":[{"award-number":["project-ID: 390685689"]}],"id":[{"id":"10.13039\/501100001659","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100000781","name":"European Research Council","doi-asserted-by":"publisher","award":["758985"],"award-info":[{"award-number":["758985"]}],"id":[{"id":"10.13039\/501100000781","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002347","name":"Bundesministerium f\u00fcr Bildung und Forschung","doi-asserted-by":"publisher","award":["01IS18025A and 01IS18037A"],"award-info":[{"award-number":["01IS18025A and 01IS18037A"]}],"id":[{"id":"10.13039\/501100002347","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100010418","name":"Institute for Information and Communications Technology Promotion","doi-asserted-by":"publisher","award":["2019-0-00079"],"award-info":[{"award-number":["2019-0-00079"]}],"id":[{"id":"10.13039\/501100010418","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100006764","name":"Technische Universit\u00e4t Berlin","doi-asserted-by":"crossref","id":[{"id":"10.13039\/501100006764","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Mach Learn"],"published-print":{"date-parts":[[2022,5]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:p>Machine learning (ML) is increasingly often used to inform high-stakes decisions. As complex ML models (e.g., deep neural networks) are often considered black boxes, a wealth of procedures has been developed to shed light on their inner workings and the ways in which their predictions come about, defining the field of \u2018explainable AI\u2019 (XAI). Saliency methods rank input features according to some measure of \u2018importance\u2019. Such methods are difficult to validate since a formal definition of feature importance is, thus far, lacking. It has been demonstrated that some saliency methods can highlight features that have no statistical association with the prediction target (suppressor variables). To avoid misinterpretations due to such behavior, we propose the actual presence of such an association as a necessary condition and objective preliminary definition for feature importance. We carefully crafted a ground-truth dataset in which all statistical dependencies are well-defined and linear, serving as a benchmark to study the problem of suppressor variables. We evaluate common explanation methods including LRP, DTD, PatternNet, PatternAttribution, LIME, Anchors, SHAP, and permutation-based methods with respect to our objective definition. We show that most of these methods are unable to distinguish important features from suppressors in this setting.<\/jats:p>","DOI":"10.1007\/s10994-022-06167-y","type":"journal-article","created":{"date-parts":[[2022,4,13]],"date-time":"2022-04-13T20:10:46Z","timestamp":1649880646000},"page":"1903-1923","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":20,"title":["Scrutinizing XAI using linear ground-truth data with suppressor variables"],"prefix":"10.1007","volume":"111","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-7237-4322","authenticated-orcid":false,"given":"Rick","family":"Wilming","sequence":"first","affiliation":[]},{"given":"C\u00e9line","family":"Budding","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3861-7685","authenticated-orcid":false,"given":"Klaus-Robert","family":"M\u00fcller","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1470-9195","authenticated-orcid":false,"given":"Stefan","family":"Haufe","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,4,13]]},"reference":[{"key":"6167_CR1","unstructured":"Adebayo, J., Gilmer, J., Muelly, M., Goodfellow, I., Hardt, M., & Kim, B. (2018). Sanity checks for saliency maps. In: Proceedings of the 32nd International Conference on Neural Information Processing Systems, Curran Associates Inc., Montr\u00e9al, Canada, NIPS\u201918, pp 9525\u20139536."},{"issue":"93","key":"6167_CR2","first-page":"1","volume":"20","author":"M Alber","year":"2019","unstructured":"Alber, M., Lapuschkin, S., Seegerer, P., H\u00e4gele, M., Sch\u00fctt, K. T., Montavon, G., et al. (2019). Innvestigate eural networks! J Mach Learn Res, 20(93), 1\u20138.","journal-title":"J Mach Learn Res"},{"key":"6167_CR3","unstructured":"Alvarez-Melis, D., & Jaakkola, T.S. (2018). On the Robustness of Interpretability Methods. arXiv:180608049 [cs, stat] ArXiv: 1806.08049."},{"key":"6167_CR4","doi-asserted-by":"crossref","unstructured":"Ancona, M., Ceolini, E., \u00d6ztireli, C., & Gross, M. (2018). Towards better understanding of gradient-based attribution methods for deep neural networks. In: ICLR.","DOI":"10.1007\/978-3-030-28954-6_9"},{"key":"6167_CR5","doi-asserted-by":"publisher","first-page":"82","DOI":"10.1016\/j.inffus.2019.12.012","volume":"58","author":"AB Arrieta","year":"2020","unstructured":"Arrieta, A. B., D\u00edaz-Rodr\u00edguez, N., Del Ser, J., Bennetot, A., Tabik, S., Barbado, A., Garc\u00eda, S., Gil-L\u00f3pez, S., Molina, D., Benjamins, R., et al. (2020). Explainable artificial intelligence (xai): Concepts, taxonomies, opportunities and challenges toward responsible ai. Information Fusion, 58, 82\u2013115.","journal-title":"Information Fusion"},{"issue":"7","key":"6167_CR6","doi-asserted-by":"publisher","first-page":"e0130140","DOI":"10.1371\/journal.pone.0130140","volume":"10","author":"S Bach","year":"2015","unstructured":"Bach, S., Binder, A., Montavon, G., Klauschen, F., M\u00fcller, K.-R., & Samek, W. (2015). On pixel-wise explanations for non-linear classifier decisions by layer-wise relevance propagation. PLOS One, 10(7), e0130140.","journal-title":"PLOS One"},{"key":"6167_CR7","first-page":"1803","volume":"11","author":"D Baehrens","year":"2010","unstructured":"Baehrens, D., Schroeter, T., Harmeling, S., Kawanabe, M., Hansen, K., & M\u00fcller, K. R. (2010). How to explain individual classification decisions. The Journal of Machine Learning Research, 11, 1803\u20131831.","journal-title":"The Journal of Machine Learning Research"},{"key":"6167_CR8","doi-asserted-by":"crossref","unstructured":"Binder, A., Bach, S., Montavon, G., M\u00fcller, K. R., & Samek, W. (2016). Layer-Wise Relevance Propagation for Deep Neural Network Architectures. In K. J. Kim & N. Joukov (Eds.), Information Science and Applications (ICISA) 2016 (pp. 913\u2013922). Lecture Notes in Electrical Engineering: Springer, Singapore.","DOI":"10.1007\/978-981-10-0557-2_87"},{"issue":"1","key":"6167_CR9","doi-asserted-by":"publisher","first-page":"5","DOI":"10.1023\/A:1010933404324","volume":"45","author":"L Breiman","year":"2001","unstructured":"Breiman, L. (2001). Random forests. Machine Learning, 45(1), 5\u201332.","journal-title":"Machine Learning"},{"issue":"1","key":"6167_CR10","doi-asserted-by":"publisher","first-page":"35","DOI":"10.1177\/001316447403400105","volume":"34","author":"AJ Conger","year":"1974","unstructured":"Conger, A. J. (1974). A revised definition for suppressor variables: a guide to their identification and interpretation, a revised definition for suppressor variables: A guide to their identification and interpretation. Educational and Psychological Measurement, 34(1), 35\u201346.","journal-title":"Educational and Psychological Measurement"},{"key":"6167_CR11","doi-asserted-by":"publisher","first-page":"108194","DOI":"10.1016\/j.patcog.2021.108194","volume":"121","author":"AK Dombrowski","year":"2022","unstructured":"Dombrowski, A. K., Anders, C. J., M\u00fcller, K. R., & Kessel, P. (2022). Towards robust explanations for deep neural networks. Pattern Recognition, 121, 108194.","journal-title":"Pattern Recognition"},{"key":"6167_CR12","unstructured":"Doshi-Velez, F., & Kim, B. (2017). Towards a rigorous science of interpretable machine learning. arXiv:170208608 [cs, stat] ArXiv: 1702.08608."},{"issue":"177","key":"6167_CR13","first-page":"1","volume":"20","author":"A Fisher","year":"2019","unstructured":"Fisher, A., Rudin, C., & Dominici, F. (2019). All models are wrong, but many are useful: Learning a variable\u2019s importance by studying an entire class of prediction models simultaneously. Journal of Machine Learning Research, 20(177), 1\u201381.","journal-title":"Journal of Machine Learning Research"},{"key":"6167_CR14","doi-asserted-by":"crossref","unstructured":"Fong, R.C., & Vedaldi, A. (2017). Interpretable explanations of black boxes by meaningful perturbation. In: Proceedings of the IEEE International Conference on Computer Vision, pp 3429\u20133437.","DOI":"10.1109\/ICCV.2017.371"},{"issue":"2","key":"6167_CR15","doi-asserted-by":"publisher","first-page":"127","DOI":"10.1198\/000313005X41337","volume":"59","author":"L Friedman","year":"2005","unstructured":"Friedman, L., & Wall, M. (2005). Graphical views of suppression and multicollinearity in multiple linear regression. The American Statistician, 59(2), 127\u2013136.","journal-title":"The American Statistician"},{"key":"6167_CR16","first-page":"585","volume":"20","author":"A Gretton","year":"2007","unstructured":"Gretton, A., Fukumizu, K., Teo, C. H., Song, L., Sch\u00f6lkopf, B., Smola, A. J., et al. (2007). A kernel statistical test of independence. Nips Citeseer, 20, 585\u2013592.","journal-title":"Nips Citeseer"},{"key":"6167_CR17","doi-asserted-by":"publisher","first-page":"96","DOI":"10.1016\/j.neuroimage.2013.10.067","volume":"87","author":"S Haufe","year":"2014","unstructured":"Haufe, S., Meinecke, F., G\u00f6rgen, K., D\u00e4hne, S., Haynes, J. D., Blankertz, B., & Bie\u00dfmann, F. (2014). On the interpretation of weight vectors of linear models in multivariate neuroimaging. NeuroImage, 87, 96\u2013110.","journal-title":"NeuroImage"},{"key":"6167_CR18","unstructured":"Hooker, S., Erhan, D., Kindermans, P.J., & Kim, B. (2019). A benchmark for interpretability methods in deep neural networks. In: Wallach H, Larochelle H, Beygelzimer A, d\u2019Alch\u00e9-Buc F, Fox E, Garnett R (eds) Advances in Neural Information Processing Systems, Curran Associates, Inc., vol\u00a032, pp 9737\u20139748."},{"key":"6167_CR19","doi-asserted-by":"crossref","unstructured":"Horst, P., Col\u00a0Wallin, P., Col\u00a0Guttman, L., Brim Col\u00a0Wallin, F., Clausen, J.A., Col\u00a0Reed, R., & Col\u00a0Rosenthal, E. (1941). The prediction of personal adjustment: A survey of logical problems and research techniques, with illustrative application to problems of vocational selection, school success, marriage, and crime. Social science research council.","DOI":"10.1037\/11521-000"},{"key":"6167_CR20","unstructured":"Ismail, A.A., Gunady, M., Pessoa, L., Corrada\u00a0Bravo, H., & Feizi, S. (2019). Input-cell attention reduces vanishing saliency of recurrent neural networks. In: H. Wallach, H. Larochelle, A. Beygelzimer, F. d\u2019Alch\u00e9-Buc, E. Fox, R. Garnett (eds) Advances in Neural Information Processing Systems, Curran Associates, Inc., vol\u00a032, pp 10814\u201310824."},{"key":"6167_CR21","unstructured":"Jaderberg, M., Simonyan, K., Zisserman, A., & Kavukcuoglu, K. (2015). Spatial transformer networks. In: Proceedings of the 28th International Conference on Neural Information Processing Systems-(Vol. 2, pp 2017\u20132025)."},{"key":"6167_CR22","unstructured":"Kim, B., Wattenberg, M., Gilmer, J., Cai, C., Wexler, J., Viegas, F., & Sayres, R. (2018). Interpretability beyond feature attribution: Quantitative testing with concept activation vectors (TCAV). In: International Conference on Machine Learning, PMLR, pp 2668\u20132677."},{"key":"6167_CR23","unstructured":"Kindermans, P., Sch\u00fctt, K.T., Alber, M., M\u00fcller, K., Erhan, D., Kim, B., & D\u00e4hne, S. (2018). Learning how to explain neural networks: Patternnet and patternattribution. In: ICLR."},{"key":"6167_CR24","first-page":"1097","volume":"25","author":"A Krizhevsky","year":"2012","unstructured":"Krizhevsky, A., Sutskever, I., & Hinton, G. E. (2012). Imagenet classification with deep convolutional neural networks. Advances in Neural Information Processing Systems, 25, 1097\u20131105.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"6167_CR25","unstructured":"Lage, I., Ross, A., Gershman, S.J., Kim, B., & Doshi-Velez, F. (2018). Human-in-the-Loop Interpretability Prior. In: S. Bengio,H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, R. Garnett (eds) Advances in Neural Information Processing Systems 31, Curran Associates, Inc., pp 10159\u201310168."},{"issue":"1","key":"6167_CR26","doi-asserted-by":"publisher","first-page":"1096","DOI":"10.1038\/s41467-019-08987-4","volume":"10","author":"S Lapuschkin","year":"2019","unstructured":"Lapuschkin, S., W\u00e4ldchen, S., Binder, A., Montavon, G., Samek, W., & M\u00fcller, K. R. (2019). Unmasking Clever Hans predictors and assessing what machines really learn. Nature Communications, 10(1), 1096.","journal-title":"Nature Communications"},{"issue":"7553","key":"6167_CR27","doi-asserted-by":"publisher","first-page":"436","DOI":"10.1038\/nature14539","volume":"521","author":"Y LeCun","year":"2015","unstructured":"LeCun, Y., Bengio, Y., & Hinton, G. (2015). Deep learning. Nature, 521(7553), 436\u2013444.","journal-title":"Nature"},{"issue":"3","key":"6167_CR28","doi-asserted-by":"publisher","first-page":"31","DOI":"10.1145\/3236386.3241340","volume":"16","author":"ZC Lipton","year":"2018","unstructured":"Lipton, Z. C. (2018). The mythos of model interpretability: In machine learning, the concept of interpretability is both important and slippery. Queue, 16(3), 31\u201357.","journal-title":"Queue"},{"key":"6167_CR29","unstructured":"Lundberg, S.M., & Lee, S.I. (2017). A Unified Approach to Interpreting Model Predictions. In: I. Guyon, U.V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, R. Garnett (eds) Advances in Neural Information Processing Systems 30, Curran Associates, Inc., pp 4765\u20134774."},{"key":"6167_CR30","doi-asserted-by":"publisher","first-page":"211","DOI":"10.1016\/j.patcog.2016.11.008","volume":"65","author":"G Montavon","year":"2017","unstructured":"Montavon, G., Bach, S., Binder, A., Samek, W., & M\u00fcller, K. R. (2017). Explaining nonlinear classification decisions with deep Taylor decomposition. Pattern Recognition, 65, 211\u2013222.","journal-title":"Pattern Recognition"},{"key":"6167_CR31","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.dsp.2017.10.011","volume":"73","author":"G Montavon","year":"2018","unstructured":"Montavon, G., Samek, W., & M\u00fcller, K. R. (2018). Methods for interpreting and understanding deep neural networks. Digital Signal Processing, 73, 1\u201315.","journal-title":"Digital Signal Processing"},{"issue":"44","key":"6167_CR32","doi-asserted-by":"publisher","first-page":"22071","DOI":"10.1073\/pnas.1900654116","volume":"116","author":"WJ Murdoch","year":"2019","unstructured":"Murdoch, W. J., Singh, C., Kumbier, K., Abbasi-Asl, R., & Yu, B. (2019). Definitions, methods, and applications in interpretable machine learning. Proceedings of the National Academy of Sciences, 116(44), 22071\u201322080.","journal-title":"Proceedings of the National Academy of Sciences"},{"key":"6167_CR33","unstructured":"Nguyen, A.p., & Mart\u00ednez, M.R. (2020). On quantitative aspects of model interpretability. arXiv:200707584 [cs, stat] ArXiv: 2007.07584."},{"key":"6167_CR34","doi-asserted-by":"crossref","unstructured":"Park, D.H., Hendricks, L.A., Akata, Z., Rohrbach, A., Schiele, B., Darrell, T., & Rohrbach, M. (2018). Multimodal explanations: Justifying decisions and pointing to the evidence. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR).","DOI":"10.1109\/CVPR.2018.00915"},{"key":"6167_CR35","unstructured":"Pedregosa, F., Varoquaux, G., Gramfort, A., Michel, V., Thirion, B., Grisel, O., Blondel, M., Prettenhofer, P., Weiss, R., Dubourg, V., & others. (2011). Scikit-learn: Machine learning in Python. Journal of Machine Learning Research 12:2825\u20132830"},{"key":"6167_CR36","doi-asserted-by":"crossref","unstructured":"Poursabzi-Sangdeh, F., Goldstein, D.G., Hofman, J.M., Wortman\u00a0Vaughan, J.W., & Wallach, H. (2021). Manipulating and measuring model interpretability. In: Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems, pp 1\u201352.","DOI":"10.1145\/3411764.3445315"},{"key":"6167_CR37","doi-asserted-by":"crossref","unstructured":"Ribeiro, M.T., Singh, S., & Guestrin, C. (2016). \u201c why should i trust you?\u201d explaining the predictions of any classifier. In: Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining, pp 1135\u20131144.","DOI":"10.1145\/2939672.2939778"},{"key":"6167_CR38","unstructured":"Ribeiro, M.T., Singh, S., & Guestrin, C. (2018). Anchors: High-Precision Model-Agnostic Explanations. In: AAAI."},{"issue":"5","key":"6167_CR39","doi-asserted-by":"publisher","first-page":"206","DOI":"10.1038\/s42256-019-0048-x","volume":"1","author":"C Rudin","year":"2019","unstructured":"Rudin, C. (2019). Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nature Machine Intelligence, 1(5), 206\u2013215.","journal-title":"Nature Machine Intelligence"},{"issue":"11","key":"6167_CR40","doi-asserted-by":"publisher","first-page":"2660","DOI":"10.1109\/TNNLS.2016.2599820","volume":"28","author":"W Samek","year":"2016","unstructured":"Samek, W., Binder, A., Montavon, G., Lapuschkin, S., & M\u00fcller, K. R. (2016). Evaluating the visualization of what a deep neural network has learned. IEEE Transactions on Neural Networks and Learning Systems, 28(11), 2660\u20132673.","journal-title":"IEEE Transactions on Neural Networks and Learning Systems"},{"key":"6167_CR41","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-28954-6","volume-title":"Explainable AI: Interpreting, explaining and visualizing deep learning","author":"W Samek","year":"2019","unstructured":"Samek, W., Montavon, G., Vedaldi, A., Hansen, L. K., & M\u00fcller, K. R. (2019). Explainable AI: Interpreting, explaining and visualizing deep learning (Vol. 11700). New York: Springer."},{"issue":"3","key":"6167_CR42","doi-asserted-by":"publisher","first-page":"247","DOI":"10.1109\/JPROC.2021.3060483","volume":"109","author":"W Samek","year":"2021","unstructured":"Samek, W., Montavon, G., Lapuschkin, S., Anders, C. J., & M\u00fcller, K. R. (2021). Explaining deep neural networks and beyond: A review of methods and applications. Proceedings of the IEEE, 109(3), 247\u2013278.","journal-title":"Proceedings of the IEEE"},{"key":"6167_CR43","unstructured":"Schmidt, P., & Biessmann, F. (2019). Quantifying interpretability and trust in machine learning systems. arXiv:190108558 [cs, stat] ArXiv: 1901.08558."},{"issue":"28","key":"6167_CR44","first-page":"307","volume":"2","author":"LS Shapley","year":"1953","unstructured":"Shapley, L. S. (1953). A value for n-person games. Contributions to the theory of games, 2(28), 307\u2013317.","journal-title":"Contributions to the theory of games"},{"issue":"7676","key":"6167_CR45","doi-asserted-by":"publisher","first-page":"354","DOI":"10.1038\/nature24270","volume":"550","author":"D Silver","year":"2017","unstructured":"Silver, D., Schrittwieser, J., Simonyan, K., Antonoglou, I., Huang, A., Guez, A., Hubert, T., Baker, L., Lai, M., Bolton, A., Chen, Y., Lillicrap, T., Hui, F., Sifre, L., van den Driessche, G., Graepel, T., & Hassabis, D. (2017). Mastering the game of Go without human knowledge. Nature, 550(7676), 354\u2013359.","journal-title":"Nature"},{"key":"6167_CR46","unstructured":"Simonyan, K., & Zisserman, A. (2015). Very deep convolutional networks for large-scale image recognition. arXiv:14091556 [cs] ArXiv: 1409.1556."},{"key":"6167_CR47","unstructured":"Simonyan, K., Vedaldi, A., & Zisserman, A. (2013). Deep inside convolutional networks: Visualising image classification models and saliency maps. arXiv preprint arXiv:13126034."},{"key":"6167_CR48","unstructured":"Sixt, L., Granz, M., & Landgraf, T. (2020). When explanations lie: Why many modified bp attributions fail. In: International Conference on Machine Learning, PMLR, pp 9046\u20139057."},{"key":"6167_CR49","unstructured":"Springenberg, J.T., Dosovitskiy, A., Brox, T., & Riedmiller, M.A. (2015). Striving for simplicity: The all convolutional net. CoRR arXiv:1412.6806."},{"key":"6167_CR50","unstructured":"Tjoa, E., & Guan, C. (2020). Quantifying explainability of saliency methods in deep neural networks. arXiv:200902899 [cs] ArXiv: 2009.02899."},{"key":"6167_CR51","unstructured":"Yang, M., & Kim, B. (2019). Benchmarking attribution methods with relative feature importance. arXiv:190709701 [cs, stat] ArXiv: 1907.09701."},{"key":"6167_CR52","doi-asserted-by":"crossref","unstructured":"Zeiler, M. D., & Fergus, R. (2014). Visualizing and Understanding Convolutional Networks. In D. Fleet, T. Pajdla, B. Schiele, & T. Tuytelaars (Eds.), Computer Vision - ECCV 2014 (pp. 818\u2013833). Lecture Notes in Computer Science: Springer International Publishing, Cham.","DOI":"10.1007\/978-3-319-10590-1_53"},{"key":"6167_CR53","doi-asserted-by":"crossref","unstructured":"Zien, A., Kr\u00e4mer, N., Sonnenburg, S., & R\u00e4tsch, G. (2009). The Feature Importance Ranking Measure. In W. Buntine, M. Grobelnik, D. Mladeni\u0107, & J. Shawe-Taylor (Eds.), Machine Learning and Knowledge Discovery in Databases (pp. 694\u2013709). Lecture Notes in Computer Science: Springer, Berlin, Heidelberg.","DOI":"10.1007\/978-3-642-04174-7_45"},{"issue":"3","key":"6167_CR54","doi-asserted-by":"publisher","first-page":"647","DOI":"10.1007\/s10115-013-0679-x","volume":"41","author":"E \u0160trumbelj","year":"2014","unstructured":"\u0160trumbelj, E., & Kononenko, I. (2014). Explaining prediction models and individual predictions with feature contributions. Knowledge and Information Systems, 41(3), 647\u2013665.","journal-title":"Knowledge and Information Systems"}],"container-title":["Machine Learning"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10994-022-06167-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10994-022-06167-y\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10994-022-06167-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,5,20]],"date-time":"2022-05-20T20:16:46Z","timestamp":1653077806000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10994-022-06167-y"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,4,13]]},"references-count":54,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2022,5]]}},"alternative-id":["6167"],"URL":"https:\/\/doi.org\/10.1007\/s10994-022-06167-y","relation":{},"ISSN":["0885-6125","1573-0565"],"issn-type":[{"value":"0885-6125","type":"print"},{"value":"1573-0565","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,4,13]]},"assertion":[{"value":"2 May 2021","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"30 November 2021","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"13 March 2022","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"13 April 2022","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflicts of interest\/competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflicts of interest\/Competing interests"}}]}}