{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,14]],"date-time":"2026-04-14T20:14:35Z","timestamp":1776197675423,"version":"3.50.1"},"reference-count":278,"publisher":"Springer Science and Business Media LLC","issue":"5","license":[{"start":{"date-parts":[[2021,11,18]],"date-time":"2021-11-18T00:00:00Z","timestamp":1637193600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2021,11,18]],"date-time":"2021-11-18T00:00:00Z","timestamp":1637193600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Artif Intell Rev"],"published-print":{"date-parts":[[2022,6]]},"DOI":"10.1007\/s10462-021-10088-y","type":"journal-article","created":{"date-parts":[[2021,11,18]],"date-time":"2021-11-18T17:04:09Z","timestamp":1637255049000},"page":"3503-3568","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":669,"title":["Explainable artificial intelligence: a comprehensive review"],"prefix":"10.1007","volume":"55","author":[{"given":"Dang","family":"Minh","sequence":"first","affiliation":[]},{"given":"H. Xiang","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Y. Fen","family":"Li","sequence":"additional","affiliation":[]},{"given":"Tan N.","family":"Nguyen","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2021,11,18]]},"reference":[{"key":"10088_CR1","doi-asserted-by":"crossref","unstructured":"Abdollahi B, Nasraoui O (2018) Transparency in fair machine learning: the case of explainable recommender systems. In: Human and machine learning. Springer, Berlin, pp 21?35","DOI":"10.1007\/978-3-319-90403-0_2"},{"key":"10088_CR2","unstructured":"ACM (2020) ACM conference on fairness, accountability, and transparency. https:\/\/fatconference.org. Accessed 24 Jan 2020"},{"key":"10088_CR3","doi-asserted-by":"publisher","first-page":"52138","DOI":"10.1109\/ACCESS.2018.2870052","volume":"6","author":"A Adadi","year":"2018","unstructured":"Adadi A, Berrada M (2018) Peeking inside the black-box: a survey on explainable artificial intelligence (XAI). IEEE Access 6:52138?52160","journal-title":"IEEE Access"},{"key":"10088_CR4","first-page":"9505","volume":"31","author":"J Adebayo","year":"2018","unstructured":"Adebayo J, Gilmer J, Muelly M, Goodfellow I, Hardt M, Kim B (2018) Sanity checks for saliency maps. Adv Neural Inf Process Syst 31:9505?9515","journal-title":"Adv Neural Inf Process Syst"},{"issue":"1","key":"10088_CR5","doi-asserted-by":"publisher","first-page":"95","DOI":"10.1007\/s10115-017-1116-3","volume":"54","author":"P Adler","year":"2018","unstructured":"Adler P, Falk C, Friedler SA, Nix T, Rybeck G, Scheidegger C, Smith B, Venkatasubramanian S (2018) Auditing black-box models for indirect influence. Knowl Inf Syst 54(1):95?122","journal-title":"Knowl Inf Syst"},{"issue":"7","key":"10088_CR6","doi-asserted-by":"publisher","first-page":"1571","DOI":"10.1007\/s00521-012-1048-5","volume":"22","author":"FC Adriana da Costa","year":"2013","unstructured":"Adriana da Costa FC, Vellasco MMB, Tanscheit R (2013) Fuzzy rules extraction from support vector machines for multi-class classification. Neural Comput Appl 22(7):1571?1580","journal-title":"Neural Comput Appl"},{"issue":"2","key":"10088_CR7","doi-asserted-by":"publisher","first-page":"249","DOI":"10.1007\/s10115-018-1183-0","volume":"58","author":"M Ahmed","year":"2019","unstructured":"Ahmed M (2019) Data summarization: a survey. Knowl Inf Syst 58(2):249?273","journal-title":"Knowl Inf Syst"},{"issue":"1","key":"10088_CR8","first-page":"1086","volume":"26","author":"Y Ahn","year":"2019","unstructured":"Ahn Y, Lin YR (2019) Fairsight: visual analytics for fairness in decision making. IEEE Trans Vis Comput Graph 26(1):1086?1095","journal-title":"IEEE Trans Vis Comput Graph"},{"key":"10088_CR9","unstructured":"AI (2019) Ethics for autonomous systems. https:\/\/www.journals.elsevier.com\/artificial-intelligence\/call-for-papers\/special-issue-on-ethics-for-autonomous-systems. Accessed 3 Mar 2020"},{"key":"10088_CR10","unstructured":"AI (2020) Explainable artificial intelligence. https:\/\/www.journals.elsevier.com\/artificial-intelligence\/call-for-papers\/special-issue-on-explainable-artificial-intelligence. Accessed 3 Mar 2020"},{"key":"10088_CR11","unstructured":"Akula AR, Todorovic S, Chai JY, Zhu SC (2019) Natural language interaction with explainable AI models. In: CVPR workshops, pp 87?90"},{"issue":"194","key":"10088_CR12","first-page":"1","volume":"21","author":"M Al-Shedivat","year":"2020","unstructured":"Al-Shedivat M, Dubey A, Xing E (2020) Contextual explanation networks. J Mach Learn Res 21(194):1?44","journal-title":"J Mach Learn Res"},{"key":"10088_CR13","doi-asserted-by":"publisher","first-page":"185","DOI":"10.1016\/j.neunet.2020.07.010","volume":"130","author":"P Angelov","year":"2020","unstructured":"Angelov P, Soares E (2020) Towards explainable deep neural networks (xDNN). Neural Netw 130:185?194","journal-title":"Neural Netw"},{"key":"10088_CR14","doi-asserted-by":"publisher","first-page":"66","DOI":"10.1016\/j.proeng.2016.08.081","volume":"153","author":"H Anysz","year":"2016","unstructured":"Anysz H, Zbiciak A, Ibadov N (2016) The influence of input data standardization method on prediction accuracy of artificial neural networks. Proc Eng 153:66?70","journal-title":"Proc Eng"},{"key":"10088_CR15","doi-asserted-by":"crossref","unstructured":"Arras L, Arjona-Medina J, Widrich M, Montavon G (2019) Explaining and interpreting lstms. In: Explainable AI: interpreting, explaining and visualizing deep learning, vol 11700, p 211","DOI":"10.1007\/978-3-030-28954-6_11"},{"key":"10088_CR16","doi-asserted-by":"publisher","first-page":"82","DOI":"10.1016\/j.inffus.2019.12.012","volume":"58","author":"AB Arrieta","year":"2020","unstructured":"Arrieta AB, D\u00edaz-Rodr\u00edguez N, Del Ser J, Bennetot A, Tabik S, Barbado A, Garc\u00eda S, Gil-L\u00f3pez S, Molina D, Benjamins R et al (2020) Explainable artificial intelligence (XAI): concepts, taxonomies, opportunities and challenges toward responsible AI. Inf Fusion 58:82?115","journal-title":"Inf Fusion"},{"issue":"4","key":"10088_CR17","doi-asserted-by":"publisher","first-page":"305","DOI":"10.1007\/s10799-016-0270-8","volume":"18","author":"S Asadi","year":"2017","unstructured":"Asadi S, Nilashi M, Husin ARC, Yadegaridehkordi E (2017) Customers perspectives on adoption of cloud computing in banking sector. Inf Technol Manag 18(4):305?330","journal-title":"Inf Technol Manag"},{"key":"10088_CR18","doi-asserted-by":"crossref","unstructured":"Assaf R, Giurgiu I, Bagehorn F, Schumann A (2019) Mtex-cnn: Multivariate time series explanations for predictions with convolutional neural networks. In: 2019 IEEE international conference on data mining (ICDM). IEEE, pp 952?957","DOI":"10.1109\/ICDM.2019.00106"},{"key":"10088_CR19","doi-asserted-by":"crossref","unstructured":"Bang JS, Lee MH, Fazli S, Guan C, Lee SW (2021) Spatio-spectral feature representation for motor imagery classification using convolutional neural networks. IEEE Trans Neural Netw Learn Syst","DOI":"10.1109\/TNNLS.2020.3048385"},{"issue":"43","key":"10088_CR20","doi-asserted-by":"publisher","first-page":"1798","DOI":"10.21105\/joss.01798","volume":"4","author":"H Baniecki","year":"2019","unstructured":"Baniecki H, Biecek P (2019) modelStudio: Interactive studio with explanations for ML predictive models. J Open Source Softw 4(43):1798","journal-title":"J Open Source Softw"},{"key":"10088_CR21","doi-asserted-by":"crossref","unstructured":"Baron B, Musolesi M (2020) Interpretable machine learning for privacy-preserving pervasive systems. IEEE Pervasive Comput","DOI":"10.1109\/MPRV.2019.2918540"},{"key":"10088_CR22","doi-asserted-by":"crossref","unstructured":"Bau D, Zhou B, Khosla A, Oliva A, Torralba A (2017) Network dissection: Quantifying interpretability of deep visual representations. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 6541?6549","DOI":"10.1109\/CVPR.2017.354"},{"key":"10088_CR23","doi-asserted-by":"publisher","first-page":"587","DOI":"10.1162\/tacl_a_00041","volume":"6","author":"EM Bender","year":"2018","unstructured":"Bender EM, Friedman B (2018) Data statements for natural language processing: toward mitigating system bias and enabling better science. Trans Assoc Comput Linguist 6:587?604","journal-title":"Trans Assoc Comput Linguist"},{"key":"10088_CR24","doi-asserted-by":"publisher","first-page":"883","DOI":"10.1016\/j.ins.2020.08.109","volume":"546","author":"X Bi","year":"2021","unstructured":"Bi X, Zhang C, He Y, Zhao X, Sun Y, Ma Y (2021) Explainable time?frequency convolutional neural network for microseismic waveform classification. Inf Sci 546:883?896","journal-title":"Inf Sci"},{"key":"10088_CR25","doi-asserted-by":"publisher","first-page":"105532","DOI":"10.1016\/j.knosys.2020.105532","volume":"194","author":"A Blanco-Justicia","year":"2020","unstructured":"Blanco-Justicia A, Domingo-Ferrer J, Mart\u00ednez S, S\u00e1nchez D (2020) Machine learning explainability via microaggregation and shallow decision trees. Knowl-Based Syst 194:105532","journal-title":"Knowl-Based Syst"},{"key":"10088_CR26","unstructured":"BMVC (2020) Interpretable & explainable machine vision. https:\/\/arxiv.org\/html\/1909.07245. Accessed 3 Mar 2020"},{"issue":"12","key":"10088_CR27","doi-asserted-by":"publisher","first-page":"2411","DOI":"10.3390\/app9122411","volume":"9","author":"G Bologna","year":"2019","unstructured":"Bologna G (2019) A simple convolutional neural network with rule extraction. Appl Sci 9(12):2411","journal-title":"Appl Sci"},{"issue":"2","key":"10088_CR28","doi-asserted-by":"publisher","first-page":"257","DOI":"10.1016\/j.clsr.2018.01.004","volume":"34","author":"M Butterworth","year":"2018","unstructured":"Butterworth M (2018) The ICO and artificial intelligence: the role of fairness in the GDPR framework. Comput Law Secur Rev 34(2):257?268","journal-title":"Comput Law Secur Rev"},{"issue":"1","key":"10088_CR29","first-page":"551","volume":"20","author":"T Campbell","year":"2019","unstructured":"Campbell T, Broderick T (2019) Automated scalable Bayesian inference via Hilbert coresets. J Mach Learn Res 20(1):551?588","journal-title":"J Mach Learn Res"},{"key":"10088_CR30","doi-asserted-by":"publisher","first-page":"218180","DOI":"10.1109\/ACCESS.2020.3041040","volume":"8","author":"HE Cao","year":"2020","unstructured":"Cao HE, Sarlin R, Jung A (2020) Learning explainable decision rules via maximum satisfiability. IEEE Access 8:218180?218185","journal-title":"IEEE Access"},{"key":"10088_CR31","volume-title":"Data protection: a practical guide to UK and EU law","author":"P Carey","year":"2018","unstructured":"Carey P (2018) Data protection: a practical guide to UK and EU law. Oxford University Press, Inc, Oxford"},{"issue":"3","key":"10088_CR32","doi-asserted-by":"publisher","first-page":"e15","DOI":"10.23915\/distill.00015","volume":"4","author":"S Carter","year":"2019","unstructured":"Carter S, Armstrong Z, Schubert L, Johnson I, Olah C (2019) Activation atlas. Distill 4(3):e15","journal-title":"Distill"},{"issue":"8","key":"10088_CR33","doi-asserted-by":"publisher","first-page":"832","DOI":"10.3390\/electronics8080832","volume":"8","author":"DV Carvalho","year":"2019","unstructured":"Carvalho DV, Pereira EM, Cardoso JS (2019a) Machine learning interpretability: a survey on methods and metrics. Electronics 8(8):832","journal-title":"Electronics"},{"issue":"8","key":"10088_CR34","doi-asserted-by":"publisher","first-page":"832","DOI":"10.3390\/electronics8080832","volume":"8","author":"DV Carvalho","year":"2019","unstructured":"Carvalho DV, Pereira EM, Cardoso JS (2019b) Machine learning interpretability: a survey on methods and metrics. Electronics 8(8):832","journal-title":"Electronics"},{"issue":"2","key":"10088_CR35","doi-asserted-by":"publisher","first-page":"330","DOI":"10.1007\/s12559-019-09634-2","volume":"12","author":"A Ceni","year":"2020","unstructured":"Ceni A, Ashwin P, Livi L (2020) Interpreting recurrent neural networks behaviour via excitable network attractors. Cogn Comput 12(2):330?356","journal-title":"Cogn Comput"},{"key":"10088_CR36","doi-asserted-by":"crossref","unstructured":"Chakraborty S, Tomsett R, Raghavendra R, Harborne D, Alzantot M, Cerutti F, Srivastava M, Preece A, Julier S, Rao RM et al (2017) Interpretability of deep learning models: a survey of results. In: 2017 IEEE SmartWorld, ubiquitous intelligence & computing, advanced & trusted computed, scalable computing & communications, cloud & big data computing, internet of people and smart city innovation (SmartWorld\/SCALCOM\/UIC\/ATC\/CBDCom\/IOP\/SCI). IEEE, pp 1?6","DOI":"10.1109\/UIC-ATC.2017.8397411"},{"issue":"12","key":"10088_CR37","doi-asserted-by":"publisher","first-page":"5017","DOI":"10.1109\/TIP.2015.2475625","volume":"24","author":"TH Chan","year":"2015","unstructured":"Chan TH, Jia K, Gao S, Lu J, Zeng Z, Ma Y (2015) PCANet: a simple deep learning baseline for image classification? IEEE Trans Image Process 24(12):5017?5032","journal-title":"IEEE Trans Image Process"},{"key":"10088_CR38","unstructured":"Chen J, Song L, Wainwright MJ, Jordan MI (2018) L-shapley and c-shapley: efficient model interpretation for structured data. In: International conference on learning representations"},{"key":"10088_CR39","doi-asserted-by":"crossref","unstructured":"Chen J, Vaughan J, Nair V, Sudjianto A (2020a) Adaptive explainable neural networks (AxNNs). Available at SSRN 3569318","DOI":"10.2139\/ssrn.3569318"},{"key":"10088_CR40","doi-asserted-by":"publisher","first-page":"104332","DOI":"10.1016\/j.ijmedinf.2020.104332","volume":"145","author":"Y Chen","year":"2020","unstructured":"Chen Y, Yu C, Liu X, Xi T, Xu G, Sun Y, Zhu F, Shen B (2020b) PCLiON: an ontology for data standardization and sharing of prostate cancer associated lifestyles. Int J Med Inform 145:104332","journal-title":"Int J Med Inform"},{"key":"10088_CR41","doi-asserted-by":"crossref","unstructured":"Chen H, Lundberg S, Lee SI (2021) Explaining models by propagating Shapley values of local components. In: Explainable AI in Healthcare and Medicine. Springer, Berlin, pp 261?270","DOI":"10.1007\/978-3-030-53352-6_24"},{"key":"10088_CR42","unstructured":"Choi E, Bahadori MT, Kulas JA, Schuetz A, Stewart WF, Sun J (2016) Retain: an interpretable predictive model for healthcare using reverse time attention mechanism. In: Advances in Neural Information Processing Systems, pp 3512?3520"},{"issue":"9","key":"10088_CR43","doi-asserted-by":"publisher","first-page":"1197","DOI":"10.1093\/neuonc\/noz095","volume":"21","author":"KS Choi","year":"2019","unstructured":"Choi KS, Choi SH, Jeong B (2019) Prediction of IDH genotype in gliomas with dynamic susceptibility contrast perfusion MR imaging using an explainable recurrent neural network. Neuro Oncol 21(9):1197?1209","journal-title":"Neuro Oncol"},{"key":"10088_CR44","doi-asserted-by":"crossref","unstructured":"Choi H, Som A, Turaga P (2020) AMC-loss: angular margin contrastive loss for improved explainability in image classification. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition workshops, pp 838?839","DOI":"10.1109\/CVPRW50498.2020.00427"},{"issue":"4","key":"10088_CR45","doi-asserted-by":"publisher","first-page":"84","DOI":"10.1109\/MCG.2018.042731661","volume":"38","author":"J Choo","year":"2018","unstructured":"Choo J, Liu S (2018) Visual analytics for explainable deep learning. IEEE Comput Graph Appl 38(4):84?92","journal-title":"IEEE Comput Graph Appl"},{"key":"10088_CR46","unstructured":"CIM I (2021) Explainable and trustworthy artificial intelligence. https:\/\/sites.google.com\/view\/special-issue-on-xai-ieee-cim. Accessed 1 Aug 2021"},{"key":"10088_CR47","first-page":"426","volume":"128","author":"VG Comizio","year":"2011","unstructured":"Comizio VG, Petrasic KL, Lee HY (2011) Regulators take steps to eliminate differences in thrift, bank and holding company reporting requirements. Banking LJ 128:426","journal-title":"Banking LJ"},{"key":"10088_CR48","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.ins.2012.10.039","volume":"225","author":"P Cortez","year":"2013","unstructured":"Cortez P, Embrechts MJ (2013) Using sensitivity analysis and visualization techniques to open black box data mining models. Inf Sci 225:1?17","journal-title":"Inf Sci"},{"key":"10088_CR49","doi-asserted-by":"crossref","unstructured":"Craven MW, Shavlik JW (2014) Learning symbolic rules using artificial neural networks. In: Proceedings of the tenth international conference on machine learning, pp 73?80","DOI":"10.1016\/B978-1-55860-307-3.50016-2"},{"key":"10088_CR50","doi-asserted-by":"crossref","unstructured":"Daglarli E (2020) Explainable artificial intelligence (XAI) approaches and deep meta-learning models. In: Advances and applications in deep learning, p 79","DOI":"10.5772\/intechopen.92172"},{"key":"10088_CR51","doi-asserted-by":"publisher","first-page":"138872","DOI":"10.1109\/ACCESS.2019.2941376","volume":"7","author":"J Dai","year":"2019","unstructured":"Dai J, Chen C, Li Y (2019) A backdoor attack against lstm-based text classification systems. IEEE Access 7:138872?138878","journal-title":"IEEE Access"},{"key":"10088_CR52","doi-asserted-by":"publisher","first-page":"96","DOI":"10.1016\/j.compind.2018.03.020","volume":"99","author":"LM Dang","year":"2018","unstructured":"Dang LM, Hassan SI, Im S, Mehmood I, Moon H (2018) Utilizing text recognition for the defects extraction in sewers CCTV inspection videos. Comput Ind 99:96?109","journal-title":"Comput Ind"},{"issue":"7","key":"10088_CR53","doi-asserted-by":"publisher","first-page":"768","DOI":"10.3390\/electronics8070768","volume":"8","author":"LM Dang","year":"2019","unstructured":"Dang LM, Piran M, Han D, Min K, Moon H et al (2019) A survey on internet of things and cloud computing for healthcare. Electronics 8(7):768","journal-title":"Electronics"},{"key":"10088_CR54","unstructured":"Darpa (2020) Explainable artificial intelligence (XAI). https:\/\/www.darpa.mil\/program\/explainable-artificial-intelligence. Accessed 24 Jan 2020"},{"key":"10088_CR55","doi-asserted-by":"publisher","first-page":"40","DOI":"10.1016\/j.procs.2020.02.255","volume":"168","author":"T De","year":"2020","unstructured":"De T, Giri P, Mevawala A, Nemani R, Deo A (2020) Explainable AI: a hybrid approach to generate human-interpretable explanation for deep learning prediction. Procedia Comput Sci 168:40?48","journal-title":"Procedia Comput Sci"},{"issue":"7","key":"10088_CR56","first-page":"1829","volume":"119","author":"A Deeks","year":"2019","unstructured":"Deeks A (2019) The judicial demand for explainable artificial intelligence. Columbia Law Rev 119(7):1829?1850","journal-title":"Columbia Law Rev"},{"issue":"5","key":"10088_CR57","doi-asserted-by":"publisher","first-page":"893","DOI":"10.1007\/s11222-014-9461-5","volume":"25","author":"A Deleforge","year":"2015","unstructured":"Deleforge A, Forbes F, Horaud R (2015) High-dimensional regression with gaussian mixtures and partially-latent response variables. Stat Comput 25(5):893?911","journal-title":"Stat Comput"},{"issue":"4","key":"10088_CR58","doi-asserted-by":"publisher","first-page":"277","DOI":"10.1007\/s41060-018-0144-8","volume":"7","author":"H Deng","year":"2019","unstructured":"Deng H (2019) Interpreting tree ensembles with intrees. Int J Data Sci Anal 7(4):277?287","journal-title":"Int J Data Sci Anal"},{"issue":"5","key":"10088_CR59","doi-asserted-by":"publisher","first-page":"33","DOI":"10.1109\/MCG.2019.2924636","volume":"39","author":"V Dibia","year":"2019","unstructured":"Dibia V, Demiralp \u00c7 (2019) Data2vis: automatic generation of data visualizations using sequence-to-sequence recurrent neural networks. IEEE Comput Graph Appl 39(5):33?46","journal-title":"IEEE Comput Graph Appl"},{"key":"10088_CR60","doi-asserted-by":"publisher","first-page":"1561","DOI":"10.1016\/j.procs.2018.08.129","volume":"126","author":"L Ding","year":"2018","unstructured":"Ding L (2018) Human knowledge in constructing AI systems?neural logic networks approach towards an explainable AI. Procedia Comput Sci 126:1561?1570","journal-title":"Procedia Comput Sci"},{"key":"10088_CR61","doi-asserted-by":"crossref","unstructured":"Dingen D, van?t Veer M, Houthuizen P, Mestrom EH, Korsten EH, Bouwman AR, Van\u00a0Wijk J (2018) Regressionexplorer: interactive exploration of logistic regression models with subgroup analysis. IEEE Trans Vis Comput Graph 25(1):246?255","DOI":"10.1109\/TVCG.2018.2865043"},{"key":"10088_CR62","unstructured":"DMKD (2021) Data mining and knowledge discovery. https:\/\/www.springer.com\/journal\/10618\/updates\/18745970. Aceessed 1 Aug 2021"},{"issue":"11","key":"10088_CR63","doi-asserted-by":"publisher","first-page":"6373","DOI":"10.1007\/s11042-015-2576-7","volume":"75","author":"DP Dogra","year":"2016","unstructured":"Dogra DP, Ahmed A, Bhaskar H (2016) Smart video summarization using mealy machine-based trajectory modelling for surveillance applications. Multimed Tools Appl 75(11):6373?6401","journal-title":"Multimed Tools Appl"},{"key":"10088_CR64","unstructured":"Doran D, Schulz S, Besold TR (2017) What does explainable AI really mean? A new conceptualization of perspectives. arXiv preprint arXiv:171000794"},{"key":"10088_CR65","doi-asserted-by":"crossref","unstructured":"DuMouchel W (2002) Data squashing: constructing summary data sets. In: Handbook of massive data sets. Springer, Cham, pp 579?591","DOI":"10.1007\/978-1-4615-0005-6_16"},{"issue":"16","key":"10088_CR66","doi-asserted-by":"publisher","first-page":"6434","DOI":"10.3390\/su12166434","volume":"12","author":"C Dunn","year":"2020","unstructured":"Dunn C, Moustafa N, Turnbull B (2020) Robustness evaluations of sustainable machine learning models against data poisoning attacks in the internet of things. Sustainability 12(16):6434","journal-title":"Sustainability"},{"key":"10088_CR67","unstructured":"Dziugaite GK, Ben-David S, Roy DM (2020) Enforcing interpretability and its statistical impacts: trade-offs between accuracy and interpretability. arXiv preprint arXiv:201013764"},{"key":"10088_CR68","doi-asserted-by":"publisher","first-page":"113141","DOI":"10.1016\/j.dss.2019.113141","volume":"127","author":"C Eiras-Franco","year":"2019","unstructured":"Eiras-Franco C, Guijarro-Berdi\u00f1as B, Alonso-Betanzos A, Bahamonde A (2019) A scalable decision-tree-based method to explain interactions in dyadic data. Decis Support Syst 127:113141","journal-title":"Decis Support Syst"},{"key":"10088_CR69","unstructured":"Electronics (2019) Interpretable deep learning in electronics, computer science and medical imaging. https:\/\/www.mdpi.com\/journal\/electronics\/special_issues\/interpretable_deep_learning. Accessed 3 Mar 2020"},{"issue":"1","key":"10088_CR70","doi-asserted-by":"publisher","first-page":"157","DOI":"10.1007\/s10994-013-5337-8","volume":"98","author":"H Elghazel","year":"2015","unstructured":"Elghazel H, Aussem A (2015) Unsupervised feature selection with ensemble learning. Mach Learn 98(1):157?180","journal-title":"Mach Learn"},{"issue":"1","key":"10088_CR71","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1186\/s12911-019-0874-0","volume":"19","author":"R Elshawi","year":"2019","unstructured":"Elshawi R, Al-Mallah MH, Sakr S (2019) On the interpretability of machine learning-based model for predicting hypertension. BMC Med Inform Decis Mak 19(1):1?32","journal-title":"BMC Med Inform Decis Mak"},{"key":"10088_CR72","doi-asserted-by":"publisher","first-page":"121","DOI":"10.1016\/j.patcog.2016.03.028","volume":"58","author":"SM Erfani","year":"2016","unstructured":"Erfani SM, Rajasegarar S, Karunasekera S, Leckie C (2016) High-dimensional and large-scale anomaly detection using a linear one-class SVM with deep learning. Pattern Recogn 58:121?134","journal-title":"Pattern Recogn"},{"key":"10088_CR73","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-98131-4","volume-title":"Explainable and interpretable models in computer vision and machine learning","author":"HJ Escalante","year":"2018","unstructured":"Escalante HJ, Escalera S, Guyon I, Bar\u00f3 X, G\u00fc\u00e7l\u00fct\u00fcrk Y, G\u00fc\u00e7l\u00fc U, van Gerven M, van Lier R (2018) Explainable and interpretable models in computer vision and machine learning. Springer, Cham"},{"key":"10088_CR74","doi-asserted-by":"publisher","first-page":"1010","DOI":"10.1016\/j.promfg.2019.06.094","volume":"34","author":"CA Escobar","year":"2019","unstructured":"Escobar CA, Morales-Menendez R (2019) Process-monitoring-for-quality?a model selection criterion for support vector machine. Procedia Manuf 34:1010?1017","journal-title":"Procedia Manuf"},{"issue":"4","key":"10088_CR75","doi-asserted-by":"publisher","first-page":"1006","DOI":"10.1109\/TNNLS.2017.2648880","volume":"29","author":"X Fang","year":"2017","unstructured":"Fang X, Xu Y, Li X, Lai Z, Wong WK, Fang B (2017) Regularized label relaxation linear regression. IEEE Trans Neural Netwo Learn Syst 29(4):1006?1018","journal-title":"IEEE Trans Neural Netwo Learn Syst"},{"issue":"2","key":"10088_CR76","doi-asserted-by":"publisher","first-page":"71","DOI":"10.1109\/MRA.2019.2904644","volume":"26","author":"H Felzmann","year":"2019","unstructured":"Felzmann H, Fosch-Villaronga E, Lutz C, Tamo-Larrieux A (2019) Robots and transparency: the multiple dimensions of transparency in the context of robot technologies. IEEE Robotics Autom Mag 26(2):71?78","journal-title":"IEEE Robotics Autom Mag"},{"issue":"1","key":"10088_CR77","doi-asserted-by":"publisher","first-page":"69","DOI":"10.1109\/MCI.2018.2881645","volume":"14","author":"A Fernandez","year":"2019","unstructured":"Fernandez A, Herrera F, Cordon O, del Jesus MJ, Marcelloni F (2019) Evolutionary fuzzy systems for explainable artificial intelligence: why, when, what for, and where to? IEEE Comput Intell Mag 14(1):69?81","journal-title":"IEEE Comput Intell Mag"},{"key":"10088_CR78","unstructured":"FGCS (2021) Future generation computer systems. https:\/\/www.journals.elsevier.com\/future-generation-computer-systems\/call-for-papers\/explainable-artificial-intelligence-for-healthcare. Accessed 1 Aug 2021"},{"issue":"1","key":"10088_CR79","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1038\/s41598-020-79139-8","volume":"11","author":"JC Forte","year":"2021","unstructured":"Forte JC, Mungroop HE, de Geus F, van der Grinten ML, Bouma HR, Pettil\u00e4 V, Scheeren TW, Nijsten MW, Mariani MA, van der Horst IC et al (2021) Ensemble machine learning prediction and variable importance analysis of 5-year mortality after cardiac valve and CABG operations. Sci Rep 11(1):1?11","journal-title":"Sci Rep"},{"key":"10088_CR80","doi-asserted-by":"crossref","unstructured":"Fr\u00e4mling K (2020) Decision theory meets explainable AI. In: International workshop on explainable, transparent autonomous agents and multi-agent systems. Springer, Cham, pp 57?74","DOI":"10.1007\/978-3-030-51924-7_4"},{"key":"10088_CR81","doi-asserted-by":"publisher","first-page":"531","DOI":"10.1016\/j.patcog.2017.09.038","volume":"74","author":"AJ Gallego","year":"2018","unstructured":"Gallego AJ, Calvo-Zaragoza J, Valero-Mas JJ, Rico-Juan JR (2018) Clustering-based k-nearest neighbor classification for large-scale data with neural codes representation. Pattern Recogn 74:531?543","journal-title":"Pattern Recogn"},{"issue":"1","key":"10088_CR82","doi-asserted-by":"publisher","first-page":"190","DOI":"10.1016\/j.media.2015.06.008","volume":"24","author":"B Gaonkar","year":"2015","unstructured":"Gaonkar B, Shinohara RT, Davatzikos C, Initiative ADN et al (2015) Interpreting support vector machine models for multivariate group wise analysis in neuroimaging. Med Image Anal 24(1):190?204","journal-title":"Med Image Anal"},{"key":"10088_CR83","doi-asserted-by":"publisher","first-page":"125562","DOI":"10.1109\/ACCESS.2019.2937521","volume":"7","author":"I Garc\u00eda-Magari\u00f1o","year":"2019","unstructured":"Garc\u00eda-Magari\u00f1o I, Muttukrishnan R, Lloret J (2019) Human-centric AI for trustworthy IoT systems with explainable multilayer perceptrons. IEEE Access 7:125562?125574","journal-title":"IEEE Access"},{"key":"10088_CR84","unstructured":"Gartner (2020) Gartner identifies the top 10 strategic technology trends for 2020. https:\/\/www.gartner.com\/en\/newsroom\/press-releases\/2019-10-21-gartner-identifies-the-top-10-strategic-technology-trends-for-2020. Accessed 24 Jan 2020"},{"key":"10088_CR85","doi-asserted-by":"crossref","unstructured":"Ghorbani A, Abid A, Zou J (2019) Interpretation of neural networks is fragile. In: Proceedings of the AAAI conference on artificial intelligence, vol 33, pp 3681?3688","DOI":"10.1609\/aaai.v33i01.33013681"},{"key":"10088_CR86","doi-asserted-by":"publisher","first-page":"e340","DOI":"10.7717\/peerj-cs.340","volume":"7","author":"S Gite","year":"2021","unstructured":"Gite S, Khatavkar H, Kotecha K, Srivastava S, Maheshwari P, Pandey N (2021) Explainable stock prices prediction from financial news articles using sentiment analysis. PeerJ Comput Sci 7:e340","journal-title":"PeerJ Comput Sci"},{"key":"10088_CR87","unstructured":"Google (2021) Google what-if toolkit. https:\/\/pair-code.github.io\/what-if-tool\/. Accessed 26 Apr 2021"},{"key":"10088_CR88","doi-asserted-by":"crossref","unstructured":"Gronauer S, Diepold K (2021) Multi-agent deep reinforcement learning: a survey. Artif Intell Rev 1?49","DOI":"10.1007\/s10462-021-09996-w"},{"key":"10088_CR89","doi-asserted-by":"publisher","first-page":"101858","DOI":"10.1016\/j.artmed.2020.101858","volume":"107","author":"D Gu","year":"2020","unstructured":"Gu D, Su K, Zhao H (2020a) A case-based ensemble learning system for explainable breast cancer recurrence prediction. Artif Intell Med 107:101858","journal-title":"Artif Intell Med"},{"key":"10088_CR90","doi-asserted-by":"crossref","unstructured":"Gu R, Wang G, Song T, Huang R, Aertsen M, Deprest J, Ourselin S, Vercauteren T, Zhang S (2020b) Ca-net: comprehensive attention convolutional neural networks for explainable medical image segmentation. IEEE Trans Med Imaging","DOI":"10.1109\/TMI.2020.3035253"},{"issue":"5","key":"10088_CR91","doi-asserted-by":"publisher","first-page":"93","DOI":"10.1145\/3236009","volume":"51","author":"R Guidotti","year":"2019","unstructured":"Guidotti R, Monreale A, Ruggieri S, Turini F, Giannotti F, Pedreschi D (2019) A survey of methods for explaining black box models. ACM Comput Surv (CSUR) 51(5):93","journal-title":"ACM Comput Surv (CSUR)"},{"key":"10088_CR92","doi-asserted-by":"crossref","unstructured":"Gulati P, Hu Q, Atashzar SF (2021) Toward deep generalization of peripheral EMG-based human-robot interfacing: a hybrid explainable solution for neurorobotic systems. IEEE Robotics Autom Lett","DOI":"10.1109\/LRA.2021.3062320"},{"key":"10088_CR93","doi-asserted-by":"publisher","first-page":"95","DOI":"10.1016\/j.commatsci.2018.12.056","volume":"160","author":"S Guo","year":"2019","unstructured":"Guo S, Yu J, Liu X, Wang C, Jiang Q (2019) A predicting model for properties of steel using the industrial big data based on machine learning. Comput Mater Sci 160:95?104","journal-title":"Comput Mater Sci"},{"issue":"6","key":"10088_CR94","doi-asserted-by":"publisher","first-page":"39","DOI":"10.1109\/MCOM.001.2000050","volume":"58","author":"W Guo","year":"2020","unstructured":"Guo W (2020) Explainable artificial intelligence for 6G: improving trust between human and machine. IEEE Commun Mag 58(6):39?45","journal-title":"IEEE Commun Mag"},{"issue":"8","key":"10088_CR95","first-page":"15","volume":"163","author":"B Gupta","year":"2017","unstructured":"Gupta B, Rawat A, Jain A, Arora A, Dhami N (2017) Analysis of various decision tree algorithms for classification in data mining. Int J Comput Appl 163(8):15?19","journal-title":"Int J Comput Appl"},{"key":"10088_CR96","unstructured":"H2oai (2017) Comparative performance analysis of neural networks architectures on h2o platform for various activation functions. In: 2017 IEEE International young scientists forum on applied physics and engineering (YSF). IEEE, pp 70?73"},{"issue":"4","key":"10088_CR97","doi-asserted-by":"publisher","first-page":"482","DOI":"10.1109\/TPAMI.2005.78","volume":"27","author":"B Haasdonk","year":"2005","unstructured":"Haasdonk B (2005) Feature space interpretation of SVMs with indefinite kernels. IEEE Trans Pattern Anal Mach Intell 27(4):482?492","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"issue":"9","key":"10088_CR98","doi-asserted-by":"publisher","first-page":"28","DOI":"10.1109\/MC.2018.3620965","volume":"51","author":"H Hagras","year":"2018","unstructured":"Hagras H (2018) Toward human-understandable, explainable AI. Computer 51(9):28?36","journal-title":"Computer"},{"key":"10088_CR99","unstructured":"Hara S, Hayashi K (2018) Making tree ensembles interpretable: a Bayesian model selection approach. In: International conference on artificial intelligence and statistics. PMLR, pp 77?85"},{"key":"10088_CR100","doi-asserted-by":"publisher","first-page":"5747","DOI":"10.1007\/s10462-020-09833-6","volume":"53","author":"J Hatwell","year":"2020","unstructured":"Hatwell J, Gaber MM, Azad RMA (2020) Chirps: explaining random forest classification. Artif Intell Rev 53:5747?5788","journal-title":"Artif Intell Rev"},{"issue":"9","key":"10088_CR101","doi-asserted-by":"publisher","first-page":"4595","DOI":"10.1016\/j.eswa.2015.01.068","volume":"42","author":"I Hatzilygeroudis","year":"2015","unstructured":"Hatzilygeroudis I, Prentzas J (2015) Symbolic-neural rule based reasoning and explanation. Expert Syst Appl 42(9):4595?4609","journal-title":"Expert Syst Appl"},{"key":"10088_CR102","doi-asserted-by":"crossref","unstructured":"Hendricks LA, Akata Z, Rohrbach M, Donahue J, Schiele B, Darrell T (2016) Generating visual explanations. In: European conference on computer vision. Springer, Cham, pp 3?19","DOI":"10.1007\/978-3-319-46493-0_1"},{"issue":"5","key":"10088_CR103","doi-asserted-by":"publisher","first-page":"1503","DOI":"10.1007\/s10618-014-0368-8","volume":"28","author":"A Henelius","year":"2014","unstructured":"Henelius A, Puolam\u00e4ki K, Bostr\u00f6m H, Asker L, Papapetrou P (2014) A peek into the black box: exploring classifiers by randomization. Data Min Knowl Disc 28(5):1503?1529","journal-title":"Data Min Knowl Disc"},{"key":"10088_CR104","doi-asserted-by":"crossref","unstructured":"Hind M, Wei D, Campbell M, Codella NC, Dhurandhar A, Mojsilovi? A, Natesan\u00a0Ramamurthy K, Varshney KR (2019) TED: teaching AI to explain its decisions. In: Proceedings of the 2019 AAAI\/ACM conference on AI, ethics, and society, pp 123?129","DOI":"10.1145\/3306618.3314273"},{"key":"10088_CR105","unstructured":"Hoffman RR, Mueller ST, Klein G, Litman J (2018) Metrics for explainable AI: challenges and prospects. arXiv preprint arXiv:181204608"},{"issue":"2","key":"10088_CR106","doi-asserted-by":"publisher","first-page":"119","DOI":"10.1007\/s40708-016-0042-6","volume":"3","author":"A Holzinger","year":"2016","unstructured":"Holzinger A (2016) Interactive machine learning for health informatics: when do we need the human-in-the-loop? Brain Inform 3(2):119?131","journal-title":"Brain Inform"},{"issue":"4","key":"10088_CR107","doi-asserted-by":"publisher","first-page":"e1312","DOI":"10.1002\/widm.1312","volume":"9","author":"A Holzinger","year":"2019","unstructured":"Holzinger A, Langs G, Denk H, Zatloukal K, M\u00fcller H (2019) Causability and explainability of artificial intelligence in medicine. Wiley Interdiscip Rev Data Min Knowl Discov 9(4):e1312","journal-title":"Wiley Interdiscip Rev Data Min Knowl Discov"},{"key":"10088_CR108","doi-asserted-by":"publisher","first-page":"28","DOI":"10.1016\/j.inffus.2021.01.008","volume":"71","author":"A Holzinger","year":"2021","unstructured":"Holzinger A, Malle B, Saranti A, Pfeifer B (2021a) Towards multi-modal causability with graph neural networks enabling information fusion for explainable AI. Inf Fusion 71:28?37","journal-title":"Inf Fusion"},{"key":"10088_CR109","doi-asserted-by":"crossref","unstructured":"Holzinger A, Weippl E, Tjoa AM, Kieseberg P (2021b) Digital transformation for sustainable development goals (SDGS)?a security, safety and privacy perspective on AI. In: International cross-domain conference for machine learning and knowledge. Springer, Cham, pp 103?107","DOI":"10.1007\/978-3-030-84060-0_1"},{"key":"10088_CR110","doi-asserted-by":"crossref","unstructured":"Hu K, Orghian D, Hidalgo C (2018a) Dive: a mixed-initiative system supporting integrated data exploration workflows. In: Proceedings of the workshop on human-in-the-loop data analytics, pp 1?7","DOI":"10.1145\/3209900.3209910"},{"key":"10088_CR111","doi-asserted-by":"crossref","unstructured":"Hu R, Andreas J, Darrell T, Saenko K (2018b) Explainable neural computation via stack neural module networks. In: Proceedings of the European conference on computer vision (ECCV), pp 53?69","DOI":"10.1007\/978-3-030-01234-2_4"},{"key":"10088_CR112","doi-asserted-by":"crossref","unstructured":"Huang Q, Katsman I, He H, Gu Z, Belongie S, Lim SN (2019) Enhancing adversarial example transferability with an intermediate level attack. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 4733?4742","DOI":"10.1109\/ICCV.2019.00483"},{"key":"10088_CR113","doi-asserted-by":"crossref","unstructured":"Huisman M, van Rijn JN, Plaat A (2021) A survey of deep meta-learning. Artif Intell Rev 1?59","DOI":"10.1007\/s10462-021-10004-4"},{"key":"10088_CR114","doi-asserted-by":"crossref","unstructured":"IBM (2019) AI fairness 360: an extensible toolkit for detecting and mitigating algorithmic bias. IBM J Res Dev 63(4\/5):4?1","DOI":"10.1147\/JRD.2019.2942287"},{"key":"10088_CR115","unstructured":"ICAPS (2020) Explainable planning. https:\/\/icaps20.icaps-conference.org\/workshops\/xaip\/. Accessed 3 Mar 2020"},{"key":"10088_CR116","unstructured":"ICCV (2019) Interpretating and explaining visual artificial intelligence models. http:\/\/xai.unist.ac.kr\/workshop\/2019\/. Accessed 3 Mar 2020"},{"key":"10088_CR117","unstructured":"ICML (2021) Theoretic foundation, criticism, and application trend of explainable AI. https:\/\/icml2021-xai.github.io\/. Accessed 1 Aug 2021"},{"key":"10088_CR118","unstructured":"IDC (2020) Worldwide spending on artificial intelligence systems will be nearly 98 billion dollars in 2023. https:\/\/www.idc.com\/getdoc.jsp?containerId=prUS45481219. Accessed 24 Jan 2020"},{"key":"10088_CR119","unstructured":"IJCAI (2019) Explainable artificial intelligence(XAI). https:\/\/sites.google.com\/view\/xai2019\/home. Accessed 3 Mar 2020"},{"issue":"7","key":"10088_CR120","doi-asserted-by":"publisher","first-page":"1291","DOI":"10.1109\/TFUZZ.2019.2917124","volume":"28","author":"MA Islam","year":"2019","unstructured":"Islam MA, Anderson DT, Pinar AJ, Havens TC, Scott G, Keller JM (2019) Enabling explainable fusion in deep learning with fuzzy integral neural networks. IEEE Trans Fuzzy Syst 28(7):1291?1300","journal-title":"IEEE Trans Fuzzy Syst"},{"key":"10088_CR121","doi-asserted-by":"publisher","first-page":"25195","DOI":"10.1109\/ACCESS.2019.2899901","volume":"7","author":"NU Islam","year":"2019","unstructured":"Islam NU, Lee S (2019) Interpretation of deep CNN based on learning feature reconstruction with feedback weights. IEEE Access 7:25195?25208","journal-title":"IEEE Access"},{"key":"10088_CR122","unstructured":"IUI (2019) Explainable smart systems. https:\/\/explainablesystems.comp.nus.edu.sg\/2019\/. Accessed 3 Mar 2020"},{"key":"10088_CR123","doi-asserted-by":"crossref","unstructured":"Ivanovs M, Kadikis R, Ozols K (2021) Perturbation-based methods for explaining deep neural networks: a survey. Pattern Recognit Lett","DOI":"10.1016\/j.patrec.2021.06.030"},{"issue":"7","key":"10088_CR124","doi-asserted-by":"publisher","first-page":"86","DOI":"10.1145\/2611567","volume":"57","author":"H Jagadish","year":"2014","unstructured":"Jagadish H, Gehrke J, Labrinidis A, Papakonstantinou Y, Patel JM, Ramakrishnan R, Shahabi C (2014) Big data and its technical challenges. Commun ACM 57(7):86?94","journal-title":"Commun ACM"},{"issue":"4","key":"10088_CR125","doi-asserted-by":"publisher","first-page":"885","DOI":"10.1007\/s11634-016-0276-4","volume":"12","author":"S Janitza","year":"2018","unstructured":"Janitza S, Celik E, Boulesteix AL (2018) A computationally fast variable importance test for random forests for high-dimensional data. Adv Data Anal Classif 12(4):885?915","journal-title":"Adv Data Anal Classif"},{"key":"10088_CR126","doi-asserted-by":"publisher","first-page":"18670","DOI":"10.1109\/ACCESS.2021.3051171","volume":"9","author":"YJ Jung","year":"2021","unstructured":"Jung YJ, Han SH, Choi HJ (2021) Explaining CNN and RNN using selective layer-wise relevance propagation. IEEE Access 9:18670?18681","journal-title":"IEEE Access"},{"key":"10088_CR127","doi-asserted-by":"publisher","first-page":"174","DOI":"10.1016\/j.neunet.2020.05.035","volume":"129","author":"JRB Junior","year":"2020","unstructured":"Junior JRB (2020) Graph embedded rules for explainable predictions in data streams. Neural Netw 129:174?192","journal-title":"Neural Netw"},{"key":"10088_CR128","doi-asserted-by":"crossref","unstructured":"Juuti M, Szyller S, Marchal S, Asokan N (2019) PRADA: protecting against DNN model stealing attacks. In: 2019 IEEE European symposium on security and privacy (EuroS&P). IEEE, pp 512?527","DOI":"10.1109\/EuroSP.2019.00044"},{"issue":"1","key":"10088_CR129","doi-asserted-by":"publisher","first-page":"13","DOI":"10.1109\/TLT.2018.2789900","volume":"11","author":"A Kapelner","year":"2018","unstructured":"Kapelner A, Soterwood J, Nessaiver S, Adlof S (2018) Predicting contextual informativeness for vocabulary learning. IEEE Trans Learn Technol 11(1):13?26","journal-title":"IEEE Trans Learn Technol"},{"issue":"5","key":"10088_CR130","doi-asserted-by":"publisher","first-page":"1671","DOI":"10.1007\/s10115-019-01389-4","volume":"62","author":"I Karlsson","year":"2020","unstructured":"Karlsson I, Rebane J, Papapetrou P, Gionis A (2020) Locally and globally explainable time series tweaking. Knowl Inf Syst 62(5):1671?1700","journal-title":"Knowl Inf Syst"},{"key":"10088_CR131","doi-asserted-by":"crossref","unstructured":"Keane MT, Kenny EM (2019) How case-based reasoning explains neural networks: A theoretical analysis of XAI using post-hoc explanation-by-example from a survey of ANN-CBR twin-systems. In: International conference on case-based reasoning. Springer, Cham, pp 155?171","DOI":"10.1007\/978-3-030-29249-2_11"},{"key":"10088_CR132","doi-asserted-by":"publisher","first-page":"17001","DOI":"10.1109\/ACCESS.2019.2893141","volume":"7","author":"BM Keneni","year":"2019","unstructured":"Keneni BM, Kaur D, Al Bataineh A, Devabhaktuni VK, Javaid AY, Zaientz JD, Marinier RP (2019) Evolving rule-based explainable artificial intelligence for unmanned aerial vehicles. IEEE Access 7:17001?17016","journal-title":"IEEE Access"},{"key":"10088_CR133","doi-asserted-by":"publisher","first-page":"103459","DOI":"10.1016\/j.artint.2021.103459","volume":"294","author":"EM Kenny","year":"2021","unstructured":"Kenny EM, Ford C, Quinn M, Keane MT (2021) Explaining black-box classifiers using post-hoc explanations-by-example: the effect of explanations and error-rates in XAI user studies. Artif Intell 294:103459","journal-title":"Artif Intell"},{"key":"10088_CR134","doi-asserted-by":"crossref","unstructured":"Kim J, Canny J (2018) Explainable deep driving by visualizing causal attention. In: Explainable and interpretable models in computer vision and machine learning. Springer, Cham, pp 173?193","DOI":"10.1007\/978-3-319-98131-4_8"},{"key":"10088_CR135","doi-asserted-by":"crossref","unstructured":"Kindermans PJ, Hooker S, Adebayo J, Alber M, Sch\u00fctt KT, D\u00e4hne S, Erhan D, Kim B (2019) The (un) reliability of saliency methods. In: Explainable AI: interpreting, explaining and visualizing deep learning. Springer, Cham, pp 267?280","DOI":"10.1007\/978-3-030-28954-6_14"},{"key":"10088_CR136","doi-asserted-by":"crossref","unstructured":"Kiritz N, Sarfati P (2018) Supervisory guidance on model risk management (SR 11-7) versus enterprise-wide model risk management for deposit-taking institutions (E-23): a detailed comparative analysis. Available at SSRN 3332484","DOI":"10.2139\/ssrn.3332484"},{"key":"10088_CR137","unstructured":"Koh PW, Liang P (2017) Understanding black-box predictions via influence functions. In: International conference on machine learning. PMLR, pp 1885?1894"},{"key":"10088_CR138","doi-asserted-by":"publisher","first-page":"18","DOI":"10.3389\/fdata.2021.660206","volume":"4","author":"I Kolyshkina","year":"2021","unstructured":"Kolyshkina I, Simoff S (2021) Interpretability of machine learning solutions in public healthcare: the CRISP-ML approach. Front Big Data 4:18","journal-title":"Front Big Data"},{"key":"10088_CR139","doi-asserted-by":"crossref","unstructured":"Konig R, Johansson U, Niklasson L (2008) G-REX: a versatile framework for evolutionary data mining. In: 2008 IEEE international conference on data mining workshops. IEEE, pp 971?974","DOI":"10.1109\/ICDMW.2008.117"},{"key":"10088_CR140","doi-asserted-by":"publisher","first-page":"106993","DOI":"10.1016\/j.knosys.2021.106993","volume":"222","author":"AV Konstantinov","year":"2021","unstructured":"Konstantinov AV, Utkin LV (2021) Interpretable machine learning with an ensemble of gradient boosting machines. Knowl Based Syst 222:106993","journal-title":"Knowl Based Syst"},{"key":"10088_CR141","doi-asserted-by":"publisher","first-page":"202","DOI":"10.1016\/j.ins.2021.02.031","volume":"564","author":"P Krishnamurthy","year":"2021","unstructured":"Krishnamurthy P, Sarmadi A, Khorrami F (2021) Explainable classification by learning human-readable sentences in feature subsets. Inf Sci 564:202?219","journal-title":"Inf Sci"},{"key":"10088_CR142","doi-asserted-by":"crossref","unstructured":"Kumari B, Swarnkar T (2020) Importance of data standardization methods on stock indices prediction accuracy. In: Advanced computing and intelligent engineering. Springer, Cham, pp 309?318","DOI":"10.1007\/978-981-15-1081-6_26"},{"key":"10088_CR143","doi-asserted-by":"publisher","first-page":"346","DOI":"10.1016\/j.jvcir.2019.03.010","volume":"60","author":"CCJ Kuo","year":"2019","unstructured":"Kuo CCJ, Zhang M, Li S, Duan J, Chen Y (2019) Interpretable convolutional neural networks via feedforward design. J Vis Commun Image Represent 60:346?359","journal-title":"J Vis Commun Image Represent"},{"key":"10088_CR144","doi-asserted-by":"publisher","first-page":"103473","DOI":"10.1016\/j.artint.2021.103473","volume":"296","author":"M Langer","year":"2021","unstructured":"Langer M, Oster D, Speith T, Hermanns H, K\u00e4stner L, Schmidt E, Sesing A, Baum K (2021) What do we want from explainable artificial intelligence (XAI)??A stakeholder perspective on XAI and a conceptual model guiding interdisciplinary XAI research. Artif Intell 296:103473","journal-title":"Artif Intell"},{"issue":"1","key":"10088_CR145","doi-asserted-by":"publisher","first-page":"4","DOI":"10.1007\/s12975-017-0531-9","volume":"9","author":"PA Lapchak","year":"2018","unstructured":"Lapchak PA, Zhang JH (2018) Data standardization and quality management. Transl Stroke Res 9(1):4?8","journal-title":"Transl Stroke Res"},{"issue":"1","key":"10088_CR146","first-page":"3938","volume":"17","author":"S Lapuschkin","year":"2016","unstructured":"Lapuschkin S, Binder A, Montavon G, M\u00fcller KR, Samek W (2016) The LRP toolbox for artificial neural networks. J Mach Learn Res 17(1):3938?3942","journal-title":"J Mach Learn Res"},{"issue":"1","key":"10088_CR147","doi-asserted-by":"publisher","first-page":"98","DOI":"10.1080\/10618600.2017.1349663","volume":"27","author":"P Latouche","year":"2018","unstructured":"Latouche P, Robin S, Ouadah S (2018) Goodness of fit of logistic regression models for random graphs. J Comput Graph Stat 27(1):98?109","journal-title":"J Comput Graph Stat"},{"issue":"1","key":"10088_CR148","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1038\/s41467-020-17431-x","volume":"11","author":"SM Lauritsen","year":"2020","unstructured":"Lauritsen SM, Kristensen M, Olsen MV, Larsen MS, Lauritsen KM, J\u00f8rgensen MJ, Lange J, Thiesson B (2020) Explainable artificial intelligence model to predict acute critical illness from electronic health records. Nat Commun 11(1):1?11","journal-title":"Nat Commun"},{"key":"10088_CR149","doi-asserted-by":"crossref","unstructured":"Lawless WF, Mittu R, Sofge D, Hiatt L (2019) Artificial intelligence, autonomy, and human-machine teams: interdependence, context, and explainable AI. AI Mag 40(3)","DOI":"10.1609\/aimag.v40i3.2866"},{"key":"10088_CR150","doi-asserted-by":"crossref","unstructured":"Lee D, Mulrow J, Haboucha CJ, Derrible S, Shiftan Y (2019) Attitudes on autonomous vehicle adoption using interpretable gradient boosting machine. Transp Res Rec, p 0361198119857953","DOI":"10.1177\/0361198119857953"},{"key":"10088_CR151","doi-asserted-by":"publisher","first-page":"106","DOI":"10.1016\/j.enbuild.2015.09.002","volume":"108","author":"K Li","year":"2015","unstructured":"Li K, Hu C, Liu G, Xue W (2015) Building?s electricity consumption prediction using optimized artificial neural networks and principal component analysis. Energy Build 108:106?113","journal-title":"Energy Build"},{"key":"10088_CR152","doi-asserted-by":"crossref","unstructured":"Liang S, Sabri AQM, Alnajjar F, Loo CK (2021) Autism spectrum self-stimulatory behaviours classification using explainable temporal coherency deep features and SVM classifier. IEEE Access","DOI":"10.1109\/ACCESS.2021.3061455"},{"issue":"1","key":"10088_CR153","doi-asserted-by":"publisher","first-page":"121","DOI":"10.1007\/s11634-015-0213-y","volume":"11","author":"C Liberati","year":"2017","unstructured":"Liberati C, Camillo F, Saporta G (2017) Advances in credit scoring: combining performance and interpretation in kernel discriminant analysis. Adv Data Anal Classif 11(1):121?138","journal-title":"Adv Data Anal Classif"},{"key":"10088_CR154","doi-asserted-by":"crossref","unstructured":"Lin YC, Lee YC, Tsai WC, Beh WK, Wu AYA (2020) Explainable deep neural network for identifying cardiac abnormalities using class activation map. In: 2020 Computing in cardiology. IEEE, pp 1?4","DOI":"10.22489\/CinC.2020.072"},{"issue":"3","key":"10088_CR155","doi-asserted-by":"publisher","first-page":"31","DOI":"10.1145\/3236386.3241340","volume":"16","author":"ZC Lipton","year":"2018","unstructured":"Lipton ZC (2018) The mythos of model interpretability. Queue 16(3):31?57","journal-title":"Queue"},{"issue":"7","key":"10088_CR156","doi-asserted-by":"publisher","first-page":"1269","DOI":"10.1109\/TMM.2016.2557061","volume":"18","author":"YJ Liu","year":"2016","unstructured":"Liu YJ, Ma C, Zhao G, Fu X, Wang H, Dai G, Xie L (2016) An interactive spiraltape video summarization. IEEE Trans Multimed 18(7):1269?1282","journal-title":"IEEE Trans Multimed"},{"key":"10088_CR157","doi-asserted-by":"publisher","first-page":"S34","DOI":"10.1016\/j.jbi.2017.05.023","volume":"75","author":"Z Liu","year":"2017","unstructured":"Liu Z, Tang B, Wang X, Chen Q (2017) De-identification of clinical notes via recurrent neural network and conditional random field. J Biomed Inform 75:S34?S42","journal-title":"J Biomed Inform"},{"issue":"6","key":"10088_CR158","doi-asserted-by":"publisher","first-page":"102099","DOI":"10.1016\/j.ipm.2019.102099","volume":"57","author":"P Liu","year":"2020","unstructured":"Liu P, Zhang L, Gulla JA (2020) Dynamic attention-based explainable recommendation with textual and visual fusion. Inf Process Manag 57(6):102099","journal-title":"Inf Process Manag"},{"issue":"12","key":"10088_CR159","doi-asserted-by":"publisher","first-page":"3071","DOI":"10.1109\/TPAMI.2018.2868685","volume":"41","author":"M Long","year":"2018","unstructured":"Long M, Cao Y, Cao Z, Wang J, Jordan MI (2018) Transferable representation learning with deep adaptation networks. IEEE Trans Pattern Anal Mach Intell 41(12):3071?3085","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"issue":"1","key":"10088_CR160","doi-asserted-by":"publisher","first-page":"1483","DOI":"10.2991\/ijcis.d.200910.002","volume":"13","author":"M Loor","year":"2020","unstructured":"Loor M, De Tr\u00e9 G (2020) Contextualizing support vector machine predictions. Int J Comput Intell Syst 13(1):1483?1497","journal-title":"Int J Comput Intell Syst"},{"key":"10088_CR161","doi-asserted-by":"publisher","first-page":"179","DOI":"10.1016\/j.neucom.2015.03.112","volume":"174","author":"X Luo","year":"2016","unstructured":"Luo X, Chang X, Ban X (2016) Regression and classification using extreme learning machine based on L1-norm and L2-norm. Neurocomputing 174:179?186","journal-title":"Neurocomputing"},{"issue":"2","key":"10088_CR162","doi-asserted-by":"publisher","first-page":"161","DOI":"10.1007\/s41095-017-0077-5","volume":"3","author":"Y Ma","year":"2017","unstructured":"Ma Y, Chen W, Ma X, Xu J, Huang X, Maciejewski R, Tung AK (2017) EasySVM: a visual analysis approach for open-box support vector machines. Comput Vis Media 3(2):161?175","journal-title":"Comput Vis Media"},{"issue":"12","key":"10088_CR163","doi-asserted-by":"publisher","first-page":"4797","DOI":"10.1021\/acs.molpharmaceut.9b00520","volume":"16","author":"M Manica","year":"2019","unstructured":"Manica M, Oskooei A, Born J, Subramanian V, S\u00e1ez-Rodr\u00edguez J, Rodriguez Martinez M (2019) Toward explainable anticancer compound sensitivity prediction via multimodal attention-based convolutional encoders. Mol Pharm 16(12):4797?4806","journal-title":"Mol Pharm"},{"issue":"1","key":"10088_CR164","doi-asserted-by":"publisher","first-page":"64","DOI":"10.2106\/JBJS.20.00875","volume":"103","author":"ML Martini","year":"2021","unstructured":"Martini ML, Neifert SN, Gal JS, Oermann EK, Gilligan JT, Caridi JM (2021) Drivers of prolonged hospitalization following spine surgery: a game-theory-based approach to explaining machine learning models. JBJS 103(1):64?73","journal-title":"JBJS"},{"key":"10088_CR165","doi-asserted-by":"crossref","unstructured":"Maweu BM, Dakshit S, Shamsuddin R, Prabhakaran B (2021) CEFEs: a CNN explainable framework for ECG signals. Artif Intell Med 102059","DOI":"10.1016\/j.artmed.2021.102059"},{"key":"10088_CR166","doi-asserted-by":"crossref","unstructured":"Meske C, Bunde E, Schneider J, Gersch M (2020) Explainable artificial intelligence: objectives, stakeholders, and future research opportunities. Inf Syst Manag 1?11","DOI":"10.1080\/10580530.2020.1849465"},{"key":"10088_CR167","unstructured":"Microsoft (2021) Azure model interpretability. https:\/\/docs.microsoft.com\/en-us\/azure\/machine-learning\/how-to-machine-learning-interpretability. Accessed 26 Apr 2021"},{"key":"10088_CR168","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.artint.2018.07.007","volume":"267","author":"T Miller","year":"2019","unstructured":"Miller T (2019) Explanation in artificial intelligence: insights from the social sciences. Artif Intell 267:1?38","journal-title":"Artif Intell"},{"key":"10088_CR169","doi-asserted-by":"publisher","first-page":"55392","DOI":"10.1109\/ACCESS.2018.2868970","volume":"6","author":"DL Minh","year":"2018","unstructured":"Minh DL, Sadeghi-Niaraki A, Huy HD, Min K, Moon H (2018) Deep learning approach for short-term stock trends prediction based on two-stream gated recurrent unit network. IEEE Access 6:55392?55404","journal-title":"IEEE Access"},{"key":"10088_CR170","doi-asserted-by":"crossref","unstructured":"Mohit, Kumari AC, Sharma M (2019) A novel approach to text clustering using shift k-medoid. Int J Soc Comput Cyber Phys Syst 2(2):106?118","DOI":"10.1504\/IJSCCPS.2019.10021894"},{"key":"10088_CR171","doi-asserted-by":"crossref","unstructured":"Molnar C, Casalicchio G, Bischl B (2019) Quantifying model complexity via functional decomposition for better post-hoc interpretability. In: Joint European conference on machine learning and knowledge discovery in databases. Springer, Cham, pp 193?204","DOI":"10.1007\/978-3-030-43823-4_17"},{"key":"10088_CR172","doi-asserted-by":"publisher","first-page":"211","DOI":"10.1016\/j.patcog.2016.11.008","volume":"65","author":"G Montavon","year":"2017","unstructured":"Montavon G, Lapuschkin S, Binder A, Samek W, M\u00fcller KR (2017) Explaining nonlinear classification decisions with deep Taylor decomposition. Pattern Recogn 65:211?222","journal-title":"Pattern Recogn"},{"key":"10088_CR173","doi-asserted-by":"publisher","first-page":"113941","DOI":"10.1016\/j.eswa.2020.113941","volume":"165","author":"M Moradi","year":"2021","unstructured":"Moradi M, Samwald M (2021) Post-hoc explanation of black-box classifiers using confident itemsets. Expert Syst Appl 165:113941","journal-title":"Expert Syst Appl"},{"key":"10088_CR174","unstructured":"Mordvintsev A, Olah C, Tyka M (2015) Inceptionism: going deeper into neural networks, 2015. https:\/\/research.googleblog.com\/2015\/06\/inceptionism-going-deeper-into-neural.html"},{"issue":"07","key":"10088_CR175","doi-asserted-by":"publisher","first-page":"119","DOI":"10.1109\/MC.2021.3074263","volume":"54","author":"H Muller","year":"2021","unstructured":"Muller H, Mayrhofer MT, Van Veen EB, Holzinger A (2021) The ten commandments of ethical medical AI. Computer 54(07):119?123","journal-title":"Computer"},{"key":"10088_CR176","doi-asserted-by":"crossref","unstructured":"Musto C, de\u00a0Gemmis M, Lops P, Semeraro G (2020) Generating post hoc review-based natural language justifications for recommender systems. User Model User Adapt Interact 1?45","DOI":"10.1007\/s11257-020-09270-8"},{"key":"10088_CR177","doi-asserted-by":"crossref","unstructured":"Neto MP, Paulovich FV (2020) Explainable matrix?visualization for global and local interpretability of random forest classification ensembles. IEEE Trans Vis Comput Graph","DOI":"10.1109\/TVCG.2020.3030354"},{"issue":"2","key":"10088_CR178","doi-asserted-by":"publisher","first-page":"1","DOI":"10.24191\/srj.v15i2.9347","volume":"15","author":"SF Ng","year":"2018","unstructured":"Ng SF, Chew YM, Chng PE, Ng KS (2018) An insight of linear regression analysis. Sci Res J 15(2):1?16","journal-title":"Sci Res J"},{"key":"10088_CR179","doi-asserted-by":"publisher","first-page":"506","DOI":"10.1016\/j.cma.2019.05.052","volume":"354","author":"TN Nguyen","year":"2019","unstructured":"Nguyen TN, Lee S, Nguyen-Xuan H, Lee J (2019) A novel analysis-prediction approach for geometrically nonlinear problems using group method of data handling. Comput Methods Appl Mech Eng 354:506?526","journal-title":"Comput Methods Appl Mech Eng"},{"key":"10088_CR180","unstructured":"Nguyen DT, Kasmarik KE, Abbass HA (2020a) Towards interpretable neural networks: an exact transformation to multi-class multivariate decision trees. arXiv preprint arXiv:200304675"},{"key":"10088_CR181","doi-asserted-by":"publisher","first-page":"103377","DOI":"10.1016\/j.finel.2019.103377","volume":"171","author":"TN Nguyen","year":"2020","unstructured":"Nguyen TN, Nguyen-Xuan H, Lee J (2020b) A novel data-driven nonlinear solver for solid mechanics using time series forecasting. Finite Elem Anal Des 171:103377","journal-title":"Finite Elem Anal Des"},{"key":"10088_CR182","unstructured":"NIPS (2017) Interpreting, explaining and visualizing deep learning. http:\/\/www.interpretable-ml.org\/nips2017workshop\/. Accessed 3 Mar 2020"},{"key":"10088_CR183","doi-asserted-by":"publisher","first-page":"64","DOI":"10.1016\/j.eswa.2019.02.012","volume":"126","author":"J Obregon","year":"2019","unstructured":"Obregon J, Kim A, Jung JY (2019) RuleCOSI: combination and simplification of production rules from boosted decision trees for imbalanced classification. Expert Syst Appl 126:64?82","journal-title":"Expert Syst Appl"},{"issue":"3","key":"10088_CR184","doi-asserted-by":"publisher","first-page":"e10","DOI":"10.23915\/distill.00010","volume":"3","author":"C Olah","year":"2018","unstructured":"Olah C, Satyanarayan A, Johnson I, Carter S, Schubert L, Ye K, Mordvintsev A (2018) The building blocks of interpretability. Distill 3(3):e10","journal-title":"Distill"},{"key":"10088_CR185","unstructured":"Oracle (2021) Oracle skater. https:\/\/oracle.github.io\/Skater\/overview.html. Accessed 26 Apr 2021"},{"issue":"2","key":"10088_CR186","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s13201-020-01356-3","volume":"11","author":"K Ostad-Ali-Askari","year":"2021","unstructured":"Ostad-Ali-Askari K, Shayannejad M (2021) Computation of subsurface drain spacing in the unsteady conditions using artificial neural networks (ANN). Appl Water Sci 11(2):1?9","journal-title":"Appl Water Sci"},{"issue":"1","key":"10088_CR187","doi-asserted-by":"publisher","first-page":"134","DOI":"10.1007\/s12205-016-0572-8","volume":"21","author":"K Ostad-Ali-Askari","year":"2017","unstructured":"Ostad-Ali-Askari K, Shayannejad M, Ghorbanizadeh-Kharazi H (2017) Artificial neural network for modeling nitrate pollution of groundwater in marginal area of Zayandeh-rood river, Isfahan, Iran. KSCE J Civ Eng 21(1):134?140","journal-title":"KSCE J Civ Eng"},{"issue":"1","key":"10088_CR188","doi-asserted-by":"publisher","first-page":"e1968","DOI":"10.1002\/rcs.1968","volume":"15","author":"S Osullivan","year":"2019","unstructured":"Osullivan S, Nevejans N, Allen C, Blyth A, Leonard S, Pagallo U, Holzinger K, Holzinger A, Sajid MI, Ashrafian H (2019) Legal, regulatory, and ethical frameworks for development of standards in artificial intelligence (AI) and autonomous robotic surgery. Int J Med Robotics Comput Assist Surg 15(1):e1968","journal-title":"Int J Med Robotics Comput Assist Surg"},{"issue":"2","key":"10088_CR189","doi-asserted-by":"publisher","first-page":"389","DOI":"10.5194\/soil-6-389-2020","volume":"6","author":"J Padarian","year":"2020","unstructured":"Padarian J, McBratney AB, Minasny B (2020) Game theory interpretation of digital soil mapping convolutional neural networks. Soil 6(2):389?397","journal-title":"Soil"},{"issue":"3","key":"10088_CR190","doi-asserted-by":"publisher","first-page":"441","DOI":"10.1007\/s11023-019-09502-w","volume":"29","author":"A P\u00e1ez","year":"2019","unstructured":"P\u00e1ez A (2019) The pragmatic turn in explainable artificial intelligence (XAI). Mind Mach 29(3):441?459","journal-title":"Mind Mach"},{"key":"10088_CR191","unstructured":"Pan X, Tang F, Dong W, Ma C, Meng Y, Huang F, Lee TY, Xu C (2019) Content-based visual summarization for image collections. IEEE Transa Vis Comput Graph"},{"key":"10088_CR192","doi-asserted-by":"crossref","unstructured":"Park DH, Hendricks LA, Akata Z, Rohrbach A, Schiele B, Darrell T, Rohrbach M (2018) Multimodal explanations: justifying decisions and pointing to the evidence. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 8779?8788","DOI":"10.1109\/CVPR.2018.00915"},{"key":"10088_CR193","doi-asserted-by":"publisher","first-page":"207","DOI":"10.1016\/j.media.2019.03.007","volume":"54","author":"C Payer","year":"2019","unstructured":"Payer C, Stern D, Bischof H, Urschler M (2019) Integrating spatial configuration into heatmap regression based CNNs for landmark localization. Med Image Anal 54:207?219","journal-title":"Med Image Anal"},{"issue":"6","key":"10088_CR194","doi-asserted-by":"publisher","first-page":"697","DOI":"10.1038\/s41431-020-0596-x","volume":"28","author":"D Peloquin","year":"2020","unstructured":"Peloquin D, DiMaio M, Bierer B, Barnes M (2020) Disruptive and avoidable: GDPR challenges to secondary research uses of data. Eur J Hum Genet 28(6):697?705","journal-title":"Eur J Hum Genet"},{"key":"10088_CR195","doi-asserted-by":"publisher","first-page":"113","DOI":"10.1016\/j.neucom.2018.11.094","volume":"342","author":"M Polato","year":"2019","unstructured":"Polato M, Aiolli F (2019) Boolean kernels for rule based interpretation of support vector machines. Neurocomputing 342:113?124","journal-title":"Neurocomputing"},{"key":"10088_CR196","unstructured":"PR (2019) Explainable deep learning for efficient and robust pattern recognition. https:\/\/www.journals.elsevier.com\/pattern-recognition\/call-for-papers\/call-for-paper-on-special-issue-on-explainable-deep-learning. Accessed 3 Mar 2020"},{"issue":"5","key":"10088_CR197","doi-asserted-by":"publisher","first-page":"74","DOI":"10.1109\/MSEC.2019.2925649","volume":"17","author":"S Raaijmakers","year":"2019","unstructured":"Raaijmakers S (2019) Artificial intelligence for law enforcement: challenges and opportunities. IEEE Secur Priv 17(5):74?77","journal-title":"IEEE Secur Priv"},{"issue":"1","key":"10088_CR198","doi-asserted-by":"publisher","first-page":"137","DOI":"10.1007\/s11747-019-00710-5","volume":"48","author":"A Rai","year":"2020","unstructured":"Rai A (2020) Explainable AI: from black box to glass box. J Acad Mark Sci 48(1):137?141","journal-title":"J Acad Mark Sci"},{"key":"10088_CR199","doi-asserted-by":"publisher","first-page":"221","DOI":"10.1016\/j.ins.2020.05.126","volume":"540","author":"D Rajapaksha","year":"2020","unstructured":"Rajapaksha D, Bergmeir C, Buntine W (2020) LoRMIkA: local rule-based model interpretability with k-optimal associations. Inf Sci 540:221?241","journal-title":"Inf Sci"},{"issue":"1","key":"10088_CR200","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1038\/s41746-018-0029-1","volume":"1","author":"A Rajkomar","year":"2018","unstructured":"Rajkomar A, Oren E, Chen K, Dai AM, Hajaj N, Hardt M, Liu PJ, Liu X, Marcus J, Sun M et al (2018) Scalable and accurate deep learning with electronic health records. NPJ Digit Med 1(1):1?10","journal-title":"NPJ Digit Med"},{"issue":"3","key":"10088_CR201","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3324916","volume":"28","author":"X Ren","year":"2019","unstructured":"Ren X, Xing Z, Xia X, Lo D, Wang X, Grundy J (2019) Neural network-based detection of self-admitted technical debt: from performance to explainability. ACM Trans Softw Eng Methodol (TOSEM) 28(3):1?45","journal-title":"ACM Trans Softw Eng Methodol (TOSEM)"},{"key":"10088_CR202","doi-asserted-by":"crossref","unstructured":"Ribeiro MT, Singh S, Guestrin C (2016) ?Why should I trust you?? explaining the predictions of any classifier. In: Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining, pp 1135?1144","DOI":"10.1145\/2939672.2939778"},{"key":"10088_CR203","doi-asserted-by":"publisher","first-page":"23","DOI":"10.1016\/j.cag.2019.08.012","volume":"85","author":"PC Ribeiro","year":"2019","unstructured":"Ribeiro PC, Schardong GG, Barbosa SD, de Souza CS, Lopes H (2019) Visual exploration of an ensemble of classifiers. Comput Graph 85:23?41","journal-title":"Comput Graph"},{"issue":"5","key":"10088_CR204","doi-asserted-by":"publisher","first-page":"206","DOI":"10.1038\/s42256-019-0048-x","volume":"1","author":"C Rudin","year":"2019","unstructured":"Rudin C (2019) Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nat Mach Intell 1(5):206?215","journal-title":"Nat Mach Intell"},{"key":"10088_CR205","doi-asserted-by":"crossref","unstructured":"Sabol P, Sin\u010d\u00e1k P, Hartono P, Ko\u010dan P, Benetinov\u00e1 Z, Blich\u00e1rov\u00e1 A, Verb\u00f3ov\u00e1 \u013d, \u0160tammov\u00e1 E, Sabolov\u00e1-Fabianov\u00e1 A, Ja\u0161kov\u00e1 A (2020) Explainable classifier for improving the accountability in decision-making for colorectal cancer diagnosis from histopathological images. J Biomed Inform 109:103523","DOI":"10.1016\/j.jbi.2020.103523"},{"key":"10088_CR206","doi-asserted-by":"publisher","first-page":"124","DOI":"10.1016\/j.inffus.2020.03.013","volume":"61","author":"O Sagi","year":"2020","unstructured":"Sagi O, Rokach L (2020) Explainable decision forest: transforming a decision forest into an interpretable tree. Inf Fusion 61:124?138","journal-title":"Inf Fusion"},{"key":"10088_CR207","doi-asserted-by":"crossref","unstructured":"Salmeron JL, Correia MB, Palos-Sanchez PR (2019) Complexity in forecasting and predictive models. Complexity 2019","DOI":"10.1155\/2019\/8160659"},{"issue":"1","key":"10088_CR208","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1186\/s12859-018-2451-4","volume":"19","author":"H Sanz","year":"2018","unstructured":"Sanz H, Valim C, Vegas E, Oller JM, Reverter F (2018) SVM-RFE: selection and visualization of the most relevant features through non-linear kernels. BMC Bioinform 19(1):1?18","journal-title":"BMC Bioinform"},{"issue":"1","key":"10088_CR209","doi-asserted-by":"publisher","first-page":"21","DOI":"10.1109\/TVCG.2016.2598466","volume":"23","author":"A Sarvghad","year":"2016","unstructured":"Sarvghad A, Tory M, Mahyar N (2016) Visualizing dimension coverage to support exploratory analysis. IEEE Trans Visual Comput Graph 23(1):21?30","journal-title":"IEEE Trans Visual Comput Graph"},{"key":"10088_CR210","doi-asserted-by":"crossref","unstructured":"Schneeberger D, St\u00f6ger K, Holzinger A (2020) The European legal framework for medical AI. In: International cross-domain conference for machine learning and knowledge extraction. Springer, Cham, pp 209?226","DOI":"10.1007\/978-3-030-57321-8_12"},{"issue":"2","key":"10088_CR211","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3158230","volume":"8","author":"JZ Self","year":"2018","unstructured":"Self JZ, Dowling M, Wenskovitch J, Crandell I, Wang M, House L, Leman S, North C (2018) Observation-level and parametric interaction for high-dimensional data analysis. ACM Trans Interact Intell Syst (TIIS) 8(2):1?36","journal-title":"ACM Trans Interact Intell Syst (TIIS)"},{"issue":"2","key":"10088_CR212","doi-asserted-by":"publisher","first-page":"336","DOI":"10.1007\/s11263-019-01228-7","volume":"128","author":"RR Selvaraju","year":"2020","unstructured":"Selvaraju RR, Cogswell M, Das A, Vedantam R, Parikh D, Batra D (2020) Grad-cam: visual explanations from deep networks via gradient-based localization. Int J Comput Vis 128(2):336?359","journal-title":"Int J Comput Vis"},{"key":"10088_CR213","doi-asserted-by":"publisher","first-page":"103457","DOI":"10.1016\/j.artint.2021.103457","volume":"294","author":"M Setzu","year":"2021","unstructured":"Setzu M, Guidotti R, Monreale A, Turini F, Pedreschi D, Giannotti F (2021) Glocalx-from local to global explanations of black box AI models. Artif Intell 294:103457","journal-title":"Artif Intell"},{"issue":"6","key":"10088_CR214","doi-asserted-by":"publisher","first-page":"1094","DOI":"10.1109\/TKDE.2018.2854193","volume":"31","author":"L Shi","year":"2018","unstructured":"Shi L, Teng Z, Wang L, Zhang Y, Binder A (2018) Deepclue: visual interpretation of text-based deep stock prediction. IEEE Trans Knowl Data Eng 31(6):1094?1108","journal-title":"IEEE Trans Knowl Data Eng"},{"key":"10088_CR215","unstructured":"Shrikumar A, Greenside P, Kundaje A (2017) Learning important features through propagating activation differences. In: International conference on machine learning. PMLR, pp 3145?3153"},{"key":"10088_CR216","doi-asserted-by":"publisher","first-page":"188","DOI":"10.1016\/j.eswa.2019.04.029","volume":"130","author":"N Singh","year":"2019","unstructured":"Singh N, Singh P, Bhagat D (2019) A rule extraction approach from support vector machines for diagnosing hypertension among diabetics. Expert Syst Appl 130:188?205","journal-title":"Expert Syst Appl"},{"issue":"6","key":"10088_CR217","doi-asserted-by":"publisher","first-page":"52","DOI":"10.3390\/jimaging6060052","volume":"6","author":"A Singh","year":"2020","unstructured":"Singh A, Sengupta S, Lakshminarayanan V (2020) Explainable deep learning models in medical image analysis. J Imaging 6(6):52","journal-title":"J Imaging"},{"issue":"1","key":"10088_CR218","doi-asserted-by":"publisher","first-page":"857","DOI":"10.1007\/s11042-018-5749-3","volume":"78","author":"S Song","year":"2019","unstructured":"Song S, Huang H, Ruan T (2019) Abstractive text summarization using LSTM-CNN based deep learning. Multimed Tools Appl 78(1):857?875","journal-title":"Multimed Tools Appl"},{"key":"10088_CR219","unstructured":"SP (2019) Explainable AI on emerging multimedia technologies. https:\/\/www.journals.elsevier.com\/signal-processing-image-communication\/call-for-papers\/emerging-multimedia-technologies. Accessed 3 Mar 2020"},{"issue":"1","key":"10088_CR220","first-page":"1064","volume":"26","author":"T Spinner","year":"2019","unstructured":"Spinner T, Schlegel U, Sch\u00e4fer H, El-Assady M (2019) explAIner: a visual analytics framework for interactive and explainable machine learning. IEEE Trans Vis Comput Graph 26(1):1064?1074","journal-title":"IEEE Trans Vis Comput Graph"},{"key":"10088_CR221","unstructured":"Statista (2020) Revenues from the artificial intelligence software market worldwide from 2018 to 2025. https:\/\/www.statista.com\/statistics\/607716\/worldwide-artificial-intelligence-market-revenues\/. Accessed 24 Jan 2020"},{"key":"10088_CR222","doi-asserted-by":"crossref","unstructured":"Stoji\u0107 A, Stani\u0107 N, Vukovi\u0107 G, Stani\u0161i\u0107 S, Peri\u0161i\u0107 M, \u0160o\u0161tari\u0107 A, Lazi\u0107 L (2019) Explainable extreme gradient boosting tree-based prediction of toluene, ethylbenzene and xylene wet deposition. Sci Total Environ 653:140?147","DOI":"10.1016\/j.scitotenv.2018.10.368"},{"issue":"1","key":"10088_CR223","doi-asserted-by":"publisher","first-page":"667","DOI":"10.1109\/TVCG.2017.2744158","volume":"24","author":"H Strobelt","year":"2017","unstructured":"Strobelt H, Gehrmann S, Pfister H, Rush AM (2017) Lstmvis: a tool for visual analysis of hidden state dynamics in recurrent neural networks. IEEE Trans Vis Comput Graph 24(1):667?676","journal-title":"IEEE Trans Vis Comput Graph"},{"issue":"1","key":"10088_CR224","doi-asserted-by":"publisher","first-page":"353","DOI":"10.1109\/TVCG.2018.2865044","volume":"25","author":"H Strobelt","year":"2018","unstructured":"Strobelt H, Gehrmann S, Behrisch M, Perer A, Pfister H, Rush AM (2018) SEQ2SEQ-VIS: a visual debugging tool for sequence-to-sequence models. IEEE Trans Vis Comput Graph 25(1):353?363","journal-title":"IEEE Trans Vis Comput Graph"},{"issue":"3","key":"10088_CR225","doi-asserted-by":"publisher","first-page":"647","DOI":"10.1007\/s10115-013-0679-x","volume":"41","author":"E \u0160trumbelj","year":"2014","unstructured":"\u0160trumbelj E, Kononenko I (2014) Explaining prediction models and individual predictions with feature contributions. Knowl Inf Syst 41(3):647?665","journal-title":"Knowl Inf Syst"},{"issue":"5","key":"10088_CR226","doi-asserted-by":"publisher","first-page":"828","DOI":"10.1109\/TEVC.2019.2890858","volume":"23","author":"J Su","year":"2019","unstructured":"Su J, Vargas DV, Sakurai K (2019) One pixel attack for fooling deep neural networks. IEEE Trans Evol Comput 23(5):828?841","journal-title":"IEEE Trans Evol Comput"},{"key":"10088_CR227","doi-asserted-by":"crossref","unstructured":"Swartout WR, Moore JD (1993) Explanation in second generation expert systems. In: Second generation expert systems. Springer, Cham, pp 543?585","DOI":"10.1007\/978-3-642-77927-5_24"},{"key":"10088_CR228","doi-asserted-by":"crossref","unstructured":"Tan Q, Ye M, Ma AJ, Yang B, Yip TCF, Wong GLH, Yuen PC (2020) Explainable uncertainty-aware convolutional recurrent neural network for irregular medical time series. IEEE Trans Neural Netw Learn Syst","DOI":"10.1109\/TNNLS.2020.3025813"},{"key":"10088_CR229","doi-asserted-by":"crossref","unstructured":"Tjoa E, Guan C (2020) A survey on explainable artificial intelligence (XAI): Toward medical XAI. IEEE Trans Neural Netw Learn Syst","DOI":"10.1109\/TNNLS.2020.3027314"},{"issue":"1","key":"10088_CR230","doi-asserted-by":"publisher","first-page":"131","DOI":"10.1109\/TVCG.2016.2598470","volume":"23","author":"C Turkay","year":"2016","unstructured":"Turkay C, Kaya E, Balcisoy S, Hauser H (2016) Designing progressive and interactive analytics processes for high-dimensional data analysis. IEEE Trans Vis Comput Graph 23(1):131?140","journal-title":"IEEE Trans Vis Comput Graph"},{"key":"10088_CR231","unstructured":"UberAccident (2020) What happens when self-driving cars kill people. https:\/\/www.forbes.com\/sites\/cognitiveworld\/2019\/09\/26\/what-happens-with-self-driving-cars-kill-people\/#4b798bcc405c. Accessed 17 Mar 2020"},{"issue":"10","key":"10088_CR232","doi-asserted-by":"publisher","first-page":"e0164568","DOI":"10.1371\/journal.pone.0164568","volume":"11","author":"V Van Belle","year":"2016","unstructured":"Van Belle V, Van Calster B, Van Huffel S, Suykens JA, Lisboa P (2016) Explaining support vector machines: a color based nomogram. PLoS ONE 11(10):e0164568","journal-title":"PLoS ONE"},{"key":"10088_CR233","unstructured":"Van\u00a0Lent M, Fisher W, Mancuso M (2004) An explainable artificial intelligence system for small-unit tactical behavior. In: Proceedings of the national conference on artificial intelligence. AAAI Press; MIT Press, Menlo Park, London, pp 900?907"},{"key":"10088_CR234","doi-asserted-by":"publisher","first-page":"4099","DOI":"10.1109\/TIP.2021.3069296","volume":"30","author":"H Van Luong","year":"2021","unstructured":"Van Luong H, Joukovsky B, Deligiannis N (2021) Designing interpretable recurrent neural networks for video reconstruction via deep unfolding. IEEE Trans Image Process 30:4099?4113","journal-title":"IEEE Trans Image Process"},{"issue":"2133","key":"10088_CR235","first-page":"20180083","volume":"376","author":"M Veale","year":"2018","unstructured":"Veale M, Binns R, Edwards L (2018) Algorithms that remember: model inversion attacks and data protection law. Philos Trans Royal Soc A Math Phys Eng Sci 376(2133):20180083","journal-title":"Philos Trans Royal Soc A Math Phys Eng Sci"},{"key":"10088_CR236","unstructured":"Vellido A (2019) The importance of interpretability and visualization in machine learning for applications in medicine and health care. Neural Comput Appl 1?15"},{"key":"10088_CR237","doi-asserted-by":"publisher","first-page":"103404","DOI":"10.1016\/j.artint.2020.103404","volume":"291","author":"J Waa","year":"2021","unstructured":"Waa J, Nieuwburg E, Cremers A, Neerincx M (2021) Evaluating XAI: a comparison of rule-based and example-based explanations. Artif Intell 291:103404","journal-title":"Artif Intell"},{"issue":"2","key":"10088_CR238","doi-asserted-by":"publisher","first-page":"76","DOI":"10.1093\/idpl\/ipx005","volume":"7","author":"S Wachter","year":"2017","unstructured":"Wachter S, Mittelstadt B, Floridi L (2017) Why a right to explanation of automated decision-making does not exist in the general data protection regulation. Int Data Privacy Law 7(2):76?99","journal-title":"Int Data Privacy Law"},{"key":"10088_CR239","doi-asserted-by":"crossref","unstructured":"Wang SC (2003) Artificial neural network. In: Interdisciplinary computing in java programming. Springer, Cham, pp 81?100","DOI":"10.1007\/978-1-4615-0377-4_5"},{"key":"10088_CR240","doi-asserted-by":"crossref","unstructured":"Wang B, Gong NZ (2018) Stealing hyperparameters in machine learning. In: 2018 IEEE symposium on security and privacy (SP). IEEE, pp 36?52","DOI":"10.1109\/SP.2018.00038"},{"issue":"12","key":"10088_CR241","doi-asserted-by":"publisher","first-page":"3395","DOI":"10.1109\/TKDE.2016.2606428","volume":"28","author":"H Wang","year":"2016","unstructured":"Wang H, Yeung DY (2016) Towards Bayesian deep learning: a framework and some existing methods. IEEE Trans Knowl Data Eng 28(12):3395?3408","journal-title":"IEEE Trans Knowl Data Eng"},{"issue":"1","key":"10088_CR242","first-page":"171","volume":"25","author":"Y Wang","year":"2017","unstructured":"Wang Y, Aghaei F, Zarafshani A, Qiu Y, Qian W, Zheng B (2017) Computer-aided classification of mammographic masses using visually sensitive image features. J Xray Sci Technol 25(1):171?186","journal-title":"J Xray Sci Technol"},{"issue":"9","key":"10088_CR243","doi-asserted-by":"publisher","first-page":"2568","DOI":"10.1162\/neco_a_01111","volume":"30","author":"Q Wang","year":"2018","unstructured":"Wang Q, Zhang K, Ororbia AG II, Xing X, Liu X, Giles CL (2018) An empirical evaluation of rule extraction from recurrent neural networks. Neural Comput 30(9):2568?2591","journal-title":"Neural Comput"},{"key":"10088_CR244","doi-asserted-by":"publisher","first-page":"18","DOI":"10.1016\/j.ijar.2018.12.013","volume":"106","author":"C Wang","year":"2019","unstructured":"Wang C, Shi Y, Fan X, Shao M (2019a) Attribute reduction based on k-nearest neighborhood rough sets. Int J Approx Reason 106:18?31","journal-title":"Int J Approx Reason"},{"key":"10088_CR245","doi-asserted-by":"crossref","unstructured":"Wang F, Kaushal R, Khullar D (2019b) Should health care demand interpretable artificial intelligence or accept ?black box? medicine? Ann Intern Med","DOI":"10.7326\/M19-2548"},{"key":"10088_CR246","unstructured":"Wang S, Zhou T, Bilmes J (2019c) Bias also matters: bias attribution for deep neural network explanation. In: International conference on machine learning. PMLR, pp 6659?6667"},{"key":"10088_CR247","doi-asserted-by":"publisher","first-page":"188","DOI":"10.1016\/j.asoc.2019.01.015","volume":"77","author":"Y Wang","year":"2019","unstructured":"Wang Y, Wang D, Geng N, Wang Y, Yin Y, Jin Y (2019d) Stacking-based ensemble learning of decision trees for interpretable prostate cancer detection. Appl Soft Comput 77:188?204","journal-title":"Appl Soft Comput"},{"key":"10088_CR248","doi-asserted-by":"crossref","unstructured":"Wasilow S, Thorpe JB (2019) Artificial intelligence, robotics, ethics, and the military: a Canadian perspective. AI Mag 40(1)","DOI":"10.1609\/aimag.v40i1.2848"},{"key":"10088_CR249","doi-asserted-by":"crossref","unstructured":"Weitz K, Schiller D, Schlagowski R, Huber T, Andr\u00e9 E (2020) ?Let me explain!?: exploring the potential of virtual agents in explainable AI interaction design. J Multimodal User Interfaces 1?12","DOI":"10.1007\/s12193-020-00332-0"},{"key":"10088_CR250","doi-asserted-by":"crossref","unstructured":"Wickstr\u00f8m KK, \u00d8yvindMikalsen K, Kampffmeyer M, Revhaug A, Jenssen R (2020) Uncertainty-aware deep ensembles for reliable and explainable predictions of clinical time series. IEEE J Biomed Health Inform","DOI":"10.1109\/JBHI.2020.3042637"},{"key":"10088_CR251","doi-asserted-by":"crossref","unstructured":"Williford JR, May BB, Byrne J (2020) Explainable face recognition. In: European Conference on computer vision. Springer, Cham, pp 248?263","DOI":"10.1007\/978-3-030-58621-8_15"},{"issue":"3","key":"10088_CR252","doi-asserted-by":"publisher","first-page":"254","DOI":"10.1007\/s10791-009-9112-1","volume":"13","author":"Q Wu","year":"2010","unstructured":"Wu Q, Burges CJ, Svore KM, Gao J (2010) Adapting boosting for information retrieval measures. Inf Retr 13(3):254?270","journal-title":"Inf Retr"},{"issue":"7","key":"10088_CR253","doi-asserted-by":"publisher","first-page":"9625","DOI":"10.1007\/s11042-016-3569-x","volume":"76","author":"J Wu","year":"2017","unstructured":"Wu J, Zhong Sh, Jiang J, Yang Y (2017) A novel clustering method for static video summarization. Multimed Tools Appl 76(7):9625?9641","journal-title":"Multimed Tools Appl"},{"key":"10088_CR254","doi-asserted-by":"crossref","unstructured":"Wu M, Hughes M, Parbhoo S, Zazzi M, Roth V, Doshi-Velez F (2018) Beyond sparsity: tree regularization of deep models for interpretability. In: Proceedings of the AAAI conference on artificial intelligence, vol 32","DOI":"10.1609\/aaai.v32i1.11501"},{"key":"10088_CR255","unstructured":"Xu J, Zhang Z, Friedman T, Liang Y, Broeck G (2018) A semantic loss function for deep learning with symbolic knowledge. In: International conference on machine learning. PMLR, pp 5502?5511"},{"issue":"1","key":"10088_CR256","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1038\/s41467-019-13647-8","volume":"10","author":"Y Yamamoto","year":"2019","unstructured":"Yamamoto Y, Tsuzuki T, Akatsuka J, Ueki M, Morikawa H, Numata Y, Takahara T, Tsuyuki T, Tsutsumi K, Nakazawa R et al (2019) Automated acquisition of explainable knowledge from unannotated histopathology images. Nat Commun 10(1):1?9","journal-title":"Nat Commun"},{"key":"10088_CR257","unstructured":"Yang SCH, Shafto P (2017) Explainable artificial intelligence via Bayesian teaching. In: NIPS 2017 workshop on teaching machines, robots, and humans, pp 127?137"},{"key":"10088_CR258","doi-asserted-by":"crossref","unstructured":"Yang Z, Zhang A, Sudjianto A (2020) Enhancing explainability of neural networks through architecture constraints. IEEE Trans Neural Netw Learn Syst","DOI":"10.1109\/TNNLS.2020.3007259"},{"issue":"7","key":"10088_CR259","first-page":"1407","volume":"28","author":"M Yeganejou","year":"2019","unstructured":"Yeganejou M, Dick S, Miller J (2019) Interpretable deep convolutional fuzzy classifier. IEEE Trans Fuzzy Syst 28(7):1407?1419","journal-title":"IEEE Trans Fuzzy Syst"},{"key":"10088_CR260","unstructured":"Yosinski J, Clune J, Nguyen A, Fuchs T, Lipson H (2015) Understanding neural networks through deep visualization. arXiv preprint arXiv:150606579"},{"key":"10088_CR261","doi-asserted-by":"publisher","first-page":"93","DOI":"10.1016\/j.eswa.2016.10.017","volume":"68","author":"M Yousefi-Azar","year":"2017","unstructured":"Yousefi-Azar M, Hamey L (2017) Text summarization using unsupervised deep learning. Expert Syst Appl 68:93?105","journal-title":"Expert Syst Appl"},{"key":"10088_CR262","doi-asserted-by":"crossref","unstructured":"Yu H, Yang S, Gu W, Zhang S (2017) Baidu driving dataset and end-to-end reactive control model. In: 2017 IEEE intelligent vehicles symposium (IV). IEEE, pp 341?346","DOI":"10.1109\/IVS.2017.7995742"},{"key":"10088_CR263","doi-asserted-by":"publisher","first-page":"107131","DOI":"10.1016\/j.patcog.2019.107131","volume":"105","author":"J Yuan","year":"2020","unstructured":"Yuan J, Xiong HC, Xiao Y, Guan W, Wang M, Hong R, Li ZY (2020) Gated CNN: Integrating multi-scale feature layers for object detection. Pattern Recogn 105:107131","journal-title":"Pattern Recogn"},{"key":"10088_CR264","doi-asserted-by":"publisher","first-page":"106779","DOI":"10.1016\/j.knosys.2021.106779","volume":"218","author":"D Zeltner","year":"2021","unstructured":"Zeltner D, Schmid B, Csisz\u00e1r G, Csisz\u00e1r O (2021) Squashing activation functions in benchmark tests: towards a more explainable artificial intelligence using continuous-valued logic. Knowl Based Syst 218:106779","journal-title":"Knowl Based Syst"},{"issue":"1","key":"10088_CR265","doi-asserted-by":"publisher","first-page":"27","DOI":"10.1631\/FITEE.1700808","volume":"19","author":"Qs Zhang","year":"2018","unstructured":"Zhang Qs, Zhu SC (2018) Visual interpretability for deep learning: a survey. Fronti Inf Technol Electron Eng 19(1):27?39","journal-title":"Fronti Inf Technol Electron Eng"},{"issue":"1","key":"10088_CR266","doi-asserted-by":"publisher","first-page":"364","DOI":"10.1109\/TVCG.2018.2864499","volume":"25","author":"J Zhang","year":"2018","unstructured":"Zhang J, Wang Y, Molino P, Li L, Ebert DS (2018a) Manifold: a model-agnostic framework for interpretation and diagnosis of machine learning models. IEEE Trans Vis Comput Graph 25(1):364?373","journal-title":"IEEE Trans Vis Comput Graph"},{"key":"10088_CR267","doi-asserted-by":"crossref","unstructured":"Zhang Q, Nian\u00a0Wu Y, Zhu SC (2018b) Interpretable convolutional neural networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 8827?8836","DOI":"10.1109\/CVPR.2018.00920"},{"key":"10088_CR268","doi-asserted-by":"crossref","unstructured":"Zhang Q, Yang Y, Ma H, Wu YN (2019) Interpreting CNNs via decision trees. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 6261?6270","DOI":"10.1109\/CVPR.2019.00642"},{"key":"10088_CR269","doi-asserted-by":"crossref","unstructured":"Zhang A, Teng L, Alterovitz G (2020a) An explainable machine learning platform for pyrazinamide resistance prediction and genetic feature identification of mycobacterium tuberculosis. J Am Med Inform Assoc","DOI":"10.1093\/jamia\/ocaa233"},{"issue":"7","key":"10088_CR270","doi-asserted-by":"publisher","first-page":"1744","DOI":"10.1109\/TMM.2019.2963592","volume":"22","author":"M Zhang","year":"2020","unstructured":"Zhang M, You H, Kadam P, Liu S, Kuo CCJ (2020b) Pointhop: an explainable machine learning method for point cloud classification. IEEE Trans Multimed 22(7):1744?1755","journal-title":"IEEE Trans Multimed"},{"key":"10088_CR271","doi-asserted-by":"crossref","unstructured":"Zhang W, Tang S, Su J, Xiao J, Zhuang Y (2020c) Tell and guess: cooperative learning for natural image caption generation with hierarchical refined attention. Multimed Tools Appl 1?16","DOI":"10.1007\/s11042-020-08832-7"},{"key":"10088_CR272","doi-asserted-by":"crossref","unstructured":"Zhang Z, Beck MW, Winkler DA, Huang B, Sibanda W, Goyal H et al (2018c) Opening the black box of neural networks: methods for interpreting neural network models in clinical applications. Ann Transl Med 6(11)","DOI":"10.21037\/atm.2018.05.32"},{"issue":"8","key":"10088_CR273","doi-asserted-by":"publisher","first-page":"4544","DOI":"10.1109\/TGRS.2016.2543748","volume":"54","author":"W Zhao","year":"2016","unstructured":"Zhao W, Du S (2016) Spectral-spatial feature extraction for hyperspectral image classification: a dimension reduction and deep learning approach. IEEE Trans Geosci Remote Sens 54(8):4544?4554","journal-title":"IEEE Trans Geosci Remote Sens"},{"key":"10088_CR274","doi-asserted-by":"publisher","first-page":"227","DOI":"10.1016\/j.patrec.2019.12.020","volume":"131","author":"S Zheng","year":"2020","unstructured":"Zheng S, Ding C (2020) A group lasso based sparse KNN classifier. Pattern Recogn Lett 131:227?233","journal-title":"Pattern Recogn Lett"},{"issue":"7","key":"10088_CR275","doi-asserted-by":"publisher","first-page":"914","DOI":"10.1631\/FITEE.1700822","volume":"20","author":"Xl Zheng","year":"2019","unstructured":"Zheng Xl, Zhu My, Li Qb, Chen Cc, Tan Yc (2019) FinBrain: when finance meets AI 2.0. Front Inf Technol Electron Eng 20(7):914?924","journal-title":"Front Inf Technol Electron Eng"},{"issue":"9","key":"10088_CR276","doi-asserted-by":"publisher","first-page":"2131","DOI":"10.1109\/TPAMI.2018.2858759","volume":"41","author":"B Zhou","year":"2018","unstructured":"Zhou B, Bau D, Oliva A, Torralba A (2018a) Interpreting deep visual representations via network dissection. IEEE Trans Pattern Anal Mach Intell 41(9):2131?2145","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"issue":"1","key":"10088_CR277","doi-asserted-by":"publisher","first-page":"51","DOI":"10.1007\/s10845-015-1089-6","volume":"29","author":"X Zhou","year":"2018","unstructured":"Zhou X, Jiang P, Wang X (2018b) Recognition of control chart patterns using fuzzy SVM with a hybrid kernel function. J Intell Manuf 29(1):51?67","journal-title":"J Intell Manuf"},{"issue":"1","key":"10088_CR278","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1631\/FITEE.1601883","volume":"18","author":"Yt Zhuang","year":"2017","unstructured":"Zhuang Yt, Wu F, Chen C, Pan Yh (2017) Challenges and opportunities: from big data to knowledge in AI 2.0. Front Inf Technol Electron Eng 18(1):3?14","journal-title":"Front Inf Technol Electron Eng"}],"container-title":["Artificial Intelligence Review"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10462-021-10088-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10462-021-10088-y\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10462-021-10088-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,2,9]],"date-time":"2023-02-09T12:00:09Z","timestamp":1675944009000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10462-021-10088-y"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,11,18]]},"references-count":278,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2022,6]]}},"alternative-id":["10088"],"URL":"https:\/\/doi.org\/10.1007\/s10462-021-10088-y","relation":{},"ISSN":["0269-2821","1573-7462"],"issn-type":[{"value":"0269-2821","type":"print"},{"value":"1573-7462","type":"electronic"}],"subject":[],"published":{"date-parts":[[2021,11,18]]},"assertion":[{"value":"18 November 2021","order":1,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}