{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,26]],"date-time":"2025-03-26T19:55:48Z","timestamp":1743018948119,"version":"3.40.3"},"publisher-location":"Cham","reference-count":60,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031467776"},{"type":"electronic","value":"9783031467783"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-46778-3_12","type":"book-chapter","created":{"date-parts":[[2023,11,24]],"date-time":"2023-11-24T06:02:05Z","timestamp":1700805725000},"page":"127-140","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Forcing the\u00a0Network to\u00a0Use Human Explanations in\u00a0Its Inference Process"],"prefix":"10.1007","author":[{"given":"Javier","family":"Via\u00f1a","sequence":"first","affiliation":[]},{"given":"Andrew","family":"Vanderburg","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,11,25]]},"reference":[{"key":"12_CR1","unstructured":"Gunning, D.: Explainable artificial intelligence (XAI). Defense Advanced Research Projects Agency (DARPA), nd Web, vol. 2, no. 2 (2017)"},{"key":"12_CR2","series-title":"Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence)","doi-asserted-by":"publisher","first-page":"563","DOI":"10.1007\/978-3-030-32236-6_51","volume-title":"Natural Language Processing and Chinese Computing","author":"F Xu","year":"2019","unstructured":"Xu, F., Uszkoreit, H., Du, Y., Fan, W., Zhao, D., Zhu, J.: Explainable AI: a brief survey on history, research areas, approaches and challenges. In: Tang, J., Kan, M.-Y., Zhao, D., Li, S., Zan, H. (eds.) NLPCC 2019. LNCS (LNAI), vol. 11839, pp. 563\u2013574. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-32236-6_51"},{"key":"12_CR3","doi-asserted-by":"publisher","first-page":"52138","DOI":"10.1109\/ACCESS.2018.2870052","volume":"6","author":"A Adadi","year":"2018","unstructured":"Adadi, A., Berrada, M.: Peeking inside the black-box: a survey on explainable artificial intelligence (XAI). IEEE Access 6, 52138\u201352160 (2018)","journal-title":"IEEE Access"},{"issue":"5","key":"12_CR4","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3236009","volume":"51","author":"R Guidotti","year":"2018","unstructured":"Guidotti, R., Monreale, A., Ruggieri, S., Turini, F., Giannotti, F., Pedreschi, D.: A survey of methods for explaining black box models. ACM Comput. Surv. (CSUR) 51(5), 1\u201342 (2018)","journal-title":"ACM Comput. Surv. (CSUR)"},{"key":"12_CR5","doi-asserted-by":"publisher","first-page":"185","DOI":"10.1016\/j.neunet.2020.07.010","volume":"130","author":"P Angelov","year":"2020","unstructured":"Angelov, P., Soares, E.: Towards explainable deep neural networks (xDNN). Neural Netw. 130, 185\u2013194 (2020). ID: 271125","journal-title":"Neural Netw."},{"key":"12_CR6","unstructured":"Turner, H., Gedeon, T.D.: Extracting meaning from neural networks. In: Proceedings 13th International Conference on AI, vol. 1, pp. 243\u2013252 (1993)"},{"key":"12_CR7","doi-asserted-by":"publisher","unstructured":"Thrun, S.: Explanation-based neural network learning. In: Thrun, S. (ed.) Explanation-Based Neural Network Learning: A Lifelong Learning Approach. The Kluwer International Series in Engineering and Computer Science, vol. 357, pp. 19\u201348. Springer, Boston (1996). https:\/\/doi.org\/10.1007\/978-1-4613-1381-6_2. ID: Thrun 1996","DOI":"10.1007\/978-1-4613-1381-6_2"},{"issue":"3","key":"12_CR8","first-page":"50","volume":"38","author":"B Goodman","year":"2017","unstructured":"Goodman, B., Flaxman, S.: European union regulations on algorithmic decision-making and a \u201cright to explanation\u2019\u2019. AI Mag. 38(3), 50\u201357 (2017)","journal-title":"AI Mag."},{"issue":"2","key":"12_CR9","doi-asserted-by":"publisher","first-page":"76","DOI":"10.1093\/idpl\/ipx005","volume":"7","author":"S Wachter","year":"2017","unstructured":"Wachter, S., Mittelstadt, B., Floridi, L.: Why a right to explanation of automated decision-making does not exist in the general data protection regulation. Int. Data Priv. Law 7(2), 76\u201399 (2017)","journal-title":"Int. Data Priv. Law"},{"issue":"1","key":"12_CR10","doi-asserted-by":"publisher","first-page":"72","DOI":"10.1109\/MCI.2021.3129960","volume":"17","author":"R Hamon","year":"2022","unstructured":"Hamon, R., Junklewitz, H., Sanchez, I., Malgieri, G., De Hert, P.: Bridging the gap between AI and explainability in the GDPR: towards trustworthiness-by-design in automated decision-making. IEEE Comput. Intell. Mag. 17(1), 72\u201385 (2022)","journal-title":"IEEE Comput. Intell. Mag."},{"key":"12_CR11","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"56","DOI":"10.1007\/978-3-030-49760-6_4","volume-title":"Design, User Experience, and Usability. Design for Contemporary Interactive Environments","author":"JJ Ferreira","year":"2020","unstructured":"Ferreira, J.J., Monteiro, M.S.: What are people doing about XAI user experience? A survey on AI explainability research and practice. In: Marcus, A., Rosenzweig, E. (eds.) HCII 2020. LNCS, vol. 12201, pp. 56\u201373. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-49760-6_4"},{"key":"12_CR12","doi-asserted-by":"publisher","first-page":"103143","DOI":"10.1016\/j.trc.2021.103143","volume":"128","author":"A Alwosheel","year":"2021","unstructured":"Alwosheel, A., van Cranenburgh, S., Chorus, C.G.: Why did you predict that? Towards explainable artificial neural networks for travel demand analysis. Transp. Res. Part C: Emerg. Technol. 128, 103143 (2021). ID: 271729","journal-title":"Transp. Res. Part C: Emerg. Technol."},{"key":"12_CR13","doi-asserted-by":"crossref","unstructured":"Markus, A.F., Kors, J.A., Rijnbeek, P.R.: The role of explainability in creating trustworthy artificial intelligence for health care: A comprehensive survey of the terminology, design choices, and evaluation strategies. J. Biomed. Inform. 113 (2021)","DOI":"10.1016\/j.jbi.2020.103655"},{"issue":"2","key":"12_CR14","doi-asserted-by":"publisher","first-page":"699","DOI":"10.1109\/TMI.2020.3035253","volume":"40","author":"G Ran","year":"2021","unstructured":"Ran, G., et al.: CA-net: comprehensive attention convolutional neural networks for explainable medical image segmentation. IEEE Trans. Med. Imaging 40(2), 699\u2013711 (2021)","journal-title":"IEEE Trans. Med. Imaging"},{"issue":"3","key":"12_CR15","doi-asserted-by":"publisher","first-page":"3002","DOI":"10.1007\/s10489-021-02572-3","volume":"52","author":"CM Dasari","year":"2022","unstructured":"Dasari, C.M., Bhukya, R.: Explainable deep neural networks for novel viral genome prediction. Appl. Intell. 52(3), 3002\u20133017 (2022)","journal-title":"Appl. Intell."},{"key":"12_CR16","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"464","DOI":"10.1007\/978-3-030-00934-2_52","volume-title":"Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2018","author":"C Biffi","year":"2018","unstructured":"Biffi, C., et al.: Learning interpretable anatomical features through deep generative models: application to cardiac remodeling. In: Frangi, A.F., Schnabel, J.A., Davatzikos, C., Alberola-L\u00f3pez, C., Fichtinger, G. (eds.) MICCAI 2018. LNCS, vol. 11071, pp. 464\u2013471. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-00934-2_52"},{"issue":"6","key":"12_CR17","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TMAG.2021.3063141","volume":"57","author":"H Sasaki","year":"2021","unstructured":"Sasaki, H., Hidaka, Y., Igarashi, H.: Explainable deep neural network for design of electric motors. IEEE Trans. Magn. 57(6), 1\u20134 (2021)","journal-title":"IEEE Trans. Magn."},{"key":"12_CR18","doi-asserted-by":"publisher","first-page":"476","DOI":"10.1016\/j.procir.2018.12.008","volume":"80","author":"J Grezmak","year":"2019","unstructured":"Grezmak, J., Wang, P., Sun, C., Gao, R.X.: Explainable convolutional neural network for gearbox fault diagnosis. Procedia CIRP 80, 476\u2013481 (2019)","journal-title":"Procedia CIRP"},{"issue":"6","key":"12_CR19","doi-asserted-by":"publisher","first-page":"4036","DOI":"10.1109\/TII.2020.3012989","volume":"17","author":"MS Kim","year":"2020","unstructured":"Kim, M.S., Yun, J.P., Park, P.: An explainable convolutional neural network for fault diagnosis in linear motion guide. IEEE Trans. Ind. Inform. 17(6), 4036\u20134045 (2020)","journal-title":"IEEE Trans. Ind. Inform."},{"key":"12_CR20","unstructured":"Jeyakumar, J.V., Noor, J., Cheng, Y.-H., Garcia, L., Srivastava, M.: How can I explain this to you? an empirical study of deep neural network explanation methods. In: Advances in Neural Information Processing Systems, vol. 33, pp. 4211\u20134222 (2020)"},{"key":"12_CR21","series-title":"Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence)","doi-asserted-by":"publisher","first-page":"155","DOI":"10.1007\/978-3-030-29249-2_11","volume-title":"Case-Based Reasoning Research and Development","author":"MT Keane","year":"2019","unstructured":"Keane, M.T., Kenny, E.M.: How case-based reasoning explains neural networks: a theoretical analysis of XAI using Post-Hoc explanation-by-example from a survey of ANN-CBR twin-systems. In: Bach, K., Marling, C. (eds.) ICCBR 2019. LNCS (LNAI), vol. 11680, pp. 155\u2013171. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-29249-2_11"},{"key":"12_CR22","unstructured":"Zhang, Q., Yang, Y., Liu, Y., Wu, Y.N., Zhu, S.-C.: Unsupervised learning of neural networks to explain neural networks. arXiv preprint arXiv:1805.07468 (2018)"},{"key":"12_CR23","unstructured":"Sarker, M.K., Xie, N., Doran, D., Raymer, M., Hitzler, P.: Explaining trained neural networks with semantic web technologies: First steps. arXiv preprint arXiv:1710.04324 (2017)"},{"key":"12_CR24","doi-asserted-by":"crossref","unstructured":"Le, T., Wang, S., Lee, D.: GRACE: generating concise and informative contrastive sample to explain neural network model\u2019s prediction. In: Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pp. 238\u2013248 (2020)","DOI":"10.1145\/3394486.3403066"},{"issue":"1","key":"12_CR25","first-page":"3","volume":"16","author":"Z-H Zhou","year":"2003","unstructured":"Zhou, Z.-H., Jiang, Y., Chen, S.-F.: Extracting symbolic rules from trained neural network ensembles. AI Commun. 16(1), 3\u201315 (2003)","journal-title":"AI Commun."},{"key":"12_CR26","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"818","DOI":"10.1007\/978-3-319-10590-1_53","volume-title":"Computer Vision \u2013 ECCV 2014","author":"MD Zeiler","year":"2014","unstructured":"Zeiler, M.D., Fergus, R.: Visualizing and understanding convolutional networks. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol. 8689, pp. 818\u2013833. Springer, Cham (2014). https:\/\/doi.org\/10.1007\/978-3-319-10590-1_53"},{"key":"12_CR27","unstructured":"Springenberg, J.T., Dosovitskiy, A., Brox, T., Riedmiller, M.A.: Striving for simplicity: the all convolutional net. In: Bengio, Y., LeCun, Y. (eds.) 3rd International Conference on Learning Representations, ICLR, San Diego, CA, USA, 7\u20139 May 2015, Workshop Track Proceedings (2015)"},{"key":"12_CR28","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"63","DOI":"10.1007\/978-3-319-44781-0_8","volume-title":"Artificial Neural Networks and Machine Learning \u2013 ICANN 2016","author":"A Binder","year":"2016","unstructured":"Binder, A., Montavon, G., Lapuschkin, S., M\u00fcller, K.-R., Samek, W.: Layer-wise relevance propagation for neural networks with local renormalization layers. In: Villa, A.E.P., Masulli, P., Pons Rivero, A.J. (eds.) ICANN 2016. LNCS, vol. 9887, pp. 63\u201371. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-44781-0_8"},{"key":"12_CR29","doi-asserted-by":"publisher","first-page":"211","DOI":"10.1016\/j.patcog.2016.11.008","volume":"65","author":"G Montavon","year":"2017","unstructured":"Montavon, G., Lapuschkin, S., Binder, A., Samek, W., M\u00fcller, K.-R.: Explaining nonlinear classification decisions with deep Taylor decomposition. Pattern Recogn. 65, 211\u2013222 (2017)","journal-title":"Pattern Recogn."},{"key":"12_CR30","unstructured":"Kindermans, P.-J., et al.: Learning how to explain neural networks: PatternNet and PatternAttributionn. arXiv preprint arXiv:1705.05598 (2017)"},{"key":"12_CR31","doi-asserted-by":"crossref","unstructured":"Neumeier, M., Botsch, M., Tollk\u00fchn, A., Berberich, T.: Variational autoencoder-based vehicle trajectory prediction with an interpretable latent space, pp. 820\u2013827 (2021)","DOI":"10.1109\/ITSC48978.2021.9565120"},{"key":"12_CR32","doi-asserted-by":"publisher","first-page":"115842","DOI":"10.1016\/j.eswa.2021.115842","volume":"186","author":"J-Y Kim","year":"2021","unstructured":"Kim, J.-Y., Cho, S.-B.: Explainable prediction of electric energy demand using a deep autoencoder with interpretable latent space. Expert Syst. Appl. 186, 115842 (2021). ID: 271506","journal-title":"Expert Syst. Appl."},{"key":"12_CR33","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"525","DOI":"10.1007\/978-3-031-18840-4_37","volume-title":"Discovery Science","author":"F Bodria","year":"2022","unstructured":"Bodria, F., Guidotti, R., Giannotti, F., Pedreschi, D.: Interpretable latent space to enable counterfactual explanations. In: Pascal, P., Ienco, D. (eds.) DS 2022. LNCS, vol. 13601, pp. 525\u2013540. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-18840-4_37"},{"key":"12_CR34","doi-asserted-by":"crossref","unstructured":"B\u00f6lat, K., Kumbasar, T.: Interpreting variational autoencoders with fuzzy logic: a step towards interpretable deep learning based fuzzy classifiers. In: IEEE International Conference on Fuzzy Systems (FUZZ-IEEE), pp. 1\u20137 (2020)","DOI":"10.1109\/FUZZ48607.2020.9177631"},{"key":"12_CR35","doi-asserted-by":"crossref","unstructured":"Bharadwaj, P., Li, M., Demanet, L.: Redatuming physical systems using symmetric autoencoders. Phys. Rev. Res. 4(2) (2022)","DOI":"10.1103\/PhysRevResearch.4.023118"},{"issue":"5","key":"12_CR36","doi-asserted-by":"publisher","first-page":"206","DOI":"10.1038\/s42256-019-0048-x","volume":"1","author":"C Rudin","year":"2019","unstructured":"Rudin, C.: Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nat. Mach. Intell. 1(5), 206\u2013215 (2019)","journal-title":"Nat. Mach. Intell."},{"key":"12_CR37","series-title":"Lecture Notes in Networks and Systems","doi-asserted-by":"publisher","first-page":"298","DOI":"10.1007\/978-3-031-16038-7_29","volume-title":"Applications of Fuzzy Techniques","author":"J Via\u00f1a","year":"2023","unstructured":"Via\u00f1a, J., Ralescu, S., Kreinovich, V., Ralescu, A., Cohen, K.: Single hidden layer CEFYDRA: cluster-first explainable FuzzY-based deep self-reorganizing algorithm. In: Dick, S., Kreinovich, V., Lingras, P. (eds.) NAFIPS 2022. LNNS, vol. 500, pp. 298\u2013307. Springer, Cham (2023). https:\/\/doi.org\/10.1007\/978-3-031-16038-7_29"},{"key":"12_CR38","series-title":"Lecture Notes in Networks and Systems","doi-asserted-by":"publisher","first-page":"308","DOI":"10.1007\/978-3-031-16038-7_30","volume-title":"Applications of Fuzzy Techniques","author":"J Via\u00f1a","year":"2023","unstructured":"Via\u00f1a, J., Ralescu, S., Kreinovich, V., Ralescu, A., Cohen, K.: Multiple hidden layered CEFYDRA: cluster-first explainable FuzzY-based deep self-reorganizing algorithm. In: Dick, S., Kreinovich, V., Lingras, P. (eds.) NAFIPS 2022. LNNS, vol. 500, pp. 308\u2013322. Springer, Cham (2023). https:\/\/doi.org\/10.1007\/978-3-031-16038-7_30"},{"key":"12_CR39","series-title":"Lecture Notes in Networks and Systems","doi-asserted-by":"publisher","first-page":"323","DOI":"10.1007\/978-3-031-16038-7_31","volume-title":"Applications of Fuzzy Techniques","author":"J Via\u00f1a","year":"2023","unstructured":"Via\u00f1a, J., Ralescu, S., Kreinovich, V., Ralescu, A., Cohen, K.: Initialization and plasticity of CEFYDRA: cluster-first explainable FuzzY-based deep self-reorganizing algorithm. In: Dick, S., Kreinovich, V., Lingras, P. (eds.) NAFIPS 2022. LNNS, vol. 500, pp. 323\u2013335. Springer, Cham (2023). https:\/\/doi.org\/10.1007\/978-3-031-16038-7_31"},{"key":"12_CR40","unstructured":"Park, J.H., Shin, C.-K., Im, K.H., Park, S.C.: A local weighting method to the integration of neural network and case based reasoning. In: Neural Networks for Signal Processing XI: Proceedings of the 2001 IEEE Signal Processing Society Workshop (IEEE Cat. No. 01TH8584), pp. 33\u201342. IEEE (2001)"},{"key":"12_CR41","series-title":"Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence)","doi-asserted-by":"publisher","first-page":"15","DOI":"10.1007\/978-3-030-01081-2_2","volume-title":"Case-Based Reasoning Research and Development","author":"K Amin","year":"2018","unstructured":"Amin, K., Kapetanakis, S., Althoff, K.-D., Dengel, A., Petridis, M.: Answering with cases: a CBR approach to deep learning. In: Cox, M.T., Funk, P., Begum, S. (eds.) ICCBR 2018. LNCS (LNAI), vol. 11156, pp. 15\u201327. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01081-2_2"},{"key":"12_CR42","doi-asserted-by":"publisher","first-page":"113200","DOI":"10.1016\/j.eswa.2020.113200","volume":"147","author":"L Corbat","year":"2020","unstructured":"Corbat, L., Nauval, M., Henriet, J., Lapayre, J.-C.: A fusion method based on deep learning and case-based reasoning which improves the resulting medical image segmentations. Expert Syst. Appl. 147, 113200 (2020)","journal-title":"Expert Syst. Appl."},{"issue":"6","key":"12_CR43","doi-asserted-by":"publisher","first-page":"2610","DOI":"10.1109\/TNNLS.2020.3007259","volume":"32","author":"Z Yang","year":"2021","unstructured":"Yang, Z., Zhang, A., Sudjianto, A.: Enhancing explainability of neural networks through architecture constraints. IEEE Trans. Neural Netw. Learn. Syst. 32(6), 2610\u20132621 (2021)","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"12_CR44","unstructured":"Rieger, L., Singh, C., Murdoch, W., Yu, B.: Interpretations are useful: penalizing explanations to align neural networks with prior knowledge. In: International Conference on Machine Learning, pp. 8116\u20138126. PMLR (2020)"},{"key":"12_CR45","doi-asserted-by":"crossref","unstructured":"Shavlik, J.W., Towell, G.G.: Combining explanation-based learning and artificial neural networks. In: Proceedings of the Sixth International Workshop on Machine Learning, pp. 90\u201392. Elsevier (1989)","DOI":"10.1016\/B978-1-55860-036-2.50032-1"},{"issue":"9","key":"12_CR46","doi-asserted-by":"publisher","first-page":"607","DOI":"10.1038\/s43588-021-00132-w","volume":"1","author":"PJ Blazek","year":"2021","unstructured":"Blazek, P.J., Lin, M.M.: Explainable neural networks that simulate reasoning. Nature Comput. Sci. 1(9), 607\u2013618 (2021). ID: Blazek 2021","journal-title":"Nature Comput. Sci."},{"key":"12_CR47","doi-asserted-by":"crossref","unstructured":"Fel, T., Vigouroux, D., Cad\u00e8ne, R., Serre, T.: How good is your explanation? Algorithmic stability measures to assess the quality of explanations for deep neural networks. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 720\u2013730 (2022)","DOI":"10.1109\/WACV51458.2022.00163"},{"key":"12_CR48","unstructured":"Johs, A.J., Lutts, M., Weber, R.O.: Measuring explanation quality in XCBR. In: Proceedings of the 26th International Conference on Case-Based Reasoning, p. 75. Springer, Heidelberg (2018)"},{"key":"12_CR49","doi-asserted-by":"crossref","unstructured":"Pedreschi, D., Giannotti, F., Guidotti, R., Monreale, A., Ruggieri, S., Turini, F.: Meaningful explanations of black box AI decision systems. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33, pp. 9780\u20139784 (2019)","DOI":"10.1609\/aaai.v33i01.33019780"},{"key":"12_CR50","doi-asserted-by":"publisher","first-page":"108194","DOI":"10.1016\/j.patcog.2021.108194","volume":"121","author":"A-K Dombrowski","year":"2022","unstructured":"Dombrowski, A.-K., Anders, C.J., M\u00fcller, K.-R., Kessel, P.: Towards robust explanations for deep neural networks. Pattern Recogn. 121, 108194 (2022)","journal-title":"Pattern Recogn."},{"issue":"7623","key":"12_CR51","doi-asserted-by":"publisher","first-page":"20","DOI":"10.1038\/538020a","volume":"538","author":"D Castelvecchi","year":"2016","unstructured":"Castelvecchi, D.: Can we open the black box of AI? Nat. News 538(7623), 20 (2016)","journal-title":"Nat. News"},{"key":"12_CR52","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.artint.2018.07.007","volume":"267","author":"T Miller","year":"2019","unstructured":"Miller, T.: Explanation in artificial intelligence: insights from the social sciences. Artif. Intell. 267, 1\u201338 (2019)","journal-title":"Artif. Intell."},{"key":"12_CR53","doi-asserted-by":"publisher","first-page":"103459","DOI":"10.1016\/j.artint.2021.103459","volume":"294","author":"EM Kenny","year":"2021","unstructured":"Kenny, E.M., Ford, C., Quinn, M., Keane, M.T.: Explaining black-box classifiers using post-hoc explanations-by-example: the effect of explanations and error-rates in xai user studies. Artif. Intell. 294, 103459 (2021)","journal-title":"Artif. Intell."},{"key":"12_CR54","unstructured":"Flickr. www.flicker.com. Accessed 16 Nov 2022"},{"key":"12_CR55","doi-asserted-by":"crossref","unstructured":"K\u00fcmmerer, M., Wallis, T.S.A., Gatys, L.A., Bethge, M.: Understanding low- and high-level contributions to fixation prediction. In: 2017 IEEE International Conference on Computer Vision (ICCV), pp. 4799\u20134808 (2017)","DOI":"10.1109\/ICCV.2017.513"},{"key":"12_CR56","unstructured":"Bylinskii, Z., et al.: MIT saliency benchmark"},{"key":"12_CR57","doi-asserted-by":"crossref","unstructured":"Jiang, M., Huang, S., Duan, J., Zhao, Q.: SALICON: Saliency in context. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1072\u20131080 (2015)","DOI":"10.1109\/CVPR.2015.7298710"},{"key":"12_CR58","doi-asserted-by":"crossref","unstructured":"Judd, T., Ehinger, K., Durand, F., Torralba, A.: Learning to predict where humans look. In: IEEE International Conference on Computer Vision (ICCV) (2009)","DOI":"10.1109\/ICCV.2009.5459462"},{"key":"12_CR59","unstructured":"Via\u00f1a, J.: ForcedNet for image reconstruction (2022). https:\/\/github.com\/JavierVianaAi\/forcednets-image-reconstruction"},{"key":"12_CR60","unstructured":"Matthias Kummerer. DeepGaze (2022). https:\/\/github.com\/matthias-k\/DeepGaze"}],"container-title":["Lecture Notes in Networks and Systems","Fuzzy Information Processing 2023"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-46778-3_12","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,11,24]],"date-time":"2023-11-24T06:04:02Z","timestamp":1700805842000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-46778-3_12"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031467776","9783031467783"],"references-count":60,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-46778-3_12","relation":{},"ISSN":["2367-3370","2367-3389"],"issn-type":[{"type":"print","value":"2367-3370"},{"type":"electronic","value":"2367-3389"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"25 November 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"This work was supported by two NASA Grants, the NASA Extremely Precise Radial Velocity Foundation Science Program (No. 80NSSC22K0848) and the NASA Astrophysical Data Analysis Program (No. 80NSSC22K1408).","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Financial Support and Sponsorship"}},{"value":"NAFIPS","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"North American Fuzzy Information Processing Society Annual Conference","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Cincinnati, OH","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"USA","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"31 May 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2 June 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"nafips2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/nafips2023.com\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}