{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,11]],"date-time":"2026-04-11T01:19:58Z","timestamp":1775870398398,"version":"3.50.1"},"publisher-location":"Cham","reference-count":28,"publisher":"Springer International Publishing","isbn-type":[{"value":"9783031090363","type":"print"},{"value":"9783031090370","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-09037-0_8","type":"book-chapter","created":{"date-parts":[[2022,6,1]],"date-time":"2022-06-01T04:26:16Z","timestamp":1654057576000},"page":"84-95","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":32,"title":["Metrics for\u00a0Saliency Map Evaluation of\u00a0Deep Learning Explanation Methods"],"prefix":"10.1007","author":[{"given":"Tristan","family":"Gomez","sequence":"first","affiliation":[]},{"given":"Thomas","family":"Fr\u00e9our","sequence":"additional","affiliation":[]},{"given":"Harold","family":"Mouch\u00e8re","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,6,2]]},"reference":[{"key":"8_CR1","unstructured":"Adebayo, J., Gilmer, J., Goodfellow, I., Kim, B.: Local explanation methods for deep neural networks lack sensitivity to parameter values (2018)"},{"key":"8_CR2","doi-asserted-by":"publisher","unstructured":"Alqaraawi, A., Schuessler, M., Wei\u00df, P., Costanza, E., Berthouze, N.: Evaluating saliency map explanations for convolutional neural networks: a user study. In: IUI 2020, pp. 275\u2013285. Association for Computing Machinery, New York (2020). https:\/\/doi.org\/10.1145\/3377325.3377519","DOI":"10.1145\/3377325.3377519"},{"key":"8_CR3","doi-asserted-by":"publisher","unstructured":"Chattopadhay, A., Sarkar, A., Howlader, P., Balasubramanian, V.N.: Grad-CAM++: generalized gradient-based visual explanations for deep convolutional networks. In: 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), pp. 839\u2013847 (2018). https:\/\/doi.org\/10.1109\/WACV.2018.00097","DOI":"10.1109\/WACV.2018.00097"},{"key":"8_CR4","unstructured":"Chen, C., Li, O., Barnett, A., Su, J., Rudin, C.: This looks like that: deep learning for interpretable image recognition. In: NeurIPS (2019)"},{"key":"8_CR5","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"619","DOI":"10.1007\/978-3-030-85616-8_36","volume-title":"Human-Computer Interaction \u2013 INTERACT 2021","author":"M Chromik","year":"2021","unstructured":"Chromik, M., Butz, A.: Human-XAI interaction: a review and design principles for explanation user interfaces. In: Ardito, C., et al. (eds.) INTERACT 2021. LNCS, vol. 12933, pp. 619\u2013640. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-85616-8_36"},{"key":"8_CR6","doi-asserted-by":"publisher","unstructured":"Desai, S., Ramaswamy, H.G.: Ablation-CAM: visual explanations for deep convolutional network via gradient-free localization. In: 2020 IEEE Winter Conference on Applications of Computer Vision (WACV), pp. 972\u2013980 (2020). https:\/\/doi.org\/10.1109\/WACV45572.2020.9093360","DOI":"10.1109\/WACV45572.2020.9093360"},{"key":"8_CR7","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"56","DOI":"10.1007\/978-3-030-49760-6_4","volume-title":"Design, User Experience, and Usability. Design for Contemporary Interactive Environments","author":"JJ Ferreira","year":"2020","unstructured":"Ferreira, J.J., Monteiro, M.S.: What are people doing about XAI user experience? A survey on AI explainability research and practice. In: Marcus, A., Rosenzweig, E. (eds.) HCII 2020. LNCS, vol. 12201, pp. 56\u201373. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-49760-6_4"},{"key":"8_CR8","doi-asserted-by":"crossref","unstructured":"Ghosh, S., Shet, R., Amon, P., Hutter, A., Kaup, A.: Robustness of deep convolutional neural networks for image degradations. In: 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 2916\u20132920. IEEE (2018)","DOI":"10.1109\/ICASSP.2018.8461907"},{"key":"8_CR9","unstructured":"Gomez, T., Ling, S., Fr\u00e9our, T., Mouch\u00e8re, H.: Improve the interpretability of attention: a fast, accurate, and interpretable high-resolution attention model (2021)"},{"key":"8_CR10","unstructured":"Guo, C., Pleiss, G., Sun, Y., Weinberger, K.Q.: On calibration of modern neural networks. In: Precup, D., Teh, Y.W. (eds.) Proceedings of the 34th International Conference on Machine Learning. Proceedings of Machine Learning Research, 06\u201311 August 2017, vol. 70, pp. 1321\u20131330. PMLR (2017)"},{"key":"8_CR11","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"8_CR12","unstructured":"Hu, T., Qi, H.: See better before looking closer: weakly supervised data augmentation network for fine-grained visual classification. CoRR abs\/1901.09891 (2019)"},{"key":"8_CR13","doi-asserted-by":"crossref","unstructured":"Huang, Z., Li, Y.: Interpretable and accurate fine-grained recognition via region grouping (2020)","DOI":"10.1109\/CVPR42600.2020.00869"},{"key":"8_CR14","doi-asserted-by":"crossref","unstructured":"Jung, H., Oh, Y.: LIFT-CAM: towards better explanations for class activation mapping. arXiv arXiv:2102.05228 (2021)","DOI":"10.1109\/ICCV48922.2021.00137"},{"key":"8_CR15","doi-asserted-by":"publisher","unstructured":"Kenny, E.M., Ford, C., Quinn, M., Keane, M.T.: Explaining black-box classifiers using post-hoc explanations-by-example: the effect of explanations and error-rates in XAI user studies. Artif. Intell. 294, 103459 (2021). https:\/\/doi.org\/10.1016\/j.artint.2021.103459","DOI":"10.1016\/j.artint.2021.103459"},{"key":"8_CR16","unstructured":"Lundberg, S.M., Lee, S.I.: A unified approach to interpreting model predictions. In: Proceedings of the 31st International Conference on Neural Information Processing Systems, pp. 4768\u20134777 (2017)"},{"key":"8_CR17","unstructured":"McInnes, L., Healy, J., Melville, J.: UMAP: uniform manifold approximation and projection for dimension reduction (2020)"},{"key":"8_CR18","doi-asserted-by":"crossref","unstructured":"Nauta, M., van Bree, R., Seifert, C.: Neural prototype trees for interpretable fine-grained image recognition (2021)","DOI":"10.1109\/CVPR46437.2021.01469"},{"key":"8_CR19","unstructured":"Nixon, J., Dusenberry, M.W., Zhang, L., Jerfel, G., Tran, D.: Measuring calibration in deep learning. In: CVPR Workshops, vol. 2 (2019)"},{"key":"8_CR20","unstructured":"Petsiuk, V., Das, A., Saenko, K.: RISE: randomized input sampling for explanation of black-box models (2018)"},{"key":"8_CR21","doi-asserted-by":"crossref","unstructured":"Selvaraju, R.R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., Batra, D.: Grad-CAM: visual explanations from deep networks via gradient-based localization. In: 2017 IEEE International Conference on Computer Vision (ICCV), pp. 618\u2013626 (2017)","DOI":"10.1109\/ICCV.2017.74"},{"key":"8_CR22","unstructured":"Slack, D., Hilgard, A., Singh, S., Lakkaraju, H.: Reliable post hoc explanations: modeling uncertainty in explainability. In: Advances in Neural Information Processing Systems, vol. 34 (2021)"},{"key":"8_CR23","unstructured":"Smilkov, D., Thorat, N., Kim, B., Vi\u00e9gas, F., Wattenberg, M.: SmoothGrad: removing noise by adding noise (2017)"},{"key":"8_CR24","doi-asserted-by":"crossref","unstructured":"Tsai, C.H., Brusilovsky, P.: Evaluating visual explanations for similarity-based recommendations: user perception and performance, pp. 22\u201330. Association for Computing Machinery, New York (2019)","DOI":"10.1145\/3320435.3320465"},{"key":"8_CR25","doi-asserted-by":"publisher","unstructured":"van der Waa, J., Nieuwburg, E., Cremers, A., Neerincx, M.: Evaluating XAI: a comparison of rule-based and example-based explanations. Artif. Intell. 291, 103404 (2021). https:\/\/doi.org\/10.1016\/j.artint.2020.103404","DOI":"10.1016\/j.artint.2020.103404"},{"key":"8_CR26","unstructured":"Wah, C., Branson, S., Welinder, P., Perona, P., Belongie, S.: The Caltech-UCSD Birds-200-2011 Dataset. Technical report, CNS-TR-2011-001, California Institute of Technology (2011)"},{"key":"8_CR27","doi-asserted-by":"crossref","unstructured":"Wang, H., et al.: Score-CAM: score-weighted visual explanations for convolutional neural networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 24\u201325 (2020)","DOI":"10.1109\/CVPRW50498.2020.00020"},{"key":"8_CR28","unstructured":"Zhang, J., Kailkhura, B., Han, T.Y.J.: Mix-n-Match: ensemble and compositional methods for uncertainty calibration in deep learning. In: Singh, A., et al. (eds.) Proceedings of the 37th International Conference on Machine Learning. Proceedings of Machine Learning Research, 13\u201318 July 2020, vol. 119, pp. 11117\u201311128. PMLR (2020)"}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition and Artificial Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-09037-0_8","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,6,27]],"date-time":"2023-06-27T15:25:56Z","timestamp":1687879556000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-09037-0_8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031090363","9783031090370"],"references-count":28,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-09037-0_8","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"2 June 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICPRAI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Pattern Recognition and Artificial Intelligence","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Paris","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"France","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"1 June 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"3 June 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"3","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icprai2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/icprai2022.sciencesconf.org\/1.6.If","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}