{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,19]],"date-time":"2026-02-19T12:42:05Z","timestamp":1771504925724,"version":"3.50.1"},"reference-count":62,"publisher":"Springer Science and Business Media LLC","issue":"5","license":[{"start":{"date-parts":[[2024,11,23]],"date-time":"2024-11-23T00:00:00Z","timestamp":1732320000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2024,11,23]],"date-time":"2024-11-23T00:00:00Z","timestamp":1732320000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"name":"Fraunhofer-Institut f\u00fcr Techno- und Wirtschaftsmathematik ITWM"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int J Comput Vis"],"published-print":{"date-parts":[[2025,5]]},"abstract":"<jats:title>Abstract<\/jats:title>\n          <jats:p>In this paper, we present an approach for evaluating attribution maps, which play a central role in interpreting the predictions of convolutional neural networks (CNNs). We show that the widely used insertion\/deletion metrics are susceptible to distribution shifts that affect the reliability of the ranking. Our method proposes to replace pixel modifications with adversarial perturbations, which provides a more robust evaluation framework. By using smoothness and monotonicity measures, we illustrate the effectiveness of our approach in correcting distribution shifts. In addition, we conduct the most comprehensive quantitative and qualitative assessment of attribution maps to date. Introducing baseline attribution maps as sanity checks, we find that our metric is the only contender to pass all checks. Using Kendall\u2019s <jats:inline-formula>\n              <jats:alternatives>\n                <jats:tex-math>$$\\tau $$<\/jats:tex-math>\n                <mml:math xmlns:mml=\"http:\/\/www.w3.org\/1998\/Math\/MathML\">\n                  <mml:mi>\u03c4<\/mml:mi>\n                <\/mml:math>\n              <\/jats:alternatives>\n            <\/jats:inline-formula> rank correlation coefficient, we show the increased consistency of our metric across 15 dataset-architecture combinations. Of the 16 attribution maps tested, our results clearly show SmoothGrad to be the best map currently available. This research makes an important contribution to the development of attribution maps by providing a reliable and consistent evaluation framework. To ensure reproducibility, we will provide the code along with our results.<\/jats:p>","DOI":"10.1007\/s11263-024-02282-6","type":"journal-article","created":{"date-parts":[[2024,11,23]],"date-time":"2024-11-23T11:49:47Z","timestamp":1732362587000},"page":"2392-2409","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["Reliable Evaluation of Attribution Maps in CNNs: A Perturbation-Based Approach"],"prefix":"10.1007","volume":"133","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-7523-5694","authenticated-orcid":false,"given":"Lars","family":"Nieradzik","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9821-1636","authenticated-orcid":false,"given":"Henrike","family":"Stephani","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1327-1243","authenticated-orcid":false,"given":"Janis","family":"Keuper","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,23]]},"reference":[{"key":"2282_CR1","unstructured":"Adebayo, J., Gilmer, J., Muelly, M., Goodfellow, I., Hardt, M., & Kim, B. (2018). Sanity checks for saliency maps. In: Bengio, S., Wallach, H., Larochelle, H., Grauman, K., Cesa-Bianchi, N., Garnett, R. (eds.) Advances in neural information processing systems, vol. 31. Curran Associates, Inc. https:\/\/proceedings.neurips.cc\/paper\/2018\/file\/294a8ed24b1ad22ec2e7efea049b8737-Paper.pdf"},{"key":"2282_CR2","doi-asserted-by":"publisher","first-page":"14410","DOI":"10.1109\/ACCESS.2018.2807385","volume":"6","author":"N Akhtar","year":"2018","unstructured":"Akhtar, N., & Mian, A. (2018). Threat of adversarial attacks on deep learning in computer vision: A survey. IEEE Access, 6, 14410\u201314430.","journal-title":"IEEE Access"},{"key":"2282_CR3","unstructured":"Ancona, M., Ceolini, E., \u00d6ztireli, A.C., & Gross, M.H. (2017). A unified view of gradient-based attribution methods for deep neural networks. arXiv:1711.06104"},{"key":"2282_CR4","unstructured":"Byun, S.-Y., & Lee, W. (2023). Recipro-CAM: Fast gradient-free visual explanations for convolutional neural networks"},{"key":"2282_CR5","doi-asserted-by":"publisher","first-page":"679","DOI":"10.1109\/TPAMI.1986.4767851","volume":"6","author":"J Canny","year":"1986","unstructured":"Canny, J. (1986). A computational approach to edge detection. IEEE Transactions on Pattern Analysis and Machine Intelligence, 6, 679\u2013698.","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence"},{"key":"2282_CR6","doi-asserted-by":"crossref","unstructured":"Carlini, N., & Wagner, D.A. (2016). Towards evaluating the robustness of neural networks. arXiv:1608.04644","DOI":"10.1109\/SP.2017.49"},{"key":"2282_CR7","doi-asserted-by":"publisher","unstructured":"Chattopadhay, A., Sarkar, A., Howlader, P., & Balasubramanian, V.N. (2018). Grad-CAM++: Generalized gradient-based visual explanations for deep convolutional networks. In: 2018 IEEE winter conference on applications of computer vision (WACV). IEEE. https:\/\/doi.org\/10.1109\/wacv.2018.00097","DOI":"10.1109\/wacv.2018.00097"},{"key":"2282_CR8","doi-asserted-by":"publisher","unstructured":"Deng, J., Dong, W., Socher, R., Li, L.-J., Li, K., & Fei-Fei, L. (2009). Imagenet: A large-scale hierarchical image database. In: 2009 IEEE conference on computer vision and pattern recognition, pp. 248\u2013255 . https:\/\/doi.org\/10.1109\/CVPR.2009.5206848","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"2282_CR9","doi-asserted-by":"publisher","unstructured":"Desai, S., & Ramaswamy, H.G. (2020). Ablation-cam: Visual explanations for deep convolutional network via gradient-free localization. In: 2020 IEEE winter conference on applications of computer vision (WACV), pp. 972\u2013980. https:\/\/doi.org\/10.1109\/WACV45572.2020.9093360","DOI":"10.1109\/WACV45572.2020.9093360"},{"key":"2282_CR10","doi-asserted-by":"crossref","unstructured":"Ding, X., Zhang, X., Ma, N., Han, J., Ding, G., & Sun, J. (2021). Repvgg: Making VGG-style convnets great again. arXiv:2101.03697","DOI":"10.1109\/CVPR46437.2021.01352"},{"key":"2282_CR11","doi-asserted-by":"publisher","unstructured":"Englebert, A., Cornu, O., & De\u00a0Vleeschouwer, C. (2022). Poly-CAM: High resolution class activation map for convolutional neural networks. arXiv. https:\/\/doi.org\/10.48550\/ARXIV.2204.13359 . arXiv:2204.13359","DOI":"10.48550\/ARXIV.2204.13359"},{"issue":"2","key":"2282_CR12","doi-asserted-by":"publisher","first-page":"303","DOI":"10.1007\/s11263-009-0275-4","volume":"88","author":"M Everingham","year":"2009","unstructured":"Everingham, M., Gool, L. V., Williams, C. K. I., Winn, J., & Zisserman, A. (2009). The pascal visual object classes (VOC) challenge. International Journal of Computer Vision, 88(2), 303\u2013338. https:\/\/doi.org\/10.1007\/s11263-009-0275-4","journal-title":"International Journal of Computer Vision"},{"key":"2282_CR13","unstructured":"Fernandez, F.-G. (2020). TorchCAM: Class activation explorer. GitHub."},{"key":"2282_CR14","doi-asserted-by":"crossref","unstructured":"Fong, R., & Vedaldi, A. (2017). Interpretable explanations of black boxes by meaningful perturbation. arXiv:1704.03296","DOI":"10.1109\/ICCV.2017.371"},{"key":"2282_CR15","doi-asserted-by":"crossref","unstructured":"Fong, R., Patrick, M., & Vedaldi, A. (2019). Understanding deep networks via extremal perturbations and smooth masks. CoRR arXiv:1910.08485","DOI":"10.1109\/ICCV.2019.00304"},{"key":"2282_CR16","unstructured":"Fu, R., Hu, Q., Dong, X., Guo, Y., Gao, Y., & Li, B. (2020). Axiom-based grad-cam: Towards accurate visualization and explanation of cnns. CoRR arXiv:2008.02312"},{"key":"2282_CR17","doi-asserted-by":"crossref","unstructured":"Gilpin, L.H., Bau, D., Yuan, B.Z., Bajwa, A., Specter, M., & Kagal, L. (2018). Explaining explanations: An overview of interpretability of machine learning. In: 2018 IEEE 5th international conference on data science and advanced analytics (DSAA), pp. 80\u201389. IEEE","DOI":"10.1109\/DSAA.2018.00018"},{"key":"2282_CR18","doi-asserted-by":"crossref","unstructured":"Gomez, T., Fr\u00e9our, T., & Mouch\u00e8re, H. (2022). Metrics for saliency map evaluation of deep learning explanation methods. CoRR arXiv:2201.13291","DOI":"10.1007\/978-3-031-09037-0_8"},{"key":"2282_CR19","doi-asserted-by":"publisher","unstructured":"Goodfellow, I.J., Shlens, J., & Szegedy, C. (2014). Explaining and harnessing adversarial examples. arXiv . https:\/\/doi.org\/10.48550\/ARXIV.1412.6572 . https:\/\/arxiv.org\/abs\/1412.6572","DOI":"10.48550\/ARXIV.1412.6572"},{"key":"2282_CR20","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., & Sun, J. (2015). Deep residual learning for image recognition. CoRR arXiv:1512.03385","DOI":"10.1109\/CVPR.2016.90"},{"key":"2282_CR21","doi-asserted-by":"crossref","unstructured":"Huang, G., Liu, Z., Weinberger, K.Q. (2018). Densely connected convolutional networks. CoRR arXiv:1608.06993","DOI":"10.1109\/CVPR.2017.243"},{"key":"2282_CR22","doi-asserted-by":"publisher","first-page":"5875","DOI":"10.1109\/TIP.2021.3089943","volume":"30","author":"P-T Jiang","year":"2021","unstructured":"Jiang, P.-T., Zhang, C.-B., Hou, Q., Cheng, M.-M., & Wei, Y. (2021). Layercam: Exploring hierarchical class activation maps for localization. IEEE Transactions on Image Processing, 30, 5875\u20135888. https:\/\/doi.org\/10.1109\/TIP.2021.3089943","journal-title":"IEEE Transactions on Image Processing"},{"key":"2282_CR23","doi-asserted-by":"crossref","unstructured":"Kapishnikov, A., Venugopalan, S., Avci, B., Wedin, B., Terry, M., & Bolukbasi, T. (2021). Guided integrated gradients: An adaptive path method for removing noise. CoRR arXiv:2106.09788","DOI":"10.1109\/CVPR46437.2021.00501"},{"key":"2282_CR24","unstructured":"LeCun, Y., Bengio, Y., et al. (1998). Convolutional networks for images, speech, and time series. In: The handbook of brain theory and neural networks. Convolutional networks for images, speech, and time series, pp. 255\u2013258. MIT Press, Cambridge, MA, USA."},{"key":"2282_CR25","unstructured":"Li, Y., Cheng, M., Hsieh, C., & Lee, T.C.M. (2021). A review of adversarial attack and defense for classification methods. CoRR arXiv:2111.09961"},{"key":"2282_CR26","doi-asserted-by":"publisher","unstructured":"Li, H., Li, Z., Ma, R., & Wu, T. (2022). FD-CAM: Improving faithfulness and discriminability of visual explanation for CNNs. arXiv. https:\/\/doi.org\/10.48550\/ARXIV.2206.08792 . arXiv:2206.08792","DOI":"10.48550\/ARXIV.2206.08792"},{"issue":"1","key":"2282_CR27","doi-asserted-by":"publisher","first-page":"72","DOI":"10.1109\/TDSC.2018.2874243","volume":"18","author":"B Liang","year":"2021","unstructured":"Liang, B., Li, H., Su, M., Li, X., Shi, W., & Wang, X. (2021). Detecting adversarial image examples in deep neural networks with adaptive noise reduction. IEEE Transactions on Dependable and Secure Computing, 18(1), 72\u201385. https:\/\/doi.org\/10.1109\/TDSC.2018.2874243","journal-title":"IEEE Transactions on Dependable and Secure Computing"},{"key":"2282_CR28","doi-asserted-by":"crossref","unstructured":"Lin, T., Maire, M., Belongie, S.J., Bourdev, L.D., Girshick, R.B., Hays, J., Perona, P., Ramanan, D., Doll\u00e1r, P., & Zitnick, C.L. (2014). Microsoft COCO: Common objects in context. CoRR arXiv:1405.0312","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"2282_CR29","doi-asserted-by":"crossref","unstructured":"Liu, Z., Mao, H., Wu, C., Feichtenhofer, C., Darrell, T., & Xie, S. (2022). A convnet for the 2020s. CoRR arXiv:2201.03545","DOI":"10.1109\/CVPR52688.2022.01167"},{"key":"2282_CR30","doi-asserted-by":"publisher","unstructured":"Madry, A., Makelov, A., Schmidt, L., Tsipras, D., & Vladu, A. (2017). Towards deep learning models resistant to adversarial attacks. arXiv. https:\/\/doi.org\/10.48550\/ARXIV.1706.06083 . arXiv:1706.06083","DOI":"10.48550\/ARXIV.1706.06083"},{"key":"2282_CR31","unstructured":"Madry, A., Makelov, A., Schmidt, L., Tsipras, D., & Vladu, A. (2019). Towards deep learning models resistant to adversarial attacks."},{"key":"2282_CR32","doi-asserted-by":"crossref","unstructured":"Muhammad, M.B., & Yeasin, M. (2020). Eigen-cam: Class activation map using principal components. CoRR arXiv:2008.00299","DOI":"10.1109\/IJCNN48605.2020.9206626"},{"key":"2282_CR33","unstructured":"Naidu, R., Ghosh, A., Maurya, Y., K, S.R.N., & Kundu, S.S. (2020). IS-CAM: Integrated score-cam for axiomatic-based explanations. CoRR arXiv:2010.03023"},{"issue":"1","key":"2282_CR34","doi-asserted-by":"publisher","first-page":"45","DOI":"10.1177\/1536867x0200200103","volume":"2","author":"R Newson","year":"2002","unstructured":"Newson, R. (2002). Parameters behind \u201cnonparametric\u2019\u2019 statistics: Kendall\u2019s tau, somers\u2019 d and median differences. The Stata Journal: Promoting communications on statistics and Stata, 2(1), 45\u201364. https:\/\/doi.org\/10.1177\/1536867x0200200103","journal-title":"The Stata Journal: Promoting communications on statistics and Stata"},{"key":"2282_CR35","doi-asserted-by":"crossref","unstructured":"Nieradzik, L., Stephani, H., Sieburg-Rockel, J., Helmling, S., Olbrich, A., & Keuper, J. (2024). Challenging the black box: A comprehensive evaluation of attribution maps of CNN applications in agriculture and forestry. In: Proceedings of the 19th international joint conference on computer vision, imaging and computer graphics theory and applications, Volume 2: VISAPP, 2024, pp. 483\u2013492","DOI":"10.5220\/0012363400003660"},{"key":"2282_CR36","unstructured":"Omeiza, D., Speakman, S., Cintas, C., & Weldemariam, K. (2019). Smooth grad-cam++: An enhanced inference level visualization technique for deep convolutional neural network models. CoRR arXiv:1908.01224"},{"key":"2282_CR37","unstructured":"PAIR, G. (2022). Saliency library. GitHub"},{"key":"2282_CR38","doi-asserted-by":"crossref","unstructured":"Parkhi, O.M., Vedaldi, A., Zisserman, A., & Jawahar, C.V. (2012). Cats and dogs. In: IEEE conference on computer vision and pattern recognition.","DOI":"10.1109\/CVPR.2012.6248092"},{"key":"2282_CR39","doi-asserted-by":"publisher","unstructured":"Petsiuk, V., Das, A., & Saenko, K. (2018). RISE: Randomized input sampling for explanation of black-box models. arXiv . https:\/\/doi.org\/10.48550\/ARXIV.1806.07421 . arXiv:1806.07421","DOI":"10.48550\/ARXIV.1806.07421"},{"key":"2282_CR40","unstructured":"Petsiuk, V., Das, A., & Saenko, K. (2018). RISE: Randomized input sampling for explanation of black-box models. CoRR arXiv:1806.07421"},{"key":"2282_CR41","doi-asserted-by":"crossref","unstructured":"Poppi, S., Cornia, M., Baraldi, L., & Cucchiara, R. (2021). Revisiting the evaluation of class activation mapping for explainability: A novel metric and experimental analysis. CoRR arXiv:2104.10252","DOI":"10.1109\/CVPRW53098.2021.00260"},{"key":"2282_CR42","doi-asserted-by":"publisher","unstructured":"Raatikainen, L., & Rahtu, E. (2022). The weighting game: Evaluating quality of explainability methods. arXiv . https:\/\/doi.org\/10.48550\/ARXIV.2208.06175 . arXiv:2208.06175","DOI":"10.48550\/ARXIV.2208.06175"},{"key":"2282_CR43","unstructured":"Rauber, J., Brendel, W., & Bethge, M. (2017). Foolbox v0.8.0: A python toolbox to benchmark the robustness of machine learning models. CoRR arXiv:1707.04131"},{"key":"2282_CR44","doi-asserted-by":"crossref","unstructured":"Ribeiro, M.T., Singh, S., & Guestrin, C. (2016). \"why should I trust you?\": Explaining the predictions of any classifier. CoRR arXiv:1602.04938","DOI":"10.18653\/v1\/N16-3020"},{"issue":"2","key":"2282_CR45","doi-asserted-by":"publisher","first-page":"336","DOI":"10.1007\/s11263-019-01228-7","volume":"128","author":"RR Selvaraju","year":"2019","unstructured":"Selvaraju, R. R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., & Batra, D. (2019). Grad-CAM: Visual explanations from deep networks via gradient-based localization. International Journal of Computer Vision, 128(2), 336\u2013359. https:\/\/doi.org\/10.1007\/s11263-019-01228-7","journal-title":"International Journal of Computer Vision"},{"key":"2282_CR46","unstructured":"Shi, X., Khademi, S., Li, Y., & Gemert, J. (2020). Zoom-cam: Generating fine-grained pixel annotations from image labels. CoRR arXiv:2010.08644"},{"key":"2282_CR47","unstructured":"Simonyan, K., Vedaldi, A., & Zisserman, A. (2014). Deep inside convolutional networks: Visualising image classification models and saliency maps. In: Bengio, Y., LeCun, Y. (eds.) 2nd international conference on learning representations, ICLR 2014, Banff, AB, Canada, Apr 14\u201316, 2014, Workshop track proceedings. arXiv:1312.6034"},{"key":"2282_CR48","unstructured":"Smilkov, D., Thorat, N., Kim, B., Vi\u00e9gas, F.B., & Wattenberg, M. (2017). Smoothgrad: Removing noise by adding noise. CoRR arXiv:1706.03825"},{"key":"2282_CR49","unstructured":"Springenberg, J.T., Dosovitskiy, A., Brox, T., & Riedmiller, M.A. (2015). Striving for simplicity: The all convolutional net. In: Bengio, Y., LeCun, Y. (eds.) 3rd International conference on learning representations, ICLR 2015, San Diego, CA, USA, May 7\u20139, 2015, Workshop track proceedings. arXiv:1412.6806"},{"key":"2282_CR50","unstructured":"Su, J., Vargas, D.V., & Sakurai, K. (2017). One pixel attack for fooling deep neural networks. CoRR arXiv:1710.08864"},{"key":"2282_CR51","unstructured":"Sundararajan, M., Taly, A., & Yan, Q. (2017). Axiomatic attribution for deep networks. CoRR arXiv:1703.01365 ."},{"key":"2282_CR52","unstructured":"Tan, M., & Le, Q.V. (2019). Efficientnet: Rethinking model scaling for convolutional neural networks. CoRR arXiv:1905.11946"},{"key":"2282_CR53","doi-asserted-by":"publisher","unstructured":"Wang, H., et al. (2020) Score-CAM: Score-weighted visual explanations for convolutional neural networks. In: 2020 IEEE\/CVF conference on computer vision and pattern recognition workshops (CVPRW), pp. 111\u2013119. Seattle, WA, USA. https:\/\/doi.org\/10.1109\/CVPRW50498.2020.00020.","DOI":"10.1109\/CVPRW50498.2020.00020"},{"key":"2282_CR54","doi-asserted-by":"crossref","unstructured":"Wang, X., Peng, Y., Lu, L., Lu, Z., Bagheri, M., & Summers, R.M. (2017). Chestx-ray8: Hospital-scale chest x-ray database and benchmarks on weakly-supervised classification and localization of common thorax diseases. In: CVPR, pp. 3462\u20133471. IEEE Computer Society. http:\/\/dblp.uni-trier.de\/db\/conf\/cvpr\/cvpr2017.html#WangPLLBS17","DOI":"10.1109\/CVPR.2017.369"},{"key":"2282_CR55","doi-asserted-by":"crossref","unstructured":"Wang, H., Wang, Z., Du, M., Yang, F., Zhang, Z., Ding, S., Mardziel, P., & Hu, X. (2020). Score-CAM: Score-weighted visual explanations for convolutional neural networks.","DOI":"10.1109\/CVPRW50498.2020.00020"},{"key":"2282_CR56","doi-asserted-by":"crossref","unstructured":"Xu, W., Evans, D., & Qi, Y. (2017). Feature squeezing: Detecting adversarial examples in deep neural networks. CoRR arXiv:1704.01155","DOI":"10.14722\/ndss.2018.23198"},{"key":"2282_CR57","doi-asserted-by":"crossref","unstructured":"Xu, S., Venugopalan, S., & Sundararajan, M. (2020). Attribution in scale and space. CoRR arXiv:2004.03383","DOI":"10.1109\/CVPR42600.2020.00970"},{"key":"2282_CR58","doi-asserted-by":"crossref","unstructured":"Zar\u00e1ndy, \u00c1., Rekeczky, C., Szolgay, P., & Chua, L.O. (2015). Overview of cnn research: 25 years history and the current trends. In: 2015 IEEE international symposium on circuits and systems (ISCAS), pp. 401\u2013404. IEEE","DOI":"10.1109\/ISCAS.2015.7168655"},{"key":"2282_CR59","doi-asserted-by":"crossref","unstructured":"Zeiler, M.D., & Fergus, R. (2013). Visualizing and understanding convolutional networks. CoRR arXiv:1311.2901","DOI":"10.1007\/978-3-319-10590-1_53"},{"key":"2282_CR60","doi-asserted-by":"crossref","unstructured":"Zhang, J., Lin, Z., Brandt, J., Shen, X., & Sclaroff, S. (2016). Top-down neural attention by excitation backprop. CoRR arXiv:1608.00507","DOI":"10.1007\/978-3-319-46493-0_33"},{"key":"2282_CR61","doi-asserted-by":"crossref","unstructured":"Zhang, Q., Rao, L., & Yang, Y. (2021). Group-cam: Group score-weighted visual explanations for deep convolutional networks. CoRR arXiv:2103.13859","DOI":"10.1109\/CVPRW50498.2020.00020"},{"issue":"5","key":"2282_CR62","doi-asserted-by":"publisher","first-page":"726","DOI":"10.1109\/TETCI.2021.3100641","volume":"5","author":"Y Zhang","year":"2021","unstructured":"Zhang, Y., Ti\u0148o, P., Leonardis, A., & Tang, K. (2021). A survey on neural network interpretability. IEEE Transactions on Emerging Topics in Computational Intelligence, 5(5), 726\u2013742.","journal-title":"IEEE Transactions on Emerging Topics in Computational Intelligence"}],"container-title":["International Journal of Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11263-024-02282-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11263-024-02282-6\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11263-024-02282-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,4,17]],"date-time":"2025-04-17T06:00:02Z","timestamp":1744869602000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11263-024-02282-6"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,23]]},"references-count":62,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2025,5]]}},"alternative-id":["2282"],"URL":"https:\/\/doi.org\/10.1007\/s11263-024-02282-6","relation":{},"ISSN":["0920-5691","1573-1405"],"issn-type":[{"value":"0920-5691","type":"print"},{"value":"1573-1405","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,11,23]]},"assertion":[{"value":"8 September 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"20 October 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"23 November 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The author(s) declare that they have no Conflict of interest to disclose.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}