{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,11]],"date-time":"2026-03-11T11:53:55Z","timestamp":1773230035333,"version":"3.50.1"},"reference-count":42,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2025,11,1]],"date-time":"2025-11-01T00:00:00Z","timestamp":1761955200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2025,11,1]],"date-time":"2025-11-01T00:00:00Z","timestamp":1761955200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2025,11,1]],"date-time":"2025-11-01T00:00:00Z","timestamp":1761955200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2025,11,1]],"date-time":"2025-11-01T00:00:00Z","timestamp":1761955200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2025,11,1]],"date-time":"2025-11-01T00:00:00Z","timestamp":1761955200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2025,11,1]],"date-time":"2025-11-01T00:00:00Z","timestamp":1761955200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,11,1]],"date-time":"2025-11-01T00:00:00Z","timestamp":1761955200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/501100004835","name":"Zhejiang University","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100004835","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["42072232"],"award-info":[{"award-number":["42072232"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Computer Vision and Image Understanding"],"published-print":{"date-parts":[[2025,11]]},"DOI":"10.1016\/j.cviu.2025.104523","type":"journal-article","created":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T02:26:15Z","timestamp":1760063175000},"page":"104523","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":0,"special_numbering":"C","title":["IP-CAM: Class activation mapping based on importance weights and principal-component weights for better and simpler visual explanations"],"prefix":"10.1016","volume":"261","author":[{"given":"Wenyi","family":"Zhang","sequence":"first","affiliation":[]},{"given":"Haoran","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Xisheng","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Xiaohua","family":"Shen","sequence":"additional","affiliation":[]},{"given":"Lejun","family":"Zou","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.cviu.2025.104523_b1","doi-asserted-by":"crossref","first-page":"14","DOI":"10.1016\/j.inffus.2021.11.008","article-title":"CLEVR-XAI: A benchmark dataset for the ground truth evaluation of neural network explanations","volume":"81","author":"Arras","year":"2022","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.cviu.2025.104523_b2","series-title":"2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops","first-page":"776","article-title":"Analysis of efficient CNN design techniques for semantic segmentation","author":"Briot","year":"2018"},{"key":"10.1016\/j.cviu.2025.104523_b3","series-title":"2018 IEEE Winter Conference on Applications of Computer Vision","first-page":"839","article-title":"Grad-CAM++: Generalized gradient-based visual explanations for deep convolutional networks","author":"Chattopadhay","year":"2018"},{"key":"10.1016\/j.cviu.2025.104523_b4","series-title":"Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2023","first-page":"173","article-title":"AME-CAM: attentive multiple-exit CAM for\u00a0weakly supervised segmentation on\u00a0MRI brain tumor","author":"Chen","year":"2023"},{"issue":"3","key":"10.1016\/j.cviu.2025.104523_b5","doi-asserted-by":"crossref","first-page":"2385","DOI":"10.1109\/TII.2022.3156658","article-title":"Denoising aggregation of graph neural networks by using principal component analysis","volume":"19","author":"Dong","year":"2023","journal-title":"IEEE Trans. Ind. Inform."},{"key":"10.1016\/j.cviu.2025.104523_b6","series-title":"An image is worth 16x16 words: transformers for image recognition at scale","author":"Dosovitskiy","year":"2021"},{"issue":"4","key":"10.1016\/j.cviu.2025.104523_b7","doi-asserted-by":"crossref","first-page":"89","DOI":"10.1007\/s00138-024-01567-7","article-title":"Poly-cam: high resolution class activation map for convolutional neural networks","volume":"35","author":"Englebert","year":"2024","journal-title":"Mach. Vis. Appl."},{"key":"10.1016\/j.cviu.2025.104523_b8","series-title":"2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition","first-page":"10697","article-title":"Attention branch network: Learning of attention mechanism for visual explanation","author":"Fukui","year":"2019"},{"issue":"7","key":"10.1016\/j.cviu.2025.104523_b9","doi-asserted-by":"crossref","first-page":"3163","DOI":"10.1109\/JBHI.2022.3153671","article-title":"A convolutional neural network and graph convolutional network based framework for classification of breast histopathological images","volume":"26","author":"Gao","year":"2022","journal-title":"IEEE J. Biomed. Heal. Inform."},{"key":"10.1016\/j.cviu.2025.104523_b10","series-title":"2016 IEEE Conference on Computer Vision and Pattern Recognition","first-page":"770","article-title":"Deep residual learning for image recognition","author":"He","year":"2016"},{"key":"10.1016\/j.cviu.2025.104523_b11","series-title":"2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition","first-page":"7132","article-title":"Squeeze-and-excitation networks","author":"Hu","year":"2018"},{"key":"10.1016\/j.cviu.2025.104523_b12","doi-asserted-by":"crossref","DOI":"10.1016\/j.knosys.2022.109287","article-title":"Augmented score-CAM: High resolution visual interpretations for deep neural networks","volume":"252","author":"Ibrahim","year":"2022","journal-title":"Knowl.-Based Syst."},{"key":"10.1016\/j.cviu.2025.104523_b13","series-title":"Pattern Recognition and Computer Vision","first-page":"261","article-title":"SCC-CAM: weakly supervised segmentation on\u00a0brain tumor MRI with\u00a0similarity constraint and\u00a0causality","author":"Jiao","year":"2025"},{"issue":"2065","key":"10.1016\/j.cviu.2025.104523_b14","doi-asserted-by":"crossref","DOI":"10.1098\/rsta.2015.0202","article-title":"Principal component analysis: a review and recent developments","volume":"374","author":"Jolliffe","year":"2016","journal-title":"Philos. Trans. R. Soc. A: Math. Phys. Eng. Sci."},{"issue":"6245","key":"10.1016\/j.cviu.2025.104523_b15","doi-asserted-by":"crossref","first-page":"255","DOI":"10.1126\/science.aaa8415","article-title":"Machine learning: Trends, perspectives, and prospects","volume":"349","author":"Jordan","year":"2015","journal-title":"Science"},{"key":"10.1016\/j.cviu.2025.104523_b16","article-title":"Imagenet classification with deep convolutional neural networks","volume":"vol. 25","author":"Krizhevsky","year":"2012"},{"key":"10.1016\/j.cviu.2025.104523_b17","series-title":"2021 IEEE\/CVF International Conference on Computer Vision","first-page":"1335","article-title":"LFI-CAM: Learning feature importance for better visual explanation","author":"Lee","year":"2021"},{"key":"10.1016\/j.cviu.2025.104523_b18","doi-asserted-by":"crossref","DOI":"10.1016\/j.knosys.2023.110473","article-title":"Probabilistic orthogonal-signal-corrected principal component analysis","volume":"268","author":"Lee","year":"2023","journal-title":"Knowl.-Based Syst."},{"key":"10.1016\/j.cviu.2025.104523_b19","series-title":"FD-CAM: Improving faithfulness and discriminability of visual explanation for CNNs","author":"Li","year":"2022"},{"key":"10.1016\/j.cviu.2025.104523_b20","first-page":"1","article-title":"Pivotal-aware principal component analysis","author":"Li","year":"2023","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"issue":"5","key":"10.1016\/j.cviu.2025.104523_b21","doi-asserted-by":"crossref","first-page":"2486","DOI":"10.1109\/TGRS.2016.2645610","article-title":"Accurate object localization in remote sensing images based on convolutional neural networks","volume":"55","author":"Long","year":"2017","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"10.1016\/j.cviu.2025.104523_b22","doi-asserted-by":"crossref","first-page":"648","DOI":"10.1016\/j.neucom.2021.01.138","article-title":"Bayesian networks for interpretable machine learning and optimization","volume":"456","author":"Mihaljevi\u0107","year":"2021","journal-title":"Neurocomputing"},{"key":"10.1016\/j.cviu.2025.104523_b23","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1016\/j.artint.2018.07.007","article-title":"Explanation in artificial intelligence: Insights from the social sciences","volume":"267","author":"Miller","year":"2019","journal-title":"Artificial Intelligence"},{"key":"10.1016\/j.cviu.2025.104523_b24","series-title":"Interpretable Machine Learning","author":"Molnar","year":"2020"},{"key":"10.1016\/j.cviu.2025.104523_b25","series-title":"2020 International Joint Conference on Neural Networks","first-page":"1","article-title":"Eigen-CAM: class activation map using principal components","author":"Muhammad","year":"2020"},{"key":"10.1016\/j.cviu.2025.104523_b26","doi-asserted-by":"crossref","DOI":"10.1016\/j.knosys.2021.107473","article-title":"Exploring deep features and ECG attributes to detect cardiac rhythm classes","volume":"232","author":"Murat","year":"2021","journal-title":"Knowl.-Based Syst."},{"key":"10.1016\/j.cviu.2025.104523_b27","series-title":"IS-CAM: Integrated score-CAM for axiomatic-based explanations","author":"Naidu","year":"2020"},{"key":"10.1016\/j.cviu.2025.104523_b28","series-title":"Smooth grad-CAM++: An enhanced inference level visualization technique for deep convolutional neural network models","author":"Omeiza","year":"2019"},{"key":"10.1016\/j.cviu.2025.104523_b29","series-title":"Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining","first-page":"1135","article-title":"\u201cWhy should I trust you?\u201d: Explaining the predictions of any classifier","author":"Ribeiro","year":"2016"},{"issue":"3","key":"10.1016\/j.cviu.2025.104523_b30","doi-asserted-by":"crossref","first-page":"211","DOI":"10.1007\/s11263-015-0816-y","article-title":"ImageNet large scale visual recognition challenge","volume":"115","author":"Russakovsky","year":"2015","journal-title":"Int. J. Comput. Vis."},{"key":"10.1016\/j.cviu.2025.104523_b31","series-title":"ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing","first-page":"1775","article-title":"Integrated grad-cam: Sensitivity-aware visual explanation of deep convolutional networks via integrated gradient-based scoring","author":"Sattarzadeh","year":"2021"},{"key":"10.1016\/j.cviu.2025.104523_b32","series-title":"2017 IEEE International Conference on Computer Vision","first-page":"618","article-title":"Grad-CAM: Visual explanations from deep networks via gradient-based localization","author":"Selvaraju","year":"2017"},{"key":"10.1016\/j.cviu.2025.104523_b33","series-title":"Very deep convolutional networks for large-scale image recognition","author":"Simonyan","year":"2015"},{"key":"10.1016\/j.cviu.2025.104523_b34","series-title":"3rd International Conference on Learning Representations","first-page":"1","article-title":"Very deep convolutional networks for large-scale image recognition","author":"Simonyan","year":"2015"},{"key":"10.1016\/j.cviu.2025.104523_b35","series-title":"2022 IEEE\/CVF Winter Conference on Applications of Computer Vision","first-page":"1474","article-title":"Interpretable semantic photo geolocation","author":"Theiner","year":"2022"},{"issue":"7","key":"10.1016\/j.cviu.2025.104523_b36","doi-asserted-by":"crossref","first-page":"1863","DOI":"10.1007\/s10115-022-01690-9","article-title":"Trusting deep learning natural-language models via local and global explanations","volume":"64","author":"Ventura","year":"2022","journal-title":"Knowl. Inf. Syst."},{"key":"10.1016\/j.cviu.2025.104523_b37","series-title":"SS-CAM: Smoothed score-CAM for sharper visual feature localization","author":"Wang","year":"2020"},{"key":"10.1016\/j.cviu.2025.104523_b38","series-title":"2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops","first-page":"111","article-title":"Score-CAM: Score-weighted visual explanations for convolutional neural networks","author":"Wang","year":"2020"},{"issue":"1","key":"10.1016\/j.cviu.2025.104523_b39","doi-asserted-by":"crossref","first-page":"37","DOI":"10.1016\/0169-7439(87)80084-9","article-title":"Principal component analysis","volume":"2","author":"Wold","year":"1987","journal-title":"Chemometr. Intell. Lab. Syst."},{"issue":"1","key":"10.1016\/j.cviu.2025.104523_b40","doi-asserted-by":"crossref","first-page":"511","DOI":"10.1007\/s10115-023-01959-7","article-title":"Feature importance measure of a multilayer perceptron based on the presingle-connection layer","volume":"66","author":"Zhang","year":"2024","journal-title":"Knowl. Inf. Syst."},{"key":"10.1016\/j.cviu.2025.104523_b41","doi-asserted-by":"crossref","DOI":"10.1016\/j.cviu.2024.104101","article-title":"Opti-CAM: Optimizing saliency maps for interpretability","volume":"248","author":"Zhang","year":"2024","journal-title":"Comput. Vis. Image Underst."},{"key":"10.1016\/j.cviu.2025.104523_b42","series-title":"2016 IEEE Conference on Computer Vision and Pattern Recognition","first-page":"2921","article-title":"Learning deep features for discriminative localization","author":"Zhou","year":"2016"}],"container-title":["Computer Vision and Image Understanding"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S1077314225002462?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S1077314225002462?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,3,10]],"date-time":"2026-03-10T14:50:19Z","timestamp":1773154219000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S1077314225002462"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,11]]},"references-count":42,"alternative-id":["S1077314225002462"],"URL":"https:\/\/doi.org\/10.1016\/j.cviu.2025.104523","relation":{},"ISSN":["1077-3142"],"issn-type":[{"value":"1077-3142","type":"print"}],"subject":[],"published":{"date-parts":[[2025,11]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"IP-CAM: Class activation mapping based on importance weights and principal-component weights for better and simpler visual explanations","name":"articletitle","label":"Article Title"},{"value":"Computer Vision and Image Understanding","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.cviu.2025.104523","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2025 Elsevier Inc. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"104523"}}