{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,23]],"date-time":"2026-03-23T19:00:37Z","timestamp":1774292437413,"version":"3.50.1"},"reference-count":38,"publisher":"MDPI AG","issue":"9","license":[{"start":{"date-parts":[[2022,4,19]],"date-time":"2022-04-19T00:00:00Z","timestamp":1650326400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Remote Sensing"],"abstract":"<jats:p>There is a growing need for algorithms to automatically detect objects in satellite images. Object detection algorithms using deep learning have demonstrated a significant improvement in object detection performance. However, deep-learning models have difficulty in interpreting the features for inference. This difficulty is practically problematic when analyzing earth-observation images, which are often used as evidence for public decision-making. In addition, for the same reason, it is difficult to set an explicit policy or criteria to improve the models. To deal with these challenges, we introduce a feature attribution method that defines an approximate model and calculates the attribution of input features to the output of a deep-learning model. For the object detection models of satellite images with complex textures, we propose a method to visualize the basis of inference using pixel-wise feature attribution. Furthermore, we propose new methods for model evaluation, regularization, and data selection, based on feature attribution. Experimental results demonstrate the feasibility of the proposed methods for basis visualization and model evaluation. Moreover, the results illustrate that the model using the proposed regularization method can avoid over-fitting and achieve higher performance, and the proposed data selection method allows for the efficient selection of new training data.<\/jats:p>","DOI":"10.3390\/rs14091970","type":"journal-article","created":{"date-parts":[[2022,4,20]],"date-time":"2022-04-20T00:22:43Z","timestamp":1650414163000},"page":"1970","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":22,"title":["SHAP-Based Interpretable Object Detection Method for Satellite Imagery"],"prefix":"10.3390","volume":"14","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-3711-6053","authenticated-orcid":false,"given":"Hiroki","family":"Kawauchi","sequence":"first","affiliation":[{"name":"Department of Civil Engineering, The University of Tokyo, 7-3-1 Hongo, Bunkyo-ku, Tokyo 113-8656, Japan"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5789-252X","authenticated-orcid":false,"given":"Takashi","family":"Fuse","sequence":"additional","affiliation":[{"name":"Department of Civil Engineering, The University of Tokyo, 7-3-1 Hongo, Bunkyo-ku, Tokyo 113-8656, Japan"}]}],"member":"1968","published-online":{"date-parts":[[2022,4,19]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"1631","DOI":"10.1109\/TGRS.2007.890808","article-title":"Hurricane disaster assessments with image-driven data mining in high-resolution satellite imagery","volume":"45","author":"Barnes","year":"2007","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_2","doi-asserted-by":"crossref","first-page":"212","DOI":"10.1016\/j.isprsjprs.2006.02.002","article-title":"Updating topographic mapping in Great Britain using imagery from high-resolution satellite sensors","volume":"60","author":"Holland","year":"2006","journal-title":"ISPRS J. Photogramm. Remote Sens."},{"key":"ref_3","first-page":"W24","article-title":"Vehicles detection from very high resolution satellite imagery","volume":"36","author":"Gerhardinger","year":"2005","journal-title":"Int. Arch. Photogramm. Remote Sens."},{"key":"ref_4","doi-asserted-by":"crossref","first-page":"296","DOI":"10.1016\/j.isprsjprs.2019.11.023","article-title":"Object detection in optical remote sensing images: A survey and a new benchmark","volume":"159","author":"Li","year":"2020","journal-title":"ISPRS J. Photogramm. Remote Sens."},{"key":"ref_5","first-page":"1097","article-title":"Imagenet classification with deep convolutional neural networks","volume":"25","author":"Krizhevsky","year":"2012","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"ref_6","doi-asserted-by":"crossref","unstructured":"Raina, R., Madhavan, A., and Ng, A.Y. (2009, January 14\u201318). Large-scale deep unsupervised learning using graphics processors. Proceedings of the 26th Annual International Conference on Machine Learning, Montreal, QC, Canada.","DOI":"10.1145\/1553374.1553486"},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"436","DOI":"10.1038\/nature14539","article-title":"Deep learning","volume":"521","author":"LeCun","year":"2015","journal-title":"Nature"},{"key":"ref_8","first-page":"3152676","article-title":"The eu general data protection regulation (gdpr)","volume":"Volume 10","author":"Voigt","year":"2017","journal-title":"A Practical Guide"},{"key":"ref_9","first-page":"44","article-title":"DARPA\u2019s explainable artificial intelligence (XAI) program","volume":"40","author":"Gunning","year":"2019","journal-title":"AI Mag."},{"key":"ref_10","doi-asserted-by":"crossref","first-page":"52138","DOI":"10.1109\/ACCESS.2018.2870052","article-title":"Peeking inside the black-box: A survey on explainable artificial intelligence (XAI)","volume":"6","author":"Adadi","year":"2018","journal-title":"IEEE Access"},{"key":"ref_11","unstructured":"Lundberg, S.M., and Lee, S.I. (2017, January 4\u20139). A unified approach to interpreting model predictions. Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA."},{"key":"ref_12","unstructured":"Simonyan, K., Vedaldi, A., and Zisserman, A. (2013). Deep inside convolutional networks: Visualising image classification models and saliency maps. arXiv."},{"key":"ref_13","doi-asserted-by":"crossref","unstructured":"Bach, S., Binder, A., Montavon, G., Klauschen, F., M\u00fcller, K.R., and Samek, W. (2015). On pixel-wise explanations for nonlinear classifier decisions by layer-wise relevance propagation. PLoS ONE, 10.","DOI":"10.1371\/journal.pone.0130140"},{"key":"ref_14","doi-asserted-by":"crossref","first-page":"211","DOI":"10.1016\/j.patcog.2016.11.008","article-title":"Explaining nonlinear classification decisions with deep taylor decomposition","volume":"65","author":"Montavon","year":"2017","journal-title":"Pattern Recognit."},{"key":"ref_15","unstructured":"Shrikumar, A., Greenside, P., Shcherbina, A., and Kundaje, A. (2016). Not just a black box: Learning important features through propagating activation differences. arXiv."},{"key":"ref_16","unstructured":"Shrikumar, A., Greenside, P., and Kundaje, A. (2017, January 6\u201311). Learning Important Features Through Propagating Activation Differences. Proceedings of the International Conference on Machine Learning, Sydney, Australia."},{"key":"ref_17","unstructured":"Springenberg, J., Dosovitskiy, A., Brox, T., and Riedmiller, M. (2014). Striving for Simplicity: The All Convolutional Net. arXiv."},{"key":"ref_18","unstructured":"Sundararajan, M., Taly, A., and Yan, Q. (2017, January 6\u201311). Axiomatic Attribution for Deep Networks. Proceedings of the International Conference on Machine Learning (ICML), Sydney, Australia."},{"key":"ref_19","doi-asserted-by":"crossref","unstructured":"Zeiler, M.D., and Fergus, R. (2014). Visualizing and understanding convolutional networks. European Conference on Computer Vision, Springer.","DOI":"10.1007\/978-3-319-10590-1_53"},{"key":"ref_20","unstructured":"Zintgraf, L.M., Cohen, T.S., Adel, T., and Welling, M. (2017). Visualizing deep neural network decisions: Prediction difference analysis. arXiv."},{"key":"ref_21","unstructured":"Gudovskiy, D., Hodgkinson, A., Yamaguchi, T., Ishii, Y., and Tsukizawa, S. (2018). Explain to fix: A framework to interpret and correct dnn object detector predictions. arXiv."},{"key":"ref_22","doi-asserted-by":"crossref","unstructured":"Fukui, H., Hirakawa, T., Yamashita, T., and Fujiyoshi, H. (2019, January 15\u201320). Attention branch network: Learning of attention mechanism for visual explanation. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA.","DOI":"10.1109\/CVPR.2019.01096"},{"key":"ref_23","unstructured":"Ancona, M., Ceolini, E., \u00d6ztireli, C., and Gross, M. (May, January 30). Towards better understanding of gradient-based attribution methods for Deep Neural Networks. Proceedings of the 6th International Conference on Learning Representations (ICLR). Arxiv-Computer Science, Vancouver, BC, Canada."},{"key":"ref_24","doi-asserted-by":"crossref","unstructured":"Ribeiro, M.T., Singh, S., and Guestrin, C. (2016, January 13\u201317). \u201cWhy should I trust you?\u201d Explaining the predictions of any classifier. Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA.","DOI":"10.1145\/2939672.2939778"},{"key":"ref_25","unstructured":"Smilkov, D., Thorat, N., Kim, B., Vi\u00e9gas, F., and Wattenberg, M. (2017). Smoothgrad: Removing noise by adding noise. arXiv."},{"key":"ref_26","doi-asserted-by":"crossref","unstructured":"Petsiuk, V., Jain, R., Manjunatha, V., Morariu, V.I., Mehra, A., Ordonez, V., and Saenko, K. (2021, January 20\u201325). Black-box explanation of object detectors via saliency maps. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA.","DOI":"10.1109\/CVPR46437.2021.01128"},{"key":"ref_27","doi-asserted-by":"crossref","first-page":"303","DOI":"10.1007\/s11263-009-0275-4","article-title":"The pascal visual object classes (voc) challenge","volume":"88","author":"Everingham","year":"2010","journal-title":"Int. J. Comput. Vis."},{"key":"ref_28","doi-asserted-by":"crossref","first-page":"261","DOI":"10.1007\/s11263-019-01247-4","article-title":"Deep learning for generic object detection: A survey","volume":"128","author":"Liu","year":"2020","journal-title":"Int. J. Comput. Vis."},{"key":"ref_29","doi-asserted-by":"crossref","unstructured":"Lewis, D.D., and Gale, W.A. (1994). A sequential algorithm for training text classifiers. SIGIR\u201994, Springer.","DOI":"10.1007\/978-1-4471-2099-5_1"},{"key":"ref_30","doi-asserted-by":"crossref","first-page":"113","DOI":"10.1007\/s11263-014-0781-x","article-title":"Multi-class active learning by uncertainty sampling with diversity maximization","volume":"113","author":"Yang","year":"2015","journal-title":"Int. J. Comput. Vis."},{"key":"ref_31","doi-asserted-by":"crossref","first-page":"212209","DOI":"10.1109\/ACCESS.2020.3040290","article-title":"VAID: An Aerial Image Dataset for Vehicle Detection and Classification","volume":"8","author":"Lin","year":"2020","journal-title":"IEEE Access"},{"key":"ref_32","doi-asserted-by":"crossref","unstructured":"Mundhenk, T.N., Konjevod, G., Sakla, W.A., and Boakye, K. (2016). A large contextual dataset for classification, detection and counting of cars with deep learning. European Conference on Computer Vision, Springer.","DOI":"10.1007\/978-3-319-46487-9_48"},{"key":"ref_33","unstructured":"Redmon, J., and Farhadi, A. (2018). Yolov3: An incremental improvement. arXiv."},{"key":"ref_34","doi-asserted-by":"crossref","unstructured":"He, K., Gkioxari, G., Doll\u00e1r, P., and Girshick, R. (2017, January 22\u201329). Mask r-cnn. Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy.","DOI":"10.1109\/ICCV.2017.322"},{"key":"ref_35","doi-asserted-by":"crossref","first-page":"98","DOI":"10.4236\/jcc.2018.611009","article-title":"A vehicle detection method for aerial image based on YOLO","volume":"6","author":"Lu","year":"2018","journal-title":"J. Comput. Commun."},{"key":"ref_36","unstructured":"He, K., Zhang, X., Ren, S., and Sun, J. (July, January 26). Deep residual learning for image recognition. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA."},{"key":"ref_37","doi-asserted-by":"crossref","unstructured":"Lin, T.Y., Maire, M., Belongie, S., Hays, J., Perona, P., Ramanan, D., Doll\u00e1r, P., and Zitnick, C.L. (2014). Microsoft coco: Common objects in context. European Conference on Computer Vision, Springer.","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"ref_38","doi-asserted-by":"crossref","first-page":"125210","DOI":"10.1109\/ACCESS.2020.3005161","article-title":"Small and slim deep convolutional neural network for mobile device","volume":"8","author":"Winoto","year":"2020","journal-title":"IEEE Access"}],"container-title":["Remote Sensing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2072-4292\/14\/9\/1970\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T22:57:01Z","timestamp":1760137021000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2072-4292\/14\/9\/1970"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,4,19]]},"references-count":38,"journal-issue":{"issue":"9","published-online":{"date-parts":[[2022,5]]}},"alternative-id":["rs14091970"],"URL":"https:\/\/doi.org\/10.3390\/rs14091970","relation":{},"ISSN":["2072-4292"],"issn-type":[{"value":"2072-4292","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,4,19]]}}}