{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,2]],"date-time":"2026-04-02T09:41:35Z","timestamp":1775122895422,"version":"3.50.1"},"reference-count":38,"publisher":"MDPI AG","issue":"3","license":[{"start":{"date-parts":[[2024,1,26]],"date-time":"2024-01-26T00:00:00Z","timestamp":1706227200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China (NSFC)","doi-asserted-by":"publisher","award":["41930112"],"award-info":[{"award-number":["41930112"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Remote Sensing"],"abstract":"<jats:p>The airborne and satellite-based synthetic aperture radar enables the acquisition of high-resolution SAR oceanographic images in which even the outlines of ships can be identified. The detection of ship targets from SAR images has a wide range of applications. Due to the density of ships in SAR images, the extreme imbalance between foreground and background clutter, and the diversity of target sizes, achieving lightweight and highly accurate multi-scale ship target detection remains a great challenge. To this end, this paper proposed an attention mechanism for multi-scale receptive fields convolution block (AMMRF). AMMRF not only makes full use of the location information of the feature map to accurately capture the regions in the feature map that are useful for detection results, but also effectively captures the relationship between the feature map channels, so as to better learn the relationship between the ship and the background. Based on this, a new YOLOv7-based ship target detection method, You Only Look Once SAR Ship Identification (YOLO-SARSI), was proposed, which acquires the abstract semantic information extracted from the high-level convolution while retaining the detailed semantic information extracted from the low-level convolution. Compared to the deep learning detection methods proposed by previous authors, our method is more lightweight, only 18.43 M. We examined the effectiveness of our method on two SAR image public datasets: the High-Resolution SAR Images Dataset (HRSID) and the Large-Scale SAR Ship Detection Dataset-v1.0 (LS-SSDD-V1.0). The results show that the average accuracy (AP50) of the detection method YOLO-SARSI proposed in this paper on the HRSID and LS-SSDD-V1.0 datasets is 2.6% and 3.9% higher than that of YOLOv7, respectively.<\/jats:p>","DOI":"10.3390\/rs16030486","type":"journal-article","created":{"date-parts":[[2024,1,26]],"date-time":"2024-01-26T10:54:27Z","timestamp":1706266467000},"page":"486","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":37,"title":["A Lightweight SAR Image Ship Detection Method Based on Improved Convolution and YOLOv7"],"prefix":"10.3390","volume":"16","author":[{"ORCID":"https:\/\/orcid.org\/0009-0003-9134-3008","authenticated-orcid":false,"given":"Hongdou","family":"Tang","sequence":"first","affiliation":[{"name":"The College of Mechanical and Electrical Engineering, Chengdu University of Technology, Chengdu 610059, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1649-073X","authenticated-orcid":false,"given":"Song","family":"Gao","sequence":"additional","affiliation":[{"name":"The College of Mechanical and Electrical Engineering, Chengdu University of Technology, Chengdu 610059, China"}]},{"given":"Song","family":"Li","sequence":"additional","affiliation":[{"name":"The College of Mechanical and Electrical Engineering, Chengdu University of Technology, Chengdu 610059, China"}]},{"given":"Pengyu","family":"Wang","sequence":"additional","affiliation":[{"name":"The College of Mechanical and Electrical Engineering, Chengdu University of Technology, Chengdu 610059, China"}]},{"given":"Jiqiu","family":"Liu","sequence":"additional","affiliation":[{"name":"The College of Mechanical and Electrical Engineering, Chengdu University of Technology, Chengdu 610059, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1333-0354","authenticated-orcid":false,"given":"Simin","family":"Wang","sequence":"additional","affiliation":[{"name":"The College of Mechanical and Electrical Engineering, Chengdu University of Technology, Chengdu 610059, China"}]},{"given":"Jiang","family":"Qian","sequence":"additional","affiliation":[{"name":"The School of Resources and Environment, University of Electronic Science and Technology of China, Chengdu 611731, China"}]}],"member":"1968","published-online":{"date-parts":[[2024,1,26]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"1925","DOI":"10.1109\/LGRS.2016.2618604","article-title":"A Modified CFAR Algorithm Based on Object Proposals for Ship Target Detection in SAR Images","volume":"13","author":"Dai","year":"2016","journal-title":"IEEE Geosci. Remote Sens. Lett."},{"key":"ref_2","doi-asserted-by":"crossref","first-page":"083501","DOI":"10.1117\/1.JRS.8.083501","article-title":"Multiscale salient region detection and salient map generation for synthetic aperture radar image","volume":"8","author":"Tang","year":"2014","journal-title":"J. Appl. Remote Sens."},{"key":"ref_3","doi-asserted-by":"crossref","first-page":"391","DOI":"10.1109\/LGRS.2009.2037341","article-title":"On a Novel Approach Using MLCC and CFAR for the Improvement of Ship Detection by Synthetic Aperture Radar","volume":"7","author":"Hwang","year":"2010","journal-title":"IEEE Geosci. Remote Sens. Lett."},{"key":"ref_4","first-page":"84","article-title":"ImageNet Classification with Deep Convolutional Neural Networks","volume":"60","author":"Krizhevsky","year":"2017","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"ref_5","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., and Sun, J. (2016, January 27\u201330). Deep Residual Learning for Image Recognition. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA.","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref_6","doi-asserted-by":"crossref","first-page":"20881","DOI":"10.1109\/ACCESS.2018.2825376","article-title":"A Densely Connected End-to-End Neural Network for Multiscale and Multiscene SAR Ship Detection","volume":"6","author":"Jiao","year":"2018","journal-title":"IEEE Access"},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"8983","DOI":"10.1109\/TGRS.2019.2923988","article-title":"Dense Attention Pyramid Networks for Multi-Scale Ship Detection in SAR Images","volume":"57","author":"Cui","year":"2019","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_8","doi-asserted-by":"crossref","unstructured":"Zhang, T., and Zhang, X. (2019). High-Speed Ship Detection in SAR Images Based on a Grid Convolutional Neural Network. Remote Sens., 11.","DOI":"10.3390\/rs11101206"},{"key":"ref_9","doi-asserted-by":"crossref","first-page":"666","DOI":"10.1109\/JSTARS.2021.3137390","article-title":"Ships Detection in SAR Images Based on Anchor-Free Model with Mask Guidance Features","volume":"15","author":"Qu","year":"2022","journal-title":"IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens."},{"key":"ref_10","doi-asserted-by":"crossref","first-page":"4162","DOI":"10.1007\/s10489-022-03683-1","article-title":"Small-Target Ship Detection in SAR Images Based on Densely Connected Deep Neural Network with Attention in Complex Scenes","volume":"53","author":"Sun","year":"2022","journal-title":"Appl. Intell."},{"key":"ref_11","first-page":"1","article-title":"A High-Effective Implementation of Ship Detector for SAR Images","volume":"19","author":"Gao","year":"2022","journal-title":"IEEE Geosci. Remote Sens. Lett."},{"key":"ref_12","doi-asserted-by":"crossref","unstructured":"Wang, S., Gao, S., Zhou, L., Liu, R., Zhang, H., Liu, J., Jia, Y., and Qian, J. (2022). YOLO-SD: Small Ship Detection in SAR Images by Multi-Scale Convolution and Feature Transformer Module. Remote Sens., 14.","DOI":"10.3390\/rs14205268"},{"key":"ref_13","doi-asserted-by":"crossref","first-page":"6623","DOI":"10.1109\/TGRS.2020.2978268","article-title":"A Patch-to-Pixel Convolutional Neural Network for Small Ship Detection with PolSAR Images","volume":"58","author":"Jin","year":"2020","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_14","doi-asserted-by":"crossref","unstructured":"Chen, P., Li, Y., Zhou, H., Liu, B., and Liu, P. (2020). Detection of Small Ship Objects Using Anchor Boxes Cluster and Feature Pyramid Network Model for SAR Imagery. J. Mar. Sci. Eng., 8.","DOI":"10.3390\/jmse8020112"},{"key":"ref_15","doi-asserted-by":"crossref","first-page":"1331","DOI":"10.1109\/TGRS.2020.3005151","article-title":"An Anchor-Free Method Based on Feature Balancing and Refinement Network for Multiscale Ship Detection in SAR Images","volume":"59","author":"Fu","year":"2021","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_16","doi-asserted-by":"crossref","first-page":"104812","DOI":"10.1016\/j.ssci.2020.104812","article-title":"Deep learning for autonomous ship-oriented small ship detection","volume":"130","author":"Chen","year":"2020","journal-title":"Saf. Sci."},{"key":"ref_17","doi-asserted-by":"crossref","first-page":"107787","DOI":"10.1016\/j.patcog.2020.107787","article-title":"A CenterNet++ model for ship detection in SAR images","volume":"112","author":"Guo","year":"2021","journal-title":"Pattern Recognit."},{"key":"ref_18","doi-asserted-by":"crossref","first-page":"42301","DOI":"10.1007\/s11432-017-9405-6","article-title":"A coupled convolutional neural network for small and densely clustered ship detection in SAR images","volume":"62","author":"Zhao","year":"2018","journal-title":"Sci. China Inf. Sci."},{"key":"ref_19","doi-asserted-by":"crossref","first-page":"6434","DOI":"10.1109\/TIP.2021.3093795","article-title":"SlimConv: Reducing Channel Redundancy in Convolutional Neural Networks by Weights Flipping","volume":"30","author":"Qiu","year":"2021","journal-title":"IEEE Trans. Image Process."},{"key":"ref_20","doi-asserted-by":"crossref","unstructured":"Wang, C.Y., Bochkovskiy, A., and Liao, H.Y.M. (2022). YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. arXiv.","DOI":"10.1109\/CVPR52729.2023.00721"},{"key":"ref_21","doi-asserted-by":"crossref","unstructured":"Girshick, R. (2015, January 7\u201313). Fast R-CNN. Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile.","DOI":"10.1109\/ICCV.2015.169"},{"key":"ref_22","doi-asserted-by":"crossref","unstructured":"Lin, T.Y., Maire, M., Belongie, S., Bourdev, L., Girshick, R., Hays, J., Perona, P., Ramanan, D., Zitnick, C.L., and Doll\u00e1r, P. (2015). Microsoft COCO: Common Objects in Context. arXiv.","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"ref_23","doi-asserted-by":"crossref","first-page":"303","DOI":"10.1007\/s11263-009-0275-4","article-title":"The Pascal Visual Object Classes (VOC) Challenge","volume":"88","author":"Everingham","year":"2010","journal-title":"Int. J. Comput. Vis."},{"key":"ref_24","doi-asserted-by":"crossref","unstructured":"Huang, M., Xu, Y., Qian, L., Shi, W., Zhang, Y., Bao, W., Wang, N., Liu, X., and Xiang, X. (2021). The QXS-SAROPT Dataset for Deep Learning in SAR-Optical Data Fusion. arXiv.","DOI":"10.34133\/2021\/9841456"},{"key":"ref_25","doi-asserted-by":"crossref","first-page":"120234","DOI":"10.1109\/ACCESS.2020.3005861","article-title":"HRSID: A High-Resolution SAR Images Dataset for Ship Detection and Instance Segmentation","volume":"8","author":"Wei","year":"2020","journal-title":"IEEE Access"},{"key":"ref_26","doi-asserted-by":"crossref","unstructured":"Zhang, T., Zhang, X., Ke, X., Zhan, X., Shi, J., Wei, S., Pan, D., Li, J., Su, H., and Zhou, Y. (2020). LS-SSDD-v1.0: A Deep Learning Dataset Dedicated to Small Ship Detection from Large-Scale Sentinel-1 SAR Images. Remote Sens., 12.","DOI":"10.3390\/rs12182997"},{"key":"ref_27","doi-asserted-by":"crossref","unstructured":"Hou, Q., Zhou, D., and Feng, J. (2021). Coordinate Attention for Efficient Mobile Network Design. arXiv.","DOI":"10.1109\/CVPR46437.2021.01350"},{"key":"ref_28","doi-asserted-by":"crossref","unstructured":"Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., Erhan, D., Vanhoucke, V., and Rabinovich, A. (2014). Going Deeper with Convolutions. arXiv.","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"ref_29","doi-asserted-by":"crossref","unstructured":"Huang, G., Liu, Z., Van Der Maaten, L., and Weinberger, K.Q. (2017, January 21\u201326). Densely Connected Convolutional Networks. Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA.","DOI":"10.1109\/CVPR.2017.243"},{"key":"ref_30","doi-asserted-by":"crossref","unstructured":"Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C.Y., and Berg, A.C. (2016, January 11\u201314). SSD: Single Shot MultiBox Detector. Proceedings of the Computer Vision\u2014ECCV 2016: 14th European Conference, Amsterdam, The Netherlands.","DOI":"10.1007\/978-3-319-46448-0_2"},{"key":"ref_31","doi-asserted-by":"crossref","unstructured":"Cai, Z., and Vasconcelos, N. (2018, January 18\u201323). Cascade R-CNN: Delving Into High Quality Object Detection. Proceedings of the 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA.","DOI":"10.1109\/CVPR.2018.00644"},{"key":"ref_32","doi-asserted-by":"crossref","unstructured":"Ren, S., He, K., Girshick, R., and Sun, J. (2016). Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. arXiv.","DOI":"10.1109\/TPAMI.2016.2577031"},{"key":"ref_33","doi-asserted-by":"crossref","unstructured":"He, K., Gkioxari, G., Doll\u00e1r, P., and Girshick, R. (2018). Mask R-CNN. arXiv.","DOI":"10.1109\/ICCV.2017.322"},{"key":"ref_34","doi-asserted-by":"crossref","unstructured":"Lin, T.Y., Goyal, P., Girshick, R., He, K., and Doll\u00e1r, P. (2018). Focal Loss for Dense Object Detection. arXiv.","DOI":"10.1109\/ICCV.2017.324"},{"key":"ref_35","doi-asserted-by":"crossref","unstructured":"Liu, Z., Lin, Y., Cao, Y., Hu, H., Wei, Y., Zhang, Z., Lin, S., and Guo, B. (2021). Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. arXiv.","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"ref_36","unstructured":"Chen, K., Wang, J., Pang, J., Cao, Y., Xiong, Y., Li, X., Sun, S., Feng, W., Liu, Z., and Xu, J. (2019). MMDetection: Open MMLab Detection Toolbox and Benchmark. arXiv."},{"key":"ref_37","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., and Gelly, S. (2021). An Image Is Worth 16 \u00d7 16 Words: Transformers for Image Recognition at Scale. arXiv."},{"key":"ref_38","doi-asserted-by":"crossref","first-page":"336","DOI":"10.1007\/s11263-019-01228-7","article-title":"Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization","volume":"128","author":"Selvaraju","year":"2019","journal-title":"Int. J. Comput. Vis."}],"container-title":["Remote Sensing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2072-4292\/16\/3\/486\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T13:50:02Z","timestamp":1760104202000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2072-4292\/16\/3\/486"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,1,26]]},"references-count":38,"journal-issue":{"issue":"3","published-online":{"date-parts":[[2024,2]]}},"alternative-id":["rs16030486"],"URL":"https:\/\/doi.org\/10.3390\/rs16030486","relation":{},"ISSN":["2072-4292"],"issn-type":[{"value":"2072-4292","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,1,26]]}}}