{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T01:52:58Z","timestamp":1769910778749,"version":"3.49.0"},"reference-count":46,"publisher":"MDPI AG","issue":"24","license":[{"start":{"date-parts":[[2023,12,18]],"date-time":"2023-12-18T00:00:00Z","timestamp":1702857600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Remote Sensing"],"abstract":"<jats:p>The advent of deep learning has significantly propelled the utilization of neural networks for Synthetic Aperture Radar (SAR) ship detection in recent years. However, there are two main obstacles in SAR detection. Challenge 1: The multiscale nature of SAR ships. Challenge 2: The influence of intricate near-shore environments and the interference of clutter noise in offshore areas, especially affecting small-ship detection. Existing neural network-based approaches attempt to tackle these challenges, yet they often fall short in effectively addressing small-ship detection across multiple scales and complex backgrounds simultaneously. To overcome these challenges, we propose a novel network called SwinT-FRM-ShipNet. Our method introduces an integrated feature extractor, Swin-T-YOLOv5l, which combines Swin Transformer and YOLOv5l. The extractor is designed to highlight the differences between the complex background and the target by encoding both local and global information. Additionally, a feature pyramid IEFR-FPN, consisting of the Information Enhancement Module (IEM) and the Feature Refinement Module (FRM), is proposed to enrich the flow of spatial contextual information, fuse multiresolution features, and refine representations of small and multiscale ships. Furthermore, we introduce recursive gated convolutional prediction heads (GCPH) to explore the potential of high-order spatial interactions and add a larger-sized prediction head to focus on small ships. Experimental results demonstrate the superior performance of our method compared to mainstream approaches on the SSDD and SAR-Ship-Dataset. Our method achieves an F1 score, mAP0.5, and mAP0.5:0.95 of 96.5% (+0.9), 98.2% (+1.0%), and 75.4% (+3.3%), respectively, surpassing the most competitive algorithms.<\/jats:p>","DOI":"10.3390\/rs15245780","type":"journal-article","created":{"date-parts":[[2023,12,18]],"date-time":"2023-12-18T10:04:47Z","timestamp":1702893887000},"page":"5780","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":17,"title":["A New Deep Neural Network Based on SwinT-FRM-ShipNet for SAR Ship Detection in Complex Near-Shore and Offshore Environments"],"prefix":"10.3390","volume":"15","author":[{"ORCID":"https:\/\/orcid.org\/0009-0002-8687-6793","authenticated-orcid":false,"given":"Zhuhao","family":"Lu","sequence":"first","affiliation":[{"name":"The Shanghai Key Laboratory of Multidimensional Information Processing, East China Normal University, Shanghai 200241, China"}]},{"given":"Pengfei","family":"Wang","sequence":"additional","affiliation":[{"name":"The Shanghai Key Laboratory of Multidimensional Information Processing, East China Normal University, Shanghai 200241, China"}]},{"given":"Yajun","family":"Li","sequence":"additional","affiliation":[{"name":"The Shanghai Key Laboratory of Multidimensional Information Processing, East China Normal University, Shanghai 200241, China"}]},{"given":"Baogang","family":"Ding","sequence":"additional","affiliation":[{"name":"The Shanghai Key Laboratory of Multidimensional Information Processing, East China Normal University, Shanghai 200241, China"}]}],"member":"1968","published-online":{"date-parts":[[2023,12,18]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"759","DOI":"10.1109\/JPROC.2012.2220511","article-title":"Very-High-Resolution Airborne Synthetic Aperture Radar Imaging: Signal Processing and Applications","volume":"101","author":"Reigber","year":"2013","journal-title":"Proc. IEEE"},{"key":"ref_2","doi-asserted-by":"crossref","first-page":"6","DOI":"10.1109\/MGRS.2013.2248301","article-title":"A tutorial on synthetic aperture radar","volume":"1","author":"Moreira","year":"2013","journal-title":"IEEE Geosci. Remote Sens. Mag."},{"key":"ref_3","doi-asserted-by":"crossref","first-page":"4806","DOI":"10.1109\/TGRS.2016.2551720","article-title":"Target Classification Using the Deep Convolutional Networks for SAR Images","volume":"54","author":"Chen","year":"2016","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_4","doi-asserted-by":"crossref","first-page":"568","DOI":"10.1080\/07038992.2001.10854896","article-title":"Automatic Detection of Ships in RADARSAT-1 SAR Imagery","volume":"27","author":"Wackerman","year":"2001","journal-title":"Can. J. Remote Sens."},{"key":"ref_5","doi-asserted-by":"crossref","first-page":"1092","DOI":"10.1109\/TGRS.2010.2071879","article-title":"Ship surveillance with TerraSAR-X","volume":"49","author":"Brusch","year":"2010","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_6","doi-asserted-by":"crossref","first-page":"2797","DOI":"10.1109\/JSTARS.2014.2321708","article-title":"Spaceborne radar imaging of maritime moving targets with the Cosmo-SkyMed SAR system","volume":"7","author":"Martorella","year":"2014","journal-title":"IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens."},{"key":"ref_7","doi-asserted-by":"crossref","unstructured":"Crisp, D.J. (2013, January 9\u201312). A Ship Detection System for RADARSAT-2 Dual-Pol Multi-Look Imagery Implemented in the ADSS. Proceedings of the 2013 International Conference on Radar, Adelaide, Australia.","DOI":"10.1109\/RADAR.2013.6652006"},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"2824","DOI":"10.1109\/TGRS.2018.2877821","article-title":"Ship Detection from PolSAR Imagery Using the Complete Polarimetric Covariance Difference Matrix","volume":"57","author":"Zhang","year":"2018","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_9","doi-asserted-by":"crossref","first-page":"41","DOI":"10.1016\/j.isprsjprs.2019.08.009","article-title":"Ship detection from PolSAR imagery using the ambiguity removal polarimetric notch filter","volume":"157","author":"Zhang","year":"2019","journal-title":"ISPRS J. Photogramm. Remote Sens."},{"key":"ref_10","doi-asserted-by":"crossref","unstructured":"Yang, B., and Zhang, H. (2022). A CFAR Algorithm Based on Monte Carlo Method for Millimeter-Wave Radar Road Traffic Target Detection. Remote Sens., 14.","DOI":"10.3390\/rs14081779"},{"key":"ref_11","doi-asserted-by":"crossref","first-page":"4811","DOI":"10.1109\/TGRS.2017.2701813","article-title":"CFAR Ship Detection in Nonhomogeneous Sea Clutter Using Polarimetric SAR Data Based on the Notch Filter","volume":"55","author":"Gao","year":"2017","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_12","doi-asserted-by":"crossref","first-page":"648","DOI":"10.1109\/36.581981","article-title":"A model for extremely heterogeneous clutter","volume":"35","author":"Frery","year":"1997","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_13","doi-asserted-by":"crossref","first-page":"529","DOI":"10.1109\/LGRS.2017.2654450","article-title":"An Intensity-Space Domain CFAR Method for Ship Detection in HR SAR Images","volume":"14","author":"Wang","year":"2017","journal-title":"IEEE Geosci. Remote Sens. Lett."},{"key":"ref_14","first-page":"1531","article-title":"A CFAR algorithm for non-Gaussian clutter based on mixture of K distributions","volume":"15","author":"Ye","year":"2018","journal-title":"IEEE Geosci. Remote Sens. Lett."},{"key":"ref_15","unstructured":"Drenkow, N., Sani, N., Shpitser, I., and Unberath, M. (2021). A systematic review of robustness in deep learning for computer vision: Mind the gap?. arXiv."},{"key":"ref_16","doi-asserted-by":"crossref","first-page":"966","DOI":"10.3390\/make3040048","article-title":"Analysis of Explainers of Black Box Deep Neural Networks for Computer Vision: A Survey","volume":"3","author":"Buhrmester","year":"2021","journal-title":"Mach. Learn. Knowl. Extr."},{"key":"ref_17","doi-asserted-by":"crossref","first-page":"5078","DOI":"10.1109\/TSP.2017.2712124","article-title":"On Multiple Covariance Equality Testing with Application to SAR Change Detection","volume":"65","author":"Ciuonzo","year":"2017","journal-title":"IEEE Trans. Signal Process."},{"key":"ref_18","doi-asserted-by":"crossref","first-page":"84","DOI":"10.1145\/3065386","article-title":"Imagenet classification with deep convolutional neural networks","volume":"60","author":"Krizhevsky","year":"2017","journal-title":"Commun. ACM"},{"key":"ref_19","doi-asserted-by":"crossref","first-page":"6999","DOI":"10.1109\/TNNLS.2021.3084827","article-title":"A Survey of Convolutional Neural Networks: Analysis, Applications, and Prospects","volume":"33","author":"Li","year":"2022","journal-title":"IEEE Trans. Neural Networks Learn. Syst."},{"key":"ref_20","doi-asserted-by":"crossref","first-page":"1904","DOI":"10.1109\/TPAMI.2015.2389824","article-title":"Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition","volume":"37","author":"He","year":"2015","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_21","doi-asserted-by":"crossref","unstructured":"Girshick, R. (2015, January 7\u201313). Fast r-cnn. Proceedings of the IEEE International Conference on Computer Vision 2015, Washington, DC, USA.","DOI":"10.1109\/ICCV.2015.169"},{"key":"ref_22","unstructured":"Ren, S., He, K., Girshick, R., and Sun, J. (2015, January 7\u201312). Faster R-CNN: Towards real-time object detection with region proposal networks. Proceedings of the 28th International Conference on Neural Information Processing Systems, Montreal, Canada."},{"key":"ref_23","doi-asserted-by":"crossref","unstructured":"Cai, Z., and Vasconcelos, N. (2018, January 18\u201322). Cascade R-CNN: Delving into High Quality Object Detection. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2018, Salt Lake City, UT, USA.","DOI":"10.1109\/CVPR.2018.00644"},{"key":"ref_24","doi-asserted-by":"crossref","unstructured":"Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C.Y., and Berg, A.C. (2016, January 11\u201314). SSSD: Single Shot Multibox Detector. Proceedings of the Computer Vision\u2013ECCV 2016, 14th European Conference, Amsterdam, The Netherlands. Proceedings, Part I 14.","DOI":"10.1007\/978-3-319-46448-0_2"},{"key":"ref_25","unstructured":"Bochkovskiy, A., Wang, C.Y., and Liao, H.Y.M. (2020). Yolov4: Optimal speed and accuracy of object detection. arXiv."},{"key":"ref_26","doi-asserted-by":"crossref","unstructured":"Wang, C.Y., Bochkovskiy, A., and Liao, H.Y.M. (2022). YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. arXiv.","DOI":"10.1109\/CVPR52729.2023.00721"},{"key":"ref_27","unstructured":"Chen, J., Lu, Y., Yu, Q., Luo, X., Adeli, E., Wang, Y., Lu, L., Yuille, A.L., and Zhou, Y. (2021). Transunet: Transformers make strong encoders for medical image segmentation. arXiv."},{"key":"ref_28","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., and Gelly, S. (2020). An image is worth 16 \u00d7 16 words: Transformers for image recognition at scale. arXiv."},{"key":"ref_29","unstructured":"Zhu, X., Su, W., Lu, L., Li, B., Wang, X., and Dai, J. (2020). Deformable DETR: Deformable transformers for end-to-end object detection. arXiv."},{"key":"ref_30","doi-asserted-by":"crossref","unstructured":"Liu, Z., Lin, Y., Cao, Y., Hu, H., Wei, Y., Zhang, Z., Lin, S., and Guo, B. (2021, January 11\u201317). Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. Proceedings of the IEEE\/CVF International Conference on Computer Vision, Montreal, BC, Canada.","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"ref_31","doi-asserted-by":"crossref","unstructured":"Cao, H., Wang, Y., Chen, J., Jiang, D., Zhang, X., Tian, Q., and Wang, M. (2022, January 24\u201328). Swin-Unet: Unet-Like Pure Transformer for Medical Image Segmentation. Proceedings of the European Conference on Computer Vision 2022, Tel Aviv, Israel.","DOI":"10.1007\/978-3-031-25066-8_9"},{"key":"ref_32","doi-asserted-by":"crossref","first-page":"1107","DOI":"10.1109\/36.193786","article-title":"SAR calibration: An overview","volume":"30","author":"Freeman","year":"1992","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_33","doi-asserted-by":"crossref","first-page":"5772","DOI":"10.1109\/TGRS.2020.2969979","article-title":"A Rotational Libra R-CNN Method for Ship Detection","volume":"58","author":"Guo","year":"2020","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_34","doi-asserted-by":"crossref","first-page":"107787","DOI":"10.1016\/j.patcog.2020.107787","article-title":"A CenterNet++ model for ship detection in SAR images","volume":"112","author":"Guo","year":"2021","journal-title":"Pattern Recognit."},{"key":"ref_35","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1109\/TGRS.2020.3034752","article-title":"Remote sensing image change detection with transformers","volume":"60","author":"Chen","year":"2021","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_36","doi-asserted-by":"crossref","first-page":"379","DOI":"10.1109\/TGRS.2020.2997200","article-title":"Ship detection in large-scale SAR images via spatial shuffle-group enhance attention","volume":"59","author":"Cui","year":"2020","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_37","doi-asserted-by":"crossref","first-page":"8983","DOI":"10.1109\/TGRS.2019.2923988","article-title":"Dense Attention Pyramid Networks for Multi-Scale Ship Detection in SAR Images","volume":"57","author":"Cui","year":"2019","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_38","first-page":"5203213","article-title":"A Novel Multidimensional Domain Deep Learning Network for SAR Ship Detection","volume":"60","author":"Li","year":"2022","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_39","doi-asserted-by":"crossref","unstructured":"Dai, W., Mao, Y., Yuan, R., Liu, Y., Pu, X., and Li, C. (2020). A Novel Detector Based on Convolution Neural Networks for Multiscale SAR Ship Detection in Complex Background. Sensors, 20.","DOI":"10.3390\/s20092547"},{"key":"ref_40","first-page":"10353","article-title":"Hornet: Efficient high-order spatial interactions with recursive gated convolutions","volume":"35","author":"Rao","year":"2022","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"ref_41","doi-asserted-by":"crossref","unstructured":"Li, J., Qu, C., and Shao, J. (2017, January 13\u201314). Ship detection in SAR images based on an improved faster R-CNN. Proceedings of the 2017 SAR in Big Data Era: Models, Methods and Applications (BIGSARDATA), Beijing, China.","DOI":"10.1109\/BIGSARDATA.2017.8124934"},{"key":"ref_42","doi-asserted-by":"crossref","unstructured":"Wang, Y., Wang, C., Zhang, H., Dong, Y., and Wei, S. (2019). A SAR Dataset of Ship Detection for Deep Learning under Complex Backgrounds. Remote Sens., 11.","DOI":"10.3390\/rs11070765"},{"key":"ref_43","doi-asserted-by":"crossref","unstructured":"Zhu, X., Lyu, S., Wang, X., and Zhao, Q. (2021, January 11\u201317). TPH-YOLOv5: Improved YOLOv5 Based on Transformer Prediction Head for Object Detection on Drone-captured Scenarios. Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV) Workshops, Montreal, BC, Canada.","DOI":"10.1109\/ICCVW54120.2021.00312"},{"key":"ref_44","unstructured":"Reis, D., Kupec, J., Hong, J., and Daoudi, A. (2023). Real-Time Flying Object Detection with YOLOv8. arXiv."},{"key":"ref_45","doi-asserted-by":"crossref","unstructured":"Woo, S., Park, J., Lee, J.Y., and Kweon, I.S. (2018, January 8\u201314). Cbam: Convolutional Block Attention Module. Proceedings of the European conference on computer vision (ECCV) 2018, Munich, Germany.","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"ref_46","unstructured":"Keskar, N.S., and Socher, R. (2017). Improving generalization performance by switching from adam to sgd. arXiv."}],"container-title":["Remote Sensing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2072-4292\/15\/24\/5780\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T21:40:55Z","timestamp":1760132455000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2072-4292\/15\/24\/5780"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,12,18]]},"references-count":46,"journal-issue":{"issue":"24","published-online":{"date-parts":[[2023,12]]}},"alternative-id":["rs15245780"],"URL":"https:\/\/doi.org\/10.3390\/rs15245780","relation":{},"ISSN":["2072-4292"],"issn-type":[{"value":"2072-4292","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,12,18]]}}}