{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,12]],"date-time":"2025-10-12T02:35:08Z","timestamp":1760236508178,"version":"build-2065373602"},"reference-count":34,"publisher":"MDPI AG","issue":"23","license":[{"start":{"date-parts":[[2021,11,29]],"date-time":"2021-11-29T00:00:00Z","timestamp":1638144000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61975043"],"award-info":[{"award-number":["61975043"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Remote Sensing"],"abstract":"<jats:p>Due to the complexity of scene interference and the variability of ship scale and position, automatic ship detection in remote sensing images makes for challenging research. The existing deep networks rarely design receptive fields that fit the target scale based on training data. Moreover, most of them ignore the effective retention of position information in the feature extraction process, which reduces the contribution of features to subsequent classification. To overcome these limitations, we propose a novel ship detection framework combining the dilated rate selection and attention-guided feature representation strategies, which can efficiently detect ships of different scales under the interference of complex environments such as clouds, sea clutter and mist. Specifically, we present a dilated convolution parameter search strategy to adaptively select the dilated rate for the multi-branch extraction architecture, adaptively obtaining context information of different receptive fields without sacrificing the image resolution. Moreover, to enhance the spatial position information of the feature maps, we calculate the correlation of spatial points from the vertical and horizontal directions and embed it into the channel compression coding process, thus generating the multi-dimensional feature descriptors which are sensitive to direction and position characteristics of ships. Experimental results on the Airbus dataset demonstrate that the proposed method achieves state-of-the-art performance compared with other detection models.<\/jats:p>","DOI":"10.3390\/rs13234840","type":"journal-article","created":{"date-parts":[[2021,12,1]],"date-time":"2021-12-01T01:45:02Z","timestamp":1638323102000},"page":"4840","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":3,"title":["Ship Detection via Dilated Rate Search and Attention-Guided Feature Representation"],"prefix":"10.3390","volume":"13","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-4418-605X","authenticated-orcid":false,"given":"Jianming","family":"Hu","sequence":"first","affiliation":[{"name":"Research Center for Space Optical Engineering, Harbin Institute of Technology, Harbin 150001, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5504-8480","authenticated-orcid":false,"given":"Xiyang","family":"Zhi","sequence":"additional","affiliation":[{"name":"Research Center for Space Optical Engineering, Harbin Institute of Technology, Harbin 150001, China"}]},{"given":"Tianjun","family":"Shi","sequence":"additional","affiliation":[{"name":"Research Center for Space Optical Engineering, Harbin Institute of Technology, Harbin 150001, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7202-2127","authenticated-orcid":false,"given":"Lijian","family":"Yu","sequence":"additional","affiliation":[{"name":"Research Center for Space Optical Engineering, Harbin Institute of Technology, Harbin 150001, China"}]},{"given":"Wei","family":"Zhang","sequence":"additional","affiliation":[{"name":"Research Center for Space Optical Engineering, Harbin Institute of Technology, Harbin 150001, China"}]}],"member":"1968","published-online":{"date-parts":[[2021,11,29]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1016\/j.rse.2017.12.033","article-title":"Vessel detection and classification from spaceborne optical images: A literature survey","volume":"207","author":"Kanjir","year":"2018","journal-title":"Remote Sens. Environ."},{"key":"ref_2","doi-asserted-by":"crossref","first-page":"46","DOI":"10.1016\/j.neucom.2018.03.049","article-title":"Saliency detection based on foreground appearance and background-prior","volume":"301","author":"Li","year":"2018","journal-title":"Neurocomputing"},{"key":"ref_3","doi-asserted-by":"crossref","unstructured":"Hu, J., Zhi, X., Zhang, W., Ren, L., and Bruzzone, L. (2020). Salient Ship Detection via Background Prior and Foreground Constraint in Remote Sensing Images. Remote Sens., 12.","DOI":"10.3390\/rs12203370"},{"key":"ref_4","doi-asserted-by":"crossref","first-page":"1665","DOI":"10.1109\/LGRS.2017.2727515","article-title":"Fully convolutional network with task partitioning for inshore ship detection in optical remote sensing images","volume":"14","author":"Lin","year":"2017","journal-title":"IEEE Geosci. Remote Sens. Lett."},{"key":"ref_5","doi-asserted-by":"crossref","unstructured":"Hu, J., Zhi, X., Shi, T., Zhang, W., Cui, Y., and Zhao, S. (2021). PAG-YOLO: A Portable Attention-Guided YOLO Network for Small Ship Detection. Remote Sens., 13.","DOI":"10.3390\/rs13163059"},{"key":"ref_6","doi-asserted-by":"crossref","first-page":"751","DOI":"10.1109\/LGRS.2018.2882551","article-title":"Squeeze and excitation rank faster R-CNN for ship detection in SAR images","volume":"16","author":"Lin","year":"2018","journal-title":"IEEE Geosci. Remote Sens. Lett."},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"5485","DOI":"10.1109\/TII.2020.3033170","article-title":"HDCB-Net: A Neural Network with the Hybrid Dilated Convolution for Pixel-Level Crack Detection on Concrete Bridges","volume":"17","author":"Jiang","year":"2020","journal-title":"IEEE Trans. Ind. Inform."},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"520","DOI":"10.1016\/j.tics.2007.09.009","article-title":"The role of context in object recognition","volume":"11","author":"Oliva","year":"2007","journal-title":"Trends Cogn. Sci."},{"key":"ref_9","doi-asserted-by":"crossref","unstructured":"Jeon, M., and Jeong, Y.S. (2020). Compact and accurate scene text detector. Appl. Sci., 10.","DOI":"10.3390\/app10062096"},{"key":"ref_10","doi-asserted-by":"crossref","unstructured":"Vu, T., Van Nguyen, C., Pham, T.X., Luu, T.M., and Yoo, C.D. (2018, January 8\u201314). Fast and efficient image quality enhancement via desubpixel convolutional neural networks. Proceedings of the European Conference on Computer Vision (ECCV) Workshops, Munich, Germany.","DOI":"10.1007\/978-3-030-11021-5_16"},{"key":"ref_11","doi-asserted-by":"crossref","first-page":"835","DOI":"10.1016\/j.ins.2020.09.003","article-title":"CNN-based encoder-decoder networks for salient object detection: A comprehensive review and recent advances","volume":"546","author":"Ji","year":"2021","journal-title":"Inf. Sci."},{"key":"ref_12","doi-asserted-by":"crossref","unstructured":"Zhang, S., Wu, R., Xu, K., Wang, J., and Sun, W. (2019). R-CNN-based ship detection from high resolution remote sensing imagery. Remote Sens., 11.","DOI":"10.3390\/rs11060631"},{"key":"ref_13","doi-asserted-by":"crossref","unstructured":"Liu, S., and Huang, D. (2018, January 8\u201314). Receptive field block net for accurate and fast object detection. Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany.","DOI":"10.1007\/978-3-030-01252-6_24"},{"key":"ref_14","doi-asserted-by":"crossref","first-page":"107098","DOI":"10.1016\/j.patcog.2019.107098","article-title":"Multi-model ensemble with rich spatial information for object detection","volume":"99","author":"Xu","year":"2020","journal-title":"Pattern Recognit."},{"key":"ref_15","doi-asserted-by":"crossref","first-page":"82832","DOI":"10.1109\/ACCESS.2020.2991439","article-title":"Dilated convolution and feature fusion SSD network for small object detection in remote sensing images","volume":"8","author":"Qu","year":"2020","journal-title":"IEEE Access"},{"key":"ref_16","doi-asserted-by":"crossref","unstructured":"Li, Z., Peng, C., Yu, G., Zhang, X., Deng, Y., and Sun, J. (2018, January 8\u201314). Detnet: Design backbone for object detection. Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany.","DOI":"10.1007\/978-3-030-01240-3_21"},{"key":"ref_17","doi-asserted-by":"crossref","first-page":"1392","DOI":"10.1109\/TMI.2019.2950051","article-title":"Dense dilated network with probability regularized walk for vessel detection","volume":"39","author":"Mou","year":"2019","journal-title":"IEEE Trans. Med. Imaging"},{"key":"ref_18","doi-asserted-by":"crossref","first-page":"1254","DOI":"10.1109\/34.730558","article-title":"A model of saliency-based visual attention for rapid scene analysis","volume":"20","author":"Itti","year":"1998","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_19","unstructured":"Mnih, V., Heess, N., and Graves, A. (2014, January 8\u201313). Recurrent models of visual attention. Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada."},{"key":"ref_20","doi-asserted-by":"crossref","unstructured":"Fu, J., Zheng, H., and Mei, T. (2017, January 21\u201326). Look closer to see better: Recurrent attention convolutional neural network for fine-grained image recognition. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA.","DOI":"10.1109\/CVPR.2017.476"},{"key":"ref_21","doi-asserted-by":"crossref","unstructured":"Hu, J., Shen, L., and Sun, G. (2018, January 18\u201323). Squeeze-and-excitation networks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA.","DOI":"10.1109\/CVPR.2018.00745"},{"key":"ref_22","doi-asserted-by":"crossref","unstructured":"Woo, S., Park, J., Lee, J.Y., and Kweon, I.S. (2018, January 8\u201314). Cbam: Convolutional block attention module. Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany.","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"ref_23","doi-asserted-by":"crossref","first-page":"4460","DOI":"10.1109\/TNNLS.2020.3017939","article-title":"IAUnet: Global context-aware feature learning for person reidentification","volume":"32","author":"Hou","year":"2020","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"ref_24","doi-asserted-by":"crossref","unstructured":"Yuan, Y., Chen, X., and Wang, J. (2020). Object-contextual representations for semantic segmentation. Computer Vision\u2013ECCV 2020, Proceedings of the 16th European Conference, Glasgow, UK, 23\u201328 August 2020, Springer.","DOI":"10.1007\/978-3-030-58539-6_11"},{"key":"ref_25","doi-asserted-by":"crossref","unstructured":"Liu, J., Li, C., Liang, F., Lin, C., Sun, M., Yan, J., Ouyang, W., and Xu, D. (2021, January 19\u201325). Inception convolution with efficient dilation search. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA.","DOI":"10.1109\/CVPR46437.2021.01132"},{"key":"ref_26","doi-asserted-by":"crossref","unstructured":"Guo, Z., Zhang, X., Mu, H., Heng, W., Liu, Z., Wei, Y., and Sun, J. (2020, January 23\u201328). Single path one-shot neural architecture search with uniform sampling. Proceedings of the European Conference on Computer Vision, Glasgow, UK.","DOI":"10.1007\/978-3-030-58517-4_32"},{"key":"ref_27","unstructured":"Fang, J., Sun, Y., Peng, K., Zhang, Q., Li, Y., Liu, W., and Wang, X. (2020). Fast neural network adaptation via parameter remapping and architecture search. arXiv."},{"key":"ref_28","unstructured":"Liu, Y., Sun, Y., Xue, B., Zhang, M., Yen, G.G., and Tan, K.C. (2021). A survey on evolutionary neural architecture search. IEEE Trans. Neural Netw. Learn. Syst., 1\u201321."},{"key":"ref_29","doi-asserted-by":"crossref","unstructured":"Zhao, H., Shi, J., Qi, X., Wang, X., and Jia, J. (2017, January 21\u201326). Pyramid scene parsing network. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA.","DOI":"10.1109\/CVPR.2017.660"},{"key":"ref_30","doi-asserted-by":"crossref","unstructured":"Hou, Q., Zhou, D., and Feng, J. (2021, January 19\u201325). Coordinate attention for efficient mobile network design. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA.","DOI":"10.1109\/CVPR46437.2021.01350"},{"key":"ref_31","first-page":"91","article-title":"Faster r-cnn: Towards real-time object detection with region proposal networks","volume":"28","author":"Ren","year":"2015","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"ref_32","unstructured":"Bochkovskiy, A., Wang, C.Y., and Liao, H.Y.M. (2020). Yolov4: Optimal speed and accuracy of object detection. arXiv."},{"key":"ref_33","doi-asserted-by":"crossref","unstructured":"Nie, T., Han, X., He, B., Li, X., Liu, H., and Bi, G. (2020). Ship detection in panchromatic optical remote sensing images based on visual saliency and multi-dimensional feature description. Remote Sens., 12.","DOI":"10.3390\/rs12010152"},{"key":"ref_34","doi-asserted-by":"crossref","first-page":"7147","DOI":"10.1109\/TGRS.2018.2848901","article-title":"HSF-Net: Multiscale deep feature embedding for ship detection in optical remote sensing imagery","volume":"56","author":"Li","year":"2018","journal-title":"IEEE Trans. Geosci. Remote Sens."}],"container-title":["Remote Sensing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2072-4292\/13\/23\/4840\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T07:37:12Z","timestamp":1760168232000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2072-4292\/13\/23\/4840"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,11,29]]},"references-count":34,"journal-issue":{"issue":"23","published-online":{"date-parts":[[2021,12]]}},"alternative-id":["rs13234840"],"URL":"https:\/\/doi.org\/10.3390\/rs13234840","relation":{},"ISSN":["2072-4292"],"issn-type":[{"type":"electronic","value":"2072-4292"}],"subject":[],"published":{"date-parts":[[2021,11,29]]}}}