{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,11]],"date-time":"2025-10-11T01:18:25Z","timestamp":1760145505801,"version":"build-2065373602"},"reference-count":67,"publisher":"MDPI AG","issue":"15","license":[{"start":{"date-parts":[[2024,7,24]],"date-time":"2024-07-24T00:00:00Z","timestamp":1721779200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100001809","name":"Natural Science Foundation of China","doi-asserted-by":"publisher","award":["U21A20139"],"award-info":[{"award-number":["U21A20139"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Remote Sensing"],"abstract":"<jats:p>Cables are vital load-bearing components of cable-stayed bridges. Surface defects can lead to internal corrosion and fracturing, significantly impacting the stability of the bridge structure. The detection of surface defects from bridge cable images faces numerous challenges, including shadow disturbances due to uneven lighting and difficulties in addressing multiscale defect features. To address these challenges, this paper proposes a novel and cost-effective deep learning segmentation network, named Trans-DCN, to detect defects in the surface of the bridge cable. The network leverages an efficient Transformer-based encoder and integrates multiscale features to overcome the limitations associated with local feature inadequacy. The decoder implements an atrous Deformable Convolution (DCN) pyramid and dynamically fuses low-level feature information to perceive the complex distribution of defects. The effectiveness of Trans-DCN is evaluated by comparing it with state-of-the-art segmentation baseline models using a dataset comprising cable bridge defect images. Experimental results demonstrate that our network outperforms the state-of-the-art network SegFormer, achieving a 27.1% reduction in GFLOPs, a 1.2% increase in mean Intersection over Union, and a 1.5% increase in the F1 score. Ablation experiments confirmed the effectiveness of each module within our network, further substantiating the significant validity and advantages of Trans-DCN in the task of bridge cable defect segmentation. The network proposed in this paper provides an effective solution for downstream cable bridge image analysis.<\/jats:p>","DOI":"10.3390\/rs16152711","type":"journal-article","created":{"date-parts":[[2024,7,24]],"date-time":"2024-07-24T14:55:47Z","timestamp":1721832947000},"page":"2711","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":3,"title":["Trans-DCN: A High-Efficiency and Adaptive Deep Network for Bridge Cable Surface Defect Segmentation"],"prefix":"10.3390","volume":"16","author":[{"given":"Zhihai","family":"Huang","sequence":"first","affiliation":[{"name":"School of Civil and Transportation Engineering, Guangdong University of Technology, Guangzhou 510006, China"}]},{"given":"Bo","family":"Guo","sequence":"additional","affiliation":[{"name":"School of Architecture and Urban Planning, Shenzhen University, Shenzhen 518060, China"}]},{"given":"Xiaolong","family":"Deng","sequence":"additional","affiliation":[{"name":"School of Civil and Transportation Engineering, Guangdong University of Technology, Guangzhou 510006, China"}]},{"given":"Wenchao","family":"Guo","sequence":"additional","affiliation":[{"name":"School of Civil and Transportation Engineering, Guangdong University of Technology, Guangzhou 510006, China"}]},{"given":"Xing","family":"Min","sequence":"additional","affiliation":[{"name":"Guangdong Metro Design Research Institute Co., Ltd., Guangzhou 510499, China"}]}],"member":"1968","published-online":{"date-parts":[[2024,7,24]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"44485","DOI":"10.1109\/ACCESS.2019.2961755","article-title":"Particle swarm optimization-based SVM for classification of cable surface defects of the cable-stayed bridges","volume":"8","author":"Li","year":"2019","journal-title":"IEEE Access"},{"key":"ref_2","doi-asserted-by":"crossref","first-page":"254","DOI":"10.1016\/j.jsv.2016.04.025","article-title":"Vibration characteristics and damage detection in a suspension bridge","volume":"375","author":"Wickramasinghe","year":"2016","journal-title":"J. Sound Vib."},{"key":"ref_3","doi-asserted-by":"crossref","first-page":"297","DOI":"10.1177\/1475921706067742","article-title":"Feature extraction for defect detection in strands by guided ultrasonic waves","volume":"5","author":"Rizzo","year":"2006","journal-title":"Struct. Health Monit."},{"key":"ref_4","doi-asserted-by":"crossref","first-page":"1077","DOI":"10.1016\/j.optlaseng.2009.04.016","article-title":"Applications of optical fibre Bragg gratings sensing technology-based smart stay cables","volume":"47","author":"Li","year":"2009","journal-title":"Opt. Lasers Eng."},{"doi-asserted-by":"crossref","unstructured":"Cho, K.H., Jin, Y.H., Kim, H.M., Moon, H., Koo, J.C., and Choi, H.R. (2013, January 17\u201320). Caterpillar-based cable climbing robot for inspection of suspension bridge hanger rope. Proceedings of the 2013 IEEE International Conference on Automation Science and Engineering (CASE), Madison, WI, USA.","key":"ref_5","DOI":"10.1109\/CoASE.2013.6653913"},{"key":"ref_6","doi-asserted-by":"crossref","first-page":"269","DOI":"10.1139\/tcsme-2011-0016","article-title":"Climbing model and obstacle-climbing performance of a cable inspection robot for a cable-stayed bridge","volume":"35","author":"Xu","year":"2011","journal-title":"Trans. Can. Soc. Mech. Eng."},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"75","DOI":"10.1007\/s10846-020-01266-1","article-title":"A climbing robot for steel bridge inspection","volume":"102","author":"Nguyen","year":"2021","journal-title":"J. Intell. Robot. Syst."},{"key":"ref_8","first-page":"10","article-title":"Lightweight Semantic Segmentation Network for RGB-D Image Based on Attention Mechanism","volume":"43","author":"Sun","year":"2022","journal-title":"Packag. Eng."},{"key":"ref_9","first-page":"6","article-title":"Lightweight and multi-pose face recognition method based on deep learning","volume":"40","author":"Gong","year":"2020","journal-title":"J. Comput. Appl."},{"key":"ref_10","doi-asserted-by":"crossref","first-page":"101882","DOI":"10.1016\/j.aei.2023.101882","article-title":"Defect-aware transformer network for intelligent visual surface defect detection","volume":"55","author":"Shang","year":"2023","journal-title":"Adv. Eng. Inform."},{"unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., and Gelly, S. (2020). An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv.","key":"ref_11"},{"unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, L., and Polosukhin, I. (2017). Attention is all you need. arXiv.","key":"ref_12"},{"unstructured":"Simonyan, K., and Zisserman, A. (2015, January 7\u20139). Very deep convolutional networks for large-scale image recognition. Proceedings of the 3rd International Conference on Learning Representations (ICLR 2015), San Diego, CA, USA.","key":"ref_13"},{"doi-asserted-by":"crossref","unstructured":"Dai, J., Qi, H., Xiong, Y., Li, Y., Zhang, G., Hu, H., and Wei, Y. (2017, January 22\u201329). Deformable Convolutional Networks. Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy.","key":"ref_14","DOI":"10.1109\/ICCV.2017.89"},{"key":"ref_15","doi-asserted-by":"crossref","first-page":"36","DOI":"10.1016\/j.optlastec.2018.07.012","article-title":"Cable surface damage detection in cable-stayed bridges using optical techniques and image mosaicking","volume":"110","author":"Li","year":"2019","journal-title":"Opt. Laser Technol."},{"key":"ref_16","doi-asserted-by":"crossref","first-page":"170","DOI":"10.1016\/j.engstruct.2018.05.084","article-title":"Emerging artificial intelligence methods in structural engineering","volume":"171","author":"Salehi","year":"2018","journal-title":"Eng. Struct."},{"doi-asserted-by":"crossref","unstructured":"Luo, K., Kong, X., Zhang, J., Hu, J., Li, J., and Tang, H. (2023). Computer vision-based bridge inspection and monitoring: A review. Sensors, 23.","key":"ref_17","DOI":"10.3390\/s23187863"},{"key":"ref_18","doi-asserted-by":"crossref","first-page":"16","DOI":"10.1016\/j.engstruct.2017.10.057","article-title":"Visual data classification in post-event building reconnaissance","volume":"155","author":"Yeum","year":"2018","journal-title":"Eng. Struct."},{"doi-asserted-by":"crossref","unstructured":"Xu, Z., Wang, Y., Hao, X., and Fan, J. (2023). Crack Detection of Bridge Concrete Components Based on Large-Scene Images Using an Unmanned Aerial Vehicle. Sensors, 23.","key":"ref_19","DOI":"10.3390\/s23146271"},{"doi-asserted-by":"crossref","unstructured":"Chen, J., Wang, H., Tu, C.L., Wang, X.S., and Li, X.D. (2021, January 14\u201317). Surface Defect Detection of Cable Based on Threshold Image Difference. Proceedings of the 2021 IEEE Far East NDT New Technology & Application Forum (FENDT), Kunming, China.","key":"ref_20","DOI":"10.1109\/FENDT54151.2021.9749676"},{"doi-asserted-by":"crossref","unstructured":"Hu, J., He, H., Liao, G., and Hu, G. (2022). Study on Image Processing of Bridge Cable Surface Defect Detection System. Advances in Precision Instruments and Optical Engineering, Proceedings of the International Conference on Precision Instruments and Optical Engineering, Chengdu, China, 25\u201327 August 2021, Springer.","key":"ref_21","DOI":"10.1007\/978-981-16-7258-3_42"},{"key":"ref_22","doi-asserted-by":"crossref","first-page":"513","DOI":"10.1016\/j.measurement.2009.12.022","article-title":"A SVM-based pipeline leakage detection and pre-warning system","volume":"43","author":"Qu","year":"2010","journal-title":"Measurement"},{"key":"ref_23","doi-asserted-by":"crossref","first-page":"04020038","DOI":"10.1061\/(ASCE)CP.1943-5487.0000918","article-title":"Machine learning for crack detection: Review and model performance comparison","volume":"34","author":"Hsieh","year":"2020","journal-title":"J. Comput. Civ. Eng."},{"key":"ref_24","doi-asserted-by":"crossref","first-page":"361","DOI":"10.1111\/mice.12263","article-title":"Deep learning-based crack damage detection using convolutional neural networks","volume":"32","author":"Cha","year":"2017","journal-title":"Comput.-Aided Civ. Infrastruct. Eng."},{"key":"ref_25","doi-asserted-by":"crossref","first-page":"52","DOI":"10.1016\/j.autcon.2018.11.028","article-title":"Autonomous concrete crack detection using deep fully convolutional neural network","volume":"99","author":"Dung","year":"2019","journal-title":"Autom. Constr."},{"unstructured":"Dong, S., Tan, H., Liu, C., and Hu, X. (2024). Apparent disease detection of bridges based on improved YOLOv5s. J. Chongqing Univ., 1\u201312. Available online: https:\/\/link.cnki.net\/urlid\/50.1044.N.20230331.1847.002.","key":"ref_26"},{"key":"ref_27","doi-asserted-by":"crossref","first-page":"1137","DOI":"10.1109\/TPAMI.2016.2577031","article-title":"Faster R-CNN: Towards real-time object detection with region proposal networks","volume":"39","author":"Ren","year":"2016","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"doi-asserted-by":"crossref","unstructured":"Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C.Y., and Berg, A.C. (2016). Ssd: Single shot multibox detector. Computer Vision\u2014ECCV 2016, Proceedings of the 14th European Conference, Amsterdam, The Netherlands, 11\u201314 October 2016, Springer. Proceedings, Part I 14.","key":"ref_28","DOI":"10.1007\/978-3-319-46448-0_2"},{"unstructured":"Bochkovskiy, A., Wang, C.Y., and Liao, H.Y.M. (2020). Yolov4: Optimal speed and accuracy of object detection. arXiv.","key":"ref_29"},{"doi-asserted-by":"crossref","unstructured":"Zeiler, M.D., and Fergus, R. (2014). Visualizing and understanding convolutional networks. Computer Vision\u2014ECCV 2014, Proceedings of the 13th European Conference, Zurich, Switzerland, 6\u201312 September 2014, Springer. Proceedings, Part I 13.","key":"ref_30","DOI":"10.1007\/978-3-319-10590-1_53"},{"key":"ref_31","doi-asserted-by":"crossref","first-page":"051101","DOI":"10.1007\/s11432-017-9189-6","article-title":"Survey of recent progress in semantic image segmentation with CNNs","volume":"61","author":"Geng","year":"2018","journal-title":"Sci. China Inf. Sci."},{"key":"ref_32","doi-asserted-by":"crossref","first-page":"16878132221122770","DOI":"10.1177\/16878132221122770","article-title":"Review of artificial intelligence-based bridge damage detection","volume":"14","author":"Zhang","year":"2022","journal-title":"Adv. Mech. Eng."},{"key":"ref_33","doi-asserted-by":"crossref","first-page":"640","DOI":"10.1109\/TPAMI.2016.2572683","article-title":"Fully convolutional networks for semantic segmentation","volume":"39","author":"Long","year":"2017","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_34","doi-asserted-by":"crossref","first-page":"302","DOI":"10.1016\/j.neucom.2019.11.118","article-title":"A brief survey on semantic segmentation with deep learning","volume":"406","author":"Hao","year":"2020","journal-title":"Neurocomputing"},{"doi-asserted-by":"crossref","unstructured":"Shi, J., Dang, J., Cui, M., Zuo, R., Shimizu, K., Tsunoda, A., and Suzuki, Y. (2021). Improvement of damage segmentation based on pixel-level data balance using vgg-unet. Appl. Sci., 11.","key":"ref_35","DOI":"10.3390\/app11020518"},{"doi-asserted-by":"crossref","unstructured":"Zhou, Z., Rahman Siddiquee, M.M., Tajbakhsh, N., and Liang, J. (2018, January 20). Unet++: A nested u-net architecture for medical image segmentation. Proceedings of the Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support, Granada, Spain.","key":"ref_36","DOI":"10.1007\/978-3-030-00889-5_1"},{"key":"ref_37","doi-asserted-by":"crossref","first-page":"102973","DOI":"10.1016\/j.autcon.2019.102973","article-title":"Vision based pixel-level bridge structural damage detection using a link ASPP network","volume":"110","author":"Deng","year":"2020","journal-title":"Autom. Constr."},{"doi-asserted-by":"crossref","unstructured":"Zheng, S., Lu, J., Zhao, H., Zhu, X., Luo, Z., Wang, Y., Fu, Y., Feng, J., Xiang, T., and Torr, P.H. (2021, January 19\u201325). Rethinking Semantic Segmentation from a Sequence-to-Sequence Perspective with Transformers. Proceedings of the 2021 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA.","key":"ref_38","DOI":"10.1109\/CVPR46437.2021.00681"},{"unstructured":"Chen, J., Lu, Y., Yu, Q., Luo, X., Adeli, E., Wang, Y., Lu, L., Yuille, A.L., and Zhou, Y. (2021). Transunet: Transformers make strong encoders for medical image segmentation. arXiv.","key":"ref_39"},{"key":"ref_40","doi-asserted-by":"crossref","first-page":"106669","DOI":"10.1016\/j.engappai.2023.106669","article-title":"Semantic segmentation using Vision Transformers: A survey","volume":"126","author":"Thisanke","year":"2023","journal-title":"Eng. Appl. Artif. Intell."},{"doi-asserted-by":"crossref","unstructured":"Liu, Z., Lin, Y., Cao, Y., Hu, H., Wei, Y., Zhang, Z., Lin, S., and Guo, B. (2021, January 10\u201317). Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada.","key":"ref_41","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"ref_42","doi-asserted-by":"crossref","first-page":"4408715","DOI":"10.1109\/TGRS.2022.3144165","article-title":"Swin transformer embedding UNet for remote sensing image semantic segmentation","volume":"60","author":"He","year":"2022","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"doi-asserted-by":"crossref","unstructured":"Wang, W., Xie, E., Li, X., Fan, D.P., Song, K., Liang, D., Lu, T., Luo, P., and Shao, L. (2021, January 10\u201317). Pyramid Vision Transformer: A Versatile Backbone for Dense Prediction Without Convolutions. Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada.","key":"ref_43","DOI":"10.1109\/ICCV48922.2021.00061"},{"unstructured":"Howard, A.G., Zhu, M., Chen, B., Kalenichenko, D., Wang, W., Weyand, T., Andreetto, M., and Adam, H. (2017). Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv.","key":"ref_44"},{"doi-asserted-by":"crossref","unstructured":"Chollet, F. (2017, January 21\u201326). Xception: Deep Learning With Depthwise Separable Convolutions. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA.","key":"ref_45","DOI":"10.1109\/CVPR.2017.195"},{"doi-asserted-by":"crossref","unstructured":"Fang, M., Liang, X., Fu, F., Song, Y., and Shao, Z. (2020). Attention mechanism based semi-supervised multi-gain image fusion. Symmetry, 12.","key":"ref_46","DOI":"10.3390\/sym12030451"},{"doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., and Sun, J. (2016, January 27\u201330). Deep Residual Learning for Image Recognition. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA.","key":"ref_47","DOI":"10.1109\/CVPR.2016.90"},{"unstructured":"Islam, M.A., Jia, S., and Bruce, N.D. (2019, January 6\u20139). How much Position Information Do Convolutional Neural Networks Encode?. Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA.","key":"ref_48"},{"unstructured":"Chen, L.C., Papandreou, G., Schroff, F., and Adam, H. (2017). Rethinking Atrous Convolution for Semantic Image Segmentation. arXiv.","key":"ref_49"},{"doi-asserted-by":"crossref","unstructured":"Yang, M., Yu, K., Zhang, C., Li, Z., and Yang, K. (2018, January 18\u201323). DenseASPP for Semantic Segmentation in Street Scenes. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA.","key":"ref_50","DOI":"10.1109\/CVPR.2018.00388"},{"doi-asserted-by":"crossref","unstructured":"Wang, W., Dai, J., Chen, Z., Huang, Z., Li, Z., Zhu, X., Hu, X., Lu, T., Lu, L., and Li, H. (2023, January 17\u201324). Internimage: Exploring large-scale vision foundation models with deformable convolutions. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada.","key":"ref_51","DOI":"10.1109\/CVPR52729.2023.01385"},{"doi-asserted-by":"crossref","unstructured":"Milletari, F., Navab, N., and Ahmadi, S.A. (2016, January 25\u201328). V-Net: Fully Convolutional Neural Networks for Volumetric Medical Image Segmentation. Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA.","key":"ref_52","DOI":"10.1109\/3DV.2016.79"},{"doi-asserted-by":"crossref","unstructured":"Ronneberger, O., Fischer, P., and Brox, T. (2015). U-net: Convolutional networks for biomedical image segmentation. Medical Image Computing and Computer-Assisted Intervention\u2013MICCAI 2015, Proceedings of the 18th International Conference, Munich, Germany, 5\u20139 October 2015, Springer. Proceedings, Part III 18.","key":"ref_53","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"ref_54","doi-asserted-by":"crossref","first-page":"2481","DOI":"10.1109\/TPAMI.2016.2644615","article-title":"Segnet: A deep convolutional encoder-decoder architecture for image segmentation","volume":"39","author":"Badrinarayanan","year":"2017","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"doi-asserted-by":"crossref","unstructured":"Zhao, H., Shi, J., Qi, X., Wang, X., and Jia, J. (2017, January 14\u201319). Pyramid Scene Parsing Network. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA.","key":"ref_55","DOI":"10.1109\/CVPR.2017.660"},{"doi-asserted-by":"crossref","unstructured":"Chen, L.C., Zhu, Y., Papandreou, G., Schroff, F., and Adam, H. (2018, January 8\u201314). Encoder-decoder with atrous separable convolution for semantic image segmentation. Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany.","key":"ref_56","DOI":"10.1007\/978-3-030-01234-2_49"},{"doi-asserted-by":"crossref","unstructured":"Fu, J., Liu, J., Tian, H., Li, Y., Bao, Y., Fang, Z., and Lu, H. (2019, January 15\u201320). Dual Attention Network for Scene Segmentation. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA.","key":"ref_57","DOI":"10.1109\/CVPR.2019.00326"},{"unstructured":"Howard, A., Sandler, M., Chu, G., Chen, L.C., Chen, B., Tan, M., Wang, W., Zhu, Y., Pang, R., and Vasudevan, V. (November, January 27). Searching for MobileNetV3. Proceedings of the IEEE International Conference on Computer Vision (ICCV), Seoul, Republic of Korea.","key":"ref_58"},{"unstructured":"Guo, M.H., Lu, C.Z., Hou, Q., Liu, Z., Cheng, M.M., and Hu, S.M. (2022). SegNeXt: Rethinking Convolutional Attention Design for Semantic Segmentation. arXiv.","key":"ref_59"},{"key":"ref_60","first-page":"12077","article-title":"SegFormer: Simple and efficient design for semantic segmentation with transformers","volume":"34","author":"Xie","year":"2021","journal-title":"Adv. Neural Inf. Process. Syst."},{"doi-asserted-by":"crossref","unstructured":"Cao, H., Wang, Y., Chen, J., Jiang, D., Zhang, X., Tian, Q., and Wang, M. (2022, January 23\u201327). Swin-Unet: Unet-Like Pure Transformer for Medical Image Segmentation. Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel.","key":"ref_61","DOI":"10.1007\/978-3-031-25066-8_9"},{"doi-asserted-by":"crossref","unstructured":"Zhu, L., Wang, X., Ke, Z., Zhang, W., and Lau, R.W. (2023, January 17\u201324). BiFormer: Vision Transformer with Bi-Level Routing Attention. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada.","key":"ref_62","DOI":"10.1109\/CVPR52729.2023.00995"},{"doi-asserted-by":"crossref","unstructured":"Xia, C., Wang, X., Lv, F., Hao, X., and Shi, Y. (2024). ViT-CoMer: Vision Transformer with Convolutional Multi-scale Feature Interaction for Dense Predictions. arXiv.","key":"ref_63","DOI":"10.1109\/CVPR52733.2024.00525"},{"doi-asserted-by":"crossref","unstructured":"Xie, S., Girshick, R., Doll\u00e1r, P., Tu, Z., and He, K. (2017, January 21\u201326). Aggregated residual transformations for deep neural networks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA.","key":"ref_64","DOI":"10.1109\/CVPR.2017.634"},{"doi-asserted-by":"crossref","unstructured":"Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., and Chen, L.C. (2018, January 18\u201323). Mobilenetv2: Inverted residuals and linear bottlenecks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA.","key":"ref_65","DOI":"10.1109\/CVPR.2018.00474"},{"doi-asserted-by":"crossref","unstructured":"Woo, S., Debnath, S., Hu, R., Chen, X., Liu, Z., Kweon, I.S., and Xie, S. (2023, January 17\u201324). Convnext v2: Co-designing and scaling convnets with masked autoencoders. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada.","key":"ref_66","DOI":"10.1109\/CVPR52729.2023.01548"},{"doi-asserted-by":"crossref","unstructured":"Liu, Z., Hu, H., Lin, Y., Yao, Z., Xie, Z., Wei, Y., Ning, J., Cao, Y., Zhang, Z., and Dong, L. (2022, January 18\u201324). Swin transformer v2: Scaling up capacity and resolution. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA.","key":"ref_67","DOI":"10.1109\/CVPR52688.2022.01170"}],"container-title":["Remote Sensing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2072-4292\/16\/15\/2711\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T15:22:34Z","timestamp":1760109754000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2072-4292\/16\/15\/2711"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,7,24]]},"references-count":67,"journal-issue":{"issue":"15","published-online":{"date-parts":[[2024,8]]}},"alternative-id":["rs16152711"],"URL":"https:\/\/doi.org\/10.3390\/rs16152711","relation":{},"ISSN":["2072-4292"],"issn-type":[{"type":"electronic","value":"2072-4292"}],"subject":[],"published":{"date-parts":[[2024,7,24]]}}}