{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,15]],"date-time":"2026-01-15T03:25:50Z","timestamp":1768447550820,"version":"3.49.0"},"reference-count":49,"publisher":"MDPI AG","issue":"8","license":[{"start":{"date-parts":[[2023,4,8]],"date-time":"2023-04-08T00:00:00Z","timestamp":1680912000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"name":"Key Projects of Global Change and Response of Ministry of Science and Technology of China","award":["2020YFA0608203"],"award-info":[{"award-number":["2020YFA0608203"]}]},{"name":"Key Projects of Global Change and Response of Ministry of Science and Technology of China","award":["2023YFS0366"],"award-info":[{"award-number":["2023YFS0366"]}]},{"name":"Key Projects of Global Change and Response of Ministry of Science and Technology of China","award":["2021YFS0335"],"award-info":[{"award-number":["2021YFS0335"]}]},{"name":"Key Projects of Global Change and Response of Ministry of Science and Technology of China","award":["2022ZDZX0001"],"award-info":[{"award-number":["2022ZDZX0001"]}]},{"name":"Key Projects of Global Change and Response of Ministry of Science and Technology of China","award":["2023YFG0020"],"award-info":[{"award-number":["2023YFG0020"]}]},{"name":"Key Projects of Global Change and Response of Ministry of Science and Technology of China","award":["FY-APP-2021.0304"],"award-info":[{"award-number":["FY-APP-2021.0304"]}]},{"name":"Key Projects of Global Change and Response of Ministry of Science and Technology of China","award":["2023NSFSC0751"],"award-info":[{"award-number":["2023NSFSC0751"]}]},{"name":"Science and Technology Support Project of Sichuan Province","award":["2020YFA0608203"],"award-info":[{"award-number":["2020YFA0608203"]}]},{"name":"Science and Technology Support Project of Sichuan Province","award":["2023YFS0366"],"award-info":[{"award-number":["2023YFS0366"]}]},{"name":"Science and Technology Support Project of Sichuan Province","award":["2021YFS0335"],"award-info":[{"award-number":["2021YFS0335"]}]},{"name":"Science and Technology Support Project of Sichuan Province","award":["2022ZDZX0001"],"award-info":[{"award-number":["2022ZDZX0001"]}]},{"name":"Science and Technology Support Project of Sichuan Province","award":["2023YFG0020"],"award-info":[{"award-number":["2023YFG0020"]}]},{"name":"Science and Technology Support Project of Sichuan Province","award":["FY-APP-2021.0304"],"award-info":[{"award-number":["FY-APP-2021.0304"]}]},{"name":"Science and Technology Support Project of Sichuan Province","award":["2023NSFSC0751"],"award-info":[{"award-number":["2023NSFSC0751"]}]},{"name":"Fengyun Satellite Application Advance Plan","award":["2020YFA0608203"],"award-info":[{"award-number":["2020YFA0608203"]}]},{"name":"Fengyun Satellite Application Advance Plan","award":["2023YFS0366"],"award-info":[{"award-number":["2023YFS0366"]}]},{"name":"Fengyun Satellite Application Advance Plan","award":["2021YFS0335"],"award-info":[{"award-number":["2021YFS0335"]}]},{"name":"Fengyun Satellite Application Advance Plan","award":["2022ZDZX0001"],"award-info":[{"award-number":["2022ZDZX0001"]}]},{"name":"Fengyun Satellite Application Advance Plan","award":["2023YFG0020"],"award-info":[{"award-number":["2023YFG0020"]}]},{"name":"Fengyun Satellite Application Advance Plan","award":["FY-APP-2021.0304"],"award-info":[{"award-number":["FY-APP-2021.0304"]}]},{"name":"Fengyun Satellite Application Advance Plan","award":["2023NSFSC0751"],"award-info":[{"award-number":["2023NSFSC0751"]}]},{"name":"Natural Science Foundation of Sichuan Province","award":["2020YFA0608203"],"award-info":[{"award-number":["2020YFA0608203"]}]},{"name":"Natural Science Foundation of Sichuan Province","award":["2023YFS0366"],"award-info":[{"award-number":["2023YFS0366"]}]},{"name":"Natural Science Foundation of Sichuan Province","award":["2021YFS0335"],"award-info":[{"award-number":["2021YFS0335"]}]},{"name":"Natural Science Foundation of Sichuan Province","award":["2022ZDZX0001"],"award-info":[{"award-number":["2022ZDZX0001"]}]},{"name":"Natural Science Foundation of Sichuan Province","award":["2023YFG0020"],"award-info":[{"award-number":["2023YFG0020"]}]},{"name":"Natural Science Foundation of Sichuan Province","award":["FY-APP-2021.0304"],"award-info":[{"award-number":["FY-APP-2021.0304"]}]},{"name":"Natural Science Foundation of Sichuan Province","award":["2023NSFSC0751"],"award-info":[{"award-number":["2023NSFSC0751"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Remote Sensing"],"abstract":"<jats:p>With the development of deep learning and remote sensing technologies in recent years, many semantic segmentation methods based on convolutional neural networks (CNNs) have been applied to road extraction. However, previous deep learning-based road extraction methods primarily used RGB imagery as an input and did not take advantage of the spectral information contained in hyperspectral imagery. These methods can produce discontinuous outputs caused by objects with similar spectral signatures to roads. In addition, the images obtained from different Earth remote sensing sensors may have different spatial resolutions, enhancing the difficulty of the joint analysis. This work proposes the Multiscale Fusion Attention Network (MSFANet) to overcome these problems. Compared to traditional road extraction frameworks, the proposed MSFANet fuses information from different spectra at multiple scales. In MSFANet, multispectral remote sensing data is used as an additional input to the network, in addition to RGB remote sensing data, to obtain richer spectral information. The Cross-source Feature Fusion Module (CFFM) is used to calibrate and fuse spectral features at different scales, reducing the impact of noise and redundant features from different inputs. The Multiscale Semantic Aggregation Decoder (MSAD) fuses multiscale features and global context information from the upsampling process layer by layer, reducing information loss during the multiscale feature fusion. The proposed MSFANet network was applied to the SpaceNet dataset and self-annotated images from Chongzhou, a representative city in China. Our MSFANet performs better over the baseline HRNet by a large margin of +6.38 IoU and +5.11 F1-score on the SpaceNet dataset, +3.61 IoU and +2.32 F1-score on the self-annotated dataset (Chongzhou dataset). Moreover, the effectiveness of MSFANet was also proven by comparative experiments with other studies.<\/jats:p>","DOI":"10.3390\/rs15081978","type":"journal-article","created":{"date-parts":[[2023,4,10]],"date-time":"2023-04-10T03:19:54Z","timestamp":1681096794000},"page":"1978","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":16,"title":["MSFANet: Multiscale Fusion Attention Network for Road Segmentation of Multispectral Remote Sensing Data"],"prefix":"10.3390","volume":"15","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-9380-4806","authenticated-orcid":false,"given":"Zhonggui","family":"Tong","sequence":"first","affiliation":[{"name":"School of Automation Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China"}]},{"given":"Yuxia","family":"Li","sequence":"additional","affiliation":[{"name":"School of Automation Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China"}]},{"given":"Jinglin","family":"Zhang","sequence":"additional","affiliation":[{"name":"School of Automation Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9875-9853","authenticated-orcid":false,"given":"Lei","family":"He","sequence":"additional","affiliation":[{"name":"School of Software Engineering, Chengdu University of Information Technology, Chengdu 610225, China"}]},{"given":"Yushu","family":"Gong","sequence":"additional","affiliation":[{"name":"School of Automation Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China"}]}],"member":"1968","published-online":{"date-parts":[[2023,4,8]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","first-page":"288","DOI":"10.1016\/j.isprsjprs.2020.08.019","article-title":"BT-RoadNet: A boundary and topologically-aware neural network for road extraction from high-resolution remote sensing imagery","volume":"168","author":"Zhou","year":"2020","journal-title":"ISPRS J. Photogramm. Remote Sens."},{"key":"ref_2","doi-asserted-by":"crossref","first-page":"111418","DOI":"10.1016\/j.rse.2019.111418","article-title":"Remote sensing and GIS techniques for reconstructing the military fort system on the Roman boundary (Tunisian section) and identifying archaeological sites","volume":"236","author":"Bachagha","year":"2020","journal-title":"Remote Sens. Environ."},{"key":"ref_3","doi-asserted-by":"crossref","unstructured":"Jia, J., Sun, H., Jiang, C., Karila, K., Karjalainen, M., Ahokas, E., Khoramshahi, E., Hu, P., Chen, C., and Xue, T. (2021). Review on active and passive remote sensing techniques for road extraction. Remote Sens., 13.","DOI":"10.3390\/rs13214235"},{"key":"ref_4","doi-asserted-by":"crossref","first-page":"5063","DOI":"10.1109\/LRA.2022.3154052","article-title":"csBoundary: City-Scale Road-Boundary Detection in Aerial Images for High-Definition Maps","volume":"7","author":"Xu","year":"2022","journal-title":"IEEE Robot. Autom. Lett."},{"key":"ref_5","doi-asserted-by":"crossref","first-page":"540","DOI":"10.1109\/TVT.2013.2281199","article-title":"A sensor-fusion drivable-region and lane-detection system for autonomous vehicle navigation in challenging road scenarios","volume":"63","author":"Li","year":"2013","journal-title":"IEEE Trans. Veh. Technol."},{"key":"ref_6","doi-asserted-by":"crossref","unstructured":"Aboah, A. (2021, January 19\u201325). A Vision-Based System for Traffic Anomaly Detection Using Deep Learning and Decision Trees. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, Nashville, TN, USA.","DOI":"10.1109\/CVPRW53098.2021.00475"},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"1073","DOI":"10.1016\/S0167-8655(02)00054-5","article-title":"Geographic information system updating using remote sensing images","volume":"23","author":"Bonnefon","year":"2002","journal-title":"Pattern Recognit. Lett."},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"5489","DOI":"10.1109\/JSTARS.2020.3023549","article-title":"Road extraction methods in high-resolution remote sensing images: A comprehensive review","volume":"13","author":"Lian","year":"2020","journal-title":"IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens."},{"key":"ref_9","doi-asserted-by":"crossref","first-page":"121","DOI":"10.1023\/B:VISI.0000013086.45688.5d","article-title":"A Gibbs point process for road extraction from remotely sensed images","volume":"57","author":"Stoica","year":"2004","journal-title":"Int. J. Comput. Vis."},{"key":"ref_10","unstructured":"Bacher, U., and Mayer, H. (2005, January 29\u201330). Automatic road extraction from multispectral high resolution satellite images. Proceedings of the CMRT05, Vienna, Austria."},{"key":"ref_11","doi-asserted-by":"crossref","first-page":"44","DOI":"10.1111\/j.1477-9730.2006.00353.x","article-title":"Road extraction based on fuzzy logic and mathematical morphology from pan-sharpened Ikonos images","volume":"21","author":"Mohammadzadeh","year":"2006","journal-title":"Photogramm. Rec."},{"key":"ref_12","doi-asserted-by":"crossref","unstructured":"Maurya, R., Gupta, P., and Shukla, A.S. (2011, January 3\u20135). Road extraction using k-means clustering and morphological operations. Proceedings of the 2011 International Conference on Image Information Processing, Shimla, Himachal Pradesh, India.","DOI":"10.1109\/ICIIP.2011.6108839"},{"key":"ref_13","doi-asserted-by":"crossref","first-page":"1365","DOI":"10.14358\/PERS.70.12.1365","article-title":"Road extraction using SVM and image segmentation","volume":"70","author":"Song","year":"2004","journal-title":"Photogramm. Eng. Remote Sens."},{"key":"ref_14","doi-asserted-by":"crossref","first-page":"1192","DOI":"10.1109\/TIP.2005.864232","article-title":"Road extraction from aerial images using a region competition algorithm","volume":"15","author":"Amo","year":"2006","journal-title":"IEEE Trans. Image Process."},{"key":"ref_15","doi-asserted-by":"crossref","unstructured":"Yager, N., and Sowmya, A. (2003, January 25\u201327). Support vector machines for road extraction from remotely sensed images. Proceedings of the Computer Analysis of Images and Patterns: 10th International Conference, CAIP 2003, Groningen, The Netherlands.","DOI":"10.1007\/978-3-540-45179-2_36"},{"key":"ref_16","doi-asserted-by":"crossref","first-page":"539","DOI":"10.1109\/TGRS.2004.841395","article-title":"A Bayesian approach to classification of multiresolution remote sensing data","volume":"43","author":"Storvik","year":"2005","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_17","doi-asserted-by":"crossref","unstructured":"Long, J., Shelhamer, E., and Darrell, T. (2015, January 7\u201312). Fully convolutional networks for semantic segmentation. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA.","DOI":"10.1109\/CVPR.2015.7298965"},{"key":"ref_18","doi-asserted-by":"crossref","unstructured":"Chaurasia, A., and Culurciello, E. (2017, January 10\u201313). Linknet: Exploiting encoder representations for efficient semantic segmentation. Proceedings of the 2017 IEEE Visual Communications and Image Processing (VCIP), St. Petersburg, FL, USA.","DOI":"10.1109\/VCIP.2017.8305148"},{"key":"ref_19","doi-asserted-by":"crossref","unstructured":"Ronneberger, O., Fischer, P., and Brox, T. (2015, January 5\u20139). U-net: Convolutional networks for biomedical image segmentation. Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany.","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"ref_20","unstructured":"Ke, K., Zhao, Y., Jiang, B., Cheng, T., Xiao, B., Liu, D., Yadong, M., Wang, X., Liu, W., and Wang, J. (November, January 27). High-resolution representations for labeling pixels and regions. Proceedings of the IEEE International Conference on Computer Vision, Seoul, Republic of Korea."},{"key":"ref_21","first-page":"1","article-title":"DDU-Net: Dual-decoder-U-Net for road extraction using high-resolution remote sensing images","volume":"60","author":"Wang","year":"2022","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_22","doi-asserted-by":"crossref","first-page":"108549","DOI":"10.1016\/j.patcog.2022.108549","article-title":"SDUNet: Road extraction via spatial enhanced and densely connected UNet","volume":"126","author":"Yang","year":"2022","journal-title":"Pattern Recognit."},{"key":"ref_23","doi-asserted-by":"crossref","first-page":"7209","DOI":"10.1109\/TGRS.2019.2912301","article-title":"Road detection and centerline extraction via deep recurrent convolutional neural network U-Net","volume":"57","author":"Yang","year":"2019","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_24","doi-asserted-by":"crossref","first-page":"6302","DOI":"10.1109\/JSTARS.2021.3083055","article-title":"DA-RoadNet: A dual-attention network for road extraction from high resolution satellite imagery","volume":"14","author":"Wan","year":"2021","journal-title":"IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens."},{"key":"ref_25","doi-asserted-by":"crossref","unstructured":"Huan, H., Sheng, Y., Zhang, Y., and Liu, Y. (2022). Strip Attention Networks for Road Extraction. Remote Sens., 14.","DOI":"10.3390\/rs14184516"},{"key":"ref_26","doi-asserted-by":"crossref","unstructured":"Ma, W., Karaku\u015f, O., and Rosin, P.L. (2022). AMM-FuseNet: Attention-based multi-modal image fusion network for land cover mapping. Remote Sens., 14.","DOI":"10.20944\/preprints202207.0347.v1"},{"key":"ref_27","doi-asserted-by":"crossref","unstructured":"Zhao, H., Shi, J., Qi, X., Wang, X., and Jia, J. (2017, January 21\u201326). Pyramid scene parsing network. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA.","DOI":"10.1109\/CVPR.2017.660"},{"key":"ref_28","doi-asserted-by":"crossref","unstructured":"Chen, L.C., Zhu, Y., Papandreou, G., Schroff, F., and Adam, H. (2018, January 8\u201314). Encoder-decoder with atrous separable convolution for semantic image segmentation. Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany.","DOI":"10.1007\/978-3-030-01234-2_49"},{"key":"ref_29","doi-asserted-by":"crossref","unstructured":"Zhou, L., and Zhang, C. (2018, January 18\u201322). Linknet with pretrained encoder and dilated convolution for high resolution satellite imagery road extraction. Proceedings of the CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Salt Lake City, UT, USA.","DOI":"10.1109\/CVPRW.2018.00034"},{"key":"ref_30","doi-asserted-by":"crossref","unstructured":"He, S., Bastani, F., Jagwani, S., Alizadeh, M., Balakrishnan, H., Chawla, S., Elshrif, M.M., Madden, S., and Sadeghi, M.A. (2020, January 23\u201328). Sat2graph: Road graph extraction through graph-tensor encoding. Proceedings of the European Conference on Computer Vision, Glasgow, UK.","DOI":"10.1007\/978-3-030-58586-0_4"},{"key":"ref_31","doi-asserted-by":"crossref","unstructured":"Xie, Y., Miao, F., Zhou, K., and Peng, J. (2019). HsgNet: A road extraction network based on global perception of high-order spatial information. ISPRS Int. J. Geo Inf., 8.","DOI":"10.3390\/ijgi8120571"},{"key":"ref_32","first-page":"1","article-title":"DBRANet: Road extraction by dual-branch encoder and regional attention decoder","volume":"19","author":"Chen","year":"2021","journal-title":"IEEE Geosci. Remote Sens. Lett."},{"key":"ref_33","first-page":"1","article-title":"NL-LinkNet: Toward lighter but more accurate road extraction with nonlocal operations","volume":"19","author":"Wang","year":"2021","journal-title":"IEEE Geosci. Remote Sens. Lett."},{"key":"ref_34","doi-asserted-by":"crossref","unstructured":"Wang, X., Girshick, R., Gupta, A., and He, K. (2018, January 18\u201322). Non-local neural networks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA.","DOI":"10.1109\/CVPR.2018.00813"},{"key":"ref_35","doi-asserted-by":"crossref","unstructured":"Hu, J., Shen, L., and Sun, G. (2018, January 18\u201322). Squeeze-and-excitation networks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA.","DOI":"10.1109\/CVPR.2018.00745"},{"key":"ref_36","unstructured":"Huang, Z., Wang, X., Huang, L., Huang, C., Wei, Y., and Liu, W. (November, January 27). Ccnet: Criss-cross attention for semantic segmentation. Proceedings of the IEEE\/CVF International Conference on Computer Vision, Seoul, Republic of Korea."},{"key":"ref_37","doi-asserted-by":"crossref","unstructured":"Fu, J., Liu, J., Tian, H., Li, Y., Bao, Y., Fang, Z., and Lu, H. (2019, January 15\u201320). Dual attention network for scene segmentation. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA.","DOI":"10.1109\/CVPR.2019.00326"},{"key":"ref_38","unstructured":"Li, H., Xiong, P., An, J., and Wang, L. (2018, January 3\u20136). Pyramid attention network for semantic segmentation. Proceedings of the British Machine Vision Conference, Newcastle, UK."},{"key":"ref_39","unstructured":"Rudner, T.G., Ru\u00dfwurm, M., Fil, J., Pelich, R., Bischke, B., Kopa\u010dkov\u00e1, V., and Bili\u0144ski, P. (February, January 27). Multi3net: Segmenting flooded buildings via fusion of multiresolution, multisensor, and multitemporal satellite imagery. Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA."},{"key":"ref_40","doi-asserted-by":"crossref","first-page":"3463","DOI":"10.1109\/JSTARS.2022.3165005","article-title":"A crossmodal multiscale fusion network for semantic segmentation of remote sensing data","volume":"15","author":"Ma","year":"2022","journal-title":"IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens."},{"key":"ref_41","doi-asserted-by":"crossref","unstructured":"Lei, T., Li, L., Lv, Z., Zhu, M., Du, X., and Nandi, A.K. (2021). Multi-modality and multi-scale attention fusion network for land cover classification from VHR remote sensing images. Remote Sens., 13.","DOI":"10.3390\/rs13183771"},{"key":"ref_42","first-page":"1","article-title":"Deep multimodal fusion network for semantic segmentation using remote sensing image and LiDAR data","volume":"60","author":"Sun","year":"2021","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_43","doi-asserted-by":"crossref","unstructured":"Cao, Z., Diao, W., Sun, X., Lyu, X., Yan, M., and Fu, K. (2021). C3net: Cross-modal feature recalibrated, cross-scale semantic aggregated and compact network for semantic segmentation of multi-modal high-resolution aerial images. Remote Sens., 13.","DOI":"10.3390\/rs13030528"},{"key":"ref_44","first-page":"102910","article-title":"Multispectral panoptic segmentation: Exploring the beach setting with worldview-3 imagery","volume":"112","author":"Santana","year":"2022","journal-title":"Int. J. Appl. Earth Obs. Geoinf."},{"key":"ref_45","unstructured":"Van Etten, A., Lindenbaum, D., and Bacastow, T.M. (2018). Spacenet: A remote sensing dataset and challenge series. arXiv."},{"key":"ref_46","doi-asserted-by":"crossref","unstructured":"Batra, A., Singh, S., Pang, G., Basu, S., Jawahar, C., and Paluri, M. (2019, January 15\u201320). Improved road connectivity by joint learning of orientation and segmentation. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA.","DOI":"10.1109\/CVPR.2019.01063"},{"key":"ref_47","doi-asserted-by":"crossref","unstructured":"Selvaraju, R.R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., and Batra, D. (2017, January 22\u201329). Grad-CAM: Visual Explanations From Deep Networks via Gradient-Based Localization. Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy.","DOI":"10.1109\/ICCV.2017.74"},{"key":"ref_48","unstructured":"Zhou, B., Khosla, A., Lapedriza, A., Oliva, A., and Torralba, A. (July, January 26). Learning deep features for discriminative localization. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA."},{"key":"ref_49","doi-asserted-by":"crossref","unstructured":"Chu, X., Chen, L., and Yu, W. (2022, January 18\u201324). NAFSSR: Stereo Image Super-Resolution Using NAFNet. Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA.","DOI":"10.1109\/CVPRW56347.2022.00130"}],"container-title":["Remote Sensing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2072-4292\/15\/8\/1978\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T19:12:39Z","timestamp":1760123559000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2072-4292\/15\/8\/1978"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,4,8]]},"references-count":49,"journal-issue":{"issue":"8","published-online":{"date-parts":[[2023,4]]}},"alternative-id":["rs15081978"],"URL":"https:\/\/doi.org\/10.3390\/rs15081978","relation":{},"ISSN":["2072-4292"],"issn-type":[{"value":"2072-4292","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,4,8]]}}}