{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,16]],"date-time":"2026-04-16T23:28:44Z","timestamp":1776382124020,"version":"3.51.2"},"reference-count":49,"publisher":"MDPI AG","issue":"12","license":[{"start":{"date-parts":[[2023,6,15]],"date-time":"2023-06-15T00:00:00Z","timestamp":1686787200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"name":"Key Project of Tianjin Natural Science Foundation","award":["21JCZDJC00670"],"award-info":[{"award-number":["21JCZDJC00670"]}]},{"name":"Key Project of Tianjin Natural Science Foundation","award":["No. 2021ZH04"],"award-info":[{"award-number":["No. 2021ZH04"]}]},{"name":"Key Project of Tianjin Natural Science Foundation","award":["2022-40"],"award-info":[{"award-number":["2022-40"]}]},{"name":"Key Project of Tianjin Natural Science Foundation","award":["2020-02"],"award-info":[{"award-number":["2020-02"]}]},{"name":"Key Project of Tianjin Natural Science Foundation","award":["No. 41601446"],"award-info":[{"award-number":["No. 41601446"]}]},{"name":"National Engineering Laboratory for Digital Construction and Evaluation Technology of Urban Rail Transit","award":["21JCZDJC00670"],"award-info":[{"award-number":["21JCZDJC00670"]}]},{"name":"National Engineering Laboratory for Digital Construction and Evaluation Technology of Urban Rail Transit","award":["No. 2021ZH04"],"award-info":[{"award-number":["No. 2021ZH04"]}]},{"name":"National Engineering Laboratory for Digital Construction and Evaluation Technology of Urban Rail Transit","award":["2022-40"],"award-info":[{"award-number":["2022-40"]}]},{"name":"National Engineering Laboratory for Digital Construction and Evaluation Technology of Urban Rail Transit","award":["2020-02"],"award-info":[{"award-number":["2020-02"]}]},{"name":"National Engineering Laboratory for Digital Construction and Evaluation Technology of Urban Rail Transit","award":["No. 41601446"],"award-info":[{"award-number":["No. 41601446"]}]},{"name":"Tianjin Transportation Science and Technology Development Project","award":["21JCZDJC00670"],"award-info":[{"award-number":["21JCZDJC00670"]}]},{"name":"Tianjin Transportation Science and Technology Development Project","award":["No. 2021ZH04"],"award-info":[{"award-number":["No. 2021ZH04"]}]},{"name":"Tianjin Transportation Science and Technology Development Project","award":["2022-40"],"award-info":[{"award-number":["2022-40"]}]},{"name":"Tianjin Transportation Science and Technology Development Project","award":["2020-02"],"award-info":[{"award-number":["2020-02"]}]},{"name":"Tianjin Transportation Science and Technology Development Project","award":["No. 41601446"],"award-info":[{"award-number":["No. 41601446"]}]},{"name":"National Natural Science Foundation of China Grant","award":["21JCZDJC00670"],"award-info":[{"award-number":["21JCZDJC00670"]}]},{"name":"National Natural Science Foundation of China Grant","award":["No. 2021ZH04"],"award-info":[{"award-number":["No. 2021ZH04"]}]},{"name":"National Natural Science Foundation of China Grant","award":["2022-40"],"award-info":[{"award-number":["2022-40"]}]},{"name":"National Natural Science Foundation of China Grant","award":["2020-02"],"award-info":[{"award-number":["2020-02"]}]},{"name":"National Natural Science Foundation of China Grant","award":["No. 41601446"],"award-info":[{"award-number":["No. 41601446"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Remote Sensing"],"abstract":"<jats:p>Synthetic aperture radar (SAR) imagery has the advantages of all-day and all-weather observation. However, due to the imaging mechanism of microwaves, it is difficult for nonexperts to interpret SAR images. Transferring SAR imagery into optical imagery can better improve the interpretation of SAR data and support the further fusion research of multi-source remote sensing. Methods based on generative adversarial networks (GAN) have been proven to be effective in SAR-to-optical translation tasks. To further improve the translation results of SAR data, we propose a method of an adjacent dual-decoder UNet (ADD-UNet) based on conditional GAN (cGAN) for SAR-to-optical translation. The proposed network architecture adds an adjacent scale of the decoder to the UNet, and the multi-scale feature aggregation of the two decoders improves structures, details, and edge sharpness of generated images while introducing fewer parameters compared with UNet++. In addition, we combine multi-scale structure similarity (MS-SSIM) loss and L1 loss as loss functions with cGAN loss together to help preserve structures and details. The experimental results demonstrate the superiority of our method compared with several state-of-the-art methods.<\/jats:p>","DOI":"10.3390\/rs15123125","type":"journal-article","created":{"date-parts":[[2023,6,15]],"date-time":"2023-06-15T02:03:19Z","timestamp":1686794599000},"page":"3125","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":12,"title":["ADD-UNet: An Adjacent Dual-Decoder UNet for SAR-to-Optical Translation"],"prefix":"10.3390","volume":"15","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-0375-9947","authenticated-orcid":false,"given":"Qingli","family":"Luo","sequence":"first","affiliation":[{"name":"State Key Laboratory of Precision Measuring Technology and Instruments, Tianjin University, No. 92, Weijin Road, Nankai District, Tianjin 300072, China"}]},{"given":"Hong","family":"Li","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Precision Measuring Technology and Instruments, Tianjin University, No. 92, Weijin Road, Nankai District, Tianjin 300072, China"}]},{"given":"Zhiyuan","family":"Chen","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Precision Measuring Technology and Instruments, Tianjin University, No. 92, Weijin Road, Nankai District, Tianjin 300072, China"}]},{"given":"Jian","family":"Li","sequence":"additional","affiliation":[{"name":"State Key Laboratory of Precision Measuring Technology and Instruments, Tianjin University, No. 92, Weijin Road, Nankai District, Tianjin 300072, China"}]}],"member":"1968","published-online":{"date-parts":[[2023,6,15]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","unstructured":"Singh, P., and Komodakis, N. (2018, January 22\u201327). Cloud-Gan: Cloud Removal for Sentinel-2 Imagery Using a Cyclic Consistent Generative Adversarial Networks. Proceedings of the IGARSS 2018-2018 IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain.","DOI":"10.1109\/IGARSS.2018.8519033"},{"key":"ref_2","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1109\/TGRS.2021.3131035","article-title":"Cloud Removal in Remote Sensing Images Using Generative Adversarial Networks and SAR-to-Optical Image Translation","volume":"60","author":"Darbaghshahi","year":"2021","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_3","doi-asserted-by":"crossref","unstructured":"Nie, H., Fu, Z., Tang, B.-H., Li, Z., Chen, S., and Wang, L. (2022). A Dual-Generator Translation Network Fusing Texture and Structure Features for SAR and Optical Image Matching. Remote Sens., 14.","DOI":"10.3390\/rs14122946"},{"key":"ref_4","doi-asserted-by":"crossref","unstructured":"Zhou, X., Zhang, C., and Li, S. (2006, January 16\u201319). A Perceptive Uniform Pseudo-Color Coding Method of SAR Images. Proceedings of the 2006 CIE International Conference on Radar, Shanghai, China.","DOI":"10.1109\/ICR.2006.343253"},{"key":"ref_5","doi-asserted-by":"crossref","unstructured":"Li, Z., Liu, J., and Huang, J. (2008, January 12\u201314). Dynamic Range Compression and Pseudo-Color Presentation Based on Retinex for SAR Images. Proceedings of the 2008 International Conference on Computer Science and Software Engineering, Wuhan, China.","DOI":"10.1109\/CSSE.2008.1459"},{"key":"ref_6","doi-asserted-by":"crossref","unstructured":"Deng, Q., Chen, Y., Zhang, W., and Yang, J. (2008, January 27\u201330). Colorization for Polarimetric SAR Image Based on Scattering Mechanisms. Proceedings of the 2008 Congress on Image and Signal Processing, Sanya, China.","DOI":"10.1109\/CISP.2008.366"},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"2256","DOI":"10.1109\/TGRS.2020.3004353","article-title":"Super-Resolution Mapping Based on Spatial\u2013Spectral Correlation for Spectral Imagery","volume":"59","author":"Wang","year":"2020","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_8","doi-asserted-by":"crossref","first-page":"6044","DOI":"10.1109\/TGRS.2020.3010826","article-title":"Target-Constrained Interference-Minimized Band Selection for Hyperspectral Target Detection","volume":"59","author":"Shang","year":"2021","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_9","doi-asserted-by":"crossref","first-page":"88","DOI":"10.1007\/s10915-022-01939-z","article-title":"Scientific Machine Learning Through Physics\u2013Informed Neural Networks: Where We Are and What\u2019s Next","volume":"92","author":"Cuomo","year":"2022","journal-title":"J. Sci. Comput."},{"key":"ref_10","doi-asserted-by":"crossref","first-page":"6136","DOI":"10.1038\/s41467-021-26434-1","article-title":"Physics-Informed Learning of Governing Equations from Scarce Data","volume":"12","author":"Chen","year":"2021","journal-title":"Nat. Commun."},{"key":"ref_11","doi-asserted-by":"crossref","first-page":"139","DOI":"10.1145\/3422622","article-title":"Generative Adversarial Networks","volume":"63","author":"Goodfellow","year":"2020","journal-title":"Commun. ACM"},{"key":"ref_12","unstructured":"Mirza, M., and Osindero, S. (2014). Conditional Generative Adversarial Nets 2014. arXiv."},{"key":"ref_13","doi-asserted-by":"crossref","unstructured":"Isola, P., Zhu, J.-Y., Zhou, T., and Efros, A.A. (2017, January 21\u201326). Image-to-Image Translation with Conditional Adversarial Networks. Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA.","DOI":"10.1109\/CVPR.2017.632"},{"key":"ref_14","doi-asserted-by":"crossref","unstructured":"Zhu, J.-Y., Park, T., Isola, P., and Efros, A.A. (2017, January 22\u201329). Unpaired Image-To-Image Translation Using Cycle-Consistent Adversarial Networks. Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy.","DOI":"10.1109\/ICCV.2017.244"},{"key":"ref_15","unstructured":"Kim, J., Kim, M., Kang, H., and Lee, K.H. (, January 26\u201330April). U-GAT-IT: Unsupervised Generative Attentional Networks with Adaptive Layer-Instance Normalization for Image-to-Image Translation. Proceedings of the International Conference on Learning Representations, Addis Ababa, Ethiopia."},{"key":"ref_16","doi-asserted-by":"crossref","unstructured":"Niu, X., Yang, D., Yang, K., Pan, H., and Dou, Y. (2018, January 21\u201322). Image Translation Between High-Resolution Remote Sensing Optical and SAR Data Using Conditional GAN. Proceedings of the Advances in Multimedia Information Processing\u2014PCM 2018: 19th Pacific-Rim Conference on Multimedia, Hefei, China.","DOI":"10.1007\/978-3-030-00764-5_23"},{"key":"ref_17","doi-asserted-by":"crossref","first-page":"1811","DOI":"10.1109\/JSTARS.2018.2803212","article-title":"Exploring the Potential of Conditional Adversarial Networks for Optical and SAR Image Matching","volume":"11","author":"Merkle","year":"2018","journal-title":"IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens."},{"key":"ref_18","unstructured":"Fu, S., Xu, F., and Jin, Y.-Q. (2019). Translating SAR to Optical Images for Assisted Interpretation 2019. arXiv."},{"key":"ref_19","doi-asserted-by":"crossref","unstructured":"Fuentes Reyes, M., Auer, S., Merkle, N., Henry, C., and Schmitt, M. (2019). SAR-to-Optical Image Translation Based on Conditional Generative Adversarial Networks\u2014Optimization, Opportunities and Limits. Remote Sens., 11.","DOI":"10.3390\/rs11172067"},{"key":"ref_20","doi-asserted-by":"crossref","first-page":"129136","DOI":"10.1109\/ACCESS.2019.2939649","article-title":"SAR-to-Optical Image Translation Using Supervised Cycle-Consistent Adversarial Networks","volume":"7","author":"Wang","year":"2019","journal-title":"IEEE Access"},{"key":"ref_21","doi-asserted-by":"crossref","first-page":"70925","DOI":"10.1109\/ACCESS.2020.2987105","article-title":"Feature-Guided SAR-to-Optical Image Translation","volume":"8","author":"Zhang","year":"2020","journal-title":"IEEE Access"},{"key":"ref_22","doi-asserted-by":"crossref","unstructured":"Zhang, Q., Liu, X., Liu, M., Zou, X., Zhu, L., and Ruan, X. (2021). Comparative Analysis of Edge Information and Polarization on SAR-to-Optical Translation Based on Conditional Generative Adversarial Networks. Remote Sens., 13.","DOI":"10.3390\/rs13010128"},{"key":"ref_23","first-page":"1","article-title":"Multiscale Generative Adversarial Network Based on Wavelet Feature Learning for SAR-to-Optical Image Translation","volume":"60","author":"Li","year":"2022","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_24","first-page":"1","article-title":"Hybrid CGAN: Coupling Global and Local Features for SAR-to-Optical Image Translation","volume":"60","author":"Wang","year":"2022","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"ref_25","doi-asserted-by":"crossref","unstructured":"Guo, J., He, C., Zhang, M., Li, Y., Gao, X., and Song, B. (2021). Edge-Preserving Convolutional Generative Adversarial Networks for SAR-to-Optical Image Translation. Remote Sens., 13.","DOI":"10.3390\/rs13183575"},{"key":"ref_26","doi-asserted-by":"crossref","unstructured":"Pathak, D., Krahenbuhl, P., Donahue, J., Darrell, T., and Efros, A.A. (2016, January 27\u201330). Context Encoders: Feature Learning by Inpainting. Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA.","DOI":"10.1109\/CVPR.2016.278"},{"key":"ref_27","doi-asserted-by":"crossref","unstructured":"Leibe, B., Matas, J., Sebe, N., and Welling, M. (2016, January 11\u201314). Generative Image Modeling Using Style and Structure Adversarial Networks. Proceedings of the Computer Vision\u2014ECCV 2016, Amsterdam, The Netherlands.","DOI":"10.1007\/978-3-319-46454-1"},{"key":"ref_28","doi-asserted-by":"crossref","unstructured":"Leibe, B., Matas, J., Sebe, N., and Welling, M. (2016, January 11\u201314). Learning Temporal Transformations from Time-Lapse Videos. Proceedings of the Computer Vision\u2014ECCV 2016, Amsterdam, The Netherlands.","DOI":"10.1007\/978-3-319-46466-4"},{"key":"ref_29","doi-asserted-by":"crossref","unstructured":"Yoo, D., Kim, N., Park, S., Paek, A.S., and Kweon, I.S. (2016, January 11\u201314). Pixel-Level Domain Transfer. Proceedings of the Computer Vision\u2014ECCV 2016, Amsterdam, The Netherlands.","DOI":"10.1007\/978-3-319-46484-8_31"},{"key":"ref_30","doi-asserted-by":"crossref","first-page":"504","DOI":"10.1126\/science.1127647","article-title":"Reducing the Dimensionality of Data with Neural Networks","volume":"313","author":"Hinton","year":"2006","journal-title":"Science"},{"key":"ref_31","doi-asserted-by":"crossref","unstructured":"Johnson, J., Alahi, A., and Fei-Fei, L. (2016, January 11\u201314). Perceptual Losses for Real-Time Style Transfer and Super-Resolution. Proceedings of the Computer Vision\u2014ECCV 2016, Amsterdam, The Netherlands.","DOI":"10.1007\/978-3-319-46475-6_43"},{"key":"ref_32","doi-asserted-by":"crossref","unstructured":"Ronneberger, O., Fischer, P., and Brox, T. (2015, January 5\u20139). U-Net: Convolutional Networks for Biomedical Image Segmentation. Proceedings of the Medical Image Computing and Computer-Assisted Intervention\u2014MICCAI 2015, Munich, Germany.","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"ref_33","doi-asserted-by":"crossref","unstructured":"Gatys, L.A., Ecker, A.S., and Bethge, M. (2015). A Neural Algorithm of Artistic Style 2015. arXiv.","DOI":"10.1167\/16.12.326"},{"key":"ref_34","doi-asserted-by":"crossref","unstructured":"Wang, T.-C., Liu, M.-Y., Zhu, J.-Y., Tao, A., Kautz, J., and Catanzaro, B. (2018, January 18\u201322). High-Resolution Image Synthesis and Semantic Manipulation With Conditional GANs. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA.","DOI":"10.1109\/CVPR.2018.00917"},{"key":"ref_35","doi-asserted-by":"crossref","first-page":"1856","DOI":"10.1109\/TMI.2019.2959609","article-title":"UNet++: Redesigning Skip Connections to Exploit Multiscale Features in Image Segmentation","volume":"39","author":"Zhou","year":"2020","journal-title":"IEEE Trans. Med. Imaging"},{"key":"ref_36","doi-asserted-by":"crossref","first-page":"107453","DOI":"10.1016\/j.patcog.2020.107453","article-title":"Generative Adversarial Classifier for Handwriting Characters Super-Resolution","volume":"107","author":"Qian","year":"2020","journal-title":"Pattern Recognit."},{"key":"ref_37","doi-asserted-by":"crossref","first-page":"107249","DOI":"10.1016\/j.patcog.2020.107249","article-title":"Identity-Aware CycleGAN for Face Photo-Sketch Synthesis and Recognition","volume":"102","author":"Fang","year":"2020","journal-title":"Pattern Recognit."},{"key":"ref_38","doi-asserted-by":"crossref","first-page":"107085","DOI":"10.1016\/j.patcog.2019.107085","article-title":"Semi-Supervised Cross-Modal Image Generation with Generative Adversarial Networks","volume":"100","author":"Li","year":"2020","journal-title":"Pattern Recognit."},{"key":"ref_39","doi-asserted-by":"crossref","first-page":"570","DOI":"10.1016\/j.patcog.2019.05.017","article-title":"Toward Learning a Unified Many-to-Many Mapping for Diverse Image Translation","volume":"93","author":"Xu","year":"2019","journal-title":"Pattern Recognit."},{"key":"ref_40","doi-asserted-by":"crossref","first-page":"107097","DOI":"10.1016\/j.patcog.2019.107097","article-title":"Disentangled Representation Learning and Residual GAN for Age-Invariant Face Verification","volume":"100","author":"Zhao","year":"2020","journal-title":"Pattern Recognit."},{"key":"ref_41","doi-asserted-by":"crossref","first-page":"107350","DOI":"10.1016\/j.patcog.2020.107350","article-title":"GAN-Based Person Search via Deep Complementary Classifier with Center-Constrained Triplet Loss","volume":"104","author":"Yao","year":"2020","journal-title":"Pattern Recognit."},{"key":"ref_42","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., and Efros, A.A. (2016, January 11\u201314). Colorful Image Colorization. Proceedings of the Computer Vision\u2014ECCV 2016, Amsterdam, The Netherlands.","DOI":"10.1007\/978-3-319-46487-9_40"},{"key":"ref_43","doi-asserted-by":"crossref","first-page":"108208","DOI":"10.1016\/j.patcog.2021.108208","article-title":"SAR-to-Optical Image Translation Based on Improved CGAN","volume":"121","author":"Yang","year":"2022","journal-title":"Pattern Recognit."},{"key":"ref_44","doi-asserted-by":"crossref","first-page":"60338","DOI":"10.1109\/ACCESS.2020.2977103","article-title":"A SAR-to-Optical Image Translation Method Based on Conditional Generation Adversarial Network (CGAN)","volume":"8","author":"Li","year":"2020","journal-title":"IEEE Access"},{"key":"ref_45","doi-asserted-by":"crossref","first-page":"600","DOI":"10.1109\/TIP.2003.819861","article-title":"Image Quality Assessment: From Error Visibility to Structural Similarity","volume":"13","author":"Wang","year":"2004","journal-title":"IEEE Trans. Image Process."},{"key":"ref_46","doi-asserted-by":"crossref","first-page":"47","DOI":"10.1109\/TCI.2016.2644865","article-title":"Loss Functions for Image Restoration With Neural Networks","volume":"3","author":"Zhao","year":"2017","journal-title":"IEEE Trans. Comput. Imaging"},{"key":"ref_47","unstructured":"Wang, Z., Simoncelli, E.P., and Bovik, A.C. (2003, January 9\u201312). Multiscale Structural Similarity for Image Quality Assessment. Proceedings of the Thrity-Seventh Asilomar Conference on Signals, Systems & Computers, Pacific Grove, CA, USA."},{"key":"ref_48","doi-asserted-by":"crossref","unstructured":"Schmitt, M., Hughes, L.H., and Zhu, X.X. (2018). The SEN1-2 Dataset for Deep Learning in SAR-Optical Data Fusion 2018. arXiv.","DOI":"10.5194\/isprs-annals-IV-1-141-2018"},{"key":"ref_49","doi-asserted-by":"crossref","unstructured":"Mao, X., Li, Q., Xie, H., Lau, R.Y.K., Wang, Z., and Paul Smolley, S. (2017, January 22\u201329). Least Squares Generative Adversarial Networks. Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy.","DOI":"10.1109\/ICCV.2017.304"}],"container-title":["Remote Sensing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2072-4292\/15\/12\/3125\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T19:55:17Z","timestamp":1760126117000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2072-4292\/15\/12\/3125"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,6,15]]},"references-count":49,"journal-issue":{"issue":"12","published-online":{"date-parts":[[2023,6]]}},"alternative-id":["rs15123125"],"URL":"https:\/\/doi.org\/10.3390\/rs15123125","relation":{},"ISSN":["2072-4292"],"issn-type":[{"value":"2072-4292","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,6,15]]}}}