{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,3]],"date-time":"2026-02-03T17:49:25Z","timestamp":1770140965432,"version":"3.49.0"},"reference-count":35,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2024,6,17]],"date-time":"2024-06-17T00:00:00Z","timestamp":1718582400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,6,17]],"date-time":"2024-06-17T00:00:00Z","timestamp":1718582400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"Hebei Graduate Innovation Funding Project of China","award":["CXZZBS2022148"],"award-info":[{"award-number":["CXZZBS2022148"]}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61971373"],"award-info":[{"award-number":["61971373"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100003787","name":"Natural Science Foundation of Hebei Province","doi-asserted-by":"publisher","award":["F2023105001"],"award-info":[{"award-number":["F2023105001"]}],"id":[{"id":"10.13039\/501100003787","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2025,2]]},"DOI":"10.1007\/s00371-024-03519-8","type":"journal-article","created":{"date-parts":[[2024,6,17]],"date-time":"2024-06-17T15:02:33Z","timestamp":1718636553000},"page":"1999-2010","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":6,"title":["MDSCN: multiscale depthwise separable convolutional network for underwater graphics restoration"],"prefix":"10.1007","volume":"41","author":[{"given":"Shiyu","family":"Li","sequence":"first","affiliation":[]},{"given":"Zehao","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Meijing","family":"Gao","sequence":"additional","affiliation":[]},{"given":"Yang","family":"Bai","sequence":"additional","affiliation":[]},{"given":"Haozheng","family":"Yin","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,6,17]]},"reference":[{"key":"3519_CR1","doi-asserted-by":"publisher","first-page":"731","DOI":"10.3390\/rs13040731","volume":"13","author":"B Chen","year":"2021","unstructured":"Chen, B., Xia, M., Huang, J.: MFANet: a multi-level feature aggregation network for semantic segmentation of land cover. Remote Sens. 13, 731 (2021)","journal-title":"Remote Sens."},{"key":"3519_CR2","doi-asserted-by":"publisher","first-page":"2022","DOI":"10.1080\/01431161.2020.1849852","volume":"42","author":"M Xia","year":"2021","unstructured":"Xia, M., Wang, T., Zhang, Y., et al.: Cloud\/shadow segmentation based on global attention feature fusion residual network for remote sensing imagery. Remote Sens. 42, 2022\u20132045 (2021)","journal-title":"Remote Sens."},{"key":"3519_CR3","first-page":"1377","volume":"36","author":"RM Cong","year":"2020","unstructured":"Cong, R.M., Zhang, Y.M., Zhang, C., et al.: Research progress of deep learning driven underwater image enhancement and restoration. Signal Process. 36, 1377\u20131389 (2020)","journal-title":"Signal Process."},{"issue":"2","key":"3519_CR4","doi-asserted-by":"publisher","first-page":"241","DOI":"10.3390\/jmse10020241","volume":"10","author":"K Hu","year":"2022","unstructured":"Hu, K., Weng, C., Zhang, Y., et al.: An overview of underwater vision enhancement: from traditional methods to recent deep learning. J. Mar. Sci. Eng. 10(2), 241 (2022)","journal-title":"J. Mar. Sci. Eng."},{"key":"3519_CR5","doi-asserted-by":"crossref","unstructured":"Garcia, R., Nicosevici, T., Cufi, C.: On the way to solve lighting problems in underwater imaging. In: Proceedings of MTS\/IEEE Oceans, pp. 1018\u20131024 (2002)","DOI":"10.1109\/OCEANS.2002.1192107"},{"key":"3519_CR6","doi-asserted-by":"crossref","unstructured":"Zuiderveld, K.: Contrast limited adaptive histogram equalization. In: Graphics Gems IV, pp. 474\u2013485. Academic Press (1994)","DOI":"10.1016\/B978-0-12-336156-1.50061-6"},{"issue":"12","key":"3519_CR7","doi-asserted-by":"publisher","first-page":"2341","DOI":"10.1109\/TPAMI.2010.168","volume":"33","author":"K He","year":"2011","unstructured":"He, K., Sun, J., Tang, X.: Single image haze removal using dark channel prior. IEEE Trans. Pattern Anal. Mach. Intell. 33(12), 2341\u20132353 (2011)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"3519_CR8","doi-asserted-by":"crossref","unstructured":"Drews Jr, P., Nascimento, E., Moraes, F., et al.: Transmission estimation in underwater single images. In: IEEE International Conference on Computer Vision Workshops, pp. 825\u2013830 (2013)","DOI":"10.1109\/ICCVW.2013.113"},{"issue":"2","key":"3519_CR9","doi-asserted-by":"publisher","first-page":"239","DOI":"10.1007\/s11554-014-0399-9","volume":"10","author":"B Jiang","year":"2015","unstructured":"Jiang, B., Woodell, G.A., Jobson, D.J.: Novel multi-scale retinex with color restoration on graphics processing unit. J. Real-time Image Process. 10(2), 239\u2013253 (2015)","journal-title":"J. Real-time Image Process."},{"key":"3519_CR10","unstructured":"https:\/\/github.com\/AomanHao\/Matlab-Image-Dehaze-Enhance"},{"issue":"8","key":"3519_CR11","doi-asserted-by":"publisher","first-page":"4499","DOI":"10.1109\/TNNLS.2021.3116209","volume":"34","author":"Z Xie","year":"2023","unstructured":"Xie, Z., Zhang, W., Sheng, B., et al.: BaGFN: broad attentive graph fusion network for high-order feature interactions. IEEE Trans. Neural Netw. Learn. Syst. 34(8), 4499\u20134513 (2023)","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"3519_CR12","doi-asserted-by":"publisher","first-page":"50","DOI":"10.1109\/TMM.2021.3120873","volume":"25","author":"X Lin","year":"2023","unstructured":"Lin, X., Sun, S., Huang, W., et al.: EAPT: efficient attention pyramid transformer for image processing. IEEE Trans. Multimed. 25, 50\u201361 (2023)","journal-title":"IEEE Trans. Multimed."},{"key":"3519_CR13","doi-asserted-by":"crossref","unstructured":"Perez, J., Attanasio, A. C., Nechyporenko, N., et al.: A deep learning approach for underwater image enhancement. In: Biomedical Applications Based on Natural and Artificial Computing, pp.183\u2013192 (2017)","DOI":"10.1007\/978-3-319-59773-7_19"},{"key":"3519_CR14","doi-asserted-by":"crossref","unstructured":"Wang, Y., Zhang, J., Cao, Y., et al.: A deep CNN method for underwater image enhancement. In: 2017 IEEE International Conference on Image Processing (ICIP), pp. 1382\u20131386. Beijing, China (2017)","DOI":"10.1109\/ICIP.2017.8296508"},{"issue":"3","key":"3519_CR15","doi-asserted-by":"publisher","first-page":"323","DOI":"10.1109\/LSP.2018.2792050","volume":"25","author":"C Li","year":"2018","unstructured":"Li, C., Guo, G., Guo, C.: emerging from water: underwater image color correction based on weakly supervised color transfer. IEEE Signal Process. Lett. 25(3), 323\u2013327 (2018)","journal-title":"IEEE Signal Process. Lett."},{"key":"3519_CR16","first-page":"1","volume":"60","author":"R Chen","year":"2022","unstructured":"Chen, R., Cai, Z., Cao, W.: MFFN: an underwater sensing scene image enhancement method based on multiscale feature fusion network. IEEE Trans. Geosci. Remote Sens. 60, 1\u201312 (2022)","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"issue":"1","key":"3519_CR17","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3511021","volume":"19","author":"P Sharma","year":"2023","unstructured":"Sharma, P., Bisht, I., Sur, A.: Wavelength-based attributed deep neural network for underwater image restoration. ACM Trans. Multimed. Comput. Commun. Appl. 19(1), 1\u201323 (2023)","journal-title":"ACM Trans. Multimed. Comput. Commun. Appl."},{"issue":"1","key":"3519_CR18","first-page":"387","volume":"3","author":"J Li","year":"2018","unstructured":"Li, J., Skinner, K.A., Eustice, R.M., et al.: WaterGAN: unsupervised generative network to enable real-time color correction of monocular underwater images. IEEE Robot. Autom. Lett. 3(1), 387\u2013394 (2018)","journal-title":"IEEE Robot. Autom. Lett."},{"key":"3519_CR19","doi-asserted-by":"crossref","unstructured":"Fabbri, C., Jahidul, I. M., Sattar, J.: Enhancing underwater imagery using generative adversarial networks. In: IEEE International Conference on Robotics and Automation, pp. 7159\u20137165 (2018)","DOI":"10.1109\/ICRA.2018.8460552"},{"key":"3519_CR20","unstructured":"Islam, M. J., Luo, P., Sattar, J.: simultaneous enhancement and super-resolution of underwater imagery for improved visual perception robotics: Science and Systems XVI. Preprint at https:\/\/arxiv.org\/abs\/2002.01155v1 (2020)"},{"key":"3519_CR21","doi-asserted-by":"publisher","first-page":"54241","DOI":"10.1109\/ACCESS.2018.2870854","volume":"6","author":"N Li","year":"2019","unstructured":"Li, N., Zheng, Z., Zhang, S., et al.: The synthesis of unpaired underwater images using a multistyle generative adversarial network. IEEE Access 6, 54241\u201354257 (2019)","journal-title":"IEEE Access"},{"key":"3519_CR22","unstructured":"Wang, N., Zhou, Y., Han, F., et al.: UWGAN: underwater gan for real-world underwater color restoration and dehazing. arXiv Preprint, arXiv:1912.10269 (2019)"},{"key":"3519_CR23","doi-asserted-by":"publisher","first-page":"6816","DOI":"10.1109\/TIP.2022.3216208","volume":"31","author":"Q Qi","year":"2022","unstructured":"Qi, Q., Li, K., Zheng, H., et al.: SGUIE-net: semantic attention guided underwater image enhancement with multi-scale perception. IEEE Trans. Image Process. 31, 6816\u20136830 (2022)","journal-title":"IEEE Trans. Image Process."},{"key":"3519_CR24","doi-asserted-by":"publisher","first-page":"4922","DOI":"10.1109\/TIP.2022.3190209","volume":"31","author":"R Liu","year":"2022","unstructured":"Liu, R., Jiang, Z., Yang, S., et al.: Twin adversarial contrastive learning for underwater image enhancement and beyond. IEEE Trans. Image Process. 31, 4922\u20134936 (2022)","journal-title":"IEEE Trans. Image Process."},{"issue":"10","key":"3519_CR25","doi-asserted-by":"publisher","first-page":"6584","DOI":"10.1109\/TCSVT.2022.3174817","volume":"30","author":"Z Jiang","year":"2022","unstructured":"Jiang, Z., Li, Z., Yang, S., et al.: Target oriented perceptual adversarial fusion network for underwater image enhancement. IEEE Trans. Circuits Syst. Video Technol. 30(10), 6584\u20136598 (2022)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"issue":"10","key":"3519_CR26","doi-asserted-by":"publisher","first-page":"7719","DOI":"10.1109\/TNNLS.2022.3146004","volume":"34","author":"Y Zhou","year":"2023","unstructured":"Zhou, Y., Chen, Z., Li, P., et al.: FSAD-Net feedback spatial attention dehazing network. IEEE Trans. Neural Netw. Learn. Syst. 34(10), 7719\u20137733 (2023)","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"3519_CR27","unstructured":"Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint, arXiv:1609.04747 (2016)"},{"key":"3519_CR28","unstructured":"Loshchilov, I., Hutter, F.: Decoupled weight decay regularization. arXiv preprint, arXiv:1711.05101 (2017)"},{"key":"3519_CR29","unstructured":"Zhang, M., Lucas, J., Ba, J., et al.: Lookahead optimizer: k steps forward, 1 step back. Advances in neural information processing systems, vol. 32. Preprint at arXiv:1907.08610v1 (2019)"},{"key":"3519_CR30","doi-asserted-by":"crossref","unstructured":"Yong, H., Huang, J., Hua, X., et al.: Gradient centralization: a new optimization technique for deep neural networks. arXiv preprint,\narXiv:2004.01461 (2020)","DOI":"10.1007\/978-3-030-58452-8_37"},{"key":"3519_CR31","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1016\/j.neunet.2017.12.012","volume":"107","author":"S Elfwing","year":"2018","unstructured":"Elfwing, S., Uchibe, E., Doya, K.: Sigmoid-weighted linear units for neural network function approximation in reinforcement learning. Neural Netw. 107, 3\u201311 (2018)","journal-title":"Neural Netw."},{"key":"3519_CR32","unstructured":"https:\/\/gitee.com\/raining-li\/underwater-image-processing\/tree\/master"},{"key":"3519_CR33","doi-asserted-by":"crossref","unstructured":"Song, W., Wang, Y., Huang, D., et al.: A rapid scene depth estimation model based on underwater light attenuation prior for underwater image restoration. pp. 678\u2013688. Springer International Publishing (2018)","DOI":"10.1007\/978-3-030-00776-8_62"},{"key":"3519_CR34","doi-asserted-by":"crossref","unstructured":"Zhu J.Y., Park, T., Isola, P. et al.: Unpaired image-to-image translation using cycle-consistent adversarial networks. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2223\u20132232 (2017)","DOI":"10.1109\/ICCV.2017.244"},{"key":"3519_CR35","unstructured":"https:\/\/li-chongyi.github.io\/proj_benchmark.html"}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-024-03519-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-024-03519-8\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-024-03519-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,2,12]],"date-time":"2025-02-12T14:53:04Z","timestamp":1739371984000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-024-03519-8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,6,17]]},"references-count":35,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2025,2]]}},"alternative-id":["3519"],"URL":"https:\/\/doi.org\/10.1007\/s00371-024-03519-8","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"value":"0178-2789","type":"print"},{"value":"1432-2315","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,6,17]]},"assertion":[{"value":"27 May 2024","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"17 June 2024","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}