{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,26]],"date-time":"2026-02-26T15:33:26Z","timestamp":1772120006873,"version":"3.50.1"},"reference-count":55,"publisher":"Springer Science and Business Media LLC","issue":"6","license":[{"start":{"date-parts":[[2025,4,2]],"date-time":"2025-04-02T00:00:00Z","timestamp":1743552000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,4,2]],"date-time":"2025-04-02T00:00:00Z","timestamp":1743552000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"the National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62076247"],"award-info":[{"award-number":["62076247"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["SIViP"],"published-print":{"date-parts":[[2025,6]]},"DOI":"10.1007\/s11760-025-04012-y","type":"journal-article","created":{"date-parts":[[2025,4,4]],"date-time":"2025-04-04T08:16:51Z","timestamp":1743754611000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Hybrid sparse and dense attentions of similar regions for image denoising"],"prefix":"10.1007","volume":"19","author":[{"given":"Dai-Qiang","family":"Chen","sequence":"first","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,4,2]]},"reference":[{"issue":"12","key":"4012_CR1","doi-asserted-by":"publisher","first-page":"3736","DOI":"10.1109\/TIP.2006.881969","volume":"15","author":"M Elad","year":"2006","unstructured":"Elad, M., Aharon, M.: Image denoising via sparse and Redundant representations over Learned Dictionaries. IEEE Trans. Image Process. 15(12), 3736\u20133745 (2006)","journal-title":"IEEE Trans. Image Process."},{"issue":"8","key":"4012_CR2","doi-asserted-by":"publisher","first-page":"2080","DOI":"10.1109\/TIP.2007.901238","volume":"16","author":"K Dabov","year":"2007","unstructured":"Dabov, K., Foi, A., Katkovnik, V., et al.: Image denoising by sparse 3-D transform-domain collaborative filtering. IEEE Trans. Image Process. 16(8), 2080\u20132095 (2007)","journal-title":"IEEE Trans. Image Process."},{"issue":"7","key":"4012_CR3","first-page":"3142","volume":"26","author":"Z Kai","year":"2016","unstructured":"Kai, Z., Zuo, W., Chen, Y., et al.: Beyond a Gaussian denoiser: residual learning of deep CNN for image denoising. IEEE Trans. Image Process. 26(7), 3142\u20133155 (2016)","journal-title":"IEEE Trans. Image Process."},{"issue":"9","key":"4012_CR4","doi-asserted-by":"publisher","first-page":"4608","DOI":"10.1109\/TIP.2018.2839891","volume":"27","author":"K Zhang","year":"2018","unstructured":"Zhang, K., Zuo, W., Zhang, L.: FFDNet: toward a fast and flexible solution for CNN based image denoising. IEEE Trans. Image Process. 27(9), 4608\u20134622 (2018)","journal-title":"IEEE Trans. Image Process."},{"issue":"7","key":"4012_CR5","doi-asserted-by":"publisher","first-page":"2480","DOI":"10.1109\/TPAMI.2020.2968521","volume":"43","author":"Y Zhang","year":"2021","unstructured":"Zhang, Y., Tian, Y., Kong, Y., et al.: Residual dense network for image restoration. IEEE Trans. Pattern Anal. Mach. Intell. 43(7), 2480\u20132495 (2021)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"1","key":"4012_CR6","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/S1063-5203(02)00511-0","volume":"14","author":"I Daubechies","year":"2003","unstructured":"Daubechies, I., Han, B., Ron, A., et al.: Framelets: MRA-based constructions of wavelet frames. Appl. Comput. Harmon. Anal. 14(1), 1\u201346 (2003)","journal-title":"Appl. Comput. Harmon. Anal."},{"issue":"8","key":"4012_CR7","doi-asserted-by":"publisher","first-page":"3336","DOI":"10.1109\/TIP.2014.2323127","volume":"23","author":"J Zhang","year":"2014","unstructured":"Zhang, J., Zhao, D., Gao, W.: Group-based sparse representation for image restoration. IEEE Trans. Image Process. 23(8), 3336\u20133351 (2014)","journal-title":"IEEE Trans. Image Process."},{"issue":"2","key":"4012_CR8","doi-asserted-by":"publisher","first-page":"461","DOI":"10.1007\/s10915-016-0205-x","volume":"69","author":"DQ Chen","year":"2016","unstructured":"Chen, D.Q.: Data-driven tight frame learning scheme based on local and non-local sparsity with application to image recovery. J. Sci. Comput. 69(2), 461\u2013486 (2016)","journal-title":"J. Sci. Comput."},{"key":"4012_CR9","volume-title":"Learning a Deep Convolutional Network for Image Super-Resolution","author":"D Chao","year":"2014","unstructured":"Chao, D., Chen, C.L., He, K., et al.: Learning a Deep Convolutional Network for Image Super-Resolution. Springer International Publishing, ECCV (2014)"},{"key":"4012_CR10","doi-asserted-by":"crossref","unstructured":"Lim, B., Son, S., Kim, H., et al.: Enhanced Deep Residual Networks for Single Image Super-Resolution[C]. IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) (2017)","DOI":"10.1109\/CVPRW.2017.151"},{"key":"4012_CR11","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Li, K., Li, K., Wang, L., Zhong, B., Fu, Y.: Image super-resolution using very deep residual channel attention networks. In: ECCV (2018)","DOI":"10.1007\/978-3-030-01234-2_18"},{"key":"4012_CR12","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., et al.: Identity mappings in deep residual networks. European Conference on Computer Vision. Springer, Cham, 630\u2013645 (2016)","DOI":"10.1007\/978-3-319-46493-0_38"},{"key":"4012_CR13","doi-asserted-by":"publisher","first-page":"153","DOI":"10.1016\/j.neucom.2019.09.031","volume":"376","author":"W Liu","year":"2020","unstructured":"Liu, W., Zhao, Y., Liu, M., et al.: Triple-adjacent-frame generative network for blind video motion deblurring. Neurocomputing 376, 153\u2013165 (2020)","journal-title":"Neurocomputing"},{"key":"4012_CR14","doi-asserted-by":"publisher","DOI":"10.1016\/j.displa.2023.102577","volume":"80","author":"W Yi","year":"2023","unstructured":"Yi, W., Dong, L., Liu, M., et al.: Frequency-guidance collaborative triple-branch network for single image dehazing. Displays 80, 102577 (2023)","journal-title":"Displays"},{"issue":"28","key":"4012_CR15","doi-asserted-by":"publisher","first-page":"71619","DOI":"10.1007\/s11042-024-18502-7","volume":"83","author":"W Yi","year":"2024","unstructured":"Yi, W., Dong, L., Liu, M., et al.: SID-Net: single image dehazing network using adversarial and contrastive learning. Multimedia Tools Appl. 83(28), 71619\u201371638 (2024)","journal-title":"Multimedia Tools Appl."},{"key":"4012_CR16","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., Polosukhin, I.: Attention is all you need. In: NIPS (2017)"},{"key":"4012_CR17","doi-asserted-by":"crossref","unstructured":"Wang, X., Girshick, R., Gupta, A., et al.: Non-local Neural Networks. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00813"},{"key":"4012_CR18","doi-asserted-by":"crossref","unstructured":"Fu, J., Liu, J., Tian, H., et al.: Dual Attention Network for Scene Segmentation. In: IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2020)","DOI":"10.1109\/CVPR.2019.00326"},{"key":"4012_CR19","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2023.126494","volume":"551","author":"W Yi","year":"2023","unstructured":"Yi, W., Dong, L., Liu, M., et al.: Semi-supervised progressive dehazing network using unlabeled contrastive guidance. Neurocomputing 551, 126494 (2023)","journal-title":"Neurocomputing"},{"key":"4012_CR20","doi-asserted-by":"crossref","unstructured":"Liu, Z., Lin, Y., Cao, Y., Hu, H., Wei, Y., Zhang, Z., Lin, S., Guo, B.: Swin transformer: Hierarchical vision transformer using shifted windows. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"4012_CR21","doi-asserted-by":"crossref","unstructured":"Hassani, A., Walton, S., Li, J., et al.: Neighborhood attention transformer. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition(CVPR), 6185\u20136194 (2023)","DOI":"10.1109\/CVPR52729.2023.00599"},{"key":"4012_CR22","unstructured":"Kitaev, N., Kaiser, Levskaya, A.: Reformer: The efficient transformer. In: International Conference on Learning Representations(ICLR) (2019)"},{"key":"4012_CR23","unstructured":"Yu, Q., Xia, Y., Bai, Y., Lu, Y., Yuille, A.L., Shen, W.: Glance-and-gaze vision transformer. In: NeurIPS (2021)"},{"key":"4012_CR24","doi-asserted-by":"crossref","unstructured":"Tai, Y., Yang, J., Liu, X., et al.: MemNet: a persistent memory network for image restoration. In: IEEE International Conference on Computer Vision (2017)","DOI":"10.1109\/ICCV.2017.486"},{"issue":"15","key":"4012_CR25","doi-asserted-by":"publisher","first-page":"17449","DOI":"10.1007\/s10489-022-03157-4","volume":"52","author":"W Yi","year":"2022","unstructured":"Yi, W., Dong, L., Liu, M., et al.: Gated residual feature attention network for real-time Dehazing. Appl. Intell. 52(15), 17449\u201317464 (2022)","journal-title":"Appl. Intell."},{"issue":"12","key":"4012_CR26","doi-asserted-by":"publisher","first-page":"9423","DOI":"10.1109\/TPAMI.2024.3419007","volume":"46","author":"Y Cui","year":"2024","unstructured":"Cui, Y., Ren, W., Cao, X., et al.: Revitalizing convolutional network for image restoration. IEEE Trans. Pattern Anal. Mach. Intell. 46(12), 9423\u20139438 (2024)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"4012_CR27","doi-asserted-by":"publisher","DOI":"10.1016\/j.neunet.2024.106165","volume":"173","author":"W Yi","year":"2024","unstructured":"Yi, W., Dong, L., Liu, M., Hui, M., Kong, L., Zhao, Y.: Priors-assisted dehazing network with attention supervision and detail preservation. Neural Netw. 173, 106165 (2024)","journal-title":"Neural Netw."},{"key":"4012_CR28","doi-asserted-by":"crossref","unstructured":"Dai, T., Cai, J., Zhang, Y., et al.: Second-order attention network for single image super-resolution. In: IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2019)","DOI":"10.1109\/CVPR.2019.01132"},{"key":"4012_CR29","unstructured":"Zhang, Y., Li, K., Li, K., et al.: Residual non-local attention networks for image restoration. In: ICLR (2019)"},{"key":"4012_CR30","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2023.121130","volume":"235","author":"W Yi","year":"2024","unstructured":"Yi, W., Dong, L., Liu, M., et al.: Towards compact single image dehazing via task-related contrastive network. Expert Syst. Appl. 235, 121130 (2024)","journal-title":"Expert Syst. Appl."},{"key":"4012_CR31","doi-asserted-by":"publisher","first-page":"16771","DOI":"10.1007\/s00521-022-07319-w","volume":"34","author":"W Yi","year":"2022","unstructured":"Yi, W., Dong, L., Liu, M., et al.: DCNet: dual-cascade network for single image dehazing. Neural Comput. Appl. 34, 16771\u201316783 (2022)","journal-title":"Neural Comput. Appl."},{"key":"4012_CR32","doi-asserted-by":"crossref","unstructured":"Cui, Y., Ren, W., Cao, X., Knoll, A.: Focal network for image restoration. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 13001\u201313011 (2023)","DOI":"10.1109\/ICCV51070.2023.01195"},{"key":"4012_CR33","doi-asserted-by":"crossref","unstructured":"Cui, Y., Ren, W., Knoll, A.: Omni-Kernel network for image restoration. In: Proceedings of the AAAI Conference on Artificial Intelligence 38(2), 1426\u20131434 (2024)","DOI":"10.1609\/aaai.v38i2.27907"},{"key":"4012_CR34","doi-asserted-by":"publisher","first-page":"2293","DOI":"10.1007\/s00371-023-02917-8","volume":"40","author":"W Yi","year":"2024","unstructured":"Yi, W., Dong, L., Liu, M., et al.: MFAF-Net: image dehazing with multi-level features and adaptive fusion[J]. Vis. Comput. 40, 2293\u20132307 (2024)","journal-title":"Vis. Comput."},{"key":"4012_CR35","doi-asserted-by":"crossref","unstructured":"Cui, Y., Tao, Y., Bing, Z., et al.: Selective frequency network for image restoration. In: ICLR (2023)","DOI":"10.24963\/ijcai.2024\/80"},{"key":"4012_CR36","doi-asserted-by":"crossref","unstructured":"Dong, L., Xu, S., Xu, B.: Speech-transformer: A no-recurrence sequence-to-sequence model for speech recognition. In: ICASSP (2018)","DOI":"10.1109\/ICASSP.2018.8462506"},{"key":"4012_CR37","doi-asserted-by":"crossref","unstructured":"Al-Rfou, R., Choe, D., Constant, N., et al.: Character-level language modeling with deeper self-attention. In: Proceedings of the AAAI Conference on Artificial Intelligence. 33(01), 3159\u20133166 (2019)","DOI":"10.1609\/aaai.v33i01.33013159"},{"key":"4012_CR38","unstructured":"Buades, A., Coll, B., Morel, J.M.: A non-local algorithm for image denoising. IEEE Comput. Vis. Pattern Recognit. (2005)"},{"issue":"12","key":"4012_CR39","doi-asserted-by":"publisher","first-page":"3207","DOI":"10.1007\/s11263-023-01843-5","volume":"131","author":"Y Mei","year":"2023","unstructured":"Mei, Y., Fan, Y., Zhang, Y., et al.: Pyramid attention network for image restoration[J]. Int. J. Comput. Vision 131(12), 3207\u20133225 (2023)","journal-title":"Int. J. Comput. Vision"},{"key":"4012_CR40","unstructured":"Ramachandran, P., Parmar, N., Vaswani, A., Bello, I., Levskaya, A., Shlens, J.: Stand-alone self-attention in vision models. arXiv:1906.05909 (2019)"},{"key":"4012_CR41","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., et al.: An image is worth 16x16 words: Transformers for image recognition at scale. In: ICLR (2021)"},{"key":"4012_CR42","unstructured":"Ho, J., Kalchbrenner, N., Weissenborn, D., Salimans, T.: Axial attention in multidimensional transformers. CoRR (2019)"},{"issue":"6","key":"4012_CR43","doi-asserted-by":"publisher","first-page":"6896","DOI":"10.1109\/TPAMI.2020.3007032","volume":"45","author":"Z Huang","year":"2023","unstructured":"Huang, Z., Wang, X., Wei, Y., Huang, L., Shi, H., Liu, W., Huang, T.S.: Ccnet: Criss-cross attention for semantic segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 45(6), 6896\u20136908 (2023)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"4012_CR44","doi-asserted-by":"crossref","unstructured":"Dong, X., Bao, J., Chen, D., et al.: CSWin transformer: a general vision transformer backbone with cross-shaped windows. In: IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2022)","DOI":"10.1109\/CVPR52688.2022.01181"},{"issue":"9","key":"4012_CR45","doi-asserted-by":"publisher","first-page":"5735","DOI":"10.1109\/LRA.2023.3300254","volume":"8","author":"Y Cui","year":"2023","unstructured":"Cui, Y., Knoll, A.: PSNet: towards efficient image restoration with self-attention. IEEE Robot. Autom. Lett. 8(9), 5735\u20135742 (2023)","journal-title":"IEEE Robot. Autom. Lett."},{"key":"4012_CR46","doi-asserted-by":"crossref","unstructured":"Tu, Z., Talebi, H., Zhang, H., Yang, F., Milanfar, P., Bovik, A., Li, Y.: Maxvit: Multi-axis vision transformer. In: ECCV (2022)","DOI":"10.1007\/978-3-031-20053-3_27"},{"key":"4012_CR47","doi-asserted-by":"publisher","first-page":"47","DOI":"10.1007\/s11760-023-02708-7","volume":"18","author":"DQ Chen","year":"2024","unstructured":"Chen, D.Q.: Down-scale simplified non-local attention networks with application to image denoising. SIViP 18, 47\u201354 (2024)","journal-title":"SIViP"},{"key":"4012_CR48","unstructured":"Timofte, R., Gu, S., Wu, J., et al.: Ntire 2018 challenge on single image super-resolution: Methods and results. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, 852\u2013863 (2018)"},{"key":"4012_CR49","doi-asserted-by":"crossref","unstructured":"Huang, J.B., Singh, A., Ahuja, N.: Single image super-resolution from transformed self-exemplars. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 5197\u20135206 (2015)","DOI":"10.1109\/CVPR.2015.7299156"},{"key":"4012_CR50","doi-asserted-by":"publisher","first-page":"416","DOI":"10.1109\/ICCV.2001.937655","volume":"2","author":"D Martin","year":"2001","unstructured":"Martin, D., Fowlkes, C., Tal, D., et al.: A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics[C]. Proc. Eighth IEEE Int. Conf. Comput. Vis. (ICCV). 2, 416\u2013423 (2001)","journal-title":"Proc. Eighth IEEE Int. Conf. Comput. Vis. (ICCV)."},{"key":"4012_CR51","unstructured":"Franzen, R.: Kodak lossless true color image suite. source: http:\/\/r0k.us\/graphics\/kodak, 4(2) (1999)"},{"key":"4012_CR52","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2021.106949","volume":"226","author":"C Tian","year":"2021","unstructured":"Tian, C., Xu, Y., Zuo, W., et al.: Designing and training of a dual CNN for image denoising. Knowl.-Based Syst. 226, 106949 (2021)","journal-title":"Knowl.-Based Syst."},{"key":"4012_CR53","doi-asserted-by":"publisher","first-page":"4363","DOI":"10.1109\/TIP.2022.3183835","volume":"31","author":"JW Soh","year":"2022","unstructured":"Soh, J.W., Cho, N.I.: Variational deep image restoration. IEEE Trans. Image Process. 31, 4363\u20134376 (2022)","journal-title":"IEEE Trans. Image Process."},{"key":"4012_CR54","doi-asserted-by":"crossref","unstructured":"Xu, J., Yuan, M., Yan, D.M., et al.: Deep unfolding multi-scale regularizer network for image denoising. Comput. Vis. Media 9(2), 335\u2013350 (2023)","DOI":"10.1007\/s41095-022-0277-5"},{"key":"4012_CR55","doi-asserted-by":"crossref","unstructured":"Dabov, K., Foi, A., Katkovnik, V., et al.: Color image denoising via sparse 3D collaborative filtering with grouping constraint in luminance-chrominance space. In: 2007 IEEE International Conference on Image Processing (2007)","DOI":"10.1109\/ICIP.2007.4378954"}],"container-title":["Signal, Image and Video Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-025-04012-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11760-025-04012-y\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-025-04012-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,5,19]],"date-time":"2025-05-19T06:36:54Z","timestamp":1747636614000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11760-025-04012-y"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,4,2]]},"references-count":55,"journal-issue":{"issue":"6","published-print":{"date-parts":[[2025,6]]}},"alternative-id":["4012"],"URL":"https:\/\/doi.org\/10.1007\/s11760-025-04012-y","relation":{"has-preprint":[{"id-type":"doi","id":"10.21203\/rs.3.rs-4154630\/v1","asserted-by":"object"}]},"ISSN":["1863-1703","1863-1711"],"issn-type":[{"value":"1863-1703","type":"print"},{"value":"1863-1711","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,4,2]]},"assertion":[{"value":"23 March 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"16 February 2025","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"2 March 2025","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"2 April 2025","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"We declared no potential Conflict of interest with respect to the research, author-ship, and publication of this article.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"434"}}