{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,21]],"date-time":"2025-02-21T03:17:55Z","timestamp":1740107875986,"version":"3.37.3"},"reference-count":85,"publisher":"Springer Science and Business Media LLC","issue":"6","license":[{"start":{"date-parts":[[2024,12,1]],"date-time":"2024-12-01T00:00:00Z","timestamp":1733011200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,12,1]],"date-time":"2024-12-01T00:00:00Z","timestamp":1733011200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["61971007 and 61571013","61971007 and 61571013","61971007 and 61571013"],"award-info":[{"award-number":["61971007 and 61571013","61971007 and 61571013","61971007 and 61571013"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimedia Systems"],"published-print":{"date-parts":[[2024,12]]},"DOI":"10.1007\/s00530-024-01572-w","type":"journal-article","created":{"date-parts":[[2024,12,2]],"date-time":"2024-12-02T16:44:38Z","timestamp":1733157878000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["CGMAformer: CNN and gated multi axial-sparse transformer feature fusion network for image deraining"],"prefix":"10.1007","volume":"30","author":[{"given":"Yongsheng","family":"Qiu","sequence":"first","affiliation":[]},{"given":"Yuanyao","family":"Lu","sequence":"additional","affiliation":[]},{"given":"Yuantao","family":"Wang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,12,2]]},"reference":[{"key":"1572_CR1","doi-asserted-by":"publisher","first-page":"3525","DOI":"10.1109\/TIP.2022.3172208","volume":"31","author":"L Liao","year":"2022","unstructured":"Liao, L., Chen, W., Xiao, J., Wang, Z., Lin, C.-W., Satoh, S.: Unsupervised foggy scene understanding via self spatial-temporal label diffusion. IEEE Trans. Image Process. 31, 3525\u20133540 (2022)","journal-title":"IEEE Trans. Image Process."},{"key":"1572_CR2","doi-asserted-by":"crossref","unstructured":"Wang, W., Chen, C., Wang, J., Zha, S., Zhang, Y., Li, J.: Med-danet: dynamic architecture network for efficient medical volumetric segmentation. In: European Conference on Computer Vision, pp. 506\u2013522. Springer (2022)","DOI":"10.1007\/978-3-031-19803-8_30"},{"key":"1572_CR3","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2022.3142822","author":"Z Huang","year":"2022","unstructured":"Huang, Z., Wu, J., Lv, C.: Efficient deep reinforcement learning with imitative expert priors for autonomous driving. IEEE Trans. Neural Netw. Learn. Syst. (2022). https:\/\/doi.org\/10.1109\/TNNLS.2022.3142822","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"1572_CR4","first-page":"1","volume":"72","author":"L Wang","year":"2022","unstructured":"Wang, L., Qin, H., Zhou, X., Lu, X., Zhang, F.: R-yolo: a robust object detector in adverse weather. IEEE Trans. Instrum. Meas. 72, 1\u201311 (2022)","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"1572_CR5","doi-asserted-by":"publisher","first-page":"1487","DOI":"10.3390\/rs15061487","volume":"15","author":"Y Xi","year":"2023","unstructured":"Xi, Y., Jia, W., Miao, Q., Feng, J., Liu, X., Li, F.: Coderainnet: collaborative deraining network for drone-view object detection in rainy weather conditions. Remote Sens. 15, 1487 (2023)","journal-title":"Remote Sens."},{"issue":"4","key":"1572_CR6","doi-asserted-by":"publisher","first-page":"9541","DOI":"10.1109\/LRA.2022.3192200","volume":"7","author":"K Wang","year":"2022","unstructured":"Wang, K., Wang, T., Qu, J., Jiang, H., Li, Q., Chang, L.: An end-to-end cascaded image deraining and object detection neural network. IEEE Robot. Autom. Lett. 7(4), 9541\u20139548 (2022)","journal-title":"IEEE Robot. Autom. Lett."},{"key":"1572_CR7","doi-asserted-by":"crossref","unstructured":"Rai, S.N., Saluja, R., Arora, C., Balasubramanian, V.N., Subramanian, A., Jawahar, C.: Fluid: few-shot self-supervised image deraining. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision(WACV), Waikoloa, HI, USA, pp. 3077\u20133086 (2022)","DOI":"10.1109\/WACV51458.2022.00049"},{"key":"1572_CR8","doi-asserted-by":"crossref","unstructured":"Yu, Y., Yang, W., Tan, Y.-P., Kot, A.C.: Towards robust rain removal against adversarial attacks: a comprehensive benchmark analysis and beyond. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6003\u20136012 (2022)","DOI":"10.1109\/CVPR52688.2022.00592"},{"key":"1572_CR9","first-page":"12077","volume":"34","author":"E Xie","year":"2021","unstructured":"Xie, E., Wang, W., Yu, Z., Anandkumar, A., Alvarez, J.M., Luo, P.: Segformer: simple and efficient design for semantic segmentation with transformers. Adv. Neural Inf. Process. Syst. 34, 12077\u201312090 (2021)","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"1572_CR10","doi-asserted-by":"crossref","unstructured":"Chen, X., Pan, J., Lu, J., Fan, Z., Li, H.: Hybrid CNN-transformer feature fusion for single image deraining. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 37, no. 1, pp. 378\u2013386 (2023)","DOI":"10.1609\/aaai.v37i1.25111"},{"key":"1572_CR11","doi-asserted-by":"crossref","unstructured":"Chen, L.-C., Zhu, Y., Papandreou, G., Schroff, F., Adam, H.: Encoder\u2013decoder with atrous separable convolution for semantic image segmentation. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 801\u2013818 (2018)","DOI":"10.1007\/978-3-030-01234-2_49"},{"issue":"1","key":"1572_CR12","first-page":"53","volume":"2","author":"P Liu","year":"2009","unstructured":"Liu, P., Xu, J., Liu, J., Tang, X.: Pixel based temporal analysis using chromatic property for removing rain from videos. Comput. Inf. Sci. 2(1), 53\u201360 (2009)","journal-title":"Comput. Inf. Sci."},{"key":"1572_CR13","doi-asserted-by":"crossref","unstructured":"Chen, Y.-L., Hsu, C.-T.: A generalized low-rank appearance model for spatio-temporally correlated rain streaks. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1968\u20131975 (2013)","DOI":"10.1109\/ICCV.2013.247"},{"key":"1572_CR14","doi-asserted-by":"crossref","unstructured":"Li, M., Xie, Q., Zhao, Q., Wei, W., Gu, S., Tao, J., Meng, D.: Video rain streak removal by multiscale convolutional sparse coding. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6644\u20136653 (2018)","DOI":"10.1109\/CVPR.2018.00695"},{"key":"1572_CR15","doi-asserted-by":"crossref","unstructured":"Wei, W., Yi, L., Xie, Q., Zhao, Q., Meng, D., Xu, Z.: Should we encode rain streaks in video as deterministic or stochastic? In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2535\u20132544 (2017)","DOI":"10.1109\/ICCV.2017.275"},{"key":"1572_CR16","doi-asserted-by":"crossref","unstructured":"Ren, D., Zuo, W., Hu, Q., Zhu, P., Meng, D.: Progressive image deraining networks: a better and simpler baseline. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3932\u20133941 (2019)","DOI":"10.1109\/CVPR.2019.00406"},{"key":"1572_CR17","doi-asserted-by":"crossref","unstructured":"Fu, X., Qi, Q., Zha, Z.-J., Zhu, Y., Ding, X.: Rain streak removal via dual graph convolutional network. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 35, no. 2, pp. 1352\u20131360 (2021)","DOI":"10.1609\/aaai.v35i2.16224"},{"key":"1572_CR18","doi-asserted-by":"crossref","unstructured":"Yi, Q., Li, J., Dai, Q., Fang, F., Zhang, G., Zeng, T.: Structure-preserving deraining with residue channel prior guidance. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 4218\u20134227 (2021)","DOI":"10.1109\/ICCV48922.2021.00420"},{"issue":"10","key":"1572_CR19","doi-asserted-by":"publisher","first-page":"3981","DOI":"10.1109\/TCSVT.2020.3044887","volume":"31","author":"K Jiang","year":"2021","unstructured":"Jiang, K., Wang, Z., Yi, P., Chen, C., Han, Z., Lu, T., Huang, B., Jiang, J.: Decomposition makes better rain removal: an improved attention-guided deraining network. IEEE Trans. Circuits Syst. Video Technol. 31(10), 3981\u20133995 (2021). https:\/\/doi.org\/10.1109\/TCSVT.2020.3044887","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"1572_CR20","doi-asserted-by":"publisher","first-page":"1622","DOI":"10.1109\/TMM.2021.3068833","volume":"24","author":"Y Yang","year":"2022","unstructured":"Yang, Y., Guan, J., Huang, S., Wan, W., Xu, Y., Liu, J.: End-to-end rain removal network based on progressive residual detail supplement. IEEE Trans. Multimed. 24, 1622\u20131636 (2022). https:\/\/doi.org\/10.1109\/TMM.2021.3068833","journal-title":"IEEE Trans. Multimed."},{"key":"1572_CR21","doi-asserted-by":"crossref","unstructured":"Zamir, S.W., Arora, A., Khan, S., Hayat, M., Khan, F.S., Yang, M.: Restormer: efficient transformer for high-resolution image restoration. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5718\u20135729 (2022)","DOI":"10.1109\/CVPR52688.2022.00564"},{"key":"1572_CR22","doi-asserted-by":"crossref","unstructured":"Luo, Y., Xu, Y., Ji, H.: Removing rain from a single image via discriminative sparse coding. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 3397\u20133405 (2015)","DOI":"10.1109\/ICCV.2015.388"},{"key":"1572_CR23","doi-asserted-by":"crossref","unstructured":"Fu, X., Huang, J., Zeng, D., Huang, Y., Ding, X., Paisley, J.: Removing rain from single images via a deep detail network. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1715\u20131723 (2017)","DOI":"10.1109\/CVPR.2017.186"},{"key":"1572_CR24","doi-asserted-by":"crossref","unstructured":"Zamir, S.W., Arora, A., Khan, S., Hayat, M., Khan, F.S., Yang, M.-H., Shao, L.: Multi-stage progressive image restoration. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 14816\u201314826 (2021)","DOI":"10.1109\/CVPR46437.2021.01458"},{"issue":"11","key":"1572_CR25","doi-asserted-by":"publisher","first-page":"12978","DOI":"10.1109\/TPAMI.2022.3183612","volume":"45","author":"J Xiao","year":"2023","unstructured":"Xiao, J., Fu, X., Liu, A., Wu, F., Zha, Z.-J.: Image de-raining transformer. IEEE Trans. Pattern Anal. Mach. Intell. 45(11), 12978\u201312995 (2023). https:\/\/doi.org\/10.1109\/TPAMI.2022.3183612","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1572_CR26","doi-asserted-by":"crossref","unstructured":"Chen, X., Li, H., Li, M., Pan, J.: Learning a sparse transformer network for effective image deraining. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5896\u20135905 (2023)","DOI":"10.1109\/CVPR52729.2023.00571"},{"key":"1572_CR27","doi-asserted-by":"crossref","unstructured":"Wang, W., Xie, E., Li, X., Fan, D.-P., Song, K., Liang, D., Lu, T., Luo, P., Shao, L.: Pyramid vision transformer: a versatile backbone for dense prediction without convolutions. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 548\u2013558 (2021)","DOI":"10.1109\/ICCV48922.2021.00061"},{"key":"1572_CR28","doi-asserted-by":"crossref","unstructured":"Wang, Z., Cun, X., Bao, J., Zhou, W., Liu, J., Li, H.: Uformer: A general u-shaped transformer for image restoration. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 17662\u201317672 (2022)","DOI":"10.1109\/CVPR52688.2022.01716"},{"key":"1572_CR29","doi-asserted-by":"crossref","unstructured":"Liang, J., Cao, J., Sun, G., Zhang, K., Van Gool, L., Timofte, R.: Swinir: Image restoration using swin transformer. In: Proceedings of the IEEE International Conference on Computer Vision Workshops, pp. 1833\u20131844 (2021)","DOI":"10.1109\/ICCVW54120.2021.00210"},{"key":"1572_CR30","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, \u0141., Polosukhin, I.: Attention is all you need. Adv. Neural Inf. Process. Syst. 30 (2017)"},{"key":"1572_CR31","doi-asserted-by":"crossref","unstructured":"Li, R., Cheong, L.-F., Tan, R.T.: Heavy rain image restoration: Integrating physics model and conditional adversarial learning. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1633\u20131642 (2019)","DOI":"10.1109\/CVPR.2019.00173"},{"key":"1572_CR32","doi-asserted-by":"crossref","unstructured":"Yang, W., Tan, R.T., Feng, J., Liu, J., Guo, Z., Yan, S.: Deep joint rain detection and removal from a single image. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1685\u20131694 (2017)","DOI":"10.1109\/CVPR.2017.183"},{"key":"1572_CR33","doi-asserted-by":"crossref","unstructured":"Garg, K., Nayar, S.K.: Detection and removal of rain from videos. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2004. CVPR 2004, vol. 1 (2004)","DOI":"10.1109\/CVPR.2004.1315077"},{"issue":"4","key":"1572_CR34","doi-asserted-by":"publisher","first-page":"1742","DOI":"10.1109\/TIP.2011.2179057","volume":"21","author":"L-W Kang","year":"2011","unstructured":"Kang, L.-W., Lin, C.-W., Fu, Y.-H.: Automatic single-image-based rain streaks removal via image decomposition. IEEE Trans. Image Process. 21(4), 1742\u20131755 (2011)","journal-title":"IEEE Trans. Image Process."},{"key":"1572_CR35","doi-asserted-by":"crossref","unstructured":"Chen, Y.-L., Hsu, C.-T.: A generalized low-rank appearance model for spatio-temporally correlated rain streaks. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1968\u20131975 (2013)","DOI":"10.1109\/ICCV.2013.247"},{"key":"1572_CR36","doi-asserted-by":"crossref","unstructured":"Gu, S., Meng, D., Zuo, W., Zhang, L.: Joint convolutional analysis and synthesis sparse representation for single image layer separation. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1717\u20131725 (2017)","DOI":"10.1109\/ICCV.2017.189"},{"issue":"6","key":"1572_CR37","doi-asserted-by":"publisher","first-page":"2944","DOI":"10.1109\/TIP.2017.2691802","volume":"26","author":"X Fu","year":"2017","unstructured":"Fu, X., Huang, J., Ding, X., Liao, Y., Paisley, J.: Clearing the skies: a deep network architecture for single-image rain removal. IEEE Trans. Image Process. 26(6), 2944\u20132956 (2017). https:\/\/doi.org\/10.1109\/TIP.2017.2691802","journal-title":"IEEE Trans. Image Process."},{"key":"1572_CR38","doi-asserted-by":"crossref","unstructured":"Cui, Y., Ren, W., Yang, S., Cao, X., Knoll, A.: Irnext: rethinking convolutional network design for image restoration. In: International Conference on Machine Learning (2023). https:\/\/api.semanticscholar.org\/CorpusID:260927679","DOI":"10.1109\/ICCV51070.2023.01195"},{"key":"1572_CR39","doi-asserted-by":"crossref","unstructured":"Cui, Y., Tao, Y., Bing, Z., Ren, W., Gao, X., Cao, X., Huang, K., Knoll, A.: Selective frequency network for image restoration. In: International Conference on Learning Representations (2023). https:\/\/api.semanticscholar.org\/CorpusID:259298517","DOI":"10.1109\/ICCV51070.2023.01195"},{"issue":"2","key":"1572_CR40","doi-asserted-by":"publisher","first-page":"1093","DOI":"10.1109\/TPAMI.2023.3330416","volume":"46","author":"Y Cui","year":"2024","unstructured":"Cui, Y., Ren, W., Cao, X., Knoll, A.: Image restoration via frequency selection. IEEE Trans. Pattern Anal. Mach. Intell. 46(2), 1093\u20131108 (2024). https:\/\/doi.org\/10.1109\/TPAMI.2023.3330416","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1572_CR41","doi-asserted-by":"crossref","unstructured":"Wei, W., Meng, D., Zhao, Q., Xu, Z., Wu, Y.: Semi-supervised transfer learning for image rain removal. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3872\u20133881 (2019)","DOI":"10.1109\/CVPR.2019.00400"},{"key":"1572_CR42","doi-asserted-by":"crossref","unstructured":"Yang, F., Yang, H., Fu, J., Lu, H., Guo, B.: Learning texture transformer network for image super-resolution. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5790\u20135799 (2020)","DOI":"10.1109\/CVPR42600.2020.00583"},{"key":"1572_CR43","doi-asserted-by":"crossref","unstructured":"Chen, H., Wang, Y., Guo, T., Xu, C., Deng, Y., Liu, Z., Ma, S., Xu, C., Xu, C., Gao, W.: Pre-trained image processing transformer. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 12294\u201312305 (2021)","DOI":"10.1109\/CVPR46437.2021.01212"},{"key":"1572_CR44","doi-asserted-by":"crossref","unstructured":"Cai, Y., Zhang, W., Wu, Y., Jin, C.: Fusionformer: a concise unified feature fusion transformer for 3d pose estimation. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 38, no. 2, pp. 900\u2013908 (2024)","DOI":"10.1609\/aaai.v38i2.27849"},{"key":"1572_CR45","unstructured":"Devlin, J., Chang, M.-W., Lee, K., Toutanova, K.: Bert: pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018)"},{"issue":"10","key":"1572_CR46","doi-asserted-by":"publisher","first-page":"5562","DOI":"10.1109\/TCSVT.2023.3260274","volume":"33","author":"X Wen","year":"2023","unstructured":"Wen, X., Nie, W., Liu, J., Su, Y.: Mrft: multiscale recurrent fusion transformer based prior knowledge for bit-depth enhancement. IEEE Trans. Circuits Syst. Video Technol. 33(10), 5562\u20135575 (2023). https:\/\/doi.org\/10.1109\/TCSVT.2023.3260274","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"1572_CR47","doi-asserted-by":"crossref","unstructured":"Zhang, J., Huang, J., Luo, Z., Zhang, G., Zhang, X., Lu, S.: Da-detr: Domain adaptive detection transformer with information fusion. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 23787\u201323798 (2023)","DOI":"10.1109\/CVPR52729.2023.02278"},{"key":"1572_CR48","unstructured":"Touvron, H., Cord, M., Douze, M., Massa, F., Sablayrolles, A., J\u00e9gou, H.: Training data-efficient image transformers & distillation through attention. In: International Conference on Machine Learning, pp. 10347\u201310357. PMLR (2021)"},{"key":"1572_CR49","doi-asserted-by":"crossref","unstructured":"Yuan, L., Chen, Y., Wang, T., Yu, W., Shi, Y., Jiang, Z.-H., Tay, F.E., Feng, J., Yan, S.: Tokens-to-token vit: training vision transformers from scratch on imagenet. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 558\u2013567 (2021)","DOI":"10.1109\/ICCV48922.2021.00060"},{"key":"1572_CR50","doi-asserted-by":"crossref","unstructured":"Yuan, K., Guo, S., Liu, Z., Zhou, A., Yu, F., Wu, W.: Incorporating convolution designs into visual transformers. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 559\u2013568 (2021)","DOI":"10.1109\/ICCV48922.2021.00062"},{"key":"1572_CR51","doi-asserted-by":"crossref","unstructured":"Jiang, K., Wang, Z., Chen, C., Wang, Z., Cui, L., Lin, C.-W.: Magic elf: Image deraining meets association learning and transformer. arXiv preprint arXiv:2207.10455 (2022)","DOI":"10.1145\/3503161.3547760"},{"key":"1572_CR52","doi-asserted-by":"crossref","unstructured":"Suganuma, M., Liu, X., Okatani, T.: Attention-based adaptive selection of operations for image restoration in the presence of unknown combined distortions. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 9031\u20139040 (2019)","DOI":"10.1109\/CVPR.2019.00925"},{"key":"1572_CR53","doi-asserted-by":"publisher","unstructured":"Ren, W., Ma, L., Zhang, J., Pan, J., Cao, X., Liu, W., Yang, M.-H.: Gated fusion network for single image dehazing. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3253\u20133261 (2018). https:\/\/doi.org\/10.1109\/CVPR.2018.00343","DOI":"10.1109\/CVPR.2018.00343"},{"key":"1572_CR54","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., et al.: An image is worth 16x16 words: transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)"},{"key":"1572_CR55","doi-asserted-by":"crossref","unstructured":"Wang, H., Zhu, Y., Green, B., Adam, H., Yuille, A., Chen, L.-C.: Axial-deeplab: stand-alone axial-attention for panoptic segmentation. In: European Conference on Computer Vision, pp. 108\u2013126. Springer, (2020)","DOI":"10.1007\/978-3-030-58548-8_7"},{"key":"1572_CR56","doi-asserted-by":"crossref","unstructured":"Valanarasu, J.M.J., Oza, P., Hacihaliloglu, I., Patel, V.M.: Medical transformer: gated axial-attention for medical image segmentation. In: Medical Image Computing and Computer Assisted Intervention\u2013MICCAI 2021: 24th International Conference, Strasbourg, France, September 27\u2013October 1, 2021, Proceedings, Part I 24, pp. 36\u201346. Springer (2021)","DOI":"10.1007\/978-3-030-87193-2_4"},{"key":"1572_CR57","unstructured":"Zhao, G., Lin, J., Zhang, Z., Ren, X., Su, Q., Sun, X.: Explicit sparse transformer: Concentrated attention through explicit selection. arXiv preprint arXiv:1912.11637 (2019)"},{"key":"1572_CR58","doi-asserted-by":"crossref","unstructured":"Jiang, K., Wang, Z., Yi, P., Chen, C., Huang, B., Luo, Y., Ma, J., Jiang, J.: Multi-scale progressive fusion network for single image deraining. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 8343\u20138352 (2020)","DOI":"10.1109\/CVPR42600.2020.00837"},{"key":"1572_CR59","doi-asserted-by":"crossref","unstructured":"Li, G., He, X., Zhang, W., Chang, H., Dong, L., Lin, L.: Non-locally enhanced encoder-decoder network for single image de-raining. In: Proceedings of the 26th ACM International Conference on Multimedia, pp. 1056\u20131064 (2018)","DOI":"10.1145\/3240508.3240636"},{"key":"1572_CR60","doi-asserted-by":"crossref","unstructured":"Wang, T., Yang, X., Xu, K., Chen, S., Zhang, Q., Lau, R.W.H.: Spatial attentive single-image deraining with a high quality real rain dataset. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 12262\u201312271 (2019)","DOI":"10.1109\/CVPR.2019.01255"},{"issue":"11","key":"1572_CR61","doi-asserted-by":"publisher","first-page":"3943","DOI":"10.1109\/TCSVT.2019.2920407","volume":"30","author":"H Zhang","year":"2020","unstructured":"Zhang, H., Sindagi, V., Patel, V.M.: Image de-raining using a conditional generative adversarial network. IEEE Trans. Circuits Syst. Video Technol. 30(11), 3943\u20133956 (2020). https:\/\/doi.org\/10.1109\/TCSVT.2019.2920407","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"1572_CR62","doi-asserted-by":"crossref","unstructured":"Yasarla, R., Sindagi, V.A., Patel, V.M.: Syn2real transfer learning for image deraining using gaussian processes. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2723\u20132733 (2020)","DOI":"10.1109\/CVPR42600.2020.00280"},{"issue":"13","key":"1572_CR63","doi-asserted-by":"publisher","first-page":"800","DOI":"10.1049\/el:20080522","volume":"44","author":"Q Huynh-Thu","year":"2008","unstructured":"Huynh-Thu, Q., Ghanbari, M.: Scope of validity of psnr in image\/video quality assessment. Electron. Lett. 44(13), 800\u2013801 (2008)","journal-title":"Electron. Lett."},{"issue":"4","key":"1572_CR64","doi-asserted-by":"publisher","first-page":"600","DOI":"10.1109\/TIP.2003.819861","volume":"13","author":"Z Wang","year":"2004","unstructured":"Wang, Z., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P.: Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 13(4), 600\u2013612 (2004). https:\/\/doi.org\/10.1109\/TIP.2003.819861","journal-title":"IEEE Trans. Image Process."},{"issue":"3","key":"1572_CR65","doi-asserted-by":"publisher","first-page":"209","DOI":"10.1109\/LSP.2012.2227726","volume":"20","author":"A Mittal","year":"2013","unstructured":"Mittal, A., Soundararajan, R., Bovik, A.C.: Making a \u201ccompletely blind\u2019\u2019 image quality analyzer. IEEE Signal Process. Lett. 20(3), 209\u2013212 (2013). https:\/\/doi.org\/10.1109\/LSP.2012.2227726","journal-title":"IEEE Signal Process. Lett."},{"key":"1572_CR66","doi-asserted-by":"crossref","unstructured":"Venkatanath, N., Praneeth, D., Bh, M.C., Channappayya, S.S., Medasani, S.S.: Blind image quality evaluation using perception based features. In: 2015 Twenty First National Conference on Communications (NCC), pp. 1\u20136. IEEE (2015)","DOI":"10.1109\/NCC.2015.7084843"},{"key":"1572_CR67","doi-asserted-by":"crossref","unstructured":"Mittal, A., Moorthy, A.K., Bovik, A.C.: Blind\/referenceless image spatial quality evaluator. In: Proceedings of the 2011 Conference Record of the Forty Fifth Asilomar Conference on Signals, Systems and Computers (ASILOMAR), pp. 723\u2013727 (2011)","DOI":"10.1109\/ACSSC.2011.6190099"},{"issue":"1","key":"1572_CR68","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1186\/s13640-019-0479-7","volume":"2019","author":"X Chen","year":"2019","unstructured":"Chen, X., Zhang, Q., Lin, M., Yang, G., He, C.: No-reference color image quality assessment: from entropy to perceptual quality. EURASIP J. Image Video Process. 2019(1), 1\u201314 (2019)","journal-title":"EURASIP J. Image Video Process."},{"key":"1572_CR69","doi-asserted-by":"crossref","unstructured":"Li, Y., Tan, R.T., Guo, X., Lu, J., Brown, M.S.: Rain streak removal using layer priors. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2736\u20132744 (2016)","DOI":"10.1109\/CVPR.2016.299"},{"key":"1572_CR70","doi-asserted-by":"crossref","unstructured":"Li, X., Wu, J., Lin, Z., Liu, H., Zha, H.: Recurrent squeeze-and-excitation context aggregation net for single image deraining. In: Proceedings of the European Conference on Computer Vision, pp. 254\u2013269 (2018)","DOI":"10.1007\/978-3-030-01234-2_16"},{"key":"1572_CR71","doi-asserted-by":"crossref","unstructured":"Wang, H., Xie, Q., Zhao, Q., Meng, D.: A model-driven deep neural network for single image rain removal. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3100\u20133109 (2020)","DOI":"10.1109\/CVPR42600.2020.00317"},{"key":"1572_CR72","first-page":"324","volume":"39","author":"RA Bradley","year":"1952","unstructured":"Bradley, R.A., Terry, M.E.: Rank analysis of incomplete block designs the method of paired comparisons. Biometrika 39, 324\u2013345 (1952)","journal-title":"Biometrika"},{"key":"1572_CR73","doi-asserted-by":"publisher","unstructured":"Abdelhamed, A., Lin, S., Brown, M.S.: A high-quality denoising dataset for smartphone cameras. In: 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 1692\u20131700 (2018). https:\/\/doi.org\/10.1109\/CVPR.2018.00182","DOI":"10.1109\/CVPR.2018.00182"},{"issue":"8","key":"1572_CR74","doi-asserted-by":"publisher","first-page":"2080","DOI":"10.1109\/TIP.2007.901238","volume":"16","author":"K Dabov","year":"2007","unstructured":"Dabov, K., Foi, A., Katkovnik, V., Egiazarian, K.: Image denoising by sparse 3-d transform-domain collaborative filtering. IEEE Trans. Image Process. 16(8), 2080\u20132095 (2007). https:\/\/doi.org\/10.1109\/TIP.2007.901238","journal-title":"IEEE Trans. Image Process."},{"issue":"7","key":"1572_CR75","doi-asserted-by":"publisher","first-page":"3142","DOI":"10.1109\/TIP.2017.2662206","volume":"26","author":"K Zhang","year":"2017","unstructured":"Zhang, K., Zuo, W., Chen, Y., Meng, D., Zhang, L.: Beyond a gaussian denoiser: residual learning of deep cnn for image denoising. IEEE Trans. Image Process. 26(7), 3142\u20133155 (2017). https:\/\/doi.org\/10.1109\/TIP.2017.2662206","journal-title":"IEEE Trans. Image Process."},{"key":"1572_CR76","first-page":"1688","volume":"32","author":"Z Yue","year":"2019","unstructured":"Yue, Z., Yong, H., Zhao, Q., Meng, D., Zhang, L.: Variational denoising network: toward blind noise modeling and removal. Adv. Neural Inf. Process. Syst. 32, 1688\u20131699 (2019)","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"1572_CR77","doi-asserted-by":"crossref","unstructured":"Yue, Z., Zhao, Q., Zhang, L., Meng, D.: Dual adversarial network: toward real-world noise removal and noise generation. In: Computer Vision\u2013ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part X 16, pp. 41\u201358. Springer (2020)","DOI":"10.1007\/978-3-030-58607-2_3"},{"key":"1572_CR78","doi-asserted-by":"crossref","unstructured":"Zamir, S.W., Arora, A., Khan, S., Hayat, M., Khan, F.S., Yang, M.-H., Shao, L.: Cycleisp: real image restoration via improved data synthesis. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2693\u20132702 (2020)","DOI":"10.1109\/CVPR42600.2020.00277"},{"key":"1572_CR79","doi-asserted-by":"crossref","unstructured":"Zamir, S.W., Arora, A., Khan, S., Hayat, M., Khan, F.S., Yang, M.-H., Shao, L.: Learning enriched features for real image restoration and enhancement. In: Computer Vision\u2013ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part XXV 16, pp. 492\u2013511. Springer (2020)","DOI":"10.1007\/978-3-030-58595-2_30"},{"key":"1572_CR80","doi-asserted-by":"publisher","unstructured":"Ren, C., He, X., Wang, C., Zhao, Z.: Adaptive consistency prior based deep network for image denoising. In: 2021 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 8592\u20138602 (2021). https:\/\/doi.org\/10.1109\/CVPR46437.2021.00849","DOI":"10.1109\/CVPR46437.2021.00849"},{"key":"1572_CR81","doi-asserted-by":"crossref","unstructured":"Lin, T.-Y., Maire, M., Belongie, S., Hays, J., Perona, P., Ramanan, D., Doll\u00e1r, P., Zitnick, C.L.: Microsoft coco: Common objects in context. In: Computer Vision\u2013ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6-12, 2014, Proceedings, Part V 13, pp. 740\u2013755. Springer (2014)","DOI":"10.1007\/978-3-319-10602-1_48"},{"issue":"6","key":"1572_CR82","doi-asserted-by":"publisher","first-page":"1137","DOI":"10.1109\/TPAMI.2016.2577031","volume":"39","author":"S Ren","year":"2017","unstructured":"Ren, S., He, K., Girshick, R., Sun, J.: Faster r-cnn: towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 39(6), 1137\u20131149 (2017). https:\/\/doi.org\/10.1109\/TPAMI.2016.2577031","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1572_CR83","unstructured":"Redmon, J., Farhadi, A.: Yolov3: an incremental improvement. arXiv preprint arXiv:1804.02767 (2018)"},{"key":"1572_CR84","doi-asserted-by":"crossref","unstructured":"Wang, C.-Y., Bochkovskiy, A., Liao, H.-Y.M.: Yolov7: trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7464\u20137475 (2023)","DOI":"10.1109\/CVPR52729.2023.00721"},{"key":"1572_CR85","doi-asserted-by":"crossref","unstructured":"Long, J., Shelhamer, E., Darrell, T.: Fully convolutional networks for semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3431\u20133440 (2015)","DOI":"10.1109\/CVPR.2015.7298965"}],"container-title":["Multimedia Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-024-01572-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00530-024-01572-w\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-024-01572-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,12,16]],"date-time":"2024-12-16T09:23:37Z","timestamp":1734341017000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00530-024-01572-w"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,12]]},"references-count":85,"journal-issue":{"issue":"6","published-print":{"date-parts":[[2024,12]]}},"alternative-id":["1572"],"URL":"https:\/\/doi.org\/10.1007\/s00530-024-01572-w","relation":{},"ISSN":["0942-4962","1432-1882"],"issn-type":[{"type":"print","value":"0942-4962"},{"type":"electronic","value":"1432-1882"}],"subject":[],"published":{"date-parts":[[2024,12]]},"assertion":[{"value":"4 May 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"7 November 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"2 December 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"379"}}