{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,1]],"date-time":"2026-02-01T03:03:30Z","timestamp":1769915010885,"version":"3.49.0"},"reference-count":52,"publisher":"Springer Science and Business Media LLC","issue":"6-7","license":[{"start":{"date-parts":[[2024,5,10]],"date-time":"2024-05-10T00:00:00Z","timestamp":1715299200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,5,10]],"date-time":"2024-05-10T00:00:00Z","timestamp":1715299200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62202142"],"award-info":[{"award-number":["62202142"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["SIViP"],"published-print":{"date-parts":[[2024,8]]},"DOI":"10.1007\/s11760-024-03224-y","type":"journal-article","created":{"date-parts":[[2024,5,10]],"date-time":"2024-05-10T12:01:31Z","timestamp":1715342491000},"page":"5173-5187","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["EFRG-SRGAN: combining augmented features for real-world super-resolution"],"prefix":"10.1007","volume":"18","author":[{"given":"Yibing","family":"Yao","sequence":"first","affiliation":[]},{"given":"Zhisheng","family":"Cui","sequence":"additional","affiliation":[]},{"given":"Dakai","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Miaohui","family":"Zhang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,5,10]]},"reference":[{"key":"3224_CR1","doi-asserted-by":"crossref","unstructured":"Dong, C., et al.: Image super-resolution using deep convolutional networks. IEEE Transact. Pattern Anal. Mach. Intel. (2015)","DOI":"10.1109\/TPAMI.2015.2439281"},{"key":"3224_CR2","doi-asserted-by":"publisher","DOI":"10.1038\/s41592-019-0677-3","author":"N Vogt","year":"2019","unstructured":"Vogt, N.: Super-resolution Raman imaging. Nat. Methods (2019). https:\/\/doi.org\/10.1038\/s41592-019-0677-3","journal-title":"Nat. Methods"},{"key":"3224_CR3","doi-asserted-by":"publisher","DOI":"10.1109\/TGRS.2023.3330876","author":"B Liu","year":"2023","unstructured":"Liu, B., et al.: RAN: Region-aware network for remote sensing image super-resolution. IEEE Transact. Geosci. Remote Sens. (2023). https:\/\/doi.org\/10.1109\/TGRS.2023.3330876","journal-title":"IEEE Transact. Geosci. Remote Sens."},{"key":"3224_CR4","doi-asserted-by":"publisher","DOI":"10.1109\/TGRS.2023.3267271","author":"Y Liu","year":"2023","unstructured":"Liu, Y., et al.: Distilling knowledge from super resolution for efficient remote sensing salient object detection. IEEE Transact. Geosci. Remote Sens. (2023). https:\/\/doi.org\/10.1109\/TGRS.2023.3267271","journal-title":"IEEE Transact. Geosci. Remote Sens."},{"key":"3224_CR5","doi-asserted-by":"publisher","DOI":"10.1016\/j.jnca.2020.102691","author":"K Guo","year":"2020","unstructured":"Guo, K., et al.: Towards efficient motion-blurred public security video super-resolution based on back-projection networks. J. Netw. Comput. Appl. (2020). https:\/\/doi.org\/10.1016\/j.jnca.2020.102691","journal-title":"J. Netw. Comput. Appl."},{"key":"3224_CR6","doi-asserted-by":"crossref","unstructured":"Dong, C., Loy, C.C., He, K., Tang, X.: Learning a deep convolutional network for image super-resolution. In: European conference on computer vision (ECCV). (2014)","DOI":"10.1007\/978-3-319-10593-2_13"},{"key":"3224_CR7","doi-asserted-by":"crossref","unstructured":"Ledig, C., et al.: Photo-realistic single image super-resolution using a generative adversarial network. In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR). (2017)","DOI":"10.1109\/CVPR.2017.19"},{"key":"3224_CR8","doi-asserted-by":"crossref","unstructured":"Wang, X., et al.: Real-ESRGAN: Training real-world blind super-resolution with pure synthetic data. In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR). (2021)","DOI":"10.1109\/ICCVW54120.2021.00217"},{"key":"3224_CR9","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2020.2982166","author":"Z Wang","year":"2020","unstructured":"Wang, Z., Chen, J., Hoi, S.C.H.: Deep learning for image super-resolution: A survey. IEEE Transact. Pattern Anal. Mach. Intell. (2020). https:\/\/doi.org\/10.1109\/TPAMI.2020.2982166","journal-title":"IEEE Transact. Pattern Anal. Mach. Intell."},{"key":"3224_CR10","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2022.3204461","author":"C Saharia","year":"2022","unstructured":"Saharia, C., et al.: Image super-resolution via iterative refinement. IEEE Transact. Pattern Anal. Mach. Intell. (2022). https:\/\/doi.org\/10.1109\/TPAMI.2022.3204461","journal-title":"IEEE Transact. Pattern Anal. Mach. Intell."},{"key":"3224_CR11","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2022.3203009","author":"A Liu","year":"2022","unstructured":"Liu, A., et al.: Blind image super-resolution: A survey and beyond. IEEE Transact. Pattern Anal. Mach. Intel. (2022). https:\/\/doi.org\/10.1109\/TPAMI.2022.3203009","journal-title":"IEEE Transact. Pattern Anal. Mach. Intel."},{"key":"3224_CR12","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2022.10.007","author":"DC Lepcha","year":"2023","unstructured":"Lepcha, D.C., et al.: Image super-resolution: A comprehensive review, recent trends, challenges and applications. Inf. Fusion (2023). https:\/\/doi.org\/10.1016\/j.inffus.2022.10.007","journal-title":"Inf. Fusion"},{"key":"3224_CR13","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2023.01.024","author":"G Gendy","year":"2023","unstructured":"Gendy, G., He, G., Sabor, N.: Lightweight image super-resolution based on deep learning: State-of-the-art and future directions. Inf. Fusion (2023). https:\/\/doi.org\/10.1016\/j.inffus.2023.01.024","journal-title":"Inf. Fusion"},{"key":"3224_CR14","doi-asserted-by":"crossref","unstructured":"Dong, C. Loy, C.C. Tang, X.: Accelerating the super-resolution convolutional neural network. In: European conference on computer vision (ECCV). (2016)","DOI":"10.1007\/978-3-319-46475-6_25"},{"key":"3224_CR15","doi-asserted-by":"crossref","unstructured":"Kim, J., Lee, J.K., Lee, K.M.: Accurate image super-resolution using very deep convolutional networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR). (2016)","DOI":"10.1109\/CVPR.2016.182"},{"key":"3224_CR16","doi-asserted-by":"crossref","unstructured":"Kim, J., Lee, J.K., Lee, KM.: Deeply-recursive convolutional network for image super-resolution. In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR). (2016)","DOI":"10.1109\/CVPR.2016.181"},{"key":"3224_CR17","doi-asserted-by":"crossref","unstructured":"Lai, W-S., et al.: Deep laplacian pyramid networks for fast and accurate super-resolution. In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR). (2017)","DOI":"10.1109\/CVPR.2017.618"},{"key":"3224_CR18","doi-asserted-by":"crossref","unstructured":"Lim, B., et al.: Enhanced deep residual networks for single image super-resolution. In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR) workshops. (2017)","DOI":"10.1109\/CVPRW.2017.151"},{"key":"3224_CR19","doi-asserted-by":"crossref","unstructured":"Zhang, Y., et al.: Image super-resolution using very deep residual channel attention networks. In: Proceedings of the European conference on computer vision (ECCV). (2018)","DOI":"10.1007\/978-3-030-01234-2_18"},{"key":"3224_CR20","doi-asserted-by":"crossref","unstructured":"Huang, G., et al. Densely connected convolutional networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR). (2017)","DOI":"10.1109\/CVPR.2017.243"},{"key":"3224_CR21","doi-asserted-by":"crossref","unstructured":"He, K., et al.: Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR). (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"3224_CR22","unstructured":"Goodfellow, I., et al.: Generative adversarial nets. Adv. Neural Inf. Process. Syst. (2014)"},{"key":"3224_CR23","doi-asserted-by":"crossref","unstructured":"Wang, X., et al.: ESRGAN: Enhanced super-resolution generative adversarial networks. In: Proceedings of the European conference on computer vision (eccv) workshops. (2018)","DOI":"10.1007\/978-3-030-11021-5_5"},{"key":"3224_CR24","doi-asserted-by":"crossref","unstructured":"Zhang, W., et al.: RankSRGAN: Generative adversarial networks with ranker for image super-resolution. In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR). (2019)","DOI":"10.1109\/ICCV.2019.00319"},{"key":"3224_CR25","first-page":"5998","volume":"30","author":"A Vaswani","year":"2017","unstructured":"Vaswani, A., et al.: Attention is all you need. Adv. Neural. Inf. Process. Syst. 30, 5998 (2017)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"3224_CR26","doi-asserted-by":"crossref","unstructured":"Wei, M., Zhang, X.: Super-resolution neural operator. In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR). (2023)","DOI":"10.1109\/CVPR52729.2023.01750"},{"key":"3224_CR27","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2022.3229689","author":"J-N Su","year":"2022","unstructured":"Su, J.-N., et al.: Global learnable attention for single image super-resolution. IEEE Transact. Pattern Anal. Mach. Intell. (2022). https:\/\/doi.org\/10.1109\/TPAMI.2022.3229689","journal-title":"IEEE Transact. Pattern Anal. Mach. Intell."},{"key":"3224_CR28","doi-asserted-by":"publisher","first-page":"4905","DOI":"10.1109\/TIP.2021.3077135","volume":"30","author":"Y Yan","year":"2021","unstructured":"Yan, Y., et al.: SRGAT: Single image super-resolution with graph attention network. IEEE Transact. Image Process. 30, 4905 (2021)","journal-title":"IEEE Transact. Image Process."},{"key":"3224_CR29","unstructured":"Mnih, V., Heess, N., Graves, A. Recurrent models of visual attention. Adv. Neural Inf. Process. Syst. (2014)"},{"key":"3224_CR30","doi-asserted-by":"crossref","unstructured":"Dai, T., et al.: Second-order attention network for single image super-resolution. In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR). (2019)","DOI":"10.1109\/CVPR.2019.01132"},{"key":"3224_CR31","doi-asserted-by":"crossref","unstructured":"Wang, X., et al.: Recovering realistic texture in image super-resolution by deep spatial feature transform. In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR). (2018)","DOI":"10.1109\/CVPR.2018.00070"},{"key":"3224_CR32","doi-asserted-by":"crossref","unstructured":"Niu, B., et al.: Single image super-resolution via a holistic attention network. European conference on computer vision (ECCV). (2020)","DOI":"10.1007\/978-3-030-58610-2_12"},{"key":"3224_CR33","doi-asserted-by":"crossref","unstructured":"Zhao, H., et al.: Efficient image super-resolution using pixel attention. In: Proceedings of the European conference on computer vision (ECCV) workshops. (2020)","DOI":"10.1007\/978-3-030-67070-2_3"},{"key":"3224_CR34","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2022.3176526","author":"G Yin","year":"2022","unstructured":"Yin, G., et al.: Conditional hyper-network for blind super-resolution with multiple degradations. IEEE Transact. Image Process. (2022). https:\/\/doi.org\/10.1109\/TIP.2022.3176526","journal-title":"IEEE Transact. Image Process."},{"key":"3224_CR35","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2022.3206870","author":"Y Huang","year":"2022","unstructured":"Huang, Y., et al.: Transitional learning: Exploring the transition states of degradation for blind super-resolution. IEEE Transact. Pattern Anal. Mach. Intell. (2022). https:\/\/doi.org\/10.1109\/TPAMI.2022.3206870","journal-title":"IEEE Transact. Pattern Anal. Mach. Intell."},{"key":"3224_CR36","doi-asserted-by":"publisher","DOI":"10.1109\/JAS.2022.105914","author":"X Wang","year":"2022","unstructured":"Wang, X., Ma, J., Jiang, J.: Contrastive learning for blind super-resolution via a distortion-specific network. IEEE\/CAA J. Autom. Sin. (2022). https:\/\/doi.org\/10.1109\/JAS.2022.105914","journal-title":"IEEE\/CAA J. Autom. Sin."},{"key":"3224_CR37","doi-asserted-by":"publisher","first-page":"1646","DOI":"10.1109\/83.650118","volume":"6","author":"M Elad","year":"1997","unstructured":"Elad, M., Feuer, A.: Restoration of a single superresolution image from several blurred, noisy, and undersampled measured images. IEEE Transact. Image Process. 6, 1646 (1997)","journal-title":"IEEE Transact. Image Process."},{"key":"3224_CR38","doi-asserted-by":"crossref","unstructured":"Wei, Y., et al.: Unsupervised real-world image super resolution via domain-distance aware training. In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR). (2021)","DOI":"10.1109\/CVPR46437.2021.01318"},{"key":"3224_CR39","doi-asserted-by":"crossref","unstructured":"Agustsson, E., Timofte, R.: Ntire 2017 challenge on single image super-resolution: Dataset and study. In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR) workshops. (2017)","DOI":"10.1109\/CVPRW.2017.150"},{"key":"3224_CR40","unstructured":"Timofte, R., et al.: Ntire 2017 challenge on single image super-resolution: Methods and results. In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR) workshops. (2017)"},{"key":"3224_CR41","doi-asserted-by":"crossref","unstructured":"Bevilacqua, M., et al.: Low-complexity single-image super-resolution based on nonnegative neighbor embedding. (2012)","DOI":"10.5244\/C.26.135"},{"key":"3224_CR42","doi-asserted-by":"crossref","unstructured":"Zeyde, R., Elad, M., Protter, M.: On single image scale-up using sparse-representations. In: Curves and surfaces: 7th international conference, Springer, Berlin Heidelberg. (2012)","DOI":"10.1007\/978-3-642-27413-8_47"},{"key":"3224_CR43","unstructured":"Martin, D., et al.: A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics. In: Proceedings eighth IEEE international conference on computer vision (ICCV). (2001)"},{"key":"3224_CR44","doi-asserted-by":"crossref","unstructured":"Huang, J-B., Singh, A., Ahuja, N.: Single image super-resolution from transformed self-exemplars. In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR). (2015)","DOI":"10.1109\/CVPR.2015.7299156"},{"key":"3224_CR45","doi-asserted-by":"crossref","unstructured":"Fujimoto, A., et al.: Manga109 dataset and creation of metadata. In: Proceedings of the 1st international workshop on comics analysis, processing and understanding. (2016)","DOI":"10.1145\/3011549.3011551"},{"key":"3224_CR46","doi-asserted-by":"publisher","DOI":"10.1109\/LSP.2012.2227726","author":"A Mittal","year":"2012","unstructured":"Mittal, A., Soundararajan, R., Bovik, A.C.: Making a \u201ccompletely blind\u201d image quality analyzer. IEEE Signal Process. Lett. (2012). https:\/\/doi.org\/10.1109\/LSP.2012.2227726","journal-title":"IEEE Signal Process. Lett."},{"key":"3224_CR47","doi-asserted-by":"crossref","unstructured":"Yang, S, et al.: Maniqa: Multi-dimension attention network for no-reference image quality assessment. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (CVPR). (2022)","DOI":"10.1109\/CVPRW56347.2022.00126"},{"key":"3224_CR48","doi-asserted-by":"crossref","unstructured":"Zhang, K., et al.: Designing a practical degradation model for deep blind image super-resolution.In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR). (2021)","DOI":"10.1109\/ICCV48922.2021.00475"},{"key":"3224_CR49","doi-asserted-by":"crossref","unstructured":"Conde, M.V., et al.: Swin2SR: Swinv2 transformer for compressed image super-resolution and restoration. In: European conference on computer vision(ECCV). (2022)","DOI":"10.1007\/978-3-031-25063-7_42"},{"key":"3224_CR50","doi-asserted-by":"crossref","unstructured":"Chen, X., et al.: Activating more pixels in image super-resolution transformer. In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR). (2023)","DOI":"10.1109\/CVPR52729.2023.02142"},{"key":"3224_CR51","first-page":"19","volume":"19","author":"X Yang","year":"2023","unstructured":"Yang, X., et al.: HIFGAN: A high-frequency information-based generative adversarial network for image super-resolution. ACM Transact. Multimed. Comput. Commun. Appl. 19, 19 (2023)","journal-title":"ACM Transact. Multimed. Comput. Commun. Appl."},{"key":"3224_CR52","doi-asserted-by":"crossref","unstructured":"Jeevan, P., et al. WaveMixSR: Resource-efficient neural network for image super-resolution. In: Proceedings of the IEEE\/CVF winter conference on applications of computer vision. (2024)","DOI":"10.1109\/WACV57701.2024.00578"}],"container-title":["Signal, Image and Video Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-024-03224-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11760-024-03224-y\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-024-03224-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,7,15]],"date-time":"2024-07-15T08:24:07Z","timestamp":1721031847000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11760-024-03224-y"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,5,10]]},"references-count":52,"journal-issue":{"issue":"6-7","published-print":{"date-parts":[[2024,8]]}},"alternative-id":["3224"],"URL":"https:\/\/doi.org\/10.1007\/s11760-024-03224-y","relation":{},"ISSN":["1863-1703","1863-1711"],"issn-type":[{"value":"1863-1703","type":"print"},{"value":"1863-1711","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,5,10]]},"assertion":[{"value":"5 March 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"11 April 2024","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"15 April 2024","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"10 May 2024","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}