{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,6]],"date-time":"2026-05-06T16:09:03Z","timestamp":1778083743749,"version":"3.51.4"},"reference-count":40,"publisher":"MDPI AG","issue":"4","license":[{"start":{"date-parts":[[2025,4,11]],"date-time":"2025-04-11T00:00:00Z","timestamp":1744329600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"name":"National Natural Science Foundation of China","award":["62062048"],"award-info":[{"award-number":["62062048"]}]},{"name":"National Natural Science Foundation of China","award":["62263017"],"award-info":[{"award-number":["62263017"]}]},{"name":"National Natural Science Foundation of China","award":["202201AT070113"],"award-info":[{"award-number":["202201AT070113"]}]},{"name":"Yunnan Department of Science and Technology Project","award":["62062048"],"award-info":[{"award-number":["62062048"]}]},{"name":"Yunnan Department of Science and Technology Project","award":["62263017"],"award-info":[{"award-number":["62263017"]}]},{"name":"Yunnan Department of Science and Technology Project","award":["202201AT070113"],"award-info":[{"award-number":["202201AT070113"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Entropy"],"abstract":"<jats:p>Single-image super-resolution (SISR) based on GANs has achieved significant progress. However, these methods still face challenges when reconstructing locally consistent textures due to a lack of semantic understanding of image categories. This highlights the necessity of focusing on contextual information comprehension and the acquisition of high-frequency details in model design. To address this issue, we propose the Semantic Prior-Improved GAN (SP-IGAN) framework, which incorporates additional contextual semantic information into the Real-ESRGAN model. The framework consists of two branches. The main branch introduces a Graph Convolutional Channel Attention (GCCA) module to transform channel dependencies into adjacency relationships between feature vertices, thereby enhancing pixel associations. The auxiliary branch strengthens the correlation between semantic category information and regional textures in the Residual-in-Residual Dense Block (RRDB) module. The auxiliary branch employs a pretrained segmentation model to accurately extract regional semantic information from the input low-resolution image. This information is injected into the RRDB module through Spatial Feature Transform (SFT) layers, generating more accurate and semantically consistent texture details. Additionally, a wavelet loss is incorporated into the loss function to capture high-frequency details that are often overlooked. The experimental results demonstrate that the proposed SP-IGAN outperforms state-of-the-art (SOTA) super-resolution models across multiple public datasets. For the X4 super-resolution task, SP-IGAN achieves a 0.55 dB improvement in Peak Signal-to-Noise Ratio (PSNR) and a 0.0363 increase in Structural Similarity Index (SSIM) compared to the baseline model Real-ESRGAN.<\/jats:p>","DOI":"10.3390\/e27040414","type":"journal-article","created":{"date-parts":[[2025,4,11]],"date-time":"2025-04-11T05:38:26Z","timestamp":1744349906000},"page":"414","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["SP-IGAN: An Improved GAN Framework for Effective Utilization of Semantic Priors in Real-World Image Super-Resolution"],"prefix":"10.3390","volume":"27","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-6245-2673","authenticated-orcid":false,"given":"Meng","family":"Wang","sequence":"first","affiliation":[{"name":"Faculty of Information Engineering and Automation, Kunming University of Science and Technology, Kunming 650500, China"}]},{"given":"Zhengnan","family":"Li","sequence":"additional","affiliation":[{"name":"Faculty of Information Engineering and Automation, Kunming University of Science and Technology, Kunming 650500, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7478-773X","authenticated-orcid":false,"given":"Haipeng","family":"Liu","sequence":"additional","affiliation":[{"name":"Faculty of Information Engineering and Automation, Kunming University of Science and Technology, Kunming 650500, China"},{"name":"Yunnan Province Key Laboratory of Computer, Kunming University of Science and Technology, Kunming 650500, China"}]},{"given":"Zhaoyu","family":"Chen","sequence":"additional","affiliation":[{"name":"Faculty of Information Engineering and Automation, Kunming University of Science and Technology, Kunming 650500, China"}]},{"given":"Kewei","family":"Cai","sequence":"additional","affiliation":[{"name":"Faculty of Information Engineering and Automation, Kunming University of Science and Technology, Kunming 650500, China"}]}],"member":"1968","published-online":{"date-parts":[[2025,4,11]]},"reference":[{"key":"ref_1","doi-asserted-by":"crossref","unstructured":"Al-Mekhlafi, H., and Liu, S. (2024). Single image super-resolution: A comprehensive review and recent insight. Front. Comput. Sci., 18.","DOI":"10.1007\/s11704-023-2588-9"},{"key":"ref_2","doi-asserted-by":"crossref","first-page":"104110","DOI":"10.1016\/j.earscirev.2022.104110","article-title":"A comprehensive review on deep learning based remote sensing image super-resolution methods","volume":"232","author":"Wang","year":"2022","journal-title":"Earth Sci. Rev."},{"key":"ref_3","first-page":"168","article-title":"Super Resolution with Interpolation-based Method: A Review","volume":"9","author":"Azam","year":"2022","journal-title":"IJRAR Int. J. Res. Anal. Rev. (IJRAR)"},{"key":"ref_4","doi-asserted-by":"crossref","unstructured":"Ooi, Y.K., and Ibrahim, H. (2021). Deep Learning Algorithms for Single Image Super-Resolution: A Systematic Review. Electronics, 10.","DOI":"10.3390\/electronics10070867"},{"key":"ref_5","first-page":"1977","article-title":"Image super-resolution based on generative adversarial networks: A brief review","volume":"64","author":"Fu","year":"2020","journal-title":"Comput. Mater. Contin."},{"key":"ref_6","doi-asserted-by":"crossref","first-page":"295","DOI":"10.1109\/TPAMI.2015.2439281","article-title":"Image Super-Resolution Using Deep Convolutional Networks","volume":"38","author":"Dong","year":"2015","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"ref_7","doi-asserted-by":"crossref","unstructured":"Kim, J., Lee, J.K., and Lee, K.M. (2016, January 27\u201330). Accurate Image Super-Resolution Using Very Deep Convolutional Networks. Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA.","DOI":"10.1109\/CVPR.2016.182"},{"key":"ref_8","doi-asserted-by":"crossref","unstructured":"Lim, B., Son, S., Kim, H., Nah, S., and Lee, K.M. (2017, January 21\u201326). Enhanced Deep Residual Networks for Single Image Super-Resolution. Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Honolulu, HI, USA.","DOI":"10.1109\/CVPRW.2017.151"},{"key":"ref_9","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Li, K., Li, K., Wang, L., Zhong, B., and Fu, Y. (2018, January 8\u201314). Image Super-Resolution Using Very Deep Residual Channel Attention Networks. Proceedings of the 2018 European Conference on Computer Vision (ECCV), Munich, Germany.","DOI":"10.1007\/978-3-030-01234-2_18"},{"key":"ref_10","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Tian, Y., Kong, Y., Zhong, B., and Fu, Y. (2018, January 18\u201323). Residual Dense Network for Image Super-Resolution. Proceedings of the 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA.","DOI":"10.1109\/CVPR.2018.00262"},{"key":"ref_11","unstructured":"Li, X., Ren, Y., Jin, X., Lan, C., Wang, X., Zeng, W., Wang, X., and Chen, Z. (2023). Diffusion Models for Image Restoration and Enhancement\u2014A Comprehensive Survey. arXiv."},{"key":"ref_12","unstructured":"Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., and Bengio, Y. (2014). Generative Adversarial Nets. Advances in Neural Information Processing Systems 27 (NIPS 2014), MIT Press."},{"key":"ref_13","doi-asserted-by":"crossref","unstructured":"Ledig, C., Theis, L., Husz\u00e1r, F., Caballero, J., Cunningham, A., Acosta, A., Aitken, A., Tejani, A., Totz, J., and Wang, Z. (2017, January 21\u201326). Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network. Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA.","DOI":"10.1109\/CVPR.2017.19"},{"key":"ref_14","doi-asserted-by":"crossref","unstructured":"Wang, X., Yu, K., Wu, S., Gu, J., Liu, Y., Dong, C., Qiao, Y., and Loy, C.C. (2018, January 8\u201314). ESRGAN: Enhanced Super-Resolution Generative Adversarial Networks. Proceedings of the 2018 European Conference on Computer Vision (ECCV) Workshops, Munich, Germany.","DOI":"10.1007\/978-3-030-11021-5_5"},{"key":"ref_15","doi-asserted-by":"crossref","unstructured":"Zhang, W., Liu, Y., Dong, C., and Qiao, Y. (2019, January 27\u201328). RANKSRGAN: Generative Adversarial Networks with Ranker for Image Super-Resolution. Proceedings of the 2019 IEEE\/CVF International Conference on Computer Vision Workshop (ICCVW), Seoul, Republic of Korea.","DOI":"10.1109\/ICCV.2019.00319"},{"key":"ref_16","doi-asserted-by":"crossref","unstructured":"Schonfeld, E., Schiele, B., and Khoreva, A. (2020, January 13\u201319). A U-Net Based Discriminator for Generative Adversarial Networks. Proceedings of the 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA.","DOI":"10.1109\/CVPR42600.2020.00823"},{"key":"ref_17","doi-asserted-by":"crossref","unstructured":"Zhang, K., Liang, J., Van Gool, L., and Timofte, R. (2021, January 11\u201317). Designing a Practical Degradation Model for Deep Blind Image Super-Resolution. Proceedings of the 2021 IEEE\/CVF International Conference on Computer Vision, Montreal, BC, Canada.","DOI":"10.1109\/ICCV48922.2021.00475"},{"key":"ref_18","doi-asserted-by":"crossref","unstructured":"Fritsche, M., Gu, S., and Timofte, R. (2019, January 27\u201328). Frequency Separation for Real-World Super-Resolution. Proceedings of the 2019 IEEE\/CVF International Conference on Computer Vision Workshop (ICCVW), Seoul, Republic of Korea.","DOI":"10.1109\/ICCVW.2019.00445"},{"key":"ref_19","doi-asserted-by":"crossref","unstructured":"Wang, X., Xie, L., Dong, C., and Shan, Y. (2021, January 11\u201317). Real-ESRGAN: Training Real-World Blind Super-Resolution with Pure Synthetic Data. Proceedings of the 2021 IEEE\/CVF International Conference on Computer Vision, Montreal, BC, Canada.","DOI":"10.1109\/ICCVW54120.2021.00217"},{"key":"ref_20","doi-asserted-by":"crossref","unstructured":"Wang, L., Wang, Y., Dong, X., Xu, Q., Yang, J., An, W., and Guo, Y. (2021, January 20\u201325). Unsupervised Degradation Representation Learning for Blind Super-Resolution. Proceedings of the 2021 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA.","DOI":"10.1109\/CVPR46437.2021.01044"},{"key":"ref_21","doi-asserted-by":"crossref","unstructured":"Vo, K.D., and Bui, L.T. (2023). StarSRGAN: Improving Real-World Blind Super-Resolution. arXiv.","DOI":"10.24132\/CSRN.3301.9"},{"key":"ref_22","doi-asserted-by":"crossref","unstructured":"Rakotonirina, N.C., and Rasoanaivo, A. (2020, January 4\u20138). ESRGAN+: Further Improving Enhanced Super-Resolution Generative Adversarial Network. Proceedings of the ICASSP 2020\u20142020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain.","DOI":"10.1109\/ICASSP40776.2020.9054071"},{"key":"ref_23","doi-asserted-by":"crossref","unstructured":"Korkmaz, C., and Tekalp, A.M. (2024). Training Transformer Models by Wavelet Losses Improves Quantitative and Visual Performance in Single Image Super-Resolution. arXiv.","DOI":"10.1109\/CVPRW63382.2024.00660"},{"key":"ref_24","doi-asserted-by":"crossref","first-page":"48","DOI":"10.1016\/j.neucom.2021.03.091","article-title":"A Review on the Attention Mechanism of Deep Learning","volume":"452","author":"Niu","year":"2021","journal-title":"Neurocomputing"},{"key":"ref_25","doi-asserted-by":"crossref","unstructured":"Hu, J., Shen, L., and Sun, G. (2018, January 18\u201323). Squeeze-and-Excitation Networks. Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA.","DOI":"10.1109\/CVPR.2018.00745"},{"key":"ref_26","unstructured":"Buades, A., Coll, B., and Morel, J.M. (2005, January 20\u201325). A non-local algorithm for image denoising. Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR\u201905), San Diego, CA, USA."},{"key":"ref_27","doi-asserted-by":"crossref","unstructured":"Woo, S., Park, J., Lee, J.Y., and Kweon, I.S. (2018, January 8\u201314). CBAM: Convolutional Block Attention Module. Proceedings of the 2018 European Conference on Computer Vision (ECCV), Munich, Germany.","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"ref_28","doi-asserted-by":"crossref","unstructured":"Dai, T., Cai, J., Zhang, Y., Xia, S.T., and Zhang, L. (2019, January 15\u201320). Second-Order Attention Network for Single Image Super-Resolution. Proceedings of the 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA.","DOI":"10.1109\/CVPR.2019.01132"},{"key":"ref_29","doi-asserted-by":"crossref","unstructured":"Wang, X., Yu, K., Dong, C., and Loy, C.C. (2018, January 18\u201323). Recovering Realistic Texture in Image Super-Resolution by Deep Spatial Feature Transform. Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA.","DOI":"10.1109\/CVPR.2018.00070"},{"key":"ref_30","doi-asserted-by":"crossref","unstructured":"Park, S.H., Moon, Y.S., and Cho, N.I. (2023, January 17\u201324). Perception-Oriented Single Image Super-Resolution Using Optimal Objective Estimation. Proceedings of the 2023 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada.","DOI":"10.1109\/CVPR52729.2023.00172"},{"key":"ref_31","doi-asserted-by":"crossref","first-page":"9774","DOI":"10.1109\/ACCESS.2022.3144406","article-title":"Flexible Style Image Super-Resolution Using Conditional Objective","volume":"10","author":"Park","year":"2022","journal-title":"IEEE Access"},{"key":"ref_32","doi-asserted-by":"crossref","unstructured":"Ma, C., Rao, Y., Cheng, Y., Chen, C., Lu, J., and Zhou, J. (2020, January 13\u201319). Structure-preserving super resolution with gradient guidance. Proceedings of the 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA.","DOI":"10.1109\/CVPR42600.2020.00779"},{"key":"ref_33","doi-asserted-by":"crossref","unstructured":"Li, B., Li, X., Zhu, H., Jin, Y., Feng, R., Zhang, Z., and Chen, Z. (2024). SeD: Semantic-Aware Discriminator for Image Super-Resolution. arXiv.","DOI":"10.1109\/CVPR52733.2024.02436"},{"key":"ref_34","doi-asserted-by":"crossref","first-page":"5008812","DOI":"10.1109\/TIM.2023.3248111","article-title":"AGCA: An Adaptive Graph Channel Attention Module for Steel Surface Defect Detection","volume":"72","author":"Xiang","year":"2023","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"ref_35","doi-asserted-by":"crossref","unstructured":"Chen, X., Wang, X., Zhou, J., Qiao, Y., and Dong, C. (2023, January 17\u201324). Activating More Pixels in Image Super-Resolution Transformer. Proceedings of the 2023 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada.","DOI":"10.1109\/CVPR52729.2023.02142"},{"key":"ref_36","doi-asserted-by":"crossref","unstructured":"Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., Erhan, D., Vanhoucke, V., and Rabinovich, A. (2015, January 7\u201312). Going Deeper with Convolutions. Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2015), Boston, MA, USA.","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"ref_37","doi-asserted-by":"crossref","unstructured":"Zhou, D., Hou, Q., Chen, Y., Feng, J., and Yan, S. (2020). Rethinking Bottleneck Structure for Efficient Mobile Network Design. Computer Vision\u2014ECCV 2020, Proceedings of the 16th European Conference, Glasgow, UK, 23\u201328 August 2020, Springer International Publishing.","DOI":"10.1007\/978-3-030-58580-8_40"},{"key":"ref_38","unstructured":"Lin, M. (2013). Network in Network. arXiv."},{"key":"ref_39","unstructured":"Kipf, T.N., and Welling, M. (2016). Semi-Supervised Classification with Graph Convolutional Networks. arXiv."},{"key":"ref_40","unstructured":"Aghelan, A., Amiryan, A., Zarghani, A., and Rouhani, M. (2024). IG-CFAT: An Improved GAN-Based Framework for Effectively Exploiting Transformers in Real-World Image Super-Resolution. arXiv."}],"container-title":["Entropy"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/1099-4300\/27\/4\/414\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,9]],"date-time":"2025-10-09T17:12:42Z","timestamp":1760029962000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/1099-4300\/27\/4\/414"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,4,11]]},"references-count":40,"journal-issue":{"issue":"4","published-online":{"date-parts":[[2025,4]]}},"alternative-id":["e27040414"],"URL":"https:\/\/doi.org\/10.3390\/e27040414","relation":{},"ISSN":["1099-4300"],"issn-type":[{"value":"1099-4300","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,4,11]]}}}