{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,3]],"date-time":"2026-02-03T06:34:13Z","timestamp":1770100453419,"version":"3.49.0"},"reference-count":57,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2026,2,2]],"date-time":"2026-02-02T00:00:00Z","timestamp":1769990400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,2,2]],"date-time":"2026-02-02T00:00:00Z","timestamp":1769990400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"DOI":"10.1007\/s11042-026-21213-w","type":"journal-article","created":{"date-parts":[[2026,2,2]],"date-time":"2026-02-02T18:26:29Z","timestamp":1770056789000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Improving generative adversarial network generalization for facialexpression synthesis"],"prefix":"10.1007","volume":"85","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-2661-4855","authenticated-orcid":false,"given":"Arbish","family":"Akram","sequence":"first","affiliation":[]},{"given":"Nazar","family":"Khan","sequence":"additional","affiliation":[]},{"given":"Arif","family":"Mahmood","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,2,2]]},"reference":[{"key":"21213_CR1","unstructured":"Mirza M, Osindero S (2014) Conditional generative adversarial nets. arXiv:1411.1784"},{"key":"21213_CR2","unstructured":"Karras T, Aila T, Laine S, Lehtinen J (2017) Progressive growing of gans for improved quality, stability, and variation. arXiv:1710.10196"},{"key":"21213_CR3","unstructured":"Ho J, Jain A, Abbeel P (2020) Denoising diffusion probabilistic models. 33:6840\u20136851"},{"key":"21213_CR4","doi-asserted-by":"crossref","unstructured":"Brooks T, Holynski A, Efros AA (2023) Instructpix2pix: learning to follow image editing instructions, pp 18392\u201318402","DOI":"10.1109\/CVPR52729.2023.01764"},{"key":"21213_CR5","doi-asserted-by":"crossref","unstructured":"Brack M, Friedrich F, Kornmeier K, Tsaban L, Schramowski P, Kersting K, Passos A (2024) LEDITS++: limitless image editing using text-to-image models, pp 8861\u20138870","DOI":"10.1109\/CVPR52733.2024.00846"},{"key":"21213_CR6","doi-asserted-by":"crossref","unstructured":"Karras T, Laine S, Aila T (2019) A style-based generator architecture for generative adversarial networks. In: IEEE conference on computer vision and pattern recognition, pp 4401\u20134410","DOI":"10.1109\/CVPR.2019.00453"},{"key":"21213_CR7","unstructured":"Karras T, Aittala M, Laine S, H\u00e4rk\u00f6nen E, Hellsten J, Lehtinen J, Aila T (2021) Alias-free generative adversarial networks. 34:852\u2013863"},{"issue":"15","key":"21213_CR8","doi-asserted-by":"publisher","first-page":"1454","DOI":"10.1073\/pnas.1322355111","volume":"111","author":"S Du","year":"2014","unstructured":"Du S, Tao Y, Martinez AM (2014) Compound facial expressions of emotion. Proc Natl Acad Sci 111(15):1454\u20131462","journal-title":"Proc Natl Acad Sci"},{"key":"21213_CR9","doi-asserted-by":"crossref","unstructured":"Isola P, Zhu J-Y, Zhou T, Efros AA (2017) Image-to-image translation with conditional adversarial networks. In: IEEE conference on computer vision and pattern recognition, pp 1125\u20131134","DOI":"10.1109\/CVPR.2017.632"},{"key":"21213_CR10","doi-asserted-by":"crossref","unstructured":"Zhu J-Y, Park T, Isola P, Efros AA (2017) Unpaired image-to-image translation using cycle-consistent adversarial networks. In: IEEE international conference on computer vision, pp 2223\u20132232","DOI":"10.1109\/ICCV.2017.244"},{"key":"21213_CR11","doi-asserted-by":"crossref","unstructured":"Choi Y, Choi M, Kim M, Ha J-W, Kim S, Choo J (2018) StarGAN: unified generative adversarial networks for multi-domain image-to-image translation. In: IEEE conference on computer vision and pattern recognition, pp 8789\u20138797","DOI":"10.1109\/CVPR.2018.00916"},{"key":"21213_CR12","unstructured":"Kim T, Cha M, Kim H, Lee JK, Kim J (2017) Learning to discover cross-domain relations with generative adversarial networks. PMLR, pp 1857\u20131865"},{"key":"21213_CR13","doi-asserted-by":"crossref","unstructured":"Huang R, Zhang S, Li T, He R (2017) Beyond face rotation: global and local perception gan for photorealistic and identity preserving frontal view synthesis. In: IEEE conference on computer vision and pattern recognition, pp 2439\u20132448","DOI":"10.1109\/ICCV.2017.267"},{"key":"21213_CR14","doi-asserted-by":"crossref","unstructured":"Di X, Sindagi VA, Patel VM (2018) GP-GAN: gender preserving gan for synthesizing faces from landmarks. In: International conference on pattern recognition. IEEE, pp 1079\u20131084","DOI":"10.1109\/ICPR.2018.8545081"},{"key":"21213_CR15","doi-asserted-by":"publisher","first-page":"4002","DOI":"10.1109\/TMM.2021.3111501","volume":"24","author":"J Huang","year":"2021","unstructured":"Huang J, Jing L, Tan Z, Kwong S (2021) Multi-density sketch-to-image translation network. IEEE Trans Multimedia 24:4002\u20134015","journal-title":"IEEE Trans Multimedia"},{"key":"21213_CR16","doi-asserted-by":"crossref","unstructured":"Rashid MM, Wu S, Nie Y, Li G (2023) High-fidelity facial expression transfer using part-based local-global conditional GANs. Vis Comp, 1\u201312","DOI":"10.1007\/s00371-023-03035-1"},{"key":"21213_CR17","doi-asserted-by":"publisher","first-page":"14981","DOI":"10.1007\/s11042-019-08346-x","volume":"79","author":"Y Zhao","year":"2020","unstructured":"Zhao Y, Zheng Z, Wang C, Gu Z, Fu M, Yu Z, Zheng H, Wang N, Zheng B (2020) Fine-grained facial image-to-image translation with an attention-based pipeline generative adversarial framework. Multimed Tools Appl 79:14981\u201315000","journal-title":"Multimed Tools Appl"},{"issue":"5","key":"21213_CR18","doi-asserted-by":"publisher","first-page":"1433","DOI":"10.1007\/s11263-019-01256-3","volume":"128","author":"N Khan","year":"2020","unstructured":"Khan N, Akram A, Mahmood A, Ashraf S, Murtaza K (2020) Masked linear regression for learning local receptive fields for facial expression synthesis. Int J Comput Vision 128(5):1433\u20131454","journal-title":"Int J Comput Vision"},{"issue":"3","key":"21213_CR19","doi-asserted-by":"publisher","first-page":"698","DOI":"10.1007\/s11263-019-01210-3","volume":"128","author":"A Pumarola","year":"2020","unstructured":"Pumarola A, Agudo A, Martinez AM, Sanfeliu A, Moreno-Noguer F (2020) GANimation: one-shot anatomically consistent facial animation. Int J Comput Vision 128(3):698\u2013713","journal-title":"Int J Comput Vision"},{"key":"21213_CR20","doi-asserted-by":"crossref","unstructured":"Wu R, Zhang G, Lu S, Chen T (2020) Cascade EF-GAN: progressive facial expression editing with local focuses. In: IEEE conference on computer vision and pattern recognition, pp 5021\u20135030","DOI":"10.1109\/CVPR42600.2020.00507"},{"key":"21213_CR21","doi-asserted-by":"crossref","unstructured":"Chen Y-C, Xu X, Jia J (2020) Domain adaptive image-to-image translation. In: IEEE conference on computer vision and pattern recognition, pp 5274\u20135283","DOI":"10.1109\/CVPR42600.2020.00532"},{"key":"21213_CR22","doi-asserted-by":"crossref","unstructured":"Akram A, Khan N (2023) US-GAN: on the importance of ultimate skip connection for facial expression synthesis. Multimed Tools Appl, 1\u20131","DOI":"10.1007\/s11042-023-15268-2"},{"key":"21213_CR23","doi-asserted-by":"crossref","unstructured":"Akram A, Khan N (2023) SARGAN: spatial attention-based residuals for facial expression manipulation. IEEE Trans Circuits Syst Video Technol, 1\u20131","DOI":"10.1109\/TCSVT.2023.3255243"},{"key":"21213_CR24","doi-asserted-by":"crossref","unstructured":"Herranz L, Jiang S, Li X (2016) Scene recognition with CNNs: objects, scales and dataset bias. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 571\u2013579","DOI":"10.1109\/CVPR.2016.68"},{"key":"21213_CR25","doi-asserted-by":"crossref","unstructured":"Torralba A, Efros AA (2011) Unbiased look at dataset bias. In: Computer Vision and Pattern Recognition (CVPR), 2011 IEEE Conference On. IEEE, pp 1521\u20131528","DOI":"10.1109\/CVPR.2011.5995347"},{"key":"21213_CR26","doi-asserted-by":"crossref","unstructured":"Wang T, Zhang Y, Fan Y, Wang J, Chen Q (2022) High-fidelity GAN inversion for image attribute editing, pp 11379\u201311388","DOI":"10.1109\/CVPR52688.2022.01109"},{"key":"21213_CR27","doi-asserted-by":"crossref","unstructured":"Hu X, Huang Q, Shi Z, Li S, Gao C, Sun L, Li Q (2022) Style transformer for image inversion and editing, pp 11337\u201311346","DOI":"10.1109\/CVPR52688.2022.01105"},{"key":"21213_CR28","doi-asserted-by":"crossref","unstructured":"Abdal R, Qin Y, Wonka P (2020) Image2StyleGAN++: how to edit the embedded images?, pp 8296\u20138305","DOI":"10.1109\/CVPR42600.2020.00832"},{"issue":"4","key":"21213_CR29","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3450626.3459838","volume":"40","author":"O Tov","year":"2021","unstructured":"Tov O, Alaluf Y, Nitzan Y, Patashnik O, Cohen-Or D (2021) Designing an encoder for StyleGAN image manipulation. ACM Trans Graphics (TOG) 40(4):1\u201314","journal-title":"ACM Trans Graphics (TOG)"},{"key":"21213_CR30","doi-asserted-by":"crossref","unstructured":"Alaluf Y, Tov O, Mokady R, Gal R, Bermano A (2022) HyperStyle: StyleGAN inversion with hypernetworks for real image editing, pp 18511\u201318521","DOI":"10.1109\/CVPR52688.2022.01796"},{"key":"21213_CR31","doi-asserted-by":"crossref","unstructured":"Pehlivan H, Dalva Y, Dundar A (2023) StyleRes: transforming the residuals for real image editing with StyleGAN, pp 1828\u20131837","DOI":"10.1109\/CVPR52729.2023.00182"},{"key":"21213_CR32","doi-asserted-by":"crossref","unstructured":"Bobkov D, Titov V, Alanov A, Vetrov D (2024) The devil is in the details: StyleFeatureEditor for detail-rich StyleGAN inversion and high quality image editing, pp 9337\u20139346","DOI":"10.1109\/CVPR52733.2024.00892"},{"key":"21213_CR33","doi-asserted-by":"crossref","unstructured":"Akram A, Khan N (2021) Pixel-based facial expression synthesis. In: International conference on pattern recognition","DOI":"10.1109\/ICPR48806.2021.9413065"},{"key":"21213_CR34","doi-asserted-by":"crossref","unstructured":"Akram A, Khan N (2023) LSRF: localized and sparse receptive fields for linear facial expression synthesis based on global face context. Multimed Tools Appl, 1\u20131","DOI":"10.1007\/s11042-023-16822-8"},{"issue":"15","key":"21213_CR35","doi-asserted-by":"publisher","first-page":"1454","DOI":"10.1073\/pnas.1322355111","volume":"111","author":"S Du","year":"2014","unstructured":"Du S, Tao Y, Martinez AM (2014) Compound facial expressions of emotion. Proc Natl Acad Sci 111(15):1454\u20131462","journal-title":"Proc Natl Acad Sci"},{"key":"21213_CR36","unstructured":"Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, Courville A, Bengio Y (2014) Generative adversarial nets. In: Advances in neural information processing systems, pp 2672\u20132680"},{"key":"21213_CR37","doi-asserted-by":"crossref","unstructured":"Newell A, Yang K, Deng J (2016) Stacked hourglass networks for human pose estimation. In: European conference on computer vision. Springer, pp 483\u2013499","DOI":"10.1007\/978-3-319-46484-8_29"},{"key":"21213_CR38","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1016\/j.neucom.2021.03.048","volume":"446","author":"J Kim","year":"2021","unstructured":"Kim J, Li G, Yun I, Jung C, Kim J (2021) Edge and identity preserving network for face super-resolution. Neurocomputing 446:11\u201322","journal-title":"Neurocomputing"},{"key":"21213_CR39","doi-asserted-by":"crossref","unstructured":"Yu X, Fernando B, Ghanem B, Porikli F, Hartley R (2018) Face super-resolution guided by facial component heatmaps. In: European conference on computer vision, pp 217\u2013233","DOI":"10.1007\/978-3-030-01240-3_14"},{"key":"21213_CR40","doi-asserted-by":"crossref","unstructured":"Bulat A, Tzimiropoulos G (2017) How far are we from solving the 2D & 3D Face Alignment problem? (and a dataset of 230,000 3D facial landmarks). In: IEEE international conference on computer vision, pp 1021\u20131030","DOI":"10.1109\/ICCV.2017.116"},{"key":"21213_CR41","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, Sun J (2015) Delving deep into rectifiers: surpassing human-level performance on ImageNet classification. In: IEEE conference on computer vision and pattern recognition, pp 1026\u20131034","DOI":"10.1109\/ICCV.2015.123"},{"key":"21213_CR42","doi-asserted-by":"publisher","first-page":"1219","DOI":"10.1109\/TIP.2020.3043093","volume":"30","author":"C Chen","year":"2020","unstructured":"Chen C, Gong D, Wang H, Li Z, Wong K-YK (2020) Learning spatial attention for face super-resolution. IEEE Trans Image Process 30:1219\u20131231","journal-title":"IEEE Trans Image Process"},{"key":"21213_CR43","doi-asserted-by":"crossref","unstructured":"Wang T-C, Liu M-Y, Zhu J-Y, Tao A, Kautz J, Catanzaro B (2018) High-resolution image synthesis and semantic manipulation with conditional GANs. In: IEEE conference on computer vision and pattern recognition, pp 8798\u20138807","DOI":"10.1109\/CVPR.2018.00917"},{"key":"21213_CR44","unstructured":"Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. arXiv:1409.1556"},{"key":"21213_CR45","doi-asserted-by":"crossref","unstructured":"Lee C-H, Liu Z, Wu L, Luo P (2020) MaskGAN: towards diverse and interactive facial image manipulation. In: IEEE conference on computer vision and pattern recognition, pp 5549\u20135558","DOI":"10.1109\/CVPR42600.2020.00559"},{"key":"21213_CR46","doi-asserted-by":"crossref","unstructured":"Nirkin Y, Keller Y, Hassner T (2019) FSGAN: subject agnostic face swapping and reenactment. In: IEEE conference on computer vision and pattern recognition, pp 7184\u20137193","DOI":"10.1109\/ICCV.2019.00728"},{"key":"21213_CR47","doi-asserted-by":"crossref","unstructured":"Chen C, Li X, Yang L, Lin X, Zhang L, Wong K-YK (2021) Progressive semantic-aware style transformation for blind face restoration. In: IEEE conference on computer vision and pattern recognition, pp 11896\u201311905","DOI":"10.1109\/CVPR46437.2021.01172"},{"key":"21213_CR48","unstructured":"Lim JH, Ye JC (2017) Geometric GAN. arXiv:1705.02894"},{"key":"21213_CR49","doi-asserted-by":"crossref","unstructured":"Rombach R, Blattmann A, Lorenz D, Esser P, Ommer B (2022) High-resolution image synthesis with latent diffusion models. In: IEEE conference on computer vision and pattern recognition, pp 10684\u201310695","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"21213_CR50","unstructured":"Kingma DP, Ba J (2014) Adam: a method for stochastic optimization. arXiv:1412.6980"},{"key":"21213_CR51","doi-asserted-by":"crossref","unstructured":"Afifi M, Brubaker MA, Brown MS (2021) HistoGAN: controlling colors of GAN-generated and real images via color histograms. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 7941\u20137950","DOI":"10.1109\/CVPR46437.2021.00785"},{"key":"21213_CR52","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: IEEE conference on computer vision and pattern recognition, pp 770\u2013778","DOI":"10.1109\/CVPR.2016.90"},{"key":"21213_CR53","unstructured":"Amos B, Ludwiczuk B, Satyanarayanan M (2016) OpenFace: a general-purpose face recognition library with mobile applications. Technical report, CMU-CS-16-118, CMU School of Computer Science"},{"key":"21213_CR54","unstructured":"Agnolucci L, Galteri L, Bertini M (2024) Quality-aware image-text alignment for real-world image quality assessment. arXiv:2403.11176"},{"key":"21213_CR55","unstructured":"Radford A, Kim JW, Hallacy C, Ramesh A, Goh G, Agarwal S, Sastry G, Askell A, Mishkin P, Clark J et al (2021) Learning transferable visual models from natural language supervision. PMLR, pp 8748\u20138763"},{"key":"21213_CR56","unstructured":"Heusel M, Ramsauer H, Unterthiner T, Nessler B, Hochreiter S (2017) GANs trained by a two time-scale update rule converge to a local nash equilibrium. 30"},{"key":"21213_CR57","doi-asserted-by":"crossref","unstructured":"Szegedy C, Vanhoucke V, Ioffe S, Shlens J, Wojna Z (2016) Rethinking the inception architecture for computer vision, pp 2818\u20132826","DOI":"10.1109\/CVPR.2016.308"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-026-21213-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-026-21213-w","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-026-21213-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,2,2]],"date-time":"2026-02-02T18:26:36Z","timestamp":1770056796000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-026-21213-w"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,2,2]]},"references-count":57,"journal-issue":{"issue":"2","published-online":{"date-parts":[[2026,2]]}},"alternative-id":["21213"],"URL":"https:\/\/doi.org\/10.1007\/s11042-026-21213-w","relation":{},"ISSN":["1573-7721"],"issn-type":[{"value":"1573-7721","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,2,2]]},"assertion":[{"value":"30 October 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"4 October 2025","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"13 November 2025","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"2 February 2026","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"Not applicable.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical Approval"}},{"value":"Not applicable.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent to Participate"}},{"value":"Not applicable.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent to Publish"}},{"value":"Not applicable. The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":5,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing Interests"}}],"article-number":"89"}}