{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,18]],"date-time":"2026-03-18T16:10:10Z","timestamp":1773850210405,"version":"3.50.1"},"reference-count":36,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2023,6,6]],"date-time":"2023-06-06T00:00:00Z","timestamp":1686009600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,6,6]],"date-time":"2023-06-06T00:00:00Z","timestamp":1686009600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"published-print":{"date-parts":[[2024,1]]},"DOI":"10.1007\/s11042-023-15268-2","type":"journal-article","created":{"date-parts":[[2023,6,6]],"date-time":"2023-06-06T18:03:21Z","timestamp":1686074601000},"page":"7231-7247","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":13,"title":["US-GAN: on the importance of ultimate skip connection for facial expression synthesis"],"prefix":"10.1007","volume":"83","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-2661-4855","authenticated-orcid":false,"given":"Arbish","family":"Akram","sequence":"first","affiliation":[]},{"given":"Nazar","family":"Khan","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,6,6]]},"reference":[{"key":"15268_CR1","doi-asserted-by":"crossref","unstructured":"Akram A, Khan N (2021) Pixel-based facial expression synthesis. In: International conference on pattern recognition. IEEE, pp 9733\u20139739","DOI":"10.1109\/ICPR48806.2021.9413065"},{"key":"15268_CR2","unstructured":"Arjovsky M, Chintala S, Bottou L (2017) Wasserstein generative adversarial networks. In: International conference on machine learning, pp 214\u2013223"},{"key":"15268_CR3","doi-asserted-by":"publisher","first-page":"1219","DOI":"10.1109\/TIP.2020.3043093","volume":"30","author":"C Chen","year":"2020","unstructured":"Chen C, Gong D, Wang H, Li Z, Wong K-Y K (2020) Learning spatial attention for face super-resolution. IEEE Trans Image Process 30:1219\u20131231","journal-title":"IEEE Trans Image Process"},{"key":"15268_CR4","doi-asserted-by":"crossref","unstructured":"Chen Y-C, Xu X, Jia J (2020) Domain adaptive image-to-image translation. In: IEEE Conference on computer vision and pattern recognition, pp 5274\u20135283","DOI":"10.1109\/CVPR42600.2020.00532"},{"key":"15268_CR5","doi-asserted-by":"crossref","unstructured":"Chen Y, Tai Y, Liu X, Shen C, Yang J (2018) FSRNet: end-to-end learning face super-resolution with facial priors. In: IEEE Conference on computer vision and pattern recognition, pp 2492\u20132501","DOI":"10.1109\/CVPR.2018.00264"},{"key":"15268_CR6","doi-asserted-by":"crossref","unstructured":"Choi Y, Choi M, Kim M, Ha J-W, Kim S, Choo J (2018) StarGAN: unified generative adversarial networks for multi-domain image-to-image translation. In: IEEE Conference on computer vision and pattern recognition, pp 8789\u20138797","DOI":"10.1109\/CVPR.2018.00916"},{"key":"15268_CR7","doi-asserted-by":"crossref","unstructured":"d\u2019Apolito S, Paudel D P, Huang Z, Romero A, Van Gool L (2021) GANmut: learning interpretable conditional space for gamut of emotions. In: IEEE Conference on computer vision and pattern recognition, pp 568\u2013577","DOI":"10.1109\/CVPR46437.2021.00063"},{"key":"15268_CR8","doi-asserted-by":"crossref","unstructured":"Ding H, Sricharan K, Chellappa R (2018) ExprGAN: facial expression editing with controllable expression intensity. In: Proceedings of the AAAI conference on artificial intelligence, vol 32","DOI":"10.1609\/aaai.v32i1.12277"},{"issue":"15","key":"15268_CR9","doi-asserted-by":"publisher","first-page":"E1454","DOI":"10.1073\/pnas.1322355111","volume":"111","author":"S Du","year":"2014","unstructured":"Du S, Tao Y, Martinez A M (2014) Compound facial expressions of emotion. Proc Natl Acad Sci 111(15):E1454\u2013E1462","journal-title":"Proc Natl Acad Sci"},{"key":"15268_CR10","doi-asserted-by":"crossref","unstructured":"Fabian Benitez-Quiroz C, Srinivasan R, Martinez A M (2016) EmotioNet: an accurate, real-time algorithm for the automatic annotation of a million facial expressions in the wild. In: IEEE Conference on computer vision and pattern recognition, pp 5562\u20135570","DOI":"10.1109\/CVPR.2016.600"},{"issue":"3","key":"15268_CR11","doi-asserted-by":"publisher","first-page":"121","DOI":"10.1007\/BF00342633","volume":"20","author":"K Fukushima","year":"1975","unstructured":"Fukushima K (1975) Cognitron: a self-organizing multilayered neural network. Biol Cybern 20(3):121\u2013136","journal-title":"Biol Cybern"},{"key":"15268_CR12","doi-asserted-by":"crossref","unstructured":"Gao Y, Wei F, Bao J, Gu S, Chen D, Wen F, Lian Z (2021) High-fidelity and arbitrary face editing. In: IEEE Conference on computer vision and pattern recognition, pp 16115\u201316124","DOI":"10.1109\/CVPR46437.2021.01585"},{"key":"15268_CR13","unstructured":"Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, Courville A, Bengio Y (2014) Generative adversarial nets. In: Advances in neural information processing systems, pp 2672\u20132680"},{"key":"15268_CR14","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: IEEE Conference on computer vision and pattern recognition, pp 770\u2013778","DOI":"10.1109\/CVPR.2016.90"},{"key":"15268_CR15","doi-asserted-by":"crossref","unstructured":"Huang G, Liu Z, van der Maaten L, Weinberger K Q (2017) Densely connected convolutional networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition","DOI":"10.1109\/CVPR.2017.243"},{"key":"15268_CR16","doi-asserted-by":"crossref","unstructured":"Isola P, Zhu J-Y, Zhou T, Efros A A (2017) Image-to-image translation with conditional adversarial networks. In: IEEE Conference on computer vision and pattern recognition, pp 1125\u20131134","DOI":"10.1109\/CVPR.2017.632"},{"key":"15268_CR17","doi-asserted-by":"crossref","unstructured":"Johnson J, Alahi A, Fei-Fei L (2016) Perceptual losses for real-time style transfer and super-resolution. In: European Conference on computer vision. Springer, pp 694\u2013711","DOI":"10.1007\/978-3-319-46475-6_43"},{"issue":"5","key":"15268_CR18","doi-asserted-by":"publisher","first-page":"1433","DOI":"10.1007\/s11263-019-01256-3","volume":"128","author":"N Khan","year":"2020","unstructured":"Khan N, Akram A, Mahmood A, Ashraf S, Murtaza K (2020) Masked linear regression for learning local receptive fields for facial expression synthesis. Int J Comput Vis 128(5):1433\u20131454","journal-title":"Int J Comput Vis"},{"key":"15268_CR19","unstructured":"Kingma D P, Ba J (2014) Adam: a method for stochastic optimization. arXiv:1412.6980"},{"issue":"8","key":"15268_CR20","doi-asserted-by":"publisher","first-page":"1377","DOI":"10.1080\/02699930903485076","volume":"24","author":"O Langner","year":"2010","unstructured":"Langner O, Dotsch R, Bijlstra G, Wigboldus Daniel HJ, Hawk S T, Van Knippenberg AD (2010) Presentation and validation of the Radboud Faces Database. Cogn Emotion 24(8):1377\u20131388","journal-title":"Cogn Emotion"},{"key":"15268_CR21","doi-asserted-by":"crossref","unstructured":"Liu M, Ding Y, Xia M, Liu X, Ding E, Zuo W, Wen S (2019) Stgan: a unified selective transfer network for arbitrary image attribute editing. In: IEEE Conference on computer vision and pattern recognition, pp 3673\u20133682","DOI":"10.1109\/CVPR.2019.00379"},{"key":"15268_CR22","volume-title":"The karolinska directed emotional faces - KDEF, CD ROM","author":"D Lundqvist","year":"1998","unstructured":"Lundqvist D, Flykt A, \u00d6hman A (1998) The karolinska directed emotional faces - KDEF, CD ROM. Department of Clinical Neuroscience, Psychology section, Karolinska Institutet, Stockholm, Sweden"},{"key":"15268_CR23","first-page":"2802","volume":"29","author":"X Mao","year":"2016","unstructured":"Mao X, Shen C, Yang Y-B (2016) Image restoration using very deep convolutional encoder-decoder networks with symmetric skip connections. Adv Neur Inform Process Syst 29:2802\u20132810","journal-title":"Adv Neur Inform Process Syst"},{"key":"15268_CR24","unstructured":"Mirza M, Osindero S (2014) Conditional generative adversarial nets. ArXiv:1411.1784"},{"issue":"1","key":"15268_CR25","doi-asserted-by":"publisher","first-page":"18","DOI":"10.1109\/TAFFC.2017.2740923","volume":"10","author":"A Mollahosseini","year":"2017","unstructured":"Mollahosseini A, Hasani B, Mahoor M H (2017) AffectNet: a database for facial expression, valence, and arousal computing in the wild. IEEE Trans Affect Comput 10(1):18\u201331","journal-title":"IEEE Trans Affect Comput"},{"key":"15268_CR26","unstructured":"Perarnau G, van de Weijer J, Raducanu B, \u00c1lvarez J M (2016) Invertible conditional GANs for image editing. ArXiv:1611.06355"},{"issue":"3","key":"15268_CR27","doi-asserted-by":"publisher","first-page":"698","DOI":"10.1007\/s11263-019-01210-3","volume":"128","author":"A Pumarola","year":"2020","unstructured":"Pumarola A, Agudo A, Martinez A M, Sanfeliu A, Moreno-Noguer F (2020) GANimation: one-shot anatomically consistent facial animation. Int J Comput Vis 128(3):698\u2013713","journal-title":"Int J Comput Vis"},{"key":"15268_CR28","unstructured":"Qiao F, Yao N, Jiao Z, Li Z, Chen H, Wang H (2018) Geometry-contrastive generative adversarial network for facial expression synthesis. ArXiv:1802.01822"},{"key":"15268_CR29","doi-asserted-by":"crossref","unstructured":"Ronneberger O, Fischer P, Brox T (2015) U-Net: convolutional networks for biomedical image segmentation. In: International conference on medical image computing and computer-assisted intervention. Springer, pp 234\u2013241","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"15268_CR30","doi-asserted-by":"crossref","unstructured":"Shen W, Liu R (2017) Learning residual images for face attribute manipulation. In: IEEE Conference on computer vision and pattern recognition, pp 4030\u20134038","DOI":"10.1109\/CVPR.2017.135"},{"key":"15268_CR31","doi-asserted-by":"crossref","unstructured":"Tang J, Shao Z, Ma L (2021) EGGAN: learning latent space for fine-grained expression manipulation. IEEE MultiMedia","DOI":"10.1109\/ICME46284.2020.9102852"},{"key":"15268_CR32","unstructured":"Ulyanov D, Vedaldi A, Lempitsky V (2016) Instance normalization: the missing ingredient for fast stylization. ArXiv:1607.08022"},{"key":"15268_CR33","unstructured":"Wu P-W, Lin Y-J, Chang C-H, Chang E Y, Liao S-W (2019) RelGAN: multi-domain image-to-image translation via relative attributes. In: IEEE International conference on computer vision, pp 5914\u20135922"},{"key":"15268_CR34","doi-asserted-by":"crossref","unstructured":"Wu R, Zhang G, Lu S, Chen T (2020) Cascade EF-GAN: progressive facial expression editing with local focuses. In: IEEE Conference on computer vision and pattern recognition, pp 5021\u20135030","DOI":"10.1109\/CVPR42600.2020.00507"},{"key":"15268_CR35","doi-asserted-by":"crossref","unstructured":"Yi Z, Zhang H, Tan P, Gong M (2017) DualGAN: unsupervised dual learning for image-to-image translation. In: IEEE International conference on computer vision, pp 2849\u20132857","DOI":"10.1109\/ICCV.2017.310"},{"key":"15268_CR36","doi-asserted-by":"crossref","unstructured":"Zhu J-Y, Park T, Isola P, Efros A A (2017) Unpaired image-to-image translation using cycle-consistent adversarial networks. In: IEEE International conference on computer vision, pp 2223\u20132232","DOI":"10.1109\/ICCV.2017.244"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-023-15268-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-023-15268-2\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-023-15268-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,1,8]],"date-time":"2024-01-08T07:08:27Z","timestamp":1704697707000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-023-15268-2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,6,6]]},"references-count":36,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2024,1]]}},"alternative-id":["15268"],"URL":"https:\/\/doi.org\/10.1007\/s11042-023-15268-2","relation":{},"ISSN":["1380-7501","1573-7721"],"issn-type":[{"value":"1380-7501","type":"print"},{"value":"1573-7721","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,6,6]]},"assertion":[{"value":"14 April 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"1 September 2022","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 April 2023","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 June 2023","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}}]}}