{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,21]],"date-time":"2025-02-21T17:19:08Z","timestamp":1740158348759,"version":"3.37.3"},"reference-count":34,"publisher":"Wiley","license":[{"start":{"date-parts":[[2023,11,14]],"date-time":"2023-11-14T00:00:00Z","timestamp":1699920000000},"content-version":"unspecified","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100015758","name":"Adama Science and Technology University","doi-asserted-by":"publisher","award":["ASTU\/SM-R\/239\/21"],"award-info":[{"award-number":["ASTU\/SM-R\/239\/21"]}],"id":[{"id":"10.13039\/501100015758","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Journal of Electrical and Computer Engineering"],"published-print":{"date-parts":[[2023,11,14]]},"abstract":"<jats:p>The fundamental challenge in video generation is not only generating high-quality image sequences but also generating consistent frames with no abrupt shifts. With the development of generative adversarial networks (GANs), great progress has been made in image generation tasks which can be used for facial expression synthesis. Most previous works focused on synthesizing frontal and near frontal faces and manual annotation. However, considering only the frontal and near frontal area is not sufficient for many real-world applications, and manual annotation fails when the video is incomplete. AffineGAN, a recent study, uses affine transformation in latent space to automatically infer the expression intensity value; however, this work requires extraction of the feature of the target ground truth image, and the generated sequence of images is also not sufficient. To address these issues, this study is proposed to infer the expression of intensity value automatically without the need to extract the feature of the ground truth images. The local dataset is prepared with frontal and with two different face positions (the left and right sides). Average content distance metrics of the proposed solution along with different experiments have been measured, and the proposed solution has shown improvements. The proposed method has improved the ACD-I of affine GAN from 1.606\u2009\u00b1\u20090.018 to 1.584\u2009\u00b1\u20090.00, ACD-C of affine GAN from 1.452\u2009\u00b1\u20090.008 to 1.430\u2009\u00b1\u20090.009, and ACD-G of affine GAN from 1.769\u2009\u00b1\u20090.007 to 1.744\u2009\u00b1\u20090.01, which is far better than AffineGAN. This work concludes that integrating self-attention into the generator network improves a quality of the generated images sequences. In addition, evenly distributing values based on frame size to assign expression intensity value improves the consistency of image sequences being generated. It also enables the generator to generate different frame size videos while remaining within the range [0, 1].<\/jats:p>","DOI":"10.1155\/2023\/6645356","type":"journal-article","created":{"date-parts":[[2023,11,14]],"date-time":"2023-11-14T20:20:10Z","timestamp":1699993210000},"page":"1-13","source":"Crossref","is-referenced-by-count":1,"title":["Attention-Based Image-to-Video Translation for Synthesizing Facial Expression Using GAN"],"prefix":"10.1155","volume":"2023","author":[{"ORCID":"https:\/\/orcid.org\/0009-0001-6310-4515","authenticated-orcid":true,"given":"Kidist","family":"Alemayehu","sequence":"first","affiliation":[{"name":"School of Electrical Engineering and Computing, Department of Computer Science and Engineering, Adama Science and Technology University, Adama, Ethiopia"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6658-5142","authenticated-orcid":true,"given":"Worku","family":"Jifara","sequence":"additional","affiliation":[{"name":"School of Electrical Engineering and Computing, Department of Computer Science and Engineering, Adama Science and Technology University, Adama, Ethiopia"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7948-8285","authenticated-orcid":true,"given":"Demissie","family":"Jobir","sequence":"additional","affiliation":[{"name":"School of Electrical Engineering and Computing, Department of Electronics and Communications Engineering, Adama Science and Technology University, Adama 1888, Ethiopia"}]}],"member":"311","reference":[{"issue":"11","key":"1","doi-asserted-by":"crossref","first-page":"139","DOI":"10.1145\/3422622","article-title":"Generative adversarial networks","volume":"63","author":"I. Goodfellow","year":"2020","journal-title":"Communications of the ACM"},{"first-page":"294","article-title":"Image super-resolution using very deep residual channel attention networks","author":"Y. Zhang","key":"2"},{"key":"3","doi-asserted-by":"publisher","DOI":"10.1109\/tmm.2019.2962317"},{"first-page":"7354","article-title":"Self-attention generative adversarial networks","author":"H. Zhang","key":"4"},{"author":"P. Isola","key":"5","article-title":"Image-to-image translation with conditional adversarial networks"},{"author":"J. Y. Zhu","key":"6","article-title":"Unpaired image-to-image translation using cycle-consistent adversarial networks"},{"key":"7","doi-asserted-by":"publisher","DOI":"10.1109\/access.2021.3053408"},{"author":"Y. Zhou","key":"8","article-title":"Image2GIF: generating cinemagraphs using recurrent deep Q-networks"},{"article-title":"Video-to-Video synthesis","year":"2018","author":"M. Liu","key":"9"},{"first-page":"283","article-title":"Facial expression synthesis by u-net conditional generative adversarial networks","author":"X. Wang","key":"10"},{"author":"L. Fan","key":"11","article-title":"Controllable image-to-video translation: a case study on facial expression generation"},{"key":"12","article-title":"Learning to forecast and refine residual motion for image-to-video generation","volume-title":"Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)","author":"L. Zhao","year":"2018"},{"key":"13","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-020-01308-z"},{"key":"14","doi-asserted-by":"publisher","DOI":"10.1109\/tpami.2003.1227983"},{"first-page":"2505","article-title":"Facial image-to-video translation by a hidden affine transformation","author":"G. Shen","key":"15"},{"key":"16","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2014.01.005"},{"key":"17","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2019.107111"},{"author":"M. Chen","key":"18","article-title":"Double encoder conditional GAN for facial expression synthesis"},{"author":"Y. Choi","key":"19","article-title":"StarGAN: unified generative adversarial networks for multi-domain image-to-image translation"},{"key":"20","doi-asserted-by":"publisher","DOI":"10.1109\/tifs.2021.3050065"},{"article-title":"Geometry-Contrastive GAN for Facial Expression Transfer","year":"2018","author":"F. Qiao","key":"21"},{"article-title":"Geometry guided adversarial facial expression synthesis","author":"L. Song","key":"22","doi-asserted-by":"crossref","DOI":"10.1145\/3240508.3240612"},{"author":"Z. Lu","key":"23","article-title":"Conditional expression synthesis with face parsing transformation"},{"key":"24","doi-asserted-by":"crossref","DOI":"10.1007\/978-3-030-01249-6_50","article-title":"GANimation: anatomically-aware facial animation from a single image","volume-title":"Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)","author":"A. Pumarola","year":"2018"},{"issue":"10\u201011","key":"25","doi-asserted-by":"crossref","first-page":"2629","DOI":"10.1007\/s11263-020-01338-7","article-title":"Slidergan: synthesizing expressive face images by sliding 3d blendshape parameters","volume":"128","author":"E. Ververas","year":"2020","journal-title":"International Journal of Computer Vision"},{"article-title":"ExprGAN: facial expression editing with controllable expression intensity","author":"H. Ding","key":"26","doi-asserted-by":"crossref","DOI":"10.1609\/aaai.v32i1.12277"},{"author":"S. Qian","key":"27","article-title":"Make a face: towards arbitrary high fidelity face manipulation"},{"article-title":"An Attention-Based Image-To-Video Translation for Synthesizing Facial Expression Using GAN","year":"2022","author":"A. Kidist","key":"28"},{"author":"H. Zhang","key":"29","article-title":"Self-attention generative adversarial networks"},{"key":"30","doi-asserted-by":"publisher","DOI":"10.3389\/fninf.2020.611666"},{"key":"31","doi-asserted-by":"publisher","DOI":"10.1109\/tpami.2019.2913372"},{"first-page":"1622","article-title":"MRI reconstruction via cascaded channel-wise attention network","author":"Q. Huang","key":"32"},{"author":"D. P. Kingma","key":"33","article-title":"A method for stochastic optimization"},{"first-page":"1021","article-title":"How far are we from solving the 2D & 3D Face Alignment problem?(and a dataset of 230, 000 3D facial landmarks)","author":"A. Bulat","key":"34"}],"container-title":["Journal of Electrical and Computer Engineering"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/downloads.hindawi.com\/journals\/jece\/2023\/6645356.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/downloads.hindawi.com\/journals\/jece\/2023\/6645356.xml","content-type":"application\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/downloads.hindawi.com\/journals\/jece\/2023\/6645356.pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,11,14]],"date-time":"2023-11-14T20:20:15Z","timestamp":1699993215000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.hindawi.com\/journals\/jece\/2023\/6645356\/"}},"subtitle":[],"editor":[{"given":"Nihal F. F.","family":"Areed","sequence":"additional","affiliation":[]}],"short-title":[],"issued":{"date-parts":[[2023,11,14]]},"references-count":34,"alternative-id":["6645356","6645356"],"URL":"https:\/\/doi.org\/10.1155\/2023\/6645356","relation":{},"ISSN":["2090-0155","2090-0147"],"issn-type":[{"type":"electronic","value":"2090-0155"},{"type":"print","value":"2090-0147"}],"subject":[],"published":{"date-parts":[[2023,11,14]]}}}