{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,7]],"date-time":"2026-04-07T16:56:16Z","timestamp":1775580976520,"version":"3.50.1"},"reference-count":145,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"5","license":[{"start":{"date-parts":[[2024,5,1]],"date-time":"2024-05-01T00:00:00Z","timestamp":1714521600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0\/"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Pattern Anal. Mach. Intell."],"published-print":{"date-parts":[[2024,5]]},"DOI":"10.1109\/tpami.2024.3350004","type":"journal-article","created":{"date-parts":[[2024,1,15]],"date-time":"2024-01-15T21:10:11Z","timestamp":1705353011000},"page":"3557-3576","source":"Crossref","is-referenced-by-count":80,"title":["Face Generation and Editing With StyleGAN: A Survey"],"prefix":"10.1109","volume":"46","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-7252-9267","authenticated-orcid":false,"given":"Andrew","family":"Melnik","sequence":"first","affiliation":[{"name":"Bielefeld Univeristy, Bielefeld, Germany"}]},{"ORCID":"https:\/\/orcid.org\/0009-0006-9415-4634","authenticated-orcid":false,"given":"Maksim","family":"Miasayedzenkau","sequence":"additional","affiliation":[{"name":"Banuba, Minsk, Belarus"}]},{"ORCID":"https:\/\/orcid.org\/0009-0006-8799-0634","authenticated-orcid":false,"given":"Dzianis","family":"Makaravets","sequence":"additional","affiliation":[{"name":"Banuba, Minsk, Belarus"}]},{"ORCID":"https:\/\/orcid.org\/0009-0009-7831-1953","authenticated-orcid":false,"given":"Dzianis","family":"Pirshtuk","sequence":"additional","affiliation":[{"name":"Banuba, Minsk, Belarus"}]},{"given":"Eren","family":"Akbulut","sequence":"additional","affiliation":[{"name":"Bielefeld Univeristy, Bielefeld, Germany"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-3406-1738","authenticated-orcid":false,"given":"Dennis","family":"Holzmann","sequence":"additional","affiliation":[{"name":"Bielefeld Univeristy, Bielefeld, Germany"}]},{"given":"Tarek","family":"Renusch","sequence":"additional","affiliation":[{"name":"Bielefeld Univeristy, Bielefeld, Germany"}]},{"given":"Gustav","family":"Reichert","sequence":"additional","affiliation":[{"name":"Bielefeld Univeristy, Bielefeld, Germany"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1703-1906","authenticated-orcid":false,"given":"Helge","family":"Ritter","sequence":"additional","affiliation":[{"name":"Bielefeld Univeristy, Bielefeld, Germany"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00813"},{"key":"ref2","article-title":"GAN folks","year":"2021"},{"key":"ref3","first-page":"12104","article-title":"Training generative adversarial networks with limited data","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Karras"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00209"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00832"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00905"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1145\/3550454.3555436"},{"key":"ref8","article-title":"My style examples","year":"2022"},{"key":"ref9","article-title":"Toonify.photos","year":"2020"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1145\/3528223.3530164"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00618"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00480"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01149"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1523\/JNEUROSCI.2609-12.2012"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.3389\/fnhum.2017.00150"},{"key":"ref16","article-title":"Facial features report","year":"2023"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00453"},{"key":"ref18","article-title":"GAN folks medium","year":"2021"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.02008"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/FG52635.2021.9666998"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2015.2461462"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.167"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-25063-7_13"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.48550\/ARXIV.2106.12423"},{"key":"ref25","article-title":"Faces: Ai blitz XIII solutions","author":"Melnik","year":"2022"},{"key":"ref26","first-page":"8748","article-title":"Learning transferable visual models from natural language supervision","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Radford"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00073"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.02135"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2019.2933590"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2018.2890017"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1109\/TCYB.2019.2924589"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2018.2869688"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2019.2942514"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1145\/3450626.3459860"},{"key":"ref35","article-title":"3D cartoon face generation with controllable expressions from a single GAN image","author":"Wang","year":"2022"},{"key":"ref36","article-title":"Fine-tuning StyleGAN2 for cartoon face generation","author":"Back","year":"2021"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1016\/j.cviu.2022.103525"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1145\/3425780"},{"key":"ref39","article-title":"D. nostalgia","year":"2022"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1109\/JSTSP.2020.3002101"},{"key":"ref41","article-title":"Unconstrained facial expression transfer using style-based generator","author":"Yang","year":"2019"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.01380"},{"key":"ref43","first-page":"9759","article-title":"Unsupervised depth estimation, 3D face rotation and replacement","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Moniz"},{"key":"ref44","article-title":"MobileNets: Efficient convolutional neural networks for mobile vision applications","author":"Howard","year":"2017"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00533"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58542-6_11"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.425"},{"key":"ref48","article-title":"Progressive growing of GANs for improved quality, stability, and variation","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Karras"},{"key":"ref49","first-page":"1755","article-title":"Dlib-ml: A machine learning toolkit","volume":"10","author":"King","year":"2009","journal-title":"J. Mach. Learn. Res."},{"key":"ref50","doi-asserted-by":"publisher","DOI":"10.1145\/3422622"},{"key":"ref51","article-title":"On convergence and stability of GANs","author":"Kodali","year":"2017"},{"key":"ref52","article-title":"Alias-free generative adversarial networks (styleGAN3)","year":"2021"},{"key":"ref53","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01336"},{"key":"ref54","first-page":"214","article-title":"Wasserstein generative adversarial networks","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Arjovsky"},{"key":"ref55","article-title":"Geometric GAN","author":"Lim","year":"2017"},{"key":"ref56","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00652"},{"key":"ref57","article-title":"Generating images with perceptual similarity metrics based on deep networks","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Dosovitskiy"},{"key":"ref58","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00068"},{"key":"ref59","article-title":"Very deep convolutional networks for large-scale image recognition","author":"Simonyan","year":"2014","journal-title":"arXiv:1409.1556"},{"key":"ref60","doi-asserted-by":"publisher","DOI":"10.1145\/3065386"},{"key":"ref61","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00482"},{"key":"ref62","article-title":"GANs trained by a two time-scale update rule converge to a local nash equilibrium","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Heusel"},{"key":"ref63","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.308"},{"key":"ref64","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-022-01598-5"},{"key":"ref65","doi-asserted-by":"publisher","DOI":"10.1111\/cgf.14503"},{"key":"ref66","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01267"},{"key":"ref67","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00232"},{"key":"ref68","doi-asserted-by":"publisher","DOI":"10.1145\/3450626.3459838"},{"key":"ref69","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00664"},{"key":"ref70","doi-asserted-by":"publisher","DOI":"10.1145\/3544777"},{"key":"ref71","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01796"},{"key":"ref72","article-title":"Precise recovery of latent vectors from generative adversarial networks","author":"Lipton","year":"2017"},{"key":"ref73","article-title":"Official portrait of president-elect barack obama","author":"Garfield","year":"2009"},{"key":"ref74","article-title":"Actors headshots female white late twenties","author":"Photography","year":"2016"},{"key":"ref75","doi-asserted-by":"publisher","DOI":"10.48550\/arXiv.1603.08155"},{"key":"ref76","article-title":"Collaborative learning for faster styleGAN embedding","author":"Guan","year":"2020"},{"key":"ref77","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2003.819861"},{"key":"ref78","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2022.3167305"},{"key":"ref79","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58520-4_35"},{"key":"ref80","article-title":"Adversarial feature learning","author":"Donahue","year":"2016"},{"key":"ref81","article-title":"MyStyle: A personalized generative prior","author":"Melnik","year":"2022"},{"key":"ref82","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00926"},{"key":"ref83","first-page":"9841","article-title":"Ganspac: Discovering interpretable GAN controls","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"H\u00e4rk\u00f6nen"},{"key":"ref84","article-title":"StyleAlign: Analysis and applications of aligned styleGAN models","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Wu"},{"key":"ref85","article-title":"Resolution dependent GAN interpolation for controllable image synthesis between domains","author":"Pinkney","year":"2020"},{"key":"ref86","first-page":"29710","article-title":"Blendgan: Implicitly GAN blending for arbitrary stylized face generation","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst.","author":"Liu"},{"key":"ref87","first-page":"30599","article-title":"Towards robust blind face restoration with codebook lookup transformer","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst.","author":"Zhou"},{"key":"ref88","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-19797-0_8"},{"key":"ref89","article-title":"Neural discrete representation learning","author":"van den Oord","year":"2017"},{"key":"ref90","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.262"},{"key":"ref91","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00165"},{"key":"ref92","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2019\/129"},{"key":"ref93","doi-asserted-by":"publisher","DOI":"10.1109\/ICDM.2019.00089"},{"key":"ref94","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2017.2737423"},{"key":"ref95","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2018.2868350"},{"key":"ref96","article-title":"Fakeapp"},{"key":"ref97","article-title":"FaceSwapNet: Landmark guided many-to-many face reenactment","author":"Zhang","year":"2019"},{"key":"ref98","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00512"},{"key":"ref99","article-title":"How deepfake technology can change the movie industry","year":"2023"},{"key":"ref100","doi-asserted-by":"publisher","DOI":"10.1016\/j.cviu.2022.103526"},{"key":"ref101","doi-asserted-by":"publisher","DOI":"10.1109\/SIPROCESS.2017.8124497"},{"key":"ref102","doi-asserted-by":"publisher","DOI":"10.5220\/0011677300003417"},{"key":"ref103","article-title":"End-to-end speech-driven facial animation with temporal GANs","volume-title":"Proc. Brit. Mach. Vis. Conf.","author":"Vougioukas"},{"key":"ref104","article-title":"Video-to-video synthesis","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst.","author":"Wang"},{"key":"ref105","doi-asserted-by":"publisher","DOI":"10.2139\/ssrn.4170552"},{"key":"ref106","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i2.25280"},{"key":"ref107","doi-asserted-by":"publisher","DOI":"10.1142\/S0218001493000339"},{"key":"ref108","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.244"},{"key":"ref109","doi-asserted-by":"publisher","DOI":"10.1145\/3503250"},{"key":"ref110","article-title":"StyleNeRF: A style-based 3D-aware generator for high-resolution image synthesis","author":"Gu","year":"2021"},{"key":"ref111","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01314"},{"key":"ref112","doi-asserted-by":"publisher","DOI":"10.1016\/j.imavis.2024.104911"},{"key":"ref113","doi-asserted-by":"publisher","DOI":"10.1109\/cvpr52729.2023.00197"},{"key":"ref114","article-title":"Diffused heads: Diffusion models beat GANs on talking-face generation","author":"Stypu\u0142kowski","year":"2023"},{"key":"ref115","doi-asserted-by":"publisher","DOI":"10.1109\/cvpr52729.2023.01764"},{"key":"ref116","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00355"},{"key":"ref117","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00589"},{"key":"ref118","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00588"},{"key":"ref119","article-title":"Llama-adapter: Efficient fine-tuning of language models with zero-init attention","author":"Zhang","year":"2023"},{"key":"ref120","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.01913"},{"key":"ref121","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00439"},{"key":"ref122","doi-asserted-by":"publisher","DOI":"10.1145\/3528233.3530738"},{"key":"ref123","doi-asserted-by":"publisher","DOI":"10.1145\/3588432.3591500"},{"key":"ref124","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW59228.2023.00495"},{"key":"ref125","doi-asserted-by":"publisher","DOI":"10.1109\/ICCVW60793.2023.00333"},{"key":"ref126","doi-asserted-by":"publisher","DOI":"10.1109\/ICCVW60793.2023.00341"},{"key":"ref127","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.01223"},{"key":"ref128","article-title":"Exploiting diffusion prior for real-world image super-resolution","author":"Wang","year":"2023"},{"key":"ref129","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00743"},{"key":"ref130","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00170"},{"key":"ref131","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00590"},{"key":"ref132","doi-asserted-by":"publisher","DOI":"10.1145\/3592450"},{"key":"ref133","article-title":"StyleGAN-T: Unlocking the power of GANs for fast large-scale text-to-image synthesis","author":"Sauer","year":"2023"},{"key":"ref134","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00976"},{"key":"ref135","article-title":"Photorealistic text-to-image diffusion models with deep language understanding","author":"Saharia","year":"2022"},{"key":"ref136","article-title":"Hierarchical text-conditional image generation with clip latents","author":"Ramesh","year":"2022"},{"key":"ref137","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.01618"},{"key":"ref138","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01565"},{"key":"ref139","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00442"},{"key":"ref140","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.02011"},{"key":"ref141","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.01621"},{"key":"ref142","article-title":"In-n-out: Face video inversion and editing with volumetric decomposition","author":"Xu","year":"2023"},{"key":"ref143","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00041"},{"key":"ref144","doi-asserted-by":"publisher","DOI":"10.1145\/3550454.3555437"},{"key":"ref145","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.01920"}],"container-title":["IEEE Transactions on Pattern Analysis and Machine Intelligence"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/34\/10490207\/10399793.pdf?arnumber=10399793","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,4,9]],"date-time":"2024-04-09T19:37:57Z","timestamp":1712691477000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10399793\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,5]]},"references-count":145,"journal-issue":{"issue":"5"},"URL":"https:\/\/doi.org\/10.1109\/tpami.2024.3350004","relation":{},"ISSN":["0162-8828","2160-9292","1939-3539"],"issn-type":[{"value":"0162-8828","type":"print"},{"value":"2160-9292","type":"electronic"},{"value":"1939-3539","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,5]]}}}