{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,7]],"date-time":"2026-04-07T16:16:55Z","timestamp":1775578615328,"version":"3.50.1"},"reference-count":69,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"4","license":[{"start":{"date-parts":[[2022,4,1]],"date-time":"2022-04-01T00:00:00Z","timestamp":1648771200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2022,4,1]],"date-time":"2022-04-01T00:00:00Z","timestamp":1648771200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2022,4,1]],"date-time":"2022-04-01T00:00:00Z","timestamp":1648771200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"name":"Research Grants Council of Hong Kong","award":["24206219"],"award-info":[{"award-number":["24206219"]}]},{"name":"CUHK Faculty of Engineering","award":["RSFS"],"award-info":[{"award-number":["RSFS"]}]},{"name":"SenseTime Collaborative"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Pattern Anal. Mach. Intell."],"published-print":{"date-parts":[[2022,4,1]]},"DOI":"10.1109\/tpami.2020.3034267","type":"journal-article","created":{"date-parts":[[2020,10,27]],"date-time":"2020-10-27T20:08:51Z","timestamp":1603829331000},"page":"2004-2018","source":"Crossref","is-referenced-by-count":361,"title":["InterFaceGAN: Interpreting the Disentangled Face Representation Learned by GANs"],"prefix":"10.1109","volume":"44","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-3801-6705","authenticated-orcid":false,"given":"Yujun","family":"Shen","sequence":"first","affiliation":[{"name":"Department of Information Engineering, The Chinese University of Hong Kong, Hong Kong"}]},{"given":"Ceyuan","family":"Yang","sequence":"additional","affiliation":[{"name":"Department of Information Engineering, The Chinese University of Hong Kong, Hong Kong"}]},{"given":"Xiaoou","family":"Tang","sequence":"additional","affiliation":[{"name":"Department of Information Engineering, The Chinese University of Hong Kong, Hong Kong"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4030-0684","authenticated-orcid":false,"given":"Bolei","family":"Zhou","sequence":"additional","affiliation":[{"name":"Department of Information Engineering, The Chinese University of Hong Kong, Hong Kong"}]}],"member":"263","reference":[{"key":"ref1","first-page":"1","article-title":"Progressive growing of GANs for improved quality, stability, and variation","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Karras"},{"key":"ref2","first-page":"1","article-title":"Generative adversarial nets","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst.","author":"Goodfellow"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00813"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00453"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00917"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00926"},{"key":"ref7","first-page":"5969","article-title":"Fader networks: Manipulating images by sliding attributes","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst.","author":"Lample"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1145\/3306346.3323023"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.19"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-11021-5_5"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.728"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00457"},{"key":"ref13","first-page":"1","article-title":"Video-to-video synthesis","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst.","author":"Wang"},{"key":"ref14","first-page":"1","article-title":"Few-shot video-to-video synthesis","author":"Wang","year":"2019"},{"key":"ref15","first-page":"214","article-title":"Wasserstein generative adversarial networks","volume-title":"Proc. 34th Int. Conf. Mach. Learn.","author":"Arjovsky"},{"key":"ref16","first-page":"1","article-title":"Improved training of wasserstein GANs","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst.","author":"Gulrajani"},{"key":"ref17","article-title":"Began: Boundary equilibrium generative adversarial networks","author":"Berthelot","year":"2017","journal-title":"arXiv: 1703.10717"},{"key":"ref18","first-page":"1","article-title":"Spectral normalization for generative adversarial networks","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Miyato"},{"key":"ref19","first-page":"7354","article-title":"Self-attention generative adversarial networks","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Zhang"},{"key":"ref20","first-page":"1","article-title":"Large scale GAN training for high fidelity natural image synthesis","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Brock"},{"key":"ref21","first-page":"1540","article-title":"Metrics for deep generative models","volume-title":"Proc. 21st Int. Conf. Artif. Intell. Statist.","author":"Chen"},{"key":"ref22","first-page":"1","article-title":"Latent space oddity: On the curvature of deep generative models","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Arvanitidis"},{"key":"ref23","article-title":"Latent space non-linear statistics","author":"Kuhnel","year":"2018"},{"key":"ref24","first-page":"1","article-title":"Feature-based metrics for exploring the latent space of generative models","volume-title":"Proc. Int. Conf. Learn. Representations Workshop","author":"Laine"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW.2018.00071"},{"key":"ref26","first-page":"600","article-title":"Optimizing the latent space of generative networks","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Bojanowski"},{"key":"ref27","first-page":"1","article-title":"Gan dissection: Visualizing and understanding generative adversarial networks","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Bau"},{"key":"ref28","first-page":"1","article-title":"Unsupervised representation learning with deep convolutional generative adversarial networks","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Radford"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.645"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00584"},{"key":"ref31","first-page":"1","article-title":"On the \u201dsteerability\u201d of generative adversarial networks","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Jahanian"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-020-01429-5"},{"key":"ref33","first-page":"2642","article-title":"Conditional image synthesis with auxiliary classifier GANs","volume-title":"Proc. 34th Int. Conf. Mach. Learn.","author":"Odena"},{"key":"ref34","first-page":"1","article-title":"InfoGAN: Interpretable representation learning by information maximizing generative adversarial nets","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst.","author":"Chen"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.141"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.430"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00702"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01249-6_11"},{"key":"ref39","article-title":"FaceFeat-GAN: A two-stage approach for identity-preserving face synthesis","author":"Shen","year":"2018","journal-title":"arXiv: 1812.01288"},{"key":"ref40","first-page":"1","article-title":"Semantically decomposing the latent spaces of generative adversarial networks","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Donahue"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00092"},{"key":"ref42","first-page":"1","article-title":"Invertible conditional GANs for image editing","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst. Workshop","author":"Perarnau"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46454-1_36"},{"key":"ref44","first-page":"1","article-title":"Precise recovery of latent vectors from generative adversarial networks","volume-title":"Proc. Int. Conf. Learn. Representations Workshop","author":"Lipton"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2018.2875194"},{"key":"ref46","first-page":"1","article-title":"Invertibility of convolutional generative networks from partial measurements","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst.","author":"Ma"},{"key":"ref47","first-page":"4432","article-title":"Image2styleGAN: How to embed images into the styleGAN latent space?","volume-title":"Proc. Int. Conf. Comput. Vis.","author":"Rameen"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00832"},{"key":"ref49","first-page":"1","article-title":"Adversarially learned inference","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Dumoulin"},{"key":"ref50","first-page":"1","article-title":"Adversarial feature learning","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Donahue"},{"key":"ref51","article-title":"Lia: Latently invertible autoencoder with adversarial learning","author":"Zhu","year":"2019","journal-title":"arXiv : 1906.08090"},{"key":"ref52","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00460"},{"key":"ref53","first-page":"1","article-title":"Inverting layers of a large generator","volume-title":"Proc. Int. Conf. Learn. Representations Workshop","author":"Bau"},{"key":"ref54","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58520-4_35"},{"key":"ref55","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00308"},{"key":"ref56","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58536-5_16"},{"key":"ref57","article-title":"Learning to generate images of outdoor scenes from attributes and semantic layouts","author":"Karacan","year":"2016"},{"key":"ref58","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00244"},{"key":"ref59","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.632"},{"key":"ref60","first-page":"1","article-title":"Unsupervised image-to-image translation networks","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst.","author":"Liu"},{"key":"ref61","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.244"},{"key":"ref62","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01219-9_11"},{"key":"ref63","first-page":"1","article-title":"Toward multimodal image-to-image translation","volume-title":"Proc. Int. Conf. Neural Inf. Process. Syst.","author":"Zhu"},{"key":"ref64","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00916"},{"key":"ref65","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00821"},{"key":"ref66","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.425"},{"key":"ref67","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref68","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00179"},{"key":"ref69","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58542-6_11"}],"container-title":["IEEE Transactions on Pattern Analysis and Machine Intelligence"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/34\/9729045\/09241434.pdf?arnumber=9241434","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,1,9]],"date-time":"2024-01-09T23:57:02Z","timestamp":1704844622000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9241434\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,4,1]]},"references-count":69,"journal-issue":{"issue":"4"},"URL":"https:\/\/doi.org\/10.1109\/tpami.2020.3034267","relation":{},"ISSN":["0162-8828","2160-9292","1939-3539"],"issn-type":[{"value":"0162-8828","type":"print"},{"value":"2160-9292","type":"electronic"},{"value":"1939-3539","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,4,1]]}}}