{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,8]],"date-time":"2026-02-08T16:52:48Z","timestamp":1770569568072,"version":"3.49.0"},"reference-count":209,"publisher":"Springer Science and Business Media LLC","issue":"5","license":[{"start":{"date-parts":[[2023,4,4]],"date-time":"2023-04-04T00:00:00Z","timestamp":1680566400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,4,4]],"date-time":"2023-04-04T00:00:00Z","timestamp":1680566400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Sci. China Inf. Sci."],"published-print":{"date-parts":[[2023,5]]},"DOI":"10.1007\/s11432-022-3679-0","type":"journal-article","created":{"date-parts":[[2023,4,13]],"date-time":"2023-04-13T17:03:04Z","timestamp":1681405384000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":25,"title":["Survey on leveraging pre-trained generative adversarial networks for image editing and restoration"],"prefix":"10.1007","volume":"66","author":[{"given":"Ming","family":"Liu","sequence":"first","affiliation":[]},{"given":"Yuxiang","family":"Wei","sequence":"additional","affiliation":[]},{"given":"Xiaohe","family":"Wu","sequence":"additional","affiliation":[]},{"given":"Wangmeng","family":"Zuo","sequence":"additional","affiliation":[]},{"given":"Lei","family":"Zhang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,4,4]]},"reference":[{"key":"3679_CR1","unstructured":"Goodfellow I, Pouget-Abadie J, Mirza M, et al. Generative adversarial nets. In: Proceedings of International Conference on Neural Information Processing Systems, 2014"},{"key":"3679_CR2","unstructured":"Denton E L, Chintala S, Fergus R, et al. Deep generative image models using a laplacian pyramid of adversarial networks. In: Proceedings of International Conference on Neural Information Processing Systems, 2015"},{"key":"3679_CR3","unstructured":"Radford A, Metz L, Chintala S. Unsupervised representation learning with deep convolutional generative adversarial networks. In: Proceedings of International Conference on Learning Representations, 2016"},{"key":"3679_CR4","doi-asserted-by":"crossref","unstructured":"Zhang H, Xu T, Li H, et al. StackGAN: text to photo-realistic image synthesis with stacked generative adversarial networks. In: Proceedings of IEEE International Conference on Computer Vision, 2017. 5907\u20135915","DOI":"10.1109\/ICCV.2017.629"},{"key":"3679_CR5","unstructured":"Zhang H, Goodfellow I, Metaxas D, et al. Self-attention generative adversarial networks. In: Proceedings of International Conference on Learning Representations, 2019. 7354\u20137363"},{"key":"3679_CR6","doi-asserted-by":"crossref","unstructured":"Mao X, Li Q, Xie H, et al. Least squares generative adversarial networks. In: Proceedings of IEEE International Conference on Computer Vision, 2017. 2794\u20132802","DOI":"10.1109\/ICCV.2017.304"},{"key":"3679_CR7","unstructured":"Berthelot D, Schumm T, Metz L. BEGAN: boundary equilibrium generative adversarial networks. 2017. ArXiv:1703.10717"},{"key":"3679_CR8","unstructured":"Jolicoeur-Martineau A. The relativistic discriminator: a key element missing from standard gan. In: Proceedings of International Conference on Learning Representations, 2019"},{"key":"3679_CR9","unstructured":"Arjovsky M, Chintala S, Bottou L. Wasserstein generative adversarial networks. In: Proceedings of International Conference on Learning Representations, 2017. 214\u2013223"},{"key":"3679_CR10","unstructured":"Gulrajani I, Ahmed F, Arjovsky M, et al. Improved training of wasserstein GANs. In: Proceedings of International Conference on Neural Information Processing Systems, 2017. 5769\u20135779"},{"key":"3679_CR11","unstructured":"Miyato T, Kataoka T, Koyama M, et al. Spectral normalization for generative adversarial networks. In: Proceedings of International Conference on Learning Representations, 2018"},{"key":"3679_CR12","unstructured":"Karras T, Aila T, Laine S, et al. Progressive growing of GANs for improved quality, stability, and variation. In: Proceedings of International Conference on Learning Representations, 2018"},{"key":"3679_CR13","unstructured":"Brock A, Donahue J, Simonyan K. Large scale GAN training for high fidelity natural image synthesis. In: Proceedings of International Conference on Learning Representations, 2018"},{"key":"3679_CR14","doi-asserted-by":"crossref","unstructured":"Karras T, Laine S, Aila T. A style-based generator architecture for generative adversarial networks. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2019. 4401\u20134410","DOI":"10.1109\/CVPR.2019.00453"},{"key":"3679_CR15","doi-asserted-by":"crossref","unstructured":"Karras T, Laine S, Aittala M, et al. Analyzing and improving the image quality of StyleGAN. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2020. 8110\u20138119","DOI":"10.1109\/CVPR42600.2020.00813"},{"key":"3679_CR16","unstructured":"Karras T, Aittala M, Hellsten J, et al. Training generative adversarial networks with limited data. 2020. ArXiv:2006.06676"},{"key":"3679_CR17","unstructured":"Karras T, Aittala M, Laine S, et al. Alias-free generative adversarial networks. In: Proceedings of International Conference on Neural Information Processing Systems, 2021"},{"key":"3679_CR18","doi-asserted-by":"crossref","unstructured":"Isola P, Zhu J Y, Zhou T, et al. Image-to-image translation with conditional adversarial networks. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2017. 1125\u20131134","DOI":"10.1109\/CVPR.2017.632"},{"key":"3679_CR19","doi-asserted-by":"crossref","unstructured":"Zhu J Y, Park T, Isola P, et al. Unpaired image-to-image translation using cycle-consistent adversarial networks. In: Proceedings of IEEE International Conference on Computer Vision, 2017. 2223\u20132232","DOI":"10.1109\/ICCV.2017.244"},{"key":"3679_CR20","doi-asserted-by":"crossref","unstructured":"Choi Y, Choi M, Kim M, et al. StarGAN: unified generative adversarial networks for multi-domain image-to-image translation. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2018. 8789\u20138797","DOI":"10.1109\/CVPR.2018.00916"},{"key":"3679_CR21","doi-asserted-by":"publisher","first-page":"5464","DOI":"10.1109\/TIP.2019.2916751","volume":"28","author":"Z He","year":"2019","unstructured":"He Z, Zuo W, Kan M, et al. AttGAN: facial attribute editing by only changing what you want. IEEE Trans Image Process, 2019, 28: 5464\u20135478","journal-title":"IEEE Trans Image Process"},{"key":"3679_CR22","doi-asserted-by":"crossref","unstructured":"Liu M, Ding Y, Xia M, et al. STGAN: a unified selective transfer network for arbitrary image attribute editing. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2019. 3673\u20133682","DOI":"10.1109\/CVPR.2019.00379"},{"key":"3679_CR23","doi-asserted-by":"crossref","unstructured":"Choi Y, Uh Y, Yoo J, et al. StarGAN v2: diverse image synthesis for multiple domains. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2020. 8188\u20138197","DOI":"10.1109\/CVPR42600.2020.00821"},{"key":"3679_CR24","doi-asserted-by":"crossref","unstructured":"Ledig C, Theis L, Husz\u00e1r F, et al. Photo-realistic single image super-resolution using a generative adversarial network. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2017. 4681\u20134690","DOI":"10.1109\/CVPR.2017.19"},{"key":"3679_CR25","doi-asserted-by":"crossref","unstructured":"Wang X, Yu K, Wu S, et al. ESRGAN: enhanced super-resolution generative adversarial networks. In: Proceedings of European Conference on Computer Vision, 2018","DOI":"10.1007\/978-3-030-11021-5_5"},{"key":"3679_CR26","doi-asserted-by":"crossref","unstructured":"Wang X, Xie L, Dong C, et al. Real-ESRGAN: training real-world blind super-resolution with pure synthetic data. In: Proceedings of IEEE International Conference on Computer Vision Workshops, 2021. 1905\u20131914","DOI":"10.1109\/ICCVW54120.2021.00217"},{"key":"3679_CR27","doi-asserted-by":"crossref","unstructured":"Zhang K, Liang J, van Gool L, et al. Designing a practical degradation model for deep blind image super-resolution. In: Proceedings of IEEE International Conference on Computer Vision, 2021. 4791\u20134800","DOI":"10.1109\/ICCV48922.2021.00475"},{"key":"3679_CR28","doi-asserted-by":"crossref","unstructured":"Kupyn O, Budzan V, Mykhailych M, et al. DeblurGAN: blind motion deblurring using conditional adversarial networks. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2018. 8183\u20138192","DOI":"10.1109\/CVPR.2018.00854"},{"key":"3679_CR29","doi-asserted-by":"crossref","unstructured":"Kupyn O, Martyniuk T, Wu J, et al. DeblurGAN-v2: deblurring (orders-of-magnitude) faster and better. In: Proceedings of IEEE International Conference on Computer Vision, 2019. 8878\u20138887","DOI":"10.1109\/ICCV.2019.00897"},{"key":"3679_CR30","doi-asserted-by":"crossref","unstructured":"Zheng S, Zhu Z, Zhang X, et al. Distribution-induced bidirectional generative adversarial network for graph representation learning. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2020. 7224\u20137233","DOI":"10.1109\/CVPR42600.2020.00725"},{"key":"3679_CR31","doi-asserted-by":"crossref","unstructured":"Zhu H, Peng X, Chandrasekhar V, et al. DehazeGAN: when image dehazing meets differential programming. In: Proceedings of International Joint Conference on Artificial Intelligence, 2018. 1234\u20131240","DOI":"10.24963\/ijcai.2018\/172"},{"key":"3679_CR32","doi-asserted-by":"publisher","first-page":"829","DOI":"10.1109\/TCYB.2019.2955092","volume":"51","author":"H Zhu","year":"2019","unstructured":"Zhu H, Cheng Y, Peng X, et al. Single-image dehazing via compositional adversarial network. IEEE Trans Cybern, 2019, 51: 829\u2013838","journal-title":"IEEE Trans Cybern"},{"key":"3679_CR33","doi-asserted-by":"crossref","unstructured":"Mehta A, Sinha H, Narang P, et al. HiDeGAN: a hyperspectral-guided image dehazing GAN. In: Proceedings of IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2020. 212\u2013213","DOI":"10.1109\/CVPRW50498.2020.00114"},{"key":"3679_CR34","doi-asserted-by":"crossref","unstructured":"Dong Y, Liu Y, Zhang H, et al. FD-GAN: generative adversarial networks with fusion-discriminator for single image dehazing. In: Proceedings of AAAI Conference on Artificial Intelligence, 2020. 10729\u201310736","DOI":"10.1609\/aaai.v34i07.6701"},{"key":"3679_CR35","doi-asserted-by":"crossref","unstructured":"Liu Z, Luo P, Wang X, et al. Deep learning face attributes in the wild. In: Proceedings of IEEE International Conference on Computer Vision, 2015. 3730\u20133738","DOI":"10.1109\/ICCV.2015.425"},{"key":"3679_CR36","unstructured":"Voynov A, Babenko A. Unsupervised discovery of interpretable directions in the GAN latent space. In: Proceedings of International Conference on Learning Representations, 2020. 9786\u20139796"},{"key":"3679_CR37","unstructured":"Yu F, Seff A, Zhang Y, et al. LSUN: construction of a large-scale image dataset using deep learning with humans in the loop. 2015. ArXiv:1506.03365"},{"key":"3679_CR38","doi-asserted-by":"crossref","unstructured":"Zhu J, Shen Y, Zhao D, et al. In-domain GAN inversion for real image editing. In: Proceedings of European Conference on Computer Vision, 2020. 592\u2013608","DOI":"10.1007\/978-3-030-58520-4_35"},{"key":"3679_CR39","doi-asserted-by":"crossref","unstructured":"Rudin L I, Osher S, Fatemi E. Nonlinear total variation based noise removal algorithms. Physica D-Nonlinear Phenomena, 1992, 60: 259\u2013268","DOI":"10.1016\/0167-2789(92)90242-F"},{"key":"3679_CR40","doi-asserted-by":"crossref","unstructured":"Buades A, Coll B, Morel J M. A non-local algorithm for image denoising. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2005. 60\u201365","DOI":"10.1109\/CVPR.2005.38"},{"key":"3679_CR41","doi-asserted-by":"publisher","first-page":"3736","DOI":"10.1109\/TIP.2006.881969","volume":"15","author":"M Elad","year":"2006","unstructured":"Elad M, Aharon M. Image denoising via sparse and redundant representations over learned dictionaries. IEEE Trans Image Process, 2006, 15: 3736\u20133745","journal-title":"IEEE Trans Image Process"},{"key":"3679_CR42","doi-asserted-by":"publisher","first-page":"1754","DOI":"10.1007\/s11263-021-01431-5","volume":"129","author":"B Li","year":"2021","unstructured":"Li B, Gou Y, Gu S, et al. You only look yourself: unsupervised and untrained single image dehazing neural network. Int J Comput Vis, 2021, 129: 1754\u20131767","journal-title":"Int J Comput Vis"},{"key":"3679_CR43","doi-asserted-by":"crossref","unstructured":"Shoshan A, Mechrez R, Zelnik-Manor L. Dynamic-Net: tuning the objective without re-training for synthesis tasks. In: Proceedings of IEEE International Conference on Computer Vision, 2019. 3215\u20133223","DOI":"10.1109\/ICCV.2019.00331"},{"key":"3679_CR44","unstructured":"Gou Y, Li B, Liu Z, et al. CLEARER: multi-scale neural architecture search for image restoration. In: Proceedings of International Conference on Neural Information Processing Systems, 2020, 33: 17129\u201317140"},{"key":"3679_CR45","unstructured":"Bau D, Zhu J Y, Strobelt H, et al. GAN dissection: visualizing and understanding generative adversarial networks. In: Proceedings of International Conference on Learning Representations, 2019"},{"key":"3679_CR46","doi-asserted-by":"crossref","unstructured":"Bau D, Zhu J Y, Wulff J, et al. Seeing what a GAN cannot generate. In: Proceedings of IEEE International Conference on Computer Vision, 2019. 4502\u20134511","DOI":"10.1109\/ICCV.2019.00460"},{"key":"3679_CR47","doi-asserted-by":"crossref","unstructured":"Goetschalckx L, Andonian A, Oliva A, et al. GANalyze: toward visual definitions of cognitive image properties. In: Proceedings of IEEE International Conference on Computer Vision, 2019. 5744\u20135753","DOI":"10.1109\/ICCV.2019.00584"},{"key":"3679_CR48","unstructured":"H\u00e4rk\u00f6nen E, Hertzmann A, Lehtinen J, et al. GANSpace: discovering interpretable GAN controls. In: Proceedings of International Conference on Neural Information Processing Systems, 2020"},{"key":"3679_CR49","unstructured":"Suzuki R, Koyama M, Miyato T, et al. Spatially controllable image synthesis with internal representation collaging. 2018. ArXiv:1811.10153"},{"key":"3679_CR50","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3306346.3323023","volume":"38","author":"D Bau","year":"2019","unstructured":"Bau D, Strobelt H, Peebles W, et al. Semantic photo manipulation with a generative image prior. ACM Trans Graph, 2019, 38: 1\u201311","journal-title":"ACM Trans Graph"},{"key":"3679_CR51","doi-asserted-by":"crossref","unstructured":"Tewari A, Elgharib M, Bharaj G, et al. StyleRig: rigging StyleGAN for 3D control over portrait images. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2020. 6142\u20136151","DOI":"10.1109\/CVPR42600.2020.00618"},{"key":"3679_CR52","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3447648","volume":"40","author":"R Abdal","year":"2021","unstructured":"Abdal R, Zhu P, Mitra N J, et al. StyleFlow: attribute-conditioned exploration of StyleGAN-generated images using conditional continuous normalizing flows. ACM Trans Graph, 2021, 40: 1\u201321","journal-title":"ACM Trans Graph"},{"key":"3679_CR53","doi-asserted-by":"crossref","unstructured":"Menon S, Damian A, Hu S, et al. Pulse: self-supervised photo upsampling via latent space exploration of generative models. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2020. 2437\u20132445","DOI":"10.1109\/CVPR42600.2020.00251"},{"key":"3679_CR54","doi-asserted-by":"crossref","unstructured":"Richardson E, Alaluf Y, Patashnik O, et al. Encoding in style: a StyleGAN encoder for image-to-image translation. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2021. 2287\u20132296","DOI":"10.1109\/CVPR46437.2021.00232"},{"key":"3679_CR55","doi-asserted-by":"crossref","unstructured":"Chan K C, Wang X, Xu X, et al. GLEAN: generative latent bank for large-factor image super-resolution. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2021. 14245\u201314254","DOI":"10.1109\/CVPR46437.2021.01402"},{"key":"3679_CR56","doi-asserted-by":"crossref","unstructured":"Wang X, Li Y, Zhang H, et al. Towards real-world blind face restoration with generative facial prior. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2021. 9168\u20139178","DOI":"10.1109\/CVPR46437.2021.00905"},{"key":"3679_CR57","doi-asserted-by":"crossref","unstructured":"Yang T, Ren P, Xie X, et al. GAN prior embedded network for blind face restoration in the wild. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2021. 672\u2013681","DOI":"10.1109\/CVPR46437.2021.00073"},{"key":"3679_CR58","doi-asserted-by":"publisher","first-page":"436","DOI":"10.1038\/nature14539","volume":"521","author":"Y LeCun","year":"2015","unstructured":"LeCun Y, Bengio Y, Hinton G. Deep learning. Nature, 2015, 521: 436\u2013444","journal-title":"Nature"},{"key":"3679_CR59","doi-asserted-by":"crossref","unstructured":"Deng J, Dong W, Socher R, et al. ImageNet: a large-scale hierarchical image database. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2009. 248\u2013255","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"3679_CR60","doi-asserted-by":"crossref","unstructured":"Lee C H, Liu Z, Wu L, et al. MaskGAN: towards diverse and interactive facial image manipulation. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2020. 5549\u20135558","DOI":"10.1109\/CVPR42600.2020.00559"},{"key":"3679_CR61","doi-asserted-by":"publisher","first-page":"2278","DOI":"10.1109\/5.726791","volume":"86","author":"Y Lecun","year":"1998","unstructured":"Lecun Y, Bottou L, Bengio Y, et al. Gradient-based learning applied to document recognition. Proc IEEE, 1998, 86: 2278\u20132324","journal-title":"Proc IEEE"},{"key":"3679_CR62","unstructured":"Netzer Y, Wang T, Coates A, et al. Reading digits in natural images with unsupervised feature learning. In: Proceedings of_NIPS Workshop on Deep Learning and Unsupervised Feature Learning, 2011"},{"key":"3679_CR63","unstructured":"Krizhevsky A. Learning multiple layers of features from tiny images. 2009. https:\/\/www.cs.toronto.edu\/kriz\/learning-features-2009-TR.pdf"},{"key":"3679_CR64","doi-asserted-by":"crossref","unstructured":"Liu Z, Yan S, Luo P, et al. Fashion landmark detection in the wild. In: Proceedings of European Conference on Computer Vision, 2016. 229\u2013245","DOI":"10.1007\/978-3-319-46475-6_15"},{"key":"3679_CR65","doi-asserted-by":"crossref","unstructured":"Cordts M, Omran M, Ramos S, et al. The cityscapes dataset for semantic urban scene understanding. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2016. 3213\u20133223","DOI":"10.1109\/CVPR.2016.350"},{"key":"3679_CR66","doi-asserted-by":"crossref","unstructured":"Shao S, Li Z, Zhang T, et al. Objects365: a large-scale, high-quality dataset for object detection. In: Proceedings of IEEE International Conference on Computer Vision, 2019. 8430\u20138439","DOI":"10.1109\/ICCV.2019.00852"},{"key":"3679_CR67","doi-asserted-by":"publisher","first-page":"1452","DOI":"10.1109\/TPAMI.2017.2723009","volume":"40","author":"B Zhou","year":"2017","unstructured":"Zhou B, Lapedriza A, Khosla A, et al. Places: a 10 million image database for scene recognition. IEEE Trans Pattern Anal Mach Intell, 2017, 40: 1452\u20131464","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"3679_CR68","unstructured":"Krasin I, Duerig T, Alldrin N, et al. OpenImages: a public dataset for large-scale multi-label and multi-class image classification. 2017. https:\/\/storage.googleapis.com\/openimages\/web\/index.html"},{"key":"3679_CR69","unstructured":"Salimans T, Goodfellow I, Zaremba W, et al. Improved techniques for training GANs. In: Proceedings of International Conference on Neural Information Processing Systems, 2016"},{"key":"3679_CR70","doi-asserted-by":"crossref","unstructured":"Szegedy C, Vanhoucke V, Ioffe S, et al. Rethinking the inception architecture for computer vision. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2016. 2818\u20132826","DOI":"10.1109\/CVPR.2016.308"},{"key":"3679_CR71","doi-asserted-by":"crossref","unstructured":"Gurumurthy S, Sarvadevabhatla S R K, Babu R V. DeliGAN: generative adversarial networks for diverse and limited data. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2017. 166\u2013174","DOI":"10.1109\/CVPR.2017.525"},{"key":"3679_CR72","unstructured":"Che T, Li Y, Jacob A P, et al. Mode regularized generative adversarial networks. In: Proceedings of International Conference on Learning Representations, 2017"},{"key":"3679_CR73","unstructured":"Zhou Z, Zhang W, Wang J. Inception score, label smoothing, gradient vanishing and \u2212log(D(x)) alternative. 2017. ArXiv:1708.01729"},{"key":"3679_CR74","unstructured":"Zhou Z, Cai H, Rong S, et al. Activation maximization generative adversarial nets. In: Proceedings of International Conference on Learning Representations, 2018"},{"key":"3679_CR75","unstructured":"Heusel M, Ramsauer H, Unterthiner T, et al. GANs trained by a two time-scale update rule converge to a local nash equilibrium. In: Proceedings of International Conference on Neural Information Processing Systems, 2017"},{"key":"3679_CR76","doi-asserted-by":"publisher","first-page":"22","DOI":"10.1007\/s10851-014-0506-3","volume":"51","author":"N Bonneel","year":"2015","unstructured":"Bonneel N, Rabin J, Peyr\u00e9 G, et al. Sliced and radon Wasserstein barycenters of measures. J Math Imaging Vision, 2015, 51: 22\u201345","journal-title":"J Math Imaging Vision"},{"key":"3679_CR77","unstructured":"Kolouri S, Nadjahi K, Simsekli U, et al. Generalized sliced Wasserstein distances. 2019. ArXiv:1902.00434"},{"key":"3679_CR78","doi-asserted-by":"crossref","unstructured":"Shmelkov K, Schmid C, Alahari K. How good is my GAN? In: Proceedings of European Conference on Computer Vision, 2018. 213\u2013229","DOI":"10.1007\/978-3-030-01216-8_14"},{"key":"3679_CR79","unstructured":"Kynk\u00e4\u00e4nniemi T, Karras T, Laine S, et al. Improved precision and recall metric for assessing generative models. 2019. ArXiv:1904.06991"},{"key":"3679_CR80","unstructured":"Khrulkov V, Oseledets I. Geometry score: a method for comparing generative adversarial networks. In: Proceedings of International Conference on Learning Representations, 2018. 2621\u20132629"},{"key":"3679_CR81","doi-asserted-by":"publisher","first-page":"600","DOI":"10.1109\/TIP.2003.819861","volume":"13","author":"Z Wang","year":"2004","unstructured":"Wang Z, Bovik A C, Sheikh H R, et al. Image quality assessment: from error visibility to structural similarity. IEEE Trans Image Process, 2004, 13: 600\u2013612","journal-title":"IEEE Trans Image Process"},{"key":"3679_CR82","doi-asserted-by":"crossref","unstructured":"Zhang R, Isola P, Efros A A, et al. The unreasonable effectiveness of deep features as a perceptual metric. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2018. 586\u2013595","DOI":"10.1109\/CVPR.2018.00068"},{"key":"3679_CR83","doi-asserted-by":"publisher","first-page":"41","DOI":"10.1016\/j.cviu.2018.10.009","volume":"179","author":"A Borji","year":"2019","unstructured":"Borji A. Pros and cons of GAN evaluation measures. Comput Vision Image Understanding, 2019, 179: 41\u201365","journal-title":"Comput Vision Image Understanding"},{"key":"3679_CR84","first-page":"1","volume":"54","author":"Z Wang","year":"2022","unstructured":"Wang Z, She Q, Ward T E. Generative adversarial networks in computer vision. ACM Comput Surv, 2022, 54: 1\u201338","journal-title":"ACM Comput Surv"},{"key":"3679_CR85","unstructured":"Kang M, Shin J, Park J. StudioGAN: a taxonomy and benchmark of gans for image synthesis. 2022. ArXiv:2206.09479"},{"key":"3679_CR86","unstructured":"Mescheder L, Geiger A, Nowozin S. Which training methods for GANs do actually converge? In: Proceedings of International Conference on Learning Representations, 2018. 3481\u20133490"},{"key":"3679_CR87","doi-asserted-by":"crossref","unstructured":"Huang X, Belongie S. Arbitrary style transfer in real-time with adaptive instance normalization. In: Proceedings of IEEE International Conference on Computer Vision, 2017. 1501\u20131510","DOI":"10.1109\/ICCV.2017.167"},{"key":"3679_CR88","unstructured":"Tancik M, Srinivasan P, Mildenhall B, et al. Fourier features let networks learn high frequency functions in low dimensional domains. In: Proceedings of International Conference on Neural Information Processing Systems, 2020. 7537\u20137547"},{"key":"3679_CR89","unstructured":"Mirza M, Osindero S. Conditional generative adversarial nets. 2014. ArXiv:1411.1784"},{"key":"3679_CR90","unstructured":"Perarnau G, van de Weijer J, Raducanu B, et al. Invertible conditional GANs for image editing. In: Proceedings of NeurIPSW, 2016"},{"key":"3679_CR91","doi-asserted-by":"crossref","unstructured":"Abdal R, Qin Y, Wonka P. Image2StyleGAN: how to embed images into the stylegan latent space? In: Proceedings of IEEE International Conference on Computer Vision, 2019. 4432\u20134441","DOI":"10.1109\/ICCV.2019.00453"},{"key":"3679_CR92","unstructured":"Liu Y, Li Q, Sun Z, et al. Style intervention: how to achieve spatial disentanglement with style-based generators? 2020. ArXiv:2011.09699"},{"key":"3679_CR93","doi-asserted-by":"crossref","unstructured":"Wu Z, Lischinski D, Shechtman E. Stylespace analysis: disentangled controls for StyleGAN image generation. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2021. 12863\u201312872","DOI":"10.1109\/CVPR46437.2021.01267"},{"key":"3679_CR94","doi-asserted-by":"crossref","unstructured":"Xu J, Xu H, Ni B, et al. Hierarchical style-based networks for motion synthesis. In: Proceedings of European Conference on Computer Vision, 2020. 178\u2013194","DOI":"10.1007\/978-3-030-58621-8_11"},{"key":"3679_CR95","doi-asserted-by":"crossref","unstructured":"Zhang L, Bai X, Gao Y. SalS-GAN: spatially-adaptive latent space in StyleGAN for real image embedding. In: Proceedings of ACM International Conference on Multimedia, 2021. 5176\u20135184","DOI":"10.1145\/3474085.3475633"},{"key":"3679_CR96","unstructured":"Zhu P, Abdal R, Qin Y, et al. Improved StyleGAN embedding: where are the good latents? 2020. ArXiv:2012.09036"},{"key":"3679_CR97","doi-asserted-by":"crossref","unstructured":"Abdal R, Qin Y, Wonka P. Image2StyleGAN++: how to edit the embedded images? In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2020. 8296\u20138305","DOI":"10.1109\/CVPR42600.2020.00832"},{"key":"3679_CR98","doi-asserted-by":"crossref","unstructured":"Kang K, Kim S, Cho S. GAN inversion for out-of-range images with geometric transformations. In: Proceedings of IEEE International Conference on Computer Vision, 2021. 13941\u201313949","DOI":"10.1109\/ICCV48922.2021.01368"},{"key":"3679_CR99","doi-asserted-by":"crossref","unstructured":"Cherepkov A, Voynov A, Babenko A. Navigating the GAN parameter space for semantic image editing. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2021. 3671\u20133680","DOI":"10.1109\/CVPR46437.2021.00367"},{"key":"3679_CR100","unstructured":"Feng Q, Shah V, Gadde R, et al. Near perfect GAN inversion. 2022. ArXiv:2202.11833"},{"key":"3679_CR101","unstructured":"Donahue J, Kr\u00e4henb\u00fchl P, Darrell T. Adversarial feature learning. In: Proceedings of International Conference on Learning Representations, 2017"},{"key":"3679_CR102","unstructured":"Dumoulin V, Belghazi I, Poole B, et al. Adversarially learned inference. 2016. ArXiv:1606.00704"},{"key":"3679_CR103","doi-asserted-by":"crossref","unstructured":"Zhu J Y, Kr\u00e4henb\u00fchl P, Shechtman E, et al. Generative visual manipulation on the natural image manifold. In: Proceedings of European Conference on Computer Vision, 2016. 597\u2013613","DOI":"10.1007\/978-3-319-46454-1_36"},{"key":"3679_CR104","doi-asserted-by":"publisher","first-page":"1967","DOI":"10.1109\/TNNLS.2018.2875194","volume":"30","author":"A Creswell","year":"2019","unstructured":"Creswell A, Bharath A A. Inverting the generator of a generative adversarial network. IEEE Trans Neural Netw Learn Syst, 2019, 30: 1967\u20131974","journal-title":"IEEE Trans Neural Netw Learn Syst"},{"key":"3679_CR105","unstructured":"Lipton Z C, Tripathi S. Precise recovery of latent vectors from generative adversarial networks. In: Proceedings of International Conference on Learning Representations Workshops, 2017"},{"key":"3679_CR106","doi-asserted-by":"crossref","unstructured":"Shah V, Hegde C. Solving linear inverse problems using GAN priors: an algorithm with provable guarantees. In: Proceedings of International Conference on Acoustics, Speech and Signal Processing, 2018. 4609\u20134613","DOI":"10.1109\/ICASSP.2018.8462233"},{"key":"3679_CR107","unstructured":"Ma F, Ayaz U, Karaman S. Invertibility of convolutional generative networks from partial measurements. In: Proceedings of International Conference on Neural Information Processing Systems, 2018. 9651\u20139660"},{"key":"3679_CR108","doi-asserted-by":"crossref","unstructured":"Raj A, Li Y, Bresler Y. GAN-based projector for faster recovery with convergence guarantees in linear inverse problems. In: Proceedings of IEEE International Conference on Computer Vision, 2019. 5602\u20135611","DOI":"10.1109\/ICCV.2019.00570"},{"key":"3679_CR109","unstructured":"Bau D, Zhu J Y, Wulff J, et al. Inverting layers of a large generator. In: Proceedings of International Conference on Learning Representations Workshops, 2019. 4"},{"key":"3679_CR110","doi-asserted-by":"crossref","unstructured":"Shen Y, Gu J, Tang X, et al. Interpreting the latent space of GANs for semantic face editing. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2020. 9243\u20139252","DOI":"10.1109\/CVPR42600.2020.00926"},{"key":"3679_CR111","doi-asserted-by":"crossref","unstructured":"Daras G, Odena A, Zhang H, et al. Your local GAN: designing two dimensional local attention mechanisms for generative models. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2020. 14531\u201314539","DOI":"10.1109\/CVPR42600.2020.01454"},{"key":"3679_CR112","doi-asserted-by":"crossref","unstructured":"Gu J, Shen Y, Zhou B. Image processing using multi-code GAN prior. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2020. 3012\u20133021","DOI":"10.1109\/CVPR42600.2020.00308"},{"key":"3679_CR113","doi-asserted-by":"publisher","first-page":"2459","DOI":"10.1007\/s11263-020-01310-5","volume":"128","author":"R Anirudh","year":"2020","unstructured":"Anirudh R, Thiagarajan J J, Kailkhura B, et al. MimicGAN: robust projection onto image manifolds with corruption mimicking. Int J Comput Vis, 2020, 128: 2459\u20132477","journal-title":"Int J Comput Vis"},{"key":"3679_CR114","doi-asserted-by":"publisher","first-page":"7474","DOI":"10.1109\/TPAMI.2021.3115428","volume":"44","author":"X Pan","year":"2022","unstructured":"Pan X, Zhan X, Dai B, et al. Exploiting deep generative prior for versatile image restoration and manipulation. IEEE Trans Pattern Anal Mach Intell, 2022, 44: 7474\u20137489","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"3679_CR115","doi-asserted-by":"crossref","unstructured":"Viazovetskyi Y, Ivashkin V, Kashin E. StyleGAN2 distillation for feed-forward image manipulation. In: Proceedings of European Conference on Computer Vision, 2020. 170\u2013186","DOI":"10.1007\/978-3-030-58542-6_11"},{"key":"3679_CR116","doi-asserted-by":"crossref","unstructured":"Collins E, Bala R, Price B, et al. Editing in style: uncovering the local semantics of GANs. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2020. 5771\u20135780","DOI":"10.1109\/CVPR42600.2020.00581"},{"key":"3679_CR117","doi-asserted-by":"crossref","unstructured":"Pidhorskyi S, Adjeroh D A, Doretto G. Adversarial latent autoencoders. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2020. 14104\u201314113","DOI":"10.1109\/CVPR42600.2020.01411"},{"key":"3679_CR118","doi-asserted-by":"crossref","unstructured":"Huh M, Zhang R, Zhu J Y, et al. Transforming and projecting images into class-conditional generative networks. In: Proceedings of European Conference on Computer Vision, 2020. 17\u201334","DOI":"10.1007\/978-3-030-58536-5_2"},{"key":"3679_CR119","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3414685.3417826","volume":"39","author":"Y Nitzan","year":"2020","unstructured":"Nitzan Y, Bermano A, Li Y, et al. Face identity disentanglement via latent space mapping. ACM Trans Graph, 2020, 39: 1\u201314","journal-title":"ACM Trans Graph"},{"key":"3679_CR120","unstructured":"Aberdam A, Simon D, Elad M. When and how can deep generative models be inverted? 2020. ArXiv:2006.15555"},{"key":"3679_CR121","unstructured":"Guan S, Tai Y, Ni B, et al. Collaborative learning for faster StyleGAN embedding. 2020. ArXiv:2007.01758"},{"key":"3679_CR122","doi-asserted-by":"crossref","unstructured":"Shen Y, Zhou B. Closed-form factorization of latent semantics in GANs. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2021. 1532\u20131540","DOI":"10.1109\/CVPR46437.2021.00158"},{"key":"3679_CR123","doi-asserted-by":"crossref","unstructured":"Xu Y, Shen Y, Zhu J, et al. Generative hierarchical features from synthesizing images. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2021. 4432\u20134442","DOI":"10.1109\/CVPR46437.2021.00441"},{"key":"3679_CR124","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3414685.3417803","volume":"39","author":"A Tewari","year":"2020","unstructured":"Tewari A, Elgharib M, R M B, et al. PIE: portrait image embedding for semantic control. ACM Trans Graph, 2020, 39: 1\u201314","journal-title":"ACM Trans Graph"},{"key":"3679_CR125","unstructured":"Bartz C, Bethge J, Yang H, et al. One model to reconstruct them all: a novel way to use the stochastic noise in StyleGAN. In: Proceedings of British Machine Vision Association, 2020"},{"key":"3679_CR126","doi-asserted-by":"crossref","unstructured":"Wang H P, Yu N, Fritz M. Hijack-GAN: unintended-use of pretrained, black-box GANs. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2021. 7872\u20137881","DOI":"10.1109\/CVPR46437.2021.00778"},{"key":"3679_CR127","unstructured":"Zhuang P, Koyejo O O, Schwing A. Enjoy your editing: controllable GANs for image editing via latent space navigation. In: Proceedings of International Conference on Learning Representations, 2021"},{"key":"3679_CR128","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3450626.3459805","volume":"40","author":"Y Alaluf","year":"2021","unstructured":"Alaluf Y, Patashnik O, Cohen-Or D. Only a matter of style: age transformation using a style-based regression model. ACM Trans Graph, 2021, 40: 1\u201312","journal-title":"ACM Trans Graph"},{"key":"3679_CR129","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3450626.3459838","volume":"40","author":"O Tov","year":"2021","unstructured":"Tov O, Alaluf Y, Nitzan Y, et al. Designing an encoder for StyleGAN image manipulation. ACM Trans Graph, 2021, 40: 1\u201314","journal-title":"ACM Trans Graph"},{"key":"3679_CR130","doi-asserted-by":"crossref","unstructured":"Patashnik O, Wu Z, Shechtman E, et al. StyleCLIP: text-driven manipulation of StyleGAN imagery. In: Proceedings of IEEE International Conference on Computer Vision, 2021. 2085\u20132094","DOI":"10.1109\/ICCV48922.2021.00209"},{"key":"3679_CR131","unstructured":"Chai L, Wulff J, Isola P. Using latent space regression to analyze and leverage compositionality in GANs. In: Proceedings of International Conference on Learning Representations, 2021"},{"key":"3679_CR132","doi-asserted-by":"crossref","unstructured":"Chai L, Zhu J Y, Shechtman E, et al. Ensembling with deep generative views. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2021. 14997\u201315007","DOI":"10.1109\/CVPR46437.2021.01475"},{"key":"3679_CR133","doi-asserted-by":"crossref","unstructured":"Alaluf Y, Patashnik O, Cohen-Or D. ReStyle: a residual-based StyleGAN encoder via iterative refinement. In: Proceedings of IEEE International Conference on Computer Vision, 2021. 6711\u20136720","DOI":"10.1109\/ICCV48922.2021.00664"},{"key":"3679_CR134","doi-asserted-by":"publisher","first-page":"3267","DOI":"10.1109\/TIP.2022.3167305","volume":"31","author":"T Wei","year":"2022","unstructured":"Wei T, Chen D, Zhou W, et al. E2Style: improve the efficiency and effectiveness of StyleGAN inversion. IEEE Trans Image Process, 2022, 31: 3267\u20133280","journal-title":"IEEE Trans Image Process"},{"key":"3679_CR135","doi-asserted-by":"crossref","unstructured":"Xu Y, Du Y, Xiao W, et al. From continuity to editability: inverting GANs with consecutive images. In: Proceedings of IEEE International Conference on Computer Vision, 2021. 13910\u201313918","DOI":"10.1109\/ICCV48922.2021.01365"},{"key":"3679_CR136","doi-asserted-by":"crossref","unstructured":"Wang T, Zhang Y, Fan Y, et al. High-fidelity GAN inversion for image attribute editing. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2022. 11379\u201311388","DOI":"10.1109\/CVPR52688.2022.01109"},{"key":"3679_CR137","doi-asserted-by":"crossref","unstructured":"Schwettmann S, Hernandez E, Bau D, et al. Toward a visual concept vocabulary for GAN latent space. In: Proceedings of IEEE International Conference on Computer Vision, 2021. 6804\u20136812","DOI":"10.1109\/ICCV48922.2021.00673"},{"key":"3679_CR138","doi-asserted-by":"crossref","unstructured":"Alaluf Y, Tov O, Mokady R, et al. HyperStyle: StyleGAN inversion with hypernetworks for real image editing. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2022. 18511\u201318521","DOI":"10.1109\/CVPR52688.2022.01796"},{"key":"3679_CR139","doi-asserted-by":"crossref","unstructured":"Peebles W, Zhu J Y, Zhang R, et al. GAN-supervised dense visual alignment. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2022. 13470\u201313481","DOI":"10.1109\/CVPR52688.2022.01311"},{"key":"3679_CR140","doi-asserted-by":"crossref","unstructured":"Dinh T M, Tran A T, Nguyen R, et al. HyperInverter: improving StyleGAN inversion via hypernetwork. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2022. 11389\u201311398","DOI":"10.1109\/CVPR52688.2022.01110"},{"key":"3679_CR141","doi-asserted-by":"crossref","unstructured":"Alaluf Y, Patashnik O, Wu Z, et al. Third time\u2019s the charm? Image and video editing with StyleGAN3. 2022. ArXiv:2201.13433","DOI":"10.1007\/978-3-031-25063-7_13"},{"key":"3679_CR142","doi-asserted-by":"crossref","unstructured":"Fr\u00fchst\u00fcck A, Singh K K, Shechtman E, et al. InsetGAN for full-body image generation. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2022. 7723\u20137732","DOI":"10.1109\/CVPR52688.2022.00757"},{"key":"3679_CR143","doi-asserted-by":"crossref","unstructured":"Wu Y, Yang Y L, Jin X. HairMapper: removing hair from portraits using gans. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2022. 4227\u20134236","DOI":"10.1109\/CVPR52688.2022.00419"},{"key":"3679_CR144","doi-asserted-by":"crossref","unstructured":"Parmar G, Li Y, Lu J, et al. Spatially-adaptive multilayer selection for GAN inversion and editing. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2022. 11399\u201311409","DOI":"10.1109\/CVPR52688.2022.01111"},{"key":"3679_CR145","doi-asserted-by":"crossref","unstructured":"Zhou B, Zhao H, Puig X, et al. Scene parsing through ADE20K dataset. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2017. 633\u2013641","DOI":"10.1109\/CVPR.2017.544"},{"key":"3679_CR146","doi-asserted-by":"crossref","unstructured":"Chen B C, Chen C S, Hsu W H. Cross-age reference coding for age-invariant face recognition and retrieval. In: Proceedings of European Conference on Computer Vision, 2014. 768\u2013783","DOI":"10.1007\/978-3-319-10599-4_49"},{"key":"3679_CR147","doi-asserted-by":"crossref","unstructured":"Lin T Y, Maire M, Belongie S, et al. Microsoft COCO: common objects in context. In: Proceedings of European Conference on Computer Vision, 2014. 740\u2013755","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"3679_CR148","unstructured":"Wah C, Branson S, Welinder P, et al. The Caltech-UCSD birds-200-2011 dataset. 2011. http:\/\/www.vision.caltech.edu\/visipedia\/CUB-200.html"},{"key":"3679_CR149","unstructured":"Anonymous, The Danbooru Community, Branwen G. Danbooru2021: a large-scale crowdsourced and tagged anime illustration dataset. 2021. https:\/\/www.gwern.net\/Danbooru"},{"key":"3679_CR150","doi-asserted-by":"crossref","unstructured":"Nilsback M E, Zisserman A. Automated flower classification over a large number of classes. In: Proceedings of the 6th Indian Conference on Computer Vision, Graphics & Image Processing, 2008. 722\u2013729","DOI":"10.1109\/ICVGIP.2008.47"},{"key":"3679_CR151","unstructured":"Huang G B, Mattar M, Berg T, et al. Labeled faces in the wild: a database forstudying face recognition in unconstrained environments. In: Proceedings of Workshop on Faces in \u2018Real-Life\u2019 Images: Detection, Alignment, and Recognition, 2008"},{"key":"3679_CR152","doi-asserted-by":"crossref","unstructured":"Skorokhodov I, Sotnikov G, Elhoseiny M. Aligning latent and image spaces to connect the unconnectable. In: Proceedings of IEEE International Conference on Computer Vision, 2021. 14144\u201314153","DOI":"10.1109\/ICCV48922.2021.01388"},{"key":"3679_CR153","doi-asserted-by":"publisher","first-page":"1332","DOI":"10.1126\/science.aab3050","volume":"350","author":"B M Lake","year":"2015","unstructured":"Lake B M, Salakhutdinov R, Tenenbaum J B. Human-level concept learning through probabilistic program induction. Science, 2015, 350: 1332\u20131338","journal-title":"Science"},{"key":"3679_CR154","unstructured":"Zhou B, Lapedriza A, Xiao J, et al. Learning deep features for scene recognition using places database. In: Proceedings of International Conference on Neural Information Processing Systems, 2014"},{"key":"3679_CR155","doi-asserted-by":"crossref","unstructured":"Parkhi O M, Vedaldi A, Zisserman A, et al. Cats and dogs. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2012. 3498\u20133505","DOI":"10.1109\/CVPR.2012.6248092"},{"key":"3679_CR156","doi-asserted-by":"publisher","first-page":"e0196391","DOI":"10.1371\/journal.pone.0196391","volume":"13","author":"S R Livingstone","year":"2018","unstructured":"Livingstone S R, Russo F A. The Ryerson Audio-Visual Database of Emotional Speech and Song (RAVDESS): a dynamic, multimodal set of facial and vocal expressions in North American English. Plos One, 2018, 13: e0196391","journal-title":"Plos One"},{"key":"3679_CR157","doi-asserted-by":"crossref","unstructured":"Krause J, Stark M, Deng J, et al. 3D object representations for fine-grained categorization. In: Proceedings of IEEE International Conference on Computer Vision Workshops, 2013. 554\u2013561","DOI":"10.1109\/ICCVW.2013.77"},{"key":"3679_CR158","doi-asserted-by":"crossref","unstructured":"Naik N, Philipoom J, Raskar R, et al. Streetscore-predicting the perceived safety of one million streetscapes. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition Workshops, 2014. 779\u2013785","DOI":"10.1109\/CVPRW.2014.121"},{"key":"3679_CR159","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/2601097.2601101","volume":"33","author":"P Y Laffont","year":"2014","unstructured":"Laffont P Y, Ren Z, Tao X, et al. Transient attributes for high-level understanding and editing of outdoor scenes. ACM Trans Graph, 2014, 33: 1\u201311","journal-title":"ACM Trans Graph"},{"key":"3679_CR160","doi-asserted-by":"crossref","unstructured":"Yu A, Grauman K. Fine-grained visual comparisons with local learning. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2014. 192\u2013199","DOI":"10.1109\/CVPR.2014.32"},{"key":"3679_CR161","doi-asserted-by":"publisher","first-page":"503","DOI":"10.1007\/BF01589116","volume":"45","author":"D C Liu","year":"1989","unstructured":"Liu D C, Nocedal J. On the limited memory BFGS method for large scale optimization. Math Programming, 1989, 45: 503\u2013528","journal-title":"Math Programming"},{"key":"3679_CR162","unstructured":"Kingma D P, Ba J. Adam: a method for stochastic optimization. 2014. ArXiv:1412.6980"},{"key":"3679_CR163","doi-asserted-by":"crossref","unstructured":"Deng J, Guo J, Xue N, et al. ArcFace: additive angular margin loss for deep face recognition. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2019. 4690\u20134699","DOI":"10.1109\/CVPR.2019.00482"},{"key":"3679_CR164","doi-asserted-by":"crossref","unstructured":"Huang Y, Wang Y, Tai Y, et al. CurricularFace: adaptive curriculum learning loss for deep face recognition. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2020. 5901\u20135910","DOI":"10.1109\/CVPR42600.2020.00594"},{"key":"3679_CR165","doi-asserted-by":"crossref","unstructured":"He K, Fan H, Wu Y, et al. Momentum contrast for unsupervised visual representation learning. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2020. 9729\u20139738","DOI":"10.1109\/CVPR42600.2020.00975"},{"key":"3679_CR166","unstructured":"Donahue J, Simonyan K. Large scale adversarial representation learning. In: Proceedings of International Conference on Neural Information Processing Systems, 2019. 32"},{"key":"3679_CR167","unstructured":"Kingma D P, Dhariwal P. Glow: generative flow with invertible 1 \u00d7 1 convolutions. In: Proceedings of International Conference on Neural Information Processing Systems, 2018. 10236\u201310245"},{"key":"3679_CR168","unstructured":"Ho J, Jain A, Abbeel P. Denoising diffusion probabilistic models. In: Proceedings of International Conference on Neural Information Processing Systems, 2020. 6840\u20136851"},{"key":"3679_CR169","doi-asserted-by":"crossref","unstructured":"Tousi A, Jeong H, Han J, et al. Automatic correction of internal units in generative neural networks. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2021. 7932\u20137940","DOI":"10.1109\/CVPR46437.2021.00784"},{"key":"3679_CR170","doi-asserted-by":"crossref","unstructured":"Bau D, Zhou B, Khosla A, et al. Network dissection: quantifying interpretability of deep visual representations. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2017. 6541\u20136549","DOI":"10.1109\/CVPR.2017.354"},{"key":"3679_CR171","doi-asserted-by":"crossref","unstructured":"Carter S, Armstrong Z, Schubert L, et al. Activation atlas. Distill, 2019. https:\/\/distill.pub\/2019\/activation-atlas","DOI":"10.23915\/distill.00015"},{"key":"3679_CR172","doi-asserted-by":"crossref","unstructured":"Bau D, Liu S, Wang T, et al. Rewriting a deep generative model. In: Proceedings of European Conference on Computer Vision, 2020. 351\u2013369","DOI":"10.1007\/978-3-030-58452-8_21"},{"key":"3679_CR173","doi-asserted-by":"publisher","first-page":"1377","DOI":"10.1080\/02699930903485076","volume":"24","author":"O Langner","year":"2010","unstructured":"Langner O, Dotsch R, Bijlstra G, et al. Presentation and validation of the Radboud Faces Database. Cognition Emotion, 2010, 24: 1377\u20131388","journal-title":"Cognition Emotion"},{"key":"3679_CR174","unstructured":"Ramesh A, Choi Y, LeCun Y. A spectral regularizer for unsupervised disentanglement. 2018. ArXiv:1812.01161"},{"key":"3679_CR175","unstructured":"Chen X, Duan Y, Houthooft R, et al. InfoGAN: interpretable representation learning by information maximizing generative adversarial nets. In: Proceedings of International Conference on Neural Information Processing Systems, 2016. 2172\u20132180"},{"key":"3679_CR176","doi-asserted-by":"crossref","unstructured":"Peebles W, Peebles J, Zhu J Y, et al. The hessian penalty: a weak prior for unsupervised disentanglement. In: Proceedings of European Conference on Computer Vision, 2020. 581\u2013597","DOI":"10.1007\/978-3-030-58539-6_35"},{"key":"3679_CR177","doi-asserted-by":"crossref","unstructured":"Zhu X, Xu C, Tao D. Learning disentangled representations with latent variation predictability. In: Proceedings of European Conference on Computer Vision, 2020. 684\u2013700","DOI":"10.1007\/978-3-030-58607-2_40"},{"key":"3679_CR178","doi-asserted-by":"crossref","unstructured":"Zhu X, Xu C, Tao D. Where and what? Examining interpretable disentangled representations. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2021. 5861\u20135870","DOI":"10.1109\/CVPR46437.2021.00580"},{"key":"3679_CR179","doi-asserted-by":"crossref","unstructured":"Wei Y, Shi Y, Liu X, et al. Orthogonal jacobian regularization for unsupervised disentanglement in image generation. In: Proceedings of IEEE International Conference on Computer Vision, 2021. 6721\u20136730","DOI":"10.1109\/ICCV48922.2021.00665"},{"key":"3679_CR180","doi-asserted-by":"crossref","unstructured":"He Z, Kan M, Shan S. EigenGAN: layer-wise eigen-learning for GANs. In: Proceedings of IEEE International Conference on Computer Vision, 2021. 14408\u201314417","DOI":"10.1109\/ICCV48922.2021.01414"},{"key":"3679_CR181","unstructured":"Jahanian A, Chai L, Isola P. On the \u201csteerability\u201d of generative adversarial networks. In: Proceedings of International Conference on Learning Representations, 2020"},{"key":"3679_CR182","unstructured":"Zhu J, Shen Y, Xu Y, et al. Region-based semantic factorization in GANs. 2022. ArXiv:2202.09649"},{"key":"3679_CR183","unstructured":"Wang B, Ponce C R. A geometric analysis of deep generative image models and its applications. In: Proceedings of International Conference on Learning Representations, 2021"},{"key":"3679_CR184","doi-asserted-by":"crossref","unstructured":"Tzelepis C, Tzimiropoulos G, Patras I. WarpedGANSpace: finding non-linear RBF paths in GAN latent space. In: Proceedings of IEEE International Conference on Computer Vision, 2021. 6393\u20136402","DOI":"10.1109\/ICCV48922.2021.00633"},{"key":"3679_CR185","doi-asserted-by":"crossref","unstructured":"Wang X, Yu K, Dong C, et al. Deep network interpolation for continuous imagery effect transition. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2019. 1692\u20131701","DOI":"10.1109\/CVPR.2019.00179"},{"key":"3679_CR186","doi-asserted-by":"crossref","unstructured":"Selvaraju R R, Cogswell M, Das A, et al. Grad-CAM: visual explanations from deep networks via gradient-based localization. In: Proceedings of IEEE International Conference on Computer Vision, 2017. 618\u2013626","DOI":"10.1109\/ICCV.2017.74"},{"key":"3679_CR187","unstructured":"Pan X, Dai B, Liu Z, et al. Do 2D GANs know 3D shape? Unsupervised 3D shape reconstruction from 2D image GANs. 2020. ArXiv:2011.00844"},{"key":"3679_CR188","doi-asserted-by":"crossref","unstructured":"Zhang J, Chen X, Cai Z, et al. Unsupervised 3D shape completion through GAN inversion. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2021. 1768\u20131777","DOI":"10.1109\/CVPR46437.2021.00181"},{"key":"3679_CR189","unstructured":"Kingma D P, Welling M. Auto-encoding variational Bayes. 2013. ArXiv:1312.6114"},{"key":"3679_CR190","unstructured":"van den Oord A, Kalchbrenner N, Kavukcuoglu K. Pixel recurrent neural networks. In: Proceedings of International Conference on Machine Learning, 2016. 1747\u20131756"},{"key":"3679_CR191","unstructured":"Ramesh A, Dhariwal P, Nichol A, et al. Hierarchical text-conditional image generation with clip latents. 2022. ArXiv:2204.06125"},{"key":"3679_CR192","doi-asserted-by":"crossref","unstructured":"Saharia C, Chan W, Saxena S, et al. Photorealistic text-to-image diffusion models with deep language understanding. 2022. ArXiv:2205.11487","DOI":"10.1145\/3528233.3530757"},{"key":"3679_CR193","first-page":"5866","volume":"44","author":"D Zhang","year":"2021","unstructured":"Zhang D, Han J, Cheng G, et al. Weakly supervised object localization and detection: a survey. IEEE Trans Pattern Anal Mach Intell, 2021, 44: 5866\u20135885","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"3679_CR194","doi-asserted-by":"publisher","first-page":"84","DOI":"10.1109\/MSP.2017.2749125","volume":"35","author":"J Han","year":"2018","unstructured":"Han J, Zhang D, Cheng G, et al. Advanced deep-learning techniques for salient and category-specific object detection: a survey. IEEE Signal Process Mag, 2018, 35: 84\u2013100","journal-title":"IEEE Signal Process Mag"},{"key":"3679_CR195","unstructured":"Zhang D, Tian H, Han J. Few-cost salient object detection with adversarial-paced learning. In: Proceedings of International Conference on Neural Information Processing Systems, 2020. 33: 12236\u201312247"},{"key":"3679_CR196","doi-asserted-by":"crossref","unstructured":"Frid-Adar M, Klang E, Amitai M, et al. Synthetic data augmentation using GAN for improved liver lesion classification. In: Proceedings of International Symposium on Biomedical Imaging, 2018. 289\u2013293","DOI":"10.1109\/ISBI.2018.8363576"},{"key":"3679_CR197","doi-asserted-by":"crossref","unstructured":"Huang S W, Lin C T, Chen S P, et al. AugGAN: cross domain adaptation with gan-based data augmentation. In: Proceedings of European Conference on Computer Vision, 2018. 718\u2013731","DOI":"10.1007\/978-3-030-01240-3_44"},{"key":"3679_CR198","doi-asserted-by":"crossref","unstructured":"Zhang Y, Ling H, Gao J, et al. DatasetGAN: efficient labeled data factory with minimal human effort. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2021. 10145\u201310155","DOI":"10.1109\/CVPR46437.2021.01001"},{"key":"3679_CR199","doi-asserted-by":"crossref","unstructured":"Han M, Zheng H, Wang C, et al. Leveraging GAN priors for few-shot part segmentation. In: Proceedings of ACM International Conference on Multimedia, 2022. 1339\u20131347","DOI":"10.1145\/3503161.3548398"},{"key":"3679_CR200","doi-asserted-by":"publisher","first-page":"30","DOI":"10.1016\/j.media.2019.01.010","volume":"54","author":"T Schlegl","year":"2019","unstructured":"Schlegl T, Seeb\u00f6ck P, Waldstein S M, et al. f-AnoGAN: fast unsupervised anomaly detection with generative adversarial networks. Med Image Anal, 2019, 54: 30\u201344","journal-title":"Med Image Anal"},{"key":"3679_CR201","unstructured":"Dunn I, Pouget H, Melham T, et al. Adaptive generation of unrestricted adversarial inputs. 2019. ArXiv:1905.02463"},{"key":"3679_CR202","unstructured":"Wang X, He K, Hopcroft J E. At-GAN: a generative attack model for adversarial transferring on generative adversarial nets. 2019. ArXiv:1904.07793"},{"key":"3679_CR203","doi-asserted-by":"crossref","unstructured":"Ojha U, Li Y, Lu J, et al. Few-shot image generation via cross-domain correspondence. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2021. 10743\u201310752","DOI":"10.1109\/CVPR46437.2021.01060"},{"key":"3679_CR204","unstructured":"Gu J, Liu L, Wang P, et al. StyleNeRF: a style-based 3D aware generator for high-resolution image synthesis. In: Proceedings of International Conference on Learning Representations, 2022"},{"key":"3679_CR205","doi-asserted-by":"crossref","unstructured":"He J, Shi W, Chen K, et al. GCFSR: a generative and controllable face super resolution method without facial and GAN priors. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2022. 1889\u20131898","DOI":"10.1109\/CVPR52688.2022.00193"},{"key":"3679_CR206","doi-asserted-by":"crossref","unstructured":"Li X, Chen C, Lin X, et al. From face to natural image: learning real degradation for blind image super-resolution. In: Proceedings of European Conference on Computer Vision, 2022","DOI":"10.1007\/978-3-031-19797-0_22"},{"key":"3679_CR207","doi-asserted-by":"crossref","unstructured":"Li B, Liu X, Hu P, et al. All-in-one image restoration for unknown corruption. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 2022. 17452\u201317462","DOI":"10.1109\/CVPR52688.2022.01693"},{"key":"3679_CR208","unstructured":"Lyu Z, Xu X, Yang C, et al. Accelerating diffusion models via early stop of the diffusion process. 2022. ArXiv:2205.12524"},{"key":"3679_CR209","doi-asserted-by":"crossref","unstructured":"Grover A, Dhar M, Ermon S. Flow-GAN: combining maximum likelihood and adversarial learning in generative models. In: Proceedings of AAAI Conference on Artificial Intelligence, 2018","DOI":"10.1609\/aaai.v32i1.11829"}],"container-title":["Science China Information Sciences"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11432-022-3679-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11432-022-3679-0\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11432-022-3679-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,18]],"date-time":"2024-10-18T07:30:24Z","timestamp":1729236624000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11432-022-3679-0"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,4,4]]},"references-count":209,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2023,5]]}},"alternative-id":["3679"],"URL":"https:\/\/doi.org\/10.1007\/s11432-022-3679-0","relation":{},"ISSN":["1674-733X","1869-1919"],"issn-type":[{"value":"1674-733X","type":"print"},{"value":"1869-1919","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,4,4]]},"assertion":[{"value":"29 July 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"25 October 2022","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"7 December 2022","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"4 April 2023","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}],"article-number":"151101"}}