{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,27]],"date-time":"2026-01-27T21:00:53Z","timestamp":1769547653763,"version":"3.49.0"},"reference-count":43,"publisher":"Springer Science and Business Media LLC","issue":"5","license":[{"start":{"date-parts":[[2023,8,22]],"date-time":"2023-08-22T00:00:00Z","timestamp":1692662400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,8,22]],"date-time":"2023-08-22T00:00:00Z","timestamp":1692662400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/100000001","name":"National Science Foundation","doi-asserted-by":"publisher","award":["IIS 1527200 and 1941613"],"award-info":[{"award-number":["IIS 1527200 and 1941613"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2024,5]]},"DOI":"10.1007\/s00371-023-03051-1","type":"journal-article","created":{"date-parts":[[2023,8,22]],"date-time":"2023-08-22T17:02:47Z","timestamp":1692723767000},"page":"3553-3568","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":9,"title":["Transforming the latent space of StyleGAN for real face editing"],"prefix":"10.1007","volume":"40","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-2979-0849","authenticated-orcid":false,"given":"Heyi","family":"Li","sequence":"first","affiliation":[]},{"given":"Jinlong","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Xinyu","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Yunzhi","family":"Bai","sequence":"additional","affiliation":[]},{"given":"Huayan","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Klaus","family":"Mueller","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,8,22]]},"reference":[{"issue":"11","key":"3051_CR1","doi-asserted-by":"publisher","first-page":"139","DOI":"10.1145\/3422622","volume":"63","author":"I Goodfellow","year":"2020","unstructured":"Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., Bengio, Y.: Generative adversarial networks. Commun. ACM 63(11), 139\u2013144 (2020)","journal-title":"Commun. ACM"},{"key":"3051_CR2","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aila, T.: A style-based generator architecture for generative adversarial networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4401\u20134410 (2019)","DOI":"10.1109\/CVPR.2019.00453"},{"key":"3051_CR3","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aittala, M., Hellsten, J., Lehtinen, J., Aila, T.: Analyzing and improving the image quality of stylegan. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8110\u20138119 (2020)","DOI":"10.1109\/CVPR42600.2020.00813"},{"key":"3051_CR4","doi-asserted-by":"crossref","unstructured":"Shen, Y., Gu, J., Tang, X., Zhou, B.: Interpreting the latent space of gans for semantic face editing. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 9243\u20139252 (2020)","DOI":"10.1109\/CVPR42600.2020.00926"},{"key":"3051_CR5","unstructured":"Shen, Y., Yang, C., Tang, X., Zhou, B.: Interfacegan: interpreting the disentangled face representation learned by gans. IEEE Trans. Pattern Anal. Mach. Intell. (2020)"},{"key":"3051_CR6","first-page":"9841","volume":"33","author":"E H\u00e4rk\u00f6nen","year":"2020","unstructured":"H\u00e4rk\u00f6nen, E., Hertzmann, A., Lehtinen, J., Paris, S.: Ganspace: discovering interpretable gan controls. Adv. Neural Inf. Process. Syst. 33, 9841\u20139850 (2020)","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"3051_CR7","doi-asserted-by":"crossref","unstructured":"Collins, E., Bala, R., Price, B., Susstrunk, S.: Editing in style: uncovering the local semantics of gans. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5771\u20135780 (2020)","DOI":"10.1109\/CVPR42600.2020.00581"},{"key":"3051_CR8","doi-asserted-by":"crossref","unstructured":"Shoshan, A., Bhonker, N., Kviatkovsky, I., Medioni, G.: Gan-control: explicitly controllable gans. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 14083\u201314093 (2021)","DOI":"10.1109\/ICCV48922.2021.01382"},{"key":"3051_CR9","doi-asserted-by":"crossref","unstructured":"Su, W., Ye, H., Chen, S.-Y., Gao, L., Fu, H.: Drawinginstyles: portrait image generation and editing with spatially conditioned stylegan. IEEE Trans. Vis. Comput. Graph. (2022)","DOI":"10.1109\/TVCG.2022.3178734"},{"key":"3051_CR10","doi-asserted-by":"crossref","unstructured":"Shi, Y., Yang, X., Wan, Y., Shen, X.: Semanticstylegan: learning compositional generative priors for controllable image synthesis and editing. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11254\u201311264 (2022)","DOI":"10.1109\/CVPR52688.2022.01097"},{"key":"3051_CR11","doi-asserted-by":"crossref","unstructured":"Abdal, R., Qin, Y., Wonka, P.: Image2stylegan++: how to edit the embedded images? In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8296\u20138305 (2020)","DOI":"10.1109\/CVPR42600.2020.00832"},{"key":"3051_CR12","doi-asserted-by":"crossref","unstructured":"Zhu, J., Shen, Y., Zhao, D., Zhou, B.: In-domain gan inversion for real image editing. In: European Conference on Computer Vision, pp. 592\u2013608. Springer (2020)","DOI":"10.1007\/978-3-030-58520-4_35"},{"issue":"3","key":"3051_CR13","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3447648","volume":"40","author":"R Abdal","year":"2021","unstructured":"Abdal, R., Zhu, P., Mitra, N.J., Wonka, P.: Styleflow: attribute-conditioned exploration of stylegan-generated images using conditional continuous normalizing flows. ACM Trans. Graph. (ToG) 40(3), 1\u201321 (2021)","journal-title":"ACM Trans. Graph. (ToG)"},{"issue":"6","key":"3051_CR14","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3414685.3417803","volume":"39","author":"A Tewari","year":"2020","unstructured":"Tewari, A., Elgharib, M., Bernard, F., Seidel, H.-P., P\u00e9rez, P., Zollh\u00f6fer, M., Theobalt, C.: Pie: portrait image embedding for semantic control. ACM Trans. Graph. (TOG) 39(6), 1\u201314 (2020)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"3051_CR15","doi-asserted-by":"publisher","first-page":"209","DOI":"10.1016\/j.neunet.2021.10.017","volume":"145","author":"X Hou","year":"2022","unstructured":"Hou, X., Zhang, X., Liang, H., Shen, L., Lai, Z., Wan, J.: Guidedstyle: attribute knowledge guided style manipulation for semantic face editing. Neural Netw. 145, 209\u2013220 (2022)","journal-title":"Neural Netw."},{"key":"3051_CR16","doi-asserted-by":"crossref","unstructured":"Abdal, R., Qin, Y., Wonka, P.: Image2stylegan: how to embed images into the stylegan latent space? In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 4432\u20134441 (2019)","DOI":"10.1109\/ICCV.2019.00453"},{"key":"3051_CR17","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, \u0141., Polosukhin, I.: Attention is all you need. Adv. Neural Inf. Process. Syst. 30 (2017)"},{"key":"3051_CR18","doi-asserted-by":"crossref","unstructured":"Richardson, E., Alaluf, Y., Patashnik, O., Nitzan, Y., Azar, Y., Shapiro, S., Cohen-Or, D.: Encoding in style: a stylegan encoder for image-to-image translation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2287\u20132296 (2021)","DOI":"10.1109\/CVPR46437.2021.00232"},{"key":"3051_CR19","unstructured":"Mirza, M., Osindero, S.: Conditional generative adversarial nets. arXiv preprint arXiv:1411.1784 (2014)"},{"key":"3051_CR20","doi-asserted-by":"crossref","unstructured":"Xia, W., Zhang, Y., Yang, Y., Xue, J.-H., Zhou, B., Yang, M.-H.: Gan inversion: a survey. IEEE Trans. Pattern Anal. Mach. Intell. (2022)","DOI":"10.1109\/TPAMI.2022.3181070"},{"key":"3051_CR21","unstructured":"robertluxemburg: Git repository: stylegan2encoder (2020). https:\/\/github.com\/robertluxemburg\/stylegan2encoder"},{"issue":"4","key":"3051_CR22","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3450626.3459838","volume":"40","author":"O Tov","year":"2021","unstructured":"Tov, O., Alaluf, Y., Nitzan, Y., Patashnik, O., Cohen-Or, D.: Designing an encoder for stylegan image manipulation. ACM Trans. Graph. (TOG) 40(4), 1\u201314 (2021)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"3051_CR23","doi-asserted-by":"crossref","unstructured":"Alaluf, Y., Patashnik, O., Cohen-Or, D.: Restyle: a residual-based stylegan encoder via iterative refinement. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV) (2021)","DOI":"10.1109\/ICCV48922.2021.00664"},{"key":"3051_CR24","doi-asserted-by":"crossref","unstructured":"Roich, D., Mokady, R., Bermano, A.H., Cohen-Or, D.: Pivotal tuning for latent-based editing of real images. ACM Trans. Graph. (2021)","DOI":"10.1145\/3544777"},{"key":"3051_CR25","doi-asserted-by":"crossref","unstructured":"Alaluf, Y., Tov, O., Mokady, R., Gal, R., Bermano, A.: Hyperstyle: stylegan inversion with hypernetworks for real image editing. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 18511\u201318521 (2022)","DOI":"10.1109\/CVPR52688.2022.01796"},{"key":"3051_CR26","doi-asserted-by":"crossref","unstructured":"Tewari, A., Elgharib, M., Bharaj, G., Bernard, F., Seidel, H.-P., P\u00e9rez, P., Zollhofer, M., Theobalt, C.: Stylerig: rigging stylegan for 3d control over portrait images. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 6142\u20136151 (2020)","DOI":"10.1109\/CVPR42600.2020.00618"},{"issue":"9","key":"3051_CR27","doi-asserted-by":"publisher","first-page":"2907","DOI":"10.1007\/s00371-021-02198-z","volume":"37","author":"Y Ju","year":"2021","unstructured":"Ju, Y., Zhang, J., Mao, X., Xu, J.: Adaptive semantic attribute decoupling for precise face image editing. Vis. Comput. 37(9), 2907\u20132918 (2021)","journal-title":"Vis. Comput."},{"key":"3051_CR28","doi-asserted-by":"crossref","unstructured":"Lin, C., Xiong, S., Lu, X.: Disentangled face editing via individual walk in personalized facial semantic field. Vis. Comput. 1\u201310 (2022)","DOI":"10.1007\/s00371-022-02708-7"},{"key":"3051_CR29","doi-asserted-by":"crossref","unstructured":"Shen, Y., Zhou, B.: Closed-form factorization of latent semantics in gans. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 1532\u20131540 (2021)","DOI":"10.1109\/CVPR46437.2021.00158"},{"key":"3051_CR30","unstructured":"Zhu, P., Abdal, R., Qin, Y., Wonka, P.: Improved stylegan embedding: where are the good latents? arXiv preprint arXiv:2012.09036 (2020)"},{"key":"3051_CR31","unstructured":"Liu, Y., Li, Q., Sun, Z., Tan, T.: Style intervention: How to achieve spatial disentanglement with style-based generators? arXiv preprint arXiv:2011.09699 (2020)"},{"key":"3051_CR32","doi-asserted-by":"crossref","unstructured":"Wu, Z., Lischinski, D., Shechtman, E.: Stylespace analysis: disentangled controls for stylegan image generation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12863\u201312872 (2021)","DOI":"10.1109\/CVPR46437.2021.01267"},{"key":"3051_CR33","doi-asserted-by":"crossref","unstructured":"Zhang, B., Gu, S., Zhang, B., Bao, J., Chen, D., Wen, F., Wang, Y., Guo, B.: Styleswin: transformer-based gan for high-resolution image generation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11304\u201311314 (2022)","DOI":"10.1109\/CVPR52688.2022.01102"},{"key":"3051_CR34","doi-asserted-by":"crossref","unstructured":"Xu, Y., Yin, Y., Jiang, L., Wu, Q., Zheng, C., Loy, C.C., Dai, B., Wu, W.: Transeditor: transformer-based dual-space gan for highly controllable facial editing. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 7683\u20137692 (2022)","DOI":"10.1109\/CVPR52688.2022.00753"},{"key":"3051_CR35","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., Uszkoreit, J., Houlsby, N.: An image is worth 16x16 words: transformers for image recognition at scale. In: International Conference on Learning Representations (2021)"},{"key":"3051_CR36","unstructured":"Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Hochreiter, S.: Gans trained by a two time-scale update rule converge to a local nash equilibrium. Adv. Neural Inf. Process. Syst. 30 (2017)"},{"key":"3051_CR37","unstructured":"Kynk\u00e4\u00e4nniemi, T., Karras, T., Laine, S., Lehtinen, J., Aila, T.: Improved precision and recall metric for assessing generative models. Adv. Neural Inf. Process. Syst. 32 (2019)"},{"key":"3051_CR38","unstructured":"Naeem, M.F., Oh, S.J., Uh, Y., Choi, Y., Yoo, J.: Reliable fidelity and diversity metrics for generative models. In: International Conference on Machine Learning, pp. 7176\u20137185. PMLR (2020)"},{"key":"3051_CR39","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., Efros, A.A., Shechtman, E., Wang, O.: The unreasonable effectiveness of deep features as a perceptual metric. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 586\u2013595 (2018)","DOI":"10.1109\/CVPR.2018.00068"},{"key":"3051_CR40","doi-asserted-by":"crossref","unstructured":"Rothe, R., Timofte, R., Gool, L.V.: Dex: deep expectation of apparent age from a single image. In: IEEE International Conference on Computer Vision Workshops (ICCVW) (2015)","DOI":"10.1109\/ICCVW.2015.41"},{"key":"3051_CR41","doi-asserted-by":"crossref","unstructured":"Deng, Y., Yang, J., Xu, S., Chen, D., Jia, Y., Tong, X.: Accurate 3d face reconstruction with weakly-supervised learning: from single image to image set. In: IEEE Computer Vision and Pattern Recognition Workshops (2019)","DOI":"10.1109\/CVPRW.2019.00038"},{"key":"3051_CR42","doi-asserted-by":"publisher","DOI":"10.1109\/AVSS.2009.58","volume-title":"A 3D Face Model for Pose and Illumination Invariant Face Recognition","author":"P Paysan","year":"2009","unstructured":"Paysan, P., Knothe, R., Amberg, B., Romdhani, S., Vetter, T.: A 3D Face Model for Pose and Illumination Invariant Face Recognition. IEEE, Genova, Italy (2009)"},{"issue":"6","key":"3051_CR43","doi-asserted-by":"publisher","first-page":"1294","DOI":"10.1109\/TPAMI.2018.2837742","volume":"41","author":"Y Guo","year":"2019","unstructured":"Guo, Y., Zhang, J., Cai, J., Jiang, B., Zheng, J.: Cnn-based real-time dense face reconstruction with inverse-rendered photo-realistic face images. IEEE Trans. Pattern Anal. Mach. Intell. 41(6), 1294\u20131307 (2019)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-023-03051-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-023-03051-1\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-023-03051-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,4,19]],"date-time":"2024-04-19T09:13:59Z","timestamp":1713518039000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-023-03051-1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,8,22]]},"references-count":43,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2024,5]]}},"alternative-id":["3051"],"URL":"https:\/\/doi.org\/10.1007\/s00371-023-03051-1","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"value":"0178-2789","type":"print"},{"value":"1432-2315","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,8,22]]},"assertion":[{"value":"28 June 2023","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"22 August 2023","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they do not have any commercial or associative interest that represents a conflict of interest in connection with the work submitted.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}