{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,7,1]],"date-time":"2025-07-01T16:46:17Z","timestamp":1751388377568,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":49,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,7,13]],"date-time":"2024-07-13T00:00:00Z","timestamp":1720828800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0\/"}],"funder":[{"DOI":"10.13039\/https:\/\/doi.org\/10.13039\/501100003725","name":"National Research Foundation of Korea","doi-asserted-by":"publisher","award":["RS-2023-00280400"],"award-info":[{"award-number":["RS-2023-00280400"]}],"id":[{"id":"10.13039\/https:\/\/doi.org\/10.13039\/501100003725","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/https:\/\/doi.org\/10.13039\/501100010418","name":"Institute for Information and communications Technology Promotion","doi-asserted-by":"publisher","award":["2019-0-01906;2021-0-02068"],"award-info":[{"award-number":["2019-0-01906;2021-0-02068"]}],"id":[{"id":"10.13039\/https:\/\/doi.org\/10.13039\/501100010418","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,7,13]]},"DOI":"10.1145\/3641519.3657480","type":"proceedings-article","created":{"date-parts":[[2024,7,12]],"date-time":"2024-07-12T10:39:28Z","timestamp":1720780768000},"page":"1-11","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":1,"title":["Toonify3D: StyleGAN-based 3D Stylized Face Generator"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-1442-9399","authenticated-orcid":false,"given":"Wonjong","family":"Jang","sequence":"first","affiliation":[{"name":"POSTECH, South Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1593-4626","authenticated-orcid":false,"given":"Yucheol","family":"Jung","sequence":"additional","affiliation":[{"name":"POSTECH, South Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2162-4627","authenticated-orcid":false,"given":"Hyomin","family":"Kim","sequence":"additional","affiliation":[{"name":"POSTECH, South Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1006-8351","authenticated-orcid":false,"given":"Gwangjin","family":"Ju","sequence":"additional","affiliation":[{"name":"POSTECH, South Korea"}]},{"ORCID":"https:\/\/orcid.org\/0009-0009-8329-3083","authenticated-orcid":false,"given":"Chaewon","family":"Son","sequence":"additional","affiliation":[{"name":"POSTECH, South Korea"}]},{"ORCID":"https:\/\/orcid.org\/0009-0003-6916-9912","authenticated-orcid":false,"given":"Jooeun","family":"Son","sequence":"additional","affiliation":[{"name":"POSTECH, South Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8159-4271","authenticated-orcid":false,"given":"Seungyong","family":"Lee","sequence":"additional","affiliation":[{"name":"POSTECH, South Korea"}]}],"member":"320","published-online":{"date-parts":[[2024,7,13]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"3DScanStore. 2023. 3DScanStore. https:\/\/www.3dscanstore.com"},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00442"},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"crossref","unstructured":"Rameen Abdal Peihao Zhu Niloy\u00a0J. Mitra and Peter Wonka. 2021. StyleFlow: Attribute-Conditioned Exploration of StyleGAN-Generated Images Using Conditional Continuous Normalizing Flows. ACM Trans. Graph. (2021).","DOI":"10.1145\/3447648"},{"key":"e_1_3_2_1_4_1","volume-title":"Proc. CVPR.","author":"Abrevaya Victoria\u00a0Fernandez","year":"2020","unstructured":"Victoria\u00a0Fernandez Abrevaya, Adnane Boukhayma, Philip\u00a0H.S. Torr, and Edmond Boyer. 2020. Cross-Modal Deep Face Normals With Deactivable Skip Connections. In Proc. CVPR."},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2007.383165"},{"key":"e_1_3_2_1_6_1","unstructured":"David Beniaguev. 2022. Synthetic Faces High Quality (SFHQ) dataset. https:\/\/github.com\/SelfishGene\/SFHQ-dataset"},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1145\/311535.311556"},{"key":"e_1_3_2_1_8_1","unstructured":"Blender. 2023. Blender - a 3D modelling and rendering package. http:\/\/www.blender.org"},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01565"},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00574"},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW.2019.00038"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1145\/3528223.3530164"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"crossref","unstructured":"Pablo Garrido Michael Zollh\u00f6fer Dan Casas Levi Valgaerts Kiran Varanasi Patrick P\u00e9rez and Christian Theobalt. 2016. Reconstruction of personalized 3D face rigs from monocular video. ACM Trans. Graph. (2016).","DOI":"10.1145\/2890493"},{"key":"e_1_3_2_1_14_1","volume-title":"Proc. ICLR.","author":"Gu Jiatao","year":"2021","unstructured":"Jiatao Gu, Lingjie Liu, Peng Wang, and Christian Theobalt. 2021. StyleNeRF: A style-based 3D aware generator for high-resolution image synthesis. In Proc. ICLR."},{"key":"e_1_3_2_1_15_1","unstructured":"Xiaoguang Han Chang Gao and Yizhou Yu. 2017. DeepSketch2Face: a deep learning based sketching system for 3D face and caricature modeling. ACM Trans. Graph. (2017)."},{"key":"e_1_3_2_1_16_1","volume-title":"Example-based photometric stereo: Shape reconstruction with general, varying brdfs","author":"Hertzmann Aaron","year":"2005","unstructured":"Aaron Hertzmann and Steven\u00a0M Seitz. 2005. Example-based photometric stereo: Shape reconstruction with general, varying brdfs. IEEE TPAMI (2005)."},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"crossref","unstructured":"Wonjong Jang Gwangjin Ju Yucheol Jung Jiaolong Yang Xin Tong and Seungyong Lee. 2021. StyleCariGAN: Caricature generation via StyleGAN feature map modulation. ACM Trans. Graph. (2021).","DOI":"10.1145\/3476576.3476683"},{"key":"e_1_3_2_1_18_1","volume-title":"Pixel-in-Pixel Net: Towards Efficient Facial Landmark Detection in the Wild. IJCV","author":"Jin Haibo","year":"2021","unstructured":"Haibo Jin, Shengcai Liao, and Ling Shao. 2021. Pixel-in-Pixel Net: Towards Efficient Facial Landmark Detection in the Wild. IJCV (2021)."},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1145\/3550469.3555422"},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1145\/3528233.3530748"},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00453"},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00813"},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01037"},{"key":"e_1_3_2_1_24_1","volume-title":"Proc. CVPR.","author":"Lan Yushi","year":"2023","unstructured":"Yushi Lan, Xuyi Meng, Shuai Yang, Chen\u00a0Change Loy, and Bo Dai. 2023. E3DGE: Self-Supervised Geometry-Aware Encoder for Style-based 3D GAN Inversion. In Proc. CVPR."},{"key":"e_1_3_2_1_25_1","unstructured":"Tianye Li Timo Bolkart Michael\u00a0J Black Hao Li and Javier Romero. 2017. Learning a model of facial shape and expression from 4D scans.ACM Trans. Graph. (2017)."},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01864"},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01060"},{"key":"e_1_3_2_1_28_1","volume-title":"Segmentation in style: Unsupervised semantic image segmentation with stylegan and clip. arXiv preprint arXiv:2107.12518","author":"Pakhomov Daniil","year":"2021","unstructured":"Daniil Pakhomov, Sanchit Hira, Narayani Wagle, Kemar\u00a0E Green, and Nassir Navab. 2021. Segmentation in style: Unsupervised semantic image segmentation with stylegan and clip. arXiv preprint arXiv:2107.12518 (2021)."},{"key":"e_1_3_2_1_29_1","volume-title":"Do 2d gans know 3d shape? unsupervised 3d shape reconstruction from 2d image gans. arXiv preprint arXiv:2011.00844","author":"Pan Xingang","year":"2020","unstructured":"Xingang Pan, Bo Dai, Ziwei Liu, Chen\u00a0Change Loy, and Ping Luo. 2020. Do 2d gans know 3d shape? unsupervised 3d shape reconstruction from 2d image gans. arXiv preprint arXiv:2011.00844 (2020)."},{"key":"e_1_3_2_1_30_1","volume-title":"A shading-guided generative implicit model for shape-accurate 3d-aware image synthesis. NeurIPS","author":"Pan Xingang","year":"2021","unstructured":"Xingang Pan, Xudong Xu, Chen\u00a0Change Loy, Christian Theobalt, and Bo Dai. 2021. A shading-guided generative implicit model for shape-accurate 3d-aware image synthesis. NeurIPS (2021)."},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1109\/AVSS.2009.58"},{"key":"e_1_3_2_1_32_1","volume-title":"Proc. NeurIPS workshop on Machine Learning for Creativity and Design.","author":"Pinkney NM","year":"2020","unstructured":"Justin\u00a0NM Pinkney and Doron Adler. 2020. Resolution dependent GAN interpolation for controllable image synthesis between domains. In Proc. NeurIPS workshop on Machine Learning for Creativity and Design."},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01010"},{"key":"e_1_3_2_1_34_1","volume-title":"Proc. ICML.","author":"Radford Alec","year":"2021","unstructured":"Alec Radford, Jong\u00a0Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, 2021. Learning transferable visual models from natural language supervision. In Proc. ICML."},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00232"},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"crossref","unstructured":"Daniel Roich Ron Mokady Amit\u00a0H Bermano and Daniel Cohen-Or. 2022. Pivotal Tuning for Latent-based Editing of Real Images. ACM Trans. Graph. (2022).","DOI":"10.1145\/3544777"},{"key":"e_1_3_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.175"},{"key":"e_1_3_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00659"},{"key":"e_1_3_2_1_39_1","volume-title":"Interfacegan: Interpreting the disentangled face representation learned by gans","author":"Shen Yujun","year":"2020","unstructured":"Yujun Shen, Ceyuan Yang, Xiaoou Tang, and Bolei Zhou. 2020. Interfacegan: Interpreting the disentangled face representation learned by gans. IEEE TPAMI (2020)."},{"key":"e_1_3_2_1_40_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00619"},{"key":"e_1_3_2_1_41_1","volume-title":"Proc. ICLR.","author":"Simonyan Karen","year":"2015","unstructured":"Karen Simonyan and Andrew Zisserman. 2015. Very deep convolutional networks for large-scale image recognition. In Proc. ICLR."},{"key":"e_1_3_2_1_42_1","volume-title":"Proc. ICCV Workshops.","author":"Tewari Ayush","year":"2017","unstructured":"Ayush Tewari, Michael Zollhofer, Hyeongwoo Kim, Pablo Garrido, Florian Bernard, Patrick Perez, and Christian Theobalt. 2017. Mofa: Model-based deep convolutional face autoencoder for unsupervised monocular reconstruction. In Proc. ICCV Workshops."},{"key":"e_1_3_2_1_43_1","doi-asserted-by":"crossref","unstructured":"Omer Tov Yuval Alaluf Yotam Nitzan Or Patashnik and Daniel Cohen-Or. 2021. Designing an encoder for stylegan image manipulation. ACM Trans. Graph. (2021).","DOI":"10.1145\/3476576.3476706"},{"key":"e_1_3_2_1_44_1","volume-title":"3D cartoon face generation with controllable expressions from a single GAN image. arXiv preprint arXiv:2207.14425","author":"Wang Hao","year":"2022","unstructured":"Hao Wang, Guosheng Lin, Steven\u00a0CH Hoi, and Chunyan Miao. 2022. 3D cartoon face generation with controllable expressions from a single GAN image. arXiv preprint arXiv:2207.14425 (2022)."},{"key":"e_1_3_2_1_45_1","volume-title":"Proc. ICLR.","author":"Wu Zongze","year":"2022","unstructured":"Zongze Wu, Yotam Nitzan, Eli Shechtman, and Dani Lischinski. 2022. StyleAlign: Analysis and Applications of Aligned StyleGAN Models. In Proc. ICLR."},{"key":"e_1_3_2_1_46_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00441"},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.00754"},{"key":"e_1_3_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01001"},{"key":"e_1_3_2_1_49_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58520-4_35"}],"event":{"name":"SIGGRAPH '24: Special Interest Group on Computer Graphics and Interactive Techniques Conference","sponsor":["SIGGRAPH ACM Special Interest Group on Computer Graphics and Interactive Techniques"],"location":"Denver CO USA","acronym":"SIGGRAPH '24"},"container-title":["Special Interest Group on Computer Graphics and Interactive Techniques Conference Conference Papers"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3641519.3657480","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3641519.3657480","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T00:05:50Z","timestamp":1750291550000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3641519.3657480"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,7,13]]},"references-count":49,"alternative-id":["10.1145\/3641519.3657480","10.1145\/3641519"],"URL":"https:\/\/doi.org\/10.1145\/3641519.3657480","relation":{},"subject":[],"published":{"date-parts":[[2024,7,13]]},"assertion":[{"value":"2024-07-13","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}