{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,21]],"date-time":"2026-03-21T02:12:43Z","timestamp":1774059163490,"version":"3.50.1"},"publisher-location":"Cham","reference-count":79,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031728471","type":"print"},{"value":"9783031728488","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,11,29]],"date-time":"2024-11-29T00:00:00Z","timestamp":1732838400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,11,29]],"date-time":"2024-11-29T00:00:00Z","timestamp":1732838400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-72848-8_11","type":"book-chapter","created":{"date-parts":[[2024,11,28]],"date-time":"2024-11-28T13:38:10Z","timestamp":1732801090000},"page":"178-196","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["MagicMirror: Fast and\u00a0High-Quality Avatar Generation with\u00a0a\u00a0Constrained Search Space"],"prefix":"10.1007","author":[{"given":"Armand","family":"Comas-Massagu\u00e9","sequence":"first","affiliation":[]},{"given":"Di","family":"Qiu","sequence":"additional","affiliation":[]},{"given":"Menglei","family":"Chai","sequence":"additional","affiliation":[]},{"given":"Marcel","family":"B\u00fchler","sequence":"additional","affiliation":[]},{"given":"Amit","family":"Raj","sequence":"additional","affiliation":[]},{"given":"Ruiqi","family":"Gao","sequence":"additional","affiliation":[]},{"given":"Qiangeng","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Mark","family":"Matthews","sequence":"additional","affiliation":[]},{"given":"Paulo","family":"Gotardo","sequence":"additional","affiliation":[]},{"given":"Sergio","family":"Orts-Escolano","sequence":"additional","affiliation":[]},{"given":"Thabo","family":"Beeler","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,29]]},"reference":[{"key":"11_CR1","doi-asserted-by":"crossref","unstructured":"Athar, S., Xu, Z., Sunkavalli, K., Shechtman, E., Shu, Z.: RigNeRF: fully controllable neural 3D portraits. In: CVPR 2022, pp. 20332\u201320341 (2022)","DOI":"10.1109\/CVPR52688.2022.01972"},{"key":"11_CR2","doi-asserted-by":"crossref","unstructured":"Bai, Z., Cui, Z., Liu, X., Tan, P.: Riggable 3D face reconstruction via in-network optimization. In: CVPR 2021, pp. 6216\u20136225 (2021)","DOI":"10.1109\/CVPR46437.2021.00615"},{"key":"11_CR3","doi-asserted-by":"crossref","unstructured":"Bai, Z., et al.: Learning personalized high quality volumetric head avatars from monocular RGB videos. In: CVPR 2023, pp. 16890\u201316900 (2023)","DOI":"10.1109\/CVPR52729.2023.01620"},{"key":"11_CR4","doi-asserted-by":"crossref","unstructured":"Barron, J.T., Mildenhall, B., Verbin, D., Srinivasan, P.P., Hedman, P.: Mip-Nerf 360: unbounded anti-aliased neural radiance fields. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5470\u20135479 (2022)","DOI":"10.1109\/CVPR52688.2022.00539"},{"key":"11_CR5","doi-asserted-by":"crossref","unstructured":"Blanz, V., Vetter, T.: A morphable model for the synthesis of 3d faces. In: SIGGRAPH 1999, pp. 187\u2013194 (1999)","DOI":"10.1145\/311535.311556"},{"key":"11_CR6","unstructured":"Bojanowski, P., Joulin, A., Lopez-Paz, D., Szlam, A.: Optimizing the latent space of generative networks. In: Proceedings of the 35th International Conference on Machine Learning, pp. 2640\u20133498 (2018)"},{"key":"11_CR7","doi-asserted-by":"crossref","unstructured":"B\u00fchler, M.C., et al.: Preface: a data-driven volumetric prior for few-shot ultra high-resolution face synthesis. In: ICCV 2023, pp. 3379\u20133390 (2023)","DOI":"10.1109\/ICCV51070.2023.00315"},{"key":"11_CR8","doi-asserted-by":"crossref","unstructured":"Cao, C., et al.: Authentic volumetric avatars from a phone scan. ACM Trans. Graph. 41(4), 163:1\u2013163:19 (2022)","DOI":"10.1145\/3528223.3530143"},{"key":"11_CR9","doi-asserted-by":"crossref","unstructured":"Cao, C., Wu, H., Weng, Y., Shao, T., Zhou, K.: Real-time facial animation with image-based dynamic avatars. ACM Trans. Graph. 35(4), 126:1\u2013126:12 (2016)","DOI":"10.1145\/2897824.2925873"},{"key":"11_CR10","doi-asserted-by":"crossref","unstructured":"Cao, Y., Cao, Y., Han, K., Shan, Y., Wong, K.K.: Dreamavatar: text-and-shape guided 3D human avatar generation via diffusion models. CoRR abs\/2304.00916 (2023)","DOI":"10.1109\/CVPR52733.2024.00097"},{"key":"11_CR11","doi-asserted-by":"crossref","unstructured":"Chan, E.R., et al.: Efficient geometry-aware 3d generative adversarial networks. In: CVPR 2022, pp. 16102\u201316112 (2022)","DOI":"10.1109\/CVPR52688.2022.01565"},{"key":"11_CR12","doi-asserted-by":"crossref","unstructured":"Chan, E.R., et al.: Generative novel view synthesis with 3D-aware diffusion models. In: ICCV 2023, pp. 4194\u20134206 (2023)","DOI":"10.1109\/ICCV51070.2023.00389"},{"key":"11_CR13","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"142","DOI":"10.1007\/978-3-030-58558-7_9","volume-title":"Computer Vision \u2013 ECCV 2020","author":"B Chaudhuri","year":"2020","unstructured":"Chaudhuri, B., Vesdapunt, N., Shapiro, L., Wang, B.: Personalized face modeling for improved face reconstruction and motion retargeting. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12350, pp. 142\u2013160. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58558-7_9"},{"key":"11_CR14","doi-asserted-by":"crossref","unstructured":"Chen, R., Chen, Y., Jiao, N., Jia, K.: Fantasia3D: disentangling geometry and appearance for high-quality text-to-3D content creation. In: ICCV 2023, pp. 22189\u201322199 (2023)","DOI":"10.1109\/ICCV51070.2023.02033"},{"key":"11_CR15","doi-asserted-by":"crossref","unstructured":"Chen, S., Liu, F., Lai, Y., Rosin, P.L., Li, C., Fu, H., Gao, L.: Deepfaceediting: deep face generation and editing with disentangled geometry and appearance control. ACM Trans. Graph. 40(4), 90:1\u201390:15 (2021)","DOI":"10.1145\/3476576.3476648"},{"key":"11_CR16","doi-asserted-by":"crossref","unstructured":"Dan\u011b\u010dek, R., Black, M.J., Bolkart, T.: EMOCA: emotion driven monocular face capture and animation. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01967"},{"key":"11_CR17","doi-asserted-by":"crossref","unstructured":"Deitke, M., et al.: Objaverse: a universe of annotated 3D objects. In: CVPR (2023)","DOI":"10.1109\/CVPR52729.2023.01263"},{"key":"11_CR18","unstructured":"Du, Y., Li, S., Mordatch, I.: Compositional visual generation with energy based models. In: Neural Information Processing Systems (2020). https:\/\/api.semanticscholar.org\/CorpusID:214223619"},{"key":"11_CR19","doi-asserted-by":"crossref","unstructured":"Gafni, G., Thies, J., Zollh\u00f6fer, M., Nie\u00dfner, M.: Dynamic neural radiance fields for monocular 4D facial avatar reconstruction. In: CVPR 2021, pp. 8649\u20138658 (2021)","DOI":"10.1109\/CVPR46437.2021.00854"},{"key":"11_CR20","doi-asserted-by":"crossref","unstructured":"Gal, R., Patashnik, O., Maron, H., Bermano, A.H., Chechik, G., Cohen-Or, D.: StyleGAN-nada: clip-guided domain adaptation of image generators. ACM Trans. Graph. 41(4), 141:1\u2013141:13 (2022)","DOI":"10.1145\/3528223.3530164"},{"key":"11_CR21","doi-asserted-by":"crossref","unstructured":"Garrido, P., et al.: Reconstruction of personalized 3D face rigs from monocular video. ACM Trans. Graph. 35(3), 28:1\u201328:15 (2016)","DOI":"10.1145\/2890493"},{"key":"11_CR22","unstructured":"Han, X., et al.: HeadSculpt: crafting 3D head avatars with text. In: NeurIPS 2023 (2023)"},{"key":"11_CR23","doi-asserted-by":"crossref","unstructured":"Haque, A., Tancik, M., Efros, A.A., Holynski, A., Kanazawa, A.: Instruct-NeRF2NeRF: editing 3D scenes with instructions. In: CVPR (2023)","DOI":"10.1109\/ICCV51070.2023.01808"},{"key":"11_CR24","unstructured":"Ho, J., Jain, A., Abbeel, P.: Denoising Diffusion Probabilistic Models (2020)"},{"key":"11_CR25","unstructured":"Ho, J., Salimans, T.: Classifier-free diffusion guidance. CoRR abs\/2207.12598 (2022)"},{"key":"11_CR26","doi-asserted-by":"crossref","unstructured":"Hong, F., Zhang, M., Pan, L., Cai, Z., Yang, L., Liu, Z.: Avatarclip: zero-shot text-driven generation and animation of 3D avatars. ACM Trans. Graph. 41(4), 161:1\u2013161:19 (2022)","DOI":"10.1145\/3528223.3530094"},{"key":"11_CR27","unstructured":"Hu, E.J., et al.: Lora: low-rank adaptation of large language models. In: ICLR 2022 (2022)"},{"key":"11_CR28","doi-asserted-by":"crossref","unstructured":"Hu, L., et al.: Avatar digitization from a single image for real-time rendering. ACM Trans. Graph. 36(6), 195:1\u2013195:14 (2017)","DOI":"10.1145\/3130800.31310887"},{"key":"11_CR29","doi-asserted-by":"crossref","unstructured":"Huang, X., Shao, R., Zhang, Q., Zhang, H., Feng, Y., Liu, Y., Wang, Q.: HumanNorm: learning normal diffusion model for high-quality and realistic 3d human generation. CoRR abs\/2310.01406 (2023)","DOI":"10.1109\/CVPR52733.2024.00437"},{"key":"11_CR30","unstructured":"Huang, Y., Wang, J., Shi, Y., Qi, X., Zha, Z., Zhang, L.: Dreamtime: an improved optimization strategy for text-to-3D content creation. CoRR abs\/2306.12422 (2023)"},{"key":"11_CR31","unstructured":"Huang, Y., et al.: Dreamwaltz: make a scene with complex 3d animatable avatars. In: NeurIPS 2023 (2023)"},{"key":"11_CR32","doi-asserted-by":"crossref","unstructured":"Ichim, A.E., Bouaziz, S., Pauly, M.: Dynamic 3D avatar creation from hand-held video input. ACM Trans. Graph. 34(4), 45:1\u201345:14 (2015)","DOI":"10.1145\/2766974"},{"key":"11_CR33","doi-asserted-by":"crossref","unstructured":"Jiang, R., et al.: Avatarcraft: transforming text into neural human avatars with parameterized shape and pose control. In: ICCV 2023, pp. 14325\u201314336 (2023)","DOI":"10.1109\/ICCV51070.2023.01322"},{"key":"11_CR34","unstructured":"Jun, H., Nichol, A.: Shap-e: generating conditional 3D implicit functions. CoRR abs\/2305.02463 (2023)"},{"key":"11_CR35","doi-asserted-by":"crossref","unstructured":"Karnewar, A., Vedaldi, A., Novotn\u00fd, D., Mitra, N.J.: HOLODIFFUSION: training a 3D diffusion model using 2D images. In: CVPR 2023, pp. 18423\u201318433 (2023)","DOI":"10.1109\/CVPR52729.2023.01767"},{"issue":"12","key":"11_CR36","doi-asserted-by":"publisher","first-page":"4217","DOI":"10.1109\/TPAMI.2020.2970919","volume":"43","author":"T Karras","year":"2021","unstructured":"Karras, T., Laine, S., Aila, T.: A style-based generator architecture for generative adversarial networks. IEEE Trans. Pattern Anal. Mach. Intell. 43(12), 4217\u20134228 (2021)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"11_CR37","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aittala, M., Hellsten, J., Lehtinen, J., Aila, T.: Analyzing and improving the image quality of stylegan. In: CVPR 2020, pp. 8107\u20138116 (2020)","DOI":"10.1109\/CVPR42600.2020.00813"},{"key":"11_CR38","doi-asserted-by":"crossref","unstructured":"Kerbl, B., Kopanas, G., Leimk\u00fchler, T., Drettakis, G.: 3D Gaussian splatting for real-time radiance field rendering. ACM Trans. Graph. 42(4), 139:1\u2013139:14 (2023)","DOI":"10.1145\/3592433"},{"issue":"4","key":"11_CR39","doi-asserted-by":"publisher","first-page":"163","DOI":"10.1145\/3197517.3201283","volume":"37","author":"H Kim","year":"2018","unstructured":"Kim, H., et al.: Deep video portraits. ACM Trans. Graph. 37(4), 163 (2018)","journal-title":"ACM Trans. Graph."},{"key":"11_CR40","unstructured":"Kirstain, Y., Polyak, A., Singer, U., Matiana, S., Penna, J., Levy, O.: Pick-a-pic: an open dataset of user preferences for text-to-image generation. ArXiv abs\/2305.01569 (2023). https:\/\/api.semanticscholar.org\/CorpusID:258437096"},{"key":"11_CR41","unstructured":"Kolotouros, N., Alldieck, T., Zanfir, A., Bazavan, E.G., Fieraru, M., Sminchisescu, C.: Dreamhuman: animatable 3D avatars from text. In: NeurIPS 2023 (2023)"},{"key":"11_CR42","doi-asserted-by":"crossref","unstructured":"Lin, C., et al.: Magic3D: high-resolution text-to-3D content creation. In: CVPR 2023, pp. 300\u2013309 (2023)","DOI":"10.1109\/CVPR52729.2023.00037"},{"key":"11_CR43","unstructured":"Liu, N., Li, S., Du, Y., Tenenbaum, J.B., Torralba, A.: Learning to compose visual relations. ArXiv abs\/2111.09297 (2021). https:\/\/api.semanticscholar.org\/CorpusID:244270027"},{"key":"11_CR44","unstructured":"Liu, N., Li, S., Du, Y., Torralba, A., Tenenbaum, J.B.: Compositional visual generation with composable diffusion models. ArXiv abs\/2206.01714 (2022). https:\/\/api.semanticscholar.org\/CorpusID:249375227"},{"key":"11_CR45","doi-asserted-by":"crossref","unstructured":"Liu, R., Wu, R., Hoorick, B.V., Tokmakov, P., Zakharov, S., Vondrick, C.: Zero-1-to-3: Zero-shot one image to 3d object. In: ICCV 2023, pp. 9264\u20139275 (2023)","DOI":"10.1109\/ICCV51070.2023.00853"},{"key":"11_CR46","unstructured":"Liu, Y., et al.: Syncdreamer: generating multiview-consistent images from a single-view image. CoRR abs\/2309.03453 (2023)"},{"key":"11_CR47","doi-asserted-by":"crossref","unstructured":"Loper, M., Mahmood, N., Romero, J., Pons-Moll, G., Black, M.J.: SMPL: a skinned multi-person linear model. ACM Trans. Graph. 34(6), 248:1\u2013248:16 (2015)","DOI":"10.1145\/2816795.2818013"},{"key":"11_CR48","doi-asserted-by":"crossref","unstructured":"Luo, S., Hu, W.: Diffusion probabilistic models for 3D point cloud generation. In: CVPR 2021, pp. 2837\u20132845 (2021)","DOI":"10.1109\/CVPR46437.2021.00286"},{"key":"11_CR49","doi-asserted-by":"crossref","unstructured":"Mendiratta, M., et al.: Avatarstudio: text-driven editing of 3D dynamic human head avatars. ACM Trans. Graph. 42(6), 226:1\u2013226:18 (2023)","DOI":"10.1145\/3618368"},{"key":"11_CR50","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"405","DOI":"10.1007\/978-3-030-58452-8_24","volume-title":"Computer Vision \u2013 ECCV 2020","author":"B Mildenhall","year":"2020","unstructured":"Mildenhall, B., Srinivasan, P.P., Tancik, M., Barron, J.T., Ramamoorthi, R., Ng, R.: NeRF: Representing scenes as neural radiance fields for view synthesis. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12346, pp. 405\u2013421. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58452-8_24"},{"key":"11_CR51","unstructured":"Nichol, A., Jun, H., Dhariwal, P., Mishkin, P., Chen, M.: Point-e: a system for generating 3D point clouds from complex prompts. CoRR abs\/2212.08751 (2022)"},{"key":"11_CR52","unstructured":"Poole, B., Jain, A., Barron, J.T., Mildenhall, B.: Dreamfusion: text-to-3D using 2D diffusion. In: ICLR 2023 (2023)"},{"key":"11_CR53","unstructured":"Radford, A., et al.: Learning transferable visual models from natural language supervision. In: ICML 2021, vol.\u00a0139, pp. 8748\u20138763 (2021)"},{"key":"11_CR54","doi-asserted-by":"crossref","unstructured":"Raj, A., et\u00a0al.: DreamBooth3D: subject-driven text-to-3D generation (2023)","DOI":"10.1109\/ICCV51070.2023.00223"},{"key":"11_CR55","doi-asserted-by":"crossref","unstructured":"Raj, A., et al.: Pixel-aligned volumetric avatars. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11733\u201311742 (2021)","DOI":"10.1109\/CVPR46437.2021.01156"},{"key":"11_CR56","doi-asserted-by":"crossref","unstructured":"Rombach, R., Blattmann, A., Lorenz, D., Esser, P., Ommer, B.: High-resolution image synthesis with latent diffusion models. In: CVPR 2022, pp. 10674\u201310685 (2022)","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"11_CR57","doi-asserted-by":"crossref","unstructured":"Ruiz, N., Li, Y., Jampani, V., Pritch, Y., Rubinstein, M., Aberman, K.: DreamBooth: fine tuning text-to-image diffusion models for subject-driven generation. In: CVPR 2023, pp. 22500\u201322510 (2023)","DOI":"10.1109\/CVPR52729.2023.02155"},{"key":"11_CR58","unstructured":"Shen, T., Gao, J., Yin, K., Liu, M., Fidler, S.: Deep marching tetrahedra: a hybrid representation for high-resolution 3D shape synthesis. In: NeurIPS 2021, pp. 6087\u20136101 (2021)"},{"issue":"4","key":"11_CR59","doi-asserted-by":"publisher","first-page":"2004","DOI":"10.1109\/TPAMI.2020.3034267","volume":"44","author":"Y Shen","year":"2022","unstructured":"Shen, Y., Yang, C., Tang, X., Zhou, B.: InterFaceGAN: interpreting the disentangled face representation learned by GANs. IEEE Trans. Pattern Anal. Mach. Intell. 44(4), 2004\u20132018 (2022)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"11_CR60","unstructured":"Shi, Y., Wang, P., Ye, J., Long, M., Li, K., Yang, X.: MVDream: multi-view diffusion for 3D generation. CoRR abs\/2308.16512 (2023)"},{"key":"11_CR61","doi-asserted-by":"crossref","unstructured":"Shue, J.R., Chan, E.R., Po, R., Ankner, Z., Wu, J., Wetzstein, G.: 3D neural field generation using triplane diffusion. In: CVPR 2023, pp. 20875\u201320886 (2023)","DOI":"10.1109\/CVPR52729.2023.02000"},{"key":"11_CR62","doi-asserted-by":"crossref","unstructured":"Tang, J., et al.: Make-it-3D: high-fidelity 3D creation from a single image with diffusion prior. In: ICCV 2023, pp. 22762\u201322772 (2023)","DOI":"10.1109\/ICCV51070.2023.02086"},{"key":"11_CR63","doi-asserted-by":"crossref","unstructured":"Tewari, A., et al.: FML: face model learning from videos. In: CVPR 2019, pp. 10812\u201310822 (2019)","DOI":"10.1109\/CVPR.2019.01107"},{"key":"11_CR64","doi-asserted-by":"crossref","unstructured":"Thies, J., Zollh\u00f6fer, M., Stamminger, M., Theobalt, C., Nie\u00dfner, M.: Face2face: Real-time face capture and reenactment of RGB videos. In: CVPR 2016, pp. 2387\u20132395 (2016)","DOI":"10.1109\/CVPR.2016.262"},{"key":"11_CR65","doi-asserted-by":"crossref","unstructured":"Trevithick, A., et al.: Real-time radiance fields for single-image portrait view synthesis. ACM Trans. Graph. 42(4), 135:1\u2013135:15 (2023)","DOI":"10.1145\/3592460"},{"key":"11_CR66","doi-asserted-by":"crossref","unstructured":"Tsalicoglou, C., Manhardt, F., Tonioni, A., Niemeyer, M., Tombari, F.: TextMesh: generation of realistic 3d meshes from text prompts. CoRR abs\/2304.12439 (2023)","DOI":"10.1109\/3DV62453.2024.00154"},{"key":"11_CR67","doi-asserted-by":"crossref","unstructured":"Wang, C., Jiang, R., Chai, M., He, M., Chen, D., Liao, J.: NeRF-Art: text-driven neural radiance fields stylization. IEEE TVCG (2023)","DOI":"10.1109\/TVCG.2023.3283400"},{"key":"11_CR68","first-page":"1","volume":"01","author":"C Wang","year":"2023","unstructured":"Wang, C., Jiang, R., Chai, M., He, M., Chen, D., Liao, J.: Nerf-art: text-driven neural radiance fields stylization. IEEE Trans. Vis. Comput. Graph. 01, 1\u201315 (2023)","journal-title":"IEEE Trans. Vis. Comput. Graph."},{"key":"11_CR69","doi-asserted-by":"crossref","unstructured":"Wang, H., Du, X., Li, J., Yeh, R.A., Shakhnarovich, G.: Score Jacobian chaining: lifting pretrained 2D diffusion models for 3D generation. In: CVPR 2023, pp. 12619\u201312629 (2023)","DOI":"10.1109\/CVPR52729.2023.01214"},{"key":"11_CR70","doi-asserted-by":"crossref","unstructured":"Wang, T., et al.: RODIN: a generative model for sculpting 3D digital avatars using diffusion. In: CVPR 2023, pp. 4563\u20134573 (2023)","DOI":"10.1109\/CVPR52729.2023.00443"},{"key":"11_CR71","unstructured":"Wang, Z., et al.: ProlificDreamer: high-fidelity and diverse text-to-3D generation with variational score distillation. In: NeurIPS 2023 (2023)"},{"key":"11_CR72","unstructured":"Watson, D., Chan, W., Martin-Brualla, R., Ho, J., Tagliasacchi, A., Norouzi, M.: Novel view synthesis with diffusion models. In: ICLR 2023 (2023)"},{"issue":"4","key":"11_CR73","doi-asserted-by":"publisher","first-page":"77","DOI":"10.1145\/2010324.1964972","volume":"30","author":"T Weise","year":"2011","unstructured":"Weise, T., Bouaziz, S., Li, H., Pauly, M.: Realtime performance-based facial animation. ACM Trans. Graph. 30(4), 77 (2011)","journal-title":"ACM Trans. Graph."},{"key":"11_CR74","unstructured":"Xu, Y., Yang, Z., Yang, Y.: SEEAvatar: photorealistic text-to-3D avatar generation with constrained geometry and appearance. CoRR abs\/2312.08889 (2023)"},{"key":"11_CR75","unstructured":"Zeng, Y., Lu, Y., Ji, X., Yao, Y., Zhu, H., Cao, X.: Avatarbooth: high-quality and customizable 3d human avatar generation. CoRR abs\/2306.09864 (2023)"},{"key":"11_CR76","doi-asserted-by":"crossref","unstructured":"Zhang, L., et al.: DreamFace: progressive generation of animatable 3D faces under text guidance. ACM Trans. Graph. 42(4), 138:1\u2013138:16 (2023)","DOI":"10.1145\/3592094"},{"key":"11_CR77","doi-asserted-by":"crossref","unstructured":"Zheng, Y., Abrevaya, V.F., B\u00fchler, M.C., Chen, X., Black, M.J., Hilliges, O.: I M avatar: implicit morphable head avatars from videos. In: CVPR 2022, pp. 13535\u201313545 (2022)","DOI":"10.1109\/CVPR52688.2022.01318"},{"key":"11_CR78","doi-asserted-by":"crossref","unstructured":"Zhou, Z., Tulsiani, S.: SparseFusion: distilling view-conditioned diffusion for 3D reconstruction. In: CVPR 2023, pp. 12588\u201312597 (2023)","DOI":"10.1109\/CVPR52729.2023.01211"},{"key":"11_CR79","unstructured":"Zhu, P., Abdal, R., Femiani, J., Wonka, P.: Mind the gap: domain gap control for single shot domain adaptation for generative adversarial networks. In: ICLR 2022 (2022)"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-72848-8_11","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,28]],"date-time":"2024-11-28T14:07:17Z","timestamp":1732802837000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-72848-8_11"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,29]]},"ISBN":["9783031728471","9783031728488"],"references-count":79,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-72848-8_11","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,11,29]]},"assertion":[{"value":"29 November 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}