{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,19]],"date-time":"2026-02-19T15:22:43Z","timestamp":1771514563136,"version":"3.50.1"},"publisher-location":"Cham","reference-count":52,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031726392","type":"print"},{"value":"9783031726408","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,10,29]],"date-time":"2024-10-29T00:00:00Z","timestamp":1730160000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,10,29]],"date-time":"2024-10-29T00:00:00Z","timestamp":1730160000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-72640-8_23","type":"book-chapter","created":{"date-parts":[[2024,10,28]],"date-time":"2024-10-28T09:34:20Z","timestamp":1730108060000},"page":"409-425","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":8,"title":["Human Hair Reconstruction with\u00a0Strand-Aligned 3D Gaussians"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-9880-9531","authenticated-orcid":false,"given":"Egor","family":"Zakharov","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8883-9972","authenticated-orcid":false,"given":"Vanessa","family":"Sklyarova","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6077-4540","authenticated-orcid":false,"given":"Michael","family":"Black","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1822-1501","authenticated-orcid":false,"given":"Giljoo","family":"Nam","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0056-9825","authenticated-orcid":false,"given":"Justus","family":"Thies","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5068-3474","authenticated-orcid":false,"given":"Otmar","family":"Hilliges","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,10,29]]},"reference":[{"issue":"3","key":"23_CR1","first-page":"413","volume":"20","author":"C Cao","year":"2013","unstructured":"Cao, C., Weng, Y., Zhou, S., Tong, Y., Zhou, K.: Facewarehouse: a 3D facial expression database for visual computing. IEEE Trans. Visual Comput. Graphics 20(3), 413\u2013425 (2013)","journal-title":"IEEE Trans. Visual Comput. Graphics"},{"issue":"6","key":"23_CR2","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/2816795.2818112","volume":"34","author":"M Chai","year":"2015","unstructured":"Chai, M., Luo, L., Sunkavalli, K., Carr, N., Hadap, S., Zhou, K.: High-quality hair modeling from a single portrait photo. ACM Trans. Graphics 34(6), 1\u201310 (2015)","journal-title":"ACM Trans. Graphics"},{"key":"23_CR3","doi-asserted-by":"crossref","unstructured":"Chai, M., Shao, T., Wu, H., Weng, Y., Zhou, K.: AutoHair: fully automatic hair modeling from a single image. ACM Trans. Graph. 35, 116:1\u2013116:12 (2016)","DOI":"10.1145\/2897824.2925961"},{"key":"23_CR4","doi-asserted-by":"crossref","unstructured":"Chen, Y., et al.: Monogaussianavatar: monocular gaussian point-based head avatar. arXiv preprint arXiv:2312.04558 (2023)","DOI":"10.1145\/3641519.3657499"},{"key":"23_CR5","doi-asserted-by":"crossref","unstructured":"Chiang, M.J.Y., Bitterli, B., Tappan, C., Burley, B.: A practical and controllable hair and fur model for production path tracing. In: Computer Graphics Forum, vol.\u00a035, pp. 275\u2013283. Wiley Online Library (2016)","DOI":"10.1111\/cgf.12830"},{"key":"23_CR6","unstructured":"Community, B.O.: Blender - a 3D modelling and rendering package. Blender Foundation, Stichting Blender Foundation, Amsterdam (2023). http:\/\/www.blender.org"},{"key":"23_CR7","doi-asserted-by":"crossref","unstructured":"Daviet, G.: Interactive hair simulation on the GPU using ADMM. In: ACM SIGGRAPH 2023 Conference Proceedings, pp. 1\u201311 (2023)","DOI":"10.1145\/3588432.3591551"},{"key":"23_CR8","doi-asserted-by":"crossref","unstructured":"Dhamo, H., et al.: Headgas: real-time animatable head avatars via 3D gaussian splatting. arXiv preprint arXiv:2312.02902 (2023)","DOI":"10.1007\/978-3-031-72627-9_26"},{"key":"23_CR9","unstructured":"Epic Games: Unreal engine. https:\/\/www.unrealengine.com"},{"key":"23_CR10","doi-asserted-by":"crossref","unstructured":"Fascione, L., et al.: Path tracing in production. In: ACM SIGGRAPH 2018 Courses, pp. 1\u201379 (2018)","DOI":"10.1145\/3214834.3214864"},{"issue":"4","key":"23_CR11","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3072959.3073630","volume":"36","author":"Y Fei","year":"2017","unstructured":"Fei, Y., Maia, H.T., Batty, C., Zheng, C., Grinspun, E.: A multi-scale model for simulating liquid-hair interactions. ACM Trans. Graphics (TOG) 36(4), 1\u201317 (2017)","journal-title":"ACM Trans. Graphics (TOG)"},{"key":"23_CR12","unstructured":"Garbin, S.J., et al.: Voltemorph: realtime, controllable and generalisable animation of volumetric representations. arXiv preprint arXiv:2208.00949 (2022)"},{"key":"23_CR13","doi-asserted-by":"crossref","unstructured":"Hsu, J., Wang, T., Pan, Z., Gao, X., Yuksel, C., Wu, K.: Sag-free initialization for strand-based hybrid hair simulation. ACM Transactions on Graphics (Proceedings of SIGGRAPH 2023) 42(4) (2023)","DOI":"10.1145\/3592143"},{"key":"23_CR14","doi-asserted-by":"crossref","unstructured":"Jiang, Y., et al.: Hifi4g: high-fidelity human performance rendering via compact gaussian splatting. arXiv:abs\/2312.03461 (2023)","DOI":"10.1109\/CVPR52733.2024.01866"},{"key":"23_CR15","unstructured":"Karras, T., Aittala, M., Aila, T., Laine, S.: Elucidating the design space of diffusion-based generative models. In: Advances in Neural Information Processing Systems (NeurIPS) (2022)"},{"key":"23_CR16","doi-asserted-by":"crossref","unstructured":"Kerbl, B., Kopanas, G., Leimk\u00fchler, T., Drettakis, G.: 3D gaussian splatting for real-time radiance field rendering. ACM Trans. Graphics 42(4) (2023)","DOI":"10.1145\/3592433"},{"key":"23_CR17","doi-asserted-by":"crossref","unstructured":"Kirschstein, T., et al.: Nersemble: multi-view radiance field reconstruction of human heads. ACM Trans. Graph. 42(4) (2023)","DOI":"10.1145\/3592455"},{"key":"23_CR18","doi-asserted-by":"crossref","unstructured":"Li, T., Bolkart, T., Black, M.J., Li, H., Romero, J.: Learning a model of facial shape and expression from 4D scans. ACM Trans. Graphics, (Proc. SIGGRAPH Asia) 36(6), 194:1\u2013194:17 (2017)","DOI":"10.1145\/3130800.3130813"},{"key":"23_CR19","doi-asserted-by":"crossref","unstructured":"Lin, C.H., Ma, W.C., Torralba, A., Lucey, S.: Barf: bundle-adjusting neural radiance fields. In: 2021 IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 5721\u20135731 (2021)","DOI":"10.1109\/ICCV48922.2021.00569"},{"issue":"4","key":"23_CR20","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3450626.3459863","volume":"40","author":"S Lombardi","year":"2021","unstructured":"Lombardi, S., Simon, T., Schwartz, G., Zollhoefer, M., Sheikh, Y., Saragih, J.: Mixture of volumetric primitives for efficient neural rendering. ACM Trans. Graphics (ToG) 40(4), 1\u201313 (2021)","journal-title":"ACM Trans. Graphics (ToG)"},{"key":"23_CR21","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3450626.3459863","volume":"40","author":"S Lombardi","year":"2021","unstructured":"Lombardi, S., Simon, T., Schwartz, G., Zollhoefer, M., Sheikh, Y., Saragih, J.M.: Mixture of volumetric primitives for efficient neural rendering. ACM Trans. Graphics (TOG) 40, 1\u201313 (2021)","journal-title":"ACM Trans. Graphics (TOG)"},{"key":"23_CR22","unstructured":"Luo, H., et al.: Gaussianhair: hair modeling and rendering with light-aware gaussians, vol. abs\/2402.10483 (2024)"},{"key":"23_CR23","doi-asserted-by":"crossref","unstructured":"Luo, L., Li, H., Paris, S., Weise, T., Pauly, M., Rusinkiewicz, S.: Multi-view hair capture using orientation fields. In: 2012 IEEE Conference on Computer Vision and Pattern Recognition, pp. 1490\u20131497. IEEE (2012)","DOI":"10.1109\/CVPR.2012.6247838"},{"issue":"4","key":"23_CR24","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/2461912.2462026","volume":"32","author":"L Luo","year":"2013","unstructured":"Luo, L., Li, H., Rusinkiewicz, S.: Structure-aware hair capture. ACM Trans. Graphics 32(4), 1\u201312 (2013)","journal-title":"ACM Trans. Graphics"},{"key":"23_CR25","doi-asserted-by":"crossref","unstructured":"Luo, L., Zhang, C., Zhang, Z., Rusinkiewicz, S.: Wide-baseline hair capture using strand-based refinement. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 265\u2013272 (2013)","DOI":"10.1109\/CVPR.2013.41"},{"key":"23_CR26","doi-asserted-by":"crossref","unstructured":"Nam, G., Wu, C., Kim, M.H., Sheikh, Y.: Strand-accurate multi-view hair capture. In: 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 155\u2013164 (2019)","DOI":"10.1109\/CVPR.2019.00024"},{"key":"23_CR27","doi-asserted-by":"crossref","unstructured":"Paris, S., Brice\u00f1o, H.M., Sillion, F.X.: Capture of hair geometry from multiple images. ACM SIGGRAPH 2004 Papers (2004)","DOI":"10.1145\/1186562.1015784"},{"issue":"3","key":"23_CR28","doi-asserted-by":"publisher","first-page":"30","DOI":"10.1145\/1360612.1360629","volume":"27","author":"S Paris","year":"2008","unstructured":"Paris, S., et al.: Hair photobooth: geometric and photometric acquisition of real hairstyles. ACM Trans. Graphics 27(3), 30 (2008)","journal-title":"ACM Trans. Graphics"},{"key":"23_CR29","doi-asserted-by":"crossref","unstructured":"Piuze, E., Kry, P.G., Siddiqi, K.: Generalized helicoids for modeling hair geometry. In: Computer Graphics Forum, vol.\u00a030, pp. 247\u2013256. Wiley Online Library (2011)","DOI":"10.1111\/j.1467-8659.2011.01856.x"},{"key":"23_CR30","unstructured":"Poole, B., Jain, A., Barron, J.T., Mildenhall, B.: Dreamfusion: Text-to-3D using 2D diffusion. ICLR abs\/2209.14988 (2023)"},{"key":"23_CR31","doi-asserted-by":"crossref","unstructured":"Qian, S., Kirschstein, T., Schoneveld, L., Davoli, D., Giebenhain, S., Nie\u00dfner, M.: Gaussianavatars: photorealistic head avatars with rigged 3d gaussians. arXiv preprint arXiv:2312.02069 (2023)","DOI":"10.1109\/CVPR52733.2024.01919"},{"key":"23_CR32","unstructured":"Rivero, A., Athar, S., Shu, Z., Samaras, D.: Rig3DGS: creating controllable portraits from casual monocular videos. arXiv preprint arXiv:2402.03723 (2024)"},{"key":"23_CR33","doi-asserted-by":"crossref","unstructured":"Rosu, R.A., Saito, S., Wang, Z., Wu, C., Behnke, S., Nam, G.: Neural strands: learning hair geometry and appearance from multi-view images. In: European Conference on Computer Vision (2022)","DOI":"10.1007\/978-3-031-19827-4_5"},{"key":"23_CR34","unstructured":"Saito, S., Schwartz, G., Simon, T., Li, J., Nam, G.: Relightable gaussian codec avatars. arXiv:abs\/2312.03704 (2023)"},{"key":"23_CR35","doi-asserted-by":"crossref","unstructured":"Sch\u00f6nberger, J.L., Frahm, J.M.: Structure-from-motion revisited. In: Conference on Computer Vision and Pattern Recognition (CVPR) (2016)","DOI":"10.1109\/CVPR.2016.445"},{"issue":"4","key":"23_CR36","first-page":"1","volume":"42","author":"Y Shen","year":"2023","unstructured":"Shen, Y., et al.: Ct2hair: high-fidelity 3D hair modeling using computed tomography. ACM Trans. Graphics 42(4), 1\u201313 (2023)","journal-title":"ACM Trans. Graphics"},{"issue":"7","key":"23_CR37","doi-asserted-by":"publisher","first-page":"3250","DOI":"10.1109\/TVCG.2020.2968433","volume":"27","author":"Y Shen","year":"2020","unstructured":"Shen, Y., Zhang, C., Fu, H., Zhou, K., Zheng, Y.: Deepsketchhair: deep sketch-based 3D hair modeling. IEEE Trans. Visual Comput. Graphics 27(7), 3250\u20133263 (2020)","journal-title":"IEEE Trans. Visual Comput. Graphics"},{"key":"23_CR38","doi-asserted-by":"crossref","unstructured":"Sklyarova, V., Chelishev, J., Dogaru, A., Medvedev, I., Lempitsky, V., Zakharov, E.: Neural haircut: Prior-guided strand-based hair reconstruction. In: Proceedings of IEEE International Conference on Computer Vision (ICCV) (2023)","DOI":"10.1109\/ICCV51070.2023.01810"},{"key":"23_CR39","doi-asserted-by":"crossref","unstructured":"Sklyarova, V., Zakharov, E., Hilliges, O., Black, M.J., Thies, J.: HAAR: text-conditioned generative model of 3D strand-based human hairstyles. ArXiv (2023)","DOI":"10.1109\/CVPR52733.2024.00450"},{"key":"23_CR40","doi-asserted-by":"crossref","unstructured":"Wang, P., Liu, L., Liu, Y., Theobalt, C., Komura, T., Wang, W.: Neus: Learning neural implicit surfaces by volume rendering for multi-view reconstruction. In: Advances in Neural Information Processing Systems (NeurIPS) (2022)","DOI":"10.1109\/ICCV51070.2023.00305"},{"key":"23_CR41","doi-asserted-by":"crossref","unstructured":"Wang, Z., et al.: Neuwigs: a neural dynamic model for volumetric hair capture and animation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 8641\u20138651 (2023)","DOI":"10.1109\/CVPR52729.2023.00835"},{"key":"23_CR42","doi-asserted-by":"crossref","unstructured":"Wang, Z., et al.: HVH: learning a hybrid neural volumetric representation for dynamic hair performance capture. In: 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6133\u20136144 (2021)","DOI":"10.1109\/CVPR52688.2022.00605"},{"key":"23_CR43","unstructured":"Woo, M., Neider, J., Davis, T., Shreiner, D.: OpenGL programming guide: the official guide to learning OpenGL, version 1.2. Addison-Wesley Longman Publishing Co., Inc. (1999)"},{"key":"23_CR44","unstructured":"Xiang, J., Gao, X., Guo, Y., Zhang, J.: Flashavatar: high-fidelity digital avatar rendering at 300fps (2023)"},{"key":"23_CR45","doi-asserted-by":"crossref","unstructured":"Xing, J., et al.: Hairbrush for immersive data-driven hair modeling. In: Proceedings of the 32Nd Annual ACM Symposium on User Interface Software and Technology, pp. 263\u2013279 (2019)","DOI":"10.1145\/3332165.3347876"},{"key":"23_CR46","doi-asserted-by":"crossref","unstructured":"Xu, Y., et al.: Gaussian head avatar: ultra high-fidelity head avatar via dynamic gaussians (2023)","DOI":"10.1109\/CVPR52733.2024.00189"},{"issue":"5","key":"23_CR47","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/1618452.1618512","volume":"28","author":"C Yuksel","year":"2009","unstructured":"Yuksel, C., Schaefer, S., Keyser, J.: Hair meshes. ACM Trans. Graphics 28(5), 1\u20137 (2009)","journal-title":"ACM Trans. Graphics"},{"issue":"4","key":"23_CR48","doi-asserted-by":"publisher","first-page":"156\u20131","DOI":"10.1145\/3072959.3073627","volume":"36","author":"M Zhang","year":"2017","unstructured":"Zhang, M., Chai, M., Wu, H., Yang, H., Zhou, K.: A data-driven approach to four-view image-based hair modeling. ACM Trans. Graphics 36(4), 156\u20131 (2017)","journal-title":"ACM Trans. Graphics"},{"issue":"6","key":"23_CR49","first-page":"1","volume":"37","author":"M Zhang","year":"2018","unstructured":"Zhang, M., Wu, P., Wu, H., Weng, Y., Zheng, Y., Zhou, K.: Modeling hair from an RGB-D camera. ACM Trans. Graphics 37(6), 1\u201310 (2018)","journal-title":"ACM Trans. Graphics"},{"key":"23_CR50","unstructured":"Zhao, Z., Bao, Z., Li, Q., Qiu, G., Liu, K.: Psavatar: a point-based morphable shape model for real-time head avatar creation with 3D gaussian splatting. arXiv preprint arXiv:2401.12900 (2024)"},{"issue":"6","key":"23_CR51","first-page":"1","volume":"42","author":"Y Zhou","year":"2023","unstructured":"Zhou, Y., Chai, M., Pepe, A., Gross, M., Beeler, T.: Groomgen: a high-quality generative hair model using hierarchical latent representations. ACM Trans. Graphics (TOG) 42(6), 1\u201316 (2023)","journal-title":"ACM Trans. Graphics (TOG)"},{"key":"23_CR52","doi-asserted-by":"crossref","unstructured":"Zielonka, W., Bolkart, T., Thies, J.: Instant volumetric head avatars. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4574\u20134584 (2023)","DOI":"10.1109\/CVPR52729.2023.00444"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-72640-8_23","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,30]],"date-time":"2024-11-30T10:29:21Z","timestamp":1732962561000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-72640-8_23"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,29]]},"ISBN":["9783031726392","9783031726408"],"references-count":52,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-72640-8_23","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,10,29]]},"assertion":[{"value":"29 October 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}