{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,6]],"date-time":"2025-11-06T11:46:33Z","timestamp":1762429593785,"version":"3.40.5"},"publisher-location":"Cham","reference-count":66,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031732010"},{"type":"electronic","value":"9783031732027"}],"license":[{"start":{"date-parts":[[2024,11,21]],"date-time":"2024-11-21T00:00:00Z","timestamp":1732147200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,11,21]],"date-time":"2024-11-21T00:00:00Z","timestamp":1732147200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-73202-7_27","type":"book-chapter","created":{"date-parts":[[2024,11,20]],"date-time":"2024-11-20T14:16:31Z","timestamp":1732112191000},"page":"465-483","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Relightable Neural Actor with\u00a0Intrinsic Decomposition and\u00a0Pose Control"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-5055-500X","authenticated-orcid":false,"given":"Diogo","family":"Carbonera Luvizon","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1630-2006","authenticated-orcid":false,"given":"Vladislav","family":"Golyanik","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9146-4403","authenticated-orcid":false,"given":"Adam","family":"Kortylewski","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3899-7515","authenticated-orcid":false,"given":"Marc","family":"Habermann","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6104-6625","authenticated-orcid":false,"given":"Christian","family":"Theobalt","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,21]]},"reference":[{"key":"27_CR1","unstructured":"Easymocap - make human motion capture easier. Github (2021). https:\/\/github.com\/zju3dv\/EasyMocap"},{"key":"27_CR2","doi-asserted-by":"crossref","unstructured":"Alldieck, T., Magnor, M., Xu, W., Theobalt, C., Pons-Moll, G.: Video based reconstruction of 3D people models. In: Computer Vision and Pattern Recognition (CVPR), pp. 8387\u20138397 (2018)","DOI":"10.1109\/CVPR.2018.00875"},{"key":"27_CR3","doi-asserted-by":"crossref","unstructured":"Alldieck, T., Xu, H., Sminchisescu, C.: imghum: implicit generative models of 3D human shape and articulated pose. In: International Conference on Computer Vision (ICCV), pp. 5461\u20135470 (2021)","DOI":"10.1109\/ICCV48922.2021.00541"},{"issue":"4","key":"27_CR4","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3450626.3459829","volume":"40","author":"S Bi","year":"2021","unstructured":"Bi, S., et al.: Deep relightable appearance models for animatable faces. ACM Trans. Graph. (TOG) 40(4), 1\u201315 (2021)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"27_CR5","doi-asserted-by":"crossref","unstructured":"Boss, M., Braun, R., Jampani, V., Barron, J.T., Liu, C., Lensch, H.: Nerd: neural reflectance decomposition from image collections. In: International Conference on Computer Vision (ICCV), pp. 12684\u201312694 (2021)","DOI":"10.1109\/ICCV48922.2021.01245"},{"key":"27_CR6","series-title":"LNCS","doi-asserted-by":"publisher","first-page":"606","DOI":"10.1007\/978-3-031-19781-9_35","volume-title":"ECCV 2022","author":"Z Chen","year":"2022","unstructured":"Chen, Z., Liu, Z.: Relighting4D: neural relightable human from videos. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13674, pp. 606\u2013623. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19781-9_35"},{"key":"27_CR7","doi-asserted-by":"crossref","unstructured":"De\u00a0Aguiar, E., Theobalt, C., Thrun, S., Seidel, H.P.: Automatic conversion of mesh animations into skeleton-based animations. In: Computer Graphics Forum (EUROGRAPHICS), vol.\u00a027, pp. 389\u2013397 (2008)","DOI":"10.1111\/j.1467-8659.2008.01136.x"},{"key":"27_CR8","doi-asserted-by":"crossref","unstructured":"Debevec, P.E., Malik, J.: Recovering high dynamic range radiance maps from photographs. In: ACM SIGGRAPH 2008 classes, pp. 1\u201310 (2008)","DOI":"10.1145\/1401132.1401174"},{"key":"27_CR9","doi-asserted-by":"crossref","unstructured":"Deng, B., Wang, Y., Wetzstein, G.: Lumigan: unconditional generation of relightable 3D human faces. In: International Conference on 3D Vision (3DV), pp. 302\u2013312. IEEE (2024)","DOI":"10.1109\/3DV62453.2024.00081"},{"key":"27_CR10","doi-asserted-by":"publisher","unstructured":"d\u2019Eon, E., Luebke, D., Enderton, E.: Efficient Rendering of Human Skin. In: Kautz, J., Pattanaik, S. (eds.) Rendering Techniques. The Eurographics Association (2007). https:\/\/doi.org\/10.2312\/EGWR\/EGSR07\/147-157","DOI":"10.2312\/EGWR\/EGSR07\/147-157"},{"issue":"6","key":"27_CR11","first-page":"1","volume":"38","author":"K Guo","year":"2019","unstructured":"Guo, K., et al.: The relightables: volumetric performance capture of humans with realistic relighting. ACM Trans. Graph. (TOG) 38(6), 1\u201319 (2019)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"27_CR12","doi-asserted-by":"publisher","unstructured":"Habermann, M., Liu, L., Xu, W., Pons-Moll, G., Zollhoefer, M., Theobalt, C.: Hdhumans: a hybrid approach for high-fidelity digital humans. In: Proceedings of ACM Computer Graphics and Interaction Technology, vol. 6. no. 3 (2023). https:\/\/doi.org\/10.1145\/3606927","DOI":"10.1145\/3606927"},{"key":"27_CR13","doi-asserted-by":"crossref","unstructured":"Habermann, M., Liu, L., Xu, W., Zollhoefer, M., Pons-Moll, G., Theobalt, C.: Real-time deep dynamic characters. ACM Trans. Graph. 40(4) (2021)","DOI":"10.1145\/3476576.3476653"},{"issue":"2","key":"27_CR14","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3311970","volume":"38","author":"M Habermann","year":"2019","unstructured":"Habermann, M., Xu, W., Zollhoefer, M., Pons-Moll, G., Theobalt, C.: LiveCap: real-time human performance capture from monocular video. ACM Trans. Graph. (TOG) 38(2), 1\u201317 (2019)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"27_CR15","doi-asserted-by":"crossref","unstructured":"Habermann, M., Xu, W., Zollhofer, M., Pons-Moll, G., Theobalt, C.: Deepcap: monocular human performance capture using weak supervision. In: Computer Vision and Pattern Recognition (CVPR), pp. 5052\u20135063 (2020)","DOI":"10.1109\/CVPR42600.2020.00510"},{"key":"27_CR16","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Computer Vision and Pattern Recognition (CVPR), pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"27_CR17","doi-asserted-by":"crossref","unstructured":"Iqbal, U., Caliskan, A., Nagano, K., Khamis, S., Molchanov, P., Kautz, J.: Rana: relightable articulated neural avatars. In: International Conference on Computer Vision (ICCV), pp. 23142\u201323153 (2023)","DOI":"10.1109\/ICCV51070.2023.02115"},{"key":"27_CR18","doi-asserted-by":"crossref","unstructured":"Iwase, S., et al.: Relightablehands: efficient neural relighting of articulated hand models. In: Computer Vision and Pattern Recognition (CVPR), pp. 16663\u201316673 (2023)","DOI":"10.1109\/CVPR52729.2023.01599"},{"key":"27_CR19","series-title":"LNCS","doi-asserted-by":"publisher","first-page":"388","DOI":"10.1007\/978-3-031-19787-1_22","volume-title":"ECCV","author":"C Ji","year":"2022","unstructured":"Ji, C., Yu, T., Guo, K., Liu, J., Liu, Y.: Geometry-aware single-image full-body human relighting. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV. LNCS, vol. 13676, pp. 388\u2013405. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19787-1_22"},{"key":"27_CR20","series-title":"LNCS","doi-asserted-by":"publisher","first-page":"402","DOI":"10.1007\/978-3-031-19824-3_24","volume-title":"ECCV 2022","author":"W Jiang","year":"2022","unstructured":"Jiang, W., Yi, K.M., Samei, G., Tuzel, O., Ranjan, A.: NeuMan: neural human radiance field from a single video. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13692, pp. 402\u2013418. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19824-3_24"},{"key":"27_CR21","unstructured":"Jiang, Y., Habermann, M., Golyanik, V., Theobalt, C.: Hifecap: monocular high-fidelity and expressive capture of human performances. In: British Machine Vision Conference (BMVC) (2022)"},{"key":"27_CR22","doi-asserted-by":"crossref","unstructured":"Jin, H., et al.: Tensoir: tensorial inverse rendering. In: Computer Vision and Pattern Recognition (CVPR), pp. 165\u2013174 (2023)","DOI":"10.1109\/CVPR52729.2023.00024"},{"key":"27_CR23","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. CoRR abs\/1412.6980 (2014). https:\/\/api.semanticscholar.org\/CorpusID:6628106"},{"key":"27_CR24","unstructured":"Kwon, Y., Kim, D., Ceylan, D., Fuchs, H.: Neural human performer: learning generalizable radiance fields for human performance rendering. In: Advances in Neural Information Processing Systems (NeurIPS), vol. 34, pp. 24741\u201324752 (2021)"},{"key":"27_CR25","unstructured":"Kwon, Y., Liu, L., Fuchs, H., Habermann, M., Theobalt, C.: DELIFFAS: deformable light fields for fast avatar synthesis. In: Advances in Neural Information Processing Systems (2023)"},{"key":"27_CR26","doi-asserted-by":"crossref","unstructured":"Ledig, C., et\u00a0al.: Photo-realistic single image super-resolution using a generative adversarial network. In: Computer Vision and Pattern Recognition (CVPR), pp. 4681\u20134690 (2017)","DOI":"10.1109\/CVPR.2017.19"},{"key":"27_CR27","doi-asserted-by":"publisher","unstructured":"Lewis, J.P., Cordner, M., Fong, N.: Pose space deformation: a unified approach to shape interpolation and skeleton-driven deformation. In: SIGGRAPH, pp. 165\u2013172. ACM Press\/Addison-Wesley Publishing Co., USA (2000). https:\/\/doi.org\/10.1145\/344779.344862","DOI":"10.1145\/344779.344862"},{"issue":"4","key":"27_CR28","first-page":"1","volume":"41","author":"G Li","year":"2022","unstructured":"Li, G., Meka, A., Mueller, F., Buehler, M.C., Hilliges, O., Beeler, T.: EyeNeRF: a hybrid representation for photorealistic synthesis, animation and relighting of human eyes. ACM Trans. Graph. (TOG) 41(4), 1\u201316 (2022)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"27_CR29","doi-asserted-by":"crossref","unstructured":"Li, G., et al.: Capturing relightable human performances under general uncontrolled illumination. In: Computer Graphics Forum, vol.\u00a032, pp. 275\u2013284. Wiley Online Library (2013)","DOI":"10.1111\/cgf.12047"},{"key":"27_CR30","doi-asserted-by":"crossref","unstructured":"Lin, S., Ryabtsev, A., Sengupta, S., Curless, B.L., Seitz, S.M., Kemelmacher-Shlizerman, I.: Real-time high-resolution background matting. In: Computer Vision and Pattern Recognition (CVPR), pp. 8762\u20138771 (2021)","DOI":"10.1109\/CVPR46437.2021.00865"},{"key":"27_CR31","doi-asserted-by":"crossref","unstructured":"Liu, G., Reda, F.A., Shih, K.J., Wang, T.C., Tao, A., Catanzaro, B.: Image inpainting for irregular holes using partial convolutions. In: European Conference on Computer Vision (ECCV) (2018)","DOI":"10.1007\/978-3-030-01252-6_6"},{"issue":"6","key":"27_CR32","first-page":"1","volume":"40","author":"L Liu","year":"2021","unstructured":"Liu, L., Habermann, M., Rudnev, V., Sarkar, K., Gu, J., Theobalt, C.: Neural actor: neural free-view synthesis of human actors with pose control. ACM Trans. Graph. (TOG) 40(6), 1\u201316 (2021)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"27_CR33","doi-asserted-by":"crossref","unstructured":"Liu, Y., et al.: NeRO: neural geometry and BRDF reconstruction of reflective objects from multiview images. In: SIGGRAPH (2023)","DOI":"10.1145\/3592134"},{"key":"27_CR34","doi-asserted-by":"crossref","unstructured":"Loper, M., Mahmood, N., Romero, J., Pons-Moll, G., Black, M.J.: SMPL: a skinned multi-person linear model. ACM Trans. Graph. (SIGGRAPH Asia) 34(6), 248:1\u2013248:16 (2015)","DOI":"10.1145\/2816795.2818013"},{"key":"27_CR35","series-title":"LNCS","doi-asserted-by":"publisher","first-page":"153","DOI":"10.1007\/978-3-031-19790-1_10","volume-title":"ECCV 2022","author":"L Lyu","year":"2022","unstructured":"Lyu, L., Tewari, A., Leimkuehler, T., Habermann, M., Theobalt, C.: Neural radiance transfer fields for relightable novel-view synthesis with global illumination. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13677, pp. 153\u2013169. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19790-1_10"},{"key":"27_CR36","doi-asserted-by":"publisher","unstructured":"Meka, A., et al.: Deep relightable textures: volumetric performance capture with neural rendering. ACM Trans. Graph. (TOG) 39(6) (2020). https:\/\/doi.org\/10.1145\/3414685.3417814","DOI":"10.1145\/3414685.3417814"},{"key":"27_CR37","doi-asserted-by":"crossref","unstructured":"Mihajlovic, M., Saito, S., Bansal, A., Zollhoefer, M., Tang, S.: COAP: compositional articulated occupancy of people. In: Computer Vision and Pattern Recognition (CVPR) (2022)","DOI":"10.1109\/CVPR52688.2022.01285"},{"key":"27_CR38","doi-asserted-by":"crossref","unstructured":"Mihajlovic, M., Zhang, Y., Black, M.J., Tang, S.: Leap: learning articulated occupancy of people. In: Computer Vision and Pattern Recognition (CVPR), pp. 10461\u201310471 (2021)","DOI":"10.1109\/CVPR46437.2021.01032"},{"key":"27_CR39","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"405","DOI":"10.1007\/978-3-030-58452-8_24","volume-title":"Computer Vision \u2013 ECCV 2020","author":"B Mildenhall","year":"2020","unstructured":"Mildenhall, B., Srinivasan, P.P., Tancik, M., Barron, J.T., Ramamoorthi, R., Ng, R.: NeRF: representing scenes as neural radiance fields for view synthesis. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12346, pp. 405\u2013421. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58452-8_24"},{"key":"27_CR40","doi-asserted-by":"crossref","unstructured":"Pang, H., Zhu, H., Kortylewski, A., Theobalt, C., Habermann, M.: Ash: animatable gaussian splats for efficient and photoreal human rendering. In: Computer Vision and Pattern Recognition (CVPR), pp. 1165\u20131175 (2024)","DOI":"10.1109\/CVPR52733.2024.00117"},{"key":"27_CR41","doi-asserted-by":"crossref","unstructured":"Pavlakos, G., et al.: Expressive body capture: 3D hands, face, and body from a single image. In: Computer Vision and Pattern Recognition (CVPR) (2019)","DOI":"10.1109\/CVPR.2019.01123"},{"key":"27_CR42","doi-asserted-by":"crossref","unstructured":"Peng, S., et al.: Neural body: Implicit neural representations with structured latent codes for novel view synthesis of dynamic humans. In: Computer Vision and Pattern Recognition (CVPR) (2021)","DOI":"10.1109\/CVPR46437.2021.00894"},{"key":"27_CR43","doi-asserted-by":"publisher","unstructured":"Ramamoorthi, R., Hanrahan, P.: A signal-processing framework for inverse rendering. In: Conference on Computer Graphics and Interactive Techniques, pp. 117\u2013128. SIGGRAPH 2001, Association for Computing Machinery, New York, NY, USA (2001). https:\/\/doi.org\/10.1145\/383259.383271","DOI":"10.1145\/383259.383271"},{"key":"27_CR44","doi-asserted-by":"crossref","unstructured":"Ranjan, A., Yi, K.M., Chang, J.H.R., Tuzel, O.: Facelit: neural 3D relightable faces. In: Computer Vision and Pattern Recognition (CVPR), pp. 8619\u20138628 (2023)","DOI":"10.1109\/CVPR52729.2023.00833"},{"key":"27_CR45","doi-asserted-by":"crossref","unstructured":"Remelli, E., et\u00a0al.: Drivable volumetric avatars using texel-aligned features. In: ACM SIGGRAPH 2022 Conference Proceedings (2022)","DOI":"10.1145\/3528233.3530740"},{"issue":"4","key":"27_CR46","doi-asserted-by":"publisher","first-page":"91","DOI":"10.1145\/3386569.3392493","volume":"39","author":"G Schwartz","year":"2020","unstructured":"Schwartz, G., et al.: The eyes have it: an integrated eye and face model for photorealistic facial animation. ACM Trans. Graph. (TOG) 39(4), 91\u20131 (2020)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"27_CR47","doi-asserted-by":"crossref","unstructured":"Shu, Z., Sahasrabudhe, M., Guler, R.A., Samaras, D., Paragios, N., Kokkinos, I.: Deforming autoencoders: unsupervised disentangling of shape and appearance. In: European Conference on Computer Vision (ECCV), pp. 650\u2013665 (2018)","DOI":"10.1007\/978-3-030-01249-6_40"},{"key":"27_CR48","doi-asserted-by":"crossref","unstructured":"Shysheya, A., et\u00a0al.: Textured neural avatars. In: Computer Vision and Pattern Recognition (CVPR), pp. 2387\u20132397 (2019)","DOI":"10.1109\/CVPR.2019.00249"},{"key":"27_CR49","doi-asserted-by":"publisher","unstructured":"Srinivasan, P.P., Deng, B., Zhang, X., Tancik, M., Mildenhall, B., Barron, J.T.: Nerv: neural reflectance and visibility fields for relighting and view synthesis. In: Computer Vision and Pattern Recognition (CVPR), pp. 7491\u20137500 (2021). https:\/\/doi.org\/10.1109\/CVPR46437.2021.00741","DOI":"10.1109\/CVPR46437.2021.00741"},{"key":"27_CR50","doi-asserted-by":"crossref","unstructured":"Tan, F., et al.: VoLux-GAN: a generative model for 3d face synthesis with HDRI relighting. In: ACM SIGGRAPH 2022 Conference Proceedings, pp.\u00a01\u20139 (2022)","DOI":"10.1145\/3528233.3530751"},{"key":"27_CR51","unstructured":"Tewari, A., et al.: Advances in neural rendering. In: Computer Graphics Forum (EG STAR 2022) (2022)"},{"key":"27_CR52","doi-asserted-by":"publisher","unstructured":"Walter, B., Marschner, S.R., Li, H., Torrance, K.E.: Microfacet models for refraction through rough surfaces. In: Kautz, J., Pattanaik, S. (eds.) Rendering Techniques. The Eurographics Association (2007). https:\/\/doi.org\/10.2312\/EGWR\/EGSR07\/195-206","DOI":"10.2312\/EGWR\/EGSR07\/195-206"},{"key":"27_CR53","unstructured":"Wang, P., Liu, L., Liu, Y., Theobalt, C., Komura, T., Wang, W.: NeuS: learning neural implicit surfaces by volume rendering for multi-view reconstruction. In: Advances in Neural Information Processing Systems (NeurIPS) (2021)"},{"key":"27_CR54","series-title":"LNCS","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/978-3-031-19824-3_1","volume-title":"ECCV 2022","author":"S Wang","year":"2022","unstructured":"Wang, S., Schwarz, K., Geiger, A., Tang, S.: ARAH: animatable volume rendering of articulated human SDFs. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13692, pp. 1\u201319. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19824-3_1"},{"key":"27_CR55","unstructured":"Wang, T.C., et al.: Video-to-video synthesis. In: Conference on Neural Information Processing Systems (NeurIPS) (2018)"},{"key":"27_CR56","doi-asserted-by":"crossref","unstructured":"Wang, Y., Holynski, A., Zhang, X., Zhang, X.: Sunstage: portrait reconstruction and relighting using the sun as a light stage. In: Computer Vision and Pattern Recognition (CVPR), pp. 20792\u201320802 (2023)","DOI":"10.1109\/CVPR52729.2023.01992"},{"key":"27_CR57","doi-asserted-by":"crossref","unstructured":"Weng, C.Y., Curless, B., Srinivasan, P.P., Barron, J.T., Kemelmacher-Shlizerman, I.: HumanNeRF: free-viewpoint rendering of moving people from monocular video. In: Computer Vision and Pattern Recognition (CVPR), pp. 16210\u201316220 (2022)","DOI":"10.1109\/CVPR52688.2022.01573"},{"key":"27_CR58","doi-asserted-by":"publisher","unstructured":"Wuu, C.H., et al.: Multiface: a dataset for neural face rendering. In: arXiv (2022). https:\/\/doi.org\/10.48550\/ARXIV.2207.11243","DOI":"10.48550\/ARXIV.2207.11243"},{"key":"27_CR59","doi-asserted-by":"crossref","unstructured":"Yang, H., et al.: Towards practical capture of high-fidelity relightable avatars. In: SIGGRAPH Asia 2023 Conference Papers, pp. 1\u201311 (2023)","DOI":"10.1145\/3610548.3618138"},{"key":"27_CR60","doi-asserted-by":"crossref","unstructured":"Zhang, K., Luan, F., Li, Z., Snavely, N.: IRON: inverse rendering by optimizing neural SDFs and materials from photometric images. In: Computer Vision and Pattern Recognition (CVPR), pp. 5565\u20135574 (2022)","DOI":"10.1109\/CVPR52688.2022.00548"},{"key":"27_CR61","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., Efros, A.A., Shechtman, E., Wang, O.: The unreasonable effectiveness of deep features as a perceptual metric. In: Computer Vision and Pattern Recognition (CVPR) (2018)","DOI":"10.1109\/CVPR.2018.00068"},{"key":"27_CR62","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3478513.3480500","volume":"40","author":"X Zhang","year":"2021","unstructured":"Zhang, X., Srinivasan, P.P., Deng, B., Debevec, P., Freeman, W.T., Barron, J.T.: NeRFactor: neural factorization of shape and reflectance under an unknown illumination. ACM Trans. Graph. (TOG) 40, 1\u201318 (2021)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"27_CR63","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Sun, J., He, X., Fu, H., Jia, R., Zhou, X.: Modeling indirect illumination for inverse rendering. In: Computer Vision and Pattern Recognition (CVPR), pp. 18643\u201318652 (2022)","DOI":"10.1109\/CVPR52688.2022.01809"},{"key":"27_CR64","doi-asserted-by":"crossref","unstructured":"Zheng, R., Li, P., Wang, H., Yu, T.: Learning visibility field for detailed 3D human reconstruction and relighting. In: Computer Vision and Pattern Recognition (CVPR), pp. 216\u2013226 (2023)","DOI":"10.1109\/CVPR52729.2023.00029"},{"key":"27_CR65","doi-asserted-by":"crossref","unstructured":"Zhi, Y., Qian, S., Yan, X., Gao, S.: Dual-space nerf: learning animatable avatars and scene lighting in separate spaces. In: International Conference on 3D Vision (3DV) (2022)","DOI":"10.1109\/3DV57658.2022.00048"},{"key":"27_CR66","doi-asserted-by":"crossref","unstructured":"Zhou, T., et al.: Relightable neural human assets from multi-view gradient illuminations. In: Computer Vision and Pattern Recognition (CVPR), pp. 4315\u20134327 (2023)","DOI":"10.1109\/CVPR52729.2023.00420"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-73202-7_27","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,20]],"date-time":"2024-11-20T15:12:08Z","timestamp":1732115528000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-73202-7_27"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,21]]},"ISBN":["9783031732010","9783031732027"],"references-count":66,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-73202-7_27","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024,11,21]]},"assertion":[{"value":"21 November 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}