{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,27]],"date-time":"2025-08-27T15:55:37Z","timestamp":1756310137791,"version":"3.40.3"},"publisher-location":"Cham","reference-count":99,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031729515"},{"type":"electronic","value":"9783031729522"}],"license":[{"start":{"date-parts":[[2024,10,1]],"date-time":"2024-10-01T00:00:00Z","timestamp":1727740800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,10,1]],"date-time":"2024-10-01T00:00:00Z","timestamp":1727740800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-72952-2_20","type":"book-chapter","created":{"date-parts":[[2024,9,30]],"date-time":"2024-09-30T05:02:02Z","timestamp":1727672522000},"page":"341-361","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["MetaCap: Meta-learning Priors from\u00a0Multi-view Imagery for\u00a0Sparse-View Human Performance Capture and\u00a0Rendering"],"prefix":"10.1007","author":[{"given":"Guoxing","family":"Sun","sequence":"first","affiliation":[]},{"given":"Rishabh","family":"Dabral","sequence":"additional","affiliation":[]},{"given":"Pascal","family":"Fua","sequence":"additional","affiliation":[]},{"given":"Christian","family":"Theobalt","sequence":"additional","affiliation":[]},{"given":"Marc","family":"Habermann","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,10,1]]},"reference":[{"key":"20_CR1","unstructured":"https:\/\/web.twindom.com\/"},{"key":"20_CR2","unstructured":"Antoniou, A., Edwards, H., Storkey, A.: How to train your MAML. arXiv preprint arXiv:1810.09502 (2018)"},{"key":"20_CR3","doi-asserted-by":"crossref","unstructured":"Barron, J.T., Mildenhall, B., Tancik, M., Hedman, P., Martin-Brualla, R., Srinivasan, P.P.: Mip-NeRF: a multiscale representation for anti-aliasing neural radiance fields. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.00580"},{"key":"20_CR4","doi-asserted-by":"crossref","unstructured":"Barron, J.T., Mildenhall, B., Verbin, D., Srinivasan, P.P., Hedman, P.: Mip-NeRF 360: unbounded anti-aliased neural radiance fields. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.00539"},{"key":"20_CR5","doi-asserted-by":"crossref","unstructured":"Barron, J.T., Mildenhall, B., Verbin, D., Srinivasan, P.P., Hedman, P.: Zip-NeRF: anti-aliased grid-based neural radiance fields. In: ICCV (2023)","DOI":"10.1109\/ICCV51070.2023.01804"},{"key":"20_CR6","doi-asserted-by":"crossref","unstructured":"B\u00fchler, M.C., et al.: Preface: a data-driven volumetric prior for few-shot ultra high-resolution face synthesis. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 3402\u20133413 (2023)","DOI":"10.1109\/ICCV51070.2023.00315"},{"key":"20_CR7","doi-asserted-by":"publisher","first-page":"333","DOI":"10.1007\/978-3-031-19824-3_20","volume-title":"European Conference on Computer Vision (ECCV)","author":"A Chen","year":"2022","unstructured":"Chen, A., Xu, Z., Geiger, A., Yu, J., Su, H.: TensoRF: tensorial radiance fields. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13692, pp. 333\u2013350. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19824-3_20"},{"key":"20_CR8","doi-asserted-by":"crossref","unstructured":"Chen, A., et al.: MVSNeRF: fast generalizable radiance field reconstruction from multi-view stereo. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 14124\u201314133 (2021)","DOI":"10.1109\/ICCV48922.2021.01386"},{"issue":"4","key":"20_CR9","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/2766945","volume":"34","author":"A Collet","year":"2015","unstructured":"Collet, A., et al.: High-quality streamable free-viewpoint video. ACM Trans. Graph. (ToG) 34(4), 1\u201313 (2015)","journal-title":"ACM Trans. Graph. (ToG)"},{"key":"20_CR10","doi-asserted-by":"crossref","unstructured":"Davydov, A., Remizova, A., Constantin, V., Honari, S., Salzmann, M., Fua, P.: Adversarial parametric pose prior. In: IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2022)","DOI":"10.1109\/CVPR52688.2022.01072"},{"key":"20_CR11","doi-asserted-by":"crossref","unstructured":"De\u00a0Luigi, L., Li, R., Guillard, B., Salzmann, M., Fua, P.: DrapeNet: garment generation and self-supervised draping. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (2023)","DOI":"10.1109\/CVPR52729.2023.00146"},{"key":"20_CR12","unstructured":"Finn, C., Abbeel, P., Levine, S.: Model-agnostic meta-learning for fast adaptation of deep networks. In: International Conference on Machine Learning, pp. 1126\u20131135. PMLR (2017)"},{"key":"20_CR13","doi-asserted-by":"crossref","unstructured":"Fridovich-Keil, S., Yu, A., Tancik, M., Chen, Q., Recht, B., Kanazawa, A.: Plenoxels: radiance fields without neural networks. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.00542"},{"key":"20_CR14","unstructured":"Gropp, A., Yariv, L., Haim, N., Atzmon, M., Lipman, Y.: Implicit geometric regularization for learning shapes. In: Proceedings of Machine Learning and Systems 2020, pp. 3569\u20133579 (2020)"},{"key":"20_CR15","unstructured":"Gu, J., et al.: NerfDiff: single-image view synthesis with nerf-guided distillation from 3D-aware diffusion. In: International Conference on Machine Learning (2023)"},{"key":"20_CR16","doi-asserted-by":"crossref","unstructured":"Wang, G., Chen, Z., Loy, C.C., Liu, Z.: SparseNeRF: distilling depth ranking for few-shot novel view synthesis. Technical report (2023)","DOI":"10.1109\/ICCV51070.2023.00832"},{"issue":"6","key":"20_CR17","first-page":"1","volume":"38","author":"K Guo","year":"2019","unstructured":"Guo, K., et al.: The relightables: volumetric performance capture of humans with realistic relighting. ACM Trans. Graph. (ToG) 38(6), 1\u201319 (2019)","journal-title":"ACM Trans. Graph. (ToG)"},{"issue":"3","key":"20_CR18","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3606927","volume":"6","author":"M Habermann","year":"2023","unstructured":"Habermann, M., Liu, L., Xu, W., Pons-Moll, G., Zollhoefer, M., Theobalt, C.: HDHumans: a hybrid approach for high-fidelity digital humans. Proc. ACM Comput. Graph. Interact. Tech. 6(3), 1\u201323 (2023)","journal-title":"Proc. ACM Comput. Graph. Interact. Tech."},{"issue":"4","key":"20_CR19","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3450626.3459749","volume":"40","author":"M Habermann","year":"2021","unstructured":"Habermann, M., Liu, L., Xu, W., Zollhoefer, M., Pons-Moll, G., Theobalt, C.: Real-time deep dynamic characters. ACM Trans. Graph. 40(4), 1\u201316 (2021)","journal-title":"ACM Trans. Graph."},{"issue":"2","key":"20_CR20","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3311970","volume":"38","author":"M Habermann","year":"2019","unstructured":"Habermann, M., Xu, W., Zollhoefer, M., Pons-Moll, G., Theobalt, C.: LiveCap: real-time human performance capture from monocular video. ACM Trans. Graph. (TOG) 38(2), 1\u201317 (2019)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"20_CR21","doi-asserted-by":"crossref","unstructured":"Habermann, M., Xu, W., Zollhoefer, M., Pons-Moll, G., Theobalt, C.: DeepCap: monocular human performance capture using weak supervision. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE (2020)","DOI":"10.1109\/CVPR42600.2020.00510"},{"issue":"1","key":"20_CR22","doi-asserted-by":"publisher","first-page":"974","DOI":"10.1109\/TVCG.2017.2744238","volume":"24","author":"M Hadwiger","year":"2017","unstructured":"Hadwiger, M., Al-Awami, A.K., Beyer, J., Agus, M., Pfister, H.: SparseLeap: efficient empty space skipping for large-scale volume rendering. IEEE Trans. Vis. Comput. Graph. 24(1), 974\u2013983 (2017)","journal-title":"IEEE Trans. Vis. Comput. Graph."},{"issue":"9","key":"20_CR23","first-page":"5149","volume":"44","author":"T Hospedales","year":"2021","unstructured":"Hospedales, T., Antoniou, A., Micaelli, P., Storkey, A.: Meta-learning in neural networks: a survey. IEEE Trans. Pattern Anal. Mach. Intell. 44(9), 5149\u20135169 (2021)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"20_CR24","doi-asserted-by":"crossref","unstructured":"Huang, Y., et al.: TeCH: text-guided reconstruction of lifelike clothed humans. In: International Conference on 3D Vision (3DV) (2024)","DOI":"10.1109\/3DV62453.2024.00152"},{"key":"20_CR25","doi-asserted-by":"publisher","first-page":"492","DOI":"10.1007\/978-1-4612-4380-9_35","volume-title":"Breakthroughs in Statistics: Methodology and Distribution","author":"PJ Huber","year":"1992","unstructured":"Huber, P.J.: Robust estimation of a location parameter. In: Kotz, S., Johnson, N.L. (eds.) Breakthroughs in Statistics: Methodology and Distribution. SSS, pp. 492\u2013518. Springer, New York (1992). https:\/\/doi.org\/10.1007\/978-1-4612-4380-9_35"},{"key":"20_CR26","doi-asserted-by":"publisher","first-page":"402","DOI":"10.1007\/978-3-031-19824-3_24","volume-title":"European Conference on Computer Vision (ECCV)","author":"W Jiang","year":"2022","unstructured":"Jiang, W., Yi, K.M., Samei, G., Tuzel, O., Ranjan, A.: NeuMan: neural human radiance field from a single video. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13692, pp. 402\u2013418. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19824-3_24"},{"key":"20_CR27","unstructured":"Jiang, Y., Habermann, M., Golyanik, V., Theobalt, C.: HiFECap: monocular high-fidelity and expressive capture of human performances. In: BMVC (2022)"},{"key":"20_CR28","doi-asserted-by":"crossref","unstructured":"Johnson, E.C., Habermann, M., Shimada, S., Golyanik, V., Theobalt, C.: Unbiased 4D: monocular 4D reconstruction with a neural deformation model. In: Computer Vision and Pattern Recognition Workshops (CVPRW) (2023)","DOI":"10.1109\/CVPRW59228.2023.00701"},{"key":"20_CR29","doi-asserted-by":"crossref","unstructured":"Kerbl, B., Kopanas, G., Leimk\u00fchler, T., Drettakis, G.: 3D Gaussian splatting for real-time radiance field rendering. ACM Trans. Graph. 42(4) (2023). https:\/\/repo-sam.inria.fr\/fungraph\/3d-gaussian-splatting\/","DOI":"10.1145\/3592433"},{"key":"20_CR30","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization (2017)"},{"key":"20_CR31","unstructured":"Kwon, Y., Kim, D., Ceylan, D., Fuchs, H.: Neural human performer: learning generalizable radiance fields for human performance rendering. In: Advances in Neural Information Processing Systems, vol. 34, pp. 24741\u201324752 (2021)"},{"key":"20_CR32","unstructured":"Kwon, Y., Liu, L., Fuchs, H., Habermann, M., Theobalt, C.: DELIFFAS: deformable light fields for fast avatar synthesis. In: Advances in Neural Information Processing Systems (2023)"},{"key":"20_CR33","unstructured":"Li, K., Malik, J.: Learning to optimize. arXiv preprint arXiv:1606.01885 (2016)"},{"key":"20_CR34","doi-asserted-by":"publisher","first-page":"419","DOI":"10.1007\/978-3-031-19824-3_25","volume-title":"European Conference on Computer Vision","author":"R Li","year":"2022","unstructured":"Li, R., et al.: TAVA: template-free animatable volumetric actors. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13692, pp. 419\u2013436. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19824-3_25"},{"key":"20_CR35","doi-asserted-by":"publisher","unstructured":"Li, Y., Habermann, M., Thomaszewski, B., Coros, S., Beeler, T., Theobalt, C.: Deep physics-aware inference of cloth deformation for monocular human performance capture. In: 2021 International Conference on 3D Vision (3DV), Los Alamitos, CA, USA, pp. 373\u2013384. IEEE Computer Society (2021). https:\/\/doi.org\/10.1109\/3DV53792.2021.00047. https:\/\/doi.ieeecomputersociety.org\/10.1109\/3DV53792.2021.00047","DOI":"10.1109\/3DV53792.2021.00047"},{"key":"20_CR36","doi-asserted-by":"publisher","first-page":"322","DOI":"10.1007\/978-3-031-19769-7_19","volume-title":"European Conference on Computer Vision (ECCV)","author":"Z Li","year":"2022","unstructured":"Li, Z., Zheng, Z., Zhang, H., Ji, C., Liu, Y.: AvatarCap: animatable avatar conditioned monocular human volumetric capture. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13661, pp. 322\u2013341. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19769-7_19"},{"issue":"6","key":"20_CR37","first-page":"1","volume":"40","author":"L Liu","year":"2021","unstructured":"Liu, L., Habermann, M., Rudnev, V., Sarkar, K., Gu, J., Theobalt, C.: Neural actor: neural free-view synthesis of human actors with pose control. ACM Trans. Graph. 40(6), 1\u201316 (2021). (ACM SIGGRAPH Asia)","journal-title":"ACM Trans. Graph."},{"key":"20_CR38","doi-asserted-by":"publisher","first-page":"210","DOI":"10.1007\/978-3-031-19824-3_13","volume-title":"ECCV 2022","author":"X Long","year":"2022","unstructured":"Long, X., Lin, C., Wang, P., Komura, T., Wang, W.: SparseNeuS: fast generalizable neural surface reconstruction from sparse views. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13692, pp. 210\u2013227. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19824-3_13"},{"key":"20_CR39","doi-asserted-by":"crossref","unstructured":"Loper, M., Mahmood, N., Romero, J., Pons-Moll, G., Black, M.J.: SMPL: a skinned multi-person linear model. ACM Trans. Graph. 34(6), 248:1\u2013248:16 (2015). (Proc. SIGGRAPH Asia)","DOI":"10.1145\/2816795.2818013"},{"key":"20_CR40","unstructured":"Luvizon, D., Golyanik, V., Kortylewski, A., Habermann, M., Theobalt, C.: Relightable neural actor with intrinsic decomposition and pose control. In: European Conference on Computer Vision (ECCV) (2024)"},{"key":"20_CR41","doi-asserted-by":"crossref","unstructured":"Ma, Q., et al.: Learning to dress 3D people in generative clothing. In: Computer Vision and Pattern Recognition (CVPR) (2020)","DOI":"10.1109\/CVPR42600.2020.00650"},{"key":"20_CR42","doi-asserted-by":"crossref","unstructured":"Mescheder, L., Oechsle, M., Niemeyer, M., Nowozin, S., Geiger, A.: Occupancy networks: learning 3D reconstruction in function space. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4460\u20134470 (2019)","DOI":"10.1109\/CVPR.2019.00459"},{"key":"20_CR43","doi-asserted-by":"publisher","first-page":"179","DOI":"10.1007\/978-3-031-19784-0_11","volume-title":"European Conference on Computer Vision","author":"M Mihajlovic","year":"2022","unstructured":"Mihajlovic, M., Bansal, A., Zollhoefer, M., Tang, S., Saito, S.: KeypointNeRF: generalizing image-based volumetric avatars using relative spatial encoding of keypoints. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13675, pp. 179\u2013197. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19784-0_11"},{"key":"20_CR44","doi-asserted-by":"publisher","first-page":"405","DOI":"10.1007\/978-3-030-58452-8_24","volume-title":"ECCV","author":"B Mildenhall","year":"2020","unstructured":"Mildenhall, B., Srinivasan, P.P., Tancik, M., Barron, J.T., Ramamoorthi, R., Ng, R.: NeRF: representing scenes as neural radiance fields for view synthesis. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.M. (eds.) ECCV 2020. LNCS, vol. 12346, pp. 405\u2013421. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58452-8_24"},{"issue":"4","key":"20_CR45","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3528223.3530127","volume":"41","author":"T M\u00fcller","year":"2022","unstructured":"M\u00fcller, T., Evans, A., Schied, C., Keller, A.: Instant neural graphics primitives with a multiresolution hash encoding. ACM Trans. Graph. (ToG) 41(4), 1\u201315 (2022)","journal-title":"ACM Trans. Graph. (ToG)"},{"key":"20_CR46","doi-asserted-by":"crossref","unstructured":"Newcombe, R.A., Fox, D., Seitz, S.M.: DynamicFusion: reconstruction and tracking of non-rigid scenes in real-time. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 343\u2013352 (2015)","DOI":"10.1109\/CVPR.2015.7298631"},{"key":"20_CR47","unstructured":"Nichol, A., Achiam, J., Schulman, J.: On first-order meta-learning algorithms. arXiv preprint arXiv:1803.02999 (2018)"},{"key":"20_CR48","doi-asserted-by":"crossref","unstructured":"Niemeyer, M., Barron, J.T., Mildenhall, B., Sajjadi, M.S., Geiger, A., Radwan, N.: RegNeRF: regularizing neural radiance fields for view synthesis from sparse inputs. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5480\u20135490 (2022)","DOI":"10.1109\/CVPR52688.2022.00540"},{"key":"20_CR49","doi-asserted-by":"crossref","unstructured":"Palafox, P., Sarafianos, N., Tung, T., Dai, A.: SPAMs: structured implicit parametric models. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01251"},{"key":"20_CR50","doi-asserted-by":"crossref","unstructured":"Pan, X., Yang, Z., Ma, J., Zhou, C., Yang, Y.: TransHuman: a transformer-based human representation for generalizable neural human rendering. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 3544\u20133555 (2023)","DOI":"10.1109\/ICCV51070.2023.00328"},{"key":"20_CR51","doi-asserted-by":"crossref","unstructured":"Pang, H., Zhu, H., Kortylewski, A., Theobalt, C., Habermann, M.: ASH: animatable Gaussian splats for efficient and photoreal human rendering. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1165\u20131175 (2024)","DOI":"10.1109\/CVPR52733.2024.00117"},{"key":"20_CR52","doi-asserted-by":"crossref","unstructured":"Park, J.J., Florence, P., Straub, J., Newcombe, R., Lovegrove, S.: DeepSDF: learning continuous signed distance functions for shape representation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 165\u2013174 (2019)","DOI":"10.1109\/CVPR.2019.00025"},{"key":"20_CR53","doi-asserted-by":"crossref","unstructured":"Pavlakos, G., et al.: Expressive body capture: 3D hands, face, and body from a single image. In: Proceedings IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 10975\u201310985 (2019)","DOI":"10.1109\/CVPR.2019.01123"},{"key":"20_CR54","doi-asserted-by":"crossref","unstructured":"Peng, S., et al.: Animatable neural radiance fields for modeling dynamic human bodies. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.01405"},{"key":"20_CR55","unstructured":"Peng, S., et al.: Animatable neural implicit surfaces for creating avatars from videos. arXiv preprint arXiv:2203.08133 (2022)"},{"key":"20_CR56","doi-asserted-by":"crossref","unstructured":"Peng, S., et al.: Neural body: implicit neural representations with structured latent codes for novel view synthesis of dynamic humans. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00894"},{"key":"20_CR57","unstructured":"Rajeswaran, A., Finn, C., Kakade, S.M., Levine, S.: Meta-learning with implicit gradients. In: Advances in Neural Information Processing Systems, vol. 32 (2019)"},{"key":"20_CR58","doi-asserted-by":"crossref","unstructured":"Remelli, E., et al.: Drivable volumetric avatars using texel-aligned features. In: ACM SIGGRAPH 2022 Conference Proceedings (2022)","DOI":"10.1145\/3528233.3530740"},{"key":"20_CR59","doi-asserted-by":"crossref","unstructured":"Saito, S., Huang, Z., Natsume, R., Morishima, S., Kanazawa, A., Li, H.: PIFu: pixel-aligned implicit function for high-resolution clothed human digitization. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 2304\u20132314 (2019)","DOI":"10.1109\/ICCV.2019.00239"},{"key":"20_CR60","doi-asserted-by":"crossref","unstructured":"Saito, S., Simon, T., Saragih, J., Joo, H.: PIFuHD: multi-level pixel-aligned implicit function for high-resolution 3D human digitization. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 84\u201393 (2020)","DOI":"10.1109\/CVPR42600.2020.00016"},{"key":"20_CR61","doi-asserted-by":"crossref","unstructured":"Saito, S., Yang, J., Ma, Q., Black, M.J.: SCANimate: weakly supervised learning of skinned clothed avatar networks. In: Proceedings IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2021)","DOI":"10.1109\/CVPR46437.2021.00291"},{"key":"20_CR62","doi-asserted-by":"crossref","unstructured":"Shao, R., et al.: FloRen: real-time high-quality human performance rendering via appearance flow using sparse RGB cameras. In: SIGGRAPH Asia 2022 Conference Papers, pp. 1\u201310 (2022)","DOI":"10.1145\/3550469.3555409"},{"key":"20_CR63","doi-asserted-by":"crossref","unstructured":"Shao, R., et al.: DoubleField: bridging the neural surface and radiance fields for high-fidelity human reconstruction and rendering. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01541"},{"key":"20_CR64","doi-asserted-by":"publisher","first-page":"702","DOI":"10.1007\/978-3-031-19824-3_41","volume-title":"European Conference on Computer Vision","author":"R Shao","year":"2022","unstructured":"Shao, R., Zheng, Z., Zhang, H., Sun, J., Liu, Y.: DiffuStereo: high quality human reconstruction via diffusion-based stereo using sparse cameras. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13692, pp. 702\u2013720. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19824-3_41"},{"key":"20_CR65","doi-asserted-by":"crossref","unstructured":"Shen, K., et al.: X-avatar: expressive human avatars. In: Computer Vision and Pattern Recognition (CVPR) (2023)","DOI":"10.1109\/CVPR52729.2023.01622"},{"key":"20_CR66","doi-asserted-by":"crossref","unstructured":"Shetty, A., Habermann, M., Sun, G., Luvizon, D., Golyanik, V., Theobalt, C.: Holoported characters: real-time free-viewpoint rendering of humans from sparse RGB cameras (2023)","DOI":"10.1109\/CVPR52733.2024.00121"},{"key":"20_CR67","doi-asserted-by":"crossref","unstructured":"Shuai, Q., et al.: Novel view synthesis of human interactions from sparse multi-view videos. In: SIGGRAPH Conference Proceedings (2022)","DOI":"10.1145\/3528233.3530704"},{"key":"20_CR68","unstructured":"Sitzmann, V., Chan, E., Tucker, R., Snavely, N., Wetzstein, G.: MetaSDF: meta-learning signed distance functions. In: Advances in Neural Information Processing Systems, vol. 33, pp. 10136\u201310147 (2020)"},{"key":"20_CR69","doi-asserted-by":"crossref","unstructured":"Stoll, C., Hasler, N., Gall, J., Seidel, H.P., Theobalt, C.: Fast articulated motion tracking using a sums of Gaussians body model. In: 2011 International Conference on Computer Vision, pp. 951\u2013958. IEEE (2011)","DOI":"10.1109\/ICCV.2011.6126338"},{"key":"20_CR70","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"246","DOI":"10.1007\/978-3-030-58548-8_15","volume-title":"Computer Vision \u2013 ECCV 2020","author":"Z Su","year":"2020","unstructured":"Su, Z., Xu, L., Zheng, Z., Yu, T., Liu, Y., Fang, L.: RobustFusion: human volumetric capture with data-driven visual cues using a RGBD camera. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020, Part IV. LNCS, vol. 12349, pp. 246\u2013264. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58548-8_15"},{"key":"20_CR71","doi-asserted-by":"crossref","unstructured":"Sun, G., et al.: Neural free-viewpoint performance rendering under complex human-object interactions. In: Proceedings of the 29th ACM International Conference on Multimedia, pp. 4651\u20134660 (2021)","DOI":"10.1145\/3474085.3475442"},{"key":"20_CR72","doi-asserted-by":"crossref","unstructured":"Tancik, M., et al.: Learned initializations for optimizing coordinate-based neural representations. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00287"},{"key":"20_CR73","doi-asserted-by":"crossref","unstructured":"Tretschk, E., et al.: State of the art in dense monocular non-rigid 3D reconstruction. In: Computer Graphics Forum (Eurographics State of the Art Reports) (2023)","DOI":"10.1111\/cgf.14774"},{"issue":"12","key":"20_CR74","doi-asserted-by":"publisher","first-page":"5097","DOI":"10.1109\/TVCG.2022.3202503","volume":"29","author":"K Wang","year":"2022","unstructured":"Wang, K., Peng, S., Zhou, X., Yang, J., Zhang, G.: NerfCap: human performance capture with dynamic neural radiance fields. IEEE Trans. Vis. Comput. Graph. 29(12), 5097\u20135110 (2022)","journal-title":"IEEE Trans. Vis. Comput. Graph."},{"key":"20_CR75","doi-asserted-by":"crossref","unstructured":"Wang, K., Zhang, G., Cong, S., Yang, J.: Clothed human performance capture with a double-layer neural radiance fields. In: Computer Vision and Pattern Recognition (CVPR) (2023)","DOI":"10.1109\/CVPR52729.2023.02021"},{"key":"20_CR76","unstructured":"Wang, P., Liu, L., Liu, Y., Theobalt, C., Komura, T., Wang, W.: NeuS: learning neural implicit surfaces by volume rendering for multi-view reconstruction. In: NeurIPS (2021)"},{"key":"20_CR77","unstructured":"Wang, S., Mihajlovic, M., Ma, Q., Geiger, A., Tang, S.: MetaAvatar: learning animatable clothed human models from few depth images. In: Advances in Neural Information Processing Systems (2021)"},{"key":"20_CR78","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/978-3-031-19824-3_1","volume-title":"European Conference on Computer Vision","author":"S Wang","year":"2022","unstructured":"Wang, S., Schwarz, K., Geiger, A., Tang, S.: ARAH: animatable volume rendering of articulated human SDFs. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13692, pp. 1\u201319. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19824-3_1"},{"key":"20_CR79","doi-asserted-by":"crossref","unstructured":"Wang, Y., Han, Q., Habermann, M., Daniilidis, K., Theobalt, C., Liu, L.: NeuS2: fast learning of neural implicit surfaces for multi-view reconstruction. arXiv preprint arXiv:2212.05231 (2022)","DOI":"10.1109\/ICCV51070.2023.00305"},{"key":"20_CR80","doi-asserted-by":"crossref","unstructured":"Weng, C.Y., Curless, B., Srinivasan, P.P., Barron, J.T., Kemelmacher-Shlizerman, I.: HumanNeRF: free-viewpoint rendering of moving people from monocular video. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 16210\u201316220 (2022)","DOI":"10.1109\/CVPR52688.2022.01573"},{"key":"20_CR81","doi-asserted-by":"crossref","unstructured":"Xiang, D., et al.: Drivable avatar clothing: faithful full-body telepresence with dynamic clothing driven by sparse RGB-D input. In: SIGGRAPH Asia 2023 Conference Papers, pp. 1\u201311 (2023)","DOI":"10.1145\/3610548.3618136"},{"key":"20_CR82","doi-asserted-by":"crossref","unstructured":"Xiang, D., Prada, F., Wu, C., Hodgins, J.: MonoClothCap: towards temporally coherent clothing capture from monocular RGB video. In: 2020 International Conference on 3D Vision (3DV), pp. 322\u2013332. IEEE (2020)","DOI":"10.1109\/3DV50981.2020.00042"},{"key":"20_CR83","doi-asserted-by":"crossref","unstructured":"Xiu, Y., Yang, J., Cao, X., Tzionas, D., Black, M.J.: ECON: explicit clothed humans optimized via normal integration. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 512\u2013523 (2023)","DOI":"10.1109\/CVPR52729.2023.00057"},{"key":"20_CR84","doi-asserted-by":"crossref","unstructured":"Xiu, Y., Yang, J., Tzionas, D., Black, M.J.: ICON: implicit clothed humans obtained from normals. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 13296\u201313306 (2022)","DOI":"10.1109\/CVPR52688.2022.01294"},{"key":"20_CR85","doi-asserted-by":"publisher","unstructured":"Xu, W., et al.: MonoPerfCap: human performance capture from monocular video. ACM Trans. Graph. 37(2), 27:1\u201327:15 (2018). https:\/\/doi.org\/10.1145\/3181973","DOI":"10.1145\/3181973"},{"key":"20_CR86","doi-asserted-by":"crossref","unstructured":"Xue, Y., et al.: NSF: neural surface field for human modeling from monocular depth. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV) (2023)","DOI":"10.1109\/ICCV51070.2023.01382"},{"key":"20_CR87","doi-asserted-by":"crossref","unstructured":"Yang, J., Pavone, M., Wang, Y.: FreeNeRF: improving few-shot neural rendering with free frequency regularization (2023)","DOI":"10.1109\/CVPR52729.2023.00798"},{"key":"20_CR88","doi-asserted-by":"crossref","unstructured":"Yu, A., Li, R., Tancik, M., Li, H., Ng, R., Kanazawa, A.: PlenOctrees for real-time rendering of neural radiance fields. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.00570"},{"key":"20_CR89","doi-asserted-by":"crossref","unstructured":"Yu, A., Ye, V., Tancik, M., Kanazawa, A.: pixelNeRF: neural radiance fields from one or few images. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00455"},{"key":"20_CR90","doi-asserted-by":"crossref","unstructured":"Yu, T., et al.: DoubleFusion: real-time capture of human performances with inner body shapes from a single depth sensor. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7287\u20137296 (2018)","DOI":"10.1109\/CVPR.2018.00761"},{"key":"20_CR91","doi-asserted-by":"publisher","unstructured":"Zhang, R., Isola, P., Efros, A.A., Shechtman, E., Wang, O.: The unreasonable effectiveness of deep features as a perceptual metric. In: Conference on Computer Vision and Pattern Recognition (CVPR), Los Alamitos, CA, USA, pp. 586\u2013595. IEEE Computer Society (2018). https:\/\/doi.org\/10.1109\/CVPR.2018.00068. https:\/\/doi.ieeecomputersociety.org\/10.1109\/CVPR.2018.00068","DOI":"10.1109\/CVPR.2018.00068"},{"issue":"6","key":"20_CR92","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3550454.3555451","volume":"41","author":"F Zhao","year":"2022","unstructured":"Zhao, F., et al.: Human performance modeling and rendering via neural animated mesh. ACM Trans. Graph. (TOG) 41(6), 1\u201317 (2022)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"20_CR93","doi-asserted-by":"crossref","unstructured":"Zhao, F., et al.: HumanNeRF: efficiently generated human radiance field from sparse inputs. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 7743\u20137753 (2022)","DOI":"10.1109\/CVPR52688.2022.00759"},{"key":"20_CR94","doi-asserted-by":"crossref","unstructured":"Zheng, Y., et al.: DeepMultiCap: performance capture of multiple characters using sparse multiview cameras. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 6239\u20136249 (2021)","DOI":"10.1109\/ICCV48922.2021.00618"},{"key":"20_CR95","doi-asserted-by":"crossref","unstructured":"Zheng, Z., Huang, H., Yu, T., Zhang, H., Guo, Y., Liu, Y.: Structured local radiance fields for human avatar modeling. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2022)","DOI":"10.1109\/CVPR52688.2022.01543"},{"key":"20_CR96","doi-asserted-by":"crossref","unstructured":"Zheng, Z., Yu, T., Liu, Y., Dai, Q.: PaMIR: parametric model-conditioned implicit representation for image-based human reconstruction (2021)","DOI":"10.1109\/TPAMI.2021.3050505"},{"key":"20_CR97","doi-asserted-by":"crossref","unstructured":"Zheng, Z., Yu, T., Wei, Y., Dai, Q., Liu, Y.: DeepHuman: 3D human reconstruction from a single image. In: The IEEE International Conference on Computer Vision (ICCV) (2019)","DOI":"10.1109\/ICCV.2019.00783"},{"key":"20_CR98","doi-asserted-by":"crossref","unstructured":"Zhu, H., Zhan, F., Theobalt, C., Habermann, M.: TriHuman: a real-time and controllable tri-plane representation for detailed human geometry and appearance synthesis (2023)","DOI":"10.1145\/3697140"},{"key":"20_CR99","doi-asserted-by":"publisher","first-page":"1617","DOI":"10.1109\/TMM.2020.3001506","volume":"23","author":"X Zuo","year":"2020","unstructured":"Zuo, X., et al.: SparseFusion: dynamic human avatar modeling from sparse RGBD images. IEEE Trans. Multimedia 23, 1617\u20131629 (2020)","journal-title":"IEEE Trans. Multimedia"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-72952-2_20","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,28]],"date-time":"2024-11-28T21:40:34Z","timestamp":1732830034000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-72952-2_20"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,1]]},"ISBN":["9783031729515","9783031729522"],"references-count":99,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-72952-2_20","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024,10,1]]},"assertion":[{"value":"1 October 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}