{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,21]],"date-time":"2026-03-21T02:11:25Z","timestamp":1774059085182,"version":"3.50.1"},"publisher-location":"Cham","reference-count":84,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031197680","type":"print"},{"value":"9783031197697","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-19769-7_19","type":"book-chapter","created":{"date-parts":[[2022,10,22]],"date-time":"2022-10-22T11:40:06Z","timestamp":1666438806000},"page":"322-341","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":28,"title":["AvatarCap: Animatable Avatar Conditioned Monocular Human Volumetric Capture"],"prefix":"10.1007","author":[{"given":"Zhe","family":"Li","sequence":"first","affiliation":[]},{"given":"Zerong","family":"Zheng","sequence":"additional","affiliation":[]},{"given":"Hongwen","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Chaonan","family":"Ji","sequence":"additional","affiliation":[]},{"given":"Yebin","family":"Liu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,10,23]]},"reference":[{"key":"19_CR1","doi-asserted-by":"crossref","unstructured":"Alldieck, T., Pons-Moll, G., Theobalt, C., Magnor, M.: Tex2shape: Detailed full human body geometry from a single image. In: ICCV. pp. 2293\u20132303 (2019)","DOI":"10.1109\/ICCV.2019.00238"},{"issue":"4","key":"19_CR2","first-page":"1","volume":"40","author":"T Bagautdinov","year":"2021","unstructured":"Bagautdinov, T., Wu, C., Simon, T., Prada, F., Shiratori, T., Wei, S.E., Xu, W., Sheikh, Y., Saragih, J.: Driving-signal aware full-body avatars. TOG 40(4), 1\u201317 (2021)","journal-title":"Driving-signal aware full-body avatars. TOG"},{"key":"19_CR3","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"311","DOI":"10.1007\/978-3-030-58536-5_19","volume-title":"Computer Vision \u2013 ECCV 2020","author":"BL Bhatnagar","year":"2020","unstructured":"Bhatnagar, B.L., Sminchisescu, C., Theobalt, C., Pons-Moll, G.: Combining implicit function learning and parametric models for 3D human reconstruction. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12347, pp. 311\u2013329. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58536-5_19"},{"issue":"3","key":"19_CR4","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/1360612.1360698","volume":"27","author":"D Bradley","year":"2008","unstructured":"Bradley, D., Popa, T., Sheffer, A., Heidrich, W., Boubekeur, T.: Markerless garment capture. TOG 27(3), 1\u20139 (2008)","journal-title":"TOG"},{"issue":"3","key":"19_CR5","doi-asserted-by":"publisher","first-page":"402","DOI":"10.1109\/TPAMI.2009.32","volume":"32","author":"T Brox","year":"2009","unstructured":"Brox, T., Rosenhahn, B., Gall, J., Cremers, D.: Combined region and motion-based 3d tracking of rigid and articulated objects. IEEE T-PAMI 32(3), 402\u2013415 (2009)","journal-title":"IEEE T-PAMI"},{"key":"19_CR6","doi-asserted-by":"crossref","unstructured":"Burov, A., Nie\u00dfner, M., Thies, J.: Dynamic surface function networks for clothed human bodies. In: ICCV, pp. 10754\u201310764 (2021)","DOI":"10.1109\/ICCV48922.2021.01058"},{"key":"19_CR7","doi-asserted-by":"crossref","unstructured":"Chen, X., Zheng, Y., Black, M.J., Hilliges, O., Geiger, A.: Snarf: differentiable forward skinning for animating non-rigid neural implicit shapes. In: ICCV, pp. 11594\u201311604 (2021)","DOI":"10.1109\/ICCV48922.2021.01139"},{"key":"19_CR8","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"612","DOI":"10.1007\/978-3-030-58571-6_36","volume-title":"Computer Vision \u2013 ECCV 2020","author":"B Deng","year":"2020","unstructured":"Deng, B., et al.: NASA neural articulated shape approximation. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12352, pp. 612\u2013628. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58571-6_36"},{"key":"19_CR9","doi-asserted-by":"crossref","unstructured":"Dong, Z., Guo, C., Song, J., Chen, X., Geiger, A., Hilliges, O.: Pina: learning a personalized implicit neural avatar from a single RGB-D video sequence. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01982"},{"issue":"4","key":"19_CR10","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/2897824.2925969","volume":"35","author":"M Dou","year":"2016","unstructured":"Dou, M., et al.: Fusion4d: real-time performance capture of challenging scenes. TOG 35(4), 1\u201313 (2016)","journal-title":"TOG"},{"key":"19_CR11","doi-asserted-by":"crossref","unstructured":"Gabeur, V., Franco, J.S., Martin, X., Schmid, C., Rogez, G.: Moulding humans: non-parametric 3d human shape estimation from single images. In: ICCV, pp. 2232\u20132241 (2019)","DOI":"10.1109\/ICCV.2019.00232"},{"key":"19_CR12","doi-asserted-by":"crossref","unstructured":"Gall, J., Stoll, C., De Aguiar, E., Theobalt, C., Rosenhahn, B., Seidel, H.P.: Motion capture using joint skeleton tracking and surface estimation. In: CVPR, pp. 1746\u20131753. IEEE (2009)","DOI":"10.1109\/CVPRW.2009.5206755"},{"key":"19_CR13","unstructured":"Goodfellow, I., et al.: Generative adversarial nets. NeurIPS 27 (2014)"},{"issue":"4","key":"19_CR14","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/2185520.2185531","volume":"31","author":"P Guan","year":"2012","unstructured":"Guan, P., Reiss, L., Hirshberg, D.A., Weiss, A., Black, M.J.: Drape: dressing any person. TOG 31(4), 1\u201310 (2012)","journal-title":"TOG"},{"key":"19_CR15","doi-asserted-by":"crossref","unstructured":"Guo, C., Chen, X., Song, J., Hilliges, O.: Human performance capture from monocular video in the wild. In: 3DV, pp. 889\u2013898. IEEE (2021)","DOI":"10.1109\/3DV53792.2021.00097"},{"key":"19_CR16","doi-asserted-by":"crossref","unstructured":"Guo, K., Xu, F., Wang, Y., Liu, Y., Dai, Q.: Robust non-rigid motion tracking and surface reconstruction using l0 regularization. In: ICCV, pp. 3083\u20133091 (2015)","DOI":"10.1109\/ICCV.2015.353"},{"issue":"3","key":"19_CR17","doi-asserted-by":"publisher","first-page":"32:1","DOI":"10.1145\/3083722","volume":"36","author":"K Guo","year":"2017","unstructured":"Guo, K., Xu, F., Yu, T., Liu, X., Dai, Q., Liu, Y.: Real-time geometry, albedo and motion reconstruction using a single RGBD camera. TOG 36(3), 32:1-32:13 (2017)","journal-title":"TOG"},{"issue":"4","key":"19_CR18","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3450626.3459749","volume":"40","author":"M Habermann","year":"2021","unstructured":"Habermann, M., Liu, L., Xu, W., Zollhoefer, M., Pons-Moll, G., Theobalt, C.: Real-time deep dynamic characters. TOG 40(4), 1\u201316 (2021)","journal-title":"TOG"},{"issue":"2","key":"19_CR19","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3311970","volume":"38","author":"M Habermann","year":"2019","unstructured":"Habermann, M., Xu, W., Zollhoefer, M., Pons-Moll, G., Theobalt, C.: Livecap: real-time human performance capture from monocular video. TOG 38(2), 1\u201317 (2019)","journal-title":"TOG"},{"key":"19_CR20","doi-asserted-by":"crossref","unstructured":"Habermann, M., Xu, W., Zollhofer, M., Pons-Moll, G., Theobalt, C.: Deepcap: monocular human performance capture using weak supervision. In: CVPR, pp. 5052\u20135063 (2020)","DOI":"10.1109\/CVPR42600.2020.00510"},{"key":"19_CR21","first-page":"9276","volume":"33","author":"T He","year":"2020","unstructured":"He, T., Collomosse, J., Jin, H., Soatto, S.: Geo-PIFU: geometry and pixel aligned implicit functions for single-view human reconstruction. NeurIPS 33, 9276\u20139287 (2020)","journal-title":"NeurIPS"},{"key":"19_CR22","doi-asserted-by":"crossref","unstructured":"He, T., Xu, Y., Saito, S., Soatto, S., Tung, T.: Arch++: animation-ready clothed human reconstruction revisited. In: ICCV, pp. 11046\u201311056 (2021)","DOI":"10.1109\/ICCV48922.2021.01086"},{"key":"19_CR23","doi-asserted-by":"crossref","unstructured":"He, Y., et al.: Challencap: Monocular 3d capture of challenging human performances using multi-modal references. In: CVPR, pp. 11400\u201311411 (2021)","DOI":"10.1109\/CVPR46437.2021.01124"},{"key":"19_CR24","doi-asserted-by":"crossref","unstructured":"Hong, Y., Zhang, J., Jiang, B., Guo, Y., Liu, L., Bao, H.: Stereopifu: depth aware clothed human digitization via stereo vision. In: CVPR, pp. 535\u2013545 (2021)","DOI":"10.1109\/CVPR46437.2021.00060"},{"key":"19_CR25","doi-asserted-by":"crossref","unstructured":"Huang, Z., Xu, Y., Lassner, C., Li, H., Tung, T.: Arch: animatable reconstruction of clothed humans. In: CVPR, pp. 3093\u20133102 (2020)","DOI":"10.1109\/CVPR42600.2020.00316"},{"key":"19_CR26","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"362","DOI":"10.1007\/978-3-319-46484-8_22","volume-title":"Computer Vision \u2013 ECCV 2016","author":"M Innmann","year":"2016","unstructured":"Innmann, M., Zollh\u00f6fer, M., Nie\u00dfner, M., Theobalt, C., Stamminger, M.: VolumeDeform: real-time volumetric non-rigid reconstruction. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9912, pp. 362\u2013379. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46484-8_22"},{"key":"19_CR27","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"64","DOI":"10.1007\/978-3-030-11018-5_6","volume-title":"Computer Vision \u2013 ECCV 2018 Workshops","author":"AS Jackson","year":"2019","unstructured":"Jackson, A.S., Manafas, C., Tzimiropoulos, G.: 3D human body reconstruction from a single image via volumetric regression. In: Leal-Taix\u00e9, L., Roth, S. (eds.) 3d human body reconstruction from a single image via volumetric regression. LNCS, vol. 11132, pp. 64\u201377. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-11018-5_6"},{"key":"19_CR28","doi-asserted-by":"crossref","unstructured":"Kolotouros, N., Pavlakos, G., Black, M.J., Daniilidis, K.: Learning to reconstruct 3D human pose and shape via model-fitting in the loop. In: ICCV, pp. 2252\u20132261 (2019)","DOI":"10.1109\/ICCV.2019.00234"},{"key":"19_CR29","doi-asserted-by":"crossref","unstructured":"Leroy, V., Franco, J.S., Boyer, E.: Multi-view dynamic shape refinement using local temporal integration. In: ICCV, pp. 3094\u20133103 (2017)","DOI":"10.1109\/ICCV.2017.336"},{"key":"19_CR30","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"324","DOI":"10.1007\/978-3-030-01237-3_20","volume-title":"Computer Vision \u2013 ECCV 2018","author":"C Li","year":"2018","unstructured":"Li, C., Zhao, Z., Guo, X.: ArticulatedFusion: real-time reconstruction of motion, geometry and segmentation using a single depth camera. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11212, pp. 324\u2013340. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01237-3_20"},{"issue":"5","key":"19_CR31","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/1618452.1618521","volume":"28","author":"H Li","year":"2009","unstructured":"Li, H., Adams, B., Guibas, L.J., Pauly, M.: Robust single-view geometry and motion reconstruction. TOG 28(5), 1\u201310 (2009)","journal-title":"TOG"},{"issue":"6","key":"19_CR32","first-page":"1","volume":"32","author":"H Li","year":"2013","unstructured":"Li, H., Vouga, E., Gudym, A., Luo, L., Barron, J.T., Gusev, G.: 3D self-portraits. TOG 32(6), 1\u20139 (2013)","journal-title":"TOG"},{"key":"19_CR33","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"49","DOI":"10.1007\/978-3-030-58592-1_4","volume-title":"Computer Vision \u2013 ECCV 2020","author":"R Li","year":"2020","unstructured":"Li, R., Xiu, Y., Saito, S., Huang, Z., Olszewski, K., Li, H.: Monocular real-time volumetric performance capture. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12368, pp. 49\u201367. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58592-1_4"},{"key":"19_CR34","doi-asserted-by":"crossref","unstructured":"Li, Z., Yu, T., Pan, C., Zheng, Z., Liu, Y.: Robust 3d self-portraits in seconds. In: CVPR, pp. 1344\u20131353 (2020)","DOI":"10.1109\/CVPR42600.2020.00142"},{"key":"19_CR35","doi-asserted-by":"crossref","unstructured":"Li, Z., Yu, T., Zheng, Z., Guo, K., Liu, Y.: Posefusion: pose-guided selective fusion for single-view human volumetric capture. In: CVPR. pp. 14162\u201314172 (2021)","DOI":"10.1109\/CVPR46437.2021.01394"},{"issue":"6","key":"19_CR36","first-page":"1","volume":"40","author":"L Liu","year":"2021","unstructured":"Liu, L., Habermann, M., Rudnev, V., Sarkar, K., Gu, J., Theobalt, C.: Neural actor: neural free-view synthesis of human actors with pose control. TOG 40(6), 1\u201316 (2021)","journal-title":"TOG"},{"issue":"3","key":"19_CR37","first-page":"407","volume":"16","author":"Y Liu","year":"2009","unstructured":"Liu, Y., Dai, Q., Xu, W.: A point-cloud-based multiview stereo algorithm for free-viewpoint video. TVCG 16(3), 407\u2013418 (2009)","journal-title":"TVCG"},{"key":"19_CR38","doi-asserted-by":"crossref","unstructured":"Liu, Y., Stoll, C., Gall, J., Seidel, H.P., Theobalt, C.: Markerless motion capture of interacting characters using multi-view image segmentation. In: CVPR, pp. 1249\u20131256. IEEE (2011)","DOI":"10.1109\/CVPR.2011.5995424"},{"issue":"6","key":"19_CR39","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/2816795.2818013","volume":"34","author":"M Loper","year":"2015","unstructured":"Loper, M., Mahmood, N., Romero, J., Pons-Moll, G., Black, M.J.: SMPL: a skinned multi-person linear model. TOG 34(6), 1\u201316 (2015)","journal-title":"TOG"},{"issue":"4","key":"19_CR40","first-page":"163","volume":"21","author":"WE Lorensen","year":"1987","unstructured":"Lorensen, W.E., Cline, H.E.: Marching cubes: a high resolution 3D surface construction algorithm. TOG 21(4), 163\u2013169 (1987)","journal-title":"TOG"},{"key":"19_CR41","doi-asserted-by":"crossref","unstructured":"Ma, Q., Saito, S., Yang, J., Tang, S., Black, M.J.: Scale: modeling clothed humans with a surface codec of articulated local elements. In: CVPR, pp. 16082\u201316093 (2021)","DOI":"10.1109\/CVPR46437.2021.01582"},{"key":"19_CR42","doi-asserted-by":"crossref","unstructured":"Ma, Q., Yang, J., Ranjan, A., Pujades, S., Pons-Moll, G., Tang, S., Black, M.J.: Learning to dress 3d people in generative clothing. In: CVPR. pp. 6469\u20136478 (2020)","DOI":"10.1109\/CVPR42600.2020.00650"},{"key":"19_CR43","doi-asserted-by":"crossref","unstructured":"Ma, Q., Yang, J., Tang, S., Black, M.J.: The power of points for modeling humans in clothing. In: ICCV, pp. 10974\u201310984 (2021)","DOI":"10.1109\/ICCV48922.2021.01079"},{"key":"19_CR44","unstructured":"Magnenat-Thalmann, N., Laperrire, R., Thalmann, D.: Joint-dependent local deformations for hand animation and object grasping. In: In Proceedings on Graphics Interface. Citeseer (1988)"},{"key":"19_CR45","doi-asserted-by":"crossref","unstructured":"Mescheder, L., Oechsle, M., Niemeyer, M., Nowozin, S., Geiger, A.: Occupancy networks: learning 3D reconstruction in function space. In: CVPR, pp. 4460\u20134470 (2019)","DOI":"10.1109\/CVPR.2019.00459"},{"key":"19_CR46","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"405","DOI":"10.1007\/978-3-030-58452-8_24","volume-title":"Computer Vision \u2013 ECCV 2020","author":"B Mildenhall","year":"2020","unstructured":"Mildenhall, B., Srinivasan, P.P., Tancik, M., Barron, J.T., Ramamoorthi, R., Ng, R.: NeRF: representing scenes as neural radiance fields for view synthesis. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12346, pp. 405\u2013421. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58452-8_24"},{"key":"19_CR47","doi-asserted-by":"crossref","unstructured":"Mustafa, A., Kim, H., Guillemaut, J.Y., Hilton, A.: General dynamic scene reconstruction from multiple view video. In: ICCV, pp. 900\u2013908 (2015)","DOI":"10.1109\/ICCV.2015.109"},{"key":"19_CR48","doi-asserted-by":"crossref","unstructured":"Natsume, R., et al.: Siclope: Silhouette-based clothed people. In: CVPR, pp. 4480\u20134490 (2019)","DOI":"10.1109\/CVPR.2019.00461"},{"key":"19_CR49","doi-asserted-by":"crossref","unstructured":"Newcombe, R.A., Fox, D., Seitz, S.M.: Dynamicfusion: reconstruction and tracking of non-rigid scenes in real-time. In: CVPR, pp. 343\u2013352 (2015)","DOI":"10.1109\/CVPR.2015.7298631"},{"key":"19_CR50","doi-asserted-by":"crossref","unstructured":"Peng, S., Dong, J., Wang, Q., Zhang, S., Shuai, Q., Zhou, X., Bao, H.: Animatable neural radiance fields for modeling dynamic human bodies. In: ICCV, pp. 14314\u201314323 (2021)","DOI":"10.1109\/ICCV48922.2021.01405"},{"issue":"4","key":"19_CR51","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3072959.3073711","volume":"36","author":"G Pons-Moll","year":"2017","unstructured":"Pons-Moll, G., Pujades, S., Hu, S., Black, M.J.: Clothcap: seamless 4D clothing capture and retargeting. TOG 36(4), 1\u201315 (2017)","journal-title":"TOG"},{"key":"19_CR52","doi-asserted-by":"crossref","unstructured":"Saito, S., Huang, Z., Natsume, R., Morishima, S., Kanazawa, A., Li, H.: PIFU: pixel-aligned implicit function for high-resolution clothed human digitization. In: ICCV, pp. 2304\u20132314 (2019)","DOI":"10.1109\/ICCV.2019.00239"},{"key":"19_CR53","doi-asserted-by":"crossref","unstructured":"Saito, S., Simon, T., Saragih, J., Joo, H.: Pifuhd: multi-level pixel-aligned implicit function for high-resolution 3d human digitization. In: CVPR, June 2020","DOI":"10.1109\/CVPR42600.2020.00016"},{"key":"19_CR54","doi-asserted-by":"crossref","unstructured":"Saito, S., Yang, J., Ma, Q., Black, M.J.: Scanimate: weakly supervised learning of skinned clothed avatar networks. In: CVPR, pp. 2886\u20132897 (2021)","DOI":"10.1109\/CVPR46437.2021.00291"},{"key":"19_CR55","doi-asserted-by":"crossref","unstructured":"Shao, R., et al.: Doublefield: Bridging the neural surface and radiance fields for high-fidelity human reconstruction and rendering. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01541"},{"key":"19_CR56","doi-asserted-by":"crossref","unstructured":"Slavcheva, M., Baust, M., Cremers, D., Ilic, S.: Killingfusion: non-rigid 3d reconstruction without correspondences. In: CVPR, pp. 1386\u20131395 (2017)","DOI":"10.1109\/CVPR.2017.581"},{"key":"19_CR57","doi-asserted-by":"crossref","unstructured":"Slavcheva, M., Baust, M., Ilic, S.: Sobolevfusion: 3D reconstruction of scenes undergoing free non-rigid motion. In: CVPR, pp. 2646\u20132655. IEEE, Salt Lake City, June 2018","DOI":"10.1109\/CVPR.2018.00280"},{"key":"19_CR58","doi-asserted-by":"crossref","unstructured":"Smith, D., Loper, M., Hu, X., Mavroidis, P., Romero, J.: Facsimile: fast and accurate scans from an image in less than a second. In: ICCV, pp. 5330\u20135339 (2019)","DOI":"10.1109\/ICCV.2019.00543"},{"issue":"6","key":"19_CR59","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/1882261.1866161","volume":"29","author":"C Stoll","year":"2010","unstructured":"Stoll, C., Gall, J., De Aguiar, E., Thrun, S., Theobalt, C.: Video-based reconstruction of animatable human characters. TOG 29(6), 1\u201310 (2010)","journal-title":"TOG"},{"key":"19_CR60","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"246","DOI":"10.1007\/978-3-030-58548-8_15","volume-title":"Computer Vision \u2013 ECCV 2020","author":"Z Su","year":"2020","unstructured":"Su, Z., Xu, L., Zheng, Z., Yu, T., Liu, Y., Fang, L.: RobustFusion: human volumetric capture with data-driven visual cues using a RGBD camera. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12349, pp. 246\u2013264. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58548-8_15"},{"issue":"3","key":"19_CR61","doi-asserted-by":"publisher","first-page":"80-es","DOI":"10.1145\/1276377.1276478","volume":"26","author":"RW Sumner","year":"2007","unstructured":"Sumner, R.W., Schmid, J., Pauly, M.: Embedded deformation for shape manipulation. TOG 26(3), 80-es (2007)","journal-title":"TOG"},{"key":"19_CR62","doi-asserted-by":"crossref","unstructured":"Suo, X., et al.: Neuralhumanfvv: real-time neural volumetric human performance rendering using RGB cameras. In: CVPR, pp. 6226\u20136237 (2021)","DOI":"10.1109\/CVPR46437.2021.00616"},{"key":"19_CR63","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"20","DOI":"10.1007\/978-3-030-01234-2_2","volume-title":"Computer Vision \u2013 ECCV 2018","author":"G Varol","year":"2018","unstructured":"Varol, G., et al.: BodyNet: volumetric inference of 3D human body shapes. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11211, pp. 20\u201338. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01234-2_2"},{"key":"19_CR64","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"430","DOI":"10.1007\/978-3-030-58565-5_26","volume-title":"Computer Vision \u2013 ECCV 2020","author":"L Wang","year":"2020","unstructured":"Wang, L., Zhao, X., Yu, T., Wang, S., Liu, Y.: NormalGAN: learning detailed 3D human from a single RGB-D image. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12365, pp. 430\u2013446. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58565-5_26"},{"key":"19_CR65","unstructured":"Wang, S., Mihajlovic, M., Ma, Q., Geiger, A., Tang, S.: Metaavatar: learning animatable clothed human models from few depth images. NeurIPS 34 (2021)"},{"key":"19_CR66","doi-asserted-by":"crossref","unstructured":"Wang, T.C., Liu, M.Y., Zhu, J.Y., Tao, A., Kautz, J., Catanzaro, B.: High-resolution image synthesis and semantic manipulation with conditional GANs. In: CVPR, pp. 8798\u20138807 (2018)","DOI":"10.1109\/CVPR.2018.00917"},{"issue":"6","key":"19_CR67","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3478513.3480545","volume":"40","author":"D Xiang","year":"2021","unstructured":"Xiang, D., et al.: Modeling clothing as a separate layer for an animatable human avatar. TOG 40(6), 1\u201315 (2021)","journal-title":"TOG"},{"key":"19_CR68","doi-asserted-by":"crossref","unstructured":"Xiang, D., Prada, F., Wu, C., Hodgins, J.: Monoclothcap: towards temporally coherent clothing capture from monocular RGB video. In: 3DV, pp. 322\u2013332. IEEE (2020)","DOI":"10.1109\/3DV50981.2020.00042"},{"key":"19_CR69","doi-asserted-by":"crossref","unstructured":"Xiu, Y., Yang, J., Tzionas, D., Black, M.J.: Icon: implicit clothed humans obtained from normals. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01294"},{"issue":"2","key":"19_CR70","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3181973","volume":"37","author":"W Xu","year":"2018","unstructured":"Xu, W., et al.: Monoperfcap: human performance capture from monocular video. TOG 37(2), 1\u201315 (2018)","journal-title":"TOG"},{"key":"19_CR71","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"828","DOI":"10.1007\/978-3-642-33709-3_59","volume-title":"Computer Vision \u2013 ECCV 2012","author":"G Ye","year":"2012","unstructured":"Ye, G., Liu, Y., Hasler, N., Ji, X., Dai, Q., Theobalt, C.: Performance capture of interacting characters with handheld kinects. In: Fitzgibbon, A., Lazebnik, S., Perona, P., Sato, Y., Schmid, C. (eds.) ECCV 2012. LNCS, vol. 7573, pp. 828\u2013841. Springer, Heidelberg (2012). https:\/\/doi.org\/10.1007\/978-3-642-33709-3_59"},{"key":"19_CR72","doi-asserted-by":"crossref","unstructured":"Yu, T., et al.: Bodyfusion: real-time capture of human motion and surface geometry using a single depth camera. In: ICCV, Venice, pp. 910\u2013919. IEEE (2017)","DOI":"10.1109\/ICCV.2017.104"},{"key":"19_CR73","doi-asserted-by":"crossref","unstructured":"Yu, T., et al.: Function4d: real-time human volumetric capture from very sparse consumer RGBD sensors. In: CVPR, pp. 5746\u20135756 (2021)","DOI":"10.1109\/CVPR46437.2021.00569"},{"key":"19_CR74","doi-asserted-by":"crossref","unstructured":"Yu, T., et al.: Doublefusion: real-time capture of human performances with inner body shapes from a single depth sensor. In: CVPR, Salt Lake City, pp. 7287\u20137296. IEEE, June 2018","DOI":"10.1109\/CVPR.2018.00761"},{"key":"19_CR75","doi-asserted-by":"crossref","unstructured":"Yu, T., et al.: Simulcap: single-view human performance capture with cloth simulation. In: CVPR, pp. 5499\u20135509. IEEE (2019)","DOI":"10.1109\/CVPR.2019.00565"},{"key":"19_CR76","doi-asserted-by":"crossref","unstructured":"Zhang, H., et al.: Pymaf: 3D human pose and shape regression with pyramidal mesh alignment feedback loop. In: ICCV, pp. 11446\u201311456 (2021)","DOI":"10.1109\/ICCV48922.2021.01125"},{"key":"19_CR77","doi-asserted-by":"crossref","unstructured":"Zheng, Y., et al.: Deepmulticap: performance capture of multiple characters using sparse multiview cameras. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.00618"},{"key":"19_CR78","doi-asserted-by":"crossref","unstructured":"Zheng, Z., Yu, T., Dai, Q., Liu, Y.: Deep implicit templates for 3D shape representation. In: CVPR, pp. 1429\u20131439 (2021)","DOI":"10.1109\/CVPR46437.2021.00148"},{"key":"19_CR79","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"389","DOI":"10.1007\/978-3-030-01240-3_24","volume-title":"Computer Vision \u2013 ECCV 2018","author":"Z Zheng","year":"2018","unstructured":"Zheng, Z., et al.: HybridFusion: real-time performance capture using a single depth sensor and sparse IMUs. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) Hybridfusion: real-time performance capture using a single depth sensor and sparse imus. LNCS, vol. 11213, pp. 389\u2013406. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01240-3_24"},{"key":"19_CR80","doi-asserted-by":"crossref","unstructured":"Zheng, Z., Yu, T., Liu, Y., Dai, Q.: Pamir: parametric model-conditioned implicit representation for image-based human reconstruction. IEEE T-PAMI (2021)","DOI":"10.1109\/TPAMI.2021.3050505"},{"key":"19_CR81","doi-asserted-by":"crossref","unstructured":"Zheng, Z., Yu, T., Wei, Y., Dai, Q., Liu, Y.: Deephuman: 3D human reconstruction from a single image. In: ICCV, pp. 7739\u20137749 (2019)","DOI":"10.1109\/ICCV.2019.00783"},{"key":"19_CR82","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"492","DOI":"10.1007\/978-3-030-58607-2_29","volume-title":"Computer Vision \u2013 ECCV 2020","author":"T Zhi","year":"2020","unstructured":"Zhi, T., Lassner, C., Tung, T., Stoll, C., Narasimhan, S.G., Vo, M.: TexMesh: reconstructing detailed human texture and geometry from RGB-D Video. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12355, pp. 492\u2013509. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58607-2_29"},{"key":"19_CR83","doi-asserted-by":"crossref","unstructured":"Zhu, H., Zuo, X., Wang, S., Cao, X., Yang, R.: Detailed human shape estimation from a single image by hierarchical mesh deformation. In: CVPR, pp. 4491\u20134500 (2019)","DOI":"10.1109\/CVPR.2019.00462"},{"issue":"4","key":"19_CR84","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/2601097.2601165","volume":"33","author":"M Zollh\u00f6fer","year":"2014","unstructured":"Zollh\u00f6fer, M., et al.: Real-time non-rigid reconstruction using an RGB-D camera. TOG 33(4), 1\u201312 (2014)","journal-title":"TOG"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-19769-7_19","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,13]],"date-time":"2024-03-13T14:04:58Z","timestamp":1710338698000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-19769-7_19"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031197680","9783031197697"],"references-count":84,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-19769-7_19","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"23 October 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}