{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T21:57:13Z","timestamp":1775080633325,"version":"3.50.1"},"reference-count":77,"publisher":"Association for Computing Machinery (ACM)","issue":"6","license":[{"start":{"date-parts":[[2024,11,19]],"date-time":"2024-11-19T00:00:00Z","timestamp":1731974400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62172364"],"award-info":[{"award-number":["62172364"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["ACM Trans. Graph."],"published-print":{"date-parts":[[2024,12,19]]},"abstract":"<jats:p>This paper aims to address the challenge of reconstructing long volumetric videos from multi-view RGB videos. Recent dynamic view synthesis methods leverage powerful 4D representations, like feature grids or point cloud sequences, to achieve high-quality rendering results. However, they are typically limited to short (1~2s) video clips and often suffer from large memory footprints when dealing with longer videos. To solve this issue, we propose a novel 4D representation, named Temporal Gaussian Hierarchy, to compactly model long volumetric videos. Our key observation is that there are generally various degrees of temporal redundancy in dynamic scenes, which consist of areas changing at different speeds. Motivated by this, our approach builds a multi-level hierarchy of 4D Gaussian primitives, where each level separately describes scene regions with different degrees of content change, and adaptively shares Gaussian primitives to represent unchanged scene content over different temporal segments, thus effectively reducing the number of Gaussian primitives. In addition, the tree-like structure of the Gaussian hierarchy allows us to efficiently represent the scene at a particular moment with a subset of Gaussian primitives, leading to nearly constant GPU memory usage during the training or rendering regardless of the video length. Moreover, we design a Compact Appearance Model that mixes diffuse and view-dependent Gaussians to further minimize the model size while maintaining the rendering quality. We also develop a rasterization pipeline of Gaussian primitives based on the hardware-accelerated technique to improve rendering speed. Extensive experimental results demonstrate the superiority of our method over alternative methods in terms of training cost, rendering speed, and storage usage. To our knowledge, this work is the first approach capable of efficiently handling hours of volumetric video data while maintaining state-of-the-art rendering quality.<\/jats:p>","DOI":"10.1145\/3687919","type":"journal-article","created":{"date-parts":[[2024,11,19]],"date-time":"2024-11-19T15:46:04Z","timestamp":1732031164000},"page":"1-18","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":21,"title":["Representing Long Volumetric Video with Temporal Gaussian Hierarchy"],"prefix":"10.1145","volume":"43","author":[{"ORCID":"https:\/\/orcid.org\/0009-0002-6098-4198","authenticated-orcid":false,"given":"Zhen","family":"Xu","sequence":"first","affiliation":[{"name":"State Key Lab of CAD and CG, Zhejiang University, Hangzhou, China"},{"name":"Zhejiang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2696-9664","authenticated-orcid":false,"given":"Yinghao","family":"Xu","sequence":"additional","affiliation":[{"name":"Stanford University, California, United States of America"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7220-7789","authenticated-orcid":false,"given":"Zhiyuan","family":"Yu","sequence":"additional","affiliation":[{"name":"Department of Mathematics, Hong Kong University of Science and Technology, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6546-4525","authenticated-orcid":false,"given":"Sida","family":"Peng","sequence":"additional","affiliation":[{"name":"Zhejiang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4053-8510","authenticated-orcid":false,"given":"Jiaming","family":"Sun","sequence":"additional","affiliation":[{"name":"Zhejiang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2662-0334","authenticated-orcid":false,"given":"Hujun","family":"Bao","sequence":"additional","affiliation":[{"name":"State Key Laboratory of CAD &amp; CG, Zhejiang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1926-5597","authenticated-orcid":false,"given":"Xiaowei","family":"Zhou","sequence":"additional","affiliation":[{"name":"State Key Lab of CAD and CG, Zhejiang University, Hangzhou, China"}]}],"member":"320","published-online":{"date-parts":[[2024,11,19]]},"reference":[{"key":"e_1_2_1_1_1","doi-asserted-by":"crossref","unstructured":"Naveed Ahmed Christian Theobalt Christian Rossl Sebastian Thrun and Hans-Peter Seidel. 2008. Dense correspondence finding for parametrization-free animation reconstruction from video.","DOI":"10.1109\/CVPR.2008.4587758"},{"key":"e_1_2_1_2_1","doi-asserted-by":"crossref","unstructured":"Kara-Ali Aliev Artem Sevastopolsky Maria Kolos Dmitry Ulyanov and Victor Lempitsky. 2020. Neural point-based graphics. In ECCV.","DOI":"10.1007\/978-3-030-58542-6_42"},{"key":"e_1_2_1_3_1","volume-title":"Deep-deform: Learning non-rigid rgb-d reconstruction with semi-supervised data. In CVPR.","author":"Bozic Aljaz","year":"2020","unstructured":"Aljaz Bozic, Michael Zollhofer, Christian Theobalt, and Matthias Nie\u00dfner. 2020. Deep-deform: Learning non-rigid rgb-d reconstruction with semi-supervised data. In CVPR."},{"key":"e_1_2_1_4_1","doi-asserted-by":"crossref","unstructured":"Ang Cao and Justin Johnson. 2023. HexPlane: A Fast Representation for Dynamic Scenes. (2023).","DOI":"10.1109\/CVPR52729.2023.00021"},{"key":"e_1_2_1_5_1","volume-title":"Free-viewpoint video of human actors. ACM transactions on graphics (TOG) 22, 3","author":"Carranza Joel","year":"2003","unstructured":"Joel Carranza, Christian Theobalt, Marcus A Magnor, and Hans-Peter Seidel. 2003. Free-viewpoint video of human actors. ACM transactions on graphics (TOG) 22, 3 (2003), 569--577."},{"key":"e_1_2_1_6_1","volume-title":"European Conference on Computer Vision. Springer.","author":"Chen Anpei","year":"2022","unstructured":"Anpei Chen, Zexiang Xu, Andreas Geiger, Jingyi Yu, and Hao Su. 2022. Tensorf: Tensorial radiance fields. In European Conference on Computer Vision. Springer."},{"key":"e_1_2_1_7_1","volume-title":"Mvsnerf: Fast generalizable radiance field reconstruction from multi-view stereo. In ICCV.","author":"Chen Anpei","year":"2021","unstructured":"Anpei Chen, Zexiang Xu, Fuqiang Zhao, Xiaoshuai Zhang, Fanbo Xiang, Jingyi Yu, and Hao Su. 2021. Mvsnerf: Fast generalizable radiance field reconstruction from multi-view stereo. In ICCV."},{"key":"e_1_2_1_8_1","volume-title":"Mobilenerf: Exploiting the polygon rasterization pipeline for efficient neural field rendering on mobile architectures. In CVPR.","author":"Chen Zhiqin","year":"2023","unstructured":"Zhiqin Chen, Thomas Funkhouser, Peter Hedman, and Andrea Tagliasacchi. 2023. Mobilenerf: Exploiting the polygon rasterization pipeline for efficient neural field rendering on mobile architectures. In CVPR."},{"key":"e_1_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1145\/2766945"},{"key":"e_1_2_1_10_1","volume-title":"2021 IEEE. In CVF International Conference on Computer Vision (ICCV). 14304--14314","author":"Du Yilun","year":"2021","unstructured":"Yilun Du, Yinan Zhang, Hong-Xing Yu, Joshua B Tenenbaum, and Jiajun Wu. 2021. Neural radiance flow for 4d view synthesis and video processing. In 2021 IEEE. In CVF International Conference on Computer Vision (ICCV). 14304--14314."},{"key":"e_1_2_1_11_1","volume-title":"4D Gaussian Splatting: Towards Efficient Novel View Synthesis for Dynamic Scenes. arXiv preprint arXiv:2402.03307","author":"Duan Yuanxing","year":"2024","unstructured":"Yuanxing Duan, Fangyin Wei, Qiyu Dai, Yuhang He, Wenzheng Chen, and Baoquan Chen. 2024. 4D Gaussian Splatting: Towards Efficient Novel View Synthesis for Dynamic Scenes. arXiv preprint arXiv:2402.03307 (2024)."},{"key":"e_1_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1145\/3550469.3555383"},{"key":"e_1_2_1_13_1","volume-title":"Benjamin Recht, and Angjoo Kanazawa.","author":"Fridovich-Keil Sara","year":"2023","unstructured":"Sara Fridovich-Keil, Giacomo Meanti, Frederik Rahb\u00e6k Warburg, Benjamin Recht, and Angjoo Kanazawa. 2023. K-Planes: Explicit Radiance Fields in Space, Time, and Appearance. In CVPR."},{"key":"e_1_2_1_14_1","volume-title":"Plenoxels: Radiance fields without neural networks. In CVPR.","author":"Fridovich-Keil Sara","year":"2022","unstructured":"Sara Fridovich-Keil, Alex Yu, Matthew Tancik, Qinhong Chen, Benjamin Recht, and Angjoo Kanazawa. 2022. Plenoxels: Radiance fields without neural networks. In CVPR."},{"key":"e_1_2_1_15_1","volume-title":"Fastnerf: High-fidelity neural rendering at 200fps.","author":"Garbin Stephan J","year":"2021","unstructured":"Stephan J Garbin, Marek Kowalski, Matthew Johnson, Jamie Shotton, and Julien Valentin. 2021. Fastnerf: High-fidelity neural rendering at 200fps."},{"key":"e_1_2_1_16_1","volume-title":"light, and material decomposition from images using monte carlo rendering and denoising. NeuRIPS","author":"Hasselgren Jon","year":"2022","unstructured":"Jon Hasselgren, Nikolai Hofmann, and Jacob Munkberg. 2022. Shape, light, and material decomposition from images using monte carlo rendering and denoising. NeuRIPS (2022)."},{"key":"e_1_2_1_17_1","doi-asserted-by":"crossref","unstructured":"Peter Hedman Pratul P Srinivasan Ben Mildenhall Jonathan T Barron and Paul Debevec. 2021. Baking neural radiance fields for real-time view synthesis.","DOI":"10.1109\/ICCV48922.2021.00582"},{"key":"e_1_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1109\/JRPROC.1952.273898"},{"key":"e_1_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.381"},{"key":"e_1_2_1_20_1","volume-title":"Deformable 3d gaussian splatting for animatable human avatars. arXiv preprint arXiv:2312.15059","author":"Jung HyunJun","year":"2023","unstructured":"HyunJun Jung, Nikolas Brasch, Jifei Song, Eduardo Perez-Pellitero, Yiren Zhou, Zhihao Li, Nassir Navab, and Benjamin Busam. 2023. Deformable 3d gaussian splatting for animatable human avatars. arXiv preprint arXiv:2312.15059 (2023)."},{"key":"e_1_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1145\/3592433"},{"key":"e_1_2_1_22_1","volume-title":"Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980","author":"Kingma Diederik P","year":"2014","unstructured":"Diederik P Kingma and Jimmy Ba. 2014. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)."},{"key":"e_1_2_1_23_1","unstructured":"Georgios Kopanas Bernhard Kerbl Antoine Gu\u00e9don and Jonathon Luiten. 2024. 3D Gaussian Splatting Tutorial. https:\/\/3dgstutorial.github.io\/ International Conference on 3D Vision Tutorial."},{"key":"e_1_2_1_24_1","volume-title":"Imagenet classification with deep convolutional neural networks. Advances in neural information processing systems 25","author":"Krizhevsky Alex","year":"2012","unstructured":"Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton. 2012. Imagenet classification with deep convolutional neural networks. Advances in neural information processing systems 25 (2012)."},{"key":"e_1_2_1_25_1","volume-title":"Tetra-nerf: Representing neural radiance fields using tetrahedra. In ICCV.","author":"Kulhanek Jonas","year":"2023","unstructured":"Jonas Kulhanek and Torsten Sattler. 2023. Tetra-nerf: Representing neural radiance fields using tetrahedra. In ICCV."},{"key":"e_1_2_1_26_1","volume-title":"Pulsar: Efficient sphere-based neural rendering. In CVPR.","author":"Lassner Christoph","year":"2021","unstructured":"Christoph Lassner and Michael Zollhofer. 2021. Pulsar: Efficient sphere-based neural rendering. In CVPR."},{"key":"e_1_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.00544"},{"key":"e_1_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00416"},{"key":"e_1_2_1_29_1","volume-title":"High-Fidelity and Real-Time Novel View Synthesis for Dynamic Scenes. In SIGGRAPH Asia Conference Proceedings.","author":"Lin Haotong","year":"2023","unstructured":"Haotong Lin, Sida Peng, Zhen Xu, Tao Xie, Xingyi He, Hujun Bao, and Xiaowei Zhou. 2023b. High-Fidelity and Real-Time Novel View Synthesis for Dynamic Scenes. In SIGGRAPH Asia Conference Proceedings."},{"key":"e_1_2_1_30_1","volume-title":"Efficient Neural Radiance Fields for Interactive Free-viewpoint Video. In SIGGRAPH Asia Conference Proceedings.","author":"Lin Haotong","year":"2022","unstructured":"Haotong Lin, Sida Peng, Zhen Xu, Yunzhi Yan, Qing Shuai, Hujun Bao, and Xiaowei Zhou. 2022. Efficient Neural Radiance Fields for Interactive Free-viewpoint Video. In SIGGRAPH Asia Conference Proceedings."},{"key":"e_1_2_1_31_1","volume-title":"Gaussian-flow: 4d reconstruction with dynamic 3d gaussian particle. arXiv preprint arXiv:2312.03431","author":"Lin Youtian","year":"2023","unstructured":"Youtian Lin, Zuozhuo Dai, Siyu Zhu, and Yao Yao. 2023a. Gaussian-flow: 4d reconstruction with dynamic 3d gaussian particle. arXiv preprint arXiv:2312.03431 (2023)."},{"key":"e_1_2_1_32_1","volume-title":"Antonio Torralba, Sanja Fidler, and Karsten Kreis.","author":"Ling Huan","year":"2023","unstructured":"Huan Ling, Seung Wook Kim, Antonio Torralba, Sanja Fidler, and Karsten Kreis. 2023. Align your gaussians: Text-to-4d with dynamic 3d gaussians and composed diffusion models. arXiv preprint arXiv:2312.13763 (2023)."},{"key":"e_1_2_1_33_1","volume-title":"Neural volumes: Learning dynamic renderable volumes from images. arXiv preprint arXiv:1906.07751","author":"Lombardi Stephen","year":"2019","unstructured":"Stephen Lombardi, Tomas Simon, Jason Saragih, Gabriel Schwartz, Andreas Lehrmann, and Yaser Sheikh. 2019. Neural volumes: Learning dynamic renderable volumes from images. arXiv preprint arXiv:1906.07751 (2019)."},{"key":"e_1_2_1_34_1","volume-title":"Sparseneus: Fast generalizable neural surface reconstruction from sparse views.","author":"Long Xiaoxiao","year":"2022","unstructured":"Xiaoxiao Long, Cheng Lin, Peng Wang, Taku Komura, and Wenping Wang. 2022. Sparseneus: Fast generalizable neural surface reconstruction from sparse views."},{"key":"e_1_2_1_35_1","doi-asserted-by":"crossref","unstructured":"Fan Lu Yan Xu Guang Chen Hongsheng Li Kwan-Yee Lin and Changjun Jiang. 2023. Urban radiance field representation with deformable neural mesh primitives. In ICCV.","DOI":"10.1109\/ICCV51070.2023.00049"},{"key":"e_1_2_1_36_1","volume-title":"3d geometry-aware deformable gaussian splatting for dynamic view synthesis. arXiv preprint arXiv:2404.06270","author":"Lu Zhicheng","year":"2024","unstructured":"Zhicheng Lu, Xiang Guo, Le Hui, Tianrui Chen, Min Yang, Xiao Tang, Feng Zhu, and Yuchao Dai. 2024. 3d geometry-aware deformable gaussian splatting for dynamic view synthesis. arXiv preprint arXiv:2404.06270 (2024)."},{"key":"e_1_2_1_37_1","doi-asserted-by":"crossref","unstructured":"Jonathon Luiten Georgios Kopanas Bastian Leibe and Deva Ramanan. 2024. Dynamic 3D Gaussians: Tracking by Persistent Dynamic View Synthesis. In 3DV.","DOI":"10.1109\/3DV62453.2024.00044"},{"key":"e_1_2_1_38_1","volume-title":"Occupancy Networks: Learning 3D Reconstruction in Function Space. In CVPR.","author":"Mescheder Lars","year":"2019","unstructured":"Lars Mescheder, Michael Oechsle, Michael Niemeyer, Sebastian Nowozin, and Andreas Geiger. 2019. Occupancy Networks: Learning 3D Reconstruction in Function Space. In CVPR."},{"key":"e_1_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1145\/3503250"},{"key":"e_1_2_1_40_1","unstructured":"mkkellogg. 2024. GaussianSplats3D. https:\/\/github.com\/mkkellogg\/GaussianSplats3D."},{"key":"e_1_2_1_41_1","volume-title":"Spherical harmonics","author":"M\u00fcller Claus","unstructured":"Claus M\u00fcller. 2006. Spherical harmonics. Vol. 17. Springer."},{"key":"e_1_2_1_42_1","volume-title":"Instant neural graphics primitives with a multiresolution hash encoding. ACM transactions on graphics (TOG)","author":"M\u00fcller Thomas","year":"2022","unstructured":"Thomas M\u00fcller, Alex Evans, Christoph Schied, and Alexander Keller. 2022. Instant neural graphics primitives with a multiresolution hash encoding. ACM transactions on graphics (TOG) (2022)."},{"key":"e_1_2_1_43_1","volume-title":"Dynamicfusion: Reconstruction and tracking of non-rigid scenes in real-time.","author":"Newcombe Richard A","year":"2015","unstructured":"Richard A Newcombe, Dieter Fox, and Steven M Seitz. 2015. Dynamicfusion: Reconstruction and tracking of non-rigid scenes in real-time."},{"key":"e_1_2_1_44_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00025"},{"key":"e_1_2_1_45_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00581"},{"key":"e_1_2_1_46_1","volume-title":"Hypernerf: A higher-dimensional representation for topologically varying neural radiance fields. arXiv preprint arXiv:2106.13228","author":"Park Keunhong","year":"2021","unstructured":"Keunhong Park, Utkarsh Sinha, Peter Hedman, Jonathan T Barron, Sofien Bouaziz, Dan B Goldman, Ricardo Martin-Brualla, and Steven M Seitz. 2021b. Hypernerf: A higher-dimensional representation for topologically varying neural radiance fields. arXiv preprint arXiv:2106.13228 (2021)."},{"key":"e_1_2_1_47_1","volume-title":"PyTorch: An Imperative Style","author":"Paszke Adam","unstructured":"Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, Alban Desmaison, Andreas Kopf, Edward Yang, Zachary DeVito, Martin Raison, Alykhan Tejani, Sasank Chilamkurthy, Benoit Steiner, Lu Fang, Junjie Bai, and Soumith Chintala. 2019. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In NeurIPS."},{"key":"e_1_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01018"},{"key":"e_1_2_1_49_1","volume-title":"D-nerf: Neural radiance fields for dynamic scenes.","author":"Pumarola Albert","year":"2021","unstructured":"Albert Pumarola, Enric Corona, Gerard Pons-Moll, and Francesc Moreno-Noguer. 2021b. D-nerf: Neural radiance fields for dynamic scenes."},{"key":"e_1_2_1_50_1","doi-asserted-by":"crossref","unstructured":"Ruslan Rakhimov Andrei-Timotei Ardelean Victor Lempitsky and Evgeny Burnaev. 2022. Npbg+ + : Accelerating neural point-based graphics. In CVPR.","DOI":"10.1109\/CVPR52688.2022.01550"},{"key":"e_1_2_1_51_1","volume-title":"Dreamgaussian4d: Generative 4d gaussian splatting. arXiv preprint arXiv:2312.17142","author":"Ren Jiawei","year":"2023","unstructured":"Jiawei Ren, Liang Pan, Jiaxiang Tang, Chi Zhang, Ang Cao, Gang Zeng, and Ziwei Liu. 2023. Dreamgaussian4d: Generative 4d gaussian splatting. arXiv preprint arXiv:2312.17142 (2023)."},{"key":"e_1_2_1_52_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"e_1_2_1_53_1","volume-title":"Adop: Approximate differentiable one-pixel point rendering. ACM Transactions on Graphics (ToG)","author":"R\u00fcckert Darius","year":"2022","unstructured":"Darius R\u00fcckert, Linus Franke, and Marc Stamminger. 2022. Adop: Approximate differentiable one-pixel point rendering. ACM Transactions on Graphics (ToG) (2022)."},{"key":"e_1_2_1_54_1","volume-title":"CUDA by example: an introduction to generalpurpose GPU programming","author":"Sanders Jason","unstructured":"Jason Sanders and Edward Kandrot. 2010. CUDA by example: an introduction to generalpurpose GPU programming. Addison-Wesley Professional."},{"key":"e_1_2_1_55_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.445"},{"key":"e_1_2_1_56_1","unstructured":"Dave Shreiner et al. 2009. OpenGL programming guide: the official guide to learning OpenGL versions 3.0 and 3.1. Pearson Education."},{"key":"e_1_2_1_57_1","doi-asserted-by":"publisher","DOI":"10.1145\/3528233.3530727"},{"key":"e_1_2_1_58_1","volume-title":"Compressible-composable nerf via rank-residual decomposition. NeuRIPS","author":"Tang Jiaxiang","year":"2022","unstructured":"Jiaxiang Tang, Xiaokang Chen, Jingbo Wang, and Gang Zeng. 2022. Compressible-composable nerf via rank-residual decomposition. NeuRIPS (2022)."},{"key":"e_1_2_1_59_1","doi-asserted-by":"crossref","unstructured":"Edgar Tretschk Ayush Tewari Vladislav Golyanik Michael Zollh\u00f6fer Christoph Lassner and Christian Theobalt. 2021a. Non-rigid neural radiance fields: Reconstruction and novel view synthesis of a dynamic scene from monocular video.","DOI":"10.1109\/ICCV48922.2021.01272"},{"key":"e_1_2_1_60_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.01272"},{"key":"e_1_2_1_61_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01316"},{"key":"e_1_2_1_62_1","volume-title":"Ibrnet: Learning multi-view image-based rendering.","author":"Wang Qianqian","year":"2021","unstructured":"Qianqian Wang, Zhicheng Wang, Kyle Genova, Pratul P Srinivasan, Howard Zhou, Jonathan T Barron, Ricardo Martin-Brualla, Noah Snavely, and Thomas Funkhouser. 2021. Ibrnet: Learning multi-view image-based rendering."},{"key":"e_1_2_1_63_1","volume-title":"Image quality assessment: from error visibility to structural similarity","author":"Wang Zhou","year":"2004","unstructured":"Zhou Wang, Alan C Bovik, Hamid R Sheikh, and Eero P Simoncelli. 2004a. Image quality assessment: from error visibility to structural similarity. IEEE transactions on image processing 13, 4 (2004), 600--612."},{"key":"e_1_2_1_64_1","volume-title":"Image quality assessment: from error visibility to structural similarity","author":"Wang Zhou","year":"2004","unstructured":"Zhou Wang, Alan C Bovik, Hamid R Sheikh, and Eero P Simoncelli. 2004b. Image quality assessment: from error visibility to structural similarity. IEEE transactions on image processing (2004)."},{"key":"e_1_2_1_65_1","volume-title":"4D Gaussian Splatting for Real-Time Dynamic Scene Rendering. arXiv preprint arXiv:2310.08528","author":"Wu Guanjun","year":"2023","unstructured":"Guanjun Wu, Taoran Yi, Jiemin Fang, Lingxi Xie, Xiaopeng Zhang, Wei Wei, Wenyu Liu, Qi Tian, and Wang Xinggang. 2023b. 4D Gaussian Splatting for Real-Time Dynamic Scene Rendering. arXiv preprint arXiv:2310.08528 (2023)."},{"key":"e_1_2_1_66_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00175"},{"key":"e_1_2_1_67_1","unstructured":"Rundi Wu Ben Mildenhall Philipp Henzler Keunhong Park Ruiqi Gao Daniel Watson Pratul P Srinivasan Dor Verbin Jonathan T Barron Ben Poole et al. 2023a. Reconfusion: 3d reconstruction with diffusion priors. arXiv preprint arXiv:2312.02981 (2023)."},{"key":"e_1_2_1_68_1","unstructured":"Zhen Xu Sida Peng Chen Geng Linzhan Mou Zihan Yan Jiaming Sun Hujun Bao and Xiaowei Zhou. 2024a. Relightable and Animatable Neural Avatar from Sparse-View Video. In CVPR."},{"key":"e_1_2_1_69_1","unstructured":"Zhen Xu Sida Peng Haotong Lin Guangzhao He Jiaming Sun Yujun Shen Hujun Bao and Xiaowei Zhou. 2024b. 4K4D: Real-Time 4D View Synthesis at 4K Resolution. In CVPR."},{"key":"e_1_2_1_70_1","unstructured":"Zhen Xu Tao Xie Sida Peng Haotong Lin Qing Shuai Zhiyuan Yu Guangzhao He Jiaming Sun Hujun Bao and Xiaowei Zhou. 2023. EasyVolcap: Accelerating Neural Volumetric Video Research. (2023)."},{"key":"e_1_2_1_71_1","volume-title":"Deformable 3D Gaussians for High-Fidelity Monocular Dynamic Scene Reconstruction. arXiv preprint arXiv:2309.13101","author":"Yang Ziyi","year":"2023","unstructured":"Ziyi Yang, Xinyu Gao, Wen Zhou, Shaohui Jiao, Yuqing Zhang, and Xiaogang Jin. 2023a. Deformable 3D Gaussians for High-Fidelity Monocular Dynamic Scene Reconstruction. arXiv preprint arXiv:2309.13101 (2023)."},{"key":"e_1_2_1_72_1","volume-title":"Real-time Photorealistic Dynamic Scene Representation and Rendering with 4D Gaussian Splatting. arXiv preprint arXiv 2310.10642","author":"Yang Zeyu","year":"2023","unstructured":"Zeyu Yang, Hongye Yang, Zijie Pan, Xiatian Zhu, and Li Zhang. 2023b. Real-time Photorealistic Dynamic Scene Representation and Rendering with 4D Gaussian Splatting. arXiv preprint arXiv 2310.10642 (2023)."},{"key":"e_1_2_1_73_1","unstructured":"Alex Yu Vickie Ye Matthew Tancik and Angjoo Kanazawa. 2021. pixelnerf: Neural radiance fields from one or few images."},{"key":"e_1_2_1_74_1","volume-title":"Bodyfusion: Real-time capture of human motion and surface geometry using a single depth camera.","author":"Yu Tao","year":"2017","unstructured":"Tao Yu, Kaiwen Guo, Feng Xu, Yuan Dong, Zhaoqi Su, Jianhui Zhao, Jianguo Li, Qionghai Dai, and Yebin Liu. 2017. Bodyfusion: Real-time capture of human motion and surface geometry using a single depth camera."},{"key":"e_1_2_1_75_1","volume-title":"Doublefusion: Real-time capture of human performances with inner body shapes from a single depth sensor.","author":"Yu Tao","year":"2018","unstructured":"Tao Yu, Zerong Zheng, Kaiwen Guo, Jianhui Zhao, Qionghai Dai, Hao Li, Gerard Pons-Moll, and Yebin Liu. 2018. Doublefusion: Real-time capture of human performances with inner body shapes from a single depth sensor."},{"key":"e_1_2_1_76_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00068"},{"key":"e_1_2_1_77_1","doi-asserted-by":"crossref","unstructured":"Richard Zhang Phillip Isola Alexei A Efros Eli Shechtman and Oliver Wang. 2018b. The Unreasonable Effectiveness of Deep Features as a Perceptual Metric. In CVPR.","DOI":"10.1109\/CVPR.2018.00068"}],"container-title":["ACM Transactions on Graphics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3687919","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3687919","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T01:09:57Z","timestamp":1750295397000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3687919"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,19]]},"references-count":77,"journal-issue":{"issue":"6","published-print":{"date-parts":[[2024,12,19]]}},"alternative-id":["10.1145\/3687919"],"URL":"https:\/\/doi.org\/10.1145\/3687919","relation":{},"ISSN":["0730-0301","1557-7368"],"issn-type":[{"value":"0730-0301","type":"print"},{"value":"1557-7368","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,11,19]]},"assertion":[{"value":"2024-11-19","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}