{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,4]],"date-time":"2026-03-04T17:05:52Z","timestamp":1772643952034,"version":"3.50.1"},"reference-count":45,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2025,12,27]],"date-time":"2025-12-27T00:00:00Z","timestamp":1766793600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,12,27]],"date-time":"2025-12-27T00:00:00Z","timestamp":1766793600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2026,1]]},"DOI":"10.1007\/s00371-025-04229-5","type":"journal-article","created":{"date-parts":[[2025,12,27]],"date-time":"2025-12-27T05:26:06Z","timestamp":1766813166000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Label-guided 4D Gaussian splatting for high-fidelity dynamic scene reconstruction"],"prefix":"10.1007","volume":"42","author":[{"given":"Beibei","family":"Wang","sequence":"first","affiliation":[]},{"given":"Weiwei","family":"Fu","sequence":"additional","affiliation":[]},{"given":"Tianyou","family":"Zheng","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,12,27]]},"reference":[{"issue":"1","key":"4229_CR1","doi-asserted-by":"publisher","first-page":"99","DOI":"10.1145\/3503250","volume":"65","author":"B Mildenhall","year":"2021","unstructured":"Mildenhall, B., Srinivasan, P.P., Tancik, M., Barron, J.T., Ramamoorthi, R., Ng, R.: Nerf: Representing scenes as neural radiance fields for view synthesis. Commun. ACM 65(1), 99\u2013106 (2021)","journal-title":"Commun. ACM"},{"key":"4229_CR2","doi-asserted-by":"publisher","DOI":"10.1007\/s00371-025-03974-x","author":"T Bao","year":"2025","unstructured":"Bao, T., Ye, J., Bao, Z.: L2h-nerf: low- to high-frequency-guided nerf for 3d reconstruction with a few input scenes. Vis. Comput. (2025). https:\/\/doi.org\/10.1007\/s00371-025-03974-x","journal-title":"Vis. Comput."},{"key":"4229_CR3","doi-asserted-by":"publisher","first-page":"5043","DOI":"10.1007\/s00371-024-03507-y","volume":"40","author":"T Wirth","year":"2024","unstructured":"Wirth, T., Rak, A., Buelow, M.: Nerf-ff: a plug-in method to mitigate defocus blur for runtime optimized neural radiance fields. Vis. Comput. 40, 5043\u20135055 (2024). https:\/\/doi.org\/10.1007\/s00371-024-03507-y","journal-title":"Vis. Comput."},{"key":"4229_CR4","doi-asserted-by":"publisher","first-page":"991","DOI":"10.1007\/s00371-024-03379-2","volume":"41","author":"J Yuan","year":"2025","unstructured":"Yuan, J., Fan, M., Liu, Z.: Collaborative neural radiance fields for novel view synthesis. Vis. Comput. 41, 991\u20131006 (2025). https:\/\/doi.org\/10.1007\/s00371-024-03379-2","journal-title":"Vis. Comput."},{"key":"4229_CR5","doi-asserted-by":"crossref","unstructured":"Pumarola, A., Corona, E., Pons-Moll, G., Moreno-Noguer, F.: D-nerf: Neural radiance fields for dynamic scenes. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10318\u201310327 (2021)","DOI":"10.1109\/CVPR46437.2021.01018"},{"key":"4229_CR6","doi-asserted-by":"crossref","unstructured":"Park, K., Sinha, U., Barron, J.T., Bouaziz, S., Goldman, D.B., Seitz, S.M., Martin-Brualla, R.: Nerfies: Deformable neural radiance fields. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 5865\u20135874 (2021)","DOI":"10.1109\/ICCV48922.2021.00581"},{"key":"4229_CR7","doi-asserted-by":"crossref","unstructured":"Park, K., Sinha, U., Hedman, P., Barron, J.T., Bouaziz, S., Goldman, D.B., Martin-Brualla, R., Seitz, S.M.: Hypernerf: A higher-dimensional representation for topologically varying neural radiance fields. arXiv preprint arXiv:2106.13228 (2021)","DOI":"10.1145\/3478513.3480487"},{"key":"4229_CR8","doi-asserted-by":"crossref","unstructured":"Li, Z., Niklaus, S., Snavely, N., Wang, O.: Neural scene flow fields for space-time view synthesis of dynamic scenes. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 6498\u20136508 (2021)","DOI":"10.1109\/CVPR46437.2021.00643"},{"key":"4229_CR9","doi-asserted-by":"crossref","unstructured":"Fang, J., Yi, T., Wang, X., Xie, L., Zhang, X., Liu, W., Nie\u00dfner, M., Tian, Q.: Fast dynamic radiance fields with time-aware neural voxels. In: SIGGRAPH Asia 2022 Conference Papers, pp. 1\u20139 (2022)","DOI":"10.1145\/3550469.3555383"},{"issue":"4","key":"4229_CR10","doi-asserted-by":"publisher","first-page":"139","DOI":"10.1145\/3592433","volume":"42","author":"B Kerbl","year":"2023","unstructured":"Kerbl, B., Kopanas, G., Leimk\u00fchler, T., Drettakis, G.: 3d Gaussian splatting for real-time radiance field rendering. ACM Trans. Graph. 42(4), 139 (2023)","journal-title":"ACM Trans. Graph."},{"key":"4229_CR11","doi-asserted-by":"crossref","unstructured":"Cen, J., Fang, J., Yang, C., Xie, L., Zhang, X., Shen, W., Tian, Q.: Segment any 3d gaussians. Proceedings of the AAAI Conference on Artificial Intelligence. 39, 1971\u20131979 (2025)","DOI":"10.1609\/aaai.v39i2.32193"},{"key":"4229_CR12","doi-asserted-by":"crossref","unstructured":"Huang, B., Yu, Z., Chen, A., Geiger, A., Gao, S.: 2d Gaussian splatting for geometrically accurate radiance fields. In: ACM SIGGRAPH 2024 Conference Papers, pp. 1\u201311 (2024)","DOI":"10.1145\/3641519.3657428"},{"key":"4229_CR13","unstructured":"Xu, B., Hu, J., Li, J., He, Y.: Gsurf: 3d reconstruction via signed distance fields with direct gaussian supervision. arXiv preprint arXiv:2411.15723 (2024)"},{"issue":"6","key":"4229_CR14","first-page":"1","volume":"43","author":"Z Yu","year":"2024","unstructured":"Yu, Z., Sattler, T., Geiger, A.: Gaussian opacity fields: Efficient adaptive surface reconstruction in unbounded scenes. ACM Transact. Graph. (ToG) 43(6), 1\u201313 (2024)","journal-title":"ACM Transact. Graph. (ToG)"},{"key":"4229_CR15","unstructured":"Zhang, Z., Huang, B., Jiang, H., Zhou, L., Xiang, X., Shen, S.: Quadratic gaussian splatting for efficient and detailed surface reconstruction. arXiv preprint arXiv:2411.16392 (2024)"},{"key":"4229_CR16","doi-asserted-by":"crossref","unstructured":"Zhang, S., Wu, G., Wang, X., Feng, B., Liu, W.: Dynamic 2d gaussians: Geometrically accurate radiance fields for dynamic objects. arXiv preprint arXiv:2409.14072 (2024)","DOI":"10.1145\/3746027.3755328"},{"key":"4229_CR17","doi-asserted-by":"crossref","unstructured":"Yang, Z., Gao, X., Zhou, W., Jiao, S., Zhang, Y., Jin, X.: Deformable 3d Gaussians for high-fidelity monocular dynamic scene reconstruction. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 20331\u201320341 (2024)","DOI":"10.1109\/CVPR52733.2024.01922"},{"key":"4229_CR18","doi-asserted-by":"crossref","unstructured":"Wu, G., Yi, T., Fang, J., Xie, L., Zhang, X., Wei, W., Liu, W., Tian, Q., Wang, X.: 4d Gaussian splatting for real-time dynamic scene rendering. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 20310\u201320320 (2024)","DOI":"10.1109\/CVPR52733.2024.01920"},{"key":"4229_CR19","doi-asserted-by":"crossref","unstructured":"Luiten, J., Kopanas, G., Leibe, B., Ramanan, D.: Dynamic 3d gaussians: Tracking by persistent dynamic view synthesis. In: 2024 International Conference on 3D Vision (3DV), pp. 800\u2013809 (2024). IEEE","DOI":"10.1109\/3DV62453.2024.00044"},{"key":"4229_CR20","unstructured":"Yang, Z., Yang, H., Pan, Z., Zhang, L.: Real-time photorealistic dynamic scene representation and rendering with 4d gaussian splatting. arXiv preprint arXiv:2310.10642 (2023)"},{"key":"4229_CR21","doi-asserted-by":"crossref","unstructured":"Kirillov, A., Mintun, E., Ravi, N., Mao, H., Rolland, C., Gustafson, L., Xiao, T., Whitehead, S., Berg, A.C., Lo, W.-Y.: Segment anything. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 4015\u20134026 (2023)","DOI":"10.1109\/ICCV51070.2023.00371"},{"key":"4229_CR22","doi-asserted-by":"crossref","unstructured":"Cao, A., Johnson, J.: Hexplane: A fast representation for dynamic scenes. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 130\u2013141 (2023)","DOI":"10.1109\/CVPR52729.2023.00021"},{"key":"4229_CR23","doi-asserted-by":"crossref","unstructured":"Gao, C., Saraf, A., Kopf, J., Huang, J.-B.: Dynamic view synthesis from dynamic monocular video. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 5712\u20135721 (2021)","DOI":"10.1109\/ICCV48922.2021.00566"},{"key":"4229_CR24","doi-asserted-by":"crossref","unstructured":"Li, Z., Niklaus, S., Snavely, N., Wang, O.: Neural scene flow fields for space-time view synthesis of dynamic scenes. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 6498\u20136508 (2021)","DOI":"10.1109\/CVPR46437.2021.00643"},{"key":"4229_CR25","doi-asserted-by":"crossref","unstructured":"Weng, C.-Y., Curless, B., Srinivasan, P.P., Barron, J.T., Kemelmacher-Shlizerman, I.: Humannerf: Free-viewpoint rendering of moving people from monocular video. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 16210\u201316220 (2022)","DOI":"10.1109\/CVPR52688.2022.01573"},{"key":"4229_CR26","doi-asserted-by":"crossref","unstructured":"Chen, A., Xu, Z., Geiger, A., Yu, J., Su, H.: Tensorf: Tensorial radiance fields. In: European Conference on Computer Vision, pp. 333\u2013350 (2022). Springer","DOI":"10.1007\/978-3-031-19824-3_20"},{"key":"4229_CR27","doi-asserted-by":"crossref","unstructured":"Fridovich-Keil, S., Yu, A., Tancik, M., Chen, Q., Recht, B., Kanazawa, A.: Plenoxels: Radiance fields without neural networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5501\u20135510 (2022)","DOI":"10.1109\/CVPR52688.2022.00542"},{"key":"4229_CR28","doi-asserted-by":"crossref","unstructured":"Fridovich-Keil, S., Meanti, G., Warburg, F.R., Recht, B., Kanazawa, A.: K-planes: Explicit radiance fields in space, time, and appearance. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12479\u201312488 (2023)","DOI":"10.1109\/CVPR52729.2023.01201"},{"key":"4229_CR29","doi-asserted-by":"crossref","unstructured":"Liang, Y., Khan, N., Li, Z., Nguyen-Phuoc, T., Lanman, D., Tompkin, J., Xiao, L.: Gaufre: Gaussian deformation fields for real-time dynamic novel view synthesis. In: 2025 IEEE\/CVF Winter Conference on Applications of Computer Vision (WACV), pp. 2642\u20132652 (2025). IEEE","DOI":"10.1109\/WACV61041.2025.00262"},{"key":"4229_CR30","unstructured":"Duisterhof, B.P., Mandi, Z., Yao, Y., Liu, J.-W., Seidenschwarz, J., Shou, M.Z., Ramanan, D., Song, S., Birchfield, S., Wen, B., et al.: Deformgs: Scene flow in highly deformable scenes for deformable object manipulation. arXiv preprint arXiv:2312.00583 (2023)"},{"key":"4229_CR31","doi-asserted-by":"crossref","unstructured":"Wu, G., Yi, T., Fang, J., Xie, L., Zhang, X., Wei, W., Liu, W., Tian, Q., Wang, X.: 4d Gaussian splatting for real-time dynamic scene rendering. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 20310\u201320320 (2024)","DOI":"10.1109\/CVPR52733.2024.01920"},{"key":"4229_CR32","doi-asserted-by":"crossref","unstructured":"Duan, Y., Wei, F., Dai, Q., He, Y., Chen, W., Chen, B.: 4d-rotor Gaussian splatting: towards efficient novel view synthesis for dynamic scenes. In: ACM SIGGRAPH 2024 Conference Papers, pp. 1\u201311 (2024)","DOI":"10.1145\/3641519.3657463"},{"key":"4229_CR33","doi-asserted-by":"crossref","unstructured":"Lin, Y., Dai, Z., Zhu, S., Yao, Y.: Gaussian-flow: 4d reconstruction with dynamic 3d Gaussian particle. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 21136\u201321145 (2024)","DOI":"10.1109\/CVPR52733.2024.01997"},{"key":"4229_CR34","doi-asserted-by":"crossref","unstructured":"Kratimenos, A., Lei, J., Daniilidis, K.: Dynmf: Neural motion factorization for real-time dynamic view synthesis with 3d gaussian splatting. In: European Conference on Computer Vision, pp. 252\u2013269 (2024). Springer","DOI":"10.1007\/978-3-031-72986-7_15"},{"key":"4229_CR35","doi-asserted-by":"crossref","unstructured":"Li, Z., Chen, Z., Li, Z., Xu, Y.: Spacetime Gaussian feature splatting for real-time dynamic view synthesis. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8508\u20138520 (2024)","DOI":"10.1109\/CVPR52733.2024.00813"},{"key":"4229_CR36","doi-asserted-by":"crossref","unstructured":"Katsumata, K., Vo, D.M., Nakayama, H.: A compact dynamic 3d Gaussian representation for real-time dynamic view synthesis. In: European Conference on Computer Vision, pp. 394\u2013412 (2024). Springer","DOI":"10.1007\/978-3-031-73016-0_23"},{"key":"4229_CR37","unstructured":"Guo, M., Fathi, A., Wu, J., Funkhouser, T.: Object-centric neural scene rendering. arXiv preprint arXiv:2012.08503 (2020)"},{"key":"4229_CR38","doi-asserted-by":"crossref","unstructured":"Wang, C., Chai, M., He, M., Chen, D., Liao, J.: Clip-nerf: Text-and-image driven manipulation of neural radiance fields. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3835\u20133844 (2022)","DOI":"10.1109\/CVPR52688.2022.00381"},{"key":"4229_CR39","doi-asserted-by":"crossref","unstructured":"Ye, M., Danelljan, M., Yu, F., Ke, L.: Gaussian grouping: Segment and edit anything in 3d scenes. In: European Conference on Computer Vision, pp. 162\u2013179 (2024). Springer","DOI":"10.1007\/978-3-031-73397-0_10"},{"key":"4229_CR40","unstructured":"Li, Y.-J., Gladkova, M., Xia, Y., Cremers, D.: Sadg: Segment any dynamic gaussian without object trackers. arXiv preprint arXiv:2411.19290 (2024)"},{"key":"4229_CR41","doi-asserted-by":"crossref","unstructured":"Li, T., Slavcheva, M., Zollhoefer, M., Green, S., Lassner, C., Kim, C., Schmidt, T., Lovegrove, S., Goesele, M., Newcombe, R.: Neural 3d video synthesis from multi-view video. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5521\u20135531 (2022)","DOI":"10.1109\/CVPR52688.2022.00544"},{"issue":"5","key":"4229_CR42","doi-asserted-by":"publisher","first-page":"2732","DOI":"10.1109\/TVCG.2023.3247082","volume":"29","author":"L Song","year":"2023","unstructured":"Song, L., Chen, A., Li, Z., Chen, Z., Chen, L., Yuan, J., Xu, Y., Geiger, A.: Nerfplayer: A streamable dynamic scene representation with decomposed neural radiance fields. IEEE Trans. Visual Comput. Graph. 29(5), 2732\u20132742 (2023). https:\/\/doi.org\/10.1109\/TVCG.2023.3247082","journal-title":"IEEE Trans. Visual Comput. Graph."},{"key":"4229_CR43","doi-asserted-by":"crossref","unstructured":"Attal, B., Huang, J.-B., Richardt, C., Zollhoefer, M., Kopf, J., O\u2019Toole, M., Kim, C.: HyperReel: High-fidelity 6-DoF video with ray-conditioned sampling. In: CVPR (2023)","DOI":"10.1109\/CVPR52729.2023.01594"},{"key":"4229_CR44","doi-asserted-by":"crossref","unstructured":"Liu, M., Yang, Q., Huang, H., Huang, W., Yuan, Z., Li, Z., Xu, Y.: Light4gs: Lightweight compact 4d Gaussian splatting generation via context model. arXiv preprint arXiv:2503.13948 (2025)","DOI":"10.1109\/TCSVT.2026.3656896"},{"key":"4229_CR45","doi-asserted-by":"crossref","unstructured":"Bae, J., Kim, S., Yun, Y., Lee, H., Bang, G., Uh, Y.: Per-Gaussian embedding-based deformation for deformable 3d gaussian splatting. In: European Conference on Computer Vision, pp. 321\u2013335 (2024). Springer","DOI":"10.1007\/978-3-031-72633-0_18"}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-025-04229-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-025-04229-5","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-025-04229-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,4]],"date-time":"2026-03-04T13:01:01Z","timestamp":1772629261000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-025-04229-5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,12,27]]},"references-count":45,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2026,1]]}},"alternative-id":["4229"],"URL":"https:\/\/doi.org\/10.1007\/s00371-025-04229-5","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"value":"0178-2789","type":"print"},{"value":"1432-2315","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,12,27]]},"assertion":[{"value":"18 June 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"28 October 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"27 December 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"100"}}