{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,26]],"date-time":"2025-03-26T10:01:27Z","timestamp":1742983287866,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":33,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819756117"},{"type":"electronic","value":"9789819756124"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-981-97-5612-4_8","type":"book-chapter","created":{"date-parts":[[2024,7,30]],"date-time":"2024-07-30T13:03:51Z","timestamp":1722344631000},"page":"84-95","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Improving Dynamic 3D Gaussian Splatting from Monocular Videos with Object Motion Information"],"prefix":"10.1007","author":[{"given":"Yixin","family":"Luo","sequence":"first","affiliation":[]},{"given":"Zhangjin","family":"Huang","sequence":"additional","affiliation":[]},{"given":"Xudong","family":"Huang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,7,31]]},"reference":[{"key":"8_CR1","doi-asserted-by":"crossref","unstructured":"Agarwal, A., Arora, C.: Attention attention everywhere: monocular depth prediction with skip attention. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 5861\u20135870 (2023)","DOI":"10.1109\/WACV56688.2023.00581"},{"key":"8_CR2","unstructured":"Bhat, S.F., Alhashim, I., Wonka, P.: AdaBins: depth estimation using adaptive bins. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4009\u20134018 (2021)"},{"key":"8_CR3","doi-asserted-by":"publisher","unstructured":"Bhat, S.F., Alhashim, I., Wonka, P.: LocalBins: improving depth estimation by learning local distributions. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) European Conference on Computer Vision, pp. 480\u2013496. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19769-7_28","DOI":"10.1007\/978-3-031-19769-7_28"},{"key":"8_CR4","unstructured":"Bhat, S.F., Birkl, R., Wofk, D., Wonka, P., M\u00fcller, M.: ZoeDepth: zero-shot transfer by combining relative and metric depth. arXiv preprint arXiv:2302.12288 (2023)"},{"key":"8_CR5","doi-asserted-by":"crossref","unstructured":"Cao, A., Johnson, J.: HexPlane: a fast representation for dynamic scenes. In: CVPR (2023)","DOI":"10.1109\/CVPR52729.2023.00021"},{"key":"8_CR6","doi-asserted-by":"publisher","unstructured":"Chen, A., Xu, Z., Geiger, A., Yu, J., Su, H.: TensoRF: tensorial radiance fields. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) European Conference on Computer Vision (ECCV), vol. 13692. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19824-3_20","DOI":"10.1007\/978-3-031-19824-3_20"},{"key":"8_CR7","doi-asserted-by":"crossref","unstructured":"Fang, J., et al.: Fast dynamic radiance fields with time-aware neural voxels. In: SIGGRAPH Asia 2022 Conference Papers (2022)","DOI":"10.1145\/3550469.3555383"},{"key":"8_CR8","unstructured":"Gao, H., Li, R., Tulsiani, S., Russell, B., Kanazawa, A.: Monocular dynamic view synthesis: a reality check. In: Advances in Neural Information Processing Systems (2022)"},{"key":"8_CR9","unstructured":"Katsumata, K., Vo, D.M., Nakayama, H.: An efficient 3D Gaussian representation for monocular\/multi-view dynamic scenes. arXiv preprint arXiv:2311.12897 (2023)"},{"key":"8_CR10","doi-asserted-by":"crossref","unstructured":"Kerbl, B., Kopanas, G., Leimk\u00fchler, T., Drettakis, G.: 3D Gaussian splatting for real-time radiance field rendering. ACM Trans. Graph. 42(4), 1\u201314 (2023). https:\/\/repo-sam.inria.fr\/fungraph\/3d-gaussian-splatting\/","DOI":"10.1145\/3592433"},{"key":"8_CR11","doi-asserted-by":"crossref","unstructured":"Kirillov, A., et al.: Segment anything. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 4015\u20134026 (2023)","DOI":"10.1109\/ICCV51070.2023.00371"},{"key":"8_CR12","doi-asserted-by":"crossref","unstructured":"Kopf, J., Rong, X., Huang, J.B.: Robust consistent video depth estimation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 1611\u20131621 (2021)","DOI":"10.1109\/CVPR46437.2021.00166"},{"key":"8_CR13","doi-asserted-by":"crossref","unstructured":"Li, Z., Niklaus, S., Snavely, N., Wang, O.: Neural scene flow fields for space-time view synthesis of dynamic scenes. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 6498\u20136508 (2021)","DOI":"10.1109\/CVPR46437.2021.00643"},{"key":"8_CR14","doi-asserted-by":"crossref","unstructured":"Li, Z., Wang, Q., Cole, F., Tucker, R., Snavely, N.: DynIBaR: neural dynamic image-based rendering. arXiv preprint arXiv:2211.11082 (2022)","DOI":"10.1109\/CVPR52729.2023.00416"},{"key":"8_CR15","unstructured":"Li, Z., Wang, X., Liu, X., Jiang, J.: BinsFormer: revisiting adaptive bins for monocular depth estimation. arXiv preprint arXiv:2204.00987 (2022)"},{"key":"8_CR16","doi-asserted-by":"crossref","unstructured":"Luiten, J., Kopanas, G., Leibe, B., Ramanan, D.: Dynamic 3D Gaussians: tracking by persistent dynamic view synthesis. Preprint (2023)","DOI":"10.1109\/3DV62453.2024.00044"},{"issue":"4","key":"8_CR17","doi-asserted-by":"publisher","first-page":"71","DOI":"10.1145\/3386569.3392377","volume":"39","author":"X Luo","year":"2020","unstructured":"Luo, X., Huang, J.B., Szeliski, R., Matzen, K., Kopf, J.: Consistent video depth estimation. ACM Trans. Graph. (ToG) 39(4), 71\u201381 (2020)","journal-title":"ACM Trans. Graph. (ToG)"},{"issue":"1","key":"8_CR18","doi-asserted-by":"publisher","first-page":"99","DOI":"10.1145\/3503250","volume":"65","author":"B Mildenhall","year":"2021","unstructured":"Mildenhall, B., Srinivasan, P.P., Tancik, M., Barron, J.T., Ramamoorthi, R., Ng, R.: NeRF: representing scenes as neural radiance fields for view synthesis. Commun. ACM 65(1), 99\u2013106 (2021)","journal-title":"Commun. ACM"},{"issue":"4","key":"8_CR19","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3528223.3530127","volume":"41","author":"T M\u00fcller","year":"2022","unstructured":"M\u00fcller, T., Evans, A., Schied, C., Keller, A.: Instant neural graphics primitives with a multiresolution hash encoding. ACM Trans. Graph. (TOG) 41(4), 1\u201315 (2022)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"8_CR20","doi-asserted-by":"crossref","unstructured":"Park, K., et al.: Nerfies: deformable neural radiance fields. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 5865\u20135874 (2021)","DOI":"10.1109\/ICCV48922.2021.00581"},{"key":"8_CR21","doi-asserted-by":"crossref","unstructured":"Park, K., et al.: HyperNeRF: a higher-dimensional representation for topo logically varying neural radiance fields. ACM Trans. Graph. 40(6), 1\u201312 (2021)","DOI":"10.1145\/3478513.3480487"},{"key":"8_CR22","doi-asserted-by":"crossref","unstructured":"Pumarola, A., Corona, E., Pons-Moll, G., Moreno-Noguer, F.: D-NeRF: neural radiance fields for dynamic scenes. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10318\u201310327 (2021)","DOI":"10.1109\/CVPR46437.2021.01018"},{"issue":"3","key":"8_CR23","doi-asserted-by":"publisher","first-page":"1623","DOI":"10.1109\/TPAMI.2020.3019967","volume":"44","author":"R Ranftl","year":"2020","unstructured":"Ranftl, R., Lasinger, K., Hafner, D., Schindler, K., Koltun, V.: Towards robust monocular depth estimation: mixing datasets for zero-shot cross-dataset transfer. IEEE Trans. Pattern Anal. Mach. Intell. 44(3), 1623\u20131637 (2020)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"8_CR24","doi-asserted-by":"crossref","unstructured":"Fridovich-Keil, S., Meanti, G., Warburg, F.R., Recht, B., Kanazawa, A.: K-planes: explicit radiance fields in space, time, and appearance. In: CVPR (2023)","DOI":"10.1109\/CVPR52729.2023.01201"},{"key":"8_CR25","doi-asserted-by":"crossref","unstructured":"Shao, R., Zheng, Z., Tu, H., Liu, B., Zhang, H., Liu, Y.: Tensor4D: efficient neural 4D decomposition for high-fidelity dynamic reconstruction and rendering. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2023)","DOI":"10.1109\/CVPR52729.2023.01596"},{"key":"8_CR26","doi-asserted-by":"crossref","unstructured":"Sun, C., Sun, M., Chen, H.T.: Direct voxel grid optimization: super-fast convergence for radiance fields reconstruction. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5459\u20135469 (2022)","DOI":"10.1109\/CVPR52688.2022.00538"},{"key":"8_CR27","doi-asserted-by":"crossref","unstructured":"Wu, G., et al.: 4D Gaussian splatting for real-time dynamic scene rendering. arXiv preprint arXiv:2310.08528 (2023)","DOI":"10.1109\/CVPR52733.2024.01920"},{"key":"8_CR28","doi-asserted-by":"crossref","unstructured":"Xie, T., et al.: PhysGaussian: physics-integrated 3D Gaussians for generative dynamics. arXiv preprint arXiv:2311.12198 (2023)","DOI":"10.1109\/CVPR52733.2024.00420"},{"key":"8_CR29","unstructured":"Yang, J., Gao, M., Li, Z., Gao, S., Wang, F., Zheng, F.: Track anything: segment anything meets videos. arXiv preprint arXiv:2304.11968 (2023)"},{"key":"8_CR30","doi-asserted-by":"crossref","unstructured":"Yang, Z., Gao, X., Zhou, W., Jiao, S., Zhang, Y., Jin, X.: Deformable 3D Gaussians for high-fidelity monocular dynamic scene reconstruction. arXiv preprint arXiv:2309.13101 (2023)","DOI":"10.1109\/CVPR52733.2024.01922"},{"key":"8_CR31","doi-asserted-by":"crossref","unstructured":"Yin, W., et al.: Metric3D: towards zero-shot metric 3D prediction from a single image. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 9043\u20139053 (2023)","DOI":"10.1109\/ICCV51070.2023.00830"},{"issue":"4","key":"8_CR32","first-page":"1","volume":"40","author":"Z Zhang","year":"2021","unstructured":"Zhang, Z., Cole, F., Tucker, R., Freeman, W.T., Dekel, T.: Consistent depth of moving objects in video. ACM Trans. Graph. (TOG) 40(4), 1\u201312 (2021)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"8_CR33","unstructured":"Zhou, K., et al.: DynPoint: dynamic neural point for view synthesis. In: Advances in Neural Information Processing Systems, vol. 36 (2024)"}],"container-title":["Lecture Notes in Computer Science","Advanced Intelligent Computing Technology and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-97-5612-4_8","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,25]],"date-time":"2024-11-25T02:46:41Z","timestamp":1732502801000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-97-5612-4_8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9789819756117","9789819756124"],"references-count":33,"URL":"https:\/\/doi.org\/10.1007\/978-981-97-5612-4_8","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"31 July 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICIC","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Intelligent Computing","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tianjin","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"5 August 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8 August 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"20","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icic2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/www.ic-icc.cn\/2024\/index.htm","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}