{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,25]],"date-time":"2025-10-25T19:05:45Z","timestamp":1761419145430,"version":"build-2065373602"},"reference-count":31,"publisher":"Springer Science and Business Media LLC","issue":"14","license":[{"start":{"date-parts":[[2025,9,23]],"date-time":"2025-09-23T00:00:00Z","timestamp":1758585600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,9,23]],"date-time":"2025-09-23T00:00:00Z","timestamp":1758585600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"The National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62272014","62272014","62272014","62272014","62272014"],"award-info":[{"award-number":["62272014","62272014","62272014","62272014","62272014"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["SIViP"],"published-print":{"date-parts":[[2025,12]]},"DOI":"10.1007\/s11760-025-04821-1","type":"journal-article","created":{"date-parts":[[2025,9,23]],"date-time":"2025-09-23T13:17:23Z","timestamp":1758633443000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Spatiotemporal-Modal Collaborative Calibration for Dynamic Depth Video Enhancement"],"prefix":"10.1007","volume":"19","author":[{"given":"Xiaoling","family":"Xu","sequence":"first","affiliation":[]},{"given":"Sheng","family":"Zhao","sequence":"additional","affiliation":[]},{"given":"Jian","family":"Cao","sequence":"additional","affiliation":[]},{"given":"Xiaoqun","family":"Wu","sequence":"additional","affiliation":[]},{"given":"Huilin","family":"Si","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,9,23]]},"reference":[{"key":"4821_CR1","doi-asserted-by":"crossref","unstructured":"Conti, A., Poggi, M., Cambareri, V., Mattoccia, S.: Depth on demand: Streaming dense depth from a low frame rate active sensor. In: European Conference on Computer Vision, pp. 283\u2013302 (2024). Springer","DOI":"10.1007\/978-3-031-73030-6_16"},{"key":"4821_CR2","doi-asserted-by":"publisher","first-page":"227825","DOI":"10.1109\/ACCESS.2020.3045681","volume":"8","author":"L Bai","year":"2020","unstructured":"Bai, L., Zhao, Y., Elhousni, M., Huang, X.: Depthnet: real-time lidar point cloud depth completion for autonomous vehicles. IEEE access 8, 227825\u2013227833 (2020)","journal-title":"IEEE access"},{"key":"4821_CR3","doi-asserted-by":"crossref","unstructured":"Wang, H., Wang, M., Che, Z., Xu, Z., Qiao, X., Qi, M., Feng, F., Tang, J.: Rgb-depth fusion gan for indoor depth completion. In: Proceedings of the Ieee\/cvf Conference on Computer Vision and Pattern Recognition, pp. 6209\u20136218 (2022)","DOI":"10.1109\/CVPR52688.2022.00611"},{"key":"4821_CR4","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Funkhouser, T.: Deep depth completion of a single rgb-d image. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 175\u2013185 (2018)","DOI":"10.1109\/CVPR.2018.00026"},{"key":"4821_CR5","unstructured":"Xiang, R., Zheng, F., Su, H., Zhang, Z.: 3ddepthnet: Point cloud guided depth completion network for sparse depth and single color image. arXiv preprint arXiv:2003.09175 (2020)"},{"key":"4821_CR6","doi-asserted-by":"crossref","unstructured":"Choi, K., Jeong, S., Kim, Y., Sohn, K.: Stereo-augmented depth completion from a single rgb-lidar image. In: 2021 IEEE International Conference on Robotics and Automation (ICRA), pp. 13641\u201313647 (2021). IEEE","DOI":"10.1109\/ICRA48506.2021.9561557"},{"key":"4821_CR7","doi-asserted-by":"crossref","unstructured":"Hu, M., Wang, S., Li, B., Ning, S., Fan, L., Gong, X.: Penet: Towards precise and efficient image guided depth completion. In: 2021 IEEE International Conference on Robotics and Automation (ICRA), pp. 13656\u201313662 (2021). IEEE","DOI":"10.1109\/ICRA48506.2021.9561035"},{"key":"4821_CR8","doi-asserted-by":"crossref","unstructured":"Kim, J., Kim, S., Park, J., Lee, S.: Deep cost ray fusion for sparse depth video completion. In: European Conference on Computer Vision, pp. 328\u2013346 (2024). Springer","DOI":"10.1007\/978-3-031-73347-5_19"},{"issue":"4","key":"4821_CR9","doi-asserted-by":"publisher","first-page":"124","DOI":"10.1007\/s11554-024-01491-z","volume":"21","author":"A Duarte","year":"2024","unstructured":"Duarte, A., Fernandes, F., Pereira, J.M., Moreira, C., Nascimento, J.C., Jorge, J.: Selfredepth: self-supervised real-time depth restoration for consumer-grade sensors. J. Real-Time Image Proc. 21(4), 124 (2024)","journal-title":"J. Real-Time Image Proc."},{"key":"4821_CR10","unstructured":"Shi, X., Chen, Z., Wang, H., Yeung, D.-Y., Wong, W.-K., Woo, W.-c.: Convolutional lstm network: A machine learning approach for precipitation nowcasting. Advances in neural information processing systems 28 (2015)"},{"key":"4821_CR11","doi-asserted-by":"crossref","unstructured":"Lu, K., Barnes, N., Anwar, S., Zheng, L.: From depth what can you see? depth completion via auxiliary image reconstruction. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11306\u201311315 (2020)","DOI":"10.1109\/CVPR42600.2020.01132"},{"key":"4821_CR12","doi-asserted-by":"crossref","unstructured":"Chodosh, N., Wang, C., Lucey, S.: Deep convolutional compressed sensing for lidar depth completion. In: Asian Conference on Computer Vision, pp. 499\u2013513 (2018). Springer","DOI":"10.1007\/978-3-030-20887-5_31"},{"key":"4821_CR13","doi-asserted-by":"publisher","first-page":"3429","DOI":"10.1109\/TIP.2019.2960589","volume":"29","author":"Z Huang","year":"2019","unstructured":"Huang, Z., Fan, J., Cheng, S., Yi, S., Wang, X., Li, H.: Hms-net: hierarchical multi-scale sparsity-invariant network for sparse depth completion. IEEE Trans. Image Process. 29, 3429\u20133441 (2019)","journal-title":"IEEE Trans. Image Process."},{"key":"4821_CR14","doi-asserted-by":"crossref","unstructured":"Eldesokey, A., Felsberg, M., Holmquist, K., Persson, M.: Uncertainty-aware cnns for depth completion: Uncertainty from beginning to end. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12014\u201312023 (2020)","DOI":"10.1109\/CVPR42600.2020.01203"},{"key":"4821_CR15","doi-asserted-by":"crossref","unstructured":"Ma, F., Karaman, S.: Sparse-to-dense: Depth prediction from sparse depth samples and a single image. In: 2018 IEEE International Conference on Robotics and Automation (ICRA), pp. 4796\u20134803 (2018). IEEE","DOI":"10.1109\/ICRA.2018.8460184"},{"key":"4821_CR16","doi-asserted-by":"crossref","unstructured":"Xu, Y., Zhu, X., Shi, J., Zhang, G., Bao, H., Li, H.: Depth completion from sparse lidar data with depth-normal constraints. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 2811\u20132820 (2019)","DOI":"10.1109\/ICCV.2019.00290"},{"key":"4821_CR17","doi-asserted-by":"crossref","unstructured":"Wang, H., Wang, M., Che, Z., Xu, Z., Qiao, X., Qi, M., Feng, F., Tang, J.: Rgb-depth fusion gan for indoor depth completion. In: Proceedings of the Ieee\/cvf Conference on Computer Vision and Pattern Recognition, pp. 6209\u20136218 (2022)","DOI":"10.1109\/CVPR52688.2022.00611"},{"issue":"17","key":"4821_CR18","doi-asserted-by":"publisher","first-page":"6414","DOI":"10.3390\/s22176414","volume":"22","author":"B Chen","year":"2022","unstructured":"Chen, B., Lv, X., Liu, C., Jiao, H.: Sgsnet: a lightweight depth completion network based on secondary guidance and spatial fusion. Sensors 22(17), 6414 (2022)","journal-title":"Sensors"},{"key":"4821_CR19","doi-asserted-by":"crossref","unstructured":"Sun, X., Ponce, J., Wang, Y.-X.: Revisiting deformable convolution for depth completion. in 2023 ieee. In: RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 1300\u20131306","DOI":"10.1109\/IROS55552.2023.10342026"},{"key":"4821_CR20","doi-asserted-by":"crossref","unstructured":"Wang, Y., Li, B., Zhang, G., Liu, Q., Gao, T., Dai, Y.: Lrru: Long-short range recurrent updating networks for depth completion. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 9422\u20139432 (2023)","DOI":"10.1109\/ICCV51070.2023.00864"},{"issue":"2","key":"4821_CR21","doi-asserted-by":"publisher","first-page":"1186","DOI":"10.1109\/TCSVT.2023.3292398","volume":"34","author":"Y Wang","year":"2023","unstructured":"Wang, Y., Mao, Y., Liu, Q., Dai, Y.: Decomposed guided dynamic filters for efficient rgb-guided depth completion. IEEE Trans. Circuits Syst. Video Technol. 34(2), 1186\u20131198 (2023)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"4821_CR22","doi-asserted-by":"crossref","unstructured":"Yan, Z., Wang, Z., Wang, K., Li, J., Yang, J.: Completion as enhancement: A degradation-aware selective image guided network for depth completion. In: Proceedings of the Computer Vision and Pattern Recognition Conference, pp. 26943\u201326953 (2025)","DOI":"10.1109\/CVPR52734.2025.02509"},{"key":"4821_CR23","doi-asserted-by":"crossref","unstructured":"Shi, Y., Singh, M.K., Cai, H., Porikli, F.: Decotr: Enhancing depth completion with 2d and 3d attentions. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10736\u201310746 (2024)","DOI":"10.1109\/CVPR52733.2024.01021"},{"key":"4821_CR24","doi-asserted-by":"crossref","unstructured":"Jia, X., Jian, S., Tan, Y., Che, Y., Chen, W., Liang, Z.: Gated cross-attention network for depth completion. In: ICASSP 2025-2025 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 1\u20135 (2025). IEEE","DOI":"10.1109\/ICASSP49660.2025.10889559"},{"key":"4821_CR25","doi-asserted-by":"crossref","unstructured":"Wu, K., Lin, J., Miao, J., Li, Z., Zhang, X., Xing, G., Fan, Y., Luo, J., Zhao, H., Liu, Y., et al.: Difnet: Dual-information fusion network for depth completion. Information Fusion, 103424 (2025)","DOI":"10.1016\/j.inffus.2025.103424"},{"issue":"11","key":"4821_CR26","doi-asserted-by":"publisher","first-page":"8279","DOI":"10.1007\/s11760-024-03469-7","volume":"18","author":"C Yang","year":"2024","unstructured":"Yang, C., Kong, G., Duan, X., Long, H., Zhao, J.: Space-time video super-resolution via multi-scale feature interpolation and temporal feature fusion. SIViP 18(11), 8279\u20138291 (2024)","journal-title":"SIViP"},{"issue":"1","key":"4821_CR27","doi-asserted-by":"publisher","first-page":"5","DOI":"10.1007\/s11760-024-03596-1","volume":"19","author":"C Ji","year":"2025","unstructured":"Ji, C., Xu, Y., Ke, W., Tang, L., Yan, C., Zhang, Y.: Predat-gan: a new spatiotemporal forecast model for precipitation nowcasting with weather radar echo images. SIViP 19(1), 5 (2025)","journal-title":"SIViP"},{"key":"4821_CR28","unstructured":"Qi, C.R., Su, H., Mo, K., Guibas, L.J.: Pointnet: Deep learning on point sets for 3d classification and segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 652\u2013660 (2017)"},{"key":"4821_CR29","doi-asserted-by":"crossref","unstructured":"Ku, J., Harakeh, A., Waslander, S.L.: In defense of classical image processing: Fast depth completion on the cpu. In: 2018 15th Conference on Computer and Robot Vision (CRV), pp. 16\u201322 (2018). IEEE","DOI":"10.1109\/CRV.2018.00013"},{"key":"4821_CR30","doi-asserted-by":"crossref","unstructured":"Khan, M.F.F., Troncoso\u00a0Aldas, N.D., Kumar, A., Advani, S., Narayanan, V.: Sparse to dense depth completion using a generative adversarial network with intelligent sampling strategies. In: Proceedings of the 29th Acm International Conference on Multimedia, pp. 5528\u20135536 (2021)","DOI":"10.1145\/3474085.3475688"},{"issue":"6","key":"4821_CR31","doi-asserted-by":"publisher","first-page":"3270","DOI":"10.1109\/LRA.2023.3266670","volume":"8","author":"\u00d6 Zov\u00e1thi","year":"2023","unstructured":"Zov\u00e1thi, \u00d6., P\u00e1lffy, B., Jank\u00f3, Z., Benedek, C.: St-depthnet: a spatio-temporal deep network for depth completion using a single non-repetitive circular scanning lidar. IEEE Robotics and Automation Letters 8(6), 3270\u20133277 (2023)","journal-title":"IEEE Robotics and Automation Letters"}],"container-title":["Signal, Image and Video Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-025-04821-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11760-025-04821-1\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-025-04821-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,25]],"date-time":"2025-10-25T18:58:27Z","timestamp":1761418707000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11760-025-04821-1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,9,23]]},"references-count":31,"journal-issue":{"issue":"14","published-print":{"date-parts":[[2025,12]]}},"alternative-id":["4821"],"URL":"https:\/\/doi.org\/10.1007\/s11760-025-04821-1","relation":{},"ISSN":["1863-1703","1863-1711"],"issn-type":[{"type":"print","value":"1863-1703"},{"type":"electronic","value":"1863-1711"}],"subject":[],"published":{"date-parts":[[2025,9,23]]},"assertion":[{"value":"25 July 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"2 September 2025","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"17 September 2025","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"23 September 2025","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"The authors declare no competing interests.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"1216"}}