{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,8]],"date-time":"2025-09-08T06:41:08Z","timestamp":1757313668823,"version":"3.37.3"},"reference-count":44,"publisher":"Springer Science and Business Media LLC","issue":"5","license":[{"start":{"date-parts":[[2022,4,4]],"date-time":"2022-04-04T00:00:00Z","timestamp":1649030400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,4,4]],"date-time":"2022-04-04T00:00:00Z","timestamp":1649030400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["6217071255"],"award-info":[{"award-number":["6217071255"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimedia Systems"],"published-print":{"date-parts":[[2022,10]]},"DOI":"10.1007\/s00530-022-00921-x","type":"journal-article","created":{"date-parts":[[2022,4,4]],"date-time":"2022-04-04T10:04:14Z","timestamp":1649066654000},"page":"1611-1620","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":3,"title":["Future pseudo-LiDAR frame prediction for autonomous driving"],"prefix":"10.1007","volume":"28","author":[{"given":"Xudong","family":"Huang","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2847-0349","authenticated-orcid":false,"given":"Chunyu","family":"Lin","sequence":"additional","affiliation":[]},{"given":"Haojie","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Lang","family":"Nie","sequence":"additional","affiliation":[]},{"given":"Yao","family":"Zhao","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,4,4]]},"reference":[{"key":"921_CR1","doi-asserted-by":"crossref","unstructured":"Schneider, L., Jasch, M., Fr\u00f6hlich, B., Weber, T., Franke, U., Pollefeys, M., R\u00e4tsch, M.: Multimodal neural networks: Rgb-d for semantic segmentation and object detection. In: Scandinavian Conference on Image Analysis, pp. 98\u2013109 (2017)","DOI":"10.1007\/978-3-319-59126-1_9"},{"key":"921_CR2","doi-asserted-by":"crossref","unstructured":"Zhao, J.-X., Cao, Y., Fan, D.-P., Cheng, M.-M., Li, X.-Y., Zhang, L.: Contrast prior and fluid pyramid integration for rgbd salient object detection. In: 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3922\u20133931 (2019)","DOI":"10.1109\/CVPR.2019.00405"},{"key":"921_CR3","doi-asserted-by":"crossref","unstructured":"Hu, X., Yang, K., Fei, L., Wang, K.: Acnet: Attention based network to exploit complementary features for rgbd semantic segmentation. In: 2019 IEEE International Conference on Image Processing (ICIP), pp. 1440\u20131444 (2019)","DOI":"10.1109\/ICIP.2019.8803025"},{"key":"921_CR4","doi-asserted-by":"crossref","unstructured":"Kundu, A., Yin, X., Fathi, A., Ross, D., Brewington, B., Funkhouser, T., Pantofaru, C.: Virtual multi-view fusion for 3d semantic segmentation. In: European Conference on Computer Vision, pp. 518\u2013535 (2020)","DOI":"10.1007\/978-3-030-58586-0_31"},{"key":"921_CR5","doi-asserted-by":"crossref","unstructured":"Ku, J., Mozifian, M., Lee, J., Harakeh, A., Waslander, S.L.: Joint 3d proposal generation and object detection from view aggregation. In: 2018 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 1\u20138 (2018)","DOI":"10.1109\/IROS.2018.8594049"},{"key":"921_CR6","doi-asserted-by":"crossref","unstructured":"Krispel, G., Opitz, M., Waltner, G., Possegger, H., Bischof, H.: Fuseseg: Lidar point cloud segmentation fusing multi-modal data. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 1874\u20131883 (2020)","DOI":"10.1109\/WACV45572.2020.9093584"},{"issue":"6","key":"921_CR7","doi-asserted-by":"publisher","first-page":"1573","DOI":"10.3390\/s20061573","volume":"20","author":"H Liu","year":"2020","unstructured":"Liu, H., Liao, K., Lin, C., Zhao, Y., Liu, M.: Plin: a network for pseudo-lidar point cloud interpolation. Sensors 20(6), 1573 (2020)","journal-title":"Sensors"},{"issue":"11","key":"921_CR8","doi-asserted-by":"publisher","first-page":"1231","DOI":"10.1177\/0278364913491297","volume":"32","author":"A Geiger","year":"2013","unstructured":"Geiger, A., Lenz, P., Stiller, C., Urtasun, R.: Vision meets robotics: the Kitti dataset. Int. J. Robot. Res. 32(11), 1231\u20131237 (2013)","journal-title":"Int. J. Robot. Res."},{"key":"921_CR9","doi-asserted-by":"crossref","unstructured":"Uhrig, J., Schneider, N., Schneider, L., Franke, U., Brox, T., Geiger, A.: Sparsity invariant cnns. In: 2017 International Conference on 3D Vision (3DV), pp. 11\u201320 (2017)","DOI":"10.1109\/3DV.2017.00012"},{"key":"921_CR10","doi-asserted-by":"crossref","unstructured":"Eldesokey, A., Felsberg, M., Holmquist, K., Persson, M.: Uncertainty-aware cnns for depth completion: Uncertainty from beginning to end. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12014\u201312023 (2020)","DOI":"10.1109\/CVPR42600.2020.01203"},{"key":"921_CR11","doi-asserted-by":"crossref","unstructured":"Liu, J., Gong, X.: Guided depth enhancement via anisotropic diffusion. In: Pacific-Rim Conference on Multimedia, pp. 408\u2013417 (2013)","DOI":"10.1007\/978-3-319-03731-8_38"},{"key":"921_CR12","doi-asserted-by":"crossref","unstructured":"Qiu, J., Cui, Z., Zhang, Y., Zhang, X., Liu, S., Zeng, B., Pollefeys, M.: Deeplidar: Deep surface normal guided depth prediction for outdoor scene from sparse lidar data and single color image. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3313\u20133322 (2019)","DOI":"10.1109\/CVPR.2019.00343"},{"issue":"12","key":"921_CR13","doi-asserted-by":"publisher","first-page":"5559","DOI":"10.1109\/TIP.2014.2361034","volume":"23","author":"J Park","year":"2014","unstructured":"Park, J., Kim, H., Tai, Y.-W., Brown, M.S., Kweon, I.S.: High-quality depth map upsampling and completion for rgb-d cameras. IEEE Trans. Image Process. 23(12), 5559\u20135572 (2014)","journal-title":"IEEE Trans. Image Process."},{"key":"921_CR14","doi-asserted-by":"crossref","unstructured":"Herrera, D., Kannala, J., Heikkil\u00e4, J., : Depth map inpainting under a second-order smoothness prior. In: Scandinavian Conference on Image Analysis, pp. 555\u2013566 (2013)","DOI":"10.1007\/978-3-642-38886-6_52"},{"key":"921_CR15","doi-asserted-by":"crossref","unstructured":"Ku, J., Harakeh, A., Waslander, S.L.: In defense of classical image processing: Fast depth completion on the cpu. In: 2018 15th Conference on Computer and Robot Vision (CRV), pp. 16\u201322 (2018)","DOI":"10.1109\/CRV.2018.00013"},{"key":"921_CR16","doi-asserted-by":"crossref","unstructured":"Van\u00a0Gansbeke, W., Neven, D., De\u00a0Brabandere, B., Van\u00a0Gool, L.: Sparse and noisy lidar completion with rgb guidance and uncertainty. In: 2019 16th International Conference on Machine Vision Applications (MVA), pp. 1\u20136 (2019)","DOI":"10.23919\/MVA.2019.8757939"},{"key":"921_CR17","doi-asserted-by":"crossref","unstructured":"Ma, F., Cavalheiro, G.V., Karaman, S.: Self-supervised sparse-to-dense: Self-supervised depth completion from lidar and monocular camera. In: 2019 International Conference on Robotics and Automation (ICRA), pp. 3288\u20133295 (2019)","DOI":"10.1109\/ICRA.2019.8793637"},{"key":"921_CR18","doi-asserted-by":"crossref","unstructured":"Xu, Y., Zhu, X., Shi, J., Zhang, G., Bao, H., Li, H.: Depth completion from sparse lidar data with depth-normal constraints. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 2811\u20132820 (2019)","DOI":"10.1109\/ICCV.2019.00290"},{"key":"921_CR19","doi-asserted-by":"crossref","unstructured":"Park, J., Joo, K., Hu, Z., Liu, C.-K., So\u00a0Kweon, I.: Non-local spatial propagation network for depth completion. In: Computer Vision\u2013ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part XIII 16, pp. 120\u2013136 (2020)","DOI":"10.1007\/978-3-030-58601-0_8"},{"key":"921_CR20","doi-asserted-by":"publisher","first-page":"1116","DOI":"10.1109\/TIP.2020.3040528","volume":"30","author":"J Tang","year":"2020","unstructured":"Tang, J., Tian, F.-P., Feng, W., Li, J., Tan, P.: Learning guided convolutional network for depth completion. IEEE Trans. Image Process. 30, 1116\u20131129 (2020)","journal-title":"IEEE Trans. Image Process."},{"key":"921_CR21","doi-asserted-by":"crossref","unstructured":"Meyer, S., Wang, O., Zimmer, H., Grosse, M., Sorkine-Hornung, A.: Phase-based frame interpolation for video. In: 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), vol. 2, pp. 1410\u20131418 (2015)","DOI":"10.1109\/CVPR.2015.7298747"},{"key":"921_CR22","doi-asserted-by":"crossref","unstructured":"Jiang, H., Sun, D., Jampani, V., Yang, M.-H., Learned-Miller, E., Kautz, J.: Super slomo: high quality estimation of multiple intermediate frames for video interpolation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 9000\u20139008 (2018)","DOI":"10.1109\/CVPR.2018.00938"},{"key":"921_CR23","doi-asserted-by":"crossref","unstructured":"Wang, Y., Chao, W.-L., Garg, D., Hariharan, B., Campbell, M., Weinberger, K.Q.: Pseudo-lidar from visual depth estimation: Bridging the gap in 3d object detection for autonomous driving. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8445\u20138453 (2019)","DOI":"10.1109\/CVPR.2019.00864"},{"key":"921_CR24","doi-asserted-by":"crossref","unstructured":"Lu, F., Chen, G., Qu, S., Li, Z., Liu, Y., Knoll, A.: Pointinet: Point cloud frame interpolation network. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 35, pp. 2251\u20132259 (2021)","DOI":"10.1609\/aaai.v35i3.16324"},{"key":"921_CR25","unstructured":"Lotter, W., Kreiman, G., Cox, D.: Deep predictive coding networks for video prediction and unsupervised learning. arXiv preprint arXiv:1605.08104 (2016)"},{"key":"921_CR26","doi-asserted-by":"crossref","unstructured":"Tulyakov, S., Liu, M.-Y., Yang, X., Kautz, J.: Mocogan: Decomposing motion and content for video generation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1526\u20131535 (2018)","DOI":"10.1109\/CVPR.2018.00165"},{"key":"921_CR27","first-page":"613","volume":"29","author":"C Vondrick","year":"2016","unstructured":"Vondrick, C., Pirsiavash, H., Torralba, A.: Generating videos with scene dynamics. Adv. Neural. Inf. Process. Syst. 29, 613\u2013621 (2016)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"921_CR28","doi-asserted-by":"crossref","unstructured":"Liang, X., Lee, L., Dai, W., Xing, E.P.: Dual motion gan for future-flow embedded video prediction. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1744\u20131752 (2017)","DOI":"10.1109\/ICCV.2017.194"},{"key":"921_CR29","doi-asserted-by":"crossref","unstructured":"Walker, J., Doersch, C., Gupta, A., Hebert, M.: An uncertain future: forecasting from static images using variational autoencoders. In: European Conference on Computer Vision, pp. 835\u2013851 (2016)","DOI":"10.1007\/978-3-319-46478-7_51"},{"key":"921_CR30","first-page":"64","volume":"29","author":"C Finn","year":"2016","unstructured":"Finn, C., Goodfellow, I., Levine, S.: Unsupervised learning for physical interaction through video prediction. Adv. Neural. Inf. Process. Syst. 29, 64\u201372 (2016)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"921_CR31","doi-asserted-by":"crossref","unstructured":"Hui, T.-W., Tang, X., Loy, C.C.: Liteflownet: A lightweight convolutional neural network for optical flow estimation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 8981\u20138989 (2018)","DOI":"10.1109\/CVPR.2018.00936"},{"key":"921_CR32","doi-asserted-by":"crossref","unstructured":"Luo, C., Zhan, J., Xue, X., Wang, L., Ren, R., Yang, Q.: Cosine normalization: Using cosine similarity instead of dot product in neural networks. In: International Conference on Artificial Neural Networks, pp. 382\u2013391 (2018)","DOI":"10.1007\/978-3-030-01418-6_38"},{"key":"921_CR33","doi-asserted-by":"crossref","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-net: Convolutional networks for biomedical image segmentation. In: International Conference on Medical Image Computing and Computer-assisted Intervention, pp. 234\u2013241 (2015)","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"921_CR34","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"921_CR35","doi-asserted-by":"publisher","first-page":"679","DOI":"10.1109\/TPAMI.1986.4767851","volume":"6","author":"J Canny","year":"1986","unstructured":"Canny, J.: A computational approach to edge detection. IEEE Trans. Pattern Anal. Mach. Intell. 6, 679\u2013698 (1986)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"921_CR36","doi-asserted-by":"crossref","unstructured":"Woo, S., Park, J., Lee, J.-Y., Kweon, I.S.: Cbam: Convolutional block attention module. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 3\u201319 (2018)","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"921_CR37","unstructured":"Kingma, D.P., Ba, J.: Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)"},{"key":"921_CR38","doi-asserted-by":"crossref","unstructured":"Hu, M., Wang, S., Li, B., Ning, S., Fan, L., Gong, X.: Penet: towards precise and efficient image guided depth completion. arXiv preprint arXiv:2103.00783 (2021)","DOI":"10.1109\/ICRA48506.2021.9561035"},{"key":"921_CR39","doi-asserted-by":"publisher","first-page":"5264","DOI":"10.1109\/TIP.2021.3079821","volume":"30","author":"S Zhao","year":"2021","unstructured":"Zhao, S., Gong, M., Fu, H., Tao, D.: Adaptive context-aware multi-modal network for depth completion. IEEE Trans. Image Process. 30, 5264\u20135276 (2021)","journal-title":"IEEE Trans. Image Process."},{"key":"921_CR40","doi-asserted-by":"crossref","unstructured":"Li, A., Yuan, Z., Ling, Y., Chi, W., Zhang, C., : A multi-scale guided cascade hourglass network for depth completion. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 32\u201340 (2020)","DOI":"10.1109\/WACV45572.2020.9093407"},{"key":"921_CR41","doi-asserted-by":"crossref","unstructured":"Deng, D., Zakhor, A.: Temporal lidar frame prediction for autonomous driving. In: 2020 International Conference on 3D Vision (3DV), pp. 829\u2013837 (2020)","DOI":"10.1109\/3DV50981.2020.00093"},{"key":"921_CR42","doi-asserted-by":"crossref","unstructured":"Fan, H., Su, H., Guibas, L.J.: A point set generation network for 3d object reconstruction from a single image. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 605\u2013613 (2017)","DOI":"10.1109\/CVPR.2017.264"},{"key":"921_CR43","unstructured":"Qi, C.R., Yi, L., Su, H., Guibas, L.J.: Pointnet++: deep hierarchical feature learning on point sets in a metric space. Adv. Neural Inf. Process. Syst. 30, 5099\u20135108 (2017)"},{"key":"921_CR44","doi-asserted-by":"crossref","unstructured":"Liu, H., Liao, K., Lin, C., Zhao, Y., Guo, Y.: Pseudo-lidar point cloud interpolation based on 3d motion representation and spatial supervision. IEEE Trans. Intell. Transport. Syst, 1\u201311 (2021)","DOI":"10.1109\/TITS.2021.3056048"}],"container-title":["Multimedia Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-022-00921-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00530-022-00921-x\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-022-00921-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,9,25]],"date-time":"2022-09-25T12:24:20Z","timestamp":1664108660000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00530-022-00921-x"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,4,4]]},"references-count":44,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2022,10]]}},"alternative-id":["921"],"URL":"https:\/\/doi.org\/10.1007\/s00530-022-00921-x","relation":{},"ISSN":["0942-4962","1432-1882"],"issn-type":[{"type":"print","value":"0942-4962"},{"type":"electronic","value":"1432-1882"}],"subject":[],"published":{"date-parts":[[2022,4,4]]},"assertion":[{"value":"4 September 2021","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"15 March 2022","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"4 April 2022","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}