{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,19]],"date-time":"2026-03-19T00:09:50Z","timestamp":1773878990644,"version":"3.50.1"},"reference-count":57,"publisher":"Springer Science and Business Media LLC","issue":"6","license":[{"start":{"date-parts":[[2024,4,22]],"date-time":"2024-04-22T00:00:00Z","timestamp":1713744000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,4,22]],"date-time":"2024-04-22T00:00:00Z","timestamp":1713744000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62241601"],"award-info":[{"award-number":["62241601"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2024,6]]},"DOI":"10.1007\/s00371-024-03396-1","type":"journal-article","created":{"date-parts":[[2024,4,22]],"date-time":"2024-04-22T19:02:01Z","timestamp":1713812521000},"page":"3903-3917","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":12,"title":["Latent diffusion transformer for point cloud generation"],"prefix":"10.1007","volume":"40","author":[{"given":"Junzhong","family":"Ji","sequence":"first","affiliation":[]},{"given":"Runfeng","family":"Zhao","sequence":"additional","affiliation":[]},{"given":"Minglong","family":"Lei","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,4,22]]},"reference":[{"key":"3396_CR1","first-page":"37349","volume":"35","author":"E Aiello","year":"2022","unstructured":"Aiello, E., Valsesia, D., Magli, E.: Cross-modal learning for image-guided point cloud shape completion. Adv. Neural Inf. Process. Syst. 35, 37349\u201337362 (2022)","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"3396_CR2","doi-asserted-by":"publisher","unstructured":"Cai, R., Yang, G., Averbuch-Elor. H., et\u00a0al.: Learning gradient fields for shape generation. In: European Conference on Computer Vision, pp. 364\u2013381 (2020). https:\/\/doi.org\/10.1007\/978-3-030-58580-8_22","DOI":"10.1007\/978-3-030-58580-8_22"},{"key":"3396_CR3","doi-asserted-by":"publisher","unstructured":"Chai, S., Zhuang, L., Yan, F.: Layoutdm: transformer-based diffusion model for layout generation, pp. 18349\u201318358 (2023). https:\/\/doi.org\/10.1109\/CVPR52729.2023.01760","DOI":"10.1109\/CVPR52729.2023.01760"},{"key":"3396_CR4","unstructured":"Chang, A.X., Funkhouser, T., Guibas, L. et\u00a0al.: Shapenet: an information-rich 3d model repository (2015). arXiv preprint arXiv:1512.03012"},{"key":"3396_CR5","doi-asserted-by":"publisher","unstructured":"Chang, H., Zhang, H., Jiang, L. et\u00a0al.: Maskgit: masked generative image transformer, pp. 11305\u201311315 (2022) https:\/\/doi.org\/10.1109\/CVPR52688.2022.01103","DOI":"10.1109\/CVPR52688.2022.01103"},{"key":"3396_CR6","doi-asserted-by":"publisher","unstructured":"Chen, R., Han, S., Xu, J. et\u00a0al.: Point-based multi-view stereo network. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 1538\u20131547 (2019). https:\/\/doi.org\/10.1109\/ICCV.2019.00162","DOI":"10.1109\/ICCV.2019.00162"},{"issue":"9\u201311","key":"3396_CR7","doi-asserted-by":"publisher","first-page":"2657","DOI":"10.1007\/S00371-021-02199-Y","volume":"37","author":"Z Chen","year":"2021","unstructured":"Chen, Z., Qiu, J., Sheng, B., et al.: Gpsd: generative parking spot detection using multi-clue recovery model. Vis. Comput. 37(9\u201311), 2657\u20132669 (2021). https:\/\/doi.org\/10.1007\/S00371-021-02199-Y","journal-title":"Vis. Comput."},{"issue":"11","key":"3396_CR8","doi-asserted-by":"publisher","first-page":"13489","DOI":"10.1109\/TPAMI.2023.3293885","volume":"45","author":"Z Chen","year":"2023","unstructured":"Chen, Z., Qiu, G., Li, P., et al.: Mngnas: distilling adaptive combination of multiple searched networks for one-shot neural architecture search. IEEE Trans. Pattern Anal. Mach. Intell. 45(11), 13489\u201313508 (2023). https:\/\/doi.org\/10.1109\/TPAMI.2023.3293885","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"3396_CR9","doi-asserted-by":"publisher","unstructured":"Cheng, A.C., Li, X., Liu, S., et\u00a0al.: Autoregressive 3d shape generation via canonical mapping. In: European Conference on Computer Vision, pp. 89\u2013104 (2022). https:\/\/doi.org\/10.1007\/978-3-031-20062-5_6","DOI":"10.1007\/978-3-031-20062-5_6"},{"key":"3396_CR10","doi-asserted-by":"publisher","unstructured":"Cho, J., Zala, A., Bansal, M.: Dall-eval: probing the reasoning skills and social biases of text-to-image generation models. pp 3020\u20133031 (2023). https:\/\/doi.org\/10.1109\/ICCV51070.2023.00283","DOI":"10.1109\/ICCV51070.2023.00283"},{"key":"3396_CR11","first-page":"8780","volume":"34","author":"P Dhariwal","year":"2021","unstructured":"Dhariwal, P., Nichol, A.: Diffusion models beat gans on image synthesis. Adv. Neural Inf. Process. Syst. 34, 8780\u20138794 (2021)","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"3396_CR12","unstructured":"Dinh, L., Krueger, D., Bengio, Y.: Nice: Non-linear independent components estimation. In: International Conference on Learning Representations (2015)"},{"key":"3396_CR13","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., et\u00a0al.: An image is worth 16x16 words: transformers for image recognition at scale. In: International Conference on Learning Representations (2021). https:\/\/openreview.net\/forum?id=YicbFdNTTy"},{"key":"3396_CR14","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., et\u00a0al.: An image is worth 16x16 words: transformers for image recognition at scale. In: International Conference on Learning Representations (2021)"},{"key":"3396_CR15","doi-asserted-by":"publisher","unstructured":"Groueix, T., Fisher, M., Kim, V.G., et\u00a0al.: Atlasnet: a papier-mache approach to learning 3d surface generation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 216\u2013224 (2018) https:\/\/doi.org\/10.1109\/CVPR.2018.00030","DOI":"10.1109\/CVPR.2018.00030"},{"key":"3396_CR16","first-page":"27953","volume":"35","author":"W Harvey","year":"2022","unstructured":"Harvey, W., Naderiparizi, S., Masrani, V., et al.: Flexible diffusion modeling of long videos. Adv. Neural Inf. Process. Syst. 35, 27953\u201327965 (2022)","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"3396_CR17","first-page":"6840","volume":"33","author":"J Ho","year":"2020","unstructured":"Ho, J., Jain, A., Abbeel, P.: Denoising diffusion probabilistic models. Adv. Neural Inf. Process. Syst. 33, 6840\u20136851 (2020)","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"3396_CR18","first-page":"2249","volume":"23","author":"J Ho","year":"2022","unstructured":"Ho, J., Saharia, C., Chan, W., et al.: Cascaded diffusion models for high fidelity image generation. J. Mach. Learn. Res. 23, 2249\u20132281 (2022)","journal-title":"J. Mach. Learn. Res."},{"key":"3396_CR19","doi-asserted-by":"publisher","unstructured":"Huang, R., Lam, M.W., Wang, J. et\u00a0al.: Fastdiff: a fast conditional diffusion model for high-quality speech synthesis. In: Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence, pp. 4157\u20134163 (2022). https:\/\/doi.org\/10.24963\/ijcai.2022\/577","DOI":"10.24963\/ijcai.2022\/577"},{"key":"3396_CR20","doi-asserted-by":"publisher","unstructured":"Huang, R., Zhao, Z., Liu, H. et\u00a0al.: Prodiff: Progressive fast diffusion model for high-quality text-to-speech. In: Proceedings of the 30th ACM International Conference on Multimedia, pp. 2595\u20132605, (2022). https:\/\/doi.org\/10.1145\/3503161.3547855","DOI":"10.1145\/3503161.3547855"},{"key":"3396_CR21","doi-asserted-by":"publisher","first-page":"2226","DOI":"10.1109\/TMM.2022.3144890","volume":"25","author":"N Jiang","year":"2023","unstructured":"Jiang, N., Sheng, B., Li, P., et al.: Photohelper: portrait photographing guidance via deep feature retrieval and fusion. IEEE Trans. Multimed. 25, 2226\u20132238 (2023). https:\/\/doi.org\/10.1109\/TMM.2022.3144890","journal-title":"IEEE Trans. Multimed."},{"key":"3396_CR22","first-page":"16388","volume":"33","author":"H Kim","year":"2020","unstructured":"Kim, H., Lee, H., Kang, W.H., et al.: Softflow: probabilistic framework for normalizing flow on manifolds. Adv. Neural Inf. Process. Syst. 33, 16388\u201316397 (2020)","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"3396_CR23","doi-asserted-by":"publisher","unstructured":"Kim, J., Yoo, J., Lee, J. et\u00a0al.: Setvae: learning hierarchical composition for generative modeling of set-structured data. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 15059\u201315068 (2021). https:\/\/doi.org\/10.1109\/CVPR46437.2021.01481","DOI":"10.1109\/CVPR46437.2021.01481"},{"key":"3396_CR24","doi-asserted-by":"publisher","unstructured":"Klokov, R., Boyer, E., Verbeek, J.: Discrete point flow networks for efficient point cloud generation. In: European Conference on Computer Vision, pp. 694\u2013710 (2020). https:\/\/doi.org\/10.1007\/978-3-030-58592-1_41","DOI":"10.1007\/978-3-030-58592-1_41"},{"key":"3396_CR25","doi-asserted-by":"publisher","unstructured":"Lai, X., Liu, J., Jiang, L. et\u00a0al.: Stratified transformer for 3d point cloud segmentation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8500\u20138509 (2022). https:\/\/doi.org\/10.1109\/CVPR52688.2022.00831","DOI":"10.1109\/CVPR52688.2022.00831"},{"key":"3396_CR26","unstructured":"Lee, J., Lee, Y., Kim, J. et\u00a0al.: Set transformer: a framework for attention-based permutation-invariant neural networks. In: International Conference on Machine Learning, pp. 3744\u20133753 (2019)"},{"issue":"1","key":"3396_CR27","doi-asserted-by":"publisher","first-page":"163","DOI":"10.1109\/TII.2021.3085669","volume":"18","author":"J Li","year":"2022","unstructured":"Li, J., Chen, J., Sheng, B., et al.: Automatic detection and classification system of domestic waste via multimodel cascaded convolutional neural network. IEEE Trans. Ind. Inf. 18(1), 163\u2013173 (2022). https:\/\/doi.org\/10.1109\/TII.2021.3085669","journal-title":"IEEE Trans. Ind. Inf."},{"key":"3396_CR28","doi-asserted-by":"publisher","first-page":"50","DOI":"10.1109\/TMM.2021.3120873","volume":"25","author":"X Lin","year":"2023","unstructured":"Lin, X., Sun, S., Huang, W., et al.: Eapt: efficient attention pyramid transformer for image processing. IEEE Trans. Multimed. 25, 50\u201361 (2023). https:\/\/doi.org\/10.1109\/TMM.2021.3120873","journal-title":"IEEE Trans. Multimed."},{"issue":"9","key":"3396_CR29","doi-asserted-by":"publisher","first-page":"3341","DOI":"10.1007\/S00371-022-02550-X","volume":"38","author":"Q Liu","year":"2022","unstructured":"Liu, Q., Zhao, J., Cheng, C., et al.: Pointalcr: adversarial latent GAN and contrastive regularization for point cloud completion. Vis. Comput. 38(9), 3341\u20133349 (2022). https:\/\/doi.org\/10.1007\/S00371-022-02550-X","journal-title":"Vis. Comput."},{"key":"3396_CR30","first-page":"963","volume":"32","author":"Z Liu","year":"2019","unstructured":"Liu, Z., Tang, H., Lin, Y., et al.: Point-voxel cnn for efficient 3d deep learning. Adv. Neural Inf. Process. Syst. 32, 963\u2013973 (2019)","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"3396_CR31","doi-asserted-by":"publisher","unstructured":"Luo, S., Hu, W.: Diffusion probabilistic models for 3d point cloud generation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2837\u20132845 (2021). https:\/\/doi.org\/10.1109\/CVPR46437.2021.00286","DOI":"10.1109\/CVPR46437.2021.00286"},{"key":"3396_CR32","unstructured":"Lyu, Z., Kong, Z., Xu, X. et\u00a0al.: A conditional point diffusion-refinement paradigm for 3d point cloud completion. In: International Conference on Learning Representations (2022)"},{"key":"3396_CR33","doi-asserted-by":"publisher","unstructured":"Ma, B., Liu, Y.S., Han, Z.: Reconstructing surfaces for sparse point clouds with on-surface priors. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 6315\u20136325 (2022). https:\/\/doi.org\/10.1109\/CVPR52688.2022.00621","DOI":"10.1109\/CVPR52688.2022.00621"},{"key":"3396_CR34","doi-asserted-by":"publisher","unstructured":"Peebles, W., Xie, S.: Scalable diffusion models with transformers, pp. 4172\u20134182 (2023). https:\/\/doi.org\/10.1109\/ICCV51070.2023.00387","DOI":"10.1109\/ICCV51070.2023.00387"},{"key":"3396_CR35","first-page":"13032","volume":"34","author":"S Peng","year":"2021","unstructured":"Peng, S., Jiang, C., Liao, Y., et al.: Shape as points: a differentiable poisson solver. Adv. Neural Inf. Process. Syst. 34, 13032\u201313044 (2021)","journal-title":"Adv. Neural Inf. Process. Syst."},{"issue":"8","key":"3396_CR36","doi-asserted-by":"publisher","first-page":"2195","DOI":"10.1007\/S00371-020-01979-2","volume":"37","author":"Z Qin","year":"2021","unstructured":"Qin, Z., Yin, M., Lin, Z., et al.: Three-view generation based on a single front view image for car. Vis. Comput. 37(8), 2195\u20132205 (2021). https:\/\/doi.org\/10.1007\/S00371-020-01979-2","journal-title":"Vis. Comput."},{"key":"3396_CR37","doi-asserted-by":"publisher","unstructured":"Ramasinghe, S., Khan, S., Barnes, N., et\u00a0al.: Spectral-gans for high-resolution 3d point-cloud generation. In: 2020 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 8169\u20138176 (2020). https:\/\/doi.org\/10.1109\/IROS45743.2020.9341265","DOI":"10.1109\/IROS45743.2020.9341265"},{"key":"3396_CR38","doi-asserted-by":"publisher","unstructured":"Rombach, R., Blattmann, A., Lorenz, D. et\u00a0al.: High-resolution image synthesis with latent diffusion models. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 10684\u201310695 (2022). https:\/\/doi.org\/10.1109\/CVPR52688.2022.01042","DOI":"10.1109\/CVPR52688.2022.01042"},{"key":"3396_CR39","doi-asserted-by":"crossref","unstructured":"Ruan, L., Ma, Y., Yang, H. et\u00a0al.: Mm-diffusion: learning multi-modal diffusion models for joint audio and video generation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10219\u201310228 (2023)","DOI":"10.1109\/CVPR52729.2023.00985"},{"issue":"7","key":"3396_CR40","doi-asserted-by":"publisher","first-page":"6662","DOI":"10.1109\/TCYB.2021.3079311","volume":"52","author":"B Sheng","year":"2022","unstructured":"Sheng, B., Andge Riaz Ali, P.L., Chen, C.L.P.: Improving video temporal consistency via broad learning system. IEEE Trans. Cybern. 52(7), 6662\u20136675 (2022). https:\/\/doi.org\/10.1109\/TCYB.2021.3079311","journal-title":"IEEE Trans. Cybern."},{"key":"3396_CR41","doi-asserted-by":"publisher","unstructured":"Shu, D.W., Park, S.W., Kwon, U.: 3d point cloud generative adversarial network based on tree structured graph convolutions. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 3859\u20133868 (2019). https:\/\/doi.org\/10.1109\/ICCV.2019.00396","DOI":"10.1109\/ICCV.2019.00396"},{"key":"3396_CR42","doi-asserted-by":"publisher","unstructured":"Tchapmi, L.P., Kosaraju, V., Rezatofighi, H. et\u00a0al.: Topnet: structural point cloud decoder. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 383\u2013392, (2019). https:\/\/doi.org\/10.1109\/CVPR.2019.00047","DOI":"10.1109\/CVPR.2019.00047"},{"key":"3396_CR43","first-page":"11287","volume":"34","author":"A Vahdat","year":"2021","unstructured":"Vahdat, A., Kreis, K., Kautz, J.: Score-based generative modeling in latent space. Adv. Neural Inf. Process. Syst. 34, 11287\u201311302 (2021)","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"3396_CR44","first-page":"82","volume":"29","author":"J Wu","year":"2016","unstructured":"Wu, J., Zhang, C., Xue, T., et al.: Learning a probabilistic latent space of object shapes via 3d generative-adversarial modeling. Adv. Neural Inf. Process. Syst. 29, 82\u201390 (2016)","journal-title":"Adv. Neural Inf. Process. Syst."},{"issue":"12","key":"3396_CR45","doi-asserted-by":"publisher","first-page":"14971","DOI":"10.1007\/s10489-022-04219-3","volume":"53","author":"M Xiang","year":"2023","unstructured":"Xiang, M., Ye, H., Yang, B., et al.: Multi-space and detail-supplemented attention network for point cloud completion. Appl. Intell. 53(12), 14971\u201314985 (2023). https:\/\/doi.org\/10.1007\/s10489-022-04219-3","journal-title":"Appl. Intell."},{"key":"3396_CR46","doi-asserted-by":"publisher","unstructured":"Xiang, P., Wen, X., Liu, Y.S. et\u00a0al.: Snowflakenet: point cloud completion by snowflake point deconvolution with skip-transformer. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 5499\u20135509 (2021). https:\/\/doi.org\/10.1109\/ICCV48922.2021.00545","DOI":"10.1109\/ICCV48922.2021.00545"},{"issue":"2","key":"3396_CR47","doi-asserted-by":"publisher","first-page":"2362","DOI":"10.1007\/s10489-022-03576-3","volume":"53","author":"F Xu","year":"2023","unstructured":"Xu, F., Wang, Z., Wang, H., et al.: Dynamic vehicle pose estimation and tracking based on motion feedback for lidars. Appl. Intell. 53(2), 2362\u20132390 (2023). https:\/\/doi.org\/10.1007\/s10489-022-03576-3","journal-title":"Appl. Intell."},{"key":"3396_CR48","doi-asserted-by":"publisher","unstructured":"Xu, Q., Xu, Z., Philip, J., et\u00a0al.: Point-nerf: point-based neural radiance fields. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5438\u20135448 (2022). https:\/\/doi.org\/10.1109\/CVPR52688.2022.00536","DOI":"10.1109\/CVPR52688.2022.00536"},{"key":"3396_CR49","doi-asserted-by":"publisher","unstructured":"Yang, G., Huang, X., Hao, Z. et\u00a0al.: Pointflow: 3d point cloud generation with continuous normalizing flows. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 4541\u20134550 (2019). https:\/\/doi.org\/10.1109\/ICCV.2019.00464","DOI":"10.1109\/ICCV.2019.00464"},{"key":"3396_CR50","doi-asserted-by":"publisher","unstructured":"Yang, Y., Feng, C., Shen, Y. et\u00a0al.: Foldingnet: point cloud auto-encoder via deep grid deformation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 206\u2013215, (2018). https:\/\/doi.org\/10.1109\/CVPR.2018.00029","DOI":"10.1109\/CVPR.2018.00029"},{"key":"3396_CR51","doi-asserted-by":"publisher","unstructured":"Yu, X., Rao, Y., Wang, Z., et\u00a0al.: Pointr: diverse point cloud completion with geometry-aware transformers. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 12498\u201312507 (2021). https:\/\/doi.org\/10.1109\/ICCV48922.2021.01227","DOI":"10.1109\/ICCV48922.2021.01227"},{"key":"3396_CR52","doi-asserted-by":"publisher","unstructured":"Yuan, W., Khot, T., Held, D. et\u00a0al.: Pcn: point completion network. In: International Conference on 3D Vision (3DV), pp. 728\u2013737 (2018). https:\/\/doi.org\/10.1109\/3DV.2018.00088","DOI":"10.1109\/3DV.2018.00088"},{"key":"3396_CR53","first-page":"10021","volume":"35","author":"X Zeng","year":"2022","unstructured":"Zeng, X., Vahdat, A., Williams, F., et al.: Lion: latent point diffusion models for 3d shape generation. Adv. Neural Inf. Process. Syst. 35, 10021\u201310039 (2022)","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"3396_CR54","doi-asserted-by":"publisher","unstructured":"Zhang, B., Gu, S., Zhang, B. et\u00a0al.: Styleswin: transformer-based gan for high-resolution image generation, pp. 11294\u201311304 (2022). https:\/\/doi.org\/10.1109\/CVPR52688.2022.01102","DOI":"10.1109\/CVPR52688.2022.01102"},{"key":"3396_CR55","doi-asserted-by":"publisher","unstructured":"Zhang, X., Feng, Y., Li, S., et\u00a0al.: View-guided point cloud completion. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 15890\u201315899 (2021). https:\/\/doi.org\/10.1109\/CVPR46437.2021.01563","DOI":"10.1109\/CVPR46437.2021.01563"},{"key":"3396_CR56","doi-asserted-by":"publisher","unstructured":"Zhou, L., Du, Y., Wu, J.: 3d shape generation and completion through point-voxel diffusion. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 5826\u20135835 (2021). https:\/\/doi.org\/10.1109\/ICCV48922.2021.00577","DOI":"10.1109\/ICCV48922.2021.00577"},{"key":"3396_CR57","doi-asserted-by":"publisher","DOI":"10.1109\/TVCG.2023.3236061","author":"Z Zhu","year":"2023","unstructured":"Zhu, Z., Nan, L., Xie, H., et al.: Csdn: cross-modal shape-transfer dual-refinement network for point cloud completion. IEEE Trans. Vis. Comput. Gr. (2023). https:\/\/doi.org\/10.1109\/TVCG.2023.3236061","journal-title":"IEEE Trans. Vis. Comput. Gr."}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-024-03396-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-024-03396-1\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-024-03396-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,6,6]],"date-time":"2024-06-06T11:20:43Z","timestamp":1717672843000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-024-03396-1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,4,22]]},"references-count":57,"journal-issue":{"issue":"6","published-print":{"date-parts":[[2024,6]]}},"alternative-id":["3396"],"URL":"https:\/\/doi.org\/10.1007\/s00371-024-03396-1","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"value":"0178-2789","type":"print"},{"value":"1432-2315","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,4,22]]},"assertion":[{"value":"29 March 2024","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"22 April 2024","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"Not applicable. The current study does not involve humans and animals.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical and informed consent for data used"}}]}}