{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,12]],"date-time":"2026-03-12T12:04:31Z","timestamp":1773317071439,"version":"3.50.1"},"reference-count":28,"publisher":"Springer Science and Business Media LLC","issue":"4","license":[{"start":{"date-parts":[[2024,6,21]],"date-time":"2024-06-21T00:00:00Z","timestamp":1718928000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,6,21]],"date-time":"2024-06-21T00:00:00Z","timestamp":1718928000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["No.62002039, No. 61672122"],"award-info":[{"award-number":["No.62002039, No. 61672122"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100012226","name":"Fundamental Research Funds for the Central Universities","doi-asserted-by":"publisher","award":["No.36330603"],"award-info":[{"award-number":["No.36330603"]}],"id":[{"id":"10.13039\/501100012226","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2025,3]]},"DOI":"10.1007\/s00371-024-03535-8","type":"journal-article","created":{"date-parts":[[2024,6,21]],"date-time":"2024-06-21T02:01:39Z","timestamp":1718935299000},"page":"2323-2337","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Point cloud upsampling via a coarse-to-fine network with transformer-encoder"],"prefix":"10.1007","volume":"41","author":[{"given":"Yixi","family":"Li","sequence":"first","affiliation":[]},{"given":"Yanzhe","family":"Liu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-5848-6398","authenticated-orcid":false,"given":"Rong","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Hui","family":"Li","sequence":"additional","affiliation":[]},{"given":"Na","family":"Zhao","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,6,21]]},"reference":[{"key":"3535_CR1","doi-asserted-by":"crossref","unstructured":"Li, R., Li, X., Heng, P.-A., Fu, C.-W.: Point cloud upsampling via disentangled refinement. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 344\u2013353 (2021)","DOI":"10.1109\/CVPR46437.2021.00041"},{"issue":"12","key":"3535_CR2","doi-asserted-by":"publisher","first-page":"4338","DOI":"10.1109\/TPAMI.2020.3005434","volume":"43","author":"Y Guo","year":"2020","unstructured":"Guo, Y., Wang, H., Hu, Q., Liu, H., Liu, L., Bennamoun, M.: Deep learning for 3d point clouds: a survey. IEEE Trans. Pattern Anal. Mach. Intell. 43(12), 4338\u20134364 (2020)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"3535_CR3","doi-asserted-by":"crossref","unstructured":"Qian, G., Abualshour, A., Li, G., Thabet, A., Ghanem, B.: Pu-gcn: point cloud upsampling using graph convolutional networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11683\u201311692 (2021)","DOI":"10.1109\/CVPR46437.2021.01151"},{"key":"3535_CR4","doi-asserted-by":"crossref","unstructured":"Yifan, W., Wu, S., Huang, H., Cohen-Or, D., Sorkine-Hornung, O.: Patch-based progressive 3d point set upsampling. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5958\u20135967 (2019)","DOI":"10.1109\/CVPR.2019.00611"},{"key":"3535_CR5","doi-asserted-by":"crossref","unstructured":"Li, R., Li, X., Fu, C.-W., Cohen-Or, D., Heng, P.-A.: Pu-gan: a point cloud upsampling adversarial network. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 7203\u20137212 (2019)","DOI":"10.1109\/ICCV.2019.00730"},{"key":"3535_CR6","doi-asserted-by":"crossref","unstructured":"Du, H., Yan, X., Wang, J., Xie, D., Pu, S.: Point cloud upsampling via cascaded refinement network. In: Proceedings of the Asian Conference on Computer Vision, pp. 586\u2013601 (2022)","DOI":"10.1007\/978-3-031-26319-4_7"},{"key":"3535_CR7","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2023.109796","volume":"143","author":"T Zhao","year":"2023","unstructured":"Zhao, T., Li, L., Tian, T., Ma, J., Tian, J.: Apunet: attention-guided upsampling network for sparse and non-uniform point cloud. Pattern Recognit. 143, 109796 (2023)","journal-title":"Pattern Recognit."},{"key":"3535_CR8","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, \u0141., Polosukhin, I.: Attention is all you need. Adv. Neural Inf. Proc. Syst. 30, 5998\u20136008 (2017)"},{"key":"3535_CR9","unstructured":"Devlin, J., Chang, M.-W., Lee, K., Toutanova, K.: Bert: pre-training of deep bidirectional transformers for language understanding. arXiv:1810.04805 (2018)"},{"key":"3535_CR10","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., et al.: An image is worth 16x16 words: transformers for image recognition at scale. arXiv:2010.11929 (2020)"},{"issue":"7","key":"3535_CR11","doi-asserted-by":"publisher","first-page":"2781","DOI":"10.1007\/s00371-022-02492-4","volume":"39","author":"G Cai","year":"2023","unstructured":"Cai, G., Zhu, Y., Wu, Y., Jiang, X., Ye, J., Yang, D.: A multimodal transformer to fuse images and metadata for skin disease classification. Vis. Comput. 39(7), 2781\u20132793 (2023)","journal-title":"Vis. Comput."},{"issue":"5","key":"3535_CR12","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3326362","volume":"38","author":"Y Wang","year":"2019","unstructured":"Wang, Y., Sun, Y., Liu, Z., Sarma, S.E., Bronstein, M.M., Solomon, J.M.: Dynamic graph cnn for learning on point clouds. ACM Trans. Graph. (tog) 38(5), 1\u201312 (2019)","journal-title":"ACM Trans. Graph. (tog)"},{"key":"3535_CR13","doi-asserted-by":"crossref","unstructured":"Yu, L., Li, X., Fu, C.-W., Cohen-Or, D., Heng, P.-A.: Pu-net: point cloud upsampling network. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2790\u20132799 (2018)","DOI":"10.1109\/CVPR.2018.00295"},{"key":"3535_CR14","doi-asserted-by":"crossref","unstructured":"Qian, Y., Hou, J., Kwong, S., He, Y.: Pugeo-net: a geometry-centric network for 3d point cloud upsampling. In: European Conference on Computer Vision, pp. 752\u2013769. Springer (2020)","DOI":"10.1007\/978-3-030-58529-7_44"},{"issue":"12","key":"3535_CR15","doi-asserted-by":"publisher","first-page":"4661","DOI":"10.1109\/TCSVT.2021.3099106","volume":"31","author":"D Ding","year":"2021","unstructured":"Ding, D., Qiu, C., Liu, F., Pan, Z.: Point cloud upsampling via perturbation learning. IEEE Trans. Circuits Syst. Video Technol. 31(12), 4661\u20134672 (2021)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"3535_CR16","doi-asserted-by":"crossref","unstructured":"Long, C., Zhang, W., Li, R., Wang, H., Dong, Z., Yang, B.: Pc2-pu: patch correlation and point correlation for effective point cloud upsampling. In: Proceedings of the 30th ACM International Conference on Multimedia, pp. 2191\u20132201 (2022)","DOI":"10.1145\/3503161.3547777"},{"issue":"12","key":"3535_CR17","doi-asserted-by":"publisher","first-page":"4673","DOI":"10.1109\/TCSVT.2021.3100134","volume":"31","author":"P Zhang","year":"2021","unstructured":"Zhang, P., Wang, X., Ma, L., Wang, S., Kwong, S., Jiang, J.: Progressive point cloud upsampling via differentiable rendering. IEEE Trans. Circuits Syst. Video Technol. 31(12), 4673\u20134685 (2021)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"3535_CR18","doi-asserted-by":"publisher","first-page":"7389","DOI":"10.1109\/TIP.2022.3222918","volume":"31","author":"H Liu","year":"2022","unstructured":"Liu, H., Yuan, H., Hou, J., Hamzaoui, R., Gao, W.: Pufa-gan: a frequency-aware generative adversarial network for 3d point cloud upsampling. IEEE Trans. Image Process. 31, 7389\u20137402 (2022)","journal-title":"IEEE Trans. Image Process."},{"key":"3535_CR19","doi-asserted-by":"crossref","unstructured":"Zhao, H., Jiang, L., Jia, J., Torr, P.H., Koltun, V.: Point transformer. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 16259\u201316268 (2021)","DOI":"10.1109\/ICCV48922.2021.01595"},{"key":"3535_CR20","doi-asserted-by":"publisher","first-page":"187","DOI":"10.1007\/s41095-021-0229-5","volume":"7","author":"M-H Guo","year":"2021","unstructured":"Guo, M.-H., Cai, J.-X., Liu, Z.-N., Mu, T.-J., Martin, R.R., Hu, S.-M.: Pct: point cloud transformer. Comput. Vis. Media 7, 187\u2013199 (2021)","journal-title":"Comput. Vis. Media"},{"key":"3535_CR21","doi-asserted-by":"crossref","unstructured":"Qiu, S., Anwar, S., Barnes, N.: Pu-transformer: point cloud upsampling transformer. In: Proceedings of the Asian Conference on Computer Vision, pp. 2475\u20132493 (2022)","DOI":"10.1007\/978-3-031-26319-4_20"},{"key":"3535_CR22","doi-asserted-by":"crossref","unstructured":"Wang, Q., Wu, B., Zhu, P., Li, P., Zuo, W., Hu, Q.: Eca-net: efficient channel attention for deep convolutional neural networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11534\u201311542 (2020)","DOI":"10.1109\/CVPR42600.2020.01155"},{"key":"3535_CR23","doi-asserted-by":"crossref","unstructured":"Fan, H., Su, H., Guibas, L.J.: A point set generation network for 3d object reconstruction from a single image. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 605\u2013613 (2017)","DOI":"10.1109\/CVPR.2017.264"},{"key":"3535_CR24","unstructured":"Chang, A.X., Funkhouser, T., Guibas, L., Hanrahan, P., Huang, Q., Li, Z., Savarese, S., Savva, M., Song, S., Su, H., et al.: Shapenet: an information-rich 3d model repository. arXiv:1512.03012 (2015)"},{"key":"3535_CR25","doi-asserted-by":"crossref","unstructured":"Dai, A., Chang, A. X., Savva, M., Halber, M., Funkhouser, T., Nie\u00dfner, M.: Scannet: richly-annotated 3d reconstructions of indoor scenes. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5828\u20135839 (2017)","DOI":"10.1109\/CVPR.2017.261"},{"key":"3535_CR26","doi-asserted-by":"crossref","unstructured":"Geiger, A., Lenz, P., Urtasun, R.: Are we ready for autonomous driving? The kitti vision benchmark suite. In: 2012 IEEE Conference on Computer Vision and Pattern Recognition, pp. 3354\u20133361 (2012). IEEE","DOI":"10.1109\/CVPR.2012.6248074"},{"issue":"3","key":"3535_CR27","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/2487228.2487237","volume":"32","author":"M Kazhdan","year":"2013","unstructured":"Kazhdan, M., Hoppe, H.: Screened Poisson surface reconstruction. ACM Trans. Graph. (ToG) 32(3), 1\u201313 (2013)","journal-title":"ACM Trans. Graph. (ToG)"},{"key":"3535_CR28","doi-asserted-by":"crossref","unstructured":"He, Y., Tang, D., Zhang, Y., Xue, X., Fu, Y.: Grad-pu: arbitrary-scale point cloud upsampling via gradient descent with learned distance functions. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5354\u20135363 (2023)","DOI":"10.1109\/CVPR52729.2023.00518"}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-024-03535-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-024-03535-8\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-024-03535-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,3,3]],"date-time":"2025-03-03T11:30:49Z","timestamp":1741001449000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-024-03535-8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,6,21]]},"references-count":28,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2025,3]]}},"alternative-id":["3535"],"URL":"https:\/\/doi.org\/10.1007\/s00371-024-03535-8","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"value":"0178-2789","type":"print"},{"value":"1432-2315","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,6,21]]},"assertion":[{"value":"2 June 2024","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"21 June 2024","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors have no financial or proprietary interests in any material discussed in this article.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}