{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,17]],"date-time":"2025-11-17T03:02:00Z","timestamp":1763348520998,"version":"3.37.3"},"reference-count":61,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2023,4,15]],"date-time":"2023-04-15T00:00:00Z","timestamp":1681516800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,4,15]],"date-time":"2023-04-15T00:00:00Z","timestamp":1681516800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100012165","name":"Key Technologies Research and Development Program","doi-asserted-by":"publisher","award":["2021ZD0201300"],"award-info":[{"award-number":["2021ZD0201300"]}],"id":[{"id":"10.13039\/501100012165","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2024,2]]},"DOI":"10.1007\/s00371-023-02840-y","type":"journal-article","created":{"date-parts":[[2023,4,15]],"date-time":"2023-04-15T15:02:11Z","timestamp":1681570931000},"page":"1193-1204","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["Self-supervised learning of monocular 3D geometry understanding with two- and three-view geometric constraints"],"prefix":"10.1007","volume":"40","author":[{"given":"Xiaoliang","family":"Liu","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7285-326X","authenticated-orcid":false,"given":"Furao","family":"Shen","sequence":"additional","affiliation":[]},{"given":"Jian","family":"Zhao","sequence":"additional","affiliation":[]},{"given":"Changhai","family":"Nie","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,4,15]]},"reference":[{"issue":"2","key":"2840_CR1","doi-asserted-by":"publisher","first-page":"91","DOI":"10.1023\/B:VISI.0000029664.99615.94","volume":"60","author":"DG Lowe","year":"2004","unstructured":"Lowe, D.G.: Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. (IJCV) 60(2), 91\u2013110 (2004)","journal-title":"Int. J. Comput. Vis. (IJCV)"},{"key":"2840_CR2","doi-asserted-by":"crossref","unstructured":"Bian, J., Lin, W.-Y., Matsushita, Y., Yeung, S.-K., Nguyen, T.-D., Cheng, M.-M.: Gms: Grid-based motion statistics for fast, ultra-robust feature correspondence. In: IEEE conference on computer vision and pattern recognition (CVPR), pp. 4181\u20134190 (2017)","DOI":"10.1109\/CVPR.2017.302"},{"key":"2840_CR3","volume-title":"Multiple View Geometry in Computer Vision","author":"R Hartley","year":"2003","unstructured":"Hartley, R., Zisserman, A.: Multiple View Geometry in Computer Vision. Cambridge University Press, Cambridge (2003)"},{"key":"2840_CR4","unstructured":"Horn, B.K., Schunck, B.G.: Determining optical flow. In: Techniques and Applications of Image Understanding, vol. 281, pp. 319\u2013331 (1981). International Society for Optics and Photonics, USA"},{"key":"2840_CR5","doi-asserted-by":"crossref","unstructured":"Sun, D., Roth, S., Black, M.J.: Secrets of optical flow estimation and their principles. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2432\u20132439 (2010). IEEE","DOI":"10.1109\/CVPR.2010.5539939"},{"key":"2840_CR6","unstructured":"Eigen, D., Puhrsch, C., Fergus, R.: Depth map prediction from a single image using a multi-scale deep network. In: Neural Information Processing Systems (NeurIPS) (2014)"},{"key":"2840_CR7","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2015.2505283","author":"F Liu","year":"2016","unstructured":"Liu, F., Shen, C., Lin, G., Reid, I.: Learning depth from single monocular images using deep convolutional neural fields. IEEE Trans. Pattern Recognit. Mach. Intell. (PAMI) (2016). https:\/\/doi.org\/10.1109\/TPAMI.2015.2505283","journal-title":"IEEE Trans. Pattern Recognit. Mach. Intell. (PAMI)"},{"key":"2840_CR8","doi-asserted-by":"crossref","unstructured":"Wang, C., Buenaposada, J.M., Zhu, R., Lucey, S.: Learning depth from monocular videos using direct methods. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2018)","DOI":"10.1109\/CVPR.2018.00216"},{"key":"2840_CR9","doi-asserted-by":"crossref","unstructured":"Wang, S., Clark, R., Wen, H., Trigoni, N.: Deepvo: towards end-to-end visual odometry with deep recurrent convolutional neural networks. In: IEEE International Conference on Robotics and Automation (ICRA), pp. 2043\u20132050 (2017). IEEE","DOI":"10.1109\/ICRA.2017.7989236"},{"issue":"4\u20135","key":"2840_CR10","doi-asserted-by":"publisher","first-page":"513","DOI":"10.1177\/0278364917734298","volume":"37","author":"S Wang","year":"2018","unstructured":"Wang, S., Clark, R., Wen, H., Trigoni, N.: End-to-end, sequence-to-sequence probabilistic visual odometry through deep neural networks. Int. J. Robot. Res. (IJRR) 37(4\u20135), 513\u2013542 (2018)","journal-title":"Int. J. Robot. Res. (IJRR)"},{"key":"2840_CR11","doi-asserted-by":"crossref","unstructured":"Dosovitskiy, A., Fischer, P., Ilg, E., Hausser, P., Hazirbas, C., Golkov, V., Van Der\u00a0Smagt, P., Cremers, D., Brox, T.: Flownet: Learning optical flow with convolutional networks. In: IEEE International Conference on Computer Vision (ICCV) (2015)","DOI":"10.1109\/ICCV.2015.316"},{"key":"2840_CR12","doi-asserted-by":"crossref","unstructured":"Ilg, E., Mayer, N., Saikia, T., Keuper, M., Dosovitskiy, A., Brox, T.: Flownet 2.0: evolution of optical flow estimation with deep networks. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2017)","DOI":"10.1109\/CVPR.2017.179"},{"key":"2840_CR13","doi-asserted-by":"crossref","unstructured":"Sun, D., Yang, X., Liu, M.-Y., Kautz, J.: Pwc-net: CNNS for optical flow using pyramid, warping, and cost volume. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2018)","DOI":"10.1109\/CVPR.2018.00931"},{"key":"2840_CR14","doi-asserted-by":"crossref","unstructured":"Ranjan, A., Black, M.J.: Optical flow estimation using a spatial pyramid network. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4161\u20134170 (2017)","DOI":"10.1109\/CVPR.2017.291"},{"key":"2840_CR15","doi-asserted-by":"crossref","unstructured":"Garg, R., BG, V.K., Carneiro, G., Reid, I.: Unsupervised CNN for single view depth estimation: Geometry to the rescue. In: European Conference on Computer Vision (ECCV) (2016). Springer","DOI":"10.1007\/978-3-319-46484-8_45"},{"key":"2840_CR16","doi-asserted-by":"crossref","unstructured":"Godard, C., Aodha, O.M., Brostow, G.J.: Unsupervised monocular depth estimation with left-right consistency. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2017)","DOI":"10.1109\/CVPR.2017.699"},{"key":"2840_CR17","doi-asserted-by":"crossref","unstructured":"Meister, S., Hur, J., Roth, S.: UnFlow: unsupervised learning of optical flow with a bidirectional census loss. In: Association for the Advancement of Artificial Intelligence (AAAI), New Orleans, Louisiana (2018)","DOI":"10.1609\/aaai.v32i1.12276"},{"key":"2840_CR18","doi-asserted-by":"crossref","unstructured":"Zhong, Y., Ji, P., Wang, J., Dai, Y., Li, H.: Unsupervised deep epipolar flow for stationary or dynamic scenes. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 12095\u201312104 (2019)","DOI":"10.1109\/CVPR.2019.01237"},{"key":"2840_CR19","doi-asserted-by":"crossref","unstructured":"Janai, J., Guney, F., Ranjan, A., Black, M., Geiger, A.: Unsupervised learning of multi-frame optical flow with occlusions. In: European Conference on Computer Vision (ECCV), pp. 690\u2013706 (2018)","DOI":"10.1007\/978-3-030-01270-0_42"},{"key":"2840_CR20","doi-asserted-by":"crossref","unstructured":"Yin, Z., Shi, J.: GeoNet: unsupervised learning of dense depth, optical flow and camera pose. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2018)","DOI":"10.1109\/CVPR.2018.00212"},{"key":"2840_CR21","doi-asserted-by":"crossref","unstructured":"Zou, Y., Luo, Z., Huang, J.-B.: DF-Net: unsupervised joint learning of depth and flow using cross-task consistency. In: European Conference on Computer Vision (ECCV) (2018)","DOI":"10.1007\/978-3-030-01228-1_3"},{"key":"2840_CR22","doi-asserted-by":"crossref","unstructured":"Ranjan, A., Jampani, V., Kim, K., Sun, D., Wulff, J., Black, M.J.: Competitive Collaboration: joint unsupervised learning of depth, camera motion, optical flow and motion segmentation. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2019)","DOI":"10.1109\/CVPR.2019.01252"},{"key":"2840_CR23","doi-asserted-by":"crossref","unstructured":"Chen, Y., Schmid, C., Sminchisescu, C.: Self-supervised learning with geometric constraints in monocular video: Connecting flow, depth, and camera. In: IEEE International Conference on Computer Vision (ICCV), pp. 7063\u20137072 (2019)","DOI":"10.1109\/ICCV.2019.00716"},{"key":"2840_CR24","unstructured":"Wang, Y., Yang, Z., Wang, P., Yang, Y., Luo, C., Xu, W.: Joint unsupervised learning of optical flow and depth by watching stereo videos. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2019)"},{"key":"2840_CR25","doi-asserted-by":"crossref","unstructured":"Liu, L., Zhai, G., Ye, W., Liu, Y.: Unsupervised learning of scene flow estimation fusing with local rigidity. In: Association for the Advancement of Artificial Intelligence (AAAI), pp. 876\u2013882 (2019)","DOI":"10.24963\/ijcai.2019\/123"},{"issue":"11","key":"2840_CR26","doi-asserted-by":"publisher","first-page":"1231","DOI":"10.1177\/0278364913491297","volume":"32","author":"A Geiger","year":"2013","unstructured":"Geiger, A., Lenz, P., Stiller, C., Urtasun, R.: Vision meets robotics: the KITTI dataset. Int. J. Robot. Res. (IJRR) 32(11), 1231\u20131237 (2013)","journal-title":"Int. J. Robot. Res. (IJRR)"},{"key":"2840_CR27","doi-asserted-by":"crossref","unstructured":"Geiger, A., Ziegler, J., Stiller, C.: Stereoscan: Dense 3D reconstruction in real-time. In: IEEE Intelligent Vehicles Symposium (IV), pp. 963\u2013968 (2011). IEEE","DOI":"10.1109\/IVS.2011.5940405"},{"issue":"5","key":"2840_CR28","doi-asserted-by":"publisher","first-page":"1255","DOI":"10.1109\/TRO.2017.2705103","volume":"33","author":"R Mur-Artal","year":"2017","unstructured":"Mur-Artal, R., Tard\u00f3s, J.D.: ORB-SLAM2: an open-source SLAM system for monocular, stereo, and RGB-D cameras. IEEE Trans. Robot. (TRO) 33(5), 1255\u20131262 (2017)","journal-title":"IEEE Trans. Robot. (TRO)"},{"key":"2840_CR29","doi-asserted-by":"crossref","unstructured":"Engel, J., Sch\u00f6ps, T., Cremers, D.: LSD-SLAM: large-scale direct monocular slam. In: European Conference on Computer Vision (ECCV), pp. 834\u2013849 (2014). Springer","DOI":"10.1007\/978-3-319-10605-2_54"},{"issue":"3","key":"2840_CR30","doi-asserted-by":"publisher","first-page":"611","DOI":"10.1109\/TPAMI.2017.2658577","volume":"40","author":"J Engel","year":"2017","unstructured":"Engel, J., Koltun, V., Cremers, D.: Direct sparse odometry. IEEE Trans. Pattern Recognit. Mach. Intell. (PAMI) 40(3), 611\u2013625 (2017)","journal-title":"IEEE Trans. Pattern Recognit. Mach. Intell. (PAMI)"},{"key":"2840_CR31","doi-asserted-by":"crossref","unstructured":"Laina, I., Rupprecht, C., Belagiannis, V., Tombari, F., Navab, N.: Deeper depth prediction with fully convolutional residual networks. In: 3D Vision (3DV), pp. 239\u2013248 (2016). IEEE","DOI":"10.1109\/3DV.2016.32"},{"key":"2840_CR32","doi-asserted-by":"crossref","unstructured":"Kendall, A., Grimes, M., Cipolla, R.: Posenet: a convolutional network for real-time 6-dof camera relocalization. In: IEEE International Conference on Computer Vision (ICCV), pp. 2938\u20132946 (2015)","DOI":"10.1109\/ICCV.2015.336"},{"issue":"2","key":"2840_CR33","doi-asserted-by":"publisher","first-page":"651","DOI":"10.1109\/TASE.2017.2664920","volume":"15","author":"R Li","year":"2017","unstructured":"Li, R., Liu, Q., Gui, J., Gu, D., Hu, H.: Indoor relocalization in challenging environments with dual-stream convolutional neural networks. IEEE Trans. Autom. Sci. Eng. 15(2), 651\u2013662 (2017)","journal-title":"IEEE Trans. Autom. Sci. Eng."},{"issue":"1","key":"2840_CR34","doi-asserted-by":"publisher","first-page":"18","DOI":"10.1109\/LRA.2015.2505717","volume":"1","author":"G Costante","year":"2015","unstructured":"Costante, G., Mancini, M., Valigi, P., Ciarfuglia, T.A.: Exploring representation learning with CNNS for frame-to-frame ego-motion estimation. IEEE Robot. Autom. Lett. 1(1), 18\u201325 (2015)","journal-title":"IEEE Robot. Autom. Lett."},{"key":"2840_CR35","doi-asserted-by":"crossref","unstructured":"Clark, R., Wang, S., Wen, H., Markham, A., Trigoni, N.: Vinet: visual-inertial odometry as a sequence-to-sequence learning problem. In: Association for the Advancement of Artificial Intelligence (AAAI) (2017)","DOI":"10.1609\/aaai.v31i1.11215"},{"key":"2840_CR36","doi-asserted-by":"crossref","unstructured":"Li, Z., Chen, Z., Liu, X., Jiang, J.: Depthformer: exploiting long-range correlation and local information for accurate monocular depth estimation. arXiv preprint arXiv:2203.14211 (2022)","DOI":"10.1007\/s11633-023-1458-0"},{"key":"2840_CR37","unstructured":"Li, Z., Wang, X., Liu, X., Jiang, J.: Binsformer: revisiting adaptive bins for monocular depth estimation. arXiv preprint arXiv:2204.00987 (2022)"},{"key":"2840_CR38","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., et al.: An image is worth 16x16 words: transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)"},{"key":"2840_CR39","doi-asserted-by":"crossref","unstructured":"Geiger, A., Lenz, P., Urtasun, R.: Are we ready for autonomous driving? the KITTI vision benchmark suite. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3354\u20133361 (2012). IEEE","DOI":"10.1109\/CVPR.2012.6248074"},{"key":"2840_CR40","unstructured":"Butler, D., Wulff, J., Stanley, G., Black, M.: MPI-sintel optical flow benchmark: Supplemental material. In: MPI-IS-TR-006, MPI for Intelligent Systems (2012). Citeseer"},{"key":"2840_CR41","doi-asserted-by":"crossref","unstructured":"Pilzer, A., Lathuiliere, S., Sebe, N., Ricci, E.: Refine and distill: Exploiting cycle-inconsistency and knowledge distillation for unsupervised monocular depth estimation. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 9768\u20139777 (2019)","DOI":"10.1109\/CVPR.2019.01000"},{"key":"2840_CR42","doi-asserted-by":"crossref","unstructured":"Ahmadi, A., Patras, I.: Unsupervised convolutional neural networks for motion estimation. In: IEEE International Conference on Image Processing (ICIP), pp. 1629\u20131633 (2016). IEEE","DOI":"10.1109\/ICIP.2016.7532634"},{"key":"2840_CR43","doi-asserted-by":"crossref","unstructured":"Jason, J.Y., Harley, A.W., Derpanis, K.G.: Back to basics: unsupervised learning of optical flow via brightness constancy and motion smoothness. In: European Conference on Computer Vision (ECCV), pp. 3\u201310 (2016). Springer","DOI":"10.1007\/978-3-319-49409-8_1"},{"key":"2840_CR44","doi-asserted-by":"crossref","unstructured":"Wang, Y., Yang, Y., Yang, Z., Zhao, L., Wang, P., Xu, W.: Occlusion aware unsupervised learning of optical flow. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4884\u20134893 (2018)","DOI":"10.1109\/CVPR.2018.00513"},{"key":"2840_CR45","doi-asserted-by":"crossref","unstructured":"Godard, C., Mac\u00a0Aodha, O., Brostow, G.: Digging into self-supervised monocular depth estimation. arXiv preprint arXiv:1806.01260 (2018)","DOI":"10.1109\/ICCV.2019.00393"},{"key":"2840_CR46","unstructured":"Ruder, S.: An overview of multi-task learning in deep neural networks. arXiv preprint arXiv:1706.05098 (2017)"},{"key":"2840_CR47","doi-asserted-by":"crossref","unstructured":"Zhou, T., Brown, M., Snavely, N., Lowe, D.G.: Unsupervised learning of depth and ego-motion from video. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2017)","DOI":"10.1109\/CVPR.2017.700"},{"key":"2840_CR48","doi-asserted-by":"crossref","unstructured":"Godard, C., Aodha, O.M., Firman, M., Brostow, G.J.: Digging into self-supervised monocular depth estimation. In: IEEE International Conference on Computer Vision (ICCV) (2019)","DOI":"10.1109\/ICCV.2019.00393"},{"key":"2840_CR49","doi-asserted-by":"crossref","unstructured":"Gordon, A., Li, H., Jonschkowski, R., Angelova, A.: Depth from videos in the wild: Unsupervised monocular depth learning from unknown cameras. In: IEEE International Conference on Computer Vision (ICCV), pp. 8977\u20138986 (2019)","DOI":"10.1109\/ICCV.2019.00907"},{"issue":"9","key":"2840_CR50","doi-asserted-by":"publisher","first-page":"2548","DOI":"10.1007\/s11263-021-01484-6","volume":"129","author":"J-W Bian","year":"2021","unstructured":"Bian, J.-W., Zhan, H., Wang, N., Li, Z., Zhang, L., Shen, C., Cheng, M.-M., Reid, I.: Unsupervised scale-consistent depth learning from video. Int. J. Comput. Vis. 129(9), 2548\u20132564 (2021)","journal-title":"Int. J. Comput. Vis."},{"key":"2840_CR51","doi-asserted-by":"crossref","unstructured":"Mayer, N., Ilg, E., Hausser, P., Fischer, P., Cremers, D., Dosovitskiy, A., Brox, T.: A large dataset to train convolutional networks for disparity, optical flow, and scene flow estimation. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4040\u20134048 (2016)","DOI":"10.1109\/CVPR.2016.438"},{"key":"2840_CR52","unstructured":"Bian, J.-W., Li, Z., Wang, N., Zhan, H., Shen, C., Cheng, M.-M., Reid, I.: Unsupervised scale-consistent depth and ego-motion learning from monocular video. In: Neural Information Processing Systems (NeurIPS) (2019)"},{"key":"2840_CR53","doi-asserted-by":"crossref","unstructured":"Teed, Z., Deng, J.: Raft: recurrent all-pairs field transforms for optical flow. In: European Conference on Computer Vision (ECCV), pp. 402\u2013419 (2020). Springer","DOI":"10.1007\/978-3-030-58536-5_24"},{"key":"2840_CR54","unstructured":"Luo, C., Yang, Z., Wang, P., Wang, Y., Xu, W., Nevatia, R., Yuille, A.: Every pixel counts ++: joint learning of geometry and motion with 3d holistic understanding. IEEE Transactions on Pattern Recognition and Machine Intelligence (PAMI) PP(99), 1\u20131"},{"key":"2840_CR55","doi-asserted-by":"crossref","unstructured":"Hur, J., Roth, S.: Self-supervised monocular scene flow estimation. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 7396\u20137405 (2020)","DOI":"10.1109\/CVPR42600.2020.00742"},{"key":"2840_CR56","unstructured":"Wang, G., Zhang, C., Wang, H., Wang, J., Wang, Y., Wang, X.: Unsupervised learning of depth, optical flow and pose with occlusion from 3D geometry. In: IEEE Transactions on Intelligent Transportation Systems (TITS) (2020)"},{"key":"2840_CR57","doi-asserted-by":"crossref","unstructured":"Casser, V., Pirk, S., Mahjourian, R., Angelova, A.: Depth prediction without the sensors: leveraging structure for unsupervised learning from monocular videos. In: Association for the Advancement of Artificial Intelligence (AAAI), vol. 33, pp. 8001\u20138008 (2019)","DOI":"10.1609\/aaai.v33i01.33018001"},{"key":"2840_CR58","doi-asserted-by":"crossref","unstructured":"Zhan, H., Garg, R., Saroj\u00a0Weerasekera, C., Li, K., Agarwal, H., Reid, I.: Unsupervised learning of monocular depth estimation and visual odometry with deep feature reconstruction. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2018)","DOI":"10.1109\/CVPR.2018.00043"},{"key":"2840_CR59","doi-asserted-by":"crossref","unstructured":"Liang, Y., He, F., Zeng, X., Luo, J.: An improved loop subdivision to coordinate the smoothness and the number of faces via multi-objective optimization. Integrated Computer-Aided Engineering (Preprint), pp. 1\u201319 (2022)","DOI":"10.3233\/ICA-210661"},{"issue":"2","key":"2840_CR60","doi-asserted-by":"publisher","first-page":"341","DOI":"10.1109\/TSC.2015.2501981","volume":"11","author":"Y Wu","year":"2015","unstructured":"Wu, Y., He, F., Zhang, D., Li, X.: Service-oriented feature-based data exchange for cloud-based design and manufacturing. IEEE Trans. Serv. Comput. 11(2), 341\u2013353 (2015)","journal-title":"IEEE Trans. Serv. Comput."},{"key":"2840_CR61","doi-asserted-by":"publisher","DOI":"10.1016\/j.cad.2022.103196","volume":"146","author":"Y Song","year":"2022","unstructured":"Song, Y., He, F., Duan, Y., Liang, Y., Yan, X.: A kernel correlation-based approach to adaptively acquire local features for learning 3D point clouds. Comput.-Aided Des. 146, 103196 (2022)","journal-title":"Comput.-Aided Des."}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-023-02840-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-023-02840-y\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-023-02840-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,1,23]],"date-time":"2024-01-23T19:15:01Z","timestamp":1706037301000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-023-02840-y"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,4,15]]},"references-count":61,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2024,2]]}},"alternative-id":["2840"],"URL":"https:\/\/doi.org\/10.1007\/s00371-023-02840-y","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"type":"print","value":"0178-2789"},{"type":"electronic","value":"1432-2315"}],"subject":[],"published":{"date-parts":[[2023,4,15]]},"assertion":[{"value":"1 March 2023","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"15 April 2023","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}