{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,10]],"date-time":"2025-11-10T07:06:49Z","timestamp":1762758409190,"version":"build-2065373602"},"reference-count":32,"publisher":"Springer Science and Business Media LLC","issue":"6","license":[{"start":{"date-parts":[[2025,10,14]],"date-time":"2025-10-14T00:00:00Z","timestamp":1760400000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,10,14]],"date-time":"2025-10-14T00:00:00Z","timestamp":1760400000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J Real-Time Image Proc"],"published-print":{"date-parts":[[2025,12]]},"DOI":"10.1007\/s11554-025-01784-x","type":"journal-article","created":{"date-parts":[[2025,10,14]],"date-time":"2025-10-14T18:57:50Z","timestamp":1760468270000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Real-time 3D underwater target recognition using a lightweight LiteDet-CART network"],"prefix":"10.1007","volume":"22","author":[{"given":"Zhijie","family":"Tang","sequence":"first","affiliation":[]},{"given":"Jiajun","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Tian","family":"Guo","sequence":"additional","affiliation":[]},{"given":"Ke","family":"Meng","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,10,14]]},"reference":[{"key":"1784_CR1","doi-asserted-by":"publisher","first-page":"24330","DOI":"10.1109\/TITS.2022.3203715","volume":"23","author":"J Shen","year":"2022","unstructured":"Shen, J., Zhou, W., Liu, N., Sun, H., Li, D., Zhang, Y.: An anchor-free lightweight deep convolutional network for vehicle detection in aerial images. IEEE Trans. Intell. Transp. Syst. 23, 24330\u201324342 (2022). https:\/\/doi.org\/10.1109\/TITS.2022.3203715","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"key":"1784_CR2","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2022.108873","author":"X Ning","year":"2022","unstructured":"Ning, X., Tian, W., Yu, Z., Li, W., Bai, X.: HCFNN: high-order coverage function neural network for image classification. Pattern Recogn. (2022). https:\/\/doi.org\/10.1016\/j.patcog.2022.108873","journal-title":"Pattern Recogn."},{"key":"1784_CR3","doi-asserted-by":"publisher","DOI":"10.1109\/TIM.2021.3132332","author":"J Shen","year":"2022","unstructured":"Shen, J., Liu, N., Xu, C., et al.: Finger vein recognition algorithm based on lightweight deep convolutional neural network. IEEE Trans. Instrum. Meas. (2022). https:\/\/doi.org\/10.1109\/TIM.2021.3132332","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"1784_CR4","doi-asserted-by":"publisher","DOI":"10.1109\/TIM.2023.3346488","author":"J Shen","year":"2024","unstructured":"Shen, J., Liu, N., Sun, H., et al.: An instrument indication acquisition algorithm based on lightweight deep convolutional neural network and hybrid attention fine-grained features. IEEE Trans. Instrum. Meas. (2024). https:\/\/doi.org\/10.1109\/TIM.2023.3346488","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"1784_CR5","doi-asserted-by":"crossref","unstructured":"Wu, Z., et al.: Boosting 3D single object tracking with 2D matching distillation and 3D pre-training. In: Proceedings of the European Conference on Computer Vision (ECCV 2024) (2024)","DOI":"10.1007\/978-3-031-73254-6_16"},{"key":"1784_CR6","doi-asserted-by":"publisher","unstructured":"Xia, Y.Q., Xia, Y., Li, W., Song, R., Cao, K., Stilla, U.: ASFM-Net: asymmetrical Siamese feature matching network for point completion. ACM MM (2021). https:\/\/doi.org\/10.1145\/3474085.3475348","DOI":"10.1145\/3474085.3475348"},{"key":"1784_CR7","doi-asserted-by":"publisher","unstructured":"Su, H., Maji, S., Kalogerakis, E., Learned-Miller, E.: Multi-view convolutional neural networks for 3D shape recognition. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), pp. 945\u2013953 (2015). https:\/\/doi.org\/10.1109\/ICCV.2015.114","DOI":"10.1109\/ICCV.2015.114"},{"key":"1784_CR8","doi-asserted-by":"publisher","unstructured":"Bai, S., Bai, X., Zhou, Z., Zhang, Z., Latecki, L.J.: GIFT: a real-time and scalable 3D shape search engine. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5023\u20135032 (2016). https:\/\/doi.org\/10.1109\/CVPR.2016.543","DOI":"10.1109\/CVPR.2016.543"},{"issue":"2025","key":"1784_CR9","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2024.102810","volume":"117","author":"Ju Yakun","year":"2025","unstructured":"Yakun, Ju., Xiao, Ju., Zhang, Cong, et al.: Towards marine snow removal with fusing Fourier information. Inf Fusion 117(2025), 102810 (2025). https:\/\/doi.org\/10.1016\/j.inffus.2024.102810","journal-title":"Inf Fusion"},{"issue":"2","key":"1784_CR10","doi-asserted-by":"publisher","first-page":"1728","DOI":"10.1109\/TCSVT.2024.3482548","volume":"35","author":"Z Cheng","year":"2025","unstructured":"Cheng, Z., Fan, G., Zhou, J., Gan, M., Chen, C.L.P.: FDCE-Net: underwater image enhancement with embedding frequency and dual color encoder. IEEE Trans. Circuits Syst. Video Technol. 35(2), 1728\u20131744 (2025). https:\/\/doi.org\/10.1109\/TCSVT.2024.3482548","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"1784_CR11","unstructured":"Zhao, H., et al.: Relation-Shape CNN: enhancing 3D object representations with relational shape features (2019). arXiv preprint. arXiv:1904.10333. https:\/\/arxiv.org\/abs\/1904.10333"},{"issue":"4","key":"1784_CR12","doi-asserted-by":"publisher","first-page":"90","DOI":"10.1145\/3386569.3392412","volume":"39","author":"R Hanocka","year":"2020","unstructured":"Hanocka, R., et al.: MeshCNN: a network with an edge. ACM Trans. Graphics (TOG) 39(4), 90 (2020). https:\/\/doi.org\/10.1145\/3386569.3392412","journal-title":"ACM Trans. Graphics (TOG)"},{"key":"1784_CR13","doi-asserted-by":"publisher","first-page":"5756","DOI":"10.1109\/TIP.2021.3104978","volume":"30","author":"Z Han","year":"2021","unstructured":"Han, Z., et al.: Cross-view fusion for 3D object classification. IEEE Trans. Image Process. 30, 5756\u20135771 (2021). https:\/\/doi.org\/10.1109\/TIP.2021.3104978","journal-title":"IEEE Trans. Image Process."},{"key":"1784_CR14","doi-asserted-by":"publisher","DOI":"10.1007\/s00521-023-07305-9","author":"H Yu","year":"2022","unstructured":"Yu, H., et al.: MVDAN: multi-view dual attention network for 3D object recognition. Neural Comput. Appl. (2022). https:\/\/doi.org\/10.1007\/s00521-023-07305-9","journal-title":"Neural Comput. Appl."},{"key":"1784_CR15","doi-asserted-by":"publisher","unstructured":"Feng, Y., Zhang, Z., Zhao, X., Ji, R., Yue, G.: GVCNN: group-view convolutional neural networks for 3D shape recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 264\u2013272 (2018). https:\/\/doi.org\/10.1109\/CVPR.2018.00036","DOI":"10.1109\/CVPR.2018.00036"},{"key":"1784_CR16","unstructured":"Wang, C., Pelillo, M., Siddiqi, K.: Dominant set clustering and pooling for multi-view 3D object recognition (2019). arXiv preprint. arXiv:1906.01592. https:\/\/arxiv.org\/abs\/1906.01592"},{"key":"1784_CR17","doi-asserted-by":"publisher","unstructured":"Yu, T., Meng, J., Yuan, J.: Multi-view harmonized bilinear network for 3D object recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 186\u2013194 (2018). https:\/\/doi.org\/10.1109\/CVPR.2018.00028","DOI":"10.1109\/CVPR.2018.00028"},{"issue":"1","key":"1784_CR18","doi-asserted-by":"publisher","first-page":"192","DOI":"10.1109\/JOE.2024.3458110","volume":"50","author":"J Yakun","year":"2025","unstructured":"Yakun, Ju., Li, Li., Zhong, X., et al.: Underwater surface normal reconstruction via cross-grained photometric stereo transformer. IEEE J. Ocean. Eng. 50(1), 192\u2013203 (2025). https:\/\/doi.org\/10.1109\/JOE.2024.3458110","journal-title":"IEEE J. Ocean. Eng."},{"key":"1784_CR19","unstructured":"He, K., et al.: MvACon: Multi-View Attentive Contextualization for 3D object detection (2023). arXiv preprint. arXiv:2405.12200. https:\/\/arxiv.org\/abs\/2405.12200"},{"key":"1784_CR20","doi-asserted-by":"crossref","unstructured":"Wang, T., et al.: MVContrast: Unsupervised Pretraining for Multi-view 3D Object Recognition. (2023) arXiv preprint. arXiv:2306.05361. https:\/\/arxiv.org\/abs\/2306.05361","DOI":"10.1007\/s11633-023-1430-z"},{"key":"1784_CR21","doi-asserted-by":"publisher","unstructured":"Redmon, J., et al.: YOLOv1: You Only Look Once (2016). arXiv preprint. arXiv:1506.02640. https:\/\/doi.org\/10.48550\/arXiv.1506.02640","DOI":"10.48550\/arXiv.1506.02640"},{"key":"1784_CR22","doi-asserted-by":"publisher","unstructured":"Redmon, J., et al.: YOLOv2: Better, Faster, Stronger (2017). arXiv preprint. arXiv:1612.08242. https:\/\/doi.org\/10.48550\/arXiv.1612.08242","DOI":"10.48550\/arXiv.1612.08242"},{"key":"1784_CR23","doi-asserted-by":"publisher","unstructured":"Redmon, J., et al.: YOLOv3: An Incremental Improvement (2018). arXiv preprint. arXiv:1804.02767. https:\/\/doi.org\/10.48550\/arXiv.1804.02767","DOI":"10.48550\/arXiv.1804.02767"},{"key":"1784_CR24","doi-asserted-by":"publisher","unstructured":"Bochkovskiy, A., et al.: YOLOv4: Optimal Speed and Accuracy of Object Detection (2020). arXiv preprint. arXiv:2004.10934. https:\/\/doi.org\/10.48550\/arXiv.2004.10934","DOI":"10.48550\/arXiv.2004.10934"},{"key":"1784_CR25","doi-asserted-by":"publisher","unstructured":"Zhu, X., Lyu, S., Wang, X., et al.: TPH-YOLOv5: improved YOLOv5 based on transformer prediction head for object detection on drone-captured scenarios. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision Workshops (ICCVW), 2021, pp. 2778\u20132788 (2021). https:\/\/doi.org\/10.1109\/ICCVW54120.2021.00312","DOI":"10.1109\/ICCVW54120.2021.00312"},{"key":"1784_CR26","doi-asserted-by":"publisher","unstructured":"Boudjoghra, M., et al.: Open-YOLO 3D: A Fast and Accurate Open Vocabulary 3D Instance Segmentation Method (2024). arXiv:2406.02548. https:\/\/doi.org\/10.48550\/arXiv.2406.02548","DOI":"10.48550\/arXiv.2406.02548"},{"key":"1784_CR27","volume-title":"Classification and Regression Trees","author":"L Breiman","year":"1984","unstructured":"Breiman, L., Friedman, J., Olshen, R., Stone, C.: Classification and Regression Trees. Wadsworth International Group, Belmont (1984)"},{"key":"1784_CR28","volume-title":"The Elements of Statistical Learning","author":"J Friedman","year":"2001","unstructured":"Friedman, J., Hastie, T., Tibshirani, R.: The Elements of Statistical Learning. Springer, New York (2001)"},{"key":"1784_CR29","unstructured":"Redmon, J.: 2013\u20132016. 5 Darknet: Open Source Neural Networks in C. http:\/\/pjreddie.com\/darknet\/"},{"issue":"1","key":"1784_CR30","doi-asserted-by":"publisher","first-page":"81","DOI":"10.1023\/A:1022643204877","volume":"1","author":"JR Quinlan","year":"1986","unstructured":"Quinlan, J.R.: Induction of decision trees. Mach. Learn. 1(1), 81\u2013106 (1986). https:\/\/doi.org\/10.1023\/A:1022643204877","journal-title":"Mach. Learn."},{"key":"1784_CR31","doi-asserted-by":"publisher","first-page":"85","DOI":"10.1007\/s00138-022-01337-3","volume":"33","author":"Z Tang","year":"2022","unstructured":"Tang, Z., Li, J., Huang, J., et al.: Multi-scale convolution underwater image restoration network. Mach. Vis. Appl. 33, 85 (2022). https:\/\/doi.org\/10.1007\/s00138-022-01337-3","journal-title":"Mach. Vis. Appl."},{"key":"1784_CR32","doi-asserted-by":"publisher","unstructured":"Liu, C., Li, H., et al.: A dataset and benchmark of underwater object detection for robot picking. 2021 IEEE International Conference on Multimedia & Expo Workshops (ICMEW), pp. 1\u20136 (2020). https:\/\/doi.org\/10.1109\/ICMEW53276.2021.9455997","DOI":"10.1109\/ICMEW53276.2021.9455997"}],"container-title":["Journal of Real-Time Image Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11554-025-01784-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11554-025-01784-x\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11554-025-01784-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,10]],"date-time":"2025-11-10T07:04:41Z","timestamp":1762758281000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11554-025-01784-x"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,10,14]]},"references-count":32,"journal-issue":{"issue":"6","published-print":{"date-parts":[[2025,12]]}},"alternative-id":["1784"],"URL":"https:\/\/doi.org\/10.1007\/s11554-025-01784-x","relation":{},"ISSN":["1861-8200","1861-8219"],"issn-type":[{"type":"print","value":"1861-8200"},{"type":"electronic","value":"1861-8219"}],"subject":[],"published":{"date-parts":[[2025,10,14]]},"assertion":[{"value":"24 April 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"1 October 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"14 October 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"205"}}