{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,11]],"date-time":"2025-09-11T19:12:38Z","timestamp":1757617958064,"version":"3.44.0"},"reference-count":66,"publisher":"Springer Science and Business Media LLC","issue":"4","license":[{"start":{"date-parts":[[2025,5,21]],"date-time":"2025-05-21T00:00:00Z","timestamp":1747785600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,5,21]],"date-time":"2025-05-21T00:00:00Z","timestamp":1747785600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["42074039","42074039","42074039"],"award-info":[{"award-number":["42074039","42074039","42074039"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Machine Vision and Applications"],"published-print":{"date-parts":[[2025,7]]},"DOI":"10.1007\/s00138-025-01701-z","type":"journal-article","created":{"date-parts":[[2025,5,21]],"date-time":"2025-05-21T13:05:39Z","timestamp":1747832739000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["YG-SLAM: dynamic environment-based geometric constraint point-line fusion visual SLAM system"],"prefix":"10.1007","volume":"36","author":[{"given":"Shuo","family":"Hu","sequence":"first","affiliation":[]},{"given":"Liye","family":"Zhao","sequence":"additional","affiliation":[]},{"given":"Qing","family":"Wang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,5,21]]},"reference":[{"issue":"3","key":"1701_CR1","doi-asserted-by":"publisher","first-page":"2909","DOI":"10.1109\/LRA.2024.3363534","volume":"9","author":"A Adkins","year":"2024","unstructured":"Adkins, A., Chen, T., Biswas, J.: ObVi-SLAM: Long-Term Object-Visual SLAM. IEEE Rob. Autom. Lett. 9(3), 2909\u20132916 (2024)","journal-title":"IEEE Rob. Autom. Lett."},{"key":"1701_CR2","doi-asserted-by":"crossref","unstructured":"Ma, Z., Yang, S., Li, J., et al.: Research on SLAM localization algorithm for orchard dynamic vision based on YOLOD-SLAM2. Agriculture.\u00a014(9), (2024)","DOI":"10.3390\/agriculture14091622"},{"key":"1701_CR3","doi-asserted-by":"publisher","first-page":"2046","DOI":"10.1109\/TRO.2024.3366815","volume":"40","author":"Y Ge","year":"2024","unstructured":"Ge, Y., Zhang, L., Wu, Y., et al.: PIPO-SLAM: Lightweight Visual-Inertial SLAM with preintegration merging theory and Pose-Only descriptions of multiple view Geometry. IEEE Trans. Robot. 40, 2046\u20132059 (2024)","journal-title":"IEEE Trans. Robot."},{"key":"1701_CR4","first-page":"104110","volume":"133","author":"H Wu","year":"2024","unstructured":"Wu, H., Zhan, S., Shao, X., et al.: SLG-SLAM: An integrated SLAM framework to improve accuracy using semantic information, laser and GNSS data. Int. J. Appl. Earth Obs. Geoinf. 133, 104110 (2024)","journal-title":"Int. J. Appl. Earth Obs. Geoinf."},{"issue":"11","key":"1701_CR5","doi-asserted-by":"publisher","first-page":"116304","DOI":"10.1088\/1361-6501\/ad6627","volume":"35","author":"J Wei","year":"2024","unstructured":"Wei, J., Deng, H., Wang, J., et al.: AFO-SLAM: An improved visual SLAM in dynamic scenes using acceleration of feature extraction And object detection. Meas. Sci. Technol. 35(11), 116304 (2024)","journal-title":"Meas. Sci. Technol."},{"key":"1701_CR6","doi-asserted-by":"crossref","unstructured":"Cong, P., Li, J., Liu, J., et al.: SEG-SLAM: Dynamic indoor RGB-D visual SLAM integrating geometric and YOLOv5-Based semantic Information. Sensors.\u00a024(7), (2024)","DOI":"10.3390\/s24072102"},{"issue":"6","key":"1701_CR7","doi-asserted-by":"publisher","first-page":"1874","DOI":"10.1109\/TRO.2021.3075644","volume":"37","author":"CCR E","year":"2021","unstructured":"E, C.C.R.: ORB-SLAM3: An accurate Open-Source library for visual, Visual\u2013Inertial, and multimap SLAM. IEEE Trans. Robot. 37(6), 1874\u20131890 (2021)","journal-title":"IEEE Trans. Robot."},{"issue":"3","key":"1701_CR8","doi-asserted-by":"publisher","first-page":"611","DOI":"10.1109\/TPAMI.2017.2658577","volume":"40","author":"J Engel","year":"2018","unstructured":"Engel, J., Koltun, V., Cremers, D.: Direct sparse Odometry. IEEE Trans. Pattern Anal. Mach. Intell. 40(3), 611\u2013625 (2018)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"2","key":"1701_CR9","doi-asserted-by":"publisher","first-page":"249","DOI":"10.1109\/TRO.2016.2623335","volume":"33","author":"C Forster","year":"2017","unstructured":"Forster, C., Zhang, Z., Gassner, M., et al.: SVO: Semidirect visual odometry for monocular and multicamera Systems. IEEE Trans. Robot. 33(2), 249\u2013265 (2017)","journal-title":"IEEE Trans. Robot."},{"issue":"4","key":"1701_CR10","doi-asserted-by":"publisher","first-page":"1181","DOI":"10.1007\/s11760-022-02325-w","volume":"17","author":"I Bakkouri","year":"2023","unstructured":"Bakkouri, I., Afdel, K.: MLCA2F: Multi-Level context attentional feature fusion for COVID-19 lesion segmentation from CT scans. Signal. Image Video Process. 17(4), 1181\u20131188 (2023)","journal-title":"Signal. Image Video Process."},{"issue":"8","key":"1701_CR11","doi-asserted-by":"publisher","first-page":"10743","DOI":"10.1007\/s11042-022-12242-2","volume":"81","author":"I Bakkouri","year":"2022","unstructured":"Bakkouri, I., Afdel, K., Benois-Pineau, J., et al.: BG-3DM2F: Bidirectional gated 3D multi-scale feature fusion for Alzheimer\u2019s disease diagnosis. Multimedia Tools Appl. 81(8), 10743\u201310776 (2022)","journal-title":"Multimedia Tools Appl."},{"issue":"29","key":"1701_CR12","doi-asserted-by":"publisher","first-page":"20483","DOI":"10.1007\/s11042-019-07988-1","volume":"79","author":"I Bakkouri","year":"2020","unstructured":"Bakkouri, I., Afdel, K.: Computer-aided diagnosis (CAD) system based on multi-layer feature fusion network for skin lesion recognition in dermoscopy images. Multimedia Tools Appl. 79(29), 20483\u201320518 (2020)","journal-title":"Multimedia Tools Appl."},{"issue":"10","key":"1701_CR13","doi-asserted-by":"publisher","first-page":"12939","DOI":"10.1007\/s11042-018-6267-z","volume":"78","author":"I Bakkouri","year":"2019","unstructured":"Bakkouri, I., Afdel, K.: Multi-scale CNN based on region proposals for efficient breast abnormality recognition. Multimedia Tools Appl. 78(10), 12939\u201312960 (2019)","journal-title":"Multimedia Tools Appl."},{"key":"1701_CR14","doi-asserted-by":"crossref","unstructured":"Kundu, A., Krishna, K.M., Sivaswamy, J.: Moving object detection by multi-view geometric techniques from a single camera mounted robot IEEE: St. Louis, MO, USA. pp. 4306\u20134312 (2009)","DOI":"10.1109\/IROS.2009.5354227"},{"key":"1701_CR15","doi-asserted-by":"crossref","unstructured":"Tan, W., Liu, H., Dong, Z., et al.: Robust monocular SLAM in dynamic environments IEEE: Adelaide, SA, Australia. pp. 209\u2013218 (2013)","DOI":"10.1109\/ISMAR.2013.6671781"},{"key":"1701_CR16","doi-asserted-by":"publisher","first-page":"110","DOI":"10.1016\/j.robot.2016.11.012","volume":"89","author":"Y Sun","year":"2017","unstructured":"Sun, Y., Liu, M., Meng, M.Q.H.: Improving RGB-D SLAM in dynamic environments: A motion removal approach. Robot. Auton. Syst. 89, 110\u2013122 (2017)","journal-title":"Robot. Auton. Syst."},{"issue":"4","key":"1701_CR17","doi-asserted-by":"publisher","first-page":"2263","DOI":"10.1109\/LRA.2017.2724759","volume":"2","author":"S Li","year":"2017","unstructured":"Li, S., Lee, D.: RGB-D SLAM in dynamic environments using static point Weighting. IEEE Rob. Autom. Lett. 2(4), 2263\u20132270 (2017)","journal-title":"IEEE Rob. Autom. Lett."},{"key":"1701_CR18","doi-asserted-by":"crossref","unstructured":"Wang, Y., Huang, S.: Towards dense moving object segmentation based robust dense RGB-D SLAM in dynamic scenarios IEEE: Singapore. pp. 1841\u20131846 (2014)","DOI":"10.1109\/ICARCV.2014.7064596"},{"key":"1701_CR19","doi-asserted-by":"crossref","unstructured":"Derome, M., Plyer, A., Sanfourche, M., et al.: Real-time mobile object detection using stereo IEEE: Singapore. pp. 1021\u20131026 (2014)","DOI":"10.1109\/ICARCV.2014.7064446"},{"issue":"04","key":"1701_CR20","doi-asserted-by":"publisher","first-page":"253","DOI":"10.1142\/S2301385015400026","volume":"03","author":"M Derome","year":"2015","unstructured":"Derome, M., Plyer, A., Sanfourche, M., et al.: Moving object detection in Real-Time using stereo from a mobile Platform. Unmanned Syst. 03(04), 253\u2013266 (2015)","journal-title":"Unmanned Syst."},{"key":"1701_CR21","doi-asserted-by":"crossref","unstructured":"Zhang, T., Zhang, H., Li, Y., et al.: FlowFusion: Dynamic Dense RGB-D SLAM Based on Optical Flow IEEE: Paris, France. pp. 7322\u20137328 (2020)","DOI":"10.1109\/ICRA40945.2020.9197349"},{"key":"1701_CR22","doi-asserted-by":"crossref","unstructured":"Sun, D., Yang, X., Liu, M., et al.: PWC-Net: CNNs for Optical Flow Using Pyramid, Warping, and Cost Volume IEEE: Salt Lake City, UT, USA. pp. 8934\u20138943 (2018)","DOI":"10.1109\/CVPR.2018.00931"},{"issue":"11","key":"1701_CR23","doi-asserted-by":"publisher","first-page":"e13866","DOI":"10.1111\/jfpe.13866","volume":"44","author":"W Ji","year":"2021","unstructured":"Ji, W., Gao, X., Xu, B., et al.: Apple target recognition method in complex environment based on improved YOLOv4. J. Food Process Eng. 44(11), e13866 (2021)","journal-title":"J. Food Process Eng."},{"key":"1701_CR24","doi-asserted-by":"crossref","unstructured":"Yu, C., Liu, Z., Liu, X., et al.: DS-SLAM: A Semantic Visual SLAM towards Dynamic Environments IEEE: Madrid, Spain. pp. 1168\u20131174 (2018)","DOI":"10.1109\/IROS.2018.8593691"},{"issue":"10","key":"1701_CR25","first-page":"2847","volume":"39","author":"ZH Xi","year":"2019","unstructured":"Xi, Z.H., Han, S.Q., Wang, H.X.: Simultaneous localization and semantic mapping of indoor dynamic scenes based on semantic segmentation. Comput. Appli. 39(10), 2847\u20132851 (2019)","journal-title":"Comput. Appli."},{"key":"1701_CR26","doi-asserted-by":"crossref","unstructured":"Zhao, H., Shi, J., Qi, X., et al.: Pyramid Scene Parsing Network IEEE: Honolulu, HI, USA. pp. 6230\u20136239 (2017)","DOI":"10.1109\/CVPR.2017.660"},{"key":"1701_CR27","doi-asserted-by":"publisher","first-page":"214685","DOI":"10.1109\/ACCESS.2020.3041038","volume":"8","author":"X Long","year":"2020","unstructured":"Long, X., Zhang, W., Zhao, B.: PSPNet-SLAM: A semantic SLAM detect dynamic object by pyramid scene parsing Network. IEEE Access. 8, 214685\u2013214695 (2020)","journal-title":"IEEE Access."},{"key":"1701_CR28","doi-asserted-by":"crossref","unstructured":"Peng, Y., Wang, A., Liu, J., et al.: A comparative study of semantic segmentation models for identification of grape with different Varieties. Agriculture.\u00a011(10), (2021)","DOI":"10.3390\/agriculture11100997"},{"issue":"4","key":"1701_CR29","doi-asserted-by":"publisher","first-page":"4076","DOI":"10.1109\/LRA.2018.2860039","volume":"3","author":"B Bescos","year":"2018","unstructured":"Bescos, B., F\u00e1cil, J.M., Civera, J., et al.: DynaSLAM: Tracking, mapping, and inpainting in dynamic Scenes. IEEE Rob. Autom. Lett. 3(4), 4076\u20134083 (2018)","journal-title":"IEEE Rob. Autom. Lett."},{"key":"1701_CR30","doi-asserted-by":"crossref","unstructured":"Liu, J., Abbas, I., Noor, R.S.: Development of deep Learning-Based variable rate agrochemical spraying system for targeted weeds control in strawberry Crop. Agronomy.\u00a011(8), (2021)","DOI":"10.3390\/agronomy11081480"},{"key":"1701_CR31","doi-asserted-by":"crossref","unstructured":"Ma, J., Zhao, Y., Fan, W., et al.: An improved YOLOv8 model for Lotus seedpod instance segmentation in the Lotus pond Environment. Agronomy.\u00a014(6), (2024)","DOI":"10.3390\/agronomy14061325"},{"key":"1701_CR32","doi-asserted-by":"crossref","unstructured":"Zhong, F., Wang, S., Zhang, Z., et al.: Detect-SLAM: Making Object Detection and SLAM Mutually Beneficial.\u00a0IEEE, Lake Tahoe, NV, USA (2018)","DOI":"10.1109\/WACV.2018.00115"},{"key":"1701_CR33","doi-asserted-by":"crossref","unstructured":"Zhang, Q., Chen, Q., Xu, W., et al.: Prediction of feed quantity for wheat combine harvester based on improved YOLOv5s and weight of single wheat plant without Stubble. Agriculture.\u00a014(8), (2024)","DOI":"10.3390\/agriculture14081251"},{"key":"1701_CR34","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Xu, X., Zhang, N., et al.: A semantic SLAM system for catadioptric panoramic cameras in dynamic Environments. Sensors.\u00a021(17), (2021)","DOI":"10.3390\/s21175889"},{"issue":"4","key":"1701_CR35","first-page":"390","volume":"31","author":"ZH Chen","year":"2023","unstructured":"Chen, Z.H., Wang, Z., Li, X.C.: VSLAM method based on depth camera in indoor dynamic scenes. J. Chin. Inert. Technol. 31(4), 390\u2013400 (2023)","journal-title":"J. Chin. Inert. Technol."},{"issue":"3","key":"1701_CR36","first-page":"242","volume":"50","author":"XY Du","year":"2024","unstructured":"Du, X.Y., Yuan, Q.N., Qi, J.Y., et al.: Visual SLAM method based on semantic segmentation in dynamic scenes. Comput. Eng. 50(3), 242\u2013249 (2024)","journal-title":"Comput. Eng."},{"key":"1701_CR37","doi-asserted-by":"publisher","first-page":"21160","DOI":"10.1109\/ACCESS.2022.3154086","volume":"10","author":"Z Hu","year":"2022","unstructured":"Hu, Z., Zhao, J., Luo, Y., et al.: Semantic SLAM based on improved DeepLabv3\u207a in dynamic Scenarios. IEEE Access. 10, 21160\u201321168 (2022)","journal-title":"IEEE Access."},{"key":"1701_CR38","doi-asserted-by":"crossref","unstructured":"Sandler, M., Howard, A., Zhu, M., et al.: MobileNetV2: Inverted Residuals and Linear Bottlenecks IEEE: Salt Lake City, UT, USA. pp. 4510\u20134520 (2018)","DOI":"10.1109\/CVPR.2018.00474"},{"key":"1701_CR39","doi-asserted-by":"crossref","unstructured":"Wang, Q., Qin, W., Liu, M., et al.: Semantic segmentation Model-Based boundary line recognition method for wheat Harvesting. Agriculture.\u00a014(10), (2024)","DOI":"10.3390\/agriculture14101846"},{"key":"1701_CR40","doi-asserted-by":"crossref","unstructured":"Zhang, Z., Lu, Y., Zhao, Y., et al.: TS-YOLO: An All-Day and lightweight tea canopy shoots detection Model. Agronomy.\u00a013(5), (2023)","DOI":"10.3390\/agronomy13051411"},{"key":"1701_CR41","unstructured":"Khanam, R., Hussain, M.: YOLOv11: An overview of the key architectural Enhancements. arXiv:2410.17725. (2024)"},{"issue":"2","key":"1701_CR42","doi-asserted-by":"publisher","first-page":"2628","DOI":"10.1364\/OE.480816","volume":"31","author":"H Tang","year":"2023","unstructured":"Tang, H., Liang, S., Yao, D., et al.: A visual defect detection for optics lens based on the YOLOv5 -C3CA-SPPF network model. Opt. Express 31(2), 2628\u20132643 (2023)","journal-title":"Opt. Express"},{"key":"1701_CR43","doi-asserted-by":"crossref","unstructured":"Zhang, T., Zhou, J., Liu, W., et al.: Seedling-YOLO: High-Efficiency target detection algorithm for field broccoli seedling transplanting quality based on YOLOv7-Tiny. Agronomy.\u00a014(5), (2024)","DOI":"10.3390\/agronomy14050931"},{"key":"1701_CR44","doi-asserted-by":"crossref","unstructured":"Ji, W., Wang, J., Xu, B., et al.: Apple grading based on Multi-Dimensional view processing and deep Learning. Foods 12(11), (2023)","DOI":"10.3390\/foods12112117"},{"key":"1701_CR45","doi-asserted-by":"crossref","unstructured":"Hou, Q., Zhou, D., Feng, J.: Coordinate Attention for Efficient Mobile Network Design IEEE: Nashville, TN, USA. pp. 13708\u201313717 (2021)","DOI":"10.1109\/CVPR46437.2021.01350"},{"issue":"3","key":"1701_CR46","doi-asserted-by":"publisher","first-page":"62","DOI":"10.1007\/s11554-024-01436-6","volume":"21","author":"H Li","year":"2024","unstructured":"Li, H., Li, J., Wei, H., et al.: Slim-neck by GSConv: A lightweight-design for real-time detector architectures. J. Real-Time Image Proc. 21(3), 62 (2024)","journal-title":"J. Real-Time Image Proc."},{"key":"1701_CR47","doi-asserted-by":"crossref","unstructured":"Wang, J., Chen, K., Xu, R., et al.: CARAFE: Content-Aware reassembly of FEatures IEEE: Seoul, Korea (South). pp. 3007\u20133016 (2019)","DOI":"10.1109\/ICCV.2019.00310"},{"issue":"8","key":"1701_CR48","doi-asserted-by":"publisher","first-page":"2011","DOI":"10.1109\/TPAMI.2019.2913372","volume":"42","author":"J Hu","year":"2020","unstructured":"Hu, J., Shen, L., Albanie, S., et al.: Squeeze-and-Excitation Networks. IEEE Trans. Pattern Anal. Mach. Intell. 42(8), 2011\u20132023 (2020)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1701_CR49","doi-asserted-by":"crossref","unstructured":"Zhao, S., Peng, Y., Liu, J., et al.: Tomato leaf disease diagnosis based on improved Convolution neural network by attention Module. Agriculture 11(7), (2021)","DOI":"10.3390\/agriculture11070651"},{"key":"1701_CR50","doi-asserted-by":"crossref","unstructured":"Wang, Q., Wu, B., Zhu, P., et al.: ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks. IEEE: Seattle, WA, USA. 11531\u201311539 (2020)","DOI":"10.1109\/CVPR42600.2020.01155"},{"key":"1701_CR51","doi-asserted-by":"crossref","unstructured":"Qin, Z., Zhang, P., Wu, F., et al.: FcaNet: Frequency Channel Attention Networks. IEEE: Montreal, QC, Canada. 763\u2013772 (2021)","DOI":"10.1109\/ICCV48922.2021.00082"},{"key":"1701_CR52","unstructured":"Zhang, H., Zu, K., Lu, J., et al.: EPSANet: An efficient pyramid split attention block on convolutional neural Network. arXiv:2105.14447\u00a0(2021)"},{"issue":"24","key":"1701_CR53","doi-asserted-by":"publisher","first-page":"30794","DOI":"10.1109\/JSEN.2023.3329524","volume":"23","author":"C Hu","year":"2023","unstructured":"Hu, C., Zhang, X., Li, K., et al.: EM-LSD-Based Visual-Inertial odometry with Point-Line Feature. IEEE Sens. J. 23(24), 30794\u201330804 (2023)","journal-title":"IEEE Sens. J."},{"key":"1701_CR54","doi-asserted-by":"publisher","first-page":"111245","DOI":"10.1109\/ACCESS.2019.2934521","volume":"7","author":"R Dong","year":"2019","unstructured":"Dong, R., Wei, Z., Liu, C., et al.: A novel loop closure detection method using line Features. IEEE Access. 7, 111245\u2013111256 (2019)","journal-title":"IEEE Access."},{"issue":"14","key":"1701_CR55","doi-asserted-by":"publisher","first-page":"23265","DOI":"10.1109\/JSEN.2024.3405940","volume":"24","author":"X He","year":"2024","unstructured":"He, X., Ding, L., Lan, Y.: DSK-SLAM: A dynamic SLAM system combining semantic information and a novel geometric method based on K-Means Clustering. IEEE Sens. J. 24(14), 23265\u201323279 (2024)","journal-title":"IEEE Sens. J."},{"issue":"1","key":"1701_CR56","doi-asserted-by":"publisher","first-page":"98","DOI":"10.1007\/s11263-014-0733-5","volume":"111","author":"M Everingham","year":"2015","unstructured":"Everingham, M., Eslami, S.M.A., Van Gool, L., et al.: The Pascal visual object classes challenge: A Retrospective. Int. J. Comput. Vision 111(1), 98\u2013136 (2015)","journal-title":"Int. J. Comput. Vision"},{"key":"1701_CR57","doi-asserted-by":"publisher","first-page":"125931","DOI":"10.1109\/ACCESS.2024.3424501","volume":"12","author":"W Zhen","year":"2024","unstructured":"Zhen, W., Weiwei, H., Wenlei, Y., et al.: Based on BiSeNetV2 for semantic SLAM in dynamic Scenes. IEEE Access. 12, 125931\u2013125941 (2024)","journal-title":"IEEE Access."},{"key":"1701_CR58","doi-asserted-by":"crossref","unstructured":"Palazzolo, E., Behley, J., Lottes, P., et al.: ReFusion: 3D Reconstruction in Dynamic Environments for RGB-D Cameras Exploiting Residuals IEEE: Macau, China. pp. 7855\u20137862 (2019)","DOI":"10.1109\/IROS40897.2019.8967590"},{"key":"1701_CR59","doi-asserted-by":"crossref","unstructured":"Guan, H., Qian, C., Wu, T., et al.: A dynamic scene vision SLAM method incorporating object detection and object Characterization. Sustainability 15(4), (2023)","DOI":"10.3390\/su15043048"},{"issue":"4","key":"1701_CR60","first-page":"477","volume":"5","author":"JM Li","year":"2023","unstructured":"Li, J.M., Jie, M.Y., Zhang, M., et al.: Visual SLAM system based on semantic information and geometric constraints in dynamic environment. J. Intell. Sci. Technol. 5(4), 477\u2013485 (2023)","journal-title":"J. Intell. Sci. Technol."},{"key":"1701_CR61","first-page":"1","volume":"72","author":"W Wei","year":"2023","unstructured":"Wei, W., Huang, K., Liu, X., et al.: GSL-VO: A Geometric-Semantic information enhanced lightweight visual odometry in dynamic Environments. IEEE Trans. Instrum. Meas. 72, 1\u201313 (2023)","journal-title":"IEEE Trans. Instrum. Meas."},{"issue":"2","key":"1701_CR62","doi-asserted-by":"publisher","first-page":"799","DOI":"10.1007\/s00371-024-03365-8","volume":"41","author":"H Zhu","year":"2025","unstructured":"Zhu, H., Bai, S., Shi, J., et al.: IOFusion: Instance segmentation and optical-flow guided 3D reconstruction in dynamic scenes. Visual Comput. 41(2), 799\u2013813 (2025)","journal-title":"Visual Comput."},{"key":"1701_CR63","doi-asserted-by":"crossref","unstructured":"Zhu, H., Bai, S., Shi, J., et al.: Ellipsoid-SLAM: Enhancing Dynamic Scene Understanding Through Ellipsoidal Object Representation and Trajectory tracking. Visual Comput. (2025)","DOI":"10.1007\/s00371-025-03819-7"},{"issue":"8","key":"1701_CR64","doi-asserted-by":"publisher","first-page":"6011","DOI":"10.1007\/s00521-021-06764-3","volume":"34","author":"W Wu","year":"2022","unstructured":"Wu, W., Guo, L., Gao, H., et al.: YOLO-SLAM: A semantic SLAM system towards dynamic environment with geometric constraint. Neural Comput. Appl. 34(8), 6011\u20136026 (2022)","journal-title":"Neural Comput. Appl."},{"issue":"1","key":"1701_CR65","doi-asserted-by":"publisher","first-page":"27","DOI":"10.1007\/s10846-023-02046-3","volume":"110","author":"J Li","year":"2024","unstructured":"Li, J., Dai, J., Su, Z., et al.: RGB-D based visual SLAM algorithm for indoor crowd Environment. J. Intell. Robotic Syst. 110(1), 27 (2024)","journal-title":"J. Intell. Robotic Syst."},{"key":"1701_CR66","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TIM.2023.3326234","volume":"72","author":"S Cheng","year":"2023","unstructured":"Cheng, S., Sun, C., Zhang, S., et al.: SG-SLAM: A Real-Time RGB-D visual SLAM toward dynamic scenes with semantic and geometric Information. IEEE Trans. Instrum. Meas. 72, 1\u201312 (2023)","journal-title":"IEEE Trans. Instrum. Meas."}],"container-title":["Machine Vision and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00138-025-01701-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00138-025-01701-z\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00138-025-01701-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,6]],"date-time":"2025-09-06T15:12:42Z","timestamp":1757171562000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00138-025-01701-z"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,5,21]]},"references-count":66,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2025,7]]}},"alternative-id":["1701"],"URL":"https:\/\/doi.org\/10.1007\/s00138-025-01701-z","relation":{},"ISSN":["0932-8092","1432-1769"],"issn-type":[{"type":"print","value":"0932-8092"},{"type":"electronic","value":"1432-1769"}],"subject":[],"published":{"date-parts":[[2025,5,21]]},"assertion":[{"value":"12 December 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"28 February 2025","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"29 April 2025","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"21 May 2025","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}},{"value":"Not applicable.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Institutional Review Board Statement"}},{"value":"Not applicable.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Informed Consent Statement"}}],"article-number":"82"}}