{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,7]],"date-time":"2026-02-07T13:10:26Z","timestamp":1770469826780,"version":"3.49.0"},"reference-count":41,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2025,2,17]],"date-time":"2025-02-17T00:00:00Z","timestamp":1739750400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"},{"start":{"date-parts":[[2025,2,17]],"date-time":"2025-02-17T00:00:00Z","timestamp":1739750400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["U24A20263"],"award-info":[{"award-number":["U24A20263"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100010881","name":"Suzhou Municipal Science and Technology Bureau","doi-asserted-by":"publisher","award":["No. SYG202351"],"award-info":[{"award-number":["No. SYG202351"]}],"id":[{"id":"10.13039\/501100010881","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100010881","name":"Suzhou Municipal Science and Technology Bureau","doi-asserted-by":"publisher","award":["No. SYG202129"],"award-info":[{"award-number":["No. SYG202129"]}],"id":[{"id":"10.13039\/501100010881","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J Intell Robot Syst"],"DOI":"10.1007\/s10846-025-02234-3","type":"journal-article","created":{"date-parts":[[2025,2,17]],"date-time":"2025-02-17T20:35:14Z","timestamp":1739824514000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["An Inpainting SLAM Approach for Detecting and Recovering Regions with Dynamic Objects"],"prefix":"10.1007","volume":"111","author":[{"given":"Longxin","family":"Zhang","sequence":"first","affiliation":[]},{"given":"Benlian","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Siwen","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Brett","family":"Nener","sequence":"additional","affiliation":[]},{"given":"Xu","family":"Zhou","sequence":"additional","affiliation":[]},{"given":"Mingli","family":"Lu","sequence":"additional","affiliation":[]},{"given":"Xinya","family":"Li","sequence":"additional","affiliation":[]},{"given":"Shuting","family":"Le","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,2,17]]},"reference":[{"key":"2234_CR1","doi-asserted-by":"crossref","unstructured":"Engel, J., Sch\u00f6ps, T., Cremers, D.: Lsd-slam: Large-scale direct monocular slam. In: European Conference on Computer Vision, pp. 834\u2013849 (2014). Springer","DOI":"10.1007\/978-3-319-10605-2_54"},{"issue":"5","key":"2234_CR2","doi-asserted-by":"crossref","first-page":"1147","DOI":"10.1109\/TRO.2015.2463671","volume":"31","author":"R Mur-Artal","year":"2015","unstructured":"Mur-Artal, R., Montiel, J.M.M., Tardos, J.D.: Orb-slam: a versatile and accurate monocular slam system. IEEE Trans. Rob. 31(5), 1147\u20131163 (2015)","journal-title":"IEEE Trans. Rob."},{"issue":"10","key":"2234_CR3","volume":"35","author":"H Tao","year":"2024","unstructured":"Tao, H., Zheng, Y., Wang, Y., Qiu, J., Stojanovic, V.: Enhanced feature extraction yolo industrial small object detection algorithm based on receptive-field attention and multi-scale features. Meas. Sci. Technol. 35(10), 105023 (2024)","journal-title":"Meas. Sci. Technol."},{"key":"2234_CR4","volume":"225","author":"Z Li","year":"2023","unstructured":"Li, Z., Xu, B., Wu, D., Zhao, K., Chen, S., Lu, M., Cong, J.: A yolo-ggcnn based grasping framework for mobile robots in unknown environments. Expert Syst. Appl. 225, 119993 (2023)","journal-title":"Expert Syst. Appl."},{"key":"2234_CR5","doi-asserted-by":"crossref","unstructured":"Tateno, K., Tombari, F., Laina, I., Navab, N.: Cnn-slam: Real-time dense monocular slam with learned depth prediction. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6243\u20136252 (2017)","DOI":"10.1109\/CVPR.2017.695"},{"issue":"21","key":"2234_CR6","doi-asserted-by":"crossref","first-page":"25802","DOI":"10.1007\/s10489-023-04720-3","volume":"53","author":"B Ayman","year":"2023","unstructured":"Ayman, B., Malik, M., Lotfi, B.: Dam-slam: depth attention module in a semantic visual slam based on objects interaction for dynamic environments. Appl. Intell. 53(21), 25802\u201325815 (2023)","journal-title":"Appl. Intell."},{"issue":"4","key":"2234_CR7","doi-asserted-by":"crossref","first-page":"5279","DOI":"10.1109\/JSEN.2023.3345877","volume":"24","author":"C Ruan","year":"2023","unstructured":"Ruan, C., Zang, Q., Zhang, K., Huang, K.: Dn-slam: A visual slam with orb features and nerf mapping in dynamic environments. IEEE Sens. J. 24(4), 5279\u20135287 (2023)","journal-title":"IEEE Sens. J."},{"issue":"12","key":"2234_CR8","doi-asserted-by":"crossref","first-page":"13210","DOI":"10.1109\/JSEN.2023.3270534","volume":"23","author":"J He","year":"2023","unstructured":"He, J., Li, M., Wang, Y., Wang, H.: Ovd-slam: An online visual slam for dynamic environments. IEEE Sens. J. 23(12), 13210\u201313219 (2023)","journal-title":"IEEE Sens. J."},{"key":"2234_CR9","doi-asserted-by":"crossref","unstructured":"Bailey, T., Nieto, J., Guivant, J., Stevens, M., Nebot, E.: Consistency of the ekf-slam algorithm. In: 2006 IEEE\/RSJ International Conference on Intelligent Robots and Systems, pp. 3562\u20133568 (2006). IEEE","DOI":"10.1109\/IROS.2006.281644"},{"key":"2234_CR10","doi-asserted-by":"crossref","unstructured":"Klein, G., Murray, D.: Parallel tracking and mapping for small ar workspaces. In: 2007 6th IEEE and ACM International Symposium on Mixed and Augmented Reality, pp. 225\u2013234 (2007). IEEE","DOI":"10.1109\/ISMAR.2007.4538852"},{"issue":"6","key":"2234_CR11","doi-asserted-by":"crossref","first-page":"381","DOI":"10.1145\/358669.358692","volume":"24","author":"MA Fischler","year":"1981","unstructured":"Fischler, M.A., Bolles, R.C.: Random sample consensus: a paradigm for model fitting with applications to image analysis and automated cartography. Commun. ACM 24(6), 381\u2013395 (1981)","journal-title":"Commun. ACM"},{"issue":"2","key":"2234_CR12","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3177853","volume":"51","author":"MRU Saputra","year":"2018","unstructured":"Saputra, M.R.U., Markham, A., Trigoni, N.: Visual slam and structure from motion in dynamic environments: A survey. ACM Computing Surveys (CSUR) 51(2), 1\u201336 (2018)","journal-title":"ACM Computing Surveys (CSUR)"},{"issue":"2","key":"2234_CR13","doi-asserted-by":"crossref","first-page":"796","DOI":"10.1109\/LRA.2017.2653359","volume":"2","author":"R Mur-Artal","year":"2017","unstructured":"Mur-Artal, R., Tard\u00f3s, J.D.: Visual-inertial monocular slam with map reuse. IEEE Robotics and Automation Letters 2(2), 796\u2013803 (2017)","journal-title":"IEEE Robotics and Automation Letters"},{"issue":"6","key":"2234_CR14","doi-asserted-by":"crossref","first-page":"1874","DOI":"10.1109\/TRO.2021.3075644","volume":"37","author":"C Campos","year":"2021","unstructured":"Campos, C., Elvira, R., Rodr\u00edguez, J.J.G., Montiel, J.M., Tard\u00f3s, J.D.: Orbslam3: An accurate open-source library for visual, visual-inertial, and multimap slam. IEEE Transactions on Robotics 37(6), 1874\u20131890 (2021)","journal-title":"IEEE Transactions on Robotics"},{"key":"2234_CR15","doi-asserted-by":"crossref","first-page":"110","DOI":"10.1016\/j.robot.2016.11.012","volume":"89","author":"Y Sun","year":"2017","unstructured":"Sun, Y., Liu, M., Meng, M.Q.-H.: Improving rgb-d slam in dynamic environments: A motion removal approach. Robot. Auton. Syst. 89, 110\u2013122 (2017)","journal-title":"Robot. Auton. Syst."},{"key":"2234_CR16","doi-asserted-by":"crossref","unstructured":"Sun, Y., Liu, M., Meng, M.Q.-H.: Motion removal from moving platforms: An rgb-d data-based motion detection, tracking and segmentation approach. In: 2015 IEEE International Conference on Robotics and Biomimetics (ROBIO), pp. 1377\u2013 1382 (2015). IEEE","DOI":"10.1109\/ROBIO.2015.7418963"},{"issue":"1","key":"2234_CR17","doi-asserted-by":"crossref","first-page":"27","DOI":"10.1007\/s10846-023-02046-3","volume":"110","author":"J Li","year":"2024","unstructured":"Li, J., Dai, J., Su, Z., Zhu, C.: Rgb-d based visual slam algorithm for indoor crowd environment. Journal of Intelligent & Robotic Systems 110(1), 27 (2024)","journal-title":"Journal of Intelligent & Robotic Systems"},{"issue":"3","key":"2234_CR18","doi-asserted-by":"crossref","first-page":"101","DOI":"10.1007\/s10846-024-02128-w","volume":"110","author":"C Wang","year":"2024","unstructured":"Wang, C., Aouf, N.: Depth-enhanced deep learning approach for monocular camera based 3d object detection. Journal of Intelligent & Robotic Systems 110(3), 101 (2024)","journal-title":"Journal of Intelligent & Robotic Systems"},{"key":"2234_CR19","doi-asserted-by":"crossref","unstructured":"Kaneko, M., Iwami, K., Ogawa, T., Yamasaki, T., Aizawa, K.: Mask-slam: Robust feature-based monocular slam by masking using semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 258\u2013266 (2018)","DOI":"10.1109\/CVPRW.2018.00063"},{"key":"2234_CR20","doi-asserted-by":"crossref","unstructured":"Yu, C., Liu, Z., Liu, X.-J., Xie, F., Yang, Y., Wei, Q., Fei, Q.: Ds-slam: A semantic visual slam towards dynamic environments. In: 2018 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 1168\u20131174 (2018). IEEE","DOI":"10.1109\/IROS.2018.8593691"},{"issue":"12","key":"2234_CR21","doi-asserted-by":"crossref","first-page":"2481","DOI":"10.1109\/TPAMI.2016.2644615","volume":"39","author":"V Badrinarayanan","year":"2017","unstructured":"Badrinarayanan, V., Kendall, A., Cipolla, R.: Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 39(12), 2481\u20132495 (2017)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"2234_CR22","doi-asserted-by":"crossref","unstructured":"Wang, K., Lin, Y., Wang, L., Han, L., Hua, M., Wang, X., Lian, S., Huang, B.: A unified framework for mutual improvement of slam and semantic segmentation. In: 2019 International Conference on Robotics and Automation (ICRA), pp. 5224\u20135230 (2019). IEEE","DOI":"10.1109\/ICRA.2019.8793499"},{"issue":"24","key":"2234_CR23","doi-asserted-by":"crossref","first-page":"30009","DOI":"10.1007\/s10489-023-05070-w","volume":"53","author":"Y Wei","year":"2023","unstructured":"Wei, Y., Zhou, B., Duan, Y., Liu, J., An, D.: Do-slam: research and application of semantic slam system towards dynamic environments based on object detection. Appl. Intell. 53(24), 30009\u201330026 (2023)","journal-title":"Appl. Intell."},{"issue":"4","key":"2234_CR24","doi-asserted-by":"crossref","first-page":"4076","DOI":"10.1109\/LRA.2018.2860039","volume":"3","author":"B Bescos","year":"2018","unstructured":"Bescos, B., F\u00e1cil, J.M., Civera, J., Neira, J.: Dynaslam: Tracking, mapping, and inpainting in dynamic scenes. IEEE Robotics and Automation Letters 3(4), 4076\u20134083 (2018)","journal-title":"IEEE Robotics and Automation Letters"},{"key":"2234_CR25","doi-asserted-by":"crossref","unstructured":"He, K., Gkioxari, G., Doll\u00e1r, P., Girshick, R.: Mask r-cnn. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2961\u20132969 (2017)","DOI":"10.1109\/ICCV.2017.322"},{"issue":"2","key":"2234_CR26","doi-asserted-by":"crossref","first-page":"433","DOI":"10.1109\/TRO.2020.3031267","volume":"37","author":"B Bescos","year":"2020","unstructured":"Bescos, B., Cadena, C., Neira, J.: Empty cities: A dynamic-object-invariant space for visual slam. IEEE Trans. Rob. 37(2), 433\u2013451 (2020)","journal-title":"IEEE Trans. Rob."},{"issue":"16","key":"2234_CR27","doi-asserted-by":"crossref","first-page":"16628","DOI":"10.1109\/JSEN.2022.3190999","volume":"22","author":"Y Zhang","year":"2022","unstructured":"Zhang, Y., Liu, R., Yu, H., Zhou, B., Qian, K.: Visual loop closure detection with instance segmentation and image inpainting in dynamic scenes using wearable camera. IEEE Sens. J. 22(16), 16628\u201316637 (2022)","journal-title":"IEEE Sens. J."},{"issue":"5","key":"2234_CR28","doi-asserted-by":"crossref","first-page":"1255","DOI":"10.1109\/TRO.2017.2705103","volume":"33","author":"R Mur-Artal","year":"2017","unstructured":"Mur-Artal, R., Tard\u00f3s, J.D.: Orb-slam2: An open-source slam system for monocular, stereo, and rgb-d cameras. IEEE Transactions on Robotics 33(5), 1255\u20131262 (2017)","journal-title":"IEEE Transactions on Robotics"},{"key":"2234_CR29","doi-asserted-by":"crossref","unstructured":"Wang, C.-Y., Bochkovskiy, A., Liao, H.-Y.M.: Yolov7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 7464\u2013 7475 (2023)","DOI":"10.1109\/CVPR52729.2023.00721"},{"issue":"1","key":"2234_CR30","doi-asserted-by":"crossref","first-page":"1","DOI":"10.54254\/2755-2721\/1\/2022001","volume":"1","author":"Y Chen","year":"2022","unstructured":"Chen, Y., Alifu, K., Lin, W.: Ca-yolov5 for crowded pedestrian detection. Comput. Eng. Appl. 1(1), 1\u201310 (2022)","journal-title":"Comput. Eng. Appl."},{"key":"2234_CR31","doi-asserted-by":"crossref","unstructured":"Lin, T.-Y., Maire, M., Belongie, S., Hays, J., Perona, P., Ramanan, D., Doll\u00e1r, P., Zitnick, C.L.: Microsoft coco: Common objects in context. In: Computer Vision- ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6-12, 2014, Proceedings, Part V 13, pp. 740\u2013755 (2014). Springer","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"2234_CR32","doi-asserted-by":"crossref","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-net: Convolutional networks for biomedical image segmentation. In: Medical Image Computing and Computer-Assisted Intervention-MICCAI 2015: 18th International Conference, Munich, Germany, October 5-9, 2015, Proceedings, Part III 18, pp. 234\u2013241 (2015). Springer","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"2234_CR33","doi-asserted-by":"crossref","unstructured":"Zhou, Z., Rahman Siddiquee, M.M., Tajbakhsh, N., Liang, J.: Unet++: A nested u-net architecture for medical image segmentation. In: Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support: 4th International Workshop, DLMIA 2018, and 8th International Workshop, ML-CDS 2018, Held in Conjunction with MICCAI 2018, Granada, Spain, September 20, 2018, Proceedings 4, pp. 3\u201311 (2018). Springer","DOI":"10.1007\/978-3-030-00889-5_1"},{"key":"2234_CR34","doi-asserted-by":"crossref","unstructured":"Carreira, J., Zisserman, A.: Quo vadis, action recognition? a new model and the kinetics dataset. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6299\u20136308 (2017)","DOI":"10.1109\/CVPR.2017.502"},{"issue":"11","key":"2234_CR35","first-page":"6000","volume":"30","author":"A Vaswani","year":"2017","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, L., Polosukhin, I.: Attention is all you need. Adv. Neural. Inf. Process. Syst. 30(11), 6000\u20136010 (2017)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"2234_CR36","doi-asserted-by":"crossref","unstructured":"Li, Z., Lu, C.-Z., Qin, J., Guo, C.-L., Cheng, M.-M.: Towards an end-to-end framework for flow-guided video inpainting. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 17562\u201317571 (2022)","DOI":"10.1109\/CVPR52688.2022.01704"},{"key":"2234_CR37","doi-asserted-by":"crossref","unstructured":"Chang, Y.-L., Liu, Z.Y., Lee, K.-Y., Hsu, W.: Free-form video inpainting with 3d gated convolution and temporal patchgan. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 9066\u20139075 (2019)","DOI":"10.1109\/ICCV.2019.00916"},{"key":"2234_CR38","doi-asserted-by":"crossref","unstructured":"Sturm, J., Engelhard, N., Endres, F., Burgard, W., Cremers, D.: A benchmark for the evaluation of rgb-d slam systems. In: 2012 IEEE\/RSJ International Conference on Intelligent Robots and Systems, pp. 573\u2013580 (2012). IEEE","DOI":"10.1109\/IROS.2012.6385773"},{"issue":"18","key":"2234_CR39","doi-asserted-by":"crossref","first-page":"20657","DOI":"10.1109\/JSEN.2021.3099511","volume":"21","author":"T Ran","year":"2021","unstructured":"Ran, T., Yuan, L., Zhang, J., Tang, D., He, L.: Rs-slam: A robust semantic slam in dynamic environments based on rgb-d sensor. IEEE Sens. J. 21(18), 20657\u201320664 (2021)","journal-title":"IEEE Sens. J."},{"key":"2234_CR40","doi-asserted-by":"crossref","unstructured":"Xu, N., Yang, L., Fan, Y., Yang, J., Yue, D., Liang, Y., Price, B., Cohen, S., Huang, T.: Youtube-vos: Sequence-to-sequence video object segmentation. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 585\u2013 601 (2018)","DOI":"10.1007\/978-3-030-01228-1_36"},{"key":"2234_CR41","doi-asserted-by":"crossref","unstructured":"Li, J., Wang, N., Zhang, L., Du, B., Tao, D.: Recurrent feature reasoning for image inpainting. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 7760\u20137768 (2020)","DOI":"10.1109\/CVPR42600.2020.00778"}],"container-title":["Journal of Intelligent &amp; Robotic Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10846-025-02234-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10846-025-02234-3\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10846-025-02234-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,4,5]],"date-time":"2025-04-05T18:31:09Z","timestamp":1743877869000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10846-025-02234-3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,2,17]]},"references-count":41,"journal-issue":{"issue":"1","published-online":{"date-parts":[[2025,3]]}},"alternative-id":["2234"],"URL":"https:\/\/doi.org\/10.1007\/s10846-025-02234-3","relation":{},"ISSN":["1573-0409"],"issn-type":[{"value":"1573-0409","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,2,17]]},"assertion":[{"value":"26 August 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"1 February 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"17 February 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing Interests"}},{"value":"All authors have read and agreed to the published version of the manuscript.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethics approval"}}],"article-number":"29"}}