{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,19]],"date-time":"2026-01-19T07:42:57Z","timestamp":1768808577542,"version":"3.49.0"},"reference-count":76,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2023,5,10]],"date-time":"2023-05-10T00:00:00Z","timestamp":1683676800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2023,5,10]],"date-time":"2023-05-10T00:00:00Z","timestamp":1683676800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int J Comput Intell Syst"],"abstract":"<jats:sec><jats:title>Abstract<\/jats:title><jats:p>In recent years, the development of ground robots with human-like perception capabilities has led to the use of multiple sensors, including cameras, lidars, and radars, along with deep learning techniques for detecting and recognizing objects and estimating distances. This paper proposes a computer vision-based navigation system that integrates object detection, segmentation, and monocular depth estimation using deep neural networks to identify predefined target objects and navigate towards them with a single monocular camera as a sensor. Our experiments include different sensitivity analyses to evaluate the impact of monocular cues on distance estimation. We show that this system can provide a ground robot with the perception capabilities needed for autonomous navigation in unknown indoor environments without the need for prior mapping or external positioning systems. This technique provides an efficient and cost-effective means of navigation, overcoming the limitations of other navigation techniques such as GPS-based and SLAM-based navigation.<\/jats:p><\/jats:sec><jats:sec><jats:title>Graphical Abstract<\/jats:title><\/jats:sec>","DOI":"10.1007\/s44196-023-00250-5","type":"journal-article","created":{"date-parts":[[2023,5,11]],"date-time":"2023-05-11T06:36:37Z","timestamp":1683786997000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":8,"title":["Monocular Based Navigation System for Autonomous Ground Robots Using Multiple Deep Learning Models"],"prefix":"10.1007","volume":"16","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-8002-5501","authenticated-orcid":false,"given":"Zakariae","family":"Machkour","sequence":"first","affiliation":[]},{"given":"Daniel","family":"Ortiz-Arroyo","sequence":"additional","affiliation":[]},{"given":"Petar","family":"Durdevic","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,5,10]]},"reference":[{"issue":"2","key":"250_CR1","doi-asserted-by":"publisher","first-page":"237","DOI":"10.1109\/34.982903","volume":"24","author":"GN DeSouza","year":"2002","unstructured":"DeSouza, G.N., Kak, A.C.: Vision for mobile robot navigation: a survey. IEEE Trans. Pattern Anal. Mach. Intell. 24(2), 237\u2013267 (2002)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"2","key":"250_CR2","doi-asserted-by":"publisher","first-page":"172988141983959","DOI":"10.1177\/1729881419839596","volume":"16","author":"F Rubio","year":"2019","unstructured":"Rubio, F., Valero, F., Llopis-Albert, C.: A review of mobile robots: concepts, methods, theoretical framework, and applications. Int. J. Adv. Robot. Syst. 16(2), 1729881419839596 (2019)","journal-title":"Int. J. Adv. Robot. Syst."},{"issue":"1","key":"250_CR3","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s12046-017-0776-8","volume":"43","author":"B Madhevan","year":"2018","unstructured":"Madhevan, B., Sreekumar, M.: Identification of probabilistic approaches and map-based navigation in motion planning for mobile robots. S\u0101dhan\u0101 43(1), 1\u201318 (2018)","journal-title":"S\u0101dhan\u0101"},{"key":"250_CR4","doi-asserted-by":"crossref","unstructured":"Engel, J., Sch\u00f6ps, T., Cremers, D.: Lsd-slam: large-scale direct monocular slam. In: European Conference on Computer Vision, pp. 834\u2013849. Springer (2014)","DOI":"10.1007\/978-3-319-10605-2_54"},{"issue":"5","key":"250_CR5","doi-asserted-by":"publisher","first-page":"1147","DOI":"10.1109\/TRO.2015.2463671","volume":"31","author":"R Mur-Artal","year":"2015","unstructured":"Mur-Artal, R., Montiel, J.M.M., Tardos, J.D.: Orb-slam: a versatile and accurate monocular slam system. IEEE Trans. Robot. 31(5), 1147\u20131163 (2015)","journal-title":"IEEE Trans. Robot."},{"key":"250_CR6","unstructured":"Anderson, P., Chang, A., Chaplot, D.S., Dosovitskiy, A., Gupta, S., Koltun, V., Kosecka, J., Malik, J., Mottaghi, R., Savva, M., et al.: On evaluation of embodied navigation agents. Preprint at arXiv:1807.06757 (2018)"},{"key":"250_CR7","doi-asserted-by":"crossref","unstructured":"Zhang, S., Gong, Z., Tao, B., Ding, H.: A visual servoing method based on point cloud. In: 2020 IEEE International Conference on Real-time Computing and Robotics (RCAR), pp. 369\u2013374. IEEE (2020)","DOI":"10.1109\/RCAR49640.2020.9303277"},{"key":"250_CR8","doi-asserted-by":"crossref","unstructured":"de Villiers, F., Brink, W.: Learning fine-grained control for mapless navigation. In: 2020 International SAUPEC\/RobMech\/PRASA Conference, pp. 1\u20136. IEEE (2020)","DOI":"10.1109\/SAUPEC\/RobMech\/PRASA48453.2020.9041011"},{"key":"250_CR9","doi-asserted-by":"publisher","DOI":"10.1016\/j.compeleceng.2022.107730","volume":"98","author":"J Zhang","year":"2022","unstructured":"Zhang, J., Sun, J., Wang, J., Li, Z., Chen, X.: An object tracking framework with recapture based on correlation filters and siamese networks. Comput. Electr. Eng. 98, 107730 (2022)","journal-title":"Comput. Electr. Eng."},{"issue":"7","key":"250_CR10","doi-asserted-by":"publisher","first-page":"2068","DOI":"10.3390\/s20072068","volume":"20","author":"C Debeunne","year":"2020","unstructured":"Debeunne, C., Vivet, D.: A review of visual-lidar fusion based simultaneous localization and mapping. Sensors 20(7), 2068 (2020)","journal-title":"Sensors"},{"key":"250_CR11","doi-asserted-by":"crossref","unstructured":"Cen, S.H., Newman, P.: Precise ego-motion estimation with millimeter-wave radar under diverse and challenging conditions. In: 2018 IEEE International Conference on Robotics and Automation (ICRA), pp. 6045\u20136052. IEEE (2018)","DOI":"10.1109\/ICRA.2018.8460687"},{"issue":"6","key":"250_CR12","doi-asserted-by":"publisher","first-page":"1389","DOI":"10.3390\/s19061389","volume":"19","author":"JH Rhee","year":"2019","unstructured":"Rhee, J.H., Seo, J.: Low-cost curb detection and localization system using multiple ultrasonic sensors. Sensors 19(6), 1389 (2019)","journal-title":"Sensors"},{"key":"250_CR13","unstructured":"Haseeb, M.A., Guan, J., Ristic-Durrant, D., Gr\u00e4ser, A.: Disnet: a novel method for distance estimation from monocular camera. In: 10th Planning, Perception and Navigation for Intelligent Vehicles (PPNIV18), IROS (2018)"},{"issue":"5","key":"250_CR14","doi-asserted-by":"publisher","first-page":"1437","DOI":"10.3390\/s20051437","volume":"20","author":"P Durdevic","year":"2020","unstructured":"Durdevic, P., Ortiz-Arroyo, D.: A deep neural network sensor for visual servoing in 3d spaces. Sensors 20(5), 1437 (2020)","journal-title":"Sensors"},{"key":"250_CR15","doi-asserted-by":"crossref","unstructured":"Cantrell, K.J., Miller, C.D., Morato, C.: Practical depth estimation with image segmentation and serial u-nets. In: VEHITS, pp. 406\u2013414 (2020)","DOI":"10.5220\/0009781804060414"},{"key":"250_CR16","unstructured":"Bhoi, A.: Monocular depth estimation: a survey. Preprint at arXiv:1901.09402 (2019)"},{"key":"250_CR17","doi-asserted-by":"publisher","first-page":"1612","DOI":"10.1007\/s11431-020-1582-8","volume":"63","author":"C Zhao","year":"2020","unstructured":"Zhao, C., Sun, Q., Zhang, C., Tang, Y., Qian, F.: Monocular depth estimation based on deep learning: an overview. Sci. China Technol. Sci. 63, 1612\u20131627 (2020)","journal-title":"Sci. China Technol. Sci."},{"key":"250_CR18","doi-asserted-by":"publisher","first-page":"14","DOI":"10.1016\/j.neucom.2020.12.089","volume":"438","author":"Y Ming","year":"2021","unstructured":"Ming, Y., Meng, X., Fan, C., Yu, H.: Deep learning for monocular depth estimation: a review. Neurocomputing 438, 14\u201333 (2021)","journal-title":"Neurocomputing"},{"key":"250_CR19","first-page":"2366","volume":"27","author":"D Eigen","year":"2014","unstructured":"Eigen, D., Puhrsch, C., Fergus, R.: Depth map prediction from a single image using a multi-scale deep network. Adv. Neural Inf. Process. Syst. 27, 2366\u20132374 (2014)","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"250_CR20","doi-asserted-by":"crossref","unstructured":"Godard, C., Mac\u00a0Aodha, O., Firman, M., Brostow, G.J.: Digging into self-supervised monocular depth estimation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 3828\u20133838 (2019)","DOI":"10.1109\/ICCV.2019.00393"},{"key":"250_CR21","doi-asserted-by":"crossref","unstructured":"Silberman, N., Hoiem, D., Kohli, P., Fergus, R.: Indoor segmentation and support inference from rgbd images. In: European Conference on Computer Vision, pp. 746\u2013760. Springer (2012)","DOI":"10.1007\/978-3-642-33715-4_54"},{"issue":"11","key":"250_CR22","doi-asserted-by":"publisher","first-page":"1231","DOI":"10.1177\/0278364913491297","volume":"32","author":"A Geiger","year":"2013","unstructured":"Geiger, A., Lenz, P., Stiller, C., Urtasun, R.: Vision meets robotics: the kitti dataset. Int. J. Robot. Res. 32(11), 1231\u20131237 (2013)","journal-title":"Int. J. Robot. Res."},{"key":"250_CR23","doi-asserted-by":"crossref","unstructured":"Dijk, T.V., Croon, G.D.: How do neural networks see depth in single images? In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2183\u20132191 (2019)","DOI":"10.1109\/ICCV.2019.00227"},{"key":"250_CR24","doi-asserted-by":"crossref","unstructured":"Xu, D., Wang, W., Tang, H., Liu, H., Sebe, N., Ricci, E.: Structured attention guided convolutional neural fields for monocular depth estimation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3917\u20133925 (2018)","DOI":"10.1109\/CVPR.2018.00412"},{"key":"250_CR25","doi-asserted-by":"crossref","unstructured":"Fu, H., Gong, M., Wang, C., Batmanghelich, K., Tao, D.: Deep ordinal regression network for monocular depth estimation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2002\u20132011 (2018)","DOI":"10.1109\/CVPR.2018.00214"},{"key":"250_CR26","unstructured":"Alhashim, I., Wonka, P.: High quality monocular depth estimation via transfer learning. Preprint at arXiv:1812.11941 (2018)"},{"key":"250_CR27","doi-asserted-by":"crossref","unstructured":"Huang, G., Liu, Z., Van Der\u00a0Maaten, L., Weinberger, K.Q.: Densely connected convolutional networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4700\u20134708 (2017)","DOI":"10.1109\/CVPR.2017.243"},{"key":"250_CR28","doi-asserted-by":"crossref","unstructured":"Hara, K., Kataoka, H., Satoh, Y.: Can spatiotemporal 3d cnns retrace the history of 2d cnns and imagenet? In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6546\u20136555 (2018)","DOI":"10.1109\/CVPR.2018.00685"},{"key":"250_CR29","unstructured":"Lee, J.H., Han, M.-K., Ko, D.W., Suh, I.H.: From big to small: multi-scale local planar guidance for monocular depth estimation. Preprint at arXiv:1907.10326 (2019)"},{"key":"250_CR30","doi-asserted-by":"crossref","unstructured":"Wofk, D., Ma, F., Yang, T.-J., Karaman, S., Sze, V.: Fastdepth: fast monocular depth estimation on embedded systems. In: 2019 International Conference on Robotics and Automation (ICRA), pp. 6101\u20136108. IEEE (2019)","DOI":"10.1109\/ICRA.2019.8794182"},{"key":"250_CR31","unstructured":"TeslaAI: Tesla AI,\u201d. https:\/\/www.tesla.com\/AI (2022). Accessed 6 Dec 2021"},{"key":"250_CR32","doi-asserted-by":"crossref","unstructured":"Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: unified, real-time object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 779\u2013788 (2016)","DOI":"10.1109\/CVPR.2016.91"},{"key":"250_CR33","doi-asserted-by":"crossref","unstructured":"Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C.-Y., Berg, A.C.: Ssd: single shot multibox detector. In: European Conference on Computer Vision, pp. 21\u201337. Springer (2016)","DOI":"10.1007\/978-3-319-46448-0_2"},{"key":"250_CR34","doi-asserted-by":"crossref","unstructured":"Girshick, R.: Fast r-cnn. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1440\u20131448 (2015)","DOI":"10.1109\/ICCV.2015.169"},{"key":"250_CR35","first-page":"91","volume":"28","author":"S Ren","year":"2015","unstructured":"Ren, S., He, K., Girshick, R., Sun, J.: Faster r-cnn: towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 28, 91\u201399 (2015)","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"250_CR36","doi-asserted-by":"crossref","unstructured":"He, K., Gkioxari, G., Doll\u00e1r, P., Girshick, R.: Mask r-cnn. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2961\u20132969 (2017)","DOI":"10.1109\/ICCV.2017.322"},{"key":"250_CR37","doi-asserted-by":"crossref","unstructured":"Lu, X., Li, Q., Li, B., Yan, J.: Mimicdet: Bridging the gap between one-stage and two-stage object detection. In: European Conference on Computer Vision, pp. 541\u2013557. Springer (2020)","DOI":"10.1007\/978-3-030-58568-6_32"},{"key":"250_CR38","doi-asserted-by":"publisher","first-page":"90330","DOI":"10.1109\/ACCESS.2021.3090961","volume":"9","author":"P Kumar","year":"2021","unstructured":"Kumar, P., Sharma, A., Kota, S.R.: Automatic multiclass instance segmentation of concrete damage using deep learning model. IEEE Access 9, 90330\u201390345 (2021)","journal-title":"IEEE Access"},{"issue":"8","key":"250_CR39","doi-asserted-by":"publisher","first-page":"1288","DOI":"10.3390\/rs12081288","volume":"12","author":"JR G Braga","year":"2020","unstructured":"G Braga, J.R., Peripato, V., Dalagnol, R., P Ferreira, M., Tarabalka, Y., OC Arag\u00e3o, L.E., de F Campos, Velho H., Shiguemori, E.H., Wagner, F.H.: Tree crown delineation algorithm based on a convolutional neural network. Remote Sens. 12(8), 1288 (2020)","journal-title":"Remote Sens."},{"issue":"24","key":"250_CR40","doi-asserted-by":"publisher","first-page":"4104","DOI":"10.3390\/rs12244104","volume":"12","author":"AJ Chadwick","year":"2020","unstructured":"Chadwick, A.J., Goodbody, T.R., Coops, N.C., Hervieux, A., Bater, C.W., Martens, L.A., White, B., R\u00f6eser, D.: Automatic delineation and height measurement of regenerating conifer crowns under leaf-off conditions using uav imagery. Remote Sens. 12(24), 4104 (2020)","journal-title":"Remote Sens."},{"issue":"7","key":"250_CR41","doi-asserted-by":"publisher","first-page":"753","DOI":"10.1049\/iet-its.2019.0376","volume":"14","author":"C Shen","year":"2020","unstructured":"Shen, C., Zhao, X., Liu, Z., Gao, T., Xu, J.: Joint vehicle detection and distance prediction via monocular depth estimation. IET Intell. Transp. Syst. 14(7), 753\u2013763 (2020)","journal-title":"IET Intell. Transp. Syst."},{"key":"250_CR42","doi-asserted-by":"publisher","first-page":"123107","DOI":"10.1109\/ACCESS.2019.2936126","volume":"7","author":"C Zhou","year":"2019","unstructured":"Zhou, C., Liu, Y., Sun, Q., Lasang, P.: Joint object detection and depth estimation in multiplexed image. IEEE Access 7, 123107\u2013123115 (2019)","journal-title":"IEEE Access"},{"key":"250_CR43","unstructured":"Redmon, J., Farhadi, A.: Yolov3: an incremental improvement. Preprint at arXiv:1804.02767 (2018)"},{"issue":"4","key":"250_CR44","doi-asserted-by":"publisher","first-page":"61","DOI":"10.3390\/jimaging7040061","volume":"7","author":"D Urban","year":"2021","unstructured":"Urban, D., Caplier, A.: Time-and resource-efficient time-to-collision forecasting for indoor pedestrian obstacles avoidance. J. Imaging 7(4), 61 (2021)","journal-title":"J. Imaging"},{"issue":"2","key":"250_CR45","doi-asserted-by":"publisher","first-page":"836","DOI":"10.1109\/TIP.2016.2621673","volume":"26","author":"Y Cao","year":"2016","unstructured":"Cao, Y., Shen, C., Shen, H.T.: Exploiting depth from single monocular images for object detection and semantic segmentation. IEEE Trans. Image Process. 26(2), 836\u2013846 (2016)","journal-title":"IEEE Trans. Image Process."},{"key":"250_CR46","unstructured":"Li, S., Xu, C., Xing, J., Ning, Y., Chen, Y.: Sdod: real-time segmenting and detecting 3d object by depth. Preprint at arXiv:2001.09425 (2020)"},{"issue":"9","key":"250_CR47","doi-asserted-by":"publisher","first-page":"2478","DOI":"10.1109\/TMM.2018.2798282","volume":"20","author":"B Kang","year":"2018","unstructured":"Kang, B., Lee, Y., Nguyen, T.Q.: Depth-adaptive deep neural network for semantic segmentation. IEEE Trans. Multimed. 20(9), 2478\u20132490 (2018)","journal-title":"IEEE Trans. Multimed."},{"issue":"10","key":"250_CR48","doi-asserted-by":"publisher","first-page":"2024","DOI":"10.1109\/TPAMI.2015.2505283","volume":"38","author":"F Liu","year":"2015","unstructured":"Liu, F., Shen, C., Lin, G., Reid, I.: Learning depth from single monocular images using deep convolutional neural fields. IEEE Trans. Pattern Anal. Mach. Intell. 38(10), 2024\u20132039 (2015)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"250_CR49","unstructured":"Lafferty, J.D., McCallum, A., Pereira, F.C.N.: Conditional random fields: probabilistic models for segmenting and labeling sequence data. In: Proceedings of the Eighteenth International Conference on Machine Learning. ICML \u201901, pp. 282\u2013289. Morgan Kaufmann Publishers Inc., San Francisco (2001)"},{"key":"250_CR50","doi-asserted-by":"crossref","unstructured":"Girshick, R., Donahue, J., Darrell, T., Malik, J.: Rich feature hierarchies for accurate object detection and semantic segmentation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 580\u2013587 (2014)","DOI":"10.1109\/CVPR.2014.81"},{"key":"250_CR51","unstructured":"Girshick, R., Radosavovic, I., Gkioxari, G., Doll\u00e1r, P., He, K.: Detectron. https:\/\/github.com\/facebookresearch\/detectron (2018)"},{"issue":"8","key":"250_CR52","doi-asserted-by":"publisher","first-page":"1795","DOI":"10.3390\/s19081795","volume":"19","author":"X Lin","year":"2019","unstructured":"Lin, X., S\u00e1nchez-Escobedo, D., Casas, J.R., Pard\u00e0s, M.: Depth estimation and semantic segmentation from a single rgb image using a hybrid convolutional neural network. Sensors 19(8), 1795 (2019)","journal-title":"Sensors"},{"key":"250_CR53","doi-asserted-by":"crossref","unstructured":"Wang, Y., Chao, W.-L., Garg, D., Hariharan, B., Campbell, M., Weinberger, K.Q.: Pseudo-lidar from visual depth estimation: Bridging the gap in 3d object detection for autonomous driving. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8445\u20138453 (2019)","DOI":"10.1109\/CVPR.2019.00864"},{"key":"250_CR54","doi-asserted-by":"crossref","unstructured":"Zhang, S., Wen, L., Bian, X., Lei, Z., Li, S.Z.: Single-shot refinement neural network for object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4203\u20134212 (2018)","DOI":"10.1109\/CVPR.2018.00442"},{"key":"250_CR55","doi-asserted-by":"crossref","unstructured":"Godard, C., Mac\u00a0Aodha, O., Brostow, G.J.: Unsupervised monocular depth estimation with left-right consistency. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 270\u2013279 (2017)","DOI":"10.1109\/CVPR.2017.699"},{"key":"250_CR56","doi-asserted-by":"crossref","unstructured":"Atapour-Abarghouei, A., Breckon, T.P.: Monocular segment-wise depth: monocular depth estimation based on a semantic segmentation prior. In: 2019 IEEE International Conference on Image Processing (ICIP), pp. 4295\u20134299. IEEE (2019)","DOI":"10.1109\/ICIP.2019.8803551"},{"key":"250_CR57","doi-asserted-by":"crossref","unstructured":"Ros, G., Sellart, L., Materzynska, J., Vazquez, D., Lopez, A.M.: The synthia dataset: a large collection of synthetic images for semantic segmentation of urban scenes. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3234\u20133243 (2016)","DOI":"10.1109\/CVPR.2016.352"},{"key":"250_CR58","doi-asserted-by":"crossref","unstructured":"Hu, J., Zhang, Y., Okatani, T.: Visualization of convolutional neural networks for monocular depth estimation. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 3869\u20133878 (2019)","DOI":"10.1109\/ICCV.2019.00397"},{"key":"250_CR59","doi-asserted-by":"crossref","unstructured":"Laina, I., Rupprecht, C., Belagiannis, V., Tombari, F., Navab, N.: Deeper depth prediction with fully convolutional residual networks. In: 2016 Fourth International Conference on 3D Vision (3DV), pp. 239\u2013248 IEEE (2016)","DOI":"10.1109\/3DV.2016.32"},{"key":"250_CR60","doi-asserted-by":"crossref","unstructured":"Hu, J., Ozay, M., Zhang, Y., Okatani, T.: Revisiting single image depth estimation: toward higher resolution maps with accurate object boundaries. In: 2019 IEEE Winter Conference on Applications of Computer Vision (WACV), pp. 1043\u20131051. IEEE (2019)","DOI":"10.1109\/WACV.2019.00116"},{"key":"250_CR61","doi-asserted-by":"crossref","unstructured":"Kanayama, H., Ueda, T., Ito, H., Yamamoto, K.: Two-mode mapless visual navigation of indoor autonomous mobile robot using deep convolutional neural network. In: 2020 IEEE\/SICE International Symposium on System Integration (SII), pp. 536\u2013541. IEEE (2020)","DOI":"10.1109\/SII46433.2020.9025851"},{"key":"250_CR62","doi-asserted-by":"publisher","first-page":"117527","DOI":"10.1109\/ACCESS.2021.3107041","volume":"9","author":"C-Y Tsai","year":"2021","unstructured":"Tsai, C.-Y., Nisar, H., Hu, Y.-C.: Mapless lidar navigation control of wheeled mobile robots based on deep imitation learning. IEEE Access 9, 117527\u2013117541 (2021)","journal-title":"IEEE Access"},{"key":"250_CR63","doi-asserted-by":"crossref","unstructured":"Nguyen, A., Tran, Q.D.: Autonomous navigation with mobile robots using deep learning and the robot operating system. In: Robot Operating System (ROS), pp. 177\u2013195. Springer (2021)","DOI":"10.1007\/978-3-030-75472-3_5"},{"key":"250_CR64","doi-asserted-by":"crossref","unstructured":"Xiong, Y., Zhang, X., Peng, J., Yu, W.: 3d depth map based optimal motion control for wheeled mobile robot. In: 2017 IEEE International Conference on Systems, Man, and Cybernetics (SMC), pp. 2045\u20132050. IEEE (2017)","DOI":"10.1109\/SMC.2017.8122920"},{"issue":"2","key":"250_CR65","doi-asserted-by":"publisher","first-page":"203","DOI":"10.1007\/s41315-021-00177-0","volume":"5","author":"C Li","year":"2021","unstructured":"Li, C., Li, B., Wang, R., Zhang, X.: A survey on visual servoing for wheeled mobile robots. Int. J. Intell. Robot. Appl. 5(2), 203\u2013218 (2021)","journal-title":"Int. J. Intell. Robot. Appl."},{"key":"250_CR66","doi-asserted-by":"crossref","unstructured":"Islam, S., Dias, J., Sunda-Meya, A.: On the design and development of vision-based autonomous mobile manipulation. In: IECON 2021\u201347th Annual Conference of the IEEE Industrial Electronics Society, pp. 1\u20136. IEEE (2021)","DOI":"10.1109\/IECON48115.2021.9589584"},{"key":"250_CR67","unstructured":"Xiao, X., Liu, B., Warnell, G., Stone, P.: Motion planning and control for mobile robot navigation using machine learning: a survey"},{"issue":"1","key":"250_CR68","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s10846-021-01540-w","volume":"104","author":"Z Machkour","year":"2022","unstructured":"Machkour, Z., Ortiz-Arroyo, D., Durdevic, P.: Classical and deep learning based visual servoing systems: a survey on state of the art. J. Intell. Robot. Syst. 104(1), 1\u201327 (2022)","journal-title":"J. Intell. Robot. Syst."},{"key":"250_CR69","doi-asserted-by":"crossref","unstructured":"Furtado, J.S., Liu, H.H., Lai, G., Lacheray, H., Desouza-Coelho, J.: Comparative analysis of optitrack motion capture systems. In: Advances in Motion Sensing and Control for Robotic Applications, pp. 15\u201331. Springer (2019)","DOI":"10.1007\/978-3-030-17369-2_2"},{"key":"250_CR70","doi-asserted-by":"crossref","unstructured":"Sturm, J., Engelhard, N., Endres, F., Burgard, W., Cremers, D.: A benchmark for the evaluation of rgb-d slam systems. In: 2012 IEEE\/RSJ International Conference on Intelligent Robots and Systems, pp. 573\u2013580. IEEE (2012)","DOI":"10.1109\/IROS.2012.6385773"},{"issue":"3","key":"250_CR71","doi-asserted-by":"publisher","first-page":"1623","DOI":"10.1109\/TPAMI.2020.3019967","volume":"44","author":"R Ranftl","year":"2022","unstructured":"Ranftl, R., Lasinger, K., Hafner, D., Schindler, K., Koltun, V.: Towards robust monocular depth estimation: mixing datasets for zero-shot cross-dataset transfer. IEEE Trans. Pattern Anal. Mach. Intell. 44(3), 1623\u20131637 (2022)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"250_CR72","unstructured":"Palou\u00a0Visa, G.: Monocular depth estimation in images and sequences using occlusion cues (2014)"},{"issue":"8","key":"250_CR73","doi-asserted-by":"publisher","first-page":"1333","DOI":"10.3390\/math8081333","volume":"8","author":"Y-S Tsai","year":"2020","unstructured":"Tsai, Y.-S., Hsu, L.-H., Hsieh, Y.-Z., Lin, S.-S.: The real-time depth estimation for an occluded person based on a single image and openpose method. Mathematics 8(8), 1333 (2020)","journal-title":"Mathematics"},{"key":"250_CR74","doi-asserted-by":"crossref","unstructured":"Cao, Z., Simon, T., Wei, S.-E., Sheikh, Y.: Realtime multi-person 2d pose estimation using part affinity fields. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7291\u20137299 (2017)","DOI":"10.1109\/CVPR.2017.143"},{"key":"250_CR75","unstructured":"Jang, H., Ko, Y., Lee, Y., Jeon, M.: Light robust monocular depth estimation for outdoor environment via monochrome and color camera fusion. Preprint at arXiv:2202.12108 (2022)"},{"key":"250_CR76","unstructured":"Tian, Y., Hu, X.: Monocular depth estimation based on a single image: a literature review. In: Twelfth International Conference on Graphics and Image Processing (ICGIP 2020), vol. 11720, p. 117201. International Society for Optics and Photonics (2021)"}],"container-title":["International Journal of Computational Intelligence Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s44196-023-00250-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s44196-023-00250-5\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s44196-023-00250-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,20]],"date-time":"2024-10-20T05:43:34Z","timestamp":1729403014000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s44196-023-00250-5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,5,10]]},"references-count":76,"journal-issue":{"issue":"1","published-online":{"date-parts":[[2023,12]]}},"alternative-id":["250"],"URL":"https:\/\/doi.org\/10.1007\/s44196-023-00250-5","relation":{},"ISSN":["1875-6883"],"issn-type":[{"value":"1875-6883","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,5,10]]},"assertion":[{"value":"20 April 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"12 April 2023","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"10 May 2023","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The author(s) declare(s) that there is no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"79"}}