{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,11]],"date-time":"2026-04-11T02:06:03Z","timestamp":1775873163546,"version":"3.50.1"},"reference-count":24,"publisher":"MDPI AG","issue":"19","license":[{"start":{"date-parts":[[2023,9,28]],"date-time":"2023-09-28T00:00:00Z","timestamp":1695859200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"name":"Korea Institute for Advancement of Technology (KIAT) grant funded by the Korea Government (MOTID)","award":["P0008473"],"award-info":[{"award-number":["P0008473"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Remote Sensing"],"abstract":"<jats:p>In this paper, a novel DRL algorithm based on a DQN is proposed for multiple mobile robots to find optimized paths. The multiple robots\u2019 states are the inputs of the DQN. The DQN estimates the Q-value of the agents\u2019 actions. After selecting the action with the maximum Q-value, the multiple robots\u2019 actions are calculated and sent to them. Then, the robots will explore the area and detect the obstacles. In the area, there are static obstacles. The robots should detect the static obstacles using a LiDAR sensor. The other moving robots are recognized as dynamic obstacles that need to be avoided. The robots will give feedback on the reward and the robots\u2019 new states. A positive reward will be given when a robot successfully arrives at its goal point. If it is in a free space, zero reward will be given. If the robot collides with a static obstacle or other robots or reaches its start point, it will receive a negative reward. Multiple robots explore safe paths to the goals at the same time, in order to improve learning efficiency. If a robot collides with an obstacle or other robots, it will stop and wait for the other robots to complete their exploration tasks. The episode will end when all robots find safe paths to reach their goals or when all of them have collisions. This collaborative behavior can reduce the risk of collisions between robots, enhance overall efficiency, and help avoid multiple robots attempting to navigate through the same unsafe path simultaneously. Moreover, storage space is used to store the optimal safe paths of all robots. Finally, the multi-robots will learn the policy to find the optimized paths to go to the goal points. The goal of the simulations and experiment is to make multiple robots efficiently and safely move to their goal points.<\/jats:p>","DOI":"10.3390\/rs15194757","type":"journal-article","created":{"date-parts":[[2023,9,29]],"date-time":"2023-09-29T05:48:13Z","timestamp":1695966493000},"page":"4757","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":10,"title":["Sensing and Navigation for Multiple Mobile Robots Based on Deep Q-Network"],"prefix":"10.3390","volume":"15","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-2484-4585","authenticated-orcid":false,"given":"Yanyan","family":"Dai","sequence":"first","affiliation":[{"name":"Robotics Department, Yeungnam University, Gyeongsan 38541, Republic of Korea"}]},{"given":"Seokho","family":"Yang","sequence":"additional","affiliation":[{"name":"Robotics Department, Yeungnam University, Gyeongsan 38541, Republic of Korea"}]},{"given":"Kidong","family":"Lee","sequence":"additional","affiliation":[{"name":"Robotics Department, Yeungnam University, Gyeongsan 38541, Republic of Korea"}]}],"member":"1968","published-online":{"date-parts":[[2023,9,28]]},"reference":[{"key":"ref_1","unstructured":"Mustafa, K., Botteghi, N., Sirmacek, B., Poel, M., and Stramigioli, S. (2019). The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, Copernicus Publications."},{"key":"ref_2","unstructured":"Linh, K., Cornelius, M., and Jens, L. (2020, January 20\u201324). Deep-Reinforcement-Learning-Based Semantic Navigation of Mobile Robots in Dynamic Environments. Proceedings of the IEEE International Conference on Automation Science and Engineering, Mexico City, Mexico."},{"key":"ref_3","doi-asserted-by":"crossref","first-page":"3","DOI":"10.1177\/0278364919887447","article-title":"Learning dexterous in hand manipulation","volume":"39","author":"Andrychowicz","year":"2020","journal-title":"Int. J. Robot. Res."},{"key":"ref_4","doi-asserted-by":"crossref","first-page":"1143","DOI":"10.1109\/LRA.2020.2966414","article-title":"Learning Robust Control Policies for End-toEnd Autonomous Driving from Data-Driven Simulation","volume":"5","author":"Amini","year":"2020","journal-title":"IEEE Robot. Autom. Lett."},{"key":"ref_5","unstructured":"Liu, X., Chen, Y.R., Li, H.R., Li, B.Y., and Zhao, D.B. (2023). Cross-domain Random Pre-training with Prototypes for Reinforcement Learning. arXiv."},{"key":"ref_6","doi-asserted-by":"crossref","unstructured":"Wu, J.D., Huang, W.H., Boer, N., Mo, Y.H., He, X.K., and Lv, C. (2022, January 8\u201312). Safe Decision-making for Lane-change of Autonomous Ve-hicles via Human Demonstration-aided Reinforcement Learning. Proceedings of the 2022 IEEE 25th International Conference on Intelligent Transportation Systems (ITSC), Macau, China.","DOI":"10.1109\/ITSC55140.2022.9921872"},{"key":"ref_7","doi-asserted-by":"crossref","unstructured":"Hu, T., Luo, B., and Yang, C. (2021). Multi-objective optimization for autonomous driving strategy based on Deep Q Network. Discov. Artif. Intell., 1.","DOI":"10.1007\/s44163-021-00011-3"},{"key":"ref_8","doi-asserted-by":"crossref","unstructured":"Zeng, J., Ju, R., Qin, L., Yin, Q., and Hu, C. (2019). Navigation in unknown dynamic environments based on deep reinforcement learning. Sensors, 19.","DOI":"10.3390\/s19183837"},{"key":"ref_9","doi-asserted-by":"crossref","unstructured":"Tan, J., Zhang, T.N., Counmans, E., Iscen, A., Bai, Y.F., Hafner, D., Bohez, S., and Vanhoucke, V. (2018). Sim-to-real: Learning agile locomotion for quadruped robots. arXiv.","DOI":"10.15607\/RSS.2018.XIV.010"},{"key":"ref_10","unstructured":"Surmann, H., Jestel, C., Marchel, R., Musberg, F., Elhadj, H., and Ardani, M. (2020). Deep Reinforcement learning for real autonomous mobile robot navigation in indoor environments. arXiv."},{"key":"ref_11","doi-asserted-by":"crossref","unstructured":"Lee, M.R., and Yusuf, S.H. (2022). Mobile Robot Navigation Using Deep Reinforcement Learning. Sensors, 10.","DOI":"10.3390\/pr10122748"},{"key":"ref_12","doi-asserted-by":"crossref","unstructured":"Orr, J., and Dutta, A. (2023). Multi-Agent Deep Reinforcement Learning for Multi-Robot Applications: A Survey. Sensors, 23.","DOI":"10.3390\/s23073625"},{"key":"ref_13","doi-asserted-by":"crossref","unstructured":"Cai, Z.X., Liang, Z., and Ren, J. (2021). MRDRL-ROS: A Multi Robot Deep Reinforcement Learning Platform Based on Robot Operating System. J. Phys. Conf. Ser., 2113.","DOI":"10.1088\/1742-6596\/2113\/1\/012086"},{"key":"ref_14","doi-asserted-by":"crossref","unstructured":"Chen, W., Zhou, S., Pan, Z., Zheng, H., and Liu, Y. (2019). Mapless Collaborative Navigation for a Multi-Robot System Based on the Deep Reinforcement Learning. Appl. Sci., 9.","DOI":"10.3390\/app9204198"},{"key":"ref_15","doi-asserted-by":"crossref","unstructured":"Jestel, C., Surmann, H., Stenzel, J., Urbann, O., and Brehler, M. (2021, January 4\u20136). Obataining Robust Control and Navigation Policies for Multi-robot Navigation via Deep Reinforcement Learning. Proceedings of the International Conference on Automation, Robotics and Applications, Prague, Czech Republic.","DOI":"10.1109\/ICARA51699.2021.9376457"},{"key":"ref_16","doi-asserted-by":"crossref","unstructured":"Han, R., Chen, S., and Hao, Q. (August, January 31). Cooperative multi-robot navigation in dynamic environment with deep reinforcement learning. Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Paris, France.","DOI":"10.1109\/ICRA40945.2020.9197209"},{"key":"ref_17","doi-asserted-by":"crossref","first-page":"3160","DOI":"10.1109\/JSAC.2021.3088718","article-title":"Deep reinforcement learning based three-dimensional area coverage with UAV swarm","volume":"39","author":"Mou","year":"2021","journal-title":"IEEE J. Sel. Areas Commun."},{"key":"ref_18","doi-asserted-by":"crossref","first-page":"417","DOI":"10.1177\/00202940211000380","article-title":"A real-time HIL control system on rotary inverted pendulum hardware platform based on double deep Q-network","volume":"54","author":"Dai","year":"2021","journal-title":"Meas. Control."},{"key":"ref_19","doi-asserted-by":"crossref","unstructured":"Salla, A.E., Abdou, M., Perot, E., and Yogamani, S.K.J.A. (2017). Deep Reinforcement Learning framework for Autonomous Driving. Electron. Imaging, 29.","DOI":"10.2352\/ISSN.2470-1173.2017.19.AVM-023"},{"key":"ref_20","doi-asserted-by":"crossref","unstructured":"Qiu, X., Wan, K., and Li, F. (2019, January 22\u201324). Autonomous Robot Navigation in Dynamic Environment Using Deep Reinforcement Learning. Proceedings of the 2019 IEEE 2nd International Conference on Automation, Electronics and Electrical Engineering (AUTEEE), Shenyang, China.","DOI":"10.1109\/AUTEEE48671.2019.9033166"},{"key":"ref_21","doi-asserted-by":"crossref","unstructured":"Rahman, M.D.M., Rashid, S.M.H., and Hossain, M.M. (2018). Implementation of Q learning and deep Q network for controlling a self-balancing robot model. Robot. Biomim., 5.","DOI":"10.1186\/s40638-018-0091-9"},{"key":"ref_22","doi-asserted-by":"crossref","unstructured":"Sumanans, M., Petronis, A., Bucinskas, V., Dzedzickis, A., Virzonix, D., and Morkvenaite-Vilkonciene, I. (2022). Deep Q-Learning in Robotics: Improvement of Accuracy and Repeatability. Sensors, 22.","DOI":"10.3390\/s22103911"},{"key":"ref_23","doi-asserted-by":"crossref","unstructured":"Zhou, S., Liu, X., Xu, Y., and Guo, J. (2018, January 11\u201313). A Deep Q-network (DQN) Based Path Planning Method for Mobile Robots. Proceedings of the 2018 IEEE International Conference on Information and Automation (ICIA), Wuyishan, China.","DOI":"10.1109\/ICInfA.2018.8812452"},{"key":"ref_24","doi-asserted-by":"crossref","first-page":"1673","DOI":"10.1631\/FITEE.2200109","article-title":"A deep Q-learning network based active object detection model with a novel training algorithm for service robots","volume":"23","author":"Liu","year":"2022","journal-title":"Front. Inf. Technol. Electron. Eng."}],"container-title":["Remote Sensing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2072-4292\/15\/19\/4757\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,10]],"date-time":"2025-10-10T21:01:05Z","timestamp":1760130065000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2072-4292\/15\/19\/4757"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,9,28]]},"references-count":24,"journal-issue":{"issue":"19","published-online":{"date-parts":[[2023,10]]}},"alternative-id":["rs15194757"],"URL":"https:\/\/doi.org\/10.3390\/rs15194757","relation":{},"ISSN":["2072-4292"],"issn-type":[{"value":"2072-4292","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,9,28]]}}}