{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T00:05:20Z","timestamp":1743033920779,"version":"3.40.3"},"publisher-location":"Cham","reference-count":18,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031807596"},{"type":"electronic","value":"9783031807602"}],"license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-80760-2_27","type":"book-chapter","created":{"date-parts":[[2025,2,12]],"date-time":"2025-02-12T03:06:24Z","timestamp":1739329584000},"page":"425-439","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Multi-agent Path Planning for\u00a0Logistics Cargo Environment Using LSTM Based Reinforcement Learning"],"prefix":"10.1007","author":[{"given":"Gun Rae","family":"Cho","sequence":"first","affiliation":[]},{"given":"Sungho","family":"Park","sequence":"additional","affiliation":[]},{"given":"Eui-Jung","family":"Jung","sequence":"additional","affiliation":[]},{"given":"Hyunseok","family":"Shin","sequence":"additional","affiliation":[]},{"given":"So Eun","family":"Son","sequence":"additional","affiliation":[]},{"given":"Yong","family":"Choi","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,2,13]]},"reference":[{"key":"27_CR1","unstructured":"Badia, A.P., et\u00a0al.: Never give up: Learning directed exploration strategies. arXiv preprint arXiv:2002.06038 (2020)"},{"key":"27_CR2","doi-asserted-by":"publisher","first-page":"59","DOI":"10.1016\/j.proeng.2014.12.098","volume":"96","author":"F Ducho\u0148","year":"2014","unstructured":"Ducho\u0148, F., et al.: Path planning with modified a star algorithm for a mobile robot. Procedia Eng. 96, 59\u201369 (2014)","journal-title":"Procedia Eng."},{"issue":"8","key":"27_CR3","doi-asserted-by":"publisher","first-page":"1735","DOI":"10.1162\/neco.1997.9.8.1735","volume":"9","author":"S Hochreiter","year":"1997","unstructured":"Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735\u20131780 (1997)","journal-title":"Neural Comput."},{"issue":"3","key":"27_CR4","doi-asserted-by":"publisher","first-page":"385","DOI":"10.1109\/TSM.2022.3168702","volume":"35","author":"S Hong","year":"2022","unstructured":"Hong, S., Hwang, I., Jang, Y.J.: Practical Q-learning-based route-guidance and vehicle assignment for OHT systems in semiconductor fabs. IEEE Trans. Semicond. Manuf. 35(3), 385\u2013396 (2022)","journal-title":"IEEE Trans. Semicond. Manuf."},{"key":"27_CR5","doi-asserted-by":"crossref","unstructured":"Hu, Y., Yang, L., Lou, Y.: Path planning with Q-learning. J. Phys. Conf. Ser. 1948, 012038 (2021)","DOI":"10.1088\/1742-6596\/1948\/1\/012038"},{"issue":"4","key":"27_CR6","doi-asserted-by":"publisher","first-page":"1199","DOI":"10.1080\/00207543.2019.1614692","volume":"58","author":"I Hwang","year":"2020","unstructured":"Hwang, I., Jang, Y.J.: Q ($$\\lambda $$) learning-based dynamic route guidance algorithm for overhead hoist transport systems in semiconductor fabs. Int. J. Prod. Res. 58(4), 1199\u20131221 (2020)","journal-title":"Int. J. Prod. Res."},{"key":"27_CR7","unstructured":"Kulkarni, T.D., Narasimhan, K., Saeedi, A., Tenenbaum, J.: Hierarchical deep reinforcement learning: integrating temporal abstraction and intrinsic motivation. In: Advances in Neural Information Processing Systems, vol. 29 (2016)"},{"key":"27_CR8","unstructured":"Lillicrap, T.P., et al.: Continuous control with deep reinforcement learning. arXiv preprint arXiv:1509.02971 (2015)"},{"key":"27_CR9","doi-asserted-by":"crossref","unstructured":"Liu, Y., Chen, M., Huang, H.: Multi-agent pathfinding based on improved cooperative a in kiva system. In: 2019 5th International Conference on Control, Automation and Robotics (ICCAR), pp. 633\u2013638. IEEE (2019)","DOI":"10.1109\/ICCAR.2019.8813319"},{"key":"27_CR10","doi-asserted-by":"crossref","unstructured":"Liu, Z., Chen, B., Zhou, H., Koushik, G., Hebert, M., Zhao, D.: MAPPER: multi-agent path planning with evolutionary reinforcement learning in mixed dynamic environments. In: 2020 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 11748\u201311754. IEEE (2020)","DOI":"10.1109\/IROS45743.2020.9340876"},{"key":"27_CR11","doi-asserted-by":"publisher","DOI":"10.1016\/j.asoc.2020.106796","volume":"97","author":"A Maoudj","year":"2020","unstructured":"Maoudj, A., Hentout, A.: Optimal path planning approach based on Q-learning algorithm for mobile robots. Appl. Soft Comput. 97, 106796 (2020)","journal-title":"Appl. Soft Comput."},{"key":"27_CR12","unstructured":"Mnih, V., et al.: Playing Atari with deep reinforcement learning. arXiv preprint arXiv:1312.5602 (2013)"},{"issue":"5","key":"27_CR13","doi-asserted-by":"publisher","first-page":"952","DOI":"10.2514\/1.G006129","volume":"45","author":"MR Schlichting","year":"2022","unstructured":"Schlichting, M.R., Notter, S., Fichter, W.: Long short-term memory for spatial encoding in multi-agent path planning. J. Guid. Control. Dyn. 45(5), 952\u2013961 (2022)","journal-title":"J. Guid. Control. Dyn."},{"key":"27_CR14","doi-asserted-by":"crossref","unstructured":"Van\u00a0Hasselt, H., Guez, A., Silver, D.: Deep reinforcement learning with double Q-learning. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol.\u00a030 (2016)","DOI":"10.1609\/aaai.v30i1.10295"},{"issue":"4","key":"27_CR15","doi-asserted-by":"publisher","first-page":"6932","DOI":"10.1109\/LRA.2020.3026638","volume":"5","author":"B Wang","year":"2020","unstructured":"Wang, B., Liu, Z., Li, Q., Prorok, A.: Mobile robot path planning in dynamic environments through globally guided reinforcement learning. IEEE Rob. Autom. Lett. 5(4), 6932\u20136939 (2020)","journal-title":"IEEE Rob. Autom. Lett."},{"key":"27_CR16","first-page":"1","volume":"2022","author":"W Wang","year":"2022","unstructured":"Wang, W., Wu, Z., Luo, H., Zhang, B.: Path planning method of mobile robot using improved deep reinforcement learning. J. Electric. Comput. Eng. 2022, 1\u20137 (2022)","journal-title":"J. Electric. Comput. Eng."},{"key":"27_CR17","doi-asserted-by":"crossref","unstructured":"Wu, R., et\u00a0al.: Optimization path and design of intelligent logistics management system based on ROS robot. J. Rob. 2023 (2023)","DOI":"10.1155\/2023\/9505155"},{"issue":"3","key":"27_CR18","doi-asserted-by":"publisher","first-page":"177","DOI":"10.1049\/trit.2020.0024","volume":"5","author":"Y Yang","year":"2020","unstructured":"Yang, Y., Juntao, L., Lingling, P.: Multi-robot path planning based on a deep reinforcement learning DQN algorithm. CAAI Trans. Intell. Technol. 5(3), 177\u2013183 (2020)","journal-title":"CAAI Trans. Intell. Technol."}],"container-title":["Communications in Computer and Information Science","Innovative Intelligent Industrial Production and Logistics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-80760-2_27","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,2,12]],"date-time":"2025-02-12T03:06:33Z","timestamp":1739329593000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-80760-2_27"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"ISBN":["9783031807596","9783031807602"],"references-count":18,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-80760-2_27","relation":{},"ISSN":["1865-0929","1865-0937"],"issn-type":[{"type":"print","value":"1865-0929"},{"type":"electronic","value":"1865-0937"}],"subject":[],"published":{"date-parts":[[2025]]},"assertion":[{"value":"13 February 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"The authors have no competing interests to declare that are relevant to the content of this article.","order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Disclosure of Interests"}},{"value":"IN4PL","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Innovative Intelligent Industrial Production and Logistics","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Porto","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Portugal","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"22 November 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 November 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"5","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"in4pl2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/in4pl.scitevents.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}