{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,24]],"date-time":"2026-03-24T00:21:14Z","timestamp":1774311674434,"version":"3.50.1"},"publisher-location":"Cham","reference-count":30,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031632143","type":"print"},{"value":"9783031632150","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-3-031-63215-0_10","type":"book-chapter","created":{"date-parts":[[2024,6,18]],"date-time":"2024-06-18T15:16:08Z","timestamp":1718723768000},"page":"129-141","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":3,"title":["Strategizing the Shallows: Leveraging Multi-Agent Reinforcement Learning for Enhanced Tactical Decision-Making in Littoral Naval Warfare"],"prefix":"10.1007","author":[{"given":"Lauri","family":"Vasankari","sequence":"first","affiliation":[]},{"given":"Kalle","family":"Saastamoinen","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,6,19]]},"reference":[{"issue":"2","key":"10_CR1","doi-asserted-by":"publisher","first-page":"267","DOI":"10.1002\/1520-6750(199503)42:2<267::AID-NAV3220420209>3.0.CO;2-Y","volume":"42","author":"WP Hughes Jr","year":"1995","unstructured":"Hughes, W.P., Jr.: A salvo model of warships in missile combat used to evaluate their staying power. Naval Res. Logis. (NRL) 42(2), 267\u2013289 (1995)","journal-title":"Naval Res. Logis. (NRL)"},{"issue":"5","key":"10_CR2","doi-asserted-by":"publisher","first-page":"830","DOI":"10.1287\/opre.1040.0195","volume":"53","author":"MJ Armstrong","year":"2005","unstructured":"Armstrong, M.J.: A stochastic salvo model for naval surface combat. Oper. Res. 53(5), 830\u2013841 (2005)","journal-title":"Oper. Res."},{"issue":"1","key":"10_CR3","doi-asserted-by":"publisher","first-page":"66","DOI":"10.1002\/nav.20187","volume":"54","author":"MJ Armstrong","year":"2007","unstructured":"Armstrong, M.J.: Effective attacks in the salvo combat model: salvo sizes and quantities of targets. Naval Res. Logis. (NRL) 54(1), 66\u201377 (2007)","journal-title":"Naval Res. Logis. (NRL)"},{"key":"10_CR4","doi-asserted-by":"crossref","unstructured":"Dupuy, T.: The Quantified Judgment Method of Analysis of Historical Combat Data: A Monograph, Historical Evaluation and Research Organization (1974)","DOI":"10.1007\/978-1-4757-0958-2_18"},{"key":"10_CR5","unstructured":"Mooren, E. M., Reinforcement learning applications to combat identification (Doctoral dissertation, Monterey, California: Naval Postgraduate School), (2017)"},{"key":"10_CR6","doi-asserted-by":"crossref","unstructured":"Wang, J., Wang, J., He, J., Wang, G., Wang, M.: Research on naval air defense intelligent operations on deep reinforcement learning. In: 2022 34th Chinese Control and Decision Conference (CCDC), pp. 2246\u20132252. IEEE, August 2022","DOI":"10.1109\/CCDC55256.2022.10034115"},{"key":"10_CR7","volume":"8","author":"M Rempel","year":"2021","unstructured":"Rempel, M., Cai, J.: A review of approximate dynamic programming applications within military operations research. Oper. Res. Perspect. 8, 100204 (2021)","journal-title":"Oper. Res. Perspect."},{"key":"10_CR8","doi-asserted-by":"crossref","unstructured":"Sztykgold, A., Coppin, G., Hudry, O.: Dynamic optimization of the strength ratio during a terrestrial conflict. In: 2007 IEEE International Symposium on Approximate Dynamic Programming and Reinforcement Learning, pp. 241\u2013246. IEEE. April 2007","DOI":"10.1109\/ADPRL.2007.368194"},{"key":"10_CR9","unstructured":"Kemple, W., Porter, G., Vodantis, N.: Littoral warfare simulation experiment. In: Proceedings of the 1996 Command and Control Research and Technology Symposium: Command and Control in the Information Age, pp. 536\u2013543 (1996)"},{"key":"10_CR10","unstructured":"Mahan, A.T.: The influence of sea power upon history, 1660-1783, Good Press\/Methuen London [England] (2020\/1965)"},{"issue":"2","key":"10_CR11","first-page":"30","volume":"68","author":"M Vego","year":"2015","unstructured":"Vego, M.: On littoral warfare. Naval War College Rev. 68(2), 30\u201368 (2015)","journal-title":"Naval War College Rev."},{"key":"10_CR12","unstructured":"Tver, D.F.: Ocean and marine dictionary (1979)"},{"key":"10_CR13","unstructured":"V\u00e4nsk\u00e4, V.: Nykyaikainen merisodank\u00e4ynti. Julkaisusarja 2: Tutkimusselosteita nro 10 (2021)"},{"key":"10_CR14","unstructured":"Johnson, B., et al.: Mapping artificial intelligence to the naval tactical kill chain. Naval Eng. J. 135(1), 155\u2013166 (2023)"},{"key":"10_CR15","unstructured":"Christiansen, B.J.: Littoral combat vessels analysis and comparison of designs (Doctoral dissertation, Monterey, California. Naval Postgraduate School) (2008)"},{"key":"10_CR16","unstructured":"Albrecht, S.V., Christianos, F., Sch\u00e4fer, L.: Multi-agent Reinforcement Learning: Foundations and Modern Approaches. Massachusetts Institute of Technology, Cambridge (2023)"},{"key":"10_CR17","doi-asserted-by":"publisher","unstructured":"Gronauer, S., Diepold, K.: Multi-agent deep reinforcement learning: a survey. Artif. Intell. Rev. 1\u201349 (2022). https:\/\/doi.org\/10.1007\/s10462-021-09996-w","DOI":"10.1007\/s10462-021-09996-w"},{"key":"10_CR18","doi-asserted-by":"crossref","unstructured":"Shoham, Y., Leyton-Brown, K.: Multiagent Systems: Algorithmic, Game-Theoretic, and Logical Foundations. Cambridge University Press (2008)","DOI":"10.1017\/CBO9780511811654"},{"key":"10_CR19","unstructured":"Sutton, R.S., Barto, A.G.: Reinforcement Learning: An Introduction. MIT press (2018)"},{"issue":"10","key":"10_CR20","doi-asserted-by":"publisher","first-page":"1095","DOI":"10.1073\/pnas.39.10.1095","volume":"39","author":"LS Shapley","year":"1953","unstructured":"Shapley, L.S.: Stochastic games. Proc. Natl. Acad. Sci. 39(10), 1095\u20131100 (1953)","journal-title":"Proc. Natl. Acad. Sci."},{"issue":"2","key":"10_CR21","doi-asserted-by":"publisher","first-page":"100","DOI":"10.1109\/TSSC.1968.300136","volume":"4","author":"PE Hart","year":"1968","unstructured":"Hart, P.E., Nilsson, N.J., Raphael, B.: A formal basis for the heuristic determination of minimum cost paths. IEEE Trans. Syst. Sci. Cybern. 4(2), 100\u2013107 (1968)","journal-title":"IEEE Trans. Syst. Sci. Cybern."},{"issue":"7","key":"10_CR22","doi-asserted-by":"publisher","first-page":"3625","DOI":"10.3390\/s23073625","volume":"23","author":"J Orr","year":"2023","unstructured":"Orr, J., Dutta, A.: Multi-agent deep reinforcement learning for multi-robot applications: a survey. Sensors 23(7), 3625 (2023)","journal-title":"Sensors"},{"key":"10_CR23","first-page":"24611","volume":"35","author":"C Yu","year":"2022","unstructured":"Yu, C., et al.: The surprising effectiveness of PPO in cooperative multi-agent games. Adv. Neural. Inf. Process. Syst. 35, 24611\u201324624 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"10_CR24","unstructured":"Mnih, V., et al.: Playing atari with deep reinforcement learning, arXiv preprint arXiv:1312.5602 (2013)"},{"key":"10_CR25","doi-asserted-by":"crossref","unstructured":"Van Hasselt, H., Guez, A., Silver, D.: Deep reinforcement learning with double q-learning. In: Proceedings of the AAAI Conference on Artificial Intelligence 30(1), March 2016","DOI":"10.1609\/aaai.v30i1.10295"},{"key":"10_CR26","unstructured":"van Hasselt, H.P., Guez, A., Hessel, M., Mnih, V., Silver, D.: Learning values across many orders of magnitude. Adv. Neural Inf. Process. Syst. 29 (2016)"},{"key":"10_CR27","unstructured":"Schaul, T., Quan, J., Antonoglou, I., Silver, D.: Prioritized experience replay, arXiv preprint arXiv:1511.05952 (2015)"},{"key":"10_CR28","unstructured":"Plappert, M., et al.: Parameter space noise for exploration, arXiv preprint arXiv:1706.01905 (2017)"},{"key":"10_CR29","unstructured":"Ba, J. L., Kiros, J.R., Hinton, G.E.: Layer normalization. arXiv preprint arXiv:1607.06450 (2016)"},{"key":"10_CR30","unstructured":"HEADQUARTERS OF THE COMMANDER IN CHIEF, Radar pickets and methods of combating suicide attacks off okinawa (1945)"}],"container-title":["IFIP Advances in Information and Communication Technology","Artificial Intelligence Applications and Innovations"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-63215-0_10","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,6,18]],"date-time":"2024-06-18T15:17:42Z","timestamp":1718723862000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-63215-0_10"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9783031632143","9783031632150"],"references-count":30,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-63215-0_10","relation":{},"ISSN":["1868-4238","1868-422X"],"issn-type":[{"value":"1868-4238","type":"print"},{"value":"1868-422X","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"19 June 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"AIAI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"IFIP International Conference on Artificial Intelligence Applications and Innovations","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Corfu","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Greece","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 June 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"30 June 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"20","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"aiai2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/ifipaiai.org\/2024\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}