{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,30]],"date-time":"2026-01-30T05:09:26Z","timestamp":1769749766380,"version":"3.49.0"},"reference-count":43,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2025,1,3]],"date-time":"2025-01-03T00:00:00Z","timestamp":1735862400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,3]],"date-time":"2025-01-03T00:00:00Z","timestamp":1735862400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["SN COMPUT. SCI."],"DOI":"10.1007\/s42979-024-03606-6","type":"journal-article","created":{"date-parts":[[2025,1,3]],"date-time":"2025-01-03T12:59:16Z","timestamp":1735909156000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["A Reinforcement Learning Approach for Routing in Marine Communication Network of Fishing Vessels"],"prefix":"10.1007","volume":"6","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-7018-9476","authenticated-orcid":false,"given":"Simi","family":"Surendran","sequence":"first","affiliation":[]},{"given":"Alberto","family":"Montresor","sequence":"additional","affiliation":[]},{"given":"Maneesha","family":"Vinodini Ramesh","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,1,3]]},"reference":[{"key":"3606_CR1","doi-asserted-by":"crossref","unstructured":"Rao SN, Ramesh MV, Rangan V. Mobile infrastructure for coastal region offshore communications and networks. In: Proc. of the IEEE Global Humanitarian Technology Conf. (GHTC);2016. pp. 99\u2013104. IEEE.","DOI":"10.1109\/GHTC.2016.7857266"},{"key":"3606_CR2","doi-asserted-by":"publisher","first-page":"48236","DOI":"10.1109\/ACCESS.2019.2909921","volume":"7","author":"K-LA Yau","year":"2019","unstructured":"Yau K-LA, Syed AR, Hashim W, Qadir J, Wu C, Hassan N. Maritime networking: bringing internet to the sea. IEEE Access. 2019;7:48236\u201355.","journal-title":"IEEE Access"},{"key":"3606_CR3","doi-asserted-by":"publisher","DOI":"10.1016\/j.measen.2021.100051","volume":"16","author":"PK Shrivastava","year":"2021","unstructured":"Shrivastava PK, Vishwamitra L. Comparative analysis of proactive and reactive routing protocols in vanet environment. Meas Sens. 2021;16: 100051.","journal-title":"Meas Sens"},{"key":"3606_CR4","doi-asserted-by":"publisher","DOI":"10.1016\/j.adhoc.2020.102157","volume":"103","author":"B-S Kim","year":"2020","unstructured":"Kim B-S, Ullah S, Kim KH, Roh B, Ham J-H, Kim K-I. An enhanced geographical routing protocol based on multi-criteria decision making method in mobile ad-hoc networks. Ad Hoc Netw. 2020;103: 102157.","journal-title":"Ad Hoc Netw"},{"key":"3606_CR5","doi-asserted-by":"publisher","first-page":"27552","DOI":"10.1109\/ACCESS.2021.3058388","volume":"9","author":"RA Nazib","year":"2021","unstructured":"Nazib RA, Moh S. Reinforcement learning-based routing protocols for vehicular ad hoc networks: a comparative survey. IEEE Access. 2021;9:27552\u201387.","journal-title":"IEEE Access"},{"key":"3606_CR6","doi-asserted-by":"publisher","first-page":"2155","DOI":"10.1109\/TNSM.2024.3352014","volume":"21","author":"AA Okine","year":"2024","unstructured":"Okine AA, Adam N, Naeem F, Kaddoum G. Multi-agent deep reinforcement learning for packet routing in tactical mobile sensor networks. IEEE Trans Netw Serv Manag. 2024;21:2155\u20132169.","journal-title":"IEEE Trans Netw Serv Manag"},{"issue":"4","key":"3606_CR7","doi-asserted-by":"publisher","first-page":"1787","DOI":"10.1007\/s11276-022-03198-0","volume":"29","author":"D Prabhu","year":"2023","unstructured":"Prabhu D, Alageswaran R, Miruna Joe Amali S. Multiple agent based reinforcement learning for energy efficient routing in wsn. Wirel Netw. 2023;29(4):1787\u201397.","journal-title":"Wirel Netw"},{"key":"3606_CR8","doi-asserted-by":"publisher","DOI":"10.1016\/j.comnet.2023.109562","volume":"222","author":"CS Nandyala","year":"2023","unstructured":"Nandyala CS, Kim H-W, Cho H-S. Qtar: a q-learning-based topology-aware routing protocol for underwater wireless sensor networks. Comput Netw. 2023;222:109562.","journal-title":"Comput Netw"},{"key":"3606_CR9","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2023.123089","volume":"246","author":"X Yang","year":"2024","unstructured":"Yang X, Yan J, Wang D, Xu Y, Hua G. Woad3qn-rp: an intelligent routing protocol in wireless sensor networks-a swarm intelligence and deep reinforcement learning based approach. Expert Syst Appl. 2024;246: 123089.","journal-title":"Expert Syst Appl"},{"key":"3606_CR10","doi-asserted-by":"publisher","first-page":"55916","DOI":"10.1109\/ACCESS.2019.2913776","volume":"7","author":"Z Mammeri","year":"2019","unstructured":"Mammeri Z. Reinforcement learning based routing in networks: review and classification of approaches. IEEE Access. 2019;7:55916\u201350.","journal-title":"IEEE Access"},{"key":"3606_CR11","unstructured":"Boyan JA, Littman ML. Packet routing in dynamically changing networks: a reinforcement learning approach. In: Advances in neural information processing systems; 1994. pp. 671\u2013678. Citeseer."},{"key":"3606_CR12","unstructured":"Choi SP, Yeung D.-Y. Predictive q-routing: a memory-based reinforcement learning approach to adaptive traffic control. In: Advances in neural information processing systems; 1996. pp. 945\u2013951."},{"issue":"3","key":"3606_CR13","doi-asserted-by":"publisher","first-page":"360","DOI":"10.1109\/TSMCA.2005.846390","volume":"35","author":"J Dowling","year":"2005","unstructured":"Dowling J, Curran E, Cunningham R, Cahill V. Using feedback in collaborative reinforcement learning to adaptively optimize manet routing. IEEE Trans Syst Man Cybern-Part A Syst Humans. 2005;35(3):360\u201372.","journal-title":"IEEE Trans Syst Man Cybern-Part A Syst Humans"},{"key":"3606_CR14","doi-asserted-by":"publisher","first-page":"304","DOI":"10.1016\/j.comcom.2019.11.011","volume":"150","author":"J Liu","year":"2020","unstructured":"Liu J, Wang Q, He C, Jaffr\u00e8s-Runser K, Xu Y, Li Z, Xu Y. Qmr: Q-learning based multi-objective optimization routing protocol for flying ad hoc networks. Comput Commun. 2020;150:304\u201316.","journal-title":"Comput Commun"},{"key":"3606_CR15","doi-asserted-by":"crossref","unstructured":"Li R, Li F, Li X, Wang Y. Qgrid: Q-learning based routing protocol for vehicular ad hoc networks. In: 2014 IEEE 33rd Int. Performance Computing and Communications Conf. (IPCCC); 2014. pp. 1\u20138.IEEE.","DOI":"10.1109\/PCCC.2014.7017079"},{"issue":"5","key":"3606_CR16","doi-asserted-by":"publisher","first-page":"940","DOI":"10.1016\/j.adhoc.2010.11.006","volume":"9","author":"A F\u00f6rster","year":"2011","unstructured":"F\u00f6rster A, Murphy AL. Froms: a failure tolerant and mobility enabled multicast routing paradigm with reinforcement learning for wsns. Ad Hoc Netw. 2011;9(5):940\u201365.","journal-title":"Ad Hoc Netw"},{"issue":"4","key":"3606_CR17","doi-asserted-by":"publisher","first-page":"1025","DOI":"10.3390\/s20041025","volume":"20","author":"Y Lu","year":"2020","unstructured":"Lu Y, He R, Chen X, Lin B, Yu C. Energy-efficient depth-based opportunistic routing with q-learning for underwater wireless sensor networks. Sensors. 2020;20(4):1025.","journal-title":"Sensors"},{"issue":"13","key":"3606_CR18","doi-asserted-by":"publisher","first-page":"13589","DOI":"10.1109\/JSEN.2022.3175994","volume":"22","author":"R Zhu","year":"2022","unstructured":"Zhu R, Jiang Q, Huang X, Li D, Yang Q. A reinforcement-learning-based opportunistic routing protocol for energy-efficient and void-avoided uasns. IEEE Sens J. 2022;22(13):13589\u2013601.","journal-title":"IEEE Sens J"},{"key":"3606_CR19","doi-asserted-by":"publisher","DOI":"10.1016\/j.asoc.2021.107486","volume":"108","author":"Y Chai","year":"2021","unstructured":"Chai Y, Zeng X-J. A multi-objective dyna-q based routing in wireless mesh network. Appl Soft Comput. 2021;108: 107486.","journal-title":"Appl Soft Comput"},{"key":"3606_CR20","doi-asserted-by":"crossref","unstructured":"Elwhishi A, Ho P-H, Naik K, Shihada B. Arbr: Adaptive reinforcement-based routing for dtn. In: 2010 IEEE 6th Int. Conf. on Wireless and Mobile Computing, Networking and Communications; 2010. pp. 376\u2013385.IEEE.","DOI":"10.1109\/WIMOB.2010.5645040"},{"issue":"10","key":"3606_CR21","doi-asserted-by":"publisher","first-page":"9805","DOI":"10.1109\/JIOT.2020.2989924","volume":"7","author":"X Li","year":"2020","unstructured":"Li X, Hu X, Zhang R, Yang L. Routing protocol design for underwater optical wireless sensor networks: A multiagent reinforcement learning approach. IEEE Internet Things J. 2020;7(10):9805\u201318.","journal-title":"IEEE Internet Things J"},{"issue":"3","key":"3606_CR22","doi-asserted-by":"publisher","first-page":"2756","DOI":"10.1109\/TVT.2021.3058282","volume":"70","author":"Y Zhang","year":"2021","unstructured":"Zhang Y, Zhang Z, Chen L, Wang X. Reinforcement learning-based opportunistic routing protocol for underwater acoustic sensor networks. IEEE Trans Veh Technol. 2021;70(3):2756\u201370.","journal-title":"IEEE Trans Veh Technol"},{"key":"3606_CR23","doi-asserted-by":"crossref","unstructured":"Lin S-C, Akyildiz IF, Wang P, Luo M. Qos-aware adaptive routing in multi-layer hierarchical software defined networks: a reinforcement learning approach. In: 2016 IEEE Int. Conf. on Services Computing (SCC); 2016. pp. 25\u201333. IEEE.","DOI":"10.1109\/SCC.2016.12"},{"issue":"4","key":"3606_CR24","doi-asserted-by":"publisher","first-page":"3185","DOI":"10.1109\/TNSE.2020.3017751","volume":"7","author":"Y-R Chen","year":"2020","unstructured":"Chen Y-R, Rezapour A, Tzeng W-G, Tsai S-C. Rl-routing: an sdn routing algorithm based on deep reinforcement learning. IEEE Trans Netw Sci Eng. 2020;7(4):3185\u201399.","journal-title":"IEEE Trans Netw Sci Eng"},{"issue":"5","key":"3606_CR25","doi-asserted-by":"publisher","first-page":"10026","DOI":"10.3390\/s150510026","volume":"15","author":"D Ye","year":"2015","unstructured":"Ye D, Zhang M, Yang Y. A multi-agent framework for packet routing in wireless sensor networks. Sensors. 2015;15(5):10026\u201347.","journal-title":"Sensors"},{"key":"3606_CR26","unstructured":"Kumar S, Miikkulainen R. Dual reinforcement q-routing: An on-line adaptive routing algorithm. In: Proc. of the Artificial Neural Networks in Engineering Conf.; 1997. pp. 231\u2013238. Citeseer."},{"issue":"3","key":"3606_CR27","doi-asserted-by":"publisher","first-page":"861","DOI":"10.1016\/j.adhoc.2012.09.008","volume":"11","author":"D Macone","year":"2013","unstructured":"Macone D, Oddi G, Pietrabissa A. Mq-routing: mobility-, gps-and energy-aware routing protocol in manets for disaster relief scenarios. Ad Hoc Netw. 2013;11(3):861\u201378.","journal-title":"Ad Hoc Netw"},{"issue":"6","key":"3606_CR28","doi-asserted-by":"publisher","first-page":"796","DOI":"10.1109\/TMC.2010.28","volume":"9","author":"T Hu","year":"2010","unstructured":"Hu T, Fei Y. Qelar: a machine-learning-based adaptive routing protocol for energy-efficient and lifetime-extended underwater sensor networks. IEEE Trans Mob Comput. 2010;9(6):796\u2013809.","journal-title":"IEEE Trans Mob Comput"},{"key":"3606_CR29","doi-asserted-by":"publisher","first-page":"269","DOI":"10.1016\/j.comnet.2015.06.016","volume":"88","author":"N Coutinho","year":"2015","unstructured":"Coutinho N, Matos R, Marques C, Reis A, Sargento S, Chakareski J, Kassler A. Dynamic dual-reinforcement-learning routing strategies for quality of experience-aware wireless mesh networking. Comput Netw. 2015;88:269\u201385.","journal-title":"Comput Netw"},{"issue":"10","key":"3606_CR30","doi-asserted-by":"publisher","first-page":"2258","DOI":"10.1109\/LCOMM.2017.2656879","volume":"21","author":"W-S Jung","year":"2017","unstructured":"Jung W-S, Yim J, Ko Y-B. Qgeo: Q-learning-based geographic ad hoc routing protocol for unmanned robotic networks. IEEE Commun Lett. 2017;21(10):2258\u201361.","journal-title":"IEEE Commun Lett"},{"issue":"6","key":"3606_CR31","doi-asserted-by":"publisher","first-page":"1431","DOI":"10.1587\/transcom.E93.B.1431","volume":"93","author":"C Wu","year":"2010","unstructured":"Wu C, Kumekawa K, Kato T. Distributed reinforcement learning approach for vehicular ad hoc networks. IEICE Trans Commun. 2010;93(6):1431\u201342.","journal-title":"IEICE Trans Commun"},{"key":"3606_CR32","doi-asserted-by":"publisher","first-page":"196","DOI":"10.1016\/j.comnet.2015.08.019","volume":"91","author":"Y Saleem","year":"2015","unstructured":"Saleem Y, Yau K-LA, Mohamad H, Ramli N, Rehmani MH. Smart: a spectrum-aware cluster-based routing scheme for distributed cognitive radio networks. Comput Netw. 2015;91:196\u2013224.","journal-title":"Comput Netw"},{"key":"3606_CR33","doi-asserted-by":"publisher","DOI":"10.1016\/j.comnet.2021.108379","volume":"198","author":"LAL Costa","year":"2021","unstructured":"Costa LAL, Kunst R, Freitas EP. Q-fanet: improved q-learning based routing protocol for fanets. Comput Netw. 2021;198: 108379.","journal-title":"Comput Netw"},{"issue":"3","key":"3606_CR34","doi-asserted-by":"publisher","first-page":"1985","DOI":"10.1109\/JIOT.2021.3089759","volume":"9","author":"MY Arafat","year":"2021","unstructured":"Arafat MY, Moh S. A q-learning-based topology-aware routing protocol for flying ad hoc networks. IEEE Internet of Things J. 2021;9(3):1985\u20132000.","journal-title":"IEEE Internet of Things J"},{"key":"3606_CR35","doi-asserted-by":"publisher","DOI":"10.1016\/j.adhoc.2021.102589","volume":"121","author":"I Lahsen-Cherif","year":"2021","unstructured":"Lahsen-Cherif I, Zitoune L, V\u00e8que V. Energy efficient routing for wireless mesh networks with directional antennas: When q-learning meets ant systems. Ad Hoc Netw. 2021;121: 102589.","journal-title":"Ad Hoc Netw"},{"key":"3606_CR36","doi-asserted-by":"publisher","first-page":"6434","DOI":"10.1109\/JIOT.2021.3113128","volume":"9","author":"S Lee","year":"2021","unstructured":"Lee S, Yu H, Lee H. Multi-agent q-learning based multi-uav wireless networks for maximizing energy efficiency: deployment and power control strategy design. IEEE Internet of Things J. 2021;9:6434\u20136442.","journal-title":"IEEE Internet of Things J"},{"key":"3606_CR37","first-page":"598","volume":"10","author":"S Simi","year":"2018","unstructured":"Simi S, Ramesh MV. A reinforcement learning approach for improving internet connectivity in maritime network. J Adv Res Dyn Control Syst. 2018;10:598\u2013604.","journal-title":"J Adv Res Dyn Control Syst"},{"key":"3606_CR38","doi-asserted-by":"crossref","unstructured":"Rao SN, Raj D, Aiswarya S, Unni S. Realizing cost-effective marine internet for fishermen. In: 2016 14th Int. Symposium on Modeling and Optimization in Mobile, Ad Hoc, and Wireless Networks (WiOpt); 2016. pp. 1\u20135. IEEE.","DOI":"10.1109\/WIOPT.2016.7492904"},{"key":"3606_CR39","doi-asserted-by":"crossref","unstructured":"Unni S, Raj D, Sasidhar K, Rao S. Performance measurement and analysis of long range wi-fi network for over-the-sea communication. In: 2015 13th Int. Symposium on Modeling and Optimization in Mobile, Ad Hoc, and Wireless Networks (WiOpt); 2015. pp. 36\u201341. IEEE.","DOI":"10.1109\/WIOPT.2015.7151030"},{"key":"3606_CR40","doi-asserted-by":"crossref","unstructured":"Dhivvya J, Rao SN, Simi S. Towards maximizing throughput and coverage of a novel heterogeneous maritime communication network. In: Proc. of the 18th ACM Int. Symposium on Mobile Ad Hoc Networking and Computing; 2017. p. 39. ACM.","DOI":"10.1145\/3084041.3084077"},{"key":"3606_CR41","doi-asserted-by":"publisher","first-page":"5031","DOI":"10.1109\/ACCESS.2023.3235387","volume":"11","author":"S Surendran","year":"2023","unstructured":"Surendran S, Ramesh MV, Montresor A, Montag MJ. Link characterization and edge-centric predictive modeling in an ocean network. IEEE Access. 2023;11:5031\u201346.","journal-title":"IEEE Access"},{"issue":"3","key":"3606_CR42","first-page":"279","volume":"8","author":"CJ Watkins","year":"1992","unstructured":"Watkins CJ, Dayan P. Q-learning Mach Learn. 1992;8(3):279\u201392.","journal-title":"Q-learning. Mach Learn"},{"key":"3606_CR43","doi-asserted-by":"publisher","DOI":"10.1016\/j.compeleceng.2020.106781","volume":"87","author":"S Surendran","year":"2020","unstructured":"Surendran S, Ramesh MV, Montag MJ, Montresor A. Modelling communication capability and node reorientation in offshore communication network. Comput Electr Eng. 2020;87: 106781.","journal-title":"Comput Electr Eng"}],"container-title":["SN Computer Science"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s42979-024-03606-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s42979-024-03606-6\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s42979-024-03606-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,1,3]],"date-time":"2025-01-03T13:07:51Z","timestamp":1735909671000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s42979-024-03606-6"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,1,3]]},"references-count":43,"journal-issue":{"issue":"1","published-online":{"date-parts":[[2025,1]]}},"alternative-id":["3606"],"URL":"https:\/\/doi.org\/10.1007\/s42979-024-03606-6","relation":{},"ISSN":["2661-8907"],"issn-type":[{"value":"2661-8907","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,1,3]]},"assertion":[{"value":"30 May 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 December 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"3 January 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"On behalf of all authors, the corresponding author states that there is no Conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of Interest"}}],"article-number":"62"}}