{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,23]],"date-time":"2026-02-23T21:22:02Z","timestamp":1771881722830,"version":"3.50.1"},"reference-count":30,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/legalcode"}],"funder":[{"name":"European Union under Italian National Recovery and Resilience Plan (NRRP) of NextGenerationEU, Partnership on \u201cTelecommunications of the Future\u201d (Program \u201cRESTART\u201d","award":["PE00000001"],"award-info":[{"award-number":["PE00000001"]}]},{"name":"PRIN Project \u201cResilient Delivery of Real-Time Interactive Services Over NextG Compute-Dense Mobile Networks\u201d","award":["E53D2300055000"],"award-info":[{"award-number":["E53D2300055000"]}]},{"name":"Funds from U.S. National Science Foundation as Specified in the RINGS Program","award":["CNS-2148315"],"award-info":[{"award-number":["CNS-2148315"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Netw."],"published-print":{"date-parts":[[2026]]},"DOI":"10.1109\/ton.2025.3649614","type":"journal-article","created":{"date-parts":[[2025,12,30]],"date-time":"2025-12-30T18:37:07Z","timestamp":1767119827000},"page":"2653-2668","source":"Crossref","is-referenced-by-count":0,"title":["A Flexible Multi-Agent Deep Reinforcement Learning Framework for Dynamic Routing and Scheduling of Latency-Critical Services"],"prefix":"10.1109","volume":"34","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-0365-8575","authenticated-orcid":false,"given":"Vincenzo Norman","family":"Vitale","sequence":"first","affiliation":[{"name":"University of Naples Federico II, Naples, Italy"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6050-4150","authenticated-orcid":false,"given":"Antonia Maria","family":"Tulino","sequence":"additional","affiliation":[{"name":"University of Naples Federico II, Naples, Italy"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4779-4763","authenticated-orcid":false,"given":"Andreas F.","family":"Molisch","sequence":"additional","affiliation":[{"name":"University of Southern California, Los Angeles, CA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6713-5861","authenticated-orcid":false,"given":"Jaime","family":"Llorca","sequence":"additional","affiliation":[{"name":"ECE Department, New York University, Brooklyn, NY, USA"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/6GNet54646.2022.9830429"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/JSAC.2016.2621398"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1145\/3397166.3409149"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/tmc.2025.3613949"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/ICIN60470.2024.10494452"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/CDC.1990.204000"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-79995-2"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/TNET.2017.2783846"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/TNET.2021.3070699"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/TAC.2018.2874671"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/TNET.2022.3179349"},{"key":"ref12","first-page":"671","article-title":"Packet routing in dynamically changing networks: A reinforcement learning approach","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"6","author":"Boyan"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2019.2913776"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1016\/B978-1-55860-335-6.50027-1"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/COMST.2019.2916583"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1145\/3529375"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/TSMC.2020.3012832"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/ICNP52444.2021.9651934"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1512\/iumj.1957.6.56038"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1016\/S0004-3702(98)00023-X"},{"key":"ref21","volume-title":"Reinforcement Learning: An Introduction","author":"Sutton","year":"2018"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1007\/BF00992698"},{"key":"ref23","first-page":"1057","article-title":"Policy gradient methods for reinforcement learning with function approximation","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"12","author":"Sutton"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1038\/nature14236"},{"key":"ref25","first-page":"6379","article-title":"Multi-agent actor-critic for mixed cooperative-competitive environments","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"30","author":"Lowe"},{"key":"ref26","article-title":"Continuous control with deep reinforcement learning","author":"Lillicrap","year":"2015","journal-title":"arXiv:1509.02971"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.3390\/fi16040119"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1145\/3492866.3549712"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/TCCN.2023.3288133"},{"key":"ref30","article-title":"Asynchronous methods for deep reinforcement learning","author":"Mnih","year":"2016","journal-title":"arXiv:1602.01783"}],"container-title":["IEEE Transactions on Networking"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/10723154\/11317935\/11318654.pdf?arnumber=11318654","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,2,23]],"date-time":"2026-02-23T20:47:15Z","timestamp":1771879635000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11318654\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026]]},"references-count":30,"URL":"https:\/\/doi.org\/10.1109\/ton.2025.3649614","relation":{},"ISSN":["2998-4157"],"issn-type":[{"value":"2998-4157","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026]]}}}