{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,7,3]],"date-time":"2025-07-03T05:40:01Z","timestamp":1751521201367,"version":"3.41.0"},"reference-count":25,"publisher":"IEEE","license":[{"start":{"date-parts":[[2025,6,2]],"date-time":"2025-06-02T00:00:00Z","timestamp":1748822400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,6,2]],"date-time":"2025-06-02T00:00:00Z","timestamp":1748822400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025,6,2]]},"DOI":"10.1109\/vnc64509.2025.11054106","type":"proceedings-article","created":{"date-parts":[[2025,7,2]],"date-time":"2025-07-02T17:41:50Z","timestamp":1751478110000},"page":"1-8","source":"Crossref","is-referenced-by-count":0,"title":["Multi-Agent Reinforcement Learning Scheduling to Support Low Latency in Teleoperated Driving"],"prefix":"10.1109","author":[{"given":"Giacomo","family":"Avanzi","sequence":"first","affiliation":[{"name":"University of Padova,Department of Information Engineering,Italy"}]},{"given":"Marco","family":"Giordani","sequence":"additional","affiliation":[{"name":"University of Padova,Department of Information Engineering,Italy"}]},{"given":"Michele","family":"Zorzi","sequence":"additional","affiliation":[{"name":"University of Padova,Department of Information Engineering,Italy"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/MCOM.001.1900411"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/JIOT.2020.3028766"},{"journal-title":"White Paper","article-title":"C-V2X Use Cases Volume II: Examples and Service Level Requirements","year":"2020","key":"ref3"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/MCOM.2016.1600071CM"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/MNET.001.2100237"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/WCNC51071.2022.9771590"},{"article-title":"Towards Decentralized Predictive Quality of Service in Next-Generation Vehicular Networks","volume-title":"IEEE Information Theory and Applications Workshop (ITA)","author":"Bragato","key":"ref7"},{"key":"ref8","doi-asserted-by":"crossref","DOI":"10.1109\/GLOBECOM52923.2024.10901349","article-title":"Federated Reinforcement Learning to Optimize Teleoperated Driving Networks","volume-title":"IEEE Global Communications Conference","author":"Bragato"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/JPROC.2019.2957798"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/JSAC.2021.3078498"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/VTC2020-Fall49728.2020.9348718"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/COMST.2018.2828880"},{"issue":"3&4","key":"ref13","first-page":"128","article-title":"Recent development and applications of SUMO - Simulation of Urban MObility","volume":"5","author":"Krajzewicz","year":"2012","journal-title":"International Journal On Advances in Systems and Measurements"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/TVT.2014.2317803"},{"volume-title":"Draco 3D Data Compression","year":"2017","key":"ref15"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1145\/3532577.3532605"},{"journal-title":"TS 38.300","article-title":"NR and NG-RAN Overall Description (Release 15)","year":"2018","key":"ref17"},{"volume-title":"Multi-Agent Reinforcement Learning: Foundations and Modern Approaches.","year":"2024","author":"Albrecht","key":"ref18"},{"key":"ref19","doi-asserted-by":"crossref","DOI":"10.1287\/moor.27.4.819.297","article-title":"The complexity of decentralized control of Markov decision processes","volume-title":"Proceedings of the Sixteenth Conference on Uncertainty in Artificial Intelligence","author":"Bernstein"},{"article-title":"Proximal policy optimization algorithms","year":"2017","author":"Schulman","key":"ref20"},{"article-title":"Trust region policy optimization","volume-title":"Proceedings of the 32nd International Conference on Machine Learning","author":"Schulman","key":"ref21"},{"article-title":"High-dimensional continuous control using generalized advantage estimation","volume-title":"Proceedings of the International Conference on Learning Representations (ICLR)","author":"Schulman","key":"ref22"},{"volume-title":"Is independent learning all you need in the starcraft multi-agent challenge?","year":"2020","author":"de Witt","key":"ref23"},{"article-title":"The surprising effectiveness of PPO in cooperative multi-agent games","volume-title":"Proceedings of the 36th International Conference on Neural Information Processing Systems","author":"Yu","key":"ref24"},{"article-title":"An introduction to centralized training for decentralized execution in cooperative multi-agent reinforcement learning","year":"2024","author":"Amato","key":"ref25"}],"event":{"name":"2025 IEEE Vehicular Networking Conference (VNC)","start":{"date-parts":[[2025,6,2]]},"location":"Porto, Portugal","end":{"date-parts":[[2025,6,4]]}},"container-title":["2025 IEEE Vehicular Networking Conference (VNC)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/11054063\/11054055\/11054106.pdf?arnumber=11054106","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,7,3]],"date-time":"2025-07-03T05:17:35Z","timestamp":1751519855000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11054106\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,6,2]]},"references-count":25,"URL":"https:\/\/doi.org\/10.1109\/vnc64509.2025.11054106","relation":{},"subject":[],"published":{"date-parts":[[2025,6,2]]}}}