{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,12]],"date-time":"2026-05-12T07:01:48Z","timestamp":1778569308008,"version":"3.51.4"},"reference-count":29,"publisher":"Springer Science and Business Media LLC","issue":"4","license":[{"start":{"date-parts":[[2025,5,9]],"date-time":"2025-05-09T00:00:00Z","timestamp":1746748800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"},{"start":{"date-parts":[[2025,5,9]],"date-time":"2025-05-09T00:00:00Z","timestamp":1746748800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"}],"funder":[{"name":"National Key Research and Development Program of China","award":["2022YFF0605205"],"award-info":[{"award-number":["2022YFF0605205"]}]},{"DOI":"10.13039\/501100001809","name":"Natural Science Foundation of China","doi-asserted-by":"crossref","award":["61931001"],"award-info":[{"award-number":["61931001"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100005090","name":"Beijing Nova Program","doi-asserted-by":"publisher","award":["20230484364"],"award-info":[{"award-number":["20230484364"]}],"id":[{"id":"10.13039\/501100005090","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Peer-to-Peer Netw. Appl."],"published-print":{"date-parts":[[2025,7]]},"DOI":"10.1007\/s12083-025-02003-x","type":"journal-article","created":{"date-parts":[[2025,5,9]],"date-time":"2025-05-09T02:26:24Z","timestamp":1746757584000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Attentional value-factorization-based resource allocation and performance evaluation for intelligent connected vehicles"],"prefix":"10.1007","volume":"18","author":[{"given":"Chao","family":"Cai","sequence":"first","affiliation":[]},{"given":"Jiahui","family":"Qiu","sequence":"additional","affiliation":[]},{"given":"Bin","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Xiangyun","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Xuanhan","family":"Zhu","sequence":"additional","affiliation":[]},{"given":"Yang","family":"Li","sequence":"additional","affiliation":[]},{"given":"Quan","family":"Yuan","sequence":"additional","affiliation":[]},{"given":"Lili","family":"Li","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,5,9]]},"reference":[{"key":"2003_CR1","doi-asserted-by":"crossref","unstructured":"Huang W et al (2024) Manipulating voice assistants eavesdropping via inherent vulnerability unveiling in mobile systems. IEEE Trans Mobile Comput","DOI":"10.1109\/TMC.2024.3401096"},{"key":"2003_CR2","doi-asserted-by":"crossref","unstructured":"Huang W, Tang W, Jiang H, Zhang Y (2024) Recognizing voice spoofing attacks via acoustic nonlinearity dissection for mobile devices. IEEE Trans Mobile Comput","DOI":"10.1109\/TMC.2024.3411791"},{"issue":"12","key":"2003_CR3","doi-asserted-by":"publisher","first-page":"3878","DOI":"10.1109\/jsac.2023.3322841","volume":"41","author":"H Lu","year":"2023","unstructured":"Lu H et al (2023) FL-AMM: Federated Learning Augmented Map Matching with heterogeneous cellular moving trajectories. IEEE J Sel Areas Commun 41(12):3878\u20133892. https:\/\/doi.org\/10.1109\/jsac.2023.3322841","journal-title":"IEEE J Sel Areas Commun"},{"key":"2003_CR4","unstructured":"Cao H, et al. (2024) Security analysis of wifi-based sensing systems: threats from perturbation attacks. arXiv:2404.15587"},{"issue":"24","key":"2003_CR5","doi-asserted-by":"publisher","first-page":"21820","DOI":"10.1109\/jiot.2023.3299934","volume":"10","author":"W Yue","year":"2023","unstructured":"Yue W, Li C, Duan P, Yu FR (2023) Revolution on wheels: a survey on the positive and negative impacts of connected and automated vehicles in era of mixed autonomy. IEEE Int Things J 10(24):21820\u201321835. https:\/\/doi.org\/10.1109\/jiot.2023.3299934","journal-title":"IEEE Int Things J"},{"key":"2003_CR6","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/jiot.2024.3359673","volume":"100","author":"Z Zheng","year":"2024","unstructured":"Zheng Z et al (2024) Capacity of vehicular networks in mixed traffic with CAVs and human-driven vehicles. IEEE Intern Things J 100:1\u20131. https:\/\/doi.org\/10.1109\/jiot.2024.3359673","journal-title":"IEEE Intern Things J"},{"issue":"24","key":"2003_CR7","doi-asserted-by":"publisher","first-page":"39579","DOI":"10.1109\/jiot.2024.3444044","volume":"11","author":"X Fu","year":"2024","unstructured":"Fu X et al (2024) HierNet: a hierarchical resource allocation method for vehicle platooning networks. IEEE Internet Things J 11(24):39579\u201339592. https:\/\/doi.org\/10.1109\/jiot.2024.3444044","journal-title":"IEEE Internet Things J"},{"issue":"9","key":"2003_CR8","doi-asserted-by":"publisher","first-page":"2753","DOI":"10.3390\/s24092753","volume":"24","author":"J-W Kim","year":"2024","unstructured":"Kim J-W, Kim J-W, Lee J (2024) Intelligent resource allocation scheme using reinforcement learning for efficient data transmission in VANET. Sensors 24(9):2753. https:\/\/doi.org\/10.3390\/s24092753","journal-title":"Sensors"},{"key":"2003_CR9","doi-asserted-by":"publisher","unstructured":"Cao L, Yin H (2021) Resource allocation for vehicle platooning in 5G NR-V2X via deep reinforcement learning. arXiv Cornell University. https:\/\/doi.org\/10.1109\/blackseacom52164.2021.9527765","DOI":"10.1109\/blackseacom52164.2021.9527765"},{"key":"2003_CR10","doi-asserted-by":"publisher","unstructured":"Strunz M, Heinovski J, Dressler F (2021) CoOP: V2V-based cooperative overtaking for platoons on freeways. 2021 IEEE Intelligent Transportation Systems Conference (ITSC) 2021. https:\/\/doi.org\/10.1109\/itsc48978.2021.9565122","DOI":"10.1109\/itsc48978.2021.9565122"},{"issue":"7","key":"2003_CR11","doi-asserted-by":"publisher","first-page":"8834","DOI":"10.1109\/tits.2021.3086809","volume":"23","author":"V Vegamoor","year":"2022","unstructured":"Vegamoor V, Rathinam S, Darbha S (2022) String stability of connected vehicle platoons under lossy V2V communication. IEEE Trans Intell Transp Syst 23(7):8834\u20138845. https:\/\/doi.org\/10.1109\/tits.2021.3086809","journal-title":"IEEE Trans Intell Transp Syst"},{"key":"2003_CR12","doi-asserted-by":"publisher","unstructured":"Bhardwaj A, Agnihotri S (2020) Multicast protocols for D2D. Wiley 5G Ref, pp 1\u201318. https:\/\/doi.org\/10.1002\/9781119471509.w5gref183","DOI":"10.1002\/9781119471509.w5gref183"},{"issue":"17","key":"2003_CR13","doi-asserted-by":"publisher","first-page":"5658","DOI":"10.3390\/s24175658","volume":"24","author":"I Narayanasamy","year":"2024","unstructured":"Narayanasamy I, Rajamanickam V (2024) A cascaded multi-agent reinforcement learning-based resource allocation for cellular-V2X vehicular platooning networks. Sensors 24(17):5658. https:\/\/doi.org\/10.3390\/s24175658","journal-title":"Sensors"},{"key":"2003_CR14","doi-asserted-by":"publisher","first-page":"1719","DOI":"10.1109\/ojcoms.2022.3211340","volume":"3","author":"N Banitalebi","year":"2022","unstructured":"Banitalebi N, Azmi P, Mokari N, Arani AH, Yanikomeroglu H (2022) Distributed learning-based resource allocation for self-organizing C-V2X communication in cellular networks. IEEE Open J Commun Soc 3:1719\u20131736. https:\/\/doi.org\/10.1109\/ojcoms.2022.3211340","journal-title":"IEEE Open J Commun Soc"},{"issue":"19","key":"2003_CR15","doi-asserted-by":"publisher","first-page":"10145","DOI":"10.3390\/app121910145","volume":"12","author":"HT Trinh","year":"2022","unstructured":"Trinh HT, Bae S-H, Tran DQ (2022) Deep reinforcement learning for vehicle platooning at a signalized intersection in mixed traffic with partial detection. Appl Sci 12(19):10145. https:\/\/doi.org\/10.3390\/app121910145","journal-title":"Appl Sci"},{"issue":"5","key":"2003_CR16","doi-asserted-by":"publisher","first-page":"1874","DOI":"10.3390\/s22051874","volume":"22","author":"J Fu","year":"2022","unstructured":"Fu J, Qin X, Huang Y, Tang L, Liu Y (2022) Deep reinforcement learning-based resource allocation for cellular vehicular network mode 3 with underlay approach. Sensors 22(5):1874. https:\/\/doi.org\/10.3390\/s22051874","journal-title":"Sensors"},{"issue":"20","key":"2003_CR17","doi-asserted-by":"publisher","first-page":"4191","DOI":"10.3390\/electronics12204191","volume":"12","author":"B Hou","year":"2023","unstructured":"Hou B et al (2023) SoC-VRP: A deep-reinforcement-learning-based vehicle route planning mechanism for service-oriented cooperative ITS. Electronics 12(20):4191\u20134191. https:\/\/doi.org\/10.3390\/electronics12204191","journal-title":"Electronics"},{"issue":"23","key":"2003_CR18","doi-asserted-by":"publisher","first-page":"9459","DOI":"10.3390\/s22239459","volume":"22","author":"S Yu","year":"2022","unstructured":"Yu S, Lee JW (2022) Deep reinforcement learning based resource allocation for D2D communications underlay cellular networks. Sensors 22(23):9459\u20139459. https:\/\/doi.org\/10.3390\/s22239459","journal-title":"Sensors"},{"issue":"3","key":"2003_CR19","doi-asserted-by":"publisher","first-page":"1295","DOI":"10.3390\/s23031295","volume":"23","author":"D Han","year":"2023","unstructured":"Han D, So J (2023) Energy-efficient resource allocation based on deep Q-Network in V2V communications. Sensors 23(3):1295. https:\/\/doi.org\/10.3390\/s23031295","journal-title":"Sensors"},{"issue":"2","key":"2003_CR20","doi-asserted-by":"publisher","first-page":"360","DOI":"10.3390\/electronics12020360","volume":"12","author":"X Li","year":"2023","unstructured":"Li X, Chen G, Wu G, Sun Z, Chen G (2023) Research on multi-agent D2D communication resource allocation algorithm based on A2C. Electronics 12(2):360\u2013360. https:\/\/doi.org\/10.3390\/electronics12020360","journal-title":"Electronics"},{"key":"2003_CR21","doi-asserted-by":"publisher","unstructured":"Erg\u00fcn S (2024) Resource allocation optimization for effective vehicle network communications using multi-agent deep reinforcement learning. J Dyn Game. https:\/\/doi.org\/10.3934\/jdg.2024017","DOI":"10.3934\/jdg.2024017"},{"key":"2003_CR22","doi-asserted-by":"publisher","unstructured":"Fu X, Yuan Q, Zhuang Z, Li Y, Liao J, Zhao D (2024) TacNet: a tactic-interactive resource allocation method for vehicular networks. IEEE Internet Things J 1\u20131. https:\/\/doi.org\/10.1109\/jiot.2023.3345853","DOI":"10.1109\/jiot.2023.3345853"},{"issue":"1","key":"2003_CR23","doi-asserted-by":"publisher","first-page":"645","DOI":"10.1109\/tmc.2022.3220720","volume":"23","author":"S Duan","year":"2022","unstructured":"Duan S et al (2022) MOTO: mobility-aware online task offloading with adaptive load balancing in small-cell MEC. IEEE Trans Mob Comput 23(1):645\u2013659. https:\/\/doi.org\/10.1109\/tmc.2022.3220720","journal-title":"IEEE Trans Mob Comput"},{"issue":"8","key":"2003_CR24","doi-asserted-by":"publisher","first-page":"1996","DOI":"10.1109\/TPDS.2021.3134647","volume":"33","author":"Y Deng","year":"2022","unstructured":"Deng Y et al (2022) AUCTION: automated and quality-aware client selection framework for efficient federated learning. IEEE Trans Parallel Distrib Syst 33(8):1996\u20132009. https:\/\/doi.org\/10.1109\/TPDS.2021.3134647","journal-title":"IEEE Trans Parallel Distrib Syst"},{"key":"2003_CR25","unstructured":"3GPP (2018) Study on enhancement of 3GPP support for 5G V2X services"},{"key":"2003_CR26","unstructured":"van Seijen H, Fatemi M, Romoff J, Laroche R, Barnes T, Tsang J (2017) Hybrid reward architecture for reinforcement learning. arXiv (Cornell Univ) 30:5392\u20135402"},{"key":"2003_CR27","doi-asserted-by":"publisher","unstructured":"L\u00f3pez A, Matsui Y (2006) RTP payload format for 3rd Generation Partnership Project (3GPP) Timed Text. https:\/\/doi.org\/10.17487\/rfc4396","DOI":"10.17487\/rfc4396"},{"key":"2003_CR28","unstructured":"Parvini M, Javan MR, Mokari N, Abbasi B, Jorswieck EA (2021) AoI-aware resource allocation for platoon-based C-V2X networks via multi-agent multi-task reinforcement learning. arXiv:2105.04196"},{"issue":"1","key":"2003_CR29","first-page":"4292","volume":"21","author":"T Rashid","year":"2018","unstructured":"Rashid T, Samvelyan M, Schroeder C, Farquhar G, Foerster J, Whiteson S (2018) QMIX: monotonic value function factorisation for deep multi-agent reinforcement learning. J Mach Learn Res 21(1):4292\u20134301","journal-title":"J Mach Learn Res"}],"container-title":["Peer-to-Peer Networking and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s12083-025-02003-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s12083-025-02003-x\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s12083-025-02003-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,27]],"date-time":"2025-09-27T15:21:17Z","timestamp":1758986477000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s12083-025-02003-x"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,5,9]]},"references-count":29,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2025,7]]}},"alternative-id":["2003"],"URL":"https:\/\/doi.org\/10.1007\/s12083-025-02003-x","relation":{},"ISSN":["1936-6442","1936-6450"],"issn-type":[{"value":"1936-6442","type":"print"},{"value":"1936-6450","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,5,9]]},"assertion":[{"value":"9 April 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"28 April 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"9 May 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing Interests"}},{"value":"This study did not require ethical approval.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethics Approval and Consent to Participate"}},{"value":"Not applicable.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent for Publication"}}],"article-number":"170"}}