{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,20]],"date-time":"2026-03-20T16:20:56Z","timestamp":1774023656747,"version":"3.50.1"},"reference-count":29,"publisher":"IEEE","license":[{"start":{"date-parts":[[2022,10,26]],"date-time":"2022-10-26T00:00:00Z","timestamp":1666742400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2022,10,26]],"date-time":"2022-10-26T00:00:00Z","timestamp":1666742400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022,10,26]]},"DOI":"10.1109\/wf-iot54382.2022.10152199","type":"proceedings-article","created":{"date-parts":[[2023,6,23]],"date-time":"2023-06-23T23:49:29Z","timestamp":1687564169000},"page":"1-7","source":"Crossref","is-referenced-by-count":20,"title":["Resource Allocation for Mobile Metaverse with the Internet of Vehicles over 6G Wireless Communications: A Deep Reinforcement Learning Approach"],"prefix":"10.1109","author":[{"given":"Terence Jie","family":"Chua","sequence":"first","affiliation":[{"name":"Graduate College, Nanyang Technological University,Singapore"}]},{"given":"Wenhan","family":"Yu","sequence":"additional","affiliation":[{"name":"Graduate College, Nanyang Technological University,Singapore"}]},{"given":"Jun","family":"Zhao","sequence":"additional","affiliation":[{"name":"School of Computer Science &#x0026; Engineering, Nanyang Technological University,Singapore"}]}],"member":"263","reference":[{"key":"ref1","article-title":"All one needs to know about metaverse: A complete survey on technological singularity, virtual ecosystem, and research agenda","author":"Lee","year":"2021","journal-title":"arXiv preprint"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/ICC45855.2022.9838422"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/ICC45855.2022.9838492"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/ICC45855.2022.9838736"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/jiot.2022.3232845"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/WF-IoT54382.2022.10152057"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/GLOCOM.2017.8254650"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/icc42927.2021.9500494"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/INFOCOM.2018.8486241"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/INFOCOM41043.2020.9155517"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/TCOMM.2018.2850303"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/TVT.2020.2995877"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/INFOCOM.2016.7524340"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1109\/INFOCOM.2017.8057116"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-73100-7_12"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/MNET.2019.1800418"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.3010896"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/TNET.2015.2487344"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1038\/nature14236"},{"key":"ref20","volume-title":"Reinforcement learning: An introduction","author":"Sutton","year":"2018"},{"key":"ref21","first-page":"1928","article-title":"Asynchronous methods for deep reinforcement learning","volume-title":"International conference on machine learning","author":"Mnih"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/TVT.2021.3071983"},{"key":"ref23","article-title":"Continuous control with deep reinforcement learning","author":"Lillicrap","year":"2015","journal-title":"arXiv preprint"},{"key":"ref24","first-page":"1861","article-title":"Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor","volume-title":"International Conference on Machine Learning","author":"Haarnoja"},{"key":"ref25","first-page":"1995","article-title":"Dueling network architectures for deep reinforcement learning","volume-title":"International Conference on Machine Learning","author":"Wang"},{"key":"ref26","volume-title":"Openai baselines: ACKTR & A2C","author":"Wu","year":"2017"},{"key":"ref27","article-title":"Prox-imal policy optimization algorithms","author":"Schulman","year":"2017","journal-title":"arXiv preprint"},{"key":"ref28","volume-title":"Introduction to radio propagation for. fixed and mobile commu-nications","author":"Doble","year":"1996"},{"key":"ref29","article-title":"Adam: A method for stochastic optimization","author":"Kingma","year":"2014","journal-title":"arXiv preprint"}],"event":{"name":"2022 IEEE 8th World Forum on Internet of Things (WF-IoT)","location":"Yokohama, Japan","start":{"date-parts":[[2022,10,26]]},"end":{"date-parts":[[2022,11,11]]}},"container-title":["2022 IEEE 8th World Forum on Internet of Things (WF-IoT)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/10151825\/10152024\/10152199.pdf?arnumber=10152199","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,1,24]],"date-time":"2024-01-24T04:12:33Z","timestamp":1706069553000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10152199\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,10,26]]},"references-count":29,"URL":"https:\/\/doi.org\/10.1109\/wf-iot54382.2022.10152199","relation":{},"subject":[],"published":{"date-parts":[[2022,10,26]]}}}