{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,14]],"date-time":"2026-03-14T16:54:01Z","timestamp":1773507241955,"version":"3.50.1"},"reference-count":19,"publisher":"IEEE","license":[{"start":{"date-parts":[[2021,12,1]],"date-time":"2021-12-01T00:00:00Z","timestamp":1638316800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2021,12,1]],"date-time":"2021-12-01T00:00:00Z","timestamp":1638316800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2021,12,1]],"date-time":"2021-12-01T00:00:00Z","timestamp":1638316800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021,12]]},"DOI":"10.1109\/gcwkshps52748.2021.9682160","type":"proceedings-article","created":{"date-parts":[[2022,1,24]],"date-time":"2022-01-24T21:06:05Z","timestamp":1643058365000},"page":"1-6","source":"Crossref","is-referenced-by-count":16,"title":["Cell On\/Off Parameter Optimization for Saving Energy via Reinforcement Learning"],"prefix":"10.1109","author":[{"given":"Minsuk","family":"Choi","sequence":"first","affiliation":[]},{"given":"Kyungrae","family":"Kim","sequence":"additional","affiliation":[]},{"given":"Hongjun","family":"Jang","sequence":"additional","affiliation":[]},{"given":"Hyokyung","family":"Woo","sequence":"additional","affiliation":[]},{"given":"Joan S.","family":"Pujol-Roig","sequence":"additional","affiliation":[]},{"given":"Yue","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Hunje","family":"Yeon","sequence":"additional","affiliation":[]},{"given":"Sunghyun","family":"Choi","sequence":"additional","affiliation":[]},{"given":"Seowoo","family":"Jang","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/CAMAD.2019.8858474"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2020.2995057"},{"key":"ref12","article-title":"Deep reinforcement learning for cell on\/off energy saving on wireless networks","author":"pujol","year":"2021","journal-title":"IEEE Global Communications Conference (Globecom)"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/VTCSpring.2015.7145603"},{"key":"ref14","article-title":"Deep reinforcement learning for intelligent transportation systems: A survey","author":"haydari","year":"2020","journal-title":"IEEE Transactions on Intelligent Transportation Systems (T-ITS)"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1109\/TITS.2021.3054625"},{"key":"ref16","article-title":"Real2sim: Data-driven modeling of radio access networks","author":"jang","year":"2021"},{"key":"ref17","article-title":"Proximal policy optimization algorithms","author":"schulman","year":"2017"},{"key":"ref18","article-title":"Continuous control with deep reinforcement learning","author":"lillicrap","year":"2015"},{"key":"ref19","article-title":"Trust region policy optimization","author":"schulman","year":"2015","journal-title":"International Conference on Machine Learning (ICML)"},{"key":"ref4","article-title":"Ai-powered energy management","year":"2021"},{"key":"ref3","article-title":"Ava energy efficiency - harness ai for more sustainable energy use","year":"2021","journal-title":"NokiaAVA Energy Efficiency brochure Brochure EN pdf"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/PIMRC.2019.8904155"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/VTCFall.2018.8690555"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/NCA.2019.8935062"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/WCNC.2019.8885818"},{"key":"ref2","article-title":"Building autonomous driving telecom networks -white paper for the huawei softcom ai solution","year":"2019"},{"key":"ref1","article-title":"The case for committing to greener telecom networks","author":"lee","year":"2020"},{"key":"ref9","article-title":"Drag: Deep reinforcement learning based base station activation in heterogeneous networks","author":"ye","year":"2019","journal-title":"IEEE Transactions on Mobile Computing (TMC)"}],"event":{"name":"2021 IEEE Globecom Workshops (GC Wkshps)","location":"Madrid, Spain","start":{"date-parts":[[2021,12,7]]},"end":{"date-parts":[[2021,12,11]]}},"container-title":["2021 IEEE Globecom Workshops (GC Wkshps)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9681915\/9681831\/09682160.pdf?arnumber=9682160","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,5,10]],"date-time":"2022-05-10T16:57:52Z","timestamp":1652201872000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9682160\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,12]]},"references-count":19,"URL":"https:\/\/doi.org\/10.1109\/gcwkshps52748.2021.9682160","relation":{},"subject":[],"published":{"date-parts":[[2021,12]]}}}