{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T04:16:49Z","timestamp":1750220209590,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":26,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,6,28]],"date-time":"2022-06-28T00:00:00Z","timestamp":1656374400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/100014013","name":"UK Research and Innovation","doi-asserted-by":"publisher","award":["EP\/S022961\/1"],"award-info":[{"award-number":["EP\/S022961\/1"]}],"id":[{"id":"10.13039\/100014013","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2022,6,28]]},"DOI":"10.1145\/3538637.3538866","type":"proceedings-article","created":{"date-parts":[[2022,6,22]],"date-time":"2022-06-22T16:33:05Z","timestamp":1655915585000},"page":"374-382","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":4,"title":["Beobench"],"prefix":"10.1145","author":[{"given":"Arduin","family":"Findeis","sequence":"first","affiliation":[{"name":"University of Cambridge"}]},{"given":"Fiodar","family":"Kazhamiaka","sequence":"additional","affiliation":[{"name":"Stanford University"}]},{"given":"Scott","family":"Jeen","sequence":"additional","affiliation":[{"name":"University of Cambridge"}]},{"given":"Srinivasan","family":"Keshav","sequence":"additional","affiliation":[{"name":"University of Cambridge"}]}],"member":"320","published-online":{"date-parts":[[2022,6,28]]},"reference":[{"key":"e_1_3_2_1_1_1","first-page":"29304","article-title":"Deep Reinforcement Learning at the Edge of the Statistical Precipice","volume":"34","author":"Agarwal Rishabh","year":"2021","unstructured":"Rishabh Agarwal , Max Schwarzer , Pablo Samuel Castro , Aaron C. Courville , and Marc Bellemare . 2021 . Deep Reinforcement Learning at the Edge of the Statistical Precipice . Advances in Neural Information Processing Systems 34 (2021), 29304 -- 29320 . Rishabh Agarwal, Max Schwarzer, Pablo Samuel Castro, Aaron C. Courville, and Marc Bellemare. 2021. Deep Reinforcement Learning at the Edge of the Statistical Precipice. Advances in Neural Information Processing Systems 34 (2021), 29304--29320.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.26868\/25222708.2021.30380"},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1080\/19401493.2021.1986574"},{"key":"e_1_3_2_1_4_1","volume-title":"arXiv preprint arXiv:1606.01540","author":"Brockman Greg","year":"2016","unstructured":"Greg Brockman , Vicki Cheung , Ludwig Pettersson , Jonas Schneider , John Schulman , Jie Tang , and Wojciech Zaremba . 2016. Open AI Gym . arXiv preprint arXiv:1606.01540 ( 2016 ). arXiv:1606.01540 Greg Brockman, Vicki Cheung, Ludwig Pettersson, Jonas Schneider, John Schulman, Jie Tang, and Wojciech Zaremba. 2016. OpenAI Gym. arXiv preprint arXiv:1606.01540 (2016). arXiv:1606.01540"},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.35833\/MPCE.2020.000552"},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.enbuild.2018.03.051"},{"key":"e_1_3_2_1_7_1","volume-title":"Witte","author":"Crawley Drury B.","year":"2001","unstructured":"Drury B. Crawley , Linda K. Lawrie , Frederick C. Winkelmann , Walter F. Buhl , Y. Joe Huang , Curtis O. Pedersen , Richard K. Strand , Richard J. Liesen , Daniel E. Fisher , and Michael J . Witte . 2001 . EnergyPlus: Creating a New-Generation Building Energy Simulation Program. Energy and buildings 33, 4 (2001), 319--331. Drury B. Crawley, Linda K. Lawrie, Frederick C. Winkelmann, Walter F. Buhl, Y. Joe Huang, Curtis O. Pedersen, Richard K. Strand, Richard J. Liesen, Daniel E. Fisher, and Michael J. Witte. 2001. EnergyPlus: Creating a New-Generation Building Energy Simulation Program. Energy and buildings 33, 4 (2001), 319--331."},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1145\/3486611.3488729"},{"key":"e_1_3_2_1_9_1","volume-title":"RLlib: Abstractions for Distributed Reinforcement Learning. In International Conference on Machine Learning. PMLR, 3053--3062","author":"Liang Eric","year":"2018","unstructured":"Eric Liang , Richard Liaw , Robert Nishihara , Philipp Moritz , Roy Fox , Ken Goldberg , Joseph Gonzalez , Michael Jordan , and Ion Stoica . 2018 . RLlib: Abstractions for Distributed Reinforcement Learning. In International Conference on Machine Learning. PMLR, 3053--3062 . Eric Liang, Richard Liaw, Robert Nishihara, Philipp Moritz, Roy Fox, Ken Goldberg, Joseph Gonzalez, Michael Jordan, and Ion Stoica. 2018. RLlib: Abstractions for Distributed Reinforcement Learning. In International Conference on Machine Learning. PMLR, 3053--3062."},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1145\/3365984.3365985"},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1016\/S1474-6670(17)43628-7"},{"key":"e_1_3_2_1_12_1","unstructured":"mechyai. 2022. RL - EmsPy. https:\/\/github.com\/mechyai\/RL-EmsPy  mechyai. 2022. RL - EmsPy. https:\/\/github.com\/mechyai\/RL-EmsPy"},{"key":"e_1_3_2_1_13_1","volume-title":"Reinforcement Learning Testbed for Power-Consumption Optimization. In Asian Simulation Conference. Springer, 45--59","author":"Moriyama Takao","year":"2018","unstructured":"Takao Moriyama , Giovanni De Magistris , Michiaki Tatsubori , Tu-Hoa Pham , Asim Munawar , and Ryuki Tachibana . 2018 . Reinforcement Learning Testbed for Power-Consumption Optimization. In Asian Simulation Conference. Springer, 45--59 . Takao Moriyama, Giovanni De Magistris, Michiaki Tatsubori, Tu-Hoa Pham, Asim Munawar, and Ryuki Tachibana. 2018. Reinforcement Learning Testbed for Power-Consumption Optimization. In Asian Simulation Conference. Springer, 45--59."},{"key":"e_1_3_2_1_14_1","volume-title":"GridLearn: Multiagent Reinforcement Learning for Grid-Aware Building Energy Management. arXiv:2110.06396 [cs] (Oct","author":"Pigott Aisling","year":"2021","unstructured":"Aisling Pigott , Constance Crozier , Kyri Baker , and Zoltan Nagy . 2021. GridLearn: Multiagent Reinforcement Learning for Grid-Aware Building Energy Management. arXiv:2110.06396 [cs] (Oct . 2021 ). arXiv:2110.06396 [cs] Aisling Pigott, Constance Crozier, Kyri Baker, and Zoltan Nagy. 2021. GridLearn: Multiagent Reinforcement Learning for Grid-Aware Building Energy Management. arXiv:2110.06396 [cs] (Oct. 2021). arXiv:2110.06396 [cs]"},{"key":"e_1_3_2_1_15_1","volume-title":"Stable-Baselines3: Reliable Reinforcement Learning Implementations. Journal of Machine Learning Research","author":"Raffin Antonin","year":"2021","unstructured":"Antonin Raffin , Ashley Hill , Adam Gleave , Anssi Kanervisto , Maximilian Ernestus , and Noah Dormann . 2021. Stable-Baselines3: Reliable Reinforcement Learning Implementations. Journal of Machine Learning Research ( 2021 ). Antonin Raffin, Ashley Hill, Adam Gleave, Anssi Kanervisto, Maximilian Ernestus, and Noah Dormann. 2021. Stable-Baselines3: Reliable Reinforcement Learning Implementations. Journal of Machine Learning Research (2021)."},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.3390\/app11083518"},{"key":"e_1_3_2_1_17_1","volume-title":"Proximal Policy Optimization Algorithms. arXiv:1707.06347 [cs] (Aug","author":"Schulman John","year":"2017","unstructured":"John Schulman , Filip Wolski , Prafulla Dhariwal , Alec Radford , and Oleg Klimov . 2017. Proximal Policy Optimization Algorithms. arXiv:1707.06347 [cs] (Aug . 2017 ). arXiv:1707.06347 [cs] John Schulman, Filip Wolski, Prafulla Dhariwal, Alec Radford, and Oleg Klimov. 2017. Proximal Policy Optimization Algorithms. arXiv:1707.06347 [cs] (Aug. 2017). arXiv:1707.06347 [cs]"},{"key":"e_1_3_2_1_18_1","volume-title":"Barto","author":"Sutton Richard S.","year":"2018","unstructured":"Richard S. Sutton and Andrew G . Barto . 2018 . Reinforcement Learning : An Introduction. MIT press . Richard S. Sutton and Andrew G. Barto. 2018. Reinforcement Learning: An Introduction. MIT press."},{"key":"e_1_3_2_1_19_1","first-page":"15032","article-title":"Pettingzoo: Gym for Multi-Agent Reinforcement Learning","volume":"34","author":"Terry J.","year":"2021","unstructured":"J. Terry , Benjamin Black , Nathaniel Grammel , Mario Jayakumar , Ananth Hari , Ryan Sullivan , Luis S. Santos , Clemens Dieffendahl , Caroline Horsch , and Rodrigo Perez-Vicente . 2021 . Pettingzoo: Gym for Multi-Agent Reinforcement Learning . Advances in Neural Information Processing Systems 34 (2021), 15032 -- 15043 . J. Terry, Benjamin Black, Nathaniel Grammel, Mario Jayakumar, Ananth Hari, Ryan Sullivan, Luis S. Santos, Clemens Dieffendahl, Caroline Horsch, and Rodrigo Perez-Vicente. 2021. Pettingzoo: Gym for Multi-Agent Reinforcement Learning. Advances in Neural Information Processing Systems 34 (2021), 15032--15043.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_20_1","volume-title":"CityLearn: Standardizing Research in Multi-Agent Reinforcement Learning for Demand Response and Urban Energy Management. arXiv:2012.10504 [cs] (Dec","author":"Vazquez-Canteli Jose R.","year":"2020","unstructured":"Jose R. Vazquez-Canteli , Sourav Dey , Gregor Henze , and Zoltan Nagy . 2020. CityLearn: Standardizing Research in Multi-Agent Reinforcement Learning for Demand Response and Urban Energy Management. arXiv:2012.10504 [cs] (Dec . 2020 ). arXiv:2012.10504 [cs] Jose R. Vazquez-Canteli, Sourav Dey, Gregor Henze, and Zoltan Nagy. 2020. CityLearn: Standardizing Research in Multi-Agent Reinforcement Learning for Demand Response and Urban Energy Management. arXiv:2012.10504 [cs] (Dec. 2020). arXiv:2012.10504 [cs]"},{"key":"e_1_3_2_1_21_1","volume-title":"Reinforcement learning for demand response: A review of algorithms and modeling techniques. Applied energy 235","author":"V\u00e1zquez-Canteli Jos\u00e9 R","year":"2019","unstructured":"Jos\u00e9 R V\u00e1zquez-Canteli and Zolt\u00e1n Nagy . 2019. Reinforcement learning for demand response: A review of algorithms and modeling techniques. Applied energy 235 ( 2019 ), 1072--1089. Jos\u00e9 R V\u00e1zquez-Canteli and Zolt\u00e1n Nagy. 2019. Reinforcement learning for demand response: A review of algorithms and modeling techniques. Applied energy 235 (2019), 1072--1089."},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1145\/3061639.3062224"},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1145\/3408308.3427614"},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1109\/JIOT.2021.3078462"},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.1145\/3408308.3431119"},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1145\/3276774.3276775"}],"event":{"name":"e-Energy '22: The Thirteenth ACM International Conference on Future Energy Systems","sponsor":["SIGEnergy ACM Special Interest Group on Energy Systems and Informatics"],"location":"Virtual Event","acronym":"e-Energy '22"},"container-title":["Proceedings of the Thirteenth ACM International Conference on Future Energy Systems"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3538637.3538866","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3538637.3538866","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T19:03:02Z","timestamp":1750186982000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3538637.3538866"}},"subtitle":["a toolkit for unified access to building simulations for reinforcement learning"],"short-title":[],"issued":{"date-parts":[[2022,6,28]]},"references-count":26,"alternative-id":["10.1145\/3538637.3538866","10.1145\/3538637"],"URL":"https:\/\/doi.org\/10.1145\/3538637.3538866","relation":{},"subject":[],"published":{"date-parts":[[2022,6,28]]},"assertion":[{"value":"2022-06-28","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}