{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T04:16:09Z","timestamp":1750220169264,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":16,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,12,1]],"date-time":"2022-12-01T00:00:00Z","timestamp":1669852800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"Vingroup Innovation Foundation (VINIF), Vingroup Big Data Institute (VinBigdata)","award":["VINIF.2021.STS.22"],"award-info":[{"award-number":["VINIF.2021.STS.22"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2022,12]]},"DOI":"10.1145\/3568562.3568579","type":"proceedings-article","created":{"date-parts":[[2022,11,29]],"date-time":"2022-11-29T00:25:01Z","timestamp":1669681501000},"page":"39-46","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":1,"title":["Benchmarking Gradient Estimation Mechanisms in Evolution Strategies for Solving Black-Box Optimization Functions and Reinforcement Learning Problems"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-5876-4092","authenticated-orcid":false,"given":"Thai Bao","family":"Tran","sequence":"first","affiliation":[{"name":"University of Information Technology, Viet Nam and Vietnam National University Ho Chi Minh City, Vietnam"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6768-1950","authenticated-orcid":false,"given":"Ngoc Hoang","family":"Luong","sequence":"additional","affiliation":[{"name":"University of Information Technology, Viet Nam and Vietnam National University Ho Chi Minh City, Vietnam"}]}],"member":"320","published-online":{"date-parts":[[2022,12]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"Mirrored Sampling and Sequential Selection for Evolution Strategies. In PPSN","author":"Brockhoff Dimo","year":"2010","unstructured":"Dimo Brockhoff , Anne Auger , Nikolaus Hansen , Dirk\u00a0 V. Arnold , and Tim Hohm . 2010 . Mirrored Sampling and Sequential Selection for Evolution Strategies. In PPSN 2010. https:\/\/doi.org\/10.1007\/978-3-642-15844-5_2 10.1007\/978-3-642-15844-5_2 Dimo Brockhoff, Anne Auger, Nikolaus Hansen, Dirk\u00a0V. Arnold, and Tim Hohm. 2010. Mirrored Sampling and Sequential Selection for Evolution Strategies. In PPSN 2010. https:\/\/doi.org\/10.1007\/978-3-642-15844-5_2"},{"key":"e_1_3_2_1_2_1","unstructured":"Greg Brockman Vicki Cheung Ludwig Pettersson Jonas Schneider John Schulman Jie Tang and Wojciech Zaremba. 2016. OpenAI Gym. CoRR abs\/1606.01540(2016). arXiv:1606.01540http:\/\/arxiv.org\/abs\/1606.01540  Greg Brockman Vicki Cheung Ludwig Pettersson Jonas Schneider John Schulman Jie Tang and Wojciech Zaremba. 2016. OpenAI Gym. CoRR abs\/1606.01540(2016). arXiv:1606.01540http:\/\/arxiv.org\/abs\/1606.01540"},{"key":"e_1_3_2_1_3_1","volume-title":"Structured Evolution with Compact Architectures for Scalable Policy Optimization. In ICML","author":"Choromanski Krzysztof","year":"2018","unstructured":"Krzysztof Choromanski , Mark Rowland , Vikas Sindhwani , Richard\u00a0 E. Turner , and Adrian Weller . 2018 . Structured Evolution with Compact Architectures for Scalable Policy Optimization. In ICML 2018. Krzysztof Choromanski, Mark Rowland, Vikas Sindhwani, Richard\u00a0E. Turner, and Adrian Weller. 2018. Structured Evolution with Compact Architectures for Scalable Policy Optimization. In ICML 2018."},{"key":"e_1_3_2_1_4_1","volume-title":"Benchmarking Deep Reinforcement Learning for Continuous Control. In ICML","author":"Duan Yan","year":"2016","unstructured":"Yan Duan , Xi Chen , Rein Houthooft , John Schulman , and Pieter Abbeel . 2016 . Benchmarking Deep Reinforcement Learning for Continuous Control. In ICML 2016. Yan Duan, Xi Chen, Rein Houthooft, John Schulman, and Pieter Abbeel. 2016. Benchmarking Deep Reinforcement Learning for Continuous Control. In ICML 2016."},{"key":"e_1_3_2_1_5_1","unstructured":"Nikolaus Hansen. 2016. The CMA Evolution Strategy: A Tutorial. CoRR abs\/1604.00772(2016).  Nikolaus Hansen. 2016. The CMA Evolution Strategy: A Tutorial. CoRR abs\/1604.00772(2016)."},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1162\/106365601750190398"},{"key":"e_1_3_2_1_7_1","volume-title":"Self-Guided Evolution Strategies with Historical Estimated Gradients. In IJCAI","author":"Liu Fei-Yu","year":"2020","unstructured":"Fei-Yu Liu , Zi-Niu Li , and Chao Qian . 2020 . Self-Guided Evolution Strategies with Historical Estimated Gradients. In IJCAI 2020. Fei-Yu Liu, Zi-Niu Li, and Chao Qian. 2020. Self-Guided Evolution Strategies with Historical Estimated Gradients. In IJCAI 2020."},{"key":"e_1_3_2_1_8_1","volume-title":"ICML","author":"Maheswaranathan Niru","year":"2019","unstructured":"Niru Maheswaranathan , Luke Metz , George Tucker , Dami Choi , and Jascha Sohl-Dickstein . 2019 . Guided evolutionary strategies: augmenting random search with surrogate gradients . In ICML 2019. Niru Maheswaranathan, Luke Metz, George Tucker, Dami Choi, and Jascha Sohl-Dickstein. 2019. Guided evolutionary strategies: augmenting random search with surrogate gradients. In ICML 2019."},{"key":"e_1_3_2_1_9_1","volume-title":"NeurIPS","author":"Mania Horia","year":"2018","unstructured":"Horia Mania , Aurelia Guy , and Benjamin Recht . 2018. Simple random search of static linear policies is competitive for reinforcement learning . In NeurIPS 2018 . 1805\u20131814. Horia Mania, Aurelia Guy, and Benjamin Recht. 2018. Simple random search of static linear policies is competitive for reinforcement learning. In NeurIPS 2018. 1805\u20131814."},{"key":"e_1_3_2_1_10_1","volume-title":"Human-level control through deep reinforcement learning. Nat. 518, 7540","author":"Mnih Volodymyr","year":"2015","unstructured":"Volodymyr Mnih , Koray Kavukcuoglu , David Silver , Andrei\u00a0 A. Rusu , Joel Veness , Marc\u00a0 G. Bellemare , Alex Graves , Martin\u00a0 A. Riedmiller , Andreas Fidjeland , Georg Ostrovski , Stig Petersen , Charles Beattie , Amir Sadik , Ioannis Antonoglou , Helen King , Dharshan Kumaran , Daan Wierstra , Shane Legg , and Demis Hassabis . 2015. Human-level control through deep reinforcement learning. Nat. 518, 7540 ( 2015 ), 529\u2013533. Volodymyr Mnih, Koray Kavukcuoglu, David Silver, Andrei\u00a0A. Rusu, Joel Veness, Marc\u00a0G. Bellemare, Alex Graves, Martin\u00a0A. Riedmiller, Andreas Fidjeland, Georg Ostrovski, Stig Petersen, Charles Beattie, Amir Sadik, Ioannis Antonoglou, Helen King, Dharshan Kumaran, Daan Wierstra, Shane Legg, and Demis Hassabis. 2015. Human-level control through deep reinforcement learning. Nat. 518, 7540 (2015), 529\u2013533."},{"key":"e_1_3_2_1_11_1","volume-title":"Challenges in High-Dimensional Reinforcement Learning with Evolution Strategies. In PPSN","author":"M\u00fcller Nils","year":"2018","unstructured":"Nils M\u00fcller and Tobias Glasmachers . 2018 . Challenges in High-Dimensional Reinforcement Learning with Evolution Strategies. In PPSN 2018. https:\/\/doi.org\/10.1007\/978-3-319-99259-4_33 10.1007\/978-3-319-99259-4_33 Nils M\u00fcller and Tobias Glasmachers. 2018. Challenges in High-Dimensional Reinforcement Learning with Evolution Strategies. In PPSN 2018. https:\/\/doi.org\/10.1007\/978-3-319-99259-4_33"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10208-015-9296-2"},{"key":"e_1_3_2_1_13_1","volume-title":"Enhancing Twin Delayed Deep Deterministic Policy Gradient with Cross-Entropy Method. In NICS 2021","author":"Nguyen Hieu\u00a0Trung","year":"2021","unstructured":"Hieu\u00a0Trung Nguyen , Khang Tran , and Ngoc\u00a0Hoang Luong . 2021 . Enhancing Twin Delayed Deep Deterministic Policy Gradient with Cross-Entropy Method. In NICS 2021 . https:\/\/doi.org\/10.1109\/NICS54270.2021.9701549 10.1109\/NICS54270.2021.9701549 Hieu\u00a0Trung Nguyen, Khang Tran, and Ngoc\u00a0Hoang Luong. 2021. Enhancing Twin Delayed Deep Deterministic Policy Gradient with Cross-Entropy Method. In NICS 2021. https:\/\/doi.org\/10.1109\/NICS54270.2021.9701549"},{"key":"e_1_3_2_1_14_1","unstructured":"Art\u00a0B. Owen. 2013. Monte Carlo theory methods and examples.  Art\u00a0B. Owen. 2013. Monte Carlo theory methods and examples."},{"key":"e_1_3_2_1_15_1","volume-title":"ICLR","author":"Pourchot Alo\u00efs","year":"2019","unstructured":"Alo\u00efs Pourchot and Olivier Sigaud . 2019 . CEM-RL: Combining evolutionary and gradient-based methods for policy search . In ICLR 2019,. Alo\u00efs Pourchot and Olivier Sigaud. 2019. CEM-RL: Combining evolutionary and gradient-based methods for policy search. In ICLR 2019,."},{"key":"e_1_3_2_1_16_1","unstructured":"Tim Salimans Jonathan Ho Xi Chen and Ilya Sutskever. 2017. Evolution Strategies as a Scalable Alternative to Reinforcement Learning. CoRR abs\/1703.03864(2017). arXiv:1703.03864http:\/\/arxiv.org\/abs\/1703.03864  Tim Salimans Jonathan Ho Xi Chen and Ilya Sutskever. 2017. Evolution Strategies as a Scalable Alternative to Reinforcement Learning. CoRR abs\/1703.03864(2017). arXiv:1703.03864http:\/\/arxiv.org\/abs\/1703.03864"}],"event":{"name":"SoICT 2022: The 11th International Symposium on Information and Communication Technology","acronym":"SoICT 2022","location":"Hanoi Vietnam"},"container-title":["The 11th International Symposium on Information and Communication Technology"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3568562.3568579","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3568562.3568579","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T19:00:39Z","timestamp":1750186839000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3568562.3568579"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,12]]},"references-count":16,"alternative-id":["10.1145\/3568562.3568579","10.1145\/3568562"],"URL":"https:\/\/doi.org\/10.1145\/3568562.3568579","relation":{},"subject":[],"published":{"date-parts":[[2022,12]]},"assertion":[{"value":"2022-12-01","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}