{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,27]],"date-time":"2025-10-27T21:07:19Z","timestamp":1761599239050,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":18,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,5,12]],"date-time":"2022-05-12T00:00:00Z","timestamp":1652313600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022,5,12]]},"DOI":"10.1145\/3543712.3543722","type":"proceedings-article","created":{"date-parts":[[2022,9,20]],"date-time":"2022-09-20T22:24:32Z","timestamp":1663712672000},"page":"188-194","source":"Crossref","is-referenced-by-count":4,"title":["Split Feature Space Ensemble Method using Deep Reinforcement Learning for Algorithmic Trading"],"prefix":"10.1145","author":[{"given":"Marcell","family":"N\u00e9meth","sequence":"first","affiliation":[{"name":"Department of Telecommunications and Media Informatics, Budapest University of Technology and Economics, Hungary"}]},{"given":"G\u00e1bor","family":"Sz\u0171cs","sequence":"additional","affiliation":[{"name":"Department of Telecommunications and Media Informatics, Budapest University of Technology and Economics, Hungary"}]}],"member":"320","published-online":{"date-parts":[[2022,9,20]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"Multi-agent deep reinforcement learning for liquidation strategy analysis. arXiv preprint arXiv:1906.11046","author":"Bao W.","year":"2019","unstructured":"Bao , W. , & Liu , X. Y. ( 2019 ). Multi-agent deep reinforcement learning for liquidation strategy analysis. arXiv preprint arXiv:1906.11046 . Bao, W., & Liu, X. Y. (2019). Multi-agent deep reinforcement learning for liquidation strategy analysis. arXiv preprint arXiv:1906.11046."},{"issue":"5","key":"e_1_3_2_1_2_1","first-page":"679","article-title":"A Markovian Decision Process","volume":"6","author":"Bellman R.","year":"1957","unstructured":"Bellman , R. ( 1957 ). A Markovian Decision Process . Journal of Mathematics and Mechanics. 6 ( 5 ): 679 - 684 . JSTOR 24900506. Bellman, R. (1957). A Markovian Decision Process. Journal of Mathematics and Mechanics. 6 (5): 679-684. JSTOR 24900506.","journal-title":"Journal of Mathematics and Mechanics."},{"key":"e_1_3_2_1_3_1","volume-title":"Gonzalez Nieto J., Reif W., Wang G., Indulska J. (eds) Autonomic and Trusted Computing. ATC","author":"Dusparic I.","year":"2009","unstructured":"Dusparic I. , Cahill V. ( 2009 ) Using Reinforcement Learning for Multi-policy Optimization in Decentralized Autonomic Systems - An Experimental Evaluation . In: Gonzalez Nieto J., Reif W., Wang G., Indulska J. (eds) Autonomic and Trusted Computing. ATC 2009. Lecture Notes in Computer Science, vol 5586. Springer , Berlin, Heidelberg. https:\/\/doi.org\/10.1007\/978-3-642-02704-8_9 Dusparic I., Cahill V. (2009) Using Reinforcement Learning for Multi-policy Optimization in Decentralized Autonomic Systems - An Experimental Evaluation. In: Gonzalez Nieto J., Reif W., Wang G., Indulska J. (eds) Autonomic and Trusted Computing. ATC 2009. Lecture Notes in Computer Science, vol 5586. Springer, Berlin, Heidelberg. https:\/\/doi.org\/10.1007\/978-3-642-02704-8_9"},{"key":"e_1_3_2_1_4_1","volume-title":"A Deep Ensemble Multi-Agent Reinforcement Learning Approach for Air Traffic Control. arXiv preprint arXiv:2004.01387","author":"Ghosh S.","year":"2020","unstructured":"Ghosh , S. , Laguna , S. , Lim , S.H. , Wynter , L. , & Poonawala , H.A. ( 2020 ). A Deep Ensemble Multi-Agent Reinforcement Learning Approach for Air Traffic Control. arXiv preprint arXiv:2004.01387 . Ghosh, S., Laguna, S., Lim, S.H., Wynter, L., & Poonawala, H.A. (2020). A Deep Ensemble Multi-Agent Reinforcement Learning Approach for Air Traffic Control. arXiv preprint arXiv:2004.01387."},{"key":"e_1_3_2_1_5_1","volume-title":"Financial Trading as a Game: A Deep Reinforcement Learning Approach.\u00a0 ArXiv, abs\/1807.02787","author":"Huang C.","year":"2018","unstructured":"Huang , C. ( 2018 ). Financial Trading as a Game: A Deep Reinforcement Learning Approach.\u00a0 ArXiv, abs\/1807.02787 . Huang, C. (2018). Financial Trading as a Game: A Deep Reinforcement Learning Approach.\u00a0 ArXiv, abs\/1807.02787."},{"key":"e_1_3_2_1_6_1","volume-title":"Actor-Critic Algorithms","author":"Konda","year":"2001","unstructured":"Konda , Vijay & Tsitsiklis, John . ( 2001 ). Actor-Critic Algorithms . Society for Industrial and Applied Mathematics . 42. Konda, Vijay & Tsitsiklis, John. (2001). Actor-Critic Algorithms. Society for Industrial and Applied Mathematics. 42."},{"volume-title":"A Multiagent Approach to Q-Learning for Daily Stock Trading,\" in\u00a0 IEEE Transactions on Systems, Man, and Cybernetics - Part A: Systems and Humans","author":"Lee J. W.","key":"e_1_3_2_1_7_1","unstructured":"Lee , J. W. , Park J. , and Hong E. , \" A Multiagent Approach to Q-Learning for Daily Stock Trading,\" in\u00a0 IEEE Transactions on Systems, Man, and Cybernetics - Part A: Systems and Humans , vol. 37 , no. 6, pp. 864-877, Nov. 2007, doi: 10.1109\/TSMCA.2007.904825. Lee, J. W., Park J., and Hong E., \"A Multiagent Approach to Q-Learning for Daily Stock Trading,\" in\u00a0 IEEE Transactions on Systems, Man, and Cybernetics - Part A: Systems and Humans, vol. 37, no. 6, pp. 864-877, Nov. 2007, doi: 10.1109\/TSMCA.2007.904825."},{"key":"e_1_3_2_1_8_1","volume-title":"Continuous control with deep reinforcement learning.\u202fCoRR, abs\/1509.02971. https:\/\/arxiv.org\/abs\/1509.02971","author":"Lillicrap T.","year":"2016","unstructured":"Lillicrap , T. , Hunt , J. , Pritzel , A. , Heess , N. , Erez , T. , Tassa , Y. , Silver , D. , & Wierstra , D. ( 2016 ). Continuous control with deep reinforcement learning.\u202fCoRR, abs\/1509.02971. https:\/\/arxiv.org\/abs\/1509.02971 Lillicrap, T., Hunt, J., Pritzel, A., Heess, N., Erez, T., Tassa, Y., Silver, D., & Wierstra, D. (2016). Continuous control with deep reinforcement learning.\u202fCoRR, abs\/1509.02971. https:\/\/arxiv.org\/abs\/1509.02971"},{"key":"e_1_3_2_1_9_1","volume-title":"ArXiv abs\/2011.09607","author":"Liu X.Y.","year":"2020","unstructured":"Liu , X.Y. , Yang , H. , Chen , Q. , Zhang , R. , Yang , L. , Xiao , B. , Wang , C.D. ( 2020 ). FinRL: A Deep Reinforcement Learning Library for Automated Stock Trading in Quantitative Finance , ArXiv abs\/2011.09607 Liu, X.Y., Yang, H., Chen, Q., Zhang, R., Yang, L., Xiao, B., Wang, C.D. (2020). FinRL: A Deep Reinforcement Learning Library for Automated Stock Trading in Quantitative Finance, ArXiv abs\/2011.09607"},{"key":"e_1_3_2_1_10_1","volume-title":"International conference on machine learning (pp. 1928-1937)","author":"Mnih V.","year":"2016","unstructured":"Mnih , V. , Badia , A. P. , Mirza , M. , Graves , A. , Lillicrap , T. , Harley , T. , Lillicrap , T. P. , Silver , D. & Kavukcuoglu , K. ( 2016 ). Asynchronous methods for deep reinforcement learning . In International conference on machine learning (pp. 1928-1937) . PMLR. Mnih, V., Badia, A. P., Mirza, M., Graves, A., Lillicrap, T., Harley, T., Lillicrap, T. P., Silver, D. & Kavukcuoglu, K. (2016). Asynchronous methods for deep reinforcement learning. In International conference on machine learning (pp. 1928-1937). PMLR."},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1038\/nature14236"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"crossref","unstructured":"Mosavi A.; Faghan Y.; Ghamisi P.; Duan P.; Ardabili S.F.; Salwana E.; Band S.S. Comprehensive Review of Deep Reinforcement Learning Methods and Applications in Economics.\u00a0 Mathematics\u00a02020 \u00a08 1640. https:\/\/doi.org\/10.3390\/math8101640  Mosavi A.; Faghan Y.; Ghamisi P.; Duan P.; Ardabili S.F.; Salwana E.; Band S.S. Comprehensive Review of Deep Reinforcement Learning Methods and Applications in Economics.\u00a0 Mathematics\u00a02020 \u00a08 1640. https:\/\/doi.org\/10.3390\/math8101640","DOI":"10.3390\/math8101640"},{"key":"e_1_3_2_1_14_1","volume-title":"Proximal Policy Optimization Algorithms.\u00a0 ArXiv, abs\/1707.06347","author":"Schulman J.","year":"2017","unstructured":"Schulman , J. , Wolski , F. , Dhariwal , P. , Radford , A. , & Klimov , O. ( 2017 ). Proximal Policy Optimization Algorithms.\u00a0 ArXiv, abs\/1707.06347 . Schulman, J., Wolski, F., Dhariwal, P., Radford, A., & Klimov, O. (2017). Proximal Policy Optimization Algorithms.\u00a0 ArXiv, abs\/1707.06347."},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"crossref","first-page":"2123","DOI":"10.11118\/actaun201664062123","article-title":"Optimized Indicators of Technical Analysis on the New York Stock Exchange","volume":"64","author":"\u0160ir\u016f\u010dek","year":"2016","unstructured":"\u0160ir\u016f\u010dek , Martin & \u0160\u00edma, Karel . ( 2016 ). Optimized Indicators of Technical Analysis on the New York Stock Exchange . Acta Universitatis Agriculturae et Silviculturae Mendelianae Brunensis. 64. 2123 - 2131 . 10.11118\/actaun201664062123. \u0160ir\u016f\u010dek, Martin & \u0160\u00edma, Karel. (2016). Optimized Indicators of Technical Analysis on the New York Stock Exchange. Acta Universitatis Agriculturae et Silviculturae Mendelianae Brunensis. 64. 2123-2131. 10.11118\/actaun201664062123.","journal-title":"Acta Universitatis Agriculturae et Silviculturae Mendelianae Brunensis."},{"key":"e_1_3_2_1_16_1","volume-title":"Reinforcement Learning","author":"Sutton R.","year":"1998","unstructured":"Sutton , R. , Barto , A. ( 1998 ). Reinforcement Learning . MIT Press . ISBN 978-0-585-02445-5. Archived from the original on 2017-03-30. Sutton, R., Barto, A. (1998). Reinforcement Learning. MIT Press. ISBN 978-0-585-02445-5. Archived from the original on 2017-03-30."},{"key":"e_1_3_2_1_17_1","volume-title":"Ensemble Algorithms in Reinforcement Learning","author":"Wiering","year":"2008","unstructured":"Wiering , Marco & Van Hasselt, Hado . ( 2008 ). Ensemble Algorithms in Reinforcement Learning . IEEE transactions on systems, man, and cybernetics. Part B, Cybernetics : a publication of the IEEE Systems, Man, and Cybernetics Society . 38. 930-6. 10.1109\/TSMCB.2008.920231. Wiering, Marco & Van Hasselt, Hado. (2008). Ensemble Algorithms in Reinforcement Learning. IEEE transactions on systems, man, and cybernetics. Part B, Cybernetics : a publication of the IEEE Systems, Man, and Cybernetics Society. 38. 930-6. 10.1109\/TSMCB.2008.920231."},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"crossref","unstructured":"Yang Hongyang and Liu Xiao-Yang and Zhong Shan and Walid Anwar Deep Reinforcement Learning for Automated Stock Trading: An Ensemble Strategy (2020) http:\/\/dx.doi.org\/10.2139\/ssrn.3690996  Yang Hongyang and Liu Xiao-Yang and Zhong Shan and Walid Anwar Deep Reinforcement Learning for Automated Stock Trading: An Ensemble Strategy (2020) http:\/\/dx.doi.org\/10.2139\/ssrn.3690996","DOI":"10.2139\/ssrn.3690996"},{"issue":"2","key":"e_1_3_2_1_19_1","doi-asserted-by":"crossref","first-page":"25","DOI":"10.3905\/jfds.2020.1.030","article-title":"Deep reinforcement learning for trading","volume":"2","author":"Zhang Z.","year":"2020","unstructured":"Zhang , Z. , Zohren , S. , & Roberts , S. ( 2020 ). Deep reinforcement learning for trading . The Journal of Financial Data Science , 2 ( 2 ), 25 - 40 . Zhang, Z., Zohren, S., & Roberts, S. (2020). Deep reinforcement learning for trading. The Journal of Financial Data Science, 2(2), 25-40.","journal-title":"The Journal of Financial Data Science"}],"event":{"name":"ICCTA 2022: 2022 8th International Conference on Computer Technology Applications","acronym":"ICCTA 2022","location":"Kapfenberg Austria"},"container-title":["2022 8th International Conference on Computer Technology Applications"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3543712.3543722","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3543712.3543722","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T19:00:14Z","timestamp":1750186814000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3543712.3543722"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,5,12]]},"references-count":18,"alternative-id":["10.1145\/3543712.3543722","10.1145\/3543712"],"URL":"https:\/\/doi.org\/10.1145\/3543712.3543722","relation":{},"subject":[],"published":{"date-parts":[[2022,5,12]]}}}