{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,10,22]],"date-time":"2024-10-22T21:13:47Z","timestamp":1729631627828,"version":"3.28.0"},"reference-count":21,"publisher":"IEEE","license":[{"start":{"date-parts":[[2021,10,25]],"date-time":"2021-10-25T00:00:00Z","timestamp":1635120000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2021,10,25]],"date-time":"2021-10-25T00:00:00Z","timestamp":1635120000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2021,10,25]],"date-time":"2021-10-25T00:00:00Z","timestamp":1635120000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021,10,25]]},"DOI":"10.1109\/mlsp52302.2021.9596155","type":"proceedings-article","created":{"date-parts":[[2021,11,15]],"date-time":"2021-11-15T17:53:58Z","timestamp":1636998838000},"page":"1-6","source":"Crossref","is-referenced-by-count":3,"title":["Improving Deep Reinforcement Learning for Financial Trading Using Deep Adaptive Group-Based Normalization"],"prefix":"10.1109","author":[{"given":"Angelos","family":"Nalmpantis","sequence":"first","affiliation":[]},{"given":"Nikolaos","family":"Passalis","sequence":"additional","affiliation":[]},{"given":"Avraam","family":"Tsantekidis","sequence":"additional","affiliation":[]},{"given":"Anastasios","family":"Tefas","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref10","doi-asserted-by":"crossref","DOI":"10.1609\/aaai.v32i1.11796","article-title":"Rainbow: Combining improvements in deep reinforcement learning","author":"hessel","year":"2018","journal-title":"Proc AAAI Conf Artificial Intelligence"},{"key":"ref11","first-page":"5982","article-title":"Controllable neural story plot generation via reward shaping","author":"pradyumna","year":"2019","journal-title":"Proc Int Joint Conf Artificial Intelligence"},{"key":"ref12","article-title":"Critical learning periods in deep neural networks","author":"achille","year":"2017","journal-title":"ArXiv Preprint"},{"key":"ref13","article-title":"Deep adaptive input normalization for time series forecasting","author":"passalis","year":"2019","journal-title":"IEEE Trans on Neural Networks and Learning Systems"},{"journal-title":"Layer normalization","year":"2016","author":"ba","key":"ref14"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1007\/s10994-019-05788-0"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/TSMCB.2008.920231"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.123"},{"key":"ref18","article-title":"Fixing weight decay regularization in adam","volume":"abs 1711 5101","author":"loshchilov","year":"2017","journal-title":"CoRR"},{"key":"ref19","first-page":"80","article-title":"Eligibility traces for off-policy policy evaluation","author":"precup","year":"2000","journal-title":"Computer Science Department Faculty Publication Series"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2016.2522401"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/CBI.2017.23"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP.2019.8683161"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/SSCI.2017.8280812"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2019.112891"},{"key":"ref7","article-title":"A deep reinforcement learning framework for the financial portfolio management problem","author":"jiang","year":"2017","journal-title":"ArXiv Preprint"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/TSP.2019.2907260"},{"key":"ref1","volume":"217","author":"vidyamurthy","year":"2004","journal-title":"Pairs Trading Quantitative Methods and Analysis"},{"key":"ref9","article-title":"Proximal policy optimization algorithms","author":"schulman","year":"2017","journal-title":"ar Xiv preprint"},{"journal-title":"What matters in on-policy reinforcement learning? a large-scale empirical study","year":"2020","author":"andrychowicz","key":"ref20"},{"key":"ref21","article-title":"Batch normalization: Accelerating deep network training by reducing internal covariate shift","volume":"abs 1502 3167","author":"ioffe","year":"2015","journal-title":"CoRR"}],"event":{"name":"2021 IEEE 31st International Workshop on Machine Learning for Signal Processing (MLSP)","start":{"date-parts":[[2021,10,25]]},"location":"Gold Coast, Australia","end":{"date-parts":[[2021,10,28]]}},"container-title":["2021 IEEE 31st International Workshop on Machine Learning for Signal Processing (MLSP)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9596063\/9596068\/09596155.pdf?arnumber=9596155","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,11,12]],"date-time":"2023-11-12T12:37:47Z","timestamp":1699792667000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9596155\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,10,25]]},"references-count":21,"URL":"https:\/\/doi.org\/10.1109\/mlsp52302.2021.9596155","relation":{},"subject":[],"published":{"date-parts":[[2021,10,25]]}}}