{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,2]],"date-time":"2026-01-02T07:24:08Z","timestamp":1767338648702,"version":"3.37.3"},"reference-count":28,"publisher":"IEEE","license":[{"start":{"date-parts":[[2022,7,18]],"date-time":"2022-07-18T00:00:00Z","timestamp":1658102400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2022,7,18]],"date-time":"2022-07-18T00:00:00Z","timestamp":1658102400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61602149"],"award-info":[{"award-number":["61602149"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022,7,18]]},"DOI":"10.1109\/ijcnn55064.2022.9892166","type":"proceedings-article","created":{"date-parts":[[2022,9,30]],"date-time":"2022-09-30T19:56:04Z","timestamp":1664567764000},"page":"1-8","source":"Crossref","is-referenced-by-count":3,"title":["A Multi-Agent Deep Reinforcement Learning Framework for VWAP Strategy Optimization"],"prefix":"10.1109","author":[{"given":"Jiaqi","family":"Ye","sequence":"first","affiliation":[{"name":"College of Computer and Information Hohai University,Nanjing,China"}]},{"given":"Xiaodong","family":"Li","sequence":"additional","affiliation":[{"name":"College of Computer and Information Hohai University,Nanjing,China"}]},{"given":"Yingying","family":"Wang","sequence":"additional","affiliation":[{"name":"College of Computer and Information Hohai University,Nanjing,China"}]}],"member":"263","reference":[{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-47358-7_7"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1109\/IJCNN48605.2020.9207663"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/TITS.2019.2901791"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/TITS.2020.3023788"},{"key":"ref14","first-page":"3646","article-title":"A multi-agent reinforcement learning model of common-pool resource appropriation","author":"perolat","year":"0","journal-title":"Proceedings of the 31st International Conference on Neural Information Processing Systems"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.5220\/0006393400170026"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/IJCNN52387.2021.9534348"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.29007\/g87p"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2020\/623"},{"key":"ref19","article-title":"A modularized and scalable multi-agent reinforcement learning-based system for financial portfolio management","author":"huang","year":"2021","journal-title":"ArXiv Preprint"},{"key":"ref28","doi-asserted-by":"crossref","first-page":"529","DOI":"10.1038\/nature14236","article-title":"Human-level control through deep reinforcement learning","volume":"518","author":"mnih","year":"2015","journal-title":"Nature"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1145\/1143844.1143929"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-60990-0_12"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.21314\/JOR.2001.041"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-59410-7_42"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/CIFEr.2014.6924109"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2020\/627"},{"key":"ref7","article-title":"Double deep q-learning for optimal execution","author":"ning","year":"2018","journal-title":"ArXiv Preprint"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1111\/j.1540-6261.1988.tb02591.x"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1016\/S1386-4181(97)00012-8"},{"journal-title":"Algorithmic Trading and DMA","year":"2018","author":"johnson","key":"ref1"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1145\/3383455.3422570"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1145\/988772.988801"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1016\/j.jbankfin.2007.09.023"},{"key":"ref24","article-title":"Multi-agent deep reinforcement learning for liquidation strategy analysis","author":"bao","year":"2019","journal-title":"ArXiv Preprint"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1016\/S1386-4181(01)00023-4"},{"key":"ref26","doi-asserted-by":"crossref","first-page":"1095","DOI":"10.1073\/pnas.39.10.1095","article-title":"Stochastic games","volume":"39","author":"shapley","year":"0","journal-title":"Proceedings of the National Academy of Sciences"},{"key":"ref25","doi-asserted-by":"crossref","first-page":"887","DOI":"10.1016\/S1574-6526(07)03024-6","article-title":"Multi-agent systems","volume":"3","author":"van der hoek","year":"2008","journal-title":"Foundations of Artificial Intelligence"}],"event":{"name":"2022 International Joint Conference on Neural Networks (IJCNN)","start":{"date-parts":[[2022,7,18]]},"location":"Padua, Italy","end":{"date-parts":[[2022,7,23]]}},"container-title":["2022 International Joint Conference on Neural Networks (IJCNN)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9891857\/9889787\/09892166.pdf?arnumber=9892166","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,11,3]],"date-time":"2022-11-03T23:00:37Z","timestamp":1667516437000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9892166\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,7,18]]},"references-count":28,"URL":"https:\/\/doi.org\/10.1109\/ijcnn55064.2022.9892166","relation":{},"subject":[],"published":{"date-parts":[[2022,7,18]]}}}