{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,4]],"date-time":"2025-10-04T08:01:53Z","timestamp":1759564913933,"version":"3.28.0"},"reference-count":16,"publisher":"IEEE","license":[{"start":{"date-parts":[[2019,11,1]],"date-time":"2019-11-01T00:00:00Z","timestamp":1572566400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2019,11,1]],"date-time":"2019-11-01T00:00:00Z","timestamp":1572566400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2019,11,1]],"date-time":"2019-11-01T00:00:00Z","timestamp":1572566400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2019,11]]},"DOI":"10.1109\/cis-ram47153.2019.9095802","type":"proceedings-article","created":{"date-parts":[[2020,5,19]],"date-time":"2020-05-19T17:09:55Z","timestamp":1589908195000},"page":"386-391","source":"Crossref","is-referenced-by-count":11,"title":["Trajectory-Tracking Control of Robotic Systems via Deep Reinforcement Learning"],"prefix":"10.1109","author":[{"given":"Shansi","family":"Zhang","sequence":"first","affiliation":[]},{"given":"Chao","family":"Sun","sequence":"additional","affiliation":[]},{"given":"Zhi","family":"Feng","sequence":"additional","affiliation":[]},{"given":"Guoqiang","family":"Hu","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref10","article-title":"Massively parallel methods for deep reinforcement learning","author":"nair","year":"2015","journal-title":"International Conference on Machine Learning"},{"key":"ref11","article-title":"Asynchronous Methods for Deep Reinforcement Learning","author":"mnih","year":"2016","journal-title":"International Conference on Machine Learning"},{"key":"ref12","article-title":"Distributed Prioritized Experience Replay","author":"horgan","year":"2018","journal-title":"arXiv preprint arXiv 1803 00933"},{"key":"ref13","first-page":"388","article-title":"Kinematic Modeling and Simulation of a SCARA Robot by Using Solid Dynamics and Verification by MATLAB\/Simulink","volume":"37","author":"mahdi","year":"2009","journal-title":"European Journal of Scientific Research"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1016\/j.robot.2016.01.002"},{"key":"ref15","article-title":"Prioritized Experience Replay","author":"schaul","year":"2015","journal-title":"arXiv preprint arXiv 1511 05952"},{"key":"ref16","article-title":"Deep-Mimic: Example-Guided Deep Reinforcement Learning of Physics-Based Character Skills","author":"peng","year":"2018","journal-title":"arXiv preprint arXiv 1804 02671"},{"key":"ref4","article-title":"Dueling Network Architectures for Deep Reinforcement Learning","author":"wang","year":"2015","journal-title":"arXiv preprint arXiv 1511 06841"},{"key":"ref3","doi-asserted-by":"crossref","DOI":"10.1609\/aaai.v30i1.10295","article-title":"Deep Reinforcement Learning with Double Q-learning","author":"van hasselt","year":"2016","journal-title":"Proceedings of the AAAI Conference on Artificial Intelligence"},{"key":"ref6","article-title":"Continuous control with deep reinforcement learning","author":"lillicrap","year":"2016","journal-title":"International Conference on Learning Representations"},{"key":"ref5","doi-asserted-by":"crossref","DOI":"10.1609\/aaai.v32i1.11796","article-title":"Rainbow: Combining Improvements in Deep Reinforcement Learning","author":"hessel","year":"2018","journal-title":"Proceedings of the AAAI Conference on Artificial Intelligence"},{"key":"ref8","article-title":"Learning Dexterous In-Hand Manipulation","author":"openai","year":"2018","journal-title":"arXiv preprint arXiv 1808 02194"},{"key":"ref7","article-title":"Deterministic Policy Gradient Algorithms","author":"silver","year":"2014","journal-title":"International Conference on Machine Learning"},{"key":"ref2","article-title":"Playing Atari with Deep Reinforcement Learning","author":"mnih","year":"2013","journal-title":"arXiv preprint arXiv 1312 5602"},{"key":"ref1","doi-asserted-by":"crossref","first-page":"529","DOI":"10.1038\/nature14236","article-title":"Human-level control through deep reinforcement learning","volume":"518","author":"mnih","year":"2015","journal-title":"Nature"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/IROS.2017.8202134"}],"event":{"name":"2019 IEEE International Conference on Cybernetics and Intelligent Systems (CIS) and IEEE Conference on Robotics, Automation and Mechatronics (RAM)","start":{"date-parts":[[2019,11,18]]},"location":"Bangkok, Thailand","end":{"date-parts":[[2019,11,20]]}},"container-title":["2019 IEEE International Conference on Cybernetics and Intelligent Systems (CIS) and IEEE Conference on Robotics, Automation and Mechatronics (RAM)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9094049\/9095541\/09095802.pdf?arnumber=9095802","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,10,24]],"date-time":"2022-10-24T03:35:38Z","timestamp":1666582538000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9095802\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019,11]]},"references-count":16,"URL":"https:\/\/doi.org\/10.1109\/cis-ram47153.2019.9095802","relation":{},"subject":[],"published":{"date-parts":[[2019,11]]}}}