{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,9,6]],"date-time":"2024-09-06T07:12:17Z","timestamp":1725606737301},"reference-count":23,"publisher":"IEEE","license":[{"start":{"date-parts":[[2022,9,1]],"date-time":"2022-09-01T00:00:00Z","timestamp":1661990400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2022,9,1]],"date-time":"2022-09-01T00:00:00Z","timestamp":1661990400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022,9,1]]},"DOI":"10.1109\/icac55051.2022.9911177","type":"proceedings-article","created":{"date-parts":[[2022,10,10]],"date-time":"2022-10-10T20:25:08Z","timestamp":1665433508000},"page":"1-6","source":"Crossref","is-referenced-by-count":0,"title":["Collision-free Path Planning For Welding Manipulator Via Deep Reinforcement Learning"],"prefix":"10.1109","author":[{"given":"Bo","family":"Hu","sequence":"first","affiliation":[{"name":"Guangdong University of Technology,Guangdong Provincial Key Laboratory of Cyber-Physical System,Guangzhou,China"}]},{"given":"Tao","family":"Wang","sequence":"additional","affiliation":[{"name":"Guangdong University of Technology,Guangdong Provincial Key Laboratory of Cyber-Physical System,Guangzhou,China"}]},{"given":"Chong","family":"Chen","sequence":"additional","affiliation":[{"name":"Guangdong University of Technology,Guangdong Provincial Key Laboratory of Cyber-Physical System,Guangzhou,China"}]},{"given":"Yintao","family":"Xu","sequence":"additional","affiliation":[{"name":"Guangdong University of Technology,Guangdong Provincial Key Laboratory of Cyber-Physical System,Guangzhou,China"}]},{"given":"Lianglun","family":"Cheng","sequence":"additional","affiliation":[{"name":"Guangdong University of Technology,Guangdong Provincial Key Laboratory of Cyber-Physical System,Guangzhou,China"}]}],"member":"263","reference":[{"journal-title":"Distributional Soft Actor-Critic Off-Policy Reinforcement Learning for Addressing Value Estimation Errors","year":"2020","author":"duan","key":"ref10"},{"key":"ref11","article-title":"Collision-free path planning for welding manipulator via hybrid algorithm of deep reinforcement learning and inverse kinematics","author":"zhong","year":"2021","journal-title":"Complex & Intelligent Systems"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2017.7989385"},{"journal-title":"Motion Planner Augmented Reinforcement Learning for Robot Manipulation in Obstructed Environments","year":"2020","author":"yamada","key":"ref13"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.15607\/RSS.2018.XIV.009"},{"key":"ref15","article-title":"Learning dexterous in-hand manipulation","author":"marcin","year":"2020","journal-title":"The International Journal of Robotics Research"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/ICRA.2019.8793789"},{"journal-title":"Data-efficient hierarchical reinforcement learning","year":"2021","author":"lee","key":"ref17"},{"journal-title":"Distributed Distributional Deterministic Policy Gradients","year":"2018","author":"barth-maron","key":"ref18"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/WCSP.2019.8928110"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1007\/s10846-020-01247-4"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/ETFA.2019.8869344"},{"key":"ref6","article-title":"Identification of the space posture of arc welding guns with rotational arc sensors in GMAW","author":"jian","year":"2017","journal-title":"International Journal of Advanced Manufacturing Technology"},{"key":"ref5","first-page":"1","article-title":"Reinforcement learning-based collision-free path planner for redundant robot in narrow duct","author":"hua","year":"2020","journal-title":"Journal of Intelligent Manufacturing"},{"journal-title":"Emergence of locomotion behaviours in rich environments","year":"2017","author":"heess","key":"ref8"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1016\/j.rcim.2019.101929"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1016\/j.compind.2020.103326"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1016\/j.rcim.2010.08.006"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1145\/3072959.3073602"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/ICRoM.2015.7367810"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/IROS40897.2019.8967959"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/ROBIO49542.2019.8961852"},{"journal-title":"Soft Actor-Critic Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor[C]","year":"2018","author":"haarnoja","key":"ref23"}],"event":{"name":"2022 27th International Conference on Automation and Computing (ICAC)","start":{"date-parts":[[2022,9,1]]},"location":"Bristol, United Kingdom","end":{"date-parts":[[2022,9,3]]}},"container-title":["2022 27th International Conference on Automation and Computing (ICAC)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9911058\/9911059\/09911177.pdf?arnumber=9911177","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,11,4]],"date-time":"2022-11-04T01:27:24Z","timestamp":1667525244000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9911177\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,9,1]]},"references-count":23,"URL":"https:\/\/doi.org\/10.1109\/icac55051.2022.9911177","relation":{},"subject":[],"published":{"date-parts":[[2022,9,1]]}}}