{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,29]],"date-time":"2025-10-29T03:48:15Z","timestamp":1761709695852,"version":"3.44.0"},"reference-count":32,"publisher":"IEEE","license":[{"start":{"date-parts":[[2019,7,1]],"date-time":"2019-07-01T00:00:00Z","timestamp":1561939200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2019,7,1]],"date-time":"2019-07-01T00:00:00Z","timestamp":1561939200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2019,7]]},"DOI":"10.1109\/indin41052.2019.8972157","type":"proceedings-article","created":{"date-parts":[[2020,1,31]],"date-time":"2020-01-31T00:16:43Z","timestamp":1580429803000},"page":"1775-1780","source":"Crossref","is-referenced-by-count":8,"title":["A framework for scheduling in cloud manufacturing with deep reinforcement learning"],"prefix":"10.1109","author":[{"given":"Yongkui","family":"Liu","sequence":"first","affiliation":[{"name":"Xidian University,Center for Smart Manufacturing Systems,Xi&#x2019;an,China"}]},{"given":"Lin","family":"Zhang","sequence":"additional","affiliation":[{"name":"Xidian University,Center for Smart Manufacturing Systems,Xi&#x2019;an,China"}]},{"given":"Lihui","family":"Wang","sequence":"additional","affiliation":[{"name":"Xidian University,Center for Smart Manufacturing Systems,Xi&#x2019;an,China"}]},{"given":"Yingying","family":"Xiao","sequence":"additional","affiliation":[{"name":"Beijing Institute of Electronic System Engineering,State Key Laboratory of Intelligent, Manufacturing System Technology,Beijing,China"}]},{"given":"Xun","family":"Xu","sequence":"additional","affiliation":[{"name":"The University of Auckland,Department of Mechanical Engineering,Auckland,New Zealand"}]},{"given":"Mei","family":"Wang","sequence":"additional","affiliation":[{"name":"Beijing Institute of Electronic System Engineering,State Key Laboratory of Intelligent, Manufacturing System Technology,Beijing,China"}]}],"member":"263","reference":[{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2018.2872674"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-04182-3_26"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1007\/s10586-007-0035-6"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1016\/j.rcim.2016.09.008"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1038\/nature16961"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1038\/nature24270"},{"key":"ref13","first-page":"201","article-title":"Playing Atari with deep reinforcement learning","author":"mnih","year":"2013","journal-title":"Proceedings of the Workshops at the 26th Neural Information Processing Systems 2013"},{"key":"ref14","doi-asserted-by":"crossref","first-page":"529","DOI":"10.1038\/nature14236","article-title":"Human-level control through deep reinforcement learning","volume":"518","author":"mnih","year":"2015","journal-title":"Nature"},{"key":"ref15","first-page":"2094","article-title":"Deep reinforcement learning with double Q learning","author":"van hasselt","year":"2016","journal-title":"Proceedings of the AAAI Conference on Artificial Intelligence Phoenix"},{"key":"ref16","first-page":"1476","article-title":"Increasing the action gap: New operators for reinforcement learning","author":"bellemare","year":"2016","journal-title":"Proceedings of the AAAI Conference on Artificial Intelligence"},{"key":"ref17","first-page":"322","article-title":"Prioritized experience replay","author":"schaul","year":"2016","journal-title":"Proceedings of International Conference on Learning Representations"},{"key":"ref18","first-page":"1995","article-title":"Dueling network architectures for deep reinforcement learning","author":"wang","year":"2016","journal-title":"Proceedings of the International Conference on Machine Learning"},{"key":"ref19","article-title":"Deep recurrent q-learning for partially observable mdps","author":"hausknecht","year":"2015","journal-title":"AAAI Fall Symposium Series"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1016\/j.procir.2018.03.212"},{"key":"ref4","article-title":"Deep reinforcement learning: An overview","author":"li","year":"2017","journal-title":"arXiv preprint arXiv 1701 07717"},{"key":"ref27","first-page":"675","article-title":"Reinforcement learning integrated with simulation for job-shop scheduling system","volume":"22","author":"pan","year":"2007","journal-title":"Control and Decision"},{"key":"ref3","article-title":"Scheduling in cloud manufacturing: state-of-the-art and research challenges","author":"liu","year":"2018","journal-title":"International Journal of Production Research"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1007\/s00170-015-7350-5"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1016\/j.jpdc.2017.05.001"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1145\/3005745.3005750"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/JSYST.2015.2438054"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1007\/s00170-017-1167-3"},{"key":"ref2","first-page":"2226","article-title":"A revisit to cloud manufacturing","volume":"29","author":"liu","year":"2018","journal-title":"China Mechanical Engineering"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1115\/1.4034186"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1080\/17517575.2012.683812"},{"key":"ref20","first-page":"1","article-title":"Continuous control with deep reinforcement learning","author":"lillicrap","year":"2016","journal-title":"International Conference on Learning Representations"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/ROBOT.1995.525285"},{"key":"ref21","first-page":"1928","article-title":"Asynchronous methods for deep reinforcement learning","author":"mnih","year":"2016","journal-title":"International Conference on Machine Learning"},{"key":"ref24","first-page":"83","article-title":"Flow-shop scheduling based on reinforcement learning algorithm","volume":"1","author":"stef\u00e1n","year":"2003","journal-title":"Production Systems and Information Engineering"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1016\/j.engappai.2004.08.018"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1016\/j.engappai.2004.08.018"},{"key":"ref25","doi-asserted-by":"crossref","first-page":"280","DOI":"10.1109\/ICII.2001.983070","article-title":"Reinforcement learning approach to re-entrant manufacturing system scheduling","volume":"3","author":"liu","year":"2001","journal-title":"IEEE International Conferences on Info-Tech and Info-Net Proceedings (Cat No 01EX479)"}],"event":{"name":"2019 IEEE 17th International Conference on Industrial Informatics (INDIN)","start":{"date-parts":[[2019,7,22]]},"location":"Helsinki, Finland","end":{"date-parts":[[2019,7,25]]}},"container-title":["2019 IEEE 17th International Conference on Industrial Informatics (INDIN)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/8961950\/8972012\/08972157.pdf?arnumber=8972157","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,5]],"date-time":"2025-09-05T18:09:30Z","timestamp":1757095770000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/8972157\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019,7]]},"references-count":32,"URL":"https:\/\/doi.org\/10.1109\/indin41052.2019.8972157","relation":{},"subject":[],"published":{"date-parts":[[2019,7]]}}}