{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,24]],"date-time":"2025-11-24T20:13:15Z","timestamp":1764015195074,"version":"3.45.0"},"reference-count":31,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61803169"],"award-info":[{"award-number":["61803169"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Fundamental Research Funds for Central Universities","award":["2662018JC029"],"award-info":[{"award-number":["2662018JC029"]}]},{"name":"Hubei Key Research and Development Program of China","award":["2024BBB055","2024BAA008"],"award-info":[{"award-number":["2024BBB055","2024BAA008"]}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Automat. Sci. Eng."],"published-print":{"date-parts":[[2025]]},"DOI":"10.1109\/tase.2025.3627282","type":"journal-article","created":{"date-parts":[[2025,10,30]],"date-time":"2025-10-30T18:05:51Z","timestamp":1761847551000},"page":"24036-24052","source":"Crossref","is-referenced-by-count":0,"title":["An Adaptive Meta-Reinforcement Learning Framework for Dynamic Flexible Job Shop Scheduling"],"prefix":"10.1109","volume":"22","author":[{"ORCID":"https:\/\/orcid.org\/0009-0008-1791-5588","authenticated-orcid":false,"given":"Lincong","family":"Wu","sequence":"first","affiliation":[{"name":"College of Informatics, Huazhong Agricultural University, Wuhan, Hubei, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0887-4952","authenticated-orcid":false,"given":"Xiaoxia","family":"Li","sequence":"additional","affiliation":[{"name":"College of Informatics, Huazhong Agricultural University, Wuhan, Hubei, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9885-6031","authenticated-orcid":false,"given":"Xin","family":"Lu","sequence":"additional","affiliation":[{"name":"School of Computing and Creative Industries, Leeds Trinity University, Leeds, U.K."}]},{"given":"Zaiwen","family":"Feng","sequence":"additional","affiliation":[{"name":"College of Informatics, Huazhong Agricultural University, Wuhan, Hubei, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9581-4215","authenticated-orcid":false,"given":"Yanguo","family":"Jing","sequence":"additional","affiliation":[{"name":"Institute of Business, Industry and Leadership, University of Cumbria, Lancaster, U.K."}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1049\/iet-cim.2018.0009"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.3390\/app12031491"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/WSC.2010.5678946"},{"key":"ref4","first-page":"1","article-title":"Solving integrated process planning and scheduling problem with constructive meta-heuristics","volume":"340","author":"Zhang","year":"2016","journal-title":"Inf. Sci."},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1109\/TII.2022.3189725"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2023.3306421"},{"key":"ref7","first-page":"1621","article-title":"Learning to dispatch for job shop scheduling via deep reinforcement learning","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Zhang"},{"key":"ref8","article-title":"Learning to adapt in dynamic, real-world environments through meta-reinforcement learning","author":"Nagabandi","year":"2018","journal-title":"arXiv:1803.11347"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/TASLP.2022.3167258"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-33377-4_27"},{"key":"ref11","article-title":"DRL: Deep reinforcement learning for intelligent robot control\u2013concept, literature, and future","volume-title":"arXiv:2105.13806","author":"Dargazany","year":"2021"},{"key":"ref12","doi-asserted-by":"crossref","DOI":"10.1016\/j.cor.2024.106929","article-title":"A literature review of reinforcement learning methods applied to job-shop scheduling problems","volume":"175","author":"Zhang","year":"2025","journal-title":"Comput. Oper. Res."},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1016\/j.asoc.2022.109717"},{"issue":"10","key":"ref14","doi-asserted-by":"crossref","first-page":"4029","DOI":"10.3390\/app14104029","article-title":"Research on multi-objective flexible job shop scheduling problem with setup and handling based on an improved shuffled frog leaping algorithm","volume":"14","author":"Kong","year":"2024","journal-title":"Appl. Sci."},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1007\/s00170-010-2579-5"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1016\/j.ejor.2023.05.017"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2022.117796"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2019.2938548"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1088\/1742-6596\/2825\/1\/012013"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1007\/s11227-024-06741-2"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.3390\/electronics13183696"},{"key":"ref22","article-title":"Learning context-aware task reasoning for efficient meta-reinforcement learning","author":"Wang","year":"2020","journal-title":"arXiv:2003.01373"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1080\/00207543.2022.2027041"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/TASE.2019.2924444"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1016\/j.ifacol.2021.08.321"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/JIOT.2023.3241222"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1186\/s13677-023-00440-8"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/TCSI.2023.3240702"},{"key":"ref29","first-page":"1126","article-title":"Model-agnostic meta-learning for fast adaptation of deep networks","volume-title":"Proc. 34th Int. Conf. Mach. Learn.","volume":"70","author":"Finn"},{"key":"ref30","article-title":"Performance-weighed policy sampling for meta-reinforcement learning","author":"Ahmed","year":"2020","journal-title":"arXiv:2012.06016"},{"key":"ref31","article-title":"Meta learning shared hierarchies","author":"Frans","year":"2017","journal-title":"arXiv:1710.09767"}],"container-title":["IEEE Transactions on Automation Science and Engineering"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/8856\/10839176\/11222732.pdf?arnumber=11222732","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,24]],"date-time":"2025-11-24T19:03:13Z","timestamp":1764010993000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/11222732\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"references-count":31,"URL":"https:\/\/doi.org\/10.1109\/tase.2025.3627282","relation":{},"ISSN":["1545-5955","1558-3783"],"issn-type":[{"type":"print","value":"1545-5955"},{"type":"electronic","value":"1558-3783"}],"subject":[],"published":{"date-parts":[[2025]]}}}