{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,15]],"date-time":"2025-12-15T14:20:42Z","timestamp":1765808442751,"version":"build-2065373602"},"reference-count":317,"publisher":"MDPI AG","issue":"3","license":[{"start":{"date-parts":[[2025,1,24]],"date-time":"2025-01-24T00:00:00Z","timestamp":1737676800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100001871","name":"Funda\u00e7\u00e3o para a Ci\u00eancia e a Tecnologia (FCT)","doi-asserted-by":"publisher","award":["10.54499\/UIDB\/50022\/2020","10.54499\/UIDP\/50022\/2020","10.54499\/2020.08776.BD"],"award-info":[{"award-number":["10.54499\/UIDB\/50022\/2020","10.54499\/UIDP\/50022\/2020","10.54499\/2020.08776.BD"]}],"id":[{"id":"10.13039\/501100001871","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["Applied Sciences"],"abstract":"<jats:p>This paper presents a systematic review on reinforcement learning approaches for combinatorial optimization problems based on real-world industrial applications. While this topic is increasing in popularity, explicit implementation details are not always available in the literature. The main objective of this paper is characterizing the agent\u2013environment interactions, namely, the state space representation, action space mapping and reward design. Also, the main limitations for practical implementation and the needed future developments are identified. The literature selected covers a wide range of industrial combinatorial optimization problems, found in the IEEE Xplore, Scopus and Web of Science databases. A total of 715 unique papers were extracted from the query. Then, out-of-scope applications, reviews, surveys and papers with insufficient implementation details were removed. This resulted in a total of 298 papers that align with the focus of the review with sufficient implementation details. The state space representation shows the most variety, while the reward design is based on combinations of different modules. The presented studies use a large variety of features and strategies. However, one of the main limitations is that even with state-of-the-art complex models the scalability issues of increasing problem complexity cannot be fully solved. No methods were used to assess risk of biases or automatically synthesize the results.<\/jats:p>","DOI":"10.3390\/app15031211","type":"journal-article","created":{"date-parts":[[2025,1,24]],"date-time":"2025-01-24T10:46:07Z","timestamp":1737715567000},"page":"1211","update-policy":"https:\/\/doi.org\/10.3390\/mdpi_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["A Systematic Review on Reinforcement Learning for Industrial Combinatorial Optimization Problems"],"prefix":"10.3390","volume":"15","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-6285-8737","authenticated-orcid":false,"given":"Miguel S. E.","family":"Martins","sequence":"first","affiliation":[{"name":"IDMEC, Instituto Superior T\u00e9cnico, Universidade de Lisboa, 1049-001 Lisbon, Portugal"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8030-4746","authenticated-orcid":false,"given":"Jo\u00e3o M. C.","family":"Sousa","sequence":"additional","affiliation":[{"name":"IDMEC, Instituto Superior T\u00e9cnico, Universidade de Lisboa, 1049-001 Lisbon, Portugal"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7961-1004","authenticated-orcid":false,"given":"Susana","family":"Vieira","sequence":"additional","affiliation":[{"name":"IDMEC, Instituto Superior T\u00e9cnico, Universidade de Lisboa, 1049-001 Lisbon, Portugal"}]}],"member":"1968","published-online":{"date-parts":[[2025,1,24]]},"reference":[{"key":"ref_1","unstructured":"Sutton, R.S., and Barto, A.G. (2020). Reinforcement Learning, MIT Press."},{"key":"ref_2","doi-asserted-by":"crossref","unstructured":"Patel, P.P., and Jhaveri, R.H. (2016, January 29\u201330). Soft computing techniques to address various issues in wireless sensor networks: A survey. Proceedings of the IEEE International Conference on Computing, Communication and Automation, ICCCA 2016, Greater Noida, India.","DOI":"10.1109\/CCAA.2016.7813753"},{"key":"ref_3","doi-asserted-by":"crossref","first-page":"1928","DOI":"10.1002\/er.4333","article-title":"A new generation of AI: A review and perspective on machine learning technologies applied to smart energy and electric power systems","volume":"43","author":"Cheng","year":"2019","journal-title":"Int. J. Energy Res."},{"key":"ref_4","unstructured":"Cunha, B., Madureira, A.M., Fonseca, B., and Coelho, D. (2018, January 13\u201315). Deep Reinforcement Learning as a Job Shop Scheduling Solver: A Literature Review. Proceedings of the 18th International Conference on Hybrid Intelligent Systems (HIS 2018), Porto, Portugal."},{"key":"ref_5","doi-asserted-by":"crossref","first-page":"137184","DOI":"10.1109\/ACCESS.2019.2942390","article-title":"Machine learning for 5G\/B5G mobile and wireless communications: Potential, limitations, and future directions","volume":"7","author":"Lee","year":"2019","journal-title":"IEEE Access"},{"key":"ref_6","doi-asserted-by":"crossref","first-page":"13","DOI":"10.1016\/j.arcontrol.2020.08.003","article-title":"On the requirements of digital twin-driven autonomous maintenance","volume":"50","author":"Khan","year":"2020","journal-title":"Annu. Rev. Control"},{"key":"ref_7","doi-asserted-by":"crossref","first-page":"209320","DOI":"10.1109\/ACCESS.2020.3038605","article-title":"A Gentle Introduction to Reinforcement Learning and its Application in Different Fields","volume":"8","author":"Naeem","year":"2020","journal-title":"IEEE Access"},{"key":"ref_8","doi-asserted-by":"crossref","unstructured":"Quach, H.N., Yeom, S., and Kim, K. (2020, January 17\u201319). Survey on reinforcement learning based efficient routing in SDN. Proceedings of the 9th International Conference on Smart Media and Applications, Jeju, Republic of Korea.","DOI":"10.1145\/3426020.3426072"},{"key":"ref_9","doi-asserted-by":"crossref","first-page":"98","DOI":"10.1016\/j.comcom.2021.07.014","article-title":"Reinforcement and deep reinforcement learning for wireless Internet of Things: A survey","volume":"178","author":"Frikha","year":"2021","journal-title":"Comput. Commun."},{"key":"ref_10","doi-asserted-by":"crossref","first-page":"2064","DOI":"10.1109\/COMST.2021.3102580","article-title":"Leveraging Deep Reinforcement Learning for Traffic Engineering: A Survey","volume":"23","author":"Xiao","year":"2021","journal-title":"IEEE Commun. Surv. Tutor."},{"key":"ref_11","doi-asserted-by":"crossref","first-page":"5772","DOI":"10.1080\/00207543.2022.2104180","article-title":"Reinforcement learning applied to production planning and control","volume":"61","author":"Esteso","year":"2023","journal-title":"Int. J. Prod. Res."},{"key":"ref_12","doi-asserted-by":"crossref","unstructured":"Torres, A.d.R., Andreiana, D.S., Rold\u00e1n, \u00c1.O., Bustos, A.H., and Galicia, L.E.A. (2022). A Review of Deep Reinforcement Learning Approaches for Smart Manufacturing in Industry 4.0 and 5.0 Framework. Appl. Sci., 12.","DOI":"10.3390\/app122312377"},{"key":"ref_13","doi-asserted-by":"crossref","first-page":"244","DOI":"10.1016\/j.jmsy.2023.07.014","article-title":"Reinforcement and deep reinforcement learning-based solutions for machine maintenance planning, scheduling policies, and optimization","volume":"70","author":"Ogunfowora","year":"2023","journal-title":"J. Manuf. Syst."},{"key":"ref_14","doi-asserted-by":"crossref","first-page":"n71","DOI":"10.1136\/bmj.n71","article-title":"The PRISMA 2020 statement: An updated guideline for reporting systematic reviews","volume":"372","author":"Page","year":"2021","journal-title":"BMJ"},{"key":"ref_15","doi-asserted-by":"crossref","first-page":"e1230","DOI":"10.1002\/cl2.1230","article-title":"PRISMA2020: An R package and Shiny app for producing PRISMA 2020-compliant flow diagrams, with interactivity for optimised digital transparency and Open Synthesis","volume":"18","author":"Haddaway","year":"2022","journal-title":"Campbell Syst. Rev."},{"key":"ref_16","doi-asserted-by":"crossref","unstructured":"Bai, Y., and Lv, Y. (2022, January 7\u201310). Reinforcement Learning-based Job Shop Scheduling for Remanufacturing Production. Proceedings of the IEEE International Conference on Industrial Engineering and Engineering Management, Kuala Lumpur, Malaysia.","DOI":"10.1109\/IEEM55944.2022.9989643"},{"key":"ref_17","doi-asserted-by":"crossref","first-page":"e13315","DOI":"10.1111\/exsy.13315","article-title":"Addressing deadlock in large-scale, complex rail networks via multi-agent deep reinforcement learning","volume":"42","author":"Bretas","year":"2023","journal-title":"Expert Syst."},{"key":"ref_18","doi-asserted-by":"crossref","unstructured":"Chang, J., Yu, D., Zhou, Z., He, W., and Zhang, L. (2022). Hierarchical Reinforcement Learning for Multi-Objective Real-Time Flexible Scheduling in a Smart Shop Floor. Machines, 10.","DOI":"10.3390\/machines10121195"},{"key":"ref_19","doi-asserted-by":"crossref","first-page":"1638","DOI":"10.1109\/TVLSI.2021.3097712","article-title":"A Reinforcement Learning-Based Framework for Solving the IP Mapping Problem","volume":"29","author":"Chen","year":"2021","journal-title":"IEEE Trans. Very Large Scale Integr. Syst."},{"key":"ref_20","doi-asserted-by":"crossref","unstructured":"Danino, T., Ben-Shimol, Y., and Greenberg, S. (2023). Container Allocation in Cloud Environment Using Multi-Agent Deep Reinforcement Learning. Electronics, 12.","DOI":"10.3390\/electronics12122614"},{"key":"ref_21","doi-asserted-by":"crossref","unstructured":"Geng, N., Lan, T., Aggarwal, V., Yang, Y., and Xu, M. (2020, January 13\u201316). A Multi-agent Reinforcement Learning Perspective on Distributed Traffic Engineering. Proceedings of the 2020 IEEE 28th International Conference on Network Protocols (ICNP), Madrid, Spain.","DOI":"10.1109\/ICNP49622.2020.9259413"},{"key":"ref_22","doi-asserted-by":"crossref","first-page":"186474","DOI":"10.1109\/ACCESS.2020.3029868","article-title":"Research on adaptive job shop scheduling problems based on dueling double DQN","volume":"8","author":"Han","year":"2020","journal-title":"IEEE Access"},{"key":"ref_23","doi-asserted-by":"crossref","first-page":"2285","DOI":"10.1109\/LCOMM.2021.3074756","article-title":"Dynamic Resource Configuration for Low-Power IoT Networks: A Multi-Objective Reinforcement Learning Method","volume":"25","author":"Huang","year":"2021","journal-title":"IEEE Commun. Lett."},{"key":"ref_24","doi-asserted-by":"crossref","first-page":"1695","DOI":"10.1109\/TPDS.2021.3124670","article-title":"Performance and Cost-Efficient Spark Job Scheduling Based on Deep Reinforcement Learning in Cloud Computing Environments","volume":"33","author":"Islam","year":"2022","journal-title":"IEEE Trans. Parallel Distrib. Syst."},{"key":"ref_25","doi-asserted-by":"crossref","first-page":"99","DOI":"10.1016\/j.ejor.2012.03.020","article-title":"Reinforcement learning for joint pricing, lead-time and scheduling decisions in make-to-order systems","volume":"221","author":"Li","year":"2012","journal-title":"Eur. J. Oper. Res."},{"key":"ref_26","doi-asserted-by":"crossref","unstructured":"Li, X., Fang, Y., Pan, C., Cai, Y., and Zhou, M. (2023). Resource Scheduling for UAV-Assisted Failure-Prone MEC in Industrial Internet. Drones, 7.","DOI":"10.3390\/drones7040259"},{"key":"ref_27","doi-asserted-by":"crossref","unstructured":"Liu, W., Wu, S., Zhu, H., and Zhang, H. (2021, January 1\u20134). An Integration Method of Heterogeneous Models for Process Scheduling Based on Deep Q-Learning Integration Agent. Proceedings of the 16th IEEE Conference on Industrial Electronics and Applications, ICIEA 2021, Chengdu, China.","DOI":"10.1109\/ICIEA51954.2021.9516381"},{"key":"ref_28","doi-asserted-by":"crossref","unstructured":"Ma, S., Ilyushkin, A., Stegehuis, A., and Iosup, A. (2017, January 17\u201321). Ananke: A Q-Learning-Based Portfolio Scheduler for Complex Industrial Workflows. Proceedings of the 2017 IEEE International Conference on Autonomic Computing, ICAC 2017, Columbus, OH, USA.","DOI":"10.1109\/ICAC.2017.21"},{"key":"ref_29","doi-asserted-by":"crossref","first-page":"10810","DOI":"10.1016\/j.ifacol.2020.12.2866","article-title":"Reinforcement learning for dual-resource constrained scheduling","volume":"53","author":"Martins","year":"2020","journal-title":"IFAC-PapersOnLine"},{"key":"ref_30","doi-asserted-by":"crossref","unstructured":"Moon, J., Yang, M., and Jeong, J. (2021). A novel approach to the job shop scheduling problem based on the deep Q-network in a cooperative multi-access edge computing ecosystem. Sensors, 21.","DOI":"10.3390\/s21134553"},{"key":"ref_31","doi-asserted-by":"crossref","first-page":"4171","DOI":"10.1007\/s10586-022-03630-2","article-title":"A novel deep reinforcement learning scheme for task scheduling in cloud computing","volume":"25","author":"Siddesha","year":"2022","journal-title":"Clust. Comput."},{"key":"ref_32","doi-asserted-by":"crossref","first-page":"194","DOI":"10.1016\/j.promfg.2020.01.026","article-title":"Production flow control through the use of reinforcement learning","volume":"38","author":"Silva","year":"2019","journal-title":"Procedia Manuf."},{"key":"ref_33","doi-asserted-by":"crossref","unstructured":"Williem, R.S., and Setiawan, K. (2011, January 10\u201312). Reinforcement learning combined with radial basis function neural network to solve job-shop scheduling problem. Proceedings of the APBITM 2011\u20142011 IEEE International Summer Conference of Asia Pacific Business Innovation and Technology Management, Dalian, China.","DOI":"10.1109\/APBITM.2011.5996285"},{"key":"ref_34","doi-asserted-by":"crossref","first-page":"106306","DOI":"10.1016\/j.cor.2023.106306","article-title":"A DRL based approach for adaptive scheduling of one-of-a-kind production","volume":"158","author":"Wang","year":"2023","journal-title":"Comput. Oper. Res."},{"key":"ref_35","doi-asserted-by":"crossref","unstructured":"Xu, N., and Bu, T.M. (2022, January 9\u201313). Policy network for solving flexible job shop scheduling problem with setup times and rescoure constraints. Proceedings of the GECCO 2022 Companion\u20142022 Genetic and Evolutionary Computation Conference, Boston, MA, USA.","DOI":"10.1145\/3520304.3529084"},{"key":"ref_36","doi-asserted-by":"crossref","unstructured":"Yang, Y., Chen, X., Yang, M., Guo, W., and Jiang, P. (2024). Designing an Industrial Product Service System for Robot-Driven Sanding Processing Line: A Reinforcement Learning Based Approach. Machines, 12.","DOI":"10.3390\/machines12020136"},{"key":"ref_37","doi-asserted-by":"crossref","first-page":"102230","DOI":"10.1016\/j.aei.2023.102230","article-title":"A multi-agent double Deep-Q-network based on state machine and event stream for flexible job shop scheduling problem","volume":"58","author":"Yuan","year":"2023","journal-title":"Adv. Eng. Inform."},{"key":"ref_38","doi-asserted-by":"crossref","unstructured":"Yu, L., Yu, P.S., Duan, Y., and Qiao, H. (2022). A resource scheduling method for reliable and trusted distributed composite services in cloud environment based on deep reinforcement learning. Front. Genet., 13.","DOI":"10.3389\/fgene.2022.964784"},{"key":"ref_39","doi-asserted-by":"crossref","unstructured":"Zhang, C., Odonkor, P., Zheng, S., Khorasgani, H., Serita, S., Gupta, C., and Wang, H. (2020, January 10\u201313). Dynamic Dispatching for Large-Scale Heterogeneous Fleet via Multi-agent Deep Reinforcement Learning. Proceedings of the 2020 IEEE International Conference on Big Data, Big Data 2020, Atlanta, GA, USA.","DOI":"10.1109\/BigData50022.2020.9378191"},{"key":"ref_40","doi-asserted-by":"crossref","unstructured":"Zhang, M., Lu, Y., Hu, Y., Amaitik, N., and Xu, Y. (2022). Dynamic Scheduling Method for Job-Shop Manufacturing Systems by Deep Reinforcement Learning with Proximal Policy Optimization. Sustainability, 14.","DOI":"10.3390\/su14095177"},{"key":"ref_41","doi-asserted-by":"crossref","first-page":"122995","DOI":"10.1109\/ACCESS.2021.3110242","article-title":"Dynamic Jobshop Scheduling Algorithm Based on Deep Q Network","volume":"9","author":"Zhao","year":"2021","journal-title":"IEEE Access"},{"key":"ref_42","doi-asserted-by":"crossref","first-page":"1445","DOI":"10.3934\/mbe.2024062","article-title":"An actor-critic framework based on deep reinforcement learning for addressing flexible job shop scheduling problems","volume":"21","author":"Zhao","year":"2024","journal-title":"Math. Biosci. Eng."},{"key":"ref_43","doi-asserted-by":"crossref","first-page":"174","DOI":"10.1109\/JIOT.2023.3300011","article-title":"Enhancing AIoT Device Association With Task Offloading in Aerial MEC Networks","volume":"11","author":"Chen","year":"2024","journal-title":"IEEE Internet Things J."},{"key":"ref_44","doi-asserted-by":"crossref","unstructured":"Gao, Y., Wu, W., Nan, H., Sun, Y., and Si, P. (2020, January 7\u201311). Deep Reinforcement Learning based Task Scheduling in Mobile Blockchain for IoT Applications. Proceedings of the ICC 2020\u20142020 IEEE International Conference on Communications (ICC), Dublin, Ireland.","DOI":"10.1109\/ICC40277.2020.9148888"},{"key":"ref_45","doi-asserted-by":"crossref","unstructured":"Geurtsen, M., Adan, I., and Atan, Z. (2022, January 11\u201314). Dynamic Scheduling of Maintenance by a Reinforcement Learning Approach\u2014A Semiconductor Simulation Study. Proceedings of the Winter Simulation Conference, Singapore.","DOI":"10.1109\/WSC57314.2022.10015402"},{"key":"ref_46","unstructured":"Gong, Y., Sun, S., Wei, Y., and Song, M. (2021, January 25\u201328). Deep Reinforcement Learning for Edge Computing Resource Allocation in Blockchain Network Slicing Broker Framework. Proceedings of the IEEE Vehicular Technology Conference, Helsinki, Finland."},{"key":"ref_47","doi-asserted-by":"crossref","first-page":"380","DOI":"10.1109\/TNET.2022.3194953","article-title":"Delay-Oriented Scheduling in 5G Downlink Wireless Networks Based on Reinforcement Learning With Partial Observations","volume":"31","author":"Hao","year":"2023","journal-title":"IEEE\/ACM Trans. Netw."},{"key":"ref_48","doi-asserted-by":"crossref","unstructured":"Lamprecht, R., Wurst, F., and Huber, M.F. (2021, January 21\u201323). Reinforcement Learning based Condition-oriented Maintenance Scheduling for Flow Line Systems. Proceedings of the IEEE International Conference on Industrial Informatics (INDIN), Palma de Mallorca, Spain.","DOI":"10.1109\/INDIN45523.2021.9557373"},{"key":"ref_49","doi-asserted-by":"crossref","unstructured":"Lei, K., Guo, P., Wang, Y., Xiong, J., and Zhao, W. (2022, January 18\u201323). An End-to-end Hierarchical Reinforcement Learning Framework for Large-scale Dynamic Flexible Job-shop Scheduling Problem. Proceedings of the International Joint Conference on Neural Networks, Padua, Italy.","DOI":"10.1109\/IJCNN55064.2022.9892005"},{"key":"ref_50","doi-asserted-by":"crossref","unstructured":"Li, Y.L., Fadda, E., Manerba, D., Roohnavazfar, M., Tadei, R., and Terzo, O. (2022). Online Single-Machine Scheduling via Reinforcement Learning. Recent Advances in Computational Optimization, Springer.","DOI":"10.1007\/978-3-030-82397-9_5"},{"key":"ref_51","doi-asserted-by":"crossref","first-page":"4179","DOI":"10.1109\/JIOT.2021.3102831","article-title":"LSTM-Characterized Deep Reinforcement Learning for Continuous Flight Control and Resource Allocation in UAV-Assisted Sensor Network","volume":"9","author":"Li","year":"2022","journal-title":"IEEE Internet Things J."},{"key":"ref_52","doi-asserted-by":"crossref","first-page":"2932","DOI":"10.1016\/j.ifacol.2022.10.177","article-title":"Dynamic scheduling of a due date constrained flow shop with Deep Reinforcement Learning","volume":"55","author":"Marchesano","year":"2022","journal-title":"IFAC-PapersOnLine"},{"key":"ref_53","doi-asserted-by":"crossref","first-page":"173","DOI":"10.1080\/23335777.2021.1900922","article-title":"Q-Learning Based Optimisation Framework for Real-Time Mixed-Task Scheduling","volume":"8","author":"Meng","year":"2022","journal-title":"Cyber-Phys. Syst."},{"key":"ref_54","doi-asserted-by":"crossref","first-page":"227","DOI":"10.1016\/j.procir.2022.02.183","article-title":"A Q-Learning algorithm for flexible job shop scheduling in a real-world manufacturing scenario","volume":"106","author":"Palacio","year":"2022","journal-title":"Procedia CIRP"},{"key":"ref_55","doi-asserted-by":"crossref","unstructured":"Raeissi, M.M., Brooks, N., and Farinelli, A. (November, January 30). A Balking Queue Approach for Modeling Human-Multi-Robot Interaction for Water Monitoring. Proceedings of the PRIMA 2017: Principles and Practice of Multi-Agent Systems\u201420th International Conference, Nice, France. 10621 LNAI.","DOI":"10.1007\/978-3-319-69131-2_13"},{"key":"ref_56","doi-asserted-by":"crossref","first-page":"3979","DOI":"10.1007\/s00170-019-03940-7","article-title":"Modeling, planning, and scheduling of shop-floor assembly process with dynamic cyber-physical interactions: A case study for CPS-based smart industrial robot production","volume":"105","author":"Tan","year":"2019","journal-title":"Int. J. Adv. Manuf. Technol."},{"key":"ref_57","doi-asserted-by":"crossref","unstructured":"Tassel, P., Kov\u00e1cs, B., Gebser, M., Schekotihin, K., Kohlenbrein, W., and Schrott-Kostwein, P. (2022, January 13\u201324). Reinforcement Learning of Dispatching Strategies for Large-Scale Industrial Scheduling. Proceedings of the International Conference on Automated Planning and Scheduling, ICAPS, Virtual.","DOI":"10.1609\/icaps.v32i1.19852"},{"key":"ref_58","doi-asserted-by":"crossref","first-page":"1501","DOI":"10.1109\/TCE.2023.3332454","article-title":"Sustainable Fog-Assisted Intelligent Monitoring Framework for Consumer Electronics in Industry 5.0 Applications","volume":"70","author":"Tripathy","year":"2024","journal-title":"IEEE Trans. Consum. Electron."},{"key":"ref_59","doi-asserted-by":"crossref","unstructured":"Thomas, T.E., Koo, J., Chaterji, S., and Bagchi, S. (2018, January 3\u20137). Minerva: A reinforcement learning-based technique for optimal scheduling and bottleneck detection in distributed factory operations. Proceedings of the 2018 10th International Conference on Communication Systems and Networks, COMSNETS 2018, Bengaluru, India.","DOI":"10.1109\/COMSNETS.2018.8328189"},{"key":"ref_60","doi-asserted-by":"crossref","first-page":"518","DOI":"10.1016\/j.jmsy.2022.07.016","article-title":"Opportunistic maintenance scheduling with deep reinforcement learning","volume":"64","author":"Valet","year":"2022","journal-title":"J. Manuf. Syst."},{"key":"ref_61","doi-asserted-by":"crossref","unstructured":"Xing, Y., Yang, L., Hu, X., Mei, C., Wang, H., and Li, J. (2023, January 14\u201316). 6G Deterministic Network Technology Based on Hierarchical Reinforcement Learning Framework. Proceedings of the IEEE International Symposium on Broadband Multimedia Systems and Broadcasting, BMSB, Beijing, China.","DOI":"10.1109\/BMSB58369.2023.10211210"},{"key":"ref_62","doi-asserted-by":"crossref","first-page":"2593","DOI":"10.1007\/s10845-023-02161-w","article-title":"Smart scheduling of dynamic job shop based on discrete event simulation and deep reinforcement learning","volume":"35","author":"Wang","year":"2023","journal-title":"J. Intell. Manuf."},{"key":"ref_63","doi-asserted-by":"crossref","unstructured":"Yang, D., Gong, K., Zhang, W., Guo, K., and Chen, J. (2023, January 18\u201321). enDRTS: Deep Reinforcement Learning Based Deterministic Scheduling for Chain Flows in TSN. Proceedings of the 2023 International Conference on Networking and Network Applications (NaNA), Qingdao, China.","DOI":"10.1109\/NaNA60121.2023.00047"},{"key":"ref_64","unstructured":"Zhang, Z., Li, S., Yan, X., and Zhang, L. (2018, January 10\u201313). Self-organizing network control with a TD learning algorithm. Proceedings of the IEEE International Conference on Industrial Engineering and Engineering Management, Singapore."},{"key":"ref_65","doi-asserted-by":"crossref","unstructured":"Zhang, T., Shen, S., Mao, S., and Chang, G.K. (2020, January 7\u201311). Delay-aware Cellular Traffic Scheduling with Deep Reinforcement Learning. Proceedings of the GLOBECOM 2020\u20142020 IEEE Global Communications Conference, Taipei, Taiwan.","DOI":"10.1109\/GLOBECOM42002.2020.9322560"},{"key":"ref_66","doi-asserted-by":"crossref","first-page":"8999","DOI":"10.1109\/TII.2022.3178410","article-title":"Distributed Real-Time Scheduling in Cloud Manufacturing by Deep Reinforcement Learning","volume":"18","author":"Zhang","year":"2022","journal-title":"IEEE Trans. Ind. Inform."},{"key":"ref_67","doi-asserted-by":"crossref","first-page":"5526","DOI":"10.1109\/JIOT.2023.3306803","article-title":"Cooperative Partial Task Offloading and Resource Allocation for IIoT Based on Decentralized Multi-Agent Deep Reinforcement Learning","volume":"11","author":"Zhang","year":"2024","journal-title":"IEEE Internet Things J."},{"key":"ref_68","doi-asserted-by":"crossref","unstructured":"Zhu, Y., Sun, L., Wang, J., Huang, R., and Jia, X. (2023). Deep Reinforcement Learning-Based Joint Scheduling of 5G and TSN in Industrial Networks. Electronics , 12.","DOI":"10.3390\/electronics12122686"},{"key":"ref_69","doi-asserted-by":"crossref","first-page":"2513","DOI":"10.1007\/s10845-011-0580-y","article-title":"Dynamic scheduling for multi-site companies: A decisional approach based on reinforcement multi-agent learning","volume":"23","author":"Aissani","year":"2012","journal-title":"J. Intell. Manuf."},{"key":"ref_70","doi-asserted-by":"crossref","first-page":"62459","DOI":"10.1109\/ACCESS.2022.3182009","article-title":"Deep Reinforcement Learning Based Routing in IP Media Broadcast Networks: Feasibility and Performance","volume":"10","author":"Amaral","year":"2022","journal-title":"IEEE Access"},{"key":"ref_71","doi-asserted-by":"crossref","unstructured":"Bulbul, N.S., and Fischer, M. (2022, January 4\u20138). Reinforcement Learning assisted Routing for Time Sensitive Networks. Proceedings of the GLOBECOM 2022\u20142022 IEEE Global Communications Conference, Rio de Janeiro, Brazil.","DOI":"10.1109\/GLOBECOM48099.2022.10001630"},{"key":"ref_72","doi-asserted-by":"crossref","first-page":"61","DOI":"10.1109\/MNET.001.1800505","article-title":"Improving cognitive ability of edge intelligent IIoT through machine learning","volume":"33","author":"Chen","year":"2019","journal-title":"IEEE Netw."},{"key":"ref_73","doi-asserted-by":"crossref","first-page":"674","DOI":"10.1016\/j.robot.2008.12.001","article-title":"Multi-robot task allocation through vacancy chain scheduling","volume":"57","author":"Dahl","year":"2009","journal-title":"Robot. Auton. Syst."},{"key":"ref_74","doi-asserted-by":"crossref","unstructured":"Farahani, A., Genga, L., and DIjkman, R. (2021, January 17\u201320). Online Multimodal Transportation Planning using Deep Reinforcement Learning. Proceedings of the 2021 IEEE International Conference on Systems, Man and Cybernetics, Melbourne, Australia.","DOI":"10.1109\/SMC52423.2021.9658943"},{"key":"ref_75","doi-asserted-by":"crossref","first-page":"7931","DOI":"10.1109\/TITS.2021.3096537","article-title":"A Distributed Model-Free Ride-Sharing Approach for Joint Matching, Pricing, and Dispatching Using Deep Reinforcement Learning","volume":"22","author":"Haliem","year":"2021","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"key":"ref_76","doi-asserted-by":"crossref","first-page":"6255","DOI":"10.1109\/TII.2021.3130255","article-title":"CeCO: Cost-Efficient Computation Offloading of IoT Applications in Green Industrial Fog Networks","volume":"18","author":"Hazra","year":"2022","journal-title":"IEEE Trans. Ind. Inform."},{"key":"ref_77","doi-asserted-by":"crossref","first-page":"68","DOI":"10.1016\/j.procir.2023.06.013","article-title":"Delivery scheduling in meat industry using reinforcement learning","volume":"118","author":"Pargmann","year":"2023","journal-title":"Procedia CIRP"},{"key":"ref_78","doi-asserted-by":"crossref","first-page":"109650","DOI":"10.1016\/j.cie.2023.109650","article-title":"A Cooperative Hierarchical Deep Reinforcement Lerning based Multi-agent Method for Distributed Job Shop Scheduling Problem with Random Job Arrivals","volume":"185","author":"Huang","year":"2023","journal-title":"Comput. Ind. Eng."},{"key":"ref_79","doi-asserted-by":"crossref","first-page":"106982","DOI":"10.1016\/j.compchemeng.2020.106982","article-title":"A deep reinforcement learning approach for chemical production scheduling","volume":"141","author":"Hubbs","year":"2020","journal-title":"Comput. Chem. Eng."},{"key":"ref_80","doi-asserted-by":"crossref","first-page":"1007","DOI":"10.1109\/TII.2023.3272661","article-title":"Large-Scale Dynamic Scheduling for Flexible Job-Shop With Random Arrivals of New Jobs by Hierarchical Reinforcement Learning","volume":"20","author":"Lei","year":"2024","journal-title":"IEEE Trans. Ind. Inform."},{"key":"ref_81","doi-asserted-by":"crossref","first-page":"13572","DOI":"10.1109\/TCYB.2021.3111082","article-title":"Deep Reinforcement Learning for Solving the Heterogeneous Capacitated Vehicle Routing Problem","volume":"52","author":"Li","year":"2022","journal-title":"IEEE Trans. Cybern."},{"key":"ref_82","doi-asserted-by":"crossref","first-page":"4151","DOI":"10.1109\/TNSM.2022.3191748","article-title":"DRL-Based Long-Term Resource Planning for Task Offloading Policies in Multiserver Edge Computing Networks","volume":"19","author":"Li","year":"2022","journal-title":"IEEE Trans. Netw. Serv. Manag."},{"key":"ref_83","doi-asserted-by":"crossref","first-page":"18077","DOI":"10.1007\/s00500-023-09269-4","article-title":"Optimizing warehouse logistics scheduling strategy using soft computing and advanced machine learning techniques","volume":"27","author":"Li","year":"2023","journal-title":"Soft Comput."},{"key":"ref_84","doi-asserted-by":"crossref","first-page":"8218","DOI":"10.1109\/TII.2022.3216295","article-title":"Bi-Dueling DQN Enhanced Two-stage Scheduling for Augmented Surveillance in Smart EMS","volume":"19","author":"Liang","year":"2022","journal-title":"IEEE Trans. Ind. Inform."},{"key":"ref_85","doi-asserted-by":"crossref","first-page":"107744","DOI":"10.1109\/ACCESS.2019.2932047","article-title":"Which Channel to Ask My Question? Personalized Customer Service Request Stream Routing Using Deep Reinforcement Learning","volume":"7","author":"Liu","year":"2019","journal-title":"IEEE Access"},{"key":"ref_86","doi-asserted-by":"crossref","first-page":"2276","DOI":"10.1109\/JIOT.2020.3015772","article-title":"Communication-Efficient Federated Learning and Permissioned Blockchain for Digital Twin Edge Networks","volume":"8","author":"Lu","year":"2021","journal-title":"IEEE Internet Things J."},{"key":"ref_87","doi-asserted-by":"crossref","unstructured":"M\u00e9ndez-Hern\u00e1ndez, B.M., Rodr\u00edguez-Bazan, E.D., Martinez-Jimenez, Y., Libin, P., and Now\u00e9, A. (2019, January 17\u201319). A Multi-objective Reinforcement Learning Algorithm for JSSP. Proceedings of the 28th International Conference on Artificial Neural Networks, Munich, Germany.","DOI":"10.1007\/978-3-030-30487-4_44"},{"key":"ref_88","doi-asserted-by":"crossref","unstructured":"Mhaisen, N., Fetais, N., and Massoud, A. (2020, January 2\u20135). Real-Time Scheduling for Electric Vehicles Charging\/Discharging Using Reinforcement Learning. Proceedings of the 2020 IEEE International Conference on Informatics, IoT, and Enabling Technologies, ICIoT 2020, Doha, Qatar.","DOI":"10.1109\/ICIoT48696.2020.9089471"},{"key":"ref_89","doi-asserted-by":"crossref","first-page":"714","DOI":"10.1007\/s10696-023-09496-9","article-title":"A reinforcement learning\/ad-hoc planning and scheduling mechanism for flexible and sustainable manufacturing systems","volume":"36","author":"Paraschos","year":"2024","journal-title":"Flex. Serv. Manuf. J."},{"key":"ref_90","first-page":"1420","article-title":"A Reinforcement Learning Approach to Robust Scheduling of Semiconductor Manufacturing Facilities","volume":"17","author":"Park","year":"2020","journal-title":"IEEE Trans. Autom. Sci. Eng."},{"key":"ref_91","doi-asserted-by":"crossref","unstructured":"Roy, S.B., and Tan, E. (2023, January 20\u201323). Multi-hop Computational Offloading with Reinforcement Learning for Industrial IoT Networks. Proceedings of the 2023 IEEE 97th Vehicular Technology Conference (VTC2023-Spring), Florence, Italy.","DOI":"10.1109\/VTC2023-Spring57618.2023.10200149"},{"key":"ref_92","unstructured":"Schneider, J.G., Boyan, J.A., and Moore, A.W. (1998, January 18). Stochastic Production Scheduling to meet Demand Forecasts. Proceedings of the 37th IEEE Conference on Decision & Control, Tampa, FL, USA."},{"key":"ref_93","doi-asserted-by":"crossref","first-page":"2173","DOI":"10.1109\/TASE.2023.3289915","article-title":"Digital Twin-Driven Reinforcement Learning Method for Marine Equipment Vehicles Scheduling Problem","volume":"21","author":"Shen","year":"2024","journal-title":"IEEE Trans. Autom. Sci. Eng."},{"key":"ref_94","doi-asserted-by":"crossref","first-page":"1457","DOI":"10.1109\/TASE.2023.3248229","article-title":"Stochastic Economic Lot Scheduling via Self-Attention Based Deep Reinforcement Learning","volume":"21","author":"Song","year":"2024","journal-title":"IEEE Trans. Autom. Sci. Eng."},{"key":"ref_95","doi-asserted-by":"crossref","first-page":"11030","DOI":"10.1109\/TII.2023.3242769","article-title":"Energy and Performance-Efficient Dynamic Consolidate VMs Using Deep-Q Neural Network","volume":"19","author":"Tong","year":"2023","journal-title":"IEEE Trans. Ind. Inform."},{"key":"ref_96","doi-asserted-by":"crossref","first-page":"418","DOI":"10.3390\/make5020025","article-title":"A Reinforcement Learning Approach for Scheduling Problems with Improved Generalization through Order Swapping","volume":"5","author":"Vivekanandan","year":"2023","journal-title":"Mach. Learn. Knowl. Extr."},{"key":"ref_97","doi-asserted-by":"crossref","first-page":"1050","DOI":"10.1109\/TCCN.2024.3358565","article-title":"Priority-Aware Deployment of Autoscaling Service Function Chains based On Deep Reinforcement Learning","volume":"10","author":"Yu","year":"2024","journal-title":"IEEE Trans. Cogn. Commun. Netw."},{"key":"ref_98","doi-asserted-by":"crossref","first-page":"849","DOI":"10.1109\/TII.2019.2950809","article-title":"Reinforcement Learning for Real-Time Pricing and Scheduling Control in EV Charging Stations","volume":"17","author":"Wang","year":"2021","journal-title":"IEEE Trans. Ind. Inform."},{"key":"ref_99","doi-asserted-by":"crossref","first-page":"446","DOI":"10.1109\/JSAC.2022.3227097","article-title":"User Scheduling and Task Offloading in Multi-Tier Computing 6G Vehicular Network","volume":"41","author":"Zhang","year":"2023","journal-title":"IEEE J. Sel. Areas Commun."},{"key":"ref_100","doi-asserted-by":"crossref","first-page":"118","DOI":"10.1109\/IOTM.001.2200264","article-title":"QoS-Driven Distributed Cooperative Data Offloading and Heterogeneous Resource Scheduling for IIoT","volume":"6","author":"Zhang","year":"2023","journal-title":"IEEE Internet Things Mag."},{"key":"ref_101","doi-asserted-by":"crossref","unstructured":"Zhang, J., Kong, L., and Zhang, H. (2023, January 7\u20139). Coordinated Ride-hailing Order Scheduling and Charging for Autonomous Electric Vehicles Based on Deep Reinforcement Learning. Proceedings of the 2023 IEEE IAS Industrial and Commercial Power System Asia, I and CPS Asia 2023, Chongqing, China.","DOI":"10.1109\/ICPSAsia58343.2023.10294915"},{"key":"ref_102","doi-asserted-by":"crossref","first-page":"105308","DOI":"10.1016\/j.autcon.2024.105308","article-title":"Scheduling optimization of electric ready mixed concrete vehicles using an improved model-based reinforcement learning","volume":"160","author":"Chen","year":"2024","journal-title":"Autom. Constr."},{"key":"ref_103","doi-asserted-by":"crossref","first-page":"1308","DOI":"10.1109\/JIOT.2020.3003398","article-title":"Soft Actor-Critic DRL for Live Transcoding and Streaming in Vehicular Fog-Computing-Enabled IoV","volume":"8","author":"Fu","year":"2021","journal-title":"IEEE Internet Things J."},{"key":"ref_104","doi-asserted-by":"crossref","first-page":"12029","DOI":"10.1109\/JIOT.2021.3132849","article-title":"Cost-Efficient and Quality-of-Experience-Aware Player Request Scheduling and Rendering Server Allocation for Edge-Computing-Assisted Multiplayer Cloud Gaming","volume":"9","author":"Gao","year":"2022","journal-title":"IEEE Internet Things J."},{"key":"ref_105","doi-asserted-by":"crossref","unstructured":"Huang, Y., Sun, Y., and Ding, Z. (2022, January 8\u201311). Renewable Energy Integration Driven Charging Scheme for Electric Vehicle Based Large Scale Delivery System. Proceedings of the 2022 IEEE\/IAS Industrial and Commercial Power System Asia (I&CPS Asia), Shanghai, China.","DOI":"10.1109\/ICPSAsia55496.2022.9949661"},{"key":"ref_106","doi-asserted-by":"crossref","unstructured":"Ingalalli, A., Kamalasadan, S., Dong, Z., Bharati, G., and Chakraborty, S. (2022, January 9\u201314). An Extended Q-Routing-based Event-driven Dynamic Reconfiguration of Networked Microgrids. Proceedings of the 2022 IEEE Industry Applications Society Annual Meeting (IAS), Detroit, MI, USA.","DOI":"10.1109\/IAS54023.2022.9939942"},{"key":"ref_107","doi-asserted-by":"crossref","first-page":"47863","DOI":"10.1109\/ACCESS.2023.3268754","article-title":"Multi-Agent Learning and Bargaining Scheme for Cooperative Spectrum Sharing Process","volume":"11","author":"Kim","year":"2023","journal-title":"IEEE Access"},{"key":"ref_108","doi-asserted-by":"crossref","first-page":"116222","DOI":"10.1016\/j.eswa.2021.116222","article-title":"Deep reinforcement learning based scheduling within production plan in semiconductor fabrication","volume":"191","author":"Lee","year":"2022","journal-title":"Expert Syst. Appl."},{"key":"ref_109","doi-asserted-by":"crossref","first-page":"4419","DOI":"10.1080\/00207543.2022.2142314","article-title":"Reinforcement learning-based dynamic production-logistics-integrated tasks allocation in smart factories","volume":"61","author":"Lei","year":"2023","journal-title":"Int. J. Prod. Res."},{"key":"ref_110","doi-asserted-by":"crossref","first-page":"9175","DOI":"10.1109\/JIOT.2020.3006090","article-title":"Learning-Based Autonomous Scheduling for AoI-Aware Industrial Wireless Networks","volume":"7","author":"Li","year":"2020","journal-title":"IEEE Internet Things J."},{"key":"ref_111","doi-asserted-by":"crossref","first-page":"15725","DOI":"10.1109\/JIOT.2022.3151862","article-title":"Predictive Maintenance Model for IIoT-Based Manufacturing: A Transferable Deep Reinforcement Learning Approach","volume":"9","author":"Ong","year":"2022","journal-title":"IEEE Internet Things J."},{"key":"ref_112","doi-asserted-by":"crossref","unstructured":"Onishi, T., Takahashi, E., Nishikawa, Y., and Maruyama, S. (2023, January 8\u201311). AppDAS: An Application QoS-Aware Distributed Antenna Selection for 5G Industrial Applications. Proceedings of the 2023 IEEE 20th Consumer Communications & Networking Conference (CCNC), Las Vegas, NV, USA.","DOI":"10.1109\/CCNC51644.2023.10059796"},{"key":"ref_113","doi-asserted-by":"crossref","first-page":"851","DOI":"10.1109\/JRFID.2022.3211555","article-title":"A Parallel Learning Approach for the Flexible Job Shop Scheduling Problem","volume":"6","author":"Peng","year":"2022","journal-title":"IEEE J. Radio Freq. Identif."},{"key":"ref_114","doi-asserted-by":"crossref","first-page":"1853","DOI":"10.1109\/TNSM.2020.3002828","article-title":"Cooperative Network Model for Joint Mobile Sink Scheduling and Dynamic Buffer Management Using Q-Learning","volume":"17","author":"Redhu","year":"2020","journal-title":"IEEE Trans. Netw. Serv. Manag."},{"key":"ref_115","doi-asserted-by":"crossref","first-page":"e5919","DOI":"10.1002\/cpe.5919","article-title":"Deep and reinforcement learning for automated task scheduling in large-scale cloud computing systems","volume":"33","author":"Rjoub","year":"2021","journal-title":"Concurr. Comput. Pract. Exp."},{"key":"ref_116","doi-asserted-by":"crossref","first-page":"102406","DOI":"10.1016\/j.rcim.2022.102406","article-title":"Multi-agent deep reinforcement learning based Predictive Maintenance on parallel machines","volume":"78","author":"Kubler","year":"2022","journal-title":"Robot. Comput. Integr. Manuf."},{"key":"ref_117","doi-asserted-by":"crossref","first-page":"1600","DOI":"10.1109\/TII.2022.3189725","article-title":"Flexible Job-Shop Scheduling via Graph Neural Network and Deep Reinforcement Learning","volume":"19","author":"Song","year":"2023","journal-title":"IEEE Trans. Ind. Inform."},{"key":"ref_118","doi-asserted-by":"crossref","unstructured":"Tan, L., Hai, X., Ma, K., Fan, D., Qiu, H., and Feng, Q. (2023, January 16\u201319). Digital Twin-Enabled Decision-Making Framework for Multi-UAV Mission Planning: A Multiagent Deep Reinforcement Learning Perspective. Proceedings of the IECON 2023\u201449th Annual Conference of the IEEE Industrial Electronics Society, Singapore.","DOI":"10.1109\/IECON51785.2023.10312492"},{"key":"ref_119","doi-asserted-by":"crossref","unstructured":"Waschneck, B., Reichstaller, A., Belzner, L., Altenmuller, T., Bauernhansl, T., Knapp, A., and Kyek, A. (May, January 30). Deep reinforcement learning for semiconductor production scheduling. Proceedings of the 2018 29th Annual SEMI Advanced Semiconductor Manufacturing Conference, ASMC 2018, Saratoga Springs, NY, USA.","DOI":"10.1109\/ASMC.2018.8373191"},{"key":"ref_120","first-page":"805","article-title":"A dynamic scheduling method with Conv-Dueling and generalized representation based on reinforcement learning","volume":"14","author":"Xia","year":"2023","journal-title":"Int. J. Ind. Eng. Comput."},{"key":"ref_121","doi-asserted-by":"crossref","first-page":"8242","DOI":"10.1109\/TII.2022.3217477","article-title":"Workflow Scheduling in Serverless Edge Computing for the Industrial Internet of Things: A Learning Approach","volume":"19","author":"Xie","year":"2022","journal-title":"IEEE Trans. Ind. Inform."},{"key":"ref_122","doi-asserted-by":"crossref","unstructured":"Xu, Y., and Zhao, J. (2022, January 22\u201324). Actor-Critic with Transformer for Cloud Computing Resource Three Stage Job Scheduling. Proceedings of the 2022 7th International Conference on Cloud Computing and Big Data Analytics, ICCCBDA 2022, Chengdu, China.","DOI":"10.1109\/ICCCBDA55098.2022.9778883"},{"key":"ref_123","first-page":"1142","article-title":"Reinforcement Learning-Based Mobile Edge Computing and Transmission Scheduling for Video Surveillance","volume":"10","author":"Yan","year":"2022","journal-title":"IEEE Trans. Emerg. Top. Comput."},{"key":"ref_124","doi-asserted-by":"crossref","unstructured":"Wang, S., Li, J., and Luo, Y. (2021, January 17\u201319). Smart Scheduling for Flexible and Hybrid Production with Multi-Agent Deep Reinforcement Learning. Proceedings of the 2021 IEEE 2nd International Conference on Information Technology, Big Data and Artificial Intelligence, ICIBA 2021, Chongqing, China.","DOI":"10.1109\/ICIBA52610.2021.9688235"},{"key":"ref_125","doi-asserted-by":"crossref","unstructured":"Wang, Z., and Liao, W. (2023, January 18\u201321). Job Shop Scheduling Problem Using Proximal Policy Optimization. Proceedings of the 2023 IEEE International Conference on Industrial Engineering and Engineering Management, IEEM 2023, Singapore.","DOI":"10.1109\/IEEM58616.2023.10406397"},{"key":"ref_126","doi-asserted-by":"crossref","first-page":"84258","DOI":"10.1109\/ACCESS.2020.2992127","article-title":"A novel on-demand charging strategy based on swarm reinforcement learning in WRSNs","volume":"8","author":"Wei","year":"2020","journal-title":"IEEE Access"},{"key":"ref_127","doi-asserted-by":"crossref","first-page":"1090","DOI":"10.1109\/TGCN.2022.3187128","article-title":"Delay-Aware Edge-Terminal Collaboration in Green Internet of Vehicles: A Multiagent Soft Actor-Critic Approach","volume":"7","author":"Wu","year":"2023","journal-title":"IEEE Trans. Green Commun. Netw."},{"key":"ref_128","first-page":"6889","article-title":"MobiCharger: Optimal Scheduling for Cooperative EV-to-EV Dynamic Wireless Charging","volume":"22","author":"Yan","year":"2023","journal-title":"IEEE Trans. Mob. Comput."},{"key":"ref_129","doi-asserted-by":"crossref","unstructured":"Zisgen, H., Miltenberger, R., Hochhaus, M., and St\u00f6hr, N. (2023, January 10\u201313). Dynamic Scheduling of Gantry Robots using Simulation and Reinforcement Learning. Proceedings of the 2023 Winter Simulation Conference (WSC), San Antonio, TX, USA.","DOI":"10.1109\/WSC60868.2023.10407159"},{"key":"ref_130","doi-asserted-by":"crossref","first-page":"109802","DOI":"10.1016\/j.cie.2023.109802","article-title":"The application of heterogeneous graph neural network and deep reinforcement learning in hybrid flow shop scheduling problem","volume":"187","author":"Zhao","year":"2024","journal-title":"Comput. Ind. Eng."},{"key":"ref_131","doi-asserted-by":"crossref","first-page":"122","DOI":"10.1016\/j.jmsy.2023.11.010","article-title":"Multirobot collaborative task dynamic scheduling based on multiagent reinforcement learning with heuristic graph convolution considering robot service performance","volume":"72","author":"Zhou","year":"2024","journal-title":"J. Manuf. Syst."},{"key":"ref_132","unstructured":"Felder, M., Steiner, D., Busch, P., Trat, M., Sun, C., Bender, J., and Ovtcharova, J. (March, January 28). Energy-Flexible Job-Shop Scheduling Using Deep Reinforcement Learning. Proceedings of the Conference on Production Systems and Logistics, Santiago de Quer\u00e9taro, Mexico."},{"key":"ref_133","doi-asserted-by":"crossref","unstructured":"Lara-C\u00e1rdenas, E., Silva-G\u00e1lves, A., Ortiz-Bayliss, J.C., Amaya, I., Cruz-Duarte, J.M., and Terashima-Mar\u00edn, H. (2020, January 1\u20134). Exploring Reward-based Hyper-heuristics for the Job-shop Scheduling Problem. Proceedings of the 2020 IEEE Symposium Series on Computational Intelligence (SSCI), Canberra, Australia.","DOI":"10.1109\/SSCI47803.2020.9308131"},{"key":"ref_134","unstructured":"Qu, S., Jie, W., and Shivani, G. (2016, January 6\u20139). Learning adaptive dispatching rules for a manufacturing process system by using reinforcement learning approach. Proceedings of the IEEE International Conference on Emerging Technologies and Factory Automation, ETFA, Berlin, Germany."},{"key":"ref_135","doi-asserted-by":"crossref","first-page":"346","DOI":"10.1109\/JIOT.2021.3098917","article-title":"Profit Maximizing Smart Manufacturing over AI-Enabled Configurable Blockchains","volume":"9","author":"Teng","year":"2022","journal-title":"IEEE Internet Things J."},{"key":"ref_136","first-page":"100471","article-title":"Logistics-involved task scheduling in cloud manufacturing with offline deep reinforcement learning","volume":"34","author":"Wang","year":"2023","journal-title":"J. Ind. Inf. Integr."},{"key":"ref_137","doi-asserted-by":"crossref","unstructured":"Klein, N., and Prunte, J. (2022, January 7\u201310). A New Deep Reinforcement Learning Algorithm for the Online Stochastic Profitable Tour Problem. Proceedings of the 2022 IEEE International Conference on Industrial Engineering and Engineering Management (IEEM), Kuala Lumpur, Malaysia.","DOI":"10.1109\/IEEM55944.2022.9989933"},{"key":"ref_138","doi-asserted-by":"crossref","first-page":"111","DOI":"10.1109\/MNET.2019.1800254","article-title":"Intelligent Edge Computing for IoT-Based Energy Management in Smart Cities","volume":"33","author":"Liu","year":"2019","journal-title":"IEEE Netw."},{"key":"ref_139","doi-asserted-by":"crossref","first-page":"11388","DOI":"10.1109\/ACCESS.2024.3355269","article-title":"Reinforcement Learning for Two-Stage Permutation Flow Shop Scheduling\u2014A Real-World Application in Household Appliance Production","volume":"12","author":"Muller","year":"2024","journal-title":"IEEE Access"},{"key":"ref_140","doi-asserted-by":"crossref","first-page":"10004","DOI":"10.1109\/TII.2022.3230691","article-title":"Deep Reinforcement Learning-Based Rescue Resource Distribution Scheduling of Storm Surge Inundation Emergency Logistics","volume":"19","author":"Wang","year":"2023","journal-title":"IEEE Trans. Ind. Inform."},{"key":"ref_141","doi-asserted-by":"crossref","first-page":"469","DOI":"10.1109\/TII.2022.3158651","article-title":"Distributed Multiagent Deep Reinforcement Learning for Multiline Dynamic Bus Timetable Optimization","volume":"19","author":"Yan","year":"2023","journal-title":"IEEE Trans. Ind. Inform."},{"key":"ref_142","doi-asserted-by":"crossref","unstructured":"Zhou, Y., Li, X., Luo, J., Yuan, M., Zeng, J., and Yao, J. (2022, January 6\u20139). Learning to Optimize DAG Scheduling in Heterogeneous Environment. Proceedings of the IEEE International Conference on Mobile Data Management, Paphos, Cyprus.","DOI":"10.1109\/MDM55031.2022.00040"},{"key":"ref_143","doi-asserted-by":"crossref","unstructured":"Chen, Q., Zheng, Z., Hu, C., Wang, D., and Liu, F. (2019, January 7\u201310). Data-driven task allocation for multi-task transfer learning on the edge. Proceedings of the International Conference on Distributed Computing Systems, Dallas, TX, USA.","DOI":"10.1109\/ICDCS.2019.00107"},{"key":"ref_144","doi-asserted-by":"crossref","first-page":"2048","DOI":"10.1109\/ACCESS.2022.3233875","article-title":"A Seed Scheduling Method with a Reinforcement Learning for a Coverage Guided Fuzzing","volume":"11","author":"Choi","year":"2023","journal-title":"IEEE Access"},{"key":"ref_145","doi-asserted-by":"crossref","first-page":"132","DOI":"10.1016\/j.procir.2019.03.142","article-title":"Collaborative optimization of service scheduling for industrial cloud robotics based on knowledge sharing","volume":"83","author":"Du","year":"2019","journal-title":"Procedia CIRP"},{"key":"ref_146","doi-asserted-by":"crossref","first-page":"68","DOI":"10.1080\/08839514.2018.1525852","article-title":"Approximate Q-Learning for Stacking Problems with Continuous Production and Retrieval","volume":"33","author":"Fechter","year":"2019","journal-title":"Appl. Artif. Intell."},{"key":"ref_147","doi-asserted-by":"crossref","unstructured":"Fu, F., Kang, Y., Zhang, Z., and Yu, F.R. (2020, January 6\u20139). Transcoding for live streaming-based on vehicular fog computing: An actor-critic DRL approach. Proceedings of the IEEE INFOCOM 2020\u2014IEEE Conference on Computer Communications Workshops (INFOCOM WKSHPS), Toronto, ON, Canada.","DOI":"10.1109\/INFOCOMWKSHPS50562.2020.9163056"},{"key":"ref_148","doi-asserted-by":"crossref","unstructured":"Iwamura, K., Mayumi, N., Tanimizu, Y., and Sugimura, N. (September, January 31). A study on real-time scheduling for holonic manufacturing systems\u2014Determination of utility values based on multi-agent reinforcement learning. Proceedings of the 4th International Conference on Industrial Applications of Holonic and Multi-Agent Systems, HoloMAS 2009, Linz, Austria.","DOI":"10.1007\/978-3-642-03668-2_13"},{"key":"ref_149","doi-asserted-by":"crossref","first-page":"970","DOI":"10.1109\/TCCN.2020.2992628","article-title":"Deep Reinforcement Learning-Based Spectrum Allocation in Integrated Access and Backhaul Networks","volume":"6","author":"Lei","year":"2020","journal-title":"IEEE Trans. Cogn. Commun. Netw."},{"key":"ref_150","doi-asserted-by":"crossref","unstructured":"Li, X., Luo, W., Yuan, M., Wang, J., Lu, J., Wang, J., Lu, J., and Zeng, J. (2021, January 19\u201322). Learning to optimize industry-scale dynamic pickup and delivery problems. Proceedings of the International Conference on Data Engineering, Chania, Greece.","DOI":"10.1109\/ICDE51399.2021.00283"},{"key":"ref_151","doi-asserted-by":"crossref","first-page":"1269","DOI":"10.1080\/0951192X.2022.2025623","article-title":"Reinforcement learning based optimal decision making towards product lifecycle sustainability","volume":"35","author":"Liu","year":"2022","journal-title":"Int. J. Comput. Integr. Manuf."},{"key":"ref_152","doi-asserted-by":"crossref","unstructured":"Ma, S., Ruan, J., Du, Y., Bucknall, R., and Liu, Y. (2024). An End-to-End Deep Reinforcement Learning Based Modular Task Allocation Framework for Autonomous Mobile Systems. IEEE Trans. Autom. Sci. Eng., 1\u201315.","DOI":"10.1109\/TASE.2024.3367237"},{"key":"ref_153","doi-asserted-by":"crossref","unstructured":"Melnik, M., Dolgov, I., and Nasonov, D. (2020, January 2\u20134). Hybrid intellectual scheme for scheduling of heterogeneous workflows based on evolutionary approach and reinforcement learning. Proceedings of the IJCCI 2020\u201412th International Joint Conference on Computational Intelligence, Budapest, Hungary.","DOI":"10.5220\/0010112802000211"},{"key":"ref_154","doi-asserted-by":"crossref","unstructured":"Muller-Zhang, Z., and Kuhn, T. (2022, January 6\u20139). A Digital Twin-based Approach Performing Integrated Process Planning and Scheduling for Service-based Production. Proceedings of the IEEE International Conference on Emerging Technologies and Factory Automation, ETFA, Stuttgart, Germany.","DOI":"10.1109\/ETFA52439.2022.9921643"},{"key":"ref_155","doi-asserted-by":"crossref","first-page":"163325","DOI":"10.1109\/ACCESS.2021.3131799","article-title":"A Hybrid Reinforcement Learning-Based Model for the Vehicle Routing Problem in Transportation Logistics","volume":"9","author":"Phiboonbanakit","year":"2021","journal-title":"IEEE Access"},{"key":"ref_156","doi-asserted-by":"crossref","unstructured":"Song, G., Xia, M., and Zhang, D. (2023). Deep Reinforcement Learning for Risk and Disaster Management in Energy-Efficient Marine Ranching. Energies, 16.","DOI":"10.3390\/en16166092"},{"key":"ref_157","doi-asserted-by":"crossref","unstructured":"Szwarcfiter, C., Herer, Y.T., and Shtub, A. (2023). Balancing Project Schedule, Cost, and Value under Uncertainty: A Reinforcement Learning Approach. Algorithms, 16.","DOI":"10.3390\/a16080395"},{"key":"ref_158","doi-asserted-by":"crossref","unstructured":"Troch, A., Mannens, E., and Mercelis, S. (2023, January 7\u20139). Solving the Storage Location Assignment Problem Using Reinforcement Learning. Proceedings of the 2023 the 8th International Conference on Mathematics and Artificial Intelligence, Chongqing, China.","DOI":"10.1145\/3594300.3594314"},{"key":"ref_159","doi-asserted-by":"crossref","first-page":"167944","DOI":"10.1109\/ACCESS.2019.2953498","article-title":"Reinforcement learning for service function chain reconfiguration in NFV-SDN metro-core optical networks","volume":"7","author":"Troia","year":"2019","journal-title":"IEEE Access"},{"key":"ref_160","doi-asserted-by":"crossref","unstructured":"Zhang, J., Lv, Y., Li, Y., and Liu, J. (2022, January 9\u201311). An Improved QMIX-Based AGV Scheduling Approach for Material Handling Towards Intelligent Manufacturing. Proceedings of the 2022 IEEE 20th International Conference on Embedded and Ubiquitous Computing, EUC 2022, Wuhan, China.","DOI":"10.1109\/EUC57774.2022.00018"},{"key":"ref_161","doi-asserted-by":"crossref","first-page":"103415","DOI":"10.1016\/j.tre.2024.103415","article-title":"Efficient inventory routing for Bike-Sharing Systems: A combinatorial reinforcement learning framework","volume":"182","author":"Guo","year":"2024","journal-title":"Transp. Res. Part E Logist. Transp. Rev."},{"key":"ref_162","doi-asserted-by":"crossref","first-page":"107644","DOI":"10.1016\/j.asoc.2021.107644","article-title":"Production scheduling in industrial mining complexes with incoming new information using tree search and deep reinforcement learning","volume":"110","author":"Kumar","year":"2021","journal-title":"Appl. Soft Comput."},{"key":"ref_163","doi-asserted-by":"crossref","first-page":"5945","DOI":"10.1080\/00207543.2024.2304021","article-title":"A graph-based approach for integrating massive data in container terminals with application to scheduling problem","volume":"62","author":"Liu","year":"2024","journal-title":"Int. J. Prod. Res."},{"key":"ref_164","doi-asserted-by":"crossref","unstructured":"Lu, Y., Fang, S., Niu, T., Chen, G., and Liao, R. (2023, January 21\u201325). Battery Swapping Strategy for Electric Transfer-Vehicles in Seaport: A Deep Q-Network Approach. Proceedings of the 2023 IEEE\/IAS 59th Industrial and Commercial Power Systems Technical Conference (I&CPS), Las Vegas, NV, USA.","DOI":"10.1109\/ICPS57144.2023.10142090"},{"key":"ref_165","doi-asserted-by":"crossref","first-page":"114399","DOI":"10.1016\/j.eswa.2020.114399","article-title":"A reinforcement learning-based algorithm for the aircraft maintenance routing problem","volume":"169","author":"Ruan","year":"2021","journal-title":"Expert Syst. Appl."},{"key":"ref_166","doi-asserted-by":"crossref","first-page":"1426","DOI":"10.1109\/JIOT.2022.3206491","article-title":"Long-Term Matching Optimization With Federated Neural Temporal Difference Learning in Mobility-on-Demand Systems","volume":"10","author":"Sun","year":"2023","journal-title":"IEEE Internet Things J."},{"key":"ref_167","doi-asserted-by":"crossref","first-page":"925","DOI":"10.1109\/TMC.2022.3213125","article-title":"Towards Accessible Shared Autonomous Electric Mobility with Dynamic Deadlines","volume":"23","author":"Wang","year":"2024","journal-title":"IEEE Trans. Mob. Comput."},{"key":"ref_168","doi-asserted-by":"crossref","first-page":"3875","DOI":"10.1007\/s10845-023-02208-y","article-title":"Deep reinforcement learning for dynamic scheduling of energy-efficient automated guided vehicles","volume":"35","author":"Zhang","year":"2024","journal-title":"J. Intell. Manuf."},{"key":"ref_169","doi-asserted-by":"crossref","first-page":"492","DOI":"10.1016\/j.jmsy.2023.12.008","article-title":"Automated guided vehicle dispatching and routing integration via digital twin with deep reinforcement learning","volume":"72","author":"Zhang","year":"2024","journal-title":"J. Manuf. Syst."},{"key":"ref_170","doi-asserted-by":"crossref","unstructured":"Gankin, D., Mayer, S., Zinn, J., Vogel-Heuser, B., and Endisch, C. (2021, January 7\u201310). Modular Production Control with Multi-Agent Deep Q-Learning. Proceedings of the IEEE International Conference on Emerging Technologies and Factory Automation, ETFA, Vasteras, Sweden.","DOI":"10.1109\/ETFA45728.2021.9613177"},{"key":"ref_171","doi-asserted-by":"crossref","unstructured":"St\u00f6ckermann, P., Immordino, A., Altenm\u00fcller, T., and Seidel, G. (2023, January 10\u201313). Dispatching in Real Frontend Fabs With Industrial Grade Discrete-Event Simulations by Deep Reinforcement Learning with Evolution Strategies. Proceedings of the 2023 Winter Simulation Conference (WSC), San Antonio, TX, USA.","DOI":"10.1109\/WSC60868.2023.10408625"},{"key":"ref_172","doi-asserted-by":"crossref","first-page":"106294","DOI":"10.1016\/j.cor.2023.106294","article-title":"A deep multi-agent reinforcement learning approach to solve dynamic job shop scheduling problem","volume":"159","author":"Liu","year":"2023","journal-title":"Comput. Oper. Res."},{"key":"ref_173","doi-asserted-by":"crossref","unstructured":"Farag, H., Gidlund, M., and Stefanovic, C. (2021, January 12\u201316). A Deep Reinforcement Learning Approach for Improving Age of Information in Mission-Critical IoT. Proceedings of the 2021 IEEE Global Conference on Artificial Intelligence and Internet of Things, GCAIoT 2021, Dubai, United Arab Emirates.","DOI":"10.1109\/GCAIoT53516.2021.9692982"},{"key":"ref_174","doi-asserted-by":"crossref","unstructured":"Lee, S., Cho, Y., and Lee, Y.H. (2020). Injection mold production sustainable scheduling using deep reinforcement learning. Sustainability, 12.","DOI":"10.3390\/su12208718"},{"key":"ref_175","doi-asserted-by":"crossref","first-page":"2189","DOI":"10.1007\/s10489-021-02483-3","article-title":"Multi-robot exploration in task allocation problem","volume":"52","author":"Alitappeh","year":"2022","journal-title":"Appl. Intell."},{"key":"ref_176","doi-asserted-by":"crossref","unstructured":"Ao, W., Zhang, G., Li, Y., and Jin, D. (2023, January 6\u201310). Learning to Solve Grouped 2D Bin Packing Problems in the Manufacturing Industry. Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Long Beach, CA, USA.","DOI":"10.1145\/3580305.3599860"},{"key":"ref_177","doi-asserted-by":"crossref","first-page":"105439","DOI":"10.1016\/j.engappai.2022.105439","article-title":"Machine learning approach for truck-drones based last-mile delivery in the era of industry 4.0","volume":"116","author":"Arishi","year":"2022","journal-title":"Eng. Appl. Artif. Intell."},{"key":"ref_178","doi-asserted-by":"crossref","unstructured":"Fang, J., Rao, Y., Luo, Q., and Xu, J. (2023). Solving One-Dimensional Cutting Stock Problems with the Deep Reinforcement Learning. Mathematics, 11.","DOI":"10.3390\/math11041028"},{"key":"ref_179","doi-asserted-by":"crossref","first-page":"23398","DOI":"10.1007\/s10489-023-04604-6","article-title":"The 3D bin packing problem for multiple boxes and irregular items based on deep Q-network","volume":"53","author":"Liu","year":"2023","journal-title":"Appl. Intell."},{"key":"ref_180","doi-asserted-by":"crossref","first-page":"10251","DOI":"10.1016\/j.eswa.2012.02.176","article-title":"SmartGantt\u2014An intelligent system for real time rescheduling based on relational reinforcement learning","volume":"39","author":"Palombarini","year":"2012","journal-title":"Expert Syst. Appl."},{"key":"ref_181","doi-asserted-by":"crossref","first-page":"4434","DOI":"10.1080\/00207543.2021.2002963","article-title":"End-to-end on-line rescheduling from Gantt chart images using deep reinforcement learning","volume":"60","author":"Palombarini","year":"2022","journal-title":"Int. J. Prod. Res."},{"key":"ref_182","unstructured":"Saroliya, U., Arima, E., Liu, D., and Schulz, M. (November, January 31). Hierarchical Resource Partitioning on Modern GPUs: A Reinforcement Learning Approach. Proceedings of the IEEE International Conference on Cluster Computing, ICCC, Santa Fe, NM, USA."},{"key":"ref_183","doi-asserted-by":"crossref","unstructured":"Servadei, L., Zheng, J., Arjona-Medina, J., Werner, M., Esen, V., Hochreiter, S., Ecker, W., and Wille, R. (2020, January 16\u201320). Cost optimization at early stages of design using deep reinforcement learning. Proceedings of the MLCAD 2020\u20142020 ACM\/IEEE Workshop on Machine Learning for CAD, Reykjavik, Iceland.","DOI":"10.1145\/3380446.3430619"},{"key":"ref_184","doi-asserted-by":"crossref","first-page":"421","DOI":"10.1080\/00207543.2023.2188646","article-title":"Scheduling a multi-agent flow shop with two scenarios and release dates","volume":"62","author":"Wang","year":"2023","journal-title":"Int. J. Prod. Res."},{"key":"ref_185","doi-asserted-by":"crossref","first-page":"44","DOI":"10.23919\/JCC.fa.2023-0034.202308","article-title":"Intelligent preamble allocation for coexistence of mMTC\/URLLC devices: A hierarchical Q-learning based approach","volume":"20","author":"Wang","year":"2023","journal-title":"China Commun."},{"key":"ref_186","doi-asserted-by":"crossref","unstructured":"Wang, Z., Chen, Y., Liu, C., Lin, W., and Yang, L. (2023, January 16\u201319). Guided Reinforce Learning Through Spatial Residual Value for Online 3D Bin Packing. Proceedings of the IECON 2023\u201449th Annual Conference of the IEEE Industrial Electronics Society, Singapore.","DOI":"10.1109\/IECON51785.2023.10312036"},{"key":"ref_187","doi-asserted-by":"crossref","first-page":"7042","DOI":"10.1109\/JSEN.2023.3345947","article-title":"Multi-Agent Reinforcement Learning-Based Routing Protocol for Underwater Wireless Sensor Networks with Value of Information","volume":"24","author":"Wang","year":"2024","journal-title":"IEEE Sens. J."},{"key":"ref_188","doi-asserted-by":"crossref","first-page":"5057","DOI":"10.1109\/TNNLS.2021.3068828","article-title":"Learning Improvement Heuristics for Solving Routing Problems","volume":"33","author":"Wu","year":"2022","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"ref_189","doi-asserted-by":"crossref","first-page":"102028","DOI":"10.1016\/j.aei.2023.102028","article-title":"Towards reliable robot packing system based on deep reinforcement learning","volume":"57","author":"Xiong","year":"2023","journal-title":"Adv. Eng. Inform."},{"key":"ref_190","doi-asserted-by":"crossref","unstructured":"Yuan, J., Zhang, J., Cai, Z., and Yan, J. (2023, January 6\u201310). Towards Variance Reduction for Reinforcement Learning of Industrial Decision-making Tasks: A Bi-Critic based Demand-Constraint Decoupling Approach. Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Long Beach, CA, USA.","DOI":"10.1145\/3580305.3599527"},{"key":"ref_191","doi-asserted-by":"crossref","first-page":"78482","DOI":"10.1109\/ACCESS.2020.2989353","article-title":"A method for deploying distributed denial of service attack defense strategies on edge servers using reinforcement learning","volume":"8","author":"Zhang","year":"2020","journal-title":"IEEE Access"},{"key":"ref_192","doi-asserted-by":"crossref","first-page":"8427","DOI":"10.1109\/TII.2022.3218645","article-title":"A Reinforcement Learning Driven Cooperative Meta-Heuristic Algorithm for Energy-Efficient Distributed No-Wait Flow-Shop Scheduling with Sequence-Dependent Setup Time","volume":"19","author":"Zhao","year":"2022","journal-title":"IEEE Trans. Ind. Inform."},{"key":"ref_193","doi-asserted-by":"crossref","first-page":"109925","DOI":"10.1016\/j.cie.2024.109925","article-title":"An improved deep Q-learning algorithm for a trade-off between energy consumption and productivity in batch scheduling","volume":"188","author":"Zheng","year":"2024","journal-title":"Comput. Ind. Eng."},{"key":"ref_194","doi-asserted-by":"crossref","unstructured":"Zhang, J., Liu, Y., Qin, X., and Xu, X. (2021, January 13\u201316). Energy-Efficient Federated Learning Framework for Digital Twin-Enabled Industrial Internet of Things. Proceedings of the IEEE International Symposium on Personal, Indoor and Mobile Radio Communications, PIMRC, Helsinki, Finland.","DOI":"10.1109\/PIMRC50174.2021.9569716"},{"key":"ref_195","unstructured":"Yang, Z., Yang, S., Song, S., Zhang, W., Song, R., Cheng, J., and Li, Y. (October, January 27). PackerBot: Variable-Sized Product Packing with Heuristic Deep Reinforcement Learning. Proceedings of the IEEE International Conference on Intelligent Robots and Systems, Prague, Czech Republic."},{"key":"ref_196","doi-asserted-by":"crossref","first-page":"68760","DOI":"10.1109\/ACCESS.2023.3292548","article-title":"A Memetic Algorithm with Reinforcement Learning for Sociotechnical Production Scheduling","volume":"11","author":"Grumbach","year":"2022","journal-title":"IEEE Access"},{"key":"ref_197","doi-asserted-by":"crossref","first-page":"446","DOI":"10.1016\/j.ejor.2023.01.017","article-title":"A general deep reinforcement learning hyperheuristic framework for solving combinatorial optimization problems","volume":"309","author":"Kallestad","year":"2023","journal-title":"Eur. J. Oper. Res."},{"key":"ref_198","doi-asserted-by":"crossref","first-page":"102357","DOI":"10.1016\/j.rcim.2022.102357","article-title":"Probing an intelligent predictive maintenance approach with deep learning and augmented reality for machine tools in IoT-enabled manufacturing","volume":"77","author":"Liu","year":"2022","journal-title":"Robot. Comput. Integr. Manuf."},{"key":"ref_199","doi-asserted-by":"crossref","first-page":"1296","DOI":"10.1109\/TSC.2022.3157145","article-title":"Optimizing Data Center Energy Efficiency via Event-Driven Deep Reinforcement Learning","volume":"16","author":"Ran","year":"2023","journal-title":"IEEE Trans. Serv. Comput."},{"key":"ref_200","doi-asserted-by":"crossref","unstructured":"Shafiq, S., Mayr-Dorn, C., Mashkoor, A., and Egyed, A. (2020, January 8\u201311). Towards Optimal Assembly Line Order Sequencing with Reinforcement Learning: A Case Study. Proceedings of the IEEE International Conference on Emerging Technologies and Factory Automation, ETFA, Vienna, Austria.","DOI":"10.1109\/ETFA46521.2020.9211982"},{"key":"ref_201","doi-asserted-by":"crossref","first-page":"41","DOI":"10.1007\/s10723-023-09673-y","article-title":"Data Security Aware and Effective Task Offloading Strategy in Mobile Edge Computing","volume":"21","author":"Tong","year":"2023","journal-title":"J. Grid Comput."},{"key":"ref_202","doi-asserted-by":"crossref","first-page":"1342","DOI":"10.1109\/TII.2020.2993858","article-title":"Vehicle to Grid Frequency Regulation Capacity Optimal Scheduling for Battery Swapping Station Using Deep Q-Network","volume":"17","author":"Wang","year":"2021","journal-title":"IEEE Trans. Ind. Inform."},{"key":"ref_203","doi-asserted-by":"crossref","first-page":"121050","DOI":"10.1016\/j.eswa.2023.121050","article-title":"A Q-learning-based hyper-heuristic evolutionary algorithm for the distributed flexible job-shop scheduling problem with crane transportation","volume":"234","author":"Zhang","year":"2023","journal-title":"Expert Syst. Appl."},{"key":"ref_204","doi-asserted-by":"crossref","first-page":"2771","DOI":"10.1109\/TCYB.2023.3312647","article-title":"Hybrid Residual Multiexpert Reinforcement Learning for Spatial Scheduling of High-Density Parking Lots","volume":"54","author":"Hou","year":"2024","journal-title":"IEEE Trans. Cybern."},{"key":"ref_205","doi-asserted-by":"crossref","first-page":"209","DOI":"10.1109\/TCOMM.2023.3322736","article-title":"Deep Reinforcement Learning Based Resource Allocation in Delay-Tolerance-Aware 5G Industrial IoT Systems","volume":"72","author":"Wang","year":"2024","journal-title":"IEEE Trans. Commun."},{"key":"ref_206","doi-asserted-by":"crossref","unstructured":"Yeh, Y.H., Chen, S.Y.H., Chen, H.M., Tu, D.Y., Fang, G.Q., Kuo, Y.C., and Chen, P.Y. (2023, January 16\u201319). DPRoute: Deep Learning Framework for Package Routing. Proceedings of the 2023 28th Asia and South Pacific Design Automation Conference (ASP-DAC), Tokyo, Japan.","DOI":"10.1145\/3566097.3567902"},{"key":"ref_207","doi-asserted-by":"crossref","unstructured":"Perin, G., Nophut, D., Badia, L., and Fitzek, F.H. (2020, January 28\u201331). Maximizing Airtime Efficiency for Reliable Broadcast Streams in WMNs with Multi-Armed Bandits. Proceedings of the 2020 11th IEEE Annual Ubiquitous Computing, Electronics and Mobile Communication Conference, UEMCON 2020, New York, NY, USA.","DOI":"10.1109\/UEMCON51285.2020.9298050"},{"key":"ref_208","doi-asserted-by":"crossref","first-page":"106349","DOI":"10.1016\/j.cor.2023.106349","article-title":"Learning to schedule heuristics for the simultaneous stochastic optimization of mining complexes","volume":"159","author":"Yaakoubi","year":"2023","journal-title":"Comput. Oper. Res."},{"key":"ref_209","doi-asserted-by":"crossref","first-page":"4276","DOI":"10.1109\/TII.2019.2908210","article-title":"Smart Manufacturing Scheduling with Edge Computing Using Multiclass Deep Q Network","volume":"15","author":"Lin","year":"2019","journal-title":"IEEE Trans. Ind. Inform."},{"key":"ref_210","doi-asserted-by":"crossref","first-page":"101390","DOI":"10.1109\/ACCESS.2021.3097254","article-title":"Deep Reinforcement Learning for Minimizing Tardiness in Parallel Machine Scheduling with Sequence Dependent Family Setups","volume":"9","author":"Paeng","year":"2021","journal-title":"IEEE Access"},{"key":"ref_211","doi-asserted-by":"crossref","unstructured":"Yang, F., Tian, J., Feng, T., Xu, F., Qiu, C., and Zhao, C. (2021, January 7\u201311). Blockchain-Enabled Parallel Learning in Industrial Edge-Cloud Network: A Fuzzy DPoSt-PBFT Approach. Proceedings of the 2021 IEEE Globecom Workshops, GC Wkshps 2021, Madrid, Spain.","DOI":"10.1109\/GCWkshps52748.2021.9681977"},{"key":"ref_212","doi-asserted-by":"crossref","first-page":"410","DOI":"10.2507\/IJSIMM20-2-CO10","article-title":"Application of machine learning and rule scheduling in a job-shop production control system","volume":"20","author":"Zhao","year":"2021","journal-title":"Int. J. Simul. Model."},{"key":"ref_213","doi-asserted-by":"crossref","first-page":"5064","DOI":"10.1109\/TVT.2022.3224443","article-title":"Learning Aided Joint Sensor Activation and Mobile Charging Vehicle Scheduling for Energy-Efficient WRSN-Based Industrial IoT","volume":"72","author":"Chen","year":"2023","journal-title":"IEEE Trans. Veh. Technol."},{"key":"ref_214","unstructured":"Dai, B., Ren, T., Niu, J., Hu, Z., Hu, S., and Qiu, M. (October, January 30). A Distributed Computation Offloading Scheduling Framework based on Deep Reinforcement Learning. Proceedings of the 19th IEEE International Symposium on Parallel and Distributed Processing with Applications, 11th IEEE International Conference on Big Data and Cloud Computing, 14th IEEE International Conference on Social Computing and Networking and 11th IEEE Internation, New York, NY, USA."},{"key":"ref_215","doi-asserted-by":"crossref","first-page":"55","DOI":"10.1016\/j.procir.2016.11.011","article-title":"Optimized Adaptive Scheduling of a Manufacturing Process System with Multi-skill Workforce and Multiple Machine Types: An Ontology-based, Multi-agent Reinforcement Learning Approach","volume":"57","author":"Qu","year":"2016","journal-title":"Procedia CIRP"},{"key":"ref_216","doi-asserted-by":"crossref","unstructured":"Simonis, H. (2020). Leveraging Reinforcement Learning, Constraint Programming and Local Search: A Case Study in Car Manufacturing. Principles and Practice of Constraint Programming, Springer International Publishing.","DOI":"10.1007\/978-3-030-58475-7"},{"key":"ref_217","doi-asserted-by":"crossref","first-page":"7684","DOI":"10.1109\/LRA.2022.3184795","article-title":"Multi-Agent Reinforcement Learning for Real-Time Dynamic Production Scheduling in a Robot Assembly Cell","volume":"7","author":"Johnson","year":"2022","journal-title":"IEEE Robot. Autom. Lett."},{"key":"ref_218","doi-asserted-by":"crossref","unstructured":"Rudolf, T., Fl\u00f6gel, D., Sch\u00fcrmann, T., S\u00fc\u00df, S., Schwab, S., and Hohmann, S. (2023, January 1\u20134). ReACT: Reinforcement Learning for Controller Parametrization Using B-Spline Geometries. Proceedings of the 2023 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Honolulu, Oahu, HI, USA.","DOI":"10.1109\/SMC53992.2023.10394648"},{"key":"ref_219","doi-asserted-by":"crossref","unstructured":"Sun, M., Wang, X., Liu, X., Wu, S., Zhou, X., and Ouyang, C.X. (2021, January 24\u201326). A Multi-agent Reinforcement Learning Routing Protocol in Mobile Robot Network. Proceedings of the 2021 4th International Conference on Information Communication and Signal Processing, ICICSP 2021, Shanghai, China.","DOI":"10.1109\/ICICSP54369.2021.9611881"},{"key":"ref_220","doi-asserted-by":"crossref","first-page":"1034","DOI":"10.1109\/TC.2024.3350243","article-title":"Edge Generation Scheduling for DAG Tasks using Deep Reinforcement Learning","volume":"73","author":"Sun","year":"2024","journal-title":"IEEE Trans. Comput."},{"key":"ref_221","doi-asserted-by":"crossref","first-page":"939","DOI":"10.1109\/TASE.2023.3235742","article-title":"Heuristics Integrated Deep Reinforcement Learning for Online 3D Bin Packing","volume":"21","author":"Yang","year":"2024","journal-title":"IEEE Trans. Autom. Sci. Eng."},{"key":"ref_222","doi-asserted-by":"crossref","first-page":"63","DOI":"10.2478\/fcds-2024-0005","article-title":"Online Three-Dimensional Bin Packing: A DRL Algorithm with the Buffer Zone","volume":"49","author":"Zhang","year":"2024","journal-title":"Found. Comput. Decis. Sci."},{"key":"ref_223","doi-asserted-by":"crossref","unstructured":"Zhou, Y., Yan, S., and Peng, M. (2019, January 11\u201312). Content placement with unknown popularity in fog radio access networks. Proceedings of the IEEE International Conference on Industrial Internet Cloud, ICII 2019, Orlando, FL, USA.","DOI":"10.1109\/ICII.2019.00068"},{"key":"ref_224","doi-asserted-by":"crossref","unstructured":"Chen, S., Jiang, C., Li, J., Xiang, J., and Xiao, W. (2021). Improved deep q-network for user-side battery energy storage charging and discharging strategy in industrial parks. Entropy, 23.","DOI":"10.3390\/e23101311"},{"key":"ref_225","doi-asserted-by":"crossref","unstructured":"Ding, L., Lin, Z., and Yan, G. (2020, January 18\u201321). Multi-agent Deep Reinforcement Learning Algorithm for Distributed Economic Dispatch in Smart Grid. Proceedings of the IECON 2020 The 46th Annual Conference of the IEEE Industrial Electronics Society, Singapore.","DOI":"10.1109\/IECON43393.2020.9255238"},{"key":"ref_226","doi-asserted-by":"crossref","first-page":"848","DOI":"10.1109\/TWC.2023.3282909","article-title":"Reinforcement Learning-Based Resource Allocation for Coverage Continuity in High Dynamic UAV Communication Networks","volume":"23","author":"Li","year":"2024","journal-title":"IEEE Trans. Wirel. Commun."},{"key":"ref_227","doi-asserted-by":"crossref","first-page":"982","DOI":"10.1049\/rpg2.12650","article-title":"Low-carbon economic dispatch of the combined heat and power-virtual power plants: A improved deep reinforcement learning-based approach","volume":"17","author":"Tan","year":"2023","journal-title":"IET Renew. Power Gener."},{"key":"ref_228","doi-asserted-by":"crossref","unstructured":"Van Den Bovenkamp, N., Giraldo, J.S., Salazar Duque, E.M., Vergara, P.P., Konstantinou, C., and Palensky, P. (2023, January 25\u201329). Optimal Energy Scheduling of Flexible Industrial Prosumers via Reinforcement Learning. Proceedings of the 2023 IEEE Belgrade PowerTech, PowerTech 2023, Belgrade, Serbia.","DOI":"10.1109\/PowerTech55446.2023.10202699"},{"key":"ref_229","doi-asserted-by":"crossref","first-page":"458","DOI":"10.1016\/j.adhoc.2011.07.015","article-title":"InRout\u2014A QoS aware route selection algorithm for industrial wireless sensor networks","volume":"10","author":"Villaverde","year":"2012","journal-title":"Ad Hoc Netw."},{"key":"ref_230","doi-asserted-by":"crossref","unstructured":"Xu, J., Zhu, K., and Wang, R. (2019, January 11\u201313). RF aerially charging scheduling for UAV Fleet: AAA Q-learning approach. Proceedings of the 2019 15th International Conference on Mobile Ad-Hoc and Sensor Networks, MSN 2019, Shenzhen, China.","DOI":"10.1109\/MSN48538.2019.00046"},{"key":"ref_231","doi-asserted-by":"crossref","unstructured":"Ludeke, R., and Heyns, P.S. (2023, January 5\u20137). Towards a Deep Reinforcement Learning based approach for real time decision making and resource allocation for Prognostics and Health Management applications. Proceedings of the 2023 IEEE International Conference on Prognostics and Health Management, ICPHM 2023, Montreal, QC, Canada.","DOI":"10.1109\/ICPHM57936.2023.10194168"},{"key":"ref_232","doi-asserted-by":"crossref","first-page":"81","DOI":"10.1109\/LRA.2022.3222996","article-title":"Planning Irregular Object Packing via Hierarchical Reinforcement Learning","volume":"8","author":"Huang","year":"2023","journal-title":"IEEE Robot. Autom. Lett."},{"key":"ref_233","doi-asserted-by":"crossref","unstructured":"Puche, A.V., and Lee, S. (2022, January 23\u201327). Online 3D Bin Packing Reinforcement Learning Solution with Buffer. Proceedings of the IEEE International Conference on Intelligent Robots and Systems, Kyoto, Japan.","DOI":"10.1109\/IROS47612.2022.9982095"},{"key":"ref_234","doi-asserted-by":"crossref","unstructured":"Wu, Y., and Yao, L. (2021, January 24\u201326). Research on the Problem of 3D Bin Packing under Incomplete Information Based on Deep Reinforcement Learning. Proceedings of the 2021 International Conference on E-Commerce and E-Management, ICECEM 2021, Dalian, China.","DOI":"10.1109\/ICECEM54757.2021.00016"},{"key":"ref_235","doi-asserted-by":"crossref","first-page":"238","DOI":"10.1109\/LCOMM.2023.3336463","article-title":"A Hybrid DRL-Based Adaptive Traffic Matching Strategy for Transmitting and Computing in MEC-Enabled IIoT","volume":"28","author":"Chen","year":"2024","journal-title":"IEEE Commun. Lett."},{"key":"ref_236","doi-asserted-by":"crossref","unstructured":"Ho, T.M., Nguyen, K.K., and Cheriet, M. (2022, January 10\u201313). Game Theoretic Reinforcement Learning Framework For Industrial Internet of Things. Proceedings of the IEEE Wireless Communications and Networking Conference, WCNC, Austin, TX, USA.","DOI":"10.1109\/WCNC51071.2022.9771864"},{"key":"ref_237","doi-asserted-by":"crossref","first-page":"491","DOI":"10.1109\/TII.2022.3177415","article-title":"Service Function Chaining in Industrial Internet of Things With Edge Intelligence: A Natural Actor-Critic Approach","volume":"19","author":"Li","year":"2023","journal-title":"IEEE Trans. Ind. Inform."},{"key":"ref_238","doi-asserted-by":"crossref","first-page":"5173","DOI":"10.1109\/JIOT.2021.3109955","article-title":"Deep-Reinforcement-Learning-Based Predictive Maintenance Model for Effective Resource Management in Industrial IoT","volume":"9","author":"Ong","year":"2022","journal-title":"IEEE Internet Things J."},{"key":"ref_239","doi-asserted-by":"crossref","first-page":"2487","DOI":"10.1109\/JSAC.2021.3087264","article-title":"Age of Information Aware VNF Scheduling in Industrial IoT Using Deep Reinforcement Learning","volume":"39","author":"Akbari","year":"2021","journal-title":"IEEE J. Sel. Areas Commun."},{"key":"ref_240","doi-asserted-by":"crossref","first-page":"488","DOI":"10.1177\/09544054231180513","article-title":"A digital twin-driven dynamic path planning approach for multiple automatic guided vehicles based on deep reinforcement learning","volume":"238","author":"Bao","year":"2024","journal-title":"Proc. Inst. Mech. Eng. Part B J. Eng. Manuf."},{"key":"ref_241","doi-asserted-by":"crossref","unstructured":"Gowri, A.S., and Shanth I Bala, P. (2019, January 29\u201330). An agent based resource provision for IoT through machine learning in Fog computing. Proceedings of the 2019 IEEE International Conference on System, Computation, Automation and Networking, ICSCAN 2019, Pondicherry, India.","DOI":"10.1109\/ICSCAN.2019.8878821"},{"key":"ref_242","doi-asserted-by":"crossref","first-page":"4801","DOI":"10.1109\/TNSM.2021.3102621","article-title":"Multi-Agent and Cooperative Deep Reinforcement Learning for Scalable Network Automation in Multi-Domain SD-EONs","volume":"18","author":"Li","year":"2021","journal-title":"IEEE Trans. Netw. Serv. Manag."},{"key":"ref_243","doi-asserted-by":"crossref","first-page":"170","DOI":"10.1007\/s12083-022-01348-x","article-title":"Deep reinforcement learning-based task scheduling and resource allocation for NOMA-MEC in Industrial Internet of Things","volume":"16","author":"Lin","year":"2023","journal-title":"Peer-to-Peer Netw. Appl."},{"key":"ref_244","doi-asserted-by":"crossref","first-page":"3426","DOI":"10.1109\/TCOMM.2020.2974738","article-title":"A Deep Reinforcement Learning-Based Transcoder Selection Framework for Blockchain-Enabled Wireless D2D Transcoding","volume":"68","author":"Liu","year":"2020","journal-title":"IEEE Trans. Commun."},{"key":"ref_245","doi-asserted-by":"crossref","first-page":"71752","DOI":"10.1109\/ACCESS.2020.2987820","article-title":"Actor-critic deep reinforcement learning for solving job shop scheduling problems","volume":"8","author":"Liu","year":"2020","journal-title":"IEEE Access"},{"key":"ref_246","doi-asserted-by":"crossref","first-page":"15707","DOI":"10.1109\/TITS.2022.3145375","article-title":"Option-Based Multi-Agent Reinforcement Learning for Painting With Multiple Large-Sized Robots","volume":"23","author":"Liu","year":"2022","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"key":"ref_247","doi-asserted-by":"crossref","first-page":"302","DOI":"10.23919\/JCC.2023.00.009","article-title":"Task offloading optimization for AGVs with fixed routes in industrial IoT environment","volume":"20","author":"Liu","year":"2023","journal-title":"China Commun."},{"key":"ref_248","doi-asserted-by":"crossref","first-page":"115473","DOI":"10.1016\/j.apenergy.2020.115473","article-title":"Multi-agent deep reinforcement learning based demand response for discrete manufacturing systems energy management","volume":"276","author":"Lu","year":"2020","journal-title":"Appl. Energy"},{"key":"ref_249","doi-asserted-by":"crossref","first-page":"2753","DOI":"10.1007\/s10586-019-03042-9","article-title":"A multi-objective trade-off framework for cloud resource scheduling based on the Deep Q-network algorithm","volume":"23","author":"Peng","year":"2020","journal-title":"Clust. Comput."},{"key":"ref_250","doi-asserted-by":"crossref","first-page":"146492","DOI":"10.1109\/ACCESS.2021.3123213","article-title":"Packet Delivery Maximization Using Deep Reinforcement Learning-Based Transmission Scheduling for Industrial Cognitive Radio Systems","volume":"9","author":"Thanh","year":"2021","journal-title":"IEEE Access"},{"key":"ref_251","doi-asserted-by":"crossref","first-page":"5208","DOI":"10.1109\/TCOMM.2022.3187146","article-title":"Multiagent Deep Reinforcement Learning for Cost- and Delay-Sensitive Virtual Network Function Placement and Routing","volume":"70","author":"Wang","year":"2022","journal-title":"IEEE Trans. Commun."},{"key":"ref_252","doi-asserted-by":"crossref","unstructured":"Budak, A.F., Bhansali, P., Liu, B., Sun, N., Pan, D.Z., and Kashyap, C.V. (2021, January 5\u20139). DNN-Opt: An RL Inspired Optimization for Analog Circuit Sizing using Deep Neural Networks. Proceedings of the Design Automation Conference, San Francisco, CA, USA.","DOI":"10.1109\/DAC18074.2021.9586139"},{"key":"ref_253","doi-asserted-by":"crossref","first-page":"825","DOI":"10.1109\/TASE.2018.2862380","article-title":"Scheduling Semiconductor Testing Facility by Using Cuckoo Search Algorithm with Reinforcement Learning and Surrogate Modeling","volume":"16","author":"Cao","year":"2019","journal-title":"IEEE Trans. Autom. Sci. Eng."},{"key":"ref_254","unstructured":"Stuckey, P.J. (2021). SeaPearl: A Constraint Programming Solver Guided by Reinforcement Learning. Integration of Constraint Programming, Artificial Intelligence, and Operations Research, Springer International Publishing."},{"key":"ref_255","doi-asserted-by":"crossref","first-page":"3659","DOI":"10.1109\/TASE.2021.3129439","article-title":"Learning-Based Grey Wolf Optimizer for Stochastic Flexible Job Shop Scheduling","volume":"19","author":"Lin","year":"2022","journal-title":"IEEE Trans. Autom. Sci. Eng."},{"key":"ref_256","doi-asserted-by":"crossref","first-page":"6663","DOI":"10.1109\/TCYB.2022.3210228","article-title":"Learning-Based Cuckoo Search Algorithm to Schedule a Flexible Job Shop With Sequencing Flexibility","volume":"53","author":"Lin","year":"2023","journal-title":"IEEE Trans. Cybern."},{"key":"ref_257","doi-asserted-by":"crossref","first-page":"121723","DOI":"10.1016\/j.eswa.2023.121723","article-title":"A DQL-NSGA-III algorithm for solving the flexible job shop dynamic scheduling problem","volume":"237","author":"Tang","year":"2024","journal-title":"Expert Syst. Appl."},{"key":"ref_258","doi-asserted-by":"crossref","first-page":"541","DOI":"10.1109\/TASE.2022.3162653","article-title":"A Dynamic Scheduling Framework for Byproduct Gas System Combining Expert Knowledge and Production Plan","volume":"20","author":"Wang","year":"2023","journal-title":"IEEE Trans. Autom. Sci. Eng."},{"key":"ref_259","doi-asserted-by":"crossref","first-page":"3254","DOI":"10.1109\/TNET.2023.3276363","article-title":"Reinforcement Learning-Based Particle Swarm Optimization for End-to-End Traffic Scheduling in TSN-5G Networks","volume":"31","author":"Wang","year":"2023","journal-title":"IEEE\/ACM Trans. Netw."},{"key":"ref_260","doi-asserted-by":"crossref","first-page":"120571","DOI":"10.1016\/j.eswa.2023.120571","article-title":"A knowledge-driven cooperative scatter search algorithm with reinforcement learning for the distributed blocking flow shop scheduling problem","volume":"230","author":"Zhao","year":"2023","journal-title":"Expert Syst. Appl."},{"key":"ref_261","doi-asserted-by":"crossref","unstructured":"Ma, N., Wang, Z., Ba, Z., Li, X., Yang, N., Yang, X., and Zhang, H. (2023). Hierarchical Reinforcement Learning for Crude Oil Supply Chain Scheduling. Algorithms, 16.","DOI":"10.3390\/a16070354"},{"key":"ref_262","doi-asserted-by":"crossref","first-page":"873","DOI":"10.1080\/01605682.2018.1468862","article-title":"Simulated annealing with improved reheating and learning for the post enrolment course timetabling problem","volume":"70","author":"Goh","year":"2019","journal-title":"J. Oper. Res. Soc."},{"key":"ref_263","doi-asserted-by":"crossref","unstructured":"Fairee, S., Khompatraporn, C., Prom-on, S., and Sirinaovakul, B. (2019, January 10\u201313). Combinatorial Artificial Bee Colony Optimization with Reinforcement Learning Updating for Travelling Salesman Problem. Proceedings of the 2019 16th International Conference on Electrical Engineering\/Electronics, Computer, Telecommunications and Information Technology (ECTI-CON), Pattaya, Thailand.","DOI":"10.1109\/ECTI-CON47248.2019.8955176"},{"key":"ref_264","doi-asserted-by":"crossref","unstructured":"Durst, P., Jia, X., and Li, L. (2023, January 24\u201326). Multi-Objective Optimization of AGV Real-Time Scheduling Based on Deep Reinforcement Learning. Proceedings of the 42nd Chinese Control Conference, Tianjin, China.","DOI":"10.23919\/CCC58697.2023.10240797"},{"key":"ref_265","doi-asserted-by":"crossref","unstructured":"Wang, L., Yang, C., Wang, X., Li, J., Wang, Y., and Wang, Y. (2019, January 23\u201325). Integrated Resource Scheduling for User Experience Enhancement: A Heuristically Accelerated DRL. Proceedings of the 2019 11th International Conference on Wireless Communications and Signal Processing, WCSP 2019, Xi\u2019an, China.","DOI":"10.1109\/WCSP.2019.8927970"},{"key":"ref_266","doi-asserted-by":"crossref","first-page":"23179","DOI":"10.1007\/s10489-023-04774-3","article-title":"Integrating short-term stochastic production planning updating with mining fleet management in industrial mining complexes: An actor-critic reinforcement learning approach","volume":"53","author":"Carvalho","year":"2023","journal-title":"Appl. Intell."},{"key":"ref_267","doi-asserted-by":"crossref","first-page":"104069","DOI":"10.1016\/j.autcon.2021.104069","article-title":"Automating look-ahead schedule generation for construction using linked-data based constraint checking and reinforcement learning","volume":"134","author":"Soman","year":"2022","journal-title":"Autom. Constr."},{"key":"ref_268","doi-asserted-by":"crossref","first-page":"2639","DOI":"10.1109\/TNSM.2023.3240396","article-title":"GRL-PS: Graph embedding-based DRL approach for adaptive path selection","volume":"20","author":"Wei","year":"2023","journal-title":"IEEE Trans. Netw. Serv. Manag."},{"key":"ref_269","doi-asserted-by":"crossref","first-page":"2798","DOI":"10.1109\/TITS.2021.3099477","article-title":"Space-Air-Ground Integrated Multi-Domain Network Resource Orchestration Based on Virtual Network Architecture: A DRL Method","volume":"23","author":"Zhang","year":"2022","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"key":"ref_270","doi-asserted-by":"crossref","first-page":"110695","DOI":"10.1016\/j.asoc.2023.110695","article-title":"Q-learning-based hyper-heuristic evolutionary algorithm for the distributed assembly blocking flowshop scheduling problem","volume":"146","author":"Zhang","year":"2023","journal-title":"Appl. Soft Comput."},{"key":"ref_271","doi-asserted-by":"crossref","first-page":"1322","DOI":"10.1109\/TII.2022.3167380","article-title":"A Deep Reinforcement Learning Framework Based on an Attention Mechanism and Disjunctive Graph Embedding for the Job-Shop Scheduling Problem","volume":"19","author":"Chen","year":"2023","journal-title":"IEEE Trans. Ind. Inform."},{"key":"ref_272","first-page":"5103","article-title":"Deep Reinforcement Learning-Based Job Shop Scheduling of Smart Manufacturing","volume":"73","author":"Elsayed","year":"2022","journal-title":"Comput. Mater. Contin."},{"key":"ref_273","doi-asserted-by":"crossref","unstructured":"Farahani, A., Elzakker, M.V., Genga, L., Troubil, P., and Dijkman, R. (2023, January 4\u20138). Relational Graph Attention-Based Deep Reinforcement Learning: An Application to Flexible Job Shop Scheduling with Sequence-Dependent Setup Times. Proceedings of the 17th International Conference, LION 17, Nice, France.","DOI":"10.1007\/978-3-031-44505-7_24"},{"key":"ref_274","doi-asserted-by":"crossref","first-page":"1937","DOI":"10.1007\/s11431-022-2413-5","article-title":"Digital twin-enabled adaptive scheduling strategy based on deep reinforcement learning","volume":"66","author":"Gan","year":"2023","journal-title":"Sci. China Technol. Sci."},{"key":"ref_275","doi-asserted-by":"crossref","first-page":"121756","DOI":"10.1016\/j.eswa.2023.121756","article-title":"An end-to-end deep reinforcement learning method based on graph neural network for distributed job-shop scheduling problem","volume":"238","author":"Huang","year":"2024","journal-title":"Expert Syst. Appl."},{"key":"ref_276","doi-asserted-by":"crossref","unstructured":"Lee, J.H., and Kim, H.J. (2022, January 11\u201314). Imitation Learning for Real-Time Job Shop Scheduling Using Graph-Based Representation. Proceedings of the 2022 Winter Simulation Conference, Singapore.","DOI":"10.1109\/WSC57314.2022.10015445"},{"key":"ref_277","doi-asserted-by":"crossref","first-page":"6836","DOI":"10.1109\/TSMC.2023.3287655","article-title":"Dynamic Job-Shop Scheduling Problems Using Graph Neural Network and Deep Reinforcement Learning","volume":"53","author":"Liu","year":"2023","journal-title":"IEEE Trans. Syst. Man Cybern. Syst."},{"key":"ref_278","doi-asserted-by":"crossref","unstructured":"Zhao, X., Song, W., Li, Q., Shi, H., Kang, Z., and Zhang, C. (2022, January 4\u20137). A Deep Reinforcement Learning Approach for Resource-Constrained Project Scheduling. Proceedings of the 2022 IEEE Symposium Series on Computational Intelligence, SSCI 2022, Singapore.","DOI":"10.1109\/SSCI51031.2022.10022122"},{"key":"ref_279","doi-asserted-by":"crossref","first-page":"132053","DOI":"10.1109\/ACCESS.2021.3114967","article-title":"RECCE: Deep Reinforcement Learning for Joint Routing and Scheduling in Time-Constrained Wireless Networks","volume":"9","author":"Chilukuri","year":"2021","journal-title":"IEEE Access"},{"key":"ref_280","doi-asserted-by":"crossref","first-page":"17473","DOI":"10.1007\/s00500-023-09159-9","article-title":"Reinforcement learning-based multi-objective energy-efficient task scheduling in fog-cloud industrial IoT-based systems","volume":"27","author":"Vijayalakshmi","year":"2023","journal-title":"Soft Comput."},{"key":"ref_281","doi-asserted-by":"crossref","first-page":"11562","DOI":"10.1109\/JSEN.2024.3364683","article-title":"Q Learning-Based Routing Protocol with Accelerating Convergence for Underwater Wireless Sensor Networks","volume":"24","author":"Wang","year":"2024","journal-title":"IEEE Sens. J."},{"key":"ref_282","doi-asserted-by":"crossref","unstructured":"Yan, Z., Du, H., Zhang, J., and Li, G. (2021, January 1\u20134). Cherrypick: Solving the Steiner Tree Problem in Graphs using Deep Reinforcement Learning. Proceedings of the 16th IEEE Conference on Industrial Electronics and Applications, ICIEA 2021, Chengdu, China.","DOI":"10.1109\/ICIEA51954.2021.9516291"},{"key":"ref_283","doi-asserted-by":"crossref","first-page":"5672227","DOI":"10.1155\/2021\/5672227","article-title":"Application of Deep Reinforcement Learning Algorithm in Uncertain Logistics Transportation Scheduling","volume":"2021","author":"Yuan","year":"2021","journal-title":"Comput. Intell. Neurosci."},{"key":"ref_284","doi-asserted-by":"crossref","unstructured":"Zhong, C., Jia, H., Wan, H., and Zhao, X. (2021, January 19\u201322). DRLS: A Deep Reinforcement Learning Based Scheduler for Time-Triggered Ethernet. Proceedings of the International Conference on Computer Communications and Networks, ICCCN, Athens, Greece.","DOI":"10.1109\/ICCCN52240.2021.9522239"},{"key":"ref_285","first-page":"26","article-title":"Reinforcement Learning Guided Detailed Routing for Custom Circuits","volume":"1","author":"Chen","year":"2023","journal-title":"Proc. Int. Symp. Phys. Des."},{"key":"ref_286","doi-asserted-by":"crossref","unstructured":"Da Costa, P., Zhang, Y., Akcay, A., and Kaymak, U. (2021, January 18\u201322). Learning 2-opt Local Search from Heuristics as Expert Demonstrations. Proceedings of the International Joint Conference on Neural Networks, Shenzhen, China.","DOI":"10.1109\/IJCNN52387.2021.9533697"},{"key":"ref_287","doi-asserted-by":"crossref","unstructured":"He, X., Zhuge, X., Dang, F., Xu, W., and Yang, Z. (2023, January 17\u201320). DeepScheduler: Enabling Flow-Aware Scheduling in Time-Sensitive Networking. Proceedings of the IEEE INFOCOM 2023\u2014IEEE Conference on Computer Communications, New York, NY, USA.","DOI":"10.1109\/INFOCOM53939.2023.10228875"},{"key":"ref_288","doi-asserted-by":"crossref","first-page":"15652","DOI":"10.1109\/TITS.2023.3253552","article-title":"Neural Airport Ground Handling","volume":"24","author":"Wu","year":"2023","journal-title":"IEEE Trans. Intell. Transp. Syst."},{"key":"ref_289","doi-asserted-by":"crossref","first-page":"11578","DOI":"10.1109\/JIOT.2021.3130474","article-title":"Online Partial Offloading and Task Scheduling in SDN-Fog Networks with Deep Recurrent Reinforcement Learning","volume":"9","author":"Baek","year":"2022","journal-title":"IEEE Internet Things J."},{"key":"ref_290","doi-asserted-by":"crossref","unstructured":"Elsayed, M., and Erol-Kantarci, M. (2018, January 9\u201313). Deep Reinforcement Learning for Reducing Latency in Mission Critical Services. Proceedings of the 2018 IEEE Global Communications Conference, GLOBECOM 2018, Abu Dhabi, United Arab Emirates.","DOI":"10.1109\/GLOCOM.2018.8647289"},{"key":"ref_291","doi-asserted-by":"crossref","first-page":"43","DOI":"10.1109\/MDAT.2022.3145344","article-title":"Deep Reinforcement Learning for Optimization at Early Design Stages","volume":"40","author":"Servadei","year":"2023","journal-title":"IEEE Des. Test"},{"key":"ref_292","doi-asserted-by":"crossref","first-page":"292","DOI":"10.1109\/JSAC.2019.2959183","article-title":"Virtual Network Function Placement Optimization with Deep Reinforcement Learning","volume":"38","author":"Solozabal","year":"2020","journal-title":"IEEE J. Sel. Areas Commun."},{"key":"ref_293","doi-asserted-by":"crossref","first-page":"517","DOI":"10.1007\/s10479-022-04788-z","article-title":"An improved transformer model with multi-head attention and attention to attention for low-carbon multi-depot vehicle routing problem","volume":"339","author":"Zou","year":"2024","journal-title":"Ann. Oper. Res."},{"key":"ref_294","doi-asserted-by":"crossref","first-page":"13929","DOI":"10.1007\/s00500-020-04769-z","article-title":"An evaluation of Monte Carlo-based hyper-heuristic for interaction testing of industrial embedded software applications","volume":"24","author":"Ahmed","year":"2020","journal-title":"Soft Comput."},{"key":"ref_295","unstructured":"Li, Y., Fadda, E., Manerba, D., Tadei, R., and Terzo, O. (2020, January 6\u20139). Reinforcement Learning Algorithms for Online Single-Machine Scheduling. Proceedings of the 2020 Federated Conference on Computer Science and Information Systems, FedCSIS 2020, Sofia, Bulgaria."},{"key":"ref_296","doi-asserted-by":"crossref","first-page":"2129","DOI":"10.1109\/TII.2022.3211622","article-title":"Real-Time Virtual Machine Scheduling in Industry IoT Network: A Reinforcement Learning Method","volume":"19","author":"Ma","year":"2023","journal-title":"IEEE Trans. Ind. Inform."},{"key":"ref_297","doi-asserted-by":"crossref","first-page":"976","DOI":"10.1109\/TII.2018.2883991","article-title":"Traffic and Computation Co-Offloading with Reinforcement Learning in Fog Computing for Industrial Applications","volume":"15","author":"Wang","year":"2019","journal-title":"IEEE Trans. Ind. Inform."},{"key":"ref_298","doi-asserted-by":"crossref","first-page":"1077","DOI":"10.1109\/TGCN.2021.3061789","article-title":"Multi-energy scheduling of an industrial integrated energy system by reinforcement learning-based differential evolution","volume":"5","author":"Xu","year":"2021","journal-title":"IEEE Trans. Green Commun. Netw."},{"key":"ref_299","doi-asserted-by":"crossref","first-page":"012050","DOI":"10.1088\/1742-6596\/1873\/1\/012050","article-title":"A New Heuristic Reinforcement Learning for Container Relocation Problem","volume":"1873","author":"Jiang","year":"2021","journal-title":"J. Phys. Conf. Ser."},{"key":"ref_300","doi-asserted-by":"crossref","first-page":"117519","DOI":"10.1016\/j.apenergy.2021.117519","article-title":"Applying reinforcement learning and tree search to the unit commitment problem","volume":"302","year":"2021","journal-title":"Appl. Energy"},{"key":"ref_301","doi-asserted-by":"crossref","unstructured":"Revadekar, A., Soni, R., and Nimkar, A.V. (2020, January 1\u20133). QORAl: Q Learning based Delivery Optimization for Pharmacies. Proceedings of the 2020 11th International Conference on Computing, Communication and Networking Technologies, ICCCNT 2020, Kharagpur, India.","DOI":"10.1109\/ICCCNT49239.2020.9225589"},{"key":"ref_302","doi-asserted-by":"crossref","first-page":"855","DOI":"10.1007\/s10845-020-01612-y","article-title":"Designing an adaptive production control system using reinforcement learning","volume":"32","author":"Kuhnle","year":"2021","journal-title":"J. Intell. Manuf."},{"key":"ref_303","doi-asserted-by":"crossref","first-page":"70","DOI":"10.1016\/j.cie.2009.08.005","article-title":"Learning and adaptation of a policy for dynamic order acceptance in make-to-order manufacturing","volume":"58","author":"Arredondo","year":"2010","journal-title":"Comput. Ind. Eng."},{"key":"ref_304","doi-asserted-by":"crossref","first-page":"264","DOI":"10.1109\/MNET.011.2000644","article-title":"Customized Slicing for 6G: Enforcing Artificial Intelligence on Resource Management","volume":"35","author":"Guan","year":"2021","journal-title":"IEEE Netw."},{"key":"ref_305","doi-asserted-by":"crossref","unstructured":"Kan, H., Shuai, L., Chen, H., and Zhang, W. (2020, January 13\u201316). Automated Guided Logistics Handling Vehicle Path Routing under Multi-Task Scenarios. Proceedings of the 2020 IEEE International Conference on Mechatronics and Automation, ICMA 2020, Beijing, China.","DOI":"10.1109\/ICMA49215.2020.9233747"},{"key":"ref_306","doi-asserted-by":"crossref","unstructured":"Ghaleb, M., Namoura, H.A., and Taghipour, S. (2021, January 24\u201327). Reinforcement Learning-based Real-time Scheduling under Random Machine Breakdowns and Other Disturbances: A Case Study. Proceedings of the Annual Reliability and Maintainability Symposium, Orlando, FL, USA.","DOI":"10.1109\/RAMS48097.2021.9605791"},{"key":"ref_307","doi-asserted-by":"crossref","first-page":"106749","DOI":"10.1016\/j.cie.2020.106749","article-title":"Deep reinforcement learning based AGVs real-time scheduling with mixed rule for flexible shop floor in industry 4.0","volume":"149","author":"Hu","year":"2020","journal-title":"Comput. Ind. Eng."},{"key":"ref_308","doi-asserted-by":"crossref","first-page":"3020","DOI":"10.1109\/TASE.2021.3104716","article-title":"Real-Time Scheduling for Dynamic Partial-No-Wait Multiobjective Flexible Job Shop by Deep Reinforcement Learning","volume":"19","author":"Luo","year":"2022","journal-title":"IEEE Trans. Autom. Sci. Eng."},{"key":"ref_309","doi-asserted-by":"crossref","unstructured":"Wu, J., Zhang, G., Nie, J., Peng, Y., and Zhang, Y. (2021). Deep Reinforcement Learning for Scheduling in an Edge Computing-Based Industrial Internet of Things. Wireless Communications and Mobile Computing, John Wiley and Sons Ltd.","DOI":"10.1155\/2021\/8017334"},{"key":"ref_310","unstructured":"Song, Q., Lei, S., Sun, W., and Zhang, Y. (April, January 29). Adaptive federated learning for digital twin driven industrial internet of things. Proceedings of the IEEE Wireless Communications and Networking Conference, WCNC, Nanjing, China."},{"key":"ref_311","doi-asserted-by":"crossref","first-page":"110658","DOI":"10.1016\/j.asoc.2023.110658","article-title":"A Reinforcement Learning-Artificial Bee Colony algorithm for Flexible Job-shop Scheduling Problem with Lot Streaming","volume":"146","author":"Li","year":"2023","journal-title":"Appl. Soft Comput."},{"key":"ref_312","doi-asserted-by":"crossref","first-page":"1529","DOI":"10.1016\/j.enconman.2005.08.012","article-title":"Application of Q-learning with temperature variation for bidding strategies in market based power systems","volume":"47","year":"2006","journal-title":"Energy Convers. Manag."},{"key":"ref_313","doi-asserted-by":"crossref","first-page":"5617","DOI":"10.1109\/TII.2019.2941771","article-title":"Latency and Lifetime Enhancements in Industrial Wireless Sensor Networks: A Q-Learning Approach for Graph Routing","volume":"16","author":"Kunzel","year":"2020","journal-title":"IEEE Trans. Ind. Inform."},{"key":"ref_314","unstructured":"Lu, H., Zhang, X., and Yang, S. (May, January 26). A Learning-based Iterative Method for Solving Vehicle Routing Problems. Proceedings of the International Conference on Learning Representations, Virtual."},{"key":"ref_315","doi-asserted-by":"crossref","first-page":"81161","DOI":"10.1109\/ACCESS.2021.3084217","article-title":"RIATA: A Reinforcement Learning-Based Intelligent Routing Update Scheme for Future Generation IoT Networks","volume":"9","author":"Nain","year":"2021","journal-title":"IEEE Access"},{"key":"ref_316","doi-asserted-by":"crossref","first-page":"29102","DOI":"10.1109\/JIOT.2024.3406044","article-title":"Distributed DDPG-Based Resource Allocation for Age of Information Minimization in Mobile Wireless-Powered Internet of Things","volume":"11","author":"Zheng","year":"2024","journal-title":"IEEE Internet Things J."},{"key":"ref_317","doi-asserted-by":"crossref","unstructured":"Liu, X., Xu, J., Zheng, K., Zhang, G., Liu, J., and Shiratori, N. (IEEE Trans. Wirel. Commun., 2024). Throughput Maximization with an AoI Constraint in Energy Harvesting D2D-enabled Cellular Networks: An MSRA-TD3 Approach, IEEE Trans. Wirel. Commun., early access.","DOI":"10.1109\/TWC.2024.3509475"}],"container-title":["Applied Sciences"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/www.mdpi.com\/2076-3417\/15\/3\/1211\/pdf","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,8]],"date-time":"2025-10-08T10:35:47Z","timestamp":1759919747000},"score":1,"resource":{"primary":{"URL":"https:\/\/www.mdpi.com\/2076-3417\/15\/3\/1211"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,1,24]]},"references-count":317,"journal-issue":{"issue":"3","published-online":{"date-parts":[[2025,2]]}},"alternative-id":["app15031211"],"URL":"https:\/\/doi.org\/10.3390\/app15031211","relation":{},"ISSN":["2076-3417"],"issn-type":[{"type":"electronic","value":"2076-3417"}],"subject":[],"published":{"date-parts":[[2025,1,24]]}}}