4.7 Article

Multi-objective reinforcement learning framework for dynamic flexible job shop scheduling problem with uncertain events

期刊

APPLIED SOFT COMPUTING
卷 131, 期 -, 页码 -

出版社

ELSEVIER
DOI: 10.1016/j.asoc.2022.109717

关键词

scheduling problem; Real-time processing framework; Deep reinforcement learning; Local search algorithm; Dynamic multi-objective flexible job shop

资金

  1. Natural Science Foundation of China
  2. Science and Technology Department of Sichuan Province, China
  3. Luzhou Science and Technology Innovation R&D Program, China
  4. Foundation of Science and Technology on Communication Security Laboratory, China
  5. [62072319]
  6. [2022YFG0041]
  7. [2021CDLZ-11]
  8. [6142103190415]

向作者/读者索取更多资源

This research proposes a new dynamic multi-objective flexible job shop scheduling problem and designs a scheduling algorithm based on deep reinforcement learning. Experimental results demonstrate that the algorithm outperforms other methods in terms of performance improvement.
The economic benefits for manufacturing companies will be influenced by how it handles potential dynamic events and performs multi-objective real-time scheduling for existing dynamic events. Based on these, we propose a new dynamic multi-objective flexible job shop scheduling problem (DMFJSP) to simulate realistic production environment. Six dynamic events are involved in the problem including job insertion, job cancellation, job operation modification, machine addition, machine tool replacement and machine breakdown. As well as three objectives of longest job processing time (makespan), average machine utilization and average job processing delay rate with a set of constraints are also raised in the study. Then this research designs a novel dynamic multi-objective scheduling algorithm based on deep reinforcement learning. The algorithm uses two deep Q-learning networks and a real-time processing framework to process each dynamic event and generate complete scheduling scheme. In addition, an improved local search algorithm is adopted to further optimize the scheduling results and the idea of combination is used to make the scheduling rules more comprehensive. Experiments on 27 instances show the superiority and stability of our approach compared to each proposed combined rule, well-known scheduling rules and standard deep Q-learning based algorithms. Compared to the current optimal deep Q-learning method, the maximum performance improvement for our three objectives are approximately 57%, 164% and 28%.(c) 2022 Published by Elsevier B.V.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据