4.7 Article

A reinforcement learning approach to parameter estimation in dynamic job shop scheduling

期刊

COMPUTERS & INDUSTRIAL ENGINEERING
卷 110, 期 -, 页码 75-82

出版社

PERGAMON-ELSEVIER SCIENCE LTD
DOI: 10.1016/j.cie.2017.05.026

关键词

Reinforcement learning; Q-factor; Dynamic job shop scheduling; Variable neighborhood search

向作者/读者索取更多资源

In this paper, reinforcement learning (RL) with a Q-factor algorithm is used to enhance performance of the scheduling method proposed for dynamic job shop scheduling (DJSS) problem which considers random job arrivals and machine breakdowns. In fact, parameters of an optimization process at any rescheduling point are selected by continually improving policy which comes from RL. The scheduling method is based on variable neighborhood search (VNS) which is introduced to address the DJSS problem. A new approach is also introduced to calculate reward values in learning processes based on quality of selected parameters. The proposed method is compared with general Variable neighborhood search and some common dispatching rules that have been widely used in the literature for the DJSS problem. Results illustrate the high performance of the proposed method in a simulated environment. (C) 2017 Published by Elsevier Ltd.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据