4.7 Article

A reinforcement learning approach to parameter estimation in dynamic job shop scheduling

Journal

COMPUTERS & INDUSTRIAL ENGINEERING
Volume 110, Issue -, Pages 75-82

Publisher

PERGAMON-ELSEVIER SCIENCE LTD
DOI: 10.1016/j.cie.2017.05.026

Keywords

Reinforcement learning; Q-factor; Dynamic job shop scheduling; Variable neighborhood search

Ask authors/readers for more resources

In this paper, reinforcement learning (RL) with a Q-factor algorithm is used to enhance performance of the scheduling method proposed for dynamic job shop scheduling (DJSS) problem which considers random job arrivals and machine breakdowns. In fact, parameters of an optimization process at any rescheduling point are selected by continually improving policy which comes from RL. The scheduling method is based on variable neighborhood search (VNS) which is introduced to address the DJSS problem. A new approach is also introduced to calculate reward values in learning processes based on quality of selected parameters. The proposed method is compared with general Variable neighborhood search and some common dispatching rules that have been widely used in the literature for the DJSS problem. Results illustrate the high performance of the proposed method in a simulated environment. (C) 2017 Published by Elsevier Ltd.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available