4.7 Article

Reinforcement learning for joint pricing, lead-time and scheduling decisions in make-to-order systems

期刊

EUROPEAN JOURNAL OF OPERATIONAL RESEARCH
卷 221, 期 1, 页码 99-109

出版社

ELSEVIER
DOI: 10.1016/j.ejor.2012.03.020

关键词

Pricing; Reinforcement learning (RL); Scheduling; Q-learning; Simulation-based optimization; Semi-Markov Decision Problem (SMDP)

向作者/读者索取更多资源

The paper investigates a problem faced by a make-to-order (MTO) firm that has the ability to reject or accept orders, and set prices and lead-times to influence demands. Inventory holding costs for early completed orders, tardiness costs for late delivery orders, order rejection costs, manufacturing variable costs, and fixed costs are considered. In order to maximize the expected profits in an infinite planning horizon with stochastic demands, the firm needs to make decisions from the following aspects: which orders to accept or reject, the trade-off between price and lead-time, and the potential for increased demand against capacity constraints. We model the problem as a Semi-Markov Decision Problem (SMDP) and develop a reinforcement learning (RL) based Q-learning algorithm (QLA) for the problem. In addition, we build a discrete-event simulation model to validate the performance of the QLA, and compare the experimental results with two benchmark policies, the First-Come-First-Serve (FCFS) policy and a threshold heuristic policy. It is shown that the QLA outperforms the existing policies. (C) 2012 Elsevier B.V. All rights reserved.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据