4.6 Article

Reducing the Learning Time of Reinforcement Learning for the Supervisory Control of Discrete Event Systems

期刊

IEEE ACCESS
卷 11, 期 -, 页码 59840-59853

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/ACCESS.2023.3285432

关键词

Discrete event system; linear temporal logic; supervisory control theory; reinforcement learning

向作者/读者索取更多资源

Reinforcement learning can obtain supervisory controller for discrete-event systems and address the limitations of large training data requirement and neglecting uncontrollable events by applying supervisory control theory and reducing exploration space. The proposed method achieves a nonblocking supervisor for all specifications with shorter learning time compared to existing methods.
Reinforcement learning (RL) can obtain the supervisory controller for discrete-event systems modeled by finite automata and temporal logic. The published methods often have two limitations. First, a large number of training data are required to learn the RL controller. Second, the RL algorithms do not consider uncontrollable events, which are essential for supervisory control theory (SCT). To address the limitations, we first apply SCT to find the supervisors for the specifications modeled by automata. These supervisors remove illegal training data violating these specifications and hence reduce the exploration space of the RL algorithm. For the remaining specifications modeled by temporal logic, the RL algorithm is applied to search for the optimal control decision within the confined exploration space. Uncontrollable events are considered by the RL algorithm as uncertainties in the plant model. The proposed method can obtain a nonblocking supervisor for all specifications with less learning time than the published methods.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据