4.7 Article

Evolving population method for real-time reinforcement learning

期刊

EXPERT SYSTEMS WITH APPLICATIONS
卷 229, 期 -, 页码 -

出版社

PERGAMON-ELSEVIER SCIENCE LTD
DOI: 10.1016/j.eswa.2023.120493

关键词

Reinforcement learning; Deep Q network; Monte Carlo tree search; Real-time reinforcement learning; Genetic algorithm

向作者/读者索取更多资源

Reinforcement learning is promising in machine learning, but its applicability in real-time environment is limited due to short response time, high computational complexity, and learning instability. This paper proposes a new method called Evolving Population, which improves reinforcement learning performance by optimizing hyperparameters and available actions. The method utilizes an iterative structure based on evolutionary strategy to optimize these elements, and its performance is validated in an environment with real-time properties and large branching factors.
Reinforcement learning has recently been recognized as a promising means of machine learning, but its applica-bility remains limited in real-time environment due to its short response time, high computational complexity, and instability in learning. Although researchers devised several measures in attempts to press beyond the horizon, the problems consisting of large branching factors with real-time properties still stays unconquered, demanding a new method for reinforcement learning as a whole. In this paper, we propose Evolving Population. This method improves the performance of reinforcement learning by optimizing hyperparameters and available actions. This method uses an iterative structure based on an evolutionary strategy to optimize these elements. We validate the performance of our method in an environment with real-time properties and large branching factors.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据