期刊
KNOWLEDGE-BASED SYSTEMS
卷 258, 期 -, 页码 -出版社
ELSEVIER
DOI: 10.1016/j.knosys.2022.109998
关键词
Q-learning; Reinforcement learning; Uncertainty
资金
- Science and Technology Research Project
- Jiangxi Education Department, Startup Project of Doctor Scientific Research
- Jiangxi University of Science and Technology
- [GJJ180442]
- [2022205200100595]
In this paper, a variant of Q-learning, named uncertainty quantification based Q-learning, is proposed by introducing the hedonistic expected value (HEV) to increase the probability of outputting an optimal partial order in online reinforcement learning. The weights assigned by HEV to the successors are compatible with the existing operators, and the prediction of the return is not only the sum over the weights succeeding the operator but also over the weights following HEV through re-weighting. The proposed algorithm with HEV demonstrates favorable performance in practice.
In online reinforcement learning, operators predict the return by weighting the successors' estimated value. However, due to the lack of uncertainty quantification, weights assigned by operators are affected by the potentially biased estimations. As a result, the partial order of estimated values is ineffective. To increase the probability of outputting an optimal partial order, this paper introduces the hedonistic expected value (HEV), an upper bound of the return's expectation to quantify the uncer-tainty. Notably, for compatibility reasons, some complex operators are rewritten as the weighted-sum forms. Based on the weighted-sum form of the operator, the variant Q-learning, namely uncertainty quantification based Q-learning is proposed in this paper. In the proposed algorithm, the weights assigned by HEV of the successors are compatible with the existing operators. The prediction of the return is not only the sum over the weights succeeding the operator but also over the weights following HEV through re-weighting. The greediness of the re-weighted operator is unchanged, and the contraction mapping indicates the convergence can be maintained. We demonstrate that the proposed algorithm with HEV performs favorably in practice. (c) 2022 Elsevier B.V. All rights reserved.
作者
我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。
推荐
暂无数据