4.6 Article

Hierarchical Reinforcement Learning Framework in Geographic Coordination for Air Combat Tactical Pursuit

期刊

ENTROPY
卷 25, 期 10, 页码 -

出版社

MDPI
DOI: 10.3390/e25101409

关键词

hierarchical reinforcement learning; meta-learning; reward design; decision

向作者/读者索取更多资源

This paper proposes a hierarchical reinforcement learning framework for air combat training to address the non-convergence problem caused by the curse of dimensionality in the state space during air combat tactical pursuit. By using hierarchical reinforcement learning, three-dimensional problems are transformed into two-dimensional problems, resulting in improved training performance compared to other baselines. To further enhance overall learning performance, a meta-learning-based algorithm is established with a corresponding reward function designed to improve the agent's performance in the air combat tactical chase scenario. The results demonstrate that the proposed framework achieves better performance than the baseline approach.
This paper proposes an air combat training framework based on hierarchical reinforcement learning to address the problem of non-convergence in training due to the curse of dimensionality caused by the large state space during air combat tactical pursuit. Using hierarchical reinforcement learning, three-dimensional problems can be transformed into two-dimensional problems, improving training performance compared to other baselines. To further improve the overall learning performance, a meta-learning-based algorithm is established, and the corresponding reward function is designed to further improve the performance of the agent in the air combat tactical chase scenario. The results show that the proposed framework can achieve better performance than the baseline approach.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据