4.6 Article

Resilient adaptive optimal control of distributed multi-agent systems using reinforcement learning

期刊

IET CONTROL THEORY AND APPLICATIONS
卷 12, 期 16, 页码 2165-2174

出版社

WILEY
DOI: 10.1049/iet-cta.2018.0029

关键词

game theory; Riccati equations; adaptive control; learning (artificial intelligence); multi-agent systems; distributed control; H control; resilient adaptive optimal control; distributed multiagent systems; model-free reinforcement learning; control protocol; leader-follower multiagent systems; optimal control protocols; adversarial input; adverse effects; learning outcome; intact agents; unified RL-based; control frameworks; corrupted sensory data; actual sensory information; leader state; distributed observer; compromised agent; off-policy RL algorithm; control problem; H control problem

向作者/读者索取更多资源

This study presents a unified resilient model-free reinforcement learning (RL) based distributed control protocol for leader-follower multi-agent systems. Although RL has been successfully used to learn optimal control protocols for multi-agent systems, the effects of adversarial inputs are ignored. It is shown in this study, however, that their adverse effects can propagate across the network and impact the learning outcome of other intact agents. To alleviate this problem, a unified RL-based distributed control frameworks is developed for both homogeneous and heterogeneous multi-agent systems to prevent corrupted sensory data from propagating across the network. To this end, only the leader communicates its actual sensory information and other agents estimate the leader' state using a distributed observer and communicate this estimation to their neighbours to achieve consensus on the leader state. The observer cannot be physically affected by any adversarial input. To further improve resiliency, distributed H8 control protocols are designed to attenuate the effect of the adversarial inputs on the compromised agent itself. An off-policy RL algorithm is developed to learn the solutions of the game algebraic Riccati equations arising from solving the H8 control problem. No knowledge of the agent's dynamics is required and it is shown that the proposed RL-based H8 control protocol is resilient against adversarial inputs.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据