4.7 Article

Adaptive fuzzy sliding-mode consensus control of nonlinear under-actuated agents in a near-optimal reinforcement learning framework

向作者/读者索取更多资源

This study presents a new framework that combines Adaptive Fuzzy Sliding-Mode Control (AF-SMC) with off-policy Reinforcement Learning (RL) algorithm to control nonlinear under-actuated agents. The framework uses graphical games to achieve near-optimal leader-follower consensus. The coefficients of the sliding variables are adaptively tuned policies to achieve an optimal compromise between tracking performance and control efforts.
This study presents a new framework for merging the Adaptive Fuzzy Sliding-Mode Control (AF-SMC) with an off-policy Reinforcement Learning (RL) algorithm to control nonlinear under-actuated agents. In particular, a near-optimal leader-follower consensus is considered, and a new method is proposed using the framework of graphical games. In the proposed technique, the sliding variables' coefficients are considered adaptively tuned policies to achieve an optimal compromise between the sat-isfactory tracking performance and the allowable control efforts. Contrary to the conventional off-policy RL algorithms for consensus control of multi-agent systems, the proposed method does not require par-tial knowledge of the system dynamics to initialize the RL process. Furthermore, an actor-critic fuzzy methodology is employed to approximate optimal policies using the measured input/output data. There-fore, using the tuned sliding vector, the control input for each agent is generated which includes a fuzzy term, a robust term, and a saturation compensating term. In particular, the fuzzy system approximates a nonlinear function, and the robust part of the input compensates for any possible mismatches. Fur-thermore, the saturation compensating gain prevents instability due to any possible actuator saturation. Based on the local sliding variables, the fuzzy singletons, the bounds of the approximation errors, and the compensating gains are adaptively tuned. Closed-loop asymptotic stability is proved using the sec-ond Lyapunov theorem and Barbalat's lemma. The method's efficacy is verified by consensus control of multiple REMUS AUVs in the vertical plane. (C) 2022 The Franklin Institute. Published by Elsevier Ltd. All rights reserved.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据