4.6 Article

Online Adaptive Policy Learning Algorithm for H∞ State Feedback Control of Unknown Affine Nonlinear Discrete-Time Systems

Journal

IEEE TRANSACTIONS ON CYBERNETICS
Volume 44, Issue 12, Pages 2706-2718

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TCYB.2014.2313915

Keywords

Adaptive dynamic programming; H-infinity control; neural networks; nonlinear discrete-time system; zero-sum game

Funding

  1. National Natural Science Foundation of China [61034005, 61104010]
  2. National High Technology Research and Development Program of China [2012AA040104]

Ask authors/readers for more resources

The problem of H-infinity state feedback control of affine nonlinear discrete-time systems with unknown dynamics is investigated in this paper. An online adaptive policy learning algorithm (APLA) based on adaptive dynamic programming (ADP) is proposed for learning in real-time the solution to the Hamilton-Jacobi-Isaacs (HJI) equation, which appears in the H-infinity control problem. In the proposed algorithm, three neural networks (NNs) are utilized to find suitable approximations of the optimal value function and the saddle point feedback control and disturbance policies. Novel weight updating laws are given to tune the critic, actor, and disturbance NNs simultaneously by using data generated in real-time along the system trajectories. Considering NN approximation errors, we provide the stability analysis of the proposed algorithm with Lyapunov approach. Moreover, the need of the system input dynamics for the proposed algorithm is relaxed by using a NN identification scheme. Finally, simulation examples show the effectiveness of the proposed algorithm.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available