3.8 Proceedings Paper

Cost Inference of Discrete-time Linear Quadratic Control Policies using Human-Behaviour Learning

Publisher

IEEE
DOI: 10.1109/CODIT55151.2022.9804118

Keywords

-

Funding

  1. Royal Academy of Engineering
  2. Office of the Chief Science Adviser for National Security under the UK Intelligence Community Postdoctoral Research Fellowship programme

Ask authors/readers for more resources

In this paper, a cost inference algorithm for discrete-time systems using human-behaviour learning is proposed. The approach is inspired in the complementary learning that exhibits the neocortex, hippocampus, and striatum learning systems to achieve complex decision making. The main objective is to infer the hidden cost function from expert's data associated to the hippocampus (off-policy data) and transfer it to the neocortex for policy generalization (on-policy data) in different systems and environments. The neocortex is modelled by a Qlearning and a least-squares identification algorithms for onpolicy learning and system identification. The cost inference is obtained using a one-step gradient descent rule and an inverse optimal control algorithm. Convergence of the cost inference algorithm is discussed using Lyapunov recursions. Simulations verify the effectiveness of the approach.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

3.8
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available