4.6 Article

Sample Efficient Learning of Path Following and Obstacle Avoidance Behavior for Quadrotors

期刊

IEEE ROBOTICS AND AUTOMATION LETTERS
卷 3, 期 4, 页码 3852-3859

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/LRA.2018.2856922

关键词

Collision avoidance; deep learning in robotics and automation

类别

资金

  1. Swiss National Science Foundation [UFO 200021L_153644]
  2. NWO Domain Applied Sciences
  3. Swiss National Science Foundation (SNF) [200021L_153644] Funding Source: Swiss National Science Foundation (SNF)

向作者/读者索取更多资源

In this letter, we propose an algorithm for the training of neural network control policies for quadrotors. The learned control policy computes control commands directly from sensor inputs and is, hence, computationally efficient. An imitation learning algorithm produces a policy that reproduces the behavior of a supervisor. The supervisor provides demonstrations of path following and collision avoidance maneuvers. Due to the generalization ability of neural networks, the resulting policy performs local collision avoidance, while following a global reference path. The algorithm uses a time-free model-predictive path-following controller as a supervisor. The controller generates demonstrations by following few example paths. This enables an easy-to-implement learning algorithm that is robust to errors of the model used in the model-predictive controller. The policy is trained on the real quadrotor, which requires collision-free exploration around the example path. An adapted version of the supervisor is used to enable exploration. Thus, the policy can be trained from a relatively small number of examples on the real quadrotor, making the training sample efficient.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据