3.8 Proceedings Paper

Where to Look Next: Learning Viewpoint Recommendations for Informative Trajectory Planning

出版社

IEEE
DOI: 10.1109/ICRA46639.2022.9812190

关键词

-

资金

  1. National Police of the Netherlands
  2. Dutch Science Foundation NWO-TTW within the Veni project HARMONIA [18165]

向作者/读者索取更多资源

This paper presents an information-aware policy trained through deep reinforcement learning to guide a trajectory optimization planner for information gathering in search missions. In simulation tests, the method performs well and outperforms existing methods in multiple aspects.
Search missions require motion planning and navigation methods for information gathering that continuously replan based on new observations of the robot's surroundings. Current methods for information gathering, such as Monte Carlo Tree Search, are capable of reasoning over long horizons, but they are computationally expensive. An alternative for fast online execution is to train, offline, an information gathering policy, which indirectly reasons about the information value of new observations. However, these policies lack safety guarantees and do not account for the robot dynamics. To overcome these limitations we train an information-aware policy via deep reinforcement learning, that guides a receding-horizon trajectory optimization planner. In particular, the policy continuously recommends a reference viewpoint to the local planner, such that the resulting dynamically feasible and collision-free trajectories lead to observations that maximize the information gain and reduce the uncertainty about the environment. In simulation tests in previously unseen environments, our method consistently outperforms greedy next-best-view policies and achieves competitive performance compared to Monte Carlo Tree Search, in terms of information gains and coverage time, with a reduction in execution time by three orders of magnitude.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

3.8
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据