4.7 Article

Rule-Based Reinforcement Learning for Efficient Robot Navigation With Space Reduction

Journal

IEEE-ASME TRANSACTIONS ON MECHATRONICS
Volume 27, Issue 2, Pages 846-857

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TMECH.2021.3072675

Keywords

Navigation; Robots; Trajectory; Space exploration; Mobile robots; Task analysis; Reinforcement learning; Hex-grid; robot navigation; rule-based reinforcement learning; space reduction

Funding

  1. National Natural Science Foundation of China [71732003, 62073160, 62006111]
  2. National Key Research and Development Program of China [2018AAA0101100]
  3. Synergistic Innovation Center of Jiangsu Modern Agricultural Equipment and Technology [4091600002]
  4. Australian Research Councils's Discovery Projects funding scheme [DP190101566]

Ask authors/readers for more resources

This article introduces a rule-based reinforcement learning (RuRL) algorithm for efficient navigation. By employing a wall-following rule to generate a closed-loop trajectory, a reduction rule to shrink the trajectory, and the Pledge rule to guide the exploration strategy, RuRL achieves improved navigation performance in real robot navigation experiments.
For real-world deployments, it is critical to allow robots to navigate in complex environments autonomously. Traditional methods usually maintain an internal map of the environment, and then design several simple rules, in conjunction with a localization and planning approach, to navigate through the internal map. These approaches often involve a variety of assumptions and prior knowledge. In contrast, recent reinforcement learning (RL) methods can provide a model-free, self-learning mechanism as the robot interacts with an initially unknown environment, but are expensive to deploy in real-world scenarios due to inefficient exploration. In this article, we focus on efficient navigation with the RL technique and combine the advantages of these two kinds of methods into a rule-based RL (RuRL) algorithm for reducing the sample complexity and cost of time. First, we use the rule of wall-following to generate a closed-loop trajectory. Second, we employ a reduction rule to shrink the trajectory, which in turn effectively reduces the redundant exploration space. Besides, we give the detailed theoretical guarantee that the optimal navigation path is still in the reduced space. Third, in the reduced space, we utilize the Pledge rule to guide the exploration strategy for accelerating the RL process at the early stage. Experiments conducted on real robot navigation problems in hex-grid environments demonstrate that RuRL can achieve improved navigation performance.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available