4.6 Article

Reinforcement Learning-Based Complete Area Coverage Path Planning for a Modified hTrihex Robot

期刊

SENSORS
卷 21, 期 4, 页码 -

出版社

MDPI
DOI: 10.3390/s21041067

关键词

reconfigurable robot; tiling robots; reinforcement learning; complete coverage planing; energy path planning

资金

  1. National Robotics Programme under its Robotics Enabling Capabilities and Technologies [1922500051]
  2. National Robotics Programme under Robot Domain Specific [1922200058]

向作者/读者索取更多资源

Cleaning robots must achieve complete area coverage, and tiling robots offer an innovative solution to this problem. This study proposes a complete area coverage planning module for a modified honeycomb-shaped tiling robot based on deep reinforcement learning, which simultaneously generates tiling shapes and trajectories with minimum overall cost.
One of the essential attributes of a cleaning robot is to achieve complete area coverage. Current commercial indoor cleaning robots have fixed morphology and are restricted to clean only specific areas in a house. The results of maximum area coverage are sub-optimal in this case. Tiling robots are innovative solutions for such a coverage problem. These new kinds of robots can be deployed in the cases of cleaning, painting, maintenance, and inspection, which require complete area coverage. Tiling robots' objective is to cover the entire area by reconfiguring to different shapes as per the area requirements. In this context, it is vital to have a framework that enables the robot to maximize the area coverage while minimizing energy consumption. That means it is necessary for the robot to cover the maximum area with the least number of shape reconfigurations possible. The current paper proposes a complete area coverage planning module for the modified hTrihex, a honeycomb-shaped tiling robot, based on the deep reinforcement learning technique. This framework simultaneously generates the tiling shapes and the trajectory with minimum overall cost. In this regard, a convolutional neural network (CNN) with long short term memory (LSTM) layer was trained using the actor-critic experience replay (ACER) reinforcement learning algorithm. The simulation results obtained from the current implementation were compared against the results that were generated through traditional tiling theory models that included zigzag, spiral, and greedy search schemes. The model presented in the current paper was also compared against other methods where this problem was considered as a traveling salesman problem (TSP) solved through genetic algorithm (GA) and ant colony optimization (ACO) approaches. Our proposed scheme generates a path with a minimized cost at a lesser time.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据