4.4 Article

Soft Actor-Critic for Navigation of Mobile Robots

期刊

出版社

SPRINGER
DOI: 10.1007/s10846-021-01367-5

关键词

Soft actor-critic; Deep deterministic policy gradients; Deep reinforcement learning; Navigation for mobile robots

向作者/读者索取更多资源

This paper studies two deep reinforcement learning techniques for mobile robot navigation, comparing the SAC algorithm with the DDPG algorithm. The networks use laser range findings, previous velocity, and relative position as inputs, with linear and angular velocity as outputs. The SAC algorithm demonstrates superior performance compared to DDPG in navigating mobile robots.
This paper provides a study of two deep reinforcement learning techniques for application in navigation of mobile robots, one of the techniques is the Soft Actor Critic (SAC) that is compared with the Deep Deterministic Policy Gradients (DDPG) algorithm in the same situation. In order to make a robot to arrive at a target in an environment, both networks have 10 laser range findings, the previous linear and angular velocity, and relative position and angle of the mobile robot to the target are used as the network inputs. As outputs, the networks have the linear and angular velocity of the mobile robot. The reward function created was designed in a way to only give a positive reward to the agent when it gets to the target and a negative reward when colliding with any object. The proposed architecture was applied successfully in two simulated environments, and a comparison between the two referred techniques was made using the results obtained as a basis and it was demonstrated that the SAC algorithm has a superior performance for the navigation of mobile robots than the DDPG algorithm (Code available at https://github.com/dranaju/project).

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.4
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据