4.5 Article

An enhanced deep deterministic policy gradient algorithm for intelligent control of robotic arms

期刊

FRONTIERS IN NEUROINFORMATICS
卷 17, 期 -, 页码 -

出版社

FRONTIERS MEDIA SA
DOI: 10.3389/fninf.2023.1096053

关键词

robotic arm; intelligent control; reward function; experience replay mechanism; deep deterministic policy gradient algorithm; artificial intelligence; machine learning

向作者/读者索取更多资源

To address the issues of poor robustness and adaptability in traditional control methods, the deep deterministic policy gradient (DDPG) algorithm is improved by incorporating a hybrid function with multiple rewards. The experience replay mechanism of DDPG is also enhanced through a combination of priority sampling and uniform sampling, resulting in accelerated convergence. Experimental results in a simulation environment demonstrate that the improved DDPG algorithm achieves accurate control of robot arm motion, with a higher success rate of 91.27% compared to the original DDPG algorithm, thereby exhibiting improved environmental adaptability.
Aiming at the poor robustness and adaptability of traditional control methods for different situations, the deep deterministic policy gradient (DDPG) algorithm is improved by designing a hybrid function that includes different rewards superimposed on each other. In addition, the experience replay mechanism of DDPG is also improved by combining priority sampling and uniform sampling to accelerate the DDPG's convergence. Finally, it is verified in the simulation environment that the improved DDPG algorithm can achieve accurate control of the robot arm motion. The experimental results show that the improved DDPG algorithm can converge in a shorter time, and the average success rate in the robotic arm end-reaching task is as high as 91.27%. Compared with the original DDPG algorithm, it has more robust environmental adaptability.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.5
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据