4.6 Article

An Object Recognition Grasping Approach Using Proximal Policy Optimization With YOLOv5

Journal

IEEE ACCESS
Volume 11, Issue -, Pages 87330-87343

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/ACCESS.2023.3305339

Keywords

Manipulators; Grasping; Robots; Wheels; Kinematics; Deep learning; Reinforcement learning; Deep reinforcement learning; manipulator; object grasping; proximal policy optimization; YOLOv5

Ask authors/readers for more resources

This paper proposes an object recognition grasping approach using Proximal Policy Optimization (PPO) with You Only Look Once v5 (YOLOv5) to overcome the problems of traditional grasping methods for mobile manipulators. The approach combines a vision recognition algorithm with a deep reinforcement learning algorithm to achieve object recognition grasping. Experimental results show that the proposed method outperforms the original YOLOv4 model in terms of object recognition speed and achieves higher detection precision and lower hardware requirements. The proposed method also outperforms the SAC and TRPO algorithms in object grasping, with the average reward of the PPO algorithm improved significantly compared to the other algorithms.
Aiming at the problems of traditional grasping methods for mobile manipulators, such as single application scenarios, low accuracy, and complex grasping tasks, this paper proposes an object recognition grasping approach using Proximal Policy Optimization (PPO) with You Only Look Once v5 (YOLOv5), which combines a vision recognition algorithm with a deep reinforcement learning algorithm to achieve object recognition grasping. First, YOLOv5 is adopted to identify the object and obtain the location information. Second, the PPO algorithm is used for object grasping to obtain the grasping strategy. Third, the PPO algorithm is compared with the Soft Actor-Critic (SAC) and Trust Region Policy Optimization (TRPO) algorithms in batches 16 and 128, respectively. The average reward training results of the PPO, SAC, and TRPO algorithms are obtained in our work. Experimental results show that the proposed method, in terms of object recognition speed, outperforms the original YOLOv4 model. The YOLOv5 model achieves 96% precision on our own built recognition dataset, which has higher detection precision and lower hardware requirements than the YOLOv4 model. Our proposed method outperforms SAC and TRPO algorithms in object grasping, and the average reward of the PPO algorithm is improved by 93.3% and 41% compared to SAC and TRPO algorithms, respectively. Finally, through the comparison of ablation experiments, our method has the highest accuracy and mean average precision (mAP)@0.5 value of 92.3%. We demonstrate in actual physical experiments that the grasping success rate under our proposed approach reaches 100%, providing a new research strategy for object grasping by the robot manipulator.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available