4.6 Article

Sim2real Learning of Obstacle Avoidance for Robotic Manipulators in Uncertain Environments

Journal

IEEE ROBOTICS AND AUTOMATION LETTERS
Volume 7, Issue 1, Pages 65-72

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/LRA.2021.3116700

Keywords

Robots; Manipulators; Collision avoidance; Three-dimensional displays; Task analysis; Planning; Adaptation models; Planning under uncertainty; collision avoidance; reinforcement learning; robotic manipulator; unified representation

Categories

Funding

  1. NSFC [U2001206]
  2. Guangdong Talent Program [2019JC05X328]
  3. Guangdong Science and Technology Program [2020A0505100064]
  4. DEGP Key Project [2018KZDXM058]
  5. Shenzhen Science and Technology Program [RCJC20200714114435012, JCYJ20210324120213036]

Ask authors/readers for more resources

This paper proposes a unified representation for obstacle avoidance of robotic manipulators in unstructured environments using sim-to-real deep reinforcement learning. The unified representation is achieved through a vision-based actor-critic framework with a bounding box predictor module. The end-to-end model of the unified representation outperforms state-of-the-art techniques in sim-to-real adaptation and scene generalization.
Obstacle avoidance for robotic manipulators can be challenging when they operate in unstructured environments. This problem is probed with the sim-to-real (sim2real) deep reinforcement learning, such that a moving policy of the robotic arm is learnt in a simulator and then adapted to the real world. However, the problem of sim2real adaptation is notoriously difficult. To this end, this work proposes (1) a unified representation of obstacles and targets to capture the underlying dynamics of the environment while allowing generalization to unseen goals and (2) a flexible end-to-end model combining the unified representation with the deep reinforcement learning control module that can be trained by interacting with the environment. Such a representation is agnostic to the shape and appearance of the underlying objects, which simplifies and unifies the scene representation in both simulated and real worlds. We implement this idea with a vision-based actor-critic framework by devising a bounding box predictor module. The predictor estimates the 3D bounding boxes of obstacles and targets from the RGB-D input. The features extracted by the predictor are fed into the policy network, and all the modules are jointly trained. This makes the policy learn object-aware scene representation, which leads to a data-efficient learning of the obstacle avoidance policy. Our experiments in simulated environment and the real-world show that the end-to-end model of the unified representation achieves better sim2real adaption and scene generalization than state-of-the-art techniques.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available