4.5 Article

Goal-aware generative adversarial imitation learning from imperfect demonstration for robotic cloth manipulation

Journal

ROBOTICS AND AUTONOMOUS SYSTEMS
Volume 158, Issue -, Pages -

Publisher

ELSEVIER
DOI: 10.1016/j.robot.2022.104264

Keywords

Generative adversarial imitation learning; Robotic cloth manipulation; Deep reinforcement learning

Funding

  1. JSPS, Japan KAKENHI [21H03522]
  2. JSPS, Japan Research Fellow Grant [20J11948]

Ask authors/readers for more resources

Generative Adversarial Imitation Learning (GAIL) is a method that can learn policies from demonstrations without explicitly defining the reward function. This paper proposes Goal-Aware Generative Adversarial Imitation Learning (GA-GAIL), which addresses the issue of imperfect demonstration data by introducing a second discriminator to distinguish the goal state. GA-GAIL extends the standard GAIL framework to robustly learn desirable policies even from imperfect demonstrations.
Generative Adversarial Imitation Learning (GAIL) can learn policies without explicitly defining the reward function from demonstrations. GAIL has the potential to learn policies with high-dimensional observations as input, e.g., images. By applying GAIL to a real robot, perhaps robot policies can be obtained for daily activities like washing, folding clothes, cooking, and cleaning. However, human demonstration data are often imperfect due to mistakes, which degrade the performance of the resulting policies. We address this issue by focusing on the following features: (1) many robotic tasks are goal-reaching tasks, and (2) labeling such goal states in demonstration data is relatively easy. With these in mind, this paper proposes Goal-Aware Generative Adversarial Imitation Learning (GA-GAIL), which trains a policy by introducing a second discriminator to distinguish the goal state in parallel with the first discriminator that indicates the demonstration data. This extends a standard GAIL framework to more robustly learn desirable policies even from imperfect demonstrations through a goal-state discriminator that promotes achieving the goal state. Furthermore, GA-GAIL employs the Entropy -maximizing Deep P-Network (EDPN) as a generator, which considers both the smoothness and causal entropy in the policy update, to achieve stable policy learning from two discriminators. Our proposed method was successfully applied to two real-robotic cloth-manipulation tasks: turning a handkerchief over and folding clothes. We confirmed that it learns cloth-manipulation policies without task-specific reward function design. Video of the real experiments are available at this URL.(c) 2022 Elsevier B.V. All rights reserved.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.5
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available