4.7 Article

Concept2Robot: Learning manipulation concepts from instructions and human demonstrations

Journal

INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH
Volume 40, Issue 12-14, Pages 1419-1434

Publisher

SAGE PUBLICATIONS LTD
DOI: 10.1177/02783649211046285

Keywords

Learning from demonstration; manipulation; human-robot interaction; natural language understanding

Categories

Funding

  1. Toyota Research Institute (TRI)

Ask authors/readers for more resources

This research aims to endow a robot with the ability to learn manipulation concepts by leveraging large-scale video datasets of humans performing manipulation actions. By using a two-stage learning process, the robot is able to generalize over a variety of tasks and environments without the need for manual task-specific rewards. The multi-task policy learns to perform a large percentage of manipulation tasks, showcasing the ability to generalize over different environments and instructions.
We aim to endow a robot with the ability to learn manipulation concepts that link natural language instructions to motor skills. Our goal is to learn a single multi-task policy that takes as input a natural language instruction and an image of the initial scene and outputs a robot motion trajectory to achieve the specified task. This policy has to generalize over different instructions and environments. Our insight is that we can approach this problem through learning from demonstration by leveraging large-scale video datasets of humans performing manipulation actions. Thereby, we avoid more time-consuming processes such as teleoperation or kinesthetic teaching. We also avoid having to manually design task-specific rewards. We propose a two-stage learning process where we first learn single-task policies through reinforcement learning. The reward is provided by scoring how well the robot visually appears to perform the task. This score is given by a video-based action classifier trained on a large-scale human activity dataset. In the second stage, we train a multi-task policy through imitation learning to imitate all the single-task policies. In extensive simulation experiments, we show that the multi-task policy learns to perform a large percentage of the 78 different manipulation tasks on which it was trained. The tasks are of greater variety and complexity than previously considered robot manipulation tasks. We show that the policy generalizes over variations of the environment. We also show examples of successful generalization over novel but similar instructions.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available