4.6 Article

Learning Pneumatic Non-Prehensile Manipulation With a Mobile Blower

Journal

IEEE ROBOTICS AND AUTOMATION LETTERS
Volume 7, Issue 3, Pages 8471-8478

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/LRA.2022.3187833

Keywords

Deep learning methods; reinforcement learning; mobile manipulation

Categories

Funding

  1. Princeton School of Engineering
  2. National Science Foundation [IIS-1815070, DGE-1656466]

Ask authors/readers for more resources

This study investigates pneumatic non-prehensile manipulation (i.e., blowing) as an efficient method for moving scattered objects into a target receptacle. Using a multi-frequency version of the spatial action maps framework in the context of deep reinforcement learning, the research tackles the challenges of adapting to unexpected changes, maintaining fine-grained control, and inferring long-range plans. Experimental results demonstrate that blowing achieves better performance than pushing and that the proposed policies improve performance over baselines. The study also shows that the system encourages emergent specialization between low-level fine-grained control and high-level planning.
We investigate pneumatic non-prehensile manipulation (i.e., blowing) as a means of efficiently moving scattered objects into a target receptacle. Due to the chaotic nature of aerodynamic forces, a blowing controller must i) continually adapt to unexpected changes from its actions, ii) maintain fine-grained control, since the slightest misstep can result in large unintended consequences (e.g., scatter objects already in a pile), and iii) infer long-range plans (e.g., move the robot to strategic blowing locations). We tackle these challenges in the context of deep reinforcement learning, introducing a multi-frequency version of the spatial action maps framework. This allows for efficient learning of vision-based policies that effectively combine high-level planning and low-level closed-loop control for dynamic mobile manipulation. Experiments show that our system learns efficient behaviors for the task, demonstrating in particular that blowing achieves better downstream performance than pushing, and that our policies improve performance over baselines. Moreover, we show that our system naturally encourages emergent specialization between the different subpolicies spanning low-level fine-grained control and high-level planning. On a real mobile robot equipped with a miniature air blower, we show that our simulation-trained policies transfer well to a real environment and can generalize to novel objects.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available