Journal
IEEE ROBOTICS AND AUTOMATION LETTERS
Volume 7, Issue 3, Pages 8471-8478Publisher
IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/LRA.2022.3187833
Keywords
Deep learning methods; reinforcement learning; mobile manipulation
Categories
Funding
- Princeton School of Engineering
- National Science Foundation [IIS-1815070, DGE-1656466]
Ask authors/readers for more resources
This study investigates pneumatic non-prehensile manipulation (i.e., blowing) as an efficient method for moving scattered objects into a target receptacle. Using a multi-frequency version of the spatial action maps framework in the context of deep reinforcement learning, the research tackles the challenges of adapting to unexpected changes, maintaining fine-grained control, and inferring long-range plans. Experimental results demonstrate that blowing achieves better performance than pushing and that the proposed policies improve performance over baselines. The study also shows that the system encourages emergent specialization between low-level fine-grained control and high-level planning.
We investigate pneumatic non-prehensile manipulation (i.e., blowing) as a means of efficiently moving scattered objects into a target receptacle. Due to the chaotic nature of aerodynamic forces, a blowing controller must i) continually adapt to unexpected changes from its actions, ii) maintain fine-grained control, since the slightest misstep can result in large unintended consequences (e.g., scatter objects already in a pile), and iii) infer long-range plans (e.g., move the robot to strategic blowing locations). We tackle these challenges in the context of deep reinforcement learning, introducing a multi-frequency version of the spatial action maps framework. This allows for efficient learning of vision-based policies that effectively combine high-level planning and low-level closed-loop control for dynamic mobile manipulation. Experiments show that our system learns efficient behaviors for the task, demonstrating in particular that blowing achieves better downstream performance than pushing, and that our policies improve performance over baselines. Moreover, we show that our system naturally encourages emergent specialization between the different subpolicies spanning low-level fine-grained control and high-level planning. On a real mobile robot equipped with a miniature air blower, we show that our simulation-trained policies transfer well to a real environment and can generalize to novel objects.
Authors
I am an author on this paper
Click your name to claim this paper and add it to your profile.
Reviews
Recommended
No Data Available