4.5 Article

VisuoSpatial Foresight for physical sequential fabric manipulation

Journal

AUTONOMOUS ROBOTS
Volume 46, Issue 1, Pages 175-199

Publisher

SPRINGER
DOI: 10.1007/s10514-021-10001-0

Keywords

Deformable object manipulation; Model-based reinforcement learning; Self-supervised learning

Funding

  1. NSF National Robotics Initiative [1734633]
  2. Siemens
  3. Google
  4. Amazon Robotics
  5. Toyota Research Institute
  6. Autodesk
  7. ABB
  8. Samsung
  9. Knapp
  10. Loccioni
  11. Intel
  12. Comcast
  13. Cisco
  14. Hewlett-Packard
  15. PhotoNeo
  16. NVidia
  17. Intuitive Surgical
  18. Graduate Fellowships for STEM Diversity
  19. NSF GRFP

Ask authors/readers for more resources

Robotic fabric manipulation has applications in various fields, but existing techniques are often task-specific and hard to generalize. Researchers leverage the Visual Foresight framework to efficiently reuse learned fabric dynamics for different sequential fabric manipulation tasks.
Robotic fabric manipulation has applications in home robotics, textiles, senior care and surgery. Existing fabric manipulation techniques, however, are designed for specific tasks, making it difficult to generalize across different but related tasks. We build upon the Visual Foresight framework to learn fabric dynamics that can be efficiently reused to accomplish different sequential fabric manipulation tasks with a single goal-conditioned policy. We extend our earlier work on VisuoSpatial Foresight (VSF), which learns visual dynamics on domain randomized RGB images and depth maps simultaneously and completely in simulation. In this earlier work, we evaluated VSF on multi-step fabric smoothing and folding tasks against 5 baseline methods in simulation and on the da Vinci Research Kit surgical robot without any demonstrations at train or test time. A key finding was that depth sensing significantly improves performance: RGBD data yields an 80% improvement in fabric folding success rate in simulation over pure RGB data. In this work, we vary 4 components of VSF, including data generation, visual dynamics model, cost function, and optimization procedure. Results suggest that training visual dynamics models using longer, corner-based actions can improve the efficiency of fabric folding by 76% and enable a physical sequential fabric folding task that VSF could not previously perform with 90% reliability. Code, data, videos, and supplementarymaterial are available at https://sites.google.com/ view/ fabric- vsf/.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.5
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available