3.8 Proceedings Paper

VP-GO: A 'Light' Action-Conditioned Visual Prediction Model for Grasping Objects

出版社

IEEE
DOI: 10.1109/ICARM54641.2022.9959321

关键词

-

资金

  1. CHIST-ERA project HEAP
  2. Creativ'Lab in Loria

向作者/读者索取更多资源

In this paper, VP-GO, a light stochastic action-conditioned visual prediction model, is proposed for robotic grasping of unknown soft objects. By decomposing semantic actions into elementary movements, compatibility with existing models and datasets is ensured. A new open dataset called PandaGrasp is also provided for visual prediction of object grasping.
Visual prediction models are promising solutions for visual-based robotic grasping of cluttered, unknown soft objects. Previous models from the literature are computationally greedy, which limits reproducibility; although some consider stochasticity in the prediction model, it is often too weak to catch the reality of robotics experiments involving grasping such objects. Furthermore, previous work focused on elementary movements that are not efficient to reason in terms of more complex semantic actions. To address these limitations, we propose VP-GO, a light stochastic action-conditioned visual prediction model. We propose a hierarchical decomposition of semantic grasping and manipulation actions into elementary end-effector movements, to ensure compatibility with existing models and datasets for visual prediction of robotic actions such as RoboNet. We also record and release a new open dataset for visual prediction of object grasping, called PandaGrasp. Our model can be pre-trained on RoboNet and fine-tuned on PandaGrasp, and performs similarly to more complex models in terms of signal prediction metrics. Qualitatively, it outperforms when predicting the outcome of complex grasps performed by our robot.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

3.8
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据