4.6 Article

Hands-Free Human-Robot Interaction Using Multimodal Gestures and Deep Learning in Wearable Mixed Reality

期刊

IEEE ACCESS
卷 9, 期 -, 页码 55448-55464

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/ACCESS.2021.3071364

关键词

Robots; Task analysis; Three-dimensional displays; Virtual reality; Service robots; Object detection; Deep learning; Deep learning; eye gazing; hands-free interaction; head gestures; human– robot interaction; mixed reality; object detection

资金

  1. Republic of Korea's Ministry of Science and ICT (MSIT) through the High-Potential Individuals Global Training Program [2020-0-01532]
  2. National Research Foundation of Korea (NRF) - Ministry of Education [2019R1I1A3A01059082]
  3. National Research Foundation of Korea [2019R1I1A3A01059082] Funding Source: Korea Institute of Science & Technology Information (KISTI), National Science & Technology Information Service (NTIS)

向作者/读者索取更多资源

This study introduces a novel hands-free interaction method using multimodal gestures and deep learning for human-robot interaction in mixed reality environments. The method supports coarse-to-fine interactions, utilizing eye gazing for searching and previewing and head gestures for selection and 3D manipulation. Deep learning-based object detection helps estimate initial object positioning, and virtual object-based indirect manipulation enables more intuitive and efficient control of the robot compared to traditional methods.
This study proposes a novel hands-free interaction method using multimodal gestures such as eye gazing and head gestures and deep learning for human-robot interaction (HRI) in mixed reality (MR) environments. Since human operators hold some objects for conducting tasks, there are many constrained situations where they cannot use their hands for HRI interactions. To provide more effective and intuitive task assistance, the proposed hands-free method supports coarse-to-fine interactions. Eye gazing-based interaction is used for coarse interactions such as searching and previewing of target objects, and head gesture interactions are used for fine interactions such as selection and 3D manipulation. In addition, deep learning-based object detection is applied to estimate the initial positioning of physical objects to be manipulated by the robot. The result of object detection is then combined with 3D spatial mapping in the MR environment for supporting accurate initial object positioning. Furthermore, virtual object-based indirect manipulation is proposed to support more intuitive and efficient control of the robot, compared with traditional direct manipulation (e.g., joint-based and end effector-based manipulations). In particular, a digital twin, the synchronized virtual robot of the real robot, is used to provide a preview and simulation of the real robot to manipulate it more effectively and accurately. Two case studies were conducted to confirm the originality and advantages of the proposed hands-free HRI: (1) performance evaluation of initial object positioning and (2) comparative analysis with traditional direct robot manipulations. The deep learning-based initial positioning reduces much effort for robot manipulation using eye gazing and head gestures. The object-based indirect manipulation also supports more effective HRI than previous direct interaction methods.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据