4.8 Article

Toward Proactive Human-Robot Collaborative Assembly: A Multimodal Transfer-Learning-Enabled Action Prediction Approach

Journal

IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS
Volume 69, Issue 8, Pages 8579-8588

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TIE.2021.3105977

Keywords

Robots; Three-dimensional displays; Collaboration; Service robots; Skeleton; Videos; Visualization; Action recognition; human-robot collaboration; multimodal intelligence; transfer learning

Funding

  1. Research Committee of The Hong Kong Polytechnic University under Departmental General Research Fund under Grant G-UAHH
  2. Human Subjects Ethics Sub-Committee at the Hong Kong Polytechnic University [HSEARS20201110002]
  3. Laboratory for Artificial Intelligence in Design, under Hong Kong Special Administrative Region [RP2-1]

Ask authors/readers for more resources

This article proposes a multimodal transfer-learning-enabled action prediction approach, serving as a prerequisite for achieving proactive human-robot collaborative assembly. By leveraging intelligent action recognition and transfer-learning models, ongoing human actions can be predicted and rapidly converted into industrial assembly operations. A dynamic decision-making mechanism allows mobile robots to assist operators in a proactive manner. Experimental results demonstrate that the proposed approach outperforms other state-of-the-art methods in efficient action prediction.
Human-robot collaborative assembly (HRCA) is vital for achieving high-level flexible automation for mass personalization in today's smart factories. However, existing works in both industry and academia mainly focus on the adaptive robot planning, while seldom consider human operator's intentions in advance. Hence, it hinders the HRCA transition toward a proactive manner. To overcome the bottleneck, this article proposes a multimodal transfer-learning-enabled action prediction approach, serving as the prerequisite to ensure the proactive HRCA. First, a multimodal intelligence-based action recognition approach is proposed to predict ongoing human actions by leveraging the visual stream and skeleton stream with short-time input frames. Second, a transfer-learning-enabled model is adapted to transfer learnt knowledge from daily activities to industrial assembly operations rapidly for online operator intention analysis. Third, a dynamic decision-making mechanism, including robotic decision and motion control, is described to allow mobile robots to assist operators in a proactive manner. Finally, an aircraft bracket assembly task is demonstrated in the laboratory environment, and the comparative study result shows that the proposed approach outperforms other state-of-the-art ones for efficient action prediction.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.8
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available