4.7 Article

Arbitrary-View Human Action Recognition: A Varying-View RGB-D Action Dataset

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TCSVT.2020.2975845

关键词

Skeleton; Three-dimensional displays; Sensors; Videos; Two dimensional displays; Dictionaries; Robots; Human action recognition; varying-view RGB-D action dataset; cross-view recognition; arbitrary-view recognition; HRI

资金

  1. Natural Science Foundation of China (NSFC) [61673088]
  2. 111 Project [B17008]

向作者/读者索取更多资源

Arbitrary-view action recognition remains a challenging problem due to view changes and visual occlusions. To address this issue, researchers have collected a large-scale RGB-D action dataset with diverse data types, rich action performances, and different viewpoints, providing valuable and challenging data for evaluating arbitrary-view recognition.
Current researches of action recognition which focus on single-view and multi-view recognition can hardly satisfy the requirements of human-robot interaction (HRI) applications for recognizing human actions from arbitrary views. Arbitrary-view recognition is still a challenging issue due to view changes and visual occlusions. In addition, the lack of datasets also sets up barriers. To provide data for arbitrary-view action recognition, we collect a new large-scale RGB-D action dataset for arbitrary-view action analysis, including RGB videos, depth and skeleton sequences. The dataset includes action samples captured in 8 fixed viewpoints and varying-view sequences which cover the entire 360 degrees view angles. In total, 118 persons are invited to act 40 action categories. Our dataset involves more participants, more viewpoints and a large number of samples. More importantly, it is the first dataset containing the entire 360 degrees varying-view sequences. The dataset provides sufficient data for multi-view, cross-view and arbitrary-view action analysis. Besides, we propose a View-guided Skeleton CNN (VS-CNN) to tackle the problem of arbitrary-view action recognition. Experiment results show that the VS-CNN achieves superior performance, and our dataset provides valuable but challenging data for the evaluation of arbitrary-view recognition.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据