4.6 Article

Pairwise-Covariance Multi-view Discriminant Analysis for Robust Cross-View Human Action Recognition

期刊

IEEE ACCESS
卷 9, 期 -, 页码 76097-76111

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/ACCESS.2021.3082142

关键词

Feature extraction; Training; Three-dimensional displays; Neural networks; Cameras; Deep learning; Correlation; Multi-view analysis; action recognition; deep learning; cross-view recognition

资金

  1. Air Force Office of Scientific Research [FA2386-17-1-4056]

向作者/读者索取更多资源

This paper presents a novel method for Human Action Recognition (HAR) under different camera viewpoints using deep learning techniques, achieving robust performance across various datasets, especially for harder classes. The proposed pc-MvDA approach constructs a common feature space to maintain view-invariant features among separated camera views, leading to consistent performance gains in experimental results.
Human action recognition (HAR) under different camera viewpoints is the most critical requirement for practical deployment. In this paper, we propose a novel method that leverages successful deep learning-based features for action representation and multi-view analysis to accomplish robust HAR under viewpoint changes. Specifically, we investigate various deep learning techniques, from 2D CNNs to 3D CNNs to capture spatial and temporal characteristics of actions at each separated camera view. A common feature space is then constructed to keep view-invariant features among extracted streams. This is carried out by learning a set of linear transformations that project private features into the common space in which the classes are well distinguished from each other. To this end, we first adopt Multi-view Discriminant Analysis (MvDA). The original MvDA suffers from odd situations in which the most class-discrepant common space could not be found because its objective is overly concentrated on pushing classes from the global mean but unaware of the distance between specific pairs of adjoining classes. We then introduce a pairwise-covariance maximizing extension that takes pairwise distances between classes into account, namely pc-MvDA. The novel method also differs in the way that could be more favorably applied for large high-dimensional multi-view datasets. Extensive experimental results on four datasets (IXMAS, MuHAVi, MICAGes, NTU RGB+D) show that pc-MvDA achieves consistent performance gain, especially for harder classes. The code is publicly available for research purpose at https://github.com/inspiros/pcmvda.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据