4.6 Article

Pairwise-Covariance Multi-view Discriminant Analysis for Robust Cross-View Human Action Recognition

Journal

IEEE ACCESS
Volume 9, Issue -, Pages 76097-76111

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/ACCESS.2021.3082142

Keywords

Feature extraction; Training; Three-dimensional displays; Neural networks; Cameras; Deep learning; Correlation; Multi-view analysis; action recognition; deep learning; cross-view recognition

Funding

  1. Air Force Office of Scientific Research [FA2386-17-1-4056]

Ask authors/readers for more resources

This paper presents a novel method for Human Action Recognition (HAR) under different camera viewpoints using deep learning techniques, achieving robust performance across various datasets, especially for harder classes. The proposed pc-MvDA approach constructs a common feature space to maintain view-invariant features among separated camera views, leading to consistent performance gains in experimental results.
Human action recognition (HAR) under different camera viewpoints is the most critical requirement for practical deployment. In this paper, we propose a novel method that leverages successful deep learning-based features for action representation and multi-view analysis to accomplish robust HAR under viewpoint changes. Specifically, we investigate various deep learning techniques, from 2D CNNs to 3D CNNs to capture spatial and temporal characteristics of actions at each separated camera view. A common feature space is then constructed to keep view-invariant features among extracted streams. This is carried out by learning a set of linear transformations that project private features into the common space in which the classes are well distinguished from each other. To this end, we first adopt Multi-view Discriminant Analysis (MvDA). The original MvDA suffers from odd situations in which the most class-discrepant common space could not be found because its objective is overly concentrated on pushing classes from the global mean but unaware of the distance between specific pairs of adjoining classes. We then introduce a pairwise-covariance maximizing extension that takes pairwise distances between classes into account, namely pc-MvDA. The novel method also differs in the way that could be more favorably applied for large high-dimensional multi-view datasets. Extensive experimental results on four datasets (IXMAS, MuHAVi, MICAGes, NTU RGB+D) show that pc-MvDA achieves consistent performance gain, especially for harder classes. The code is publicly available for research purpose at https://github.com/inspiros/pcmvda.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available