4.7 Article

BoMW: Bag of Manifold Words for One-Shot Learning Gesture Recognition From Kinect

Journal

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TCSVT.2017.2721108

Keywords

Gesture recognition; covariance descriptor; Riemannian manifold; reproducing kernel Hilbert space; kernel sparse coding

Funding

  1. National Natural Science Foundation of China [61572155, 61672188]
  2. Key Research and Development Program of Shandong Province [2016GGX101021]
  3. HIT Outstanding Young Talents Program
  4. Major State Basic Research Development Program of China (973 Program) [2015CB351804]
  5. Natural Science Foundation of China [61403116]
  6. China Postdoctoral Science Foundation [2014M560507]
  7. U.K. EPSRC [EP/N508664/1, EP/R007187/1, EP/N011074/1]
  8. Royal Society-Newton Advanced Fellowship [NA160342]
  9. EPSRC [EP/N508664/1, EP/N011074/1, EP/R007187/1] Funding Source: UKRI

Ask authors/readers for more resources

In this paper, we study one-shot learning gesture recognition on RGB-D data recorded from Microsoft's Kinect. To this end, we propose a novel bag of manifold words (BoMW)based feature representation on symmetric positive definite (SPD) manifolds. In particular, we use covariance matrices to extract local features from RGB-D data due to its compact representation ability as well as the convenience of fusing both RGB and depth information. Since covariance matrices are SPD matrices and the space spanned by them is the SPD manifold, traditional learning methods in the Euclidean space, such as sparse coding, cannot be directly applied to them. To overcome this problem, we propose a unified framework to transfer the sparse coding on SPD manifolds to the one on the Euclidean space, which enables any existing learning method to be used. After building BoMW representation on a video from each gesture class, a nearest neighbor classifier is adopted to perform the one-shot learning gesture recognition. Experimental results on the ChaLearn gesture data set demonstrate the outstanding performance of the proposed one-shot learning gesture recognition method compared against the state-of-the-art methods. The effectiveness of the proposed feature extraction method is also validated on a new RGB-D action recognition data set.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available