4.3 Article

3D SMoSIFT: three-dimensional sparse motion scale invariant feature transform for activity recognition from RGB-D videos

期刊

JOURNAL OF ELECTRONIC IMAGING
卷 23, 期 2, 页码 -

出版社

SPIE-SOC PHOTO-OPTICAL INSTRUMENTATION ENGINEERS
DOI: 10.1117/1.JEI.23.2.023017

关键词

three-dimensional sparse motion scale-invariant feature transform; bag of words model; spatiotemporal feature; optical flow; RGB-D data

资金

  1. National Natural Science Foundation of China [61172128]
  2. National Key Basic Research Program of China [2012CB316304]
  3. New Century Excellent Talents in University [NCET-12-0768]
  4. fundamental research funds for the central universities [2013JBZ003]
  5. Program for Innovative Research Team in University of Ministry of Education of China [IRT201206]
  6. Beijing Higher Education Young Elite Teacher Project [YETP0544]
  7. Research Fund for the Doctoral Program of Higher Education of China [20120009110008]

向作者/读者索取更多资源

Human activity recognition based on RGB-D data has received more attention in recent years. We propose a spatiotemporal feature named three-dimensional (3D) sparse motion scale-invariant feature transform (SIFT) from RGB-D data for activity recognition. First, we build pyramids as scale space for each RGB and depth frame, and then use Shi-Tomasi corner detector and sparse optical flow to quickly detect and track robust keypoints around the motion pattern in the scale space. Subsequently, local patches around keypoints, which are extracted from RGB-D data, are used to build 3D gradient and motion spaces. Then SIFT-like descriptors are calculated on both 3D spaces, respectively. The proposed feature is invariant to scale, transition, and partial occlusions. More importantly, the running time of the proposed feature is fast so that it is well-suited for real-time applications. We have evaluated the proposed feature under a bag of words model on three public RGB-D datasets: one-shot learning Chalearn Gesture Dataset, Cornell Activity Dataset-60, and MSR Daily Activity 3D dataset. Experimental results show that the proposed feature outperforms other spatiotemporal features and are comparative to other state-of-the-art approaches, even though there is only one training sample for each class. (C) 2014 SPIE and IS&T

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.3
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据