3.8 Proceedings Paper

MARS: Motion-Augmented RGB Stream for Action Recognition

出版社

IEEE
DOI: 10.1109/CVPR.2019.00807

关键词

-

资金

  1. ERC advanced grant Allegro
  2. ANR (AVENUE) [ANR-18-CE23-0011]
  3. Agence Nationale de la Recherche (ANR) [ANR-18-CE23-0011] Funding Source: Agence Nationale de la Recherche (ANR)

向作者/读者索取更多资源

Most state-of-the-art methods for action recognition consist of a two-stream architecture with 3D convolutions: an appearancestreamfor RGB frames and a motion stream for opticalflowframes. Although combiningflow with RGB improves the performance, the cost of computing accurate optical flow is high, and increases action recognition latency. This limits the usage of two-stream approaches in real-worldapplicationsrequiringlow latency. In thispaper we introduce two learning approachesto train a standard 3D CNN, operating on RGB frames, that mimics the motion stream, and as a result avoidsflow computation at test time. First,by minimizing a feature-based loss compared to the Flow stream, we show that the network reproduces the motion stream with high fidelity. Second, to leverage both appearance and motion information effectively, we train with a linear combination of the feature-basedloss and the standardcross-entropy loss for action recognition. We denote the stream trainedusing this combined loss as MotionAugmented RGB Stream (MARS). As a single stream, MARS performs better than RGB or Flow alone, for instance with 72.7% accuracy on Kinetics comparedto 72.0% and 65.6% with RGB and Flow streams respectively.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

3.8
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据