4.5 Article

Learning Video Actions in Two Stream Recurrent Neural Network

期刊

PATTERN RECOGNITION LETTERS
卷 151, 期 -, 页码 200-208

出版社

ELSEVIER
DOI: 10.1016/j.patrec.2021.08.017

关键词

Action recognition; Two-stream deep network; LSTM; Feature fusion

资金

  1. Kuwait Foundation for the Advancement of Sciences, Kuwait [PR18-18QS-01]

向作者/读者索取更多资源

This paper investigates the use of LSTM networks for human action recognition in videos. By learning the fusion of spatial and temporal feature streams with LSTM, a novel two-stream deep network is proposed, efficiently capturing long range temporal dependencies in video streams.
The paper investigates the Long short term memory (LSTM) networks for human action recognition in videos. In spite of significant progress in the field, recognizing actions in real-world videos is a challenging task due to the spatial and temporal variations within and across video clips. We propose a novel two-stream deep network for action recognition by applying the LSTM for learning the fusion of spatial and temporal feature streams. The LSTM type of Recurrent neural network by design possess unique capability to preserve long range context in temporal streams. The proposed method capitalizes on LSTMs memory attribute to fuse the input streams in high-dimensional space exploring the spatial and temporal correlations. The temporal stream input is defined on the LSTM learned deep features summarizing the input frame sequence. Our approach of combining the convolutional features based spatial stream and the deep features based temporal stream in LSTM network efficiently captures the long range temporal dependencies in video streams. We perform primary evaluation of the proposed approach on UCF101, HMBD51 and Kinetics400 datasets achieving competitive recognition accuracy of 93.1%, 71.3% and 74.6% respectively. (c) 2021 Elsevier B.V. All rights reserved.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.5
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据