4.7 Article

Adaptive detrending to accelerate convolutional gated recurrent unit training for contextual video recognition

期刊

NEURAL NETWORKS
卷 105, 期 -, 页码 356-370

出版社

PERGAMON-ELSEVIER SCIENCE LTD
DOI: 10.1016/j.neunet.2018.05.009

关键词

Detrending; Normalization; Internal covariate shift; Convolutional neural networks (CNNs); Recurrent neural networks (RNNs); Convolutional recurrent neural networks (ConvRNNs)

资金

  1. National Research Foundation of Korea (NRF) - Korea government (MSIP) [2014R1A2A2A01005491]
  2. Okinawa Institute of Science and Technology Graduate University, Japan

向作者/读者索取更多资源

Video image recognition has been extensively studied with rapid progress recently. However, most methods focus on short-term rather than long-term (contextual) video recognition. Convolutional recurrent neural networks (ConvRNNs) provide robust spatio-temporal information processing capabilities for contextual video recognition, but require extensive computation that slows down training. Inspired by normalization and detrending methods, in this paper we propose adaptive detrending'' (AD) for temporal normalization in order to accelerate the training of ConvRNNs, especially of convolutional gated recurrent unit (ConvGRU). For each neuron in a recurrent neural network (RNN), AD identifies the trending change within a sequence and subtracts it, removing the internal covariate shift. In experiments testing for contextual video recognition with ConvGRU, results show that (1) ConvGRU clearly outperforms feed-forward neural networks, (2) AD consistently and significantly accelerates training and improves generalization, (3) performance is further improved when AD is coupled with other normalization methods, and most importantly, (4) the more long-term contextual information is required, the more AD outperforms existing methods. (c) 2018 Elsevier Ltd. All rights reserved.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据