4.7 Article

Video representation learning for temporal action detection using global-local attention

Journal

PATTERN RECOGNITION
Volume 134, Issue -, Pages -

Publisher

ELSEVIER SCI LTD
DOI: 10.1016/j.patcog.2022.109135

Keywords

Temporal action detection; Video representation; Untrimmed video analysis

Funding

  1. National Natural Sci- ence Foundation of China
  2. Key Research and Development Program in the Shaanxi Province of China
  3. Natural Science Basic Research Program of Shaanxi
  4. [61976167]
  5. [U19B2030]
  6. [62101416]
  7. [2021GY082]
  8. [2022JQ- 708]

Ask authors/readers for more resources

Video representation is crucial for temporal action detection, with different requirements for action classification and action localization. This paper proposes a Global-Local Attention (GLA) mechanism to produce a powerful video representation without additional parameters. GLA enhances the discriminability and localization ability of video representation through global and local attention mechanisms, achieving state-of-the-art performance.
Video representation is of significant importance for temporal action detection. The two sub-tasks of temporal action detection, i.e., action classification and action localization, have different requirements for video representation. Specifically, action classification requires video representations to be highly dis-criminative, so that action features and background features are as dissimilar as possible. For action local-ization, it is crucial to obtain information about the action itself and the surrounding context for accurate prediction of action boundaries. However, the previous methods failed to extract the optimal representa-tions for the two sub-tasks, whose representations for both sub-tasks are obtained in a similar way. In this paper, a Global-Local Attention (GLA) mechanism is proposed to produce a more powerful video rep-resentation for temporal action detection without introducing additional parameters. The global attention mechanism predicts each action category by integrating features in the entire video that are similar to the action while suppressing other features, thus enhancing the discriminability of video representation during the training process. The local attention mechanism uses a Gaussian weighting function to inte-grate each action and its surrounding contextual information, thereby enabling precise localization of the action. The effectiveness of GLA is demonstrated on THUMOS'14 and ActivityNet-1.3 with a simple one -stage action detection network, achieving state-of-the-art performance among the methods using only RGB images as input. The inference speed of the proposed model reaches 1373 FPS on a single Nvidia Titan Xp GPU. The generalizability of GLA to other detection architectures is verified using R-C3D and Decouple-SSAD, both of which achieve consistent improvements. The experimental results demonstrate that designing representations with different properties for the two sub-tasks leads to better performance for temporal action detection compared to the representations obtained in a similar way.(c) 2022 Elsevier Ltd. All rights reserved.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available