4.7 Article

Multi-scale skeleton adaptive weighted GCN for skeleton-based human action recognition in IoT

Journal

APPLIED SOFT COMPUTING
Volume 104, Issue -, Pages -

Publisher

ELSEVIER
DOI: 10.1016/j.asoc.2021.107236

Keywords

Internet of things (IoT); Skeleton-based human action recognition; Graph convolution network (GCN); Graph topology; Multi-scale

Funding

  1. 111 project, China [B17007]
  2. Director Funds of Beijing Key Laboratory of Network System Architecture and Convergence, China [2017BKLNSAC-ZJ-01]

Ask authors/readers for more resources

This paper proposes a multi-scale skeleton adaptive weighted graph convolution network (MSAWGCN) for skeleton-based action recognition, which extracts more abundant spatial features of skeletons through multi-scale skeleton graph convolution network and adopts a simple graph vertex fusion strategy to learn the graph topology adaptively.
Skeleton-based human action recognition has become a hot topic due to its potential advantages. Graph convolution network (GCN) has obtained remarkable performances in the modeling of skeleton-based human action recognition in IoT. In order to capture robust spatial-temporal features from the human skeleton, a powerful feature extractor is essential. However, Most GCN-based methods use the fixed graph topology. Besides, only a single-scale feature is used, and the multi-scale information is ignored. In this paper, we propose a multi-scale skeleton adaptive weighted graph convolution network (MSAWGCN) for skeleton-based action recognition. Specifically, a multi-scale skeleton graph convolution network is adopted to extract more abundant spatial features of skeletons. Moreover, we develop a simple graph vertex fusion strategy, which can learn the latent graph topology adaptively by replacing the handcrafted adjacency matrix with a learnable matrix. According to different sampling strategies, weighted learning method is adopted to enrich features while aggregating. Experiments on three large datasets illustrate that the proposed method achieves comparable performances to state-of-the-art methods. Our proposed method attains an improvement of 0.9% and 0.7% respectively over the recent GCN-based method on the NTU RGB+D and Kinetics dataset. (c) 2021 Elsevier B.V. All rights reserved.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available