Hybrid features for skeleton-based action recognition based on network fusion

Authors: Chen, Z., Pan, J., Yang, X. and Qin, H.

Journal: Computer Animation and Virtual Worlds

Volume: 31

Issue: 4-5

eISSN: 1546-427X

ISSN: 1546-4261

DOI: 10.1002/cav.1952

Abstract:

In recent years, the topic of skeleton-based human action recognition has attracted significant attention from researchers and practitioners in graphics, vision, animation, and virtual environments. The most fundamental issue is how to learn an effective and accurate representation from spatiotemporal action sequences towards improved performance, and this article aims to address the aforementioned challenge. In particular, we design a novel method of hybrid features' extraction based on the construction of multistream networks and their organic fusion. First, we train a convolution neural networks (CNN) model to learn CNN-based features with the raw skeleton coordinates and their temporal differences serving as input signals. The attention mechanism is injected into the CNN model to weigh more effective and important information. Then, we employ long short-term memory (LSTM) to obtain long-term temporal features from action sequences. Finally, we generate the hybrid features by fusing the CNN and LSTM networks, and we classify action types with the hybrid features. The extensive experiments are performed on several large-scale publically available databases, and promising results demonstrate the efficacy and effectiveness of our proposed framework.

https://eprints.bournemouth.ac.uk/34481/

Source: Scopus

Hybrid features for skeleton-based action recognition based on network fusion

Authors: Chen, Z., Pan, J., Yang, X. and Qin, H.

Journal: Computer Animation and Virtual Worlds

Volume: 31

Issue: 4-5

ISSN: 1546-4261

Abstract:

© 2020 John Wiley & Sons, Ltd. In recent years, the topic of skeleton-based human action recognition has attracted significant attention from researchers and practitioners in graphics, vision, animation, and virtual environments. The most fundamental issue is how to learn an effective and accurate representation from spatiotemporal action sequences towards improved performance, and this article aims to address the aforementioned challenge. In particular, we design a novel method of hybrid features' extraction based on the construction of multistream networks and their organic fusion. First, we train a convolution neural networks (CNN) model to learn CNN-based features with the raw skeleton coordinates and their temporal differences serving as input signals. The attention mechanism is injected into the CNN model to weigh more effective and important information. Then, we employ long short-term memory (LSTM) to obtain long-term temporal features from action sequences. Finally, we generate the hybrid features by fusing the CNN and LSTM networks, and we classify action types with the hybrid features. The extensive experiments are performed on several large-scale publically available databases, and promising results demonstrate the efficacy and effectiveness of our proposed framework.

https://eprints.bournemouth.ac.uk/34481/

Source: BURO EPrints