An Attention-Aware Model for Human Action Recognition on Tree-Based Skeleton Sequences

被引:1
|
作者
Ding, Runwei [1 ]
Liu, Chang [1 ]
Liu, Hong [1 ,2 ]
机构
[1] Peking Univ, Shenzhen Grad Sch, Shenzhen, Peoples R China
[2] Peking Univ, Key Lab Machine Percept, Beijing, Peoples R China
来源
SOCIAL ROBOTICS, ICSR 2018 | 2018年 / 11357卷
基金
中国国家自然科学基金;
关键词
Human action recognition; Skeleton; Attention-ware model; Tri-directional Tree Traversal Map (TTTM);
D O I
10.1007/978-3-030-05204-1_56
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Skeleton-based human action recognition (HAR) has attracted a lot of research attentions because of robustness to variations of locations and appearances. However, most existing methods treat the whole skeleton as a fixed pattern, in which the importance of different skeleton joints for action recognition is not considered. In this paper, a novel CNN-based attention-ware network is proposed. First, to describe the semantic meaning of skeletons and learn the discriminative joints over time, an attention generate network named Global Attention Network (GAN) is proposed to generate attention masks. Then, to encode the spatial structure of skeleton sequences, we design a tree-based traversal (TTTM) rule, which can represent the skeleton structure, as a convolution unit of main network. Finally, the GAN and main network are cascaded as a whole network which is trained in an end-to-end manner. Experiments show that the TTTM and GAN are supplemented each other, and the whole network achieves an efficient improvement over the state-of-the-arts, e.g., the classification accuracy of this network was 83.6% and 89.5% on NTU-RGBD CV and CS dataset, which outperforms any other methods.
引用
收藏
页码:569 / 579
页数:11
相关论文
共 50 条
  • [1] Skeleton-based attention-aware spatial-temporal model for action detection and recognition
    Cui, Ran
    Zhu, Aichun
    Wu, Jingran
    Hua, Gang
    IET COMPUTER VISION, 2020, 14 (05) : 177 - 184
  • [2] Human-centered attention-aware networks for action recognition
    Liu, Shuai
    Li, Yating
    Fu, Weina
    INTERNATIONAL JOURNAL OF INTELLIGENT SYSTEMS, 2022, 37 (12) : 10968 - 10987
  • [3] Context-Aware Cross-Attention for Skeleton-Based Human Action Recognition
    Fan, Yanbo
    Weng, Shuchen
    Zhang, Yong
    Shi, Boxin
    Zhang, Yi
    IEEE ACCESS, 2020, 8 (08): : 15280 - 15290
  • [4] Attention-Aware Sampling via Deep Reinforcement Learning for Action Recognition
    Dong, Wenkai
    Zhang, Zhaoxiang
    Tan, Tieniu
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 8247 - 8254
  • [5] Action Recognition by an Attention-Aware Temporal Weighted Convolutional Neural Network
    Wang, Le
    Zang, Jinliang
    Zhang, Qilin
    Niu, Zhenxing
    Hua, Gang
    Zheng, Nanning
    SENSORS, 2018, 18 (07)
  • [6] STAP: Spatial-Temporal Attention-Aware Pooling for Action Recognition
    Nguyen, Tam V.
    Song, Zheng
    Yan, Shuicheng
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2015, 25 (01) : 77 - 86
  • [7] Identifying the key frames: An attention-aware sampling method for action recognition
    Dong, Wenkai
    Zhang, Zhaoxiang
    Song, Chunfeng
    Tan, Tieniu
    PATTERN RECOGNITION, 2022, 130
  • [8] AttentionXML: Label Tree-based Attention-Aware Deep Model for High-Performance Extreme Multi-Label Text Classification
    You, Ronghui
    Zhang, Zihan
    Wang, Ziye
    Dai, Suyang
    Mamitsuka, Hiroshi
    Zhu, Shanfeng
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [9] Skeleton-Based Human Action Recognition With Global Context-Aware Attention LSTM Networks
    Liu, Jun
    Wang, Gang
    Duan, Ling-Yu
    Abdiyeva, Kamila
    Kot, Alex C.
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2018, 27 (04) : 1586 - 1599
  • [10] Action Tree Convolutional Networks: Skeleton-Based Human Action Recognition
    Liu, Wenjie
    Zhang, Ziyi
    Han, Bing
    Zhu, Chenhui
    ADVANCES IN MULTIMEDIA INFORMATION PROCESSING, PT III, 2018, 11166 : 783 - 792