Multi-grained clip focus for skeleton-based action recognition

被引:10
|
作者
Qiu, Helei [1 ]
Hou, Biao [1 ]
机构
[1] Xidian Univ, Sch Artificial Intelligence, Int Res Ctr Intelligent Percept & Computat, Key Lab Intelligent Percept & Image Understanding,, Xian 710071, Shaanxi, Peoples R China
基金
中国国家自然科学基金;
关键词
Action recognition; Skeleton; Multi-grain; Self-attention;
D O I
10.1016/j.patcog.2023.110188
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Joint-level and part-level information are crucial for modeling actions with different granularity. In addition, the relevant information on different joints between consecutive frames is very useful for skeleton-based action recognition. To effectively capture the action information, a new multi-grained clip focus network (MGCF-Net) is proposed. Firstly, the skeleton sequence is divided into multiple clips, each containing several consecutive frames. According to the structure of the human body, each clip is divided into several tuples. Then an intra-clip attention module is proposed to capture intra-clip action information. Specifically, multi-head self-attention is divided into two parts, obtaining relevant information at the joint and part levels, and integrating the information captured from these two parts to obtain multi-grained contextual features. In addition, an inter clip focus module is used to capture the key information of several consecutive sub-actions, which will help to distinguish similar actions. On two large-scale benchmarks for skeleton-based action recognition, our method achieves the most advanced performance, and its effectiveness has been verified.
引用
收藏
页数:9
相关论文
共 50 条
  • [31] Bootstrapped Representation Learning for Skeleton-Based Action Recognition
    Moliner, Olivier
    Huang, Sangxia
    Astrom, Kalle
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2022, 2022, : 4153 - 4163
  • [32] Convolutional relation network for skeleton-based action recognition
    Zhu, Jiagang
    Zou, Wei
    Zhu, Zheng
    Hu, Yiming
    NEUROCOMPUTING, 2019, 370 : 109 - 117
  • [33] SKELETON-BASED ACTION RECOGNITION WITH CONVOLUTIONAL NEURAL NETWORKS
    Li, Chao
    Zhong, Qiaoyong
    Xie, Di
    Pu, Shiliang
    2017 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA & EXPO WORKSHOPS (ICMEW), 2017,
  • [34] A Spatiotemporal Fusion Network For Skeleton-Based Action Recognition
    Bao, Wenxia
    Wang, Junyi
    Yang, Xianjun
    Chen, Hemu
    2024 3RD INTERNATIONAL CONFERENCE ON IMAGE PROCESSING AND MEDIA COMPUTING, ICIPMC 2024, 2024, : 347 - 352
  • [35] Memory Attention Networks for Skeleton-Based Action Recognition
    Li, Ce
    Xie, Chunyu
    Zhang, Baochang
    Han, Jungong
    Zhen, Xiantong
    Chen, Jie
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (09) : 4800 - 4814
  • [36] SkeleTR: Towards Skeleton-based Action Recognition in the Wild
    Duan, Haodong
    Xu, Mingze
    Shuai, Bing
    Modolo, Davide
    Tu, Zhuowen
    Tighe, Joseph
    Bergamo, Alessandro
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 13588 - 13598
  • [37] Memory Attention Networks for Skeleton-based Action Recognition
    Xie, Chunyu
    Li, Ce
    Zhang, Baochang
    Chen, Chen
    Han, Jungong
    Liu, Jianzhuang
    PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2018, : 1639 - 1645
  • [38] SKELETON-BASED ACTION RECOGNITION USING LSTM AND CNN
    Li, Chuankun
    Wang, Pichao
    Wang, Shuang
    Hou, Yonghong
    Li, Wanqing
    2017 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA & EXPO WORKSHOPS (ICMEW), 2017,
  • [39] Pose Encoding for Robust Skeleton-Based Action Recognition
    Demisse, Girum G.
    Papadopoulos, Konstantinos
    Aouada, Djamila
    Ottersten, Bjorn
    PROCEEDINGS 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW), 2018, : 301 - 307
  • [40] Hypergraph Neural Network for Skeleton-Based Action Recognition
    Hao, Xiaoke
    Li, Jie
    Guo, Yingchun
    Jiang, Tao
    Yu, Ming
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2021, 30 : 2263 - 2275