LGCANet: lightweight hand pose estimation network based on HRNet

被引:0
|
作者
Pan, Xiaoying [1 ,2 ]
Li, Shoukun [1 ,2 ]
Wang, Hao [3 ,4 ]
Wang, Beibei [1 ,2 ]
Wang, Haoyi [5 ]
机构
[1] Xian Univ Posts & Telecommun, Sch Comp Sci & Technol, Xian 710121, Shaanxi, Peoples R China
[2] Xian Univ Posts & Telecommun, Shaanxi Key Lab Network Data Anal & Intelligent Pr, Xian 710121, Shaanxi, Peoples R China
[3] Northwestern Polytech Univ, Sch Software, Xian 710072, Shaanxi, Peoples R China
[4] Northwestern Polytech Univ, Natl Engn Lab Air Earth Sea Integrat Big Data Appl, Xian 710121, Shaanxi, Peoples R China
[5] Southwest Univ, Westa Coll, Chongqing, Peoples R China
来源
JOURNAL OF SUPERCOMPUTING | 2024年 / 80卷 / 13期
关键词
Hand pose estimation; High-resolution network; Multi-scale feature fusion; Lightweight network;
D O I
10.1007/s11227-024-06226-2
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Hand pose estimation is a fundamental task in computer vision with applications in virtual reality, gesture recognition, autonomous driving, and virtual surgery. Keypoint detection often relies on deep learning methods and high-resolution feature map representations to achieve accurate detection. The HRNet framework serves as the basis, but it presents challenges in terms of extensive parameter count and demanding computational complexity due to high-resolution representations. To mitigate these challenges, we propose a lightweight keypoint detection network called LGCANet (Lightweight Ghost-Coordinate Attention Network). This network primarily consists of a lightweight feature extraction head for initial feature extraction and multiple lightweight foundational network modules called GCAblocks. GCAblocks introduce linear transformations to generate redundant feature maps while concurrently considering inter-channel relationships and long-range positional information using a coordinate attention mechanism. Validation on the RHD dataset and the COCO-WholeBody-Hand dataset shows that LGCANet reduces the number of parameters by 65.9% and GFLOPs by 72.6% while preserving the accuracy and improves the detection speed.
引用
收藏
页码:19351 / 19373
页数:23
相关论文
共 50 条
  • [31] Attention! A Lightweight 2D Hand Pose Estimation Approach
    Santavas, Nicholas
    Kansizoglou, Ioannis
    Bampis, Loukas
    Karakasis, Evangelos
    Gasteratos, Antonios
    IEEE SENSORS JOURNAL, 2021, 21 (10) : 11488 - 11496
  • [32] DB-HRNet: Dual Branch High-Resolution Network for Human Pose Estimation
    Wang, Yanxia
    Wang, Renjie
    Shi, Hu
    IEEE ACCESS, 2023, 11 : 120628 - 120641
  • [33] Hand Pose Estimation with Attention-and-Sequence Network
    Hu, Tianping
    Wang, Wenhai
    Lu, Tong
    ADVANCES IN MULTIMEDIA INFORMATION PROCESSING, PT I, 2018, 11164 : 556 - 566
  • [34] InterNet plus : A Light Network for Hand Pose Estimation
    Liu, Yang
    Jiang, Jie
    Sun, Jiahao
    Wang, Xianghan
    SENSORS, 2021, 21 (20)
  • [35] InterNet+: A light network for hand pose estimation
    Liu, Yang
    Jiang, Jie
    Sun, Jiahao
    Wang, Xianghan
    Jiang, Jie (JieJiang@nudt.edu.cn), 1600, MDPI (21)
  • [36] Hand pose estimation with multi-scale network
    Zhongxu Hu
    Youmin Hu
    Bo Wu
    Jie Liu
    Dongmin Han
    Thomas Kurfess
    Applied Intelligence, 2018, 48 : 2501 - 2515
  • [37] Enhanced Human Pose Estimation with Attention-Augmented HRNet
    Zhang, Junjie
    Yang, Haojie
    Deng, Yancong
    6TH INTERNATIONAL CONFERENCE ON IMAGE PROCESSING AND MACHINE VISION, IPMV 2024, 2024, : 88 - 93
  • [38] Hand pose estimation with multi-scale network
    Hu, Zhongxu
    Hu, Youmin
    Wu, Bo
    Liu, Jie
    Han, Dongmin
    Kurfess, Thomas
    APPLIED INTELLIGENCE, 2018, 48 (08) : 2501 - 2515
  • [39] Lightweight Network-Based End-to-End Pose Estimation for Noncooperative Targets
    Liu Jiahui
    Zhang Yonghe
    Zhang Wenxiu
    LASER & OPTOELECTRONICS PROGRESS, 2024, 61 (14)
  • [40] LFSimCC: Spatial fusion lightweight network for human pose estimation
    Zheng, Qian
    Guo, Hualing
    Yin, Yunhua
    Zheng, Bin
    Jiang, Hongxu
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2024, 99