Leveraging vision transformers and entropy-based attention for accurate micro-expression recognition

被引:0
|
作者
Yibo Zhang [1 ]
Weiguo Lin [3 ]
Yuanfa Zhang [1 ]
Junfeng Xu [1 ]
Yan Xu [1 ]
机构
[1] Communication University of China,School of Computer and Cyberspace Security
[2] Emergency General Hospital,Department of Nephrology
[3] North China Institute of Science and Technology,School of Computing
关键词
Vision transformer; Micro-expression recognition; Agent attention;
D O I
10.1038/s41598-025-98610-y
中图分类号
学科分类号
摘要
Micro-expressions are difficult to fake and inherently truthful, making micro-expression recognition technology widely applicable across various domains. With the development of artificial intelligence, the accuracy and efficiency of micro-expression recognition systems have been significantly improved. However, the short duration and subtle facial movement changes present significant challenges to real-time recognition and accuracy. To address these issues, this paper proposes a novel micro-expression recognition method based on the Vision Transformer. First, a new model called HTNet with LAPE (hierarchical transformer network with learnable absolute position embedding) is introduced to improve the model’s capacity for capturing subtle facial features, thereby enhancing the accuracy of micro-expression recognition. Second, an entropy-based selection agent attention is proposed to reduce the model parameters and computational effort while preserving its learning capability. Finally, a diffusion model is utilized for data augmentation to expand the micro-expression sample size, further enhancing the model’s generalization, accuracy, and robustness. Extensive experiments conducted on multiple datasets validate the framework’s effectiveness and highlight its potential in real-world applications.
引用
收藏
相关论文
共 50 条
  • [31] Micro-Expression Recognition Based on Spatiotemporal Gabor Filters
    Lin, Chenhan
    Long, Fei
    Huang, JianMing
    Li, Jun
    2018 8TH INTERNATIONAL CONFERENCE ON INFORMATION SCIENCE AND TECHNOLOGY (ICIST 2018), 2018, : 487 - 491
  • [32] Micro-expression recognition based on differential feature fusion
    Ziyang Shang
    Penghai Wang
    Xinfu Li
    Multimedia Tools and Applications, 2024, 83 : 11111 - 11126
  • [33] Micro-Expression Recognition Based on Temporal Sensitive Transformer
    Li, Daxiang
    Qiao, Nannan
    Liu, Ying
    2024 6TH INTERNATIONAL CONFERENCE ON NATURAL LANGUAGE PROCESSING, ICNLP 2024, 2024, : 484 - 489
  • [34] A survey: facial micro-expression recognition
    Madhumita Takalkar
    Min Xu
    Qiang Wu
    Zenon Chaczko
    Multimedia Tools and Applications, 2018, 77 : 19301 - 19325
  • [35] Inceptr: micro-expression recognition integrating inception-CBAM and vision transformer
    Haoliang Zhou
    Shucheng Huang
    Yuqiao Xu
    Multimedia Systems, 2023, 29 : 3863 - 3876
  • [36] Inceptr: micro-expression recognition integrating inception-CBAM and vision transformer
    Zhou, Haoliang
    Huang, Shucheng
    Xu, Yuqiao
    MULTIMEDIA SYSTEMS, 2023, 29 (6) : 3863 - 3876
  • [37] Micro-Expression Recognition Using Convolutional Variational Attention Transformer (ConVAT) With Multihead Attention Mechanism
    Bin Talib, Hafiz Khizer
    Xu, Kaiwei
    Cao, Yanlong
    Xu, Yuan Ping
    Xu, Zhijie
    Zaman, Muhammad
    Akhunzada, Adnan
    IEEE ACCESS, 2025, 13 : 20054 - 20070
  • [38] AU-assisted Graph Attention Convolutional Network for Micro-Expression Recognition
    Xie, Hong-Xia
    Lo, Ling
    Shuai, Hong-Han
    Cheng, Wen-Huang
    MM '20: PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, 2020, : 2871 - 2880
  • [39] Apex Frame Spotting Using Attention Networks for Micro-Expression Recognition System
    Yee, Ng Lai
    Zulkifley, Mohd Asyraf
    Saputro, Adhi Harmoko
    Abdani, Siti Raihanah
    CMC-COMPUTERS MATERIALS & CONTINUA, 2022, 73 (03): : 5331 - 5348
  • [40] Investigating LSTM for Micro-Expression Recognition
    Bai, Mengjiong
    Goecke, Roland
    COMPANION PUBLICATON OF THE 2020 INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION (ICMI '20 COMPANION), 2020, : 7 - 11