Region Attentive Action Unit Intensity Estimation With Uncertainty Weighted Multi-Task Learning

被引:4
|
作者
Chen, Haifeng [1 ,2 ]
Jiang, Dongmei [1 ,2 ]
Zhao, Yong [1 ,2 ]
Wei, Xiaoyong [2 ,3 ]
Lu, Ke [2 ,4 ]
Sahli, Hichem [5 ,6 ]
机构
[1] Northwestern Polytech Univ, Sch Comp Sci, Shaanxi Key Lab Speech & Image Informat Proc, Natl Engn Lab Integrated Aerosp Ground Ocean Big D, Xian 710072, Peoples R China
[2] Peng Cheng Lab, Shenzhen 518055, Guangdong, Peoples R China
[3] Sichuan Univ, Sch Comp Sci, Chengdu 610065, Peoples R China
[4] Univ Chinese Acad Sci, Sch Engn Sci, Beijing 100049, Peoples R China
[5] Vrije Univ Brussel VUB, Dept Elect & Informat ETRO, AVSP Res Lab, B-1050 Brussels, Belgium
[6] Interuniv Microelect Ctr IMEC, B-3001 Heverlee, Belgium
基金
中国国家自然科学基金;
关键词
Action unit intensity; relation learning; multi-head self-attention; multi-task learning; data uncertainty; FACIAL EXPRESSION; PATCHES; MACHINE;
D O I
10.1109/TAFFC.2021.3139101
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Facial action units (AUs) refer to a comprehensive set of atomic facial muscle movements. Recent works have focused on exploring complementary information by learning the relationships among AUs. Most existing approaches process AU co-occurrence and enhance AU recognition by learning the dependencies among AUs from labels, however, the complementary information among features of different AUs are ignored. Moreover, ground truth annotations suffer from a large intra-class variance and their associated intensity levels may vary depending on the annotators' experience. In this paper, we propose the Region Attentive AU intensity estimation method with Uncertainty Weighted Multi-task Learning (RA-UWML). A RoI-Net is first used to extract features from the pre-defined facial patches where the AUs locate. Then, we use the co-occurrence of AUs using both within patch and between patches representation learning. Within a given patch, we propose sharing representation learning in a multi-task manner. To achieve complementarity and avoid redundancy between different image patches, we propose to use a multi-head self-attention mechanism to adaptively and attentively encode each patch specific representation. Moreover, the AU intensity is represented as a Gaussian distribution, instead of a single value, where the mean value indicates the most likely AU intensity and the variance indicates the uncertainty of the estimated AU intensity. The estimated variances are leveraged to automatically weight the loss of each AU in the multitask learning model. In extensive experiments on the Disfa, Fera2015 and Feafa benchmarks, it is shown that the proposed AU intensity estimation model achieves better results compared to the state-of-the-art models.
引用
收藏
页码:2033 / 2047
页数:15
相关论文
共 50 条
  • [31] Multi-Task Learning with Group Information for Human Action Recognition
    Qian, Li
    Wu, Song
    Pu, Nan
    Xu, Shulin
    Xiao, Guoqiang
    NINTH INTERNATIONAL CONFERENCE ON GRAPHIC AND IMAGE PROCESSING (ICGIP 2017), 2018, 10615
  • [32] Multi-Task Deep Reinforcement Learning for Continuous Action Control
    Yang, Zhaoyang
    Merrick, Kathryn
    Abbass, Hussein
    Jin, Lianwen
    PROCEEDINGS OF THE TWENTY-SIXTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 3301 - 3307
  • [33] Enhancing stance detection through sequential weighted multi-task learning
    Alturayeif, Nora
    Luqman, Hamzah
    Ahmed, Moataz
    SOCIAL NETWORK ANALYSIS AND MINING, 2023, 14 (01)
  • [34] Facial Action Unit detection based on multi-task learning strategy for unlabeled facial images in the wild
    Shang, Ziqiao
    Liu, Bin
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 253
  • [35] MULTI-TASK LEARNING OF GENERALIZABLE REPRESENTATIONS FOR VIDEO ACTION RECOGNITION
    Yao, Zhiyu
    Wang, Yunbo
    Long, Mingsheng
    Wang, Jianmin
    Yu, Philip S.
    Sun, Jiaguang
    2020 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2020,
  • [36] ADAPTIVELY WEIGHTED MULTI-TASK LEARNING USING INVERSE VALIDATION LOSS
    Abbas, Waseem
    Taj, Murtaza
    2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 1408 - 1412
  • [37] MULTI-TASK LEARNING OF EMOTION RECOGNITION AND FACIAL ACTION UNIT DETECTION WITH ADAPTIVELY WEIGHTS SHARING NETWORK
    Wang, Chu
    Zeng, Jiabei
    Shan, Shiguang
    Chen, Xilin
    2019 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2019, : 56 - 60
  • [38] Multi-task learning とmulti-stream monocular depth estimation using integrated model with multi-task learning and multi-stream
    Takamine, Michiru
    Endo, Satoshi
    Transactions of the Japanese Society for Artificial Intelligence, 2021, 36 (05): : 1 - 9
  • [39] Multi-task contrastive learning for semi-supervised medical image segmentation with multi-scale uncertainty estimation
    Xing, Chengcheng
    Dong, Haoji
    Xi, Heran
    Ma, Jiquan
    Zhu, Jinghua
    PHYSICS IN MEDICINE AND BIOLOGY, 2023, 68 (18):
  • [40] A multi-task learning framework for gas detection and concentration estimation
    Liu, Huixiang
    Li, Qing
    Gu, Yu
    NEUROCOMPUTING, 2020, 416 : 28 - 37