Region Attentive Action Unit Intensity Estimation With Uncertainty Weighted Multi-Task Learning

被引:4
|
作者
Chen, Haifeng [1 ,2 ]
Jiang, Dongmei [1 ,2 ]
Zhao, Yong [1 ,2 ]
Wei, Xiaoyong [2 ,3 ]
Lu, Ke [2 ,4 ]
Sahli, Hichem [5 ,6 ]
机构
[1] Northwestern Polytech Univ, Sch Comp Sci, Shaanxi Key Lab Speech & Image Informat Proc, Natl Engn Lab Integrated Aerosp Ground Ocean Big D, Xian 710072, Peoples R China
[2] Peng Cheng Lab, Shenzhen 518055, Guangdong, Peoples R China
[3] Sichuan Univ, Sch Comp Sci, Chengdu 610065, Peoples R China
[4] Univ Chinese Acad Sci, Sch Engn Sci, Beijing 100049, Peoples R China
[5] Vrije Univ Brussel VUB, Dept Elect & Informat ETRO, AVSP Res Lab, B-1050 Brussels, Belgium
[6] Interuniv Microelect Ctr IMEC, B-3001 Heverlee, Belgium
基金
中国国家自然科学基金;
关键词
Action unit intensity; relation learning; multi-head self-attention; multi-task learning; data uncertainty; FACIAL EXPRESSION; PATCHES; MACHINE;
D O I
10.1109/TAFFC.2021.3139101
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Facial action units (AUs) refer to a comprehensive set of atomic facial muscle movements. Recent works have focused on exploring complementary information by learning the relationships among AUs. Most existing approaches process AU co-occurrence and enhance AU recognition by learning the dependencies among AUs from labels, however, the complementary information among features of different AUs are ignored. Moreover, ground truth annotations suffer from a large intra-class variance and their associated intensity levels may vary depending on the annotators' experience. In this paper, we propose the Region Attentive AU intensity estimation method with Uncertainty Weighted Multi-task Learning (RA-UWML). A RoI-Net is first used to extract features from the pre-defined facial patches where the AUs locate. Then, we use the co-occurrence of AUs using both within patch and between patches representation learning. Within a given patch, we propose sharing representation learning in a multi-task manner. To achieve complementarity and avoid redundancy between different image patches, we propose to use a multi-head self-attention mechanism to adaptively and attentively encode each patch specific representation. Moreover, the AU intensity is represented as a Gaussian distribution, instead of a single value, where the mean value indicates the most likely AU intensity and the variance indicates the uncertainty of the estimated AU intensity. The estimated variances are leveraged to automatically weight the loss of each AU in the multitask learning model. In extensive experiments on the Disfa, Fera2015 and Feafa benchmarks, it is shown that the proposed AU intensity estimation model achieves better results compared to the state-of-the-art models.
引用
收藏
页码:2033 / 2047
页数:15
相关论文
共 50 条
  • [1] Attentive Task Interaction Network for Multi-Task Learning
    Sinodinos, Dimitrios
    Armanfard, Narges
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 2885 - 2891
  • [2] On Multi-task Learning for Facial Action Unit Detection
    Zhang, Xiao
    Mahoor, Mohammad H.
    Nielsen, Rodney D.
    PROCEEDINGS OF 2013 28TH INTERNATIONAL CONFERENCE ON IMAGE AND VISION COMPUTING NEW ZEALAND (IVCNZ 2013), 2013, : 202 - 207
  • [3] Attentive Multi-task Deep Reinforcement Learning
    Bram, Timo
    Brunner, Gino
    Richter, Oliver
    Wattenhofer, Roger
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2019, PT III, 2020, 11908 : 134 - 149
  • [4] Attentive multi-task learning for group itinerary recommendation
    Lei Chen
    Jie Cao
    Huanhuan Chen
    Weichao Liang
    Haicheng Tao
    Guixiang Zhu
    Knowledge and Information Systems, 2021, 63 : 1687 - 1716
  • [5] Attentive multi-task learning for group itinerary recommendation
    Chen, Lei
    Cao, Jie
    Chen, Huanhuan
    Liang, Weichao
    Tao, Haicheng
    Zhu, Guixiang
    KNOWLEDGE AND INFORMATION SYSTEMS, 2021, 63 (07) : 1687 - 1716
  • [6] Uncertainty Regularized Multi-Task Learning
    Meshgi, Kourosh
    Mirzaei, Maryam Sadat
    Sekine, Satoshi
    PROCEEDINGS OF THE 12TH WORKSHOP ON COMPUTATIONAL APPROACHES TO SUBJECTIVITY, SENTIMENT & SOCIAL MEDIA ANALYSIS, 2022, : 78 - 88
  • [7] Facial Action Unit Intensity Prediction via Hard Multi-Task Metric Learning for Kernel Regression
    Nicolle, Jeremie
    Bailly, Kevin
    Chetouani, Mohamed
    2015 11TH IEEE INTERNATIONAL CONFERENCE AND WORKSHOPS ON AUTOMATIC FACE AND GESTURE RECOGNITION (FG), VOL. 6, 2015,
  • [8] Co-Attentive Multi-Task Learning for Explainable Recommendation
    Chen, Zhongxia
    Wang, Xiting
    Xie, Xing
    Wu, Tong
    Bu, Guoqing
    Wang, Yining
    Chen, Enhong
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 2137 - 2143
  • [9] Multi-Task Decouple Learning With Hierarchical Attentive Point Process
    Wu, Weichang
    Zhang, Xiaolu
    Zhao, Shiwan
    Fu, Chilin
    Zhou, Jun
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (04) : 1741 - 1757
  • [10] Pose-independent Facial Action Unit Intensity Regression Based on Multi-task Deep Transfer Learning
    Zhou, Yuqian
    Pi, Jimin
    Shi, Bertram E.
    2017 12TH IEEE INTERNATIONAL CONFERENCE ON AUTOMATIC FACE AND GESTURE RECOGNITION (FG 2017), 2017, : 872 - 877