Toward Mathematical Representation of Emotion: A Deep Multitask Learning Method Based On Multimodal Recognition

被引:1
|
作者
Harata, Seiichi [1 ]
Sakuma, Takuto [1 ]
Kato, Shohei [1 ]
机构
[1] Nagoya Inst Technol, Nagoya, Aichi, Japan
关键词
Affective Computing; Deep Neural Networks; Multimodal Fusion; Multitask Learning; Emotional Space;
D O I
10.1145/3395035.3425254
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
To emulate human emotions in agents, the mathematical representation of emotion (an emotional space) is essential for each component, such as emotion recognition, generation, and expression. In this study, we aim to acquire a modality-independent emotional space by extracting shared emotional information from different modalities. We propose a method of acquiring an emotional space by integrating multimodalities on a DNN and combining the emotion recognition task and the unification task. The emotion recognition task learns the representation of emotions, and the unification task learns an identical emotional space from each modality. Through the experiments with audio-visual data, we confirmed that there are differences in emotional spaces acquired from unimodality, and the proposed method can acquire a joint emotional space. We also indicated that the proposed method could adequately represent emotions in a low-dimensional emotional space, such as in five or six dimensions, under this paper's experimental conditions.
引用
收藏
页码:47 / 51
页数:5
相关论文
共 50 条
  • [31] Multimodal Multitask Emotion Recognition using Images, Texts and Tags
    Fortin, Mathieu Page
    Chaib-draa, Brahim
    PROCEEDINGS OF THE ACM WORKSHOP ON CROSSMODAL LEARNING AND APPLICATION (WCRML'19), 2019, : 3 - 10
  • [32] Towards Learning a Joint Representation from Transformer in Multimodal Emotion Recognition
    Deng, James J.
    Leung, Clement H. C.
    BRAIN INFORMATICS, BI 2021, 2021, 12960 : 179 - 188
  • [33] A Survey of Deep Learning-Based Multimodal Emotion Recognition: Speech, Text, and Face
    Lian, Hailun
    Lu, Cheng
    Li, Sunan
    Zhao, Yan
    Tang, Chuangao
    Zong, Yuan
    ENTROPY, 2023, 25 (10)
  • [34] Audio-Video Based Multimodal Emotion Recognition Using SVMs and Deep Learning
    Sun, Bo
    Xu, Qihua
    He, Jun
    Yu, Lejun
    Li, Liandong
    Wei, Qinglan
    PATTERN RECOGNITION (CCPR 2016), PT II, 2016, 663 : 621 - 631
  • [35] Transformer-Based Self-Supervised Multimodal Representation Learning for Wearable Emotion Recognition
    Wu, Yujin
    Daoudi, Mohamed
    Amad, Ali
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2024, 15 (01) : 157 - 172
  • [36] Multitask Representation Learning for Multimodal Estimation of Depression Level
    Qureshi, Syed Arbaaz
    Saha, Sriparna
    Hasanuzzaman, Mohammed
    Dias, Gael
    IEEE INTELLIGENT SYSTEMS, 2019, 34 (05) : 45 - 52
  • [37] An Emotion Recognition Method Using Speech Signals Based on Deep Learning
    Byun, Sung-woo
    Shin, Bo-ra
    Lee, Seok-Pil
    BASIC & CLINICAL PHARMACOLOGY & TOXICOLOGY, 2019, 124 : 181 - 182
  • [38] Multi-Label Multimodal Emotion Recognition With Transformer-Based Fusion and Emotion-Level Representation Learning
    Le, Hoai-Duy
    Lee, Guee-Sang
    Kim, Soo-Hyung
    Kim, Seungwon
    Yang, Hyung-Jeong
    IEEE ACCESS, 2023, 11 : 14742 - 14751
  • [39] Joint Representation and Recognition for Ship-Radiated Noise Based on Multimodal Deep Learning
    Yuan, Fei
    Ke, Xiaoquan
    Cheng, En
    JOURNAL OF MARINE SCIENCE AND ENGINEERING, 2019, 7 (11)
  • [40] Chinese Multimodal Emotion Recognition in Deep and Traditional Machine Learning Approaches
    Miao, Haotian
    Zhang, Yifei
    Li, Weipeng
    Zhang, Haoran
    Wang, Daling
    Feng, Shi
    2018 FIRST ASIAN CONFERENCE ON AFFECTIVE COMPUTING AND INTELLIGENT INTERACTION (ACII ASIA), 2018,