PGIF: A Personality-Guided Iterative Feedback Graph Network for Multimodal Conversational Emotion Recognition

被引:0
|
作者
Xie, Yunhe [1 ]
Mao, Rui [2 ]
机构
[1] Harbin Inst Technol, Fac Comp, Harbin 150001, Peoples R China
[2] Nanyang Technol Univ, Coll Comp & Data Sci, Singapore 639798, Singapore
关键词
Emotion recognition; Iterative methods; Pragmatics; Feature extraction; Vectors; Semantics; Oral communication; Long short term memory; Correlation; Context modeling; interlocutor-induced pragmatic variation; iterative feedback fusion mechanism; multimodal conversational emotion recognition (MCER);
D O I
10.1109/TCSS.2024.3523322
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Multimodal emotion recognition in conversation (MERC) aims to identify emotions in target utterances from multimodal records, drawing significant attention for its value in conversational artificial intelligence. While early research focuses on exploring conversational context, recent efforts emphasize integrating multimodal cues. Existing methods focus on modeling the impact of conversational context on emotion recognition while neglecting the role of the speaker's personality factors. Furthermore, these approaches often suffer from inefficiencies in information transfer due to full-utterance connectivity and fail to leverage multiple fusion modes for complementary benefits. To address these issues, we propose a personality-guided iterative feedback graph network (PGIF) for MERC. PGIF incorporates personality information as a specialized modality to enhance the feature space for emotional inference. We utilize a graph network to model information flow, integrating interlocutor-aware contextual information by considering interlocutor dependencies between utterances. Additionally, we employ a dialogue discourse parser to directly model semantic relationships between utterances. Our iterative feedback fusion mechanism explicitly simulates emotional interactions between feature-level and decision-level modalities, improving inference without ground truth labels through iterative refinement. PGIF demonstrates improvements of 1.94% and 1.42% over state-of-the-art methods on the IEMOCAP and MELD datasets, respectively. Ablation studies further validate the effectiveness of PGIF's mechanisms, while its manipulation of input features and global fusion strategies ensures compatibility with existing approaches.
引用
收藏
页数:13
相关论文
共 50 条
  • [21] GraphCFC: A Directed Graph Based Cross-Modal Feature Complementation Approach for Multimodal Conversational Emotion Recognition
    Li, Jiang
    Wang, Xiaoping
    Lv, Guoqing
    Zeng, Zhigang
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 77 - 89
  • [22] Ugan: Uncertainty-Guided Graph Augmentation Network for EEG Emotion Recognition
    Chen, Bianna
    Chen, C. L. Philip
    Zhang, Tong
    IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2024,
  • [23] ICON: Interactive Conversational Memory Network for Multimodal Emotion Detection
    Hazarika, Devamanyu
    Poria, Soujanya
    Mihalcea, Rada
    Cambria, Erik
    Zimmermann, Roger
    2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), 2018, : 2594 - 2604
  • [24] Context- and Knowledge-Aware Graph Convolutional Network for Multimodal Emotion Recognition
    Fu, Yahui
    Okada, Shogo
    Wang, Longbiao
    Guo, Lili
    Song, Yaodong
    Liu, Jiaxing
    Dang, Jianwu
    IEEE MULTIMEDIA, 2022, 29 (03) : 91 - 99
  • [25] Dense Graph Convolutional With Joint Cross-Attention Network for Multimodal Emotion Recognition
    Cheng, Cheng
    Liu, Wenzhe
    Feng, Lin
    Jia, Ziyu
    IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2024, 11 (05): : 6672 - 6683
  • [26] Revisiting Disentanglement and Fusion on Modality and Context in Conversational Multimodal Emotion Recognition
    Li, Bobo
    Fei, Hao
    Liao, Lizi
    Zhao, Yu
    Teng, Chong
    Chua, Tat-Seng
    Ji, Donghong
    Li, Fei
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 5923 - 5934
  • [27] Emotion Recognition Based on Feedback Weighted Fusion of Multimodal Emotion Data
    Wei, Wei
    Jia, Qingxuan
    Feng, Yongli
    2017 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND BIOMIMETICS (IEEE ROBIO 2017), 2017, : 1682 - 1687
  • [28] LR-GCN: Latent Relation-Aware Graph Convolutional Network for Conversational Emotion Recognition
    Ren, Minjie
    Huang, Xiangdong
    Li, Wenhui
    Song, Dan
    Nie, Weizhi
    IEEE TRANSACTIONS ON MULTIMEDIA, 2022, 24 : 4422 - 4432
  • [29] A cross-modal fusion network based on graph feature learning for multimodal emotion recognition
    Cao Xiaopeng
    Zhang Linying
    Chen Qiuxian
    Ning Hailong
    Dong Yizhuo
    The Journal of China Universities of Posts and Telecommunications, 2024, 31 (06) : 16 - 25
  • [30] PhyMER: Physiological Dataset for Multimodal Emotion Recognition With Personality as a Context
    Pant, Sudarshan
    Yang, Hyung-Jeong
    Lim, Eunchae
    Kim, Soo-Hyung
    Yoo, Seok-Bong
    IEEE ACCESS, 2023, 11 : 107638 - 107656