PGIF: A Personality-Guided Iterative Feedback Graph Network for Multimodal Conversational Emotion Recognition

被引:0
|
作者
Xie, Yunhe [1 ]
Mao, Rui [2 ]
机构
[1] Harbin Inst Technol, Fac Comp, Harbin 150001, Peoples R China
[2] Nanyang Technol Univ, Coll Comp & Data Sci, Singapore 639798, Singapore
关键词
Emotion recognition; Iterative methods; Pragmatics; Feature extraction; Vectors; Semantics; Oral communication; Long short term memory; Correlation; Context modeling; interlocutor-induced pragmatic variation; iterative feedback fusion mechanism; multimodal conversational emotion recognition (MCER);
D O I
10.1109/TCSS.2024.3523322
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Multimodal emotion recognition in conversation (MERC) aims to identify emotions in target utterances from multimodal records, drawing significant attention for its value in conversational artificial intelligence. While early research focuses on exploring conversational context, recent efforts emphasize integrating multimodal cues. Existing methods focus on modeling the impact of conversational context on emotion recognition while neglecting the role of the speaker's personality factors. Furthermore, these approaches often suffer from inefficiencies in information transfer due to full-utterance connectivity and fail to leverage multiple fusion modes for complementary benefits. To address these issues, we propose a personality-guided iterative feedback graph network (PGIF) for MERC. PGIF incorporates personality information as a specialized modality to enhance the feature space for emotional inference. We utilize a graph network to model information flow, integrating interlocutor-aware contextual information by considering interlocutor dependencies between utterances. Additionally, we employ a dialogue discourse parser to directly model semantic relationships between utterances. Our iterative feedback fusion mechanism explicitly simulates emotional interactions between feature-level and decision-level modalities, improving inference without ground truth labels through iterative refinement. PGIF demonstrates improvements of 1.94% and 1.42% over state-of-the-art methods on the IEMOCAP and MELD datasets, respectively. Ablation studies further validate the effectiveness of PGIF's mechanisms, while its manipulation of input features and global fusion strategies ensures compatibility with existing approaches.
引用
收藏
页数:13
相关论文
共 50 条
  • [41] A Multi-Level Alignment and Cross-Modal Unified Semantic Graph Refinement Network for Conversational Emotion Recognition
    Zhang, Xiaoheng
    Cui, Weigang
    Hu, Bin
    Li, Yang
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2024, 15 (03) : 1553 - 1566
  • [42] Masked Graph Learning With Recurrent Alignment for Multimodal Emotion Recognition in Conversation
    Meng, Tao
    Zhang, Fuchen
    Shou, Yuntao
    Shao, Hongen
    Ai, Wei
    Li, Keqin
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2024, 32 : 4298 - 4312
  • [43] MULTIMODAL EMOTION RECOGNITION WITH CAPSULE GRAPH CONVOLUTIONAL BASED REPRESENTATION FUSION
    Liu, Jiaxing
    Chen, Sen
    Wang, Longbiao
    Liu, Zhilei
    Fu, Yahui
    Guo, Lili
    Dang, Jianwu
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 6339 - 6343
  • [44] Meaningful Multimodal Emotion Recognition Based on Capsule Graph Transformer Architecture
    Filali, Hajar
    Boulealam, Chafik
    El Fazazy, Khalid
    Mahraz, Adnane Mohamed
    Tairi, Hamid
    Riffi, Jamal
    INFORMATION, 2025, 16 (01)
  • [45] Progressive graph convolution network for EEG emotion recognition
    Zhou, Yijin
    Li, Fu
    Li, Yang
    Ji, Youshuo
    Shi, Guangming
    Zheng, Wenming
    Zhang, Lijian
    Chen, Yuanfang
    Cheng, Rui
    NEUROCOMPUTING, 2023, 544
  • [46] Causal Graph Convolutional Neural Network for Emotion Recognition
    Kong, Wanzeng
    Qiu, Min
    Li, Menghang
    Jin, Xuanyu
    Zhu, Li
    IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2023, 15 (04) : 1686 - 1693
  • [47] Graph Reasoning-Based Emotion Recognition Network
    Gao, Qinquan
    Zeng, Hanxin
    Li, Gen
    Tong, Tong
    IEEE ACCESS, 2021, 9 : 6488 - 6497
  • [48] Conversational Emotion Recognition Using Self-Attention Mechanisms and Graph Neural Networks
    Lian, Zheng
    Tao, Jianhua
    Liu, Bin
    Huang, Jian
    Yang, Zhanlei
    Li, Rongjun
    INTERSPEECH 2020, 2020, : 2347 - 2351
  • [49] A Multimodal Low Complexity Neural Network Approach for Emotion Recognition
    Aguinaga, Adrian Rodriguez
    Ramirez, Margarita Ramirez
    Soto, Maria del Consuelo Salgado
    Cisnero, Maria de los Angeles Quezada
    HUMAN BEHAVIOR AND EMERGING TECHNOLOGIES, 2024, 2024
  • [50] Multimodal Emotion Recognition Based on Ensemble Convolutional Neural Network
    Huang, Haiping
    Hu, Zhenchao
    Wang, Wenming
    Wu, Min
    IEEE ACCESS, 2020, 8 : 3265 - 3271