PGIF: A Personality-Guided Iterative Feedback Graph Network for Multimodal Conversational Emotion Recognition

被引:0
|
作者
Xie, Yunhe [1 ]
Mao, Rui [2 ]
机构
[1] Harbin Inst Technol, Fac Comp, Harbin 150001, Peoples R China
[2] Nanyang Technol Univ, Coll Comp & Data Sci, Singapore 639798, Singapore
关键词
Emotion recognition; Iterative methods; Pragmatics; Feature extraction; Vectors; Semantics; Oral communication; Long short term memory; Correlation; Context modeling; interlocutor-induced pragmatic variation; iterative feedback fusion mechanism; multimodal conversational emotion recognition (MCER);
D O I
10.1109/TCSS.2024.3523322
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Multimodal emotion recognition in conversation (MERC) aims to identify emotions in target utterances from multimodal records, drawing significant attention for its value in conversational artificial intelligence. While early research focuses on exploring conversational context, recent efforts emphasize integrating multimodal cues. Existing methods focus on modeling the impact of conversational context on emotion recognition while neglecting the role of the speaker's personality factors. Furthermore, these approaches often suffer from inefficiencies in information transfer due to full-utterance connectivity and fail to leverage multiple fusion modes for complementary benefits. To address these issues, we propose a personality-guided iterative feedback graph network (PGIF) for MERC. PGIF incorporates personality information as a specialized modality to enhance the feature space for emotional inference. We utilize a graph network to model information flow, integrating interlocutor-aware contextual information by considering interlocutor dependencies between utterances. Additionally, we employ a dialogue discourse parser to directly model semantic relationships between utterances. Our iterative feedback fusion mechanism explicitly simulates emotional interactions between feature-level and decision-level modalities, improving inference without ground truth labels through iterative refinement. PGIF demonstrates improvements of 1.94% and 1.42% over state-of-the-art methods on the IEMOCAP and MELD datasets, respectively. Ablation studies further validate the effectiveness of PGIF's mechanisms, while its manipulation of input features and global fusion strategies ensures compatibility with existing approaches.
引用
收藏
页数:13
相关论文
共 50 条
  • [1] Topics Guided Multimodal Fusion Network for Conversational Emotion Recognition
    Yuan, Peicong
    Cai, Guoyong
    Chen, Ming
    Tang, Xiaolv
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT III, ICIC 2024, 2024, 14877 : 250 - 262
  • [2] Multi-loop graph convolutional network for multimodal conversational emotion recognition
    Ren, Minjie
    Huang, Xiangdong
    Li, Wenhui
    Liu, Jing
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2023, 94
  • [3] Directed Acyclic Graph Network for Conversational Emotion Recognition
    Shen, Weizhou
    Wu, Siyue
    Yang, Yunyi
    Quan, Xiaojun
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 1 (ACL-IJCNLP 2021), 2021, : 1551 - 1560
  • [4] MALN: Multimodal Adversarial Learning Network for Conversational Emotion Recognition
    Ren, Minjie
    Huang, Xiangdong
    Liu, Jing
    Liu, Ming
    Li, Xuanya
    Liu, An-An
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2023, 33 (11) : 6965 - 6980
  • [5] SDR-GNN: Spectral Domain Reconstruction Graph Neural Network for incomplete multimodal learning in conversational emotion recognition
    Fu, Fangze
    Ai, Wei
    Yang, Fan
    Shou, Yuntao
    Meng, Tao
    Li, Keqin
    KNOWLEDGE-BASED SYSTEMS, 2025, 309
  • [6] Adaptive Graph Learning for Multimodal Conversational Emotion Detection
    Tu, Geng
    Xie, Tian
    Liang, Bin
    Wang, Hongpeng
    Xu, Ruifeng
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 17, 2024, : 19089 - 19097
  • [7] A graph neural network with context filtering and feature correction for conversational emotion recognition
    Gan, Chenquan
    Zheng, Jiahao
    Zhu, Qingyi
    Jain, Deepak Kumar
    Štruc, Vitomir
    Information Sciences, 2024, 658
  • [8] A graph neural network with context filtering and feature correction for conversational emotion recognition
    Gan, Chenquan
    Zheng, Jiahao
    Zhu, Qingyi
    Kumar, Deepak
    Struc, Vitomir
    INFORMATION SCIENCES, 2024, 658
  • [9] Hierarchical heterogeneous graph network based multimodal emotion recognition in conversation
    Peng, Junyin
    Tang, Hong
    Zheng, Wenbin
    MULTIMEDIA SYSTEMS, 2025, 31 (01)
  • [10] DGSNet: Dual Graph Structure Network for Emotion Recognition in Multimodal Conversations
    Tang, Shimin
    Wang, Changjian
    Tian, Fengyu
    Xu, Kele
    Xu, Minpeng
    2023 IEEE 35TH INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE, ICTAI, 2023, : 78 - 85