SAPBERT: Speaker-Aware Pretrained BERT for Emotion Recognition in Conversation

被引:2
|
作者
Lim, Seunguook [1 ]
Kim, Jihie [1 ]
机构
[1] Dongguk Univ Seoul, Dept Artificial Intelligence, 30 Pildong Ro 1 Gil, Seoul 04620, South Korea
关键词
natural language processing; motion recognition in conversation; dialogue modeling; pre-training; hierarchical BERT;
D O I
10.3390/a16010008
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Emotion recognition in conversation (ERC) is receiving more and more attention, as interactions between humans and machines increase in a variety of services such as chat-bot and virtual assistants. As emotional expressions within a conversation can heavily depend on the contextual information of the participating speakers, it is important to capture self-dependency and inter-speaker dynamics. In this study, we propose a new pre-trained model, SAPBERT, that learns to identify speakers in a conversation to capture the speaker-dependent contexts and address the ERC task. SAPBERT is pre-trained with three training objectives including Speaker Classification (SC), Masked Utterance Regression (MUR), and Last Utterance Generation (LUG). We investigate whether our pre-trained speaker-aware model can be leveraged for capturing speaker-dependent contexts for ERC tasks. Experiments show that our proposed approach outperforms baseline models through demonstrating the effectiveness and validity of our method.
引用
收藏
页数:16
相关论文
共 50 条
  • [1] Speaker-Aware Interactive Graph Attention Network for Emotion Recognition in Conversation
    Jia, Zhaohong
    Shi, Yunwei
    Liu, Weifeng
    Huang, Zhenhua
    Sun, Xiao
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2023, 22 (12)
  • [2] Speaker-aware cognitive network with cross-modal attention for multimodal emotion recognition in conversation
    Guo, Lili
    Song, Yikang
    Ding, Shifei
    KNOWLEDGE-BASED SYSTEMS, 2024, 296
  • [3] Speaker-Aware Speech Emotion Recognition by Fusing Amplitude and Phase Information
    Guo, Lili
    Wang, Longbiao
    Dang, Jianwu
    Liu, Zhilei
    Guan, Haotian
    MULTIMEDIA MODELING (MMM 2020), PT I, 2020, 11961 : 14 - 25
  • [4] Speaker-aware Cross-modal Fusion Architecture for Conversational Emotion Recognition
    Zhao, Huan
    Li, Bo
    Zhang, Zixing
    INTERSPEECH 2023, 2023, : 2718 - 2722
  • [5] Global-View and Speaker-Aware Emotion Cause Extraction in Conversations
    An, Jiaming
    Ding, Zixiang
    Li, Ke
    Xia, Rui
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2023, 31 : 3814 - 3823
  • [6] Speaker-Aware Multi-Task Learning for Automatic Speech Recognition
    Pironkov, Gueorgui
    Dupont, Stephane
    Dutoit, Thierry
    2016 23RD INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2016, : 2900 - 2905
  • [7] Speaker-Aware BERT for Multi-Turn Response Selection in Retrieval-Based Chatbots
    Gu, Jia-Chen
    Li, Tianda
    Liu, Quan
    Ling, Zhen-Hua
    Su, Zhiming
    Wei, Si
    Zhu, Xiaodan
    CIKM '20: PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, 2020, : 2041 - 2044
  • [8] Speaker-Aware Linear Discriminant Analysis in Speaker Verification
    Zheng, Naijun
    Wu, Xixin
    Zhong, Jinghua
    Liu, Xunying
    Meng, Helen
    INTERSPEECH 2020, 2020, : 3012 - 3016
  • [9] Speaker-Aware Anti-spoofing
    Liu, Xuechen
    Sahidullah, Md
    Lee, Kong Aik
    Kinnunen, Tomi
    INTERSPEECH 2023, 2023, : 2498 - 2502
  • [10] SPEAKER-AWARE SPEECH-TRANSFORMER
    Fan, Zhiyun
    Li, Jie
    Zhou, Shiyu
    Xu, Bo
    2019 IEEE AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING WORKSHOP (ASRU 2019), 2019, : 222 - 229