HiBERT: Detecting the illogical patterns with hierarchical BERT for multi-turn dialogue reasoning

被引:1
|
作者
Wang, Xu [1 ,2 ,3 ]
Zhang, Hainan [4 ]
Zhao, Shuai [5 ,6 ]
Chen, Hongshen [4 ]
Cheng, Bo [5 ]
Ding, Zhuoye [4 ]
Xu, Sulong [4 ]
Yan, Weipeng [4 ]
Lan, Yanyan [7 ]
机构
[1] Hebei Univ Technol, Sch Artificial Intelligence, Tianjin 300401, Peoples R China
[2] Hebei Prov Key Lab Big Data Comp, Tianjin 300401, Peoples R China
[3] Hebei Engn Res Ctr Data Driven Ind Intelligent, Tianjin 300401, Peoples R China
[4] JD Com, Data Sci Lab, Beijing, Peoples R China
[5] Beijing Univ Posts & Telecommun, State Key Lab Networking & Switching Technol, Beijing, Peoples R China
[6] Guangxi Key Lab Cryptog & Informat Secur, Guilin, Peoples R China
[7] Tsinghua Univ, Inst AI Ind Res, Beijing, Peoples R China
基金
中国国家自然科学基金; 国家重点研发计划;
关键词
Dialogue reasoning; Neural networks; BERT; Multi -turn dialogue;
D O I
10.1016/j.neucom.2022.12.038
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Dialogue reasoning is a new task beyond the traditional dialogue system, because it requires recognizing not only semantic relevance but also logical consistency between the candidate response and the dia-logue history. Like "all happy families are happy alike, all unhappy families are unhappy in their own way", various illogical patterns exist in the data. For example, some candidate responses use many sim-ilar words but with contradicted meanings with history; while some candidates may employ totally dif-ferent words but convey consistent meanings. Therefore, an ideal dialogue reasoning model should gather clues from both coarse-grained utterance-level and fine-grained word-level to determine the log-ical relation between candidates and the dialogue history. However, traditional models mainly rely on the widely used BERT to read all the history and candidates word by word but ignore the utterance-level sig-nals, which cannot well capture various illogical patterns in this task. To tackle this problem, we propose a novel Hierarchical BERT (HiBERT) to recognize both utterance-level and word-level illogical patterns in this paper. Specifically, BERT is firstly utilized to encode the dialogue history and each candidate response as the contextualized representation. Secondly, hierarchical reasoning architecture is conducted with this contextualized representation to obtain the word-level and the utterance-level attention distributions, respectively. In detail, we utilize the word-grained attention mechanism to obtain the word-level repre-sentation, and propose two different types of attention function, i.e, hard attention and soft attention, to obtain the utterance-grained representation. Finally, we fuse both the word-grained representation and the utterance-grained representation to calculate the logical ranking scores for the given candidate. Experimental results on two public dialogue datasets show that our model obtains higher ranking mea-sures than the widely used BERT model, validating the effectiveness of hierarchical reading of HiBERT. Further analysis on the impact of context length and attention weights shows that the HiBERT actually has the ability to recognize different illogical patterns.(c) 2022 Elsevier B.V. All rights reserved.
引用
收藏
页码:167 / 177
页数:11
相关论文
共 50 条
  • [1] MuTual: A Dataset for Multi-Turn Dialogue Reasoning
    Cui, Leyang
    Wu, Yu
    Liu, Shujie
    Zhang, Yue
    Zhou, Ming
    58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020), 2020, : 1406 - 1416
  • [2] Debiasing Counterfactual Context With Causal Inference for Multi-Turn Dialogue Reasoning
    Wang, Xu
    Zhang, Hainan
    Zhao, Shuai
    Chen, Hongshen
    Ding, Zhuoye
    Wan, Zhiguo
    Cheng, Bo
    Lan, Yanyan
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2024, 32 : 1125 - 1132
  • [3] HKA: A HIERARCHICAL KNOWLEDGE ATTENTION MECHANISM FOR MULTI-TURN DIALOGUE SYSTEM
    Song, Jian
    Zhang, Kailai
    Zhou, Xuesi
    Wu, Ji
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 3512 - 3516
  • [4] FCM: A Fine-grained Comparison Model for Multi-turn Dialogue Reasoning
    Wang, Xu
    Zhang, Hainan
    Zhao, Shuai
    Zou, Yanyan
    Chen, Hongshen
    Ding, Zhuoye
    Cheng, Bo
    Lan, Yanyan
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 4284 - 4293
  • [5] Multi-turn Dialogue Generation Model with Dialogue Structure
    Jiang X.-T.
    Wang Z.-Q.
    Li S.-S.
    Zhou G.-D.
    Ruan Jian Xue Bao/Journal of Software, 2022, 33 (11): : 4239 - 4250
  • [6] Multi-turn dialogue model based on the improved hierarchical recurrent attention network
    Miao J.
    Wu J.
    International Journal for Engineering Modelling, 2021, 34 (02) : 17 - 29
  • [7] An Efficient Approach based on BERT and Recurrent Neural Network for Multi-turn Spoken Dialogue Understanding
    Xiong, Weixing
    Ma, Li
    Liao, Hongtao
    ICAART: PROCEEDINGS OF THE 12TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE, VOL 2, 2020, : 793 - 800
  • [8] HSAN: A HIERARCHICAL SELF-ATTENTION NETWORK FOR MULTI-TURN DIALOGUE GENERATION
    Kong, Yawei
    Zhang, Lu
    Ma, Can
    Cao, Cong
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 7433 - 7437
  • [9] Model of Multi-turn Dialogue in Emotional Chatbot
    Kao, Chien-Hao
    Chen, Chih-Chieh
    Tsai, Yu-Tza
    2019 INTERNATIONAL CONFERENCE ON TECHNOLOGIES AND APPLICATIONS OF ARTIFICIAL INTELLIGENCE (TAAI), 2019,
  • [10] A Context-Aware Hierarchical BERT Fusion Network for Multi-turn Dialog Act Detection
    Wu, Ting-Wei
    Su, Ruolin
    Juang, Biing-Hwang
    INTERSPEECH 2021, 2021, : 1239 - 1243