Hierarchical Recurrent Attention Network for Response Generation

被引:0
|
作者
Xing, Chen [1 ,2 ,5 ]
Wu, Yu [3 ]
Wu, Wei [4 ]
Huang, Yalou [1 ,2 ]
Zhou, Ming [4 ]
机构
[1] Nankai Univ, Coll Comp & Control Engn, Tianjin, Peoples R China
[2] Nankai Univ, Coll Software, Tianjin, Peoples R China
[3] Beihang Univ, State Key Lab Software Dev Environm, Beijing, Peoples R China
[4] Microsoft Res, Beijing, Peoples R China
[5] Microsoft Res Asia, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We study multi-turn response generation in chatbots where a response is generated according to a conversation context. Existing work has modeled the hierarchy of the context, but does not pay enough attention to the fact that words and utterances in the context are differentially important. As a result, they may lose important information in context and generate irrelevant responses. We propose a hierarchical recurrent attention network (BRAN) to model both the hierarchy and the importance variance in a unified framework. In HRAN, a hierarchical attention mechanism attends to important parts within and among utterances with word level attention and utterance level attention respectively. Empirical studies on both automatic evaluation and human judgment show that HRAN can significantly outperform state-of-the-art models for context based response generation.
引用
收藏
页码:5610 / 5617
页数:8
相关论文
共 50 条
  • [21] Hierarchical Attention Network for Image Captioning
    Wang, Weixuan
    Chen, Zhihong
    Hu, Haifeng
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 8957 - 8964
  • [22] Hierarchical Attention Network for Action Segmentation
    Gammulle, Harshala
    Denman, Simon
    Sridharan, Sridha
    Fookes, Clinton
    PATTERN RECOGNITION LETTERS, 2020, 131 : 442 - 448
  • [23] Semantically Conditioned Dialog Response Generation via Hierarchical Disentangled Self-Attention
    Chen, Wenhu
    Chen, Jianshu
    Qin, Pengda
    Yan, Xifeng
    Wang, William Yang
    57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 3696 - 3709
  • [24] Hierarchical Recurrent Attention Networks for Structured Online Maps
    Homayounfar, Namdar
    Ma, Wei-Chiu
    Lakshmikanth, Shrinidhi Kowshika
    Urtasun, Raquel
    2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 3417 - 3426
  • [25] A Recurrent Attention Network for Judgment Prediction
    Yang, Ze
    Wang, Pengfei
    Zhang, Lei
    Shou, Linjun
    Xu, Wenwen
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2019: TEXT AND TIME SERIES, PT IV, 2019, 11730 : 253 - 266
  • [26] A Hierarchical Attention Headline Generation Model with a Filter
    Xie, Jiehang
    Wang, Xiaoming
    Wang, Xinyan
    Pang, Guangyao
    ADVANCED MULTIMEDIA AND UBIQUITOUS ENGINEERING, 2020, 590 : 332 - 338
  • [27] Threat intelligence ATT&CK extraction based on the attention transformer hierarchical recurrent neural network
    Liu, Chenjing
    Wang, Junfeng
    Chen, Xiangru
    APPLIED SOFT COMPUTING, 2022, 122
  • [28] HAN-ReGRU: hierarchical attention network with residual gated recurrent unit for emotion recognition in conversation
    Hui Ma
    Jian Wang
    Lingfei Qian
    Hongfei Lin
    Neural Computing and Applications, 2021, 33 : 2685 - 2703
  • [29] HAN-ReGRU: hierarchical attention network with residual gated recurrent unit for emotion recognition in conversation
    Ma, Hui
    Wang, Jian
    Qian, Lingfei
    Lin, Hongfei
    NEURAL COMPUTING & APPLICATIONS, 2021, 33 (07): : 2685 - 2703
  • [30] Threat intelligence ATT&CK extraction based on the attention transformer hierarchical recurrent neural network
    Liu, Chenjing
    Wang, Junfeng
    Chen, Xiangru
    Applied Soft Computing, 2022, 122