DUMA: Reading Comprehension With Transposition Thinking

被引:22
|
作者
Zhu, Pengfei [1 ]
Zhang, Zhuosheng [1 ]
Zhao, Hai [1 ]
Li, Xiaoguang [2 ]
机构
[1] Shanghai Jiao Tong Univ, Dept Comp Sci & Engn, Key Lab, Shanghai Educ Commiss Intelligent Interact & Cogn, Shanghai 200240, Peoples R China
[2] Huawei Noahs Ark Lab, Shenzhen 518129, Peoples R China
基金
中国国家自然科学基金;
关键词
Task analysis; Training; Transformers; Speech processing; Bit error rate; Bidirectional control; Context modeling; Attention network; machine reading comprehension; pre-trained language model; MODEL;
D O I
10.1109/TASLP.2021.3138683
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Multi-choice Machine Reading Comprehension (MRC) requires models to decide the correct answer from a set of answer options when given a passage and a question. Thus, in addition to a powerful Pre-trained Language Model (PrLM) as an encoder, multi-choice MRC especially relies on a matching network design that is supposed to effectively capture the relationships among the triplet of passage, question, and answers. While the newer and more powerful PrLMs have shown their strengths even without the support from a matching network, we propose a new DUal Multi-head Co-Attention (DUMA) model. It is inspired by the human transposition thinking process solving the multi-choice MRC problem by considering each other's focus from the standpoint of passage and question. The proposed DUMA has been shown to be effective and is capable of generally promoting PrLMs. Our proposed method is evaluated on two benchmark multi-choice MRC tasks, DREAM, and RACE. Our results show that in terms of powerful PrLMs, DUMA can further boost the models to obtain higher performance.
引用
收藏
页码:269 / 279
页数:11
相关论文
共 50 条