Contextualized Embeddings based Transformer Encoder for Sentence Similarity Modeling in Answer Selection Task

被引:0
|
作者
Laskar, Md Tahmid Rahman [1 ,3 ]
Huang, Jimmy [2 ,3 ]
Hoque, Enamul [2 ]
机构
[1] York Univ, Dept Elect Engn & Comp Sci, Toronto, ON, Canada
[2] York Univ, Sch Informat Technol, Toronto, ON, Canada
[3] York Univ, Informat Retrieval & Knowledge Management Res Lab, Toronto, ON, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
Answer Selection; Transformer Encoder; Contextualized Embeddings; ELMo; BERT; RoBERTa; Deep Learning;
D O I
暂无
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Word embeddings that consider context have attracted great attention for various natural language processing tasks in recent years. In this paper, we utilize contextualized word embeddings with the transformer encoder for sentence similarity modeling in the answer selection task. We present two different approaches (feature-based and fine-tuning-based) for answer selection. In the feature-based approach, we utilize two types of contextualized embeddings, namely the Embeddings from Language Models (ELMo) and the Bidirectional Encoder Representations from Transformers (BERT) and integrate each of them with the transformer encoder. We find that integrating these contextual embeddings with the transformer encoder is effective to improve the performance of sentence similarity modeling. In the second approach, we fine-tune two pre-trained transformer encoder models for the answer selection task. Based on our experiments on six datasets, we find that the fine-tuning approach outperforms the feature-based approach on all of them. Among our fine-tuning-based models, the Robustly Optimized BERT Pretraining Approach (RoBERTa) model results in new state-of-the-art performance across five datasets.
引用
收藏
页码:5505 / 5514
页数:10
相关论文
共 50 条
  • [1] Question Classification Using Universal Sentence Encoder and Deep Contextualized Transformer
    Arif, Najam
    Latif, Seemab
    Latif, Rabia
    2021 14TH INTERNATIONAL CONFERENCE ON DEVELOPMENTS IN ESYSTEMS ENGINEERING (DESE), 2021, : 206 - 211
  • [2] Question-Answer Sentence Graph for Joint Modeling Answer Selection
    Iyer, Roshni G.
    Thuy Vu
    Moschitti, Alessandro
    Sun, Yizhou
    17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 968 - 979
  • [3] Modeling Context in Answer Sentence Selection Systems on a Latency Budget
    Han, Rujun
    Soldaini, Luca
    Moschitti, Alessandro
    16TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2021), 2021, : 3005 - 3010
  • [4] Context-Aware Transformer Pre-Training for Answer Sentence Selection
    Di Liello, Luca
    Garg, Siddhant
    Moschitti, Alessandro
    61ST CONFERENCE OF THE THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 2, 2023, : 458 - 468
  • [5] Entity-aware answer sentence selection for question answering with transformer-based language models
    Zahra Abbasiantaeb
    Saeedeh Momtazi
    Journal of Intelligent Information Systems, 2022, 59 : 755 - 777
  • [6] Entity-aware answer sentence selection for question answering with transformer-based language models
    Abbasiantaeb, Zahra
    Momtazi, Saeedeh
    JOURNAL OF INTELLIGENT INFORMATION SYSTEMS, 2022, 59 (03) : 755 - 777
  • [7] Entity-aware answer sentence selection for question answering with transformer-based language models
    Abbasiantaeb, Zahra
    Momtazi, Saeedeh
    Journal of Intelligent Information Systems, 2022, 59 (03): : 755 - 777
  • [8] Comparing RNN and Transformer Context Representations in the Czech Answer Selection Task
    Medved, Marek
    Sabol, Radoslav
    Horak, Ales
    ICAART: PROCEEDINGS OF THE 14TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE - VOL 3, 2022, : 388 - 394
  • [9] CollabAS2: Enhancing Arabic Answer Sentence Selection Using Transformer-Based Collaborative Models
    Aouichat, Asma
    Guessoum, Ahmed
    ARABIAN JOURNAL FOR SCIENCE AND ENGINEERING, 2024,
  • [10] TANDA: Transfer and Adapt Pre-Trained Transformer Models for Answer Sentence Selection
    Garg, Siddhant
    Vu, Thuy
    Moschitti, Alessandro
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 7780 - 7788