A Comparative Study on Pre-Trained Models Based on BERT

被引:0
|
作者
Zhang, Minghua [1 ]
机构
[1] Northeastern Univ, Khoury Coll Comp Sci, Beijing, Peoples R China
关键词
Self-Supervised Learning; PTM; NLP; BERT;
D O I
10.1109/ICNLP60986.2024.10692659
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The introduction of pre-trained models (PTMs) especially Bidirectional Encoder Representation from Transformer (BERT) [1] brought significant improvements in Natural Language Processing (NLP) tasks and demonstrated the power of transfer learning in large language models. The state-of-the-art performance of BERT on eleven NLP tasks inspired many researchers to focus on building variants based on BERT. This survey is going to collect and investigate the NLP-PTMs researches especially the ones motivated by BERT, concentrating on three main tasks: classifications of their research objects and research methods, and an experimental analysis. The collected papers are going to be classified based on different criteria for each task and provide detailed explanations of why certain research is classified into certain type. In the end, based on the investigation, a future direction for the development of PTMs in NLP is suggested.
引用
收藏
页码:326 / 330
页数:5
相关论文
共 50 条
  • [1] μBERT: Mutation Testing using Pre-Trained Language Models
    Degiovanni, Renzo
    Papadakis, Mike
    2022 IEEE 15TH INTERNATIONAL CONFERENCE ON SOFTWARE TESTING, VERIFICATION AND VALIDATION WORKSHOPS (ICSTW 2022), 2022, : 160 - 169
  • [2] Pre-trained language models evaluating themselves - A comparative study
    Koch, Philipp
    Assenmacher, Matthias
    Heumann, Christian
    PROCEEDINGS OF THE THIRD WORKSHOP ON INSIGHTS FROM NEGATIVE RESULTS IN NLP (INSIGHTS 2022), 2022, : 180 - 187
  • [3] Leveraging Pre-trained BERT for Audio Captioning
    Liu, Xubo
    Mei, Xinhao
    Huang, Qiushi
    Sun, Jianyuan
    Zhao, Jinzheng
    Liu, Haohe
    Plumbley, Mark D.
    Kilic, Volkan
    Wang, Wenwu
    2022 30TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO 2022), 2022, : 1145 - 1149
  • [4] Patent classification with pre-trained Bert model
    Kahraman, Selen Yuecesoy
    Durmusoglu, Alptekin
    Dereli, Tuerkay
    JOURNAL OF THE FACULTY OF ENGINEERING AND ARCHITECTURE OF GAZI UNIVERSITY, 2024, 39 (04): : 2485 - 2496
  • [5] A comparative study of pre-trained models in breast ultrasound image segmentation
    Honi, Dhafer G.
    Nsaif, Mohammed
    Szathmary, Laszlo
    Szeghalmy, Szilvia
    2024 IEEE 3RD CONFERENCE ON INFORMATION TECHNOLOGY AND DATA SCIENCE, CITDS 2024, 2024, : 81 - 86
  • [6] Fine-Tuning BERT-Based Pre-Trained Models for Arabic Dependency Parsing
    Al-Ghamdi, Sharefah
    Al-Khalifa, Hend
    Al-Salman, Abdulmalik
    APPLIED SCIENCES-BASEL, 2023, 13 (07):
  • [7] Traceability Transformed: Generating more Accurate Links with Pre-Trained BERT Models
    Lin, Jinfeng
    Liu, Yalin
    Zeng, Qingkai
    Jiang, Meng
    Cleland-Huang, Jane
    2021 IEEE/ACM 43RD INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING (ICSE 2021), 2021, : 324 - 335
  • [8] Research on Chinese Intent Recognition Based on BERT pre-trained model
    Zhang, Pan
    Huang, Li
    2020 5TH INTERNATIONAL CONFERENCE ON MATHEMATICS AND ARTIFICIAL INTELLIGENCE (ICMAI 2020), 2020, : 128 - 132
  • [9] miProBERT: identification of microRNA promoters based on the pre-trained model BERT
    Wang, Xin
    Gao, Xin
    Wang, Guohua
    Li, Dan
    BRIEFINGS IN BIOINFORMATICS, 2023, 24 (03)
  • [10] BERT-siRNA: siRNA target prediction based on BERT pre-trained interpretable model
    Xu, Jiayu
    Xu, Nan
    Xie, Weixin
    Zhao, Chengkui
    Yu, Lei
    Feng, Weixing
    GENE, 2024, 910