A Comparative Study on Pre-Trained Models Based on BERT

被引:0
|
作者
Zhang, Minghua [1 ]
机构
[1] Northeastern Univ, Khoury Coll Comp Sci, Beijing, Peoples R China
关键词
Self-Supervised Learning; PTM; NLP; BERT;
D O I
10.1109/ICNLP60986.2024.10692659
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The introduction of pre-trained models (PTMs) especially Bidirectional Encoder Representation from Transformer (BERT) [1] brought significant improvements in Natural Language Processing (NLP) tasks and demonstrated the power of transfer learning in large language models. The state-of-the-art performance of BERT on eleven NLP tasks inspired many researchers to focus on building variants based on BERT. This survey is going to collect and investigate the NLP-PTMs researches especially the ones motivated by BERT, concentrating on three main tasks: classifications of their research objects and research methods, and an experimental analysis. The collected papers are going to be classified based on different criteria for each task and provide detailed explanations of why certain research is classified into certain type. In the end, based on the investigation, a future direction for the development of PTMs in NLP is suggested.
引用
收藏
页码:326 / 330
页数:5
相关论文
共 50 条
  • [41] A Comparative Study of Different Pre-Trained DeepLearning Models and Custom CNN for Pancreatic Tumor Detection
    Zavalsiz, Muhammed Talha
    Alhajj, Sleiman
    Sailunaz, Kashfia
    Ozyer, Tansel
    Alhajj, Reda
    INTERNATIONAL ARAB JOURNAL OF INFORMATION TECHNOLOGY, 2023, 20 (3A) : 515 - 526
  • [42] Pre-trained transformer-based language models for Sundanese
    Wilson Wongso
    Henry Lucky
    Derwin Suhartono
    Journal of Big Data, 9
  • [43] An Extensive Study on Pre-trained Models for Program Understanding and Generation
    Zeng, Zhengran
    Ta, Hanzhuo
    Zhang, Haotian
    Li, Jing
    Zhang, Yuqun
    Zhang, Lingming
    PROCEEDINGS OF THE 31ST ACM SIGSOFT INTERNATIONAL SYMPOSIUM ON SOFTWARE TESTING AND ANALYSIS, ISSTA 2022, 2022, : 39 - 51
  • [44] TinyMIM: An Empirical Study of Distilling MIM Pre-trained Models
    Ren, Sucheng
    Wei, Fangyun
    Zhang, Zhang
    Hu, Han
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 3687 - 3697
  • [45] A Study of Pre-trained Language Models in Natural Language Processing
    Duan, Jiajia
    Zhao, Hui
    Zhou, Qian
    Qiu, Meikang
    Liu, Meiqin
    2020 IEEE INTERNATIONAL CONFERENCE ON SMART CLOUD (SMARTCLOUD 2020), 2020, : 116 - 121
  • [46] Pre-trained CNNs Models for Content based Image Retrieval
    Ahmed, Ali
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2021, 12 (07) : 200 - 206
  • [47] Pre-trained transformer-based language models for Sundanese
    Wongso, Wilson
    Lucky, Henry
    Suhartono, Derwin
    JOURNAL OF BIG DATA, 2022, 9 (01)
  • [48] BERT for Sentiment Analysis: Pre-trained and Fine-Tuned Alternatives
    Souza, Frederico Dias
    de Oliveira e Souza Filho, Joao Baptista
    COMPUTATIONAL PROCESSING OF THE PORTUGUESE LANGUAGE, PROPOR 2022, 2022, 13208 : 209 - 218
  • [49] Syntax-BERT: Improving Pre-trained Transformers with Syntax Trees
    Bai, Jiangang
    Wang, Yujing
    Chen, Yiren
    Yang, Yaming
    Bai, Jing
    Yu, Jing
    Tong, Yunhai
    16TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2021), 2021, : 3011 - 3020
  • [50] Lattice-BERT: Leveraging Multi-Granularity Representations in Chinese Pre-trained Language Models
    Lai, Yuxuan
    Liu, Yijia
    Feng, Yansong
    Huang, Songfang
    Zhao, Dongyan
    2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), 2021, : 1716 - 1731