A Comparative Study on Pre-Trained Models Based on BERT

被引:0
|
作者
Zhang, Minghua [1 ]
机构
[1] Northeastern Univ, Khoury Coll Comp Sci, Beijing, Peoples R China
关键词
Self-Supervised Learning; PTM; NLP; BERT;
D O I
10.1109/ICNLP60986.2024.10692659
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The introduction of pre-trained models (PTMs) especially Bidirectional Encoder Representation from Transformer (BERT) [1] brought significant improvements in Natural Language Processing (NLP) tasks and demonstrated the power of transfer learning in large language models. The state-of-the-art performance of BERT on eleven NLP tasks inspired many researchers to focus on building variants based on BERT. This survey is going to collect and investigate the NLP-PTMs researches especially the ones motivated by BERT, concentrating on three main tasks: classifications of their research objects and research methods, and an experimental analysis. The collected papers are going to be classified based on different criteria for each task and provide detailed explanations of why certain research is classified into certain type. In the end, based on the investigation, a future direction for the development of PTMs in NLP is suggested.
引用
收藏
页码:326 / 330
页数:5
相关论文
共 50 条
  • [31] GT-Finder: Classify the family of glucose transporters with pre-trained BERT language models
    Shah, Syed Muazzam Ali
    Taju, Semmy Wellem
    Quang-Thai Ho
    Trinh-Trung-Duong Nguyen
    Ou, Yu-Yen
    COMPUTERS IN BIOLOGY AND MEDICINE, 2021, 131
  • [32] Refining Pre-Trained Motion Models
    Sun, Xinglong
    Harley, Adam W.
    Guibas, Leonidas J.
    2024 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA 2024, 2024, : 4932 - 4938
  • [33] Comparative Analysis of Pre-trained CNN Models for Neurobiological Disease Classification
    Munir, Uwasila Binte
    Al Mamun, Shamim
    APPLIED INTELLIGENCE AND INFORMATICS, AII 2023, 2024, 2065 : 339 - 354
  • [34] Efficiently Robustify Pre-Trained Models
    Jain, Nishant
    Behl, Harkirat
    Rawat, Yogesh Singh
    Vineet, Vibhav
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION, ICCV, 2023, : 5482 - 5492
  • [35] Pre-trained Models for Sonar Images
    Valdenegro-Toro, Matias
    Preciado-Grijalva, Alan
    Wehbe, Bilal
    OCEANS 2021: SAN DIEGO - PORTO, 2021,
  • [36] Inference-based No-Learning Approach on Pre-trained BERT Model Retrieval
    Pham, Huu-Long
    Mibayashi, Ryota
    Yamamoto, Takehiro
    Kato, Makoto P.
    Yamamoto, Yusuke
    Shoji, Yoshiyuki
    Ohshima, Hiroaki
    2024 IEEE INTERNATIONAL CONFERENCE ON BIG DATA AND SMART COMPUTING, IEEE BIGCOMP 2024, 2024, : 234 - 241
  • [37] Pre-Trained Language Models and Their Applications
    Wang, Haifeng
    Li, Jiwei
    Wu, Hua
    Hovy, Eduard
    Sun, Yu
    ENGINEERING, 2023, 25 : 51 - 65
  • [38] A Data Cartography based MixUp for Pre-trained Language Models
    Park, Seo Yeon
    Caragea, Cornelia
    NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 4244 - 4250
  • [39] Deep learning approaches for online signature authentication: a comparative study of pre-trained CNN models
    Swamy, M. Ranga
    Vijayalakshmi, P.
    Rajendran, V
    ENGINEERING RESEARCH EXPRESS, 2025, 7 (01):
  • [40] Measuring Fairness with Biased Rulers: A Comparative Study on Bias Metrics for Pre-trained Language Models
    Delobelle, Pieter
    Tokpo, Ewoenam Kwaku
    Calders, Toon
    Berendt, Bettina
    NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 1693 - 1706