A Comparative Study on Pre-Trained Models Based on BERT

被引:0
|
作者
Zhang, Minghua [1 ]
机构
[1] Northeastern Univ, Khoury Coll Comp Sci, Beijing, Peoples R China
关键词
Self-Supervised Learning; PTM; NLP; BERT;
D O I
10.1109/ICNLP60986.2024.10692659
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The introduction of pre-trained models (PTMs) especially Bidirectional Encoder Representation from Transformer (BERT) [1] brought significant improvements in Natural Language Processing (NLP) tasks and demonstrated the power of transfer learning in large language models. The state-of-the-art performance of BERT on eleven NLP tasks inspired many researchers to focus on building variants based on BERT. This survey is going to collect and investigate the NLP-PTMs researches especially the ones motivated by BERT, concentrating on three main tasks: classifications of their research objects and research methods, and an experimental analysis. The collected papers are going to be classified based on different criteria for each task and provide detailed explanations of why certain research is classified into certain type. In the end, based on the investigation, a future direction for the development of PTMs in NLP is suggested.
引用
收藏
页码:326 / 330
页数:5
相关论文
共 50 条
  • [21] MF-BERT: Multimodal Fusion in Pre-Trained BERT for Sentiment Analysis
    He, Jiaxuan
    Hu, Haifeng
    IEEE SIGNAL PROCESSING LETTERS, 2022, 29 : 454 - 458
  • [22] Chinese Grammatical Correction Using BERT-based Pre-trained Model
    Wang, Hongfei
    Kurosawa, Michiki
    Katsumatat, Satoru
    Komachi, Mamoru
    1ST CONFERENCE OF THE ASIA-PACIFIC CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 10TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (AACL-IJCNLP 2020), 2020, : 163 - 168
  • [23] Text clustering based on pre-trained models and autoencoders
    Xu, Qiang
    Gu, Hao
    Ji, ShengWei
    FRONTIERS IN COMPUTATIONAL NEUROSCIENCE, 2024, 17
  • [24] Pre-trained Convolutional Networks and Generative Statistical Models: A Comparative Study in Large Datasets
    Michael, John
    Teixeira, Luis F.
    PATTERN RECOGNITION AND IMAGE ANALYSIS (IBPRIA 2017), 2017, 10255 : 69 - 75
  • [25] Learning to Remove: Towards Isotropic Pre-trained BERT Embedding
    Liang, Yuxin
    Cao, Rui
    Zheng, Jie
    Ren, Jie
    Gao, Ling
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2021, PT V, 2021, 12895 : 448 - 459
  • [26] A pre-trained BERT for Korean medical natural language processing
    Yoojoong Kim
    Jong-Ho Kim
    Jeong Moon Lee
    Moon Joung Jang
    Yun Jin Yum
    Seongtae Kim
    Unsub Shin
    Young-Min Kim
    Hyung Joon Joo
    Sanghoun Song
    Scientific Reports, 12
  • [27] A pre-trained BERT for Korean medical natural language processing
    Kim, Yoojoong
    Kim, Jong-Ho
    Lee, Jeong Moon
    Jang, Moon Joung
    Yum, Yun Jin
    Kim, Seongtae
    Shin, Unsub
    Kim, Young-Min
    Joo, Hyung Joon
    Song, Sanghoun
    SCIENTIFIC REPORTS, 2022, 12 (01)
  • [28] Transfer Learning from Pre-trained BERT for Pronoun Resolution
    Bao, Xingce
    Qiao, Qianqian
    GENDER BIAS IN NATURAL LANGUAGE PROCESSING (GEBNLP 2019), 2019, : 82 - 88
  • [29] BERT is to NLP what AlexNet is to CV: Can Pre-Trained Language Models Identify Analogies?
    Ushio, Asahi
    Espinosa-Anke, Luis
    Schockaert, Steven
    Camacho-Collados, Jose
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, 2021, : 3609 - 3624
  • [30] Is BERT a Cross-Disciplinary Knowledge Learner? A Surprising Finding of Pre-trained Models' Transferability
    Kao, Wei-Tsung
    Lee, Hung-yi
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 2195 - 2208