SCIBERT: A Pretrained Language Model for Scientific Text

被引:0
|
作者
Beltagy, Iz [1 ]
Lo, Kyle [1 ]
Cohan, Arman [1 ]
机构
[1] Allen Inst Artificial Intelligence, Seattle, WA 98103 USA
关键词
CORPUS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Obtaining large-scale annotated data for NLP tasks in the scientific domain is challenging and expensive. We release SCIBERT, a pretrained language model based on BERT (Devlin et al., 2019) to address the lack of highquality, large-scale labeled scientific data. SCIBERT leverages unsupervised pretraining on a large multi-domain corpus of scientific publications to improve performance on downstream scientific NLP tasks. We evaluate on a suite of tasks including sequence tagging, sentence classification and dependency parsing, with datasets from a variety of scientific domains. We demonstrate statistically significant improvements over BERT and achieve new state-of-the-art results on several of these tasks.
引用
收藏
页码:3615 / 3620
页数:6
相关论文
共 50 条
  • [1] Pretrained Language Models for Text Generation: A Survey
    Li, Junyi
    Tang, Tianyi
    Zhao, Wayne Xin
    Wen, Ji-Rong
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 4492 - 4499
  • [2] Unsupervised Graph-Text Mutual Conversion with a Unified Pretrained Language Model
    Xu, Yi
    Sheng, Shuqian
    Qi, Jiexing
    Fu, Luoyi
    Lin, Zhouhan
    Wang, Xinbing
    Zhou, Chenghu
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 5130 - 5144
  • [3] Progressive Generation of Long Text with Pretrained Language Models
    Tan, Bowen
    Yang, Zichao
    Al-Shedivat, Maruan
    Xing, Eric P.
    Hu, Zhiting
    2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), 2021, : 4313 - 4324
  • [4] Your fairness may vary: Pretrained language model fairness in toxic text classification
    Baldini, Ioana
    Wei, Dennis
    Ramamurthy, Karthikeyan Natesan
    Yurochkin, Mikhail
    Singh, Moninder
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), 2022, : 2245 - 2262
  • [5] Leveraging Text-to-Text Pretrained Language Models for Question Answering in Chemistry
    Tran, Dan
    Pascazio, Laura
    Akroyd, Jethro
    Mosbach, Sebastian
    Kraft, Markus
    ACS OMEGA, 2024, 9 (12): : 13883 - 13896
  • [6] Incremental Text-to-Speech Synthesis Using Pseudo Lookahead With Large Pretrained Language Model
    Saeki, Takaaki
    Takamichi, Shinnosuke
    Saruwatari, Hiroshi
    IEEE SIGNAL PROCESSING LETTERS, 2021, 28 : 857 - 861
  • [7] Dense Text Retrieval Based on Pretrained Language Models: A Survey
    Zhao, Wayne Xin
    Liu, Jing
    Ren, Ruiyang
    Wen, Ji-Rong
    ACM TRANSACTIONS ON INFORMATION SYSTEMS, 2024, 42 (04)
  • [8] A comparative study of pretrained language models for long clinical text
    Li, Yikuan
    Wehbe, Ramsey M.
    Ahmad, Faraz S.
    Wang, Hanyin
    Luo, Yuan
    JOURNAL OF THE AMERICAN MEDICAL INFORMATICS ASSOCIATION, 2022, : 340 - 347
  • [9] A Survey of Abstractive Text Summarization Utilising Pretrained Language Models
    Syed, Ayesha Ayub
    Gaol, Ford Lumban
    Boediman, Alfred
    Matsuo, Tokuro
    Budiharto, Widodo
    INTELLIGENT INFORMATION AND DATABASE SYSTEMS, ACIIDS 2022, PT I, 2022, 13757 : 532 - 544
  • [10] Pretrained Models with Adversarial Training for Named Entity Recognition in Scientific Text
    Ma, Hangchao
    Zhang, You
    Wang, Jin
    2022 INTERNATIONAL CONFERENCE ON ASIAN LANGUAGE PROCESSING (IALP 2022), 2022, : 259 - 264