SCIBERT: A Pretrained Language Model for Scientific Text

被引:0
|
作者
Beltagy, Iz [1 ]
Lo, Kyle [1 ]
Cohan, Arman [1 ]
机构
[1] Allen Inst Artificial Intelligence, Seattle, WA 98103 USA
关键词
CORPUS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Obtaining large-scale annotated data for NLP tasks in the scientific domain is challenging and expensive. We release SCIBERT, a pretrained language model based on BERT (Devlin et al., 2019) to address the lack of highquality, large-scale labeled scientific data. SCIBERT leverages unsupervised pretraining on a large multi-domain corpus of scientific publications to improve performance on downstream scientific NLP tasks. We evaluate on a suite of tasks including sequence tagging, sentence classification and dependency parsing, with datasets from a variety of scientific domains. We demonstrate statistically significant improvements over BERT and achieve new state-of-the-art results on several of these tasks.
引用
收藏
页码:3615 / 3620
页数:6
相关论文
共 50 条
  • [21] A Survey on Model Compression and Acceleration for Pretrained Language Models
    Xu, Canwen
    McAuley, Julian
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 9, 2023, : 10566 - 10575
  • [22] The Classification of Short Scientific Texts Using Pretrained BERT Model
    Danilov, Gleb
    Ishankulov, Timur
    Kotik, Konstantin
    Orlov, Yuriy
    Shifrin, Mikhail
    Potapov, Alexander
    PUBLIC HEALTH AND INFORMATICS, PROCEEDINGS OF MIE 2021, 2021, 281 : 83 - 87
  • [23] Text Summarization with Pretrained Encoders
    Liu, Yang
    Lapata, Mirella
    2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 3730 - 3740
  • [24] A Novel Pretrained General-purpose Vision Language Model for the Vietnamese Language
    Dinh Anh Vu
    Quang Nhat Minh Pham
    Giang Son Tran
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2024, 23 (05)
  • [25] ReGen: Reinforcement Learning for Text and Knowledge Base Generation using Pretrained Language Models
    Dognin, Pierre L.
    Padhi, Inkit
    Melnyk, Igor
    Das, Payel
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 1084 - 1099
  • [26] Few-shot Knowledge Graph-to-Text Generation with Pretrained Language Models
    Li, Junyi
    Tang, Tianyi
    Zhao, Wayne Xin
    Wei, Zhicheng
    Yuan, Nicholas Jing
    Wen, Ji-Rong
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 1558 - 1568
  • [27] The evidential model of a scientific text
    Bazhenova, Elena A.
    Kotyurova, Maria P.
    VESTNIK TOMSKOGO GOSUDARSTVENNOGO UNIVERSITETA FILOLOGIYA-TOMSK STATE UNIVERSITY JOURNAL OF PHILOLOGY, 2023, 81 : 269 - 278
  • [28] Comparative Study of Multiclass Text Classification in Research Proposals Using Pretrained Language Models
    Lee, Eunchan
    Lee, Changhyeon
    Ahn, Sangtae
    APPLIED SCIENCES-BASEL, 2022, 12 (09):
  • [29] ChestXRayBERT: A Pretrained Language Model for Chest Radiology Report Summarization
    Cai, Xiaoyan
    Liu, Sen
    Han, Junwei
    Yang, Libin
    Liu, Zhenguo
    Liu, Tianming
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 845 - 855
  • [30] Predicting Immune Escape with Pretrained Protein Language Model Embeddings
    Swanson, Kyle
    Chang, Howard
    Zou, James
    MACHINE LEARNING IN COMPUTATIONAL BIOLOGY, VOL 200, 2022, 200