SCIBERT: A Pretrained Language Model for Scientific Text

被引:0
|
作者
Beltagy, Iz [1 ]
Lo, Kyle [1 ]
Cohan, Arman [1 ]
机构
[1] Allen Inst Artificial Intelligence, Seattle, WA 98103 USA
关键词
CORPUS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Obtaining large-scale annotated data for NLP tasks in the scientific domain is challenging and expensive. We release SCIBERT, a pretrained language model based on BERT (Devlin et al., 2019) to address the lack of highquality, large-scale labeled scientific data. SCIBERT leverages unsupervised pretraining on a large multi-domain corpus of scientific publications to improve performance on downstream scientific NLP tasks. We evaluate on a suite of tasks including sequence tagging, sentence classification and dependency parsing, with datasets from a variety of scientific domains. We demonstrate statistically significant improvements over BERT and achieve new state-of-the-art results on several of these tasks.
引用
收藏
页码:3615 / 3620
页数:6
相关论文
共 50 条
  • [31] Adapting Pretrained Text-to-Text Models for Long Text Sequences
    Xiong, Wenhan
    Gupta, Anchit
    Toshniwal, Shubham
    Mehdad, Yashar
    Yih, Wen-tau
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 5566 - 5578
  • [32] COVIDSum: A linguistically enriched SciBERT-based summarization model for COVID-19 scientific papers
    Cai, Xiaoyan
    Liu, Sen
    Yang, Libin
    Lu, Yan
    Zhao, Jintao
    Shen, Dinggang
    Liu, Tianming
    JOURNAL OF BIOMEDICAL INFORMATICS, 2022, 127
  • [33] Adapting Pretrained Representations for Text Mining
    Meng, Yu
    Huang, Jiaxin
    Zhang, Yu
    Han, Jiawei
    PROCEEDINGS OF THE 28TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2022, 2022, : 4806 - 4807
  • [34] DN at SemEval-2023 Task 12: Low-Resource Language Text Classification via Multilingual Pretrained Language Model Fine-tuning
    Daniil, Homskiy
    Narek, Maloyan
    17TH INTERNATIONAL WORKSHOP ON SEMANTIC EVALUATION, SEMEVAL-2023, 2023, : 1537 - 1541
  • [35] A Survey of Pretrained Language Models
    Sun, Kaili
    Luo, Xudong
    Luo, Michael Y.
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, PT II, 2022, 13369 : 442 - 456
  • [36] Identifying Speakers in Dialogue Transcripts: A Text-based Approach Using Pretrained Language Models
    Nguyen, Minh
    Dernoncourt, Franck
    Yoon, Seunghyun
    Deilamsalehy, Hanieh
    Tana, Hao
    Rossi, Ryan
    Trani, Quan Hung
    Bui, Trung
    Nguyen, Thien Huu
    INTERSPEECH 2024, 2024, : 3799 - 3803
  • [37] Can Pretrained Language Models Generate Persuasive, Faithful, and Informative Ad Text for Product Descriptions?
    Koto, Fajri
    Lau, Jey Han
    Baldwin, Timothy
    PROCEEDINGS OF THE 5TH WORKSHOP ON E-COMMERCE AND NLP (ECNLP 5), 2022, : 234 - 243
  • [38] CLIPSONIC: TEXT-TO-AUDIO SYNTHESIS WITH UNLABELED VIDEOS AND PRETRAINED LANGUAGE-VISION MODELS
    Dong, Hao-Wen
    Liu, Xiaoyu
    Pons, Jordi
    Bhattacharya, Gautam
    Pascual, Santiago
    Serra, Joan
    Berg-Kirkpatrick, Taylor
    McAuley, Julian
    2023 IEEE WORKSHOP ON APPLICATIONS OF SIGNAL PROCESSING TO AUDIO AND ACOUSTICS, WASPAA, 2023,
  • [39] SCINLI: A Corpus for Natural Language Inference on Scientific Text
    Sadat, Mobashir
    Caragea, Cornelia
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 7399 - 7409
  • [40] An Unsupervised Clinical Acronym Disambiguation Method Based on Pretrained Language Model
    Wei, Siwen
    Yuan, Chi
    Li, Zixuan
    Wang, Huaiyu
    HEALTH INFORMATION PROCESSING, CHIP 2023, 2023, 1993 : 270 - 284