SCIBERT: A Pretrained Language Model for Scientific Text

被引:0
|
作者
Beltagy, Iz [1 ]
Lo, Kyle [1 ]
Cohan, Arman [1 ]
机构
[1] Allen Inst Artificial Intelligence, Seattle, WA 98103 USA
关键词
CORPUS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Obtaining large-scale annotated data for NLP tasks in the scientific domain is challenging and expensive. We release SCIBERT, a pretrained language model based on BERT (Devlin et al., 2019) to address the lack of highquality, large-scale labeled scientific data. SCIBERT leverages unsupervised pretraining on a large multi-domain corpus of scientific publications to improve performance on downstream scientific NLP tasks. We evaluate on a suite of tasks including sequence tagging, sentence classification and dependency parsing, with datasets from a variety of scientific domains. We demonstrate statistically significant improvements over BERT and achieve new state-of-the-art results on several of these tasks.
引用
收藏
页码:3615 / 3620
页数:6
相关论文
共 50 条
  • [41] Reusing a Pretrained Language Model on Languages with Limited Corpora for Unsupervised NMT
    Chronopoulou, Alexandra
    Stojanovski, Dario
    Fraser, Alexander
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 2703 - 2711
  • [42] On the Effectiveness of Adapter-based Tuning for Pretrained Language Model Adaptation
    He, Ruidan
    Liu, Linlin
    Ye, Hai
    Tan, Qingyu
    Ding, Bosheng
    Cheng, Liying
    Low, Jia-Wei
    Bing, Lidong
    Si, Luo
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 1 (ACL-IJCNLP 2021), 2021, : 2208 - 2222
  • [43] Layerwised multimodal knowledge distillation for vision-language pretrained model
    Wang, Jin
    Liao, Dawei
    Zhang, You
    Xu, Dan
    Zhang, Xuejie
    NEURAL NETWORKS, 2024, 175
  • [44] KBioXLM: A Knowledge-anchored Biomedical Multilingual Pretrained Language Model
    Geng, Lei
    Yan, Xu
    Cao, Ziqiang
    Li, Juntao
    Li, Wenjie
    Li, Sujian
    Zhou, Xinjie
    Yang, Yang
    Zhang, Jun
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 11239 - 11250
  • [45] Unsupervised Domain Adaptation of a Pretrained Cross-Lingual Language Model
    Li, Juntao
    He, Ruidan
    Ye, Hai
    Ng, Hwee Tou
    Bing, Lidong
    Yan, Rui
    PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 3672 - 3678
  • [46] Pretrained domain-specific language model for natural language processing tasks in the AEC domain
    Zheng, Zhe
    Lu, Xin-Zheng
    Chen, Ke-Yin
    Zhou, Yu-Cheng
    Lin, Jia-Rui
    COMPUTERS IN INDUSTRY, 2022, 142
  • [47] A Hybrid Citation Recommendation Model With SciBERT and GraphSAGE
    Dinh, Thi N.
    Pham, Phu
    Nguyen, Giang L.
    Nguyen, Ngoc Thanh
    Vo, Bay
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2025, 55 (02): : 852 - 863
  • [48] Chinese Prosodic Structure Prediction Based on a Pretrained Language Representation Model
    Zhang P.
    Lu C.
    Wang R.
    Zhang, Pengyuan (zhangpengyuan@hccl.ioa.ac.cn), 1600, Tianjin University (53): : 265 - 271
  • [49] Pretrained Transformers for Text Ranking: BERT and Beyond
    Yates, Andrew
    Nogueira, Rodrigo
    Lin, Jimmy
    SIGIR '21 - PROCEEDINGS OF THE 44TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2021, : 2666 - 2668
  • [50] Language model adaptation for language and dialect identification of text
    Jauhiainen, T.
    Linden, K.
    Jauhiainen, H.
    NATURAL LANGUAGE ENGINEERING, 2019, 25 (05) : 561 - 583