共 50 条
How to Fine-Tune BERT for Text Classification?
被引:647
|作者:
Sun, Chi
[1
]
Qiu, Xipeng
[1
]
Xu, Yige
[1
]
Huang, Xuanjing
[1
]
机构:
[1] Fudan Univ, Sch Comp Sci, Shanghai Key Lab Intelligent Informat Proc, 825 Zhangheng Rd, Shanghai, Peoples R China
来源:
关键词:
Transfer learning;
BERT;
Text classification;
D O I:
10.1007/978-3-030-32381-3_16
中图分类号:
TP18 [人工智能理论];
学科分类号:
081104 ;
0812 ;
0835 ;
1405 ;
摘要:
Language model pre-training has proven to be useful in learning universal language representations. As a state-of-the-art language model pre-training model, BERT (Bidirectional Encoder Representations from Transformers) has achieved amazing results in many language understanding tasks. In this paper, we conduct exhaustive experiments to investigate different fine-tuning methods of BERT on text classification task and provide a general solution for BERT fine-tuning. Finally, the proposed solution obtains new state-of-the-art results on eight widely-studied text classification datasets.
引用
收藏
页码:194 / 206
页数:13
相关论文