Universal Language Model Fine-tuning for Text Classification

被引:0
|
作者
Howard, Jeremy [1 ]
Ruder, Sebastian [2 ,3 ]
机构
[1] Univ San Francisco, Fast Ai, San Francisco, CA 94117 USA
[2] NUI Galway, Insight Ctr, Galway, Ireland
[3] Aylien Ltd, Dublin, Ireland
基金
爱尔兰科学基金会;
关键词
D O I
暂无
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Inductive transfer learning has greatly impacted computer vision, but existing approaches in NLP still require task-specific modifications and training from scratch. We propose Universal Language Model Fine-tuning (ULMFiT), an effective transfer learning method that can be applied to any task in NLP, and introduce techniques that are key for fine-tuning a language model. Our method significantly outperforms the state-of-the-art on six text classification tasks, reducing the error by 18-24% on the majority of datasets. Furthermore, with only 100 labeled examples, it matches the performance of training from scratch on 100x more data. We opensource our pretrained models and code(1).
引用
收藏
页码:328 / 339
页数:12
相关论文
共 50 条
  • [31] Compressing BERT for Binary Text Classification via Adaptive Truncation before Fine-Tuning
    Zhang, Xin
    Fan, Jing
    Hei, Mengzhe
    APPLIED SCIENCES-BASEL, 2022, 12 (23):
  • [32] Improving Pretrained Language Model Fine-Tuning With Noise Stability Regularization
    Hua, Hang
    Li, Xingjian
    Dou, Dejing
    Xu, Cheng-Zhong
    Luo, Jiebo
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2025, 36 (01) : 1898 - 1910
  • [33] MultiFiT: Efficient Multi-lingual Language Model Fine-tuning
    Eisenschlos, Julian
    Ruder, Sebastian
    Czapla, Piotr
    Kardas, Marcin
    Gugger, Sylvain
    Howard, Jeremy
    2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 5702 - 5707
  • [34] MediBioDeBERTa: Biomedical Language Model With Continuous Learning and Intermediate Fine-Tuning
    Kim, Eunhui
    Jeong, Yuna
    Choi, Myung-Seok
    IEEE ACCESS, 2023, 11 : 141036 - 141044
  • [35] Automatic text simplification for French: model fine-tuning for simplicity assessment and simpler text generation
    Lucía Ormaechea
    Nikos Tsourakis
    International Journal of Speech Technology, 2024, 27 (4) : 957 - 976
  • [36] FINE-TUNING TRANSFER LEARNING MODEL IN WOVEN FABRIC PATTERN CLASSIFICATION
    Noprisson H.
    Ermatita E.
    Abdiansah A.
    Ayumi V.
    Purba M.
    Setiawan H.
    International Journal of Innovative Computing, Information and Control, 2022, 18 (06): : 1885 - 1894
  • [37] Auto-Encoder Classification Model for Water Crystals with Fine-Tuning
    Mahmoud, Hanan A. Hosni A.
    Hakami, Nada Ali
    CRYSTALS, 2022, 12 (11)
  • [38] Implementing a provincial-level universal daily industrial carbon emissions prediction by fine-tuning the large language model
    Feng, Zhengyuan
    Sun, Yuheng
    Ning, Jun
    Tang, Shoujuan
    Liu, Guangxin
    Liu, Fangtao
    Li, Yang
    Shi, Lei
    APPLIED ENERGY, 2025, 383
  • [39] A transformer fine-tuning strategy for text dialect identification
    Mohammad Ali Humayun
    Hayati Yassin
    Junaid Shuja
    Abdullah Alourani
    Pg Emeroylariffion Abas
    Neural Computing and Applications, 2023, 35 : 6115 - 6124
  • [40] A transformer fine-tuning strategy for text dialect identification
    Humayun, Mohammad Ali
    Yassin, Hayati
    Shuja, Junaid
    Alourani, Abdullah
    Abas, Pg Emeroylariffion
    NEURAL COMPUTING & APPLICATIONS, 2023, 35 (08): : 6115 - 6124