Calibrated Language Model Fine-Tuning for In- and Out-of-Distribution Data

被引:0
|
作者
Kong, Lingkai [1 ]
Jiang, Haoming [1 ]
Zhuang, Yuchen [1 ]
Lyu, Jie [1 ]
Zhao, Tuo [1 ]
Zhang, Chao [1 ]
机构
[1] Georgia Inst Technol, Atlanta, GA USA
基金
美国国家科学基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Fine-tuned pre-trained language models can suffer from severe miscalibration for both in-distribution and out-of-distribution (OOD) data due to over-parameterization. To mitigate this issue, we propose a regularized fine-tuning method. Our method introduces two types of regularization for better calibration: (1) On-manifold regularization, which generates pseudo on-manifold samples through interpolation within the data manifold. Augmented training with these pseudo samples imposes a smoothness regularization to improve in-distribution calibration. (2) Off-manifold regularization, which encourages the model to output uniform distributions for pseudo off-manifold samples to address the over-confidence issue for OOD data. Our experiments demonstrate that the proposed method outperforms existing calibration methods for text classification in terms of expectation calibration error, misclassification detection, and OOD detection on six datasets. Our code can be found at https://github.com/Lingkai-Kong/Calibrated-BERT-Fine-Tuning.
引用
收藏
页码:1326 / 1340
页数:15
相关论文
共 50 条
  • [1] Model Reprogramming Outperforms Fine-tuning on Out-of-distribution Data in Text-Image Encoders
    Geng, Andrew
    Chen, Pin-Yu
    [J]. IEEE CONFERENCE ON SAFE AND TRUSTWORTHY MACHINE LEARNING, SATML 2024, 2024, : 552 - 568
  • [2] How Does Fine-Tuning Impact Out-of-Distribution Detection for Vision-Language Models?
    Yifei Ming
    Yixuan Li
    [J]. International Journal of Computer Vision, 2024, 132 : 596 - 609
  • [3] How Does Fine-Tuning Impact Out-of-Distribution Detection for Vision-Language Models?
    Ming, Yifei
    Li, Yixuan
    [J]. INTERNATIONAL JOURNAL OF COMPUTER VISION, 2024, 132 (02) : 596 - 609
  • [4] Fine-Tuning Deteriorates General Textual Out-of-Distribution Detection by Distorting Task-Agnostic Features
    Chen, Sishuo
    Yang, Wenkai
    Bi, Xiaohan
    Sun, Xu
    [J]. 17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 564 - 579
  • [5] Fine-tuning vs From Scratch: Do Vision & Language Models Have Similar Capabilities on Out-of-Distribution Visual Question Answering?
    Jensen, Kristian Norgaard
    Plank, Barbara
    [J]. LREC 2022: THIRTEEN INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2022, : 1496 - 1508
  • [6] Calibrated Out-of-Distribution Detection with a Generic Representation
    Vojir, Tomas
    Sochman, Jan
    Aljundi, Rahaf
    Matas, Jiri
    [J]. 2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS, ICCVW, 2023, : 4509 - 4518
  • [7] Data Fine-Tuning
    Chhabra, Saheb
    Majumdar, Puspita
    Vatsa, Mayank
    Singh, Richa
    [J]. THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 8223 - 8230
  • [8] Foundation Models and Fine-Tuning: A Benchmark for Out of Distribution Detection
    Cappio Borlino, Francesco
    Lu, Lorenzo
    Tommasi, Tatiana
    [J]. IEEE ACCESS, 2024, 12 : 79401 - 79414
  • [9] Comprehensive Review of Large Language Model Fine-Tuning
    Zhang, Qintong
    Wang, Yuchao
    Wang, Hexi
    Wang, Junxin
    Chen, Hai
    [J]. Computer Engineering and Applications, 2024, 60 (17) : 17 - 33
  • [10] Patent classification by fine-tuning BERT language model
    Lee, Jieh-Sheng
    Hsiang, Jieh
    [J]. WORLD PATENT INFORMATION, 2020, 61