Knowledge distillation for BERT unsupervised domain adaptation

被引:12
|
作者
Ryu, Minho [1 ]
Lee, Geonseok [2 ]
Lee, Kichun [2 ]
机构
[1] SK Telecom, Seoul, South Korea
[2] Hanyang Univ, Dept Ind Engn, Seoul, South Korea
基金
新加坡国家研究基金会;
关键词
Language model; Knowledge distillation; Domain adaptation;
D O I
10.1007/s10115-022-01736-y
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A pre-trained language model, BERT, has brought significant performance improvements across a range of natural language processing tasks. Since the model is trained on a large corpus of diverse topics, it shows robust performance for domain shift problems in which data distributions at training (source data) and testing (target data) differ while sharing similarities. Despite its great improvements compared to previous models, it still suffers from performance degradation due to domain shifts. To mitigate such problems, we propose a simple but effective unsupervised domain adaptation method, adversarial adaptation with distillation (AAD), which combines the adversarial discriminative domain adaptation (ADDA) framework with knowledge distillation. We evaluate our approach in the task of cross-domain sentiment classification on 30 domain pairs, advancing the state-of-the-art performance for unsupervised domain adaptation in text sentiment classification.
引用
收藏
页码:3113 / 3128
页数:16
相关论文
共 50 条
  • [1] Knowledge distillation for BERT unsupervised domain adaptation
    Minho Ryu
    Geonseok Lee
    Kichun Lee
    [J]. Knowledge and Information Systems, 2022, 64 : 3113 - 3128
  • [2] Joint Progressive Knowledge Distillation and Unsupervised Domain Adaptation
    Nguyen-Meidine, Le Thanh
    Granger, Eric
    Kiran, Madhu
    Dolz, Jose
    Blais-Morin, Louis-Antoine
    [J]. 2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [3] Bilateral Knowledge Distillation for Unsupervised Domain Adaptation of Semantic Segmentation
    Wang, Yunnan
    Li, Jianxun
    [J]. 2022 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2022, : 10177 - 10184
  • [4] Unsupervised Multi-Target Domain Adaptation Through Knowledge Distillation
    Nguyen-Meidine, L. T.
    Belal, A.
    Kiran, M.
    Dolz, J.
    Blais-Morin, L-A
    Granger, E.
    [J]. 2021 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2021), 2021, : 1338 - 1346
  • [5] Unsupervised domain adaptation for lip reading based on cross-modal knowledge distillation
    Takashima, Yuki
    Takashima, Ryoichi
    Tsunoda, Ryota
    Aihara, Ryo
    Takiguchi, Tetsuya
    Ariki, Yasuo
    Motoyama, Nobuaki
    [J]. EURASIP JOURNAL ON AUDIO SPEECH AND MUSIC PROCESSING, 2021, 2021 (01)
  • [6] Unsupervised domain adaptation for lip reading based on cross-modal knowledge distillation
    Yuki Takashima
    Ryoichi Takashima
    Ryota Tsunoda
    Ryo Aihara
    Tetsuya Takiguchi
    Yasuo Ariki
    Nobuaki Motoyama
    [J]. EURASIP Journal on Audio, Speech, and Music Processing, 2021
  • [7] Geometric Knowledge Embedding for unsupervised domain adaptation
    Wu, Hanrui
    Yan, Yuguang
    Ye, Yuzhong
    Ng, Michael K.
    Wu, Qingyao
    [J]. KNOWLEDGE-BASED SYSTEMS, 2020, 191
  • [8] Prior Knowledge Guided Unsupervised Domain Adaptation
    Sun, Tao
    Lu, Cheng
    Ling, Haibin
    [J]. COMPUTER VISION - ECCV 2022, PT XXXIII, 2022, 13693 : 639 - 655
  • [9] Domain Adaptation and Summary Distillation for Unsupervised Query Focused Summarization
    Du, Jiancheng
    Gao, Yang
    [J]. IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (03) : 1044 - 1055
  • [10] Unsupervised Domain Adaptation Method Based on Domain-Invariant Features Evaluation and Knowledge Distillation for Bearing Fault Diagnosis
    Sun, Kong
    Bo, Lin
    Ran, Haoting
    Tang, Zhi
    Bi, Yuanliang
    [J]. IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2023, 72