Identifying the Limits of Cross-Domain Knowledge Transfer for Pretrained Models

被引:0
|
作者
Wu, Zhengxuan [1 ]
Liu, Nelson F. [1 ]
Potts, Christopher [1 ]
机构
[1] Stanford Univ, Stanford, CA 94305 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
There is growing evidence that pretrained language models improve task-specific fine-tuning even where the task examples are radically different from those seen in training. We study an extreme case of transfer learning by providing a systematic exploration of how much transfer occurs when models are denied any information about word identity via random scrambling. In four classification tasks and two sequence labeling tasks, we evaluate LSTMs using GloVe embeddings, BERT, and baseline models. Among these models, we find that only BERT shows high rates of transfer into our scrambled domains, and for classification but not sequence labeling tasks. Our analyses seek to explain why transfer succeeds for some tasks but not others, to isolate the separate contributions of pretraining versus fine-tuning, to show that the fine-tuning process is not merely learning to unscramble the scrambled inputs, and to quantify the role of word frequency. Furthermore, our results suggest that current benchmarks may overestimate the degree to which current models actually understand language.
引用
收藏
页码:100 / 110
页数:11
相关论文
共 50 条
  • [21] Prototypical Cross-domain Knowledge Transfer for Cervical Dysplasia Visual Inspection
    Zhang, Yichen
    Yin, Yifang
    Zhang, Ying
    Liu, Zhenguang
    Wang, Zheng
    Zimmermann, Roger
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 1504 - 1514
  • [22] Multimodal Hate Speech Detection via Cross-Domain Knowledge Transfer
    Yang, Chuanpeng
    Zhu, Fuqing
    Liu, Guihua
    Han, Jizhong
    Hu, Songlin
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, : 4505 - 4514
  • [23] Heterogeneous Knowledge Transfer via Domain Regularization for Improving Cross-Domain Collaborative Filtering
    Zang, Yizhou
    Hu, Xiaohua
    2017 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2017, : 3968 - 3974
  • [24] Domain transfer via cross-domain analogy
    Klenk, Matthew
    Forbus, Ken
    COGNITIVE SYSTEMS RESEARCH, 2009, 10 (03) : 240 - 250
  • [25] A cross-domain trust inferential transfer model for cross-domain Industrial Internet of Things
    Wu, Xu
    Liang, Junbin
    ICT EXPRESS, 2023, 9 (05): : 761 - 768
  • [26] Dual-Level Adaptive and Discriminative Knowledge Transfer for Cross-Domain Recognition
    Meng, Min
    Lan, Mengcheng
    Yu, Jun
    Wu, Jigang
    Liu, Ligang
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 2266 - 2279
  • [27] Cross-Domain Knowledge Transfer for Incremental Deep Learning in Facial Expression Recognition
    Sugianto, Nehemia
    Tjondronegoro, Dian
    2019 7TH INTERNATIONAL CONFERENCE ON ROBOT INTELLIGENCE TECHNOLOGY AND APPLICATIONS (RITA), 2019, : 205 - 209
  • [28] Iterative Transfer Knowledge Distillation and Channel Pruning for Unsupervised Cross-Domain Compression
    Wang, Zhiyuan
    Shi, Long
    Mei, Zhen
    Zhao, Xiang
    Wang, Zhe
    Li, Jun
    WEB INFORMATION SYSTEMS AND APPLICATIONS, WISA 2024, 2024, 14883 : 3 - 15
  • [29] Knowledge transfer by domain-independent user latent factor for cross-domain recommender systems
    Sahu, Ashish Kumar
    Dwivedi, Pragya
    FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2020, 108 : 320 - 333
  • [30] Autonomous Cross-Domain Knowledge Transfer in Lifelong Policy Gradient Reinforcement Learning
    Ammar, Haitham Bou
    Eaton, Eric
    Luna, Jose Marcio
    Ruvolo, Paul
    PROCEEDINGS OF THE TWENTY-FOURTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE (IJCAI), 2015, : 3345 - 3351