Identifying the Limits of Cross-Domain Knowledge Transfer for Pretrained Models

被引:0
|
作者
Wu, Zhengxuan [1 ]
Liu, Nelson F. [1 ]
Potts, Christopher [1 ]
机构
[1] Stanford Univ, Stanford, CA 94305 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
There is growing evidence that pretrained language models improve task-specific fine-tuning even where the task examples are radically different from those seen in training. We study an extreme case of transfer learning by providing a systematic exploration of how much transfer occurs when models are denied any information about word identity via random scrambling. In four classification tasks and two sequence labeling tasks, we evaluate LSTMs using GloVe embeddings, BERT, and baseline models. Among these models, we find that only BERT shows high rates of transfer into our scrambled domains, and for classification but not sequence labeling tasks. Our analyses seek to explain why transfer succeeds for some tasks but not others, to isolate the separate contributions of pretraining versus fine-tuning, to show that the fine-tuning process is not merely learning to unscramble the scrambled inputs, and to quantify the role of word frequency. Furthermore, our results suggest that current benchmarks may overestimate the degree to which current models actually understand language.
引用
收藏
页码:100 / 110
页数:11
相关论文
共 50 条
  • [41] Knowledge Transformation for Cross-Domain Sentiment Classification
    Li, Tao
    Sindhwani, Vikas
    Ding, Chris
    Zhang, Yi
    PROCEEDINGS 32ND ANNUAL INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2009, : 716 - 717
  • [42] Knowledge Transformation by Cross-Domain Belief Propagation
    Wang, Fei
    Li, Tao
    2009 IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS (ICDMW 2009), 2009, : 441 - 446
  • [43] Cross-Domain Product Search with Knowledge Graph
    Zhu, Rui
    Zhao, Yiming
    Qu, Wei
    Liu, Zhongyi
    Li, Chenliang
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 3746 - 3755
  • [44] Knowledge-Reinforced Cross-Domain Recommendation
    Huang, Ling
    Huang, Xiao-Dong
    Zou, Han
    Gao, Yuefang
    Wang, Chang-Dong
    Yu, Philip S.
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024,
  • [45] DBpedia: A Multilingual Cross-Domain Knowledge Base
    Mendes, Pablo N.
    Jakob, Max
    Bizer, Christian
    LREC 2012 - EIGHTH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2012, : 1813 - 1817
  • [46] Cross-domain knowledge distillation for text classification
    Zhang, Shaokang
    Jiang, Lei
    Tan, Jianlong
    NEUROCOMPUTING, 2022, 509 : 11 - 20
  • [47] Cross-Domain Transfer of EEG to EEG or ECG Learning for CNN Classification Models
    Yang, Chia-Yen
    Chen, Pin-Chen
    Huang, Wen-Chen
    SENSORS, 2023, 23 (05)
  • [48] Cross-Domain Transfer of Generative Explanations Using Text-to-Text Models
    Erliksson, Karl Fredrik
    Arpteg, Anders
    Matskin, Mihhail
    Payberah, Amir H.
    NATURAL LANGUAGE PROCESSING AND INFORMATION SYSTEMS (NLDB 2021), 2021, 12801 : 76 - 89
  • [49] Formal Cross-Domain Ontologization of Human Knowledge
    Isaeva, Ekaterina
    Bakhtin, Vadim
    Tararkov, Andrey
    INFORMATION TECHNOLOGY AND SYSTEMS, ICITS 2020, 2020, 1137 : 94 - 103
  • [50] Cross-domain recommendation via knowledge distillation
    Li, Xiuze
    Huang, Zhenhua
    Wu, Zhengyang
    Wang, Changdong
    Chen, Yunwen
    KNOWLEDGE-BASED SYSTEMS, 2025, 311