Mosaicking to Distill: Knowledge Distillation from Out-of-Domain Data

被引:0
|
作者
Fang, Gongfan [1 ,4 ]
Bao, Yifan [1 ]
Song, Jie [1 ]
Wang, Xinchao [2 ]
Xie, Donglin [1 ]
Shen, Chengchao [3 ]
Song, Mingli [1 ]
机构
[1] Zhejiang Univ, Hangzhou, Peoples R China
[2] Natl Univ Singapore, Singapore, Singapore
[3] Cent South Univ, Changsha, Peoples R China
[4] Alibaba Zhejiang Univ Joint Inst Frontier Technol, Hangzhou, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Knowledge distillation (KD) aims to craft a compact student model that imitates the behavior of a pre-trained teacher in a target domain. Prior KD approaches, despite their gratifying results, have largely relied on the premise that in-domain data is available to carry out the knowledge transfer. Such an assumption, unfortunately, in many cases violates the practical setting, since the original training data or even the data domain is often unreachable due to privacy or copyright reasons. In this paper, we attempt to tackle an ambitious task, termed as out-of-domain knowledge distillation (OOD-KD), which allows us to conduct KD using only OOD data that can be readily obtained at a very low cost. Admittedly, OOD-KD is by nature a highly challenging task due to the agnostic domain gap. To this end, we introduce a handy yet surprisingly efficacious approach, dubbed as MosaicKD. The key insight behind MosaicKD lies in that, samples from various domains share common local patterns, even though their global semantic may vary significantly; these shared local patterns, in turn, can be re-assembled analogous to mosaic tiling, to approximate the in-domain data and to further alleviating the domain discrepancy. In MosaicKD, this is achieved through a four-player min-max game, in which a generator, a discriminator, a student network, are collectively trained in an adversarial manner, partially under the guidance of a pre-trained teacher. We validate MosaicKD over classification and semantic segmentation tasks across various benchmarks, and demonstrate that it yields results much superior to the state-of-the-art counterparts on OOD data. Our code is available at https://github.com/zju-vipa/MosaicKD.
引用
收藏
页数:13
相关论文
共 50 条
  • [21] Learning from noisy out-of-domain corpus using dataless classification
    Jin, Yiping
    Wanvarie, Dittaya
    Le, Phu T., V
    NATURAL LANGUAGE ENGINEERING, 2022, 28 (01) : 39 - 69
  • [22] IN-DOMAIN AND OUT-OF-DOMAIN DATA AUGMENTATION TO IMPROVE CHILDREN'S SPEAKER VERIFICATION SYSTEM IN LIMITED DATA SCENARIO
    Shahnawazuddin, S.
    Ahmad, Waquar
    Adiga, Nagaraj
    Kumar, Avinash
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 7554 - 7558
  • [23] An Empirical Study on Explanations in Out-of-Domain Settings
    Chrysostomou, George
    Aletras, Nikolaos
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 6920 - 6938
  • [24] Out-of-domain FrameNet Semantic Role Labeling
    Hartmann, Silvana
    Kuznetsov, Ilia
    Martin, Teresa
    Gurevych, Iryna
    15TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2017), VOL 1: LONG PAPERS, 2017, : 471 - 482
  • [25] Assessing Out-of-Domain Language Model Performance from Few Examples
    Singhal, Prasann
    Forristal, Jarad
    Ye, Xi
    Durrett, Greg
    17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 2385 - 2397
  • [26] Distill on the Go: Online knowledge distillation in self-supervised learning
    Bhat, Prashant
    Arani, Elahe
    Zonooz, Bahram
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2021, 2021, : 2672 - 2681
  • [27] Divide and Distill: New Outlooks on Knowledge Distillation for Environmental Sound Classification
    Tripathi, Achyut Mani
    Pandey, Om Jee
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2023, 31 : 1100 - 1113
  • [28] Practical and Efficient Out-of-Domain Detection with Adversarial Learning
    Wang, Bo
    Mine, Tsunenori
    37TH ANNUAL ACM SYMPOSIUM ON APPLIED COMPUTING, 2022, : 853 - 862
  • [29] Efficient Out-of-Domain Detection for Sequence to Sequence Models
    Vazhentsev, Artem
    Tsvigun, Akim
    Vashurin, Roman
    Petrakov, Sergey
    Vasilev, Daniil
    Panov, Maxim
    Panchenko, Alexander
    Shelmanov, Artem
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 1430 - 1454
  • [30] Extraction of Specific Arguments from Chinese Financial News with out-of-domain Samples
    Luo, Yu
    Zou, Xinyi
    Liu, Di
    Peng, Wanwan
    Wu, Xiaohua
    PROCEEDINGS OF THE 10TH INTERNATIONAL CONFERENCE OF INFORMATION AND COMMUNICATION TECHNOLOGY, 2021, 183 : 288 - 294