Cross-Modal Knowledge Distillation with Dropout-Based Confidence

被引:0
|
作者
Cho, Won Ik [1 ]
Kim, Jeunghun [2 ,3 ]
Kim, Nam Soo [2 ,3 ]
机构
[1] Samsung Adv Inst Technol, Samsung Elect, Suwon, South Korea
[2] Seoul Natl, Dept Elect & Comp Engn, Seoul, South Korea
[3] Seoul Natl, INMC, Seoul, South Korea
关键词
dropout; uncertainty; cross-modal distillation; spoken language understanding;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In cross-modal distillation, e.g., from text-based inference modules to spoken language understanding module, it is difficult to determine the teacher's influence due to the different nature of both modalities that bring the heterogeneity in the aspect of uncertainty. Though error rate or entropy-based schemes have been suggested to cope with the heuristics of time-based scheduling, the confidence of the teacher inference has not been necessarily taken into deciding the teacher's influence. In this paper, we propose a dropout-based confidence that decides the teacher's confidence and to-student influence of the loss. On the widely used spoken language understanding dataset, Fluent Speech Command, we show that our weight decision scheme enhances performance in combination with the conventional scheduling strategies, displaying a maximum 20% relative error reduction concerning the model with no distillation.
引用
收藏
页码:653 / 657
页数:5
相关论文
共 50 条
  • [21] Visual-to-EEG cross-modal knowledge distillation for continuous emotion recognition
    Zhang, Su
    Tang, Chuangao
    Guan, Cuntai
    [J]. PATTERN RECOGNITION, 2022, 130
  • [22] Visual-to-EEG cross-modal knowledge distillation for continuous emotion recognition
    Zhang, Su
    Tang, Chuangao
    Guan, Cuntai
    [J]. PATTERN RECOGNITION, 2022, 130
  • [23] Creating Something from Nothing: Unsupervised Knowledge Distillation for Cross-Modal Hashing
    Hu, Hengtong
    Xie, Lingxi
    Hong, Richang
    Tian, Qi
    [J]. 2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, : 3120 - 3129
  • [24] Weakly Supervised Exaggeration Transfer for Caricature Generation With Cross-Modal Knowledge Distillation
    Tong, Shuo
    Liu, Han
    He, Yuxin
    Du, Chenxiao
    Wang, Wenqing
    Guo, Runyuan
    Liu, Jingyun
    [J]. IEEE COMPUTER GRAPHICS AND APPLICATIONS, 2024, 44 (04) : 98 - 112
  • [25] XKD: Cross-Modal Knowledge Distillation with Domain Alignment for Video Representation Learning
    Sarkar, Pritam
    Etemad, Ali
    [J]. THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 13, 2024, : 14875 - 14885
  • [26] Cross-Modal Graph Knowledge Representation and Distillation Learning for Land Cover Classification
    Wang, Wenzhen
    Liu, Fang
    Liao, Wenzhi
    Xiao, Liang
    [J]. IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2023, 61
  • [27] A Precise and Scalable Indoor Positioning System Using Cross-Modal Knowledge Distillation
    Rizk, Hamada
    Elmogy, Ahmed
    Rihan, Mohamed
    Yamaguchi, Hirozumi
    [J]. Sensors, 2024, 24 (22)
  • [28] Learning an Augmented RGB Representation with Cross-Modal Knowledge Distillation for Action Detection
    Dai, Rui
    Das, Srijan
    Bremond, Francois
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 13033 - 13044
  • [29] Cross-modal distillation for flood extent mapping
    Garg, Shubhika
    Feinstein, Ben
    Timnat, Shahar
    Batchu, Vishal
    Dror, Gideon
    Rosenthal, Adi Gerzi
    Gulshan, Varun
    [J]. ENVIRONMENTAL DATA SCIENCE, 2023, 2
  • [30] Cross-Modal Object Detection Based on a Knowledge Update
    Gao, Yueqing
    Zhou, Huachun
    Chen, Lulu
    Shen, Yuting
    Guo, Ce
    Zhang, Xinyu
    [J]. SENSORS, 2022, 22 (04)