Dropout Distillation

被引:0
|
作者
Bulo, Samuel Rota [1 ]
Porzi, Lorenzo [1 ]
Kontschieder, Peter [2 ,3 ]
机构
[1] FBK Irst, Trento, Italy
[2] Mapillary, Graz, Austria
[3] Microsoft Res, Cambridge, England
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Dropout is a popular stochastic regularization technique for deep neural networks that works by randomly dropping (i.e. zeroing) units from the network during training. This randomization process allows to implicitly train an ensemble of exponentially many networks sharing the same parametrization, which should be averaged at test time to deliver the final prediction. A typical workaround for this intractable averaging operation consists in scaling the layers undergoing dropout randomization. This simple rule called "standard dropout" is efficient, but might degrade the accuracy of the prediction. In this work we introduce a novel approach, coined "dropout distillation", that allows us to train a predictor in a way to better approximate the intractable, but preferable, averaging process, while keeping under control its computational efficiency. We are thus able to construct models that are as efficient as standard dropout, or even more efficient, while being more accurate. Experiments on standard benchmark datasets demonstrate the validity of our method, yielding consistent improvements over conventional dropout.
引用
收藏
页数:9
相关论文
共 50 条
  • [1] Self-knowledge distillation via dropout
    Lee, Hyoje
    Park, Yeachan
    Seo, Hyun
    Kang, Myungjoo
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2023, 233
  • [2] A dynamic dropout self-distillation method for object segmentation
    Chen, Lei
    Cao, Tieyong
    Zheng, Yunfei
    Wang, Yang
    Zhang, Bo
    Yang, Jibin
    COMPLEX & INTELLIGENT SYSTEMS, 2025, 11 (01)
  • [3] Cross-Modal Knowledge Distillation with Dropout-Based Confidence
    Cho, Won Ik
    Kim, Jeunghun
    Kim, Nam Soo
    PROCEEDINGS OF 2022 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC), 2022, : 653 - 657
  • [4] β-Dropout: A Unified Dropout
    Liu, Lei
    Luo, Yuhao
    Shen, Xu
    Sun, Mingzhai
    Li, Bin
    IEEE ACCESS, 2019, 7 : 36140 - 36153
  • [5] Cons-KD: Dropout-Robust Knowledge Distillation for CTC-Based Automatic Speech Recognition
    Won Yoon, Ji
    Lee, Hyeonseung
    Yeon Kang, Ju
    Soo Kim, Nam
    IEEE ACCESS, 2024, 12 : 131136 - 131146
  • [6] DROPOUT
    CLAUS, EP
    AMERICAN JOURNAL OF PHARMACEUTICAL EDUCATION, 1966, 30 (02) : 267 - &
  • [7] THE DROPOUT
    Walters, Ben
    SIGHT AND SOUND, 2022, 32 (04): : 91 - 91
  • [8] The Dropout
    Smith, Judith E.
    Showalter, Michael
    Gregorini, Francesca
    Watson, Erica
    JOURNAL OF AMERICAN HISTORY, 2022, 109 (03) : 724 - 727
  • [9] Dropout Rates And Reasons For Dropout In Patients On Clozapine
    Mishra, Eepsita
    Grover, Sandeep
    Chakrabarti, Subho
    INDIAN JOURNAL OF PSYCHIATRY, 2022, 64
  • [10] Dropout intention: a valid predictor of actual dropout?
    Findeisen, Stefanie
    Brodsky, Alexander
    Michaelis, Christian
    Schimmelpenningh, Beatrice
    Seifried, Juergen
    EMPIRICAL RESEARCH IN VOCATIONAL EDUCATION AND TRAINING, 2024, 16 (01)