Dropout Rademacher complexity of deep neural networks

被引:51
|
作者
Gao, Wei [1 ,2 ]
Zhou, Zhi-Hua [1 ,2 ]
机构
[1] Nanjing Univ, Natl Key Lab Novel Software Technol, Nanjing 210023, Jiangsu, Peoples R China
[2] Nanjing Univ, Collaborat Innovat Ctr Novel Software Technol & I, Nanjing 210023, Jiangsu, Peoples R China
基金
中国国家自然科学基金;
关键词
artificial intelligence; machine learning; deep learning; dropout; Rademacher complexity; BOUNDS;
D O I
10.1007/s11432-015-5470-z
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Great successes of deep neural networks have been witnessed in various real applications. Many algorithmic and implementation techniques have been developed; however, theoretical understanding of many aspects of deep neural networks is far from clear. A particular interesting issue is the usefulness of dropout, which was motivated from the intuition of preventing complex co-adaptation of feature detectors. In this paper, we study the Rademacher complexity of different types of dropouts, and our theoretical results disclose that for shallow neural networks (with one or none hidden layer) dropout is able to reduce the Rademacher complexity in polynomial, whereas for deep neural networks it can amazingly lead to an exponential reduction.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] Dropout Rademacher complexity of deep neural networks
    Wei GAO
    Zhi-Hua ZHOU
    [J]. Science China(Information Sciences), 2016, 59 (07) : 173 - 184
  • [2] Dropout Rademacher complexity of deep neural networks
    Wei Gao
    Zhi-Hua Zhou
    [J]. Science China Information Sciences, 2016, 59
  • [3] Rademacher dropout: An adaptive dropout for deep neural network via optimizing generalization gap
    Wang, Haotian
    Yang, Wenjing
    Zhao, Zhenyu
    Luo, Tingjin
    Wang, Ji
    Tang, Yuhua
    [J]. NEUROCOMPUTING, 2019, 357 : 177 - 187
  • [4] Selective Dropout for Deep Neural Networks
    Barrow, Erik
    Eastwood, Mark
    Jayne, Chrisina
    [J]. NEURAL INFORMATION PROCESSING, ICONIP 2016, PT III, 2016, 9949 : 519 - 528
  • [5] Variational Dropout Sparsifies Deep Neural Networks
    Molchanov, Dmitry
    Ashukha, Arsenii
    Vetrov, Dmitry
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 70, 2017, 70
  • [6] Regularization of deep neural networks with spectral dropout
    Khan, Salman H.
    Hayat, Munawar
    Porikli, Fatih
    [J]. NEURAL NETWORKS, 2019, 110 : 82 - 90
  • [7] Jumpout : Improved Dropout for Deep Neural Networks with ReLUs
    Wang, Shengjie
    Zhou, Tianyi
    Bilmes, Jeff A.
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [8] Excitation Dropout: Encouraging Plasticity in Deep Neural Networks
    Andrea Zunino
    Sarah Adel Bargal
    Pietro Morerio
    Jianming Zhang
    Stan Sclaroff
    Vittorio Murino
    [J]. International Journal of Computer Vision, 2021, 129 : 1139 - 1152
  • [9] Excitation Dropout: Encouraging Plasticity in Deep Neural Networks
    Zunino, Andrea
    Bargal, Sarah Adel
    Morerio, Pietro
    Zhang, Jianming
    Sclaroff, Stan
    Murino, Vittorio
    [J]. INTERNATIONAL JOURNAL OF COMPUTER VISION, 2021, 129 (04) : 1139 - 1152
  • [10] Dropout with Tabu Strategy for Regularizing Deep Neural Networks
    Ma, Zongjie
    Sattar, Abdul
    Zhou, Jun
    Chen, Qingliang
    Su, Kaile
    [J]. COMPUTER JOURNAL, 2020, 63 (07): : 1031 - 1038