Dropout Rademacher complexity of deep neural networks

被引:51
|
作者
Gao, Wei [1 ,2 ]
Zhou, Zhi-Hua [1 ,2 ]
机构
[1] Nanjing Univ, Natl Key Lab Novel Software Technol, Nanjing 210023, Jiangsu, Peoples R China
[2] Nanjing Univ, Collaborat Innovat Ctr Novel Software Technol & I, Nanjing 210023, Jiangsu, Peoples R China
基金
中国国家自然科学基金;
关键词
artificial intelligence; machine learning; deep learning; dropout; Rademacher complexity; BOUNDS;
D O I
10.1007/s11432-015-5470-z
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Great successes of deep neural networks have been witnessed in various real applications. Many algorithmic and implementation techniques have been developed; however, theoretical understanding of many aspects of deep neural networks is far from clear. A particular interesting issue is the usefulness of dropout, which was motivated from the intuition of preventing complex co-adaptation of feature detectors. In this paper, we study the Rademacher complexity of different types of dropouts, and our theoretical results disclose that for shallow neural networks (with one or none hidden layer) dropout is able to reduce the Rademacher complexity in polynomial, whereas for deep neural networks it can amazingly lead to an exponential reduction.
引用
收藏
页数:12
相关论文
共 50 条
  • [31] A Review on Dropout Regularization Approaches for Deep Neural Networks within the Scholarly Domain
    Salehin, Imrus
    Kang, Dae-Ki
    [J]. ELECTRONICS, 2023, 12 (14)
  • [32] ISING-DROPOUT: A REGULARIZATION METHOD FOR TRAINING AND COMPRESSION OF DEEP NEURAL NETWORKS
    Salehinejad, Hojjat
    Valaee, Shahrokh
    [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 3602 - 3606
  • [33] Dropout in Neural Networks Simulates the Paradoxical Effects of Deep Brain Stimulation on Memory
    Tan, Shawn Zheng Kai
    Du, Richard
    Perucho, Jose Angelo Udal
    Chopra, Shauhrat S.
    Vardhanabhuti, Varut
    Lim, Lee Wei
    [J]. FRONTIERS IN AGING NEUROSCIENCE, 2020, 12
  • [34] Batch Normalization and Dropout Regularization in Training Deep Neural Networks with Label Noise
    Rusiecki, Andrzej
    [J]. INTELLIGENT SYSTEMS DESIGN AND APPLICATIONS, ISDA 2021, 2022, 418 : 57 - 66
  • [35] Online Arabic Handwriting Recognition with Dropout applied in Deep Recurrent Neural Networks
    Maalej, Rania
    Tagougui, Najiba
    Kherallah, Monji
    [J]. PROCEEDINGS OF 12TH IAPR WORKSHOP ON DOCUMENT ANALYSIS SYSTEMS, (DAS 2016), 2016, : 417 - 421
  • [36] Rademacher margin complexity
    Wang, Liwei
    Feng, Jufu
    [J]. LEARNING THEORY, PROCEEDINGS, 2007, 4539 : 620 - +
  • [37] A Deep Connection Between the Vapnik-Chervonenkis Entropy and the Rademacher Complexity
    Anguita, Davide
    Ghio, Alessandro
    Oneto, Luca
    Ridella, Sandro
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2014, 25 (12) : 2202 - 2211
  • [38] Interpretability vs. Complexity: The Friction in Deep Neural Networks
    Amorim, Jose P.
    Abreu, Pedro H.
    Reyes, Mauricio
    Santos, Joao
    [J]. 2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [39] ANNEALED DROPOUT TRAINING OF DEEP NETWORKS
    Rennie, Steven J.
    Goel, Vaibhava
    Thomas, Samuel
    [J]. 2014 IEEE WORKSHOP ON SPOKEN LANGUAGE TECHNOLOGY SLT 2014, 2014, : 159 - 164
  • [40] A Dropout Distribution Model on Deep Networks
    Li, Fengqi
    Yang, Helin
    [J]. EIGHTH INTERNATIONAL CONFERENCE ON DIGITAL IMAGE PROCESSING (ICDIP 2016), 2016, 10033