Dropout Rademacher complexity of deep neural networks

被引:51
|
作者
Gao, Wei [1 ,2 ]
Zhou, Zhi-Hua [1 ,2 ]
机构
[1] Nanjing Univ, Natl Key Lab Novel Software Technol, Nanjing 210023, Jiangsu, Peoples R China
[2] Nanjing Univ, Collaborat Innovat Ctr Novel Software Technol & I, Nanjing 210023, Jiangsu, Peoples R China
基金
中国国家自然科学基金;
关键词
artificial intelligence; machine learning; deep learning; dropout; Rademacher complexity; BOUNDS;
D O I
10.1007/s11432-015-5470-z
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Great successes of deep neural networks have been witnessed in various real applications. Many algorithmic and implementation techniques have been developed; however, theoretical understanding of many aspects of deep neural networks is far from clear. A particular interesting issue is the usefulness of dropout, which was motivated from the intuition of preventing complex co-adaptation of feature detectors. In this paper, we study the Rademacher complexity of different types of dropouts, and our theoretical results disclose that for shallow neural networks (with one or none hidden layer) dropout is able to reduce the Rademacher complexity in polynomial, whereas for deep neural networks it can amazingly lead to an exponential reduction.
引用
收藏
页数:12
相关论文
共 50 条
  • [41] Surprising properties of dropout in deep networks
    Helmbold, David P.
    Long, Philip M.
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2018, 18
  • [42] Universal Approximation in Dropout Neural Networks
    Manita, Oxana A.
    Peletier, Mark A.
    Portegies, Jacobus W.
    Sanders, Jaron
    Senen-Cerda, Albert
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2022, 23
  • [43] Understanding Dropout for Graph Neural Networks
    Shu, Juan
    Xi, Bowei
    Li, Yu
    Wu, Fan
    Kamhoua, Charles
    Ma, Jianzhu
    [J]. COMPANION PROCEEDINGS OF THE WEB CONFERENCE 2022, WWW 2022 COMPANION, 2022, : 1128 - 1138
  • [44] Adversarial Dropout for Recurrent Neural Networks
    Park, Sungrae
    Song, Kyungwoo
    Ji, Mingi
    Lee, Wonsung
    Moon, Il-Chul
    [J]. THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 4699 - 4706
  • [45] Dropout Algorithms for Recurrent Neural Networks
    Watt, Nathan
    du Plessis, Mathys C.
    [J]. PROCEEDINGS OF THE ANNUAL CONFERENCE OF THE SOUTH AFRICAN INSTITUTE OF COMPUTER SCIENTISTS AND INFORMATION TECHNOLOGISTS (SAICSIT 2018), 2018, : 72 - 78
  • [46] GUIDE: Training Deep Graph Neural Networks via Guided Dropout Over Edges
    Wang, Jie
    Liang, Jianqing
    Liang, Jiye
    Yao, Kaixuan
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (04) : 4465 - 4477
  • [47] Reliable Prediction Errors for Deep Neural Networks Using Test-Time Dropout
    Cortes-Ciriano, Isidro
    Bender, Andreas
    [J]. JOURNAL OF CHEMICAL INFORMATION AND MODELING, 2019, 59 (07) : 3330 - 3339
  • [48] Random image frequency aggregation dropout in image classification for deep convolutional neural networks
    Nam, Ju-Hyeon
    Lee, Sang-Chul
    [J]. COMPUTER VISION AND IMAGE UNDERSTANDING, 2023, 232
  • [49] Controlled Dropout: a Different Approach to Using Dropout on Deep Neural Network
    Ko, ByungSoo
    Kim, Han-Gyu
    Oh, Kyo-Joong
    Choi, Ho-Jin
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON BIG DATA AND SMART COMPUTING (BIGCOMP), 2017, : 358 - 362
  • [50] On the Rademacher Complexity of Weighted Automata
    Balle, Borja
    Mohri, Mehryar
    [J]. ALGORITHMIC LEARNING THEORY, ALT 2015, 2015, 9355 : 179 - 193