Shakedrop Regularization for Deep Residual Learning

被引:68
|
作者
Yamada, Yoshihiro [1 ]
Iwamura, Masakazu [1 ]
Akiba, Takuya [2 ]
Kise, Koichi [1 ]
机构
[1] Osaka Prefecture Univ, Grad Sch Engn, Osaka 5998531, Japan
[2] Preferred Networks Inc, Chiyoda Ku, Tokyo 1000004, Japan
来源
IEEE ACCESS | 2019年 / 7卷
关键词
Computer vision; image classification; neural networks;
D O I
10.1109/ACCESS.2019.2960566
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Overfitting is a crucial problem in deep neural networks, even in the latest network architectures. In this paper, to relieve the overfitting effect of ResNet and its improvements (i.e., Wide ResNet, PyramidNet, and ResNeXt), we propose a new regularization method called ShakeDrop regularization. ShakeDrop is inspired by Shake-Shake, which is an effective regularization method, but can be applied to ResNeXt only. ShakeDrop is more effective than Shake-Shake and can be applied not only to ResNeXt but also ResNet, Wide ResNet, and PyramidNet. An important key is to achieve stability of training. Because effective regularization often causes unstable training, we introduce a training stabilizer, which is an unusual use of an existing regularizer. Through experiments under various conditions, we demonstrate the conditions under which ShakeDrop works well.
引用
收藏
页码:186126 / 186136
页数:11
相关论文
共 50 条
  • [1] Adaptive Regularization via Residual Smoothing in Deep Learning Optimization
    Cho, Junghee
    Kwon, Junseok
    Hong, Byung-Woo
    IEEE ACCESS, 2019, 7 : 122889 - 122899
  • [2] Controlling TV Regularization with Deep Learning
    Toth, Balazs
    Toth, Marton
    Kacso, Agota
    Racz, Gergely
    Szirmay-Kalos, Laszlo
    2018 IEEE NUCLEAR SCIENCE SYMPOSIUM AND MEDICAL IMAGING CONFERENCE PROCEEDINGS (NSS/MIC), 2018,
  • [3] Improving Interpretability and Regularization in Deep Learning
    Wu, Chunyang
    Gales, Mark J. F.
    Ragni, Anton
    Karanasou, Penny
    Sim, Khe Chai
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2018, 26 (02) : 256 - 265
  • [4] Regularization by deep learning in signal processing
    Villamarin, Carlos Ramirez
    Suazo, Erwin
    Oraby, Tamer
    SIGNAL IMAGE AND VIDEO PROCESSING, 2024, 18 (05) : 4425 - 4433
  • [5] DEEP INHOMOGENEOUS REGULARIZATION FOR TRANSFER LEARNING
    Wang, Wen
    Zhai, Wei
    Cao, Yang
    2020 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2020, : 221 - 225
  • [6] Regularization Learning Networks: Deep Learning for Tabular Datasets
    Shavitt, Ira
    Segal, Eran
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [7] Guidelines for the Regularization of Gammas in Batch Normalization for Deep Residual Networks
    Kim, Bum Jun
    Choi, Hyeyeon
    Jang, Hyeonah
    Kim, Sang Woo
    ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2024, 15 (03)
  • [8] Flexible Regularization Approaches for Fairness in Deep Learning
    Olfat, Matt
    Mintz, Yonatan
    2020 59TH IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2020, : 3389 - 3394
  • [9] Multistage Fusion and Dissimilarity Regularization for Deep Learning
    Cho, Young-Rae
    Shin, Seungjun
    Yim, Sung-Hyuk
    Cho, Hyun-Woong
    Song, Woo-Jin
    2017 IEEE INTERNATIONAL CONFERENCE ON MULTISENSOR FUSION AND INTEGRATION FOR INTELLIGENT SYSTEMS (MFI), 2017, : 586 - 591
  • [10] Online Deep Equilibrium Learning for Regularization by Denoising
    Liu, Jiaming
    Xu, Xiaojian
    Gan, Weijie
    Shoushtari, Shirin
    Kamilov, Ulugbek S.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,