Shakedrop Regularization for Deep Residual Learning

被引:68
|
作者
Yamada, Yoshihiro [1 ]
Iwamura, Masakazu [1 ]
Akiba, Takuya [2 ]
Kise, Koichi [1 ]
机构
[1] Osaka Prefecture Univ, Grad Sch Engn, Osaka 5998531, Japan
[2] Preferred Networks Inc, Chiyoda Ku, Tokyo 1000004, Japan
来源
IEEE ACCESS | 2019年 / 7卷
关键词
Computer vision; image classification; neural networks;
D O I
10.1109/ACCESS.2019.2960566
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Overfitting is a crucial problem in deep neural networks, even in the latest network architectures. In this paper, to relieve the overfitting effect of ResNet and its improvements (i.e., Wide ResNet, PyramidNet, and ResNeXt), we propose a new regularization method called ShakeDrop regularization. ShakeDrop is inspired by Shake-Shake, which is an effective regularization method, but can be applied to ResNeXt only. ShakeDrop is more effective than Shake-Shake and can be applied not only to ResNeXt but also ResNet, Wide ResNet, and PyramidNet. An important key is to achieve stability of training. Because effective regularization often causes unstable training, we introduce a training stabilizer, which is an unusual use of an existing regularizer. Through experiments under various conditions, we demonstrate the conditions under which ShakeDrop works well.
引用
收藏
页码:186126 / 186136
页数:11
相关论文
共 50 条
  • [21] Implicit Regularization in Deep Learning May Not Be Explainable by Norms
    Razin, Noam
    Cohen, Nadav
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [22] Learning Credible Deep Neural Networks with Rationale Regularization
    Du, Mengnan
    Liu, Ninghao
    Yang, Fan
    Hu, Xia
    2019 19TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM 2019), 2019, : 150 - 159
  • [23] Revisiting Consistency Regularization for Deep Partial Label Learning
    Wu, Dong-Dong
    Wang, Deng-Bao
    Zhang, Min-Ling
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [24] Deep Learning and Regularization Algorithms for Malicious Code Classification
    Wang, Haojun
    Long, Haixia
    Wang, Ailan
    Liu, Tianyue
    Fu, Haiyan
    IEEE ACCESS, 2021, 9 : 91512 - 91523
  • [25] Fault-Tolerant Deep Learning Using Regularization
    Joardar, Biresh Kumar
    Arka, Aqeeb Iqbal
    Doppa, Janardhan Rao
    Pande, Partha Pratim
    2022 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER AIDED DESIGN, ICCAD, 2022,
  • [26] Feature Regularization and Deep Learning for Human Resource Recommendation
    Wang, Haoxiang
    Liang, Guihuang
    Zhang, Xingming
    IEEE ACCESS, 2018, 6 : 39415 - 39421
  • [27] Improving Regularization of Deep Learning Models in Fundus Analysis
    Hsu, Wei-Wen
    Chang, Yao-Chung
    Lee, Wei-Min
    Huang, Yu-Chuan
    Lu, Da-Wen
    2023 ASIA PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE, APSIPA ASC, 2023, : 898 - 901
  • [28] Rectilinear Building Footprint Regularization Using Deep Learning
    Schuegraf, Philipp
    Li, Zhixin
    Tian, Jiaojiao
    Shan, Jie
    Bittner, Ksenia
    ISPRS ANNALS OF THE PHOTOGRAMMETRY, REMOTE SENSING AND SPATIAL INFORMATION SCIENCES: VOLUME X-2-2024, 2024, : 217 - 222
  • [29] SHADE: INFORMATION-BASED REGULARIZATION FOR DEEP LEARNING
    Blot, Michael
    Robert, Thomas
    Thome, Nicolas
    Cord, Matthieu
    2018 25TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2018, : 813 - 817
  • [30] Deep Learning Meets Sparse Regularization: A signal processing perspective
    Parhi R.
    Nowak R.D.
    IEEE Signal Processing Magazine, 2023, 40 (06) : 63 - 74