Shakedrop Regularization for Deep Residual Learning

被引:68
|
作者
Yamada, Yoshihiro [1 ]
Iwamura, Masakazu [1 ]
Akiba, Takuya [2 ]
Kise, Koichi [1 ]
机构
[1] Osaka Prefecture Univ, Grad Sch Engn, Osaka 5998531, Japan
[2] Preferred Networks Inc, Chiyoda Ku, Tokyo 1000004, Japan
来源
IEEE ACCESS | 2019年 / 7卷
关键词
Computer vision; image classification; neural networks;
D O I
10.1109/ACCESS.2019.2960566
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Overfitting is a crucial problem in deep neural networks, even in the latest network architectures. In this paper, to relieve the overfitting effect of ResNet and its improvements (i.e., Wide ResNet, PyramidNet, and ResNeXt), we propose a new regularization method called ShakeDrop regularization. ShakeDrop is inspired by Shake-Shake, which is an effective regularization method, but can be applied to ResNeXt only. ShakeDrop is more effective than Shake-Shake and can be applied not only to ResNeXt but also ResNet, Wide ResNet, and PyramidNet. An important key is to achieve stability of training. Because effective regularization often causes unstable training, we introduce a training stabilizer, which is an unusual use of an existing regularizer. Through experiments under various conditions, we demonstrate the conditions under which ShakeDrop works well.
引用
收藏
页码:186126 / 186136
页数:11
相关论文
共 50 条
  • [31] Combining Explicit and Implicit Regularization for Efficient Learning in Deep Networks
    Zhao, Dan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [32] Deep Residual Feature Learning for Action Prediction
    Guo, Shuangshuang
    Qing, Laiyun
    Miao, Jun
    Duan, Lijuan
    2018 IEEE FOURTH INTERNATIONAL CONFERENCE ON MULTIMEDIA BIG DATA (BIGMM), 2018,
  • [33] Deep Residual Learning for Facial Emotion Recognition
    Mishra, Sagar
    Joshi, Basanta
    Paudyal, Rajendra
    Chaulagain, Duryodhan
    Shakya, Subarna
    MOBILE COMPUTING AND SUSTAINABLE INFORMATICS, 2022, 68 : 301 - 313
  • [34] Deep Residual Learning in Spiking Neural Networks
    Fang, Wei
    Yu, Zhaofei
    Chen, Yanqi
    Huang, Tiejun
    Masquelier, Timothee
    Tian, Yonghong
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [35] Deep Residual Learning in the JPEG Transform Domain
    Ehrlich, Max
    Davis, Larry
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 3483 - 3492
  • [36] Invertible Residual Blocks in Deep Learning Networks
    Wang, Ruhua
    An, Senjian
    Liu, Wanquan
    Li, Ling
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (07) : 10167 - 10173
  • [37] Multi-level Distance Regularization for Deep Metric Learning
    Kim, Yonghyun
    Park, Wonpyo
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 1827 - 1835
  • [38] DEEP RESIDUAL LEARNING FOR COMPRESSED SENSING MRI
    Lee, Dongwook
    Yoo, Jaejun
    Ye, Jong Chul
    2017 IEEE 14TH INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING (ISBI 2017), 2017, : 15 - 18
  • [39] Deep Residual Learning for Image Recognition: A Survey
    Shafiq, Muhammad
    Gu, Zhaoquan
    APPLIED SCIENCES-BASEL, 2022, 12 (18):
  • [40] Variational Characterizations of Local Entropy and Heat Regularization in Deep Learning
    Trillos, Nicolas Garcia
    Kaplan, Zachary
    Sanz-Alonso, Daniel
    ENTROPY, 2019, 21 (05)