Shakedrop Regularization for Deep Residual Learning

被引:68
|
作者
Yamada, Yoshihiro [1 ]
Iwamura, Masakazu [1 ]
Akiba, Takuya [2 ]
Kise, Koichi [1 ]
机构
[1] Osaka Prefecture Univ, Grad Sch Engn, Osaka 5998531, Japan
[2] Preferred Networks Inc, Chiyoda Ku, Tokyo 1000004, Japan
来源
IEEE ACCESS | 2019年 / 7卷
关键词
Computer vision; image classification; neural networks;
D O I
10.1109/ACCESS.2019.2960566
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Overfitting is a crucial problem in deep neural networks, even in the latest network architectures. In this paper, to relieve the overfitting effect of ResNet and its improvements (i.e., Wide ResNet, PyramidNet, and ResNeXt), we propose a new regularization method called ShakeDrop regularization. ShakeDrop is inspired by Shake-Shake, which is an effective regularization method, but can be applied to ResNeXt only. ShakeDrop is more effective than Shake-Shake and can be applied not only to ResNeXt but also ResNet, Wide ResNet, and PyramidNet. An important key is to achieve stability of training. Because effective regularization often causes unstable training, we introduce a training stabilizer, which is an unusual use of an existing regularizer. Through experiments under various conditions, we demonstrate the conditions under which ShakeDrop works well.
引用
收藏
页码:186126 / 186136
页数:11
相关论文
共 50 条
  • [41] Loan Default Prediction with Deep Learning and Muddling Label Regularization
    Jiang, Weiwei
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2022, E105D (07) : 1340 - 1342
  • [42] Deep Residual Reinforcement Learning (Extended Abstract)
    Zhang, Shangtong
    Boehmer, Wendelin
    Whiteson, Shimon
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 4869 - 4873
  • [43] HIERARCHICAL QUADRUPLET NET FOR DEEP METRIC LEARNING AND NETWORK REGULARIZATION
    Zheng, Su
    Tang, Wenqi
    He, Zicheng
    Chen, Jialin
    Wang, Lingli
    Zhou, Xuegong
    Feng, Zhi-Hua
    2019 CHINA SEMICONDUCTOR TECHNOLOGY INTERNATIONAL CONFERENCE (CSTIC), 2019,
  • [44] Deep Into the Domain Shift: Transfer Learning Through Dependence Regularization
    Ma, Shumin
    Yuan, Zhiri
    Wu, Qi
    Huang, Yiyan
    Hu, Xixu
    Leung, Cheuk Hang
    Wang, Dongdong
    Huang, Zhixiang
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (10) : 14409 - 14423
  • [45] Learning Deep Networks from Noisy Labels with Dropout Regularization
    Jindal, Ishan
    Nokleby, Matthew
    Chen, Xuewen
    2016 IEEE 16TH INTERNATIONAL CONFERENCE ON DATA MINING (ICDM), 2016, : 967 - 972
  • [46] Deep learning of multimodal networks with topological regularization for drug repositioning
    Ohnuki, Yuto
    Akiyama, Manato
    Sakakibara, Yasubumi
    JOURNAL OF CHEMINFORMATICS, 2024, 16 (01):
  • [47] Learning deep hierarchical and temporal recurrent neural networks with residual learning
    Tehseen Zia
    Assad Abbas
    Usman Habib
    Muhammad Sajid Khan
    International Journal of Machine Learning and Cybernetics, 2020, 11 : 873 - 882
  • [48] Deep Residual Learning With Dilated Causal Convolution Extreme Learning Machine
    Sasou, Akira
    IEEE ACCESS, 2021, 9 : 165708 - 165718
  • [49] Learning deep hierarchical and temporal recurrent neural networks with residual learning
    Zia, Tehseen
    Abbas, Assad
    Habib, Usman
    Khan, Muhammad Sajid
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2020, 11 (04) : 873 - 882
  • [50] REGULARIZATION OF INVERSE PROBLEMS: DEEP EQUILIBRIUM MODELS VERSUS BILEVEL LEARNING
    Riccio, Danilo
    Ehrhardt, Matthias j.
    Benning, Martin
    NUMERICAL ALGEBRA CONTROL AND OPTIMIZATION, 2023,