Enhance the Performance of Deep Neural Networks via L2 Regularization on the Input of Activations

被引:0
|
作者
Guang Shi
Jiangshe Zhang
Huirong Li
Changpeng Wang
机构
[1] Xi’an Jiaotong University,School of Mathematics and Statistics
[2] Shangluo University,Department of Mathematics and Computer Application
[3] Chang’an University,School of Mathematics and Information Science
来源
Neural Processing Letters | 2019年 / 50卷
关键词
Neural networks; ReLU; Saturation phenomenon; L2 regularization;
D O I
暂无
中图分类号
学科分类号
摘要
Deep neural networks (DNNs) are witnessing increasing attention in machine learning. However, the information propagation is becoming increasingly difficult as the networks get deeper, which makes the optimization of DNN extremely hard. One reason of this difficulty is saturation of hidden units. In this paper, we propose a novel methodology named RegA to decrease the influences of saturation on ReLU-DNNs (DNNs with ReLU). Instead of changing the activation functions or the initialization strategy, our methodology explicitly encourage the pre-activation to be out of the saturation region. Specifically, we add an auxiliary objective induced by L2-norm of the pre-activation values to the optimization problem. The auxiliary objective could help to active more units and promote effective information propagation in ReLU-DNNs. By conducting experiments on several large-scale real datasets, we demonstrate better representations could be learned by using RegA and the method help ReLU-DNNs get better performance on convergence and accuracy.
引用
收藏
页码:57 / 75
页数:18
相关论文
共 50 条
  • [1] Enhance the Performance of Deep Neural Networks via L2 Regularization on the Input of Activations
    Shi, Guang
    Zhang, Jiangshe
    Li, Huirong
    Wang, Changpeng
    NEURAL PROCESSING LETTERS, 2019, 50 (01) : 57 - 75
  • [2] On the training dynamics of deep networks with L2 regularization
    Lewkowycz, Aitor
    Gur-Ari, Guy
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [3] Deep neural networks with L1 and L2 regularization for high dimensional corporate credit risk prediction
    Yang, Mei
    Lim, Ming K.
    Qu, Yingchi
    Li, Xingzhi
    Ni, Du
    EXPERT SYSTEMS WITH APPLICATIONS, 2023, 213
  • [4] A Hybrid Improved Neural Networks Algorithm Based on L2 and Dropout Regularization
    Xie, Xiaoyun
    Xie, Ming
    Moshayedi, Ata Jahangir
    Skandari, Mohammad Hadi Noori
    MATHEMATICAL PROBLEMS IN ENGINEERING, 2022, 2022
  • [5] A Hybrid Improved Neural Networks Algorithm Based on L2 and Dropout Regularization
    Xie, Xiaoyun
    Xie, Ming
    Moshayedi, Ata Jahangir
    Skandari, Mohammad Hadi Noori
    MATHEMATICAL PROBLEMS IN ENGINEERING, 2022, 2022
  • [6] Towards Robustness of Deep Neural Networks via Regularization
    Li, Yao
    Min, Martin Renqiang
    Lee, Thomas
    Yu, Wenchao
    Kruus, Erik
    Wang, Wei
    Hsieh, Cho-Jui
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 7476 - 7485
  • [7] L2 regularized deep convolutional neural networks for fire detection
    Roy, Sanjiban Sekhar
    Goti, Vatsal
    Sood, Aditya
    Roy, Harsh
    Gavrila, Tania
    Floroian, Dan
    Paraschiv, Nicolae
    Mohammadi-Ivatloo, Behnam
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2022, 43 (02) : 1799 - 1810
  • [8] Enhance the Hidden Structure of Deep Neural Networks by Double Laplacian Regularization
    Fan, Yetian
    Yang, Wenyu
    Song, Bo
    Yan, Peilei
    Kang, Xiaoning
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2023, 70 (08) : 3114 - 3118
  • [9] A smoothed monotonic regression via L2 regularization
    Oleg Sysoev
    Oleg Burdakov
    Knowledge and Information Systems, 2019, 59 : 197 - 218
  • [10] A smoothed monotonic regression via L2 regularization
    Sysoev, Oleg
    Burdakov, Oleg
    KNOWLEDGE AND INFORMATION SYSTEMS, 2019, 59 (01) : 197 - 218