Structure injected weight normalization for training deep networks

被引:0
|
作者
Xu Yuan
Xiangjun Shen
Sumet Mehta
Teng Li
Shiming Ge
Zhengjun Zha
机构
[1] Jiangsu University,School of Computer Science and Telecommunication Engineering
[2] Anhui University,School of electrical engineering and Automation
[3] Chinese academy of sciences,Institute of information engineering, CAS
[4] University of science and technology of china,School of data science
来源
Multimedia Systems | 2022年 / 28卷
关键词
Weight normalization; Structural learning; Sparsity measurement; Neuron measurement;
D O I
暂无
中图分类号
学科分类号
摘要
Weight normalization (WN) can help to stabilize the distribution of activations over layers, which boost the performance of DNNs in generalization. In this paper, we further propose deep structural weight normalization (DSWN) methods to inject the network structure measurements into the WN to fully acknowledge the data propagation through the neural network. In DSWN, two novel structural measurements are developed to impose regularity on each network weight using different penalty matrices. One is sparsity measurement (DSWN-SM). In this measurement, L1,2 weight regularization is applied in our proposed model to promote competition for features between network weights to obtain a sparsity network and finally prune the network. The other is neuron measurement (DSWN-NM). It uses L2 norm of column weight to scale up or down the importance of each intermediate neuron, which leads to accelerating the speed of network convergence. Extensive experiments on several benchmark image datasets using fully connected network and convolution neural network are performed, and the proposed DSWN-SM and DSWN-NM methods are compared with state-of-the-art sparsity and weight normalization methods. The results show that DSWN-SM can reduce the number of trainable parameters while guaranteeing high accuracy, whereas DSWN-NM can accelerate the convergence while improving the performance of deep networks.
引用
收藏
页码:433 / 444
页数:11
相关论文
共 50 条
  • [1] Structure injected weight normalization for training deep networks
    Yuan, Xu
    Shen, Xiangjun
    Mehta, Sumet
    Li, Teng
    Ge, Shiming
    Zha, Zhengjun
    [J]. MULTIMEDIA SYSTEMS, 2022, 28 (02) : 433 - 444
  • [2] Centered Weight Normalization in Accelerating Training of Deep Neural Networks
    Huang, Lei
    Liu, Xianglong
    Liu, Yang
    Lang, Bo
    Tao, Dacheng
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 2822 - 2830
  • [3] Weight Normalization: A Simple Reparameterization to Accelerate Training of Deep Neural Networks
    Salimans, Tim
    Kingma, Diederik P.
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 29 (NIPS 2016), 2016, 29
  • [4] Is normalization indispensable for training deep neural networks?
    Shao, Jie
    Hu, Kai
    Wang, Changhu
    Xue, Xiangyang
    Raj, Bhiksha
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [5] Scaling-Based Weight Normalization for Deep Neural Networks
    Yuan, Qunyong
    Xiao, Nanfeng
    [J]. IEEE ACCESS, 2019, 7 : 7286 - 7295
  • [6] SPARSE DEEP NEURAL NETWORKS USING L1,∞-WEIGHT NORMALIZATION
    Wen, Ming
    Xu, Yixi
    Zheng, Yunling
    Yang, Zhouwang
    Wang, Xiao
    [J]. STATISTICA SINICA, 2021, 31 (03) : 1397 - 1414
  • [7] Batch Normalization and Dropout Regularization in Training Deep Neural Networks with Label Noise
    Rusiecki, Andrzej
    [J]. INTELLIGENT SYSTEMS DESIGN AND APPLICATIONS, ISDA 2021, 2022, 418 : 57 - 66
  • [8] Online Normalization for Training Neural Networks
    Chiley, Vitaliy
    Sharapov, Ilya
    Kosson, Atli
    Koster, Urs
    Reece, Ryan
    de la Fuente, Sofia Samaniego
    Subbiah, Vishal
    James, Michael
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [9] L1-Norm Batch Normalization for Efficient Training of Deep Neural Networks
    Wu, Shuang
    Li, Guoqi
    Deng, Lei
    Liu, Liu
    Wu, Dong
    Xie, Yuan
    Shi, Luping
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2019, 30 (07) : 2043 - 2051
  • [10] NORMALIZATION EFFECTS ON DEEP NEURAL NETWORKS
    Yu, Jiahui
    Spiliopoulos, Konstantinos
    [J]. FOUNDATIONS OF DATA SCIENCE, 2023, 5 (03): : 389 - 465