Transformed l1 regularization for learning sparse deep neural networks

被引:58
|
作者
Ma, Rongrong [1 ]
Miao, Jianyu [2 ]
Niu, Lingfeng [3 ]
Zhang, Peng [4 ]
机构
[1] Univ Chinese Acad Sci, Sch Math Sci, Beijing 100049, Peoples R China
[2] Henan Univ Technol, Coll Informat Sci & Engn, Zhengzhou 450001, Henan, Peoples R China
[3] Univ Chinese Acad Sci, Sch Econ & Management, Beijing 100190, Peoples R China
[4] Ant Financial Serv Grp, Hangzhou 310012, Zhejiang, Peoples R China
基金
中国国家自然科学基金;
关键词
Deep neural networks; Non-convex regularization; Transformed l(1); Group sparsity; VARIABLE SELECTION; REPRESENTATION; MINIMIZATION; DROPOUT;
D O I
10.1016/j.neunet.2019.08.015
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep Neural Networks (DNNs) have achieved extraordinary success in numerous areas. However, DNNs often carry a large number of weight parameters, leading to the challenge of heavy memory and computation costs. Overfitting is another challenge for DNNs when the training data are insufficient. These challenges severely hinder the application of DNNs in resource-constrained platforms. In fact, many network weights are redundant and can be removed from the network without much loss of performance. In this paper, we introduce a new non-convex integrated transformed l(1) regularizer to promote sparsity for DNNs, which removes redundant connections and unnecessary neurons simultaneously. Specifically, we apply the transformed l(1) regularizer to the matrix space of network weights and utilize it to remove redundant connections. Besides, group sparsity is integrated to remove unnecessary neurons. An efficient stochastic proximal gradient algorithm is presented to solve the new model. To the best of our knowledge, this is the first work to develop a non-convex regularizer in sparse optimization based method to simultaneously promote connection-level and neuron-level sparsity for DNNs. Experiments on public datasets demonstrate the effectiveness of the proposed method. (C) 2019 Elsevier Ltd. All rights reserved.
引用
收藏
页码:286 / 298
页数:13
相关论文
共 50 条
  • [1] Compact Deep Neural Networks with l1,1 and l1,2 Regularization
    Ma, Rongrong
    Niu, Lingfeng
    [J]. 2018 18TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS (ICDMW), 2018, : 1248 - 1254
  • [2] SPARSE DEEP NEURAL NETWORKS USING L1,∞-WEIGHT NORMALIZATION
    Wen, Ming
    Xu, Yixi
    Zheng, Yunling
    Yang, Zhouwang
    Wang, Xiao
    [J]. STATISTICA SINICA, 2021, 31 (03) : 1397 - 1414
  • [3] A Simple Neural Network for Sparse Optimization With l1 Regularization
    Ma, Litao
    Bian, Wei
    [J]. IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, 2021, 8 (04): : 3430 - 3442
  • [4] Sparse smooth group L0°L1/2 regularization method for convolutional neural networks
    Quasdane, Mohamed
    Ramchoun, Hassan
    Masrour, Tawfik
    [J]. KNOWLEDGE-BASED SYSTEMS, 2024, 284
  • [5] Learning Optimized Structure of Neural Networks by Hidden Node Pruning With L1 Regularization
    Xie, Xuetao
    Zhang, Huaqing
    Wang, Junze
    Chang, Qin
    Wang, Jian
    Pal, Nikhil R.
    [J]. IEEE TRANSACTIONS ON CYBERNETICS, 2020, 50 (03) : 1333 - 1346
  • [7] Group sparse regularization for deep neural networks
    Scardapane, Simone
    Comminiello, Danilo
    Hussain, Amir
    Uncini, Aurelio
    [J]. NEUROCOMPUTING, 2017, 241 : 81 - 89
  • [8] Sparse synthesis regularization with deep neural networks
    Obmann, Daniel
    Schwab, Johannes
    Haltmeier, Markus
    [J]. 2019 13TH INTERNATIONAL CONFERENCE ON SAMPLING THEORY AND APPLICATIONS (SAMPTA), 2019,
  • [9] SPARSE REPRESENTATION LEARNING OF DATA BY AUTOENCODERS WITH L1/2 REGULARIZATION
    Li, F.
    Zurada, J. M.
    Wu, W.
    [J]. NEURAL NETWORK WORLD, 2018, 28 (02) : 133 - 147
  • [10] A novel method for financial distress prediction based on sparse neural networks with L1/2 regularization
    Chen, Ying
    Guo, Jifeng
    Huang, Junqin
    Lin, Bin
    [J]. INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2022, 13 (07) : 2089 - 2103