A NOVEL LAYERWISE PRUNING METHOD FOR MODEL REDUCTION OF FULLY CONNECTED DEEP NEURAL NETWORKS

被引:0
|
作者
Mauch, Lukas [1 ]
Yang, Bin [1 ]
机构
[1] Univ Stuttgart, Inst Signal Proc & Syst Theory, Stuttgart, Germany
关键词
Deep neural networks; model reduction; pruning; parameter adaptation;
D O I
暂无
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Deep neural networks (DNN) are powerful models for many pattern recognition tasks, yet they tend to have many layers and many neurons resulting in a high computational complexity. This limits their application to high-performance computing platfonns. In order to evaluate a trained DNN on a lowerperformance computing platfonn like a mobile or embedded device, model reduction techniques which shrink the network size and reduce the number of parameters without considerable perfonnance degradation perfonnance are highly desirable. In this paper, we start with a trained fully connected DNN and show how to reduce the network complexity by a novellayerwise pruning method. We show that if some neurons are pruned and the remaining parameters (weights and biases) are adapted correspondingly to correct the errors introduced by pruning, the model reduction can be done almost without performance loss. The main contribution of our pruning method is a closed-fonn solution that only makes use of the first and second order moments of the layer outputs and, therefore, only needs unlabeled data. Using three benchmark datasets, we compare our pruning method with the low-rank approximation approach.
引用
收藏
页码:2382 / 2386
页数:5
相关论文
共 50 条
  • [1] A Clustering Method for Pruning Fully Connected Neural Network
    Li, Gang
    Qian, Xingsan
    Ye, Chunming
    Zhao, Lin
    [J]. ADVANCED RESEARCH ON INDUSTRY, INFORMATION SYSTEMS AND MATERIAL ENGINEERING, PTS 1-7, 2011, 204-210 : 600 - 603
  • [2] DEEP LEARNING BASED METHOD FOR PRUNING DEEP NEURAL NETWORKS
    Li, Lianqiang
    Zhu, Jie
    Sun, Ming-Ting
    [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA & EXPO WORKSHOPS (ICMEW), 2019, : 312 - 317
  • [3] A New Pruning Method to Train Deep Neural Networks
    Guo, Haonan
    Ren, Xudie
    Li, Shenghong
    [J]. COMMUNICATIONS, SIGNAL PROCESSING, AND SYSTEMS, 2018, 423 : 767 - 775
  • [4] Anonymous Model Pruning for Compressing Deep Neural Networks
    Zhang, Lechun
    Chen, Guangyao
    Shi, Yemin
    Zhang, Quan
    Tan, Mingkui
    Wang, Yaowei
    Tian, Yonghong
    Huang, Tiejun
    [J]. THIRD INTERNATIONAL CONFERENCE ON MULTIMEDIA INFORMATION PROCESSING AND RETRIEVAL (MIPR 2020), 2020, : 161 - 164
  • [5] Automatic model selection for fully connected neural networks
    Laredo D.
    Ma S.F.
    Leylaz G.
    Schütze O.
    Sun J.-Q.
    [J]. International Journal of Dynamics and Control, 2020, 8 (04) : 1063 - 1079
  • [6] A novel and efficient model pruning method for deep convolutional neural networks by evaluating the direct and indirect effects of filters
    Zheng, Yongbin
    Sun, Peng
    Ren, Qian
    Xu, Wanying
    Zhu, Di
    [J]. NEUROCOMPUTING, 2024, 569
  • [7] A GA-Based Pruning Fully Connected Network for Tuned Connections in Deep Networks
    Khatami, Amin
    Kebria, Parham M.
    Jalali, Seyed Mohammad Jafar
    Khosravi, Abbas
    Nazari, Asef
    Shamszadeh, Marjan
    Thanh Nguyen
    Nahavandi, Saeid
    [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN AND CYBERNETICS (SMC), 2019, : 3492 - 3497
  • [8] Reduction of Fully Connected Parameters by Pruning and Matrix Factorization
    Nishino, Shunsuke
    Sudo, Kyoko
    [J]. Journal of the Institute of Image Electronics Engineers of Japan, 2023, 52 (04): : 527 - 530
  • [9] EasiEdge: A Novel Global Deep Neural Networks Pruning Method for Efficient Edge Computing
    Yu, Fang
    Cui, Li
    Wang, Pengcheng
    Han, Chuanqi
    Huang, Ruoran
    Huang, Xi
    [J]. IEEE INTERNET OF THINGS JOURNAL, 2021, 8 (03): : 1259 - 1271
  • [10] A Novel Clustering-Based Filter Pruning Method for Efficient Deep Neural Networks
    Wei, Xiaohui
    Shen, Xiaoxian
    Zhou, Changbao
    Yue, Hengshan
    [J]. ALGORITHMS AND ARCHITECTURES FOR PARALLEL PROCESSING, ICA3PP 2020, PT II, 2020, 12453 : 245 - 258