Automatic Channel Pruning with Hyper-parameter Search and Dynamic Masking

被引:3
|
作者
Li, Baopu [1 ]
Fan, Yanwen [2 ]
Pan, Zhihong [1 ]
Bian, Yuchen [3 ]
Zhang, Gang [2 ]
机构
[1] Baidu USA LLC, Sunnyvale, CA 94089 USA
[2] Baidu Inc, VIS Dept, Beijing, Peoples R China
[3] Baidu Res, Beijing, Peoples R China
关键词
Model compression; Network pruning; Auto ML;
D O I
10.1145/3474085.3475370
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Modern deep neural network models tend to be large and computationally intensive. One typical solution to this issue is model pruning. However, most current model pruning algorithms depend on hand crafted rules or need to input the pruning ratio beforehand. To overcome this problem, we propose a learning based automatic channel pruning algorithm for deep neural network, which is inspired by recent automatic machine learning (Auto ML). A two objectives' pruning problem that aims for the weights and the remaining channels for each layer is first formulated. An alternative optimization approach is then proposed to derive the channel numbers and weights simultaneously. In the process of pruning, we utilize a searchable hyper-parameter, remaining ratio, to denote the number of channels in each convolution layer, and then a dynamic masking process is proposed to describe the corresponding channel evolution. To adjust the trade-off between accuracy of a model and the pruning ratio of floating point operations, a new loss function is further introduced. Extensive experimental results on benchmark datasets demonstrate that our scheme achieves competitive results for neural network pruning.
引用
收藏
页码:2121 / 2129
页数:9
相关论文
共 50 条
  • [1] Random search for hyper-parameter optimization
    Département D'Informatique et de Recherche Opérationnelle, Université de Montréal, Montréal, QC, H3C 3J7, Canada
    J. Mach. Learn. Res., (281-305):
  • [2] Random Search for Hyper-Parameter Optimization
    Bergstra, James
    Bengio, Yoshua
    JOURNAL OF MACHINE LEARNING RESEARCH, 2012, 13 : 281 - 305
  • [3] Total Variation with Automatic Hyper-Parameter Estimation
    Nascimento, Jacinto
    Sanches, Joao
    2008 30TH ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY, VOLS 1-8, 2008, : 443 - +
  • [4] Automatic CNN Compression Based on Hyper-parameter Learning
    Tian, Nannan
    Liu, Yong
    Wang, Weiping
    Meng, Dan
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [5] Effectiveness of Random Search in SVM hyper-parameter tuning
    Mantovani, Rafael G.
    Rossi, Andre L. D.
    Vanschoren, Joaquin
    Bischl, Bernd
    de Carvalho, Andre C. P. L. F.
    2015 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2015,
  • [6] A novel automatic hyper-parameter estimation for penalized PET reconstruction
    Kim, Kyungsang
    Li, Quanzheng
    2019 IEEE NUCLEAR SCIENCE SYMPOSIUM AND MEDICAL IMAGING CONFERENCE (NSS/MIC), 2019,
  • [7] Experienced Optimization with Reusable Directional Model for Hyper-Parameter Search
    Hu, Yi-Qi
    Yu, Yang
    Zhou, Zhi-Hua
    PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2018, : 2276 - 2282
  • [8] Automatic Hyper-parameter Tuning for Soft Sensor Modeling based on Dynamic Deep Neural Network
    Wang, Kangcheng
    Shang, Chao
    Yang, Fan
    Jiang, Yongheng
    Huang, Dexian
    2017 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2017, : 989 - 994
  • [9] KGTuner: Efficient Hyper-parameter Search for Knowledge Graph Learning
    Zhang, Yongqi
    Zhou, Zhanke
    Yao, Quanming
    Li, Yong
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 2715 - 2735
  • [10] Quadratic optimization for the hyper-parameter based on maximum entropy search
    Li, Yuqi
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2023, 45 (03) : 4991 - 5006