Automatic Channel Pruning with Hyper-parameter Search and Dynamic Masking

被引:3
|
作者
Li, Baopu [1 ]
Fan, Yanwen [2 ]
Pan, Zhihong [1 ]
Bian, Yuchen [3 ]
Zhang, Gang [2 ]
机构
[1] Baidu USA LLC, Sunnyvale, CA 94089 USA
[2] Baidu Inc, VIS Dept, Beijing, Peoples R China
[3] Baidu Res, Beijing, Peoples R China
关键词
Model compression; Network pruning; Auto ML;
D O I
10.1145/3474085.3475370
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Modern deep neural network models tend to be large and computationally intensive. One typical solution to this issue is model pruning. However, most current model pruning algorithms depend on hand crafted rules or need to input the pruning ratio beforehand. To overcome this problem, we propose a learning based automatic channel pruning algorithm for deep neural network, which is inspired by recent automatic machine learning (Auto ML). A two objectives' pruning problem that aims for the weights and the remaining channels for each layer is first formulated. An alternative optimization approach is then proposed to derive the channel numbers and weights simultaneously. In the process of pruning, we utilize a searchable hyper-parameter, remaining ratio, to denote the number of channels in each convolution layer, and then a dynamic masking process is proposed to describe the corresponding channel evolution. To adjust the trade-off between accuracy of a model and the pruning ratio of floating point operations, a new loss function is further introduced. Extensive experimental results on benchmark datasets demonstrate that our scheme achieves competitive results for neural network pruning.
引用
收藏
页码:2121 / 2129
页数:9
相关论文
共 50 条
  • [31] Hyper-parameter Search in Support Vector Machines using PSO with Cellular Fitness Approximation
    Yamada, Shinichi
    Neshatian, Kourosh
    2017 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (SSCI), 2017, : 748 - 755
  • [32] Research on Hyper-Parameter Optimization of Activity Recognition Algorithm Based on Improved Cuckoo Search
    Tong, Yu
    Yu, Bo
    ENTROPY, 2022, 24 (06)
  • [33] Gradient Hyper-parameter Optimization for Manifold Regularization
    Becker, Cassiano O.
    Ferreira, Paulo A. V.
    2013 12TH INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA 2013), VOL 2, 2013, : 339 - 344
  • [34] Hyper-Parameter in Hidden Markov Random Field
    Lim, Johan
    Yu, Donghyeon
    Pyun, Kyungsuk
    KOREAN JOURNAL OF APPLIED STATISTICS, 2011, 24 (01) : 177 - 183
  • [35] Bayesian Optimization for Accelerating Hyper-parameter Tuning
    Vu Nguyen
    2019 IEEE SECOND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND KNOWLEDGE ENGINEERING (AIKE), 2019, : 302 - 305
  • [36] On hyper-parameter selection for guaranteed convergence of RMSProp
    Liu, Jinlan
    Xu, Dongpo
    Zhang, Huisheng
    Mandic, Danilo
    COGNITIVE NEURODYNAMICS, 2022, 18 (6) : 3227 - 3237
  • [37] Efficient Hyper-parameter Optimization with Cubic Regularization
    Shen, Zhenqian
    Yang, Hansi
    Li, Yong
    Kwok, James
    Yao, Quanming
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [38] A Comparative study of Hyper-Parameter Optimization Tools
    Shekhar, Shashank
    Bansode, Adesh
    Salim, Asif
    2021 IEEE ASIA-PACIFIC CONFERENCE ON COMPUTER SCIENCE AND DATA ENGINEERING (CSDE), 2021,
  • [39] FastTuning: Enabling Fast and Efficient Hyper-Parameter Tuning With Partitioning and Parallelism of Search Space
    Li, Xiaqing
    Guo, Qi
    Zhang, Guangyan
    Ye, Siwei
    He, Guanhua
    Yao, Yiheng
    Zhang, Rui
    Hao, Yifan
    Du, Zidong
    Zheng, Weimin
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2024, 35 (07) : 1174 - 1188
  • [40] ONLINE HYPER-PARAMETER TUNING FOR THE CONTEXTUAL BANDIT
    Bouneffouf, Djallel
    Claeys, Emmanuelle
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 3445 - 3449