Automatic Network Pruning via Hilbert-Schmidt Independence Criterion Lasso under Information Bottleneck Principle

被引:2
|
作者
Guo, Song [1 ]
Zhang, Lei [1 ]
Zheng, Xiawu [1 ,2 ]
Wang, Yan [3 ]
Li, Yuchao [4 ]
Chao, Fei [1 ]
Wu, Chenglin [5 ]
Zhang, Shengchuan [1 ]
Ji, Rongrong [1 ,2 ,6 ,7 ]
机构
[1] Xiamen Univ, Sch Informat, Dept Artificial Intelligence,Minist Educ China, Key Lab Multimedia Trusted Percept & Efficient Co, Xiamen, Peoples R China
[2] Peng Cheng Lab, Shenzhen, Peoples R China
[3] Samsara Inc, San Francisco, CA USA
[4] Alibaba Grp, Hangzhou, Peoples R China
[5] Deep Wisdom Inc, Xiamen, Peoples R China
[6] Xiamen Univ, Inst Artificial Intelligence, Xiamen, Peoples R China
[7] Xiamen Univ, Fujian Engn Res Ctr Trusted Artificial Intelligen, Xiamen, Peoples R China
基金
中国国家自然科学基金; 国家重点研发计划;
关键词
D O I
10.1109/ICCV51070.2023.01601
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Most existing neural network pruning methods handcrafted their importance criteria and structures to prune. This constructs heavy and unintended dependencies on heuristics and expert experience for both the objective and the parameters of the pruning approach. In this paper, we try to solve this problem by introducing a principled and unified framework based on Information Bottleneck (IB) theory, which further guides us to an automatic pruning approach. Specifically, we first formulate the channel pruning problem from an IB perspective, and then implement the IB principle by solving a Hilbert-Schmidt Independence Criterion (HSIC) Lasso problem under certain conditions. Based on the theoretical guidance, we then provide an automatic pruning scheme by searching for global penalty coefficients. Verified by extensive experiments, our method yields state-of-the-art performance on various benchmark networks and datasets. For example, with VGG-16, we achieve a 60%-FLOPs reduction by removing 76% of the parameters, with an improvement of 0.40% in top-1 accuracy on CIFAR-10. With ResNet-50, we achieve a 56%-FLOPs reduction by removing 50% of the parameters, with a small loss of 0.08% in the top-1 accuracy on ImageNet. The code is available at https://github.com/sunggo/APIB.
引用
收藏
页码:17412 / 17423
页数:12
相关论文
共 46 条
  • [1] UNSUPERVISED FEATURE SELECTION WITH HILBERT-SCHMIDT INDEPENDENCE CRITERION LASSO
    Wang, Tinghua
    Hu, Zhenwei
    Zhou, Huiying
    [J]. INTERNATIONAL JOURNAL OF INNOVATIVE COMPUTING INFORMATION AND CONTROL, 2023, 19 (03): : 927 - 939
  • [2] Sequence Alignment with the Hilbert-Schmidt Independence Criterion
    Campbell, Jordan
    Lewis, J. P.
    Seol, Yeongho
    [J]. PROCEEDINGS CVMP 2018: THE 15TH ACM SIGGRAPH EUROPEAN CONFERENCE ON VISUAL MEDIA PRODUCTION, 2018,
  • [3] Robust Learning with the Hilbert-Schmidt Independence Criterion
    Greenfeld, Daniel
    Shalit, Uri
    [J]. 25TH AMERICAS CONFERENCE ON INFORMATION SYSTEMS (AMCIS 2019), 2019,
  • [4] Test of conditional independence in factor models via Hilbert-Schmidt independence criterion
    Xu, Kai
    Cheng, Qing
    [J]. JOURNAL OF MULTIVARIATE ANALYSIS, 2024, 199
  • [5] Sensitivity maps of the Hilbert-Schmidt independence criterion
    Perez-Suay, Adrian
    Camps-Valls, Gustau
    [J]. APPLIED SOFT COMPUTING, 2018, 70 : 1054 - 1063
  • [6] Nystrom M -Hilbert-Schmidt Independence Criterion
    Kalinke, Florian
    Szabo, Zoltan
    [J]. UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, 2023, 216 : 1005 - 1015
  • [7] Sparse Hilbert-Schmidt Independence Criterion Regression
    Poignard, Benjamin
    Yamada, Makoto
    [J]. INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 108, 2020, 108 : 538 - 547
  • [8] Kernel Learning with Hilbert-Schmidt Independence Criterion
    Wang, Tinghua
    Li, Wei
    He, Xianwen
    [J]. PATTERN RECOGNITION (CCPR 2016), PT I, 2016, 662 : 720 - 730
  • [9] Robust Learning with the Hilbert-Schmidt Independence Criterion
    Greenfeld, Daniel
    Shalit, Uri
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
  • [10] DIBAD: A Disentangled Information Bottleneck Adversarial Defense Method Using Hilbert-Schmidt Independence Criterion for Spectrum Security
    Zhang, Sicheng
    Yang, Yandie
    Zhou, Ziyao
    Sun, Zhi
    Lin, Yun
    [J]. IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2024, 19 : 3879 - 3891