Compressing Neural Networks using the Variational Information Bottleneck

被引:0
|
作者
Dai, Bin [1 ]
Zhu, Chen [2 ]
Guo, Baining [3 ]
Wipf, David [3 ]
机构
[1] Tsinghua Univ, Inst Adv Study, Beijing, Peoples R China
[2] Univ Maryland, Dept Comp Sci, College Pk, MD 20742 USA
[3] Microsoft, Beijing, Peoples R China
关键词
SELECTION;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Neural networks can be compressed to reduce memory and computational requirements, or to increase accuracy by facilitating the use of a larger base architecture. In this paper we focus on pruning individual neurons, which can simultaneously trim model size, FLOPs, and run-time memory. To improve upon the performance of existing compression algorithms we utilize the information bottleneck principle instantiated via a tractable variational bound. Minimization of this information theoretic bound reduces the redundancy between adjacent layers by aggregating useful information into a subset of neurons that can be preserved. In contrast, the activations of disposable neurons are shut off via an attractive form of sparse regularization that emerges naturally from this framework, providing tangible advantages over traditional sparsity penalties without contributing additional tuning parameters to the energy landscape. We demonstrate state-of-theart compression rates across an array of datasets and network architectures.
引用
收藏
页数:10
相关论文
共 50 条
  • [1] Sentiment Analysis via Deep Multichannel Neural Networks With Variational Information Bottleneck
    Gu, Tong
    Xu, Guoliang
    Luo, Jiangtao
    [J]. IEEE ACCESS, 2020, 8 : 121014 - 121021
  • [2] Information Bottleneck Theory on Convolutional Neural Networks
    Li, Junjie
    Liu, Ding
    [J]. NEURAL PROCESSING LETTERS, 2021, 53 (02) : 1385 - 1400
  • [3] Information Bottleneck Theory on Convolutional Neural Networks
    Junjie Li
    Ding Liu
    [J]. Neural Processing Letters, 2021, 53 : 1385 - 1400
  • [4] Training quantum neural networks using the quantum information bottleneck method
    Catli, Ahmet Burak
    Wiebe, Nathan
    [J]. JOURNAL OF PHYSICS A-MATHEMATICAL AND THEORETICAL, 2024, 57 (37)
  • [5] Variational Predictive Information Bottleneck
    Alemi, Alexander A.
    [J]. SYMPOSIUM ON ADVANCES IN APPROXIMATE BAYESIAN INFERENCE, VOL 118, 2019, 118
  • [6] Markov Information Bottleneck to Improve Information Flow in Stochastic Neural Networks
    Thanh Tang Nguyen
    Choi, Jaesik
    [J]. ENTROPY, 2019, 21 (10)
  • [7] Distributed Deep Variational Information Bottleneck
    Zaidi, Abdellatif
    Aguerri, Inaki Estella
    [J]. PROCEEDINGS OF THE 21ST IEEE INTERNATIONAL WORKSHOP ON SIGNAL PROCESSING ADVANCES IN WIRELESS COMMUNICATIONS (IEEE SPAWC2020), 2020,
  • [8] Cell Variational Information Bottleneck Network
    Zhai, Zhonghua
    [J]. ASIAN CONFERENCE ON MACHINE LEARNING, VOL 222, 2023, 222
  • [9] Information Bottleneck in Control Tasks with Recurrent Spiking Neural Networks
    Vasu, Madhavun Candadai
    Izquierdo, Eduardo J.
    [J]. ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2017, PT I, 2017, 10613 : 236 - 244
  • [10] Anti-Spoofing Using Transfer Learning with Variational Information Bottleneck
    Eom, Youngsik
    Lee, Yeonghyeon
    Um, Ji Sub
    Kim, Hoirin
    [J]. INTERSPEECH 2022, 2022, : 3568 - 3572