Compressing neural networks with two-layer decoupling

被引:0
|
作者
De Jonghe, Joppe [1 ]
Usevich, Konstantin [2 ]
Dreesen, Philippe [3 ]
Ishteva, Mariya [1 ]
机构
[1] Katholieke Univ Leuven, Dept Comp Sci, Geel, Belgium
[2] Univ Lorraine, CNRS, Nancy, France
[3] Maastricht Univ, DACS, Maastricht, Netherlands
关键词
tensor; tensor decomposition; decoupling; compression; neural network; MODEL COMPRESSION; ACCELERATION;
D O I
10.1109/CAMSAP58249.2023.10403509
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
The single-layer decoupling problem has recently been used for the compression of neural networks. However, methods that are based on the single-layer decoupling problem only allow the compression of a neural network to a single flexible layer. As a result, compressing more complex networks leads to worse approximations of the original network due to only having one flexible layer. Having the ability to compress to more than one flexible layer thus allows to better approximate the underlying network compared to compression into only a single flexible layer. Performing compression into more than one flexible layer corresponds to solving a multi-layer decoupling problem. As a first step towards general multi-layer decoupling, this work introduces a method for solving the two-layer decoupling problem in the approximate case. This method enables the compression of neural networks into two flexible layers.
引用
下载
收藏
页码:226 / 230
页数:5
相关论文
共 50 条
  • [31] Fast Convergence in Learning Two-Layer Neural Networks with Separable Data
    Taheri, Hossein
    Thrampoulidis, Christos
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 8, 2023, : 9944 - 9952
  • [32] Fast and Provable Algorithms for Learning Two-Layer Polynomial Neural Networks
    Soltani, Mohammadreza
    Hegde, Chinmay
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2019, 67 (13) : 3361 - 3371
  • [33] Synchronizability of two-layer correlation networks
    Wei, Xiang
    Wu, Xiaoqun
    Lu, Jun-An
    Wei, Juan
    Zhao, Junchan
    Wang, Yisi
    CHAOS, 2021, 31 (10)
  • [34] Untrusted Caches in Two-layer Networks
    Zewail, Ahmed A.
    Yener, Aylin
    2019 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY (ISIT), 2019, : 1 - 5
  • [35] Provable Identifiability of Two-Layer ReLU Neural Networks via LASSO Regularization
    Li G.
    Wang G.
    Ding J.
    IEEE Transactions on Information Theory, 2023, 69 (09) : 5921 - 5935
  • [36] On the learning dynamics of two-layer quadratic neural networks for understanding deep learning
    TAN Zhenghao
    CHEN Songcan
    Frontiers of Computer Science, 2022, 16 (03)
  • [37] Cumulant-based blind identification of two-layer feedforward neural networks
    Dai, Xianhua
    Dianzi Yu Xinxi Xuebao/Journal of Electronics and Information Technology, 2002, 24 (01):
  • [38] New method of training two-layer sigmoid neural networks using regularization
    Krutikov, V. N.
    Kazakovtsev, L. A.
    Shkaberina, G. Sh
    Kazakovtsev, V. L.
    INTERNATIONAL WORKSHOP ADVANCED TECHNOLOGIES IN MATERIAL SCIENCE, MECHANICAL AND AUTOMATION ENGINEERING - MIP: ENGINEERING - 2019, 2019, 537
  • [39] On the learning dynamics of two-layer quadratic neural networks for understanding deep learning
    Zhenghao Tan
    Songcan Chen
    Frontiers of Computer Science, 2022, 16
  • [40] Self-Regularity of Output Weights for Overparameterized Two-Layer Neural Networks
    Gamarnik, David
    Kizildag, Eren C.
    Zadik, Ilias
    2021 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY (ISIT), 2021, : 819 - 824