Sharp asymptotics on the compression of two-layer neural networks

被引:0
|
作者
Amani, Mohammad Hossein [1 ]
Bombari, Simone [2 ]
Mondelli, Marco [2 ]
Pukdee, Rattana [3 ]
Rini, Stefano [4 ]
机构
[1] Ecole Polytech Fed Lausanne, Lausanne, Switzerland
[2] IST Austria, Klosterneuburg, Austria
[3] CMU, Pittsburgh, PA USA
[4] NYCU, Hsinchu, Taiwan
关键词
D O I
10.1109/ITW54588.2022.9965870
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, we study the compression of a target two-layer neural network with N nodes into a compressed network with M < N nodes. More precisely, we consider the setting in which the weights of the target network are i.i.d. sub-Gaussian, and we minimize the population L-2 loss between the outputs of the target and of the compressed network, under the assumption of Gaussian inputs. By using tools from high-dimensional probability, we show that this non-convex problem can be simplified when the target network is sufficiently over-parameterized, and provide the error rate of this approximation as a function of the input dimension and N. In this mean-field limit, the simplified objective, as well as the optimal weights of the compressed network, does not depend on the realization of the target network, but only on expected scaling factors. Furthermore, for networks with ReLU activation, we conjecture that the optimum of the simplified optimization problem is achieved by taking weights on the Equiangular Tight Frame (ETF), while the scaling of the weights and the orientation of the ETF depend on the parameters of the target network. Numerical evidence is provided to support this conjecture.
引用
下载
收藏
页码:588 / 593
页数:6
相关论文
共 50 条
  • [1] Plasticity of two-layer fast neural networks
    Alexeev, AA
    Dorogov, AY
    JOURNAL OF COMPUTER AND SYSTEMS SCIENCES INTERNATIONAL, 1999, 38 (05) : 786 - 791
  • [2] Compressing neural networks with two-layer decoupling
    De Jonghe, Joppe
    Usevich, Konstantin
    Dreesen, Philippe
    Ishteva, Mariya
    2023 IEEE 9TH INTERNATIONAL WORKSHOP ON COMPUTATIONAL ADVANCES IN MULTI-SENSOR ADAPTIVE PROCESSING, CAMSAP, 2023, : 226 - 230
  • [3] On the Structure of Two-Layer Cellular Neural Networks
    Ban, Jung-Chao
    Chang, Chih-Hung
    Lin, Song-Sun
    DIFFERENTIAL AND DIFFERENCE EQUATIONS WITH APPLICATI ONS, 2013, 47 : 265 - 273
  • [4] Templates and algorithms for two-layer cellular neural Networks
    Yang, ZH
    Nishio, Y
    Ushida, A
    PROCEEDING OF THE 2002 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-3, 2002, : 1946 - 1951
  • [5] Two-Layer Feedback Neural Networks with Associative Memories
    Wu Gui-Kun
    Zhao Hong
    CHINESE PHYSICS LETTERS, 2008, 25 (11) : 3871 - 3874
  • [6] Two-layer stabilization of continuous neural networks with feedbacks
    Dudnikov, EE
    CYBERNETICS AND SYSTEMS, 2002, 33 (04) : 325 - 340
  • [7] Structural synthesis of fast two-layer neural networks
    A. Yu. Dorogov
    Cybernetics and Systems Analysis, 2000, 36 : 512 - 519
  • [8] Benign Overfitting in Two-layer Convolutional Neural Networks
    Cao, Yuan
    Chen, Zixiang
    Belkin, Mikhail
    Gu, Quanquan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [9] Structural synthesis of two-layer rapid neural networks
    Dorogov, A.Yu.
    Kibernetika i Sistemnyj Analiz, 2000, (04): : 47 - 57
  • [10] On the symmetries in the dynamics of wide two-layer neural networks
    Hajjar, Karl
    Chizat, Lenaic
    ELECTRONIC RESEARCH ARCHIVE, 2023, 31 (04): : 2175 - 2212