Generalization error bounds for iterative recovery algorithms unfolded as neural networks

被引:1
|
作者
Schnoor, Ekkehard [1 ]
Behboodi, Arash [2 ]
Rauhut, Holger [1 ]
机构
[1] Rhein Westfal TH Aachen, Chair Math Informat Proc, Aachen, Germany
[2] Rhein Westfal TH Aachen, Inst Theoret Informat Technol, Aachen, Germany
关键词
compressive sensing; neural networks; iterative soft thresholding; generalization; Rademacher complexity; SAMPLE COMPLEXITY; OVERCOMPLETE DICTIONARIES; SPARSE;
D O I
10.1093/imaiai/iaad023
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
Motivated by the learned iterative soft thresholding algorithm (LISTA), we introduce a general class of neural networks suitable for sparse reconstruction from few linear measurements. By allowing a wide range of degrees of weight-sharing between the flayers, we enable a unified analysis for very different neural network types, ranging from recurrent ones to networks more similar to standard feedforward neural networks. Based on training samples, via empirical risk minimization, we aim at learning the optimal network parameters and thereby the optimal network that reconstructs signals from their low-dimensional linear measurements. We derive generalization bounds by analyzing the Rademacher complexity of hypothesis classes consisting of such deep networks, that also take into account the thresholding parameters. We obtain estimates of the sample complexity that essentially depend only linearly on the number of parameters and on the depth. We apply our main result to obtain specific generalization bounds for several practical examples, including different algorithms for (implicit) dictionary learning, and convolutional neural networks.
引用
收藏
页数:33
相关论文
共 50 条
  • [41] NEW ITERATIVE INVERSE SCATTERING ALGORITHMS BASED ON NEURAL NETWORKS
    LEE, HJ
    AHN, CH
    PARK, CS
    JEONG, BS
    LEE, SY
    IEEE TRANSACTIONS ON MAGNETICS, 1994, 30 (05) : 3641 - 3643
  • [42] Generalization error bounds for the logical analysis of data
    Anthony, Martin
    DISCRETE APPLIED MATHEMATICS, 2012, 160 (10-11) : 1407 - 1415
  • [43] Generalization error bounds using Wasserstein distances
    Lopez, Adrian Tovar
    Jog, Varun
    2018 IEEE INFORMATION THEORY WORKSHOP (ITW), 2018, : 205 - 209
  • [44] Generalization error bounds using unlabeled data
    Kääriäinen, M
    LEARNING THEORY, PROCEEDINGS, 2005, 3559 : 127 - 142
  • [45] Generalization error bounds for threshold decision lists
    Anthony, M
    JOURNAL OF MACHINE LEARNING RESEARCH, 2004, 5 : 189 - 217
  • [46] Renyi Divergence Based Bounds on Generalization Error
    Modak, Eeshan
    Asnani, Himanshu
    Prabhakaran, Vinod M.
    2021 IEEE INFORMATION THEORY WORKSHOP (ITW), 2021,
  • [47] Non-vacuous Generalization Bounds for Adversarial Risk in Stochastic Neural Networks
    Mustafa, Waleed
    Liznerski, Philipp
    Ledent, Antoine
    Wagner, Dennis
    Wang, Puyu
    Kloft, Marius
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 238, 2024, 238
  • [48] Generalization Bounds for Some Ordinal Regression Algorithms
    Agarwal, Shivani
    ALGORITHMIC LEARNING THEORY, PROCEEDINGS, 2008, 5254 : 7 - 21
  • [49] Infinite Kernel Learning: Generalization Bounds and Algorithms
    Liu, Yong
    Liao, Shizhong
    Lin, Hailun
    Yue, Yinliang
    Wang, Weiping
    THIRTY-FIRST AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 2280 - 2286
  • [50] GENERALIZATION BY NEURAL NETWORKS
    SHEKHAR, S
    AMIN, MB
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 1992, 4 (02) : 177 - 185