Deep Neural Network Approximation Theory

被引:71
|
作者
Elbrachter, Dennis [1 ]
Perekrestenko, Dmytro [2 ]
Grohs, Philipp [1 ,3 ]
Boelcskei, Helmut [2 ]
机构
[1] Univ Vienna, Dept Math, A-1090 Vienna, Austria
[2] Swiss Fed Inst Technol, Chair Math Informat Sci, CH-8092 Zurich, Switzerland
[3] Univ Vienna, Res Platform Data Sci Univ Vienna, A-1090 Vienna, Austria
基金
奥地利科学基金会;
关键词
Neural networks; Complexity theory; Function approximation; Dictionaries; Approximation error; Two dimensional displays; Training data; machine learning; function approximation; rate-distortion theory; nonlinear approximation; metric entropy; BOUNDS; REPRESENTATIONS; ENTROPY; SMOOTH; BASES;
D O I
10.1109/TIT.2021.3062161
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper develops fundamental limits of deep neural network learning by characterizing what is possible if no constraints are imposed on the learning algorithm and on the amount of training data. Concretely, we consider Kolmogorov-optimal approximation through deep neural networks with the guiding theme being a relation between the complexity of the function (class) to be approximated and the complexity of the approximating network in terms of connectivity and memory requirements for storing the network topology and the associated quantized weights. The theory we develop establishes that deep networks are Kolmogorov-optimal approximants for markedly different function classes, such as unit balls in Besov spaces and modulation spaces. In addition, deep networks provide exponential approximation accuracy-i.e., the approximation error decays exponentially in the number of nonzero weights in the network-of the multiplication operation, polynomials, sinusoidal functions, and certain smooth functions. Moreover, this holds true even for one-dimensional oscillatory textures and the Weierstrass function-a fractal function, neither of which has previously known methods achieving exponential approximation accuracy. We also show that in the approximation of sufficiently smooth functions finite-width deep networks require strictly smaller connectivity than finite-depth wide networks.
引用
收藏
页码:2581 / 2623
页数:43
相关论文
共 50 条
  • [1] Constructive Deep ReLU Neural Network Approximation
    Herrmann, Lukas
    Opschoor, Joost A. A.
    Schwab, Christoph
    [J]. JOURNAL OF SCIENTIFIC COMPUTING, 2022, 90 (02)
  • [2] Constructive Deep ReLU Neural Network Approximation
    Lukas Herrmann
    Joost A. A. Opschoor
    Christoph Schwab
    [J]. Journal of Scientific Computing, 2022, 90
  • [3] Exponential convergence of the deep neural network approximation for analytic functions
    E, Weinan
    Wang, Qingcan
    [J]. SCIENCE CHINA-MATHEMATICS, 2018, 61 (10) : 1733 - 1740
  • [4] Exponential convergence of the deep neural network approximation for analytic functions
    Weinan E
    Qingcan Wang
    [J]. Science China Mathematics, 2018, 61 : 1733 - 1740
  • [5] The deep neural network solver for B-spline approximation
    Wen, Zepeng
    Luo, Jiaqi
    Kang, Hongmei
    [J]. COMPUTER-AIDED DESIGN, 2024, 169
  • [6] Deep Neural Network Approximation of Nonlinear Model Predictive Control
    Cao, Yankai
    Gopaluni, R. Bhushan
    [J]. IFAC PAPERSONLINE, 2020, 53 (02): : 11319 - 11324
  • [7] Incremental Deep Neural Network Pruning based on Hessian Approximation
    Li, Li
    Li, Zhu
    Li, Yue
    Kathariya, Birendra
    Bhattacharyya, Shuvra
    [J]. 2019 DATA COMPRESSION CONFERENCE (DCC), 2019, : 590 - 590
  • [8] Adaptive Approximation and Generalization of Deep Neural Network with Intrinsic Dimensionality
    Nakada, Ryumei
    Imaizumi, Masaaki
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2020, 21
  • [9] A Neural Network Monte Carlo Approximation for Expected Utility Theory
    Zhu, Yichen
    Escobar-Anel, Marcos
    [J]. JOURNAL OF RISK AND FINANCIAL MANAGEMENT, 2021, 14 (07)
  • [10] The Gap between Theory and Practice in Function Approximation with Deep Neural Networks
    Adcock, Ben
    Dexter, Nick
    [J]. SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE, 2021, 3 (02): : 624 - 655