Uniform Convergence of Deep Neural Networks With Lipschitz Continuous Activation Functions and Variable Widths

被引:0
|
作者
Xu, Yuesheng [1 ]
Zhang, Haizhang [2 ]
机构
[1] Old Dominion Univ, Dept Math & Stat, Norfolk, VA 23529 USA
[2] Sun Yat Sen Univ, Sch Math Zhuhai, Zhuhai 519082, Peoples R China
基金
美国国家科学基金会; 美国国家卫生研究院; 中国国家自然科学基金;
关键词
Convergence; Vectors; Artificial neural networks; Kernel; Training; Deep learning; Uniform convergence; deep neural networks; convolutional neural networks; Lipschitz continuous activation functions; variable widths; RELU NETWORKS; ERROR-BOUNDS;
D O I
10.1109/TIT.2024.3439136
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We consider deep neural networks (DNNs) with a Lipschitz continuous activation function and with weight matrices of variable widths. We establish a uniform convergence analysis framework in which sufficient conditions on weight matrices and bias vectors together with the Lipschitz constant are provided to ensure uniform convergence of DNNs to a meaningful function as the number of their layers tends to infinity. In the framework, special results on uniform convergence of DNNs with a fixed width, bounded widths and unbounded widths are presented. In particular, as convolutional neural networks are special DNNs with weight matrices of increasing widths, we put forward conditions on the mask sequence which lead to uniform convergence of the resulting convolutional neural networks. The Lipschitz continuity assumption on the activation functions allows us to include in our theory most of commonly used activation functions in applications.
引用
收藏
页码:7125 / 7142
页数:18
相关论文
共 50 条
  • [31] Global Convergence for Cohen-Grossberg Neural Networks with Discontinuous Activation Functions
    Wang, Yanyan
    Zhou, Jianping
    ABSTRACT AND APPLIED ANALYSIS, 2012,
  • [32] Global robust exponential stability for Hopfield neural networks with non-Lipschitz activation functions
    Hongtao Yu
    Huaiqin Wu
    Journal of Mathematical Sciences, 2012, 187 (4) : 511 - 523
  • [33] Convergence of deep convolutional neural networks
    Xu, Yuesheng
    Zhang, Haizhang
    NEURAL NETWORKS, 2022, 153 : 553 - 563
  • [34] Robust exponential stability for interval neural networks with delays and non-Lipschitz activation functions
    Huaiqin Wu
    Feng Tao
    Leijie Qin
    Rui Shi
    Lijun He
    Nonlinear Dynamics, 2011, 66 : 479 - 487
  • [35] Robust exponential stability for interval neural networks with delays and non-Lipschitz activation functions
    Wu, Huaiqin
    Tao, Feng
    Qin, Leijie
    Shi, Rui
    He, Lijun
    NONLINEAR DYNAMICS, 2011, 66 (04) : 479 - 487
  • [36] CONSTRUCTION OF NEURAL NETWORKS TO APPROXIMATE ARBITRARY CONTINUOUS-FUNCTIONS OF ONE VARIABLE
    CHOI, CH
    CHOI, JY
    ELECTRONICS LETTERS, 1992, 28 (02) : 151 - 153
  • [37] Approximating smooth functions by deep neural networks with sigmoid activation function
    Langer, Sophie
    JOURNAL OF MULTIVARIATE ANALYSIS, 2021, 182
  • [38] Simple Electromagnetic Analysis Against Activation Functions of Deep Neural Networks
    Takatoi, Go
    Sugawara, Takeshi
    Sakiyama, Kazuo
    Li, Yang
    APPLIED CRYPTOGRAPHY AND NETWORK SECURITY WORKSHOPS, ACNS 2020, 2020, 12418 : 181 - 197
  • [39] Smooth Function Approximation by Deep Neural Networks with General Activation Functions
    Ohn, Ilsang
    Kim, Yongdai
    ENTROPY, 2019, 21 (07)
  • [40] Influences of variable scales and activation functions on the performances of multilayer feedforward neural networks
    Gao, DQ
    Yang, GX
    PATTERN RECOGNITION, 2003, 36 (04) : 869 - 878