Generalization Bounds of Deep Neural Networks With τ -Mixing Samples

被引:0
|
作者
Liu, Liyuan [1 ,2 ]
Chen, Yaohui [3 ]
Li, Weifu [1 ,2 ]
Wang, Yingjie [4 ]
Gu, Bin [5 ]
Zheng, Feng [6 ]
Chen, Hong [1 ,2 ]
机构
[1] Huazhong Agr Univ, Coll Informat, Wuhan 430070, Peoples R China
[2] Minist Educ, Engn Res Ctr Intelligent Technol Agr, Wuhan 430070, Peoples R China
[3] Huazhong Agr Univ, Coll Engn, Wuhan 430070, Peoples R China
[4] China Univ Petr East China, Coll Control Sci & Engn, Qingdao 266580, Peoples R China
[5] Jilin Univ, Sch Artificial Intelligence, Changchun 130012, Peoples R China
[6] Southern Univ Sci & Technol, Dept Comp Sci & Engn, Shenzhen 518055, Peoples R China
基金
中国国家自然科学基金;
关键词
Estimation; Convergence; Analytical models; Artificial neural networks; Time series analysis; Vectors; Robustness; Lips; Learning systems; Hidden Markov models; tau-mixing; covering number; deep neural networks (DNNs); generalization bounds; TIME-SERIES; INEQUALITIES; SEQUENCES;
D O I
10.1109/TNNLS.2025.3526235
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks (DNNs) have shown an astonishing ability to unlock the complicated relationships among the inputs and their responses. Along with empirical successes, some approximation analysis of DNNs has also been provided to understand their generalization performance. However, the existing analysis depends heavily on the independently identically distribution (i.i.d.) assumption of observations, which may be too ideal and often violated in real-world applications. To relax the i.i.d. assumption, this article develops the covering number-based concentration estimation to establish generalization bounds of DNNs with tau -mixing samples, where the dependency between samples is much general including alpha-mixing process as a special case. By assigning a specific parameter value to the tau -mixing process, our results are consistent with the existing convergence analysis under the i.i.d. case. Experiments on simulated data validate the theoretical findings.
引用
收藏
页数:15
相关论文
共 50 条
  • [1] Theoretical Investigation of Generalization Bounds for Adversarial Learning of Deep Neural Networks
    Qingyi Gao
    Xiao Wang
    Journal of Statistical Theory and Practice, 2021, 15
  • [2] Generalization bounds for neural ordinary differential equations and deep residual networks
    Marion, Pierre
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [3] Theoretical Investigation of Generalization Bounds for Adversarial Learning of Deep Neural Networks
    Gao, Qingyi
    Wang, Xiao
    JOURNAL OF STATISTICAL THEORY AND PRACTICE, 2021, 15 (02)
  • [4] Generalization Bounds of Stochastic Gradient Descent for Wide and Deep Neural Networks
    Cao, Yuan
    Gu, Quanquan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [5] On Generalization Bounds of a Family of Recurrent Neural Networks
    Chen, Minshuo
    Li, Xingguo
    Zhao, Tuo
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 108, 2020, 108 : 1233 - 1242
  • [6] Generalization and risk bounds for recurrent neural networks
    Cheng, Xuewei
    Huang, Ke
    Ma, Shujie
    NEUROCOMPUTING, 2025, 616
  • [7] Generalization Bounds for Neural Networks via Approximate Description Length
    Daniely, Amit
    Granot, Elad
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [8] Norm-based Generalization Bounds for Sparse Neural Networks
    Galanti, Tomer
    Xu, Mengjia
    Galanti, Liane
    Poggio, Tomaso
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [9] Computing Nonvacuous Generalization Bounds for Deep (Stochastic) Neural Networks with Many More Parameters than Training Data
    Dziugaite, Gintare Karolina
    Roy, Daniel M.
    CONFERENCE ON UNCERTAINTY IN ARTIFICIAL INTELLIGENCE (UAI2017), 2017,
  • [10] Algorithm-Dependent Generalization Bounds for Overparameterized Deep Residual Networks
    Frei, Spencer
    Cao, Yuan
    Gu, Quanquan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32