Generalization Bounds of Deep Neural Networks With τ -Mixing Samples

被引:0
|
作者
Liu, Liyuan [1 ,2 ]
Chen, Yaohui [3 ]
Li, Weifu [1 ,2 ]
Wang, Yingjie [4 ]
Gu, Bin [5 ]
Zheng, Feng [6 ]
Chen, Hong [1 ,2 ]
机构
[1] Huazhong Agr Univ, Coll Informat, Wuhan 430070, Peoples R China
[2] Minist Educ, Engn Res Ctr Intelligent Technol Agr, Wuhan 430070, Peoples R China
[3] Huazhong Agr Univ, Coll Engn, Wuhan 430070, Peoples R China
[4] China Univ Petr East China, Coll Control Sci & Engn, Qingdao 266580, Peoples R China
[5] Jilin Univ, Sch Artificial Intelligence, Changchun 130012, Peoples R China
[6] Southern Univ Sci & Technol, Dept Comp Sci & Engn, Shenzhen 518055, Peoples R China
基金
中国国家自然科学基金;
关键词
Estimation; Convergence; Analytical models; Artificial neural networks; Time series analysis; Vectors; Robustness; Lips; Learning systems; Hidden Markov models; tau-mixing; covering number; deep neural networks (DNNs); generalization bounds; TIME-SERIES; INEQUALITIES; SEQUENCES;
D O I
10.1109/TNNLS.2025.3526235
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks (DNNs) have shown an astonishing ability to unlock the complicated relationships among the inputs and their responses. Along with empirical successes, some approximation analysis of DNNs has also been provided to understand their generalization performance. However, the existing analysis depends heavily on the independently identically distribution (i.i.d.) assumption of observations, which may be too ideal and often violated in real-world applications. To relax the i.i.d. assumption, this article develops the covering number-based concentration estimation to establish generalization bounds of DNNs with tau -mixing samples, where the dependency between samples is much general including alpha-mixing process as a special case. By assigning a specific parameter value to the tau -mixing process, our results are consistent with the existing convergence analysis under the i.i.d. case. Experiments on simulated data validate the theoretical findings.
引用
收藏
页数:15
相关论文
共 50 条
  • [21] Generalization Comparison of Deep Neural Networks via Output Sensitivity
    Forouzesh, Mahsa
    Salehi, Farnood
    Thiran, Patrick
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 7411 - 7418
  • [22] Quantitative analysis of the generalization ability of deep feedforward neural networks
    Yang, Yanli
    Li, Chenxia
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2021, 40 (03) : 4867 - 4876
  • [23] Learning Cartographic Building Generalization with Deep Convolutional Neural Networks
    Feng, Yu
    Thiemann, Frank
    Sester, Monika
    ISPRS INTERNATIONAL JOURNAL OF GEO-INFORMATION, 2019, 8 (06)
  • [24] Improving generalization of deep neural networks by leveraging margin distribution
    Lyu, Shen-Huan
    Wang, Lu
    Zhou, Zhi-Hua
    NEURAL NETWORKS, 2022, 151 : 48 - 60
  • [25] Improving the Generalization of Deep Neural Networks in Seismic Resolution Enhancement
    Zhang, Haoran
    Alkhalifah, Tariq
    Liu, Yang
    Birnie, Claire
    Di, Xi
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2023, 20
  • [26] Sparsity-aware generalization theory for deep neural networks
    Muthukumar, Ramchandran
    Sulam, Jeremias
    THIRTY SIXTH ANNUAL CONFERENCE ON LEARNING THEORY, VOL 195, 2023, 195
  • [27] Seismic signal augmentation to improve generalization of deep neural networks
    Zhu, Weiqiang
    Mousavi, S. Mostafa
    Beroza, Gregory C.
    MACHINE LEARNING IN GEOSCIENCES, 2020, 61 : 151 - 177
  • [28] A type of generalization error induced by initialization in deep neural networks
    Zhang, Yaoyu
    Xu, Zhi-Qin John
    Luo, Tao
    Ma, Zheng
    MATHEMATICAL AND SCIENTIFIC MACHINE LEARNING, VOL 107, 2020, 107 : 144 - 164
  • [29] On the turnpike to design of deep neural networks: Explicit depth bounds
    Faulwasser, Timm
    Hempel, Arne-Jens
    Streif, Stefan
    IFAC JOURNAL OF SYSTEMS AND CONTROL, 2024, 30
  • [30] GENERALIZATION BY NEURAL NETWORKS
    SHEKHAR, S
    AMIN, MB
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 1992, 4 (02) : 177 - 185