Improving generalization of deep neural networks by leveraging margin distribution

被引:9
|
作者
Lyu, Shen-Huan [1 ]
Wang, Lu [1 ]
Zhou, Zhi-Hua [1 ]
机构
[1] Nanjing Univ, Natl Key Lab Novel Software Technol, Nanjing 210023, Peoples R China
关键词
Deep neural network; Margin theory; Generalization;
D O I
10.1016/j.neunet.2022.03.019
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent research has used margin theory to analyze the generalization performance for deep neural networks (DNNs). The existed results are almost based on the spectrally-normalized minimum margin. However, optimizing the minimum margin ignores a mass of information about the entire margin distribution, which is crucial to generalization performance. In this paper, we prove a generalization upper bound dominated by the statistics of the entire margin distribution. Compared with the minimum margin bounds, our bound highlights an important measure for controlling the complexity, which is the ratio of the margin standard deviation to the expected margin. We utilize a convex margin distribution loss function on the deep neural networks to validate our theoretical results by optimizing the margin ratio. Experiments and visualizations confirm the effectiveness of our approach and the correlation between generalization gap and margin ratio. (c) 2022 Elsevier Ltd. All rights reserved.
引用
收藏
页码:48 / 60
页数:13
相关论文
共 50 条
  • [11] Robust Large Margin Deep Neural Networks
    Sokolic, Jure
    Giryes, Raja
    Sapiro, Guillermo
    Rodrigues, Miguel R. D.
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2017, 65 (16) : 4265 - 4280
  • [12] Pairwise Margin Maximization for Deep Neural Networks
    Weinstein, Berry
    Fine, Shai
    Hel-Or, Yacov
    20TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA 2021), 2021, : 263 - 270
  • [13] Improving Hashing by Leveraging Multiple Layers of Deep Networks
    Luo, Xin
    Chen, Zhen-Duo
    Du, Gao-Yuan
    Xu, Xin-Shun
    NEURAL INFORMATION PROCESSING, ICONIP 2017, PT I, 2017, 10634 : 597 - 607
  • [14] Improving Stability and Generalization of Magnetic Anomaly Detection Using Deep Convolutional Siamese Neural Networks
    Chen, Zijie
    Miao, Linliang
    Yang, Xiaofei
    Ouyang, Jun
    IEEE SENSORS JOURNAL, 2024, 24 (15) : 24466 - 24482
  • [15] Ensemble Approach for Improving Generalization Ability of Neural Networks
    Ahmed, Shaib
    Razib, Md. Razibul Islam
    Alam, Mohammed Shamsul
    Alam, Mohammad Shafiul
    Huda, Mohammad Nurul
    2013 INTERNATIONAL CONFERENCE ON INFORMATICS, ELECTRONICS & VISION (ICIEV), 2013,
  • [16] Improving the generalization ability of neural networks by interval arithmetic
    Ishibuchi, H
    Nii, M
    1998 SECOND INTERNATIONAL CONFERENCE ON KNOWLEDGE-BASED INTELLIGENT ELECTRONIC SYSTEMS, KES'98 PROCEEDINGS, VOL 1, 1998, : 231 - 236
  • [17] Hybrid methodology for improving generalization performance of neural networks
    Yang, Zhong-Jin
    PROCEEDINGS OF 2006 INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND CYBERNETICS, VOLS 1-7, 2006, : 3167 - 3172
  • [18] A Survey on Leveraging Deep Neural Networks for Object Tracking
    Krebs, Sebastian
    Duraisamy, Bharanidhar
    Flohr, Fabian
    2017 IEEE 20TH INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS (ITSC), 2017,
  • [19] iDropout: Leveraging Deep Taylor Decomposition for the Robustness of Deep Neural Networks
    Schreckenberger, Christian
    Bartelt, Christian
    Stuckenschmidt, Heiner
    ON THE MOVE TO MEANINGFUL INTERNET SYSTEMS: OTM 2019 CONFERENCES, 2019, 11877 : 113 - 126
  • [20] Generalization Bounds of Deep Neural Networks With τ -Mixing Samples
    Liu, Liyuan
    Chen, Yaohui
    Li, Weifu
    Wang, Yingjie
    Gu, Bin
    Zheng, Feng
    Chen, Hong
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2025,