Improving generalization of deep neural networks by leveraging margin distribution

被引:9
|
作者
Lyu, Shen-Huan [1 ]
Wang, Lu [1 ]
Zhou, Zhi-Hua [1 ]
机构
[1] Nanjing Univ, Natl Key Lab Novel Software Technol, Nanjing 210023, Peoples R China
关键词
Deep neural network; Margin theory; Generalization;
D O I
10.1016/j.neunet.2022.03.019
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent research has used margin theory to analyze the generalization performance for deep neural networks (DNNs). The existed results are almost based on the spectrally-normalized minimum margin. However, optimizing the minimum margin ignores a mass of information about the entire margin distribution, which is crucial to generalization performance. In this paper, we prove a generalization upper bound dominated by the statistics of the entire margin distribution. Compared with the minimum margin bounds, our bound highlights an important measure for controlling the complexity, which is the ratio of the margin standard deviation to the expected margin. We utilize a convex margin distribution loss function on the deep neural networks to validate our theoretical results by optimizing the margin ratio. Experiments and visualizations confirm the effectiveness of our approach and the correlation between generalization gap and margin ratio. (c) 2022 Elsevier Ltd. All rights reserved.
引用
收藏
页码:48 / 60
页数:13
相关论文
共 50 条
  • [21] VOVU: A Method for Predicting Generalization in Deep Neural Networks
    Wang, Juan
    Ge, Liangzhu
    Liu, Guorui
    Li, Guoyan
    MATHEMATICAL PROBLEMS IN ENGINEERING, 2021, 2021
  • [22] Abstraction Mechanisms Predict Generalization in Deep Neural Networks
    Gain, Alex
    Siegelmann, Hava
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
  • [23] A generalization bound of deep neural networks for dependent data
    Do, Quan Huu
    Nguyen, Binh T.
    Ho, Lam Si Tung
    STATISTICS & PROBABILITY LETTERS, 2024, 208
  • [24] Abstraction Mechanisms Predict Generalization in Deep Neural Networks
    Gain, Alex
    Siegelmann, Hava
    25TH AMERICAS CONFERENCE ON INFORMATION SYSTEMS (AMCIS 2019), 2019,
  • [25] Margin distribution bounds on generalization
    Shawe-Taylor, J
    Cristianini, N
    COMPUTATIONAL LEARNING THEORY, 1999, 1572 : 263 - 273
  • [26] EvoAug: improving generalization and interpretability of genomic deep neural networks with evolution-inspired data augmentations
    Lee, Nicholas Keone
    Tang, Ziqi
    Toneyan, Shushan
    Koo, Peter K.
    GENOME BIOLOGY, 2023, 24 (01)
  • [27] EvoAug: improving generalization and interpretability of genomic deep neural networks with evolution-inspired data augmentations
    Nicholas Keone Lee
    Ziqi Tang
    Shushan Toneyan
    Peter K. Koo
    Genome Biology, 24
  • [28] Improving Deep Neural Networks with Multilayer Maxout Networks
    Sun, Weichen
    Su, Fei
    Wang, Leiquan
    2014 IEEE VISUAL COMMUNICATIONS AND IMAGE PROCESSING CONFERENCE, 2014, : 334 - 337
  • [29] Penalized AdaBoost: Improving the Generalization Error of Gentle AdaBoost through a Margin Distribution
    Wu, Shuqiong
    Nagahashi, Hiroshi
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2015, E98D (11): : 1906 - 1915
  • [30] Spatial Data Augmentation: Improving the Generalization of Neural Networks for Pansharpening
    Chen, Lihui
    Vivone, Gemine
    Nie, Zihao
    Chanussot, Jocelyn
    Yang, Xiaomin
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2023, 61