Improving generalization of deep neural networks by leveraging margin distribution

被引:9
|
作者
Lyu, Shen-Huan [1 ]
Wang, Lu [1 ]
Zhou, Zhi-Hua [1 ]
机构
[1] Nanjing Univ, Natl Key Lab Novel Software Technol, Nanjing 210023, Peoples R China
关键词
Deep neural network; Margin theory; Generalization;
D O I
10.1016/j.neunet.2022.03.019
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent research has used margin theory to analyze the generalization performance for deep neural networks (DNNs). The existed results are almost based on the spectrally-normalized minimum margin. However, optimizing the minimum margin ignores a mass of information about the entire margin distribution, which is crucial to generalization performance. In this paper, we prove a generalization upper bound dominated by the statistics of the entire margin distribution. Compared with the minimum margin bounds, our bound highlights an important measure for controlling the complexity, which is the ratio of the margin standard deviation to the expected margin. We utilize a convex margin distribution loss function on the deep neural networks to validate our theoretical results by optimizing the margin ratio. Experiments and visualizations confirm the effectiveness of our approach and the correlation between generalization gap and margin ratio. (c) 2022 Elsevier Ltd. All rights reserved.
引用
收藏
页码:48 / 60
页数:13
相关论文
共 50 条
  • [41] Leveraging Stochasticity for In Situ Learning in Binarized Deep Neural Networks
    Pyle, Steven D.
    Sapp, Justin D.
    DeMara, Ronald F.
    COMPUTER, 2019, 52 (05) : 30 - 39
  • [42] Leveraging the Bhattacharyya coefficient for uncertainty quantification in deep neural networks
    Van Molle, Pieter
    Verbelen, Tim
    Vankeirsbilck, Bert
    De Vylder, Jonas
    Diricx, Bart
    Kimpe, Tom
    Simoens, Pieter
    Dhoedt, Bart
    NEURAL COMPUTING & APPLICATIONS, 2021, 33 (16): : 10259 - 10275
  • [43] Leveraging Deep Neural Networks for Massive MIMO Data Detection
    Nguyen, Ly V.
    Nguyen, Nhan T.
    Tran, Nghi H.
    Juntti, Markku
    Swindlehurst, A. Lee
    Nguyen, Duy H. N.
    IEEE WIRELESS COMMUNICATIONS, 2023, 30 (01) : 174 - 180
  • [44] Leveraging the Bhattacharyya coefficient for uncertainty quantification in deep neural networks
    Pieter Van Molle
    Tim Verbelen
    Bert Vankeirsbilck
    Jonas De Vylder
    Bart Diricx
    Tom Kimpe
    Pieter Simoens
    Bart Dhoedt
    Neural Computing and Applications, 2021, 33 : 10259 - 10275
  • [45] Generalization Analysis of Pairwise Learning for Ranking With Deep Neural Networks
    Huang, Shuo
    Zhou, Junyu
    Feng, Han
    Zhou, Ding-Xuan
    NEURAL COMPUTATION, 2023, 35 (06) : 1135 - 1158
  • [46] Generalization Comparison of Deep Neural Networks via Output Sensitivity
    Forouzesh, Mahsa
    Salehi, Farnood
    Thiran, Patrick
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 7411 - 7418
  • [47] Quantitative analysis of the generalization ability of deep feedforward neural networks
    Yang, Yanli
    Li, Chenxia
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2021, 40 (03) : 4867 - 4876
  • [48] Learning Cartographic Building Generalization with Deep Convolutional Neural Networks
    Feng, Yu
    Thiemann, Frank
    Sester, Monika
    ISPRS INTERNATIONAL JOURNAL OF GEO-INFORMATION, 2019, 8 (06)
  • [49] Sparsity-aware generalization theory for deep neural networks
    Muthukumar, Ramchandran
    Sulam, Jeremias
    THIRTY SIXTH ANNUAL CONFERENCE ON LEARNING THEORY, VOL 195, 2023, 195
  • [50] Seismic signal augmentation to improve generalization of deep neural networks
    Zhu, Weiqiang
    Mousavi, S. Mostafa
    Beroza, Gregory C.
    MACHINE LEARNING IN GEOSCIENCES, 2020, 61 : 151 - 177