How does Weight Correlation Affect the Generalisation Ability of Deep Neural Networks?

被引:0
|
作者
Jin, Gaojie [1 ]
Yi, Xinping [1 ]
Zhang, Liang [2 ,3 ]
Zhang, Lijun [2 ,4 ]
Schewe, Sven [1 ]
Huang, Xiaowei [1 ]
机构
[1] Univ Liverpool, Liverpool, England
[2] Chinese Acad Sci, Inst Software, State Key Lab Comp Sci, Beijing, Peoples R China
[3] Univ Chinese Acad Sci, Beijing, Peoples R China
[4] Inst Intellegence Software, Guangzhou, Peoples R China
基金
英国工程与自然科学研究理事会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper studies the novel concept of weight correlation in deep neural networks and discusses its impact on the networks' generalisation ability. For fully-connected layers, the weight correlation is defined as the average cosine similarity between weight vectors of neurons, and for convolutional layers, the weight correlation is defined as the cosine similarity between filter matrices. Theoretically, we show that, weight correlation can, and should, be incorporated into the PAC Bayesian framework for the generalisation of neural networks, and the resulting generalisation bound is monotonic with respect to the weight correlation. We formulate a new complexity measure, which lifts the PAC Bayes measure with weight correlation, and experimentally confirm that it is able to rank the generalisation errors of a set of networks more precisely than existing measures. More importantly, we develop a new regulariser for training, and provide extensive experiments that show that the generalisation error can be greatly reduced with our novel approach.
引用
收藏
页数:11
相关论文
共 50 条
  • [41] How to Boost Deep Neural Networks for Computer Vision
    Ha, Soonhoi
    2023 60TH ACM/IEEE DESIGN AUTOMATION CONFERENCE, DAC, 2023,
  • [42] How Does the Experimental Setting Affect the Conclusions of Neural Encoding Models?
    Zhang, Xiaohan
    Wang, Shaonan
    Zong, Chengqing
    LREC 2022: THIRTEEN INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2022, : 6397 - 6404
  • [43] Intergroup Empathy: How Does Race Affect Empathic Neural Responses?
    Chiao, Joan Y.
    Mathur, Vani A.
    CURRENT BIOLOGY, 2010, 20 (11) : R478 - R480
  • [44] Latent Weight Quantization for Integerized Training of Deep Neural Networks
    Fei, Wen
    Dai, Wenrui
    Zhang, Liang
    Zhang, Luoming
    Li, Chenglin
    Zou, Junni
    Xiong, Hongkai
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2025, 47 (04) : 2816 - 2832
  • [45] Centered Weight Normalization in Accelerating Training of Deep Neural Networks
    Huang, Lei
    Liu, Xianglong
    Liu, Yang
    Lang, Bo
    Tao, Dacheng
    2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 2822 - 2830
  • [46] Scaling-Based Weight Normalization for Deep Neural Networks
    Yuan, Qunyong
    Xiao, Nanfeng
    IEEE ACCESS, 2019, 7 : 7286 - 7295
  • [47] SqueezeBlock: A Transparent Weight Compression Scheme for Deep Neural Networks
    Song, Mo
    Wu, Jiajun
    Ding, Yuhao
    So, Hayden Kwok-Hay
    2023 INTERNATIONAL CONFERENCE ON FIELD PROGRAMMABLE TECHNOLOGY, ICFPT, 2023, : 238 - 243
  • [48] Layer-Wise Weight Decay for Deep Neural Networks
    Ishii, Masato
    Sato, Atsushi
    IMAGE AND VIDEO TECHNOLOGY (PSIVT 2017), 2018, 10749 : 276 - 289
  • [49] Hybrid pooling for enhancement of generalization ability in deep convolutional neural networks
    Tong, Zhiqiang
    Tanaka, Gouhei
    NEUROCOMPUTING, 2019, 333 : 76 - 85
  • [50] Understanding the Ability of Deep Neural Networks to Count Connected Components in Images
    Guan, Shuyue
    Loew, Murray
    2020 IEEE APPLIED IMAGERY PATTERN RECOGNITION WORKSHOP (AIPR): TRUSTED COMPUTING, PRIVACY, AND SECURING MULTIMEDIA, 2020,