How does Weight Correlation Affect the Generalisation Ability of Deep Neural Networks?

被引:0
|
作者
Jin, Gaojie [1 ]
Yi, Xinping [1 ]
Zhang, Liang [2 ,3 ]
Zhang, Lijun [2 ,4 ]
Schewe, Sven [1 ]
Huang, Xiaowei [1 ]
机构
[1] Univ Liverpool, Liverpool, England
[2] Chinese Acad Sci, Inst Software, State Key Lab Comp Sci, Beijing, Peoples R China
[3] Univ Chinese Acad Sci, Beijing, Peoples R China
[4] Inst Intellegence Software, Guangzhou, Peoples R China
基金
英国工程与自然科学研究理事会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper studies the novel concept of weight correlation in deep neural networks and discusses its impact on the networks' generalisation ability. For fully-connected layers, the weight correlation is defined as the average cosine similarity between weight vectors of neurons, and for convolutional layers, the weight correlation is defined as the cosine similarity between filter matrices. Theoretically, we show that, weight correlation can, and should, be incorporated into the PAC Bayesian framework for the generalisation of neural networks, and the resulting generalisation bound is monotonic with respect to the weight correlation. We formulate a new complexity measure, which lifts the PAC Bayes measure with weight correlation, and experimentally confirm that it is able to rank the generalisation errors of a set of networks more precisely than existing measures. More importantly, we develop a new regulariser for training, and provide extensive experiments that show that the generalisation error can be greatly reduced with our novel approach.
引用
收藏
页数:11
相关论文
共 50 条
  • [21] How does enactment affect the ability to follow instructions in working memory?
    Richard J. Allen
    Amanda H. Waterman
    Memory & Cognition, 2015, 43 : 555 - 561
  • [22] Simulating speech processing with cochlear implants: How does channel interaction affect learning in neural networks?
    Grimm, Robert
    Pettinato, Michele
    Gillis, Steven
    Daelemans, Walter
    PLOS ONE, 2019, 14 (02):
  • [23] How does momentum benefit deep neural networks architecture design? A few case studies
    Wang, Bao
    Xia, Hedi
    Nguyen, Tan
    Osher, Stanley
    RESEARCH IN THE MATHEMATICAL SCIENCES, 2022, 9 (03)
  • [24] How does momentum benefit deep neural networks architecture design? A few case studies
    Bao Wang
    Hedi Xia
    Tan Nguyen
    Stanley Osher
    Research in the Mathematical Sciences, 2022, 9
  • [25] How Does a Firms's Knowledge-Application Ability Affect Its Innovation Ability?
    Tao Ye
    Yu Jian
    Zhang Shilong
    2009 SECOND INTERNATIONAL SYMPOSIUM ON KNOWLEDGE ACQUISITION AND MODELING: KAM 2009, VOL 2, 2009, : 403 - +
  • [26] How does the presence of neural probes affect extracellular potentials?
    Buccino, Alessio Paolo
    Kuchta, Miroslav
    Jaeger, Karoline Horgmo
    Ness, Torbjorn Vefferstad
    Berthet, Pierre
    Mardal, Kent-Andre
    Cauwenberghs, Gert
    Tveito, Aslak
    JOURNAL OF NEURAL ENGINEERING, 2019, 16 (02)
  • [27] How Does Numerosity Adaptation Affect Neural Numerosity Selectivity?
    Tsouli, Andromachi
    Cai, Yuxuan
    Harvey, Ben M.
    te Pas, Susan F.
    van der Smagt, Maarten J.
    Dumoulin, Serge O.
    PERCEPTION, 2019, 48 : 96 - 97
  • [28] Modelling of surface roughness and grinding forces using artificial neural networks with assessment of the ability to data generalisation
    Lipinski, D.
    Balasz, B.
    Rypina, L.
    INTERNATIONAL JOURNAL OF ADVANCED MANUFACTURING TECHNOLOGY, 2018, 94 (1-4): : 1335 - 1347
  • [29] Modelling of surface roughness and grinding forces using artificial neural networks with assessment of the ability to data generalisation
    D. Lipiński
    B. Bałasz
    Ł. Rypina
    The International Journal of Advanced Manufacturing Technology, 2018, 94 : 1335 - 1347
  • [30] Quantitative analysis of the generalization ability of deep feedforward neural networks
    Yang, Yanli
    Li, Chenxia
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2021, 40 (03) : 4867 - 4876