Neural Redshift: Random Networks are not Random Functions

被引:0
|
作者
Teney, Damien [1 ]
Nicolicioiu, Armand Mihai [2 ]
Hartmann, Valentin [3 ]
Abbasnejad, Ehsan [4 ]
机构
[1] Idiap Res Inst, Martigny, Switzerland
[2] Swiss Fed Inst Technol, Zurich, Switzerland
[3] Ecole Polytech Fed Lausanne, Lausanne, Switzerland
[4] Univ Adelaide, Adelaide, SA, Australia
基金
澳大利亚研究理事会;
关键词
COMPLEXITY;
D O I
10.1109/CVPR52733.2024.00458
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Our understanding of the generalization capabilities of neural networks (NNs) is still incomplete. Prevailing explanations are based on implicit biases of gradient descent (GD) but they cannot account for the capabilities of models from gradient-free methods [9] nor the simplicity bias recently observed in untrained networks [29]. This paper seeks other sources of generalization in NNs. Findings. To understand the inductive biases provided by architectures independently from GD, we examine untrained, random-weight networks. Even simple MLPs show strong inductive biases: uniform sampling in weight space yields a very biased distribution of functions in terms of complexity. But unlike common wisdom, NNs do not have an inherent "simplicity bias". This property depends on components such as ReLUs, residual connections, and layer normalizations. Alternative architectures can be built with a bias for any level of complexity. Transformers also inherit all these properties from their building blocks. Implications. We provide a fresh explanation for the success of deep learning independent from gradient-based training. It points at promising avenues for controlling the solutions implemented by trained models.
引用
收藏
页码:4786 / 4796
页数:11
相关论文
共 50 条
  • [31] The Separation Capacity of Random Neural Networks
    Dirksen, Sjoerd
    Genzel, Martin
    Jacques, Laurent
    Stollenwerk, Alexander
    Journal of Machine Learning Research, 2022, 23
  • [32] Random neural networks for the adaptive control of packet networks
    Gellman, Michael
    Liu, Peixiang
    ARTIFICIAL NEURAL NETWORKS - ICANN 2006, PT 1, 2006, 4131 : 313 - 320
  • [33] Random relations, random utilities, and random functions
    Regenwetter, M
    Marley, AAJ
    JOURNAL OF MATHEMATICAL PSYCHOLOGY, 2001, 45 (06) : 864 - 912
  • [34] Learning Drug Functions from Chemical Structures with Convolutional Neural Networks and Random Forests
    Meyer, Jesse G.
    Liu, Shengchao
    Miller, Ian J.
    Coon, Joshua J.
    Gitter, Anthony
    JOURNAL OF CHEMICAL INFORMATION AND MODELING, 2019, 59 (10) : 4438 - 4449
  • [35] Green's functions for random resistor networks
    Bhattacharjee, Sayak
    Ramola, Kabir
    PHYSICAL REVIEW E, 2023, 108 (04)
  • [36] ON FORCING FUNCTIONS IN KAUFFMAN RANDOM BOOLEAN NETWORKS
    STAUFFER, D
    JOURNAL OF STATISTICAL PHYSICS, 1987, 46 (3-4) : 789 - 794
  • [38] Routes to chaos in neural networks with random weights
    Albers, DJ
    Sprott, JC
    Dechert, WD
    INTERNATIONAL JOURNAL OF BIFURCATION AND CHAOS, 1998, 8 (07): : 1463 - 1478
  • [39] Random neural networks with multiple classes of signals
    Gelenbe, E
    Fourneau, JM
    NEURAL COMPUTATION, 1999, 11 (04) : 953 - 963
  • [40] Random-Forest-Inspired Neural Networks
    Wang, Suhang
    Aggarwal, Charu
    Liu, Huan
    ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2018, 9 (06)