Random ReLU Neural Networks as Non-Gaussian Processes

被引:0
|
作者
Parhi, Rahul [1 ]
Bohra, Pakshal [2 ]
El Biari, Ayoub [2 ]
Pourya, Mehrsa [2 ]
Unser, Michael [2 ]
机构
[1] Department of Electrical and Computer Engineering, University of California, La Jolla, San Diego,CA,92093, United States
[2] Biomedical Imaging Group, École polytechnique fédérale de Lausanne, Lausanne,CH-1015, Switzerland
基金
欧洲研究理事会;
关键词
Differential equations - Gaussian distribution - Gaussian noise (electronic) - Impulse noise - Neural networks - Random processes - Random variables - Stochastic systems;
D O I
暂无
中图分类号
学科分类号
摘要
We consider a large class of shallow neural networks with randomly initialized parameters and rectified linear unit activation functions. We prove that these random neural networks are well-defined non-Gaussian processes. As a by-product, we demonstrate that these networks are solutions to stochastic differential equations driven by impulsive white noise (combinations of random Dirac measures). These processes are parameterized by the law of the weights and biases as well as the density of activation thresholds in each bounded region of the input domain. We prove that these processes are isotropic and wide-sense self-similar with Hurst exponent 3/2. We also derive a remarkably simple closed-form expression for their autocovariance function. Our results are fundamentally different from prior work in that we consider a non-asymptotic viewpoint: The number of neurons in each bounded region of the input domain (i.e., the width) is itself a random variable with a Poisson law with mean proportional to the density parameter. Finally, we show that, under suitable hypotheses, as the expected width tends to infinity, these processes can converge in law not only to Gaussian processes, but also to non-Gaussian processes depending on the law of the weights. Our asymptotic results provide a new take on several classical results (wide networks converge to Gaussian processes) as well as some new ones (wide networks can converge to non-Gaussian processes). ©2025 Rahul Parhi, Pakshal Bohra, Ayoub El Biari, Mehrsa Pourya, and Michael Unser.
引用
收藏
相关论文
共 50 条
  • [1] Random ReLU Neural Networks as Non-Gaussian Processes
    Parhi, Rahul
    Bohra, Pakshal
    El Biari, Ayoub
    Pourya, Mehrsa
    Unser, Michael
    JOURNAL OF MACHINE LEARNING RESEARCH, 2025, 26 : 1 - 31
  • [2] Non-Gaussian processes and neural networks at finite widths
    Yaida, Sho
    MATHEMATICAL AND SCIENTIFIC MACHINE LEARNING, VOL 107, 2020, 107 : 165 - 192
  • [3] OPTIMUM CLASSIFICATION OF NON-GAUSSIAN PROCESSES USING NEURAL NETWORKS
    BLACKNELL, D
    WHITE, RG
    IEE PROCEEDINGS-VISION IMAGE AND SIGNAL PROCESSING, 1994, 141 (01): : 56 - 66
  • [4] Representation of nonlinear random transformations by non-Gaussian stochastic neural networks
    Turchetti, Claudio
    Crippa, Paolo
    Pirani, Massimiliano
    Biagetti, Giorgio
    IEEE TRANSACTIONS ON NEURAL NETWORKS, 2008, 19 (06): : 1033 - 1060
  • [5] INFORMATION RATES ON NON-GAUSSIAN RANDOM PROCESSES
    WATANABE, Y
    ELECTRONICS & COMMUNICATIONS IN JAPAN, 1972, 55 (02): : 32 - 38
  • [6] TRANSFORMATION OF NON-GAUSSIAN RANDOM PROCESSES BY RADIO DEVICES
    SAFIULLIN, NZ
    CHABDAROV, SM
    TELECOMMUNICATIONS AND RADIO ENGINEERING, 1978, 32-3 (04) : 114 - 116
  • [7] Sample Path Asymmetries in Non-Gaussian Random Processes
    Baxevani, Anastassia
    Podgorski, Krzysztof
    Wegener, Jorg
    SCANDINAVIAN JOURNAL OF STATISTICS, 2014, 41 (04) : 1102 - 1123
  • [9] PROBABILITIES OF ERRORS OF RECOGNITION OF NON-GAUSSIAN RANDOM PROCESSES
    KOZIN, IV
    ENGINEERING CYBERNETICS, 1973, 10 (05): : 910 - 915
  • [10] INFLUENCE OF A FILTER ON NON-GAUSSIAN RANDOM-PROCESSES
    CYGAN, D
    FRANZ, J
    SODER, G
    AEU-ARCHIV FUR ELEKTRONIK UND UBERTRAGUNGSTECHNIK-INTERNATIONAL JOURNAL OF ELECTRONICS AND COMMUNICATIONS, 1986, 40 (06): : 377 - 384