On the expected behaviour of noise regularised deep neural networks as Gaussian processes

被引:2
|
作者
Pretorius, Arnu [1 ]
Kamper, Herman [2 ]
Kroon, Steve [1 ]
机构
[1] Stellenbosch Univ, Comp Sci Div, Stellenbosch, South Africa
[2] Stellenbosch Univ, Dept Elect & Elect Engn, Stellenbosch, South Africa
关键词
Neural networks; Gaussian processes; Signal propagation; Noise regularisation;
D O I
10.1016/j.patrec.2020.06.027
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent work has established the equivalence between deep neural networks and Gaussian processes (GPs), resulting in so-called neural network Gaussian processes (NNGPs). The behaviour of these models depends on the initialisation of the corresponding network. In this work, we consider the impact of noise regularisation (e.g. dropout) on NNGPs, and relate their behaviour to signal propagation theory in noise regularised deep neural networks. For ReLU activations, we find that the best performing NNGPs have kernel parameters that correspond to a recently proposed initialisation scheme for noise regularised ReLU networks. In addition, we show how the noise influences the covariance matrix of the NNGP, producing a stronger prior towards simple functions away from the training points. We verify our theoretical findings with experiments on MNIST and CIFAR-10 as well as on synthetic data. (C) 2020 Elsevier B.V. All rights reserved.
引用
收藏
页码:75 / 81
页数:7
相关论文
共 50 条
  • [1] Deep Neural Networks as Point Estimates for Deep Gaussian Processes
    Dutordoir, Vincent
    Hensman, James
    van der wilk, Mark
    Ek, Carl Henrik
    Ghahramani, Zoubin
    Durrande, Nicolas
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [2] Wide neural networks with bottlenecks are deep gaussian processes
    Agrawal, Devanshu
    Papamarkou, Theodore
    Hinkle, Jacob
    [J]. Journal of Machine Learning Research, 2020, 21
  • [3] Wide Neural Networks with Bottlenecks are Deep Gaussian Processes
    Agrawal, Devanshu
    Papamarkou, Theodore
    Hinkle, Jacob
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2020, 21
  • [4] Regularised Estimators for Fractional Gaussian Noise
    Vivero, Oskar
    Heath, William P.
    [J]. 49TH IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2010, : 5025 - 5030
  • [5] Neural networks: A replacement for Gaussian processes?
    Lilley, M
    Frean, M
    [J]. INTELLIGENT DATA ENGINEERING AND AUTOMATED LEARNING IDEAL 2005, PROCEEDINGS, 2005, 3578 : 195 - 202
  • [6] Wide Neural Networks as Gaussian Processes: Lessons from Deep Equilibrium Models
    Gao, Tianxiang
    Huo, Xiaokai
    Liu, Hailiang
    Gao, Hongyang
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [7] Global inducing point variational posteriors for Bayesian neural networks and deep Gaussian processes
    Ober, Sebastian W.
    Aitchison, Laurence
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [8] Passivity analysis for neural networks perturbed by Poisson noise and Gaussian noise
    Ma, Xiang
    Song, Bo
    Fin, Feng
    [J]. PROCEEDINGS OF THE 38TH CHINESE CONTROL CONFERENCE (CCC), 2019, : 8778 - 8782
  • [9] Upper bounds on the expected training errors of neural network regressions for a Gaussian noise
    Hagiwara, K
    Kuno, K
    Usui, S
    [J]. ICONIP'98: THE FIFTH INTERNATIONAL CONFERENCE ON NEURAL INFORMATION PROCESSING JOINTLY WITH JNNS'98: THE 1998 ANNUAL CONFERENCE OF THE JAPANESE NEURAL NETWORK SOCIETY - PROCEEDINGS, VOLS 1-3, 1998, : 502 - 505
  • [10] Enhanced data efficiency using deep neural networks and Gaussian processes for aerodynamic design optimization
    Renganathan, S. Ashwin
    Maulik, Romit
    Ahuja, Jai
    [J]. AEROSPACE SCIENCE AND TECHNOLOGY, 2021, 111