Riemannian Laplace approximations for Bayesian neural networks

被引:0
|
作者
Bergamin, Federico [1 ]
Moreno-Munoz, Pablo [1 ]
Hauberg, Soren [1 ]
Arvanitidis, Georgios [1 ]
机构
[1] Tech Univ Denmark, DTU Compute, Sect Cognit Syst, Lyngby, Denmark
基金
欧洲研究理事会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Bayesian neural networks often approximate the weight-posterior with a Gaussian distribution. However, practical posteriors are often, even locally, highly non-Gaussian, and empirical performance deteriorates. We propose a simple parametric approximate posterior that adapts to the shape of the true posterior through a Riemannian metric that is determined by the log-posterior gradient. We develop a Riemannian Laplace approximation where samples naturally fall into weight-regions with low negative log-posterior. We show that these samples can be drawn by solving a system of ordinary differential equations, which can be done efficiently by leveraging the structure of the Riemannian metric and automatic differentiation. Empirically, we demonstrate that our approach consistently improves over the conventional Laplace approximation across tasks. We further show that, unlike the conventional Laplace approximation, our method is not overly sensitive to the choice of prior, which alleviates a practical pitfall of current approaches.
引用
收藏
页数:30
相关论文
共 50 条
  • [1] Invariance Learning in Deep Neural Networks with Differentiable Laplace Approximations
    Immer, Alexander
    van der Ouderaa, Tycho F. A.
    Ratsch, Gunnar
    Fortuin, Vincent
    van der Wilk, Mark
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [2] Mixture approximations to Bayesian networks
    Tresp, V
    Haft, M
    Hofmann, R
    UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, PROCEEDINGS, 1999, : 639 - 646
  • [3] Parallelized integrated nested Laplace approximations for fast Bayesian inference
    Lisa Gaedke-Merzhäuser
    Janet van Niekerk
    Olaf Schenk
    Håvard Rue
    Statistics and Computing, 2023, 33
  • [4] Parallelized integrated nested Laplace approximations for fast Bayesian inference
    Gaedke-Merzhaeuser, Lisa
    van Niekerk, Janet
    Schenk, Olaf
    Rue, Havard
    STATISTICS AND COMPUTING, 2023, 33 (01)
  • [5] Laplace approximations and Bayesian information criteria in possibly misspecified models
    Miyata, Yoichi
    COMMUNICATIONS IN STATISTICS-THEORY AND METHODS, 2017, 46 (24) : 12240 - 12258
  • [6] Riemannian Residual Neural Networks
    Katsman, Isay
    Chen, Eric M.
    Holalkere, Sidhanth
    Asch, Anna
    Lou, Aaron
    Lim, Ser-Nam
    De Sa, Christopher
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [7] Approximations and adaptability of neural networks
    Shi, K
    Fei, SH
    Lin, C
    2005 IEEE International Conference on Granular Computing, Vols 1 and 2, 2005, : 253 - 255
  • [8] Riemannian Curvature of Deep Neural Networks
    Kaul, Piyush
    Lall, Brejesh
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (04) : 1410 - 1416
  • [9] Principles of Riemannian Geometry in Neural Networks
    Hauser, Michael
    Ray, Asok
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 30 (NIPS 2017), 2017, 30
  • [10] Stability of Neural Networks on Riemannian Manifolds
    Wang, Zhiyang
    Ruiz, Luana
    Ribeiro, Alejandro
    29TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO 2021), 2021, : 1845 - 1849