Variational Bayesian Neural Networks via Resolution of Singularities

被引:0
|
作者
Wei, Susan [1 ]
Lau, Edmund [1 ]
机构
[1] Univ Melbourne, Sch Math & Stat, Parkville, Australia
基金
澳大利亚研究理事会;
关键词
Normalizing flow; Real log canonical threshold; Singular learning theory; Singular models; Test log-likelihood; Variational free energy; Variational generalization error; Variational inference; ERROR; COMPLEXITY;
D O I
10.1080/10618600.2024.2325455
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
In this work, we advocate for the importance of singular learning theory (SLT) as it pertains to the theory and practice of variational inference in Bayesian neural networks (BNNs). To begin, we lay to rest some of the confusion surrounding discrepancies between downstream predictive performance measured via the test log predictive density and the variational objective. Next, we use the SLT-corrected asymptotic form for singular posterior distributions to inform the design of the variational family itself. Specifically, we build upon the idealized variational family introduced in Bhattacharya, Pati, and Plummer which is theoretically appealing but practically intractable. Our proposal takes shape as a normalizing flow where the base distribution is a carefully-initialized generalized gamma. We conduct experiments comparing this to the canonical Gaussian base distribution and show improvements in terms of variational free energy and variational generalization error. Supplemental appendices and code for the article are available online.
引用
收藏
页数:11
相关论文
共 50 条
  • [1] Variational Learning of Bayesian Neural Networks via Bayesian Dark Knowledge
    Shen, Gehui
    Chen, Xi
    Deng, Zhihong
    [J]. PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 2037 - 2043
  • [2] Resolution of singularities via deep complex-valued neural networks
    Nitta, Tohru
    [J]. MATHEMATICAL METHODS IN THE APPLIED SCIENCES, 2018, 41 (11) : 4170 - 4178
  • [3] Multi-Resolution Model Compression for Deep Neural Networks: A Variational Bayesian Approach
    Xia, Chengyu
    Guo, Huayan
    Ma, Haoyu
    Tsang, Danny H. K.
    Lau, Vincent K. N.
    [J]. IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2024, 72 : 1944 - 1959
  • [4] Collapsed Variational Bounds for Bayesian Neural Networks
    Tomczak, Marcin B.
    Swaroop, Siddharth
    Foong, Andrew Y. K.
    Turner, Richard E.
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [5] Representation Learning via Variational Bayesian Networks
    Barkan, Oren
    Caciularu, Avi
    Rejwan, Idan
    Katz, Ori
    Weill, Jonathan
    Malkiel, Itzik
    Koenigstein, Noam
    [J]. PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021, 2021, : 78 - 88
  • [6] Multiplicative Normalizing Flows for Variational Bayesian Neural Networks
    Louizos, Christos
    Welling, Max
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 70, 2017, 70
  • [7] Structured Dropout Variational Inference for Bayesian Neural Networks
    Son Nguyen
    Duong Nguyen
    Khai Nguyen
    Khoat Than
    Hung Bui
    Nhat Ho
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021,
  • [8] Structured Variational Learning of Bayesian Neural Networks with Horseshoe Priors
    Ghosh, Soumya
    Yao, Jiayu
    Doshi-Velez, Finale
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80
  • [9] Resolution of Singularities Introduced by Hierarchical Structure in Deep Neural Networks
    Nitta, Tohru
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2017, 28 (10) : 2282 - 2293
  • [10] Variational Bayesian Super Resolution
    Babacan, S. Derin
    Molina, Rafael
    Katsaggelos, Aggelos K.
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2011, 20 (04) : 984 - 999