Invariance Learning in Deep Neural Networks with Differentiable Laplace Approximations

被引:0
|
作者
Immer, Alexander [1 ,2 ]
van der Ouderaa, Tycho F. A. [3 ]
Ratsch, Gunnar [1 ]
Fortuin, Vincent [1 ,4 ]
van der Wilk, Mark [3 ]
机构
[1] Swiss Fed Inst Technol, Dept Comp Sci, Zurich, Switzerland
[2] Max Planck Inst Intelligent Syst, Tubingen, Germany
[3] Imperial Coll London, Dept Comp, London, England
[4] Univ Cambridge, Dept Engn, Cambridge, England
基金
瑞士国家科学基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Data augmentation is commonly applied to improve performance of deep learning by enforcing the knowledge that certain transformations on the input preserve the output. Currently, the data augmentation parameters are chosen by human effort and costly cross-validation, which makes it cumbersome to apply to new datasets. We develop a convenient gradient-based method for selecting the data augmentation without validation data during training of a deep neural network. Our approach relies on phrasing data augmentation as an invariance in the prior distribution on the functions of a neural network, which allows us to learn it using Bayesian model selection. This has been shown to work in Gaussian processes, but not yet for deep neural networks. We propose a differentiable Kronecker-factored Laplace approximation to the marginal likelihood as our objective, which can be optimised without human supervision or validation data. We show that our method can successfully recover invariances present in the data, and that this improves generalisation and data efficiency on image datasets.
引用
收藏
页数:15
相关论文
共 50 条
  • [41] Deep learning with coherent VCSEL neural networks
    Chen, Zaijun
    Sludds, Alexander
    Davis III, Ronald
    Christen, Ian
    Bernstein, Liane
    Ateshian, Lamia
    Heuser, Tobias
    Heermeier, Niels
    Lott, James A.
    Reitzenstein, Stephan
    Hamerly, Ryan
    Englund, Dirk
    NATURE PHOTONICS, 2023, 17 (08) : 723 - +
  • [42] Learning hidden chemistry with deep neural networks
    Nguyen, Tien-Cuong
    Nguyen, Van-Quyen
    Ngo, Van-Linh
    Than, Quang-Khoat
    Pham, Tien-Lam
    COMPUTATIONAL MATERIALS SCIENCE, 2021, 200
  • [43] Learning Structured Sparsity in Deep Neural Networks
    Wen, Wei
    Wu, Chunpeng
    Wang, Yandan
    Chen, Yiran
    Li, Hai
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 29 (NIPS 2016), 2016, 29
  • [44] Representational Distance Learning for Deep Neural Networks
    McClure, Patrick
    Kriegeskorte, Nikolaus
    FRONTIERS IN COMPUTATIONAL NEUROSCIENCE, 2016, 10
  • [45] Learning hidden elasticity with deep neural networks
    Chen, Chun-Teh
    Gu, Grace X.
    PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2021, 118 (31)
  • [46] Neural networks and deep learning: a brief introduction
    Georgevici, Adrian Iustin
    Terblanche, Marius
    INTENSIVE CARE MEDICINE, 2019, 45 (05) : 712 - 714
  • [47] Evolving Deep Neural Networks for Continuous Learning
    Atamanczuk, Bruna
    Karadas, Kurt Arve Skipenes
    Agrawal, Bikash
    Chakravorty, Antorweep
    FRONTIERS OF ARTIFICIAL INTELLIGENCE, ETHICS, AND MULTIDISCIPLINARY APPLICATIONS, FAIEMA 2023, 2024, : 3 - 16
  • [48] Evolutionary neural networks for deep learning: a review
    Ma, Yongjie
    Xie, Yirong
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2022, 13 (10) : 3001 - 3018
  • [49] Deep supervised learning with mixture of neural networks
    Hu, Yaxian
    Luo, Senlin
    Han, Longfei
    Pan, Limin
    Zhang, Tiemei
    ARTIFICIAL INTELLIGENCE IN MEDICINE, 2020, 102
  • [50] On the Expressivity of Neural Networks for Deep Reinforcement Learning
    Dong, Kefan
    Luo, Yuping
    Yu, Tianhe
    Finn, Chelsea
    Ma, Tengyu
    25TH AMERICAS CONFERENCE ON INFORMATION SYSTEMS (AMCIS 2019), 2019,