Stochastic Gradient Hamiltonian Monte Carlo

被引:0
|
作者
Chen, Tianqi [1 ]
Fox, Emily B. [1 ]
Guestrin, Carlos [1 ]
机构
[1] Univ Washington, MODE Lab, Seattle, WA 98195 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Hamiltonian Monte Carlo (HMC) sampling methods provide a mechanism for defining distant proposals with high acceptance probabilities in a Metropolis-Hastings framework, enabling more efficient exploration of the state space than standard random-walk proposals. The popularity of such methods has grown significantly in recent years. However, a limitation of HMC methods is the required gradient computation for simulation of the Hamiltonian dynamical system-such computation is infeasible in problems involving a large sample size or streaming data. Instead, we must rely on a noisy gradient estimate computed from a subset of the data. In this paper, we explore the properties of such a stochastic gradient HMC approach. Surprisingly, the natural implementation of the stochastic approximation can be arbitrarily bad. To address this problem we introduce a variant that uses second-order Langevin dynamics with a friction term that counteracts the effects of the noisy gradient, maintaining the desired target distribution as the invariant distribution. Results on simulated data validate our theory. We also provide an application of our methods to a classification task using neural networks and to online Bayesian matrix factorization.
引用
收藏
页码:1683 / 1691
页数:9
相关论文
共 50 条
  • [1] A Hybrid Stochastic Gradient Hamiltonian Monte Carlo Method
    Zhang, Chao
    Li, Zhijian
    Shen, Zebang
    Xie, Jiahao
    Qian, Hui
    [J]. THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 10842 - 10850
  • [2] Decentralized Stochastic Gradient Langevin Dynamics and Hamiltonian Monte Carlo
    Gurbuzbalaban, Mert
    Gao, Xuefeng
    Hu, Yuanhan
    Zhu, Lingjiong
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2021, 22
  • [3] Decentralized stochastic gradient langevin dynamics and hamiltonian Monte Carlo
    Gürbüzbalaban, Mert
    Gao, Xuefeng
    Hu, Yuanhan
    Zhu, Lingjiong
    [J]. Journal of Machine Learning Research, 2021, 22
  • [4] Stochastic Gradient Hamiltonian Monte Carlo for non-convex learning
    Chau, Huy N.
    Rasonyi, Miklos
    [J]. STOCHASTIC PROCESSES AND THEIR APPLICATIONS, 2022, 149 : 341 - 368
  • [5] Stochastic Gradient Hamiltonian Monte Carlo Methods with Recursive Variance Reduction
    Zou, Difan
    Xu, Pan
    Gu, Quanquan
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [6] Stochastic gradient Hamiltonian Monte Carlo with variance reduction for Bayesian inference
    Zhize Li
    Tianyi Zhang
    Shuyu Cheng
    Jun Zhu
    Jian Li
    [J]. Machine Learning, 2019, 108 : 1701 - 1727
  • [7] Stochastic gradient Hamiltonian Monte Carlo with variance reduction for Bayesian inference
    Li, Zhize
    Zhang, Tianyi
    Cheng, Shuyu
    Zhu, Jun
    Li, Jian
    [J]. MACHINE LEARNING, 2019, 108 (8-9) : 1701 - 1727
  • [8] Stochastic Fractional Hamiltonian Monte Carlo
    Ye, Nanyang
    Zhu, Zhanxing
    [J]. PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2018, : 3019 - 3025
  • [9] Stochastic approximation Hamiltonian Monte Carlo
    Yun, Jonghyun
    Shin, Minsuk
    Hoon Jin, Ick
    Liang, Faming
    [J]. JOURNAL OF STATISTICAL COMPUTATION AND SIMULATION, 2020, 90 (17) : 3135 - 3156
  • [10] Inference in Deep Gaussian Processes using Stochastic Gradient Hamiltonian Monte Carlo
    Havasi, Marton
    Hernandez-Lobato, Jose Miguel
    Jose Murillo-Fuentes, Juan
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31