Weak approximation of transformed stochastic gradient MCMC

被引:0
|
作者
Soma Yokoi
Takuma Otsuka
Issei Sato
机构
[1] The University of Tokyo,Department of Complexity Science and Engineering, Graduate School of Frontier Sciences
[2] RIKEN,NTT Communication Science Laboratories
[3] NTT Corporation,Department of Computer Science, Graduate School of Information Science and Technology
[4] The University of Tokyo,undefined
来源
Machine Learning | 2020年 / 109卷
关键词
Stochastic gradient MCMC; Transform; Convergence analysis; Itô process;
D O I
暂无
中图分类号
学科分类号
摘要
Stochastic gradient Langevin dynamics (SGLD) is a computationally efficient sampler for Bayesian posterior inference given a large scale dataset and a complex model. Although SGLD is designed for unbounded random variables, practical models often incorporate variables within a bounded domain, such as non-negative or a finite interval. The use of variable transformation is a typical way to handle such a bounded variable. This paper reveals that several mapping approaches commonly used in the literature produce erroneous samples from theoretical and empirical perspectives. We show that the change of random variable in discretization using an invertible Lipschitz mapping function overcomes the pitfall as well as attains the weak convergence, while the other methods are numerically unstable or cannot be justified theoretically. Experiments demonstrate its efficacy for widely-used models with bounded latent variables, including Bayesian non-negative matrix factorization and binary neural networks.
引用
收藏
页码:1903 / 1923
页数:20
相关论文
共 50 条
  • [1] Weak approximation of transformed stochastic gradient MCMC
    Yokoi, Soma
    Otsuka, Takuma
    Sato, Issei
    [J]. MACHINE LEARNING, 2020, 109 (9-10) : 1903 - 1923
  • [2] Distributed Stochastic Gradient MCMC
    Ahn, Sungjin
    Shahbaba, Babak
    Welling, Max
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 32 (CYCLE 2), 2014, 32 : 1044 - 1052
  • [3] Structured Stochastic Gradient MCMC
    Alexos, Antonios
    Boyd, Alex
    Mandt, Stephan
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022, : 414 - 434
  • [4] A Complete Recipe for Stochastic Gradient MCMC
    Ma, Yi-An
    Chen, Tianqi
    Fox, Emily B.
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 28 (NIPS 2015), 2015, 28
  • [5] Control variates for stochastic gradient MCMC
    Baker, Jack
    Fearnhead, Paul
    Fox, Emily B.
    Nemeth, Christopher
    [J]. STATISTICS AND COMPUTING, 2019, 29 (03) : 599 - 615
  • [6] Control variates for stochastic gradient MCMC
    Jack Baker
    Paul Fearnhead
    Emily B. Fox
    Christopher Nemeth
    [J]. Statistics and Computing, 2019, 29 : 599 - 615
  • [7] Stochastic Gradient MCMC with Stale Gradients
    Chen, Changyou
    Ding, Nan
    Li, Chunyuan
    Zhang, Yizhe
    Carin, Lawrence
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 29 (NIPS 2016), 2016, 29
  • [8] Stochastic Gradient Geodesic MCMC Methods
    Liu, Chang
    Zhu, Jun
    Song, Yang
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 29 (NIPS 2016), 2016, 29
  • [9] WEAK CONVERGENCE RATES OF POPULATION VERSUS SINGLE-CHAIN STOCHASTIC APPROXIMATION MCMC ALGORITHMS
    Song, Qifan
    Wu, Mingqi
    Liang, Faming
    [J]. ADVANCES IN APPLIED PROBABILITY, 2014, 46 (04) : 1059 - 1083