Learning Deep Generative Models With Doubly Stochastic Gradient MCMC

被引:8
|
作者
Du, Chao [1 ]
Zhu, Jun [1 ]
Zhang, Bo [1 ]
机构
[1] Tsinghua Univ, Dept Comp Sci & Technol, State Key Lab Intelligent Technol & Syst, Tsinghua Natl Lab Informat Sci & Technol, Beijing 100084, Peoples R China
关键词
Bayesian methods; deep generative models (DGMs); deep learning; Markov chain Monte Carlo (MCMC); stochastic gradient; BACKPROPAGATION; INFERENCE; ALGORITHM; NETWORKS;
D O I
10.1109/TNNLS.2017.2688499
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep generative models (DGMs), which are often organized in a hierarchical manner, provide a principled framework of capturing the underlying causal factors of data. Recent work on DGMs focussed on the development of efficient and scalable variational inference methods that learn a single model under some mean-field or parameterization assumptions. However, little work has been done on extending Markov chain Monte Carlo (MCMC) methods to Bayesian DGMs, which enjoy many advantages compared with variational methods. We present doubly stochastic gradient MCMC, a simple and generic method for (approximate) Bayesian inference of DGMs in a collapsed continuous parameter space. At each MCMC sampling step, the algorithm randomly draws a mini-batch of data samples to estimate the gradient of log-posterior and further estimates the intractable expectation over hidden variables via a neural adaptive importance sampler, where the proposal distribution is parameterized by a deep neural network and learnt jointly along with the sampling process. We demonstrate the effectiveness of learning various DGMs on a wide range of tasks, including density estimation, data generation, and missing data imputation. Our method outperforms many state-of-the-art competitors.
引用
收藏
页码:3084 / 3096
页数:13
相关论文
共 50 条
  • [21] Generative chemistry: drug discovery with deep learning generative models
    Bian, Yuemin
    Xie, Xiang-Qun
    [J]. JOURNAL OF MOLECULAR MODELING, 2021, 27 (03)
  • [22] Deep Generative Learning via Variational Gradient Flow
    Gao, Yuan
    Jiao, Yuling
    Wang, Yang
    Wang, Yao
    Yang, Can
    Zhang, Shunkang
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [23] Generative chemistry: drug discovery with deep learning generative models
    Yuemin Bian
    Xiang-Qun Xie
    [J]. Journal of Molecular Modeling, 2021, 27
  • [24] Non-convex Learning via Replica Exchange Stochastic Gradient MCMC
    Deng, Wei
    Feng, Qi
    Gao, Liyao
    Liang, Faming
    Lin, Guang
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
  • [25] Non-convex Learning via Replica Exchange Stochastic Gradient MCMC
    Deng, Wei
    Feng, Qi
    Gao, Liyao
    Liang, Faming
    Lin, Guang
    [J]. 25TH AMERICAS CONFERENCE ON INFORMATION SYSTEMS (AMCIS 2019), 2019,
  • [26] LEARNING SPARSE STRUCTURED ENSEMBLES WITH STOCHASTIC GRADIENT MCMC SAMPLING AND NETWORK PRUNING
    Zhang, Yichi
    Ou, Zhijian
    [J]. 2018 IEEE 28TH INTERNATIONAL WORKSHOP ON MACHINE LEARNING FOR SIGNAL PROCESSING (MLSP), 2018,
  • [27] Composite Functional Gradient Learning of Generative Adversarial Models
    Johnson, Rie
    Zhang, Tong
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80
  • [28] Learning Deep Generative Models for Queuing Systems
    Ojeda, Cesar
    Cvejoski, Kostadin
    Georgiev, Bodgan
    Bauckhage, Christian
    Schuecker, Jannis
    Sanchez, Ramses J.
    [J]. THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 9214 - 9222
  • [29] Weak approximation of transformed stochastic gradient MCMC
    Soma Yokoi
    Takuma Otsuka
    Issei Sato
    [J]. Machine Learning, 2020, 109 : 1903 - 1923
  • [30] On Connecting Stochastic Gradient MCMC and Differential Privacy
    Li, Bai
    Chen, Changyou
    Liu, Hao
    Carin, Lawrence
    [J]. 22ND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 89, 2019, 89 : 557 - 566