On the Theory of Variance Reduction for Stochastic Gradient Monte Carlo

被引:0
|
作者
Chatterji, Niladri S. [1 ]
Flammarion, Nicolas [2 ]
Ma, Yi-An [2 ]
Bartlett, Peter L. [2 ,3 ]
Jordan, Michael I. [2 ,3 ]
机构
[1] Univ Calif Berkeley, Dept Phys, Berkeley, CA 94720 USA
[2] Univ Calif Berkeley, Div Comp Sci, Berkeley, CA 94720 USA
[3] Univ Calif Berkeley, Dept Stat, Berkeley, CA 94720 USA
关键词
LANGEVIN;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We provide convergence guarantees in Wasserstein distance for a variety of variance-reduction methods: SAGA Langevin diffusion, SVRG Langevin diffusion and control-variate under-damped Langevin diffusion. We analyze these methods under a uniform set of assumptions on the log-posterior distribution, assuming it to be smooth, strongly convex and Hessian Lipschitz. This is achieved by a new proof technique combining ideas from finite-sum optimization and the analysis of sampling methods. Our sharp theoretical bounds allow us to identify regimes of interest where each method performs better than the others. Our theory is verified with experiments on real-world and synthetic datasets.
引用
收藏
页数:10
相关论文
共 50 条