A Two-Stage Pretraining Algorithm for Deep Boltzmann Machines

被引:0
|
作者
Cho, KyungHyun [1 ]
Raiko, Tapani [1 ]
Ilin, Alexander [1 ]
Karhunen, Juha [1 ]
机构
[1] Aalto Univ, Sch Sci, Dept Informat & Comp Sci, Helsinki, Finland
关键词
Deep Boltzmann Machine; Deep Learning; Pretraining;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A deep Boltzmann machine (DBM) is a recently introduced Markov random field model that has multiple layers of hidden units. It has been shown empirically that it is difficult to train a DBM with approximate maximum-likelihood learning using the stochastic gradient unlike its simpler special case, restricted Boltzmann machine (RBM). In this paper, we propose a novel pretraining algorithm that consists of two stages; obtaining approximate posterior distributions over hidden units from a simpler model and maximizing the variational lower-bound given the fixed hidden posterior distributions. We show empirically that the proposed method overcomes the difficulty in training DBMs from randomly initialized parameters and results in a better, or comparable, generative model when compared to the conventional pretraining algorithm.
引用
收藏
页码:106 / 113
页数:8
相关论文
共 50 条