Training restricted Boltzmann machines: An introduction

被引:303
|
作者
Fischer, Asja [1 ,2 ]
Igel, Christian [2 ]
机构
[1] Ruhr Univ Bochum, Inst Neuroinformat, D-44780 Bochum, Germany
[2] Univ Copenhagen, Dept Comp Sci, DK-2100 Copenhagen, Denmark
关键词
Restricted Boltzmann machines; Markov random fields; Markov chains; Gibbs sampling; Neural networks; Contrastive divergence learning; Parallel tempering; LEARNING ALGORITHM;
D O I
10.1016/j.patcog.2013.05.025
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Restricted Boltzmann machines (RBMs) are probabilistic graphical models that can be interpreted as stochastic neural networks. They have attracted much attention as building blocks for the multi-layer learning systems called deep belief networks, and variants and extensions of RBMs have found application in a wide range of pattern recognition tasks. This tutorial introduces RBMs from the viewpoint of Markov random fields, starting with the required concepts of undirected graphical models. Different learning algorithms for RBMs, including contrastive divergence learning and parallel tempering, are discussed. As sampling from RBMs, and therefore also most of their learning algorithms, are based on Markov chain Monte Carlo (MCMC) methods, an introduction to Markov chains and MCMC techniques is provided. Experiments demonstrate relevant aspects of RBM training. (C) 2013 Elsevier Ltd. All rights reserved.
引用
收藏
页码:25 / 39
页数:15
相关论文
共 50 条
  • [1] Training Restricted Boltzmann Machines
    Fischer, Asja
    [J]. KUNSTLICHE INTELLIGENZ, 2015, 29 (04): : 441 - 444
  • [2] Wasserstein Training of Restricted Boltzmann Machines
    Montavon, Gregoire
    Mueller, Klaus-Robert
    Cuturi, Marco
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 29 (NIPS 2016), 2016, 29
  • [3] On better training the infinite restricted Boltzmann machines
    Xuan Peng
    Xunzhang Gao
    Xiang Li
    [J]. Machine Learning, 2018, 107 : 943 - 968
  • [4] The Streaming Approach to Training Restricted Boltzmann Machines
    Duda, Piotr
    Rutkowski, Leszek
    Woldan, Piotr
    Najgebauer, Patryk
    [J]. ARTIFICIAL INTELLIGENCE AND SOFT COMPUTING (ICAISC 2021), PT I, 2021, 12854 : 308 - 317
  • [5] On better training the infinite restricted Boltzmann machines
    Peng, Xuan
    Gao, Xunzhang
    Li, Xiang
    [J]. MACHINE LEARNING, 2018, 107 (06) : 943 - 968
  • [6] Enhanced Gradient for Training Restricted Boltzmann Machines
    Cho, KyungHyun
    Raiko, Tapani
    Ilin, Alexander
    [J]. NEURAL COMPUTATION, 2013, 25 (03) : 805 - 831
  • [7] TRAINING RESTRICTED BOLTZMANN MACHINES WITH AUXILIARY FUNCTION APPROACH
    Kameoka, Hirokazu
    Takamune, Norihiro
    [J]. 2014 IEEE INTERNATIONAL WORKSHOP ON MACHINE LEARNING FOR SIGNAL PROCESSING (MLSP), 2014,
  • [8] Average Contrastive Divergence for Training Restricted Boltzmann Machines
    Ma, Xuesi
    Wang, Xiaojie
    [J]. ENTROPY, 2016, 18 (01):
  • [9] Infinite Swapping Algorithm for Training Restricted Boltzmann Machines
    Hult, Henrik
    Nyquist, Pierre
    Ringqvist, Carl
    [J]. MONTE CARLO AND QUASI-MONTE CARLO METHODS, MCQMC 2018, 2020, 324 : 285 - 307
  • [10] Analysis on Noisy Boltzmann Machines and Noisy Restricted Boltzmann Machines
    Lu, Wenhao
    Leung, Chi-Sing
    Sum, John
    [J]. IEEE ACCESS, 2021, 9 : 112955 - 112965