Auxiliary gradient-based sampling algorithms

被引:16
|
作者
Titsias, Michalis K. [1 ]
Papaspiliopoulos, Omiros [2 ,3 ,4 ]
机构
[1] Athens Univ Econ & Business, Athens, Greece
[2] Inst Catalana Recerca Estudis Avancats, Barcelona, Spain
[3] Univ Pompeu Fabra, Barcelona, Spain
[4] Barcelona Grad Sch Econ, Barcelona, Spain
关键词
Latent Gaussian models; Machine learning; Markov chain Monte Carlo sampling; Peskun ordering; Scalability; MCMC METHODS; METROPOLIS;
D O I
10.1111/rssb.12269
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
We introduce a new family of Markov chain Monte Carlo samplers that combine auxiliary variables, Gibbs sampling and Taylor expansions of the target density. Our approach permits the marginalization over the auxiliary variables, yielding marginal samplers, or the augmentation of the auxiliary variables, yielding auxiliary samplers. The well-known Metropolis-adjusted Langevin algorithm MALA and preconditioned Crank-Nicolson-Langevin algorithm pCNL are shown to be special cases. We prove that marginal samplers are superior in terms of asymptotic variance and demonstrate cases where they are slower in computing time compared with auxiliary samplers. In the context of latent Gaussian models we propose new auxiliary and marginal samplers whose implementation requires a single tuning parameter, which can be found automatically during the transient phase. Extensive experimentation shows that the increase in efficiency (measured as the effective sample size per unit of computing time) relative to (optimized implementations of) pCNL, elliptical slice sampling and MALA ranges from tenfold in binary classification problems to 25 fold in log-Gaussian Cox processes to 100 fold in Gaussian process regression, and it is on a par with Riemann manifold Hamiltonian Monte Carlo sampling in an example where that algorithm has the same complexity as the aforementioned algorithms. We explain this remarkable improvement in terms of the way that alternative samplers try to approximate the eigenvalues of the target. We introduce a novel Markov chain Monte Carlo sampling scheme for hyperparameter learning that builds on the auxiliary samplers. The MATLAB code for reproducing the experiments in the paper is publicly available and an on-line supplement to this paper contains additional experiments and implementation details.
引用
收藏
页码:749 / 767
页数:19
相关论文
共 50 条
  • [1] Gradient-based Algorithms for Machine Teaching
    Wang, Pei
    Nagrecha, Kabir
    Vasconcelos, Nuno
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 1387 - 1396
  • [2] Yield prediction for crops by gradient-based algorithms
    Mahesh, Pavithra
    Soundrapandiyan, Rajkumar
    [J]. PLOS ONE, 2024, 19 (08):
  • [3] Gradient-based genetic algorithms in image registration
    Maslov, IV
    Gertner, I
    [J]. AUTOMATIC TARGET RECOGNITION XI, 2001, 4379 : 509 - 520
  • [4] Orthogonal wavelets and stochastic gradient-based algorithms
    Attallah, S
    Najim, M
    [J]. PROCEEDINGS OF THE IEEE-SP INTERNATIONAL SYMPOSIUM ON TIME-FREQUENCY AND TIME-SCALE ANALYSIS, 1996, : 405 - 408
  • [5] Gradient-based Sampling: An Adaptive Importance Sampling for Least-squares
    Zhu, Rong
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 29 (NIPS 2016), 2016, 29
  • [6] Robust multiscale algorithms for gradient-based motion estimation
    Lu, Qing-Hua
    Zhang, Xian-Min
    [J]. INTERNATIONAL JOURNAL OF IMAGING SYSTEMS AND TECHNOLOGY, 2007, 17 (06) : 333 - 340
  • [7] Selective coefficient update of gradient-based adaptive algorithms
    Aboulnasr, T
    Mayyas, K
    [J]. 1997 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOLS I - V: VOL I: PLENARY, EXPERT SUMMARIES, SPECIAL, AUDIO, UNDERWATER ACOUSTICS, VLSI; VOL II: SPEECH PROCESSING; VOL III: SPEECH PROCESSING, DIGITAL SIGNAL PROCESSING; VOL IV: MULTIDIMENSIONAL SIGNAL PROCESSING, NEURAL NETWORKS - VOL V: STATISTICAL SIGNAL AND ARRAY PROCESSING, APPLICATIONS, 1997, : 1929 - 1932
  • [8] Empirical Characterization of Discretization Error in Gradient-based Algorithms
    Bachrach, Jonathan
    Beal, Jacob
    Horowitz, Joshua
    Qumsiyeh, Dany
    [J]. SASO 2008: SECOND IEEE INTERNATIONAL CONFERENCE ON SELF-ADAPTIVE AND SELF-ORGANIZING SYSTEMS, PROCEEDINGS, 2008, : 203 - 212
  • [9] Hybridization of the multi-objective evolutionary algorithms and the gradient-based algorithms
    Hu, XL
    Huang, ZC
    Wang, ZF
    [J]. CEC: 2003 CONGRESS ON EVOLUTIONARY COMPUTATION, VOLS 1-4, PROCEEDINGS, 2003, : 870 - 877
  • [10] Gradient-based optimization algorithms for networks of reconfigurable sensors
    de Groot, T. H.
    Krasnov, O. A.
    Yarovoy, A. G.
    [J]. CONTROL ENGINEERING PRACTICE, 2014, 29 : 74 - 85