Simple approximate MAP inference for Dirichlet processes mixtures

被引:14
|
作者
Raykov, Yordan P. [1 ]
Boukouvalas, Alexis [2 ]
Little, Max A. [1 ,3 ]
机构
[1] Aston Univ, Birmingham, W Midlands, England
[2] Univ Manchester, Manchester, Lancs, England
[3] MIT, Cambridge, MA 02139 USA
来源
ELECTRONIC JOURNAL OF STATISTICS | 2016年 / 10卷 / 02期
关键词
Bayesian nonparametrics; clustering; Gaussian mixture model;
D O I
10.1214/16-EJS1196
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
The Dirichlet process mixture model (DPMM) is a ubiquitous, flexible Bayesian nonparametric statistical model. However, full probabilistic inference in this model is analytically intractable, so that computationally intensive techniques such as Gibbs sampling are required. As a result, DPMM-based methods, which have considerable potential, are restricted to applications in which computational resources and time for inference is plentiful. For example, they would not be practical for digital signal processing on embedded hardware, where computational resources are at a serious premium. Here, we develop a simplified yet statistically rigorous approximate maximum a-posteriori (MAP) inference algorithm for DPMMs. This algorithm is as simple as DP-means clustering, solves the MAP problem as well as Gibbs sampling, while requiring only a fraction of the computational effort.+ Unlike related small variance asymptotics (SVA), our method is non-degenerate and so inherits the "rich get richer" property of the Dirichlet process. It also retains a non-degenerate closed-form likelihood which enables out-of-sample calculations and the use of standard tools such as cross-validation. We illustrate the benefits of our algorithm on a range of examples and contrast it to variational, SVA and sampling approaches from both a computational complexity perspective as well as in terms of clustering performance. We demonstrate the wide applicabiity of our approach by presenting an approximate MAP inference method for the infinite hidden Markov model whose performance contrasts favorably with a recently proposed hybrid SVA approach. Similarly, we show how our algorithm can applied to a semiparametric mixed-effects regression model where the random effects distribution is modelled using an infinite mixture model, as used in longitudinal progression modelling in population health science. Finally, we propose directions for future research on approximate MAP inference in Bayesian nonparametrics.
引用
收藏
页码:3548 / 3578
页数:31
相关论文
共 50 条
  • [1] Variational Inference for Dirichlet Process Mixtures
    Blei, David M.
    Jordan, Michael I.
    BAYESIAN ANALYSIS, 2006, 1 (01): : 121 - 143
  • [2] Mixtures of Dirichlet processes according to a Dirichlet process
    Carota, C
    AMERICAN STATISTICAL ASSOCIATION - 1996 PROCEEDINGS OF THE SECTION ON BAYESIAN STATISTICAL SCIENCE, 1996, : 310 - 313
  • [3] COMPUTATIONS OF MIXTURES OF DIRICHLET PROCESSES
    KUO, L
    SIAM JOURNAL ON SCIENTIFIC AND STATISTICAL COMPUTING, 1986, 7 (01): : 60 - 71
  • [4] Dependent mixtures of Dirichlet processes
    Hatjispyros, Spyridon J.
    Nicoleris, Theodoros
    Walker, Stephen G.
    COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2011, 55 (06) : 2011 - 2025
  • [5] Fast approximate inference for variable selection in Dirichlet process mixtures, with an application to pan-cancer proteomics
    Crook, Oliver M.
    Gatto, Laurent
    Kirk, Paul D. W.
    STATISTICAL APPLICATIONS IN GENETICS AND MOLECULAR BIOLOGY, 2019, 18 (06)
  • [6] Exactness of Approximate MAP Inference in Continuous MRFs
    Ruozzi, Nicholas
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 28 (NIPS 2015), 2015, 28
  • [7] Supervised Hierarchical Dirichlet Processes with Variational Inference
    Zhang, Cheng
    Ek, Carl Henrik
    Gratal, Xavi
    Pokorny, Florian T.
    Kjellstrom, Hedvig
    2013 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW), 2013, : 254 - 261
  • [8] Hybrid Parallel Inference for Hierarchical Dirichlet Processes
    Omoto, Tsukasa
    Eguchi, Koji
    Tora, Shotaro
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2014, E97D (04): : 815 - 820
  • [9] Bayesian inference for dynamic models with dirichlet process mixtures
    Caron, Francois
    Davy, Manuel
    Doucet, Arnaud
    Duflos, Emmanuel
    Vanheeghe, Philippe
    2006 9th International Conference on Information Fusion, Vols 1-4, 2006, : 138 - 145
  • [10] Dirichlet to Neumann map for domains with corners and approximate boundary conditions
    Halpern, Laurence
    Lafitte, Olivier
    JOURNAL OF COMPUTATIONAL AND APPLIED MATHEMATICS, 2007, 204 (02) : 505 - 514