On convergence and parameter selection of the EM and DA-EM algorithms for Gaussian mixtures

被引:27
|
作者
Yu, Jian [1 ]
Chaomurilige, Chaomu [1 ]
Yang, Miin-Shen [2 ]
机构
[1] Beijing Jiaotong Univ, Beijing Key Lab Traff Data Anal & Min, Beijing, Peoples R China
[2] Chung Yuan Christian Univ, Dept Appl Math, Chungli 32023, Taiwan
关键词
Expectation & maximization (EM) algorithm; Deterministic annealing EM (DA-EM); GAUSSIAN mixtures; Self-annealing; Convergence; Parameter selection; MAXIMUM-LIKELIHOOD; MODELS;
D O I
10.1016/j.patcog.2017.12.014
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The expectation & maximization (EM) for Gaussian mixtures is popular as a clustering algorithm. However, the EM algorithm is sensitive to initial values, and so Ueda and Nakano [4] proposed the deterministic annealing EM (DA-EM) algorithm to improve it. In this paper, we investigate theoretical behaviors of the EM and DA-EM algorithms. We first derive a general Jacobian matrix of the DA-EM algorithm with respect to posterior probabilities. We then propose a theoretical lower bound for initialization of the annealing parameter in the DA-EM algorithm. On the other hand, some researches mentioned that the EM algorithm exhibits a self-annealing behavior, that is, the equal posterior probability with small random perturbations can avoid the EM algorithm to output the mass center for Gaussian mixtures. However, there is no theoretical analysis on this self-annealing property. Since the DA-EM will become the EM when the annealing parameter is 1, according to the Jacobian matrix of the DA-EM, we can prove the self-annealing property of the EM algorithm for Gaussian mixtures. Based on these results, we give not only convergence behaviors of the equal posterior probabilities and initialization lower bound of the temperature parameter of the DA-EM, but also a theoretical explanation why the EM algorithm for Gaussian mixtures exhibits a self-annealing behavior. (C) 2017 Elsevier Ltd. All rights reserved.
引用
收藏
页码:188 / 203
页数:16
相关论文
共 50 条
  • [41] The EM Algorithm is Adaptively-Optimal for Unbalanced Symmetric Gaussian Mixtures
    Weinberger, Nir
    Bresler, Guy
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2021, 22
  • [42] A Precise Hard-Cut EM Algorithm for Mixtures of Gaussian Processes
    Chen, Ziyi
    Ma, Jinwen
    Zhou, Yatong
    [J]. INTELLIGENT COMPUTING METHODOLOGIES, 2014, 8589 : 68 - 75
  • [43] Global Convergence of the EM Algorithm for Mixtures of Two Component Linear Regression
    Kwon, Jeongyeol
    Qian, Wei
    Caramanis, Constantine
    Chen, Yudong
    Davis, Damek
    [J]. CONFERENCE ON LEARNING THEORY, VOL 99, 2019, 99
  • [44] EM algorithms for multivariate Gaussian mixture models with truncated and censored data
    Lee, Gyemin
    Scott, Clayton
    [J]. COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2012, 56 (09) : 2816 - 2829
  • [45] New EM-type algorithms for the Heckman selection model
    Zhao, Jun
    Kim, Hea-Jung
    Kim, Hyoung-Moon
    [J]. COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2020, 146
  • [46] CONVERGENCE IN NORM FOR ALTERNATING EXPECTATION-MAXIMIZATION (EM) TYPE ALGORITHMS
    HERO, AO
    FESSLER, JA
    [J]. STATISTICA SINICA, 1995, 5 (01) : 41 - 54
  • [47] Approximate EM algorithms for parameter and state estimation in nonlinear stochastic models
    Goodwin, Graham C.
    Aguero, Juan C.
    [J]. 2005 44TH IEEE CONFERENCE ON DECISION AND CONTROL & EUROPEAN CONTROL CONFERENCE, VOLS 1-8, 2005, : 368 - 373
  • [48] Towards EM-style algorithms for a posteriori optimization of normal mixtures
    Ristad, ES
    Yianilos, PN
    [J]. 1998 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY - PROCEEDINGS, 1998, : 241 - 241
  • [49] Diffusion Scheme of Distributed EM Algorithm for Gaussian Mixtures over Random Networks
    Weng, Yang
    Xie, Lihua
    Xiao, Wendong
    [J]. 2009 IEEE INTERNATIONAL CONFERENCE ON CONTROL AND AUTOMATION, VOLS 1-3, 2009, : 1529 - +
  • [50] An effective EM algorithm for mixtures of Gaussian processes via the MCMC sampling and approximation
    Wu, Di
    Ma, Jinwen
    [J]. NEUROCOMPUTING, 2019, 331 : 366 - 374