On convergence and parameter selection of the EM and DA-EM algorithms for Gaussian mixtures

被引:27
|
作者
Yu, Jian [1 ]
Chaomurilige, Chaomu [1 ]
Yang, Miin-Shen [2 ]
机构
[1] Beijing Jiaotong Univ, Beijing Key Lab Traff Data Anal & Min, Beijing, Peoples R China
[2] Chung Yuan Christian Univ, Dept Appl Math, Chungli 32023, Taiwan
关键词
Expectation & maximization (EM) algorithm; Deterministic annealing EM (DA-EM); GAUSSIAN mixtures; Self-annealing; Convergence; Parameter selection; MAXIMUM-LIKELIHOOD; MODELS;
D O I
10.1016/j.patcog.2017.12.014
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The expectation & maximization (EM) for Gaussian mixtures is popular as a clustering algorithm. However, the EM algorithm is sensitive to initial values, and so Ueda and Nakano [4] proposed the deterministic annealing EM (DA-EM) algorithm to improve it. In this paper, we investigate theoretical behaviors of the EM and DA-EM algorithms. We first derive a general Jacobian matrix of the DA-EM algorithm with respect to posterior probabilities. We then propose a theoretical lower bound for initialization of the annealing parameter in the DA-EM algorithm. On the other hand, some researches mentioned that the EM algorithm exhibits a self-annealing behavior, that is, the equal posterior probability with small random perturbations can avoid the EM algorithm to output the mass center for Gaussian mixtures. However, there is no theoretical analysis on this self-annealing property. Since the DA-EM will become the EM when the annealing parameter is 1, according to the Jacobian matrix of the DA-EM, we can prove the self-annealing property of the EM algorithm for Gaussian mixtures. Based on these results, we give not only convergence behaviors of the equal posterior probabilities and initialization lower bound of the temperature parameter of the DA-EM, but also a theoretical explanation why the EM algorithm for Gaussian mixtures exhibits a self-annealing behavior. (C) 2017 Elsevier Ltd. All rights reserved.
引用
收藏
页码:188 / 203
页数:16
相关论文
共 50 条
  • [1] Parameter convergence for EM and MM algorithms
    Vaida, F
    [J]. STATISTICA SINICA, 2005, 15 (03) : 831 - 840
  • [2] On convergence properties of the EM algorithm for gaussian mixtures
    Xu, L
    Jordan, MI
    [J]. NEURAL COMPUTATION, 1996, 8 (01) : 129 - 151
  • [4] On the correct convergence of the EM algorithm for Gaussian mixtures
    Ma, JW
    Fu, SQ
    [J]. PATTERN RECOGNITION, 2005, 38 (12) : 2602 - 2611
  • [5] Singularity and Slow Convergence of the EM algorithm for Gaussian Mixtures
    Park, Hyeyoung
    Ozeki, Tomoko
    [J]. NEURAL PROCESSING LETTERS, 2009, 29 (01) : 45 - 59
  • [6] Asymptotic convergence rate of the EM algorithm for Gaussian mixtures
    Ma, JW
    Xu, L
    Jordan, MI
    [J]. NEURAL COMPUTATION, 2000, 12 (12) : 2881 - 2907
  • [7] Research on correct convergence of the EM algorihm for Gaussian mixtures
    Fu, SQ
    Cao, BY
    Ma, JW
    [J]. ICONIP'02: PROCEEDINGS OF THE 9TH INTERNATIONAL CONFERENCE ON NEURAL INFORMATION PROCESSING: COMPUTATIONAL INTELLIGENCE FOR THE E-AGE, 2002, : 2660 - 2664
  • [8] Singularity and Slow Convergence of the EM algorithm for Gaussian Mixtures
    Hyeyoung Park
    Tomoko Ozeki
    [J]. Neural Processing Letters, 2009, 29 : 45 - 59
  • [9] EM algorithms for Gaussian mixtures with split-and-merge operation
    Zhang, ZH
    Chen, CB
    Sun, J
    Chan, KL
    [J]. PATTERN RECOGNITION, 2003, 36 (09) : 1973 - 1983
  • [10] Improved convergence guarantees for learning Gaussian mixture models by EM and gradient EM
    Segol, Nimrod
    Nadler, Boaz
    [J]. ELECTRONIC JOURNAL OF STATISTICS, 2021, 15 (02): : 4510 - 4544