k-means as a variational EM approximation of Gaussian mixture models

被引:34
|
作者
Luecke, Joerg [1 ]
Forster, Dennis [1 ]
机构
[1] Carl von Ossietzky Univ Oldenburg, Machine Learning Lab, Ammerlander Heerstr 114-118, D-26129 Oldenburg, Germany
关键词
k-means; Gaussian mixture models; Expectation maximization; Variational methods; Free energy; ALGORITHM;
D O I
10.1016/j.patrec.2019.04.001
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We show that k-means (Lloyd's algorithm) is obtained as a special case when truncated variational EM approximations are applied to Gaussian mixture models (GMM) with isotropic Gaussians. In contrast to the standard way to relate k-means and GMMs, the provided derivation shows that it is not required to consider Gaussians with small variances or the limit case of zero variances. There are a number of consequences that directly follow from our approach: (A) k-means can be shown to increase a free energy (a.k.a. ELBO) associated with truncated distributions and this free energy can directly be reformulated in terms of the k-means objective; (B) k-means generalizations can directly be derived by considering the 2nd closest, 3rd closest etc. cluster in addition to just the closest one; and (C) the embedding of k-means into a free energy framework allows for theoretical interpretations of other k-means generalizations in the literature. In general, truncated variational EM provides a natural and rigorous quantitative link between k-means-like clustering and GMM clustering algorithms which may be very relevant for future theoretical and empirical studies. (C) 2019 Elsevier B.V. All rights reserved.
引用
收藏
页码:349 / 356
页数:8
相关论文
共 50 条
  • [41] Credit rating algorithm of corporate bonds based on Gaussian process mixture model and improved K-means
    Xia W.
    [J]. Journal of Combinatorial Mathematics and Combinatorial Computing, 2023, 117 : 159 - 168
  • [42] A SAMPLING APPROXIMATION FOR LARGE-SCALE K-MEANS
    Phoungphol, Piyaphol
    [J]. ICAART: PROCEEDINGS OF THE 4TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE, VOL 1, 2012, : 324 - 327
  • [43] An approximation algorithm for the uniform capacitated k-means problem
    Han, Lu
    Xu, Dachuan
    Du, Donglei
    Zhang, Dongmei
    [J]. JOURNAL OF COMBINATORIAL OPTIMIZATION, 2022, 44 (03) : 1812 - 1823
  • [44] Efficient k-Means plus plus Approximation with MapReduce
    Xu, Yujie
    Qu, Wenyu
    Li, Zhiyang
    Min, Geyong
    Li, Keqiu
    Liu, Zhaobin
    [J]. IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2014, 25 (12) : 3135 - 3144
  • [45] A local search approximation algorithm for k-means clustering
    Kanungo, T
    Mount, DM
    Netanyahu, NS
    Piatko, CD
    Silverman, R
    Wu, AY
    [J]. COMPUTATIONAL GEOMETRY-THEORY AND APPLICATIONS, 2004, 28 (2-3): : 89 - 112
  • [46] An efficient approximation to the K-means clustering for massive data
    Capo, Marco
    Perez, Aritz
    Lozano, Jose A.
    [J]. KNOWLEDGE-BASED SYSTEMS, 2017, 117 : 56 - 69
  • [47] An approximation algorithm for the uniform capacitated k-means problem
    Lu Han
    Dachuan Xu
    Donglei Du
    Dongmei Zhang
    [J]. Journal of Combinatorial Optimization, 2022, 44 : 1812 - 1823
  • [48] APPROXIMATION ALGORITHM FOR SPHERICAL k-MEANS PROBLEM WITH PENALTY
    Wu, Chenchen
    Lv, Wei
    Wang, Yujie
    Xu, Dachuan
    [J]. JOURNAL OF INDUSTRIAL AND MANAGEMENT OPTIMIZATION, 2022, 18 (04) : 2277 - 2287
  • [49] k-Means plus plus under approximation stability
    Agarwal, Manu
    Jaiswal, Ragesh
    Pal, Arindam
    [J]. THEORETICAL COMPUTER SCIENCE, 2015, 588 : 37 - 51
  • [50] A new iterative initialization of EM algorithm for Gaussian mixture models
    You, Jie
    Li, Zhaoxuan
    Du, Junli
    [J]. PLOS ONE, 2023, 18 (04):