k-means as a variational EM approximation of Gaussian mixture models

被引:34
|
作者
Luecke, Joerg [1 ]
Forster, Dennis [1 ]
机构
[1] Carl von Ossietzky Univ Oldenburg, Machine Learning Lab, Ammerlander Heerstr 114-118, D-26129 Oldenburg, Germany
关键词
k-means; Gaussian mixture models; Expectation maximization; Variational methods; Free energy; ALGORITHM;
D O I
10.1016/j.patrec.2019.04.001
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We show that k-means (Lloyd's algorithm) is obtained as a special case when truncated variational EM approximations are applied to Gaussian mixture models (GMM) with isotropic Gaussians. In contrast to the standard way to relate k-means and GMMs, the provided derivation shows that it is not required to consider Gaussians with small variances or the limit case of zero variances. There are a number of consequences that directly follow from our approach: (A) k-means can be shown to increase a free energy (a.k.a. ELBO) associated with truncated distributions and this free energy can directly be reformulated in terms of the k-means objective; (B) k-means generalizations can directly be derived by considering the 2nd closest, 3rd closest etc. cluster in addition to just the closest one; and (C) the embedding of k-means into a free energy framework allows for theoretical interpretations of other k-means generalizations in the literature. In general, truncated variational EM provides a natural and rigorous quantitative link between k-means-like clustering and GMM clustering algorithms which may be very relevant for future theoretical and empirical studies. (C) 2019 Elsevier B.V. All rights reserved.
引用
收藏
页码:349 / 356
页数:8
相关论文
共 50 条
  • [1] K-means and gaussian mixture modeling with a separation constraint
    Jiang, He
    Arias-Castro, Ery
    [J]. COMMUNICATIONS IN STATISTICS-SIMULATION AND COMPUTATION, 2024,
  • [2] The Informativeness of k-Means for Learning Mixture Models
    Liu, Zhaoqiang
    Tan, Vincent Y. F.
    [J]. 2018 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY (ISIT), 2018, : 26 - 30
  • [3] The Informativeness of k-Means for Learning Mixture Models
    Liu, Zhaoqiang
    Tan, Vincent Y. F.
    [J]. IEEE TRANSACTIONS ON INFORMATION THEORY, 2019, 65 (11) : 7460 - 7479
  • [4] In search of deterministic methods for initializing K-means and Gaussian mixture clustering
    Su, Ting
    Dy, Jennifer G.
    [J]. INTELLIGENT DATA ANALYSIS, 2007, 11 (04) : 319 - 338
  • [5] Application of k-means and Gaussian mixture model for classification of seismic activities in Istanbul
    Kuyuk, H. S.
    Yildirim, E.
    Dogan, E.
    Horasan, G.
    [J]. NONLINEAR PROCESSES IN GEOPHYSICS, 2012, 19 (04) : 411 - 419
  • [6] Variational learning for Gaussian mixture models
    Nasios, Nikolaos
    Bors, Adrian G.
    [J]. IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART B-CYBERNETICS, 2006, 36 (04): : 849 - 862
  • [7] On k-means iterations and Gaussian clusters
    de Amorim, Renato Cordeiro
    Makarenkov, Vladimir
    [J]. NEUROCOMPUTING, 2023, 553
  • [8] A refined approximation for Euclidean k-means
    Grandoni, Fabrizio
    Ostrovsky, Rafail
    Rabani, Yuval
    Schulman, Leonard J.
    Venkat, Rakesh
    [J]. INFORMATION PROCESSING LETTERS, 2022, 176
  • [9] Mixture of robust Gaussian processes and its hard-cut EM algorithm with variational bounding approximation
    Li, Tao
    Wu, Di
    Ma, Jinwen
    [J]. NEUROCOMPUTING, 2021, 452 : 224 - 238
  • [10] Can clustering scale sublinearly with its clusters? A variational EM acceleration of GMMs and k-means
    Forster, Dennis
    Luecke, Joerg
    [J]. INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 84, 2018, 84