Subspace K-means clustering

被引:0
|
作者
Marieke E. Timmerman
Eva Ceulemans
Kim De Roover
Karla Van Leeuwen
机构
[1] University of Groningen,Heymans Institute for Psychology, Psychometrics & Statistics
[2] K.U. Leuven,Educational Sciences
[3] K.U. Leuven,Parenting and Special Education
来源
Behavior Research Methods | 2013年 / 45卷
关键词
Cluster analysis; Cluster recovery; Multivariate data; Reduced ; -means; means; Factorial ; -means; Mixtures of factor analyzers; MCLUST;
D O I
暂无
中图分类号
学科分类号
摘要
To achieve an insightful clustering of multivariate data, we propose subspace K-means. Its central idea is to model the centroids and cluster residuals in reduced spaces, which allows for dealing with a wide range of cluster types and yields rich interpretations of the clusters. We review the existing related clustering methods, including deterministic, stochastic, and unsupervised learning approaches. To evaluate subspace K-means, we performed a comparative simulation study, in which we manipulated the overlap of subspaces, the between-cluster variance, and the error variance. The study shows that the subspace K-means algorithm is sensitive to local minima but that the problem can be reasonably dealt with by using partitions of various cluster procedures as a starting point for the algorithm. Subspace K-means performs very well in recovering the true clustering across all conditions considered and appears to be superior to its competitor methods: K-means, reduced K-means, factorial K-means, mixtures of factor analyzers (MFA), and MCLUST. The best competitor method, MFA, showed a performance similar to that of subspace K-means in easy conditions but deteriorated in more difficult ones. Using data from a study on parental behavior, we show that subspace K-means analysis provides a rich insight into the cluster characteristics, in terms of both the relative positions of the clusters (via the centroids) and the shape of the clusters (via the within-cluster residuals).
引用
收藏
页码:1011 / 1023
页数:12
相关论文
共 50 条
  • [41] STRONG CONSISTENCY OF K-MEANS CLUSTERING
    POLLARD, D
    ANNALS OF STATISTICS, 1981, 9 (01): : 135 - 140
  • [42] An Improved K-means Clustering Algorithm
    Wang Yintong
    Li Wanlong
    Gao Rujia
    2012 WORLD AUTOMATION CONGRESS (WAC), 2012,
  • [43] Granular K-means Clustering Algorithm
    Zhou, Chenglong
    Chen, Yuming
    Zhu, Yidong
    Computer Engineering and Applications, 2023, 59 (13) : 317 - 324
  • [44] Locally Private k-Means Clustering
    Stemmer, Uri
    PROCEEDINGS OF THE THIRTY-FIRST ANNUAL ACM-SIAM SYMPOSIUM ON DISCRETE ALGORITHMS (SODA'20), 2020, : 548 - 559
  • [45] k-means clustering for persistent homology
    Cao, Yueqi
    Leung, Prudence
    Monod, Anthea
    ADVANCES IN DATA ANALYSIS AND CLASSIFICATION, 2024,
  • [46] The global k-means clustering algorithm
    Likas, A
    Vlassis, N
    Verbeek, JJ
    PATTERN RECOGNITION, 2003, 36 (02) : 451 - 461
  • [47] K-Means Clustering With Incomplete Data
    Wang, Siwei
    Li, Miaomiao
    Hu, Ning
    Zhu, En
    Hu, Jingtao
    Liu, Xinwang
    Yin, Jianping
    IEEE ACCESS, 2019, 7 : 69162 - 69171
  • [48] K-means - Laplacian clustering revisited
    Rengasamy, Sundar
    Murugesan, Punniyamoorthy
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2022, 107
  • [49] Improved K-means clustering algorithm
    Zhang, Zhe
    Zhang, Junxi
    Xue, Huifeng
    CISP 2008: FIRST INTERNATIONAL CONGRESS ON IMAGE AND SIGNAL PROCESSING, VOL 5, PROCEEDINGS, 2008, : 169 - 172
  • [50] The LINEX Weighted k-Means Clustering
    Ahmadzadehgoli, Narges
    Mohammadpour, Adel
    Behzadi, Mohammad Hassan
    JOURNAL OF STATISTICAL THEORY AND APPLICATIONS, 2019, 18 (02): : 147 - 154