Improving K-means by an Agglomerative Method and Density Peaks

被引:0
|
作者
Nigro, Libero [1 ]
Cicirelli, Franco [2 ]
机构
[1] Univ Calabria, DIMES, I-87036 Arcavacata Di Rende, Italy
[2] CNR Natl Res Council Italy, Inst High Performance Comp & Networking ICAR, I-87036 Arcavacata Di Rende, Italy
关键词
Clustering problem; K-means; Agglomerative clustering; Density peaks; !text type='Java']Java[!/text; Parallel streams; Multi-core machines; Benchmark datasets; CLUSTERING-ALGORITHM;
D O I
10.1007/978-981-19-9225-4_26
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
K-means is one of the most used clustering algorithms in many application domains including image segmentation, text mining, bioinformatics, machine learning and artificial intelligence. Its strength derives from its simplicity and efficiency. K-means clustering quality, though, usually is lowdue to its "modus operandi" and local semantics, that is, its main ability to fine-tune a solution which ultimately depends on the adopted centroids' initialization method. This paper proposes a novel approach and supporting tool named ADKM which improves K-means behavior through a new centroid initialization algorithm which exploits the concepts of agglomerative clustering and density peaks. ADKM is currently implemented in Java on top of parallel streams, which can boost the execution efficiency on a multicoremachine with shared memory. The paper demonstrates by practical experiments on a collection of benchmark datasets that ADKM outperforms, by time efficiency and reliable clustering, the standard K-means algorithm, although iterated a large number of times, and its behavior is comparable to that of more sophisticated clustering algorithms. Finally, conclusions are presented together with an indication of further work.
引用
收藏
页码:343 / 359
页数:17
相关论文
共 50 条
  • [21] Improving K-means by aggregation field model
    Lu, Zhimao
    Zhang, Qi
    Massinanke, Sambourouand
    Lang, Jun
    ICIC Express Letters, 2013, 7 (08): : 2293 - 2298
  • [22] Improving the performance of k-means for color quantization
    Celebi, M. Emre
    IMAGE AND VISION COMPUTING, 2011, 29 (04) : 260 - 271
  • [23] Statistically Improving K-means Clustering Performance
    Ihsanoglu, Abdullah
    Zaval, Mounes
    32ND IEEE SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE, SIU 2024, 2024,
  • [24] Kernel Penalized K-means: A feature selection method based on Kernel K-means
    Maldonado, Sebastian
    Carrizosa, Emilio
    Weber, Richard
    INFORMATION SCIENCES, 2015, 322 : 150 - 160
  • [25] Improving K-means clustering method in fault diagnosis based on SOM network
    Chen, Anhua
    Pan, Yang
    Jiang, Lingli
    Journal of Networks, 2013, 8 (03) : 680 - 687
  • [26] A combined K-means and hierarchical clustering method for improving the clustering efficiency of microarray
    Chen, TS
    Tsai, TH
    Chen, YT
    Lin, CC
    Chen, RC
    Li, SY
    Chen, HY
    ISPACS 2005: PROCEEDINGS OF THE 2005 INTERNATIONAL SYMPOSIUM ON INTELLIGENT SIGNAL PROCESSING AND COMMUNICATION SYSTEMS, 2005, : 405 - 408
  • [27] Supervised kernel density estimation K-means
    Bortoloti, Frederico Damasceno
    de Oliveira, Elias
    Ciarelli, Patrick Marques
    EXPERT SYSTEMS WITH APPLICATIONS, 2021, 168
  • [28] Smoothed Analysis of the k-Means Method
    Arthur, David
    Manthey, Bodo
    Roeglin, Heiko
    JOURNAL OF THE ACM, 2011, 58 (05)
  • [29] How Fast is the k-means Method?
    Har-Peled, Sariel
    Sadri, Bardia
    PROCEEDINGS OF THE SIXTEENTH ANNUAL ACM-SIAM SYMPOSIUM ON DISCRETE ALGORITHMS, 2005, : 877 - 885
  • [30] The New K-Means Initialization Method
    Brejna, Bartosz
    Pietranik, Marcin
    Kozierkiewicz, Adrianna
    COMPUTATIONAL COLLECTIVE INTELLIGENCE, PT I, ICCCI 2024, 2024, 14810 : 372 - 381