Adaptive local Principal Component Analysis improves the clustering of high-dimensional data

被引:2
|
作者
Migenda, Nico [1 ,3 ]
Moeller, Ralf [2 ]
Schenck, Wolfram [1 ]
机构
[1] Bielefeld Univ Appl Sci & Arts, Ctr Appl Data Sci CfADS, Bielefeld, Germany
[2] Bielefeld Univ, Fac Technol, Comp Engn Grp, Bielefeld, Germany
[3] Schulstr 10, D-33330 Gutersloh, Germany
关键词
High-dimensional clustering; Potential function; Adaptive learning rate; Ranking criteria; Neural network-based PCA; Mixture PCA; Local PCA; LEARNING ALGORITHM; DECOMPOSITION; CONVERGENCE;
D O I
10.1016/j.patcog.2023.110030
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In local Principal Component Analysis (PCA), a distribution is approximated by multiple units, each repre-senting a local region by a hyper-ellipsoid obtained through PCA. We present an extension for local PCA which adaptively adjusts both the learning rate of each unit and the potential function which guides the competition between the local units. Our local PCA method is an online neural network method where unit centers and shapes are modified after the presentation of each data point. For several benchmark distributions, we demonstrate that our method improves the overall quality of clustering, especially for high-dimensional distributions where many conventional methods do not perform satisfactorily. Our online method is also well suited for the processing of streaming data: The two adaptive mechanisms lead to a quick reorganization of the clustering when the underlying distribution changes.
引用
收藏
页数:16
相关论文
共 50 条
  • [1] Principal component analysis for sparse high-dimensional data
    Raiko, Tapani
    Ilin, Alexander
    Karhunen, Juha
    [J]. NEURAL INFORMATION PROCESSING, PART I, 2008, 4984 : 566 - 575
  • [2] Multilevel Functional Principal Component Analysis for High-Dimensional Data
    Zipunnikov, Vadim
    Caffo, Brian
    Yousem, David M.
    Davatzikos, Christos
    Schwartz, Brian S.
    Crainiceanu, Ciprian
    [J]. JOURNAL OF COMPUTATIONAL AND GRAPHICAL STATISTICS, 2011, 20 (04) : 852 - 873
  • [3] On principal component analysis for high-dimensional XCSR
    Behdad, Mohammad
    French, Tim
    Barone, Luigi
    Bennamoun, Mohammed
    [J]. EVOLUTIONARY INTELLIGENCE, 2012, 5 (02) : 129 - 138
  • [4] Cauchy robust principal component analysis with applications to high-dimensional data sets
    Fayomi, Aisha
    Pantazis, Yannis
    Tsagris, Michail
    Wood, Andrew T. A.
    [J]. STATISTICS AND COMPUTING, 2024, 34 (01)
  • [5] Cauchy robust principal component analysis with applications to high-dimensional data sets
    Aisha Fayomi
    Yannis Pantazis
    Michail Tsagris
    Andrew T. A. Wood
    [J]. Statistics and Computing, 2024, 34
  • [6] Exploring high-dimensional biological data with sparse contrastive principal component analysis
    Boileau, Philippe
    Hejazi, Nima S.
    Dudoit, Sandrine
    [J]. BIOINFORMATICS, 2020, 36 (11) : 3422 - 3430
  • [7] High-dimensional principal component analysis with heterogeneous missingness
    Zhu, Ziwei
    Wang, Tengyao
    Samworth, Richard J.
    [J]. JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES B-STATISTICAL METHODOLOGY, 2022, 84 (05) : 2000 - 2031
  • [8] PRINCIPAL COMPONENT ANALYSIS IN VERY HIGH-DIMENSIONAL SPACES
    Lee, Young Kyung
    Lee, Eun Ryung
    Park, Byeong U.
    [J]. STATISTICA SINICA, 2012, 22 (03) : 933 - 956
  • [9] Test for high-dimensional outliers with principal component analysis
    Nakayama, Yugo
    Yata, Kazuyoshi
    Aoshima, Makoto
    [J]. JAPANESE JOURNAL OF STATISTICS AND DATA SCIENCE, 2024,
  • [10] Forecasting High-Dimensional Covariance Matrices Using High-Dimensional Principal Component Analysis
    Shigemoto, Hideto
    Morimoto, Takayuki
    [J]. AXIOMS, 2022, 11 (12)