Subspace clustering by directly solving Discriminative K-means

被引:2
|
作者
Gao, Chenhui [1 ]
Chen, Wenzhi [1 ]
Nie, Feiping [2 ]
Yu, Weizhong [2 ]
Yan, Feihu [1 ]
机构
[1] Zhejiang Univ, Hangzhou 310027, Peoples R China
[2] Northwestern Polytech Univ, Xian 710072, Peoples R China
关键词
Subspace clustering; K-means; Unsupervised learning;
D O I
10.1016/j.knosys.2022.109452
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Applications in many domains such as text mining and natural language processing need to deal with high-dimensional data. High-dimensional data may present better clustering characteristics on a selected low-dimensional subspace. Subspace clustering is to project the data onto a low-dimensional subspace before clustering. Traditional subspace clustering methods employ eigenvalue decomposition to find the projection of the input data and perform K-means or kernel K-means to obtain the clustering matrix. This kind of methods is not only inefficient, but also adopts a two-step method to generate an approximate solution. Although Discriminative K-means (DisKmeans) integrates dimensionality reduction and clustering into a joint framework and solves the optimization problem by kernel K -means, such method needs to find the centroids in the kernel space and class labels iteratively and has a square time complexity. Accordingly, in this paper, we propose an algorithm, namely Fast DisKmeans (FDKM), to obtain the cluster indicator matrix in a direct way. Moreover, our proposed method has a linear time complexity, which is a significant reduction compared with the squared time complexity of DisKmeans. We also demonstrate that solving the object function of DisKmeans is equivalent to representing the cluster assignment matrix by a low-dimensional linear mapping of the data. Based on this observation, we propose the second algorithm, namely Iterative Fast DisKmeans (IFDKM), which also has a linear time complexity. A series of experiments were conducted on several datasets, and the experimental results showed the superior performance of FDKM and IFDKM. (c) 2022 Elsevier B.V. All rights reserved.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] Discriminative k-Means Clustering
    Arandjelovic, Ognjen
    [J]. 2013 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2013,
  • [2] Subspace K-means clustering
    Marieke E. Timmerman
    Eva Ceulemans
    Kim De Roover
    Karla Van Leeuwen
    [J]. Behavior Research Methods, 2013, 45 : 1011 - 1023
  • [3] Subspace K-means clustering
    Timmerman, Marieke E.
    Ceulemans, Eva
    De Roover, Kim
    Van Leeuwen, Karla
    [J]. BEHAVIOR RESEARCH METHODS, 2013, 45 (04) : 1011 - 1023
  • [4] Discriminative K-Means Laplacian Clustering
    Chao, Guoqing
    [J]. NEURAL PROCESSING LETTERS, 2019, 49 (01) : 393 - 405
  • [5] Discriminative K-Means Laplacian Clustering
    Guoqing Chao
    [J]. Neural Processing Letters, 2019, 49 : 393 - 405
  • [6] Fuzzy K-Means Clustering With Discriminative Embedding
    Nie, Feiping
    Zhao, Xiaowei
    Wang, Rong
    Li, Xuelong
    Li, Zhihui
    [J]. IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2022, 34 (03) : 1221 - 1230
  • [7] A Heuristically Weighting K-Means Algorithm for Subspace Clustering
    Li, Boyang
    Jiang, Qingshan
    Chen, Lifei
    [J]. 2008 2ND INTERNATIONAL CONFERENCE ON ANTI-COUNTERFEITING, SECURITY AND IDENTIFICATION, 2008, : 268 - +
  • [8] Supplier categorization with K-means type subspace clustering
    Zhang, XJ
    Huang, JZ
    Qian, DP
    Xu, J
    Jing, LP
    [J]. FRONTIERS OF WWW RESEARCH AND DEVELOPMENT - APWEB 2006, PROCEEDINGS, 2006, 3841 : 226 - 237
  • [9] Deep Clustering: On the Link Between Discriminative Models and K-Means
    Jabi, Mohammed
    Pedersoli, Marco
    Mitiche, Amar
    Ayed, Ismail Ben
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2021, 43 (06) : 1887 - 1896
  • [10] On the performance of feature weighting K-means for text subspace clustering
    Jing, LP
    Ng, MK
    Xu, J
    Huang, JZX
    [J]. ADVANCES IN WEB-AGE INFORMATION MANAGEMENT, PROCEEDINGS, 2005, 3739 : 502 - 512