Nonsmooth Penalized Clustering via lp Regularized Sparse Regression

被引:18
|
作者
Niu, Lingfeng [1 ,2 ]
Zhou, Ruizhi [1 ,2 ]
Tian, Yingjie [1 ,2 ]
Qi, Zhiquan [1 ,2 ]
Zhang, Peng [3 ]
机构
[1] Chinese Acad Sci, Res Ctr Fictitious Econ & Data Sci, Beijing 100190, Peoples R China
[2] Chinese Acad Sci, Key Lab Big Data Min & Knowledge Management, Beijing 100190, Peoples R China
[3] Univ Technol Sydney, Ctr QCIS, Sydney, NSW 2007, Australia
基金
北京市自然科学基金; 中国国家自然科学基金;
关键词
l(p)-norm; clustering analysis; nonconvex optimization; nonsmooth optimization; penalized regression; SELECTION; PERFORMANCE; MODEL; INTELLIGENCE; EVOLUTIONARY; METHODOLOGY; ALGORITHMS; SIGNALS; NUMBER; TESTS;
D O I
10.1109/TCYB.2016.2546965
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Clustering has been widely used in data analysis. A majority of existing clustering approaches assume that the number of clusters is given in advance. Recently, a novel clustering framework is proposed which can automatically learn the number of clusters from training data. Based on these works, we propose a nonsmooth penalized clustering model via l(p)( 0 < p < 1) regularized sparse regression. In particular, this model is formulated as a nonsmooth nonconvex optimization, which is based on over-parameterization and utilizes an l(p)norm-based regularization to control the tradeoff between the model fit and the number of clusters. We theoretically prove that the new model can guarantee the sparseness of cluster centers. To increase its practicality for practical use, we adhere to an easy-to-compute criterion and follow a strategy to narrow down the search interval of cross validation. To address the non-smoothness and nonconvexness of the cost function, we propose a simple smoothing trust region algorithm and present its convergent and computational complexity analysis. Numerical studies on both simulated and practical data sets provide support to our theoretical results and demonstrate the advantages of our new method.
引用
收藏
页码:1423 / 1433
页数:11
相关论文
共 50 条
  • [1] A fast algorithm for nonsmooth penalized clustering
    Zhou, Ruizhi
    Shen, Xin
    Niu, Lingfeng
    NEUROCOMPUTING, 2018, 273 : 583 - 592
  • [2] Sparse factor regression via penalized maximum likelihood estimation
    Hirose, Kei
    Imada, Miyuki
    STATISTICAL PAPERS, 2018, 59 (02) : 633 - 662
  • [3] Sparse factor regression via penalized maximum likelihood estimation
    Kei Hirose
    Miyuki Imada
    Statistical Papers, 2018, 59 : 633 - 662
  • [4] Sparse subspace clustering via smoothed lp minimization
    Dong, Wenhua
    Wu, Xiao-jun
    Kittler, Josef
    PATTERN RECOGNITION LETTERS, 2019, 125 : 206 - 211
  • [5] Subspace Clustering via Variance Regularized Ridge Regression
    Peng, Chong
    Kang, Zhao
    Cheng, Qiang
    30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 682 - 691
  • [6] Sparse regularized local regression
    Vidaurre, Diego
    Bielza, Concha
    Larranaga, Pedro
    COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2013, 62 : 122 - 135
  • [7] SPARSE REGULARIZED FUZZY REGRESSION
    Rapaic, Danilo
    Krstanovic, Lidija
    Ralevic, Nebojsa
    Obradovic, Ratko
    Klipa, Djuro
    APPLICABLE ANALYSIS AND DISCRETE MATHEMATICS, 2019, 13 (02) : 583 - 604
  • [8] On Regularized Sparse Logistic Regression
    Zhang, Mengyuan
    Liu, Kai
    23RD IEEE INTERNATIONAL CONFERENCE ON DATA MINING, ICDM 2023, 2023, : 1535 - 1540
  • [9] Nonconvex clustering via l0 fusion penalized regression
    Chen, Huangyue
    Kong, Lingchen
    Li, Yan
    PATTERN RECOGNITION, 2022, 128
  • [10] Feature selection for probabilistic load forecasting via sparse penalized quantile regression
    Yi WANG
    Dahua GAN
    Ning ZHANG
    Le XIE
    Chongqing KANG
    Journal of Modern Power Systems and Clean Energy, 2019, 7 (05) : 1200 - 1209