An augmented Lagrangian approach for sparse principal component analysis

被引:64
|
作者
Lu, Zhaosong [1 ]
Zhang, Yong [1 ]
机构
[1] Simon Fraser Univ, Dept Math, Burnaby, BC V5A 1S6, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
Sparse PCA; Augmented Lagrangian method; Nonmonotone gradient methods; Nonsmooth minimization; GRADIENT METHODS; ALGORITHM;
D O I
10.1007/s10107-011-0452-4
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Principal component analysis (PCA) is a widely used technique for data analysis and dimension reduction with numerous applications in science and engineering. However, the standard PCA suffers from the fact that the principal components (PCs) are usually linear combinations of all the original variables, and it is thus often difficult to interpret the PCs. To alleviate this drawback, various sparse PCA approaches were proposed in the literature (Cadima and Jolliffe in J Appl Stat 22:203-214, 1995; d'Aspremont et al. in J Mach Learn Res 9:1269-1294, 2008; d'Aspremont et al. SIAM Rev 49:434-448, 2007; Jolliffe in J Appl Stat 22:29-35, 1995; Journ,e et al. in J Mach Learn Res 11:517-553, 2010; Jolliffe et al. in J Comput Graph Stat 12:531-547, 2003; Moghaddam et al. in Advances in neural information processing systems 18:915-922, MIT Press, Cambridge, 2006; Shen and Huang in J Multivar Anal 99(6):1015-1034, 2008; Zou et al. in J Comput Graph Stat 15(2):265-286, 2006). Despite success in achieving sparsity, some important properties enjoyed by the standard PCA are lost in these methods such as uncorrelation of PCs and orthogonality of loading vectors. Also, the total explained variance that they attempt to maximize can be too optimistic. In this paper we propose a new formulation for sparse PCA, aiming at finding sparse and nearly uncorrelated PCs with orthogonal loading vectors while explaining as much of the total variance as possible. We also develop a novel augmented Lagrangian method for solving a class of nonsmooth constrained optimization problems, which is well suited for our formulation of sparse PCA. We show that it converges to a feasible point, and moreover under some regularity assumptions, it converges to a stationary point. Additionally, we propose two nonmonotone gradient methods for solving the augmented Lagrangian subproblems, and establish their global and local convergence. Finally, we compare our sparse PCA approach with several existing methods on synthetic (Zou et al. in J Comput Graph Stat 15(2):265-286, 2006), Pitprops (Jeffers in Appl Stat 16:225-236, 1967), and gene expression data (Chin et al in Cancer Cell 10:529C-541C, 2006), respectively. The computational results demonstrate that the sparse PCs produced by our approach substantially outperform those by other methods in terms of total explained variance, correlation of PCs, and orthogonality of loading vectors. Moreover, the experiments on random data show that our method is capable of solving large-scale problems within a reasonable amount of time.
引用
收藏
页码:149 / 193
页数:45
相关论文
共 50 条
  • [1] An augmented Lagrangian approach for sparse principal component analysis
    Zhaosong Lu
    Yong Zhang
    [J]. Mathematical Programming, 2012, 135 : 149 - 193
  • [2] An exact approach to sparse principal component analysis
    Alessio Farcomeni
    [J]. Computational Statistics, 2009, 24 : 583 - 604
  • [3] An exact approach to sparse principal component analysis
    Farcomeni, Alessio
    [J]. COMPUTATIONAL STATISTICS, 2009, 24 (04) : 583 - 604
  • [4] Sparse principal component analysis
    Zou, Hui
    Hastie, Trevor
    Tibshirani, Robert
    [J]. JOURNAL OF COMPUTATIONAL AND GRAPHICAL STATISTICS, 2006, 15 (02) : 265 - 286
  • [5] A block coordinate descent approach for sparse principal component analysis
    Zhao, Qian
    Meng, Deyu
    Xu, Zongben
    Gao, Chenqiang
    [J]. NEUROCOMPUTING, 2015, 153 : 180 - 190
  • [6] A plug-in approach to sparse and robust principal component analysis
    Luca Greco
    Alessio Farcomeni
    [J]. TEST, 2016, 25 : 449 - 481
  • [7] A plug-in approach to sparse and robust principal component analysis
    Greco, Luca
    Farcomeni, Alessio
    [J]. TEST, 2016, 25 (03) : 449 - 481
  • [8] Robust sparse principal component analysis
    ZHAO Qian
    MENG DeYu
    XU ZongBen
    [J]. Science China(Information Sciences), 2014, 57 (09) : 175 - 188
  • [9] Multilinear Sparse Principal Component Analysis
    Lai, Zhihui
    Xu, Yong
    Chen, Qingcai
    Yang, Jian
    Zhang, David
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2014, 25 (10) : 1942 - 1950
  • [10] Robust Sparse Principal Component Analysis
    Croux, Christophe
    Filzmoser, Peter
    Fritz, Heinrich
    [J]. TECHNOMETRICS, 2013, 55 (02) : 202 - 214