Using principal component analysis and correspondence analysis for estimation in latent variable models

被引:10
|
作者
Lynn, HS [1 ]
McCulloch, CE
机构
[1] Rho Inc, Chapel Hill, NC 27514 USA
[2] Cornell Univ, Biometr Unit, Dept Stat Sci, Ithaca, NY 14853 USA
关键词
consistency; correspondence analysis; incidental parameters; principal component analysis;
D O I
10.2307/2669399
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
Correspondence analysis (CA) and principal component analysis (PCA) are often used to describe multivariate data. In certain applications they have been used for estimation in latent variable models. The theoretical basis for such inference is assessed in generalized linear models where the linear predictor equals alpha(j) + x(i)beta(j) or a(j) - b(j) (x(i) - u(j))(2), (i = 1, ..., n; j = 1, ..., m), and x(i) is treated as a latent fixed effect. The PCA and CA eigenvectors/column scores are evaluated as estimators of beta(j) and u(j) and as estimators of u(j). With m fixed and n up arrow infinity, consistent estimators cannot be obtained due to the incidental parameters problem unless sufficient "moment" conditions are imposed on x(i). PCA is equivalent to maximum likelihood estimation for the linear Gaussian model and gives a consistent estimator of beta(j) (up to a scale change) when the second sample moment of x(i) is positive and finite in the limit. It is inconsistent for Poisson and Bernoulli distributions, but when b(j) is constant, its first and/or second eigenvectors can consistently estimate u(j) (up to a location and scale change) for the quadratic Gaussian model. In contrast, the CA estimator is always inconsistent. For finite samples, however, the CA column scores often have high correlations with the u(j)'s, especially when the response curves are spread out relative to one another. The correlations obtained from PCA are usually weaker, although the second PCA eigenvector can sometimes do much better than the first eigenvector, and for incidence data with tightly clustered response curves its performance is comparable to that of CA. For small sample sizes, PCA and particularly CA are competitive alternatives to maximum likelihood and may be preferred because of their computational ease.
引用
收藏
页码:561 / 572
页数:12
相关论文
共 50 条
  • [41] Application of the Principal Component Analysis for the reduction of variable delayed
    Dasal, Kazimierz
    Poplawski, Tomasz
    Rusek, Karolina
    PRZEGLAD ELEKTROTECHNICZNY, 2012, 88 (9A): : 130 - 133
  • [42] PRINCIPAL COMPONENT ANALYSIS OF KINETIC-MODELS
    VAJDA, S
    VALKO, P
    TURANYI, T
    INTERNATIONAL JOURNAL OF CHEMICAL KINETICS, 1985, 17 (01) : 55 - 81
  • [43] PROJECTED PRINCIPAL COMPONENT ANALYSIS IN FACTOR MODELS
    Fan, Jianqing
    Liao, Yuan
    Wang, Weichen
    ANNALS OF STATISTICS, 2016, 44 (01): : 219 - 254
  • [44] Watermarking of 3D models using principal component analysis
    Kalivas, A
    Tefas, A
    Pitas, I
    2003 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOL V, PROCEEDINGS: SENSOR ARRAY & MULTICHANNEL SIGNAL PROCESSING AUDIO AND ELECTROACOUSTICS MULTIMEDIA SIGNAL PROCESSING, 2003, : 676 - 679
  • [45] Watermarking of 3D models using principal component analysis
    Kalivas, A
    Tefas, A
    Pitas, I
    2003 INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, VOL I, PROCEEDINGS, 2003, : 637 - 640
  • [46] Identification of latent variables in a semantic odor profile database using principal component analysis
    Zarzo, Manuel
    Stanton, David T.
    CHEMICAL SENSES, 2006, 31 (08) : 713 - 724
  • [47] Identification of latent variables in a semantic odor profile database using principal component analysis
    Zarzo, M.
    Stanton, D.
    CHEMICAL SENSES, 2006, 31 (05) : A144 - A144
  • [48] Determination of component concentrations in models of exhaled air samples using principal component analysis and canonical correlation analysis
    Kistenev, Yu. V.
    Borisov, A. V.
    Shapovalov, A. V.
    INTERNATIONAL CONFERENCE ON ATOMIC AND MOLECULAR PULSED LASERS XII, 2015, 9810
  • [49] Latent Variable Models for Hippocampal Sequence Analysis
    Ackermann, Etienne
    Kemere, Caleb
    Maboudi, Kourosh
    Diba, Kamran
    2017 FIFTY-FIRST ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS, AND COMPUTERS, 2017, : 724 - 728
  • [50] Integrative Analysis using Coupled Latent Variable Models for Individualizing Prognoses
    Sehulam, Peter
    Saria, Suchi
    JOURNAL OF MACHINE LEARNING RESEARCH, 2016, 17 : 1 - 35