Tensor Decompositions for Learning Latent Variable Models

被引:0
|
作者
Anandkumar, Animashree [1 ]
Ge, Rong [2 ]
Hsu, Daniel [3 ]
Kakade, Sham M. [2 ]
Telgarsky, Matus [4 ]
机构
[1] Univ Calif Irvine, Irvine, CA 92697 USA
[2] Microsoft Res, Cambridge, MA 02142 USA
[3] Columbia Univ, Dept Comp Sci, New York, NY 10027 USA
[4] Rutgers State Univ, Dept Stat, Piscataway, NJ 08854 USA
关键词
latent variable models; tensor decompositions; mixture models; topic models; method of moments; power method; INDEPENDENT COMPONENT ANALYSIS; FIXED-POINT ALGORITHMS; MAXIMUM-LIKELIHOOD; MIXTURES; EM; IDENTIFIABILITY; APPROXIMATION; EIGENVALUES; RANK;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This work considers a computationally and statistically efficient parameter estimation method for a wide class of latent variable models-including Gaussian mixture models, hidden Markov models, and latent Dirichlet allocation-which exploits a certain tensor structure in their low-order observable moments (typically, of second- and third-order). Specifically, parameter estimation is reduced to the problem of extracting a certain (orthogonal) decomposition of a symmetric tensor derived from the moments; this decomposition can be viewed as a natural generalization of the singular value decomposition for matrices. Although tensor decompositions are generally intractable to compute, the decomposition of these specially structured tensors can be efficiently obtained by a variety of approaches, including power iterations and maximization approaches (similar to the case of matrices). A detailed analysis of a robust tensor power method is provided, establishing an analogue of Wedin's perturbation theorem for the singular vectors of matrices. This implies a robust and computationally tractable estimation approach for several popular latent variable models.
引用
收藏
页码:2773 / 2832
页数:60
相关论文
共 50 条
  • [21] Diversity-Promoting Bayesian Learning of Latent Variable Models
    Xie, Pengtao
    Zhu, Jun
    Xing, Eric P.
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 48, 2016, 48
  • [22] A COMPARISON OF DISCRETE LATENT VARIABLE MODELS FOR SPEECH REPRESENTATION LEARNING
    Zhou, Henry
    Baevski, Alexei
    Auli, Michael
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 3050 - 3054
  • [23] Empirical Study of the Benefits of Overparameterization in Learning Latent Variable Models
    Buhai, Rares-Darius
    Halpern, Yoni
    Kim, Yoon
    Risteski, Andrej
    Sontag, David
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
  • [24] FlowPrior: Learning Expressive Priors for Latent Variable Sentence Models
    Ding, Xiaoan
    Gimpel, Kevin
    2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), 2021, : 3242 - 3258
  • [25] Flexible latent variable models for multi-task learning
    Zhang, Jian
    Ghahramani, Zoubin
    Yang, Yiming
    MACHINE LEARNING, 2008, 73 (03) : 221 - 242
  • [26] Learning latent variable structured prediction models with Gaussian perturbations
    Bello, Kevin
    Honorio, Jean
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [27] Flexible latent variable models for multi-task learning
    Jian Zhang
    Zoubin Ghahramani
    Yiming Yang
    Machine Learning, 2008, 73 : 221 - 242
  • [28] Learning latent variable models from distributed and abstracted data
    Zhang, Xiaofeng
    Cheung, William K.
    Li, C. H.
    INFORMATION SCIENCES, 2011, 181 (14) : 2964 - 2988
  • [29] Unsupervised learning in radiology using novel latent variable models
    Carrivick, L
    Prabhu, S
    Goddard, P
    Rossiter, J
    2005 IEEE COMPUTER SOCIETY CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, VOL 2, PROCEEDINGS, 2005, : 854 - 859
  • [30] Building blocks for variational Bayesian learning of latent variable models
    Raiko, Tapani
    Valpola, Harri
    Harva, Markus
    Karhunen, Juha
    Journal of Machine Learning Research, 2007, 8 : 155 - 201