Tensor Decompositions for Learning Latent Variable Models

被引:0
|
作者
Anandkumar, Animashree [1 ]
Ge, Rong [2 ]
Hsu, Daniel [3 ]
Kakade, Sham M. [2 ]
Telgarsky, Matus [4 ]
机构
[1] Univ Calif Irvine, Irvine, CA 92697 USA
[2] Microsoft Res, Cambridge, MA 02142 USA
[3] Columbia Univ, Dept Comp Sci, New York, NY 10027 USA
[4] Rutgers State Univ, Dept Stat, Piscataway, NJ 08854 USA
关键词
latent variable models; tensor decompositions; mixture models; topic models; method of moments; power method; INDEPENDENT COMPONENT ANALYSIS; FIXED-POINT ALGORITHMS; MAXIMUM-LIKELIHOOD; MIXTURES; EM; IDENTIFIABILITY; APPROXIMATION; EIGENVALUES; RANK;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This work considers a computationally and statistically efficient parameter estimation method for a wide class of latent variable models-including Gaussian mixture models, hidden Markov models, and latent Dirichlet allocation-which exploits a certain tensor structure in their low-order observable moments (typically, of second- and third-order). Specifically, parameter estimation is reduced to the problem of extracting a certain (orthogonal) decomposition of a symmetric tensor derived from the moments; this decomposition can be viewed as a natural generalization of the singular value decomposition for matrices. Although tensor decompositions are generally intractable to compute, the decomposition of these specially structured tensors can be efficiently obtained by a variety of approaches, including power iterations and maximization approaches (similar to the case of matrices). A detailed analysis of a robust tensor power method is provided, establishing an analogue of Wedin's perturbation theorem for the singular vectors of matrices. This implies a robust and computationally tractable estimation approach for several popular latent variable models.
引用
收藏
页码:2773 / 2832
页数:60
相关论文
共 50 条
  • [1] Tensor decompositions for learning latent variable models
    Electrical Engineering and Computer Science, University of California, Irvine, 2200 Engineering Hall, Irvine
    CA
    92697, United States
    不详
    MA
    02142, United States
    不详
    NY
    10027, United States
    不详
    NJ
    08854, United States
    J. Mach. Learn. Res., (2773-2832):
  • [2] Tensor Decompositions for Learning Latent Variable Models (A Survey for ALT)
    Anandkumar, Anima
    Ge, Rong
    Hsu, Daniel
    Kakade, Sham M.
    Telgarsky, Matus
    ALGORITHMIC LEARNING THEORY, ALT 2015, 2015, 9355 : 19 - 38
  • [3] Online Tensor Methods for Learning Latent Variable Models
    Huang, Furong
    Niranjan, U. N.
    Hakeem, Mohammad Umar
    Anandkumar, Animashree
    JOURNAL OF MACHINE LEARNING RESEARCH, 2015, 16 : 2797 - 2835
  • [4] Learning Binary Latent Variable Models: A Tensor Eigenpair Approach
    Jaffe, Ariel
    Weiss, Roi
    Carmi, Shai
    Kluger, Yuval
    Nadler, Boaz
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80
  • [5] Learning Latent Variable Models with Regularization
    Peng, Jing
    INTERNATIONAL CONFERENCE ON ELECTRICAL, COMPUTER AND ENERGY TECHNOLOGIES (ICECET 2021), 2021, : 2099 - 2104
  • [6] Learning Diagonal Gaussian Mixture Models and Incomplete Tensor Decompositions
    Guo, Bingni
    Nie, Jiawang
    Yang, Zi
    VIETNAM JOURNAL OF MATHEMATICS, 2022, 50 (02) : 421 - 446
  • [7] Learning Diagonal Gaussian Mixture Models and Incomplete Tensor Decompositions
    Bingni Guo
    Jiawang Nie
    Zi Yang
    Vietnam Journal of Mathematics, 2022, 50 : 421 - 446
  • [8] Learning Latent Variable Gaussian Graphical Models
    Meng, Zhaoshi
    Eriksson, Brian
    Hero, Alfred O., III
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 32 (CYCLE 2), 2014, 32 : 1269 - 1277
  • [9] Learning Latent Variable Models with Discriminant Regularization
    Peng, Jing
    Aved, Alex J.
    AGENTS AND ARTIFICIAL INTELLIGENCE, ICAART 2020, 2021, 12613 : 378 - 398
  • [10] Active Learning for Discrete Latent Variable Models
    Jha, Aditi
    Ashwood, Zoe C.
    Pillow, Jonathan W.
    NEURAL COMPUTATION, 2024, 36 (03) : 437 - 474