Deterministic Inference of Topic Models via Maximal Latent State Replication

被引:3
|
作者
Rugeles, Daniel [1 ]
Hai, Zhen [3 ]
Dash, Manoranjan [2 ]
Cong, Gao [1 ]
机构
[1] Nanyang Technol Univ, Sch Comp Sci & Engn, Singapore 639798, Singapore
[2] Natl Univ Singapore, Sch Comp, Singapore 119077, Singapore
[3] ASTAR, Inst Infocomm Res, Singapore 138632, Singapore
关键词
Inference algorithms; Computational modeling; Sampling methods; Probabilistic logic; Mathematical model; Resource management; Convergence; Topic models; gibbs sampling; deterministic inference; distributable inference; latent state replication; PARALLEL;
D O I
10.1109/TKDE.2020.3000559
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Probabilistic topic models, such as latent dirichlet allocation (LDA), are often used to discover hidden semantic structure of a collection of documents. In recent years, various inference algorithms have been developed to cope with learning of topic models, among which Gibbs sampling methods remain a popular choice. In this paper, we aim to improve the inference of topic models based on the Gibbs sampling framework. We extend a state augmentation based Gibbs sampling method by maximizing the replications of latent states, and propose a new generic deterministic inference method, named maximal latent state replication (MAX), for learning of a family of probabilistic topic models. One key benefit of the proposed method lies in the deterministic nature for inference, which may help to improve its running efficiency as well as predictive perplexity. We have conducted extensive experiments on real-life publicly available datasets, and the results have validated that our proposed method MAX significantly outperforms state-of-the-art baselines for inference of existing well-known topic models.
引用
收藏
页码:1684 / 1695
页数:12
相关论文
共 50 条
  • [1] Scalable inference of topic evolution via models for latent geometric structures
    Yurochkin, Mikhail
    Fan, Zhiwei
    Guha, Aritra
    Koutris, Paraschos
    Nguyen, XuanLong
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [2] Deterministic approximate inference techniques for conditionally Gaussian state space models
    Zoeter, Onno
    Heskes, Tom
    STATISTICS AND COMPUTING, 2006, 16 (03) : 279 - 292
  • [3] Deterministic approximate inference techniques for conditionally Gaussian state space models
    Onno Zoeter
    Tom Heskes
    Statistics and Computing, 2006, 16 : 279 - 292
  • [4] A Latent Concept Topic Model for Robust Topic Inference Using Word Embeddings
    Hu, Weihua
    Tsujii, Jun'ichi
    PROCEEDINGS OF THE 54TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2016), VOL 2, 2016, : 380 - 386
  • [5] Efficient State-Space Inference of Periodic Latent Force Models
    Reece, Steven
    Ghosh, Siddhartha
    Rogers, Alex
    Roberts, Stephen
    Jennings, Nicholas R.
    JOURNAL OF MACHINE LEARNING RESEARCH, 2014, 15 : 2337 - 2397
  • [6] Efficient state-space inference of periodic latent force models
    Reece, Steven
    Ghosh, Siddhartha
    Rogers, Alex
    Roberts, Stephen
    Jennings, Nicholas R.
    Journal of Machine Learning Research, 2014, 15 : 2337 - 2397
  • [7] Inference and Decoding of Motor Cortex Low-Dimensional Dynamics via Latent State-Space Models
    Aghagolzadeh, Mehdi
    Truccolo, Wilson
    IEEE TRANSACTIONS ON NEURAL SYSTEMS AND REHABILITATION ENGINEERING, 2016, 24 (02) : 272 - 282
  • [8] Latent Gaussian models for topic modeling
    20160501866993
    (1) Department of Electrical and Computer Engineering, Duke University, Durham; NC; 27708, United States, 1600, Amazon; Digile; Facebook; Google; Springer (Microtome Publishing):
  • [9] Latent Gaussian Models for Topic Modeling
    Hu, Changwei
    Ryu, Eunsu
    Carlson, David
    Wang, Yingjian
    Carin, Lawrence
    ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 33, 2014, 33 : 393 - 401
  • [10] Provable Algorithms for Inference in Topic Models
    Arora, Sanjeev
    Ge, Rong
    Koehler, Frederic
    Ma, Tengyu
    Moitra, Ankur
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 48, 2016, 48