Deterministic Inference of Topic Models via Maximal Latent State Replication

被引:3
|
作者
Rugeles, Daniel [1 ]
Hai, Zhen [3 ]
Dash, Manoranjan [2 ]
Cong, Gao [1 ]
机构
[1] Nanyang Technol Univ, Sch Comp Sci & Engn, Singapore 639798, Singapore
[2] Natl Univ Singapore, Sch Comp, Singapore 119077, Singapore
[3] ASTAR, Inst Infocomm Res, Singapore 138632, Singapore
关键词
Inference algorithms; Computational modeling; Sampling methods; Probabilistic logic; Mathematical model; Resource management; Convergence; Topic models; gibbs sampling; deterministic inference; distributable inference; latent state replication; PARALLEL;
D O I
10.1109/TKDE.2020.3000559
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Probabilistic topic models, such as latent dirichlet allocation (LDA), are often used to discover hidden semantic structure of a collection of documents. In recent years, various inference algorithms have been developed to cope with learning of topic models, among which Gibbs sampling methods remain a popular choice. In this paper, we aim to improve the inference of topic models based on the Gibbs sampling framework. We extend a state augmentation based Gibbs sampling method by maximizing the replications of latent states, and propose a new generic deterministic inference method, named maximal latent state replication (MAX), for learning of a family of probabilistic topic models. One key benefit of the proposed method lies in the deterministic nature for inference, which may help to improve its running efficiency as well as predictive perplexity. We have conducted extensive experiments on real-life publicly available datasets, and the results have validated that our proposed method MAX significantly outperforms state-of-the-art baselines for inference of existing well-known topic models.
引用
收藏
页码:1684 / 1695
页数:12
相关论文
共 50 条
  • [41] Faster inference from state space models via GPU computing
    Fagard-Jenkin, Calliste
    Thomas, Len
    ECOLOGICAL INFORMATICS, 2024, 80
  • [42] Inference in components of variance models with low replication
    Hall, P
    Yao, QW
    ANNALS OF STATISTICS, 2003, 31 (02): : 414 - 441
  • [43] Deterministic Remote State Preparation via the χ State
    张佩
    李娴
    马松雅
    瞿治国
    Communications in Theoretical Physics, 2017, 67 (05) : 498 - 506
  • [44] Deterministic Remote State Preparation via the χ State
    Zhang, Pei
    Li, Xian
    Ma, Song-Ya
    Qu, Zhi-Guo
    COMMUNICATIONS IN THEORETICAL PHYSICS, 2017, 67 (05) : 498 - 506
  • [45] Neural field models for latent state inference: Application to large-scale neuronal recordings
    Rule, Michael E.
    Schnoerr, David
    Hennig, Matthias H.
    Sanguinetti, Guido
    PLOS COMPUTATIONAL BIOLOGY, 2019, 15 (11)
  • [46] Efficient and Deterministic Scheduling for Parallel State Machine Replication
    Mendizabal, Odorico M.
    De Moura, Ruda S. T.
    Dotti, Fernando Luis
    Pedone, Fernando
    2017 31ST IEEE INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM (IPDPS), 2017, : 748 - 757
  • [47] Topic-aware latent models for representation learning on networks
    Celikkanat, Abdulkadir
    Malliaros, Fragkiskos D.
    PATTERN RECOGNITION LETTERS, 2021, 144 : 89 - 96
  • [48] Progressive EM for Latent Tree Models and Hierarchical Topic Detection
    Chen, Peixian
    Zhang, Nevin L.
    Poon, Leonard K. M.
    Chen, Zhourong
    THIRTIETH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2016, : 1498 - 1504
  • [49] Prior-aware Composition Inference for Spectral Topic Models
    Lee, Moontae
    Bindel, David
    Mimno, David
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 108, 2020, 108 : 4258 - 4267
  • [50] Inference and Learning in Evidential Discrete Latent Markov Models
    Ramasso, Emmanuel
    IEEE TRANSACTIONS ON FUZZY SYSTEMS, 2017, 25 (05) : 1102 - 1114