Deterministic Inference of Topic Models via Maximal Latent State Replication

被引:3
|
作者
Rugeles, Daniel [1 ]
Hai, Zhen [3 ]
Dash, Manoranjan [2 ]
Cong, Gao [1 ]
机构
[1] Nanyang Technol Univ, Sch Comp Sci & Engn, Singapore 639798, Singapore
[2] Natl Univ Singapore, Sch Comp, Singapore 119077, Singapore
[3] ASTAR, Inst Infocomm Res, Singapore 138632, Singapore
关键词
Inference algorithms; Computational modeling; Sampling methods; Probabilistic logic; Mathematical model; Resource management; Convergence; Topic models; gibbs sampling; deterministic inference; distributable inference; latent state replication; PARALLEL;
D O I
10.1109/TKDE.2020.3000559
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Probabilistic topic models, such as latent dirichlet allocation (LDA), are often used to discover hidden semantic structure of a collection of documents. In recent years, various inference algorithms have been developed to cope with learning of topic models, among which Gibbs sampling methods remain a popular choice. In this paper, we aim to improve the inference of topic models based on the Gibbs sampling framework. We extend a state augmentation based Gibbs sampling method by maximizing the replications of latent states, and propose a new generic deterministic inference method, named maximal latent state replication (MAX), for learning of a family of probabilistic topic models. One key benefit of the proposed method lies in the deterministic nature for inference, which may help to improve its running efficiency as well as predictive perplexity. We have conducted extensive experiments on real-life publicly available datasets, and the results have validated that our proposed method MAX significantly outperforms state-of-the-art baselines for inference of existing well-known topic models.
引用
收藏
页码:1684 / 1695
页数:12
相关论文
共 50 条
  • [21] Latent Topic Models of Surface Syntactic Information
    Basili, Roberto
    Giannone, C.
    Croce, Danilo
    Domeniconi, C.
    AI(STAR)IA 2011: ARTIFICIAL INTELLIGENCE AROUND MAN AND BEYOND, 2011, 6934 : 225 - +
  • [22] Latent Topic Networks: A Versatile Probabilistic Programming Framework for Topic Models
    Foulds, James
    Kumar, Shachi H.
    Getoor, Lise
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 37, 2015, 37 : 777 - 786
  • [23] Estimation and inference in parametric deterministic frontier models
    Amsler, Christine
    Leonard, Michael
    Schmidt, Peter
    JOURNAL OF PRODUCTIVITY ANALYSIS, 2013, 40 (03) : 293 - 305
  • [24] Bayesian Inference of Deterministic Population Growth Models
    Carvalho, Luiz Max
    Struchiner, Claudio J.
    Bastos, Leonardo S.
    INTERDISCIPLINARY BAYESIAN STATISTICS, EBEB 2014, 2015, 118 : 217 - 228
  • [25] Estimation and inference in parametric deterministic frontier models
    Christine Amsler
    Michael Leonard
    Peter Schmidt
    Journal of Productivity Analysis, 2013, 40 : 293 - 305
  • [26] Generic inference in latent Gaussian process models
    Bonilla, Edwin V.
    Krauth, Karl
    Dezfouli, Amir
    Journal of Machine Learning Research, 2019, 20
  • [27] Generic Inference in Latent Gaussian Process Models
    Bonilla, Edwin V.
    Krauth, Karl
    Dezfouli, Amir
    JOURNAL OF MACHINE LEARNING RESEARCH, 2019, 20
  • [28] Spectral Latent Variable Models for perceptual inference
    Kanaujia, Atul
    Sminchisescu, Cristian
    Metaxas, Dimitris
    2007 IEEE 11TH INTERNATIONAL CONFERENCE ON COMPUTER VISION, VOLS 1-6, 2007, : 142 - +
  • [29] Inference on latent factor models for informative censoring
    Ungolo, Francesco
    van den Heuvel, Edwin R.
    STATISTICAL METHODS IN MEDICAL RESEARCH, 2022, 31 (05) : 801 - 820
  • [30] Paired Feature Constraints for Latent Dirichlet Topic Models
    Sristy, Nagesh Bhattu
    Somayajulu, D. V. L. N.
    Subramanyam, R. B. V.
    2013 INTERNATIONAL CONFERENCE OF SOFT COMPUTING AND PATTERN RECOGNITION (SOCPAR), 2013, : 270 - 275