Latent Dirichlet mixture model

被引:12
|
作者
Chien, Jen-Tzung [1 ]
Lee, Chao-Hsi [1 ]
Tan, Zheng-Hua [2 ]
机构
[1] Natl Chiao Tung Univ, Dept Elect & Comp Engn, Hsinchu, Taiwan
[2] Aalborg Univ, Dept Elect Syst, Aalborg, Denmark
关键词
Bayesian learning; Topic model; Dirichlet mixture model; PRIORS;
D O I
10.1016/j.neucom.2017.08.029
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Text representation based on latent topic model is seen as a non-Gaussian problem where the observed words and latent topics are multinomial variables and the topic proportionals are Dirichlet variables. Traditional topic model is established by introducing a single Dirichlet prior to characterize the topic proportionals. The words in a text document are represented by a random mixture of semantic topics. However, in real world, a single Dirichlet distribution may not faithfully reflect the variations of topic proportionals estimated from the heterogeneous documents. To address these variations, we propose a new latent variable model where latent topics and their proportionals are learned by incorporating the prior based on Dirichlet mixture model. The resulting latent Dirichlet mixture model (LDMM) is constructed for topic clustering as well as document clustering. Multiple Dirichlets provide a solution to build structural latent variables in learning representation over a variety of topics. This study carries out the inference for LDMM according to the variational Bayes and the collapsed variational Bayes. Such an unsupervised LDMM is further extended to a supervised LDMM for text classification. Experiments on document representation, summarization and classification show the merit of structural prior in LDMM topic models. (C) 2017 Elsevier B. V. All rights reserved.
引用
收藏
页码:12 / 22
页数:11
相关论文
共 50 条
  • [1] A comparison of the performance of latent Dirichlet allocation and the Dirichlet multinomial mixture model on short text
    Mazarura, Jocelyn
    de Waal, Alta
    [J]. 2016 PATTERN RECOGNITION ASSOCIATION OF SOUTH AFRICA AND ROBOTICS AND MECHATRONICS INTERNATIONAL CONFERENCE (PRASA-ROBMECH), 2016,
  • [2] AUGMENTED LATENT DIRICHLET ALLOCATION (LDA) TOPIC MODEL WITH GAUSSIAN MIXTURE TOPICS
    Prabhudesai, Kedar S.
    Mainsah, Boyla O.
    Collins, Leslie M.
    Throckmorton, Chandra S.
    [J]. 2018 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2018, : 2451 - 2455
  • [3] Inference for the Number of Topics in the Latent Dirichlet Allocation Model via Bayesian Mixture Modeling
    Chen, Zhe
    Doss, Hani
    [J]. JOURNAL OF COMPUTATIONAL AND GRAPHICAL STATISTICS, 2019, 28 (03) : 567 - 585
  • [4] Supervised latent Dirichlet allocation with a mixture of sparse softmax
    Li, Xiaoxu
    Ma, Zhanyu
    Peng, Pai
    Guo, Xiaowei
    Huang, Feiyue
    Wang, Xiaojie
    Guo, Jun
    [J]. NEUROCOMPUTING, 2018, 312 : 324 - 335
  • [5] Latent Dirichlet allocation mixture models for nucleotide sequence analysis
    Wang, Bixuan
    Mount, Stephen M.
    [J]. NAR GENOMICS AND BIOINFORMATICS, 2024, 6 (03)
  • [6] Sampling the Dirichlet mixture model with slices
    Walker, Stephen G.
    [J]. COMMUNICATIONS IN STATISTICS-SIMULATION AND COMPUTATION, 2007, 36 (01) : 45 - 54
  • [7] Dirichlet Latent Variable Model: A Dynamic Model Based on Dirichlet Prior for Audio Processing
    Kumar, Anurendra
    Guha, Tanaya
    Ghosh, Prasanta Kumar
    [J]. IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2019, 27 (05) : 919 - 931
  • [8] Indexing by Latent Dirichlet Allocation and an Ensemble Model
    Wang, Yanshan
    Lee, Jae-Sung
    Choi, In-Chan
    [J]. JOURNAL OF THE ASSOCIATION FOR INFORMATION SCIENCE AND TECHNOLOGY, 2016, 67 (07) : 1736 - 1750
  • [9] A Latent Dirichlet Model for Unsupervised Entity Resolution
    Bhattacharya, Indrajit
    Getoor, Lise
    [J]. PROCEEDINGS OF THE SIXTH SIAM INTERNATIONAL CONFERENCE ON DATA MINING, 2006, : 47 - 58
  • [10] Research on dirichlet process mixture model for clustering
    Zhang, Biyao
    Zhang, Kaisong
    Zhong, Luo
    Zhang, Xuanya
    [J]. Ingenierie des Systemes d'Information, 2019, 24 (02): : 183 - 189