GraphBTM: Graph Enhanced Autoencoded Variational Inference for Biterm Topic Model

被引:0
|
作者
Zhu, Qile [1 ]
Feng, Zheng [1 ]
Li, Xiaolin [1 ]
机构
[1] Univ Florida, NSF Ctr Big Learning, Large Scale Intelligent Syst Lab, Gainesville, FL 32611 USA
基金
美国国家卫生研究院; 美国国家科学基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Discovering the latent topics within texts has been a fundamental task for many applications. However, conventional topic models suffer different problems in different settings. The Latent Dirichlet Allocation (LDA) may not work well for short texts due to the data sparsity (i.e., the sparse word co-occurrence patterns in short documents). The Biterm Topic Model (BTM) learns topics by modeling the word-pairs named biterms in the whole corpus. This assumption is very strong when documents are long with rich topic information and do not exhibit the transitivity of biterms. In this paper, we propose a novel way called GraphBTM to represent biterms as graphs and design Graph Convolutional Networks (GCNs) with residual connections to extract transitive features from biterms. To overcome the data sparsity of LDA and the strong assumption of BTM, we sample a fixed number of documents to form a mini-corpus as a training instance. We also propose a dataset called All News extracted from (Thompson, 2017), in which documents are much longer than 20 Newsgroups. We present an amortized variational inference method for GraphBTM. Our method generates more coherent topics compared with previous approaches. Experiments show that the sampling strategy improves performance by a large margin.
引用
收藏
页码:4663 / 4672
页数:10
相关论文
共 50 条
  • [31] An Intelligent Web Service Discovery Framework Based on Improved Biterm Topic Model
    Yuan, Yuan
    Du, Yegang
    Pan, Jun
    IEEE Access, 2024, 12 : 144437 - 144455
  • [32] On some provably correct cases of variational inference for topic models
    Awasthi, Pranjal
    Risteski, Andrej
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 28 (NIPS 2015), 2015, 28
  • [33] A Study on Stochastic Variational Inference for Topic Modeling with Word Embeddings
    Ozaki, Kana
    Kobayashie, Ichiro
    COMPUTACION Y SISTEMAS, 2022, 26 (03): : 1225 - 1232
  • [34] Public perception of cultural ecosystem services in historic districts based on biterm topic model
    Pan, Ying
    Nik Hashim, Nik Hazwani
    Goh, Hong Ching
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [35] Topic Model for Graph Mining
    Xuan, Junyu
    Lu, Jie
    Zhang, Guangquan
    Luo, Xiangfeng
    IEEE TRANSACTIONS ON CYBERNETICS, 2015, 45 (12) : 2792 - 2803
  • [36] User graph topic model
    Akhtar, Nadeem
    Beg, M. M. Sufyan
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2019, 36 (03) : 2229 - 2240
  • [37] The graph embedded topic model
    Liang, Dingge
    Corneli, Marco
    Bouveyron, Charles
    Latouche, Pierre
    NEUROCOMPUTING, 2023, 562
  • [38] Graph Contrastive Topic Model
    Luo, Zheheng
    Liu, Lei
    Ananiadou, Sophia
    Xie, Qianqian
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 255
  • [39] Bayesian Variational Inference for Exponential Random Graph Models
    Tan, Linda S. L.
    Friel, Nial
    JOURNAL OF COMPUTATIONAL AND GRAPHICAL STATISTICS, 2020, 29 (04) : 910 - 928
  • [40] Enhanced Multiple Model GPB2 Filtering Using Variational Inference
    Li, Xi
    Liu, Yi
    Mihaylova, Lyudmila
    Yang, Le
    Weddell, Steve
    Guo, Fucheng
    2019 22ND INTERNATIONAL CONFERENCE ON INFORMATION FUSION (FUSION 2019), 2019,