Benchmarking Neural Topic Models: An Empirical Study

被引:0
|
作者
Thanh-Nam Doan [1 ]
Tuan-Anh Hoang [2 ]
机构
[1] Univ Tennessee, Chattanooga, TN USA
[2] VNU Univ Sci, 334 Nguyen Trai, Hanoi, Vietnam
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Neural topic modeling approach has been attracting much attention recently as it is able to leverage the advantages of both neural networks and probabilistic topic models. Previous works have proposed several models that are based on this framework and obtained impressive experimental results compared to traditional probabilistic models. However, the reported result is not consistent across the works, making them hard for gaining a rigorous assessment of these approaches. This work aims to address this issue by offering an extensive empirical evaluation of typical neural topic models in different aspects using large, diverse datasets as well as a thorough set of metrics. Precisely, we examine the performance of these models in three tasks, namely uncovering cohesive topics, modeling the input documents, and representing them for downstream classification. Our results show that while the neural topic models are better in the first and the third tasks, the traditional probabilistic models are still a strong baseline and are better in the second task in many cases. These findings give us more insights for choosing off-the-shelf topic modeling toolboxes in different contexts, as well as for designing more comprehensive evaluation for neural topic models.
引用
收藏
页码:4363 / 4368
页数:6
相关论文
共 50 条
  • [1] An Empirical Study on Crosslingual Transfer in Probabilistic Topic Models
    Hao, Shudong
    Paul, Michael J.
    [J]. COMPUTATIONAL LINGUISTICS, 2020, 46 (01) : 95 - 134
  • [2] Empirical study on variational inference methods for topic models
    Chi, Jinjin
    Ouyang, Jihong
    Li, Ximing
    Li, Changchun
    [J]. JOURNAL OF EXPERIMENTAL & THEORETICAL ARTIFICIAL INTELLIGENCE, 2018, 30 (01) : 129 - 142
  • [3] Neural Topic Models for Hierarchical Topic Detection and Visualization
    Pham, Dang
    Le, Than M., V
    [J]. MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2021: RESEARCH TRACK, PT III, 2021, 12977 : 35 - 51
  • [4] Coherence Regularization for Neural Topic Models
    Krasnashchok, Katsiaryna
    Cherif, Aymen
    [J]. ADVANCES IN NEURAL NETWORKS - ISNN 2019, PT I, 2019, 11554 : 426 - 433
  • [5] Structured Neural Topic Models for Reviews
    Esmaeili, Babak
    Huang, Hongyi
    Wallace, Byron C.
    van de Meent, Jan-Willem
    [J]. 22ND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 89, 2019, 89
  • [6] Investigating task performance of probabilistic topic models: an empirical study of PLSA and LDA
    Yue Lu
    Qiaozhu Mei
    ChengXiang Zhai
    [J]. Information Retrieval, 2011, 14 : 178 - 203
  • [7] Investigating task performance of probabilistic topic models: an empirical study of PLSA and LDA
    Lu, Yue
    Mei, Qiaozhu
    Zhai, ChengXiang
    [J]. INFORMATION RETRIEVAL, 2011, 14 (02): : 178 - 203
  • [8] Benchmarking test of empirical root water uptake models
    dos Santos, Marcos Alex
    van Lier, Quirijn de Jong
    van Dam, Jos C.
    Freire Bezerra, Andre Herman
    [J]. HYDROLOGY AND EARTH SYSTEM SCIENCES, 2017, 21 (01) : 473 - 493
  • [9] Empirical Study of Tweets Topic Classification Using Transformer-Based Language Models
    Mandal, Ranju
    Chen, Jinyan
    Becken, Susanne
    Stantic, Bela
    [J]. INTELLIGENT INFORMATION AND DATABASE SYSTEMS, ACIIDS 2021, 2021, 12672 : 340 - 350
  • [10] An Empirical study to understand the Compositional Prowess of Neural Dialog Models
    Kumar, Vinayshekhar Bannihatti
    Kumar, Vaibhav
    Bhutani, Mukul
    Rudnicky, Alexander
    [J]. PROCEEDINGS OF THE THIRD WORKSHOP ON INSIGHTS FROM NEGATIVE RESULTS IN NLP (INSIGHTS 2022), 2022, : 154 - 158