Supervised N-gram Topic Model

被引:12
|
作者
Kawamae, Noriaki [1 ]
机构
[1] NTT Comware, Mihama Ku, 1-6 Nakase, Chiba 2610023, Japan
关键词
Nonparametric Bayes models; Nonparametric Dirichlet process; Topic models; Latent variable models; Graphical models; Sentiment analysis; N-gram topic model;
D O I
10.1145/2556195.2559895
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We propose a Bayesian nonparametric topic model that represents relationships between given labels and the corresponding words/phrases, as found in supervised articles. Unlike existing supervised topic models, our proposal, supervised N-gram topic model (SNT), focuses on both the number of topics and power-law distribution in the word frequencies for topic-specific N-grams. To achieve this goal, SNT takes a Bayesian nonparametric approach to topic sampling; it assigns a topic to each token using Chinese restaurant process (CRP), and generates a word distribution jointly with the given variable in textual order, and then forms each N-gram word as a hierarchy of Pitman-Yor process (PYP) priors. CRP can help SNT to automatically estimate the appropriate number of topics, which impacts the quality of topic specific words, N-grams, and observed value distribution. Since PYP recovers the exact formulation of interpolated Kneser-Ney, one of the best smoothing approaches for N-gram language models, it can allow SNT to generate more interpretable N-grams that the alternatives. Experiments on labeled text data show that SNT is useful as a generative model for discovering more phrases that better complement human experts than existing alternatives and provide more domain specific knowledge. The results show that SNT can be applied to various tasks such as automatic annotation.
引用
收藏
页码:473 / 482
页数:10
相关论文
共 50 条
  • [21] A language independent n-gram model for word segmentation
    Kang, Seung-Shik
    Hwang, Kyu-Baek
    [J]. AI 2006: ADVANCES IN ARTIFICIAL INTELLIGENCE, PROCEEDINGS, 2006, 4304 : 557 - +
  • [22] Analysis of N-gram model on Telugu Document Classification
    Rani, B. Padmaja
    Vardhan, B. Vishnu
    Durga, A. Kanaka
    Reddy, L. Pratap
    Babu, A. Vinaya
    [J]. 2008 IEEE CONGRESS ON EVOLUTIONARY COMPUTATION, VOLS 1-8, 2008, : 3199 - +
  • [23] Extended N-gram Model for Analysis of Polish Texts
    Banasiak, Dariusz
    Mierzwa, Jaroslaw
    Sterna, Antoni
    [J]. MAN-MACHINE INTERACTIONS 5, ICMMI 2017, 2018, 659 : 355 - 364
  • [24] N-gram MalGAN: Evading machine learning detection via feature n-gram
    Zhu, Enmin
    Zhang, Jianjie
    Yan, Jijie
    Chen, Kongyang
    Gao, Chongzhi
    [J]. DIGITAL COMMUNICATIONS AND NETWORKS, 2022, 8 (04) : 485 - 491
  • [25] N-gram MalGAN:Evading machine learning detection via feature n-gram
    Enmin Zhu
    Jianjie Zhang
    Jijie Yan
    Kongyang Chen
    Chongzhi Gao
    [J]. Digital Communications and Networks., 2022, 8 (04) - 491
  • [26] Topic-dependent N-gram models based on Optimization of Context Lengths in LDA
    Nakamura, Akira
    Hayamizu, Satoru
    [J]. 11TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2010 (INTERSPEECH 2010), VOLS 3 AND 4, 2010, : 3066 - 3069
  • [27] N-gram模型综述
    尹陈
    吴敏
    [J]. 计算机系统应用, 2018, 27 (10) : 33 - 38
  • [28] N-gram over Context
    Kawamae, Noriaki
    [J]. PROCEEDINGS OF THE 25TH INTERNATIONAL CONFERENCE ON WORLD WIDE WEB (WWW'16), 2016, : 1045 - 1055
  • [29] N-gram similarity and distance
    Kondrak, Grzegorz
    [J]. STRING PROCESSING AND INFORMATION RETRIEVAL, PROCEEDINGS, 2005, 3772 : 115 - 126
  • [30] BIGRAM VS N-GRAM
    HALPIN, P
    [J]. BYTE, 1988, 13 (08): : 26 - 26