Latent Support Measure Machines for Bag-of-Words Data Classification

被引:0
|
作者
Yoshikawa, Yuya [1 ]
Iwata, Tomoharu [2 ]
Sawada, Hiroshi [3 ]
机构
[1] Nara Inst Sci & Technol, Nara 6300192, Japan
[2] NTT Commun Sci Labs, Kyoto 6190237, Japan
[3] NTT Serv Evolut Labs, Yokosuka, Kanagawa 2390847, Japan
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In many classification problems, the input is represented as a set of features, e.g., the bag-of-words (BoW) representation of documents. Support vector machines (SVMs) are widely used tools for such classification problems. The performance of the SVMs is generally determined by whether kernel values between data points can be de fined properly. However, SVMs for BoW representations have a major weakness in that the co-occurrence of different but semantically similar words cannot be reflected in the kernel calculation. To overcome the weakness, we propose a kernel-based discriminative classifier for BoW data, which we call the latent support measure machine (latent SMM). With the latent SMM, a latent vector is associated with each vocabulary term, and each document is represented as a distribution of the latent vectors for words appearing in the document. To represent the distributions efficiently, we use the kernel embeddings of distributions that hold high order moment information about distributions. Then the latent SMM finds a separating hyperplane that maximizes the margins between distributions of different classes while estimating latent vectors for words to improve the classification performance. In the experiments, we show that the latent SMM achieves state-of-the-art accuracy for BoW text classification, is robust with respect to its own hyper-parameters, and is useful to visualize words.
引用
收藏
页数:9
相关论文
共 50 条
  • [1] Release 'Bag-of-Words' Assumption of Latent Dirichlet Allocation
    Xuan, Junyu
    Lu, Jie
    Zhang, Guangquan
    Luo, Xiangfeng
    [J]. FOUNDATIONS OF INTELLIGENT SYSTEMS (ISKE 2013), 2014, 277 : 83 - 92
  • [2] Graph-based bag-of-words for classification
    Silva, Fernanda B.
    Werneck, Rafael de O.
    Goldenstein, Siome
    Tabbone, Salvatore
    Torres, Ricardo da S.
    [J]. PATTERN RECOGNITION, 2018, 74 : 266 - 285
  • [3] Persistence Bag-of-Words for Topological Data Analysis
    Zielinski, Bartosz
    Lipinski, Michal
    Juda, Mateusz
    Zeppelzauer, Matthias
    Dlotko, Pawel
    [J]. PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 4489 - 4495
  • [4] Sequential Bag-of-Words model for human action classification
    Liu, Hong
    Tang, Hao
    Xiao, Wei
    Guo, ZiYi
    Tian, Lu
    Gao, Yuan
    [J]. CAAI TRANSACTIONS ON INTELLIGENCE TECHNOLOGY, 2016, 1 (02) : 125 - 136
  • [5] Internet Traffic Classification based on bag-of-words model
    Zhang, Yin
    Zhou, Yi
    Chen, Kai
    [J]. 2012 IEEE GLOBECOM WORKSHOPS (GC WKSHPS), 2012, : 736 - 741
  • [6] Time series classification with temporal bag-of-words model
    Gui, Zi-Wen
    Yeh, Yi-Ren
    [J]. Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 2014, 8916 : 145 - 153
  • [7] Bag-of-words representation for biomedical time series classification
    Wang, Jin
    Liu, Ping
    She, Mary F. H.
    Nahavandi, Saeid
    Kouzani, Abbas
    [J]. BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2013, 8 (06) : 634 - 644
  • [8] Importance of feature locations in bag-of-words image classification
    Lazic, Nevena
    Aarabi, Parham
    [J]. 2007 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOL I, PTS 1-3, PROCEEDINGS, 2007, : 641 - 644
  • [9] Image classification by visual bag-of-words refinement and reduction
    Lu, Zhiwu
    Wang, Liwei
    Wen, Ji-Rong
    [J]. NEUROCOMPUTING, 2016, 173 : 373 - 384
  • [10] Cross-Domain Matching for Bag-of-Words Data via Kernel Embeddings of Latent Distributions
    Yoshikawa, Yuya
    Iwata, Tomoharu
    Sawada, Hiroshi
    Yamada, Takeshi
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 28 (NIPS 2015), 2015, 28