Mixture Variational Autoencoder of Boltzmann Machines for Text Processing

被引:0
|
作者
Guilherme Gomes, Bruno [1 ]
Murai, Fabricio [1 ]
Goussevskaia, Olga [1 ]
Couto Da Silva, Ana Paula [1 ]
机构
[1] Univ Fed Minas Gerais, Belo Horizonte, MG, Brazil
关键词
D O I
10.1007/978-3-030-80599-9_5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Variational autoencoders (VAEs) have been successfully used to learn good representations in unsupervised settings, especially for image data. More recently, mixture variational autoencoders (MVAEs) have been proposed to enhance the representation capabilities of VAEs by assuming that data can come from a mixture distribution. In this work, we adapt MVAEs for text processing by modeling each component's joint distribution of latent variables and document's bag-of-words as a graphical model known as the Boltzmann Machine, popular in natural language processing for performing well in a number of tasks. The proposed model, MVAE-BM, can learn text representations from unlabeled data without requiring pre-trained word embeddings. We evaluate the representations obtained by MVAE-BM on six corpora w.r.t. the perplexity metric and accuracy on binary and multi-class text classification. Despite its simplicity, our results show that MVAE-BM's performance is on par with or superior to that of modern deep learning techniques such as BERT and RoBERTa. Last, we show that the mapping to mixture components learned by the model lends itself naturally to document clustering.
引用
收藏
页码:46 / 56
页数:11
相关论文
共 50 条
  • [31] Condition-Transforming Variational Autoencoder for Generating Diverse Short Text Conversations
    Ruan, Yu-Ping
    Ling, Zhen-Hua
    Zhu, Xiaodan
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2020, 19 (06)
  • [32] Condition-Transforming Variational Autoencoder for Generating Diverse Short Text Conversations
    Ruan, Yu-Ping
    Ling, Zhen-Hua
    Zhu, Xiaodan
    Ruan, Yu-Ping (ypruan@mail.ustc.edu.cn); Ling, Zhen-Hua (zhling@ustc.edu.cn), 1600, Association for Computing Machinery (19):
  • [33] Complex Text Processing by the Temporal Context Machines
    Weng, Juyang
    Zhang, Qi
    Chi, Mingmin
    Xue, Xiangyang
    2009 IEEE 8TH INTERNATIONAL CONFERENCE ON DEVELOPMENT AND LEARNING, 2009, : 220 - +
  • [34] Text-mining the NeuroSynth corpus using Deep Boltzmann Machines
    Monti, Ricardo
    Lorenz, Romy
    Leech, Robert
    Anagnostopoulos, Christoforos
    Montana, Giovanni
    2016 6TH INTERNATIONAL WORKSHOP ON PATTERN RECOGNITION IN NEUROIMAGING (PRNI), 2016, : 13 - 16
  • [35] Unsupervised Clustering through Gaussian Mixture Variational AutoEncoder with Non-Reparameterized Variational Inference and Std Annealing
    Li, Zhihan
    Zhao, Youjian
    Xu, Haowen
    Chen, Wenxiao
    Xu, Shangqing
    Li, Yilin
    Pei, Dan
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [36] Dirichlet Variational Autoencoder
    Joo, Weonyoung
    Lee, Wonsung
    Park, Sungrae
    Moon, Il-Chul
    PATTERN RECOGNITION, 2020, 107
  • [37] Grammar Variational Autoencoder
    Kusner, Matt J.
    Paige, Brooks
    Hernandez-Lobato, Jose Miguel
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 70, 2017, 70
  • [38] The Autoencoding Variational Autoencoder
    Cemgil, A. Taylan
    Ghaisas, Sumedh
    Dvijotham, Krishnamurthy
    Gowal, Sven
    Kohli, Pushmeet
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [39] Quantum variational autoencoder
    Khoshaman, Amir
    Vinci, Walter
    Denis, Brandon
    Andriyash, Evgeny
    Amin, Mohammad H.
    QUANTUM SCIENCE AND TECHNOLOGY, 2019, 4 (01)
  • [40] Variational Selective Autoencoder
    Gong, Yu
    Hajimirsadeghi, Hossein
    He, Jiawei
    Nawhal, Megha
    Durand, Thibaut
    Mori, Greg
    SYMPOSIUM ON ADVANCES IN APPROXIMATE BAYESIAN INFERENCE, VOL 118, 2019, 118