Mixture Variational Autoencoder of Boltzmann Machines for Text Processing

被引:0
|
作者
Guilherme Gomes, Bruno [1 ]
Murai, Fabricio [1 ]
Goussevskaia, Olga [1 ]
Couto Da Silva, Ana Paula [1 ]
机构
[1] Univ Fed Minas Gerais, Belo Horizonte, MG, Brazil
关键词
D O I
10.1007/978-3-030-80599-9_5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Variational autoencoders (VAEs) have been successfully used to learn good representations in unsupervised settings, especially for image data. More recently, mixture variational autoencoders (MVAEs) have been proposed to enhance the representation capabilities of VAEs by assuming that data can come from a mixture distribution. In this work, we adapt MVAEs for text processing by modeling each component's joint distribution of latent variables and document's bag-of-words as a graphical model known as the Boltzmann Machine, popular in natural language processing for performing well in a number of tasks. The proposed model, MVAE-BM, can learn text representations from unlabeled data without requiring pre-trained word embeddings. We evaluate the representations obtained by MVAE-BM on six corpora w.r.t. the perplexity metric and accuracy on binary and multi-class text classification. Despite its simplicity, our results show that MVAE-BM's performance is on par with or superior to that of modern deep learning techniques such as BERT and RoBERTa. Last, we show that the mapping to mixture components learned by the model lends itself naturally to document clustering.
引用
收藏
页码:46 / 56
页数:11
相关论文
共 50 条
  • [1] Variational quantum Boltzmann machines
    Christa Zoufal
    Aurélien Lucchi
    Stefan Woerner
    Quantum Machine Intelligence, 2021, 3
  • [2] Variational quantum Boltzmann machines
    Zoufal, Christa
    Lucchi, Aurelien
    Woerner, Stefan
    QUANTUM MACHINE INTELLIGENCE, 2021, 3 (01)
  • [3] Semisupervised Text Classification by Variational Autoencoder
    Xu, Weidi
    Tan, Ying
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (01) : 295 - 308
  • [4] Discriminative Mixture Variational Autoencoder for Semisupervised Classification
    Chen, Jian
    Du, Lan
    Liao, Leiyao
    IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (05) : 3032 - 3046
  • [5] Syntax-Infused Variational Autoencoder for Text Generation
    Zhang, Xinyuan
    Yang, Yi
    Yuan, Siyang
    Shen, Dinghan
    Carin, Lawrence
    57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 2069 - 2078
  • [6] Text Generation with Syntax-Enhanced Variational Autoencoder
    Yuan, Weijie
    Ding, Linyi
    Meng, Kui
    Liu, Gongshen
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [7] Semisupervised Classification With Sequence Gaussian Mixture Variational Autoencoder
    Wang, Shuangqing
    Yu, Jianbo
    Li, Zhi
    Chai, Tianyou
    IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2024, 71 (09) : 11540 - 11548
  • [8] A Variational Autoencoder Mixture Model for Online Behavior Recommendation
    Nguyen, Minh-Duc
    Cho, Yoon-Sik
    IEEE ACCESS, 2020, 8 (08) : 132736 - 132747
  • [9] HGMVAE: hierarchical disentanglement in Gaussian mixture variational autoencoder
    Zhou, Jiashuang
    Liu, Yongqi
    Du, Xiaoqin
    VISUAL COMPUTER, 2024, 40 (10): : 7491 - 7502
  • [10] Variational Autoencoder With Optimizing Gaussian Mixture Model Priors
    Guo, Chunsheng
    Zhou, Jialuo
    Chen, Huahua
    Ying, Na
    Zhang, Jianwu
    Zhou, Di
    IEEE ACCESS, 2020, 8 : 43992 - 44005