Mixture Variational Autoencoder of Boltzmann Machines for Text Processing

被引:0
|
作者
Guilherme Gomes, Bruno [1 ]
Murai, Fabricio [1 ]
Goussevskaia, Olga [1 ]
Couto Da Silva, Ana Paula [1 ]
机构
[1] Univ Fed Minas Gerais, Belo Horizonte, MG, Brazil
关键词
D O I
10.1007/978-3-030-80599-9_5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Variational autoencoders (VAEs) have been successfully used to learn good representations in unsupervised settings, especially for image data. More recently, mixture variational autoencoders (MVAEs) have been proposed to enhance the representation capabilities of VAEs by assuming that data can come from a mixture distribution. In this work, we adapt MVAEs for text processing by modeling each component's joint distribution of latent variables and document's bag-of-words as a graphical model known as the Boltzmann Machine, popular in natural language processing for performing well in a number of tasks. The proposed model, MVAE-BM, can learn text representations from unlabeled data without requiring pre-trained word embeddings. We evaluate the representations obtained by MVAE-BM on six corpora w.r.t. the perplexity metric and accuracy on binary and multi-class text classification. Despite its simplicity, our results show that MVAE-BM's performance is on par with or superior to that of modern deep learning techniques such as BERT and RoBERTa. Last, we show that the mapping to mixture components learned by the model lends itself naturally to document clustering.
引用
收藏
页码:46 / 56
页数:11
相关论文
共 50 条
  • [41] Longitudinal Variational Autoencoder
    Ramchandran, Siddharth
    Tikhonov, Gleb
    Kujanpaa, Kalle
    Koskinen, Miika
    Lahdesmaki, Harri
    24TH INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS (AISTATS), 2021, 130
  • [42] The Variational InfoMax AutoEncoder
    Crescimanna, Vincenzo
    Graham, Bruce
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [43] Generative autoencoder to prevent overregularization of variational autoencoder
    Ko, YoungMin
    Ko, SunWoo
    Kim, YoungSoo
    ETRI JOURNAL, 2024,
  • [44] Spoken-Text-Style Transfer with Conditional Variational Autoencoder and Content Word Storage
    Yoshioka, Daiki
    Yasuda, Yusuke
    Matsunaga, Noriyuki
    Ohtani, Yamato
    Toda, Tomoki
    INTERSPEECH 2022, 2022, : 4576 - 4580
  • [45] A Novel Model for Ship Trajectory Anomaly Detection Based on Gaussian Mixture Variational Autoencoder
    Xie, Lei
    Guo, Tao
    Chang, Jiliang
    Wan, Chengpeng
    Hu, Xinyuan
    Yang, Yang
    Ou, Changkui
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2023, 72 (11) : 13826 - 13835
  • [46] Mixture autoregressive and spectral attention network for multispectral image compression based on variational autoencoder
    Kong, Fanqiang
    Ren, Guanglong
    Hu, Yunfang
    Li, Dan
    Hu, Kedi
    VISUAL COMPUTER, 2024, 40 (09): : 6295 - 6318
  • [47] Enhancing Graph Variational Autoencoder for Short Text Topic Modeling with Mutual Information Maximization
    Ge, Yuhang
    Hu, Xuegang
    2022 IEEE INTERNATIONAL CONFERENCE ON KNOWLEDGE GRAPH (ICKG), 2022, : 64 - 70
  • [48] A GAUSSIAN MIXTURE VARIATIONAL AUTOENCODER-BASED APPROACH FOR DESIGNING PHONONIC BANDGAP METAMATERIALS
    Wang, Zihan
    Xian, Weikang
    Baccouche, M. Ridha
    Lanzerath, Horst
    Li, Ying
    Xu, Hongyi
    PROCEEDINGS OF ASME 2021 INTERNATIONAL DESIGN ENGINEERING TECHNICAL CONFERENCES AND COMPUTERS AND INFORMATION IN ENGINEERING CONFERENCE, IDETC-CIE2021, VOL 3B, 2021,
  • [49] Video anomaly detection and localization via Gaussian Mixture Fully Convolutional Variational Autoencoder
    Fan, Yaxiang
    Wen, Gongjian
    Li, Deren
    Qiu, Shaohua
    Levine, Martin D.
    Xiao, Fei
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2020, 195
  • [50] Case2vec: joint variational autoencoder for case text embedding representation
    Ran Song
    Shengxiang Gao
    Zhengtao Yu
    Yafei Zhang
    Gaofeng Zhou
    International Journal of Machine Learning and Cybernetics, 2021, 12 : 2517 - 2528