A Transformer-Based Variational Autoencoder for Sentence Generation

被引:24
|
作者
Liu, Danyang [1 ]
Liu, Gongshen [1 ]
机构
[1] Shanghai Jiao Tong Univ, Sch Elect Informat & Elect Engn, Shanghai, Peoples R China
基金
中国国家自然科学基金;
关键词
variational autoencoder; text generation; self-attention; transformer;
D O I
10.1109/ijcnn.2019.8852155
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The variational autoencoder(VAE) has been proved to be a most efficient generative model, but its applications in natural language tasks have not been fully developed. A novel variational autoencoder for natural texts generation is presented in this paper. Compared to the previously introduced variational autoencoder for natural text where both the encoder and decoder are RNN-based, we propose a new transformer-based architecture and augment the decoder with an LSTM language model layer to fully exploit information of latent variables. We also propose some methods to deal with problems during training time, such as KL divergency collapsing and model degradation. In the experiment, we use random sampling and linear interpolation to test our model. Results show that the generated sentences by our approach are more meaningful and the semantics are more coherent in the latent space.
引用
收藏
页数:7
相关论文
共 50 条
  • [31] VTAE: Variational Transformer Autoencoder With Manifolds Learning
    Shamsolmoali, Pourya
    Zareapoor, Masoumeh
    Zhou, Huiyu
    Tao, Dacheng
    Li, Xuelong
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2023, 32 : 4486 - 4500
  • [32] LayoutDM: Transformer-based Diffusion Model for Layout Generation
    Chai, Shang
    Zhuang, Liansheng
    Yan, Fengying
    [J]. 2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 18349 - 18358
  • [33] Arabic Paraphrase Generation Using Transformer-Based Approaches
    Al-Shameri, Noora Aref
    Al-Khalifa, Hend S.
    [J]. IEEE ACCESS, 2024, 12 : 121896 - 121914
  • [34] Variational transformer-based anomaly detection approach for multivariate time series
    Wang, Xixuan
    Pi, Dechang
    Zhang, Xiangyan
    Liu, Hao
    Guo, Chang
    [J]. MEASUREMENT, 2022, 191
  • [35] Multidimensional degradation data generation method based on variational autoencoder
    Lin, Yanhui
    Li, Chunbo
    [J]. Beijing Hangkong Hangtian Daxue Xuebao/Journal of Beijing University of Aeronautics and Astronautics, 2023, 49 (10): : 2617 - 2627
  • [36] Stochastic Wasserstein Autoencoder for Probabilistic Sentence Generation
    Bahuleyan, Hareesh
    Mou, Lili
    Zhou, Hao
    Vechtomova, Olga
    [J]. 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, 2019, : 4068 - 4076
  • [37] Variational Autoencoder Based Synthetic Data Generation for Imbalanced Learning
    Wan, Zhiqiang
    Zhang, Yazhou
    He, Haibo
    [J]. 2017 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (SSCI), 2017, : 1500 - 1506
  • [38] Urdu Text Reuse Detection at Phrasal level using Sentence Transformer-based approach
    Mehak, Gull
    Muneer, Iqra
    Nawab, Rao Muhammad Adeel
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2023, 234
  • [39] The interactive reading task: Transformer-based automatic item generation
    Attali, Yigal
    Runge, Andrew
    LaFlair, Geoffrey T.
    Yancey, Kevin
    Goodwin, Sarah
    Park, Yena
    von Davier, Alina A.
    [J]. FRONTIERS IN ARTIFICIAL INTELLIGENCE, 2022, 5
  • [40] Sparse Transformer-Based Sequence Generation for Visual Object Tracking
    Tian, Dan
    Liu, Dong-Xin
    Wang, Xiao
    Hao, Ying
    [J]. IEEE Access, 2024, 12 : 154418 - 154425