TTVAE: Transformer-based generative modeling for tabular data generation

被引:1
|
作者
Wang, Alex X. [1 ]
Nguyen, Binh P. [1 ,2 ]
机构
[1] Victoria Univ Wellington, Sch Math & Stat, Wellington 6012, New Zealand
[2] Ho Chi Minh City Open Univ, Fac Informat Technol, 97 Vo Van Tan,Dist 3, Ho Chi Minh City 70000, Vietnam
关键词
Generative AI; Tabular data; Transformer; Latent space interpolation; SMOTE;
D O I
10.1016/j.artint.2025.104292
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Tabular data synthesis presents unique challenges, with Transformer models remaining underexplored despite the applications of Variational Autoencoders and Generative Adversarial Networks. To address this gap, we propose the Transformer-based Tabular Variational AutoEncoder (TTVAE), leveraging the attention mechanism for capturing complex data distributions. The inclusion of the attention mechanism enables our model to understand complex relationships among heterogeneous features, a task often difficult for traditional methods. TTVAE facilitates the integration of interpolation within the latent space during the data generation process. Specifically, TTVAE is trained once, establishing a low-dimensional representation of real data, and then various latent interpolation methods can efficiently generate synthetic latent points. Through extensive experiments on diverse datasets, TTVAE consistently achieves state-of-the-art performance, highlighting its adaptability across different feature types and data sizes. This innovative approach, empowered by the attention mechanism and the integration of interpolation, addresses the complex challenges of tabular data synthesis, establishing TTVAE as a powerful solution.
引用
收藏
页数:17
相关论文
共 50 条
  • [41] Transformer-based conditional generative transfer learning network for cross domain fault diagnosis under limited data
    Li, Xiang
    Ma, Jun
    Wu, Jiande
    Li, Zhuorui
    Tan, Zhengzhou
    SCIENTIFIC REPORTS, 2025, 15 (01):
  • [42] The research landscape on generative artificial intelligence: a bibliometric analysis of transformer-based models
    Sekli, Giulio Marchena
    KYBERNETES, 2024,
  • [43] TRANSFORMER-BASED DOMAIN ADAPTATION FOR EVENT DATA CLASSIFICATION
    Zhao, Junwei
    Zhang, Shiliang
    Huang, Tiejun
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 4673 - 4677
  • [44] Classification of hyperspectral and LiDAR data by transformer-based enhancement
    Pan, Jiechen
    Shuai, Xing
    Xu, Qing
    Dai, Mofan
    Zhang, Guoping
    Wang, Guo
    REMOTE SENSING LETTERS, 2024, 15 (10) : 1074 - 1084
  • [45] The interactive reading task: Transformer-based automatic item generation
    Attali, Yigal
    Runge, Andrew
    LaFlair, Geoffrey T.
    Yancey, Kevin
    Goodwin, Sarah
    Park, Yena
    von Davier, Alina A.
    FRONTIERS IN ARTIFICIAL INTELLIGENCE, 2022, 5
  • [46] A transformer-based approach for improving app review response generation
    Zhang, Weizhe
    Gu, Wenchao
    Gao, Cuiyun
    Lyu, Michael R.
    SOFTWARE-PRACTICE & EXPERIENCE, 2023, 53 (02): : 438 - 454
  • [47] Transformer-based protein generation with regularized latent space optimization
    Egbert Castro
    Abhinav Godavarthi
    Julian Rubinfien
    Kevin Givechian
    Dhananjay Bhaskar
    Smita Krishnaswamy
    Nature Machine Intelligence, 2022, 4 : 840 - 851
  • [48] Sparse Transformer-Based Sequence Generation for Visual Object Tracking
    Tian, Dan
    Liu, Dong-Xin
    Wang, Xiao
    Hao, Ying
    IEEE ACCESS, 2024, 12 : 154418 - 154425
  • [49] Synthetic seismocardiogram generation using a transformer-based neural network
    Nikbakht, Mohammad
    Gazi, Asim H.
    Zia, Jonathan
    An, Sungtae
    Lin, David J.
    Inan, Omer T.
    Kamaleswaran, Rishikesan
    JOURNAL OF THE AMERICAN MEDICAL INFORMATICS ASSOCIATION, 2023, 30 (07) : 1266 - 1273
  • [50] Transformer-based protein generation with regularized latent space optimization
    Castro, Egbert
    Godavarthi, Abhinav
    Rubinfien, Julian
    Givechian, Kevin
    Bhaskar, Dhananjay
    Krishnaswamy, Smita
    NATURE MACHINE INTELLIGENCE, 2022, 4 (10) : 840 - 851