TTVAE: Transformer-based generative modeling for tabular data generation

被引:1
|
作者
Wang, Alex X. [1 ]
Nguyen, Binh P. [1 ,2 ]
机构
[1] Victoria Univ Wellington, Sch Math & Stat, Wellington 6012, New Zealand
[2] Ho Chi Minh City Open Univ, Fac Informat Technol, 97 Vo Van Tan,Dist 3, Ho Chi Minh City 70000, Vietnam
关键词
Generative AI; Tabular data; Transformer; Latent space interpolation; SMOTE;
D O I
10.1016/j.artint.2025.104292
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Tabular data synthesis presents unique challenges, with Transformer models remaining underexplored despite the applications of Variational Autoencoders and Generative Adversarial Networks. To address this gap, we propose the Transformer-based Tabular Variational AutoEncoder (TTVAE), leveraging the attention mechanism for capturing complex data distributions. The inclusion of the attention mechanism enables our model to understand complex relationships among heterogeneous features, a task often difficult for traditional methods. TTVAE facilitates the integration of interpolation within the latent space during the data generation process. Specifically, TTVAE is trained once, establishing a low-dimensional representation of real data, and then various latent interpolation methods can efficiently generate synthetic latent points. Through extensive experiments on diverse datasets, TTVAE consistently achieves state-of-the-art performance, highlighting its adaptability across different feature types and data sizes. This innovative approach, empowered by the attention mechanism and the integration of interpolation, addresses the complex challenges of tabular data synthesis, establishing TTVAE as a powerful solution.
引用
收藏
页数:17
相关论文
共 50 条
  • [31] A transformer-based approach to Nigerian Pidgin text generation
    Garba, Kabir
    Kolajo, Taiwo
    Agbogun, Joshua B.
    International Journal of Speech Technology, 2024, 27 (04) : 1027 - 1037
  • [32] Transformer-based image generation from scene graphs
    Sortino, Renato
    Palazzo, Simone
    Rundo, Francesco
    Spampinato, Concetto
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2023, 233
  • [33] Distance Correlation GAN: Fair Tabular Data Generation with Generative Adversarial Networks
    Rajabi, Amirarsalan
    Garibay, Ozlem Ozmen
    ARTIFICIAL INTELLIGENCE IN HCI, AI-HCI 2023, PT I, 2023, 14050 : 431 - 445
  • [34] LayoutDM: Transformer-based Diffusion Model for Layout Generation
    Chai, Shang
    Zhuang, Liansheng
    Yan, Fengying
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 18349 - 18358
  • [35] Arabic Paraphrase Generation Using Transformer-Based Approaches
    Al-Shameri, Noora Aref
    Al-Khalifa, Hend S.
    IEEE ACCESS, 2024, 12 : 121896 - 121914
  • [36] Tabular transformer generative adversarial network for heterogeneous distribution in healthcare
    Ha Ye Jin Kang
    Minsam Ko
    Kwang Sun Ryu
    Scientific Reports, 15 (1)
  • [37] Transformer-Based Deep Neural Language Modeling for Construct-Specific Automatic Item Generation
    Hommel, Bjoern E.
    Wollang, Franz-Josef M.
    Kotova, Veronika
    Zacher, Hannes
    Schmukle, Stefan C.
    PSYCHOMETRIKA, 2022, 87 (02) : 749 - 772
  • [38] TRANSFORMER-BASED ACOUSTIC MODELING FOR HYBRID SPEECH RECOGNITION
    Wang, Yongqiang
    Mohamed, Abdelrahman
    Le, Duc
    Liu, Chunxi
    Xiao, Alex
    Mahadeokar, Jay
    Huang, Hongzhao
    Tjandra, Andros
    Zhang, Xiaohui
    Zhang, Frank
    Fuegen, Christian
    Zweig, Geoffrey
    Seltzer, Michael L.
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 6874 - 6878
  • [39] Transformer-based Acoustic Modeling for Streaming Speech Synthesis
    Wu, Chunyang
    Xiu, Zhiping
    Shi, Yangyang
    Kalinli, Ozlem
    Fuegen, Christian
    Koehler, Thilo
    He, Qing
    INTERSPEECH 2021, 2021, : 146 - 150
  • [40] Transformer-Based Deep Neural Language Modeling for Construct-Specific Automatic Item Generation
    Björn E. Hommel
    Franz-Josef M. Wollang
    Veronika Kotova
    Hannes Zacher
    Stefan C. Schmukle
    Psychometrika, 2022, 87 : 749 - 772