Implications of data topology for deep generative models

被引:0
|
作者
Jin, Yinzhu [1 ]
Mcdaniel, Rory [1 ]
Tatro, N. Joseph [2 ]
Catanzaro, Michael J. [3 ]
Smith, Abraham D. [3 ,4 ]
Bendich, Paul [3 ,5 ]
Dwyer, Matthew B. [1 ]
Fletcher, P. Thomas [1 ,6 ]
机构
[1] Univ Virginia, Dept Comp Sci, Charlottesville, VA 22903 USA
[2] STR Vis & Image Understanding Grp, Woburn, MA USA
[3] Geometr Data Analyt Inc, Durham, NC USA
[4] Univ Wisconsin Stout, Math Stats & CS Dept, Menomonie, WI USA
[5] Duke Univ, Dept Math, Durham, NC USA
[6] Univ Virginia, Dept Elect & Comp Engn, Charlottesville, VA USA
来源
基金
美国国家科学基金会;
关键词
data topology; generative model; variational autoencoder (VAE); diffusion probabilistic models (DDPM); topological data analysis;
D O I
10.3389/fcomp.2024.1260604
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Many deep generative models, such as variational autoencoders (VAEs) and generative adversarial networks (GANs), learn an immersion mapping from a standard normal distribution in a low-dimensional latent space into a higher-dimensional data space. As such, these mappings are only capable of producing simple data topologies, i.e., those equivalent to an immersion of Euclidean space. In this work, we demonstrate the limitations of such latent space generative models when trained on data distributions with non-trivial topologies. We do this by training these models on synthetic image datasets with known topologies (spheres, torii, etc.). We then show how this results in failures of both data generation as well as data interpolation. Next, we compare this behavior to two classes of deep generative models that in principle allow for more complex data topologies. First, we look at chart autoencoders (CAEs), which construct a smooth data manifold from multiple latent space chart mappings. Second, we explore score-based models, e.g., denoising diffusion probabilistic models, which estimate gradients of the data distribution without resorting to an explicit mapping to a latent space. Our results show that these models do demonstrate improved ability over latent space models in modeling data distributions with complex topologies, however, challenges still remain.
引用
收藏
页数:15
相关论文
共 50 条
  • [1] Deep Generative Design: Integration of Topology Optimization and Generative Models
    Oh, Sangeun
    Jung, Yongsu
    Kim, Seongsin
    Lee, Ikjin
    Kang, Namwoo
    [J]. JOURNAL OF MECHANICAL DESIGN, 2019, 141 (11)
  • [2] Deep Generative Models for Synthetic Data: A Survey
    Eigenschink, Peter
    Reutterer, Thomas
    Vamosi, Stefan
    Vamosi, Ralf
    Sun, Chang
    Kalcher, Klaudius
    [J]. IEEE ACCESS, 2023, 11 : 47304 - 47320
  • [3] Data-driven topology design using a deep generative model
    Yamasaki, Shintaro
    Yaji, Kentaro
    Fujita, Kikuo
    [J]. STRUCTURAL AND MULTIDISCIPLINARY OPTIMIZATION, 2021, 64 (03) : 1401 - 1420
  • [4] Data-driven topology design using a deep generative model
    Shintaro Yamasaki
    Kentaro Yaji
    Kikuo Fujita
    [J]. Structural and Multidisciplinary Optimization, 2021, 64 : 1401 - 1420
  • [5] Efficient Ring-Topology Decentralized Federated Learning with Deep Generative Models for Medical Data in eHealthcare Systems
    Wang, Zhao
    Hu, Yifan
    Yan, Shiyang
    Wang, Zhihao
    Hou, Ruijie
    Wu, Chao
    [J]. ELECTRONICS, 2022, 11 (10)
  • [6] On oversampling imbalanced data with deep conditional generative models
    Fajardo, Val Andrei
    Findlay, David
    Jaiswal, Charu
    Yin, Xinshang
    Houmanfar, Roshanak
    Xie, Honglei
    Liang, Jiaxi
    She, Xichen
    Emerson, D. B.
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2021, 169
  • [7] Deep Generative Models for Relational Data with Side Information
    Hu, Changwei
    Rai, Piyush
    Carin, Lawrence
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 70, 2017, 70
  • [8] The application of deep generative models in urban form generation based on topology: a review
    Lin, Bo
    Jabi, Wassim
    Corcoran, Padraig
    Lannon, Simon
    [J]. ARCHITECTURAL SCIENCE REVIEW, 2024, 67 (03) : 189 - 204
  • [9] Disease variant prediction with deep generative models of evolutionary data
    Frazer, Jonathan
    Notin, Pascal
    Dias, Mafalda
    Gomez, Aidan
    Min, Joseph K.
    Brock, Kelly
    Gal, Yarin
    Marks, Debora S.
    [J]. NATURE, 2021, 599 (7883) : 91 - +
  • [10] Neurosymbolic Deep Generative Models for Sequence Data with Relational Constraints
    Young, Halley
    Du, Maxwell
    Bastani, Osbert
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,