Identifiability of deep generative models without auxiliary information

被引:0
|
作者
Kivva, Bohdan [1 ]
Rajendran, Goutham [1 ]
Ravikumar, Pradeep [2 ]
Aragam, Bryon [1 ]
机构
[1] Univ Chicago, Chicago, IL 60637 USA
[2] Carnegie Mellon Univ, Pittsburgh, PA 15213 USA
关键词
INDEPENDENT COMPONENT ANALYSIS; INFERENCE; MIXTURES;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We prove identifiability of a broad class of deep latent variable models that (a) have universal approximation capabilities and (b) are the decoders of variational autoencoders that are commonly used in practice. Unlike existing work, our analysis does not require weak supervision, auxiliary information, or conditioning in the latent space. Specifically, we show that for a broad class of generative (i.e. unsupervised) models with universal approximation capabilities, the side information u is not necessary: We prove identifiability of the entire generative model where we do not observe u and only observe the data x. The models we consider match autoencoder architectures used in practice that leverage mixture priors in the latent space and ReLU/leaky-ReLU activations in the encoder, such as VaDE and MFC-VAE. Our main result is an identifiability hierarchy that significantly generalizes previous work and exposes how different assumptions lead to different "strengths" of identifiability, and includes certain "vanilla" VAEs with isotropic Gaussian priors as a special case. For example, our weakest result establishes (unsupervised) identifiability up to an affine transformation, and thus partially resolves an open problem regarding model identifiability raised in prior work. These theoretical results are augmented with experiments on both simulated and real data.
引用
收藏
页数:15
相关论文
共 50 条
  • [1] Auxiliary Deep Generative Models
    Maaloe, Lars
    Sonderby, Casper Kaae
    Sonderby, Soren Kaae
    Winther, Ole
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 48, 2016, 48
  • [2] Toward the Identifiability of Comparative Deep Generative Models
    Lopez, Romain
    Huetter, Jan-Christian
    Hajiramezanali, Ehsan
    Pritchard, Jonathan K.
    Regev, Aviv
    CAUSAL LEARNING AND REASONING, VOL 236, 2024, 236 : 868 - 912
  • [3] Deep Generative Models for Relational Data with Side Information
    Hu, Changwei
    Rai, Piyush
    Carin, Lawrence
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 70, 2017, 70
  • [4] Prediction and confidence intervals for nonlinear measurement error models without identifiability information
    Huwang, L
    Hwang, JTG
    STATISTICS & PROBABILITY LETTERS, 2002, 58 (04) : 355 - 362
  • [5] Diversity in Deep Generative Models and Generative AI
    Turinici, Gabriel
    MACHINE LEARNING, OPTIMIZATION, AND DATA SCIENCE, LOD 2023, PT II, 2024, 14506 : 84 - 93
  • [6] Deep generative models in DataSHIELD
    Stefan Lenz
    Moritz Hess
    Harald Binder
    BMC Medical Research Methodology, 21
  • [7] Learning Deep Generative Models
    Salakhutdinov, Ruslan
    ANNUAL REVIEW OF STATISTICS AND ITS APPLICATION, VOL 2, 2015, 2 : 361 - 385
  • [8] AIC-GAN: An Auxiliary Information Classification GAN for Learning Deep Models
    Yang, Hao
    Zhou, Yun
    2020 CHINESE AUTOMATION CONGRESS (CAC 2020), 2020, : 6106 - 6111
  • [9] Metrics for Deep Generative Models
    Chen, Nutan
    Klushyn, Alexej
    Kurle, Richard
    Jiang, Xueyan
    Bayer, Justin
    van der Smagt, Patrick
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 84, 2018, 84
  • [10] Asymmetric deep generative models
    Partaourides, Harris
    Chatzis, Sotirios P.
    NEUROCOMPUTING, 2017, 241 : 90 - 96