Posterior Consistency for Missing Data in Variational Autoencoders

被引:0
|
作者
Sudak, Timur [1 ]
Tschiatschek, Sebastian [1 ,2 ]
机构
[1] Univ Vienna, Fac Comp Sci, Vienna, Austria
[2] Univ Vienna, Res Network Data Sci, Vienna, Austria
来源
MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: RESEARCH TRACK, ECML PKDD 2023, PT II | 2023年 / 14170卷
关键词
Variational Autoencoders; Missing Data;
D O I
10.1007/978-3-031-43415-0_30
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We consider the problem of learning Variational Autoencoders (VAEs), i.e., a type of deep generative model, from data with missing values. Such data is omnipresent in real-world applications of machine learning because complete data is often impossible or too costly to obtain. We particularly focus on improving a VAE's amortized posterior inference, i.e., the encoder, which in the case of missing data can be susceptible to learning inconsistent posterior distributions regarding the missingness. To this end, we provide a formal definition of posterior consistency and propose an approach for regularizing an encoder's posterior distribution which promotes this consistency. We observe that the proposed regularization suggests a different training objective than that typically considered in the literature when facing missing values. Furthermore, we empirically demonstrate that our regularization leads to improved performance in missing value settings in terms of reconstruction quality and downstream tasks utilizing uncertainty in the latent space. This improved performance can be observed for many classes of VAEs including VAEs equipped with normalizing flows.
引用
收藏
页码:508 / 524
页数:17
相关论文
共 50 条
  • [21] Variational autoencoders learn transferrable representations of metabolomics data
    Gomari, Daniel P.
    Schweickart, Annalise
    Cerchietti, Leandro
    Paietta, Elisabeth
    Fernandez, Hugo
    Al-Amin, Hassen
    Suhre, Karsten
    Krumsiek, Jan
    COMMUNICATIONS BIOLOGY, 2022, 5 (01)
  • [22] Seismic labeled data expansion using variational autoencoders
    Li, Kunhong
    Chen, Song
    Hu, Guangmin
    ARTIFICIAL INTELLIGENCE IN GEOSCIENCES, 2020, 1 : 24 - 30
  • [23] Variational autoencoders learn transferrable representations of metabolomics data
    Daniel P. Gomari
    Annalise Schweickart
    Leandro Cerchietti
    Elisabeth Paietta
    Hugo Fernandez
    Hassen Al-Amin
    Karsten Suhre
    Jan Krumsiek
    Communications Biology, 5
  • [24] A Generation of Enhanced Data by Variational Autoencoders and Diffusion Modeling
    Kim, Young-Jun
    Lee, Seok-Pil
    ELECTRONICS, 2024, 13 (07)
  • [25] Variational Mode Decomposition with Missing Data
    Choi, Guebin
    Oh, Hee-Seok
    Lee, Youngjo
    Kim, Donghoh
    Yu, Kyungsang
    KOREAN JOURNAL OF APPLIED STATISTICS, 2015, 28 (02) : 159 - 174
  • [26] Mixture variational autoencoders
    Jiang, Shuoran
    Chen, Yarui
    Yang, Jucheng
    Zhang, Chuanlei
    Zhao, Tingting
    PATTERN RECOGNITION LETTERS, 2019, 128 : 263 - 269
  • [27] Imputing missing indoor air quality data via variational convolutional autoencoders: Implications for ventilation management of subway metro systems
    Loy-Benitez, Jorge
    Heo, SungKu
    Yoo, ChangKyoo
    BUILDING AND ENVIRONMENT, 2020, 182
  • [28] An Introduction to Variational Autoencoders
    Kingma, Diederik P.
    Welling, Max
    FOUNDATIONS AND TRENDS IN MACHINE LEARNING, 2019, 12 (04): : 4 - 89
  • [29] Subitizing with Variational Autoencoders
    Wever, Rijnder
    Runia, Tom F. H.
    COMPUTER VISION - ECCV 2018 WORKSHOPS, PT III, 2019, 11131 : 617 - 627
  • [30] Mixtures of Variational Autoencoders
    Ye, Fei
    Bors, Adrian G.
    2020 TENTH INTERNATIONAL CONFERENCE ON IMAGE PROCESSING THEORY, TOOLS AND APPLICATIONS (IPTA), 2020,