Affine Variational Autoencoders

被引:1
|
作者
Bidart, Rene [1 ,2 ]
Wong, Alexander [1 ,2 ]
机构
[1] Waterloo Artificial Intelligence Inst, Waterloo, ON, Canada
[2] Univ Waterloo, Waterloo, ON, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
Deep learning; Variational autoencoders; Image generation; Perturbation;
D O I
10.1007/978-3-030-27202-9_42
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Variational autoencoders (VAEs) have in recent years become one of the most powerful approaches to learning useful latent representations of data in an unsupervised manner. However, a major challenge with VAEs is that they have tremendous difficulty in generalizing to data that deviate from the training set (e.g., perturbed image variants). Normally data augmentation is leveraged to overcome this limitation; however, this is not only computational expensive but also necessitates the construction of more complex models. In this study, we introduce the notion of affine variational autoencoders (AVAEs), which extends upon the conventional VAE architecture through the introduction of affine layers. More specifically, within the AVAE architecture an affine layer perturbs the input image prior to the encoder, and a second affine layer performs an inverse perturbation to the output of the decoder. The parameters of the affine layers are learned to enable the AVAE to encode images at canonical perturbations, resulting in a better reconstruction and a disentangled latent space without the need for data augmentation or the use of more complex models. Experimental results demonstrate the efficacy of the proposed VAE architecture for generalizing to images in the MNIST validation set under affine perturbations without the need for data augmentation, demonstrating significantly reduced loss when compared to conventional VAEs.
引用
收藏
页码:461 / 472
页数:12
相关论文
共 50 条
  • [1] Mixture variational autoencoders
    Jiang, Shuoran
    Chen, Yarui
    Yang, Jucheng
    Zhang, Chuanlei
    Zhao, Tingting
    [J]. PATTERN RECOGNITION LETTERS, 2019, 128 : 263 - 269
  • [2] An Introduction to Variational Autoencoders
    Kingma, Diederik P.
    Welling, Max
    [J]. FOUNDATIONS AND TRENDS IN MACHINE LEARNING, 2019, 12 (04): : 4 - 89
  • [3] Mixtures of Variational Autoencoders
    Ye, Fei
    Bors, Adrian G.
    [J]. 2020 TENTH INTERNATIONAL CONFERENCE ON IMAGE PROCESSING THEORY, TOOLS AND APPLICATIONS (IPTA), 2020,
  • [4] Subitizing with Variational Autoencoders
    Wever, Rijnder
    Runia, Tom F. H.
    [J]. COMPUTER VISION - ECCV 2018 WORKSHOPS, PT III, 2019, 11131 : 617 - 627
  • [5] Variational Laplace Autoencoders
    Park, Yookoon
    Kim, Chris Dongjoo
    Kim, Gunhee
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [6] Diffusion Variational Autoencoders
    Rey, Luis A. Perez
    Menkovski, Vlado
    Portegies, Jim
    [J]. PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 2704 - 2710
  • [7] Overdispersed Variational Autoencoders
    Shah, Harshil
    Barber, David
    Botev, Aleksandar
    [J]. 2017 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2017, : 1109 - 1116
  • [8] Ladder Variational Autoencoders
    Sonderby, Casper Kaae
    Raiko, Tapani
    Maaloe, Lars
    Sonderby, Soren Kaae
    Winther, Ole
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 29 (NIPS 2016), 2016, 29
  • [9] Tree Variational Autoencoders
    Manduchi, Laura
    Vandenhirtz, Moritz
    Ryser, Alain
    Vogt, Julia E.
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [10] Clockwork Variational Autoencoders
    Saxena, Vaibhav
    Ba, Jimmy
    Hafner, Danijar
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34