Learning Disentangled Discrete Representations

被引:1
|
作者
Friede, David [1 ]
Reimers, Christian [2 ]
Stuckenschmidt, Heiner [1 ]
Niepert, Mathias [3 ,4 ]
机构
[1] Univ Mannheim, Mannheim, Germany
[2] Max Planck Inst Biogeochem, Jena, Germany
[3] Univ Stuttgart, Stuttgart, Germany
[4] NEC Labs Europe, Heidelberg, Germany
关键词
Categorical VAE; Disentanglement;
D O I
10.1007/978-3-031-43421-1_35
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent successes in image generation, model-based reinforcement learning, and text-to-image generation have demonstrated the empirical advantages of discrete latent representations, although the reasons behind their benefits remain unclear. We explore the relationship between discrete latent spaces and disentangled representations by replacing the standard Gaussian variational autoencoder (VAE) with a tailored categorical variational autoencoder. We show that the underlying grid structure of categorical distributions mitigates the problem of rotational invariance associated with multivariate Gaussian distributions, acting as an efficient inductive prior for disentangled representations. We provide both analytical and empirical findings that demonstrate the advantages of discrete VAEs for learning disentangled representations. Furthermore, we introduce the first unsupervised model selection strategy that favors disentangled representations.
引用
收藏
页码:593 / 609
页数:17
相关论文
共 50 条
  • [1] Learning Disentangled Joint Continuous and Discrete Representations
    Dupont, Emilien
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [2] Learning Disentangled Representations for Recommendation
    Ma, Jianxin
    Zhou, Chang
    Cui, Peng
    Yang, Hongxia
    Zhu, Wenwu
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [3] Learning Disentangled Representations of Negation and Uncertainty
    Vasilakes, Jake
    Zerva, Chrysoula
    Miwa, Makoto
    Ananiadou, Sophia
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 8380 - 8397
  • [4] Domain Agnostic Learning with Disentangled Representations
    Peng, Xingchao
    Huang, Zijun
    Sun, Ximeng
    Saenko, Kate
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [5] A Contrastive Objective for Learning Disentangled Representations
    Kahana, Jonathan
    Hoshen, Yedid
    COMPUTER VISION, ECCV 2022, PT XXVI, 2022, 13686 : 579 - 595
  • [6] Learning disentangled representations in the imaging domain
    Liu, Xiao
    Sanchez, Pedro
    Thermos, Spyridon
    O'Neil, Alison Q.
    Tsaftaris, Sotirios A.
    MEDICAL IMAGE ANALYSIS, 2022, 80
  • [7] A Commentary on the Unsupervised Learning of Disentangled Representations
    Locatello, Francesco
    Bauer, Stefan
    Lucie, Mario
    Raetsch, Gunnar
    Gelly, Sylvain
    Schoelkopf, Bernhard
    Bachem, Olivier
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 13681 - 13684
  • [8] On Learning Disentangled Representations for Gait Recognition
    Zhang, Ziyuan
    Tran, Luan
    Liu, Feng
    Liu, Xiaoming
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (01) : 345 - 360
  • [9] Learning Disentangled Representations with the Wasserstein Autoencoder
    Gaujac, Benoit
    Feige, Ilya
    Barber, David
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2021: RESEARCH TRACK, PT III, 2021, 12977 : 69 - 84
  • [10] Learning Disentangled Representations of Video with Missing Data
    Massague, Armand Comas
    Zhang, Chi
    Feric, Zlatan
    Camps, Octavia
    Yu, Rose
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33