Learning Disentangled Discrete Representations

被引:1
|
作者
Friede, David [1 ]
Reimers, Christian [2 ]
Stuckenschmidt, Heiner [1 ]
Niepert, Mathias [3 ,4 ]
机构
[1] Univ Mannheim, Mannheim, Germany
[2] Max Planck Inst Biogeochem, Jena, Germany
[3] Univ Stuttgart, Stuttgart, Germany
[4] NEC Labs Europe, Heidelberg, Germany
来源
MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: RESEARCH TRACK, ECML PKDD 2023, PT IV | 2023年 / 14172卷
关键词
Categorical VAE; Disentanglement;
D O I
10.1007/978-3-031-43421-1_35
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent successes in image generation, model-based reinforcement learning, and text-to-image generation have demonstrated the empirical advantages of discrete latent representations, although the reasons behind their benefits remain unclear. We explore the relationship between discrete latent spaces and disentangled representations by replacing the standard Gaussian variational autoencoder (VAE) with a tailored categorical variational autoencoder. We show that the underlying grid structure of categorical distributions mitigates the problem of rotational invariance associated with multivariate Gaussian distributions, acting as an efficient inductive prior for disentangled representations. We provide both analytical and empirical findings that demonstrate the advantages of discrete VAEs for learning disentangled representations. Furthermore, we introduce the first unsupervised model selection strategy that favors disentangled representations.
引用
收藏
页码:593 / 609
页数:17
相关论文
共 50 条
  • [21] KNOWLEDGE ROUTER: Learning Disentangled Representations for Knowledge Graphs
    Zhang, Shuai
    Rao, Xi
    Tay, Yi
    Zhang, Ce
    2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), 2021, : 1 - 10
  • [22] Temporal Consistency Objectives Regularize the Learning of Disentangled Representations
    Valvano, Gabriele
    Chartsias, Agisilaos
    Leo, Andrea
    Tsaftaris, Sotirios A.
    DOMAIN ADAPTATION AND REPRESENTATION TRANSFER AND MEDICAL IMAGE LEARNING WITH LESS LABELS AND IMPERFECT DATA, DART 2019, MIL3ID 2019, 2019, 11795 : 11 - 19
  • [23] A Sober Look at the Unsupervised Learning of Disentangled Representations and their Evaluation
    Locatello, Francesco
    Bauer, Stefan
    Lucic, Mario
    Ratsch, Gunnar
    Gelly, Sylvain
    Schoelkopf, Bernhard
    Bachem, Olivier
    JOURNAL OF MACHINE LEARNING RESEARCH, 2020, 21
  • [24] Learning disentangled representations for controllable human motion prediction
    Gu, Chunzhi
    Yu, Jun
    Zhang, Chao
    PATTERN RECOGNITION, 2024, 146
  • [25] Learning disentangled representations via product manifold projection
    Fumero, Marco
    Cosmo, Luca
    Melzi, Simone
    Rodola, Emanuele
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [26] Independent Subspace Analysis for Unsupervised Learning of Disentangled Representations
    Stuhmer, Jan
    Turner, Richard E.
    Nowozin, Sebastian
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 108, 2020, 108
  • [27] Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations
    Locatello, Francesco
    Bauer, Stefan
    Lucic, Mario
    Ratsch, Gunnar
    Gelly, Sylvain
    Scholkopf, Bernhard
    Bachem, Olivier
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [28] DEVIAS: Learning Disentangled Video Representations of Action and Scene
    Bae, Kyungho
    Ahn, Geo
    Kim, Youngrae
    Choi, Jinwoo
    COMPUTER VISION - ECCV 2024, PT XXXVII, 2025, 15095 : 431 - 448
  • [29] Conditional Mutual Information for Disentangled Representations in Reinforcement Learning
    Dunion, Mhairi
    McInroe, Trevor
    Luck, Kevin Sebastian
    Hanna, Josiah P.
    Albrecht, Stefano V.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [30] Leveraging Relational Information for Learning Weakly Disentangled Representations
    Valenti, Andrea
    Bacciu, Davide
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,