Compositional Generalization in Unsupervised Compositional Representation Learning: A Study on Disentanglement and Emergent Language

被引:0
|
作者
Xu, Zhenlin [1 ]
Niethammer, Marc [1 ]
Raffel, Colin [1 ]
机构
[1] Univ North Carolina Chapel Hill, Dept Comp Sci, Chapel Hill, NC 27599 USA
关键词
RECOGNITION;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep learning models struggle with compositional generalization, i.e. the ability to recognize or generate novel combinations of observed elementary concepts. In hopes of enabling compositional generalization, various unsupervised learning algorithms have been proposed with inductive biases that aim to induce compositional structure in learned representations (e.g. disentangled representation and emergent language learning). In this work, we evaluate these unsupervised learning algorithms in terms of how well they enable compositional generalization. Specifically, our evaluation protocol focuses on whether or not it is easy to train a simple model on top of the learned representation that generalizes to new combinations of compositional factors. We systematically study three unsupervised representation learning algorithms - beta-VAE, beta-TCVAE, and emergent language (EL) autoencoders - on two datasets that allow directly testing compositional generalization. We find that directly using the bottleneck representation with simple models and few labels may lead to worse generalization than using representations from layers before or after the learned representation itself. In addition, we find that the previously proposed metrics for evaluating the levels of compositionality are not correlated with the actual compositional generalization in our framework. Surprisingly, we find that increasing pressure to produce a disentangled representation (e.g. increasing beta in the beta-VAE) produces representations with worse generalization, while representations from EL models show strong compositional generalization. Motivated by this observation, we further investigate the advantages of using EL to induce compositional structure in unsupervised representation learning, finding that it shows consistently stronger generalization than disentanglement models, especially when using less unlabeled data for unsupervised learning and fewer labels for downstream tasks. Taken together, our results shed new light onto the compositional generalization behavior of different unsupervised learning algorithms with a new setting to rigorously test this behavior, and suggest the potential benefits of developing EL learning algorithms for more generalizable representations.
引用
收藏
页数:14
相关论文
共 50 条
  • [21] Compositional Networks Enable Systematic Generalization for Grounded Language Understanding
    Kuo, Yen-Ling
    Katz, Boris
    Barbu, Andrei
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 216 - 226
  • [22] Learning Compositional Tasks from Language Instructions
    Logeswaran, Lajanugen
    Carvalho, Wilka
    Lee, Honglak
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 11, 2023, : 13300 - 13308
  • [23] PMGNet: Disentanglement and entanglement benefit mutually for compositional zero-shot learning
    Liu, Yu
    Li, Jianghao
    Zhang, Yanyi
    Jia, Qi
    Wang, Weimin
    Pu, Nan
    Sebe, Nicu
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2024, 249
  • [24] Unsupervised Learning of Camera Pose with Compositional Re-estimation
    Nabavi, Seyed Shahabeddin
    Hosseinzadeh, Mehrdad
    Fahimi, Ramin
    Wang, Yang
    2020 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2020, : 11 - 20
  • [25] Improving Compositional Generalization using Iterated Learning and Simplicial Embeddings
    Ren, Yi
    Lavoie, Samuel
    Galkin, Mikhail
    Sutherland, Danica J.
    Courville, Aaron
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [26] Compositional generalization through meta sequence-to-sequence learning
    Lake, Brenden M.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [27] Compositional Scene Representation Learning via Reconstruction: A Survey
    Yuan, Jinyang
    Chen, Tonglin
    Li, Bin
    Xue, Xiangyang
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (10) : 11540 - 11560
  • [28] Unbiased Semantic Representation Learning Based on Causal Disentanglement for Domain Generalization
    Jin, Xuanyu
    Li, Ni
    Kong, Wangzeng
    Tang, Jiajia
    Yang, Bing
    ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2024, 20 (08)
  • [29] Learning a Hierarchical Compositional Representation of Multiple Object Classes
    Leonardis, Ales
    2009 IEEE COMPUTER SOCIETY CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPR WORKSHOPS 2009), VOLS 1 AND 2, 2009, : 529 - 529
  • [30] Unsupervised Learning of Compositional Scene Representations from Multiple Unspecified Viewpoints
    Yuan, Jinyang
    Li, Bin
    Xue, Xiangyang
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 8971 - 8979