Compositional Generalization in Unsupervised Compositional Representation Learning: A Study on Disentanglement and Emergent Language

被引:0
|
作者
Xu, Zhenlin [1 ]
Niethammer, Marc [1 ]
Raffel, Colin [1 ]
机构
[1] Univ North Carolina Chapel Hill, Dept Comp Sci, Chapel Hill, NC 27599 USA
关键词
RECOGNITION;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep learning models struggle with compositional generalization, i.e. the ability to recognize or generate novel combinations of observed elementary concepts. In hopes of enabling compositional generalization, various unsupervised learning algorithms have been proposed with inductive biases that aim to induce compositional structure in learned representations (e.g. disentangled representation and emergent language learning). In this work, we evaluate these unsupervised learning algorithms in terms of how well they enable compositional generalization. Specifically, our evaluation protocol focuses on whether or not it is easy to train a simple model on top of the learned representation that generalizes to new combinations of compositional factors. We systematically study three unsupervised representation learning algorithms - beta-VAE, beta-TCVAE, and emergent language (EL) autoencoders - on two datasets that allow directly testing compositional generalization. We find that directly using the bottleneck representation with simple models and few labels may lead to worse generalization than using representations from layers before or after the learned representation itself. In addition, we find that the previously proposed metrics for evaluating the levels of compositionality are not correlated with the actual compositional generalization in our framework. Surprisingly, we find that increasing pressure to produce a disentangled representation (e.g. increasing beta in the beta-VAE) produces representations with worse generalization, while representations from EL models show strong compositional generalization. Motivated by this observation, we further investigate the advantages of using EL to induce compositional structure in unsupervised representation learning, finding that it shows consistently stronger generalization than disentanglement models, especially when using less unlabeled data for unsupervised learning and fewer labels for downstream tasks. Taken together, our results shed new light onto the compositional generalization behavior of different unsupervised learning algorithms with a new setting to rigorously test this behavior, and suggest the potential benefits of developing EL learning algorithms for more generalizable representations.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] Deciphering the Role of Representation Disentanglement: Investigating Compositional Generalization in CLIP Models
    Abbasi, Reza
    Rohban, Mohammad Hossein
    Baghshah, Mahdieh Soleymani
    COMPUTER VISION - ECCV 2024, PT LXXXIX, 2025, 15147 : 35 - 50
  • [2] UNSUPERVISED LEARNING OF COMPOSITIONAL SPARSE CODE FOR NATURAL IMAGE REPRESENTATION
    Hong, Yi
    Si, Zhangzhang
    Hu, Wenze
    Zhu, Song-Chun
    Wu, Ying Nian
    QUARTERLY OF APPLIED MATHEMATICS, 2014, 72 (02) : 373 - 406
  • [3] Compositional Generalization with Grounded Language Models
    Wold, Sondre
    Simon, Etienne
    Georges, Lucas
    Charpentier, Gabriel
    Kostylev, Egor V.
    Velldal, Erik
    Ovrelid, Lilja
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 3447 - 3460
  • [4] Compositional Generalization in Spoken Language Understanding
    Ray, Avik
    Shen, Yilin
    Jin, Hongxia
    INTERSPEECH 2023, 2023, : 750 - 754
  • [5] Compositional Generalization in Grounded Language Learning via Induced Model Sparsity
    Spilsbury, Sam
    Ilin, Alexander
    NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES: PROCEEDINGS OF THE STUDENT RESEARCH WORKSHOP, 2022, : 143 - 155
  • [6] Representation Learning of Compositional Data
    Avalos-Fernandez, Marta
    Nock, Richard
    Ong, Cheng Soon
    Rouar, Julien
    Sun, Ke
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [7] Unsupervised Learning of Compositional Energy Concepts
    Du, Yilun
    Li, Shuang
    Sharma, Yash
    Tenenbaum, Joshua B.
    Mordatch, Igor
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021,
  • [8] Curriculum learning for human compositional generalization
    Dekker, Ronald B.
    Otto, Fabian
    Summerfield, Christopher
    PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2022, 119 (41)
  • [9] Learning Algebraic Recombination for Compositional Generalization
    Liu, Chenyao
    An, Shengnan
    Lin, Zeqi
    Liu, Qian
    Chen, Bei
    Lou, Jian-Guang
    Wen, Lijie
    Zheng, Nanning
    Zhang, Dongmei
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 1129 - 1144
  • [10] Hierarchical Poset Decoding for Compositional Generalization in Language
    Guo, Yinuo
    Lin, Zeqi
    Lou, Jian-Guang
    Zhang, Dongmei
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33