Preference-Conditioned Language-Guided Abstraction

被引:2
|
作者
Peng, Andi [1 ]
Bobu, Andreea [2 ]
Li, Belinda Z. [1 ]
Sumers, Theodore R. [3 ]
Sucholutsky, Ilia [3 ]
Kumar, Nishanth [1 ]
Grifths, Thomas L. [3 ]
Shah, Julie A. [1 ]
机构
[1] MIT, Cambridge, MA 02139 USA
[2] Boston Dynam AI Inst, Cambridge, MA USA
[3] Princeton, Princeton, NJ USA
关键词
state abstraction; learning from human input; human preferences;
D O I
10.1145/3610977.3634930
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Learning from demonstrations is a common way for users to teach robots, but it is prone to spurious feature correlations. Recent work constructs state abstractions, i.e. visual representations containing task-relevant features, from language as a way to perform more generalizable learning. However, these abstractions also depend on a user's preference for what matters in a task, which may be hard to describe or infeasible to exhaustively specify using language alone. How do we construct abstractions to capture these latent preferences? We observe that how humans behave reveals how they see the world. Our key insight is that changes in human behavior inform us that there are diferences in preferences for how humans see the world, i.e. their state abstractions. In this work, we propose using language models (LMs) to query for those preferences directly given knowledge that a change in behavior has occurred. In our framework, we use the LM in two ways: frst, given a text description of the task and knowledge of behavioral change between states, we query the LM for possible hidden preferences; second, given the most likely preference, we query the LM to construct the state abstraction. In this framework, the LM is also able to ask the human directly when uncertain about its own estimate. We demonstrate our framework's ability to construct efective preference-conditioned abstractions in simulated experiments, a user study, as well as on a real Spot robot performing mobile manipulation tasks.
引用
收藏
页码:572 / 581
页数:10
相关论文
共 50 条
  • [31] Language-Guided Transformer for Federated Multi-Label Classification
    Liu, I-Jieh
    Lin, Ci-Siang
    Yang, Fu-En
    Wang, Yu-Chiang Frank
    arXiv, 2023,
  • [32] LapsCore: Language-guided Person Search via Color Reasoning
    Wu, Yushuang
    Yan, Zizheng
    Han, Xiaoguang
    Li, Guanbin
    Zou, Changqing
    Cui, Shuguang
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 1604 - 1613
  • [33] Towards Language-Guided Visual Recognition via Dynamic Convolutions
    Luo, Gen
    Zhou, Yiyi
    Sun, Xiaoshuai
    Wu, Yongjian
    Gao, Yue
    Ji, Rongrong
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2024, 132 (01) : 1 - 19
  • [34] LucIE: Language-guided local image editing for fashion images
    Wen, Huanglu
    You, Shaodi
    Fu, Ying
    COMPUTATIONAL VISUAL MEDIA, 2025, 11 (01): : 179 - 194
  • [35] Language-Guided Transformer for Federated Multi-Label Classification
    Liu, I-Jieh
    Lin, Ci-Siang
    Yang, Fu-En
    Wang, Yu-Chiang Frank
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 12, 2024, : 13882 - 13890
  • [36] Language-Guided Progressive Attention for Visual Grounding in Remote Sensing Images
    Li, Ke
    Wang, Di
    Xu, Haojie
    Zhong, Haodi
    Wang, Cong
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2024, 62 : 1 - 1
  • [37] LANDMARK: language-guided representation enhancement framework for scene graph generation
    Xiaoguang Chang
    Teng Wang
    Shaowei Cai
    Changyin Sun
    Applied Intelligence, 2023, 53 : 26126 - 26138
  • [38] LPN: Language-Guided Prototypical Network for Few-Shot Classification
    Cheng, Kaihui
    Yang, Chule
    Liu, Xiao
    Guan, Naiyang
    Wang, Zhiyuan
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2025, 35 (01) : 632 - 642
  • [39] Language-Guided Face Animation by Recurrent StyleGAN-Based Generator
    Hang, Tiankai
    Yang, Huan
    Liu, Bei
    Fu, Jianlong
    Geng, Xin
    Guo, Baining
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 9216 - 9227
  • [40] DenseCLIP: Language-Guided Dense Prediction with Context-Aware Prompting
    Rao, Yongming
    Zhao, Wenliang
    Chen, Guangyi
    Tang, Yansong
    Zhu, Zheng
    Huang, Guan
    Zhou, Jie
    Lu, Jiwen
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 18061 - 18070