Preference-Conditioned Language-Guided Abstraction

被引:2
|
作者
Peng, Andi [1 ]
Bobu, Andreea [2 ]
Li, Belinda Z. [1 ]
Sumers, Theodore R. [3 ]
Sucholutsky, Ilia [3 ]
Kumar, Nishanth [1 ]
Grifths, Thomas L. [3 ]
Shah, Julie A. [1 ]
机构
[1] MIT, Cambridge, MA 02139 USA
[2] Boston Dynam AI Inst, Cambridge, MA USA
[3] Princeton, Princeton, NJ USA
关键词
state abstraction; learning from human input; human preferences;
D O I
10.1145/3610977.3634930
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Learning from demonstrations is a common way for users to teach robots, but it is prone to spurious feature correlations. Recent work constructs state abstractions, i.e. visual representations containing task-relevant features, from language as a way to perform more generalizable learning. However, these abstractions also depend on a user's preference for what matters in a task, which may be hard to describe or infeasible to exhaustively specify using language alone. How do we construct abstractions to capture these latent preferences? We observe that how humans behave reveals how they see the world. Our key insight is that changes in human behavior inform us that there are diferences in preferences for how humans see the world, i.e. their state abstractions. In this work, we propose using language models (LMs) to query for those preferences directly given knowledge that a change in behavior has occurred. In our framework, we use the LM in two ways: frst, given a text description of the task and knowledge of behavioral change between states, we query the LM for possible hidden preferences; second, given the most likely preference, we query the LM to construct the state abstraction. In this framework, the LM is also able to ask the human directly when uncertain about its own estimate. We demonstrate our framework's ability to construct efective preference-conditioned abstractions in simulated experiments, a user study, as well as on a real Spot robot performing mobile manipulation tasks.
引用
收藏
页码:572 / 581
页数:10
相关论文
共 50 条
  • [41] A language-guided cross-modal semantic fusion retrieval method
    Zhu, Ligu
    Zhou, Fei
    Wang, Suping
    Shi, Lei
    Kou, Feifei
    Li, Zeyu
    Zhou, Pengpeng
    SIGNAL PROCESSING, 2025, 234
  • [42] CLUE: Contrastive language-guided learning for referring video object segmentation
    Gao, Qiqi
    Zhong, Wanjun
    Li, Jie
    Zhao, Tiejun
    PATTERN RECOGNITION LETTERS, 2024, 178 : 115 - 121
  • [43] Language-guided Multi-Modal Fusion for Video Action Recognition
    Hsiao, Jenhao
    Li, Yikang
    Ho, Chiuman
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW 2021), 2021, : 3151 - 3155
  • [44] Language-guided Semantic Mapping and Mobile Manipulation in Partially Observable Environments
    Patki, Siddharth
    Fahnestock, Ethan
    Howard, Thomas M.
    Walter, Matthew R.
    CONFERENCE ON ROBOT LEARNING, VOL 100, 2019, 100
  • [45] LASO: Language-guided Affordance Segmentation on 3D Object
    Li, Yicong
    Zhao, Na
    Xiao, Junbin
    Feng, Chun
    Wang, Xiang
    Chua, Tat-seng
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 14251 - 14260
  • [46] Scaling Up and Distilling Down: Language-Guided Robot Skill Acquisition
    Ha, Huy
    Florence, Pete
    Song, Shuran
    CONFERENCE ON ROBOT LEARNING, VOL 229, 2023, 229
  • [47] GVCCI: Lifelong Learning of Visual Grounding for Language-Guided Robotic Manipulation
    Kim, Junghyun
    Kang, Gi-Cheon
    Kim, Jaein
    Shin, Suyeon
    Zhang, Byoung-Tak
    2023 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, IROS, 2023, : 952 - 959
  • [48] LANDMARK: language-guided representation enhancement framework for scene graph generation
    Chang, Xiaoguang
    Wang, Teng
    Cai, Shaowei
    Sun, Changyin
    APPLIED INTELLIGENCE, 2023, 53 (21) : 26126 - 26138
  • [49] Language-Guided Controller Synthesis for Discrete-Time Linear Systems
    Gol, Ebru Aydin
    Lazar, Mircea
    Belta, Calin
    HSCC 12: PROCEEDINGS OF THE 15TH ACM INTERNATIONAL CONFERENCE ON HYBRID SYSTEMS: COMPUTATION AND CONTROL, 2012, : 95 - 104
  • [50] Language-Guided Traffic Simulation via Scene-Level Diffusion
    Zhong, Ziyuan
    Rempe, Davis
    Chen, Yuxiao
    Ivanovic, Boris
    Cao, Yulong
    Xu, Danfei
    Pavone, Marco
    Ray, Baishakhi
    CONFERENCE ON ROBOT LEARNING, VOL 229, 2023, 229