TY - GEN
T1 - Preference-Conditioned Language-Guided Abstraction
AU - Peng, Andi
AU - Bobu, Andreea
AU - Li, Belinda Z.
AU - Sumers, Theodore R.
AU - Sucholutsky, Ilia
AU - Kumar, Nishanth
AU - Griffiths, Thomas L.
AU - Shah, Julie A.
N1 - Publisher Copyright:
© 2024 IEEE Computer Society. All rights reserved.
PY - 2024/3/11
Y1 - 2024/3/11
N2 - Learning from demonstrations is a common way for users to teach robots, but it is prone to spurious feature correlations. Recent work constructs state abstractions, i.e. visual representations containing task-relevant features, from language as a way to perform more generalizable learning. However, these abstractions also depend on a user's preference for what matters in a task, which may be hard to describe or infeasible to exhaustively specify using language alone. How do we construct abstractions to capture these latent preferences? We observe that how humans behave reveals how they see the world. Our key insight is that changes in human behavior inform us that there are differences in preferences for how humans see the world, i.e. their state abstractions. In this work, we propose using language models (LMs) to query for those preferences directly given knowledge that a change in behavior has occurred. In our framework, we use the LM in two ways: first, given a text description of the task and knowledge of behavioral change between states, we query the LM for possible hidden preferences; second, given the most likely preference, we query the LM to construct the state abstraction. In this framework, the LM is also able to ask the human directly when uncertain about its own estimate. We demonstrate our framework's ability to construct effective preference-conditioned abstractions in simulated experiments, a user study, as well as on a real Spot robot performing mobile manipulation tasks.
AB - Learning from demonstrations is a common way for users to teach robots, but it is prone to spurious feature correlations. Recent work constructs state abstractions, i.e. visual representations containing task-relevant features, from language as a way to perform more generalizable learning. However, these abstractions also depend on a user's preference for what matters in a task, which may be hard to describe or infeasible to exhaustively specify using language alone. How do we construct abstractions to capture these latent preferences? We observe that how humans behave reveals how they see the world. Our key insight is that changes in human behavior inform us that there are differences in preferences for how humans see the world, i.e. their state abstractions. In this work, we propose using language models (LMs) to query for those preferences directly given knowledge that a change in behavior has occurred. In our framework, we use the LM in two ways: first, given a text description of the task and knowledge of behavioral change between states, we query the LM for possible hidden preferences; second, given the most likely preference, we query the LM to construct the state abstraction. In this framework, the LM is also able to ask the human directly when uncertain about its own estimate. We demonstrate our framework's ability to construct effective preference-conditioned abstractions in simulated experiments, a user study, as well as on a real Spot robot performing mobile manipulation tasks.
KW - human preferences
KW - learning from human input
KW - state abstraction
UR - http://www.scopus.com/inward/record.url?scp=85188449734&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85188449734&partnerID=8YFLogxK
U2 - 10.1145/3610977.3634930
DO - 10.1145/3610977.3634930
M3 - Conference contribution
AN - SCOPUS:85188449734
T3 - ACM/IEEE International Conference on Human-Robot Interaction
SP - 572
EP - 581
BT - HRI 2024 - Proceedings of the 2024 ACM/IEEE International Conference on Human-Robot Interaction
PB - IEEE Computer Society
T2 - 19th Annual ACM/IEEE International Conference on Human-Robot Interaction, HRI 2024
Y2 - 11 March 2024 through 15 March 2024
ER -