TY - GEN
T1 - TidyBot
T2 - 2023 IEEE/RSJ International Conference on Intelligent Robots and Systems, IROS 2023
AU - Wu, Jimmy
AU - Antonova, Rika
AU - Kan, Adam
AU - Lepert, Marion
AU - Zeng, Andy
AU - Song, Shuran
AU - Bohg, Jeannette
AU - Rusinkiewicz, Szymon
AU - Funkhouser, Thomas
N1 - Publisher Copyright:
© 2023 IEEE.
PY - 2023
Y1 - 2023
N2 - For a robot to personalize physical assistance effectively, it must learn user preferences that can be generally reapplied to future scenarios. In this work, we investigate personalization of household cleanup with robots that can tidy up rooms by picking up objects and putting them away. A key challenge is determining the proper place to put each object, as people's preferences can vary greatly depending on personal taste or cultural background. For instance, one person may prefer storing shirts in the drawer, while another may prefer them on the shelf. We aim to build systems that can learn such preferences from just a handful of examples via prior interactions with a particular person. We show that robots can combine language-based planning and perception with the few-shot summarization capabilities of large language models (LLMs) to infer generalized user preferences that are broadly applicable to future interactions. This approach enables fast adaptation and achieves 91.2% accuracy on unseen objects in our benchmark dataset. We also demonstrate our approach on a real-world mobile manipulator called TidyBot, which successfully puts away 85.0% of objects in real-world test scenarios.
AB - For a robot to personalize physical assistance effectively, it must learn user preferences that can be generally reapplied to future scenarios. In this work, we investigate personalization of household cleanup with robots that can tidy up rooms by picking up objects and putting them away. A key challenge is determining the proper place to put each object, as people's preferences can vary greatly depending on personal taste or cultural background. For instance, one person may prefer storing shirts in the drawer, while another may prefer them on the shelf. We aim to build systems that can learn such preferences from just a handful of examples via prior interactions with a particular person. We show that robots can combine language-based planning and perception with the few-shot summarization capabilities of large language models (LLMs) to infer generalized user preferences that are broadly applicable to future interactions. This approach enables fast adaptation and achieves 91.2% accuracy on unseen objects in our benchmark dataset. We also demonstrate our approach on a real-world mobile manipulator called TidyBot, which successfully puts away 85.0% of objects in real-world test scenarios.
UR - http://www.scopus.com/inward/record.url?scp=85165644163&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85165644163&partnerID=8YFLogxK
U2 - 10.1109/IROS55552.2023.10341577
DO - 10.1109/IROS55552.2023.10341577
M3 - Conference contribution
AN - SCOPUS:85165644163
T3 - IEEE International Conference on Intelligent Robots and Systems
SP - 3546
EP - 3553
BT - 2023 IEEE/RSJ International Conference on Intelligent Robots and Systems, IROS 2023
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 1 October 2023 through 5 October 2023
ER -