TY - GEN
T1 - Language-Guided World Models A Model-Based Approach to AI Control
AU - Zhang, Alex
AU - Nguyen, Khanh
AU - Tuyls, Jens
AU - Lin, Albert
AU - Narasimhan, Karthik
N1 - Publisher Copyright:
© 2024 Association for Computational Linguistics.
PY - 2024
Y1 - 2024
N2 - This paper introduces the concept of Language-Guided World Models (LWMs)—probabilistic models that can simulate environments by reading texts. Agents equipped with these models provide humans with more extensive and efficient control, allowing them to simultaneously alter agent behaviors in multiple tasks via natural verbal communication. In this work, we take initial steps in developing robust LWMs that can generalize to compositionally novel language descriptions. We design a challenging world modeling benchmark based on the game of MESSENGER (Hanjie et al., 2021), featuring evaluation settings that require varying degrees of compositional generalization. Our experiments reveal the lack of generalizability of the state-of-the-art Transformer model, as it offers marginal improvements in simulation quality over a no-text baseline. We devise a more robust model by fusing the Transformer with the EMMA attention mechanism (Hanjie et al., 2021). Our model substantially outperforms the Transformer and approaches the performance of a model with an oracle semantic parsing and grounding capability. To demonstrate the practicality of this model in improving AI safety and transparency, we simulate a scenario in which the model enables an agent to present plans to a human before execution, and to revise plans based on their language feedback.
AB - This paper introduces the concept of Language-Guided World Models (LWMs)—probabilistic models that can simulate environments by reading texts. Agents equipped with these models provide humans with more extensive and efficient control, allowing them to simultaneously alter agent behaviors in multiple tasks via natural verbal communication. In this work, we take initial steps in developing robust LWMs that can generalize to compositionally novel language descriptions. We design a challenging world modeling benchmark based on the game of MESSENGER (Hanjie et al., 2021), featuring evaluation settings that require varying degrees of compositional generalization. Our experiments reveal the lack of generalizability of the state-of-the-art Transformer model, as it offers marginal improvements in simulation quality over a no-text baseline. We devise a more robust model by fusing the Transformer with the EMMA attention mechanism (Hanjie et al., 2021). Our model substantially outperforms the Transformer and approaches the performance of a model with an oracle semantic parsing and grounding capability. To demonstrate the practicality of this model in improving AI safety and transparency, we simulate a scenario in which the model enables an agent to present plans to a human before execution, and to revise plans based on their language feedback.
UR - http://www.scopus.com/inward/record.url?scp=85199222874&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85199222874&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85199222874
T3 - SpLU-RoboNLP 2024 - 4th Workshop on Spatial Language Understanding and Grounded Communication for Robotics, Proceedings of the Workshop
SP - 1
EP - 16
BT - SpLU-RoboNLP 2024 - 4th Workshop on Spatial Language Understanding and Grounded Communication for Robotics, Proceedings of the Workshop
A2 - Kordjamshidi, Parisa
A2 - Wang, Xin Eric
A2 - Zhang, Yue
A2 - Ma, Ziqiao
A2 - Inan, Mert
PB - Association for Computational Linguistics (ACL)
T2 - 4th Workshop on Spatial Language Understanding and Grounded Communication for Robotics, SpLU-RoboNLP 2024
Y2 - 11 August 2024 through 16 August 2024
ER -