General agents need world models
General agents need world models
Are world models a necessary ingredient for flexible, goal-directed behaviour, or is model-free learning sufficient? We provide a formal answer to this question, showing that any agent capable of generalizing to multi-step goal-directed tasks must have learned a predictive model of its environment. We show that this model can be extracted from the agent's policy, and that increasing the agents performance or the complexity of the goals it can achieve requires learning increasingly accurate world models. This has a number of consequences: from developing safe and general agents, to bounding agent capabilities in complex environments, and providing new algorithms for eliciting world models from agents.
Jonathan Richens、David Abel、Alexis Bellot、Tom Everitt
自动化基础理论
Jonathan Richens,David Abel,Alexis Bellot,Tom Everitt.General agents need world models[EB/OL].(2025-06-02)[2025-06-29].https://arxiv.org/abs/2506.01622.点此复制
评论