Transformer World Model for Sample Efficient Multi-Agent Reinforcement Learning
Transformer World Model for Sample Efficient Multi-Agent Reinforcement Learning
We present the Multi-Agent Transformer World Model (MATWM), a novel transformer-based world model designed for multi-agent reinforcement learning in both vector- and image-based environments. MATWM combines a decentralized imagination framework with a semi-centralized critic and a teammate prediction module, enabling agents to model and anticipate the behavior of others under partial observability. To address non-stationarity, we incorporate a prioritized replay mechanism that trains the world model on recent experiences, allowing it to adapt to agents' evolving policies. We evaluated MATWM on a broad suite of benchmarks, including the StarCraft Multi-Agent Challenge, PettingZoo, and MeltingPot. MATWM achieves state-of-the-art performance, outperforming both model-free and prior world model approaches, while demonstrating strong sample efficiency, achieving near-optimal performance in as few as 50K environment interactions. Ablation studies confirm the impact of each component, with substantial gains in coordination-heavy tasks.
Azad Deihim、Eduardo Alonso、Dimitra Apostolopoulou
计算技术、计算机技术
Azad Deihim,Eduardo Alonso,Dimitra Apostolopoulou.Transformer World Model for Sample Efficient Multi-Agent Reinforcement Learning[EB/OL].(2025-06-23)[2025-07-02].https://arxiv.org/abs/2506.18537.点此复制
评论