Maximum Causal Entropy IRL in Mean-Field Games and GNEP Framework for Forward RL
Maximum Causal Entropy IRL in Mean-Field Games and GNEP Framework for Forward RL
This paper explores the use of Maximum Causal Entropy Inverse Reinforcement Learning (IRL) within the context of discrete-time stationary Mean-Field Games (MFGs) characterized by finite state spaces and an infinite-horizon, discounted-reward setting. Although the resulting optimization problem is non-convex with respect to policies, we reformulate it as a convex optimization problem in terms of state-action occupation measures by leveraging the linear programming framework of Markov Decision Processes. Based on this convex reformulation, we introduce a gradient descent algorithm with a guaranteed convergence rate to efficiently compute the optimal solution. Moreover, we develop a new method that conceptualizes the MFG problem as a Generalized Nash Equilibrium Problem (GNEP), enabling effective computation of the mean-field equilibrium for forward reinforcement learning (RL) problems and marking an advancement in MFG solution techniques. We further illustrate the practical applicability of our GNEP approach by employing this algorithm to generate data for numerical MFG examples.
Berkay Anahtarci、Can Deha Kariksiz、Naci Saldi
自动化基础理论计算技术、计算机技术
Berkay Anahtarci,Can Deha Kariksiz,Naci Saldi.Maximum Causal Entropy IRL in Mean-Field Games and GNEP Framework for Forward RL[EB/OL].(2025-07-19)[2025-08-06].https://arxiv.org/abs/2401.06566.点此复制
评论