Back to Explorer
Research PaperResearchia:202602.25035[Artificial Intelligence > AI]

Recurrent Structural Policy Gradient for Partially Observable Mean Field Games

Clarisse Wibault

Abstract

Mean Field Games (MFGs) provide a principled framework for modeling interactions in large population models: at scale, population dynamics become deterministic, with uncertainty entering only through aggregate shocks, or common noise. However, algorithmic progress has been limited since model-free methods are too high variance and exact methods scale poorly. Recent Hybrid Structural Methods (HSMs) use Monte Carlo rollouts for the common noise in combination with exact estimation of the expected return, conditioned on those samples. However, HSMs have not been scaled to Partially Observable settings. We propose Recurrent Structural Policy Gradient (RSPG), the first history-aware HSM for settings involving public information. We also introduce MFAX, our JAX-based framework for MFGs. By leveraging known transition dynamics, RSPG achieves state-of-the-art performance as well as an order-of-magnitude faster convergence and solves, for the first time, a macroeconomics MFG with heterogeneous agents, common noise and history-aware policies. MFAX is publicly available at: https://github.com/CWibault/mfax.


Source: arXiv:2602.20141v1 - http://arxiv.org/abs/2602.20141v1 PDF: https://arxiv.org/pdf/2602.20141v1 Original Link: http://arxiv.org/abs/2602.20141v1

Submission:2/25/2026
Comments:0 comments
Subjects:AI; Artificial Intelligence
Original Source:
View Original PDF
arXiv: This paper is hosted on arXiv, an open-access repository
Was this helpful?

Discussion (0)

Please sign in to join the discussion.

No comments yet. Be the first to share your thoughts!

Recurrent Structural Policy Gradient for Partially Observable Mean Field Games | Researchia | Researchia