Continuous Strategy Replicator Dynamics for Multi--Agent Learning
The problem of multi-agent learning and adaptation has attracted a great deal of attention in recent years. It has been suggested that the dynamics of multi agent learning can be studied using replicator equations from population biology. Most existing studies so far have been limited to discrete strategy spaces with a small number of available actions. In many cases, however, the choices available to agents are better characterized by continuous spectra. This paper suggests a generalization of the replicator framework that allows to study the adaptive dynamics of Q-learning agents with continuous strategy spaces. Instead of probability vectors, agents strategies are now characterized by probability measures over continuous variables. As a result, the ordinary differential equations for the discrete case are replaced by a system of coupled integral–differential replicator equations that describe the mutual evolution of individual agent strategies. We derive a set of functional equations describing the steady state of the replicator dynamics, examine their solutions for several two-player games, and confirm our analytical results using simulations.
💡 Research Summary
The paper addresses a gap in the theoretical analysis of multi‑agent reinforcement learning (MARL) when agents choose actions from a continuous set. While previous work has successfully linked stateless Q‑learning with replicator dynamics, those studies were confined to discrete action spaces. The authors propose a continuous‑strategy extension of the replicator framework that captures the adaptive dynamics of Boltzmann‑exploring Q‑learning agents.
Key modeling steps: each agent i maintains a Q‑function Q_i(x,t) defined over the interval x∈
Comments & Academic Discussion
Loading comments...
Leave a Comment