Review of Mean-Field Theory for the Robotic Swarms Common Target Problem
Passos, Y. T.; Nunes, J. V. N. S.
DOI 10.5433/1679-0375.2025.v46.53577
Citation Semin., Ciênc. Exatas Tecnol. 2025, v. 46: e53577
Received: August 15, 2025 Received in revised for: October 24, 2025 Accepted: October 30, 2025 Available online: November 30, 2025
Abstract:
This review provides a literature overview of the usage of Mean-Field Theory (MFT) for modelling the performance estimation of algorithms solving the common target problem in robotic swarm navigation. This problem involves coordinating multiple robots to converge simultaneously towards a single target, which can result in congestion and degraded performance. Its main objective is to minimise the maximum time required for all robots to reach and then depart from the target. The challenge lies in ensuring effective coordination among the robots to prevent conflicts and optimise resource utilisation. MFT may accurately predict navigation metrics, such as agent arrival and departure times, target location density, and system throughput under congestion, and it may enable control strategies based on distributed feedback. This literature overview illustrates that MFT is an effective and promising mathematical tool to mitigate the issues of high dimensionality and complexity of interactions present in systems consisting of a large number of agents.
Keywords: robotic swarms, mean-field theory, common target problem, differential equations
Introduction
Robotics is an engineering and computer science field dedicated to the design and production of robots. The aim of robotics is to design machines that can carry out certain tasks and interact with the world, augmenting human abilities and systematising processes.
The domain of robotics spans a broad spectrum of fields, ranging from industrial robots to medical and space robots. The use of swarm-based approaches in the development of robotics systems has attracted long-standing interest (Bonabeau et al., 1999; Couzin et al., 2002). Swarms consists of large numbers of collaborating robots to conduct complicated tasks. One of the major challenges facing this field is the formulation of modelling and control methods that scale favourably with swarm size. Considering the swarm as a continuum presents one solution to this problem.
Among the complex tasks addressed in robotic swarms, the common target problem stands out as a particularly challenging scenario that directly exemplifies the need for scalable modelling and control. A pervasive and highly challenging issue in robotic swarm navigation under such conditions is the Common Target Problem. This problem arises when multiple robots attempt to reach the same target simultaneously, creating congestion scenarios that may degrade performance (Yuri Tavares Passos et al., 2023).
The core challenge is coordinating the convergence of large groups of robots into a confined space; any lack of coordination would lead to conflicts and inefficient utilisation of resources. Although the mathematical formulation and constraint definition are elucidated, optimising the dynamics of long-term experiments such as the common target problem poses substantial control difficulties. The escalating complexity associated with an increasing number of robots frequently renders centralised control strategies impractical, thereby underscoring the necessity for approaches that enhance system scalability and robustness.
Therefore, this study focuses on this specific problem within robotic swarms. Let a swarm of \(N\) robots be modelled such that the dynamics of each robot \(i \in \{1,\dots,N\}\) at position \(\mathbf{r}_{i}(t) = (r_{ix}(t), r_{iy}(t), r_{iz}(t))\) at time \(t\) are governed by attractive and repulsive forces. Let \(\dot{\mathbf{r}}_{i} = \frac{d \mathbf{r}_{i}}{d t}\) and \(\ddot{\mathbf{r}}_{i} = \frac{d \dot{\mathbf{r}}_{i}}{d t}\). The force \(\mathbf{F}_{A}(\mathbf{T})\) attracts the robot towards a target located at \(\mathbf{T}\) while \(\mathbf{F}_{R}\) is the repulsive force applied to avoid bumping into other robots. The working and target areas are circles with radius \(D\) and \(\epsilon\), respectively, centred on \(\mathbf{T}\) with \(D > \epsilon\). Let \(t_{ti}\) and \(t_{fi}\) be the time to reach the target area and leave the working region, respectively, and \(t_{i} = t_{ti} + t_{fi}\). After a robot leaves the working area, it proceeds to a new target located at \(\mathbf{N}_i\). All robots are assumed to be circles of radius \(d_{r}\). Hence, they have to stay apart from each other by at least \(2d_{r}\). The maximum time between all robots, \(\displaystyle{\max_{i \in \{1, \dots, N\}}t_{i}}\), and their total work must to be minimised, that is,
The coefficient \(K_{\text{res}}\) determines the magnitude of the resultant force obtained by the attractive and repulsive forces, while \(K_{\text{dp}}\) governs motion dissipation, reducing oscillations and smoothing trajectories as the robots move towards their targets. In essence, the equation (1) quantifies the overall cost or effort for a robotic swarm to complete a common task, specifically aiming to minimise the longest time any single robot takes to reach and return from a target, while also considering the accumulated cost of movement over time.
Typically, two categories of optimisation criteria can be used in task allocation to achieve a stated system objective: maximisation (reward-based) or minimisation (cost-based). The first category is employed when there is a reward for assigning a task to any robot, while the second is more applicable in practice when a cost exists for performing a task and can be further subcategorised as follows. To tackle these optimal control challenges, a variety of strategies have been proposed for traffic management and collision avoidance within multi-robot environments.
Within this context, Mean-Field Theory (MFT) provides a robust mathematical tool for modelling collective behaviour by aggregating all agents (Lasry & Lions, 2007; Carmona & Delarue, 2018; Bensoussan et al., 2013), replacing intricate, agent-specific dynamics with macroscopic approximations often expressed through partial and ordinary differential equations. Therefore, an analysis of MFT-based approaches for addressing the common target problem in robotic swarms is presented in this work. These observations lead to the objective of this paper, which is guided by two primary questions:
How has MFT been utilised in the modelling and performance evaluation of algorithms designed for addressing the common target problem in robotic swarms?
What are the principal challenges, advantages, and potential applications identified in the literature?
Every subsequent step, from the development of a search plan to data analysis, adhered to these questions as guiding principles.
Materials and methods
To guarantee reproducibility, the following eligibility criteria were defined:
Studies addressing modelling, control, performance estimation, or behaviour analysis using MFT in multi-agent systems focused on robotic swarms.
Only research published in English was included, with no restrictions on publication date to encompass both recent advances and foundational work.
A systematic search strategy was developed to identify relevant studies. Search terms included combinations of “Mean-Field Control”, “Mean-Field Games”, or “Collective Dynamics” with “Robotic Swarms”, “Multi-agent Systems”, or “Collective Robotics”, alongside keywords such as “Performance Estimation”, “Modelling”, “Control Strategies”, “Behaviour Analysis”, “Throughput”, and “Congestion”. To maintain focus, works that did not explicitly apply MFT to robotic swarms were excluded, as were those from unrelated domains, extended abstracts without full texts, and duplicates.
Results and discussion
The principal significance of MFT resides in its capability to circumvent the curse of dimensionality inherent to the control of numerous interacting agents by representing the system as a continuous probability distribution. This abstraction markedly diminishes the computational strain, facilitating the design of scalable and computationally tractable control strategies.
By concentrating upon the average behaviour, MFT enables the prediction and control of aggregate swarm dynamics without the need to track the individual state of every robot. Figure 1 illustrates this behaviour, showing both the agent’s influence on the collective and the simplification that the mean-field formulation introduces into the interaction structure.
Adapted from "Mean field multi-agent reinforcement learning" by Y. Yang, R. Luo, M. Li, M. Zhou, W. Zhang & J. Wang, (Yang et al., 2018), Proceedings of the 35th International Conference on Machine Learning.
MFT treats energy as a shared resource, formulating optimisation problems that minimise total swarm power rather than individual consumption, yielding efficient trajectories, balanced density and power-saving behaviours that extend mission life. The mathematical engines that turn local rules into global behaviour were examined.
MFT elevates performance in complex tasks, for instance, area coverage, search and rescue (Menec et al., 2024) and environmental monitoring, by distilling swarm behaviour into concise mathematical forecasts. It supplies unified frameworks that predict collective motion, enabling cooperative strategies for thorough coverage, precise localisation and efficient resource use. Framing tasks as Mean-Field Control or Game problems yields global strategies that raise success rates while curbing congestion and redundant effort (Ornia et al., 2022). MFT further underpins task allocation (Cui et al., 2023), coverage, self-assembly , consensus, flocking (Elamvazhuthi et al., 2021) and strategic interactions.
Continuous settings rely on Fokker–Planck Partial Differential Equations (PDE). When agents pursue selfish optima, a Mean-Field Game emerges, coupling Hamilton–Jacobi–Bellman and Fokker–Planck equations. Essentially, these equations model how the density or distribution of a large population of agents evolves over time and space, accounting for their movement, ensuring that the total number of agents is conserved within the system (Burger et al., 2014).
Thus, MFT moves beyond theory to become indispensable for scalable, energy-aware and task-focused swarms. It replaces individual control with unified continuum models.
MFT becomes an adequate method of simulation and control of collective behaviours in robotic swarms using modelling techniques borrowed from natural sciences—for example, astrophysics (Hazra et al., 2023), biology (Couzin et al., 2002) and statistical mechanics (Vicsek et al., 1995). Thus, MFT assumes that individual agent behaviour can be approximated by the mean behaviour of all agents within a system. Hence, simplifying the macroscopic description and making it much more accessible for analysis. Under this approximation, two major theoretical frameworks have emerged developed: Mean-Field Game (MFG) and Mean-Field Control (MFC) problems.
Despite substantial research growth and notable progress in applying MFT to robotic swarms, a more focused synthesis in the review literature remains valuable. While existing reviews have comprehensively explored MFT in robotics (Elamvazhuthi et al., 2019), their scope has not specifically extended to a thorough assessment of MFT’s role in performance evaluation for common target scenarios, nor have they systematically compared different control strategies amidst congestion and conflicting agent preferences.
Consequently, although the foundational context of MFT in multi-agent systems is well covered, a dedicated review addressing the practical implications of MFT in handling these complex swarm behaviours, particularly concerning performance evaluation and specific control strategy comparisons, is timely and complementary to the existing body of work.
Following the established framework, the ensuing sections are systematically structured according to the classification of mean-field models. The initial section delves into the application in robotic swarms of two primary categories of mean-field models, infinite-dimensional continuous and finite-dimensional discrete, facilitating a comparative analysis between these two paradigms.
The second section addresses mean-field models that originate from the fields of physics and statistical mechanics, which elucidate the mechanisms by which complex physical concepts are examined via mean-field approximations. Having framed the challenge, how MFT enables the modelling and control of collective behaviours in robotic swarms is analysed below.
This review uniquely bridges the gap between MFT theory and practical swarm robotics challenges, providing validated solutions for congestion and collision avoidance areas largely overlooked in existing literature.To better clarify the distinction between existing reviews on Mean Field Theory applied to robotic swarms and the current review, the main differences can be observed in Table 1.
| Reference | Main Objectives | Main Themes | Practical Implications |
|---|---|---|---|
| Elamvazhuthi and Berman (2019) | To provide a fundamental understanding and overview of MFT applications in multi-agent systems. |
⋅ Theoretical foundations of MFT. ⋅ General applicability in robotics. |
The authors articulate theoretical advantages; however, they establish a less explicit connection to detailed validation and practical application. |
| Vicsek et al. (1995) | To model collective motion and phase transitions in systems of self-driven particles using statistical physics. |
⋅ Local alignment rules. ⋅ Phase transition to global order. ⋅ Noise-driven fluctuations. |
Demonstrates how local interaction rules lead to global coherence; forms the basis for swarm coordination models in robotics. |
| Bensoussan et al. (2013) | To introduce mean-field games and mean-field type control theory. |
⋅ Mathematical foundations of MFG and MFC. ⋅ Applications to large-scale systems. |
Provides rigorous theoretical background, but lacks direct application to robotic swarms. |
| Carmona and Delarue (2018) | To present a probabilistic theory of mean-field games and applications. |
⋅ Stochastic analysis of MFG. ⋅ Coupled HJB and FP equations. |
High mathematical complexity; limited focus on robotic systems or empirical validation. |
| Lasry and Lions (2007) | To establish the foundational theory of mean-field games. |
⋅ Nash equilibria in large populations. ⋅ Coupled PDE systems (HJB + FP). |
Introduces core MFG concepts; not tailored to robotic swarm applications. |
| Transporte et al. (2022) | To propose a top-down control strategy for robotic swarms using mean-field PDEs. |
⋅ Density control via velocity fields. ⋅ Input-to-State Stability (ISS). |
Demonstrates scalable control with ISS guarantees; relevant to common target problems. |
| This Review | To provide a pragmatic and concentrated analysis of the utility of MFT in addressing the challenges related to performance estimation and control in robotic swarms. |
⋅ Performance assessment using MFT. ⋅ Comparison of control strategies. ⋅ Prediction of navigation metrics. ⋅ Validation via direct simulation. |
The current review serves as a conduit between the overarching principles of MFT and its empirically validated application in addressing particular congestion and conflict issues within robotic swarms, a domain that has not been comprehensively analysed in preceding reviews. |
Mean-field models in robotic swarms
Building on this theoretical framework, it is evident that the inherent complexities of robotic swarms, which are characterised by a large number of interacting agents, provide both tremendous opportunities and difficult obstacles. While this paradigm has great real-world application potential, notably in terms of scalability, robustness, and decentralised function, it is also fraught with complications and challenges.
Efficiently managing the synchronisation of large, distributed teams is crucial for maximising the potential of these systems. This fundamental duality, the appealing promise and the complex coordination challenges, serves as a foundation for recognising the need of advanced modelling and control frameworks in this rapidly evolving domain.
The transition from individual dynamics to collective behaviour is critical to scalable analysis and control. Recent progress demonstrates its integration with modern methods. For example, Zheng et al. (2022) proposes a “top-down” control strategy for robotic swarms, utilising mean-field PDEs to model and control the swarm’s global density.
The approach generates velocity fields that guide the robots’ distribution to a desired target profile, with robustness guaranteed through Input-to-State Stability (ISS) analysis against perturbations and modelling errors. Although the velocity field design is centralised, the implementation on individual robots is fully distributed, making the solution practical for large-scale systems. This MFT-based control method is highly pertinent for resolving the common target problem in robotic swarms, as it optimises collective movement, positively impacting the arrival time at the destination and managing distribution to mitigate congestion and overcrowding.
In Dogbe et al. (2010) the work is highly relevant to the application of MFT in robotic swarms. It proposes a framework for modelling the dynamics of pedestrian crowds using differential games, the dynamics of an individual pedestrian is described by the state vector \(X(t):=(x,v)\), where \(x\) represents the position and \(v\) the velocity. The equations of motion are \(\frac{d}{dt}x(t) = v(t)\) and \(\frac{d}{dt}v(t) = \gamma(t),\) where \(\gamma(t)\) is the acceleration vector that acts as the control.
The general cost functional for a trajectory, considering transport costs and a terminal cost, is given by \[J(x,v)=\int_{0}^{T}L(t,x(t),v(t))dt+\phi(T,x(T)),\] where \(L(t,x,v)\) is the instantaneous cost function (running cost) and \(\phi(T,x(T))\) is the terminal cost at the final position of the pedestrian at time \(T\), reflecting the quality or effort to reach this final location. MFG is used to simplify the complex interactions among a large number of individuals, allowing each pedestrian’s decisions to be modelled in relation to the statistical distribution of all others.
The Hamilton-Jacobi-Bellman (HJB) equation (2) below (Dogbe et al., 2010) is the mathematical apparatus governing the optimal dynamics of this agent continuum. Fundamentally, it describes how the optimal value (or minimum cost-to-go) for an individual agent changes over time and space, given its dynamics and the costs it incurs, serving as a cornerstone for optimal control problems in MFG; they emerge as a necessary condition for each agent to minimise its describes the optimal cost \(u(t,x)\) for an individual pedestrian, given by
The evolution of the pedestrian population density \(\rho(t,x)\) is described by the Fokker-Planck equation (3) below (continuity equation), which describes how the probability \(\rho(t,x)\) changes due to the flow of agents:
The HJB equation (4) below, for the optimal control problem of a representative agent, which now depends on the population density
\[\frac{\partial u}{\partial t}(t,x) + H(x,\nabla u(t,x),\rho(t,x)) = 0, \text{in } (0,T) \times \Omega, \label{eq:hjb1} \tag{4} \]with the terminal condition \(u(T,x) = \phi(x,\rho(T,x))\). The Hamiltonian \(H(x,p,\rho)\) is defined as \(H(x,p,\rho) = \max_{v\in\mathbb{R}^{d}}\{-v\cdot p-L(x,v,\rho)\}.\) Here, the cost function \(L(x,v,\rho)\) now depends on the population density \(\rho\).The Fokker-Planck equation (5) below represent the evolution of the population density, which depends on the value function \(u\)
\[\frac{\partial \rho}{\partial t}(t,x) - \nabla \cdot (\rho(t,x) \nabla_{p} H(x,\nabla u(t,x),\rho(t,x))) = 0, \text{in } (0,T) \times \Omega, \label{eq:fokkerplanck} \tag{5} \]with the initial condition \(\rho(0,x) = \rho_{0}(x)\).
Having mapped the landscape, the continuum limit where PDEs replace discrete agents was zoomed into the last subsection.
Infinite-dimensional mean-field models
This section details the fundamental models extracted from the reviewed literature, which form the theoretical basis for the analysis of robotic swarm systems. The models, primarily derived from PDE, address congestion control and common target problems through mean-field approaches.
In these models, control generally focuses on managing the distribution of agents, which can have continuous or discrete states. It was assumed that the robots have local perception, allowing them to assess the density of robots in their neighbourhood or other aggregated properties of the swarm. Direct communication between robots is often reduced or absent, with interactions mediated by the mean field. These challenges have shifted attention towards the phenomena of diffusion, subdiffusion, and superdiffusion, which are recognised for introducing characteristics such as chaos and heightened dimensionality in robotic swarm simulations. This methodology effectively substitutes intricate, agent-specific dynamics with macroscopic approximations derived from population averages, commonly formulated through ordinary or partial differential equations such as the Fokker-Planck equation (5) (Auricchio et al., 2022).
Notably, Lévy flight within the realm of robotics, which emulates the foraging pathways of animals such as albatrosses, bumblebees, and deer, have been integrated into the depiction of complex dynamics. Beyond the basic paradigm of Brownian motion, characterized by Jean Perrin as the continuous, unpredictable motion of particles immersed in a liquid, Perrin articulated this phenomenon in French, stating: "Ils vont et viennent en tournoyant, montent, descendent, remontent encore, sans tendre aucunement vers le repos." This phrase describes the dynamic capacity of particles to move back and forth, as depicted two-dimensionally in Figure 2 and three-dimensionally in Figure 3.
| (a) | (b) |
![]() | ![]() |

Trajectory visualizations illustrate that Brownian motion follows a relatively compact and diffusive pathway, consisting of small, random steps, thereby maintaining proximity to its origin. Conversely, a Lévy flight displays intermittent, substantial leaps, allowing it to traverse greater distances from its point of origin within the same number of steps. This superdiffusive behavior is facilitated by the flight’s heavy-tailed step-length distribution (a Lévy stable distribution), enabling more effective spatial exploration compared to Brownian motion, which is characterized by normally distributed step-lengths, resulting in a linear scaling of mean squared displacement over time. To model these complex phenomena, fractional calculus is utilised as a critical mathematical instrument. This facilitates the inclusion of memory effects and non-local interactions, thereby refining mean-field equations to more accurately depict anomalous diffusion and swarm exploration.
The fundamental equations presented by Estrada-rodriguez et al. (2020) are crucial for modelling robotic swarm system dynamics, encompassing individual movement, stopping, and interactions. The microscopic movement of individuals is governed by the \(N\)-particle probability density function \(\sigma\), which evolves according to the kinetic equations: \[\frac{\partial\sigma}{\partial t} +\sum_{i=1}^{N}\Bigl( \frac{\partial\sigma}{\partial\tau_{i}} +c\,\theta_{i}\!\cdot\!\nabla_{\!x_{i}}\sigma \Bigr) =-\sum_{i=1}^{N}\beta_{i}\sigma,\] where \(\sigma=\sigma(x_{1},\dots,x_{N},t,\theta_{1},\dots,\theta_{N},\tau_{1},\dots,\tau_{N})\) is the \(N\)-particle probability density; \(x_{i}\in\mathbb{R}^{n}\) is the position of robot \(i\); \(\theta_{i}\in S=\{v\in\mathbb{R}^{n}:|v|=1\}\) is its direction of motion; \(\tau_{i}\ge 0\) is its elapsed run time since the last reorientation; \(c>0\) is the constant speed; \(\beta_{i}\) is the stopping frequency defined below. The probability that robot \(i\) has not stopped after run time \(\tau_i\) follows the power law \[\psi_{i}(x_{i},\tau_{i}) =\left( \frac{\zeta_{0}(x_{i})}{\zeta_{0}(x_{i})+\tau_{i}} \right)^{\!\alpha}, \qquad 1<\alpha<2,\] with \(\zeta_{0}(x_{i})>0\) a spatially-varying scale parameter. The stopping frequency is then \[\beta_{i}(x_{i},\tau_{i}) =-\frac{1}{\psi_{i}}\frac{\partial\psi_{i}}{\partial\tau_{i}} =\frac{\varphi_{i}(x_{i},\tau_{i})}{\psi_{i}(x_{i},\tau_{i})},\] where \(\varphi_{i}(x_{i},\tau_{i}) =-\partial\psi_{i}/\partial\tau_{i},\) is the hazard rate. Upon stopping, a robot chooses a new direction \(\theta_{i}^{*}\) via tumbling or alignment. The turning kernel is \[T_{i}\phi(\theta_{i}^{*}) =\int_{S} k(x_{i},t,\theta_{i};\theta_{i}^{*})\,\phi(\theta_{i})\,d\theta_{i},\] where \(k(x_{i},t,\theta_{i};\theta_{i}^{*})\) is the transition probability density from direction \(\theta_{i}\) to \(\theta_{i}^{*}\). The mean direction of the neighbours at position \(x_{i}\) is \[\Lambda_{i}(x_{i},\theta_{i},t) =\frac{\mathcal{J}(x_{i},t)}{|\mathcal{J}(x_{i},t)|},\] \[\mathcal{J}(x_{i},t) =\int_{\mathbb{R}^{n}}\!\int_{S} K(|x_{j}-x_{i}|)\,p(x_{j},t,\theta_{j})\,\theta_{j}\, d\theta_{j}\,dx_{j},\] with \(K(|x_{j}-x_{i}|)\) an influence kernel (decaying with distance); \(p(x_{j},t,\theta_{j})\) the one-particle density of robots at \((x_{j},\theta_{j})\). For the two-particle density \(\sigma(x_{1},x_{2},t,\theta_{1},\theta_{2},\tau_{1},\tau_{2})\), the governing equation (6) is
The work of Huang et al. (2020) rigorously justifies the mean-field limit of an \(N\)-particle system, deriving the Vlasov-Poisson-Fokker-Planck (VPFP) (Huang et al., 2020) equations from a regularised microscopic \(N\)-particle system. The core of their work involves several key mathematical formulations that describe the particle dynamics at both microscopic and macroscopic levels along with specific interaction kernels. It follows that the microscopic \(N\)-particle system, subject to Brownian motions and interacting through a Newtonian potential in \(\mathbb{R}^{3}\), is described by the following system of Stochastic Differential Equations (SDE):
The paper of Huang et al. (2020) also considers a first-order stochastic system for particles: \[dx_{i}=\frac{1}{N-1}\sum_{j\ne i}^{N}k(x_{i}-x_{j})dt+\sqrt{2\sigma}dB_{i},\] for \(i=1,...,N.\) The continuous description of the dynamics for this first-order system is given by the following non-linear PDE for the spatial density \(f(x,t)\): \[\frac{\partial f(x,t)}{\partial t} + \nabla \cdot (f(k*f)) = \sigma \Delta_{x} f.\]
Specific interaction kernels are presented for various applications. For fluid dynamics, the Biot-Savart kernel is given by: \[k(x)=\frac{1}{2\pi}\left(\frac{-x_{2}}{|x|^{2}},\frac{x_{1}}{|x|^{2}}\right).\]
For chemotaxis models, the Poisson kernel is often used. Thus, the paper introduces the notation for trajectories on phase space for the time evolution of the \(N\)-interacting Newtonian particles: \(\Phi_{t}:=(X_{t},V_{t}):=(x_{1}^{t},...,x_{N}^{t},v_{1}^{t},...v_{N}^{t}),\) which explicitly states the Newtonian system with a noise term coupled to the velocity as a system of SDEs. It is essentially a detailed restatement of the equation (7) with explicit time dependence.
Furthermore, the equations of Auricchio et al. (2022) describe a novel Fokker-Planck-type model designed to simulate manufacturing processes through the collective dynamics of a large set of agents, aiming for uniform coverage of a target domain. The mesoscopic model in the study of Auricchio et al. (2022) translating the agent dynamics to a type of the same as in equation (5), with a constant diffusion and a discontinuous drift, expressed in divergence form as:
In equation (8) below, the function characterising the drift is expressed by:
The same equilibrium configuration as in the equation (11) can also be obtained from an alternative formula similar to the equation (12) below, with a variable diffusion coefficient:
The values \(m_1\) and \(m_2\), which quantify the percentages of mass for the Gaussian and uniform parts, respectively, are determined by solving the following linear system resulting from mass conservation and continuity conditions:
Similarly, Huang et al. (2006) addresses stochastic dynamic games in large populations where agents are weakly coupled, using the Nash Certainty Equivalence (NCE) Principle, to derive decentralised control synthesis via McKean-Vlasov systems. The individual agent optimal control problem is defined by minimising a cost functional \(J\), considering the population distribution \(P_s\): \[J(t,x,v,P) = \inf_{u_{t}} E_{t,x,v} \left[ \int_{t}^{\infty} r(s, X_{s}, v_{s}, P_{s})ds + S(X_{s}) \right].\]
The optimal value function \(V(t,x,v,P)\) for the representative agent is governed by the equation (15) below:
The evolution of the population distribution \(P(t,x,v)\) is described by the equation (16) below, which is of Fokker-Planck type:
The usage of the overbar observed with the Fokker-Planck type equation means an averaged quantity, as in the term \(\bar{u}(t,x,v,P)\). Here, \(\bar{u}\) denotes a collective or spatially averaged quantity, often referred to as the mean field. This structure is critical because it introduces non-linearity into the system dynamics: the local transport (advection) of particles is not only governed by local states \((x,v)\) but is also influenced by the aggregated, mean behaviour of the entire population distribution \(P\). This dependency is fundamental to models describing self-organising systems and collective dynamics.
In the specific context of a Linear Quadratic Gaussian (LQG) framework, the individual agent dynamics is given by the SDE: \[dx_{t}=(A_{0}x_{t}+A_{1}\bar{x}_{t}+B u_{t})dt+Fdw_{t}.\]
The quadratic cost functional for each individual agent in the LQG problem is \[J(u) = E \left[ \int_{0}^{\infty} (||x_{t}||_{Q}^{2}+||u_{t}||_{R}^{2})dt \right].\]
The HJB equation (15) for the value function \(V(x)\) in this LQG setting becomes the quadratic equation (17) below:
In the stationary HJB equation for the Linear-Quadratic-Gaussian (LQG) control problem, the terms \(\bar{x}\) and \(\bar{Q}\) denote quantities linked to state uncertainty. \(\bar{x}\) signifies the conditional mean state, the optimal estimate of the true system state \(x\), typically supplied by a Kalman filter. Its presence in the quadratic cost function \((\bar{x}T\bar{Q}\bar{ x})\) and the drift term demonstrates the fundamental separation principle of LQG control. Furthermore, \(\bar{Q}\) is the cost weighting matrix specifically associated with the estimated state \(\bar{x}\). While \(Q\) penalises the deviation of the true state \(x\), \(\bar{Q}\) allows the cost function to also penalise or reward outcomes based on the estimated or perceived state, which is crucial for systems operating under partial observation.
From the equation (17), the optimal control law \(u_t\) for the individual agent is derived as \(u_{t} = -R^{-1}B^{T}\nabla V(x_{t}).\) Applying this optimal control yields the linearised state equation (18) below for \(x_t\):
In the equation (18), \(\bar{x}_{t}\) denotes the estimated state at time \(t\). This equation describes the evolution of the actual state \(x_{t}\) when subject to the optimal control policy. Although the system state is \(x_{t}\), the optimal control input is calculated based only on the best available estimate, \(\bar{x}_{t}\), because the true state is corrupted by noise. The appearance of \(\bar{x}_{t}\) in the drift term, \(A_{1}\bar{x}_{t}\) , therefore explicitly confirms that the controller is basing its corrective action on its filtered knowledge of the system, a key feature of optimal stochastic feedback control.
The resulting McKean-Vlasov SDE for the individual state \(x_t\), incorporating the mean-field effect (\(E[x_t]\)), is \[dx_{t}=(A_{0}x_{t}+A_{1}E[x_{t}]-B R^{-1}B^{T}P x_{t})dt+Fdw_{t}.\]
Thus, the evolution of the mean of the state \(E[x_t]\) is governed by a linear Ordinary Differential Equation (ODE) as in the equation (19) below:
Another study related to infinite-dimensional mean-field model is the MFG concept. Its primary objective, as introduced by (Lasry et al., 2007), is to tackle the formidable complexity found in systems that involve an exceptionally large number of rational agents. This approach seeks to make such intricate systems manageable for analysis and modelling.
MFG theory eschews the computationally intractable task of modelling every pairwise interaction, a problem of exponential complexity, in favour of a paradigm shift: it models the decision-making of a single, representative agent. This agent does not react to specific counterparts but instead responds to the aggregate and anonymised influence of the entire population, termed the Mean-Field.
This innovative strategy reduces an ostensibly high-dimensional problem to a coupled system of two PDEs: one governing the individual optimal strategy, equation (20), and the other describing the temporal evolution of the population distribution, equation (21). Thus, Lasry et al. (2007) seek to establish and validate a rigorous yet tractable mathematical framework capable of identifying equilibrium solutions in complex collective dynamics that would otherwise be analytically unattainable from the equations
Here, \(u(t,x)\) is the value function (or function of the cost to go) for a representative agent, and \(m(t,x)\) is the density of agents in the state space. The time horizon is denoted by \(T\) and \(\mathbb{T}^d\) (a \(d\)-dimensional torus) as the spatial domain is considered periodic.
The Hamiltonian function \(H(x,p)\) describes the dynamics of the individual agent and the cost associated with its control:
Lasry et al. (2007) also discuss two advanced mathematical concepts in control theory: classical optimal control and MFG. Classical optimal control uses the coupling of state PDEs with backward-propagating adjoint equations to optimise the system trajectory. In strategic settings, MFG refines this approach by seeking equilibrium through coupled PDE systems. The text emphasises the importance of robustness alongside optimality, highlighting the role of Input-to-State Stability (ISS), which ensures system stability and bounded error amid external disturbances like sensor noise or model inaccuracies. These methods, while distinct in their goals – optimality and robustness – utilise mean-field abstraction to derive control laws that ensure efficiency, scalability, and reliability in complex systems.
Zheng et al. (2022) describe the evolution of the swarm’s mean-field density \(\rho(x,t)\) over a bounded spatial domain \(\Omega \subset \mathbb{R}^d\) by the mean-field PDE
The optimal control law \(u(x,t)\) that minimises the objective function equation (26) is derived from the equation (23) as a feedback law
The full characterisation of the optimal solution also requires the adjoint (costate) equation for \(\lambda(x,t)\), that is, the Lagrange multipliers corresponding to the constraints of the minimisation of the cost in the equation (26):
By substituting the optimal control equation (28) back into the swarm dynamics equation (25), the total velocity field \(v(x,t)\) of the swarm can be expressed as \[v(x,t) = -A(x) - \sigma^2 \frac{\nabla \rho(x,t)}{\rho(x,t)} - B(x)B(x)^T \nabla V(x,t). \]
The paper of (Zheng et al., 2022) also leverages the concept of Input-to-State Stability (ISS) to show that the perturbed closed-loop system is stable.
Moreover, (Ornia et al., 2022) explores how MFT can be used to reformulate a stochastic multi-agent foraging problem into a deterministic autonomous system, enabling the analysis of limit behaviours and optimality guarantees. The individual agent state evolution (position \(X_t\)) in the multi-agent system is described by the SDE
Here, \(v_c\) is a constant velocity, \(\beta_d(C(X_t))\) represents a density-dependent velocity, \(\gamma_c C(X_t)\) is a consumption or decay term, and \(W_t\) is a Wiener process with a diffusion coefficient \(\sigma\). The pheromone concentration \(C(x)\) at a location \(x \in {T}^d\) is a function of the positions of the agents. In the discrete \(N\)-agent system, it is defined as \[C(x) = \frac{1}{N}\sum_{i=1}^{N} \frac{1}{\epsilon^d} \chi_{B_\epsilon(x_i)}(x), \] where \(\chi_{B_\epsilon(x_i)}(x)\) is the characteristic function for a ball of radius \(\epsilon\) centred at \(x_i\). The macroscopic behaviour of the agent swarm is described by the equation (31) below for the agent density \(m(t,x)\):
Here, \(\alpha\) is the deposition rate of pheromone by agents, and \(\lambda\) is the decay rate of pheromone.
The deterministic part of the velocity field \(V(C(x))\) for the agents which is influenced by the pheromone concentration \(C(x)\), derived from the SDE in the equation (30), is expressed by
In essence, the equation (32) models the agents’ velocity field as a combination of a constant base velocity, a velocity component that responds positively to pheromone concentration, and another component that acts as a decay or consumption factor related to the pheromone concentration. The paper then proceeds to analyse the steady-state behaviour of the system. In the steady state, the time derivatives are zero, leading to a coupled system for the stationary density \(m(x)\) and concentration \(C(x)\):
From the equation (34), a direct relationship between the steady-state density and concentration is found: \(m(x) = \frac{\lambda}{\alpha} C(x).\) Substituting this relation into equation (33) yields a single differential equation for the steady-state concentration \(C(x)\):
The implementation of MFT in robotic swarms relies not on a singular model, but rather on a diverse range of mathematical tools selected according to the task’s intrinsic characteristics. Such adaptability is essential, as it manages the otherwise insurmountable complexity of multi-agent systems by substituting pairwise interactions with the effect of an anonymous mean field exerted on a representative agent. In continuous spatial settings, this abstraction manifests as a PDE, and the careful selection of the governing PDE is the initial decisive step toward effective control. In scenarios where agent motion is predominantly deterministic, such as formation flight, the continuity equation, an example is equation (3), is identified as the appropriate mechanism, determining the evolution of the swarm’s numerical density \(\rho(x,t)\).
By contrast, in environments characterised by stochasticity or inherent uncertainty, such as exploratory missions, the equation (31) becomes essential, characterising the evolution of the probability density \(w(x,t)\) (Carmona et al., 2018) of an agent presence in a given state.
This distinction is not merely descriptive; it underpins the development of advanced control strategies. The resulting PDE, in whatever form it may take, serves as the basis for determining an optimal trajectory through adjoint-based optimisation or ensuring robustness against real-world disturbances via Input-to-State Stability (ISS) analysis.
Hence, the ability to select the appropriate mean-field model enables MFT to not only represent, but also regulate a broad spectrum of collective behaviours in a scalable and reliable manner.
Finite-dimensional mean-field approximations
In (Elamvazhuthi et al., 2019), the authors propose a dual-faceted modelling approach wherein finite-dimensional systems are employed to capture low-order statistics, and infinite-dimensional limits are utilised to ensure precision for extensive populations. Both continuous-time and discrete-time frameworks are considered simultaneously, resulting in a well-structured hierarchy of representations.
Continuous-time
The macroscopic evolution of an ensemble of agents, also known as Kolmogorov forward equation, is encoded in the probability-mass vector \(x(t)\in\mathcal P(\mathcal V)\), whose components obey:
Here, \(\dot x(t)\) denotes the temporal rate of change of the probabilities; each entry \(x(t)\) is the distribution of the random variables and \(\mathbb P(X_i(t) = v )=x_v(t))\) is the likelihood that an arbitrary agent occupies state \(v\) at time \(t\) that determines the time evolution of the probability densities.
The scalar control parameter \(u_e(t)\) represents the transition rate along the directed edge \(e\), while the matrix \(B_e\in\mathbb R^{M\times M}\) encodes the net effect of that transition on the probability vector: its entries satisfy \(B_e^{\,ij}=-1\) when \(i=j\) is the source vertex of \(e\), \(B_e^{\,ij}=+1\) when \(i\) is the target vertex and \(j\) the source, and vanish otherwise. Consider an example with a three-state Markov chain. In this case, this construction is given by
(Elamvazhuthi et al., 2019) rigorously justify the deterministic approximation by establishing the mean-field limit when approximating large finite-dimensional systems with simpler infinite-dimensional models. Specifically, when individual transition rates adopt the form
Generalising the feedback control law, (Bensoussan et al., 2018) proposes an arbitrary distribution stabilisation approach through \(u_e(t) = f_e(x_{S(e)}(t))\), where \(f_e\) is any continuously differentiable, non-decreasing mapping satisfying \(f_e(y)=0\) if and only if \(y = x_{S(e)}^d\). This construction guarantees global asymptotic convergence to the desired macroscopic profile.
To encompass mass-action kinetics, the mean-field representation extends naturally to chemical reaction networks through
The mathematical structure of the equation (41) mirrors that of the population dynamics system, both sharing the fundamental form \(\dot{x}(t) = \sum [\text{rate functions} \times \text{state-dependent terms}]\). This structural similarity underscores the unifying power of the mean-field perspective, demonstrating how this theoretical framework provides a common mathematical foundation for describing collective behaviour across diverse physical, biological and chemical systems.
Discrete-time
When agent states evolve according to a Markov chain, the mean-field evolution is captured by the difference equation (42) below, proposed by Bensoussan et al. (2018):
In summary, discrete-time systems exhibit significant advantages over their continuous-time counterparts within the realm of contemporary technological applications. These advantages are primarily attributed to their exceptional computational efficiency, inherent numerical robustness, and fundamental alignment with modern digital architectures.
While continuous systems require complex and potentially unstable numerical integration, discrete models function based on straightforward algebraic principles, thereby offering enhanced stability and predictability.
Although constraints such as aliasing necessitate meticulous design consideration, they are both well-understood and quantifiable. Consequently, the adoption of discrete-time frameworks signifies a pivotal and irreversible paradigm shift, establishing them as the essential cornerstone for future advancements in engineering and the applied sciences.
Mean-field models from statistical physics
The seminal work by Thouless et al. (1977) rectified inconsistencies in the Sherrington-Kirkpatrick “spin glass” model at low temperatures by developing an MFT that considers both the average agent state and its fluctuations, resulting in a thermodynamically stable description. For robotic swarms, this implies the need for mean-field models that capture the distribution and heterogeneity of agent states, not just their average behaviour.
Bonnet (2019) presents a rigorous mathematical framework for the optimal control of multi-agent systems in the mean-field limit, employing optimal transport theory. The central innovation lies in generalising Pontryagin’s Maximum Principle to the Wasserstein space, providing the necessary conditions to determine controls that optimise the behaviour of an agent density.
The work by Bonnet (2019) establishes the conditions for the spatial regularity of optimal controls, a crucial step to ensure the applicability of MFC to systems with a finite number of agents (microscopic). It also explores control strategies to guarantee alignment (flocking) in cooperative systems, even in the face of communication failures, which has direct application in developing robust algorithms for robotic swarms. That work develops a complex mathematical structure, with its most representative formulae defining the models at different scales (micro and macro) and the optimal control problem.
The microscopic model (\(N\)-agents) is described by the controlled \(N\)-agent dynamics, which expresses how the dynamics of each agent \(i\) are influenced by the other agents and by an individual control \(u_i(t)\):
where \(x_i\) is the state of an agent \(i\), \(x(t)\) is the state vector of all \(N\) agents, \(v^N\) is the interaction velocity field, and \(u_i(t)\) is the applied control.
The Macroscopic Model (Mean-Field Limit) is given by the controlled continuity equation (44) below, which describes the evolution of the probability density \(\mu(t)\) of finding an agent in a given state, in the limit as \(N \rightarrow \infty\):
The Optimal Control Problem in Wasserstein Space is formulated as the General Optimisation Problem (GOP), seeking to minimise a cost subject to the continuity equation (45) below:
The optimality condition (Pontryagin’s maximum principle) is expressed by the Forward-Backward Hamiltonian System. This central result of Bonnet (2019) generalises Pontryagin’s maximum principle, stating that an optimal trajectory and control (\(\mu^*(t)\) and \(u^*(t)\)) must satisfy a system of equations for the state and co-state, along with a maximisation condition.
Bonnet (2019) Research in robotic swarms has been extended to encompass advanced algorithmic methodologies which are aimed at optimising planning and decision making within multi-agent systems. The experimental validation of mean-field theory in robotic swarms has been demonstrated across a variety of collective tasks.
Through the careful comparison of quantitative metrics, such as cost and efficiency rates, spatial distributions, and task performance, researchers have shown that mean-field models can accurately predict the emergent behaviour of real-world robotic swarms. For a comprehensive comparison of the models, the Table 2 has been created within this systematic literature review, structured according to the following key criteria.
| Feature | HJB/FP Models (Mathematical Components) | Physics-Informed Emergent Models (e.g., Spin/Vicsek/Kuramoto) | MFC/MFG Frameworks (MFT Control/Equilibrium) |
|---|---|---|---|
| Primary Objective | HJB: Find the optimal control policy \(u^*\) for individual agents (stochastic optimisation). FP: Model and control the evolution of the probability density \(\rho(x,t)\). | Describe and explain emergent collective behaviours based on local rules; analyse phase transitions and self-organisation. | MFC: Provide an optimal control law for the density distribution \(\rho\) (coupled global optimisation). MFG: Find the Nash Equilibrium (optimal individual strategies \(\leftrightarrow\) consistent population density). |
| Modelling Scale | HJB: Microscopic (focus on individual path \(x(t)\)). FP: Macroscopic (focus on density \(\rho(x,t)\)). | Microscopically defined (local interactions) with macroscopically emergent outcomes (e.g., global polarisation). | Intrinsically macroscopic (treats the swarm as a continuous density in the limit \(N\to\infty\)). |
| Nature of Solution | HJB: A “value function” \(V(x,t)\) (optimal cost). FP: A probability distribution function \(p(x,t)\). | An order parameter characterising the system’s phase state (e.g., polarisation, average magnetisation). | MFC: Optimal control law \(u^*\) for the collective system. MFG: The coupled, self-consistent pair \((V^*, \rho^*)\) (solutions to the coupled HJB and FP PDEs). |
| Computational Complexity | HJB: High; suffers from the “curse of dimensionality” for finite \(N\). FP: Requires solving PDEs (scales better with \(N\)). | Direct (agent-based) simulations are scalable \(O(N)\); equilibrium analysis can be NP-hard. | Requires solving coupled, non-linear PDEs (HJB and FP) in the continuous space; high analytical complexity. |
| Information Requirements | HJB: Precise dynamical model, cost functional. FP: Interaction rules and diffusion/noise function. | Only requires the definition of local interaction rules and a noise/temperature parameter. | Requires individual agent cost functions and interaction kernels, assuming agents respond to the mean field \(\rho\). |
| Control Type | HJB: Centralised (in finite systems). FP: Density control, decentralisable. | Inherently decentralised (based on local neighbourhood interactions). | MFC: Global optimisation (centralised). MFG: Decentralised strategy (agents react to the mean field \(\rho\), not specific individuals). |
| Typical Application | HJB: Optimal trajectory planning for small groups. FP: Area coverage, density management, congestion control. | Collective decision-making (quorum sensing), aggregation, pattern formation, stability analysis. | Large-scale congestion control, resource optimisation (MFC), strategic and competitive interactions (MFG). |
The Table 3 serves to systematically delineate the core conceptual and mathematical differences between the two primary MFT frameworks: Mean Field Control (MFC) and Mean Field Game (MFG), it maps the distinct primary focus and solution processes of each, enabling a clear taxonomic classification of the extant literature on robotic swarms and providing the necessary theoretical foundation for discussing how each framework contributes to the estimation of performance metrics in target acquisition scenarios.
| Framework | Primary Focus | Optimisation Eq. (HJB-like) | Dynamics Eq. (FP-like) | Solution Process |
|---|---|---|---|---|
| MFC (Control) | Global cost minimisation \(J(\rho)\) | Variational (optimality condition) | Controlled FP | Find the control law \(u^*\) that minimises \(J(\rho^*)\), where \(\rho^*\) is the resulting distribution under that law. |
| MFG (Game) | Nash equilibrium | HJB (Hamilton–Jacobi–Bellman) | FP (Fokker–Planck) | Fixed-point solution: find the pair \((V^*, \rho^*)\) that satisfies both individual optimality and population consistency simultaneously. |
Conclusions
Drawing together the ideas above, the applicability of MFT as a foundational methodology for the scalable estimation of performance in robotic swarms was investigated. This review of the literature demonstrated that MFT provides a robust and promising mathematical framework for addressing the challenges of high-dimensionality and complex interactions in multi-agent systems.
A substantial increase in scholarly investigation was noted in this area, particularly from 2010 onwards, indicating MFT’s growing significance and potential. Among the various applications explored, those related to optimal control, coordination, and emergent behaviour in robotic systems were particularly prominent. A recurring conclusion is MFT’s capability to simplify the analysis of complex systems, facilitating the derivation of collective behaviours from individual interaction rules.
However, interpreting MFT outcomes in robotic swarms needs a meticulous assessment of the extent to which the predicted collective behaviour corresponds with empirical observations. Discrepancies may arise from the disparity between idealised models and the intricacy of real-world systems.
While MFT’s principal advantage resides in its scalability (where the computational effort per robot ideally remains unaffected by swarm size), several inherent limitations are encountered by the method. These include: the assumption of a large number of agents \(N\), which results in a diminished precision for small swarms due to significant stochastic fluctuations; the presumption of agent homogeneity in classical models, contrasting with observed variations in reality; the nature of the mean approximation, which tends to smooth out critical fluctuations and the effect of outliers; the mathematical complexity intrinsic to the MFT model, often manifesting as a non-linear PDE-based or an MFG system, posing difficulties for real-time resolution; and the incongruity between theoretical global mean-field interactions and practical local interactions, necessitating theoretical justification for how locality impacts model validity.
Despite these challenges, the strengths identified, such as the ability to simulate large-scale systems with computational efficiency, are indisputable. The insights acquired will subsequently be used from this review to estimate key performance metrics including total simulation time, target arrival time, and target departure time, for a swarm algorithm based on potential fields as those developed by (Passos et al., 2023). These estimates will be validated against outcomes from direct simulations, aiming to enhance the existing state-of-the-art and advance the practical application of MFT in robotic systems.
Acknowledgments
The authors thank the Federal University of Recôncavo da Bahia (UFRB), FAPESB and CNPq for their support and studentships.
Author contributions
J. V. N. de S. Nunes contributed to the Conceptualisation, Investigation, Formal Analysis, Visualisation, Writing and Original Draft. Y. T. dos Passos contributed to the Conceptualisation, Methodology, Supervision, Writing and Revision.
Conflicts of interest
The authors declare that no competing financial or commercial interests exist in relation to the work described in this manuscript.
References
Auricchio, F., Toscani, G., & Zanella, M. (2022). Fokker-Planck Modeling of Many-Agent Systems in Swarm Manufacturing: Asymptotic Analysis and Numerical Results. ResearchGate. https://doi.org/10.13140/RG.2.2.25950.72006
Bensoussan, A., Huang, T., & Laurière, M. (2018). Mean field control and mean field game models with several populations. arXiv, 2, 1–32. https://doi.org/10.48550/arXiv.1810.00783
Bensoussan, A., Frehse, J., & Yam, P. (2013). Mean Field Games and Mean Field Type Control Theory. Springer. https://doi.org/10.1007/978-1-4614-8508-7
Bonabeau, E., Dorigo, M., & Theraulaz, G. (1999). Swarm Intelligence: From Natural to Artificial Systems. Oxford University Press. https://doi.org/10.1086/393972
Bonnet, B. (2019). Optimal Control in Wasserstein Spaces. (Doctoral thesis, Aix-Marseille Université). https://hal.science/tel-02361353
Burger, M., Di Francesco, M., Markowich, P. A., & Wolfram, M.-T. (2014). Mean Field Games with Nonlinear Mobilities in Pedestrian Dynamics. Discrete and Continuous Dynamical Systems – B, 19(5), 1311–1333. https://doi.org/10.3934/dcdsb.2014.19.1311
Carmona, R., & Delarue, F. (2018). Probabilistic Theory of Mean Field Games with Applications I–II. Springer. https://doi.org/10.1007/978-3-319-58920-6
Couzin, I. D., Krause, J., James, R., Ruxton, G. D., & Franks, N. R. (2002). Collective Memory and Spatial Sorting in Animal Groups. Journal of Theoretical Biology, 218(1), 1–35. https://doi.org/10.1006/jtbi.2002.3065
Cui, K., Li, M., Fabian, C., & Koeppl, H. (2023). Scalable task-driven robotic swarm control via collision avoidance and learning mean-field control. In IEEE International Conference on Robotics and Automation (ICRA). https://doi.org/10.1109/ICRA48891.2023.10161498
Dogbé, C. (2010). Modeling Crowd Dynamics by the Mean-Field Limit Approach. Mathematical and Computer Modelling, 52(9), 1506–1520. https://doi.org/10.1016/j.mcm.2010.06.012
Elamvazhuthi, K., & Berman, S. (2019). Mean-Field Models in Swarm Robotics: A Survey. Bioinspiration & Biomimetics, 15(1), 015001. https://doi.org/10.1088/1748-3190/ab49a4
Elamvazhuthi, K., Kakish, Z., Shirsat, A., & Berman, S. (2021). Controllability and Stabilization for Herding a Robotic Swarm Using a Leader: A Mean-Field Approach. IEEE Transactions on Robotics, 37(2), 418–432. https://doi.org/10.1109/TRO.2020.3031237
Estrada-Rodriguez, G., & Gimperlein, H. (2020). Interacting Particles with Lévy Strategies: Limits of Transport Equations for Swarm Robotic Systems. SIAM Journal on Applied Mathematics, 80(1), 476–498. https://doi.org/10.1137/18M1205327
Hazra, G., Nandy, D., Kitchatinov, L., & Choudhuri, A. R. (2023). Mean Field Models of Flux Transport Dynamo and Meridional Circulation in the Sun and Stars. Space Science Reviews, 219(5). https://doi.org/10.1007/s11214-023-00982-y
Huang, H., & Liu, J.-G. (2020). On the Mean-Field Limit for the Vlasov–Poisson–Fokker–Planck System. Journal of Statistical Physics, 25(12), 1915–1965. https://doi.org/10.1007/s10955-020-02648-3
Huang, M., Malhamé, R., & Caines, P. (2006). Large population stochastic dynamic games: Closed-loop McKean–Vlasov systems and the Nash certainty equivalence principle. Communications in Information and Systems, 6(3), 221–252. https://doi.org/10.4310/CIS.2006.v6.n3.a5
Karafyllis, I., Jiang, Z.-P., & Athanasiou, G. (2010). Nash equilibrium and robust stability in dynamic games: A small-gain perspective. Computers & Mathematics with Applications, 60(11), 2936–2952. https://doi.org/10.1016/j.camwa.2010.09.054
Lasry, J.-M., & Lions, P.-L. (2007). Mean Field Games. Japanese Journal of Mathematics, 2, 229–260. https://doi.org/10.1007/s11537-007-0657-8
Ménec, S. L. (2024). Swarm guidance based on mean field game concepts. International Game Theory Review, 26(2), 244008. https://doi.org/10.1142/S0219198924400085
Ornia, D. J., Zufiria, P. J., & Mazo, M. Jr. (2022). Mean Field Behavior of Collaborative Multiagent Foragers. IEEE Transactions on Robotics, 38(4), 2151–2165. https://doi.org/10.1109/TRO.2022.3152691
Passos, Y. T., Duquesne, X., & Marcolino, L. S. (2023). Congestion control algorithms for robotic swarms with a common target based on the throughput of the target area. Robotics and Autonomous Systems, 159, 104284. https://doi.org/10.1016/j.robot.2022.104284
Thouless, D. J., Anderson, P. W., & Palmer, R. G. (1977). Solution of “Solvable Model of a Spin Glass”. The Philosophical Magazine, 35(3), 593–601. https://doi.org/10.1080/14786437708235992
Vicsek, T., Czirók, A., Ben-Jacob, E., Cohen, I., & Shochet, O. (1995). Novel Type of Phase Transition in a System of Self-Driven Particles. Physical Review Letters, 75(6), 1226–1229. https://doi.org/10.1103/PhysRevLett.75.1226
Yang, Y., Luo, R., Li, M., Zhou, M., Zhang, W., & Wang, J. (2018). Mean Field Multi-Agent Reinforcement Learning. In J. Dy & A. Krause (Eds.), Proceedings of the 35th International Conference on Machine Learning (Vol. 80, pp. 5571–5580).
Zheng, T., Han, Q., & Lin, H. (2022). Transporting Robotic Swarms via Mean-Field Feedback Control. IEEE Transactions on Automatic Control, 67(8), 4170–4177. https://doi.org/10.1109/TAC.2021.3108672

