Second order local minimal-time Mean Field Games
Abstract.
The paper considers a forward-backward system of parabolic PDEs arising in a Mean Field Game (MFG) model where every agent controls the drift of a trajectory subject to Brownian diffusion, trying to escape a given bounded domain in minimal expected time. Agents are constrained by a bound on the drift depending on the density of other agents at their location. Existence for a finite time horizon is proven via a fixed point argument, but the natural setting for this problem is in infinite time horizon. Estimates are needed to treat the limit , and the asymptotic behavior of the solution obtained in this way is also studied. This passes through classical parabolic arguments and specific computations for MFGs. Both the Fokker–Planck equation on the density of agents and the Hamilton–Jacobi–Bellman equation on the value function display Dirichlet boundary conditions as a consequence of the fact that agents stop as soon as they reach . The initial datum for the density is given, and the long-time limit of the value function is characterized as the solution of a stationary problem.
Key words and phrases:
Mean Field Games, congestion games, parabolic PDEs, MFG system, existence of solutions, asymptotic behavior2020 Mathematics Subject Classification:
35Q89, 35K40, 35B40, 35A01, 35D301. Introduction
Introduced around 2006 by Jean-Michel Lasry and Pierre-Louis Lions [21, 22, 23] and at the same time by Peter Caines, Minyi Huang, and Roland Malhamé [14, 15, 16], the theory of Mean Field Games (MFGs, for short) describes the interaction of a continuum of players, assumed to be rational, indistinguishable, and negligible, when each one tries to solve a dynamical control problem influenced only by the average behavior of the other players (through a mean-field type interaction, using the physicists’ terminology). The Nash equilibrium in these continuous games is described by a system of PDEs: a Hamilton–Jacobi–Bellman equation for the value function of the control problem of each player, where the distribution (density) of the players appears, coupled with a continuity equation describing the evolution of such a density, where the velocity field is the optimal one in order to solve the control problem, and is therefore related to the gradient of the value function. This system is typically forward-backward in nature: the density evolves forward in time starting from a given initial datum, and the value function backward in time, according to Bellman’s dynamical programming principle, and its final value at a given time horizon is usually known.
The literature about MFG theory is quickly growing and many references are available. The 6-year course given by P.-L. Lions at Collège de France, for which video-recording is available in French [26], explains well the birth of the theory, but the reader can also refer to the lecture notes by P. Cardaliaguet [6], based on the same course.
In most of the MFG models studied so far the agents consider a fixed time interval and optimize a trajectory (where is the state space) trying to minimize a cost of the form where denotes the distribution of players at time . The function is typically increasing in and, in some sense, in . This means that high velocities are costly, and passing through areas where the population is strongly concentrated is also costly. Some MFGs, called MFGs of congestion (see, for instance, [1]), consider costs which include a product of the form (for some exponents ), which means that high velocities are costly, and that they are even more costly in the presence of high concentrations. These models present harder mathematical difficulties compared to those where the cost is decomposed into . Indeed, in many cases the latter MFG admits a variational formulation: equilibria can be found by minimizing a global energy among all possible evolutions (hence, they are potential games). This allows to prove the existence of the equilibrium via semicontinuity methods, and we refer to [4] and [30] for a detailed discussion of this branch of MFG theory.
When the MFG has no variational interpretation, then the existence of a solution is usually obtained via fixed-point theorems, but these theorems require much more regularity. Roughly speaking, given an evolution one computes the corresponding value function as a solution to a Hamilton–Jacobi–Bellman equation and, given , one computes a new density evolution by following an evolution equation. We need existence, uniqueness, and stability results for these equations in order to find a fixed point . This usually requires regularity of the velocity field , which is difficult to prove, and can be essentially only obtained in two different frameworks: either the dependence of the cost functions on the distribution is highly regularizing (which usually means that it is non-local, and passes through averaged quantities such as convolutions ), or diffusion of the agents is taken into account, transforming the optimal control problem into a stochastic one. In this latter case, agents minimize where the process follows and represents a standard Brownian motion.
In [27], the second and third authors of the present paper introduced a different class of models, called minimal-time MFGs. The main difference is that instead of considering a cost for the players penalizing both the velocity and the density, and minimizing the integral of such a cost on a fixed time interval , the dynamics is subject to a constraint where the maximal velocity of the agents cannot exceed a quantity depending on the density , and the goal of each agent is to arrive to a given target as soon as possible. In the typical situation, the target of the agents is the boundary of the domain where the evolution occurs. This can model, for instance, an evacuation phenomenon in crowd motion. The system that one obtains is the following
(1.1) |
where the function denotes the maximal speed that agents can have at point at time , i.e., the dynamics is constrained to satisfy . Ideally, one would like to choose to be a non-increasing function of the density itself, such as . This choice is what is done in the well-known Hughes’ model for crowd motion [17, 18]. Indeed, this model is very similar to Hughes’, which also considers agents who aim at leaving in minimal time a bounded domain under a congestion-dependent constraint on their speeds.
The main difference between the model in [27] (from which the present paper stems) and Hughes’ is that, in the latter, at each time, an agent moves in the optimal direction to the boundary assuming that the distribution of agents remains constant, whereas in [27] and here agents take into account the future evolution of the distribution of agents in the computation of their optimal trajectories. This accounts for the time derivative in the Hamilton–Jacobi–Bellman equation from (1.1), which is the main difference between (1.1) and the equations describing the motion of agents in Hughes’ model and stands for the anticipation of future behavior of other agents.
Another crucial (and disappointing) similarity between the above MFG system and Hughes’ model is the fact that general mathematical results do not exist in the case and more generally in the local case (except few results in the Hughes case in 1D). Indeed, the lack of regularity makes the model too hard to study, and the MFG case is not variational.
In some sense the closest MFG model to this one is the one with multiplicative costs in [1] (MFG with congestion). Indeed, an constraint can be seen as a limit as of an integral penalization
Note that the boundaries of the time interval have been omitted on purpose from the above integral, since the model in [1] is set on a fixed time horizon but this is not part of our setting. For MFG with congestion, [1] presents not only existence but also uniqueness results, under the assumption that the exponents appearing in the running cost satisfy a certain inequality. Unfortunately this inequality is never satisfied in the limit as above, and it is not surprising that in our work we are not able to establish uniqueness results for our MFG system. Additional results for MFG with congestion were presented, for instance, in [12], under a smallness assumption on the time horizon, but this assumption cannot be made here, as the model is exactly meant to consider the case where a time horizon is not fixed.
Because of these difficulties, [27] studied the case of a non-local dependence of w.r.t. (say, , for a non-increasing function and a positive convolution kernel ), and proved existence of an equilibrium, characterized it as a solution of a non-local MFG system, and analyzed some examples, including numerical simulations. Instead, in the present paper we want to study the local case with diffusion.
This means that we will consider a local dependence , and each agent solves a stochastic control problem
where denotes a standard Brownian motion and the Brownian motions for all players are assumed to be mutually independent. Defining the corresponding value function , from classical results on stochastic optimal control (see [10, Chapter IV]), under suitable assumptions, the optimal control is given in feedback form by
(a definition which has to be carefully adapted to the case ); moreover, the value function solves the Hamilton–Jacobi–Bellman equation
for . Hence, we know the drift of the optimal stochastic processes followed by each agent, and this allows to write the Fokker–Planck equation solved by the law of this process. Putting together all this information, we obtain the following MFG system
(1.2) |
where is an open and bounded set, whose boundary will be supposed to be of class in this paper, is a fixed constant, , and is the initial density. The Dirichlet condition on comes as usual from the fact that, for agents who are already on the boundary, the remaining time to reach it is zero, and the Dirichlet condition on comes from the fact that we stop the evolution of a particle as soon as it touches the boundary (absorbing boundary conditions).
A crucial difference with the previous paper [27] concerns the time horizon. If we suppose that is bounded from below in the model without diffusion, it is not difficult to see that all agents will have left the domain after a common finite time, so that the final value of is not really relevant, and the problem can be studied on a finite interval . This is not the case when there is diffusion, as a density following a Fokker–Planck equation with a bounded drift cannot fully vanish in finite time. As a consequence, the model should be studied on the unbounded interval . For every time there is still mass everywhere, but this mass decreases to as , which suggests that the value function should converge to a function, that we call , which is the value function for the corresponding control problem with no mass, i.e. when . Since in this control problem is independent of time, is a function of only and solves a stationary Hamilton–Jacobi–Bellman equation which takes the form of an elliptic PDE
with Dirichlet boundary conditions on . It is then reasonable to investigate whether solutions of the above system satisfy further and as .
In order to study the above system, we will first study an artificial finite-horizon setting, where we stop the game at time , choose a penalization with on , and look at the stochastic optimal control problem
This gives rise to the MFG system
(1.3) |
which corresponds to (1.2) with the unbounded time interval replaced by and the additional final condition . We will prove the existence of a solution of the system for finite (note that for this system, as well as for its infinite-horizon counterpart, we are not able to prove uniqueness), and then consider the limit as . In order to guarantee suitable bounds, we just need to choose a sequence of final data , possibly depending on , which is uniformly bounded. We will then get at the limit a solution of the limit system which automatically satisfies (in the sense of uniform convergence) and (this convergence being both uniform and strong in ).
The paper is organized as follows. After this introduction, Section 2 presents the tools that we need to study the two separate equations appearing in System (1.3) on a finite horizon, which come from the classical theory of parabolic equations. Section 3 is devoted to the existence of solutions of (1.3). After providing a precise definition of solution of (1.3) taking care of the case , we use the estimates of Section 2 to prove existence via a fixed-point argument based on Kakutani’s theorem. Section 4 concerns the limit . In this section, some estimates of Section 2 need to be made more precise, in order to see how constants depend on the time horizon . In this way we are able to prove existence of a limit of the solutions of (1.3) as the time horizon tends to and that this limit solves the limit system (1.2). Then we consider the asymptotic behavior of a solution of (1.2) as , proving first in and, thanks to a parabolic regularization argument, also in . To prove convergence in , which is true for general Fokker–Planck systems under very mild assumptions, we exploit the MFG nature of the system, i.e. the coupling between the two equations, which also provides exponential decrease. We then consider the limit in time of , and prove that any bounded solution of this equation, once we know , can only converge as to the stationary function . This convergence is a priori very weak, but we are able to improve it into , and to prove that the uniform convergences of both and occur exponentially fast. The paper is then completed by an appendix, which details some global estimates for a large class of parabolic equations, including the estimates that we use to prove uniform convergence in time of and to and respectively. These estimates are not surprising and not difficult to prove, using standard Moser iterations, but are not easy to find in the literature under the sole assumption of boundedness of the drift term in the divergence. The computations and the results are essentially the same as in the appendix of [7], but the boundary conditions are different.
2. Preliminary results
This section presents some preliminary results on Fokker–Planck and Hamilton–Jacobi–Bellman equations which are useful for the analysis of the Mean Field Game systems (1.2) and (1.3). We recall that, in the whole paper, denotes an open and bounded set whose boundary is assumed to be . Even though some of the results presented in this preliminary section also hold without the smoothness assumption on (such as existence and uniqueness results for both Fokker–Planck and Hamilton–Jacobi–Bellman equations in Propositions 2.2 and 2.5), this assumption is first used to obtain higher regularity of solutions of Hamilton–Jacobi–Bellman equations in Proposition 2.5 and is required for almost all of the subsequent results, including in particular our main results in Sections 3 and 4, as a consequence of the need of higher regularity of .
2.1. Fokker–Planck equation
We recall some results on the Fokker–Planck equation on a bounded domain in finite time horizon ,
(2.1) |
where is a given velocity field. We will only focus on the case where is bounded, an assumption which is satisfied in the cases of interest for this paper and which strongly simplifies the analysis. The results presented in this short section are a mixture of classical results (for which we mainly refer to [8, Section 7.1] or [19, Chapter III]), recent results obtained by Porretta in [28], and extra computations which are not original but are difficult to find in the literature, which we present in the Appendix.
Definition 2.1.
Let , , and . We say that is a weak solution of (2.1) if, for every such that and , one has
(2.2) |
We observe that, whenever equality (2.2) holds for functions, and if we have further that for some with , then we also have
(2.3) |
for every and, by density, for every such that . Of course it is well-known that, in case is more regular, other test functions can also be accepted, and that if then the equation is satisfied in a classical sense.
We now state a proposition summarizing all the main results that we will use.
Proposition 2.2.
Let , , and be a given non-negative initial datum. Then (2.1) admits a unique weak solution . In addition, we have and with , as well as and for all and for all , and the norms of and in the above spaces are bounded by quantities only depending on . Moreover, for every , we also have and .
Of course we do not provide a full proof of the above results, but we explain below how to deduce the different parts of the statement from the most well-known literature and the relevant references.
Proof.
The definition of the solution is exactly the one used in [28], where the key assumption is . In our case, where is bounded, this assumption is satisfied as soon as . One of the main results of [28] is exactly the uniqueness of the solution in this class, and this can be applied to the present setting. The same paper also guarantees the estimates , , , and the bound.
Existence is not included in [28] but in the particular case it is easy to obtain by regularization and compactness. Indeed, one can apply the classical theory of [19, Chapter III] to an approximated initial datum, and obtain a sequence of solutions: the bounds of [28], which only depend on the initial norm in this setting, allow to obtain the compactness we need to pass the PDE to the limit. Note that this argument is specific to the case since, otherwise, we would need to control the norm of , which is non-trivial.
By approximating and by smooth functions and with , the corresponding solution of (2.1) satisfies thanks to the classical maximum principle, and then this property passes to the limit and also applies to the unique weak solution corresponding to the original and .
The local bound can be obtained thanks to the Appendix of the present paper (even if we stress that similar computations are nowadays standard). For simplicity, the bound is presented under the assumption , , and not . Yet, the time-space summability already stated in the claim allows to deduce for a.e. , and if we choose we obtain the desired bound. Once we know that is locally (in time) (in space), it is also locally (in time) (in space), and hence the classical theory of [19, Chapter III] provides the last estimates of the statement. ∎
2.2. Hamilton–Jacobi–Bellman equation
We consider the non-linear Hamilton–Jacobi–Bellman equation in a finite time horizon
(2.4) |
where is a given function.
Definition 2.3.
Let , , and . We say that is a weak solution of (2.4) if, for every such that and , one has
(2.5) |
As we did after Definition 2.1, we observe that, if (2.5) holds for every as before, and if we assume further that for some with , then we also have
(2.6) |
for every and, by density, for every such that .
Remark 2.4.
The next proposition gathers the main results on solutions of (2.4) that will be needed in the paper.
Proposition 2.5.
Let , , and . Then (2.4) admits a unique weak solution . In addition, we have , and the norms of in and are bounded by quantities depending only on , , , , an upper bound on , and .
Moreover, if a.e. in , then the unique solution also satisfies a.e. in . If , , and a.e. in , then there exists a constant depending on , , and such that a.e. on .
Finally, if , then , , and the norms of in these spaces are bounded by quantities depending only on , , , , an upper bound on , and .
The results stated in Proposition 2.5 are classical and follow from more general results for nonlinear pseudo-monotone operators. Similarly to Proposition 2.2, we explain below how they can be retrieved from the relevant literature.
Proof.
Existence of a weak solution for follows from [29, Theorem 2.1] and the corresponding bounds on the norms of are a consequence of [29, Lemma 4.1], whereas uniqueness follows from [9, Theorem 2.4].
The positivity of when is classical for smooth solutions and can be obtained by an easy application of the maximum principle for parabolic equations. For solutions of HJB obtained as value functions of a stochastic control problem, the result is also straightforward, as the quantity which is minimized is positive. In our context of weak solutions, it can be deduced by applying, for instance, [2, Theorem 1] to , after changing time orientation and paying attention to the observation at the end of the proof (page 98) that the inequality is enough (indeed, the source term in the HJB equation has the good sign to preserve positivity).
The upper bound on under the positivity assumption on and and the fact that can be obtained by applying a parabolic comparison principle (see [24, Theorem 9.1] for the smooth case) to and , where is the solution of the torsion equation in with Dirichlet boundary conditions.
Finally, higher regularity of when can be obtained in a straightforward manner by noticing that , i.e., satisfies a linear backwards heat equation in with source term . The conclusions then follow from classical improved regularity results for heat equations (such as [8, Section 7.1, Theorem 5] and [19, Chapter III, § 6, Equation (6.10) and Theorem 6.1]). ∎
We next state, for future reference, a standard parabolic comparison principle for (2.4) (see, e.g., [9, Corollary 2.2]).
Proposition 2.6.
Let be two solutions of (2.4) with , with final data such that . Then
3. The MFG system with a finite time horizon
We now consider the MFG system with a finite time horizon (1.3). One of the difficulties in the analysis of (1.3) is that the velocity field in the continuity equation depends on , which is defined only when . In order to handle this difficulty, we make use of the following definition of weak solution.
Definition 3.1.
Let , , be continuous and bounded, , and . We say that is a weak solution of (1.3) with initial condition and final condition if there exists such that and a.e. on and such that is a solution of the Fokker–Planck equation (2.1) with initial datum and vector field on in the sense of Definition 2.1, and is a solution of the Hamilton–Jacobi–Bellman equation (2.4) with final datum and in the sense of Definition 2.3 on the same domain.
Remark 3.2.
If is a weak solution of (1.3) and is any function satisfying the properties stated in Definition 3.1, then we have wherever . The introduction of the function in Definition 3.1 has the advantages of providing a meaning to the first equation of (1.3) and handling its velocity field even when , which might a priori happen in a set of positive measure.
The main result of this section is the following.
Theorem 3.3.
Let , , be continuous and bounded, , and . Then there exists a weak solution of (1.3) with initial condition and final condition .
The proof of Theorem 3.3 relies on a fixed-point argument on the velocity field of the Fokker–Planck equation in (1.3). Before turning to the proof, we need some continuity results on solutions of (2.1) with respect to the velocity field and on solutions of (2.4) with respect to the function , which we state and prove now.
Proposition 3.4.
Let and . Given , let be a sequence in such that as . For , let (resp. ) be the unique weak solution of (2.1) in with velocity field (resp. ). Then in as .
Proof.
Since converges weakly- to in , there exists a constant such that for every and thus, by Proposition 2.2, there exists depending only on , , , and such that, for every ,
(3.1) |
It follows from (3.1) and Aubin–Lions Lemma (see, e.g., [32, Corollary 4]) that is relatively compact in . Let be a limit point of and a subsequence of converging to in .
The weak convergence of in together with the strong convergence of in allow to pass to the limit the drift term in the equation and we then easily obtain that is a weak solution of (2.1). By the uniqueness of such solution from Proposition 2.2, one concludes . In particular, is the unique limit point of the relatively compact sequence in , which yields the result. ∎
Proposition 3.5.
Let and . Given , let be a sequence in such that as . For , let (resp. ) be the unique weak solution of (2.4) in with (resp. ). Then in as .
Proof.
Again, there exists a constant such that for every and thus, by Proposition 2.5, there exists depending only on , , , , , and such that
(3.2) |
Hence, by Aubin–Lions Lemma (see, e.g., [32, Corollary 4]), is relatively compact in . Let be a limit point of and be a subsequence of converging to in . By (3.2), we also have .
Now, because of the non-linearity in the equation, we prefer to provide details on how to pass it to the limit. For every and every such that and , one has
Since in and in , one obtains, letting , that
Hence is a weak solution of (2.4) and, by the uniqueness of solutions of (2.4) from Proposition 2.5, one deduces that . Thus is the unique limit point in of the relatively compact sequence , yielding the conclusion. ∎
We now recall the statement of Kakutani’s fixed point theorem (see, e.g., [13, § 7, Theorem 8.6]), which is used in the proof of Theorem 3.3.
Theorem 3.6 (Kakutani’s fixed point theorem).
Let be a compact convex subset of a locally convex topological vector space , and let be a set-valued map in , i.e., associates, with each , a set . Assume that is upper semi-continuous and that, for every , is non-empty, compact, and convex. Then admits a fixed point in , i.e., there exists such that .
We are finally in position to provide the proof of Theorem 3.3.
Proof of Theorem 3.3.
Let be an upper bound on . We endow the space with its weak- topology and consider the ball of radius given by
Note that is clearly convex and, by the Banach–Alaoglu theorem, is a compact subset of .
Let be the function that associates, with each , the unique weak solution of (2.1) with initial condition . Note that, by Proposition 3.4, is continuous with respect to the weak- topology of and the strong topology of . Similarly, we define as the function that associates, with each , the unique weak solution of (2.4) with terminal condition . Proposition 3.5 ensures that is continuous with respect to the weak- topology of and the strong topology of .
We define the set-valued map that, with each , associates the set given by
In order to prove the existence of a weak solution of (1.3), we first prove the existence of a fixed point of the set-valued map , i.e., of a such that . This is done by applying Kakutani’s fixed point theorem to the set-valued map . To do so, we first need to verify some properties of and its graph defined by
Claim 1.
For every , the set is non-empty and convex.
Proof.
It is immediate to verify that is convex. To prove that it is non-empty, let , , and . Then, the function defined for a.e. by
clearly satisfies . ∎
Claim 2.
The graph is a closed subset of .
Proof.
Let be a sequence in converging weakly- in to a point . We want to prove , i.e., .
Define, for , the functions and by and and, similarly, let and . Since is continuous with respect to the weak- topology of and the strong topology of , one deduces in as . Hence, up to extracting subsequences (which we still denote using the same notation for simplicity), one has a.e. in . Since is continuous, we deduce a.e. in , and it follows in . The continuity of with respect to the weak- topology of and the strong topology of implies in as .
From the weak convergence of to , the convexity of the function , and the (strong) convergence of to , the inequality gives at the limit
(3.3) |
Claim 3.
For every , the set is compact.
Proof.
This is a consequence of the fact that is a closed subset of the compact set . ∎
Thanks to Claims 2 and 3, it follows from [3, Proposition 1.4.8] that the set-valued map is upper semi-continuous. Using this fact and Claims 1 and 3, it follows from Kakutani’s fixed point theorem that admits a fixed point . Let and . Using the facts that and are solutions of (2.1) and (2.4), respectively, and that , it is immediate to verify, using Definitions 2.1, 2.3, and 3.1, that is a weak solution of (1.3) with initial condition and final condition , as required. ∎
4. The MFG system with an infinite time horizon
Now that we have established in Section 3 the existence of solutions to the Mean Field Game system (1.3) in a finite time horizon , we consider in this section the Mean Field Game system (1.2) with an infinite time horizon. Let us first provide the definition of a weak solution in this setting.
Definition 4.1.
Let , be continuous and bounded, and . We say that is a weak solution of (1.2) with initial condition if and if there exists such that and a.e. on and such that, for every , is a solution of the Fokker–Planck equation (2.1) with initial datum and vector field on in the sense of Definition 2.1 and is a solution of the Hamilton–Jacobi–Bellman equation (2.4) with in the sense of Definition 2.3 on the same domain111Note that Definition 2.3 requires to fix a final value, and we did not define the notion of solution independently of the final value . This could be formalized as “there exists such that is a solution of (2.4)”. Yet, since the function will be finally continuous as a function valued into , the final datum on will be necessarily given by its own value ..
Notice that, with respect to Definition 3.1, we make the additional requirement that . This is done mainly for three reasons. Firstly, boundedness of the solution of a Hamilton–Jacobi–Bellman equation is a condition usually required in order to ensure that this solution is the value function of an optimal control problem (see, e.g., [5, Theorem 8.1.10] and [10, Chapter II, Corollary 9.1]). Secondly, the strategy we use in this section to prove existence of a solution of (1.2), based on a limit argument from solutions of (1.3) in finite time horizon as , allows us to ensure that the function we construct is indeed bounded. Finally, boundedness of is an important property in order to establish the results on the the asymptotic behavior of solutions to (1.2) provided in Theorem 4.2 and Propositions 4.5 and 4.6.
4.1. Existence of solutions and their asymptotic behavior
From now on, we let denote the solution of the (stationary) Hamilton–Jacobi–Bellman equation
(4.1) |
with Dirichlet boundary conditions on . Existence of such a solution follows from standard results on elliptic equations, and is continuous in the closure of and and positive in (see, e.g., [11, Theorem 15.12], [20, Chapter 4, Section 8], [25]; these results require additional regularity properties on but they can be easily adapted to a boundary using the techniques from [11, Section 15.6] or [20, Chapter 4, pp. 309–310]). Uniqueness of follows also from classical arguments for elliptic equations based on the maximum principle: the difference of two solutions of (4.1) is zero on and satisfies in , and hence the maximum principle from [11, Theorem 10.9] allows one to conclude that in .
The main result of this section is the following.
Theorem 4.2.
Let . Then, there exists at least one solution to the Mean Field Game system with infinite time horizon (1.2).
In addition, any such solution satisfies
and the above convergences hold uniformly.
The sequel of this section is devoted to the proof of Theorem 4.2. Let us start by giving an idea of the proof. First, we will construct solutions to the problem with infinite time horizon as limits of solutions of the problem with finite time horizon by letting go to . Then, to prove the long-time uniform convergence of the solutions, we shall make a crucial use of some regularity results for parabolic equations. More precisely, we will use local maximum principles for Fokker–Planck and for (forward) Hamilton–Jacobi–Bellman equations; roughly speaking, these results state that the norm of solutions of such equations at some time is controlled by some norms of the same solution at some previous time . The results we use are proved in Appendix A, see Proposition A.1 and Corollaries A.2 and A.3.
We start with a lemma that gathers some useful estimates. These estimates have already been discussed in Section 2, but we need now to track possible dependencies of the constant on the time horizon .
Lemma 4.3.
Proof.
Step 1. A preliminary estimate.
Let us start with giving an estimate on the gradient of . First, multiplying by the equation satisfied by and integrating on for a fixed , we find
Therefore, since is bounded, we have
(4.4) |
for some depending on , , , and for every with . Note that, from Proposition 2.5, is bounded in terms of , , and .
Step 2. Bound on .
We define, for ,
We differentiate to obtain
(4.5) |
Using Young’s inequality, we find that there are depending only on , , such that
This implies, for any ,
(4.6) |
We integrate (4.6) for to get
Using (4.4) yields the bound (4.2) for . To get the bound (4.2) for , we use (4.6) with . The result follows, with a constant also depending on .
Step 3. Bound in .
The next lemma shows that the time derivative of is equal to . Differentiating the average value of the value function is a classical computation in Mean Field Game theory. Since here the value function is an exit time, it is expected that it should decrease with rate , and one can guess the result from the fact that the total mass of the agents in this model is not fixed but decreases in time and is equal to .
Lemma 4.4.
Proof.
Let us fix two instants of times , with . On the interval we can use as a test function in (2.3) and in (2.6) since both and are continuous as curves valued in , belong to , and their time-derivatives belong to . We subtract the two equalities that we obtain, which provides
After canceling the terms with and using we are left with
It is then easy to see, by approximation via smooth functions, that for every pair such that and , we have
We are then left with
which is equivalent to the claim. ∎
We are now in position to prove Theorem 4.2.
Proof of Theorem 4.2.
Let be fixed.
Step 1. Existence.
For , we let denote a solution of (1.3) with , with initial datum for and with final datum for , where is any family of non-negative functions, bounded in .
Recall that, by Proposition 2.5, is bounded independently of . Let be fixed. Lemma 4.3 implies that, as soon as , is bounded in independently of . Moreover, because owing to Proposition 2.5, we can apply Aubin–Lions Lemma to the sequence to get that, up to extraction, it converges strongly in to some limit . Up to another extraction, we ensure that the convergence of also holds pointwise.
Using Aubin–Lions Lemma for the sequence as in the proof of Proposition 3.4, we find that, up to another extraction, it converges strongly to a limit in and weakly in . The solutions are associated with a bounded vector field , which will converge weakly- in to a vector field . Using the same arguments as in the proof of Theorem 3.3, Claim 2, we can pass to the limit in the equation to find that the pair solves (1.2).
Step 2. Long-time behavior of .
Let be a solution of (1.2), as built in the previous step. The integral version of Lemma 4.4, which is valid for , also applies to at the limit, and we have
hence, for all , we have
Moreover, using the fact that is non-increasing, we get, integrating the relation from Lemma 4.4,
from which we get that there are such that
Now, let us denote . This is well defined for all . We have
and, using Young’s inequality, we get that there is (depending on and ) such that
Hence
Now, let be close enough to so that . Let . By classical interpolation arguments on spaces, one has
where . Now that we have that the norm of goes to zero as goes to , Corollary A.2 gives us that the norm of also goes to zero when goes to .
Step 3. Long-time behavior of .
We now turn to the convergence of as . Let be a sequence of positive real numbers diverging to . Define
Then, solves
Using the same estimates as in the first step, we find that, up to a subsequence, converges to some in the sense, that satisfies
where we have used the uniform convergence as from the previous step in order to get the convergence of to as . We now want to prove . From the boundedness of , the function is also bounded.
Let be fixed. Let be the solutions of
(4.7) |
with homogeneous Dirichlet boundary conditions and final data and , where and is the restriction to of the solution of the torsion equation in (with a domain that contains , say ) with Dirichlet boundary conditions. We recall that the existence of is guaranteed by Proposition 2.5.
The parabolic comparison principle, Proposition 2.6, implies that, for every ,
Let us prove that converge to , the stationary solution of (4.7), as goes to . To get this, let us show that the sequences of functions and are non-decreasing and non-increasing respectively, in the sense that and on for every .
Let be fixed and let . Because (4.7) is autonomous, and are both solutions of (4.7) on , with final data and respectively. However, because for (as recalled in Proposition 2.5), we have . To phrase it differently, and are solutions of the same equation with ordered final data, hence, we can apply the comparison principle Proposition 2.6 to find that on .
Similarly , we have that and solve (4.7) on , with final data and . By a standard comparison principle, we have that . Therefore, we can apply the parabolic comparison principle Proposition 2.6 to get that on .
Therefore, owing to theses monotonicities, the sequences and converge a.e. as goes to to functions that do not depend on the variable (this last fact comes from the equality , which is true because (4.7) is autonomous and because the solutions are unique). Moreover, arguing as in the first step, we have that these limiting functions are solutions of (4.7). The only stationary solution of (4.7) being , we get that for every . We have thus proven that
in the sense.
Let us prove that this convergence is actually uniform. To this aim, observe that is a weak solution of
where is bounded. Then, for every such that , using Corollary A.3, we find that
Integrating this for , we find
Because goes to zero in the sense and is bounded, observing that converges uniformly to zero (this comes from the uniform convergence to zero of from Step 2), we obtain that goes to zero uniformly, whence
in the sense. ∎
4.2. Improved convergence results
In the previous section, Theorem 4.2 proved the existence of solutions to the MFG system with infinite time horizon (1.2) and characterized the asymptotic behavior of any such solution by providing uniform convergence and . We want here to improve this result in two ways: first, we will prove that this convergence is actually exponential (in what concerns this requires a very small extra assumption on the function ); second, we will prove that the convergence of to as , in addition to being uniform, is also a strong convergence in . This last result is natural to evoke, because of the role played by in the dynamics.
Proposition 4.5.
Suppose that the function is Hölder continuous. Then, there exist constants (depending on , , and ), such that we have, for any ,
Proof.
The exponential convergence of to is indeed part of the proof of Theorem 4.2, since we proved that, for close to , the norm of tends exponentially to , and we then used the parabolic regularization estimate .
Thanks to the assumption that is Hölder continuous, up to modifying the coefficient in the exponent, we obtain , where .
We need now to discuss the exponential convergence of . Let us fix a time and define
We will use a comparison principle between and . The functions solve
where the time-derivative term is actually since they are functions of the variable only. If we set , the functions solve a linear PDE of the form
where the vector fields are such that . In particular, if we note that, for large enough, we have , we have . Hence, for we have
and for
Let us look now at the boundary conditions of relative to the parabolic domain . If is large enough, using the uniform convergence , we can infer for every . Moreover, we also have for every and every . The inequalities are opposite for , i.e. we have for or . This implies, by the maximum principle in [2] (see [2, Theorem 1], adapted to this backward equation, and using again the version with the inequality presented at the end of the proof, page 98), the inequalities , i.e.
This shows , for a new constant . ∎
We can now pass to the following statement, which proves the convergence of the gradient of .
Proposition 4.6.
Proof.
We first observe that, by Lemma 4.3, the family of functions is bounded in . This, together with the uniform convergence to , implies that one has the weak convergence in as .
In order to conclude the proof, it suffices to show as . Since in as , one has
Let us argue by contradiction and assume that there exists and an increasing sequence with as such that
(4.8) |
for every . Recall that, from (4.6) in the proof of Lemma 4.3, there exists depending only on , , , and such that
for every and . Combining this with (4.8), one obtains that there exists depending only on , , and such that
(4.9) |
for every and . By Lemma 4.3, there exists a constant depending only on , , , , and such that
In particular, for every , there exists such that . Hence is bounded in and thus, up to extracting subsequences (which we still denote by and for simplicity), converges strongly in as . Since as , the strong limit of in is necessarily , and thus, in particular, as . This, however, contradicts (4.9), and establishes the desired result. ∎
Appendix A Regularizing effects of parabolic equations
This appendix is concerned with the regularizing properties of a class of parabolic equations including both the Fokker–Planck and the Hamilton–Jacobi–Bellman equations we consider in this paper. More precisely, we consider the increase of the exponent of the integrability in space of the solution of the system. As recalled in the introduction, the computations and results presented here are very similar to those from the appendix of [7], the main difference lying in the boundary condition. The main result of this appendix is the following.
Proposition A.1.
Let . Let with , , on , such that
(A.1) |
Then, for every , every number and such that and , there is , depending only on such that
The same result is true omitting the assumption if the PDE (A.1) is satisfied as an equality instead of an inequality.
The proof follows a standard method based on Moser’s iterations that will be detailed here. This appendix is included for completeness: the experienced reader will recognize well-known computations, which are simplified in this setting thanks in particular to the Dirichlet boundary conditions we use.
Proof.
Let be as in the proposition. For , we define
We also define if (here is the Sobolev exponent in dimension ). When we set (but any number larger than and smaller than could be used here). Moreover, we set
Step 1. estimates.
Let us start with proving that, for , there is depending on and on the norms of , such that, for every ,
(A.2) |
In order to do so, we differentiate with respect to , to get
Now, owing to Young’s inequality, we can find depending only on , , , such that
(note that we replaced the coefficient with , as these two numbers are equivalent up to multiplicative constants as far as ). Moreover, thanks again to a Young inequality, we have
Therefore, up to increasing , we get
Now, owing to the Gagliardo–Nirenberg–Sobolev inequality, we have, for some ,
Hence
Let us denote . Then, the above equation gives us
which we can rewrite in order to get (A.2).
Step 2. Estimates on .
We show in this step that, for and for , we have
(A.3) |
for some depending on and on the norms of .
The relation (A.2) provides
Let us take . We integrate the above inequality for to get:
Taking the power and using its subadditivity yields
We replace by so as to re-write the above inequality as
We multiply by and integrate this for in order to obtain
We then use in order to obtain
and finally we use , which provides the desired inequality.
Step 3. Higher integrability estimates.
Let us now show that, for , there is (depending on the same quantities as in the previous steps), such that
(A.4) |
First of all, integrating (A.2) for , and discharging the final value , we obtain
Combining this with (A.3), we get
Up to enlarging the constant and using , we can write the above inequality in a simpler form, i.e.
hence, (A.4) holds true, after raising to the power and including in the constant .
Step 4. Iterations.
We conclude the proof in this step by proving that, for as in the statement of the proposition, there is such that
(A.5) |
We denote
and
Then, (A.4) gives us that
Hence, up to replacing the constant with a larger one so as to suppose , we find
We observe that we have as a consequence of the logarithmic estimate
Therefore, we obtain
Hence, thanks to , we obtain (A.5). This concludes the proof. ∎
Corollary A.2.
Let . Let . Let be a positive distributional solution of
satisfying the following mild regularity assumption: is obtained as a measurable curve of functions of the variable, which is such that is continuous in time for every (note that we do not restrict to ). Then, for every and , there is , depending only on , , , such that we have
for every with .
Proof.
To prove this estimate the only important point is to regularize the equation so as to apply Proposition A.1. In order for the proof to be self-contained, we detail a two-step approximation argument.
We convolve the equation by an approximation of the identity and to apply Proposition A.1. However, convolving will not preserve the Dirichlet boundary conditions, so we first have to extend by zero on a bigger set.
We define to be a open bounded regular set such that , where is unit ball in .
We define if , and elsewhere. Let be an approximation of the identity whose support is included in . We define (here, is the convolution in space only). It is a function which is smooth in and continuous in . We then convolve in time as well, taking an approximation of the identity whose support is included in . Defining we have now a function which is smooth in time and space. It satisfies, in the classical sense,
with . Moreover, the norm of is bounded independently of and . Then, is positive, regular and is a (classical) subsolution of a Fokker–Planck equation with regular coefficients, hence we can apply Proposition A.1. We then take the limit , and we observe that we have
for every , since is continuous. Then, we have
from standard properties of convolutions (with the possibility, of course, that this limit and this norm take the value ). ∎
Corollary A.3.
Let . Let and let be solution (in the weak sense) of
with Dirichlet boundary conditions and initial datum .
Then, for every , and there is , depending only on such that
for every with .
Proof.
Let be and such that and in the norm. Assume moreover that we have and . Let be the solution of
with Dirichlet boundary condition and with initial datum , where is a smooth approximation of .
Then, is smooth enough to apply Proposition A.1 to , to get, for as in the statement of the corollary,
(A.6) |
Then, as goes to , converges (the arguments to prove this are standard and based on the weak convergence of ) to a solution (in the weak sense) of
with Dirichlet boundary conditions and with initial datum . The convergence is also strong in the sense. Because this solution is unique, it necessarily coincides with the original solution of the statement. In order to obtain the result, we need to pass to the limit the inequality (A.6). The left-hand side can easily be dealt with by semicontinuity, while for the right-hand side, we suppose and we use strong convergence. Since this convergence is in space-time, we have convergence of the right-hand side only for a.e. . Yet, using the fact that the solution is continuous in time as a function valued into , the result extends to all . The inequality for implies that with , up to modifying the constant in a way depending on . ∎
The reader may observe that we used different regularization strategies to prove the two above corollaries. Indeed, the linear behavior of the Fokker–Planck equation allowed to directly regularize the solution (up to modifying the drift vector field: we convolve the solution and define a new drift vector field which preserves the same bound, a trick which is completely standard for curves in the Wasserstein space, see for instance [31, Chapter 5]). This is not possible for the Hamilton–Jacobi–Bellman equation. However, when uniqueness of the solution is known, regularizing the coefficients and the data of the equation is another option, and it is what we did in our last corollary.
Acknowledgments. The authors wish to thank many colleagues for useful discussions and suggestions, and in particular Alessio Porretta. Without the comments he made after a talk the second author gave on the topic of the present paper, the strategy to achieve convergence to a solution in the limit would have been completely different, the result less general, and the time needed to achieve it much longer.
The authors acknowledge the financial support of French ANR project “MFG”, reference ANR-16-CE40-0015-01, and of a public grant as part of the “Investissement d’avenir” project, reference ANR-11-LABX-0056-LMH, LabEx LMH, PGMO project VarPDEMFG. The first author was also partially supported by the by the French IDEXLYON project Impulsion “Optimal Transport and Congestion Games” PFI 19IA106udl and the second author was also partially supported by the Hadamard Mathematics LabEx (LMH) through the grant number ANR-11-LABX-0056-LMH in the “Investissement d’avenir” project.
References
- [1] Y. Achdou and A. Porretta. Mean field games with congestion. Ann. Inst. H. Poincaré Anal. Non Linéaire, 35(2):443–480, 2018.
- [2] D. G. Aronson and J. Serrin. Local behavior of solutions of quasilinear parabolic equations. Arch. Rational Mech. Anal., 25:81–122, 1967.
- [3] J.-P. Aubin and H. Frankowska. Set-valued analysis. Modern Birkhäuser Classics. Birkhäuser Boston, Inc., Boston, MA, 2009. Reprint of the 1990 edition.
- [4] J.-D. Benamou, G. Carlier, and F. Santambrogio. Variational mean field games. In Active particles. Vol. 1. Advances in theory, models, and applications, Model. Simul. Sci. Eng. Technol., pages 141–171. Birkhäuser/Springer, Cham, 2017.
- [5] P. Cannarsa and C. Sinestrari. Semiconcave functions, Hamilton–Jacobi equations, and optimal control. Progress in Nonlinear Differential Equations and their Applications, 58. Birkhäuser Boston, Inc., Boston, MA, 2004.
- [6] P. Cardaliaguet. Notes on mean field games (from P.-L. Lions’ lectures at Collège de France). Available at https://www.ceremade.dauphine.fr/~cardaliaguet/MFG20130420.pdf, 2013.
- [7] P. Cardaliaguet, J.-M. Lasry, P.-L. Lions, and A. Porretta. Long time average of mean field games with a nonlocal coupling. SIAM J. Control Optim., 51(5):3558–3591, 2013.
- [8] L. C. Evans. Partial differential equations, volume 19 of Graduate Studies in Mathematics. American Mathematical Society, Providence, RI, 1998.
- [9] F. Feo. A remark on uniqueness of weak solutions for some classes of parabolic problems. Ric. Mat., 63(1, suppl.):S143–S155, 2014.
- [10] W. H. Fleming and H. M. Soner. Controlled Markov processes and viscosity solutions, volume 25 of Stochastic Modelling and Applied Probability. Springer, New York, second edition, 2006.
- [11] D. Gilbarg and N. S. Trudinger. Elliptic partial differential equations of second order. Classics in Mathematics. Springer-Verlag, Berlin, 2001. Reprint of the 1998 edition.
- [12] D. A. Gomes and V. K. Voskanyan. Short-time existence of solutions for mean-field games with congestion. Journal of the London Mathematical Society, 92(3):778–799, 11 2015.
- [13] A. Granas and J. Dugundji. Fixed point theory. Springer Monographs in Mathematics. Springer-Verlag, New York, 2003.
- [14] M. Huang, P. E. Caines, and R. P. Malhamé. Individual and mass behaviour in large population stochastic wireless power control problems: centralized and Nash equilibrium solutions. In 42nd IEEE Conference on Decision and Control, 2003. Proceedings, volume 1, pages 98–103. IEEE, 2003.
- [15] M. Huang, P. E. Caines, and R. P. Malhamé. Large-population cost-coupled LQG problems with nonuniform agents: individual-mass behavior and decentralized -Nash equilibria. IEEE Trans. Automat. Control, 52(9):1560–1571, 2007.
- [16] M. Huang, R. P. Malhamé, and P. E. Caines. Large population stochastic dynamic games: closed-loop McKean–Vlasov systems and the Nash certainty equivalence principle. Commun. Inf. Syst., 6(3):221–251, 2006.
- [17] R. L. Hughes. A continuum theory for the flow of pedestrians. Transportation Research Part B: Methodological, 36(6):507–535, jul 2002.
- [18] R. L. Hughes. The flow of human crowds. In Annual review of fluid mechanics, Vol. 35, volume 35 of Annu. Rev. Fluid Mech., pages 169–182. Annual Reviews, Palo Alto, CA, 2003.
- [19] O. A. Ladyženskaja, V. A. Solonnikov, and N. N. Ural\cprimeceva. Linear and quasilinear equations of parabolic type. Translations of Mathematical Monographs, Vol. 23. American Mathematical Society, Providence, R.I., 1968. Translated from the Russian by S. Smith.
- [20] O. A. Ladyzhenskaya and N. N. Ural’tseva. Linear and quasilinear elliptic equations. Academic Press, New York-London, 1968. Translated from the Russian by Scripta Technica, Inc, Translation editor: Leon Ehrenpreis.
- [21] J.-M. Lasry and P.-L. Lions. Jeux à champ moyen. I. Le cas stationnaire. C. R. Math. Acad. Sci. Paris, 343(9):619–625, 2006.
- [22] J.-M. Lasry and P.-L. Lions. Jeux à champ moyen. II. Horizon fini et contrôle optimal. C. R. Math. Acad. Sci. Paris, 343(10):679–684, 2006.
- [23] J.-M. Lasry and P.-L. Lions. Mean field games. Jpn. J. Math., 2(1):229–260, 2007.
- [24] G. M. Lieberman. Second order parabolic differential equations. World Scientific Publishing Co., Inc., River Edge, NJ, 1996.
- [25] P.-L. Lions. Résolution de problèmes elliptiques quasilinéaires. Arch. Rational Mech. Anal., 74(4):335–353, 1980.
- [26] P.-L. Lions. Courses at Collège de France, 2006–2012. http://www.college-de-france.fr/site/pierre-louis-lions/_course.htm.
- [27] G. Mazanti and F. Santambrogio. Minimal-time mean field games. Math. Models Methods Appl. Sci., 29(8):1413–1464, 2019.
- [28] A. Porretta. Weak solutions to Fokker-Planck equations and mean field games. Arch. Ration. Mech. Anal., 216(1):1–62, 2015.
- [29] M. M. Porzio. Existence of solutions for some “noncoercive” parabolic equations. Discrete Contin. Dynam. Systems, 5(3):553–568, 1999.
- [30] F. Santambrogio. Lecture notes on variational mean field games. Preprint cvgmt. http://cvgmt.sns.it/paper/4646/.
- [31] F. Santambrogio. Optimal transport for applied mathematicians, volume 87 of Progress in Nonlinear Differential Equations and their Applications. Birkhäuser/Springer, Cham, 2015. Calculus of variations, PDEs, and modeling.
- [32] J. Simon. Compact sets in the space . Ann. Mat. Pura Appl. (4), 146:65–96, 1987.