sectioning
Nonlinear Markov Chains with Finite State Space: Invariant Distributions and Long-Term Behaviour
Abstract
Nonlinear Markov chains with finite state space have been introduced in Kolokoltsov (2010) [9]. The characteristic property of these processes is that the transition probabilities do not only depend on the state, but also on the distribution of the process. In this paper we provide first results regarding their invariant distributions and long-term behaviour. We will show that under a continuity assumption an invariant distribution exists. Moreover, we provide a sufficient criterion for the uniqueness of the invariant distribution that relies on the Brouwer degree. Thereafter, we will present examples of peculiar limit behaviour that cannot occur for classical linear Markov chains. Finally, we present for the case of small state spaces sufficient (and easy-to-verify) criteria for the ergodicity of the process.
1 Introduction
Nonlinear Markov processes are a particular class of stochastic processes, where the transition probabilities do not only depend on the state, but also on the distribution of the process. McKean [10] introduced these type of processes to tackle mechanical transport problems. Thereafter they have been studied by several authors (see the monographs of Kolokoltsov [9] and Sznitman [17]). Recently, the close connection to continuous time mean field games led to significant progress in the analysis of McKean-Vlasov SDEs, in particular the control of these systems (see for example [4, 14]).
In this paper, we consider a special class of these processes, namely, nonlinear Markov chains in continuous time with a finite state space and provide first insights regarding the long-term behaviour of these processes. Nonlinear Markov chains with finite state space arise naturally, in particular in evolutionary biology, epidemiology and game theory. Namely, the replicator dynamics, several infection models, but also the dynamics of learning procedures in game theory are nonlinear Markov chains [9]. Moreover, also the population’s dynamics in mean field games with finite state and action space are nonlinear Markov chains [12].
The main focus of this paper lies in the characterization of the long-term behaviour of these processes. We show that always an invariant distribution exists and provide a sufficient criterion for the uniqueness of this invariant distribution. Thereafter, we turn to the long-term behaviour, where we first illustrate by two examples that the limit behaviour is much more complex than for classical Markov chains. More precisely, we show that the marginal distributions of a nonlinear Markov chain might be periodic and that irreducibility of the generator does not necessarily imply ergodicity. Then we provide easy-to-verify sufficient criteria for ergodicity for small state spaces (two or three states). All conditions that we propose are simple and rely only on the shape of the nonlinear generator, not on the shape of the transition probabilities.
The long-term behaviour of nonlinear Markov chains in continuous time with a finite state space has not been analysed before. The closest contribution in the literature are ergodicity criteria for nonlinear Markov processes in discrete time [3, 16]. These criteria are a generalization of Dobrushin’s ergodicity condition and the proofs crucially rely on the sequential nature of the problem.
The rest of the paper is structured as follows: In Section 2 we review the relevant definitions and notation. In Section 3 we present the results on existence and uniqueness of the invariant distribution. In Section 4 we provide examples of limit behaviour that cannot arise in the context of classical Markov chains. In Section 5 we present the ergodicity results for small state spaces. The Appendix A contains the proofs of two technical results.
2 Continuous Time Nonlinear Markov Chains with Finite State Space
This section gives a short overview over the relevant definitions, notations and preliminary facts regarding nonlinear Markov chains. For more details regarding these processes we refer the reader to [9, Chapter 1]. Moreover, it introduces the relevant notions to characterize the long-term behaviour of these processes.
Let be the state space of the nonlinear Markov chain and denote by the probability simplex over . A nonlinear Markov chain is characterized by a continuous family of nonlinear transition probabilities which is a family of stochastic matrices that depends continuously on and such that the nonlinear Chapman-Kolmogorov equation
is satisfied. As usual is interpreted as the probability that the process is in state at time given that the initial state was and the initial distribution of the process was . Such a family yields a nonlinear Markov semigroup of continuous transformations of via
Also has the usual interpretation that it represents the marginal distribution of the process at time when the initial distribution is . A nonlinear Markov chain with initial distribution is then given as the time-inhomogeneous Markov chain with initial distribution and transition probabilities .
As in the theory of classical continuous time Markov chains the infinitesimal generator will be the cornerstone of the description and analysis of such processes: Let be differentiable in and , then the (nonlinear) infinitesimal generator of the semigroup is given by a transition rate matrix function such that for we have for all and .
By [9, Section 1.1] any differentiable nonlinear semigroup has a nonlinear infinitesimal generator. However, the converse problem is more important: Given a transition rate matrix function (that is a function such that is a transition rate matrix for all ) is there a nonlinear Markov semigroup (and thus a nonlinear Markov chain) such that is the nonlinear infinitesimal generator of the process? Relying on the semigroup identity this problem is equivalent to the following Cauchy problem: Is there, for any a solution of
such that is a continuous function ranging from to itself and such that for all and .
In the monograph [9] the problem to construct a semigroup from a given generator is treated in a very general setting. Here, we present a result with easy-to-verify conditions tailored for the specific situation of nonlinear Markov chains with finite state space. The proof of the result, which relies on classical arguments from ODE theory, is presented in the appendix.
Theorem 2.1.
Let be a transition rate matrix function such that is Lipschitz continuous for all . Then there is a unique Markov semigroup such that is the infinitesimal generator for .
In this paper we are now mainly interested in the characterization of the long-term behaviour of nonlinear Markov chains: We say that is an invariant distribution if and thus also . An equivalent condition with respect to the generator is that a vector is an invariant distribution if it solves .
We say that a nonlinear Markov chain with nonlinear semigroup is strongly ergodic if there exists an such that for all we have
3 Existence and Uniqueness of the Invariant Distribution
The invariant distributions of a nonlinear Markov chain are exactly the fixed points of the set-valued map
Using Kakutani’s fixed point theorem, we directly obtain the existence of an invariant distribution for any generator:
Theorem 3.1.
Let be a nonlinear generator such that the map is continuous. Then the nonlinear Markov chain with generator has an invariant distribution.
Proof.
By [7, Theorem 5.3] the set of all invariant distributions given a fixed generator matrix is the convex hull of the invariant distributions given the recurrent communication classes of . Therefore, the values of the map are non-empty, convex and compact. Moreover, the graph of the map is closed: Let be a converging sequence such that . Denote its limit by . Then for all . By continuity of we have , which implies . Thus, Kakutani’s fixed point theorem yields a fixed point of the map , which is an invariant distribution given . ∎
If is irreducible for all , the sets will be singletons [1, Theorem 4.2]. Let denote this point. We remark that there are explicit representation formulas for (e.g. [13, 15]). With these insights we provide the following sufficient criterion for the uniqueness of the invariant distribution:
Theorem 3.2.
Assume that is irreducible for all . Furthermore, assume that is continuously differentiable and that the matrix
is non-singular for all . Then there is a unique invariant distribution.
Proof.
We first note that any invariant distribution of a nonlinear Markov chain with generator is an invariant distribution of a classical Markov chain with generator . Since any invariant distribution of a classical Markov chain with generator has to satisfy that all components are strictly positive [1, Theorem 4.2], no invariant distribution of lies on the boundary of . Therefore, we only need to ensure the existence of a unique invariant distribution in the interior of .
The set is homeomorphic to with
where the continuous bijections are given as the restrictions of
Define by . By the chain rule we obtain
The matrix is, by assumption, non-singular for all . Thus,
Since , , and det are continuous functions, we obtain that also the function is continuous. Thus, the intermediate value theorem yields that has uniform sign over .
Furthermore, we note that by assumption is in particular non-singular for all . Thus, is a non-critical value of .
The map given by
is continuous. Furthermore, : Indeed, a point either satisfies for some or . However, by [1, Theorem 4.2], all components of the invariant distribution for an irreducible generator are strictly positive. Thus, we obtain in the first case that and in the second case that the sum of all components is strictly negative, which in both cases implies that .
With these preparations we can make use of the Brouwer degree (see [5, Section 1.1 and 1.2]), namely we obtain that
Since for continuously differentiable maps and regular values the degree is given by
we obtain that
Because the determinant has uniform sign over , we obtain that consists of exactly one element. Thus, there is a unique stationary point for the nonlinear Markov chain with nonlinear generator . ∎
Example.
We illustrate the use of the result in an example: Consider a nonlinear Markov chain with the following generator
where all constants are strictly positive. This nonlinear Markov chain arises in a mean field game model of consumer choice with congestion effects (see [12], also for detailed calculations). In this setting the invariant distributions are given as the solution(s) of the nonlinear equation , for which closed form solutions are hard or impossible to obtain. However, it is possible to verify that the matrix is non-singular for all yielding a unique invariant distribution. This information can in particular be used, to obtain certain characteristic properties of the solutions.
4 Examples for Peculiar Limit Behaviour
The following examples show that the limit behaviour for nonlinear Markov chains (also in the case of small state spaces) is more complex than for classical continuous time Markov chains. In particular, the marginal distributions might not converge, but are periodic and a nonlinear Markov chain with an irreducible nonlinear generator might not be strongly ergodic, but we observe convergence towards several different invariant distributions.
4.1 An Example with Periodic Marginal Distributions
Let and set for all the matrix as follows
where is if is true and else. Since all transition rates on are Lipschitz continuous functions, there is an extension of on for all , which is again Lipschitz continuous. Thus, a nonlinear Markov chain with generator exists. The ordinary differential equation characterizing the marginals on reads
Thus, for any neighbourhood of the first two components of the marginal behave like the classical harmonic oscillator. Therefore, there are initial distributions such that the marginals are periodic. An example is the initial distribution for which the marginals are plotted in Figure 1.

4.2 An Example of a Nonlinear Markov Chain with Irreducible Generator that is not Strongly Ergodic
Let
This matrix is irreducible for all since and for all .
The ordinary differential equation describing the marginals for the initial condition is given by
We obtain that there are three stationary points , and and the following convergence behaviour:
-
•
Since the function is strictly positive on , the trajectories will for all initial conditions converge towards .
-
•
Since the function is strictly negative on , the trajectories will for all initial conditions converge towards .
-
•
Since the function is strictly positive on , the trajectories will for all initial conditions converge towards .
-
•
Since the function is strictly negative on , the trajectories will for all initial conditions converge towards .
This behaviour is visualized in Figure 2, where several trajectories for different initial conditions are plotted.

5 Sufficient Criteria for Ergodicity for Small State Spaces
Although the limit behaviour is more complex for nonlinear Markov chains, we still obtain sufficient criteria for ergodicity in the case of a small number of states. Here, we present these criteria, discuss applicability as well as the problems that occur for larger state spaces.
Theorem 5.1.
Let and assume that defined via
is continuous. Furthermore, assume that is the unique stationary point given . Then, the nonlinear Markov chain is strongly ergodic.
Proof.
An equilibrium point is characterized by the property that . By flow invariance of for the ordinary differential equation (see the proof of Theorem 2.1), which implies that , this property is equivalent to the fact that .
Since and since we have a unique equilibrium point, we obtain that and for all . Since is continuous, we obtain that is non-vanishing on and and has uniform sign on each of these sets. Since is a conservative generator we moreover obtain that and . Thus, we obtain that for all and for all . This in turn yields that is flow invariant for .
Fix . Then the systems and are equivalent in the sense that for all , : Indeed, let be a solution of the differential equation with initial condition . By flow invariance of for (see Theorem 2.1), we have for all . Thus, is equivalent to
(1) |
Therefore, is indeed a solution of . For the converse implication we first note that because is conservative for all the last equation of (1) is the first equation multiplied by . If satisfies , , then, by flow invariance, for all . Thus, the function satisfies .
The desired convergence statement directly follows from for all and for all . ∎
We also obtain a sufficient criterion for the case of three states. The proof technique is similar to the two state case. Indeed, we first show that our system is equivalent to a two-dimensional system, for which we can then use standard tools for two-dimensional dynamical systems exploiting that the dynamical system has a particular shape since is a conservative generator.
As mentioned, we obtain for systems with three states that given the function is a solution of , if and only of is a solution of
where
(2) |
and . Indeed, the proof is analogous to the proof for the two state case, the central adjustment is to prove the flow invariance of for instead of the flow invariance of for . This statement is proven in the appendix (Lemma A.1).
To show the desired convergence statement, we now rely on the Poincaré-Bendixson Theorem [18, Chapter 7], which characterizes the -limit sets of a trajectory with initial condition :
Theorem 5.2.
Let be a simply connected and bounded region such that there is a continuously differentiable function satisfying (2) on . Let be the unique stationary point given . Furthermore, assume that
-
(a)
is non-vanishing for all and has uniform sign on ,
-
(b)
it holds that
or it holds that
Then, the nonlinear Markov chain is strongly ergodic.
Proof.
Since the set is flow invariant for , any trajectory will stay in this set. Since the set is compact, we obtain by [18, Lemma 6.6] that lies . Since there is, by assumption, only one stationary point we can apply the Poincaré-Bendixson Theorem [18, Theorem 7.16]. It yields that one of the following three cases holds:
-
(i)
-
(ii)
is a regular periodic orbit
-
(iii)
consists of (finitely many) fixed points and non-closed orbits such that .
By condition (a) and Bedixson’s criterion [8, Theorem 3.5] the case (ii) is not possible. Since, by condition (b), the point is not a saddle point, there is no homoclinic path joining to itself. Therefore, since is the only stationary point, also case (iii) is not possible. Thus, . Since the considered trajectory lies in the compact set , we moreover obtain by [18, Lemma 6.7] that
∎
Remark 5.3.
The equivalence of the considered systems and systems on some subset of as well as the construction performed in Section 4.1 hint the general problem for a larger number of states (). It might happen that the dynamics of the nonlinear Markov chain describe a classical “chaotic” nonlinear system like the Lorentz system. In other words, the difficulties that arise in the classical theory of dynamical systems might also arise here, for which reason criteria for a larger number of states are more complex.
Example.
Theorem 5.2 now yields strong ergodicity of the nonlinear Markov chain introduced in the end of Section 3. In this setting the function is given by
and we moreover have for all as well as
for all and, thus, in particular for the unique invariant distribution. Therefore, by Theorem 5.2 we obtain strong ergodicity.
Appendix A Appendix
Proof of Theorem 2.1.
We first note that
is Lipschitz continuous on : Indeed, let be a Lipschitz constant for all functions () simultaneously. Moreover, since is compact there is a finite constant
Thus, we have
By McShane’s extension theorem [11] there is a Lipschitz continuous extension of . Let us fix an arbitrary . By the classical existence and uniqueness theorem for ordinary differential equations, we obtain that there is a unique solution of of .
As a next step we show that the vectors lie for all in the Bouligand tangent cone
where the second line follows from [2, Proposition 5.1.7]: Indeed, since for all interior points of the condition is trivially satisfied, it suffices to consider the boundary points . These points satisfy that there is at least one such that . Since the only non-positive column entry of (which is ) gets weight , the vector will have non-negative entries at each such that . Since is conservative, we moreover obtain that
Thus, for all . Therefore, we obtain, by the classical flow invariance statement for ordinary differential equations ([19, Theorem 10.XVI]), that the solution satisfies for all . Thus, is also the unique solution of . The continuity of follows from a classical general dependence theorem [19, Theorem 12.VII]. ∎
Lemma A.1.
The set is flow invariant for .
Proof.
The statement follows from [6, Lemma 1]. This lemma states that for an open set and a family of continuously differentiable functions () the set
is flow invariant for whenever for any there is an such that and
Indeed, in our case we have
and the boundary points of this set either satisfy for at least one or . Since is conservative and irreducible, we obtain in the first case and in the second case. Therefore, the claim follows. ∎
References
- Asmussen [2003] S. Asmussen. Applied Probability and Queues, volume 51 of Stochastic Modelling and Applied Probability. Springer-Verlag, New York, 2nd edition, 2003. ISBN 0-387-00211-1.
- Aubin and Cellina [1984] J.-P. Aubin and A. Cellina. Differential Inclusions: Set-Valued Maps and Viability Theory, volume 264 of Grundlehren der mathematischen Wissenschaften. Springer-Verlag, Berlin, Heidelberg, 1984. 10.1007/978-3-642-69512-4.
- Butkovsky [2014] O. A. Butkovsky. On Ergodic Properties of Nonlinear Markov Chains and Stochastic McKean-Vlasov Equations. Theory Probab. Its Appl., 58(4):661–674, 2014. 10.1137/S0040585X97986825.
- Carmona and Delarue [2015] R. Carmona and F. Delarue. Forward-backward stochastic differential equations and controlled McKean-Vlasov dynamics. An. Probab., 43(5):2647–2700, 2015. 10.1214/14-AOP946.
- Deimling [1985] K. Deimling. Nonlinear Functional Analysis. Springer-Verlag, Berlin, 1985. ISBN 3-540-13928-1.
- Fernandes and Zanolin [1987] M. L. Fernandes and F. Zanolin. Remarks on Strongly Flow-Invariant Sets. J. Math. Anal. Appl., 128(1):176–188, 1987. 10.1016/0022-247X(87)90223-X.
- Iosifescu [1980] M. Iosifescu. Finite Markov Processes and Their Applications. Wiley series in probability and mathematical statistics. John Wiley & Sons, Ltd., Chichester, New York, Brisbane, Toronto, 1980. ISBN 0-471-27677-4.
- Jordan and Smith [2005] D. W. Jordan and P. Smith. Nonlinear Ordinary Differential Equations: An Introduction to Dynamical Systems, volume 2 of Oxford Applied and Engineering Mathematics. Oxford University Press, Oxford, 3rd edition, 2005. ISBN 0-19-856562-3.
- Kolokoltsov [2010] V. N. Kolokoltsov. Nonlinear Markov Processes and Kinetic Equations, volume 182 of Cambridge Tracts in Mathematics. Cambridge University Press, Cambridge, 2010. ISBN 978-0-521-11184-3.
- McKean [1966] H. P. McKean, Jr. A Class of Markov Processes Associated with Nonlinear Parabolic Equations. Proc. Natl. Acad. Sci. U.S.A., 56(6):1907–1911, 1966. 10.1073/pnas.56.6.1907.
- McShane [1934] E. J. McShane. Extension of range of functions. Bulletin of the American Mathematical Society, 40(12):837–842, 1934.
- Neumann [2019] B. A. Neumann. Stationary Equilibria of Mean Field Games with Finite State and Action Space : Existence, Computation, Stability, and a Myopic Adjustment Process. PhD thesis, Universität Hamburg, 2019. URL http://ediss.sub.uni-hamburg.de/volltexte/2020/10313/.
- Neumann [2020] B. A. Neumann. Stationary Equilibria of Mean Field Games with Finite State and Action Space. Dyn. Games Appl., 2020. 10.1007/s13235-019-00345-9.
- Pham and Wei [2017] H. Pham and X. Wei. Dynamic Programming for Optimal Control of Stochastic McKean-Vlasov Dynamics. SIAM J. Control Optim., 55(2):1069–1101, 2017. 10.1137/16M1071390.
- Resnick [1992] S. I. Resnick. Adventures in Stochastic Processes. Birkhäuser, Boston, 1992. ISBN 0-8176-3591-2.
- Saburov [2016] M. Saburov. Ergodicity of nonlinear Markov operators on the finite dimensional space. Nonlinear Anal. Theory Methods Appl., 143:105–119, 2016. 10.1016/j.na.2016.05.006.
- Sznitman [1991] A.-S. Sznitman. Topics in propagation of chaos. In P.-L. Hennequin, editor, Ecole d’Eté de Probabilités de Saint-Flour XIX — 1989, volume 1464 of Lecture Notes in Mathematics, pages 165–251, Berlin, Heidelberg, 1991. Springer. 10.1007/BFb0085169.
- Teschl [2012] G. Teschl. Ordinary Differential Equations and Dynamical Systems, volume 140 of Graduate Studies in Mathematics. American Mathematical Society, Providence, Rhode Island, 2012. ISBN 978-0-8218-8328-0.
- Walter [1998] W. Walter. Ordinary Differential Equations, volume 182 of Graduate Texts in Mathematics. Springer-Verlag, New York, 1998. ISBN 0-387-98459-3.