This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

\CDat

Probability measures on graph trajectories

Michael J. Catanzaro Department of Mathematics, Iowa State University, Ames, IA 50011 [email protected] Vladimir Y. Chernyak Department of Chemistry, Wayne State University, Detroit, MI 48202 [email protected]  and  John R. Klein Department of Mathematics, Wayne State University, Detroit, MI 48202 [email protected]
Abstract.

The aim of this note is to construct a probability measure on the space of trajectories in a continuous time Markov chain having a finite state diagram, or more generally which admits a global bound on its degree and rates. Our approach is elementary. Our main intention is to fill a gap in the literature and to give some additional details in the proof of [CCK-fluctuation, prop. 4.8].

1. Introduction

As is well known, Markov chains model random walks on graphs. Let Γ\Gamma be a directed graph. Its set of vertices Γ0\Gamma_{0} represent the states of the system and its edges Γ1\Gamma_{1} indicate transitions between states. There are two flavors of random walk: those in discrete time and those in continuous time. This note will consider the continuous time variant.

The dynamics of continuous time random walk are encoded by a master equation

p(t)=(t)p(t),p^{\prime}(t)=\mathbb{H}(t)p(t)\,,

where (t)\mathbb{H}(t) is a time dependent matrix of transition rates and p(t)p(t) is a 1-parameter family of probability distributions on Γ0\Gamma_{0}. The solutions to the equation describe the time evolution of probability. For vertices ii and jj, the matrix entry (t)ij\mathbb{H}(t)_{ij} is the instantaneous rate of change at time tt in jumping from state ii to state jj along the set of edges of Γ\Gamma having initial vertex ii and terminal vertex jj. The operator \mathbb{H} is called the master operator; its off diagonal entries are non-negative and the sum of the entries in any column add to zero.

Given a continuous time Markov chain with state diagram Γ\Gamma, our goal here will be to construct a probability distribution on the space of trajectories in Γ\Gamma. By a trajectory in Γ\Gamma, we mean a path of contiguous edges equipped with jump times at each vertex of the path. Note that such a probability distribution amounts to a description of the stochastic process associated with the Markov chain.

Remark 1.1.

We apologize to the reader in advance for our somewhat unconventional treatment: two of us are algebraic topologists and one is a chemical physicist.

2. Preliminaries

For a set TT, let (T2)\binom{T}{2} denote the set of its non-empty subsets of cardinality 22. An undirected graph consists of data

X:=(X0,X1,δ),X:=(X_{0},X_{1},\delta)\,,

in which X0X_{0} is the set of vertices, X1X_{1} is the set of edges and

δ:X1(X02)\delta\colon\!X_{1}\to\tbinom{X_{0}}{2}

is a function. We will always assume that XX is locally finite in the sense that the function δ\delta is a finite-to-one. With this definition multiple edges connecting a pair of distinct vertices are permitted, but we do not permit loop edges, i.e., edges which connect a vertex to itself.

A directed graph Γ\Gamma is defined in a similar way, but where now δ\delta is replaced by a function d:Γ1Γ0(2)d\colon\!\Gamma_{1}\to\Gamma_{0}(2), where Γ0(2):=Γ0×Γ0Δ\Gamma_{0}(2):=\Gamma_{0}\times\Gamma_{0}\setminus\Delta, i.e., the cartesian product with its diagonal deleted. We write d=(d0,d1)d=(d_{0},d_{1}), where di:ΓΓ0d_{i}\colon\!\Gamma\to\Gamma_{0} is is the function which assigns to a directed edge its source, respectively target. Note that the canonical map π:Γ0(2)(Γ02)\pi\colon\!\Gamma_{0}(2)\to\binom{\Gamma_{0}}{2} is a double cover, and the composition δ:=πd\delta:=\pi\circ d defines the underlying undirected graph.

Example 2.1.

Given an undirected graph XX, we may construct its double. This is the directed graph

DX:=Γ=(Γ0,Γ1,d),DX:=\Gamma=(\Gamma_{0},\Gamma_{1},d)\,,

in which Γ0=X0\Gamma_{0}=X_{0} and Γ1\Gamma_{1} is the set of ordered pairs (i,α)X0×X1(i,\alpha)\in X_{0}\times X_{1} in which iδ(α)i\in\delta(\alpha). The function d:Γ1Γ0(2)d\colon\!\Gamma_{1}\to\Gamma_{0}(2) is given by d(i,α)=(i,j)d(i,\alpha)=(i,j), where δ(α)={i,j}\delta(\alpha)=\{i,j\}.

Remark 2.2.

Let 𝒢\mathcal{G} be the category of undirected graphs. An object is an undirected graph and a morphism f:(G0,G1,δ)(H0,H1,δ)f\colon\!(G_{0},G_{1},\delta)\to(H_{0},H_{1},\delta^{\prime}) consists of functions fi:GiHif_{i}\colon\!G_{i}\to H_{i}, i=0,1i=0,1 such that δf1(α)=f0(δα)\delta^{\prime}f_{1}(\alpha)=f_{0}(\delta\alpha). Similarly, one has the category 𝒢+\mathcal{G}^{+} of directed graphs. Then we have an adjoint functor pair

U:𝒢+𝒢:DU\colon\!\mathcal{G}^{+}\leftrightarrows\mathcal{G}:D

where UU is the forgetful functor and DD is given by the double.

2.1. Markov chains

Let Γ\Gamma be a directed graph. A continuous time Markov chain with state diagram Γ\Gamma is an assignment of a continuous function

kα:[0,),k_{\alpha}\colon\!\mathbb{R}\to[0,\infty)\,,

to each edge αΓ1\alpha\in\Gamma_{1}. The function kαk_{\alpha} is called the transition rate of α\alpha. If d(α)=(i,j)d(\alpha)=(i,j), then kαk_{\alpha} is to be interpreted as the instantaneous rate of change of probability in jumping from ii to jj along α\alpha.

Remark 2.3.

The foundational material on Markov chains can be found in the texts of Norris [Norris] and Stroock [STR14]. When the transition rates are constant, the Markov chain is said to be time homogeneous. When the rates are not constant, the chain is said to be time inhomogeneous.111In contrast with the homogeneous case, the literature on the inhomogeneous case is scant, with the known results making strong additional assumptions. The only foundational work we are of aware of that treats the time inhomogeneous case is Stroock’s text (cf. [STR14, §5.5.2]).

Remark 2.4.

The canonical map ΓDUΓ\Gamma\to DU\Gamma is an embedding. Given a Markov chain on Γ\Gamma, one has a canonical extension to DUΓDU\Gamma by defining the rates to be zero on those edges which aren’t in Γ\Gamma. The Markovian dynamics of the two chains coincide. From this standpoint, there is nothing to lose by assuming that Γ=DX\Gamma=DX for some undirected graph XX.

If Γ\Gamma is infinite, we also require the following growth constraints.

Definition 2.5 (Rate Bound).

For each t>0t>0, there exists a constant RR, possibly depending on tt, such that

kα(s)Rk_{\alpha}(s)\leq R

for 0st0\leq s\leq t and every αΓ1\alpha\in\Gamma_{1}.

Definition 2.6 (Degree Bound).

Let deg:Γ0\deg\colon\!\Gamma_{0}\to\mathbb{N} be the function which assigns to a vertex its degree, i.e., the number of edges meeting it. There is a positive integer DD such that

deg(i)D,for all iΓ0.\deg(i)\leq D\,,\quad\text{for all }\quad i\in\Gamma_{0}\,.

Observe that when Γ\Gamma is a finite, both conditions hold automatically.

2.2. The master equation

Then the rates define a time dependent square matrix =(t)\mathbb{H}=\mathbb{H}(t), as follows. For iji\neq j, set

hij=d(α)=(i,j)kα,h_{ij}=\sum_{d(\alpha)=(i,j)}k_{\alpha}\,,

where the sum is interpreted as zero when d1(i,j)d^{-1}(i,j) is the empty set. Then the matrix entries of \mathbb{H} are given by

ij={hij,ij;ihi, if i=j,\mathbb{H}_{ij}=\begin{cases}h_{ij}\,,\qquad&i\neq j\,;\\ -\sum_{\ell\neq i}h_{\ell i}\,,\quad\text{ if }&i=j\,,\end{cases}

where the indices range over i,jΓ0i,j\in\Gamma_{0}. The time dependent matrix \mathbb{H} is called the master operator. Associated with \mathbb{H} is a linear, first order ordinary differential equation

(1) p(t)=p(t),p^{\prime}(t)=\mathbb{H}p(t),\qquad

in which p(t)p(t) is a one parameter family of (probability) distributions on the set of vertices Γ0\Gamma_{0}. Equation (1) is called the (forward) Kolmogorov equation or the master equation [STR14, eqn. 5.5.2]. Its solutions describe the evolution of an initial distribution p(0)p(0).

Remark 2.7.

If Γ=DX\Gamma=DX and the transition rates are constant with value 1, then \mathbb{H} is the graph Laplacian of XX and (1) is a combinatorial version of the heat (diffusion) equation.

Remark 2.8.

The forward Kolmogorov equation is often written in the literature in adjoint form, i.e., as

q(t)=q(t)𝕎,q^{\prime}(t)=q(t)\mathbb{W}\,,

where q(t)=p(t)q(t)=p(t)^{*} and 𝕎=\mathbb{W}=\mathbb{H}^{*} are the transposed matrices. The backward equation (which we will not consider here) is

q(t)=𝕎q(t).q^{\prime}(t)=\mathbb{W}q(t)\,.

2.3. Trajectories

A path of length nn in Γ\Gamma consists of a sequence of edges

α:=(α1,,αn),\alpha_{\bullet}:=(\alpha_{1},\dots,\alpha_{n})\,,

such that d1(αk)=d0(αk+1)d_{1}(\alpha_{k})=d_{0}(\alpha_{k+1}) for 1k<n1\leq k<n. We let

ik(α):=iki_{k}(\alpha_{\bullet}):=i_{k}

denote the kk-th vertex of the path, i.e., ik=d0(αk))i_{k}=d_{0}(\alpha_{k})) if knk\leq n and in+1=d1(αn)i_{n+1}=d_{1}(\alpha_{n}).

A trajectory of length nn and duration t>0t>0 is a pair

(α,t),(\alpha_{\bullet},t_{\bullet})\,,

such that α\alpha_{\bullet} is a path of length nn and t=(t1,,tn)t_{\bullet}=(t_{1},\dots,t_{n}) is a sequence of real numbers satisfying

0t1tnt.0\leq t_{1}\leq\cdots\leq t_{n}\leq t\,.

In what follows, it will be convenient to set

t0:=0 and tn+1=:t.t_{0}:=0\quad\text{ and }\quad t_{n+1}=:t\,.
Remark 2.9.

For a vertex ik=ik(α)i_{k}=i_{k}(\alpha_{\bullet}) of the path α)\alpha_{\bullet}), the number tkt_{k} is called the jump time and the number wk:=tktk1w_{k}:=t_{k}-t_{k-1} is called wait time.

3. The probability of a trajectory

Let (Γ,k)(\Gamma,k_{\bullet}) be as in the previous section. Given a vertex iΓ0i\in\Gamma_{0} and an interval [a,b][a,b], the escape rate at ii is

ui(a,b):=exp(d1(α)=iabkα(s)𝑑s)=exp(abhii(s)𝑑s).u_{i}(a,b):=\exp\left(-\sum_{d_{1}(\alpha)=i}\int_{a}^{b}k_{\alpha}(s)\,ds\right)=\exp\left(\int_{a}^{b}h_{ii}(s)\,ds\right)\,.

Fix an initial probability distribution q:Γ0+q\colon\!\Gamma_{0}\to\mathbb{R}_{+}. For jΓ0j\in\Gamma_{0}, set qj:=q(j)q_{j}:=q(j).

Let

𝒯(Γ,n,t)\mathcal{T}(\Gamma,n,t)

denote the set of trajectories of Γ\Gamma having length nn. Define a function

f:𝒯(Γ,n,t)+f\colon\!\mathcal{T}(\Gamma,n,t)\to\mathbb{R}_{+}

by the formula

f(α,t)\displaystyle f(\alpha_{\bullet},t_{\bullet}) =qi1ui1(0,t1)kα1(t1)uin(tn1,tn)kαn(tn)uin+1(tn,t),\displaystyle=q_{i_{1}}u_{i_{1}}(0,t_{1})k_{\alpha_{1}}(t_{1})\cdots u_{i_{n}}(t_{n-1},t_{n})k_{\alpha_{n}}(t_{n})u_{i_{n+1}}(t_{n},t)\,,
=qi1m=1n+1uim(tm1,tm)m=1nkαm(tm)\displaystyle=q_{i_{1}}\prod_{m=1}^{n+1}u_{i_{m}}(t_{m-1},t_{m})\prod_{m=1}^{n}k_{\alpha_{m}}(t_{m})

(compare [Bellac, eqn. 1.112] in the constant rate case).222The function ff is a discrete analogue of the Onsager-Machlup Lagrangian [Onsager-Machlup].

Consider the master equation

p(t)=p(t),p(0)=q.p^{\prime}(t)=\mathbb{H}p(t),\quad p(0)=q\,.

Let

𝒫(Γ,n)\mathcal{P}(\Gamma,n)

denote the set of paths of length nn and let 𝒫i(Γ,n)𝒫(Γ,n)\mathcal{P}^{i}(\Gamma,n)\subset\mathcal{P}(\Gamma,n) denote the subset of those paths which have terminus iΓ0i\in\Gamma_{0}.

Theorem 3.1.

The formal solution to the master equation is the vector valued function p(t)p(t) whose component at iΓ0i\in\Gamma_{0} is given by the expression

pi(t)=n=0α𝒫i(Γ,n)0t0tn0t2f(α,t)𝑑t1𝑑tn.p_{i}(t)=\sum_{n=0}^{\infty}\sum_{\scriptscriptstyle\alpha_{\bullet}\in\mathcal{P}^{i}(\Gamma,n)}\int_{0}^{t}\!\!\!\int_{0}^{t_{n}}\!\!\!\cdots\!\!\!\int_{0}^{t_{2}}f(\alpha_{\bullet},t_{\bullet})\,dt_{1}\cdots dt_{n}\,.
Proof.

Write =A0+A1\mathbb{H}=A_{0}+A_{1}, where A0A_{0} is the diagonal matrix with entries hiih_{ii}. For ϵ>0\epsilon>0, set ϵ=A0+ϵA1\mathbb{H}_{\epsilon}=A_{0}+\epsilon A_{1}. Consider the equation

(2) p˙=Hϵp,p(0)=q.\dot{p}=H_{\epsilon}p,\quad p(0)=q\,.

We seek a formal solution p=p0+ϵp1+ϵ2p2+p=p^{0}+\epsilon p^{1}+\epsilon^{2}p^{2}+\cdots with p0(0)=qp^{0}(0)=q and pn(0)=0p^{n}(0)=0 for n>0n>0. Once such a solution is found, we set ϵ=1\epsilon=1 to obtain the formal solution to the master equation.

Expanding (2) in ϵ\epsilon, we obtain the linear system

(3) p˙n=A0pn+A1pn1,n=0,1,2,\dot{p}^{n}=A_{0}p^{n}+A_{1}p^{n-1}\,,\quad n=0,1,2,\dots

where by convention p1:=0p^{-1}:=0.

For iΓ0i\in\Gamma_{0}, the ii-th equation of the system is the first order linear differential equation

(4) p˙in=hiipin+jihijpjn1.\dot{p}_{i}^{n}=h_{ii}p_{i}^{n}+\sum_{j\neq i}h_{ij}p_{j}^{n-1}\,.

If n=0n=0, the system is uncoupled and separation of variables gives

pi0=qie0thii(t1)𝑑t1=qiui(0,t).p_{i}^{0}=q_{i}e^{\int_{0}^{t}h_{ii}(t_{1})dt_{1}}=q_{i}u_{i}(0,t)\,.

For n>0n>0, the solution to (4) can be iteratively solved using the integrating factor. The first iteration gives

pin(t)\displaystyle p_{i}^{n}(t) =ji0tetnthii𝑑tn1hij(tn)pjn1(tn)𝑑tn\displaystyle=\sum_{j\neq i}\int_{0}^{t}e^{\int_{t_{n}}^{t}h_{ii}\,dt_{n-1}}h_{ij}(t_{n})p_{j}^{n-1}(t_{n})\,dt_{n}
=αn0tuin+1(tn,t)kαn(tn)pinn1(tn)𝑑tn.\displaystyle=\sum_{\alpha_{n}}\int_{0}^{t}u_{i_{n+1}}(t_{n},t)k_{\alpha_{n}}(t_{n})p_{i_{n}}^{n-1}(t_{n})\,dt_{n}\,.

where the second sum is over all edges αn\alpha_{n} with terminus i=in+1i=i_{n+1} and whose source is denoted in the integrand by ini_{n}. We then repeat the procedure using pin1p_{i}^{n-1} in place of pinp_{i}^{n}, to obtain

pin(t)=α0t0tnuin+1(tn,t)kαn(tn)uin(tn1,tn)kαn1(tn1)pin1n2(tn1)𝑑tn1𝑑tn,p_{i}^{n}(t)=\sum_{\alpha_{\bullet}}\int_{0}^{t}\!\!\!\int_{0}^{t_{n}}u_{i_{n+1}}(t_{n},t)k_{\alpha_{n}}(t_{n})u_{i_{n}}(t_{n-1},t_{n})k_{\alpha_{n-1}}(t_{n-1})p_{i_{n-1}}^{n-2}(t_{n-1})\,dt_{n-1}dt_{n}\,,

where the sum is indexed over all paths of length two α=(αn1,αn)\alpha_{\bullet}=(\alpha_{n-1},\alpha_{n}) satisfying d(αn1)=(in1,in)d(\alpha_{n-1})=(i_{n-1},i_{n}), d(αn)=(in,in+1)d(\alpha_{n})=(i_{n},i_{n+1}), and in+1=ii_{n+1}=i. Applying this procedure a total of nn times results in the desired expression for pin(t)p^{n}_{i}(t). ∎

Let 𝒯(Γ,t)\mathcal{T}(\Gamma,t) denote the space of trajectories of duration tt of arbitrary length.

Corollary 3.2.

Let p(t)p(t) denote the formal solution to the master equation. Then p(t)p(t) is a probability distribution on Γ0\Gamma_{0} for every t0t\geq 0. In particular, the function ff is a probability density on 𝒯(Γ,t)\mathcal{T}(\Gamma,t).

Proof.

As the rate bound holds, there is a constant C>0C>0, independent of nn, such f(α,t)Cnf(\alpha_{\bullet},t_{\bullet})\leq C^{n} for all trajectories of length nn. As the degree bound holds, there is global bound DD on the degree function, so the number of paths of length nn terminating at a vertex ii is at most DnD^{n}. Consequently,

0α0t0tn0t2f(α,t)𝑑t1𝑑tn(CDt)nn!,0\leq\sum_{\alpha_{\bullet}}\int_{0}^{t}\!\!\!\int_{0}^{t_{n}}\!\!\!\cdots\!\!\!\int_{0}^{t_{2}}f(\alpha_{\bullet},t_{\bullet})\,dt_{1}\cdots dt_{n}\leq\frac{(CDt)^{n}}{n!}\,,

where the sum ranges over paths of length nn with terminus ii. Here, we have used the fact that tn/n!t^{n}/n! is the volume of the nn-simplex 0t1tnt0\leq t_{1}\leq\cdots t_{n}\leq t. By the comparison test, npin(t)\sum_{n}p^{n}_{i}(t) converges. Therefore p(t)=npn(t)p(t)=\sum_{n}p^{n}(t) also converges.

Let 𝟏:Γ0\mathbf{1}\colon\!\Gamma_{0}\to\mathbb{R} be the row vector which is identically one at every vertex. It will suffice to show that 𝟏p(t)=1\mathbf{1}\cdot p(t)=1. Observe that 𝟏=0\mathbf{1}\cdot\mathbb{H}=0, since the entries of \mathbb{H} in any column add to zero. Then for all tt we have

ddt(𝟏p(t))=𝟏p(t)=𝟏p(t)=0.\frac{d}{dt}(\mathbf{1}\cdot p(t))=\mathbf{1}\cdot p^{\prime}(t)=\mathbf{1}\cdot\mathbb{H}p(t)=0\,.

Consequently, 𝟏p(t)\mathbf{1}\cdot p(t) is a constant. But 𝟏p(0)=1\mathbf{1}\cdot p(0)=1, hence 𝟏p(t)=1\mathbf{1}\cdot p(t)=1 for all tt. ∎

4. Fundamental solutions

Consider the master equation with initial distribution p(0)=δip(0)=\delta_{i} for a fixed vertex ii, where δi(j)=δij\delta_{i}(j)=\delta_{ij} is the Kronecker delta function. The solution to this equation is called a fundamental solution and will be denoted by u(i,t)u(i,t).

In this case the density ff is supported on the set of trajectories (α,t)(\alpha_{\bullet},t_{\bullet}) with initial vertex ii. Let 𝒯i(Γ,t)𝒯(Γ,t)\mathcal{T}_{i}(\Gamma,t)\subset\mathcal{T}(\Gamma,t) denote the subspace of trajectories which start at the vertex ii.

Corollary 4.1.

With respect to this assumption, the function ff is a probability density on 𝒯i(Γ,t)\mathcal{T}_{i}(\Gamma,t).

Remark 4.2.

The general solution p(t)p(t) to the master equation with initial condition p(0)=qp(0)=q is obtained from the fundamental solutions using the identity

(5) p(t)=jΓ0qju(j,t).p(t)=\sum_{j\in\Gamma_{0}}q_{j}u(j,t)\,.
Definition 4.3.

Define the propagator K:Γ0×Γ0×++K\colon\!\Gamma_{0}\times\Gamma_{0}\times\mathbb{R}_{+}\to\mathbb{R}_{+} by

K(i,j,t)=uj(i,t),K(i,j,t)=u_{j}(i,t)\,,

i.e., the probability of the set of trajectories of duration tt which start at vertex ii and terminate at vertex jj. Note the initial condition K(i,j,0)=δijK(i,j,0)=\delta_{ij}.

Setting ψ(x,t):=px(t)\psi(x,t):=p_{x}(t), equation (5) becomes

ψ(y,t)=xΓ0K(x,y,t)ψ(x,0)=xΓ0K(x,y,t)ψ(x,0),\psi(y,t)=\sum_{x\in\Gamma_{0}}K(x,y,t)\psi(x,0)=\int_{x\in\Gamma_{0}}K(x,y,t)\psi(x,0)\,,

which is familiar to the physics literature. By Theorem 3.1, we obtain the path integral representation

K(x,y,t)=n=0α𝒫xy(Γ,n)0t0tn0t2f(α,t)𝑑t1𝑑tn,K(x,y,t)=\sum_{n=0}^{\infty}\sum_{\scriptscriptstyle\alpha_{\bullet}\in\mathcal{P}^{y}_{x}(\Gamma,n)}\int_{0}^{t}\!\!\!\int_{0}^{t_{n}}\!\!\!\cdots\!\!\!\int_{0}^{t_{2}}f(\alpha_{\bullet},t_{\bullet})\,dt_{1}\cdots dt_{n}\,,

where 𝒫xy(Γ,n)\mathcal{P}^{y}_{x}(\Gamma,n) is the set of paths of length nn which start at xx and terminate at yy. Furthermore, the series converges if Γ\Gamma satisfies the rate and degree bounds.

Example 4.4.

Let XX be an rr-regular graph, i.e., the number of edges meeting each vertex is rr. Assume that the rates kk_{\bullet} are constant with value one. Then the master operator \mathbb{H} is the negative of the graph Laplacian. In this instance elementary to check that f(α,t)=δiertf(\alpha_{\bullet},t_{\bullet})=\delta_{i}e^{-rt}. Let πn(i,j)\pi_{n}(i,j) be the number of paths of length nn in Γ=DX\Gamma=DX from ii to jj. Then a straightforward calculation shows

K(i,j,t)=ertn=0πn(i,j)tnn!.K(i,j,t)=e^{-rt}\sum_{n=0}^{\infty}\pi_{n}(i,j)\frac{t^{n}}{n!}\,.

In this case, KK is the combinatorial heat kernel.

As Γ\Gamma is rr-regular, there are precisely rnr^{n} paths of length nn which start at ii. Set

ϕn(i,j):=πn(i,j)rn.\phi_{n}(i,j):=\frac{\pi_{n}(i,j)}{r^{n}}\,.

Then ϕn(i,j)\phi_{n}(i,j) is the probability of the set of paths (not trajectories), which end at jj after nn-jumps, given that such paths start at ii (where the probability of jumping across an edge meeting any vertex is 1/r1/r).

Let

P(n)=(rt)nertn!.P(n)=\frac{(rt)^{n}e^{-rt}}{n!}\,.

Then PP is the Poisson probability mass function with parameter λ=rt\lambda=rt. Consequently,

K(i,j,t)=n=0ϕn(i,j)P(n)=𝔼[Φij],K(i,j,t)=\sum_{n=0}^{\infty}\phi_{n}(i,j)P(n)={\mathbb{E}}[\Phi_{ij}]\,,

is the (Poisson) expected value of the random variable Φij(n):=ϕn(i,j)\Phi_{ij}(n):=\phi_{n}(i,j). Summarizing, the continuous time random walk on an rr-regular graph with uniform rate 1/r1/r may be thought of as a discrete time random walk subordinated to a Poisson process (cf. [Feller, chap. X§7]).

References