This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

A First Look at First-Passage Processes

S. Redner [email protected] Santa Fe Institute, 1399 Hyde Park Road, Santa Fe, New Mexico 87501, USA
Abstract

These notes are based on the lectures that I gave (virtually) at the Bruneck Summer School in 2021 on first-passage processes and some applications of the basic theory. I begin by defining a first-passage process and presenting the connection between the first-passage probability and the familiar occupation probability. Some basic features of first passage on the semi-infinite line and a finite interval are then discussed, such as splitting probabilities and first-passage times. I also treat the fundamental connection between first passage and electrostatics. A number of applications of first-passage processes are then presented, including the hitting probability for a sphere in greater than two dimensions, reaction rate theory and its extension to receptors on a cell surface, first-passage inside an infinite absorbing wedge in two dimensions, stochastic hunting processes in one dimension, the survival of a diffusing particle in an expanding interval, and finally the dynamics of the classic birth-death process.

journal: Physica A

1 What is a First-Passage Process?

The first-passage probability is defined as the probability that a diffusing particle or a random walk first reaches a given site (or set of sites) at a specified time. Typical examples of first-passage processes include: fluorescence quenching, in which light emission by a fluorescent molecule stops when it reacts with a quencher; integrate-and-fire neurons, in which a neuron fires when a fluctuating voltage level first reaches a specified level; and the execution of buy/sell orders when a stock price first reaches a threshold. To appreciate why first-passage phenomena might be relevant practically, consider the following example. You are an investor who buys stock in a company at a price of $100. Suppose that this price fluctuates daily by ±$1\pm\$1. You will sell if the stock price reaches $200 and if the stock price reaches $0, the company has gone bankrupt and you’ve lost all your investment. What it the probability of doubling your investment or losing your entire investment? How long will it take before one of these two events occurs? These are the types of questions that are the purview of first-passage phenomena. Much of the material covered here is discussed more detail in this monograph [1], and in other general reviews and texts on probability theory and stochastic processes [2, 3, 4, 5].

2 First-Passage and Occupation Probabilities

Let’s start by deriving the formal relation between the first-passage probability and the familiar occupation probability. For concreteness, consider a random walk in discrete space and discrete time. We define P(𝐫,t)P(\mathbf{r},t) as the occupation probability; this is the probability that a random walk is at site 𝐫\mathbf{r} at time tt when it starts at the origin. Similarly, let F(𝐫,t)F(\mathbf{r},t) be the first-passage probability, namely, the probability that the random walk first visits 𝐫\mathbf{r} at time tt with the same initial condition. Clearly F(𝐫,t)F(\mathbf{r},t) decays more rapidly in time than P(𝐫,t)P(\mathbf{r},t) because once a random walk reaches 𝐫\mathbf{r}, there can be no further contribution to F(𝐫,t)F(\mathbf{r},t), although the same walk may still contribute to P(𝐫,t)P(\mathbf{r},t).

Refer to caption
Figure 1: Schematic diagrammatic relation between the occupation probability of a random walk (whose propagation is represented by a wavy line) and the first-passage probability (straight line).

It is convenient to write P(𝐫,t)P(\mathbf{r},t) in terms of F(𝐫,t)F(\mathbf{r},t) and then invert this relation to find F(𝐫,t)F(\mathbf{r},t). For a random walk to be located at 𝐫\mathbf{r} at time tt, the walk must first reach 𝐫\mathbf{r} at some earlier time step tt^{\prime} and then return to 𝐫\mathbf{r} after ttt-t^{\prime} additional steps (Fig. 1). This connection between F(𝐫,t)F(\mathbf{r},t) and P(𝐫,t)P(\mathbf{r},t) may be expressed as the convolution

P(𝐫,t)=δ𝐫,0δt,0+ttF(𝐫,t)P(0,tt),P(\mathbf{r},t)=\delta_{\mathbf{r},0}\delta_{t,0}+\sum_{t^{\prime}\leq t}F(\mathbf{r},t^{\prime})\,P(0,t-t^{\prime})\,, (2.1)

where δt,0\delta_{t,0} is the Kronecker delta function. This equation expresses the fact that if a random walk is at 𝐫\mathbf{r} at time tt, it must have first reached 𝐫\mathbf{r} at some earlier time tt^{\prime} (which could even be tt). If the walk reached 𝐫\mathbf{r} at a time earlier than tt, then it must return to 𝐫\mathbf{r} (and any number of such returns could occur) in the remaining time ttt-t^{\prime}. The probability for this set of events is expressed by P(0,tt)P(0,t-t^{\prime}). The delta function term accounts for the initial condition that the walk starts at 𝐫=0\mathbf{r}=0.

The above convolution is most conveniently solved by introducing the generating functions,

P(𝐫,z)=t=0P(𝐫,t)zt,F(𝐫,z)=t=0F(𝐫,t)zt.\displaystyle P(\mathbf{r},z)=\sum_{t=0}^{\infty}P(\mathbf{r},t)z^{t}\,,\qquad F(\mathbf{r},z)=\sum_{t=0}^{\infty}F(\mathbf{r},t)z^{t}\,.

For a random walk in continuous time, we would merely replace the sum over discrete time in Eq. (2.1) by an integral and then use the Laplace transform. However, the asymptotic results would be identical. To solve for the first-passage probability, we multiply Eq. (2.1) by ztz^{t} and sum over all tt. We thereby find that the generating functions for PP and FF are related by

P(𝐫,z)=δ𝐫,0+F(𝐫,z)P(0,z).P(\mathbf{r},z)=\delta_{\mathbf{r},0}+F(\mathbf{r},z)P(0,z)\,. (2.2)

Thus we obtain the fundamental connection between the generating functions

F(𝐫,z)={P(𝐫,z)P(0,z)𝐫0,11P(0,z),𝐫=0.F(\mathbf{r},z)=\begin{cases}{\displaystyle\frac{P(\mathbf{r},z)}{P(0,z)}}&\mathbf{r}\neq 0\,,\\[11.38109pt] {\displaystyle 1-\frac{1}{P(0,z)}},&\mathbf{r}=0\,.\end{cases} (2.3)

The important point is that the first-passage probability can be determined solely from the occupation probability. Many profound results about random walks in infinite space can be obtained from the fundamental relation (2.3) (see, e.g., [6, 7, 8, 9]). Our focus here will be on random walks or diffusion in confined geometries that reflect important physical constraints.

3 The Half Line

Suppose that a diffusing particle starts at x0>0x_{0}>0 on the infinite half line [0,][0,\infty] and is absorbed when it reaches the origin. Does the particle ever reach the origin? If so, when does this particle first reach the origin? To answer these questions, we have to solve the diffusion equation for the concentration c(x,t)c(x,t), subject to the initial condition c(x,t=0)=δ(xx0)c(x,t\!=\!0)=\delta(x-x_{0}), and the boundary condition c(x=0,t>0)=0c(x\!=\!0,t\!>\!0)=0; the latter enforces the absorption of the particle when it reaches the origin.

A standard approach to solve this problem is to first take the Laplace transform of the diffusion equation and then solve for the Green’s function of this transformed equation. Then one inverts the Laplace transform of the Green’s function to obtain the concentration c(x,t)c(x,t) in the time domain. The diffusive flux to the origin gives the probability that the particle reaches the origin at time tt. Because of the absorbing boundary condition, when the particle does reach the origin, it is removed from the system. Thus the diffusive flux corresponds to the probability for the particle to reach the origin for the first time—namely, the first-passage probability to the origin.

Refer to caption
Figure 2: Concentration profile of a diffusing particle on the absorbing infinite half line (0,)(0,\infty) at Dt=10Dt=10, with x0=2x_{0}=2 (solid curve). Also shown are the component Gaussian (blue dashed) and image anti-Gaussian (red dot-dash) and their superposition in the physical region x>0x>0.

A more fun way to solve this problem is by invoking the familiar image method from electrostatics. In this method, a diffusing particle that starts at x0>0x_{0}>0 and is subject to an absorbing boundary condition at x=0x=0 is equivalent to removing the boundary altogether and introducing an image “antiparticle” that is initially at x=x0x=-x_{0}. Because the concentration is clearly equal to zero at the origin by symmetry, this image antiparticle effectively imposes the absorbing boundary condition c(x=0,t)=0c(x\!=\!0,t)=0. The initial particle and the image antiparticle both diffuse freely on [,][-\infty,\infty] and their superposition gives a resultant concentration c(x,t)c(x,t) for x>0x>0 that solves the original problem.

Hence the concentration for a diffusing particle on the positive half-line is the sum of a Gaussian centered at x0x_{0} and an anti-Gaussian centered at x0-x_{0}:

c(x,t)=14πDt[e(xx0)2/4Dte(x+x0)2/4Dt].c(x,t)=\frac{1}{\sqrt{4\pi Dt}}\left[e^{-(x-x_{0})^{2}/4Dt}-e^{-(x+x_{0})^{2}/4Dt}\right]. (3.1)

This concentration profile has a linear dependence on xx near the origin and a Gaussian tail for x/Dt1x/\sqrt{Dt}\gg 1, as illustrated in Fig. 2. Because the initial condition is normalized, the first-passage probability to the origin at time tt is just the diffusive flux to this point. From the above expression for c(x,t)c(x,t), we find

F(0,t)=+Dc(x,t)x|x=0=x04πDt3ex02/4Dtt.\displaystyle F(0,t)=+D\frac{\partial c(x,t)}{\partial x}\bigg{|}_{x=0}=\frac{x_{0}}{\sqrt{4\pi Dt^{3}}}\,\,e^{-x_{0}^{2}/4Dt}\qquad t\to\infty\,. (3.2)

This fundamental and simple formula has a number of striking implications:

  1. 1.

    The particle is sure to reach to the origin because 0F(0,t)𝑑t=1\int_{0}^{\infty}F(0,t)\,dt=1. That is, eventual absorption is certain.

  2. 2.

    The average time for the particle to reach the origin is infinite! This fact arises because the first-passage probability has the long-time algebraic tail F(0,t)x0/4πDt3F(0,t)\to x_{0}/\sqrt{4\pi Dt^{3}} for tx02/Dt\gg x_{0}^{2}/D. This dichotomy between hitting the origin with certainty but taking an infinite average time to do so underlies many of the intriguing features of one-dimensional diffusion.

  3. 3.

    The typical time to reach the origin is finite. We can define the term typical time in a precise way as follows: As a preliminary, define the typical position of the particle, xTx_{T}, as

    xTc(x,t)𝑑x=12.\displaystyle\int_{x_{T}}^{\infty}c(x,t)\,dx=\tfrac{1}{2}\,. (3.3)

    That is, one half of the total probability lies in the range beyond xTx_{T} and one half lies in the range [0,xT][0,x_{T}]. Substituting the concentration profile (3.1) into the above integral and performing the integral leads to the transcendental equation

    erfc(xTx04Dt)erfc(xT+x04Dt)=1.\displaystyle\text{erfc}\left(\frac{x_{T}-x_{0}}{\sqrt{4Dt}}\right)-\text{erfc}\left(\frac{x_{T}+x_{0}}{\sqrt{4Dt}}\right)=1\,. (3.4)

    This equation can only be solved numerically and the salient feature is that xTx_{T} monotonically decreases with time and reaches zero at a time that is roughly 1.1x02/D1.1\,x_{0}^{2}/D; this defines the typical hitting time.

  4. 4.

    Even though the average time to reach the origin is infinite, the number of times the origin is reached in a time tt is proportional to t\sqrt{t}. This result follows directly from the probability distribution of a freely diffusing particle. At large times, the bulk of Gaussian distribution for a particle that starts at x0x_{0} will spread past the origin. Each site that is within the Gaussian envelope will have been visited of the order of t\sqrt{t} times. This last fact will be relevant for our discussion of the reaction rate of the sphere in Sec. 6.

4 The Finite Interval

We now turn to the first-passage properties in a finite interval. The reason for focusing on the finite interval is that the basic first-passage questions in this geometry have many profound implications. Moreover, the interval geometry is sufficiently simple that many results can be readily derived. Let us begin by outlining the basic questions that we will address. We consider a diffusing particle that starts at some point x0x_{0} within the interval [0,L][0,L], with absorbing boundary conditions at both ends of the interval. Eventually the particle is absorbed, and our goal is to characterize the time dependence of this absorption. Basic first-passage questions include:

  1. 1.

    What is the time dependence of the survival probability S(t)S(t)? This is the probability that a diffusing particle does not touch either absorbing boundary before time tt.

  2. 2.

    What is the time dependence of the first-passage, or exit probabilities, to either 0 or to LL as a function of x0x_{0}? Integrating these probabilities over all time gives the eventual hitting, or splitting probability to a specified boundary. What is the dependence of the splitting probability to 0 or to LL as a function of the starting position?

  3. 3.

    What is the average exit time, that is, the average time until the particle hits either of the absorbing boundaries as a function of starting position? What are the conditional exit times, that is, the average time to hit a specified boundary (without ever touching the other boundary) as a function of the starting position?

To answer the first question, we need to solve the diffusion equation in the interval, subject to the initial condition that a particle starts at x0x_{0}, and with absorbing boundary conditions at both ends. This is a standard exercise and the result for the concentration is

c(x,t)=n12Lsinnπx0LsinnπxLen2π2Dt/L2.c(x,t)=\sum_{n\geq 1}^{\infty}\frac{2}{L}\sin\frac{n\pi x_{0}}{L}\,\sin\frac{n\pi x}{L}\;e^{-n^{2}\pi^{2}\,Dt/L^{2}}~{}. (4.1)

Since the large-nn eigenmodes decay more rapidly in time, the most slowly decaying eigenmode dominates in the long-time limit. As a result, the survival probability, which is the spatial integral of the concentration over the interval, asymptotically decays as

S(t)eDπ2t/L2.S(t)\sim e^{-D\pi^{2}t/L^{2}}\,. (4.2)

For answering the second question, it is mathematically simpler to work in the Laplace transform domain. Applying the Laplace transform to the diffusion equation recasts it as

sc(x,s)c(x,t=0)=Dc′′(x,s),sc(x,s)-c(x,t=0)=Dc^{\prime\prime}(x,s)\,, (4.3)

where the prime denotes differentiation with respect to xx. The argument ss indicates that c(x,s)c(x,s) is the Laplace transform. Within the standard Green’s function approach, the homogeneous equation in each subdomain (0,x0)(0,x_{0}) and (x0,L)(x_{0},L) has elementary solutions of the form c(x,s)=Aexp(xs/D)+Bexp(xs/D)c(x,s)=A\exp(x\sqrt{{s/D}})+B\exp(-x\sqrt{{s/D}}), with the constants AA and BB determined by the boundary conditions. Because the absorbing boundary condition at x=0x=0 mandates an antisymmetric combination of exponentials and because the form of the Green’s function as xLx\to L must be identical to that as x0x\to 0, we can be immediately write

c<(x,s)=Asinh(sDx)x<x0,c>(x,s)=Bsinh(sD(Lx))x>x0,\displaystyle\begin{split}c_{<}(x,s)&=A\sinh\left(\sqrt{\frac{s}{D}}\,x\right)\hskip 36.135ptx<x_{0}\,,\\[5.69054pt] c_{>}(x,s)&=B\sinh\left(\sqrt{\frac{s}{D}}(L-x)\right)\quad x>x_{0}\,,\end{split} (4.4)

for the subdomain Green’s functions c<c_{<} and c>c_{>}, where AA and BB are constants.

We now impose the continuity condition c<(x0,s)=c>(x0,s)c_{<}(x_{0},s)=c_{>}(x_{0},s) and the jump condition that is obtained by integrating Eq. (4.3) over an infinitesimal interval that includes x0x_{0}:

c(x,s)|x=x0+c(x,s)|x=x0=1/D,c^{\prime}(x,s)\big{|}_{x=x_{0}^{+}}-c^{\prime}(x,s)\big{|}_{x=x_{0}^{-}}=-1/D\,,

to finally obtain

c(x,s)=sinh(sDx<)sinh(sD(Lx>))sDsinh(sDL).c(x,s)=\frac{\displaystyle\sinh\left(\sqrt{\frac{s}{D}}\,x_{<}\right)\sinh\left(\sqrt{\frac{s}{D}}(L-x_{>})\right)}{\displaystyle\sqrt{sD}\sinh\left(\sqrt{\frac{s}{D}}\,L\right)}\,. (4.5)

From this Green’s function, the Laplace transform of the fluxes at x=0x=0 and at x=Lx=L are

j(s|x0)\displaystyle j_{-}(s|x_{0}) +Dc(x,s)x|x=0=sinh(sD(Lx0))/sinh(sDL).\displaystyle\equiv+D\frac{\partial c(x,s)}{\partial x}\Bigg{|}_{x=0}={\displaystyle\sinh\left({\sqrt{\frac{s}{D}}}(L-x_{0})\right)}\bigg{/}{\displaystyle\sinh\left({\sqrt{\frac{s}{D}}}\,L\right)}\,. (4.6a)
j+(s|x0)\displaystyle j_{+}(s|x_{0}) Dc(x,s)x|x=L=sinh(sDx0)/sinh(sDL).\displaystyle\equiv-D\frac{\partial c(x,s)}{\partial x}\Bigg{|}_{x=L}={\displaystyle\sinh\left(\sqrt{\frac{s}{D}}\,x_{0}\right)}\bigg{/}{\displaystyle\sinh\left(\sqrt{\frac{s}{D}}\,L\right)}\,. (4.6b)

The subsidiary argument x0x_{0} in jj emphasizes that the flux depends on the initial particle position. Since the initial condition is normalized, the magnitude of the flux to each boundary is identical to the respective first-passage probabilities.

For s=0s=0, these Laplace transforms are just the time-integrated first-passage probabilities to 0 and at LL. These quantities therefore coincide with the respective splitting probabilities, (x0){\mathcal{E}}_{-}(x_{0}) and +(x0){\mathcal{E}}_{+}(x_{0}), namely, the probabilities to eventually hit the left and the right ends of the interval as a function of the initial position x0x_{0}:

(x0)=j(s=0|x0)=1x0L,+(x0)=j+(s=0|x0)=x0L.\displaystyle\begin{split}{\mathcal{E}}_{-}(x_{0})&=j_{-}(s\!=\!0|x_{0})=1-\frac{x_{0}}{L}\,,\\[5.69054pt] {\mathcal{E}}_{+}(x_{0})&=j_{+}(s\!=\!0|x_{0})=\frac{x_{0}}{L}\,.\end{split} (4.7)

Thus the splitting probabilities are given by an amazingly simple formula—the probability of reaching one endpoint is just the fractional distance to the other endpoint!

It is instructive to also derive these splitting probabilities by the backward Kolmogorov approach [1, 2]. The word backward reflects the feature that the initial condition becomes the dependent variable, rather than the current position of the particle. As we shall see, this method provides a powerful tool for determining first-passage properties. Physically, we obtain the eventual hitting probability +(x0){\mathcal{E}}_{+}(x_{0}) to the right boundary by summing the probabilities for all paths that start at x0x_{0} and reach LL without touching 0. Thus

+(x0)=pathsΠx0L,{\mathcal{E}}_{+}(x_{0})=\sum_{\text{paths}}\Pi_{x_{0}\to L}\,, (4.8)

where Πx0L\Pi_{x_{0}\to L} denotes the probability of a path from x0x_{0} to LL that avoids 0. As illustrated in Fig. 3, the sum over all such paths can be decomposed into the outcome after one step and the sum over all path remainders from the intermediate point xx^{\prime} to LL. This gives

+(x0)=12pathsΠx0+δxL+12paths′′Πx0δxL=12[+(x0+δx)++(x0δx)].\displaystyle\mathcal{E}_{+}(x_{0})=\tfrac{1}{2}\sum_{\text{paths}^{\prime}}\Pi_{x_{0}+\delta x\to L}+\tfrac{1}{2}\,\sum_{\text{paths}^{\prime\prime}}\Pi_{x_{0}-\delta x\to L}=\tfrac{1}{2}[{\mathcal{E}}_{+}(x_{0}\!+\!\delta x)+{\mathcal{E}}_{+}(x_{0}\!-\!\delta x)]\,. (4.9)

Here δx\delta x is the length of a single random-walk step and paths and paths′′ indicate, respectively, all paths that start at x0±δxx_{0}\pm\delta x and reach LL without touching 0.

Refer to caption
Figure 3: Schematic decomposition of a random walk path from x0x_{0} to LL into the outcome after one step (red) and the remainder from xx^{\prime} to LL. The factors 1/21/2 account for the probabilities associated with the first step of the decomposed paths.

Equation (4.9) reduces to Δ(2)±(x0)=0\Delta^{(2)}{\mathcal{E}}_{\pm}(x_{0})=0, where Δ(2)\Delta^{(2)} is the discrete second-difference operator, Δ(2)f(x)f(xδx)2f(x)+f(x+δx)\Delta^{(2)}f(x)\equiv f(x-\delta x)-2f(x)+f(x+\delta x). This difference equation is subject to the boundary conditions +(0)=0{\mathcal{E}}_{+}(0)=0, +(L)=1{\mathcal{E}}_{+}(L)=1. The solution is simply

+(x0)=xL,\displaystyle{\mathcal{E}}_{+}(x_{0})=\frac{x}{L}\,, (4.10)

and correspondingly (x)=1xL{\mathcal{E}}_{-}(x)=1-\frac{x}{L}.

It is worth mentioning that there is an even simpler way to determine the exit probabilities by the martingale method [10]. This solution relies on the fact that the motion of the random walk in the interval is a “fair game” at any time. Thus the average position of the walk is time independent. Formally, a martingale is a process in which the average value of a random variable at time t+δtt+\delta t equals the average value of this variable at time tt.

For the present example, at t=0t=0, the average position of the particle is x=x0\langle x\rangle=x_{0}. At infinite time, the walk is either at the left end or the right end of the interval, with respective probabilities +(x0){\mathcal{E}}_{+}(x_{0}) or (x0){\mathcal{E}}_{-}(x_{0}). Thus the average position at infinite time is x=0×(x0)+L×+(x0)\langle x\rangle=0\times{\mathcal{E}}_{-}(x_{0})+L\times{\mathcal{E}}_{+}(x_{0}). Since the initial average position equals the final average position, we immediately recover (4.10).

For a biased random walk with a probability pp of hopping to the right and probability qq of hopping to the left, the analog of Eq. (4.9) for the splitting probability is

(x0)=p+(x0+δx)+q+(x0δx),\displaystyle{\mathcal{E}}(x_{0})=p{\mathcal{E}}_{+}(x_{0}\!+\!\delta x)+q{\mathcal{E}}_{+}(x_{0}\!-\!\delta x)\,, (4.11)

with solution

+(x0)=1evx0/D1evL/D1eu0Pe1ePe,\displaystyle{\mathcal{E}}_{+}(x_{0})=\frac{1-e^{-vx_{0}/D}}{1-e^{-vL/D}}\equiv\frac{1-e^{-u_{0}P\!e}}{1-e^{-P\!e}}\,, (4.12)

where v=(pq)δxv=(p-q)\delta x, D=δx2/2D=\delta x^{2}/2, u0=x0/Lu_{0}=x_{0}/L, and Pe=vL/DP\!e=vL/D is the Péclet number, which is a dimensionless measure of the influence of the bias relative to diffusive fluctuations. When the Péclet number is large, the exit probability clearly reflects the strong influence of the bias (Fig. 4).

Refer to caption
Figure 4: Dependence of the exit probability to x=Lx=L on u0=x0/Lu_{0}=x_{0}/L in the interval [0,L][0,L] for various Péclet numbers PeP\!e.

We now extend the backward Kolmogorov approach to determine the average exit time from the finite interval. We distinguish between the unconditional average exit time, namely, the average time for a particle to reach either end of the interval, and the conditional average exit time, namely, the average time for a particle to reach, say, the right end of the interval without ever touching the other end.

In close analogy with (4.8), the unconditional exit time satisfies

t(x0)=paths(Πt)x0±L.t(x_{0})=\sum_{\text{paths}}(\Pi t)_{x_{0}\to\pm L}\,. (4.13)

That is, to compute the average unconditional exit time, we take the time for a path to go from x0x_{0} to ±L\pm L times the probability of this path and sum over all possible paths. Using the same decomposition that which led to (4.9), the unconditional exit time satisfies

t(x0)\displaystyle t(x_{0}) =12pathsΠx0+δx±L(tx0+dx±L+dt)+12paths′′Πx0δx±L(tx0dx±L+dt).\displaystyle=\tfrac{1}{2}\!\!\sum_{\text{paths}^{\prime}}\!\!\Pi_{x_{0}+\delta x\to\pm L}(t_{x_{0}+dx\to\pm L}\!+\!dt)+\tfrac{1}{2}\!\!\sum_{\text{paths}^{\prime\prime}}\!\!\Pi_{x_{0}-\delta x\to\pm L}(t_{x_{0}-dx\to\pm L}\!+\!dt)\,. (4.14)

Notice that the term

12pathsΠx0+δx±Ltx0+dx±L\displaystyle\tfrac{1}{2}\!\!\sum_{\text{paths}^{\prime}}\!\!\Pi_{x_{0}+\delta x\to\pm L}\;t_{x_{0}+dx\to\pm L}

has exactly the same form as (4.13), so that the above expression is merely 12t(x0+δx)\frac{1}{2}t(x_{0}+\delta x). A similar identification holds for the analogous term

12paths′′Πx0δx±Ltx0dx±L.\displaystyle\tfrac{1}{2}\!\!\sum_{\text{paths}^{\prime\prime}}\!\!\Pi_{x_{0}-\delta x\to\pm L}\;t_{x_{0}-dx\to\pm L}\,.

Finally, the terms multiplying the factor dtdt in (4.14) just gives the probability of all possible paths from x0x_{0} to either end of the interval; this probability is clearly equal to 1. Thus we have

t(x0)=12[t(x0+δx)+t(x0δx)]+dt.\displaystyle t(x_{0})=\tfrac{1}{2}[t(x_{0}\!+\!\delta x)+t(x_{0}\!-\!\delta x)]+dt\,. (4.15)

In the continuum limit, we expand (4.15) in a Taylor series to second order in δx\delta x to give

δx22t′′=dt.\displaystyle\frac{\delta x^{2}}{2}t^{\prime\prime}=-dt\,.

Now identifying δx2/(2dt)\delta x^{2}/(2dt) as the diffusion coefficient DD, the equation for the unconditional exit time reduces to Dt′′=1D\,t^{\prime\prime}=-1, subject to the boundary conditions t(0)=t(L)=0t(0)=t(L)=0. The solution is (see Fig. 5)

t(x0)=x0(Lx0)2D.t(x_{0})=\frac{x_{0}(L-x_{0})}{2D}\,. (4.16)

Notice that this exit time is of the order of LL for a particle that starts a distance of the order of one from an absorbing boundary and of the order of L2L^{2} for a particle that starts near the middle of the interval.

Refer to caption
Figure 5: Unconditional and conditional average exit times from the finite interval [0,L][0,L], normalized by L2/2DL^{2}/2D, as a function of the dimensionless initial position x0/Lx_{0}/L.

Now let’s calculate the conditional exit time to the right boundary when starting from x0x_{0}, t+(x0)t_{+}(x_{0}). By definition this conditional time is

t+(x0)=paths+(Πt)x0Lpaths+Πx0L=paths+(Πt)x0L+(x0).\displaystyle t_{+}(x_{0})=\frac{\sum_{\text{paths}_{+}}(\Pi t)_{x_{0}\to L}}{\sum_{\text{paths}_{+}}\Pi_{x_{0}\to L}}=\frac{\sum_{\text{paths}_{+}}(\Pi t)_{x_{0}\to L}}{{\mathcal{E}}_{+}(x_{0})}\,. (4.17)

That is, the conditional exit time is the time for a path to start at x0x_{0} and reach LL without touching 0 multiplies by the probability of this path, summed over all such allowable paths. Here the subscript + on the word paths indicates that only paths that go from x0x_{0} to LL without touching 0 are included. Since the total probability of all these restricted paths is less than 1, we need to divide by this total probability, +(x0){\mathcal{E}}_{+}(x_{0}), to obtain the properly normalized conditional exit time. Thus, for the quantity +(x0)t+(x0){\mathcal{E}}_{+}(x_{0})t_{+}(x_{0}), we have

+(x0)t+(x0)\displaystyle{\mathcal{E}}_{+}(x_{0})t_{+}(x_{0}) =paths+(Πt)x0L\displaystyle=\sum_{\text{paths}_{+}}(\Pi t)_{x_{0}\to L}
=12paths+Π(tx0+dxL+dt)+12paths+′′Π(tx0dxL+dt)\displaystyle=\tfrac{1}{2}\!\!\sum_{\text{paths}^{\prime}_{+}}\Pi(t_{x_{0}+dx\to L}+dt)+\tfrac{1}{2}\!\!\sum_{\text{paths}^{\prime\prime}_{+}}\Pi(t_{x_{0}-dx\to L}+dt)
=12[+(x0+dx)t+(x0+dx)]+12[+(x0dx)t+(x0dx)]\displaystyle=\tfrac{1}{2}\left[{\mathcal{E}}_{+}(x_{0}\!+\!dx)t_{+}(x_{0}\!+\!dx)\right]+\tfrac{1}{2}\left[{\mathcal{E}}_{+}(x_{0}\!-\!dx)t_{+}(x_{0}\!-\!dx)\right]
++(x0)dt.\displaystyle\hskip 158.99377pt+{\mathcal{E}}_{+}(x_{0})\,dt\,. (4.18)

In the continuum limit, the above equation reduces to D(+t+)′′=+D({\mathcal{E}}_{+}t_{+})^{\prime\prime}=-{\mathcal{E}}_{+}, with solution

t+(x0)=L2x026D.\displaystyle t_{+}(x_{0})=\frac{L^{2}-x_{0}^{2}}{6D}\,. (4.19)

From this result, we also immediately find t(x0)=t+(Lx0)t_{-}(x_{0})=t_{+}(L-x_{0}). The dependences of all the exit times on the starting position are illustrated in Fig. 5.

5 Connection with Electrostatics

One of the alluring features of first-passage processes is its intimate connection to electrostatics. By this connection, one can recast an electrostatic problem in a given geometry as a first-passage problem in the same geometry. With this perspective, it is possible to solve seemingly difficult first-passage problems in a simple way by this electrostatic connection.

To illustrate the basic principle, consider the following general problem. Suppose that a diffusing particle starts at some point 𝐫0\mathbf{r}_{0} inside an arbitrary bounded domain. At the boundary of this domain, the particle is absorbed. Eventually, all of the initial probability is absorbed on the boundary and we ask: what is the exit probability at some arbitrary point 𝐫B\mathbf{r}_{B} on the domain boundary? Formally, we have to solve the diffusion equation in this domain, subject to the appropriate initial and boundary conditions:

{c(𝐫,t)t=D2c(𝐫,t)c(𝐫,t=0)=δ(𝐫𝐫0)c(𝐫,t)|𝐫B=0\displaystyle\begin{cases}\displaystyle{\frac{\partial c(\mathbf{r},t)}{\partial t}}=D\nabla^{2}c(\mathbf{r},t)\\[5.69054pt] c(\mathbf{r},t=0)=\delta(\mathbf{r}-\mathbf{r}_{0})\\[2.84526pt] c(\mathbf{r},t)\big{|}_{\mathbf{r}_{B}}=0\end{cases} (5.1)

Then the exit probability at the boundary point 𝐫B\mathbf{r}_{B} is given by

(𝐫B)=D0c(𝐫,t)𝐧^|𝐫Bdt,\displaystyle\mathcal{E}(\mathbf{r}_{B})=-D\int_{0}^{\infty}\frac{\partial c(\mathbf{r},t)}{\partial\hat{\mathbf{n}}}\bigg{|}_{\mathbf{r}_{B}}\,dt\,, (5.2)

where 𝐧^\hat{\mathbf{n}} is the outward normal to the surface of the domain at 𝐫B\mathbf{r}_{B}.

Let’s look critically at this calculation. We are attempting to solve a partial differential equation in some domain (which may well be difficult), then take the result of this calculation and integrate over all time. That is, we really don’t need that exit probability at all times, but merely the time integral of the exit probability. This observation suggests that it will be useful to take the original problem (5.1) and integrate it over all time. To simplify what emerges, we also define the time integrated concentration, 𝒞(𝐫)0c(𝐫,t)𝑑t\mathcal{C}(\mathbf{r})\equiv\int_{0}^{\infty}c(\mathbf{r},t)\,dt. Performing this time integration on Eq. (5.1) leads to

{δ(𝐫𝐫0)=D2𝒞(𝐫)𝒞(𝐫)|𝐫B=0\displaystyle\begin{cases}-\delta(\mathbf{r}-\mathbf{r}_{0})=D\nabla^{2}\mathcal{C}(\mathbf{r})\\ \mathcal{C}(\mathbf{r})\big{|}_{\mathbf{r}_{B}}=0\end{cases}

The delta function on the left-hand side is what remains when we integrate the time derivative in (5.1) over all time. At t=t=\infty the concentration is zero, while at t=0t=0, we merely have the initial condition. But notice that in terms of the time-integrated concentration, the exit probability may be written as

(𝐫B)=D𝒞(𝐫,t)𝐧^|𝐫B.\displaystyle\mathcal{E}(\mathbf{r}_{B})=-D\frac{\partial\mathcal{C}(\mathbf{r},t)}{\partial\hat{\mathbf{n}}}\bigg{|}_{\mathbf{r}_{B}}\,. (5.3)

Thus we arrive at the fundamental result:

(𝐫B)\displaystyle\mathcal{E}(\mathbf{r}_{B}) =the electric field at 𝐫B on the surface of a grounded conductor,\displaystyle=\text{the\ electric\ field\ at\ }\mathbf{r}_{B}\text{\ on the surface of a grounded conductor,}
when a point charge of magnitude 1/(DΩd) is placed at 𝐫0.\displaystyle\hskip 14.45377pt\text{when a point charge of magnitude\ }1/(D\Omega_{d})\text{\ is placed at\ }\mathbf{r}_{0}\,.

Here Ωd\Omega_{d} is the surface area of a dd-dimensional sphere; this factor is needed to convert the prefactor DD in the diffusion equation to the correct prefactor in the Laplace equation. Thus a given first-passage problem can be expressed as an equivalent electrostatic problem in the same geometry.

6 Hitting a Sphere and Reaction Rate Theory

What is the probability H(r)H(r) that a diffusing particle eventually hits a sphere of radius aa, when the particle starts at a distance r>ar>a from the origin? One can determine this hitting probability in the standard way by solving the diffusion equation exterior to the sphere, computing the flux to the sphere, and then integrating over all time. However, it is much simpler to use the connection between first passage and electrostatics. Indeed, by a direct extension of Eq. (4.9) to three dimensions, the hitting probability satisfies (here for a discrete random walk in Cartesian coordinates for simplicity)

H(x,y)\displaystyle H(x,y) =16[H(x+1,y,z)+H(x1,y,z)+H(x,y+1,z)\displaystyle=\tfrac{1}{6}\big{[}H(x+1,y,z)+H(x-1,y,z)+H(x,y+1,z)
+H(x,y1,z)+H(x,y,z+1)+H(x,y,z1)].\displaystyle\hskip 36.135pt+H(x,y-1,z)+H(x,y,z+1)+H(x,y,z-1)\big{]}\,. (6.1)

Let us now take the continuum limit so that we can work in spherical coordinates. Then the above discrete difference equation becomes 2H=0\nabla^{2}H=0, subject to the boundary conditions H(r=a)=1H(r\!=\!a)=1 and H()=0H(\infty)=0. The solution is

H(r)=ar.\displaystyle H(r)=\frac{a}{r}\,. (6.2)

Amazingly simple!

Now let’s treat a related problem that is fundamental in chemical kinetics. Suppose that there is an initially uniform concentration of particles exterior to an absorbing sphere of radius aa. A fundamental kinetic characteristic of the absorbing sphere is its reaction rate kk, namely, the efficiency at which this sphere captures particles. Formally, kk is defined as the number of particles absorbed per unit time divided by the initial concentration. This normalization ensures that the reaction rate is a quantity that is intrinsic to the system. By dimensional analysis, the reaction rate as defined above has units of (length)/d{}^{d}/time. Moreover the reaction rate can only be a function of intrinsic parameters of the system, namely, the diffusion coefficient DD and the sphere radius aa. Since the units of DD are length2/time, we infer that kk must have units of Dad2D\,a^{d-2}. Thus the reaction rate k=ADad2k=A\,D\,a^{d-2}, where AA is a constant of the order of 1. This example shows the power of dimensional analysis in obtaining a non-trivial physical property of multi-particle system.

This simple result has some surprising implications. First, in three dimensions, the reaction rate is linear in aa; it is not proportional to the cross-sectional area of the sphere. Second for d<2d<2, the reaction rate increases when the radius of the absorbing sphere is decreased! This nonsensical result indicates that there is a basic problem with classic chemical kinetics for d<2d<2. This pathology arises because the concentration field exterior to the absorber never reaches a steady state for d<2d<2. Instead, the absorption rate of the sphere is time dependent. In contrast, for d>2d>2, a steady state concentration field does arise. Once this steady state is reached, it is a trivial exercise to compute the steady-state density by solving the Laplace equation rather than the diffusion equation, and thereby obtain the steady-state flux to the sphere. From these steps, one finds the exact reaction rate

k=4πDa.\displaystyle k=4\pi Da\,. (6.3)

Armed with the basic result for the reaction rate, we now turn to a much more profound problem that is fundamental to living systems, namely, how many receptors should there be on the surface of a cell? One might expect that much of the cell surface should be covered by receptors so that its detection efficiency is high. On the other hand, one can imagine that receptors are complex and evolutionarily expensive machines. Based on cost considerations only, it would be advantageous for a cell to minimize the number of receptors. What is the appropriate balance between these two competing attributes? As a first step to address this question, we want to compute the reaction rate of a cell that is sensitive to its environment only at the locations of the receptors. We model the cell as a sphere of radius aa in which most of the surface is reflecting. However, on the sphere surface there are also NN circular domains of radius ss that are absorbing (Fig. 6). We view these NN absorbing circles as the receptors on the cell surface. What is the reaction rate of this toy model of a cell? If most of the sphere surface is reflecting, one might anticipate that the reaction efficiency of the cell will be poor. Surprisingly, the reaction efficiency of the sphere with NN absorbing receptors is almost as good as a perfectly absorbing sphere, even when the area fraction covered by the receptors is vanishingly small! This realization is the brilliant insight of the article by Berg and Purcell that was far ahead of its time [11]. Here I outline their argument.

Refer to caption
Figure 6: Cartoon of a cell surface with 4 receptors (gray ovals). A diffusing particle first hits the cell (green trajectory) and then makes 3 non-independent hits of the surface (blue), before an independent hit (red). The particle lands on a receptor after one more non-independent hit.

The first step in their argument relies on the feature that if a diffusing particle hits the surface of a sphere, it will hit again many times before diffusing away; this point was discussed above in Sec. 3. There are two types of subsequent hitting events: (i) the particle rises a distance less than ss above the surface before hitting it again, and (ii) the particle rises a distance greater than ss above the surface before hitting it again (Fig. 6). In the former case, if the particle initially misses a receptor, it will likely miss upon the second encounter. In the latter case, if the particle misses a receptor initially, we have no information about whether the second encounter will hit or miss a receptor. Thus the rise distance ss demarcates the regime of dependent subsequent hits and independent subsequent hits. It is the latter events that are relevant for estimating the reaction rate. We thus use the height ss as the criterion for determining the number of times that a particle independently hits the cell surface before diffusing away.

When a particle is a distance ss above the cell surface, the probability that it eventually hits the cell again is, using Eq. (6.2),

ps=aa+s.\displaystyle p_{s}=\frac{a}{a+s}\,. (6.4a)
Thus the probability that the diffusing particle independently hits the cell nn times before diffusing away is psn(1ps)p_{s}^{n}(1-p_{s}). Correspondingly, the average number of independent hits to the surface is
n=n>0npsn(1ps)=(1ps)1=a+ssas.\displaystyle\langle n\rangle=\sum_{n>0}np_{s}^{n}\;(1-p_{s})=(1-p_{s})^{-1}=\frac{a+s}{s}\sim\frac{a}{s}\,. (6.4b)

The probability for a diffusing particle to not land on a receptor in a single independent hitting event is

β=1Nπs24πa2,\displaystyle\beta=1-\frac{N\pi s^{2}}{4\pi a^{2}}\,, (6.4c)

namely, the area fraction of the surface that is not covered by receptors. Thus the probability pescapep_{\rm escape} that a diffusing particle that reaches the cell but never lands on a receptor is the probability that the particle always misses a receptor in each of its independent hitting attempts. This is

pescape=n>0(βps)n(1ps)=1ps1βps=4a4a+Ns.\displaystyle p_{\rm escape}=\sum_{n>0}(\beta p_{s})^{n}(1-p_{s})=\frac{1-p_{s}}{1-\beta p_{s}}=\frac{4a}{4a+Ns}\,. (6.4d)

The probability that a diffusing particle ultimately hits a receptor thus is

phit=1pescape=Ns4a+Ns.\displaystyle p_{\rm hit}=1-p_{\rm escape}=\frac{Ns}{4a+Ns}\,. (6.4e)

The final result for the reaction rate of a cell of radius aa that is covered by NN receptors, each of radius ss, is

k=4πDa×Ns4a+Ns4πDa×efficiency.\displaystyle k=4\pi Da\times\frac{Ns}{4a+Ns}\equiv 4\pi Da\times\text{efficiency}\,. (6.5)

To understand the implication of this result, let’s use some numbers that typify a cell: a=1a=1 micron, s=10s=10 nanometers, and N=104N=10^{4}. The area fraction covered by the receptors is roughly 10310^{-3}, but the absorption efficiency of the cell is roughly 2/3! Evidently, Mother Nature is very smart to not waste resources on endowing a cell with too many receptors.

7 Wedge Domains

We now turn to the first-passage properties of diffusion in a two-dimensional wedge domain with absorption when the particle hits the wedge boundary. One of our motivations for studying this system is that first passage in the wedge can be mapped onto a simple diffusive capture problem in one dimension that we’ll treat in the next section. We will obtain first-passage properties in the wedge geometry both by direct solution of the diffusion equation and also, for two dimensions, in a more aesthetically pleasing fashion by conformal transformation techniques, in conjunction with the electrostatic formulation. We also present a heuristic extension of the electrostatic approach that allows us to infer, with little additional computational effort, time-dependent first-passage properties in the wedge from corresponding time-integrated properties.

Solution to the Diffusion Equation

We first solve the diffusion equation in the wedge to determine the survival probability of a diffusing particle. While the exact Green’s function for this system is well known [12, 13], we adopt the strategy of choosing an initial condition that allows us to eliminate angular variables and deal with an effective radial problem. This simplification is appropriate if we are interested only in asymptotic first-passage properties.

The diffusion equation for the two-dimensional wedge geometry in plane polar co-ordinates is

ct=D(2cr2+1rcr+1r22cθ2),\frac{\partial c}{\partial t}=D\left(\frac{\partial^{2}c}{\partial r^{2}}+\frac{1}{r}\frac{\partial c}{\partial r}+\frac{1}{r^{2}}\frac{\partial^{2}c}{\partial\theta^{2}}\right), (7.1)

where c=c(r,θ,t)c=c(r,\theta,t) is the particle concentration at (r,θ)(r,\theta) at time tt, DD is the diffusion coefficient, and the boundary conditions are c=0c=0 at θ=0,Θ\theta=0,\Theta, where Θ\Theta is the wedge opening angle. To reduce this two-dimensional problem to an effective one-dimensional radial problem, note that the exact Green’s function can be written as an eigenfunction expansion in which the angular dependence is a sum of sine waves of the form sin(nπθ/Θ)\sin(n\pi\theta/\Theta), such that an integral number of half-wavelengths fit within (0,Θ)(0,\Theta) to satisfy the absorbing boundary conditions [13]. In this series, each sine wave is multiplied by a conjugate decaying function of time, in which the decay rate increases with nn. In the long time limit, only the lowest term in this expansion dominates the survival probability. Consequently, we obtain the long-time behavior by choosing an initial condition whose angular dependence is a half sine-wave in the wedge. This ensures that the time-dependent problem will contain only this single term in the Fourier series.

We therefore define c(r,θ,t=0)=πδ(rr0)/(2Θr0)×sin(πθ/Θ)c(r,\theta,t=0)=\pi\delta(r-r_{0})/(2\Theta r_{0})\times\sin(\pi\theta/\Theta). With this initial distribution function and after the Laplace transform is applied, the diffusion equation (7.1) becomes

sc(r,θ,s)π2Θr0δ(rr0)sin(πθ/Θ)=D(c2r2+1rcr+1r22cθ2),\displaystyle sc(r,\theta,s)-\frac{\pi}{2\Theta r_{0}}\delta(r-r_{0})\sin(\pi\theta/\Theta)=D\left(\frac{\partial c^{2}}{\partial r^{2}}+\frac{1}{r}\frac{\partial c}{\partial r}+\frac{1}{r^{2}}\frac{\partial^{2}c}{\partial\theta^{2}}\right)\,,

where c=c(r,θ,s)c=c(r,\theta,s). Substituting in the ansatz c(r,θ,s)=R(r,s)sin(πθ/Θ)c(r,\theta,s)=R(r,s)\sin(\pi\theta/\Theta), the angular dependence may now be separated and reduces the system to an effective one-dimensional radial problem. By introducing the dimensionless co-ordinate x=rs/Dx=r\sqrt{s/D}, we find the modified Bessel equation for the remaining radial co-ordinate,

R′′(x,s)+1xR(x,s)(1+ν2x2)R(x,s)=ν2Dx0δ(xx0),R^{\prime\prime}(x,s)+\frac{1}{x}R^{\prime}(x,s)-\left(1+\frac{\nu^{2}}{x^{2}}\right)R(x,s)=-\frac{\nu}{2Dx_{0}}\delta(x-x_{0}), (7.2)

where ν=π/Θ\nu=\pi/\Theta and the prime now denotes differentiation with respect to xx.

The general solution for xx0x\neq x_{0} is a superposition of modified Bessel functions of order ν\nu. Since the domain is unbounded, the interior Green’s function (x<x0x<x_{0}) involves only IνI_{\nu}, since KνK_{\nu} diverges as x0x\to 0, while the exterior Green’s function (x>x0x>x_{0}) involves only KνK_{\nu}, since IνI_{\nu} diverges as xx\to\infty. By imposing continuity at x=x0x=x_{0}, we find that the Green’s function has the symmetric form R(x,s)=AIν(x<)Kν(x>)R(x,s)=AI_{\nu}(x_{<})K_{\nu}(x_{>}), with the constant AA determined by the joining condition that arises from integrating Eq. (7.2) over an infinitesimal radial range that includes r0r_{0}. This gives

R>x=x0R<x=x0=ν2Dx0,\displaystyle R_{>}^{\prime}\mid_{x=x_{0}}-R_{<}^{\prime}\mid_{x=x_{0}}=-\frac{\nu}{2Dx_{0}}\,,

from which A=ν/2DA=\nu/2D. Therefore the radial Green’s function in the wedge is

R(x,s)=ν2DIν(x<)Kν(x>),R(x,s)=\frac{\nu}{2D}\,I_{\nu}(x_{<})\,K_{\nu}(x_{>}), (7.3)

and its Laplace inverse has the relatively simple closed form [12, 13]

R(r,t)=ν4Dte(r2+r02)/4DtIν(rr02Dt).R(r,t)=\frac{\nu}{4Dt}\,\,e^{-(r^{2}+r_{0}^{2})/4Dt}\,\,I_{\nu}\left(\frac{rr_{0}}{2Dt}\right). (7.4)

With this radial Green’s function, the asymptotic survival probability is

S(t)0Θsin(νθ)𝑑θ0rR(r,t)𝑑r.S(t)\sim\int_{0}^{\Theta}\sin(\nu\theta)\,d\theta\,\int_{0}^{\infty}r\,R(r,t)\,dr. (7.5)

We can estimate this integral by noting that the radial distance over which the concentration is appreciable extends to the order of Dt\sqrt{Dt}. This provides a cutoff rDtr\approx\sqrt{Dt} in the radial integral in Eq. (7.5), within which the Gaussian factors in R(r,t)R(r,t) can be replaced by one. Using the small-argument expansion of the Bessel function, we then obtain

S(t)=0Θsin(νθ)𝑑θ0rR(r,t)𝑑r0Dt1Dt(rr0Dt)νr𝑑r(r0Dt)π/Θ.\displaystyle S(t)=\int_{0}^{\Theta}\!\!\!\sin(\nu\theta)\,d\theta\,\int_{0}^{\infty}\!\!\!r\,R(r,t)\,dr\propto\int_{0}^{\sqrt{Dt}}\!\frac{1}{Dt}\left(\frac{rr_{0}}{Dt}\right)^{\nu}\,r\,dr\sim\left(\frac{r_{0}}{\sqrt{Dt}}\right)^{\pi/\Theta}. (7.6)

The basic result is that the survival probability of a diffusing particle in a wedge of opening angle Θ\Theta decays with time as

S(t)tπ/2Θtα.S(t)\sim t^{-\pi/2\Theta}\equiv t^{-\alpha}. (7.7)

The striking feature of this formula is that the exponent α\alpha depends on the wedge opening angle in a non-trivial way, with α\alpha\to\infty as Θ0\Theta\to 0 and α1/4\alpha\to 1/4 for Θ2π\Theta\to 2\pi.

Conformal Transformations and Electrostatic Methods

Let’s now solve the same wedge problem by exploiting conformal transformations, together with the connection between first passage and electrostatics. To set the stage for the wedge geometry, consider the first-passage probability for a diffusing particle in two dimensions to an absorbing infinite line. This problem may also be solved elegantly by the electrostatic formulation. In this approach, the time-integrated concentration 𝒞(x,y)=0c(x,y,t)𝑑t\mathcal{C}(x,y)=\int_{0}^{\infty}c(x,y,t)\,dt obeys the Laplace equation

2𝒞(z)=12πDδ(zz0),\displaystyle\nabla^{2}\mathcal{C}(z)=-\frac{1}{2\pi D}\;\delta(z-z_{0})\,,

where z=(x,y)z=(x,y) is the complex co-ordinate and the factor 1/(2πD)1/(2\pi D) ensures the correct normalization. Using the image method for two-dimensional electrostatics, we find that the complex potential is

𝒞(z)=12πDlnzz0zz0=12πDlnxx0+i(yy0)xx0+i(y+y0),\mathcal{C}(z)=\frac{1}{2\pi D}\ln\frac{z-z_{0}}{z-z_{0}^{*}}=\frac{1}{2\pi D}\ln\frac{x-x_{0}+i(y-y_{0})}{x-x_{0}+i(y+y_{0})}, (7.8)

where the asterisk denotes complex conjugation. Finally, the time-integrated flux that is absorbed at xx coincides with the electric field at this point. This is

(x|x0,y0)\displaystyle{\mathcal{E}}(x|x_{0},y_{0}) =D𝒞(x,y)y|y=0\displaystyle=-D\,\frac{\partial\mathcal{C}(x,y)}{\partial y}\bigg{|}_{y=0}
=12π(1y0+i(xx0)+1y0i(xx0))\displaystyle=\frac{1}{2\pi}\left(\frac{1}{y_{0}+i(x-x_{0})}+\frac{1}{y_{0}-i(x-x_{0})}\right)
=1πy0(xx0)2+y02.\displaystyle=\frac{1}{\pi}\frac{y_{0}}{(x-x_{0})^{2}+y_{0}^{2}}\,. (7.9)

We now use a conformal transformation to extend the result for the hitting probability to the infinite line to the hitting probability in the wedge. Consider the transformation w=f(z)=zπ/Θw=f(z)=z^{\pi/\Theta} that maps the interior of the wedge of opening angle Θ\Theta to the upper half plane. In complex co-ordinates, the electrostatic potential in the wedge is

𝒞(z)=12πDlnzπ/Θz0π/Θzπ/Θ(z0)π/Θ.\mathcal{C}(z)=\frac{1}{2\pi D}\ln\frac{z^{\pi/\Theta}-z_{0}^{\pi/\Theta}}{z^{\pi/\Theta}-(z_{0}^{*})^{\pi/\Theta}}. (7.10)

From this expression and using the analogy between electrostatics and first passage, we can extract time-integrated first-passage properties in the wedge. For example, the probability of being absorbed at a distance xx from the wedge apex, when a particle begins at a unit distance from the apex along the wedge bisector, is just the electric field at this point

(x|x0,y0)=1Θxπ/Θ11+x2π/Θx(1+π/Θ)for x.\displaystyle{\mathcal{E}}(x|x_{0},y_{0})=\frac{1}{\Theta}\frac{x^{\pi/\Theta-1}}{1+x^{2\pi/\Theta}}\to x^{-(1+\pi/\Theta)}\qquad\text{for~{}}x\to\infty\,. (7.11)

Although the electrostatic formulation ostensibly gives only time-integrated first-passage properties, we can adapt it to also give time-dependent features. This adaptation is based on the following re-interpretation of the equivalence between electrostatics and diffusion: an electrostatic system with a point charge and specified boundary conditions is identical to a diffusive system in the same geometry and boundary conditions, in which a continuous source of particles is fed in at the location of the charge starting at time t=t=-\infty. Suppose now that the particle source is “turned on” at t=0t=0. Then, in a near zone that extends out to a distance of the order of Dt\sqrt{Dt} from the source, the concentration has sufficient time to reach its steady-state value. Within this zone, the diffusive solution converges to the Laplacian solution. Outside this zone, however, the concentration is close to zero. This almost-Laplacian solution provides the time integral of the survival probability up to time tt. We can then deduce the survival probability by differentiating the concentration that is due to this finite-duration source.

Thus suppose that a constant source of diffusing particles at z0z_{0} inside the absorbing wedge is turned on at t=0t=0. Within the region where the concentration has had time to reach the steady state, |z0|<|z|<Dt|z_{0}|<|z|<\sqrt{Dt}, the density profile is approximately equal to the Laplacian solution, 𝒞(z)|z|π/Θ\mathcal{C}(z)\sim|z|^{-\pi/\Theta}. We can neglect the angular dependence of 𝒞(z)\mathcal{C}(z) in this zone, as this dependence is immaterial for the survival probability. Conversely, for |z|>Dt|z|>\sqrt{Dt}, the particle concentration is vanishingly small because a particle is unlikely to diffuse such a large distance. From the analogy between electrostatics and first passage, the near-zone density profile is just the same as the time integral of the diffusive concentration. Thus, by using the equivalence between the spatial integral of this near-zone concentration in the wedge and the time integral of the survival probability, we have

0tS(t)𝑑t0Dtrπ/Θr𝑑rt1π/2Θ.\int_{0}^{t}S(t^{\prime})\,dt^{\prime}\approx\int_{0}^{\sqrt{Dt}}r^{-\pi/\Theta}\,r\,dr\sim t^{1-\pi/2\Theta}. (7.12)

Since the total density injected into the system equals tt, the survival probability in the wedge is roughly 0tS(t)𝑑t/tt1π/2Θ/t\int_{0}^{t}S(t^{\prime})\,dt^{\prime}/t\sim t^{1-\pi/2\Theta}/t, which gives S(t)tπ/2ΘS(t)\sim t^{-\pi/2\Theta}.

8 Stochastic Hunting in One Dimension

What is the time dependence of the survival probability of a diffusing lamb that is hunted by NN diffusing lions? We define this survival probability as SN(t)S_{N}(t). This toy problem is most interesting in a one-dimensional geometry where all the lions are located to one side of the lamb. It is known [14] that this survival probability asymptotically decays algebraically with time,

SN(t)tβN,\displaystyle S_{N}(t)\sim t^{-\beta_{N}}\,, (8.1)

and the goal is to compute the decay exponent βN\beta_{N}. As we shall discuss, the decay exponent is known for N=1N=1 and N=2N=2 only: β1=12\beta_{1}=\frac{1}{2}, and β2=34\beta_{2}=\frac{3}{4}. For N3N\geq 3, βN\beta_{N} grows slowly with NN, and numerical simulations give β30.91\beta_{3}\approx 0.91, β41.03\beta_{4}\approx 1.03, and β101.4\beta_{10}\approx 1.4. The focus of this section is to derive β2=34\beta_{2}=\frac{3}{4} by a simple geometric approach and to develop some analytical understanding of the dependence of βN\beta_{N} on NN for NN\to\infty.

Let us begin by treating a lamb that starts at x0>0x_{0}>0 and a single lion that starts at x=0x=0. For simplicity, the diffusivities of the lamb and the lion are assumed to both equal DD. The separation between the lamb and the lion thus diffuses with diffusion coefficient 2D2D. When this separation reaches zero, the lamb has been eaten. This problem is just the classic first-passage problem on the positive infinite line, except that the diffusion coefficient is 2D2D. The probability that the lamb survives until time tt is the same as the first-passage time being greater than tt. This probability therefore is

S1(t)\displaystyle S_{1}(t) =tx08πDt3ex02/8Dt𝑑t=erf(x08Dt)x08πDt.\displaystyle=\int_{t}^{\infty}\!\frac{x_{0}}{\sqrt{8\pi Dt^{\prime 3}}}\;e^{-x_{0}^{2}/8Dt^{\prime}}\,dt^{\prime}=\text{erf}\biggl{(}\frac{x_{0}}{\sqrt{8Dt}}\biggr{)}\sim\frac{x_{0}}{\sqrt{8\pi Dt}}\,. (8.2)

Thus the survival probability of the lamb asymptotically decays as t1/2t^{-1/2}. While the lamb is sure to die, its average lifetime is infinite. Thus a single diffusing lion is not a particularly good hunter and it might starve before eating the lamb.

What happens when there are N=2N=2 lions? We again assume that the lions start from the origin while the lamb starts at x0>0x_{0}>0, and that the diffusivities of all particles are the same. Let us label the positions of the lions as x1x_{1} and x2x_{2}, and the position of the lamb as x3x_{3} The lamb survives up to time tt if the conditions x3>x2x_{3}>x_{2} and x3>x1x_{3}>x_{1} always hold. We can give an insightful geometric interpretation of this problem by viewing the motion of the three particles on the line as equivalent to the motion of a single effective particle in three dimensions with coordinates (x1,x2,x3)(x_{1},x_{2},x_{3}). The constraints x3>x2x_{3}>x_{2} and x3>x1x_{3}>x_{1} mean that the effective particle in three-space remains to the left of the plane x2=x3x_{2}=x_{3} and behind the plane x1=x3x_{1}=x_{3} (Fig. 7(a)). This allowed region is a wedge of opening angle Θ\Theta that is defined by the intersection of these two planes. If the particle hits one of the planes, then one of the lions has eaten the lamb.

Refer to caption
Refer to caption
Figure 7: (a) The allowed region for the effective particle that corresponds to the motion of the lamb and two lions. (b) A view of the constraint planes perpendicular to the (1,1,1) axis. The constraint plane x1=x2x_{1}=x_{2} is also shown

This mapping therefore provides the lamb survival probability, since the survival probability of a diffusing particle within this absorbing wedge asymptotically decays as Swedge(t)tπ/2ΘS_{\rm wedge}(t)\sim t^{-\pi/2\Theta}. What is the opening angle of this wedge? We can determine this angle in a simple way by also including the plane x1=x2x_{1}=x_{2} and then viewing the system along the (1,1,1)(1,1,1) axis (Fig. 7(b)). It is then clear that the wedge angle is 2π/32\pi/3. Substituting this result in Eq. (7.7), we find that β2=34\beta_{2}=\frac{3}{4}. Notice that S2(t)>S1(t)2S_{2}(t)>S_{1}(t)^{2}. This inequality reflects the fact that the incremental threat to the lamb from the second lion is less than the first.

In general, we can map the motion of the lamb and NN lions in one dimension to a single effective particle in N+1N+1 dimensions, with absorption when the effective particle hits any of the NN constraint planes. However, the calculation of the survival probability of the effective particle within the domain where the effective particle is confined—known as a Weyl chamber—appears to be intractable.

Refer to caption
Figure 8: Space-time trajectories of 6 lions and 1 lamb. Although the motion of each lion is isotropic, the trajectory of the lion (whose identity may change) that happens to be closest to the lamb (solid red curve) tends to move towards the lamb.

While the problem for N=3,4,N=3,4,\ldots lions is difficult, the problem becomes much simpler for large NN. To determine the lamb survival probability for large NN, we only need to focus on the lion closest to the lamb, because this last lion ultimately kills the lamb. As shown in Fig. 8, the individual identity of this last lion can change with time due to the crossing of different lion trajectories. For large NN, there is a systematic bias of the motion of the last lion, xlast(t)x_{\rm last}(t). This bias becomes stronger for increasing NN, so that xlast(t)x_{\rm last}(t) becomes smoother as NN increases (Fig. 9). This gradual approach of the last lion to the lamb is the mechanism which leads to the survival probability of the lamb decaying as tβNt^{-\beta_{N}}, with βN\beta_{N} a slowly increasing function of NN.

Refer to caption
Figure 9: Position of the last lion versus time for 4, 64, and 1024 lions.

To estimate the location of this last lion when N1N\gg 1 lions are initially at the origin, we use the extreme statistics condition [15]

xlastN4πDtex2/4Dt𝑑x=1.\int_{x_{\rm last}}^{\infty}\frac{N}{\sqrt{4\pi Dt}}\,e^{-x^{2}/4Dt}\,\,dx=1. (8.3)

Equation (8.3) states that one lion out of an initial group of NN is in the range [xlast,][x_{\rm last},\infty]. Although the integral in Eq. (8.3) can be expressed in terms of the complementary error function, it is instructive to evaluate it approximately in a self consistent way by writing x=xlast+ϵx=x_{\rm last}+\epsilon and re-expressing the integrand in terms of ϵ\epsilon. We thus find

0N4πDtexlast2/4Dtexlastϵ/2Dteϵ2/4Dt𝑑ϵ=1.\int_{0}^{\infty}\frac{N}{\sqrt{4\pi Dt}}\,e^{-x_{\rm last}^{2}/4Dt}\,e^{-x_{\rm last}\epsilon/2Dt}\,e^{-\epsilon^{2}/4Dt}\,\,d\epsilon=1.

Now the second term in the integrand,

exlastϵ/2Dt,\displaystyle e^{-x_{\rm last}\epsilon/2Dt}\,,

is non-negligible for ϵ<2Dt/xlast\epsilon<2Dt/x_{\rm last}. Over this range of ϵ\epsilon, the third exponential factor is of the order of

eDt/xlast2.\displaystyle e^{-Dt/x_{\rm last}^{2}}\,.

If we use the result for xlastx_{\rm last} in Eq. (8.5), the above exponential factor becomes

e1/(4lnN),\displaystyle e^{-1/(4\ln N)}\,,

which is very close to 1 for large N. If we thus ignore this term, the integral above reduces to simple exponential decay, with the result

N4πDtexlast2/4Dt2Dtxlast=1.\frac{N}{\sqrt{4\pi Dt}}\,e^{-x_{\rm last}^{2}/4Dt}\,\frac{2Dt}{x_{\rm last}}=1. (8.4)

We now define y=xlast/4Dty=x_{\rm last}/\sqrt{4Dt} and M=N/4πM=N/\sqrt{4\pi}, so that the above condition can be simplified to yexp(ey2)=My\exp(e^{y^{2}})=M, whose asymptotic solution is

y=lnM(114ln(lnM)lnM+).y=\sqrt{\ln M\,}\Bigl{(}1-\frac{1}{4}\frac{\ln(\ln M)}{\ln M}+\ldots\Bigr{)}.

To lowest order, this gives

xlast(t)4DlnNtANt,x_{\rm last}(t)\sim\sqrt{4D\ln N\,t}\equiv\sqrt{A_{N}t}, (8.5)

for NN finite. For N=N=\infty, xlast(t)x_{\rm last}(t) would always equal tt if an infinite number of discrete random walking lions were initially at the origin. A more suitable initial condition therefore is a concentration c0c_{0} of lions that are uniformly distributed from -\infty to 0. In this case, only Nc02DtN\propto\sqrt{c^{2}_{0}Dt} of the lions are “dangerous,” that is, within a diffusion distance from the edge of the pack and thus potential candidates for eating the lamb. Consequently, for NN\to\infty, the leading behavior of xlast(t)x_{\rm last}(t) becomes

xlast(t)2Dln(c02Dt)t.x_{\rm last}(t)\sim\sqrt{2D\ln(c^{2}_{0}Dt)\,t}\,. (8.6)

An important feature of the time dependence of xlastx_{\rm last} is that fluctuations in this quantity decrease for large NN (Fig. 9). Therefore the lamb and NN diffusing lions can be recast as a two-body system of a lamb and an absorbing boundary which deterministically advances toward the lamb according to xlast(t)=ANtx_{\rm last}(t)=\sqrt{A_{N}t}. This determinism is what makes the problem for large NN tractable.

To solve this effective two-body problem, it is convenient to change coordinates from (x,t)(x,t) to (=xxlast(t),t)(^{\prime}=x-x_{\rm last}(t),t) to fix the absorbing boundary at the origin. By this construction, the diffusion equation for the lamb probability distribution is transformed to the convection-diffusion equation

p(x,t)txlast2tp(x,t)x=D2p(x,t)x2,(0x<)\frac{\partial p(x^{\prime},t)}{\partial t}-\frac{x_{\rm last}}{2t}\frac{\partial p(x^{\prime},t)}{\partial x^{\prime}}=D\frac{\partial^{2}p(x^{\prime},t)}{\partial x^{\prime 2}}\,,\quad(0\leq x^{\prime}<\infty) (8.7)

with the absorbing boundary condition p(x=0,t)=0p(x^{\prime}=0,t)=0. In this reference frame that is fixed on the average position of the last lion, the second term in Eq. (8.7) accounts for the bias of the lamb towards the absorber with an effective speed xlast/2t-x_{\rm last}/2t. Because xlastANtx_{\rm last}\sim\sqrt{A_{N}t} and xDtx^{\prime}\sim\sqrt{Dt} have the same time dependence, the lamb survival probability acquires a nontrivial dependence on the dimensionless parameter AN/DA_{N}/D. This behavior arises whenever there is a coincidence of fundamental length scales in the system (see, e.g., [16] for other such examples).

Equation (8.7) can be transformed into the parabolic cylinder equation by first introducing the dimensionless length ξ=x/xlast\xi=x^{\prime}/x_{\rm last} and making the following scaling ansatz for the lamb probability density,

p(x,t)tβN1/2𝒞(ξ).p(x^{\prime},t)\sim t^{-\beta_{N}-1/2}\mathcal{C}(\xi). (8.8)

The power law prefactor in Eq. (8.8) ensures that the integral of p(x,t)p(x^{\prime},t) over all space, namely the survival probability, decays as tβNt^{-\beta_{N}}, and 𝒞(ξ)\mathcal{C}(\xi) expresses the spatial dependence of the lamb probability distribution in scaled length units. This ansatz codifies the fact that the probability density is not a function of xx^{\prime} and tt separately, but is a function only of the dimensionless ratio x/xlast(t)x^{\prime}/x_{\rm last}(t).

Substituting Eq. (8.8) into Eq. (8.7), we obtain

DANd2𝒞dξ2+12(ξ+1)d𝒞dξ+(βN+12)𝒞=0.\frac{D}{A_{N}}\frac{d^{2}{\mathcal{C}}}{d\xi^{2}}+\frac{1}{2}(\xi+1)\frac{d\mathcal{C}}{d\xi}+\bigl{(}\beta_{N}+\frac{1}{2}\bigr{)}\mathcal{C}=0. (8.9)

By introducing η=(ξ+1)AN/2D\eta=(\xi+1)\sqrt{A_{N}/2D} and 𝒞(ξ)=eη2/4𝒟(η)\mathcal{C}(\xi)=e^{-\eta^{2}/4}\,{\mathcal{D}}(\eta) in Eq. (8.9), we are led to the parabolic cylinder equation of order 2βN2\beta_{N} [17]

d2𝒟2βNdη2+[2βN+12η24]𝒟2βN=0,\frac{d^{2}{\mathcal{D}}_{2\beta_{N}}}{d\eta^{2}}+\biggl{[}2\beta_{N}+\frac{1}{2}-\frac{\eta^{2}}{4}\biggr{]}{\mathcal{D}}_{2\beta_{N}}=0, (8.10)

subject to the boundary condition, 𝒟2βN(η)=0{\mathcal{D}}_{2\beta_{N}}(\eta)=0 for both η=AN/2D\eta=\sqrt{A_{N}/2D} and η=\eta=\infty. Equation (8.10) has the form of a Schrödinger equation for a quantum particle of energy 2βN+122\beta_{N}+\frac{1}{2} in a harmonic oscillator potential η2/4\eta^{2}/4 for η>AN/2D\eta>\sqrt{A_{N}/2D}, but with an infinite barrier at η=AN/D\eta=\sqrt{A_{N}/D} [18]. For the long-time behavior, we need to find the ground state energy in this potential. For N1N\gg 1, we may approximate this energy as the potential at the classical turning point, that is, 2βN+12η2/42\beta_{N}+\frac{1}{2}\simeq\eta^{2}/4. We therefore obtain βNAN/16D\beta_{N}\sim A_{N}/16D. Using the value of ANA_{N} given in Eqs. (8.5) and (8.6) the decay exponent is

βN{14lnNNfinite18lntN=.\beta_{N}\sim\begin{cases}\tfrac{1}{4}\ln N&N~{}\text{finite}\\[5.69054pt] \tfrac{1}{8}\ln t&N=\infty\,.\end{cases} (8.11)

The latter dependence of βN\beta_{N} implies that for NN\to\infty, the survival probability has the log-normal form

S(t)exp(18ln2t).S_{\infty}(t)\sim\exp\left(-\tfrac{1}{8}\,\ln^{2}t\right). (8.12)

The important feature of the exponent βN\beta_{N} is its very slow increase with NN. That is, each successive lion that is added to the hunt has a decreasing influence on the survival of the lamb. Indeed, only a small subset of the lions for large NN actually have a chance to catch the lamb.

9 The Expanding Interval

We have seen that the survival probability S(t)S(t) of a diffusing particle in a fixed-length absorbing interval of length LL asymptotically decays as eπ2Dt/L2e^{-\pi^{2}Dt/L^{2}}. What happens if the interval length grows with time, L(t)tαL(t)\sim t^{\alpha}? This simple question illustrates the relative effects of diffusion and the motion of the boundary on first-passage properties. This interplay is a classic problem in the first-passage literature, especially when the boundary motion matches that of diffusion. Solutions to this problem have been obtained by a variety of methods (see, e.g., [19, 20, 21, 22]). Here we give a physics-based approach that is based on [23].

It is easy to infer the survival probability for a slowly expanding interval. Here, slowly expanding means that the interval length grows slower than diffusion. Consequently, the probability distribution of the particle spreads faster than the interval grows and so that the survival probability should decay rapidly with time. Using an adiabatic approximation that one typically encounters in basic quantum mechanics, we will show that S(t)exp[At(12α)]S(t)\sim\exp\big{[}-At^{(1-2\alpha)}\big{]}, where AA is a constant. Conversely, for the rapidly expanding interval, α>1/2\alpha>1/2, the particle is unlikely reach the either end of the interval and the probability distribution is close to that for free diffusion. This is the basis of the free approximation that leads to a non-zero limiting value for S(t)S(t) as tt\to\infty.

In the marginal case where the interval expands at the same rate as diffusion, L(t)=AtL(t)=\sqrt{At}, a new dimensionless parameter arises—the ratio of the diffusion length to the interval length. As we shall show, this leads to S(t)S(t) decaying as a non-universal power-law in time, S(t)tβS(t)\sim t^{-\beta}, with β=(¯A,D)\beta=\b{(}A,D) diverging for A/D1A/D\ll 1 and approaching zero for A/D1A/D\gg 1.

Slowly Expanding Interval

For L(t)DtL(t)\ll\sqrt{Dt}, we invoke the adiabatic approximation [18], in which the spatial dependence of the concentration for an interval of length L(t)L(t) is assumed to be identical to that of the static diffusion equation at the instantaneous value of LL. This assumption is based on the expectation that the concentration in a slowly expanding interval is always close to that of a fixed-size interval. Thus we write

c(x,t)f(t)cos(πx2L(t))cad(x,t),c(x,t)\simeq f(t)\,\cos\left(\frac{\pi x}{2L(t)}\right)\equiv c_{\rm ad}(x,t), (9.1)

with f(t)f(t) to be determined. The corresponding survival probability is

S(t)L(t)L(t)cad(x,t)𝑑x=4πf(t)L(t).S(t)\approx\int_{-L(t)}^{L(t)}~{}c_{\rm ad}(x,t)\,dx=\frac{4}{\pi}~{}f(t)L(t). (9.2)

For convenience, we now define the interval boundaries as [L(t),L(t)][-L(t),L(t)]. To obtain f(t)f(t), we substitute approximation (9.1) into the diffusion equation, as in separation of variables, to give

dfdt=(Dπ24L2)f(πx2L2)dLdttan(πx2L)f.\frac{df}{dt}=-\left(\frac{D\pi^{2}}{4L^{2}}\right)f-\left(\frac{\pi x}{2L^{2}}\right)\,\frac{dL}{dt}\,\tan\left(\frac{\pi x}{2L}\right)\,f. (9.3)

Notice that variable separation does not strictly hold, since the equation for f(t)f(t) also involves xx. However, when L(t)L(t) increases as (At)α(At)^{\alpha} with α<1/2\alpha<1/2, the second term on the right-hand side is negligible. Thus we drop this second term and solve the simplified form of (9.3). We thereby find that the controlling factor of f(t)f(t) is given by

f(t)exp[Dπ240tdtL2(t)]=exp[Dπ24(12α)A2αt12α].\displaystyle f(t)\sim\exp\left[-\frac{D\pi^{2}}{4}\int_{0}^{t}\,\frac{dt^{\prime}}{L^{2}(t^{\prime})}\right]=\exp\left[-\frac{D\pi^{2}}{4(1-2\alpha)A^{2\alpha}}\,t^{1-2\alpha}\right]. (9.4)

Notice that f(t)f(t) reduces to a pure exponential decay for a fixed-length interval, while for α1/2\alpha\to 1/2, Eq. (9.4) suggests a more slowly decaying functional form for S(t)S(t).

Rapidly Expanding Interval

For a rapidly expanding interval, the escape rate from the system is small and the absorbing boundaries should eventually become irrelevant. We therefore expect that the concentration profile should approach the Gaussian distribution of free diffusion at long times [23]. We may then account for the slow decay of the survival probability by augmenting the Gaussian with an overall decaying amplitude. This free approximation is a nice example in which the existence of widely separated time scales, Dt\sqrt{Dt} and L(t)L(t), suggests the nature of the approximation itself.

According to the free approximation, we write

c(x,t)S(t)4πDtex2/4Dtcfree(x,t).\displaystyle c(x,t)\approx\frac{S(t)}{\sqrt{4\pi Dt}}\,e^{-x^{2}/4Dt}\equiv c_{\rm free}(x,t)\,.

Although this concentration does not satisfy the absorbing boundary condition, the inconsistency is negligible at large times, since the density is exponentially small at the interval boundaries. We may now find the time dependence of the survival probability by equating the probability flux to the interval boundaries, 2D|cx|2D|\frac{\partial c}{\partial x}|, to the loss of probability within the interval. For L(t)=(At)αL(t)=(At)^{\alpha}, this flux is

S(t)Aα4πDtα3/2exp(A2α4Dt2α1),\frac{S(t)A^{\alpha}}{\sqrt{4\pi D}}\;t^{\alpha-3/2}\,\exp\left(-\frac{A^{2\alpha}}{4D}t^{2\alpha-1}\right), (9.5)

which rapidly goes to zero for α>1/2\alpha>1/2. Since this flux equals dSdt-\frac{dS}{dt}, it follows that the survival probability approaches a non-zero limiting value for α>1/2\alpha>1/2, and that this limiting value goes to zero as α1/2\alpha\to 1/2. Explicitly,

dS(t)dt=1S(t)Btα3/2exp(Ct2α1),\displaystyle\frac{dS(t)}{dt}=-\frac{1}{S(t)}\;B\,t^{\alpha-3/2}\;\exp\left(-C\,t^{2\alpha-1}\right)\,, (9.6)

where BAα/4πDB\equiv A^{\alpha}/\sqrt{4\pi D} and CA2α/(4D)C\equiv A^{2\alpha}/(4D). We now introduce y=Ct2α1y=Ct^{2\alpha-1} and change the integration variable from tt to yy. After some straightforward steps we have

lnS(t=)\displaystyle\ln S(t\!=\!\infty) =B2α10(yC)1/2eydyy\displaystyle=-\frac{B}{2\alpha-1}\int_{0}^{\infty}\left(\frac{y}{C}\right)^{1/2}e^{-y}\;\frac{dy}{y}
=BC1/212α1Γ(12),\displaystyle=-\frac{B}{C^{1/2}}\frac{1}{2\alpha-1}\,\Gamma\left(\tfrac{1}{2}\right)\,, (9.7)

where Γ()\Gamma(\cdot) is the Euler gamma function. Thus a diffusing particle has a non-zero probability to survive forever when the interval grows fast enough. This ultimate survival probability rapidly goes to zero as α1/2\alpha\to 1/2 from above.

Marginally Expanding Interval

For the marginal case of α=1/2\alpha=1/2, the adiabatic and the free approximations are ostensibly no longer appropriate, since L(t)=AtL(t)=\sqrt{At} and Dt\sqrt{Dt} have a fixed ratio. However, for A/D1A/D\ll 1 and A/D1A/D\gg 1, we might hope that these methods could still be useful. Thus we continue to apply these heuristic approximations in their respective domains of validity, A/D1A/D\ll 1 and A/D1A/D\gg 1, and check their accuracies a posteriori. We will see that the survival probability exponents predicted by these two approximations are each quite close to the exact result except for A/D1A/D\approx 1.

When the adiabatic approximation is applied, the second term in Eq. (9.3) is, in principle, non-negligible for α=1/2\alpha=1/2. However, for ADA\ll D, the interval still expands more slowly (in amplitude) than free diffusion and the error made by neglecting the second term in Eq. (9.3) may still be small. The solution to this crudely truncated equation immediately gives f(t)f(t), which, when substituted into approximately (9.2) leads to Sad(t)tβadS_{\rm ad}(t)\approx t^{-\beta_{\rm ad}}, with

βadDπ24A12.\beta_{\rm ad}\approx\frac{D\pi^{2}}{4A}-\frac{1}{2}. (9.8)

The trailing factor of 1/2-1/2 should not be taken very seriously, because the neglected term in Eq. (9.3) leads to additional corrections to βad\beta_{\rm ad} that are also of the order of 1.

Refer to caption
Figure 10: The survival probability exponent β\beta for the marginal case L(t)=(At)1/2L(t)=(At)^{1/2}. Shown is the numerically exact value of β\beta (red, solid) from solution of Eq. (9.13), together with the predictions from the adiabatic and the free approximations, Eqs.(9.8) and (9.10) respectively (dashed curves).

Similarly for ADA\gg D, the free approximation gives

dSdt2D|c(x,t)x|x=At=S(t)tA4πDeA/4D.\displaystyle\frac{dS}{dt}\approx-2D\bigg{|}\frac{\partial c(x,t)}{\partial x}\bigg{|}_{x=\sqrt{At}}\;=\,-\frac{S(t)}{t}\sqrt{\frac{A}{4\pi D}}\,e^{-A/4D}. (9.9)

This again leads to the non-universal power law for the survival probability, SfreetβfreeS_{\rm free}\sim t^{-\beta_{\rm free}}, with

βfree=A4πDeA/4D.\beta_{\rm free}=\sqrt{\frac{A}{4\pi D}}\,e^{-A/4D}. (9.10)

As shown in Fig. 10, these approximations are surprisingly accurate over much of the range of A/DA/D.

To complete our discussion, we outline a first-principles analysis for the survival probability of a diffusing particle in a marginally expanding interval [23]. When L(t)=AtL(t)=\sqrt{At}, a natural scaling hypothesis is to write the density in terms of the two dimensionless variables

ξxL(t)andρ=AD.\displaystyle\xi\equiv\frac{x}{L(t)}\qquad{\rm and}\qquad\rho=\sqrt{\frac{A}{D}}\,.

We now seek solutions for the concentration in the form,

c(x,t)=tβ1/2𝒞ρ(ξ),c(x,t)=t^{-\beta-1/2}\;\mathcal{C}_{\rho}(\xi), (9.11)

where 𝒞ρ(ξ)\mathcal{C}_{\rho}(\xi) is a two-variable scaling function that encodes the spatial dependence. The power law prefactor ensures that the survival probability, namely, the spatial integral of c(x,t)c(x,t), decays as tβt^{-\beta}, as defined at the outset of this section.

After substituting Eq. (9.11) into the diffusion equation, the scaling function satisfies the ordinary differential equation

1ρ2d2𝒞dξ2+ξ2d𝒞dξ+(β+12)𝒞=0.\displaystyle\frac{1}{\rho^{2}}\frac{d^{2}\mathcal{C}}{d\xi^{2}}+\frac{\xi}{2}\frac{d\mathcal{C}}{d\xi}+\left(\beta+\frac{1}{2}\right)\mathcal{C}=0\,.

Then by introducing η=ξρ/2\eta=\xi\,\sqrt{\rho/2} and 𝒞(ξ)=eη2/4𝒟(η)\mathcal{C}(\xi)=e^{-\eta^{2}/4}\,{\mathcal{D}}(\eta), we transform this into the parabolic cylinder equation [25]

d𝒟dη2+[2β+12η24]𝒟=0.\frac{d{\mathcal{D}}}{d\eta^{2}}+\left[2\beta+\frac{1}{2}-\frac{\eta^{2}}{4}\right]{\mathcal{D}}=0. (9.12)

When the range of η\eta is unbounded, this equation has solutions for quantized values of the energy eigenvalue E=2β+12=12E=2\beta+\frac{1}{2}=\frac{1}{2}, 32\frac{3}{2}, 52,\frac{5}{2},\ldots [18].

For our interval problem, the range of η\eta is restricted to |η|A/2D|\eta|\leq\sqrt{A/2D}. In the equivalent quantum mechanical system, this corresponds to a particle in a harmonic-oscillator potential for |η|<A/2D|\eta|<\sqrt{A/2D} and an infinite potential for |η|>A/2D|\eta|>\sqrt{A/2D}. For this geometry, a spatially symmetric solution to Eq. (9.12), appropriate for the long-time limit for an arbitrary starting point, is

𝒟(η)12[𝒟2β(η)+𝒟2β(η)],\displaystyle{\mathcal{D}}(\eta)\equiv\tfrac{1}{2}\big{[}{\mathcal{D}}_{2\beta}(\eta)+{\mathcal{D}}_{2\beta}(-\eta)\big{]}\,,

where 𝒟ν(η){\mathcal{D}}_{\nu}(\eta) is the parabolic cylinder function of order ν\nu. Finally, the relation between the decay exponent β\beta and A/D\sqrt{A/D} is determined implicitly by the absorbing boundary condition, namely,

𝒟2β(A/2D)+𝒟2β(A/2D)=0.{\mathcal{D}}_{2\beta}(\sqrt{A/2D})+{\mathcal{D}}_{2\beta}(-\sqrt{A/2D})=0. (9.13)

This condition for β=β(A/D)\beta=\beta(A/D) simplifies in the limiting cases A/D1A/D\ll 1 and A/D1A/D\gg 1. In the former, the exponent β\beta is large and the second two terms in the brackets in Eq. (9.12) can be neglected. Equivalently, the physical range of η\eta is small, so that the potential plays a negligible role. The solution to this limiting free-particle equation is just the cosine function, and the boundary condition immediately gives the limiting expression of Eq. (9.8), but without the subdominant term of 1/2-1/2. In the latter case of ADA\gg D, β0\beta\to 0 and Eq. (9.12) approaches the Schrödinger equation for the ground state of the harmonic oscillator. In this case, a detailed analysis of the differential equation reproduces the limiting exponent of Eq. (9.10) (see [23] for details). These provide rigorous justification for the limiting values of the decay exponent β\beta which we obtained by heuristic means.

The Khintchine Iterated Logarithm Law

In the marginal situation of L(t)=(At)1/2L(t)=(At)^{1/2}, we have seen that the survival probability S(t)S(t) decays as a power law tβ(A,D)t^{-\beta(A,D)}, with β0\beta\to 0 as A/DA/D\to\infty. This decay becomes progressively slower as AA increases. On the other hand, when L(t)tαL(t)\propto t^{\alpha}, with α\alpha strictly greater than 1/2, the survival probability at infinite time is greater than zero. This leads to the following natural question: what is the nature of the transition between certain death, defined as S(t)=0S(t\to\infty)=0, and a non-zero survival survival probability, S(t)>0S(t\to\infty)>0?

The answer to this question is surprisingly rich. There is an infinite sequence of transitions, where L(t)L(t) acquires additional iterated logarithmic time dependences, which define regimes where S(t)S(t) assumes progressively slower functional forms. The first term in this series is known as the Khintchine iterated logarithm law ([24, 3]). While the Khintchine law has been obtained by rigorous methods, we can also obtain this intriguing result, as well as the infinite sequence of transitions, with relatively little computation by the free approximation.

Because we anticipate that the transition between life and death occurs when L(t)L(t) grows slightly faster than (At)1/2(At)^{1/2}, we make the hypothesis that Au(t)A\propto u(t), with u(t)u(t) growing slower than a power law in tt. Now that L(t)L(t) increases more rapidly than the diffusion length (Dt)1/2(Dt)^{1/2}, the free approximation should be asymptotically exact, since it already works extremely well when L(t)=(At)1/2L(t)=(At)^{1/2} with AA large. Within this approximation, we rewrite Eq. (9.9) as

lnS(t)tdttA4πDeA/4D.\ln S(t)\sim-\int^{t}\frac{dt^{\prime}}{t^{\prime}}\sqrt{\frac{A}{4\pi D}}\,e^{-A/4D}. (9.14)

Here we neglect the lower limit, since the free approximation is valid only as tt\to\infty, where the short-time behavior is irrelevant. In this form, it is clear that for L=(At)1/2L=(At)^{1/2}, lnS\ln S decreases by an infinite amount for tt\to\infty because of the divergence of the integral. Thus S(t)0S(t\to\infty)\to 0. To make the integral converge, the other factors in the integral must somehow cancel the logarithmic divergence that arises from the factor dt/tdt/t. Accordingly, let us substitute L(t)=4Dtu(t)L(t)=\sqrt{4Dt\,u(t)} into the approximation (9.14). This gives

lnS(t)1πtdttu(t)eu(t).\ln S(t)\sim-\frac{1}{\sqrt{\pi}}\int^{t}\frac{dt^{\prime}}{t^{\prime}}\sqrt{u(t^{\prime})}\,e^{-u(t^{\prime})}.

To simplify this integral, it is helpful to define x=lntx=\ln t so that

S(x)1πlnt𝑑xu(x)eu(x).S(x)\sim-\frac{1}{\sqrt{\pi}}\int^{\ln t}dx\,\sqrt{u(x)}\,e^{-u(x)}. (9.15)

To lowest order, it is clear that if we choose u(x)=λlnxu(x)=\lambda\ln x with λ>1\lambda>1, the integral converges as tt\to\infty. Thus the asymptotic survival probability is positive. Conversely, for λ1\lambda\leq 1, the integral diverges and the particle surely dies. In this latter case, evaluation of the integral to lowest order gives

S(t)exp[(lnt)1λλlnlntπ(1λ)]λ<1.S(t)\sim\exp\left[-\frac{(\ln t)^{1-\lambda}\sqrt{\lambda\ln\ln t}}{\sqrt{\pi}(1-\lambda)}\right]\qquad\lambda<1. (9.16)

This decay is slower than any power law, but faster than any power of logarithm, that is, tβ<S(t)<(lnt)γt^{-\beta}<S(t)<(\ln t)^{-\gamma} for β0\beta\to 0 and γ\gamma\to\infty.

What happens in the marginal case of λ=1\lambda=1? Here we can refine the criterion between life and death still further by incorporating into u(x)u(x) a correction that effectively cancels the subdominant factor u(x)\sqrt{u(x)} in Eq. (9.15). We therefore define u(x)u(x) such that eu(x)=1/x(lnx)μe^{-u(x)}=1/x(\ln x)^{\mu}. Then in terms of y=lnxy=\ln x, Eq. (9.15) becomes

S(y)1πlnlntdyyμ1/2(1+μlnyy).S(y)\sim-\frac{1}{\sqrt{\pi}}\int^{\ln\ln t}\frac{dy}{y^{\mu-1/2}}\left(1+\frac{\mu\ln y}{y}\right). (9.17)

This integral now converges for μ>3/2\mu>3/2 and diverges for μ3/2\mu\leq 3/2. In the latter case, the survival probability now lies between the bounds (lnt)γ<S(t)<(lnlnt)δ(\ln t)^{-\gamma}<S(t)<(\ln\ln t)^{-\delta} for γ0\gamma\to 0 and δ\delta\to\infty. At this level of approximation, we conclude that when the cage length grows faster than

aL(t)=4Dt(lnlnt+32lnlnlnt+)aL^{*}(t)=\sqrt{4Dt\left(\ln\ln t+\tfrac{3}{2}\ln\ln\ln t+\ldots\right)} (9.18a)
a diffusing particle has a non-zero asymptotic survival probability, while for a interval that expands as LL^{*}, there is an extremely slow decay of the survival probability.

By incorporating successively finer corrections into u(x)u(x) and following the same logic that led to Eq. (9.17), an infinite series of correction terms can be generated in the expression for L(t)L^{*}(t). By this approach, the ultimate life-death transition corresponds to an ultra-slow decay in which S(t)S(t) has the form S(t)limn1/lnntS(t)\sim\lim_{n\to\infty}1/\ln_{n}t, where ln2tlnlnt\ln_{2}t\equiv\ln\ln t and lnntlnlnn1t\ln_{n}t\equiv\ln\ln_{n-1}t. It is remarkable that the physically motivated and relatively naive free approximation can generate such an intricate solution. As a final note, P. Erdös sharpened the result of Eq. (9.18a) considerably and found that L(t)L^{*}(t) has the infinite series representation

L(t)=4Dt(ln2t+32ln3t+ln4t+ln5t+),L^{*}(t)=\sqrt{4Dt\left(\ln_{2}t+\tfrac{3}{2}\ln_{3}t+\ln_{4}t+\ln_{5}t+\ldots\right)}\,, (9.18b)

in which only the coefficient of the term multiplying ln3t\ln_{3}t is different than 1.

10 Birth-Death Dynamics

As our last topic, we determine the kinetics of the birth-death process. We imagine a collection of identical particles, each of which gives birth to an identical offspring with rate λ\lambda, and each particle can independently die with rate μ\mu. The goal is to determine the time dependence of the population size. As one can easily imagine, this is a classic model for a variety of biological processes and there is vast literature on this general topic (see, e.g., [26, 27]).

The most interesting case physically is the symmetric situation of equal birth and death rates for each particle, λ=μ\lambda=\mu, so that the average population is static. For μ>λ\mu>\lambda, the population size decreases as e(μλ)te^{-(\mu-\lambda)t}, which quickly goes to zero. In the opposite case, the population grows exponentially with time and an additional mechanism is needed to cut off this growth. For λ=μ\lambda=\mu, the average population is fixed, but the time dependence of the distribution of the number of particle exhibits non-trivial kinetics on the positive infinite line. We can alternatively view the birth-death process as a continuous-time random walk on the line, but with birth and death rates for the entire population that are linear functions of nn. That is, the overall process is symmetric but moves faster for a larger population.

Let nn denote the number of particles in the population. The time dependence of the average number of particles obeys the rate equation n˙=(λμ)n=0\langle\dot{n}\rangle=(\lambda-\mu)\langle n\rangle=0, where the overdot denotes the time derivative. Thus the average number of particles is conserved, as is clear from the condition λ=μ\lambda=\mu. That is, the birth-death process for λ=μ\lambda=\mu is a martingale. More meaningful information is obtained from the full population distribution. For simplicity in the ensuing formulas, we now set λ=μ=1\lambda=\mu=1 without loss of generality. Let Pn(t)P_{n}(t) denote the probability that the population consists of nn particles at time tt. This probability distribution changes in time according to

P˙n=(n1)Pn12nPn+(n+1)Pn+1,\displaystyle\dot{P}_{n}=(n-1)P_{n-1}-2nP_{n}+(n+1)P_{n+1}\,, (10.1)

where we define P1=0P_{-1}=0, so that this equation is valid for all n0n\geq 0. For the standard continuous-time random walk, the corresponding master equation is P˙n=Pn12Pn+Pn+1\dot{P}_{n}=P_{n-1}-2P_{n}+P_{n+1}. We know that this random walk eventually hits the origin, but that the average time to do so is infinite. We want to find the behavior of these two first-passage properties for the birth-death process.

A convenient and powerful way to solve the master equation (10.1) is by the generating function method. We first define the generating function

g(z,t)=n0Pn(t)zn,g(z,t)=\sum_{n\geq 0}P_{n}(t)z^{n}\,,

then take each of the equations for P˙n\dot{P}_{n}, multiply it by znz^{n}, and then sum over all nn. In doing so, we will encounter terms from the right-hand side of (10.1), for example, the second term on the right, that looks like

n02nPnzn,\displaystyle\sum_{n\geq 0}2nP_{n}\,z^{n},

which we can recast as

2zzn0Pnzn=2zgz.\displaystyle 2z\frac{\partial}{\partial z}\sum_{n\geq 0}P_{n}\,z^{n}=2z\frac{\partial g}{\partial z}\,.

By this device of converting multiplication by nn to differentiation for all three terms on the right-hand side of (10.1), we recast (10.1) as

gt=(z22z+1)gz=(1z)2gz,g_{t}=(z^{2}-2z+1)g_{z}=(1-z)^{2}g_{z}\,, (10.2)

where the subscripts now denote partial differentiation and the arguments of gg are not written for compactness.

This first-order partial differential equation can be simplified further by defining the variable yy via dy=dz/(1z)2dy=dz/(1-z)^{2}, which implies that y=1/(1z)y=1/(1-z), or equivalently, z=1y1z=1-y^{-1}. In terms of the variable yy, (10.2) is converted to the classic wave equation gt=gyg_{t}=g_{y}. This equation has the general solution g=F(t+y)g=F(t+y), where the function FF is, in principle, arbitrary, and whose explicit form is fixed by the initial condition. Let us specialize to the simple case of the single-particle initial condition, namely, Pn(t=0)=δn,1P_{n}(t=0)=\delta_{n,1}. This immediately leads to g(z,t=0)=zg(z,t\!=\!0)=z. Then at t=0t=0 the function FF is simply given by F(y)=zF(y)=z. However, we must express the right-hand side in terms of the true dependent variable yy, which means that F(y)=1y1F(y)=1-y^{-1}. Thus for any t0t\geq 0 the generating function is

g(z,t)=F(y+t)=11t+y.\displaystyle g(z,t)=F(y+t)=1-\frac{1}{t+y}\,. (10.3)

To extract the individual terms in the power-series representation of the generating function, we now need to re-express this function in terms of zz:

g(z,t)\displaystyle g(z,t) =11t+1/(1z)=11z(1+t)(1zx)\displaystyle=1-\frac{1}{t+1/(1-z)}=1-\frac{1-z}{(1+t)(1-zx)}
=11z1+tn0(xz)n\displaystyle=1-\frac{1-z}{1+t}\sum_{n\geq 0}(xz)^{n}
=111+tn0(xz)n+z1+tn0(xz)n,\displaystyle=1-\frac{1}{1+t}\sum_{n\geq 0}(xz)^{n}+\frac{z}{1+t}\sum_{n\geq 0}(xz)^{n}\,, (10.4)

where, for notational simplicity, we introduce xt/(1+t)x\equiv t/(1+t). From the last line of the above, we can immediately extract all the Pn(t)P_{n}(t) and obtain the well-known formulas:

P0(t)=t1+tPn(t)=tn1(1+t)n+1n1.\displaystyle P_{0}(t)=\frac{t}{1+t}\qquad\quad P_{n}(t)=\frac{t^{n-1}}{(1+t)^{n+1}}\qquad n\geq 1\,. (10.5)

With these results, we now obtain the first-passage properties of the birth-death process. The quantity P0(t)P_{0}(t) may be interpreted as the probability that the population has gone extinct by time tt, while S(t)=1P0(t)=1/(1+t)S(t)=1-P_{0}(t)=1/(1+t) is the probability that the population survives up to to time tt. Thus extinction is sure to occur, but the average extinction time is infinite, just as for isotropic diffusion. The main distinction with isotropic diffusion is that S(t)t1/2S(t)\sim t^{-1/2} for diffusion, while S(t)t1S(t)\sim t^{-1} for the birth-death process. Thus survival is less likely when the hopping rate is a linearly increasing function of nn.

Concluding Comments

These lecture notes have given a whirlwind tour through some basic and some not-so-basic aspects of first-passage processes. At the level of fundamentals, I presented some classic results about first passage in the simplest geometries of the infinite half line and the finite interval, including first-passage probabilities, first-passage times, and splitting probabilities. I also discussed the intriguing connection between first passage and electrostatics. I then presented a number of applications. Some, like the reaction rate of a cell and the birth-death process are classic and have many immediate applications. Some, like the survival of a diffusing lamb that is hunted by NN diffusing lions and survival of a diffusing particle in a growing interval may seem somewhat idiosyncratic. However, the solution methods are quite generic and may prove useful in many other settings. I hope that the uninitiated reader will enjoy learning about some of these applications of first-passage processes and will be inspired to delve further into this fascinating topic.

Much of the material in Secs. 8 and 9 stems from joint work with Paul Krapivsky. I thank him for pleasant collaborations on these projects, as well as pointing out Ref. [28] to me. I also thank the National Science Foundation for financial support over many years that helped advance some of the topics discussed in these notes, most recently through NSF grant DMR-1910736.

References

  • [1] S. Redner, A Guide to First-Passage Processes (Cambridge University Press, Cambridge, UK 2001).
  • [2] A. J. Bray, S. N. Majumdar, and G. Schehr, Adv. Phys. 62 225 (2015).
  • [3] W. Feller, An Introduction to Probability Theory and Its Applications, Vol. I, 3rd ed. (John Wiley, New York, 1968).
  • [4] N. G. van Kampen, Stochastic Processes in Physics and Chemistry, 2nd ed., (North-Holland, Amsterdam, 1997).
  • [5] S. Karlin and H. M. Taylor, A First Course in Stochastic Processes, 2nd ed. (Academic Press, New York, 2014).
  • [6] E. Montroll, Proceedings of the Symposium on Applied Mathematics (American Mathematical Society, Providence, RI, 1965), 16, 193 (1965).
  • [7] E. Montroll and G. H. Weiss, J. Math. Phys. 6, 167 1965.
  • [8] G. H. Weiss, Aspects and Applications of the Random Walk, (North-Holland, Amsterdam, 1964).
  • [9] B. D. Hughes, Random Walks and Random Environments, (Oxford University Press, New York, 1995).
  • [10] P. G. Doyle and J. L. Snell, Random Walks and Electric Networks Carus Mathematical Monographs #22 (Mathematical Association of America, Oberlin, OH, 1984).
  • [11] H. C. Berg and E. M. Purcell, Biophys. J. 20, 193 (1977).
  • [12] A. Sommerfeld, Math. Annalen (Leipzig) 45, 263 (1894).
  • [13] H. S. Carslaw and J. C Jaeger, Conduction of Heat in Solids 2nd2^{\rm nd} edition, (Oxford University Press, Oxford, 1959).
  • [14] M. Bramson and D. Griffeath, in Random Walks, Brownian Motion, and Interacting Particle Systems: A Festschrift in Honor of Frank Spitzer, 153 eds. R. Durrett and H. Kesten (Birkhäuser, Boston, MA, 1991).
  • [15] J. Galambos, The Asymptotic Theory of Extreme Order Statistics. (Krieger, Malabar, FL, 1987).
  • [16] G. I. Barenblatt, Scaling, Self-similarity, and Intermediate Asymptotics, (Cambridge University Press, Cambridge, UK, 1996).
  • [17] C. M. Bender, & S. A. Orszag, Advanced Mathematical Methods for Scientists and Engineers (McGraw-Hill, New York, 1978).
  • [18] See, for example, L. I. Schiff, Quantum Mechanics, (McGraw-Hill, New York, 1968).
  • [19] L. Breiman, Proc. Fifth Berkeley Symp. Math. Statist. and Probab. 2, 9 (1966).
  • [20] H. E. Daniels, J. Appl. Prob. 6, 399 (1969).
  • [21] K. Uchiyama, Z. Wahrsch. verw. Gebiete 54, 75 (1980).
  • [22] P. Salminen, Adv. Appl. Prob. 20, 411 (1988).
  • [23] P. L. Krapivsky and S. Redner, Am. J. Phys. 64, 546 (1996).
  • [24] A. Khintchine, Fundamental Mathematicae 6, 9 (1924).
  • [25] M. Abramowitz, & I. A. Stegun, Handbook of Mathematical Functions (Dover, New York, 1972).
  • [26] D. G. Kendall, J. Roy. Statist. Soc. Ser. B 11, 230 (1949).
  • [27] P. L. Krapivsky, S. Redner and E. Ben-Naim, A Kinetic View of Statistical Physics (Cambridge University Press, Cambridge, UK, 2010).
  • [28] P. Erdös, Ann. Math. 42, 119 (1942).