This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Most transient random walks have infinitely many cut times

Noah Halberstam Department of Pure Mathematics and Mathematical Statistics, University of Cambridge. Email: [email protected]    Tom Hutchcroft The Division of Physics, Mathematics and Astronomy, California Institute of Technology. Email: [email protected]
Abstract

We prove that if (Xn)n0(X_{n})_{n\geq 0} is a random walk on a transient graph such that the Green’s function decays at least polynomially along the random walk, then (Xn)n0(X_{n})_{n\geq 0} has infinitely many cut times almost surely. This condition applies in particular to any graph of spectral dimension strictly larger than 22. In fact, our proof applies to general (possibly nonreversible) Markov chains satisfying a similar decay condition for the Green’s function that is sharp for birth-death chains. We deduce that a conjecture of Diaconis and Freedman (Ann. Probab. 1980) holds for the same class of Markov chains, and resolve a conjecture of Benjamini, Gurel-Gurevich, and Schramm (Ann. Probab. 2011) on the existence of infinitely many cut times for random walks of positive speed.

1 Introduction

Let (xn)n0(x_{n})_{n\geq 0} be a sequence taking values in some set Ω\Omega. A cut time of (xn)n1(x_{n})_{n\geq 1} is a time n0n\in\operatorname{\mathbb{Z}}_{\geq 0} for which the sets {xi:in}\{x_{i}:i\leq n\} and {xi:i>n}\{x_{i}:i>n\} are disjoint. The study of cut times of random walks was initiated by Erdős and Taylor in 1960 [16], who proved lower bounds on the densities of cut times for simple random walks on the integer lattices d\operatorname{\mathbb{Z}}^{d} for d5d\geq 5, showing that in this case the doubly infinite random walk has a positive density of cut times. The lower dimensional cases d=3,4d=3,4 are more complicated, with the singly infinite random walk having an infinite, density zero set of cut times and the doubly infinite random walk having no cut times almost surely; see [22, 23, 12, 21, 15, 10, 9] for highlights of the literature and [24] for an overview. Extending these results beyond the simple random walk on d\operatorname{\mathbb{Z}}^{d}, James and Peres [19] and Blanchere [6] proved that every centered, finite-range random walk on a transient Cayley graph has infinitely many cut times almost surely; see also also the recent work [27] for a more robust analysis. The proofs of these results rely on delicate estimates on the gradient of the Green’s function that are not available in more general settings, with the works [19, 6] also employing a case analysis of the different possible transitive low-dimensional geometries.

Indeed, while transience is of course a necessary condition for a random walk to have infinitely many cut times, the converse implication quickly breaks down once we leave the transitive setting: James, Lyons and Peres [18] constructed an example of a birth-death chain that is transient but has finitely many cut times almost surely (see also [13]), and Benjamini, Gurel-Gurevich, and Schramm [3] showed that the same behaviour is possible for random walks on bounded degree graphs. On the other hand, Benjamini, Gurel-Gurevich, and Schramm [3] also prove that a graph is transient if and only if the expected number of cut times of the random walk is infinite, which suggests that most ‘non-pathological’ transient random walks should indeed have infinitely many cut times. It is also known that the set of edges crossed by a random walk always spans a recurrent graph almost surely [2, 4], a property that holds trivially when there are infinitely many cut times.

In this paper, we prove a new, very easily satisfied criterion for a transient Markov chain to have infinitely many cut times almost surely, applying in particular to any Markov chain in which the Green’s function decays at least polynomially along a trajectory of the chain. Our result demonstrates that most transient Markov chains arising in examples will have infinitely many cut times almost surely, and, in particular, provides a simple and unified treatment of the transitive locally finite case.

We now state our main theorem. Let M=(Ω,P)M=(\Omega,P) be an irreducible Markov chain consisting of countable state space Ω\Omega and transition kernel PP. For each xΩx\in\Omega, we write x\mathbb{P}_{x} and 𝔼x\mathbb{E}_{x} for probabilities and expectations taken with respect to the law of the Markov chain trajectory (Xn)n0(X_{n})_{n\geq 0} started at xx, and write 𝔾(x,y)\mathbb{G}(x,y) for the Green’s function 𝔾(x,y)=n0Pn(x,y)=𝔼xn0𝟙(Xn=y)\mathbb{G}(x,y)=\sum_{n\geq 0}P^{n}(x,y)=\mathbb{E}_{x}\sum_{n\geq 0}\mathbbm{1}(X_{n}=y). We say that a sequence of non-negative numbers (an)n0(a_{n})_{n\geq 0} decays at least polynomially as nn\to\infty if there exists a constant c>0c>0 and an integer NN such that annca_{n}\leq n^{-c} for every nNn\geq N.

Theorem 1.1.

Let M=(Ω,P)M=(\Omega,P) be a countable Markov chain and let (Xn)n0(X_{n})_{n\geq 0} be a trajectory of MM started at some state xΩx\in\Omega. If there exists a decreasing bijection Φ:[0,)(0,1]\Phi:[0,\infty)\rightarrow(0,1] such that

011u(1logΦ1(u))du= and lim supn𝔾(Xn,Xm)Φ(n)< a.s. for every m0\int_{0}^{1}\frac{1}{u(1\vee\log\Phi^{-1}(u))}\ \mathrm{d}u=\infty\qquad\text{ and }\qquad\limsup_{n\to\infty}\frac{\mathbb{G}(X_{n},X_{m})}{\Phi(n)}<\infty\quad\text{ a.s.\ for every $m\geq 0$} (1)

then the trajectory (Xn)n0(X_{n})_{n\geq 0} has infinitely many cut times almost surely. In particular, the same conclusion holds if 𝔾(Xn,Xm)\mathbb{G}(X_{n},X_{m}) decays at least polynomially as nn\to\infty for each fixed m0m\geq 0 almost surely.

We stress that the Φ1(u)\Phi^{-1}(u) term appearing in (1) denotes the inverse of Φ\Phi rather than its reciprocal. Note that if the Markov chain is irreducible we can replace the decay condition appearing here with the condition that lim supnΦ(n)1𝔾(Xn,X0)<\limsup_{n\to\infty}\Phi(n)^{-1}\mathbb{G}(X_{n},X_{0})<\infty a.s.

Remark 1.2.

Theorem 1.1 applies to some decay rates that are slightly slower than polynomial, such as that given by Φ(n)=exp(lognloglogn)\Phi(n)=\exp({-\frac{\log n}{\log\log n}}). In Section 4, we discuss how the results of Csáki, Földes, and Révész [13] imply that the integral condition of Theorem 1.1 is sharp for birth-death chains and hence cannot be improved in general.

Theorem 1.1 easily implies various sufficient conditions for a Markov chain trajectory to have infinitely many cut times almost surely. One particularly simple such condition is as follows.

Corollary 1.3.

Let M=(Ω,P)M=(\Omega,P) be a countable Markov chain and let X=(Xn)n0X=(X_{n})_{n\geq 0} be a trajectory of MM started at some state xΩx\in\Omega. If for each yΩy\in\Omega there exist constants C=Cxy<C=C_{xy}<\infty and d=dxy>2d=d_{xy}>2 such that Pn(x,y)Cnd/2P^{n}(x,y)\leq Cn^{-d/2} for every n1n\geq 1, then XX has infinitely many cut times almost surely.

Note that if MM is irreducible then the hypothesis of this corollary is equivalent to the on-diagonal heat kernel estimate Pn(x,x)=O(nd/2)P^{n}(x,x)=O(n^{-d/2}) holding for some d>2d>2; for graphs, this is (by definition) equivalent to the spectral dimension of the graph being strictly larger than 22. As such, Corollary 1.3 is already sufficient to treat most natural examples of transient graphs arising in examples.

Proof of Corollary 1.3 given Theorem 1.1.

Fix x,yΩx,y\in\Omega and suppose that C<C<\infty and d>2d>2 are such that Pn(x,y)Cnd/2P^{n}(x,y)\leq Cn^{-d/2} for every n1n\geq 1. We have by the Markov property that

𝔼x[𝔾(Xn,y)]=𝔼x#{visits to y after time n}=m=nPm(x,y)Cm=nmd/22Cd2n(d2)/2\mathbb{E}_{x}\left[\mathbb{G}(X_{n},y)\right]=\mathbb{E}_{x}\#\{\text{visits to $y$ after time $n$}\}=\sum_{m=n}^{\infty}P^{m}(x,y)\leq C\sum_{m=n}^{\infty}m^{-d/2}\leq\frac{2C}{d-2}n^{-(d-2)/2}

for every n1n\geq 1, and hence by Borel-Cantelli that

𝔾(X2k,y)k22(d2)k/2 for all sufficiently large k almost surely.\mathbb{G}(X_{2^{k}},y)\leq k^{2}2^{-(d-2)k/2}\qquad\text{ for all sufficiently large $k$ almost surely.}

If τ\tau denotes the first time after time 2k2^{k} that XX hits yy then the stopped process (𝔾(Xm,y))m=2kτ(\mathbb{G}(X_{m},y))_{m=2^{k}}^{\tau} is a non-negative martingale, and it follows by the optional stopping theorem that

x(there exists m2k such that 𝔾(Xm,y)k42(d2)k/2𝔾(X2n,y)k22(d2)k/2)1k2\mathbb{P}_{x}\left(\text{there exists }m\geq 2^{k}\text{ such that }\mathbb{G}(X_{m},y)\geq k^{4}2^{-(d-2)k/2}\mid\mathbb{G}(X_{2^{n}},y)\leq k^{2}2^{-(d-2)k/2}\right)\leq\frac{1}{k^{2}}

for all sufficiently large kk. Thus, a further application of Borel-Cantelli yields that

𝔾(Xn,y)(log2n)4(n2)(d2)/2\mathbb{G}(X_{n},y)\leq(\log_{2}n)^{4}\left(\frac{n}{2}\right)^{-(d-2)/2} (2)

for all sufficiently large nn almost surely. Since yy was arbitrary, the hypotheses of Theorem 1.1 are satisfied and XX has infinitely many cut times almost surely. ∎

As mentioned above, earlier results concerning random walks on groups relied on relatively fine control of the Green’s function and its gradient, which was used to prove the existence of infinitely many cut times via a second moment argument. The far weaker and more distributed nature of our decay hypothesis causes this second moment argument to break down. Instead, we compare expectations and conditional expectations of certain special types of cut times as the process (𝔾(Xn,X0))n0(\mathbb{G}(X_{n},X_{0}))_{n\geq 0} crosses a small exponential scale [ek1,ek][e^{-k-1},e^{-k}]. Roughly speaking, this allows us to integrate all of the available information across time, compensating for the looser information. See Section 2 for details.

Superdiffusive random walks have infinitely many cut times. As an application of Theorem 1.1, we also prove that walks on graphs and networks (i.e. reversible Markov chains) satisfying a weak superdiffusivity condition have infinitely many cut times almost surely. Given a network N=(V,E,c)N=(V,E,c) with underlying graph (V,E)(V,E) and conductances c:E(0,)c:E\rightarrow(0,\infty), we define the conductance c(v)c(v) of a vertex vv to be the total conductance of all oriented edges emanating from vv.

Theorem 1.4.

Let N=(V,E,c)N=(V,E,c) be a locally finite, connected network with infvc(v)>0\inf_{v}c(v)>0 and let XX be a random walk on NN. If there exists r>3/2r>3/2 such that

lim infnd(X0,Xn)n1/2(logn)r>0 almost surely,\liminf_{n\to\infty}\frac{d(X_{0},X_{n})}{n^{1/2}(\log n)^{r}}>0\qquad\text{ almost surely}, (3)

then XX has infinitely many cut times almost surely.

This result resolves a conjecture of Benjamini, Gurel-Gurevich, and Schramm [3], who asked whether random walks on graphs with positive linear lim inf\liminf speed have infinitely many cut times almost surely. For bounded degree graphs where the walk has positive speed, our proof yields that the walk has a positive density of cut times a.s., yielding a very strong version of their conjecture.

Theorem 1.5.

Let GG be a bounded degree graph and let XX be a random walk on GG.

Iflim infn1nd(X0,Xn)>0a.s.thenlim infn1n#{0mn:m is a cut time for X}>0a.s.\text{If}\quad\liminf_{n\to\infty}\frac{1}{n}d(X_{0},X_{n})>0\quad\text{a.s.}\quad\text{then}\quad\liminf_{n\to\infty}\frac{1}{n}\#\{0\leq m\leq n:m\text{ is a cut time for $X$}\}>0\quad\text{a.s.}

Note that Theorem 1.4 is not an immediate consequence of Theorem 1.1, as we are not aware of any general result allowing us to deduce Green’s function decay estimates from distance estimates without further assumptions on the graph: the Varopoulos-Carne inequality [32, 11] tells us that pm(Xn,X0)p_{m}(X_{n},X_{0}) is small when d(X0,Xn)d(X_{0},X_{n}) is much larger than m1/2m^{1/2}, but does not give any control whatsoever of the large-time contribution to the Green’s function mn2pm(Xn,X0)\sum_{m\geq n^{2}}p_{m}(X_{n},X_{0}). To circumvent this obstacle, we consider adding a spatially-dependent killing to our network. We tune the rate of killing to be weak enough that the walk has a positive chance to live forever when superdiffusive, and strong enough that we can control the decay of the killed Green’s function along the walk. We prove that this killed walk has infinitely many cut times almost surely on the event that it survives forever, from which Theorem 1.4 easily follows.

The Diaconis-Freedman conjecture. Let M=(Ω,P)M=(\Omega,P) be a transient Markov chain, and let X=(Xn)n0X=(X_{n})_{n\geq 0} be a trajectory of MM. The partially exchangeable σ\sigma-algebra of XX is defined to be the exchangeable σ\sigma-algebra generated by the sequence of increments ((Xn,Xn+1))n0((X_{n},X_{n+1}))_{n\geq 0}, that is, the set of events that are determined by the sequence of increments and that are invariant under permutations of this sequence that fix all but finitely many terms. This σ\sigma-algebra arises naturally in the work of Diaconis and Freedman [14], who proved that every partially exchangeable sequence of random variables can be expressed as a Markov process in a random environment. This can be thought of as a partially-exchangeable version of de Finetti’s theorem and plays an important role in the theory of reinforced random walks [1, 31]. Their study of the partially exchangeable σ\sigma-algebra led Diaconis and Freedman to make the following conjecture. Given a trajectory X=(Xn)n0X=(X_{n})_{n\geq 0}, we define the crossing number of an ordered pair of states (x,y)(x,y) to be the number of integers nn such that (Xn,Xn+1)=(x,y)(X_{n},X_{n+1})=(x,y).

Conjecture 1.6 (Diaconis-Freedman 1980).

Let XX be a trajectory of a transient Markov chain. Then the partially exchangeable σ\sigma-algebra of XX is generated by the crossing numbers of XX.

An equivalent statement of this conjecture is that if we condition on the crossing numbers then the resulting process has trivial exchangeable σ\sigma-algebra almost surely. Note that there is a close analogy between this conjecture and the problem of computing the Poisson boundary for lamplighter random walks [30, 20]. As observed in [19], it is easily seen that the Diaconis-Freedman conjecture holds whenever XX has infinitely many cut times almost surely. As such, our main results imply that the Diaconis-Freedman conjecture holds for most transient Markov chains arising in examples.

Corollary 1.7.

Let M=(Ω,P)M=(\Omega,P) be an irreducible transient Markov chain with trajectory (Xn)n0(X_{n})_{n\geq 0}. If (M,x)(M,x) satisfies the hypotheses of either Theorem 1.1 or Corollary 1.3 then the partially exchangeable σ\sigma-algebra of XX is generated by its crossing numbers.

Organisation. Section 2 contains the proof of our main theorem, Theorem 1.1. First, in Section 2.1, we describe the overarching strategy behind the proof of Theorem 1.1 and give a proof in the much simpler special case in which the Green’s function decays exponentially along the random walk. We then introduce relevant technical preliminaries in Sections 2.2 and 2.3 before proving Theorem 1.1 in Section 2.4. Finally, we prove our results concerning superdiffusive walks in Section 3 and prove that Theorem 1.1 is sharp for birth-death chains in Section 4.

Notation. Given a sequence of real numbers (zn)n0(z_{n})_{n\geq 0}, we will often write (zn)n0(z^{*}_{n})_{n\geq 0} for the associated sequence of running minima zn=min0mnzmz_{n}^{*}=\min_{0\leq m\leq n}z_{m}.

2 Proof of the main theorem

In this section we prove our main theorem, Theorem 1.1. We will work mostly under the additional assumption that MM is irreducible, locally finite (i.e. that there are finitely many possible transitions from each state), and has P(x,x)=0P(x,x)=0 except possibly for one absorbing state \dagger, before showing that the general case follows from this case at the end of the proof. It will be convenient to work throughout with the hitting probabilities

(x,y)=x(hit y)=𝔾(x,y)𝔾(y,y)\mathbb{H}(x,y)=\mathbb{P}_{x}(\text{hit $y$})=\frac{\mathbb{G}(x,y)}{\mathbb{G}(y,y)}

rather than the Green’s function. This can be done with minimal changes to each of the other statements since (Xn,y)\mathbb{H}(X_{n},y) decays at the same rate as 𝔾(Xn,y)\mathbb{G}(X_{n},y) for each fixed yy.

Let us now give some relevant definitions. We define a Markov chain with killing to be a tuple M=(Ω,P,)M=(\Omega,P,\dagger) where Ω\Omega is a countable state space, P:Ω×Ω[0,1]P:\Omega\times\Omega\rightarrow[0,1] is the transition kernel and Ω\dagger\in\Omega is a distinguished graveyard state satisfying p(,)=1p(\dagger,\dagger)=1. We say that a Markov chain with killing is locally finite if the set {v:p(u,v)>0}\{v:p(u,v)>0\} is finite for every uΩu\in\Omega and say that a Markov chain with killing is irreducible if for every u,vΩ{}u,v\in\Omega\setminus\{\dagger\} there exists nn\in\operatorname{\mathbb{N}} such that Pn(u,v)>0P^{n}(u,v)>0. We say the chain is transient if every state other than \dagger is visited at most finitely many times almost surely. Given a trajectory XX of a Markov chain with killing, we define for each xΩx\in\Omega the hitting time τx=inf{n0:Xn=x}\tau_{x}=\inf\{n\geq 0:X_{n}=x\}, and say that a trajectory of the chain is killed if τ\stretchrelX<\tau_{\stretchrel*{{{\dagger}}}{X}}<\infty.

Theorem 2.1.

Let M=(Ω,P,)M=(\Omega,P,\dagger) be a transient, locally finite, irreducible Markov chain with killing such that P(x,x)=0P(x,x)=0 for every xx\neq\dagger, let X=(Xn)n0X=(X_{n})_{n\geq 0} be a trajectory of MM, and let ϕ:[0,)[0,)\phi:[0,\infty)\rightarrow[0,\infty) be an increasing bijection such that

n=111log(ϕ1(n))=.\sum_{n=1}^{\infty}\frac{1}{1\vee\log\left(\phi^{-1}(n)\right)}=\infty. (4)

If the event 𝒢={lim supneϕ(n)(Xn,X0)<}\mathscr{G}=\{\limsup_{n\to\infty}e^{\phi(n)}\mathbb{H}(X_{n},X_{0})<\infty\} has positive probability, then XX is either killed or has infinitely many cut times almost surely conditional on 𝒢\mathscr{G}.

Note that (25) becomes equivalent to (1) when Φ(x)=eϕ(x)\Phi(x)=e^{-\phi(x)} as established in the following lemma; we found the condition in terms of Φ\Phi given in Theorem 1.1 to be easier to think about in examples, while the condition in terms of ϕ\phi given in Theorem 2.1 is better suited to the proof.

Lemma 2.2.

Let Φ:[0,)(0,1]\Phi:[0,\infty)\to(0,1] be a decreasing bijection and let ϕ=logΦ\phi=-\log\Phi. Then

011u(1logΦ1(u))du= if and only if n=11(1logϕ1(n))=.\int_{0}^{1}\frac{1}{u(1\vee\log\Phi^{-1}(u))}\mathrm{d}u=\infty\qquad\text{ if and only if }\qquad\sum_{n=1}^{\infty}\frac{1}{(1\vee\log\phi^{-1}(n))}=\infty.
Proof of Lemma 2.2.

We will prove that if the integral involving Φ\Phi diverges then the sum involving ϕ\phi diverges, this being the only direction of the lemma that we need. The reverse direction is proved similarly. Since Φ\Phi is decreasing, we have that

011u(1logΦ1(u))du=k=1ekek+11ulog(1Φ1(u))duk=1ek+1ek(1logΦ1(ek+1))=k=1e(1logΦ1(ek+1)),\int_{0}^{1}\frac{1}{u(1\vee\log\Phi^{-1}(u))}\mathrm{d}u=\sum_{k=1}^{\infty}\int_{e^{-k}}^{e^{-k+1}}\frac{1}{u\log(1\vee\Phi^{-1}(u))}\mathrm{d}u\\ \leq\sum_{k=1}^{\infty}\frac{e^{-k+1}}{e^{-k}(1\vee\log\Phi^{-1}(e^{-k+1}))}=\sum_{k=1}^{\infty}\frac{e}{(1\vee\log\Phi^{-1}(e^{-k+1}))}, (5)

and the claim follows since Φ1(ek+1)=ϕ1(k1)\Phi^{-1}(e^{-k+1})=\phi^{-1}(k-1). ∎

2.1 The overarching strategy and the special case of exponential decay

In this section we describe the high-level strategy underlying Theorem 1.1 and present a proof in the much simpler case of an exponentially decaying hitting probability process (Xn)\mathbb{H}(X_{n}). We then document the issues that arise when attempting to extend this method to the subexponential case and outline how we overcome them.

The high-level idea is to construct a function F:Ω[0,)F:\Omega\rightarrow[0,\infty) such that there are infinitely many times nn when the trajectory (Xm)(X_{m}) of the irreducible Markov chain M=(Ω,P)M=(\Omega,P) satisfies

1. F(Xn)<minm<nF(Xm)F(X_{n})<\min_{m<n}F(X_{m}).       and       2. F(Xm)F(Xn)F(X_{m})\leq F(X_{n}) for m>n.m>n.

In other words, at each of these times nn, the process F(Xi)F(X_{i}) must drop lower than it has previously, and this drop must be a permadrop, i.e. F(Xi)F(X_{i}) must not recover to any level achieved prior to the drop. Indeed, if both 1. and 2. hold then the walk cannot return to any vertex it has previously visited and therefore has a cut time at nn. For the first of these two properties to hold infinitely often, it is sufficient that the process (F(Xn))n0(F(X_{n}))_{n\geq 0} converges to zero, and given transience of the Markov chain, a candidate such as F(x)=d(o,x)1F(x)=d(o,x)^{-1} would suffice. Indeed, studying graph distances appears to be a particularly natural choice in the superdiffusive regime. Unfortunately, there seem to be very limited tools available to prove that this function yields infinitely many permadrops, even when the random walk has positive speed.

These considerations make it natural to instead study the decay of hitting probabilities along the random walk: when the chain is transient the hitting probability process ((Xn,X0))n0(\mathbb{H}(X_{n},X_{0}))_{n\geq 0} automatically tends to zero, and we can use the fact that the process is a martingale to attempt to analyse the number of permadrops. Indeed, Benjamini, Gurel-Gurevich and Schramm [3] used martingale techniques to show that the expected number of permadrops of this process is always infinite when the chain is transient and hence that every transient chain has infinitely many cut times in expectation. Thus, a natural approach to the cut times problem is to find sufficient conditions for the number of permadrops of this process to be infinite almost surely.

Let us first consider the special case in which MM is irreducible and (Xn,X0)\mathbb{H}(X_{n},X_{0}) decays exponentially. Note that this case is already sufficient to resolve the conjecture of Benjamini, Gurel-Gurevich, and Schramm [3] in conjunction with the spatially-dependent-killing argument of Section 3.

Proposition 2.3.

If MM is irreducible and the hitting probability process Zn:=(Xn,X0)Z_{n}:=\mathbb{H}(X_{n},X_{0}) decays exponentially in the sense that lim infn1nlog1/Zn>0\liminf_{n\to\infty}\frac{1}{n}\log 1/Z_{n}>0 a.s. then XX has infinitely many cut times a.s.

The proof of this proposition will rely on Lévy’s zero-one law [26], which is a special case of the martingale convergence theorem.

Lemma 2.4 (Lévy’s zero-one law).

Let (Ω,F,)(\Omega,F,\operatorname{\mathbb{P}}) be a probability space, and let 𝔼\mathbb{E} denote expectation with respect to \operatorname{\mathbb{P}}. Let (Fn)n0(F_{n})_{n\geq 0} be a filtration and let AA be an F=nFnF_{\infty}=\cup_{n}F_{n} measurable event. Then

limk[AFk]=𝟙Aalmost surely.\lim_{k\rightarrow\infty}\mathbb{P}[A\mid F_{k}]=\mathbbm{1}_{A}\qquad\text{almost surely.}
Proof of Proposition 2.3.

Since MM is irreducible, ZnZ_{n} is positive for every n0n\geq 0. Since ZZ also decays exponentially almost surely, the sequence Zn=infmnZmZ^{*}_{n}=\inf_{m\leq n}Z_{m} is also positive and decays exponentially almost surely. In particular, there exists a [0,1][0,1]-valued random variable α\alpha satisfying α<1\alpha<1 almost surely such that

Zn=ZnαZn1Z_{n}=Z_{n}^{*}\leq\alpha Z^{*}_{n-1} (6)

for infinitely many nn.

For each a(0,1)a\in(0,1), define the sequence of times (tna)n0(t_{n}^{a})_{n\geq 0} recursively by setting t0a=0t_{0}^{a}=0 and for each n1n\geq 1 setting

tna=inf{m>tn1a:ZmaZm1}t_{n}^{a}=\inf\{m>t_{n-1}^{a}:Z_{m}\leq aZ^{*}_{m-1}\}

with the convention that inf=\inf\emptyset=\infty. Let Ea={tna<n0}E^{a}=\{t_{n}^{a}<\infty\ \forall n\geq 0\} be the event that there are infinitely many drops of size at least aa and for each n1n\geq 1 consider the permadrop event Ana={tna<A_{n}^{a}=\{t_{n}^{a}<\infty and Zm<a1ZtnaZ_{m}<a^{-1}Z_{t_{n}^{a}} for every mtna}m\geq t^{a}_{n}\}, so that Zm<Ztna1Z_{m}<Z^{*}_{t_{n}^{a}-1} for every mtnam\geq t_{n}^{a} on the event AnaA^{a}_{n}. Let 𝒜a\mathcal{A}^{a} be the event that infinitely many of the events AnaA^{a}_{n} hold, so that 𝒜aEa\mathcal{A}^{a}\subseteq E^{a} and XX has infinitely many cut times whenever 𝒜a\mathcal{A}^{a} holds. Fix a(0,1)a\in(0,1) and suppose that EaE^{a} occurs with positive probability. Since the filtration has the σ\sigma-algebra generated by the entire random walk as its union, Lévy’s zero-one law implies that

limn(tna< and mn s.t. Ama occurstn)=𝟙(𝒜a)\lim_{n\rightarrow\infty}\operatorname{\mathbb{P}}(t_{n}^{a}<\infty\text{ and }\exists m\geq n\text{ s.t. }A_{m}^{a}\text{ occurs}\mid\mathcal{F}_{t_{n}})=\mathbbm{1}(\mathcal{A}^{a})

almost surely. On the other hand, since ZZ is a supermartingale, we have by optional stopping that

(tna< and mn s.t. Ama occurstn)(Anatn)𝟙(tna<)(1a)𝟙(tna<)\operatorname{\mathbb{P}}(t_{n}^{a}<\infty\text{ and }\exists m\geq n\text{ s.t. }A_{m}^{a}\text{ occurs}\mid\mathcal{F}_{t_{n}})\geq\operatorname{\mathbb{P}}(A_{n}^{a}\mid\mathcal{F}_{t_{n}})\mathbbm{1}(t_{n}^{a}<\infty)\geq(1-a)\mathbbm{1}(t_{n}^{a}<\infty)

almost surely for each n1n\geq 1. Since the latter estimate is bounded away from zero as nn\to\infty on the event EaE^{a}, we deduce that 𝒜a\mathcal{A}^{a} holds almost surely conditional on EaE^{a} and hence that XX has infinitely many cut times almost surely conditional on EaE^{a}. The claim follows since a(0,1)a\in(0,1) was arbitrary and the countable union k1E(k1)/k\bigcup_{k\geq 1}E^{(k-1)/k} has probability 11 by (6). ∎

Problems in the subexponential case. As we have just seen, it is straightforward to show that ZnZ_{n} has infinitely many permadrops whenever it decays exponentially: the large decay rate guarantees an infinite supply of drops of a constant relative size, and the optional stopping theorem bounds the probability of each of these drops being a permadrop below by a constant. This constant lower bound means we can rely on soft techniques like Lévy’s zero-one law to deduce that permadrop events occur infinitely often without having to worry about their dependencies. However, even if we did have to think about dependencies, we could choose the drops far enough away from each other such that we could easily control the correlations between their recovery events. (The exact argument is somewhat subtle: it is not necessarily true that the correlations are small, but the conditional probability of there being a permadrop on one scale given what has happened on previous scales is bounded away from 0.)

When we move to the subexponential case, this argument quickly begins to break down. Indeed, the best we were able to do by optimizing the above approach was to handle the case of stretched-exponential decay Zn=eΘ(nz)Z_{n}=e^{-\Theta(n^{z})} for z>1/2z>1/2. Let us now overview the problems that arise when attempting to perform such an optimization. First, without access to Lévy’s zero-one law, we now have to consider correlations between recovery events. Perhaps more significantly, however, subexponential decay gives us only very loose information about the local behaviour of the hitting probability process. We know the extent to which it must decrease over long periods of time, but have relatively little structural information about how this decrease occurs or about the positions and sizes of the drops: the overall fall in value of the process could be made up of frequent small drops, rare large drops, or any combination thereof. Consider for instance the case of stretched exponential decay Zn=eΘ(nz)Z_{n}=e^{-\Theta(n^{z})} for z(0,1)z\in(0,1). This decay could be achieved by drops by a factor of size 1nz11-n^{z-1} at a positive density of times, or, say, by halving at each time of the form n1/zn^{1/z}. The only restriction is that we cannot have too much of the decay made up of very small drops, as this would contradict the assumed decay of the process.

In an attempt to adapt the arguments used in the exponential decay case, a natural starting place would be to attempt to extract a sparse sequence of roughly independent drops of guaranteed size. For instance, in the stretched exponential decay case Zn=eΘ(nz)Z_{n}=e^{-\Theta(n^{z})}, we can set up the infinite sequence of stopping times

tn=inf{m>tn1:Zm<anZtn1 and Zm<(1nz1)Zm}t_{n}=\inf\left\{m>t_{n-1}:Z_{m}<a_{n}Z_{t_{n-1}}\text{ and }Z_{m}<(1-n^{z^{\prime}-1})Z_{m}^{*}\right\}

for some decreasing sequence (an)(a_{n}) very slowly converging to 0, and z(0,z)z^{\prime}\in(0,z), where the sequence ana_{n} should be chosen to allow us to safely ignore dependencies between successive steps. It turns out that this works well for z>1/2z>1/2: a deterministic argument proves that if (Zm)(Z_{m}) has only finitely many drops of any constant relative size, then for nn large enough, the drop at time tnt_{n} must approximately have size at least 1n(z1)/z1-n^{(z^{\prime}-1)/z}, and optional stopping allows us to control the dependencies between the recovery events. Optional stopping then gives a n(z1)/zn^{(z^{\prime}-1)/z} probability of the drop at time tnt_{n} being a permadrop, and a simple generalisation of Borel-Cantelli then implies that there are infinitely many permadrops almost surely. For z12z\leq\frac{1}{2}, however, the sequence n(z1)/zn^{(z^{\prime}-1)/z} has a convergent sum and the argument breaks down. At this stage we are very far from handling polynomial decay!

Addressing the problems. To get results when the process decays slower than en1/2e^{-n^{1/2}}, we can no longer just extract sparse sequences and must begin to consider neighbouring drops and the interactions between their recovery events. We attempted to employ a second moment method, bounding each (AiAj)\operatorname{\mathbb{P}}(A_{i}\cap A_{j}) from above where AiA_{i} is the probability that the iith drop is a permadrop. Unfortunately, due to the looseness of the information that we have regarding the locations and sizes of the drops, this method proved difficult to implement and did not seem capable of producing optimal results. To overcome the outlined issues, we instead analyse the path of the hitting probability process as it traverses a series of spatial scales. At each scale we upper bound the expected number of large permadrops conditional on there being at least one, and simultaneously lower bound the unconditional expected number of large permadrops. We modulate the definition of “large” across scales to ensure that the former quantity is not too large and the latter is not too small: we need the threshold for the drop sizes we consider to be small enough that we get an adequate supply of drops to lower bound the unconditional expectation while being large enough to prevent an accumulation of drops amplifying the conditional expectation. Once we have done this with a well-chosen choice of thresholding function, comparing these two quantities allows us to lower bound the probability that there is a permadrop on each scale; considering a whole scale simultaneously, rather than individual pairs of drops, allowed us to tackle the flexibility present in the structure of the decay. The Borel-Cantelli counterpart then has a natural application demonstrating that there are infinitely many permadrops when the decay of the hitting probability is strong enough.

Rather than working directly with the hitting probability process of the Markov chain, we work with an augmented continuous time process which we call the drawbridge process. This makes the hitting probability process a continuous martingale away from 11 and lets us use optional stopping to get exact expressions for permadrop probabilities rather than one-sided inequalities: this is important since we need to prove both upper and lower bounds on relevant expectations. As mentioned above, we will work primarily in the setting of locally finite Markov chains that are irreducible bar the presence of a graveyard state, before deducing a result for general Markov chains via a simple reduction argument.

2.2 The drawbridge process

At several points in our analysis we will want to apply the optional stopping theorem to get equalities rather than one-sided inequalities, making it convenient to work with continuous rather than discrete martingales. For the random walk on a graph, it is well-known that one can embed the discrete-time random walk inside a continuous-time continuous process by considering Brownian motion on an appropriately constructed metric graph known as the cable graph [17, 28]. We now construct a similar way of embedding a non-reversible locally finite Markov chain inside a continuous Markov process, which we call the drawbridge process, and hence of embedding the discrete-time hitting probability process inside a continuous martingale. While there are precedents for considering similar processes [25], it appears to be much less well known than the cable process, and we give a fairly detailed introduction to keep the paper self-contained.

Before giving a precise definition let us first give the intuition behind the name. Let M=(Ω,P,)M=(\Omega,P,\dagger) be a locally finite Markov chain with killing and suppose that P(x,x)=0P(x,x)=0 for every xx\neq\dagger. Consider the corresponding directed graph GG with vertex set Ω\Omega and with a directed edge from a vertex uu to a vertex vv if uvu\neq v and P(u,v)>0P(u,v)>0. We can make this abstract graph physical, in some sense, by assigning the positive real length 1/P(u,v)1/P(u,v) to each directed edge (u,v)(u,v). While it is nonsensical to think of a Brownian motion which can only travel in one direction, we can recover restrictions in motion through the use of “drawbridges”. More specifically, we envision Brownian motion on a modified version of the metric graph, in which one places a “drawbridge” along each directed edge of the metric graph. Each drawbridge has two states, raised and lowered. When the drawbridge at (u,v)(u,v) is raised, the connection between the part of the edge near vv and the vertex vv itself is severed, and the Brownian motion cannot cross from vv onto the edge (u,v)(u,v). Conversely, when (u,v)(u,v) is lowered it is possible for the Brownian motion to enter the edge from either uu or vv. For each vertex uu, we call the drawbridges across the edges emanating from uu in the corresponding directed graph the outgoing drawbridges from uu. The drawbridge process will be defined by taking the Brownian motion on this metric graph and raising and lowering drawbridges as the Brownian motion moves so that, at each time, the outgoing drawbridges from the last vertex it visited are lowered and all other drawbridges are raised.

We now make this precise. Let M=(Ω,P,)M=(\Omega,P,\dagger) be a locally finite Markov chain with killing such that P(x,x)=0P(x,x)=0 for every xx\neq\dagger. For each state xΩx\in\Omega, we define the set of outgoing states xx^{\rightarrow} to be {yΩ{x}:P(x,y)>0}\{y\in\Omega\setminus\{x\}:P(x,y)>0\}. We define the star graph S[x]S[x] to be the metric graph with vertex set {x}x\{x\}\cup x^{\rightarrow}, with edge set {{x,y}:yx}\{\{x,y\}:y\in x^{\rightarrow}\}, and with edge lengths 1/P(x,y)1/P(x,y), so that S[]S[\dagger] is the metric graph consisting of the single vertex {}\{\dagger\} and no edges. In an abuse of notation, we will identify vertices in the star graph with their corresponding states; the precise meaning will be clear from context. We construct the metric space 𝒮\mathcal{S} from the disjoint union 𝒮=yΩS[y]={(x,y):xΩ,yS[x]}\mathcal{S}^{\sqcup}=\sqcup_{y\in\Omega}S[y]=\{(x,y):x\in\Omega,y\in S[x]\} by gluing together (u,u)(u,u) and (v,u)(v,u) for every vΩv\in\Omega and uvu\in v^{\rightarrow}. Note that every point in 𝒮\mathcal{S} has a unique representation of the form (x,y)(x,y) where xΩx\in\Omega and yS[x]xy\in S[x]\setminus x^{\rightarrow}.

Let (x0,y)𝒮(x_{0},y)\in\mathcal{S} be such that x0Ωx_{0}\in\Omega and yS[x]x0y\in S[x]\setminus x_{0}^{\rightarrow}. We construct the drawbridge process on 𝒮\mathcal{S} starting at (x,y)(x,y) as follows. First we start a Brownian motion (Bt0)t0(B^{0}_{t})_{t\geq 0} on 𝒮\mathcal{S}^{\sqcup} starting at (x0,y)(x_{0},y) at time 𝒯(0)=0\mathcal{T}(0)=0 and run until the stopping time 𝒯(1)=inf{t>𝒯(0):Bt0{x0}×x0}\mathcal{T}(1)=\inf\{t>\mathcal{T}(0):B^{0}_{t}\in\{x_{0}\}\times x_{0}^{\rightarrow}\}, so that if 𝒯(1)<\mathcal{T}(1)<\infty then B𝒯(1)0=(x0,x1)B^{0}_{\mathcal{T}(1)}=(x_{0},x_{1}) for some x1x0x_{1}\in x_{0}^{\rightarrow}. If 𝒯(1)\mathcal{T}(1) is finite, we then run a Brownian motion (Bt1)t=𝒯(1)𝒯(2)(B^{1}_{t})_{t=\mathcal{T}(1)}^{\mathcal{T}(2)} on 𝒮[x1]\mathcal{S}[x_{1}], started from (x1,x1)(x_{1},x_{1}) at time 𝒯(1)\mathcal{T}(1) and run until the stopping time 𝒯(2)=inf{t>𝒯(1):Bt1{x1}×x1}\mathcal{T}(2)=\inf\{t>\mathcal{T}(1):B^{1}_{t}\in\{x_{1}\}\times x_{1}^{\rightarrow}\}. We iterate this construction to generate a possibly infinite sequence ((Bti)𝒯(i)t𝒯(i+1))i((B^{i}_{t})_{\mathcal{T}(i)\leq t\leq\mathcal{T}(i+1)})_{i}, noting that 𝒯(i)\mathcal{T}(i) is almost surely finite whenever xix_{i}\neq\dagger. If the sequence terminates because 𝒯(i)=\mathcal{T}(i)=\infty for some iNi\in N, which almost surely happens exactly when the process first visits the graveyard state \dagger, then we define 𝒯(j)=\mathcal{T}(j)=\infty for j>ij>i, set τ\stretchrelX=i1\tau_{\stretchrel*{{{\dagger}}}{X}}=i-1 and 𝒯=𝒯(τ)\mathcal{T}_{\dagger}=\mathcal{T}(\tau_{\dagger}). If the sequence does not terminate, we set τ\stretchrelX=𝒯=\tau_{\stretchrel*{{{\dagger}}}{X}}=\mathcal{T}_{\dagger}=\infty. Finally, we construct the drawbridge process (𝒳t)t0(\mathcal{X}_{t})_{t\geq 0} by concatenating, in order, the images of the paths of the Brownian motions (Bi)0iτ\stretchrelX(B^{i})_{0\leq i\leq\tau_{\stretchrel*{{{\dagger}}}{X}}} in 𝒮\mathcal{S} under the gluing map 𝒮𝒮\mathcal{S}^{\sqcup}\rightarrow\mathcal{S}.

We let 𝐏x,y\mathbf{P}_{x,y}, 𝐄x,y\mathbf{E}_{x,y} denote probability and expectation with respect to the law of 𝒳\mathcal{X} started at (x,y)(x,y) and write 𝐏x=𝐏x,x\mathbf{P}_{x}=\mathbf{P}_{x,x} and 𝐄x=𝐄x,x\mathbf{E}_{x}=\mathbf{E}_{x,x}. Observe that if 𝒳\mathcal{X} is started at (x,x)(x,x) for some xΩx\in\Omega then the discrete-time process X=(Xn)n=0τX=(X_{n})_{n=0}^{\tau_{\dagger}} defined by (𝒳𝒯(n))n=0τ=(Xn,Xn)n=0τ(\mathcal{X}_{\mathcal{T}(n)})_{n=0}^{\tau_{\dagger}}=(X_{n},X_{n})_{n=0}^{\tau_{\dagger}} has the distribution of a trajectory of the Markov chain stopped when it hits the graveyard state \dagger. Thus, if we fix an arbitrary ‘origin’ state oo\neq\dagger and let 𝒯o=inf{t0:𝒳t=(o,o)}\mathcal{T}_{o}=\inf\{t\geq 0:\mathcal{X}_{t}=(o,o)\} be the hitting time of oo, setting 𝒯o=\mathcal{T}_{o}=\infty if oo is never hit, then the hitting probability (x,y)=𝐏x,y(𝒯o<)\mathcal{H}(x,y)=\mathbf{P}_{x,y}(\mathcal{T}_{o}<\infty) satisfies (x,x)=(x)=(x,o)\mathcal{H}(x,x)=\mathbb{H}(x)=\mathbb{H}(x,o) for all xΩx\in\Omega. We define (𝒵t)t0=((𝒳t))t0(\mathcal{Z}_{t})_{t\geq 0}=(\mathcal{H}(\mathcal{X}_{t}))_{t\geq 0} and (Zn)n0=((Xn))n0=(𝒵𝒯(n))n0(Z_{n})_{n\geq 0}=(\mathbb{H}(X_{n}))_{n\geq 0}=(\mathcal{Z}_{\mathcal{T}(n)})_{n\geq 0}, noting that if σ1σ2\sigma_{1}\leq\sigma_{2} are stopping times for 𝒳\mathcal{X} such that 𝒳to\mathcal{X}_{t}\neq o almost surely for every σ1<t<σ2\sigma_{1}<t<\sigma_{2} then (𝒵t)t=σ1σ2(\mathcal{Z}_{t})_{t=\sigma_{1}}^{\sigma_{2}} is a continuous time martingale with respect to its natural filtration.

2.3 Deterministic preliminaries

We now set up the notation to record the behaviour of the running minima of a non-negative sequence as it converges to zero across a series of exponential scales. Recall that for each sequence (zn)n1(z_{n})_{n\geq 1} we write (zn)n0=(minmnzm)n0(z_{n}^{*})_{n\geq 0}=(\min_{m\leq n}z_{m})_{n\geq 0} for the associated sequence of running minima. Given a non-negative sequence (zn)n0(z_{n})_{n\geq 0} converging to 0 with z0>0z_{0}>0, we construct a sequence of logarithmic scales over which to analyse and control its behaviour. We record the associated notation in the following definition.

Definition 2.5 (Notation for drops at scale kk).

Fix a non-negative sequence z=(zn)n0z=(z_{n})_{n\geq 0} with zn>0z_{n}>0 and zn0z_{n}\to 0 as nn\to\infty. Let k0=k0(z)=2logz01k_{0}=k_{0}(z)=\lceil 2\log z_{0}^{-1}\rceil, and for each k1k\geq 1 define the kkth scale interval In=[ek1,ek]I_{n}=[e^{-{k-1}},e^{-k}]. For each kk0k\geq k_{0} we define the set DkD_{k} by adjoining the set of running minima on the kkth scale to the endpoints of the corresponding interval IkI_{k} so that

Dk=Dk(z)=({zm:m1}Ik){ek,ek1}D_{k}=D_{k}(z)=(\{z_{m}^{*}:m\geq 1\}\cap I_{k})\cup\{e^{-k},e^{-k-1}\}

for each kk0k\geq k_{0}. We define Nk=Nk(z)=|Dk(z)|1N_{k}=N_{k}(z)=\lvert D_{k}(z)\rvert-1 and label the elements of Dk(z)D_{k}(z) in decreasing order as (di,k)1iNk=(di,k(z))0iNk(d_{i,k})_{1\leq i\leq N_{k}}=(d_{i,k}(z))_{0\leq i\leq N_{k}} so that

ek=d0,k>d1,k>>dNk,k=ek1 and i=0Nk1di+1,kdi,k=e1.e^{-k}=d_{0,k}>d_{1,k}>\cdots>d_{N_{k},k}=e^{-k-1}\qquad\text{ and }\qquad\prod_{i=0}^{N_{k}-1}\frac{d_{i+1,k}}{d_{i},k}=e^{-1}.

We call the pairs (di,k,di+1,k)(d_{i,k},d_{i+1,k}) for 0iNk10\leq i\leq N_{k}-1 the drops of zz on scale kk. When context makes clear which sequence zz we are referring to, we will drop it from our notation. Similarly, when it is clear that we are referring to a particular scale kk we will drop the second subscript on the di,kd_{i,k} by writing di=di,kd_{i}=d_{i,k}.

We begin the proof by proving the following deterministic lemma. Roughly speaking, this lemma says that for sequences which decay to zero sufficiently quickly, we can define a threshold between large and small drops in such a way that the following hold:

  1. 1.

    For a good proportion of scales, a good proportion of the decay is made up of large drops.

  2. 2.

    The large drops are large enough that there cannot be too many such drops at any particular scale.

The actual threshold we will use will be a simple function of the ψ\psi which is outputted by this lemma. Later, we will apply this lemma to the hitting probability process. Given an increasing function ψ:[0,)[0,)\psi:[0,\infty)\to[0,\infty) and a sequence of non-negative numbers (zn)n0(z_{n})_{n\geq 0}, we say that (zn)n0(z_{n})_{n\geq 0} is ψ\psi-good on scale kk if

{di+1,kdi,k:0i<Nk is such that di+1,kdi,kexp[k2ψ1(k)]}{di+1,kdi,k:0i<Nk is such that di+1,kdi,k>exp[k2ψ1(k)]},\prod\left\{\frac{d_{i+1,k}}{d_{i,k}}:0\leq i<N_{k}\text{ is such that }\frac{d_{i+1,k}}{d_{i,k}}\leq\exp\left[-\frac{k}{2\psi^{-1}(k)}\right]\right\}\\ \leq\prod\left\{\frac{d_{i+1,k}}{d_{i,k}}:0\leq i<N_{k}\text{ is such that }\frac{d_{i+1,k}}{d_{i,k}}>\exp\left[-\frac{k}{2\psi^{-1}(k)}\right]\right\},

or equivalently if

{di+1,kdi,k:0i<Nk is such that di+1,kdi,kexp[k2ψ1(k)]}e1/2,\prod\left\{\frac{d_{i+1,k}}{d_{i,k}}:0\leq i<N_{k}\text{ is such that }\frac{d_{i+1,k}}{d_{i,k}}\leq\exp\left[-\frac{k}{2\psi^{-1}(k)}\right]\right\}\leq e^{-1/2},

where we write {Ai:iI}=iIAi\prod\{A_{i}:i\in I\}=\prod_{i\in I}A_{i} for reasons of legibility. That is, (zn)n0(z_{n})_{n\geq 0} is ψ\psi-good on scale kk if at least half of the total decay across the scale comes from drops of size at least Ψ(k):=ek/2ψ1(k)\Psi(k):=e^{-k/2\psi^{-1}(k)} in a geometric sense. Note that ψ1\psi^{-1} denotes the inverse of ψ\psi defined by ψ1(x)=min{y0:ψ(y)x}\psi^{-1}(x)=\min\{y\geq 0:\psi(y)\geq x\}; we will typically think of ψ\psi as being a slowly growing function so that ψ1\psi^{-1} satisfies ψ1(x)x\psi^{-1}(x)\gg x.

Lemma 2.6 (Good, well-separated scales).

Let ϕ:[0,)[0,)\phi:[0,\infty)\rightarrow[0,\infty) be an increasing bijection such that

k=111log(ϕ1(k))=.\sum_{k=1}^{\infty}\frac{1}{1\vee\log\left(\phi^{-1}(k)\right)}=\infty. (7)

Then there exist an increasing bijection ψ:[0,)[0,)\psi:[0,\infty)\rightarrow[0,\infty) satisfying ψ(x)ϕ(x)\psi(x)\leq\phi(x) and ψ(x)x\psi(x)\leq\sqrt{x} for every x0x\geq 0 and a strictly increasing function a:a:\operatorname{\mathbb{N}}\rightarrow\operatorname{\mathbb{N}} satisfying

limna(k)a(k1)=+ and k=111log(ψ1(a(k)))=,\lim_{n\rightarrow\infty}a(k)-a(k-1)=+\infty\qquad\text{ and }\qquad\sum_{k=1}^{\infty}\frac{1}{1\vee\log(\psi^{-1}(a(k)))}=\infty, (8)

such that if (zn)n0(z_{n})_{n\geq 0} is a sequence of positive reals with z0>0z_{0}>0 satisfying lim supneϕ(n)zn<\limsup_{n\to\infty}e^{\phi(n)}z_{n}<\infty, then

lim infk1k#{1rk:z is ψ-good on scale a(r)}12.\liminf_{k\rightarrow\infty}\frac{1}{k}\#\left\{1\leq r\leq k:\text{$z$ is $\psi$-good on scale $a(r)$}\right\}\geq\frac{1}{2}. (9)

As mentioned above, the function ψ\psi, which we think of as “ϕ\phi with some room”, is used to define the threshold for a drop on scale kk to be “large”, with zz being ψ\psi-good on scale kk precisely when a good proportion of the total decay on this scale comes from drops that are larger than this threshold. Meanwhile, the sequence aa is used to take a sparse sequence of spatial scales so that we can safely ignore dependencies between scales while keeping various series divergent so that we can still hope to conclude via Borel-Cantelli.

The proof of Lemma 2.6 will rely on the following elementary analytic facts.

Lemma 2.7.

Suppose that f:[0,)f:\operatorname{\mathbb{N}}\rightarrow[0,\infty) is a decreasing function satisfying n=1f(n)=\sum_{n=1}^{\infty}f(n)=\infty.

  1. 1.

    If AA\subseteq\operatorname{\mathbb{N}} has positive density in the sense that lim infN1Nn=1N𝟙(nA)>0\liminf_{N\to\infty}\frac{1}{N}\sum_{n=1}^{N}\mathbbm{1}(n\in A)>0 then

    nAf(n)= and lim infNn=1N𝟙(nA)f(n)n=1Nf(n)>0.\sum_{n\in A}f(n)=\infty\qquad\text{ and }\qquad\liminf_{N\to\infty}\frac{\sum_{n=1}^{N}\mathbbm{1}(n\in A)f(n)}{\sum_{n=1}^{N}f(n)}>0.
  2. 2.

    There exists a convex, strictly increasing function a:a:\operatorname{\mathbb{N}}\rightarrow\operatorname{\mathbb{N}} with limna(n)a(n1)=\lim_{n\rightarrow\infty}a(n)-a(n-1)=\infty such that n=1f(a(n))=\sum_{n=1}^{\infty}f(a(n))=\infty.

Proof of Lemma 2.7.

Fix ff as in the statement of the lemma. We begin with the first statement. Let AA\subseteq\operatorname{\mathbb{N}} be such that lim infN1Nn=1N𝟙(nA)>0\liminf_{N\to\infty}\frac{1}{N}\sum_{n=1}^{N}\mathbbm{1}(n\in A)>0 and let kk be such that lim infN1Nn=1N𝟙(nA)>2/k\liminf_{N\to\infty}\frac{1}{N}\sum_{n=1}^{N}\mathbbm{1}(n\in A)>2/k, so that there exists 0\ell_{0} such that n=k1k+1𝟙(nA)2k1k1=k1\sum_{n=k^{\ell-1}}^{k^{\ell}+1}\mathbbm{1}(n\in A)\geq 2k^{\ell-1}-k^{\ell-1}=k^{\ell-1} for every 0\ell\geq\ell_{0}. Letting Nk0+1N\geq k^{\ell_{0}+1} and setting 1=logkN\ell_{1}=\lfloor\log_{k}N\rfloor, we have that

n=k0+1N𝟙(nA)f(n)=011f(k+1)n=k+1k+1𝟙(nA)=011kf(k+1)\sum_{n=k^{\ell_{0}}+1}^{N}\mathbbm{1}(n\in A)f(n)\geq\sum_{\ell=\ell_{0}}^{\ell_{1}-1}f(k^{\ell+1})\sum_{n=k^{\ell}+1}^{k^{\ell+1}}\mathbbm{1}(n\in A)\geq\sum_{\ell=\ell_{0}}^{\ell_{1}-1}k^{\ell}f(k^{\ell+1})

and that

n=k0+1Nf(n)=01k+1f(k)k0+1f(k0)+k2=011kf(k+1).\sum_{n=k^{\ell_{0}}+1}^{N}f(n)\leq\sum_{\ell=\ell_{0}}^{\ell_{1}}k^{\ell+1}f(k^{\ell})\leq k^{\ell_{0}+1}f(k^{\ell_{0}})+k^{2}\sum_{\ell=\ell_{0}}^{\ell_{1}-1}k^{\ell}f(k^{\ell+1}).

Since ff was assumed to be divergent, it follows that

lim infNn=1N𝟙(nA)f(n)n=1Nf(n)1k2>0,\liminf_{N\to\infty}\frac{\sum_{n=1}^{N}\mathbbm{1}(n\in A)f(n)}{\sum_{n=1}^{N}f(n)}\geq\frac{1}{k^{2}}>0,

and hence that nAf(n)=\sum_{n\in A}f(n)=\infty as claimed.

We now prove the second statement. The first statement implies that for any a,b1a,b\geq 1 there exists m=m(a,b)1m=m(a,b)\geq 1 such that n=1mf(a+bn)1\sum_{n=1}^{m}f(a+bn)\geq 1. This fact allows us to recursively construct a pair of integer sequences (bi)i0(b_{i})_{i\geq 0} and (di)i0(d_{i})_{i\geq 0} by setting b0=0b_{0}=0 and recursively defining

di=min{m:n=1mf(bi+2in)1} and bi+1=bi+2idi for each i0,d_{i}=\min\Bigl{\{}m:\sum_{n=1}^{m}f(b_{i}+2^{i}n)\geq 1\Bigr{\}}\qquad\text{ and }\qquad b_{i+1}=b_{i}+2^{i}d_{i}\qquad\text{ for each $i\geq 0$},

and we observe that both did_{i} and bib_{i} must be finite for i0i\geq 0. We must then have

i=1n=1dif(bi+2in)=n=1f(a(n))=,\sum_{i=1}^{\infty}\sum_{n=1}^{d_{i}}f(b_{i}+2^{i}n)=\sum_{n=1}^{\infty}f(a(n))=\infty,

where a(n)a(n) is the convex, strictly increasing sequence defined by

(a(1),a(2),)=(b1+2,b1+22,,b1+2d1,b2+22,b2+222,,b2+22d2,b3+23,b3+232,)(a(1),a(2),\ldots)=(b_{1}+2,b_{1}+2\cdot 2,\ldots,b_{1}+2\cdot d_{1},b_{2}+2^{2},b_{2}+2^{2}\cdot 2,\ldots,b_{2}+2^{2}\cdot d_{2},b_{3}+2^{3},b_{3}+2^{3}\cdot 2,\ldots)

This sequence has increasing increments tending to infinity by construction, completing the proof. ∎

We now apply Lemma 2.7 to prove Lemma 2.6.

Proof of Lemma 2.6.

We may assume without loss of generality that ϕ(x)x\phi(x)\leq\sqrt{x} for every x0x\geq 0, replacing ϕ\phi with ϕ~=min{ϕ,x}\tilde{\phi}=\min\{\phi,\sqrt{x}\} otherwise. Indeed, since ϕ\phi is increasing, we have that ϕ~1(y)max{ϕ1(y),y2}\tilde{\phi}^{-1}(y)\leq\max\{\phi^{-1}(y),y^{2}\}, and we have by the Cauchy condensation test that

k=111logmax{ϕ1(k),k2}= if and only if k=12k1max{logϕ1(2k),klog4}=.\sum_{k=1}^{\infty}\frac{1}{1\vee\log\max\{\phi^{-1}(k),k^{2}\}}=\infty\quad\text{ if and only if }\quad\sum_{k=1}^{\infty}\frac{2^{k}}{1\vee\max\{\log\phi^{-1}(2^{k}),k\log 4\}}=\infty.

If there are infinitely many kk such that 4kϕ1(2k)4^{k}\geq\phi^{-1}(2^{k}) then the right hand series trivially diverges, while if not then it diverges as a consequence of the Cauchy condensation test applied to k=111logϕ1(k)\sum_{k=1}^{\infty}\frac{1}{1\vee\log\phi^{-1}(k)}.

We begin by applying Lemma 2.7 to the function f(k)=1/1logϕ1(k)f(k)=1/1\vee\log\phi^{-1}(k) to give a strictly increasing function a(k):a(k):\operatorname{\mathbb{N}}\rightarrow\operatorname{\mathbb{N}} such that limka(k)a(k1)=\lim_{k\rightarrow\infty}a(k)-a(k-1)=\infty and

k=111logϕ1(a(k))=.\sum_{k=1}^{\infty}\frac{1}{1\vee\log\phi^{-1}(a(k))}=\infty. (10)

Extend aa arbitrarily to an increasing bijection a:[0,)[0,)a:[0,\infty)\to[0,\infty) and define ψ:[0,)[0,)\psi:[0,\infty)\to[0,\infty) to be the inverse of the increasing bijection

ψ1(x)=8ϕ1(a(8a1(x))),\psi^{-1}(x)=8\phi^{-1}\Bigl{(}a\left(8a^{-1}(x)\right)\Bigr{)},

so that ψ\psi is strictly increasing, bounded above by ϕ\phi, and satisfies

k=111log(ψ1(a(k)))=k=111(log8+log(ϕ1(a(8k))))=\sum_{k=1}^{\infty}\frac{1}{1\vee\log(\psi^{-1}(a(k)))}=\sum_{k=1}^{\infty}\frac{1}{1\vee(\log 8+\log(\phi^{-1}(a(8k))))}=\infty

by Lemma 2.7. Since ϕ\phi and aa are increasing and ϕ(x)x\phi(x)\leq\sqrt{x} for every x0x\geq 0 we also have that ψ1(x)8x2\psi^{-1}(x)\geq 8x^{2} and ψ(x)x\psi(x)\leq\sqrt{x} for every x0x\geq 0.

Let (zn)n0(z_{n})_{n\geq 0} be a sequence of positive numbers and let CC be such that znCeϕ(n)z_{n}\leq Ce^{-\phi(n)} for every n1n\geq 1. We will use the notation of Definition 2.5. Observe that if kk0k\geq k_{0} is such that zz is not ψ\psi-good on scale kk then we must have that

#{i:di+1,kdi,k>exp[12ψ1(k)]}>loge1/2logexp(k/(2ψ1(k))=1kψ1(k).\#\left\{i:\frac{d_{i+1,k}}{d_{i,k}}>\exp\left[{-\frac{1}{2\psi^{-1}(k)}}\right]\right\}>\frac{\log e^{-1/2}}{\log\exp(-k/(2\psi^{-1}(k))}=\frac{1}{k}\psi^{-1}(k).

Discounting the endpoints of the interval, it follows that

#({zm:m0}(ek1,ek))>1kψ1(k)212kψ1(k)\#\bigl{(}\{z_{m}^{*}:m\geq 0\}\cap(e^{-k-1},e^{-k})\bigr{)}>\frac{1}{k}\psi^{-1}(k)-2\geq\frac{1}{2k}\psi^{-1}(k)

whenever zz is not ψ\psi-good on scale kk, where we used that ψ1(k)8k2\psi^{-1}(k)\geq 8k^{2} in the final inequality. Let BB be the set of positive integers kk0k\geq k_{0} such that zz is not ψ\psi-good on scale a(k)a(k) and let

A={kk0:1k|B{k0,,k}|12}.A=\left\{k\geq k_{0}:\frac{1}{k}|B\cap\{k_{0},\ldots,k\}|\geq\frac{1}{2}\right\}.

We wish to prove that AA is finite, and observe that AA is finite if and only if ABA\cap B is finite. For each kABk\in A\cap B with k4k0k\geq 4k_{0}, we have that |B{k/4,k}|k/4|B\cap\{\lfloor k/4\rfloor,\ldots k\}|\geq k/4 and hence, since ψ1\psi^{-1} is increasing, that

|{n:znea(k)1}|\displaystyle|\{n:z_{n}\geq e^{-a(k)-1}\}| iB{k0,,k}|{zm}(ea(i)1,ea(i))|\displaystyle\geq\sum_{i\in B\cap\{k_{0},\ldots,k\}}\lvert\{z_{m}^{*}\}\cap(e^{-a(i)-1},e^{-a(i)})\rvert
12iB{k0,,k}1kψ1(a(i))18ψ1(a(k/4)).\displaystyle\geq\frac{1}{2}\sum_{i\in B\cap\{k_{0},\ldots,k\}}\frac{1}{k}\psi^{-1}(a(i))\geq\frac{1}{8}\psi^{-1}(a(\lfloor k/4\rfloor)).

As such, for each kABk\in A\cap B with k4k04k\geq 4k_{0}\geq 4, there must exist n18ψ1(a(k/4))n\geq\frac{1}{8}\psi^{-1}(a(\lfloor k/4\rfloor)) such that znea(k)1z_{n}\geq e^{-a(k)-1}. On the other hand, for such nn, we also have that

ϕ(n)ϕ(18ψ1(a(k/4)))=a(8k/4)a(2k),\phi(n)\geq\phi\left(\frac{1}{8}\psi^{-1}(a(\lfloor k/4\rfloor))\right)=a\left(8\lfloor k/4\rfloor\right)\geq a(2k),

and since a(2k)a(k)a(2k)-a(k)\to\infty and lim supneϕ(n)zn<\limsup_{n\to\infty}e^{\phi(n)}z_{n}<\infty, we deduce that ABA\cap B is finite as claimed. ∎

2.4 Proof of Theorem 2.1

We now apply the deterministic tools from Section 2.3 to prove Theorem 2.1. Let us fix for the remainder of this subsection a transient Markov chain with killing M=(Ω,P,)M=(\Omega,P,\dagger) with distinguished origin vertex oo such that MM is irreducible, locally finite, and satisfies P(x,x)=0P(x,x)=0 for every xx\neq\dagger. Fix X0Ω{}X_{0}\in\Omega\setminus\{\dagger\}, let (𝒳t)t0(\mathcal{X}_{t})_{t\geq 0} be the drawbridge process started at (X0,X0)(X_{0},X_{0}), let (Xn)n0(X_{n})_{n\geq 0} be the associated discrete-time trajectory of the Markov process, and let (𝒵t)t0(\mathcal{Z}_{t})_{t\geq 0} and (Zn)n0(Z_{n})_{n\geq 0} be the associated continuous- and discrete-time hitting probability processes as defined in Section 2.2. We write 𝐏\mathbf{P} and 𝐄\mathbf{E} for probabilities and expectations taken with respect to the joint law of these processes.

Fix an increasing bijection ϕ:[0,)[0,)\phi:[0,\infty)\to[0,\infty) as in the statement of the theorem, and let aa and ψ\psi be as in Lemma 2.6. For each k1k\geq 1 and 0iNk10\leq i\leq N_{k}-1 we say that the drop (di,k,di+1,k)(d_{i,k},d_{i+1,k}) is a large drop if

di+1,kdi,kΨ(k):=exp[k2ψ1(k)]\frac{d_{i+1,k}}{d_{i,k}}\leq\Psi(k):=\exp\left[-\frac{k}{2\psi^{-1}(k)}\right]

and there exists nn such that Zn=di+1,kZ_{n}^{*}=d_{i+1,k}. (The latter condition can fail if di+1,k=ek1d_{i+1,k}=e^{-k-1}.) If (di,k,di+1,k)(d_{i,k},d_{i+1,k}) is a large drop and ZZ first hits di+1,kd_{i+1,k} at some time nn, we say that (di,k,di+1,k)(d_{i,k},d_{i+1,k}) is a large permadrop if we have additionally that

𝒵t<di,k for every t𝒯(n).\mathcal{Z}_{t}<d_{i,k}\text{ for every $t\geq\mathcal{T}(n)$.}

We say that an arbitrary pair of values (a,b)(a,b) in [ek1,ek][e^{-k-1},e^{-k}] with a>ba>b is a large drop or large permadrop if (a,b)=(di,k,di+1,k)(a,b)=(d_{i,k},d_{i+1,k}) for some large drop or large permadrop (di,k,di+1,k)(d_{i,k},d_{i+1,k}) as appropriate.

Given kk0k\geq k_{0} and i1i\geq 1, we write τi=τi,k\tau_{i}=\tau_{i,k} for the iith time the discrete process ZZ reaches a new running minimum smaller than eke^{-k}, and write 𝒯i=𝒯(τi,k)\mathcal{T}_{i}=\mathcal{T}(\tau_{i,k}) for the corresponding time for the continuous process 𝒵\mathcal{Z}, noting that 𝒯i,k\mathcal{T}_{i,k} is a stopping time for 𝒳\mathcal{X} for each i,k1i,k\geq 1. Note that when 1iNk1\leq i\leq N_{k}, τi,k\tau_{i,k} can be defined equivalently as the first time that Zndi,kZ_{n}\leq d_{i,k}. Recall that if ρ\rho is a stopping time for 𝒳\mathcal{X} then ρ\mathcal{F}_{\rho} denotes the σ\sigma-algebra generated by (𝒳t)t=0ρ(\mathcal{X}_{t})_{t=0}^{\rho}. Given such a stopping time, we lighten notation by writing 𝐄ρ[]=𝐄[ρ]\mathbf{E}_{\rho}[{}\cdot{}]=\mathbf{E}[\ \cdot\mid\mathcal{F}_{\rho}] and 𝐏ρ[]=𝐏[ρ]\mathbf{P}_{\rho}[{}\cdot{}]=\mathbf{P}[\ \cdot\mid\mathcal{F}_{\rho}].

The following two estimates on the distribution of the random variable Rk:=#{0iNk1:(di,k,di+1,k) is a large permadrop}R_{k}:=\#\{0\leq i\leq N_{k}-1:(d_{i,k},d_{i+1,k})\text{ is a large permadrop}\} lie at the heart of the paper.

Proposition 2.8.

The estimate

𝐄𝒯1,k[Rk]14𝐏𝒯1,k(Z is ψ-good on scale k and there exists n such that ek1<Znek3/4)\mathbf{E}_{\mathcal{T}_{1,k}}[R_{k}]\geq\frac{1}{4}\mathbf{P}_{\mathcal{T}_{1,k}}\Bigl{(}\text{\emph{$Z$ is $\psi$-good on scale $k$ and there exists $n$ such that $e^{-k-1}<Z_{n}^{*}\leq e^{-k-3/4}$}}\Bigr{)} (11)

holds almost surely for every k1k\geq 1.

Proposition 2.9.

There exists a universal constant CC such that the estimate

𝐄𝒯1,k[Rk](2+Clog2ψ1(k)k)𝐏𝒯1,k(Rk1)\mathbf{E}_{\mathcal{T}_{1,k}}[R_{k}]\leq\left(2+C\log\frac{2\psi^{-1}(k)}{k}\right)\mathbf{P}_{\mathcal{T}_{1,k}}(R_{k}\geq 1) (12)

holds almost surely for every k1k\geq 1.

Proof of Proposition 2.8.

To lighten notation, we drop kks from subscripts wherever possible. We can use the optional stopping theorem to compute

𝐄𝒯1[R]\displaystyle\mathbf{E}_{\mathcal{T}_{1}}[R] 𝐄𝒯1[i0𝟙(Zτi+1>ek1)𝐏𝒯i+1((di,di+1) is a large permadrop)]\displaystyle\geq\mathbf{E}_{\mathcal{T}_{1}}\left[\sum_{i\geq 0}\mathbbm{1}(Z_{\tau_{i+1}}>e^{-k-1})\mathbf{P}_{\mathcal{T}_{i+1}}((d_{i},d_{i+1})\text{ is a large permadrop})\right]
=𝐄𝒯1[i=0Nk2(1di+1di)𝟙(di+1diΨ)]\displaystyle=\mathbf{E}_{\mathcal{T}_{1}}\left[\sum_{i=0}^{N_{k}-2}\left(1-\frac{d_{i+1}}{d_{i}}\right)\mathbbm{1}\left(\frac{d_{i+1}}{d_{i}}\leq\Psi\right)\right] (13)

and applying the inequality 1xlogx1-x\geq\log x yields that

𝐄𝒯1[R]\displaystyle\mathbf{E}_{\mathcal{T}_{1}}[R] 𝐄𝒯1[log{didi+1:0iNk2,di+1diΨ}].\displaystyle\geq\mathbf{E}_{\mathcal{T}_{1}}\left[\log\prod\left\{\frac{d_{i}}{d_{i+1}}:0\leq i\leq N_{k}-2,\frac{d_{i+1}}{d_{i}}\leq\Psi\right\}\right]. (14)

When ZZ is ψ\psi-good on scale kk and there exists nn such that ek1<Znek3/4e^{-k-1}<Z_{n}^{*}\leq e^{-k-3/4} we have that

{didi+1:0iNk1,di+1diΨ}e1/2 and dNk1dNke1/4,\prod\left\{\frac{d_{i}}{d_{i+1}}:0\leq i\leq N_{k}-1,\frac{d_{i+1}}{d_{i}}\leq\Psi\right\}\geq e^{1/2}\qquad\text{ and }\qquad\frac{d_{N_{k}-1}}{d_{N_{k}}}\leq e^{1/4},

so that the claim follows from (14). ∎

Proof of Proposition 2.9.

We fix a scale Ik=[ek1,ek]I_{k}=[e^{-k-1},e^{-k}] and calculate the conditional expectation of the number of large permadrops on that scale given there is at least one. Since k1k\geq 1 is fixed throughout, we will drop it from notation when possible. We condition on the location on the first permadrop being (a,b)Ik(a,b)\subset I_{k} as follows. Let RR^{\prime} be the number of large permadrops in the scale excluding possibly the last drop, so that

R=0iNk2𝟙((di,di+1) is a large permadrop),R^{\prime}=\sum_{0\leq i\leq N_{k}-2}\mathbbm{1}((d_{i},d_{i+1})\text{ is a large permadrop}),

and let R′′=(R1)0R^{\prime\prime}=(R^{\prime}-1)\vee 0 be the amount that RR^{\prime} exceeds 11. Given an arbitrary pair ek1b<aeke^{-k-1}\leq b<a\leq e^{-k}, we say that (a,b)(a,b) is the first large permadrop if (a,b)=(di,di+1)(a,b)=(d_{i},d_{i+1}) for some 0iNk10\leq i\leq N_{k}-1 such that the pair (di,di+1)(d_{i},d_{i+1}) is a large permadrop and (dj,dj+1)(d_{j},d_{j+1}) is not a large permadrop for any j<ij<i. We write τb\tau_{b} for the first time ZnZ_{n} hits bb (letting τb=\tau_{b}=\infty if this never occurs), write 𝒯b=𝒯(τb)\mathcal{T}_{b}=\mathcal{T}(\tau_{b}), and seek to upper bound the conditional expectation

𝐄𝒯b[𝟙((a,b) is the first large permadrop)R′′].\mathbf{E}_{\mathcal{T}_{b}}\left[\mathbbm{1}((a,b)\text{ is the first large permadrop})R^{\prime\prime}\right].

If b=ek1b=e^{-k-1} then this conditional expectation is zero, so assume b>ek1b>e^{-k-1}.

Note that if a large drop (di,di+1)(d_{i},d_{i+1}) is not a permadrop then there must exist a recovery time at which 𝒵\mathcal{Z} hits did_{i} for the first time after 𝒯i+1\mathcal{T}_{i+1}. Let L=L(a,b)={(di,di+1):1iNk2L=L(a,b)=\{(d_{i},d_{i+1}):1\leq i\leq N_{k}-2, dibd_{i}\leq b, and di+1/diΨ(k)}d_{i+1}/d_{i}\leq\Psi(k)\} be the set of large drops on the scale kk after (a,b)(a,b) and possibly excluding the last drop, let K=K(a,b)K=K(a,b) be the event that (a,b)(a,b) is a drop and that every previous drop in the scale recovers before time 𝒯b\mathcal{T}_{b}, and observe that

𝟙((a,b) is the first large permadrop)R′′=𝟙(K)𝟙((a,b) is a large permadrop)(di,di+1)L𝟙((di,di+1) is a large permadrop).\mathbbm{1}((a,b)\text{ is the first large permadrop})R^{\prime\prime}\\ =\mathbbm{1}(K)\mathbbm{1}((a,b)\text{ is a large permadrop})\sum_{(d_{i},d_{i+1})\in L}\mathbbm{1}((d_{i},d_{i+1})\text{ is a large permadrop}).

Since 𝒯i\mathcal{T}_{i} is a stopping time for each i1i\geq 1, we can use the optional stopping theorem to compute that if b/aΨ(k)b/a\leq\Psi(k) then

𝐄𝒯b[𝟙((a,b) is the first large permadrop)R′′]\displaystyle\mathbf{E}_{\mathcal{T}_{b}}\left[\mathbbm{1}((a,b)\text{ is the first large permadrop})R^{\prime\prime}\right]
=𝐄𝒯b[𝟙(K)(di,di+1)L𝟙((a,b) and (di,di+1) are permadrops)]\displaystyle\hskip 85.35826pt=\mathbf{E}_{\mathcal{T}_{b}}\biggl{[}\mathbbm{1}(K)\sum_{(d_{i},d_{i+1})\in L}\mathbbm{1}((a,b)\text{ and }(d_{i},d_{i+1})\text{ are permadrops})\biggr{]}
=𝐄𝒯b[𝟙(K)(di,di+1)L𝟙(𝒵t<a for t(𝒯b,𝒯i+1) and 𝒵t<di for t>𝒯i+1)]\displaystyle\hskip 85.35826pt=\mathbf{E}_{\mathcal{T}_{b}}\bigg{[}\mathbbm{1}(K)\sum_{(d_{i},d_{i+1})\in L}\mathbbm{1}(\mathcal{Z}_{t}<a\text{ for }t\in(\mathcal{T}_{b},\mathcal{T}_{i+1})\text{ and }\mathcal{Z}_{t}<d_{i}\text{ for }t>\mathcal{T}_{i+1})\bigg{]}
=𝐄𝒯b[𝟙(K)(di,di+1)L(1di+1di)𝟙(𝒵t<a for t(𝒯b,𝒯i+1))]\displaystyle\hskip 85.35826pt=\mathbf{E}_{\mathcal{T}_{b}}\bigg{[}\mathbbm{1}(K)\sum_{(d_{i},d_{i+1})\in L}\bigl{(}1-\frac{d_{i+1}}{d_{i}}\bigr{)}\mathbbm{1}\bigl{(}\mathcal{Z}_{t}<a\text{ for }t\in(\mathcal{T}_{b},\mathcal{T}_{i+1})\bigr{)}\bigg{]}
=𝐄𝒯b[𝟙(K)(di,di+1)L(1di+1di)𝟙((𝒵t)t>𝒯b hits di+1 before a)],\displaystyle\hskip 85.35826pt=\mathbf{E}_{\mathcal{T}_{b}}\bigg{[}\mathbbm{1}(K)\sum_{(d_{i},d_{i+1})\in L}\big{(}1-\frac{d_{i+1}}{d_{i}}\big{)}\mathbbm{1}((\mathcal{Z}_{t})_{t>\mathcal{T}_{b}}\text{ hits }d_{i+1}\text{ before }a)\bigg{]}, (15)

where the third equality follows by taking conditional expectations with respect to 𝒯i+1\mathcal{F}_{\mathcal{T}_{i+1}} for the term in the summation corresponding to (di,di+1)(d_{i},d_{i+1}) and then applying optional stopping.

Write Ψ=Ψ(k)\Psi=\Psi(k) and let =(b)\mathscr{L}=\mathscr{L}(b) be the set of all finite sets SS of ordered pairs of numbers in [ek1,ek][e^{-k-1},e^{-k}] satisfying the following conditions:

  1. 1.

    If (x,y)S(x,y)\in S then xbx\leq b and y/xΨy/x\leq\Psi.

  2. 2.

    If (x,y)(x,y) and (z,w)(z,w) are distinct elements of SS then the open intervals (y,x)(y,x) and (w,z)(w,z) are disjoint.

If we consider the (random) function F:[0,)F:\mathscr{L}\to[0,\infty) defined by

F(S)=(x,y)S(1yx)𝟙((𝒵t)t>𝒯b hits y before a),F(S)=\sum_{(x,y)\in S}\big{(}1-\frac{y}{x}\big{)}\mathbbm{1}((\mathcal{Z}_{t})_{t>\mathcal{T}_{b}}\text{ hits }y\text{ before }a),

then we can rewrite (15) as

𝐄𝒯b[𝟙((a,b) is the first permadrop)Rk′′]𝐄𝒯b[𝟙(K)F(L)],\mathbf{E}_{\mathcal{T}_{b}}\left[\mathbbm{1}((a,b)\text{ is the first permadrop})R_{k}^{\prime\prime}\right]\leq\mathbf{E}_{\mathcal{T}_{b}}\left[\mathbbm{1}(K)F(L)\right], (16)

and we claim that the inequality

F(S)(1Ψ2)i=11/logΨ𝟙((𝒵t)t>𝒯b hits Ψib before a),F(S)\leq(1-\Psi^{2})\sum_{i=1}^{\lceil-1/\log\Psi\rceil}\mathbbm{1}((\mathcal{Z}_{t})_{t>\mathcal{T}_{b}}\text{ hits }\Psi^{i}b\text{ before }a), (17)

is satisfied deterministically for every SS\in\mathscr{L}.

Before proving the claimed inequality (17), let us first see how it implies (12). Substituting (17) into (16) yields that

𝐄𝒯b[𝟙((a,b) is the first permadrop)R′′]\displaystyle\mathbf{E}_{\mathcal{T}_{b}}\left[\mathbbm{1}((a,b)\text{ is the first permadrop})R^{\prime\prime}\right] (1Ψ2)𝐄𝒯b𝟙(K)i=11/logΨ𝟙((𝒵t)t>𝒯b hits Ψib before a)\displaystyle\leq(1-\Psi^{2})\mathbf{E}_{\mathcal{T}_{b}}{\mathbbm{1}(K)\sum_{i=1}^{\lceil-1/\log\Psi\rceil}\mathbbm{1}((\mathcal{Z}_{t})_{t>\mathcal{T}_{b}}\text{ hits }\Psi^{i}b\text{ before }a)}
=(1Ψ2)𝐄𝒯b[𝟙(K)i=11/logΨabaΨib]\displaystyle=(1-\Psi^{2})\mathbf{E}_{\mathcal{T}_{b}}\left[\mathbbm{1}(K)\sum_{i=1}^{\lceil-1/\log\Psi\rceil}\frac{a-b}{a-\Psi^{i}b}\right]
=(1Ψ2)aba𝟙(K)i=11/logΨ11(b/a)Ψi,\displaystyle=(1-\Psi^{2})\frac{a-b}{a}\mathbbm{1}(K)\sum_{i=1}^{\lceil-1/\log\Psi\rceil}\frac{1}{1-(b/a)\Psi^{i}},

where we have applied optional stopping in the second inequality. Since 1/(1xΨi)1/(1-x\Psi^{i}) is an increasing function of xx it follows that if bΨab\leq\Psi a then

𝐄𝒯b[𝟙((a,b) is the first permadrop)R′′]\displaystyle\mathbf{E}_{\mathcal{T}_{b}}\left[\mathbbm{1}((a,b)\text{ is the first permadrop})R^{\prime\prime}\right] abb𝟙(K)(1Ψ2)i=11/logΨ11Ψi+1.\displaystyle\leq\frac{a-b}{b}\mathbbm{1}(K)(1-\Psi^{2})\sum_{i=1}^{-\lceil 1/\log\Psi\rceil}\frac{1}{1-\Psi^{i+1}}.

Now, we have by calculus that 1Ψi+1(i+1)(1Ψ)1-\Psi^{i+1}\geq(i+1)(1-\Psi) for every 1i1/logΨ1\leq i\leq-\lceil 1/\log\Psi\rceil and hence that there exists a universal constant CC such that

𝐄𝒯b[𝟙((a,b) is the first permadrop)R′′]abb𝟙(K)1Ψ21Ψi=11/logΨ1i+1Cabb𝟙(K)log2ψ1(k)k,\mathbf{E}_{\mathcal{T}_{b}}\left[\mathbbm{1}((a,b)\text{ is the first permadrop})R^{\prime\prime}\right]\leq\frac{a-b}{b}\mathbbm{1}(K)\frac{1-\Psi^{2}}{1-\Psi}\sum_{i=1}^{-\lceil 1/\log\Psi\rceil}\frac{1}{i+1}\leq C\frac{a-b}{b}\mathbbm{1}(K)\log\frac{2\psi^{-1}(k)}{k},

where we used the definition of Ψ=Ψ(k)\Psi=\Psi(k) in the final inequality. Since we also have by optional stopping that

𝐏𝒯b((a,b) is the first permadrop))=abb𝟙(K),\mathbf{P}_{\mathcal{T}_{b}}\left((a,b)\text{ is the first permadrop})\right)=\frac{a-b}{b}\mathbbm{1}(K), (18)

we can take expectations over 𝒯b\mathcal{F}_{\mathcal{T}_{b}} conditional on 𝒯1\mathcal{F}_{\mathcal{T}_{1}} to deduce that

𝐄𝒯1[𝟙((a,b) is the first large permadrop)Rk′′]Clog[2ψ1(k)k]𝐏𝒯1((a,b) is the first large permadrop).\mathbf{E}_{{\mathcal{T}_{1}}}\left[\mathbbm{1}((a,b)\text{ is the first large permadrop})R^{\prime\prime}_{k}\right]\\ \leq C\log\left[\frac{2\psi^{-1}(k)}{k}\right]\cdot\mathbf{P}_{{\mathcal{T}_{1}}}\left((a,b)\text{ is the first large permadrop}\right). (19)

We stress that this holds for every pair of real numbers a>ba>b in [ek1,ek][e^{-k-1},e^{-k}], but that both sides will be equal to zero for all but countably many such pairs. Summing over the countable set of pairs giving a non-zero contribution yields that

𝐄𝒯1[R]2𝐏𝒯1(R1)+𝐄𝒯1[R′′](2+Clog2ψ1(k)k)𝐏𝒯1(R1)\mathbf{E}_{\mathcal{T}_{1}}[R]\leq 2\mathbf{P}_{{\mathcal{T}_{1}}}(R\geq 1)+\mathbf{E}_{\mathcal{T}_{1}}[R^{\prime\prime}]\leq\left(2+C\log\frac{2\psi^{-1}(k)}{k}\right)\mathbf{P}_{\mathcal{T}_{1}}(R\geq 1)

as claimed.

It remains to prove (17). Given a set SS\in\mathscr{L}, we say SS is slack if there exists an element (x,y)S(x,y)\in S such that y/x<Ψ2y/x<\Psi^{2} and taut otherwise. Observe that if SS\in\mathscr{L} is slack and (x,y)S(x,y)\in S satisfies y/x<Ψ2y/x<\Psi^{2} then the set S=S{(x,Ψx),(Ψx,y)}{(x,y)}S^{\prime}=S\cup\{(x,\Psi x),(\Psi x,y)\}\setminus\{(x,y)\} also belongs to \mathscr{L} and satisfies F(S)F(S)F(S)\leq F(S^{\prime}). Indeed, the latter inequality follows from the pointwise inequality

𝟙((𝒵t)t>𝒯b hits yi before a)(1yixi)𝟙((𝒵t)t>𝒯b hits Ψxi before a)(1Ψxixi)+𝟙((𝒵t)t>𝒯b hits yi before a)(1yiΨxi).\mathbbm{1}((\mathcal{Z}_{t})_{t>\mathcal{T}_{b}}\text{ hits }y_{i}\text{ before }a)\Bigl{(}1-\frac{y_{i}}{x_{i}}\Bigr{)}\\ \leq\mathbbm{1}((\mathcal{Z}_{t})_{t>\mathcal{T}_{b}}\text{ hits }\Psi x_{i}\text{ before }a)\Bigl{(}1-\frac{{\Psi x_{i}}}{x_{i}}\Bigr{)}+\mathbbm{1}((\mathcal{Z}_{t})_{t>\mathcal{T}_{b}}\text{ hits }y_{i}\text{ before }a)\Bigl{(}1-\frac{y_{i}}{\Psi x_{i}}\Bigr{)}.

To verify this inequality, note that if the indicator on the left is one, then so are both indicators on the right, and when all three indicators are equal to one, the inequality is equivalent to the elementary inequality

1Ψxixi+1yiΨxi(1yixi)=Ψ(1Ψ)xi(1Ψ)yiΨxiΨ(1Ψ)xi(1Ψ)Ψ2xiΨxi=(1Ψ)201-\frac{\Psi x_{i}}{x_{i}}+1-\frac{y_{i}}{\Psi x_{i}}-\left(1-\frac{y_{i}}{x_{i}}\right)=\frac{\Psi(1-\Psi)x_{i}-(1-\Psi)y_{i}}{\Psi x_{i}}\geq\frac{\Psi(1-\Psi)x_{i}-(1-\Psi)\Psi^{2}x_{i}}{\Psi x_{i}}=(1-\Psi)^{2}\geq 0

which holds since yi<Ψ2xiy_{i}<\Psi^{2}x_{i}. Given a slack set SS\in\mathscr{L}, we can therefore iterate this operation until we obtain a taut set SS^{\bullet} with F(S)F(S)F(S)\leq F(S^{\bullet}); this iterative process must terminate after finitely many steps since |S|=|S|+1|S^{\prime}|=|S|+1 and every set in \mathscr{L} contains at most 1/logΨ\lceil-1/\log\Psi\rceil pairs of points. Enumerate the pairs of points of SS^{\bullet} in decreasing order as (x1,y1)(x_{1},y_{1}), (x2,y2)(x_{2},y_{2}), …, (x,y)(x_{\ell},y_{\ell}). Since every pair (x,y)S(x,y)\in S^{\bullet} satisfies y/xΨ2y/x\leq\Psi^{2} and every two distinct pairs of points in SS^{\bullet} span disjoint open intervals of [ek1,b][e^{-k-1},b] we must have that yiΨiby_{i}\leq\Psi^{i}b for every 1i1\leq i\leq\ell and hence that 1/logΨ\ell\leq\lceil-1/\log\Psi\rceil as previously mentioned. It follows that

F(S)F(S\scaleobj0.8)=i=1𝟙((𝒵t)t>𝒯b hits yi before a)(1yixi)i=11/logΨ𝟙((𝒵t)t>𝒯b hits Ψib before a)(1Ψ2),F(S)\leq F(S^{\scaleobj{0.8}{\bullet}})=\sum_{i=1}^{\ell}\mathbbm{1}((\mathcal{Z}_{t})_{t>\mathcal{T}_{b}}\text{ hits }y_{i}\text{ before }a)\big{(}1-\frac{y_{i}}{x_{i}}\big{)}\\ \leq\sum_{i=1}^{\lceil-1/\log\Psi\rceil}\mathbbm{1}((\mathcal{Z}_{t})_{t>\mathcal{T}_{b}}\text{ hits }\Psi^{i}b\text{ before }a)(1-\Psi^{2}),

as claimed, where we used that SS^{\bullet} is taut in the second inequality. ∎

With these bounds in hand, we can now prove Theorem 2.1. The proof will apply the Borel-Cantelli counterpart [8] which is an extension of the second Borel-Cantelli lemma to dependent events.

Lemma 2.10 (Borel-Cantelli Counterpart).

If (En)n0(E_{n})_{n\geq 0} is an increasing sequence of events satisfying the divergence condition n1(EnEn1c)=\sum_{n\geq 1}\operatorname{\mathbb{P}}(E_{n}\mid E_{n-1}^{c})=\infty, then (n1En)=1\operatorname{\mathbb{P}}(\bigcup_{n\geq 1}E_{n})=1.

Setting En=1inAiE_{n}=\cup_{1\leq i\leq n}A_{i} for n1n\geq 1 where (Ai)i1(A_{i})_{i\geq 1} is an arbitrary sequence of events, the Borel-Cantelli counterpart implies in particular that

 if n1(An(i=1n1Ai)c)= then (n1An)=1.\text{ if }\qquad\sum_{n\geq 1}\operatorname{\mathbb{P}}\big{(}A_{n}\mid(\cup_{i=1}^{n-1}A_{i})^{c}\big{)}=\infty\qquad\text{ then }\qquad\operatorname{\mathbb{P}}\Biggl{(}\,\bigcup_{n\geq 1}A_{n}\Biggr{)}=1. (20)
Proof of Theorem 2.1.

We assume that τ=\tau_{\dagger}=\infty with positive probability, the claim holding vacuously otherwise. We continue to use ϕ\phi, ψ\psi, aa, and k0k_{0} as defined above and let 𝒢\mathscr{G} be the event that Zneϕ(n)Z_{n}\leq e^{-\phi(n)} for all sufficiently large nn. Since each permadrop gives rise to a distinct cut time, it suffices to prove that kRk=\sum_{k}R_{k}=\infty almost surely on the event that ZZ is not killed. For each k1k\geq 1, let Ak={Ra(k)1}A_{k}=\{R_{a(k)}\geq 1\} be the event that there is at least one large permadrop on the a(k)a(k)th scale, let BkB_{k} be the event that there exists nn such that ea(k)1<Znea(k)3/4e^{-a(k)-1}<Z_{n}^{*}\leq e^{-a(k)-3/4}, and let CkC_{k} be the event that there exists nn such that ea(k1)1ea(k)Zn<aa(k1)1\sqrt{e^{-a(k-1)-1}e^{-a(k)}}\leq Z_{n}^{*}<a^{-a(k-1)-1} . If (BkCk)c(B_{k}\cap C_{k})^{c} holds for infinitely many kk then ZZ is either killed or satisfies Zn+1e1/4ZnZ_{n+1}^{*}\leq e^{-1/4}Z_{n}^{*} for infinitely many nn, in which case the claim follows from the proof of Proposition 2.3. As such, it suffices to prove that

𝐏(k=k1Ak𝒢 holds, τ=, and BkCk holds for all sufficiently large k)=1 for every k1k0\mathbf{P}\biggl{(}\bigcup_{k=k_{1}}^{\infty}A_{k}\mid\mathscr{G}\text{ holds, }\tau_{\dagger}=\infty,\text{ and }B_{k}\cap C_{k}\text{ holds for all sufficiently large $k$}\biggr{)}=1\text{ for every $k_{1}\geq k_{0}$}

whenever the event being conditioned on has positive probability. We will assume for contradiction that there exists k1k0k_{1}\geq k_{0} such that this does not hold, and fix such a k1k_{1} for the remainder of the proof.

For each kk0k\geq k_{0}, let GkG_{k} be the event that ZZ is ψ\psi-good on scale a(k)a(k). It follows from Propositions 2.8 and 2.9 that there exists a universal constant c>0c>0 such that

𝐏𝒯1,a(k)(Ak)c𝐏𝒯1,a(k)(GkBk)1+logψ1(a(k))\mathbf{P}_{\mathcal{T}_{1,a(k)}}(A_{k})\geq\frac{c\cdot\mathbf{P}_{\mathcal{T}_{1,a(k)}}(G_{k}\cap B_{k})}{1+\log\psi^{-1}(a(k))} (21)

for every k1k\geq 1, where we used that ψ(x)x\psi(x)\leq\sqrt{x} to bound log(ψ1(a(k))/a(k))12logψ1(a(k))\log(\psi^{-1}(a(k))/a(k))\geq\frac{1}{2}\log\psi^{-1}(a(k)). For each kk0k\geq k_{0}, let FkF_{k} be the event that for each <k\ell<k and 0iNa()10\leq i\leq N_{a(\ell)-1} such that the drop (di,a(),di+1,a())(d_{i,a(\ell)},d_{i+1,a(\ell)}) is not a permadrop, the process 𝒵\mathcal{Z} hits di,a()d_{i,a(\ell)} at a time between 𝒯i+1,a()\mathcal{T}_{i+1,a(\ell)} and 𝒯1,a(k)\mathcal{T}_{1,a(k)}. The event FkF_{k} is constructed precisely to force decorrelation between AkA_{k} and (i=k1k1Ai)c(\bigcup_{i=k_{1}}^{k-1}A_{i})^{c}. Indeed, the intersection FkCki=k1k1AiF_{k}\cap C_{k}\setminus\cup_{i=k_{1}}^{k-1}A_{i} is measurable with respect to 𝒯1,a(k)\mathcal{F}_{\mathcal{T}_{1,a(k)}} and we can apply (21) to deduce that

𝐏(AkFkCki=k1k1Ai)\displaystyle\mathbf{P}(A_{k}\cap F_{k}\cap C_{k}\setminus\cup_{i=k_{1}}^{k-1}A_{i}) =𝐄[𝟙(FkCki=k1k1Ai)𝐏𝒯1,a(k)(Ak)]\displaystyle=\mathbf{E}\left[\mathbbm{1}(F_{k}\cap C_{k}\setminus\cup_{i=k_{1}}^{k-1}A_{i})\mathbf{P}_{\mathcal{T}_{1,a(k)}}(A_{k})\right]
c1+logψ1(a(k))𝐄[𝟙(FkCki=k1k1Ai)𝐏𝒯1,a(k)(GkBk)]\displaystyle\geq\frac{c}{1+\log\psi^{-1}(a(k))}\mathbf{E}\left[\mathbbm{1}(F_{k}\cap C_{k}\setminus\cup_{i=k_{1}}^{k-1}A_{i})\mathbf{P}_{\mathcal{T}_{1,a(k)}}(G_{k}\cap B_{k})\right]
=c1+logψ1(a(k))𝐏(FkCkGkBki=k1k1Ai)\displaystyle=\frac{c}{1+\log\psi^{-1}(a(k))}\mathbf{P}(F_{k}\cap C_{k}\cap G_{k}\cap B_{k}\setminus\cup_{i=k_{1}}^{k-1}A_{i}) (22)

for every kk0k\geq k_{0}. On the other hand, it follows by optional stopping that

𝐏(FkcCkGkBki=k1kAi)𝐏(FkcCk)ea(k)+a(k1)+1\mathbf{P}(F_{k}^{c}\cap C_{k}\cap G_{k}\cap B_{k}\setminus\cup_{i=k_{1}}^{k}A_{i})\leq\mathbf{P}(F_{k}^{c}\cap C_{k})\leq\sqrt{e^{-a(k)+a(k-1)+1}}

and hence that

𝐏(Aki=k1k1Ai)\displaystyle\mathbf{P}(A_{k}\setminus\cup_{i=k_{1}}^{k-1}A_{i}) c1+logψ1(a(k))(𝐏(CkGkBki=k1k1Ai)ea(k)+a(k1)+1)0.\displaystyle\geq\frac{c}{1+\log\psi^{-1}(a(k))}\left(\mathbf{P}(C_{k}\cap G_{k}\cap B_{k}\setminus\cup_{i=k_{1}}^{k-1}A_{i})-\sqrt{e^{-a(k)+a(k-1)+1}}\right)\vee 0. (23)

We deduce by linearity of expectation that

k=k1𝐏(Aki=k1k1Ai)c𝐄[k=k1(𝟙(CkGkBki=k1k1Ai)ea(k)+a(k1)+1)01+logψ1(a(k))].\sum_{k=k_{1}}^{\ell}\mathbf{P}(A_{k}\setminus\cup_{i=k_{1}}^{k-1}A_{i})\geq c\cdot\mathbf{E}\left[\sum_{k=k_{1}}^{\ell}\frac{(\mathbbm{1}(C_{k}\cap G_{k}\cap B_{k}\setminus\cup_{i=k_{1}}^{k-1}A_{i})-\sqrt{e^{-a(k)+a(k-1)+1}})\vee 0}{1+\log\psi^{-1}(a(k))}\right]. (24)

On the event that 𝒢i=k1Ai\mathscr{G}\setminus\cup_{i=k_{1}}^{\infty}A_{i} holds and BkCkB_{k}\cap C_{k} holds for all sufficiently large kk (which has positive probability by assumption), we have by choice of ψ\psi in Lemma 2.6 that lim infk1k=k1k𝟙(CkGkBki=k1Ai)>0\liminf_{k\to\infty}\frac{1}{k}\sum_{\ell=k_{1}}^{k}\mathbbm{1}(C_{k}\cap G_{k}\cap B_{k}\setminus\cup_{i=k_{1}}^{\infty}A_{i})>0 and hence by Lemma 2.7 that there exists an almost surely positive η>0\eta>0 and almost surely finite k2k_{2} such that

k=k1(𝟙(CkGkBki=k1k1Ai)1+logψ1(a(k))ηk=k111+logψ1(a(k))\sum_{k=k_{1}}^{\ell}\frac{(\mathbbm{1}(C_{k}\cap G_{k}\cap B_{k}\setminus\cup_{i=k_{1}}^{k-1}A_{i})}{1+\log\psi^{-1}(a(k))}\geq\eta\sum_{k=k_{1}}^{\ell}\frac{1}{1+\log\psi^{-1}(a(k))}

for every k2\ell\geq k_{2}. Since a(k)a(k1)a(k)-a(k-1)\to\infty as kk\to\infty, the other term in (24) is of lower order than this and we deduce that

k=k1𝐏(Aki=k1k1Ai)=.\sum_{k=k_{1}}^{\infty}\mathbf{P}(A_{k}\setminus\cup_{i=k_{1}}^{k-1}A_{i})=\infty.

It follows from the Borel-Cantelli counterpart that 𝐏(k=k1Ak)=1\mathbf{P}(\cup_{k=k_{1}}^{\infty}A_{k})=1, contradicting the definition of k1k_{1}. ∎

2.5 Completing the proof of Theorem 1.1

In this section we deduce Theorem 1.1 from Theorem 2.1. Note that the proof establishes a slightly stronger claim giving the almost sure existence of infinitely many cut times on the event that hitting probabilities decay quickly, without needing to assume that the latter occurs almost surely.

Proof of Theorem 1.1.

Let M=(Ω,P,)M=(\Omega,P,\dagger) be a transient Markov chain with killing, let X=(Xn)n0X=(X_{n})_{n\geq 0} be a trajectory of MM, and let ϕ:[0,)[0,)\phi:[0,\infty)\rightarrow[0,\infty) be a strictly increasing function such that

n=111log(ϕ1(n))=.\sum_{n=1}^{\infty}\frac{1}{1\vee\log\left(\phi^{-1}(n)\right)}=\infty. (25)

It suffices by Lemma 2.2 to prove that if the event 𝒢={lim supneϕ(n)(Xn,Xm)<\mathscr{G}=\{\limsup_{n\to\infty}e^{\phi(n)}\mathbb{H}(X_{n},X_{m})<\infty for every m0m\geq 0 such that Xm}X_{m}\neq\dagger\} has positive probability, then XX is either killed or has infinitely many cut times almost surely conditional on 𝒢\mathscr{G}. Compared to this statement, Theorem 2.1 has three additional hypotheses: that MM is locally finite, that MM is irreducible, and that P(x,x)=0P(x,x)=0 for every xx\neq\dagger. We will show that these assumptions can each be removed via a simple reduction argument.

Removing the condition that P(x,x)=0P(x,x)=0 for every xx\neq\dagger: First suppose that MM is irreducible and locally finite but does not necessarily satisfy P(x,x)=0P(x,x)=0 for every xx\neq\dagger. Since MM is irreducible and transient, P(x,x)1P(x,x)\neq 1 for every xx\neq\dagger. Consider the Markov chain M=(Ω,P,)M^{\prime}=(\Omega,P^{\prime},\dagger) where PP^{\prime} is the transition matrix defined by P(x,x)=0P^{\prime}(x,x)=0 for every xx\neq\dagger and

P(x,y)=P(x,y)1P(x,x) for every xΩ{} and yΩ{x}.P^{\prime}(x,y)=\frac{P(x,y)}{1-P(x,x)}\qquad\text{ for every $x\in\Omega\setminus\{\dagger\}$ and $y\in\Omega\setminus\{x\}$.}

We can couple trajectories XX and YY of MM and MM^{\prime} so that XX visits the same states as YY in the same order but possibly includes additional steps where it stays at the same non-graveyard vertex for more than one consecutive step. In particular, if YY has infinitely many cut times then XX does also. Since the hitting probabilities for MM and MM^{\prime} are equal and Yn{Xm:mn}Y_{n}\in\{X_{m}:m\geq n\} for every n1n\geq 1, if the event 𝒢\mathscr{G} holds for XX then the analogous event holds for YY also, and the claim follows from Theorem 2.1.

Removing the condition that MM is irreducible: Now suppose that MM is locally finite but not necessarily irreducible, and does not necessarily satisfy P(x,x)=0P(x,x)=0 for every xx\neq\dagger. For each communicating class C{}C\neq\{\dagger\} of MM we can define a Markov chain with killing MC=(C{},PC,)M_{C}=(C\cup\{\dagger\},P_{C},\dagger), where PC(u,v)=P(u,v)P_{C}(u,v)=P(u,v) for each u,vCu,v\in C and PC(u,)=vCP(u,v)P_{C}(u,\dagger)=\sum_{v\notin C}P(u,v) for each uCu\in C. When a trajectory XX of the original Markov chain MM enters a communicating class C{}C\neq\{\dagger\}, it can be coupled with a trajectory of MCM_{C} up to the first time that it leaves CC, at which time the coupled trajectory of MCM_{C} is killed. Observe that a trajectory of MM must either pass though infinitely many communicating classes or enter some final communicating class CfC_{f}. If Cf={}C_{f}=\{\dagger\}, the trajectory is killed and there is nothing to prove. Each time the trajectory (Xn)(X_{n}) enters a new communicating class C{}C\neq\{\dagger\}, the coupling with a trajectory of MCM_{C} together with the previous part of the proof implies that, conditional on 𝒢\mathscr{G}, the walk will almost surely either stay in CC forever and have infinitely many cut times or leave CC. Thus, if 𝒢\mathscr{G} holds and XX eventually stays in a single communicating class, then it is either killed or has infinitely many cut times almost surely. On the other hand, if XX visits infinitely many communicating classes then the set of times at which it enters a new communicating class constitute an infinite set of cut times, so that the claim also holds in this case.

Removing the condition that MM is locally finite: We now let MM be arbitrary; it remains only to remove the restriction that it is locally finite. We assume that trajectory XX starts at a non-recurrent state X0ΩX_{0}\in\Omega, the claim holding vacuously otherwise. We merge all the recurrent communicating classes of MM into the single state \dagger to give a Markov chain with killing M=(Ω,P,)M^{\prime}=(\Omega^{\prime},P^{\prime},\dagger), noting that we can couple trajectories of MM and MM^{\prime} such that they are identical up to the first time the two trajectories enter a recurrent communicating class (which corresponds to be killed in MM^{\prime}). We enumerate the states in Ω{}\Omega^{\prime}\setminus\{\dagger\} as (yi)i1(y_{i})_{i\geq 1} and for each state yΩy\in\Omega define y={zΩ:P(y,z)>0}y^{\rightarrow}=\{z\in\Omega:P(y,z)>0\}. Fix ε>0\varepsilon>0. Since every state in Ω{}\Omega^{\prime}\setminus\{\dagger\} is transient, we can select for each i0i\geq 0 a subset LiL_{i} of the states in yiy_{i}^{\rightarrow} such that yiLiy_{i}^{\rightarrow}\setminus L_{i} is finite and the trajectory (Xn)(X_{n}) on MM^{\prime} starting at X0X_{0} satisfies

(j such that Xj=yi and Xj+1Li)<ε2i.\operatorname{\mathbb{P}}(\exists j\in\operatorname{\mathbb{N}}\text{ such that }X_{j}=y_{i}\text{ and }X_{j+1}\in L_{i})<\varepsilon 2^{-i}.

It follows by a union bound that the event ={i,j\mathscr{L}=\{\exists i,j\in\operatorname{\mathbb{N}} such that Xj=yiX_{j}=y_{i} and Xj+1Li}X_{j+1}\in L_{i}\} that the trajectory ever makes a transition of this type has probability at most ε\varepsilon. We construct a new Markov chain with killing M′′=(Ω,P′′,)M^{\prime\prime}=(\Omega^{\prime},P^{\prime\prime},\dagger) where, for each i1i\geq 1, transitions from yiy_{i} to LiL_{i} are redirected to the graveyard state. That is, for each i1i\geq 1, we set P′′(yi,v)=0P^{\prime\prime}(y_{i},v)=0 for every vLiv\in L_{i}, set P′′(yi,v)=P(yi,v)P^{\prime\prime}(y_{i},v)=P^{\prime}(y_{i},v) for each vLi{}v\notin L_{i}\cup\{\dagger\}, and set P′′(yi,)=P(yi,)+vLiP(yi,v)P^{\prime\prime}(y_{i},\dagger)=P^{\prime}(y_{i},\dagger)+\sum_{v\in L_{i}}P^{\prime}(y_{i},v). This construction ensures that M′′M^{\prime\prime} is locally finite. We can couple trajectories XX on MM^{\prime} and YY on M′′M^{\prime\prime} to be identical up until the time that XX makes a transition from yiy_{i} to LiL_{i} for some i1i\geq 1, after which YY is killed. It follows from this coupling that

M(x,y)M′′(x,y) for every x,yΩ{},\mathbb{H}^{M}(x,y)\geq\mathbb{H}^{M^{\prime\prime}}(x,y)\qquad\text{ for every $x,y\in\Omega^{\prime}\setminus\{\dagger\}$}, (26)

and hence under this coupling that M′′(Yn,Ym)M′′(Xn,Xm)\mathbb{H}^{M^{\prime\prime}}(Y_{n},Y_{m})\leq\mathbb{H}^{M^{\prime\prime}}(X_{n},X_{m}) whenever nmn\geq m is such that YnY_{n}\neq\dagger. Since M′′M^{\prime\prime} is locally finite it follows that, under this coupling, YY is either killed or has infinitely many cut times on the event 𝒢={lim supneϕ(n)(Xn,Xm)<\mathscr{G}=\{\limsup_{n\to\infty}e^{\phi(n)}\mathbb{H}(X_{n},X_{m})<\infty for every m0m\geq 0 such that Xm}X_{m}\neq\dagger\}. The claim follows since XX and YY coincide forever with probability at least 1ε1-\varepsilon and ε>0\varepsilon>0 was arbitrary. ∎

3 Superdiffusive walks

In this section we prove Theorem 1.4, which states that random walks on networks satisfying a mild superdiffusivity condition have infinitely many cut times almost surely. It will once again be convenient to work within a more general framework that allows for random walks to be killed. We define a network with killing to be a tuple N=(V,E,c,K)N=(V,E,c,K) where (V,E,c)(V,E,c) is a network and K:V[0,)K:V\to[0,\infty) is a killing function. Given a network with killing N=(V,E,c,K)N=(V,E,c,K), the random walk on NN is the Markov chain with state space V{}V\cup\{\dagger\} and with transition matrix defined by

P(u,v)=c(u,v)c(u)+K(u)for u,vVP(u,)=K(u)c(u)+K(u)for uV, and P(,)=1,P(u,v)=\frac{c(u,v)}{c(u)+K(u)}\quad\text{for $u,v\in V$}\qquad P(u,\dagger)=\frac{K(u)}{c(u)+K(u)}\quad\text{for $u\in V$},\quad\text{ and }\quad P(\dagger,\dagger)=1,

where c(u)c(u) denotes the total conductance of all oriented edges emanating from uu and c(u,v)c(u,v) denotes the total conductance of all oriented edges with tail uu and head vv. We will follow the standard practice of writing pn(u,v)=Pn(u,v)p_{n}(u,v)=P^{n}(u,v) for transition probabilities.

The starting point of our analysis is the following well-known theorem of Varopoulos and Carne [32, 11] (see also [29]). While usually stated without allowing for killing, the same proof applies equally well to networks with killing; the important thing is that PP satisfies the self-adjointness relation (c(u)+K(u))P(u,v)=(c(v)+K(v))P(v,u)(c(u)+K(u))P(u,v)=(c(v)+K(v))P(v,u) for every u,vVu,v\in V and that the restriction of PP to VV is substochastic.

Theorem 3.1 (Varopoulos-Carne Inequality).

The transition probabilities pn(x,y)p_{n}(x,y) of a simple random walk on a network with killing N=(V,E,c,K)N=(V,E,c,K) satisfy

pn(x,y)c(y)+K(y)c(x)+K(x)exp[d(x,y)22n]ρn.p_{n}(x,y)\leq\sqrt{\frac{c(y)+K(y)}{c(x)+K(x)}}\exp\left[-\frac{d(x,y)^{2}}{2n}\right]\rho^{n}. (27)

for every x,yVx,y\in V and n1n\geq 1, where ρ\rho is the spectral radius of the restriction of PP to VV.

We will bound the spectral radius term trivially by 11 in all our applications of this inequality.

While we would naively like to use Varoupoulos-Carne together with our superdiffusivity hypothesis to obtain bounds on the decay of the Green’s function along the random walk, and conclude by applying Theorem 1.1, unfortunately, this does not seem to be possible in general. Indeed, while it is possible to obtain bounds on the small-time and medium-time transition probabilities of the walk using the Varopoulos-Carne inequality, this inequality gives us no control of the large-time contributions to the Green’s function. In our efforts to circumvent this issue, we will establish some rather general conditions under which we can compare the decay of 𝔾(Xn,X0)\mathbb{G}(X_{n},X_{0}) and pn(Xn,X0)p_{n}(X_{n},X_{0}) that may be of independent interest.

3.1 Comparing pn(Xn,X0)p_{n}(X_{n},X_{0}) and 𝔾(Xn,X0)\mathbb{G}(X_{n},X_{0}) assuming superpolynomial decay

The first step of our proof is to give conditions under which the a.s. rates of decay of pn(Xn,X0)p_{n}(X_{n},X_{0}) and 𝔾(Xn,X0)\mathbb{G}(X_{n},X_{0}) can be compared. Given a connected network with killing NN, we say that NN satisfies the superpolynomial decay condition if

limnlogsupuVpn(u,v)logn=0 for some (and hence every) vV.\lim_{n\to\infty}\frac{\log\sup_{u\in V}p_{n}(u,v)}{\log n}=0\qquad\text{ for some (and hence every) $v\in V$.} (SPD)
Proposition 3.2.

Let NN be a network with killing and let X=(Xn)n0X=(X_{n})_{n\geq 0} be a random walk started at oo. If the transition probabilities to oo satisfy the superpolynomial decay condition (SPD) then

limnlogpn(Xn,X0)log𝔾(Xn,X0)=1\lim_{n\to\infty}\frac{\log p_{n}(X_{n},X_{0})}{\log\mathbb{G}(X_{n},X_{0})}=1

almost surely, with the convention that this ratio is equal to 11 when Xn=X_{n}=\dagger.

This proposition is not really needed for Theorem 1.4, since the superpolynomial decay hypothesis (SPD) would already suffice to deduce the claim from Theorem 1.1. It will, however, be used more seriously in the proof of Theorem 1.5. For random walks on finitely generated groups with positive speed, which always satisfy (SPD) by [29, Corollary 6.32], Proposition 3.2 implies that the Avéz entropy and exponential decay rate of the Green’s function coincide, recovering a result of Benjamini and Peres [5, Proposition 6.2]. Similar results for groups that are not finitely generated have been obtained in [7].

Proposition 3.2 will be deduced from the following elementary observation.

Lemma 3.3.

The transition probabilities pn(x,y)p_{n}(x,y) of a simple random walk (Xn)n0(X_{n})_{n\geq 0} on a network with killing N=(V,E,c,K)N=(V,E,c,K) satisfy

𝔼[pm(Xn,X0)pn(Xn,X0)](Xm)+(Xn=)2,\mathbb{E}\left[\frac{p_{m}(X_{n},X_{0})}{p_{n}(X_{n},X_{0})}\right]\leq\mathbb{P}(X_{m}\neq\dagger)+\mathbb{P}(X_{n}=\dagger)\leq 2, (28)

for every xVx\in V and m,n0m,n\geq 0, with the convention that the ratio is 11 when Xn=X_{n}=\dagger.

Proof of Lemma 3.3.

Let AA be the set of vertices xVx\in V such that pn(X0,x)>0p_{n}(X_{0},x)>0. Then we have that

𝔼[pm(Xn,X0)pn(Xn,X0)𝟙(Xn)]=𝔼[pm(X0,Xn)pn(X0,Xn)𝟙(Xn)]=xApn(X0,x)pm(X0,x)pn(X0,x)=xApm(X0,x)(Xm),\mathbb{E}\left[\frac{p_{m}(X_{n},X_{0})}{p_{n}(X_{n},X_{0})}\mathbbm{1}(X_{n}\neq\dagger)\right]=\mathbb{E}\left[\frac{p_{m}(X_{0},X_{n})}{p_{n}(X_{0},X_{n})}\mathbbm{1}(X_{n}\neq\dagger)\right]\\ =\sum_{x\in A}p_{n}(X_{0},x)\frac{p_{m}(X_{0},x)}{p_{n}(X_{0},x)}=\sum_{x\in A}p_{m}(X_{0},x)\leq\mathbb{P}(X_{m}\neq\dagger),

which is easily seen to imply the claim. ∎

Proof of Proposition 3.2.

Using Lemma 3.3, an application of the Borel-Cantelli Lemma implies that there exists an almost surely finite random variable γ\gamma such that

pm(Xn,X0)γ(m+1)2(n+1)2pn(Xn,X0)p_{m}(X_{n},X_{0})\leq\gamma(m+1)^{2}(n+1)^{2}p_{n}(X_{n},X_{0}) (29)

for every n,m0n,m\geq 0. Fix ε>0\varepsilon>0, let n1n\geq 1 and let N=pn(Xn,X0)εN=\lceil p_{n}(X_{n},X_{0})^{-\varepsilon}\rceil. We deduce by summing (29) over 0mN0\leq m\leq N that

𝔾(Xn,X0)=m=0pm(Xn,X0)γ(N+1)3(n+1)2pn(Xn,X0)+m=N+1pm(Xn,X0).\mathbb{G}(X_{n},X_{0})=\sum_{m=0}^{\infty}p_{m}(X_{n},X_{0})\leq\gamma(N+1)^{3}(n+1)^{2}p_{n}(X_{n},X_{0})+\sum_{m=N+1}^{\infty}p_{m}(X_{n},X_{0}).

Since pm(Xn,X0)supvpm(v,X0)p_{m}(X_{n},X_{0})\leq\sup_{v}p_{m}(v,X_{0}) decays superpolynomially in mm by (SPD) we can write this estimate in asymptotic notation as

𝔾(Xn,X0)pn(Xn,X0)13εo(1)+pn(Xn,X0)ω(1) a.s. as n for each fixed ε>0,\mathbb{G}(X_{n},X_{0})\leq p_{n}(X_{n},X_{0})^{1-3\varepsilon-o(1)}+p_{n}(X_{n},X_{0})^{\omega(1)}\qquad\text{ a.s.\ as $n\to\infty$ for each fixed $\varepsilon>0$,}

where o(1)o(1) and ω(1)\omega(1) denote quantities tending to 0 and ++\infty respectively. The claim follows since ε>0\varepsilon>0 was arbitrary and the inequality 𝔾(Xn,X0)pn(Xn,X0)\mathbb{G}(X_{n},X_{0})\geq p_{n}(X_{n},X_{0}) holds trivially. ∎

Since pn(Xn,X0)p_{n}(X_{n},X_{0}) decays superpolynomially under the superdiffusivity assumption (3) by Varopoulos-Carne and we only require polynomial decay of 𝔾(Xn,X0)\mathbb{G}(X_{n},X_{0}) to apply Theorem 1.1, to prove Theorem 1.4 it would suffice for us to have a much weaker comparison of the two quantities than that provided by Proposition 3.2. Such comparison inequalities can be provided by the proof of Proposition 3.2 under much weaker assumptions on transition probabilities that are only barely stronger than transience. For example, this argument is able to handle the ballistic case under the mild additional assumption that there exists c>0c>0 such that

supupn(u,v)Cvn(logn)1+c for every vV and n2,\sup_{u}p_{n}(u,v)\leq\frac{C_{v}}{n(\log n)^{1+c}}\qquad\text{ for every $v\in V$ and $n\geq 2$}, (30)

where CvC_{v} is a finite constant depending on the choice of vv. Unfortunately, we believe that such transition probability estimates need not hold in general, even when the random walk has positive speed. Indeed, identifying the origin of 2\operatorname{\mathbb{Z}}^{2} with the root of a binary tree gives an example where the random walk has positive liminf speed almost surely but where pn(0,0)p_{n}(0,0) is at least the probability that the walk makes an excursion of length nn from the origin to itself in 2\operatorname{\mathbb{Z}}^{2}, which is of order n1(logn)2n^{-1}(\log n)^{-2}. Replacing 2\operatorname{\mathbb{Z}}^{2} in this example by a tree of slightly superquadratic growth should allow one to construct examples where the random walk has positive speed but where (30) does not hold for any c>0c>0; we do not pursue this further here. We believe that there exist examples where the random walk has positive speed but where 𝔾(Xn,X0)\mathbb{G}(X_{n},X_{0}) decays very slowly, but this seems to require a more involved construction.

3.2 Spatially-dependent killing and the proof of Theorem 1.4

We now describe how we circumvent the issue discussed at the end of the previous subsection by introducing spatially dependent killing to our network, where we will take K(x)K(x) to be a function of the distance of xx from some fixed origin vertex oo. We will show under the hypotheses of Theorem 1.4 that this killing function can be chosen to decay sufficiently quickly that the random walk has a positive probability never to be killed, but decay sufficiently slowly that the resulting network with killing satisfies (SPD).

We begin by finding the marginal rate of decay under which the resulting network with killing automatically satisfies (SPD). Given a network N=(V,E,c)N=(V,E,c) and a fixed origin vertex oo, we write x=2d(o,x)\langle x\rangle=2\vee d(o,x) for each xVx\in V to avoid division by zero.

Lemma 3.4.

Let N=(V,E,c)N=(V,E,c) be a network with cmin=infxVc(x)>0c_{\min}=\inf_{x\in V}c(x)>0, fix a vertex oVo\in V, let γ\gamma\in\operatorname{\mathbb{R}} and let K:V[0,)K:V\to[0,\infty) be the killing function defined by K(x)=c(x)min{1,K(x)=c(x)\min\{1, x2(logx)γ}\langle x\rangle^{-2}(\log\langle x\rangle)^{\gamma}\}. Then there exists a positive constant c=c(γ)c=c(\gamma) such that

pn(x,o)8c(o)cminexp[c(logn)γ/2]p_{n}(x,o)\leq\sqrt{\frac{8c(o)}{c_{\min}}}\exp\Big{[}-c\,(\log n)^{\gamma/2}\Big{]}

for every xVx\in V and n2n\geq 2. In particular, if γ>2\gamma>2 then (V,E,c,K)(V,E,c,K) satisfies (SPD).

The rough idea behind this lemma is as follows: Suppose we run a random walk for time nn started at some vertex xx. If d(o,Xm)nd(o,X_{m})\gg\sqrt{n} for some 0mn0\leq m\leq n then the probability of hitting the origin at time nn is small as a consequence of Varopoulos-Carne. On the other hand, if this never happens, the higher rate of killing ensures that the walk is killed before time nn with high probability and is therefore unlikely to hit the origin at time nn.

Proof of Lemma 3.4.

Let x\mathbb{P}_{x} denote the law of the random walk X=(Xn)n0X=(X_{n})_{n\geq 0} on the network with killing (V,E,c,K)(V,E,c,K) started at some fixed vertex xVx\in V, and let τ\tau_{\dagger} denote the time the walk is killed (i.e. first visits the graveyard state \dagger). We define d(o,)=d(o,\dagger)=\infty and decompose

x(Xn=o)=x(Xn=o and d(o,Xm)>r for some 0mn)+x(Xn=o and d(o,Xm)r for every 0mn)\mathbb{P}_{x}(X_{n}=o)=\mathbb{P}_{x}(X_{n}=o\text{ and }d(o,X_{m})>r\text{ for some }0\leq m\leq n)\\ +\mathbb{P}_{x}(X_{n}=o\text{ and }d(o,X_{m})\leq r\text{ for every $0\leq m\leq n$}) (31)

for each n,r2n,r\geq 2, where rr is a parameter we will optimize over at the end of the proof. We begin by analysing the first term on the right hand side of (31). Let κ\kappa be the stopping time κ:=inf{m0:d(o,Xm)>r}\kappa:=\inf\{m\geq 0:d(o,X_{m})>r\}. We apply the strong Markov property at κ\kappa together with Varopoulos-Carne to give that

x(Xn=o and d(o,Xm)>r for some 0mn)\displaystyle\mathbb{P}_{x}(X_{n}=o\text{ and }d(o,X_{m})>r\text{ for some }0\leq m\leq n) m=0nzVx(κ=m,Xκ=z)z(Xnm=o)\displaystyle\leq\sum_{m=0}^{n}\sum_{z\in V}\mathbb{P}_{x}(\kappa=m,X_{\kappa}=z)\mathbb{P}_{z}(X_{n-m}=o)
c(o)+K(o)c(z)+K(z)exp[r22n]2c(o)cminexp[r22n],\displaystyle\leq\sqrt{\frac{c(o)+K(o)}{c(z)+K(z)}}\exp\left[-\frac{r^{2}}{2n}\right]\leq\sqrt{\frac{2c(o)}{c_{\min}}}\exp\left[-\frac{r^{2}}{2n}\right],

where cmin=infzVc(z)c_{\min}=\inf_{z\in V}c(z), and where the final inequality follows by definition of KK. We now turn our attention to the second term on the right hand side of (31). Each time the walk makes a step at distance at most rr it is killed with probability at least 12(1r2(logr)γ)\frac{1}{2}(1\wedge r^{-2}(\log r)^{\gamma}). Letting c1=c2(γ)c_{1}=c_{2}(\gamma) be a positive constant such that this probability is at least c1r2(logr)γc_{1}r^{-2}(\log r)^{\gamma}, we deduce that

x(Xn=o and d(o,Xm)r for every 0mn)\displaystyle\mathbb{P}_{x}\bigl{(}X_{n}=o\text{ and }d(o,X_{m})\leq r\text{ for every $0\leq m\leq n$}\bigr{)} x(τ>n and d(o,Xm)r for every 0mn)\displaystyle\leq\mathbb{P}_{x}\bigl{(}\tau_{\dagger}>n\text{ and }d(o,X_{m})\leq r\text{ for every $0\leq m\leq n$}\bigr{)}
(1c1(logr)γr2)nexp[c1(logr)γnr2],\displaystyle\leq\left(1-\frac{c_{1}(\log r)^{\gamma}}{r^{2}}\right)^{n}\leq\exp\left[-\frac{c_{1}(\log r)^{\gamma}n}{r^{2}}\right],

where we used the inequality 1tet1-t\leq e^{-t} in the final inequality. Substituting these two estimates into (31) yields that

x(Xn=o)2c(o)cmin(exp[r22n]+exp[c1(logr)γnr2]),\mathbb{P}_{x}(X_{n}=o)\leq\sqrt{\frac{2c(o)}{c_{\min}}}\left(\exp\left[-\frac{r^{2}}{2n}\right]+\exp\left[-\frac{c_{1}(\log r)^{\gamma}n}{r^{2}}\right]\right),

and the claim follows by taking r=n1/2(logn)γ/4r=\lceil n^{1/2}(\log n)^{\gamma/4}\rceil. ∎

Let N=(V,E,c)N=(V,E,c) be a network, let oo be a vertex of NN, and let X=(Xn)n0X=(X_{n})_{n\geq 0} be the random walk on NN. Let r>0r>0 and let 𝒮r\mathscr{S}_{r} be the event that

lim infnd(o,Xn)n1/2(logn)r>0.\liminf_{n\to\infty}\frac{d(o,X_{n})}{n^{1/2}(\log n)^{r}}>0.

We next wish to show that for any choice of rr, we can choose the killing function KK as in Lemma 3.4 such that if 𝒮r\mathscr{S}_{r} holds, the walk does not “feel” the effects of the killing. More precisely, we can ensure the killing function decays quickly enough such that conditional on the path of the walk, the walk almost surely has a positive probability of never getting killed. To formulate this lemma, let us first note that we can couple the random walks on (V,E,c)(V,E,c) and (V,E,c,K)(V,E,c,K) so that they coincide up until the killing time τ\tau_{\dagger}. Writing XX for the unkilled walk and writing 𝐏x\mathbf{P}_{x} for the joint law of XX and τ\tau_{\dagger} when XX is started at xVx\in V, this coupling is determined by the equality

𝐏x(τ=nX)=K(Xn1)i=0n2(1K(Xi)).\mathbf{P}_{x}(\tau_{\dagger}=n\mid X)=K(X_{n-1})\prod_{i=0}^{n-2}(1-K(X_{i})).
Lemma 3.5.

Let N=(V,E,c)N=(V,E,c) be a network with cmin=infxVc(x)>0c_{\min}=\inf_{x\in V}c(x)>0, fix a vertex oVo\in V, let γ\gamma\in\operatorname{\mathbb{R}}, and let K:V[0,)K:V\to[0,\infty) be the killing function defined by K(x)=c(x)min{1,K(x)=c(x)\min\{1, x2(logx)γ}\langle x\rangle^{-2}(\log\langle x\rangle)^{\gamma}\}. If XX is a random walk on NN and γ+1<2r\gamma+1<2r, then 𝐏x(τ=X)>0\mathbf{P}_{x}(\tau_{\dagger}=\infty\mid X)>0 almost surely on the event 𝒮r\mathscr{S}_{r}.

Proof of Lemma 3.5.

We can write the conditional probability 𝐏x(τ=X)\mathbf{P}_{x}(\tau_{\dagger}=\infty\mid X) as an infinite product

𝐏x(τ=X)=i=0(1K(Xi)), which is positive if and only if i=0K(Xi)<.\mathbf{P}_{x}(\tau_{\dagger}=\infty\mid X)=\prod_{i=0}^{\infty}(1-K(X_{i})),\quad\text{ which is positive if and only if }\quad\sum_{i=0}^{\infty}K(X_{i})<\infty.

We have by calculus that there exists a random variable α\alpha taking values in [1,][1,\infty] that is finite on the event 𝒮r\mathscr{S}_{r} and satisfies K(Xn)α(logn)γ2rn1K(X_{n})\leq\alpha(\log n)^{\gamma-2r}n^{-1} for every n1n\geq 1, and it follows that if 2r>1+γ2r>1+\gamma then i=0K(Xi)<\sum_{i=0}^{\infty}K(X_{i})<\infty on the event 𝒮r\mathscr{S}_{r} as required. ∎

We are now ready to complete the proofs of Theorems 1.4 and 1.5.

Proof of Theorem 1.4.

Let r>3/2r>3/2 and 2<γ<2r12<\gamma<2r-1. Let N=(V,E,c)N=(V,E,c) be a network with cmin=infxVc(x)>0c_{\min}=\inf_{x\in V}c(x)>0, fix a vertex oVo\in V, and let K:V[0,)K:V\to[0,\infty) be the killing function defined by K(x)=c(x)min{1,K(x)=c(x)\min\{1, x2(logx)γ}\langle x\rangle^{-2}(\log\langle x\rangle)^{\gamma}\}. Couple the random walk XX on NN with the killing time τ\tau_{\dagger} as above, write XX^{\dagger} for the killed walk, and assume that the superdiffusivity event 𝒮r\mathscr{S}_{r} has positive probability. Let pnp^{\dagger}_{n} and 𝔾\mathbb{G}_{\dagger} denote transition probabilities and the Green’s function with respect to the killed network N=(V,E,c,K)N_{\dagger}=(V,E,c,K). Lemma 3.4 implies that NN_{\dagger} satisfies the superpolynomial decay condition (SPD), and we deduce from Proposition 3.2 that

limnlogpn(Xn,X0)log𝔾(Xn,X0)=1\lim_{n\to\infty}\frac{\log p^{\dagger}_{n}(X_{n}^{\dagger},X_{0}^{\dagger})}{\log\mathbb{G}_{\dagger}(X_{n}^{\dagger},X_{0}^{\dagger})}=1 (32)

almost surely, where the ratio is considered to be equal to 11 when Xn=X_{n}^{\dagger}=\dagger. Varopoulos-Carne yields that

pn(Xn,X0)\displaystyle p^{\dagger}_{n}(X_{n}^{\dagger},X_{0}^{\dagger}) =exp[Ω((logn)2r)]\displaystyle=\exp\left[-\Omega\left((\log n)^{2r}\right)\right] as nn\to\infty
when 𝒮r\mathscr{S}_{r} holds, and hence by (32) that
𝔾(Xn,X0)\displaystyle\mathbb{G}_{\dagger}(X_{n}^{\dagger},X_{0}^{\dagger}) =exp[Ω((logn)2r)]\displaystyle=\exp\left[-\Omega\left((\log n)^{2r}\right)\right] as nn\to\infty

almost surely on the event 𝒮r\mathscr{S}_{r}. (Here we recall that Ω(f(n))\Omega(f(n)) denotes a quantity that is lower bonded by a (possibly random) positive multiple of f(n)f(n) for large values of nn.) Since r>3/2>1/2r>3/2>1/2, this decay is superpolynomial, and it follows from Theorem 1.1 that XX^{\dagger} is either killed or has infinitely many cut times almost surely on the event 𝒮r\mathscr{S}_{r}. Since we also have that the conditional probability 𝐏x(τ=X)\mathbf{P}_{x}(\tau_{\dagger}=\infty\mid X) is almost surely positive on the event 𝒮r\mathscr{S}_{r}, we deduce that XX has infinitely many cut times almost surely on the event 𝒮r\mathscr{S}_{r} as claimed. ∎

Proof of Theorem 1.5.

Since GG has bounded degrees and the walk has positive liminf speed almost surely, it follows as above that we can take a bounded killing function KK so that the walk has a.s. positive conditional probability not to be killed and the killed Green’s function 𝔾(Xn,Xo)\mathbb{G}_{\dagger}(X_{n}^{\dagger},X_{o}^{\dagger}) decays exponentially. On the other hand, since the degrees and the killing function are both bounded, there exists a positive constant cc such that 𝔾(Xn+1,Xo)c𝔾(Xn,Xo)\mathbb{G}_{\dagger}(X_{n+1}^{\dagger},X_{o}^{\dagger})\geq c\cdot\mathbb{G}_{\dagger}(X_{n}^{\dagger},X_{o}^{\dagger}) for every n0n\geq 0 such that Xn+1X_{n+1}^{\dagger}\neq\dagger. Combined with exponential decay this implies that if we define Aa={n:Zn+1aminmnZm}A_{a}=\{n:Z_{n+1}^{\dagger}\leq a\min_{m\leq n}Z_{m}^{\dagger}\}, where Zn=𝔾(Xn,Xo)Z_{n}^{\dagger}=\mathbb{G}_{\dagger}(X_{n}^{\dagger},X_{o}^{\dagger}), and define 𝒜a\mathcal{A}_{a} to be the event that AaA_{a} has positive lim inf\liminf density then k=1𝒜(k1)/k\bigcup_{k=1}^{\infty}\mathcal{A}_{(k-1)/k} has probability 11. On the other hand, by optional stopping, for each n1n\geq 1 the conditional probability that nn is a cut time given everything the walk has done up to time nn is bounded below by 1a1-a whenever nAan\in A_{a}. From here the claim follows easily by standard arguments and we omit the details. ∎

4 Sharpness for birth-death chains

In this final section, we demonstrate that the integral condition given in Theorem 1.1 is sharp by comparing our results to those of Csáki, Földes, and Révész [13] on the cut times of birth-death chains. Throughout this section, (Xn)n0(X_{n})_{n\geq 0} will denote a random walk on 0\operatorname{\mathbb{Z}}_{\geq 0} with transition probabilities of the form

Ei:=(Xn+1=i+1Xn=i)=1(Xn+1=i1Xn=i)={1if i=01/2+piotherwise,E_{i}:=\operatorname{\mathbb{P}}(X_{n+1}=i+1\mid X_{n}=i)=1-\operatorname{\mathbb{P}}(X_{n+1}=i-1\mid X_{n}=i)=\begin{cases}1&\text{if }i=0\\ 1/2+p_{i}&\text{otherwise}\end{cases},

where 1/2<pi<1/2-1/2<p_{i}<1/2 for each i1i\geq 1. For each m0m\geq 0, define

D(m)=1+j=1i=1j(1Em+i1).D(m)=1+\sum_{j=1}^{\infty}\prod_{i=1}^{j}\bigg{(}\frac{1}{E_{m+i}}-1\bigg{)}.

The aforementioned work [13] establishes the following dichotomy. (Here we rephrase their theorem in terms of cut times and omit the strengthened conclusion concerning strong cut points.)

Theorem 4.1 ([13, Theorem 1.1]).

Let (Xn)n0(X_{n})_{n\geq 0} be a transient birth-death chain as defined above with 0pi<1/20\leq p_{i}<1/2 for each i1i\geq 1.

  • If n=2(D(n)logn)1<\sum_{n=2}^{\infty}(D(n)\log n)^{-1}<\infty, then (Xn)(X_{n}) has finitely many cut times a.s.

  • If D(n)n(logn)1/2D(n)\leq n(\log n)^{1/2} and n=2(D(n)logn)1=\sum_{n=2}^{\infty}(D(n)\log n)^{-1}=\infty, then (Xn)(X_{n}) has infinitely many cut times a.s.

We use this Theorem to prove the following partial converse of Theorem 1.1. We let 𝔾(n)=𝔾(n,0)\mathbb{G}(n)=\mathbb{G}(n,0) denote the Green’s function associated with (Xn)(X_{n}) and say a function F:[0,)(0,)F:[0,\infty)\rightarrow(0,\infty) is eventually log-convex if there exists r0r\geq 0 such that the restriction of FF to the interval [r,)[r,\infty) is log-convex.

Proposition 4.2.

Given any decreasing differentiable bijection Φ:[0,)(0,1]\Phi:[0,\infty)\rightarrow(0,1] that is eventually log-convex and satisfies

011u(1logΦ1(u))du<,\int_{0}^{1}\frac{1}{u(1\vee\log\Phi^{-1}(u))}\ \mathrm{d}u<\infty,

there exists a nearest-neighbour random walk (Xn)n0(X_{n})_{n\geq 0} on 0\operatorname{\mathbb{Z}}_{\geq 0} with Green’s function 𝔾(n)=𝔾(n,0)\mathbb{G}(n)=\mathbb{G}(n,0), such that lim supnΦ(n)1𝔾(Xn)<\limsup_{n\to\infty}\Phi(n)^{-1}\mathbb{G}(X_{n})<\infty and (Xn)n0(X_{n})_{n\geq 0} has at most finitely many cut times almost surely.

In the proof of this proposition, we will utilize the following two elementary identities relating the quantities pnp_{n}, 𝔾(n)\mathbb{G}(n), and D(n)D(n) for each n1n\geq 1:

D(n1)=𝔾(n1)𝔾(n1)𝔾(n)D(n-1)=\frac{\mathbb{G}(n-1)}{\mathbb{G}(n-1)-\mathbb{G}(n)} (33)

and pn=12𝔾(n1)+𝔾(n+1)2𝔾(n)𝔾(n1)𝔾(n+1).p_{n}=\frac{1}{2}\frac{\mathbb{G}(n-1)+\mathbb{G}(n+1)-2\mathbb{G}(n)}{\mathbb{G}(n-1)-\mathbb{G}(n+1)}. (34)

The first identity follows from [13, (2.1)] and the elementary identity (n+1,n)=(n+1)/(n)\mathbb{H}(n+1,n)=\mathbb{H}(n+1)/\mathbb{H}(n), where (n)\mathbb{H}(n) is the probability that (Xm)(X_{m}) will hit 0 when X0=nX_{0}=n, and the second identity follows from [13, (2.2)] together with the first.

Plugging (33) into n=2(D(n)logn)1\sum_{n=2}^{\infty}(D(n)\log n)^{-1}, we observe that their summation criterion is roughly related to our integral condition by a change of variables. We prove Proposition 4.2 by formalising this relationship for a walk whose Green’s function is an appropriate transformation of the input function Φ\Phi. We then conclude by proving a very weak lower bound on the displacement of the walk from 0.

Proof of Proposition 4.2.

Let ff be the decreasing, log-convex function f(x):=elog(x+2)f(x):=e^{-\sqrt{\log(x+2)}}, and let M2M\geq 2 be the smallest integer such that the restriction of Φ\Phi to [M,)[M,\infty) is log-convex. We begin by defining the function Φ~:[0,)(0,)\widetilde{\Phi}:[0,\infty)\rightarrow(0,\infty) by

Φ~(x)=Φ((x+M)4)f(x),\widetilde{\Phi}(x)=\Phi((x+M)^{4})f(x),

and noting some its properties. First, observe that Φ~Φ\widetilde{\Phi}\leq\Phi is strictly positive, strictly decreasing, log-convex and differentiable. Moreover, since Φ~(x)Φ((x+M)4)f(x)\widetilde{\Phi}(x)\leq\Phi((x+M)^{4})\wedge f(x), we also have that Φ~1(x)(Φ1(x)1/4M)f1(x)\widetilde{\Phi}^{-1}(x)\geq(\Phi^{-1}(x)^{1/4}-M)\vee f^{-1}(x), and hence that there exists a C<C<\infty finite such that

011u(1logΦ~1(u))du\displaystyle\int_{0}^{1}\frac{1}{u(1\vee\log\widetilde{\Phi}^{-1}(u))}\ \mathrm{d}u 011u(1log[(Φ1(x)1/4M)f1(x)])du\displaystyle\leq\int_{0}^{1}\frac{1}{u(1\vee\log[(\Phi^{-1}(x)^{1/4}-M)\vee f^{-1}(x)])}\ \mathrm{d}u
=01min{1u(1logf1(x)),1u(1log[(Φ1(x)1/4M)])}du\displaystyle=\int_{0}^{1}\min\left\{\frac{1}{u(1\vee\log f^{-1}(x))},\,\frac{1}{u(1\vee\log[(\Phi^{-1}(x)^{1/4}-M)])}\right\}\ \mathrm{d}u
C+C011u(1logΦ1(x))du<,\displaystyle\leq C+C\int_{0}^{1}\frac{1}{u(1\vee\log\Phi^{-1}(x))}\mathrm{d}u<\infty, (35)

where for functions F{Φ~,f}F\in\{\widetilde{\Phi},f\}, we use the convention that 1logF1(u)=11\vee\log F^{-1}(u)=1 when F1(u)F^{-1}(u) is not defined. We also note that the logarithmic derivative (logΦ~)(\log\widetilde{\Phi})^{\prime} of Φ~\widetilde{\Phi}, which is increasing by log-convexity of Φ~\widetilde{\Phi}, satisfies the inequality

ddxlogΦ~(x)ddxlogf(x)=12(x+2)log(x+2).-\frac{\mathrm{d}}{\mathrm{d}x}\log\widetilde{\Phi}(x)\geq-\frac{\mathrm{d}}{\mathrm{d}x}\log f(x)=\frac{1}{2(x+2)\sqrt{\log(x+2)}}. (36)

We now use the function Φ~\widetilde{\Phi} to define a Markov chain satisfying the desired properties. For i1i\geq 1, we define

pi=12Φ~(n1)+Φ~(n+1)2Φ~(n)Φ~(n1)Φ~(n+1),p_{i}=\frac{1}{2}\frac{\widetilde{\Phi}(n-1)+\widetilde{\Phi}(n+1)-2\widetilde{\Phi}(n)}{\widetilde{\Phi}(n-1)-\widetilde{\Phi}(n+1)}, (37)

which is non-negative since Φ~\widetilde{\Phi} is convex and strictly less than 1/21/2 since Φ~\widetilde{\Phi} is strictly decreasing. We can therefore define a nearest-neighbour random walk (Xn)n0(X_{n})_{n\geq 0} on the integers with X0=0X_{0}=0 and with transition probabilities

(Xn+1=i+1Xn=i)=1(Xn+1=i1Xn=i)=12+pifor i1,\operatorname{\mathbb{P}}(X_{n+1}=i+1\mid X_{n}=i)=1-\operatorname{\mathbb{P}}(X_{n+1}=i-1\mid X_{n}=i)=\frac{1}{2}+p_{i}\qquad\text{for }i\geq 1,

and (Xn+1=1Xn=0)=1\operatorname{\mathbb{P}}(X_{n+1}=1\mid X_{n}=0)=1. Comparing (37) and (34), it follows by induction on nn that the Green’s function of this Markov chain is given by

𝔾(n)=CΦ~(n)for n0,\mathbb{G}(n)=C\widetilde{\Phi}(n)\qquad\text{for }n\geq 0,

for some constant C=𝔾(0)/Φ~(0)C=\mathbb{G}(0)/\widetilde{\Phi}(0) independent of nn. Therefore, to complete the proof, it suffices to show that

lim supnΦ~(Xn)Φ(n)lim supnΦ((Xn+M)4)Φ(n)< a.s.\limsup_{n\to\infty}\frac{\widetilde{\Phi}(X_{n})}{\Phi(n)}\leq\limsup_{n\to\infty}\frac{\Phi((X_{n}+M)^{4})}{\Phi(n)}<\infty\quad\text{ a.s.} (38)

and that XX has at most finitely many cut times almost surely.

We first apply Theorem 4.1 to prove that XX has finitely many cut times almost surely. Let N3N\geq 3 be large enough such that Φ(N1)<f(2)\Phi(N-1)<f(2). We can calculate

n=N1D(n)logn=n=NΦ~(n)Φ~(n+1)Φ~(n)lognn=NΦ~(n)Φ~(n)lognN1Φ~(x)Φ~(x)logxdx0Φ(N1)duulogΦ~1(u)<,\sum_{n=N}^{\infty}\frac{1}{D(n)\log n}=\sum_{n=N}^{\infty}\frac{\widetilde{\Phi}(n)-\widetilde{\Phi}(n+1)}{\widetilde{\Phi}(n)\log n}\\ \leq\sum_{n=N}^{\infty}\frac{-\widetilde{\Phi}^{\prime}(n)}{\widetilde{\Phi}(n)\log n}\leq\int_{N-1}^{\infty}\frac{-\widetilde{\Phi}^{\prime}(x)}{\widetilde{\Phi}(x)\log x}\ \mathrm{d}x\leq\int_{0}^{\Phi(N-1)}\frac{\mathrm{d}u}{u\log\widetilde{\Phi}^{-1}(u)}<\infty,

where the first equality follows from (33), the first inequality is by convexity, the second follows by integral comparison as (d/dx)[logΦ~(x)](\mathrm{d}/\mathrm{d}x)[\log\widetilde{\Phi}(x)] is increasing, the third follows by the substitution u=Φ~(x)u=\widetilde{\Phi}(x) and the inequality Φ~Φ\widetilde{\Phi}\leq\Phi, and the fourth follows from (4). The claim then follows from Theorem 4.1.

Finally, we prove (38). As Φ~\widetilde{\Phi} is decreasing, it suffices to show that

infmnXmn1/2o(1)a.s. as n  and hence that lim infnXnn1/4>1a.s.\inf_{m\geq n}X_{m}\geq n^{1/2-o(1)}\quad\text{a.s.\ as $n\to\infty$ }\qquad\text{ and hence that }\qquad\liminf_{n\rightarrow\infty}\frac{X_{n}}{n^{1/4}}>1\quad\text{a.s.} (39)

Since Φ~\widetilde{\Phi} is log-convex, Φ~(m+1)Φ~(m)Φ~(1)/Φ~(0)\widetilde{\Phi}(m+1)\geq\widetilde{\Phi}(m)\widetilde{\Phi}(1)/\widetilde{\Phi}(0) for every m1m\geq 1. For each m0m\geq 0, define Hm=|{n0:Xn=m}|H_{m}=\lvert\{n\geq 0:X_{n}=m\}\rvert. By [13, Lemma B], for each m0m\geq 0, HmH_{m} is a geometric random variable with parameter

1/μm:=1+2pm2Φ~(m)Φ~(m+1)Φ~(m)12Φ~(m)Φ~(m+1)Φ~(m)Φ~(m+1)2Φ~(m)Φ~(1)Φ~(m+1)2Φ~(0)Φ~(m+1)Φ~(1)4Φ~(0)(m+2)log(m+2),1/\mu_{m}:=\frac{1+2p_{m}}{2}\cdot\frac{\widetilde{\Phi}(m)-\widetilde{\Phi}(m+1)}{\widetilde{\Phi}(m)}\geq\frac{1}{2}\frac{\widetilde{\Phi}(m)-\widetilde{\Phi}(m+1)}{\widetilde{\Phi}(m)}\\ \geq\frac{-\widetilde{\Phi}^{\prime}(m+1)}{2\widetilde{\Phi}(m)}\geq-\frac{\widetilde{\Phi}(1)\widetilde{\Phi}^{\prime}(m+1)}{2\widetilde{\Phi}(0)\widetilde{\Phi}(m+1)}\geq\frac{\widetilde{\Phi}(1)}{4\widetilde{\Phi}(0)(m+2)\sqrt{\log(m+2)}},

where the second inequality follows by convexity of Φ~\widetilde{\Phi} and the final inequality follows from (36). Since each HmH_{m} is a geometric random variable with mean of order m1+o(1)m^{1+o(1)}, it follows by an elementary application of Borel-Cantelli that maxmnHm=n1+o(1)\max_{m\leq n}H_{m}=n^{1+o(1)} almost surely as nn\to\infty, and hence that maxmnXmn1/2o(1)\max_{m\leq n}X_{m}\geq n^{1/2-o(1)} almost surely as nn\to\infty. On the other hand, letting τn\tau_{n} be the hitting time of nn for each n1n\geq 1, we have by optional stopping that

(X visits m after τn)𝔾(n)𝔾(m)=Φ~(n)Φ~(m)f(n)f(m)\mathbb{P}(\text{$X$ visits $m$ after $\tau_{n}$})\leq\frac{\mathbb{G}(n)}{\mathbb{G}(m)}=\frac{\widetilde{\Phi}(n)}{\widetilde{\Phi}(m)}\leq\frac{f(n)}{f(m)}

for each 0mn0\leq m\leq n. Since f(2k)/f(2(1ε)k)f(2^{k})/f(\lfloor 2^{(1-\varepsilon)k}\rfloor) is superpolynomially small in kk for each fixed ε>0\varepsilon>0, we deduce by a further simple Borel-Cantelli argument that infmnXm=(maxmnXm)1o(1)n1/2o(1)\inf_{m\geq n}X_{m}=(\max_{m\leq n}X_{m})^{1-o(1)}\geq n^{1/2-o(1)} almost surely as nn\to\infty, completing the proof. ∎

Acknowledgements

TH thanks his former advisor Asaf Nachmias for introducing him to the problem in 2013. NH has been supported by the doctoral training centre, Cambridge Mathematics of Information (CMI).

References

  • [1] O. Angel, N. Crawford, and G. Kozma. Localization for linearly edge reinforced random walks. Duke Math. J., 163(5):889–921, 2014.
  • [2] I. Benjamini, O. Gurel-Gurevich, and R. Lyons. Recurrence of random walk traces. Ann. Probab., 35(2):732–738, 2007.
  • [3] I. Benjamini, O. Gurel-Gurevich, and O. Schramm. Cutpoints and resistance of random walk paths. Ann. Probab., 39(3):1122–1136, 2011.
  • [4] I. Benjamini and J. Hermon. Recurrence of Markov chain traces. Ann. Inst. Henri Poincaré Probab. Stat., 56(1):734–759, 2020.
  • [5] I. Benjamini and Y. Peres. Tree-indexed random walks on groups and first passage percolation. Probab. Theory Related Fields, 98(1):91–112, 1994.
  • [6] S. Blachère. Cut times for random walks on the discrete Heisenberg group. Ann. Inst. H. Poincaré Probab. Statist., 39(4):621–638, 2003.
  • [7] S. Blachère, P. Haïssinsky, and P. Mathieu. Asymptotic entropy and Green speed for random walks on countable groups. Ann. Probab., 36(3):1134–1152, 2008.
  • [8] F. T. Bruss. A counterpart of the Borel-Cantelli lemma. J. Appl. Probab., 17(4):1094–1101, 1980.
  • [9] K. Burdzy and G. F. Lawler. Nonintersection exponents for Brownian paths. I. Existence and an invariance principle. Probab. Theory Related Fields, 84(3):393–410, 1990.
  • [10] K. Burdzy and G. F. Lawler. Nonintersection exponents for Brownian paths. II. Estimates and applications to a random fractal. Ann. Probab., 18(3):981–1009, 1990.
  • [11] T. K. Carne. A transmutation formula for Markov chains. Bull. Sci. Math. (2), 109(4):399–405, 1985.
  • [12] M. C. Cranston and T. S. Mountford. An extension of a result of Burdzy and Lawler. Probab. Theory Related Fields, 89(4):487–502, 1991.
  • [13] E. Csáki, A. Földes, and P. Révész. On the number of cutpoints of the transient nearest neighbor random walk on the line. J. Theoret. Probab., 23(2):624–638, 2010.
  • [14] P. Diaconis and D. Freedman. de Finetti’s theorem for Markov chains. Ann. Probab., 8(1):115–130, 1980.
  • [15] B. Duplantier. Intersections of random walks. A direct renormalization approach. Comm. Math. Phys., 117(2):279–329, 1988.
  • [16] P. Erdős and S. J. Taylor. Some intersection properties of random walk paths. Acta Math. Acad. Sci. Hungar., 11:231–248, 1960.
  • [17] M. Folz. Volume growth and stochastic completeness of graphs. Trans. Amer. Math. Soc., 366(4):2089–2119, 2014.
  • [18] N. James, R. Lyons, and Y. Peres. A transient Markov chain with finitely many cutpoints. In Probability and statistics: essays in honor of David A. Freedman, volume 2 of Inst. Math. Stat. (IMS) Collect., pages 24–29. Inst. Math. Statist., Beachwood, OH, 2008.
  • [19] N. James and Y. Peres. Cutpoints and exchangeable events for random walks. Teor. Veroyatnost. i Primenen., 41(4):854–868, 1996.
  • [20] V. A. Kaĭmanovich and A. M. Vershik. Random walks on discrete groups: boundary and entropy. Ann. Probab., 11(3):457–490, 1983.
  • [21] G. F. Lawler. Escape probabilities for slowly recurrent sets. Probab. Theory Related Fields, 94(1):91–117, 1992.
  • [22] G. F. Lawler. Cut times for simple random walk. Electron. J. Probab., 1:no. 13, approx. 24 pp., 1996.
  • [23] G. F. Lawler. Hausdorff dimension of cut points for Brownian motion. Electron. J. Probab., 1:no. 2, approx. 20 pp., 1996.
  • [24] G. F. Lawler. Intersections of random walks. Modern Birkhäuser Classics. Birkhäuser/Springer, New York, 2013. Reprint of the 1996 edition.
  • [25] A. Lejay. Simulating a diffusion on a graph. Application to reservoir engineering. Monte Carlo Methods Appl., 9(3):241–255, 2003.
  • [26] P. Lévy. Théorie de l’addition des variables aléatoires. Monographies des probabilités ; calcul des probabilités et ses applications. Gauthier-Villars, Paris, 2. ed. edition, 1954.
  • [27] C. H. Lo, M. V. Menshikov, and A. R. Wade. Cutpoints of non-homogeneous random walks. ALEA - Latin American Journal of Probability and Mathematical Statistics, 2021.
  • [28] T. Lupu. From loop clusters and random interlacements to the free field. Ann. Probab., 44(3):2117–2146, 2016.
  • [29] R. Lyons and Y. Peres. Probability on Trees and Networks, volume 42 of Cambridge Series in Statistical and Probabilistic Mathematics. Cambridge University Press, New York, 2016. Available at https://rdlyons.pages.iu.edu/.
  • [30] R. Lyons and Y. Peres. Poisson boundaries of lamplighter groups: proof of the Kaimanovich-Vershik conjecture. J. Eur. Math. Soc. (JEMS), 23(4):1133–1160, 2021.
  • [31] F. Merkl, A. Öry, and S. W. W. Rolles. The ‘magic formula’ for linearly edge-reinforced random walks. Statist. Neerlandica, 62(3):345–363, 2008.
  • [32] N. T. Varopoulos. Long range estimates for Markov chains. Bull. Sci. Math. (2), 109(3):225–252, 1985.