This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Is ‘being above the median’ a noise sensitive property?

Daniel Ahlberg111Department of Mathematics, Stockholm University.  and Daniel de la Riva222Department of Mathematics, University of British Columbia.
Abstract

Assign independent weights to the edges of the square lattice, from the uniform distribution on {a,b}\{a,b\} for some 0<a<b<0<a<b<\infty. The weighted graph induces a random metric on 2{\mathbb{Z}}^{2}. Let TnT_{n} denote the distance between (0,0)(0,0) and (n,0)(n,0) in this metric. The distribution of TnT_{n} has a well-defined median. Itai Benjamini asked in 2011 if the sequence of Boolean functions encoding whether TnT_{n} exceeds its median is noise sensitive? In this paper we present the first progress on Benjamini’s problem. More precisely, we study the minimal weight along any path crossing an n×nn\times n-square horizontally and whose vertical fluctuation is smaller than n1/22n^{1/22}, and show that for this observable, ‘being above the median’ is a noise sensitive property.

Keywords: First-passage percolation; noise sensitivity; moderate deviations.

Funding: Research in part supported by the Swedish Research Council through grant 2021-03964.

1 Introduction

Central in statistical physics is the notion of a phase transition, i.e. a sudden change of behaviour as some parameter of the model is changed. As a consequence, configurations that correlate well on a microscopic scale may look radically different on a macroscopic scale, if they correspond to different sides of the transition. However, it is also possible for highly correlated configurations to behave differently, despite having the same law. A formal framework, in the context of Boolean functions, in which questions like this could be studied was introduced in a seminal paper by Benjamini, Kalai and Schramm [10]. Let ω{0,1}n\omega\in\{0,1\}^{n} be chosen uniformly at random, and obtain ωε\omega^{\varepsilon} from ω\omega by independently resampling the coordinates with probability ε(0,1)\varepsilon\in(0,1). A sequence (fn)n1(f_{n})_{n\geq 1} of Boolean functions fn:{0,1}n{0,1}f_{n}:\{0,1\}^{n}\to\{0,1\} is said to be noise sensitive if for every ε>0\varepsilon>0

𝔼[fn(ω)fn(ωε)]𝔼[fn(ω)]20as n.{\mathbb{E}}\big{[}f_{n}(\omega)f_{n}(\omega^{\varepsilon})\big{]}-{\mathbb{E}}[f_{n}(\omega)]^{2}\to 0\quad\text{as }n\to\infty. (1)

In [10], the authors gave the first example of noise sensitivity, in particular establishing noise sensitivity in planar Bernoulli percolation at criticality. In order to do this, they developed methods by which noise sensitivity could be established, that remain relevant to this day. Later works have established analogous results for percolation in the continuum, based on Poisson [2, 5, 35] and Gaussian processes [30], as well as in the context of random graphs [36]. For these models, central observables have a Boolean outcome. For many other models in the realm of random spatial processes, the main observables are not Boolean, but real-valued functions on the space of configurations. This is the case for a variety of disordered systems, polymer models, and spatial growth models such as first- and last-passage percolation.

In September 2011, at the doctoral defense of the first author, Itai Benjamini proposed a natural approach to explore the sensitivity to small perturbations of real-valued observables. The approach can be synthesised briefly with the words: Is ‘being above the median’ a noise sensitive property? The purpose of this paper is to present the first progress on Benjamini’s problem.

1.1 Model description and main result

In the most well-studied setting, first-passage percolation is the study of the random metric space that arise by assigning non-negative independent random weights, from some common distribution FF, to the edges of the 2{\mathbb{Z}}^{2} nearest-neighbour lattice. For simplicity, we shall in this paper stick to the planar setting, and we shall for most of the paper assume that FF is supported on {a,b}\{a,b\} for some 0<a<b<0<a<b<\infty. The edge weights induce a metric TT on 2{\mathbb{Z}}^{2} as follows: For u,v2u,v\in{\mathbb{Z}}^{2}, set

T(u,v):=inf{T(Γ):Γ is a path from u to v},whereT(Γ):=eΓωe.T(u,v):=\inf\big{\{}T(\Gamma):\Gamma\text{ is a path from $u$ to $v$}\big{\}},\quad\text{where}\quad T(\Gamma):=\sum_{e\in\Gamma}\omega_{e}.

The infimum in T(u,v)T(u,v) is known to be attained for some finite path, although this path does not have to be unique. We let π(u,v)\pi(u,v) denote this path, and apply some deterministic rule for selecting one in case it is not unique.

For n1n\geq 1, set Tn:=T(0,n𝐞1)T_{n}:=T(0,n{\bf e}_{1}) and πn:=π(0,n𝐞1)\pi_{n}:=\pi(0,n{\bf e}_{1}) for brevity, where 𝐞1:=(1,0){\bf e}_{1}:=(1,0) denote the first coordinate vector. A standard consequence of the Subadditive Ergodic Theorem is the existence of a constant μ\mu, known as the time constant, such that almost surely

Tnnμas n.\frac{T_{n}}{n}\to\mu\quad\text{as }n\to\infty. (2)

Also πn\pi_{n} is known to be of linear order, although it is not known to have a well-defined asymptotic speed. In approaching the problem of the current paper, one soon requests a finer description of the order of fluctuations, both for TnT_{n} around its mean, and πn\pi_{n} away from the coordinate axis. Predictions from the physics literature [33], which have been established for related models of last-passage percolation [8, 32], suggest that fluctuations of TnT_{n} are order n1/3n^{1/3} and transversal fluctuations of πn\pi_{n} are order n2/3n^{2/3}.

The approach proposed by Itai Benjamini, in September 2011, to explore questions of noise sensitivity in the context of first-passage percolation (and for other real-valued observables) can be described as follows: The distance TnT_{n} is a random variable, whose distribution may be unknown. This distribution has a median mnm_{n}, and for large nn, this median can be expected to split the distribution of TnT_{n} roughly in half, i.e. that

(Tn<mn)(Tn>mn)12.{\mathbb{P}}(T_{n}<m_{n})\approx{\mathbb{P}}(T_{n}>m_{n})\approx\frac{1}{2}. (3)

Under the assumption that the weight distribution FF is supported on {a,b}\{a,b\}, for some 0<a<b<0<a<b<\infty, the event that TnT_{n} exceeds its median can be encoded as a Boolean function. If (3) holds, then this function is non-degenerate. It is thus possible, as proposed by Benjamini, to investigate whether TnT_{n} exceeding its median is a noise sensitive property, within the framework of Boolean functions.

In this paper we shall present the first progress on Benjamini’s problem. We have not been able to answer the question as it has been formulated above, for reasons that we shall elaborate upon below. As stated the question thus remains open. Indeed, although (3) trivially holds for continuous weight distributions, it seems to remain unknown whether, uniformly in nn,

(Tn<mn)>0and(Tn>mn)>0,{\mathbb{P}}(T_{n}<m_{n})>0\quad\text{and}\quad{\mathbb{P}}(T_{n}>m_{n})>0,

for some median of TnT_{n}, when FF is discrete. See, however, [15, 21] for results in this direction.

In order to circumvent the difficulties faced above, we shall make two simplifications to Benjamini’s problem. First, we replace point-to-point passage times by horizontal crossing times of squares, and hence increase the symmetry in the problem. Second, we restrict the transversal fluctuations allowed by paths crossing the squares. Given k1k\geq 1, let 𝒫k(n)\mathcal{P}_{k}(n) denote the set of nearest-neighbour paths contained the ‘square’ [0,n]×[0,n1][0,n]\times[0,n-1] that connect the left side to the right, and whose vertical displacement is at most kk, and set

τ(n,k):=inf{T(Γ):Γ𝒫k(n)}.\tau(n,k):=\inf\big{\{}T(\Gamma):\Gamma\in\mathcal{P}_{k}(n)\big{\}}. (4)

Our main result is the following theorem, which makes the first progress on Benjamini’s problem. Our result is formulated for an arbitrary quantile of the crossing variable, and not just its median.

Theorem 1.1.

Suppose that FF is supported on {a,b}\{a,b\} for some 0<a<b<0<a<b<\infty. Let α<1/22\alpha<1/22 and β(0,1)\beta\in(0,1) be fixed. For any sequence (kn)n1(k_{n})_{n\geq 1} such that knnαk_{n}\leq n^{\alpha}, and for any β\beta-quantile qβq_{\beta} of τ(n,kn)\tau(n,k_{n}), we have

limn(τ(n,kn)<qβ)=1limn(τ(n,kn)>qβ)=β,\lim_{n\to\infty}{\mathbb{P}}\big{(}\tau(n,k_{n})<q_{\beta}\big{)}=1-\lim_{n\to\infty}{\mathbb{P}}\big{(}\tau(n,k_{n})>q_{\beta}\big{)}=\beta,

and the sequence (fn)n1(f_{n})_{n\geq 1} of functions fn:=𝟙{τ(n,kn)>qβ}f_{n}:=\mathbbm{1}_{\{\tau(n,k_{n})>q_{\beta}\}} is noise sensitive.

We remark that the analogous result holds for the square replaced by an n×nn\times n-torus, and τ(n,kn)\tau(n,k_{n}) replaced by the minimal weight among all circuits crossing the torus horizontally, and whose transversal fluctuations are bounded by knk_{n}. Moreover, while we here focus on the planar setting, an analogous statement holds also in dimensions d3d\geq 3, with a stronger restriction on the growth of knk_{n}. In both cases, adapting the proof of Theorem 1.1 is straightforward.

Related to the study of noise sensitivity is the notion of ‘chaos’, that stems from the physics literature on spin-glasses [13, 26]. In the context of first-passage percolation, chaos refers to the sensitivity of the distance-minimising path πn\pi_{n} as opposed to the distance TnT_{n}. The first rigorous evidence of chaos was obtained by Chatterjee in two preprints [16, 17], later combined into a book [18]. That the first-passage metric is chaotic was established only recently, in work of Ahlberg, Deijfen and Sfragara [3]. To state this result, let πnε\pi_{n}^{\varepsilon} denote the distance-minimising path between the origin and n𝐞1n{\bf e}_{1} with respect to the perturbed weights ωε\omega^{\varepsilon}. If, for instance, FF is continuous and has finite moment of order 2+log2+\log, then

𝔼[|πnπnε|]=o(n).{\mathbb{E}}\big{[}|\pi_{n}\cap\pi_{n}^{\varepsilon}|\big{]}=o(n).

Significantly more precise results, determining the rate at which ε=ε(n)\varepsilon=\varepsilon(n) is allowed to decay with nn, have been obtained by Ganguly and Hammond [27] for certain integrable models of last-passage percolation; see also [4] for related results.

Our result differ from the above in that it addresses the sensitivity of the metric TT as opposed to the structure minimising TT, and (to our knowledge) this result is the first of its kind for a (supercritical) spatial growth model. Note, however, related work of Damron, Hanson, Harper and Lam [20] that establish the existence of exceptional times in a dynamical version of critical first-passage percolation.

1.2 A tale of influences

A key result from the original paper of Benjamini, Kalai and Schramm [10] gives a criterion for a sequence of Boolean functions to be noise sensitive in terms of the notion of influences. The influence of bits is central in computer science, and has its origin in social choice theory. The influence of a bit i{1,2,,n}i\in\{1,2,\ldots,n\} of a function f:{0,1}n{0,1}f:\{0,1\}^{n}\to\{0,1\} is defined as the probability that the bit is desisive for the outcome of the function, i.e.

Infi(f):=(f(ω)f(σiω)),\textup{Inf}_{i}(f):={\mathbb{P}}\big{(}f(\omega)\neq f(\sigma_{i}\omega)\big{)}, (5)

where σi:{0,1}n{0,1}n\sigma_{i}:\{0,1\}^{n}\to\{0,1\}^{n} is the operator that flips the entry at position ii. The criterion, which has come to be known as the BKS Theorem, states that if

i=1nInfi(fn)20as n,\sum_{i=1}^{n}\textup{Inf}_{i}(f_{n})^{2}\to 0\quad\text{as }n\to\infty, (6)

then the sequence (fn)n1(f_{n})_{n\geq 1} is noise sensitive.

Apart from the computation of influences, as the BKS Theorem invites to, there are other methods by which noise sensitivity may be established. The main development has occurred with applications to Bernoulli percolation in mind: A method involving the revealment of randomised algorithms was developed by Schramm and Steif [38]; The Fourier spectrum of critical percolation was analysed by Garban, Pete and Schramm [28]; A probabilistic approach was taken by Tassion and Vanneuville [39], inspired by Kesten’s scaling relations. Neither of these routes seem easy to follow in our context. Moreover, for monotone functions (which we are concerned with here) the criterion in (6) is both necessary and sufficient for a sequence to be noise sensitive. So, either directly or indirectly, verifying (6) is inevitable. This will, hence, be the route we take.

Let us start with a general observation. For functions f:{0,1}nf:\{0,1\}^{n}\to{\mathbb{R}} that are Lipschitz, i.e. have bounded differences |f(ω)f(σiω)|c|f(\omega)-f(\sigma_{i}\omega)|\leq c for some constant c>0c>0 and all ii, if changing the value of a bit ii takes ff from below to above its median mm (or vice versa), then ff must have been within distance cc from the median. In particular, we have the distributional bound

Infi(𝟙{f>m})(|fm|c).\textup{Inf}_{i}(\mathbbm{1}_{\{f>m\}})\leq{\mathbb{P}}\big{(}|f-m|\leq c\big{)}. (7)

Standard variance bounds for functions that are Lipschitz (a.k.a. having bounded differences) with constant cc give Var(f)c2n\textup{Var}(f)\leq c^{2}n; see [12, Corollary 3.2]. Hence the above distributional bound gives a bound of order 1/n1/\sqrt{n} at best. This would amount to an upper bound on the sum of influences squared being a non-vanishing constant. Hence, it cannot in general be sufficient to bound the influences simply considering the distribution of ff.

Note that, regardless if we consider TnT_{n} or τ(n,k)\tau(n,k), changing the value of an edge may affect the observable by at most ±(ba)\pm(b-a), meaning that they are both Lipschitz. Hence, a simple distributional bound as in (7) will not suffice. Using an observation from [11], we may link influential edges to edges on the geodesic. Recall that πn\pi_{n} is the path (a path in case of multiple) attaining the infimum in TnT_{n}. Then,

Infe(𝟙{Tn>mn})=2(ωe=a,e pivotal)2(eπn,|Tnmn|ba).\textup{Inf}_{e}(\mathbbm{1}_{\{T_{n}>m_{n}\}})=2\,{\mathbb{P}}\big{(}\omega_{e}=a,e\text{ pivotal}\big{)}\leq 2\,{\mathbb{P}}\big{(}e\in\pi_{n},|T_{n}-m_{n}|\leq b-a\big{)}. (8)

The predictions from KPZ universality suggest that |Tnmn|ba|T_{n}-m_{n}|\leq b-a should occur with probability order n1/3n^{-1/3}, and that a typical edge being on the geodesic has probability order n2/3n^{-2/3}. For most edges within distance n2/3n^{2/3} or the coordinate axis the influence is thus order 1/n1/n, and for edges further away it is negligible. This amounts to a bound on the sum of influences squared that vanishes with nn.

The above heuristic is merely conjectural, and we are nowhere close to establish statements like this in first-passage percolation. It is generally not even known whether

(|Tnmn|ba)0as n{\mathbb{P}}\big{(}|T_{n}-m_{n}|\leq b-a\big{)}\to 0\quad\text{as }n\to\infty

for some median mnm_{n} of TnT_{n}. However, a result by Pemantle and Peres [37] implies such a statement for exponentially distributed edge weights.

In a first attempt to simplify Benjamini’s problem it is tempting to replace the point-to-point passage times with the left-right crossing times of rectangles. Let 𝒫(n,m)\mathcal{P}(n,m) denote the set of nearest-neighbour paths contained in the rectangle R(n,m):=[0,n]×[0,m1]R(n,m):=[0,n]\times[0,m-1] that connect the left side to the right. We define the crossing time of the rectangle R(n,m)R(n,m) as

t(n,m):=inf{T(Γ):Γ𝒫(n,m)}.t(n,m):=\inf\big{\{}T(\Gamma):\Gamma\in\mathcal{P}(n,m)\big{\}}. (9)

In particular, we let tn:=t(n,n)t_{n}:=t(n,n) denote the crossing time of the ‘square’ R(n,n)R(n,n). The increasing level of symmetry attained in this way is manifested in that all horizontal and all vertical bonds of the square an effect of tnt_{n} that is roughly equal.333Admittedly, the square may have to be replaced by a torus, and the crossing time of the square by the circumference of the torus, for this to be fully rigorous. From the linear upper bound on the length of a geodesic due to Kesten [34], a bound analogous to (8) would give

Infe(𝟙{tn>mn})C1n(|tnmn|ba),\textup{Inf}_{e}(\mathbbm{1}_{\{t_{n}>m_{n}\}})\leq C\,\frac{1}{n}\,{\mathbb{P}}\big{(}|t_{n}-m_{n}|\leq b-a\big{)},

and hence

eInfe(𝟙{tn>mn})2C(|tnmn|ba).\sum_{e}\textup{Inf}_{e}(\mathbbm{1}_{\{t_{n}>m_{n}\}})^{2}\leq C\,{\mathbb{P}}\big{(}|t_{n}-m_{n}|\leq b-a\big{)}. (10)

This shows that even if we spread out the contribution coming from ‘being on the geodesic’, only considering the contribution from the geodesic, and not the distributional properties of the crossing time, will not suffice in order to deduce noise sensitivity.

Again, it remains unknown whether the probability in the right-hand side of (10) vanishes as nn\to\infty. In fact, also the weaker question whether the variance of tnt_{n} diverges as nn\to\infty remains unknown; the best lower bound gives a constant. We refer the reader to the recent work of Damron, Houdré and Özdemir [22] for a further discussion in this direction.

1.3 Distributional control over restricted paths

We shall circumvent the above mentioned difficulties in calculating the influences by imposing a restriction on the transversal fluctuations of the paths admissible for crossing the square R(n,n)R(n,n).

The restriction on transversal fluctuations does not result in a lower asymptotic velocity by which the square is crossed, as long as the allowed fluctuations diverge with nn; see [1, 14]. That is, for any diverging sequence (kn)n1(k_{n})_{n\geq 1} we have, almost surely,

μ=limnTnn=limntnn=limnτ(n,kn)n.\mu=\lim_{n\to\infty}\frac{T_{n}}{n}=\lim_{n\to\infty}\frac{t_{n}}{n}=\lim_{n\to\infty}\frac{\tau(n,k_{n})}{n}.

However, it is expected that the restriction does have an effect on a lower order. For kk fixed, on the other hand, one may show that there exists μk>μ\mu_{k}>\mu such that almost surely

limnτ(n,k)n=μk.\lim_{n\to\infty}\frac{\tau(n,k)}{n}=\mu_{k}.

To see how the transversal restriction will help in calculating the influences, let us consider the case when k=1k=1. Note that τ(n,1)\tau(n,1) is the sum of nn independent binomial random variables with parameters nn and 1/21/2. Using moderate deviation estimates of the binomial distribution we are able to compute the asymptotic behaviour of the quantiles of their minimum as well as the influences. Note how the case k=1k=1 is reminiscent of the classical Tribes function, introduced by Ben-Or and Linial [9], but with polynomial-sized tribes as opposed to logarithmic. We treat the case k=1k=1 in detail in Section 2, as special case of a larger family of polynomial Tribes functions.

For k2k\geq 2 we may express τ(n,k)\tau(n,k) as the minimum of nk+1n-k+1 identically distributed, but dependent, variables as follows. Let Ri(n,k)R_{i}(n,k) denote the rectangle [0,n]×[i,i+k1][0,n]\times[i,i+k-1], and ti(n,k)t_{i}(n,k) the horizontal crossing time of Ri(n,k)R_{i}(n,k). Since every path in 𝒫k(n)\mathcal{P}_{k}(n) may fluctuate vertically at most kk, it has to be contained in Ri(n,k)R_{i}(n,k) for some i=0,1,,nki=0,1,\ldots,n-k. It follows that

τ(n,k)=mini=0,1,,nkti(n,k).\tau(n,k)=\min_{i=0,1,\ldots,n-k}t_{i}(n,k).

For fixed kk, the distribution of these variables is asymptotically Gaussian, as proved (in parallel) by Ahlberg [1] and Chatterjee and Dey [14]. In fact, the latter paper shows that the asymptotic normality continues to hold for k=k(n)k=k(n) growing slower than n1/3n^{1/3}. The asymptotic normality will not be sufficient in itself, as we will need to peak into the tail of the distribution, in that τ(n,k)\tau(n,k) is a minimum of a large number of variables. For that reason, we shall need to combine the approach from [14] with a Cramér-type moderate deviations theorem for triangular arrays (Theorem 3.1), in order to obtain a moderate deviations theorem for first-passage percolation across thin rectangles (Theorem 4.1). With the moderate deviations estimates at hand, we will be able to approximate the asymptotic behaviour of quantiles and influences for the restricted crossing time τ(n,k)\tau(n,k), and prove Theorem 1.1.

We remark that the asymptotic normality is in itself not central to our approach. The relevant part is that it allows us to bound the influence of an edge by a rare enough event, whose probability we may compute. As a by-product of our proof we obtain the following estimate on the fluctuations on τ(n,k)\tau(n,k).

Theorem 1.2.

Suppose that FF is supported on {a,b}\{a,b\} for some 0<a<b<0<a<b<\infty. For any α<1/22\alpha<1/22 and any sequence (kn)n1(k_{n})_{n\geq 1} such that knnαk_{n}\leq n^{\alpha} we have

supx0(τ(n,kn)[x,x+c])=o(1n1/22).\sup_{x\geq 0}{\mathbb{P}}\big{(}\tau(n,k_{n})\in[x,x+c]\big{)}=o\Big{(}\frac{1}{n^{1/22}}\Big{)}.

While we here focus on weight distributions supported on two points, we remark that our proof of the above theorem goes through without change for bounded weight distributions. Apart from a result by Pemantle and Peres [37] for exponentially distributed edge weights, it remains unknown whether for every c>0c>0 we have

supx0(Tn[x,x+c])0as n.\sup_{x\geq 0}{\mathbb{P}}\big{(}T_{n}\in[x,x+c]\big{)}\to 0\quad\text{as }n\to\infty. (11)

It would be interesting to establish (11) for a large class of weight distributions.

The analogous problem for geodesics is the well-known ‘midpoint problem’, which was posed by Benjamini, Kalai and Schramm in [11]. Interestingly, this problem has been solved for continuous weight with finite mean by Ahlberg and Hoffman [6]. Their result shows that for every edge ee we have

(eπ(n𝐞1,n𝐞1))0as n.{\mathbb{P}}\big{(}e\in\pi(-n{\bf e}_{1},n{\bf e}_{1})\big{)}\to 0\quad\text{as }n\to\infty. (12)

Earlier work of Damron and Hanson [19] gave a conditional proof under plausible, but unverified, assumptions on the asymptotic shape. In more recent work, Dembin, Elboim and Peled [23] derive polynomial rates on the decay in (12) for a more restrictive class of weight distributions. However, as mentioned above, without progress on the problem in (11), these results are insufficient for making further progress on Benjamini’s problem.

1.4 Organisation of the paper

The rest of this paper is organised as follows. In Section 2 we showcase out approach by considering a polynomial version of the classical Tribes function. In Section 3 we prove a Cramér-type moderate deviations theorem for triangular arrays. In Section 4 we apply the moderate deviations theorem to prove a moderate deviations theorem for first-passage percolation across thin rectangles, which will allow us to analyse the asymptotic behaviour of the quantiles of our main observable. In Section 5 we derive a preliminary version of our main theorem, in which we consider the minimum crossing time across disjoint rectangles. Finally, in Section 6, we prove our main results, and in Section 7 we elaborate upon some open problems.

2 Polynomial Tribes

In this section we illustrate our method in a simplified context. We shall prove that ‘being above the median’ is a noise sensitive property for a class of functions that generalises the classical function known as Tribes; see e.g. [29]. For every λ(0,1)\lambda\in(0,1) we partition [n]:={1,2,,n}[n]:=\{1,2,\ldots,n\} into blocks of length λ:=nλ\ell_{\lambda}:=\lfloor n^{\lambda}\rfloor, and perhaps some leftover debris. We refer to each block as a tribe. Given ω{0,1}[n]\omega\in\{0,1\}^{[n]}, we define SjS_{j} as the sum of the coordinates of the jjth tribe, for each of the mλ:=n/λm_{\lambda}:=\lfloor n/\ell_{\lambda}\rfloor tribes. Finally, let

Sλ:=max1jmλSjS^{\lambda}:=\max_{1\leq j\leq m_{\lambda}}S_{j} (13)

denote the maximal number of 1s in any tribe. Note that we have suppressed the dependence on nn in the above notation. Note, moreover, that the choice λ=1/2\lambda=1/2 coincides with the weight of a left-right crossing of a square when k=1k=1.

For each β(0,1)\beta\in(0,1), let qλ,βq_{\lambda,\beta} denote any β\beta-quantile of (the distribution of) SλS^{\lambda}. Define fλ,βf_{\lambda,\beta} to be the indicator function of the event {Sλ>qλ,β}\{S^{\lambda}>q_{\lambda,\beta}\}, i.e. that at least one tribe contains more than qλ,βq_{\lambda,\beta} 1s.

Naturally, our idea will be to study the behavior of (fn,βλ=0){\mathbb{P}}(f_{n,\beta}^{\lambda}=0), and if the sequence of functions {fn,βλ}\{f_{n,\beta}^{\lambda}\} is Noise Sensitive. More precisely, we get the following result

Proposition 2.1.

For every λ,β(0,1)\lambda,\beta\in(0,1) we have, as nn\to\infty, that fλ,βf_{\lambda,\beta} is noise sensitive and

(fλ,β=0)β.{\mathbb{P}}(f_{\lambda,\beta}=0)\to\beta.

Since the number of 1s in a tribe follows a binomial distribution, and since our function asks for the maximal number of 1s in any tribe, we shall in the proof of the proposition make use of known estimates on the tail of the centred binomial. Let XnX_{n} denote a binomially distributed random variable with parameters nn and 1/21/2. The following estimates date back to the work of Bahadur [7]; see also [31]: For any sequence xnx_{n} satisfying 1xnn1/61\ll x_{n}\ll n^{1/6} we have, as nn\to\infty, that

(Xnn/2+xnn/2)\displaystyle{\mathbb{P}}\big{(}X_{n}\geq n/2+x_{n}\sqrt{n}/2\big{)} =(1+o(1))1xn2πexp(xn2/2),\displaystyle=(1+o(1))\frac{1}{x_{n}\sqrt{2\pi}}\exp\big{(}-x_{n}^{2}/2\big{)}, (14)
(Xn=n/2+xnn/2)\displaystyle{\mathbb{P}}\big{(}X_{n}=\lfloor n/2+x_{n}\sqrt{n}/2\rfloor\big{)} =(1+o(1))2πnexp(xn2/2).\displaystyle=(1+o(1))\frac{\sqrt{2}}{\sqrt{\pi n}}\exp\big{(}-x_{n}^{2}/2\big{)}. (15)

We begin with a couple of lemmas determining the correct order of the β\beta-quantiles of the generalised tribes function. For λ,β(0,1)\lambda,\beta\in(0,1) and n1n\geq 1, let sλ,β=sλ,β(n)s_{\lambda,\beta}=s_{\lambda,\beta}(n) be defined as

sλ,β:=λ2+λ22(1λ)lognloglogn2log(4π(1λ)logβ1).s_{\lambda,\beta}:=\frac{\ell_{\lambda}}{2}+\frac{\sqrt{\ell_{\lambda}}}{2}\sqrt{2(1-\lambda)\log n-\log\log n-2\log\big{(}\sqrt{4\pi\big{(}1-\lambda\big{)}}\log\beta^{-1}\big{)}}.
Lemma 2.2.

For every λ,β(0,1)\lambda,\beta\in(0,1) we have (Sλsλ,β)β{\mathbb{P}}(S^{\lambda}\leq s_{\lambda,\beta})\to\beta as nn\to\infty.

Proof.

For {Sλsλ,β}\{S^{\lambda}\leq s_{\lambda,\beta}\} to occur, we need that all tribes to contain at most sλ,βs_{\lambda,\beta} 1s. Since tribes are disjoint it follows by independence that

(Sλsλ,β)=(1(Xn>sλ,β))mλ.{\mathbb{P}}\big{(}S^{\lambda}\leq s_{\lambda,\beta}\big{)}=\Big{(}1-{\mathbb{P}}\big{(}X_{\ell_{n}}>s_{\lambda,\beta}\big{)}\Big{)}^{m_{\lambda}}.

By (14), and since by (15) the probability of attaining the value sλ,β\lfloor s_{\lambda,\beta}\rfloor is of a lower order, we have that

(Xn>sλ,β)=(1+o(1))log1/βn1λ.{\mathbb{P}}\big{(}X_{\ell_{n}}>s_{\lambda,\beta}\big{)}=(1+o(1))\frac{\log 1/\beta}{n^{1-\lambda}}.

Since mλ=(1+o(1))n1λm_{\lambda}=(1+o(1))n^{1-\lambda}, we thus obtain, as nn\to\infty, that

(Sλsλ,β)=(1(1+o(1))log1/βn1λ)mλexp(log1β)=β,{\mathbb{P}}\big{(}S^{\lambda}\leq s_{\lambda,\beta}\big{)}=\Big{(}1-(1+o(1))\frac{\log 1/\beta}{n^{1-\lambda}}\Big{)}^{m_{\lambda}}\to\exp\Big{(}-\log\frac{1}{\beta}\Big{)}=\beta,

as required. ∎

Lemma 2.3.

For every λ,β(0,1)\lambda,\beta\in(0,1) and ε>0\varepsilon>0 small enough we have that any β\beta-quantile qλ,βq_{\lambda,\beta} of SλS^{\lambda} satisfies, for all sufficiently large nn, that

sλ,βε<qλ,β<sλ,β+ε.s_{\lambda,\beta-\varepsilon}<q_{\lambda,\beta}<s_{\lambda,\beta+\varepsilon}.
Proof.

Fix λ,β(0,1)\lambda,\beta\in(0,1) and ε>0\varepsilon>0 such that 0<βε<β+ε<10<\beta-\varepsilon<\beta+\varepsilon<1. By Lemma 2.2, for large nn we have

(Sλsλ,βε)βε/2,{\mathbb{P}}\big{(}S^{\lambda}\leq s_{\lambda,\beta-\varepsilon}\big{)}\leq\beta-\varepsilon/2,

which implies that qλ,β>sλ,βεq_{\lambda,\beta}>s_{\lambda,\beta-\varepsilon}. We similarly obtain, again from Lemma 2.2, that

(Sλsλ,β+ε)(Sλ>sλ,β+ε/2)1βε/4,{\mathbb{P}}\big{(}S^{\lambda}\geq s_{\lambda,\beta+\varepsilon}\big{)}\leq{\mathbb{P}}\big{(}S^{\lambda}>s_{\lambda,\beta+\varepsilon/2}\big{)}\leq 1-\beta-\varepsilon/4,

which shows that qλ,β<sλ,β+εq_{\lambda,\beta}<s_{\lambda,\beta+\varepsilon}, for large values of nn. ∎

With these estimates at hand, we now prove Proposition 2.1.

Proof of Proposition 2.1.

Fix λ,β(0,1)\lambda,\beta\in(0,1) and let qλ,βq_{\lambda,\beta} be any β\beta-quantile of SλS^{\lambda}. By Lemmas 2.2 and 2.3 we have for small enough ε>0\varepsilon>0 and all large nn that

(fλ,β=0)=(Sλqβ,λ)(Sλsλ,β+ε)β+2ε.{\mathbb{P}}\big{(}f_{\lambda,\beta}=0\big{)}={\mathbb{P}}\big{(}S^{\lambda}\leq q_{\beta,\lambda}\big{)}\leq{\mathbb{P}}\big{(}S^{\lambda}\leq s_{\lambda,\beta+\varepsilon}\big{)}\leq\beta+2\varepsilon.

Analogously we obtain the lower bound

(fλ,β=0)=(Sλqβ,λ)(Sλsλ,βε)β2ε.{\mathbb{P}}\big{(}f_{\lambda,\beta}=0\big{)}={\mathbb{P}}\big{(}S^{\lambda}\leq q_{\beta,\lambda}\big{)}\geq{\mathbb{P}}\big{(}S^{\lambda}\leq s_{\lambda,\beta-\varepsilon}\big{)}\geq\beta-2\varepsilon.

Since ε>0\varepsilon>0 was arbitrary, it follows that (fλ,β=0)β{\mathbb{P}}(f_{\lambda,\beta}=0)\to\beta as nn\to\infty.

To prove that the sequence is noise sensitive we aim to prove that the sum of square influences tends to zero as nn\to\infty. Noise sensitivity will then follow from the BKS Theorem.

First note that bits not part of any tribe have zero influence. In addition, all remaining influences are equal due to symmetry. It will hence suffice to bound the influence of the first bit of the first tribe. For this bit to be decisive there have to be precisely qλ,β\lfloor q_{\lambda,\beta}\rfloor 1s among the remaining λ1\ell_{\lambda}-1 bits of the first tribe, as well as no other tribe with more than qλ,βq_{\lambda,\beta} 1s. Since a particular tribe is unlikely to exceed qλ,βq_{\lambda,\beta}, the probability of the latter approaches β\beta as nn\to\infty. Consequently, by independence between tribes,

Inf1(fλ,β)=(β+o(1))(Xλ1=qλ,β),\textup{Inf}_{1}(f_{\lambda,\beta})=(\beta+o(1)){\mathbb{P}}\big{(}X_{\ell_{\lambda}-1}=\lfloor q_{\lambda,\beta}\rfloor\big{)},

where XnX_{n} again is a centred binomial of nn trials. Using (15) and Lemma 2.3 we obtain for fixed values of λ,β(0,1)\lambda,\beta\in(0,1) that

Inf1(fλ,β)lognn1λ/2.\textup{Inf}_{1}(f_{\lambda,\beta})\asymp\frac{\sqrt{\log n}}{n^{1-\lambda/2}}.

Squaring the influences thus gives that

i[n]Infi(fλ,β)lognn1λ.\sum_{i\in[n]}\textup{Inf}_{i}(f_{\lambda,\beta})^{\asymp}\frac{\log n}{n^{1-\lambda}}.

For fixed λ,β(0,1)\lambda,\beta\in(0,1) the BKS Theorem hence implies that fλ,βf_{\lambda,\beta} is noise sensitive, as nn\to\infty. ∎

Due to the connection between the generalised tribes function and the left-right crossing of a square of height k=1k=1, the reader can note that by Proposition 2.1, for λ=1/2\lambda=1/2, β(0,1)\beta\in(0,1) and any β\beta-quantile qβq_{\beta} of V1nV_{1}^{n}, gives us (V1nqβ)β,{\mathbb{P}}(V^{n}_{1}\leq q_{\beta})\rightarrow\beta, and that the indicator 𝟙{V1n>qβ}\mathbbm{1}_{\{V^{n}_{1}>q_{\beta}\}} is noise sensitive as nn\to\infty.

3 Cramér-type moderate deviations for triangular arrays

In this section we state and prove a Cramér-type result for the moderate deviations of a sum of independent random variables. The result is different from Cramér’s classical result in that it applies to triangular arrays of independent, but not necessarily identically distributed, random variables. In particular, the distributions of the existing random variables are allowed to vary as more variables are included. As mentioned in the introduction, this will be one of the key steps to prove noise sensitivity in the context of first-passage percolation.

Theorem 3.1.

For every m1m\geq 1, let X1(m),X2(m),,Xm(m)X^{(m)}_{1},X_{2}^{(m)},\ldots,X^{(m)}_{m} be a sequence of independent random variables with mean zero and finite variance, and set

σm:=i=1mVar(Xi(m))m.\sigma_{m}:=\sqrt{\frac{\sum_{i=1}^{m}\textup{Var}(X^{(m)}_{i})}{m}}.

Suppose that σm1\sigma_{m}\geq 1 and that there exist global constants δ[0,1)\delta\in[0,1) and C1C\geq 1 such that for every m1m\geq 1 and i=1,2,,mi=1,2,\ldots,m, and all j2j\geq 2, we have

𝔼[|Xi(m)|j]j!(Cσm)(1+δ)j.{\mathbb{E}}\big{[}\big{|}X_{i}^{(m)}\big{|}^{j}\big{]}\leq j!\,(C\sigma_{m})^{(1+\delta)j}. (16)

Let FmF_{m} be the distribution function of the normalised sum (X1(m)++Xm(m))/(σmm)(X^{(m)}_{1}+...+X^{(m)}_{m})/(\sigma_{m}\sqrt{m}). Then, assuming σmδm1/6\sigma_{m}^{\delta}\ll m^{1/6}, we have for 1xm1/6/σmδ1\ll x\ll m^{1/6}/\sigma_{m}^{\delta} that

1Fm(x)=[1+O(σm3δx3m)][1Φ(x)].1-F_{m}(x)=\Big{[}1+O\Big{(}\sigma_{m}^{3\delta}\frac{x^{3}}{\sqrt{m}}\Big{)}\Big{]}\big{[}1-\Phi(x)\big{]}.

Our proof will follow closely the proof of Cramér’s Theorem as presented by Feller [25, Chapter XVI.7]. As is usual in the proof of theorems of this type, the proof will follow from the analysis of moment and cumulant generating functions. The moment generating function of a random variable XX is the function f(s):=𝔼[esX]f(s):={\mathbb{E}}[e^{sX}], and the cumulant generating function is defined as ψ(s):=logf(s)\psi(s):=\log f(s). These functions are not well defined for all random variables XX, but when they are, in a vicinity of the origin, they provide useful information of the random variable.

Before we tend to the proof of the above theorem, we prove a lemma regarding the regularity of the cumulant generating function of a random variable.

Lemma 3.2.

Let XX be a random variable with mean zero, variance σ2\sigma^{2} and third moment μ3\mu_{3}. Suppose there exists constants δ0\delta\geq 0 and γ>0\gamma>0 such that for all j2j\geq 2

𝔼[|X|j]j!γ(1+δ)j.{\mathbb{E}}\big{[}|X|^{j}\big{]}\leq j!\,\gamma^{(1+\delta)j}. (17)

Then, the moment generating function f(s)=𝔼[esX]f(s)={\mathbb{E}}[e^{sX}] and the cumulant generating function ψ(s)=logf(s)\psi(s)=\log f(s) are well-defined and continuously differentiable of all orders for |s|<1/γ1+δ|s|<1/\gamma^{1+\delta}. Moreover, there exists a global constant C>0C>0, not depending on the distribution of XX, such that for |s|1/(2γ1+δ)|s|\leq 1/(2\gamma^{1+\delta})

|ψ(s)12σ2s216μ3s3|\displaystyle\big{|}\psi(s)-\frac{1}{2}\sigma^{2}s^{2}-\frac{1}{6}\mu_{3}s^{3}\big{|} Cγ4(1+δ)|s|4,\displaystyle\leq C\gamma^{4(1+\delta)}|s|^{4},
|ψ(s)σ2s12μ3s2|\displaystyle\big{|}\psi^{\prime}(s)-\sigma^{2}s-\frac{1}{2}\mu_{3}s^{2}\big{|} Cγ4(1+δ)|s|3,\displaystyle\leq C\gamma^{4(1+\delta)}|s|^{3},
|ψ′′(s)σ2μ3s|\displaystyle\big{|}\psi^{\prime\prime}(s)-\sigma^{2}-\mu_{3}s\big{|} Cγ4(1+δ)|s|2.\displaystyle\leq C\gamma^{4(1+\delta)}|s|^{2}.
Proof.

Let μj:=𝔼[Xj]\mu_{j}:={\mathbb{E}}[X^{j}] denote the jjth moment of XX. Then, by (17), we have for k1k\geq 1 and |s|1/(2γ1+δ)|s|\leq 1/(2\gamma^{1+\delta})

|f(s)1j=2kμjj!sj|jk+1|μj|j!|s|jjk+1(γ1+δ|s|)j2(γ1+δ|s|)k+1.\Big{|}f(s)-1-\sum_{j=2}^{k}\frac{\mu_{j}}{j!}s^{j}\Big{|}\leq\sum_{j\geq k+1}\frac{|\mu_{j}|}{j!}|s|^{j}\leq\sum_{j\geq k+1}\big{(}\gamma^{1+\delta}|s|\big{)}^{j}\leq 2\big{(}\gamma^{1+\delta}|s|\big{)}^{k+1}. (18)

In the same way we find that there exist global constants C,C′′>0C^{\prime},C^{\prime\prime}>0 such that for |s|1/(2γ1+δ)|s|\leq 1/(2\gamma^{1+\delta})

|f(s)σ2s12μ3s2|Cγ4(1+δ)|s|3and|f′′(s)σ2μ3s|C′′γ4(1+δ)|s|2.\big{|}f^{\prime}(s)-\sigma^{2}s-\frac{1}{2}\mu_{3}s^{2}\big{|}\leq C^{\prime}\gamma^{4(1+\delta)}|s|^{3}\quad\text{and}\quad\big{|}f^{\prime\prime}(s)-\sigma^{2}-\mu_{3}s\big{|}\leq C^{\prime\prime}\gamma^{4(1+\delta)}|s|^{2}. (19)

By (18) we have, in particular, that |f(s)1|2(γ1+δ|s|)2|f(s)-1|\leq 2(\gamma^{1+\delta}|s|)^{2} and |f(s)112σ2s216μ3s3|2(γ1+δ|s|)4|f(s)-1-\frac{1}{2}\sigma^{2}s^{2}-\frac{1}{6}\mu_{3}s^{3}|\leq 2(\gamma^{1+\delta}|s|)^{4}. Since ψ(s)=logf(s)\psi(s)=\log f(s) and log(1+x)=x+O(x2)\log(1+x)=x+O(x^{2}) we obtain for |s|1/(2γ1+δ)|s|\leq 1/(2\gamma^{1+\delta}) that

|ψ(s)12σ2s216μ3s3||ψ(s)(f(s)1)|+|f(s)112σ2s216μ3s3|C(γ1+δ|s|)4,\big{|}\psi(s)-\frac{1}{2}\sigma^{2}s^{2}-\frac{1}{6}\mu_{3}s^{3}\big{|}\leq\big{|}\psi(s)-(f(s)-1)\big{|}+\big{|}f(s)-1-\frac{1}{2}\sigma^{2}s^{2}-\frac{1}{6}\mu_{3}s^{3}\big{|}\leq C(\gamma^{1+\delta}|s|)^{4},

for some constant CC not depending on the distribution of XX. Moreover, differentiation yields

ψ(s)=f(s)f(s)andψ′′(s)=f′′(s)f(s)f(s)2f(s)2.\psi^{\prime}(s)=\frac{f^{\prime}(s)}{f(s)}\quad\text{and}\quad\psi^{\prime\prime}(s)=\frac{f^{\prime\prime}(s)f(s)-f^{\prime}(s)^{2}}{f(s)^{2}}.

Hence, since 11+x=1+O(x)\frac{1}{1+x}=1+O(x), and since |f(s)|4γ2(1+δ)|s||f^{\prime}(s)|\leq 4\gamma^{2(1+\delta)}|s|, we obtain by (19) that for |s|1/(2γ1+δ)|s|\leq 1/(2\gamma^{1+\delta})

|ψ(s)σ2s12μ3s2||f(s)||1f(s)1|+|f(s)σ2s12μ3s2|Cγ4(1+δ)|s|3,\big{|}\psi^{\prime}(s)-\sigma^{2}s-\frac{1}{2}\mu_{3}s^{2}\big{|}\leq|f^{\prime}(s)|\Big{|}\frac{1}{f(s)}-1\Big{|}+\big{|}f^{\prime}(s)-\sigma^{2}s-\frac{1}{2}\mu_{3}s^{2}\big{|}\leq C\gamma^{4(1+\delta)}|s|^{3},

for some global constant CC not depending on the distribution of XX. Finally, using (19), and that |f′′(s)|4γ2(1+δ)|f^{\prime\prime}(s)|\leq 4\gamma^{2(1+\delta)} and |ψ(s)|4γ2(1+δ)|s||\psi^{\prime}(s)|\leq 4\gamma^{2(1+\delta)}|s|, we obtain that for |s|1/(2γ1+δ)|s|\leq 1/(2\gamma^{1+\delta})

|ψ′′(s)σ2μ3s||f′′(s)||1f(s)1|+|f′′(s)σ2μ3s|+|ψ(s)|2Cγ4(1+δ)|s|2,\big{|}\psi^{\prime\prime}(s)-\sigma^{2}-\mu_{3}s\big{|}\leq|f^{\prime\prime}(s)|\Big{|}\frac{1}{f(s)}-1\Big{|}+\big{|}f^{\prime\prime}(s)-\sigma^{2}-\mu_{3}s\big{|}+|\psi^{\prime}(s)|^{2}\leq C\gamma^{4(1+\delta)}|s|^{2},

for some global constant CC not depending on the distribution of XX. ∎

Proof of Theorem 3.1.

Although we will be working with triangular arrays, where the distribution of all variables are allowed to change in each step, we shall throughout the proof suppress the dependence on mm in order to ease the notation. For instance, we shall for a given value of m1m\geq 1 and i=1,2,,mi=1,2,\ldots,m denote by GiG_{i} the distribution function of Xi=Xi(m)X_{i}=X_{i}^{(m)}, although the distribution is allowed to vary with mm. Moreover, we shall let fi(s):=𝔼[esXi]f_{i}(s):={\mathbb{E}}[e^{sX_{i}}] denote the moment generating function and ψi(s):=logfi(s)\psi_{i}(s):=\log f_{i}(s) the cumulant generating function of GiG_{i}. From Lemma 3.2, by assumption (16), it follows that fi(s)f_{i}(s) and ψi(s)\psi_{i}(s) are well-defined and smooth for |s|<(Cσm)(1+δ)|s|<(C\sigma_{m})^{-(1+\delta)}, and we let

ψ(s):=1mi=1mψi(s).\psi(s):=\frac{1}{m}\sum_{i=1}^{m}\psi_{i}(s).

Note that for each m1m\geq 1 and 1im1\leq i\leq m the first two derivatives of ψi\psi_{i} are again given by

ψi(s)=fi(s)fi(s)andψi′′(s)=fi′′(s)fi(s)fi(s)2f(s)2.\psi_{i}^{\prime}(s)=\frac{f_{i}^{\prime}(s)}{f_{i}(s)}\quad\text{and}\quad\psi_{i}^{\prime\prime}(s)=\frac{f_{i}^{\prime\prime}(s)f_{i}(s)-f^{\prime}_{i}(s)^{2}}{f(s)^{2}}.

An application of the Cauchy-Schwartz inequality shows that

fi(s)2=𝔼[XiesXi]2𝔼[|Xi|esXi]2𝔼[Xi2esXi]𝔼[esXi]=fi′′(s)fi(s),f^{\prime}_{i}(s)^{2}={\mathbb{E}}\big{[}X_{i}e^{sX_{i}}\big{]}^{2}\leq{\mathbb{E}}\big{[}|X_{i}|e^{sX_{i}}\big{]}^{2}\leq{\mathbb{E}}\big{[}X_{i}^{2}e^{sX_{i}}\big{]}{\mathbb{E}}\big{[}e^{sX_{i}}\big{]}=f^{\prime\prime}_{i}(s)f_{i}(s),

so that ψi′′(s)0\psi_{i}^{\prime\prime}(s)\geq 0 on the domain where it is defined. In fact, since GiG_{i} has mean zero, the first inequality is strict and ψi′′(s)>0\psi_{i}^{\prime\prime}(s)>0, unless GiG_{i} also has zero variance. Since σm21\sigma_{m}^{2}\geq 1 by assumption, it follows that not all GiG_{i} may have zero variance, and so that

ψ′′(s)=1mi=1mψi′′(s)>0\psi^{\prime\prime}(s)=\frac{1}{m}\sum_{i=1}^{m}\psi^{\prime\prime}_{i}(s)>0

on its domain. Since ψi(0)=0\psi_{i}^{\prime}(0)=0 for each ii it follows that ψ(s)\psi^{\prime}(s) is positive and strictly increasing on the interval (0,1/(Cσm)1+δ)(0,1/(C\sigma_{m})^{1+\delta}). Consequently, for s>0s>0 and x>0x>0 the relation

mψ(s)=σmx,\sqrt{m}\psi^{\prime}(s)=\sigma_{m}x, (20)

establishes a 1-1 correspondence between ss and xx. From Lemma 3.2 we obtain that

|xσmms|=|1σm2ψ(s)s|=O(σm1+3δ|s|2),\Big{|}\frac{x}{\sigma_{m}\sqrt{m}}-s\Big{|}=\Big{|}\frac{1}{\sigma_{m}^{2}}\psi^{\prime}(s)-s\Big{|}=O(\sigma_{m}^{1+3\delta}|s|^{2}),

so that for s=o(1/σm1+3δ)s=o(1/\sigma_{m}^{1+3\delta}) we have

xσmm=(1+o(1))s.\frac{x}{\sigma_{m}\sqrt{m}}=(1+o(1))s. (21)

We shall henceforth assume that xx and ss satisfy (20) and that s=o(1/σm1+3δ)s=o(1/\sigma_{m}^{1+3\delta}), so that also (21) holds.

Following the steps of Feller, we next associate a new probability distribution ViV_{i} with the distribution GiG_{i} defined by

Vi(dy)=eψi(s)esyGi(dy),V_{i}(dy)=e^{-\psi_{i}(s)}e^{sy}\,G_{i}(dy), (22)

where ss is chosen accordingly to our previous restrictions. Analogously to the function fif_{i}, we define the moment generating function of ViV_{i} as

νi(ζ):=eζyVi(dy)=fi(ζ+s)fi(s).\nu_{i}(\zeta):=\int e^{\zeta y}\,V_{i}(dy)=\frac{f_{i}(\zeta+s)}{f_{i}(s)}.

It follows by differentiation that ViV_{i} has expectation ψi(s)\psi_{i}^{\prime}(s) and variance ψi′′(s)\psi_{i}^{\prime\prime}(s). Now, let FmF_{m}^{\star} denote the the non-normalized version of FmF_{m}, i.e. the cumulative distribution function of the sum of the mm independent variables distributed as G1,,GmG_{1},\ldots,G_{m}, and let UmU_{m}^{\star} denote ditto for mm independent variables distributed as V1,,VmV_{1},\ldots,V_{m}. Then UmU_{m}^{\star} has expectation mψ(s)m\psi^{\prime}(s) and variance mψ′′(s)m\psi^{\prime\prime}(s). Also, by comparing the moment generating functions, we observe that UmU_{m}^{\star} and FmF_{m}^{\star} satisfy a relation similar to (22) in that

Um(dy)=emψ(s)esyFm(dy).U_{m}^{\star}(dy)=e^{-m\psi(s)}e^{sy}\,F_{m}^{\star}(dy).

It follows that

1Fm(x)=1Fm(xσmm)=emψ(s)xσmmesyUm(dy).1-F_{m}(x)=1-F_{m}^{\star}(x\sigma_{m}\sqrt{m})=e^{m\psi(s)}\int_{x\sigma_{m}\sqrt{m}}^{\infty}e^{-sy}\,U_{m}^{\star}(dy). (23)

The proof will now proceed in two steps. We first analyse the expression obtained from (23) when substituting UmU_{m}^{\star} by the normal distribution with the same mean and variance. Second, we evaluate the relative error committed by this operation. So, we define AsA_{s} to be the quantity obtained by substituting UmU_{m}^{\star} by N(mψ(s),mψ′′(s))N(m\psi^{\prime}(s),m\psi^{\prime\prime}(s)) in the right-hand side of (23). Using the substitution of variables y=mψ(s)+zmψ′′(s)y=m\psi^{\prime}(s)+z\sqrt{m\psi^{\prime\prime}(s)}, and the relation in (20), we have that

As:=emψ(s)xσmmesy12πmψ′′(s)e(ymψ(s))2/(2mψ′′(s))𝑑y=em[ψ(s)sψ(s)+12s2ψ′′(s)]12π0e(z+smψ′′(s))2/2𝑑z.\begin{split}A_{s}&:=e^{m\psi(s)}\int_{x\sigma_{m}\sqrt{m}}^{\infty}e^{-sy}\frac{1}{\sqrt{2\pi m\psi^{\prime\prime}(s)}}e^{-(y-m\psi^{\prime}(s))^{2}/(2m\psi^{\prime\prime}(s))}\,dy\\ &\;=e^{m[\psi(s)-s\psi^{\prime}(s)+\frac{1}{2}s^{2}\psi^{\prime\prime}(s)]}\frac{1}{\sqrt{2\pi}}\int_{0}^{\infty}e^{-(z+s\sqrt{m\psi^{\prime\prime}(s)})^{2}/2}\,dz.\end{split} (24)

We are now interested in the behavior of

m[ψ(s)sψ(s)+12s2ψ′′(s)]=i=1m[ψi(s)sψi(s)+12s2ψi′′(s)].m\Big{[}\psi(s)-s\psi^{\prime}(s)+\frac{1}{2}s^{2}\psi^{\prime\prime}(s)\Big{]}=\sum_{i=1}^{m}\Big{[}\psi_{i}(s)-s\psi_{i}^{\prime}(s)+\frac{1}{2}s^{2}\psi_{i}^{\prime\prime}(s)\Big{]}.

Lemma 3.2, applied to each term in the sum, gives that for s=o(1/σm1+3δ)s=o(1/\sigma_{m}^{1+3\delta})

m[ψ(s)sψ(s)+12s2ψ′′(s)]=m6μ3s3+O(mσm4(1+δ)s4)=O(mσm3(1+δ)s3),m\Big{[}\psi(s)-s\psi^{\prime}(s)+\frac{1}{2}s^{2}\psi^{\prime\prime}(s)\Big{]}=\frac{m}{6}\mu_{3}s^{3}+O(m\sigma_{m}^{4(1+\delta)}s^{4})=O(m\sigma_{m}^{3(1+\delta)}s^{3}), (25)

where μ3\mu_{3} is the average of the third moments of the distributions G1,,GmG_{1},\ldots,G_{m}. The above expression vanishes as mm\to\infty if s=o(1/(m1/3σm1+δ))s=o(1/(m^{1/3}\sigma_{m}^{1+\delta})). We note that, under the assumption that σmδ=o(m1/6)\sigma_{m}^{\delta}=o(m^{1/6}), which is assumed, s=o(1/(m1/3σm1+δ))s=o(1/(m^{1/3}\sigma_{m}^{1+\delta})) is a stronger condition than s=o(1/σm1+3δ)s=o(1/\sigma_{m}^{1+3\delta}).

Since ey=1+O(y)e^{y}=1+O(y) for small values of |y||y|, it follows from (24) and (25) that for s=o(1/(m1/3σm1+δ))s=o(1/(m^{1/3}\sigma_{m}^{1+\delta}))

As=(1+O(mσm3(1+δ)s3))[1Φ(x¯)],A_{s}=\Big{(}1+O\big{(}m\sigma_{m}^{3(1+\delta)}s^{3}\big{)}\Big{)}\big{[}1-\Phi(\bar{x})\big{]},

where x¯:=smψ′′(s)\bar{x}:=s\sqrt{m\psi^{\prime\prime}(s)}. Hence, if s=o(1/(m1/3σm1+δ))s=o(1/(m^{1/3}\sigma_{m}^{1+\delta})), which by (21) is equivalent to x=o(m1/6/σmδ)x=o(m^{1/6}/\sigma_{m}^{\delta}), we obtain from (21) that

As=[1+O(σm3δx3m)][1Φ(x¯)].A_{s}=\bigg{[}1+O\bigg{(}\sigma_{m}^{3\delta}\frac{x^{3}}{\sqrt{m}}\bigg{)}\bigg{]}\big{[}1-\Phi(\bar{x})\big{]}. (26)

We now want to verify that we can substitute x¯\bar{x} by xx in (26). Observe that, by (20), we have

|xx¯|=m|1σmψ(s)sψ′′(s)|,|x-\bar{x}|=\sqrt{m}\,\Big{|}\frac{1}{\sigma_{m}}\psi^{\prime}(s)-s\sqrt{\psi^{\prime\prime}(s)}\Big{|},

Using that 1+y=1+12y+O(y2)\sqrt{1+y}=1+\frac{1}{2}y+O(y^{2}) for |y||y| small, we obtain from Lemma 3.2 that for s=o(1/σm1+3δ)s=o(1/\sigma_{m}^{1+3\delta})

sψ′′(s)=σms1+μ3σm2s+O(σm2+4δs2)=σms+μ32σms2+O(σm3+4δs3)+O(σm3+6δs3).s\sqrt{\psi^{\prime\prime}(s)}=\sigma_{m}s\sqrt{1+\frac{\mu_{3}}{\sigma_{m}^{2}}s+O(\sigma_{m}^{2+4\delta}s^{2})}=\sigma_{m}s+\frac{\mu_{3}}{2\sigma_{m}}s^{2}+O(\sigma_{m}^{3+4\delta}s^{3})+O(\sigma_{m}^{3+6\delta}s^{3}).

Another application of Lemma 3.2 thus gives, for s=o(1/σm1+3δ)s=o(1/\sigma_{m}^{1+3\delta}),

|1σmψ(s)sψ′′(s)|=O(σm3+6δs3).\Big{|}\frac{1}{\sigma_{m}}\psi^{\prime}(s)-s\sqrt{\psi^{\prime\prime}(s)}\Big{|}=O(\sigma_{m}^{3+6\delta}s^{3}).

Hence, for s=o(1/(m1/3σm1+δ))s=o(1/(m^{1/3}\sigma_{m}^{1+\delta})), which by (21) is equivalent to x=o(m1/6/σmδ)x=o(m^{1/6}/\sigma_{m}^{\delta}), (20) gives

|xx¯|=mO(σm3+6δs3)=O(σm6δx3m),|x-\bar{x}|=\sqrt{m}\,O(\sigma_{m}^{3+6\delta}s^{3})=O\bigg{(}\sigma_{m}^{6\delta}\frac{x^{3}}{m}\bigg{)}, (27)

from which we conclude that |xx¯|=o(x)|x-\bar{x}|=o(x).

Denote by φ(y)\varphi(y) the density of the standard normal distribution. Recall that as yy\rightarrow\infty

φ(y)1Φ(y)=(1+o(1))y.\frac{\varphi(y)}{1-\Phi(y)}=(1+o(1))y. (28)

Integrating the above expression between xx and x¯\bar{x} we find via (27) that for xx\rightarrow\infty such that x=o(m1/6/σmδ)x=o(m^{1/6}/\sigma_{m}^{\delta}),

|log1Φ(x¯)1Φ(x)|=O(|x+x¯||x¯x|)=O(σm6δx4m)=O(σm3δx3m),\Big{|}\log\frac{1-\Phi(\bar{x})}{1-\Phi(x)}\Big{|}=O\big{(}|x+\bar{x}||\bar{x}-x|\big{)}=O\Big{(}\sigma_{m}^{6\delta}\frac{x^{4}}{m}\Big{)}=O\Big{(}\sigma_{m}^{3\delta}\frac{x^{3}}{\sqrt{m}}\Big{)},

and finally, since ey=1+O(y)e^{y}=1+O(y) for |y||y| small, we obtain that

1Φ(x¯)1Φ(x)=1+O(σm3δx3m).\frac{1-\Phi(\bar{x})}{1-\Phi(x)}=1+O\Big{(}\sigma_{m}^{3\delta}\frac{x^{3}}{\sqrt{m}}\Big{)}.

Now, together with (26),\eqref{eq:As4}, we have that if xx\rightarrow\infty with x=o(m1/6/σmδ)x=o(m^{1/6}/\sigma_{m}^{\delta}), then

As=[1+O(σm3δx3m)][1Φ(x)].A_{s}=\bigg{[}1+O\bigg{(}\sigma_{m}^{3\delta}\frac{x^{3}}{\sqrt{m}}\bigg{)}\bigg{]}\big{[}1-\Phi(x)\big{]}. (29)

It remains to estimate the error committed by substituting UmU_{m}^{\star} by the N(mψ(s),mψ′′(s))N(m\psi^{\prime}(s),m\psi^{\prime\prime}(s)) distribution in the right-hand side of (23). Let YiY_{i} denote a generic random variable distributed according to ViV_{i}. Recall that YiY_{i} has mean ψi(s)\psi_{i}^{\prime}(s) and variance ψi′′(s)\psi_{i}^{\prime\prime}(s). Let Φs\Phi_{s} denote the cumulative distribution function of the N(mψ(s),mψ′′(s))N(m\psi^{\prime}(s),m\psi^{\prime\prime}(s)) distribution. By the Berry-Esseen Theorem (for non-identically distributed variables) we have that for all yy that

|Um(y)Φs(y)|3(mψ′′(s))3/2i=1m𝔼[|Yiψi(s)|3].\big{|}U_{m}^{\star}(y)-\Phi_{s}(y)\big{|}\leq 3\big{(}m\psi^{\prime\prime}(s)\big{)}^{-3/2}\sum_{i=1}^{m}{\mathbb{E}}\big{[}|Y_{i}-\psi^{\prime}_{i}(s)|^{3}\big{]}. (30)

Integration by parts, using (20), (23) and (30), gives

|1Fm(x)As|=emψ(s)|xσmmesyUm(dy)xσmmesyΦs(dy)|emψ(s)([esy|Um(y)Φs(y)|]mψ(s)+smψ(s)esy|Um(y)Φs(y)|𝑑y)6(mψ′′(s))3/2em[ψ(s)sψ(s)]i=1m𝔼[|Yiψi(s)|3].\begin{split}\big{|}1-F_{m}(x)-A_{s}\big{|}&=e^{m\psi(s)}\bigg{|}\int_{x\sigma_{m}\sqrt{m}}^{\infty}e^{-sy}\,U_{m}^{\star}(dy)-\int_{x\sigma_{m}\sqrt{m}}^{\infty}e^{-sy}\,\Phi_{s}(dy)\bigg{|}\\ &\leq e^{m\psi(s)}\bigg{(}-\Big{[}e^{-sy}\big{|}U^{*}_{m}(y)-\Phi_{s}(y)\big{|}\Big{]}_{m\psi^{\prime}(s)}^{\infty}+s\int_{m\psi^{\prime}(s)}^{\infty}e^{-sy}\big{|}U^{*}_{m}(y)-\Phi_{s}(y)\big{|}\,dy\bigg{)}\\ &\leq 6\big{(}m\psi^{\prime\prime}(s)\big{)}^{-3/2}\,e^{m[\psi(s)-s\psi^{\prime}(s)]}\,\sum_{i=1}^{m}{\mathbb{E}}\big{[}|Y_{i}-\psi^{\prime}_{i}(s)|^{3}\big{]}.\end{split}

We may bound the central absolute third moment, for s=o(1/σm1+3δ)s=o(1/\sigma_{m}^{1+3\delta}), as

𝔼[|Yiψi(s)|3]23(𝔼[|Yi|3]+|ψi(s)|3)8𝔼[Yi4]3/4+O(σm3(1δ)),{\mathbb{E}}\big{[}|Y_{i}-\psi_{i}^{\prime}(s)|^{3}\big{]}\leq 2^{3}\big{(}{\mathbb{E}}\big{[}|Y_{i}|^{3}\big{]}+|\psi_{i}^{\prime}(s)|^{3}\big{)}\leq 8\,{\mathbb{E}}\big{[}Y_{i}^{4}\big{]}^{3/4}+O\big{(}\sigma_{m}^{3(1-\delta)}\big{)},

where the second inequality follows from Jensen’s inequality. By an expansion similar as before, we obtain for s=o(1/σm1+3δ)s=o\big{(}1/\sigma_{m}^{1+3\delta}\big{)} that fi(s)=1+o(1)f_{i}(s)=1+o(1) and fi(4)(s)=O(σm4(1+δ))f_{i}^{(4)}(s)=O(\sigma_{m}^{4(1+\delta)}), and hence that

𝔼[Yi4]=fi(4)(s)fi(s)=O(σm4(1+δ)).{\mathbb{E}}\big{[}Y_{i}^{4}\big{]}=\frac{f_{i}^{(4)}(s)}{f_{i}(s)}=O(\sigma_{m}^{4(1+\delta)}).

Moreover, for s=o(1/σm1+3δ)s=o\big{(}1/\sigma_{m}^{1+3\delta}\big{)}, we have ψ′′(s)=σm2(1+O(σm1+3δs))=σm2(1+o(1))\psi^{\prime\prime}(s)=\sigma_{m}^{2}(1+O(\sigma_{m}^{1+3\delta}s))=\sigma_{m}^{2}(1+o(1)), which gives

|1Fm(x)As|=O(σm3δ/m)em[ψ(s)sψ(s)].\big{|}1-F_{m}(x)-A_{s}\big{|}=O(\sigma_{m}^{3\delta}/\sqrt{m})\cdot e^{m[\psi(s)-s\psi^{\prime}(s)]}. (31)

Next, we recall from (24) and the definition of x¯\bar{x} that

As=em[ψ(s)sψ(s)]ex¯2/2[1Φ(x¯)].A_{s}=e^{m[\psi(s)-s\psi^{\prime}(s)]}\,e^{\bar{x}^{2}/2}\big{[}1-\Phi(\bar{x})\big{]}.

By (28), and the observation that x/x¯=1+o(1)x/\bar{x}=1+o(1) for x=o(m1/6/σmδ)x=o(m^{1/6}/\sigma_{m}^{\delta}), we find that

As=(1+o(1))12πxem[ψ(s)sψ(s)],A_{s}=(1+o(1))\frac{1}{\sqrt{2\pi}x}e^{m[\psi(s)-s\psi^{\prime}(s)]},

and hence, together with (31), that

|1Fm(x)As|=O(σm3δxm)As.\big{|}1-F_{m}(x)-A_{s}\big{|}=O\Big{(}\sigma_{m}^{3\delta}\frac{x}{\sqrt{m}}\Big{)}A_{s}.

In conclusion,

1Fm(x)=[1+O(σm3δxm)]As,1-F_{m}(x)=\Big{[}1+O\Big{(}\sigma_{m}^{3\delta}\frac{x}{\sqrt{m}}\Big{)}\Big{]}A_{s},

which, together with (29) completes the proof. ∎

4 Moderate deviations in first-passage percolation

We now proceed to derive a moderate deviations theorem for first-passage percolation across thin rectangles. The result will follow from the moderate deviations theorem for triangular arrays (Theorem 3.1) via an approach of Chatterjee and Dey [14].

Recall the definition, in (9), that t(n,k)t(n,k) denotes the left-right crossing time of the rectangle R(n,k)=[0,n]×[0,k]R(n,k)=[0,n]\times[0,k]. By the rectangle being ‘thin’ refers to the height satisfying knαk\ll n^{\alpha} for some α<2/3\alpha<2/3. For the proof to go through, we will have to restrict the height even further.

Since we shall foremost be interested in the lower tail, i.e. deviations of t(n,k)t(n,k) below its mean, we state the theorem accordingly. An analogous statement holds for the upper tail, i.e. for deviations above the mean.

Theorem 4.1.

Suppose that FF is supported on {a,b}\{a,b\} for some 0<a<b<0<a<b<\infty. Let α<1/10\alpha<1/10 and suppose that knnαk_{n}\ll n^{\alpha}. Then, for 1xn(110α)/181\ll x\ll n^{(1-10\alpha)/18} we have

(t(n,kn)𝔼[t(n,kn)]<xVar(t(n,kn)))=[1Φ(x)][1+o(x3n(110α)/6)].{\mathbb{P}}\Big{(}t(n,k_{n})-{\mathbb{E}}[t(n,k_{n})]<-x\sqrt{\textup{Var}(t(n,k_{n}))}\Big{)}=\Big{[}1-\Phi(x)\Big{]}\bigg{[}1+o\bigg{(}\frac{x^{3}}{n^{(1-10\alpha)/6}}\bigg{)}\bigg{]}.

Moreover, the analogous statement holds for the upper tail.

While we here focus on weight distributions supported on two points, let us mention that the proof of the above theorem goes through without modification for weight distributions with bounded support.

4.1 First-passage percolation across thin rectangles

First-passage percolation on rectangular subsets of the square lattice have previously been considered by Ahlberg [1] and Chatterjee and Dey [14]. In both papers the authors prove asymptotic normality for the crossing time of thin rectangles, though by different means. In [1] the author adopts a regenerative approach that applies for fixed kk, but fails for rectangles with height growing polynomially in nn. In [14] the authors develop a different approximation scheme that works for rectangles with height kn=o(nα)k_{n}=o(n^{\alpha}) for some α<1/3\alpha<1/3. It is the latter approach, from [14], that will be of interest to us here, as it will apply to rectangles of growing height.

The idea from [14] is to chop the rectangle [0,n]×[0,k1][0,n]\times[0,k-1] up into smaller pieces, and approximate the crossing time of the original rectangle with the sum of the crossing times of the shorter stubs. This approximates the crossing time of the original rectangle with a sum of independent variables with roughly the same distribution. Chatterjee and Dey show in [14] that if the number of independent variables is large in comparison to the width of the original rectangle, then the error committed in the approximation can be controlled.

We shall below adopt their approach in the proof of Theorem 4.1. Their argument will here require a somewhat stronger restriction on the rate at which the rectangle grows. This restriction arises from the gap in upper and lower bounds on the moments of the crossing times, which will force us to apply Theorem 3.1 with some δ>0\delta>0. The following two lemmas (from [14]) bound the central moments on rectangle crossing times, and will be used in the proof of Theorem 4.1.

Lemma 4.2.

Then there exists c>0c>0 such that for all n,k1n,k\geq 1

cnkVar(t(n,k))1cn.c\frac{n}{k}\leq\textup{Var}\big{(}t(n,k)\big{)}\leq\frac{1}{c}n.
Proof.

This is Proposition 1.3 of [14]. ∎

The next result, also from [14], is a bound on central moments.

Lemma 4.3.

There exists C>0C>0 such that for all j2j\geq 2, n1n\geq 1 and knk\leq\sqrt{n} we have

𝔼[|t(n,k)𝔼[t(n,k)]|j](Cj)jnj/2.{\mathbb{E}}\Big{[}\big{|}t(n,k)-{\mathbb{E}}[t(n,k)]\big{|}^{j}\Big{]}\leq(Cj)^{j}n^{j/2}.
Proof.

This is more precise version of Proposition 5.1 in [14], which is obtained by combining Lemmas 5.4 and 5.5 of the same paper. ∎

4.2 Proof of Theorem 4.1

Fix α<1/10\alpha<1/10 and suppose that knnαk_{n}\leq n^{\alpha} for large values of nn. Let γ(1/2,1)\gamma\in(1/2,1) be a parameter to be determined later, and set mn:=n1γm_{n}:=\lfloor n^{1-\gamma}\rfloor and n:=n/mn\ell_{n}:=\lfloor n/m_{n}\rfloor. We partition the interval [0,n][0,n] into mnm_{n} subintervals of length either n\ell_{n} or n+1\ell_{n}+1 (where consecutive intervals share endpoints). Denote the intervals by I1,I2,,ImnI_{1},I_{2},\ldots,I_{m_{n}} and let YiY_{i} denote the left-right crossing time of the rectangle Ii×[0,kn1]I_{i}\times[0,k_{n}-1]. Since the intervals are disjoint (except for their boundary points) the resulting variables Y1,Y2,,YmnY_{1},Y_{2},\ldots,Y_{m_{n}} are independent and distributed as t(n,kn)t(\ell_{n},k_{n}) or t(n+1,kn)t(\ell_{n}+1,k_{n}), depending on the length of the corresponding interval.

Since every path crossing [0,n]×[0,kn1][0,n]\times[0,k_{n}-1] from left to right can be partitioned into paths crossing the intervals I1,I2,,ImnI_{1},I_{2},\ldots,I_{m_{n}}, it follows that the sum of the YiY_{i}’ s is a lower bound on t(n,kn)t(n,k_{n}). Moreover, since the edge weights are bounded by b>0b>0, and since there are no more that knk_{n} edges along the boundary between two consecutive rectangles Ii×[0,kn1]I_{i}\times[0,k_{n}-1] and Ii+1×[0,kn1]I_{i+1}\times[0,k_{n}-1], we obtain that

i=1mnYit(n,kn)i=1mnYi+bmnkn.\sum_{i=1}^{m_{n}}Y_{i}\;\leq\;t(n,k_{n})\;\leq\;\sum_{i=1}^{m_{n}}Y_{i}+b\,m_{n}k_{n}. (32)

Let Xi:=Yi𝔼[Yi]X_{i}:=Y_{i}-{\mathbb{E}}[Y_{i}] be the centered version of YiY_{i}, and set Sn:=i=1mnXiS_{n}:=\sum_{i=1}^{m_{n}}X_{i}. Taking expectation in (32), and subtracting the result from the same, yields

Snbmnknt(n,kn)𝔼[t(n,kn)]Sn+bmnkn.S_{n}-b\,m_{n}k_{n}\;\leq\;t(n,k_{n})-{\mathbb{E}}[t(n,k_{n})]\;\leq\;S_{n}+b\,m_{n}k_{n}. (33)

For n1n\geq 1, let

σn:=1mnVar(Sn).\sigma_{n}:=\sqrt{\frac{1}{m_{n}}\textup{Var}(S_{n})}.

Since nnγ\ell_{n}\sim n^{\gamma} and k=o(nα)k=o(n^{\alpha}), it follows from Lemma 4.2 that there exists c>0c>0 so that for all n1n\geq 1

cnγαVar(Xi)1cnγ,c\,n^{\gamma-\alpha}\leq\textup{Var}(X_{i})\leq\frac{1}{c}\,n^{\gamma},

and hence that

cn(γα)/2σn1cnγ/2.\sqrt{c}\,n^{(\gamma-\alpha)/2}\leq\sigma_{n}\leq\frac{1}{\sqrt{c}}\,n^{\gamma/2}. (34)

Moreover, by the reverse triangle inequality,

|𝔼[(t(n,kn)𝔼[t(n,kn)])2]𝔼[Sn2]|𝔼[(t(n,kn)𝔼[t(n,kn)]Sn)2]bmnkn,\bigg{|}\sqrt{{\mathbb{E}}\big{[}(t(n,k_{n})-{\mathbb{E}}[t(n,k_{n})])^{2}\big{]}}-\sqrt{{\mathbb{E}}\big{[}S_{n}^{2}\big{]}}\bigg{|}\leq\sqrt{{\mathbb{E}}\Big{[}\big{(}t(n,k_{n})-{\mathbb{E}}[t(n,k_{n})]-S_{n}\big{)}^{2}\Big{]}}\leq b\,m_{n}k_{n},

which with (34) gives

|Var(t(n,kn))σnmn1|bknmncn(γα)/2=O(1n(2γ13α)/2).\bigg{|}\frac{\sqrt{\textup{Var}(t(n,k_{n}))}}{\sigma_{n}\sqrt{m_{n}}}-1\bigg{|}\leq\frac{b\,k_{n}\sqrt{m_{n}}}{\sqrt{c}\,n^{(\gamma-\alpha)/2}}=O\Big{(}\frac{1}{n^{(2\gamma-1-3\alpha)/2}}\Big{)}. (35)

The above is o(1)o(1) under the condition that 2γ>1+3α2\gamma>1+3\alpha.

From Lemma 4.3 and (34) we obtain in turn (since γ>2α\gamma>2\alpha) that

𝔼[|Xi|j](Cj)jnγj/2j!(Ce)jσn(1+α/(γα))j.{\mathbb{E}}\big{[}|X_{i}|^{j}\big{]}\leq(Cj)^{j}n^{\gamma j/2}\leq j!(Ce)^{j}\sigma_{n}^{(1+\alpha/(\gamma-\alpha))j}. (36)

In particular, this means that X1,X2,,XmnX_{1},X_{2},\ldots,X_{m_{n}} satisfy (16) with δ=α/(γα)\delta=\alpha/(\gamma-\alpha). We note, in addition, that

σnα/(γα)mn1/6=O(n(γ/2)α/(γα)n(1γ)/6)=O(1n(1γ)/6αγ/[2(γα)]).\frac{\sigma_{n}^{\alpha/(\gamma-\alpha)}}{m_{n}^{1/6}}=O\Big{(}\frac{n^{(\gamma/2)\alpha/(\gamma-\alpha)}}{n^{(1-\gamma)/6}}\Big{)}=O\Big{(}\frac{1}{n^{(1-\gamma)/6-\alpha\gamma/[2(\gamma-\alpha)]}}\Big{)}. (37)

Let β1:=(2γ13α)/2\beta_{1}:=(2\gamma-1-3\alpha)/2 and β2:=(1γ)/6αγ/[2(γα)]\beta_{2}:=(1-\gamma)/6-\alpha\gamma/[2(\gamma-\alpha)] denote the exponents in the right-hand sides of (35) and (37), respectively. Now, set γ=2/3\gamma=2/3. This gives

β1=19α6andβ2>110α18,\beta_{1}=\frac{1-9\alpha}{6}\quad\text{and}\quad\beta_{2}>\frac{1-10\alpha}{18},

which for α<1/10\alpha<1/10 are strictly positive. Hence, Theorem 3.1 applies and gives that for 1xnβ21\ll x\ll n^{\beta_{2}} that

(Snσnmn<x)=[1+O(x3n3β2)][1Φ(x)].{\mathbb{P}}\bigg{(}\frac{S_{n}}{\sigma_{n}\sqrt{m_{n}}}<-x\bigg{)}=\Big{[}1+O\Big{(}\frac{x^{3}}{n^{3\beta_{2}}}\Big{)}\Big{]}\big{[}1-\Phi(x)\big{]}. (38)

Now, let

x±:=x(1±bmnknxVar(t(n,kn)))Var(t(n,kn))σnmn.x_{\pm}:=x\bigg{(}1\pm\frac{b\,m_{n}k_{n}}{x\sqrt{\textup{Var}(t(n,k_{n}))}}\bigg{)}\frac{\sqrt{\textup{Var}(t(n,k_{n}))}}{\sigma_{n}\sqrt{m_{n}}}.

By (33) we see that

(t(n,kn)𝔼[t(n,kn)]Var(t(n,kn))<x)(Snσnmn<x)=[1+O(x3n3β2)][1Φ(x)].{\mathbb{P}}\bigg{(}\frac{t(n,k_{n})-{\mathbb{E}}[t(n,k_{n})]}{\sqrt{\textup{Var}(t(n,k_{n}))}}<-x\bigg{)}\leq{\mathbb{P}}\bigg{(}\frac{S_{n}}{\sigma_{n}\sqrt{m_{n}}}<-x_{-}\bigg{)}=\Big{[}1+O\Big{(}\frac{x_{-}^{3}}{n^{3\beta_{2}}}\Big{)}\Big{]}\big{[}1-\Phi(x_{-})\big{]}.

Analogously, we obtain

(t(n,kn)𝔼[t(n,kn)]Var(t(n,kn))<x)(Snσnmn<x+)=[1+O(x+3n3β2)][1Φ(x+)].{\mathbb{P}}\bigg{(}\frac{t(n,k_{n})-{\mathbb{E}}[t(n,k_{n})]}{\sqrt{\textup{Var}(t(n,k_{n}))}}<-x\bigg{)}\geq{\mathbb{P}}\bigg{(}\frac{S_{n}}{\sigma_{n}\sqrt{m_{n}}}<-x_{+}\bigg{)}=\Big{[}1+O\Big{(}\frac{x_{+}^{3}}{n^{3\beta_{2}}}\Big{)}\Big{]}\big{[}1-\Phi(x_{+})\big{]}.

Finally, by (35), we have

x±=x(1+O(1nβ1)),x_{\pm}=x\Big{(}1+O\Big{(}\frac{1}{n^{\beta_{1}}}\Big{)}\Big{)},

which give 1Φ(x±)=[1+O(nβ1)][1Φ(x)]1-\Phi(x_{\pm})=[1+O(n^{-\beta_{1}})][1-\Phi(x)], which by the established bounds on β1\beta_{1} and β2\beta_{2} completes the proof.

5 The analysis of independent rectangle crossings

In this section we formulate and prove a preliminary version of our main theorem, concerning the minimum crossing time of a large number of disjoint rectangles. Since the rectangles are disjoint, the corresponding crossing times are independent, which facilitates the analysis.

Recall that t(n,k)t(n,k) denotes the crossing time of the rectangle [0,n]×[0,k1][0,n]\times[0,k-1], and that ti(n,k)t_{i}(n,k) denotes the translation of t(n,k)t(n,k) along the vector (0,i)(0,i), so that ti(n,k)t_{i}(n,k) is the crossing time of the rectangle [0,n]×[i,i+k1][0,n]\times[i,i+k-1]. Finally, set

τ(n,k):=min{t(i1)k(n,k):i=1,2,,n/k}.\tau^{\star}(n,k):=\min\big{\{}t_{(i-1)k}(n,k):i=1,2,\ldots,\lfloor n/k\rfloor\big{\}}.

Note that the different rectangles are disjoint and together tile the square [0,n]×[0,n1][0,n]\times[0,n-1]. Consequently, τ(n,k)\tau^{\star}(n,k) is the minimum of n/k\lfloor n/k\rfloor independent copies of t(n,k)t(n,k), and this independence will facilitate the analysis of τ(n,k)\tau^{\star}(n,k). We remark that for the choice k=1k=1 we have τ(n,k)=τ(n,k)\tau^{\star}(n,k)=\tau(n,k), and the two are equivalent to the polynomial Tribes function on n2n^{2} bits with λ=1/2\lambda=1/2.

Theorem 5.1.

Suppose that FF is supported on {a,b}\{a,b\} for some 0<a<b<0<a<b<\infty. Suppose that kn=o(n1/22)k_{n}=o(n^{1/22}). For every β(0,1)\beta\in(0,1), and any β\beta-quantile qβq_{\beta} of τ(n,kn)\tau^{\star}(n,k_{n}), we have

limn(τ(n,kn)<qβ)=1limn(τ(n,kn)>qβ)=β,\lim_{n\to\infty}{\mathbb{P}}\big{(}\tau^{\star}(n,k_{n})<q_{\beta}\big{)}=1-\lim_{n\to\infty}{\mathbb{P}}\big{(}\tau^{\star}(n,k_{n})>q_{\beta}\big{)}=\beta,

and the function fn:=𝟙{τ(n,kn)>qβ}f_{n}^{\star}:=\mathbbm{1}_{\{\tau^{\star}(n,k_{n})>q_{\beta}\}} is noise sensitive.

We remark that the above theorem remains true for kn=o(n1/11)k_{n}=o(n^{1/11}). However, in order to be able to use one of the lemmas below (Lemma 5.4) also in the proof of Theorem 1.1 in the next section, we impose the restriction kn=o(n1/22)k_{n}=o(n^{1/22}) for the conclusion of the lemma to be stronger.

Similarly as in Section 2, when analysing the generalised tribes function, we split the proof of Theorem 5.1 into several lemmas. These lemmas will also be important in the deduction of Theorem 1.1 in the next section.

Given a sequence (kn)n1(k_{n})_{n\geq 1}, set n:=n/kn\ell_{n}:=\lfloor n/k_{n}\rfloor. For β(0,1)\beta\in(0,1) let

d(n,β):=2lognlog(2logn)2log(2πlog(1/β)),d(n,\beta):=\sqrt{2\log\ell_{n}-\log(2\log\ell_{n})-2\log\big{(}\sqrt{2\pi}\log(1/\beta)\big{)}},

and set

uβ=uβ(n):=𝔼[t(n,kn)]d(n,1β)Var(t(n,kn)).u_{\beta}=u_{\beta}(n):={\mathbb{E}}[t(n,k_{n})]-d(n,1-\beta)\cdot\sqrt{\textup{Var}(t(n,k_{n}))}. (39)

The first couple of lemmas determine the asymptotic growth of the quantiles of τ(n,kn)\tau^{\star}(n,k_{n}).

Lemma 5.2.

Suppose that kn=o(n1/22)k_{n}=o(n^{1/22}). For every β(0,1)\beta\in(0,1), and any β\beta-quantile qβq_{\beta} of τ(n,kn)\tau^{\star}(n,k_{n}), we have

limn(τ(n,kn)<uβ)=β.\lim_{n\to\infty}{\mathbb{P}}\big{(}\tau^{\star}(n,k_{n})<u_{\beta}\big{)}=\beta.
Proof.

Note first that due to independence we have

(τ(n,kn)uβ)=[1(t(n,kn)<uβ)]n.{\mathbb{P}}\big{(}\tau^{\star}(n,k_{n})\geq u_{\beta}\big{)}=\big{[}1-{\mathbb{P}}\big{(}t(n,k_{n})<u_{\beta}\big{)}\big{]}^{\ell_{n}}.

Since d(n,1β)d(n,1-\beta) grows logarithmically in nn, Theorem 4.1 applies and gives that

(t(n,kn)<uβ)=(1+o(1))[1Φ(d(n,1β))].{\mathbb{P}}\big{(}t(n,k_{n})<u_{\beta}\big{)}=\big{(}1+o(1)\big{)}\big{[}1-\Phi\big{(}d(n,1-\beta)\big{)}\big{]}.

By the tail behaviour of the Gaussian distribution, in (28), we obtain

(t(n,kn)<uβ)=(1+o(1))12πd(n,1β)ed(n,1β)2/2=(1+o(1))log(1/(1β))n.{\mathbb{P}}\big{(}t(n,k_{n})<u_{\beta}\big{)}=\big{(}1+o(1)\big{)}\frac{1}{\sqrt{2\pi}\,d(n,1-\beta)}e^{-d(n,1-\beta)^{2}/2}=\big{(}1+o(1)\big{)}\frac{\log(1/(1-\beta))}{\ell_{n}}.

We conclude that

(τ(n,kn)uβ)=[1(1+o(1))log(1/(1β))n]n1β{\mathbb{P}}\big{(}\tau^{\star}(n,k_{n})\geq u_{\beta}\big{)}=\bigg{[}1-\big{(}1+o(1)\big{)}\frac{\log(1/(1-\beta))}{\ell_{n}}\bigg{]}^{\ell_{n}}\to 1-\beta

as nn\to\infty, as required. ∎

Next we relate the quantiles of τ(n,kn)\tau^{\star}(n,k_{n}) with uβu_{\beta}.

Lemma 5.3.

Suppose that kn=o(n1/22)k_{n}=o(n^{1/22}). Fix β(0,1)\beta\in(0,1) and ε>0\varepsilon>0 so that 0<βε<β+ε<10<\beta-\varepsilon<\beta+\varepsilon<1. Then, for any β\beta-quantile qβq_{\beta} of τ(n,kn)\tau^{\star}(n,k_{n}), for large nn we have

uβε<qβ<uβ+ε.u_{\beta-\varepsilon}<q_{\beta}<u_{\beta+\varepsilon}.
Proof.

Fix β(0,1)\beta\in(0,1) and ε>0\varepsilon>0 so that 0<βε<β+ε<10<\beta-\varepsilon<\beta+\varepsilon<1. By Lemma 5.2 we have for all large nn that

(τ(n,kn)uβε)(τ(n,kn)<uβε/2)βε/4.{\mathbb{P}}\big{(}\tau^{\star}(n,k_{n})\leq u_{\beta-\varepsilon}\big{)}\leq{\mathbb{P}}\big{(}\tau^{\star}(n,k_{n})<u_{\beta-\varepsilon/2}\big{)}\leq\beta-\varepsilon/4.

Consequently, for large values of nn we have that uβεu_{\beta-\varepsilon} is too small to be a β\beta-quantile of τ(n,kn)\tau^{\star}(n,k_{n}), and hence that uβε<qβu_{\beta-\varepsilon}<q_{\beta}. Similarly, again by Lemma 5.2, we have for large nn that

(τ(n,kn)uβ+ε)β+ε/2,{\mathbb{P}}\big{(}\tau^{\star}(n,k_{n})\geq u_{\beta+\varepsilon}\big{)}\geq\beta+\varepsilon/2,

and hence that qβ<uβ+εq_{\beta}<u_{\beta+\varepsilon}. ∎

Our final lemma is a uniform bound on the probability that a the left-right crossing time of a rectangle is contained in a bounded interval.

Lemma 5.4.

Suppose that kn=o(n1/22)k_{n}=o(n^{1/22}). For every β(0,1)\beta\in(0,1) and c>0c>0 we have

supxuβ(t(n,kn)[xc,x))=o(1n23/22).\sup_{x\leq u_{\beta}}{\mathbb{P}}\Big{(}t(n,k_{n})\in[x-c,x)\Big{)}=o\Big{(}\frac{1}{n^{23/22}}\Big{)}.
Proof.

Let v=v(n):=𝔼[t(n,kn)]4lognVar(t(n,kn))v=v(n):={\mathbb{E}}[t(n,k_{n})]-\sqrt{4\log n}\cdot\sqrt{\textup{Var}(t(n,k_{n}))}. By Theorem 4.1 and (28) we have that

(t(n,kn)<v)=(1+o(1))[1Φ(4logn)]=O(1n2).{\mathbb{P}}\big{(}t(n,k_{n})<v\big{)}=\big{(}1+o(1)\big{)}\big{[}1-\Phi(\sqrt{4\log n})\big{]}=O\Big{(}\frac{1}{n^{2}}\Big{)}.

Hence, it remains to show that

supx[v,uβ](t(n,kn)[tc,t))=o(1n23/22).\sup_{x\in[v,u_{\beta}]}{\mathbb{P}}\big{(}t(n,k_{n})\in[t-c,t)\Big{)}=o\Big{(}\frac{1}{n^{23/22}}\Big{)}.

We first rewrite

(t(n,kn)[xc,x))=(t(n,kn)<x)(t(n,kn)<xc).{\mathbb{P}}\big{(}t(n,k_{n})\in[x-c,x)\big{)}={\mathbb{P}}\big{(}t(n,k_{n})<x\big{)}-{\mathbb{P}}\big{(}t(n,k_{n})<x-c\big{)}.

Next we introduce a scaling function hh as

h(x)=(x𝔼[t(n,kn)])/Var(t(n,kn)),h(x)=(x-{\mathbb{E}}[t(n,k_{n})])/\sqrt{\textup{Var}(t(n,k_{n}))}, (40)

and note that h(x)h(x) is negative on [v,uβ][v,u_{\beta}]. Applying Theorem 4.1 (twice) with α=1/22\alpha=1/22 shows that the above expression equals

(1+o(1n1/11))[1Φ(h(x))](1+o(1n1/11))[1Φ(h(xc))]\Big{(}1+o\Big{(}\frac{1}{n^{1/11}}\Big{)}\Big{)}\big{[}1-\Phi\big{(}-h(x)\big{)}\big{]}-\Big{(}1+o\Big{(}\frac{1}{n^{1/11}}\Big{)}\Big{)}\big{[}1-\Phi\big{(}-h(x-c)\big{)}\big{]}

Rearranging the terms above gives

12πh(x)h(xc)ey2/2𝑑y+o(1n1/11)[1Φ(h(x))].\frac{1}{\sqrt{2\pi}}\int_{-h(x)}^{-h(x-c)}e^{-y^{2}/2}\,dy+o\Big{(}\frac{1}{n^{1/11}}\Big{)}\big{[}1-\Phi(-h(x))\big{]}.

The above expression is increasing in xx, and maximal over the given interval for x=uβx=u_{\beta}. This gives the further upper bound

cVar(t(n,kn))ed(n,1β)2/2+o(1n1/11)[1Φ(d(n,1β))].\frac{c}{\sqrt{\textup{Var}(t(n,k_{n}))}}e^{-d(n,1-\beta)^{2}/2}+o\Big{(}\frac{1}{n^{1/11}}\Big{)}\big{[}1-\Phi\big{(}d(n,1-\beta)\big{)}\big{]}.

Since n/kn\ell\sim n/k_{n} and Var(t(n,kn))n/kn\textup{Var}(t(n,k_{n}))\geq n/k_{n}, by Lemma 4.2, we obtain by definition of dd and (28) that

supx[v,uβ](t(n,kn)[xc,x))=O(lognn30/22)+o(1n23/22),\sup_{x\in[v,u_{\beta}]}{\mathbb{P}}\big{(}t(n,k_{n})\in[x-c,x)\big{)}=O\Big{(}\frac{\sqrt{\log n}}{n^{30/22}}\Big{)}+o\Big{(}\frac{1}{n^{23/22}}\Big{)},

as required. ∎

We now finally have the tools to prove Theorem 5.1.

Proof of Theorem 5.1.

Given ε>0\varepsilon>0 we have, by Lemmas 5.2 and 5.3, that for large nn

β2ε(τ(n,kn)<uβε)(τ(n,kn)<qβ)(τ(n,kn)qβ)(τ(n,kn)<uβ+ε)β+2ε.\beta-2\varepsilon\leq{\mathbb{P}}\big{(}\tau^{\star}(n,k_{n})<u_{\beta-\varepsilon}\big{)}\leq{\mathbb{P}}\big{(}\tau^{\star}(n,k_{n})<q_{\beta}\big{)}\leq{\mathbb{P}}\big{(}\tau^{\star}(n,k_{n})\leq q_{\beta}\big{)}\leq{\mathbb{P}}\big{(}\tau^{\star}(n,k_{n})<u_{\beta+\varepsilon}\big{)}\leq\beta+2\varepsilon.

Since ε>0\varepsilon>0 was arbitrary, we conclude that

limn(τ(n,kn)<qβ)=limn(τ(n,kn)qβ)=β.\lim_{n\to\infty}{\mathbb{P}}\big{(}\tau^{\star}(n,k_{n})<q_{\beta}\big{)}=\lim_{n\to\infty}{\mathbb{P}}\big{(}\tau^{\star}(n,k_{n})\leq q_{\beta}\big{)}=\beta.

In order to prove that the function (or, more precisely, sequence of functions) fn=𝟙{τ(n,kn)>qβ}f_{n}^{\star}=\mathbbm{1}_{\{\tau^{\star}(n,k_{n})>q_{\beta}\}} is noise sensitive, we aim to bound the influences of the individual edges, to show that the sum of influences squared tends to zero as nn\to\infty. The conclusion will then follow from the BKS Theorem.

First note that since the n\ell_{n} rectangles are disjoint, each edge is contained in at most one rectangle. Moreover, changing the value of an edge may affect the crossing time of the rectangle it is contained in, but not the crossing times of the remaining rectangles. In particular, edges not contained in any rectangle have influence zero. Since all rectangles are of equal dimensions, it will suffice to bound the influence of an edge contained in the first rectangle [0,n]×[0,kn1][0,n]\times[0,k_{n}-1]. So, fix an edge ee in this rectangle.

To estimate the influence of ee, note that since being pivotal does not depend on the value of the edge itself, we have

Infe(fn)=2(e pivotal)(ωe=a)=2(e pivotal,ωe=a).\textup{Inf}_{e}(f_{n}^{\star})=2\,{\mathbb{P}}(e\text{ pivotal})\,{\mathbb{P}}(\omega_{e}=a)=2\,{\mathbb{P}}(e\text{ pivotal},\omega_{e}=a).

Next, note that if there exists a left-right distance-minimising path of the rectangle not containing ee, then increasing the weight at ee has no effect on t1(n,kn)t_{1}(n,k_{n}). However, if every left-right distance-minimising path of the rectangle contains ee, then increasing ωe\omega_{e} from aa to bb will change t0(n,kn)t_{0}(n,k_{n}) by an amount at most bab-a. Hence, on the event that ee is pivotal and ωe=a\omega_{e}=a, we have that t0(n,kn)[qβ(ba),qβ)t_{0}(n,k_{n})\in[q_{\beta}-(b-a),q_{\beta}), while the remaining rectangles all have crossing time at least qβq_{\beta}. It follows, in particular, that

Infe(fn)2(t(n,kn)[qβ(ba),qβ)).\textup{Inf}_{e}(f_{n}^{\star})\leq 2\,{\mathbb{P}}\big{(}t(n,k_{n})\in[q_{\beta}-(b-a),q_{\beta})\big{)}.

Next, fix ε>0\varepsilon>0 such that β+ε<1\beta+\varepsilon<1. By Lemma 5.3, we have qβ<uβ+εq_{\beta}<u_{\beta+\varepsilon} for large nn. Consequently, it follows from Lemma 5.4 that

Infe(fn)=o(1n23/22).\textup{Inf}_{e}(f_{n}^{\star})=o\Big{(}\frac{1}{n^{23/22}}\Big{)}. (41)

The n\ell_{n} rectangles are all contained in the square [0,n]×[0,n1][0,n]\times[0,n-1]. Since the square consists of O(n2)O(n^{2}) edges, there exists a constant C>0C>0 such that

eEInfe(fn)2Cn2maxeEInfe(fn)2.\sum_{e\in E}\textup{Inf}_{e}(f_{n}^{\star})^{2}\leq Cn^{2}\max_{e\in E}\textup{Inf}_{e}(f_{n}^{\star})^{2}.

Together with (41) we get that

eEInfe(fn)2=n2o(1n23/11)=o(1n1/11).\sum_{e\in E}\textup{Inf}_{e}(f_{n}^{\star})^{2}=n^{2}\cdot o\Big{(}\frac{1}{n^{23/11}}\Big{)}=o\Big{(}\frac{1}{n^{1/11}}\Big{)}.

The desired conclusion now follows from the BKS Theorem. ∎

6 Proof of main results

We won’t be able to derive an as precise description of the asymptotics for β\beta-quantiles of τ(n,k)\tau(n,k) as for those of τ(n,k)\tau^{\star}(n,k). Nevertheless, having done much of the ground work in the previous section, we will be able to finish up the proof of Theorem 1.1 without much effort.

Proof of Theorem 1.1.

By definition of a quantile we have for any β\beta-quantile qβq_{\beta} of τ(n,kn)\tau(n,k_{n}) that

(τ(n,kn)qβ)βand(τ(n,kn)<qβ)β.{\mathbb{P}}\big{(}\tau(n,k_{n})\leq q_{\beta}\big{)}\geq\beta\quad\text{and}\quad{\mathbb{P}}\big{(}\tau(n,k_{n})<q_{\beta}\big{)}\leq\beta.

Since by definition we have τ(n,kn)τ(n,kn)\tau(n,k_{n})\leq\tau^{\star}(n,k_{n}), it follows that for every β\beta-quantile qβq_{\beta} of τ(n,kn)\tau(n,k_{n}) there exists a β\beta-quantile qβq_{\beta}^{\star} of τ(n,kn)\tau^{\star}(n,k_{n}) such that qβqβq_{\beta}\leq q_{\beta}^{\star}. Fix ε>0\varepsilon>0 so that β+ε<1\beta+\varepsilon<1. Then, qβ<uβ+εq_{\beta}<u_{\beta+\varepsilon} for large nn by Lemma 5.3. It thus follows that

β(τ(n,kn)qβ)(τ(n,kn)<qβ)+supxuβ+ε(τ(n,kn)[xc,x)).\beta\leq{\mathbb{P}}\big{(}\tau(n,k_{n})\leq q_{\beta}\big{)}\leq{\mathbb{P}}\big{(}\tau(n,k_{n})<q_{\beta}\big{)}+\sup_{x\leq u_{\beta+\varepsilon}}{\mathbb{P}}\big{(}\tau(n,k_{n})\in[x-c,x)\big{)}.

The definition of a quantile, the union bound, and Lemma 5.4 give the upper bound

β+nsupxuβ+ε(t(n,kn)[xc,x))=β+o(1n1/22).\beta+n\cdot\sup_{x\leq u_{\beta+\varepsilon}}{\mathbb{P}}\big{(}t(n,k_{n})\in[x-c,x)\big{)}=\beta+o\Big{(}\frac{1}{n^{1/22}}\Big{)}.

This proves the first part of the theorem.

In order to prove the second part of the second part we aim once again to bound the individual influences. Let ee be an edge, and recall that

Infe(fn)=2(e pivotal)(ωe=a)=2(e pivotal,ωe=a).\textup{Inf}_{e}(f_{n})=2\,{\mathbb{P}}(e\text{ pivotal})\,{\mathbb{P}}(\omega_{e}=a)=2\,{\mathbb{P}}(e\text{ pivotal},\omega_{e}=a).

Each edge in the square [0,n]×[0,n1][0,n]\times[0,n-1] is contained in at most knk_{n} translates of the rectangle [0,n]×[0,kn1][0,n]\times[0,k_{n}-1]. Changing the value of the edge may affect the crossing time of the rectangles it is contained in, but not the crossing times of the remaining rectangles. More precisely, increasing the weight at ee from aa to bb will affect ti(n,kn)t_{i}(n,k_{n}) if and only if every left-right distance minimising path of the rectangle (0,i)+[0,n]×[0,kn1](0,i)+[0,n]\times[0,k_{n}-1] contains ee. In this case, the change can result in an increase of at most bab-a. It follows by the union bound that

Infe(fn)2kn(t(n,kn)[qβ(ba),qβ)),\textup{Inf}_{e}(f_{n})\leq 2k_{n}\,{\mathbb{P}}\big{(}t(n,k_{n})\in[q_{\beta}-(b-a),q_{\beta})\big{)},

which by Lemma 5.4 gives

maxeEInfe(fn)=o(1/n).\max_{e\in E}\textup{Inf}_{e}(f_{n})=o(1/n).

Consequently,

eEInfe(fn)2Cn2maxeEInfe(fn)2=o(1).\sum_{e\in E}\textup{Inf}_{e}(f_{n})^{2}\leq Cn^{2}\max_{e\in E}\textup{Inf}_{e}(f_{n})^{2}=o(1).

Thus, the conclusion of the theorem follows from the BKS Theorem. ∎

Although not necessary, let us also provide a rough estimate on the quantiles of τ(n,kn)\tau(n,k_{n}). For β(0,1)\beta\in(0,1) let uβu_{\beta} be defined as in (39) and set

u¯β:=𝔼[t(n,kn)]Var(t(n,kn))2lognlog(2logn)2log(2πβ).\bar{u}_{\beta}:={\mathbb{E}}[t(n,k_{n})]-\sqrt{\textup{Var}(t(n,k_{n}))}\sqrt{2\log n-\log(2\log n)-2\log\big{(}\sqrt{2\pi}\beta\big{)}}.

We claim that for fixed β(0,1)\beta\in(0,1) and ε>0\varepsilon>0 so that 0<βε<β+ε<10<\beta-\varepsilon<\beta+\varepsilon<1, any β\beta-quantile qβq_{\beta} of τ(n,kn)\tau(n,k_{n}) satisfies for large nn that

u¯βε<qβ<uβ+ε.\bar{u}_{\beta-\varepsilon}<q_{\beta}<u_{\beta+\varepsilon}. (42)

Recall that, by construction, we have τ(n,kn)τ(n,kn)\tau(n,k_{n})\leq\tau^{\star}(n,k_{n}). So that for every β\beta-quantile qβq_{\beta} there exists a β\beta-quantile qβq^{\star}_{\beta} of τ(n,kn)\tau^{\star}(n,k_{n}) such that qβqβq_{\beta}\leq q_{\beta}^{\star}. The upper bound in (42) is thus immediate from Lemma 5.3.

For the lower bound, let NβN_{\beta} denote the number of rectangles with crossing time less than u¯β\bar{u}_{\beta}, i.e. let Nβ:=#{i=1,2,,nkn:ti1(n,kn)<u¯β}N_{\beta}:=\#\{i=1,2,\ldots,n-k_{n}:t_{i-1}(n,k_{n})<\bar{u}_{\beta}\}. Then,

(τ(n,kn)<u¯β)=(Nβ1).{\mathbb{P}}\big{(}\tau(n,k_{n})<\bar{u}_{\beta}\big{)}={\mathbb{P}}(N_{\beta}\geq 1).

Theorem 4.1 gives that

(t(n,kn)<u¯β)=(1+o(1))βn.{\mathbb{P}}\big{(}t(n,k_{n})<\bar{u}_{\beta}\big{)}=\big{(}1+o(1)\big{)}\frac{\beta}{n}.

Markov’s inequality hence gives that

(τ(n,kn)<u¯β)n(t(n,kn)<u¯β)=(1+o(1))β.{\mathbb{P}}\big{(}\tau(n,k_{n})<\bar{u}_{\beta}\big{)}\leq n\,{\mathbb{P}}\big{(}t(n,k_{n})<\bar{u}_{\beta}\big{)}=\big{(}1+o(1)\big{)}\beta.

We obtain for large nn that

(τ(n,kn)u¯βε)(τ(n,kn)<u¯βε/2)βε/4.{\mathbb{P}}\big{(}\tau(n,k_{n})\leq\bar{u}_{\beta-\varepsilon}\big{)}\leq{\mathbb{P}}\big{(}\tau(n,k_{n})<\bar{u}_{\beta-\varepsilon/2}\big{)}\leq\beta-\varepsilon/4.

This shows that u¯βε\bar{u}_{\beta-\varepsilon} is too small to be a β\beta-quantile for τ(n,kn)\tau(n,k_{n}) when nn is large, and hence proved the lower bound in (42).

Proof of Theorem 1.2.

We begin with the observation that

(τ(n,k)[xc,x])=(i=0nk1{ti(n,k)[xc,x]}{tj(n,k)xc,ji}).{\mathbb{P}}\big{(}\tau(n,k)\in[x-c,x]\big{)}={\mathbb{P}}\bigg{(}\bigcup_{i=0}^{n-k-1}\big{\{}t_{i}(n,k)\in[x-c,x]\big{\}}\cap\big{\{}t_{j}(n,k)\geq x-c,\forall j\neq i\big{\}}\bigg{)}.

For each ii we may find n/k1\lfloor n/k\rfloor-1 indices jj for which the rectangles corresponding to the variables tj(n,k)t_{j}(n,k) are disjoint, and disjoint of the rectangle corresponding to ti(n,k)t_{i}(n,k). The corresponding crossing times are thus independent, and exercising the union bound, we obtain that

(τ(n,k)[xc,x])n(t(n,k)[xc,x])(t(n,k)xc)n/k1.{\mathbb{P}}\big{(}\tau(n,k)\in[x-c,x]\big{)}\leq n\,{\mathbb{P}}\big{(}t(n,k)\in[x-c,x]\big{)}\,{\mathbb{P}}\big{(}t(n,k)\geq x-c\big{)}^{\lfloor n/k\rfloor-1}. (43)

We shall bound both probabilities in the above right-hand side using Theorem 4.1.

Fix α<1/22\alpha<1/22 and set β=11/22\beta=1-1/22 so that β<1α\beta<1-\alpha. Let

y:=𝔼[t(n,kn)]Var(t(n,kn))2βlogn.y:={\mathbb{E}}[t(n,k_{n})]-\sqrt{\textup{Var}(t(n,k_{n}))}\cdot\sqrt{2\beta\log n}.

Then, Theorem 4.1 and (28) give

(t(n,kn)yc)=1(1+o(1))14πβlognnβ,{\mathbb{P}}\big{(}t(n,k_{n})\geq y-c\big{)}=1-(1+o(1))\frac{1}{\sqrt{4\pi\beta\log n}\cdot n^{\beta}}, (44)

and hence that

(t(n,kn)yc)n/kn1exp(n1αβ/4πβlogn),{\mathbb{P}}\big{(}t(n,k_{n})\geq y-c\big{)}^{\lfloor n/k_{n}\rfloor-1}\leq\exp\Big{(}-n^{1-\alpha-\beta}/\sqrt{4\pi\beta\log n}\Big{)}, (45)

which decays faster than any polynomial since β<1α\beta<1-\alpha.

The reminder of the proof will closely follow that of Lemma 5.4. Let again h(x)h(x) be defined as in (40). Then, by an analogous calculation as that leading to (44), we obtain that

(t(n,kn)<h1(4logn))=O(1n2).{\mathbb{P}}\big{(}t(n,k_{n})<h^{-1}(\sqrt{4\log n})\big{)}=O\Big{(}\frac{1}{n^{2}}\Big{)}. (46)

A calculation analogous to those in Lemma 5.4 gives that for h1(4logn)xyh^{-1}(\sqrt{4\log n})\leq x\leq y we have

(t(n,kn)[xc,x])=12πh(x)h(xc)ez2/2𝑑z+o(1n1/11)[1Φ(h(x))],{\mathbb{P}}\big{(}t(n,k_{n})\in[x-c,x]\big{)}=\frac{1}{\sqrt{2\pi}}\int_{-h(x)}^{-h(x-c)}e^{-z^{2}/2}\,dz+o\Big{(}\frac{1}{n^{1/11}}\Big{)}\big{[}1-\Phi(-h(x))\big{]},

which is maximal for x=yx=y. Together with (46) we thus get, for some constant C<C<\infty, that

supxy(t(n,kn)[xc,x])Cn1/11+βlogn=Cn1+1/22logn.\sup_{x\leq y}{\mathbb{P}}\big{(}t(n,k_{n})\in[x-c,x]\big{)}\leq\frac{C}{n^{1/11+\beta}\sqrt{\log n}}=\frac{C}{n^{1+1/22}\sqrt{\log n}}. (47)

Finally, combining (43), (45) and (47), we obtain that

supx0(τ(n,k)[xc,x])Cn1/22logn,\sup_{x\geq 0}{\mathbb{P}}\big{(}\tau(n,k)\in[x-c,x]\big{)}\leq\frac{C}{n^{1/22}\sqrt{\log n}},

as required. ∎

7 Further directions

We will devote this last section to indicate some future directions of research and open problems related to Benjamini’s problem and the work of this paper.

We started out with the problem of whether ‘being above the median’ is a noise sensitive property for the point-to-point passage time Tn=T(0,n𝐞1)T_{n}=T(0,n{\bf e}_{1}). Due to the limited understanding of fluctuations in first-passage percolation, we have had to resort to restricting the problem in order to make progress. This led us, in the introduction, to call for

supx0(Tn[x,x+c])0as n,\sup_{x\geq 0}{\mathbb{P}}\big{(}T_{n}\in[x,x+c]\big{)}\to 0\quad\text{as }n\to\infty,

for every c>0c>0.

More precise results regarding the nature of fluctuations have been established in related models of spatial growth, such as increasing subsequences in the place, last-passage percolation with exponential or geometric weights, Brownian last-passage percolation, as well as for the largest eigenvalue of random matrices. It appears as if these results are in themselves insufficient to answer Benjamini’s question. In addition, these settings do not fit into the framework of Boolean functions. Hence, solving Benjamini’s problem in these settings remains an interesting open problem.

Another relevant question regards the relation between noise sensitivity of being above a certain quantile of some real-valued sequence of functions fn:{0,1}nf_{n}:\{0,1\}^{n}\to{\mathbb{R}}, and the asymptotic independence of fn(ω)f_{n}(\omega) and fn(ωε)f_{n}(\omega^{\varepsilon}). In particular, given that

(fn(ω)>q,fn(ωε)>q)(fn(ωε))20as n{\mathbb{P}}\big{(}f_{n}(\omega)>q,f_{n}(\omega^{\varepsilon})>q\big{)}-{\mathbb{P}}\big{(}f_{n}(\omega^{\varepsilon})\big{)}^{2}\to 0\quad\text{as }n\to\infty (48)

for every quantile qq of fnf_{n}, is it then also true that

Corr(fn(ω),fn(ωε))0as n?\textup{Corr}\big{(}f_{n}(\omega),f_{n}(\omega^{\varepsilon})\big{)}\to 0\quad\text{as }n\to\infty? (49)

For many sequences it is natural to expect that the mean of fnf_{n} corresponds to one of its quantiles, and thus that if (48) holds, then the signs of fn(ω)𝔼[fn]f_{n}(\omega)-{\mathbb{E}}[f_{n}] and fn(ωε)𝔼[fn]f_{n}(\omega^{\varepsilon})-{\mathbb{E}}[f_{n}] are asymptotically independent, and hence that (49) should hold. We do not know whether this is true in general.

Finally, the reader may wonder why we consider the restricted square crossing time τ(n,k)\tau(n,k) now that already the rectangle crossing time t(n,k)t(n,k) is known to obey a Gaussian central limit theorem. Well, for fixed kk we expect that t(n,k)t(n,k) being above its median is a noise stable property, and hence not noise sensitive. Indeed, the case k=1k=1 coincides with the classical Majority function on nn bits, which is well-known to be noise stable; see e.g. [29]. For diverging sequences (kn)n1(k_{n})_{n\geq 1} we conjecture that ‘being above the median’ is a noise sensitive property for t(n,kn)t(n,k_{n}). We motivate this by an heuristic calculation similar to (8), which suggests that for a given edge ee

Infe(𝟙{t(n,k)>m})(eπn)(|t(n,k)m|ba)1k1Var(t(n,k)),\textup{Inf}_{e}(\mathbbm{1}_{\{t(n,k)>m\}})\,\asymp\,{\mathbb{P}}\big{(}e\in\pi_{n}\big{)}\,{\mathbb{P}}\big{(}|t(n,k)-m|\leq b-a\big{)}\,\asymp\,\frac{1}{k}\frac{1}{\sqrt{\textup{Var}(t(n,k))}},

and hence that (there are about nknk influential edges)

eInfe(𝟙{t(n,k)>m})2nk1Var(t(n,k)).\sum_{e}\textup{Inf}_{e}(\mathbbm{1}_{\{t(n,k)>m\}})^{2}\asymp\frac{n}{k}\frac{1}{\textup{Var}(t(n,k))}.

It is believed that Var(t(n,k))n/k\textup{Var}(t(n,k))\asymp n/\sqrt{k} whenever k=o(n2/3)k=o(n^{2/3}), and this has been proved to be the case in a related model by Dey, Joseph and Peled [24]. However, in first-passage percolation, the best bounds only give Var(t(n,k))n/k\textup{Var}(t(n,k))\geq n/k, which would give a constant upper bound on the sum of influences squared; see [14]. Hence, one would need to improve upon the variance bound in order to establish noise sensitivity of the rectangle crossing variables.

References

  • [1] Daniel Ahlberg “Asymptotics of first-passage percolation on one-dimensional graphs” In Adv. in Appl. Probab. 47.1, 2015, pp. 182–209
  • [2] Daniel Ahlberg, Erik Broman, Simon Griffiths and Robert Morris “Noise sensitivity in continuum percolation” In Israel J. Math. 201.2, 2014, pp. 847–899
  • [3] Daniel Ahlberg, Maria Deijfen and Matteo Sfragara “Chaos, concentration and multiple valleys in first-passage percolation” In arXiv preprint arXiv:2302.11367, 2023
  • [4] Daniel Ahlberg, Maria Deijfen and Matteo Sfragara “From stability to chaos in last-passage percolation” In arXiv preprint arXiv:2302.11379, 2023
  • [5] Daniel Ahlberg, Simon Griffiths, Robert Morris and Vincent Tassion “Quenched Voronoi percolation” In Advances in Mathematics. 286, 2016, pp. 889–911
  • [6] Daniel Ahlberg and Christopher Hoffman “Random coalescing geodesics in first-passage percolation” In arXiv preprint arXiv:1609.02447, 2016
  • [7] R.. Bahadur “Some approximations to the binomial distribution function” In Ann. Math. Statist. 31, 1960, pp. 43–54
  • [8] Jinho Baik, Percy Deift and Kurt Johansson “On the distribution of the length of the longest increasing subsequence of random permutations” In J. Amer. Math. Soc. 12.4, 1999, pp. 1119–1178
  • [9] Michael Ben-Or and Nathan Linial “Collective coin flipping, robust voting schemes and minima of Banzhaf values” In 26th Annual Symposium on Foundations of Computer Science (sfcs 1985), 1985, pp. 408–416 IEEE
  • [10] I. Benjamini, G. Kalai and O. Schramm “Noise sensitivity of Boolean functions and applications to percolation” In Publications Mathématiques de L’Institut des Hautes Scientifiques 90.1, 1999, pp. 5–43
  • [11] Itai Benjamini, Gil Kalai and Oded Schramm “First passage percolation has sublinear distance variance” In Ann. Probab. 31.4, 2003, pp. 1970–1978
  • [12] Stéphane Boucheron, Gábor Lugosi and Pascal Massart “Concentration inequalities” A nonasymptotic theory of independence, With a foreword by Michel Ledoux Oxford University Press, Oxford, 2013, pp. x+481
  • [13] Alan J Bray and Michael A Moore “Chaotic nature of the spin-glass phase” In Physical review letters 58.1 APS, 1987, pp. 57
  • [14] S. Chatterjee and P. Dey “Central limit theorem for first-passage percolation time across thin cylinders” In Probability Theory and Related Fields 156, 2013, pp. 613–663
  • [15] Sourav Chatterjee “A general method for lower bounds on fluctuations of random variables” In Ann. Probab. 47.4, 2019, pp. 2140–2171
  • [16] Sourav Chatterjee “Chaos, concentration, and multiple valleys” In arXiv, 2008 arXiv:0810.4221
  • [17] Sourav Chatterjee “Disorder chaos and multiple valleys in spin glasses” In arXiv, 2009 arXiv:0907.3381
  • [18] Sourav Chatterjee “Superconcentration and Related Topics” Springer International Publishing, 2014
  • [19] Michael Damron and Jack Hanson “Bigeodesics in first-passage percolation” In Comm. Math. Phys. 349.2, 2017, pp. 753–776
  • [20] Michael Damron, Jack Hanson, David Harper and Wai-Kit Lam “Transitions for exceptional times in dynamical first-passage percolation” In Probab. Theory Related Fields 185.3-4, 2023, pp. 1039–1085
  • [21] Michael Damron, Jack Hanson, Christian Houdré and Chen Xu “Lower bounds for fluctuations in first-passage percolation for general distributions” In Ann. Inst. Henri Poincaré Probab. Stat. 56.2, 2020, pp. 1336–1357
  • [22] Michael Damron, Christian Houdré and Alperen Özdemir “Fluctuation bounds for first-passage percolation on the square, tube, and torus” In arXiv, 2022
  • [23] Barbara Dembin, Dor Elboim and Ron Peled “Coalescence of geodesics and the BKS midpoint problem in planar first-passage percolation” In arXiv preprint arXiv:2204.02332, 2022
  • [24] Partha Dey, Mathew Joseph and Ron Peled “Longest increasing path within the critical strip” In arXiv preprint arXiv:1808.08407, 2018
  • [25] William Feller “An introduction to probability theory and its applications. Vol. II.”, Second edition New York: John Wiley & Sons Inc., 1971, pp. xxiv+669
  • [26] Daniel S Fisher and David A Huse “Ordered phase of short-range Ising spin-glasses” In Physical review letters 56.15 APS, 1986, pp. 1601
  • [27] Shirshendu Ganguly and Alan Hammond “Stability and chaos in dynamical last passage percolation” In arXiv, 2020 arXiv:2010.05837
  • [28] Christophe Garban, Gábor Pete and Oded Schramm “The Fourier spectrum of critical percolation” In Acta Mathematica 205, 2008
  • [29] Christophe Garban and Jeffrey E. Steif “Noise Sensitivity of Boolean Functions and Percolation”, Institute of Mathematical Statistics Textbooks Cambridge University Press, 2014
  • [30] Christophe Garban and Hugo Vanneuville “Bargmann-Fock percolation is noise sensitive” In Electronic Journal of Probability 25, 2020, pp. 1–20
  • [31] Christina Goldschmidt, Simon Griffiths and Alex Scott “Moderate deviations of subgraph counts in the Erdös-Rényi random graphs G(n,m)G(n,m) and G(n,p)G(n,p) In arXiv, 2019 arXiv:1902.06830
  • [32] Kurt Johansson “Transversal fluctuations for increasing subsequences on the plane” In Probab. Theory Related Fields 116.4, 2000, pp. 445–456
  • [33] Mehran Kardar, Giorgio Parisi and Yi-Chen Zhang “Dynamic scaling of growing interfaces” In Phys. Rev. Lett. 56, 1986, pp. 889–892
  • [34] Harry Kesten “On the time constant and path length of first-passage percolation” In Advances in Applied Probability 12.4 Cambridge University Press, 1980, pp. 848–863
  • [35] Günter Last, Giovanni Peccati and D. Yogeshwaran “Phase transitions and noise sensitivity on the Poisson space via stopping sets and decision trees” In arXiv, 2021 arXiv:2101.07180
  • [36] Eyal Lubetzky and Jeffrey E. Steif “Strong noise sensitivity and random graphs” In Ann. Probab. 43.6, 2015, pp. 3239–3278
  • [37] R. Pemantle and Y. Peres “Planar first-passage percolation times are not tight” In Probability and phase transition (Cambridge, 1993) 420, NATO Adv. Sci. Inst. Ser. C: Math. Phys. Sci. Kluwer Acad. Publ., Dordrecht, 1994, pp. 261–264
  • [38] O. Schramm and J.E. Steif “Quantitative noise sensitivity and exceptional times for percolation” In Annals of Mathematics 171.2, 2010, pp. 619–672
  • [39] Vincent Tassion and Hugo Vanneuville “Noise sensitivity of percolation via differential inequalities” In arXiv, 2020 arXiv:2011.04572