This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Random Composition of L-S-V Maps Sampled Over Large Parameter Ranges

Christopher Bose Department of Mathematics and Statistics, University of Victoria, PO BOX 1700 STN CSC, Victoria, B.C., Canada V8W 2Y2 Anthony Quas Department of Mathematics and Statistics, University of Victoria, PO BOX 1700 STN CSC, Victoria, B.C., Canada V8W 2Y2 Matteo Tanzi Courant Institute of Mathematical Sciences, New York University, New York, NY 10012, USA
Abstract

Liverani-Saussol-Vaienti (L-S-V) maps form a family of piecewise differentiable dynamical systems on [0,1][0,1] depending on one parameter ω+\omega\in\mathbb{R}^{+}. These maps are everywhere expanding apart from a neutral fixed point. It is well known that depending on the amount of expansion close to the neutral point, they have either an absolutely continuous invariant probability measure and polynomial decay of correlations (ω<1\omega<1), or a unique physical measure that is singular and concentrated at the neutral point (ω>1\omega>1). In this paper, we study the composition of L-S-V maps whose parameters are randomly sampled from a range in +\mathbb{R}^{+}, and where these two contrasting behaviours are mixed. We show that if the parameters ω<1\omega<1 are sampled with positive probability, then the stationary measure of the random system is absolutely continuous; the annealed decay rate of correlations is close (or in some cases equal) to the fastest rate of decay among those of the sampled systems; and suitably rescaled Birkhoff averages converge to limit laws. In contrast to previous studies where ω[0,1]\omega\in[0,1], we allow ω>1\omega>1 in our sampling distribution. We also show that one can obtain similar decay of correlation rates for ω+\omega\in\mathbb{R}^{+}, when sampling is done with respect to a family of smooth, heavy-tailed distributions.

1 Introduction

Let us consider the one-parameter family of L-S-V maps [17] {fω}ω+\{f_{\omega}\}_{\omega\in\mathbb{R}^{+}}, fω:[0,1][0,1]f_{\omega}:[0,1]\rightarrow[0,1], defined as

fω(x)={x(1+2ωxω)x[0,12)2x1x[12,1]f_{\omega}(x)=\left\{\begin{array}[]{cr}x(1+2^{\omega}x^{\omega})&x\in[0,\tfrac{1}{2})\\ 2x-1&x\in[\tfrac{1}{2},1]\end{array}\right. (1)

These maps were introduced as a simplified version of the previously studied Pomeau-Manneville family xx+x1+ω(mod 1)x\rightarrow x+x^{1+\omega}(\textnormal{mod}\,1) [20], retaining the essential property of having a neutral fixed point at x=0x=0 but all fωf_{\omega} having the same two intervals of monotonicity and uniformly expanding, affine, right-hand branch***It is reasonable to expect that the results one can obtain for L-S-V maps hold for the original P-M maps. However we take advantage of the affine second branch in L-S-V at various points in the proofs so extending our arguments to the non-affine case is likely to be technically complicated. Since then, L-S-V maps have become a standard example of dynamical systems with intermittency, alternating stretches of time where orbits exhibit chaotic behaviour, and long stretches of time where they are almost constant and close to zero. The uniform expansion of the maps away from the fixed point x=0x=0 is responsible for the chaotic behaviour, while the fact that fω(0)=1f^{\prime}_{\omega}(0)=1 implies that it takes a long time before orbits can escape the vicinity of x=0x=0. It is well known that for ω(0,1)\omega\in(0,1), fωf_{\omega} has a unique physical probability measure which is absolutely continuous and exhibits polynomial decay of correlations where the rate of decay depends on ω\omega [23]. For ω1\omega\geq 1, fωf_{\omega} does not have any invariant absolutely continuous probability measure, but it has a unique invariant absolutely continuous infinite measure, and the Dirac delta concentrated at zero is the unique probability physical measure. This is due to the fact that for ω1\omega\geq 1, the repulsion in a neighbourhood of zero is so small that, asymptotically, almost every orbit spends most of its time arbitrarily close to zero.

After fixing a compact interval [α,β]+[\alpha,\beta]\subset\mathbb{R}^{+} with αβ\alpha\leq\beta, we study the composition of maps sampled randomly with respect to a given probability measure ν\nu on [α,β][\alpha,\beta], and characterise the average statistical properties of their orbits (annealed results). The case where [α,β](0,1][\alpha,\beta]\subset(0,1] has been previously considered. In [4] and [3] the authors show that if the distribution is discrete and samples only finitely many values in [α,β][\alpha,\beta], the annealed correlations decay at a rate equal to that of the system with the smallest ω[α,β]\omega\in[\alpha,\beta] which is sampled with positive probability. Other works deal with quenched results that, in contrast with annealed results, establish decay of correlations and convergence to limit laws for almost every sequence of maps sampled with respect to the given measure on [α,β](0,1][\alpha,\beta]\subset(0,1] [6] (see also [5] for quenched results on random composition of unimodal maps). In [2], [19], and [18] an arbitrary composition of maps (sequential random systems) from [α,β](0,1)[\alpha,\beta]\subset(0,1), where α\alpha and β\beta satisfy some additional technical conditions, is shown to give decay of correlations at a speed bounded above by that of the system β\beta, satisfy central limit theorems, and large deviations principles.

In contrast with the above cited works, we focus on the case where α<1<β\alpha<1<\beta and thus maps with both ω<1\omega<1 and ω>1\omega>1 are composed. In this case, there is competition between two contrasting behaviours: maps with ω<1\omega<1 tend to spread mass over the whole space, while maps with ω>1\omega>1, although still expanding on most of the space, tend to accumulate mass at the neutral fixed point. We show that it is enough for maps with ω<1\omega<1 to be sampled with positive probability, to ensure that the average random system has an absolutely continuous stationary probability measure, polynomial decay of annealed correlations, and convergence to limit laws.

The case of discrete ν\nu (i.e. ν=piδωi\nu=\sum p_{i}\delta_{\omega_{i}}) was previously treated in [3] for β1\beta\leq 1 using the Young tower approach [23] to find estimates on the decay of correlations. In this analysis, the base of the skew product is conjugated to a piecewise affine and uniformly expanding system. This allows one, after inducing on a suitable subset of the phase space, to reduce the analysis to the study of a Gibbs-Markov system [1] with countably many invertible branches. This construction is not possible in the case of a non-discrete ν\nu where one can find uncountably many inverse branches all defined on measurable sets of zero measure that cover a set of positive measure, thus obstructing the construction of a countable Gibbs-Markov structure.

Our main approach in this paper is based on the renewal theory of operators which was introduced by Sarig [21] and further developed by Gouëzel [10]), and which can deal with any measure ν\nu on [α,β]+[\alpha,\beta]\subset\mathbb{R}^{+}. For the machinery to work, it is crucial to bound the distortion of the composition of different maps with parameters chosen arbitrarily from [α,β][\alpha,\beta]. For a single (deterministic) map this was done by Young [23] using a direct calculation, valid for all ω>0\omega>0. For random maps, distortion estimates have been obtained in the case β1\beta\leq 1 using the Koebe principle and non-positive Schwartzian derivative as in [11] and [4]. However, this technique for random maps fails when β>1\beta>1 since fωf_{\omega} has points where the Schwartzian derivative is positive when ω>1\omega>1. In the following we give a direct estimate of the distortion, more in the spirit of [23], that encompasses the general case [α,β]+[\alpha,\beta]\subset\mathbb{R}^{+}, including β>1\beta>1.

In the case where the interval of parameters is unbounded, i.e. β=\beta=\infty, the approach above does not work as a uniform bound on the distortion is missing. However, in the special situation where ν\nu is an absolutely continuous measure on [α,+)[\alpha,+\infty) with very regular density, we present a second approach that exploits the continuous distribution in parameter space by looking at the system as a mixture between a diffusion process, and a deterministic uniformly expanding system. The estimates on the (annealed) decay of correlations are obtained using the theory of Markov chains with subgeometric rates of convergence [22]. We are going to present this approach in the case where the distribution of ν\nu has fat polynomial tails at infinity, meaning that large parameters are sampled with high probability.

2 Setting and Results

Let us consider the one-parameter family of L-S-V maps {fω}ω+\{f_{\omega}\}_{\omega\in\mathbb{R}^{+}} as in (1) above. Given α,β+\alpha,\beta\in\mathbb{R}^{+} with α<1\alpha<1 and β>α\beta>\alpha, we consider a probability measure ν\nu on the compact interval [α,β][\alpha,\beta], and ν0\nu^{\mathbb{N}_{0}} the product measure on Ω:=[α,β]0\Omega:=[\alpha,\beta]^{\mathbb{N}_{0}}. We assume without loss of generality that α\alpha belongs to the topological support of ν\nu (that is ν([α,α+δ])>0\nu([\alpha,\alpha+\delta])>0 for any δ>0\delta>0).

These data define the random dynamical system taking skew-product form F:Ω×[0,1]Ω×[0,1]F:\Omega\times[0,1]\rightarrow\Omega\times[0,1]

F(ω,x)=(σ𝝎,fω0x),𝝎=(ω0ω1)Ω,x[0,1]F(\omega,x)=(\sigma\boldsymbol{\omega},f_{\omega_{0}}x),\quad\boldsymbol{\omega}=(\omega_{0}\omega_{1}...)\in\Omega,\;x\in[0,1] (2)

where the reference measure on Ω×[0,1]\Omega\times[0,1] is =ν0m\mathbb{P}=\nu^{\mathbb{N}_{0}}\otimes m, with mm the Lebesgue measure on [0,1][0,1]. Given 𝝎Ω\boldsymbol{\omega}\in\Omega and nn\in\mathbb{N}, we denote by f𝝎n=fωn1fω0f_{\boldsymbol{\omega}}^{n}=f_{\omega_{n-1}}\circ...\circ f_{\omega_{0}}. We are going to describe the annealed statistical properties of this random system, i.e. the statistical properties of FF averaged with respect to the reference measure ν0\nu^{\mathbb{N}_{0}}. In the following, for a measurable AΩ×[0,1]A\subset\Omega\times[0,1] we denote by

τA(𝝎,x):=inf{n|Fn(𝝎,x)A},\tau_{A}(\boldsymbol{\omega},x):=\inf\{n\in\mathbb{N}|\;F^{n}(\boldsymbol{\omega},x)\in A\},

and for a measurable J[0,1]J\subset[0,1] we denote with an abuse of notation

τJ(𝝎,x):=inf{n|Fn(𝝎,x)Ω×J}.\tau_{J}(\boldsymbol{\omega},x):=\inf\{n\in\mathbb{N}|\;F^{n}(\boldsymbol{\omega},x)\in\Omega\times J\}.

Decay of correlations.

We show that the random system has an absolutely continuous stationary probability measure and exhibits decay of correlations at polynomial speed.

Theorem 2.1.

Let 0<α<10<\alpha<1, αβ\alpha\leq\beta and let ν\nu be a probability measure on [α,β][\alpha,\beta] such that α\alpha lies in the topological support of ν\nu. Then there is an absolutely continuous stationary measure π\pi, i.e. ν0π\nu^{\mathbb{N}_{0}}\otimes\pi is invariant under FF. For ψL([0,1],m)\psi\in L^{\infty}([0,1],m), φLip([0,1],)\varphi\in\mathrm{Lip}([0,1],\mathbb{R}) and all 1<γ<1α1<\gamma<\frac{1}{\alpha}, we have

|ψFn(𝝎,x)φ(x)𝑑(𝝎,x)φ(x)𝑑m(x)ψ(x)𝑑π(x)|𝒪(1nγ1).\left|\int\psi\circ F^{n}(\boldsymbol{\omega},x)\varphi(x)d\mathbb{P}(\boldsymbol{\omega},x)-\int\varphi(x)dm(x)\int\psi(x)d\pi(x)\right|\leq\mathcal{O}\left(\frac{1}{n^{\gamma-1}}\right).

It is natural here to compute correlation integrals with respect to the reference measure \mathbb{P} since one does not, in general, know π\pi explicitly. Note that φ(x)𝑑m(x)=φ(x)𝑑(𝝎,x)\int\varphi(x)dm(x)=\int\varphi(x)d\mathbb{P}(\boldsymbol{\omega},x) so, heuristically, we interpret this result as weak convergence of the sequence ψFn\psi\circ F^{n} to ψ(x)𝑑π(x)\int\psi(x)d\pi(x), measured against Lipschitz test functions. An elementary calculation shows \mathbb{P} (and mm) above can be replaced by the stationary measure ν0π\nu^{\mathbb{N}_{0}}\otimes\pi in both integrals, obtaining the same decay rates with respect to the stationary measure, more in line with results in classical probability.

We prove this theorem using the renewal theory for operators (see Section 5.1). This approach is more standard in the study of dynamical systems with nonuniform hyperbolic properties, and relies on the bound on distortion given in Section 3. However, we have to deal with the fact that the induced map we are going to study, FY:=FτYF_{Y}:=F^{\tau_{Y}} with Y:=Ω×[1/2,1]Y:=\Omega\times[1/2,1], is not Gibbs-Markov, as is usually assumed to be the case. In fact, there is an uncountable partition of sets with zero measure that are mapped bijectively onto YY by FYF_{Y} and that cover a set of positive measure. Therefore, we have to prove the spectral properties of the operators involved directly.

In Section 6 we consider the case where ν\nu is absolutely continuous with a power law distribution, and prove a decay of correlations statement for bounded measurable functions. We look at the diffusion process induced by the skew product map FF on the vertical fibre [0,1][0,1] and use the theory of Markov chains with subgeometric rates of convergence to their stationary state. This approach does not use the bound on the distortion, whose role in the arguments is played by the diffusion and gives decay of correlations for arbitrary LL^{\infty} functions.

For 0<α<10<\alpha<1 and ε>0\varepsilon>0, let να,ε\nu_{\alpha,\varepsilon} denote the measure on [α,)[\alpha,\infty) defined by

να,ε(A)=Aεαεtε+1𝑑t,\nu_{\alpha,\varepsilon}(A)=\int_{A}\frac{\varepsilon\alpha^{\varepsilon}}{t^{\varepsilon+1}}\,dt,

for any A[α,)A\subset[\alpha,\infty). Alternatively, να,ε\nu_{\alpha,\varepsilon} is characterized by να,ε(t,)=(tα)ε\nu_{\alpha,\varepsilon}(t,\infty)=(\frac{t}{\alpha})^{-\varepsilon} for any tαt\geq\alpha.

Theorem 2.2 (LL^{\infty} decay of correlations for power law parameter distributions).

Let 0<α<10<\alpha<1 and let ε>0\varepsilon>0 and consider the random composition of LSV maps where the parameters are chosen i.i.d. from να,ε\nu_{\alpha,\varepsilon}. The corresponding Markov chain has a stationary probability distribution, π\pi. Let φ,ψL[0,1]\varphi,\psi\in L^{\infty}[0,1]. Then for all 1<γ<1α1<\gamma<\frac{1}{\alpha},

|ψFn(𝝎,x)φ(x)𝑑m(x)𝑑να,ε(𝝎)φ(x)𝑑m(x)ψ(y)𝑑π(y)|=o(1nγ1).\left|\int\int\psi\circ F^{n}(\boldsymbol{\omega},x)\varphi(x)\,dm(x)\,d\nu_{\alpha,\varepsilon}^{\mathbb{N}}(\boldsymbol{\omega})-\int\varphi(x)\,dm(x)\int\psi(y)\,d\pi(y)\right|=\textnormal{o}\left(\frac{1}{n^{\gamma-1}}\right).

Convergence to limit laws.

Let φ:[0,1]\varphi:[0,1]\rightarrow\mathbb{R} be a Lipschitz function. Define φY:Ω×[1/2,1]\varphi_{Y}:\Omega\times[1/2,1]\rightarrow\mathbb{R} by

φY(𝝎,x):=i=0τY(𝝎,x)1φπ2Fi(𝝎,x),\varphi_{Y}(\boldsymbol{\omega},x):=\sum_{i=0}^{\tau_{Y}(\boldsymbol{\omega},x)-1}\varphi\circ\pi_{2}\circ F^{i}(\boldsymbol{\omega},x),

where π2\pi_{2} denotes projection to the second coordinate.

We denote the Birkhoff sums with respect to FYF_{Y} by

Snψ:=ψ+ψFY++ψFYn1S^{n}\psi:=\psi+\psi\circ F_{Y}+...+\psi\circ F_{Y}^{n-1}
Theorem 2.3.

Let 0<α<10<\alpha<1, α1/2\alpha\neq 1/2, and α<β\alpha<\beta. Let ν\nu be a probability measure supported on [α,β][\alpha,\beta] with α\alpha in the topological support of ν\nu. Let φ\varphi be a Lipschitz function on [0,1][0,1] with φ(0)0\varphi(0)\neq 0. Then there is a stable law 𝒵\mathcal{Z} a sequence AnA^{\prime}_{n}, and a sequence BnB^{\prime}_{n} such that

limnSnφYAnBn𝒵\lim_{n\rightarrow\infty}\frac{S^{n}\varphi_{Y}-A^{\prime}_{n}}{B^{\prime}_{n}}\rightarrow\mathcal{Z}

where the convergence is in distribution. Furthermore

  • i)

    if α<1/2\alpha<1/2, then 𝒵\mathcal{Z} is a Gaussian;

  • ii)

    if α>1/2\alpha>1/2, then 𝒵\mathcal{Z} is a stable law with index 1/α1/\alpha.

These limit laws assuming a compact parameter range are proved using the Nagaev-Guivarc’h approach as outlined in [9] and [12]. The spectral properties of the transfer operator of FYF_{Y} will play a crucial role to this end. The case α=12\alpha=\frac{1}{2} is not addressed in this work, and is likely to be delicate since even for the deterministic map fω=12f_{\omega=\frac{1}{2}} the analysis in [9] (Section 1.3) derives a Gaussian limit under normalization BnnlognB_{n}^{\prime}\sim\sqrt{n\log n} whereas for ω>1/2\omega>1/2 the limit is a stable law similar to the above. One can therefore expect that when α=12\alpha=\frac{1}{2}, precise properties of the distribution of ν\nu near 12\frac{1}{2} will be needed to derive limit theorems in the random case, in contrast to the simpler results stated above.

For φ(0)>0\varphi(0)>0 our argument exploits the approximationFor a more precise statement, see the proof of Theorem 2.3 in Section 5. φYτYφ(0)\varphi_{Y}\approx\tau_{Y}\cdot\varphi(0) in distribution. A similar argument holds when φ(0)<0\varphi(0)<0. When φ(0)=0\varphi(0)=0 the estimate on φY\varphi_{Y} is more delicate, even in the deterministic case, so we do not consider that case in the current work.

Finally, we do not consider limit theorems in the case of unbounded parameter range and heavy tails, such as να,ε\nu_{\alpha,\varepsilon}. Although there may be path to obtain this via the Nagaev-Guivarc’h approach, the details will require a complete reworking of the bounded range case. We suggest this for possible investigation in the future.

3 Bound on the distortion

Definition 3.1.

Suppose f:[0,1][0,1]f:[0,1]\rightarrow[0,1] is a piecewise differentiable map with f>0f^{\prime}>0 on any J[0,1]J\subset[0,1] such that f|Jf|_{J} is differentiable. The distortion of ff on JJ is defined as

Dist(f|J)=supx,yJlog|Dxf||Dyf|.\operatorname{Dist}(f|J)=\sup_{x,y\in J}\log\frac{|D_{x}f|}{|D_{y}f|}.

We restrict our attention to J[0,1/2)J\subset[0,1/2) and for the moment write (abusing notation),

fω1=(fω|[0,1/2))1.f_{\omega}^{-1}=(f_{\omega}|_{[0,1/2)})^{-1}.
Proposition 3.2.

There is K>0K>0 such that for any ωΩ=[α,β]0\omega\in\Omega=[\alpha,\beta]^{\mathbb{N}_{0}}, I[1/2,1]I^{\prime}\subset[1/2,1], and nn\in\mathbb{N}

Dist(f𝝎n|(f𝝎n)1(I))KlogsupIinfI.\operatorname{Dist}(f^{n}_{\boldsymbol{\omega}}|(f_{\boldsymbol{\omega}}^{n})^{-1}(I^{\prime}))\leq K\log\frac{\sup I^{\prime}}{\inf I^{\prime}}.

As an immediate corollary to the previous proposition we obtain:

Corollary 3.3.

There is K>0K^{\prime}>0 such that for any 𝛚Ω\boldsymbol{\omega}\in\Omega and interval [x,y][0,1)[x,y]\subset[0,1) with f𝛚nf_{\boldsymbol{\omega}}^{n}, mapping [x,y][x,y] bijectively to f𝛚n([x,y])[1/2,1)f_{\boldsymbol{\omega}}^{n}([x,y])\subset[1/2,1)

Dist(f𝝎n|[x,y])K|f𝝎n(x)f𝝎n(y)|.\operatorname{Dist}(f^{n}_{\boldsymbol{\omega}}|[x,y])\leq K^{\prime}|f_{\boldsymbol{\omega}}^{n}(x)-f_{\boldsymbol{\omega}}^{n}(y)|.
Proof.

There is kk\in\mathbb{N} and numbers 0=n0<n1<n2<<nk=n0=n_{0}<n_{1}<n_{2}<...<n_{k}=n where nin_{i} is the ii-th return of all points in [x,y][x,y] to the set [1/2,1][1/2,1]. From Proposition 3.2, we get that for every 0ik10\leq i\leq k-1

Dist(fσni1𝝎nini1|f𝝎ni1([x,y]))K′′|f𝝎ni(x)f𝝎ni(y)|.\operatorname{Dist}(f^{n_{i}-n_{i-1}}_{\sigma^{n_{i-1}}\boldsymbol{\omega}}|f^{n_{i-1}}_{\boldsymbol{\omega}}([x,y]))\leq K^{\prime\prime}|f^{n_{i}}_{\boldsymbol{\omega}}(x)-f^{n_{i}}_{\boldsymbol{\omega}}(y)|.

Since |f𝝎ni(x)f𝝎ni(y)|2(ki)|f𝝎n(x)f𝝎n(y)||f^{n_{i}}_{\boldsymbol{\omega}}(x)-f^{n_{i}}_{\boldsymbol{\omega}}(y)|\leq 2^{-(k-i)}|f^{n}_{\boldsymbol{\omega}}(x)-f^{n}_{\boldsymbol{\omega}}(y)| one gets

Dist(f𝝎n|[x,y])\displaystyle\operatorname{Dist}(f^{n}_{\boldsymbol{\omega}}|[x,y]) i=1kDist(fσni1𝝎nini1|f𝝎ni1([x,y]))\displaystyle\leq\sum_{i=1}^{k}\operatorname{Dist}(f^{n_{i}-n_{i-1}}_{\sigma^{n_{i-1}}\boldsymbol{\omega}}|f^{n_{i-1}}_{\boldsymbol{\omega}}([x,y]))
K′′i=1k2(ki)|f𝝎n(x)f𝝎n(y)|\displaystyle\leq K^{\prime\prime}\sum_{i=1}^{k}2^{-(k-i)}|f^{n}_{\boldsymbol{\omega}}(x)-f^{n}_{\boldsymbol{\omega}}(y)|
2K′′|f𝝎n(x)f𝝎n(y)|.\displaystyle\leq 2K^{\prime\prime}|f^{n}_{\boldsymbol{\omega}}(x)-f^{n}_{\boldsymbol{\omega}}(y)|.

To prove the proposition we use the following lemma.

Lemma 3.4.

For any closed interval J(0,1]J\subset(0,1] define

𝒟(J)=(1+β)logsupJinfJ.\mathcal{D}(J)=(1+\beta)\log\frac{\sup J}{\inf J}.

Then,

maxω[α,β]{𝒟(fω1(J))+Dist(fω|fω1(J))}𝒟(J).\max_{\omega\in[\alpha,\beta]}\left\{\mathcal{D}(f_{\omega}^{-1}(J))+\operatorname{Dist}(f_{\omega}|f_{\omega}^{-1}(J))\right\}\leq\mathcal{D}(J). (3)
Proof.

Fix J=[a,b]J=[a,b]. For ω[α,β]\omega\in[\alpha,\beta]

𝒟(fω1(J))+Dist(fω|fω1(J))\displaystyle\mathcal{D}(f_{\omega}^{-1}(J))+\operatorname{Dist}(f_{\omega}|f_{\omega}^{-1}(J)) =(1+β)logfω1(b)fω1(a)+logDfω1(b)fωDfω1(a)fω\displaystyle=(1+\beta)\log\frac{f_{\omega}^{-1}(b)}{f_{\omega}^{-1}(a)}+\log\frac{D_{f_{\omega}^{-1}(b)}f_{\omega}}{D_{f_{\omega}^{-1}(a)}f_{\omega}}

where in the estimates above we used the fact that (fω)(f_{\omega})^{\prime} is positive and monotonically increasing. Calling a1:=fω1(a)a_{-1}:=f_{\omega}^{-1}(a) and b1:=fω1(b)b_{-1}:=f_{\omega}^{-1}(b) all we need to show is that

(1+β)logb1a1+logDb1fωDa1fω(1+β)logfω(b1)fω(a1).(1+\beta)\log\frac{b_{-1}}{a_{-1}}+\log\frac{D_{b_{-1}}f_{\omega}}{D_{a_{-1}}f_{\omega}}\leq(1+\beta)\log\frac{f_{\omega}(b_{-1})}{f_{\omega}(a_{-1})}. (4)

The left-hand side equals

LHS=log((b1a1)(1+β)1+(ω+1)2ωb1ω1+(ω+1)2ωa1ω)\displaystyle\mbox{LHS}=\log\left(\left(\frac{b_{-1}}{a_{-1}}\right)^{(1+\beta)}\frac{1+(\omega+1)2^{\omega}b_{-1}^{\omega}}{1+(\omega+1)2^{\omega}a_{-1}^{\omega}}\right)

and the right-hand side equals

RHS=log((b1a1)(1+β)(1+2ωb1ω1+2ωa1ω)(1+β)).\displaystyle\mbox{RHS}=\log\left(\left(\frac{b_{-1}}{a_{-1}}\right)^{(1+\beta)}\left(\frac{1+2^{\omega}b_{-1}^{\omega}}{1+2^{\omega}a_{-1}^{\omega}}\right)^{(1+\beta)}\right).
LHS-RHS =log(1+(ω+1)2ωb1ω1+(ω+1)2ωa1ω(1+2ωa1ω1+2ωb1ω)(1+β))\displaystyle=\log\left(\frac{1+(\omega+1)2^{\omega}b_{-1}^{\omega}}{1+(\omega+1)2^{\omega}a_{-1}^{\omega}}\left(\frac{1+2^{\omega}a_{-1}^{\omega}}{1+2^{\omega}b_{-1}^{\omega}}\right)^{(1+\beta)}\right)
=log((1+2ωa1ω)(1+β)(1+(ω+1)2ωa1ω)(1+(ω+1)2ωb1ω)(1+2ωb1ω)(1+β))\displaystyle=\log\left(\frac{(1+2^{\omega}a_{-1}^{\omega})^{(1+\beta)}}{(1+(\omega+1)2^{\omega}a_{-1}^{\omega})}\frac{(1+(\omega+1)2^{\omega}b_{-1}^{\omega})}{(1+2^{\omega}b_{-1}^{\omega})^{(1+\beta)}}\right)

The fact that LHS-RHS0\leq 0 follows from the fact that

g(x)=(1+x)(1+β)1+(ω+1)x.g(x)=\frac{(1+x)^{(1+\beta)}}{1+(\omega+1)x}.

is non-decreasing as an elementary derivative calculation shows, and the fact that LHS-RHS=logg(2ωa1ω)g(2ωb1ω)=\log\frac{g(2^{\omega}a_{-1}^{\omega})}{g(2^{\omega}b_{-1}^{\omega})} with b1a1b_{-1}\geq a_{-1}. This concludes the proof of (4). ∎

Proof of Proposition 3.2.

Fix 𝝎Ω\boldsymbol{\omega}\in\Omega and nn\in\mathbb{N} and let II^{\prime} be a sub-interval of [12,1][\frac{1}{2},1]. Define J0=fω01fωn11(I)J_{0}=f_{\omega_{0}}^{-1}\circ\ldots\circ f_{\omega_{n-1}}^{-1}(I^{\prime}) and for every 1in11\leq i\leq n-1 and

Ji:=f𝝎i(J0)=fωi1fωn11(I).J_{i}:=f_{\boldsymbol{\omega}}^{i}(J_{0})=f_{\omega_{i}}^{-1}\circ...\circ f_{\omega_{n-1}}^{-1}(I^{\prime}).

From the definition of distortion and monotonicity of fωf_{\omega} and (fω)(f_{\omega})^{\prime}

Dist(f𝝎n|J0)=i=0n1Dist(fωi|Ji).\operatorname{Dist}(f_{\boldsymbol{\omega}}^{n}|J_{0})=\sum_{i=0}^{n-1}\operatorname{Dist}(f_{\omega_{i}}|J_{i}).

Repeatedly applying (3)

𝒟(I)\displaystyle\mathcal{D}(I^{\prime}) 𝒟(Jn1)+Dist(fωn1|Jn1)𝒟(J0)+i=0n1Dist(fωi|Ji)\displaystyle\geq\mathcal{D}(J_{n-1})+\operatorname{Dist}(f_{\omega_{n-1}}|J_{n-1})\geq\mathcal{D}(J_{0})+\sum_{i=0}^{n-1}\operatorname{Dist}(f_{\omega_{i}}|J_{i})

which implies that

i=0n1Dist(fωi|Ji)𝒟(I)=(1+β)logsupIinfI.\sum_{i=0}^{n-1}\operatorname{Dist}(f_{\omega_{i}}|J_{i})\leq\mathcal{D}(I^{\prime})=(1+\beta)\log\frac{\sup I^{\prime}}{\inf I^{\prime}}.

4 Estimates on the Return Times to the Inducing Set

Let Y=Ω×[1/2,1]Y=\Omega\times[1/2,1] and Y\mathbb{P}_{Y} the probability measure obtained by restricting and then normalizing \mathbb{P} to YY. We use the following condition on the tails of the return time to YY in the renewal theory approach

γ>1 such that Y({(𝝎,x)Y:τY(𝝎,x)>n})=𝒪(nγ).\exists\gamma>1\mbox{ such that }\mathbb{P}_{Y}\left(\{(\boldsymbol{\omega},x)\in Y:\;\tau_{Y}(\boldsymbol{\omega},x)>n\}\right)=\mathcal{O}(n^{-\gamma}). (C1)

In the diffusion driven case we are going to use the following, closely related condition: there is a non-decreasing function R:0+R:\mathbb{N}\rightarrow\mathbb{R}^{+}_{0} such that

n=1Y({(𝝎,x)Y:τY(𝝎,x)=n})R(n)<.\sum_{n=1}^{\infty}\mathbb{P}_{Y}\left(\{(\boldsymbol{\omega},x)\in Y:\;\tau_{Y}(\boldsymbol{\omega},x)=n\}\right)\cdot R(n)<\infty. (C2)

In fact, for t>1t>1, if condition (C1) holds for all γ<t\gamma<t then condition (C2) holds for R(n)=nγR(n)=n^{\gamma} for all γ<t\gamma<t. To see this, we use summation by parts, showing

n=1(τ=n)R(n)\displaystyle\sum_{n=1}^{\infty}\mathbb{P}(\tau=n)R(n) =n=1((τ>n1)(τ>n))R(n)\displaystyle=\sum_{n=1}^{\infty}(\mathbb{P}(\tau>n-1)-\mathbb{P}(\tau>n))R(n)
=R(1)+n=1(τ>n)(R(n+1)R(n))\displaystyle=R(1)+\sum_{n=1}^{\infty}\mathbb{P}(\tau>n)(R(n+1)-R(n))

Notice that, having fixed the family {fω}ω[α,β]\{f_{\omega}\}_{\omega\in[\alpha,\beta]}, (C1) and (C2) are conditions on the measure ν\nu. Sharp bounds for the expressions in (C1) and (C2) have been obtained for various ν\nu in [6]. Below we prove the following proposition

Proposition 4.1.

Assume that ν\nu is a probability measure supported on [α,β][\alpha,\beta] with 0<α<10<\alpha<1 lying in the topological support of ν\nu. Then conditions (C1) and (C2) hold. In particular

  1. 1.

    For any 1<γ<1α1<\gamma<\frac{1}{\alpha}, Condition (C1) holds and Condition (C2) holds with R(n)=nγR(n)=n^{\gamma};

  2. 2.

    if ν({α})>0\nu(\{\alpha\})>0, then Condition (C1) holds with γ=1α\gamma=\frac{1}{\alpha} and Condition (C2) holds with R(n)=n1/αR(n)=n^{1/\alpha}.

Conditions (C1) and (C2) concern the return times of orbits to the inducing set under the map FF. They can be verified on a case-by-case basis in the following way (see also [6]). Call fω,1:=fω|[0,1/2)f_{\omega,1}:=f_{\omega}|_{[0,1/2)} the first invertible branch of fωf_{\omega}, and g:=2x1g:=2x-1 mod 1 on [1/2,1][1/2,1] the second branch common to all the maps in the family. Define xn(𝝎):=fω11fωn11(1/2)x_{n}(\boldsymbol{\omega}):=f^{-1}_{\omega_{1}}...f^{-1}_{\omega_{n-1}}(1/2). Then {(𝝎,x)Y:τY(𝝎,x)>n}={(𝝎,y):y[1/2,g1(xn(𝝎))]}.\{(\boldsymbol{\omega},x)\in Y:\;\tau_{Y}(\boldsymbol{\omega},x)>n\}=\{(\boldsymbol{\omega},y):y\in[1/2,g^{-1}(x_{n}(\boldsymbol{\omega}))]\}.

This implies that

Y((𝝎,x)Y:τY(𝝎,x)>n)=12𝔼[xn(𝝎)],\mathbb{P}_{Y}((\boldsymbol{\omega},x)\in Y\colon\tau_{Y}(\boldsymbol{\omega},x)>n)=\tfrac{1}{2}\mathbb{E}[x_{n}(\boldsymbol{\omega})], (5)

where the expectation 𝔼\mathbb{E} is with respect to ν\nu^{\mathbb{N}}. The equality above together with computations as in [4] allow us to prove Proposition 4.1.

Proof of Proposition 4.1.

From the assumptions, α:=minsupp(ν)>0\alpha:=\min\mbox{supp}(\nu)>0. Let 1<γ<1α1<\gamma<\frac{1}{\alpha}, so that from the definition of topological support, p0:=ν([α,γ1])>0p_{0}:=\nu([\alpha,\gamma^{-1}])>0. It is known that if 𝝎\boldsymbol{\omega} is a sequence such that j=1n𝟏{ωj[α,γ1]}>M\sum_{j=1}^{n}\boldsymbol{1}_{\{\omega_{j}\in[\alpha,\gamma^{-1}]\}}>M, then xn(𝝎)Mγx_{n}(\boldsymbol{\omega})\leq\lfloor M\rfloor^{-\gamma} (see for example [4]) and is a consequence of the monotonicity found in the family of functions {fω}ω+\{f_{\omega}\}_{\omega\in\mathbb{R}^{+}}. It follows from the Hoeffding concentration inequality for i.i.d. and bounded random variables that

ν(𝝎:1ni=1n𝟏{ωi[α,γ1]}<p0ε)e2nε2\nu^{\mathbb{N}}\left(\boldsymbol{\omega}:\;\frac{1}{n}\sum_{i=1}^{n}\boldsymbol{1}_{\{\omega_{i}\in[\alpha,\gamma^{-1}]\}}<p_{0}-\varepsilon\right)\leq e^{-2n\varepsilon^{2}}

so that calling

An:={𝝎:i=1n𝟏{ωi[α,γ1)}(𝝎)<n(p0ε)},A_{n}:=\left\{\boldsymbol{\omega}:\,\sum_{i=1}^{n}\boldsymbol{1}_{\{\omega_{i}\in[\alpha,\gamma^{-1})\}}(\boldsymbol{\omega})<n(p_{0}-\varepsilon)\right\}, (6)

ν(An)e2nε2\nu^{\mathbb{N}}(A_{n})\leq e^{-2n\varepsilon^{2}}. Now, using (5):

Y({(𝝎,x):τY(𝝎,x)>n})\displaystyle\mathbb{P}_{Y}(\{(\boldsymbol{\omega},x)\colon\tau_{Y}(\boldsymbol{\omega},x)>n\}) =12𝑑ν(𝝎)xn(𝝎)\displaystyle=\frac{1}{2}\int d\nu^{\mathbb{N}}(\boldsymbol{\omega})x_{n}(\boldsymbol{\omega})
=12[An𝑑ν(𝝎)+Anc𝑑ν(𝝎)]xn(𝝎)\displaystyle=\frac{1}{2}\left[\int_{A_{n}}d\nu^{\mathbb{N}}(\boldsymbol{\omega})+\int_{A_{n}^{c}}d\nu^{\mathbb{N}}(\boldsymbol{\omega})\right]x_{n}(\boldsymbol{\omega})
12(e2nε2+(n(p0ε))γ)=𝒪(nγ).\displaystyle\leq\tfrac{1}{2}(e^{-2n\varepsilon^{2}}+(n(p_{0}-\varepsilon))^{-\gamma})=\mathcal{O}(n^{-\gamma}).

establishing Condition (C1) for any γ<1α\gamma<\frac{1}{\alpha}. As pointed out above, this implies that Condition (C2) holds with R(n)=nγR(n)=n^{\gamma} for any γ<1α\gamma<\frac{1}{\alpha}.

If ν({α})>0\nu(\{\alpha\})>0, then one can repeat the above reasoning with γ=1α\gamma=\frac{1}{\alpha} to obtain

Y({(𝝎,x):τY(𝝎,x)>n})𝒪(n1α).\mathbb{P}_{Y}(\{(\boldsymbol{\omega},x)\colon\tau_{Y}(\boldsymbol{\omega},x)>n\})\leq\mathcal{O}(n^{-\frac{1}{\alpha}}).

5 Renewal Theory Approach

In this section we prove Theorem 2.1 on correlation decay, and Theorem 2.3 on convergence to stable laws. In Section 5.1 we treat the correlation decay using the renewal theory for transfer operators, as introduced by Sarig [21] and further developed by Gouëzel in [10], while in Section 5.2 we use the Nagaev-Guivarc’h approach to prove convergence to limit laws.

5.1 Decay of Correlations via the Renewal Theory for Transfer Operators

We apply the following general theorem that can be found in [10]. Let us denote the open unit disk by 𝔻={z:|z|<1}\mathbb{D}=\{z\in\mathbb{C}:\;|z|<1\}.

Theorem 5.1 (Theorem 1.1 [10]).

Let {Tn}n0\{T_{n}\}_{n\geq 0} be bounded operators on a Banach space (,)(\mathcal{B},\|\cdot\|) such that T(z)=Id+n1znTnT(z)=\operatorname{Id}+\sum_{n\geq 1}z^{n}T_{n} converges for every z𝔻z\in\mathbb{D}. Assume that:

  • for every z𝔻z\in\mathbb{D}, T(z)=(IdR(z))1T(z)=(\operatorname{Id}-R(z))^{-1} where R(z)=n1znRnR(z)=\sum_{n\geq 1}z^{n}R_{n} and {Rn}n1\{R_{n}\}_{n\geq 1} are bounded operators on \mathcal{B} such that n1Rn<+\sum_{n\geq 1}\|R_{n}\|<+\infty;

  • 1 is a simple isolated eigenvalue of R(1)R(1);

  • for every z𝔻¯\{1}z\in\overline{\mathbb{D}}\backslash\{1\}, IdR(z)\operatorname{Id}-R(z) is invertible.

Let Π\Pi be the eigenprojection of R(1)R(1) at 1. If knRk=𝒪(1/nγ)\sum_{k\geq n}\|R_{k}\|=\mathcal{O}(1/n^{\gamma}) for some γ>1\gamma>1 and ΠR(1)Π0\Pi R^{\prime}(1)\Pi\neq 0, then for all nn

Tn=1μΠ+1μ2k=n+1+Πk+EnT_{n}=\frac{1}{\mu}\Pi+\frac{1}{\mu^{2}}\sum_{k=n+1}^{+\infty}\Pi_{k}+E_{n}

where μ\mu is given by ΠR(1)Π=μΠ\Pi R^{\prime}(1)\Pi=\mu\Pi, Πn=l>nΠRlΠ\Pi_{n}=\sum_{l>n}\Pi R_{l}\Pi and EnE_{n} is a bounded operator satisfying

En={𝒪(1/nγ)if γ>2𝒪(logn/n2)if γ=2𝒪(1/n2γ2)if 2>γ>1\|E_{n}\|=\left\{\begin{array}[]{ll}\mathcal{O}(1/n^{\gamma})&\mbox{if }\gamma>2\\ \mathcal{O}(\log n/n^{2})&\mbox{if }\gamma=2\\ \mathcal{O}(1/n^{2\gamma-2})&\mbox{if }2>\gamma>1\end{array}\right.

To apply the above result to our context, consider FY:=FτY:YYF_{Y}:=F^{\tau_{Y}}:Y\rightarrow Y, and PY:L1(Y,Y)L1(Y,Y)P_{Y}:L^{1}(Y,\mathbb{P}_{Y})\rightarrow L^{1}(Y,\mathbb{P}_{Y}) its transfer operator. Define operators Tn,Rn:L1(Y,Y)L1(Y,Y)T_{n},R_{n}:L^{1}(Y,\mathbb{P}_{Y})\rightarrow L^{1}(Y,\mathbb{P}_{Y}) for every n0n\in\mathbb{N}_{0}, in the following way

T0=Id,R0=0,Tnφ=χYPFn(χYφ),Rnφ=PFn(φχ{τY=n})T_{0}=\operatorname{Id},\quad R_{0}=0,\quad T_{n}\varphi=\chi_{Y}P_{F}^{n}(\chi_{Y}\varphi),\quad R_{n}\varphi=P_{F}^{n}(\varphi\chi_{\{\tau_{Y}=n\}})

for nn\in\mathbb{N}, where PF:L1(X,)L1(X,)P_{F}:L^{1}(X,\mathbb{P})\rightarrow L^{1}(X,\mathbb{P}) denotes the transfer operator for the skew product FF.

Corollary 5.2.

Assume there is a Banach space (,)(\mathcal{B},\|\cdot\|), L1(Y,Y)\mathcal{B}\subset L^{1}(Y,\mathbb{P}_{Y}), such that the operators {Tn}n0\{T_{n}\}_{n\geq 0} and {Rn}n1\{R_{n}\}_{n\geq 1} satisfy the hypotheses of Theorem 5.1. Then for any h1h_{1}\in\mathcal{B} supported on YY, and any h2L(Y)h_{2}\in L^{\infty}(Y),

|h1h2Fn𝑑Yh1𝑑Yh2𝑑π|𝒪(n1γ).\left|\int h_{1}h_{2}\circ F^{n}d\mathbb{P}_{Y}-\int h_{1}d\mathbb{P}_{Y}\int h_{2}d\pi\right|\leq\mathcal{O}(n^{1-\gamma}).

where π\pi is the stationary measure for FF on XX.

Let us say a few words about how this result follows from the previous theorem, given the definitions above. We will see (in the proof of Proposition 5.3 below) that PY=R(1)P_{Y}=R(1), μ=1/π[12,1]\mu=1/\pi[\frac{1}{2},1] and dπY=ψ0dYd\pi_{Y}=\psi_{0}d\mathbb{P}_{Y}, with Πh=(h𝑑Y)ψ0\Pi h=(\int hd\mathbb{P}_{Y})\psi_{0} and the relation πYπ[12,1]=π|[12,1]\pi_{Y}\cdot\pi[\frac{1}{2},1]=\pi|_{[\frac{1}{2},1]}. We can then write

h1h2Fn𝑑Y\displaystyle\int h_{1}h_{2}\circ F^{n}d\mathbb{P}_{Y} =2χYh1h2Fn𝑑\displaystyle=2\int\chi_{Y}h_{1}h_{2}\circ F^{n}d\mathbb{P}
=2PFn(χYh1)h2𝑑\displaystyle=2\int P_{F}^{n}(\chi_{Y}h_{1})h_{2}d\mathbb{P}
=2χYPFn(χYh1)h2𝑑\displaystyle=2\int\chi_{Y}P_{F}^{n}(\chi_{Y}h_{1})h_{2}d\mathbb{P}
=Tn(h1)h2𝑑Y\displaystyle=\int T_{n}(h_{1})h_{2}d\mathbb{P}_{Y}

We now apply the expansion of TnT_{n} to obtain

h1h2Fn𝑑Y=π[12,1](h1𝑑Y)ψ0h2dY+H.O.T=h1𝑑Yh2𝑑π+H.O.T,\int h_{1}h_{2}\circ F^{n}d\mathbb{P}_{Y}=\pi[\frac{1}{2},1]\left(\int h_{1}d\mathbb{P}_{Y}\right)\psi_{0}h_{2}d\mathbb{P}_{Y}+\textnormal{H.O.T}=\int h_{1}d\mathbb{P}_{Y}\int h_{2}d\pi+H.O.T,

where the higher order terms arise from the second and third terms in the expansion of TnT_{n} and both decay with a rate upper bounded by 𝒪(n1γ)\mathcal{O}(n^{1-\gamma}) for all three ranges of γ\gamma in Theorem 5.1).

Corollary 5.2 now gives Theorem 2.1 for the restricted case of ψ\psi and φ\varphi supported on [12,1][\frac{1}{2},1] since dY=2|[12,1]d\mathbb{P}_{Y}=2\mathbb{P}|_{[\frac{1}{2},1]} in both integrals. We can extend to the case where φ\varphi is supported on [0,1][0,1] as follows: Set h1(ω,x)=φ(2x1)h_{1}(\omega,x)=\varphi(2x-1) for x[12,1]x\in[\frac{1}{2},1], h1(ω,x)=0h_{1}(\omega,x)=0 for x[0,12]x\in[0,\frac{1}{2}] and h2(ω,x)=ψ(x)h_{2}(\omega,x)=\psi(x). Observe that PFh1(ω,x)=12φ(x)P_{F}h_{1}(\omega,x)=\frac{1}{2}\varphi(x) so h1𝑑Y=2h1𝑑=φ𝑑\int h_{1}\,d\mathbb{P}_{Y}=2\int\,h_{1}d\mathbb{P}=\int\varphi d\mathbb{P} and the correlation integral above becomes

h1h2Fn𝑑Y=φψFn1𝑑,\int h_{1}h_{2}\circ F^{n}d\mathbb{P}_{Y}=\int\varphi\psi\circ F^{n-1}d\mathbb{P},

leading to the result stated in Theorem 2.1 provided ψ\psi is supported on [12,1][\frac{1}{2},1]. The final extension to fully supported ψ\psi can be established using the method detailed in Gouëzel [10], Theorem 6.9.

The proposition below shows that the hypotheses of Theorem 5.1 and Corollary 5.2 are satisfied in our setup by the Banach space \mathcal{B} of functions on YY that are a) constant w.r.t. ω\omega and b) Lipschitz w.r.t. the spatial variable xx. To simplify notation, we indicate these functions in terms of xx only and write

:={φ:[1/2,1]:|φ|Lip<}\mathcal{B}:=\{\varphi:[1/2,1]\rightarrow\mathbb{C}:\;|\varphi|_{\mathrm{Lip}}<\infty\}

where

|φ|Lip=supx,y[1/2,1]xy|φ(x)φ(y)||xy||\varphi|_{\mathrm{Lip}}=\sup_{\begin{subarray}{c}x,y\in[1/2,1]\\ x\neq y\end{subarray}}\frac{|\varphi(x)-\varphi(y)|}{|x-y|}

is the Lipschitz semi-norm, and \mathcal{B} is endowed with the norm φ:=|φ|Lip+|φ|\|\varphi\|:=|\varphi|_{\mathrm{Lip}}+|\varphi|_{\infty}. Notice that if φ\varphi\in\mathcal{B} the function PFφP_{F}\varphi is also constant w.r.t. ω\omega. In particular, we can compute the skew product transfer operator as

PFφ(x,ω)=PFφ(x)=[α,β]Pγφ(x)𝑑ν(γ),P_{F}\varphi(x,\omega)=P_{F}\varphi(x)=\int_{[\alpha,\beta]}P_{\gamma}\varphi(x)d\nu(\gamma),

where PγP_{\gamma} denotes the transfer operator associated to fγf_{\gamma} on [0,1][0,1] with respect to mm. We will see this leads to a useful simplification when computing higher powers PFkP_{F}^{k} and the induced transfer operator below.

Proposition 5.3.

Let \mathcal{B} be as above. The maps TnT_{n} and RnR_{n} are bounded operators on \mathcal{B}. Suppose that condition (C1) is satisfied for some γ>1\gamma>1. The series T(z)=Id+n1znTnT(z)=\operatorname{Id}+\sum_{n\geq 1}z^{n}T_{n} converges on 𝔻\mathbb{D}, and:

  • (i)

    IdR(z)\operatorname{Id}-R(z) is invertible for every z𝔻¯\{1}z\in\overline{\mathbb{D}}\backslash\{1\};

  • (ii)

    T(z)=(IdR(z))1T(z)=(\operatorname{Id}-R(z))^{-1} for z𝔻z\in\mathbb{D}, where R(z)=n1znRnR(z)=\sum_{n\geq 1}z^{n}R_{n} and n1Rn<\sum_{n\geq 1}\|R_{n}\|<\infty;

  • (iii)

    R(1)R(1) has a spectral gap, i.e. there is Π\Pi with Π2=Π\Pi^{2}=\Pi, dimΠ=1\dim\Im\Pi=1, and there is NN satisfying ΠN=NΠ=0\Pi N=N\Pi=0, σ(N)<1\sigma(N)<1, such that R(1)=Π+NR(1)=\Pi+N; The non-degeneracy condition ΠR(1)Π0\Pi R^{\prime}(1)\Pi\neq 0 holds.

  • (iv)

    k>nRk=𝒪(nγ)\sum_{k>n}\|R_{k}\|=\mathcal{O}(n^{-\gamma}).

Proof of Proposition 5.3.

Fix any kk\in\mathbb{N}, a nkn\geq k, and any sequence (ω0ωn1)(\omega_{0}\ldots\omega_{n-1}) and write fnf^{n} for f(ω0ωn1)nf^{n}_{(\omega_{0}\ldots\omega_{n-1})}. Define 𝒥(ω0ωn1)k\mathcal{J}^{k}_{(\omega_{0}\ldots\omega_{n-1})} to be the collection of maximal subintervals of [12,1][\frac{1}{2},1] where fnf^{n} is continuous, returning to [12,1][\frac{1}{2},1] for the kkth time at time nn under fnf^{n}. If J𝒥(ω0ωn1)kJ\in\mathcal{J}^{k}_{(\omega_{0}\ldots\omega_{n-1})}, the points of JJ have the same return times up to the kkth return and JJ is mapped injectively and onto [12,1][\frac{1}{2},1] under fnf^{n}. Pick ψ\psi\in\mathcal{B}, then

PYkψ=[α,β]0𝑑ν0(ω0ωn1)nkJPωn1Pω0(ψχJ)(x)P^{k}_{Y}\psi=\int_{[\alpha,\beta]^{\mathbb{N}_{0}}}d\nu^{\mathbb{N}_{0}}(\omega_{0}\ldots\omega_{n-1}\ldots)\sum_{n\geq k}\sum_{J}P_{\omega_{n-1}}\ldots P_{\omega_{0}}\left(\psi\chi_{J}\right)(x)

where the second sum is over J𝒥(ω0ωn1)kJ\in\mathcal{J}^{k}_{(\omega_{0}\ldots\omega_{n-1})} and again we indicate by PωiP_{\omega_{i}} the transfer operator of fωif_{\omega_{i}}. We are going to prove that PYP_{Y} satisfies a Lasota-Yorke inequality. First of all notice that

Pωn1Pω0(ψχJ)(x)=ψf(ω0ωn1),Jn(x)xf(ω0ωn1),Jn(x)P_{\omega_{n-1}}\circ\ldots\circ P_{\omega_{0}}(\psi\chi_{J})(x)=\psi\circ f_{(\omega_{0}\ldots\omega_{n-1}),J}^{-n}(x)\partial_{x}f_{(\omega_{0}\ldots\omega_{n-1}),J}^{-n}(x)

with f(ω0ωn1),Jnf_{(\omega_{0}\ldots\omega_{n-1}),J}^{n} indicating the restriction of the function to JJ. Given two points x,y[1/2,1]x,y\in[1/2,1]

|ψf(ω0ωn1),Jn(x)xf(ω0ωn1),Jn(x)ψf(ω0ωn1),Jn(y)xf(ω0ωn1),J,n(y)|\displaystyle\left|\psi\circ f_{(\omega_{0}\ldots\omega_{n-1}),J}^{-n}(x)\partial_{x}f_{(\omega_{0}\ldots\omega_{n-1}),J}^{-n}(x)-\psi\circ f_{(\omega_{0}\ldots\omega_{n-1}),J}^{-n}(y)\partial_{x}f_{(\omega_{0}\ldots\omega_{n-1}),J,}^{-n}(y)\right|\leq (7)
|xf(ω0ωn1),Jn(x)||ψf(ω0ωn1),Jn(x)ψf(ω0ωn1),Jn(y)|+\displaystyle\quad\quad\leq\left|\partial_{x}f_{(\omega_{0}\ldots\omega_{n-1}),J}^{-n}(x)\right|\left|\psi\circ f_{(\omega_{0}\ldots\omega_{n-1}),J}^{-n}(x)-\psi\circ f_{(\omega_{0}\ldots\omega_{n-1}),J}^{-n}(y)\right|+
+|ψf(ω0ωn1),Jn(y)||xf(ω0ωn1),Jn(x)xf(ω0ωn1),Jn(y)|=\displaystyle\quad\quad\quad+\left|\psi\circ f_{(\omega_{0}\ldots\omega_{n-1}),J}^{-n}(y)\right|\left|\partial_{x}f_{(\omega_{0}\ldots\omega_{n-1}),J}^{-n}(x)-\partial_{x}f_{(\omega_{0}\ldots\omega_{n-1}),J}^{-n}(y)\right|=
=:A+B.\displaystyle\quad\quad=:A+B.

To bound AA we show first that there is M>0M>0 uniform in the choice of JJ such that

|xf(ω0ωn1),Jn(x)|M|J|.|\partial_{x}f_{(\omega_{0}\ldots\omega_{n-1}),J}^{-n}(x)|\leq M|J|.

In fact, recall that JJ is mapped bijectively onto [12,1][\frac{1}{2},1]. By the mean value theorem, there is ξ[12,1]\xi\in[\frac{1}{2},1] such that

2|J|=xf(ω0ωn1),Jn(ξ).2|J|=\partial_{x}f_{(\omega_{0}\ldots\omega_{n-1}),J}^{-n}(\xi).

Combining this with the bound on distortion from Corollary 3.3, there is a constant KK such that

|xf(ω0ωn1),Jn(x)|\displaystyle|\partial_{x}f_{(\omega_{0}\ldots\omega_{n-1}),J}^{-n}(x)| =|xf(ω0ωn1),Jn(x)xf(ω0ωn1),Jn(ξ))||xf(ω0ωn1),Jn(ξ)|\displaystyle=\left|\frac{\partial_{x}f_{(\omega_{0}\ldots\omega_{n-1}),J}^{-n}(x)}{\partial_{x}f_{(\omega_{0}\ldots\omega_{n-1}),J}^{-n}(\xi))}\right|\left|\partial_{x}f_{(\omega_{0}\ldots\omega_{n-1}),J}^{-n}(\xi)\right|
=|xf(ω0ωn1),Jn(f(ω0ωn1),Jn(ξ))xf(ω0ωn1),Jn(f(ω0ωn1),Jn(x))||xf(ω0ωn1),Jn(ξ)|\displaystyle=\left|\frac{\partial_{x}f_{(\omega_{0}\ldots\omega_{n-1}),J}^{n}(f_{(\omega_{0}\ldots\omega_{n-1}),J}^{-n}(\xi))}{\partial_{x}f_{(\omega_{0}\ldots\omega_{n-1}),J}^{n}(f_{(\omega_{0}\ldots\omega_{n-1}),J}^{-n}(x))}\right||\partial_{x}f_{(\omega_{0}\ldots\omega_{n-1}),J}^{-n}(\xi)|
K|J|\displaystyle\leq K|J|

This and the fact that

|ψf(ω0ωn1),Jn(x)ψf(ω0ωn1),Jn(y)|2k|ψ|Lip|xy||\psi\circ f^{-n}_{(\omega_{0}\ldots\omega_{n-1}),J}(x)-\psi\circ f^{-n}_{(\omega_{0}\ldots\omega_{n-1}),J}(y)|\leq 2^{-k}|\psi|_{\mathrm{Lip}}|x-y|

imply that

AM|J|2k|ψ|Lip|xy|.A\leq M|J|2^{-k}|\psi|_{\mathrm{Lip}}|x-y|.

To bound BB first of all notice that

|ψf(ω0ωn1),Jn(y)|\displaystyle\left|\psi\circ f_{(\omega_{0}\ldots\omega_{n-1}),J}^{-n}(y)\right| |J|1J|ψ|+2k|ψ|Lip\displaystyle\leq|J|^{-1}\int_{J}|\psi|+2^{-k}|\psi|_{\mathrm{Lip}}

where we used that the diameter of JJ is less then 2k2^{-k} given that its points have kk common return times with respect to sequences from the cylinder [ω0ωn1][\omega_{0}\ldots\omega_{n-1}]. Then notice that

|xf(ω0ωn1),Jn(x)xf(ω0ωn1),Jn(y)|\displaystyle\left|\partial_{x}f_{(\omega_{0}\ldots\omega_{n-1}),J}^{-n}(x)-\partial_{x}f_{(\omega_{0}\ldots\omega_{n-1}),J}^{-n}(y)\right|
=|xf(ω0ωn1),Jn(x)||1xf(ω0ωn1),Jn(y)xf(ω0ωn1),Jn(x)|\displaystyle=|\partial_{x}f_{(\omega_{0}\ldots\omega_{n-1}),J}^{-n}(x)|\left|1-\frac{\partial_{x}f_{(\omega_{0}\ldots\omega_{n-1}),J}^{-n}(y)}{\partial_{x}f_{(\omega_{0}\ldots\omega_{n-1}),J}^{-n}(x)}\right|
=|xf(ω0ωn1),Jn(x)||1xf(ω0ωn1),Jn(f(ω0ωn1),Jn(x))xf(ω0ωn1),Jn(f(ω0ωn1),Jn(y))|\displaystyle=|\partial_{x}f_{(\omega_{0}\ldots\omega_{n-1}),J}^{-n}(x)|\left|1-\frac{\partial_{x}f_{(\omega_{0}\ldots\omega_{n-1}),J}^{n}(f_{(\omega_{0}\ldots\omega_{n-1}),J}^{-n}(x))}{\partial_{x}f_{(\omega_{0}\ldots\omega_{n-1}),J}^{n}(f_{(\omega_{0}\ldots\omega_{n-1}),J}^{-n}(y))}\right|
M|J||xy|,\displaystyle\leq M^{\prime}|J||x-y|,

where we made use of Proposition 3.2 in the last line. We conclude that

BM(J|ψ|𝑑m(x)+|J|2k|ψ|Lip)|xy|B\leq M^{\prime}\left(\int_{J}|\psi|dm(x)+|J|2^{-k}|\psi|_{\mathrm{Lip}}\right)|x-y|

Putting the bounds for AA and BB together, we conclude that there is M′′>0M^{\prime\prime}>0 (uniform) such that for every kk\in\mathbb{N}, nkn\geq k, (ω0ωn1)(\omega_{0}\ldots\omega_{n-1}), and J𝒥(ω0ωn1)kJ\in\mathcal{J}^{k}_{(\omega_{0}\ldots\omega_{n-1})}

|Pωn1Pω0(ψχJ)|LipM′′(J|ψ|𝑑m(x)+|J|2k|ψ|Lip).|P_{\omega_{n-1}}\circ\ldots\circ P_{\omega_{0}}(\psi\chi_{J})|_{\mathrm{Lip}}\leq M^{\prime\prime}\left(\int_{J}|\psi|dm(x)+|J|2^{-k}|\psi|_{\mathrm{Lip}}\right). (8)

Applying this to 𝒥ω0ωn11\mathcal{J}^{1}_{\omega_{0}\ldots\omega_{n-1}}, which consists of a single interval, Jω0,,ωn11J^{1}_{\omega_{0},\ldots,\omega_{n-1}} say, we see

|Pωn1Pω0(χJ)|LipM′′|Jω0ωn11|.|P_{\omega_{n-1}}\circ\ldots\circ P_{\omega_{0}}(\cdot\chi_{J})|_{\text{Lip}}\leq M^{\prime\prime}|J^{1}_{\omega_{0}\ldots\omega_{n-1}}|.

We note

Rn()=[α,β]nPωn1Pω0(χJω0ωn11)dνn,R_{n}(\cdot)=\int_{[\alpha,\beta]^{n}}P_{\omega_{n-1}}\circ\ldots\circ P_{\omega_{0}}(\cdot\chi_{J^{1}_{\omega_{0}\ldots\omega_{n-1}}})\,d\nu^{n},

so that

RnM′′[α,β]n|Jω0ωn11|𝑑νn.\|R_{n}\|\leq M^{\prime\prime}\int_{[\alpha,\beta]^{n}}|J^{1}_{\omega_{0}\ldots\omega_{n-1}}|\,d\nu^{n}.

Since for each (ω0,)[α,β]0(\omega_{0},\ldots)\in[\alpha,\beta]^{\mathbb{N}_{0}}, the sets Jω0ωn11J^{1}_{\omega_{0}\ldots\omega_{n-1}} form a countable partition of [12,1][\frac{1}{2},1], it follows that Rn<\sum\|R_{n}\|<\infty as required.

By the same calculation,

k>nRk\displaystyle\sum_{k>n}\|R_{k}\| M′′[α,β]0k>n|Jω0ωk11|dν0(𝝎)\displaystyle\leq M^{\prime\prime}\int_{[\alpha,\beta]^{\mathbb{N}_{0}}}\sum_{k>n}|J^{1}_{\omega_{0}\ldots\omega_{k-1}}|\,d\nu^{\mathbb{N}_{0}}(\boldsymbol{\omega})
=M′′Y({(𝝎,x)Y:τY(𝝎,x)>n})=𝒪(nγ)\displaystyle=M^{\prime\prime}\,\mathbb{P}_{Y}(\{(\boldsymbol{\omega},x)\in Y\colon\tau_{Y}(\boldsymbol{\omega},x)>n\})=\mathcal{O}(n^{-\gamma})

by condition (C1), establishing point (iv).

For every 𝝎=(ω0ω1)\boldsymbol{\omega}=(\omega_{0}\omega_{1}\ldots), let 𝒥𝝎k\mathcal{J}^{k}_{\boldsymbol{\omega}} be nk𝒥(ω0ωn1)k\bigcup_{n\geq k}\mathcal{J}^{k}_{(\omega_{0}\ldots\omega_{n-1})}, the countable partition of [12,1][\frac{1}{2},1] according to the first kk returns to [12,1][\frac{1}{2},1] under the maps fωnf_{\omega}^{n}. One obtains

|PYkψ|Lip[α,β]0𝑑ν0(ω0ωn1)nkJ|Pωn1Pω0(ψχJ)|LipM′′nk[α,β]n𝑑νn(ω0ωn1)J𝒥(ω0ωn1)k(2k|J||ψ|Lip+J|ψ|)=M′′[α,β]0𝑑ν0(𝝎)J𝒥𝝎k(2k|J||ψ|Lip+J|ψ|)=M′′2k|ψ|Lip[α,β]0𝑑ν0(𝝎)J𝒥𝝎k|J|+M′′[α,β]0𝑑ν0(𝝎)J𝒥𝝎kJ|ψ|=M′′2k|ψ|Lip12+M′′[1/2,1]|ψ|,\begin{split}|P^{k}_{Y}\psi|_{\mathrm{Lip}}&\leq\int_{[\alpha,\beta]^{\mathbb{N}_{0}}}d\nu^{\mathbb{N}_{0}}(\omega_{0}\ldots\omega_{n-1}\ldots)\sum_{n\geq k}\sum_{J}|P_{\omega_{n-1}}\ldots P_{\omega_{0}}\left(\psi\chi_{J}\right)|_{\mathrm{Lip}}\\ &\leq M^{\prime\prime}\sum_{n\geq k}\int_{[\alpha,\beta]^{n}}d\nu^{n}(\omega_{0}\ldots\omega_{n-1})\sum_{J\in\mathcal{J}^{k}_{(\omega_{0}\ldots\omega_{n-1})}}\left(2^{-k}|J||\psi|_{\mathrm{Lip}}+\int_{J}|\psi|\right)\\ &=M^{\prime\prime}\int_{[\alpha,\beta]^{\mathbb{N}_{0}}}d\nu^{\mathbb{N}_{0}}(\boldsymbol{\omega})\sum_{J\in\mathcal{J}^{k}_{\boldsymbol{\omega}}}\left(2^{-k}|J||\psi|_{\mathrm{Lip}}+\int_{J}|\psi|\right)\\ &=M^{\prime\prime}2^{-k}|\psi|_{\mathrm{Lip}}\int_{[\alpha,\beta]^{\mathbb{N}_{0}}}d\nu^{\mathbb{N}_{0}}(\boldsymbol{\omega})\sum_{J\in\mathcal{J}^{k}_{\boldsymbol{\omega}}}|J|+M^{\prime\prime}\int_{[\alpha,\beta]^{\mathbb{N}_{0}}}d\nu^{\mathbb{N}_{0}}(\boldsymbol{\omega})\sum_{J\in\mathcal{J}^{k}_{\boldsymbol{\omega}}}\int_{J}|\psi|\\ &=M^{\prime\prime}2^{-k}|\psi|_{\mathrm{Lip}}\,\frac{1}{2}+M^{\prime\prime}\int_{[1/2,1]}|\psi|,\end{split} (9)

where in the first line, the sum is over those intervals whose kkth return occurs at time nn. An elementary calculation shows

111/21/21PYkψ𝑑m|PYkψ|LipPYkψ(x)111/21/21PYkψ𝑑m+|PYkψ|Lip,\frac{1}{1-1/2}\int_{1/2}^{1}P^{k}_{Y}\psi\,dm-|P^{k}_{Y}\psi|_{\mathrm{Lip}}\leq P^{k}_{Y}\psi(x)\leq\frac{1}{1-1/2}\int_{1/2}^{1}P^{k}_{Y}\psi\,dm+|P^{k}_{Y}\psi|_{\mathrm{Lip}},

for every x[1/2,1]x\in[1/2,1] which, when combined with |1/21PYkψ|=|1/21ψ|1/2|ψ||\int_{1/2}^{1}P^{k}_{Y}\psi|=|\int_{1/2}^{1}\psi|\leq 1/2|\psi|_{\infty}, yields |PYkψ||ψ|+|PYkψ|Lip|P^{k}_{Y}\psi|_{\infty}\leq|\psi|_{\infty}+|P^{k}_{Y}\psi|_{\mathrm{Lip}}. Combining this with the estimate above gives the following Lasota-Yorke inequality.

PYkψM′′2kψ+(M′′+1)|ψ|.\|P^{k}_{Y}\psi\|\leq\frac{M^{\prime\prime}}{2^{k}}\|\psi\|+(M^{\prime\prime}+1)|\psi|_{\infty}. (10)

The operator PYP_{Y} is therefore quasi-compact (by Hennion’s theorem [13]).

We now show that PYP_{Y} has unique fixed point, giving rise to an invariant mixing measure ν0μ\nu^{\mathbb{N}_{0}}\otimes\mu for FYF_{Y} where μ\mu is absolutely continuous with respect to Lebesgue. In particular, consider the cone of positive functions 𝒞a={ψ:[1/2,1]+>0:ψ(x)/ψ(y)ea|xy|}\mathcal{C}_{a}=\{\psi:[1/2,1]\rightarrow\mathbb{R}^{+}>0:\;\psi(x)/\psi(y)\leq e^{a|x-y|}\}. Since ψ𝒞a\psi\in\mathcal{C}_{a} is independent of ω\omega

PYψ(x)=[α,β]0n1Pωn1Pω0(ψχJ(ω0ωn1))(x)dν0(ω0ωn1).P_{Y}\psi(x)=\int_{[\alpha,\beta]^{\mathbb{N}_{0}}}\sum_{n\geq 1}P_{\omega_{n-1}}\ldots P_{\omega_{0}}\left(\psi\chi_{J_{(\omega_{0}\ldots\omega_{n-1})}}\right)(x)d\nu^{\mathbb{N}_{0}}(\omega_{0}\ldots\omega_{n-1}\ldots).

Since f(ω0ωn1)nf^{n}_{(\omega_{0}\ldots\omega_{n-1})} has bounded distortion and expansion both uniform in nn and (ω0ωn1)(\omega_{0}\ldots\omega_{n-1}), there is a>0a>0 sufficiently large such that if ψ𝒞a\psi\in\mathcal{C}_{a}

Pωn1Pω0(ψχJ(ω0ωn1))(x)Pωn1Pω0(ψχJ(ω0ωn1))(y)ea|xy|x,y[1/2,1],n,(ω0ωn1)\frac{P_{\omega_{n-1}}\ldots P_{\omega_{0}}\left(\psi\chi_{J_{(\omega_{0}\ldots\omega_{n-1})}}\right)(x)}{P_{\omega_{n-1}}\ldots P_{\omega_{0}}\left(\psi\chi_{J_{(\omega_{0}\ldots\omega_{n-1})}}\right)(y)}\leq e^{a^{\prime}|x-y|}\;\;\;\;\forall x,y\in[1/2,1],\;\forall n,\;\forall(\omega_{0}\ldots\omega_{n-1})

with a<aa^{\prime}<a. It is easy to conclude that if ψ𝒞a\psi\in\mathcal{C}_{a}, then PYψ𝒞aP_{Y}\psi\in\mathcal{C}_{a^{\prime}}. It is standard to conclude that PYP_{Y} fixes a direction in 𝒞a\mathcal{C}_{a}, giving a mixing invariant absolutely continuous probability measure dπY=ψ0dYd\pi_{Y}=\psi_{0}d\mathbb{P}_{Y} (see [15] pp 244-250). It is also easy to see that ψ\psi\in\mathcal{B}. This fact together with quasi-compactness of PYP_{Y} and mixing implies that the operator has a spectral gap.

Point (i) We first deal with |z|<1|z|<1. Pick z𝔻z\in\mathbb{D}. Notice that

R(z)kφ(x)\displaystyle R(z)^{k}\varphi(x) =n1znJPωn1Pω0(φχJ)dν(𝝎),\displaystyle=\sum_{n\geq 1}\int z^{n}\sum_{J}P_{\omega_{n-1}}\circ\ldots\circ P_{\omega_{0}}(\varphi\chi_{J})\,d\nu^{\mathbb{N}}(\boldsymbol{\omega}),

where the summation is over 𝒥ω0,,ωn1k\mathcal{J}^{k}_{\omega_{0},\ldots,\omega_{n-1}}, the collection of intervals (depending on 𝝎\boldsymbol{\omega}) returning to YY for the kkth time at the nnth step as above. The terms corresponding to AA and BB in (7) are scaled by |z|n|z|^{n}, where nkn\geq k is the time of the kkth return. In particular, arguing as in (9), we see R(z)kM|z|k\|R(z)^{k}\|\leq M|z|^{k}, where MM does not depend on kk or zz. Hence for |z|<1|z|<1, n0R(z)n\sum_{n\geq 0}R(z)^{n} is convergent, and IdR(z)\operatorname{Id}-R(z) is invertible.

For |z|=1|z|=1, one can show that R(z)R(z) satisfies a Lasota-Yorke inequality as we did for R(1)=PYR(1)=P_{Y}. This implies that the essential spectrum of R(z)R(z) is contained in the open disk 𝔻\mathbb{D}, and IdR(z)\operatorname{Id}-R(z) is invertible if and only if 11 is not an eigenvalue of R(z)R(z). Now let zz lie on the unit circle but z1z\neq 1. Suppose for a contradiction that φ\varphi is a Lipschitz function that is an eigenfunction of R(z)R(z) with eigenvalue 1. Then taking absolute values and using the triangle inequality, we see R(1)|φ||φ|R(1)|\varphi|\geq|\varphi|, this inequality being strict unless for each xx, almost every term contributing to R(z)φR(z)\varphi has the same argument. Since R(1)|φ|(x)𝑑m(x)=|φ(x)|𝑑m(x)\int R(1)|\varphi|(x)\,dm(x)=\int|\varphi(x)|\,dm(x), this implies R(1)|φ|=|φ|R(1)|\varphi|=|\varphi|, so that |φ||\varphi| is the leading eigenvector, gg say, of PYP_{Y}. We write φ(x)=g(x)h(x)\varphi(x)=g(x)h(x) where |h(x)|=1|h(x)|=1 for all xx. Since g𝒞ag\in\mathcal{C}_{a}, it is bounded below, so that hh must be Lipschitz. The condition for equality in the triangle inequality implies that h(FY(𝝎,x))=h(x)zτY(𝝎,x)h(F_{Y}(\boldsymbol{\omega},x))=h(x)z^{\tau_{Y}(\boldsymbol{\omega},x)} for Y\mathbb{P}_{Y}-a.e. (𝝎,x)(\boldsymbol{\omega},x). In particular for ν\nu^{\mathbb{N}}-a.e. 𝝎\boldsymbol{\omega}, for mm-a.e. xx in each Jω0,ωn11J^{1}_{\omega_{0},\ldots\omega_{n-1}}, we have

h(FY(𝝎,x))=h(x)zn.h(F_{Y}(\boldsymbol{\omega},x))=h(x)z^{n}.

Since both sides of the equality are Lipschitz functions, this holds for all xJω0,ωn11x\in J^{1}_{\omega_{0},\ldots\omega_{n-1}}. Since the Jω0,ωn11J^{1}_{\omega_{0},\ldots\omega_{n-1}} are arbitrarily short for large nn, we see that varJ(h)\operatorname{var}_{J}(h) becomes arbitrarily small. But the above equality and the fact that FY(𝝎,)F_{Y}(\boldsymbol{\omega},\cdot) maps JJ onto [12,1][\frac{1}{2},1] imply that varJ(h)=var[12,1](h)\operatorname{var}_{J}(h)=\operatorname{var}_{[\frac{1}{2},1]}(h). Hence hh is a constant and we may assume φ=g\varphi=g. Now R(z)g=R(1)g=gR(z)g=R(1)g=g implies that z=1z=1.

Point (ii) One can verify the renewal equations

Tn\displaystyle T_{n} =T0Rn+T1Rn1++Tn1R1\displaystyle=T_{0}R_{n}+T_{1}R_{n-1}+\ldots+T_{n-1}R_{1}
Tn\displaystyle T_{n} =RnT0+Rn1T1++R1Tn1.\displaystyle=R_{n}T_{0}+R_{n-1}T_{1}+\ldots+R_{1}T_{n-1}.

Now since n1R(z)n\sum_{n\geq 1}R(z)^{n} converges, we deduce that (IdR(z))(\operatorname{Id}-R(z)) is invertible. The coefficient multiplying znz^{n} in this last sum is

k=1ni1++ik=nRi1Ri2Rik=Tn,\sum_{k=1}^{n}\sum_{i_{1}+\ldots+i_{k}=n}R_{i_{1}}R_{i_{2}}\ldots R_{i_{k}}=T_{n},

so T(z)=(IdR(z))1T(z)=(\operatorname{Id}-R(z))^{-1}. This also establishes the boundedness of TnT_{n} for each nn.

Point (iii) Notice that R(1)=n=1Rn=PYR(1)=\sum_{n=1}^{\infty}R_{n}=P_{Y}, so that there is the required spectral gap and simple eigenvector with eigenvalue 1 as shown above. We have R(1)=n=1nRn=n=1mnRmR^{\prime}(1)=\sum_{n=1}^{\infty}nR_{n}=\sum_{n=1}^{\infty}\sum_{m\geq n}R_{m}, so that this is a bounded operator by point (iv). Note also that R(1)R(1) preserves integrals, so that Π\Pi also preserves integrals and the operators RnR_{n} preserve the class of non-negative functions. Since R(1)=R(1)+n>1(n1)RnR^{\prime}(1)=R(1)+\sum_{n>1}(n-1)R_{n}, It follows that ΠR(1)Π\Pi R^{\prime}(1)\Pi is non-zero as required. ∎

5.2 Central Limit Theorem and Convergence to Stable Laws

The fact that PYP_{Y} has a spectral gap allows us to use the Nagaev-Guivarc’h method (as outlined in [1] [12]) to show that suitably rescaled Birkhoff sums converge to stable laws. In this approach, one first proves that suitably rescaled Birkhoff sums of observables in the induced system (FYF_{Y}) converge to stable laws, and then use this result to show that the same limit theorem holds for the unfolded system (FF). We are not going to give full details of the proofs, but we are going to present some computations needed in the particular case that we are treating, and direct the reader to the relevant literature for the remaining standard part of the argument.

Let us call

SYnψ:=ψ+ψFY++ψFYn1S^{n}_{Y}\psi:=\psi+\psi\circ F_{Y}+\ldots+\psi\circ F_{Y}^{n-1}

the Birkhoff sums with respect to FYF_{Y}. We assume, as in the statement of Theorem 2.3 that φ\varphi is a Lipschitz function on [0,1][0,1] satisfying φ(0)>0\varphi(0)>0. Recall that given φ\varphi, φY(𝝎,x)=i=0τY(𝝎,x)1φFi(𝝎,x)\varphi_{Y}(\boldsymbol{\omega},x)=\sum_{i=0}^{\tau_{Y}(\boldsymbol{\omega},x)-1}\varphi\circ F^{i}(\boldsymbol{\omega},x), and GG is the cumulative distribution function of φY\varphi_{Y} with respect to Y\mathbb{P}_{Y}: that is G(t)=Y({(𝝎,x)Ω×[12,1]:φY(𝝎,x)t}G(t)=\mathbb{P}_{Y}(\{(\boldsymbol{\omega},x)\in\Omega\times[\frac{1}{2},1]\colon\varphi_{Y}(\boldsymbol{\omega},x)\leq t\}.

Proposition 5.4.

Assume GG is such that L(t):=tp(1G(t))L(t):=t^{p}(1-G(t)) with p2p\neq 2 is slowly varying as t+t\to+\infty and G(t)=0G(t)=0 for large negative tt. Then there is a stable law 𝒵\mathcal{Z} a sequence AnA_{n}, and a sequence BnB_{n} satisfying nL(Bn)=BnpnL(B_{n})=B_{n}^{p} such that

limnSYnφYAnBn𝒵\lim_{n\rightarrow\infty}\frac{S^{n}_{Y}\varphi_{Y}-A_{n}}{B_{n}}\rightarrow\mathcal{Z}

where the convergence is in distribution. If p>2p>2, then 𝒵\mathcal{Z} is a Gaussian, while if p<2p<2, 𝒵\mathcal{Z} is a stable law of index pp.

Proof.

Calling PY,t():=PY(eitφY)P_{Y,t}(\cdot):=P_{Y}(e^{it\varphi_{Y}}\cdot), it follows that

𝔼[eitSYnφY]=PY,tn1𝑑.\mathbb{E}[e^{itS^{n}_{Y}\varphi_{Y}}]=\int P^{n}_{Y,t}1d\mathbb{P}.

The relation above provides the foundation to the Nagaev-Guivarc’h approach that recovers information on the characteristic function of SYnS^{n}_{Y} from spectral properties of the family (PY,t)t(P_{Y,t})_{t}.

Step 1

The operator PY,tP_{Y,t} has a spectral gap in \mathcal{B} for all sufficiently small tt. We show how to prove this fact directly with computations analogous to those carried out in the proof of point (iii) in Proposition 5.3. However, in that proposition φY=τY\varphi_{Y}=\tau_{Y}, i.e. φ=1\varphi=1, and therefore φY\varphi_{Y} is piecewise constant on every fibre Ω×{x}\Omega\times\{x\}. When this is not the case, φY\varphi_{Y} remains piecewise Lipschitz on fibres, where the fibres are partitioned according to the return time to YY. However, the Lipschitz constants on these pieces are typically not bounded, so the previous argument needs some extra care.

One needs to estimate |PY,tk(ψ)|Lip|P^{k}_{Y,t}(\psi)|_{\mathrm{Lip}}, given by

PY,tkψ(x)=𝑑ν0(𝝎)nkJPωn1Pω0(χJeitSkφYψ)(x),P^{k}_{Y,t}\psi(x)=\int d\nu^{\mathbb{N}_{0}}(\boldsymbol{\omega})\sum_{n\geq k}\sum_{J}P_{\omega_{n-1}}\cdots P_{\omega_{0}}(\chi_{J}e^{itS_{k}\varphi_{Y}}\psi)(x),

where the JJ summation is over 𝒥ω0,,ωn1k\mathcal{J}^{k}_{\omega_{0},\ldots,\omega_{n-1}}, the collection of intervals returning to YY for the kkth time at time nn. Now fix x<yx<y in [12,1][\frac{1}{2},1]; 𝝎[α,β]0\boldsymbol{\omega}\in[\alpha,\beta]^{\mathbb{N}_{0}}, knk\leq n and J𝒥ω0,,ωn1kJ\in\mathcal{J}^{k}_{\omega_{0},\ldots,\omega_{n-1}}. Let x,yx^{\prime},y^{\prime} be respectively the points in (f𝝎n)1(x)(f_{\boldsymbol{\omega}}^{n})^{-1}(x) and (f𝝎n)1(y)(f_{\boldsymbol{\omega}}^{n})^{-1}(y) that lie in JJ. Then the contribution to PY,tkψ(y)PY,tkψ(x)P^{k}_{Y,t}\psi(y)-P^{k}_{Y,t}\psi(x) coming from the interval JJ is

eitSkφY(y)ψ(y)(f𝝎n)(y)eitSkφY(x)ψ(x)(f𝝎n)(x).\frac{e^{itS_{k}\varphi_{Y}(y^{\prime})}\psi(y^{\prime})}{(f_{\boldsymbol{\omega}}^{n})^{\prime}(y^{\prime})}-\frac{e^{itS_{k}\varphi_{Y}(x^{\prime})}\psi(x^{\prime})}{(f_{\boldsymbol{\omega}}^{n})^{\prime}(x^{\prime})}.

We estimate the absolute value of this quantity by

t|SkφY(y)SkφY(x)|ψLip(f𝝎n)(y)+|ψ(y)ψ(x)|(f𝝎n)(y)+|ψ(x)||1(f𝝎n)(y)1(f𝝎n)(x)|\frac{t|S_{k}\varphi_{Y}(y^{\prime})-S_{k}\varphi_{Y}(x^{\prime})|\,\|\psi\|_{\mathrm{Lip}}}{(f_{\boldsymbol{\omega}}^{n})^{\prime}(y^{\prime})}+\frac{|\psi(y^{\prime})-\psi(x^{\prime})|}{(f_{\boldsymbol{\omega}}^{n})^{\prime}(y^{\prime})}+|\psi(x^{\prime})|\left|\frac{1}{(f_{\boldsymbol{\omega}}^{n})^{\prime}(y^{\prime})}-\frac{1}{(f_{\boldsymbol{\omega}}^{n})^{\prime}(x^{\prime})}\right|

The combined contribution from the second and third terms as nn runs over k,k+1,k,k+1,\ldots and JJ runs over 𝒥𝝎k\mathcal{J}^{k}_{\boldsymbol{\omega}} and integrated over 𝝎\boldsymbol{\omega} is estimated exactly as in Proposition 5.3. It remains to estimate the combined contribution from the first term. Since the fωf_{\omega}’s are non-contracting and the kkth return time is nn, the first term |SkφY(y)SkφY(x)||S_{k}\varphi_{Y}(y^{\prime})-S_{k}\varphi_{Y}(x^{\prime})| may be estimated by n|φ|Lip||xy|n|\varphi|_{\mathrm{Lip}}||x-y|. As before, the distortion estimates give 1/(f𝝎n)(y)K|J|1/(f_{\boldsymbol{\omega}}^{n})^{\prime}(y)\leq K|J|. Hence the combined contribution to |PY,tk(ψ)(y)PY,tk(ψ)(x)||P^{k}_{Y,t}(\psi)(y)-P^{k}_{Y,t}(\psi)(x)| coming from the first terms in the above display (as nn, JJ and 𝝎\boldsymbol{\omega} vary) is bounded above by

KtφLipψLip|xy|𝑑ν0(𝝎)nJ𝒥ω0,,ωn1kn|J|\displaystyle Kt\|\varphi\|_{\mathrm{Lip}}\|\psi\|_{\mathrm{Lip}}|x-y|\int d\nu^{\mathbb{N}_{0}}(\boldsymbol{\omega})\sum_{n}\sum_{J\in\mathcal{J}^{k}_{\omega_{0},\ldots,\omega_{n-1}}}n|J|
=\displaystyle= KtφLipψLip|xy|τk(𝝎,x)𝑑(𝝎,x),\displaystyle Kt\|\varphi\|_{\mathrm{Lip}}\|\psi\|_{\mathrm{Lip}}|x-y|\int\tau_{k}(\boldsymbol{\omega},x)\,d\mathbb{P}(\boldsymbol{\omega},x),

where τk(𝝎,x)\tau_{k}(\boldsymbol{\omega},x) denotes the kk-th return time to YY. The measure \mathbb{P} agrees with ν0πY\nu^{\mathbb{N}_{0}}\otimes\pi_{Y} (where πY\pi_{Y} is the absolutely continuous invariant measure for FYF_{Y}, i.e. the density of πY\pi_{Y} is the unique fixed point of PYP_{Y} on \mathcal{B}) up to a multiplicative factor that is uniformly bounded above and below. Hence the above displayed quantity may be estimated by

KtφLipψLip|xy|τk(𝝎,x)𝑑ν0πY\displaystyle\leq K^{\prime}t\|\varphi\|_{\mathrm{Lip}}\|\psi\|_{\mathrm{Lip}}|x-y|\int\tau_{k}(\boldsymbol{\omega},x)\,d\nu^{\mathbb{N}_{0}}\otimes\pi_{Y}
=kKtφLipψLip|xy|τ1(𝝎,x)𝑑ν0πY\displaystyle=kK^{\prime}t\|\varphi\|_{\mathrm{Lip}}\|\psi\|_{\mathrm{Lip}}|x-y|\int\tau_{1}(\boldsymbol{\omega},x)\,d\nu^{\mathbb{N}_{0}}\otimes\pi_{Y}

By Kac’s Lemma, the integral is finite, so the above reduces to CtψLip|xy|Ct\|\psi\|_{\mathrm{Lip}}|x-y|. Taking tt sufficiently small, we see that the additional contribution to the estimate of |PY,tk(ψ)|Lip|P^{k}_{Y,t}(\psi)|_{\mathrm{Lip}} from that appearing in the earlier proposition is CtψLipCt\|\psi\|_{\mathrm{Lip}}. In particular, for sufficiently small tt, one obtains a Lasota-Yorke inequality analogous to the one in (10).

Step 2

The family of operators (PY,t)(P_{Y,t}) (acting on \mathcal{B}) is continuous in tt. To see this, we estimate (PY,tPY,s)(ψ)Lip\|(P_{Y,t}-P_{Y,s})(\psi)\|_{\mathrm{Lip}}. As above, |(PY,tPY,s)(ψ)(y)(PY,tPY,s)(ψ)(x)||(P_{Y,t}-P_{Y,s})(\psi)(y)-(P_{Y,t}-P_{Y,s})(\psi)(x)| is expressed as an integral over ν0\nu^{\mathbb{N}_{0}} of a countable sum (one term for each nn). Fix x<yx<y in [12,1][\frac{1}{2},1], 𝝎\boldsymbol{\omega} and nn as above. Since we are considering first returns, there is exactly one interval, JJ, in 𝒥ω0,,ωn11\mathcal{J}^{1}_{\omega_{0},\ldots,\omega_{n-1}}. Letting xx^{\prime} and yy^{\prime} be the preimages of xx and yy under (f𝝎n)1(f_{\boldsymbol{\omega}}^{n})^{-1} in JJ, the corresponding contribution to Δ:=|(PY,tPY,s)(ψ)(y)(PY,tPY,s)(ψ)(x)|\Delta:=|(P_{Y,t}-P_{Y,s})(\psi)(y)-(P_{Y,t}-P_{Y,s})(\psi)(x)| is estimated by

|eisφY(y)(ei(ts)φY(y)1)ψ(y)(f𝝎n)(y)eisφY(x)(ei(ts)φY(x)1)ψ(x)(f𝝎n)(x)|\left|\frac{e^{is\varphi_{Y}(y^{\prime})}(e^{i(t-s)\varphi_{Y}(y^{\prime})}-1)\psi(y^{\prime})}{(f_{\boldsymbol{\omega}}^{n})^{\prime}(y^{\prime})}-\frac{e^{is\varphi_{Y}(x^{\prime})}(e^{i(t-s)\varphi_{Y}(x^{\prime})}-1)\psi(x^{\prime})}{(f_{\boldsymbol{\omega}}^{n})^{\prime}(x^{\prime})}\right|

(where we dropped the 𝝎\boldsymbol{\omega}-dependence of φY\varphi_{Y} for clarity of the notation), which we bound as a difference of products using the triangle inequality as usual, giving rise to the sum of four terms. We use the estimates |(ei(ts)φY(y)ei(ts)φY(x))/(f𝝎n)(ξ)|n|ts|KφLip|yx||(e^{i(t-s)\varphi_{Y}(y^{\prime})}-e^{i(t-s)\varphi_{Y}(x^{\prime})})/(f_{\boldsymbol{\omega}}^{n})^{\prime}(\xi)|\leq n|t-s|K\|\varphi\|_{\mathrm{Lip}}|y-x| for any ξ\xi between xx^{\prime} and yy^{\prime}; |ei(ts)φY(x)1|n|ts|KφLip|e^{i(t-s)\varphi_{Y}(x^{\prime})}-1|\leq n|t-s|K\|\varphi\|_{\mathrm{Lip}}; |1/(f𝝎n)(y)1/(f𝝎n)(x)|K|J||xy||1/(f_{\boldsymbol{\omega}}^{n})^{\prime}(y^{\prime})-1/(f_{\boldsymbol{\omega}}^{n})(x^{\prime})|\leq K|J||x-y| (obtained from the distortion bound). Combining these estimates, we find that the contribution to Δ\Delta coming from the nnth interval Jω0,,ωn1J_{\omega_{0},\ldots,\omega_{n-1}} from each of the four terms described above is bounded above by Cn|ts|φLip|xy||J|ψLipCn|t-s|\|\varphi\|_{\mathrm{Lip}}|x-y||J|\|\psi\|_{\mathrm{Lip}}. Summing over nn and integrating, we find

PY,tPY,sLip\displaystyle\|P_{Y,t}-P_{Y,s}\|_{\mathrm{Lip}} C|ts|φLipΩnn|Jω0,,ωn1|dν0(𝝎)\displaystyle\leq C|t-s|\|\varphi\|_{\mathrm{Lip}}\int_{\Omega}\sum_{n}n|J_{\omega_{0},\ldots,\omega_{n-1}}|\,d\nu^{\mathbb{N}_{0}}(\boldsymbol{\omega})
=C|ts|φLipΩ×[12,1]τ1(𝝎,x)𝑑m(x)𝑑ν0(𝝎).\displaystyle=C|t-s|\|\varphi\|_{\mathrm{Lip}}\int_{\Omega\times[\frac{1}{2},1]}\tau_{1}(\boldsymbol{\omega},x)\,dm(x)\,d\nu^{\mathbb{N}_{0}}(\boldsymbol{\omega}).

As in Step 1, the integral is finite using Kac’s Lemma, so that PY,tPY,sLipC|ts|\|P_{Y,t}-P_{Y,s}\|_{\mathrm{Lip}}\leq C^{\prime}|t-s|.

Step 3

The spectral gap of PY,tP_{Y,t} together with continuity of the family {PY,t}t\{P_{Y,t}\}_{t} implies that PY,tP_{Y,t} has a simple eigenvalue λ(t)\lambda(t) such that λ(t)1\lambda(t)\rightarrow 1 as t0t\rightarrow 0.

If p>2p>2, then the distribution random variable with distribution GG is square summable, and results from [16] show that the limit law is a Gaussian (see also [12]).

If p<2p<2, an expansion of λ(t)\lambda(t) in tt can be obtained following the proof of Theorem 5.1 from [1]. The expansion will depend on the distribution GG. One of the main requirements here is that GG is in the domain of attraction of a stable law or, equivalently, that there are L(x)L(x), a slowly varying function as x+x\rightarrow+\infty, and c1,c20c_{1},c_{2}\geq 0 with c1+c2>0c_{1}+c_{2}>0 such that G(x)=1xpL(x)(c1+o(1))G(x)=1-x^{-p}L(x)(c_{1}+o(1)) as x+x\rightarrow+\infty and G(x)=|x|pL(x)(c2+o(1))G(x)=|x|^{-p}L(-x)(c_{2}+o(1)) for xx\rightarrow-\infty. These conditions are ensured by the assumptions on GG with c2=0c_{2}=0. One can then conclude the statement of the proposition as outlined in Theorem 6.1 from [1]. The gist of the argument is that PY,1Bnn1\int P_{Y,\frac{1}{B_{n}}}^{n}1 approaches λ(1Bn)n\lambda(\frac{1}{B_{n}})^{n}. This information and the expansion of λ(t)\lambda(t) found at the previous step can be used to determine the limit for the law of SYnφYAnBn\frac{S^{n}_{Y}\varphi_{Y}-A_{n}}{B_{n}} when nn\rightarrow\infty. ∎

The following general theorem allows us to \sayunfold the limit theorem from the induced system FYF_{Y} to the original system FF.

Theorem 5.5 (Theorem 4.8 [12]).

Let T:XXT:X\rightarrow X be an ergodic probability preserving map w.r.t. the measure μ\mu, let (Mn)(M_{n}) and (Bn)(B_{n}) be two sequences of integers which are regularly varying with positive indices, let AnA_{n}\in\mathbb{R}, and let YXY\subset X be a subset of positive μ\mu measure. Denote by μY\mu_{Y} the probability measure induced on YY. Let τY\tau_{Y} be the first return time to YY, TY:=TτYT_{Y}:=T^{\tau_{Y}} the induced transformation, φ:X\varphi:X\rightarrow\mathbb{R} a measurable function, φY(x):=i=0τY(x)1φTi(x)\varphi_{Y}(x):=\sum_{i=0}^{\tau_{Y}(x)-1}\varphi\circ T^{i}(x), SYnφY=i=0n1φYTYi(x)S_{Y}^{n}\varphi_{Y}=\sum_{i=0}^{n-1}\varphi_{Y}\circ T_{Y}^{i}(x). Assume that (SYnφYAn)/Bn𝒵(S_{Y}^{n}\varphi_{Y}-A_{n})/B_{n}\rightarrow\mathcal{Z} in distribution, that MnM_{n} is such that (SYnτYn/μ(Y))/Mn(S_{Y}^{n}\tau_{Y}-n/\mu(Y))/M_{n} is tight and max0kMn|SYkφY|/Bn\max_{0\leq k\leq M_{n}}|S^{k}_{Y}\varphi_{Y}|/B_{n} tends in probability to zero. Then (SnφAnμ(Y))/Bnμ(Y)𝒵(S^{n}\varphi-A_{\lfloor{n\mu(Y)}\rfloor})/B_{\lfloor{n\mu(Y)}\rfloor}\rightarrow\mathcal{Z} in distribution.

Before proving Theorem 2.3, we show that the tails of φY\varphi_{Y} and τY\tau_{Y} satisfy the hypotheses of Proposition 5.4. Recall that ν\nu is a probability measure supported on [α,β][\alpha,\beta] with α<1\alpha<1; α\alpha in the topological support of ν\nu; and φ\varphi is a Lipschitz function on [0,1][0,1]. For the rest of this argument we will assume φ(0)>0\varphi(0)>0. Then there is a bounded number of times one can have φ(xn)<0\varphi(x_{n})<0 which implies the condition G(t)=0G(t)=0 for all t<t0t<t_{0}. The case φ(0)<0\varphi(0)<0 will be addressed at the very end of this section. The following proposition shows that the the distribution GG of φY\varphi_{Y} is a regularly varying function independently of the choice of ν\nu, i.e. it can always be written as the product of a monomial and a slowly varying function. Furthermore, the index of the distribution depends only on the minimum of the topological support of ν\nu.

Proposition 5.6.

Let ν\nu and φ\varphi be as above. Then G(t):=Y{(𝛚,x):φY(𝛚,x)t}G(t):=\mathbb{P}_{Y}\{(\boldsymbol{\omega},x):\varphi_{Y}(\boldsymbol{\omega},x)\leq t\} satisfies 1G(t)=1/t1/αL(t)1-G(t)=1/t^{1/\alpha}L(t), where L(t)L(t) is a slowly varying function.

The proof of this proposition requires a couple of lemmas. Let

q(x)=x12φ(t)t[(2t)γ𝑑ν(γ)]𝑑t.q(x)=\int_{x}^{\frac{1}{2}}\frac{\varphi(t)}{t\left[\int(2t)^{\gamma}\,d\nu(\gamma)\right]}\,dt.

In the function above, t[(2t)γ𝑑ν(γ)]t\left[\int(2t)^{\gamma}\,d\nu(\gamma)\right] is the average size (integrating γ\gamma) of a one-step displacement fγ(t)tf_{\gamma}(t)-t. The random i.i.d. nature of the composed maps implies that the number of steps needed to escape from the point x0x\approx 0 to the inducing set, concentrates around a given value, and q(x)q(x) is expected to be a proxy for the value around which φY(𝝎,x)\varphi_{Y}(\boldsymbol{\omega},x^{\prime}) concentrates, where xx^{\prime} is the unique preimage of xx in YY. As a convenient abuse of notation in the following computations, we will write, φY(𝝎,x)\varphi_{Y}(\boldsymbol{\omega},x) when x0x\approx 0, with the convention that this means φY(𝝎,x)\varphi_{Y}(\boldsymbol{\omega},x^{\prime}). The following lemmas make the above heuristic precise.

Lemma 5.7.

For all δ>0\delta>0, there exists an x0x_{0} such that if xx0x\leq x_{0}, then

ν{𝝎:φY(𝝎,z)q(x)}1δ\displaystyle\nu^{\mathbb{N}}\{\boldsymbol{\omega}:\varphi_{Y}(\boldsymbol{\omega},z)\geq q(x)\}\geq 1-\delta for all z(1δ)xz\leq(1-\delta)x
ν{𝝎:φY(𝝎,z)q(x)}δx\displaystyle\nu^{\mathbb{N}}\{\boldsymbol{\omega}:\varphi_{Y}(\boldsymbol{\omega},z)\geq q(x)\}\leq\delta x for all z(1+δ)x.\displaystyle\quad\text{for all $z\geq(1+\delta)x$}.

In particular, if xx0x\leq x_{0} and t=q(x)t=q(x), then

(12δ)xY({(𝝎,y):φY(𝝎,y)>t})(1+2δ)x.(1-2\delta)x\leq\mathbb{P}_{Y}(\{(\boldsymbol{\omega},y)\colon\varphi_{Y}(\boldsymbol{\omega},y)>t\})\leq(1+2\delta)x.

To prove this lemma we first show a concentration result for the number of steps needed to escape from the interval [en,en1][e^{-\sqrt{n}},e^{-\sqrt{n-1}}].

Lemma 5.8.

Let the probability measure ν\nu on [α,β][\alpha,\beta] be as above. Let InI_{n} denote the interval [en,en1)[e^{-\sqrt{n}},e^{-\sqrt{n-1}}) and fix η>0\eta>0. For x[en,fα(en))x\in[e^{-\sqrt{n}},f_{\alpha}(e^{-\sqrt{n}})), let Tn(𝛚,x)=min{k:fωk1fω0(x)>en1}T_{n}(\boldsymbol{\omega},x)=\min\{k\colon f_{\omega_{k-1}}\circ\ldots\circ f_{\omega_{0}}(x)>e^{-\sqrt{n-1}}\}, that is the number of steps spent in InI_{n}.

For all sufficiently large nn and all x[en,fα(en))x\in[e^{-\sqrt{n}},f_{\alpha}(e^{-\sqrt{n}})),

ν{𝝎:Tn(𝝎,x)[(1η)Nn,(1+η)Nn]}12n,\nu^{\mathbb{N}}\{\boldsymbol{\omega}:T_{n}(\boldsymbol{\omega},x)\in[(1-\eta)N_{n},(1+\eta)N_{n}]\}\geq 1-2^{-n},

where Nn=(en1en)/[en(2en)γ𝑑ν(γ)]N_{n}=(e^{-\sqrt{n-1}}-e^{-\sqrt{n}})/[e^{-\sqrt{n}}\int(2e^{-\sqrt{n}})^{\gamma}\,d\nu(\gamma)].

Notice that when a random orbit enters InI_{n} from In+1I_{n+1} the first point of InI_{n} that is encountered lies in [en,fα(en))[e^{-\sqrt{n}},f_{\alpha}(e^{-\sqrt{n}})). Hence the lemma is estimating the number of steps spent in InI_{n} when entering from the left.

Proof.

Let an=ena_{n}=e^{-\sqrt{n}}. If xInx\in I_{n}, the size of each step is x(2x)γx(2x)^{\gamma} where γ[α,β]\gamma\in[\alpha,\beta]. We approximate this above and below by an1(2an1)γa_{n-1}(2a_{n-1})^{\gamma} and an(2an)γa_{n}(2a_{n})^{\gamma}. The ratio of these step sizes is bounded above by (an1/an)1+β(a_{n-1}/a_{n})^{1+\beta}, which approaches 1 as nn\to\infty.

This allows us to compare the displacement of xx under the composition fγkfγ1f_{\gamma_{k}}\circ\ldots\circ f_{\gamma_{1}} (that is fγkfγ1(x)xf_{\gamma_{k}}\circ\ldots\circ f_{\gamma_{1}}(x)-x) to j=1kan(2an)γj\sum_{j=1}^{k}a_{n}(2a_{n})^{\gamma_{j}}, the sum of the displacements fγj(an)anf_{\gamma_{j}}(a_{n})-a_{n}, provided that fγjfγ1(x)f_{\gamma_{j}}\circ\ldots\circ f_{\gamma_{1}}(x) remains in InI_{n} for j=0,,k1j=0,\ldots,k-1. It is convenient (to maximize the strength of the conclusion from Hoeffding’s inequality) to rescale the displacements by a multiplicative factor of 2αan(1+α)=2αe(1+α)n2^{-\alpha}a_{n}^{-(1+\alpha)}=2^{-\alpha}e^{(1+\alpha)\sqrt{n}}.

Let (Xi)i(X_{i})_{i\in\mathbb{N}} be a sequence of i.i.d. random variables, where Xi=(2an)γiαX_{i}=(2a_{n})^{\gamma_{i}-\alpha} and γi\gamma_{i} is distributed according to ν\nu. In particular, 0Xi10\leq X_{i}\leq 1 for each ii. Similarly, set Yi=(2an1)γiαY_{i}=(2a_{n-1})^{\gamma_{i}-\alpha}, so that for xInx\in I_{n}, the step size x(2x)γix(2x)^{\gamma_{i}} satisfies 2αan1+αXix(2x)γi2αan11+αYi2^{\alpha}a_{n}^{1+\alpha}X_{i}\leq x(2x)^{\gamma_{i}}\leq 2^{\alpha}a_{n-1}^{1+\alpha}Y_{i}. We use 𝐏\mathbf{P} and 𝐄\mathbf{E} to refer to the distribution on the (Xi)(X_{i}) and (Yi)(Y_{i}) in order to distinguish from \mathbb{P} used elsewhere in the paper. Notice also that Nn=(an1an)/[2αan1+α𝐄X]N_{n}=(a_{n-1}-a_{n})/[2^{\alpha}a_{n}^{1+\alpha}\mathbf{E}X] (we drop the nn subscript in the following computations and note that the random variables here are X=(2an)γαX=(2a_{n})^{\gamma-\alpha} and Y=(2an1)γαY=(2a_{n-1})^{\gamma-\alpha}).

If we define (dependent) random variables DkD_{k} by Dk=fγkfγ1(x)fγk1fγ1(x)D_{k}=f_{\gamma_{k}}\circ\ldots\circ f_{\gamma_{1}}(x)-f_{\gamma_{k-1}}\circ\ldots f_{\gamma_{1}}(x), then provided fγkfγ1(x)f_{\gamma_{k}}\circ\ldots\circ f_{\gamma_{1}}(x) remains below an1a_{n-1} we have D1++DkY1++YkD_{1}+\ldots+D_{k}\leq Y_{1}+\ldots+Y_{k}. Hence for x[en,fα(en)]x\in[e^{-\sqrt{n}},f_{\alpha}(e^{-\sqrt{n}})], 2ααn11+α(Y1++Y(1η)N)<an1fα(an)2^{\alpha}\alpha_{n-1}^{1+\alpha}(Y_{1}+\ldots+Y_{(1-\eta)N})<a_{n-1}-f_{\alpha}(a_{n}) implies fγ(1η)Nfγ1(x)<an1f_{\gamma_{(1-\eta)N}}\circ\ldots f_{\gamma_{1}}(x)<a_{n-1}. For x[en,fα(en))x\in[e^{-\sqrt{n}},f_{\alpha}(e^{-\sqrt{n}})), we have

𝐏(fγ(1η)Nfγ1(x)>an1)\displaystyle\mathbf{P}(f_{\gamma_{(1-\eta)N}}\circ\ldots\circ f_{\gamma_{1}}(x)>a_{n-1})
𝐏(2αan11+α(Y1++Y(1η)N)>an1fα(an))\displaystyle\leq\mathbf{P}\left(2^{\alpha}a_{n-1}^{1+\alpha}(Y_{1}+\ldots+Y_{(1-\eta)N})>a_{n-1}-f_{\alpha}(a_{n})\right)
=𝐏(Y1++Y(1η)N(1η)N𝐄Y>(anan1)1+α(an1fα(an))𝐄X(1η)(an1an)𝐄Y)\displaystyle=\mathbf{P}\left(\frac{Y_{1}+\ldots+Y_{(1-\eta)N}}{(1-\eta)N}-\mathbf{E}Y>\left(\frac{a_{n}}{a_{n-1}}\right)^{1+\alpha}\frac{(a_{n-1}-f_{\alpha}(a_{n}))\mathbf{E}X}{(1-\eta)(a_{n-1}-a_{n})}-\mathbf{E}Y\right)
𝐏(Y1++Y(1η)N(1η)N𝐄Y>η2𝐄X)\displaystyle\leq\mathbf{P}\left(\frac{Y_{1}+\ldots+Y_{(1-\eta)N}}{(1-\eta)N}-\mathbf{E}Y>\tfrac{\eta}{2}\mathbf{E}X\right)
e2(1η)N(η2𝐄X)2,\displaystyle\leq e^{-2(1-\eta)N(\frac{\eta}{2}\mathbf{E}X)^{2}},

The equality follows from the expression for NnN_{n} and a simple manipulation of the inequality in parentheses. The second inequality holds for all sufficiently large nn using the facts an/an1=1+o(1)a_{n}/a_{n-1}=1+o(1), (an1fα(an))/(an1an)=1+o(1)(a_{n-1}-f_{\alpha}(a_{n}))/(a_{n-1}-a_{n})=1+o(1) and 𝐄X=(1+o(1))𝐄Y\mathbf{E}X=(1+o(1))\mathbf{E}Y, which follows from the comparison of step sizes mentioned above; the final inequality is an application from Hoeffding’s inequality. Since N(an1an)/(2αan1+α)N\geq(a_{n-1}-a_{n})/(2^{\alpha}a_{n}^{1+\alpha}), we see that Neαn/(21+αn)N\geq e^{\alpha\sqrt{n}}/(2^{1+\alpha}n), while γα\gamma-\alpha takes values in the range [0,α4][0,\frac{\alpha}{4}] with positive probability, pp say, so that (𝐄X)2>p2eα2n(\mathbf{E}X)^{2}>p^{2}e^{-\frac{\alpha}{2}\sqrt{n}}. Thus we see the probability of spending less than (1η)N(1-\eta)N steps in InI_{n} is (much) smaller than 2n12^{-n-1} for large nn.

An analogous argument based on studying the probability that 2αan1+α(X1++X(1+η)N)<an1an2^{\alpha}a_{n}^{1+\alpha}(X_{1}+\ldots+X_{(1+\eta)N})<a_{n-1}-a_{n} shows that the probability of spending more than (1+η)N(1+\eta)N steps in InI_{n} is smaller than 2n12^{-n-1} for large nn. ∎

Proof of Lemma 5.7.

Let δ>0\delta>0 be fixed. Given n>0n>0, let x[en,en1)x\in[e^{-\sqrt{n}},e^{-\sqrt{n-1}}), m=α2/(4β2)nm=\lceil\alpha^{2}/(4\beta^{2})n\rceil and p=nδnp=\lfloor n-\delta\sqrt{n}\rfloor. Notice that ap(1+δ)xa_{p}\leq(1+\delta)x for large nn. Let EE be the event that starting from apa_{p}, the number of steps spent in IjI_{j} is at most eδα/9Nje^{\delta\alpha/9}N_{j} for each j=m+1,,pj=m+1,\ldots,p. Applying Lemma 5.8 with η\eta taken to be eδα/91e^{\delta\alpha/9}-1, provided nn (and hence mm) is sufficiently large, EE has probability at least 12m+11-2^{-m+1}. After leaving Im+1I_{m+1}, the position exceeds eme^{-\sqrt{m}} and the number of steps before hitting YY is bounded above by eβme^{\beta\sqrt{m}}, so that the contribution to φY\varphi_{Y} from this part of the orbit is bounded above by eβmφe^{\beta\sqrt{m}}\|\varphi\|. Similarly, for 𝝎E\boldsymbol{\omega}\in E, the contribution from IpIm+1I_{p}\cup\ldots\cup I_{m+1} is bounded above by eδα/9j=m+1pNj(φ(aj)+φLip(aj1aj))e^{\delta\alpha/9}\sum_{j=m+1}^{p}N_{j}\big{(}\varphi(a_{j})+\|\varphi\|_{\text{Lip}}(a_{j-1}-a_{j})\big{)}. We claim that for all sufficiently large nn (with mm and pp related to nn as above),

eδα/9j=m+1pNj(φ(aj)+φLip(aj1aj))+eβmφ<x12φ(t)t(2t)γ𝑑ν(γ)𝑑t.e^{\delta\alpha/9}\sum_{j=m+1}^{p}N_{j}\Big{(}\varphi(a_{j})+\|\varphi\|_{\text{Lip}}(a_{j-1}-a_{j})\Big{)}+e^{\beta\sqrt{m}}\|\varphi\|<\int_{x}^{\frac{1}{2}}\frac{\varphi(t)}{\int t(2t)^{\gamma}\,d\nu(\gamma)}\,dt. (11)

One may check that (2t)γ𝑑ν(γ)=(2t)α+ε(t)\int(2t)^{\gamma}\,d\nu(\gamma)=(2t)^{\alpha+\varepsilon(t)}, where ε(t)0\varepsilon(t)\to 0 as t0t\to 0. In particular, the right side of (11) exceeds CeαnCe^{\alpha\sqrt{n}} for all large nn (and xInx\in I_{n}). By the choice of mm, eβmφ=o(eαn)e^{\beta\sqrt{m}}\|\varphi\|=o(e^{\alpha\sqrt{n}}). Similarly, for all sufficiently large nn, Nje(α+12)jN_{j}\leq e^{(\alpha+\frac{1}{2})\sqrt{j}} for all jmj\geq m. Also aj1aj=o(ej)a_{j-1}-a_{j}=o(e^{-\sqrt{j}}), so that j=m+1pNj(aj1aj)=o(ne(α12)n)\sum_{j=m+1}^{p}N_{j}(a_{j-1}-a_{j})=o(ne^{(\alpha-\frac{1}{2})\sqrt{n}}). To show (11) it therefore suffices to show

j=m+1pNjφ(aj)eδα/9x12φ(t)t(2t)γ𝑑ν(t)𝑑t.\sum_{j=m+1}^{p}N_{j}\varphi(a_{j})\leq e^{-\delta\alpha/9}\int_{x}^{\frac{1}{2}}\frac{\varphi(t)}{\int t(2t)^{\gamma}\,d\nu(t)}\,dt. (12)

To see this, set h(t)=φ(t)/t(2t)γ𝑑ν(t)h(t)=\varphi(t)/\int t(2t)^{\gamma}\,d\nu(t) and observe

j=m+1pNjφ(aj)=j=m+1p(aj1aj)h(aj)\sum_{j=m+1}^{p}N_{j}\varphi(a_{j})=\sum_{j=m+1}^{p}(a_{j-1}-a_{j})h(a_{j})

This Riemann sum is

(1+o(1))apamh(t)𝑑t,(1+o(1))\int_{a_{p}}^{a_{m}}h(t)\,dt,

Notice that provided nn is sufficiently large, apeδ/2xa_{p}\geq e^{\delta/2}x. We also see that for any b>1b>1, for all sufficiently small tt, h(bt)bα/3bα+1h(t)h(bt)\leq\frac{b^{\alpha/3}}{b^{\alpha+1}}h(t), so that for large nn (and corresponding mm and pp),

apamh(t)𝑑t\displaystyle\int_{a_{p}}^{a_{m}}h(t)\,dt =eδ/2eδ/2apeδ/2amh(eδ/2t)𝑑t\displaystyle=e^{\delta/2}\int_{e^{-\delta/2}a_{p}}^{e^{-\delta/2}a_{m}}h(e^{\delta/2}t)\,dt
eδα/3eδ/2apeδ/2amh(t)𝑑t\displaystyle\leq e^{-\delta\alpha/3}\int_{e^{-\delta/2}a_{p}}^{e^{-\delta/2}a_{m}}h(t)\,dt
eδα/3xeδ/2amh(t)𝑑t.\displaystyle\leq e^{-\delta\alpha/3}\int_{x}^{e^{-\delta/2}a_{m}}h(t)\,dt.

Since

eδ/2am12h(t)𝑑t=o(x12h(t)𝑑t),\int_{e^{-\delta/2}a_{m}}^{\frac{1}{2}}h(t)\,dt=o\left(\int_{x}^{\frac{1}{2}}h(t)\,dt\right),

the claimed inequalities (12) and hence (11) follow. We have therefore shown that for 𝝎E\boldsymbol{\omega}\in E, φY(𝝎,ap)<q(x)\varphi_{Y}(\boldsymbol{\omega},a_{p})<q(x). Now if z(1+δ)xz\geq(1+\delta)x, since zapz\geq a_{p}, the same bound applies to φY(𝝎,z)\varphi_{Y}(\boldsymbol{\omega},z). Hence we have established the second statement of the lemma. The first statement is proved in a very similar way, by using Lemma 5.8 to give lower bounds on the time spent in Iq,,ImI_{q},\ldots,I_{m} where q=n+δnq=\lceil n+\delta\sqrt{n}\,\rceil. ∎

Proof of Proposition 5.6.

Since for x[12,1]x\in[\frac{1}{2},1], fγ(x)=2x1f_{\gamma}(x)=2x-1 for all γ\gamma, Lemma 5.7 shows that Y({𝝎,x):φY(𝝎,x)>t})=q1(t)(1+o(1))\mathbb{P}_{Y}(\{\boldsymbol{\omega},x)\colon\varphi_{Y}(\boldsymbol{\omega},x)>t\})=q^{-1}(t)(1+o(1)). It therefore suffices to show that q1(t)q^{-1}(t) is a regularly varying function with order 1/α-1/\alpha. However, Theorem 1.5.12 of [7] shows that if g(x)g(x) is regularly varying of order α\alpha as xx\to\infty, then g1(x)g^{-1}(x) is regularly varying of order 1/α1/\alpha. Applying this to g(t)=q(1/t)g(t)=q(1/t), it suffices to show that limxg(bx)/g(x)=bα\lim_{x\to\infty}g(bx)/g(x)=b^{\alpha} for each b>0b>0. A calculation shows

g(x)=2x1u(2u)γ𝑑ν(γ)𝑑u.g(x)=\int_{2}^{x}\frac{1}{u\mathbf{\int}(\frac{2}{u})^{\gamma}\,d\nu(\gamma)}\,du.

Now set h(s)=esγ𝑑ν(γ)h(s)=\int e^{-s\gamma}\,d\nu(\gamma). If s=(1λ)r+λps=(1-\lambda)r+\lambda p, noticing that 11λ\frac{1}{1-\lambda} and 1λ\frac{1}{\lambda} are conjugate Hölder exponents, we have

h(s)\displaystyle h(s) =e(1λ)rγeλpγ𝑑ν(γ)\displaystyle=\int e^{-(1-\lambda)r\gamma}e^{-\lambda p\gamma}\,d\nu(\gamma)
(erγ𝑑ν(γ))1λ(epγ𝑑ν(γ))λ\displaystyle\leq\left(\int e^{-r\gamma}\,d\nu(\gamma)\right)^{1-\lambda}\left(\int e^{-p\gamma}\,d\nu(\gamma)\right)^{\lambda}
=h(r)1λh(p)λ,\displaystyle=h(r)^{1-\lambda}h(p)^{\lambda},

so that hh is log-convex. Since for all ε>0\varepsilon>0, ν([α,α+ε])e(α+ε)sh(s)eαs\nu([\alpha,\alpha+\varepsilon])e^{-(\alpha+\varepsilon)s}\leq h(s)\leq e^{-\alpha s} for large ss, it follows that (logh)(s)α(\log h)^{\prime}(s)\to-\alpha as ss\to\infty and h(s+l)/h(s)eαlh(s+l)/h(s)\to e^{-\alpha l} as ss\to\infty. It follows that u(2u)γ𝑑ν(γ)=elog(u/2)γ𝑑ν(γ)u\mapsto\int(\frac{2}{u})^{\gamma}\,d\nu(\gamma)=\int e^{-\log(u/2)\gamma}d\nu(\gamma) is regularly varying with index α-\alpha as uu\to\infty, in fact

elog(λu/2)γ𝑑ν(γ)elog(u/2)γ𝑑ν(γ)=h(log(u/2)+logλ)h(log(u/2))eαlogλ=λα.\frac{\int e^{-\log(\lambda u/2)\gamma}d\nu(\gamma)}{\int e^{-\log(u/2)\gamma}d\nu(\gamma)}=\frac{h(\log(u/2)+\log\lambda)}{h(\log(u/2))}\rightarrow e^{-\alpha\log\lambda}=\lambda^{-\alpha}.

Now it follows from Theorem 1.5.11 (i) of [7] that since [h(log(u/2))]1[h(\log(u/2))]^{-1} is regularly varying with index α\alpha, for σ=1(α+1)\sigma=-1\geq-(\alpha+1)

xσ+1[h(log(x/2))]12xuσ[h(log(u/2))]1𝑑u=[h(log(x/2))]1g(x)σ+α+1=α.\frac{x^{\sigma+1}[h(\log(x/2))]^{-1}}{\int_{2}^{x}u^{-\sigma}[h(\log(u/2))]^{-1}du}=\frac{[h(\log(x/2))]^{-1}}{g(x)}\rightarrow\sigma+\alpha+1=\alpha.

and therefore g(x)g(x) is regularly varying with index α\alpha as required. ∎

Proof of Theorem 2.3.

We can restrict to the case φ𝑑π(x)=0\int\varphi\,d\pi(x)=0, where π\pi is the stationary measure. By Proposition 5.6, the cumulative distribution function, GφG_{\varphi} of φY\varphi_{Y} satisfies 1Gφ(t)=Lφ(t)t1α1-G_{\varphi}(t)=L_{\varphi}(t)t^{-\frac{1}{\alpha}}, where LφL_{\varphi} is slowly varying. We showed in the proof of Lemma 5.7 that 1Gφ(t)=(1+o(1))qφ1(t)1-G_{\varphi}(t)=(1+o(1))q_{\varphi}^{-1}(t) where qφ(x)=x12φ(z)/[z(2z)γ𝑑ν(γ)]𝑑zq_{\varphi}(x)=\int_{x}^{\frac{1}{2}}\varphi(z)/[\int z(2z)^{\gamma}\,d\nu(\gamma)]\,dz. The same proposition applies if φ\varphi is replaced by 𝟏\mathbf{1} so that the cumulative distribution function, GτG_{\tau} of τY\tau_{Y} satisfies 1Gτ(t)=Lτ(t)t1α1-G_{\tau}(t)=L_{\tau}(t)t^{-\frac{1}{\alpha}}, where LτL_{\tau} is again slowly varying. As before, 1Gτ1-G_{\tau} behaves asymptotically as the inverse function of qτ(x)=x121/[z(2z)γ𝑑ν(γ)]𝑑xq_{\tau}(x)=\int_{x}^{\frac{1}{2}}1/[\int z(2z)^{\gamma}\,d\nu(\gamma)]\,dx. Since qφ(x)=φ(0)qτ(x)(1+o(1))q_{\varphi}(x)=\varphi(0)q_{\tau}(x)(1+o(1)) and both are regularly varying of order α-\alpha, we see that qφ(φ(0)1αx)qτ(x)q_{\varphi}(\varphi(0)^{\frac{1}{\alpha}}x)\sim q_{\tau}(x) and 1Gτ(t)(1Gφ(t))/φ(0)1α1-G_{\tau}(t)\sim(1-G_{\varphi}(t))/\varphi(0)^{\frac{1}{\alpha}}.

Let BnB_{n} be the solution of nLφ(Bn)=Bn1αnL_{\varphi}(B_{n})=B_{n}^{\frac{1}{\alpha}} as in the statement of Proposition 5.4. By that proposition, we obtain the existence of AnA_{n} and AnA_{n}^{\prime} such that (SYnφYAn)/Bn(S_{Y}^{n}\varphi_{Y}-A_{n})/B_{n} and (SYnτYAn)/Bn(S_{Y}^{n}\tau_{Y}-A_{n}^{\prime})/B_{n} converge to stable laws. By Kac’s Lemma, we may assume An=n/π(Y)A_{n}^{\prime}=n/\pi(Y).

Since (SYnτYn/π(Y))/Bn(S_{Y}^{n}\tau_{Y}-n/\pi(Y))/B_{n} converges in distribution, then it is also tight. Since φY\varphi_{Y} is integrable with YφY𝑑πY=φ𝑑π^=0\int_{Y}\varphi_{Y}d\pi_{Y}=\int\varphi d\hat{\pi}=0, where π^\hat{\pi} is the unfolding of πY\pi_{Y} to [0,1][0,1] and π^\hat{\pi} equals, up to a normalizing factor, the invariant measure π\pi. By Birkhoff’s ergodic theorem SYBnφY/BnS_{Y}^{B_{n}}\varphi_{Y}/{B_{n}} tends to zero almost everywhere. This implies that max0kBn|SYkφY|/Bn\max_{0\leq k\leq B_{n}}|S_{Y}^{k}\varphi_{Y}|/B_{n} tends to zero almost surely (and thus in probability). To see this, consider 𝒜\mathcal{A} the set where the Birkhoff averages have limit zero. For x𝒜x\in\mathcal{A}, call kn(x)k_{n}(x) the sequence of indices realizing the maxima. The sequence (kn(x))(k_{n}(x)) is non-decreasing. If kn(x)k_{n}(x) is eventually constant, then |SYkn(x)φY(x)|/Bn0|S_{Y}^{k_{n}(x)}\varphi_{Y}(x)|/B_{n}\rightarrow 0. If not, then

|SYkn(x)φY(x)|Bn=kn(x)Bn|SYkn(x)φY(x)|kn(x)0\frac{|S_{Y}^{k_{n}(x)}\varphi_{Y}(x)|}{B_{n}}=\frac{k_{n}(x)}{B_{n}}\frac{|S_{Y}^{k_{n}(x)}\varphi_{Y}(x)|}{k_{n}(x)}\rightarrow 0

By Theorem 5.5 the conclusion of the theorem follows when φ(0)>0\varphi(0)>0. ∎

Our final step it to consider the case φ(0)<0\varphi(0)<0. One simply observes that the statement of the Theorem 5.5 is invariant under φφ\varphi\rightarrow-\varphi, AnAnA^{\prime}_{n}\rightarrow-A^{\prime}_{n} and 𝒵𝒵\mathcal{Z}\rightarrow-\mathcal{Z}, another stable law, and apply the previous argument.

6 Diffusion Driven Decay of Correlations

In this section we consider a family of distributions ν\nu on the parameter space, with unbounded support. More precisely, for α,ε>0\alpha,\varepsilon>0, να,ε\nu_{\alpha,\varepsilon} is supported on [α,)[\alpha,\infty), and given by

να,ε[0,t]=F(t):={1(tα)εif tα;0if t<α.\nu_{\alpha,\varepsilon}[0,t]=F(t):=\begin{cases}1-\left(\frac{t}{\alpha}\right)^{-\varepsilon}&\text{if $t\geq\alpha$;}\\ 0&\text{if $t<\alpha$.}\end{cases}

Notice that since the measure has unbounded support, the considerations made in sections 4 and 5 on the uniform bound of the distortion, do not apply. Furthermore, the fat polynomial tails of the distribution of ν\nu, imply that arbitrarily large parameters are sampled somewhat frequently.

6.1 Markov chains with subgeometric rates of convergence

The arguments in sections 4 and 5 do not apply to this setup. Instead, we exploit the absolute continuity of ν\nu to translate the deterministic process (2) into a Markov process on the state space [0,1][0,1] and apply a theorem on Markov chains (see [22]) to prove the existence of a stationary measure and to estimate the asymptotic rate of correlation decay. To this end, consider the stationary Markov process (Xt)t0(X_{t})_{t\in\mathbb{N}_{0}} on the probability space (Ω×[0,1],(Ω×[0,1]),)(\Omega\times[0,1],\mathcal{B}(\Omega\times[0,1]),\mathbb{P}) taking values in [0,1][0,1], where Xt(𝝎,x0)=ΠFt(𝝎,x0)X_{t}(\boldsymbol{\omega},x_{0})=\Pi\circ F^{t}(\boldsymbol{\omega},x_{0}). Calling Γx:[α,β][0,1]\Gamma_{x}:[\alpha,\beta]\rightarrow[0,1] the map Γx(ω)=fω(x)\Gamma_{x}(\omega)=f_{\omega}(x), then the transition probability P(x,A)=(XtA|Xt1=x)P(x,A)=\mathbb{P}\left(X_{t}\in A|\;X_{t-1}=x\right), is given by P(x,)=(Γx)νP(x,\cdot)=(\Gamma_{x})_{*}\nu, i.e. the push-forward under Γx\Gamma_{x} of the probability ν\nu on the parameter space, or more concretely P(x,B)=ν{γ:fγ(x)B}P(x,B)=\nu\{\gamma\colon f_{\gamma}(x)\in B\}. The Markov chain defines an evolution of measures, which we denote by QQ: Qμ(B)=[0,1]P(x,B)𝑑μ(x)Q\mu(B)=\int_{[0,1]}P(x,B)\,d\mu(x), that is if XnX_{n} has distribution μ\mu, then Xn+1X_{n+1} has distribution QμQ\mu. In case μ\mu is an absolutely continuous measure, with density ρ\rho, QμQ\mu is again absolutely continuous with density

Pρ(x)=ρ(fγ|[0,12]1(x))fγ(fγ|[0,12]1(x))𝑑ν(γ)+ρ(x+12)2.P\rho(x)=\int\frac{\rho({f_{\gamma}|_{[0,\frac{1}{2}]}}^{-1}(x))}{f_{\gamma}^{\prime}({f_{\gamma}|_{[0,\frac{1}{2}]}}^{-1}(x))}\,d\nu(\gamma)+\frac{\rho(\frac{x+1}{2})}{2}. (13)

For A[0,1]A\subset[0,1], we define τA(𝝎,x)=inf{n|Xn(𝝎,x)A}\tau_{A}(\boldsymbol{\omega},x)=\inf\{n\in\mathbb{N}|\;X_{n}(\boldsymbol{\omega},x)\in A\}, and let us call x\mathbb{P}_{x}, the probability measure \mathbb{P} conditioned on {X0=x}\{X_{0}=x\}.

Definition 6.1.

A Markov chain is ψ\psi-irreducible if there is a measure ψ\psi such that for every A(S)A\in\mathcal{B}(S) with ψ(A)>0\psi(A)>0 and every xSx\in S

x(τA<)>0.\mathbb{P}_{x}\left(\tau_{A}<\infty\right)>0.
Lemma 6.2.

The Markov chain {Xt}t0\{X_{t}\}_{t\in\mathbb{N}_{0}} is ψ\psi-irreducible and aperiodic.

Proof.

The only problem with these two properties is that P(0,)=δ0P(0,\cdot)=\delta_{0}. However, this can be fixed by removing the set {0}{hn(12):n0}\{0\}\cup\{h^{-n}(\tfrac{1}{2}):n\geq 0\} from the state space, where h:[12,1][0,1]h:[\frac{1}{2},1]\rightarrow[0,1] is h(x)=2x1h(x)=2x-1 mod 11, and redefining the σ\sigma-algebras and transition probabilities accordingly. In this modified state space, every orbit eventually enters (0,1/2)(0,1/2) and is spread by the diffusion. This implies that any set of positive Lebesgue measure is eventually visited by a random orbit originating from any given xx. Aperiodicity is a consequence of the two branches of all the maps fγf_{\gamma} being onto. ∎

Definition 6.3.

Given a probability distribution a:0+a:\mathbb{N}\rightarrow\mathbb{R}^{+}_{0}, and a nontrivial measure νa\nu_{a}, a set C(S)C\in\mathcal{B}(S) is νa\nu_{a}-petite if

Ka(x,):=n=1a(n)Pn(x,)νa(),xC.K_{a}(x,\cdot):=\sum_{n=1}^{\infty}a(n)P^{n}(x,\cdot)\geq\nu_{a}(\cdot),\quad\quad\forall x\in C.
Definition 6.4.

A function r:0r:\mathbb{N}_{0}\rightarrow\mathbb{R} is a subgeometric rate function if there is a non-decreasing r0:0r_{0}:\mathbb{N}_{0}\rightarrow\mathbb{R} with r0(1)2r_{0}(1)\geq 2 and log(r0(n))/n0\log(r_{0}(n))/n\rightarrow 0 as nn\rightarrow\infty such that

lim infnr(n)r0(n)>0,lim supnr(n)r0(n)<.\liminf_{n\rightarrow\infty}\frac{r(n)}{r_{0}(n)}>0,\quad\quad\limsup_{n\rightarrow\infty}\frac{r(n)}{r_{0}(n)}<\infty.

The main result we are going to use in this section is the following theorem on subgeometric rates of convergence of ergodic Markov chains.

Theorem 6.5 ([22]).

Suppose that {Xt}t0\{X_{t}\}_{t\in\mathbb{N}_{0}} is a discrete time Markov process on a general state space SS endowed with a countably generated σ\sigma-field (S)\mathcal{B}(S) which is ψ\psi-irreducible and aperiodic. Suppose that there is a petite set C(S)C\in\mathcal{B}(S) and a subgeometric rate function r:+r:\mathbb{N}\rightarrow\mathbb{R}^{+} (see above for the definitions) such that

supxC𝔼x[n=0τC1r(n)]<.\sup_{x\in C}\mathbb{E}_{x}\left[\sum_{n=0}^{\tau_{C}-1}r(n)\right]<\infty. (14)

where τC:=inf{t:XtC}\tau_{C}:=\inf\{t\in\mathbb{N}:\;X_{t}\in C\}, and 𝔼x\mathbb{E}_{x} is the expectation with respect to the probability law of the Markov process conditioned on X0=xX_{0}=x. Then the Markov process admits a stationary measure π\pi, and for almost every point xSx\in S

limnr(n)Pn(x,)πTV=0,\lim_{n\rightarrow\infty}r(n)\|P^{n}(x,\cdot)-\pi\|_{TV}=0,

where TV\|\cdot\|_{TV} denotes the total variation norm difference between the two measures. Furthermore, if (14) holds and λ\lambda is a probability measure on SS that satisfies

𝔼λ[n=0τC1r(n)]<,\mathbb{E}_{\lambda}\left[\sum_{n=0}^{\tau_{C}-1}r(n)\right]<\infty, (15)

then

r(n)Pn(x,)πTV𝑑λ(x)0.r(n)\int\|P^{n}(x,\cdot)-\pi\|_{TV}d\lambda(x)\rightarrow 0.

For the remainder of this section, let α\alpha and ε\varepsilon be fixed. Let b<12b<\frac{1}{2} be chosen so that fα(b)>3+b4f_{\alpha}(b)>\frac{3+b}{4}.

We compute the one step probability transition density starting from a point x<12x<\frac{1}{2}, which will be denoted px(y)p_{x}(y) (with support [x,fα(x)][x,f_{\alpha}(x)]). For x<yfα(x)x<y\leq f_{\alpha}(x), define γx(y)\gamma_{x}(y) to be the value of γ\gamma such that fγ(x)=yf_{\gamma}(x)=y. That is γ\gamma satisfies x(1+(2x)γ)=yx(1+(2x)^{\gamma})=y, or (2x)γ=yxx(2x)^{\gamma}=\frac{y-x}{x}, giving an explicit form: γx(y)=logxyx/log12x\gamma_{x}(y)=\log\frac{x}{y-x}/\log\frac{1}{2x}, where we took reciprocals of both numerator and denominator to ensure positivity of the logarithms. Note that this is a decreasing function of yy. We can then obtain px(y)p_{x}(y) by

px(y)\displaystyle p_{x}(y) =limh0F(γx(y))F(γx(y+h))h\displaystyle=\lim_{h\to 0}\frac{F(\gamma_{x}(y))-F(\gamma_{x}(y+h))}{h}
=F(γx(y))γx(y)\displaystyle=-F^{\prime}(\gamma_{x}(y))\cdot\gamma_{x}^{\prime}(y)
=εαε(γx(y))1+ε1(yx)log(12x)\displaystyle=\frac{\varepsilon\alpha^{\varepsilon}}{(\gamma_{x}(y))^{1+\varepsilon}}\cdot\frac{1}{(y-x)\log(\frac{1}{2x})}
=εαε(log(12x))ε(log(xyx))1+ε(yx).\displaystyle=\frac{\varepsilon\alpha^{\varepsilon}\big{(}\log(\frac{1}{2x})\big{)}^{\varepsilon}}{\big{(}\log(\frac{x}{y-x})\big{)}^{1+\varepsilon}(y-x)}.

Recall that px()p_{x}(\cdot) is supported on [x,fα(x)][x,f_{\alpha}(x)]. By the choice of bb and since b<x<12b<x<\frac{1}{2}, this contains the sub-interval [b+12,b+34][\frac{b+1}{2},\frac{b+3}{4}]. We will denote by pxn(y)p^{n}_{x}(y) the nn-step probability transition density.

In this case where ν\nu has a power law distribution, we can write the transition operator (13) of the Markov process in the form:

Pρ(y)=012px(y)ρ(x)𝑑x+12ρ(y+12),P\rho(y)=\int_{0}^{\frac{1}{2}}p_{x}(y)\rho(x)dx+\tfrac{1}{2}\rho\big{(}\tfrac{y+1}{2}\big{)},

where px(y)p_{x}(y) is taken to be 0 if yxy\leq x or y>fα(x)y>f_{\alpha}(x).

Remark 6.6.

For any fixed b<12b<\frac{1}{2}, the set

C:=[fα1(b),b]C:=[f_{\alpha}^{-1}(b),b]

is a petite set. In the case where fα(b)>b+34f_{\alpha}(b)>\frac{b+3}{4} that we are considering, it is not hard to see that px2(y)p^{2}_{x}(y) is uniformly bounded below for xCx\in C and y[12,b+12]y\in[\frac{1}{2},\frac{b+1}{2}]. Therefore, in Definition 6.3 one can pick a(n)=1a(n)=1 for n=2n=2 and zero otherwise, with νa\nu_{a} a multiple of the Lebesgue measure on [12,b+12][\frac{1}{2},\frac{b+1}{2}].

6.2 Outline of the proof of Theorem 2.2

The rest of the section is mostly dedicated to estimating the return times to CC and to finding rates rr for which (14) holds.

Notice that b+12\frac{b+1}{2} is the preimage of bb under the second branch. Two facts play an important role in the arguments that follow: by the particular choice of CC, any random orbit starting from (0,b)(0,b) will pass through CC before landing to the right of it; points in H:=[12,b+12)H:=[\frac{1}{2},\frac{b+1}{2}) are mapped to (0,b)(0,b), i.e. to the left of the petite set CC. Any point in (b,1)(b,1) must visit HH before hitting CC.

Refer to caption
Figure 1: Schematic indicating the petite set CC and the subintervals WW and HH.

Step 1. First we show that random orbits originating from CC tend to end up to the left of CC quite fast. More precisely, given any initial condition to the right of bb, the probability that a random orbit hasn’t entered HH by time tt is exponentially small in tt. This is the content of Proposition 6.13 and Lemma 6.14 below which need several other lemmas to be proven.

Step 2. Then we study the hitting times to CC for orbits originating on (0,b)(0,b). For xx close to 0, by Lemma 5.7 the time to hit CC is of the rough order xαx^{-\alpha}. Since for zHz\in H, fγ(z)=2z1<bf_{\gamma}(z)=2z-1<b for all γ\gamma, the distribution of times to enter CC after hitting HH is determined by the distribution of positions at which HH is hit. Lemma 6.15 assembles the prior facts to show that XτHX_{\tau_{H}}, conditioned on starting at xCx\in C is absolutely continuous with density bounded above uniformly in xx. This ensures that the distribution of the time between hitting HH and entering CC has a polynomial tail which, compared to the exponential tails of the times estimated in Step 1, dominate the statistics of the returns to CC. In Proposition 6.15 we put together all the estimates and obtain subgeometric rates rr for which (14) holds.

Step 3. Finally, we apply Theorem 6.5 to prove Theorem 2.2.

6.3 Step 1

In order to show that the hitting times to HH have exponential tails, we divide (b,1)(b,1) into the intervals W:=[b,b+12)W:=[b,\frac{b+1}{2}) and [b+12,1)[\frac{b+1}{2},1). Notice that WW is such that for each y[b+12,1)y\in[\frac{b+1}{2},1), the iterates of yy under the fγf_{\gamma}’s remain in the right branch until they hit WW. We are going to show that

  • i)

    one can control the density of the random orbits originating from CC at the stopping time τW\tau_{W}, that is the first entry to WW (lemmas 6.9, 6.14);

  • ii)

    after each return to WW, at least a fixed fraction of the random orbits will enter HH and then be mapped to the left of the petite set (Lemma 6.12);

  • iii)

    the times between consecutive entries to WW before hitting HH have at most exponential tails (Lemma 6.11).

Lemma 6.7.

Let b<12b<\frac{1}{2} satisfy fα(b)>b+34f_{\alpha}(b)>\frac{b+3}{4}. There is r>0r>0 such that for every x(b,12)x\in(b,\frac{1}{2})

r(x):=(b+12,1)px(y)𝑑y>rr(x):=\int_{(\frac{b+1}{2},1)}p_{x}(y)dy>r

and r1r\rightarrow 1 when b12b\rightarrow\frac{1}{2}.

Proof.

Notice that for x[b,12)x\in[b,\frac{1}{2}), r(x)=F(γx(b+12))r(x)=F(\gamma_{x}(\frac{b+1}{2})), which is strictly positive, and approaches 1 as x12x\to\frac{1}{2}. ∎

We show that there exists C#>0C_{\#}>0 such that for all x(b,12)x\in(b,\frac{1}{2}), conditioned on X0=xX_{0}=x and X1>b+12X_{1}>\frac{b+1}{2} (so that X2>bX_{2}>b) then the distribution of XτWX_{\tau_{W}} has density bounded between C#1C_{\#}^{-1} and C#C_{\#}, where τW\tau_{W} is the time of the first return to WW. To prove this, we first need an estimate on pxp_{x}:

Lemma 6.8.

There exists KK such that for all x(b,12)x\in(b,\frac{1}{2}), and all y,y[b+12,fα(x)]y,y^{\prime}\in[\frac{b+1}{2},f_{\alpha}(x)] satisfying 2y1y<y2y-1\leq y^{\prime}<y (that is, for any γ\gamma, fγ(y)y<yf_{\gamma}(y)\leq y^{\prime}<y), one has px(y)/px(y)p_{x}(y^{\prime})/p_{x}(y) is bounded between 1K\frac{1}{K} and KK.

Proof.

We have

px(y)px(y)=(log(xyx))1+ε(yx)(log(xyx))1+ε(yx).\frac{p_{x}(y)}{p_{x}(y^{\prime})}=\frac{\big{(}\log(\frac{x}{y^{\prime}-x})\big{)}^{1+\varepsilon}(y^{\prime}-x)}{\big{(}\log(\frac{x}{y-x})\big{)}^{1+\varepsilon}(y-x)}.

Clearly the ratio (yx)/(yx)(y^{\prime}-x)/(y-x) is uniformly bounded above and below for x12x\leq\frac{1}{2} and y,y[b+12,1]y,y^{\prime}\in[\frac{b+1}{2},1] so it suffices to establish log(xyx)/log(xyx)\log(\frac{x}{y^{\prime}-x})/\log(\frac{x}{y-x}) is uniformly bounded above and below for xx, yy and yy^{\prime} as in the statement of the claim. It is clear that the numerator exceeds the denominator, so it suffices to give an upper bound.

There exist positive numbers aa and AA such that a(2u)log1u1A(2u)a(2-u)\leq\log\frac{1}{u-1}\leq A(2-u) for all u[b+1,2]u\in[b+1,2]. Applying this with uu taken to be y/xy^{\prime}/x and y/xy/x, we see that log(xyx)/log(xyx)A(2yx)/a(2yx)\log(\frac{x}{y^{\prime}-x})/\log(\frac{x}{y-x})\leq A(2-\frac{y^{\prime}}{x})/a(2-\frac{y}{x}), so that it suffices to give a uniform upper bound for (2xy)/(2xy)(2x-y^{\prime})/(2x-y). Since y2y1y^{\prime}\geq 2y-1, we see (2xy)/(2xy)(1+2x2y)/(2xy)=2+(12x)/(2xy)(2x-y^{\prime})/(2x-y)\leq(1+2x-2y)/(2x-y)=2+(1-2x)/(2x-y).

Finally 2xy2xfα(x)=x(1(2x)α)b(1(2x)α)αb(12x)2x-y\geq 2x-f_{\alpha}(x)=x(1-(2x)^{\alpha})\geq b(1-(2x)^{\alpha})\geq\alpha b(1-2x), giving the required upper bound for (2xy)/(2xy)(2x-y^{\prime})/(2x-y). ∎

Lemma 6.9.

There exists C#>1C_{\#}>1 such that for all x(b,12)x\in(b,\frac{1}{2}) the density ρx,W()\rho_{x,W}(\cdot) of XτWX_{\tau_{W}} conditioned on X0=xX_{0}=x, X1>b+12X_{1}>\frac{b+1}{2} satisfies 1C#ρx,W(y)C#\frac{1}{C}_{\#}\leq\rho_{x,W}(y)\leq C_{\#} for all yW=[b,b+12)y\in W=[b,\frac{b+1}{2}).

Proof.

We establish this by showing that there exists K>0K>0 such that for all x(b,12)x\in(b,\frac{1}{2}) and all z,z[b,b+12]z,z^{\prime}\in[b,\frac{b+1}{2}], ρx,W(z)/ρx,W(z)K\rho_{x,W}(z)/\rho_{x,W}(z^{\prime})\leq K.

We first observe that

ρx,W(z)=n=12npx(2n1+z2n)/x(X1>b+12).\rho_{x,W}(z)=\sum_{n=1}^{\infty}2^{-n}p_{x}\left(\frac{2^{n}-1+z}{2^{n}}\right)/\mathbb{P}_{x}(X_{1}>\tfrac{b+1}{2}). (16)

However, since pxp_{x} is supported on [x,fα(x)][x,f_{\alpha}(x)], there are only finitely many non-trivial terms in the sum. Also if z<zz<z^{\prime} both belong to [b,b+12][b,\frac{b+1}{2}], then zz+12z^{\prime}\leq\frac{z+1}{2} so

2n1+z2n<2n1+z2n2n+11+z2n+1.\frac{2^{n}-1+z}{2^{n}}<\frac{2^{n}-1+z^{\prime}}{2^{n}}\leq\frac{2^{n+1}-1+z}{2^{n+1}}.

Hence the number of non-trivial terms in the summation for ρx,W(z)\rho_{x,W}(z) is at least the number of terms in the summation for ρx,W(z)\rho_{x,W}(z^{\prime}) and at most one more.

Now suppose that nn is the largest number such that 2n1+z2nfα(x)\frac{2^{n}-1+z^{\prime}}{2^{n}}\leq f_{\alpha}(x). Since 2n1+z2n=2(2n+11+z2n+1)1\frac{2^{n}-1+z}{2^{n}}=2(\frac{2^{n+1}-1+z}{2^{n+1}})-1, The previous lemma establishes

px(2j1+z2j)Kpx(2j1+z2j) for j=1,,n.p_{x}\left(\frac{2^{j}-1+z}{2^{j}}\right)\leq K\cdot p_{x}\left(\frac{2^{j}-1+z^{\prime}}{2^{j}}\right)\text{\quad for $j=1,\ldots,n$.}

If 2n+11+z2n+1fα(x)\frac{2^{n+1}-1+z}{2^{n+1}}\leq f_{\alpha}(x), then since

2n1+z2n2n+11+z2n+1<2n+11+z2n+1\frac{2^{n}-1+z^{\prime}}{2^{n}}\leq\frac{2^{n+1}-1+z}{2^{n+1}}<\frac{2^{n+1}-1+z^{\prime}}{2^{n+1}}

and 2n1+z2n=2(2n+11+z2n+1)1\frac{2^{n}-1+z^{\prime}}{2^{n}}=2(\frac{2^{n+1}-1+z^{\prime}}{2^{n+1}})-1, the previous lemma implies

px(2n+11+z2n+1)Kpx(2n1+z2n).p_{x}\left(\frac{2^{n+1}-1+z}{2^{n+1}}\right)\leq K\cdot p_{x}\left(\frac{2^{n}-1+z^{\prime}}{2^{n}}\right).

Summing these inequalities yields the desired claim. ∎

The next lemma shows that if the starting density ρ\rho on (b,12)(b,\frac{1}{2}) is controlled as in equation (17) below, then the condition is invariant under the transition operator.

Lemma 6.10.

Let σ(0,1)\sigma\in(0,1). There exists b0<12b_{0}<\frac{1}{2} such that for all b(b0,12)b\in(b_{0},\frac{1}{2}) and for every density ρ\rho on (b,12)(b,\frac{1}{2}) satisfying

ρ(x)K(xb)(logbxb)1+ε,\rho(x)\leq\frac{K}{(x-b)\left(\log\frac{b}{x-b}\right)^{1+\varepsilon}}, (17)

then

Pρ(y)σK(yb)(logbyb)1+εP\rho(y)\leq\frac{\sigma K}{(y-b)\left(\log\frac{b}{y-b}\right)^{1+\varepsilon}}

for any y(b,12)y\in(b,\frac{1}{2}).

Furthermore, for any b<12b<\frac{1}{2}, there exists MM such that if ρ\rho satisfies (17), then Pρ(y)MP\rho(y)\leq M for all y[12,b+12)y\in[\frac{1}{2},\frac{b+1}{2}).

Proof.

Let y12y\leq\frac{1}{2}. We start by noticing that for any 0<s<c0<s<c

1(logcs)1+εs=dds1ε(logcs)ε.\frac{1}{\left(\log\frac{c}{s}\right)^{1+\varepsilon}s}=\frac{d}{ds}\frac{1}{\varepsilon}\left(\log\frac{c}{s}\right)^{-\varepsilon}.

Since px(y)>0p_{x}(y)>0 for any x[b,12)x\in[b,\frac{1}{2}) and any x<yx<y; and using the hypothesis on ρ\rho

b12px(y)ρ(x)𝑑x\displaystyle\int_{b}^{\frac{1}{2}}p_{x}(y)\rho(x)dx εαεbyK(log12x)ε(logxyx)1+ε(yx)(xb)(logbxb)1+ε𝑑x\displaystyle\leq\varepsilon\alpha^{\varepsilon}\int_{b}^{y}\frac{K\big{(}\log\tfrac{1}{2x}\big{)}^{\varepsilon}}{\big{(}\log\frac{x}{y-x}\big{)}^{1+\varepsilon}(y-x)(x-b)\left(\log\frac{b}{x-b}\right)^{1+\varepsilon}}dx
εαεK(log12b)εby1(logbyx)1+ε(yx)(xb)(logbxb)1+ε𝑑x.\displaystyle\leq\varepsilon\alpha^{\varepsilon}K\big{(}\log\tfrac{1}{2b}\big{)}^{\varepsilon}\int_{b}^{y}\frac{1}{\big{(}\log\frac{b}{y-x}\big{)}^{1+\varepsilon}(y-x)(x-b)\left(\log\frac{b}{x-b}\right)^{1+\varepsilon}}dx.

Set z=b+y2z=\frac{b+y}{2} and notice that the integrand is symmetric about zz. Integrating by parts, the integral becomes

I(y):=\displaystyle I(y):= 2C(ε)[(logbyx)ε(1(xb)(logbxb)1+ε)]zy\displaystyle-2C(\varepsilon)\left[\left(\log\frac{b}{y-x}\right)^{-\varepsilon}\cdot\left(\frac{1}{(x-b)(\log\frac{b}{x-b})^{1+\varepsilon}}\right)\right]_{z}^{y}
+2C(ε)zy(logbyx)εddx(1(xb)(logbxb)1+ε)𝑑x\displaystyle\quad+2C(\varepsilon)\int_{z}^{y}\left(\log\frac{b}{y-x}\right)^{-\varepsilon}\cdot\frac{d}{dx}\left(\frac{1}{(x-b)(\log\frac{b}{x-b})^{1+\varepsilon}}\right)dx
2C(ε)(logbyz)ε1(zb)(logbzb)1+ε\displaystyle\leq 2C(\varepsilon)\left(\log\frac{b}{y-z}\right)^{-\varepsilon}\cdot\frac{1}{(z-b)\left(\log\frac{b}{z-b}\right)^{1+\varepsilon}}
4C(ε)(logbyz)ε1(yb)(logbyb)1+ε\displaystyle\leq 4C(\varepsilon)\left(\log\frac{b}{y-z}\right)^{-\varepsilon}\frac{1}{(y-b)\left(\log\frac{b}{y-b}\right)^{1+\varepsilon}}
4Kαε(log12b)ε(logb12b4)ε1(yb)(logbyb)1+ε,\displaystyle\leq 4K\alpha^{\varepsilon}\big{(}\log\tfrac{1}{2b}\big{)}^{\varepsilon}\left(\log\frac{b}{\frac{1-2b}{4}}\right)^{-\varepsilon}\frac{1}{(y-b)\left(\log\frac{b}{y-b}\right)^{1+\varepsilon}},

where C(ε)=αεK(log12b)εC(\varepsilon)=\alpha^{\varepsilon}K\big{(}\log\tfrac{1}{2b}\big{)}^{\varepsilon} and where, in the first inequality, we used the fact that derivative in the second line is negative for all x[b,12]x\in[b,\frac{1}{2}] for all bb sufficiently close to 12\frac{1}{2}. Since

limb124(log12b)ε(logb12b4)ε=0,\lim_{b\rightarrow\frac{1}{2}}4\big{(}\log\tfrac{1}{2b}\big{)}^{\varepsilon}\left(\log\frac{b}{\frac{1-2b}{4}}\right)^{-\varepsilon}=0,

the first conclusion follows.

For y12y\geq\frac{1}{2},

I(y)2εC(ε)b121(logbyx)1+ε(yx)(xb)(logbxb)1+ε𝑑x.I(y)\leq 2\varepsilon C(\varepsilon)\int_{b}^{\frac{1}{2}}\frac{1}{\big{(}\log\frac{b}{y-x}\big{)}^{1+\varepsilon}(y-x)(x-b)\left(\log\frac{b}{x-b}\right)^{1+\varepsilon}}dx.

For any c(b,12)c\in(b,\frac{1}{2}), 1/((logbyx)1+ε(yx))1/((\log\frac{b}{y-x})^{1+\varepsilon}(y-x)) is uniformly bounded above for x[b,c]x\in[b,c] and y[12,b+12]y\in[\frac{1}{2},\frac{b+1}{2}] and 1/((xb)(logbxb)1+ε)1/((x-b)(\log\frac{b}{x-b})^{1+\varepsilon}) is integrable. Similarly, on [c,12][c,\frac{1}{2}], 1/((xb)(logbxb)1+ε)1/((x-b)(\log\frac{b}{x-b})^{1+\varepsilon}) is bounded above and the functions 1/((logbyx)1+ε(yx))1/((\log\frac{b}{y-x})^{1+\varepsilon}(y-x)) for y[12,b+12]y\in[\frac{1}{2},\frac{b+1}{2}] are integrable over [c,12][c,\frac{1}{2}] with integral uniformly bounded in yy, so that the second conclusion holds. ∎

The following lemma shows that the gaps between consecutive entries to W=[b,b+12)W=[b,\frac{b+1}{2}) are dominated by a random variable ZZ with an exponential tail. If X0WX_{0}\in W, we define τW,i\tau_{W,i} to be the time of the iith re-entry to WW. That is τW,0=0\tau_{W,0}=0 and τW,i+1=min{n>τW,i:Xn1W,XnW}\tau_{W,i+1}=\min\{n>\tau_{W,i}\colon X_{n-1}\not\in W,X_{n}\in W\}. Recall that H=[12,b+12)H=[\frac{1}{2},\frac{b+1}{2}) is the subset mapped to the left of the petite set. We study the distribution of some random variables conditioned on the event {τW,1τH}\{\tau_{W,1}\leq\tau_{H}\}. Notice that since HWH\subset W, this is the event that the process leaves WW before first hitting HH.

Lemma 6.11.

There exists an integer-valued random variable ZZ such that for each absolutely continuous distribution on (b,12)(b,\frac{1}{2}) with density ρ\rho bounded as in (17), there exists a random variable YρY_{\rho} such that the following properties are satisfied:

  1. 1.

    If X0X_{0} is continuously distributed on [b,12)[b,\frac{1}{2}) with density ρ\rho, then conditional on {τW,1τH}\{\tau_{W,1}\leq\tau_{H}\}, τW,1Yρ\tau_{W,1}\leq Y_{\rho}.

  2. 2.

    For all n>0n>0 and all ρ\rho satisfying (17),

    ρ(Yρn|τW,1τH)(Zn);\mathbb{P}_{\rho}(Y_{\rho}\geq n|\tau_{W,1}\leq\tau_{H})\leq\mathbb{P}(Z\geq n);
  3. 3.

    There exists a>0a>0, such that 𝔼eaZ<\mathbb{E}e^{aZ}<\infty;

  4. 4.

    Let X0X_{0} be absolutely continuously distributed on [b,12][b,\frac{1}{2}] with density ρ\rho satisfying (17). For any n2n\geq 2, conditioned on {τW,1τH}\{\tau_{W,1}\leq\tau_{H}\}; and given that Yρ=nY_{\rho}=n, the distribution of XτW,1X_{\tau_{W,1}}, the position at which the system reenters WW, is absolutely continuous with density bounded above and below by constants that do not depend on ρ\rho or nn.

Proof.

Let ρ\rho be a probability density on [b,12][b,\frac{1}{2}] satisfying (17). Let X0X_{0} be distributed with density ρ\rho. Notice that by Lemma 6.7, the event {τW,1τH}\{\tau_{W,1}\leq\tau_{H}\}, that is that the system leaves WW before entering HH, has probability bounded below by a constant r>0r>0. We define YY by

Yρ={0if τH<τW,1;τW,1+𝖦𝖾𝗈𝗆(12)otherwise,Y_{\rho}=\begin{cases}0&\text{if $\tau_{H}<\tau_{W,1}$;}\\ \tau_{W,1}+\mathsf{Geom}(\tfrac{1}{2})&\text{otherwise,}\end{cases}

where 𝖦𝖾𝗈𝗆(12)\mathsf{Geom}(\tfrac{1}{2}) denotes an independent geometric random variable with parameter 12\frac{1}{2}, so that conclusion 1 is evident. To establish conclusions 2 and 3, it suffices to show that there exists c>0c>0 such that for all ρ\rho satisfying (17), one has ρ(τW,1n)enc\mathbb{P}_{\rho}(\tau_{W,1}\geq n)\leq e^{-nc} for all nn. It then follows that there exists c>0c^{\prime}>0 such that (Yρn)enc\mathbb{P}(Y_{\rho}\geq n)\leq e^{-nc^{\prime}} for all nn. Then ZZ can be defined by Z=nZ=n with probability enc/re^{-nc^{\prime}}/r for all n>n0n>n_{0} where n0n_{0} is chosen so that n>n0enc/r<1\sum_{n>n_{0}}e^{-nc^{\prime}}/r<1; and Z=n0Z=n_{0} with probability 1n>n0enc/r1-\sum_{n>n_{0}}e^{-nc^{\prime}}/r.

Notice that in order that τW,13n\tau_{W,1}\geq 3n, at least one of the following must occur: the system must remain in WW for nn steps; the system must exit WW to a point above (b+2n1)/2n(b+2^{n}-1)/2^{n} (so that it takes nn or more steps to re-enter WW); or the geometric random variable must take a value of nn or above. Then (τW,13n)\mathbb{P}(\tau_{W,1}\geq 3n) is at most the sum of these three probabilities. The first of these has probability at most (1r)n(1-r)^{n} by Lemma 6.7. The third event has probability 2n2^{-n}.

For the second event, note that for x[b,12]x\in[b,\frac{1}{2}], it is only possible that fγ(x)(b+2n1)/2nf_{\gamma}(x)\geq(b+2^{n}-1)/2^{n} if x>(b+2n1)/2n+1x>(b+2^{n}-1)/2^{n+1}, that is, if x(121b2n+1,12)x\in(\frac{1}{2}-\frac{1-b}{2^{n+1}},\frac{1}{2}). Define the operator P[b,12]P_{[b,\frac{1}{2}]}, mapping L1([b,12])L^{1}([b,\frac{1}{2}]) to itself by P[b,12]f(x)=𝟏[b,12](x)Pf(x)P_{[b,\frac{1}{2}]}f(x)=\mathbf{1}_{[b,\frac{1}{2}]}(x)Pf(x). By Lemma 6.10, the function g(x)=n=1P[b,12]n(ρ)(x)g(x)=\sum_{n=1}^{\infty}P_{[b,\frac{1}{2}]}^{n}(\rho)(x) satisfies

g(x)11σK(xb)(logbxb)1+ε,g(x)\leq\frac{1}{1-\sigma}\cdot\frac{K}{(x-b)\left(\log\frac{b}{x-b}\right)^{1+\varepsilon}}, (18)

which is bounded above in a neighbourhood of 12\frac{1}{2} by some number cc^{\prime} (which does not depend on the initial distribution). Hence (τ[12h,12]<τWc)<hc\mathbb{P}(\tau_{[\frac{1}{2}-h,\frac{1}{2}]}<\tau_{W^{c}})<hc^{\prime} for all small hh. In particular, the probability of hitting (121b2n+2,12)(\frac{1}{2}-\frac{1-b}{2^{n+2}},\frac{1}{2}) is bounded above by a constant multiple of 2n2^{-n}, where the constant does not depend on ρ\rho.

To establish conclusion 4, suppose we are given that τW,1τH\tau_{W,1}\leq\tau_{H} and Yρ=nY_{\rho}=n. We additionally condition on the time taken for the system to exit WW and the location of the system prior to exiting WW. We establish bounds on the density of XτW,1X_{\tau_{W,1}} based on this additional information. Then the bounds without this additional conditioning are simply a convex combination of these bounds.

Thus suppose that the system exits WW for the first time at time kk, and we are given that Xk1=x[b,12]X_{k-1}=x\in[b,\frac{1}{2}]. Since Yρ=nY_{\rho}=n, the number of steps to reenter WW after leaving is one of 1, 2, …, nkn-k. That is XkX_{k} belongs to one of the intervals

Ij=(b+2j12j,b+2nj+112nj+1],I_{j}=\left(\frac{b+2^{j}-1}{2^{j}},\frac{b+2^{n-j+1}-1}{2^{n-j+1}}\right],

for jj in the range 1 to min{nk,l}\min\{n-k,l\} where ll is such that fα(x)Ilf_{\alpha}(x)\in I_{l}. Since fα(b)>b+34f_{\alpha}(b)>\frac{b+3}{4}, we have l3l\geq 3. Since typically fα(x)<maxIlf_{\alpha}(x)<\max I_{l}, one knows that XkX_{k} may only occupy a (possibly small, depending on xx) portion of IlI_{l}, namely Il=Il[0,fα(x))I_{l}^{-}=I_{l}\cap[0,f_{\alpha}(x)). Hence if τW,1=k+l\tau_{W,1}=k+l, one sees that XτW,1X_{\tau_{W,1}} is restricted to a possibly small sub-interval of WW (and therefore its conditional density may not be bounded away from zero). This is the reason that we introduced the geometric random variable: to ensure that the return time variable, YρY_{\rho} does not completely determine τW,1\tau_{W,1} and thereby overly constrain the location of XτW,1X_{\tau_{W,1}}.

Conditioned on Xk1=xX_{k-1}=x, the event {Xkb+12,τW,1=k+j}\{X_{k}\geq\frac{b+1}{2},\tau_{W,1}=k+j\} has probability P(x,Ij)P(x,I_{j}) for j=1,,l1j=1,\ldots,l-1 and {Xkb+12,τW,1=k+l}\{X_{k}\geq\frac{b+1}{2},\tau_{W,1}=k+l\} has probability P(x,Il)P(x,I_{l}^{-}). We therefore have (Yρ=n,Xk>b+12|Xk1=x)\mathbb{P}(Y_{\rho}=n,X_{k}>\tfrac{b+1}{2}|X_{k-1}=x) is

{j=1l12(nkj+1)P(x,Ij)+2(nkl+1)P(x,Il)if nk+l;j=1nk12(nkj+1)P(x,Ij)otherwise.\begin{cases}\sum_{j=1}^{l-1}2^{-(n-k-j+1)}P(x,I_{j})+2^{-(n-k-l+1)}P(x,I_{l}^{-})&\text{if $n\geq k+l$;}\\ \sum_{j=1}^{n-k-1}2^{-(n-k-j+1)}P(x,I_{j})&\text{otherwise.}\end{cases}

Now an application of Bayes’ theorem, together with Lemma 6.8, shows that conditioned on Xk1=xX_{k-1}=x, Xkb+12X_{k}\geq\frac{b+1}{2} and Y=nY=n, the distribution of XkX_{k} on each interval above is uniform up to a multiplicative factor of the fixed constant KK, except for IlI_{l} in which the density drops off to 0, but has the property that on IlI_{l}^{-}, the density is within a factor of KK of that on Il1I_{l-1}. Since in the next τW,1k\tau_{W,1}-k steps, the interval is mapped linearly onto [b,b+12][b,\frac{b+1}{2}], the distribution of XτW,1X_{\tau_{W,1}} is uniform up to a multiplicative factor of 2K2K on [b,b+12][b,\frac{b+1}{2}]. ∎

The following lemma states that after each return to the set WW, at least a fixed positive proportion of the mass ends up in the subset H=[12,b+12)H=[\frac{1}{2},\frac{b+1}{2}) of WW. Recall that τW,i\tau_{W,i} is defined to be the time of the iith reentry to WW.

Lemma 6.12.

There exists q>0q^{\prime}>0 such that for every x(b,b+12)x\in(b,\frac{b+1}{2})

x(τHτW,1)>q.\mathbb{P}_{x}\left(\tau_{H}\leq\tau_{W,1}\right)>q^{\prime}.
Proof.

If x[12,b+12)x\in[\frac{1}{2},\frac{b+1}{2}), then τH=0\tau_{H}=0 and τW,1>0\tau_{W,1}>0 so that x(τHτW,1)=1\mathbb{P}_{x}(\tau_{H}\leq\tau_{W,1})=1. If x<12x<\frac{1}{2} then by point (i) in Lemma 6.7,

x(X1>b+12)>r.\mathbb{P}_{x}\left(X_{1}>\tfrac{b+1}{2}\right)>r.

It then follows from Lemma 6.9 that XτW,1X_{\tau_{W,1}} is absolutely continuous with density at least rC1rC^{-1}. In particular, x(XτW,1H)>rbC\mathbb{P}_{x}(X_{\tau_{W,1}}\in H)>\frac{rb}{C}, so that x(τHτW,1)rbC\mathbb{P}_{x}(\tau_{H}\leq\tau_{W,1})\geq\frac{rb}{C}. ∎

We now show that the entry time into HH has exponential tails.

Proposition 6.13.

Given any probability measure μ\mu with density ρ\rho satisfying (17), there is c>0c^{\prime}>0 such that for any tt\in\mathbb{N},

μ(τH>t)ect.\mathbb{P}_{\mu}(\tau_{H}>t)\leq e^{-c^{\prime}t}.
Proof.

First of all define random variables Yi:=τW,iτW,i1Y_{i}:=\tau_{W,i}-\tau_{W,i-1}, and the random variable NN equal to the number of reentries to WW up to the first entry to HH (recalling that HWH\subset W). With these definitions, the time to enter HH is given by τH=i=1NYi+V\tau_{H}=\sum_{i=1}^{N}Y_{i}+V, where VV is the number of steps spent in [b,12][b,\frac{1}{2}] after the NNth reentry to WW prior to entering HH (note that VV may be 0 if the first point of WW that the system enters on the NNth visit is in HH).

Notice that if the constant KK in Lemma 6.10 is chosen sufficiently large, then conclusion 4 of Lemma 6.11 shows that conditional on {τW,1<τH}\{\tau_{W,1}<\tau_{H}\}, the density of XτW,1X_{\tau_{W,1}} on [b,12][b,\frac{1}{2}] satisfies (17).

Lemma 6.11 and the Markov property, together with Strassen’s theorem [14, Theorem 2.4], imply that the sequence of random variables (Yi)(Y_{i}) may be coupled with a sequence (Zi)(Z_{i}) of independent identically distributed random variables with exponential tails in such a way that YiZiY_{i}\leq Z_{i} for i=1,,Ni=1,\ldots,N. Lemma 6.12 implies that NN also has exponential tails.

Let m>𝔼Zm>\mathbb{E}Z. A Chernoff bound (see [8, Section 2.7]) shows that there exists δ>0\delta>0 such that

μ(i=1nZi>mn)eδn.\mathbb{P}_{\mu}\left(\sum_{i=1}^{n}Z_{i}>mn\right)\leq e^{-\delta n}. (19)

The event {τH>t}\{\tau_{H}>t\} is a subset of {V>t2}{N>t2m}{Nt2m,Z1++ZN>t2}\{V>\frac{t}{2}\}\cup\{N>\frac{t}{2m}\}\cup\{N\leq\frac{t}{2m},\,Z_{1}+\ldots+Z_{N}>\frac{t}{2}\} and this is a subset of {V>t2}{N>t2m}{Z1++Zt2m>mt2m}\{V>\frac{t}{2}\}\cup\{N>\frac{t}{2m}\}\cup\{Z_{1}+\ldots+Z_{\lfloor\frac{t}{2m}\rfloor}>m\lfloor\frac{t}{2m}\rfloor\}. For the first set (V>t/2)(1r)t/2\mathbb{P}(V>t/2)\leq(1-r)^{\lfloor t/2\rfloor} by Lemma 6.7. For the second set,

μ(N>t2m)ec2t2m,\mathbb{P}_{\mu}\left(N>\tfrac{t}{2m}\right)\leq e^{-c_{2}\frac{t}{2m}},

since NN has exponential tails, and for the third

μ(Z1++Zt2m>mt2m)eδt2m\mathbb{P}_{\mu}\left(Z_{1}+\ldots+Z_{\lfloor\frac{t}{2m}\rfloor}>m\lfloor\tfrac{t}{2m}\rfloor\right)\leq e^{-\delta\frac{t}{2m}}

from (19). ∎

The following lemma proves that Proposition 6.13 can be applied to the density of XτWX_{\tau_{W}}.

Lemma 6.14.

There exists a<1a<1 and n0n_{0} such that for all xx in the petite set C=[fα1b,b)C=[f_{\alpha}^{-1}b,b), one has x(τW>n)an\mathbb{P}_{x}(\tau_{W}>n)\leq a^{n} for all nn0n\geq n_{0}

There exists a K>0K>0 such that for all xx in the petite set CC, the distribution of XτWX_{\tau_{W}}, the position at the first entrance to WW, is absolutely continuous with density satisfying

ρ(y)K(logbyb)1+ε(yb)\rho(y)\leq\frac{K}{(\log\frac{b}{y-b})^{1+\varepsilon}(y-b)} (20)

for all yWy\in W (that is, the density satisfies the condition (17) extended to the full interval WW).

Proof.

Let d(fα1b,b)d\in(f_{\alpha}^{-1}b,b) be chosen so that fα(d)<12f_{\alpha}(d)<\frac{1}{2}. For any x[fα1b,d)x\in[f_{\alpha}^{-1}b,d), the time to enter [d,fα(d)][d,f_{\alpha}(d)] is bounded above by a geometric random variable (with parameter not depending on xx since P(x,[d,fα(d)])P(fα1b,[d,fα(d)])>0P(x,[d,f_{\alpha}(d)])\geq P(f_{\alpha}^{-1}b,[d,f_{\alpha}(d)])>0 for all x[fα1b,d]x\in[f_{\alpha}^{-1}b,d]). For any x[d,b)x\in[d,b), the time to enter [b,fα(b)][b,f_{\alpha}(b)] is bounded above by another geometric random variable by an identical argument. On the interval (b+12,fα(b)](\frac{b+1}{2},f_{\alpha}(b)], the time to enter WW is uniformly bounded above. Summing these contributions gives the required geometric upper bounds on τW\tau_{W}.

We look at the distribution of XτWX_{\tau_{W}}. If x[fα1b,d]x\in[f_{\alpha}^{-1}b,d], then we study

n=1𝟏Cc(P𝟏[fα1b,d])nδx,\sum_{n=1}^{\infty}\mathbf{1}_{C^{c}}(P\mathbf{1}_{[f_{\alpha}^{-1}b,d]})^{n}\delta_{x}, (21)

the contribution to the density on WW coming from points that stay in [fα1b,d][f_{\alpha}^{-1}b,d] until they enter CcC^{c} (necessarily into WW since fα(d)<12f_{\alpha}(d)<\frac{1}{2}). (Note that this contribution is trivial if x>dx>d.)

If x[fα1b,d]x\in[f_{\alpha}^{-1}b,d], we showed above that the number of steps before leaving the interval is bounded above by a geometric random variable; and px(y)p_{x}(y) is bounded above for (x,y)[fα1b,d]×W(x,y)\in[f_{\alpha}^{-1}b,d]\times W. Combining these, we see that the density of the contribution in (21) is uniformly bounded above.

For the remaining part of the distribution, we condition on the last point of [d,b)[d,b) that is visited. We show that for all x[d,b)x\in[d,b), conditional on leaving CC in a single step, the density of XτWX_{\tau_{W}} conditioned on X0=xX_{0}=x and X1bX_{1}\geq b satisfies a bound of the form (20). First note, that for x[d,b)x\in[d,b) and yb+12y\geq\frac{b+1}{2}, px(y)p_{x}(y) is uniformly bounded above, so that when it is mapped under iteration of the second branch back to WW, it gives a contribution that is uniformly bounded above (similar to (16)). It therefore suffices to show that for x[d,b]x\in[d,b], px(y|yW)p_{x}(y|y\in W) satisfies a bound of the form (20) with a KK that does not depend on xx. Since the probability of entering WW in one step is bounded below for these xx’s, it suffices to show the existence of a KK such that for all x[d,b]x\in[d,b] and all yWy\in W, px(y)K/[(yb)(logbyb)1+ε]p_{x}(y)\leq K/[(y-b)(\log\frac{b}{y-b})^{1+\varepsilon}]. For xx in a small interval [bδ,b][b-\delta,b] and yy in a small interval [b,b+δ][b,b+\delta], one may check that px(y)p_{x}(y) is increasing in xx, so that px(y)pb(y)p_{x}(y)\leq p_{b}(y) as required. If either xx or yy lies outside this range, there is a uniform upper bound on px(y)p_{x}(y), establishing the required inequality. ∎

6.4 Step 2

Finally, we verify condition (14) of Theorem 6.5.

Proposition 6.15.

Suppose that 0<α<10<\alpha<1 is the minimum of the support of the measure ν\nu and let 1<γ<1α1<\gamma<\frac{1}{\alpha}. For every xx belonging to the petite set C=[fα1b,b]C=[f_{\alpha}^{-1}b,b],

supxC𝔼x[τCγ]<.\sup_{x\in C}\mathbb{E}_{x}\left[\tau_{C}^{\gamma}\right]<\infty.
Proof.

Notice that we can bound the return time to the petite set τC\tau_{C} as τCτW+τWH+τHC\tau_{C}\leq\tau_{W}+\tau_{W\to H}+\tau_{H\rightarrow C} where τW\tau_{W} is the first time to hit WW, τWH\tau_{W\to H} is the random time after then that it takes to hit HH (this may be 0), and τHC\tau_{H\rightarrow C} is the random time needed to go back to CC starting from when HH is hit. It follows from Minkowski’s inequality that if supxC𝔼x[τWγ]\sup_{x\in C}\mathbb{E}_{x}[\tau_{W}^{\gamma}], supxC𝔼x[τWHγ]\sup_{x\in C}\mathbb{E}_{x}[\tau_{W\to H}^{\gamma}] and supxC𝔼x[τHCγ]\sup_{x\in C}\mathbb{E}_{x}[\tau_{H\rightarrow C}^{\gamma}] are finite, then also supxC𝔼x[τCγ]\sup_{x\in C}\mathbb{E}_{x}[\tau_{C}^{\gamma}] is finite.

By Lemma 6.14, supxC𝔼x[τWγ]<\sup_{x\in C}\mathbb{E}_{x}[\tau_{W}^{\gamma}]<\infty. To show that supxC𝔼x[τWHγ]<\sup_{x\in C}\mathbb{E}_{x}[\tau_{W\to H}^{\gamma}]<\infty, Lemma 6.14 shows that the distribution of XτWX_{\tau_{W}} satisfies (20) and the conclusion follows from Proposition 6.13.

To estimate 𝔼xτHCγ\mathbb{E}_{x}\tau_{H\to C}^{\gamma}, we need to control the distribution of XτHX_{\tau_{H}}. The system may enter HH in a number of ways: from CC without previously entering WW; by making a number of visits to WW and then entering direct from WW; or by making a number of visits to WW and then entering from [b+12,1][\frac{b+1}{2},1]. For direct entry from CC, conditioning on the last point visited in CC, the density on HH is uniformly bounded above. Similarly, conditioning on the last point visited in WW before entering HH from the right, Lemma 6.9 gives a uniform upper bound on the contribution to the density of XτHX_{\tau_{H}}. For the points entering from WW, Lemmas 6.9 and 6.14 ensure that on entry to [b,12)[b,\frac{1}{2}), the density satisfies (17) (with a fixed KK). The sum of the densities prior to exiting [b,12)[b,\frac{1}{2}) is estimated in (18) and this gives a bound on the density of the last position in WW before exiting, which is 1/(1σ)1/(1-\sigma) times the bound in (17). Then the second part of Lemma 6.10 gives a uniform upper bound on this contribution to the density on HH of XτHX_{\tau_{H}}. Taken together, we have shown that the distribution of XτHX_{\tau_{H}} (and therefore the density of its image on [0,b)[0,b)) is absolutely continuous with density bounded above by a constant that is independent of xx. Since for z[0,b)z\in[0,b), τCτ[12,1]\tau_{C}\leq\tau_{[\frac{1}{2},1]}, and Proposition 4.1 established condition (C2), we see 𝔼xτHCγ<\mathbb{E}_{x}\tau_{H\to C}^{\gamma}<\infty as required. ∎

6.5 Step 3

Proof of Theorem 2.2.

Let 1<γ<1α1<\gamma<\frac{1}{\alpha} and let r(n)=nγ1r(n)=n^{\gamma-1}, so that r(n)r(n) is a subgeometric sequence with j=0n1r(n)=(1+o(1))nγγ\sum_{j=0}^{n-1}r(n)=(1+o(1))\frac{n^{\gamma}}{\gamma}. Now Proposition 6.15 shows that the hypotheses of Theorem 6.5 are satisfied for this sequence r(n)r(n), so that

Pn(x,)πTV𝑑m(x)=o(n1γ).\int\|P^{n}(x,\cdot)-\pi\|_{\text{TV}}\,dm(x)=o(n^{1-\gamma}).

Finally, for φ,ψL([0,1])\varphi,\psi\in L^{\infty}([0,1]),

|φ(x)ψ(f𝝎nx)𝑑m(x)𝑑να,ε(𝝎)φ(x)𝑑m(x)ψ(y)𝑑π(y)|\displaystyle\left|\int\int\varphi(x)\psi(f_{\boldsymbol{\omega}}^{n}x)\,dm(x)\,d\nu_{\alpha,\varepsilon}^{\mathbb{N}}(\boldsymbol{\omega})-\int\varphi(x)\,dm(x)\int\psi(y)\,d\pi(y)\right|
=|φ(x)(ψ(y)Pn(x,dy)ψ(y)𝑑π(y))𝑑m(x)|\displaystyle=\left|\int\varphi(x)\left(\int\psi(y)\,P^{n}(x,dy)-\int\psi(y)\,d\pi(y)\right)\,dm(x)\right|
φψPn(x,)πTV𝑑m(x)=o(n1γ).\displaystyle\leq\|\varphi\|_{\infty}\|\psi\|_{\infty}\int\|P^{n}(x,\cdot)-\pi\|_{\text{TV}}\,dm(x)=o(n^{1-\gamma}).

References

  • AD [01] J. Aaronson and M. Denker, Local limit theorems for partial sums of stationary random sequences generated by Gibbs-Markov maps, Stoch. Dyn. (2001), 193–237.
  • AHN+ [15] R. Aimino, H. Hu, M. Nicol, S. Vaienti, et al., Polynomial loss of memory for maps of the interval with a neutral fixed point, Discrete and Continuous Dynamical Systems (2015), no. 35(3), 793–806.
  • BB [16] W. Bahsoun and C. Bose, Mixing rates and limit theorems for random intermittent maps, Nonlinearity 29 (2016), 1417.
  • BBD [14] W. Bahsoun, C. Bose, and Y. Duan, Decay of correlation for random intermittent maps, Nonlinearity 27 (2014), 1543.
  • BBMD [02] V. Baladi, M. Benedicks, and V. Maume-Deschamps, Almost sure rates of mixing for iid unimodal maps, Annales scientifiques de l’école normale supérieure, vol. 35, Elsevier, 2002, pp. 77–126.
  • BBR [19] W. Bahsoun, C. Bose, and M. Ruziboev, Quenched decay of correlations for slowly mixing systems, Transactions of the American Mathematical Society 372 (2019), no. 9, 6547–6587.
  • BGT [87] N. Bingham, C. Goldie, and J. Teugels, Regular variation, Encyclopedia of mathematics and its applications, Cambridge, 1987.
  • Dur [19] R. Durrett, Probability theory and examples, 5th edition ed., Cambridge, 2019.
  • [9] S. Gouëzel, Central limit theorem and stable laws for intermittent maps, Probability Theory and Related Fields 128 (2004), no. 1, 82–122.
  • [10] S. Gouëzel, Sharp polynomial estimates for the decay of correlations, Israel J. Math. 139 (2004), 29–65.
  • Gou [07]  , Statistical properties of a skew product with a curve of neutral points, Erg. Thry. Dyn. Sys. 27 (2007), 123–151.
  • Gou [15] S. Gouëzel, Limit theorems in dynamical systems using the spectral method, Hyperbolic dynamics, fluctuations and large deviations, vol. 89, Amer. Math. Soc., Providence, RI, 2015, pp. 161–193.
  • Hen [93] H. Hennion, Sur un théorème spectral et son application aux noyaux lipchitziens, Proc. Amer. Math. Soc. 118 (1993), 627–634.
  • Lig [12] T. M. Liggett, Interacting particle systems, vol. 276, Springer Science &amp; Business Media, 2012.
  • Liv [95] C. Liverani, Decay of correlations, Annals of Mathematics-Second Series 142 (1995), no. 2, 239–302.
  • Liv [96]  , Central limit theorem for deterministic systems, International Conference on Dynamical Systems (Montevideo, 1995), vol. 362, 1996, pp. 56–75.
  • LSV [99] C. Liverani, B. Saussol, and S. Vaienti, A probabilistic approach to intermittency, Ergodic Theory Dynam. Systems 19 (1999), 671–685.
  • NP [19] M. Nicol and F. Perez Pereira, Large deviations and central limit theorems for sequential and random systems of intermittent maps, arXiv preprint arXiv:1909.07435 (2019).
  • NTV [18] M. Nicol, A. Török, and S. Vaienti, Central limit theorems for sequential and random intermittent dynamical systems, Ergodic Theory Dynam. Systems 38 (2018), 1127–1153.
  • PM [80] Y. Pomeau and P. Manneville, Intermittent transition to turbulence in dissipative dynamical systems, Comm. Math. Phys. 74 (1980), no. 2, 189–197.
  • Sar [02] O. Sarig, Subexponential decay of correlations, Invent. Math. 150 (2002), 629–653.
  • TT [94] P. Tuominen and R. L. Tweedie, Subgeometric rates of convergence of ff-ergodic Markov chains, Adv. Appl. Prob. 26 (1994), 775–798.
  • You [99] L.-S. Young, Recurrence times and rates of mixing, Isr. J. Math. 110 (1999), 153–188.