theoremdummy \aliascntresetthetheorem \newaliascntpropositiondummy \aliascntresettheproposition \newaliascntcorollarydummy \aliascntresetthecorollary \newaliascntlemmadummy \aliascntresetthelemma \newaliascntdefinitiondummy \aliascntresetthedefinition \newaliascntexampledummy \aliascntresettheexample \newaliascntassumptiondummy \aliascntresettheassumption \newaliascntremarkdummy \aliascntresettheremark
Counterbalancing steps at random
in a random walk
Abstract
A random walk with counterbalanced steps is a process of partial sums whose steps are given recursively as follows. For each , with a fixed probability , is a new independent sample from some fixed law , and with complementary probability , counterbalances a previous step, with a uniform random pick from . We determine the asymptotic behavior of in terms of and the first two moments of . Our approach relies on a coupling with a reinforcement algorithm due to H.A. Simon, and on properties of random recursive trees and Eulerian numbers, which may be of independent interest. The method can be adapted to the situation where the step distribution belongs to the domain of attraction of a stable law.
Keywords:
Reinforcement, random walk, random recursive tree, Eulerian numbers, Yule-Simon model.
Mathematics Subject Classification: 60F05; 60G50; 05C05; 05A05.
1 Introduction
In short, the purpose of the present work is to investigate long time effects of an algorithm for counterbalancing steps in a random walk. As we shall first explain, our motivation stems from a nearest neighbor process on the integer lattice, known as the elephant random walk.
The elephant random walk is a stochastic process with memory on , which records the trajectory of an elephant that makes steps with unit length left or right at each positive integer time. It has been introduced by Schütz and Trimper [31] and triggered a growing interest in the recent years; see, for instance, [4, 5, 15, 16, 22, 23], and also [2, 6, 8, 14, 20, 26] for related models. The dynamics depend on a parameter and can be described as follows. Let us assume that the first step of the elephant is a Rademacher variable, that is equals or with probability . For each time , the elephant remembers a step picked uniformly at random among those it made previously, and decides either to repeat it with probability , or to make the opposite step with complementary probability . Obviously, each step of the elephant has then the Rademacher law, although the sequence of steps is clearly not stationary.
Roughly speaking, it seems natural to generalize these dynamics and allow steps to have an arbitrary distribution on , say . In this direction, Kürsten [23] pointed at an equivalent way of describing the dynamics of the elephant random walk which makes such generalization non trivial222 Note that merely replacing the Rademacher distribution for the first step of the elephant by would not be interesting, as one would then just get the evolution of the elephant random walk multiplied by some random factor with law .. Let , and imagine a walker who makes at each time a step which is either, with probability , a new independent random variable with law , or, with probability , a repetition of one of his preceding steps picked uniformly at random. It is immediately checked that when is the Rademacher distribution, then the walker follows the same dynamics as the elephant random walk with parameter . When is an isotropic stable law, this is the model referred to as the shark random swim by Businger [14], and more generally, when is arbitrary, this is the step reinforced random walk that has been studied lately in e.g. [7, 9, 10, 11].
The model of Kürsten yields an elephant random walk only with parameter ; nonetheless the remaining range can be obtained by a simple modification. Indeed, let again and imagine now an repentant walker who makes at each time a step which is either, with probability , a new independent random variable with law , or, with probability , the opposite of one of his previous steps picked uniformly at random. When is the Rademacher distribution, we simply get the dynamics of the elephant random walk with parameter .
More formally, we consider a sequence of i.i.d. real random variables with some given law and a sequence of i.i.d. Bernoulli variables with parameter , which we assume furthermore independent of . We construct a counterbalanced sequence by interpreting each as a counterbalancing event and each as an innovation event. Specifically, we agree that for definiteness and denote the number of innovations after steps by
We introduce a sequence of independent variables, where each has the uniform distribution on , and which is also independent of and . We then define recursively
(1) |
Note that the same step can be counterbalanced several times, and also that certain steps counterbalance previous steps which in turn already counterbalanced earlier ones. The process
which records the positions of the repentant walker as a function of time, is called here a random walk with counterbalanced steps. Note that for , i.e. when no counterbalancing events occur, is just a usual random walk with i.i.d. steps.
In short, we are interested in understanding how counterbalancing steps affects the asymptotic behavior of random walks. We first introduce some notation. Recall that denotes the distribution of the first step and write
for the moment of order of whenever . To start with, we point out that if the first moment is finite, then the algorithm (1) yields the recursive equation
with the initial condition . It follows easily that
see e.g. Lemma 4.1.2 in [18]. Our first result about the ballistic behavior should therefore not come as a surprise.
Proposition \theproposition.
Let . If , then there is the convergence in probability
We see in particular that counterbalancing steps reduces the asymptotic velocity of a random walk by a factor . The velocity is smaller when the innovation rate is smaller (i.e. when counterbalancing events have a higher frequency), and vanishes as approaches .
The main purpose of this work is to establish the asymptotic normality when has a finite second moment.
Theorem \thetheorem.
Let . If , then there is the convergence in distribution
where the right-hand side denotes a centered Gaussian variable parametrized by mean and variance.
It is interesting to observe that the variance of the Gaussian limit depends linearly on the square of the first moment and the second moment of only, although not just on the variance (except, of course, for ). Furthermore, it is not always a monotonous function333 For instance, in the simplest case when is a Dirac point mass, i.e. , then the variance is given by and reaches its maximum for . At the opposite, when is centered, i.e. , the variance is given by and hence increases with . of the innovation rate , and does not vanish when tends to either.
Actually, our proofs of Proposition 1 and Theorem 1 provide a much finer analysis than what is encapsulated by those general statements. Indeed, we shall identify the main actors for the evolution of and their respective contributions to its asymptotic behavior. In short, we shall see that the ballistic behavior stems from those of the variables that have been used just once by the algorithm (1) (in particular, they have not yet been counterbalanced), whereas the impact of variables that occurred twice or more regardless of their signs , is asymptotically negligible as far as only velocity is concerned. Asymptotic normality is more delicate to analyze. We shall show that, roughly speaking, it results from the combination of, on the one hand, the central limit theorem for certain centered random walks, and on the other hand, Gaussian fluctuations for the asymptotic frequencies of some pattern induced by (1).
Our analysis relies on a natural coupling of the counterbalancing algorithm (1) with a basic linear reinforcement algorithm which has been introduced a long time ago by H.A. Simon [32] to explain the occurrence of certain heavy tailed distributions in a variety of empirical data. Specifically, Simon defined recursively a sequence denoted here by (beware of the difference of notation between and ) via
(2) |
We stress that the same Bernoulli variables and the same uniform variables are used to run both Simon’s algorithm (2) and (1); in particular either or . It might then seem natural to refer to (1) and (2) respectively as negative and positive reinforcement algorithms. However, in the literature, negative reinforcement usually refers to a somehow different notion (see e.g. [29]), and we shall thus avoid using this terminology.
A key observation is that (1) can be recovered from (2) as follows. Simon’s algorithm naturally encodes a genealogical forest with set of vertices and edges for all with ; see the forthcoming Figure 1. Then if the vertex belongs to an even generation of its tree component, and if belongs to an odd generation. On the other hand, the statistics of Simon’s genealogical forest can be described in terms of independent random recursive trees (see e.g. Chapter 6 in [17] for background) conditionally on their sizes. This leads us to investigate the difference between the number of vertices at even generations and the number of vertices at odd generations in a random recursive tree of size . The law of can be expressed in terms of Eulerian numbers, and properties of the latter enable us either to compute explicitly or estimate certain quantities which are crucial for the proofs of Proposition 1 and Theorem 1.
It is interesting to compare asymptotic behaviors for counterbalanced steps with those for reinforced steps. If we write for the random walk with reinforced steps, then it is known that the law of large numbers holds for , namely in when , independently of the innovation parameter . Further, regarding fluctuations when , a phase transition occurs for the critical parameter , in the sense that is diffusive for and superdiffusive for ; see [10, 11]. Despite of the natural coupling between (1) and (2), there are thus major differences444 This should not come as a surprise. In the simplest case when is the Dirac mass at , one has whereas is a truly stochastic process, even for when there is no innovation. between the asymptotic behaviors of and of : Proposition 1 shows that the asymptotic speed of depends on , and Theorem 1 that there is no such phase transition for counterbalanced steps and is always diffusive.
The phase transition for step reinforcement when has a finite second moment can be explained informally as follows; for the sake of simplicity, suppose also that is centered, i.e. . There are trees in Simon’s genealogical forest, which are overwhelmingly microscopic (i.e. of size ), whereas only a few trees reach the size . Because is centered, the contribution of microscopic trees to is of order , and that of the few largest trees of order . This is the reason why grows like when , and rather like when . For counterbalanced steps, we will see that, due to the counterbalancing mechanism, the contribution of a large tree with size is now only of order . As a consequence, the contribution to of the largest trees of Simon’s genealogical forest is only of order . This is always much smaller than the contribution of microscopic trees which remain of order . We further stress that, even though only the sizes of the trees in Simon’s genealogical forest are relevant for the analysis of the random walk with reinforced steps, the study of the random walk with counterbalanced steps is more complex and requires informations on the fine structure of those trees, not merely their sizes.
The rest of this text is organized as follows. Section 2 focusses on the purely counterbalanced case . In this situation, for each fixed , the distribution of can be expressed explicitly in terms of Eulerian numbers. Section 3 is devoted to the coupling between the counterbalancing algorithm (1) and H.A. Simon’s algorithm (2), and to the interpretation of the former in terms of a forest of random recursive trees induced by the latter. Proposition 1 and Theorem 1 are proved in Section 4, where we analyze more finely the respective contributions of some natural sub-families. Last, in Section 5, we present a stable version of Theorem 1 when belongs to the domain of attraction (without centering) of an -stable distribution for some .
2 Warm-up: the purely counterbalanced case
This section is devoted to the simpler situation555 Observe that this case without innovation has been excluded in Theorem 1. when . So for all , meaning that every step, except of course the first one, counterbalances some preceding step. The law then only plays a superficial role as it is merely relevant for the first step. For the sake of simplicity, we further focus on the case when is the Dirac mass at .
The dynamics are entirely encoded by the sequence of independent uniform variables on ; more precisely the purely counterbalanced sequence of bits is given by
(3) |
The random algorithm (3) points at a convenient representation in terms of random recursive trees. Specifically, the sequence encodes a random tree with set of vertices and set of edges . Roughly speaking, is constructed recursively by incorporating vertices one after the other and creating an edge between each new vertex and its parent which is picked uniformly at random in and independently of the other vertices. If we view as the root of and call a vertex odd (respectively, even) when its generation (i.e. its distance to the root in ) is an odd (respectively, even) number, then
Let us now introduce some notation with that respect. For every , we write for the restriction of to the set of vertices and refer to as a random recursive tree with size . We also write (respectively, ) for the number of odd (respectively, even) vertices in and set
Of course, we can also express
which is the trajectory of an elephant full of regrets (i.e. for the parameter ).
The main observation of this section is that law of the number of odd vertices is readily expressed in terms of the Eulerian numbers. Recall that denotes the number of permutations of with descents, i.e. such that . Obviously if and only if , and one has
The linear recurrence equation
(4) |
is easily derived from a recursive construction of permutations (see Theorem 1.3 in [30]); we also mention the explicit formula (see Corollary 1.3 in [30])
Lemma \thelemma.
For every , we have
with the convention666 Note that this convention is in agreement with the linear recurrence equation (4). that in the right-hand side for and .
Proof.
Consider and note from the very construction of random recursive trees that there is the identity
Indeed, the first term of the sum in the right-hand side accounts for the event that the parent of the new vertex is an odd vertex (then is an even vertex), and the second term for the event that is an even vertex (then is an odd vertex).
In terms of , this yields
which is the linear recurrence equation (4) satisfied by the Eulerian numbers. Since plainly , we conclude by iteration that for all and . Last, the formula in the statement also holds for since . ∎
Remark \theremark.
Lemma 2 is implicit in Mahmoud [24]777 Beware however that the definition of Eulerian numbers in [24] slightly differs from ours, namely there corresponds to here.. Indeed can be viewed as the number of blue balls in an analytic Friedman’s urn model started with one white ball and replacement scheme ; see Section 7.2.2 in [24]. In this setting, Lemma 2 is equivalent to the formula for the number of white balls at the bottom of page 127 in [24]. Mahmoud relied on the analysis of the differential system associated to the replacement scheme via a Riccati differential equation and inversion of generating functions. The present approach based on the linear recurrence equation (4) is more direct. Lemma 2 is also a closed relative to a result due to Najock and Heyde [27] (see also Theorem 8.6 in Mahmoud [24] and Section 6.2.4 in Drmota [17]) which states that the number of leaves in a random recursive tree with size has the same distribution as that appearing in Lemma 2.
We next point at a useful identity related to Lemma 2 which goes back to Laplace (see Exercise 51 of Chapter I in Stanley [33]) and is often attributed to Tanny [34]. For every , there is the identity in distribution
(5) |
where in the right-hand side, is a sequence of i.i.d. uniform variables on and denotes the ceiling function. We now record for future use the following consequences.
Corollary \thecorollary.
-
(i)
For every , the variable is symmetric, i.e. , and in particular .
-
(ii)
For all , one has .
-
(iii)
For all , one has .
Proof.
(i) Equivalently, the assertion claims that in a random recursive tree of size at least , the number of odd vertices and the number of even vertices have the same distribution. This is immediate from (5) and can also be checked directly from the construction.
(ii) This has been already observed by Schütz and Trimper [31] in the setting of the elephant random walk; for the sake of completeness we present a short argument. The vertex is odd (respectively, even) in if and only if its parent is an even (respectively, odd) vertex in . Hence one has
and since , this yields the recursive equation
By iteration, we conclude that for all .
(iii) Recall that the process of the fractional parts is a Markov chain on whose distribution at any fixed time is uniform on . Writing and , we see that is a sequence of i.i.d. uniform variables on and that has the uniform distribution on too.
The characteristic function of the uniform variable is
and therefore for every ,
It follows that
We can rephrase (5) as the identity in distribution
Since , the proof is completed with the elementary bound . ∎
We now conclude this section with an application of (5) to the asymptotic normality of . Since and , the classical central limit theorem immediately yields the following.
Corollary \thecorollary.
Assume and . One has
3 Genealogical trees in Simon’s algorithm
From now on, is an arbitrary probability law on and we also suppose that the innovation rate is strictly positive, . Recall the construction of the sequence from Simon’s reinforcement algorithm (2). Simon was interested in the asymptotic frequencies of variables having a given number of occurrences. Specifically, for every , we write
for the number of occurrences of the variable until the -th step of the algorithm (2), and
(6) |
for the number of such variables that have occurred exactly times. Observe also that the number of innovations satisfies the law of large numbers a.s.
Lemma \thelemma.
For every , we have
where denotes the beta function.
Lemma 3 is essentially due to H.A. Simon [32], who actually only established the convergence of the mean value. The strengthening to convergence in probability can be obtained as in [13] from a concentration argument based on the Azuma-Hoeffding’s inequality; see Section 3.1 in [28]. The right-hand side in the formula is a probability mass on known as the Yule-Simon distribution with parameter . We record for future use a couple of identities which are easily checked from the integral definition of the beta function:
(7) |
and
(8) |
For , Lemma 3 reads
(9) |
We shall also need to estimate the fluctuations, which can be derived by specializing a Gaussian limit theorem for extended Pólya urns due to Bai et al. [1].
Lemma \thelemma.
There is the convergence in distribution
Proof.
The proof relies on the observation that Simon’s algorithm can be coupled with a two-color urn governed by the same sequences of random bits and of uniform variables as follows. Imagine that we observe the outcome of Simon’s algorithm at the -step and that for each , we associate a white ball if the variable appears exactly once, and a red ball otherwise. At the initial time , the urn contains just one white ball and no red balls. At each step , a ball picked uniformly at random in the urn (in terms of Simon’s algorithm, this is given by the uniform variable ). If , then the ball picked is returned to the urn and one adds a white ball (in terms of Simon’s algorithm, this corresponds to an innovation and ). If , then the ball picked is removed from the urn and one adds two red balls (in terms of Simon’s algorithm, this corresponds to a repetition and either if the ball picked is white, or if the ball picked is red). By construction, the number of white balls in the urn coincides with the number of variables that have appeared exactly once in Simon’s algorithm (2).
We shall now check our claim in the setting of [1] by specifying the quantities which appear there. The evolution of number of white balls in the urn is governed by Equation (2.1) in [1], viz.
where if a white ball is picked and otherwise. In our framework, we further have and . If we write for the natural filtration generated by the variables , then and are independent of with
This gives in the notation (2.2) of [1]:
and finally
Our claim can now be seen as a special case of Corollary 2.1 in [1]. ∎
We shall also need a refinement of Lemma 3 in which one does not only record the number of occurrences of the variable , but more generally the genealogical structure of these occurrences. We need to introduce first some notation in that respect.
Fix and let some (i.e. the variable has already appeared at the -th step of the algorithm). Write for the increasing sequence of steps of the algorithm at which appears, where . The genealogy of occurrences of the variable until the -th step is recorded as a tree on such that for every , is an edge of if and only if , that is, if and only if the identity actually results from the fact that the algorithm repeats the variable at its -th step. Plainly, is an increasing tree with size , meaning a tree on such that the sequence of vertices along any branch from the root to a leaf is increasing. In this direction, we recall that there are increasing trees with size and that the uniform distribution of the set increasing trees with size coincides with the law of the random recursive tree of size , . See for instance Section 1.3.1 in Drmota [17].

More generally, the distribution of the entire genealogical forest given the sizes of the genealogical trees can be described as follows.
Lemma \thelemma.
Fix , , and let with sum . Then conditionally on for every , the genealogical trees are independent random recursive trees with respective sizes .
Proof.
Recall that the set for is that of the edges of the random recursive tree with size , . The well-known splitting property states that removing a given edge, say for some fixed , from produces two subtrees which in turn, conditionally on their sizes, are two independent random recursive trees. This has been observed first by Meir and Moon [25]; see also [3] and references therein for more about this property.
The genealogical trees result by removing the edges in for which and enumerating in each subtree component their vertices in the increasing order. Our statement is now easily seen by applying iteratively this splitting property. ∎
We shall also need for the proofs of Proposition 1 and Theorem 1 an argument of uniform integrability that relies in turn on the following lemma. Recall that if is a rooted tree, denotes the difference between the number of vertices at even distance from root and that at odd distance.
Lemma \thelemma.
For every , one has
and
Proof.
The first claim is a consequence of Lemma 3.6 of [9] which states that for [beware that the parameter denoted by in [9] is actually here], there exists numerical constants and such that for all .
Next, combining Jensen’s inequality with Corollary 2(iii), we get that for
Then recall that conditionally on , has the law of the random recursive tree with size , , and hence
We know from the first part that this last quantity is finite, and the proof is complete. ∎
4 Proofs of the main results
As its title indicates, the purpose of this section is to establish Proposition 1 and Theorem 1. The observation that for every and , the variable appears exactly times and its opposite exactly times until the -step of the algorithm (1), yields the identity
(10) |
which lies at the heart of our approach. We stress that in (10) as well as in related expressions that we shall use in the sequel, the sequence of i.i.d. variables and the family of genealogical trees are independent, because the latter are constructed from the sequences and only.
Actually, our proof analyzes more precisely the effects of the counterbalancing algorithm (1) by estimating specifically the contributions of certain sub-families to the asymptotic behavior of . Specifically, we set for every ,
(11) |
so that
4.1 Proof of Proposition 1
The case (no counterbalancing events) of Proposition 1 is just the weak law of large numbers, and the case (no innovations) is a consequence of Corollary 2. The case derives from the next lemma which shows more precisely that the variables that have appeared in the algorithm (1) but have not yet counterbalanced determine the ballistic behavior of , whereas those that have appeared twice or more (i.e. such that ) have a negligible impact.
Lemma \thelemma.
Assume that and recall that . Then the following limits hold in probability:
-
(i)
,
-
(ii)
.
Proof.
(i) Recall the notation (6) and that the sequence of i.i.d. variables is independent of the events . We see that there is the identity in distribution
where is the usual random walk. The claim follows readily from the law of large numbers and (9).
(ii) We first argue that for each fixed ,
(12) |
Indeed, recall that denotes the number of genealogical trees with size . It follows from Lemma 3 that conditionally on , the sub-family of such enumerated in the increasing order of the index , is given by i.i.d. copies of the random recursive tree . Hence, still conditionally on , enumerating the elements of the sub-family in the increasing order of yields independent variables, each being distributed as with and independent. We deduce from Corollary 2(i) that the variable symmetric, and since it is also integrable, it is centered. Since , this readily entails (12) by an application of the law of large numbers.
The proof can be completed by an argument of uniform integrability. In this direction, fix an arbitrarily large integer and write by the triangular inequality
where the last inequality holds for any . We see from Lemma 3 that the right-hand side converges to as uniformly in , and the rest of the proof is straightforward. ∎
4.2 Proof of Theorem 1
For (no counterbalancing events), Theorem 1 just reduces to the classical central limit theorem, so we assume . The first step of the proof consists in determining jointly the fluctuations of the components defined in (11).
Lemma \thelemma.
Assume that . Then as , the sequences of random variables
and
converge jointly in distribution towards a sequence
of independent centered Gaussian variables, where
, and
Proof.
For each , let be a sequence of i.i.d. copies of , where has the law and is independent of the random recursive tree . We further assume that these sequences are independent. Taking partial sums yields a sequence indexed by of independent random walks
For each , the family of blocks
forms a random partition of which is independent of the ’s. Recall that we are using the notation , and also from Lemma 3, that conditionally on the ’s, the genealogical trees are independent random recursive trees. We now see from the very definition (11) that for every fixed , there is the identity in distribution
where in the right-hand side, the random walks are independent of the sequence of block sizes .
Next we write, first for ,
second (since ) for , and third, for ,
with the usual summation convention that when .
Since the i.i.d. variables have mean and variance , the central limit theorem ensures that there is the convergence in distribution
(13) |
Similarly, for , each is centered with variance (by Corollary 2(i-ii)) and hence, using the notation in the statement, there is the convergence in distribution
(14) |
Plainly, the weak convergences (13) and (14) hold jointly when we agree that the limits are independent Gaussian variables.
Next, from Lemma 3 and the fact that for , the i.i.d. variables are centered with finite variance, we easily get
Finally, for , we write
On the one hand, we have from the same argument as above that
On the other hand, we already know from Lemma 3 that there is the convergence in distribution
Obviously, this convergence in law hold jointly with (13) and (14), where the limiting Gaussian variables are independent. Putting the pieces together, this completes the proof. ∎
The final step for the proof of Theorem 1 is the following lemma.
Lemma \thelemma.
We have
Proof.
We write
Since the are independent of the , we get
We evaluate the expectation in the right-hand side by conditioning first on and with . Recall from Lemma 3 that the genealogical trees and are then two random recursive trees with respective sizes and , which are further independent when . Thanks to Corollary 2(i-ii) we get
5 A stable central limit theorem
The arguments for the proof of Theorem 1 when the step distribution has a finite second moment can be adapted to the case when belongs to some stable domain of attraction; for the sake of simplicity we focus on the situation without centering. Specifically, let be a sequence of positive real numbers that is regularly varying with exponent for some , in the sense that for every , and suppose that
(15) |
where is some -stable random variable. We refer to Theorems 4 and 5 on p. 181-2 in [19] and Section 6 of Chapter 2 in [21] for necessary and sufficient conditions for (15) in terms of only. We write for the characteristic exponent of , viz.
recall that is homogeneous with exponent , i.e.
Recall the definition and properties of the Eulerian numbers from Section 2, and also the Pochhammer notation
for the rising factorial, where stands for the gamma function. We can now claim:
Theorem \thetheorem.
Assume (15). For each , we have
where is an -stable random variable with characteristic exponent given by
The proof of Theorem 5 relies on a refinement of Simon’s result (Lemma 3) to the asymptotic frequencies of genealogical trees induced by the reinforcement algorithm (2). We denote by the set of increasing trees (of arbitrary finite size), and for any , we write for its the size (number of vertices) and for the difference between its numbers of even vertices and of odd vertices. Refining (6), we also define
Lemma \thelemma.
We have
and there is the convergence in probability
Proof.
We start by claiming that for every , and every tree with size , we have
(16) |
Indeed, the distribution of the random recursive tree of size is the uniform probability measure on the set of increasing trees with size , which has elements. We deduce from Lemma 3 and the law of large numbers that
The claim (16) now follows from Lemma 3 and the identity
We now establish Theorem 5.
Proof of Theorem 5.
We denote the characteristic function of by
Fix small enough so that whenever , and then define the characteristic exponent as the continuous determination of the logarithm of on . In words, is the unique continuous function on with and such that for all . For definitiveness, we further set whenever .
Next, observe from the Markov’s inequality that for any and any
so that, thanks to Lemma 3,
In particular, since the sequence is regularly varying with exponent , for every , the events
occur with high probability as , in the sense that .
We then deduce from (10) and the fact that the variables are i.i.d. with law that for every ,
We then write, in the notation of Lemma 5,
Recall that we assume (15). According to Theorem 2.6.5 in Ibragimov and Linnik [21], is regularly varying at with exponent , and since , the Potter bounds (see Theorem 1.5.6 in [12]) show that for some constant :
(17) |
We deduce from Lemma 5 that for every fixed , there is the convergence in probability
Furthermore, still from Theorem 2.6.5 in Ibragimov and Linnik [21], we have
and we deduce by dominated convergence, using Lemma 5 and (17), that
Putting the pieces together, we have shown that
It only remains to check that the right-hand side above agrees with the formula of the statement. This follows from Lemma 2 and the fact that for every , has the uniform distribution on . ∎
References
- [1] Bai, Z. D., Hu, F., and Zhang, L.-X. Gaussian approximation theorems for urn models and their applications. Ann. Appl. Probab. 12, 4 (2002), 1149–1173.
- [2] Baur, E. On a class of random walks with reinforced memory. Journal of Statistical Physics 181 (2020), 772–802.
- [3] Baur, E., and Bertoin, J. Cutting edges at random in large recursive trees. In Stochastic analysis and applications 2014, vol. 100 of Springer Proc. Math. Stat. Springer, Cham, 2014, pp. 51–76.
- [4] Baur, E., and Bertoin, J. Elephant random walks and their connection to Pólya-type urns. Phys. Rev. E 94 (Nov 2016), 052134.
- [5] Bercu, B. A martingale approach for the elephant random walk. J. Phys. A 51, 1 (2018), 015201, 16.
- [6] Bercu, B., and Laulin, L. On the center of mass of the elephant random walk. Stochastic Processes and their Applications 133 (2021), 111 – 128.
- [7] Bertenghi, M. Asymptotic normality of superdiffusive step-reinforced random walks. arXiv:2101.00906.
- [8] Bertenghi, M. Functional limit theorems for the multi-dimensional elephant random walk. Stoch. Models 38, 1 (2022), 37–50.
- [9] Bertoin, J. Noise reinforcement for Lévy processes. Ann. Inst. H. Poincaré Probab. Statist. 56, 3 (2020), 2236–2252.
- [10] Bertoin, J. Scaling exponents of step-reinforced random walks. Probab. Theory Relat. Fields 179 (2021), 295–315.
- [11] Bertoin, J. Universality of noise reinforced Brownian motions. In In and out of equilibrium 3. Celebrating Vladas Sidoravicius, vol. 77 of Progr. Probab. Birkhäuser/Springer, Cham, [2021] ©2021, pp. 147–161.
- [12] Bingham, N. H., Goldie, C. M., and Teugels, J. L. Regular variation, vol. 27 of Encyclopedia of Mathematics and its Applications. Cambridge University Press, Cambridge, 1987.
- [13] Bollobás, B., Riordan, O., Spencer, J., and Tusnády, G. The degree sequence of a scale-free random graph process. Random Structures Algorithms 18, 3 (2001), 279–290.
- [14] Businger, S. The shark random swim (Lévy flight with memory). J. Stat. Phys. 172, 3 (2018), 701–717.
- [15] Coletti, C. F., Gava, R., and Schütz, G. M. Central limit theorem and related results for the elephant random walk. J. Math. Phys. 58, 5 (2017), 053303, 8.
- [16] Coletti, C. F., Gava, R., and Schütz, G. M. A strong invariance principle for the elephant random walk. J. Stat. Mech. Theory Exp., 12 (2017), 123207, 8.
- [17] Drmota, M. Random trees. Springer Wien, NewYork, Vienna, 2009. An interplay between combinatorics and probability.
- [18] Durrett, R. Random graph dynamics, vol. 20 of Cambridge Series in Statistical and Probabilistic Mathematics. Cambridge University Press, Cambridge, 2007.
- [19] Gnedenko, B. V., and Kolmogorov, A. N. Limit distributions for sums of independent random variables. Translated from the Russian, annotated, and revised by K. L. Chung. With appendices by J. L. Doob and P. L. Hsu. Revised edition. Addison-Wesley Publishing Co., Reading, Mass.-London-Don Mills., Ont., 1968.
- [20] Gut, A., and Stadtmüller, U. The number of zeros in elephant random walks with delays. Statist. Probab. Lett. 174 (2021), Paper No. 109112, 9.
- [21] Ibragimov, I. A., and Linnik, Y. V. Independent and stationary sequences of random variables. Wolters-Noordhoff Publishing, Groningen, 1971. With a supplementary chapter by I. A. Ibragimov and V. V. Petrov, Translation from the Russian edited by J. F. C. Kingman.
- [22] Kubota, N., and Takei, M. Gaussian fluctuation for superdiffusive elephant random walks. J. Stat. Phys. 177, 6 (2019), 1157–1171.
- [23] Kürsten, R. Random recursive trees and the elephant random walk. Phys. Rev. E 93, 3 (2016), 032111, 11.
- [24] Mahmoud, H. M. Pólya urn models. Texts in Statistical Science Series. CRC Press, Boca Raton, FL, 2009.
- [25] Meir, A., and Moon, J. Cutting down recursive trees. Mathematical Biosciences 21, 3 (1974), 173 – 181.
- [26] Miyazaki, T., and Takei, M. Limit theorems for the ‘laziest’ minimal random walk model of elephant type. J. Stat. Phys. 181, 2 (2020), 587–602.
- [27] Najock, D., and Heyde, C. C. On the number of terminal vertices in certain random trees with an application to stemma construction in philology. J. Appl. Probab. 19, 3 (1982), 675–680.
- [28] Pachon, A., Polito, F., and Sacerdote, L. Random graphs associated to some discrete and continuous time preferential attachment models. J. Stat. Phys. 162, 6 (2016), 1608–1638.
- [29] Pemantle, R. A survey of random processes with reinforcement. Probab. Surveys 4 (2007), 1–79.
- [30] Petersen, T. K. Eulerian numbers. Birkhäuser Advanced Texts: Basler Lehrbücher. [Birkhäuser Advanced Texts: Basel Textbooks]. Birkhäuser/Springer, New York, 2015. With a foreword by Richard Stanley.
- [31] Schütz, G. M., and Trimper, S. Elephants can always remember: Exact long-range memory effects in a non-Markovian random walk. Phys. Rev. E 70 (Oct 2004), 045101.
- [32] Simon, H. A. On a class of skew distribution functions. Biometrika 42, 3/4 (1955), 425–440.
- [33] Stanley, R. P. Enumerative combinatorics. Vol. 1, vol. 49 of Cambridge Studies in Advanced Mathematics. Cambridge University Press, Cambridge, 1997. With a foreword by Gian-Carlo Rota, Corrected reprint of the 1986 original.
- [34] Tanny, S. A probabilistic interpretation of Eulerian numbers. Duke Math. J. 40, 4 (12 1973), 717–722.