This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Improved bounds for coloring locally sparse hypergraphs

Fotis Iliopoulos
Institute for Advanced Study
and Princeton University
[email protected]
This material is based upon work directly supported by the IAS Fund for Math and indirectly supported by the National Science Foundation Grant No. CCF-1900460. Any opinions, findings and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the National Science Foundation. This work is also supported by the National Science Foundation Grant No. CCF-1815328.
Abstract

We show that, for every k2k\geq 2, every kk-uniform hypergaph of degree Δ\Delta and girth at least 55 is efficiently (1+o(1))(k1)(Δ/lnΔ)1/(k1)(1+o(1))(k-1)(\Delta/\ln\Delta)^{1/(k-1)}-list colorable. As an application (and to the best of our knowledge) we obtain the currently best algorithm for list-coloring random hypergraphs of bounded average degree.

1 Introduction

In hypergraph coloring one is given a hypergraph H(V,E)H(V,E) and the goal is to find an assignment of one of qq colors to each vertex vVv\in V so that no hyperedge is monochromatic. In the more general list-coloring problem, a list of qq allowed colors is specified for each vertex. A graph is qq-list-colorable if it has a list-coloring no matter how the lists are assigned to each vertex. The list chromatic number, χ(H)\chi_{\ell}(H), is the smallest qq for which HH is qq-list colorable.

Hypergraph coloring is a fundamental constraint satisfaction problem with several applications in computer science and combinatorics, that has been studied for over 60 years. In this paper we consider the task of coloring locally sparse hypergraphs and its connection to coloring sparse random hypegraphs.

A hypergraph is kk-uniform if every hyperedge contains exactly kk vertices. An ii-cycle in a kk-uniform hypergraph is a collection of ii distinct hyperedges spanned by at most i(k1)i(k-1) vertices. We say that a kk-uniform hypergraph has girth at least gg if it contains no ii-cycles for 2i<g2\leq i<g. Note that if a kk-uniform hypergraph has girth at least 33 then every two of its hyperedges have at most one vertex in common.

The main contribution of this paper is to prove the following theorem.

Theorem 1.1.

Let HH by any kk-uniform hypergraph, k2k\geq 2, of maximum degree Δ\Delta and girth at least 55. For all ϵ>0\epsilon>0, there exist a positive constant Δϵ,k\Delta_{\epsilon,k} such that if ΔΔϵ,k\Delta\geq\Delta_{\epsilon,k}, then

χ(H)(1+ϵ)(k1)(ΔlnΔ)1k1.\displaystyle\chi_{\ell}(H)\leq(1+\epsilon)(k-1)\left(\frac{\Delta}{\ln\Delta}\right)^{\frac{1}{k-1}}. (1)

Furthermore, if HH is a hypergraph on nn vertices then there exists an algorithm that constructs such a coloring in expected polynomial time in nn.

Remark 1.1.

If Δ\Delta is assumed to be constant, then the algorithm of Theorem 1.1 can be efficiently derandomized, i.e., there exists a deterministic algorithm that constructs the promised coloring in polynomial time in nn.

Theorem 1.1 is interesting for a number of reasons. First, it generalizes a well-known result of Kim [21] for coloring graphs of degree Δ\Delta and girth 55, and it implies the classical theorem of Ajtai, Komlós, Pintz, Spencer and Szemerédi [4] regarding the independence number of kk-uniform hypergraphs of degree Δ\Delta and girth 55. The latter is a seminal result in combinatorics, with applications in geometry and coding theory [22, 23, 24]. Second, Theorem 1.1 is tight up to a constant [8]. Note also that, without the girth assumption, the best possible bound [11] on the chromatic number of kk-uniform hypergraphs is O(Δ1/(k1))O(\Delta^{1/(k-1)}), i.e., it is asymptotically worse than the one of Theorem 1.1. For example, there exist graphs of degree Δ\Delta whose chromatic number is exactly Δ+1\Delta+1. Third, when it applies, Theorem 1.1 improves upon a result of Frieze and Mubayi [13] regarding the chromatic number of simple hypergraphs, who showed (1) with an unspecified large leading constant (of order at least Ω(k4)\Omega(k^{4})). Finally, Theorem 1.1 can be used to provide the currently best algorithm for list-coloring random kk-uniform hypergarphs of bounded average degree (to the best of our knowledge). We discuss the connection between locally sparse hypergraphs and sparse random hypergraphs with respect to the task of coloring in the following section.

1.1 Application to coloring pseudo-random hypergraphs

The random kk-uniform hypergraph H(k,n,p)H(k,n,p) is obtained by choosing each of the (nk){n\choose k} kk-element subsets of a vertex set VV (|V|=n|V|=n ) independently with probability pp. The chosen subsets are the hyperedges of the hypergraph. Note that for k=2k=2 we have the usual definition of the random graph G(n,p)G(n,p). We say that H(k,n,p)H(k,n,p) has a certain property AA asymptotically almost surely or with high probability, if the probability that HH(k,n,p)H\in H(k,n,p) has AA tends to 11 as nn\to\infty.

In this paper we are interested in H(k,n,d/(nk1))H(k,n,d/{n\choose k-1}), i.e., the family of random kk-uniform hypergraphs of bounded average degree dd. Specifically, we use Theorem 1.1 to prove the following theorem.

Theorem 1.2.

For any constants δ(0,1)\delta\in(0,1), k2k\geq 2, there exists dδ,k>0d_{\delta,k}>0 such that for every constant ddδ,kd\geq d_{\delta,k}, the random hypergraph H(k,n,d/(nk1))H(k,n,d/{n\choose k-1}) can be (1+δ)(k1)(d/lnd)1/(k1)(1+\delta)(k-1)(d/\ln d)^{1/(k-1)}-list-colored by a deterministic algorithm whose running time is polynomial in nn asymptotically almost surely.

Remark 1.2.

Note that, for k,dk,d constants, a very standard argument reveals that H(k,n,d/(nk1))H(k,n,d/{n\choose k-1}) is essentially equivalent to (k,n,dn/k)\mathbb{H}(k,n,dn/k), namely the uniform distribution over kk-uniform hypergraphs with nn vertices and exactly dn/kdn/k hyperedges. Thus, Theorem 1.2 extends to that model as well.

We note that previous approaches [3, 31, 13] for list-coloring random kk-uniform hypergraphs of bounded average degree dd are either randomized, or require significantly larger lists of colors per vertex in order to succeed. Indeed, for k=2k=2 the approach of Achlioptas and Molloy [3] matches the bound of Theorem 1.2 but is randomized, while for k3k\geq 3, and to the best our knowledge, our algorithm uses less colors that any algorithm for (list-)coloring random hypergraphs of bounded average degree that has been rigorously analyzed in the literature. Moreover, it is believed that all efficient algorithms require lists of size at least (1+o(1))((k1)d/lnd)1/(k1)(1+o(1))((k-1)d/\ln d)^{1/(k-1)}, as this bound corresponds to the so-called shattering threshold [1, 7, 14] for coloring sparse random hypergraphs, which is also often referred to as the “algorithmic barrier” [1]. This threshold arises in a plethora of random constraint satisfaction problems, and it corresponds to a precise phase transition in the geometry set of solutions. In all of these problems, we are not aware of any efficient algorithm that works beyond the algorithmic barrier, despite the fact that solutions exist for constraint-densities larger than the one in which the shattering phenomenon appears. We refer the reader to [1, 33] for further details.

In order to prove Theorem 1.2, we show that random kk-uniform hypergraphs of bounded average degree dd can essentially be treated as hypergraphs of girth 55 and maximum degree dd for the purposes of list-coloring, and then apply Theorem 1.1. In particular, we identify a pseudo-random family of hypergraphs which we call girth-reducible, and show that almost all kk-uniform hypergraphs of bounded average degree belong in this class. Then we show that girth-reducible hypergraphs can be colored efficiently using Theorem 1.1.

Formally, a kk-uniform hypergraph HH is κ\kappa-degenerate if the induced subhypergraph of all subsets of its vertex set has a vertex of degree at most κ\kappa. The degeneracy of a hypergraph HH is the smallest value of κ\kappa for which HH is κ\kappa-degenerate. Note that it is known that κ\kappa-degenerate hypergraphs are (κ+1)(\kappa+1)-list colorable and that the degeneracy of a hypergraph can be computed efficiently by an algorithm that repeatedly removes minimum degree vertices. Indeed, to list-color a κ\kappa-degenerate hypergraph we repeatedly find a vertex with (remaining) degree at most κ\kappa, assign to it a color that does not appear in any of its neighbors so far, and remove it from the hypergraph. Clearly, if the lists assigned to each vertex are of size at least κ+1\kappa+1 this procedure always terminates successfully.

Definition 1.3.

For δ(0,1)\delta\in(0,1), we say that a kk-uniform hypergraph H(V,E)H(V,E) of average degree dd is δ\delta-girth-reducible if its vertex set can be partitioned in two sets, UU and VUV\setminus U, such that:

  1. (a)

    UU contains all cycles of length at most 44, and all vertices of degree larger than (1+δ)d(1+\delta)d;

  2. (b)

    subhypergraph H[U]H[U] is (dlnd)1k1\left(\frac{d}{\ln d}\right)^{\frac{1}{k-1}}-degenerate;

  3. (c)

    every vertex in VUV\setminus U has at most δ(dlnd)1k1\delta\left(\frac{d}{\ln d}\right)^{\frac{1}{k-1}} neighbors in UU.

In words, a hypergraph is δ\delta-girth-reducible if its vertex set can be seen as the union of two parts: A “low-degeneracy” part, which contains all vertices of degree more than (1+δ)d(1+\delta)d and all cycles of lengths at most 44, and a “high-girth” part, which induces a hypergraph of maximum degree at most (1+δ)d(1+\delta)d and girth 55. Moreover, each vertex in the “high-girth” part has only a few neighbors in the “low-degeneracy” part.

Note that given a δ\delta-girth-reducible hypergraph we can efficiently find the promised partition (U,VU)(U,V\setminus U) as follows. We start with U:=U0U:=U_{0}, where U0U_{0} is the set of vertices that either have degree at least (1+δ)d(1+\delta)d, or they are contained in a cycle of length at most 44. Let U\partial U denote the vertices in VUV\setminus U that violate property (c). While U\partial U\neq\emptyset, update UU as U:=UUU:=U\cup\partial U. The correctness of the process lies in the fact that in each step we add to the current UU a set of vertices that must be in the low-degeneracy part of the hypergraph. Observe also that this process allows us to efficiently check whether a hypergraph is δ\delta-girth-reducible.

We prove the following theorem regarding the list-chromatic number of girth-reducible hypergraphs.

Theorem 1.4.

For any constants δ(0,1)\delta\in(0,1) and k2k\geq 2, there exists dδ,k>0d_{\delta,k}>0 such that if HH is a δ\delta-girth-reducible, kk-uniform hypergraph of average degree ddδ,kd\geq d_{\delta,k}, then

χ(H)(1+ϵ)(k1)(dlnd)1k1,\displaystyle\chi_{\ell}(H)\leq(1+\epsilon)(k-1)\left(\frac{d}{\ln d}\right)^{\frac{1}{k-1}},

where ϵ=4δ=O(δ)\epsilon=4\delta=O(\delta). Furthermore, if HH is a hypergraph on nn vertices then there exists a deterministic algorithm that constructs such a coloring in time polynomial in nn.

Proof of Theorem 1.4.

Let ϵ=4δ\epsilon=4\delta. Given lists of colors of size (1+ϵ)(k1)(dlnd)1k1(1+\epsilon)(k-1)\left(\frac{d}{\ln d}\right)^{\frac{1}{k-1}} for each vertex of HH, we first color the vertices of UU using the greedy algorithm which exploits the low degeneracy of H[U]H[U]. Now each vertex in VUV-U has at most δ(dlnd)1k1\delta\left(\frac{d}{\ln d}\right)^{\frac{1}{k-1}} forbidden colors in its list as it has at most that many neighbors in UU. We delete these colors from the list. Observe that if we manage to properly color the induced subgraph H[VU]H[V\setminus U] using colors from the updated lists, then we are done since every hyperedge with vertices both in UU and VUV\setminus U will be automatically “satisfied”, i.e., it cannot be monochromatic. Notice now that the updated list of each vertex still contains at least (1+3δ)(k1)(dlnd)1k1(1+3\delta)(k-1)\left(\frac{d}{\ln d}\right)^{\frac{1}{k-1}} colors, for sufficiently large dd. Since the induced subgraph H[VU]H[V\setminus U] is of girth at least 55 and of maximum degree at most (1+δ)d(1+\delta)d, it is efficiently (1+δ)(k1)((1+δ)dln((1+δ)d))1k1(1+\delta)(k-1)\left(\frac{(1+\delta)d}{\ln\left((1+\delta)d\right)}\right)^{\frac{1}{k-1}}-list-colorable for sufficiently large dd per Theorem 1.1 and Remark 1.1. This concludes the proof since (1+δ)(1+δ)1k1<(1+3δ)(1+\delta)(1+\delta)^{\frac{1}{k-1}}<(1+3\delta).

Moreover, we show that girth-reducibility is a pseudo-random property which is admitted by almost all sparse kk-uniform hypregraphs.

Theorem 1.5.

For any constants δ(0,1)\delta\in(0,1), k2k\geq 2, there exists dδ,k>0d_{\delta,k}>0 such that for every constant ddδ,kd\geq d_{\delta,k}, asymptotically almost surely, the random hypergraph H(k,n,d/(nk1))H(k,n,d/{n\choose k-1}) is δ\delta-girth-reducible.

Theorem 1.5 follows by simple, although somewhat technical, considerations on properties of sparse random hypergraphs, which are mainly inspired by the results of Alon, Krivelevich and Sudakov [6] and Łuczak [25]. Observe that combining Theorem 1.5 with Theorem 1.4 immediately implies Theorem 1.2.

Overall, the task of coloring locally sparse hypergraphs is inherently related to the average-case complexity of coloring. In particular, in this section we showed that Theorem 1.1 implies a robust algorithm for hypergraph coloring, namely a deterministic procedure that applies to worst-case kk-uniform hypergraphs, while at the same using a number of colors that is only a (k1)(k-1)-factor away from the algorithmic barrier for random instances (matching it for k=2k=2). We remark that this application is inspired by recent results that study the connection between local sparsity and efficient randomized algorithms for coloring sparse regular random graphs [26, 2, 10].

1.2 Technical overview

The intuition behind the proof of Theorem 1.1 comes from the following observation, which we explain in terms of graph coloring for simplicity. Let GG be a triangle-free graph of degree Δ\Delta, and assume that each of its vertices is assigned an arbitrary list of qq colors. Fix a vertex vv of GG, and consider the random experiment in which the neighborhood of vv is properly list-colored randomly. Since GG contains no triangles, this amounts to assigning to each neighbor of vv a color from its list randomly and independently. Assuming that qq:=(1+ϵ)Δ/lnΔq\geq q^{*}:=(1+\epsilon)\Delta/\ln\Delta, the expected number of available colors for vv, i.e., the colors from the list of vv that do not appear in any of its neighbors, is at least q(11/q)Δ=ω(Δϵ/2)q(1-1/q)^{\Delta}=\omega(\Delta^{\epsilon/2}). In fact, a simple concentration argument reveals that the number of available colors for vv in the end of this experiment is at least Δϵ/2\Delta^{\epsilon/2} with probability that goes to 11 as Δ\Delta grows. To put it differently, as long as qqq\geq q^{*}, the vast majority of valid ways to list-color the neighborhood of vv “leaves enough room” to color vv without creating any monochromatic edges.

A completely analogous observation regarding the ways to properly color the neighborhood of a vertex can be made for kk-uniform hypergraphs. In order to exploit it we employ the so-called semi-random method, which is the main tool behind some of the strongest graph coloring results, e.g., [15, 16, 17, 18, 20, 27, 32], including the one of Kim [21]. (See also the very recent survey [19] of Kang et. al. on the subject.) The idea is to gradually color the hypergraph in iterations until we reach a point where we can finish the coloring with a simple, e.g., greedy, algorithm. In its most basic form, each iteration consists of the following simple procedure (using graph vertex coloring as a canonical example): Assign to each vertex a color chosen uniformly at random; then uncolor any vertex that receives the same color as one of its neighbors. Using the Lovász Local Lemma [11] and concentration inequalities, one typically shows that, with positive probability, the resulting partial coloring has useful properties that allow for the continuation of the argument in the next iteration. (In fact, using the Moser-Tardos algorithm [29] this approach yields efficient, and often times deterministic [9], algorithms.) Specifically, one keeps track of certain parameters of the current partial coloring and makes sure that, in each iteration, these parameters evolve almost as if the coloring was totally random. For example, recalling the heuristic experiment of the previous paragraph, one of the parameters we would like to keep track of in our case is a lower bound on the number of available colors of each vertex in the hypergraph: If this parameter evolves “randomly” throughout the process, then the vertices that remain uncolored in the end are guaranteed to have a non-trivial number of available colors.

Applications of the semi-random method tend to be technically intense and this is even more so in our case, where we have to deal with constraints of large arity. Large constraints introduce several difficulties, but the most important one is that our algorithm has to control many parameters that interact with each other. Roughly, in order to guarantee the properties that allow for the continuation of the argument in the next iteration, for each uncolored vertex vv, each color cc in the list of vv, and each integer r[k1]r\in[k-1], we should keep track of a lower bound on the number of adjacent to vv hyperedges that have rr uncolored vertices and k1rk-1-r vertices colored cc. Clearly, these parameters are not independent of each other throughout the process, and so the main challenge is to design and analyze a coloring procedure in which all of them, simultaneously, evolve essentially randomly.

1.3 Organization of the paper

The paper is organized as follows. In Section 2 we present the necessary background. In Section 3 we present the algorithm and state the key lemmas for the proof of Theorem 1.1, while in Section 4 we give the full details. Finally, in Section 5 we prove Theorem 1.5.

2 Background and preliminaries

In this section we give some background on the technical tools that we will use in our proofs.

2.1 The Lovász Local Lemma

As we have already mentioned, one of the key tools we will use in our proof is the Lovász Local Lemma [11].

Theorem 2.1.

Consider a set ={B1,B2,,Bm}\mathcal{B}=\{B_{1},B_{2},\ldots,B_{m}\} of (bad) events. For each BB\in\mathcal{B}, let D(B){B}D(B)\subseteq\mathcal{B}\setminus\{B\} be such that Pr[BCSC¯]=Pr[B]\Pr[B\mid\bigcap_{C\in S}\overline{C}]=\Pr[B] for every S(D(B){B})S\subseteq\mathcal{B}\setminus(D(B)\cup\{B\}). If there is a function x:(0,1)x:\mathcal{B}\rightarrow(0,1) satisfying

Pr[B]x(B)CD(B)(1x(C)) for all B,\Pr[B]\leq x(B)\prod_{C\in D(B)}(1-x(C))\enspace\text{ for all $B\in\mathcal{B}$}, (2)

then the probability that none of the events in \mathcal{B} occurs is at least B(1x(B))>0\prod_{B\in\mathcal{B}}(1-x(B))>0.

In particular, we will need the following two corollaries of Theorem 2.1. For their proofs, the reader is referred to Chapter 19 in [28].

Corollary 2.2.

Consider a set ={B1,,Bm}\mathcal{B}=\{B_{1},\ldots,B_{m}\} of (bad) events. For each BB\in\mathcal{B}, let D(B){B}D(B)\subseteq\mathcal{B}\setminus\{B\} be such that Pr[BCSC¯]=Pr[B]\Pr[B\mid\bigcap_{C\in S}\overline{C}]=\Pr[B] for every S(D(B){B})S\subseteq\mathcal{B}\setminus(D(B)\cup\{B\}). If for every BB\in\mathcal{B}:

  1. (a)

    Pr[B]14\Pr[B]\leq\frac{1}{4};

  2. (b)

    CD(B)Pr[C]14\sum_{C\in D(B)}\Pr[C]\leq\frac{1}{4},

then the probability that none of the events in \mathcal{B} occurs is strictly positive.

Corollary 2.3.

Consider a set ={B1,B2,,Bm}\mathcal{B}=\{B_{1},B_{2},\ldots,B_{m}\} of (bad) events such that for each BB\in\mathcal{B}:

  1. (a)

    Pr[B]p<1\Pr[B]\leq p<1;

  2. (b)

    BB is mutually independent of a set of all but at most Δ\Delta of the other events.

If 4pΔ14p\Delta\leq 1 then with positive probability, none of the events in \mathcal{B} occur.

2.2 Talagrand’s inequality

We will also need the following version of Talagrand’s inequality [30] whose proof can be found in Chapter 20 of [28].

Theorem 2.4.

Let XX be a non-negative random variable, not identically 0, which is determined by nn independent trials T1,,TnT_{1},\ldots,T_{n}, and satisfying the following for some c,r>0c,r>0:

  1. 1.

    changing the outcome of any trial can affect XX by at most cc, and

  2. 2.

    for any ss, if XsX\geq s then there is a set of at most wsws trials whose outcomes certify that XsX\geq s,

then for any 0t𝔼[X]0\leq t\leq{\mathbb{E}}[X],

Pr[|X𝔼[X]|>t+60cw𝔼[X]]4et28c2w𝔼[X].\displaystyle\Pr[|X-{\mathbb{E}}[X]|>t+60c\sqrt{w{\mathbb{E}}[X]}]\leq 4\mathrm{e}^{-\frac{t^{2}}{8c^{2}w{\mathbb{E}}[X]}}\enspace.

3 List-coloring high-girth hypergraphs

In this section we describe the algorithm of Theorem 1.1. As we already explained, our approach is based on the semi-random method. For an excellent exposition both of the method and Kim’s result the reader is referred to [28].

We assume without loss of generality that ϵ<110\epsilon<\frac{1}{10}. Also, it will be convenient to define the parameter δ:=(1+ϵ)(k1)1\delta:=(1+\epsilon)(k-1)-1, so that the list of each vertex initially has at least (1+δ)(ΔlnΔ)1k1(1+\delta)(\frac{\Delta}{\ln\Delta})^{\frac{1}{k-1}} colors, and assume that k3k\geq 3. (The case k=2k=2 is Kim’s result.)

We analyze each iteration of our procedure using a probability distribution over the set of (possibly improper) colorings of the uncolored vertices of HH where, additionally, each vertex is either activated or deactivated. We call a pair of coloring and activation bits assignments for the uncolored vertices of hypergraph HH a state.

Let ViV_{i} denote the set of uncolored vertices in the beginning of the ii-th iteration. (Initially, all vertices are uncolored.) Also, for each vViv\in V_{i} we denote by Lv=Lv(i)L_{v}=L_{v}(i) the list of colors of vv in the beginning of the ii-th iteration. Finally, we say that a color cLvc\in L_{v} is available for vv in a state σ\sigma if assigning cc to vv does not cause any hyperedge whose initially uncolored vertices are all activated in σ\sigma to be monochromatic.

For each vertex vv, color cLvc\in L_{v} and iteration ii, we define a few quantities of interest that our algorithm will attempt to control. Let i(v)\ell_{i}(v) be the size of LvL_{v}. Further, for each r[k]r\in[k], let Di,r(v,c)D_{i,r}(v,c) denote the set of hyperedges hh that contain vv and (i) exactly rr vertices {u1,,ur}h{v}\{u_{1},\ldots,u_{r}\}\subseteq h\setminus\{v\} are uncolored and cLujc\in L_{u_{j}} for every j[r]j\in[r]; (ii) the rest k1rk-1-r vertices of hh other than vv are colored cc. We define ti,r(v,c):=|Di,r(v,c)|t_{i,r}(v,c):=|D_{i,r}(v,c)|.

As it is common in the applications of the semi-random method, we will not attempt to keep track of the values of i(v)\ell_{i}(v) and ti,r(v,c)t_{i,r}(v,c), r[k1]r\in[k-1], for every vertex vv and color cc, but rather we will focus on their extreme values. In particular, we will define appropriate Li,Ti,rL_{i},T_{i,r} such that for each ii the following property holds in the beginning of iteration ii:

Property P(i): For each vertex vViv\in V_{i}, color cLvc\in L_{v} and r[k1]r\in[k-1]:

i(v)\displaystyle\ell_{i}(v) \displaystyle\geq Li;\displaystyle L_{i};
ti,r(v,c)\displaystyle t_{i,r}(v,c) \displaystyle\leq Ti,r.\displaystyle T_{i,r}.

As a matter of fact, it would be helpful for our analysis (though not necessary) if the inequalities defined in P(i)P(i) were actually tight. Given that P(i)P(i) holds, we can always enforce this stronger property in a straightforward way as follows. First, for each vertex vv such that i(v)>Li\ell_{i}(v)>L_{i} we choose arbitrarily i(v)Li\ell_{i}(v)-L_{i} colors from its list and remove them. Then, for each vertex vv and color cLic\in L_{i} such that ti,r(v,c)<Ti,rt_{i,r}(v,c)<T_{i,r} we add to the hypergraph Ti,rti,r(v,c)T_{i,r}-t_{i,r}(v,c) new hyperedges of size r+1r+1 that contain vv and rr new “dummy” vertices. (As it will be evident from the proof, we can always assume that Li,Ti,rL_{i},T_{i,r} are integers, since our analysis is robust to replacing Li,Ti,rL_{i},T_{i,r} with Li\lfloor L_{i}\rfloor and Ti,rT_{i,r} with Ti,r\lceil T_{i,r}\rceil.) We assign each dummy vertex a list of LiL_{i} colors: Li1L_{i}-1 of them are new and do not appear in the list of any other vertex, and the last one is cc.

Remark 3.1.

Dummy vertices are only useful for the purposes of our analysis and can be removed at the end of the iteration. Indeed, one could use the technique of “equalizing coin flips” instead. For more details see e.g., [28].

Overall, without loss of generality, at each iteration ii our goal will be to guarantee that Property P(i+1)P(i+1) holds assuming Property Q(i)Q(i).

Property Q(i): For each vertex vViv\in V_{i}, color cLvc\in L_{v} and r[k1]r\in[k-1]:

i(v)\displaystyle\ell_{i}(v) =\displaystyle= Li;\displaystyle L_{i};
ti,r(v,c)\displaystyle t_{i,r}(v,c) =\displaystyle= Ti,r.\displaystyle T_{i,r}.

An iteration.

For the ii-th iteration we apply the Local Lemma with respect to the probability distribution induced by assigning to each vertex vViv\in V_{i} a color chosen uniformly at random from LvL_{v} and activating vv with probability α=KlnΔ\alpha=\frac{K}{\ln\Delta}, where K=(100k3k)1K=(100k^{3k})^{-1}. That is, we apply the Moser-Tardos algorithm in the space induced by the 2|Vi|2|V_{i}| variables corresponding to the color and activation bit of each variable in ViV_{i}. (We will define the family of bad events for each iteration shortly.)

When the execution of the Moser-Tardos algorithm terminates, we uncolor some of the vertices in ViV_{i} in order to get a new partial coloring. In particular, the partial coloring of the hypergraph, set Vi+1V_{i+1}, and the lists of colors for each uncolored vertex in the beginning of iteration i+1i+1 are induced as follows. Let σ\sigma be the output state of the application of the Moser-Tardos algorithm in the ii-th iteration. The list of each vertex vv, Lv(i+1)L_{v}(i+1), is induced from Lv(i)L_{v}(i) by removing every non-available color cLv(i)c\in L_{v}(i) for vv in σ\sigma. We obtain the partial coloring ϕ\phi for the hypergraph and set Vi+1V_{i+1} for the beginning of iteration i+1i+1 by removing the color from every vertex vViv\in V_{i} which is either deactivated or is assigned a non-available for it color in σ\sigma.

Overall, the ii-th iteration of our algorithm can be described at a high-level as follows.

  1. 1.

    Apply the Moser-Tardos algorithm to the probability space induced by assigning to each vertex vViv\in V_{i} a color chosen uniformly at random from Lv(i)L_{v}(i), and activating vv with probability α\alpha.

  2. 2.

    Let σ\sigma be the output state of the Moser-Tardos algorithm.

  3. 3.

    For each vertex vViv\in V_{i}, remove any non-available color cLv(i)c\in L_{v}(i) in σ\sigma to get a list Lv(i+1)L_{v}(i+1).

  4. 4.

    Uncolor every vertex vViv\in V_{i} that has either received a non-available color or is deactivated in σ\sigma, to get a new partial coloring ϕ\phi.

Controlling the parameters of interest.

Next we describe the recursive definitions for LiL_{i} and Ti,rT_{i,r} which, as we already explained, will determine the behavior of the parameters i(v)\ell_{i}(v) and ti,r(v,c)t_{i,r}(v,c), respectively.

Initially, L1=(1+δ)(ΔlnΔ)1k1L_{1}=(1+\delta)\left(\frac{\Delta}{\ln\Delta}\right)^{\frac{1}{k-1}}, T1,k1=ΔT_{1,k-1}=\Delta and T1,r=0T_{1,r}=0 for every r[k2]r\in[k-2]. Letting

Keepi=r=1k1(1(αLi)r)Ti,r,\displaystyle\mathrm{Keep}_{i}=\prod_{r=1}^{k-1}\left(1-\left(\frac{\alpha}{L_{i}}\right)^{r}\right)^{T_{i,r}}, (3)

we define

Li+1\displaystyle L_{i+1} =\displaystyle= LiKeepiLi2/3,\displaystyle L_{i}\cdot\mathrm{Keep}_{i}-L_{i}^{2/3}, (4)
Ti+1,r\displaystyle T_{i+1,r} =\displaystyle= j=rk1(Ti,j(jr)(Keepi(1αKeepi))r(αKeepiLi)jr)\displaystyle\sum_{j=r}^{k-1}\left(T_{i,j}\cdot{j\choose r}\left(\mathrm{Keep}_{i}\left(1-\alpha\mathrm{Keep}_{i}\right)\right)^{r}\left(\frac{\alpha\mathrm{Keep}_{i}}{L_{i}}\right)^{j-r}\right) (5)
+4k2(kr)α(α1Li)rlnΔ=1k1Ti,Li2(lnΔ)2+(j=rk1(jr)αjrTi,jLijr)2/3.\displaystyle+4k^{2(k-r)}\alpha(\alpha^{-1}L_{i})^{r}\ln\Delta\sum_{\ell=1}^{k-1}\frac{T_{i,\ell}}{L_{i}^{2\ell}(\ln\Delta)^{2\ell}}+\left(\sum_{j=r}^{k-1}{j\choose r}\alpha^{j-r}\frac{T_{i,j}}{L_{i}^{j-r}}\right)^{2/3}.

To get some intuition for the recursive definitions (4), (5), observe that Keepi\mathrm{Keep}_{i} is the probability that a color cLv(i)c\in L_{v}(i) is present in Lv(i+1)L_{v}(i+1) as well. Note further that this implies that the expected value of i+1(v,c)\ell_{i+1}(v,c) is LiKeepiL_{i}\cdot\mathrm{Keep}_{i}, a fact which motivates (4). Calculations of similar flavor for 𝔼[ti+1,r(v,c)]{\mathbb{E}}[t_{i+1,r}(v,c)] motivate (5).

The key lemmas.

We are almost ready to state the main lemmas which guarantee that our procedure eventually reaches a partial list-coloring of HH with favorable properties that allow us to extend it to a full list-coloring. Before doing so, we need to settle a subtle issue that has to do with the fact that ti+1,r(v,c)t_{i+1,r}(v,c) is not sufficiently concentrated around its expectation. To see this, notice for example that ti+1,1(v,c)t_{i+1,1}(v,c) drops to zero if vv is assigned cc. (Similarly, for r{2,,k1}r\in\{2,\ldots,k-1\}, if vv is assigned cc then ti+1,r(v,c)t_{i+1,r}(v,c) can be affected by a large amount.) To deal with this problem we will focus instead on variable ti+1,r(v,c)t_{i+1,r}^{\prime}(v,c), i.e., the number of hyperedges hh that contain vv and (i) exactly kr1k-r-1 vertices of h{v}h\setminus\{v\} are colored cc in the end of iteration ii; (ii) the rest rr vertices of h{v}h\setminus\{v\} did not retain their color during iteration ii and, further, cc would be available for them if we ignored the color assigned to vv. Observe that if cc is not assigned to vv then ti+1,r(v,c)=ti+1,r(v,c)t_{i+1,r}(v,c)=t_{i+1,r}^{{}^{\prime}}(v,c) and ti+1,r(v,c)ti+1,r(v,c)t_{i+1,r}^{\prime}(v,c)\geq t_{i+1,r}(v,c) otherwise.

The first lemma that we prove estimates the expected value of the parameters at the end of the ii-th iteration. Its proof can be found in Section 4.

Lemma 3.1.

Let Si==1k1Ti,Li2(lnΔ)2S_{i}=\sum_{\ell=1}^{k-1}\frac{T_{i,\ell}}{L_{i}^{2\ell}(\ln\Delta)^{2\ell}} and Yi,r=j=rk1Ti,jLijY_{i,r}=\sum_{j=r}^{k-1}\frac{T_{i,j}}{L_{i}^{j}}. If Q(i)Q(i) holds and for all 1<j<i,r[k1],Lj(lnΔ)20(k1),Ti,r(lnΔ)20(k1)1<j<i,r\in[k-1],L_{j}\geq(\ln\Delta)^{20(k-1)},T_{i,r}\geq(\ln\Delta)^{20(k-1)}, then, for every vertex vVi+1v\in V_{i+1} and color cLvc\in L_{v}:

  1. (a)

    𝔼[i+1(v)]=i(v)Keepi{\mathbb{E}}[\ell_{i+1}(v)]=\ell_{i}(v)\cdot\mathrm{Keep}_{i};

  2. (b)
    𝔼[ti+1,r(v,c)]\displaystyle{\mathbb{E}}[t_{i+1,r}^{\prime}(v,c)]\leq j=rk1(Ti,j(jr)(Keepi(1αKeepi))r(αKeepiLi)jr)\displaystyle\sum_{j=r}^{k-1}\left(T_{i,j}\cdot{j\choose r}\left(\mathrm{Keep}_{i}\left(1-\alpha\mathrm{Keep}_{i}\right)\right)^{r}\left(\frac{\alpha\mathrm{Keep}_{i}}{L_{i}}\right)^{j-r}\right)
    +4k2(kr)α(α1Li)rSilnΔ+O(Yi,r).\displaystyle+4k^{2(k-r)}\alpha(\alpha^{-1}L_{i})^{r}S_{i}\ln\Delta+O(Y_{i,r}).

The next step is to prove strong concentration around the mean for our random variables per the following lemma. Its proof can be found in Section 4.

Lemma 3.2.

If Q(i)Q(i) holds and Li,Ti,r(lnΔ)20(k1)L_{i},T_{i,r}\geq(\ln\Delta)^{20(k-1)}, r[k1]r\in[k-1], then for every vertex vVi+1v\in V_{i+1} and color cLvc\in L_{v},

  1. (a)

    Pr[|i+1(v)𝔼[i+1(v)]|<Li2/3]<ΔlnΔ\Pr\left[|\ell_{i+1}(v)-{\mathbb{E}}[\ell_{i+1}(v)]|<L_{i}^{2/3}\right]<\Delta^{-\ln\Delta};

  2. (b)

    Pr[ti+1,r(v,c)𝔼[ti+1,r(v,c)]>12(j=rk1(jr)αjrTi,jLijr)2/3]<ΔlnΔ\Pr\left[t_{i+1,r}^{\prime}(v,c)-{\mathbb{E}}[t_{i+1,r}^{\prime}(v,c)]>\frac{1}{2}\left(\sum_{j=r}^{k-1}{j\choose r}\alpha^{j-r}\frac{T_{i,j}}{L_{i}^{j-r}}\right)^{2/3}\right]<\Delta^{-\ln\Delta}.

Armed with Lemmas 3.13.2, a straightforward application of the symmetric Local Lemma, i.e., Corollary 2.3, reveals the following.

Lemma 3.3.

With positive probability, P(i)P(i) holds for every ii such that for all 1<j<i:Lj,Tj,r(lnΔ)20(k1)1<j<i:L_{j},T_{j,r}\geq(\ln\Delta)^{20(k-1)} and Tj,k1110k2Ljk1T_{j,k-1}\geq\frac{1}{10k^{2}}L_{j}^{k-1}.

The proof of Lemma 3.3 can be found in Section 4.

In analyzing the recursive equations (4), (5), it would be helpful if we could ignore the “error terms”. The next lemma shows that this is indeed possible. Its proof can be found in Section 4.

Lemma 3.4.

Define L1=(1+δ)(ΔlnΔ)1k1,T1,k1=ΔL_{1}^{\prime}=(1+\delta)\left(\frac{\Delta}{\ln\Delta}\right)^{\frac{1}{k-1}},T_{1,k-1}^{\prime}=\Delta, T1,r=0T_{1,r}^{\prime}=0 for r[k2]r\in[k-2], and recursively define

Li+1\displaystyle L_{i+1}^{\prime} =\displaystyle= LiKeepi,\displaystyle L_{i}^{\prime}\cdot\mathrm{Keep}_{i}, (6)
Ti+1,r\displaystyle T_{i+1,r}^{\prime} =\displaystyle= j=rk1(Ti,j(jr)(Keepi(1αKeepi))r(αKeepiLi)jr)\displaystyle\sum_{j=r}^{k-1}\left(T_{i,j}^{\prime}\cdot{j\choose r}\left(\mathrm{Keep}_{i}\cdot\left(1-\alpha\mathrm{Keep}_{i}\right)\right)^{r}\left(\frac{\alpha\mathrm{Keep}_{i}}{L_{i}^{\prime}}\right)^{j-r}\right) (7)
+4k2(kr)α(α1Li)rlnΔ=1k1Ti,Li2(lnΔ)2.\displaystyle+4k^{2(k-r)}\alpha(\alpha^{-1}L_{i})^{r}\ln\Delta\sum_{\ell=1}^{k-1}\frac{T_{i,\ell}}{L_{i}^{2\ell}(\ln\Delta)^{2\ell}}.

If for all 1<j<i1<j<i, Lj(lnΔ)20(k1)L_{j}\geq(\ln\Delta)^{20(k-1)}, Tj,r(lnΔ)20(k1)T_{j,r}\geq(\ln\Delta)^{20(k-1)} for every r[k1]r\in[k-1], and Tj,k1Ljk110k2T_{j,k-1}\geq\frac{L_{j}^{k-1}}{10k^{2}}, then

  1. (a)

    |LiLi|(Li)56|L_{i}-L_{i}^{\prime}|\leq(L_{i}^{\prime})^{\frac{5}{6}};

  2. (b)

    |Ti,rTi,r|(Ti,r)100r100r+1|T_{i,r}-T_{i,r}^{\prime}|\leq(T_{i,r}^{\prime})^{\frac{100r}{100r+1}}.

Remark 3.2.

Note that Keepi\mathrm{Keep}_{i} in Lemma 3.4 is still defined in terms of Li,Ti,rL_{i},T_{i,r} and not Li,Ti,rL_{i}^{\prime},T_{i,r}^{\prime}. Note also that in the definition of Ti+1,rT_{i+1,r}^{\prime}, the second summand is a function of Ti,,LiT_{i,\ell},L_{i}, [r1]\ell\in[r-1], and not Ti,,LiT_{i,\ell}^{\prime},L_{i}^{\prime}.

Using Lemma 3.4 we are able to prove the following in Section 4.

Lemma 3.5.

There exists i=O(lnΔlnlnΔ)i^{*}=O(\ln\Delta\ln\ln\Delta) such that

  1. (a)

    For all 1<ii,Ti,r>(lnΔ)20(k1),LiΔϵ/3(k1)(1+ϵ/2)1<i\leq i^{*},T_{i,r}>(\ln\Delta)^{20(k-1)},L_{i}\geq\Delta^{\frac{\epsilon/3}{(k-1)(1+\epsilon/2)}}, and Ti,k1110k2Lik1T_{i,k-1}\geq\frac{1}{10k^{2}}L_{i}^{k-1};

  2. (b)

    Ti+1,r110k2Li+1rT_{i^{*}+1,r}\leq\frac{1}{10k^{2}}L_{i^{*}+1}^{r}, for every r[k1]r\in[k-1] and Li+1Δϵ/3(k1)(1+ϵ/2)L_{i^{*}+1}\geq\Delta^{\frac{\epsilon/3}{(k-1)(1+\epsilon/2)}}.

Remark 3.3.

Notice that Lemmas 3.33.5 imply that with positive probability, after ii^{*} iterations the resulting proper coloring ϕ\phi satisfies property P(i+1)P(i^{*}+1), and therefore condition (b) of Lemma 3.5.

Lemmas 3.33.5 and 3.6 imply Theorem 1.1.

Lemma 3.6.

Let ii^{*} be the integer promised by Lemma 3.5, and assume property P(i+1)P(i^{*}+1) holds for the partial coloring ϕ\phi in the beginning of the (i+1)(i^{*}+1)-th iteration. Given ϕ\phi, we can find a full list-coloring of HH in expected polynomial time in the number of vertices of HH. Also, if Δ\Delta is assumed to be constant, then such a coloring can be constructed deterministically in polynomial time.

Proof of Theorem 1.1.

We carry out ii^{*} iterations of our procedure. If P(i)P(i) fails to hold for any iteration ii, then we halt. By Lemmas 3.3 and 3.5, P(i)P(i) (and, therefore, Q(i)Q(i)) holds with positive probability for each iteration and so it is possible to perform ii^{*} iterations. Further, since the application of the LLL in the proof of Lemma 3.3 is within the scope of the so-called variable setting, i.e., the setting considered by [29], the Moser-Tardos algorithm applies and terminates in expected polynomial time. (If Δ\Delta is assumed to be constant the deterministic version of the Moser-Tardos algorithm also applies and terminates in polynomial time.) Thus, we can perform ii^{*} (successful) iterations in polynomial time.

After ii^{*} iterations we apply the algorithm of Lemma 3.6 and complete the list-coloring of the input hypergraph.

3.1 Proof of Lemma 3.6

Let 𝒰ϕ\mathcal{U}_{\phi} denote the set of uncolored vertices in ϕ\phi, and 𝒰ϕ(h)\mathcal{U}_{\phi}(h) the subset of 𝒰ϕ\mathcal{U}_{\phi} that belongs to a hyperedge hh. Our goal is to color the vertices in 𝒰ϕ\mathcal{U}_{\phi} to get a proper list-coloring.

Towards that end, let Lv=Lv(ϕ)L_{v}=L_{v}(\phi) denote the list of colors for vv in ϕ\phi, and Dr(v,c):=Di+1,r(v,c)D_{r}(v,c):=D_{i^{*}+1,r}(v,c) the set of hyperedges (of size ti+1,r(v,c)t_{i^{*}+1,r}(v,c)) with rr uncolored vertices in ϕ\phi whose vertices “compete” for cc with vv, and recall the conclusion of Lemma 3.5. Let μ\mu be the probability distribution induced by giving each vertex v𝒰ϕv\in\mathcal{U}_{\phi} a color from LvL_{v} uniformly at random. For every hyperedge hh and color cc such that (i) cv𝒰ϕ(h)Lvc\in\bigcap_{v\in\mathcal{U}_{\phi}(h)}L_{v}; and (ii) ϕ(v)=c\phi(v)=c for every vertex in h𝒰ϕ(h)h\setminus\mathcal{U}_{\phi}(h), we define Ah,cA_{h,c} to be the event that all vertices of hh are colored cc. Let 𝒜\mathcal{A} be the family of these (bad) events, and observe that any elementary event (list-coloring) that does not belong in their union is a proper. In other words, if we avoid these bad events we have found a proper list-coloring of the hypergraph. Moreover, for every Ah,c𝒜A_{h,c}\in\mathcal{A}:

μ(Ah,c)1v𝒰ϕ(h)|Lv(ϕ)|<14,\displaystyle\mu\left(A_{h,c}\right)\leq\frac{1}{\prod_{v\in\mathcal{U}_{\phi}(h)}|L_{v}(\phi)|}<\frac{1}{4},

for large enough Δ\Delta, since Li+1=Li+1(Δ)Δ++L_{i^{*}+1}=L_{i^{*}+1}(\Delta)\xrightarrow{\Delta\to+\infty}+\infty.

Define

D(Ah,c):=v𝒰ϕ(h)cLvr=1k1Dr(v,c)\displaystyle D(A_{h,c}):=\bigcup_{v\in\mathcal{U}_{\phi}(h)}\bigcup_{c^{\prime}\in L_{v}}\bigcup_{r=1}^{k-1}D_{r}(v,c^{\prime})

and observe that Ah,cA_{h,c} is mutually independent of the events in 𝒜D(Ah,c)\mathcal{A}\setminus D(A_{h,c}). The existential claim of Lemma 3.6 follows from Corollary 2.2 as, for every Ah,c𝒜A_{h,c}\in\mathcal{A}:

AD(Ah,c)μ(A)\displaystyle\sum_{A\in D(A_{h,c})}\mu(A) \displaystyle\leq v𝒰ϕ(h)cLvr=1k1hDr(v,c)μ(Ah,c)\displaystyle\sum_{v\in\mathcal{U}_{\phi}(h)}\sum_{c^{\prime}\in L_{v}}\sum_{r=1}^{k-1}\sum_{h^{\prime}\in D_{r}(v,c^{\prime})}\mu\left(A_{h^{\prime},c^{\prime}}\right) (8)
=\displaystyle= v𝒰ϕ(h)cLvi=1k1hDr(v,c)1u𝒰ϕ(h)|Lu|\displaystyle\sum_{v\in\mathcal{U}_{\phi}(h)}\sum_{c^{\prime}\in L_{v}}\sum_{i=1}^{k-1}\sum_{h^{\prime}\in D_{r}(v,c^{\prime})}\frac{1}{\prod_{u\in\mathcal{U}_{\phi}(h^{\prime})}|L_{u}|}
\displaystyle\leq maxv𝒰ϕ(h)k|Lv|cLvr=1k1|Dr(v,c)|Li+1r\displaystyle\max_{v\in\mathcal{U}_{\phi}(h)}\frac{k}{|L_{v}|}\sum_{c^{\prime}\in L_{v}}\sum_{r=1}^{k-1}\frac{|D_{r}(v,c^{\prime})|}{L_{i^{*}+1}^{r}}
\displaystyle\leq k10k2maxv𝒰ϕ(h)Li+1r|Lv||Lv|Li+1r\displaystyle\frac{k}{10k^{2}}\max_{v\in\mathcal{U}_{\phi}(h)}\frac{L_{i^{*}+1}^{r}\cdot|L_{v}|}{|L_{v}|\cdot L_{i^{*}+1}^{r}} (9)
\displaystyle\leq 110<14,\displaystyle\frac{1}{10}<\frac{1}{4}, (10)

concluding the proof. Note that in (8) we used the facts that every hyperedge has at most kk vertices and Li+1Δϵ/3(k1)(1+ϵ/2)L_{i^{*}+1}\geq\Delta^{\frac{\epsilon/3}{(k-1)(1+\epsilon/2)}}, and in (9) we used the fact that |Dr(v,c)|Ti+1r110k2Li+1r|D_{r}(v,c^{\prime})|\leq T_{i^{*}+1}^{r}\leq\frac{1}{10k^{2}}L_{i^{*}+1}^{r}.

As for the algorithmic claim of the lemma, since we apply the LLL in the variable setting the Moser-Tardos algorithm applies and it terminates in expected polynomial time. Further, if Δ\Delta is assumed to be constant then its deterministic version also applies and terminates in polynomial time, concluding the proof of the lemma.

4 Hypergraph list-coloring proofs

In this section we prove Lemmas 3.13.23.33.43.5.

We start by showing a couple of important technical lemmas that will be helpful for these proofs. It will be convenient to define Ri,r=Ti,rLirR_{i,r}=\frac{T_{i,r}}{L_{i}^{r}}, Ri,r=Ti,r(Li)rR_{i,r}^{\prime}=\frac{T_{i,r}^{\prime}}{(L_{i}^{\prime})^{r}} for every r[k1]r\in[k-1].

Lemma 4.1.

If for all 1<j<i,r[k1],Lj,Tj,r(lnΔ)20(k1)1<j<i,r\in[k-1],L_{j},T_{j,r}\geq(\ln\Delta)^{20(k-1)}, then

Ri,rk2(k1r)lnΔ.\displaystyle R_{i,r}\leq k^{2(k-1-r)}\ln\Delta.

The proof of Lemma 4.1 can be found in Appendix A. A straightforward corollary of Lemma 4.1 is the following.

Corollary 4.2.

If Li,Ti,r(lnΔ)20(k1)L_{i},T_{i,r}\geq(\ln\Delta)^{20(k-1)} and Ri,k1110k2R_{i,k-1}\geq\frac{1}{10k^{2}}, then

C:=exp(Kk2(k2)1δ100k)Keepi1Kk112k2(lnΔ)k1.\displaystyle C:=\mathrm{exp}\left(-\frac{Kk^{2(k-2)}}{1-\frac{\delta}{100k}}\right)\leq\mathrm{Keep}_{i}\leq 1-\frac{K^{k-1}}{12k^{2}(\ln\Delta)^{k-1}}.
Proof.

The lower bound follows directly from (59). The upper bound follows from our assumption that Ri,k1110k2R_{i,k-1}\geq\frac{1}{10k^{2}} which implies that

Keepier=1k1αrRi,reαk1Ri,k1eKk110k2(lnΔ)k1<1Kk112k2(lnΔ)k1,\displaystyle\mathrm{Keep}_{i}\leq\mathrm{e}^{-\sum_{r=1}^{k-1}\alpha^{r}R_{i,r}}\leq\mathrm{e}^{-\alpha^{k-1}R_{i,k-1}}\leq\mathrm{e}^{-\frac{K^{k-1}}{10k^{2}(\ln\Delta)^{k-1}}}<1-\frac{K^{k-1}}{12k^{2}(\ln\Delta)^{k-1}},

for sufficiently large Δ\Delta. ∎

The proof of the following lemma can also be found in Appendix A.

Lemma 4.3.

If Lj,Tj,r(lnΔ)20(k1)L_{j},T_{j,r}\geq(\ln\Delta)^{20(k-1)} for all 1<j<i1<j<i, then for every r[k1]r\in[k-1]:

Ri,r(1αC)r(i1)lnΔ(1+δk100)k1r(1+δδk99)k1Ck1rp=rk2(p+1).\displaystyle R_{i,r}^{\prime}\leq(1-\alpha C)^{r(i-1)}\ln\Delta\cdot\frac{(1+\frac{\delta}{k^{100}})^{k-1-r}}{(1+\delta-\frac{\delta}{k^{99}})^{k-1}C^{k-1-r}}\prod_{p=r}^{k-2}(p+1).

We are now ready to prove Lemmas 3.13.23.33.4 and 3.5.

4.1 Proof of Lemma 3.1

Proof of part (a).

For every color cLv(i)c\in L_{v}(i),

Pr[cLv(i+1)]=r=1k1hDi,r(v,c)(1u(h{v})Viαi(u))=r=1k1(1(αLi)r)Ti,r=Keepi,\displaystyle\Pr[c\in L_{v}(i+1)]=\prod_{r=1}^{k-1}\prod_{h\in D_{i,r}(v,c)}\left(1-\prod_{u\in(h\setminus\{v\})\cap V_{i}}\frac{\alpha}{\ell_{i}(u)}\right)=\prod_{r=1}^{k-1}\left(1-\left(\frac{\alpha}{L_{i}}\right)^{r}\right)^{T_{i,r}}=\mathrm{Keep}_{i}, (11)

where for the second equality we used our assumption that Q(i)Q(i) holds. Therefore, the proof of the first part of the lemma follows from the linearity of expectation.

Proof of part (b) .

Recall the definition of ti+1,r(v,c)t_{i+1,r}^{\prime}(v,c) and note that only hyperedges in j=rk1Di,j(v,c)\bigcup_{j=r}^{k-1}D_{i,j}(v,c) can be potentially counted by ti+1,r(v,c)t_{i+1,r}^{\prime}(v,c). In particular, unless every uncolored vertex of an edge hDi,j(v,c)h\in D_{i,j}(v,c), jrj\geq r, is assigned the same color with vv in iteration ii, then if hh is counted by ti+1,r(c)t_{i+1,r}^{\prime}(c), it is also counted by ti+1,r(c)t_{i+1,r}(c). Therefore,

𝔼[ti+1,r(v,c)]𝔼[ti+1,r(v,c)]+O(j=rk1Ti,jLij),\displaystyle{\mathbb{E}}[t_{i+1,r}^{\prime}(v,c)]\leq{\mathbb{E}}[t_{i+1,r}(v,c)]+O\left(\sum_{j=r}^{k-1}\frac{T_{i,j}}{L_{i}^{j}}\right), (12)

and so we focus on bounding 𝔼[ti+1,r(v,c)]{\mathbb{E}}[t_{i+1,r}(v,c)].

Fix hDi,j(v,c)h\in D_{i,j}(v,c), where jrj\geq r. Our goal will be to show that

Pr[hDi+1,r(v,c)]\displaystyle\Pr[h\in D_{i+1,r}(v,c)]\leq (jr)(Keepi(1αKeepi))r(αKeepiLi)jr\displaystyle{j\choose r}\left(\mathrm{Keep}_{i}(1-\alpha\mathrm{Keep}_{i})\right)^{r}\left(\frac{\alpha\mathrm{Keep}_{i}}{L_{i}}\right)^{j-r}
+4r(jr)Keepij1αjr+1SiLijr+O(1Lij),\displaystyle+4r{j\choose r}\frac{\mathrm{Keep}_{i}^{j-1}\alpha^{j-r+1}S_{i}}{L_{i}^{j-r}}+O\left(\frac{1}{L_{i}^{j}}\right), (13)

since combining (4.1) with (12) implies the lemma. To see this, observe that

Ti,j4r(jr)Keepij1αjr+1SiLijr\displaystyle T_{i,j}\cdot 4r{j\choose r}\frac{\mathrm{Keep}_{i}^{j-1}\alpha^{j-r+1}S_{i}}{L_{i}^{j-r}} =4αr(jr)αjTi,jLijKeepij1(α1Li)rSi\displaystyle=4\alpha r{j\choose r}\cdot\alpha^{j}\frac{T_{i,j}}{L_{i}^{j}}\mathrm{Keep}_{i}^{j-1}\cdot(\alpha^{-1}L_{i})^{r}S_{i}
{4Ti,1Si,if j=r=1,4αr(jr)e(j1)(α1Li)rSiotherwise.\displaystyle\leq\begin{cases}4T_{i,1}S_{i},&\text{if }j=r=1,\\ \frac{4\alpha r{j\choose r}}{\mathrm{e}(j-1)}(\alpha^{-1}L_{i})^{r}S_{i}&\text{otherwise.}\end{cases} (14)

Note that in deriving the second part of the inequality in (14) we first used that 1xex1-x\leq\mathrm{e}^{-x} for every x0x\geq 0 in order to bound Keepi\mathrm{Keep}_{i} by exp(Ti,j/Lij)\mathrm{exp}(-T_{i,j}/L_{i}^{j}) , and then that maxxxex1e\max_{x}x\mathrm{e}^{-\ell x}\leq\frac{1}{\ell\mathrm{e}} for every \ell. Therefore,

𝔼[ti+1,r(v,c)]\displaystyle{\mathbb{E}}[t_{i+1,r}(v,c)] \displaystyle\leq j=rk1Ti,jmaxhDi,j(v,c)Pr[hDi+1,r(v,c)]\displaystyle\sum_{j=r}^{k-1}T_{i,j}\max_{h\in D_{i,j}(v,c)}\Pr[h\in D_{i+1,r}(v,c)] (16)
<\displaystyle< j=rk1(Ti,j(jr)(Keepi(1αKeepi))r(αKeepiLi)jr)\displaystyle\sum_{j=r}^{k-1}\left(T_{i,j}\cdot{j\choose r}\left(\mathrm{Keep}_{i}\left(1-\alpha\mathrm{Keep}_{i}\right)\right)^{r}\left(\frac{\alpha\mathrm{Keep}_{i}}{L_{i}}\right)^{j-r}\right)
+4k2(kr)α(α1Li)rSilnΔ+O(j=rk1Ti,jLij).\displaystyle+4k^{2(k-r)}\alpha(\alpha^{-1}L_{i})^{r}S_{i}\ln\Delta+O\left(\sum_{j=r}^{k-1}\frac{T_{i,j}}{L_{i}^{j}}\right).

for sufficiently large Δ\Delta. In deriving (16) we used (14) and the facts that:

Ti,1\displaystyle T_{i,1} =Ti,1LiLiLik2(k1r)lnΔ, according to Lemma 4.1;\displaystyle=\frac{T_{i,1}}{L_{i}}\cdot L_{i}\leq L_{i}\cdot k^{2(k-1-r)}\ln\Delta,\enspace\mbox{ according to Lemma~{}\ref{bounding_keep_lemma};}
j=rk1r(jr)e(j1)\displaystyle\sum_{j=r}^{k-1}\frac{r{j\choose r}}{\mathrm{e}(j-1)} <lnΔk2(k1r) for sufficiently large Δ and j>1.\displaystyle<\ln\Delta\cdot k^{2(k-1-r)}\mbox{ for sufficiently large $\Delta$ and $j>1$.}

Towards proving (4.1), for any vertex uh{v}u\in h\setminus\{v\}, consider the events

Eu,1\displaystyle E_{u,1} =\displaystyle= u does not retain its color and cLu(i+1),\displaystyle\text{ ``$u$ does not retain its color and $c\in L_{u}(i+1)$"},
Eu,2\displaystyle E_{u,2} =\displaystyle= u is assigned c and retains its color”.\displaystyle\text{ ``$u$ is assigned $c$ and retains its color"}.

Let also BcB_{c} be the event that vv and j1j-1 other uncolored vertices of hh receive color cc. Since we have assumed that our hypergraph is of girth at least 55, for any neighbor uu of vv and f{1,2}f\in\{1,2\} the event Eu,fE_{u,f} is mutually independent of all events Eu,,{1,2},uuE_{u^{\prime},\ell},\ell\in\{1,2\},u\neq u^{\prime}, conditional on BcB_{c} not occurring. Thus, if Pr[Eu,Bc¯]p\Pr[E_{u,\ell}\mid\overline{B_{c}}]\leq p_{\ell}, {1,2}\ell\in\{1,2\}, for every vertex uh{v}u\in h\setminus\{v\}, we obtain

Pr[hDi+1,r(v,c)](jr)p1rp2jr+Pr[Bc](jr)p1rp2jr+2kLij,\displaystyle\Pr[h\in D_{i+1,r}(v,c)]\leq{j\choose r}p_{1}^{r}p_{2}^{j-r}+\Pr[B_{c}]\leq{j\choose r}p_{1}^{r}p_{2}^{j-r}+\frac{2^{k}}{L_{i}^{j}}, (17)

since Pr[Bc]2kLij\Pr[B_{c}]\leq 2^{k}L_{i}^{-j}.

Now we claim that for any uh{v}u\in h\setminus\{v\}, and sufficiently large Δ\Delta,

Pr[Eu,2Bc¯]αKeepiLi+2(LilnΔ)j=:q2+δ2.\displaystyle\Pr[E_{u,2}\mid\overline{B_{c}}]\leq\frac{\alpha\mathrm{Keep}_{i}}{L_{i}}+\frac{2}{(L_{i}\ln\Delta)^{j}}=:q_{2}+\delta_{2}. (18)

To see this, notice that conditional on Bc¯\overline{B_{c}} the probability that uu is activated is α\alpha, it is assigned cc with probability at most 1/Li1/L_{i}, and it retains cc with probability at most

r[k1]{j}(1αrLir)Ti,r(1αjLij)Ti,j1=Keepi1αjLij.\displaystyle\prod_{r\in[k-1]\setminus\{j\}}\left(1-\frac{\alpha^{r}}{L_{i}^{r}}\right)^{T_{i,r}}\cdot\left(1-\frac{\alpha^{j}}{L_{i}^{j}}\right)^{T_{i,j}-1}=\frac{\mathrm{Keep}_{i}}{1-\frac{\alpha^{j}}{L_{i}^{j}}}. (19)

Thus,

Pr[Eu,2Bc¯]=αKeepiLi(1αjLij)αKeepiLi(1+2αjLij)αKeepiLi+2(LilnΔ)j.\displaystyle\Pr[E_{u,2}\mid\overline{B_{c}}]=\frac{\alpha\mathrm{Keep}_{i}}{L_{i}(1-\frac{\alpha^{j}}{L_{i}^{j}})}\leq\frac{\alpha\cdot\mathrm{Keep}_{i}}{L_{i}}\cdot\left(1+\frac{2\alpha^{j}}{L_{i}^{j}}\right)\leq\frac{\alpha\mathrm{Keep}_{i}}{L_{i}}+\frac{2}{(L_{i}\ln\Delta)^{j}}.

for sufficiently large Δ\Delta, concluding the proof of (18).

Further, we claim that

Pr[Eu,1Bc¯]Keepi(1αKeepi)+(2αSi+(LilnΔ)j(3+4αSi))=:q1+δ1.\displaystyle\Pr[E_{u,1}\mid\overline{B_{c}}]\leq\mathrm{Keep}_{i}(1-\alpha\mathrm{Keep}_{i})+\left(2\alpha S_{i}+(L_{i}\ln\Delta)^{-j}(3+4\alpha S_{i})\right)=:q_{1}+\delta_{1}. (20)

To show (20) we consider three cases. The first case is that uu is not activated and cLu(i+1)c\in L_{u}(i+1) (notice that these are two independent events). In this case uu will not retain its color, and observe that

Pr[cLu(i+1)Bc¯]Keepi+2(LilnΔ)j,\displaystyle\Pr[c\in L_{u}(i+1)\mid\overline{B_{c}}]\leq\mathrm{Keep}_{i}+2(L_{i}\ln\Delta)^{-j}, (21)

using essentially the same calculations as in showing (18). Thus,

Pr[u is not activated and cLu(i+1)Bc¯](1α)(Keepi+2(LilnΔ)j).\displaystyle\Pr[\text{$u$ is not activated and $c\in L_{u}(i+1)$}\mid\overline{B_{c}}]\leq(1-\alpha)\left(\mathrm{Keep}_{i}+2(L_{i}\ln\Delta)^{-j}\right). (22)

In the second case we consider the scenario where uu is activated and is assigned cc. Clearly then, the probability that cLu(i+1)c\in L_{u}(i+1) and uu does not retain cc is zero. Finally, suppose that uu is activated and is assigned a color γc\gamma\neq c. Our goal is to compute Pr[(u is activated and assigned γ)Eu,1Bc¯]\Pr[\text{($u$ is activated and assigned $\gamma$)}\wedge E_{u,1}\mid\overline{B_{c}}] for each γ\gamma so that we can sum up these probabilities over all possible γc\gamma\neq c along with (22).

For a vertex ww let FwγF_{w}^{\gamma} denote the event that ww is activated and assigned γ\gamma. Using this notation we have:

Pr[FuγEu,1Bc¯]\displaystyle\Pr[F_{u}^{\gamma}\wedge E_{u,1}\mid\overline{B_{c}}] =Pr[FuγBc¯]Pr[(γLu(i+1))(cLu(i+1))Fuγ,Bc¯]\displaystyle=\Pr[F_{u}^{\gamma}\mid\overline{B_{c}}]\cdot\Pr[(\gamma\notin L_{u}(i+1))\wedge(c\in L_{u}(i+1))\mid F_{u}^{\gamma},\overline{B_{c}}]
=αLiPr[(γLu(i+1))(cLu(i+1))Fuγ,Bc¯]\displaystyle=\frac{\alpha}{L_{i}}\cdot\Pr[(\gamma\notin L_{u}(i+1))\wedge(c\in L_{u}(i+1))\mid F_{u}^{\gamma},\overline{B_{c}}]
=αLiPr[cLu(i+1)Fuγ,Bc¯]Pr[γLu(i+1)cLu(i+1),Fuγ,Bc¯]\displaystyle=\frac{\alpha}{L_{i}}\Pr[c\in L_{u}(i+1)\mid F_{u}^{\gamma},\overline{B_{c}}]\cdot\Pr[\gamma\not\in L_{u}(i+1)\mid c\in L_{u}(i+1),F_{u}^{\gamma},\overline{B_{c}}]
=αLiPr[cLu(i+1)Bc¯]Pr[γLu(i+1)cLu(i+1),Fuγ,Bc¯]\displaystyle=\frac{\alpha}{L_{i}}\Pr[c\in L_{u}(i+1)\mid\overline{B_{c}}]\cdot\Pr[\gamma\not\in L_{u}(i+1)\mid c\in L_{u}(i+1),F_{u}^{\gamma},\overline{B_{c}}]
αLi(Keepi+2(LilnΔ)j)Pr[γLu(i+1)cLu(i+1),Fuγ,Bc¯]\displaystyle\leq\frac{\alpha}{L_{i}}\cdot(\mathrm{Keep}_{i}+2(L_{i}\ln\Delta)^{-j})\cdot\Pr[\gamma\not\in L_{u}(i+1)\mid c\in L_{u}(i+1),F_{u}^{\gamma},\overline{B_{c}}] (23)

and so below we focus on bounding for each γLu(i){c}\gamma\in L_{u}(i)\setminus\{c\} the probability that γLu(i+1)\gamma\notin L_{u}(i+1) and cLu(i+1)c\in L_{u}(i+1) conditional on that uu is activated and assigned γ\gamma and BcB_{c} did not occur. Note that in deriving (23) we used (21).

We have:

Pr[γLu(i+1)cLu(i+1),Fuγ,Bc¯]=1Pr[γLu(i+1)cLu(i+1),Fuγ,Bc¯]\displaystyle\Pr[\gamma\not\in L_{u}(i+1)\mid c\in L_{u}(i+1),F_{u}^{\gamma},\overline{B_{c}}]=1-\Pr[\gamma\in L_{u}(i+1)\mid c\in L_{u}(i+1),F_{u}^{\gamma},\overline{B_{c}}]
=1=1k1gDi,(u,γ)(1Pr[w(g{u})ViFwγcLu(i+1),Fuγ,Bc¯])\displaystyle=1-\prod_{\ell=1}^{k-1}\prod_{g\in D_{i,\ell}(u,\gamma)}\left(1-\Pr[\cap_{w\in(g\setminus\{u\})\cap V_{i}}F_{w}^{\gamma}\mid c\in L_{u}(i+1),F_{u}^{\gamma},\overline{B_{c}}]\right) (24)
=1=1k1gDi,(u,γ)(1Pr[w(g{u})ViFwγcLu(i+1),Bc¯])\displaystyle=1-\prod_{\ell=1}^{k-1}\prod_{g\in D_{i,\ell}(u,\gamma)}\left(1-\Pr[\cap_{w\in(g\setminus\{u\})\cap V_{i}}F_{w}^{\gamma}\mid c\in L_{u}(i+1),\overline{B_{c}}]\right) (25)

Note that in deriving (24) we use the fact the girth of the hypergraph is at least 55 which, in particular, implies that any two vertices that contain uu do not have any other vertex in common.

To further bound (25), we consider the probability that every vertex in (g{u})Vi(g\setminus\{u\})\cap V_{i} is activated and assigned γ\gamma, conditional on that cLu(i+1)c\in L_{u}(i+1) and Bc¯\overline{B_{c}}, for any fixed [k1]\ell\in[k-1] and gDi,(u,γ)g\in D_{i,\ell}(u,\gamma). We consider two cases depending on whether g=hg=h or not.

We start with the case where ghg\neq h. Let AgA_{g} be the event that not every vertex in (g{u})Vi(g\setminus\{u\})\cap V_{i} is activated and assigned cc. Since our hypergraph has girth at least 55 and the color activations and color assignments are independent over different vertices, we have:

Pr[w(g{u})ViFwγcLu(i+1),Bc¯]\displaystyle\Pr[\cap_{w\in(g\setminus\{u\})\cap V_{i}}F_{w}^{\gamma}\mid c\in L_{u}(i+1),\overline{B_{c}}] =Pr[w(g{u})ViFwγAg]\displaystyle=\Pr[\cap_{w\in(g\setminus\{u\})\cap V_{i}}F_{w}^{\gamma}\mid A_{g}]
=Pr[(w(g{u})ViFwγ)Ag]Pr[Ag]\displaystyle=\frac{\Pr[(\cap_{w\in(g\setminus\{u\})\cap V_{i}}F_{w}^{\gamma})\wedge A_{g}]}{\Pr[A_{g}]}
=Pr[w(g{u})ViFwγ]Pr[Ag]\displaystyle=\frac{\Pr[\cap_{w\in(g\setminus\{u\})\cap V_{i}}F_{w}^{\gamma}]}{\Pr[A_{g}]}
=αLi1αLi(αLi)+1Li2(lnΔ)2,\displaystyle=\frac{\alpha^{\ell}L_{i}^{-\ell}}{1-\alpha^{\ell}L_{i}^{-\ell}}\leq\left(\frac{\alpha}{L_{i}}\right)^{\ell}+\frac{1}{L_{i}^{2\ell}(\ln\Delta)^{2\ell}}, (26)

for sufficiently large Δ\Delta, since K<1K<1.

Next we consider the case g=hg=h. The difference here is that the event w(h{u})ViFwγ\cap_{w\in(h\setminus\{u\})\cap V_{i}}F_{w}^{\gamma} is not independent of Bc¯\overline{B_{c}} as before. However, notice that since gDi,(u,γ)g\in D_{i,\ell}(u,\gamma), hDi,j(v,c)h\in D_{i,j}(v,c) and γc\gamma\neq c, we can only have g=hg=h when j==k1j=\ell=k-1. This means that the occurrence of event the event Bc¯\overline{B_{c}} prohibits the occurrence of event event AhA_{h}. Therefore, the event w(h{u})ViFwγ\cap_{w\in(h\setminus\{u\})\cap V_{i}}F_{w}^{\gamma} is independent of the event cLu(i+1)c\in L_{u}(i+1) and, thus, we have:

Pr[w(h{u})ViFwγcLu(i+1),Bc¯]\displaystyle\Pr[\cap_{w\in(h\setminus\{u\})\cap V_{i}}F_{w}^{\gamma}\mid c\in L_{u}(i+1),\overline{B_{c}}] =Pr[w(h{u})ViFwγBc¯]\displaystyle=\Pr[\cap_{w\in(h\setminus\{u\})\cap V_{i}}F_{w}^{\gamma}\mid\overline{B_{c}}]
=Pr[w(h{u})ViFwγ]Pr[Bc¯]\displaystyle=\frac{\Pr[\cap_{w\in(h\setminus\{u\})\cap V_{i}}F_{w}^{\gamma}]}{\Pr[\overline{B_{c}}]}
=αk1Li(k1)1Li(k1)2(αLi)k1,\displaystyle=\frac{\alpha^{k-1}L_{i}^{-(k-1)}}{1-L_{i}^{-(k-1)}}\leq 2\left(\frac{\alpha}{L_{i}}\right)^{k-1}, (27)

for sufficiently large Δ\Delta.

Combining (25), (26) and (27), we are able to show the following proposition.

Proposition 4.4.

For every color γc\gamma\neq c:

Pr[γLu(i+1)cLu(i+1),Fuγ,Bc¯]1Keepi+2=1k1Ti,Li2(lnΔ)2.\displaystyle\Pr[\gamma\not\in L_{u}(i+1)\mid c\in L_{u}(i+1),F_{u}^{\gamma},\overline{B_{c}}]\leq 1-\mathrm{Keep}_{i}+2\sum_{\ell=1}^{k-1}\frac{T_{i,\ell}}{L_{i}^{2\ell}(\ln\Delta)^{2\ell}}. (28)
Proof.

Towards proving (28), a helpful observation is the following.

=1k1(1(αLi)1Li2(lnΔ)2)Ti,(u,γ)\displaystyle\prod_{\ell=1}^{k-1}\left(1-\left(\frac{\alpha}{L_{i}}\right)^{\ell}-\frac{1}{L_{i}^{2\ell}(\ln\Delta)^{2\ell}}\right)^{T_{i,\ell}(u,\gamma)} ==1k1(1(αLi))Ti,(u,γ)\displaystyle=\prod_{\ell=1}^{k-1}\left(1-\left(\frac{\alpha}{L_{i}}\right)^{\ell}\right)^{T_{i,\ell}(u,\gamma)}
×(11Li2(lnΔ)2(1(αLi)))Ti,(u,γ)\displaystyle\times\left(1-\frac{1}{L_{i}^{2\ell}(\ln\Delta)^{2\ell}\cdot(1-(\frac{\alpha}{L_{i}})^{\ell})}\right)^{T_{i,\ell}(u,\gamma)}
=Keepiexp(=1k1Ti,(u,γ)Li2(lnΔ)2(1(αLi))1)\displaystyle=\mathrm{Keep}_{i}\cdot\mathrm{exp}\left(-\sum_{\ell=1}^{k-1}\frac{T_{i,\ell}(u,\gamma)}{L_{i}^{2\ell}(\ln\Delta)^{2\ell}\cdot(1-(\frac{\alpha}{L_{i}})^{\ell})-1}\right) (29)
Keepi(132=1k1Ti,Li2(lnΔ)2(1(αLi))1)\displaystyle\geq\mathrm{Keep}_{i}\left(1-\frac{3}{2}\cdot\sum_{\ell=1}^{k-1}\frac{T_{i,\ell}}{L_{i}^{2\ell}(\ln\Delta)^{2\ell}\cdot(1-(\frac{\alpha}{L_{i}})^{\ell})-1}\right)
Keepi(132=1k1Ti,Li2(lnΔ)2Li(lnΔ))\displaystyle\geq\mathrm{Keep}_{i}\left(1-\frac{3}{2}\cdot\sum_{\ell=1}^{k-1}\frac{T_{i,\ell}}{L_{i}^{2\ell}(\ln\Delta)^{2\ell}-L_{i}^{\ell}(\ln\Delta)^{\ell}}\right)
Keepi1910=1k1Ti,Li2(lnΔ)2.\displaystyle\geq\mathrm{Keep}_{i}-\frac{19}{10}\sum_{\ell=1}^{k-1}\frac{T_{i,\ell}}{L_{i}^{2\ell}(\ln\Delta)^{2\ell}}. (30)

for sufficiently large Δ\Delta. Note that in (29) we used the fact that 1xe1x11-x\geq\mathrm{e}^{\frac{1}{x-1}} for any x2x\geq 2.

Using (26), (27) and (30), we have:

=1k1gDi,(u,γ)(1Pr[w(g{u})ViFwγcLu(i+1),Bc¯])\displaystyle\prod_{\ell=1}^{k-1}\prod_{g\in D_{i,\ell}(u,\gamma)}\left(1-\Pr[\cap_{w\in(g\setminus\{u\})\cap V_{i}}F_{w}^{\gamma}\mid c\in L_{u}(i+1),\overline{B_{c}}]\right)\geq
=1k1(1(αLi)1Li2(lnΔ)2)Ti,(u,γ)(12(αLi)k1)\displaystyle\geq\prod_{\ell=1}^{k-1}\left(1-\left(\frac{\alpha}{L_{i}}\right)^{\ell}-\frac{1}{L_{i}^{2\ell}(\ln\Delta)^{2\ell}}\right)^{T_{i,\ell}(u,\gamma)}\left(1-2\left(\frac{\alpha}{L_{i}}\right)^{k-1}\right) (31)
(Keepi1910=1k1Ti,Li2(lnΔ)2)(12(αLi)k1)\displaystyle\geq\left(\mathrm{Keep}_{i}-\frac{19}{10}\sum_{\ell=1}^{k-1}\frac{T_{i,\ell}}{L_{i}^{2\ell}(\ln\Delta)^{2\ell}}\right)\left(1-2\left(\frac{\alpha}{L_{i}}\right)^{k-1}\right) (32)
>Keepi1910=1k1Ti,Li2(lnΔ)22Keepi(αLi)k1\displaystyle>\mathrm{Keep}_{i}-\frac{19}{10}\sum_{\ell=1}^{k-1}\frac{T_{i,\ell}}{L_{i}^{2\ell}(\ln\Delta)^{2\ell}}-2\cdot\mathrm{Keep}_{i}\left(\frac{\alpha}{L_{i}}\right)^{k-1}
Keepi2=1k1Ti,Li2(lnΔ)2,\displaystyle\geq\mathrm{Keep}_{i}-2\sum_{\ell=1}^{k-1}\frac{T_{i,\ell}}{L_{i}^{2\ell}(\ln\Delta)^{2\ell}}, (33)

for sufficiently large Δ\Delta. Note that in deriving (31) we used our previous observation that (27) applies only when h=gh=g, and this can potentially happen only when =j=k1\ell=j=k-1.

Combining (25) with (33) concludes the proof.

Overall, combining (22), (23) and Proposition 4.4 we see that Pr[Eu,1Bc¯]\Pr[E_{u,1}\mid\overline{B_{c}}] is at most

(1α)Keepi+2(LilnΔ)j+αLi1Li(Keepi+2(LilnΔ)j)(1Keepi+2=1k1Ti,Li2(lnΔ)2),\displaystyle(1-\alpha)\mathrm{Keep}_{i}+2(L_{i}\ln\Delta)^{-j}+\alpha\frac{L_{i}-1}{L_{i}}(\mathrm{Keep}_{i}+2(L_{i}\ln\Delta)^{-j})\left(1-\mathrm{Keep}_{i}+2\sum_{\ell=1}^{k-1}\frac{T_{i,\ell}}{L_{i}^{2\ell}(\ln\Delta)^{2\ell}}\right),
\displaystyle\leq\, Keepi(1αKeepi)+(2αSi+(LilnΔ)j(3+4αSi)),\displaystyle\mathrm{Keep}_{i}(1-\alpha\mathrm{Keep}_{i})+\left(2\alpha S_{i}+(L_{i}\ln\Delta)^{-j}(3+4\alpha S_{i})\right),

where recall that Si==1k1Ti,Li2(lnΔ)2S_{i}=\sum_{\ell=1}^{k-1}\frac{T_{i,\ell}}{L_{i}^{2\ell}(\ln\Delta)^{2\ell}}. This concludes the proof of (20).

Finally, combining (17), (18) and  (20) we obtain

Pr[hDi+1,r(v,c)]\displaystyle\Pr[h\in D_{i+1,r}(v,c)] \displaystyle\leq (jr)q1rq2jr(1+δ1q11)r(1+δ2q21)jr+2kLij\displaystyle{j\choose r}q_{1}^{r}q_{2}^{j-r}\left(1+\delta_{1}q_{1}^{-1}\right)^{r}\left(1+\delta_{2}q_{2}^{-1}\right)^{j-r}+\frac{2^{k}}{L_{i}^{j}} (34)
\displaystyle\leq (jr)q1rq2jr(1+2rδ1q11)(1+2(jr)δ2q21)+2kLij\displaystyle{j\choose r}q_{1}^{r}q_{2}^{j-r}\left(1+2r\delta_{1}q_{1}^{-1}\right)\left(1+2(j-r)\delta_{2}q_{2}^{-1}\right)+\frac{2^{k}}{L_{i}^{j}}
\displaystyle\leq (jr)q1rq2jr+2r(jr)q1r1q2jrδ1+O(1Lij)\displaystyle{j\choose r}q_{1}^{r}q_{2}^{j-r}+2r{j\choose r}q_{1}^{r-1}q_{2}^{j-r}\delta_{1}+O\left(\frac{1}{L_{i}^{j}}\right) (35)
\displaystyle\leq (jr)q1rq2jr+4r(jr)Keepij1αjr+1SiLijr+O(1Lij),\displaystyle{j\choose r}q_{1}^{r}q_{2}^{j-r}+4r{j\choose r}\frac{\mathrm{Keep}_{i}^{j-1}\alpha^{j-r+1}S_{i}}{L_{i}^{j-r}}+O\left(\frac{1}{L_{i}^{j}}\right), (36)

concluding the proof of (4.1), which was our goal. Note that in (34) we used that Keepi\mathrm{Keep}_{i} is bounded below by a constant according to Corollary 4.2 (the lower bound only requires the assumptions of Lemma 4.1). In (35) we used the fact that δ2q21=O(1/Lij)\delta_{2}q_{2}^{-1}=O(1/L_{i}^{j}) for j2j\geq 2. (We only care about j2j\geq 2 since if j=1j=1 then r=1r=1 as well and, therefore, jr=0j-r=0.) ∎

4.2 Proof of Lemma 3.2

Let Bin(n,p)\mathrm{Bin}(n,p) denote the binomial random variable that counts the number of successes in nn Bernoulli trials, where each trial succeeds with probability pp. We will find the following lemma useful (see, e.g., Exercise 2.12 in [28]) :

Lemma 4.5.

For any c,k,nc,k,n we have

Pr[Bin(n,cn)k]ckk!.\displaystyle\Pr\left[\mathrm{Bin}\left(n,\frac{c}{n}\right)\geq k\right]\leq\frac{c^{k}}{k!}.
Proof of Part (a).

We will use Theorem 2.4 to show that that the number of colors, v¯\overline{\ell_{v}}, which are removed from LvL_{v} during iteration ii is highly concentrated.

Note that changing the assignment to any neighboring vertex of vv can change v¯\overline{\ell_{v}} by at most 11, and changing the assignment to any other vertex cannot affect v¯\overline{\ell_{v}} at all.

If v¯s\overline{\ell_{v}}\geq s, there are at most ss groups of at most k1k-1 neighbors of vv, so that each vertex in each group received the same color, and each group corresponds to a different color from LvL_{v}. Thus, the color assignments and activation choices of these vertices certify that v¯s\overline{\ell_{v}}\geq s.

Since, according to Corollary 4.2, Keepi=Ω(1)\mathrm{Keep}_{i}=\Omega(1), applying Theorem 2.4 with t=Li1.93t=L_{i}^{\frac{1.9}{3}}, w=2kw=2k, c=1c=1, we obtain

Pr[|v¯𝔼[v¯]|>Li2/3]<ΔlnΔ,\displaystyle\Pr\left[|\overline{\ell_{v}}-{\mathbb{E}}[\overline{\ell_{v}}]|>L_{i}^{2/3}\right]<\Delta^{-\ln\Delta},

for sufficiently large Δ\Delta.

Finally, 𝔼[i+1(v)]=i(v)𝔼[v¯]{\mathbb{E}}[\ell_{i+1}(v)]=\ell_{i}(v)-{\mathbb{E}}[\overline{\ell_{v}}] implies that

Pr[|i+1(v)𝔼[i+1(v)]|>Li2/3]=Pr[|v¯𝔼[v¯]|>Li2/3]<ΔlnΔ.\displaystyle\Pr\left[|\ell_{i+1}(v)-{\mathbb{E}}[\ell_{i+1}(v)]|>L_{i}^{2/3}\right]=\Pr\left[|\overline{\ell_{v}}-{\mathbb{E}}[\overline{\ell_{v}}]|>L_{i}^{2/3}\right]<\Delta^{-\ln\Delta}.

Proof of Part (b).

Recall the definition of Di,r(v,c)D_{i,r}(v,c) and let Zi,r(v,c)=j=rk1Di,j(v,c)Z_{i,r}(v,c)=\bigcup_{j=r}^{k-1}D_{i,j}(v,c). Let Xi+1,r(v,c)X_{i+1,r}(v,c) denote the number of hyperedges in Zi,r(v,c)Z_{i,r}(v,c) which (i) contain exactly rr uncolored vertices other than vv; and (ii) the rest of their vertices are assigned cc in the end of the ii-th iteration. Let also Yi+1,r(v,c)Y_{i+1,r}(v,c) be the number of these hyperedges which they contain an uncolored vertex uvu\neq v such that (i) cLu(i+1)c\notin L_{u}(i+1); and (ii) cc would still not be in Lu(i+1)L_{u}(i+1) even if we ignored the color of vv.

B definition we have ti+1,r(v,c)=Xi+1,r(v,c)Yi+1,r(v,c)t_{i+1,r}^{\prime}(v,c)=X_{i+1,r}(v,c)-Y_{i+1,r}(v,c). Therefore, by the linearity of expectation, it suffices to show that Xi+1,r(v,c)X_{i+1,r}(v,c) and Yi+1,r(v,c)Y_{i+1,r}(v,c) are both sufficiently concentrated. This is because

Pr[ti+1,r(v,c)𝔼[ti+1(v,c)]>12(j=rk1(jr)αjrTi,jLijr)2/3],\displaystyle\Pr\left[t_{i+1,r}^{\prime}(v,c)-{\mathbb{E}}[t_{i+1}^{\prime}(v,c)]>\frac{1}{2}\left(\sum_{j=r}^{k-1}{j\choose r}\alpha^{j-r}\frac{T_{i,j}}{L_{i}^{j-r}}\right)^{2/3}\right],
=\displaystyle= Pr[Xi+1,r(v,c)𝔼[Xi+1(v,c)](Yi+1,r(v,c)𝔼[Yi+1,r(v,c)])>12(j=rk1(jr)αjrTi,jLijr)2/3],\displaystyle\Pr\left[X_{i+1,r}(v,c)-{\mathbb{E}}[X_{i+1}(v,c)]-\left(Y_{i+1,r}(v,c)-{\mathbb{E}}[Y_{i+1,r}(v,c)]\right)>\frac{1}{2}\left(\sum_{j=r}^{k-1}{j\choose r}\alpha^{j-r}\frac{T_{i,j}}{L_{i}^{j-r}}\right)^{2/3}\right],

and, therefore, it is sufficient to prove that

Pr[Xi+1,r(v,c)𝔼[Xi+1,r(v,c)]>14(j=rk1(jr)αjrTi,jLijr)2/3]\displaystyle\Pr\left[X_{i+1,r}(v,c)-{\mathbb{E}}[X_{i+1,r}(v,c)]>\frac{1}{4}\left(\sum_{j=r}^{k-1}{j\choose r}\alpha^{j-r}\frac{T_{i,j}}{L_{i}^{j-r}}\right)^{2/3}\right] \displaystyle\leq 12ΔlnΔ,\displaystyle\frac{1}{2}\Delta^{-\ln\Delta}, (37)
Pr[Yi+1,r(v,c)𝔼[Yi+1,r(v,c)]<14(j=rk1(jr)αjrTi,jLijr)2/3]\displaystyle\Pr\left[Y_{i+1,r}(v,c)-{\mathbb{E}}[Y_{i+1,r}(v,c)]<-\frac{1}{4}\left(\sum_{j=r}^{k-1}{j\choose r}\alpha^{j-r}\frac{T_{i,j}}{L_{i}^{j-r}}\right)^{2/3}\right] \displaystyle\leq 12ΔlnΔ.\displaystyle\frac{1}{2}\Delta^{-\ln\Delta}. (38)

We first focus on Xi+1,r(v,c)X_{i+1,r}(v,c). Let Xi+1,r(v,c)X_{i+1,r}^{\prime}(v,c) denote the number of hyperedges in Zi,r(v,c)Z_{i,r}(v,c) which (i) contain exactly rr uncolored vertices other than vv; and (ii) the rest of their vertices were activated and assigned cc (but did not retain their color necessarily). Clearly, Xi+1,r(v,c)Xi+1,r(v,c)X_{i+1,r}^{\prime}(v,c)\geq X_{i+1,r}(v,c). Further, let Wi+1,r1(v,c)W_{i+1,r}^{1}(v,c) denote the random variable that counts all the hyperedges counted by Xi+1,r(v,c)X_{i+1,r}^{\prime}(v,c), except for those whose rr uncolored vertices (other than vv) were uncolored because they were activated and received the same color as vv. Finally, let Wi+1,r2(v,c)W_{i+1,r}^{2}(v,c) be the number of hyperedges which (i) contain exactly rr vertices that are activated and received the same color as vv; and (ii) the rest of their k1rk-1-r vertices were activated and assigned cc.

Observe that Xi+1,r(v,c)Wi+1,r1(v,c)+Wi+1,r2(v,c)X_{i+1,r}(v,c)\leq W_{i+1,r}^{1}(v,c)+W_{i+1,r}^{2}(v,c). The idea here is that we cannot directly apply Talagrand’s inequality to Xi+1,r(v,c)X_{i+1,r}(v,c) and so we consider Wi+1,r1(v,c),Wi+1,r2(v,c)W_{i+1,r}^{1}(v,c),W_{i+1,r}^{2}(v,c) instead.

First, we consider Wi+1,r1(v,c)W_{i+1,r}^{1}(v,c). Since our hypergraph is of girth at least 55, changing a choice for some vertex of a hyperedge hZi,r(v,c)h\in Z_{i,r}(v,c) can only affect whether or not the vertices of hh remain uncolored, and thus affect Wi+1,r1(v,c)W_{i+1,r}^{1}(v,c) by at most 11. Furthermore, changing a choice for a vertex outside the ones that correspond to the hyperedges in Zi,r(v,c)Z_{i,r}(v,c) can affect at most one vertex of at most one hyperedge in Zi,r(v,c)Z_{i,r}(v,c) and, therefore, can affect Wi+1,r1(v,c)W_{i+1,r}^{1}(v,c) by at most 11.

We claim now that if Wi+1,r1(v,c)sW_{i+1,r}^{1}(v,c)\geq s, then there exist at most 2k2s2k^{2}s random choices that certify this event. To see this, notice that if a hyperedge hh is counted by Wi+1,r1(v,c)W_{i+1,r}^{1}(v,c), then for every uh{v}u\in h\setminus\{v\} that remained uncolored, it must be that either uu was deactivated, or uu is contained in a hyperedge hhh^{\prime}\neq h such that all the vertices in (h{u})Vi(h^{\prime}\setminus\{u\})\cap V_{i} were activated and received the same color as uu. Moreover, the event that a variable uh{v}u\in h\setminus\{v\} was activated and received cc can be verified by the outcome of two random choices. So, overall, we can certify that hh was counted by Wi+1,r1(v,c)W_{i+1,r}^{1}(v,c) by using the outcome of at most 2k22k^{2} random choices.

Finally, observe that 𝔼[Wi+1,r1(v,c)]j=rk1(jr)αjrTi,jLijr{\mathbb{E}}[W_{i+1,r}^{1}(v,c)]\leq\sum_{j=r}^{k-1}{j\choose r}\alpha^{j-r}\frac{T_{i,j}}{L_{i}^{j-r}} and, thus, applying Theorem 2.4 with c=1c=1, w=2k2w=2k^{2} and t=(j=rk1(jr)Ti,rLijr)1.9/3t=\left(\sum_{j=r}^{k-1}{j\choose r}\frac{T_{i,r}}{L_{i}^{j-r}}\right)^{1.9/3} we obtain

Pr[|Wi+1,r1(v,c)𝔼[Wi+1,r1(v,c)]|>18(j=rk1(jr)αjrTi,jLijr)2/3]14ΔlnΔ,\displaystyle\Pr\left[\left|W_{i+1,r}^{1}(v,c)-{\mathbb{E}}[W_{i+1,r}^{1}(v,c)]\right|>\frac{1}{8}\left(\sum_{j=r}^{k-1}{j\choose r}\alpha^{j-r}\frac{T_{i,j}}{L_{i}^{j-r}}\right)^{2/3}\right]\leq\frac{1}{4}\Delta^{-\ln\Delta}, (39)

for sufficiently large Δ\Delta, since we have assumed that Ti,r(lnΔ)20(k1)T_{i,r}\geq(\ln\Delta)^{20(k-1)}.

As far as Wi+1,r2(v,c)W_{i+1,r}^{2}(v,c) is concerned, note that it is stochastically dominated by j=rk1Bin(Ti,j,(jr)αjLij)\sum_{j=r}^{k-1}\mathrm{Bin}(T_{i,j},{j\choose r}\frac{\alpha^{j}}{L_{i}^{j}}) and recall Lemma 4.1. Since Ti,j(lnΔ)20(k1)T_{i,j}\geq(\ln\Delta)^{20(k-1)} for every j[k1]j\in[k-1], Lemma 4.5 implies that

Pr[Wi+1,r2(v,c)>𝔼[Wi+1,r2(v,c)]+18(j=rk1(jr)αjrTi,jLijr)2/3]14ΔlnΔ.\displaystyle\Pr\left[W_{i+1,r}^{2}(v,c)>{\mathbb{E}}[W_{i+1,r}^{2}(v,c)]+\frac{1}{8}\left(\sum_{j=r}^{k-1}{j\choose r}\alpha^{j-r}\frac{T_{i,j}}{L_{i}^{j-r}}\right)^{2/3}\right]\leq\frac{1}{4}\Delta^{-\ln\Delta}. (40)

Combining (39) and (40) implies (37).

We follow the same approach for Yi+1,r(v,c)Y_{i+1,r}(v,c). Let Yi+1,r(v,c)Y_{i+1,r}^{\prime}(v,c) be the number of hyperedges counted by Xi+1,r(v,c)X_{i+1,r}^{\prime}(v,c) and which also contain an uncolored vertex uvu\neq v such that (i) cLu(i+1)c\notin L_{u}(i+1); and (ii) cc would still not be in Lu(i+1)L_{u}(i+1) even if we ignored the color of vv. Further, let Yi+1,r′′(v,c)Y_{i+1,r}^{{}^{\prime\prime}}(v,c) be the random variable that counts all the hyperedges counted by Yi+1,r(v,c)Y_{i+1,r}^{\prime}(v,c), except for those whose rr uncolored vertices were uncolored because they were activated and received the same color as vv, and observe that Yi+1,r(v,c)Yi+1,r(v,c)Yi+1,r′′(v,c)+Wi+1,r2(v,c)Y_{i+1,r}(v,c)\leq Y_{i+1,r}^{\prime}(v,c)\leq Y_{i+1,r}^{\prime\prime}(v,c)+W_{i+1,r}^{2}(v,c).

Moreover, if Yi+1,r′′(v,c)sY_{i+1,r}^{\prime\prime}(v,c)\geq s, then there exist at most (2k2+2k)s(2k^{2}+2k)s random choices that certify this event. To see this, observe that for each hyperedge hh counted by Yi+1,r(v,c)Y_{i+1,r}(v,c), we need the output of at most 2k22k^{2} choices to certify that it is counted by Xi+1,r(v,c)X_{i+1,r}^{\prime}(v,c), and the output of at most 2k2k extra random choices to certify that there is a vertex uh{v}u\in h\setminus\{v\} for which cLu(i+1)c\notin L_{u}(i+1), and cc would still not be in Lu(i+1)L_{u}(i+1) even if we ignored the color of vv.

Finally, 𝔼[Yi+1,r′′(v,c)]j=rk1(jr)αjrTi,jLijr{\mathbb{E}}[Y_{i+1,r}^{{}^{\prime\prime}}(v,c)]\leq\sum_{j=r}^{k-1}{j\choose r}\alpha^{j-r}\frac{T_{i,j}}{L_{i}^{j-r}} and, therefore, an almost identical argument to the case for Xi+1,r(v,c)X_{i+1,r}(v,c) implies (38).

4.3 Proof of Lemma 3.3

We use induction on ii. Property P(1)P(1) clearly holds, so we assume that property P(i)P(i) holds and we prove that with property P(i+1)P(i+1) holds with positive probability. Recall our discussion in the previous section in which we argued that we can assume without loss of generality that property Q(i)Q(i) holds.

For every vv and cLvc\in L_{v} let AvA_{v} be the event that i+1(v)<Li+1\ell_{i+1}(v)<L_{i+1} and Bv,crB_{v,c}^{r} to be the event that ti+1,r(v,c)>Ti+1,rt_{i+1,r}(v,c)>T_{i+1,r}. Clearly, if these bad events are avoided, then P(i+1)P(i+1) holds.

Since property Q(i)Q(i) holds, we have that i(v)=Li\ell_{i}(v)=L_{i}. Therefore, by (4) and Lemmas 3.13.2 we have:

Pr[Av]=Pr[i+1(v)<LiKeepLi2/3]=Pr[i+1(v)<𝔼[i+1(v)]Li2/3]<ΔlnΔ.\displaystyle\Pr[A_{v}]=\Pr\left[\ell_{i+1}(v)<L_{i}\cdot\mathrm{Keep}-L_{i}^{2/3}\right]=\Pr\left[\ell_{i+1}(v)<{\mathbb{E}}[\ell_{i+1}(v)]-L_{i}^{2/3}\right]<\Delta^{-\ln\Delta}. (41)

Similarly, by (5) and Lemmas 3.13.2 we have:

Pr[Bv,cr]\displaystyle\Pr[B_{v,c}^{r}] Pr[ti+1,r(v,c)>Ti+1,r]\displaystyle\leq\Pr[t_{i+1,r}^{\prime}(v,c)>T_{i+1,r}]
=Pr[ti+1,r(v,c)𝔼[ti+1,r(v,c)]>(j=rk1(jr)αjrTi,jLijr)2/3Cj=rk1Ti,jLij]\displaystyle=\Pr\left[t_{i+1,r}^{\prime}(v,c)-{\mathbb{E}}[t_{i+1,r}^{\prime}(v,c)]>\left(\sum_{j=r}^{k-1}{j\choose r}\alpha^{j-r}\frac{T_{i,j}}{L_{i}^{j-r}}\right)^{2/3}-C^{\prime}\cdot\sum_{j=r}^{k-1}\frac{T_{i,j}}{L_{i}^{j}}\right]
Pr[ti+1,r(v,c)𝔼[ti+1,r(v,c)]>12(j=rk1(jr)αjrTi,jLijr)2/3]\displaystyle\leq\Pr\left[t_{i+1,r}^{\prime}(v,c)-{\mathbb{E}}[t_{i+1,r}^{\prime}(v,c)]>\frac{1}{2}\left(\sum_{j=r}^{k-1}{j\choose r}\alpha^{j-r}\frac{T_{i,j}}{L_{i}^{j-r}}\right)^{2/3}\right] (42)
ΔlnΔ,\displaystyle\leq\Delta^{-\ln\Delta}, (43)

where C>0C^{\prime}>0 is the hidden constant that multiplies YiY_{i} in the statement of Lemma 3.1, Part (b). Note that in deriving (42) we used Lemma 4.1 — which implies that j=rk1Ti,jLij=O(lnΔ)\sum_{j=r}^{k-1}\frac{T_{i,j}}{L_{i}^{j}}=O(\ln\Delta) — and our assumptions that Ti,1(lnΔ)20(k1)T_{i,1}\geq(\ln\Delta)^{20(k-1)} — which implies that (j=rk1(jr)αjrTi,jLijr)2/3=o(lnΔ)\left(\sum_{j=r}^{k-1}{j\choose r}\alpha^{j-r}\frac{T_{i,j}}{L_{i}^{j-r}}\right)^{2/3}=o(\ln\Delta).

Notice now that each bad event fv{Av,Bv,cr}f_{v}\in\{A_{v},B_{v,c}^{r}\} event is determined by the colors assigned to vertices of distance at most 33 from vv. Therefore, fvf_{v} is mutually independent of all but at most (kΔ)4(1+δ)(ΔlnΔ)1k1<Δ5(k\Delta)^{4}(1+\delta)\left(\frac{\Delta}{\ln\Delta}\right)^{\frac{1}{k-1}}<\Delta^{5} other bad events. For Δ\Delta sufficiently large, ΔlnΔΔ5<14\Delta^{-\ln\Delta}\Delta^{5}<\frac{1}{4} and so the proof is concluded by applying Corollary 2.3 using (41), (43).

4.4 Proof of Lemma 3.4

Since Li<LiL_{i}<L_{i}^{\prime}, for the first part of the lemma it suffices to prove that LiLi+(Li)5/6L_{i}^{\prime}\leq L_{i}+(L_{i}^{\prime})^{5/6}. Towards that end, at first we observe that for sufficiently large Δ\Delta, Corollary 4.2 and the fact that K=1100k3kK=\frac{1}{100k^{3k}} imply:

Keepi5/6Keepi\displaystyle\mathrm{Keep}_{i}^{5/6}-\mathrm{Keep}_{i} \displaystyle\geq (1Kk112k2(lnΔ)k1)5/6(1Kk112k2(lnΔ)k1)\displaystyle\left(1-\frac{K^{k-1}}{12k^{2}(\ln\Delta)^{k-1}}\right)^{5/6}-\left(1-\frac{K^{k-1}}{12k^{2}(\ln\Delta)^{k-1}}\right) (44)
\displaystyle\geq (156Kk112k2(lnΔ)k1)(1Kk112k2(lnΔ)k1)=Kk172k2(lnΔ)k1.\displaystyle\left(1-\frac{5}{6}\cdot\frac{K^{k-1}}{12k^{2}(\ln\Delta)^{k-1}}\right)-\left(1-\frac{K^{k-1}}{12k^{2}(\ln\Delta)^{k-1}}\right)=\frac{K^{k-1}}{72k^{2}(\ln\Delta)^{k-1}}.

Note that in deriving the first inequality we used the fact that the function x5/6xx^{5/6}-x is decreasing on the interval [C,1][C,1] since KK is sufficiently small. For the second one, we used the Taylor Series for (1y)5/6(1-y)^{5/6} around y=0y=0.

We now proceed by using induction. The base case is trivial, so assume that the statement is true for ii, and consider i+1i+1. Since by our assumption Li(lnΔ)20(k1)L_{i}\geq(\ln\Delta)^{20(k-1)} we obtain

Li+1\displaystyle L_{i+1}^{\prime} =\displaystyle= KeepiLi\displaystyle\mathrm{Keep}_{i}L_{i}^{\prime}
\displaystyle\leq Keepi(Li+(Li)5/6)\displaystyle\mathrm{Keep}_{i}\left(L_{i}+(L_{i}^{\prime})^{5/6}\right)
=\displaystyle= Li+1+Li2/3+Keepi(Li)5/6\displaystyle L_{i+1}+L_{i}^{2/3}+\mathrm{Keep}_{i}(L_{i}^{\prime})^{5/6}
\displaystyle\leq Li+1+Li2/3+Keepi5/6(Li)5/6Kk172k2(lnΔ)k1(Li)5/6\displaystyle L_{i+1}+L_{i}^{2/3}+\mathrm{Keep}_{i}^{5/6}(L_{i}^{\prime})^{5/6}-\frac{K^{k-1}}{72k^{2}(\ln\Delta)^{k-1}}(L_{i}^{\prime})^{5/6}
\displaystyle\leq Li+1+(Li+1)5/6+Li2/3Kk172k2(lnΔ)k1(Li)5/6\displaystyle L_{i+1}+(L_{i+1}^{\prime})^{5/6}+L_{i}^{2/3}-\frac{K^{k-1}}{72k^{2}(\ln\Delta)^{k-1}}(L_{i}^{\prime})^{5/6}
<\displaystyle< Li+1+(Li+1)5/6,\displaystyle L_{i+1}+(L_{i+1}^{\prime})^{5/6}, (47)

for sufficiently large Δ\Delta. Note that in deriving (4.4) we used the inductive hypothesis; for (4.4) we used (44); and for (47) the fact that Li(lnΔ)20(k1)L_{i}\geq(\ln\Delta)^{20(k-1)} and the inductive hypothesis.

The proof of the second part of the lemma is conceptually similar, but more technical, so we present its proof in Appendix A.

4.5 Proof of Lemma 3.5

We proceed by induction. Let η:=ϵ/3(k1)(1+ϵ/2)\eta:=\frac{\epsilon/3}{(k-1)(1+\epsilon/2)}. We will assume that LjΔη,Tj,r(lnΔ)20(k1)L_{j}\geq\Delta^{\eta},T_{j,r}\geq(\ln\Delta)^{20(k-1)} for all 2ji<i2\leq j\leq i<i^{*}, and prove that Li+1Δη,Ti+1,r(lnΔ)20(k1)L_{i+1}\geq\Delta^{\eta},T_{i+1,r}\geq(\ln\Delta)^{20(k-1)}. Towards that end, it will be useful to focus on the family of ratios Ri,rR_{i,r}, r[k1r\in[k-1]. Note that, according to Lemma 3.4, this family is well-approximated by the family Ri,rR_{i,r}^{\prime}, r[k1]r\in[k-1]. In particular, recalling Lemma 4.3 and applying Lemma 3.4 we obtain:

Ri,r\displaystyle R_{i,r} \displaystyle\leq Ri,r1+(Ti,r)1100r+1(1(Li)1/6)r\displaystyle R_{i,r}^{\prime}\cdot\frac{1+(T_{i,r}^{\prime})^{-\frac{1}{100r+1}}}{\left(1-(L_{i}^{\prime})^{-1/6}\right)^{r}} (48)
\displaystyle\leq (1αC)r(i1)lnΔp=rk2(p+1)(1+δ1.1δk99)k1Ck1r,\displaystyle(1-\alpha C)^{r(i-1)}\ln\Delta\cdot\frac{\prod_{p=r}^{k-2}(p+1)}{(1+\delta-\frac{1.1\delta}{k^{99}})^{k-1}C^{k-1-r}},

for sufficiently large Δ\Delta, since Li,Ti,r(lnΔ)20(k1)L_{i},T_{i,r}\geq(\ln\Delta)^{20(k-1)}.

Using (48) and the fact that 11x>e1x11-\frac{1}{x}>\mathrm{e}^{-\frac{1}{x-1}} for x2x\geq 2 we can get an improved lower bound for Keepi\mathrm{Keep}_{i} as follows.

Keepi\displaystyle\mathrm{Keep}_{i} \displaystyle\geq exp(1(1δk100k)r=1k1αrRi,r)\displaystyle\mathrm{exp}\left(-\frac{1}{(1-\frac{\delta}{k^{100k}})}\sum_{r=1}^{k-1}\alpha^{r}R_{i,r}\right) (49)
\displaystyle\geq exp(1(1+δ1.2δk99)k1r=1k1(1αC)r(i1)Krp=rk2(p+1)(lnΔ)r1Ck1r),\displaystyle\mathrm{exp}\left(-\frac{1}{\left(1+\delta-\frac{1.2\delta}{k^{99}}\right)^{k-1}}\sum_{r=1}^{k-1}(1-\alpha C)^{r(i-1)}\frac{K^{r}\prod_{p=r}^{k-2}(p+1)}{(\ln\Delta)^{r-1}C^{k-1-r}}\right),

for sufficiently large Δ\Delta.

Recall that δ=(1+ϵ)(k1)1\delta=(1+\epsilon)(k-1)-1. Using (49) we get

j=1i1Keepj\displaystyle\prod_{j=1}^{i-1}\mathrm{Keep}_{j} \displaystyle\geq exp(1(1+δ1.2δk99)k1r=1k1(Krp=rk2(p+1)(lnΔ)r1Ck1rj=1i1(1αC)r(j1)))\displaystyle\mathrm{exp}\left(-\frac{1}{\left(1+\delta-\frac{1.2\delta}{k^{99}}\right)^{k-1}}\sum_{r=1}^{k-1}\left(\frac{K^{r}\prod_{p=r}^{k-2}(p+1)}{(\ln\Delta)^{r-1}C^{k-1-r}}\sum_{j=1}^{i-1}(1-\alpha C)^{r(j-1)}\right)\right) (50)
\displaystyle\geq exp(1(1+δ1.2δk99)k1r=1k1(Krp=rk2(p+1)(lnΔ)r1Ck1r11(1αC)r))\displaystyle\mathrm{exp}\left(-\frac{1}{\left(1+\delta-\frac{1.2\delta}{k^{99}}\right)^{k-1}}\sum_{r=1}^{k-1}\left(\frac{K^{r}\prod_{p=r}^{k-2}(p+1)}{(\ln\Delta)^{r-1}C^{k-1-r}}\cdot\frac{1}{1-(1-\alpha C)^{r}}\right)\right)
\displaystyle\geq exp(C(k2)(1+δk100)(k1)!lnΔ(1+δ1.2δk99)k1)\displaystyle\mathrm{exp}\left(-\frac{C^{-(k-2)}(1+\frac{\delta}{k^{100}})(k-1)!\ln\Delta}{\left(1+\delta-\frac{1.2\delta}{k^{99}}\right)^{k-1}}\right)
\displaystyle\geq exp(lnΔ(1+ϵ2)(k1))\displaystyle\mathrm{exp}\left(-\frac{\ln\Delta}{(1+\frac{\epsilon}{2})(k-1)}\right)

for sufficiently large Δ\Delta.

Using (50) we can now bound LiL_{i}^{\prime} as follows.

Li=L1j=1i1Keepj(1+δ)(ΔlnΔ)1k1Δ1(1+ϵ2)(k1)Δη,\displaystyle L_{i}^{\prime}=L_{1}^{\prime}\prod_{j=1}^{i-1}\mathrm{Keep}_{j}\geq(1+\delta)\left(\frac{\Delta}{\ln\Delta}\right)^{\frac{1}{k-1}}\Delta^{-\frac{1}{(1+\frac{\epsilon}{2})(k-1)}}\geq\Delta^{\eta}, (51)

for sufficiently large Δ\Delta. Thus, LiL_{i}^{\prime} never gets too small for the purposes of our analysis. Lemma 3.4 implies that neither does LiL_{i}.

The proof is concluded by observing that (48) implies that Ri,rR_{i,r}, r[k1]r\in[k-1], becomes smaller than 110k2\frac{1}{10k^{2}} for i=O(lnΔlnlnΔ)i=O(\ln\Delta\ln\ln\Delta).

5 A sufficient pseudo-random property for coloring

In this section we present the proof of Theorem 1.5. To do so, we build on ideas of Alon, Krivelevich and Sudakov [6] and show that the random hypergraph H(k,n,d/(nk1))H(k,n,d/{n\choose k-1}) asymptotically almost surely admits a few useful features.

The first lemma we prove states that all subgraphs of H(k,n,d/(nk1))H(k,n,d/{n\choose k-1}) with not too many vertices are sparse and, therefore, of small degeneracy.

Lemma 5.1.

For every constant k2k\geq 2, there exists dk>0d_{k}>0 such that for any constant ddkd\geq d_{k}, the random hypergraph H(k,n,d/(nk1))H(k,n,d/{n\choose k-1}) has the following property asymptotically almost surely: Every snd1k1s\leq nd^{-\frac{1}{k-1}} vertices of HH span fewer than s(d(lnd)2)1k1s\left(\frac{d}{(\ln d)^{2}}\right)^{\frac{1}{k-1}} hyperedges. Therefore, any subhypergraph of HH induced by a subset V0VV_{0}\subset V of size |V0|nd1k1|V_{0}|\leq nd^{-\frac{1}{k-1}}, is k(d(lnd)2)1k1k\left(\frac{d}{(\ln d)^{2}}\right)^{\frac{1}{k-1}}-degenerate.

Proof.

Given the statement about the sparsity of any subhypergraph of HH incudced by a set of snd1k1s\leq nd^{-\frac{1}{k-1}} vertices, the claim about its degeneracy follows from the fact that its average (and, therefore its minimum) degree is at most

kss(d(lnd)2)1k1=(d(lnd)2)1k1.\displaystyle k\cdot\frac{s}{s\left(\frac{d}{(\ln d)^{2}}\right)^{\frac{1}{k-1}}}=\left(\frac{d}{(\ln d)^{2}}\right)^{\frac{1}{k-1}}.

So we are left with proving the statement regarding sparsity.

Letting r=(d(lnd)2)1k1r=\left(\frac{d}{(\ln d)^{2}}\right)^{\frac{1}{k-1}}, we see that the probability that there exists a subset V0VV_{0}\subset V which violates the statement of the lemma is at most

i=r1k1nd1k1(ni)((ik)ri)(d(nk1))ri\displaystyle\sum_{i=r^{\frac{1}{k-1}}}^{nd^{-\frac{1}{k-1}}}{n\choose i}{{i\choose k}\choose ri}\left(\frac{d}{{n\choose k-1}}\right)^{ri} \displaystyle\leq i=r1k1nd1k1[eni(eik1r)r(d(nk1))r]i\displaystyle\sum_{i=r^{\frac{1}{k-1}}}^{nd^{-\frac{1}{k-1}}}\left[\frac{\mathrm{e}n}{i}\left(\frac{\mathrm{e}i^{k-1}}{r}\right)^{r}\left(\frac{d}{{n\choose k-1}}\right)^{r}\right]^{i} (52)
\displaystyle\leq i=r1k1nd1k1[e1+1k1(k1)(dr)1k1(eik1dr(nk1))r1k1]i\displaystyle\sum_{i=r^{\frac{1}{k-1}}}^{nd^{-\frac{1}{k-1}}}\left[\mathrm{e}^{1+\frac{1}{k-1}}(k-1)\left(\frac{d}{r}\right)^{\frac{1}{k-1}}\left(\frac{\mathrm{e}i^{k-1}d}{r{n\choose k-1}}\right)^{r-\frac{1}{k-1}}\right]^{i} (53)
=\displaystyle= o(1),\displaystyle o(1), (54)

for sufficiently large dd. Note that in the lefthand side of (52) we used the fact that any subset of vertices of size s<r1k1s<r^{\frac{1}{k-1}} cannot violate the assertion of the lemma, since it can span at most (sk)sk=sk1s<rs{s\choose k}\leq s^{k}=s^{k-1}\cdot s<rs hyperedges. In deriving (54) we used the fact that, for sufficiently large d,nd,n, every summand in (53) is at most n(k1)(d/(lnd)2)1/(k1)/2n^{-(k-1)(d/(\ln d)^{2})^{1/(k-1)}/2}, and there exist at most nn summands. Throughout our derivation we used that for any pair of positive integers α,β\alpha,\beta, we have (αβ)β(αβ)<(αeβ)β\left(\frac{\alpha}{\beta}\right)^{\beta}\leq{\alpha\choose\beta}<\left(\frac{\alpha\cdot\mathrm{e}}{\beta}\right)^{\beta}.

Next we show that, that for any constant cc, the number of vertices of H(k,n,d/(nk1))H(k,n,d/{n\choose k-1}) that have degree cc essentially behaves as a Poisson random variable with mean dd.

Lemma 5.2.

For constants c1c\geq 1, k2k\geq 2 and dd sufficiently large, let XcX_{c} denote the number of vertices of degree cc in H(k,n,d/(nk1))H(k,n,d/{n\choose k-1}). Then, asymptotically almost surely,

Xcdcedc!n(1+O(lognn)).\displaystyle X_{c}\leq\frac{d^{c}\mathrm{e}^{-d}}{c!}n\left(1+O\left(\frac{\log n}{\sqrt{n}}\right)\right).
Proof.

The lemma follows from standard ideas for estimation of the degree distribution of random graphs (see for example the proof of Theorem 3.3 in [12] for the case k=2k=2). In particular, assume that the vertices of H(k,n,d/(nk1))H(k,n,d/{n\choose k-1}) are labeled 1,2,,n1,2,\ldots,n. Then,

𝔼[Xc]\displaystyle{\mathbb{E}}[X_{c}] =\displaystyle= nPr[deg(1)=c]\displaystyle n\Pr[\mathrm{deg}(1)=c]
=\displaystyle= n((n1k1)c)(d(nk1))c(1d(nk1))(n1k1)c\displaystyle n{{{n-1\choose k-1}\choose c}}\left(\frac{d}{{n\choose k-1}}\right)^{c}\left(1-\frac{d}{{n\choose k-1}}\right)^{{n-1\choose k-1}-c}
\displaystyle\leq n((n1k1))cc!(d(nk1))cexp(((n1k1)c)d(nk1))\displaystyle n\frac{\left({n-1\choose k-1}\right)^{c}}{c!}\left(\frac{d}{{n\choose k-1}}\right)^{c}\mathrm{exp}\left(-\left({n-1\choose k-1}-c\right)\frac{d}{{n\choose k-1}}\right)
\displaystyle\leq ndcedc!(1+O(1nk1)).\displaystyle n\frac{d^{c}\mathrm{e}^{-d}}{c!}\left(1+O\left(\frac{1}{{n^{k-1}}}\right)\right). (56)

Note that in the first inequality above we used the fact that for every any pair of positive integers α,β\alpha,\beta, we have (αβ)<αββ!{\alpha\choose\beta}<\frac{\alpha^{\beta}}{\beta!}.

To show concentration of XcX_{c} around its expectation, we will use Chebyshev’s inequality. In order to do so, we need to estimate the second moment of XcX_{c} and, in particular, Pr[deg(1)=deg(2)=c]\Pr[\mathrm{deg}(1)=\mathrm{deg}(2)=c]. Letting p=d(nk1)p=\frac{d}{{n\choose k-1}} and Cc,dC_{c,d} be a sufficiently large constant, we see that Pr[deg(1)=deg(2)=c]\Pr[\mathrm{deg}(1)=\mathrm{deg}(2)=c] equals

=0c((n2k2))p(1p)(n2k2)(((n1k1)(n2k2)c)pc(1p)(n1k1)(n2k2)(c))2\displaystyle\sum_{\ell=0}^{c}{{{n-2\choose k-2}\choose\ell}}p^{\ell}\left(1-p\right)^{{n-2\choose k-2}-\ell}\left({{n-1\choose k-1}-{n-2\choose k-2}\choose c-{\ell}}p^{c-\ell}(1-p)^{{n-1\choose k-1}-{n-2\choose k-2}-(c-\ell)}\right)^{2}
\displaystyle\leq =0c((n2k2))p(1p)(n2k2)(Cc,d((n1k1)c)pc(1p)(n1k1)(n2k2)(c))2\displaystyle\sum_{\ell=0}^{c}{{{n-2\choose k-2}\choose\ell}}p^{\ell}\left(1-p\right)^{{n-2\choose k-2}-\ell}\left(C_{c,d}{{n-1\choose k-1}\choose c}p^{c}(1-p)^{{n-1\choose k-1}-{n-2\choose k-2}-(c-\ell)}\right)^{2} (57)
=\displaystyle= (((n1k1)c)pc(1p)(n1k1)c)2=0cCc,d2((n2k2))p(1p)(n2k2)+\displaystyle\left({{n-1\choose k-1}\choose c}p^{c}(1-p)^{{n-1\choose k-1}-c}\right)^{2}\sum_{\ell=0}^{c}C_{c,d}^{2}{{{n-2\choose k-2}\choose\ell}}p^{\ell}\left(1-p\right)^{-{n-2\choose k-2}+\ell}
=\displaystyle= Pr[deg(1)=c]Pr[deg(2)=c](1+O(1n)).\displaystyle\Pr[\mathrm{deg}(1)=c]\cdot\Pr[\mathrm{deg}(2)=c]\left(1+O\left(\frac{1}{n}\right)\right). (58)

Note that in deriving (57) we used that for any {1,,c1}:\ell\in\{1,\ldots,c-1\}:

((n1k1)(n2k2)c)p\displaystyle{{{n-1\choose k-1}-{n-2\choose k-2}}\choose c-\ell}\cdot p^{-\ell} \displaystyle\leq ((n1k1)(n2k2)c)((nk1)d)\displaystyle{{{n-1\choose k-1}-{n-2\choose k-2}}\choose c-\ell}\cdot\left(\frac{{n\choose k-1}}{d}\right)^{\ell}
\displaystyle\leq ((n1k1))c(c)cec((nk1)d)\displaystyle\frac{\left({n-1\choose k-1}\right)^{c-\ell}}{(c-\ell)^{c-\ell}}\mathrm{e}^{c-\ell}\cdot\left(\frac{{n\choose k-1}}{d}\right)^{\ell}
=\displaystyle= ((n1k1))c(c)cdecCc,d((n1k1)c),\displaystyle\frac{\left({n-1\choose k-1}\right)^{c}}{(c-\ell)^{c-\ell}d^{\ell}}\mathrm{e}^{c-\ell}\leq C_{c,d}\cdot{{n-1\choose k-1}\choose c},

for large enough Cc,dC_{c,d}. Moreover, in deriving (58) we use that for every {0,,c}\ell\in\{0,\ldots,c\}

Cc,d2((n2k2))p=O(n(k2)1n(k1))={O(1)if =0 ,O(1/n) otherwise,\displaystyle C_{c,d}^{2}{{{n-2\choose k-2}\choose\ell}}p^{\ell}=O\left(n^{\ell(k-2)}\cdot\frac{1}{n^{\ell(k-1)}}\right)=\begin{cases}O(1)&\text{if $\ell=0$ },\\ O(1/n)&\text{ otherwise},\end{cases}

and that:

(1p)(n2k2)+exp(d(n2k2)(n1k1))=1+O(1n).\displaystyle(1-p)^{-{n-2\choose k-2}+\ell}\leq\mathrm{exp}\left(\frac{d{n-2\choose k-2}}{{n-1\choose k-1}}\right)=1+O\left(\frac{1}{n}\right).

Therefore, letting IjI_{j} denote the indicator random variable which equals 11 if vertex jj has degree cc and 0 otherwise,

Var[Xc]\displaystyle\mathrm{Var}[X_{c}] =\displaystyle= 𝔼[Xc2](𝔼[Xc])2\displaystyle\mathbb{E}\left[X_{c}^{2}\right]-\left(\mathbb{E}[X_{c}]\right)^{2}
=\displaystyle= 𝔼[(j=1nIj)2](𝔼[Xc])2\displaystyle\mathbb{E}\left[\left(\sum_{j=1}^{n}I_{j}\right)^{2}\right]-\left(\mathbb{E}[X_{c}]\right)^{2}
=\displaystyle= 𝔼[i=1nj=1nIiIj](𝔼[Xc])2\displaystyle\mathbb{E}\left[\sum_{i=1}^{n}\sum_{j=1}^{n}I_{i}I_{j}\right]-\left(\mathbb{E}[X_{c}]\right)^{2}
=\displaystyle= i=1nj=1n(Pr[deg(i)=c,deg(j)=c]Pr[deg(1)=c]Pr[deg(2)=c])\displaystyle\sum_{i=1}^{n}\sum_{j=1}^{n}\left(\Pr[\mathrm{deg}(i)=c,\mathrm{deg}(j)=c]-\Pr[\mathrm{deg}(1)=c]\Pr[\mathrm{deg}(2)=c]\right)
\displaystyle\leq i=1nj[n]{i}(Pr[deg(i)=c,deg(j)=c]Pr[deg(1)=c]Pr[deg(2)=c])\displaystyle\sum_{i=1}^{n}\sum_{j\in[n]\setminus\{i\}}\left(\Pr[\mathrm{deg}(i)=c,\mathrm{deg}(j)=c]-\Pr[\mathrm{deg}(1)=c]\Pr[\mathrm{deg}(2)=c]\right)
+i=1nPr[deg(i)=c]\displaystyle+\sum_{i=1}^{n}\Pr[\mathrm{deg}(i)=c]
\displaystyle\leq An\displaystyle An

for some constant A=A(c,d)A=A(c,d). Note that in deriving (62) we used (58) and that i=1nPr[deg(i)=c]=𝔼[Xc]=O(n)\sum_{i=1}^{n}\Pr[\mathrm{deg}(i)=c]=\mathbb{E}[X_{c}]=O(n) according to (56).

Finally, applying the Chebyshev’s inequality, we obtain that, for any t>0t>0,

Pr[|Xc𝔼[Xc]|tn]At2,\displaystyle\Pr\left[|X_{c}-{\mathbb{E}}[X_{c}]|\geq t\sqrt{n}\right]\leq\frac{A}{t^{2}},

and, thus, the proof is concluded by choosing t=lognt=\log n.

Lemma 5.2 implies the following useful corollary.

Corollary 5.3.

For any constants δ(0,1),k2,d>0\delta\in(0,1),k\geq 2,d>0, let X=X(δ,k,d)X=X(\delta,k,d) denote the random variable equal to the number of vertices in H(k,n,d/(nk1))H(k,n,\ d/{n\choose k-1}) whose degree is in [(1+δ)d,3(k1)k1d][(1+\delta)d,3(k-1)^{k-1}d]. There exists a constant dδ>0d_{\delta}>0 such that if ddδd\geq d_{\delta} then, asymptotically almost surely, Xnd2.X\leq\frac{n}{d^{2}}.

Proof.

Let XrX_{r} denote the number of vertices of degree rr in H(k,n,d/(nk1))H(k,n,d/{n\choose k-1}). Since k,dk,d are constants, using Lemma 5.2 and Stirling’s approximation we see that, asymptotically almost surely,

r=(1+δ)d3(k1)k1dXrn(1+O(lognn))r=(1+δ)d3(k1)k1ddredr!n(1+δ)r=(1+δ)d3(k1)k1ddred2πr(re)rnd2,\displaystyle\sum_{r=(1+\delta)d}^{3(k-1)^{k-1}d}X_{r}\leq n\left(1+O\left(\frac{\log n}{\sqrt{n}}\right)\right)\sum_{r=(1+\delta)d}^{3(k-1)^{k-1}d}\frac{d^{r}\mathrm{e}^{-d}}{r!}\leq n(1+\delta)\sum_{r=(1+\delta)d}^{3(k-1)^{k-1}d}\frac{d^{r}\mathrm{e}^{-d}}{\sqrt{2\pi r}\left(\frac{r}{\mathrm{e}}\right)^{r}}\leq\frac{n}{d^{2}},

for sufficiently large dd and nn. ∎

Using Lemma 5.1 and Corollary 5.3 we show that, asymptotically almost surely, only a small fraction of vertices of H(k,n,d/(nk1))H(k,n,d/{n\choose k-1}) have degree that significantly exceeds its average degree.

Lemma 5.4.

For every constants k2k\geq 2 and δ(0,1)\delta\in(0,1), there exists dk,δ>0d_{k,\delta}>0 such that for any constant ddk,δd\geq d_{k,\delta}, all but at most 2nd2\frac{2n}{d^{2}} vertices of the random hypergraph H(k,n,d/(nk1))H(k,n,d/{n\choose k-1}) have degree at most (1+δ)d(1+\delta)d, asymptotically almost surely.

Proof.

Corollary 5.3 implies that the number of vertices with degree in the interval [(1+δ)d,3(k1)k1d][(1+\delta)d,3(k-1)^{k-1}d] is at most nd2\frac{n}{d^{2}}, for sufficiently large dd.

Suppose now there are more than nd2\frac{n}{d^{2}} vertices with degree at least 3(k1)k1d3(k-1)^{k-1}d. Denote by SS a set containing exactly nd2\frac{n}{d^{2}} such vertices. According to Lemma 5.1, asymptotically almost surely, the induced subhypergraph H[S]H[S] has at most

e(H[S])(d(lnd)2)1k1|S|=nd21k1(lnd)2k1\displaystyle e(H[S])\leq\left(\frac{d}{(\ln d)^{2}}\right)^{\frac{1}{k-1}}|S|=\frac{n}{d^{2-\frac{1}{k-1}}(\ln d)^{\frac{2}{k-1}}}

hyperedges. Therefore, the number of hyperedges between the sets of vertices SS and VSV\setminus S is at least

3(k1)k1d|S|ke(H[S])2.9(k1)k1nd=:N.\displaystyle 3(k-1)^{k-1}d|S|-ke(H[S])\geq\frac{2.9(k-1)^{k-1}n}{d}=:N.

for sufficiently large dd. However, the probability that H(k,n,d/(nk1))H(k,n,d/{n\choose k-1}) contains such a subhypergraph is at most

(nnd2)(nkd2N)(d(nk1))N(ed2)nd2(nked2Nd(nk1))N=o(1),\displaystyle{n\choose\frac{n}{d^{2}}}{\frac{n^{k}}{d^{2}}\choose N}\left(\frac{d}{{n\choose k-1}}\right)^{N}\leq\left(\mathrm{e}d^{2}\right)^{\frac{n}{d^{2}}}\left(\frac{n^{k}\mathrm{e}}{d^{2}N}\cdot\frac{d}{{n\choose k-1}}\right)^{N}=o(1),

for sufficiently large dd. Note that in deriving the final equality we used that for any pair of integers α,β\alpha,\beta, we have that (αβ)(αβ)β{\alpha\choose\beta}\geq\left(\frac{\alpha}{\beta}\right)^{\beta}. Therefore, asymptotically almost surely there are at most nd2\frac{n}{d^{2}} vertices in GG with degree greater than 3(k1)k1d3(k-1)^{k-1}d, concluding the proof. ∎

Finally, we show that the neighborhood of a typical vertex of H(k,n,d/(nk1))H(k,n,d/{n\choose k-1}) is locally tree-like.

Lemma 5.5.

For every constants k2,δ(0,1)k\geq 2,\delta\in(0,1), asymptotically almost surely, the random hypergraph H(k,n,d/(nk1))H(k,n,d/{n\choose k-1}) has a subset UV(H)U\subseteq V(H) of size at most n1δn^{1-\delta} such that the induced hypergraph H[VU]H[V\setminus U] is of girth at least 55.

Proof.

Let Y2,Y3,Y4Y_{2},Y_{3},Y_{4}, denote the number of 22-, 33- and 44-cycles in H(n,k,d/(nk1))H(n,k,d/{n\choose k-1}), respectively. A straightforward calculation reveals that for i{2,3,4}i\in\{2,3,4\}:

𝔼[Yi]s=1i(k1)(ns)((sk1)i)(d(nk1))i\displaystyle{\mathbb{E}}[Y_{i}]\leq\sum_{s=1}^{i(k-1)}{n\choose s}{{s\choose k-1}\choose i}\left(\frac{d}{{n\choose k-1}}\right)^{i} i(k1)(nei(k1))i(k1)((i(k1)k1)i)i(d(nk1))i\displaystyle\leq i(k-1)\left(\frac{n\cdot\mathrm{e}}{i(k-1)}\right)^{i(k-1)}\cdot\left(\frac{{i(k-1)\choose k-1}}{i}\right)^{i}\cdot\left(\frac{d}{{n\choose k-1}}\right)^{i}
i(k1)(nei(k1))i(k1)((ie)k1i)i(d(k1)k1nk1)i\displaystyle\leq i(k-1)\left(\frac{n\cdot\mathrm{e}}{i(k-1)}\right)^{i(k-1)}\cdot\left(\frac{(i\cdot\mathrm{e})^{k-1}}{i}\right)^{i}\cdot\left(\frac{d(k-1)^{k-1}}{n^{k-1}}\right)^{i}
=i(k1)(e2(k1)di)i=O(1).\displaystyle=i(k-1)\left(\frac{\mathrm{e}^{2(k-1)}d}{i}\right)^{i}=O(1).

By Markov’s inequality this implies that Y2+Y3+Y4n1δY_{2}+Y_{3}+Y_{4}\leq n^{1-\sqrt{\delta}} asymptotically almost surely. Denote by UU the union of all 22-, 33- and 44- cycles in HH. Then the induced subhypergraph H[VU]H[V\setminus U] has girth at least 55 and, asymptotically almost surely, |U|n1δ|U|\leq n^{1-\delta}.

We are now ready to prove Theorem 1.5.

Proof of Theorem 1.5.

Our goal will be to find a subset UVU\subset V of size |U|nd1k1|U|\leq nd^{-\frac{1}{k-1}} that (i) contains all cycles of length at most 44 and every vertex of degree more than (1+δ)d(1+\delta)d; and (ii) such that, every vertex vv in VUV\setminus U has at most 9k2(d(lnd)2)1k1=o((dlnd)1k1)9k^{2}\left(\frac{d}{(\ln d)^{2}}\right)^{\frac{1}{k-1}}=o\left(\left(\frac{d}{\ln d}\right)^{\frac{1}{k-1}}\right) neighbors in UU. Note that in this case, according to Lemma 5.1, H[U]H[U] is k(d(lnd)2)1k1k\left(\frac{d}{(\ln d)^{2}}\right)^{\frac{1}{k-1}}-degenerate, concluding the proof assuming dd is sufficiently large. A similar idea has been used in [5, 6, 25].

Towards that end, let U1U_{1} be the set of vertices of degree more than (1+δ)d(1+\delta)d, and U2U_{2} the set of vertices that are contained in a 22-,33- or a 44-cycle. Notice that U1,U2,U_{1},U_{2}, can be found in polynomial time and, according to Lemmas 5.4 and 5.5, the size of U0:=|U1U2|U_{0}:=|U_{1}\cup U_{2}| is at most 3nd2\frac{3n}{d^{2}} for sufficiently large nn and dd.

We now start with U:=U0U:=U_{0} and as long as there exists a vertex vVUv\in V\setminus U having at least 9k2(d(lnd)2)1k19k^{2}\left(\frac{d}{(\ln d)^{2}}\right)^{\frac{1}{k-1}} neighbors in UU we do the following. Let Sv={u1,u2,,uN}S_{v}=\{u_{1},u_{2},\ldots,u_{N}\} be the neighbors of vv in UU. We choose an arbitrary hyperedge hh that contains vv and u1u_{1} and update UU and SvS_{v} by defining U:=UhU:=U\cup h and Sv:=SvhS_{v}:=S_{v}\setminus h. We keep repeating this operation until SvS_{v} is empty.

This process terminates with |U|<nd1k1|U|<nd^{-\frac{1}{k-1}} because, otherwise, we would get a subset UVU\subset V of size |U|=nd1k1|U|=nd^{-\frac{1}{k-1}} spanning more than

1k(nd1k1|U0|)×9k2(d(lnd)2)1k1×1k>nd1k1×(d(lnd)2)1k1\displaystyle\frac{1}{k}\left(\frac{n}{d^{\frac{1}{k-1}}}-|U_{0}|\right)\times 9k^{2}\left(\frac{d}{(\ln d)^{2}}\right)^{\frac{1}{k-1}}\times\frac{1}{k}>\frac{n}{d^{\frac{1}{k-1}}}\times\left(\frac{d}{(\ln d)^{2}}\right)^{\frac{1}{k-1}}

hyperedges, for sufficiently large dd. According to Lemma 5.1 however, HH does not contain any such set asymptotically almost surely.

6 Acknowledgements

The author is grateful to Dimitris Achlioptas, Irit Dinur and anonymous reviewers for detailed comments and feedback.

References

  • [1] Dimitris Achlioptas and Amin Coja-Oghlan. Algorithmic barriers from phase transitions. In 49th Annual IEEE Symposium on Foundations of Computer Science, FOCS 2008, October 25-28, 2008, Philadelphia, PA, USA, pages 793–802. IEEE Computer Society, 2008.
  • [2] Dimitris Achlioptas, Fotis Iliopoulos, and Alistair Sinclair. Beyond the lovász local lemma: Point to set correlations and their algorithmic applications. In David Zuckerman, editor, 60th IEEE Annual Symposium on Foundations of Computer Science, FOCS 2019, Baltimore, Maryland, USA, November 9-12, 2019, pages 725–744. IEEE Computer Society, 2019.
  • [3] Dimitris Achlioptas and Michael Molloy. The analysis of a list-coloring algorithm on a random graph. In Proceedings 38th Annual Symposium on Foundations of Computer Science, pages 204–212. IEEE, 1997.
  • [4] Miklós Ajtai, János Komlós, Janos Pintz, Joel Spencer, and Endre Szemerédi. Extremal uncrowded hypergraphs. Journal of Combinatorial Theory, Series A, 32(3):321–335, 1982.
  • [5] Noga Alon and Michael Krivelevich. The concentration of the chromatic number of random graphs. Combinatorica, 17(3):303–313, 1997.
  • [6] Noga Alon, Michael Krivelevich, and Benny Sudakov. List coloring of random and pseudo-random graphs. Combinatorica, 19(4):453–472, 1999.
  • [7] Peter Ayre, Amin Coja-Oghlan, and Catherine Greenhill. Hypergraph coloring up to condensation. Random Structures & Algorithms, 54(4):615–652, 2019.
  • [8] Tom Bohman, Alan Frieze, and Dhruv Mubayi. Coloring h-free hypergraphs. Random Structures & Algorithms, 36(1):11–25, 2010.
  • [9] Karthekeyan Chandrasekaran, Navin Goyal, and Bernhard Haeupler. Deterministic algorithms for the Lovász local lemma. SIAM J. Comput., 42(6):2132–2155, 2013.
  • [10] Ewan Davies, Ross J Kang, François Pirot, and Jean-Sébastien Sereni. An algorithmic framework for coloring locally sparse graphs. arXiv preprint arXiv:2004.07151, 2020.
  • [11] Paul Erdős and László Lovász. Problems and results on 33-chromatic hypergraphs and some related questions. In Infinite and finite sets (Colloq., Keszthely, 1973; dedicated to P. Erdős on his 60th birthday), Vol. II, pages 609–627. Colloq. Math. Soc. János Bolyai, Vol. 10. North-Holland, Amsterdam, 1975.
  • [12] Alan Frieze and Michał Karoński. Introduction to random graphs. Cambridge University Press, 2016.
  • [13] Alan Frieze and Dhruv Mubayi. Coloring simple hypergraphs. Journal of Combinatorial Theory, Series B, 103(6):767–794, 2013.
  • [14] Marylou Gabrié, Varsha Dani, Guilhem Semerjian, and Lenka Zdeborová. Phase transitions in the q-coloring of random hypergraphs. Journal of Physics A: Mathematical and Theoretical, 50(50):505002, 2017.
  • [15] A. Johansson. Asympotic choice number for triangle free graphs., 1996.
  • [16] A. Johansson. The choice number of sparse graphs. Unpublished manuscript, 1996.
  • [17] Jeff Kahn. Asymptotics of the chromatic index for multigraphs. Journal of Combinatorial Theory, Series B, 68(2):233–254, 1996.
  • [18] Jeff Kahn. Asymptotics of the list-chromatic index for multigraphs. Random Structures & Algorithms, 17(2):117–156, 2000.
  • [19] Dong Yeap Kang, Tom Kelly, Daniela Kühn, Abhishek Methuku, and Deryk Osthus. Graph and hypergraph colouring via nibble methods: A survey. arXiv preprint arXiv:2106.13733, 2021.
  • [20] Dong Yeap Kang, Tom Kelly, Daniela Kühn, Abhishek Methuku, and Deryk Osthus. A proof of the erd\\backslashh {\{o}\} s-faber-lov\\backslash’asz conjecture. arXiv preprint arXiv:2101.04698, 2021.
  • [21] Jeong Han Kim. On Brooks’ theorem for sparse graphs. Combinatorics, Probability and Computing, 4(2):97–132, 1995.
  • [22] János Komlós, János Pintz, and Endre Szemerédi. A lower bound for heilbronn’s problem. Journal of the London Mathematical Society, 2(1):13–24, 1982.
  • [23] Alexandr Kostochka, Dhruv Mubayi, Vojtěch Rödl, and Prasad Tetali. On the chromatic number of set systems. Random Structures & Algorithms, 19(2):87–98, 2001.
  • [24] Hanno Lefmann. Sparse parity-check matrices over GF (q). Combinatorics, Probability and Computing, 14(1-2):147–169, 2005.
  • [25] Tomasz Łuczak. The chromatic number of random graphs. Combinatorica, 11(1):45–54, 1991.
  • [26] Michael Molloy. The list chromatic number of graphs with small clique number. Journal of Combinatorial Theory, Series B, 134:264–284, 2019.
  • [27] Michael Molloy and Bruce Reed. A bound on the total chromatic number. Combinatorica, 18(2):241–280, 1998.
  • [28] Michael Molloy and Bruce Reed. Graph colouring and the probabilistic method, volume 23 of Algorithms and Combinatorics. Springer-Verlag, Berlin, 2002.
  • [29] Robin A. Moser and Gábor Tardos. A constructive proof of the general Lovász local lemma. J. ACM, 57(2):Art. 11, 15, 2010.
  • [30] Michel Talagrand. Concentration of measure and isoperimetric inequalities in product spaces. Publications Mathématiques de l’Institut des Hautes Etudes Scientifiques, 81(1):73–205, 1995.
  • [31] Van H Vu. On the choice number of random hypergraphs. Combinatorics Probability and Computing, 9(1):79–95, 2000.
  • [32] Van H Vu. A general upper bound on the list chromatic number of locally sparse graphs. Combinatorics, Probability and Computing, 11(1):103–111, 2002.
  • [33] Lenka Zdeborová and Florent Krzakala. Phase transitions in the coloring of random graphs. Physical Review E, 76(3):031131, 2007.

Appendix A Proofs omitted from Section 4

In this section we prove Lemmas 4.14.3 and Lemma 3.4, Part (b).

A.1 Proof of Lemma 4.1

We proceed by induction. The case i=1i=1 is straightforward to verify since R1,r=0R_{1,r}=0 for every r[k2]r\in[k-2], while R1,k1=lnΔ(1+δ)k1R_{1,k-1}=\frac{\ln\Delta}{(1+\delta)^{k-1}}. Therefore, we inductively assume the claim for ii, and consider the case i+1i+1. Note that the inductive hypothesis implies that Keepi=Ω(1)\mathrm{Keep}_{i}=\Omega(1) since 11xe1x11-\frac{1}{x}\geq\mathrm{e}^{-\frac{1}{x-1}} for every x2x\geq 2 and, thus,

Keepiexp(r=1k1Ti,r(α1Li)r1)exp(Kk2(k2)1δ100k),\displaystyle\mathrm{Keep}_{i}\geq\mathrm{exp}\left(-\sum_{r=1}^{k-1}\frac{T_{i,r}}{(\alpha^{-1}L_{i})^{r}-1}\right)\geq\mathrm{exp}\left(-\frac{Kk^{2(k-2)}}{1-\frac{\delta}{100k}}\right), (59)

for sufficiently large Δ\Delta. Recalling (4) and (5), we have:

Ri+1,r\displaystyle R_{i+1,r} =\displaystyle= j=rk1(Ti,jLi+1r(Keepi(1αKeepi))r(jr)(αKeepiLi)jr)\displaystyle\sum_{j=r}^{k-1}\left(\frac{T_{i,j}}{L_{i+1}^{r}}\cdot\left(\mathrm{Keep}_{i}\left(1-\alpha\mathrm{Keep}_{i}\right)\right)^{r}{j\choose r}\left(\frac{\alpha\mathrm{Keep}_{i}}{L_{i}}\right)^{j-r}\right)
+1Li+1r(j=rk1(jr)αjrTi,jLijr)2/3+4k2(kr)αr+1(LiLi+1)rlnΔ=1k1Ti,Li2(lnΔ)2\displaystyle+\frac{1}{L_{i+1}^{r}}\left(\sum_{j=r}^{k-1}{j\choose r}\alpha^{j-r}\frac{T_{i,j}}{L_{i}^{j-r}}\right)^{2/3}+4k^{2(k-r)}\alpha^{-r+1}\left(\frac{L_{i}}{L_{i+1}}\right)^{r}\ln\Delta\sum_{\ell=1}^{k-1}\frac{T_{i,\ell}}{L_{i}^{2\ell}(\ln\Delta)^{2\ell}}
=\displaystyle= j=rk1(Ti,jLir(KeepiLi1/3)r(Keepi(1αKeepi))r(jr)(αKeepiLi)jr)\displaystyle\sum_{j=r}^{k-1}\left(\frac{T_{i,j}}{L_{i}^{r}\left(\mathrm{Keep}_{i}-L_{i}^{-1/3}\right)^{r}}\cdot\left(\mathrm{Keep}_{i}\left(1-\alpha\mathrm{Keep}_{i}\right)\right)^{r}{j\choose r}\left(\frac{\alpha\mathrm{Keep}_{i}}{L_{i}}\right)^{j-r}\right)
+O(1lnΔ)\displaystyle+O\left(\frac{1}{\ln\Delta}\right)
=\displaystyle= j=rk1(Ri,j(1αKeepi)r(1Li1/3Keepi)r(jr)(αKeepi)jr)+O(1lnΔ)\displaystyle\sum_{j=r}^{k-1}\left(R_{i,j}\cdot\frac{\left(1-\alpha\mathrm{Keep}_{i}\right)^{r}}{\left(1-\frac{L_{i}^{-1/3}}{\mathrm{Keep}_{i}}\right)^{r}}{j\choose r}\left(\alpha\mathrm{Keep}_{i}\right)^{j-r}\right)+O\left(\frac{1}{\ln\Delta}\right)
\displaystyle\leq j=rk1(Ri,j(1αKeepi2)r(jr)(αKeepi)jr)+O(1lnΔ)\displaystyle\sum_{j=r}^{k-1}\left(R_{i,j}\cdot\left(1-\frac{\alpha\mathrm{Keep}_{i}}{2}\right)^{r}{j\choose r}\left(\alpha\mathrm{Keep}_{i}\right)^{j-r}\right)+O\left(\frac{1}{\ln\Delta}\right)
\displaystyle\leq (1αKeepi2)r(Ri,r+j=r+1k1(jr)Ri,jαjr)+O(1lnΔ)\displaystyle\left(1-\frac{\alpha\mathrm{Keep}_{i}}{2}\right)^{r}\left(R_{i,r}+\sum_{j=r+1}^{k-1}{j\choose r}R_{i,j}\alpha^{j-r}\right)+O\left(\frac{1}{\ln\Delta}\right)
\displaystyle\leq (1αKeepi2)(k2(k1r)lnΔ+j=r+1k1(jr)k2(k1j)Kjr(lnΔ)jr1)+O(1lnΔ)\displaystyle\left(1-\frac{\alpha\mathrm{Keep}_{i}}{2}\right)\left(k^{2(k-1-r)}\ln\Delta+\sum_{j=r+1}^{k-1}{j\choose r}\frac{k^{2(k-1-j)}K^{j-r}}{(\ln\Delta)^{j-r-1}}\right)+O\left(\frac{1}{\ln\Delta}\right)
\displaystyle\leq (1αKeepi2)(k2(k1r)lnΔ+k2(k1(r+1))(r+1)K)+O(1lnΔ)\displaystyle\left(1-\frac{\alpha\mathrm{Keep}_{i}}{2}\right)\left(k^{2(k-1-r)}\ln\Delta+k^{2(k-1-(r+1))}(r+1)K\right)+O\left(\frac{1}{\ln\Delta}\right) (62)
\displaystyle\leq k2(k1r)lnΔK(Keepik2(k1r)2k2(k1(r+1))(r+1))+O(1lnΔ)\displaystyle k^{2(k-1-r)}\ln\Delta-K\left(\frac{\mathrm{Keep}_{i}k^{2(k-1-r)}}{2}-k^{2(k-1-(r+1))}(r+1)\right)+O\left(\frac{1}{\ln\Delta}\right)
\displaystyle\leq k2(k1r)lnΔ,\displaystyle k^{2(k-1-r)}\ln\Delta, (63)

for sufficiently large Δ\Delta, concluding the proof. Note that in deriving (A.1) we used the inductive hypothesis and that Li(lnΔ)20(k1)L_{i}\geq(\ln\Delta)^{20(k-1)} to obtain:

1Li+1r(j=rk1(jr)αjrTi,jLijr)2/3\displaystyle\frac{1}{L_{i+1}^{r}}\left(\sum_{j=r}^{k-1}{j\choose r}\alpha^{j-r}\frac{T_{i,j}}{L_{i}^{j-r}}\right)^{2/3} =(1Li+1r)1/3(j=rk1(jr)αjrRi,jLirLi+1r)2/3=o(1lnΔ),\displaystyle=\left(\frac{1}{L_{i+1}^{r}}\right)^{1/3}\left(\sum_{j=r}^{k-1}{j\choose r}\alpha^{j-r}R_{i,j}\frac{L_{i}^{r}}{L_{i+1}^{r}}\right)^{2/3}=o\left(\frac{1}{\ln\Delta}\right),
4k2(kr)αr+1(LiLi+1)rlnΔ=1k1Ti,Li2(lnΔ)2\displaystyle 4k^{2(k-r)}\alpha^{-r+1}\left(\frac{L_{i}}{L_{i+1}}\right)^{r}\ln\Delta\sum_{\ell=1}^{k-1}\frac{T_{i,\ell}}{L_{i}^{2\ell}(\ln\Delta)^{2\ell}} =4k2(kr)lnΔαr+1(LiLi+1)r=1k1Ri,Li(lnΔ)2\displaystyle=4k^{2(k-r)}\ln\Delta\cdot\alpha^{-r+1}\left(\frac{L_{i}}{L_{i+1}}\right)^{r}\sum_{\ell=1}^{k-1}\frac{R_{i,\ell}}{L_{i}^{\ell}(\ln\Delta)^{2\ell}}
=o(1lnΔ).\displaystyle=o\left(\frac{1}{\ln\Delta}\right).

In deriving (A.1) we used the facts that Keepi=Ω(1)\mathrm{Keep}_{i}=\Omega(1), Li,Ti,r(lnΔ)20(k1)L_{i},T_{i,r}\geq(\ln\Delta)^{20(k-1)}. In deriving (63) we used the fact that K=(100k3k)1K=(100k^{3k})^{-1} and, therefore, the second term in (62) is a negative constant, the inductive hypothesis, and the that Li(lnΔ)20(k1)L_{i}\geq(\ln\Delta)^{20(k-1)}, Keepi=Ω(1)\mathrm{Keep}_{i}=\Omega(1) and k3k\geq 3.

A.2 Proof of Lemma 4.3

We proceed by induction. The base cases are easy to verify since R1,r=0R_{1,r}^{\prime}=0 for every r[k2]r\in[k-2] and R1,k1=lnΔ(1+δ)k1R_{1,k-1}^{\prime}=\frac{\ln\Delta}{(1+\delta)^{k-1}}.

We first focus on the case r=k1r=k-1. We assume that the claim is true for i1i-1 and consider ii. Note that the inductive hypothesis, the facts that Lj(lnΔ)20(k1)L_{j}\geq(\ln\Delta)^{20(k-1)} for every 1<j<i1<j<i and KeepjC\mathrm{Keep}_{j}\geq C, imply:

4k2(kr)αr+1(LiLi+1)rlnΔ=1k1Ti,Li2(lnΔ)21(lnΔ)10(k1),\displaystyle 4k^{2(k-r)}\alpha^{-r+1}\left(\frac{L_{i}}{L_{i+1}}\right)^{r}\ln\Delta\sum_{\ell=1}^{k-1}\frac{T_{i,\ell}}{L_{i}^{2\ell}(\ln\Delta)^{2\ell}}\leq\frac{1}{(\ln\Delta)^{10(k-1)}}, (64)

for sufficiently large Δ\Delta, for every r[k1]r\in[k-1] and 1<ji1<j\leq i. Therefore, recalling (6), (7) and using (64), we have:

Ri,k1\displaystyle R_{i,k-1}^{\prime} \displaystyle\leq Ri1,k1(1αKeepi1)k1+1(lnΔ)10(k1)\displaystyle R_{i-1,k-1}^{\prime}(1-\alpha\mathrm{Keep}_{i-1})^{k-1}+\frac{1}{(\ln\Delta)^{10(k-1)}}
\displaystyle\leq Ri2,k1(1αKeepi2)k1(1αKeepi1)k1+(1αKeepi1)k1(lnΔ)10(k1)+1(lnΔ)10(k1)\displaystyle R_{i-2,k-1}^{\prime}(1-\alpha\mathrm{Keep}_{i-2})^{k-1}(1-\alpha\mathrm{Keep}_{i-1})^{k-1}+\frac{(1-\alpha\mathrm{Keep}_{i-1})^{k-1}}{(\ln\Delta)^{10(k-1)}}+\frac{1}{(\ln\Delta)^{10(k-1)}}
\displaystyle\leq Ri2,k1(1αC)2(k1)+(1αC)k1(lnΔ)10(k1)+1(lnΔ)10(k1)\displaystyle R_{i-2,k-1}^{\prime}(1-\alpha C)^{2(k-1)}+\frac{(1-\alpha C)^{k-1}}{(\ln\Delta)^{10(k-1)}}+\frac{1}{(\ln\Delta)^{10(k-1)}}
\displaystyle\leq \displaystyle\ldots
\displaystyle\leq (1αC)(i1)(k1)R1,k1+1(lnΔ)10(k1)=0i1(1αC)(k1)\displaystyle(1-\alpha C)^{(i-1)(k-1)}R_{1,k-1}+\frac{1}{(\ln\Delta)^{10(k-1)}}\sum_{\ell=0}^{i-1}(1-\alpha C)^{(k-1)\ell}
\displaystyle\leq (1αC)(i1)(k1)lnΔ(1+δ)k1+1(lnΔ)5(k1)\displaystyle(1-\alpha C)^{(i-1)(k-1)}\frac{\ln\Delta}{(1+\delta)^{k-1}}+\frac{1}{(\ln\Delta)^{5(k-1)}}
\displaystyle\leq (1αC)(i1)(k1)lnΔ(1+δδk99)k1,\displaystyle(1-\alpha C)^{(i-1)(k-1)}\frac{\ln\Delta}{(1+\delta-\frac{\delta}{k^{99}})^{k-1}},

for sufficiently large Δ\Delta, concluding the proof for the case r=k1r=k-1.

We now focus on r[k2]r\in[k-2]. We first observe that

R2,r\displaystyle R_{2,r}^{\prime} \displaystyle\leq j=rk1(T1,j(L1)r(Keep1(1αKeep1))r(jr)(αKeep1L1)jr)+1(lnΔ)10(k1)\displaystyle\sum_{j=r}^{k-1}\left(\frac{T_{1,j}^{\prime}}{(L_{1}^{\prime})^{r}}\cdot\left(\mathrm{Keep}_{1}\left(1-\alpha\mathrm{Keep}_{1}\right)\right)^{r}{j\choose r}\left(\frac{\alpha\mathrm{Keep}_{1}}{L_{1}^{\prime}}\right)^{j-r}\right)+\frac{1}{(\ln\Delta)^{10(k-1)}}
=\displaystyle= R1,k1(Keep1(1αKeep1))r(k1r)(αKeep1)k1r+1(lnΔ)10(k1)\displaystyle R_{1,k-1}^{\prime}\cdot\left(\mathrm{Keep}_{1}\left(1-\alpha\mathrm{Keep}_{1}\right)\right)^{r}{k-1\choose r}\left(\alpha\mathrm{Keep}_{1}\right)^{k-1-r}+\frac{1}{(\ln\Delta)^{10(k-1)}}
\displaystyle\leq (lnΔ)(k2r)(1+δ)k1Kk1r(k1r)+1(lnΔ)10(k1)\displaystyle\frac{(\ln\Delta)^{-(k-2-r)}}{(1+\delta)^{k-1}}K^{k-1-r}{k-1\choose r}+\frac{1}{(\ln\Delta)^{10(k-1)}}
\displaystyle\leq (1+δk100)k1r(lnΔ)(k2r)(1+δδk99)k1Kk1rp=rk2(p+1),\displaystyle\frac{(1+\frac{\delta}{k^{100}})^{k-1-r}(\ln\Delta)^{-(k-2-r)}}{(1+\delta-\frac{\delta}{k^{99}})^{k-1}}K^{k-1-r}\prod_{p=r}^{k-2}(p+1),

concluding the proof of the base cases.

Assume that the claim holds for all pairs (r,i)(r^{\prime},i^{\prime}), where r{r,,k1}r^{\prime}\in\{r,\ldots,k-1\} and ii1i^{\prime}\leq i-1. It suffices to prove that it also holds for any pair (r,i)(r,i), where i>2i>2 and r[k2]r\in[k-2]. To see this, observe that

Ri,r\displaystyle R_{i,r}^{\prime} \displaystyle\leq j=rk1(Ti1,j(Li)r(Keepi1(1αKeepi1))r(jr)(αKeepi1Li1)jr)+1(lnΔ)10(k1)\displaystyle\sum_{j=r}^{k-1}\left(\frac{T_{i-1,j}^{\prime}}{(L_{i}^{\prime})^{r}}\cdot\left(\mathrm{Keep}_{i-1}\left(1-\alpha\mathrm{Keep}_{i-1}\right)\right)^{r}{j\choose r}\left(\frac{\alpha\mathrm{Keep}_{i-1}}{L_{i-1}^{\prime}}\right)^{j-r}\right)+\frac{1}{(\ln\Delta)^{10(k-1)}} (65)
=\displaystyle= j=rk1(Ti1,j(Li1)jKeepi1jr(1αKeepi1)r(jr)αjr)+1(lnΔ)10(k1)\displaystyle\sum_{j=r}^{k-1}\left(\frac{T_{i-1,j}^{\prime}}{(L_{i-1}^{\prime})^{j}}\cdot\mathrm{Keep}_{i-1}^{j-r}\left(1-\alpha\mathrm{Keep}_{i-1}\right)^{r}{j\choose r}\alpha^{j-r}\right)+\frac{1}{(\ln\Delta)^{10(k-1)}}
=\displaystyle= (1αKeepi1)rj=rk1(Ri1,j(jr)(αKeepi1)jr)+1(lnΔ)10(k1)\displaystyle\left(1-\alpha\mathrm{Keep}_{i-1}\right)^{r}\sum_{j=r}^{k-1}\left(R_{i-1,j}^{\prime}{j\choose r}(\alpha\mathrm{Keep}_{i-1})^{j-r}\right)+\frac{1}{(\ln\Delta)^{10(k-1)}}
\displaystyle\leq (1αC)rj=rk1(Ri1,j(jr)αjr)+1(lnΔ)10(k1)\displaystyle\left(1-\alpha C\right)^{r}\sum_{j=r}^{k-1}\left(R_{i-1,j}^{\prime}{j\choose r}\alpha^{j-r}\right)+\frac{1}{(\ln\Delta)^{10(k-1)}}
\displaystyle\leq (1αC)rRi1,r\displaystyle\left(1-\alpha C\right)^{r}R_{i-1,r}^{\prime}
+j=r+1k1(jr)Kjr(lnΔ)1(jr)(1αC)j(i2)+r(1+δk100)k1j(1+δδk99)k1Ck1jp=jk2(p+1)\displaystyle+\sum_{j=r+1}^{k-1}{j\choose r}K^{j-r}(\ln\Delta)^{1-(j-r)}(1-\alpha C)^{j(i-2)+r}\frac{(1+\frac{\delta}{k^{100}})^{k-1-j}}{(1+\delta-\frac{\delta}{k^{99}})^{k-1}C^{k-1-j}}\prod_{p=j}^{k-2}(p+1)
+1(lnΔ)10(k1)\displaystyle\hphantom{\left(1-\alpha C\right)^{2r}asdfasfdsadfasdfasdfasdfsadfasdfadfasdfasfdfasdf}+\frac{1}{(\ln\Delta)^{10(k-1)}}
\displaystyle\leq (1αC)2rRi2,r\displaystyle\left(1-\alpha C\right)^{2r}R_{i-2,r}^{\prime}
+j=r+1k1(jr)Kjr(lnΔ)1(jr)(1αC)j(i2)+r(1+δk100)k1j(1+δδk99)k1Ck1jp=jk2(p+1)\displaystyle+\sum_{j=r+1}^{k-1}{j\choose r}K^{j-r}(\ln\Delta)^{1-(j-r)}(1-\alpha C)^{j(i-2)+r}\frac{(1+\frac{\delta}{k^{100}})^{k-1-j}}{(1+\delta-\frac{\delta}{k^{99}})^{k-1}C^{k-1-j}}\prod_{p=j}^{k-2}(p+1)
+j=r+1k1(jr)Kjr(lnΔ)1(jr)(1αC)j(i3)+2r(1+δk100)k1j(1+δδk99)k1Ck1jp=jk2(p+1)\displaystyle+\sum_{j=r+1}^{k-1}{j\choose r}K^{j-r}(\ln\Delta)^{1-(j-r)}(1-\alpha C)^{j(i-3)+2r}\frac{(1+\frac{\delta}{k^{100}})^{k-1-j}}{(1+\delta-\frac{\delta}{k^{99}})^{k-1}C^{k-1-j}}\prod_{p=j}^{k-2}(p+1)
+1+(1αC)r(lnΔ)10(k1)\displaystyle+\frac{1+(1-\alpha C)^{r}}{(\ln\Delta)^{10(k-1)}}
\displaystyle\leq (1αC)3rRi3,r\displaystyle\left(1-\alpha C\right)^{3r}R_{i-3,r}^{\prime} (66)
+j=r+1k1(jr)Kjr(lnΔ)1(jr)(1αC)j(i2)+r(1+δk100)k1j(1+δδk99)k1Ck1jp=jk2(p+1)\displaystyle+\sum_{j=r+1}^{k-1}{j\choose r}K^{j-r}(\ln\Delta)^{1-(j-r)}(1-\alpha C)^{j(i-2)+r}\frac{(1+\frac{\delta}{k^{100}})^{k-1-j}}{(1+\delta-\frac{\delta}{k^{99}})^{k-1}C^{k-1-j}}\prod_{p=j}^{k-2}(p+1)
+j=r+1k1(jr)Kjr(lnΔ)1(jr)(1αC)j(i3)+2r(1+δk100)k1j(1+δδk99)k1Ck1jp=jk2(p+1)\displaystyle+\sum_{j=r+1}^{k-1}{j\choose r}K^{j-r}(\ln\Delta)^{1-(j-r)}(1-\alpha C)^{j(i-3)+2r}\frac{(1+\frac{\delta}{k^{100}})^{k-1-j}}{(1+\delta-\frac{\delta}{k^{99}})^{k-1}C^{k-1-j}}\prod_{p=j}^{k-2}(p+1)
+j=r+1k1(jr)Kjr(lnΔ)1(jr)(1αC)j(i4)+3r(1+δk100)k1j(1+δδk99)k1Ck1jp=jk2(p+1)\displaystyle+\sum_{j=r+1}^{k-1}{j\choose r}K^{j-r}(\ln\Delta)^{1-(j-r)}(1-\alpha C)^{j(i-4)+3r}\frac{(1+\frac{\delta}{k^{100}})^{k-1-j}}{(1+\delta-\frac{\delta}{k^{99}})^{k-1}C^{k-1-j}}\prod_{p=j}^{k-2}(p+1)
+1+(1αC)r+(1αC)2r(lnΔ)10(k1)\displaystyle+\frac{1+(1-\alpha C)^{r}+(1-\alpha C)^{2r}}{(\ln\Delta)^{10(k-1)}}
\displaystyle\leq \displaystyle\ldots (69)
\displaystyle\leq (1αC)(i1)rR1,r\displaystyle\left(1-\alpha C\right)^{(i-1)r}R_{1,r}^{\prime}
+j=r+1k1(jr)Kjr(lnΔ)1(jr)p=jk2(p+1)(1+δk100)k1j(1+δδk99)k1Ck1j=1i1(1αC)j(i1)+r\displaystyle+\sum_{j=r+1}^{k-1}{j\choose r}K^{j-r}(\ln\Delta)^{1-(j-r)}\prod_{p=j}^{k-2}(p+1)\frac{(1+\frac{\delta}{k^{100}})^{k-1-j}}{(1+\delta-\frac{\delta}{k^{99}})^{k-1}C^{k-1-j}}\sum_{\ell=1}^{i-1}(1-\alpha C)^{j(i-\ell-1)+\ell r}
+=0i2(1αC)r(lnΔ)10(k1)\displaystyle+\frac{\sum_{\ell=0}^{i-2}(1-\alpha C)^{r\ell}}{(\ln\Delta)^{10(k-1)}}\
\displaystyle\leq j=r+1k1(jr)Kjr(lnΔ)1(jr)p=jk2(p+1)(1+δk100)k1j(1+δδk99)k1=1i1(1αC)j(i1)+r\displaystyle\sum_{j=r+1}^{k-1}{j\choose r}K^{j-r}(\ln\Delta)^{1-(j-r)}\prod_{p=j}^{k-2}(p+1)\frac{(1+\frac{\delta}{k^{100}})^{k-1-j}}{(1+\delta-\frac{\delta}{k^{99}})^{k-1}}\sum_{\ell=1}^{i-1}(1-\alpha C)^{j(i-\ell-1)+\ell r}
+O(1(lnΔ)5(k1))\displaystyle+O\left(\frac{1}{(\ln\Delta)^{5(k-1)}}\right)
=\displaystyle= j=r+1k1(jr)Kjr(lnΔ)1(jr)p=jk2(p+1)(1+δk100)k1j(1+δδk99)k1Ck1j=1i1(1αC)(i1)r+(i1)(jr)\displaystyle\sum_{j=r+1}^{k-1}{j\choose r}K^{j-r}(\ln\Delta)^{1-(j-r)}\prod_{p=j}^{k-2}(p+1)\frac{(1+\frac{\delta}{k^{100}})^{k-1-j}}{(1+\delta-\frac{\delta}{k^{99}})^{k-1}C^{k-1-j}}\sum_{\ell=1}^{i-1}(1-\alpha C)^{(i-1)r+(i-\ell-1)(j-r)}
+O(1(lnΔ)5(k1))\displaystyle+O\left(\frac{1}{(\ln\Delta)^{5(k-1)}}\right)
\displaystyle\leq (1αC)(i1)r(1+δδk99)k1j=r+1k1(jr)Kjr(lnΔ)1(jr)p=jk2(p+1)(1+δk100)k1jCk1j0(1αC)(jr)\displaystyle\frac{\left(1-\alpha C\right)^{(i-1)r}}{(1+\delta-\frac{\delta}{k^{99}})^{k-1}}\sum_{j=r+1}^{k-1}{j\choose r}K^{j-r}(\ln\Delta)^{1-(j-r)}\prod_{p=j}^{k-2}(p+1)\frac{(1+\frac{\delta}{k^{100}})^{k-1-j}}{C^{k-1-j}}\sum_{\ell\geq 0}(1-\alpha C)^{\ell(j-r)} (70)
=\displaystyle= (1αC)(i1)r(1+δδk99)k1(j=r+1k1(jr)Kjr(lnΔ)1(jr)p=jk2(p+1)Ck1j(1+δk100)k1j1(1αC)jr)\displaystyle\frac{\left(1-\alpha C\right)^{(i-1)r}}{(1+\delta-\frac{\delta}{k^{99}})^{k-1}}\left(\sum_{j=r+1}^{k-1}{j\choose r}K^{j-r}(\ln\Delta)^{1-(j-r)}\frac{\prod_{p=j}^{k-2}(p+1)}{C^{k-1-j}}\frac{(1+\frac{\delta}{k^{100}})^{k-1-j}}{1-(1-\alpha C)^{j-r}}\right) (71)
\displaystyle\leq (1αC)r(i1)lnΔ(1+δk100)k1r(1+δδk99)k1Ck1rp=rk2(p+1),\displaystyle(1-\alpha C)^{r(i-1)}\ln\Delta\cdot\frac{(1+\frac{\delta}{k^{100}})^{k-1-r}}{(1+\delta-\frac{\delta}{k^{99}})^{k-1}C^{k-1-r}}\prod_{p=r}^{k-2}(p+1), (72)

for sufficiently large Δ\Delta, concluding the proof. Note that in order to get (65) we upper bound Ri1,rR_{i-1,r}^{\prime} in the same way we upper bounded Ri,rR_{i,r}^{\prime}. We keep using the same steps to bound Ri2,r,Ri3,r,R_{i-2,r}^{\prime},R_{i-3,r}^{\prime},\ldots until we get (69). In deriving (69) we used that R1,r=0R_{1,r}=0 for every r[k2]r\in[k-2]. In going from (69) to (70) we start the summation in the last term from =0\ell=0 instead from =1\ell=1 in order to subsume the O(1/(lnΔ)5(k1))O(1/(\ln\Delta)^{5(k-1)}) error term. Finally, in going from (71) to (72) we multiply the term that corresponds to j=r+1j=r+1 in the summation by (1+δk100)k1r(1+\frac{\delta}{k^{100}})^{k-1-r} in order to subsume the terms of the summation that correspond to j>r+1j>r+1.

A.3 Proof of Lemma 3.4, Part(b)

We observe that it suffices to show that Ti,rTi,r(Ti,r)100r100r+1T_{i,r}^{\prime}\geq T_{i,r}-(T_{i,r}^{\prime})^{\frac{100r}{100r+1}} (since Ti,rTi,rT_{i,r}\geq T_{i,r}^{\prime} for every ii by definition) and proceed by using induction. Again, the base case is trivial, so we assume the statement is true for ii, and consider i+1i+1. We obtain the following.

Ti+1,r\displaystyle T_{i+1,r}^{\prime} =\displaystyle= j=rk1(Ti,j(Keepi(1αKeepi))r(jr)(αKeepiLi)jr)\displaystyle\sum_{j=r}^{k-1}\left(T_{i,j}^{\prime}\cdot\left(\mathrm{Keep}_{i}\left(1-\alpha\mathrm{Keep}_{i}\right)\right)^{r}{j\choose r}\left(\frac{\alpha\mathrm{Keep}_{i}}{L_{i}^{\prime}}\right)^{j-r}\right) (73)
+4k2(kr)α(α1Li)rlnΔ=1k1Ti,Li2(lnΔ)2\displaystyle+4k^{2(k-r)}\alpha(\alpha^{-1}L_{i})^{r}\ln\Delta\sum_{\ell=1}^{k-1}\frac{T_{i,\ell}}{L_{i}^{2\ell}(\ln\Delta)^{2\ell}}
\displaystyle\geq j=rk1((Ti,j(Ti,j)100r100r+1)(Keepi(1αKeepi))r(jr)(αKeepiLi)jr)\displaystyle\sum_{j=r}^{k-1}\left(\left(T_{i,j}-(T_{i,j}^{\prime})^{\frac{100r}{100r+1}}\right)\left(\mathrm{Keep}_{i}\left(1-\alpha\mathrm{Keep}_{i}\right)\right)^{r}{j\choose r}\left(\frac{\alpha\mathrm{Keep}_{i}}{L_{i}^{\prime}}\right)^{j-r}\right)
+4k2(kr)α(α1Li)rlnΔ=1k1Ti,Li2(lnΔ)2\displaystyle+4k^{2(k-r)}\alpha(\alpha^{-1}L_{i})^{r}\ln\Delta\sum_{\ell=1}^{k-1}\frac{T_{i,\ell}}{L_{i}^{2\ell}(\ln\Delta)^{2\ell}}
=\displaystyle= Ti+1,r(j=rk1(jr)αjrTi,jLijr)2/3\displaystyle T_{i+1,r}-\left(\sum_{j=r}^{k-1}{j\choose r}\alpha^{j-r}\frac{T_{i,j}}{L_{i}^{j-r}}\right)^{2/3} (75)
j=rk1Ti,j(Keepi(1αKeepi))r(αKeepi)jr(1Lijr1(Li)jr)\displaystyle-\sum_{j=r}^{k-1}T_{i,j}\left(\mathrm{Keep}_{i}(1-\alpha\mathrm{Keep}_{i})\right)^{r}(\alpha\mathrm{Keep}_{i})^{j-r}\left(\frac{1}{L_{i}^{j-r}}-\frac{1}{(L_{i}^{\prime})^{j-r}}\right)
j=rk1(Ti,j)100r100r+1(Keepi(1αKeepi))r(jr)(αKeepiLi)jr\displaystyle-\sum_{j=r}^{k-1}(T_{i,j}^{\prime})^{\frac{100r}{100r+1}}\left(\mathrm{Keep}_{i}\left(1-\alpha\mathrm{Keep}_{i}\right)\right)^{r}{j\choose r}\left(\frac{\alpha\mathrm{Keep}_{i}}{L_{i}^{\prime}}\right)^{j-r}
\displaystyle\geq Ti+1,r(j=rk1(jr)αjrTi,jLijr)2/3\displaystyle T_{i+1,r}-\left(\sum_{j=r}^{k-1}{j\choose r}\alpha^{j-r}\frac{T_{i,j}}{L_{i}^{j-r}}\right)^{2/3}
j=r+1k1(Keepi(1αKeepi))r(αKeepi)jrTi,jLijrO(Li16)\displaystyle-\sum_{j=r+1}^{k-1}\left(\mathrm{Keep}_{i}(1-\alpha\mathrm{Keep}_{i})\right)^{r}(\alpha\mathrm{Keep}_{i})^{j-r}\frac{T_{i,j}}{L_{i}^{j-r}}\,O\left(L_{i}^{-\frac{1}{6}}\right)
(Keepi(1αKeepi))r100r+1(Ti+1,r)100r100r+1\displaystyle-\left(\mathrm{Keep}_{i}\left(1-\alpha\mathrm{Keep}_{i}\right)\right)^{\frac{r}{100r+1}}(T_{i+1,r}^{\prime})^{\frac{100r}{100r+1}}
\displaystyle\geq Ti+1,r(1Kk11300k2(lnΔ)k1)(Ti+1,r)100r100r+1\displaystyle T_{i+1,r}-\left(1-\frac{K^{k-1}}{1300k^{2}(\ln\Delta)^{k-1}}\right)(T_{i+1,r}^{\prime})^{\frac{100r}{100r+1}}
O(j=r+1k1αjrRi,jLir1/6+Ti,r2/3+(j=r+1k1αjrRi,jLir)2r/3)\displaystyle-O\left(\sum_{j=r+1}^{k-1}\alpha^{j-r}R_{i,j}L_{i}^{r-1/6}+T_{i,r}^{2/3}+\left(\sum_{j=r+1}^{k-1}\alpha^{j-r}R_{i,j}L_{i}^{r}\right)^{2r/3}\right)
\displaystyle\geq Ti+1,r(Ti+1,r)100r100r+1,\displaystyle T_{i+1,r}-(T_{i+1,r}^{\prime})^{\frac{100r}{100r+1}}, (76)

for sufficiently large Δ\Delta, concluding the proof of the lemma. Note that in deriving (75) we used the first part of Lemma 3.4, i.e., the fact that LiLi(Li)5/6L_{i}^{\prime}-L_{i}\leq(L_{i}^{\prime})^{5/6} to obtain

1LiLi(Li)5/6Li=(1Li)1/6(1Li)1/6\displaystyle 1-\frac{L_{i}}{L_{i}^{\prime}}\leq\frac{(L_{i}^{\prime})^{5/6}}{L_{i}^{\prime}}=\left(\frac{1}{L_{i}^{\prime}}\right)^{1/6}\leq\left(\frac{1}{L_{i}}\right)^{1/6}

and the fact that

j=rk1(Ti,j)100r100r+1(Keepi(1αKeepi))r(jr)(αKeepiLi)jr\displaystyle\sum_{j=r}^{k-1}(T_{i,j}^{\prime})^{\frac{100r}{100r+1}}\left(\mathrm{Keep}_{i}\left(1-\alpha\mathrm{Keep}_{i}\right)\right)^{r}{j\choose r}\left(\frac{\alpha\mathrm{Keep}_{i}}{L_{i}^{\prime}}\right)^{j-r}
=\displaystyle= (Keepi(1αKeepi))r100r+1j=rk1(Ti,j)100r100r+1(Keepi(1αKeepi))100r2100r+1(jr)(αKeepiLi)jr\displaystyle\left(\mathrm{Keep}_{i}\left(1-\alpha\mathrm{Keep}_{i}\right)\right)^{\frac{r}{100r+1}}\sum_{j=r}^{k-1}(T_{i,j}^{\prime})^{\frac{100r}{100r+1}}\left(\mathrm{Keep}_{i}\left(1-\alpha\mathrm{Keep}_{i}\right)\right)^{\frac{100r^{2}}{100r+1}}{j\choose r}\left(\frac{\alpha\mathrm{Keep}_{i}}{L_{i}^{\prime}}\right)^{j-r}
\displaystyle\leq (Keepi(1αKeepi))r100r+1j=rk1(Ti,j)100r100r+1(Keepi(1αKeepi))100r2100r+1((jr)(αKeepiLi)jr)100r100r+1\displaystyle\left(\mathrm{Keep}_{i}\left(1-\alpha\mathrm{Keep}_{i}\right)\right)^{\frac{r}{100r+1}}\sum_{j=r}^{k-1}(T_{i,j}^{\prime})^{\frac{100r}{100r+1}}\left(\mathrm{Keep}_{i}\left(1-\alpha\mathrm{Keep}_{i}\right)\right)^{\frac{100r^{2}}{100r+1}}\left({j\choose r}\left(\frac{\alpha\mathrm{Keep}_{i}}{L_{i}^{\prime}}\right)^{j-r}\right)^{\frac{100r}{100r+1}}
\displaystyle\leq (Keepi(1αKeepi))r100r+1(j=rk1(Ti,j)(Keepi(1αKeepi))r(jr)(αKeepiLi)jr)100r100r+1\displaystyle\left(\mathrm{Keep}_{i}\left(1-\alpha\mathrm{Keep}_{i}\right)\right)^{\frac{r}{100r+1}}\left(\sum_{j=r}^{k-1}(T_{i,j}^{\prime})\left(\mathrm{Keep}_{i}\left(1-\alpha\mathrm{Keep}_{i}\right)\right)^{r}{j\choose r}\left(\frac{\alpha\mathrm{Keep}_{i}}{L_{i}^{\prime}}\right)^{j-r}\right)^{\frac{100r}{100r+1}}
=\displaystyle= (Keepi(1αKeepi))r100r+1(Ti,r+14k2(kr)α(α1Li)rlnΔ=1k1Ti,Li2(lnΔ)2)100r100r+1\displaystyle\left(\mathrm{Keep}_{i}\left(1-\alpha\mathrm{Keep}_{i}\right)\right)^{\frac{r}{100r+1}}(T_{i,r+1}^{\prime}-4k^{2(k-r)}\alpha(\alpha^{-1}L_{i})^{r}\ln\Delta\sum_{\ell=1}^{k-1}\frac{T_{i,\ell}}{L_{i}^{2\ell}(\ln\Delta)^{2\ell}})^{\frac{100r}{100r+1}}
<\displaystyle< (Keepi(1αKeepi))r100r+1(Ti,r+1)100r100r+1\displaystyle\left(\mathrm{Keep}_{i}\left(1-\alpha\mathrm{Keep}_{i}\right)\right)^{\frac{r}{100r+1}}(T_{i,r+1}^{\prime})^{\frac{100r}{100r+1}}

for sufficiently large Δ\Delta. In deriving (75) we used Lemma 4.1 to obtain that

(j=rk1(jr)αjrTi,jLijr)2/3\displaystyle\left(\sum_{j=r}^{k-1}{j\choose r}\alpha^{j-r}\frac{T_{i,j}}{L_{i}^{j-r}}\right)^{2/3} =(Ti,r+j=r+1k1(jr)αjrRi,jLir)2/3\displaystyle=\left(T_{i,r}+\sum_{j=r+1}^{k-1}{j\choose r}\alpha^{j-r}R_{i,j}L_{i}^{r}\right)^{2/3}
=O(Ti,r2/3+(j=r+1k1αjrRi,jLir)2r/3)\displaystyle=O\left(T_{i,r}^{2/3}+\left(\sum_{j=r+1}^{k-1}\alpha^{j-r}R_{i,j}L_{i}^{r}\right)^{2r/3}\right)

and, similarly, that

j=r+1k1(Keepi(1αKeepi))r(αKeepi)jrTi,jLijrO(Li16)=O(j=r+1k1αjrRi,jLir1/6).\displaystyle\sum_{j=r+1}^{k-1}\left(\mathrm{Keep}_{i}(1-\alpha\mathrm{Keep}_{i})\right)^{r}(\alpha\mathrm{Keep}_{i})^{j-r}\frac{T_{i,j}}{L_{i}^{j-r}}\,O\left(L_{i}^{-\frac{1}{6}}\right)=O\left(\sum_{j=r+1}^{k-1}\alpha^{j-r}R_{i,j}L_{i}^{r-1/6}\right).

We also used the fact that, using Corollary 4.2, we obtain

(Keepi(1αKeepi))r100r+1(1Kk112k2(lnΔ)k1)r100r+1\displaystyle\left(\mathrm{Keep}_{i}(1-\alpha\mathrm{Keep}_{i})\right)^{\frac{r}{100r+1}}\leq\left(1-\frac{K^{k-1}}{12k^{2}(\ln\Delta)^{k-1}}\right)^{\frac{r}{100r+1}} (1Kk11300k2(lnΔ)k1)\displaystyle\leq\left(1-\frac{K^{k-1}}{1300k^{2}(\ln\Delta)^{k-1}}\right)

for sufficiently large Δ\Delta. Finally, to derive (76) we observe that:

(Ti+1,r)100r100r+11300k2(lnΔ)k1\displaystyle\frac{(T_{i+1,r}^{\prime})^{\frac{100r}{100r+1}}}{1300k^{2}(\ln\Delta)^{k-1}} >11300k2(lnΔ)k1(j=rk1(Ti,j(jj)(Keepi(1αKeepi))r(αKeepiLi)jr))100r100r+1\displaystyle>\frac{1}{1300k^{2}(\ln\Delta)^{k-1}}\left(\sum_{j=r}^{k-1}\left(T_{i,j}^{\prime}{j\choose j}\left(\mathrm{Keep}_{i}(1-\alpha\mathrm{Keep}_{i})\right)^{r}\left(\frac{\alpha\mathrm{Keep}_{i}}{L_{i}^{\prime}}\right)^{j-r}\right)\right)^{\frac{100r}{100r+1}}
=Ω(11300k2(lnΔ)k1(Ti,r+j=r+1k1αjrRi,j(Li)r)100r100r+1)\displaystyle=\Omega\left(\frac{1}{1300k^{2}(\ln\Delta)^{k-1}}\left(T_{i,r}^{\prime}+\sum_{j=r+1}^{k-1}\alpha^{j-r}R_{i,j}^{\prime}(L_{i}^{\prime})^{r}\right)^{\frac{100r}{100r+1}}\right)
=ω(j=r+1k1αjrRi,jLir1/6+Ti,r2/3+(j=r+1k1αjrRi,jLir)2r/3),\displaystyle=\omega\left(\sum_{j=r+1}^{k-1}\alpha^{j-r}R_{i,j}L_{i}^{r-1/6}+T_{i,r}^{2/3}+\left(\sum_{j=r+1}^{k-1}\alpha^{j-r}R_{i,j}L_{i}^{r}\right)^{2r/3}\right), (77)

where in deriving (77) we used our assumption that Li,Ti,r(lnΔ)20(k1)L_{i},T_{i,r}\geq(\ln\Delta)^{20(k-1)} and the inductive hypothesis according to which Ri,j,LiR_{i,j}^{\prime},L_{i}^{\prime} are within a constant factor (arbitrarily close to 11) of Ri,j,LiR_{i,j},L_{i}.