The Phase Transition of Discrepancy in Random Hypergraphs
Abstract.
Motivated by the Beck-Fiala conjecture, we study the discrepancy problem in two related models of random hypergraphs on vertices and edges. In the first (edge-independent) model, a random hypergraph is constructed by fixing a parameter and allowing each of the vertices to join each of the edges independently with probability . In the parameter range in which and , we show that with high probability (w.h.p.) has discrepancy at least when , and at least when , where . In the second (edge-dependent) model, is fixed and each vertex of independently joins exactly edges uniformly at random. We obtain analogous results for this model by generalizing the techniques used for the edge-independent model with . Namely, for and , we prove that w.h.p. has discrepancy at least when , and at least when , where . Furthermore, we obtain nearly matching asymptotic upper bounds on the discrepancy in both models (when ), in the dense regime of . Specifically, we apply the partial colouring lemma of Lovett and Meka to show that w.h.p. and each have discrepancy , provided , and . This result is algorithmic, and together with the work of Bansal and Meka characterizes how the discrepancy of each random hypergraph model transitions from to as varies from to .
1. Introduction
A hypergraph111The definition of a hypergraph is equivalent to that of a set system, though we exclusively use the former terminology in this work. consists of a set of vertices together with a multiset of edges, where each edge is a subset of . We denote the size of an edge as . Note that is allowed to have duplicate edges (i.e. we may have for some ). We can bijectively represent by an -matrix , where if and if . We call the incidence matrix of . In particular, each pair of duplicate edges in corresponds to a pair of identical rows in . Moreover, we define the degree of a vertex of to be the number of edges containing that vertex, i.e. the number of ’s in the th column of . A classical problem in discrepancy theory is to find a 2-colouring of the vertices of so that every edge is as “balanced” as possible. To make this more precise, we define a colouring of to be a function . This can be extended to a map , by defining for each . We call the discrepancy of edge , and note that it measures how unbalanced the colouring is on that edge. Further, the discrepancy of colouring , denoted , is the discrepancy of the least balanced edge of . That is,
Finally, we define the discrepancy of hypergraph as
where the minimization is over all colourings of .
This and other related notions of combinatorial discrepancy have been studied from various angles and in different contexts. (For a more detailed introduction to the subject, we refer to books [18], [7] and [8].) One of the central problems in discrepancy theory in the above setting is to bound the discrepancy of a hypergraph in terms of its maximum degree . In [5], it was proven by Beck and Fiala that the discrepancy of is no larger than . Moreover, they conjectured that the correct upper bound is of the order . There has been much work in trying to improve on the original bound of [5]. Most recently, it was proven by Bukh [6] that , where is the binary iterated logarithm. This of course yields no asymptotic improvement in terms of , but to this date is the strongest upper bound known which solely depends on . If the upper bound is allowed dependence on the multiple parameters of the hypergraph, then there are results yielding improvements for hypergraphs in the correct range of parameters. For instance, Banaszczyk [3] showed that if , then —a bound which was later made algorithmic by Bansal and Meka [4]. Recently, Potukuchi [20] proved that, for -regular , , where and is the incidence matrix of .
In order to find upper bounds which depend solely on the maximum degree, restricted classes of hypergraphs have instead been studied. For example, if is assumed to be both -regular and -uniform (that is, the incidence matrix has exactly ’s in each column and row), then a folklore application of the Lovász local lemma can be used to show that there exists a colouring which achieves discrepancy (see [10, 19] for details).
Another approach is to restrict one’s attention to hypergraphs which are generated randomly. In this work, we focus on two specific random hypergraph models. Both of these models are defined as distributions over the set of hypergraphs with vertices and edges.
1.1. The Edge-Independent Model
In [15], Hoberg and Rothvoss introduced a random hypergraph model, denoted , in which a probability parameter is given (in addition to and ). Their model, which we refer to as the edge-independent model, is a distribution on hypergraphs which we describe through the following randomized procedure:
-
•
Fix the vertex set .
-
•
For each , construct edge by placing each in independently with probability .
We denote and define to be the distribution of the random hypergraph . In other words, the entries of the incidence matrix of are independent Bernoulli random variables of parameter . We write to indicate that is drawn from .
If and are functions which depend on , then we say that satisfies a property w.h.p., provided that as , where is drawn from . Often, we abuse terminology slightly and say that the random hypergraph satisfies w.h.p.
Hoberg and Rothvoss showed that, if and for some sufficiently large constant , then w.h.p. for . A natural question left open by their work is whether or not continues to have constant discrepancy when transitions from to . Potukuchi [19] provided a positive answer to this question for the special case when by showing that if for , then w.h.p. . Very recently, Altschuler and Niles-Weed [2] used Stein’s method [9] in conjunction with the second moment method to substantially generalize this result to hold when depends on . This includes the challenging case when .
1.2. The Edge-Dependent Model
A related model was introduced by Ezra and Lovett in [10]. As before, we fix and , yet we now also consider a parameter which satisfies . The edge-dependent model, denoted , is again a distribution on hypergraphs, though we describe it through a different randomized procedure:
-
•
Fix vertex set .
-
•
For each vertex , independently and u.a.r. (uniformly at random) draw with .
-
•
For each , construct edge by defining .
By setting , we define to be the distribution of the random hypergraph . In other words, the incidence matrix of is a random matrix where each column has ones and zeros. Note that the columns of are independent, but the rows are not. We define what it means for to satisfy a property w.h.p. in the same way as in the edge-independent model.
Ezra and Lovett showed that, if , then w.h.p. . Bansal and Meka [4] later showed that the factor of is redundant, thereby matching the bound claimed in the Beck-Fiala conjecture. Specifically, they showed that, for the entire range of and , w.h.p., provided . This result can be easily modified to also apply to the edge-independent model, provided the analogous condition holds.
In [12], Franks and Saks considered the more general problem of vector balancing. Their main result concerns a collection of random matrix models in which the columns are generated independently. In particular, their results apply to the sparse regime () of both the random hypergraph models we have discussed. Specifically, they show that if , then w.h.p. , provided is drawn from or for . Finally, in a very recent work, Turner et al. [22] considered the problem of vector balancing when the entries of the random matrix are each distributed as standard Gaussian random variables which are independent and identically distributed (i.i.d.). Amongst other results, they showed that the discrepancy of is w.h.p., provided .
1.3. An Overview of Our Results
All results proven in this paper are asymptotic with respect to the parameter , the number of vertices of the model. Thus, we hereby assume that , and are functions which depend on , with .
While previous results have successfully matched (or improved upon) the conjectured Beck-Fiala bound of in the random hypergraph setting, they either apply to a restricted parameter range [10, 19, 15, 12], or do not provide asymptotically tight results for the full parameter range [10, 4]. In particular, when or , the correct order of the discrepancy is unknown in either model. In this paper, we obtain (almost) matching upper and lower bounds in the dense regime of . Moreover, our upper bounds are algorithmic. In the sparse regime of , we provide the first lower bounds which apply to the full parameter range under the mild assumption that both the average edge size and degree tend to infinity with . Proving the existence of a colouring whose discrepancy matches our lower bound in the regime remains open. This problem is particularly challenging from an algorithmic perspective, as the partial colouring lemma [17] does not appear to be useful in this range of parameters, and this is the main tool used in the literature.
We now formally state our main results:
Theorem 1.1.
Suppose that is generated from with , and bounded away from . If , then w.h.p.
Moreover, if , then w.h.p.
where .
Remark 1.
This bound complements the upper bound of in [2] for where , but also implies that if is a constant, then has non-constant discrepancy for . Thus, exhibits a sharp phase transition at for constant .
We obtain analogous results regarding the edge-dependent model :
Theorem 1.2.
Suppose that is generated from with , , and bounded away from . If , then w.h.p.
Moreover, if , then w.h.p.
where .
Remark 2.
Finally, we prove an upper bound in the dense regime of which holds for both models:
Theorem 1.3.
Assume that is drawn from or with , and pick any satisfying222Let us remark at this place that by we always mean the natural logarithm. In the proof of this theorem it is convenient to use to denote logarithm of base .
(1) |
If , and , then w.h.p.
Moreover, whenever this holds, we can find a colouring of with such a discrepancy in expected polynomial time.
Remark 3.
Theorem 1.3 provides asymptotically matching bounds for the lower bounds of Theorems 1.1 and 1.2 in a broad range of the dense regime . For instance, this happens when , where is a constant, since in that case clearly satisfies (1), and , so .
Corollary 1.4.
Assume that is drawn from or , where . If and for constant , then w.h.p.
Remark 4.
This shows that in the dense regime, the main parameter of interest describing the discrepancy of each random hypergraph model is the average edge size, , opposed to , the average/maximum degree (depending on the model).
2. Preliminaries
In this section, we first provide some central-limit-type results for sums of independent random variables, which will be used in Section 3 to obtain lower bounds on the discrepancy. In the sequel, we write to denote a generic random variable distributed as a standard Gaussian with cumulative distribution function
The Berry-Esseen Theorem (see section XVI.5 in [11]), which we state below for convenience, yields a quantitative form of the Central Limit Theorem for sums of independent random variables with finite third moment.
Theorem 2.1 (Berry-Esseen).
The following holds for some universal constant . Let be independent random variables with , and for all . Consider the sum , with standard deviation , and let . Assume . Then,
Remark 5.
There has been a series of works improving upon the constant , the latest of which by Shevtsova [21] shows that in our setting. However, since we are only concerned with the asymptotic growth of discrepancy, the precise value of is not important.
Theorem 2.1 and the triangle inequality immediately yield the following corollary:
Corollary 2.2.
For any interval ,
We will apply this result to linear combinations of independent Bernoulli’s with coefficients in . More precisely, let be independent random variables with for some (where is a Bernoulli of parameter ). Given a vector , consider the sum , whose standard deviation we denote by . Under these assumptions we obtain the following bound.
Lemma 2.3.
For any bounded interval ,
(When , the right-hand side of the bound above is simply interpreted as .)
Proof.
Let . Then we centre by defining the random variables and setting
Observe that has the same standard deviation as , which we assume is non-zero (otherwise the lemma holds trivially). Moreover, if and only if , where and . Further,
and hence
Then, Corollary 2.2 immediately yields
To finalize the proof, it only remains to bound . In order to do so, we will use the inequality
(2) |
which can be found in [23]. Furthermore, note that is an even function, decreasing with . Combining this fact with (2), we get
which concludes the proof of the lemma. ∎
Now suppose there exist , and such that
(3) |
Then we can restate the upper bound of Lemma 2.3 in the following convenient way, which we use as our key tool in proving Theorems 1.1 and 1.2.
Lemma 2.4.
Suppose satisfy (3) for some , and . Then, for any bounded interval ,
(4) | ||||
(5) |
Proof.
In Theorem 1.3, we prove an upper bound on discrepancy in the dense regime (). In this parameter range, we make use of the algorithmic partial colouring lemma, a seminal result of Lovett and Meka [17] later made deterministic by Levy, Ramadas, and Rothvoss [16]. We defer the statement of this result to Lemma 4.1 of Section 4, as it will not be needed until then.
3. Lower Bounding Discrepancy
3.1. The Edge-Independent Model
We now return to the setting of hypergraph discrepancy in the context of the edge-independent model . Throughout this section, , and asymptotic statements are with respect to . We first observe that w.h.p. there are some edges containing an odd number of vertices and thus the discrepancy cannot be zero.
Proposition 3.1.
Suppose with , and bounded away from as . Then w.h.p. .
Proof.
By hypothesis, for some sufficiently small constant . Let be the edges of , and observe that the number of vertices contained in each edge is distributed as . Then the probability that has an even number of vertices is
(6) |
where we used the fact that as . Hence, the probability all the edges of contain an even number of vertices is . Therefore, w.h.p. has an edge with an odd number of vertices, and thus has discrepancy at least . ∎
Proposition 3.1 trivially implies Theorem 1.1 in the regime in which . We now use Lemma 2.4 to prove the remaining cases of Theorem 1.1 via a simple first moment argument:
Proof of Theorem 1.1.
Suppose that is generated from with , and bounded away from , as . We define
where , and choose a sufficiently small constant . To prove the theorem, it suffices to show that w.h.p. . Note that this is trivially true when , in view of Proposition 3.1. So we will assume that , and show that w.h.p. . Let be the set of all colourings , and let denote the number of colourings with discrepancy . Since the random edges of are i.i.d.,
(7) |
Note that , where denotes the indicator random variable of the event that edge contains vertex , so is distributed as in Section 2 (with and ). Hence, by applying (5) in Lemma 2.4 (with , and ) to each one of the terms of the last sum in (7), it follows that
where we also used that . Let us consider first the case that . Then, from the definition of and assuming ,
Now suppose that . In this case, we bound the factor on the right-hand side of (7) using the tighter inequality (4) in Lemma 2.4 instead of (5). Then, assuming that , we get
where we used the facts , and . In either case, , and therefore w.h.p. . ∎
3.2. The Edge-Dependent Model
In this section, we derive a lower bound on the discrepancy of a hypergraph generated from the edge-dependent model and prove Theorem 1.2. In view of our previous results for the edge-independent model, one natural approach is to compare both models and via a coupling procedure. For instance, let and for simplicity, and suppose that we can generate with and , with edge sets and , in such a way that w.h.p. for every we have , for some suitable . In particular, this implies that w.h.p. , and thus by Theorem 1.1. Unfortunately, since the standard deviation of the size of an edge is in either model, most naïve attempts to build such a coupling require , which is too large for our purposes. (In fact, it is not hard to build such a coupling with any .) As a result, while it is conceivable that a more delicate coupling argument works, we abandon this approach. Instead, we handle the dependencies of the edges of by applying a careful conditioning argument, while generalizing how we apply Lemma 2.4.
As in the edge-independent model, we first prove a constant lower bound on the discrepancy of .
Proposition 3.2.
Suppose with , and bounded away from as . Then w.h.p. .
Proof.
By hypothesis, for some sufficiently small constant . Our goal is to show that w.h.p. there is some edge in containing an odd number of vertices, so . As in the proof of Proposition 3.1, the probability that an edge has an odd number of vertices is (see (6)). Therefore, the expected number of edges with an odd number of vertices is . Similarly, the probability that two different edges (say, and ) have an odd number of vertices is . To see this, first define the bivariate generating function
For , the coefficient
gives the probability that edge has ones and edge has ones. Hence, the probability that both edges have an odd number of vertices is
which after tedious but straightforward calculations and using the facts that and , is equal to
and hence . This implies that and, by a standard application of Chebyshev’s inequality, we conclude that w.h.p. . This proves the proposition. ∎
Remark 6.
It is conceivable that in the proof of the proposition above one could obtain an upper bound on that is exponentially small in , in the same spirit as in the proof of Proposition 3.1. However, this would require some additional work due to the fact that the edges of are not formed independently.
Proposition 3.2 trivially implies Theorem 1.2 in the regime in which . To prove the remaining cases, we will generalize the ideas we used in the proof of Theorem 1.1. However, the dependencies among the edges make the argument much more delicate.
Proof of Theorem 1.2.
Suppose that is generated from with and satisfying and (as ) and with for some constant . For short, we use to denote the sample space of the distribution, i.e. the set of all possible outcomes of . Fix a constant , and define
where . Let be a sufficiently small constant. To prove Theorem 1.2, it suffices to show that w.h.p. . Note that this is trivially true when , in view of Proposition 3.2. So we will assume that , and show that w.h.p. . Note that this assumption ensures that , i.e. there are not too many more vertices than edges.
Let be the number of colourings with discrepancy . We would like to prove an analogue of (7) in order to bound , but unfortunately the random edges of are no longer i.i.d. In order to overcome this obstacle, we introduce some random variables that will play an essential role in the analysis of . For each and , let be the value of the entry of the incidence matrix of , and let denote the -th row of . Moreover, for each , we define
(8) |
In other words, counts the number of ones that appear in the -th column of below the -th row (recall that each column of has exactly ones). Note that each can be expressed as a function of the first rows of , and moreover the distribution of conditional on the outcome of can be described solely in terms of . More formally, for each , let be the sigma algebra generated by , and let denote the trivial sigma algebra. Then, for each and , is measurable with respect to , and (for )
Intuitively speaking, we would like that the above conditional probabilities remain close to (on average) as we reveal new rows of , at least for a large number of rows. In view of that, for each , we consider the event that for every
(Here from our choice of .) Observe that is a decreasing sequence of events, and each is -measurable by construction. Let . We need the following technical result, which we prove in Section 3.3.
Proposition 3.3.
Under the assumptions in the proof of Theorem 1.2, event holds w.h.p.
Now let be the set of all colourings , and pick an arbitrary . For each , let denote the event that , and let . (By convention, .) Clearly, and are -measurable. Note that, conditional upon any outcome of satisfying , the random variable is distributed as in Section 2 (with and ) and it satisfies the conditions of Lemma 2.4 (with and ). Hence, we can use that lemma to bound the conditional probability of . We first consider the sparse regime of . By Lemma 2.4, assuming and since ,
(9) |
In particular, since is -measurable and is contained in ,
Thus, for each ,
and inductively
Let . Next, we will bound from below based on the discrepancies of the first rows of when holds. First note that, since ,
(10) |
and then, by applying Markov’s inequality to the random variable ,
(11) |
Before proceeding with the proof, we consider the dense regime of . In that case, we obtain an analogue of (9) by using the tighter inequality (4) in Lemma 2.4 instead of (5). With and assuming that , we get
where we used the fact that . Reasoning as before, we obtain the following analogue of (10):
where we used the facts that and . As a result, our bound in (11) is also valid when as well. Then, in either regime ( or ),
(12) |
by (11), Proposition 3.3 and the fact . This shows that w.h.p. , and concludes the proof of Theorem 1.2. ∎
3.3. Proof of Proposition 3.3
In this section, we prove Proposition 3.3. For any , let and . Suppose that is a fixed subset of size . If is a random subset of size , then the distribution of the random variable is said to be hypergeometric with parameters and . We denote this distribution by in what follows. Now, is at least as concentrated about its expectation as the binomial distribution, (see Chapter 21 in [13] for details). As such, standard Chernoff bounds ensure the following:
Theorem 3.4.
Suppose that , and . In this case, for every ,
Let be the adjacency matrix of , which has exactly ones in each column at random positions. Let . Recall that, for each and , the random variable counts the number of ones in the -th column and below the -th row of (cf. (8)). Also recall (for ). Clearly, , so we may apply Theorem 3.4 to control the value of , and thus of . Let and . For a fixed column , our goal is to show that w.h.p. remains “close” to for all . By combining the error term in Theorem 3.4 with a naïve union bound, we can bound the probability of failure by something of the order of , which does not tend to unless . To overcome this, we need a more subtle argument in which we take the union bound over a smaller set of indices and take into account that does not change too much between two consecutive values of . This is made more precise in the following claim:
Proposition 3.5.
Assume with and . Fix . Then, with probability at least
it holds that, for all ,
(13) |
Proof.
As the columns of are identically distributed, we may assume that in what follows. We thus drop the index from the notation of for simplicity. Recall with for each .
Let , which satisfies by assumption. Our first goal is to partition the set into intervals, each of size at most . For each let , and let . Then, setting for , gives us the desired partition of . Now, let . Since , the set is contained in . Clearly, since . If , then , which implies by integrality. This contradicts the fact that . As a result, .
For each , define and let . In other words, each counts the number of ones in the -th column of below all the rows indexed by . We will prove that the variables are concentrated around their mean, and from that derive a concentration result for . Observe that must be defined slightly differently, due to divisibility issues. Fortunately, our argument will only require the analysis of , where , so this fact will cause no trouble. For ,
and therefore, for every ,
(14) |
where we also used the fact that . Now, let be the event that
A direct application of Theorem 3.4 yields
Using the fact that and the rough bound , we conclude that
(15) |
Finally, we turn our attention to . For each , we pick such that . Then, by monotonicity,
Combining this with (14) yields
As a result, event implies that for every ,
and hence
(16) |
(Note that the equation above is also valid for , since .) Dividing (16) by , we conclude that event implies that, for every ,
Our bound on in (15) completes the proof of the proposition. ∎
Corollary 3.6.
Suppose and as . Set and . Given any fixed constant and any , the following holds with probability at least . For every ,
(17) |
Proof.
Since the probability bound in the statement is asymptotic as , we will implicitly assume throughout the proof that is sufficiently large for all the inequalities therein to be valid. First, define . Clearly, . Observe that, since , we have
(18) |
In particular , and thus all the assumptions in Proposition 3.5 are satisfied. Moreover,
As a result, we can relax the inequalities in (13) to
which implies that (eventually for sufficiently large). In view of Proposition 3.5, this fails for some with probability at most
This finishes the proof of the corollary. ∎
Now we are ready to prove Proposition 3.3, which we restate below in a more explicit form for convenience.
Proposition 3.7.
Let be fixed constants with . Assume that , and as , and suppose that . Let . Then w.h.p., for every ,
(19) |
and
(20) |
Proof.
For , we say that column of is controllable if, for every , it holds that
where . Let denote the set of indices of the uncontrollable columns. Then, by Corollary 3.6 (with replaced by ),
Hence, we can apply Markov’s inequality to ensure that w.h.p. On the other hand, by applying the trivial lower bound to for each controllable column ,
w.h.p., thereby proving (19) (as ).
In order to verify that (20) holds, we first consider the regime in which . Observe then that deterministically
for each and . In particular, since (in view of (18)) and , it holds that
Thus, (20) holds in this regime, as is a fixed constant. On the other hand, if , we can apply Corollary 3.6 again, which ensures that with probability at least
we have
for all and . The proof is therefore complete. ∎
4. Upper Bounding Discrepancy—Proof of Theorem 1.3
The main tool we make use of is the algorithmic partial colouring lemma [17], as done in [4, 19, 20]. For convenience, we restate this lemma in the relevant hypergraph terminology, where we define a fractional colouring to be a relaxation of a (hypergraph) colouring, whose values are allowed to lie in the interval :
Lemma 4.1 (Partial Colouring Lemma [17]).
Suppose that is a hypergraph with edges and vertices which are coloured by some fractional colouring . Moreover, assume that and are non-negative values such that
(21) |
Under these assumptions, there exists some fractional colouring for which
-
(1)
for all , and
-
(2)
for at least vertices of .
Moreover, can be computed by a randomized algorithm in expected time
Remark 7.
In what follows, it will be convenient to refer to as the target (fractional) colouring and as the rounding parameter. We make use of Lemma 4.1 in the same way as in [4, 17, 19, 20]. In fact, we analyze the same two-phase algorithm considered by both Bansal and Meka [4] and Potukuchi [19, 20], though we must tune the relevant asymptotics carefully in order to achieve the bound claimed in Theorem 1.3. In particular, we shorten phase one and change the target discrepancy in each application of Lemma 4.1. These modifications allow us to derive more precise asymptotics in the studied range of parameters.
Let us suppose that is a hypergraph drawn from or , where . From now on, we assume that is a power of for convenience. This follows w.l.o.g. as we can always add extra vertices to which do not lie in any of the edges of . Given the lower bounds of Theorems 1.1 and 1.2, ideally we would like to compute an output colouring, with matching discrepancy. However, without finer proof techniques, this does not seem fully attainable for the full parameter range of . Let us fix . We recall the definition of as given in the statement of Theorem 1.3: For all sufficiently large, and
(22) |
Let be the target upper bound on the discrepancy that we are aiming to prove. Our argument is based on analyzing the Iterated-Colouring-Algorithm, which we now describe. Fix , , and let , and . For convenience, we refer to the below procedure as round . We define as the desired discrepancy bound to be attained in round .
-
(1)
Remove all edges of of size less than or equal to .
-
(2)
Update to be for each . Update the target colouring, which we denote by , to be the previously computed colouring restricted to (here is the identically zero function by convention).
-
(3)
If , then apply Lemma 4.1 to with the above values, yielding the fractional colouring . Otherwise, abort the round and declare an error.
-
(4)
Assuming an error did not occur, compute , such that for all , where . Afterwards, construct by restricting the edges of to , where .
We refer to the rounds as phase one of the algorithm’s execution. Note that we refer to the vertices of as inactive after round , as the value assigns to them will not change at any point onwards. We refer to the remaining vertices as being active. Observe then the following proposition:
Proposition 4.2.
For each and , it holds that
Assuming that none of the rounds yielded an error, there are exactly active vertices at the end of phase one, as . Heuristically, this means that we expect to have edges of roughly constant size. As such, we can easily complete the colouring by executing the phase two procedure for rounds , where . The phase two procedure is identical to that of phase one, with the exception that in step 2, we update to be (rather than ) for each . Analogously, we observe the following.
Proposition 4.3.
For each and , it holds that
Assuming that none of the rounds yielded an error, there are exactly active vertices at the end of phase two. In order to complete the construction of , we conclude with a post-processing phase. That is, we arbitrarily assign or to any of the vertices which remain active at the end of phase two. Finally, we round each remaining fractional value assigned by to the nearest integer within .
Let us assume that the above procedure succeeds in its execution on ; that is, it does not abort during any iteration in either phase one or two. In this case, we conclude the proof by showing the next lemma.
Lemma 4.4.
If neither phase one nor phase two fails then Theorem 1.3 holds.
Proof.
For each , let us formally extend to all of . That is, define for each , and keep unchanged on . Moreover, do the same for the target colouring . Observe then that once phase two ends, can be expressed as a sum of differences involving the partial colourings and Specifically,
Let be the time when edge becomes smaller than or if it never happens. After applying Propositions 4.2 and 4.3 we get that
The post-processing phase cannot increase the discrepancy that attains on any edge of by more than for the remaining active vertices; as we already observed, there are at most of them. The rounding of inactive vertices increases the discrepancy by at most 1. ∎
Bounding the Failure Probability
First, we recall the following lemma proven in [4] by Bansal and Meka:
Lemma 4.5 (Lemma 6 in [4]).
Suppose that is generated from and is a fixed sub-matrix of the incidence matrix of . If , and corresponds to the event in which each row of has at least 1’s, then
While this lemma is stated for the case when is generated from , the upper bound on is proven by instead bounding the probability of the analogous event when is generated from for . As such, this lemma extends to the edge independent model. We now restate it in a form which will be more convenient for our purposes.
Lemma 4.6.
Suppose that is generated from or for , whose incidence matrix we denote by . If , then define as the event in which there exists an sub-matrix of in which each row has at least 1’s. In this case,
Using this lemma, we can ensure that w.h.p. Iterated-Colouring-Algorithm will not abort during phase one (Proposition 4.7) or two (Proposition 4.8) and thus conclude the proof of Theorem 1.3.
Proposition 4.7.
If Iterated-Colouring-Algorithm inputs a hypergraph drawn from or where , then w.h.p. it does not abort during phase one, provided we assume that and .
Proof.
Given , we say that round is good, provided there are at most rows of whose size is greater than , where satisfies (22). Otherwise, we say that the round is bad. Recall that .
Now, if round is good, then we claim that Iterated-Colouring-Algorithm does not abort in iteration . To see this, it suffices to show that for sufficiently large
where , and for . Observe now that since the round is good, we get that
On the other hand, since ,
where the last line follows since , as . Thus,
We now must show that w.h.p., all of the rounds are good. Now, observe that if some round is bad, then there exists an sub-matrix of , say , in which each row of has greater than 1’s. In fact, since and , we can take a sub-matrix of (and thus of ) in which each row has at least 1’s, and whose size is . Thus, we observe the following claim:
-
(1)
If a bad round occurs, then there exists a sub-matrix of in which each row has more than 1’s.
Let us define as this latter event; namely, that there exists an sub-matrix of in which each row has more than 1’s. In order to complete the proof, it suffices to show that w.h.p., does not occur. Recall . as the number of active vertices drops by exactly half in each round. It follows that
Thus, we can apply Lemma 4.6 to ensure that
where the inequalities follow since . Now,
so
Thus, by our assumption (22) on , we get that
and
as . The proposition follows as
and
Proposition 4.8.
If Iterated-Colouring-Algorithm inputs a hypergraph drawn from or where , then w.h.p. it does not abort in phase two, provided as and .
Proof.
Suppose that for . Recall and that during phase two, for each . Thus, we get that
On the other hand, in order for an edge of to remain in , it must have greater than vertices which lie in . As a result, if Iterated-Colouring-Algorithm aborts in round , then must have at least edges of size greater than . In particular, since , this implies that the incidence matrix of has an sub-matrix in which each row has greater than 1’s. Thus, if corresponds to the event in which has an sub-matrix in which each row has greater than 1’s, then we get the following claim:
-
(1)
If Iterated-Colouring-Algorithm aborts in some round , then must occur.
As a result, in order to show that w.h.p. Iterated-Colouring-Algorithm does not abort in any round it suffices to prove that does not occur w.h.p. Now, it follows that
as . Thus, we can apply Lemma 4.6 to ensure that
and so is upper bounded by
(23) |
after applying the same simplifications as in Proposition 4.7. The proposition then follows by assumption (22) on , as
and
5. Conclusion and Open Problems
We have lower bounded the discrepancy of the random hypergraph models and for the full parameter range in which and where . In the dense regime of , we have provided asymptotically matching upper bounds, under the assumption that for some constant . These upper bounds are algorithmic, and so the main question left open by our work is whether analogous upper bounds can be proven in the sparse regime of . Our lower bounds suggest that the discrepancy is , and while we believe that a second moment argument could be used to prove the existence of such a colouring—particularly, in the edge-independent model —the partial colouring lemma does not seem to be of much use here. This leaves open whether such a colouring can be computed efficiently in this parameter range. If this is not possible, then ideally one could find a reduction to a problem which is believed to be hard on average. One candidate may be the random-lattice problem of Ajtai [1] and Goldreich et al. [14], in which a random by matrix with i.i.d. entries from is generated, and one wishes to compute a vector such that .
Acknowledgements
This work was initiated at the 2019 Graduate Research Workshop in Combinatorics, which was supported in part by NSF grant #1923238, NSA grant #H98230-18-1-0017, a generous award from the Combinatorics Foundation, and Simons Foundation Collaboration Grants #426971 (to M. Ferrara), #316262 (to S. Hartke) and #315347 (to J. Martin). We thank Aleksandar Nikolov for suggesting the problem, and Puck Rombach and Paul Horn for discussions and encouragements in the early stages of the project. Moreover, T. Masařík received funding from the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation programme Grant Agreement 714704. He completed a part of this work while he was a postdoc at Simon Fraser University in Canada, where he was supported through NSERC grants R611450 and R611368. X. Pérez-Giménez was supported in part by Simons Foundation Grant #587019.
References
- [1] Miklós Ajtai “Generating Hard Instances of Lattice Problems (Extended Abstract)” In Proceedings of the Twenty-Eighth Annual ACM Symposium on Theory of Computing, STOC ’96 Philadelphia, Pennsylvania, USA: Association for Computing Machinery, 1996, pp. 99–108 DOI: 10.1145/237814.237838
- [2] Dylan J. Altschuler and Jonathan Niles-Weed “The Discrepancy of Random Rectangular Matrices” In CoRR abs/2101.04036, 2021 arXiv:2101.04036
- [3] Wojciech Banaszczyk “Balancing vectors and Gaussian measures of n-dimensional convex bodies” In Random Structures & Algorithms 12.4, 1998, pp. 351–360 DOI: 10.1002/(SICI)1098-2418(199807)12:4<351::AID-RSA3>3.0.CO;2-S
- [4] Nikhil Bansal and Raghu Meka “On the discrepancy of random low degree set systems” In Random Structures & Algorithms 57.3 Wiley, 2020, pp. 695–705 DOI: 10.1002/rsa.20935
- [5] József Beck and Tibor Fiala ““Integer-making” theorems” In Discrete Applied Mathematics 3.1 Elsevier BV, 1981, pp. 1–8 DOI: 10.1016/0166-218x(81)90022-6
- [6] Boris Bukh “An Improvement of the Beck–Fiala Theorem” In Combinatorics, Probability and Computing 25.3 Cambridge University Press, 2016, pp. 380–398 DOI: 10.1017/S0963548315000140
- [7] Bernard Chazelle “The Discrepancy Method: Randomness and Complexity” Cambridge University Press, 2000 DOI: 10.1017/cbo9780511626371
- [8] “A Panorama of Discrepancy Theory” Springer International Publishing, 2014 DOI: 10.1007/978-3-319-04696-9
- [9] Persi Diaconis and Susan Holmes “Stein’s Method: Expository Lectures and Applications”, Institute of Mathematical Statistics Institute of Mathematical Statistics, 2004 URL: https://books.google.ca/books?id=3n-iQIU9LNEC
- [10] Esther Ezra and Shachar Lovett “On the Beck-Fiala conjecture for random set systems” In Random Structures & Algorithms 54.4 Wiley, 2018, pp. 665–675 DOI: 10.1002/rsa.20810
- [11] William Feller “An introduction to probability theory and its applications. Vol. II.”, Second edition John Wiley & Sons Inc., 1971
- [12] Cole Franks and Michael Saks “On the discrepancy of random matrices with many columns” In Random Structures & Algorithms 57.1, 2020, pp. 64–96 DOI: 10.1002/rsa.20909
- [13] Alan Frieze and Michał Karoński “Introduction to Random Graphs” Cambridge University Press, 2015 DOI: 10.1017/CBO9781316339831
- [14] Oded Goldreich, Shafi Goldwasser and Shai Halevi “Collision-Free Hashing from Lattice Problems” In Studies in Complexity and Cryptography. Miscellanea on the Interplay between Randomness and Computation: In Collaboration with Lidor Avigad, Mihir Bellare, Zvika Brakerski, Shafi Goldwasser, Shai Halevi, Tali Kaufman, Leonid Levin, Noam Nisan, Dana Ron, Madhu Sudan, Luca Trevisan, Salil Vadhan, Avi Wigderson, David Zuckerman Berlin, Heidelberg: Springer Berlin Heidelberg, 2011, pp. 30–39 DOI: 10.1007/978-3-642-22670-0_5
- [15] Rebecca Hoberg and Thomas Rothvoss “A fourier-analytic approach for the discrepancy of random set systems” In Proceedings of the Thirtieth Annual ACM-SIAM Symposium on Discrete Algorithms, 2019, pp. 2547–2556 SIAM DOI: 10.1137/1.9781611975482.156
- [16] Avi Levy, Harishchandra Ramadas and T. Rothvoss “Deterministic Discrepancy Minimization via the Multiplicative Weight Update Method” In Integer Programming and Combinatorial Optimization, 2017, pp. 380–391 DOI: 10.1007/978-3-319-59250-3_31
- [17] Shachar Lovett and Raghu Meka “Constructive Discrepancy Minimization by Walking on the Edges” In SIAM Journal on Computing 44.5 Society for Industrial & Applied Mathematics (SIAM), 2015, pp. 1573–1582 DOI: 10.1137/130929400
- [18] Jiří Matoušek “Geometric Discrepancy: An Illustrated Guide”, Algorithms and Combinatorics 18 Springer-Verlag Berlin Heidelberg, 1999 DOI: 10.1007/978-3-642-03942-3
- [19] Aditya Potukuchi “Discrepancy in random hypergraph models”, 2018 arXiv:1811.01491 [math.CO]
- [20] Aditya Potukuchi “A Spectral Bound on Hypergraph Discrepancy” In 47th International Colloquium on Automata, Languages, and Programming (ICALP 2020) 168, Leibniz International Proceedings in Informatics (LIPIcs) Dagstuhl, Germany: Schloss Dagstuhl–Leibniz-Zentrum für Informatik, 2020, pp. 93:1–93:14 DOI: 10.4230/LIPIcs.ICALP.2020.93
- [21] Irina Shevtsova “An Improvement of Convergence Rate Estimates in the Lyapunov Theorem” In Doklady Mathematics 82, 2010, pp. 862–864 DOI: 10.1134/S1064562410060062
- [22] Paxton Turner, Raghu Meka and Philippe Rigollet “Balancing Gaussian vectors in high dimension” In Conference on Learning Theory, COLT 2020, 9-12 July 2020, Virtual Event [Graz, Austria] 125, Proceedings of Machine Learning Research PMLR, 2020, pp. 3455–3486 URL: http://proceedings.mlr.press/v125/turner20a.html
- [23] J.. Williams “An Approximation to the Probability Integral” In Annals of Mathematical Statistics 17.3 The Institute of Mathematical Statistics, 1946, pp. 363–365 DOI: 10.1214/aoms/1177730951