Moser-Tardos Algorithm: Beyond Shearer’s Bound
Abstract.
In a seminal paper (Moser and Tardos, JACM’10), Moser and Tardos developed a simple and powerful algorithm to find solutions to combinatorial problems in the variable Lovász Local Lemma (LLL) setting. Kolipaka and Szegedy (Kolipaka and Szegedy, STOC’11) proved that the Moser-Tardos algorithm is efficient up to the tight condition of the abstract Lovász Local Lemma, known as Shearer’s bound. A fundamental problem around LLL is whether the efficient region of the Moser-Tardos algorithm can be further extended.
In this paper, we give a positive answer to this problem. We show that the efficient region of the Moser-Tardos algorithm goes beyond the Shearer’s bound of the underlying dependency graph, if the graph is not chordal. Otherwise, the dependency graph is chordal, and it has been shown that Shearer’s bound exactly characterizes the efficient region for such graphs (Kolipaka and Szegedy, STOC’11; He, Li, Liu, Wang and Xia, FOCS’17).
Moreover, we demonstrate that the efficient region can exceed Shearer’s bound by a constant by explicitly calculating the gaps on several infinite lattices.
The core of our proof is a new criterion on the efficiency of the Moser-Tardos algorithm which takes the intersection between dependent events into consideration. Our criterion is strictly better than Shearer’s bound whenever the intersection exists between dependent events. Meanwhile, if any two dependent events are mutually exclusive, our criterion becomes the Shearer’s bound, which is known to be tight in this situation for the Moser-Tardos algorithm (Kolipaka and Szegedy, STOC’11; Guo, Jerrum and Liu, JACM’19).
1. Introduction
Suppose is a set of bad events. If the events are mutually independent, then we can avoid all of these events simultaneously whenever no event has probability 1. Lovász Local Lemma (LLL) [14], one of the most important probabilistic methods, allows for limited dependency among the events, but still concludes that all the events can be avoided simultaneously if each individual event has a bounded probability. In the most general setting (a.k.a. abstract LLL), the dependency among is characterized by an undirected graph , called a dependency graph of , which satisfies that for any vertex , is independent of . Here stands for the set of neighbors of vertex in a given graph .
We use to denote that (i) is a dependency graph of and (ii) the probability vector of is . Given a graph , define the abstract interior to be the set consisting of all vectors such that for any . In this context, the most frequently used abstract LLL can be stated as follows:
Theorem 1.1 ([58]).
Given any graph and any probability vector , if there exist real numbers such that for any , then .
Shearer [56] obtained the strongest possible condition for abstract LLL. Let be the set of all independent sets of an undirected graph and . For each , define the quantity
is called in Shearer’s bound of if for any . Otherwise we say is beyond Shearer’s bound of . Shearer’s result can be stated as follows.
Theorem 1.2 ([56]).
For any graph and any probability vector , if and only if is in Shearer’s bound of .
Variable Lovász Local Lemma. Variable Lovász Local Lemma (VLLL) is another quite general and common setting of LLL, which applies to variable-generated event systems. In this setting, there is a set of underlying mutually independent random variables , and each event can be fully determined by some variables of them. The dependency between events and variables can be naturally characterized by a bipartite graph , known as the event-variable graph, such that edge exists if and only if .
The variable setting is important, mainly because most applications of LLL have natural underlying independent variables, such as the satisfiability of CNF formulas[30, 31, 50, 16], hypergraph coloring [49, 29], and Ramsey numbers[57, 58, 32]. In particular, the groundbreaking result by Moser and Tardos [54] on constructive LLL applies in the variable setting.
There is a natural choice for the dependency graph of variable-generated systems, called the canonical dependency graph: two events are adjacent if they share some common variables. Formally, given a bipartite graph , its base graph is defined as the graph such that for any two vertices , if and only if and share common neighbors in . If is the event-variable graph of a variable-generated system , then is the canonical dependency graph of .
Given a graph , define the variable interior to be the set consisting of all vectors such that for any variable-generated event system . Obviously, for any . In contrast with the abstract LLL, the Shearer’s bound (of the canonical dependency graph) turns out to be not tight for variable-generated systems [36]: the containment is proper if and only if is not chordal111A graph is chordal if it has no induced cycle of length at least four..
Constructive (variable) Lovász Local Lemma and Moser-Tardos algorithm. The abstract LLL and the variable LLL mentioned above are not constructive in that they do not indicate how to efficiently find an object avoiding all the bad events. In a seminal paper [54], Moser and Tardos developed an amazingly simple efficient algorithm for variable-generated systems, depicted in Algorithm 1222Throughout the paper, the Moser-Tardos algorithm is allowed to follow arbitrary selection rules., and showed that this algorithm terminates quickly under the condition in Theorem 1.1. Following the Moser-Tardos algorithm (or MT algorithm for short), a large amount of effort devoted to constructive LLL, including the remarkable works which extend the MT techniques beyond the variable setting [38, 5, 3, 4, 42, 41]. The MT algorithm has been applied to many important problems, including -SAT [31], hypergraph coloring [32], Hamiltonian cycle [32], and their counting and sampling [27, 50, 16, 18, 44, 40].
Mainly because such a simple algorithm is so powerful and general-purpose, it is one of the most intriguing and fundamental problems on constructive LLL how powerful the MT algorithm is. Given a graph , define the Moser-Tardos interior to be the set consisting of all vectors such that the MT algorithm is efficient for any variable-generated event system . Clearly, for any . A major line of follow-up works explores [46, 55, 45, 10]. The best known criterion is obtained by Kolipaka and Szegedy [45]. They extended the MT interior to the Shearer’s bound. That is, they showed that . As mentioned above, if is not chordal, is properly contained in , so it is possible to further push beyond Shearer’s bound.
In this paper, we concentrate on the following open problem:
Problem 1: does properly contain for some ? If so, for what kind of graph ?
Rather than potential applications, our main motivations are the following fundamental problems around LLL itself:
-
•
The limitation of the constructive LLL in the variable setting. In the most fascinating problems around LLL, a mysterious conjecture says that there is an algorithm which is efficient for all variable-generated systems if for some and [59]. It would be a small miracle if the conjecture is true, since if so, one can always construct a solution efficiently in the variable setting if solutions are guaranteed to exist by the LLL condition. Towards this conjecture, a good start is to show that for some , as for which is not chordal.
-
•
The limitation of the MT algorithm. The MT algorithm is one of the most intriguing topics in modern algorithm researches, not only because it is very simple and with magic power, but also because it is closely related to the famous Walksat algorithm for random -SAT. A mysterious problem about the MT algorithm is where is its true limitation [59, 10]. It is conjectured that for any [59]. To prove this conjecture, the first step is to give a positive answer to Problem 1. Moreover, due to the connection between Shearer’s bound and the Repulsive Lattice Gas model, it is conjectured that essential connection exists between statistical mechanics and the MT algorithm [59]. Whether for each is critical to this conjecture.
Remark 1.3.
To explore the power of the MT algorithm in specific applications, one may employ special structures of the applications, such as the way the variables interact, to obtain sharp bounds rather than in terms of the canonical dependency graph only. Nevertheless, characterizing the power of the MT algorithm in terms of the canonical dependency graph is a very fundamental problem and also the focus of the major line of researches [54, 55, 9, 45]. Moreover, a major difficulty to strengthen the guarantees of the MT algorithm is that the analysis should be valid for all possible variable-generated event systems. It is not quite surprising to obtain better bounds if the event system has further restrictions. To substantially improve the guarantees of the MT algorithm and provide deep insight about its dynamics, we would rather focus on the general variable LLL setting than employ the special structures in the applications.
We should emphasize that Problem 1 is still quite open! As mentioned before, it has been proved that the Shearer’s bound is not tight for variable-generated systems [36]. However, this only says that there is some probability vector beyond the Shearer’s bound such that all variable-generated event systems must have a satisfying assignment. It is unclear whether the MT algorithm can construct such an assignment efficiently.
It also has been proved that the MT algorithm can still be efficient even beyond the Shearer’s bound for some specific applications [32]. Despite its novel contribution, this result does not provide an answer to Problem 1. The result in [32] focuses on the event systems with special structures. Thus, it only implies that there is a probability vector beyond the Shearer’s bound such that the MT algorithm is efficient for some restricted variable-generated event systems . However, to show , one must prove that the MT algorithm is efficient for all possible event systems, and this is one major difficulty to resolve Problem 1.
1.1. Results and contributions
We provide a complete answer to Problem 1 (Theorem 1.5): if is not chordal, then , i.e., the efficient region of the MT algorithm goes beyond Shearer’s bound. Otherwise, , because and for chordal graphs [36].
The core of the proof of Theorem 1.5 is a new convergence criterion for the MT algorithm (Theorem 1.6), which may be of independent interest. This new criterion takes the intersection between dependent events into consideration, and is strictly better than Shearer’s bound when there exists a pair of dependent events which are not mutually exclusive.
1.1.1. Moser-Tardos algorithm: beyond Shearer’s bound
Given a dependency graph and a probability vector , we say that is on the Shearer’s boundary of if is in Shearer’s bound and is not for any . A chordless cycle in a graph is an induced cycle of length at least 4. A chordal graph is a graph without chordless cycles.
Given two vectors and , we say if the inequality holds entry-wise. Additionally, if the inequality is strict on at least one entry, we say that .
Definition 1.4 (Maximum -gap to the Shearer’s bound).
Given a dependency graph and a probability vector beyond the Shearer’s bound of , define the maximum -gap from to the Shearer’s bound of as
For convenience, we let if is in the Shearer’s bound of .
Intuitively, measures how far is from the Shearer’s bound of . One can verify that if is in the Shearer’s bound, if is on the Shearer’s boundary, and if is beyond Shearer’s bound but not on the Shearer’s boundary. Now, we are ready to state our main result.
Theorem 1.5.
For any chordal graph , , i.e., iff .
For any graph which is not chordal, if
for some disjoint chordless cycles in . In particular, there is a probability vector with satisfying the above condition, where is the length the shortest chordless cycle. This implies that contains a probability vector with .
The intuition of Theorem 1.5 is as follows. The theorem characterizes the efficient region of the MT algorithm with . It shows that if is upper bounded by a non-negative quantity related to the chordless cycles in , then the MT algorithm is efficient. Since is the set of where , our criterion is at least as good as Shearer’s bound. Moreover, for each which is not chordal, our criterion is strictly better: there exists some with satisfying our criterion. Intuitively, Theorem 1.5 implies that chordless cycles in enhance the power of the MT algorithm.
We emphasize that Theorem 1.5 provides a complete answer to Problem 1: properly contains if and only if is not chordal.
1.1.2. A new constructive LLL for non-extremal instances
Given a set of events with dependency graph , is called extremal if all pairs of dependent events are mutually exclusive, and non-extremal otherwise. Kolipaka and Szegedy [45] showed that the MT algorithm is efficient up to the Shearer’s bound. In particular, Shearer’s bound is the tight convergence criterion for extremal instances [45, 27]. Here, we provide a new convergence criterion (Theorem 1.6) which is a strict improvement of Kolipaka and Szegedy’s result: this criterion is strictly better than Shearer’s bound when the instance is non-extremal, and becomes Shearer’s bound when the instance is extremal. This criterion, named intersection LLL, is the core of our proof of Theorem 1.5.
Let be a canonical dependency graph and be a probability vector. Let be a matching of , and be another probability vector. We say that an event set is of the setting , and write , if and for each pair . Given , define as follows:
Theorem 1.6 (intersection LLL (informal)).
For any , MT algorithm terminates quickly if is in the Shearer’s bound of .
The intuition of Theorem 1.6 is as follows. For any matching in , if the intersection of events on each edge in has a lower bound , then one can subtract from the probabilities of endpoints and , and the MT algorithm is guaranteed to be efficient whenever the reduced probability vector is in the Shearer’s bound.
Remark 1.7.
In many applications of LLL [49, 31, 30, 50, 28], the dependent bad events naturally intersect with each other. For instance, in a CNF formula, if the common variables in two clauses are both either positive or negative, then the bad events corresponding to these two clauses are dependent and intersect with each other. Thus our intersection LLL may be capable of improving bounds for these applications. However, currently the improvement is weak because only the intersections between the matched events are considered in Theorem 1.6.
Nevertheless, the primary motivation of this work is to explore the power of the MT algorithm in the general variable LLL setting. This basic problem is very important in itself, besides its potential applications.
1.1.3. Application to lattices
To illustrate the application of Theorem 1.5, we estimate the efficient region of the MT algorithm on some lattices explicitly. For simplicity, we focus on symmetric probabilities, i.e., . Our lower bounds on the gaps between the efficient region of the MT algorithm and the Shearer’s bound are summarized in Table 1. For example, when the canonical dependency graph is the square lattice, the vector is on the Shearer’s boundary, and the MT algorithm is provably efficient whenever the probability of each event is at most .
1.2. Technique overview
As mentioned before, the Shearer’s bound is the tight criterion for MT algorithm on extremal instances. Thus in order to show that MT algorithm goes beyond Shearer’s bound, we need to take advantage of the intersection between dependent events. Specifically, Theorem 1.5 immediately follows from two results about non-extremal instances. One is the intersection LLL criterion (Theorem 1.6), which goes beyond Shearer’s bound whenever there are intersections between dependent events. The other result is a lower bound on the amount of intersection between dependent events for general instances (Theorem 4.1).
1.2.1. Proof overview of Theorem 1.6
Let us first remember Kolipaka and Szegedy’s argument [45], which shows that the MT algorithm is efficient up to the Shearer’s bound. We assume that is a fixed set of events with dependency graph and probabilities . The notion of a witness DAG333In the paper [45], the role of witness DAGs was played by “stable set sequences”, but the concepts are essentially the same: there is a natural bijection between stable set sequences and wdags. (abbreviated wdag) is central to their argument. A wdag is a DAG whose each node has a label from and in which two nodes and are connected by an arc if and only if or . With a resampling sequence (i.e., MT algorithm picks the events for resampling in this order), we associate a wdag on node set as follows: (a) and (b) there is an arc from to with if and only if either or (see an example in Figure 1). We say that a wdag occurs in the resampling sequence if there is subset of nodes in such that is a subgraph of induced by the nodes that have a directed path to (Figure 1 (d) is an example, where ). An useful observation is that . Here, denotes the set of all single-sink wdags (a.k.a. proper wdags) of .
We define the weight of a wdag to be . The crucial lemma in Kolipaka and Szegedy’s argument (the idea is from Moser-Tardos analysis) is that the probability of occurrence of a certain wdag is upper bounded by its weight. The idea is that we can assume (only for the analysis) that the MT algorithm has a preprocessing step where it prepares an infinite number of independent samples for each variable. These independent samples create a table , called the resampling table (see Figure 2 in Section 3.1 for an example). When the MT algorithm decides to resample variable , it picks a new sample of from the resampling table. Suppose a certain wdag occurs, then for each of its events we can determine a particular set of samples in the resampling table that must satisfy the event, where we say that is consistent with the resampling table and denote it by . Hence, .
Finally, they solved beautifully the summation of weights of proper wdags, i.e., , which turns out to converge if and only if is in the Shearer’s bound of .

Viewing Theorem 1.6 as an improvement of Kolipaka and Szegedy’s result, we begin by providing a tighter upper bound on when the instance is non-extremal (Theorem 3.7). First, note that for each wdag , there exist selection rules to make , so it is impossible to give a better upper bound on which holds for all selection rules. Our idea is to group proper wdags, and consider the sum of over a group. For example, suppose that and are dependent and . Let denote the proper wdag which consists of only one arc , and denote the proper wdag consisting of only . and cannot both occur, but they may be both consistent with a given resampling table. So the total weights of and is an overestimate of the probability that or occurs. Formally,
where the last inequality is according to the Cauchy–Schwarz inequality (see Proposition 3.3). Importantly, the upper bound holds for all selection rules.
It is crucial as well as the difficulty that our improvement over the weight of wdags should be “exponential”: since the quantity converges if and only if is in the Shearer’s bound, constant factor or even sub-exponential improvements over do not help to show the desired convergence criterion. Our exponential improvement relies on a delicate grouping and a tricky random partition of the union of across wdags.
We first state how we group proper wdags: define to be the set of proper wdags whose unique sink node is labelled with and in which there are exactly nodes labelled with . Noticing that at most one wdag in can occur, we have that
Now, we partition the space across wdags in . The notions of reversible arcs (see Definition 2.4) and a auxiliary table (see Section 3.1) are two central concepts here. Specifically, an arc in a wdag is said reversible, if the directed graph obtained from by reversing the direction of is also a wdag. The auxiliary table is a table of independent fair coins corresponding to directions of reversible arcs. We say a wdag is consistent with , denoted by if (i) ; and (ii) for each reversible arc whose direction is not consistent with , the wdag obtained by reversing the arc is not consistent with . The crucial lemma (Lemma 3.1) shows that for any certain assignment of the auxiliary table , . The point is that ’s have much less overlap with each other so that they can be viewed as a “approximate” partition of the space. By applying union bound, we get
Then we are able to provide an upper bound on which is “exponentially” smaller than (Lemma 3.4), and then complete the proof of Theorem 3.7.
The next step is to show that the tighter upper bound converges when is in the Shearer’s bound. For each vertex in the matching , we “split” vertex into two new connected vertices and . Let be the resulted dependency graph (see an example in Figure 3). Define and (see the definition of in Section 2.3). One can see that and are essentially the same: suppose , then for each , we view as the union of two mutually exclusive events and whose probabilities are and respectively. Such a representation of is of the setting . Thus, the sum of weights of proper wdags in the setting is equal to that in the setting (Proposition 3.9). So it suffices to show that our tighter upper bound is upper bounded by the sum of weights of proper wdags in the setting (Theorem 3.13). Our idea is to construct a mapping which maps each to a subset of and satisfies that:
-
(a)
distinct proper wdags of are mapped to disjoint subsets of ; and
-
(b)
for each , the bound in Lemma 3.4 is upper bounded by the sum of weights of proper wdags over the subset that is mapped to.
We present such a mapping in Definition 3.11. Conditions (a) and (b) are verified in Theorem 3.12 and Theorem 3.13 respectively.
The idea of constructing a mapping between wdags of two dependency graphs may be of independent interest, and may be applied elsewhere when we wish to show some properties about Shearer’s bound.
1.2.2. Proof overview of Theorem 4.1
The proof of Theorem 4.1 mainly consists of two parts. First, we show that there is an elementary event set which approximately achieves the minimum amount of the intersection between dependent events (Lemma 4.2). Here, we call an event elementary, if there is a subset of the domain of variable for each variable in such that happens if and only if for all variables in . We call a set of events elementary if every is elementary. Then, for elementary event sets, by applying AM-GM inequality, we obtain a lower bound on the total amount of overlap on common variables, which further implies a lower bound on the amount of intersection between dependent events (Lemma 4.5).
1.3. Related works
Beck proposed the first constructive LLL, which provides efficient algorithms for finding the perfect object avoiding all “bad” events [7]. His methods were refined and improved by a long line of research [6, 53, 13, 39]. In a groundbreaking work, Moser and Tardos proposed a new algorithm, i.e., Algorithm 1, and proved that it finds such a perfect object under the condition in Theorem 1.1 in the variable setting [54]. Pegden [55] proved that the MT algorithm efficiently converges even under the condition of the cluster expansion local lemma [9]. Kolipaka and Szegedy [45] pushed the efficient region to Shearer’s bound. The phenomenon that the MT algorithm can still be efficient beyond Shearer’s bound was known to exist for sporadic and toy examples [32]. However, such result employs the special structures in the examples and only applies to some restricted variable-generated event systems . By contrast, the results in this work applies to all variable-generated event systems.
Besides the line of research exploring the efficient region of the MT algorithm, there is a large amount of effort devoted to derandomizing or parallelizing the MT algorithm [54, 11, 34, 8, 23, 12, 35, 33] and to extending the Moser-Tardos techniques beyond the variable setting [38, 2, 41, 5, 3, 52, 42, 4].
There is a line of works studying the gap between non-constructive VLLL and Shearer’s bound [45, 36, 24, 37]. Kolipaka and Szegedy [45] obtained the first example of gap existence where the canonical dependency graph is a cycle of length 4. The paper [36] showed that Shearer’s bound is not tight for VLLL. More precisely, Shearer’s bound is tight for non-constructive VLLL if and only if the canonical dependency graph is chordal. The first paper to study quantitatively the gaps systematically is [37], which provides lower bounds on the gap when the canonical dependency graph containing many chordless cycles.
Erdös and Spencer [15] introduced the lopsided-LLL, which extends the results in [14] to lopsidependency graphs. Lopsided LLL has many interesting applications in combinatorics and theoretical computer science, such as the -SAT [31], random permutations [47], Hamiltonian cycles [1], and matchings on the complete graph [48]. Shearer’s bound is also the tight condition for the lopsided LLL [56].
LLL has a strong connection to sampling. Guo, Jerrum and Liu [27] proved that the MT algorithm indeed uniformly samples a perfect object if the instance is extremal. For extremal instances, they developed an algorithm called “partial rejection sampling” which resamples in a parallel fashion, since the occurring bad events form an independent set in the dependency graph. Actually, a series of sampling algorithms for specific problems are the parallel resampling algorithm running in the extremal case [27, 26, 22, 25]. In a celebrated work, Moitra [51] introduced a novel approach that utilizes LLL to sample -CNF solutions. This approach was then extended by several works [29, 21, 16, 17, 43, 44].
1.4. Organization of the paper.
In Section 2, we recall and introduce some definitions and notations. In Section 3, we prove Theorem 1.6. Section 4 is about the proof of Theorem 4.1, which gives a lower bound on the amount of the intersection between dependent events. In Section 5, we prove Theorem 1.5. In Section 6, we provide a explicit lower bound for the gaps between the efficient region of MT algorithm and Shearer’s bound on periodic Euclidean graphs.
2. Preliminaries
Let denote the set of non-negative integers. Let denote the set positive integers. For , we define . Throughout this section, we fix a canonical dependency graph .
2.1. Witness DAG
If for a given run, MT algorithm picks the events for resampling in this order, we say that is a resample sequence. If the algorithm never finishes, the resample sequence is infinite, and in this case we set .
Definition 2.1 (Witness DAG).
We define a witness DAG (abbreviated wdag) of to be a DAG , in which each node has a label from , and which satisfies the additional condition that for all distinct nodes there is an arc between and (in either direction) if and only if or .
We say is a proper wdag (abbreviated pwdag) if has only one sink node. Let denote the set of pwdags of .
Given a resampling sequence , we associate a wdag on the node set such that (i) and (ii) with is as an arc of if and only if either or . See Figure 1 for an example of .
Given a wdag and a set of nodes of , we define to be the induced subgraph on all nodes which has a directed path to some . Note that is also a wdag. We say that is a prefix of , denoted by , if for some node set .
Definition 2.2.
We say a wdag occurs in a resampling sequence if . Let be the indicator variable of the event that occurs in .
Similar to Lemma 12 in [45], we have that . For and , define to be the set of pwdags whose unique sink node is labelled with and in which there are exactly nodes labelled with . Let be the indicator variable of the event that there is a occurring in . It is easy to see that only one pwdag in can occur in . Thus , which further implies that
Fact 2.3.
.
2.2. Reversible arc
In the rest of this section, we fix a matching of . Given , with a slight abuse of notation, we sometimes say if there is some such that .
Definition 2.4 (Reversibility).
We say that an arc is reversible in a wdag if the directed graph obtained from by reversing the direction of the arc is still a DAG.
Furthermore, we say that is -reversible in if is reversible in and .
By definition, we have the following two observations.
Fact 2.5.
is reversible in if and only if it is the unique path from to in .
Fact 2.6.
If is reversible in a wdag of , then the directed graph obtained from by reversing the direction of is also a wdag of .
Given a pwdag , define
to be the set of nodes participating in reversible arcs, and . For , define .
2.3. Other notations
Let and be two probability vectors. Recall that is defined as
(1) |
For each where for some , define
Fact 2.7.
for each .
3. Proof of Theorem 1.6
The proof of Theorem 1.6 consists of two parts. First, we provide a tighter upper bound on the complexity of MT algorithm (Section 3.1). Then, we show that the tighter upper bound converges if is in the Shearer’s bound of (Section 3.2).
3.1. A tighter upper bound on the complexity of MT algorithm
In this subsection, we prove Theorem 3.7, which follows from Lemma 3.1 and Lemma 3.4 immediately. We first recall and introduce some concepts and notations.
Resampling Table. One key analytical technique of Moser and Tardos [54] is to precompute the randomness in a resampling table . Specifically, we can assume (only for the analysis) that MT algorithm has a preprocessing step where it draws an infinite number of independent samples for each variable . These independent samples create a table , called the resampling table (see Figure 2). MT algorithm takes that first column as the initial assignments of . Then, when is to be resampled, MT algorithm goes right in the row corresponding to and picks the sample.
Consistency with the resampling table. For a wdag , a node , and a variable , we define
Moreover, let . We say that is consistent with , denoted by , if for each node in , the event holds on . Intuitively, suppose occurs, then are the assignments of just before the time that the MT algorithm picks the event corresponding to to resample, hence must hold on . We sometimes use and instead of and respectively if is clear from the context. Besides, we use to denote that there is some such that .


Auxiliary Table. We introduce another central concept in the proof of Theorem 3.7, called the auxiliary table, which is a table of independent fair coins. Specifically, for each pair , we draw an infinite number of independent fair coins , where . These independent coins form the auxiliary table (see Figure 2). The auxiliary table is used to encode directions of -reversible arcs, according to which we partition the space .
Consistency with the resampling table and the auxiliary table. We need some notations about reversible arcs. Suppose has a unique sink node and is reversible in . Let be the DAG obtained from by reversing the direction of . We define . In other words, is the prefix of with a unique sink node . Given and a pwdag , let denote the sequence listing all nodes in with labels or in a topological order of 444It is easy to see that is well defined. That is, all topological orderings of induce the same .. Given a node in , if 555Because is a matching, there is at most one such ., we define
to be the order of in . For simplicity of notations, we will use instead of if is clear from the context.
Given a wdag , we say an -reversible arc is inconsistent with the auxiliary table if . We say is consistent with , denoted by , if (i) and (ii) for any -reversible arc inconsistent with , . We say if there is some such that .
The intuition of the notion “consistency” is as follows. Suppose in a -reversible arc in , and both and are consistent with the resampling table. But and cannot both occur. It is according to the auxiliary table to which one of and we assign .
Lemma 3.1.
For each and , .
Proof.
Fix an arbitrary assignment of and an arbitrary assignment of . Suppose , i.e, such that . We will show that there must exist some such that . This will imply the conclusion immediately.
We apply the following procedure to find such a pwdag .
By induction on , it is easy to check that and for each . Furthermore, if the procedure terminates, then in the final wdag , for every -reversible arc inconsistent with , we have that . So . In the following, we will show that the procedure always terminates, which finishes the proof.
Note that each has no more nodes than and that there are finite number of wdags in with no more nodes than , so it suffices to prove that each wdag appears at most once in the procedure.
By contradiction, assume for some . Recall that is reversible in and inconsistent with . So .
Let be the last wdag in such that . Observing that , we have such must exist. By , we have , . Therefore, . Combining with that is the inconsistent arc in which is reversed in , we have , and . Thus we have and . Note that . Combining with , we have . Combining with , we have . This is contradicted with .
∎
The following two propositions will be used in the proof of Lemma 3.4. The first proposition is an easy observation, and the second one is a direct application of the Cauchy-Schwarz inequality. For the sake of completeness, we present their proof in the appendix.
Proposition 3.2.
Given any wdag , there exists a set of disjoint -reversible arcs666We say two arc and are disjoint if their node sets are disjoint, i.e. . such that: for each ,
Proposition 3.3.
Suppose and are three independent random variables, is an event determined by , and is an event determined by . Let be four independent samples of , respectively. Then the following holds with probability at most :
-
•
is true on , is true on , and
-
•
either is false on or is false on .
Now, we are ready to show Lemma 3.4.
Lemma 3.4.
For each pwdag ,
Proof.
Let be the set of disjoint -reversible arcs defined in Proposition 3.2. Let denote the set of nodes which appears in , and consists of the other nodes. Proposition 3.2 says that for each ,
For each , let denote the event that holds on . It is easy to see that . Besides,
Claim 3.5.
If , then holds for each .
Proof.
Note that are the assignments of just before the time that the MT algorithm picks the event corresponding to to resample. MT algorithm decides to pick only if holds. Hence must hold on . ∎
Let be an arc in , where and . Then by the definition of , we have is reversible in . Let be the wdag obtained by reversing the direction of in . Recalling the definition of , one can verify that
and
For simplicity, let . We define to be the event that the following hold:
-
(a)
holds on , and holds on ;
-
(b)
If , then either is false on or is false on .
Conditioned on that , happens with probability . Condition on that , by using Proposition 3.3, happens with probability at most . Thus,
Claim 3.6.
If , then holds for each in .
Proof.
Suppose . Similar to the argument in Claim 3.5, we can see that Item (a) holds. In the following, we show Item (b) holds.
By contradiction, assume , holds on , and holds on . Then, we have in is inconsistent with and . Thus, since is a prefix of . By definition, we have , a contradiction. ∎
Since the events and depend on distinct entries of and , they are mutually independent. Therefore,
∎
Now we are ready to prove the main theorem of this subsection.
Theorem 3.7.
.
3.2. Mapping between wdags
In this section, we will prove Theorem 3.13, which provides a upper bound of in terms of .
Definition 3.8 (Homomorphic dependency graph).
Given a dependency graph and a matching of , we define a graph homomorphic to respected to as follows.
-
•
;
-
•
, each pair of vertices in are connected in .
Besides, we associate a probability vector with as follows:

In fact, and are essentially the same: suppose , then for each , we view as the union of two mutually exclusive events whose probabilities are and respectively. Such a representation of is of the setting .
We have the following proposition, whose proof can be found in the appendix.
Proposition 3.9.
Given a pwdag , recall that is the set of nodes of -reversible arcs in . Define to be the set of nodes in where is contained in an edge in . Obviously, . For simplicity of notations, we will omit from the notations if is clear from the context.
Given a pwdag , we use to represent a partition of where (some of these four sets are possibly empty). Let denote the set consisting of all such partitions. The formal definition is as follows.
Definition 3.10 (Partition).
Given a pwdag of , define
Given a wdag , there may be two or more topological ordering of . We fix an arbitrary topological ordering, and denote it by . In the following, we define an injection from to .
Definition 3.11.
Given a pwdag and , define to be a directed graph constructed as follows.
Constructing . where and . For convenience of presentation, we fix two bijections and to name nodes in . In order to distinguish between nodes in and those in , we will always use to represent the nodes of and to present the nodes of . Given , we use to denote the unique node such that (if ) or (if ).
Description of . For each node , where ,
(2) |
For each node , assuming is the node such that and is the node such that ,
(3) |
Constructing . where and
Theorem 3.12.
is an injection from to .
The proof of Theorem 3.12 is in the appendix. Now we can prove the main theorem of this subsection.
Theorem 3.13.
Proof.
Given and , let . For each in where for some , according to the definition of , (2), and (3), we have that
-
•
if , then ;
-
•
if , then ;
-
•
if , then ;
-
•
if , then .
Moreover, for each , we have . Thus, for each ,
So
where the third equality is according to the definition of . Finally,
where the second inequality is due to Theorem 3.12 and the equality is by Proposition 3.9. ∎
3.3. Putting all things together
The following lemma is implicitly proved in [45].
Lemma 3.14 ([45]).
For any undirected graph and probability vector , .
Theorem 1.6 (restated).
For any , if , then the expected number of resampling steps performed by MT algorithm is most , where is the number of events in .
4. Lower bound on the amount of intersection
In order to explore how far beyond Shearer’s bound MT algorithm is still efficient in general, we provide a lower bound on the amount of intersection between dependent events for general instances (Theorem 4.1).
We first introduce some notations. Given a bipartite graph , we call the vertex left vertex and the vertex right vertex. We call linear777The notion is not arbitrary. The bipartite graph can be represented by a hypergraph in a natural way: each right vertex is represented by a node in the hypergraph, each left vertex is represented by a hyperedge , and is in if and only if . A hypergraph is called linear if any two hyperedges share at most one node. if any two left vertices in share at most one common neighbor in . Let denote the maximum degree of , and denote the maximum degree of the left vertices in . If is clear from the context, we may omit from these notations. In addition, for a bipartite graph and a probability vector , we define888It is possible that .
and .
We use to denote that (i) is an event-variable graph of and (ii) the probability vector of is . Let be a matching of , and be another probability vector. We say that an event set is of the setting , and write , if and for each pair .
We call an event elementary, if can be written as where are subsets of the domains of variables. We call an event set elementary if all events in are elementary.
Theorem 4.1.
Let be a bipartite graph, be a probability vector, and be a collection of disjoint subsets of . For each , let denote the induced subgraph on and denote the edge set of . If all ’s are linear, then the following holds.
If , then there is a matching of satisfying that for any .
The proof of Theorem 4.1 mainly consists of two parts. First, we show that there is an elementary event set which approximately achieves the minimum amount of intersection between dependent events (Lemma 4.2). Then, for elementary event sets, by applying AM-GM inequality, we obtain a lower bound on the total amount of overlap on common variables, which further implies a lower bound on the amount of intersection between dependent events (Lemma 4.5).
Lemma 4.2.
Let be a linear bipartite graph, be the edge set of , and is a probability vector. Let denote the minimum among all event sets . Then there is an elementary event set such that .
Proof.
For simplicity, we let . Without loss of generality, we assume that each random variable is uniformly distributed over . Let be an event set where . We will replace with an elementary one by one for each , so that the resulted event set satisfies .
More precisely, fix and suppose have been replaced with elementary events respectively. For simplicity of notations, for any pair , we abbreviate , and to , and respectively. Without loss of generality, we assume depends on variables . For every , we define
for . Without loss of generality, we assume is non-decreasing. Let be the indicator of , then
For each , let
Noticing that is linear (i.e., any two events share at most one common variable), we have
(4) |
Let be an elementary event such that it happens if and only if . Here is a set of positive real numbers satisfying that
-
(i)
. That is, ;
-
(ii)
.
Claim 4.3.
Such exists. Thus so does .
Proof.
We prove a generalized statement in which can be required to be an arbitrary number in . Our proof is by induction on . The base case when is trivial. Now we assume that for any preset , there exist satisfying that
-
(i)
and
-
(ii)
.
Let denote the minimum among all such ’s. It is easy to see that and is continuous and non-increasing.
Fix an arbitrary . We define for . Obviously, and is continuous and non-decreasing. So there must exist a such that . Then let be a set of positive real numbers where
-
(i)
and
-
(ii)
.
Let . It is obvious that and . This completes the induction step. ∎
Claim 4.4.
For every , we have
Proof.
Let , and denote the indicator functions of the events , , and respectively. Since ,
Fix , then
Since is non-decreasing and , we have
According to Equation 4,
This completes the proof. ∎
Lemma 4.5.
Let be a linear bipartite graph and be a probability vector. Then for any elementary ,
where is the edge set of
Proof.
For simplicity of notation, we let stand for . Without loss of generality, we assume that each variable is uniformly distributed over . As is elementary, each can be written as where . Let be the Lebesgue measure.
The following lemma is a special case of Theorem 4.1 where and . In fact, Theorem 4.1 is proved by applying Lemma 4.6 to each separately.
Lemma 4.6.
Let be a linear bipartite graph and be a probability vector. If , then for some matching of and some satisfying that .
Proof.
Given an instance , we construct such a greedily as follows.
We maintain two sets and , which are initialized as and respectively. We do the following iteratively until becomes empty: select a edge with maximum from , add to , and delete all edges connecting or from (including ).
Let and denote and respectively. In each iteration, at most edges are deleted from and for each deleted edge , . Based on this observation, it is easy to see that
(10) |
Moreover, according to Lemma 4.2 and 4.5, it has that
(11) |
By combining Inequality 10 and 11, setting , and noting , we finish the proof. ∎
Proof of Theorem 4.1.
For each , by applying Lemma 4.6 to , we have that for some matching and some where . Note that ’s are disjoint with each other, so is still a matching. By letting and , we conclude the theorem. ∎
Remark 4.7.
Given a bipartite graph , its simplified graph is defined to be obtained from by deleting all the right nodes which only have one neighbor and combining all the right nodes with the same neighbor set. Notice that if is linear, so is its simplified graph.
Theorem 4.1 can be slightly generalized: it is sufficient that the simplified graph of instead of itself is linear.
5. The Moser-Tardos algorithm is beyond Shearer’s bound
In this section, we prove Theorem 1.5. Given a dependency graph , a vector and a chordless cycle in , define
and
Lemma 5.1.
Given , and , let be any disjoint chordless cycles in . If
then for any variable-generated event system , the expected number of resampling steps performed by MT algorithm is most .
Proof.
Fix such an instance . Define . Let denote the event-variable graph of . Let denote the induced subgraph of on . According to Remark 4.7, it is lossless to assume is a cycle of length . Thus we have
(12) |
According to Theorem 4.1, there is a matching of such that . Define as (1). We have and
Combining with (12), we have
where the last inequality is by the condition of the lemma. Thus by Definition 1.4, we have is in the Shearer’s bound of . Combining with Theorem 1.6, we have the expected number of resampling steps performed by the Moser-Tardos algorithm is most . ∎
Lemma 5.2.
Given and any chordless cycle in , there is some probability vector beyond the Shearer’s bound of and with
such that for any variable-generated event system , the expected number of resampling steps performed by MT algorithm is most .
The following two lemmas will be used in the proof of Lemma 5.2.
Lemma 5.3.
[56] holds for any extremal instance .
Lemma 5.4.
Proof of Lemma 5.2.
Let and . Let be an extremal instance defined as follows: is a variable-generated event system fully determined a set of underlying mutually independent random variables . Moreover, for each , and . According to Lemma 5.3,
Besides, according to Lemma 5.4, for any in the Shearer’s bound of ,
Thus, for any , we have that
Hence is in the Shearer’s bound of . Thus, there exists such that defined as follows is on the Shearer’s boundary of :
One can verify that
(13) |
Define
One can verify that because and . Moreover, let be large enough such that . One can verify that such must exist. We have . This is because otherwise and then
By following the proof of Lemma 5.1, we have the MT algorithm terminates at , which is contradictory with .
Moreover, is a continuous function of , because and are both continuous functions of . Combining with and , we have there must be a such that . Let . By , we have
(14) |
Combining with and (13), we have .
Fix a variable-generated event system . Define . Let denote the event-variable graph of . Let denote the induced subgraph of on . According to Remark 4.7, it is lossless to assume that is a cycle of length . Thus we have
(15) |
According to Theorem 4.1, there is a matching of such that . Define as (1). We have
Combining with (15), we have
Let
By (13) we have
Thus we have
where the last inequality is by (14). Thus by Definition 1.4, we have is in the Shearer’s bound of . Combining with Theorem 1.6, we have the expected number of resampling steps performed by the MT algorithm is most . ∎
6. Application to periodic Euclidean graphs
In this section, we explicitly calculate the gaps between our new criterion and Shearer’s bound on periodic Euclidean graphs, including several lattices that have been studied extensively in physics. It turns out the efficient region of MT algorithm can exceed significantly beyond Shearer’s bound.
A periodic Euclidean graph is a graph that is embedded into a Euclidean space naturally and has a translational unit in the sense that can be viewed as the union of periodic translations of . For example, a cycle of length 4 is a translational unit of the square lattice.
Given a dependency graph , it naturally defines a bipartite graph as follows. Regard each edge of as a variable and each vertex as an event. An event depends on a variable if and only if the vertex corresponding to is an endpoint of the edge corresponding to .
For simplicity, we only focus on symmetric probabilities, where . Given a dependency graph and a vector , remember that is on Shearer’s boundary of if is in Shearer’s bound and is not for any .
Given a dependency graph and two vertices , we use to denote the distance between and in . The following Lemma will be used.
Lemma 6.1.
Suppose is on Shearer’s boundary of . For any probability vector other than , it is in the Shearer’s bound if there exist , where , and such that the following conditions hold:
-
(a)
for each , there are at most subsets such that ;
-
(b)
if , then for each and ;
-
(c)
if , then
While Lemma 6.1 looks involved, the basic idea is simple: by contradiction, suppose there is such a vector beyond Shearer’s bound; then we apply Lemma D.1 repeatedly to transfer probability from one event to another while keeping the probability vector still beyond Shearer’s bound; finally, the vector will be changed to a vector strictly below , which makes a contradiction to the assumption that is on the Shearer’s boundary. The involved part is a transferring scheme which changes to another probability vector strictly below . We leave the proof to the appendix.
The main result of this section is as follows.
Theorem 6.2.
Let be a periodic Euclidean graph with maximum degree , and be the probability vector on Shearer’s boundary of . Suppose is a translational unit of with diameter . Let
Then for any where , the expected number of resampling steps performed by the MT algorithm is most .
Proof.
Fix any where . Let denote for . We construct a matching greedily as follows: we maintain two sets and , which are initialized as and respectively. We do the following iteratively until becomes empty: select a edge with maximum from , add to , and delete all edges connecting or from (including ). Let . Then .
Define as (1). In the remaining part of the proof, we will show that is in the Shearer’s bound. This implies the conclusion immediately by Theorem 1.6.
In fact, it is a direct application of Lemma 6.1 to show that is in the Shearer’s bound. To provide more detail, we need some notations. We use to represent vertices in , and use to represent vertices in . Let be the periodic translations of in . And we use a surjection999 is possibly not a injection, as these translations are possibly overlapped with each other. to represent how these periodic translations constitute : if the copy of in -th translation (i.e., ) is . In particular, the vertex set of , denoted by , is , and the edge set of , denoted by , is . Besides, let for . For , let . Let stand for the pairs in adjacent to . With some abuse of notation, we sometimes use to denote that for some .
The following claim says that is much smaller than even projected on a single translation. Its proof uses a similar idea to Theorem 4.6 and can be found in the appendix.
Claim 6.3.
holds for any .
We apply Theorem 6.2 to three lattices: square lattice, Hexagonal lattice, and simple cubic lattice. For square lattice, we take the square with 25 vertices as the translational unit. For Hexagonal lattice, we take a graph consisting of hexagons as the translational unit, in which there are 3,4,5,4,3 hexagons in the five columns, respectively. For simple cubic lattice, we take the cube with 27 vertices as the translational unit. The explicit gaps are summarized in Table 1. Finally, the lower bounds for these three lattices in Table 1 hold for all bipartite graphs with the given canonical dependency graph, because all such bipartite graphs are essentially the same under the reduction rules defined in [36].
References
- AFR [95] Michael Albert, Alan Frieze, and Bruce Reed. Multicoloured hamilton cycles. the electronic journal of combinatorics, 2(1):R10, 1995.
- AI [16] Dimitris Achlioptas and Fotis Iliopoulos. Random walks that find perfect objects and the lovász local lemma. Journal of ACM, 63(3):22:1–22:29, 2016.
- AIK [19] Dimitris Achlioptas, Fotis Iliopoulos, and Vladimir Kolmogorov. A local lemma for focused stochastic algorithms. SIAM Journal on Computing, 48(5):1583–1602, 2019.
- AIS [19] Dimitris Achlioptas, Fotis Iliopoulos, and Alistair Sinclair. Beyond the lovász local lemma: Point to set correlations and their algorithmic applications. In Proceedings of Symposium on Foundations of Computer Science (FOCS), pages 725–744, 2019.
- AIV [17] Dimitris Achlioptas, Fotis Iliopoulos, and Nikos Vlassis. Stochastic control via entropy compression. In Proceedings of International Colloquium on Automata, Languages, and Programming (ICALP), volume 80, pages 83:1–83:13, 2017.
- Alo [91] Noga Alon. A parallel algorithmic version of the local lemma. Random Structures and Algorithms, 2(4):367–378, 1991.
- Bec [91] József Beck. An algorithmic approach to the Lovász local lemma. Random Structures and Algorithms, 2(4):343–365, 1991.
- BFH+ [16] Sebastian Brandt, Orr Fischer, Juho Hirvonen, Barbara Keller, Tuomo Lempiäinen, Joel Rybicki, Jukka Suomela, and Jara Uitto. A lower bound for the distributed lovász local lemma. In Proceedings of Symposium on Theory of Computing (STOC), pages 479–488, 2016.
- BFPS [11] Rodrigo Bissacot, Roberto Fernández, Aldo Procacci, and Benedetto Scoppola. An improvement of the Lovász local lemma via cluster expansion. Combinatorics, Probability and Computing, 20(05):709–719, 2011.
- CCS+ [17] Jan Dean Catarata, Scott Corbett, Harry Stern, Mario Szegedy, Tomas Vyskocil, and Zheng Zhang. The Moser-Tardos resample algorithm: Where is the limit?(an experimental inquiry). In 2017 Proceedings of the Ninteenth Workshop on Algorithm Engineering and Experiments (ALENEX), pages 159–171, 2017.
- CGH [13] Karthekeyan Chandrasekaran, Navin Goyal, and Bernhard Haeupler. Deterministic algorithms for the lovász local lemma. SIAM Journal on Computing, 42(6):2132–2155, 2013.
- CPS [17] Kai-Min Chung, Seth Pettie, and Hsin-Hao Su. Distributed algorithms for the lovász local lemma and graph coloring. Distributed Computing, 30(4):261–280, 2017.
- CS [00] Artur Czumaj and Christian Scheideler. Coloring nonuniform hypergraphs: A new algorithmic approach to the general Lovász local lemma. Random Structures and Algorithms, 17(3-4):213–237, 2000.
- EL [75] Paul Erdős and László Lovász. Problems and results on 3-chromatic hypergraphs and some related questions. Infinite and finite sets, 10(2):609–627, 1975.
- ES [91] Paul Erdös and Joel Spencer. Lopsided lovász local lemma and latin transversals. Discrete Applied Mathematics, 30(2-3):151–154, 1991.
- FGYZ [20] Weiming Feng, Heng Guo, Yitong Yin, and Chihao Zhang. Fast sampling and counting -sat solutions in the local lemma regime. In Proceedings of Symposium on Theory of Computing (STOC), pages 854–867, 2020.
- FHY [20] Weiming Feng, Kun He, and Yitong Yin. Sampling constraint satisfaction solutions in the local lemma regime. arXiv preprint arXiv:2011.03915, 2020.
- FHY [21] Weiming Feng, Kun He, and Yitong Yin. Sampling constraint satisfaction solutions in the local lemma regime. In Proceedings of the 53rd Annual ACM SIGACT Symposium on Theory of Computing, pages 1565–1578, 2021.
- Gau [67] David S Gaunt. Hard-sphere lattice gases. ii. plane-triangular and three-dimensional lattices. The Journal of Chemical Physics, 46(8):3237–3259, 1967.
- GF [65] David S Gaunt and Michael E Fisher. Hard-sphere lattice gases. i. plane-square lattice. The Journal of Chemical Physics, 43(8):2840–2863, 1965.
- GGGY [20] Andreas Galanis, Leslie Ann Goldberg, Heng Guo, and Kuan Yang. Counting solutions to random CNF formulas. In ICALP, volume 168 of LIPIcs, pages 53:1–53:14, 2020.
- GH [20] Heng Guo and Kun He. Tight bounds for popping algorithms. Random Structures & Algorithms, 2020.
- Gha [16] Mohsen Ghaffari. An improved distributed algorithm for maximal independent set. In Robert Krauthgamer, editor, Proceedings of ACM-SIAM Symposium on Discrete Algorithms (SODA), pages 270–277, 2016.
- Gil [19] András Gilyén. Quantum singular value transformation & its algorithmic applications. PhD thesis, University of Amsterdam, 2019.
- GJ [18] Heng Guo and Mark Jerrum. Approximately counting bases of bicircular matroids. arXiv preprint arXiv:1808.09548, 2018.
- GJ [19] Heng Guo and Mark Jerrum. A polynomial-time approximation algorithm for all-terminal network reliability. SIAM Journal on Computing, 48(3):964–978, 2019.
- GJL [19] Heng Guo, Mark Jerrum, and Jingcheng Liu. Uniform sampling through the lovász local lemma. Journal of the ACM (JACM), 66(3):18, 2019.
- GKPT [17] Ioannis Giotis, Lefteris Kirousis, Kostas I Psaromiligkos, and Dimitrios M Thilikos. Acyclic edge coloring through the Lovász local lemma. Theoretical Computer Science, 665:40–50, 2017.
- GLLZ [19] Heng Guo, Chao Liao, Pinyan Lu, and Chihao Zhang. Counting hypergraph colorings in the local lemma regime. SIAM Journal on Computing, 48(4):1397–1424, 2019.
- GMSW [09] Heidi Gebauer, Robin A Moser, Dominik Scheder, and Emo Welzl. The lovász local lemma and satisfiability. In Efficient Algorithms, pages 30–54. Springer, 2009.
- GST [16] Heidi Gebauer, Tibor Szabó, and Gábor Tardos. The local lemma is asymptotically tight for SAT. Journal of ACM, 63(5):43, 2016.
- Har [16] David G Harris. Lopsidependency in the moser-tardos framework: beyond the lopsided Lovász local lemma. ACM Transactions on Algorithms (TALG), 13(1):17, 2016.
- Har [18] David G. Harris. Deterministic parallel algorithms for fooling polylogarithmic juntas and the lovász local lemma. ACM Transaction on Algorithms, 14(4):47:1–47:24, 2018.
- Har [19] David G. Harris. Deterministic algorithms for the lovasz local lemma: simpler, more general, and more parallel. CoRR, abs/1909.08065, 2019.
- HH [17] Bernhard Haeupler and David G. Harris. Parallel algorithms and concentration bounds for the lovász local lemma via witness dags. ACM Transaction on Algorithms, 13(4):53:1–53:25, 2017.
- HLL+ [17] Kun He, Liang Li, Xingwu Liu, Yuyi Wang, and Mingji Xia. Variable-version Lovász local lemma: Beyond shearer’s bound. In Proceedings of Symposium on Foundations of Computer Science (FOCS), pages 451–462, 2017.
- HLSZ [19] Kun He, Qian Li, Xiaoming Sun, and Jiapeng Zhang. Quantum lovász local lemma: Shearer’s bound is tight. In Proceedings of Symposium on Theory of Computing (STOC), pages 461–472, 2019.
- HS [14] David G. Harris and Aravind Srinivasan. A constructive algorithm for the lovász local lemma on permutations. In Proceedings of ACM-SIAM Symposium on Discrete Algorithms (SODA), pages 907–925, 2014.
- HSS [11] Bernhard Haeupler, Barna Saha, and Aravind Srinivasan. New constructive aspects of the lovász local lemma. Journal of ACM, 58(6):1–28, 2011.
- HSW [21] Kun He, Xiaoming Sun, and Kewen Wu. Perfect sampling for (atomic) lov’asz local lemma. arXiv preprint arXiv:2107.03932, 2021.
- HV [20] Nicholas J. A. Harvey and Jan Vondrák. An algorithmic proof of the lovász local lemma via resampling oracles. SIAM Journal on Computing, 49(2):394–428, 2020.
- IS [20] Fotis Iliopoulos and Alistair Sinclair. Efficiently list-edge coloring multigraphs asymptotically optimally. In Proceedings of ACM-SIAM Symposium on Discrete Algorithms (SODA), pages 2319–2336, 2020.
- JPV [20] Vishesh Jain, Huy Tuan Pham, and Thuy Duong Vuong. Towards the sampling lovász local lemma. CoRR, abs/2011.12196, 2020.
- JPV [21] Vishesh Jain, Huy Tuan Pham, and Thuy Duong Vuong. On the sampling lovász local lemma for atomic constraint satisfaction problems. CoRR, abs/2102.08342, 2021.
- KS [11] Kashyap Babu Rao Kolipaka and Mario Szegedy. Moser and tardos meet lovász. In Proceedings of Symposium on Theory of Computing (STOC), pages 235–244, 2011.
- KSX [12] Kashyap Babu Rao Kolipaka, Mario Szegedy, and Yixin Xu. A sharper local lemma with improved applications. In Proceedings of APPROX/RANDOM, volume 7408 of Lecture Notes in Computer Science, pages 603–614, 2012.
- LS [07] Linyuan Lu and László Székely. Using lovász local lemma in the space of random injections. the electronic journal of combinatorics, 14(1):R63, 2007.
- LS [09] Linyuan Lu and Laszlo A Szekely. A new asymptotic enumeration technique: the lovász local lemma. arXiv preprint arXiv:0905.3983, 2009.
- McD [97] Colin McDiarmid. Hypergraph colouring and the Lovász local lemma. Discrete Mathematics, 167:481–486, 1997.
- [50] Ankur Moitra. Approximate counting, the lovasz local lemma, and inference in graphical models. Journal of ACM, 66(2):10, 2019.
- [51] Ankur Moitra. Approximate counting, the Lovász local lemma, and inference in graphical models. J. ACM, 66(2):10:1–10:25, 2019.
- Mol [19] Michael Molloy. The list chromatic number of graphs with small clique number. Journal of Combinatorial Theory, Series B, 134:264–284, 2019.
- MR [98] Michael Molloy and Bruce Reed. Further algorithmic aspects of the local lemma. In Proceedings of Symposium on Theory of Computing (STOC), pages 524–529, 1998.
- MT [10] Robin A Moser and Gábor Tardos. A constructive proof of the general Lovász local lemma. Journal of ACM, 57(2):11, 2010.
- Peg [14] Wesley Pegden. An extension of the moser–tardos algorithmic local lemma. SIAM Journal on Discrete Mathematics, 28(2):911–917, 2014.
- She [85] James B. Shearer. On a problem of spencer. Combinatorica, 5(3):241–245, 1985.
- Spe [75] Joel Spencer. Ramsey’s theorem—a new lower bound. Journal of Combinatorial Theory, Series A, 18(1):108–115, 1975.
- Spe [77] Joel Spencer. Asymptotic lower bounds for Ramsey functions. Discrete Mathematics, 20:69–76, 1977.
- Sze [13] Mario Szegedy. The lovász local lemma–a survey. In International Computer Science Symposium in Russia, pages 1–11. Springer, 2013.
- Tod [99] Synge Todo. Transfer-matrix study of negative-fugacity singularity of hard-core lattice gas. International Journal of Modern Physics C, 10(04):517–529, 1999.
Appendix A Missing Proofs in Section 2
Proof of Proposition 3.2.
The following simple greedy procedure will find such a .
Obviously, for each , the procedure contains at least half of all reversible arcs where , hence at least half of nodes in . ∎
Appendix B Proof of Proposition 3.9
Given a pwdag of and a Boolean string , define to be a directed graph where , , and
It is easy to verify that is a pwdag of . Moreover, given any , there is one and only one and such that . In other words, is a bijection between and . So
where the second equality is by that , the forth equality is by the definition of , and the fifth equality is by the definition of .
Appendix C Proof of Theorem 3.12
We first verify that the image of is a subset of .
Lemma C.1.
For any and , .
Proof.
First, we prove that is a DAG. Define a total order over the set as follows: for any two distinct nodes ,
-
•
if , then in if and only if in ;
-
•
if , then in if and only if (and then ).
One can verify that is a topological order of , which means that is a DAG.
Secondly, we prove that is a wdag of . As has been shown to be a DAG, we only need to verify that: for any two distinct nodes in , there is a arc between and (in either direction) if and only if either or .
: By symmetry, suppose . If , then and for some vertex . Thus, by (2) and (3) we have and where . By , any two vertices in are connected in . In particular, . If , we have or immediately.
: Suppose are two distinct nodes where or . If , then either or in , which implies that either or . Otherwise, . Let . By (2) and (3), we have and . Therefore either or is in .
Finally, one can check that where is the unique sink of is the unique sink of . This completes the proof. ∎
In the rest of this section, we show that is injective. Given and , recall that is the sequence listing all nodes in labelled with or in the topological order. Similarly,
Definition C.2.
Given and , we use to denote the unique sequence listing all nodes in with label in in the topological order.
Claim C.3.
Suppose for some and . Let . Then for any node in ,
-
(a)
if and only if ;
-
(b)
for any other node in , if precedes in , then precedes in ;
-
(c)
if , then is next to in .
Proof.
Part (a) is immediate by Definition 3.11.
Now, we show Part (b). Suppose precedes in . Then in . Thus one can check that all the four arcs , , , and are contained in . In particular, as and . This implies that precedes in .
Finally, we prove Part (c). According to Part (b), and are adjacent in . Besides, as there is an arc in , we conclude that is next to in .
∎
Definition C.4.
For a reversible arc in , we call it -reversible in if and for some .
Claim C.5.
Suppose for some and . Let . Let be two nodes in where is next to . Then if and only if is -reversible in and .
Proof.
: Let . Assume , i.e., . By Definition 3.11, . According to Part (c) of Claim C.3, as is next to , we have and then .
Now we show that is -reversible. First, by Definition 3.11, either and , or and . What remains is to show is reversible, by Fact 2.5 which is equivalent to show that is the unique path from to in . By contradiction, assume that there is a path in where and . As , we have is not in and then should be in , which further implies that in . Similarly, we have in . So . Meanwhile, for each , if , then ; if , then in . So, it always holds that in for each . In particular, . A contradiction.
: Let and . Assume and , i.e., and . Furthermore, assume , then and . We will show that is not reversible.
Lemma C.6.
is injective.
Proof.
Fix a and a . Let denote . We show can be recovered from , which implies the injectiveness of .
First, we recover the partition . That is, given a node , we distinguish whether or . If , then according to (2). Otherwise, we have for some , hence is in . Assume the nodes in are . According to Claim C.5, we can see that the following procedure distinguishes whether or for all , including .
Secondly, we can easily recover from and . Ignoring labels, it is easy to see that is exactly the induced subgraph of on . By the way, we also get the function . For labels, we simply replace each label or with .
Finally, we recover from , and . That is, we distinguish which one of contains a given node . Assume and . Let be the node previous to in . According to Part (c) of Claim C.3, if and only if . When , if , and if . When , if , and if .
∎
Appendix D Proof of Lemma 6.1
Let denote the vector whose coordinates are all 0 except the -th that equals 1. The following lemmas will be used in the proof.
Lemma D.1.
[37] Let be a dependency graph and be a probability vector beyond the Shearer’s bound. Suppose form a shortest path from to in . Then for any , is also beyond the Shearer’s bound.
Without loss of generality, we assume that is rational for each . By contradiction, let be such a vector which is beyond Shearer’s bound. Let and . Let be a real number such that the following hold:
-
•
For each , for some . Intuitively, we cut into pieces each of size . Besides, we call such pieces positive pieces.
-
•
For each ,
for some . Intuitively, we cut into pieces each of size . We call such pieces negative pieces.
We use and to denote the set of positive pieces and negative pieces respectively.
For convenience, let if , and if . Then Condition (c) can be restated as: for , the positive pieces in are no more than the negative pieces in , i.e.,
(16) |
The basic idea of Lemma 6.1 is relatively simple: for each , we move positive pieces in to such that (i) all the positive pieces in are absorbed by the negative pieces in and (ii) the resulted probability vector is still beyond Shearer’s bound. Finally, all positive pieces will be absorbed, and we will get a vector strictly smaller than . By Lemma D.1, this vector is beyond Shearer’s bound, which makes a contradiction.
For , remember Condition (a) which says that there are at most subsets such that , and we use to represent these subsets. Let be a injection mapping each to some satisfying that (i) and (ii)
By (16), one can verify that such mapping exists. In addition, according to Condition (b), if , then .
In the following, we will apply Lemma D.1 repeatedly.
Let be , be and be . Given an injection , and where if , we construct another injection , and as follows. There are two possible cases for , and .
-
(1)
there exists such that , and there is a shortest path between and such that no vertex in is on the path;
-
(2)
For each where and each shortest path between and , there is a vertex in on the path.
For case (1), we let , , and
For case (2), there must be for some such that
-
-
, , for each ,
-
-
is on a shortest path between and for each ,
-
-
is on a shortest path between and .
We define the injection as follows.
One can verify that if and
Since is bounded, there must be a constant and such that , and there is a shortest path between and such that no vertex in is on the path. Let , and
One can verify that in both cases, is an injection from to and if .
Let be . For each , let be the unique element in . Let denote . Thus, we have
-
-
is an injection from to ,
-
-
for each ,
-
-
there is a shortest path between and such that are not on the path.
For each , define
Because is an injection, we have . Let
By is beyond Shearer’s bound and for each , we have is also beyond Shearer’s bound. For each , let
Then we have the following claim.
Claim D.2.
For , is beyond Shearer’s bound.
Proof.
We prove this claim by induction. Obviously, is beyond Shearer’s bound. In the following, we prove that if is beyond Shearer’s bound, then is also beyond Shearer’s bound.
Let
Obviously, . By is beyond Shearer’s bound, we have is also beyond Shearer’s bound. Note that there is a shortest path between and such that are not on the path. Because is beyond Shearer’s bound, by Lemma D.1, we have
is also beyond Shearer’s bound. Meanwhile, by , we have
For each , if , we have . Otherwise, , and for each . Thus, we have . Therefore,
By and for each , we have
Thus, by is beyond Shearer’s bound, we have
is also beyond Shearer’s bound. ∎
Thus, we have is beyond Shearer’s bound. It is easy to verify that , which is contradictory with that is on Shearer’s boundary.
Appendix E Missing part in the proof of Theorem 6.2
Proof of Claim 6.3.
Observe that for each , if , then one of its neighboring edge is in and satisfies that . Here, we say two edges neighboring if they share a common vertex. Besides, note that each edge has at most neighboring edges. So
(17) |
Let , , , and . In the following, we check that all the three conditions in Lemma 6.1 hold.
Condition (a). That is, we want to show for each . Observe that if , then . So
The last inequality uses the fact that if .
Condition (b). That is, we want to show for any and . This is obvious, because if , then .
Condition (c). We verify that
(19) |
On one hand, noting that , we have
(20) |