Optimal Mixing via Tensorization for Random Independent Sets on Arbitrary Trees
Abstract
We study the mixing time of the single-site update Markov chain, known as the Glauber dynamics, for generating a random independent set of a tree. Our focus is obtaining optimal convergence results for arbitrary trees. We consider the more general problem of sampling from the Gibbs distribution in the hard-core model where independent sets are weighted by a parameter ; the special case corresponds to the uniform distribution over all independent sets. Previous work of Martinelli, Sinclair and Weitz (2004) obtained optimal mixing time bounds for the complete -regular tree for all . However, Restrepo, Stefankovic, Vera, Vigoda, and Yang (2014) showed that for sufficiently large there are bounded-degree trees where optimal mixing does not hold. Recent work of Eppstein and Frishberg (2022) proved a polynomial mixing time bound for the Glauber dynamics for arbitrary trees, and more generally for graphs of bounded tree-width.
We establish an optimal bound on the relaxation time (i.e., inverse spectral gap) of for the Glauber dynamics for unweighted independent sets on arbitrary trees. We stress that our results hold for arbitrary trees and there is no dependence on the maximum degree . Interestingly, our results extend (far) beyond the uniqueness threshold which is on the order . Our proof approach is inspired by recent work on spectral independence. In fact, we prove that spectral independence holds with a constant independent of the maximum degree for any tree, but this does not imply mixing for general trees as the optimal mixing results of Chen, Liu, and Vigoda (2021) only apply for bounded degree graphs. We instead utilize the combinatorial nature of independent sets to directly prove approximate tensorization of variance via a non-trivial inductive proof.
1 Introduction
This paper studies the mixing time of the Glauber dynamics for the hard-core model assuming that the underlying graph is an arbitrary tree. In the hard-core model, we are given a graph and an activity . The model is defined on the collection of all independent sets of (regardless of size), which we denote as . Each independent set is assigned a weight where is the number of vertices contained in the independent set . The Gibbs distribution is defined on : for , let where is known as the partition function. When then every independent set has weight one and hence the Gibbs distribution is the uniform distribution over (unweighted) independent sets.
Our goal is to sample from (or estimate ) in time polynomial in . Our focus is on trees. These sampling and counting problems are computationally easy on trees using dynamic programming algorithms. Nevertheless, our interest is to understand the convergence properties of a simple Markov Chain Monte Carlo (MCMC) algorithm known as the Glauber dynamics for sampling from the Gibbs distribution.
The Glauber dynamics (also known as the Gibbs sampler) is the simple single-site update Markov chain for sampling from the Gibbs distribution of a graphical model. For the hard-core model with activity , the transitions of the Glauber dynamics are defined as follows: first, choose a random vertex . Then, with probability set and with the complementary probability set . If is an independent set, then set and otherwise set .
We consider two standard notions of convergence to stationarity. The relaxation time is the inverse spectral gap, i.e., where and are the eigenvalues of the transition matrix for the Glauber dynamics. The relaxation time is a key quantity in the running time for approximate counting algorithms (see, e.g., Štefankovič, Vempala, and Vigoda [ŠVV09]). The mixing time is the number of steps, from the worst initial state, to reach within total variation distance of the stationary distribution, which in our case is the Gibbs distribution .
We say that is the optimal relaxation time and that is the optimal mixing time (see Hayes and Sinclair [HS07] for a matching lower bound for any constant degree graph). Here, denotes the size of the underlying graph. More generally, we say the Glauber dynamics is rapidly mixing when the mixing time is .
We establish bounds on the mixing time of the Glauber dynamics by means of approximate tensorization inequalities for the variance of the hard-core model. Interestingly, our analysis utilizes nothing further than the inductive nature of the tree, e.g., we do not make any assumptions about spatial mixing properties of the Gibbs distribution. As a consequence, the bounds we obtain have no dependence on the maximum degree of the graph.
To be more specific we derive the following two group of results: We establish approximate tensorization of variance of the hard-core model on the tree for all . This implies optimal relaxation time for the Glauber dynamics. Notably this also includes the uniform distribution over independent sets, i.e., .
We can now state our main results.
Theorem 1.1.
For any -vertex tree, for any the Glauber dynamics for sampling -weighted independent sets in the hard-core model has an optimal relaxation time of .
We believe the optimal mixing results of Theorem 1.1 are related to the reconstruction threshold, which we describe now. Consider the complete -regular tree of height ; this is the rooted tree where all nodes at distance from the root have children and all nodes at distance from the root are leaves. We are interested in how the configuration at the leaves affects the configuration at the root.
Consider fixing an assignment/configuration to the leaves (i.e., specifying which leaves are fixed to occupied and which are unoccupied), we refer to this fixed assignment to the leaves as a boundary condition . Let denote the Gibbs distribution conditional on this fixed boundary condition , and let denote the marginal probability that the root is occupied in .
The uniqueness threshold measures the affect of the worst-case boundary condition on the root. For all , all , in the limit , we have ; this is known as the (tree) uniqueness region. In contrast, for there are pairs (namely, all even occupied vs. odd occupied) for which the limits are different; this is the non-uniqueness region. The uniqueness threshold is at .
In contrast, the reconstruction threshold measures the affect on the root of a random/typical boundary condition. In particular, we fix an assignment at the root and then generate the Gibbs distribution via an appropriately defined broadcasting process. Finally, we fix the boundary configuration and ask whether, in the conditional Gibbs distribution , the root has a bias towards the initial assignment . The non-reconstruction region corresponds to when we cannot infer the root’s initial value, in expectation over the choice of the boundary configuration and in the limit , see Mossel [Mos04] for a more complete introduction to reconstruction.
The reconstruction threshold is not known exactly but close bounds were established by Bhatnagar, Sly, and Tetali [BST16] and Brightwell and Winkler [BW04]; they showed that for constants and sufficiently large : , and hence is “increasing asymptotically” with whereas the uniqueness threshold is a decreasing function of . Martin [Mar03] showed that for all . As a consequence, we conjecture that Theorem 1.1 holds for all trees for all , which is close to the bound we obtain. A slowdown in the reconstruction region is known: Restrepo, Štefankovič, Vera, Vigoda, and Yang [RŠV+14] showed that there are trees for which there is a polynomial slow down for for a constant ; an explicit constant is not stated in [RŠV+14] but using the Kesten-Stigum bound one can show (by considering binary trees).
For general graphs the appropriate threshold is the uniqueness threshold, which is . In particular, for bipartite random -regular graphs the Glauber dynamics has optimal mixing in the uniqueness region by Chen, Liu and Vigoda [CLV21], and is exponentially slow in the non-uniqueness region by Mossel, Weitz, and Wormald [MWW07] (see also [GŠV16]). Moreover, for general graphs there is a computational phase transition at the uniqueness threshold: optimal mixing on all graphs of maximum degree in the uniqueness region [CLV21, CFYZ22, CE22], and NP-hardness to approximately count/sample in the non-uniqueness region by Sly [Sly10] (see also, [GŠV16, SS14]).
There are a variety of mixing results for the special case on trees, which is the focus of this paper. In terms of establishing optimal mixing time bounds for the Glauber dynamics, previous results only applied to complete, -regular trees. Seminal work of Martinelli, Sinclair, and Weitz [MSW03, MSW04] proved optimal mixing on complete -regular trees for all . The intuitive reason this holds for all is that the complete tree corresponds to one of the two extremal phases (all even boundary or all odd boundary) and hence it does not exhibit the phase co-existence which causes mixing. As mentioned earlier, [RŠV+14] shows that there is a fixed assignment for the leaves of the complete, -regular tree so that the mixing time slows down in the reconstruction region; intuitively, this boundary condition corresponds to the assignment obtained by the broadcasting process.
For more general trees the following results were known. A classical result of Berger, Kenyon, Mossel and Peres [BKMP05] proves polynomial mixing time for trees with constant maximum degree [BKMP05]. A very recent result of Eppstein and Frishberg [EF23] proved polynomial mixing time of the Glauber dynamics for graphs with bounded tree-width which includes arbitrary trees, however the polynomial exponent is for trees; see more recent work of Chen [Che24] for further improvements. For other combinatorial models, rapid mixing for the Glauber dynamics on trees with bounded maximum degree was established for -colorings in [LMP09] and edge-colorings in [DHP20].
Spectral independence is a powerful notion in the analysis of the convergence rate of Markov Chain Monte Carlo (MCMC) algorithms. For independent sets on an -vertex graph , spectral independence considers the pairwise influence matrix where ; this matrix is closely related to the vertex covariance matrix. We say that spectral independence holds if the maximum eigenvalue of for all vertex-induced subgraphs of are bounded by a constant. Spectral independence was introduced by Anari, Liu, and Oveis Gharan [ALO21]. Chen, Liu, and Vigoda [CLV21] proved that spectral independence, together with a simple condition known as marginal boundedness which is a lower bound on the marginal probability of a valid vertex-spin pair, implies optimal mixing time of the Glauber dynamics for constant-degree graphs. This has led to a flurry of optimal mixing results, e.g., [CLV22, BCC+22, Liu21, CLMM23, CG24].
The limitation of the above spectral independence results is that they only hold for graphs with constant maximum degree . There are several noteworthy results that achieve a stronger form of spectral independence which establishes optimal mixing for unbounded degree graphs [CFYZ22, CE22]; however all of these results for general graphs only achieve rapid mixing in the tree uniqueness region which corresponds to whereas we are aiming for .
The inductive approach we use to establish approximate tensorization inequalities can also be utilized to establish spectral independence. In fact, we show that spectral independence holds for any tree when , including the case where , see Section 4. Applying the results of Anari, Liu, and Oveis Gharan [ALO21] we obtain a bound on the mixing time, but with a large constant in the exponent of . For constant degree trees we obtain the following optimal mixing result by applying the results of Chen, Liu, and Vigoda [CLV21] (see also [BCC+22, CFYZ22, CE22]).
Theorem 1.2.
For all constant , all , for any tree with maximum degree , the Glauber dynamics for sampling -weighted independent sets has an optimal mixing time of .
Interestingly, combining the spectral independence results from the proof Theorem 1.2 with results of Chen, Feng, Yin, and Zhang [CFYZ24, Theorem 1.9], we are able to strengthen Theorem 1.1 by allowing larger fugacities, i.e., .
Corollary 1.3.
For any -vertex tree, for any the Glauber dynamics for sampling -weighted independent sets in the hard-core model has an optimal relaxation time of .
In the next section we recall the key functional definitions and basic properties of variance that will be useful later in the proofs. In Section 3 we prove approximate tensorization of variance which establishes Theorem 1.1. We establish spectral independence and prove Theorem 1.2 in Section 4.
Remark 1.4.
An earlier version of this paper [EHŠV23] claimed to prove mixing time for for any tree (without any constraint on the maximum degree). There was a mistake in that proof. In particular, inequality (54) is false. Moreover, Zongchen Chen pointed out a simple test function which shows that entropy tensorization and modified log-Sobolev inequality (MLSI) do not hold for the star graph with constants independent of the degree.
A recent paper of Chen, Yang, Yin, and Zhang [CYYZ24] improves Corollary 1.3 to obtain optimal relaxation time on trees for .
2 Preliminaries
2.1 Standard Definitions
Let be the transition matrix of a Markov chain with a finite state space and equilibrium distribution . For and , let denote the distribution of when the initial state of the chain satisfies . The mixing time of the Markov chain is defined by
(1) |
The transition matrix with stationary distribution is called time reversible if it satisfies the so-called detailed balance relation, i.e., for any we have . For that is time reversible the set of eigenvalues are real numbers and we denote them as . Let , then we define the relaxation time by
(2) |
The quantity is also known as the spectral gap of . We use to bound by using the following inequality
(3) |
2.2 Gibbs Distributions and Functional Analytic Definitions
For a graph and , let be the hard-core model on this graph with activity , while let be the support of , i.e., are the collection of independent sets of .
For any function , we let is the expected value of with respect to , i.e.,
Analogously, we define variance of with respect to by
We are also using the following equation for ,
For any subset , let denote the set of independent sets on . Then, let denote the marginal of on ; that is, for any , we have that
For any , any , we let be the distribution conditional on the configuration on , and let to be the support of .
For any , for any , we define the function such that for all .
Let
Let denote the variance of with respect to the conditional distribution :
(4) |
Furthermore, we let
(5) |
i.e., is the average of with respect to being distributed as in . For the sake of brevity, when , i.e., the set is a singleton, we use .
Finally, let
(6) |
i.e., is the variance of with respect to being distributed as in .
When is the following two-valued random variable:
then one formulation for the variance that will be convenient for us is
(7) |
2.3 Approximate Tensorization of Variance
To bound the convergence rate of the Glauber dynamics we consider the approximate tensorization of variance as introduced in [CMT15].
We begin with the definition of approximate tensorization of variance.
Definition 2.1 (Variance Tensorization).
A distribution with support satisfies the approximate tensorisation of Variance with constant , denoted using the predicate , if for all we have that
2.4 Decomposition of Variance
We use the following basic decomposition properties for variance. The following lemma follows from a decomposition of relative entropy, see [CP21, Lemma 3.1] (see also [BCC+22, Lemma 2.3]).
Lemma 2.2.
We utilize the following variance factorisation for product measures, see [Cap23, Eqn (4.7)].
Lemma 2.3.
Consider where . Then for all we have:
(10) |
On a first account, the reader might find it challenging to parse the expression . In that respect, (5) and (6) might be useful. Specifically, is the expectation of with respect to being distributed as in . Furthermore, corresponds to the variance of with respect to the configurations at and at , while is fixed and is distributed as in .
Proof.
We apply [Cap23, Eqn (4.7)], which reaches the same conclusion under the assumptions that and . In the current context, the reason these conditional expectation operators commute here is that, because , if we let be an independent set sampled according to distribution , then the random variables and are conditionally independent given . ∎
3 Variance Factorization
In this section we prove Theorem 1.1, establishing an optimal bound on the relaxation time for the Glauber dynamics on any tree for . We will prove this by establishing variance tensorization, see Definition 2.1.
Consider a graph and a collection of fugacities for each . Throughout this section we will assume that all the fugacities are bounded by . Consider the following more general definition of the Gibbs distribution for the hard-core model, where for an independent set ,
(11) |
Let be a tree, let be a collection of fugacities and let be the corresponding Gibbs distribution. We will establish the following variance tensorization inequality: for all
(12) |
where is a function to be determined later (in Lemma 3.1). We refer to as the “global” variance and we refer to as the “local” variance (of at ).
We will establish (12) using induction. Let be a vertex of degree in and let be the unique neighbor of . Let be the tree by removing from , i.e., is the induced subgraph of on . Let be a collection of fugacities where for and . Let be the hard-core measure on with fugacities .
Note that for we have
(13) |
Fix a function . Let be defined by
(14) |
Note that is defined on independent sets of the tree and is the natural projection of to the tree . Since , then by Lemma 2.2 we have that:
(15) |
For measure , when we condition on the configuration at , the configuration at is independent of that at . Hence, from eq. 10 for any (by setting and ) we have:
The following lemma is the main technical ingredient. It bounds the local variance at for the smaller graph in terms of the local variance at and in the original graph .
Lemma 3.1.
For and any we have:
(17) |
We now utilize the above lemma to prove the main theorem for the relaxation time. We then go back to prove Lemma 3.1.
Proof of Theorem 1.1.
Note eq. 17 is equivalent to:
(18) |
We can prove variance tensorization by induction as follows:
For the first line, we use eq. 15. The second line follows from the inductive hypothesis. For the third line, we use eq. 16 and the fact that , since is increasing and . The fourth line follows by using eq. 18. ∎
Our task now is to prove Lemma 3.1. The following technical inequality will be useful.
Lemma 3.2.
Let . Suppose satisfy . Then for all we have
(19) |
Proof.
Equation (19) is equivalent to
(20) |
We derive the above by subtracting and from both sides of equation (19) and rearranging.
A simple application of the AM-GM inequality implies that
Then, equation (20) follows from the observation that the left-hand side of the above inequality is at least , i.e., since . ∎
We can now prove the main lemma.
Proof of Lemma 3.1.
We will consider each of these local variances , , and . We will express each of them as a sum over independent sets of . We can then establish eq. 17 in a pointwise manner by considering the corresponding inequality for each independent set .
Let us begin by looking at the general definition of the expected local variance for any . Applying the definition in eq. 5 and then simplifying we obtain the following (a reader familar with the notation can apply eq. 7 to skip directly to the last line):
(21) |
Notice in eq. 21 that the only which contribute are those where is unblocked (i.e., no neighbor of is included in the independent set ) because we need that and are both independent sets and hence have positive measure in .
Let us now consider each of the local variances appearing in eq. 17 (expressed using carefully chosen summations that will allow us to prove (17) term-by-term in terms of ).
Let denote the expected local variance of at . We will use (21) (applied to instead of ); note that only where is unblocked (that is, when no neighbor of is occupied) contribute to the local variance. Moreover, we can express the expected local variance of at in terms of only those where . In particular, consider an independent set where . Note that if is blocked (i.e., ) then the local variance at is zero for this term. And for those where and is unblocked then the local variance has the same contribution as times (since ). Hence the expected local variance of at is given by
We have (since ) and . Plugging these in and simplifying we obtain the following:
(22) |
We now consider . As we did for we can express as a sum over indpendent sets where . In this case for an independent set where , consider and note that the following hold
and
Hence, we have the following:
(23) |
Finally, we consider , the expected local variance of at . We will establish a lower bound which we will denote by (note, and were identities but here we will have an inequality).
To compute , the expected local variance of at , we need to generate an independent set from . Only those where is unblocked (that is where is not in ) contribute to the local variance. We can generate by generating from (whether we add or do not add does not change the contribution to the local variance). As in eq. 21, we obtain the following:
Let denote the lower bound we obtained above:
(24) |
Plugging in (22), (23), (24) we obtain that eq. 17 follows from the following inequality:
(25) |
We will establish (25) term-by-term, that is, for each in the sums of (22), (23), (24). Fix such that and let , , and . We need to show
(26) |
Let . We will match (19) to (26), by first dividing both sides of (26) by and then choosing
Note that with this choice of and equations (19) and (26) are equivalent, and hence to prove (26) it is enough to show .
Claim 3.3.
This completes the proof of the lemma. ∎
We use the following lemma to prove Claim 3.3
Lemma 3.4.
Let and . Suppose are such that . Then
(27) |
Proof.
We will show that for we have
(28) |
and
(29) |
To see that (28) and (29) imply (27) note
where the first inequality follows from (28) and the second from (29).
Note that the constraints on imply that and . Hence . To prove (29) it is sufficient to show for
(30) |
We have
Hence is concave and we only need to check (30) for the endpoints of the interval; for LHS of (30) is zero, for the LHS of (30) has value larger than . This concludes the proof of (29).
To prove (28) note that
(31) |
Hence we only need to prove (28) for ; this simplifies to showing
(32) |
For and we have that (32) is satisfied (using interval arithmetic). Let
The critical points of are roots of
Since this polynomial is increasing when is non-negative, it has exactly one positive real root, which is in the interval . The value of on both endpoints of the interval is at least . The derivative of (a rational function) is bounded in absolute value by on the interval and hence on the entire interval (in particular at the critical point). This proves (32). ∎
We can now complete the proof of Claim 3.3.
Proof of Claim 3.3.
We will use the following substitution to simplify the expression . Note that . In terms of we have
Let and . We have
Recall, our goal is to show . First we show that for such that we have
(33) |
Note that (33) is equivalent to
(34) |
Note that the RHS of (34) is increasing in and hence it is sufficient to show (34) for the maximal value of ; this is equivalent to
the last inequality follows from
checked using Sturm sequences. This concludes the proof of (33).
4 Spectral Independence for Trees
Let be a graph. Suppose we have a set of fugacities for each , and consider the corresponding Gibbs distribution. Recall that the influence matrix has entry , where
Let be any two vertices in a graph. We have
(35) |
We have
(36) |
Let be diagonal matrix with entries . Equation (36) means that is symmetric. That also means that is symmetric (since it is obtained from by multiplying by the same diagonal matrix on the left and on the right). The entry in is
(37) |
which is equal to (take geometric mean of the sides of (37)). We will call the symmetrized influence matrix (since it is similar to the influence matrix, it has the same spectral radius).
We will prove the following result.
Lemma 4.1.
For any forest with fugacities in the spectral radius of the influence matrix of the hard-core model on is bounded by 10000.
We will prove Lemma 4.1 by induction; we will prove the following strengthened statement.
Lemma 4.2.
For any forest with fugacities in the symmetrized influence matrix of the hard-core model on satisfies
Proof.
Let be a forest. Let be a vertex of degree in . Let be the neighbor of in . Let be with removed. Let for . Let . Let be the hard-core model on with fugacities . Note that is the same as the marginalization of to , that is, for an independent set of we have
(38) |
where ranges over independent sets of that contain .
Let be the symmetrized influence matrix for and let be the symmetrized influence matrix for . Note that (38) implies that is a submatrix of (removing the column and row of corresponding to vertex yields ).
It is standard to show, e.g., see [ALO21, Lemma B.3], that
(39) |
and
(40) |
Let . Note that, using (35), we have
(41) |
From (39) and (40) we have that and have the following form ( is a matrix and is a vector):
Let be an increasing function on . We will take where and , . (We will keep and as variables to elucidate the choice of and in the proof below.) Suppose that we know and we want to conclude . Let be a diagonal matrix with entries for . We can restate our goal as follows.
Since will be an increasing function, we have .
The condition we want is equivalent (by applying row and column operations, specifically subtracting -th row/column times from the -th row/column) to
If we show
(42) |
then the conclusion will follow (adding the “know” positive semidefinite matrix and (42) (expanded with zeros to match dimensions) yields that the “want” matrix is positive semidefinite). Equation (42) is equivalent to checking if the determinant is positive (since the entry in the first row/column is positive), that is, we need to check
Hence it is sufficient (using (41)) to show
Letting the condition becomes the following
(43) |
We are going to search for that is 1) bounded from above by , 2) bounded away from on and 3) that satisfies
(44) |
Note that such will also satisfy (43) since (here the choice of is arbitrary; we just need something sufficiently small).
We will search for of the following form: . Note that (44) is invariant under scaling of and hence satisfying the upper bound of can be achieved by scaling of . Ultimately the price we will pay for being small is that is big and hence we get a weaker upper bound on the spectral radius of ; we do not optimize the constants at this point. Consider the following function (obtained as LHS of (44) divided by RHS of (44), excluding the constant term ):
(45) |
The minimum of the first term in (45) is attained for and hence setting we have that:
(46) |
For the second term in (45) by setting we have the following:
(47) |
The second inequality in (47) is obtained from the following inequality which is valid for and :
(48) |
For the above, it suffices to prove that
(49) |
We have that for :
by Bernoulli’s inequality | ||||
Finally, plugging in (46) and (47) into (45) we obtain:
(50) |
Equation (50) implies that (44) is satisfied. Recall that the statement of the lemma assumed that the fugacities are in the interval . For we have
(51) |
This implies that satisfies (43) and hence for
we have (42) and this completes the proof of Lemma 4.2 by induction. ∎
We can now complete the proof for spectral independence.
Proof of Theorem 1.2.
We apply Theorem 1.12 of [CLV21] which says that if for all pinnings we have -spectral independence and -marginally boundedness then the mixing time of the Glauber dynamics is . A pinning refers to a fixed configuration on a subset of vertices; for the hard-core model a pinning of a tree corresponds to the hard-core model on a forest which is an induced subgraph. Hence, Lemma 4.1 implies that -spectral independence holds for all pinnings with . The condition -marginally boundedness (see Definition 1.9 in [CLV21]) says that for every pinning on a subset , for every vertex , for every assignment to denoted as which has positive probability in the conditional Gibbs distribution , then the marginal probability is lower bounded as . This holds for . Hence, [CLV21, Theorem 1.12] implies Theorem 1.2. ∎
5 Proof of Corollary 1.3
We prove Corollary 1.3 by combining the spectral independence result in Lemma 4.1 and Theorem 1.1, while we utilise [CFYZ24, Theorem 1.9].
In [CFYZ24], they introduce the notion of “complete -spectral independence”. Let be the hard-core model on graph with fugacity . For , complete -spectral independence for corresponds to the following condition: For any induced subgraph of , for the hard-core model on such that each vertex has fugacity , the corresponding influence matrix has spectral radius at most .
Lemma 4.1 is equivalent to complete -spectral independence for all with . We can now prove Corollary 1.3.
Proof of Corollary 1.3:.
For the forest , let be the hard-core model with fugacity . Also, let be the spectral gap for Glauber dynamics on . Corollary 1.3 follows by showing that .
First, note that if , then Theorem 1.1 already implies . We now focus on the case where .
For the same forest , let be the hard-core model on with fugacity . Let be the spectral gap for Glauber dynamics on . From Theorem 1.1, we have that
(52) |
Lemma 4.1 and [CFYZ24, Theorem 1.9] imply the following relation between and : for and , we have
(53) |
From the above we have established Corollary 1.3. ∎
Acknowledgements
We thank Zongchen Chen for pointing out the counterexample to entropy tensorization as discussed in Remark 1.4, and the anonymous referee for pointing out Corollary 1.3.
References
- [ALO21] Nima Anari, Kuikui Liu, and Shayan Oveis Gharan. Spectral independence in high-dimensional expanders and applications to the hardcore model. SIAM Journal on Computing, 2021.
- [BCC+22] Antonio Blanca, Pietro Caputo, Zongchen Chen, Daniel Parisi, Daniel Štefankovič, and Eric Vigoda. On Mixing of Markov Chains: Coupling, Spectral Independence, and Entropy Factorization. In Proceedings of the 33rd Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), pages 3670–3692, 2022.
- [BKMP05] Noam Berger, Claire Kenyon, Elchanan Mossel, and Yuval Peres. Glauber dynamics on trees and hyperbolic graphs. Probability Theory and Related Fields, 131(3):311–340, 2005.
- [BST16] Nayantara Bhatnagar, Allan Sly, and Prasad Tetali. Decay of correlations for the hardcore model on the -regular random graph. Electron. J. Probab., 21:1–42, 2016.
- [BW04] Graham Brightwell and Peter Winkler. A second threshold for the hard-core model on a Bethe lattice. Random Structures and Algorithms, 24:303–314, 2004.
- [Cap23] Pietro Caputo. Lecture notes on Entropy and Markov Chains. Preprint, available from: http://www.mat.uniroma3.it/users/caputo/entropy.pdf, 2023.
- [CE22] Yuansi Chen and Ronen Eldan. Localization schemes: A framework for proving mixing bounds for Markov chains. In Proceedings of the 63rd Annual IEEE Symposium on Foundations of Computer Science (FOCS), pages 110–122, 2022.
- [CFYZ22] Xiaoyu Chen, Weiming Feng, Yitong Yin, and Xinyuan Zhang. Optimal mixing for two-state anti-ferromagnetic spin systems. In Proceedings of the 63rd Annual IEEE Symposium on Foundations of Computer Science (FOCS), pages 588–599, 2022.
- [CFYZ24] Xiaoyu Chen, Weiming Feng, Yitong Yin, and Xinyuan Zhang. Rapid mixing of Glauber dynamics via spectral independence for all degrees. SIAM Journal on Computing, 2024.
- [CG24] Zongchen Chen and Yuzhou Gu. Fast sampling of -matchings and -edge covers. In Proceedings of the 35th Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), pages 4972–4987, 2024.
- [Che24] Zongchen Chen. Combinatorial approach for factorization of variance and entropy in spin systems. In Proceedings of the 35th Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), pages 4988–5012, 2024.
- [CLMM23] Zongchen Chen, Kuikui Liu, Nitya Mani, and Ankur Moitra. Strong spatial mixing for colorings on trees and its algorithmic applications. In Proceedings of the 64th Annual IEEE Symposium on Foundations of Computer Science (FOCS), pages 810–845, 2023.
- [CLV21] Zongchen Chen, Kuikui Liu, and Eric Vigoda. Optimal mixing of Glauber dynamics: Entropy factorization via high-dimensional expansion. In Proceedings of the 53rd Annual ACM Symposium on Theory of Computing (STOC), pages 1537–1550, 2021.
- [CLV22] Zongchen Chen, Kuikui Liu, and Eric Vigoda. Spectral independence via stability and applications to Holant-type problems. In Proceedings of the 63rd Annual IEEE Symposium on Foundations of Computer Science (FOCS), pages 149–160, 2022.
- [CMT15] Pietro Caputo, Georg Menz, and Prasad Tetali. Approximate tensorization of entropy at high temperature. Annales de la Faculté des sciences de Toulouse: Mathématiques, 24(4):691–716, 2015.
- [CP21] Pietro Caputo and Daniel Parisi. Block factorization of the relative entropy via spatial mixing. Communications in Mathematical Physics, 388:793–818, 2021.
- [CYYZ24] Xiaoyu Chen, Xiongxin Yang, Yitong Yin, and Xinyuan Zhang. Spectral independence beyond total influence on trees and related graphs. arXiv preprint arXiv:2404.04668, 2024.
- [DHP20] Michelle Delcourt, Marc Heinrich, and Guillem Perarnau. The Glauber dynamics for edge-colorings of trees. Random Structures & Algorithms, 57(4):1050–1076, 2020.
- [EF23] David Eppstein and Daniel Frishberg. Rapid mixing of the hardcore Glauber dynamics and other Markov chains in bounded-treewidth graphs. In Proceedings of the 34th International Symposium on Algorithms and Computation (ISAAC), pages 30:1–30:13, 2023.
- [EHŠV23] Charilaos Efthymiou, Thomas P. Hayes, Daniel Štefankovič, and Eric Vigoda. Optimal mixing via tensorization for random independent sets on arbitrary trees. arXiv preprint arXiv:2307.07727, version 2, 2023.
- [GŠV16] Andreas Galanis, Daniel Štefankovič, and Eric Vigoda. Inapproximability of the partition function for the antiferromagnetic Ising and hard-core models. Combinatorics, Probability and Computing, 25(4):500–559, 2016.
- [HS07] Thomas P. Hayes and Alistair Sinclair. A general lower bound for mixing of single-site dynamics on graphs. The Annals of Applied Probability, 17(3):931 – 952, 2007.
- [Liu21] Kuikui Liu. From coupling to spectral independence and blackbox comparison with the down-up walk. In APPROX-RANDOM, pages 32:1–32:21, 2021.
- [LMP09] Brendan Lucier, Michael Molloy, and Yuval Peres. The Glauber dynamics for colourings of bounded degree trees. In Randomization and Approximation Techniques in Computer Science (RANDOM), pages 631–645, 2009.
- [Mar03] James B. Martin. Reconstruction thresholds on regular trees. In Discrete Random Walks, DIMACS Series in Discrete Mathematics and Theoretical Computer Science, page 191–204, 2003.
- [Mos04] Elchanan Mossel. Survey: Information flow on trees. In Graphs, Morphisms and Statistical Physics, DIMACS Series in Discrete Mathematics and Theoretical Computer Science, page 155–170, 2004.
- [MSW03] Fabio Martinelli, Alistair Sinclair, and Dror Weitz. The Ising model on trees: boundary conditions and mixing time. In Proceedings of the 44th Annual IEEE Symposium on Foundations of Computer Science (FOCS), pages 628–639, 2003.
- [MSW04] Fabio Martinelli, Alistair Sinclair, and Dror Weitz. Fast mixing for independent sets, colorings and other models on trees. In Proceedings of the 15th Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), pages 456–465, 2004.
- [MWW07] Elchanan Mossel, Dror Weitz, and Nicholas C. Wormald. On the hardness of sampling independent sets beyond the tree threshold. Probability Theory and Related Fields, 143:401–439, 2007.
- [RŠV+14] Ricardo Restrepo, Daniel Štefankovič, Juan C. Vera, Eric Vigoda, and Linji Yang. Phase transition for Glauber dynamics for independent sets on regular trees. SIAM Journal on Discrete Mathematics, 28:835–861, 2014.
- [Sly10] Allan Sly. Computational transition at the uniqueness threshold. In Proceedings of the 51st Annual IEEE Symposium on Foundations of Computer Science (FOCS), pages 287–296, 2010.
- [SS14] Allan Sly and Nike Sun. The computational hardness of counting in two-spin models on -regular graphs. The Annals of Probability, 42(6):2383–2416, 2014.
- [ŠVV09] Daniel Štefankovič, Santosh Vempala, and Eric Vigoda. Adaptive simulated annealing: A near-optimal connection between sampling and counting. Journal of the ACM, 56(3):1–36, 2009.