On the Houdré-Tetali conjecture about an isoperimetric constant of graphs
Houdré and Tetali defined a class of isoperimetric constants of graphs for , and conjectured a Cheeger-type inequality for of the form
where is the second smallest eigenvalue of the normalized Laplacian matrix. If true, the conjeecture would be a strengthening of the hard direction of the classical Cheeger’s inequality. Morris and Peres proved Houdré and Tetali’s conjecture up to an additional log factor, using techniques from evolving sets. We present the following related results on this conjecture.
-
1.
We provide a family of counterexamples to the conjecture of Houdré and Tetali, showing that the logarithmic factor is needed.
-
2.
We match Morris and Peres’s bound using standard spectral arguments.
-
3.
We prove that Houdré and Tetali’s conjecture is true for any constant strictly bigger than , which is also a strengthening of the hard direction of Cheeger’s inequality.
Furthermore, our results can be extended to directed graphs using Chung’s definition of eigenvalues for directed graphs.
1 Introduction
Motivated by Talagrand’s isoperimetric inequality for the hypercubes [Tal93] (see Section 1.2), Houdré and Tetali [HT04] extended Talagrand’s isoperimetric constant to general Markov chains and also to different exponents.
Definition 1.1 (Isoperimetric Constants for Markov Chains [HT04]).
Let be an irreducible Markov chain with vertex set , transition matrix and stationary distribution . For any , define the isoperimetric constant as
Let be the inner vertex boundary of . Then, for ,
Given an undirected graph or a directed graph , let be the transition matrix of the natural random walk, where in the undirected case and in the directed case. Then the isoperimetric constant for the graph is defined as .
When , this is known as the Cheeger isoperimetric constant of a Markov chain (see e.g. [LP17]) or the edge conductance/expansion of a graph (see Section 2 for definition). When , this measures the vertex expansion of an undirected graph or a directed graph. Talagrand studied the case and proved a lower bound on for Boolean hypercubes. One can view as a quantity that interpolates between edge conductance and vertex expansion, since it follows from the Cauchy-Schwarz inequality that , and Talagrand used his lower bound to recover Margulis’ theorem about edge conductance and vertex expansion on hypercubes (see Section 1.2).
For an undirected graph , one can use the second smallest eigenvalue of the matrix to give upper and lower bound on . The classical Cheeger’s inequality is
Houdré and Tetali conjectured that the same relations hold for as well when the Markov chain is reversible.
Conjecture 1.2 ([HT04]).
Let be an irreducible and reversible Markov chain. Then
It is clear from the definition that increases as decreases, and thus for all . Therefore, the Houdré-Tetali conjecture is a strengthening of the hard direction of Cheeger’s inequality. It predicts that when the hard direction of Cheeger’s inequality is tight for a graph such that , then the graph must satisfy . Or, in other words, when such as on the hypercube (see 1.7) or on the dumbbell graphs, then the hard direction of Cheeger’s inequality cannot be tight. So, the conjecture can be viewed as an improved Cheeger’s inequality in the spirit of [KLL+13, KLL17] and this is a main motivation of this work.
Morris and Peres came close to proving the conjecture with an extra logarithmic factor.
Theorem 1.3 ([MP05]).
Let be an irreducible and reversible Markov chain. Suppose that for all . Then
Their proof is based on techniques from evolving sets. They lower bounded the “boundary gauge” using , and also upper bounded the mixing rate using so that they can relate and .
1.1 Our Results
We found counterexamples to the Houdré-Tetali conjecture, showing that the extra logarithmic factor is needed.
Theorem 1.4.
There are irreducible and reversible Markov chains with
The counterexample is simple to describe, which is a weighted undirected graph with vertex set and edge weight inversely proportional to . See Section 4 for details.
On the positive side, we match the result of Morris and Peres using standard spectral arguments. We show that the simple sweep-cut algorithm can be used to output a set with satisfying the guarantee in Theorem 1.3, without the self-loop assumption. See Section 3.1.
Perhaps more interestingly, the same arguments can be used to prove that the Houdré-Tetali conjecture is true if we replace by any constant .
Theorem 1.5.
Let be an irreducible and reversible Markov chain. For any ,
Similar to the discussion after 1.2, this shows that the tight examples of the hard direction of Cheeger’s inequality must satisfy for any . Also, this provides an improved analysis of Cheeger’s inequality that if then . So this result has similar consequences as if Houdré and Tetali’s conjecture was true.
Finally, we observe that the same statement as in Theorem 1.5 can also be proved for non-reversible Markov chains, by replacing with the eigenvalue defined for directed graphs by Chung [Chu05]. See Theorem 3.2.
Remark 1.6 ( for ).
For , a simple argument shows that an inequality of the form in Theorem 1.5 cannot hold. To see it, consider the transformation for some parameter (equivalent to adding a large self loop when is small) will scale by a factor , while the second eigenvalue scales by a factor of , and so the ratio scales by as . When , adding self loops does not change the ratio. Thus, it is the first exponent where such an inequality makes sense.
1.2 Previous Work on Boolean Hypercubes
The isoperimetric constant was initially studied by Talagrand in the Boolean hypercubes. Let be the -dimensional hypercube. For a point , let be the point obtained by flipping the -th bit of . For a subset , if define and otherwise if define
so that is the size of the edge boundary of . Let be the uniform distribution on and . The classical Poincaré inequality can be stated as, for any ,
(1.1) |
Talagrand [Tal93] proved a strengthening of the Poincaré inequality: For any ,
(1.2) |
The quantity is always smaller than and can be seen as a different measure of the boundary information of . Let be the vertex boundary of . By the Cauchy-Schwarz inequality, Talagrand’s theorem implies Margulis’ theorem [Mar77] that
which was an original motivation for Talagrand to consider the quantity . More recently, both Margulis’ and Talagrand’s theorems inspired the analogs for directed graphs developed in [CS16, KMS18], to make major progresses in analyzing sublinear time algorithms for testing monotone functions. See also [EG22, EKLM22] for a proof of a Talagrand’s conjecture that further sharpens these inequalities.
The following remark clarifies the connection between and the quantities appearing in Poincaré’s inequality and Talagrand’s inequality.
Remark 1.7 ( for Hypercubes).
For the -dimensional Boolean hypercube , the stationary distribution is simply the uniform distribution . Note that the numerator in is exactly , and the Poincaré inequality translates to . Similarly, the numerator of is exactly , and the Talagrand’s inequality translates to .
Finally, we note that a parameter similar to , called , was also studied in [EG22].
2 Preliminaries
Given two functions , we use to denote the existence of a positive constant , such that always holds. We use to denote and . For positive integers , we use to denote the set . For a function , denotes the domain subset on which is nonzero. The function refers to the base logarithm.
Undirected Graphs: Let be an undirected graph. Let be a weight function on the edges. The weighted degree of a vertex is defined as . Let be a nonempty subset of vertices. The edge boundary of is defined as and be the total edge weight of . The volume of is defined as . The edge conductance of and of are defined as
In an undirected graph, the ordinary random walk has transition matrix with for every . If the graph is connected, then the stationary distribution is unique with for every . It is thus straightforward to check that and , i.e. the two definitions coincide.
Directed Graphs: Let be a directed graph. Let be a weight function on the edges. The weighted indegree of a vertex is defined as and the weighted outdegree of is defined as . In a directed graph, the ordinary random walk has transition matrix with . The stationary distribution has no easy description but is unique as long as the directed graph is strongly connected. There are different notions of directed edge conductance for directed graphs. In analyzing random walks, the standard definition is exactly as described in 1.1, and this quantity is closely related to the mixing time of random walks; see e.g. [LP17, Chu05, MP05]. In analyzing graph partitioning, there is a definition that extends the edge conductance above to directed graphs, which will not be used in this paper; see e.g. [Yos16, LTW23].
Spectral Graph Theory: Given an undirected graph with a weight function , its adjacency matrix is an matrix where the -th entry is . The Laplacian matrix is defined as , where is the diagonal degree matrix. The normalized adjacency matrix is defined as , and the normalized Laplacian matrix is defined as . Let be the eigenvalues of . It is known that with eigenvector , and
Cheeger’s inequality [Che70, AM85, Alo86] is a fundamental result in spectral graph theory that connects edge conductance of an undirected graph to the second smallest eigenvalue of its normalized Laplacian matrix:
The random walk transition matrix is similar to the normalized Laplacian matrix , and the matrix is similar to the normalized Laplacian matrix . In particular, enjoys the same spectral properties as with real eigenvalues and a quadratic form characterization of as above; see 3.1.
Chung [Chu05] defined the Laplacian matrix of a directed graph and used it to prove an analog of Cheeger’s inequality. These will be stated in Section 3.2.
3 Positive Results
To prove Theorem 1.5, we follow standard spectral arguments used in proving Cheeger-type inequalities, in Trevisan’s style. First, we start with the second eigenvector and truncate it so that the weight of its support is at most half while preserving its Rayleigh quotient. The proof of the following lemma is standard and we defer it to the end of this section.
Lemma 3.1.
Let be an irreducible and reversible Markov chain. Let be an eigenvector associated to the second smallest eigenvalue of the matrix , with . Define the truncated vector such that for all . Then
Then the plan is to prove that one of the level sets has small isoperimetric constant. We index the vertices by and order them so that for . For any , define the level set . By the construction of , it holds that for any , and so . For convenience, we rescale so that .
To prove that one of has small isoperimetric constant, we choose a uniform random and consider . We will bound the expected value of the numerator of and of the denominator of and conclude that there exists a with at most the ratio of the expected values, i.e. .
The expected value of the denominator is easy to analyze. Since we choose uniformly randomly, each vertex is included in with probability , and thus
The rest of the proof is to analyze the expected value of the numerator . For a vertex , if the random threshold is between and with , then , and so
where the second equality is by writing a telescoping sum and the third equality is by a change of summation. So, the expected numerator is
where the first inequality is by Cauchy-Schwarz, and the second inequality is by 3.1 and and .
To upper bound the inner sum of the second term, we denote and it suffices to upper bound the sum of the form with and , with a bound independent of . Let denote the supremum of the sum when . Note that by a simple scaling argument. Let be an optimal sequence that achieves the supremum of . Then,
It follows that
where the second inequality is by the mean value theorem that and the last inequality is because and . Clearly, for any , and so the inner sum of the second term in the expected numerator is at most . Putting together, this completes the proof of Theorem 1.5 as
which implies that
3.1 Recovering Morris and Peres’s Result
To recover Theorem 1.3, we follow the same arguments but add a truncation step so that the sequence above will start with . In this subsection, we plug in . As above, the main work is to upper bound the expected value of the numerator. Recall that
Let be the index such that but . Then, we can upper bound the right hand side by
To shorten the expression, let us use the notations
Summing over and using these notations, the expected numerator is
Applying Cauchy-Schwarz as before gives
where the second inequality uses 3.1 and .
To upper bound , we let and use that to upper bound the function
The partial derivative of is
Since by definition, the function increases up until and then decreases. So, the maximum is attained when with and , in which case the sum is
For , this value is increasing when increases, and so the sum is upper bounded by
where the third last equality is by L’Hôpital’s rule. Plugging this back into and , the expected numerator is
As before, the expected denominator is . Putting together,
Rearranging recovers Theorem 1.3.
3.2 Non-Reversible Markov Chains
Houdré and Tetali only formulated 1.2 for reversible Markov chains of which the eigevalues of are real. For non-reversible Markov chains, we observe that Chung’s definition of eigenvalues for directed graphs [Chu05] can be used to obtain the same results in Theorem 1.3 and Theorem 1.5.
Given a directed graph with a weight function , let be the transition matrix of the ordinary random walks on with for each edge . Suppose is strongly connected, then there is a unique stationary distribution such that . Let . Chung defined the Laplacian of the directed graph as
Since is a real symmetric matrix, its eigenvalues are real. Let be the second smallest eigenvalue of . Chung [Chu05] proved an analog of Cheeger’s inequality that
We observe that can be used to extend our results to non-reversible Markov chains.
Theorem 3.2.
Let be an irreducible Markov chain. For any ,
For ,
Note that the main proofs of Theorem 1.5 and Theorem 1.3 (i.e. computing the expected numerator) did not require the Markov chain to be reversible. The reversible assumption was only used in characterizing the second eigenvalue in 3.1. The following is an analog of 3.1 for non-reversible Markov chains using Chung’s definition of the second eigenvalue of directed graphs.
Lemma 3.3.
Let be an irreducible Markov chain. Let be an eigenvector associated to the second smallest eigenvalue of the matrix . Define the reweighted eigenvector , with . Define the truncated vector . Then
With this lemma, we can follow the proofs of Theorem 1.5 and Theorem 1.3 verbatim as in above, by defining level sets using the truncated vector and computing the expected numerator and denominator and so on.
This concludes the proof of Theorem 3.2. We will prove 3.3 in the next subsection.
3.3 Proofs of Auxiliary Lemmas
In this subsection, we prove 3.1 and 3.3. The proofs are standard but we include them for completeness.
Proof of 3.1.
For , we define . By definition of the second eigenvector, .
For , note that , as
and thus
where the last equality uses that . It follows that
The denominator is the same as the denominator in the statement. It remains to check that the numerator is also the same as the numerator in the statement. By direct calculation,
where the second equality uses and the third equality uses reversibility which gives for all , to get . ∎
Lemma 3.4 ([Chu05]).
Let be a strongly connected directed graph and be its stationary distribution. The second smallest eigenvalue of the directed Laplacian satisfies
Suppose is an eigenvector of associated with eigenvalue . Then, for the reweighted eigenvector , for all ,
Proof of 3.3.
We claim that the truncated vector satisfies
for all . Indeed, for such that ,
where the second equality is by the fact above and the last inequality is by for all due to truncation. For such that , the inequality holds trivially because
as for all by truncation. Thus the claim follows. Multiplying both sides of the claim by and then summing over all gives
This is equivalent to the statement where the sum is over pairs with . ∎
4 Counterexamples
In this section, we prove Theorem 1.4 by constructing a family of counterexamples and bounding their second eigenvalues and value. The construction is simple.
Definition 4.1 (Counterexamples).
Let be a graph with vertex set . For each , the edge weight is
where is the normalizing constant to make the graph -regular.
We will prove the two claims in Theorem 1.4 about the second smallest eigenvalue and the value. First, we analyze the second smallest eigenvalue, based on the construction that is a circulant matrix.
Lemma 4.2.
For in 4.1, the second smallest eigenvalue of is
Proof.
By our construction, the graph is cyclic that for all . So the matrix is a circulant matrix of the form
where and for all . It is well-known that an circulant matrix with first row entries has eigenvalues and corresponding eigenvectors
where are the -th roots of unity for (where denotes the imaginary number).
So, the second smallest eigenvalue of corresponds to the first -th root of unity , and
We consider two cases, when is odd and is even. When is odd, note that by definition for and so we can pair up the terms in the above equation to get
Using the definition of , it follows that
Finally we use the fact that and that for as to conclude that
When is even, the proof follows along the same lines, but we need to remove the term in the sum because . However, it only contributes a term of to the sum, which is negligible. ∎
It remains to prove that . As this graph is symmetric, our intuition is that attains its minimum at the set . In this case, for each vertex ,
which implies that
Our plan was to prove that indeed attains the minimum, but we do not have such a proof. Instead, we will work on a slightly different lower bound, which satisfies a concavity property that allows us to argue that sets of consecutive vertices attain the minimum, in order to prove the lower bound. It turns out that the proof is a bit long and we will present it in the next subsection.
4.1 Proof of Lower Bound
First, we set up some notations for the proof. Let us partition the vertex set of into two sets and with . As the graph is cyclic, we can arrange the vertices in a clockwise manner and without loss of generality we assume and . Let us divide the vertices of and into contiguous sets in the cyclic representation, and denote their sizes by and for . More explicitly, for , the vertices in and are
For two disjoint subsets , let us define . Note that , so our goal is to lower bound
For two sets , let us define the contiguous block to be the block of sets from clockwise up until , possibly going around. For example, , and note that since the sets are counted clockwise.
After we set up the notations, we start with a lower bound on by a natural function, the logarithm of the size of contiguous sets, which is the “slightly different lower bound” that we mentioned before this subsection.
Lemma 4.3.
Proof.
We prove the statement for . By definition of stated above,
By the definition of in 4.1, and so
We lower bound the inner sum by an integral, so that
Now we use the following simple inequality about decreasing numbers.
Claim 4.4.
Let be positive real numbers such that . Then
Proof.
The proof is by induction. For , the claim is clear as . Suppose that the claim is true for . Let and . For the induction step, we need to show that . Since by induction, it suffices to show that , which is equivalent to . It follows from the property of decreasing sequence that , verifying the induction step. ∎
The and in the right hand side above satisfy the assumptions of the claim, and thus
We again lower bound the inner sum by an integral so that is at least
using the definition e.g. . ∎
Next, we are going to sum up the lower bounds in 4.3 to obtain a lower bound on . To write the sum nicely, we use a simple observation on the signs of the logarithm in our sum. Let us call a contiguous block odd if there are an odd number of sets in , and even otherwise. Note that the odd blocks are exactly those with the first and last sets from the same partition or , e.g. . With this definition, the lower bound on can be written as follows.
Lemma 4.5.
Using the definitions and notations in this subsection,
where the sum is over .
Proof.
We sum the inequalities in 4.3 from . On the right hand side of the inequality in 4.3, we see that all contiguous blocks starting from or are in the sum, with the odd blocks positive and even blocks negative. Thus, summing over all , every contiguous block is counted once as it is uniquely determined by the starting and ending sets, except for the whole cycle which appears once on the right hand side for every with a negative sign. ∎
To prove a lower bound on the right hand side of 4.5, the idea is to use the following concavity property.
Lemma 4.6.
For , consider the function
where the sum is over and so depends on . Then, for all positive , the function
obtained by fixing non-negative integers as the size of the other sets and as the sum of , is concave on .
Proof.
To prove concavity, we use the second derivative test, where is concave if the second derivative is non-positive. We write as , where the consists of all the log terms which contain but not , and similarly consists of all the log terms which contain only but not . The remaining terms are in , which either contain both and or none of and . Note that these terms are independent of , because if a block contains both and then its size is the same even when we change , so these terms can be ignored when we compute derivatives.
Let us focus on first. The blocks that contain but not must be of the form or for some set . Let denote the parity of the block . Note that the parity of and are different, and so
where the sum is over . In the special case when , we violate our own notation and let in this proof; all other cases are still the same.
When , the sum equals zero and we are done, so assume . To see that is negative, we pair up the terms with and with indices taken modulo so that
where the inequality holds because and so each summand is negative (recall the special case that in this proof).
The function is handled analogously in view of the symmetry of the second derivative of the logarithm. This proves that is concave. ∎
With the concavity property, we can apply a simple “swapping/merging” argument to reduce to the case when there is only one contiguous set, i.e. , and then finish the proof.
By concavity, the function attains its minimum at one of the endpoints, and so
The next observation is that when one set has size zero, we can merge the two adjacent sets in the same partition into one. More formally, let without loss of generality, we claim that
To see this, note that and have the same values but they have different signs so the terms involving them cancel out each other, and similarly the terms involving and cancel out each other. Therefore, in , there are no terms ending with or and no terms beginning with or , and all the remaining terms have a one-to-one correspondence with the terms in .
This reduces by one. Repeating the same argument until , we see that
and thus
where we used that is upper bounded by an absolute constant.
It remains to lower bound the right hand side. Since , it follows that , and so
where the last inequality is because is decreasing for and for the last inequality clearly holds when is large enough. This concludes the proof of Theorem 1.4.
Concluding Remarks and Open Questions
We believe that the same analysis of can be extended to other generalizations of Cheeger’s inequality in [LGT12, KLL+13], and also to the directed edge conductance using the recent notions of reweighted eigenvalues in [OTZ22, KLT22, LTW23, LTW24]. We leave it as an open question to find a counterexample where the transition matrix is the simple random walk matrix of a graph.
Acknowledgements
We thank Christian Houdré and Prasad Tetali for their encouragement and the anonymous reviewers for their helpful comments.
References
- [Alo86] Noga Alon. Eigenvalues and expanders. Combinatorica, 6:83–96, 1986.
- [AM85] Noga Alon and Vitali Milman. , isoperimetric inequalities for graphs, and superconcentrators. Journal of Combinatorial Theory, Series B, 38(1):73–88, 1985.
- [Che70] Jeff Cheeger. A lower bound for the smallest eigenvalue of the laplacian. In Problems in Analysis, pages 195–199. Princeton University Press, 1970.
- [Chu05] Fan Chung. Laplacians and cheeger inequality for directed graphs. Annals of Combinatorics, 9:1–19, 2005.
- [CS16] Deeparnab Chakrabarty and C. Seshadhri. An monotonicity tester for boolean functions over the hypercube. SIAM Journal on Computing, 45(2), 2016.
- [EG22] Ronen Eldan and Renan Gross. Concentration on the boolean hypercube via pathwise stochastic analysis. Inventiones Mathematics, 230:935–994, 2022.
- [EKLM22] Ronen Eldan, Guy Kindler, Noam Lifshitz, and Dor Minzer. Isoperimetric inequalities made simpler. Technical report, ArXiv preprint, 2204.06686, 2022.
- [HT04] Christian Houdré and Prasad Tetali. Isoperimetric invariants for product Markov chains and graph products. Combinatorica, 24:359–388, 2004.
- [KLL+13] Tsz Chiu Kwok, Lap Chi Lau, Yin Tat Lee, Shayan Oveis Gharan, and Luca Trevisan. Improved cheeger’s inequality: Analysis of spectral partitioning algorithms through higher order spectral gap. In Proceedings of the 45th Annual Symposium on Theory of Computing (STOC), pages 11–20, 2013.
- [KLL17] Tsz Chiu Kwok, Lap Chi Lau, and Yin Tat Lee. Improved cheeger’s inequality and analysis of local graph partitioning using vertex expansion and expansion profile. SIAM Journal on Computing, 46(3):890–910, 2017.
- [KLT22] Tsz Chiu Kwok, Lap Chi Lau, and Kam Chuen Tung. Cheeger inequalities for vertex expansion and reweighted eigenvalues. In Proceedings of the 62nd IEEE Annual Symposium on Foundations of Computer Science (FOCS), pages 366–377, 2022.
- [KMS18] Subhash Khot, Dor Minzer, and Muli Safra. On monotonicity testing and boolean isoperimetric-type theorem. SIAM Journal on Computing, 47(6), 2018.
- [LGT12] James R. Lee, Shayan Oveis Gharan, and Luca Trevisan. Multi-way spectral partitioning and higher-order cheeger inequalities. In Proceedings of the 44th Annual Symposium on Theory of Computing (STOC), pages 1117–1130, 2012.
- [LP17] David Levin and Yuval Peres. Markov chains and mixing times, volume 107. American Mathematical Society, 2017.
- [LTW23] Lap Chi Lau, Kam Chuen Tung, and Robert Wang. Cheeger inequalities for directed graphs and hypergraphs using reweighted eigenvalues. In Proceedings of the 55th Annual ACM Symposium on Theory of Computing (STOC), pages 1834–1847, 2023.
- [LTW24] Lap Chi Lau, Kam Chuen Tung, and Robert Wang. Fast algorithms for directed graph partitioning using flows and reweighted eigenvalues. In Proceedings of the 2024 ACM-SIAM Symposium on Discrete Algorithms, SODA 2024, pages 591–624, 2024.
- [Mar77] Grigory Margulis. Probabilistic characteristics of graphs with large connectivity. Problemy Peredachi Informatsii, 10(2):101–108, 1977.
- [MP05] Ben Morris and Yuval Peres. Evolving sets, mixing and heat kernel bounds. Probability Theory and Related Fields, 133:245–266, 2005.
- [OTZ22] Sam Olesker-Taylor and Luca Zanetti. Geometric bounds on the fastest mixing Markov chain. In the 13th Innovations in Theoretical Computer Science Conference (ITCS 2022), 2022.
- [Tal93] Michel Talagrand. Isoperimetry, logarithmic sobolev inequalities on the discrete cube, and margulis’ graph connectivity theorem. Geometric and Functional Analysis, 3(3):295–314, 1993.
- [Yos16] Yuichi Yoshida. Nonlinear laplacian for digraphs and its applications to network analysis. In Proceedings of the 9th ACM International Conference on Web Search and Data Mining (WSDM), pages 483–492, 2016.