Isoperimetric Inequalities Made Simpler
Abstract
We give an alternative, simple method to prove isoperimetric inequalities over the hypercube. In particular, we show:
-
1.
An elementary proof of classical isoperimetric inequalities of Talagrand, as well as a stronger isoperimetric result conjectured by Talagrand and recently proved by Eldan and Gross.
-
2.
A strengthening of the Friedgut junta theorem, asserting that if the -moment of the sensitivity of a function is constant for some , then the function is close to a junta. In this language, Friedgut’s theorem is the special case that .
1 Introduction
1.1 Isoperimetric Inequalities Over the Hypercube
Isoperimetric inequalities are fundamental results in mathematics with a myriad of applications. The main topic of this paper is discrete isoperimetric inequalities, and more specifically isoperimetric inequalities over the hypercube . Classical results along these lines are due to Harper, Bernstein, Lindsey and Hart [Har66, Ber67, Lin64, Har76], Harper [Har66], Margulis [Mar74], Talagrand [Tal93, Tal97, Tal94] and Kahn, Kalai and Linial [KKL88].
All of these theorems discuss various measures of boundaries and relations between them for Boolean functions. For a function and a point , we define the sensitivity of at , denoted by , to be the number of coordinates such that . The vertex boundary of a function , denoted by , is the set of all ’s such that , and the edge boundary of is defined to be the set of edges of the hypercube whose endpoints are assigned different values by . The most basic isoperimetric result related to Boolean functions, known as Poincaré’s inequality, asserts that , i.e. the edge boundary of a function cannot be too small.555We remark that the term “Poincaré’s inequality” comes from the theory of functional inequalities and Markov chains and refers to a much more general result, see for example [KT14, Section 2]. In the case of graphs, it asserts that for a graph , if the first non-trivial eigenvalue of its Laplacian is at least , then . In our example, the Laplacian is where is the adjacency matrix of the Boolean hypercube, and is . The edge isoperimetric inequality for the hypercube, due to Harper, Bernstein, Lindsey and Hart [Har66, Ber67, Lin64, Har76], is a strengthening of that statement, asserting that for a fixed value of , the functions that minimize the edge boundary are indicator functions of initial segments in lexicographical orderings of the hypercube. In particular, this result implies that if is a non-constant Boolean function, then (here and throughout, means the logarithm function with base ). The vertex isoperimetric inequality, due to Harper [Har66], asserts that among all Boolean function of a given measure, the functions that minimize the size of the vertex boundary are the indicator functions of initial segments in simplicial orderings of the hypercube. Margulis [Mar74] proved a result that combines the notion of edge boundary and vertex boundary, showing that for all functions one has that .
Talagrand considered a different notion of boundary for a function , namely the quantity , which we will refer to as the Talagrand boundary of . Here and throughout, the distribution over will be uniform over the hypercube . Talagrand proved a number of results about this quantity [Tal93, Tal97], the most basic of which asserts that . Using a simple application of the Cauchy-Schwarz inequality one sees that the result of Talagrand implies Margulis’ theorem. Talagrand then went on and strengthened the result for functions with small variance, proving the following theorem:
Theorem 1.1 (Talagrand [Tal93]).
For all non-constant Boolean functions it holds that
We remark that Talagrand’s result is in fact more general and slightly stronger. First, Talagrand proves that the above inequality holds for the -biased measure of the cube if one inserts a factor of to the right hand side. The result and techniques of the current paper also generalize to the -biased measure in a straight-forward manner, and we focus on the case that for simplicity. Second, Talagrand considers a variant of the quantity wherein is defined to be for such that . The results we state below also hold with this stronger definition.
While Theorem 1.1 can be seen to be tight up to constant factors (as can be seen by considering sub-cubes), it leaves something to be desired: it is incomparable to yet another prominent isoperimetric result – the KKL theorem [KKL88]. The KKL Theorem is concerned with the influence of coordinates on a function. For a coordinate , the influence of is defined by
In words, it is the fraction of edges along direction that are in the edge boundary of . The KKL theorem [KKL88] asserts that any function must have a variable that is at least somewhat influential; more quantitatively, they proved that .
Since the isoperimetric results of Talagrand and of KKL look very different (as well as the techniques that go into their proofs), Talagrand asked whether one can establish a single isoperimetric inequality that is strong enough to capture both of these results simultaneously. In [Tal97], Talagrand conjectures such a statement, and makes some progress towards it. Namely, he shows that there exists such that for any non-constant function it holds that
(1) |
for some absolute constant . This inequality can be seen to imply a weaker result along the lines of the KKL theorem,666Namely that there is such that if , . but is not enough to fully recover it. Talagrand conjectured that the above inequality holds for , a statement that is strong enough to directly imply the KKL theorem.
1.2 Our Results
We show a unified and simple Fourier analytic approach, based on random restrictions, to establish the results mentioned above, as well as a strengthening of Friedgut’s junta theorem [Fri98]. Our results are:
-
1.
Classical isoperimetric inequalities. We give Fourier analytic proofs of the inequalities above by Margulis, Talagrand and Eldan-Gross, as well as of robust versions of them. Our proof technique also unravels a connection between these isoperimetric inequalities and other classical results in the area, such as the Friedgut junta theorem [Fri98] and Bourgain’s tail theorem [Bou02]. The fact that such purely Fourier analytic proof exists is quite surprising to us, as the quantity does not seem to relate well to Fourier analytic methods (and indeed, the proof of Talagrand proceeds by induction, whereas the proof of Eldan and Gross uses stochastic calculus).
-
2.
Strengthened junta theorems. We establish a strengthening of Friedgut’s theorem in which the assumption that the average sensitivity of is small, i.e. that is small, is replaced by the weaker condition that is small, for slightly larger than .
In the rest of this introductory section, we describe our results in more detail.
1.3 Classical Isoperimetric Inequalities and Robust Versions
We give simpler proofs for the results below, which are due to Talagrand [Tal93] and Eldan-Gross [EG].
Theorem 1.2.
There exists an absolute constant , such that for all non-constant functions we have that
Theorem 1.3.
There exists an absolute constant , such that for all non-constant functions we have that
Our technique is more general, and can be used to prove more robust versions of Theorems 1.2, 1.3. For a -coloring of the edges of the hypercube , define the red sensitivity as if , and otherwise is the number of sensitive edges adjacent to that are red. Similarly, define the blue sensitivity as if and otherwise is the number of sensitive edges adjacent to that are blue. Then, defining
we have that the left hand side of Theorems 1.2, 1.3 can be replaced by for any -coloring . Such results can be used to prove the existence of nearly bi-regular structures in the sensitivity graph of which are sometimes useful in applications (for example, in [KMS18] such structures in the directed sensitivity graph are necessary to construct monotonicity testers). To get some intuition, consider the bipartite sub-graph of the hypercube consisting of sensitive edges of , and suppose that we want to find a large sub-graph of it which is nearly regular, in the sense that the degrees of one side are between and , while the degrees of the other side are at most for some . In that case, taking the coloring that assigns to all edges red or the coloring that assigns all of the edges blue (which amount to Talagrand’s original formulation) gives us information about the typical degrees of vertices with and vertices with respectively. It doesn’t give us any information about the possible relation between the degrees of such ’s and such ’s, though. Using the coloring appropriately, one may establish a better balance between these two degrees and get useful information on the relation between them. Establishing such robust results using the techniques of [EG] is more challenging.
1.4 New Junta Theorems
Next, we study connections between Talagrand-like quantities such as , and the notion of juntas. For a point and a subset , we denote by the point in corresponding to the value of on coordinates .
Definition 1.4.
A function is called a -junta if there exists of size at most , and such that for all .
We say a function is -close to a -junta if there is a -junta such that .
On the one hand, it is possible that the quantity is constant while the function is very far from being a junta, as can be seen by taking to be the majority function. On the other hand, if we look at the expected sensitivity – that is – a well known result of Friedgut [Fri98] asserts that if , then is close to junta. More precisely, Friedgut’s result asserts that if , then for all , is -close to a -junta. In light of this contrast, it is interesting to ask what happens if an intermediate moment of the sensitivity, say the th moment for , is constant.
We prove a strengthening of Friedgut’s junta theorem [Fri98] addressing this case, morally showing that for any , bounded -moment of the sensitivity implies closeness to junta. More precisely:
Theorem 1.5.
For all there is such that the following holds. For all and , and , if is a function such that , then is -close to a -junta, where .
Theorem 1.5 is tight, and can be seen as a generalization of Friedgut’s Junta Theorem (corresponding to the case that ) that is somewhat close in spirit to Bourgain’s Junta Theorem. Indeed, for all the tribes function shows this to be tight. Here, the tribes function is the function in which we take a partition of and define if there is such that , and otherwise. We take of equal size and the probability that is roughly ; the correct choice of parameters is and where . For this function one has . Indeed, as with constant probability we have the lower bound, and the upper bound follows by Jensen’s inequality and the fact that the total influence of is .
As a consequence, Theorem 1.5 also implies a version of the KKL theorem for th moments of sensitivity:
Corollary 1.6.
For all , there is such that the following holds for all . Suppose that a function satisfies that , then there exists such that
Proof.
Assume without loss of generality that (otherwise we work with ). Take and apply Theorem 1.5 to get that is -close to a function which is a -junta for , and let be the set of coordinates that depends on. Choose uniformly, and take where and is sampled uniformly from . Then ; also, and hold with probability at least , hence we conclude that
On the other hand, write and consider a path where , and may differ only on coordinate . Then by the union bound,
and the result follows. ∎
We note that the KKL theorem is this statement for , and the version for is only stronger than the KKL theorem as always .
Remark 1.7.
We remark that a naive application of our method establishes a sub-optimal dependency between the parameters and in Theorem 1.5, but gives an interesting relation to the noise sensitivity theorem of Bourgain [Bou02]. More precisely, in Corollary 3.11 we show that if satisfies , then must be noise stable. More specifically, for , we define the -correlated distribution by taking uniformly, and for each independently, setting with probability and otherwise resampling uniformly. In this language, we show that if , and , then for all ,
From this, one may establish a weaker variant of Theorem 1.5 by appealing to Bourgain’s result [Bou02] (see also [KKO18]). The version of that theorem from [KKO18, Theorem 3.5] asserts that if the Fourier weight of a function above level is at most , then the function is -close to a -junta. To use this result, one may observe that the noise stability bound above implies that for , the Fourier weight of above level is at most , hence by [KKO18, Theorem 3.5] is -close to a -junta for all . In particular, setting , we get that is -close to a junta.
1.5 Proof Overview
In this section we give an outline of the proof of Theorems 1.2 and 1.3. Both results rely on the same simple observation, that, once combined with off-the-shelf Fourier concentration results, finishes the proof.
Given , we consider its Fourier coefficients as well as its gradient defined as . We first observe that the square root of the sensitivity of at is nothing but , and also that looking at the absolute value of the th entry in , namely at the function , one has that its average is at least the absolute value of the singleton Fourier coefficient . This suggests that if has significant singleton Fourier coefficients then the magnitude of the gradient should be large, and therefore the expected root of the sensitivity should also be large. Indeed, by a simple application of the triangle inequality one gets that
which is equal to . As the sum of all the squares of all of the Fourier coefficients is at most , we get that . In words, via a simple application of the triangle inequality we managed to show a lower bound of the quantity we are interested in by the level Fourier weight of the function . This is already a rather interesting isoperimetric consequence, albeit rather weak.
We next show how to bootstrap this basic result into Theorems 1.2 and 1.3 via random restrictions. To get some intuition, suppose that we knew our function has significant weight on level (where ); can we utilize the above logic to say anything of interest about lower bounds of ? The answer is yes, and one way to do that is via random restrictions. If has considerable weight on level , we choose a subset of coordinates by independently including each coordinate in it with probability and then restrict the coordinates inside to a uniformly chosen value . In that case, we expect the weight of to collapse from level to roughly level , which by the choice of the parameter is roughly the first Fourier level. At that point we could apply the above logic on the restricted function and get some lower bound for the expected root of the sensitivity of the restricted function. But how does this last quantity relate to ?
Thinking about it for a moment, for a fixed point , if we choose randomly as above and fix all coordinates inside , then only roughly of the coordinates sensitive on remain alive, hence we expect the sensitivity to drop by a factor of as a result of this restriction. Indeed, it can be shown that under random restrictions, the quantity drops by a factor of . Thus, combining this observation with the above reasoning suggests that we should get a lower bound of
and indeed this is the bound we establish.
2 Preliminaries
Notation.
We will use big notations throughout the paper. Sometimes, it will be easier for us to use and notations; when we say that we mean that there is an absolute constant such that . Analogously, when we say that we mean that there is an absolute constant such that .
2.1 Fourier Analysis over the Hypercube
We consider the space of real-valued functions , equipped with the inner product . It is well-known that the collection of functions defined by forms an orthonormal basis, so any function may be written as in a unique way, where . We also consider norms for , that are defined as .
For , we denote by the vector resulting from dropping the th coordinate of . We denote by the -bit vector which is on the th coordinate and is equal to on any other coordinate.
Definition 2.1.
The derivative of along direction is defined by . The gradient of , , is defined as .
Note that for a Boolean function , for each , if and only if the coordinate is sensitive on , and in this case its value is either or . Thus, , and it will be often convenient for us to use the -norm of the gradient of in place of .
Fact 2.2.
Let , and let . Then . In particular, we have that .
Next, we define the level Fourier weight of a function , the approximate level weight of a function and the Fourier weight of up to/ above level .
Definition 2.3.
The level Fourier weight of is defined to be . The approximate level weight of a function is defined to be . The weight of up to level is defined as , and the weight of above level is defined as .
For a function , we also define the level at most part of , denoted by , as . We note that by Parseval’s equality we have that .
2.2 Random Restrictions
For a function , a set of coordinates and , we define the restriction of according to as the function defined by
We refer to the coordinates of as “alive”, and refer to the coordinates of as fixed. We will often be interested in random restrictions of a function . By that, we mean that the set is chosen randomly by including in it each independently with some probability (which is specified), and is chosen uniformly. We have the following fact, which establishes a relation between the approximate level weight of and the level weight of a random restriction of in which each is included in with probability (see [KKO18, Fact 4.1] and [KKK+, Section 3.1]).
Fact 2.4.
Let and , let be a random restriction where each is included in with probability , and is chosen uniformly. Then .
Proof.
For a fixed and , we have that . Thus,
where the last transition follows by pushing the expectation over inside and using Parseval’s equality. It follows that
and as for all whose size is between and and the rest of the summands are all non-negative, the proof is concluded. ∎
2.3 Bonami’s Lemma
The degree of a function is defined as . We will need the following consequence of the Bonami-Beckner-Gross hypercontractive inequality [Bec75, Bon70, Gro75] (see also [O’D14]), showing that the -norm of a low-degree function is comparable to its -norm.
Theorem 2.5.
Let be a function of degree at most . Then .
3 Isoperimetric Inequalities on the Hypercube
In this section, we prove Theorems 1.2 and 1.3. We begin with the following very crude bound on the Talagrand boundary, and then improve it using random restrictions.
Lemma 3.1.
Let . Then .
Proof.
By the triangle inequality and Fact 2.2 we have
Lemma 3.2.
Let , and . Then .
Proof.
Let be a random restriction for , where each is included in independently with probability , and is sampled uniformly. Fix , and note that
Therefore,
(2) |
On the other hand, for each and we may apply Lemma 3.1 on to get that
Taking expectation over and we get that
where the last inequality is by Fact 2.4. Combining this with equation (2) finishes the proof. ∎
Lemma 3.3.
Let , and . Then .
Proof.
3.1 Proof of Theorem 1.2
If , then it is enough to prove a lower bound of . Applying Lemma 3.2 on ’s that are powers of between and , we get that
and by geometric sum, the left hand side is .
3.2 Proof of Theorem 1.3
Denote ; we need the following result due to [Tal96, BKS99, KK13] (the precise version we use is due to Keller and Kindler [KK13]).
Theorem 3.4.
There are , such that for all we have that
We may now give the proof of Theorem 1.3.
Proof of Theorem 1.3.
We remark that Theorem 1.3 implies a stability result for the KKL theorem. Namely, if all influences of a Boolean function are at most , then must have constant-size vertex boundary. More precisely:
Corollary 3.5.
For any there are such that the following holds. Suppose that for we have that . Then
Proof.
Note that by assumption, , so Theorem 1.3 implies that for some depending only on . Also, we have that
It follows from the Paley-Zygmund inequality that
and the claim is proved for and . ∎
3.3 Extensions
3.3.1 Other -norms
Lemma 3.1 applies not only in the case of -norms, but rather for any norm:
Lemma 3.6.
Let and . Then .
Proof.
By Jensen’s inequality and Lemma 3.1 we have
As in Lemma 3.2, the previous lemma quickly leads to the following conclusion.
Lemma 3.7.
Let , and . Then .
Using the same argument as before and replacing invocations of Lemma 3.3 with Lemma 3.7, one may conclude variants of Theorems 1.2 and 1.3 in which square roots are replaced with any power . Below are precise statements.
Theorem 3.8.
There exists an absolute constant , such that for all non-constant functions and for all we have that
Proof.
Theorem 3.9.
There exists an absolute constant , such that for all non-constant functions and for all we have that
3.3.2 Talagrand Boundary and Noise Sensitivity
We next demonstrate the connection between having small Talagrand-type boundary and being noise stable, a notion that we define next.
For a parameter and a point , the distribution of -correlated inputs with , denoted as , is defined by the following randomized process. For each independently, set with probability , and otherwise sample uniformly. The operator , known as the noise operator with noise rate , is defined as
Definition 3.10.
For , the noise stability of with noise rate is defined as .
The relation established in Lemma 3.7 between the Fourier weight of and for implies a relation between the latter quantity and noise stability, as follows:
Corollary 3.11.
There exists an absolute constant , such that for any and for any function satisfying , we have that .
Proof.
Remark 3.12.
Inspecting the proof of Corollary 3.11, we see that if is bounded away from , say , then the term may be removed. Indeed, this is true because in the end of the proof, we can replace the upper bound by the upper bound that holds if is bounded away from .
3.4 Robust Versions
Fix a -coloring . Let be the Boolean vector whose th coordinate is if and is a red sensitive edge, and let be the Boolean vector whose th coordinate is if and is a blue sensitive edge. Analogously to Lemma 3.1 we have:
Lemma 3.13.
Let , and be any coloring. Then
Proof.
By the triangle inequality , and looking at the vector we see its th entry is . Similarly, and looking at the vector we see its th entry is . It follows that
Let be the indicator that is a red influential edge and , and be the indicator that is a blue influential edge and . Then
Hence
Secondly, for any , coloring and , choosing randomly by including each coordinate independently with probability , one has
and similarly for blue sensitivity. This implies the analog of (2) for red and blue sensitivity, and repeating the argument in Lemmas 3.2, 3.3 we get that for any coloring . The robust versions of Theorems 1.2 and 1.3 now follow as we have shown that where we can either take or .
4 Improved Junta Theorems
4.1 Proof of Theorem 1.5
In this section, we prove Theorem 1.5. As explained in the introduction, if one does not care about getting tight dependency between the junta size and the parameter , then one may use Corollary 3.11 and Bourgain’s noise sensitivity theorem [Bou02, KKO18] and get a junta theorem with slightly worse parameters than those stated in Theorem 1.5. The main goal of this section is to get an optimal dependency of on .
For the proof, we will need the notion of noisy influences of a function , and we first generalize the notion of influences to real valued functions.
Definition 4.1.
Let be a function, and let be a coordinate. We define the influence of on as , where we recall that is the discrete derivative of along direction .
Definition 4.2.
The -noisy influence of a coordinate for is defined to be .
Fact 4.3.
[[O’D14]] For all we have that . ∎
Let be two absolute constants, and we think of as much larger than . Given with , we denote , . Define the candidate set of junta coordinates as
i.e. is the set of coordinates whose noisy influence at is somewhat large. Our task is to show that is -close to a junta, i.e. that
(3) |
and the rest of the proof is devoted towards this goal. We remark that the junta is similar in spirit to the junta in Bourgain’s original proof [Bou02], which is easier to see in [KKO18, Section 4] and in particular in [KKO18, Theorem 4.20]. Indeed, both here and in [KKO18, Section 4], the junta is taken to be the set of coordinates with significant noisy influence. There are many other similarities between the argument presented therein and the argument here. First, in both cases one considers a dyadic sort of partitioning of the Fourier coefficients: in [KKO18, Section 4] it is achieved via considering the difference in the noise sensitivity of at different noise rates (see [KKO18, Theorem 4.16]). In the current proof, we perform explicit dyadic partitioning and also accompany it by applications of suitable noise operators (the operators , and below). Second, both proofs also use some relation between the noise sensitivity / moments of the sensitivity and the level weight of the function (and its restrictions). While the technical execution of the proofs is different, it is hard to pinpoint the high level difference between them. For example, the “harsh” degree truncations in [KKO18] have to be replaced by the “softer” type of degree truncations that are achieved by the operators , and (see also the parameter ) to make the argument go through. Additionally, towards the end of the argument we are able to bound the sum of noisy influences in a non-trivial way, which ultimately allows us to choose as opposed to (which would have resulted in slightly worse bounds as in [KKO18, Theorem 4.20]); see Claim 4.5.
Towards proving (3), we partition the contribution of different scales of to the left hand side, and for each define
Thus, we may write
(4) |
The rest of the proof is devoted to bounding and for each . From Lemma 3.7 it follows that
(5) |
for sufficiently large . Next, we bound . Fix , and consider the operator , i.e. the noise operator that applies noise on coordinates outside , and noise on coordinates of . Note that
so it suffices to bound . We partition according to contribution of each , getting:
Thus, we have that , and the rest of the proof is devoted to upper bounding the right hand side. Define the operators and as
We establish the following claim using random restrictions and hypercontractivity:
Claim 4.4.
For all , we have that
where is an absolute constant.
Proof.
Deferred to Section 4.2. ∎
Given Claim 4.4, one may pull out outside the sum, and bound the sum on the rest of the noisy influences by to get a bound of the form . This idea is good enough to establish a junta theorem, yet it gives worse parameters.777One has to take , as opposed to as we have taken. We remark that if one only cares about getting a larger junta of size , the proof greatly simplifies. Instead, to get the tight result with respect to the junta size, we will need a more careful bound on the sum of noisy influences outside . Specifically, we use an improved bound of the sum of noisy influences that relates them to . To state this bound we define
Claim 4.5.
.
Proof.
By definition,
Partitioning the last sum according to and dyadically partitioning so that , we have that the last expression is at most
Corollary 4.6.
There exists an absolute constant such that for all ,
We may now give an upper bound on .
Claim 4.7.
For sufficiently large absolute constant and sufficiently large in comparison to , we have that .
Proof.
We can now prove Theorem 1.5.
4.2 Proof of Claim 4.4
For the proof of Claim 4.4 we need a version of the hypercontractive inequality with the noise operator, as follows:
Lemma 4.8.
[[O’D14]] Let , and . Then for all we have .
We also need the following claim, asserting that noise only decreases Talagrand’s boundary.
Claim 4.9.
For and any noise operator we have .
Proof.
and the result follows from the triangle inequality. ∎
Take to be a random restriction so that each is included in independently with probability , and consider . We calculate the contribution of to the level weight of a restriction of :
Also,
where and . Using for and we have
and so
Hence our goal of upper bounding the left hand side in Claim 4.4 reduces to upper bounding the contribution of to the level of random restrictions of . For define
Then
Bounding the contribution from .
For the first sum, we have
In the last expression we have times the -norm of the vector . Thus, using the triangle inequality in a similar way to Lemma 3.1, we may bound the last expression from above by
(6) |
Taking expectation, we get that
Fix , and note that . It follows by Jensen’s inequality that
Therefore (6) is upper bounded by . Using Claim 4.9, we have that this may be upper bounded by .
Bounding the contribution from .
We have
and we next take expectation over :
(7) |
Define the operator where and , and note that by the semi-group property of the noise operator (namely, the fact that ) we may write . Thus,
Thus, using hypercontractivity, i.e. Lemma 4.8, we have that
(8) |
In the last inequality, we used Parseval’s equality and that fact that . Note that
so plugging (4.2) into (7) yields that
Taking expectation over now gives that
∎
5 Acknowledgements
We thank an anonymous for many helpful comments and corrections to an earlier version of the paper which led to considerable improvements of the paper in all aspects.
References
- [Bec75] William Beckner. Inequalities in Fourier analysis. Annals of Mathematics, 102:159–182, 1975.
- [Ber67] Arthur J Bernstein. Maximally connected arrays on the n-cube. SIAM Journal on Applied Mathematics, 15(6):1485–1489, 1967.
- [BKS99] Itai Benjamini, Gil Kalai, and Oded Schramm. Noise sensitivity of Boolean functions and applications to percolation. Publications Mathématiques de l’Institut des Hautes Etudes Scientifiques, 90(1):5–43, 1999.
- [Bon70] Aline Bonami. Étude des coefficients de Fourier des fonctions de . Annales de l’Institut Fourier, 20(2):335–402, 1970.
- [Bou02] Jean Bourgain. On the distribution of the fourier spectrum of boolean functions. Israel Journal of Mathematics, 131(1):269–276, 2002.
- [EG] Ronen Eldan and Renan Gross. Concentration on the Boolean hypercube via pathwise stochastic analysis. In STOC 2020.
- [Fri98] Ehud Friedgut. Boolean functions with low average sensitivity depend on few coordinates. Combinatorica, 18(1):27–35, 1998.
- [Gro75] Leonard Gross. Logarithmic Sobolev inequalities. American Journal of Mathematics, 97(4):1061–1083, 1975.
- [Har66] Lawrence H Harper. Optimal numberings and isoperimetric problems on graphs. Journal of Combinatorial Theory, 1(3):385–393, 1966.
- [Har76] Sergiu Hart. A note on the edges of the n-cube. Discrete Mathematics, 14(2):157–163, 1976.
- [KK13] Nathan Keller and Guy Kindler. Quantitative relation between noise sensitivity and influences. Combinatorica, 33(1):45–71, 2013.
- [KKK+] Esty Kelman, Subhash Khot, Guy Kindler, Dor Minzer, and Muli Safra. Theorems of KKL, Friedgut, and Talagrand via random restrictions and log-sobolev inequality. In ITCS 2021.
- [KKL88] Jeff Kahn, Gil Kalai, and Nathan Linial. The influence of variables on boolean functions (extended abstract). In 29th Annual Symposium on Foundations of Computer Science, White Plains, New York, USA, 24-26 October 1988, pages 68–80, 1988.
- [KKO18] Guy Kindler, Naomi Kirshner, and Ryan O’Donnell. Gaussian noise sensitivity and fourier tails. Israel Journal of Mathematics, 225:71–109, 2018.
- [KMS18] Subhash Khot, Dor Minzer, and Muli Safra. On monotonicity testing and Boolean isoperimetric-type theorems. SIAM J. Comput., 47(6):2238–2276, 2018.
- [KT14] Christos P Kitsos and Thomas L Toulias. Inequalities for the Fisher’s information measures. Handbook of Functional Equations: Functional Inequalities, pages 281–313, 2014.
- [Lin64] John H Lindsey. Assignment of numbers to vertices. The American Mathematical Monthly, 71(5):508–516, 1964.
- [Mar74] Grigorii Aleksandrovich Margulis. Probabilistic characteristics of graphs with large connectivity. Problemy peredachi informatsii, 10(2):101–108, 1974.
- [O’D14] Ryan O’Donnell. Analysis of boolean functions. Cambridge University Press, 2014.
- [Tal93] Michel Talagrand. Isoperimetry, logarithmic Sobolev inequalities on the discrete cube, and Margulis’ graph connectivity theorem. Geometric & Functional Analysis GAFA, 3(3):295–314, 1993.
- [Tal94] Michel Talagrand. On Russo’s approximate zero-one law. The Annals of Probability, 22(3):1576–1587, 1994.
- [Tal96] Michel Talagrand. How much are increasing sets positively correlated? Combinatorica, 16(2):243–258, 1996.
- [Tal97] Michel Talagrand. On boundaries and influences. Combinatorica, 17(2):275–285, 1997.