The Fourier Transform of Restrictions of Functions on the Slice
Abstract
This paper considers the Fourier transform over the slice of the Boolean hypercube. We prove a relationship between the Fourier coefficients of a function over the slice, and the Fourier coefficients of its restrictions. As an application, we prove a Goldreich-Levin theorem for functions on the slice based on the Kushilevitz-Mansour algorithm for the Boolean hypercube.
1 Introduction
Boolean functions, or functions over , are central to computer science. One tool that has found many uses is Fourier analysis, which involves defining an orthonormal basis separate from the standard basis, but with many useful properties. One such property is that this basis consists of eigenvectors of the Boolean hypercube, the graph whose vertices are with an edge between two vectors and if they differ in exactly one coordinate. Fourier analysis over the Boolean hypercube has led to many applications, including in learning theory, property testing, pseudorandomness, voting theory, and more. For a more complete treatment, see [O’D14].
In recent years, much attention has been drawn to functions on the slice, or the subset of that consists of vectors with exactly ’s for some fixed . Like in the case of the Boolean hypercube, one can also define an orthonormal basis separate from the standard basis, over functions on the slice. In particular, one can consider the eigenvectors of the Johnson graph, the graph whose vertices are elements of the slice, with an edge between two vectors and if they differ in exactly two coordinates. Thus, this basis can be seen as an analogue of the Fourier transform, but for the slice.
Such a basis was given explicitly in [Fil16] and [Sri11], and this basis has subsequently found many applications in analogues of major theorems for the Boolean hypercube, but now for the slice. These include Friedgut’s theorem [Fil16], various versions of the invariance principle [FM19, FKMW18], the Kindler-Safra theorem [KK20], and a theorem due to Nisan and Szegedy regarding juntas [FI19]. In this paper, we continue this line of work of proving analogues for the slice.
Fourier analysis on the hypercube has been particularly useful when considering restrictions of functions. In particular, if is a function, and for , we let its restriction, , be the function defined by where denotes concatenation. In particular, one can relate the Fourier coefficients of with the Fourier coefficients of .
Recall that the orthonormal basis for the Fourier transform can be indexed by subsets of the set . Then for any and any ,
(1) |
where denotes the projection onto the vector indexed by (see Corollary 3.22 in [O’D14]).
In the case of the slice, one can also define a restriction, where given and for , we let be defined by , where denotes the Hamming weight of . If contains more than ’s, then is an empty function. In this work, we show that an identity similar to Eq. (1) holds for the slice.
In the slice, the orthonormal basis for the Fourier transform is now indexed by subsets of with a certain property. Let be the collection of subsets of of size at most such that for all , when the elements of are listed in sorted order. Then the vectors in the orthogonal basis from [Fil16] and [Sri11] can be indexed by elements of . We note that a vector indexed by the set in the orthogonal basis for the slice differs significantly from the vector indexed by the same set in the orthogonal basis for the Boolean hypercube.
The main result of this paper is the following theorem, which generalizes Eq. (1) to the slice.
Theorem 1.1.
Let , and let .
Then for any and any ,
Again, denotes the projection onto the vector indexed by . For convenience, we will let if is not a subset in or is an empty function.
Eq. (1) has many applications, one of which includes the Goldreich-Levin theorem [GL89, KM93]. Informally, this theorem says that if the Fourier coefficients of a function are concentrated on just a few vectors, then the function can be learned efficiently. As an application of Theorem 1.1, we show that a similar theorem applies in the case of the slice, following the Kushilevitz-Mansour algorithm.
Theorem 1.2.
Given query access to a function and , there is a -time algorithm that with high probability outputs a list of subsets of such that
-
•
implies that
-
•
implies that
1.1 Future Work
It would be interesting to see if Theorem 1.1 leads to other analogues of theorems for the Boolean hypercube to the slice.
It would also be interesting to see to what extent the orthonormal basis for slice can be extended to high-dimensional expanders, and additionally, if there exist analogues of theorems for the Boolean hypercube on high dimensional expanders.
Similarly to how expander graphs can be thought of as sparse approximations of the complete graph, high dimensional expanders can be thought of as sparse approximations of the Johnson graph. In recent years, many constructions of high dimensional expanders have been developed [LSV05a, LSV05b, KO18, CTZ20]. The theory of constructions of high dimensional expanders have already led to many applications in theoretical computer science, including mixing times of Markov chains [ALGV19, AL20, CGM21] and coding theory [AJQ+20, DHK+21].
Unlike regular expander graphs, one has control beyond just the second eigenvalue, as shown in work by Kaufman and Oppenheim [KO20]. In particular, the spectrum of high dimensional expanders approximates the spectrum of the Johnson graph. Both Kaufman and Oppenheim, and work by Dikstein, Dinur, Filmus, and Harsha [DDFH18] give different decompositions into approximate eigenspaces. The latter allows one to obtain an analogue of the Freidgut-Kalai-Naor theorem for high dimensional expanders. It would be interesting to obtain more explicit descriptions of approximate eigenvectors in these approximate eigenspaces, that might allow one to prove more analogues for high dimensional expanders. In particular, one hope is that the structure of the orthonormal basis for the slice can used as inspiration to construct an approximately orthonormal basis for high dimensional expanders, which can then be used to prove analogues of theorems for the Boolean hypercube, but now for high dimensional expanders.
2 Preliminaries
Denote by the set and by the set or the subset of that consists of vertices with exactly ’s for some fixed .
Given a string , denote by the Hamming weight of , that is, .
As stated in the introduction, we denote by the set of subsets of of size at most such that for all , when the elements of are listed in sorted order. Note that by Bertrand’s ballot problem, there are sets in of size , and thus .
Given a vector we define the -norm by
We define the inner product for two vectors to be
It will often be convenient to index vectors by vectors rather than integers, that is, to let be a vector in for some and . It will often be convenient to identify vectors as functions, that is, to represent as , or as where in both cases, one obtains from via .
By the Cauchy-Schwarz inequality, one can relate the -norm of a vector with its -norm as follows.
Claim 2.1.
For all ,
We denote by the identity matrix.
3 An orthonormal basis of eigenvectors for the Johnson graph
Recall that the Johnson graph is defined by where and . Let be the corresponding adjacency matrix.
In this section we give with proof an orthonormal basis of eigenvectors for the Johnson graph. The claims and theorems in this section will be used in Section 4 to prove Theorem 1.1. We stress that all of the results in this section are known, and can be found in the works of Stanley [Sta88, Sta90], Filmus [Fil16], Srinivasan [Sri11], and Filmus and Lindzey [FL20]. When illustrative, however, we give the proofs of the various statements.
We start by defining a collection of matrices that will help yield the eigenvectors of .
Let be the matrix defined by
(2) |
Let .
Let
and
For simplicity, we let .
Let and . The following fact shows that and differ only by a multiple of the identity matrix.
Claim 3.1.
Similarly, one can show that differs from and by a multiple of the identity matrix. Thus, for the rest of this paper, we focus on determining an orthonormal basis of eigenvectors for and .
The following claim shows that to determine the eigenvectors of and , it is enough to determine the nullspace of for all . This will be under the assumption that has full rank when , which we will prove in Claims 3.5 and 3.6.
Claim 3.2.
Let be a vector such that . Then,
and
The following claim shows that if , then if . Thus, in the case that , one only needs to consider the nullspace of for . In fact, we will compute when , which will be useful in Section 3.2
Claim 3.3.
Let be a vector such that for . Then,
if , and is otherwise .
Proof.
We prove this using induction on . In the base case when , the Claim follows as .
Now, assume that the statement is true for . Then,
(3) | ||||
(4) |
as desired, where the third equality follows from Claim 3.2 and the fourth equality follows from the inductive hypothesis.
If , then Eq. (3) is , and it follows that whenever by the definition of . ∎
Note that the assumption that has full-rank when combined with Claims 3.2 and 3.3 yields a subspace of eigenvectors with equal eigenvalue of dimension . For a fixed and summing over all from to , this is eigenvectors, and thus, this gives a complete basis of eigenvectors.
The following claim shows that to determine an orthogonal basis of eigenvectors of and it is enough to determine an orthogonal basis for the nullspace of for all . That is, applying to two orthogonal vectors preserves orthogonality. Recall that the eigenvectors of a symmetric matrix with different eigenvalues are orthogonal, and thus it is sufficient to determine an orthogonal basis for the nullspace of for each .
Claim 3.4.
Let and be such that and for all . Then
.
Proof.
We prove this on induction on .
When , the statement is obvious.
We note that Claims 3.1, 3.2 3.3, and 3.4 can be generalized to sequentially differential posets (of which, the collection of subsets of is an example), as is done in the work of Stanley [Sta88, Sta90].
3.1 An orthogonal basis for the nullspace of
In this section, we give an orthogonal basis for the nullspace of when . This basis was found by Srinivasan [Sri11] and independently by Filmus [Fil16]. For completeness, we include the proof that this is an orthogonal basis due to Srinivasan.
Note that
if , where if , then we remove the bottom two blocks. In particular, the fact that can be written recursively in terms of and suggests that an orthogonal basis for the nullspace of the former matrix can be obtained from orthogonal bases for the nullspace of the latter matrices.
Let be a subset in . Recall that this means that when the elements are listed in sorted order as , that for all . We associate with each a vector such that . We define recursively as follows.
(5) |
We briefly give some intuition behind restricting to only those subsets in . Consider a set of size at most that is not in . Let be in sorted order, and let be the smallest integer such that . Then either and , or and . Thus, , and is not well-defined.
By Bertrand’s ballot problem, there are sets in of size . Thus, proving that is in the nullspace of and that if and , then are both enough to obtain an orthogonal basis for the nullspace of and also show it has full rank.
We first show that is in fact in the nullspace of .
Claim 3.5.
For all , .
Proof.
We prove this using induction on and .
If , then
Thus, .
If , then by the inductive hypothesis.
If , then
where the second inequality follows from the relationship between and and the inductive hypothesis, and the third inequality also follows from the inductive hypothesis. ∎
The following claim shows that if and , then and are orthogonal. Recall that eigenvectors of a symmetric matrix with different eigenvalues are orthogonal. Thus, this claim along with Claim 3.6 are enough to prove that the given basis is orthogonal.
Claim 3.6.
For , if and , then .
Proof.
We prove this using induction on .
Consider the base case of . In particular, let and let , and without loss of generality, assume that . Then
and the claim clearly holds.
For the inductive step, assume that the statement is true for all pairs and where and . Then there are three cases.
Case 1: If neither nor contain , then by the inductive hypothesis.
Case 2: If both and contain , then
where the second equality follows from Fact 3.1 and the third inequality follows from the inductive hypothesis for and , and the fact that .
Case 3: If only contains , then
where the last inequality follows from the fact that . ∎
We summarize below.
Theorem 3.7.
The set of vectors
form an orthogonal basis of eigenvectors for , the adjacency matrix of the Johnson graph.
3.2 A computation of the norm
In this section, we compute the norm of . This combined with Theorem 3.7 will yield a set of orthonormal eigenvectors that can be used to construct a Fourier transform over the slice. The results in this section can be found in [Fil16]. We give an inductive proof that follows the framework of Srinivasan [Sri11] and Filmus and Lindzey [FL20].
We start by computing the norm of .
Claim 3.8.
Let be such that . Then,
Proof.
We prove this using induction on . For the base case, if , then
and thus,
as desired.
If , then
where the second inequality follows from the inductive hypothesis.
If , then
where the third equality follows from the fact that is an eigenvector of , and the fourth equality follows from the inductive hypothesis. ∎
Finally, we can compute .
Claim 3.9.
Finally, we give a formula definition for the Fourier coefficients of a function on the slice.
4 Restrictions of functions on the slice
In this section, we state and prove our main theorem regarding a relationship between the Fourier coefficients of a function on the slice, and the Fourier coefficients of its restrictions.
Definition 4.1 (Restriction).
Let . For , and , let (where is the Hamming weight of ) be the function defined by where denotes concatenation.
We start with the following theorem.
Theorem 4.2.
Let . Then, for all ,
Additionally, if , then
Thus, if , then
and
Proof.
If , then , and , and the theorem holds. Similarly, if , then and and the theorem also holds.
If , then as if , and again, the theorem holds. If , then recall that by Claim 3.3, and thus , and again, the theorem holds.
Thus, we can assume that and are not empty functions, and that , and in particular, has a corresponding vector in the orthogonal basis for both and .
Let be defined by if and otherwise. Additionally, let be defined by if , and otherwise.
Let . Note that because , for all ,
(6) |
We start with the first inequality. We have that , as if . Thus, the right-hand side of Eq. (6) is equal to
(7) |
where the sequence of inequalities uses the fact that unless , in which case . The first equality also uses the fact that if . By Claim 3.9, Eq. (7) is equal to
as desired.
Finally, we prove the following relationship between the Fourier coefficients of and the Fourier coefficients of restrictions of .
Corollary 4.3.
Let , and let . Then for all ,
Additionally,
Proof.
The first statement follows from Theorem 4.2, and the second follows by induction. ∎
5 A Goldreich-Levin Theorem for the slice
In this section, we prove Theorem 1.2, a Goldreich-Levin theorem for the slice. Informally, this theorem states that the large Fourier coefficients of a function can be efficiently detected. With the machinery of Corollary 4.3, the algorithm and proof of correctness are essentially the same as for the Boolean hypercube. However, some care is required in estimating quantities of the form
for a given function and a given set . For completeness, we give the algorithm.
The algorithm uses a divide and conquer strategy due to Kushilevitz and Mansour [KM93].
Much of the analysis is identical to the algorithm for the Boolean hypercube. As in the case of the Boolean hypercube, the algorithm identifies a set of Fourier coefficients with large total weight, starting with the set of all Fourier coefficients. At each step, it splits each set into two, estimates the total weights of all of the subsets, and removes those subsets with small total weight. If the algorithm is able to estimate the total weight accurately, then at any given point, it will only identify different sets. (For a complete analysis, one can also refer to Section 3.5 of [O’D14]).
Thus, it is enough to show that one can estimate and in -time to error with probability at least .
Theorem 5.1.
Given query access to a function and a set for some integer , then samples are sufficient to compute up to error with probability at least .
Proof.
By Corollary 4.3, we have
(9) |
Let be a random variable over where with probability . Note that such a probability distribution is well-defined, as . Then the right-hand side of Eq. (9) can be rewritten as
For each in , let and be independent random variables over distributed according to the absolute value of the coordinates . In particular, we let the probability that be and thus we obtain
By Claim 2.1, and the fact that for all , it follows that the random variable inside the expectation is bounded above by . Thus, by a Chernoff bound, samples are sufficient to compute up to error with probability at least . ∎
References
- [AJQ+20] V. L. Alev, F. G. Jeronimo, D. Quintana, S. Srivastava, and M. Tulsiani. List decoding of direct sum codes. In Proceedings of the 2020 ACM-SIAM Symposium on Discrete Algorithms, pages 1412–1425. SIAM, Philadelphia, PA, 2020.
- [AL20] V. L. Alev and L. C. Lau. Improved analysis of higher order random walks and applications. In STOC ’20—Proceedings of the 52nd Annual ACM SIGACT Symposium on Theory of Computing, pages 1198–1211. ACM, New York, [2020] ©2020. doi:10.1145/3357713.3384317.
- [ALGV19] N. Anari, K. Liu, S. O. Gharan, and C. Vinzant. Log-concave polynomials II: High-dimensional walks and an FPRAS for counting bases of a matroid. In STOC’19—Proceedings of the 51st Annual ACM SIGACT Symposium on Theory of Computing, pages 1–12. ACM, New York, 2019. doi:10.1145/3313276.3316385.
- [CGM21] M. Cryan, H. Guo, and G. Mousa. Modified log-Sobolev inequalities for strongly log-concave distributions. Ann. Probab., 49(1):506–525, 2021. ISSN 0091-1798. doi:10.1214/20-AOP1453.
- [CTZ20] D. Conlon, J. Tidor, and Y. Zhao. Hypergraph expanders of all uniformities from Cayley graphs. Proc. Lond. Math. Soc. (3), 121(5):1311–1336, 2020. ISSN 0024-6115. doi:10.1112/plms.12371.
- [DDFH18] Y. Dikstein, I. Dinur, Y. Filmus, and P. Harsha. Boolean function analysis on high-dimensional expanders. In Approximation, randomization, and combinatorial optimization. Algorithms and techniques, volume 116 of LIPIcs. Leibniz Int. Proc. Inform., pages Art. No. 38, 20. Schloss Dagstuhl. Leibniz-Zent. Inform., Wadern, 2018.
- [DHK+21] I. Dinur, P. Harsha, T. Kaufman, I. L. Navon, and A. Ta-Shma. List-decoding with double samplers. SIAM J. Comput., 50(2):301–349, 2021. ISSN 0097-5397. doi:10.1137/19M1276650.
- [FI19] Y. Filmus and F. Ihringer. Boolean constant degree functions on the slice are juntas. Discrete Math., 342(12):111614, 7, 2019. ISSN 0012-365X. doi:10.1016/j.disc.2019.111614.
- [Fil16] Y. Filmus. An orthogonal basis for functions over a slice of the Boolean hypercube. Electron. J. Combin., 23(1):Paper 1.23, 27, 2016.
- [FKMW18] Y. Filmus, G. Kindler, E. Mossel, and K. Wimmer. Invariance principle on the slice. ACM Trans. Comput. Theory, 10(3):Art. 11, 37, 2018. ISSN 1942-3454. doi:10.1145/3186590.
- [FL20] Y. Filmus and N. Lindzey. Murali’s basis for the slice, 2020.
- [FM19] Y. Filmus and E. Mossel. Harmonicity and invariance on slices of the Boolean cube. Probab. Theory Related Fields, 175(3-4):721–782, 2019. ISSN 0178-8051. doi:10.1007/s00440-019-00900-w.
- [GL89] O. Goldreich and L. A. Levin. A hard-core predicate for all one-way functions. In D. S. Johnson, editor, Proceedings of the 21st Annual ACM Symposium on Theory of Computing, May 14-17, 1989, Seattle, Washigton, USA, pages 25–32. ACM, 1989. doi:10.1145/73007.73010.
- [KK20] N. Keller and O. Klein. A structure theorem for almost low-degree functions on the slice. Israel J. Math., 240(1):179–221, 2020. ISSN 0021-2172. doi:10.1007/s11856-020-2062-4.
- [KM93] E. Kushilevitz and Y. Mansour. Learning decision trees using the Fourier spectrum. SIAM J. Comput., 22(6):1331–1348, 1993. ISSN 0097-5397. doi:10.1137/0222080.
- [KO18] T. Kaufman and I. Oppenheim. Construction of new local spectral high dimensional expanders. In STOC’18—Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing, pages 773–786. ACM, New York, 2018. doi:10.1145/3188745.3188782.
- [KO20] T. Kaufman and I. Oppenheim. High order random walks: beyond spectral gap. Combinatorica, 40(2):245–281, 2020. ISSN 0209-9683. doi:10.1007/s00493-019-3847-0.
- [LSV05a] A. Lubotzky, B. Samuels, and U. Vishne. Explicit constructions of Ramanujan complexes of type . European J. Combin., 26(6):965–993, 2005. ISSN 0195-6698. doi:10.1016/j.ejc.2004.06.007.
- [LSV05b] A. Lubotzky, B. Samuels, and U. Vishne. Ramanujan complexes of type . volume 149, pages 267–299. 2005. doi:10.1007/BF02772543. Probability in mathematics.
- [O’D14] R. O’Donnell. Analysis of Boolean functions. Cambridge University Press, New York, 2014. ISBN 978-1-107-03832-5. doi:10.1017/CBO9781139814782.
- [Sri11] M. K. Srinivasan. Symmetric chains, Gelfand-Tsetlin chains, and the Terwilliger algebra of the binary Hamming scheme. J. Algebraic Combin., 34(2):301–322, 2011. ISSN 0925-9899. doi:10.1007/s10801-010-0272-2.
- [Sta88] R. P. Stanley. Differential posets. J. Amer. Math. Soc., 1(4):919–961, 1988. ISSN 0894-0347. doi:10.2307/1990995.
- [Sta90] R. P. Stanley. Variations on differential posets. In Invariant theory and tableaux (Minneapolis, MN, 1988), volume 19 of IMA Vol. Math. Appl., pages 145–165. Springer, New York, 1990.