A Tighter Relation Between Hereditary Discrepancy and Determinant Lower Bound
In seminal work, Lovász, Spencer, and Vesztergombi [European J. Combin., 1986] proved a lower bound for the hereditary discrepancy of a matrix in terms of the maximum over all submatrices of . We show algorithmically that this determinant lower bound can be off by at most a factor of , improving over the previous bound of given by Matoušek [Proc. of the AMS, 2013]. Our result immediately implies , for any two set systems over satisfying . Our bounds are tight up to constants when due to a construction of Pálvölgyi [Discrete Comput. Geom., 2010] or the counterexample to Beck’s three permutation conjecture by Newman, Neiman and Nikolov [FOCS, 2012].
1 Introduction
Given a matrix , the discrepancy of is . The hereditary discrepancy of is defined as , where denotes the restriction of the matrix to columns in . For a set system , and are defined to be and , where is the incidence matrix of .
In seminal work, Lovász, Spencer, and Vesztergombi [LSV86] introduced a powerful tool, known as the determinant lower bound, for bounding hereditary discrepancy:
where denotes the restriction of to rows in and columns in . In particular, they showed that for any matrix . A reverse relation was established by Matousěk [Mat13], who showed that . However, Matousěk’s bound does not match the largest known gap of between and , given by a construction of Pálvölgyi [Pál10] or the counter-example to Beck’s three permutation conjecture [NNN12].
Our main result is the following improvement over Matousěk’s bound in [Mat13].
Theorem 1.1.
Given a matrix , one can efficiently find such that .
Restricting to an arbitrary subset of the columns of , one immediately obtains the following:
Corollary 1.2.
For any matrix , .
In light of the examples in [Pál10, NNN12] where , Theorem 1.1 is tight up to constants whenever . For the case where , one cannot hope to improve the dependence on in Theorem 1.1. In particular, the set system has , and therefore . It remains an open problem, however, whether one can improve the factor in the later regime.
Hereditary discrepancy of union of set systems.
A question of V. Sós (see [LSV86]) asks whether can be estimated in terms of and , for any set systems and over . This is, however, not possible without any dependence on or , as first shown by an example of Hoffman (Proposition 4.11 in [Mat09]). This can also be seen from the examples in [Pál10, NNN12]. In [KMV05], it was shown that when contains a single set. For more general set systems, Matousěk [Mat13] proved that , where and .
Theorem 1.1 together with Lemma 4 in [Mat13] immediately imply the following improvement of this result, whose proof is the same as in [Mat13]. For and , this bound is tight up to constants.
Theorem 1.3.
Let be a system of sets on such that . Then,
Approximating hereditary discrepancy.
It was shown in [CNN11] that cannot be approximated in polynomial time for an arbitrary matrix . The more robust notion of hereditary discrepancy, however, can be approximated within a polylog factor. The best-known result in this direction is a -approximation to hereditary discrepancy via the -norm [MNT14]. When , this approximation factor is .
Our result in Theorem 1.1 suggests a potential approach of approximating hereditary discrepancy by approximating the determinant lower bound. There has been a recent line of work in approximating the maximum subdeterminant for a given matrix . For , Nikolov [Nik15] gave a -approximation; for general values of , Anari and Vuong [AV20] showed a -approximation algorithm. If these results can be strengthened to a -approximation algorithm for general values of , then together with Theorem 1.1, one would obtain the first -approximation algorithm for hereditary discrepancy when .
Overview of proof of Theorem 1.1.
We follow the approaches in [Ban10] and [Mat13]. The key notion to prove Theorem 1.1 is that of hereditary partial vector discrepancy, which is defined as follows. Given a matrix with entries for and , we consider the following SDP for a subset and a parameter :
Define the partial vector discrepancy of , denoted as , to be the smallest value of such that is feasible, and hereditary partial vector discrepancy to be the smallest such that is feasible for any subset .
Using the above definition, we show in Lemma 2.1 of Section 2.1 that the above SDP can be rounded efficiently to obtain a coloring with discrepancy at most . We then prove in Lemma 2.3 of Section 2.2 that , from which Theorem 1.1 immediately follows. We conjecture that is the same as up to constants (Conjecture 2.4).
Notations and preliminaries.
Given a matrix , its rows will be denoted by . Define to be the matrix with rows restricted to some subset and columns restricted to some , and .
Theorem 1.4 (Freedman’s Inequality, Theorem 1.6 in [Fre75]).
Consider a real-valued martingale sequence such that , and for all , where is the filtration defined by the martingale. Assume that the sequence is uniformly bounded, i.e., almost surely for all . Now define the predictable quadratic variation process of the martingale to be for all . Then for all and and any stopping time , we have
2 Proof of Theorem 1.1
2.1 The Algorithm
The main result of this subsection is the following lemma.
Lemma 2.1.
Given a matrix , there exists a randomized algorithm that w.h.p. constructs a coloring such that . This implies that .
The algorithm in Lemma 2.1 is given in Algorithm 1. This algorithm is a variant of the random walk in [Ban10], using the SDP for hereditary partial vector discrepancy.
Since Lemma 2.1 is invariant under rescaling of the matrix , we may assume without loss of generality that that . Given a coloring , we say an element is alive if . The following lemma from [Ban10] states that the number of alive elements halves after steps.
Lemma 2.2 (Lemma 4.1 of [Ban10]).
Let be an arbitrary fractional coloring with at most alive variables. Let be the fractional coloring obtained by running algorithm 1 with for steps. Then the probability that has at least alive variables is at most .
Proof of Lemma 2.1.
We first argue that after steps, no element is alive with high probability. Divide the time horizon into epochs of size . For each epoch, Lemma 2.2 states that regardless of the past, the number of alive elements decreases by at least half with probability at least . It follows that no element is alive with high probability after epochs. Note that when no element is alive for the coloring , one can round it to a full coloring without changing the discrepancy of each set by more than .
Next we prove that with high probability, the discrepancy of each row of is at most . We consider any , and denote the discrepancy of row at the end of time step . Note that and . It follows from Freedman’s inequality (Theorem 1.4) that
So by the union bound, the discrepancy of the obtained coloring is at most with high probability. This completes the proof of Lemma 2.1. ∎
2.2 Bounding Partial Vector Discrepancy
In this subsection, we prove the following lemma which upper bounds partial vector discrepancy in terms of the determinant lower bound. The proof can be seen as a simplification of Lemma 8 in [Mat13], which gives a corresponding upper bound for vector discrepancy that is weaker by a factor of due to a bucketing argument that is not needed here.
Lemma 2.3.
For any , we have
Proof.
Recall that is the optimal value of the SDP given by
By denoting , we may rewrite this SDP as follows:
where denotes the vector with on the -th coordinate and 0 elsewhere. The dual formulation of the above SDP is given by the following:
Denote . By Slater’s condition, there exists a feasible dual solution such that and . Indeed, the dual has a feasible interior point (for example, and ) and is bounded, since we may rewrite the first constraint as
(1) |
which implies
Let be the matrix obtained from by multiplying the -th row by and be the set of columns for which . Note that , for otherwise . Since for each we have , for any vector it follows by (1):
This implies that all eigenvalues of are at least , so that . In the other direction, the Cauchy-Binet formula also gives
where the last inequality follows as each term appears times in . Since , we conclude
from which . Applying this result to all subsets of the columns of proves the lemma. ∎
We conjecture that the above Lemma 2.3 is tight up to constants.
Conjecture 2.4.
For any matrix , we have .
Acknowledgments
We thank the anonymous reviewers of SOSA 2022 for insightful comments. We also thank Aleksandar Nikolov, Nikhil Bansal and Mehtaab Sawhney for helpful discussions.
References
- [AV20] Nima Anari and Thuy-Duong Vuong. An extension of plücker relations with applications to subdeterminant maximization. In Approximation, Randomization, and Combinatorial Optimization. Algorithms and Techniques (APPROX/RANDOM 2020). Schloss Dagstuhl-Leibniz-Zentrum für Informatik, 2020.
- [Ban10] Nikhil Bansal. Constructive algorithms for discrepancy minimization. In 2010 IEEE 51st Annual Symposium on Foundations of Computer Science, pages 3–10. IEEE, 2010.
- [CNN11] Moses Charikar, Alantha Newman, and Aleksandar Nikolov. Tight hardness results for minimizing discrepancy. In Proceedings of the twenty-second annual ACM-SIAM symposium on Discrete Algorithms, pages 1607–1614. SIAM, 2011.
- [Fre75] David A Freedman. On tail probabilities for martingales. the Annals of Probability, 3(1):100–118, 1975.
- [KMV05] Jeong Han Kim, Jiří Matoušek, and Van H Vu. Discrepancy after adding a single set. Combinatorica, 25(4):499, 2005.
- [LSV86] László Lovász, Joel Spencer, and Katalin Vesztergombi. Discrepancy of set-systems and matrices. European Journal of Combinatorics, 7(2):151–160, 1986.
- [Mat09] Jiri Matousek. Geometric discrepancy: An illustrated guide, volume 18. Springer Science & Business Media, 2009.
- [Mat13] Jiří Matoušek. The determinant bound for discrepancy is almost tight. Proceedings of the American Mathematical Society, 141(2):451–460, 2013.
- [MNT14] Jiří Matoušek, Aleksandar Nikolov, and Kunal Talwar. Factorization norms and hereditary discrepancy. arXiv preprint arXiv:1408.1376, 2014.
- [Nik15] Aleksandar Nikolov. Randomized rounding for the largest simplex problem. In Proceedings of the forty-seventh annual ACM symposium on Theory of computing, pages 861–870, 2015.
- [NNN12] Alantha Newman, Ofer Neiman, and Aleksandar Nikolov. Beck’s three permutations conjecture: A counterexample and some consequences. In 2012 IEEE 53rd Annual Symposium on Foundations of Computer Science, pages 253–262. IEEE, 2012.
- [Pál10] Dömötör Pálvölgyi. Indecomposable coverings with concave polygons. Discrete & Computational Geometry, 44(3):577–588, 2010.