This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Generic Classification and Asymptotic Enumeration of Dope Matrices

Ankit Bisain
Abstract

For a complex polynomial PP of degree nn and an mm-tuple of distinct complex numbers Λ=(λ1,,λm)\Lambda=(\lambda_{1},\ldots,\lambda_{m}), the dope matrix DP(Λ)D_{P}(\Lambda) is defined as the m×(n+1)m\times(n+1) matrix (c)ij(c)_{ij} with cij=1c_{ij}=1 if P(j)(λi)=0P^{(j)}(\lambda_{i})=0 and cij=0c_{ij}=0 otherwise. We classify the set of dope matrices when the entries of Λ\Lambda are algebraically independent, resolving a conjecture of Alon, Kravitz, and O’Bryant. We also provide asymptotic upper and lower bounds on the total number of m×(n+1)m\times(n+1) dope matrices. For mm much smaller than nn, these bounds give an asymptotic estimate of the logarithm of the number of m×(n+1)m\times(n+1) dope matrices.

1 Introduction

Let P[x]P\in\mathbb{C}[x] be a polynomial of degree nn, and let Λ=(λ1,,λm)\Lambda=(\lambda_{1},\ldots,\lambda_{m}) be an mm-tuple of distinct complex numbers. Following Alon, Kravitz, and O’Bryant [1], we define the dope matrix of PP with respect to Λ\Lambda as the m×(n+1)m\times(n+1) matrix given by

DP(Λ):=(cij)i[m],j[0,n]where cij={1if P(j)(λi)=00otherwise.D_{P}(\Lambda)\vcentcolon=\left(c_{ij}\right)_{i\in[m],j\in[0,n]}\quad\text{where }c_{ij}=\begin{cases}1\ \text{if }P^{(j)}(\lambda_{i})=0\\ 0\ \text{otherwise}.\end{cases}

Hence, the dope matrix tracks the pattern of common zeroes between PP and its derivatives – that is, the set of ordered pairs (i,j)(i,j) for which we have P(j)(λi)=0P^{(j)}(\lambda_{i})=0. A matrix is called dope if it is of the form DP(Λ)D_{P}(\Lambda) for some PP and Λ\Lambda. Denote by 𝒟nm\mathcal{D}_{n}^{m} the set of m×(n+1)m\times(n+1) dope matrices.

The set of possible dope matrices for fixed Λ\Lambda and nn may depend on the values of the λi\lambda_{i}. For example, if PP is a quadratic polynomial, then the conditions P(λ1)=0P(\lambda_{1})=0, P(λ3)=0P(\lambda_{3})=0, and P(λ2)=0P^{\prime}(\lambda_{2})=0 can be simultaneously satisfied only if λ2=λ1+λ32\lambda_{2}=\frac{\lambda_{1}+\lambda_{3}}{2}. In the language of dope matrices, this is

DP((λ1,λ2,λ3))=[100010100]λ2=λ1+λ32.D_{P}((\lambda_{1},\lambda_{2},\lambda_{3}))=\begin{bmatrix}1&0&0\\ 0&1&0\\ 1&0&0\end{bmatrix}\implies\lambda_{2}=\frac{\lambda_{1}+\lambda_{3}}{2}.

For fixed Λ\Lambda and nn, define

𝒟n(Λ)={DP(Λ)P[x],degP=n}.\mathcal{D}_{n}(\Lambda)=\{D_{P}(\Lambda)\mid P\in\mathbb{C}[x],\ \deg P=n\}.

For any a,ba,b\in\mathbb{C} with b0b\neq 0 and any \mathbb{Q}-automorphism φ\varphi of \mathbb{C}, it can be shown (see [4]) that

𝒟n((λ1,,λm))=𝒟n((a+bφ(λ1),,a+bφ(λm))).\mathcal{D}_{n}((\lambda_{1},\ldots,\lambda_{m}))=\mathcal{D}_{n}((a+b\varphi(\lambda_{1}),\ldots,a+b\varphi(\lambda_{m}))). (*)

We denote the mm-tuple on the right-hand side as

a+bφ(Λ):=(a+bφ(λ1),,a+bφ(λm)).a+b\varphi(\Lambda)\vcentcolon=(a+b\varphi(\lambda_{1}),\ldots,a+b\varphi(\lambda_{m})).

Define an affine algebraic dependence of Λ\Lambda to be any rational-coefficient polynomial PP such that P(a+bλ1,a+bλ2,,a+bλm)=0P(a+b\lambda_{1},a+b\lambda_{2},\ldots,a+b\lambda_{m})=0 for all a,ba,b\in\mathbb{C}. For instance, P(x1,,xm)0P(x_{1},\ldots,x_{m})\equiv 0 is an affine algebraic dependence for any Λ\Lambda, and P(x1,x2,x3)=x2x1+x32P(x_{1},x_{2},x_{3})=x_{2}-\frac{x_{1}+x_{3}}{2} is an affine algebraic dependence of (0,1,2)(0,1,2). From now on, we refer to P0P\equiv 0 as the trivial affine algebraic dependence, and refer to all other affine algebraic dependences as nontrivial.

Call Λ\Lambda affinely algebraically independent if it has no nontrivial affine algebraic dependences. Using (*1), we can show (see Theorem 1.7) that 𝒟n(Λ)\mathcal{D}_{n}(\Lambda) is the same for any affinely algebraically independent mm-tuple Λ\Lambda. We define 𝒟ngen(m):=𝒟n(Λ)\mathcal{D}_{n}^{\operatorname{gen}(m)}\vcentcolon=\mathcal{D}_{n}(\Lambda) for any choice of affinely algebraically independent Λ\Lambda. (The notation 𝒟ngen(m)\mathcal{D}_{n}^{\operatorname{gen}(m)} is motivated by the fact that a generic mm-tuple Λ\Lambda is affinely algebraically independent.) We remark that 𝒟ngen(m)\mathcal{D}_{n}^{\operatorname{gen}(m)} is natural to consider, as a generic mm-tuple Λ\Lambda is affinely algebraically independent. In a related direction, Alon, Kravitz, and O’Bryant [1, Theorem 6] have shown that |𝒟n(Λ)|\left\lvert\mathcal{D}_{n}(\Lambda)\right\rvert is maximized exactly when 𝒟n(Λ)=𝒟ngen(m)\mathcal{D}_{n}(\Lambda)=\mathcal{D}_{n}^{\operatorname{gen}(m)}.

In Section 2, we analyze 𝒟ngen(m)\mathcal{D}_{n}^{\operatorname{gen}(m)}. Nathanson [4, Theorem 2] has characterized 𝒟ngen(m)\mathcal{D}_{n}^{\operatorname{gen}(m)} when m=1m=1, and Alon, Kravitz, and O’Bryant [1, Theorem 1] have characterized 𝒟ngen(m)\mathcal{D}_{n}^{\operatorname{gen}(m)} when m=2m=2. We first generalize the aforementioned results to a complete characterization of 𝒟ngen(m)\mathcal{D}_{n}^{\operatorname{gen}(m)} for any mm, resolving a conjecture of the latter paper [1, Conjecture 15].

Theorem 1.1.

For all positive integers m,nm,n, the set 𝒟ngen(m)\mathcal{D}_{n}^{\operatorname{gen}(m)} consists of exactly the m×(n+1)m\times(n+1) matrices with {0,1}\{0,1\} entries such that for all k[0,n]k\in[0,n], there are at most kk nonzero entries in the last k+1k+1 columns.

Using this characterization, we are able to enumerate 𝒟ngen(m)\mathcal{D}_{n}^{\operatorname{gen}(m)}.

Theorem 1.2.

The number of elements of 𝒟ngen(m)\mathcal{D}_{n}^{\operatorname{gen}(m)} with kk ones is

n+1kn+1((n+1)mk),\frac{n+1-k}{n+1}\binom{(n+1)m}{k},

and

((n+1)m1n)(m2)k=0n1((n+1)m1k)\binom{(n+1)m-1}{n}-(m-2)\sum_{k=0}^{n-1}\binom{(n+1)m-1}{k}

is the size of 𝒟ngen(m)\mathcal{D}_{n}^{\operatorname{gen}(m)}.

When m=1m=1, this sum simplifies to 2n2^{n}. When m=2m=2, it simplifies to (2n+1n)\binom{2n+1}{n}, matching a result of Alon, Kravitz, and O’Bryant [1, Corollary 2].

In Sections 3 and 4, we provide bounds on the size of 𝒟nm\mathcal{D}_{n}^{m}, giving a partial answer to a question posed in [1, Problem 16].

In Section 3, we focus on the mn2+n2m\leq\frac{n^{2}+n}{2} case. Alon, Kravitz, and O’Bryant [1, Theorem 4] find an upper bound of (mn2m+n)\binom{mn^{2}}{m+n} by applying a theorem of Rónyai, Babai, and Ganapathy [5, Theorem 1.1] regarding the number of zero-patterns of general sequences of polynomials. We improve this bound by directly applying the methods of [5] to the question at hand. We also find a lower bound on |𝒟nm|\left\lvert\mathcal{D}_{n}^{m}\right\rvert. For mn2+n2m\leq\frac{n^{2}+n}{2}, we have the following:

Theorem 1.3.

For m(t),n(t):>0>0m(t),n(t)\colon\mathbb{Z}_{>0}\to\mathbb{Z}_{>0} satisfying n(t)n(t)\to\infty and 1<m(t)n(t)2+n(t)21<m(t)\leq\frac{n(t)^{2}+n(t)}{2}, we have

(1+o(1))log(nm(mnn))log|𝒟nm|(1+o(1))log(n2m(mnn)).(1+o(1))\log\left(n^{m}\binom{mn}{n}\right)\leq\log\left\lvert\mathcal{D}_{n}^{m}\right\rvert\leq(1+o(1))\log\left(n^{2m}\binom{mn}{n}\right).

Standard asymptotic notation, such as the o(1)o(1) in the above theorem, is used throughout this paper and is explained in Section 1.1. We also freely make statements regarding the asymptotic behavior of (mnn)\binom{mn}{n}, which will follow from the asymptotic estimates of log(mnn)\log\binom{mn}{n} in Section 1.1.

The lower bound in Theorem 1.3 comes from constructing a large set of elements of 𝒟nm\mathcal{D}_{n}^{m}. One possible construction is to consider only the case when Λ\Lambda is generic, which gives a lower bound of log|𝒟ngen(m)|\log\left\lvert\mathcal{D}_{n}^{\operatorname{gen}(m)}\right\rvert.

Another construction comes from taking a “generic” polynomial PP. Consider a polynomial PP such that no two derivatives of PP have a common root and no derivative of PP has a root of multiplicity more than 11. If we construct an element of 𝒟nm\mathcal{D}_{n}^{m} corresponding to the polynomial PP one row at a time, we have around nn choices for each row. Hence, this construction gives a lower bound of around log(nm)\log(n^{m}).

If m=o(n)m=o(n), then only the log(mnn)\log\binom{mn}{n} term of the lower and upper bounds matter asymptotically. In this regime, the bounds essentially match (see Theorem 3.1 for a more precise statement) and 𝒟ngen(m)\mathcal{D}_{n}^{\operatorname{gen}(m)} accounts for the lower bound. When m=ω(n)m=\omega(n), only the log(nm)\log(n^{m}) term in the lower bound matters, and hence the generic PP construction accounts for the lower bound. The construction for the m=Θ(n)m=\Theta(n) case is a combination of the generic Λ\Lambda construction and the generic PP construction.

In Section 4, we focus on the m>n2+n2m>\frac{n^{2}+n}{2} regime. As P(j)P^{(j)} has at most njn-j roots for each jj, any dope matrix can have at most n2+n2\frac{n^{2}+n}{2} nonzero rows. Hence, the growth rate of 𝒟nm\mathcal{D}_{n}^{m} in mm decreases after mm passes the threshold of n2+n2\frac{n^{2}+n}{2}. Beyond this threshold, we have the following results:

Theorem 1.4.

For m(t),n(t):>0>0m(t),n(t)\colon\mathbb{Z}_{>0}\to\mathbb{Z}_{>0} satisfying m(t),n(t)m(t),n(t)\to\infty, we have

  1. (a)

    If n2+n2<m=Θ(n2)\frac{n^{2}+n}{2}<m=\Theta(n^{2}), then log|𝒟nm|log|𝒟nn2+n2|\log\left\lvert\mathcal{D}_{n}^{m}\right\rvert\sim\log\left\lvert\mathcal{D}_{n}^{\frac{n^{2}+n}{2}}\right\rvert;

  2. (b)

    If logm=ω(logn)\log m=\omega(\log n), then log|𝒟nm|log((mn2+n2))\log\left\lvert\mathcal{D}_{n}^{m}\right\rvert\sim\log\left(\dbinom{m}{\frac{n^{2}+n}{2}}\right).

When nn is fixed, |𝒟nm|\left\lvert\mathcal{D}_{n}^{m}\right\rvert is a polynomial in mm, and we can compute the two leading terms:

|𝒟nm|=(n2+n2)!1!2!n!(mn2+n2)+(n2+n2)!1!2!n!(1+(n1)(n2)4)(mn2+n21)+O(mn2+n22).\left\lvert\mathcal{D}_{n}^{m}\right\rvert=\frac{\left(\frac{n^{2}+n}{2}\right)!}{1!2!\cdots n!}\binom{m}{\frac{n^{2}+n}{2}}+\frac{\left(\frac{n^{2}+n}{2}\right)!}{1!2!\cdots n!}\left(1+\frac{(n-1)(n-2)}{4}\right)\binom{m}{\frac{n^{2}+n}{2}-1}+O\left(m^{\frac{n^{2}+n}{2}-2}\right).

1.1 Preliminaries and Notation

We now check that 𝒟ngen(m)\mathcal{D}_{n}^{\operatorname{gen}(m)} is well-defined.

Lemma 1.5.

If an mm-tuple Λ=(λ1,,λm)\Lambda=(\lambda_{1},\ldots,\lambda_{m}) of distinct complex numbers is affinely algebraically independent, then there exist complex numbers a,ba,b such that a+bΛa+b\Lambda is algebraically independent.

Proof.

We have that for all polynomials P[x1,,xm]{0}P\in\mathbb{Q}[x_{1},\ldots,x_{m}]\setminus\{0\}, the polynomial P(a+bΛ)[a,b]P(a+b\Lambda)\in\mathbb{C}[a,b] is not identically zero. We want to show that there exist a,ba,b\in\mathbb{C} such that P(a+bΛ)0P(a+b\Lambda)\neq 0 for all P[x1,,xn]{0}P\in\mathbb{Q}[x_{1},\ldots,x_{n}]\setminus\{0\}. We’ll show, more generally, that if SS is any countable subset of [a,b]{0}\mathbb{C}[a,b]\setminus\{0\}, then there exist a,ba,b such that Q(a,b)0Q(a,b)\neq 0 for all QSQ\in S.

View each QSQ\in S as a polynomial in bb with coefficients being polynomials in aa. For some kk, the coefficient of bkb^{k} in QQ is some nonzero polynomial RR in aa. For any aa with R(a)0R(a)\neq 0, we have that Q(a,b)Q(a,b) has finitely many roots as a polynomial in bb. For each QSQ\in S, pick some such polynomial RR, and let AQA_{Q} be the set of roots of RR. Since each AQA_{Q} is finite, we have that QSAQ\bigcup_{Q\in S}A_{Q} is countable, and hence, we can pick some a0a_{0}\in\mathbb{C} that isn’t in AQA_{Q} for any QSQ\in S.

We claim that for some b0b_{0}\in\mathbb{C}, we have Q(a0,b0)0Q(a_{0},b_{0})\neq 0 for all QSQ\in S. Let BQB_{Q} be the set of roots of Q(a0,b)Q(a_{0},b) viewed as a polynomial in bb. By the definition of a0a_{0}, we have that each BQB_{Q} is finite, so QSBQ\bigcup_{Q\in S}B_{Q} is countable. Hence, there is some b0b_{0}\in\mathbb{C} that isn’t in BQB_{Q} for any QSQ\in S, as desired. ∎

We’ll also recall a fact from [4] allowing us to equate Dn(Λ)D_{n}(\Lambda) and 𝒟n(Λ)\mathcal{D}_{n}(\Lambda^{\prime}) when we have linear maps or \mathbb{Q}-automorphisms of \mathbb{C} sending Λ\Lambda to Λ\Lambda^{\prime}.

Lemma 1.6.

[4, Theorem 5] For any \mathbb{Q}-automorphism φ\varphi and complex numbers a,ba,b with b0b\neq 0, we have 𝒟n(Λ)=𝒟n(a+bφ(Λ))\mathcal{D}_{n}(\Lambda)=\mathcal{D}_{n}(a+b\varphi(\Lambda)).

The idea of the proof is, for any polynomial P(x)=anxn+an1xn1++a0P(x)=a_{n}x^{n}+a_{n-1}x^{n-1}+\cdots+a_{0}, to consider the polynomial

Pa+bφ(x):=φ(an)(xab)n+φ(an1)(xab)n1++φ(a0).P_{a+b\varphi}(x)\vcentcolon=\varphi(a_{n})\left(\frac{x-a}{b}\right)^{n}+\varphi(a_{n-1})\left(\frac{x-a}{b}\right)^{n-1}+\cdots+\varphi(a_{0}).

This polynomial has the property that 𝒟P(Λ)=𝒟Pa+bφ(a+bΛ)\mathcal{D}_{P}(\Lambda)=\mathcal{D}_{P_{a+b\varphi}}(a+b\Lambda) for all PP. A more complete explanation can be found in [4].

We are now ready to check that 𝒟ngen(m)\mathcal{D}_{n}^{\operatorname{gen}(m)} is well-defined.

Theorem 1.7.

If Λ1\Lambda_{1} and Λ2\Lambda_{2} are affinely algebraically independent mm-tuples of complex numbers, then 𝒟n(Λ1)=𝒟n(Λ2)\mathcal{D}_{n}(\Lambda_{1})=\mathcal{D}_{n}(\Lambda_{2}).

Proof.

Let Λ\Lambda be an algebraically independent mm-tuple of complex numbers. By Lemma 1.5, there are constants a1,b1,a2,b2a_{1},b_{1},a_{2},b_{2}\in\mathbb{C} such that a1+b1Λ1a_{1}+b_{1}\Lambda_{1} and a2+b2Λ2a_{2}+b_{2}\Lambda_{2} are algebraically independent. Now, there exists some \mathbb{Q}-automorphism φ\varphi of \mathbb{C} sending a1+b1Λ1a_{1}+b_{1}\Lambda_{1} to a2+b2Λ2a_{2}+b_{2}\Lambda_{2} (see the introduction of [1] for further discussion of this fact). Hence,

𝒟n(Λ1)=𝒟n(a1+b1Λ1)=𝒟n(φ(a1+b1Λ1))=𝒟n(a2+bnΛ2)=𝒟n(Λ2),\mathcal{D}_{n}(\Lambda_{1})=\mathcal{D}_{n}(a_{1}+b_{1}\Lambda_{1})=\mathcal{D}_{n}(\varphi(a_{1}+b_{1}\Lambda_{1}))=\mathcal{D}_{n}(a_{2}+b_{n}\Lambda_{2})=\mathcal{D}_{n}(\Lambda_{2}),

as desired. ∎

To state our results on asymptotic dope matrix counts, we will use standard asymptotic notation. For functions f,g:f,g\colon\mathbb{Z}\to\mathbb{R}, we can compare the growth rates of ff and gg with the following:

  1. 1.

    f=O(g)f=O(g) if, for some constant KK, we have |f|K|g|\left\lvert f\right\rvert\leq K\left\lvert g\right\rvert for all sufficiently large tt;

  2. 2.

    f=Θ(g)f=\Theta(g) if, for some constants K,KK,K^{\prime}, we have K|g||f|K|g|K\left\lvert g\right\rvert\leq\left\lvert f\right\rvert\leq K^{\prime}\left\lvert g\right\rvert for all sufficiently large tt;

  3. 3.

    fgf\sim g if, as tt approaches \infty, we have that fg\frac{f}{g} approaches 11;

  4. 4.

    f=o(g)f=o(g) if, for any constant ε\varepsilon, we have |f|ε|g|\left\lvert f\right\rvert\leq\varepsilon\left\lvert g\right\rvert for all sufficiently large tt;

  5. 5.

    f=ω(g)f=\omega(g) if, for any constant MM, we have |f|M|g|\left\lvert f\right\rvert\geq M\left\lvert g\right\rvert for all sufficiently large tt.

We will frequently need to analyze the asymptotic behavior of n!n! and log(mnn)\log\binom{mn}{n}. We will make use of the following inequalities:

  1. 1.

    For any positive integer nn, we have nlognnlogn!nlognn\log n-n\leq\log n!\leq n\log n.

  2. 2.

    For m(t),n(t):>0>1m(t),n(t)\colon\mathbb{Z}_{>0}\to\mathbb{Z}_{>1} satisfying n(t)n(t)\to\infty, we have:

    log(mnn)=nlog(mm(m1)(m1))+O(logn)=nlogm+O(n).\log\binom{mn}{n}=n\log\left(\frac{m^{m}}{(m-1)^{(m-1)}}\right)+O(\log n)=n\log m+O(n).

    In particular, log(mnn)\log\binom{mn}{n} grows linearly when m=O(1)m=O(1) and is ω(n)\omega(n) when m=ω(1)m=\omega(1).

The inequality follows from

nlognn1nlogxdxlog2+log3++lognnlogn,n\log n-n\leq\int_{1}^{n}\log x\mathrm{d}x\leq\log 2+\log 3+\cdots+\log n\leq n\log n,

since logn!=log2+log3++logn\log n!=\log 2+\log 3+\cdots+\log n. The first asymptotic estimate on log(mnn)\log\binom{mn}{n} follows directly from Stirling’s approximation, and the second follows from the fact that 2(mm1)m1<e2\leq\left(\frac{m}{m-1}\right)^{m-1}<e for all m>1m>1.

2 Generic Dope Matrices

In this section we will prove Theorem 1.1. We first outline the proof of the m=2m=2 case from [1, Section 4], as our proof uses many ideas from it. It makes use of the following result (slightly rephrased) of Gessel and Viennot.

Theorem 2.1.

[3, Corollary 3] Let 𝒢,0\mathcal{G},\mathcal{H}\subset\mathbb{Z}_{\geq 0} be finite sets. If |𝒢[0,c]||[0,c]|\left\lvert\mathcal{G}\cap[0,c]\right\rvert\leq\left\lvert\mathcal{H}\cap[0,c]\right\rvert for every c0c\in\mathbb{Z}_{\geq 0}, then the matrix of binomial coefficients

[(gh)]g𝒢,h\left[\binom{g}{h}\right]_{g\in\mathcal{G},h\in\mathcal{H}}

has rank |𝒢|\left\lvert\mathcal{G}\right\rvert.

Outline of Proof of Theorem 1.1 when m=2m=2.

Take Λ=(0,1)\Lambda=(0,1), and let P(x)=anxn+an1xn1++a0P(x)=a_{n}x^{n}+a_{n-1}x^{n-1}+\cdots+a_{0}. For a 2×(n+1)2\times(n+1) matrix MM, we can view the equation DP(Λ)=MD_{P}(\Lambda)=M as a system of linear equations in an,,a0a_{n},\ldots,a_{0} via

P(s)(0)=0as=0andP(s)(1)=0i=0n(nis)ai=0.P^{(s)}(0)=0\iff a_{s}=0\quad\text{and}\quad P^{(s)}(1)=0\iff\sum_{i=0}^{n}\binom{n-i}{s}a_{i}=0.

When MM does not satisfy the property in Theorem 1.1, for some kk, it has at least k+1k+1 ones in the last k+1k+1 columns. Take the minimum kk with the preceding property and look at the last k+1k+1 columns. The resulting linear equations are linearly independent by Theorem 2.1.

In these linear equations, only the k+1k+1 variables an,an1,,anka_{n},a_{n-1},\ldots,a_{n-k} have nonzero coefficients, so we must have an=an1==ank=0a_{n}=a_{n-1}=\cdots=a_{n-k}=0. However, if an=0a_{n}=0, then PP cannot be a degree-nn polynomial, which gives a contradiction.

To show that any matrix MM with at most kk ones in the last k+1k+1 columns is attainable, we add all-one columns to the left end of the matrix until the number of ones is one smaller than the number of columns. Say we added cc columns, and let MM^{\prime} be the matrix with the cc columns added. Take the resulting system of equations and append an+c=1a_{n+c}=1, ensuring that our desired polynomial has degree exactly n+cn+c.

By Theorem 2.1, the resulting equations are linearly independent, and hence have a unique solution. Letting P0P_{0} be the polynomial corresponding to this solution, we have that DP0(Λ)D_{P_{0}}(\Lambda) is not the zero matrix, since P(n+c)(0)0P^{(n+c)}(0)\neq 0. Hence, DP0(Λ)D_{P_{0}}(\Lambda) must have at most n+cn+c ones, and is hence exactly MM^{\prime}. Now, DP0(c)(Λ)D_{P_{0}^{(c)}}(\Lambda) is MM, as desired. ∎

We now introduce notation that will be helpful in our proof of Theorem 1.1

Definition 2.2.

Call a matrix safe if the last k+1k+1 columns contain at most kk nonzero entries for all kk. Similarly, call a matrix almost-safe if the last k+1k+1 columns contain at most k+1k+1 nonzero entries for all kk.

Throughout this section, we will let P(x)=anxn+an1xn1++a0P(x)=a_{n}x^{n}+a_{n-1}x^{n-1}+\cdots+a_{0} denote a general polynomial with degPn\deg P\leq n. Note that the set of such PP forms a complex vector space of dimension n+1n+1 with basis given by 1,x,,xn1,x,\ldots,x^{n}. Hence, for a fixed integer ss and complex number tt, we can view P(s)(t)P^{(s)}(t) as a linear form in PP:

(a0,,an)j=0nj(j1)(j+1s)tjsaj.(a_{0},\ldots,a_{n})\mapsto\sum_{j=0}^{n}{j(j-1)\cdots(j+1-s)t^{j-s}}a_{j}.

We will use Λ=(λ1,,λm)\Lambda=(\lambda_{1},\ldots,\lambda_{m}) and 𝐒=(S1,,Sm)\mathbf{S}=(S_{1},\ldots,S_{m}) to denote arbitrary mm-tuples of complex numbers and mm-tuples of subsets of {0,1,,n}\{0,1,\ldots,n\}, respectively. Denote by P(𝐒)(Λ)P^{(\mathbf{S})}(\Lambda) the set of linear forms Ps(λi)P^{s}(\lambda_{i}) for i[m]i\in[m] and sSis\in S_{i}, and denote by M(𝐒)M(\mathbf{S}) the matrix (cij)i[m],j[0,n](c_{ij})_{i\in[m],\ j\in[0,n]}, where cijc_{ij} is 11 if jSij\in S_{i} and 0 otherwise.

2.1 Reduction to the Key Lemma

The key lemma is the following linear independence property:

Lemma 2.3.

If M(𝐒)M(\mathbf{S}) is almost-safe and Λ\Lambda is generic, then P𝐒(Λ)P^{\mathbf{S}}(\Lambda) is linearly independent.

Before proving this lemma, we will show how it implies Theorem 1.1. The proof is similar to the proof of the m=2m=2 case in [1], with Lemma 2.3 replacing the result of Gessel-Viennot.

Proof of Theorem 1.1.

First, we show that being safe is necessary. Let Λ\Lambda be a generic mm-tuple. Suppose for some 𝐒\mathbf{S} such that M(𝐒)M(\mathbf{S}) is not safe, there exists a polynomial PP such that DP(Λ)=M(𝐒)D_{P}(\Lambda)=M(\mathbf{S}).

Take the smallest kk such that the last k+1k+1 columns contain at least k+1k+1 nonzero entries. The linear forms corresponding to these entries are linear functions of an,an1,,anka_{n},a_{n-1},\ldots,a_{n-k} and are linearly independent by Lemma 2.3. Hence, an=0a_{n}=0, contradicting the assumption that PP is degree-nn.

To show that the condition is sufficient, given any safe matrix MM, we prepend columns where the top two entries are 11 and the remaining entries are 0 such that the resulting matrix MM^{\prime} has exactly n+cn+c nonzero entries, where cc is the number of prepended columns. We have that MM^{\prime} is safe.

Let 𝐒\mathbf{S} be such that M(𝐒)=MM(\mathbf{S})=M^{\prime}. By Lemma 2.3, P𝐒(Λ)P^{\mathbf{S}}(\Lambda) consists of n+cn+c independent linear forms. Appending the linear form P(n+c)(λ1)P^{(n+c)}(\lambda_{1}) corresponds to adding a one to the last column of MM^{\prime}, which makes MM^{\prime} remain almost-safe, so by Lemma 2.3, P𝐒(Λ){P(n+c)(λ1)}P^{\mathbf{S}}(\Lambda)\cup\{P^{(n+c)}(\lambda_{1})\} is a set of n+c+1n+c+1 independent linear forms. Hence, there is some nonzero polynomial P0P_{0} with degree at most n+cn+c such that P0(n+c)(λ1)=1P_{0}^{(n+c)}(\lambda_{1})=1, and for all i[m]i\in[m] and sSis\in S_{i}, we have P0(s)(λi)=0P_{0}^{(s)}(\lambda_{i})=0. Since P0(n+c)(λ1)=1P_{0}^{(n+c)}(\lambda_{1})=1, we have that P0P_{0} has degree exactly n+cn+c.

We next show that we cannot have P0(s)(λi)=0P_{0}^{(s)}(\lambda_{i})=0 for any i[m]i\in[m] and sSis\not\in S_{i}. Consider 𝐒=(S1,,Si1,Si{s},Si+1,,Sm)\mathbf{S}^{\prime}=(S_{1},\ldots,S_{i-1},S_{i}\cup\{s\},S_{i+1},\ldots,S_{m}). If P0(s)(λi)=0P_{0}^{(s)}(\lambda_{i})=0, then P0P_{0} is zero on every element of P𝐒(Λ)P^{\mathbf{S}^{\prime}}(\Lambda). However, M(𝐒)M(\mathbf{S}) is almost-safe, so P𝐒(Λ)P^{\mathbf{S}^{\prime}}(\Lambda) contains n+c+1n+c+1 linearly independent linear functions by Lemma 2.3. This forces P0P_{0} to be the zero polynomial, which is a contradiction. Thus, DP0(Λ)D_{P_{0}}(\Lambda) is exactly MM^{\prime}, and hence DP0(c)(Λ)D_{P_{0}^{(c)}}(\Lambda) is the desired matrix. ∎

2.2 Demonstration of Proof Technique

As the proof of Lemma 2.3 is fairly technical, we provide a demonstration of the proof for a small almost-safe matrix.

Consider the 𝐒\mathbf{S} corresponding to the following 3×63\times 6 matrix, which is almost-safe:

M(𝐒)=[110010000100100010].M(\mathbf{S})=\begin{bmatrix}1&1&0&0&1&0\\ 0&0&0&1&0&0\\ 1&0&0&0&1&0\end{bmatrix}.

We will show that if Λ=(0,1,t)\Lambda=(0,1,t) for some transcendental tt, then P𝐒(Λ)P^{\mathbf{S}}(\Lambda) is linearly independent.

We want to show that for generic tt, the linear forms

P(0),P(0),P(4)(0),P(3)(1),P(t),P(4)(t)P(0),P^{\prime}(0),P^{(4)}(0),P^{(3)}(1),P(t),P^{(4)}(t)

are linearly independent. It suffices to show that the result holds for at least one transcendental tt, as we can then take a \mathbb{Q}-automorphism mapping tt to any transcendental number of our choosing. The key idea is to check the special case of tt very close to 0.

Taking linear combinations, we find that the span of the aforementioned linear forms is the same as the span of

P(0),P(0),P(4)(0),P(3)(1),P(t)P(0)tP(0)t2/2,P(4)(t)P(4)(0)t.P(0),P^{\prime}(0),P^{(4)}(0),P^{(3)}(1),\frac{P(t)-P(0)-tP^{\prime}(0)}{t^{2}/2},\frac{P^{(4)}(t)-P^{(4)}(0)}{t}.

The last two forms are polynomials in tt, and can hence be continuously extended to t=0t=0. At t=0t=0, the forms are equal to

P(0),P(0),P(4)(0),P(3)(1),P(2)(0),P(5)(0).P(0),P^{\prime}(0),P^{(4)}(0),P^{(3)}(1),P^{(2)}(0),P^{(5)}(0).

These linear forms correspond to the 2×62\times 6 almost-safe matrix

[111011000100],\begin{bmatrix}1&1&1&0&1&1\\ 0&0&0&1&0&0\end{bmatrix},

and hence are linearly independent by the m=2m=2 case of the theorem. By the fact that linear independence is equivalent to nonzero determinant, we can extend the linear independence of

P(0),P(0),P(4)(0),P(3)(1),P(t)P(0)tP(0)t2/2,P(4)(t)P(4)(0)tP(0),P^{\prime}(0),P^{(4)}(0),P^{(3)}(1),\frac{P(t)-P(0)-tP^{\prime}(0)}{t^{2}/2},\frac{P^{(4)}(t)-P^{(4)}(0)}{t}

from t=0t=0 to all tt in some neighborhood of 0. Hence, they are linearly independent for some transcendental tt, implying the desired result.

One can view the argument above as showing that we can combine the roots at 0 and tt, and hence combine the corresponding rows of the matrix.

The proof in the general case consists of two parts. First, we prove a claim generalizing the choice of linear combinations in the above proof. We then generalize the t0t\to 0 argument, allowing us to combine the λi\lambda_{i}’s under certain conditions. Once this is done, Lemma 2.3 follows from repeatedly combining the λi\lambda_{i}’s.

2.3 Derivatives as Linear Combinations

We first, using Theorem 2.1, find linear combinations that limit to derivatives.

Lemma 2.4.

Fix d0d\in\mathbb{Z}_{\geq 0}. Let 𝐒=(S1,S2)\mathbf{S}=(S_{1},S_{2}), where S1,S2[0,d]S_{1},S_{2}\subset[0,d]. Suppose that in M(𝐒)M(\mathbf{S}), for all 0kd0\leq k\leq d, there are at most k+1k+1 nonzero entries in columns [dk,d][d-k,d], with equality holding for k=dk=d. Then there exist constants cs,1c_{s,1} for sS1s\in S_{1} and cs,2c_{s,2} for sS2s\in S_{2} such that the following holds:

For every polynomial PP, there exists a polynomial Q[λ,ε]Q\in\mathbb{C}[\lambda,\varepsilon] such that

P(d)(λ)=sS1cs,1P(s)(λ)s!εsd+sS2cs,2P(s)(λ+ε)s!εsd+εQ(λ,ε)P^{(d)}(\lambda)=\sum_{s\in S_{1}}c_{s,1}\frac{P^{(s)}(\lambda)}{s!}\varepsilon^{s-d}+\sum_{s\in S_{2}}c_{s,2}\frac{P^{(s)}(\lambda+\varepsilon)}{s!}\varepsilon^{s-d}+\varepsilon Q(\lambda,\varepsilon) (\star)

holds for all λ,ε\lambda,\varepsilon\in\mathbb{C}.

Proof.

Let P(x)=a0(xλ)0++ad(xλ)d+(xλ)d+1Q0(xλ)P(x)=a_{0}(x-\lambda)^{0}+\cdots+a_{d}(x-\lambda)^{d}+(x-\lambda)^{d+1}Q_{0}(x-\lambda). We will view the aia_{i} as variables. We can write P(s)P^{(s)} evaluated at λ\lambda and λ+ε\lambda+\varepsilon in the basis of the aia_{i} as follows:

P(s)(λ)s!εsd=εsdasandP(s)(λ+ε)s!=t=sn(ts)εtdat.\frac{P^{(s)}(\lambda)}{s!}\varepsilon^{s-d}=\varepsilon^{s-d}a_{s}\quad\text{and}\quad\frac{P^{(s)}(\lambda+\varepsilon)}{s!}=\sum_{t=s}^{n}\binom{t}{s}\varepsilon^{t-d}a_{t}.

Hence, the right-hand side of \star2.4 contains only linear combinations of {εtdatt[0,n]}\left\{\varepsilon^{t-d}a_{t}\mid t\in[0,n]\right\}. For any choice of the cs,2c_{s,2}’s, we can pick QQ to make the coefficient of εtdat\varepsilon^{t-d}a_{t} zero for t>dt>d, and we can pick ct,1c_{t,1} to make the coefficient of εtdat\varepsilon^{t-d}a_{t} zero for tS1t\in S_{1}. Hence, it suffices to pick cs,2c_{s,2} such that, for all t[0,d]S1t\in[0,d]\setminus S_{1}, the coefficient of εtdat\varepsilon^{t-d}a_{t} in the right-hand side of \star2.4 matches the corresponding coefficient in the left-hand side.

Let 𝒢=[0,d]S1\mathcal{G}=[0,d]\setminus S_{1} and =S2\mathcal{H}=S_{2}. For all g𝒢g\in\mathcal{G}, the coefficient of εgdag\varepsilon^{g-d}a_{g} in some term in the right-hand side of (\star2.4) is

s(gs)cs,2\sum_{s\in\mathcal{H}}\binom{g}{s}c_{s,2}

The column condition implies |S1[0,c]|+|S2[0,c]|c+1\left\lvert S_{1}\cap[0,c]\right\rvert+\left\lvert S_{2}\cap[0,c]\right\rvert\geq c+1 for all cc, so |𝒢[0,c]||[0,c]|\left\lvert\mathcal{G}\cap[0,c]\right\rvert\leq\left\lvert\mathcal{H}\cap[0,c]\right\rvert for all cc. Now, by Theorem 2.1, we can find constants ch,2c_{h,2} such that for all gg in 𝒢\mathcal{G},

h(gh)ch,2={0if gdd!if g=d,\sum_{h\in\mathcal{H}}\binom{g}{h}c_{h,2}=\begin{cases}0\quad\text{if }g\neq d\\ d!\quad\text{if }g=d,\end{cases}

as desired. ∎

2.4 Combining Roots

We make use of the following fact, which allows us to take limits of linear dependences:

Proposition 2.5.

Let γ1,,γ\gamma_{1},\ldots,\gamma_{\ell} be continuous maps from \mathbb{C} to some complex vector space. If γ1(0),,γ(0)\gamma_{1}(0),\ldots,\gamma_{\ell}(0) are linearly independent, then γ1(t),,γ(t)\gamma_{1}(t),\ldots,\gamma_{\ell}(t) are linearly independent for all tt in some neighborhood of 0.

This follows from the fact that, in a given complex vector space, linearly independent \ell-tuples of vectors form an open set.

We now prove the claim allowing us to “combine” λ1\lambda_{1} and λm\lambda_{m}. This will be the inductive step in our proof of Lemma 2.3.

Lemma 2.6.

Suppose m>1m>1, and suppose that M(𝐒)M(\mathbf{S}) is almost-safe. Define Λ=Λ{λm}\Lambda^{\prime}=\Lambda\setminus\{\lambda_{m}\} and 𝐒=(S1,,Sm1)\mathbf{S}^{\prime}=(S^{\prime}_{1},\ldots,S^{\prime}_{m-1}), where

  • Si=SiS^{\prime}_{i}=S_{i} for 2im12\leq i\leq m-1

  • sS1s\in S^{\prime}_{1} if and only if for some tst\leq s, we have |S1[t,s]|+|Sm[t,s]|st+1\left\lvert S_{1}\cap[t,s]\right\rvert+\left\lvert S_{m}\cap[t,s]\right\rvert\geq s-t+1.

Then, we have:

  1. (a)

    the matrix M(𝐒)M(\mathbf{S}^{\prime}) is almost-safe,

  2. (b)

    if P𝐒(Λ)P^{\mathbf{S}^{\prime}}(\Lambda^{\prime}) is linearly independent, then so is P𝐒(Λ)P^{\mathbf{S}}(\Lambda), and

  3. (c)

    if Λ\Lambda is generic, so is Λ\Lambda^{\prime}.

Proof.

The proof of (c) is clear. We will begin by proving (a).

Consider the row vector obtained as follows:

  1. 1.

    Add the first and mmth rows of M(𝐒)M(\mathbf{S}), obtaining some vector vv.

  2. 2.

    We index the components of vv with 0,1,,n0,1,\ldots,n. If, for some jj, the jjth component of vv is at least 22, subtract 11 from the jjth component and add 11 to the (j+1)(j+1)th component.

  3. 3.

    Repeat the previous step until it cannot be repeated anymore.

This aligns with the intuition of combining the roots of row 11 and row mm, since we expect a double root of P(j)P^{(j)} to become a root of P(j)P^{(j)} and a root of (P(j))=P(j+1)\left(P^{(j)}\right)^{\prime}=P^{(j+1)}. We claim that, no matter which choices are made during step 2, this process will always result in the row vector corresponding to S1S^{\prime}_{1}.

Let viv_{i} denote the vector obtained in step 1, and let vfv_{f} denote the vector obtained at the end of the process. We use vv to denote an arbitrary vector at any point in the process.

For each ss, consider the quantity Cs(v):=maxts(v𝟏[t,s](st+1))C_{s}(v)\vcentcolon=\max_{t\leq s}\left(v\cdot\mathbf{1}_{[t,s]}-(s-t+1)\right), where 𝟏[t,s]\mathbf{1}_{[t,s]} is the row vector that is 11 on the columns [t,s][t,s] and 0 elsewhere. During any application of step 2 above, Cs(v)C_{s}(v) is unchanged if j<sj<s, is decreased by 11 if j=sj=s, and is unchanged if j>sj>s. In the j=sj=s case, taking t=st=s, we must have that Cs(v)0C_{s}(v)\geq 0 after the application of step 2. Hence, Cs(vi)0C_{s}(v_{i})\geq 0 if and only if Cs(vf)0C_{s}(v_{f})\geq 0.

Since M(𝐒)M(\mathbf{S}) is almost-safe, we have Cn(vi)0C_{n}(v_{i})\leq 0. Hence, we must also have Cn(vf)0C_{n}(v_{f})\leq 0 – in particular, the nnth component of vfv_{f} cannot be more than 11. By definition, the jjth component of vfv_{f} cannot be more than 11 for any j<nj<n. Thus, all components of vfv_{f} are either 0 or 11.

We have vi𝟏[t,s]=|S1[t,s]|+|Sm[t,s]|v_{i}\cdot\mathbf{1}_{[t,s]}=\left\lvert S_{1}\cap[t,s]\right\rvert+\left\lvert S_{m}\cap[t,s]\right\rvert, so Cs(vi)0C_{s}(v_{i})\geq 0 if and only if sS1s\in S^{\prime}_{1}. Since Cs(vf)0C_{s}(v_{f})\geq 0 is equivalent to both Cs(vi)0C_{s}(v_{i})\geq 0 and the jjth component of vfv_{f} being nonzero, vfv_{f} is the row vector corresponding to S1S^{\prime}_{1}.

As a corollary of this alternate characterization, we have |S1|+|Sm|=|S1|\left\lvert S_{1}\right\rvert+\left\lvert S_{m}\right\rvert=\left\lvert S^{\prime}_{1}\right\rvert, and hence P𝐒(Λ)P^{\mathbf{S}}(\Lambda) and P𝐒(Λ)P^{\mathbf{S}^{\prime}}(\Lambda^{\prime}) have the same number of elements. We also have that if M(𝐒)M(\mathbf{S}) is almost-safe, then so is M(𝐒)M(\mathbf{S}^{\prime}), as all steps of the above process preserve the property that the sum of the entries of the last k+1k+1 columns is at most k+1k+1 for all kk.

We now prove (b). Recall that we use PP to denote a general polynomial of degree at most nn. We claim that for each sS1s^{\prime}\in S^{\prime}_{1}, there exist a polynomial QsQ_{s^{\prime}}, independent of ε\varepsilon but possibly dependent on Λ\Lambda and PP, and constants cs,ic_{s,i}, possibly dependent on ε\varepsilon, such that

P(s)(λ1)εQs(ε)=(sS1cs,1P(s)(λ1))+(sSmcs,mP(s)(λ1+ε)).P^{(s^{\prime})}(\lambda_{1})-\varepsilon Q_{s^{\prime}}(\varepsilon)=\left(\sum_{s\in S_{1}}c_{s,1}P^{(s)}(\lambda_{1})\right)+\left(\sum_{s\in S_{m}}c_{s,m}P^{(s)}(\lambda_{1}+\varepsilon)\right).

If sS1s\in S_{1}, the result is clear. Otherwise, take the largest tt such that |S1[t,s]|+|Sm[t,s]|st+1\left\lvert S_{1}\cap[t,s^{\prime}]\right\rvert+\left\lvert S_{m}\cap[t,s^{\prime}]\right\rvert\geq s^{\prime}-t+1. For this tt, we have |S1[t,s]|+|Sm[t,s]|=st+1\left\lvert S_{1}\cap[t,s^{\prime}]\right\rvert+\left\lvert S_{m}\cap[t,s^{\prime}]\right\rvert=s^{\prime}-t+1, as otherwise t+1t+1 would have the same property. Applying Lemma 2.4 to the polynomial P(t)P^{(t)}, the degree d=std=s^{\prime}-t, λ=λ1\lambda=\lambda_{1}, ε=λmλ1\varepsilon=\lambda_{m}-\lambda_{1}, and the sets 𝐒=({stsS1,st},{stsSm,st})\mathbf{S}=(\{s-t\mid s\in S_{1},s\geq t\},\{s-t\mid s\in S_{m},s\geq t\}) gives the desired QQ and cs,ic_{s,i}.

By our assumptions, P𝐒(Λ)P^{\mathbf{S^{\prime}}}(\Lambda^{\prime}), which is

(P(S2,,Sm1)((λ2,,λm1)))(sS1{P(s)(λ1)εQs(ε)})\left(P^{(S_{2}^{\prime},\ldots,S_{m-1}^{\prime})}((\lambda_{2},\ldots,\lambda_{m-1}))\right)\bigcup\left(\bigcup_{s^{\prime}\in S^{\prime}_{1}}\left\{P^{(s^{\prime})}(\lambda_{1})-\varepsilon Q_{s^{\prime}}(\varepsilon)\right\}\right)

evaluated at ε=0\varepsilon=0, is linearly independent. Every linear form in the above set is continuous in ε\varepsilon, and hence the above set is linearly independent for some transcendental ε0\varepsilon\neq 0 by Proposition 2.5. Its span for this ε\varepsilon is a subspace of the span of P𝐒(Λ)P^{\mathbf{S}}(\Lambda), and has dimension |P𝐒(Λ)|=|P𝐒(Λ)|\left\lvert P^{\mathbf{S}^{\prime}}(\Lambda^{\prime})\right\rvert=\left\lvert P^{\mathbf{S}}(\Lambda)\right\rvert. Hence, P𝐒(Λ)P^{\mathbf{S}}(\Lambda) is linearly independent. ∎

We can now repeatedly combine elements to prove that all almost-safe matrices are linearly independent.

Proof of Lemma 2.3.

We proceed by induction on mm. The base case of m=1m=1 is trivial. The induction step follows immediately from Lemma 2.6. ∎

2.5 Enumeration

The enumeration of generic dope matrices follows from a direct application of the cycle lemma.

Definition 2.7.

We say that a sequence of ones and zeroes is tt-dominating if for every >0\ell>0, the number of zeroes among the first \ell entries is more than tt times the number of ones.

The cycle lemma allows us to count the number of tt-dominating sequences with a given number of ones and zeroes.

Theorem 2.8.

[2, Cycle Lemma] Let a,b,ta,b,t be nonnegative integers with atba\geq tb. For any sequence p1,,pa+bp_{1},\ldots,p_{a+b} of aa zeroes and bb ones, exactly atba-tb of the cyclic shifts of the sequence are tt-dominating.

Proof of Theorem 1.2.

To prove the assertion for fixed kk, consider the map from (cij)i[m],j[0,n]𝒟ngen(m)\left(c_{ij}\right)_{i\in[m],j\in[0,n]}\in\mathcal{D}_{n}^{\operatorname{gen}(m)} to (m1)(m-1)-dominating sequences a1,,am(n+1)a_{1},\ldots,a_{m(n+1)} given by cijam(n+1j)+ic_{ij}\leftrightarrow a_{m(n+1-j)+i}. By Theorem 1.1, this is a bijection between elements of 𝒟ngen(m)\mathcal{D}_{n}^{\operatorname{gen}(m)} with kk ones and mm-dominating sequences with kk ones.

By Theorem 2.8, of the ((n+1)mk)\binom{(n+1)m}{k} length-m(n+1)m(n+1) sequences with kk ones and m(n+1)km(n+1)-k zeroes,

n+1kn+1((n+1)mk)=((n+1)m1k)(m1)((n+1)m1k1)\frac{n+1-k}{n+1}\dbinom{(n+1)m}{k}=\binom{(n+1)m-1}{k}-(m-1)\binom{(n+1)m-1}{k-1}

of them are (m1)(m-1)-dominating, using the convention (N1)=0\binom{N}{-1}=0, implying the assertion for fixed kk. Summing over 0kN0\leq k\leq N gives the desired formula for |𝒟ngen(m)|\left\lvert\mathcal{D}_{n}^{\operatorname{gen}(m)}\right\rvert. ∎

Remark 2.9.

When m=2m=2, the size of 𝒟ngen(m)\mathcal{D}_{n}^{\operatorname{gen}(m)} simplifies to (2n+1n)\binom{2n+1}{n}. In this case, a direct counting argument is possible. The earlier map gives a bijection between 𝒟ngen(2)\mathcal{D}_{n}^{\operatorname{gen}(2)} and 11-dominating {0,1}\{0,1\} sequences. Deleting the first element of the sequence and treating zeroes as ups and ones as downs, the 11-dominating {0,1}\{0,1\} sequences are in bijection with length-(2n+1)(2n+1) left factors of Dyck paths, of which there are (2n+1n)\binom{2n+1}{n}.

From the formula in Theorem 1.2, we can find good closed-form estimates for |Dngen(m)|\left\lvert D_{n}^{\operatorname{gen}(m)}\right\rvert. For m=1,2m=1,2, we have the exact formulas 2n2^{n} and (2n+1n)\binom{2n+1}{n}, respectively. For larger mm, we have the following:

Corollary 2.10.

For m3m\geq 3 and n1n\geq 1, we have

1n+1((n+1)mn)|𝒟ngen(m)|(1+1m2)21n+1((n+1)mn).\frac{1}{n+1}\dbinom{(n+1)m}{n}\leq\left\lvert\mathcal{D}_{n}^{\operatorname{gen}(m)}\right\rvert\leq\left(1+\frac{1}{m-2}\right)^{2}\frac{1}{n+1}\dbinom{(n+1)m}{n}.
Proof.

We use the formula in 1.2. The lower bound follows by considering the k=nk=n term. For the upper bound, note that the k=nk=n-\ell term is

+1n+1((n+1)mn)\displaystyle\frac{\ell+1}{n+1}\binom{(n+1)m}{n-\ell} 1n+1((n+1)mn)[(+1)n((n+1)mn)]\displaystyle\leq\frac{1}{n+1}\binom{(n+1)m}{n}\left[\frac{(\ell+1)n^{\ell}}{((n+1)m-n)^{\ell}}\right]
1n+1((n+1)mn)[(+1)(m1)].\displaystyle\leq\frac{1}{n+1}\binom{(n+1)m}{n}\left[(\ell+1)(m-1)^{-\ell}\right].

Summing over 0n0\leq\ell\leq n, we have

|𝒟ngen(m)|1n+1((n+1)mn)(=0(+1)(m1))=(1+1m2)21n+1((n+1)mn),\left\lvert\mathcal{D}_{n}^{\operatorname{gen}(m)}\right\rvert\leq\frac{1}{n+1}\binom{(n+1)m}{n}\cdot\left(\sum_{\ell=0}^{\infty}(\ell+1)(m-1)^{-\ell}\right)=\left(1+\frac{1}{m-2}\right)^{2}\frac{1}{n+1}\dbinom{(n+1)m}{n},

as desired. ∎

3 𝒟nm\mathcal{D}_{n}^{m} for Small mm

In this section, we will prove Theorem 1.3. We also remark that, combining our lower bound and upper bound, we have the following asymptotic estimate for |𝒟nm|\left\lvert\mathcal{D}_{n}^{m}\right\rvert when m=o(n)m=o(n):

Theorem 3.1 (Corollary of Theorem 1.3).

For m(t),n(t):>0>0m(t),n(t)\colon\mathbb{Z}_{>0}\to\mathbb{Z}_{>0} satisfying n(t)n(t)\to\infty and m=o(n)m=o(n), we have

log|𝒟nm|=(1+o(1))log(mnn).\log\left\lvert\mathcal{D}_{n}^{m}\right\rvert=(1+o(1))\log\binom{mn}{n}.

3.1 Upper Bound

We will prove the following upper bound.

Proposition 3.2.

For m(t),n(t):>0>0m(t),n(t)\colon\mathbb{Z}_{>0}\to\mathbb{Z}_{>0} satisfying n(t)n(t)\to\infty, we have

log|𝒟nm|(1+o(1))log(n2m(mnn)).\log\left\lvert\mathcal{D}_{n}^{m}\right\rvert\leq(1+o(1))\log\left(n^{2m}\binom{mn}{n}\right).

Let f1,,fTf_{1},\ldots,f_{T} be a sequence of polynomials in NN variables.

Definition 3.3.

We define a zero-pattern to be a subset SS of [T][T] of the form

S={a|fa(u)=0}S=\left\{a|f_{a}(u)=0\right\}

for some uNu\in\mathbb{C}^{N}.

Our proof closely follows the proof of the following result of Rónyai, Babai, and Ganapathy, which provides a bound on the number of zero-patterns of general sequences of polynomials:

Theorem 3.4.

[5, Theorem 4.1] Let f1,,fTf_{1},\ldots,f_{T} be a sequence of polynomials in NN variables, where each polynomial has degree at most dd. For any tt, the number of zero-patterns is at most

((T0)++(Tt))+(N+(Tt1)dN).\left(\binom{T}{0}+\cdots+\binom{T}{t}\right)+\binom{N+(T-t-1)d}{N}.

Our proof uses the key ideas from the proof of Theorem 3.4, with the main difference being that we can estimate the degrees of polynomials more carefully in the specific case of the sequence {P(j)(λi)}\left\{P^{(j)}(\lambda_{i})\right\}.

Proof of Proposition 3.2.

Since we assume that PP is of degree nn, we may assume PP is monic, as scaling does not affect roots. Let P=xn+an1xn1++a0P=x^{n}+a_{n-1}x^{n-1}+\cdots+a_{0} and Λ=(λ1,,λm)\Lambda=(\lambda_{1},\ldots,\lambda_{m}), where aj,λia_{j},\lambda_{i}\in\mathbb{C} are variables. We view each P(j)(λi)P^{(j)}(\lambda_{i}) as a polynomial in (a0,,an1,λ1,,λm)(a_{0},\ldots,a_{n-1},\lambda_{1},\ldots,\lambda_{m}). Let S1,,SM[m]×[0,n1]S_{1},\ldots,S_{M}\subset[m]\times[0,n-1] denote the zero-patterns of {P(j)(λi)}\{P^{(j)}(\lambda_{i})\}, where we may exclude j=nj=n since P(n)P^{(n)} is the constant polynomial (n1)!(n-1)!.

Call a zero-pattern large if it has size larger than nn, and small otherwise. The number of small zero-patterns is at most

(mn0)+(mn1)++(mnn)n(mnn).\binom{mn}{0}+\binom{mn}{1}+\cdots+\binom{mn}{n}\leq n\binom{mn}{n}.

For each large zero-pattern SkS_{k}, consider the polynomial

Qk(P,Λ):=(i,j)SkP(j)(λi).Q_{k}(P,\Lambda)\vcentcolon=\prod_{(i,j)\not\in S_{k}}P^{(j)}(\lambda_{i}).

We claim that the QkQ_{k} are linearly independent (this is proven in [5, Theorem 1.1]). Suppose, for the sake of contradiction, that some linear combination kαkQk\sum_{k}\alpha_{k}Q_{k} is identically zero, where the αk\alpha_{k} are not all zero. Consider some index \ell that maximizes |S|\left\lvert S_{\ell}\right\rvert over all \ell with α0\alpha_{\ell}\neq 0. Evaluating kαkQk\sum_{k}\alpha_{k}Q_{k} at the (P,Λ)(P,\Lambda) corresponding to the zero pattern SS_{\ell} gives α=0\alpha_{\ell}=0, which is a contradiction, as desired.

Hence, the QkQ_{k}’s are linearly independent. Furthermore, all of the monomials of the QkQ_{k}’s corresponding to large zero-patterns are in the set

{a0b0an1bn1λ1c1λmcmb0+b1++bn1mnn, 0cin2+n2}.\left\{a_{0}^{b_{0}}\cdots a_{n-1}^{b_{n-1}}\cdot\lambda_{1}^{c_{1}}\cdots\lambda_{m}^{c_{m}}\mid b_{0}+b_{1}+\cdots+b_{n-1}\leq mn-n,\ 0\leq c_{i}\leq\frac{n^{2}+n}{2}\right\}.

This is a set of size (mnn)(n2+n+22)m\binom{mn}{n}\cdot\left(\frac{n^{2}+n+2}{2}\right)^{m}, so the QkQ_{k}’s corresponding to large zero-patterns lie in a space of dimension (mnn)(n2+n+22)m\binom{mn}{n}\cdot\left(\frac{n^{2}+n+2}{2}\right)^{m}. Hence, we have

|𝒟nm|(mnn)((n2+n+22)m+n),\left\lvert\mathcal{D}_{n}^{m}\right\rvert\leq\binom{mn}{n}\cdot\left(\left(\frac{n^{2}+n+2}{2}\right)^{m}+n\right),

giving the desired bound. ∎

3.2 Lower Bound

We now establish the lower bound on |𝒟nm|\left\lvert\mathcal{D}_{n}^{m}\right\rvert from Theorem 1.3.

Proposition 3.5.

For m(t),n(t):>0>0m(t),n(t)\colon\mathbb{Z}_{>0}\to\mathbb{Z}_{>0} satisfying n(t)n(t)\to\infty and 1<m(t)n(t)2+n(t)21<m(t)\leq\frac{n(t)^{2}+n(t)}{2}, we have

log|𝒟nm|(1+o(1))log(nm(mnn)).\log\left\lvert\mathcal{D}_{n}^{m}\right\rvert\geq(1+o(1))\log\left(n^{m}\binom{mn}{n}\right).

The idea behind the construction is, for some well chosen aa, to start with an a×(n+1)a\times(n+1) matrix M𝒟ngen(a)M\in\mathcal{D}_{n}^{\operatorname{gen}(a)}, to pick PP and Λ\Lambda such that DP(Λ)=MD_{P}(\Lambda)=M, and then to append mam-a elements to Λ\Lambda. We first prove a claim allowing us to find PP such that many distinct rows can be appended to DP(Λ)D_{P}(\Lambda).

Call an m×(n+1)m\times(n+1) matrix TT-limited if each row has at most TT ones. Call an m×(n+1)m\times(n+1) safe matrix saturated if it has exactly nn ones in total. We let C(m,n,T)C(m,n,T) denote the number of m×(n+1)m\times(n+1) safe TT-limited saturated matrices.

Proposition 3.6.

Let Λ=(λ1,,λa)\Lambda=(\lambda_{1},\ldots,\lambda_{a}) be affinely algebraically independent, and MM be an a×(n+1)a\times(n+1) safe TT-limited saturated matrix. Then there is a degree-nn polynomial PP such that DP(Λ)=MD_{P}(\Lambda)=M. Furthermore, this polynomial PP has the property that for any λ\lambda\in\mathbb{C}, at most TT of the entries of DP((λ))D_{P}((\lambda)) are one.

Proof.

Using Theorem 1.1, we can pick some polynomial PP such that DP(Λ)=MD_{P}(\Lambda)=M. Fix an arbitrary λ\lambda\in\mathbb{C}. We claim that for some b[a]b\in[a], the aa-tuple Λb\Lambda_{b} obtained from Λ\Lambda by replacing λb\lambda_{b} with λ\lambda is affinely algebraically independent. If Λ\Lambda with λ\lambda appended is already affinely algebraically independent, the result is clear.

Otherwise, let Q0Q_{0} be a minimum-degree affine algebraic dependence of (λ,λ1,,λa)(\lambda,\lambda_{1},\ldots,\lambda_{a}). We claim that Q0Q_{0} divides all algebraic dependences. Suppose, for the sake of contradiction, that QQ is another affine algebraic dependence such that Q0Q_{0} does not divide QQ. Viewing Q0,QQ_{0},Q as polynomials in the first variable and taking the resultant gives a nonzero polynomial RR with R(t1+t1λ1,,t1+t2λa)=0R(t_{1}+t_{1}\lambda_{1},\ldots,t_{1}+t_{2}\lambda_{a})=0 for all t1,t2t_{1},t_{2}, contradicting the affine algebraic independence of Λ\Lambda.

Now, if we choose bb such that xbx_{b} appears in Q0Q_{0}, we find that Λb\Lambda_{b} is affinely algebraically independent. For this bb, by Theorem 1.1, DP(Λb)D_{P}(\Lambda_{b}) must have at most nn ones, so the number of ones in DP((λ))D_{P}((\lambda)) is at most the number of ones in the row of λb\lambda_{b}, which is at most TT by assumption, as desired. ∎

Now, we execute the construction mentioned earlier.

Lemma 3.7.

For integers m,n,a,Tm,n,a,T with 0amn2+nT2+T0\leq a\leq m\leq\frac{n^{2}+n}{T^{2}+T}, we have

|𝒟nm|C(a,n,T)(n+1e(T2+T)aen)ma.\left\lvert\mathcal{D}_{n}^{m}\right\rvert\geq C(a,n,T)\cdot\left(\frac{n+1}{e(T^{2}+T)}-\frac{a}{en}\right)^{m-a}.
Proof.

For each a×(n+1)a\times(n+1) safe TT-limited saturated matrix MM, we construct many matrices in 𝒟nm\mathcal{D}_{n}^{m} whose top aa rows are MM as follows:

  1. 1.

    Pick an algebraically independent aa-tuple of complex numbers λ1,,λa\lambda_{1},\ldots,\lambda_{a}. Pick PP as in Proposition 3.6 so that DP((λ1,,λa))=MD_{P}((\lambda_{1},\ldots,\lambda_{a}))=M.

  2. 2.

    For each a+1ima+1\leq i\leq m, iteratively choose λi\lambda_{i} to be any value not equal to any of λ1,,λi1\lambda_{1},\ldots,\lambda_{i-1} so that DP((λi))D_{P}((\lambda_{i})) is nonzero.

We will show that there must be many possible choices for each λi\lambda_{i} in Step 2. We have that λ\lambda is a root of the degree-(n(n+1)/2)\left(n(n+1)/2\right) polynomial j=0nP(j)\prod_{j=0}^{n}P^{(j)} of multiplicity tt(t+1)2\sum_{t}\frac{t(t+1)}{2}, where the summation is over all lengths of runs of ones, with multiplicity, in DP((λ))D_{P}((\lambda)). Since DP((λ))D_{P}((\lambda)) has at most TT nonzero entries, and x(x+1)2\frac{x(x+1)}{2} is convex, λ\lambda is a root of the degree-(n(n+1)/2)\left(n(n+1)/2\right) polynomial j=0nP(j)\prod_{j=0}^{n}P^{(j)} of multiplicity at most T(T+1)/2T(T+1)/2. Note that DP((λ))D_{P}((\lambda)) is nonzero if and only if λ\lambda is a root of j=0nP(j)\prod_{j=0}^{n}P^{(j)}. Hence, the number of possible choices for λi+1\lambda_{i+1} is at least

n(n+1)/2T(T+1)/2i=n2+nT2+Ti.\left\lceil\frac{n(n+1)/2}{T(T+1)/2}\right\rceil-i=\left\lceil\frac{n^{2}+n}{T^{2}+T}\right\rceil-i.

Since P(j)P^{(j)} has at most nn roots for any jj, there are at most nn possible values of λi+1\lambda_{i+1} corresponding to the same not-all-zero row. Thus, given a choice of λ1,,λi,\lambda_{1},\ldots,\lambda_{i}, there are at least 1n(n2+nT2+Ti)\frac{1}{n}\left(\left\lceil\frac{n^{2}+n}{T^{2}+T}\right\rceil-i\right) possibilities for the (i+1)(i+1)th row. For each a×(n+1)a\times(n+1) safe TT-limited saturated matrix, this construction gives at least

i=am1n1(n2+nT2+Ti)\prod_{i=a}^{m-1}n^{-1}\left(\left\lceil\frac{n^{2}+n}{T^{2}+T}\right\rceil-i\right)

elements in 𝒟nm\mathcal{D}_{n}^{m}. For any nonnegative integers b<kb<k, we have

(k(k1)(k+1b))1/b(k(k1)(1))1/kke,\left(k(k-1)\cdots(k+1-b)\right)^{1/b}\geq\left(k(k-1)\cdots(1)\right)^{1/k}\geq\frac{k}{e},

where the first inequality follows from the fact that the left-hand side is decreasing in bb, and the second inequality is justified in Section 1.1. Taking k=n2+nT2+Tak=\left\lceil\frac{n^{2}+n}{T^{2}+T}\right\rceil-a and b=mab=m-a, we find that the number of elements in 𝒟nm\mathcal{D}_{n}^{m} is at least

C(a,n,T)i=am1n1(n2+nT2+Ti)C(a,n,T)nam(n2+nT2+Tae)ma,C(a,n,T)\prod_{i=a}^{m-1}n^{-1}\left(\left\lceil\frac{n^{2}+n}{T^{2}+T}\right\rceil-i\right)\geq C(a,n,T)n^{a-m}\left(\frac{\left\lceil\frac{n^{2}+n}{T^{2}+T}\right\rceil-a}{e}\right)^{m-a},

which exceeds the bound required. ∎

To prove Proposition 3.5, it remains to analyze the size of the bound in Lemma 3.7.

Proposition 3.8.

Suppose positive integers aa and TT with TT divisible by 33 satisfy TnT\leq n and (a2)T/3>n(a-2)T/3>n. Then we have

C(a,n,T)1n+1(ann)(1a(n+1)2T/3).C(a,n,T)\geq\frac{1}{n+1}\binom{an}{n}\left(1-a(n+1)2^{-T/3}\right).
Proof.

The number of a×(n+1)a\times(n+1) safe matrices with exactly tt ones in the first row and nn ones in total is at most

f(t):=(nt)((a1)nnt).f(t)\vcentcolon=\binom{n}{t}\binom{(a-1)n}{n-t}.

Here, the first term counts the number of ways to distribute tt ones into the first row and the second term counts the number of ways to distribute the remaining ones. Here, we are ignoring the condition given by MM being safe, but use the fact that the last column cannot have ones.

We have, for all tn1t\leq n-1,

f(t+1)=f(t)ntt+1nt(a2)n+t<f(t)nt(a2)f(t)T/3t.f(t+1)=f(t)\cdot\frac{n-t}{t+1}\cdot\frac{n-t}{(a-2)n+t}<f(t)\frac{n}{t(a-2)}\leq f(t)\frac{T/3}{t}.

For t2T/3t\geq 2T/3, the function ff decays by a factor of at least 22 when tt increases by 11. The number of a×(n+1)a\times(n+1) safe TT-limited saturated matrices is at least, by Corollary 2.10,

1n+1(a(n+1)n)a(tT+1f(t)),\frac{1}{n+1}\binom{a(n+1)}{n}-a\left(\sum_{t\geq T+1}f(t)\right),

as the number of a×(n+1)a\times(n+1) {0,1}\{0,1\} matrices with at least tt ones in a given row is at most f(t)f(t). Since f(T+i)2iT/3f(2T/3)f(T+i)\leq 2^{-i-T/3}f(2T/3), we get

C(a,n,T)1n+1(a(n+1)n)a2T/3f(2T/3).C(a,n,T)\geq\frac{1}{n+1}\binom{a(n+1)}{n}-a2^{-T/3}f(2T/3).

By Vandermonde’s identity, we have

(a(n+1)n)(ann)=i=0nf(i)f(2T/3).\binom{a(n+1)}{n}\geq\binom{an}{n}=\sum_{i=0}^{n}f(i)\geq f(2T/3).

Hence,

C(a,n,T)1n+1(a(n+1)n)a2T/3f(2T/3)1n+1(a(n+1)n)(1a(n+1)2T/3),C(a,n,T)\geq\frac{1}{n+1}\binom{a(n+1)}{n}-a2^{-T/3}f(2T/3)\geq\frac{1}{n+1}\binom{a(n+1)}{n}\left(1-a(n+1)2^{-T/3}\right),

which exceeds the desired bound. ∎

We are now ready to prove Proposition 3.5.

Proof of Proposition 3.5.

For n(logn)2m\frac{n}{(\log n)^{2}}\geq m, use the lower bound |𝒟nm||𝒟ngen(m)|\left\lvert\mathcal{D}_{n}^{m}\right\rvert\geq\left\lvert\mathcal{D}_{n}^{\operatorname{gen}(m)}\right\rvert to get

log|𝒟nm|log|𝒟ngen(m)|log(1n+1(m(n+1)n))=(1+o(1))log(nm(mnn)),\log\left\lvert\mathcal{D}_{n}^{m}\right\rvert\geq\log\left\lvert\mathcal{D}_{n}^{\operatorname{gen}(m)}\right\rvert\geq\log\left(\frac{1}{n+1}\binom{m(n+1)}{n}\right)=(1+o(1))\log\left(n^{m}\binom{mn}{n}\right),

where the second inequality follows from Corollary 2.10 and the last estimate of the error term follows from bounds log(nm)n/logn=o(n)\log(n^{m})\leq n/\log n=o(n) and log(mnn)log(2nn)=Θ(n)\log\binom{mn}{n}\geq\log\binom{2n}{n}=\Theta(n) .

We use Lemma 3.7 for the remaining regions. For mnlognm\geq n\sqrt{\log n}, take a=na=n and T=1T=1, so C(a,n,T)1C(a,n,T)\geq 1. We have assumed mn2+n2=n2+nT2+Tm\leq\frac{n^{2}+n}{2}=\frac{n^{2}+n}{T^{2}+T}, so we may apply Lemma 3.7, which gives

log|𝒟nm|(mn)log(n+12e12)=(1+o(1))mlogn=(1+o(1))log(nm).\log\left\lvert\mathcal{D}_{n}^{m}\right\rvert\geq(m-n)\log\left(\frac{n+1}{2e}-\frac{1}{2}\right)=(1+o(1))m\log n=(1+o(1))\log(n^{m}).

This is equal to (1+o(1))log(nm(mnn))(1+o(1))\log\left(n^{m}\binom{mn}{n}\right) because we can estimate log(mnn)\log\binom{mn}{n} by

log(mnn)log((mn)n)3nlogn=o(mlogn).\log\binom{mn}{n}\leq\log\left((mn)^{n}\right)\leq 3n\log n=o(m\sqrt{\log n}).

Finally, for n(logn)2mnlogn\frac{n}{(\log n)^{2}}\leq m\leq n\sqrt{\log n}, take T/3=10logn3T/3=10\left\lceil\log n\right\rceil^{3} and a=mT/3+3a=\left\lceil\frac{m}{T/3}\right\rceil+3. For large enough nn, we have mn2+nT2+Tm\leq\frac{n^{2}+n}{T^{2}+T}, so we can apply Lemma 3.7. For large enough nn, we have:

n+1e(T2+T)aenn100T2andmam(11logn),\frac{n+1}{e(T^{2}+T)}-\frac{a}{en}\geq\frac{n}{100T^{2}}\quad\text{and}\quad m-a\geq m\left(1-\frac{1}{\log n}\right),

allowing us to estimate that, using Proposition 3.8, log(|𝒟nm|C(a,n,T))(1+o(1))log(nm)\log\left(\frac{\left\lvert\mathcal{D}_{n}^{m}\right\rvert}{C(a,n,T)}\right)\geq(1+o(1))\log\left(n^{m}\right). For sufficiently large nn, we also have

(ann)(logn)100n(mnn)and1a(n+1)2T/312,\binom{an}{n}\geq(\log n)^{-100n}\binom{mn}{n}\quad\text{and}\quad 1-a(n+1)2^{-T/3}\geq\frac{1}{2},

which will allow us to estimate logC(a,n,t)(1+o(1))log((mnn))\log C(a,n,t)\geq(1+o(1))\log\left(\binom{mn}{n}\right). For nn large enough for the aformentioned inequalities to hold, putting everything together gives

|𝒟nm|12(n+1)(logn)100n(mnn)(n100T2)m(11logn),\left\lvert\mathcal{D}_{n}^{m}\right\rvert\geq\frac{1}{2(n+1)}(\log n)^{-100n}\binom{mn}{n}\cdot\left(\frac{n}{100T^{2}}\right)^{m\left(1-\frac{1}{\log n}\right)},

which is ((mnn)nm)1+o(1)\left(\binom{mn}{n}n^{m}\right)^{1+o(1)}, as desired. ∎

4 𝒟nm\mathcal{D}_{n}^{m} for Large mm

The goal of this section is to prove Theorem 1.4.

Let V(m,n)V(m,n) denote the number of elements in 𝒟nm\mathcal{D}_{n}^{m} with no all-zero rows. As P(j)P^{(j)} has at most njn-j roots for each jj, the jjth column of any element of 𝒟nm\mathcal{D}_{n}^{m} has at most njn-j ones, and so V(m,n)=0V(m,n)=0 for all m>n2+n2m>\frac{n^{2}+n}{2}.

Proposition 4.1.

For all m,nm,n, we have

|𝒟nm|=k=0n2+n2(mk)V(k,n).\left\lvert\mathcal{D}_{n}^{m}\right\rvert=\sum_{k=0}^{\frac{n^{2}+n}{2}}\binom{m}{k}V(k,n).
Proof.

The right-hand side counts |𝒟nm|\left\lvert\mathcal{D}_{n}^{m}\right\rvert by conditioning on the number kk of not-all-zero rows, since the number of possible submatrices consisting of only those kk rows is V(k,n)V(k,n). ∎

We can now conclude the first part of Theorem 1.4.

Corollary 4.2.

We have, for m>n2+n2m>\frac{n^{2}+n}{2},

max((mn2+n2)V(n2+n2,n),|𝒟nn2+n2|)|𝒟nm|(mn2+n2)|𝒟nn2+n2|.\max\left(\binom{m}{\frac{n^{2}+n}{2}}V\left(\frac{n^{2}+n}{2},n\right),\ \left\lvert\mathcal{D}_{n}^{\frac{n^{2}+n}{2}}\right\rvert\right)\leq\left\lvert\mathcal{D}_{n}^{m}\right\rvert\leq\binom{m}{\frac{n^{2}+n}{2}}\left\lvert\mathcal{D}_{n}^{\frac{n^{2}+n}{2}}\right\rvert.

Hence, if m(t),n(t):>0>0m(t),n(t)\colon\mathbb{Z}_{>0}\to\mathbb{Z}_{>0} satisfy m(t),n(t)m(t),n(t)\to\infty, then:

  1. (a)

    if m=Θ(n2)m=\Theta(n^{2}) and m>n2+n2m>\frac{n^{2}+n}{2}, then log|𝒟nm|log|𝒟nn2+n2|\log\left\lvert\mathcal{D}_{n}^{m}\right\rvert\sim\log\left\lvert\mathcal{D}_{n}^{\frac{n^{2}+n}{2}}\right\rvert;

  2. (b)

    if logm=ω(logn)\log m=\omega(\log n), then log|𝒟nm|log((mn2+n2))\log\left\lvert\mathcal{D}_{n}^{m}\right\rvert\sim\log\left(\dbinom{m}{\frac{n^{2}+n}{2}}\right).

Proof.

The lower bound of (mn2+n2)V(n2+n2)\binom{m}{\frac{n^{2}+n}{2}}V\left(\frac{n^{2}+n}{2}\right) follows by taking the last term in Proposition 4.1, and the lower bound of |𝒟nn2+n2|\left\lvert\mathcal{D}_{n}^{\frac{n^{2}+n}{2}}\right\rvert follows from noting that |𝒟nm|\left\lvert\mathcal{D}_{n}^{m}\right\rvert is an increasing function of mm. The upper bound follows from applying the inequality

(mn2+n2)(n2+n2k)(mk)\dbinom{m}{\frac{n^{2}+n}{2}}\cdot\dbinom{\frac{n^{2}+n}{2}}{k}\geq\dbinom{m}{k}

after expanding both sides by Proposition 4.1. The asymptotic estimates follow directly from the inequalities. ∎

Proposition 4.1 also tells us that for fixed nn, |𝒟nm|\left\lvert\mathcal{D}_{n}^{m}\right\rvert is a polynomial in mm of degree n2+n2\frac{n^{2}+n}{2}. We now compute the value of V(n2+n2t,n)V\left(\frac{n^{2}+n}{2}-t,n\right) for t{0,1}t\in\{0,1\}, giving us the leading terms of this polynomial. We will first need the following strengthening of Proposition 3.6.

Proposition 4.3.

Let Λ=(λ1,,λa)\Lambda=(\lambda_{1},\ldots,\lambda_{a}) be affinely algebraically independent, and let MM be an a×(n+1)a\times(n+1) safe matrix. Note that MM is not necessarily saturated.

Let TT be a positive integer such that at most one of the rows of MM has more than TT ones. Then there is a degree-nn polynomial PP with the property that DP(Λ)=MD_{P}(\Lambda)=M, and the property that for any λΛ\lambda\in\mathbb{C}\setminus\Lambda, at most TT of the entries of DP((λ))D_{P}((\lambda)) are one.

Proof.

Append rows of the form (1,0,,0)(1,0,\ldots,0) to MM until it is saturated. Let MM^{\prime} denote the resulting matrix, and let aa^{\prime} be the number of rows of MM^{\prime}. Note that MM^{\prime} is safe. Pick complex numbers λa+1,,λa\lambda_{a+1},\ldots,\lambda_{a^{\prime}} such that Λ=(λ1,,λa)\Lambda^{\prime}=(\lambda_{1},\ldots,\lambda_{a^{\prime}}) is algebraically independent.

Now, we apply the argument of Proposition 3.6 to MM^{\prime} and Λ\Lambda^{\prime}. Consider some λΛ\lambda\in\mathbb{C}\setminus\Lambda. We claim that there exist at least two b[a]b\in[a^{\prime}] such that the aa^{\prime}-tuple Λb\Lambda_{b}^{\prime} obtained from Λ\Lambda^{\prime} by replacing λb\lambda_{b} with λ\lambda is affinely algebraically independent. If Λ\Lambda^{\prime} with λ\lambda appended is already affinely algebraically independent, the result is clear.

Otherwise, let Q0Q_{0} be a minimal-degree affine algebraic dependence of (λ,λ1,,λa)(\lambda,\lambda_{1},\ldots,\lambda_{a^{\prime}}). As in the proof of Proposition 3.6, Q0Q_{0} must divide all other affine algebraic dependences. We claim that there are at least 33 variables with a nonzero coefficient in Q0Q_{0}.

Suppose otherwise. Then, there is some nonzero two-variable polynomial QQ and distinct constants α1,α2\alpha_{1},\alpha_{2}\in\mathbb{C} such that Q(a+bα1,a+bα2)=0Q(a+b\alpha_{1},a+b\alpha_{2})=0 for all a,ba,b\in\mathbb{C}. However, substituting a=α2x+α1yα1α2a=\frac{-\alpha_{2}x+\alpha_{1}y}{\alpha_{1}-\alpha_{2}} and b=xyα1α2b=\frac{x-y}{\alpha_{1}-\alpha_{2}} implies Q(x,y)=0Q(x,y)=0 for all x,yx,y\in\mathbb{C}, so QQ is identically zero, which is a contradiction, as desired.

Now, there are at least two values of bb such that xbx_{b} has nonzero coefficient in Q0Q_{0}. For these values of bb, we have that Λb\Lambda^{\prime}_{b} is affinely algebraically independent. At least one of these choices of bb must have the property that the bbth column of MM^{\prime} has at most TT nonzero entries, implying the desired result. ∎

Proposition 4.4.

We have the following values of V(m,n)V(m,n):

V(n2+n2,n)\displaystyle V\left(\frac{n^{2}+n}{2},n\right) =(n2+n2)!1!2!n!and\displaystyle=\frac{\left(\frac{n^{2}+n}{2}\right)!}{1!2!\cdots n!}\quad\text{and}\quad
V(n2+n2,n1)\displaystyle V\left(\frac{n^{2}+n}{2},n-1\right) =(n2+n2)!1!2!n!(1+(n1)(n2)4).\displaystyle=\frac{\left(\frac{n^{2}+n}{2}\right)!}{1!2!\cdots n!}\left(1+\frac{(n-1)(n-2)}{4}\right).
Proof.

For the first formula, note that if an element of 𝒟nn2+n2\mathcal{D}_{n}^{\frac{n^{2}+n}{2}} has no all-zero rows, then each row must have exactly one nonzero entry and the jjth column must have exactly n+1jn+1-j ones. We show that all matrices satisfying the aforementioned conditions are in 𝒟nn2+n2\mathcal{D}_{n}^{\frac{n^{2}+n}{2}}. This gives the desired enumeration, as we are then placing n+1jn+1-j indistinguishable rows for each j[n]j\in[n].

To show that all such matrices are attainable, consider the polynomial (xλ1)(xλn)(x-\lambda_{1})\cdots(x-\lambda_{n}), where λ1,,λn\lambda_{1},\ldots,\lambda_{n} are affinely algebraically independent. By Proposition 3.6 on P=(xλ1)(xλn)P=(x-\lambda_{1})\cdots(x-\lambda_{n}), Λ=(λ1,,λn)\Lambda=(\lambda_{1},\ldots,\lambda_{n}), and T=1T=1, all of the roots of the derivatives of the P(j)P^{(j)} are distinct, implying the desired result.

For the second formula, note that if an element of 𝒟nn2+n21\mathcal{D}_{n}^{\frac{n^{2}+n}{2}-1} has all not-all-zero rows, then it must be one of the following:

  1. (a)

    an element of 𝒟nn2+n2\mathcal{D}_{n}^{\frac{n^{2}+n}{2}} with no all-zero rows, and with the bottom row removed;

  2. (b)

    a matrix where the jjth column has exactly n+1jn+1-j ones, some row has a nonzero entry in exactly two columns (say n+1jn+1-j and n+1jn+1-j^{\prime} with j>j+1j^{\prime}>j+1), and the remaining rows have exactly one nonzero entry each.

The j>jj^{\prime}>j condition can be assumed by symmetry; the jj+1j^{\prime}\neq j+1 condition arises since P(nj)(t)=P(nj1)(t)=0P^{(n-j^{\prime})}(t)=P^{(n-j^{\prime}-1)}(t)=0 would imply that tt is a multiplicity 22 root of P(nj)P^{(n-j^{\prime})}, giving P(nj)P^{(n-j^{\prime})} more that jj^{\prime} roots counted with multiplicity.

We show that such matrices are attainable. The result is clear for matrices in (a). For matrices in (b), applying Proposition 4.3 to the value T=1T=1, the 11-tuple Λ=(0)\Lambda=(0), and the one-row matrix with ones in columns n+1jn+1-j and n+1jn+1-j^{\prime}, we get a suitable polynomial.

Now, we count this set of matrices. There are V(n2+n2,n)V\left(\frac{n^{2}+n}{2},n\right) matrices in (a), since the bottom row is uniquely determined by the other rows. The number of matrices in (b) is, conditioning on jj and jj^{\prime},

2j+1<jn(n2+n21)!1!2!n!jj\displaystyle\sum_{2\leq j+1<j^{\prime}\leq n}\frac{\left(\frac{n^{2}+n}{2}-1\right)!}{1!2!\cdots n!}\cdot jj^{\prime} =(n2+n21)!1!2!n!j=3njj<j1j\displaystyle=\frac{\left(\frac{n^{2}+n}{2}-1\right)!}{1!2!\cdots n!}\cdot\sum_{j^{\prime}=3}^{n}j^{\prime}\sum_{j<j^{\prime}-1}j
=(n2+n21)!1!2!n!j=3nj(j1)(j2)2\displaystyle=\frac{\left(\frac{n^{2}+n}{2}-1\right)!}{1!2!\cdots n!}\cdot\sum_{j^{\prime}=3}^{n}\frac{j^{\prime}(j^{\prime}-1)(j^{\prime}-2)}{2}
=(n2+n21)!1!2!n!(n+1)n(n1)(n2)8,\displaystyle=\frac{\left(\frac{n^{2}+n}{2}-1\right)!}{1!2!\cdots n!}\cdot\frac{(n+1)n(n-1)(n-2)}{8},

giving the desired formula. ∎

The second part of Theorem 1.4 directly follows.

Corollary 4.5.

When nn is fixed, we have

|𝒟nm|=(n2+n2)!1!2!n!(mn2+n2)+(n2+n2)!1!2!n!(1+(n1)(n2)4)(mn2+n21)+O(mn2+n22).\left\lvert\mathcal{D}_{n}^{m}\right\rvert=\frac{\left(\frac{n^{2}+n}{2}\right)!}{1!2!\cdots n!}\binom{m}{\frac{n^{2}+n}{2}}+\frac{\left(\frac{n^{2}+n}{2}\right)!}{1!2!\cdots n!}\left(1+\frac{(n-1)(n-2)}{4}\right)\binom{m}{\frac{n^{2}+n}{2}-1}+O\left(m^{\frac{n^{2}+n}{2}-2}\right).

5 Future Work

In Theorem 1.3, we believe the lower bound is the correct asymptotic formula for |𝒟nm|\left\lvert\mathcal{D}_{n}^{m}\right\rvert.

Conjecture 5.1.

For m(t),n(t):>0>0m(t),n(t)\colon\mathbb{Z}_{>0}\to\mathbb{Z}_{>0} satisfying n(t)n(t)\to\infty and 1<m(t)n(t)2+n(t)21<m(t)\leq\frac{n(t)^{2}+n(t)}{2}, we have

log|𝒟nm|log(nm(mnn)).\log\left\lvert\mathcal{D}_{n}^{m}\right\rvert\sim\log\left(n^{m}\binom{mn}{n}\right).

If this conjecture is true, then if m=ω(n2)m=\omega(n^{2}), log|𝒟nm|log((mn2+n2)|𝒟nn2+n2|)\log\left\lvert\mathcal{D}_{n}^{m}\right\rvert\sim\log\left(\dbinom{m}{\frac{n^{2}+n}{2}}\cdot\left\lvert\mathcal{D}_{n}^{\frac{n^{2}+n}{2}}\right\rvert\right) follows from the inequalities in Corollary 4.2, since Stirling’s approximation on the formula for V(n2+n2,n)V\left(\frac{n^{2}+n}{2},n\right) in Proposition 4.4 would give log(V(n2+n2,n))log|𝒟nn2+n2|\log\left(V\left(\frac{n^{2}+n}{2},n\right)\right)\sim\log\left\lvert\mathcal{D}_{n}^{\frac{n^{2}+n}{2}}\right\rvert.

Our proof of upper bounds on 𝒟nm\mathcal{D}_{n}^{m} for mn2+n2m\leq\frac{n^{2}+n}{2} gives no information about the elements of the set 𝒟nm\mathcal{D}_{n}^{m}. A natural further direction to investigate is the properties of the matrices in 𝒟nm\mathcal{D}_{n}^{m}.

Question 5.2.

In terms of mm and nn, how many entries can be ones?

6 Acknowledgments

This research was conducted at the University of Minnesota Duluth REU, and was supported by Jane Street Capital, the NSA (grant number H98230-22-1-0015), and the CYAN Undergraduate Mathematics Fund at MIT.

We would like to thank Noah Kravitz for his helpful discussions and guidance throughout the research and writing process. We are also grateful to Brian Lawrence for valuable feedback on this paper, and Sean Li for helpful discussions. Finally, we would like to especially thank Joe Gallian for organizing and running the REU.

References

  • [1] Noga Alon, Noah Kravitz and Kevin O’Bryant “Counting Dope Matrices” In arXiv, 19 May 2022 URL: https://arxiv.org/abs/2205.09302
  • [2] Nachum Dershowitz and Shmuel Zaks “The cycle lemma and some applications” In European J. Combin. 11.1, 1990, pp. 35–40 DOI: 10.1016/S0195-6698(13)80053-4
  • [3] Ira Gessel and Gérard Viennot “Binomial determinants, paths, and hook length formulae” In Adv. in Math. 58.3, 1985, pp. 300–321 DOI: 10.1016/0001-8708(85)90121-5
  • [4] Melvyn B. Nathanson “Interactions of zeros of of polynomials and multiplicity matrices” In arXiv arXiv, 15 March 2022 URL: https://arxiv.org/abs/2203.02477
  • [5] Lajos Rónyai, László Babai and Murali K. Ganapathy “On the number of zero-patterns of a sequence of polynomials” In J. Amer. Math. Soc. 14.3, 2001, pp. 717–735 DOI: 10.1090/S0894-0347-01-00367-8

Department of Mathematics, Massachusetts Institute of Technology, Cambridge, MA 02139

Email Address: [email protected]