This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Hypergraph universality via branching random walks

Rajko Nenadov School of Computer Science, University of Auckland, New Zealand. Email: [email protected]. Research supported by the Marsden Fund of the Royal Society of New Zealand.
Abstract

Given a family of hypergraphs \mathcal{H}, we say that a hypergraph Γ\Gamma is \mathcal{H}-universal if it contains every HH\in\mathcal{H} as a subgraph. For D,rD,r\in\mathbb{N}, we construct an rr-uniform hypergraph with Θ(nrr/Dlogr/D(n))\Theta\left(n^{r-r/D}\log^{r/D}(n)\right) edges which is universal for the family of all rr-uniform hypergraphs with nn vertices and maximum degree at most DD. This almost matches a trivial lower bound Ω(nrr/D)\Omega(n^{r-r/D}) coming from the number of such hypergraphs.

On a high level, we follow the strategy of Alon and Capalbo used in the graph case, that is r=2r=2. The construction of Γ\Gamma is deterministic and based on a bespoke product of expanders, whereas showing that Γ\Gamma is universal is probabilistic. Two key new ingredients are a decomposition result for hypergraphs of bounded density, based on Edmond’s matroid partitioning theorem, and a tail bound for branching random walks on expanders.

1 Introduction

A graph Γ\Gamma is universal for a family of graphs \mathcal{H} if every HH\in\mathcal{H} is a subgraph of Γ\Gamma, not necessarily induced. Constructions of sparse graphs which are universal for families \mathcal{H} of interest has been studied extensively in the past decades. This includes families of graphs with bounded degree [4, 5, 6, 7, 8], which are also the central family considered in this paper, trees and forests [16, 18, 17, 23] and, more generally, graphs with bounded degeneracy [2, 33], as well as families of graphs with additional structural properties such as planar graphs [10, 22] and, more generally, graphs with small separators [13, 14, 15]. Other than being an interesting problem in its own right, motivation for studying sparse universal graphs comes from applications in VLSI circuit design [12], data storage [19], and simulation of parallel computer architecture [11], to name a few.

We are interested in the family (r)(D,n)\mathcal{H}^{(r)}(D,n) consisting of all rr-uniform hypergraphs with nn vertices and maximum degree at most DD. The case r=2r=2 was first studied by Alon, Capalbo, Kohayakawa, Rödl, Ruciński, and Szemerédi [7], where they constructed an (2)(D,n)\mathcal{H}^{(2)}(D,n)-universal graph with Θ(n21/Dlog1/D(n))\Theta\left(n^{2-1/D}\log^{1/D}(n)\right) edges. This was improved in a series of papers [4, 5, 8], culminating with the work of Alon and Capalbo [6] where they constructed a universal graph with Θ(n22/D)\Theta\left(n^{2-2/D}\right) edges. A simple counting argument based on the size of the family (2)(D,n)\mathcal{H}^{(2)}(D,n) gives a lower bound Ω(n22/D)\Omega\left(n^{2-2/D}\right), showing that the construction of Alon and Capalbo is optimal.

The hypergraph case r3r\geq 3 was considered by Parczyk and Person [34] and Hetterich, Parczyk, and Person [27]. By reducing the problem to the graph case, they showed that for even rr there exists an rr-uniform hypergraph (rr-graph for short) with Θ(nrr/D)\Theta\left(n^{r-r/D}\right) edges which is (r)(D,n)\mathcal{H}^{(r)}(D,n)-universal. They also obtained the same bound in the case rr is odd and D=2D=2. Similarly as in the graph case, a simple counting argument based on the size of the family (r)(D,n)\mathcal{H}^{(r)}(D,n) shows that this is optimal. In the case of odd rr and D>2D>2, they constructed a universal hypergraph with Θ(nr(r+1)/Δ)\Theta\left(n^{r-(r+1)/\Delta}\right) edges, where Δ=(r+1)D/r\Delta=\lceil(r+1)D/r\rceil. This falls short of the lower bound Ω(nrr/D)\Omega\left(n^{r-r/D}\right) by a polynomial factor for every such rr and DD. Our main result is the existence of universal rr-graphs which are off only by a small logarithmic factor.

Theorem 1.1.

For every r,Dr,D\in\mathbb{N} there exists C=C(r,D)>0C=C(r,D)>0 such that the following holds: For every nn\in\mathbb{N} there exists an rr-graph Γ\Gamma with

e(Γ)Cnrr/Dlogr/D(n),e(\Gamma)\leq Cn^{r-r/D}\log^{r/D}(n),

which is (r)(D,n)\mathcal{H}^{(r)}(D,n)-universal.

On a high-level, our proof follows the construction based on a product of expander graphs used by Alon and Capalbo [5, 6]. To show that the obtained hypergraph is (r)(D,n)\mathcal{H}^{(r)}(D,n)-universal, we employ two new ingredients. The first one, Lemma 2.1, provides a collection of graphs with a simple structure (namely unicyclic) which together underpin a given hypergraph HH. The construction of Γ\Gamma is tailored to make use of this. The second one is a first step towards generalising a classic result of Gilman [25] on tail bounds for random walks on expanders to branching random walks which follow a given ‘blueprint’ tree with bounded degree. This is used to show that a combination of certain graph homomorphisms together form an injection – and thus an embedding of HH in Γ\Gamma. We consider this tail bound for branching random walks to be our main technical contribution, and an interesting research direction in its own right, thus we describe it next.

1.1 Tail bound for branching random walks

In the simplest form, a random walk of length \ell on a graph GG is defined as follows: Let v1v_{1} be a vertex in V(G)V(G) chosen uniformly at random, and for each i{2,,1}i\in\{2,\ldots,\ell-1\}, sequentially, take viv_{i} to be a neighbour of vi1v_{i-1} chosen uniformly at random. In the case GG has bounded degree, this gives an efficient way of sampling \ell vertices from GG in terms of the number of random bits. Indeed, compared to log(|V(G)|)\ell\log(|V(G)|) bits required to sample \ell vertices completely uniformly and independently, random walk requires only log(|V(G)|)+(1)log(D)\log(|V(G)|)+(\ell-1)\log(D), where DD is the maximum degree in GG. This is of great importance in theoretical computer science, and it prompts the question of how much sampling vertices using a random walk resembles the uniform distribution. While the vertices in a random walk are very much dependent locally, it turns out that globally the two distributions exhibit similar phenomena. This was first observed by Ajtai, Komlós, and Szemerédi [1] who studied the probability that a random walk stays confined to a given subset. Their result was significantly strengthened by Gillman [25], who showed that that if GG is a good expander, then the probability that a random walk hits a given set SV(G)S\subseteq V(G) significantly more that |S|/|V(G)|\ell|S|/|V(G)| times is similar to what Chernoff-Hoeffding inequality gives when all vertices are sampled uniformly and independently. Since then, there has been a great interest in generalising these results and by now there is a large body of research on tail bounds for random walks on expander graphs and, more generally, finite state Markov chains (e.g. see [20, 24, 26, 29, 30, 31, 32, 35, 37]).

We propose the study of similar questions for a certain class of branching random walks. For a rooted tree TT and a graph GG, we define a random TT-walk on GG to be a homomorphism ϕ:TG\phi\colon T\to G given by the following random process. Let t1,,tNt_{1},\ldots,t_{N} be any ordering of the vertices in TT such that t1t_{1} is the root, and if tit_{i} is closer to the root than tjt_{j} then i<ji<j. Choose ϕ(t1)\phi(t_{1}) uniformly at random from V(G)V(G), and for i2i\geq 2, sequentially, choose ϕ(ti)\phi(t_{i}) uniformly at random among the set of neighbours of ϕ(pi)\phi(p_{i}), where pip_{i} is the parent of tit_{i} (thus precedes it in the considered ordering). Note that the choice of the ordering is irrelevant for the outcome of the process, as long as it satisfies the stated property. In this terminology, a random walk of length 1\ell-1 corresponds to a random PP_{\ell}-walk where PP_{\ell} is a path with \ell vertices.

Recall that an (n,d,λ)(n,d,\lambda)-graph is a dd-regular graph with nn vertices, such that the second largest absolute value of its adjacency matrix is at most λ\lambda. The following lemma is our main technical contribution.

Lemma 1.2.

There exist constants K0>1K_{0}>1 and α>0\alpha>0 such that the following holds. Let TT be a rooted tree with maximum degree DD, and suppose GG is an (n,d,λ)(n,d,\lambda)-graph with λ<αd/D\lambda<\alpha d/D. Given a non-empty subset UV(T)U\subseteq V(T) and a vertex xV(G)x\in V(G), let XX denote the number of vertices from UU which are mapped to xx in a random TT-walk on GG. Then 𝔼[X]=|U|/n\mathbb{E}[X]=|U|/n and

Pr[X>K𝔼[X]]e(KK0)𝔼[X],\Pr\left[X>K\mathbb{E}[X]\right]\leq e^{-(K-K_{0})\mathbb{E}[X]},

for every K>K0K>K_{0}.

On the one hand, Lemma 1.2 is weaker than the previously discussed result of Gillman [25] in two ways: (i) we require KK to be sufficiently large, and (ii) we only bound the number of times a particular vertex xV(G)x\in V(G) is being hit, rather than a subset of vertices. On the other hand, it is stronger in the sense that it supports any tree TT and not just a path, and it also counts the number of times a particular subset of vertices of TT hits xx, instead of the whole of TT. In the case of random walks, this corresponds to the number of times a particular set of steps has landed on xx. A common generalisation would be to consider a set of functions ft:V(G)[0,1]f_{t}:V(G)\to[0,1], one for each pair tV(T)t\in V(T), and establish a tail bound for the random variable X=tV(T)ft(ϕ(t))X=\sum_{t\in V(T)}f_{t}(\phi(t)). In the case of random walks, this has been done by Rao and Regev [35]. Another, rather obvious, open problem is to improve the lower bound on K0K_{0} for which the conclusion of Lemma 1.2 holds. We leave these as interesting directions for future research.

2 Decomposition Lemma

Given a hypergraph HH, recall the usual definition of the maximal density of HH:

m(H)=maxHHe(H)v(H).m(H)=\max_{H^{\prime}\subseteq H}\frac{e(H^{\prime})}{v(H^{\prime})}.
Lemma 2.1.

Let HH be an rr-graph with m(H)a/bm(H)\leq a/b, for some a,ba,b\in\mathbb{N} with b>r1b>r-1 and a>b/(r1)a>b/(r-1). Then there exists a family of graphs (that is, 22-graphs) H1,,HaH_{1},\ldots,H_{a} on the vertex set V(H)V(H) such that the following holds:

  1. (D1)

    Each connected component of every HiH_{i} is unicyclic and the maximum degree of HiH_{i} is at most 2D2D, where DD is the maximum degree in HH, and

  2. (D2)

    For each hyperedge hE(H)h\in E(H) there exist forests F1(h)H1,,Fa(h)HaF_{1}^{(h)}\subseteq H_{1},\ldots,F_{a}^{(h)}\subseteq H_{a} on the vertex set hh (recall that hV(H)h\subseteq V(H) is a subset of size rr) such that i=1ae(Fi(h))=b\sum_{i=1}^{a}e(F_{i}^{(h)})=b.

It is crucial that each Fi(h)F_{i}^{(h)} is a forest and not just a unicyclic graph, like what we require in 1. The proof of Lemma 2.1 uses Edmonds’ matroid partitioning theorem [21]. Recall that a finite matroid MM is a pair (E,)(E,\mathcal{I}) where EE is a finite set and \mathcal{I} is a family of subsets of EE with the following properties:

  • \emptyset\in\mathcal{I},

  • If AAEA^{\prime}\subseteq A\subseteq E and AA\in\mathcal{I}, then AA^{\prime}\in\mathcal{I}, and

  • If A,BA,B\in\mathcal{I} and |A|>|B||A|>|B|, then there exists xABx\in A\setminus B such that B{x}B\cup\{x\}\in\mathcal{I}.

The set in \mathcal{I} are referred to as independent sets.

Theorem 2.2 (Edmonds’ partitioning theorem).

Let M=(E,)M=(E,\mathcal{I}) be a finite matroid, and let

k(M)=maxSE|S|r(S),k(M)=\max_{S\subseteq E}\left\lceil\frac{|S|}{r(S)}\right\rceil,

where r(S)r(S) denotes the size of a largest independent set from \mathcal{I} which is contained in SS. Then there exists a partition E=I1I2Ik(M)E=I_{1}\cup I_{2}\cup\ldots\cup I_{k(M)} such that IiI_{i}\in\mathcal{I} for each i{1,,k(M)}i\in\{1,\ldots,k(M)\}.

We are now ready to prove Lemma 2.1.

Proof of Lemma 2.1.

Given an rr-uniform HH with m(H)a/bm(H)\leq a/b, we construct a bipartite graph BB on vertex sets UU and VV as follows. The set VV corresponds to the vertex set V(H)V(H), and for each hyperedge hHh\in H there are bb vertices in UU corresponding to hh. We put an edge between uUu\in U and vVv\in V iff the hyperedge in HH corresponding to uu contains vv.

Let \mathcal{I} be the family of all subsets XUX\subseteq U such that: (i) XX contains at most r1r-1 vertices corresponding to each hyperedge, and (ii) for each nonempty XXX^{\prime}\subseteq X we have |NB(X)||X||N_{B}(X^{\prime})|\geq|X^{\prime}|. We claim that M=(U,)M=(U,\mathcal{I}) is a matroid with \mathcal{I} being the family of independent sets. We trivially have \emptyset\in\mathcal{I}, and by the definition we have that YY\in\mathcal{I} and XYX\subseteq Y implies XX\in\mathcal{I}. It remains to verify the augmentation axiom, that is, for each X,YX,Y\in\mathcal{I} such that |X|<|Y||X|<|Y|, there exists xYXx\in Y\setminus X such that X{x}X\cup\{x\}\in\mathcal{I}. This can be seen as follows. We can assume that if XhXX_{h}\subseteq X and YhYY_{h}\subseteq Y denote subsets which corresponds to vertices associated with a hyperedge hh and |Xh||Yh||X_{h}|\leq|Y_{h}|, then XhYhX_{h}\subseteq Y_{h}. This is because each vertex in YhY_{h} has the same neighbourhood, thus we can reassign vertices such that this holds. By the definition of \mathcal{I} and Hall’s condition, there exist matchings MXM_{X} and MYM_{Y} in BB which saturate XX and YY, respectively. Now MXMYM_{X}\cup M_{Y} contains an augmenting path which gives a matching saturating X{x}X\cup\{x\} for some xYXx\in Y\setminus X. Therefore, |N(X)||X||N(X^{\prime})|\geq|X^{\prime}| for every XX{x}X^{\prime}\subseteq X\cup\{x\}. By the initial assumption, we have that the number of vertices in X{x}X\cup\{x\} corresponding to each hyperedge is, again, at most r1r-1, thus X{x}X\cup\{x\}\in\mathcal{I}.

Now that we have established that MM is a matroid, we can state a claim that is the heart of the proof of the lemma.

Claim 2.3.

There exists a partition U=U1UaU=U_{1}\cup\ldots\cup U_{a} such that U1,,UaU_{1},\ldots,U_{a}\in\mathcal{I}.

Proof of Claim 2.3.

By Edmonds’ partition theorem, it suffices to show |Z|/r(Z)a|Z|/r(Z)\leq a for every ZUZ\subseteq U, where r(Z)r(Z) denotes the rank of ZZ, that is, the size of a largest independent set from MM which is contained in ZZ.

Consider some ZUZ\subseteq U, and let ZZZ^{\prime}\subseteq Z be obtained from ZZ by removing all but at most r1r-1 vertices corresponding to each hyperedge from HH. Then r(Z)r(Z) is equal to the size of a largest matching in BB between ZZ^{\prime} and VV. By König’s theorem, r(Z)=|C|r(Z)=|C| where CC is a smallest vertex cover of the induced bipartite subgraph B[Z,V]B[Z^{\prime},V]. Let ZC=ZCZ_{C}=Z^{\prime}\cap C, Z^C=ZC\hat{Z}_{C}=Z^{\prime}\setminus C, VC=VCV_{C}=V\cap C, and V^C=VVC\hat{V}_{C}=V\setminus V_{C}. Note that there is no edge in BB between Z^C\hat{Z}_{C} and V^C\hat{V}_{C}. Moreover, the smallest vertex cover in B[Z^C,VC]B[\hat{Z}_{C},V_{C}] is of size |VC||V_{C}|, and in B[ZC,V^C]B[Z_{C},\hat{V}_{C}] of size |ZC||Z_{C}|. Therefore, a largest matching between Z^C\hat{Z}_{C} and VCV_{C} is of size |VC||V_{C}|, and a largest matching between between ZCZ_{C} and V^C\hat{V}_{C} is of size |ZC||Z_{C}|. As every hyperedge corresponding to a vertex in Z^C\hat{Z}_{C} is fully contained in VCV_{C}, we conclude

|Z^C|(r1)m(H)|VC|(r1)ab|VC|.|\hat{Z}_{C}|\leq(r-1)\cdot m(H)|V_{C}|\leq\frac{(r-1)a}{b}|V_{C}|. (1)

We have |Z|b|Z|/(r1)|Z|\leq b|Z^{\prime}|/(r-1), thus

|Z|r(Z)=|Z||ZC|+|VC|br1|Z||ZC|+|VC|=br1|ZC|+|Z^C||ZC|+|VC|br1max{1,|Z^C|/|VC|}.\frac{|Z|}{r(Z)}=\frac{|Z|}{|Z_{C}|+|V_{C}|}\leq\frac{b}{r-1}\cdot\frac{|Z^{\prime}|}{|Z_{C}|+|V_{C}|}=\frac{b}{r-1}\cdot\frac{|Z_{C}|+|\hat{Z}_{C}|}{|Z_{C}|+|V_{C}|}\leq\frac{b}{r-1}\cdot\max\{1,|\hat{Z}_{C}|/|V_{C}|\}.

From b/(r1)<ab/(r-1)<a and (1), we conclude |Z|/r(Z)a|Z|/r(Z)\leq a, as desired. ∎

For each hyperedge hHh\in H fix a cyclic ordering of its vertices. Let us denote with h(v)h(v) the successor of a vertex vhv\in h in such an ordering for a hyperedge hh. By the definition of \mathcal{I} and Hall’s condition, for every independent set XX\in\mathcal{I} there exists a matching in BB which saturates XX. Let ϕ:UiV\phi\colon U_{i}\rightarrow V denote such a matching saturating UiU_{i}. We form HiH_{i} by taking an edge {ϕ(u),hu(ϕ(u))}\{\phi(u),h_{u}(\phi(u))\} for each uUiu\in U_{i}, where huHh_{u}\in H is the hyperedge in HH corresponding to uu. A vertex is incident to at most two edges coming from each hyperedge it is part of, and every connected component of HiH_{i} contains at most one cycle (one can think of the obtained graph as being a directed graph with out-degree at most 11), thus 1 is satisifed.A forest Fi(h)F_{i}^{(h)} corresponding to the hyperedge hHh\in H is simply a (possibly empty) collection of paths given by the union of edges {ϕ(u),h(ϕ(u))}\{\phi(u),h(\phi(u))\} for uUiu\in U_{i} corresponding to hh. Note that this is indeed a forest, and not a cycle, as UiU_{i} contains at most r1r-1 vertices from UU corresponding to hh. Each vertex in UU corresponding to hh contributes exactly one edge to some Fi(h)F_{i}^{(h)}, thus 2 holds as well. ∎

3 Branching random walk on expanders

In this section we prove Lemma 1.2. The proof follows the strategy of Rao and Regev [35], with the following lemma being the key new ingredient. This lemma is also the main difference compared to [35], which deals with the simpler case of random PP_{\ell}-walks.

Lemma 3.1.

Let TT be a rooted tree with maximum degree DD, and suppose GG is an (n,d,λ)(n,d,\lambda)-graph, for some λ<d/(210D)\lambda<d/(2^{10}D). Consider a random TT-walk on GG. For a subset WV(T)W\subseteq V(T) and xV(G)x\in V(G), let Ix(W)I_{x}(W) be the indicator random variable for the event that all the vertices in WW are mapped to xx. Then for any xV(G)x\in V(G), a subset UV(T)U\subseteq V(T), and 1k|U|1\leq k\leq|U|, we have

𝔼[W(Uk)Ix(W)]i=1k(k1i1)(28|U|/n)ii!(29Dλd)ki,\mathbb{E}\left[\sum_{W\in\binom{U}{k}}I_{x}(W)\right]\leq\sum_{i=1}^{k}\binom{k-1}{i-1}\frac{(2^{8}|U|/n)^{i}}{i!}\left(2^{9}D\;\frac{\lambda}{d}\right)^{k-i}, (2)

where (Uk)\binom{U}{k} denotes the family of all kk-element subsets of UU.

In the proof of Lemma 3.1 we use the following well known property of random walks on expanders, see, e.g., [28].

Lemma 3.2.

Let GG be an (n,d,λ)(n,d,\lambda)-graph, and consider a random walk starting in a given vertex vV(G)v\in V(G). The probability that after exactly \ell steps we finish in a vertex wV(G)w\in V(G) is at most

1/n+(λ/d).1/n+(\lambda/d)^{\ell}.

The proof of Lemma 3.1 is combinatorial in nature, based on a careful encoding of a depth-first search traversal of the given tree.

Proof of Lemma 3.1.

Let us denote with rr the root of TT. For a subset XV(T)X\subseteq V(T), let XV(T)X^{\uparrow}\subseteq V(T) denote the set of all vertices vTv\in T which lie on the path from some vXv\in X to the root rr, including XX and rr (that is, all vertices ‘above’ XX in a top-down drawing of TT).

Let us first describe the overall strategy of the proof and establish some important notation. For each W(Uk)W\in\binom{U}{k}, we define the ordering σ(W)=(w0,,wk1)\sigma(W)=(w_{0},\ldots,w_{k-1}) of WW such that wiWiw_{i}\not\in W_{i}^{\uparrow} for every i{1,,k1}i\in\{1,\ldots,k-1\}, where Wi={w0,,wi1}W_{i}=\{w_{0},\ldots,w_{i-1}\}. For example, taking σ(W)\sigma(W) to be an ordering induced by the distance of a vertex from the root, tie-breaking in some arbitrary way, satisfies this property. However, it will be important for us that σ(W)\sigma(W) can be encoded efficiently, thus our algorithm for producing σ\sigma is more involved and based on depth-first search. We postpone this until it becomes relevant. Next, let hiWih_{i}\in W_{i}^{\uparrow} be the closest vertex in WiW_{i}^{\uparrow} to wiw_{i}, and let did_{i} denote their distance. Note that did_{i} is a function of ii and WW (that is, of σ(W)\sigma(W)), thus we write di(W)d_{i}(W) to signify this. Conditioned on the outcome of the T[Wi]T[W_{i}^{\uparrow}]-walk ϕ\phi, wiw_{i} is mapped onto the last vertex in a random walk of length did_{i} starting from ϕ(hi)\phi(h_{i}). Note that Pr[ϕ(w0)=x]=1/n\Pr[\phi(w_{0})=x]=1/n as the stationary distribution is uniform, thus by Lemma 3.2 we have

𝔼[Ix(W)]=Pr[Ix(W)=1]1ni=1k1(1n+(λ/d)di).\mathbb{E}[I_{x}(W)]=\Pr[I_{x}(W)=1]\leq\frac{1}{n}\prod_{i=1}^{k-1}\left(\frac{1}{n}+(\lambda/d)^{d_{i}}\right). (3)

We now use the trick of Rao and Regev [35] and ‘unroll’ the right hand side,

𝔼[Ix(W)]f{0,1}k1(1n)k|f|i:fi=1(λ/d)di,\mathbb{E}[I_{x}(W)]\leq\sum_{f\in\{0,1\}^{k-1}}\left(\frac{1}{n}\right)^{k-|f|}\prod_{i\colon f_{i}=1}(\lambda/d)^{d_{i}}, (4)

where |f||f| denotes the number of 1’s in f=(f1,,fk1)f=(f_{1},\ldots,f_{k-1}). Consider some f{0,1}k1f\in\{0,1\}^{k-1}, and let Wf¯={w0}{wi:fi=0}W_{\bar{f}}=\{w_{0}\}\cup\{w_{i}\colon f_{i}=0\}. Now fix some F(Uk|f|)F\in\binom{U}{k-|f|}, and iterate over all sets W(Uk)W\in\binom{U}{k} such that Wf¯=FW_{\bar{f}}=F. Our goal is to show

W(Uk)Wf¯=Fi:fi=1(λ/d)di(W)256k|f|(29Dλd)|f|.\sum_{\begin{subarray}{c}W\in\binom{U}{k}\\ W_{\bar{f}}=F\end{subarray}}\prod_{i\colon f_{i}=1}(\lambda/d)^{d_{i}(W)}\leq 256^{k-|f|}\left(2^{9}D\;\frac{\lambda}{d}\right)^{|f|}. (5)

Together with (4), this implies the desired bound:

𝔼[W(Uk)Ix(W)]\displaystyle\mathbb{E}\left[\sum_{W\in\binom{U}{k}}I_{x}(W)\right] (4)W(Uk)f{0,1}k1(1n)k|f|i:fi=1(λ/d)di(W)\displaystyle\stackrel{{\scriptstyle\eqref{eq:estimate_IxW}}}{{\leq}}\sum_{W\in\binom{U}{k}}\sum_{f\in\{0,1\}^{k-1}}\left(\frac{1}{n}\right)^{k-|f|}\prod_{i\colon f_{i}=1}(\lambda/d)^{d_{i}(W)}
=f{0,1}k1(1n)k|f|F(Uk|f|)W(Uk)Wf¯=Fi:fi=1(λ/d)di(W)\displaystyle=\sum_{f\in\{0,1\}^{k-1}}\left(\frac{1}{n}\right)^{k-|f|}\sum_{F\in\binom{U}{k-|f|}}\sum_{\begin{subarray}{c}W\in\binom{U}{k}\\ W_{\bar{f}}=F\end{subarray}}\prod_{i\colon f_{i}=1}(\lambda/d)^{d_{i}(W)}
(5)f{0,1}k1(28n)k|f|(|U|k|f|)(29Dλd)|f|\displaystyle\stackrel{{\scriptstyle\eqref{eq:estimate_f}}}{{\leq}}\sum_{f\in\{0,1\}^{k-1}}\left(\frac{2^{8}}{n}\right)^{k-|f|}\binom{|U|}{k-|f|}\left(2^{9}D\;\frac{\lambda}{d}\right)^{|f|}
i=0k1(k1i)(28|U|/n)ki(ki)!(29Dλd)i.\displaystyle\leq\sum_{i=0}^{k-1}\binom{k-1}{i}\frac{(2^{8}|U|/n)^{k-i}}{(k-i)!}\left(2^{9}D\;\frac{\lambda}{d}\right)^{i}.

After changing the indexing, we obtain (2).

Let us now outline the strategy for estimating (5). For a fixed f{0,1}k1f\in\{0,1\}^{k-1} with |f|1|f|\geq 1, let If={i:fi=1}I_{f}=\{i\colon f_{i}=1\}. Our aim is to show that there is an injection ψf,F\psi_{f,F} from the family of all W(Uk)W\in\binom{U}{k} such that Wf¯=FW_{\bar{f}}=F to the set of 44-tuples (B,R,(bi)iIf,(𝐜i)iIf)(B,R,(b_{i})_{i\in I_{f}},(\mathbf{c}_{i})_{i\in I_{f}}), where B,R{1,,4k}B,R\subseteq\{1,\ldots,4k\}, bi0b_{i}\in\mathbb{N}_{0}, and 𝐜i\mathbf{c}_{i} is a sequence of finite length with each element being in {1,,D}\{1,\ldots,D\}. The mapping ψf,F\psi_{f,F} also has the following important property: if ψf,F(W)=(B,R,(bi)iIf,(𝐜i)iIf)\psi_{f,F}(W)=(B,R,(b_{i})_{i\in I_{f}},(\mathbf{c}_{i})_{i\in I_{f}}), then di(W)=|𝐜i|max(1,bi)d_{i}(W)=|\mathbf{c}_{i}|\geq\max(1,b_{i}) for every iIfi\in I_{f}. We are not yet in a position to say where such a 44-tuple comes from, other than hint the reader that it is an encoding of a certain traversal of the vertices in WW within TT. Assuming we have such an injection ψf,F\psi_{f,F}, we easily obtain (5):

W(Uk)Wf=Fi:fi=1(λ/d)di(W)\displaystyle\sum_{\begin{subarray}{c}W\in\binom{U}{k}\\ W_{f}=F\end{subarray}}\prod_{i\colon f_{i}=1}(\lambda/d)^{d_{i}(W)} 44k(bi=1di=biDdi(λ/d)di)|f|\displaystyle\leq 4^{4k}\left(\sum_{b_{i}=1}^{\infty}\sum_{d_{i}=b_{i}}^{\infty}D^{d_{i}}(\lambda/d)^{d_{i}}\right)^{|f|}
=44k(bi=1κbi1κ)|f|256k|f|(256κ12κ)|f|,\displaystyle=4^{4k}\left(\sum_{b_{i}=1}^{\infty}\frac{\kappa^{b_{i}}}{1-\kappa}\right)^{|f|}\leq 256^{k-|f|}\left(\frac{256\kappa}{1-2\kappa}\right)^{|f|},

where κ=Dλ/d<1/4\kappa=D\lambda/d<1/4. Now (5) follows from 12κ>1/21-2\kappa>1/2. The importance of the property bi|𝐜i|=di(W)b_{i}\leq|\mathbf{c}_{i}|=d_{i}(W) is evident from the calculation.

It remains to describe ψf,F\psi_{f,F}. To do so, we first describe the algorithm which produces the ordering σ(W)\sigma(W).

The ordering σ(W)\sigma(W).

Fix an arbitrary ordering π\pi of the vertex set of TT, and for each vertex vTv\in T fix an ordering πv\pi_{v} on the set of children of vv. Whenever we refer to the kk-th child of a vertex vTv\in T, we mean kk-th according to πv\pi_{v}. We denote with TvT_{v} the subtree of TT rooted in vv, which we identify with its set of vertices.

Consider some W(Uk)W\in\binom{U}{k}. We define the ordering σ(W)=(w0,,wk1)\sigma(W)=(w_{0},\ldots,w_{k-1}) on WW using the depth-first search over a portion of TT, with a very specific choice of the next vertex to visit. Throughout the procedure we maintain a number of sets, sequences, and indices, initially set as S=(r)S=(r), W^=W\widehat{W}=W, B=R=B=R=\emptyset, and j=0j=0. If rWr\in W, then set w0=rw_{0}=r and j=1j=1, and remove rr from W^\widehat{W}. Intuitively, W^\widehat{W} is the set of vertices in WW which we have not yet visited, SS is the stack which keeps track of important vertices on the path from the root to the current vertex, BB tells us whether in a particular step we continued the exploration in the subtree “below” the current vertex, and RR tells us whether in a particular step in which we removed a vertex from the stack, we stayed in the subtree of the next vertex from the top of the stack (i.e. we moved to the “right” of the current vertex). Finally, whenever we move to a new vertex wjw_{j}, we record how we got to it in terms of a number bib_{i} – recording how many steps to go “back” towards the root from the current vertex – and a sequence 𝐜i\mathbf{c}_{i} – recording the sequence of moves which tell us how to move through the part of TT which has not yet been explored, in order to reach wiw_{i}.

Throughout the procedure we use tt to denote the ordinal number of the current iteration. As long as SS is not empty, repeat:

  1. (1)

    Let ss be the last vertex in SS (that is, the top of the stack).

  2. (2)

    If TsW^T_{s}\cap\widehat{W}\neq\emptyset, add tt to the set BB and choose wjTsW^w_{j}\in T_{s}\cap\widehat{W} to be the closest vertex to ss, tie-breaking according to π\pi. Let djd_{j} denote the distance from ss to wjw_{j} and 𝐜j[D]dj\mathbf{c}_{j}\in[D]^{d_{j}} a sequence describing how to get from ss to uu in TT (recall that TT has maximum degree at most DD and the children of each node are ordered). For completeness, set bj=0b_{j}=0. Add wjw_{j} to the end of SS, and remove it from W^\widehat{W}. Increase jj, and proceed to the next round.

  3. (3)

    Otherwise, we have TsW^=T_{s}\cap\widehat{W}=\emptyset.

    • Remove ss from SS. If SS is now empty terminate the procedure.

    • Let ss^{\prime} be the new vertex on the top of SS (that is, the end of SS). If TsW^=T_{s^{\prime}}\cap\widehat{W}=\emptyset, proceed to the next round.

    • Otherwise, add tt to RR. For each uTsW^u\in T_{s^{\prime}}\cap\widehat{W}, let h(u)h(u) denote the vertex on the path from ss^{\prime} to ss which is closest to uu. Note that it cannot happen that h(u)=sh(u)=s, though it could be h(u)=sh(u)=s^{\prime}.

    • Take uTsW^u\in T_{s^{\prime}}\cap\widehat{W} with h(u)h(u) closest to ss (i.e. furthest away from ss^{\prime}). If there are multiple such vertices take the one which itself is closest to ss, that is, the one for which the path from h(u)h(u) to uu is shortest (tie breaking according to π\pi).

    • If h(u)sh(u)\neq s^{\prime} then add h(u)h(u) to the end of SS.

    • Add uu to the end of SS and set wj=uw_{j}=u. Set bjb_{j} to be the distance from ss to h(u)h(u) (denoting how many steps we need to go “back” towards ss^{\prime}), djd_{j} to be the distance from h(u)h(u) to uu, and 𝐜j[D]dj\mathbf{c}_{j}\in[D]^{d_{j}} the description of how to get from h(u)h(u) to uu in TT. Increase jj, and remove uu from W^\widehat{W}. Proceed to the next round.

Observe the following two crucial properties. First, for every j{0,,k1}j\in\{0,\ldots,k-1\} for which wjw_{j} was defined in (3) we have bjdjb_{j}\leq d_{j}. If this was not the case, then we would have chosen wjw_{j} before wj1w_{j-1}. Second, the procedure terminates after t4kt\leq 4k rounds.

The mapping ψf,F\psi_{f,F}.

Fix f{0,1}k1f\in\{0,1\}^{k-1} with |f|1|f|\geq 1 and F(Uk|f|)F\subseteq\binom{U}{k-|f|}. Consider some W(Uk)W\in\binom{U}{k} such that Wf¯=FW_{\bar{f}}=F. Run the previously described algorithm on WW, and let B,R{1,,4k}B,R\subseteq\{1,\ldots,4k\}, (bi)i[k](b_{i})_{i\in[k]}, and (𝐜i)i[k](\mathbf{c}_{i})_{i\in[k]} be as given by the algorithm upon its termination. Define

ψf,F(W):=(B,R,(bi)iIf,(𝐜i)iIf).\psi_{f,F}(W):=(B,R,(b_{i})_{i\in I_{f}},(\mathbf{c}_{i})_{i\in I_{f}}).

The intuition here is that if we know ff and F=Wf¯F=W_{\bar{f}}, and for each i{1,,k1}i\in\{1,\ldots,k-1\} such that fi=1f_{i}=1 we have enough information on how to reach wiw_{i}, which is precisely what is encoded in ψf,F(W)\psi_{f,F}(W), then we can uniquely reconstruct the whole set WW. Therefore, ψf,F\psi_{f,F} is necessarily injective.

To reconstruct WW from ff, F=Wf¯F=W_{\bar{f}}, and ψf,F(W)\psi_{f,F}(W), we repeat the algorithm for determining σ(W)\sigma(W) with the following modification: in the steps corresponding to choosing wiw_{i} for iIfi\in I_{f}, instead of taking a vertex from WW we follow steps described by bib_{i} and 𝐜i\mathbf{c}_{i}. The sets BB and RR tell us exactly when this is applied. We now make this precise.

Start with S=(r)S=(r), F^=F\widehat{F}=F, and j=0j=0. If rFr\in F, then set w0=rw_{0}=r, j=1j=1, and remove rr from F^\widehat{F}. Throughout the algorithm, we again use tt do denote the ordinal number of the current iteration. As long as SS is not empty, repeat the following:

  1. (i)

    Let ss be the last vertex in SS.

  2. (ii)

    If tBt\in B:

    • If j=0j=0 or fj=0f_{j}=0, then take wjF^w_{j}\in\hat{F} to be the closest vertex to ss, with tie-breaking according to π\pi, and remove it from F^\hat{F}.

    • Otherwise, take wjw_{j} to be the vertex given by following 𝐜j\mathbf{c}_{j} from ss.

    Add wjw_{j} to the end of SS, and increase jj.

  3. (iii)

    Otherwise, we have tBt\notin B. Remove ss from SS, and if SS is empty terminate the procedure. If tRt\notin R, proceed to the next round. Else:

    1. (a)

      If fj=0f_{j}=0 (note that we cannot have j=0j=0 at this point), proceed the same as in the original algorithm: Take uTsF^u\in T_{s^{\prime}}\cap\hat{F} for which h(u)h(u) is the closest to ss, and if there are multiple such vertices take the one for which the path from h(u)h(u) to uu is shortest, tie-breaking according to π\pi. Let hj=h(u)h_{j}=h(u).

    2. (b)

      If fj=1f_{j}=1, then let hjh_{j} be the vertex bjb_{j} steps back from ss towards the root in the tree TT, and then let wjw_{j} be obtained by following 𝐜j\mathbf{c}_{j} from hjh_{j}.

    If hjsh_{j}\neq s^{\prime} add hjh_{j} to the end of SS. Add wjw_{j} to the end of SS, remove it from F^\hat{F} (relevant only if wjw_{j} was obtained in (a)), and increase jj.

By comparing the two algorithms, we see that they output the same ordering σ(W)\sigma(W). Therefore, we can uniquely reconstruct WW from ψf,F(W)\psi_{f,F}(W). ∎

With Lemma 3.1 at hand, the proof of Lemma 1.2 is identical to the proof of [35, Theorem 1.1]. We repeat the argument for convenience of the reader.

Proof of Lemma 1.2.

Consider some xV(G)x\in V(G) and a subset UV(T)U\subseteq V(T). Given a random TT-walk ϕ\phi on GG, let XX denote the number of vertices vUv\in U such that ϕ(v)=x\phi(v)=x. Our aim is to show

𝔼[eX]=q=0𝔼[Xq]q!eK0𝔼[X],\mathbb{E}\left[e^{X}\right]=\sum_{q=0}^{\infty}\frac{\mathbb{E}[X^{q}]}{q!}\leq e^{K_{0}\mathbb{E}[X]}, (6)

from which we derive the desired tail bound using Markov’s inequality,

Pr[X>K𝔼[X]]=Pr[eX>eK𝔼[X]]<𝔼[eX]eK𝔼[X](6)e(KK0)𝔼[X].\Pr\left[X>K\mathbb{E}[X]\right]=\Pr\left[e^{X}>e^{K\mathbb{E}[X]}\right]<\mathbb{E}[e^{X}]e^{-K\mathbb{E}[X]}\stackrel{{\scriptstyle\eqref{eq:moment_gen}}}{{\leq}}e^{-(K-K_{0})\mathbb{E}[X]}.

Recall the notation of Lemma 3.1: given WV(T)W\subseteq V(T), let Ix(W)I_{x}(W) be the indicator random variable for the event that all the vertices in WW are mapped to xx. When W={w}W=\{w\}, we simply write Ix(w)I_{x}(w). Note that X=wUIx(w)X=\sum_{w\in U}I_{x}(w). Consider some qq\in\mathbb{N}. Then

Xq=(wUIx(w))q=k=1q{qk}k!W(Uk)Zx(W),X^{q}=\left(\sum_{w\in U}I_{x}(w)\right)^{q}=\sum_{k=1}^{q}\genfrac{\{}{\}}{0.0pt}{}{q}{k}k!\sum_{W\in\binom{U}{k}}Z_{x}(W),

where {}\genfrac{\{}{\}}{0.0pt}{}{}{} denotes the Stirling number of the second kind. By the linearity of expectation, we have

𝔼[Xq]=k=1q{qk}k!𝔼[W(Uk)Ix(W)].\mathbb{E}[X^{q}]=\sum_{k=1}^{q}\genfrac{\{}{\}}{0.0pt}{}{q}{k}k!\;\mathbb{E}\Bigl{[}\sum_{W\in\binom{U}{k}}I_{x}(W)\Bigr{]}.

Combined with Lemma 3.1, this gives the following upper bound on 𝔼[eX]\mathbb{E}\left[e^{X}\right]:

q=0𝔼[Xq]q!1+q=11q!k=1q{qk}k!i=0k1(k1i)(28𝔼[X])ki(ki)!(29Dλd)i.\sum_{q=0}^{\infty}\frac{\mathbb{E}\left[X^{q}\right]}{q!}\leq 1+\sum_{q=1}^{\infty}\frac{1}{q!}\sum_{k=1}^{q}\genfrac{\{}{\}}{0.0pt}{}{q}{k}k!\sum_{i=0}^{k-1}\binom{k-1}{i}\frac{(2^{8}\mathbb{E}[X])^{k-i}}{(k-i)!}\left(2^{9}D\;\frac{\lambda}{d}\right)^{i}.

Rearranging the sums, we get

1+i=1(28𝔼[X])ii!k=i(k1i1)(29Dλd)kiq=k{qk}k!q!.1+\sum_{i=1}^{\infty}\frac{(2^{8}\mathbb{E}[X])^{i}}{i!}\sum_{k=i}^{\infty}\binom{k-1}{i-1}\left(2^{9}D\;\frac{\lambda}{d}\right)^{k-i}\sum_{q=k}^{\infty}\genfrac{\{}{\}}{0.0pt}{}{q}{k}\frac{k!}{q!}. (7)

Using the following identity [36, Eq. 1.94(b)],

q=k{qk}1q!=(e1)kk!<2kk!,\sum_{q=k}^{\infty}\genfrac{\{}{\}}{0.0pt}{}{q}{k}\frac{1}{q!}=\frac{(e-1)^{k}}{k!}<\frac{2^{k}}{k!},

we further upper bound (7) as

1+i=1(28𝔼[X])ii!2ik=i(k1i1)(29Dλd)ki.1+\sum_{i=1}^{\infty}\frac{(2^{8}\mathbb{E}[X])^{i}}{i!}2^{i}\sum_{k=i}^{\infty}\binom{k-1}{i-1}\left(2^{9}D\;\frac{\lambda}{d}\right)^{k-i}.

Now using the identity

k=i(k1i1)xki=1(1x)i\sum_{k=i}^{\infty}\binom{k-1}{i-1}x^{k-i}=\frac{1}{(1-x)^{i}}

for |x|<1|x|<1, the inner sum further evaluates to

(129Dλd)i<2i.\left(1-2^{9}D\;\frac{\lambda}{d}\right)^{-i}<2^{i}.

We finally get

𝔼[eX]<1+i=1(210𝔼[X])ii!=e210𝔼[X].\mathbb{E}\left[e^{X}\right]<1+\sum_{i=1}^{\infty}\frac{(2^{10}\mathbb{E}[X])^{i}}{i!}=e^{2^{10}\mathbb{E}[X]}.

4 Universal hypergraphs

In this section we prove Theorem 1.1. In fact, we prove a more general result, Theorem 4.1, on universality of bounded-degree hypergraphs with an additional bound on the density. Universality for such a family of graphs was recently studied by Alon et al. [9]. Theorem 4.1 also improves the bound in [9, Theorem 1.4].

Let (r)(q,D,n)\mathcal{H}^{(r)}(q,D,n) the family of all nn-vertex rr-graphs HH with maximum degree at most DD and density m(H)qm(H)\leq q, where m(H)m(H) is as defined in Section 2. An rr-uniform hypergraph HH with maximum degree DD and vv vertices contains at most Dv/rDv/r edges, thus m(H)D/rm(H)\leq D/r. Therefore (r)(D,n)(r)(D/r,D,n)\mathcal{H}^{(r)}(D,n)\subseteq\mathcal{H}^{(r)}(D/r,D,n), thus Theorem 4.1 implies Theorem 1.1.

Theorem 4.1.

For every r,D,nr,D,n\in\mathbb{N} and qq\in\mathbb{Q}, q>1/(r1)q>1/(r-1), there exists C=C(r,D,q)>0C=C(r,D,q)>0 and an rr-graph Γ\Gamma with

e(Γ)Cnr1/qlog1/q(n)e(\Gamma)\leq Cn^{r-1/q}\log^{1/q}(n)

edges which is (r)(q,D,n)\mathcal{H}^{(r)}(q,D,n)-universal.

Proof.

Fix smallest a,ba,b\in\mathbb{N} such that b>r1b>r-1 and q=a/bq=a/b. We first describe the construction of Γ\Gamma, and then prove that it contains every hypergraph H(r)(q,D,n)H\in\mathcal{H}^{(r)}(q,D,n). We say that HH is an (r)(\leq r)-uniform hypergraph, or (r)(\leq r)-graph for short, if every hyperedge in HH has size at most rr. For brevity, throughout the proof we ignore ceilings and floors.

Construction.

Let m=(n/log(n))1/am=(n/\log(n))^{1/a}, and let dd\in\mathbb{N} be sufficiently large with respect to DD. Let GG be an (m,d,λ)(m,d,\lambda)-graph on the vertex set [m][m], for some λ<αd/D\lambda<\alpha d/D where α\alpha is as given by Lemma 1.2. Explicit construction of such a graph, for any mm0(d)m\geq m_{0}(d), was obtained by Alon [3]. Let G2G^{2} be the graph obtained from GG by adding an edge between every two vertices at distance at most 22 in GG. Note that the maximum degree in G2G^{2} is at most d2d^{2}.

We first form an (r)(\leq r)-graph Γ\Gamma^{\prime} as follows: V(Γ)=[m]aV(\Gamma^{\prime})=[m]^{a}, and rrr^{\prime}\leq r vertices 𝐯(1),,𝐯(r)[m]a\mathbf{v}^{(1)},\ldots,\mathbf{v}^{(r^{\prime})}\in[m]^{a}, 𝐯(j)=(v1(j),,va(j))\mathbf{v}^{(j)}=(v_{1}^{(j)},\ldots,v_{a}^{(j)}), form a hyperedge if there exist forests F1,,FaF_{1},\ldots,F_{a}, each on the vertex set [r][r], such that:

  • i=1ae(Fi)b\sum_{i=1}^{a}e(F_{i})\geq b, and

  • for each i[a]i\in[a] there exists a homomorphism fi:[r]{v1(i),,vr(i)}f_{i}\colon[r]\to\{v_{1}^{(i)},\ldots,v_{r^{\prime}}^{(i)}\} of FiF_{i} in G2G^{2}.

This construction is guided by the statement of Lemma 2.1.

Next, take Γ\Gamma to be an rr-graph obtained as the (Clogn)(C\log n)-blowup of Γ\Gamma^{\prime}, for CC being a sufficiently large constant. That is, for each vertex 𝐯V(Γ)\mathbf{v}\in V(\Gamma^{\prime}) we introduce a set B𝐯B_{\mathbf{v}} of size ClognC\log n, and for each hyperedge 𝐯(1)𝐯(r)Γ\mathbf{v}^{(1)}\cdots\mathbf{v}^{(r^{\prime})}\in\Gamma^{\prime}, for some 1rr1\leq r^{\prime}\leq r, we add all the subsets of i=1rB𝐯(i)\bigcup_{i=1}^{r^{\prime}}B_{\mathbf{v}^{(i)}} of size exactly rr as hyperedges in Γ\Gamma.

Let us first count the number of edges in Γ\Gamma^{\prime}. Consider forests F1,,FaF_{1},\ldots,F_{a} on [r][r], such that i=1ae(Fi)b\sum_{i=1}^{a}e(F_{i})\geq b. Then c(F1)++c(Fa)rabc(F_{1})+\ldots+c(F_{a})\leq ra-b, where c(F)c(F) denotes the number of connected components in FF. As the maximum degree in G2G^{2} is d2d^{2}, a homomorphism of a forest FF in G2G^{2} can be chosen in at most mc(F)(d2)rc(F)m^{c(F)}(d^{2})^{r-c(F)} ways. Altogether, this gives at most

F1,,Fai=1amc(Fi)(d2)rc(Fi)rramrab(d2)ra=rad2ra(nlogn)rb/a\sum_{F_{1},\ldots,F_{a}}\prod_{i=1}^{a}m^{c(F_{i})}(d^{2})^{r-c(F_{i})}\leq r^{ra}m^{ra-b}(d^{2})^{ra}=r^{a}d^{2ra}\left(\frac{n}{\log n}\right)^{r-b/a}

hyperedges in Γ\Gamma^{\prime}, where the first sum goes over aa-tuples (F1,,Fa)(F_{1},\ldots,F_{a}) of forests on [r][r] such that c(F1)++c(Fa)rc(F_{1})+\ldots+c(F_{a})\leq r.

Each hyperedge in Γ\Gamma^{\prime} gives rise to less than (rClogn)r(rC\log n)^{r} hyperedges in Γ\Gamma. Therefore, Γ\Gamma has

O(nr1/qlog1/q(n))O\left(n^{r-1/q}\log^{1/q}(n)\right)

hyperedges.

Universality.

We now show that Γ\Gamma is (q,D,n)\mathcal{H}(q,D,n)-universal. Consider an rr-graph H(r)(d,D,n)H\in\mathcal{H}^{(r)}(d,D,n). Let H1,,HaH_{1},\ldots,H_{a} be graphs on the vertex set V(H)V(H) given by Lemma 2.1, and for each edge hHh\in H let F1(h)H1,,Fa(h)HaF_{1}^{(h)}\subseteq H_{1},\ldots,F_{a}^{(h)}\subseteq H_{a} be forests corresponding to 2. For each i[a]i\in[a], let TiT_{i} be a forest on V(H)V(H) with maximum degree at most 2D2D such that HiTi2H_{i}\subseteq T_{i}^{2}, that is, if {v,w}\{v,w\} is an edge in HiH_{i} then {v,w}\{v,w\} are at distance at most 22 in TiT_{i}. Such TiT_{i} can be obtained from HiH_{i} by replacing each cycle x1,,xx_{1},\ldots,x_{\ell} in HiH_{i} by a path x1,x,x2,x1,x3,x2,x_{1},x_{\ell},x_{2},x_{\ell-1},x_{3},x_{\ell-2},\ldots. We use the fact that if f:TiGf\colon T_{i}\to G is a homomorphism of TiT_{i} in GG, then it is also a homomorphism of HiH_{i} in G2G^{2}. Therefore, from now on we can focus on TiT_{i} instead of HiH_{i}. By adding more edges, without loss of generality we may assume each TiT_{i} is a tree.

Our aim is to iteratively find homomorphisms ϕi:TiG\phi_{i}\colon T_{i}\to G such that, for each 𝐯[m]i\mathbf{v}\in[m]^{i}, the set

S𝐯i={wV(H):ϕ1(w)=v1,,ϕi(w)=vi}S_{\mathbf{v}}^{i}=\left\{w\in V(H)\colon\phi_{1}(w)=v_{1},\ldots,\phi_{i}(w)=v_{i}\right\}

is of size

|S𝐯i|n(K/m)i,|S_{\mathbf{v}}^{i}|\leq n(K/m)^{i}, (8)

where KK0K\geq K_{0} is sufficiently large and K0K_{0} is the constant given by Lemma 1.2. Suppose that we have found such homomorphisms ϕ1,,ϕi1\phi_{1},\ldots,\phi_{i-1}, for some i{1,,a}i\in\{1,\ldots,a\}. For simplicity, we define S𝐯0=V(H)S_{\mathbf{v}}^{0}=V(H). Let ϕi\phi_{i} be a random TiT_{i}-walk in GG, with an arbitrary vertex in TiT_{i} being the root. Applying Lemma 1.2 with some xV(G)x\in V(G) and S=S𝐯i1SS=S_{\mathbf{v}}^{i-1}\cup S^{\prime}, for some 𝐯[m]i1\mathbf{v}\in[m]^{i-1} and SS^{\prime} chosen arbitrarily such that |S|=n(K/m)i1|S|=n(K/m)^{i-1} (this is a rather technical detail), we get

Pr[X(x,𝐯)>K|S|/m]e(KK0)|S|/m=o(1/n2),\Pr[X(x,\mathbf{v})>K|S|/m]\leq e^{-(K-K_{0})|S|/m}=o(1/n^{2}),

where X(x,𝐯)X(x,\mathbf{v}) denotes the number of vertices vSv\in S such that ϕi(v)=x\phi_{i}(v)=x. There are mm choices for xx and ma=o(n)m^{a}=o(n) choices for 𝐯\mathbf{v}, thus with positive probability ϕi\phi_{i} is such that X(x,𝐯)n(K/m)iX(x,\mathbf{v})\leq n(K/m)^{i} for every xV(G)x\in V(G) and 𝐯[m]i1\mathbf{v}\in[m]^{i-1}. Therefore, there exists ϕi\phi_{i} for which (8) holds.

Let now f:H[m]af\colon H\to[m]^{a} be defined as f(w)=(ϕ1(w),,ϕa(w))f(w)=(\phi_{1}(w),\ldots,\phi_{a}(w)). As ϕi\phi_{i} is a homomorphism of TiT_{i} into GG and HiTi2H_{i}\subseteq T_{i}^{2}, ϕi\phi_{i} is also a homomorphism of HiH_{i} into G2G^{2}. Consider some edge h={w1,,wr}Hh=\{w_{1},\ldots,w_{r}\}\in H. Then the restriction of ϕi\phi_{i} to Fi(h)F_{i}^{(h)} is a homomorphism of Fi(h)F_{i}^{(h)} in G2G^{2}. Since i=1ae(Fi)b\sum_{i=1}^{a}e(F_{i})\geq b by Lemma 2.1, the set of vertices 𝐯(1),,𝐯(r)\mathbf{v}^{(1)},\ldots,\mathbf{v}^{(r)} given by 𝐯(i)=(ϕ1(wi),,ϕa(wi))\mathbf{v}^{(i)}=(\phi_{1}(w_{i}),\ldots,\phi_{a}(w_{i})) for i[r]i\in[r], form a hyperedge in Γ\Gamma^{\prime}. Note that these vertices might not all be distinct, thus the hyperedge is not necessarily of size rr. Therefore, each hHh\in H is preserved by ff, that is, f(h)f(h) is a hyperedge in Γ\Gamma^{\prime}. By (8), no vertex is an image of more than O(logn)O(\log n) vertices from HH, thus we can turn ff into an injective homomorphism f:HΓf^{\prime}\colon H\to\Gamma, which finally defines a copy of HH in Γ\Gamma. ∎

A reader familiar with some of the previous work [5, 6, 9], most notably the proof of [9, Theorem 1.4], will notice similarity with the overall strategy used in the proof of Theorem 4.1. We note that both the construction of the hypergraph Γ\Gamma and tools used to show it is universal are more involved in our case.

5 Concluding remarks

Towards optimal universal hypergraphs.

The following conjecture, if true, would allow one to replace the use of Lemma 1.2 in the proof of Theorem 1.1 and obtain optimal the optimal bound O(nrr/d)O(n^{r-r/d}).

Conjecture 5.1.

For every DD\in\mathbb{N} there exists dd\in\mathbb{N} and α>0\alpha>0 such that the following holds. Suppose GG is an (n,d,λ)(n,d,\lambda)-graph for λ=Θ(d)\lambda=\Theta(\sqrt{d}). Then for any tree TT with maximum degree at most DD and any partition of V(T)=T1TV(T)=T_{1}\cup\ldots\cup T_{\ell} with |Ti|αn|T_{i}|\leq\alpha n for each I[]I\in[\ell], there exists a homomorphism ϕ:TG\phi\colon T\to G such that the restriction of ϕ\phi to each TiT_{i} is an injection.

In case =1\ell=1, this corresponds to the celebrated result by Friedman and Pippenger [23]. It is important to notice that there is no upper bound on \ell in terms of nn or any other parameter.

Branching random walks.

As discussed in Section 1.1, it would be interesting to extend Lemma 1.2 to other forms of random variables. As a first step, one could consider a random variable which counts the number of times a branching random walk has landed in a set XV(G)X\subseteq V(G) rather than on a particular vertex xV(G)x\in V(G). This would generalise the result of Gillman [25].

The main obstacle in adapting the proof of Lemma 1.2 to this case seems to be in the estimate (3). For appropriately defined random variable, analogous to the one considered in (3), one would aim to upper bound its expectation as μi=1k1(μ+(λ/d)di)\mu\prod_{i=1}^{k-1}(\mu+(\lambda/d)^{d_{i}}), where μ=|X|/n\mu=|X|/n. When |X|=1|X|=1 then this is precisely what is achieved in (3). However, the way this upper bound is derived is based on the ‘worst-case’ conditioning of how a part of TT is mapped, which is too pessimistic and does not yield a desired upper bound for larger set XX. A similar estimate is encountered [35, Lemma 3.3], however the proof there leverages the linear structure of random walks and the ability to algebraically express the desired expectation, and it is not clear whether a similar argument can be applied here.

Another interesting direction would be to consider branching random walks on graphs which are not regular.

Acknowledgment.

The author thanks Noga Alon for stimulating discussions about the problem, and Anders Martinsson for discussions about Lemma 1.2.

References

  • [1] M. Ajtai, J. Komlós, and E. Szemerédi. Deterministic simulation in LOGSPACE. In Proceedings of the Nineteenth Annual ACM Symposium on Theory of Computing, STOC ’87, page 132–140, New York, NY, USA, 1987. Association for Computing Machinery.
  • [2] P. Allen, J. Böttcher, and A. Liebenau. Universality for graphs of bounded degeneracy. arXiv preprint arXiv:2309.05468, 2023.
  • [3] N. Alon. Explicit expanders of every degree and size. Combinatorica, 41(4):447–463, 2021.
  • [4] N. Alon and V. Asodi. Sparse universal graphs. J. Comput. Appl. Math., 142(1):1–11, 2002.
  • [5] N. Alon and M. Capalbo. Sparse universal graphs for bounded-degree graphs. Random Struct. Algorithms, 31(2):123–133, 2007.
  • [6] N. Alon and M. Capalbo. Optimal universal graphs with deterministic embedding. In Proceedings of the nineteenth annual ACM-SIAM symposium on discrete algorithms, SODA 2008, San Francisco, CA, January 20–22, 2008, pages 373–378. New York, NY: Association for Computing Machinery (ACM); Philadelphia, PA: Society for Industrial and Applied Mathematics (SIAM), 2008.
  • [7] N. Alon, M. Capalbo, Y. Kohayakawa, V. Rodl, A. Rucinski, and E. Szemerédi. Universality and tolerance. In Proceedings 41st Annual Symposium on Foundations of Computer Science, pages 14–21. IEEE, 2000.
  • [8] N. Alon, M. Capalbo, Y. Kohayakawa, V. Rödl, A. Ruciński, and E. Szemerédi. Near-optimum universal graphs for graphs with bounded degrees. In International Workshop on Randomization and Approximation Techniques in Computer Science, pages 170–180. Springer, 2001.
  • [9] N. Alon, N. Dodson, C. Jackson, R. McCarty, R. Nenadov, and L. Southern. Universality for graphs with bounded density. arXiv:2311.05500.
  • [10] L. Babai, F. R. K. Chung, P. Erdős, R. L. Graham, and J. H. Spencer. On graphs which contain all sparse graphs. Ann. Discrete Math. 12, 21-26 (1982)., 1982.
  • [11] S. Bhatt, F. Chung, T. Leighton, and A. Rosenberg. Optimal simulations of tree machines. In 27th Annual Symposium on Foundations of Computer Science, pages 274–282. IEEE, 1986.
  • [12] S. Bhatt and T. Leighton. A framework for solving VLSI graph layout problems. J. Comput. Syst. Sci., 28:300–343, 1984.
  • [13] M. Capalbo. Small universal graphs for bounded-degree planar graphs. Combinatorica, 22(3):345–359, 2002.
  • [14] M. R. Capalbo and S. R. Kosaraju. Small universal graphs. In Proceedings of the 31st annual ACM symposium on theory of computing, STOC 1999. Atlanta, GA, USA, May 1–4, 1999, pages 741–749. New York, NY: ACM, Association for Computing Machinery, 1999.
  • [15] F. R. K. Chung. Separator theorems and their applications. Paths, flows, and VLSI-layout, Proc. Meet., Bonn/Ger. 1988, Algorithms Comb. 9, 17-34., 1990.
  • [16] F. R. K. Chung and R. L. Graham. On graphs which contain all small trees. J. Comb. Theory, Ser. B, 24:14–23, 1978.
  • [17] F. R. K. Chung and R. L. Graham. On universal graphs for spanning trees. J. Lond. Math. Soc., II. Ser., 27:203–211, 1983.
  • [18] F. R. K. Chung, R. L. Graham, and N. Pippenger. On graphs which contain all small trees. II. Combinatorics, Keszthely 1976, Colloq. Math. Soc. Janos Bolyai 18, 213-223 (1978)., 1978.
  • [19] F. R. K. Chung, A. L. Rosenberg, and L. Snyder. Perfect storage representations for families of data structures. SIAM J. Algebraic Discrete Methods, 4:548–565, 1983.
  • [20] I. H. Dinwoodie. A probability inequality for the occupation measure of a reversible Markov chain. Ann. Appl. Probab., 5(1):37–43, 1995.
  • [21] J. Edmonds. Minimum partition of a matroid into independent subsets. J. Res. Natl. Bur. Stand., Sect. B, 69:67–72, 1965.
  • [22] L. Esperet, G. Joret, and P. Morin. Sparse universal graphs for planarity. J. Lond. Math. Soc., II. Ser., 108(4):1333–1357, 2023.
  • [23] J. Friedman and N. Pippenger. Expanding graphs contain all small trees. Combinatorica, 7:71–76, 1987.
  • [24] A. Garg, Y. T. Lee, Z. Song, and N. Srivastava. A matrix expander Chernoff bound. In Proceedings of the 50th annual ACM SIGACT symposium on theory of computing, STOC ’18, Los Angeles, CA, USA, June 25–29, 2018, pages 1102–1114. New York, NY: Association for Computing Machinery (ACM), 2018.
  • [25] D. Gillman. A Chernoff bound for random walks on expander graphs. SIAM J. Comput., 27(4):1203–1220, 1998.
  • [26] A. D. Healy. Randomness-efficient sampling within NC1. Comput. Complexity, 17(1):3–37, 2008.
  • [27] S. Hetterich, O. Parczyk, and Y. Person. On universal hypergraphs. Electron. J. Comb., 23(4):research paper p4.28, 16, 2016.
  • [28] S. Hoory, N. Linial, and A. Wigderson. Expander graphs and their applications. Bull. Am. Math. Soc., New Ser., 43(4):439–561, 2006.
  • [29] N. Kahale. Large deviation bounds for Markov chains. Comb. Probab. Comput., 6(4):465–474, 1997.
  • [30] C. A. León and F. Perron. Optimal Hoeffding bounds for discrete reversible Markov chains. Ann. Appl. Probab., 14(2):958–970, 2004.
  • [31] P. Lezaud. Chernoff-type bound for finite markov chains. The Annals of Applied Probability, 8(3):849–867.
  • [32] A. Naor, S. Rao, and O. Regev. Concentration of Markov chains with bounded moments. Ann. Inst. Henri Poincaré, Probab. Stat., 56(3):2270–2280, 2020.
  • [33] R. Nenadov. Ramsey and universality properties of random graphs. PhD thesis, ETH Zurich, 2016.
  • [34] O. Parczyk and Y. Person. Spanning structures and universality in sparse hypergraphs. Random Struct. Algorithms, 49(4):819–844, 2016.
  • [35] S. Rao and O. Regev. A sharp tail bound for the expander random sampler. arXiv:1703.10205.
  • [36] R. P. Stanley. Enumerative Combinatorics. Cambridge Studies in Advanced Mathematics. Cambridge University Press, 2 edition, 2011.
  • [37] R. Wagner. Tail estimates for sums of variables sampled by a random walk. Comb. Probab. Comput., 17(2):307–316, 2008.