This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

A Survey on Parameterized Inapproximability: kk-Clique, kk-SetCover, and More

Xuandi Ren
Peking University
[email protected]
Abstract

In the past a few years, many interesting inapproximability results have been obtained from the parameterized perspective. This article surveys some of such results, with a focus on kk-Clique, kk-SetCover, and other related problems.

1 Introduction

Parameterization and Approximation are two natural ways to cope with NP-complete optimization problems. For Clique and SetCover, two very basic NP-complete problems whose parameterized version kk-Clique and kk-SetCover are also complete problems of W[1] and W[2], both approximation and parameterization have been studied extensively. However, the combining of parameterization and approximation remains unexplored until recent years.

In their breakthrough work, [CCK+17] showed very strong inapproximability results for kk-Clique and kk-SetCover under the hypothesis Gap-ETH. However, although maybe plausible, Gap-ETH is such a strong hypothesis that it already gives a gap in hardness of approximation. Thus it is still of great interest to prove the same lower bound under a gap-free assumption like ETH, W[1]\neqFPT and so on. Although these years have witnessed many significant developments along this way, the inapproximability of kk-Clique and kk-SetCover under gap-free hypotheses is still far beyond settled.

This article surveys some recent results, all of which are beautiful, full of smart ideas, and involve delicate algebraic or combinatorial tools. We hope to extract the relationship between different problems, capture the essence of successful attempts, and convey the ideas inside those results to readers.

1.1 Organization of the Survey

This article is organized by the problems. In Section 2, we put some preliminaries, including the definition of problems and hypotheses. In Section 3, we introduce MaxCover and MinLabel, two problems which are not only important intermediate problems in proving parameterized inapproximability, but also of great interest themselves. In Section 4 and 5, we introduce recent parameterized inapproximability results of kk-Clique and kk-SetCover, respectively.

2 Preliminaries

In this section, we first introduce some concepts in FPT approximation, then briefly describe the problems discussed in this article, and the hypotheses which the results are based on.

2.1 FPT Approximation

For a parameterized optimization problem, we use nn to denote the input size, and the parameter kk usually refers to the number of elements we need to pick to obtain an optimal solution. In some problems kk is just the optimal solution size (e.g. kk-Clique), while in other problems it is not (e.g. One-Sided kk-Biclique). By enumerating the kk elements in the solution, the brute-force algorithm usually runs in time O(nk)O(n^{k}).

An algorithm for a maximization (respectively, minimization) problem is called ρ\rho-FPT-approximation if it runs in f(k)nO(1)f(k)n^{O(1)} time for some computable function ff, and outputs a solution of size at least k/ρk/\rho (respectively, at most kρk\cdot\rho). Here ρ\rho is called approximation ratio. If an optimization problem admits no f(k)f(k)-FPT-approximation for any computable function ff, we say this problem is totally FPT inapproximable.

Note that since computing a constant-size solution is trivial, for a maximization problem, we only care about o(k)o(k)-FPT-approximation and the approximation ratio is measured in terms of only kk. However, for a minimization problem, any computable approximation ratio is non-trivial, so if totally FPT inapproximability is already established, we can also discuss approximation ratio in terms of both kk and nn.

2.2 Problems

If the input is divided into kk groups, and one is only allowed to pick at most one element from each group, we say this problem is colored, otherwise it is uncolored. For some problems (e.g. kk-Clique), the two versions are equivalent, while for some other problems (e.g. kk-Biclique) they are not equivalent at all. We will discuss the coloring in each problem’s section separately.

Now we list the problems considered in this article. There are some other problems (e.g. MaxCover) which are used as intermediate problems in proving hardness of approximation. We put their definitions in their separate sections since they are a bit more complicated.

  • 3SAT. The input is a 3-CNF formula φ\varphi with mm clauses on nn variables. The goal is to decide whether there is a satisfying assignment for φ\varphi.

  • kk-Clique. The input is an undirected graph G=(V,E)G=(V,E) with nn vertices. The goal is to decide whether there is a clique of size kk.

  • Densest kk-Subgraph. The input is an undirected graph G=(V,E)G=(V,E) with nn vertices, and the goal is to find the maximum number of edges induced by kk vertices.

  • kk-SetCover. The input is a collection of nn sets 𝒮={S1,,Sn}\mathcal{S}=\{S_{1},\ldots,S_{n}\} over universe UU. The goal is to decide whether there are kk sets in 𝒮\mathcal{S}, whose union is UU.

2.3 Hypotheses

Here we list the hypotheses which the results are based on.

W[1]\neqFPT and W[2]\neqFPT are arguably the most natural hypotheses in parameterized complexity, and are often used to derive FPT time lower bounds. Since kk-Clique and kk-SetCover are complete problems of W[1] and W[2], respectively, we directly use their intractability results in the statements of those two hypotheses here, and omit the definition of W-Hierarchy.

Hypothesis 2.1 (W[1]\neqFPT).

kk-Clique cannot be solved in f(k)nO(1)f(k)n^{O(1)} time, for any computable function ff.

Hypothesis 2.2 (W[2]\neqFPT).

kk-SetCover cannot be solved in f(k)nO(1)f(k)n^{O(1)} time, for any computable function ff.

Tighter time lower bounds like nΩ(k)n^{\Omega(k)} often involves the Exponential Time Hypothesis (ETH).

Hypothesis 2.3 (Exponential Time Hypothesis (ETH)[IP01, IPZ01, Tov84]).

3SAT cannot be solved deterministically in 2o(n)2^{o(n)} time, where nn is the number of variables. Moreover, this holds even when restricted to formulae in which m=O(n)m=O(n), and each variable appears in at most three clauses.

There are two stronger assumptions on the intractability of 3SAT, namely, the Gap Exponential Time hypothesis (Gap-ETH) and Strong Exponential Time hypothesis (SETH). Gap-ETH is useful in proving strong inapproximability results for many parameterized problems, while SETH is used to show tight time lower bounds like nko(1)n^{k-o(1)}.

Hypothesis 2.4 (Gap Exponential Time Hypothesis (Gap-ETH) [Din16, MR17]).

For some constant ε>0\varepsilon>0, there is no deterministic algorithm which runs in 2o(n)2^{o(n)} time can, given a 3SAT formula on nn variables and m=O(n)m=O(n) clauses, distinguish between the following two cases:

  • (Completeness) the formula is satisfiable.

  • (Soundness) any assignment violates more than ε\varepsilon fraction of clauses.

Note that by current state-of-the-art PCP theorem, a 3SAT formula φ\varphi on nn variables can be transformed into a constant gap 3SAT formula φ\varphi^{\prime} on only npolylog(n)n\text{polylog}(n) variables [Din07]. Therefore, assuming ETH, constant gap 3SAT cannot be solved in 2o(n/polylog(n))2^{o(n/\text{polylog}(n))} time. A big open problem is whether linear-sized PCP exists. If so, Gap-ETH would follow from ETH.

Hypothesis 2.5 (Strong Exponential Time Hypothesis (SETH) [IP01, IPZ01]).

For any ε>0\varepsilon>0, there is an integer k3k\geq 3 such that no algorithm can solve kkSAT deterministically in 2(1ε)n2^{(1-\varepsilon)n} time.

Gap-ETH and SETH both imply ETH. However, no formal relationship between them is known now.

There are also randomized versions of ETH, Gap-ETH and SETH, which also rule out randomized algorithms running in corresponding time. We do not separately list them here.

One last important hypothesis is the Parameterized Inapproximability Hypothesis (PIH).

Hypothesis 2.6 (Parameterized Inapproximability Hypothesis (PIH) [LRSZ20]).

For some constant ε>0\varepsilon>0, there is no (1+ε)(1+\varepsilon) factor FPT approximation algorithm for Colored Densest kk-Subgraph.

The factor (1+ε)(1+\varepsilon) can be replaced by any constant and is not important.

Note that if for a graph the number of edges induced by kk vertices is only (εk2)ε2(k2)\binom{\varepsilon k}{2}\approx\varepsilon^{2}\binom{k}{2}, it can not have a clique of size >εk>\varepsilon k. Thus, PIH implies kk-Clique is hard to approximate within any constant factor in FPT time. However, the reverse direction is not necessarily true (forbiddinng small clique does not imply low density of edges), and it remains an important open question that whether PIH holds if we assume kk-Clique is FPT inapproximable within any constant factor.

Another remark is that PIH can be implied from Gap-ETH. See Appendix A of [BGKM18] for a simple proof. However, deriving PIH from a gap-free hypothesis such as ETH is still open.

3 MaxCover and MinLabel

In this section, we introduce two intermediate problems which are important in proving hardness of kk-Clique and kk-SetCover.

The input is the same for both problems. It is a bipartite graph G=(U˙W,E)G=(U\dot{\cup}W,E), such that UU is partitioned into U=U1˙˙UU=U_{1}\dot{\cup}\ldots\dot{\cup}U_{\ell} and WW is partitioned into W=W1˙˙WhW=W_{1}\dot{\cup}\ldots\dot{\cup}W_{h}. We refer to UiU_{i}’s and WjW_{j}’s as left super-nodes and right super-nodes, respectively, and we refer to the maximum size of left super-nodes and right super-nodes as left alphabet size and right alphabet size, and denote them as |ΣU||\Sigma_{U}| and |ΣW||\Sigma_{W}|, respectively.

We say a MaxCover or MinLabel instance has projection property if for every i[],j[h]i\in[\ell],j\in[h], one of the following holds:

  • Every uUiu\in U_{i} has exactly one neighbor wWjw\in W_{j}.

  • There is a full bipartite graph between UiU_{i} and WjW_{j}.

The bipartite case just means there are no restrictions between UiU_{i} and WjW_{j}.

Another interesting property is called pseudo projection property, which is almost the same as projection property, except the projection direction in the first case is opposite (Every wWjw\in W_{j} has exactly one neighbor uUiu\in U_{i}).

For convenience, in an instance Γ\Gamma and for a left super-node Ui,i[]U_{i},i\in[\ell], we sometimes refer to the number of right super-nodes WjW_{j}’s, such that the edges between UiU_{i} and WjW_{j} do not form a full bipartite graph, as UiU_{i}’s degree. Similarly define it for every right super-nodes. We call the maximum degree over all UiU_{i}’s (respectively, over all WiW_{i}’s), the left degree (respectively, right degree) of Γ\Gamma.

A solution to MaxCover is a subset of vertices SWS\subseteq W formed by picking a vertex from each WjW_{j} (i.e. |SWj|=1|S\cap W_{j}|=1 for all j[h]j\in[h]). We say a labeling SS covers a left super-node UiU_{i} if there exists a vertex uiUiu_{i}\in U_{i} which is a common neighbor of all vertices in SS. The goal in MaxCover is to find a labeling that covers the maximum fraction of left super-nodes. The value of a MaxCover instance is defined as

1(maxlabelingS|{i[]|Ui is covered by S}|)\frac{1}{\ell}\left(\max_{\text{labeling}~{}S}|\{i\in[\ell]|U_{i}\text{ is covered by }S\}|\right)

A solution to MinLabel is also a subset of vertices SWS\subseteq W, but not necessarily one vertex from each WjW_{j}. We say a multi-labeling SS covers a left super-node UiU_{i} if there exists a vertex uiUiu_{i}\in U_{i} which has a neighbor in SWjS\cap W_{j} for every j[h]j\in[h]. The goal in MinLabel is to find a minimum-size multi-labeling SS that covers all the left super-nodes. The value of a MinLabel instance is defined as

1h(minmulti-labeling S which covers every Ui|S|)\frac{1}{h}\left(\min_{\begin{subarray}{c}\text{multi-labeling }S\\ \text{ which covers every }U_{i}\end{subarray}}|S|\right)

There is a remark on the relationship between projection property and pseudo projection property. If the degree of each left super-node UiU_{i} is bounded by some constant (it is the case when reducing from 3SAT, see Theorem 3.2), then a MaxCover instance with projection property can be reduced to a MaxCover instance with pseudo projection property, with a constant shrinking of the gap.

Theorem 3.1.

There is a reduction which, on input a MaxCover instance Γ=(i[]Ui,i[h]Wi,E)\Gamma=(\bigcup_{i\in[\ell]}U_{i},\bigcup_{i\in[h]}W_{i},E) with projection property, and the left degree of Γ\Gamma is bounded by a constant qq, outputs a MaxCover instance Γ=(i[q]Ui,i[h+]Wi,E)\Gamma^{\prime}=(\bigcup_{i\in[\ell\cdot q]}U^{\prime}_{i},\bigcup_{i\in[h+\ell]}W^{\prime}_{i},E^{\prime}) with pseudo projection property, such that

  • (Completeness) If MaxCover(Γ)=1(\Gamma)=1, then MaxCover(Γ)=1(\Gamma^{\prime})=1.

  • (Soundness) If MaxCover(Γ)<1ε(\Gamma)<1-\varepsilon, then MaxCover(Γ)<1ε/q(\Gamma^{\prime})<1-\varepsilon/q.

and the right degree of Γ\Gamma^{\prime} is bounded by qq.

Proof.

The reduction is straightforward: for each restriction between some Ui,i[]U_{i},i\in[\ell] and Wj,j[h]W_{j},j\in[h], build a copy of WjW_{j} on the left (there are at most q\ell\cdot q restrictions, thus that many copies), and the new right super-nodes are just the juxtaposition of UU and WW. Each left super-node is only responsible to check one restriction in Γ\Gamma. The edges between the left and the right WiW_{i} parts form either a bijection or a full bipartite graph, while the edges between the left and the right UiU_{i} parts form either an injection from UU to WW (as they do in Γ\Gamma), or a full bipartite graph, too. It’s easy to see the new instance satisfies pseudo projection property, the right degree is q\leq q, and the gap is only hurt by a factor of qq. ∎

In the following we briefly list the inapproximability results for MaxCover and MinLabel, which will be introduced detailedly in subsequent subsections. We use nn to denote |Σ||\Sigma| for simplicity.

Problem Assumption Ratio Lower Bound Holds Even Reference
MaxCover Gap-ETH Any Constant 2Ω(h)2^{\Omega(h)}
|ΣU|=|ΣW|=O(1)|\Sigma_{U}|=|\Sigma_{W}|=O(1)
=O(h)\ell=O(h)
projection property
/
r/r/\ell nΩ(r)n^{\Omega(r)}
|ΣW|=O(1)|\Sigma_{W}|=O(1)
projection property
[CCK+17]
γ\gamma nΩ(h)n^{\Omega(h)} |ΣU|(1/γ)O(1)|\Sigma_{U}|\leq(1/\gamma)^{O(1)}
W[1]\neqFPT nO(1h)n^{-O(\frac{1}{\sqrt{h}})} f(h)poly(n)f(h)\cdot\text{poly}(n) / [KN21]
ETH nO(1h3)n^{-O(\frac{1}{h^{3}})} nΩ(h)n^{\Omega(h)} /
MinLabel Gap-ETH γ1/h\gamma^{-1/h} nΩ(h)n^{\Omega(h)} |ΣU|(1/γ)O(1)|\Sigma_{U}|\leq(1/\gamma)^{O(1)} [CCK+17]

It’s worth noting that [KLM19] also showed some inapproximability results for MaxCover. However, as their parameters are specifically designed for later use of proving hardness of kk-SetCover, we defer their results to Section 5.3 instead of here.

3.1 Hardness Results Based on Gap-ETH

Results in this subsection are from [CCK+17].

A 3SAT instance can be formulated as MaxCover instance as follows. Each left super-node consists of 7 vertices, which represent the satisfying assignments of a clause. Each right super-node consists of 2 vertices, which correspond to the true/false assignment for a variable. Two vertices are linked if and only if the assignments are consistent. Therefore, Gap-ETH can be also restated as an intractability result of constant gap MaxCover:

Theorem 3.2.

Assuming Gap-ETH, there is a constant ε>0\varepsilon>0 such that no deterministic algorithm can distinguish between the following cases for an instance Γ=(i[]Ui,i[h]Wi,E)\Gamma=(\cup_{i\in[\ell]}U_{i},\cup_{i\in[h]}W_{i},E) in 2o(h)2^{o(h)} time:

  • (Completeness) MaxCover(Γ)=1(\Gamma)=1.

  • (Soundness) MaxCover(Γ)<1ε(\Gamma)<1-\varepsilon.

Moreover, this holds even when |ΣU|,|ΣW|=O(1),=Θ(h)|\Sigma_{U}|,|\Sigma_{W}|=O(1),\ell=\Theta(h) and Γ\Gamma has projection property.

Actually Gap-ETH is equivalent to the above, see Appendix E of [CCK+17].

Note that MaxCover can be solved in O(|ΣU|)O^{*}(|\Sigma_{U}|^{\ell}) or O(|ΣW|h)O^{*}(|\Sigma_{W}|^{h}) time, by just enumerating vertices picked from each left super-nodes or right super-nodes. Moreover, it can even decide whether the answer is r\geq\frac{r}{\ell} in O((r)|ΣU|r)=O((|ΣU|)r)O^{*}(\binom{\ell}{r}|\Sigma_{U}|^{r})=O^{*}((\ell|\Sigma_{U}|)^{r}) or O(|ΣW|h)O^{*}(|\Sigma_{W}|^{h}) time. As shown below, these are the best possible assuming Gap-ETH.

Theorem 3.3 (Theorem 4.2 in [CCK+17]).

Assuming Gap-ETH, there exists constants δ,ρ>0\delta,\rho>0 such that for any rρ\ell\geq r\geq\rho, no algorithm can take a MaxCover instance Γ\Gamma with \ell left super-nodes, distinguish between the following cases in O,r(|Γ|δr)O_{\ell,r}(|\Gamma|^{\delta r}) time:

  • (Completeness) MaxCover(Γ)=1(\Gamma)=1.

  • (Soundness) MaxCover(Γ)<r(\Gamma)<\frac{r}{\ell}.

This holds even when |ΣW|=O(1)|\Sigma_{W}|=O(1) and Γ\Gamma has projection property.

The theorem is straightforward when r=Θ()r=\Theta(\ell), because we can directly compress a constant number of left super-nodes into one. The interesting case is when rr is much smaller than \ell.

The proof involves a combinatorial object called disperser, which is defined as follows.

Definition 3.4 (Disperser[CW89, Zuc96a, Zuc96b]).

For positive integers m,,k,rm,\ell,k,r\in\mathbb{N} and constant ε(0,1)\varepsilon\in(0,1), an (m,,k,r,ε)(m,\ell,k,r,\varepsilon)-disperser is a collection \mathcal{I} of \ell subsets I1,,I[m]I_{1},\ldots,I_{\ell}\subseteq[m], each of size kk, such that the union of any rr different subsets from the collection has size at least (1ε)m(1-\varepsilon)m.

Dispersers with proper parameters can be constructed using random subsets with high probability.

Lemma 3.5.

For positive integers m,,rm,\ell,r\in\mathbb{N} and constant ε(0,1)\varepsilon\in(0,1), let k=3mεrk=\lceil\frac{3m}{\varepsilon r}\rceil and let I1,,II_{1},\ldots,I_{\ell} be random kk-subsets of [m][m]. If lnmr\ln\ell\leq\frac{m}{r} then ={I1,,I}\mathcal{I}=\{I_{1},\ldots,I_{\ell}\} is an (m,,k,r,ε)(m,\ell,k,r,\varepsilon)-disperser with probability at least 1em1-e^{-m}.

This above construction can also be derandomized easily. With this tool, we can compress the left super-nodes in a MaxCover instance according to those subsets. Each left super-node now corresponds to satisfying assignments of the AND of kk clauses. If there is a labeling that covers at least rr left super-nodes, from the definition of disperser we know that at least (1ε)m(1-\varepsilon)m clauses in the original 3SAT instance can be simultaneously satisfied. The size of the new instance is 2O(k)=2O(m/r)2^{O(k)}=2^{O(m/r)}, thus an algorithm for MaxCover which runs in |Γ|o(r)|\Gamma|^{o(r)} time would lead to an algorithm for constant gap 3SAT in 2o(m)2^{o(m)} time, refuting Gap-ETH.

In the other direction, we would like to rule out |Γ|o(h)|\Gamma|^{o(h)} algorithms for approximating MaxCover, where hh is the number of right super-nodes. We have the following theorem:

Theorem 3.6 (Theorem 4.3 in [CCK+17]).

Assuming Gap-ETH, there exists constants δ,ρ>0\delta,\rho>0 such that for any hρh\geq\rho and 1γ>01\geq\gamma>0, no algorithm can take a MaxCover instance Γ\Gamma with hh right super-nodes, distinguish between the following cases in Oh,γ(|Γ|δh)O_{h,\gamma}(|\Gamma|^{\delta h}) time:

  • (Completeness) MaxCover(Γ)=1(\Gamma)=1.

  • (Soundness) MaxCover(Γ)<γ(\Gamma)<\gamma.

This holds even when |ΣU|(1/γ)O(1)|\Sigma_{U}|\leq(1/\gamma)^{O(1)}.

Note that in the statement of this theorem, hh can be any fixed constant, while in the statement of Gap-ETH, hh is the number of variables which goes to infinity. Thus, a straightforward idea is to compress the variables into hh groups, each of size n/hn/h.

After grouping variables, we can also compress the clauses in order to amplify the soundness parameter to γ\gamma. Specifically, let k=ln(1γ)/εk=\ln(\frac{1}{\gamma})/\varepsilon, take =(mk)\ell=\binom{m}{k} left super-nodes, each corresponding to satisfying assignments of the AND of some kk clauses. If only (1ε)m(1-\varepsilon)m clauses in the original 3SAT instance can be satisfied, then only ((1ε)mk)\binom{(1-\varepsilon)m}{k} clauses in the new instance can be satisfied, leading to a soundness parameter ((1ε)mk)/(mk)eεk=γ\binom{(1-\varepsilon)m}{k}/\binom{m}{k}\leq e^{-\varepsilon k}=\gamma. Furthermore, |ΣU|=O(1)k(1/γ)O(1)|\Sigma_{U}|=O(1)^{k}\leq(1/\gamma)^{O(1)}.

One important thing is that we need to make sure \ell and |ΣU||\Sigma_{U}| can be bounded by |ΣW|=2n/h|\Sigma_{W}|=2^{n/h}, so that |ΣW||\Sigma_{W}| is the dominating term in |Γ||\Gamma|. Thus, γ\gamma cannot be arbitrarily small. Fortunately in its major applications (e.g. hardness of SetCover), this will not be the bottleneck.

Next we proceed to discuss hardness of MinLabel.

Theorem 3.7 (Theorem 4.4 in [CCK+17] ).

Assuming Gap-ETH, there exists constants δ,ρ>0\delta,\rho>0 such that, for any hρh\geq\rho and 1γ>01\geq\gamma>0, no algorithm can take a MinLabel instance Γ\Gamma with hh right super-nodes, distinguish between the following cases in Oh,γ(|Γ|δh)O_{h,\gamma}(|\Gamma|^{\delta h}) time:

  • (Completeness) MinLabel(Γ)=1(\Gamma)=1.

  • (Soundness) MinLabel(Γ)>γ1/h(\Gamma)>\gamma^{-1/h}.

This holds even when |ΣU|(1/γ)O(1)|\Sigma_{U}|\leq(1/\gamma)^{O(1)}.

The instance is exactly the one in Theorem 3.6. It is easy to see that in the completeness case, MaxCover=1=1 implies MinLabel=1=1. We only need to additionally argue that, in the soundness case, small MaxCover implies large MinLabel. Prove by contradiction, if MinLabelγ1/h\leq\gamma^{-1/h}, we can fix this multi-labeling, and pick a vertex uniformly at random from each right super-node to form a labeling. By proving the expected fraction of covered left super-nodes γ\geq\gamma, we have MaxCoverγ\geq\gamma. In the following we suppose the optimal right multi-labeling SWS\subseteq W covers left vertices u1U1,,uhUhu_{1}\in U_{1},\ldots,u_{h}\in U_{h}.

MaxCover\displaystyle{\sc MaxCover}\geq 𝔼[1i=1[ui is covered]]\displaystyle\mathbb{E}\left[\frac{1}{\ell}\sum_{i=1}^{\ell}[u_{i}\text{ is covered}]\right]
=\displaystyle= 1i=1j=1h|𝒩(ui,WjS)|1\displaystyle\frac{1}{\ell}\sum_{i=1}^{\ell}\prod_{j=1}^{h}|\mathcal{N}(u_{i},W_{j}\cap S)|^{-1}
\displaystyle\geq 1i=1(1hj=1h|𝒩(ui,WjS)|)h\displaystyle\frac{1}{\ell}\sum_{i=1}^{\ell}\left(\frac{1}{h}\sum_{j=1}^{h}|\mathcal{N}(u_{i},W_{j}\cap S)|\right)^{-h}
\displaystyle\geq 1i=1(|S|h)h\displaystyle\frac{1}{\ell}\sum_{i=1}^{\ell}\left(\frac{|S|}{h}\right)^{-h}
=\displaystyle= γ\displaystyle\gamma

The left alphabet size is (1/γ)O(1)(1/\gamma)^{O(1)}, as in Theorem 3.6.

3.2 Gap Producing via Threshold Graph Composition

Threshold Graph Composition is a powerful gap-producing technique. It was first proposed by Lin in his breakthrough work [Lin18], and has been used to create gap for many parameterized problems [CL19, Lin19, BBE+19, KN21].

At a high level, in TGC we compose an instance which has no gap, with a threshold graph which is oblivious to the instance, to produce a gap instance of that problem. The two main challenges of TGC are the creation of a threshold graph with desired properties, and the right way to compose the input and the threshold graph, respectively.

In this subsection, we introduce a delicate threshold graph, which is constructed via error correcting codes. This graph was proposed by Karthik et al. [KN21], and had many applications such as in proving hardness of MaxCover starting from W[1]\neqFPT or ETH (later in this Section), and in simplifying the proof of kk-SetCover inapproximability in [Lin19] (see Section 5.4).

We first formalize some definitions related to error correcting codes.

Definition 3.8 (Error Correcting Codes).

Let Σ\Sigma be a finite set, for every \ell\in\mathbb{N} a subset C:ΣrΣC:\Sigma^{r}\to\Sigma^{\ell} is an error correcting code with message length rr, block length \ell and relative distance δ\delta if for every x,yΣrx,y\in\Sigma^{r}, Δ(C(x),C(y))δ\Delta(C(x),C(y))\geq\delta. We denote then Δ(C)=δ\Delta(C)=\delta. Here Δ(x,y)=1|{i[]|xiyi}|\Delta(x,y)=\frac{1}{\ell}|\{i\in[\ell]|x_{i}\neq y_{i}\}|.

We sometimes abuse notations and treat an error correcting code as its image, i.e., CΣC\subset\Sigma^{\ell}.

Definition 3.9 (Collision Number).

The collision number of an error correcting code CC is the smallest number ss such that there exists a set SCS\subseteq C with |S|=s|S|=s, and for every j[]j\in[\ell], there are two strings x,ySx,y\in S such that xj=yjx_{j}=y_{j}. We denote this number as Col(C)Col(C).

For any error correcting code C:ΣrΣC:\Sigma^{r}\to\Sigma^{\ell} and any kk\in\mathbb{N}, we can take it to build a bipartite threshold graph G=(A˙B,E)G=(A\dot{\cup}B,E) with the following properties efficiently:

  • AA is divided into kk groups, each of size |Σ|r|\Sigma|^{r}. BB is divided into \ell groups, each of size |Σ|k|\Sigma|^{k}.

  • (Completeness) For any a1A1,akAka_{1}\in A_{1},\ldots a_{k}\in A_{k} and for any j[]j\in[\ell], there is a unique vertex bBjb\in B_{j} which is a common neighbor of {a1,,ak}\{a_{1},\ldots,a_{k}\}.

  • (Soundness) For every i[k]i\in[k] and every distinct aaAia\neq a^{\prime}\in A_{i}, for Δ(C)\Delta(C)\cdot\ell of the parts j[]j\in[\ell], we have that 𝒩(a)𝒩(a)j=\mathcal{N}(a)\cap\mathcal{N}(a^{\prime})\cap\mathcal{B}_{j}=\emptyset.

  • (Collision Property) Let XAX\subseteq A such that for every j[]j\in[\ell], there exists bBjb\in B_{j} which is a common neighbor of (at least) k+1k+1 vertices in XX. Then |X|Col(C)|X|\geq Col(C).

The graph is constructed as follows:

  • For every i[k]i\in[k], we associate AiA_{i} with all codewords in CC, i.e. each vertex in AiA_{i} is a unique codeword in the image of CC.

  • For every j[]j\in[\ell], we associate BjB_{j} with the set Σk\Sigma^{k}.

  • A vertex aAia\in A_{i} and a vertex bBjb\in B_{j} are linked if and only if aj=bia_{j}=b_{i}.

We can think of the graph as an k×k\times\ell matrix, when picking vertices from each Ai,i[k]A_{i},i\in[k], we are filling the codewords into each row of the matrix. By reading out each column of the matrix, we can pick exactly one common neighbor of them in each Bj,j[]B_{j},j\in[\ell], satisfying the completeness property.

The soundness property is also straightforward: if we pick two vertices a,aa,a^{\prime} from the same left group AiA_{i}, the two codewords differ in at least Δ(C)\Delta(C)\cdot\ell positions. For those columns we cannot “read out” any kk-bit string whose ii-th bit equals to two different characters.

As for the collision property, for a set XAX\subseteq A, if for every j[]j\in[\ell] there exists bBjb\in B_{j} which is a common neighbor of at least k+1k+1 vertices in XX, it’s easy to see that for any j[]j\in[\ell], we can pick x,yXx,y\in X such that xj=yjx_{j}=y_{j} by pigeonhole principle.

Now we describe how to compose this threshold graph with a kk-MaxCover instance Γ\Gamma where the parameter kk denotes the number of right super-nodes, to produce a Gap kk-MaxCover instance Γ\Gamma^{\prime} such that:

  • (Completeness) If MaxCover(Γ)=1(\Gamma)=1, then MaxCover(Γ)=1(\Gamma^{\prime})=1.

  • (Soundness) If MaxCover(Γ)<1(\Gamma)<1, then MaxCover(Γ)1Δ(C)(\Gamma^{\prime})\leq 1-\Delta(C).

Given a kk-MaxCover instance Γ=G=(U˙W,E)\Gamma=G=(U\dot{\cup}W,E) with pseudo projection property, where W=W1˙˙WkW=W_{1}\dot{\cup}\ldots\dot{\cup}W_{k} and U=U1˙˙UtU=U_{1}\dot{\cup}\ldots\dot{\cup}U_{t}, and a threshold graph G=(A˙B,E)G^{\prime}=(A\dot{\cup}B,E^{\prime}), where A=A1˙˙AtA=A_{1}\dot{\cup}\ldots\dot{\cup}A_{t} and B=B1˙˙BB=B_{1}\dot{\cup}\ldots\dot{\cup}B_{\ell}, w.l.o.g. assume |Σ|rmaxi=1t|Ui||\Sigma|^{r}\geq\max_{i=1}^{t}|U_{i}|, we build the new kk-MaxCover instance Γ\Gamma^{\prime} as follows.

  • Arbitrarily match every vertex uiUiu_{i}\in U_{i} to a vertex in aiAia_{i}\in A_{i} without repetitions. This can be done since |Σ|rmaxi=1t|Ui||\Sigma|^{r}\geq\max_{i=1}^{t}|U_{i}|.

  • The new right super-nodes are W1WkW_{1}\ldots W_{k}, and the new left super-nodes are B1BB_{1}\ldots B_{\ell}.

  • A right vertex wWiw\in W_{i} and a left vertex bBjb\in B_{j} are linked if and only if there exists u1U1,,utUtu_{1}\in U_{1},\ldots,u_{t}\in U_{t} such that wiw_{i} is linked to each u1utu_{1}\ldots u_{t} in GG, and bb is linked to the matching a1ata_{1}\ldots a_{t} in GG^{\prime}.

The completeness case is obvious, by picking one ww in each right super-node WiW_{i}, there is a common neighbor in each left super-node UiU_{i}. Consider their matching vertices a1ata_{1}\ldots a_{t}, there is exactly one common neighbor in each BiB_{i}.

The soundness case needs the pseudo projection property of GG, i.e., for every i[k]i\in[k] and jtj\in t, edges between WiW_{i} and UjU_{j} either form a function, or are complete. Fix any labeling w1W1,,wkWkw_{1}\in W_{1},\ldots,w_{k}\in W_{k}, there must be a super-node UjU_{j} which cannot be covered. This means there must be two left vertices w,ww,w^{\prime} mapping to different vertices in UjU_{j}. Let the two vertices be u,uUju,u^{\prime}\in U_{j}, and let the matching vertices of them be a,aAja,a^{\prime}\in A_{j}, by the soundness property of threshold graph, only in 1Δ(C)1-\Delta(C) fraction of parts j[]j\in[\ell] is there a common neighbor of a,aa,a^{\prime} in BjB_{j}. This means the labeling {w1,,wk}\{w_{1},\ldots,w_{k}\} can only cover (1Δ(C))(1-\Delta(C))\cdot\ell parts of BB, i.e., MaxCover(Γ)1Δ(C)(\Gamma^{\prime})\leq 1-\Delta(C).

Next we use this technique to prove strong inapproximability results of kk-MaxCover based on W[1]\neqFPT and ETH.

Theorem 3.10 (Theorem 4.3 in [KN21] ).

Assuming W[1]\neqFPT, for any computable function ff, there is no f(k)poly(n)f(k)\cdot\text{poly}(n) time algorithm which can take a MaxCover instance Γ\Gamma with kk right super-nodes, distinguish between the following two cases:

  • (Completeness) MaxCover(Γ)=1(\Gamma)=1.

  • (Soundness) MaxCover(Γ)nO(1k)(\Gamma)\leq n^{-O(\frac{1}{\sqrt{k}})}.

Proof.

First reduce kk-Clique to MaxCover with K=(k2)K=\binom{k}{2} right super-nodes and kk left super-nodes in the canonical way. Note that this MaxCover instance has pseudo projection property. Then take a Reed-Solomon Code to build the threshold graph. To ensure the right alphabet size (which is |Σ|k|\Sigma|^{k} here) n\leq n, we need |Σ|n1k|\Sigma|\leq n^{\frac{1}{k}}, and to ensure |Σ|rmaxi=1t|Ui|=n|\Sigma|^{r}\geq\max_{i=1}^{t}|U_{i}|=n, we need rlog|Σ|n=kr\geq\log_{|\Sigma|}n=k. Thus according to the properties of Reed-Solomon Code, the soundness parameter is 1Δ(C)=1(1r|Σ|)=nO(1k)=nO(1K)1-\Delta(C)=1-(1-\frac{r}{|\Sigma|})=n^{-O(\frac{1}{k})}=n^{-O(\frac{1}{\sqrt{K}})}. ∎

Theorem 3.11 (Theorem 4.4 in [KN21] ).

Assuming ETH, there is no no(k)n^{o(k)} time algorithm which can take a MaxCover instance Γ\Gamma with kk right super-nodes, distinguish between the following two cases:

  • (Completeness) MaxCover(Γ)=1(\Gamma)=1.

  • (Soundness) MaxCover(Γ)nO(1k3)(\Gamma)\leq n^{-O(\frac{1}{k^{3}})}.

Proof.

First reduce 3SAT to MaxCover with kk right super-nodes and t=(k1)+(k2)+(k3)t=\binom{k}{1}+\binom{k}{2}+\binom{k}{3} left super-nodes. Each right super-node corresponds to satisfying assignments of some m/km/k clauses, and thus has N=2Θ(n/k)N=2^{\Theta(n/k)} vertices in it. Each left super-node corresponds to assignments to variables which appear in exactly some one/two/three groups of clauses. This MaxCover instance also has pseudo projection property. Note that here it’s necessary to group the variables to change the number of left super-nodes from nn to (k1)+(k2)+(k3)\binom{k}{1}+\binom{k}{2}+\binom{k}{3}, because in our construction of threshold graph, the size of each new right super-node BB is |Σ|t|\Sigma|^{t}, which is too large if t=nt=n. After that, we still use Reed-Solomon Codes. Now to ensure |Σ|tN|\Sigma|^{t}\leq N we need |Σ|2n/k4|\Sigma|\leq 2^{n/k^{4}}, and to ensure |Σ|rN|\Sigma|^{r}\geq N we need rlog|Σ|N=Ω(k3)r\geq\log_{|\Sigma|}N=\Omega(k^{3}). The soundness parameter is 1Δ(C)=1(1r|Σ|)=k32n/k4=NO(1k3)1-\Delta(C)=1-(1-\frac{r}{|\Sigma|})=\frac{k^{3}}{2^{n/k^{4}}}=N^{-O(\frac{1}{k^{3}})}. ∎

4 kk-Clique

Clique is arguably the first natural combinatorial optimization problem. Its inapproximability in the NP regime is studied extensively [BGLR93, BS94, Has96, FGL+96, Gol98, FK00, Zuc07]. However, in the parameterized perspective, although kk-Clique is known to be the complete problem of W[1], and even cannot be solved in no(k)n^{o(k)} time assuming ETH [CHKX06], there is still a lot of work to do on its parameterized inapproximability.

To approximate kk-Clique to a factor of ρ\rho, we only need to compute a clique of size k/ρk/\rho, which can be trivially done in nk/ρn^{k/\rho} time. In their milestone work, Chalermsook et al. [CCK+17] showed this cannot be improved assuming Gap-ETH. However, results based on non-gap assumptions are not reached until very recently Lin [Lin21] showed constant approximating kk-Clique is W[1]-hard. He also obtained an nΩ(logk5)n^{\Omega(\sqrt[5]{\log k})} lower bound for constant approximating kk-Clique based on ETH, and this bound was recently improved to nΩ(logk)n^{\Omega(\log k)} by [LRSW21].

The following table lists current state-of-the-art inapproximability results of kk-Clique based on different hypotheses. Here ff can be any computable function.

Complexity Assumption Inapproximability Ratio Time Lower Bound Reference
W[1]\neqFPT Any constant f(k)poly(n)f(k)\cdot\text{poly}(n) [Lin21]
PIH Any constant f(k)poly(n)f(k)\cdot\text{poly}(n) /
ETH Any constant f(k)nΩ(logk)f(k)\cdot n^{\Omega(\log k)} [LRSW21]
ko(1)k^{o(1)} f(k)poly(n)f(k)\cdot\text{poly}(n)
Gap-ETH ρ=o(k)\rho=o(k) f(k)nΩ(k/ρ)f(k)\cdot n^{\Omega(k/\rho)} [CCK+17]

We shall notice that the colored version and uncolored version of kk-Clique are equivalent, because a colored version can be interpreted as an uncolored version by leaving each group an independent set, and we can make kk different copies of the original graph to transform an uncolored version to a colored version.

4.1 Reduction from MaxCover with Projection Property

The Gap-ETH hardness of kk-Clique directly follows from Theorem 3.3. Since the instance has projection property, two left vertices agree if and only if they map to the same vertex in each right super-node, and kk left vertices agree if and only if they pairwise agree. Thus, we can transform a MaxCover instance in Theorem 3.3 with kk left super-nodes to an kk-Clique instance with the same value.

Therefore, as pointed out in Theorem 3.3, even deciding if there is a clique of size r\geq r needs nΩ(r)n^{\Omega(r)} time. Here rr can be any ω(1)\omega(1), which means approximating kk-Clique to any ρ=k/r=o(k)\rho=k/r=o(k) ratio cannot be done in f(k)no(r)=f(k)no(k/ρ)f(k)\cdot n^{o(r)}=f(k)\cdot n^{o(k/\rho)} time.

Note the projection property is crucial, without which a MaxCover instance cannot be reduced to an kk-Clique instance, because the agreement test of kk left vertices cannot be decomposed locally to agreement tests of (k2)\binom{k}{2} pairs of left vertices.

One interesting thing is that the optimal inapproximability result of kk-Clique can be obtained from hypotheses other than (but similar to) Gap-ETH, see the following as an example.

Theorem 4.1.

Given an undirected graph with nn groups of O(1)O(1) vertices, each forming an independent set, if there exists a constant ε>0\varepsilon>0 such that distinguishing between the following cases cannot be done in 2o(n)2^{o(n)} time:

  • (Completeness) there is a clique of size nn.

  • (Soundness) there is no clique of size εn\varepsilon n.

then kk-Clique cannot be approximated to any ρ=o(k)\rho=o(k) ratio in f(k)no(k/ρ)f(k)\cdot n^{o(k/\rho)} time.

This assumption is a little weaker than Gap-ETH since it can be obtained through the canonical reduction from 3SAT to Clique.

The proof is almost the same as that in Theorem 3.3: just compose the groups using a disperser. Details are omitted here.

4.2 kk-VectorSum

Before introducing W[1]-hardness of constant approximating kk-Clique, we want to mention an important W[1]-complete problem, kk-VectorSum, which is used as an intermediate problem in the reduction in [Lin21].

In the kk-VectorSum problem, we are given kk groups of vectors V1,,Vk𝔽dV_{1},\ldots,V_{k}\subseteq\mathbb{F}^{d} together with a target vector t𝔽d\vec{t}\in\mathbb{F}^{d}, where 𝔽\mathbb{F} is some finite field and dd is the dimension of vectors. The goal is to decide whether there exists vectors v1V1,,vkVk\vec{v}_{1}\in V_{1},\ldots,\vec{v}_{k}\in V_{k} such that i=1kvi=t\sum_{i=1}^{k}\vec{v}_{i}=\vec{t}.

It’s easy to see kk-VectorSum with |𝔽|=O(1)|\mathbb{F}|=O(1) and d=O(klogn)d=O(k\log n) is W[1]-hard, and even does not admit no(k)n^{o(k)} time algorithms assuming ETH. The idea is to use entries of vectors to check the consistency in kk-Clique or 3SAT.

Theorem 4.2.

Assuming W[1]\neqFPT, kk-VectorSum with 𝔽=𝔽2\mathbb{F}=\mathbb{F}_{2} and d=Θ(klogn)d=\Theta(k\log n) can not be solved in f(k)poly(n)f(k)\cdot\text{poly}(n) time.

Proof.

Set K=(k2)K=\binom{k}{2} groups of vectors, each representing valid edges between an ii-th block and a jj-th block of vertices in kk-Clique. For each i[k]i\in[k], we want to make sure that, the (k1)(k-1) vertices chosen from the ii-th block are all the same one. Thus we need to do k(k2)k\cdot(k-2) equality checks, each on two (logn)(\log n)-bit strings. In short, we exploit a new entry to do each bitwise equality check: set the entry to be the unchecked bit in the two vectors involved, and set the entry to be zero in all other vectors. Let the target vector to be 𝟎\vec{\mathbf{0}}. Thus all vectors sum up to zero in this entry if and only if the two to-be-checked bits are the same. The produced KK-VectorSum instance has parameter K=Θ(k2)K=\Theta(k^{2}) and dimension d=Θ(k2logn)d=\Theta(k^{2}\log n). ∎

Theorem 4.3.

Assuming ETH, kk-VectorSum with 𝔽=𝔽2\mathbb{F}=\mathbb{F}_{2} and d=Θ(klogn)d=\Theta(k\log n) can not be solved in no(k)n^{o(k)} time.

Proof.

Divide the clauses into kk equal-sized groups. Enumerate satisfying partial assignments of each group of clauses, then we want to check the consistency of those partial assignments. Note we can assume that each variable only appears in at most 3 clauses, thus we only need to do at most 2 pair-wise equality checks (between the first and the second appearances, and between the second and the third). The equality check step is the same as that in the proof of Theorem 4.2, and is omitted here.

The produced kk-VectorSum instance has size N=2Θ(n/k)N=2^{\Theta(n/k)} and the dimension is d=O(n)=O(klogN)d=O(n)=O(k\log N). Assuming ETH, this can not be solved in No(k)N^{o(k)} time. ∎

Note that the 𝔽2\mathbb{F}_{2} can be replaced by any finite field, as long as we slightly adjust the value of an entry: both 0 if x=0x=0; cc in the first appearance and c-c in the second appearance for some constant c0c\neq 0 if x=1x=1.

4.3 Gap Producing via Hadamard Codes

In this subsection we briefly introduce how Lin [Lin21] rules out constant FPT-approximations of kk-Clique under W[1]\neqFPT.

The most essential step in the reduction is to combine kk-VectorSum, whose W[1]-hardness was shown in Theorem 4.2, with Hadamard codes to create a gap. The technique is very similar to which people used in proving weakened PCP theorem, namely, in proving NP \subseteq PCP(poly(n),1)(\text{poly(n),1}) [ALM+98].

Here follows the definition of Hadamard Code, and its two important properties.

Definition 4.4 (Walsh-Hadamard Code).

For two strings x,y{0,1}nx,y\in\{0,1\}^{n}, define x,y=i=1nxiyi(mod2)\langle x,y\rangle=\sum_{i=1}^{n}x_{i}y_{i}\pmod{2}. The Walsh-Hadamard Code is the function f:{0,1}n{0,1}2nf:\{0,1\}^{n}\to\{0,1\}^{2^{n}} that maps every string x{0,1}nx\in\{0,1\}^{n} into the string z{0,1}2nz\in\{0,1\}^{2^{n}} such that zy=x,yz_{y}=\langle x,y\rangle for every y{0,1}ny\in\{0,1\}^{n}.

  1. 1.

    (Linearity Testing, [BLR93]) Let gg be a function mapping from {0,1}n\{0,1\}^{n} to {0,1}\{0,1\}. If it can pass (1δ)(1-\delta) fraction of tests g(x)+g(x)=g(x+x),x,x{0,1}ng(x)+g(x^{\prime})=g(x+x^{\prime}),x,x^{\prime}\in\{0,1\}^{n}, then gg is at least (1δ)(1-\delta)-close to a true linear function g~:{0,1}n{0,1}\tilde{g}:\{0,1\}^{n}\to\{0,1\}, which can also be parsed as a Hadamard codeword. By setting 0δ<140\leq\delta<\frac{1}{4}, this codeword is unique since the relative distance of Hadamard Code is exactly 12\frac{1}{2}.

  2. 2.

    (Locally Decodable Property) Suppose gg is (1δ)(1-\delta)-close to a Hadamard codeword f(x)f(x) for some x{0,1}nx\in\{0,1\}^{n}, then for any y{0,1}ny\in\{0,1\}^{n}, we can recover (f(x))y(f(x))_{y} probabilistically by querying only two positions of gg. Namely, sample y{0,1}ny^{\prime}\in\{0,1\}^{n} uniformly at random, and output g(y+y)+g(y)g(y+y^{\prime})+g(y^{\prime}). This succeeds with probability at least 12δ1-2\delta by a simple union bound.

Given an kk-VectorSum instance (V1,,Vk,t)(V_{1},\ldots,V_{k},\vec{t}), we build a CSP instance on variable set X={xa1,,ak:a1,,ak𝔽d}X=\{x_{\vec{a}_{1},\ldots,\vec{a}_{k}}:\vec{a}_{1},\ldots,\vec{a}_{k}\in\mathbb{F}^{d}\}. Let v1V1,,vkVk\vec{v}_{1}\in V_{1},\ldots,\vec{v}_{k}\in V_{k} be a solution that sums up to t\vec{t} in the yes case, each variable xa1,,akx_{\vec{a}_{1},\ldots,\vec{a}_{k}} is supposed to take the value i[k]ai,vi\sum_{i\in[k]}\langle\vec{a}_{i},\vec{v}_{i}\rangle. Here we are actually concatenating the kk solution vectors into one long vector v𝔽kd\vec{v}\in\mathbb{F}^{kd}, and the concatenation of variables is supposed to be the Hadamard codeword of v\vec{v}.

There are three types of tests we want to make:

  • (T1) a1,,ak,b1,bk𝔽d\forall\vec{a}_{1},\ldots,\vec{a}_{k},\vec{b}_{1},\ldots\vec{b}_{k}\in\mathbb{F}^{d}, test whether xa1,,ak+xb1,bk=xa1+b1,,ak+bkx_{\vec{a}_{1},\ldots,\vec{a}_{k}}+x_{\vec{b}_{1},\ldots\vec{b}_{k}}=x_{\vec{a}_{1}+\vec{b}_{1},\ldots,\vec{a}_{k}+\vec{b}_{k}}.

  • (T2) i[k],a1,,ak,a𝔽d\forall i\in[k],\forall\vec{a}_{1},\ldots,\vec{a}_{k},\vec{a}\in\mathbb{F}^{d}, test whether xa1,,ai+a,,akxa1,,ak=a,vx_{\vec{a}_{1},\ldots,\vec{a}_{i}+\vec{a},\ldots,\vec{a}_{k}}-x_{\vec{a}_{1},\ldots,\vec{a}_{k}}=\langle\vec{a},\vec{v}\rangle for some vViv\in V_{i}.

  • (T3) a1,,ak,a𝔽d\forall\vec{a}_{1},\ldots,\vec{a}_{k},\vec{a}\in\mathbb{F}^{d}, test whether xa1+a,,ak+axa1,,ak=a,tx_{\vec{a}_{1}+\vec{a},\ldots,\vec{a}_{k}+\vec{a}}-x_{\vec{a}_{1},\ldots,\vec{a}_{k}}=\langle\vec{a},\vec{t}\rangle.

By linearity testing, if an assignment to XX satisfies (1δ)(1-\delta)-fraction of constraints in (T1), then XX is at least (1δ)(1-\delta)-close to a true Hadamard codeword f(u)f(\vec{u}), u𝔽kd\vec{u}\in\mathbb{F}^{kd}. The constraints (T2) and (T3) use locally decodable properties to recover f(u)(0,,a,,0),f(u)(a,,a)f(\vec{u})_{(\vec{0},\ldots,\vec{a},\ldots,\vec{0})},f(\vec{u})_{(\vec{a},\ldots,\vec{a})}, and use them to check whether u\vec{u} indeed indicates a satisfying solution of our kk-VectorSum instance.

After that, we use a slightly modified FGLSS reduction to build an 2poly(k)d2^{\text{poly}(k)\cdot d}-clique instance. We build a group for each variable indicating its possible values, and a group for each (T1) test. A variable vertex and a test vertex are linked either if they are consistent, or the test is irrelevant to the variable. Two test vertices are linked in a similar way accordingly. Two variable vertices are linked either if the values specified by them pass the (T2) and (T3) tests, or there are no such tests between them. Therefore, if there is a clique whose size is at least (1ε)(1-\varepsilon) times the maximum, the following conditions hold:

  1. 1.

    A constant fraction of (T1) tests are passed.

  2. 2.

    For a constant fraction of variables, all (T2) and (T3) tests between them are passed.

This completes the reduction.

There are still some technical details. For example, in this reduction the number of vertices is 2poly(k)d=npoly(k)2^{\text{poly}(k)\cdot d}=n^{\text{poly}(k)}, which is too large. [Lin21] sampled some random matrices to handle this. Details are omitted here.

5 kk-SetCover

SetCover, which is equivalent to the Dominating Set problem, is one of the fundamental problems in computational complexity. A simple greedy algorithm yield a (lnn)(\ln n)-approximation of this problem. On the opposite side, it was shown that (1ε)lnn(1-\varepsilon)\ln n-approximation for this problem is NP-hard for every ε>0\varepsilon>0 [DS14]. Thus, its approximability in the NP regime has been completely settled.

In the parameterized regime, kk-SetCover is the complete problem of W[2], and does not admit no(k)n^{o(k)} algorithms assuming ETH [CHKX06], not even nkεn^{k-\varepsilon} algorithms assuming SETH [PW10]. Hardness of approximation of kk-SetCover in FPT time was studied by [CL19, CCK+17, KLM19, Lin19], and currently based on Gap-ETH, ETH, W[1]\neqFPT or kk-SUM Hypothesis, kk-SetCover is hard to approximate to within a (logn)1poly(k)(\log n)^{\frac{1}{\text{poly}(k)}} factor. In one direction, we wonder if this (logn)1poly(k)(\log n)^{\frac{1}{\text{poly}(k)}} can be further improved to logn\log n, or it is already tight. In the other direction, it is also worth questioning whether the total FPT inapproximability of kk-SetCover can be based on the weaker assumption W[2]\neqFPT.

Current state-of-the-art inapproximability results of kk-SetCover are reached by two contrasting methods, namely, Distributed PCP Framework [KLM19] and Threshold Graph Composition [Lin19], which we will introduce in Section 5.3 and 5.4, respectively. See the following table for an overview of results.

Complexity Assumption Inapproximability Ratio Time Lower Bound Reference
W[1]\neqFPT (logn)ε(k)=o(1)(\log n)^{\varepsilon(k)=o(1)} f(k)poly(n)f(k)\cdot\text{poly}(n) [Lin19]
(logn)1/poly(k)(\log n)^{1/\text{poly}(k)} [KLM19]
ETH (lognloglogn)1/k\left(\frac{\log n}{\log\log n}\right)^{1/k} f(k)nΩ(k)f(k)\cdot n^{\Omega(k)} [Lin19]
(logn)1/poly(k)(\log n)^{1/\text{poly}(k)} [KLM19]
SETH (lognloglogn)1/k\left(\frac{\log n}{\log\log n}\right)^{1/k} f(k)nkεf(k)\cdot n^{k-\varepsilon} [Lin19]
(logn)1/poly(k)(\log n)^{1/\text{poly}(k)} [KLM19]
kk-SUM Hypothesis (lognloglogn)1/k\left(\frac{\log n}{\log\log n}\right)^{1/k} f(k)nk/2εf(k)\cdot n^{\lceil k/2\rceil-\varepsilon} [Lin19]
(logn)1/poly(k)(\log n)^{1/\text{poly}(k)} [KLM19]

As for the coloring, the colored version and uncolored version of (exact) kk-SetCover are also equivalent, since we can add kk elements in the universe to ensure there is a set from each group, to reduce a colored version to an uncolored version. In the other direction, taking kk copies of the sets also works, because choosing replicated sets does not contribute. In the approximating sense, since it is a minimization problem, in the soundness case when solutions of size g(k)k\leq g(k)\cdot k are ruled out, it always means we can’t find such many sets to cover the universe even when choosing from the same set is allowed. Thus there is no need to specify the coloring.

5.1 Hypercube Partition System

We first introduce hypercube partition system, which is a powerful tool used in the reduction from MinLabel to kk-SetCover [CCK+17, KLM19], and in creating the gap instance of kk-SetCover [Lin19].

Definition 5.1 (Hypercube Partition System).

The (κ,ρ)(\kappa,\rho)-hypercube partition system consists of the universe \mathcal{M} and a collection of subset {Px,y}x[ρ],y[κ]\{P_{x,y}\}_{x\in[\rho],y\in[\kappa]} where =[κ]ρ\mathcal{M}=[\kappa]^{\rho} and Px,y={z:zx=y}P_{x,y}=\{z\in\mathcal{M}:z_{x}=y\}.

The universe \mathcal{M} consists of all functions from [ρ][\rho] to [κ][\kappa], and is of size κρ\kappa^{\rho}. Each subset Px,yP_{x,y} (x[ρ],y[κ]x\in[\rho],y\in[\kappa]) consists of all functions mapping xx to yy. It can be observed that one can cover the universe by picking all κ\kappa subsets from some row x[ρ]x\in[\rho], and this is the only way to cover the universe. In other words, even if we include κ1\kappa-1 subsets from every row, it is not possible to cover the universe.

5.2 Reduction from MinLabel

Now we show how to reduce MinLabel to SetCover. This reduction preserves gap, but significantly increases the instance size.

Given a MinLabel instance Γ\Gamma with \ell left super-nodes U1,,UU_{1},\ldots,U_{\ell} and hh right super-nodes W1,,WhW_{1},\ldots,W_{h}, we build a SetCover instance as follows. Take \ell different copies of (h,|ΣU|)(h,|\Sigma_{U}|)-hypercube partition system and set the universe to be the union of \ell universes. Each set in SetCover corresponds to a right vertex in MinLabel. For a set SvS_{v} associated to a right vertex vWjv\in W_{j} and for a left vertex uUiu\in U_{i}, if there is an edge (u,v)(u,v), we include Pu,jP_{u,j} in set SvS_{v}.

In order to see there is a one-to-one mapping from a solution of Γ\Gamma to a solution of the new SetCover instance, note that for a left vertex uUiu\in U_{i}, if a right multi-labeling covers uu, by picking corresponding sets we have Pu,jP_{u,j} for all j[h]j\in[h]. Moreover, the only way to cover the universe is to include all Pu,jP_{u,j} for some row indexed by uu, so a valid SetCover solution must contain sets associated to at least one neighbor in each right super-nodes for some specific left vertex uu. The same argument applies to each of the \ell left super-nodes, because we have a different copy of the hypercube partition system for each of them.

One important thing about this reduction is that the instance size is blowed up to h|ΣU|\ell\cdot h^{|\Sigma_{U}|}, where hh is the solution size of yes instance (and also the parameter of kk-SetCover). Thus, in order to make it f(k)nO(1)\leq f(k)n^{O(1)}, the left alphabet size can not exceed lognlogk\frac{\log n}{\log k}. Following the hardness of MinLabel (Theorem 3.7) and letting 1/γ=(logn)O(1)1/\gamma=(\log n)^{O(1)}, kk-SetCover is hard to approximate to a (logn)O(1k)(\log n)^{O(\frac{1}{k})} factor assuming Gap-ETH.

5.3 Gap Producing via Distributed PCP

In this section we introduce the Distributed PCP Framework. This framework was first proposed by Abbound et al. [ARW17], and later used by Karthik et al. to rule out FPT approximation algorithms for kk-SetCover [KLM19]. The interesting part of [KLM19] is to obtain hardness of MaxCover with specific parameters, while hardness of MinLabel and kk-SetCover directly follow from reductions in [CCK+17] (see Theorem 3.7 and Section 5.2 respectively).

At a high level, in this framework, one first rewrites the problem related to the hypothesis as a communication problem, then derives a Simultaneous Message Protocol for this problem and extracts an instance of MaxCover from the transcript of the protocol.

Due to space limitations, we only introduce their W[1] and ETH results to give an overview of their methods. The ideas in SETH and kk-SUM Hypothesis results are similar, except that they involve more complicated error correcting codes and protocols.

The similarity between W[1]\neqFPT and ETH is that they both often involve agreement tests (in other words, equality tests). Starting from W[1]\neqFPT, one may want to set K=(k2)K=\binom{k}{2} groups, each containing N=|E|N=|E| elements representing valid edges. The goal is to pick an edge from each group such that the label of end points (which is of logN\log N bits length) are consistent. W[1]\neqFPT states that this problem does not admit an f(K)NO(1)f(K)N^{O(1)} algorithm. Similarly, starting from ETH, one may also want to divide the clauses into KK groups, each containing at most N=2Θ(n/K)N=2^{\Theta(n/K)} partial satisfying assignments for those clauses. The goal is also to pick an assignment from each group such that the values on each variables (which is of n=Θ(KlogN)n=\Theta(K\log N) bits length in total) are consistent. ETH states that this cannot be done in No(K)N^{o(K)} time.

We can think of the agreement test problem as a communication problem: there are KK players, each receiving an element from the corresponding group. They want to collaborate nicely to decide whether their elements in hand “agree” or not. To achieve this, they use a specific communication protocol called Simultaneous Message Protocol, which was introduced by Yao [Yao79]:

Definition 5.2 (Simultaneous Message Protocol).

We say π\pi is a (r,,s)(r,\ell,s)-efficient protocol if the following holds:

  • The protocol is one-round with public randomness. The following actions happen sequentially:

    1. 1.

      The players receive their inputs.

    2. 2.

      The players and the referee jointly toss rr random coins.

    3. 3.

      Each player on seeing the randomness deterministically sends an \ell-bit message to the referee.

    4. 4.

      Based on the randomness and the KK\cdot\ell bits sent from the players, the referee outputs accept or reject.

  • The protocol has completeness 1 and soundness ss, i.e.,

    • If their inputs indeed agree, then the referee always accepts regardless of randomness.

    • Otherwise, the referee accepts with probability at most ss.

The full version of SMP protocol also admits ww bits of advice. In the contexts of W[1]\neqFPT and ETH here, we do not need this.

Note that by repeating the protocol zz times, each time using fresh randomness, we can derive a (zr,z,sz)(z\cdot r,z\cdot\ell,s^{z})-efficient protocol from an (r,,s)(r,\ell,s)-efficient protocol.

Next we will see how to use MaxCover to simulate an SMP protocol. Then the hardness of the starting problem (kk-Clique or 3SAT) and the existence of an efficient protocol for it will lead to hardness of MaxCover.

Theorem 5.3 (Theorem 5.2 in [KLM19], slightly simplified).

An instance Π\Pi of a (kk-Clique or 3SAT) problem which admits an (r,,s)(r,\ell,s)-efficient KK-player SMP protocol can be reduced to a MaxCover instance Γ\Gamma as follows:

  • The reduction runs in time 2r+Kpoly(N,K)2^{r+K\cdot\ell}\cdot\text{poly}(N,K).

  • Γ\Gamma has KK right super-nodes of size at most NN each.

  • Γ\Gamma has 2r2^{r} left super-nodes of size at most 2K2^{K\cdot\ell} each.

  • If Π\Pi is a YES instance, then MaxCover=1=1.

  • If Π\Pi is a NO instasnce, then MaxCovers\leq s.

Proof.

Here is the construction:

  • Each right super-node corresponds to the group which a player receives an input from.

  • Each left super-node corresponds to a random string γ{0,1}r\gamma\in\{0,1\}^{r}. The left super-node UγU_{\gamma} contains one node for each possible accepting messages from the KK players, i.e., each vertex in UγU_{\gamma} corresponds to (m1,,mK)({0,1})K(m_{1},\ldots,m_{K})\in(\{0,1\}^{\ell})^{K} where in the protocol the referee accepts on seeing randomness γ\gamma and messages (m1,,mK)(m_{1},\ldots,m_{K}).

  • We add an edge between a right vertex xWjx\in W_{j} and a left vertex (m1,,mK)Uγ(m_{1},\ldots,m_{K})\in U_{\gamma} if mjm_{j} is equal to the message that player jj sends on an input xx and randomness γ\gamma in the protocol.

Detailed proofs of the desired properties are omitted here since they are rather straightforward. ∎

In the last step, we need to derive an efficient SMP protocol for kk-Clique and 3SAT. Directly sending their inputs (which is of logN\geq\log N bits length) does not seem to be a good idea because it would make the size of MaxCover to be 2r+K=2Ω(K)logN=NΩ(K)2^{r+K\cdot\ell}=2^{\Omega(K)\log N}=N^{\Omega(K)}, which is too large. What’s more, since the further reduction from MaxCover to MinLabel then to kk-SetCover will introduce a K|ΣU|K^{|\Sigma_{U}|} blow-up, we even need |ΣU|=2KlogNlogK|\Sigma_{U}|=2^{K\cdot\ell}\leq\frac{\log N}{\log K}.

Fortunately, this can be done via a simple error correcting code called good code which has constant rate and constant relative distance. To check the consistency of their inputs, we only need to check the equality of one random bit of their codes. The randomness is used to specify the index of that bit. This way we can have r=O(loglogN),=O(1),s=O(1)r=O(\log\log N),\ell=O(1),s=O(1). Within the bound that 1K(loglogNloglogK)\ell\leq\frac{1}{K}(\log\log N-\log\log K), we can repeat the protocol z=O(1KloglogN)z=O(\frac{1}{K}\log\log N) times, leading to a soundness parameter s=O(1)z=O((logN)1K)s=O(1)^{z}=O((\log N)^{\frac{1}{K}}). After the reduction to MinLabel, the ε\varepsilon in gap would become ε1K\varepsilon^{\frac{1}{K}}.

Along such a long way we finally reach a (logn)1poly(k)(\log n)^{\frac{1}{\text{poly}(k)}} inapproximability ratio for kk-SetCover with FPT time lower bound under W[1]\neqFPT and no(k)n^{o(k)} time lower bound under ETH.

5.4 Gap Producing via Threshold Graph Composition

In this subsection we introduce how to obtain (lognloglogn)1/k\left(\frac{\log n}{\log\log n}\right)^{1/k} inapproximability of kk-SetCover via threshold graph composition technique.

In general, this technique transforms an kk-SetCover instance with small universe (typically |U|=Θ(logn)|U|=\Theta(\log n)) still to an instance of kk-SetCover, increasing the size of universe to O(n)O(n) while creating a gap. In the YES case, the number of sets needed to cover the universe is still kk, but in the NO case it becomes hkh\gg k, where hh is determined by the threshold graph.

Given an kk-SetCover instance Γ=(𝒮,U)\Gamma=(\mathcal{S},U), we need a bipartite threshold graph G=(A˙B,E)G=(A\dot{\cup}B,E) with the following properties:

  1. 1.

    AA is not divided. BB is divided into \ell groups: B=B1˙˙BB=B_{1}\dot{\cup}\ldots\dot{\cup}B_{\ell}, where \ell is arbitrary.

  2. 2.

    |A|=n,|Bi|lognlog|U|,i[]|A|=n,|B_{i}|\leq\frac{\log n}{\log|U|},\forall i\in[\ell].

  3. 3.

    For any kk vertices a1,,akAa_{1},\ldots,a_{k}\in A and for any i[]i\in[\ell], there is at least one common neighbor of {a1,,ak}\{a_{1},\ldots,a_{k}\} in BiB_{i}.

  4. 4.

    For any XAX\subseteq A, if for any i[]i\in[\ell], there is at least one vertex in BiB_{i}, which is a common neighbor of at least k+1k+1 vertices in XX, then |X|h|X|\geq h.

We compose the original kk-SetCover instance Γ=(𝒮,U)\Gamma=(\mathcal{S},U) with this threshold graph GG to produce an instance Γ=(𝒮,U)\Gamma^{\prime}=(\mathcal{S}^{\prime},U^{\prime}) as follows.

  • |𝒮|=|𝒮||\mathcal{S}^{\prime}|=|\mathcal{S}|. For each S𝒮S\in\mathcal{S} we associate a new set S𝒮S^{\prime}\in\mathcal{S}^{\prime}. We also associate a vertex in AA (the left side of the threshold graph) to each set S𝒮S^{\prime}\in\mathcal{S}^{\prime}.

  • Γ\Gamma consists of \ell hypercube partition systems, one for each BiB_{i} (the right side of the threshold graph). The ii-th partition system has |Bi||B_{i}| rows and |U||U| columns, thus is of size |U||Bi||U|^{|B_{i}|}.

  • For any i[],xBi,yUi\in[\ell],x\in B_{i},y\in U, subset Px,yP_{x,y} is included in a set S𝒮S^{\prime}\in\mathcal{S}^{\prime} if and only if:

    1. 1.

      ySy\in S, i.e., set SS can cover yy in the original instance Γ\Gamma.

    2. 2.

      There is an edge between the vertex associated to SS^{\prime} and vertex xx in the threshold graph GG.

As shown above, each row in the partition system corresponds to a vertex in BiB_{i}, and each column corresponds to an element in UU. According to the properties of partition system, in order to cover the universe, we must pick all |U||U| subsets in some specific row. This, together with our construction, means there is a vertex xBix\in B_{i} such that sets correspond to its neighbors in GG can cover UU. Since the \ell hypercube partition systems are independent, this holds for each Bi,i[]B_{i},i\in[\ell].

In the YES case, kk sets are enough to cover UU, and there is at least one common neighbor of them in every Bi,i[]B_{i},i\in[\ell]. Thus the answer to the new instance is still kk.

In the NO case, at least k+1k+1 sets are need to cover UU. Consider any solution X𝒮X\subseteq\mathcal{S}^{\prime} of Γ\Gamma^{\prime}, we know that in each group BiB_{i}, there is a vertex xBix\in B_{i} such that |𝒩(x)X|k+1|\mathcal{N}(x)\cap X|\geq k+1 (because they form a solution in Γ\Gamma). Therefore, by property 4 of threshold graph, |X|h|X|\geq h as desired.

The last step is to construct a threshold graph the desired properties. [Lin19] used a specific combinatorial object called universal sets to construct it. However, the graph in Section 3.2 with proper parameters also suffices, and is simpler.

The required property is closely related to the collision property in the construction of Section 3.2. In fact, the gap hh here is just the collision number Col(C)Col(C) of the error correcting code. Thus, we want codes with large collision numbers.

Let the alphabet of the code be Σ\Sigma, then it’s easy to see the collision number cannot be larger than |Σ|+1|\Sigma|+1, since such many strings must collide in every position by pigeon principle. However, this upper bound can be reached by codes constructed from perfect hash family.

Definition 5.4 (Perfect Hash Family).

For every N,N,\ell\in\mathbb{N} and Σ\Sigma, we say that H:={hi:[N]Σ|i[]}H:=\{h_{i}:[N]\to\Sigma|i\in[\ell]\} is a [N,]|Σ|[N,\ell]_{|\Sigma|}-perfect hash family if for every subset T[N]T\subseteq[N] of size |Σ||\Sigma|, there exists an i[]i\in[\ell] such that

x,yT,xy,hi(x)hi(y)\forall x,y\in T,x\neq y,h_{i}(x)\neq h_{i}(y)

Think of an [N,]|Σ|[N,\ell]_{|\Sigma|}-perfect hash family as an N×N\times\ell matrix, where each column represents a hash function. Then the property is that for any |Σ|\leq|\Sigma| rows, there is a column with no collisions (on these rows). Regard each row as a codeword of an error correcting code from Σlog|Σ|NΣ\Sigma^{\log_{|\Sigma|}N}\to\Sigma^{\ell}, then the collision number of this error correcting code is >|Σ|>|\Sigma|, thus exactly |Σ|+1|\Sigma|+1 as desired.

Lemma 5.5 (Alon et al. [AYZ95] ).

For every N,|Σ|N,|\Sigma|\in\mathbb{N} there is a [N,2O(|Σ|)logN]|Σ|[N,2^{O(|\Sigma|)}\cdot\log N]_{|\Sigma|}-perfect hash family that can be computed in O~|Σ|(N)\tilde{O}_{|\Sigma|}(N) time.

In this threshold graph the size of a right super-node BiB_{i} is |Σ|k|\Sigma|^{k}. Assuming |U|=O(logn)|U|=O(\log n), to make |Bi|lognlog|U|=lognloglogn|B_{i}|\leq\frac{\log n}{\log|U|}=\frac{\log n}{\log\log n}, we need |Σ|(lognloglogn)1/k|\Sigma|\leq\left(\frac{\log n}{\log\log n}\right)^{1/k} and this is the best gap possible. By the above lemma, the perfect hash family (and the corresponding code) can be constructed efficiently.

Our last step is to show that, assuming different hypothesis, kk-SetCover with universe size logn\leq\log n is still hard. The reduction from 3SAT is straightforward: divide the variables into kk groups of size n/kn/k each, a group of sets corresponds to N=2n/kN=2^{n/k} possible assignments to those n/kn/k variables, and the universe is just the m=Θ(klogN)m=\Theta(k\log N) clauses. ETH (respectively, SETH) asserts that this instance cannot be solved in No(k)N^{o(k)} (respectively, NkεN^{k-\varepsilon}) time. Thus by the above threshold graph composition technique, we reach (lognloglogn)1/k\left(\frac{\log n}{\log\log n}\right)^{1/k}-inapproximability of kk-SetCover based on ETH and SETH.

The reduction from kk-Clique to kk-SetCover is a bit more complicated, as stated in the following theorem. The main idea is from Karthik et al. [KLM19].

Theorem 5.6.

There is an poly(n)\text{poly}(n) time algorithm, which can reduce an kk-Clique instance to an (k2)\binom{k}{2}-SetCover instance with universe size poly(k)logn\text{poly}(k)\log n.

Proof.

Sets in a group (i,j)(i,j) still represent edges between the ii-th block and the jj-th block (in the kk-Clique instance). In order to check the consistency of labels, we need k×lognk\times\log n hypercube partition systems, one for each (i,),i[k],[logn](i,\ell),i\in[k],\ell\in[\log n]. The (i,)(i,\ell)-th partition system is meant to check whether the \ell-th bits of labels of vertices in block ii are all the same. In an invalid solution, one may pick an edge between the ii-th and jj-th blocks, and an edge between the ii-th and jj^{\prime}-th blocks, such that the two vertices (let them be v1v_{1} and v2v_{2}) in the ii-th block are not the same. In such a case, v1v_{1} and v2v_{2} must differ in at least one bit, and thus cannot fulfill the requirements in the (i,)(i,\ell)-th partition system where \ell is the position of that bit.

Specifically, for all i[k],[logn]i\in[k],\ell\in[\log n], the (i,)(i,\ell)-th partition system contains 2 rows and (k1)(k-1) columns. The choices of rows represent the choices of the bit to be 0 or 1, and the columns test agreement of the (k1)(k-1) labels (edges between the ii-th block and each remaining blocks). For an edge between vViv\in V_{i} and wVjw\in V_{j}, we include Pv[],jP_{v[\ell],j} into its set. At last, the (k2)\binom{k}{2}-SetCover instance is the union of those klognk\cdot\log n hypercube partition systems.

The instance size is poly(n)\text{poly}(n) since there are that many edges, while the universe size is klogn(k1)2=poly(k)lognk\cdot\log n\cdot(k-1)^{2}=\text{poly}(k)\cdot\log n. ∎

The W[1]-hardness of kk-SetCover then follows. It is worth noting that in such FPT reductions, the parameter kk^{\prime} can be arbitrarily amplified as long as it is a function of kk. Assuming W[1]\neqFPT, kk-SetCover is hard to approximate to within a factor of (lognloglogn)1/(k2)\left(\frac{\log n}{\log\log n}\right)^{1/\binom{k}{2}}, then for any function ε(k)\varepsilon(k) which is o(1)o(1) as kk goes to infinity, take large enough kk^{\prime} such that ε(k)<1/(k2)\varepsilon(k^{\prime})<1/\binom{k}{2}, then for large enough nn we have (logn)ε(k)<(lognloglogn)1/(k2)(\log n)^{\varepsilon(k^{\prime})}<\left(\frac{\log n}{\log\log n}\right)^{1/\binom{k}{2}}, which means kk^{\prime}-SetCover (by padding the parameter from kk to kk^{\prime}) cannot be approximated to a factor of (logn)ε(k)(\log n)^{\varepsilon(k^{\prime})} in FPT time.

Instead of introducing their kk-SUM Hypothesis result here, we want to make some comments on this technique. Note that the maximum size of a right super-node in the threshold graph is lognlog|U|\frac{\log n}{\log|U|}. Thus when |U||U| is not as small as logn\log n, it may still be possible to obtain some inapproximability results. It remains a big open question that whether we can base total FPT inapproximability of kk-SetCover on W[2]\neqFPT. If we can construct threshold graphs with a gap such that each right super-nodes consist of only O(1)O(1) vertices, we can obtain W[2] hardness of kk-SetCover, respectively. Note that the construction in Section 3.2 does not suffice, because the size of their right super-nodes is Σk\Sigma^{k}, which is too large even if |Σ|=O(1)|\Sigma|=O(1).

Acknowledgements

I want to express my deep gratitude to Prof. Bingkai Lin, who brought me into the beautiful world of hardness of approximation, discussed with me regularly and guided me patiently. I would also like to thank my talented friends Yican Sun and Xiuhan Wang for their bright ideas and unreserved help. I really enjoy working with them.

References

  • [ALM+98] Sanjeev Arora, Carsten Lund, Rajeev Motwani, Madhu Sudan, and Mario Szegedy. Proof verification and the hardness of approximation problems. J. ACM, 45(3):501–555, may 1998.
  • [ARW17] Amir Abboud, Aviad Rubinstein, and R. Ryan Williams. Distributed PCP theorems for hardness of approximation in P. In Chris Umans, editor, 58th IEEE Annual Symposium on Foundations of Computer Science, FOCS 2017, Berkeley, CA, USA, October 15-17, 2017, pages 25–36. IEEE Computer Society, 2017.
  • [AYZ95] Noga Alon, Raphael Yuster, and Uri Zwick. Color-coding. J. ACM, 42(4):844–856, 1995.
  • [BBE+19] Arnab Bhattacharyya, Édouard Bonnet, László Egri, Suprovat Ghoshal, Karthik C. S., Bingkai Lin, Pasin Manurangsi, and Dániel Marx. Parameterized intractability of even set and shortest vector problem. Electron. Colloquium Comput. Complex., 26:115, 2019.
  • [BGKM18] Arnab Bhattacharyya, Suprovat Ghoshal, C. S. Karthik, and Pasin Manurangsi. Parameterized intractability of even set and shortest vector problem from gap-eth. Electron. Colloquium Comput. Complex., 25:57, 2018.
  • [BGLR93] Mihir Bellare, Shafi Goldwasser, Carsten Lund, and Alexander Russell. Efficient probabilistically checkable proofs and applications to approximations. In Proceedings of the twenty-fifth annual ACM symposium on Theory of computing, pages 294–304, 1993.
  • [BLR93] Manuel Blum, Michael Luby, and Ronitt Rubinfeld. Self-testing/correcting with applications to numerical problems. Journal of Computer and System Sciences, 47(3):549–595, 1993.
  • [BS94] Mihir Bellare and Madhu Sudan. Improved non-approximability results. In Proceedings of the twenty-sixth annual ACM symposium on Theory of computing, pages 184–193, 1994.
  • [CCK+17] Parinya Chalermsook, Marek Cygan, Guy Kortsarz, Bundit Laekhanukit, Pasin Manurangsi, Danupon Nanongkai, and Luca Trevisan. From gap-eth to fpt-inapproximability: Clique, dominating set, and more. In Chris Umans, editor, 58th IEEE Annual Symposium on Foundations of Computer Science, FOCS 2017, Berkeley, CA, USA, October 15-17, 2017, pages 743–754. IEEE Computer Society, 2017.
  • [CHKX06] Jianer Chen, Xiuzhen Huang, Iyad A. Kanj, and Ge Xia. Strong computational lower bounds via parameterized complexity. J. Comput. Syst. Sci., 72(8):1346–1367, 2006.
  • [CL19] Yijia Chen and Bingkai Lin. The constant inapproximability of the parameterized dominating set problem. SIAM J. Comput., 48(2):513–533, 2019.
  • [CW89] Aviad Cohen and Avi Wigderson. Dispersers, deterministic amplification, and weak random sources. In 30th Annual Symposium on Foundations of Computer Science, pages 14–19. IEEE Computer Society, 1989.
  • [Din07] Irit Dinur. The PCP theorem by gap amplification. J. ACM, 54(3):12, 2007.
  • [Din16] Irit Dinur. Mildly exponential reduction from gap 3sat to polynomial-gap label-cover. Electron. Colloquium Comput. Complex., 23:128, 2016.
  • [DS14] Irit Dinur and David Steurer. Analytical approach to parallel repetition. In David B. Shmoys, editor, Symposium on Theory of Computing, STOC 2014, New York, NY, USA, May 31 - June 03, 2014, pages 624–633. ACM, 2014.
  • [FGL+96] Uriel Feige, Shafi Goldwasser, Laszlo Lovász, Shmuel Safra, and Mario Szegedy. Interactive proofs and the hardness of approximating cliques. Journal of the ACM (JACM), 43(2):268–292, 1996.
  • [FK00] Uriel Feige and Joe Kilian. Two-prover protocols—low error at affordable rates. SIAM Journal on Computing, 30(1):324–346, 2000.
  • [Gol98] Shafi Goldwasser. Introduction to special section on probabilistic proof systems. SIAM Journal on Computing, 27(3):737, 1998.
  • [Has96] Johan Hastad. Clique is hard to approximate within n1ϵn^{1-\epsilon}. In Proceedings of 37th Conference on Foundations of Computer Science, pages 627–636. IEEE, 1996.
  • [IP01] Russell Impagliazzo and Ramamohan Paturi. On the complexity of k-sat. Journal of Computer and System Sciences, 62:367–375, 2001.
  • [IPZ01] Russell Impagliazzo, Ramamohan Paturi, and Francis Zane. Which problems have strongly exponential complexity? Journal of Computer and System Sciences, 63(4):512–530, 2001.
  • [KLM19] C. S. Karthik, Bundit Laekhanukit, and Pasin Manurangsi. On the parameterized complexity of approximating dominating set. J. ACM, 66(5):33:1–33:38, 2019.
  • [KN21] C. S. Karthik and Inbal Livni Navon. On hardness of approximation of parameterized set cover and label cover: Threshold graphs from error correcting codes. In Hung Viet Le and Valerie King, editors, 4th Symposium on Simplicity in Algorithms, SOSA 2021, Virtual Conference, January 11-12, 2021, pages 210–223. SIAM, 2021.
  • [Lin18] Bingkai Lin. The parameterized complexity of the k-biclique problem. J. ACM, 65(5):34:1–34:23, 2018.
  • [Lin19] Bingkai Lin. A simple gap-producing reduction for the parameterized set cover problem. In Christel Baier, Ioannis Chatzigiannakis, Paola Flocchini, and Stefano Leonardi, editors, 46th International Colloquium on Automata, Languages, and Programming, ICALP 2019, July 9-12, 2019, Patras, Greece, volume 132 of LIPIcs, pages 81:1–81:15. Schloss Dagstuhl - Leibniz-Zentrum für Informatik, 2019.
  • [Lin21] Bingkai Lin. Constant approximating k-clique is w[1]-hard. In Samir Khuller and Virginia Vassilevska Williams, editors, STOC ’21: 53rd Annual ACM SIGACT Symposium on Theory of Computing, Virtual Event, Italy, June 21-25, 2021, pages 1749–1756. ACM, 2021.
  • [LRSW21] Bingkai Lin, Xuandi Ren, Yican Sun, and Xiuhan Wang. On lower bounds of approximating parameterized kk-clique, 2021.
  • [LRSZ20] Daniel Lokshtanov, M. S. Ramanujan, Saket Saurabh, and Meirav Zehavi. Parameterized complexity and approximability of directed odd cycle transversal. In Shuchi Chawla, editor, Proceedings of the 2020 ACM-SIAM Symposium on Discrete Algorithms, SODA 2020, Salt Lake City, UT, USA, January 5-8, 2020, pages 2181–2200. SIAM, 2020.
  • [MR17] Pasin Manurangsi and Prasad Raghavendra. A Birthday Repetition Theorem and Complexity of Approximating Dense CSPs. In Ioannis Chatzigiannakis, Piotr Indyk, Fabian Kuhn, and Anca Muscholl, editors, 44th International Colloquium on Automata, Languages, and Programming (ICALP 2017), volume 80 of Leibniz International Proceedings in Informatics (LIPIcs), pages 78:1–78:15, Dagstuhl, Germany, 2017. Schloss Dagstuhl–Leibniz-Zentrum fuer Informatik.
  • [PW10] Mihai Patrascu and Ryan Williams. On the possibility of faster SAT algorithms. In Moses Charikar, editor, Proceedings of the Twenty-First Annual ACM-SIAM Symposium on Discrete Algorithms, SODA 2010, Austin, Texas, USA, January 17-19, 2010, pages 1065–1075. SIAM, 2010.
  • [Tov84] C. Tovey. A simplified np-complete satisfiability problem. Discret. Appl. Math., 8:85–89, 1984.
  • [Yao79] Andrew Chi-Chih Yao. Some complexity questions related to distributive computing (preliminary report). In Michael J. Fischer, Richard A. DeMillo, Nancy A. Lynch, Walter A. Burkhard, and Alfred V. Aho, editors, Proceedings of the 11h Annual ACM Symposium on Theory of Computing, April 30 - May 2, 1979, Atlanta, Georgia, USA, pages 209–213. ACM, 1979.
  • [Zuc96a] David Zuckerman. On unapproximable versions of np-complete problems. SIAM J. Comput., 25(6):1293–1304, 1996.
  • [Zuc96b] David Zuckerman. Simulating BPP using a general weak random source. Algorithmica, 16(4/5):367–391, 1996.
  • [Zuc07] David Zuckerman. Linear degree extractors and the inapproximability of max clique and chromatic number. Theory Comput., 3(1):103–128, 2007.