This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

An EPTAS for Budgeted Matroid Independent Set

Ilan Doron-Arad Computer Science Department, Technion, Haifa, Israel. [email protected]    Ariel Kulik CISPA Helmholtz Center for Information Security, Saarland Informatics Campus, Germany. [email protected]    Hadas Shachnai Computer Science Department, Technion, Haifa, Israel. [email protected]
Abstract

We consider the budgeted matroid independent set problem. The input is a ground set, where each element has a cost and a non-negative profit, along with a matroid over the elements and a budget. The goal is to select a subset of elements which maximizes the total profit subject to the matroid and budget constraints. Several well known special cases, where we have, e.g., a uniform matroid and a budget, or no matroid constraint (i.e., the classic knapsack problem), admit a fully polynomial-time approximation scheme (FPTAS). In contrast, already a slight generalization to the multi-budgeted matroid independent set problem has a PTAS but does not admit an efficient polynomial-time approximation scheme (EPTAS). This implies a PTAS for our problem, which is the best known result prior to this work.

Our main contribution is an EPTAS for the budgeted matroid independent set problem. A key idea of the scheme is to find a representative set for the instance, whose cardinality depends solely on 1/ε1/{\varepsilon}, where ε>0{\varepsilon}>0 is the accuracy parameter of the scheme. The representative set is identified via matroid basis minimization, which can be solved by a simple greedy algorithm. Our scheme enumerates over subsets of the representative set and extends each subset using a linear program. The notion of representative sets may be useful in solving other variants of the budgeted matroid independent set problem.

1 Introduction

We consider the budgeted matroid independent set (BMI) problem, defined as follows. We are given a set of elements EE, a membership oracle for a collection of independent sets 2E{\mathcal{I}}\subseteq 2^{E}, where (E,)(E,{\mathcal{I}}) is a matroid, a budget β>0\beta>0, a cost function c:E[0,β]c:E\to[0,\beta], and a profit function p:E+p:E\rightarrow\mathbb{R}_{+}. A solution for the problem is an independent set XX\in{\mathcal{I}} of total cost at most β\beta (i.e., c(X)=eXc(e)βc(X)=\sum_{e\in X}c(e)\leq\beta). The profit of a solution XX is p(X)=eXp(e)p(X)=\sum_{e\in X}p(e), and the objective is to find a solution of maximal profit.

BMI is a generalization of the classic 0/10/1-knapsack problem, which is NP-hard and therefore unlikely to admit an exact polynomial-time algorithm. Thus, there is a long line of work on finding efficient approximations for knapsack and its variants (for comprehensive surveys see, e.g., [20, 17]).

Let OPT(I)\textnormal{OPT}(I) be the value of an optimal solution for an instance II of a maximization problem Π\Pi. For α(0,1]\alpha\in(0,1], we say that 𝒜{\mathcal{A}} is an α\alpha-approximation algorithm for Π\Pi if, for any instance II of Π\Pi, 𝒜{\mathcal{A}} outputs a solution of value at least αOPT(I)\alpha\textnormal{OPT}(I). A polynomial-time approximation scheme (PTAS) for a maximization problem Π\Pi is a family of algorithms (𝒜ε)ε>0({\mathcal{A}}_{{\varepsilon}})_{{\varepsilon}>0} such that, for any ε>0{\varepsilon}>0, 𝒜ε{\mathcal{A}}_{{\varepsilon}} is a polynomial-time (1ε)(1-{\varepsilon})-approximation algorithm for Π\Pi; (𝒜ε)ε>0({\mathcal{A}}_{{\varepsilon}})_{{\varepsilon}>0} is an EPTAS if the running time of 𝒜ε{\mathcal{A}}_{{\varepsilon}} is of the form f(1ε)nO(1)f\left(\frac{1}{{\varepsilon}}\right)\cdot n^{O(1)}, where ff is an arbitrary function, and nn is the bit-length encoding size of the input instance; (𝒜ε)ε>0({\mathcal{A}}_{{\varepsilon}})_{{\varepsilon}>0} is an FPTAS if the running time of 𝒜ε{\mathcal{A}}_{{\varepsilon}} is of the form (nε)O(1)\left(\frac{n}{{\varepsilon}}\right)^{O(1)}.

Polynomial-time approximation schemes allow us to obtain almost optimal solutions for NP-hard optimization problems via a speed-accuracy trade-off. However, the strong dependency of run-times on the error parameter, ε>0{\varepsilon}>0, often renders these schemes highly impractical. Thus, a natural goal is to seek the fastest scheme for a given problem, assuming one exists. Having obtained a PTAS, the next step is to consider possible improvements to an EPTAS, or even to an FPTAS. This is the focus of our work.

For the classic knapsack problem, a very efficient FPTAS exists since the 1970’s. Lawler’s scheme [19], based on ideas from [12], achieves a running time of O(nlog1/ε+1/ε4)O(n\log 1/{\varepsilon}+1/{\varepsilon}^{4}) for a (1ε)(1-{\varepsilon})-approximation. In contrast, already the two-dimensional knapsack problem has a PTAS but does not admit an EPTAS [18]. For the well known multiple knapsack problem, Chekuri and Khanna [3] derived the first PTAS, which was later improved by Jansen to an EPTAS [13, 14]. The existence of an FPTAS is ruled out by a simple reduction from Partition [3]. More generally, resolving the complexity status of NP-hard optimization problem with respect to approximation schemes has been the focus of much research relating to resource allocation and scheduling (see, e.g., [24, 6, 8, 15, 16] and the surveys  [22, 23]).

For BMI, FPTASs are known for instances in which the matroid belongs to a restricted family. One notable example is multiple choice knapsack [19], where the elements are partitioned into groups, and a set is independent if it contains at most one element from each group. Another example is 1.51.5-dimensional knapsack [2], in which a set is independent if it contains at most kk elements for some k1k\geq 1 (i.e., a uniform matroid constraint).

There are known results also for generalizations of BMI which involve multiple budget constraints or an additional matroid constraint. Specifically, Grandoni and Zenklusen developed in [9] a PTAS for multi-budgeted matroid independent set (MBMI), a generalization of BMI in which the costs are dd-dimensional (for some constant dd\in\mathbb{N}). The PTAS of [9] is based on integrality properties of a linear programming relaxation of the problem. As MBMI generalizes the two-dimensional knapsack problem, it does not admit an EPTAS unless W[1]=FPT\textnormal{W}[1]=\textnormal{FPT} [18].

The budgeted matroid intersection problem is a generalization of BMI in which the input includes membership oracles for independent sets of two different matroids, and the solution must be an independent set of both matroids. A PTAS for budgeted matroid intersection was developed in [1]. The algorithm of [1] uses a Lagrangian relaxation along with some combinatorial properties of the problem to patch two solutions (i.e., a feasible solution with sub-optimal profit, and a non-feasible solution with high profit) into the final solution. The existence of an EPTAS (or an FPTAS) for budgeted matroid intersection is still open.

The multi-budgeted matroid intersection problem is a generalization of both multi-budgeted matroid independent set and budgeted matroid intersection, in which the cost function is dd-dimensional, and the input contains two matroids. In [4] the authors developed a PTAS for multi-budgeted matroid intersection, based on randomized rounding of a solution for a linear programming (LP) relaxation of the problem.

The budgeted matroid independent set problem is also a special case of multiple knapsack with a matroid, a variant of BMI in which the input contains mm budgets β1,,βm\beta_{1},\ldots,\beta_{m}. A solution consists of mm sets S1,,SmS_{1},\ldots,S_{m}, where the cost of the jjth set is at most the jjth budget (c(Sj)βjc(S_{j})\leq\beta_{j}), and the union of the sets is an independent set of the matroid. A PTAS for the problem (based on randomized rounding) was given in [7]. The existence of an FPTAS is ruled out, as multiple knapsack is a special case [3].

To the best of our knowledge, BMI is studied here for the first time. A PTAS for the problem follows from known results for any of the above generalizations. In all cases, the running time of the scheme is dominated by an enumeration phase which guesses the most profitable elements in an optimal solution.

Our main result is an EPTAS for BMI, thus substantially improving the running times of existing schemes for the problem. Let 𝒦=(E,,c,p,β){\mathcal{K}}=(E,{\mathcal{I}},c,p,\beta) denote a BMI instance, OPT(𝒦)\textnormal{OPT}({\mathcal{K}}) the profit of an optimal solution for 𝒦{\mathcal{K}}, and |𝒦||{\mathcal{K}}| the bit-length encoding of the instance 𝒦{\mathcal{K}}.

Theorem 1.1.

Given an instance 𝒦{\mathcal{K}} of BMI and 0<ε<120<{\varepsilon}<\frac{1}{2}, there is an algorithm that outputs a solution of profit at least (1ε)OPT(𝒦)(1-{\varepsilon})\cdot\textnormal{OPT}({\mathcal{K}}) in time 2O(ε2log1ε)poly(|𝒦)2^{O\left({\varepsilon}^{-2}\cdot\log\frac{1}{{\varepsilon}}\right)}\cdot\textnormal{poly}(|{\mathcal{K}}).

Main Technique. Our algorithm builds upon a framework of Grandoni and Zenklusen [9] for multi-budgeted matroid independent set. In [9] the authors show that a basic solution for a linear programming relaxation of the problem has only a few non-integral entries. Thus, a solution for MBMI is constructed by solving the LP relaxation and adding all the elements with non-zero integral entries to the MBMI solution. An exhaustive enumeration phase which guesses the Θ(1ε)\Theta\left(\frac{1}{{\varepsilon}}\right) most profitable elements in an optimal solution is used to mitigate the profit loss caused by discarding the (few) elements with non-integral entries in the solution for the LP. The running time of the algorithm is dominated by the |E|Θ(1ε)|E|^{\Theta(\frac{1}{{\varepsilon}})} operations required for exhaustive enumeration.

The improved running time of our algorithm is obtained by reducing the time complexity of the enumeration phase. Consider a BMI instance 𝒦=(E,,c,p,β){\mathcal{K}}=(E,{\mathcal{I}},c,p,\beta). Given some 0<ε<120<{\varepsilon}<\frac{1}{2}, we say that an element eEe\in E is profitable if p(e)>εOPT(𝒦)p(e)>{\varepsilon}\cdot\textnormal{OPT}({\mathcal{K}}). Our algorithm identifies a representative set of elements RER\subseteq E satisfying |R|f(1ε)|R|\leq f\left(\frac{1}{{\varepsilon}}\right), for a computable function ff. Furthermore, we show that the BMI instance has a solution SS where all elements in SRS\setminus R are non-profitable and p(S)(1O(ε))OPT(𝒦)p(S)\geq(1-O({\varepsilon}))\cdot\textnormal{OPT}({\mathcal{K}}). Thus, we can use enumeration to guess SRS\cap R and then extend the solution using an LP relaxation similar to [9]. Crucial to our construction of SS is the notion of profit gap, used for identifying elements that may be added to SS by solving the LP (see Section 3). Since all the elements in SRS\setminus R are non-profitable, the profit loss caused by the few non-integral entries is negligible. Moreover, since |R|f(1ε)|R|\leq f\left(\frac{1}{{\varepsilon}}\right), guessing SRS\cap R can be done in 2f(1ε)2^{f\left(\frac{1}{{\varepsilon}}\right)} steps (in fact, we obtain a slightly better running time), eliminating the dependency of enumeration on the input size. The representative set is identified via matroid basis minimization, which can be solved by a simple greedy algorithm (see Section 3).

Organization. In Section 2 we give some definitions and notation. Section 3 presents our approximation scheme, EPTAS, and its analysis. In Section 4 we give a proof of correctness for algorithm FindRep that is used as a subroutine by the scheme. We conclude in Section 5 with a summary and open problems.

2 Preliminaries

For simplicity of the notation, for any set AA and an element ee, we use A+eA+e and AeA-e to denote A{e}A\cup\{e\} and A{e}A\setminus\{e\}, respectively. Also, for any kk\in\mathbb{R} let [k]={1,2,,k}[k]=\{1,2,\ldots,\left\lfloor k\right\rfloor\}. Finally, for a function f:ABf:A\rightarrow B and a subset of elements CAC\subseteq A, we define f(C)=eCf(e)f(C)=\sum_{e\in C}f(e).

2.1 Matroids

Let EE be a finite ground set and 2E{\mathcal{I}}\subseteq 2^{E} a non-empty set containing subsets of EE called the independent sets of EE. Then, =(E,){\mathcal{M}}=(E,{\mathcal{I}}) is a matroid if the following hold:

  1. 1.

    (Hereditary Property) For all AA\in{\mathcal{I}} and BAB\subseteq A, it holds that BB\in{\mathcal{I}}.

  2. 2.

    (Exchange Property) For any A,BA,B\in{\mathcal{I}} where |A|>|B||A|>|B|, there is eABe\in A\setminus B such that B+eB+e\in{\mathcal{I}}.

The next observation follows by repeatedly applying the exchange property.

Observation 2.1.

Given a matroid (E,)(E,{\mathcal{I}}) and A,BA,B\in{\mathcal{I}}, there is DABD\subseteq A\setminus B, |D|=max{|A||B|,0}|D|=\max\{|A|-|B|,0\} such that BDB\cup D\in{\mathcal{I}}.

A basis of a matroid =(E,){\mathcal{M}}=(E,{\mathcal{I}}) is an independent set BB\in{\mathcal{I}} such that for all eEBe\in E\setminus B it holds that B+eB+e\notin{\mathcal{I}}. Given a cost function c:E+c:E\rightarrow\mathbb{R}^{+}, we say that a basis BB of {\mathcal{M}} is a minimum basis of {\mathcal{M}} w.r.t. cc if, for any basis AA of {\mathcal{M}} it holds that c(B)c(A)c(B)\leq c(A). A minimum basis of {\mathcal{M}} w.r.t. cc can be easily constructed in polynomial-time using the greedy approach (see, e.g.,  [5]). In the following we define several operations on matroids that will be useful for deriving our results.

Definition 2.2.

Let =(E,){\mathcal{M}}=(E,{\mathcal{I}}) be a matroid.

  1. 1.

    (restriction) For every FEF\subseteq E define F={A|AF}{\mathcal{I}}_{\cap F}=\{A\in{\mathcal{I}}~{}|~{}A\subseteq F\} and F=(F,F){\mathcal{M}}\cap F=(F,{\mathcal{I}}_{\cap F}).

  2. 2.

    (contraction) For every FF\in{\mathcal{I}} define /F={AEF|AF}{\mathcal{I}}_{/F}=\{A\subseteq E\setminus F~{}|~{}A\cup F\in{\mathcal{I}}\} and /F=(EF,/F){\mathcal{M}}/F=(E\setminus F,{\mathcal{I}}_{/F}).

  3. 3.

    (truncation) For every qq\in\mathbb{N} define q={A||A|q}{\mathcal{I}}_{\leq q}=\{A\in{\mathcal{I}}~{}|~{}|A|\leq q\} and []q=(E,q)[{\mathcal{M}}]_{\leq q}=(E,{\mathcal{I}}_{\leq q}).

  4. 4.

    (union) Let M1=(E1,1),,(Ek,k)M_{1}=(E_{1},{\mathcal{I}}_{1}),\ldots,(E_{k},{\mathcal{I}}_{k}) be matroids; define i[k]Mi=(E,[k],,[k])\bigvee_{i\in[k]}M_{i}=(E_{\vee,[k]},{\mathcal{I}}_{\vee,[k]}), where
    E,[k]=i[k]EiE_{\vee,[k]}=\bigcup_{i\in[k]}E_{i} and ,[k]={i[k]Fi|i[k]:Fii}{\mathcal{I}}_{\vee,[k]}=\left\{\bigcup_{i\in[k]}F_{i}~{}\big{|}~{}\forall i\in[k]:F_{i}\in{\mathcal{I}}_{i}\right\}.

The next lemma gathers known results which follow directly from the definition of a matroid (see, e.g., [21]).

Lemma 2.3.

Let =(E,){\mathcal{M}}=(E,{\mathcal{I}}) be a matroid.

  1. 1.

    For any FEF\subseteq E, the restriction of {\mathcal{M}} to FF (i.e., F{\mathcal{M}}\cap F) is a matroid.

  2. 2.

    For any FF\in{\mathcal{I}}, the contraction of {\mathcal{M}} by FF (i.e., /F{\mathcal{M}}/F) is a matroid.

  3. 3.

    For any qq\in\mathbb{N}, the truncation of {\mathcal{M}} (i.e., []q[{\mathcal{M}}]_{\leq q}) is a matroid.

  4. 4.

    Given matroids M1=(E1,1),,Mk=(Ek,k)M_{1}=(E_{1},{\mathcal{I}}_{1}),\ldots,M_{k}=(E_{k},{\mathcal{I}}_{k}), the union of M1,,MkM_{1},\ldots,M_{k} (i.e., i[k]Mi\bigvee_{i\in[k]}M_{i}) is a matroid.

2.2 Matroid polytope

Let =(E,){\mathcal{M}}=(E,{\mathcal{I}}) be a matroid. Given BB\in{\mathcal{I}}, the indicator vector of BB is the vector 𝟙B{0,1}E\mathbbm{1}^{B}\in\{0,1\}^{E}, where for all aBa\in B and bEBb\in E\setminus B we have 𝟙aB=1\mathbbm{1}^{B}_{a}=1 and 𝟙bB=0\mathbbm{1}^{B}_{b}=0, respectively. Then the matroid polytope of {\mathcal{M}} is the convex hull of the set of indicator vectors of all independent sets of {\mathcal{M}}: P=conv{𝟙B|B}P_{{\mathcal{M}}}=\textsf{conv}\{\mathbbm{1}^{B}~{}|~{}B\in{\mathcal{I}}\}. The next observation will be used in the analysis of our scheme.

Observation 2.4.

Let =(E,){\mathcal{M}}=(E,{\mathcal{I}}) be a matroid, and x¯P\bar{x}\in P_{{\mathcal{M}}}. Then {eE|x¯e=1}\{e\in E~{}|~{}\bar{x}_{e}=1\}\in{\mathcal{I}}.

3 The Algorithm

In this section we present an EPTAS for BMI. Our scheme initially handles elements of high profits. This is done by finding a small representative set out of which the scheme selects the most profitable elements in the solution. More elements, of lower profits, are then added to the solution using a linear program. For the remainder of this section, fix a BMI instance 𝒦=(E,,c,p,β){\mathcal{K}}=(E,{\mathcal{I}},c,p,\beta) and a parameter 0<ε<120<{\varepsilon}<\frac{1}{2}. W.l.o.g., for all eEe\in E we assume that {e}\{e\}\in{\mathcal{I}} (otherwise ee cannot belong to any solution for 𝒦{\mathcal{K}}).

Let H(𝒦,ε)={E|p()>εOPT(𝒦)}H({\mathcal{K}},{\varepsilon})=\{\ell\in E~{}|~{}p(\ell)>{\varepsilon}\cdot\textnormal{OPT}({\mathcal{K}})\} be the set of profitable elements in 𝒦{\mathcal{K}}, and EH(𝒦,ε)E\setminus H({\mathcal{K}},{\varepsilon}) the set of non-profitable elements; when understood from the context, we simply use H=H(𝒦,ε)H=H({\mathcal{K}},{\varepsilon}). We can easily obtain a PTAS by enumerating over all subsets of at most ε1{\varepsilon}^{-1} profitable elements to find the profitable elements in the solution. However, such exhaustive search is done in time Ω(|𝒦|ε1)\Omega\left(|{\mathcal{K}}|^{{\varepsilon}^{-1}}\right), which cannot lead to an EPTAS. Thus, we take a different approach.

A key observation is that efficient solution can be obtained without enumerating over all subsets of profitable elements. Instead, we limit our scheme to a smaller search space using the notions of replacements and representative sets. Let SS be an independent set having a bounded number of elements. A replacement of SS is a subset of elements which can replace the profitable elements in SS, resulting with an independent set of lower cost and almost the same profit. A representative set RR is a subset of elements which contains a replacement within RR for any independent set with bounded number of elements. Definitions 3.1 and 3.2 give the formal properties of replacements and representative sets, respectively. Let q(ε)=εε1q({\varepsilon})={\varepsilon}^{-{\varepsilon}^{-1}}, and recall that q(ε)={A||A|q(ε)}{\mathcal{I}}_{\leq q({\varepsilon})}=\{A\in{\mathcal{I}}~{}|~{}|A|\leq q({\varepsilon})\} (the selection of value for q(ε)q({\varepsilon}) becomes clear in the proof of Lemma 3.3).

Definition 3.1.

Given a BMI instance 𝒦=(E,,c,p,β),0<ε<12{\mathcal{K}}=(E,{\mathcal{I}},c,p,\beta),0<{\varepsilon}<\frac{1}{2}, Gq(ε)G\in{\mathcal{I}}_{\leq q({\varepsilon})}, and ZGEZ_{G}\subseteq E, we say that ZGZ_{G} is a replacement of GG for 𝒦{\mathcal{K}} and ε{\varepsilon} if the following holds:

  1. 1.

    (GH)ZGq(ε)(G\setminus H)\cup Z_{G}\in{\mathcal{I}}_{\leq q({\varepsilon})}.

  2. 2.

    c(ZG)c(GH)c(Z_{G})\leq c(G\cap H).

  3. 3.

    p((GH)ZG)(1ε)p(G)p\left((G\setminus H)\cup Z_{G}\right)\geq(1-{\varepsilon})\cdot p(G).

  4. 4.

    |ZG||GH||Z_{G}|\leq|G\cap H|.

Definition 3.2.

Given a BMI instance 𝒦=(E,,c,p,β),0<ε<12{\mathcal{K}}=(E,{\mathcal{I}},c,p,\beta),0<{\varepsilon}<\frac{1}{2} and RER\subseteq E, we say that RR is a representative set of 𝒦{\mathcal{K}} and ε{\varepsilon} if, for any Gq(ε)G\in{\mathcal{I}}_{\leq q({\varepsilon})}, there is a replacement ZGRZ_{G}\subseteq R of GG for 𝒦{\mathcal{K}} and ε{\varepsilon}.

In particular, observe that for any solution SS of 𝒦{\mathcal{K}} we have that SHS\cap H is a replacement of SS; also, EE is a representative set. In the next lemma we show that there exists an almost optimal solution in which all profitable elements belong to a given representative set. Hence, guessing the profitable elements only requires enumerating over subsets of a representative set.

Lemma 3.3.

Let 𝒦=(E,,c,p,β){\mathcal{K}}=(E,{\mathcal{I}},c,p,\beta) be a BMI instance and 0<ε<120<{\varepsilon}<\frac{1}{2}. Also, let RR be a representatives set of 𝒦{\mathcal{K}} and ε{\varepsilon}. Then, there is a solution SS of 𝒦{\mathcal{K}} such that the following holds.

  1. 1.

    SHRS\cap H\subseteq R.

  2. 2.

    p(S)(13ε)OPT(K)p\left(S\right)\geq(1-3{\varepsilon})\textnormal{OPT}(K).

We give a brief outline of the proof of Lemma 3.3. Informally, we consider the elements in some optimal solution, OPT, in non-increasing order by profit. We then partition these elements into three sets: L,JiL,J_{i^{*}} and QQ, such that the maximal profit in QQ is at most ε{\varepsilon} times the minimum profit in LL. This is the profit gap of LL and QQ. In addition, Lq(ε)L\in{\mathcal{I}}_{\leq q({\varepsilon})}, and p(Ji)εOPT(𝒦)p(J_{i^{*}})\leq{\varepsilon}\cdot\textnormal{OPT}({\mathcal{K}}). Thus, we can use ZLRZ_{L}\subseteq R as a replacement of LL, i.e., LHL\cap H will be replaced by elements in ZLZ_{L} (note that ZLZ_{L} is not necessarily a subset of the profitable elements). We now discard JiJ_{i^{*}}, and define ΔL=(LH)ZL\Delta_{L}=(L\setminus H)\cup Z_{L}. As ΔLQ\Delta_{L}\cup Q may not be an independent set, we use Observation 2.1 to construct TQ,|T||Q||L|T\subseteq Q,|T|\geq|Q|-|L| such that S=ΔLTS=\Delta_{L}\cup T\in{\mathcal{I}}. An illustration of the proof is given in Figure 1.

Refer to caption
Figure 1: The construction of the solution SS in the proof of Lemma 3.3.

Proof of Lemma 3.3: With a slight abuse of notation, we use OPT also to denote the profit of an optimal solution for 𝒦{\mathcal{K}}. Given an optimal solution, we partition a subset of the elements in the solution into ε1{\varepsilon}^{-1} disjoint sets (some sets may be empty). Specifically, let N=ε1N={\left\lceil{\varepsilon}^{-1}\right\rceil}; for all i[N]i\in[N] define

Ji={eOPT|p(e)(εiOPT(𝒦),εi1OPT(𝒦)]}.J_{i}=\big{\{}e\in\textnormal{OPT}~{}\big{|}~{}p(e)\in\big{(}{\varepsilon}^{i}\cdot\textnormal{OPT}({\mathcal{K}}),{\varepsilon}^{i-1}\cdot\textnormal{OPT}({\mathcal{K}})\big{]}\big{\}}. (1)

Let i=argmini[N]p(Ji)i^{*}=\operatorname*{arg\,min}_{i\in[N]}p(J_{i}). By (1) we have at least ε1{\varepsilon}^{-1} disjoint sets; thus, p(Ji)εOPT(𝒦)p(J_{i^{*}})\leq{\varepsilon}\cdot\textnormal{OPT}({\mathcal{K}}). Now, let L=k[i1]JkL~{}=~{}\bigcup_{k\in[i^{*}-1]}J_{k} be the subset of all elements in OPT of profits greater than εi1OPT(𝒦){\varepsilon}^{i^{*}-1}\cdot\textnormal{OPT}({\mathcal{K}}), and Q=OPT(LJi)Q=\textnormal{OPT}\setminus(L\cup J_{i^{*}}). To complete the proof of the lemma, we need several claims.

Claim 3.4.

Lq(ε)L\in{\mathcal{I}}_{\leq q({\varepsilon})}.

Proof.

Since LOPTL\subseteq\textnormal{OPT}, by the hereditary property of (E,)(E,{\mathcal{I}}) we have that LL\in{\mathcal{I}}. Also,

|L|eLp(e)εi1OPT(𝒦)=p(L)εi1OPT(𝒦)ε(i1)εN+1εε1=q(ε)|L|\leq\sum_{e\in L}\frac{p(e)}{{\varepsilon}^{i^{*}-1}\cdot\textnormal{OPT}({\mathcal{K}})}=\frac{p(L)}{{\varepsilon}^{i^{*}-1}\cdot\textnormal{OPT}({\mathcal{K}})}\leq{\varepsilon}^{-(i^{*}-1)}\leq{\varepsilon}^{-N+1}\leq{\varepsilon}^{-{\varepsilon}^{-1}}=q({\varepsilon})

The first inequality holds since p(e)εi1OPTp(e)\geq{\varepsilon}^{i^{*}-1}\cdot\textnormal{OPT} for all eLe\in L. For the second inequality, we note that LL is a solution for 𝒦{\mathcal{K}}. By the above and Definition 2.2, it follows that Lq(ε)L\in{\mathcal{I}}_{\leq q({\varepsilon})}. \square
By Claim 3.4 and as RR is a representative set, it follows that LL has a replacement ZLRZ_{L}\subseteq R. Let ΔL=(LH)ZL\Delta_{L}=(L\setminus H)\cup Z_{L}. By Property 1 of Definition 3.1, we have that ΔLq(ε)\Delta_{L}\in{\mathcal{I}}_{\leq q({\varepsilon})}, and by Definition 2.2 it holds that q(ε){\mathcal{I}}_{\leq q({\varepsilon})}\subseteq{\mathcal{I}}. Hence, ΔL\Delta_{L}\in{\mathcal{I}}. Furthermore, as QOPTQ\subseteq\textnormal{OPT}\in{\mathcal{I}}, by the hereditary property for (E,)(E,{\mathcal{I}}) we have that QQ\in{\mathcal{I}}. Therefore, by Observation 2.1, there is a subset TQΔLT\subseteq Q\setminus\Delta_{L}, where |T|=max{|Q||ΔL|,0}|T|=\max\{|Q|-|\Delta_{L}|,0\}, such that ΔLT\Delta_{L}\cup T\in{\mathcal{I}}.

Let S=ΔLTS=\Delta_{L}\cup T. We show that SS satisfies the conditions of the lemma.

Claim 3.5.

SS is a solution for 𝒦{\mathcal{K}}.

Proof.

By the definition of TT it holds that S=ΔLTS=\Delta_{L}\cup T\in{\mathcal{I}}. Moreover,

c(S)=c(ΔLT)c(ZL)+c(LH)+c(T)c(LH)+c(LH)+c(T)c(L)+c(Q)c(OPT)β.\begin{array}[]{ll}c(S)&=c(\Delta_{L}\cup T)\\ &\leq c(Z_{L})+c(L\setminus H)+c(T)\\ &\leq c(L\cap H)+c(L\setminus H)+c(T)\leq c(L)+c(Q)\\ &\leq c(\textnormal{OPT})\\ &\leq\beta.\end{array}

The second inequality holds since c(ZL)c(LH)c(Z_{L})\leq c(L\cap H) (see Property 2 of Definition 3.1). For the third inequality, recall that TQT\subseteq Q. The last inequality holds since OPT is a solution for 𝒦{\mathcal{K}}. \square

The proof of the next claim relies on the profit gap between the elements in QQ and LL.

Claim 3.6.

p(QT)εOPT(𝒦)p\left(Q\setminus T\right)\leq{\varepsilon}\cdot\textnormal{OPT}({\mathcal{K}}).

Proof.

Observe that

|QT||ΔL||ZL|+|LH||LH|+|LH||L|.|Q\setminus T|\leq|\Delta_{L}|\leq|Z_{L}|+|L\setminus H|\leq|L\cap H|+|L\setminus H|\leq|L|. (2)

The first inequality follows from the definition of TT. For the third inequality, we use Property 4 of Definition 3.1. Hence,

p(QT)\displaystyle p(Q\setminus T)\leq{} |QT|εiOPT(𝒦)|L|εiOPT(𝒦)εp(L)εOPT(𝒦).\displaystyle|Q\setminus T|\cdot{\varepsilon}^{i^{*}}\cdot\textnormal{OPT}({\mathcal{K}})\leq|L|\cdot{\varepsilon}^{i^{*}}\cdot\textnormal{OPT}({\mathcal{K}})\leq{\varepsilon}\cdot p(L)\leq{\varepsilon}\cdot\textnormal{OPT}({\mathcal{K}}).

The first inequality holds since p(e)εiOPT(𝒦)p(e)\leq{\varepsilon}^{i^{*}}\cdot\textnormal{OPT}({\mathcal{K}}) for all eQe\in Q. The second inequality is by (2). The third inequality holds since p(e)>εi1OPT(𝒦)p(e)>{\varepsilon}^{i^{*}-1}\cdot\textnormal{OPT}({\mathcal{K}}) for all eLe\in L. \square

Claim 3.7.

p(S)(13ε)OPT(𝒦)p\left(S\right)\geq(1-3{\varepsilon})\textnormal{OPT}({\mathcal{K}}).

Proof.

By Property 3 of Definition 3.1,

p(ΔL)=\displaystyle p(\Delta_{L})={} p((LH)ZL)(1ε)p(L).\displaystyle p((L\setminus H)\cup Z_{L})\geq(1-{\varepsilon})\cdot p(L). (3)

Moreover,

p(T)\displaystyle p(T){} p(Q)p(QT)p(Q)εOPT(𝒦)(1ε)p(Q)εOPT(𝒦).\displaystyle\geq p(Q)-p(Q\setminus T)\geq p(Q)-{\varepsilon}\cdot\textnormal{OPT}({\mathcal{K}})\geq(1-{\varepsilon})\cdot p(Q)-{\varepsilon}\cdot\textnormal{OPT}({\mathcal{K}}). (4)

The second inequality is by Claim 3.6. Using (3) and (4), we have that

p(ΔL)+p(T)(1ε)p(LQ)εOPT(𝒦)=(1ε)p(OPTJi)εOPT(𝒦)(13ε)OPT(𝒦).\begin{array}[]{ll}p(\Delta_{L})+p(T)&\geq(1-{\varepsilon})\cdot p(L\cup Q)-{\varepsilon}\cdot\textnormal{OPT}({\mathcal{K}})\\ &=(1-{\varepsilon})\cdot p(\textnormal{OPT}\setminus J_{i^{*}})-{\varepsilon}\cdot\textnormal{OPT}({\mathcal{K}})\\ &\geq(1-3{\varepsilon})\cdot\textnormal{OPT}({\mathcal{K}}).\end{array}

The last inequality holds since p(Ji)εOPT(𝒦)p(J_{i^{*}})\leq{\varepsilon}\cdot\textnormal{OPT}({\mathcal{K}}). Observe that S=ΔLTS=\Delta_{L}\cup T and TΔL=T\cap\Delta_{L}=\emptyset. Therefore,

p(S)=p(ΔL)+p(T)(13ε)OPT(𝒦).p(S)=p(\Delta_{L})+p(T)\geq(1-3{\varepsilon})\cdot\textnormal{OPT}({\mathcal{K}}).

\square
Finally, we note that, by (1) and the definition of QQ, QH=Q\cap H=\emptyset. Consequently, by the definition of SS, we have that SHZLS\cap H\subseteq Z_{L}. As ZLRZ_{L}\subseteq R (by Definition 3.2), it follows that SHRS\cap H\subseteq R. Hence, using Claims 3.5 and 3.7, we have the statement of the lemma. ∎

Our scheme for BMI constructs a representative set whose cardinality depends solely on ε{\varepsilon}. To this end, we first partition the profitable elements (and possibly some more elements) into a small number of profit classes, where elements from the same profit class have similar profits. The profit classes are derived from a 22-approximation α\alpha for OPT(𝒦)\textnormal{OPT}({\mathcal{K}}), which can be easily computed in polynomial time. specifically, for all r[log1ε(ε2)+1]r\in[\log_{1-{\varepsilon}}\left(\frac{{\varepsilon}}{2}\right)+1] define the rr-profit class as

Cr(α)={eE|p(e)2α((1ε)r,(1ε)r1]}.C_{r}(\alpha)=\left\{e\in E~{}\bigg{|}~{}\frac{p(e)}{2\cdot\alpha}\in\big{(}(1-{\varepsilon})^{r},(1-{\varepsilon})^{r-1}\big{]}\right\}. (5)

For each r[log1ε(ε2)+1]r\in[\log_{1-{\varepsilon}}\left(\frac{{\varepsilon}}{2}\right)+1], we define [(E,)Cr(α)]q(ε)\left[(E,{\mathcal{I}})\cap C_{r}(\alpha)\right]_{\leq q({\varepsilon})}, the corresponding matroid for the rr-profit class. We construct a representative set by computing a minimum basis w.r.t. the cost function cc, for the matroid defined as the union of the corresponding matroids of all profit classes. Note that by Lemma 2.3 the latter is a matroid. The pseudocode of algorithm FindRep, which outputs a representative set, is given in Algorithm 1.

1
2Compute a 22-approximation SS^{*} for 𝒦{\mathcal{K}} using a PTAS for BMI with parameter ε=12{\varepsilon}^{\prime}=\frac{1}{2}.
3Set α=p(S)\alpha=p(S^{*}).
Return a minimum basis w.r.t. cc of the matroid
r[log1ε(ε2)+1][(E,)Cr(α)]q(ε).\bigvee_{r\in[\log_{1-{\varepsilon}}\left(\frac{{\varepsilon}}{2}\right)+1]}\left[(E,{\mathcal{I}})\cap C_{r}(\alpha)\right]_{\leq q({\varepsilon})}.
Algorithm 1 FindRep(𝒦=(E,,c,p,β),ε)\textsf{FindRep}({\mathcal{K}}=(E,{\mathcal{I}},c,p,\beta),{\varepsilon})
Lemma 3.8.

Given a BMI instance 𝒦=(E,,c,p,β){\mathcal{K}}=(E,{\mathcal{I}},c,p,\beta) and 0<ε<120<{\varepsilon}<\frac{1}{2}, Algorithm 1 returns in time poly(|𝒦|)\textnormal{poly}(|{\mathcal{K}}|) a representative set RR of 𝒦{\mathcal{K}} and ε{\varepsilon}, such that |R|q(ε)(log1ε(ε2)+1)|R|\leq q({\varepsilon})\cdot\left(\log_{1-{\varepsilon}}\left(\frac{{\varepsilon}}{2}\right)+1\right).

In the following we outline the main ideas of the proof. Let R=FindRep(𝒦,ε)R=\textsf{FindRep}({\mathcal{K}},{\varepsilon}), and consider some Gq(ε)G\in{\mathcal{I}}_{\leq q({\varepsilon})} and e(GH)Re\in(G\cap H)\setminus R. By (5) the element ee belongs to some profit class Cr(α),r[log1ε(ε2)+1]C_{r}(\alpha),r\in[\log_{1-{\varepsilon}}\left(\frac{{\varepsilon}}{2}\right)+1]. Since RR is a minimum basis w.r.t. cc, we can use matroid properties to show that there is some bCr(α)Rb\in C_{r}(\alpha)\cap R of cost c(b)c(e)c(b)\leq c(e) such that Ga+bq(ε)G-a+b\in{\mathcal{I}}_{\leq q({\varepsilon})}. We now keep replacing elements in (GH)R(G\cap H)\setminus R in a similar manner, until no such element exists. Thus, we have a replacement of GG within RR, i.e., RR is a representative set. A formal proof of Lemma 3.8 is given in Section 4.

Our scheme uses the representative set to select profitable elements for the solution. Using a linear program, the solution is extended to include also non-profitable elements. As the exact set of non-profitable elements is unknown, we use an approximation for the optimal profit. Specifically, let OPT(𝒦)2αOPT(𝒦)\frac{\textnormal{OPT}({\mathcal{K}})}{2}\leq\alpha\leq\textnormal{OPT}({\mathcal{K}}), and denote by E(α)={eE|p(e)2εα}E(\alpha)=\{e\in E~{}|~{}p(e)\leq 2{\varepsilon}\cdot\alpha\} the set including the non-profitable elements, and possibly also profitable elements ee such that p(e)2εOPT(𝒦)p(e)\leq 2{\varepsilon}\cdot\textnormal{OPT}({\mathcal{K}}).

The LP is based on the matroid polytope of the following matroid. Given a solution FF for 𝒦{\mathcal{K}}, we define F(α)=((E,)/F)E(α){\mathcal{M}}_{F}(\alpha)=\left((E,{\mathcal{I}})/F\right)\cap E(\alpha); in other words, F(α)=(E(α),F(α)){\mathcal{M}}_{F}(\alpha)=(E(\alpha),{\mathcal{I}}_{F}(\alpha)), where F(α)={AE(α)|AF}{\mathcal{I}}_{F}(\alpha)=\{A\subseteq E(\alpha)~{}|~{}A\cup F\in{\mathcal{I}}\}. Note that F(α){\mathcal{M}}_{F}(\alpha) is indeed a matroid, by Properties 1 and 2 of Lemma 2.3. The LP formulation is given by

LP(𝒦,F,α)max\displaystyle\textnormal{LP}({\mathcal{K}},F,\alpha)~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}\max eE(α)Fx¯ep(e)\displaystyle~{}~{}~{}\sum_{e\in E(\alpha)\setminus F}\bar{x}_{e}\cdot p(e)~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{} (6)
s.t. eE(α)Fx¯ec(e)βc(F)\displaystyle~{}~{}~{}\sum_{e\in E(\alpha)\setminus F}\bar{x}_{e}\cdot c(e)\leq\beta-c(F)~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}
x¯PF(α)\displaystyle~{}~{}~{}\bar{x}\in P_{{\mathcal{M}}_{F}(\alpha)}

The linear program LP(𝒦,F,α)\textnormal{LP}({\mathcal{K}},F,\alpha) maximizes the total profit of a point in the matroid polytope of F(α){\mathcal{M}}_{F}(\alpha) (i.e., PF(α)P_{{\mathcal{M}}_{F}(\alpha)}), such that the total cost of elements is at most βc(F)\beta-c(F); that is, the residual budget after selecting for the solution the elements in FF.

Observation 3.9.

Let 𝒦=(E,,c,p,β){\mathcal{K}}=(E,{\mathcal{I}},c,p,\beta) be a BMI instance, OPT(𝒦)2αOPT(𝒦)\frac{\textnormal{OPT}({\mathcal{K}})}{2}\leq\alpha\leq\textnormal{OPT}({\mathcal{K}}), SS a solution for 𝒦{\mathcal{K}}, and x¯\bar{x} an optimal basic solution for LP(𝒦,SH,α)\textnormal{LP}({\mathcal{K}},S\cap H,\alpha). Then, eE(α)(SH)x¯ep(e)p(SH)\sum_{e\in E(\alpha)\setminus(S\cap H)}\bar{x}_{e}\cdot p(e)\geq p\left(S\setminus H\right).

It is folklore that a linear program such as LP(𝒦,F,α)\textnormal{LP}({\mathcal{K}},F,\alpha) can be solved in polynomial-time in |𝒦||{\mathcal{K}}|. As we could not find a proper reference, we include the proof of the next lemma in the appendix.

Lemma 3.10.

For any BMI instance 𝒦=(E,,c,p,β){\mathcal{K}}=(E,{\mathcal{I}},c,p,\beta), OPT(𝒦)2αOPT(𝒦)\frac{\textnormal{OPT}({\mathcal{K}})}{2}\leq\alpha\leq\textnormal{OPT}({\mathcal{K}}), and a solution FF of 𝒦{\mathcal{K}}, a basic optimal solution of LP(𝒦,F,α)\textnormal{LP}({\mathcal{K}},F,\alpha) can be found in time poly(|𝒦|)\textnormal{poly}(|{\mathcal{K}}|).

The next lemma will be useful for deriving a solution of high profit using LP(𝒦,F,α)\textnormal{LP}({\mathcal{K}},F,\alpha). The proof follows as a special case of a result of [9].

Lemma 3.11.

Let 𝒦=(E,,c,p,β){\mathcal{K}}=(E,{\mathcal{I}},c,p,\beta) be a BMI instance, OPT(𝒦)2αOPT(𝒦)\frac{\textnormal{OPT}({\mathcal{K}})}{2}\leq\alpha\leq\textnormal{OPT}({\mathcal{K}}), a solution FF of 𝒦{\mathcal{K}}, and x¯\bar{x} a basic solution for LP(𝒦,F,α)\textnormal{LP}({\mathcal{K}},F,\alpha). Then x¯\bar{x} has at most two non-integral entries.

Using the above, we have the required components for an EPTAS for BMI. Let RR be the representative set returned by FindRep(𝒦,ε)\textsf{FindRep}({\mathcal{K}},{\varepsilon}). For all solutions FRF\subseteq R with |F|ε1|F|\leq{\varepsilon}^{-1}, we find a basic optimal solution λ¯F\bar{\lambda}^{F} for LP(𝒦,F,α)\textnormal{LP}({\mathcal{K}},F,\alpha) and define CF={eE(α)F|λ¯eF=1}FC_{F}=\{e\in E(\alpha)\setminus F~{}|~{}\bar{\lambda}^{F}_{e}=1\}\cup F as the solution of FF. Our scheme iterates over the solutions CFC_{F} for all such subsets FF and chooses a solution CFC_{F^{*}} of maximal total profit. The pseudocode of the scheme is given in Algorithm 2.

1
2Construct the representative elements R=FindRep(𝒦,ε)R=\textsf{FindRep}({\mathcal{K}},{\varepsilon}).
3Compute a 22-approximation SS^{*} for 𝒦{\mathcal{K}} using a PTAS for BMI with parameter ε=12{\varepsilon}^{\prime}=\frac{1}{2}.
4Set α=p(S)\alpha=p(S^{*}).
5Initialize an empty solution AA\leftarrow\emptyset.
6for FR s.t. |F|ε1 and F is a solution of 𝒦F\subseteq R\textnormal{ s.t. }|F|\leq{\varepsilon}^{-1}\textnormal{ and }F\textnormal{ is a solution of }{\mathcal{K}}  do
7      
8      Find a basic optimal solution λ¯F\bar{\lambda}^{F} of LP(𝒦,F,α)\textnormal{LP}({\mathcal{K}},F,\alpha).
9      Let CF={eE(α)F|λ¯eF=1}FC_{F}=\{e\in E(\alpha)\setminus F~{}|~{}\bar{\lambda}^{F}_{e}=1\}\cup F.
10      if p(CF)>p(A)p\left(C_{F}\right)>p(A) then
11            
12            Update ACFA\leftarrow C_{F}
13       end if
14      
15 end for
16
Return AA.
Algorithm 2 EPTAS(𝒦=(E,,c,p,β),ε)\textsf{EPTAS}({\mathcal{K}}=(E,{\mathcal{I}},c,p,\beta),{\varepsilon})
Lemma 3.12.

Given a BMI instance 𝒦=(E,,c,p,β){\mathcal{K}}=(E,{\mathcal{I}},c,p,\beta) and 0<ε<120<{\varepsilon}<\frac{1}{2}, Algorithm 2 returns a solution for 𝒦{\mathcal{K}} of profit at least (17ε)OPT(𝒦)(1-7{\varepsilon})\cdot\textnormal{OPT}({\mathcal{K}}).

Proof.

By Lemma 3.3, there is a solution SS for 𝒦{\mathcal{K}} such that SHRS\cap H\subseteq R, and p(S)(13ε)OPT(K)p\left(S\right)\geq(1-3{\varepsilon})\textnormal{OPT}(K). As for all eSHe\in S\cap H we have p(e)>εOPT(𝒦)p(e)>{\varepsilon}\cdot\textnormal{OPT}({\mathcal{K}}), and SS is a solution for 𝒦{\mathcal{K}}, it follows that |SH|ε1|S\cap H|\leq{\varepsilon}^{-1}. We note that there is an iteration of Step 2 in which F=SHF=S\cap H; thus, in Step 2 we construct a basic optimal solution λ¯SH\bar{\lambda}^{S\cap H} of LP(𝒦,SH,α)\textnormal{LP}({\mathcal{K}},S\cap H,\alpha). We use X(A)={eE(α)A|λ¯eA=1}X(A)=\{e\in E(\alpha)\setminus A~{}|~{}\bar{\lambda}^{A}_{e}=1\} for every basic solution λ¯A\bar{\lambda}^{A} computed in Step 2 for ARA\subseteq R. Then,

p(X(SH))=\displaystyle p\left(X(S\cap H)\right)={} eE(α)(SH)s.t.λ¯eSH=1p(e)\displaystyle\sum_{e\in E(\alpha)\setminus(S\cap H)~{}\text{s.t.}~{}\bar{\lambda}^{S\cap H}_{e}=1}p(e) (7)
\displaystyle\geq{} eE(α)(SH)λ¯eSHp(e)22εOPT(𝒦)\displaystyle\sum_{e\in E(\alpha)\setminus(S\cap H)}\bar{\lambda}^{S\cap H}_{e}\cdot p(e)-2\cdot 2{\varepsilon}\cdot\textnormal{OPT}({\mathcal{K}})
\displaystyle\geq{} p(SH)4εOPT(𝒦).\displaystyle p(S\setminus H)-4{\varepsilon}\cdot\textnormal{OPT}({\mathcal{K}}).

The first inequality holds since, by Lemma 3.11, |{eE(α)(SH)|λ¯eSH(0,1)}|2|\{e\in E(\alpha)\setminus(S\cap H)~{}|~{}\bar{\lambda}^{S\cap H}_{e}\in(0,1)\}|\leq 2, and for all eE(α)e\in E(\alpha), p(e)2εα2εOPT(𝒦)p(e)\leq 2{\varepsilon}\cdot\alpha\leq 2{\varepsilon}\cdot\textnormal{OPT}({\mathcal{K}}). The second inequality follows from Observation 3.9. Now,

p(CSH)=\displaystyle p(C_{S\cap H})={} p(SH)+p(X(SH))p(S)4εOPT(𝒦)(17ε)OPT(K).\displaystyle p(S\cap H)+p\left(X(S\cap H)\right)\geq p(S)-4{\varepsilon}\cdot\textnormal{OPT}({\mathcal{K}})\geq(1-7{\varepsilon})\textnormal{OPT}(K). (8)

The first inequality uses (7). The last inequality is by Lemma 3.3.

Claim 3.13.

A=EPTAS(𝒦,ε)A=\textnormal{{EPTAS}}({\mathcal{K}},{\varepsilon}) is a solution of 𝒦{\mathcal{K}}.

Proof.

If A=A=\emptyset the claim trivially follows since \emptyset is a solution of 𝒦{\mathcal{K}}. Otherwise, by Step 2 of Algorithm 2, there is a solution FF of 𝒦{\mathcal{K}} such that A=CFA=C_{F}. By Observation 2.4, X(F)X(F) is an independent set in the matroid MF(α)M_{F}(\alpha). Also, recall that F(α)=((E,)/F)E(α){\mathcal{M}}_{F}(\alpha)=\left((E,{\mathcal{I}})/F\right)\cap E(\alpha). Hence, by Definition 2.2, we have that X(F)F=CFX(F)\cup F=C_{F}\in{\mathcal{I}}. Additionally,

c(CF)=c(F)+eX(F)c(e)c(F)+βc(F)=β.c\left(C_{F}\right)=c(F)+\sum_{e\in X(F)}c(e)\leq c(F)+\beta-c(F)=\beta.

The inequality follows from (6). By the above, we conclude that AA is a solution for 𝒦{\mathcal{K}}. \square

By Claim 3.13, Steps 2, 2 and 2 of Algorithm 2 and (8), we have that A=EPTAS(𝒦,ε)A=\textsf{EPTAS}({\mathcal{K}},{\varepsilon}) is a solution for 𝒦{\mathcal{K}} satisfying p(A)p(CSH)(17ε)OPT(K)p(A)\geq p(C_{S\cap H})\geq(1-7{\varepsilon})\textnormal{OPT}(K). This completes the proof.

Lemma 3.14.

Given a BMI instance 𝒦=(E,,c,p,β){\mathcal{K}}=(E,{\mathcal{I}},c,p,\beta) and 0<ε<120<{\varepsilon}<\frac{1}{2}, Algorithm 2 runs in time 2O(ε2log1ε)poly(|𝒦|)2^{O\left({\varepsilon}^{-2}\log\frac{1}{{\varepsilon}}\right)}\cdot\textnormal{poly}(|{\mathcal{K}}|).

Proof.

By Lemma 3.8, the time complexity of Step 2 is poly(|𝒦|)\textnormal{poly}(|{\mathcal{K}}|). Step 2 can be computed in time poly(|𝒦|)\textnormal{poly}(|{\mathcal{K}}|), by using a PTAS for BMI taking ε=12{\varepsilon}=\frac{1}{2} (see, e.g., [9]). Now, using logarithm rules we have

log1ε(ε2)+1ln(2ε)ln(1ε)+12ε1ε+13ε2.\log_{1-{\varepsilon}}\left(\frac{{\varepsilon}}{2}\right)+1\leq\frac{\ln\left(\frac{2}{{\varepsilon}}\right)}{-\ln\left(1-{\varepsilon}\right)}+1\leq\frac{2{\varepsilon}^{-1}}{{\varepsilon}}+1\leq 3{\varepsilon}^{-2}. (9)

The second inequality follows from x<ln(1x),x>1,x0x<-\ln(1-x),\forall x>-1,x\neq 0, and lny<y,y>0\ln y<y,\forall y>0. Therefore,

|R|(log1ε(ε2)+1)q(ε)3ε2εε1ε3ε1.|R|\leq\left(\log_{1-{\varepsilon}}\left(\frac{{\varepsilon}}{2}\right)+1\right)\cdot q({\varepsilon})\leq 3{\varepsilon}^{-2}\cdot{\varepsilon}^{-{\varepsilon}^{-1}}\leq{\varepsilon}^{-3{\varepsilon}^{-1}}. (10)

The first inequality is by Lemma 3.8. The second inequality is by (9). Let

W={FR||F|ε1,F,c(F)β}W=\big{\{}F\subseteq R~{}\big{|}~{}|F|\leq{\varepsilon}^{-1},F\in{\mathcal{I}},c(F)\leq\beta\big{\}}

be the set of solutions considered in Step 2 of Algorithm 2. Then,

|W|\displaystyle|W|\leq{} (|R|+1)ε1(ε3ε1+1)ε1(ε4ε2)=2O(ε2log1ε).\displaystyle\left(|R|+1\right)^{{\varepsilon}^{-1}}\leq{\left({\varepsilon}^{-3{\varepsilon}^{-1}}+1\right)}^{{\varepsilon}^{-1}}\leq{\left({\varepsilon}^{-4{\varepsilon}^{-2}}\right)}=2^{O\left({\varepsilon}^{-2}\log\frac{1}{{\varepsilon}}\right)}. (11)

The second inequality is by (10). Hence, by (11), the number of iterations of the for loop in Step 2 is bounded by 2O(ε2log1ε)2^{O\left({\varepsilon}^{-2}\log\frac{1}{{\varepsilon}}\right)}. In addition, by Lemma 3.10, the running time of each iteration is poly(|𝒦|)\textnormal{poly}(|{\mathcal{K}}|). By the above, the running time of Algorithm 2 is 2O(ε2log1ε)poly(|𝒦|)2^{O\left({\varepsilon}^{-2}\log\frac{1}{{\varepsilon}}\right)}\cdot\textnormal{poly}(|{\mathcal{K}}|). ∎

Proof of Theorem 1.1: Given a BMI instance 𝒦{\mathcal{K}} and 0<ε<120<{\varepsilon}<\frac{1}{2}, using Algorithm 2 for 𝒦{\mathcal{K}} with parameter ε7\frac{{\varepsilon}}{7} we have the desired approximation guarantee by Lemma 3.12. Also, by Lemma 3.14, the running time is 2O(ε2log1ε)poly(|𝒦|)2^{O\left({\varepsilon}^{-2}\log\frac{1}{{\varepsilon}}\right)}\cdot\textnormal{poly}(|{\mathcal{K}}|). ∎

4 Correctness of FindRep

In this section we give the proof of Lemma 3.8. The proof is based on substitution of subsets by profit classes. A substitution is closely related to replacement (see Definition 3.1); indeed, we can construct replacements for independent sets (and therefore a representative set) by specific substitutions. We start by stating several lemmas that will be used in the proof of Lemma 3.8. The first lemma refers to a generalized exchange property of matroids.

Lemma 4.1.

Let =(E,){\mathcal{M}}=(E,{\mathcal{I}}) be a matroid, A,BA,B\in{\mathcal{I}}, and aABa\in A\setminus B such that B+aB+a\notin{\mathcal{I}}. Then there is bBAb\in B\setminus A such that Aa+bA-a+b\in{\mathcal{I}}.

Proof.

By Observation 2.1, there is CAB,|C|=max{|A||B|,0}C\subseteq A\setminus B,|C|=\max\{|A|-|B|,0\} such that BCB\cup C\in{\mathcal{I}}. Then,

|BC||B|+|A||B|>|A|1=|Aa|.|B\cup C|\geq|B|+|A|-|B|>|A|-1=|A-a|. (12)

The first inequality holds since |C|=max{|A||B|,0}|C|=\max\{|A|-|B|,0\} and CB=C\cap B=\emptyset. Recall that BCB\cup C\in{\mathcal{I}}; also, by the hereditary property of {\mathcal{M}}, AaA-a\in{\mathcal{I}}. Therefore, by (12) and the exchange property of {\mathcal{M}}, there is b(BC)(Aa)b\in(B\cup C)\setminus(A-a) such that Aa+bA-a+b\in{\mathcal{I}}. As CAC\subseteq A, it follows that (BC)(Aa)(BA)+a(B\cup C)\setminus(A-a)\subseteq(B\setminus A)+a, thus b(BA)+ab\in(B\setminus A)+a. Now, assume towards contradiction that b=ab=a; then, aCa\in C since b(BC)(Aa)b\in(B\cup C)\setminus(A-a) and aABa\in A\setminus B. As BCB\cup C\in{\mathcal{I}}, by the hereditary property of {\mathcal{M}} it follows that B+aB+a\in{\mathcal{I}}. Contradiction. Hence, bab\neq a and it follows that bBAb\in B\setminus A and Aa+bA-a+b\in{\mathcal{I}}. ∎

The next lemma gives a general property of minimum bases of matroids.

Lemma 4.2.

Given a matroid =(E,){\mathcal{M}}=(E,{\mathcal{I}}) and a weight function w:E0w:E\rightarrow\mathbb{R}_{\geq 0}, let BB be a minimum basis of {\mathcal{M}} w.r.t. ww. Then, for any aEBa\in E\setminus B it holds that {eB|w(e)w(a)}+a\{e\in B~{}|~{}w(e)\leq w(a)\}+a\notin{\mathcal{I}}.

Proof.

Let D={eB|w(e)w(a)}D=\{e\in B~{}|~{}w(e)\leq w(a)\}. Assume towards contradiction that D+aD+a\in{\mathcal{I}}. Then, by Observation 2.1 there is CB(D+a),|C|=max{|B||D+a|,0}C\in B\setminus(D+a),|C|=\max\{|B|-|D+a|,0\} such that (D+a)C(D+a)\cup C\in{\mathcal{I}}; let F=(D+a)CF=(D+a)\cup C. Note that |F|=|D+a|+|C||D+a|+|B||D+a|=|B||F|=|D+a|+|C|\geq|D+a|+|B|-|D+a|=|B|. Hence, as |F||B||F|\geq|B| and BB is a basis of {\mathcal{M}}, we have that FF is a basis of {\mathcal{M}} (since all bases have the same cardinality; see, e.g., (39.2) in [21]). Noting that aBa\notin B, FaBF-a\subseteq B and |F|=|B||F|=|B|, there is eBFe\in B\setminus F such that F=B+aeF=B+a-e. As eBFe\in B\setminus F and BFBDB\setminus F\subseteq B\setminus D it holds that eBDe\in B\setminus D, and it follows that w(a)<w(e)w(a)<w(e). Therefore,

w(F)=w((D+a)C)=w(B)+w(a)w(e)<w(B)w(F)=w((D+a)\cup C)=w(B)+w(a)-w(e)<w(B) (13)

By (13), we have that FF is a basis of {\mathcal{M}} satisfying w(F)<w(B)w(F)<w(B). Contradiction (to the minimality of BB w.r.t. ww). ∎

In the next lemma we consider a simple property of bases of union matroids.

Lemma 4.3.

Let M1=(E1,1),,Mk=(Ek,k)M_{1}=(E_{1},{\mathcal{I}}_{1}),\ldots,M_{k}=(E_{k},{\mathcal{I}}_{k}) be matroids such that EiEj=i,j[k],ijE_{i}\cap E_{j}=\emptyset~{}\forall i,j\in[k],i\neq j; also, let w:E0w:E\rightarrow\mathbb{R}_{\geq 0} be a weight function, and RR a minimum basis of i[k]Mi\bigvee_{i\in[k]}M_{i} w.r.t. ww. Then, for all i[k]i\in[k], REiR\cap E_{i} is a minimum basis of MiM_{i} w.r.t. ww.111With a slight abuse of notation we use ww also for the restriction of ww on EiE_{i}.

Proof.

Let i[k]i\in[k]. For all eEiRe\in E_{i}\setminus R, it holds that REi+eR\cap E_{i}+e is not an independent set of MiM_{i} since otherwise R+eR+e is an independent set of i[k]Mi\bigvee_{i\in[k]}M_{i} by Definition 2.2, contradicting that RR is a basis of i[k]Mi\bigvee_{i\in[k]}M_{i}; we conclude that REiR\cap E_{i} is a basis of MiM_{i}. Assume towards contradiction that there is a basis BB of the matroid MiM_{i} such that w(B)<w(REi)w(B)<w(R\cap E_{i}). As EiEj=i,j[k],ijE_{i}\cap E_{j}=\emptyset~{}\forall i,j\in[k],i\neq j, by Definition 2.2 it follows that (REi)B\left(R\setminus E_{i}\right)\cup B is a basis of the matroid i[k]Mi\bigvee_{i\in[k]}M_{i}. In addition,

w((REi)B)=w(R)w(REi)+w(B)<w(R).w(\left(R\setminus E_{i}\right)\cup B)=w(R)-w(R\cap E_{i})+w(B)<w(R).

We reach a contradiction since RR is a minimum basis w.r.t. ww of i[k]Mi\bigvee_{i\in[k]}M_{i}. ∎

We now prove Lemma 3.8 using several auxiliary lemmas. For the remainder of this section, let 𝒦=(E,,c,p,β){\mathcal{K}}=(E,{\mathcal{I}},c,p,\beta) be a BMI instance, 0<ε<120<{\varepsilon}<\frac{1}{2}, R=FindRep(𝒦,ε)R=\textsf{FindRep}({\mathcal{K}},{\varepsilon}), and α\alpha the value from Step 1 of Algorithm 1. For the next lemma, recall the sets Cr(α)C_{r}(\alpha) for r[log1ε(ε2)+1]r\in[\log_{1-{\varepsilon}}\left(\frac{{\varepsilon}}{2}\right)+1] were defined in (5).

Lemma 4.4.

For all r[log1ε(ε2)+1]r\in[\log_{1-{\varepsilon}}\left(\frac{{\varepsilon}}{2}\right)+1] it holds that RCr(α)R\cap C_{r}(\alpha) is a minimum basis w.r.t. cc of the matroid =[(E,)Cr(α)]q(ε){\mathcal{M}}=\left[(E,{\mathcal{I}})\cap C_{r}(\alpha)\right]_{\leq q({\varepsilon})}, where R=FindRep(𝒦,ε)R=\textsf{FindRep}({\mathcal{K}},{\varepsilon}).

Proof.

Observe that for all r,t[log1ε(ε2)+1]r,t\in[\log_{1-{\varepsilon}}\left(\frac{{\varepsilon}}{2}\right)+1] it holds that Cr(α)Ct(α)=C_{r}(\alpha)\cap C_{t}(\alpha)=\emptyset by (5). Hence, the statement of the lemma follows from Step 1 of Algorithm 1 and Lemma 4.3. ∎

Lemma 4.5.

For all eHe\in H there is exactly one r[log1ε(ε2)+1]r\in[\log_{1-{\varepsilon}}\left(\frac{{\varepsilon}}{2}\right)+1] such that eCr(α)e\in C_{r}(\alpha).

Proof.

Let eHe\in H. Observe that:

ε2p(e)2OPT(𝒦)p(e)2αp(e)OPT(𝒦)1.\frac{{\varepsilon}}{2}\leq\frac{p(e)}{2\cdot\textnormal{OPT}({\mathcal{K}})}\leq\frac{p(e)}{2\cdot\alpha}\leq\frac{p(e)}{\textnormal{OPT}({\mathcal{K}})}\leq 1. (14)

The first inequality holds since eHe\in H. The second and third inequalities are because α\alpha is the value of a 22-approximation for the optimum of 𝒦{\mathcal{K}}; thus, OPT(𝒦)2αOPT(𝒦)\frac{\textnormal{OPT}({\mathcal{K}})}{2}\leq\alpha\leq\textnormal{OPT}({\mathcal{K}}). The last inequality is because {e}\{e\} is a solution of 𝒦{\mathcal{K}}. In addition, observe that for r0=log1εε2r_{0}={\left\lceil\log_{1-{\varepsilon}}\frac{{\varepsilon}}{2}\right\rceil} it holds that (1ε)r0ε2(1-{\varepsilon})^{r_{0}}\leq\frac{{\varepsilon}}{2} and that for r1=1r_{1}=1 it holds that (1ε)r11=1(1-{\varepsilon})^{r_{1}-1}=1. Hence, because r0log1ε(ε2)+1r_{0}\leq\left\lfloor\log_{1-{\varepsilon}}\left(\frac{{\varepsilon}}{2}\right)+1\right\rfloor, by (14), there is exactly one r[log1ε(ε2)+1]r\in[\log_{1-{\varepsilon}}\left(\frac{{\varepsilon}}{2}\right)+1] such that p(e)2α((1ε)r,(1ε)r1]\frac{p(e)}{2\alpha}\in\big{(}(1-{\varepsilon})^{r},(1-{\varepsilon})^{r-1}\big{]}; thus, by (5) it holds that eCr(α)e\in C_{r}(\alpha) and eCr(α)e\notin C_{r^{\prime}}(\alpha) for r[log1ε(ε2)+1]{r}r^{\prime}\in[\log_{1-{\varepsilon}}\left(\frac{{\varepsilon}}{2}\right)+1]\setminus\{r\}. ∎

For the proof of Lemma 3.8, we define a substitution of some independent set. The first two properties of substitution of some Gq(ε)G\in{\mathcal{I}}_{\leq q({\varepsilon})} are identical to the first two properties in the definition of a replacement of GG. However, we require that a substitution preserves the number of profitable elements in GG from each profit class, and that a substitution must be disjoint to the set of non-profitable elements in GG.

Definition 4.6.

For Gq(ε)G\in{\mathcal{I}}_{\leq q({\varepsilon})} and ZGr[log1ε(ε2)+1]Cr(α)Z_{G}\subseteq\bigcup_{r\in[\log_{1-{\varepsilon}}\left(\frac{{\varepsilon}}{2}\right)+1]}C_{r}(\alpha), we say that ZGZ_{G} is a substitution of GG if the following holds.

  1. 1.

    (GH)ZGq(ε)(G\setminus H)\cup Z_{G}\in{\mathcal{I}}_{\leq q({\varepsilon})}.

  2. 2.

    c(ZG)c(GH)c(Z_{G})\leq c(G\cap H).

  3. 3.

    For all r[log1ε(ε2)+1]r\in[\log_{1-{\varepsilon}}\left(\frac{{\varepsilon}}{2}\right)+1] it holds that |Cr(α)ZG|=|Cr(α)GH||C_{r}(\alpha)\cap Z_{G}|=|C_{r}(\alpha)\cap G\cap H|.

  4. 4.

    (GH)ZG=(G\setminus H)\cap Z_{G}=\emptyset.

Lemma 4.7.

For all Gq(ε)G\in{\mathcal{I}}_{\leq q({\varepsilon})} there is a substitution ZGZ_{G} of GG such that ZGRZ_{G}\subseteq R.

In the proof of Lemma 4.7, we assume towards a contradiction that there is no substitution for some Gq(ε)G\in{\mathcal{I}}_{\leq q({\varepsilon})} which is a subset of RR. To reach a contradiction, we take a substitution ZGZ_{G} of GG with maximal number of elements from RR and show that using Lemma 4.1 and Lemma 4.2 we can create a substitution with more elements from RR by replacing an element from ZGRZ_{G}\setminus R by an element from RZGR\setminus Z_{G}.

Proof of Lemma 4.7.

Let Gq(ε)G\in{\mathcal{I}}_{\leq q({\varepsilon})} and let ZGZ_{G} be a substitution of GG such that |ZGR||Z_{G}\cap R| is maximal among all substitutions of GG; formally, let 𝒮(G)\mathcal{S}(G) be all substitutions of GG and let ZG{Z𝒮(G)||ZR|=maxZ𝒮(G)|ZR|}Z_{G}\in\{Z\in\mathcal{S}(G)~{}|~{}|Z\cap R|=\max_{Z^{\prime}\in\mathcal{S}(G)}|Z^{\prime}\cap R|\}. Since GHG\cap H is in particular a substitution of GG it follows that 𝒮(G)\mathcal{S}(G)\neq\emptyset, and thus ZGZ_{G} is well defined. Assume towards a contradiction that there is aZGRa\in Z_{G}\setminus R; then, by Definition 4.6 there is r[log1ε(ε2)+1]r\in[\log_{1-{\varepsilon}}\left(\frac{{\varepsilon}}{2}\right)+1] such that aCr(α)a\in C_{r}(\alpha). Let ΔG=(GH)ZG\Delta_{G}=(G\setminus H)\cup Z_{G}.

Claim 4.8.

There is b(Cr(α)R)ΔGb\in(C_{r}(\alpha)\cap R)\setminus\Delta_{G} such that c(b)c(a)c(b)\leq c(a) and ΔGa+bq(ε)\Delta_{G}-a+b\in{\mathcal{I}}_{\leq q({\varepsilon})}.

Proof.

Let (Cr(α),)=[(E,)Cr(α)]q(ε)(C_{r}(\alpha),{\mathcal{I}}^{\prime})=\left[(E,{\mathcal{I}})\cap C_{r}(\alpha)\right]_{\leq q({\varepsilon})} and =(Cr(α),){\mathcal{M}}=(C_{r}(\alpha),{\mathcal{I}}^{\prime}). By Lemma 4.4 it holds that RCr(α)R\cap C_{r}(\alpha) is a minimum basis w.r.t. cc of {\mathcal{M}}. Define D={eCr(α)R|c(e)c(a)}D=\{e\in C_{r}(\alpha)\cap R~{}|~{}c(e)\leq c(a)\}. Then, since aCr(α)Ra\in C_{r}(\alpha)\setminus R, it holds that D+aD+a\notin{\mathcal{I}}^{\prime} by Lemma 4.2. In addition, by Definition 2.2 it holds that ={ACr(α)|Aq(ε)}{\mathcal{I}}^{\prime}=\{A\subseteq C_{r}(\alpha)~{}|~{}A\in{\mathcal{I}}_{\leq q({\varepsilon})}\}. Hence, D+aq(ε)D+a\notin{\mathcal{I}}_{\leq q({\varepsilon})}. Therefore, by Lemma 4.1 there is bDΔGb\in D\setminus\Delta_{G} such that ΔGa+bq(ε)\Delta_{G}-a+b\in{\mathcal{I}}_{\leq q({\varepsilon})}. The claim follows since c(b)c(a)c(b)\leq c(a) because bDb\in D.

\square

Using Claim 4.8, let bCr(α)RΔGb\in C_{r}(\alpha)\cap R\setminus\Delta_{G} such that c(b)c(a)c(b)\leq c(a) and ΔGa+bq(ε)\Delta_{G}-a+b\in{\mathcal{I}}_{\leq q({\varepsilon})}. Then, the properties of Definition 4.6 are satisfied for ZGa+bZ_{G}-a+b by the following.

  1. 1.

    (GH)(ZGa+b)=ΔGa+bq(ε)(G\setminus H)\cup(Z_{G}-a+b)=\Delta_{G}-a+b\in{\mathcal{I}}_{\leq q({\varepsilon})} by the definition of bb.

  2. 2.

    c(ZGa+b)c(ZG)c(GH)c(Z_{G}-a+b)\leq c(Z_{G})\leq c(G\cap H) because c(b)c(a)c(b)\leq c(a).

  3. 3.

    for all r[log1ε(ε2)+1]r^{\prime}\in[\log_{1-{\varepsilon}}\left(\frac{{\varepsilon}}{2}\right)+1] it holds that |Cr(α)(ZGa+b)|=|Cr(α)ZG|=|Cr(α)GH||C_{r^{\prime}}(\alpha)\cap(Z_{G}-a+b)|=|C_{r^{\prime}}(\alpha)\cap Z_{G}|=|C_{r^{\prime}}(\alpha)\cap G\cap H| because a,bCr(α)a,b\in C_{r}(\alpha).

  4. 4.

    |(GH)(ZGa+b)||(GH)(ZG)|=0|(G\setminus H)\cap(Z_{G}-a+b)|\leq|(G\setminus H)\cap(Z_{G})|=0 where the inequality follows because bΔGb\notin\Delta_{G} and the equality is since ZGZ_{G} is a substitution of GG.

By the above and Definition 4.6, it follows that ZG+abZ_{G}+a-b is a substitution of GG; that is, ZG+ab𝒮(G)Z_{G}+a-b\in\mathcal{S}(G). Moreover,

|R(ZGa+b)|>|RZG|=maxZ𝒮(G)|ZR|.|R\cap(Z_{G}-a+b)|>|R\cap Z_{G}|=\max_{Z\in\mathcal{S}(G)}|Z\cap R|. (15)

The first inequality is because aZGRa\in Z_{G}\setminus R and bRb\in R. By (15) we reach a contradiction since we found a replacement of GG with more elements in RR than ZG𝒮(G)Z_{G}\in\mathcal{S}(G), which is defined as a replacement of GG with maximal number of elements in RR. Therefore, ZGRZ_{G}\subseteq R as required. ∎

We are now ready to prove Lemma 3.8. The proof follows by showing that for any Gq(ε)G\in{\mathcal{I}}_{\leq q({\varepsilon})} a substitution of GG which is a subset of RR is in fact a replacement of GG.

Proof of Lemma 3.8: Let Gq(ε)G\in{\mathcal{I}}_{\leq q({\varepsilon})}. By Lemma 4.7, GG has a substitution ZGRZ_{G}\subseteq R. Then,

p(ZG)\displaystyle p(Z_{G})\geq{} r[log1ε(ε2)+1]p(Cr(α)ZG)\displaystyle\sum_{r\in[\log_{1-{\varepsilon}}\left(\frac{{\varepsilon}}{2}\right)+1]}p(C_{r}(\alpha)\cap Z_{G}) (16)
\displaystyle\geq{} r[log1ε(ε2)+1] s.t. Cr(α)|Cr(α)ZG|mineCr(α)p(e)\displaystyle\sum_{r\in[\log_{1-{\varepsilon}}\left(\frac{{\varepsilon}}{2}\right)+1]\text{ s.t. }C_{r}(\alpha)\neq\emptyset}|C_{r}(\alpha)\cap Z_{G}|\cdot\min_{e\in C_{r}(\alpha)}p(e)
\displaystyle\geq{} r[log1ε(ε2)+1] s.t. Cr(α)|Cr(α)GH|(1ε)maxeCr(α)p(e)\displaystyle\sum_{r\in[\log_{1-{\varepsilon}}\left(\frac{{\varepsilon}}{2}\right)+1]\text{ s.t. }C_{r}(\alpha)\neq\emptyset}|C_{r}(\alpha)\cap G\cap H|\cdot(1-{\varepsilon})\cdot\max_{e\in C_{r}(\alpha)}p(e)
\displaystyle\geq{} (1ε)p(GH).\displaystyle(1-{\varepsilon})\cdot p(G\cap H).

The third inequality is by (5) and Property 3 of Definition 4.6. The last inequality follows from Lemma 4.5. Therefore,

p((GH)ZG)\displaystyle p((G\setminus H)\cup Z_{G}) =p(GH)+p(ZG)\displaystyle=p(G\setminus H)+p(Z_{G}) (17)
p(GH)+(1ε)p(GH)\displaystyle\geq p(G\setminus H)+(1-{\varepsilon})\cdot p(G\cap H)
(1ε)p(G).\displaystyle\geq(1-{\varepsilon})\cdot p(G).

The first equality follows by Property 4 of Definition 4.6. The first inequality is by (16). Now, ZGZ_{G} satisfies Property 1 and 2 of Definition 3.1 by Properties 12 of Definition 4.6, respectively. In addition, ZGZ_{G} satisfies Property 3 of Definition 3.1 by (17). Finally, ZGZ_{G} satisfies Property 4 of Definition 3.1 by

|ZG|=r[log1ε(ε2)+1]|ZGCr(α)|r[log1ε(ε2)+1]|GHCr(α)|=|GH|.|Z_{G}|=\sum_{r\in[\log_{1-{\varepsilon}}\left(\frac{{\varepsilon}}{2}\right)+1]}|Z_{G}\cap C_{r}(\alpha)|\leq\sum_{r\in[\log_{1-{\varepsilon}}\left(\frac{{\varepsilon}}{2}\right)+1]}|G\cap H\cap C_{r}(\alpha)|=|G\cap H|.

The first inequality holds since ZGZ_{G} is a substitution of GG. The last equality follows from Lemma 4.5. We conclude that ZGZ_{G} is a replacement of GG such that ZGRZ_{G}\subseteq R; thus, RR is a representative set by Definition 3.2. By Step 1 of Algorithm 1 and Definition 2.2, it holds that |RCr(α)|q(ε)|R\cap C_{r}(\alpha)|\leq q({\varepsilon}); thus, it follows that |R|(log1ε(ε2)+1)q(ε)|R|\leq\left(\log_{1-{\varepsilon}}\left(\frac{{\varepsilon}}{2}\right)+1\right)\cdot q({\varepsilon}).

We now bound the time complexity of Algorithm FindRep. Step 1 can be computed in time poly(|𝒦|)\textnormal{poly}(|{\mathcal{K}}|) using a PTAS for BMI with an error parameter ε=12{\varepsilon}=\frac{1}{2} (see, e.g., [9]). In addition, a membership oracle for the matroid in Step 1 can be implemented in time poly(|𝒦|)\textnormal{poly}(|{\mathcal{K}}|) by Definition 2.2. Finally, the basis in Step 1 can be computed in time poly(|𝒦|)\textnormal{poly}(|{\mathcal{K}}|) using a greedy matroid basis minimization algorithm [5]. Hence, the running time of the algorithm is poly(|𝒦|)\textnormal{poly}(|{\mathcal{K}}|). ∎

5 Discussion

In this paper we showed that the budgeted matroid independent set problem admits an EPTAS, thus improving upon the previously known schemes for the problem. The speed-up is achieved by replacing the exhaustive enumeration used by existing algorithms [9, 1, 4, 7] with efficient enumeration over subsets of a representative set whose size depends only on 1/ε1/{\varepsilon}.

The representative set found by our algorithm is a minimum cost matroid basis for a union matroid. The union matroid is a union of a matroid for each profit class in the given instance. The basis itself can be easily found using a simple greedy procedure. The correctness relies on matroid exchange properties, optimality properties of minimal cost bases and a “profit gap” obtained by discarding a subset of elements from an optimal solution.

We note that our EPTAS, which directly exploits structural properties of our problem, already achieves substantial improvement over schemes developed for generalizations of BMI. Furthemore, we almost resolve the complexity status of the problem w.r.t approximation schemes. The existence of an FPTAS remains open.

The notion of representative sets can potentially be used to replace exhaustive enumeration in the PTASs for multiple knapsack with matroid constraint [7] and budgeted matroid intersection [1], leading to EPTASs for both problems. It seems that the construction of a representative set, as well as the ideas behind the main lemmas, can be adapted to the setting of the multiple knapsack with matroid problem. However, to derive an EPTAS for budgted matroid intersection, one needs to generalize first the concept of representative set, so it can be applied to matroid intersection. We leave this generalization as an interesting direction for follow-up work.

References

  • [1] Berger, A., Bonifaci, V., Grandoni, F., Schäfer, G.: Budgeted matching and budgeted matroid intersection via the gasoline puzzle. Mathematical Programming 128(1), 355–372 (2011)
  • [2] Caprara, A., Kellerer, H., Pferschy, U., Pisinger, D.: Approximation algorithms for knapsack problems with cardinality constraints. European Journal of Operational Research 123(2), 333–345 (2000)
  • [3] Chekuri, C., Khanna, S.: A polynomial time approximation scheme for the multiple knapsack problem. SIAM Journal on Computing 35(3), 713–728 (2005)
  • [4] Chekuri, C., Vondrák, J., Zenklusen, R.: Multi-budgeted matchings and matroid intersection via dependent rounding. In: Proceedings of the twenty-second annual ACM-SIAM symposium on Discrete Algorithms. pp. 1080–1097. SIAM (2011)
  • [5] Cormen, T.H., Leiserson, C.E., Rivest, R.L., Stein, C.: Introduction to algorithms. MIT press (2022)
  • [6] Das, S., Wiese, A.: On minimizing the makespan when some jobs cannot be assigned on the same machine. In: 25th Annual European Symposium on Algorithms, ESA. pp. 31:1–31:14 (2017)
  • [7] Fairstein, Y., Kulik, A., Shachnai, H.: Modular and submodular optimization with multiple knapsack constraints via fractional grouping. In: 29th Annual European Symposium on Algorithms, ESA 2021, September 6-8, 2021, Lisbon, Portugal (Virtual Conference). pp. 41:1–41:16 (2021)
  • [8] Grage, K., Jansen, K., Klein, K.M.: An EPTAS for machine scheduling with bag-constraints. In: The 31st ACM Symposium on Parallelism in Algorithms and Architectures. pp. 135–144 (2019)
  • [9] Grandoni, F., Zenklusen, R.: Approximation schemes for multi-budgeted independence systems. In: European Symposium on Algorithms. pp. 536–548. Springer (2010)
  • [10] Grötschel, M., Lovász, L., Schrijver, A.: Geometric algorithms and combinatorial optimization, vol. 2. Springer Science & Business Media (1993)
  • [11] Heydrich, S., Wiese, A.: Faster approximation schemes for the two-dimensional knapsack problem. ACM Trans. Algorithms 15(4), 47:1–47:28 (2019)
  • [12] Ibarra, O.H., Kim, C.E.: Fast approximation algorithms for the knapsack and sum of subset problems. Journal of the ACM (JACM) 22(4), 463–468 (1975)
  • [13] Jansen, K.: Parameterized approximation scheme for the multiple knapsack problem. SIAM Journal on Computing 39(4), 1392–1412 (2010)
  • [14] Jansen, K.: A fast approximation scheme for the multiple knapsack problem. In: SOFSEM 2012: Theory and Practice of Computer Science - 38th Conference on Current Trends in Theory and Practice of Computer Science, Špindlerův Mlýn, Czech Republic, January 21-27, 2012. Proceedings. pp. 313–324 (2012])
  • [15] Jansen, K., Maack, M.: An EPTAS for scheduling on unrelated machines of few different types. Algorithmica 81(10), 4134–4164 (2019)
  • [16] Jansen, K., Sinnen, O., Wang, H.: An EPTAS for scheduling fork-join graphs with communication delay. Theor. Comput. Sci. 861, 66–79 (2021)
  • [17] Kellerer, H., Pferschy, U., Pisinger, D.: Knapsack problems. Springer (2004)
  • [18] Kulik, A., Shachnai, H.: There is no EPTAS for two-dimensional knapsack. Information Processing Letters 110(16), 707–710 (2010)
  • [19] Lawler, E.L.: Fast approximation algorithms for knapsack problems. Math. Oper. Res. 4(4), 339–356 (1979)
  • [20] Pisinger, D., Toth, P.: Knapsack problems. In: Handbook of combinatorial optimization, pp. 299–428 (1998)
  • [21] Schrijver, A.: Combinatorial optimization: polyhedra and efficiency, vol. 24. Springer (2003)
  • [22] Schuurman, P., Woeginger, G.J.: Approximation schemes-a tutorial. Lectures on Scheduling (to appear) (2001)
  • [23] Shachnai, H., Tamir, T.: Polynomial time approximation schemes. In: Handbook of Approximation Algorithms and Metaheuristics, Second Edition, Volume 1: Methologies and Traditional Applications, pp. 125–156 (2018)
  • [24] Woeginger, G.J.: When does a dynamic programming formulation guarantee the existence of an FPTAS? In: Proceedings of the Tenth Annual ACM-SIAM Symposium on Discrete Algorithms, 17-19 January 1999, Baltimore, Maryland, USA. pp. 820–829 (1999)

Appendix A Solving the Linear Program

In this section we show how LP(𝒦,F,α)\textnormal{LP}({\mathcal{K}},F,\alpha) can be solved in polynomial time, thus proving Lemma 3.10.

The main idea is to show that the feasibility domain of LP(𝒦,F,α)\textnormal{LP}({\mathcal{K}},F,\alpha) is a well described polytope, and that a separation oracle for it can be implemented in polynomial time. Thus, using a known connection between separation and optimization, we obtain a polynomial-time algorithm which solves LP(𝒦,F,α)\textnormal{LP}({\mathcal{K}},F,\alpha). Our initial goal, however, is to show that LP(𝒦,F,α)\textnormal{LP}({\mathcal{K}},F,\alpha) is indeed a linear program, for which we need a characterization of the matroid polytope via linear inequalities.

Let 𝒩=(Ω,𝒥){\mathcal{N}}=(\Omega,{\mathcal{J}}) be a matroid. The matroid rank function of 𝒩{\mathcal{N}}, 𝗋𝖺𝗇𝗄𝒩:2Ω\mathsf{rank}_{{\mathcal{N}}}:2^{\Omega}\to\mathbb{N}, is defined by 𝗋𝖺𝗇𝗄𝒩(S)=max{|T||T𝒥,TS}\mathsf{rank}_{{\mathcal{N}}}(S)=\max\left\{|T|~{}\middle|~{}T\in{\mathcal{J}},~{}T\subseteq S\right\}. That is, 𝗋𝖺𝗇𝗄𝒩(S)\mathsf{rank}_{{\mathcal{N}}}(S) is the maximal size of an independent set which is also a subset of SS. The rank function is used in a characterization of a matroid polytope P𝒩P_{{\mathcal{N}}} via a system of linear inequalities.

Theorem A.1 (Corollary 40.2b in [21]).

For any matroid 𝒩=(Ω,𝒥){\mathcal{N}}=(\Omega,{\mathcal{J}}), it holds that

P𝒩={x¯0|SΩ:ωSx¯ω𝗋𝖺𝗇𝗄𝒩(S)}.P_{{\mathcal{N}}}=\left\{{\bar{x}}\in\mathbb{R}_{\geq 0}~{}\middle|~{}\forall S\subseteq\Omega:~{}~{}\sum_{\omega\in S}{\bar{x}}_{\omega}\leq\mathsf{rank}_{{\mathcal{N}}}(S)\right\}.

Let 𝒦=(E,,c,p,β){\mathcal{K}}=(E,{\mathcal{I}},c,p,\beta) be a BMI instance, FEF\subseteq E and OPT(𝒦)2αOPT(𝒦)\frac{\textnormal{OPT}({\mathcal{K}})}{2}\leq\alpha\leq\textnormal{OPT}({\mathcal{K}}). By Theorem A.1 it holds that (6) is equivalent to the following linear program.

LP(𝒦,F,α)max\displaystyle\textnormal{LP}({\mathcal{K}},F,\alpha)~{}~{}~{}~{}~{}~{}~{}\max\quad~{}~{} eE(α)Fx¯ep(e)\displaystyle\sum_{e\in E(\alpha)\setminus F}\bar{x}_{e}\cdot p(e)
s.t. eE(α)Fx¯ec(e)βc(F)\displaystyle\sum_{e\in E(\alpha)\setminus F}\bar{x}_{e}\cdot c(e)\leq\beta-c(F)~{}~{}~{}~{}~{} (18)
eSx¯e𝗋𝖺𝗇𝗄F(α)(S)\displaystyle\sum_{e\in S}{\bar{x}}_{e}\leq\mathsf{rank}_{{\mathcal{M}}_{F}(\alpha)}(S) SE(α)F\displaystyle\forall S\subseteq E(\alpha)\setminus F (19)
x¯e0\displaystyle\bar{x}_{e}\geq 0 eE(α)F\displaystyle\forall e\in E(\alpha)\setminus F (20)

We follow the definitions and techniques presented in [10]. We say a polytope PnP\subseteq\mathbb{R}^{n} is of facet complexity φ\varphi if it is the solution set of a system of linear inequalities with rational coefficients, and the encoding length of each inequality in the system is at most φ\varphi. The following is an immediate consequence of the representation of LP(𝒦,F,α)\textnormal{LP}({\mathcal{K}},F,\alpha) as a linear program.

Observation A.2.

The feasibility region of LP(𝒦,F,α)\textnormal{LP}({\mathcal{K}},F,\alpha) is of facet complexity polynomial in |𝒦||{\mathcal{K}}|.

A separating hyperplane between a polytope PnP\subseteq\mathbb{R}^{n} and x¯n{\bar{x}}\in\mathbb{R}^{n} is vector c¯n{\bar{c}}\in\mathbb{R}^{n} such that c¯x¯>maxy¯Pc¯y¯{\bar{c}}\cdot{\bar{x}}>\max_{{\bar{y}}\in P}{\bar{c}}\cdot{\bar{y}}, where c¯y¯{\bar{c}}\cdot{\bar{y}} is the inner product of c¯{\bar{c}} and y¯{\bar{y}}. With a slight abuse of notation, we say that the constraint c¯z¯L{\bar{c}}\cdot{\bar{z}}\leq L, where c¯n{\bar{c}}\in\mathbb{R}^{n} and LL\in\mathbb{R}, is a separation hyperplane between PP and x¯{\bar{x}} if c¯x¯>L{\bar{c}}\cdot{\bar{x}}>L and c¯y¯L{\bar{c}}\cdot{\bar{y}}\leq L for every y¯P{\bar{y}}\in P. A separation oracle for a polytope PnP\subseteq\mathbb{R}^{n} is an oracle which receives x¯n{\bar{x}}\in\mathbb{R}^{n} as input, and either determines that x¯P{\bar{x}}\in P or returns a separating hyperplane between PP and x¯{\bar{x}}. We use the following known connection between separation and optimization.

Theorem A.3 (Thm 6.4.9 and Remark 6.5.2 in [10]).

There is an algorithm which given n,φn,\varphi\in\mathbb{N}, c¯n{\bar{c}}\in\mathbb{R}^{n} and a separation oracle for a non-empty polytope PnP\subseteq\mathbb{R}^{n} of facet complexity φ\varphi, returns a vertex y¯{\bar{y}} of PP such that c¯y¯=maxx¯Pc¯x¯{\bar{c}}\cdot{\bar{y}}=\max_{{\bar{x}}\in P}{\bar{c}}\cdot{\bar{x}} in time polynomial in nn and φ\varphi.

Thus, to show LP(𝒦,F,α)\textnormal{LP}({\mathcal{K}},F,\alpha) can be solved in polynomial time, we need to show that a separation oracle for LP(𝒦,F,α)\textnormal{LP}({\mathcal{K}},F,\alpha) can be implemented in polymomial time. To this end, we use a known result for matroid polytopes.

Theorem A.4 (Thm 40.4 in [21]).

There is a polynomial time algorithm MatroidSeparator which given a subset of elements Ω\Omega, a membership oracle for a matroid 𝒩=(Ω,𝒥){\mathcal{N}}=(\Omega,{\mathcal{J}}) and x¯0Ω{\bar{x}}\in\mathbb{Q}_{\geq 0}^{\Omega} either determines that x¯P𝒩{\bar{x}}\in P_{{\mathcal{N}}} or returns SΩS\subseteq\Omega such that ωSx¯ω>𝗋𝖺𝗇𝗄𝒩(S)\sum_{\omega\in S}{\bar{x}}_{\omega}>\mathsf{rank}_{{\mathcal{N}}}(S).

A polynomial-time implementation of a separation oracle for LP(𝒦,F,α)\textnormal{LP}({\mathcal{K}},F,\alpha) is now straightforward.

Lemma A.5.

There is an algorithm which given a BMI instance 𝒦=(E,,c,p,β){\mathcal{K}}=(E,{\mathcal{I}},c,p,\beta), FEF\subseteq E, OPT(𝒦)2αOPT(𝒦)\frac{\textnormal{OPT}({\mathcal{K}})}{2}\leq\alpha\leq\textnormal{OPT}({\mathcal{K}}) and x¯E{\bar{x}}\in\mathbb{R}^{E} implements a separation oracle for the feasibility region of LP(𝒦,F,α)\textnormal{LP}({\mathcal{K}},F,\alpha) in polynomial time.

Proof.

To implement a separation oracle, the algorithm first checks if the input x¯{\bar{x}} satisfies constraints (18) and (20). If one of the constraints is violated, the algorithm returns the constraint as a separating hyperplane.

Next, the algorithm invokes MatroidSeparator (Theorem A.4) with the matroid F(α){\mathcal{M}}_{F}(\alpha) and the point x¯{\bar{x}}. If MatroidSeparator returns that x¯PF(α){\bar{x}}\in P_{{\mathcal{M}}_{F}(\alpha)} then the algorithm returns that x¯{\bar{x}} is in the feasibility region. Otherwise, the MatroidSeparator returns SE(α)FS\subseteq E(\alpha)\setminus F such that eSx¯e>𝗋𝖺𝗇𝗄F(α)(S)\sum_{e\in S}{\bar{x}}_{e}>\mathsf{rank}_{{\mathcal{M}}_{F}(\alpha)}(S); that is, a set SS for which (19) is violated. In this case, the algorithm returns 𝟙S\mathbbm{1}^{S} as a separating hyperplane. Observe that for every y¯{\bar{y}} in the feasibility region of LP(𝒦,F,α)\textnormal{LP}({\mathcal{K}},F,\alpha), it holds that

𝟙Sy¯=eSy¯e𝗋𝖺𝗇𝗄F(α)(S)<𝟙Sx¯,\mathbbm{1}^{S}\cdot{\bar{y}}=\sum_{e\in S}{\bar{y}}_{e}\leq\mathsf{rank}_{{\mathcal{M}}_{F}(\alpha)}(S)<\mathbbm{1}^{S}\cdot{\bar{x}},

where the first inequality is by (19). Thus, 𝟙S\mathbbm{1}^{S} is indeed a separating hyperplane.

Clearly, the separating hyperplanes returned by the algorithm are indeed separating hyperplanes. Furthermore, if the algorithm asserts that x¯{\bar{x}} is in the feasibility region then constraints (18) and (20) hold as those were explicitly checked, and (19) holds by x¯PF(α){\bar{x}}\in P_{{\mathcal{M}}_{F}(\alpha)} (Theorem A.4 and Theorem A.1). That is, x¯{\bar{x}} is indeed in the feasibility region of LP(𝒦,F,α)\textnormal{LP}({\mathcal{K}},F,\alpha). The algorithm runs in polynomial-time as each of its steps can be implemented in polynomial-time. ∎

Proof of Lemma 3.10.

Observe that the vector (0,,0)E(α)(0,\ldots,0)\in\mathbb{R}^{E(\alpha)} is in the feasibility region of LP(𝒦,F,α)\textnormal{LP}({\mathcal{K}},F,\alpha), and thus the feasibility region is not empty. By Theorem A.3 and Observation A.2, there is an algorithm which finds an optimal basic solution for LP(𝒦,F,α)\textnormal{LP}({\mathcal{K}},F,\alpha) using a polynomial number of operations and calls for a separation oracle. By Lemma A.5, the separation oracle can be implemented in polynomial time as well. Thus, an optimal basic solution for LP(𝒦,F,α)\textnormal{LP}({\mathcal{K}},F,\alpha) can be found in time polynomial in |𝒦||{\mathcal{K}}|. ∎