Structural Iterative Rounding for Generalized -Median Problems
Abstract.
This paper considers approximation algorithms for generalized -median problems. This class of problems can be informally described as -median with a constant number of extra constraints, and includes -median with outliers, and knapsack median. Our first contribution is a pseudo-approximation algorithm for generalized -median that outputs a -approximate solution, with a constant number of fractional variables. The algorithm builds on the iterative rounding framework introduced by Krishnaswamy, Li, and Sandeep for -median with outliers. The main technical innovation is allowing richer constraint sets in the iterative rounding and taking advantage of the structure of the resulting extreme points.
Using our pseudo-approximation algorithm, we give improved approximation algorithms for -median with outliers and knapsack median. This involves combining our pseudo-approximation with pre- and post-processing steps to round a constant number of fractional variables at a small increase in cost. Our algorithms achieve approximation ratios and for -median with outliers and knapsack median, respectively. These improve on the best-known approximation ratio for both problems [KLS18].
1. Introduction
Clustering is a fundamental problem in combinatorial optimization, where we wish to partition a set of data points into clusters such that points within the same cluster are more similar than points across different clusters. In this paper, we focus on generalizations of the -median problem. Recall that in this problem, we are given a set of facilities, a set of clients, a metric on , and a parameter . The goal is to choose a set of facilities to open to minimize the sum of connection costs of each client to its closest open facility. That is, to minimize the objective , where we define .
The -median problem is well-studied from the perspective of approximation algorithms, and many new algorithmic techniques have been discovered while studying it. Examples include linear program rounding [BPR+17, LS16], primal-dual algorithms [JV01], local search [AGK+04], and large data techniques [LG18, MKC+15, GLZ17, GMM+03, IQM+20]. Currently, the best approximation ratio for -median is [BPR+17], and there is a lower bound of assuming [JMS02].
Recently, there has been significant interest in generalizations of the -median problem [CKMN01, KKN+15]. One such generalization is the knapsack median problem. In knapsack median, each facility has a non-negative weight, and we are given budget . The goal is to choose a set of open facilities of total weight at most (instead of having cardinality at most ) to minimize the same objective function. That is, the open facilities must satisfy a knapsack constraint. Another commonly-studied generalization is -median with outliers, also known as robust -median. Here we open facilities , as in basic -median, but we no longer have to serve all the clients; now, we are only required to serve at least clients of our choice. Formally, the objective function is now .
Both knapsack median and -median with outliers have proven to be much more difficult than the standard -median problem. Algorithmic techniques that have been successful in approximating -median often lead to only a pseudo-approximation for these generalizations—that is, they violate the knapsack constraint or serve fewer than clients [BPR+18, CKMN01, FKRS19, IQM+20]. Obtaining “true” approximation algorithms requires new ideas beyond those of -median. Currently the best approximation ratio for both problems is due to the beautiful iterative rounding framework of Krishnaswamy, Li, and Sandeep [KLS18]. The first and only other true approximation for -median with outliers is a local search algorithm due to Ke Chen [Che08].
1.1. Generalized -median
Observe that both knapsack median and -median with outliers maintain the salient features of -median; that is, the goal is to open facilities to minimize the connection costs of served clients. These variants differ in the way we put constraints on the open facilities and served clients. In particular, in standard -median, we have a cardinality constraint on the open facilities, whereas for knapsack median the open facilities are subject to a knapsack constraint; in both cases we must serve all clients. For -median with outliers, we are constrained to open at most facilities, and serve at least clients.
In this paper, we consider a further generalization of -median that we call generalized -median (GKM). As in -median, our goal is to open facilities to minimize the connection costs of served clients. In GKM, the open facilities must satisfy given knapsack constraints, and the served clients must satisfy given coverage constraints. We define .
1.2. Our Results
The main contribution of this paper is a refined iterative rounding algorithm for GKM. Specifically, we show how to round the natural linear program (LP) relaxation of GKM to ensure all except of the variables are integral, and the objective function is increased by at most a -factor. It is not difficult to show that the iterative rounding framework in [KLS18] can be extended to show a similar result. Indeed, a -approximation for GKM with at most fractional facilities is implicit in their work. The improvement in this work is the smaller loss in the objective value.
Our improvement relies on analyzing the extreme points of certain set-cover-like LPs. These extreme points arise at the intermediate steps of our iterative rounding, and by leveraging their structural properties, we obtain our improved pseudo-approximation for GKM. This work reveals some of the structure of such extreme points, and it shows how this structure can lead to improvements.
Our second contribution is improved “true” approximation algorithms for two special cases of GKM: knapsack median and -median with outliers. For both problems, applying the pseudo-approximation algorithm for GKM gives a solution with fractional facilities. Thus, the remaining work is to round a constant number of fractional facilities to obtain an integral solution. To achieve this goal, we apply known sparsification techniques [KLS18] to pre-process the instance, and then develop new post-processing algorithms to round the final fractional facilities.
We show how to round these remaining variables for knapsack median at arbitrarily small loss, giving a -approximation, improving on the best -approximation. For -median with outliers, a more sophisticated post-processing is needed to round the fractional facilities. This procedure loses more in the approximation ratio. In the end, we obtain a -approximation, modestly improving on the best known -approximation.
1.3. Overview of Techniques
To illustrate our techniques, we first introduce a natural LP relaxations for GKM. The problem admits an integer program formulation, with variables and , where indicates that client connects to facility and indicates that facility is open. Relaxing the integrality constraints gives the linear program relaxation .
We focus on for now. The linear program is the standard -median LP with the extra side constraints. Note that may seem opposite to the intuition that we want clients to get “enough” coverage from the facilities, but that will be guaranteed by the coverage constraints below.
The constraint corresponds to the knapsack constraints on the facilities , where and . These packing constraints can be thought of as a multidimensional knapsack constraint over the facilities, and ensure that “few” facilities are opened. Next, corresponds to the coverage constraints on the clients, where for all and . These coverage constraints ensure that “enough” clients are served. E.g., having one packing constraint and one covering constraint ensures that at least clients are covered by at most facilities; this is the -median with outliers problem.
Reducing the variables in the LP: We get by eliminating the variables from , thereby reducing the number of constraints. The idea from [KLS18] is to prescribe a set of permissible facilities for each client such that is implicitly set to . The details of this reduction and the procedure for creating are given in 2.1. Using this procedure, is also a relaxation for GKM. Note that in , we use the notation for .
Now consider solving to obtain an optimal extreme point . There must be linearly independent tight constraints at , and we call these constraints the basis for . The tight constraints of interest are the constraints; in general, there are at most such tight constraints, and we have little structural understanding of the -sets.
Prior Iterative Rounding Framework: Consider the family of sets corresponding to tight constraints, so . If is a family of disjoint sets , then the tight constraints of form a face of a partition matroid polytope intersected with at most side constraints (the knapsack and coverage constraints). Using ideas from, e.g., [KLS18, GRSZ14], we can show that has at most fractional variables.
Indeed, the goal of the iterative rounding framework in [KLS18] is to control the set family to obtain an optimal extreme point where is a disjoint family. To achieve this goal, they iteratively round an auxiliary LP based on , where they have the constraint for all clients in a special set . Roughly, they regulate what clients are added to and delete constraints for some clients. The idea is that a client whose constraint is deleted must be close to some client in . Since we can serve with the facility for , and the cost is small if ’s facility is close to .
To get intuition, assume each client can pay the farthest distance to a facility in , and call this the radius of . (Precisely, clients may not be able to afford this distance, but we use this assumption to highlight the ideas behind our algorithmic decisions.) For simplicity, assume all radii are powers of two. Over time, this radius shrinks if some variables in are set to zero. Consider applying the following iterative steps until none are applicable, in which case corresponds to the tight constraints: (1) delete a constraint for if the radius of is at least that of some for and . (2) add to if and for every such that it is the case that has a radius strictly larger than . If added then remove all from where ’s radius is half or less of the radius of and .
The approximation ratio is bounded by how much a client with a deleted constraint pays to get to a facility serving a client in . After removing ’s constraint, the case to worry about is if ’s closest client is later removed from . This happens only if is added to , with having half the radius of . Thus every time we remove ’s closest client in , we guarantee that ’s cost only increases geometrically. The approximation ratio is proportional to the total distance that must travel and can be directly related to the distance of “ball-chasing” though these sets. See Figure 1.

New Framework via Structured Extreme Points: The target of our framework is to ensure that the radii decreases in the ball-chasing using a smaller factor, in particular one-quarter. This will give closer facilities for clients whose constraints are deleted and a better approximation ratio. See Figure 1. To achieve this ”quarter ball-chasing,” we can simply change half to one-quarter in step (2) above.
Making this change immediately decreases the approximation ratio; however, the challenge is that is no longer disjoint. Indeed, it can be the case that such that if their radii differ by only a one half factor. Instead, our quarter ball-chasing algorithm maintains that is not disjoint, but has a bipartite intersection graph.
The main technical challenge now is obtaining an extreme point with fractional variables, which is no longer guaranteed as when was disjoint. Indeed, if has bipartite intersection graph, then the tight constraints form a face of the intersection of two partition matroid polytopes intersected with at most side constraints. In general, we cannot upper bound the number of fractional variables arising in the extreme points of such polytopes. However, such extreme points have a nice combinatorial structure: the intersection graph can be decomposed into disjoint paths. We exploit this “chain decomposition” of extreme points arising in our iterative rounding to discover clients that can be removed from even if there is not a where has one quarter of the radius of . We continue this procedure until we are left with only fractional variables.
The main technical contribution of this work is showing how the problem can be reduced to structural characterization of extreme points corresponding to bipartite matching. This illustrates some of the structural properties of polytopes defined by -median-type problems. We hope that this helps lead to other structural characterizations of these polytopes and ultimately improved algorithms.
1.4. Organization
In §2, we introduce the auxiliary LP for GKM that our iterative rounding algorithm operates on. We note that this is the same LP used in the algorithm of [KLS18]. Then §3–5 give the pseudo-approximation for GKM. In particular, §3 describes the basic iterative rounding phase, where we iteratively update the auxiliary LP such that has a bipartite intersection graph. In §4, we characterize the structure of the resulting extreme points and use it to define a new iterative operation, which allows us to reduce the number of fractional variables to . Finally, in §5, we combine the algorithms from §3 and §4 to obtain our pseduo-approximation algorithm for GKM.
We then obtain true approximations for knapsack median and -median with outliers: in §6, we describe our framework to turn pseudo-approximation algorithms into true approximations for both problems, and apply it to knapsack median. Then in §7, we give a more involved application of the same framework to -median with outliers.
2. Auxiliary LP for Iterative Rounding
In this section, we construct the auxiliary LP, , that our algorithm will use. We note that we use the same relaxation used in [KLS18]. Recall the two goals of iterative rounding, outlined in §1.3; we want to maintain a set of clients such that has bipartite intersection graph, and should provide a good set of open facilities for the clients that are not in . Thus, we want to define to accommodate moving clients in and out of , while having the LP faithfully capture how much we think the clients outside of should pay in connection costs. For all missing proofs in this section, see §A.
2.1. Defining -balls
Our starting point is , so we assume that we have sets for all . The next proposition states that such sets can be found efficiently so that is a relaxation of GKM.
Proposition 2.1.
There exists a polynomial time algorithm that given GKM instance , duplicates facilities and outputs sets for such that .
In §1.3, we assumed the radii of the sets were powers of two. To formalize this idea, we discretize the distances to powers of (up to some random offset.) The choice of is to optimize the final approximation ratio. The main ideas of the algorithm remain the same if we discretize to powers of, say , with no random offset. Our discretization procedure is the following:
Fix some and sample the random offset such that is uniformly distributed in . Without loss of generality, we may assume that the smallest non-zero inter-point distance is . Then we define the possible discretized distances, for all .
For each , we round up to the next largest discretized distance. Let denote the rounded distances. Observe that for all . See §A for proof of the following proposition, which we use to bound the cost of discretization.
Proposition 2.2.
For all , we have
Now using the discretized distances, we can define the radius level of for all by:
One should imagine that is a ball of radius in terms of the -distances. Thus, we will often refer to as the -ball of client . Further, to accommodate “shrinking” the sets, we define the inner ball of by:
Note that we defined so that if , then .
2.2. Constructing
Our auxiliary LP will maintain three sets of clients: , and . consists of all clients, whom we have not yet decided whether we should serve them or not. Then for all clients in and , we decide to serve them fully. The difference between the clients in and is that for the former, we remove the constraint from the LP, while for the latter we still require . Thus although we commit to serving , such clients rely on to find an open facility to connect to. Using the discretized distances, radius levels, inner balls, and these three sets of clients, we are ready to define :
() | ||||
s.t. | ||||
Note that we use the rounded distances in the definition of rather than the original distances. Keeping this in mind, if and , then is the same as up to the discretized distances, so the following proposition is immediate.
Proposition 2.3.
Suppose and . Then .
We now take some time to parse the definition of . Initially, all clients are in . For clients in , we are not sure yet whether we should serve them or not. Thus for these clients, we simply require , so they can be served any amount, and in the objective, the contribution of a client from is exactly its connection cost (up to discretization) to .
The clients in correspond to the “deleted” constraints in §1.3. Importantly, for , we do not require that ; rather, we relax this condition to . Recall that we made the assumption that every client can pay the radius of its set in §1.3. To realize this idea, we require that each pays its connection costs to in the objective. Then, to serve fully, must find units of open facility to connect to beyond . Now truly pays its radius, , for this units of connections in , so we can do “ball-chasing” to to find these facilities. In this case, we say that we re-route the client to some destination.
For clients in , we require . Note that the contribution of a to the objective of is exactly its connection cost to . The purpose of is to provide destinations for .
Finally, because we have decided to fully serve all clients in and , regardless of how much they are actually served in their -balls, we imagine that they every contributes to the coverage constraints, which is reflected in .
2.3. Properties of
Throughout our algorithm, we will modify the data of - we will move clients between , and and modify the -balls and radius levels. However, we still want the data of to satisfy some consistent properties, which we call our Basic Invariants.
Definition 2.4 (Basic Invariants).
We call the following properties our Basic Invariants:
-
(1)
partitions .
-
(2)
For all , we have for all .
-
(3)
For all , we have .
-
(4)
For all , we have .
-
(5)
(Distinct Neighbors) For all , if , then . In words, if the -balls of two clients in intersect, then they differ by exactly one radius level.
We want to emphasize Basic Invariant 2.4(5), which we call the Distinct Neighbors Property. It is not difficult to see that the Distinct Neighbors Property implies that has bipartite intersection graph.
Definition 2.5 (Intersection Graph).
Let be a set family indexed by . The intersection graph of is the undirected graph with vertex set such that two vertices and are connected by an edge if any only if .
Proposition 2.6.
Suppose satisfies the Distinct Neighbors Property. Then the intersection graph of is bipartite.
The following proposition will also be useful.
Proposition 2.7.
Suppose satisfies the Distinct Neighbors Property. Then each facility is in at most two -balls for clients in .
We summarize the relevant properties of in the following lemma. The algorithm described by the lemma is exactly the steps we took in this section.
Lemma 2.8.
There exists a polynomial time algorithm that takes as input a GKM instance and outputs such that and satisfies all Basic Invariants.
2.4. Notations
Throughout this paper, we will always index facilities using and clients using .
For any client , we say that is supported on facility if . Then for any , we let be the set of all facilities supported on at least one client in .
Given a setting of the -variables of , we say a facility is fractional (with respect to the given -variables) if . Otherwise, facility is integral. Similarly, we say a client is fractional if contains only fractional facilities, and is integral otherwise. Using these definitions, for any , we can partition into , where is the subset of fractional facilities and is the subset of integral facilities. An analogous partition holds for a subset of clients , so we have .
3. Basic Iterative Rounding Phase
In this section, we describe the iterative rounding phase of our algorithm. This phase has two main goals: (a) to simplify the constraint set of , and (b) to decide which clients to serve and how to serve them. To make these two decisions, we repeatedly solve to obtain an optimal extreme point, and then use the structure of tight constraints to update , and reroute clients accordingly.
3.1. The Algorithm
Our algorithm repeatedly solves to obtain an optimal extreme point , and then performs one of the following three possible updates, based on the tight constraints:
-
(1)
If some facility is set to zero in , we delete it from the instance.
-
(2)
If constraint is tight for some , then we decide to fully serve client by moving to either or . Initially, we add to then run Algorithm 2 to decide if should be in instead.
-
(3)
If constraint is tight for some , we shrink by one radius level (so ’s new -ball is exactly .) Then we possibly move to by running Algorithm 2 for .
These steps are made formal in Algorithms 1 (IterativeRound) and 2 (ReRoute). IterativeRound relies on the subroutine ReRoute, which gives our criterion for moving a client to . This criterion for adding clients to is the key way in which our algorithm differs from that of [KLS18]. In [KLS18], the criterion used ensures that is a family of disjoint sets. In contrast, we allow -balls for clients in to intersect, as long as they satisfy the Distinct Neighbors Property from Definition 2.4(5). Thus, our algorithm allows for rich structures in the set system .
The modifications made by IterativeRound do not increase , so upon termination of our algorithm, we have an optimal extreme point to such that is still a relaxation of GKM and no non-negativity constraint, -constraint, or -constraint is tight for . This is formalized in the following theorem, whose proof is similar to [KLS18], and is deferred to Appendix B.1.
Theorem 3.1.
IterativeRound is a polynomial time algorithm that maintains all Basic Invariants, weakly decreases , and outputs an optimal extreme point to such that no -, -, or non-negativity constraint is tight.
Recall the goals from the beginning of the section: procedure IterativeRound achieves goal (a) of making simpler while maintaining the Distinct Neighbors Property. Since we moved facilities between and , achieving goal (b) means deciding which facilities to open, and guaranteeing that each client has a “close-by” open facility. (Recall from §2 that is the set of clients such that their -balls are guaranteed to contain an open facility, and are the clients which are guaranteed to be served but using facilities opened in .)
Here’s the high-level idea of how we achieve goal (b). Suppose we move from to and some from to ; we want to find a good destination for . We claim ’s facility is a good destination for . Indeed, since is now in , we can use the constraint to bound the distance of to this unit of facility by , using the facts that and , which are guaranteed by ReRoute. Of course, if is removed from later, we re-route it to some client that is at least two radius levels smaller, and can send to that client. This corresponds to the “quarter ball-chasing” of §1.3. Indeed every further re-routing step for has geometrically decreasing cost, which give a cost of . We defer the formal analysis to 5.3, after we combine IterativeRound with another (new) iterative operation, which we present in the next section.
4. Iterative Operation for Structured Extreme Points
In this section, we achieve two goals: (a) we show that the structure of the extreme points of obtained from 3.1 are highly structured, and admit a chain decomposition. Then, (b) we exploit this chain decomposition to define a new iterative operation that is applicable whenever has “many” (i.e., more than ) fractional variables. We emphasize that this characterization of the extreme points is what enables the new iterative rounding algorithm.
4.1. Chain Decomposition
A chain is a sequence of clients in where the -ball of each client contains exactly two facilities—one shared with the previous ball and other with the next.
Definition 4.1 (Chain).
A chain is a sequence of clients satisfying:
-
•
for all , and
-
•
for all .
Our chain decomposition is a partition of the fractional -clients given in the next theorem, which is our main structural characterization of the extreme points of . (Recall that a client is fractional if all facilities in are fractional; we denote the fractional clients in by .)
Theorem 4.2 (Chain Decomposition).
Suppose satisfies all Basic Invariants. Let be an extreme point of such that no -, -, or non-negativity constraint is tight. Then there exists a partition of into at most chains, along with a set of at most violating clients (clients that are not in any chain.)
The proof relies on analyzing the extreme points of a set-cover-like polytope with side constraints; we defer it to §8 and proceed instead to define the new iterative operation.
4.2. Iterative Operation for Chain Decompositions
Composing 3.1 and 4.2, consider an optimal extreme point of , and a chain decomposition. We show that if the number of fractional variables in is sufficiently large, there exists a useful structure in the chain decomposition, which we call a candidate configuration.
Definition 4.3 (Candidate Configuration).
Let be an optimal extreme point of . A candidate configuration is a pair of two clients such that:
-
(1)
-
(2)
-
(3)
Every facility in and is in at exactly two -balls for clients in
-
(4)
and
Lemma 4.4.
Suppose satisfies all Basic Invariants, and let be an optimal extreme point of such that no -, -, or non-negativity constraint is tight. If , then there exist a candidate configuration in .
Our new iterative operation is easy to state: Find a candidate configuration and move from to .
The first two properties of candidate configurations are used to re-route to . Observe a key difference between ReRoute and ConfigReRoute: In the former, if a client is moved from to , there exists a client such that and . Thus we re-route to a client at least two radius levels smaller. This corresponds to “quarter ball-chasing.” On the other hand, in ConfigReRoute, we only guarantee a client of at least one radius level smaller, which corresponds to “half ball-chasing.” This raises the worry that if all re-routings are due to ConfigReRoute, any potential gains by ReRoute are not realized in the worst case. However we show that, roughly speaking, the last two properties of candidate configurations guarantee that the more expensive re-routings of ConfigReRoute happen at most half the time. The main properties of ConfigReRoute appear in the next theorem (whose proof is in Appendix B.2).
Theorem 4.5.
ConfigReRoute is a polynomial-time algorithm that maintains all Basic Invariants and weakly decreases .
Again, we defer the analysis of the re-routing cost of ConfigReRoute to §5.3, where we analyze the interactions between ConfigReRoute and ReRoute, and present our final pseudo-approximation algorithm next.
5. Pseudo-Approximation Algorithm for GKM
The pseudo-approximation algorithm for GKM combine the iterative rounding algorithm IterativeRound from §3 with the re-routing operation ConfigReRoute from §4 to construct a solution to .
Theorem 5.1 (Pseudo-Approximation Algorithm for GKM).
There exists an polynomial time randomized algorithm PseudoApproximation that takes as input an instance of GKM and outputs a feasible solution to with at most fractional facilities and expected cost at most .
There are two main components to analyzing PseudoApproximation. First, we show that the output extreme point has fractional variables. Second, we bound the re-routing cost. The first part follows directly by combining the analogous theorems for IterativeRound and ConfigReRoute. We defer its proof to Appendix B.
Theorem 5.2.
PseudoApproximation is a polynomial time algorithm that maintains all Basic Invariants, weakly decreases , and outputs an optimal extreme point of with at most fractional variables.
5.1. Analysis of Re-Routing Cost
We now bound the re-routing cost by analyzing how evolves throughout PseudoApproximation. This is one of the main technical contributions of our paper, and it is where our richer -set and relaxed re-routing rules are used. [KLS18] prove an analogous result about the re-routing cost of their algorithm. In the language of the following theorem statement, they show that for the case . We improve on this factor by analyzing the interactions between ReRoute and ConfigReRoute. Interestingly, analyzing each of ReRoute and ConfigReRoute separately would not yield any improvement over [KLS18] in the worst case, even with our richer set . It is only by using the properties of candidate configurations and analyzing sequences of calls to ReRoute and ConfigReRoute that we get an improvement.
Theorem 5.3 (Re-Routing Cost).
Upon termination of PseudoApproximation, let be a set of open facilities and such that for all . Then for all , , where .
We will need the following discretized version of the triangle inequality.
Proposition 5.4.
Let such that and intersect. Then .
Proof.
Let . Then using the triangle inequality we can bound:

The next lemma analyzes the life-cycle of a client that enters at some point in PseudoApproximation. Our improvement over [KLS18] comes from this lemma.
Lemma 5.5.
Upon termination of PseudoApproximation, let be a set of open facilities and such that for all . Suppose client is added to at radius level during PseudoApproximation (it may be removed later.) Then upon termination of PseudoApproximation, we have , where .
Proof.
Consider a client added to with radius level . If remains in until termination, the lemma holds for because . Thus, consider the case where is later removed from in PseudoApproximation. Note that the only two operations that can possibly cause this removal are ReRoute and ConfigReRoute. We prove the lemma by induction on . If , then remains in until termination because it has the smallest possible radius level and both ReRoute and ConfigReRoute remove a client from only if there exists another client with strictly smaller radius level.
Similarly, if , we note that ReRoute removes a client from only if there exists another client with radius level at least two smaller, which is not possible for . Thus, if does not remain in until termination, there must exist some that is later added to with radius level at most such that . We know that remains in until termination since it is of the lowest radius level. Thus:
Now consider where can possibly be removed from by either ReRoute or ConfigReRoute. In the first case, is removed by ReRoute, so there exists that is added to such that and . Applying the inductive hypothesis to , we can bound:
It is easy to verify by routine calculations that given that .
For our final case, suppose is removed by ConfigReRoute. Then there exists such that and . Further, . If remains in until termination, then:
Otherwise, is removed by ReRoute at an even later time because some is added to such that and . Applying the inductive hypothesis to , we can bound:
where implies .
Now, we consider the case where is later removed by ConfigReRoute. To analyze this case, consider when was removed by ConfigReRoute. At this time, we have by definition of Candidate Configuration. Because , consider any facility . When is removed from by ConfigReRoute, we have that is in exactly two -balls for clients in , exactly and . However, after removing from , is only in one -ball for clients in - namely .
Later, at the time is removed by ConfigReRoute, it must be the case that still, so is unchanged between the time that is removed and the time that is removed. Thus the facility that was previously in must still be present in . Then this facility must be in exactly two -balls for clients in , one of which is . It must be the case that the other -ball containing , say , was added to between the removal of and .
Note that the only operation that adds clients to is ReRoute, so we consider the time between the removal of and when is added to . Refer to Figure 2. At this time, we have , and because of the facility . Then it must be the case that has strictly smaller radius level than , so . To conclude the proof, we note that due to the facility , and apply the inductive hypothesis to :
which is at most . ∎
Now using the above lemma, we can prove Theorem 5.3.
Proof of Theorem 5.3.
Consider any client that is in upon termination of PseudoApproximation. It must be the case that was called at least once during PseudoApproximation. Consider the time of the last such call to . If is added to at this time, note that its radius level from now until termination remains unchanged, so applying Lemma 5.5 gives that , as required. Otherwise, if is not added to at this time, then there must exist some such that and . Then applying Lemma 5.5 to , we have:
5.2. Putting it all Together: Pseudo-Approximation for GKM
In this section, we prove Theorem 5.1. In particular, we use the output of PseudoApproximation to construct a setting of the -variables with the desired properties.
Proof of Theorem 5.1.
Given as input an instance of GKM, our algorithm is first to run the algorithm guaranteed by Lemma 2.8 to construct from such that and satisfies all Basic Invariants. Note that we will choose later to optimize our final approximation ratio. Then we run PseudoApproximation on , which satisfies all Basic Invariants, so by Theorem 5.2, PseudoApproximation outputs in polynomial time along with an optimal solution with fractional variables.
Given , we define a setting for the -variables: for all , connect to all facilities in by setting for all . For all , we have , so connect to all facilities in . Finally, to connect every to one unit of open facilities, we use the following modification of Theorem 5.3:
Proposition 5.6.
When PseudoApproximation terminates, for all , there exists one unit of open facilities with respect to within distance of , where .
The proof of the above proposition is analogous to that of Theorem 5.3 in the case , so we omit it. To see this, note that for all , we have . This implies that each has one unit of fractional facility within distance . Following an analogous inductive argument as in Lemma 5.5 gives the desired result.
By routine calculations, it is easy to see that for all . Now, for all , we connect to all facilities in . We want to connect to one unit of open facilities, so to find the remaining units, we connect to an arbitrary units of open facilities within distance of , whose existence is guaranteed by Proposition 5.6. This completes the description of .
It is easy to verify that is feasible for , because satisfies all knapsack constraints, and every client’s contribution to the coverage constraints in is exactly its contribution in . Thus it remains to bound the cost of this solution. We claim that , because each client in and contributes the same amount to and (up to discretization), and each client has connection cost at most times its contribution to .
In conclusion, the expect cost of the solution to is at most:
Choosing to minimize gives and . ∎
6. From Pseudo-Approximation to Approximation
In this section, we leverage the pseudo-approximation algorithm for GKM defined in Section 5 to construct improved approximation algorithms for two special cases of GKM: knapsack median and -median with outliers.
Recall that knapsack median is an instance of GKM with a single arbitrary knapsack constraint and a single coverage constraint that states we must serve every client in . Similarly, -median with outiers is an instance of GKM with a single knapsack constraint, stating that we can open at most facilities, and a single coverage constraint, stating that we must serve at least clients. Note that both special cases have .
Our main results for these two problems are given by the following theorems:
Theorem 6.1 (Approximation Algorithm for Knapsack Median).
There exists a polynomial time randomized algorithm that takes as input an instance of knapsack median and parameter and in time , outputs a feasible solution to of expected cost at most .
Theorem 6.2 (Approximation Algorithm for -Median with Outliers).
There exists a polynomial time randomized algorithm that takes as input an instance of -median with outliers and parameter and in time , outputs a feasible solution to of expected cost at most .
6.1. Overview
The centerpiece for both of our approximation algorithms is the pseudo-approximation algorithm PseudoApproximation for GKM. For both of these special cases, we can obtain via PseudoApproximation a solution to with only fractional facilities and bounded re-routing cost. Now our remaining task is to turn this solution with fractional facilities into an integral one.
Unfortunately, the basic LP relaxations for knapsack median and -median with outliers have an unbounded integrality gap. To overcome this bad integrality gap, we use known sparsification tools to pre-process the given instance. Our main technical contribution in this section is a post-processing algorithm that rounds the final fractional variables at a small cost increase for the special cases of knapsack median and -median with outliers.
Thus our approximation algorithms for knapsack median and -median with outliers consist of a known pre-processing algorithm [KLS18], our new pseudo-approximation algorithm, and our new post-processing algorithm.
6.2. Approximation Algorithm for Knapsack Median
To illustrate our approach, we give the pre- and post-processing algorithms for knapsack median, which is the simpler of the two variants. Our pre-processing instance modifies the data of the input instance, so the next definition is useful to specify the input instance and pre-processed instance.
Definition 6.3 (Instance of Knapsack Median).
An instance of knapsack median is of the form , where is a set of facilities, is a set of clients, is a metric on , is the weights of the facilities, and is the budget.
Note that for knapsack median, the two side constraints in are the knapsack constraint, , and the coverage constraint, .
We utilize the same pre-processing as in [KLS18]. Roughly speaking, given a knapsack median instance , we first handle the expensive parts of the optimal solution using enumeration. Once we pre-open the facilities and decide what clients should be assigned there for this expensive part of the instance, we are left with a sub-instance, say . In , our goal is to open some more facilities to serve the remaining clients.
Roughly speaking, is the “cheap” part of the input instance. Thus, when we construct for this sub-instance, we initialize additional invariants which we call our Extra Invariants.
To state our Extra Invariants, we need to define the -ball centered at with radius for any , , and , which is the set:
Definition 6.4 (Extra Invariants for Knapsack Median).
Let , , , and be given. Then we call the following properties our Extra Invariants:
-
(1)
For all , there exists a dummy client such that with radius level . We let be the collection of these dummy clients.
-
(2)
For all that is not collocated with some , we have
-
(3)
For all , we have
-
(4)
For all and , we have:
Extra Invariant 6.4(1) guarantees that we open the set of guessed facilities in our final solution. Then for all non-guessed facilities, so the set , Extra Invariant 6.4(2) captures the idea that these facilities are “cheap.” Taken together, Extra Invariants 6.4(3) and 6.4(4) capture the idea that all remaining clients are “cheap.”
The next theorem describes our pre-processing algorithm for knapsack median, which is a convenient re-packaging of the pre-processing used in [KLS18]. The theorem essentially states that given , and , we can efficiently guess a set of clients and of facilities that capture the expensive part of the input instance . Then when we construct for the cheap sub-instance, we can obtain the Extra Invariants, and the cost of extending a solution of the sub-instance to the whole instance is bounded with respect to , which one should imagine is .
Theorem 6.5 (Pre-Processing for Knapsack Median).
Let be an instance of knapsack median. Then, given as input instance , parameters , and an upper bound on , there exists an algorithm that runs in time and outputs -many sub-instances of the form along with the data for on , a set of facilities , and a vector such that:
-
(1)
satisfies all Basic and Extra Invariants
-
(2)
The proof is implicit in [KLS18]. For completeness, we prove the analogous theorem for -median with outliers, 6.13, in §F.
We will show that if satisfies the Extra Invariants for knapsack median, then we can give a post-processing algorithm with bounded cost. It is not difficult to see that PseudoApproximation maintains the Extra Invariants as well, so we use the Extra Invariants in our post-processing.
Proposition 6.6.
PseudoApproximation maintains all Extra Invariants for knapsack median.
Now we move on to describing our post-processing algorithm. Suppose we run the pre-processing algorithm guaranteed by Theorem 6.5 to obtain satisfying all Basic- and Extra Invariants. Then we can run PseudoApproximation to obtain an optimal extreme point of with fractional facilities, and still satisfies all Basic- and Extra Invariants.
It turns out, to round these fractional facilities, it suffices to open one facility in each -ball for clients in . Then we can apply Theorem 5.3 to bound the re-routing cost. The main difficulty in this approach is that we must also round some fractional facilities down to zero to maintain the knapsack constraint.
Note that closing a facility can incur an unbounded multiplicative cost in the objective. To see this, consider a fractional facility that is almost open, so . Then suppose there exists such that and . Then ’s contribution to the objective of is . However, if we close , then ’s contribution increases to .
To bound the cost of closing facilities, we use the Extra Invariants. In particular, we use the next technical lemma, which states that if we want to close down a facility , and every client that connects to has a back-up facility to go to within distance , then closing incurs only a small increase in cost. For proof, see §C.
Lemma 6.7.
Suppose satisfies all Basic and Extra Invariants for knapsack median, and let and . Further, consider a facility and set of clients such that for all , we have and there exists some facility in within distance of . Then .
By the next proposition, rounding a facility up to one does not incur any cost increase, because every client must be fully connected.
Proposition 6.8.
Upon termination of PseudoApproximation on a knapsack median instance, we have .
Proof.
We observe that the single coverage constraint in for a knapsack median instance is of the form:
, where we use the fact that , , and partition due to Basic Invariant 2.4(1). Combining this with the constraint for all gives that for all for any feasible solution to . By assumption, no -constraint is tight upon termination of PseudoApproximation, so the proposition follows. ∎
To summarize, the goal of our post-processing algorithm is to find an integral setting of the fractional facilities in the output of PseudoApproximation such that the knapsack constraint is satisfied and there is an open facility in each -ball for clients in .
Lemma 6.9.
Upon termination of PseudoApproximation on a knapsack median instance, let be the outputted extreme point of , and suppose satisfies all Basic- and Extra Invariants. Then there exists an integral setting of the fractional facilities such that the knapsack constraint is satisfied, there is an open facility in each -ball for clients in , and every facility in is open.
Proof.
Consider the following LP:
The first constraint states that we want one open facility in each -ball for clients in , and the second states that our solution should agree on the integral facilities in .
Because satisfies all Basic Invariants, the intersection graph of is bipartite by Proposition 2.6. Then the feasible region of is a face of the intersection of two partition matroids (each side of the biparitition of defines one parititon matroid), and thus is integral.
To conclude the proof, we observe that is feasible for , so . Thus there exists an integral setting of facilities that opens one facility in each -bal for all clients in , agrees with all of ’s integral facilities, and has total weight at most . Finally, by Extra Invariant 6.4(1), , so we open every facility in . ∎
Thus, in light of Lemma 6.9, our post-processing algorithm is to enumerate over all integral settings of the fractional variables to find one that satisfies the knapsack constraint, opens one facility in each -ball for clients in , and opens . Combining our post-processing algorithm with PseudoApproximation gives the following theorem.
Theorem 6.10.
There exists a polynomial time algorithm that takes as input for knapsack median instance satisfying all Basic- and Extra Invariants and outputs a feasible solution to such that the solution opens all facilities in and has cost at most , where .
Proof.
Our algorithm is to run PseudoApproximation on and then run our post-processing algorithm, which is to enumerate over all integral settings of the fractional variables, and then output the feasible solution that opens of lowest cost (if such a solution exists.)
Let be the optimal extreme point of output by PseudoApproximation, which has fractional variables by 5.2. Because has fractional variables, our post-processing algorithm is clearly efficient, which establishes the runtime of our overall algorithm.
Note that upon termination, still satisfies all Basic- and Extra Invariants. Then by Lemma 6.9, there exists an integral setting of the fractional variables that is feasible, opens , and opens a facility in each -ball for clients in . It suffices to bound the cost of this solution. Let denote the facilities opened by this integral solution, so for all . Applying Lemma 5.3 with , we obtain that for all , where . It is easy to check that for all .
To bound the cost of the solution relative to , we must bound the cost of closing the -many facilities in . We recall that by Proposition 6.8, we have , so all clients must be fully connected in .
First we consider any client that is not supported on any facility in . Such a client is not affected by closing , so if is empty, then , which is at most times ’s contribution to . Otherwise, contains an integral facility in to connect to, so is at most ’s contribution to .
It remains to consider the clients whose -balls contain a facility in . Because there are only -many facilities in , it suffices to show that for each , the additive cost of connecting all clients supported on is at most . Here we apply Lemma 6.7 to the set of clients to obtain .
To summarize, the cost of connecting the clients not supported on is at most , and the cost of the remaining clients is , as required. ∎
Now our complete approximation for knapsack median follows from combining the pre-processing with the above theorem and tuning parameters.
Proof of Theorem 6.1.
Let . We will later choose with respect to the given to obtain the desired approximation ratio and runtime. First, we choose parameters and for our pre-processing algorithm guaranteed by Theorem 6.5. We take and . We require that is an upper bound on . Using a standard binary search idea, we can guess up to a multiplicative -factor in time , so we guess such that .
With these choices of parameters, we run the algorithm guaranteed by Theorem 6.5 to obtain many sub-instances such that one such sub-instance is of the form , where for satisfies all Basic- and Extra Invariants, and we have:
(1) |
Then for each sub-instance output by the pre-processing, we run the algorithm guaranteed by Theorem 6.10 to obtain a solution to each sub-instance. Finally, out of these solutions, we output the one that is feasible for the whole instance with smallest cost. This completes the description of our approximation algorithm for knapsack median. The runtime is , so it remains to bound the cost of the output solution and to choose the parameters and .
To bound the cost, it suffices to consider the solution output on the instance where satisfies all Basic- and Extra Invariants and Equation 1. By running the algorithm guaranteed by Theorem 6.10 on this , we obtain a feasible solution to such that , and the cost of connecting to is at most , where . To extend this solution on the sub-instance to a solution on the whole instance , we must connect to . Because , applying Equation 1 allows us to upper bound the expected cost of connecting to by:
Now choosing to minimize gives and . Thus the expected cost of this solution is at most , where . Finally, by routine calculations, we can choose so that expected cost is at most , as required. Note that the runtime of our algorithm is . ∎
6.3. Approximation Algorithm for -Median with Outliers
Our approximation algorithm for -median with outliers follows the same general steps as our algorithm for knapsack median. We state the analogous Extra Invariants for -median with outliers and pre-processing algorithm here. The only differences between the Extra Invariants for knapsack median and -median with outliers is in the final Extra Invariant.
Definition 6.11 (Instance of -Median with Outliers).
An instance of -median with outliers is of the form , where is a set of facilities, is a set of clients, is a metric on , is the number of facilities to open, and is the number of clients to serve.
Note that for -median with outliers, the two side constraints in are the knapsack constraint, , and the coverage constraint, .
Definition 6.12 (Extra Invariants for -Median with Outliers).
Let , , , and be given. Then we call the following properties our Extra Invariants:
-
(1)
For all , there exists a dummy client such that with radius level . We let be the collection of these dummy clients.
-
(2)
For all that is not collocated with some , we have
-
(3)
For all , we have
-
(4)
For every and , we have:
Theorem 6.13 (Pre-Processing for -Median with Outliers).
Let be an instance of -median with outliers with optimal solution . Then, given as input instance , parameters , and an upper bound on , there exists an algorithm that runs in time and outputs -many sub-instances of the form along with the data for on , a set of facilities , and a vector such that:
-
(1)
satisfies all Basic and Extra Invariants
-
(2)
It is easy to check that PseudoApproximation maintains all Extra Invariants for -median with outliers as well, and we have an analogous technical lemma to bound the cost of closing facilities. For proof of the lemma, see §C.
Proposition 6.14.
PseudoApproximation maintains all Extra Invariants for -median with outliers.
Lemma 6.15.
Suppose satisfies all Basic and Extra Invariants for -median with outliers, and let and . Further, consider a facility and set of clients such that for all , we have and there exists some facility in within distance of . Then .
Now we focus on the main difference between the two algorithms: the post-processing. In particular, the coverage constraint of -median with outliers introduces two difficulties in rounding the final fractional facilities: (a) we are no longer guaranteed that , and (b) we must satisfy the coverage constraint.
The difficulty with (a) is that now rounding a facility up to one can also incur an unbounded multiplicative cost in the objective. To see this, consider a fractional facility that is almost closed, so . Consider rounding this facility up to one. Then for a client that fractionally connects to in the solution , if we fully connect to , this costs . The solution here is to use Extra Invariant 6.12(2) to bound the additive cost of opening facilities.
The more troublesome issue is (b). Note that the same approach that we used to prove that there exists a good integral setting of the fractional variables in 6.9 does not work here because putting the coverage constraint in the objective of the LP could result in a solution covering the same client multiple times. Our solution to (b) is a more sophisticated post-processing algorithm that first re-routes clients in . After re-routing, we carefully pick facilities to open that do not double-cover any remaining -clients. We defer the details of our post-processing algorithm for §7. For now, we present the guarantees of our pseudo-approximation combined with post-processing:
Theorem 6.16.
There exists a polynomial time algorithm that takes as input for -median with outliers instance satisfying all Basic- and Extra Invariants and outputs a feasible set of facilities such that the cost of connecting clients to is at most , where for any constant .
7. Post-Processing for -Median with Outliers
In this section we develop the post-processing algorithm for -median with outliers that is guaranteed by 6.16. The structure of our algorithm is recursive. First, we give a procedure to round at least one fractional facility or serve at least one client. Then we recurse on the remaining instance until we obtain an integral solution.
7.1. Computing Partial Solutions
In this section, we show how to round at least one fractional facility or serve at least one client. We interpret this algorithm as computing a partial solution to the given -median with outliers instance.
The main idea of this algorithm is to re-route clients in . In particular, we maintain a subset such that for every client in , we guarantee to open an integral facility in their -ball. We also maintain a subset of -clients that we re-route; that is, we guarantee to serve them even if no open facility is in their -balls. Crucially, every client in is supported on at most one -ball for clients in . Thus, we do not have to worry about double-covering those clients when we round the facilities in .
The partial solution we output consists of one open facility for each client in (along with the facilities that happen to be integral already), and we serve the clients in , , , and the -clients supported on our open facilities. See Algorithm 5 (ComputePartial) for the formal algorithm to compute partial solutions. Note that is a parameter of ComputePartial.
We note that to define a solution for -median with outliers, it suffices to specify the set of open facilities , because we can choose the clients to serve as the closest clients to . Thus when we output a partial solution, we only output the set of open facilities.
We summarize the performance of ComputePartial with the next theorem, which we prove in §7.4. In the next section, we use 7.1 to define our recursive post-processing algorithm.
Theorem 7.1.
Let and be the input to ComputePartial. Then let be the partial solution output by ComputePartial and be the modified LP. Then satisfies all Basic- and Extra Invariants and we have:
where .
We want to satisfy all Basic- and Extra Invariants so we can continue recursing on . The second property of 7.1 allows us to extend a solution computed on with the partial solution .
7.2. Recursive Post-Processing Algorithm
To complete our post-processing algorithm, we recursively apply ComputePartial until we have an integral solution.
The main idea is that we run PseudoApproximation to obtain an optimal extreme point with fractional variables. Then using this setting of the -variables, we construct a partial solution consisting of some open facilities along with the clients that they serve. However, if there are still fractional facilities remaining, we recurse on (after the modifications by ComputePartial.) Our final solution consists of the union of all recursively computed partial solutions. See Algorithm 6 (OutliersPostProcess.)
For proof of the next lemma, see §D.
Lemma 7.2.
If , then has at most two fractional variables.
7.3. Analysis of OutliersPostProcess
In this section, we show that OutliersPostProcess satisfies the guarantees of 6.16. All missing proofs can be found in §D. We let be the output of PseudoApproximation in the first line of OutliersPostProcess and . First, we handle the base cases of OutliersPostProcess.
The easiest base is when is integral. Here, we do not need the Extra Invariants - all we need is that every client has an open facility within distance .
Lemma 7.3.
If is integral, then the output solution is feasible, contains , and connecting clients costs at most .
Now, in the other base case, is not integral, but we know that . By 7.2, we may assume without loss of generality has exactly two fractional facilities, say . Further, we may assume that the -constraint is tight, because opening more facilities can only improve the objective value. It follows that . For the sake of analysis, we define the sets and where we may assume .
It remains to bound the cost of the solution . One should imagine that we obtain this solution from by closing down the facility and opening up . First we handle the degenerate case where . In this case, and must both be co-located copies of a facility in , so opening one or the other does not change the cost. Thus, we may assume without loss of generality that . Here, we need to Extra Invariants to bound the cost of opening and closing
Lemma 7.4.
Suppose . Then the output solution is feasible, contains , and connecting clients costs at most
This completes the analysis of the base cases. To handle the recursive step, we apply 7.1.
Proof of Theorem 6.16.
First, we show that OutliersPostProcess terminates in polynomial time. It suffices to show that the number of recursive calls to ComputePartial is polynomial. To see this, note that for each recursive call, it must be the case that . In particular, there exists some non-dummy client in . Thus, we are guaranteed to remove at least once client from in each recursive call.
Now it remains to show that the output solution is feasible, contains , and connecting clients costs at most . Let be the extreme point computed by PseudoApproximation in the first line of OutliersPostProcess. If is integral or , then we are done by the above lemmas.
Then it remains to consider the case where . Let denote the input to OutliersPostProcess and the updated at the end of ComputePartial as in the statement of Theorem 7.1. We note that Theorem 7.1 implies that satisfies all Basic and Extra Invariants, so is a valid input to OutliersPostProcess. Then we may assume inductively that the recursive call to OutliersPostProcess on outputs a feasible solution to such that and the cost of connecting clients from to is at most .
Further, let be the partial solution output by ComputePartial on . Now we combine the solutions and to obtain the solution output by OutliersPostProcess. First, we check that is is feasible. This follows, because by definition of ComputePartial. Also, by the inductive hypothesis.
It remains to bound the cost of connecting clients to . Consider serving the closest clients in with and with . Because , this is enough clients. Connecting the closest clients in to costs at most by the inductive hypothesis. Now we use the guarantee of 7.1, which we recall is:
Thus, the total connection cost is at most:
Note that the additive terms which we accrue in each recursive call are still overall. This is because we keep recursing on a subset of the remaining fractional facilities – which is always – and we open/close each fractional facility at most once over all recursive calls. Thus, we can bound the additive cost of each opening/closing by . ∎
7.4. Proof of 7.1
For all missing proofs in this section, see §D. We let and denote the input to ComputePartial and the updated LP that is output at the end ComputePartial. We begin with three properties of ComputePartial that will be useful throughout our analysis.
The first is immediate by definition of ComputePartial.
Proposition 7.5.
Upon termination of ComputePartial, the set family is disjoint, and every client , intersects at most one -ball for clients in .
Proposition 7.6.
ComputePartial initializes and maintains the invariants that and
Proof.
We initialize and only add clients from to . Similarly, we initialize and only remove clients from . ∎
Lemma 7.7.
Every that is reached by the For loop remains in until termination.
Now we are ready to prove both properties of 7.1. It is not difficult to see that satisfies all Basic- and Extra Invariants by construction.
Lemma 7.8.
satisfies all Basic- and Extra Invariants.
Now it remains to show . To do this, we partition into and . For each client in , we show that its contribution to the objective of is at most its contribution to . Then for each client , either is at most times ’s contribution to or we can charge ’s connection cost to an additive term.
First, we focus on . For these clients, it suffices to show that (restricted to ) is feasible for . This is because for all , either or . The clients in contribute zero to the cost of and . This is because both and satisfy Extra Invariant 6.12(1), so every dummy client is co-located with one unit of open facility corresponding to .
Thus it remains to consider the clients . We recall that and for all , so each costs less in than in .
To complete the cost analysis of , we go back to prove feasibility. The main difficulty is showing that the coverage constraint is still satisfied. Recall that we construct by greedily opening the facility in each -ball for clients in that covers the most -clients. 7.5 ensures that this greedy choice is well-defined (because is disjoint), and that we do not double-cover any -clients.
Then by definition of greedy, we show that our partial solution covers more clients than the fractional facilities we delete. This proposition is the key to showing that the coverage constraint is still satisfied.
Proposition 7.9.
Upon termination of ComputePartial, we have .
Proof.
For each , let be the unique open facility in . By definition of , for all we have . Further, by Proposition 7.5, intersects exactly one -ball among clients in , so each for covers a unique set of clients. This implies the equality:
Combining this equality with the facts that for all , we have and for all gives the desired inequality:
∎
Finally, we can complete the analysis of . It is easy to check that constraints except for the coverage constraint are satisfied. To handle the coverage constraint, we use 7.9.
Lemma 7.10.
restricted to is feasible for .
For the final property, we must upper bound the connection cost of to . We bound the connection cost in a few steps. First, we bound the re-routing cost of . Second, we show that every has a ”back-up” facility within . This is used to bound the additive cost of connecting . Finally, for the clients in , we guarantee to open a facility in their -balls, so we also bound their additive cost.
The next lemma and corollary allow us to bound the cost of .
Lemma 7.11.
Upon termination of ComputePartial, for all , we have .
Proof.
There are a few cases to consider. First, if , then the lemma is trivial because by definition there exists an integral facility in . Otherwise, if , but , then again the lemma is trivial.
Thus it remains to consider clients . We note that such a client is initially in . If remains in until termination, then we are done, because we are guaranteed to open a facility in for all (this is exactly the set of facilities.)
For the final case, we suppose client is removed from in the iteration where we consider . Then either or there exists such that intersects both and and . Note that because the For loop considers clients in increasing order of radius level, we have
In the former case, by the Distinct Neighbors Property, we have . Further, by 7.7, we know that remains in until termination, so . Then we can upper bound:
, where in the final inequality, we use the fact that .
In the latter case, we again have . Then we can bound the distance from to by first going from to , then from to , and finally from to , where and :
∎
Corollary 7.12.
Upon termination of ComputePartial, for all , we have , where
Proof.
Apply 5.3 with . ∎
Similarly, the next lemma ensures that can also be re-routed.
Lemma 7.13.
Upon termination of ComputePartial, for all , we have .
Proof.
Because , it must be the case that we put in in some iteration of the For loop where we consider client . Thus we have and . Also, by Lemma 7.7, remains in until termination, so . Using these facts, we can bound:
∎
Using the above lemmas, we are ready to bound the connection cost of our partial solution. Note that we bound the cost of serving not with , which is the partial solution we output, but rather with . Thus, we are implicitly assuming that will be opened by the later recursive calls.
We recall that and .
To begin, we bound the cost of connecting to . By definition, every client has some facility from in its -ball. Further, by definition, and using Proposition 7.5. Thus we can apply Extra Invariant 6.12(2) to each facility in . Further, we know that , because there are only fractional facilities, and every facility in is fractional by definition. Then we can bound:
, where we apply Extra Invariant 6.12(2) for each . Thus, we have shown that the connection cost of is at most an additive .
Now we move on to the rest of , that is - the clients in , and . For the clients in , we know by Lemma 7.13 that every client has an open facility in at distance at most . Further, by definition of , each is supported on a fractional facility not in . To see this, note that for all , there exists such that , and .
Then we can use Lemma 6.15 for each fractional facility to bound the cost of connecting all -clients supported on to . For each such fractional facility , the connection cost of these clients is at most . By summing over all -many fractional , we connect all of at additive cost at most .
Finally, we handle the clients in . For convenience, we denote this set of clients by . Using Lemma 7.11, every has a facility in at distance at most . There are a few cases to consider. For the first case, we consider the clients , that is the set of -clients whose -balls contain some cheap, fractional facility. By an analogous argument as for , we can apply Lemma 6.15 to each fractional to bound the cost of connecting to by .
Then it remains to consider the clients such that . For such a client , if , then ’s contribution to the objective of is exactly , so connecting to costs at most times ’s contribution to the objective of .
Similarly, if happens to contain a facility in , then we can simply connect to the closest such facility in . Note that , so ’s connection cost in this case it as most its contribution to the objective in . In conclusion, the connection cost of is at most . Summing the costs of these different groups of clients completes the proof.
8. Chain Decompositions of Extreme Points
In this section, we prove a more general version of Theorem 4.2 that applies to set-cover-like polytopes with side constraints. In particular, we consider polytopes of the form:
, where , , and the -balls are defined as in , and is an arbitrary system of linear inequalities. We note that other than the -, and -constraints, generalizes the feasible region of by taking the system to be the knapsack constraints and coverage constraints.
Although we phrase in terms of facilities and clients, one can interpret as a set cover polytope with side constraints as saying that we must choose elements in to cover each set in the family subject to the constraints . The main result of this section is that if has bipartite intersection graph, there exists a chain decomposition of the extreme points of .
Note that can also also be interpreted as the intersection of two partition matroid polytopes or as a bipartite matching polytope both with side constraints. Our chain decomposition theorem shares some parallels with the work of Grandoni, Ravi, Singh, and Zenklusen, who studied the structure of bipartite matching polytopes with side constraints [GRSZ14].
Theorem 8.1 (General Chain Decomposition).
Suppose we have a polytope:
such that is a finite ground set of elements (facilities), is a set family indexed by (clients), and is a system of linear inequalities. Further, let be an extreme point for such that no non-negativity constraint is tight. If has bipartite intersection graph, then admits a partition into chains along with at most violating clients (clients that are not in any chain.)
We will use the following geometric fact about extreme points of polyhedra:
Fact 8.2.
Let be a polyhedron in . Then is an extreme point of if and only if there exist linearly independent constraints of that are tight at . We call such a set of constraints a basis for .
Proof of Theorem 4.2.
Let be an extreme point of such that no -, -, or non-negativity constraint is tight, and suppose satisfies the Distinct Neighbors Now consider the polytope:
, where consists of the knapsack constraints and coverage constraints of . We claim that is an extreme point of . To see this, note that is an extreme point of , so fix any basis for using tight constraints of . By assumption, this basis uses no -, -, or non-negativity constraint. In particular, it only uses constraints of that are also present in , so this basis certifies that is an extreme point of .
8.1. Proof of Theorem 8.1
Now we go back to prove the more general chain decomposition theorem: Theorem 8.1. For all missing proofs, see §E. Throughout this section, let
be a polytope satisfying the properties of Theorem 8.1. In particular, the intersection graph of is bipartite. Further, let be an extreme point of such that no non-negativity constraint is tight for .
The crux of our proof is the next lemma, which allows us to bound the complexity of the intersection graph with respect to the number of side constraints . We prove the lemma by constructing an appropriate basis for . The next definition is useful for constructing a basis.
Definition 8.3.
For any subset , let denote the maximum number of linearly independent -constraints, so the constraint set .
Lemma 8.4.
Let be an extreme point of such that no non-negativity constraint is tight. Then the number of fractional facilities in satisfies (recall that is the number of constraints of .)
Now, to find a chain decomposition of , first we find the violating clients. We note that every -ball contains at least two facilities. The violating clients will be those clients whose -balls contain strictly more than two facilities, so we let be the set of violating clients. It remains to bound the size of , which follows from a standard counting argument.
Proposition 8.5.
Now that we have decided on the violating clients, it remains to partition into the desired chains. Importantly, for all , we have . To find our chains, we consider the intersection graph of , so the intersection graph of the set family . We let denote this graph. Note that is a subgraph of the standard intersection graph, so it is also bipartite by assumption.
We consider deleting the vertices from , which breaks into some connected components, say . Let denote the vertex set of , so we have that partitions . Further, for all , every -ball for clients in contains exactly two facilities, and every facility is in at most two -balls. Translating these statements into properties of the intersection graph, we can see that every vertex of has degree at most two, and is connected, so we can conclude that each is a path or even cycle (we eliminate the odd cycle case because the intersection graph is bipartite.)
Proposition 8.6.
Each is a chain.
To complete the proof, it remains to upper bound the number of chains. To do this, we first split the inequality given by Lemma 8.4 into the contribution by each . Importantly, we observe that the ’s are disjoint for all because the ’s correspond to distinct connected components. Then we can write:
The way to interpret this inequality is that each chain, , has a budget of fractional facilities to use in its chain, but we have an extra facilities to pay for any facilities beyond each ’s allocated budget. We will show that each chain uses at least one extra facility from this surplus, which allows us to upper bound by .
For the path components, we use the fact that every path has one more vertex than edge.
Proposition 8.7.
If is a path, then .
For the even cycle components, we show that the corresponding -constraints are not linearly independent.
Proposition 8.8.
If is an even cycle, then .
Applying the above two propositions, we complete the proof by bounding :
References
- [AGK+04] Vijay Arya, Naveen Garg, Rohit Khandekar, Adam Meyerson, Kamesh Munagala, and Vinayaka Pandit. Local search heuristics for k-median and facility location problems. SIAM J. Comput., 33(3):544–562, 2004.
- [BPR+17] Jaroslaw Byrka, Thomas W. Pensyl, Bartosz Rybicki, Aravind Srinivasan, and Khoa Trinh. An improved approximation for k-median and positive correlation in budgeted optimization. ACM Trans. Algorithms, 13(2):23:1–23:31, 2017.
- [BPR+18] Jaroslaw Byrka, Thomas W. Pensyl, Bartosz Rybicki, Joachim Spoerhase, Aravind Srinivasan, and Khoa Trinh. An improved approximation algorithm for knapsack median using sparsification. Algorithmica, 80(4):1093–1114, 2018.
- [Che08] Ke Chen. A constant factor approximation algorithm for -median clustering with outliers. In SODA, pages 826–835, 2008.
- [CKMN01] Moses Charikar, Samir Khuller, David M. Mount, and Giri Narasimhan. Algorithms for facility location problems with outliers. In S. Rao Kosaraju, editor, Proceedings of the Twelfth Annual Symposium on Discrete Algorithms, January 7-9, 2001, Washington, DC, USA, pages 642–651. ACM/SIAM, 2001.
- [FKRS19] Zachary Friggstad, Kamyar Khodamoradi, Mohsen Rezapour, and Mohammad R. Salavatipour. Approximation schemes for clustering with outliers. ACM Trans. Algorithms, 15(2):26:1–26:26, 2019.
- [GLZ17] Sudipto Guha, Yi Li, and Qin Zhang. Distributed partial clustering. In Proceedings of the 29th ACM Symposium on Parallelism in Algorithms and Architectures, pages 143–152. ACM, 2017.
- [GMM+03] Sudipto Guha, Adam Meyerson, Nina Mishra, Rajeev Motwani, and Liadan O’Callaghan. Clustering data streams: Theory and practice. TKDE, 15(3):515–528, 2003.
- [GRSZ14] Fabrizio Grandoni, R. Ravi, Mohit Singh, and Rico Zenklusen. New approaches to multi-objective optimization. Math. Program., 146(1-2):525–554, 2014.
- [IQM+20] Sungjin Im, Mahshid Montazer Qaem, Benjamin Moseley, Xiaorui Sun, and Rudy Zhou. Fast noise removal for k-means clustering. In The 23rd International Conference on Artificial Intelligence and Statistics, AISTATS 2020, 26-28 August 2020, Online [Palermo, Sicily, Italy], pages 456–466, 2020.
- [JMS02] Kamal Jain, Mohammad Mahdian, and Amin Saberi. A new greedy approach for facility location problems. In Proceedings of the Thiry-fourth Annual ACM Symposium on Theory of Computing, STOC ’02, pages 731–740, New York, NY, USA, 2002. ACM.
- [JV01] Kamal Jain and Vijay V. Vazirani. Approximation algorithms for metric facility location and k-median problems using the primal-dual schema and lagrangian relaxation. J. ACM, 48(2):274–296, 2001.
- [KKN+15] Ravishankar Krishnaswamy, Amit Kumar, Viswanath Nagarajan, Yogish Sabharwal, and Barna Saha. Facility location with matroid or knapsack constraints. Math. Oper. Res., 40(2):446–459, 2015.
- [KLS18] Ravishankar Krishnaswamy, Shi Li, and Sai Sandeep. Constant approximation for k-median and k-means with outliers via iterative rounding. In Ilias Diakonikolas, David Kempe, and Monika Henzinger, editors, Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing, STOC 2018, Los Angeles, CA, USA, June 25-29, 2018, pages 646–659. ACM, 2018.
- [LG18] Shi Li and Xiangyu Guo. Distributed -clustering for data with heavy noise. In Advances in Neural Information Processing Systems, pages 7838–7846, 2018.
- [LS16] Shi Li and Ola Svensson. Approximating k-median via pseudo-approximation. SIAM J. Comput., 45(2):530–547, 2016.
- [MKC+15] Gustavo Malkomes, Matt Kusner, Wenlin Chen, Kilian Weinberger, and Benjamin Moseley. Fast distributed -center clustering with outliers on massive data. In NIPS, pages 1063–1071, 2015.
Appendix A Missing Proofs from §2: Construction of
Proof of Proposition 2.1.
Let be the given instance of GKM and be an optimal solution to .
Observe that if for all , then we can define for all . It is easy to verify in this case that is feasible for and achieves the same objective value in as achieves in , which completes the proof.
Thus our goal is to duplicate facilities in and re-allocate the - and -values appropriately until for all . To prevent confusion, let denote the original set of facilities, and let denote the modified set of facilities, where make copies of each facility in , so for each , we have copies .
Now we define and with the desired properties. For each , we assume without loss of generality that . We define and recursively:
Let and for all .
Now for , let and for all .
It is easy to verify that is feasible for (after duplicating facilities) and for all , as required. Further, it is clear that this algorithm is polynomial time. ∎
Proof of Proposition 2.2.
If , then the claim is trivial. Suppose . We can rewrite for some . Also, for convenience we define . Because is uniformly distributed in , it follows that is uniformly distributed in .
It follows, is rounded to exactly when , and otherwise is rounded to when . Thus we compute:
∎
Proof of Proposition 2.6.
Assume for contradiction that the intersection graph of is not bipartite, so there exists an odd cycle, say such that each vertex . Further, along each edge , we have , so and differ by exactly one. In particular, the radius level can either increase by one or decrease by one along each edge.
Consider traversing the cycle starting from all the way to and then back to , and count the number of increases and decreases along the way. The number of increases and decreases must be equal when we return to , but this cycle has an odd number of edges, so the number of increases and decreases cannot be the same. This is a contradiction. ∎
Proof of Proposition 2.7.
Assume for contradiction that there exists a facility such that for distinct clients . Then the intersection graph of contains an odd cycle . This contradicts the fact that the intersection graph is bipartite. ∎
Proof of Lemma 2.8.
Our algorithm is to first run the algorithm guaranteed by Lemma 2.1 to obtain and the -balls such that . Then we follow the construction in §2- that is, we randomly discretize the distances to obtain , define the - and - balls and radius levels, and initialize , and . This completes the description of .
By Proposition 2.3, we have , as required. Finally, it is easy to check that satisfies all Basic Invariants. ∎
Appendix B Missing Proofs from §3 - 5: Analysis of PseudoApproximation
In this section, we present all missing proofs from the analyses of PseudoApproximation and its sub-routines IterativeRound and ConfigReRoute.
B.1. Analysis of IterativeRound
The goal of this section is to prove Theorem 3.1. First we show that IterativeRound maintains all Basic Invariants. It is easy to see that the first three Basic Invariants are maintained by IterativeRound, so we only prove the last two.
Proof.
Consider any . Suppose the invariant holds at the beginning of IterativeRound, so initially we have . Note that a necessary condition for decreasing is that is tight at some iteration of IterativeRound, and in this case we decrease by one.
Suppose . Then , so . Thus it cannot be the case that is tight. We conclude that for all , we never decrease beyond negative one. ∎
Lemma B.2 (Basic Invariant 2.4(5): Distinct Neighbors Property).
IterativeRound maintains the Distinct Neighbors Property.
Proof.
It suffices to show that ReRoute maintains the Distinct Neighbors Property, because in IterativeRound, the only time is modified is when ReRoute is called. Thus we consider an arbitrary call to . To prevent confusion, let denote the status of before the call to ReRoute. If we do not move from to , then the invariant is clearly maintained, so we may assume that we move from to .
Then for all such that , we have . Finally, after adding to , we remove all such with , so for all remaining clients in whose -balls intersect ’s, their radius level is exactly one larger than ’s. ∎
To show that IterativeRound weakly decreases , it suffices to show that each iteration weakly decreases . We show that in any iteration of IterativeRound, the computed at the beginning of the iteration is still feasible after we update in that iteration. Then, we show that achieves the same objective value before and after the updates.
Lemma B.3.
Each iteration of IterativeRound weakly decreases .
Proof.
Consider any iteration of IterativeRound. Let be the optimal solution to computed at the beginning of the iteration. There are three possible modifications we can make to in this iteration: remove a facility from , move a client from to , or shrink a -ball for a client in . For each operation, we show that is still feasible and achieves the same objective value afterwards.
For the first operation, if we remove a facility from , then it must be the case that . Thus it is immediate that (restricted to ) is feasible after deleting , and achieves the same objective value.
Otherwise, suppose there exists such that , and we move from to and then . Thus ends up it either or at the end of this iteration. In either case, we have , so satisfies the corresponding constraint after updating . Further, because , we have:
so contribution of to the objective before is the same as its contribution after.
In the final case, suppose there exists such that , so we shrink and then . Let index the data at the end of the iteration, so . Then we have , so satisfies the corresponding constraint for whether is in or . To compare the contribution of to the objective in and , we compute:
∎
Finally, we note that if IterativeRound terminates, then it is clear that no -, -, or non-negativity constraint is tight for by definition of our iterative operations. Thus it suffices to show that IterativeRound terminates in polynomial time.
Lemma B.4.
IterativeRound terminates in polynomial time.
Proof.
It suffices to show that the number of iterations of IterativeRound is polynomial. In each iteration, we make one of three actions. We either delete a facility from , move a client from to or shrink a -ball by one radius level for a client in .
We can delete each facility from at most once, so we make at most deletions. Each client can move from to at most once, because we never move clients back from to , so we do this operations at most times.
B.2. Analysis of ConfigReRoute
To prove Lemma 4.4, which bounds the number of fractional facilities needed to have a candidate configuration, we first prove a bound on the number of factional clients needed. The bound on the number of facilities will follow.
Lemma B.5.
Suppose satisfies all Basic Invariants, and let be an optimal extreme point of such that no -, -, or non-negativity constraint is tight. If , then there exist a candidate configuration in .
Proof.
We claim that in order for to have a candidate configuration, it suffices to have a chain of length at least four in . To see this, let be a chain of length at least four. Then , and by the Distinct Neighbors Property, either or .
We only consider the former case, because both cases are analogous. Thus, if , then we claim that forms a candidate configuration. We already have the first two properties of a candidate configuration. Now we verify the last two. Because and are part of a chain, we have and . Further, has neighbors and along the chain. By Proposition 2.7, each facility in is in at most two -balls for clients in . In particular, one of the facilities in is shared by and , and the other must be shared by and . Thus, each facility in is in exactly two -balls for clients in . An analogous argument holds for , so satisfies all properties of a candidate configuration, as required.
Now suppose . By Theorem 4.2, admits a chain decomposition into at most chains and a set of at most violating clients. Then at least of the clients in belong to the chains. By averaging, there must exist a chain with size at least , as required. ∎
Lemma 4.4 is a corollary of the above lemma.
Proof of Lemma 4.4.
By the previous lemma, it suffices to show that implies that . Applying Lemma 8.4, we have:
, which combined with gives the desired result. ∎
Proof of Theorem 4.5.
It is clear that ConfigReRoute can be implemented to run in polynomial time and maintains all Basic Invariants, because ConfigReRoute only moves clients from to . Thus it remains to show that ConfigReRoute weakly decreases .
Using the same strategy as in B.3, we let denote the LP at the beginning of ConfigReRoute and the optimal extreme point of . Then we show that is feasible after the operation and achieves the same objective value.
In this call to ConfigReRoute, we move some client from to . We have , so is feasible after ConfigReRoute. Finally, moving from to does not affect its contribution to the objective. ∎
B.3. Analysis of PseudoApproximation
Proof of Theorem 5.2.
It is immediate that PseudoApproximation maintains all Basic Invariants by Theorems 3.1 and 4.5. Further, both of these sub-routines are polynomial time, so to show that PseudoApproximation runs in polynomial time, it suffices to show that the number of calls to IterativeRound and ConfigReRoute is polynomial.
In every iteration of PseudoApproximation, either we terminate or we are guaranteed to move a client from to in ConfigReRoute. Each client can be removed from only polynomially many times, because each time a client is removed, in order to be re-added to , it must be the case that we shrunk the -ball of that client. However, by Basic Invariant 2.4(4), we can shrink each -ball only polynomially many times.
Finally, upon termination of PseudoApproximation, there is no candidate configuration, so Lemma 4.4 implies that has at most fractional variables. ∎
Appendix C Missing Proofs from §6
Proof of Lemma 6.7.
We let be the closest facility to in . We show that the cost of connecting to is at most . To do so, we partition into two sets of clients: those that are far from relative to , and those that are close to . In particular, let be a constant that we choose later. Then we partition into and , where:
and
First we bound the connection cost of to using the fact that , so Extra Invariant 6.4(2) says that . Thus we compute:
Now suppose . Fix any . Then for all , we have . It follows that . Our strategy is to use Extra Invariant 6.4(4):
for all . Thus we want . To lower bound with respect to , we use our assumption that there exists some such that , where by Extra Invariant 6.4(4). Thus we have:
Further, using the triangle inequality and the fact that is the closest facility to in , we have:
Combining these two inequalities gives the lower bound .
Now we are ready to choose . Recall that we want , so it suffices to choose such that:
Routine calculations show that we can take to satisfy this inequality. Now with this choice of , we can bound:
To conclude the proof, the connection cost of is at most and the connection cost of is at most . Summing these costs gives the desired result. ∎
Proof of Lemma 6.15.
We let be the closest facility to in . We show that the cost of connecting to is at most . To do so, we partition into two sets of clients: those that are far from relative to , and those that are close to . In particular, let be a constant that we choose later. Then we partition into and , where:
, and
First we bound the connection cost of to using the fact that , so Extra Invariant 6.12(2) says that is cheap. Thus we compute:
Now suppose . Importantly, all of these clients are within distance of , so we have . Our strategy to bound the connection cost of is to leverage Extra Invariant 6.12(4), so in particular we want to use the fact:
for any . We want to choose such that and for all . To see why this is useful, for such and , we have . Then we can bound:
Now we go back and specify our choice of and , which will allow us to complete the bound of the connection costs. First we lower bound in terms of for any . We recall that by assumption there exists some such that , where by Extra Invariant 6.12(4). Thus we have:
Further, using the triangle inequality and the fact that is the closest facility to in , we have:
Combining this inequality with the upper bound on gives that for all . Then we define . This gives us for all . Now we can choose satisfying:
Taking suffices.
Using these choices of and , we can bound:
, and
Summing these two costs gives the desired result. ∎
Proof of Theorem 6.2.
Let . We will later choose with respect to the given to obtain the desired approximation ratio and runtime. First, we choose parameters and for our pre-processing algorithm guaranteed by Theorem 6.5. We take and . We require that is an upper bound on . Using a standard binary search idea, we can guess up to a multiplicative -factor in time , so we guess such that .
With these choices of parameters, we run the algorithm guaranteed by Theorem 6.13 to obtain many sub-instances such that one such sub-instance is of the form , where for satisfies all Basic- and Extra Invariants, and we have:
(2) |
Then for each sub-instance output by the pre-processing, we run the algorithm guaranteed by Theorem 6.16 to obtain a solution to each sub-instance. Finally, out of these solutions, we output the one that is feasible for the whole instance with smallest cost. This completes the description of our approximation algorithm for -median with outliers. The runtime is , so it remains to bound the cost of the output solution and to choose the parameters and and .
To bound the cost, it suffices to consider the solution output on the instance where satisfies all Basic- and Extra Invariants and Equation 2. By running the algorithm guaranteed by Theorem 6.16 on this , we obtain a feasible solution to such that , and the cost of connecting clients from to is at most , where . To extend this solution on the sub-instance to a solution on the whole instance , we must connect clients from to . Because , applying Equation 1 allows us to upper bound the expected cost of connecting clients to by:
Now choosing to minimize (note that we ignore the terms), we obtain and . We can choose sufficiently large with respect to such that is sufficiently small to guarantee
Thus the expected cost of this solution is at most , where . Finally, by routine calculations, we can choose so that expected cost is at most , as required. Note that the runtime of our algorithm is . ∎
Appendix D Missing Proofs from §7: Analysis of OutliersPostProcess
In this section we present all missing proofs from the analysis of OutliersPostProcess and its subroutine ComputePartial.
D.1. Missing Proofs from Analysis of OutliersPostProcess
Proof of Lemma 7.2.
Without loss of generality, we may assume that no facilities in are co-located with each other, so is a disjoint family. This implies that is also a disjoint family. Now we construct a basis for . For every integral facility , we add the constraint to our basis. To complete the basis, we need to add further linearly independent tight constraints.
We recall that upon termination of PseudoApproximation, no -, -, or non-negativity constraint is tight for , so the only constraints we can choose are the -constraints, the -constraint, or the coverage constraint. We claim that we cannot add any -constraint to our basis, because such a constraint is supported only on integral facilities, whose constraints we already added to the basis. However, we can add every -constraint to our basis, because their supports are disjoint and they contain no integral facilities. Thus, our partial basis consists of all tight integrality constraints and all -constraints.
Now we consider adding the -constraint to our basis. Importantly, the -constraint is linearly independent with the current partial basis only if there exists at least one fractional facility not supported on any -ball for clients in . Further, we may assume the -constraint is tight (otherwise we cannot add it anyways), so there must be at least two fractional facilities not supported on any -ball for clients in
However, we note that each -ball for clients in contains at least two fractional facilities. Because these -balls are disjoint, we have . If we cannot add the -constraint to our basis, then we are done. This is because the coverage constraint is the only further constraint we can add the the basis, so we can bound . This implies implies using the previous inequality.
Otherwise, we add the -constraint to our basis, which implies because of the two fractional facilities outside and because the -constraint and coverage constraint contribute are the only further constraints we can add. Again combining these two inequalities gives . ∎
Proof of Lemma 7.3.
Let be the set of open facilities. It is immediate that . Further, satisfies all Extra Invariants, so . Because is integral, it is clear that we open . Thus it remains to show that the connecting clients to has cost at most .
It suffices to show that connecting and to is enough clients and achieves the desired cost. Because is integral and by definition of PseudoApproximation, we have that no -, -, or non-negativity constraint is tight for . It follows, for all and for all .
Then the coverage constraint of implies:
, so this solution connects enough clients.
To bound the cost, we compare the connection cost of each client with its contribution to the objective of . For all we have , so , which is exactly ’s contribution to .
For all , we note that , so ’s contribution to is exactly . We can apply 5.3 with and set of facilities to show that for all . To conclude, the connection cost of each client is at most times its contribution to , a required. ∎
Proof of Lemma 7.4.
Let be the output solution. First, note that because , and those are the only two fractional variables. Second, because , it must be the case that , because are the only fractional facilities, and by Extra Invariant 6.12 (1), there is one unit of open facility co-located at each . Note that this implies that .
Now there are two cases, either for some , or . Note that in either case, we close and open , so we still maintain the property that for all . Thus, can apply 5.3 with and set of facilities to show that for all .
We consider connecting the clients to . First, we show that this is at least clients. We observe that and are the only fractional facilities in , and no -constraint is tight. It follows that for all , we have , , or , so we can rewrite the coverage constraint as:
Then because and by assumption, we conclude that , as required.
Now it remains to show that the cost of connecting to plus the cost of connecting to is at most . First we handle . By assumption, , so by Extra Invariant 6.12(2), we can bound:
For the clients in that are not supported on , closing does not affect their connection cost; in particular, each such client either has an integral facility in its -ball to connect to (because we open and all other facilities are integral), or its -ball is empty, and there exists an integral facility within to connect to. In both cases, each client’s connection cost is at most times its contribution to the objective of .
The only remaining cost to bound is the clients in that are supported on . Let be these clients. We show that the cost of connecting all of to is at most using Lemma 6.15. Because every client in has an open facility in within distance , Lemma 6.15 is applicable to with set of facilities and .
To summarize, the connection costs of and are at most , and the connection cost of all remaining clients in that are not supported on is at most , so the total connection cost, which is the sum of these terms, it at most the desired bound. ∎
D.2. Missing Proofs from Analysis of ComputePartial
Proof of 7.7.
Assume for contradiction that there exists that is reached by the For loop, but does not remain in until termination. Note that cannot be removed from in the iteration that it is considered in the For loop. Thus there must exist a later iteration for client, say in which is removed from . In the iteration for client , there are only two possible ways that is removed from . Either or there exists a client such that intersects both and and .
In the former case, because we consider before , it must be the case that we removed from in ’s iteration. This is a contradiction. Similarly, in the second case if such a exists, then in ’s iteration, we either remove from or add to . In either case, this is a contradiction. ∎
Proof of 7.8.
By assumption, the input to ComputePartial, , satisfies all Basic and Extra Invariants. To obtain from , we delete some clients and facilities. Thus the only change to the - and -balls for clients in is that we possibly remove some facilities from their - and -balls; importantly, the radius levels, for all clients , remain the same. Thus, it is easy to see that satisfies all Basic Invariants.
Similarly, for all remaining clients , we have not changed or , so the only Extra Invariant that requires some care to verify is Extra Invariant 6.12(1). However, we recall that to obtain , we delete all clients in from the instance, so . This is because by the assumption that satisfies all Extra Invariants. ∎
Proof of 7.10.
We note that and . Then we have , , and .
It suffices to show that all -constraints, -constraint, the -constraint, and the coverage constraint are satisfied by restricted to .
Consider any . We observe that the for all and .Then for all .
Now for any , it suffices to show that we do not delete any copies of when going from from to , but this is immediate because , and we do not delete any facility from to obtain . Thus, every -constraint is satisfied.
For the -constraint, we have . We want to show . By definition , where the final equality follows because is a disjoint collection by Proposition 7.5. Then we compute:
, as required.
Finally, for the coverage constraint, we want to show:
where , , and and , so . Thus we can re-write the coverage constraint as:
Recall that the coverage constraint of implies:
By splitting this inequality into the contribution by and , we obtain:
, where in the final inequality we use the fact that for all . Now, we recall that for all , because . We can re-write:
To show that the coverage constraint is satisfied, it suffices to show:
To see this, observe that the first sum is over all clients in supported on some facility in but none in (otherwise these clients would be in .) The second sum is over all clients in supported on some facility in . Thus, recalling that , we have:
where in the final inequality we apply 7.9. ∎
Appendix E Missing Proofs from §8: Chain Decomposition
Proof of 8.4.
We construct a basis . First, for each integral facility , we add the integrality constraint to our basis. Thus we currently have constraints in our basis.
It remains to choose further linearly independent constraints to add to our basis. Note that we have already added all tight integrality constraints to our basis, and no non-negativity constraint is tight. Then the only remaining constraints we can add are the -constraints and the constraints of .
We claim that we cannot add any -constraints, because every -constraint is of the form for the unique integral facility . Note that here we used the fact that there is no facility that is set to zero. Thus every -constraint is linearly dependent with the tight integrality constraints, which we already chose.
It follows, the only possible constraints we can choose are the -constraints and the constraints of so:
∎
Proof of 8.5.
It suffices to upper bound the quantity by , because each client in contributes at least one to this sum, and every term of this sum is non-negative (because every -ball contains at least two facilities.)
Proof of 8.6.
Consider any , which is the vertex set of . We already established that is either a path or even cycle. In both cases, we can order such that is a path in the intersection graph. We verify that satisfies both properties of a chain.
Because , we have for all , and because is an edge in the intersection graph for all , we have for all . ∎
Proof of 8.7.
Because is a path, suppose such that . There are two cases to consider.
We first handle the degenerate case where there exists such that . Note that each -ball for clients in has size exactly two, so we have . Then both facilities in are already in exactly two -balls, so and can have no other neighbors in the intersection graph. This implies that is path of length two, so , such that . To finish this case, we note that , but , because both constraints and are the same.
In the second case, for all , we have . Using the fact that each facility is in at most two -balls (Proposition 2.7), we have that each non-leaf client on the path has two facilities in its -ball - one that it shares with the previous client on the path and one that it shares with the next. For the leaf clients, they share one facility with their single neighbor in the path, and they have one facility that is not shared with any other client in . With these observations, we can bound:
∎
Proof of 8.8.
Each -ball for clients along this cycle contains exactly two facilities. Using Proposition 2.7, each client along this cycle shares one of its facilities with the previous client in the cycle and one with the next client. This implies that each facility in is in exactly two -balls. Combining these two observations gives:
Thus, to prove , it suffices to show that . We do this by showing that the constraints are not linearly independent. By assumption, is bipartite with bipartition, say . Consider the linear combination of the constraints , where every constraint indexed by has coefficient and every constraint indexed by has coefficient . Then for every facility in , it is in exactly two -balls, and these two -balls must be on opposite sides of the bipartition, so each facility in has coefficient in this linear combination. In conclusion, we have constructed a non-trivial linear combination of the constraints whose left hand side is the zero vector, so . ∎
Appendix F Proof of Theorem 6.13: -Median with Outliers Pre-Processing
The goal of this section is to prove Theorem 6.13 using the relevant theorems from [KLS18]. Note that we follow exactly the same pre-processing steps; the only difference is that we summarize the results of their pre-processing in a single theorem.
The proof of the knapsack pre-processing, 6.5, follows analogously from the pre-processing steps in [KLS18] as well.
F.1. Preliminaries
We define the notions of extended instances and sparse extended instances for -median with outliers. These definitions are useful to capture the properties of our pre-processing.
Extended instances are used to handle the fact that in our pre-processing, we will guess some facilities to pre-open. Then is the set of guessed facilities.
Definition F.1 (Extended Instance for -Median with Outliers).
An extended instance for -median with outliers is of the form , where , , , , and are defined as in a standard -median with outliers instance (see Definition 6.11), and .
As in -median with outliers, the goal is to choose a set of at most open facilities and at least clients to serve to minimize the connection costs of the served clients to the open facilities, so . However, we add the additional constraint that the set of open facilities must include .
Further, sparse extended instances give our properties for what it means for the facilities and clients to be “cheap” (see the second and third properties in the next definition, respectively.)
Definition F.2 (Sparse Extended Instance for -Median with Outliers).
Let be an extended -median with outliers instance and , be parameters. We say that is -sparse with respect to solution if the following three properties hold:
-
(1)
the cost of the solution to is at most
-
(2)
for all , we have
-
(3)
for all , we have
F.2. Sparsification
In this section, we pass from the input -median with outliers instance to a sparse extended sub-instance by guessing the expensive parts of the input instance. Then on this sparse extended sub-instance, we can strengthen . The following theorems are directly from [KLS18], so we omit the proofs in this paper. The first theorem states that we can efficiently compute a sparse extended sub-instance at the cost of a small increase in approximation ratio.
Theorem F.3.
Let be an instance of -median with outliers with optimal solution and be parameters. Then there exists a -time algorithm that given , , , and an upper bound on the cost of the optimal solution 111Note that we are given , but not the solution , outputs -many extended -median with outliers instances of the form such that , , and . Further, one such instance is -sparse with respect to the solution and satisfies:
(3) |
Once we have our sparse extended sub-instance, say , we use these sparsity properties to compute the -vector, which is needed for our Extra Invariants.
Theorem F.4.
Let be an extended -median with outliers instance and and . Suppose is -sparse instance with respect to solution to such that has cost on . Then there exists a polynomial time algorithm that takes as input , , , and and outputs satisfying:
-
(1)
For every and , we have:
-
(2)
There exists a solution to of cost at most such that if client is connected to facility , then and for any facility , the total cost of clients connected to in this solution is at most
F.3. Putting it all Together: Proving Theorem 6.13
Combining the algorithms guaranteed by these above two theorems, we show how to construct with the desired properties.
Suppose we are given a -median with outliers instance , parameters , and an upper bound of . First we run the algorithm guaranteed by Theorem F.3 to obtain -many extended -median with outliers instances. Then for each instance, we run the algorithm guaranteed by Theorem F.4 to obtain a vector for each such instance.
By Theorem F.3, let be the instance output by the first algorithm such that is -sparse with respect to the solution and satisfies Equation 3. This sub-instance will be the one that is guaranteed by Theorem 6.13, so from here we need to compute the -vector, and construct with the desired properties.
Note that the cost of solution to is exactly . It follows, on this instance , the algorithm guaranteed by Theorem F.4 outputs a vector such that for every and , we have:
, and there exists a solution, say to of cost at most such that if is connected to facility , then and for any facility , the total cost of clients connected to in this solution is at most .
It remains to construct . To do so, first we construct a strengthened LP for the instance such that is feasible for the strengthened LP, which we call :
() | |||||
s.t. | |||||
The left column of constraints are the same as and the right column of constraints are extra constraints that are valid for the solution to our sub-instance . Because these constraints are valid for the solution , the following proposition is immediate.
Proposition F.5.
.
From here, we want to carry out a similar construction as in §2, where we construct satisfying all Basic Invariants from . We note that the main difference in our procedure here when compared to §2 is how we eliminate the -variables. To compute the -balls for , we must carefully duplicate facilities to capture the constraints: . We pass from to with the next lemma.
Lemma F.6.
There exists a polynomial time algorithm that takes as input , duplicates facilities in , and outputs a vector and sets for all such that:
-
(1)
for all
-
(2)
-
(3)
-
(4)
-
(5)
For all , there is one unit of open facility co-located with in
-
(6)
For every facility not co-located with a facility in , we have
Applying the algorithm guaranteed by the above lemma to , we can obtain the -sets. Using these -balls, we proceed similarly as in §2. Thus, next we randomly discretize the distances to powers of (up to a random offset) to obtain for all . Again, the possible discretized distances are for all , and satisfies Lemma 2.2.
Then we define the radius levels and inner balls in the exact same way, so:
To complete the data of for , we need to define the sets , , and . Here we must slightly modify the construction of §2 to accommodate the set of pre-opened facilities, . To satisfy Extra Invariant 6.12(1), we create a set of dummy clients such that for each , there exists a dummy client that is co-located with such that has radius level and consists of all co-located copies of . Thus, we define , , and .
This completes the description of for sub-instance . To complete our algorithm, we output each along with , , and .
To summarize, our algorithm is to first run the algorithm guaranteed by Theorem F.3 to obtain -many sub-instances. For each sub-instance, we compute using Theorem F.4, construct , construct the -balls using Lemma F.6, and define the rest of the data of as in §2. The runtime of our algorithm is immediate, so it suffices to show that one of the outputs has the desired properties.
In particular, we consider the sub-instance output by the algorithm guaranteed by Theorem F.3 such that is -sparse with respect to the solution and satisfies Equation 3. For the remainder of this section, we consider the constructed for this specific sub-instance. To complete the proof, we verify that satisfies the two desired properties.
Proposition F.7.
satisfies all Basic and Extra Invariants.
Proof.
It is easy to verify that satisfies all Basic Invariants by construction. For the Extra Invariants, we handle them one-by-one.
Proposition F.8.
Proof.
We first show that . We have , , and . Note that the dummy clients in contribute zero to the objective of , because they are co-located with one unit of open facility in their -balls. Thus Lemma F.6 implies that there exists a feasible solution to of cost at most up to the discretization of the distances. The cost of discretization is bounded by Lemma 2.2, and immediately gives the extra -factor. The factor of is due to the cost of discretization, which is bounded by Lemma 2.2.