This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Performance Guaranteed Evolutionary Algorithm for Minimum Connected Dominating Set

Chaojie Zhu1 Yingli Ran1  Zhao Zhang1 Ding-Zhu Du2
1 College of Mathematics and Computer Sciences, Zhejiang Normal University
Jinhua, Zhejiang, 321004, China
2 Department of Computer Science, University of Texas at Dallas
Richardson, TX, 75080, USA
Corresponding author: Zhao Zhang, [email protected]
Abstract

A connected dominating set is a widely adopted model for the virtual backbone of a wireless sensor network. In this paper, we design an evolutionary algorithm for the minimum connected dominating set problem (MinCDS), whose performance is theoretically guaranteed in terms of both computation time and approximation ratio. Given a connected graph G=(V,E)G=(V,E), a connected dominating set (CDS) is a subset CVC\subseteq V such that every vertex in VCV\setminus C has a neighbor in CC, and the subgraph of GG induced by CC is connected. The goal of MinCDS is to find a CDS of GG with the minimum cardinality. We show that our evolutionary algorithm can find a CDS in expected O(n3)O(n^{3}) time which approximates the optimal value within factor (2+lnΔ)(2+\ln\Delta), where nn and Δ\Delta are the number of vertices and the maximum degree of graph GG, respectively.

Keyword: evolutionary algorithm; minimum connected dominating set; expected running time; approximation ratio.

1 Introduction

Evolutionary algorithms have been widely used in real-world applications due to their simplicity of design and generality of implementation [23, 30]. Although they have been demonstrated to be quite effective and efficient by empirical studies, their theoretical analysis is far behind. From the beginning of the 21st century, theoretical studies on evolutionary algorithms have begun to come onto stage [28]. We are particularly interested in the application of evolutionary algorithms on NP-hard problems whose performance can be theoretically guaranteed. For example, in [8], Friedrich et al. designed a multi-objective evolutionary algorithm which can find an O(lnn)O(\ln n)-approximation for the minimum set cover problem (MinSC) in expected polynomial time. This ratio matches the lower bound for the approximability of MinSC. In [31], Yu et al. gave a general framework of an evolutionary algorithm, and when applied to the kk-MinSC problem in which every set has size at most kk, their algorithm can achieve in expected polynomial time the best known approximation ratio obtained in [14]. However, we found that a lot of problems do not fall into these frameworks, including the minimum connected dominating set problem (MinCDS) studied in this paper. The motivation of this paper is to theoretically explore the power of evolutionary algorithm using the MinCDS problem as a wrench, trying to reveal more underlying relationship between evolutionary algorithm and approximation algorithm.

Given a graph G=(V,E)G=(V,E), a vertex set CVC\subseteq V is a dominating set (DS) of GG if every vertex vVCv\in V\setminus C has a neighbor uCu\in C (vv is said to be dominated by uu, and uu is said to be a dominator of vv). If furthermore, the subgraph of GG induced by CC, denoted as G[C]G[C], is connected, then CC is said to be a connected dominating set (CDS). The goal of the minimum connected dominating set problem (MinCDS) is to find a CDS with the minimum cardinality.

MinCDS has attracted extensive studies in the past two decades, due to its fundamental application in a wireless sensor network (WSN). From the beginning of this century, WSNs have become ubiquitous in various fields such as environmental monitoring, traffic control, production process, health detection, etc [27]. Because sensors in a WSN have limited capacities, information is often forwarded through multi-hop transmissions. Notice that many sensors functioning at the same time may cause a large waste of energy, as well as a lot of interferences. So, researchers suggest using CDS as a virtual backbone in a WSN [2]: information collected by a non-backbone node can be forwarded to its dominator and then shared within the whole network since the backbone nodes form a connected structure.

The MinCDS problem is NP-hard [9]. In fact, it is set-cover hard [12] and thus does not admit a polynomial-time approximation within factor (1ε)lnn(1-\varepsilon)\ln n for any real number 0<ε<10<\varepsilon<1 unless P=NPP=NP [4]. On the other hand, (lnΔ+2)(\ln\Delta+2)-approximation algorithm exists for MinCDS [17], where Δ\Delta is the maximum degree of the graph. In view of its inapproximability, this ratio is asymptotically best possible under the assumption that PNPP\neq NP.

In this paper, we show that the same approximation ratio lnΔ+2\ln\Delta+2 can be achieved for MinCDS through a multi-objective evolutionary algorithm. The purpose of this paper is not to improve the approximation ratio since it is already the best possible. The aim is to understand the mechanism behind the evolutionary algorithm and its relation with the approximation algorithm.

1.1 Related Works

Evolutionary Algorithms (EAs) have their ideas originated in Darwin’s theory of biological evolution. Basically speaking, an EA iteratively generates a set of offspring and retains only those individuals which are superior. There are a lot of versions of EAs, some of the simplest ones include single-objective versions like RLS (randomized local search) and (1+1)(1+1)-EA [22] , multi-objective versions like SEMO and GSEMO [8].

The first origin of EAs can be dated to the early 1960s [19], and EAs have been applied to various areas such as antenna design [15], bioinformatics [18] and data mining [20]. In fact, since EAs are general-purpose algorithms, which can be implemented without much knowledge about the concrete instances, they are favorable in engineering fields, especially when the structures of the problems are hard to obtain. Empirical studies show that they perform fairly well in real-world applications. On the contrary, theoretical analysis on EAs is far behind.

Recently, a lot of progress has been made on the running time and performance ratio analysis of EAs for a lot of combinatorial optimization problems. For example, Neumann et al. [22] showed that (1+1)(1+1)-EA can find a minimum spanning tree (MST) in expected O(m2(logn+logwmax))O(m^{2}(\log n+\log w_{max})) iterations, where mm is the number of edges, nn is the number of vertices, and wmaxw_{max} is the maximum edge weight of the graph. Friedrich et al. [8] studied EAs for the minimum vertex cover problem (MinVC), which is a special case of the minimum set cover problem (MinSC). They showed that even on a graph as simple as a complete bipartite graph, with a positive probability, RLS cannot find a solution within approximation ratio ε/(1ε)\varepsilon/(1-\varepsilon) in finite time (and thus the approximation ratio can be arbitrarily bad), and it takes (1+1)(1+1)-EA exponential time to obtain an optimal solution. On the contrary, GSEMO can find a (lnn+1)(\ln n+1)-approximate solution for MinSC in expected O(n2m+mn(logm+logcmax))O(n^{2}m+mn(\log m+\log c_{max})) iterations, where nn is number of elements to be covered, mm is the number of sets, and cmaxc_{max} is the maximum cost of a set. These studies demonstrate the advantage of multi-objective EAs over single-objective EAs, and lead to a series of follow-up studies using multi-objective EAs for coverage problems [25, 26, 24]. In particular, Yu et al. [31] introduced a framework of an evolutionary algorithm, and when applied to the minimum kk-set cover problem (a MinSC problem in which every set has size at most kk) can obtain ratio Hkk18k9H_{k}-\frac{k-1}{8k^{9}}, where HkH_{k} is the kkth Harmonic number. However, there are a lot of combinatorial optimization problems (including the MinCDS problem studied in this paper) which do not fall into this framework.

There are also a lot of studies of evolutionary algorithms on various other combinatorial problems, including the maximum matching problem [11], the partition problem [29], the bi-objective pseudo-Boolean problem [10], makespan scheduling [32], shortest paths [5], etc. The reader may also refer to the monograph [23].

Although many works have been devoted to understanding the behavior of EAs from a theoretical point of view, there are large quantities of combinatorial problems which have not been sufficiently studied in view of approximations by EAs. It remains to be further explored which problems could be solved by EAs with a guaranteed theoretical performance. Finding underlying relationships between evolutionary algorithms and approximation algorithms is an ongoing challenge.

For the MinCDS problem, Guhn and Khuller [12] provided a two-stage greedy algorithm with approximation ratio at most HΔ+2H_{\Delta}+2, where Δ\Delta is the maximum degree of the graph (notice that lnΔHΔlnΔ+1\ln\Delta\leq H_{\Delta}\leq\ln\Delta+1). This ratio was improved to (2+lnΔ)(2+\ln\Delta) by Ruan et al. [17] using a one-stage greedy algorithm. The most challenging part is how to deal with a non-submodular potential function. Non-submodularity is also the barrier which prevents us to use existing studies on EAs for submodular optimization problems on the MinCDS problem. New insights have to be explored.

1.2 Our Contributions

In this paper, we design an evolutionary algorithm for the MinCDS problem, and show that it can find a CDS in expected O(n3)O(n^{3}) iterations which approximates the optimal value within factor lnΔ+2\ln\Delta+2. This ratio coincides with the best known ratio for MinCDS which is obtained by a centralized greedy algorithm. The main challenge lies in the theoretical analysis: can an evolutionary algorithm successfully simulate an approximation algorithm such that both efficiency and effectiveness can be guaranteed?

As shown in [8], a single-objective evolutionary algorithm seems hopeless in approximating NP-hard problems even on very simple examples. So, in this paper, we adopt a multi-objective method, trying to minimize two evaluation functions simultaneously. The first evaluation function f1f_{1} is used to measure how far the individual is from a feasible solution. A vertex set is a CDS if and only if its corresponding individual has f1f_{1}-value being 2. The second evaluation function f2f_{2} is the size of the vertex set corresponding to the individual, which is the objective of the MinCDS problem to be minimized. The algorithm follows the general framework of a multi-objective evolutionary algorithm: generating offspring randomly and keeping the Pareto front until some criterion is met.

Although the algorithm is simple, what is most challenging is how to analyze its performance theoretically. In the centralized algorithm designed by Ruan et al. in [17], there exists a “greedy path” (by adding vertices one by one greedily), which leads to a CDS with approximation ratio at most lnΔ+2\ln\Delta+2. If our algorithm can follow such a path in generating offspring, then we are done. The question is: does such a greedy path exist in the population generated by the algorithm? We think this might be the same question asked by Friedrich et al. in their work [8], in which it was shown that for the minimum set cover problem, allowing enough time, the population contains a path which is no worse than the greedy path generated by an approximation algorithm. Furthermore, Friedrich et al. designed a potential to “track” such a path. Their analysis heavily depends on the fact that the covering function is a submodular function. However, the MinCDS problem does not belong to the realm of submodular optimization, which makes the situation much more complicated.

To realize the above idea and overcome the above difficulties, we divide the analysis into two phases, and define two potentials to track a desirable path in each of these two phases. A challenge is how to coordinate these two potentials. For that purpose, a concept of good offspring is conceived and our analysis only tracks those good events.

The algorithm used in this paper is essentially SEMO or GSEMO. But the analysis is much more challenging. This is what we think most interesting: revealing deep theoretical mechanisms embedded in a simple algorithm.

The remaining part of this paper is organized as follows. Section 2 introduces some terminologies and concepts used in this paper. The algorithm is presented in Section 3. A theoretical analysis is given in Section 4. Experiments are done in Section 5, which are compared with greedy algorithm. Section 6 concludes the paper with some discussions on future work.

2 Preliminaries

In this paper, small bold letters (such as 𝐱,𝐲,𝐳{\bf x,y,z}) are used to represent vectors. Suppose G=(V,E)G=(V,E) is a graph with vertex set V={v1,,vn}V=\{v_{1},\ldots,v_{n}\} and edge set EE. Notice that there is a one-to-one correspondence between 2V2^{V} and {0,1}n\{0,1\}^{n}: for a vertex subset CVC\subseteq V, the vector 𝐱C{0,1}n{\bf x}_{C}\in\{0,1\}^{n} corresponding to CC has its bits indexed by v1,,vnv_{1},\ldots,v_{n} and the ii-th bit of 𝐱C{\bf x}_{C} is 11 if and only if viCv_{i}\in C; for a vector 𝐱{0,1}n{\bf x}\in\{0,1\}^{n}, the vertex subset C𝐱C_{\bf x} corresponding to 𝐱\bf x consists of all those vertices with bits 1 in 𝐱\bf x. Since this paper considers evolutionary algorithm, we call such a vector as an individual.

In a multi-objective optimization problem, different objectives might conflict with each other. To compare two individuals in such a setting, dominance is a commonly used terminology. In a multi-objective optimization problem in which tt objectives f1(𝐱),,ft(𝐱)f_{1}({\bf x}),\ldots,f_{t}({\bf x}) are to be minimized, an individual 𝐱{\bf x} is said to weakly dominate an individual 𝐱{\bf x}^{\prime}, or we say that 𝐱{\bf x} is weakly better than 𝐱{\bf x}^{\prime}, denoted as 𝐱𝐱{\bf x}\succeq{\bf x^{\prime}}, if fi(𝐱)fi(𝐱)f_{i}({\bf x})\leq f_{i}({\bf x^{\prime}}) holds for each i=1,,ti=1,\ldots,t. If furthermore, there is some ii with fi(𝐱)<fi(𝐱)f_{i}({\bf x})<f_{i}({\bf x^{\prime}}), then 𝐱{\bf x} is said to dominate 𝐱{\bf x^{\prime}}, or we say that 𝐱\bf x is better than 𝐱{\bf x}^{\prime}, denoted as 𝐱𝐱{\bf x}\succ{\bf x^{\prime}}. Notice that there exist individuals which are incomparable in the above sense: 𝐱\bf x might be superior to 𝐱\bf x^{\prime} in terms of fif_{i} but inferior to 𝐱\bf x^{\prime} in terms of fjf_{j}. Two individuals are said to be incomparable if no one is weakly better than the other. A Pareto front is a set of incomparable individuals.

Since the topic of this paper is dominating set, using terminology “dominate” in both the graph theoretic setting and the multi-objective optimization setting may incur confusion. So this paper only uses “dominate” in the graph theoretic setting, and uses “better” and “weakly better” in the multi-objective optimization setting.

There are a lot of methods to mutate an individual in evolutionary algorithms. We only consider two basic mutation methods: for a 0-11 vector of dimension nn, the first one flips exactly one bit uniformly at random; while in the second one, every bit is flipped independently with probability 1/n1/n.

In a simple single-objective evolutionary algorithm with objective function ff, every iteration creates one offspring 𝐱{\bf x}^{\prime} of current individual 𝐱\bf x and maintains the one in {𝐱,𝐱}\{{\bf x},{\bf x}^{\prime}\} having the better ff-value. If the first mutation is used, then it is known as random local search (RLS). If the second mutation is used, then it is known as (1+1)(1+1)-EA.

Different from RLS and (1+1)(1+1)-EA, a simple multi-objective evolutionary algorithm maintains the Pareto front PP of those individuals which have been created up to now, and in each iteration, an individual in PP is picked uniformly at random as the parent to create one offspring. In [21], such multi-objective evolutionary algorithms using the first mutation and the second mutation are called SEMO (simple evolutionary multi-objective optimizer) and GSEMO (global SEMO), respectively. As shown in [8], transforming a single-objective problem into a multi-objective problem and then using SEMO or GSEMO on it can significantly improve the performance of the solution.

The quality of an algorithm for an NP-hard problem is usually measured by its efficiency (that is, how long it takes to output a solution) and effectiveness (that is, how accurate the solution could be). We measure the running time by the expected number of mutations taken in the algorithm, and measure the accuracy by the approximation ratio. For a minimization problem, an algorithm is said to be an α\alpha-approximation algorithm if for any instance of the problem, it can output a solution whose cost is at most α\alpha times that of an optimal solution.

3 Algorithm for MinCDS

Given a graph G=(V,E)G=(V,E) on nn vertices, a vertex set CC can be represented by a vector 𝐱C{0,1}n{\bf x}_{C}\in\{0,1\}^{n}, and a vector 𝐱{0,1}n{\bf x}\in\{0,1\}^{n} corresponds to a vertex set C𝐱C_{\bf x}. Use |𝐱||{\bf x}| to denote the number of 1’s in 𝐱{\bf x}, which is also the number of vertices in vertex set C𝐱C_{\bf x}.

The algorithm is essentially SEMO, with two objective functions f1f_{1} and f2f_{2} to be minimized simultaneously. For a vector 𝐱{0,1}n{\bf x}\in\{0,1\}^{n}, f2(𝐱)=|𝐱|f_{2}({\bf x})=|{\bf x}|. So minimizing f2f_{2} is to minimize the size of the solution. Next, we introduce f1f_{1} which is used to measure the feasibility of the solution.

For a vertex subset CVC\subseteq V, let G[C]G[C] be the subgraph of GG induced by CC, and let p(C)p(C) be the number of connected components of G[C]G[C]. Denote by ECE_{C} the set of edges incident with CC, that is, every edge in ECE_{C} has at least one end in CC. The spanning subgraph of GG on vertex set VV and edge set ECE_{C} is denoted as GCG\langle C\rangle. Let q(C)q(C) be the number of connected components of GCG\langle C\rangle. Consider a path on five vertices as an example (see Fig. 1 (a)(a)). For vertex subset C={v1,v5}C=\{v_{1},v_{5}\}, graph G[C]G[C] consists of two singletons, namely v1v_{1} and v5v_{5}, and thus p(C)=2p(C)=2. While GCG\langle C\rangle is depicted in Fig. 1 (b)(b), and thus q(C)=3q(C)=3 (notice that GCG\langle C\rangle is a spanning subgraph of GG).

v1v_{1}v2v_{2}v3v_{3}v4v_{4}v5v_{5}(a)
v1v_{1}v2v_{2}v3v_{3}v4v_{4}v5v_{5}(b)
Figure 1: An illustration for the definition of p(C)p(C) and q(C)q(C).

In the following, we shall use a vector and its corresponding vertex set interchangeably. For a vector 𝐱{0,1}n{\bf x}\in\{0,1\}^{n}, let f1(𝐱)=p(𝐱)+q(𝐱)f_{1}({\bf x})=p({\bf x})+q({\bf x}). The next lemma shows how f1f_{1} can be used to measure the feasibility of the solution.

Lemma 3.1.

A vertex set CC is a CDS of GG if and only if f1(𝐱C)=2f_{1}({\bf x}_{C})=2.

Proof.

Except for \emptyset, every vertex set CC has p(C)1p(C)\geq 1 and q(C)1q(C)\geq 1. So, f1(C)2f_{1}(C)\geq 2 and equality holds if and only if p(C)=q(C)=1p(C)=q(C)=1. A CDS CC clearly satisfies p(C)=q(C)=1p(C)=q(C)=1 and thus f1(C)=2f_{1}(C)=2. On the other hand, if CC is not a dominating set, then there is a vertex vVCv\in V\setminus C which has no neighbor in CC. In this case, vv forms a connected component itself in GCG\langle C\rangle, and thus GCG\langle C\rangle has at least two connected components. So, q(C)=1q(C)=1 holds only when CC is a dominating set. Furthermore, the condition p(C)=1p(C)=1 is equivalent to saying that G[C]G[C] is connected. So a vertex set CC with p(C)=q(C)=1p(C)=q(C)=1 must be a CDS. ∎

Let 𝐱(0)=(0,0,,0){\bf x}^{(0)}=(0,0,\ldots,0) be the vector corresponding to vertex set \emptyset. Since GG\langle\emptyset\rangle consists of nn singletons, we have q()=nq(\emptyset)=n. Clearly, p()=0p(\emptyset)=0. So f1(𝐱(0))=nf_{1}({\bf x}^{(0)})=n while f2(𝐱(0))=0f_{2}({\bf x}^{(0)})=0. We study the MinCDS problem by considering the bi-objective minimization problem min𝐱{0,1}n{(f1(𝐱),f2(𝐱))}\min_{{\bf x}\in\{0,1\}^{n}}\{(f_{1}({\bf x}),f_{2}({\bf x}))\}.

The algorithm is described in Algorithm 1. It starts from 𝐱(0){\bf x}^{(0)}, and maintains a population set PP which keeps all those incomparable individuals found so far (namely Pareto front). In each iteration, the algorithm picks an individual 𝐱\bf x in PP uniformly at random, and generates an offspring 𝐱\bf x^{\prime} by flipping one bit of 𝐱\bf x uniformly at random. If there is no individual in PP which is better than 𝐱\bf x^{\prime}, then 𝐱\bf x^{\prime} enters PP, and all those individuals which are weakly worse than 𝐱\bf x^{\prime} are removed from PP. We shall show that after some polynomial number of iterations, the population PP contains a desired CDS. The output is a vertex set CC corresponding to an individual 𝐱C{\bf x}_{C} in the final population PP with the minimum f2f_{2}-value whose f1f_{1}-value equals 22 (so CC is a CDS).

Algorithm 1 Algorithm for MinCDS

Input: A connected graph GG, the number of iterations TT.

Output: A set CC which is a CDS of GG.

1:  P𝐱𝟎P\leftarrow{\bf x^{0}}.
2:  t=0t=0
3:  while tTt\leq T do
4:     choose 𝐱{\bf x} from PP uniformly at random.
5:     create 𝐱{\bf x^{\prime}} by flipping one bit of 𝐱\bf x uniformly at random.
6:     if 𝐳P\nexists{\bf z}\in P with 𝐳𝐱{\bf z}\succ{\bf x^{\prime}} then
7:        PP{𝐱}{𝐳:𝐱𝐳,𝐳P}P\leftarrow P\cup\{{\bf x^{\prime}}\}\setminus\{{\bf z}:{\bf x^{\prime}}\succeq{\bf z},{\bf z}\in P\}
8:     end if
9:     t=t+1t=t+1
10:  end while
11:  𝐱C=arg{f2(𝐱):𝐱P,f1(𝐱)=2}{\bf x}_{C}=\arg\{f_{2}({\bf x}):{\bf x}\in P,f_{1}({\bf x})=2\}
12:  return  Vertex set CC corresponding to 𝐱C{\bf x}_{C}.

4 The Analysis

The following lemma is crucial to the analysis of our algorithm, which shows that there is a vertex the addition of which may decrease the f1f_{1}-value by roughly a multiplicative factor. The proof is similar to the one in [17], we provide it here for the completeness of this paper.

Lemma 4.1.

For any vertex set CC in a connected graph G=(V,E)G=(V,E) with f1(C)>2f_{1}(C)>2, let vCv_{C} be the vertex of VCV\setminus C satisfying vC=argmaxvVC{f1(C)f1(C{v})}v_{C}=\arg\max_{v\in V\setminus C}\{f_{1}(C)-f_{1}(C\cup\{v\})\}, then we have the following two properties:

(i)(i) f1(C{vC})f1(C)1f_{1}(C\cup\{v_{C}\})\leq f_{1}(C)-1, and

(ii)(ii) f1(C{vC})(11m)f1(C)+2m+1\displaystyle f_{1}(C\cup\{v_{C}\})\leq\left(1-\frac{1}{m}\right)f_{1}(C)+\frac{2}{m}+1,
where mm is the size of a minimum CDS.

Proof.

(i)(i) We prove property (i)(i) by distinguishing two cases.

In the case q(C)>1q(C)>1, CC is not a dominating set by the proof of Lemma 3.1. By the connectedness of GG, there exists a vertex vVCv\in V\setminus C such that vv is not adjacent with CC and vv is adjacent with a vertex uN(C)u\in N(C) where N(C)N(C) is the neighbor set of CC. Then p(C{u})p(C)p(C\cup\{u\})\leq p(C) and q(C{u})<q(C)q(C\cup\{u\})<q(C).

In the case q(C)=1q(C)=1, since f1(C)>2f_{1}(C)>2, we have p(C)>1p(C)>1. So, G[C]G[C] has at least two connected components. By q(C)=1q(C)=1, it can be seen that the closest connected components of G[C]G[C] is exactly 2-hops away from each other, which implies that there is a vertex vVCv\in V\setminus C that is adjacent to at least two connected components of G[C]G[C]. In this case, p(C{v})<p(C)p(C\cup\{v\})<p(C) and q(C{v})=q(C)=1q(C\cup\{v\})=q(C)=1.

In any case, f1(C{v})<f1(C)f_{1}(C\cup\{v\})<f_{1}(C). By the choice of vCv_{C}, and because f1f_{1} is an integer-valued set function, we have f1(C{vC})f1(C{v})f1(C)1f_{1}(C\cup\{v_{C}\})\leq f_{1}(C\cup\{v\})\leq f_{1}(C)-1.

(ii)(ii) In order to prove (ii)(ii), we show that there is a vertex in an optimal solution the addition of which satisfies the claimed inequality.

Let CC^{*} be a minimum CDS of GG. Since G[C]G[C^{*}] is connected, we may order vertices in CC^{*} as v1,v2,,vmv_{1}^{*},v_{2}^{*},\ldots,v_{m}^{*} such that for any j=1,,mj=1,\ldots,m, the subgraph of GG induced by vertex set Cj={v1,v2,,vj}C_{j}^{*}=\{v_{1}^{*},v_{2}^{*},\ldots,v_{j}^{*}\} is connected. This can be done by finding a spanning tree of G[C]G[C^{*}] and removing leaf vertices one by one, ordering the vertices in the reverse order of their removal. Due to the connectedness of G[Cj1]G[C_{j-1}^{*}], it can be seen that

p(CCj1)p(CCj)(p(C)p(C{vj}))+1.p(C\cup C_{j-1}^{*})-p(C\cup C_{j}^{*})\leq\big{(}p(C)-p(C\cup\{v_{j}^{*}\})\big{)}+1.

Furthermore, it could be proved that function q-q is submodular and thus

q(CCj1)q(CCj)q(C)q(C{vj}).q(C\cup C_{j-1}^{*})-q(C\cup C_{j}^{*})\leq q(C)-q(C\cup\{v_{j}^{*}\}).

It follows that

f1(CCj1)f1(CCj)(f1(C)f1(C{vj}))+1.f_{1}(C\cup C_{j-1}^{*})-f_{1}(C\cup C_{j}^{*})\leq\big{(}f_{1}(C)-f_{1}(C\cup\{v_{j}^{*}\})\big{)}+1. (1)

Rewrite f1(C)f1(CC)f_{1}(C)-f_{1}(C\cup C^{*}) in the following way:

f1(C)f1(CC)=j=1m(f1(CCj1)f1(CCj)),f_{1}(C)-f_{1}(C\cup C^{*})=\sum\limits_{j=1}^{m}\big{(}f_{1}(C\cup C_{j-1}^{*})-f_{1}(C\cup C_{j}^{*})\big{)},

where C0=C_{0}^{*}=\emptyset. Making use of inequality (1), and the fact f(CC)=2f(C\cup C^{*})=2 (because CCC\cup C^{*} is a CDS), we have

f1(C)2j=1m(f1(C)f1(C{vj})+1).f_{1}(C)-2\leq\sum\limits_{j=1}^{m}\big{(}f_{1}(C)-f_{1}(C\cup\{v_{j}^{*}\})+1\big{)}.

Let vj0=argmaxvj{f1(C)f1(C{vj})}v_{j_{0}}^{*}=\arg\max_{v_{j}^{*}}\{f_{1}(C)-f_{1}(C\cup\{v_{j}^{*}\})\}. Then

f1(C)f1(C{vj0})+1f1(C)2m.f_{1}(C)-f_{1}(C\cup\{v_{j_{0}}^{*}\})+1\geq\frac{f_{1}(C)-2}{m}.

By the choice of vCv_{C}, we have

f1(C)f1(C{vC})+1f1(C)f1(C{vj0})+1f1(C)2m,f_{1}(C)-f_{1}(C\cup\{v_{C}\})+1\geq f_{1}(C)-f_{1}(C\cup\{v_{j_{0}}^{*}\})+1\geq\frac{f_{1}(C)-2}{m},

and thus f1(C{vC})(11m)f1(C)+2m+1f_{1}(C\cup\{v_{C}\})\leq\left(1-\frac{1}{m}\right)f_{1}(C)+\frac{2}{m}+1. Property (ii)(ii) is proved. ∎

The next theorem analyzes the expected time to obtain a CDS with a guaranteed approximation ratio.

Theorem 4.2.

In expected n(n1)(n2)n(n-1)(n-2) iterations, the output CC of Algorithm 1 is a CDS whose size is at most (2+lnΔ)(2+\ln\Delta) times of the optimal value, where Δ\Delta is the maximum degree of graph GG, nn is the number of vertices in GG.

Proof.

Let mm be the size of a minimum CDS of GG. Notice that although we use the optimal value mm in the analysis, the algorithm does not depend on mm.

For i=2,3,,ni=2,3,\ldots,n, let BiB_{i} be the bin which contains the individual in PP whose f1f_{1}-value equals ii. Notice that at any time, each bin contains at most one individual, this is because for any two individuals created in the evolutionary process, if they have the same f1f_{1}-value, then only the one with the smaller f2f_{2}-value is kept in PP. To track a “path” of desirable mutations, we only consider those bins containing good individuals, where an individual 𝐱\bf x is said to be “good” if it satisfies the following constraint:

f1(𝐱)(n2m)(11m)|𝐱|+m+2.f_{1}({\bf x})\leq(n-2-m)\left(1-\frac{1}{m}\right)^{|{\bf x}|}+m+2. (2)

Claim 1. Any good individual 𝐱{\bf x} with f1(𝐱)2m+2f_{1}({\bf x})\geq 2m+2 has |𝐱|mlnΔ|{\bf x}|\leq m\ln\Delta.

By the definition of good individual,

2m+2f1(𝐱)(n2m)(11m)|𝐱|+m+2.2m+2\leq f_{1}({\bf x})\leq(n-2-m)\left(1-\frac{1}{m}\right)^{|{\bf x}|}+m+2.

Using (11m)|𝐱|e|𝐱|m(1-\frac{1}{m})^{|{\bf x}|}\leq e^{-\frac{|{\bf x}|}{m}}, it can be derived that

|𝐱|mln(n2mm).|{\bf x}|\leq m\ln\left(\frac{n-2-m}{m}\right).

Since every vertex can dominate at most Δ+1\Delta+1 vertices, we have n(Δ+1)mn\leq(\Delta+1)m, and thus n2mmΔ\frac{n-2-m}{m}\leq\Delta. Hence |𝐱|mlnΔ|{\bf x}|\leq m\ln\Delta, Claim 1 is proved.

Claim 2. Suppose 𝐱\bf x is a good individual. Then, an individual 𝐱{\bf x}^{\prime} with f1(𝐱)f1(𝐱)f_{1}({\bf x}^{\prime})\leq f_{1}({\bf x}) and |𝐱||𝐱||{\bf x}^{\prime}|\leq|{\bf x}| is also good.

This is because

f1(𝐱)\displaystyle f_{1}({\bf x}^{\prime}) f1(𝐱)(n2m)(11m)|𝐱|+m+2\displaystyle\leq f_{1}({\bf x})\leq(n-2-m)\left(1-\frac{1}{m}\right)^{|{\bf x}|}+m+2
(n2m)(11m)|𝐱|+m+2.\displaystyle\leq(n-2-m)\left(1-\frac{1}{m}\right)^{|{\bf x}^{\prime}|}+m+2. (3)

Claim 3. Suppose 𝐱\bf x is a good individual, let 𝐱{\bf x}^{\prime} be the vector corresponding to vertex set C𝐱{vC𝐱}C_{\bf x}\cup\{v_{C_{\bf x}}\}, where vC𝐱=argmaxvVC𝐱{f1(C𝐱)f1(C𝐱{v})}v_{C_{{\bf x}}}=\arg\max_{v\in V\setminus C_{{\bf x}}}\{f_{1}(C_{{\bf x}})-f_{1}(C_{{\bf x}}\cup\{v\})\}. Then

(2.1)(2.1) |𝐱|=|𝐱|+1|{\bf x}^{\prime}|=|{\bf x}|+1,

(2.2)(2.2) f1(𝐱)f1(𝐱)1f_{1}({\bf x^{\prime}})\leq f_{1}({\bf x})-1, and

(2.3)(2.3) 𝐱{\bf x}^{\prime} is also a good individual.

Property (2.1)(2.1) is obvious, and property (2.2)(2.2) is a consequence of Lemma 4.1 (i)(i). Next, we show property (2.3)(2.3). By Lemma 4.1 (ii)(ii),

f1(C𝐱{vC𝐱})(11m)f1(𝐱)+2m+1.f_{1}(C_{{\bf x}}\cup\{v_{C_{{\bf x}}}\})\leq\left(1-\frac{1}{m}\right)f_{1}({\bf x})+\frac{2}{m}+1.

Since 𝐱\bf x is good and |𝐱|=|𝐱|+1|{\bf x}^{\prime}|=|{\bf x}|+1, we have

f1(𝐱)\displaystyle f_{1}({\bf x^{\prime}}) (11m)((n2m)(11m)|𝐱|+m+2)+2m+1\displaystyle\leq\left(1-\frac{1}{m}\right)\left((n-2-m)\left(1-\frac{1}{m}\right)^{|{\bf x}|}+m+2\right)+\frac{2}{m}+1
=(n2m)(11m)|𝐱|+m+2.\displaystyle=(n-2-m)\left(1-\frac{1}{m}\right)^{|{\bf x^{\prime}}|}+m+2.

Hence 𝐱\bf x^{\prime} is also a good individual.

An individual 𝐱{\bf x}^{\prime} generated by the way described in Claim 2 is desirable, for convenience, we call such 𝐱{\bf x}^{\prime} to be the best offspring of 𝐱\bf x.

Next, we define an indicator JiJ_{i} for each bin BiB_{i} such that individuals contained in bins with indicator 1 are all good.

Initially, only Jn=1J_{n}=1 and all other Ji=0J_{i}=0 for i=2,3,,n1i=2,3,\ldots,n-1.

When we say that bin BiB_{i} is going to be filled with an individual 𝐱{\bf x}^{\prime}, it refers to the setting that the newly generated individual 𝐱{\bf x}^{\prime} can be put into BiB_{i}, either because BiB_{i} is empty and i=f1(𝐱)i=f_{1}({\bf x}^{\prime}), or because 𝐱{\bf x}^{\prime} can replace the individual 𝐱{\bf x} in BiB_{i} due to f1(𝐱)=f1(𝐱)=if_{1}({\bf x}^{\prime})=f_{1}({\bf x})=i and |𝐱||𝐱||{\bf x}^{\prime}|\leq|{\bf x}|.

When bin BiB_{i} is going to be filled with an individual 𝐱{\bf x}^{\prime}, its indicator JiJ_{i} is changed from 0 to 1 if one of the following two good events occur.

(i)(i) There is an individual 𝐱¯\bar{\bf x} belonging to a bin BjB_{j} with indicator Jj=1J_{j}=1 such that its best offspring 𝐱¯\bar{\bf x}^{\prime} is weakly worse than 𝐱{\bf x}^{\prime}, that is

f1(𝐱¯)f1(𝐱)and|𝐱¯||𝐱|.f_{1}(\bar{\bf x}^{\prime})\geq f_{1}({\bf x}^{\prime})\ \mbox{and}\ |\bar{\bf x}^{\prime}|\geq|{\bf x}^{\prime}|. (4)

(ii)(ii) There is an individual 𝐱¯\bar{\bf x} belonging to a bin BjB_{j} with indicator Jj=1J_{j}=1 such that

f1(𝐱¯)>f1(𝐱)and|𝐱¯||𝐱|.f_{1}(\bar{\bf x})>f_{1}({\bf x}^{\prime})\ \mbox{and}\ |\bar{\bf x}|\geq|{\bf x}^{\prime}|. (5)

It is possible that indicator JiJ_{i} can be changed from 1 to 0: if the individual contained in bin BiB_{i} is removed from the population because of the creation of a better individual, then JiJ_{i} is set to be 0.

Claim 4. Any bin BiB_{i} with Ji=1J_{i}=1 contains a good individual.

We first show that if current BiB_{i} contains a good individual, then as long as BiB_{i} continues to be non-empty, the individual contained in BiB_{i} is always good. In fact, suppose 𝐱\bf x is a good individual in BiB_{i}, it can be replaced by another individual 𝐱{\bf x}^{\prime} only when f1(𝐱)=f1(𝐱)f_{1}({\bf x^{\prime}})=f_{1}({\bf x}) and |𝐱||𝐱||{\bf x^{\prime}}|\leq|{\bf x}|. Then by Claim 2, 𝐱\bf x^{\prime} is also good.

So, as long as we can show the following property:

the time when JiJ_{i} changes from 0 to 1, the individual entering BiB_{i} is good, (6)

then all individuals contained in BiB_{i} during the span when Ji=1J_{i}=1 will be good. Next, we use induction on the process of assigning indicator 1’s to prove (6).

Initially, only Jn=1J_{n}=1, and Bn={𝐱(0)}B_{n}=\{{\bf x}^{(0)}\}. Notice that |𝐱𝟎|=0|{\bf x^{0}}|=0 and f1(𝐱𝟎)=n=(n2m)(11m)0+m+2f_{1}({\bf x^{0}})=n=(n-2-m)(1-\frac{1}{m})^{0}+m+2, so 𝐱(0){\bf x}^{(0)} is a good individual.

Suppose the claim is true for all bins with indicator 1 in current population, and after some mutation, JiJ_{i} is to be changed from 0 to 1. The first case for such a change is because of the existence of 𝐱,𝐱¯,𝐱¯{\bf x}^{\prime},\bar{\bf x},\bar{\bf x}^{\prime} satisfying (4). By induction hypothesis, 𝐱¯\bar{\bf x} is good. By Claim 3, 𝐱¯\bar{\bf x}^{\prime} is also good. Then by condition (4) and Claim 2, 𝐱{\bf x}^{\prime} is also good. The second case for such a change is because of the existence of 𝐱,𝐱¯{\bf x}^{\prime},\bar{\bf x} satisfying (5). By induction hypothesis, 𝐱¯\bar{\bf x} is good, and thus by Claim 2, 𝐱{\bf x}^{\prime} is good. Claim 4 is proved.

The following analysis is divided into two phases. For each phase, a potential will be used to track the progress of the algorithm.

Define a potential II with respect to current population PP to be the integer

I=argmin{i:Ji=1}.I=\arg\min\{i\colon J_{i}=1\}. (7)

The initial population P={𝐱(0)}P=\{{\bf x}^{(0)}\} has potential I=nI=n.

Claim 5. Potential II will never increase, and the expected time for II to be decreased by at least 1 is at most n(n1)n(n-1).

Suppose 𝐱¯\bar{\bf x} is the individual corresponding to current potential II. As long as 𝐱¯\bar{\bf x} remains in PP, the value of II will not increase. By the algorithm, 𝐱¯\bar{\bf x} can be removed from PP only when a newly generated individual 𝐱{\bf x^{\prime}} satisfies f1(𝐱)f1(𝐱¯)f_{1}({\bf x^{\prime}})\leq f_{1}(\bar{\bf x}) and |𝐱||𝐱¯||{\bf x^{\prime}}|\leq|\bar{\bf x}|. If f1(𝐱)=f1(𝐱¯)f_{1}({\bf x}^{\prime})=f_{1}(\bar{\bf x}), then 𝐱¯\bar{\bf x} is replaced by 𝐱{\bf x}^{\prime} and II remains the same. If f1(𝐱)<f1(𝐱¯)f_{1}({\bf x}^{\prime})<f_{1}(\bar{\bf x}), then by (5), the birth of 𝐱{\bf x}^{\prime} makes JiJ_{i^{\prime}} to change from 0 to 1, where i=f1(𝐱)<f1(𝐱¯)=ii^{\prime}=f_{1}({\bf x}^{\prime})<f_{1}(\bar{\bf x})=i. In this case, the new I=i<II^{\prime}=i^{\prime}<I. In any case, II will not increase.

Next, we consider the expected time to reduce II. If the mutation picks 𝐱¯\bar{\bf x} and generates its best offspring 𝐱¯\bar{\bf x}^{\prime}, then JiJ_{i} can be changed from 0 to 1, where i=f1(𝐱¯)f1(𝐱¯)1<Ii=f_{1}(\bar{\bf x}^{\prime})\leq f_{1}(\bar{\bf x})-1<I (notice that before the mutation, JiJ_{i} must be 0 since II is the smallest index for the bin with indicator 1). Notice that the probability of choosing 𝐱¯\bar{\bf x} in line 4 of Algorithm 1 is 1|P|\frac{1}{|P|}, and the probability of obtaining its best offspring 𝐱¯\bar{\bf x}^{\prime} in line 5 of Algorithm 1 is 1n\frac{1}{n}. So, the probability

Pr[I is reduced]1n|P|1n(n1),Pr[\mbox{$I$ is reduced}]\geq\frac{1}{n|P|}\geq\frac{1}{n(n-1)}, (8)

where |P|n1|P|\leq n-1 is used since f1f_{1} has at most n1n-1 distinct values, namely 2,3,,n2,3,\ldots,n, and no solutions having the same f1f_{1}-value can coexist in PP by line 7 of the algorithm.

Notice that the derivation of (8) is valid starting from “any” individual in the population achieving current potential, even when the above 𝐱¯\bar{\bf x} is removed from PP due to the entering of a better individual. So, the expected number of mutations for potential II to be decreased by at least 11 is at most n(n1)n(n-1). Claim 5 is proved.

By Claim 5, in expected n(n1)(n2)n(n-1)(n-2) iterations, II could be reduced from nn to 22. However, merely tracking potential II could not guarantee that the final solution has small size. In fact, since f1(V)=2f_{1}(V)=2, the vector 𝐱(1)=(1,1,,1){\bf x}^{(1)}=(1,1,\ldots,1) satisfies (2), and thus VV is also a good individual, whose size is clearly too large. From the algorithmic point of view, when f1(𝐱)f_{1}({\bf x}) is small, the decreasing speed of f1f_{1} implied by Lemma 4.1 (ii)(ii) will slow down. So, we shall track another potential after the f1f_{1}-value becomes smaller than 2m+22m+2. The details are given as follows.

If n<2m+2n<2m+2, then the approximation ratio 2+lnΔ2+\ln\Delta holds for the trivial solution VV. So, we assume n2m+2n\geq 2m+2 in the following. Since the initial value for II is n2m+2n\geq 2m+2, and the final value for II is 2<2m+22<2m+2, there must be a time when II jumps over 2m+22m+2. That is, there exists a mutation, before which potential I~2m+2\widetilde{I}\geq 2m+2, after which potential I~2m+1\widetilde{I}^{\prime}\leq 2m+1.

Define a second potential KK as follows.

K=argmin𝐱P{f1(𝐱):|𝐱|+f1(𝐱)mlnΔ+2m+2}.K=\arg\min\limits_{{\bf x}\in P}\{f_{1}({\bf x}):|{\bf x}|+f_{1}({\bf x})\leq m\ln\Delta+2m+2\}. (9)

Claim 6. Let 𝐱~\widetilde{\bf x}^{\prime} be the individual corresponding to I~\widetilde{I}^{\prime}. Then 𝐱~\widetilde{\bf x}^{\prime} satisfies the constraint in (9). Hence from the time when 𝐱~\widetilde{\bf x}^{\prime} enters population PP, potential KK is well-defined, and Kf1(𝐱~)=I~2m+1K\leq f_{1}(\widetilde{\bf x}^{\prime})=\widetilde{I}^{\prime}\leq 2m+1.

The core part for the proof of Claim 6 is to show that 𝐱~\widetilde{\bf x}^{\prime} satisfies the constraint in (9). Since we are considering the time when potential I~2m+2\widetilde{I}\geq 2m+2 is jumped down to I~=f1(𝐱~)2m+1\widetilde{I}^{\prime}=f_{1}(\widetilde{\bf x}^{\prime})\leq 2m+1, the creation of 𝐱~\widetilde{\bf x}^{\prime} incurs the change of JI~J_{\widetilde{I}^{\prime}} from 0 to 1. The reason for such a change is because of the existence of a bin BjB_{j} with Jj=1J_{j}=1 and the individual 𝐱¯\bar{\bf x} in BjB_{j} satisfying constraint (4) or constraint (5) (with 𝐱{\bf x}^{\prime} replaced by 𝐱~\widetilde{\bf x}^{\prime}), that is,

f1(𝐱¯)f1(𝐱~)=I~and|𝐱¯||𝐱~|f_{1}(\bar{\bf x}^{\prime})\geq f_{1}(\widetilde{\bf x}^{\prime})=\widetilde{I}^{\prime}\ \mbox{and}\ |\bar{\bf x}^{\prime}|\geq|\widetilde{\bf x}^{\prime}| (10)

or

f1(𝐱¯)>f1(𝐱~)and|𝐱¯||𝐱~|.f_{1}(\bar{\bf x})>f_{1}(\widetilde{\bf x}^{\prime})\ \mbox{and}\ |\bar{\bf x}|\geq|\widetilde{\bf x}^{\prime}|. (11)

Since Jj=1J_{j}=1, by Claim 4, 𝐱¯\bar{\bf x} is good. Since I~\widetilde{I} is the smallest index of bins with indicator 1 before the mutation, we have jI~2m+2j\geq\widetilde{I}\geq 2m+2. So 𝐱¯\bar{\bf x} is a good individual with f1(𝐱¯)=j2m+2f_{1}(\bar{\bf x})=j\geq 2m+2. Then by Claim 1,

|𝐱¯|mlnΔ.|\bar{\bf x}|\leq m\ln\Delta. (12)

Since 𝐱¯\bar{\bf x}^{\prime} is the best offspring of 𝐱¯\bar{\bf x}, we have

|𝐱¯|=|𝐱¯|+1mlnΔ+1.|\bar{\bf x}^{\prime}|=|\bar{\bf x}|+1\leq m\ln\Delta+1.

So, in the first case, namely (10), we have

|𝐱~|+f1(𝐱~)|𝐱¯|+I~(mlnΔ+1)+(2m+1)=mlnΔ+2m+2.|\widetilde{\bf x}^{\prime}|+f_{1}(\widetilde{\bf x}^{\prime})\leq|\bar{\bf x}^{\prime}|+\widetilde{I}^{\prime}\leq(m\ln\Delta+1)+(2m+1)=m\ln\Delta+2m+2.

The second case (11) is easier by using (12). Claim 6 is proved.

Claim 7. The potential KK never increases and the expected time for KK to be decreased by at least 1 is at most n(n1)n(n-1).

To prove the monotonicity of KK, similar to the proof of Claim 5, it suffices to consider a solution 𝐱¯\bar{\bf x} corresponding to current KK, and the case that 𝐱¯\bar{\bf x} is removed from PP. In such a case, a newly generated solution 𝐱{\bf x^{\prime}} satisfies f1(𝐱)f1(𝐱¯)f_{1}({\bf x^{\prime}})\leq f_{1}(\bar{\bf x}) and |𝐱||𝐱¯||{\bf x^{\prime}}|\leq|\bar{\bf x}|. Such 𝐱{\bf x^{\prime}} satisfies |𝐱|+f1(𝐱)|𝐱¯|+f1(𝐱¯)mlnΔ+2m+2|{\bf x^{\prime}}|+f_{1}({\bf x^{\prime}})\leq|\bar{\bf x}|+f_{1}(\bar{\bf x})\leq m\ln\Delta+2m+2. So 𝐱\bf x^{\prime} also satisfies the constraint in (9), and thus replacing 𝐱¯\bar{\bf x} by 𝐱{\bf x^{\prime}} will not increase KK.

To prove the second part of Claim 7, observe that if a mutation picks 𝐱¯\bar{\bf x} which corresponds to current KK and generates its best offspring 𝐱¯\bar{\bf x}^{\prime}, then by Claim 3, f1(𝐱¯)f1(𝐱¯)1<Kf_{1}(\bar{\bf x}^{\prime})\leq f_{1}(\bar{\bf x})-1<K and |𝐱¯|=|𝐱¯|+1|\bar{\bf x}^{\prime}|=|\bar{\bf x}|+1. It follows that |𝐱¯|+f1(𝐱¯)|𝐱¯|+f1(𝐱¯)mlnΔ+2m+2|\bar{\bf x}^{\prime}|+f_{1}(\bar{\bf x}^{\prime})\leq|\bar{\bf x}|+f_{1}(\bar{\bf x})\leq m\ln\Delta+2m+2 and thus 𝐱¯\bar{\bf x}^{\prime} also satisfies the constraint in (9). Hence the new potential Kf1(𝐱¯)<KK^{\prime}\leq f_{1}(\bar{\bf x}^{\prime})<K. The expected time for the occurrence of such an event is at most n(n1)n(n-1), as in the proof of Claim 5. So, Claim 7 is proved.

Claim 8. Starting from initial individual 𝐱(0){\bf x}^{(0)}, the expected time for KK reaching 2 is at most n(n1)(n2)n(n-1)(n-2).

By Claim 5, the expected time for potential II to decrease from f1(𝐱(0))=nf_{1}({\bf x}^{(0)})=n to I~=f1(𝐱~)\widetilde{I}^{\prime}=f_{1}(\widetilde{\bf x}^{\prime}) is at most (nI~)n(n1)\big{(}n-\widetilde{I}^{\prime}\big{)}n(n-1). Then by Claim 7, potential KK becomes well-defined and keeps decreasing, and the the expected time for KK to decrease from I~=f1(𝐱~)\widetilde{I}^{\prime}=f_{1}(\widetilde{\bf x}^{\prime}) to 2 is at most (I~2)n(n1)(\widetilde{I}^{\prime}-2)n(n-1). Adding them together, the total expected time for KK reaching 2 is at most n(n1)(n2)n(n-1)(n-2).

Now, we are ready to finish the proof of the theorem. Suppose 𝐱A{\bf x}^{A} is the individual corresponding to K=2K=2, then f1(𝐱A)=2f_{1}({\bf x}^{A})=2 implies that the vertex set corresponding to 𝐱A{\bf x}^{A} is a CDS. Combining this with f1(𝐱A)+|𝐱A|mlnΔ+2m+2f_{1}({\bf x}^{A})+|{\bf x}^{A}|\leq m\ln\Delta+2m+2 implies that |𝐱A|(2+lnΔ)m|{\bf x}^{A}|\leq(2+\ln\Delta)m, that is, 𝐱A{\bf x}^{A} approximates the optimal solution within a factor at most 2+lnΔ2+\ln\Delta. Furthermore, Claim 8 shows that the expected time to obtain such 𝐱A{\bf x}^{A} is at most n(n1)(n2)n(n-1)(n-2). ∎

Remark 4.3.

Although our algorithm is essentially SEMO, using GSEMO will also work. Notice that for the second mutation method, the probability that exactly one bit is flipped is (n1)1n(11n)n11e{n\choose 1}\cdot\frac{1}{n}\cdot(1-\frac{1}{n})^{n-1}\geq\frac{1}{e}. Furthermore, by the way how we define “good events” to change indicators from 0 to 1, generating an individual having more bits different from its parent does not affect the validity of the analysis. So, in expected en(n1)(n2)en(n-1)(n-2) iterations, GSEMO can find a (lnΔ+2)(\ln\Delta+2)-approximation for MinCDS.

5 Experiments

In this section, we experiment on the efficiency and the effectiveness of Algorithm 1, comparing it with the centralized greedy algorithm in [17]. We first carry out experiments on scale-free networks generated by Barabási-Albert (BA) model [1]: starting from a ring network with four vertices, at each step we add a new vertex with two edges incident to the existing vertices, following the growth and preferential attachment rule. Then we consider the Erdős-Rényi (ER) network [6]: a random graph with nn vertices such that any pair of vertices are adjacent with probability pp. Since the MinCDS problem is studied on connected graphs, to generate a connected graph, we set p=lognnp=\frac{\log n}{n} [7]. The size of the graph is expressed by the number of vertices.

All experiments are coded in Matlab and run on the identical configuration: Intel(R) Core(TM) i5-4210 CPU and 4GB of RAM.

In terms of time, the greedy algorithm is much faster. This is natural, just like the comparatively slow evolution of human being versus fast customized mechanized production. The main purpose of this section is to test the effectiveness of the evolutionary algorithm, which can be achieved within the estimated time guaranteed by the theoretical analysis, and further explore empirical potential for an improvement of time.

We first study the effect of our algorithm on BA networks. The results are shown in Fig. 2. By setting TT to be T1=n(n1)(n2)T_{1}=n(n-1)(n-2), where nn is the number of vertices of the input graph, it can be seen from Fig. 2(a) that the size of the solutions output by our algorithm are no worse than those of the greedy algorithm. Interestingly, in many cases, the output of our algorithm is better than that of the greedy algorithm. The reason might lie in the observation that greedy is a deterministic method, which can no longer improve the solution after the algorithm is completed; but for evolutionary algorithm, if more time is allowed, there is still some space for improvement.

Then, we set TT to be T2=n2T_{2}=n^{2}. As shown by Fig. 2(b), the accuracy is not decreased compared with T1=n(n1)(n2)T_{1}=n(n-1)(n-2), which implies that the empirical performance of EA might be much better than what is guaranteed by the theory. This is because our analysis only tracks one evolutionary path having a clearly-cut pattern, which may have underestimated the power of evolution through other paths. We also tried to set TT to be nlognn\log n, and it turns out that in this case, the output of the algorithm can not be always guaranteed to be a feasible solution.

Refer to caption
(a) EA versus Greedy
Refer to caption
(b) T1=n(n1)(n2)T_{1}=n(n-1)(n-2) versus T2=n2T_{2}=n^{2}
Figure 2: Experiments on BA networks.

Then, we consider the performance of Algorithm 1 on ER networks. The results are shown in Fig. 3. Similar phenomena are observed for the comparison of EA and greedy (see Fig. 3(a). However, divergence appears between T1=n(n1)(n2)T_{1}=n(n-1)(n-2) and T2=n2T_{2}=n^{2}, as shown by Fig. 3(b). The reason for this phenomenon might be the following: the power-law property possessed by BA networks already makes the solutions good enough. For example, it was shown in [3] that the minimum positive dominating set problem cannot be approximated within factor (1ε)ln(max{Δ,n})(1-\varepsilon)\ln(\max\{\Delta,\sqrt{n}\}) on a general graph, but can be approximated within a constant factor on a power-law graph. We also tried T3=n2lognT_{3}=n^{2}\log n, the computed sizes coincide with those of T1T_{1}. So, although the performance of our algorithm on ER is not as good as on BA, there is also potential to improve the time empirically.

Refer to caption
(a) EA versus Greedy
Refer to caption
(b) T1=n(n1)(n2)T_{1}=n(n-1)(n-2) versus T2=n2T_{2}=n^{2}
Figure 3: Experiments on ER networks.

6 Conclusion and Discussion

This paper shows that a multi-objective evolutionary algorithm can achieve in expected O(n3)O(n^{3}) time an approximation ratio lnΔ+2\ln\Delta+2 for the MinCDS problem, which is the best known ratio that was obtained by a centralized approximation algorithm. The evolutionary algorithm is very simple and easy to implement. However, the theoretical analysis is much more complicated. The most challenging task is how to “track” an evolutionary path which is no worse than a greedy path. Our experiments show that a desired approximate solution can indeed be found in the expected time estimated by the theoretical analysis. What is more, EA has the potential to improve the solution if more time resources are available.

Unfortunately, our method could not readily deal with the minimum weight connected dominating set problem (MinWCDS). In fact, the currently best known approximation ratio for MinWCDS is 1.35lnn1.35\ln n, obtained by Guha and Khuller in [13]. This algorithm employs a generalization of an ingenious idea called spider decomposition which was first invented to study the minimum node-weighted Steiner tree problem [16]. However, how to simulate a spider decomposition by an evolutionary algorithm is unclear.

It is interesting to find out more combinatorial optimization problems which can be effectively approximated by an evolutionary algorithm, and try to answer the following question: for which combinatorial optimization problems can evolutionary algorithms achieve a theoretically guaranteed performance, and why?

Acknowledgements

This research is supported by NSFC (U20A2068), ZJNSFC (LD19A010001), and NSF (1907472).

References

  • [1] Albert László Barabási and Réka Albert. Emergence of scaling in random networks. Science, 286(5439):509–512, 1999.
  • [2] Bevan Das and Vaduvur Bharghavan. Routing in ad-hoc networks using minimum connected dominating sets. In IEEE International Conference on Communications, ICC 97, Towards the Knowledge Millennium, Montreal, pages 376–380, 1997.
  • [3] Thang N. Dinh, Yilin Shen, Dung T. Nguyen, and My T. Thai. On the approximability of positive influence dominating set in social networks. Journal of Combinatorial Optimization, 27(3):487–503, 2014.
  • [4] Irit Dinur and David Steurer. Analytical approach to parallel repetition. In Forty-sixth Acm Symposium on Theory of Computing, 2014.
  • [5] Benjamin Doerr, Edda Happ, and Christian Klein. Tight analysis of the (1+1)-ea for the single source shortest path problem. Evolutionary Computation, 19(4):673–691, 2011.
  • [6] Paul Erdős and Alfréd Rényi. On random graph 1. Publicationes Mathematicate, 6:290–297, 1959.
  • [7] Paul Erdős and Alfréd Rényi. On the evolution of random graphs. Publications of the Mathematical Institute of the Hungarian Academy of Sciences, 5:17–61, 1960.
  • [8] Tobias Friedrich, Jun He, Nils Hebbinghaus, Frank Neumann, and Carsten Witt. Approximating covering problems by randomized search heuristics using multi-objective models. Evolutionary Computation, 18(4):617–633, 2010.
  • [9] Michael R. Garey and David S. Johnson. Computers and intractability: A guide to the theory of np-completeness. Freeman, San Francisco, 1978.
  • [10] Oliver Giel. Expected runtimes of a simple multi-objective evolutionary algorithm. In The 2003 Congress on Evolutionary Computation, CEC’03, 2004.
  • [11] Oliver Giel and Ingo Wegener. Evolutionary algorithms and the maximum matching problem. Information Processing Letters, 15(82):14–19, 2003.
  • [12] Sudipto Guha and Samir Khuller. Approximation algorithms for connected dominating sets. Algorithmica, 20(4):374–387, 1998.
  • [13] Sudipto Guha and Samir Khuller. Improved methods for approximating node weighted steiner trees and connected dominating sets. Information and Computation, 150(1):57–74, 1998.
  • [14] Refael Hassin and Asaf Levin. A better-than-greedy approximation algorithm for the minimum set cover problem. SIAM Journal on Computing, 35(1):189–200, 2005.
  • [15] Greg Hornby, Al Globus, Derek S. Linden, and Jason Lohn. Automated antenna design with evolutionary algorithms. In Proceedings of the 2006 American Institute of Aeronautics and Astronautics Conference on Space, pages 19–21, 2006.
  • [16] Philip Klein and Ramamoorthi Ravi. A nearly best-possible approximation algorithm for node-weighted steiner trees. Journal of Algorithms, 19(1):104–115, 1995.
  • [17] Ruan Lu, Hongwei Du, Xiaohua Jia, Weili Wu, and Ker I. Ko. A greedy approximation for minimum connected dominating sets. Theoretical Computer Science, 329(1-3):325–330, 2004.
  • [18] Patrick C. H. Ma, Keith C. C. Chan, Yao Xin, and David K. Y. Chiu. An evolutionary clustering algorithm for gene expression microarray data analysis. IEEE Transactions on Evolutionary Computation, 10(3):296–314, 2006.
  • [19] Tomassini Marco. Evolutionary algorithms. Studies in Computational Intelligence, 6(5):443–462, 1996.
  • [20] Anirban Mukhopadhyay, Ujjwal Maulik, Santu Bandyopadhyay, and Carlos Artemio Coello Coello. A survey of multiobjective evolutionary algorithms for data mining: Part i. IEEE transactions on evolutionary computation, 18(1):4–19, 2014.
  • [21] Frank Neumann. Expected runtimes of a simple evolutionary algorithm for the multi-objective minimum spanning tree problem. European Journal of Operational Research, 181(3):1620–1629, 2007.
  • [22] Frank Neumann and Ingo Wegener. Randomized local search, evolutionary algorithms, and the minimum spanning tree problem. Theoretical Computer Science, 378(1):32–40, 2007.
  • [23] Frank Neumann and Carsten Witt. Adelaide research and scholarship: Bioinspired computation in combinatorial optimization : algorithms and their computational complexity. 2010.
  • [24] Chao Qian, Jing Cheng Shi, Yang Yu, Ke Tang, and Zhi Hua Zhou. Parallel pareto optimization for subset selection. Proceedings of the 25th International Joint Conference on Artificial Intelligence (IJCAI’16), pages 1939–1945, 2016.
  • [25] Chao Qian, Yang Yu, and Zhi Hua Zhou. On constrained boolean pareto optimization. In Proceedings of the 24th International Joint Conference on Artificial Intelligence (IJCAI’15), pages 389–395, 2015.
  • [26] Chao Qian, Yang Yu, and Zhi Hua Zhou. Subset selection by pareto optimization. Advances in Neural Information Processing Systems 28 (NIPS’15), pages 1765–1773, 2015.
  • [27] Peng Jun Wan, Khaled Muheiddin Alzoubi, and Ophir Frieder. Distributed construction of connected dominating set in wireless ad hoc networks. Mobile Networks and Applications, 9(2):141–149, 2004.
  • [28] Ingo Wegener. Theoretical aspects of evolutionary algorithms. Lecture Notes in Computer Science, 2076:64–78, 2001.
  • [29] Carsten Witt. Worst-case and average-case approximations by simple randomized search heuristics. In Conference on Theoretical Aspects of Computer Science, 2005.
  • [30] Yao Xin. Unpacking and understanding evolutionary algorithms. In IEEE World Congress on Computational Intelligence, 2012.
  • [31] Yang Yu, Xin Yao, and Zhi Hua Zhou. On the approximation ability of evolutionary optimization with application to minimum set cover. Artificial Intelligence, 180-181:20–33, 2012.
  • [32] Yuren Zhou, Jun Zhang, and Yong Wang. Performance analysis of the (1+1) evolutionary algorithm for the multiprocessor scheduling problem. Algorithmica, 73(1):21–41, 2015.