This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

On the Robustness, Connectivity and Giant Component Size of Random K-out Graphs

Eray Can Elumar, , Mansi Sood, , and Osman Yağan Parts of the material was presented at the 2021 IEEE International Symposium on Information Theory (ISIT) [1], and the 2023 IEEE International Conference on Communications (ICC) [2]. This work was supported in part by the Office of Naval Research (ONR) through Grant N00014-21-1-2547, by the CyLab IoT Initiative, and by the National Science Foundation through Grant CCF-1617934.Eray Can Elumar is with the Department of Electrical and Computer Engineering, Carnegie Mellon University, Pittsburgh, PA 15213 USA (e-mail: [email protected]). Mansi Sood is with the Department of Electrical and Computer Engineering, Carnegie Mellon University, Pittsburgh, PA 15213 USA (e-mail: [email protected]).Osman Yağan is with the Department of Electrical and Computer Engineering and CyLab, Carnegie Mellon University, Pittsburgh, PA 15213 USA (e-mail: [email protected]).
Abstract

Random K-out graphs are garnering interest in designing distributed systems including secure sensor networks, anonymous crypto-currency networks, and differentially-private decentralized learning. In these security-critical applications, it is important to model and analyze the resilience of the network to node failures and adversarial captures. Motivated by this, we analyze how the connectivity properties of random K-out graphs vary with the network parameters KK, the number of nodes (nn), and the number of nodes that get failed or compromised (γn\gamma_{n}). In particular, we study the conditions for achieving connectivity with high probability and for the existence of a giant component with formal guarantees on the size of the largest connected component in terms of the parameters n,Kn,~{}K, and γn\gamma_{n}. Next, we analyze the property of rr-robustness which is a stronger property than connectivity and leads to resilient consensus in the presence of malicious nodes. We derive conditions on KK and nn under which the random K-out graph achieves r-robustness with high probability. We also provide extensive numerical simulations and compare our results on random K-out graphs with known results on Erdős-Rényi (ER) graphs.

Index Terms:
Connectivity, giant component, robustness, r-robustness, random graphs, random K-out graphs, security, privacy

I Introduction

I-A Motivation and Background

In recent years, the rapid proliferation of affordable sensing and computing devices has led to rapid growth in technologies powered by the IoT (Internet of Things). A key challenge in this space is to develop network models for generating a securely connected ad-hoc network in a distributed fashion while minimizing operational costs.

With its unique connectivity properties, a class of random graph models known as the random KK-out graphs has found many applications in the design of ad-hoc networks. A random K-out graph [3, 4, 5], denoted as (n;K)\mathbb{H}(n;K), is an undirected graph with nn nodes where each node forms an edge with KK distinct nodes chosen uniformly at random. Random K-out graphs are known to achieve connectivity easily, i.e., with far fewer edges (O(n))(O(n)) as compared to classical random graph models including Erdős-Rényi (ER) graphs [6, 3], random geometric graphs [7], and random key graphs [8], which all require O(nlogn)O(n\log n) edges for connectivity. In particular, it is known [5, 4] that random K-out graphs are connected with high probability (whp) when K2K\geq 2. This had led to their deployment in several applications including random key predistribution schemes for secure communication in sensor networks [9, 10, 11], differentially-private distributed averaging algorithms [12], anonymity preserving cryptocurrency networks [13], and distributed secure mapping of network addresses in next-generation internet architectures [14].

In the context of sensor networks, random K-out graphs have been used [5, 11, 15] to analyze the performance of the random pairwise key predistribution scheme and its heterogeneous variants [16, 17]. The random pairwise scheme works as follows. Before deployment, each sensor chooses KK others uniformly at random. A unique pairwise key is given to each node pair where at least one of them selects the other. After deployment, two sensors can securely communicate if they share a pairwise key. The topology of the sensor network can thus be represented by a random K-out graph; each edge of the random K-out represents a secure communication link between two sensors. Consequently, random K-out graphs have been analyzed to answer key questions on the values of the parameters nn and KK needed to achieve certain desired properties, including connectivity at the time of deployment [4, 5], connectivity under link removals [11, 15], and unassailability [18]. Despite many prior works on random K-out graphs, very little is known about its connectivity properties when some of its nodes are removed. This is an increasingly relevant problem since many IoT networks are deployed in remote and hostile environments where nodes may be captured by an adversary, or fail due to harsh conditions.

Another application of random K-out graphs is in distributed learning, where a key goal is to perform computations on user data without compromising the privacy of the users. Random K-out graphs have recently been used to construct the communication graph in a differentially-private federated averaging scheme called the GOPA (GOssip Noise for Private Averaging) protocol [12, Algorithm 1]. According to the GOPA protocol, a random K-out graph is constructed on a set of nodes, of which an unknown subset is dishonest. It was shown in [12, Theorem 3] that the privacy-utility trade-offs achieved by the GOPA protocol are tightly dependent on the subgraph on honest nodes being connected. When the subgraph on honest nodes is not connected, it was shown that the performance of GOPA is tied to the size of the connected components of honest nodes. Since dishonest users can be modeled as randomly deleted nodes, analyzing the connectivity and giant component size of random K-out graphs under node deletions is the key in understanding the performance of the GOPA protocol.

I-B Properties of Interest for Random K-out Graphs in Distributed Computing Applications

In the context of applications discussed in the previous section, we can identify several key properties of random K-out graphs that need to be well-understood for performance evaluation and efficient design of the underlying systems. We believe that the graph properties discussed here can also be useful in facilitating new applications of random K-out graphs in different fields, akin to our recent work [19] paving the way to new applications by establishing connectivity guarantees in the finite node regime.

A key metric in quantifying the utility of a network is connectivity which is defined as the existence of a path of edges between every node pair [20]. Connectivity ensures that all agents in the network can communicate with one another and no node is isolated from the network. In practice, resource constraints limit the number of links that can be established in the network. Thus, a key goal is to design a resiliently connected network while keeping the number of links to be established within operational constraints. Depending on the resource constraints and mission requirements of the application at hand, it may suffice to ensure a weaker notion of connectivity or in cases where agents may routinely fail or compromise, we may even need a stronger notion of connectivity.

In resource-constrained environments, preserving connectivity despite node failures may not be feasible and the goal instead might be to ensure that there is a large enough, connected sub-network of users, also known as a giant component. For example, it may suffice to aggregate the temperature readings of the majority of sensors deployed in a field to get an estimate of the true temperature. Another example is the power grid network where it is essential to ensure supply to the majority of the users in the event of failures.

In addition to ensuring that the network remains resiliently connected in the event of node failures, it is often desirable to ensure that consensus can be achieved even in the presence of adversarial agents. In [21], it was shown that network connectivity is not sufficient to characterize consensus when nodes use a certain class of local filtering rules. In particular, it was shown that consensus can be reached in graphs that have the property of being sufficiently robust. This is formally quantified by the property of rr-robustness, which was introduced in [21]. A graph is said to be rr-robust if, for every disjoint subset pair that partitions the graph, at least one node in one of these subsets is adjacent to at least rr nodes in the other set.

The rr-robustness property is especially useful in applications of consensus dynamics, where parameters of several agents get aligned after a sufficiently long period of local interactions. In another example, it was shown [22] that if the network is (2F+1)(2F+1)-robust (for some non-negative integer FF), then the nodes in the network can reach consensus even when there are up to FF malicious nodes in the neighborhood of every correctly-behaving node. Thus, rr-robustness is particularly important for applications based on consensus dynamics in adversarial environments. Moreover, rr-robustness is known [22] to be a stronger property than rr-connectivity and thus can provide guarantees on the connectivity of the graph when up to r1r-1 nodes in the graph are removed. Due to this importance, rr-robustness has been studied in various graph models in the random graph literature, such as the ER graph and the Barabási-Albert model [23]. rr-robustness has been studied in [24] in our prior work; and the results presented here improve upon those results by providing a sharper threshold for rr-robustness.

I-C Main Contributions

With these motivations in mind, this paper aims to fill the gaps in the literature on the connectivity and robustness properties of random K-out graphs. We provide a comprehensive set of results on the connectivity and size of the giant component of the random K-out graph when some of its nodes are dishonest, have failed, or have been captured. We further analyze the conditions required for ensuring rr-robustness of the random K-out graph. Our main contributions are summarized below:

  1. 1.

    Let (n;Kn,γn)\mathbb{H}(n;K_{n},\gamma_{n}) denote the random graph obtained after removing γn\gamma_{n} nodes, selected uniformly at random, from the random K-out graph (n;Kn)\mathbb{H}(n;K_{n}). We provide a set of conditions for KnK_{n}, nn, and γn\gamma_{n} under which (n;Kn,γn)\mathbb{H}(n;K_{n},\gamma_{n}) is connected with high probability (whp). This is done for both cases where γn=Ω(n)\gamma_{n}=\Omega(n) and γn=o(n)\gamma_{n}=o(n), respectively. Our result for γn=Ω(n)\gamma_{n}=\Omega(n) (see Theorem III.1) significantly improves a prior result [25] on the same problem and leads to a sharp zero-one law for the connectivity of the random K-out graph under node deletions. Our result for the case γn=o(n)\gamma_{n}=o({n}) (see Theorem III.2) expands the existing threshold of Kn2K_{n}\geq 2 required for connectivity by showing that the graph is connected whp for Kn2K_{n}\geq 2 even when o(n)o(\sqrt{n}) nodes are deleted.

  2. 2.

    We derive conditions on KnK_{n}, nn, γn\gamma_{n} that lead to a giant component in (n;Kn,γn)\mathbb{H}(n;K_{n},\gamma_{n}) whp and provide an upper bound on the number of nodes not contained in the giant component. This is also done for both cases γn=Ω(n)\gamma_{n}=\Omega(n) and γn=o(n)\gamma_{n}=o(n); see Theorem III.3 and Theorem III.4, respectively. An important consequence of this result is to establish Kn2K_{n}\geq 2 as a sufficient condition to ensure whp the existence of a giant component in the random K-out graph despite the removal of o(n)o(n) nodes in the network.

  3. 3.

    Using a novel proof technique, we show that K2rK\geq 2r when r2r\geq 2 is sufficient to ensure that the random K-out graph (n;K)\mathbb{H}(n;K) is rr-robust whp (see Theorem III.5). Since it is already known that (n;K)\mathbb{H}(n;K) is not rr-robust whp when K<rK<r, this result is tight up to at most a multiplicative factor of two (and it is is much tighter than the condition established in [24]).

  4. 4.

    We also provide a comparison of random K-out graphs with ER graphs. We determine that random K-out graphs are much more robust in terms of rr-robustness property, and also attain connectivity and admit a giant component with fewer edges compared to ER graphs.

  5. 5.

    Combining our theoretical results with numerical simulations, we highlight the usefulness of random K-out graphs as a topology design tool for efficient design of secure, resilient and robust distributed networks.

I-D Organization of the Paper

The rest of this article is organized as follows. In Section II, we introduce the notation used across this article and the network model, the random K-out graph, and extend this model to account for node deletions. In Section III, we present the main results along with the simulation results and provide a detailed discussion. In Section IV, we provide the proof of all Theorems presented in Section III. Conclusions are provided in Section V.

II Notations and Definitions

All random variables are defined on the same probability space (Ω,,)(\Omega,{\mathcal{F}},\mathbb{P}) and probabilistic statements are given with respect to the probability measure \mathbb{P}. The complement of an event AA is denoted by AcA^{\rm c}. The cardinality of a discrete set AA is denoted by |A||A|. The intersection of events AA and BB is denoted byABA\cap B. We refer to any mapping K:00K:\mathbb{N}_{0}\rightarrow\mathbb{N}_{0} as a scaling if it satisfies the condition 2Kn<n,n=2,3,2\leq K_{n}<n,\quad n=2,3,\ldots. All limits are understood with nn going to infinity. If the probability of an event tends to one as nn\rightarrow\infty, we say that it occurs with high probability (whp). The statements an=o(bn)a_{n}={o}(b_{n}), an=ω(bn)a_{n}=\omega(b_{n}), an=O(bn)a_{n}={O}(b_{n}), an=Θ(bn)a_{n}=\Theta(b_{n}), and an=Ω(bn)a_{n}=\Omega(b_{n}), used when comparing the asymptotic behavior of sequences {an},{bn}\{a_{n}\},\{b_{n}\}, have their meaning in the standard Landau notation. The asymptotic equivalence anbna_{n}\sim b_{n} is used to denote the fact that limnanbn=1\lim_{n\to\infty}\frac{a_{n}}{b_{n}}=1. Finally, we let d\langle d\rangle denote the mean node degree of a graph.

Definition II.1 (Random K-out Graph)

The random K-out graph is defined on the vertex set V:={v1,,vn}V:=\{v_{1},\ldots,v_{n}\} as follows. Let 𝒩:={1,2,,n}\mathcal{N}:=\{1,2,\dots,n\} denote the set vertex labels. For each i𝒩i\in\mathcal{N}, let Γn,i𝒩i\Gamma_{n,i}\subseteq\mathcal{N}\setminus i denote the set of KnK_{n} labels, selected uniformly at random, corresponding to the nodes selected by viv_{i}. It is assumed that Γn,1,,Γn,n\Gamma_{n,1},\ldots,\Gamma_{n,n} are mutually independent. Distinct nodes viv_{i} and vjv_{j} are adjacent, denoted by vivjv_{i}\sim v_{j} if at least one of them picks the other. Namely,

vivjif[jΓn,i][iΓn,j].\displaystyle v_{i}\sim v_{j}~{}~{}\quad\textrm{if}~{}~{}~{}\quad[j\in\Gamma_{n,i}]~{}\vee~{}[i\in\Gamma_{n,j}].\vspace{-1mm} (1)

The set of neighbors of node ii is denoted by 𝒱i:={j𝒩i:vivj}\mathcal{V}_{i}:=\{j\in\mathcal{N}\setminus i:v_{i}\sim v_{j}\}, and the degree of node ii is denoted as di=|𝒱i|d_{i}=|\mathcal{V}_{i}|. The random graph defined on the vertex set VV through the adjacency condition (1) is called a random K-out graph [26, 3, 5] and is denoted by (n;Kn)\mathbb{H}(n;K_{n}).

Definition II.2 (Cut)

[27, Definition 6.3] For a graph 𝒢\mathcal{G} defined on the node set 𝒩\mathcal{N}, a cut is a non-empty subset S𝒩S\subset\mathcal{N} of nodes isolated from the rest of the graph. Namely, S𝒩S\subset\mathcal{N} is a cut if there is no edge between SS and Sc=𝒩SS^{\rm c}=\mathcal{N}\setminus S. If SS is a cut, then so is ScS^{\rm c}.

Definition II.3 (Connected Components)

A pair of nodes in a graph 𝔾\mathbb{G} are said to be connected if there exists a path of edges connecting them. A connected component CiC_{i} of 𝔾\mathbb{G} is a subgraph in which any two vertices are connected to each other, and no vertex is connected to a node outside of CiC_{i}.

Definition II.4 (Giant Component)

For a graph 𝔾\mathbb{G} with nn nodes, a giant component exists if its largest connected component has size Ω(n)\Omega(n). In that case, the largest connected component is referred to as the giant component of the graph.

Definition II.5 (Connectivity)

A graph 𝔾\mathbb{G} is connected if there exists a path of edges between every pair of its vertices.

Definition II.6 (rr-connectivity)

A graph is rr-connected if it remains connected after the removal of any set of r1r-1 (or, fewer) nodes or edges.

Definition II.7 (rr-reachable Set)

[21, Definition 6] For a graph 𝔾\mathbb{G} and a subset SS of nodes S𝒩S\subset\mathcal{N}, we say SS is rr-reachable if \existsiS:|𝒱iS|r\in S:|\mathcal{V}_{i}\setminus S|\geq r, where r+r\in\mathbb{Z}^{+}. In other words, SS is an rr-reachable set if it contains a node that has at least rr neighbors outside SS.

Refer to caption
Figure 1: An example for a 1-robust graph. We see that with the subset pair S={vA,vB}S=\{v_{A},v_{B}\} and Sc={vC,vD,vE,vF}S^{c}=\{v_{C},v_{D},v_{E},v_{F}\}, both vAv_{A} and vBv_{B} in SS have only one neighbor in ScS^{c}, while vCv_{C} and vFv_{F} in ScS^{c} have only one neighbor in SS, meaning both SS and ScS^{c} are 11-reachable (but not 22-reachable). Further, it can be seen that all other subset pairs that partition the graph are also at least 11-reachable, leading to the graph in Fig. 1 being 11-robust.
Definition II.8 (rr-robust Graph)

[21, Definition 6] A graph 𝔾\mathbb{G} is rr-robust if for every pair of nonempty, disjoint subsets of 𝒩\mathcal{N}, at least one of these subset pairs is rr-reachable, where r+r\in\mathbb{Z}^{+}.

It was shown in [22] that if a graph is rr-robust, it is at least rr-connected. Thus, rr-robustness is a stronger property than rr-connectivity. It is also easy to see that when r=1r=1, the properties of rr-robustness and rr-connectivity are equivalent.

A main goal of this paper is to study the connectivity and giant component size of random K-out graphs when some of its nodes are failed, captured, or dishonest. To this end, we also consider the following model of random K-out graphs under random removal of nodes. We first let γn\gamma_{n} denote the number of removed nodes and assume, for simplicity, that they are selected uniformly at random among all nodes in VV. Further, we let DV,|D|=γnD\subset V,\ |D|=\gamma_{n} denote the set of deleted nodes. We then define (n;Kn,γn)\mathbb{H}(n;K_{n},\gamma_{n}) on the vertex set R=VDR=V\setminus D and the corresponding set of labels 𝒩R\mathcal{N}_{R}, such that distinct vertices viv_{i} and vjv_{j} (both in RR) are adjacent if they were adjacent in (n;Kn)\mathbb{H}(n;K_{n}); i.e., if [jΓn,i][iΓn,j][j\in\Gamma_{n,i}]~{}\vee~{}[i\in\Gamma_{n,j}]. For each i𝒩Ri\in\mathcal{N}_{R}, the set of labels adjacent to node viv_{i} in (n;Kn,γn)\mathbb{H}(n;K_{n},\gamma_{n}) is denoted by Γnγn,i𝒩Ri\Gamma_{n-\gamma_{n},i}\subseteq\mathcal{N}_{R}\setminus i.

III Main Results

Our main results are presented in Theorems III.1- III.5 below. Each Theorem addresses a design question as to how the parameter KnK_{n} should be chosen to satisfy the desired property on robustness, connectivity or the size of the giant component; see Table I for a summary of the main results. The results on connectivity and the size of the giant component are for (n;Kn,γn)\mathbb{H}(n;K_{n},\gamma_{n}), i.e., the random K-out graph when γn\gamma_{n} nodes are deleted, while the result on rr-robustness is given for the original graph (n;Kn)\mathbb{H}(n;K_{n}) (without any node deletion). We provide the proofs of all results in Section IV.

III-A Results on Connectivity

Let P(n,Kn,γn)=[(n;Kn,γn) is connected]P(n,K_{n},\gamma_{n})=\mathbb{P}\left[\mathbb{H}(n;K_{n},\gamma_{n})\text{ is connected}\right].

Theorem III.1

Let γn=αn\gamma_{n}=\alpha n with α\alpha in (0,1)(0,1), and consider a scaling K:00K:\mathbb{N}_{0}\to\mathbb{N}_{0} such that with c>0c>0 we have

Kncr1(α,n),wherer1(α,n)=logn1αlogα\displaystyle K_{n}\sim c\cdot r_{1}(\alpha,n),\ \ \textrm{where}\quad r_{1}(\alpha,n)=\frac{\log n}{1-\alpha-\log\alpha} (2)

is the threshold function. Then, we have

limnP(n,Kn,γn)={1,ifc>10,if0<c<1.\displaystyle\lim_{n\to\infty}P(n,K_{n},\gamma_{n})=\begin{cases}1,&\mathrm{if}\quad c>1\\ 0,&\mathrm{if}\quad 0<c<1.\end{cases} (3)

The proof of the one-law in (3), i.e., that limnP(n,Kn,γn)=1\lim_{n\to\infty}P(n,K_{n},\gamma_{n})=1 if c>1c>1, is given in Section IV. The zero-law of (3), i.e., that limnP(n,Kn,γn)=0\lim_{n\to\infty}P(n,K_{n},\gamma_{n})=0 if c<1c<1, was established previously in [25, Corollary 3.3]. There, a one-law was also provided: under (2), it was shown that limnP(n,Kn,γn)\lim_{n\to\infty}P(n,K_{n},\gamma_{n}) if c>11αc>\frac{1}{1-\alpha}, leaving a gap between the thresholds of the zero-law and the one-law. Theorem III.1 presented here fills this gap by establishing a tighter one-law, and constitutes a sharp zero-one law; e.g., when α=0.5\alpha=0.5, the one-law in [25] is given with c>2c>2, while we show that it suffices to have c>1c>1.

Theorem III.2

Consider a scaling K:00K:\mathbb{N}_{0}\to\mathbb{N}_{0}.
a) If γn=o(n)\gamma_{n}=o(\sqrt{n}), then we have

limnP(n,Kn,γn)=1,ifKn2n\displaystyle\lim_{n\to\infty}P(n,K_{n},\gamma_{n})=1,\quad\mathrm{if}\quad K_{n}\geq 2\ \ \forall n (4)

b) If γn=Ω(n)\gamma_{n}=\Omega(\sqrt{n}) and γn=o(n)\gamma_{n}=o(n), and if for some sequence wnw_{n}, it holds that

Kn=r2(γn)+ωn,wherer2(γn)=log(γn)log2+1/2K_{n}=r_{2}(\gamma_{n})+\omega_{n},\ \ \textrm{where}\quad r_{2}(\gamma_{n})=\frac{\log(\gamma_{n})}{\log 2+1/2}

is the threshold function, then we have

limnP(n,Kn,γn)=1,iflimnωn=\displaystyle\lim_{n\to\infty}P(n,K_{n},\gamma_{n})=1,\quad\mathrm{if}\quad\lim_{n\to\infty}\omega_{n}=\infty (5)

We remind that random K-out graph is known [4, 5] to be connected whp when Kn2K_{n}\geq 2. Part (a)(a) of Theorem III.2 extends this result by showing that having Kn2K_{n}\geq 2 is sufficient for the random K-out graph to remain connected whp even when o(n)o(\sqrt{n}) of its nodes (selected randomly) are deleted. We believe that this result will further facilitate the application of random K-out graphs in a wide range of applications where connectivity despite node failures is crucial.

III-B Results on the Size of the Giant Component

Let Cmax(n,Kn,γn)C_{max}(n,K_{n},\gamma_{n}) denote the set of nodes in the largest connected component of (n;Kn,γn)\mathbb{H}(n;K_{n},\gamma_{n}) and let PG(n,Kn,γn,λn):=[|Cmax(n,Kn,γn)|>nγnλn]P_{G}(n,K_{n},\gamma_{n},\lambda_{n}):=\mathbb{P}[|C_{max}(n,K_{n},\gamma_{n})|>n-\gamma_{n}-\lambda_{n}]. Namely, PG(n,Kn,γn,λn)P_{G}(n,K_{n},\gamma_{n},\lambda_{n}) is the probability that less than λn\lambda_{n} nodes are outside the largest component of (n;Kn,γn)\mathbb{H}(n;K_{n},\gamma_{n}).

Theorem III.3

Let γn=o(n)\gamma_{n}=o(n), λn=Ω(n)\lambda_{n}=\Omega(\sqrt{n}) and λn(nγn)/3\lambda_{n}\leq\lfloor(n-\gamma_{n})/3\rfloor. Consider a scaling K:00K:\mathbb{N}_{0}\to\mathbb{N}_{0} and let

r3(γn,λn)=1+log(1+γn/λn)log2+1/2r_{3}(\gamma_{n},\lambda_{n})=1+\frac{\log(1+\gamma_{n}/\lambda_{n})}{\log 2+1/2}

be the threshold function. Then, we have

limnPG(n,Kn,γn,λn)=1,ifKn>r3(γn,λn),n.\displaystyle\lim_{n\to\infty}P_{G}(n,K_{n},\gamma_{n},\lambda_{n})=1,\quad\mathrm{if}\quad K_{n}>r_{3}(\gamma_{n},\lambda_{n}),\ \ \forall n.

We remark that if λn=βn\lambda_{n}=\beta n with 0<β<1/30<\beta<1/3, then r3(γn,λn)=1+o(1)r_{3}(\gamma_{n},\lambda_{n})=1+o(1). This shows that when γn=o(n)\gamma_{n}=o(n), it suffices to have Kn2K_{n}\geq 2 for (n;Kn,γn)\mathbb{H}(n;K_{n},\gamma_{n}) to have a giant component containing (1β)n(1-\beta)n nodes for arbitrary 0<β<1/30<\beta<1/3. Put differently, by choosing Kn2K_{n}\geq 2, we ensure that even when γn=o(n)\gamma_{n}=o(n) nodes are removed, the rest of the network contains a connected component whose fractional size is arbitrarily close to 1.

Theorem III.4

Let γn=αn\gamma_{n}=\alpha n with α\alpha in (0,1)(0,1), and λn(1α)n3\lambda_{n}\leq\lfloor\frac{(1-\alpha)n}{3}\rfloor. Consider a scaling K:00K:\mathbb{N}_{0}\to\mathbb{N}_{0} and let

r4(α,λn)=1+log(1+nαλn)+α+log(1α)1α2log(1+α2)r_{4}(\alpha,\lambda_{n})=1+\frac{\log(1+\frac{n\alpha}{\lambda_{n}})+\alpha+\log(1-\alpha)}{\frac{1-\alpha}{2}-\log\left(\frac{1+\alpha}{2}\right)}

be the threshold function. Then, we have

limnPG(n,Kn,α,λ)=1,ifKn>r4(α,xn),n.\displaystyle\lim_{n\to\infty}P_{G}(n,K_{n},\alpha,\lambda)=1,\quad\mathrm{if}\quad K_{n}>r_{4}(\alpha,x_{n}),\ \ \forall n.

It can be seen from this result that KnK_{n} needs to scale as Knlog(αnλn)K_{n}\sim\log(\frac{\alpha n}{\lambda_{n}}) for a random K-out graph to have a giant component of size nλnn-\lambda_{n} when αn\alpha n of its nodes are removed (or, if each node is independently removed with probability 0<α<10<\alpha<1). We also remark that the threshold r4(α,λn)r_{4}(\alpha,\lambda_{n}) is finite when λn=Ω(n)\lambda_{n}=\Omega(n). This shows that even when a positive fraction of the nodes of the random K-out graph are removed, a finite KnK_{n} is still sufficient to have a giant component of size Ω(n)\Omega(n) in the graph. This result can be useful in applications where it is required to maintain a giant component as efficiently (i.e., with as fewest edges) as possible even when large scale node failures take place.

III-C Result on Robustness

Theorem III.5

Define

r(K)=maxr=1,2,3{limnP((n;K) is r-robust)=1}r^{\star}(K)=\max_{r=1,2,3\ldots}\{\lim_{n\to\infty}P(\mathbb{H}(n;K)\text{ is }r\text{-robust})=1\}

Then, we have

r(K)K/2r^{\star}(K)\geq\lfloor K/2\rfloor

In Theorem III.5, we establish a threshold for one-law of rr-robustness in random K-out graphs, and find that r(K)K/2r^{\star}(K)\geq\lfloor K/2\rfloor. In other words, we find that with high probability, a random K-out graph is rr-robust when K2r,r2,r+K\geq 2r,r\geq 2,r\in\mathbb{Z}^{+}, and nn\to\infty. This threshold is much smaller than the previously established threshold [24] of K>2r(log(r)+log(log(r)+1)log(2)+1/2log(1+log(2)+1/22log(r)+5/2+log(2))K>\frac{2r\left(\log(r)+\log(\log(r)+1\right)}{\log(2)+1/2-\log\left(1+\frac{\log(2)+1/2}{2\log(r)+5/2+\log(2)}\right)} which scales with rlogrr\log r. Hence, Theorem III.5 constitutes a sharper one-law for rr-robustness.

Desired Minimum KnK_{n} needed to Theorem
Property achieve the property
Connectivity γn=αn\gamma_{n}=\alpha n (1+ε)logn1αlogα(1+\varepsilon)\frac{\log n}{1-\alpha-\log\alpha} Thm. III.1
Connectivity γn=o(n)\gamma_{n}=o(\sqrt{n}) Kn2K_{n}\geq 2 Thm. III.2(a)
Connectivity γn=w(n)\gamma_{n}=w(\sqrt{n}), γn=o(n)\gamma_{n}=o(n) log(γn)log2+1/2+w(1)\frac{\log(\gamma_{n})}{\log 2+1/2}+w(1) Thm. III.2(b)
Giant Component, λn=o(n)\lambda_{n}=o(n), γn=o(n)\gamma_{n}=o(n) 1+log(1+γnλn)log(2)+1/21+\frac{\log(1+\frac{\gamma_{n}}{\lambda_{n}})}{\log(2)+1/2} Thm. III.3
Giant Component, λn=βn\lambda_{n}=\beta n, γn=o(n)\gamma_{n}=o(n) Kn2K_{n}\geq 2 Thm. III.3
Giant Component, λn<(1α)n3\lambda_{n}<\lfloor\frac{(1-\alpha)n}{3}\rfloor, γn=αn\gamma_{n}=\alpha n 1+log(1+αnλn)+α+log(1α)1α2log(1+α2)1+\frac{\log(1+\frac{\alpha n}{\lambda_{n}})+\alpha+\log(1-\alpha)}{\frac{1-\alpha}{2}-\log(\frac{1+\alpha}{2})} Thm. III.4
rr-robustness Kn2rK_{n}\geq 2r Thm. III.5
TABLE I: Summary of our main results providing a condition on KnK_{n} needed to achieve a desired property in (n;Kn,γn)\mathbb{H}(n;K_{n},\gamma_{n}) (whp) where γn\gamma_{n} denotes the number of deleted nodes. For the giant component, the desired property is defined as its size to be at least nλnn-\lambda_{n}. In the first row, the result is for any ε>0\varepsilon>0. On the fifth row, we have β\beta in (0,1/3)(0,1/3) and on the sixth row, we have α\alpha in (0,1)(0,1).

III-D Discussion

In Theorem III.1, we improve the results given in [25], and with this, we close the gap between the zero law and the one law, and hence establish a sharp zero-one law for connectivity when γn=Ω(n)\gamma_{n}=\Omega(n) nodes are deleted from (n;Kn,γn)\mathbb{H}(n;K_{n},\gamma_{n}). In Theorem III.2, we establish that the graph (n;Kn,γn)\mathbb{H}(n;K_{n},\gamma_{n}) with γn=o(n)\gamma_{n}=o(n) is connected whp when Knlog(γn)K_{n}\sim\log(\gamma_{n}); and when γn=o(n)\gamma_{n}=o(\sqrt{n}), Kn2K_{n}\geq 2 is sufficient for connectivity. The latter result is especially important, since Kn2K_{n}\geq 2 is the previously established threshold for connectivity [4]. We improve this result by showing that the graph is still connected with Kn2K_{n}\geq 2 even after o(n)o(\sqrt{n}) nodes (selected randomly) are deleted.

Since most distributed systems require connectivity in the event of node failures, our results can be useful in many applications of distributed systems, particularly when the resources on each node is limited and it is critical to achieve desired connectivity and robustness properties using as few edges as possible. For example, in wireless sensor networks, knowing the minimum conditions needed for connectivity or giant component size under such failures is crucial as it enables designing them with fewest edges possible per node [18, 28], which reduces the communication overhead and potentially the cost of the hardware on each node.

We also note that Theorems III.3 - III.4 constitute the first results concerning the giant component size of random K-out graphs under randomly deleted nodes. In particular, these results help choose the value of KnK_{n} for any anticipated level of node failure and for any given giant component size required, enabling the designs of distributed systems to compromise between efficiency, robustness, and the required giant component size. Thus, we expect these results to be useful in applications where connectivity is not a stringent condition under node failures, and instead having a certain giant component size is sufficient to continue the operation of the system.

In Theorem III.5, we establish that a random K-out graph is rr-robust whp when Kn2rK_{n}\geq 2r for any r+r\in\mathbb{Z}^{+}. This is a much sharper one-law than the previous result given in [24] where it was shown that KnK_{n} needs to scale as Knrlog(r)K_{n}\sim r\log(r) for rr-robustness. This tighter result was made possible through several novel steps introduced here. While the proofs in prior work [29, 24] also rely on finding upper bounds on the probability of having at least one subset that is not rr-reachable, they tend to utilize standard upper bounds for the binomial coefficients (nk)(enk)k{n\choose k}\leq\left(\frac{en}{k}\right)^{k} and a union bound to establish them. Instead, our proof uses extensively the Beta function B(a,b)B(a,b) and its properties to obtain tighter upper bounds on such probabilities, which then enables us to establish a much sharper one-law for rr-robustness of random K-out graphs. We believe this result will pave the way for further applications of random K-out graphs in distributed computing applications such as the design of consensus networks in the presence of adversaries.

It is also of interest to compare the threshold of rr-robustness and rr-connectivity in random K-out graphs. For Erdős-Rényi graphs, the threshold for rr-connectivity and rr-robustness have been shown [29] to coincide with each other. For random K-out graphs, we know from [4] that (n;Kn)\mathbb{H}(n;K_{n}) is rr-connected whp whenever KnrK_{n}\geq r, and it is not rr-connected whp if Kn<rK_{n}<r. This leaves a factor of 2 difference between the condition Kn2rK_{n}\geq 2r we established for rr-robustness here and the threshold of rr-connectivity. Put differently, we know from [4] and Theorem III.5 that for any r=2,3,r=2,3,\ldots

limn[(n;Kn) is r-robust]={1,ifKn2r0,ifKn<r.\displaystyle\lim_{n\to\infty}\mathbb{P}\left[\mathbb{H}(n;K_{n})\text{ is $r$-robust}\right]=\begin{cases}1,&\mathrm{if}\quad K_{n}\geq 2r\\ 0,&\mathrm{if}\quad K_{n}<r.\end{cases} (6)

For r=1r=1, it is instead known that limn[(n;Kn) is r-robust]=1\lim_{n\to\infty}\mathbb{P}\left[\mathbb{H}(n;K_{n})\text{ is $r$-robust}\right]=1 if only if Kn2r=2K_{n}\geq 2r=2. Since the currently established conditions for the zero-law and one-law of rr-robustness are not the same for random K-out graphs (unlike ER graphs where the two thresholds coincide), there is a question as to whether our threshold of 2r2r is the tightest possible for rr-robustness. This is currently an open problem and would be an interesting direction for future work, e.g., by establishing a tighter zero-law for rr-robustness that coincides with the one-law of Theorem III.5. Comparison

To put all these results in perspective, we provide comparisons of our results with the results from an Erdős-Rényi graph G(n,p)G(n,p), which is one of the most commonly used random graph models. Firstly, in terms of rr-robustness, it was shown in [29] that ER graph G(n; p) is rr-robust whp if pn=log(n)+(r1)log(log(n))+ω(1)np_{n}=\frac{\log(n)+(r-1)\log(\log(n))+\omega(1)}{n}, which translates to a mean node degree of dlog(n)+(r1)log(log(n))\langle d\rangle\sim\log(n)+(r-1)\log(\log(n)). Since the mean node degree required for random K-out graphs scales as d2r\langle d\rangle\sim 2r, we can conclude that for large nn, random K-out graphs can be ensured to be rr-robust whp at a mean node degree significantly smaller than the mean node degree required for an ER graph. In terms of connectivity, an ER graph becomes connected whp if p>logn/np>\log n/n [30], and this translates to having a mean node degree of dlogn\langle d\rangle\sim\log n. Similarly, when o(n)o(\sqrt{n}) nodes are removed, the mean node degree required for connectivity scales with dlogn\langle d\rangle\sim\log n [6]. The d\langle d\rangle required for the random K-out graph to be connected whp is much lower, with d=O(1)\langle d\rangle=O(1) when o(n)o(\sqrt{n}) nodes are removed, and dlog(γn)\langle d\rangle\sim\log(\gamma_{n}) when γn=Ω(n)\gamma_{n}=\Omega(\sqrt{n}) nodes are removed.

Refer to caption
Refer to caption
Figure 2: Comparison of maximum number of nodes outside the giant component of a random K-out graph (n;Kn,γn)\mathbb{H}(n;K_{n},\gamma_{n}) and an ER graph with same mean node degree when n=5000n=5000, γn=0.4n\gamma_{n}=0.4n (Left); and when n=50,000n=50,000 and γn=500\gamma_{n}=500 (Right). Each data-point is obtained through 1000 experiments.

Next, since we are not aware of any theoretical results on the giant component size of ER graphs under node removals, we compare the size of the giant component in random K-out graphs and ER graphs under node removals through simulations in the finite node regime. We examine the maximum number of nodes outside the giant component out of 1000 experiments of a random K-out graph (n;Kn,γn)\mathbb{H}(n;K_{n},\gamma_{n}) and an ER graph G(n,p)G(n,p) with the same mean node degree when γn\gamma_{n} nodes are removed from both graphs. To ensure that both graphs have the same mean node degree, pp in the ER graph is selected as p=2Kn/np=2K_{n}/n. The results are given in Fig. 2 for n=5000n=5000, γn=0.4n\gamma_{n}=0.4n on (Left), and n=50,000n=50,000, γn=500\gamma_{n}=500 on (Right). As can be seen, the random K-out graph tends to have fewer nodes outside of the giant component than the ER graph and this difference is more pronounced when γn\gamma_{n} is smaller.

In conclusion, we see that when both graphs have the same mean node degree, random K-out graphs are more robust than ER graphs in terms of connectivity and giant component size under random node removal, and also in terms of the rr-robustness property. This reinforces the efficiency of the K-out construction in many distributed computing applications where connectivity in the event of node failures or adversarial capture of nodes is crucial. Similarly, the fact that random K-out graphs tend to achieve rr-robustness with fewer edges per node than ER graphs (for any r=1,2,r=1,2,\ldots), makes it more suitable in applications based on distributed consensus.

III-E Simulation Results

Since our results are asymptotic in nature, i.e., they have been established in the limit nn\to\infty, an important question is whether they can also be useful in practical settings where the number nn of nodes is finite. We check the usefulness to validate Theorems III.1 - III.4 under practical settings, To answer this, we examine the probability of connectivity and the number of nodes outside the giant component for the graph (n;Kn,γn)\mathbb{H}(n;K_{n},\gamma_{n}) (random K-out graph with deleted nodes) through computer simulations in two different setups111Determining whether a graph is rr-robust is a co-NP-complete problem [29] making it not feasible to check the usefulness of Theorem III.5 through computer simulations..

In the first setup, we consider the case where the number of deleted nodes, γn=αn\gamma_{n}=\alpha n, with α\alpha in (0,1)(0,1). We generate instantiations of the random graph (n;Kn,γn)\mathbb{H}(n;K_{n},\gamma_{n}) with n=5000n=5000, varying KnK_{n} in the interval [1,25][1,25] and consider several α\alpha values in the interval [0.1,0.8][0.1,0.8]. Then, we record the empirical probability of connectivity of the graph (n;Kn,γn)\mathbb{H}(n;K_{n},\gamma_{n}) and λn\lambda_{n} from 1000 independent experiments for each (Kn,α)(K_{n},\alpha) pair. The results of this experiment are shown in Fig. 3 and Fig. 4. Fig. 3 (Left) depicts the empirical probability of connectivity of (n;Kn,γn)\mathbb{H}(n;K_{n},\gamma_{n}). The vertical lines stand for the critical threshold of connectivity obtained from Theorem III.1. In each curve, P(n,Kn,γn)P(n,K_{n},\gamma_{n}) exhibits a threshold behaviour as KnK_{n} increases, and the transition from P(n,Kn,γn)=0P(n,K_{n},\gamma_{n})=0 to P(n,Kn,γn)=1P(n,K_{n},\gamma_{n})=1 takes place around Kn=logn1αlogαK_{n}=\frac{\log n}{1-\alpha-\log\alpha}, the threshold established in (2), reinforcing the usefulness of Theorem III.1 under practical settings.

In Fig. 4, the maximum number of nodes outside the giant component in 1000 experiments is plotted for each parameter pair. For comparison, we also plot the upper bound on nγn|Cmax|n-\gamma_{n}-|C_{max}| obtained from Theorem III.4 by taking the maximum γn\gamma_{n} value that gives a threshold less than or equal to the KnK_{n} value tested in the simulation. As can be seen, for any KnK_{n} and γn\gamma_{n} value, the experimental maximum number of nodes outside the giant component is smaller than the upper bound obtained from Theorem III.4, validating the usefulness of this result in the finite node regime.

Refer to caption
Refer to caption
Figure 3: (Left) Empirical probability that (n;Kn,γn)\mathbb{H}(n;K_{n},\gamma_{n}) is connected for n=5000n=5000 calculated from 1000 experiments. The vertical lines are the theoretical thresholds given by Theorem III.1. (Right) Maximum number of nodes outside the giant component of (n;Kn,γn)\mathbb{H}(n;K_{n},\gamma_{n}) for n=50,000n=50,000 in 1000 experiments.
Refer to caption
Refer to caption
Figure 4: Maximum number of nodes outside the giant component of (n;Kn,γn)\mathbb{H}(n;K_{n},\gamma_{n}) for n=5000n=5000 and γn=0.1n\gamma_{n}=0.1n, γn=0.2n\gamma_{n}=0.2n cases (Left); and for n=5000n=5000 and γn=0.4n\gamma_{n}=0.4n, γn=0.6n\gamma_{n}=0.6n cases (Right), obtained through 1000 experiments along with respective plot of theoretical nγn|Cmax|n-\gamma_{n}-|C_{max}|.
Refer to caption
Refer to caption
Figure 5: Maximum number of nodes outside the giant component of (n;Kn,γn)\mathbb{H}(n;K_{n},\gamma_{n}) for n=50,000n=50,000 and γn=10\gamma_{n}=10 cases (Left); and for for n=50,000n=50,000 and γn=250\gamma_{n}=250 cases (Right), obtained through 1000 experiments along with the plot of theoretical nγn|Cmax|n-\gamma_{n}-|C_{max}|.

The goal of the second experimental setup is to examine the case where the number of deleted nodes is γn=o(n)\gamma_{n}=o(n). As before, we generate instantiations of the random graph (n;Kn,γn)\mathbb{H}(n;K_{n},\gamma_{n}), with n=50,000n=50,000, varying KnK_{n} in [2,5][2,5], and varying λn\lambda_{n} in [10,2000][10,2000]. For each (Kn,γn)(K_{n},\gamma_{n}) pair, the maximum number of nodes outside the giant component in 1000 experiments is recorded; if no node is outside the giant component, then it is understood that the graph is connected. In Fig. 3 (Right), the maximum number of nodes outside the giant component observed in 10001000 experiments is depicted as a function of KnK_{n}. The plots for γn=10\gamma_{n}=10 and γn=100\gamma_{n}=100 are considered to represent the case γn=o(n)\gamma_{n}=o(\sqrt{n}) in Theorem III.2(a). We see that there is at most one node outside the giant component for γn=10\gamma_{n}=10 and γn=100\gamma_{n}=100, even when Kn=2K_{n}=2. This shows that the asymptotic behavior given in Theorem III.2(a), i.e., that random K-out graph remains connected if o(no(\sqrt{n} nodes are deleted, already appears when n=50,000n=50,000. The plots for γn=1000\gamma_{n}=1000 and γn=2000\gamma_{n}=2000 are used to check the case γn=w(n)\gamma_{n}=w(\sqrt{n}) and γn=o(n)\gamma_{n}=o(n) in Theorem III.2(b). The thresholds on KnK_{n} for these γn\gamma_{n} values, obtained using Theorem III.2(b) are r2(1000)=6.79r_{2}(1000)=6.79 and r2(2000)=7.37r_{2}(2000)=7.37, rounded to two digits after decimal (the ω(1)\omega(1) term in Theorem III.2(b) is ignored due to nn having a finite value in the simulations). It is clear from the plot that when Kn4K_{n}\geq 4, the graph with γn=1000\gamma_{n}=1000 is connected, while Kn5K_{n}\geq 5 suffices to ensure connectivity when γn=2000\gamma_{n}=2000. Thus, selecting KnK_{n} above the theoretical thresholds given in III.2(b) is seen to ensure the connectivity of the graph in the finite node regime as well, supporting the usefulness of Theorem III.2(b) in practical cases.

In Fig. 5, the maximum number of nodes outside the giant component in 1000 experiments is plotted as a function of KnK_{n}. For comparison, we also plot the upper bound on n|Cmax|n-|C_{max}| obtained from Theorem III.4. In particular, for each Theorem, the maximum γn\gamma_{n} value that gives a threshold less than or equal to the KnK_{n} value tested in the simulation is found. Then, the lowest of these maximum γn\gamma_{n} values is used as the theoretical n|Cmax|n-|C_{max}| value. As can be seen, for any KnK_{n} and γn\gamma_{n} value, the experimental maximum number of nodes outside the giant component is smaller than the upper bounds obtained from Theorem III.4, reinforcing the usefulness of our results in the finite node regime.

IV Proofs of Main Results

In this section, we provide the proof of all Theorems presented in Section III.

IV-A Preliminary Steps for Proving Theorems III.1 - III.4

Since the preliminary steps related to the proofs of III.1 - III.4 are the same, in this Section we present these steps. First, recall from Section II that the metrics connectivity and the size of the giant component under node removals are defined for the graph (n;Kn,γn)\mathbb{H}(n;K_{n},\gamma_{n}), where the set DD of nodes is removed from the graph (n;Kn)\mathbb{H}(n;K_{n}); also recall that R=VDR=\mathrm{V}\setminus D. Let 𝒩R\mathcal{N}_{R} denote the set of labels of the vertex set of (n;Kn,γn)\mathbb{H}(n;K_{n},\gamma_{n}) and let n(Kn,γn;S)\mathcal{E}_{n}(K_{n},\gamma_{n};S) denote the event that S𝒩RS\subset\mathcal{N}_{R} is a cut in (n;Kn,γn)\mathbb{H}(n;K_{n},\gamma_{n}) as per Definition II.2. The event n(Kn,γn;S)\mathcal{E}_{n}(K_{n},\gamma_{n};S) occurs if no nodes in SS pick neighbors in ScS^{\rm c}, and no nodes in SS pick neighbors in ScS^{\rm c}. Note that nodes in SS or ScS^{\rm c} can still pick neighbors in the set 𝒩D\mathcal{N}_{D}. Thus, we have

n(Kn,γn;S)=iSjSc({iΓnγn,j}{jΓnγn,i}).\displaystyle\mathcal{E}_{n}(K_{n},\gamma_{n};S)=\bigcap_{i\in S}\bigcap_{j\in S^{\rm c}}\left(\left\{i\not\in\Gamma_{n-\gamma_{n},j}\right\}\cap\left\{j\notin\Gamma_{n-\gamma_{n},i}\right\}\right).

Let 𝒵(λn;Kn,γn)\mathcal{Z}(\lambda_{n};K_{n},\gamma_{n}) denote the event that (n;Kn,γn)\mathbb{H}(n;K_{n},\gamma_{n}) has no cut S𝒩RS\subset\mathcal{N}_{R} with size λn|S|nγnλn\lambda_{n}\leq|S|\leq n-\gamma_{n}-\lambda_{n} where x:00x:\mathbb{N}_{0}\rightarrow\mathbb{N}_{0} is a sequence such that λn(nγn)/2n\lambda_{n}\leq(n-\gamma_{n})/{2}\ \forall n. In other words, 𝒵(λn;Kn,γn)\mathcal{Z}(\lambda_{n};K_{n},\gamma_{n}) is the event that there are no cuts in (n;Kn,γn)\mathbb{H}(n;K_{n},\gamma_{n}) whose size falls in the range [λn,nγnλn][\lambda_{n},n-\gamma_{n}-\lambda_{n}].

Lemma IV.1

[17, Lemma 4.3] For any sequence x:00x:\mathbb{N}_{0}\rightarrow\mathbb{N}_{0} such that λn(nγn)/3\lambda_{n}\leq\lfloor(n-\gamma_{n})/3\rfloor for all nn, we have

𝒵(λn;Kn,γn)|Cmax(n,Kn,γn)|>nγnλn.\displaystyle\mathcal{Z}(\lambda_{n};K_{n},\gamma_{n})\Rightarrow|C_{max}(n,K_{n},\gamma_{n})|>n-\gamma_{n}-\lambda_{n}. (7)

Lemma IV.1 states that if the event 𝒵(λn;Kn,γn)\mathcal{Z}(\lambda_{n};K_{n},\gamma_{n}) holds, then the size of the largest connected component of (n;Kn,γn)\mathbb{H}(n;K_{n},\gamma_{n}) is greater than nγnλnn-\gamma_{n}-\lambda_{n}; i.e., there are less than λn\lambda_{n} nodes outside of the giant component of (n;Kn,γn)\mathbb{H}(n;K_{n},\gamma_{n}). Also note that (n;Kn,γn)\mathbb{H}(n;K_{n},\gamma_{n}) is connected if 𝒵(λn;Kn,γn)\mathcal{Z}(\lambda_{n};K_{n},\gamma_{n}) takes place with λn=1\lambda_{n}=1, since a graph is connected if no node is outside the giant component. In order to establish the Theorems III.1-III.4., we need to show that limn[𝒵(λn;Kn,γn)c]=0\lim_{n\to\infty}\mathbb{P}[\mathcal{Z}(\lambda_{n};K_{n},\gamma_{n})^{\rm c}]=0 with λn\lambda_{n}, KnK_{n} and γn\gamma_{n} values as stated in each Theorem. From the definition of 𝒵(λn;Kn,γn)\mathcal{Z}(\lambda_{n};K_{n},\gamma_{n}), we have

𝒵(λn;Kn,γn)\displaystyle\mathcal{Z}(\lambda_{n};K_{n},\gamma_{n}) =S𝒫n:λn|S|nγn2(n(Kn,γn;S))c,\displaystyle=\bigcap_{S\in\mathcal{P}_{n}:~{}\lambda_{n}\leq|S|\leq\lfloor\frac{n-\gamma_{n}}{2}\rfloor}\left(\mathcal{E}_{n}({K}_{n},{\gamma_{n}};S)\right)^{\rm c},

where 𝒫n\mathcal{P}_{n} is the collection of all non-empty subsets of 𝒩R\mathcal{N}_{R}. Complementing both sides and using the union bound, we get

[(𝒵(λn;Kn,γn))c]\displaystyle\mathbb{P}\left[\left(\mathcal{Z}(\lambda_{n};K_{n},\gamma_{n})\right)^{\rm c}\right] S𝒫n:λn|S|nγ2[n(Kn,γn;S)]\displaystyle\leq\sum_{S\in\mathcal{P}_{n}:\lambda_{n}\leq|S|\leq\lfloor\frac{n-\gamma}{2}\rfloor}\mathbb{P}[\mathcal{E}_{n}({K}_{n},{\gamma_{n}};S)]
=r=λnnγ2S𝒫n,r[n(Kn,γn;S)],\displaystyle=\sum_{r=\lambda_{n}}^{\left\lfloor\frac{n-\gamma}{2}\right\rfloor}\sum_{S\in\mathcal{P}_{n,r}}\mathbb{P}[\mathcal{E}_{n}({K}_{n},{\gamma_{n}};S)], (8)

where 𝒫n,r\mathcal{P}_{n,r} denotes the collection of all subsets of 𝒩R\mathcal{N}_{R} with exactly rr elements. For each r=1,,(nγn)/2r=1,\ldots,\left\lfloor(n-\gamma_{n})/2\right\rfloor, we can simplify the notation by denoting n,r(Kn,γn)=n(Kn,γn;{1,,r})\mathcal{E}_{n,r}({K}_{n},{\gamma_{n}})=\mathcal{E}_{n}({K}_{n},{\gamma_{n}};\{1,\ldots,r\}). From the exchangeability of the node labels and associated random variables, we have

[n(Kn,γn;S)]=[n,r(Kn,γn)],S𝒫n,r.\mathbb{P}[\mathcal{E}_{n}({K}_{n},{\gamma_{n}};S)]=\mathbb{P}[\mathcal{E}_{n,r}({K}_{n},{\gamma_{n}})],\quad S\in\mathcal{P}_{n,r}.

|𝒫n,r|=(nγnr)|\mathcal{P}_{n,r}|={n-\gamma_{n}\choose r}, since there are (nγnr){n-\gamma_{n}\choose r} subsets of 𝒩R\mathcal{N}_{R} with r elements. Thus, we have

S𝒫n,r[n(Kn,γn;S)]=(nγnr)[n,r(Kn,γn)].\sum_{S\in\mathcal{P}_{n,r}}\mathbb{P}[\mathcal{E}_{n}({K}_{n},{\gamma_{n}};S)]={n-\gamma_{n}\choose r}~{}\mathbb{P}[\mathcal{E}_{n,r}({K}_{n},{\gamma_{n}})].

Substituting this into (8), we obtain

[(𝒵(λn;Kn,γn))c]r=λnnγ2(nγnr)[n,r(Kn,γn)]\displaystyle\mathbb{P}\left[\left(\mathcal{Z}(\lambda_{n};K_{n},\gamma_{n})\right)^{\rm c}\right]\leq\sum_{r=\lambda_{n}}^{\left\lfloor\frac{n-\gamma}{2}\right\rfloor}{n-\gamma_{n}\choose r}\mathbb{P}[\mathcal{E}_{n,r}({K}_{n},{\gamma_{n}})] (9)

Remember that n,r(Kn,γn)\mathcal{E}_{n,r}({K}_{n},{\gamma_{n}}) is the event that the nγnrn-\gamma_{n}-r nodes in SS and rr nodes in ScS^{\rm c} do not pick each other, but they can pick nodes from the set 𝒩D\mathcal{N}_{D}. Thus, we have:

[n,r(Kn,γn)]\displaystyle\mathbb{P}[\mathcal{E}_{n,r}({K}_{n},{\gamma_{n}})] =((γn+r1Kn)(n1Kn))r((nr1Kn)(n1Kn))nγnr(γn+rn)rKn(nrn)Kn(nγnr)\displaystyle=\left(\dfrac{{\gamma_{n}+r-1\choose K_{n}}}{{n-1\choose K_{n}}}\right)^{r}\left(\dfrac{{n-r-1\choose K_{n}}}{{n-1\choose K_{n}}}\right)^{n-\gamma_{n}-r}\leq\left(\dfrac{\gamma_{n}+r}{n}\right)^{rK_{n}}\left(\dfrac{n-r}{n}\right)^{K_{n}(n-\gamma_{n}-r)} (10)

Abbreviating [𝒵(1;Kn,γn)c]\mathbb{P}\left[\mathcal{Z}(1;K_{n},\gamma_{n})^{\rm c}\right] as PZP_{Z}, we get from (9) that

PZr=λnnγn2(nγnr)(γn+rn)rKn(nrn)Kn(nγnr)\displaystyle P_{Z}\leq\sum_{r=\lambda_{n}}^{\left\lfloor\frac{n-\gamma_{n}}{2}\right\rfloor}{n-\gamma_{n}\choose r}\left(\dfrac{\gamma_{n}+r}{n}\right)^{rK_{n}}\left(\dfrac{n-r}{n}\right)^{K_{n}(n-\gamma_{n}-r)} (11)

Using the upper bound on binomials (14) again, we have

PZr=λnnγn2(nγnr)r(nγnnγnr)nγnr(γn+rn)rKn(nrn)Kn(nγnr)\displaystyle P_{Z}\leq\sum_{r=\lambda_{n}}^{\left\lfloor\frac{n-\gamma_{n}}{2}\right\rfloor}\left(\frac{n-\gamma_{n}}{r}\right)^{r}\left(\frac{n-\gamma_{n}}{n-\gamma_{n}-r}\right)^{n-\gamma_{n}-r}\left(\dfrac{\gamma_{n}+r}{n}\right)^{rK_{n}}\left(\dfrac{n-r}{n}\right)^{K_{n}(n-\gamma_{n}-r)} (12)

In order to establish the Theorems, we need to show that (12) goes to zero in the limit of large n for λn\lambda_{n}, γn\gamma_{n} and KnK_{n} values as specified in each Theorem.

Since they will be referred to frequently throughout the proofs, we also include here the following standard bounds.

1±xe±x\displaystyle 1\pm x\leq e^{\pm x} (13)
(nm)(nm)m(nnm)nm,m=1,,n\displaystyle{n\choose m}\leq\left(\frac{n}{m}\right)^{m}\left(\frac{n}{n-m}\right)^{n-m},\quad\forall m=1,\ldots,n (14)

IV-B A Proof of Theorem III.1

Recall that in Theorem III.1, we have γn=αn\gamma_{n}=\alpha n with 0<α<10<\alpha<1 and that we need λn=1\lambda_{n}=1 for connectivity. Using (13) in (12), we have

PZr=1nαn2(nαnr)rer(α+rn)rKnerKn(nαnr)n\displaystyle P_{Z}\leq\sum_{r=1}^{\left\lfloor\frac{n-\alpha n}{2}\right\rfloor}\left(\dfrac{n-\alpha n}{r}\right)^{r}e^{r}\left(\alpha+\dfrac{r}{n}\right)^{rK_{n}}e^{\frac{-rK_{n}(n-\alpha n-r)}{n}}

We will show that the right side of the above expression goes to zero as nn goes to infinity. Let

An,r,α:=(nαnr)rer(α+rn)rKnerKn(nαnr)n.A_{n,r,\alpha}:=\left(\dfrac{n-\alpha n}{r}\right)^{r}e^{r}\left(\alpha+\dfrac{r}{n}\right)^{rK_{n}}e^{\frac{-rK_{n}(n-\alpha n-r)}{n}}.

We write

PZr=1n/lognAn,r,α+r=n/lognnαn2An,r,α:=S1+S2,\displaystyle P_{Z}\leq\sum_{r=1}^{\left\lfloor n/\log n\right\rfloor}A_{n,r,\alpha}+\sum_{r=\left\lfloor n/\log n\right\rfloor}^{\left\lfloor\frac{n-\alpha n}{2}\right\rfloor}A_{n,r,\alpha}:=S_{1}+S_{2},

and show that both S1S_{1} and S2S_{2} go to zero as nn\to\infty. We start with the first summation S1S_{1}.

S1\displaystyle S_{1} r=1n/logn((1α)eneKnlog(α+1logn)Kn(1α1logn))r\displaystyle\!\begin{multlined}\leq\sum_{r=1}^{\left\lfloor n/\log n\right\rfloor}\left((1-\alpha)en\cdot e^{K_{n}\log(\alpha+\frac{1}{\log n})-K_{n}(1-\alpha-\frac{1}{\log n})}\right)^{r}\end{multlined}\leq\sum_{r=1}^{\left\lfloor n/\log n\right\rfloor}\left((1-\alpha)en\cdot e^{K_{n}\log(\alpha+\frac{1}{\log n})-K_{n}(1-\alpha-\frac{1}{\log n})}\right)^{r}

Next, assume as in the statement of Theorem III.1 that

Kn=cnlogn1αlogα,n=1,2,\displaystyle K_{n}=\frac{c_{n}\log n}{1-\alpha-\log\alpha},\quad n=1,2,\ldots (15)

for some sequence c:0+c:\mathbb{N}_{0}\to\mathbb{R}_{+} such that limncn=c\lim_{n\to\infty}c_{n}=c with c>1c>1. Also define

an\displaystyle a_{n} :=(1α)eneKnlog(α+1logn)Kn(1α1logn)\displaystyle:=(1-\alpha)en\cdot e^{K_{n}\log(\alpha+\frac{1}{\log n})-K_{n}(1-\alpha-\frac{1}{\log n})}
=(1α)enecnlogn1αlogα(1α1lognlog(α+1logn))\displaystyle=(1-\alpha)en\cdot e^{-\frac{c_{n}\log n}{1-\alpha-\log\alpha}\left(1-\alpha-\frac{1}{\log n}-\log(\alpha+\frac{1}{\log n})\right)}
=(1α)en1cnecn1αlogα(1lognlog(1+1αlogn))\displaystyle=(1-\alpha)en^{1-c_{n}}\cdot e^{\frac{c_{n}}{1-\alpha-\log\alpha}\left(1-\log n\cdot\log(1+\frac{1}{\alpha\log n})\right)}
=O(1)n1cn\displaystyle=O(1)n^{1-c_{n}}

where we substituted KnK_{n} via (15) and used the fact that lognlog(1+1αlogn)=1α+o(1)\log n\cdot\log(1+\frac{1}{\alpha\log n})=\frac{1}{\alpha}+o(1). Taking the limit as nn\to\infty and recalling that limncn=c>1\lim_{n\to\infty}c_{n}=c>1, we see that limnan=0\lim_{n\to\infty}a_{n}=0. Hence, for large nn, we have

S1r=1n/logn(an)rr=1(an)r=an1an\displaystyle S_{1}\leq\sum_{r=1}^{\left\lfloor n/\log n\right\rfloor}\left(a_{n}\right)^{r}\leq\sum_{r=1}^{\infty}\left(a_{n}\right)^{r}=\frac{a_{n}}{1-a_{n}} (16)

where the geometric sum converges by virtue of limnan=0\lim_{n\to\infty}a_{n}=0. Using this, it is clear that limnS1=0\lim_{n\to\infty}S_{1}=0.

Now, consider the second summation S2S_{2}.

S2\displaystyle S_{2} r=n/logn(nαn)/2((nαn)en/logn)r(αn+nαn2n)rKnerKnn(nαnnαn2)\displaystyle\leq\!\begin{multlined}\sum_{r=\lfloor n/\log n\rfloor}^{\left\lfloor(n-\alpha n)/2\right\rfloor}\left(\dfrac{(n-\alpha n)e}{n/\log n}\right)^{r}\left(\dfrac{\alpha n+\frac{n-\alpha n}{2}}{n}\right)^{rK_{n}}e^{\frac{-rK_{n}}{n}(n-\alpha n-\frac{n-\alpha n}{2})}\end{multlined}\sum_{r=\lfloor n/\log n\rfloor}^{\left\lfloor(n-\alpha n)/2\right\rfloor}\left(\dfrac{(n-\alpha n)e}{n/\log n}\right)^{r}\left(\dfrac{\alpha n+\frac{n-\alpha n}{2}}{n}\right)^{rK_{n}}e^{\frac{-rK_{n}}{n}(n-\alpha n-\frac{n-\alpha n}{2})} (18)
r=n/logn(nαn)/2((1α)elogneKnlog(1+α2)Kn1α2)r\displaystyle\!\begin{multlined}\leq\sum_{r=\lfloor n/\log n\rfloor}^{\left\lfloor(n-\alpha n)/2\right\rfloor}\left((1-\alpha)e\log n\cdot e^{K_{n}\log(\frac{1+\alpha}{2})-K_{n}\frac{1-\alpha}{2}}\right)^{r}\end{multlined}\leq\sum_{r=\lfloor n/\log n\rfloor}^{\left\lfloor(n-\alpha n)/2\right\rfloor}\left((1-\alpha)e\log n\cdot e^{K_{n}\log(\frac{1+\alpha}{2})-K_{n}\frac{1-\alpha}{2}}\right)^{r} (20)

Next, we define

bn\displaystyle b_{n} :=(1α)elogneKn(1α2log(1+α2))\displaystyle:=(1-\alpha)e\log n\cdot e^{-K_{n}\left(\frac{1-\alpha}{2}-\log(\frac{1+\alpha}{2})\right)} (21)
=(1α)elognecnlogn1αlogα(1α2log(1+α2))\displaystyle=(1-\alpha)e\log n\cdot e^{-\frac{c_{n}\log n}{1-\alpha-\log\alpha}\left(\frac{1-\alpha}{2}-\log(\frac{1+\alpha}{2})\right)} (22)

where we substituted for KnK_{n} via (15). Taking the limit as n{n\to\infty} we see that limnbn=0\lim_{n\to\infty}b_{n}=0 upon noting that 1α2log(1+α2)>0\frac{1-\alpha}{2}-\log(\frac{1+\alpha}{2})>0 and limncn=c>1\lim_{n\to\infty}c_{n}=c>1. With arguments similar to those used in the case of S1S_{1}, we can show that when nn is large, S2bn/(1bn)S_{2}\leq b_{n}/(1-b_{n}), leading to S2S_{2} converging to zero as nn gets large. With PZS1+S2P_{Z}\leq S_{1}+S_{2}, and both S1S_{1} and S2S_{2} converging to zero when nn is large, we establish the fact that PZP_{Z} converges to zero as nn goes to infinity. This result also yields the desired conclusion limnP(n,Kn,γn)=1\lim_{n\to\infty}P(n,K_{n},\gamma_{n})=1 in Theorem III.1 since PZ=1P(n,Kn,γn)P_{Z}=1-P(n,K_{n},\gamma_{n}).

IV-C A Proof of Theorem III.2

We will first start with part (a) of Theorem III.2.
Part a) Recall that in part (a), γn=o(n)\gamma_{n}=o(\sqrt{n}) and we need λn=1\lambda_{n}=1 for connectivity. Using this and (13) in (12), we get

PZ\displaystyle P_{Z} r=1nγn2(nγnr)r(nγnnγnr)nγnr(γn+rn)rKn(nrn)Kn(nγnr)\displaystyle\leq\!\begin{multlined}\sum_{r=1}^{\left\lfloor\frac{n-\gamma_{n}}{2}\right\rfloor}\left(\dfrac{n-\gamma_{n}}{r}\right)^{r}\left(\frac{n-\gamma_{n}}{n-\gamma_{n}-r}\right)^{n-\gamma_{n}-r}\left(\dfrac{\gamma_{n}+r}{n}\right)^{rK_{n}}\left(\frac{n-r}{n}\right)^{K_{n}(n-\gamma_{n}-r)}\end{multlined}\sum_{r=1}^{\left\lfloor\frac{n-\gamma_{n}}{2}\right\rfloor}\left(\dfrac{n-\gamma_{n}}{r}\right)^{r}\left(\frac{n-\gamma_{n}}{n-\gamma_{n}-r}\right)^{n-\gamma_{n}-r}\left(\dfrac{\gamma_{n}+r}{n}\right)^{rK_{n}}\left(\frac{n-r}{n}\right)^{K_{n}(n-\gamma_{n}-r)}
r=1nγn2(nγnr)r(1+rγnn(nγnr))nγnr(γn+rn)rKn(nrn)(Kn1)(nγnr)\displaystyle\leq\!\begin{multlined}\sum_{r=1}^{\left\lfloor\frac{n-\gamma_{n}}{2}\right\rfloor}\left(\dfrac{n-\gamma_{n}}{r}\right)^{r}\left(1+\frac{r\gamma_{n}}{n(n-\gamma_{n}-r)}\right)^{n-\gamma_{n}-r}\left(\dfrac{\gamma_{n}+r}{n}\right)^{rK_{n}}\left(\frac{n-r}{n}\right)^{(K_{n}-1)(n-\gamma_{n}-r)}\end{multlined}\sum_{r=1}^{\left\lfloor\frac{n-\gamma_{n}}{2}\right\rfloor}\left(\dfrac{n-\gamma_{n}}{r}\right)^{r}\left(1+\frac{r\gamma_{n}}{n(n-\gamma_{n}-r)}\right)^{n-\gamma_{n}-r}\left(\dfrac{\gamma_{n}+r}{n}\right)^{rK_{n}}\left(\frac{n-r}{n}\right)^{(K_{n}-1)(n-\gamma_{n}-r)}
r=1nγn2(1+γnr)r(γn+rn)r(Kn1)er(Kn1)(nγnr)n\displaystyle\leq\!\begin{multlined}\sum_{r=1}^{\left\lfloor\frac{n-\gamma_{n}}{2}\right\rfloor}\left(1+\dfrac{\gamma_{n}}{r}\right)^{r}\left(\dfrac{\gamma_{n}+r}{n}\right)^{r(K_{n}-1)}e^{\frac{-r(K_{n}-1)(n-\gamma_{n}-r)}{n}}\end{multlined}\sum_{r=1}^{\left\lfloor\frac{n-\gamma_{n}}{2}\right\rfloor}\left(1+\dfrac{\gamma_{n}}{r}\right)^{r}\left(\dfrac{\gamma_{n}+r}{n}\right)^{r(K_{n}-1)}e^{\frac{-r(K_{n}-1)(n-\gamma_{n}-r)}{n}}

We will show that the right side of the above expression goes to zero as nn goes to infinity. Let

An,r,γn:=(1+γnr)r(γn+rn)r(Kn1)er(Kn1)(nγnr)nA_{n,r,\gamma_{n}}:=\left(1+\dfrac{\gamma_{n}}{r}\right)^{r}\left(\dfrac{\gamma_{n}+r}{n}\right)^{r(K_{n}-1)}e^{\frac{-r(K_{n}-1)(n-\gamma_{n}-r)}{n}}

We write

PZr=1nAn,r,γn+r=nnγn2An,r,γn:=S1+S2,\displaystyle P_{Z}\leq\sum_{r=1}^{\left\lfloor\sqrt{n}\right\rfloor}A_{n,r,\gamma_{n}}+\sum_{r=\left\lceil\sqrt{n}\right\rceil}^{\left\lfloor\frac{n-\gamma_{n}}{2}\right\rfloor}A_{n,r,\gamma_{n}}:=S_{1}+S_{2},

and show that both S1S_{1} and S2S_{2} go to zero as nn\to\infty. We start with the first summation S1S_{1}.

S1\displaystyle S_{1} r=1n(1+γnr)r(γn+rn)r(Kn1)er(Kn1)(nγnr)n\displaystyle\leq\sum_{r=1}^{\left\lfloor\sqrt{n}\right\rfloor}\left(1+\frac{\gamma_{n}}{r}\right)^{r}\left(\frac{\gamma_{n}+r}{n}\right)^{r(K_{n}-1)}e^{\frac{-r(K_{n}-1)(n-\gamma_{n}-r)}{n}}
r=1n(elog(1+γn)+(Kn1)[log(γn+nn)nγnnn])r\displaystyle\leq\!\begin{multlined}\sum_{r=1}^{\left\lfloor\sqrt{n}\right\rfloor}\left(e^{\log\left(1+\gamma_{n}\right)+(K_{n}-1)\left[\log\left(\frac{\gamma_{n}+\sqrt{n}}{n}\right)-\frac{n-\gamma_{n}-\sqrt{n}}{n}\right]}\right)^{r}\end{multlined}\sum_{r=1}^{\left\lfloor\sqrt{n}\right\rfloor}\left(e^{\log\left(1+\gamma_{n}\right)+(K_{n}-1)\left[\log\left(\frac{\gamma_{n}+\sqrt{n}}{n}\right)-\frac{n-\gamma_{n}-\sqrt{n}}{n}\right]}\right)^{r}

Next, assume as in the statement of Theorem III.2(a) that
Kn2,nK_{n}\geq 2,\ \forall n. Also define

an\displaystyle a_{n} :=elog(1+γn)+(Kn1)[log(γn+nn)nγnnn]\displaystyle:=e^{\log\left(1+\gamma_{n}\right)+(K_{n}-1)\left[\log\left(\frac{\gamma_{n}+\sqrt{n}}{n}\right)-\frac{n-\gamma_{n}-\sqrt{n}}{n}\right]}
elog(1+γn)+log(1+γnn)log(n)enγnnn\displaystyle\leq e^{\log\left(1+\gamma_{n}\right)+\log\left(1+\frac{\gamma_{n}}{\sqrt{n}}\right)-\log\left(\sqrt{n}\right)}e^{-\frac{n-\gamma_{n}-\sqrt{n}}{n}}
=O(1)elog(1+γn)log(n)\displaystyle=O(1)e^{\log\left(1+\gamma_{n}\right)-\log\left(\sqrt{n}\right)}

Taking the limit as nn\to\infty and recalling that γn=o(n)\gamma_{n}=o(\sqrt{n}), we see that limnan=0\lim_{n\to\infty}a_{n}=0. Hence, for large nn, we have

S1r=1n(an)rr=1(an)r=an1an\displaystyle S_{1}\leq\sum_{r=1}^{\left\lfloor\sqrt{n}\right\rfloor}\left(a_{n}\right)^{r}\leq\sum_{r=1}^{\infty}\left(a_{n}\right)^{r}=\frac{a_{n}}{1-a_{n}} (23)

where the geometric sum converges by virtue of limnan=0\lim_{n\to\infty}a_{n}=0. Using this once again, it is clear from the last expression that limnS1=0\lim_{n\to\infty}S_{1}=0.

Now, consider the second summation S2S_{2}.

S2\displaystyle S_{2} r=nnγn2(erγnn+(Kn1)[log(n+γn2n)nγn2n])r\displaystyle\leq\!\begin{multlined}\sum_{r=\left\lceil\sqrt{n}\right\rceil}^{\left\lfloor\frac{n-\gamma_{n}}{2}\right\rfloor}\left(e^{\frac{r\gamma_{n}}{\sqrt{n}}+(K_{n}-1)\left[\log\left(\frac{n+\gamma_{n}}{2n}\right)-\frac{n-\gamma_{n}}{2n}\right]}\right)^{r}\end{multlined}\sum_{r=\left\lceil\sqrt{n}\right\rceil}^{\left\lfloor\frac{n-\gamma_{n}}{2}\right\rfloor}\left(e^{\frac{r\gamma_{n}}{\sqrt{n}}+(K_{n}-1)\left[\log\left(\frac{n+\gamma_{n}}{2n}\right)-\frac{n-\gamma_{n}}{2n}\right]}\right)^{r}

Again assume as in the statement of Theorem III.2(a) that Kn2K_{n}\geq 2. Next, we define=

bn\displaystyle b_{n} :=erγnn+(Kn1)[log(n+γn2n)nγn2n]\displaystyle:=e^{\frac{r\gamma_{n}}{\sqrt{n}}+(K_{n}-1)\left[\log\left(\frac{n+\gamma_{n}}{2n}\right)-\frac{n-\gamma_{n}}{2n}\right]}
erγnn+log(12)+log(nn+γn)12+γn2n\displaystyle\leq e^{\frac{r\gamma_{n}}{\sqrt{n}}+\log\left(\frac{1}{2}\right)+\log\left(\frac{n}{n+\gamma_{n}}\right)-\frac{1}{2}+\frac{\gamma_{n}}{2n}}
=O(1)elog(n)\displaystyle=O(1)e^{-\log\left(\sqrt{n}\right)}

Taking the limit as nn\to\infty and recalling that γn=o(n)\gamma_{n}=o(\sqrt{n}), we see that limnbn=0\lim_{n\to\infty}b_{n}=0. Hence, for large nn, we have

S2r=nnγn2(bn)rr=n(bn)r=(bn)n1bn\displaystyle S_{2}\leq\sum_{r=\left\lceil\sqrt{n}\right\rceil}^{\left\lfloor\frac{n-\gamma_{n}}{2}\right\rfloor}\left(b_{n}\right)^{r}\leq\sum_{r=\left\lceil\sqrt{n}\right\rceil}^{\infty}\left(b_{n}\right)^{r}=\frac{(b_{n})^{\sqrt{n}}}{1-b_{n}} (24)

where the geometric sum converges by virtue of limnbn=0\lim_{n\to\infty}b_{n}=0. Using this once again, it is clear from the last expression that limnS2=0\lim_{n\to\infty}S_{2}=0. With PZS1+S2P_{Z}\leq S_{1}+S_{2}, and both S1S_{1} and S2S_{2} converging to zero when nn is large, we establish the fact that PZP_{Z} converges to zero as nn goes to infinity. This result also yields the desired conclusion limnP(n,Kn,γn)=1\lim_{n\to\infty}P(n,K_{n},\gamma_{n})=1 in Theorem III.2(a) since PZ=1P(n,Kn,γn)P_{Z}=1-P(n,K_{n},\gamma_{n}).

Part b) We now continue with the proof of Theorem III.2(b). Recall that we had γn=Ω(n)\gamma_{n}=\Omega(\sqrt{n}) and γn=o(n)\gamma_{n}=o(n). Using this and (13) in (12), we get

PZ\displaystyle P_{Z} r=1nγn2(nγnr)r(nγnnγnr)nγnr(γn+rn)rKnerKn(nγnr)n\displaystyle\leq\!\begin{multlined}\sum_{r=1}^{\left\lfloor\frac{n-\gamma_{n}}{2}\right\rfloor}\left(\dfrac{n-\gamma_{n}}{r}\right)^{r}\left(\frac{n-\gamma_{n}}{n-\gamma_{n}-r}\right)^{n-\gamma_{n}-r}\left(\dfrac{\gamma_{n}+r}{n}\right)^{rK_{n}}e^{\frac{-rK_{n}(n-\gamma_{n}-r)}{n}}\end{multlined}\sum_{r=1}^{\left\lfloor\frac{n-\gamma_{n}}{2}\right\rfloor}\left(\dfrac{n-\gamma_{n}}{r}\right)^{r}\left(\frac{n-\gamma_{n}}{n-\gamma_{n}-r}\right)^{n-\gamma_{n}-r}\left(\dfrac{\gamma_{n}+r}{n}\right)^{rK_{n}}e^{\frac{-rK_{n}(n-\gamma_{n}-r)}{n}}
r=1nγn2e(1γn/n)r(γn+rr)r(γn+rn)r(Kn1)erKn(nγnr)n\displaystyle\leq\!\begin{multlined}\sum_{r=1}^{\left\lfloor\frac{n-\gamma_{n}}{2}\right\rfloor}e^{(1-\gamma_{n}/n)r}\left(\dfrac{\gamma_{n}+r}{r}\right)^{r}\left(\dfrac{\gamma_{n}+r}{n}\right)^{r(K_{n}-1)}e^{\frac{-rK_{n}(n-\gamma_{n}-r)}{n}}\end{multlined}\sum_{r=1}^{\left\lfloor\frac{n-\gamma_{n}}{2}\right\rfloor}e^{(1-\gamma_{n}/n)r}\left(\dfrac{\gamma_{n}+r}{r}\right)^{r}\left(\dfrac{\gamma_{n}+r}{n}\right)^{r(K_{n}-1)}e^{\frac{-rK_{n}(n-\gamma_{n}-r)}{n}} (26)
r=1nγn2exp(r[(Kn1)(log(γn+rn)+rn+γnn1)+log(1+γn)+nγn2n])\displaystyle\leq\!\begin{multlined}\sum_{r=1}^{\left\lfloor\frac{n-\gamma_{n}}{2}\right\rfloor}\exp\left(r\left[(K_{n}-1)\left(\log\left(\frac{\gamma_{n}+r}{n}\right)+\frac{r}{n}+\frac{\gamma_{n}}{n}-1\right)+\log\left(1+\gamma_{n}\right)+\frac{n-\gamma_{n}}{2n}\right]\right)\end{multlined}\sum_{r=1}^{\left\lfloor\frac{n-\gamma_{n}}{2}\right\rfloor}\exp\left(r\left[(K_{n}-1)\left(\log\left(\frac{\gamma_{n}+r}{n}\right)+\frac{r}{n}+\frac{\gamma_{n}}{n}-1\right)+\log\left(1+\gamma_{n}\right)+\frac{n-\gamma_{n}}{2n}\right]\right) (28)

Next, assume as in the statement of Theorem III.2(b) that

Kn=log(γn+1)log2+1/2+w(1),n=1,2,\displaystyle K_{n}=\frac{\log(\gamma_{n}+1)}{\log 2+1/2}+w(1),\quad n=1,2,\ldots (29)

Since Kn1>0,n=1,2,K_{n}-1>0,\quad\forall n=1,2,\ldots, and noting that rnγn2r\leq\left\lfloor\frac{n-\gamma_{n}}{2}\right\rfloor in (28), we have

(Kn1)(log(γn+rn)+rn+γnn1)(Kn1)(log(γn+nγn2n)+nγn2n+γnn1)\displaystyle(K_{n}-1)\left(\log\left(\frac{\gamma_{n}+r}{n}\right)+\frac{r}{n}+\frac{\gamma_{n}}{n}-1\right)\leq(K_{n}-1)\left(\log\left(\frac{\gamma_{n}+\frac{n-\gamma_{n}}{2}}{n}\right)+\frac{\frac{n-\gamma_{n}}{2}}{n}+\frac{\gamma_{n}}{n}-1\right) (30)

Using this, we get

PZ\displaystyle P_{Z} r=1nγn2exp(r[(Kn1)(log(n+γn2n)nγn2n)+log(1+γn)+nγn2n])\displaystyle\leq\!\begin{multlined}\sum_{r=1}^{\left\lfloor\frac{n-\gamma_{n}}{2}\right\rfloor}\exp\left(r\left[(K_{n}-1)\left(\log\left(\frac{n+\gamma_{n}}{2n}\right)-\frac{n-\gamma_{n}}{2n}\right)+\log\left(1+\gamma_{n}\right)+\frac{n-\gamma_{n}}{2n}\right]\right)\end{multlined}\sum_{r=1}^{\left\lfloor\frac{n-\gamma_{n}}{2}\right\rfloor}\exp\left(r\left[(K_{n}-1)\left(\log\left(\frac{n+\gamma_{n}}{2n}\right)-\frac{n-\gamma_{n}}{2n}\right)+\log\left(1+\gamma_{n}\right)+\frac{n-\gamma_{n}}{2n}\right]\right) (32)

Next, define

an:=e(Kn1)(log(n+γn2n)nγn2n)+log(1+γn)+nγn2n\displaystyle a_{n}:=e^{(K_{n}-1)\left(\log\left(\frac{n+\gamma_{n}}{2n}\right)-\frac{n-\gamma_{n}}{2n}\right)+\log\left(1+\gamma_{n}\right)+\frac{n-\gamma_{n}}{2n}} (33)

Recall that γn=o(n)\gamma_{n}=o(n), so we have limnγn/n=0\lim_{n\to\infty}\gamma_{n}/n=0. Using this, and substituting KnK_{n} via (29), we get

limnan\displaystyle\lim_{n\to\infty}a_{n} =limn[e(log(γn+1)log2+1/2+w(1))(log212)+log(1+γn)+12]\displaystyle=\lim_{n\to\infty}\left[e^{\left(\frac{\log(\gamma_{n}+1)}{\log 2+1/2}+w(1)\right)\cdot\left(-log2-\frac{1}{2}\right)+\log\left(1+\gamma_{n}\right)+\frac{1}{2}}\right]
=limn[ew(1)(log2+1/2)log(1+γn)+log(1+γn)+12]\displaystyle=\lim_{n\to\infty}\left[e^{-w(1)\cdot\left(log2+1/2\right)-\log\left(1+\gamma_{n}\right)+\log\left(1+\gamma_{n}\right)+\frac{1}{2}}\right]
=limn[o(1)ew(1)(log2+1/2)]=0\displaystyle=\lim_{n\to\infty}\left[o(1)e^{-w(1)\cdot\left(log2+1/2\right)}\right]=0 (34)

Hence, for large nn, we have

PZr=1nγn2(an)rr=1(an)r=an1an\displaystyle P_{Z}\leq\sum_{r=1}^{\left\lfloor\frac{n-\gamma_{n}}{2}\right\rfloor}\left(a_{n}\right)^{r}\leq\sum_{r=1}^{\infty}\left(a_{n}\right)^{r}=\frac{a_{n}}{1-a_{n}} (35)

where the geometric sum converges by virtue of limnan=0\lim_{n\to\infty}a_{n}=0. Using this, it is clear from the last expression that limnPZ=0\lim_{n\to\infty}P_{Z}=0. This result also yields the desired conclusion limnP(n,Kn,γn)=1\lim_{n\to\infty}P(n,K_{n},\gamma_{n})=1 in Theorem III.2(b) since PZ=1P(n,Kn,γn)P_{Z}=1-P(n,K_{n},\gamma_{n}). This result, combined with the proof of part a, concludes the proof of Theorem III.2.

IV-D A Proof of Theorem III.3

Recall that in Theorem III.3, we have γn=o(n)\gamma_{n}=o(n) and λn=Ω(n)\lambda_{n}=\Omega(\sqrt{n}). Using (13) in (12), we have

PZ\displaystyle P_{Z} r=λnnγn2(nγnr)r(nγnnγnr)nγnr(γn+rn)rKn(nrn)Kn(nγnr)\displaystyle\leq\!\begin{multlined}\sum_{r=\lambda_{n}}^{\left\lfloor\frac{n-\gamma_{n}}{2}\right\rfloor}\left(\dfrac{n-\gamma_{n}}{r}\right)^{r}\left(\frac{n-\gamma_{n}}{n-\gamma_{n}-r}\right)^{n-\gamma_{n}-r}\left(\dfrac{\gamma_{n}+r}{n}\right)^{rK_{n}}\left(\frac{n-r}{n}\right)^{K_{n}(n-\gamma_{n}-r)}\end{multlined}\sum_{r=\lambda_{n}}^{\left\lfloor\frac{n-\gamma_{n}}{2}\right\rfloor}\left(\dfrac{n-\gamma_{n}}{r}\right)^{r}\left(\frac{n-\gamma_{n}}{n-\gamma_{n}-r}\right)^{n-\gamma_{n}-r}\left(\dfrac{\gamma_{n}+r}{n}\right)^{rK_{n}}\left(\frac{n-r}{n}\right)^{K_{n}(n-\gamma_{n}-r)}
r=λnnγn2(nγnr)r(1+rγnn(nγnr))nγnr(γn+rn)rKn(nrn)(Kn1)(nγnr)\displaystyle\leq\!\begin{multlined}\sum_{r=\lambda_{n}}^{\left\lfloor\frac{n-\gamma_{n}}{2}\right\rfloor}\left(\dfrac{n-\gamma_{n}}{r}\right)^{r}\left(1+\frac{r\gamma_{n}}{n(n-\gamma_{n}-r)}\right)^{n-\gamma_{n}-r}\left(\dfrac{\gamma_{n}+r}{n}\right)^{rK_{n}}\left(\frac{n-r}{n}\right)^{(K_{n}-1)(n-\gamma_{n}-r)}\end{multlined}\sum_{r=\lambda_{n}}^{\left\lfloor\frac{n-\gamma_{n}}{2}\right\rfloor}\left(\dfrac{n-\gamma_{n}}{r}\right)^{r}\left(1+\frac{r\gamma_{n}}{n(n-\gamma_{n}-r)}\right)^{n-\gamma_{n}-r}\left(\dfrac{\gamma_{n}+r}{n}\right)^{rK_{n}}\left(\frac{n-r}{n}\right)^{(K_{n}-1)(n-\gamma_{n}-r)}
r=λnnγn2(nγnr)rerγnn(γn+rn)r(γn+rn)r(Kn1)er(Kn1)(nγnr)n\displaystyle\leq\!\begin{multlined}\sum_{r=\lambda_{n}}^{\left\lfloor\frac{n-\gamma_{n}}{2}\right\rfloor}\left(\dfrac{n-\gamma_{n}}{r}\right)^{r}e^{\frac{r\gamma_{n}}{n}}\left(\dfrac{\gamma_{n}+r}{n}\right)^{r}\left(\dfrac{\gamma_{n}+r}{n}\right)^{r(K_{n}-1)}e^{\frac{-r(K_{n}-1)(n-\gamma_{n}-r)}{n}}\end{multlined}\sum_{r=\lambda_{n}}^{\left\lfloor\frac{n-\gamma_{n}}{2}\right\rfloor}\left(\dfrac{n-\gamma_{n}}{r}\right)^{r}e^{\frac{r\gamma_{n}}{n}}\left(\dfrac{\gamma_{n}+r}{n}\right)^{r}\left(\dfrac{\gamma_{n}+r}{n}\right)^{r(K_{n}-1)}e^{\frac{-r(K_{n}-1)(n-\gamma_{n}-r)}{n}}
r=λnnγn2(1+γnr)r(γn+rn)r(Kn1)er(Kn1)(nγnr)n\displaystyle\leq\!\begin{multlined}\sum_{r=\lambda_{n}}^{\left\lfloor\frac{n-\gamma_{n}}{2}\right\rfloor}\left(1+\dfrac{\gamma_{n}}{r}\right)^{r}\left(\dfrac{\gamma_{n}+r}{n}\right)^{r(K_{n}-1)}e^{\frac{-r(K_{n}-1)(n-\gamma_{n}-r)}{n}}\end{multlined}\sum_{r=\lambda_{n}}^{\left\lfloor\frac{n-\gamma_{n}}{2}\right\rfloor}\left(1+\dfrac{\gamma_{n}}{r}\right)^{r}\left(\dfrac{\gamma_{n}+r}{n}\right)^{r(K_{n}-1)}e^{\frac{-r(K_{n}-1)(n-\gamma_{n}-r)}{n}}

Next, assume as in the statement of Theorem III.3 that

Kn>1+log(1+γn/λn)log2+1/2\displaystyle K_{n}>1+\frac{\log(1+\gamma_{n}/\lambda_{n})}{\log 2+1/2} (36)

Since Kn>1K_{n}>1, we have

PZ\displaystyle P_{Z} r=λnnγn2(1+γnλn)r(n+γn2n)r(Kn1)er(Kn1)(nγn)2n\displaystyle\leq\!\begin{multlined}\sum_{r=\lambda_{n}}^{\left\lfloor\frac{n-\gamma_{n}}{2}\right\rfloor}\left(1+\dfrac{\gamma_{n}}{\lambda_{n}}\right)^{r}\left(\dfrac{n+\gamma_{n}}{2n}\right)^{r(K_{n}-1)}e^{\frac{-r(K_{n}-1)(n-\gamma_{n})}{2n}}\end{multlined}\sum_{r=\lambda_{n}}^{\left\lfloor\frac{n-\gamma_{n}}{2}\right\rfloor}\left(1+\dfrac{\gamma_{n}}{\lambda_{n}}\right)^{r}\left(\dfrac{n+\gamma_{n}}{2n}\right)^{r(K_{n}-1)}e^{\frac{-r(K_{n}-1)(n-\gamma_{n})}{2n}}

We will show that the right side of the above expression goes to zero as nn goes to infinity. Let

an:=elog(1+γnλn)+(Kn1)[log(n+γn2n)nγn2n]a_{n}:=e^{\log\left(1+\frac{\gamma_{n}}{\lambda_{n}}\right)+(K_{n}-1)\left[\log\left(\frac{n+\gamma_{n}}{2n}\right)-\frac{n-\gamma_{n}}{2n}\right]}

Recall that γn=o(n)\gamma_{n}=o(n), so we have limnγn/n=0\lim_{n\to\infty}\gamma_{n}/n=0. Using this, and substituting for KnK_{n} via (36), we get

an\displaystyle a_{n} <elog(1+γnλn)+(log(1+γn/λn)log(1/2)1/2)[log(n+γn2n)nγn2n]\displaystyle<e^{\log\left(1+\frac{\gamma_{n}}{\lambda_{n}}\right)+\left(\frac{\log(1+\gamma_{n}/\lambda_{n})}{-\log(1/2)-1/2}\right)\left[\log\left(\frac{n+\gamma_{n}}{2n}\right)-\frac{n-\gamma_{n}}{2n}\right]} (37)

Taking the limit nn\to\infty, we have

limnan\displaystyle\lim_{n\to\infty}a_{n} <limnelog(1+γnλn)log(1+γn/λn)=e0=1\displaystyle<\lim_{n\to\infty}e^{\log\left(1+\frac{\gamma_{n}}{\lambda_{n}}\right)-\log(1+\gamma_{n}/\lambda_{n})}=e^{0}=1 (38)

Hence, for large nn, we have

PZr=λnnγn2(an)rr=λn(an)r=(an)λn1an\displaystyle P_{Z}\leq\sum_{r=\lambda_{n}}^{\left\lfloor\frac{n-\gamma_{n}}{2}\right\rfloor}\left(a_{n}\right)^{r}\leq\sum_{r=\lambda_{n}}^{\infty}\left(a_{n}\right)^{r}=\frac{(a_{n})^{\lambda_{n}}}{1-a_{n}} (39)

where the geometric sum converges by virtue of limnan<1\lim_{n\to\infty}a_{n}<1 and limnλn=w(1)\lim_{n\to\infty}\lambda_{n}=w(1). Using this, it is clear from the last expression that limnPZ=0\lim_{n\to\infty}P_{Z}=0. This result also yields the desired conclusion limnPG(n,Kn,γn,λn)=1\lim_{n\to\infty}P_{G}(n,K_{n},\gamma_{n},\lambda_{n})=1 in Theorem III.3 since PZ=1PG(n,Kn,γn,λn)P_{Z}=1-P_{G}(n,K_{n},\gamma_{n},\lambda_{n}). This concludes the proof of Theorem III.3.

IV-E A Proof of Theorem III.4

Recall that in Theorem III.4, we have γn=αn\gamma_{n}=\alpha n with α\alpha in (0,1)(0,1), and λn<(1α)n2\lambda_{n}<\frac{(1-\alpha)n}{2}. Using γn=αn\gamma_{n}=\alpha n in (12), we get

PZ\displaystyle P_{Z} r=λnnαn2(nαnr)r(nαnnαnr)nαnr(αn+rn)rKn(nrn)Kn(nαnr)\displaystyle\leq\!\begin{multlined}\sum_{r=\lambda_{n}}^{\left\lfloor\frac{n-\alpha n}{2}\right\rfloor}\left(\dfrac{n-\alpha n}{r}\right)^{r}\left(\frac{n-\alpha n}{n-\alpha n-r}\right)^{n-\alpha n-r}\left(\dfrac{\alpha n+r}{n}\right)^{rK_{n}}\left(\frac{n-r}{n}\right)^{K_{n}(n-\alpha n-r)}\end{multlined}\sum_{r=\lambda_{n}}^{\left\lfloor\frac{n-\alpha n}{2}\right\rfloor}\left(\dfrac{n-\alpha n}{r}\right)^{r}\left(\frac{n-\alpha n}{n-\alpha n-r}\right)^{n-\alpha n-r}\left(\dfrac{\alpha n+r}{n}\right)^{rK_{n}}\left(\frac{n-r}{n}\right)^{K_{n}(n-\alpha n-r)}
r=λnnαn2(1α)reαr(1+αnr)r(αn+rn)r(Kn1)er(Kn1)(nαnr)n\displaystyle\leq\!\begin{multlined}\sum_{r=\lambda_{n}}^{\left\lfloor\frac{n-\alpha n}{2}\right\rfloor}\left(1-\alpha\right)^{r}e^{\alpha r}\left(1+\dfrac{\alpha n}{r}\right)^{r}\left(\dfrac{\alpha n+r}{n}\right)^{r(K_{n}-1)}e^{-r(K_{n}-1)\frac{(n-\alpha n-r)}{n}}\\ \end{multlined}\sum_{r=\lambda_{n}}^{\left\lfloor\frac{n-\alpha n}{2}\right\rfloor}\left(1-\alpha\right)^{r}e^{\alpha r}\left(1+\dfrac{\alpha n}{r}\right)^{r}\left(\dfrac{\alpha n+r}{n}\right)^{r(K_{n}-1)}e^{-r(K_{n}-1)\frac{(n-\alpha n-r)}{n}}\\ (41)

Next, assume as in the statement of Theorem III.4 that

Kn>1+log(1+αnλn)+α+log(1α)1α2+log2log(1+α),n=1,2,\displaystyle K_{n}>1+\frac{\log(1+\frac{\alpha n}{\lambda_{n}})+\alpha+\log(1-\alpha)}{\frac{1-\alpha}{2}+\log 2-\log(1+\alpha)},\quad n=1,2,\ldots

Since Kn>1K_{n}>1, we have

PZ\displaystyle P_{Z} r=λnnαn2(1α)reαr(1+αnλn)r(1+α2)r(Kn1)er(Kn1)(1α2)\displaystyle\leq\!\begin{multlined}\sum_{r=\lambda_{n}}^{\left\lfloor\frac{n-\alpha n}{2}\right\rfloor}\left(1-\alpha\right)^{r}e^{\alpha r}\left(1+\dfrac{\alpha n}{\lambda_{n}}\right)^{r}\left(\dfrac{1+\alpha}{2}\right)^{r(K_{n}-1)}e^{-r(K_{n}-1)\left(\frac{1-\alpha}{2}\right)}\end{multlined}\sum_{r=\lambda_{n}}^{\left\lfloor\frac{n-\alpha n}{2}\right\rfloor}\left(1-\alpha\right)^{r}e^{\alpha r}\left(1+\dfrac{\alpha n}{\lambda_{n}}\right)^{r}\left(\dfrac{1+\alpha}{2}\right)^{r(K_{n}-1)}e^{-r(K_{n}-1)\left(\frac{1-\alpha}{2}\right)} (43)

Also define

an\displaystyle a_{n} :=eα+log(1α)+log(1+αnλn)+(Kn1)[log(1+α2)(1α2)]\displaystyle:=e^{\alpha+\log(1-\alpha)+\log(1+\frac{\alpha n}{\lambda_{n}})+(K_{n}-1)\left[\log\left(\frac{1+\alpha}{2}\right)-\left(\frac{1-\alpha}{2}\right)\right]}
<eα+log(1α)+log(1+αnλn)(α+log(1α)+log(1+αnλn))=1\displaystyle<e^{\alpha+\log(1-\alpha)+\log(1+\frac{\alpha n}{\lambda_{n}})-\left(\alpha+\log(1-\alpha)+\log(1+\frac{\alpha n}{\lambda_{n}})\right)}=1

where we substituted KnK_{n} via (IV-E). Taking the limit as nn\to\infty, we see that limnan<1\lim_{n\to\infty}a_{n}<1. Hence, for large n, we have

PZr=λnnαn2(an)rr=λn(an)r=(an)λn1an\displaystyle P_{Z}\leq\sum_{r=\lambda_{n}}^{\left\lfloor\frac{n-\alpha n}{2}\right\rfloor}\left(a_{n}\right)^{r}\leq\sum_{r=\lambda_{n}}^{\infty}\left(a_{n}\right)^{r}=\frac{(a_{n})^{\lambda_{n}}}{1-a_{n}} (44)

where the geometric sum converges by virtue of limnan<1\lim_{n\to\infty}a_{n}<1 and limnλn=w(1)\lim_{n\to\infty}\lambda_{n}=w(1). Using this, it is clear from the last expression that limnPZ=0\lim_{n\to\infty}P_{Z}=0. This result also yields the desired conclusion limnPG(n,Kn,γn,λn)=1\lim_{n\to\infty}P_{G}(n,K_{n},\gamma_{n},\lambda_{n})=1 in Theorem III.4 since PZ=1PG(n,Kn,γn,λn)P_{Z}=1-P_{G}(n,K_{n},\gamma_{n},\lambda_{n}). This concludes the proof of Theorem III.4.

IV-F Preliminaries Needed in the Proof of Theorem III.5

We start with a few definitions and properties that will be useful throughout the rest of the proof. First, let B(a,b)B(a,b) denote the beta function, Bx(a,b)B_{x}(a,b) denote the incomplete beta function, and Ix(a,b)I_{x}(a,b) denote the regularized incomplete beta function, where aa and bb are non-negative integers. These functions are defined as follows [31]:

B(a,b)\displaystyle B(a,b) =01ta1(1t)b1𝑑t=(a1)!(b1)!(a+b1)!\displaystyle=\int_{0}^{1}t^{a-1}(1-t)^{b-1}dt=\frac{(a-1)!(b-1)!}{(a+b-1)!}
Bx(a,b)\displaystyle B_{x}(a,b) =0xta1(1t)b1𝑑t,0x1\displaystyle=\int_{0}^{x}t^{a-1}(1-t)^{b-1}dt,\quad 0\leq x\leq 1
Ix(a,b)\displaystyle I_{x}(a,b) =Bx(a,b)B(a,b),0x1\displaystyle=\frac{B_{x}(a,b)}{B(a,b)},\quad 0\leq x\leq 1 (45)

Using these definitions, it can easily be shown that

I1/2(r,r)=1/2,r>0\displaystyle I_{1/2}(r,r)=1/2,\quad r>0 (46)

Proof: B(r,r)=01tr1(1t)r1𝑑t=2012tr1(1t)r1𝑑t=2B1/2(r,r)B(r,r)=\int_{0}^{1}t^{r-1}(1-t)^{r-1}dt=2\int_{0}^{\frac{1}{2}}t^{r-1}(1-t)^{r-1}dt=2B_{1/2}(r,r) where we divided the integral into two parts since the function (tt2)r1(t-t^{2})^{r-1} is symmetric around 1/2. Using the fact that B1/2(r,r)=I1/2(r,r)B(r,r)B_{1/2}(r,r)=I_{1/2}(r,r)B(r,r), we can conclude that I1/2(r,r)=1/2I_{1/2}(r,r)=1/2.

The cumulative distribution function F(a;n,p)F(a;n,p) of a Binomial random variable XB(n,p)X\sim B(n,p) can be expressed using the regularized incomplete beta function as:

F(a;n,p)\displaystyle F(a;n,p) =[Xa]=I1p(na,a+1)\displaystyle=\mathbb{P}[X\leq a]=I_{1-p}(n-a,a+1)
=(na)(na)01ptna1(1t)a𝑑t\displaystyle=(n-a){n\choose a}\int_{0}^{1-p}t^{n-a-1}(1-t)^{a}dt (47)
Lemma IV.2

[31, Eq. 8.17.4]: For a,b>0, 0x1a,b>0,\ 0\leq x\leq 1,

Ix(a,b)=1I1x(b,a)\displaystyle I_{x}(a,b)=1-I_{1-x}(b,a) (48)
Lemma IV.3

[31, Eq. 8.17.20]: For a,b>0, 0x1a,b>0,\ 0\leq x\leq 1,

Ix(a+1,b)=Ix(a,b)xa(1x)baB(a,b)\displaystyle I_{x}(a+1,b)=I_{x}(a,b)-\frac{x^{a}(1-x)^{b}}{aB(a,b)} (49)
Lemma IV.4

The equation Iα(r,r)=cαI_{\alpha}(r,r)=c\alpha has only one solution when r>0r>0, r+r\in\mathbb{Z}^{+}, 0<α1/20<\alpha\leq 1/2 and 0<c10<c\leq 1.

Proof: First, α=0\alpha=0 is a solution to this equation since when α=0\alpha=0, both Iα(r,r)I_{\alpha}(r,r) and cαc\alpha are zero. Also, when c=1c=1, α=1/2\alpha=1/2 is a solution of the equation since I1/2(r,r)=1/2I_{1/2}(r,r)=1/2. The derivative of both terms with respect to α\alpha is:

(Iα(r,r))α\displaystyle\frac{\partial(I_{\alpha}(r,r))}{\partial\alpha} =αr1(1α)r1B(r,r),(cα)α=c\displaystyle=\frac{\alpha^{r-1}(1-\alpha)^{r-1}}{B(r,r)},\quad\frac{\partial(c\alpha)}{\partial\alpha}=c (50)

It can be seen that the derivative of cαc\alpha is a constant, and (α(1α))r1B(r,r)=0\frac{\left(\alpha(1-\alpha)\right)^{r-1}}{B(r,r)}=0 when α=0\alpha=0 and (α(1α))r1B(r,r)\frac{\left(\alpha(1-\alpha)\right)^{r-1}}{B(r,r)} is monotone increasing in the range 0<α1/20<\alpha\leq 1/2. For the case where c=1c=1; α=0\alpha=0 and α=1/2\alpha=1/2 is a solution to the equation, hence for some 0<α<1/20<\alpha^{**}<1/2 that satisfies (α(1α))r1B(r,r)=1\frac{\left(\alpha^{**}(1-\alpha^{*})\right)^{r-1}}{B(r,r)}=1, it must hold that (Iα(r,r))α<(α)α=1\frac{\partial(I_{\alpha}(r,r))}{\partial\alpha}<\frac{\partial(\alpha)}{\partial\alpha}=1 when 0<α<α0<\alpha<\alpha^{**}, and (Iα(r,r))α>(α)α=1\frac{\partial(I_{\alpha}(r,r))}{\partial\alpha}>\frac{\partial(\alpha)}{\partial\alpha}=1 when α<α1/2\alpha^{**}<\alpha\leq 1/2. This is because if such α\alpha^{**} such that 0<α<1/20<\alpha^{**}<1/2 does not exist, α=1/2\alpha=1/2 can’t be a solution to the equation Iα(r,r)=cαI_{\alpha}(r,r)=c\alpha. Now, considering the case for arbitrary 0<c10<c\leq 1, since αr(1α)rB(r,r)=0\frac{\alpha^{r}(1-\alpha)^{r}}{B(r,r)}=0 is monotone increasing, there can only be one 0<α<1/20<\alpha^{*}<1/2 such that (α)r(1α)rB(r,r)=c\frac{(\alpha^{*})^{r}(1-\alpha^{*})^{r}}{B(r,r)}=c. This means that cαc\alpha is increasing faster than Iα(r,r)I_{\alpha}(r,r) in the region 0<α<α0<\alpha<\alpha^{*}, hence there can’t be a solution to Iα(r,r)=cαI_{\alpha}(r,r)=c\alpha in this region. Further, Iα(r,r)I_{\alpha}(r,r) is increasing faster than cαc\alpha in the region α<α1/2\alpha^{*}<\alpha\leq 1/2, hence there can be at most one solution to the equation Iα(r,r)=cαI_{\alpha}(r,r)=c\alpha in the region 0<α1/20<\alpha\leq 1/2. Now, consider the fact that Iα(r,r)<cαI_{\alpha^{*}}(r,r)<c\alpha^{*}, and I1/2(r,r)=1/2c/2I_{1/2}(r,r)=1/2\geq c/2 when α=1/2\alpha=1/2. Combining this with the fact that both functions are continuous, there must be at least one solution to the equation Iα(r,r)=cαI_{\alpha}(r,r)=c\alpha for 0<c10<c\leq 1 in the range α<α1/2\alpha^{*}<\alpha\leq 1/2. Combining this with previous statement (that there can be at most one solution in this range), it can be concluded that there is only one solution to the equation Iα(r,r)=cαI_{\alpha}(r,r)=c\alpha in the range 0<α1/20<\alpha\leq 1/2 where 0<c<10<c<1.

IV-G Proof of Theorem III.5

To prove Theorem III.5, we need to show that for any r+r\in\mathbb{Z}^{+}, the random K-out graph (n;Kn)\mathbb{H}(n;K_{n}) is rr-robust whp if Kn2rK_{n}\geq 2r. To do this, similar to the proof given in [29] for Erdős-Rényi graphs, we will first find an upper bound on the probability of a subset of given size being not rr-reachable, and then use this result to show that the probability of not being rr-robust goes to zero when nn\to\infty and Kn2rK_{n}\geq 2r. Different from the prior work which relied on the commonly used upper bounds for the binomial coefficients (nk)(enk)k{n\choose k}\leq\left(\frac{en}{k}\right)^{k} and the union bound [29, 24] to bound the probability of a subset of given size being not rr-reachable, our proof uses the Beta function B(a,b)B(a,b) and its properties described in the previous Section to achieve tighter bounds. This in turn enables us to establish a tighter threshold for the rr-robustness of random K-out graphs than what was previously possible; e.g., see [24].

First, let n(Kn,r;S)\mathcal{E}_{n}(K_{n},r;S) denote the event that SVS\subset V is an rr-reachable set as per Definition II.7. The event n(Kn,r;S)\mathcal{E}_{n}(K_{n},r;S) occurs if there exists at least one node in SS that is adjacent to at least rr nodes in ScS^{c}, the subset comprised of nodes outside the subset SS. Thus, we have

n(Kn,r;S)=i𝒩S{(j𝒩Sc𝟙{vivj})r}\displaystyle\mathcal{E}_{n}(K_{n},r;S)=\bigcup_{i\in\mathcal{N}_{S}}\left\{\left(\sum_{j\in\mathcal{N}_{S^{c}}}\mathds{1}\left\{v_{i}\sim v_{j}\right\}\right)\geq r\right\}

with 𝒩S\mathcal{N}_{S}, 𝒩Sc\mathcal{N}_{S^{c}} denoting the set of labels of the vertices in SS and ScS^{c}, respectively, and 𝟙{}\mathds{1}\{\} denoting the indicator function. We are also interested in the complement of this event, denoted as (n(Kn,r;S))c\left(\mathcal{E}_{n}(K_{n},r;S)\right)^{\rm c}, which occurs if all nodes in SS are adjacent to less than r nodes in ScS^{c}. This can be written as

(nc(Kn,r;S))=i𝒩S{(j𝒩Sc𝟙{vivj})<r}.\displaystyle\left(\mathcal{E}_{n}^{\rm c}(K_{n},r;S)\right)=\bigcap_{i\in\mathcal{N}_{S}}\left\{\left(\sum_{j\in\mathcal{N}_{S^{c}}}\mathds{1}\left\{v_{i}\sim v_{j}\right\}\right)<r\right\}.

Note that at least one subset in every disjoint subset pairs needs to be rr-reachable per the definition of rr-robustness, hence one of the events n(Kn,r;S)\mathcal{E}_{n}(K_{n},r;S) or n(Kn,r;S)\mathcal{E}_{n}(K_{n},r;S^{\prime}) need to hold with high probability for every disjoint subset pairs S,SS,S^{\prime} of VV. Now, let 𝒵(Kn,r)\mathcal{Z}(K_{n},r) denote the event that both subsets in at least one of the disjoint subset pairs S,SVS,S^{\prime}\subset V are rr-reachable. Thus, we have

𝒵(Kn,r)\displaystyle\mathcal{Z}(K_{n},r) =S,S𝒫n:|S|n2[(n(Kn,r;S))c(n(Kn,r;Sc))c],\displaystyle=\bigcup_{S,S^{\prime}\in\mathcal{P}_{n}:~{}|S|\leq\lfloor\frac{n}{2}\rfloor}\left[(\mathcal{E}_{n}({K}_{n},r;S))^{\rm c}\wedge(\mathcal{E}_{n}({K}_{n},r;S^{\rm c}))^{\rm c}\right],

where SS=S\cap S^{\prime}=\emptyset, 𝒫n\mathcal{P}_{n} is the collection of all non-empty subsets of VV and since for each SS we check the rr-reachability of both SS and ScS^{\rm c}, the condition |S|n2|S|\leq\lfloor\frac{n}{2}\rfloor is used to prevent counting each subset twice. From this defintion, it can be seen that the graph (n;Kn)\mathbb{H}(n;K_{n}) is rr-robust if the event 𝒵(Kn,r)\mathcal{Z}(K_{n},r) does not occur. Using union bound, we get

PZ\displaystyle P_{Z} |S|n2[(n(Kn,r;S))c(n(Kn,r;S))c]\displaystyle\leq\sum_{|S|\leq\lfloor\frac{n}{2}\rfloor}\mathbb{P}[\left(\mathcal{E}_{n}(K_{n},r;S)\right)^{\rm c}\wedge\left(\mathcal{E}_{n}(K_{n},r;S^{\prime})\right)^{\rm c}]
|S|n2[(n(Kn,r;S))c]\displaystyle\leq\sum_{|S|\leq\lfloor\frac{n}{2}\rfloor}\mathbb{P}[\left(\mathcal{E}_{n}(K_{n},r;S)\right)^{\rm c}]
=m=1n2Sm𝒫n,m[(n(Kn,r;Sm))c],\displaystyle=\sum_{m=1}^{\left\lfloor\frac{n}{2}\right\rfloor}\sum_{S_{m}\in\mathcal{P}_{n,m}}\mathbb{P}[\left(\mathcal{E}_{n}(K_{n},r;S_{m})\right)^{\rm c}], (51)

where 𝒫n,m\mathcal{P}_{n,m} denotes the collection of all subsets of VV with exactly mm elements, and let Sm𝒫n,mS_{m}\in\mathcal{P}_{n,m} be a subset of the vertex set VV with size mm, i.e. SmVS_{m}\subset V and |Sm|=m|S_{m}|=m. Further, [𝒵(Kn,r)]\mathbb{P}\left[\mathcal{Z}(K_{n},r)\right] is abbreviated as PZ:=[𝒵(Kn,r)]P_{Z}:=\mathbb{P}\left[\mathcal{Z}(K_{n},r)\right]. From the exchangeability of the node labels and associated random variables, we have

Sm𝒫n,m[(n(Kn,r;Sm))c]=(nm)[(n(Kn,r;Sm))c].\displaystyle\sum_{S_{m}\in\mathcal{P}_{n,m}}\mathbb{P}[\left(\mathcal{E}_{n}(K_{n},r;S_{m})\right)^{\rm c}]={n\choose m}\mathbb{P}[\left(\mathcal{E}_{n}(K_{n},r;S_{m})\right)^{\rm c}]. (52)

since |𝒫n,m|=(nm)|\mathcal{P}_{n,m}|={n\choose m}, as there are (nm){n\choose m} subsets of VV with mm elements. Substituting this into (51), we obtain

PZm=1n2(nm)[(n(Kn,r;Sm))c]\displaystyle P_{Z}\leq\sum_{m=1}^{\left\lfloor\frac{n}{2}\right\rfloor}{n\choose m}\mathbb{P}[\left(\mathcal{E}_{n}(K_{n},r;S_{m})\right)^{\rm c}]

Before evaluating this expression, we will start with evaluating the probability that the set SmS_{m} is not rr-reachable, abbreviated as [(n(Kn,r;Sm))c]:=PSm\mathbb{P}[\left(\mathcal{E}_{n}(K_{n},r;S_{m})\right)^{\rm c}]:=P_{S_{m}}. Since a node vSmv\in S_{m} can have neighbors in SmcS_{m}^{c} if it forms an edge with nodes in SmcS_{m}^{c} or if nodes in SmcS_{m}^{c} forms edges with node vv, let PSm,1P_{S_{m,1}} denote the probability that all nodes vSmv\in S_{m} form an edge with less than rr nodes in SmcS_{m}^{c}, and let PSm,2P_{S_{m,2}} denote the probability that for each node vSmv\in S_{m}, nodes in SmcS_{m}^{c} form less than rr edges with them. Evidently, PSmPSm,1PSm,2P_{S_{m}}\leq P_{S_{m,1}}\cdot P_{S_{m,2}}. Further, let Pvm,1P_{v_{m,1}} denote the probability that a node vSmv\in S_{m} forms an edge with less than rr nodes in SmcS_{m}^{c}, and let Pvm,2P_{v_{m,2}} denote the probability that nodes in SmcS_{m}^{c} form less than rr edges with the node vSmv\in S_{m}.

Lemma IV.5

The probability that the node vSmv\in S_{m} chooses less than rr nodes in the set SmcS_{m}^{c}, denoted as Pvm,1P_{v_{m,1}}, can be upper bounded by the cumulative distribution function F(r1;Kn,p)F(r-1;K_{n},p) of a binomial random variable with KnK_{n} trials and success probability p=nmr+1nrp=\frac{n-m-r+1}{n-r}.

Proof: For node vv, after making one selection, the number of nodes available to choose from decreases so the probability of choosing a node in SmcS_{m}^{c} changes at each selection. For example, the probability of choosing a node in SmcS_{m}^{c} in the first selection is nmn1\frac{n-m}{n-1} and the probability of choosing a node in SmcS_{m}^{c} in the second selection is nm1n2\frac{n-m-1}{n-2} if a node in SmcS_{m}^{c} was selected in the first selection and it is nmn2\frac{n-m}{n-2} otherwise. Based on this, the probability of selecting a node in SmcS_{m}^{c} at the ithi^{th} selection out of KnK_{n} selections can be expressed as nmjni\frac{n-m-j}{n-i}, 1iKn1\leq i\leq K_{n}, 0j<i0\leq j<i where jj denotes the number of nodes already chosen from the set SmcS_{m}^{c} before the ithi^{th} selection. Since we are considering the case of choosing less than rr nodes in SmcS_{m}^{c}, we have that j<rj<r, and with this constraint the lowest possible value of nmjni\frac{n-m-j}{n-i} occurs when j=r1j=r-1 and i=ri=r, and hence it is nmr+1nr\frac{n-m-r+1}{n-r}. This gives a lower bound on the probability of selecting a node in SmcS_{m}^{c} in one of the KnK_{n} selections and at the same time is an upper bound on the probability of not selecting a node in SmcS_{m}^{c} in one of the KnK_{n} selections, and hence it is an upper bound for choosing less than rr nodes.

Next, using this upper bound, we plug in n=Knn=K_{n} and p=nmr+1nrp=\frac{n-m-r+1}{n-r} to (47), then we have

Pvm,1\displaystyle P_{v_{m,1}} F(r1;m,1m1nr)=Im1nr(Knr+1,r)\displaystyle\leq F\left(r-1;m,1-\frac{m-1}{n-r}\right)=I_{\frac{m-1}{n-r}}(K_{n}-r+1,r)
=(Knr+1)(Knr1)0m1nrtKnr(1t)r1𝑑t\displaystyle=(K_{n}-r+1){K_{n}\choose r-1}\int_{0}^{\frac{m-1}{n-r}}t^{K_{n}-r}(1-t)^{r-1}dt (53)

The selections of each node in SmS_{m} are independent, hence we can use (PSm,1)=(Pvm,1)m(P_{S_{m,1}})=(P_{v_{m,1}})^{m}.

In order to find Pvm,2P_{v_{m,2}}, a node in SmcS_{m}^{c} forming an edge with the node vv can be modeled as a Bernoulli trial with probability p=Knn1p=\frac{K_{n}}{n-1} so the event that nodes in SmcS_{m}^{c} forming less than rr edges with the node vv can be represented by a Binomial model with nmn-m trials and p=Knn1p=\frac{K_{n}}{n-1}. Hence,

Pvm,2\displaystyle P_{v_{m,2}} =F(r1;nm,Knn1)=InKn1n1(nmr+1,r)\displaystyle=F\left(r-1;n-m,\frac{K_{n}}{n-1}\right)=I_{\frac{n-K_{n}-1}{n-1}}(n-m-r+1,r) (54)

Since the nodes in SmcS_{m}^{c} forming edges with nodes in SmS_{m} are not independent of the other nodes in SmS_{m}, we cannot write (PSm,2)(Pvm,2)m(P_{S_{m,2}})\leq(P_{v_{m,2}})^{m}. To find (PSm,2)(P_{S_{m,2}}), we will decompose it into the following conditional probabilities.

PSm,2=\displaystyle P_{S_{m,2}}= [dv1<r][dv2<r|dv1<r]\displaystyle\mathbb{P}[d_{v_{1}}<r]\cdot\mathbb{P}[d_{v_{2}}<r|d_{v_{1}}<r]\cdot
[dvm<r|dv1<r,dv2<r,,dvm1<r]\displaystyle\ldots\mathbb{P}[d_{v_{m}}<r|d_{v_{1}}<r,d_{v_{2}}<r,\ldots,d_{v_{m-1}}<r] (55)

where v1,v2,,vmSmv_{1},v_{2},\ldots,v_{m}\in S_{m} represent all the nodes in SmS_{m}, and dvid_{v_{i}} is used to denote the number of nodes in SmcS_{m}^{c} that form an edge with the node viv_{i}. To find an upper bound on PSm,2P_{S_{m,2}}, we consider the worst case. In the worst case, all the preceding nodes are selected by nodes in SmcS_{m}^{c} exactly r1r-1 times. This reduces the number of available edges in SmcS_{m}^{c} that can make connections with the remaining nodes in SmS_{m}, hence increases the probability of nodes in SmcS_{m}^{c} forming less than rr edges with the remaining nodes in SmS_{m}. Hence, we can write:

PSm,2\displaystyle P_{S_{m,2}} [dv1<r][dv2<r|dv1=r1]\displaystyle\leq\mathbb{P}[d_{v_{1}}<r]\cdot\mathbb{P}[d_{v_{2}}<r|d_{v_{1}}=r-1]\cdot
[dvm<r|dv1=r1,,dvm1=r1]\displaystyle\ldots\mathbb{P}[d_{v_{m}}<r|d_{v_{1}}=r-1,\ldots,d_{v_{m-1}}=r-1] (56)

Consider the general case for [dva+1<r|dv1=r1,,dva=r1]\mathbb{P}[d_{v_{a+1}}<r|d_{v_{1}}=r-1,\ldots,d_{v_{a}}=r-1] where 1am11\leq a\leq m-1. Assume that q1q_{1} nodes in SmcS_{m}^{c} formed an edge with only one node among the nodes v1,,vav_{1},\ldots,v_{a}. Similarly, assume q2q_{2} nodes in SmcS_{m}^{c} formed an edge with only two nodes, and so on (qKnq_{K_{n}} nodes in SmcS_{m}^{c} formed edges with KnK_{n} nodes in v1,,vav_{1},\ldots,v_{a}). Also define q=q1+q2++qKnq=q_{1}+q_{2}+\ldots+q_{K_{n}}. It can be seen that a=q1+2q2++KnqKnr1a=\frac{q_{1}+2q_{2}+\ldots+K_{n}q_{K_{n}}}{r-1}. Here, we have n0=nmqn_{0}=n-m-q nodes that did not use any of its selections yet, so their probability of choosing the node vav_{a} is p0=Knna1p_{0}=\frac{K_{n}}{n-a-1}. Similarly, we have n1=q1n_{1}=q_{1} nodes that used one of its selections, so their probability is p1=Kn1na1p_{1}=\frac{K_{n}-1}{n-a-1} (for nKn=qKnn_{K_{n}}=q_{K_{n}} nodes pKn=KnKnna1=0p_{K_{n}}=\frac{K_{n}-K_{n}}{n-a-1}=0).

Considering the selection of each node in SmcS_{m}^{c} as a Bernoulli trial with different probabilities, the collection of all such trials defines a Poisson Binomial distribution XPB([p0]n0,,[pKn]nKn)X\sim PB([p_{0}]^{n_{0}},\ldots,[p_{K_{n}}]^{n_{K_{n}}}). The mean of this distribution is:

μX\displaystyle\mu_{X} =(nmq)Knna1+i=1KnqiKnina1\displaystyle=(n-m-q)*\frac{K_{n}}{n-a-1}+\sum_{i=1}^{K_{n}}q_{i}\frac{K_{n}-i}{n-a-1}
=(nm)Kna(r1)na1\displaystyle=\frac{(n-m)K_{n}-a(r-1)}{n-a-1} (57)

Assuming Kn>2r2K_{n}>2r-2 and using mn/2m\leq\lfloor n/2\rfloor, we have:

μX=(2n2ma)na1(r1)+Kn2r+2na1\displaystyle\mu_{X}=\frac{(2n-2m-a)}{n-a-1}*(r-1)+\frac{K_{n}-2r+2}{n-a-1} (58)

Since the mode of the Poisson Binomial distribution satisfies ψXμX+1\psi_{X}\leq\mu_{X}+1 [32], we have ψX>r1\psi_{X}>r-1 for any 1am11\leq a\leq m-1 if Kn>2r2K_{n}>2r-2. Since for any distribution, the cumulative distribution function is equal to 1/2 at the mode of distribution, we have that [dva+1<r|dv1=r1,,dva=r1]<1/2\mathbb{P}[d_{v_{a+1}}<r|d_{v_{1}}=r-1,\ldots,d_{v_{a}}=r-1]<1/2. Hence,

PSm,2<(12)m\displaystyle P_{S_{m,2}}<\left(\frac{1}{2}\right)^{m} (59)

Now, using the fact that PSmPSm,1PSm,2P_{S_{m}}\leq P_{S_{m,1}}\cdot P_{S_{m,2}}, we have:

PSm\displaystyle P_{S_{m}} PSm,1PSm,2(Pvm,1)m(12)m\displaystyle\leq P_{S_{m,1}}\cdot P_{S_{m,2}}\leq(P_{v_{m,1}})^{m}\cdot\left(\frac{1}{2}\right)^{m}
(12Im1nr(Knr+1,r))m\displaystyle\leq\left(\frac{1}{2}I_{\frac{m-1}{n-r}}(K_{n}-r+1,r)\right)^{m}

Let Pm:=(nm)PSmP_{m}:={n\choose m}P_{S_{m}}. Then, we have:

PZm=1n2(nm)PSm=m=1n2Pm\displaystyle P_{Z}\leq\sum_{m=1}^{\left\lfloor\frac{n}{2}\right\rfloor}{n\choose m}P_{S_{m}}=\sum_{m=1}^{\left\lfloor\frac{n}{2}\right\rfloor}P_{m} (60)

We will divide the summation into three parts as follows:

PZ=m=1n/2Pm=m=1log(n)Pm+m=log(n)α(nr)Pm\displaystyle P_{Z}=\sum_{m=1}^{\lfloor n/2\rfloor}P_{m}=\sum_{m=1}^{\lfloor\log(n)\rfloor}P_{m}+\sum_{m=\lceil\log(n)\rceil}^{\lfloor\alpha^{*}(n-r)\rfloor}P_{m}
+m=α(nr)n/2Pm=P1+P2+P3\displaystyle+\sum_{m=\lceil\alpha^{*}(n-r)\rceil}^{\lfloor n/2\rfloor}P_{m}=P_{1}+P_{2}+P_{3} (61)

where α\alpha^{*} is the solution to the equation Iα(r,r)=nrneαI_{\alpha}(r,r)=\frac{n-r}{ne}\cdot\alpha in the range 0<α<120<\alpha<\frac{1}{2}. (The purpose of defining α\alpha^{*} this way will be given later in the proof.) Start with the first summation P1P_{1} and use (14) along with (nm)(enm)m{n\choose m}\leq\left(\frac{en}{m}\right)^{m}, then we have:

Pm\displaystyle P_{m} =(nm)PSm(nm)PSm,1PSm,2(nm)PSm,1\displaystyle={n\choose m}P_{S_{m}}\leq{n\choose m}P_{S_{m,1}}\cdot P_{S_{m,2}}\leq{n\choose m}P_{S_{m,1}}
(enmIm1nr(Knr+1,r))m\displaystyle\leq\left(\frac{en}{m}I_{\frac{m-1}{n-r}}(K_{n}-r+1,r)\right)^{m}
(enmB(Knr+1,r)0m1nrtKnr(1t)r1𝑑t)m\displaystyle\leq\left(\frac{en}{mB(K_{n}-r+1,r)}\int_{0}^{\frac{m-1}{n-r}}t^{K_{n}-r}(1-t)^{r-1}dt\right)^{m}
(enmB(Knr+1,r)0m1nrtKnr𝑑t)m\displaystyle\leq\left(\frac{en}{mB(K_{n}-r+1,r)}\int_{0}^{\frac{m-1}{n-r}}t^{K_{n}-r}dt\right)^{m}
(en(m1nr)Knr+1mB(Knr+1,r)(Knr+1))m\displaystyle\leq\left(\frac{en\left(\frac{m-1}{n-r}\right)^{K_{n}-r+1}}{mB(K_{n}-r+1,r)(K_{n}-r+1)}\right)^{m}
(e(1+r/m)B(Knr+1,r)(Knr+1)(m1nr)Knr)m\displaystyle\leq\left(\frac{e(1+r/m)}{B(K_{n}-r+1,r)(K_{n}-r+1)}\left(\frac{m-1}{n-r}\right)^{K_{n}-r}\right)^{m}
(e(1+r)(log(n)1nr)KnrB(Knr+1,r)(Knr+1))m:=(an)m\displaystyle\leq\left(\frac{e(1+r)\left(\frac{\log(n)-1}{n-r}\right)^{K_{n}-r}}{B(K_{n}-r+1,r)(K_{n}-r+1)}\right)^{m}:=(a_{n})^{m}

For Kn>rK_{n}>r, since B(Knr+1,r)B(K_{n}-r+1,r) and rr are finite values, we have limnan=0\underset{n\rightarrow\infty}{\lim}a_{n}=0 by virtue of limn(log(n)1nr)Knr=0\underset{n\rightarrow\infty}{\lim}\left(\frac{\log(n)-1}{n-r}\right)^{K_{n}-r}=0. Using this, we can express the summation as:

P1=m=1log(n)(an)man1(an)log(n)1an\displaystyle P_{1}=\sum_{m=1}^{\lfloor\log(n)\rfloor}(a_{n})^{m}\leq a_{n}\cdot\frac{1-(a_{n})^{\log(n)}}{1-a_{n}} (62)

where the geometric sum converges by virtue of limnan=0\lim_{n\to\infty}a_{n}=0, leading to P1P_{1} converging to zero for large nn.

Now, similarly consider the second summation P2P_{2}. Using (nm)(enm)m{n\choose m}\leq\left(\frac{en}{m}\right)^{m}, we have

Pm\displaystyle P_{m} (nm)PSm(nm)(PSm,1)m\displaystyle\leq{n\choose m}P_{S_{m}}\leq{n\choose m}(P_{S_{m,1}})^{m}
(enmIm1nr(Knr+1,r))m:=(am)m\displaystyle\leq\left(\frac{en}{m}I_{\frac{m-1}{n-r}}(K_{n}-r+1,r)\right)^{m}:=(a_{m})^{m} (63)

where am:=enmIm1nr(Knr+1,r)a_{m}:=\frac{en}{m}\cdot I_{\frac{m-1}{n-r}}(K_{n}-r+1,r). Assume that Kn>2r1K_{n}>2r-1. It can be shown that Im+rn(Knr+1)<Im+rn(r,r)I_{\frac{m+r}{n}}(K_{n}-r+1)<I_{\frac{m+r}{n}}(r,r) as a consequence of the property (49). Hence, we have

am\displaystyle a_{m} <enmIm1nr(r,r)enmImnr(r,r)\displaystyle<\frac{en}{m}\cdot I_{\frac{m-1}{n-r}}(r,r)\leq\frac{en}{m}\cdot I_{\frac{m}{n-r}}(r,r) (64)

Assume that α=α\alpha=\alpha^{*} is the solution of the equation Iα(r,r)=nrneαI_{\alpha}(r,r)=\frac{n-r}{ne}\cdot\alpha in the range 0<α<120<\alpha<\frac{1}{2}. From Lemma IV.4, we know that this equation has only one solution in this range, and Iα(r,r)α(nr)neI_{\alpha}(r,r)\leq\frac{\alpha^{*}(n-r)}{ne} for 0<αα0<\alpha\leq\alpha^{*}. Plugging in α=mnr\alpha=\frac{m}{n-r}, we have Imnr(r,r)αmneI_{\frac{m}{n-r}}(r,r)\leq\frac{\alpha^{*}m}{ne}, which leads to am<1a_{m}<1 when mα(nr)m\leq\lfloor\alpha^{*}(n-r)\rfloor. Denoting a:=maxm(am)a:=\max_{m}(a_{m}), we have

P2=m=log(n)α(nr)(am)mm=log(n)(a)malog(n)1a\displaystyle P_{2}=\sum_{m=\lceil\log(n)\rceil}^{\lfloor\alpha^{*}(n-r)\rfloor}(a_{m})^{m}\leq\sum_{m=\lceil\log(n)\rceil}^{\infty}(a)^{m}\leq\frac{a^{\log(n)}}{1-a} (65)

where the geometric sum converges by virtue of limna<1\lim_{n\to\infty}a<1, leading to P2P_{2} converging to zero as nn gets large.

Now, similarly consider the third summation P3P_{3}. Assuming Kn2r1K_{n}\geq 2r-1, we have

Pm\displaystyle P_{m} (nm)(PSm,1)m(PSm,2)m\displaystyle\leq{n\choose m}(P_{S_{m,1}})^{m}\cdot(P_{S_{m,2}})^{m}
<(12(nm)1mIm1nr(Knr+1,r))m\displaystyle<\left(\frac{1}{2}{n\choose m}^{\frac{1}{m}}I_{\frac{m-1}{n-r}}(K_{n}-r+1,r)\right)^{m}
<(12(nm)1mImnr(r,r))m:=(am)m\displaystyle<\left(\frac{1}{2}{n\choose m}^{\frac{1}{m}}I_{\frac{m}{n-r}}(r,r)\right)^{m}:=(a_{m})^{m} (66)

From the previous summation P2P_{2} and the proof of Lemma IV.4, we know that in the range αnrmn/2\lceil\alpha^{*}n-r\rceil\geq m\geq\lfloor n/2\rfloor; nm\frac{n}{m} is decreasing, Imnr(r,r)I_{\frac{m}{n-r}}(r,r) is increasing, and the overall expression nmImnr(r,r)\frac{n}{m}\cdot I_{\frac{m}{n-r}}(r,r) is also increasing. Hence, nmImnr(r,r)\frac{n}{m}\cdot I_{\frac{m}{n-r}}(r,r) takes its maximum value when m=n/2m=\lfloor n/2\rfloor. Since enm\frac{en}{m} is an upper bound of (nm)1m{n\choose m}^{\frac{1}{m}}, the expression ama_{m} also takes its maximum value when m=n/2m=\lfloor n/2\rfloor. Denoting a:=maxm(am)=an/2a:=\max_{m}(a_{m})=a_{\lfloor n/2\rfloor}, and using Stirling’s approximation limn(2nn)4nπn<4nπ\underset{n\rightarrow\infty}{\lim}{2n\choose n}\sim\frac{4^{n}}{\sqrt{\pi n}}<\frac{4^{n}}{\sqrt{\pi}} as a finer upper bound, we have

limna=limnan/2<124π12<1\displaystyle\underset{n\rightarrow\infty}{\lim}a=\underset{n\rightarrow\infty}{\lim}a_{\lfloor n/2\rfloor}<\frac{1}{2}\cdot\frac{4}{\sqrt{\pi}}\cdot\frac{1}{2}<1 (67)

From this, we have:

P3\displaystyle P_{3} m=α(nr)n2(am)nn2(a)n\displaystyle\leq\sum_{m=\lceil\alpha^{*}(n-r)\rceil}^{\lfloor\frac{n}{2}\rfloor}(a_{m})^{n}\leq\frac{n}{2}(a)^{n} (68)

where the sum converges by virtue of limna=1\underset{n\rightarrow\infty}{\lim}a=1 since limnn(a)n=0\underset{n\rightarrow\infty}{\lim}n(a)^{n}=0 in this case, leading to P3P_{3} converging to zero as nn gets large. Since P1P_{1}, P2P_{2}, and P3P_{3} all converge to zero as nn gets large, PZ=P1+P2+P3P_{Z}=P_{1}+P_{2}+P_{3} also converges to zero as nn gets large. This concludes the proof of Theorem III.5.

V Conclusion

In this paper, we provide a comprehensive set of results on the rr-robustness of the random K-out graph (n;Kn)\mathbb{H}(n;K_{n}), and the connectivity and giant component size of (n;Kn,γn)\mathbb{H}(n;K_{n},\gamma_{n}), i.e., random K-out graph with (randomly selected) γn\gamma_{n} nodes deleted. In addition to providing proofs of our results, we include computer simulations to validate our results in the finite node regime. To demonstrate the usefulness of the random K-out graphs, we compare our results on the random K-out graphs with results from ER graphs under similar settings, and determine that random K-out graphs attain rr-robustness, connectivity or the occurrence of a giant component of a given size at a significantly lower mean node degree value compared to ER graphs. These results reinforce the usefulness of random K-out graphs in applications that require a certain degree of robustness or tolerance to nodes failing, being captured, or being dishonest; such as federated learning, consensus dynamics, distributed averaging and wireless sensor networks.

References

  • [1] E. C. Elumar, M. Sood, and O. Yağan, “On the connectivity and giant component size of random k-out graphs under randomly deleted nodes,” in 2021 IEEE International Symposium on Information Theory (ISIT), 2021, pp. 2572–2577.
  • [2] E. C. Elumar and O. Yağan, “Analyzing r-robustness of random k-out graphs for the design of robust networks,” in ICC 2023 - IEEE International Conference on Communications, 2023, pp. 5946–5952.
  • [3] B. Bollobás, Random graphs.   Cambridge university press, 2001.
  • [4] T. I. Fenner and A. M. Frieze, “On the connectivity of random mm-orientable graphs and digraphs,” Combinatorica, vol. 2, no. 4, pp. 347–359, Dec 1982.
  • [5] O. Yağan and A. M. Makowski, “On the connectivity of sensor networks under random pairwise key predistribution,” IEEE Transactions on Information Theory, vol. 59, no. 9, pp. 5754–5762, Sept 2013.
  • [6] P. Erdős and A. Rényi, “On the strength of connectedness of random graphs,” Acta Math. Acad. Sci. Hungar, pp. 261–267, 1961.
  • [7] M. Penrose, Random geometric graphs.   OUP Oxford, 2003, vol. 5.
  • [8] O. Yağan and A. M. Makowski, “Zero–one laws for connectivity in random key graphs,” IEEE Transactions on Information Theory, vol. 58, no. 5, pp. 2983–2999, 2012.
  • [9] M. Sood and O. Yağan, “k-connectivity in random graphs induced by pairwise key predistribution schemes,” in 2020 IEEE International Symposium on Information Theory (ISIT), 2020, pp. 1343–1348.
  • [10] M. Sood and O. Yağan, “Towards k-connectivity in heterogeneous sensor networks under pairwise key predistribution,” in 2019 IEEE Global Communications Conference (GLOBECOM), 2019, pp. 1–6.
  • [11] O. Yağan and A. M. Makowski, “Modeling the pairwise key predistribution scheme in the presence of unreliable links,” IEEE Transactions on Information Theory, vol. 59, no. 3, pp. 1740–1760, March 2013.
  • [12] C. Sabater, A. Bellet, and J. Ramon, “Distributed differentially private averaging with improved utility and robustness to malicious parties,” 2020.
  • [13] G. Fanti, S. B. Venkatakrishnan, S. Bakshi, B. Denby, S. Bhargava, A. Miller, and P. Viswanath, “Dandelion++: Lightweight cryptocurrency networking with formal anonymity guarantees,” Proc. ACM Meas. Anal. Comput. Syst., vol. 2, no. 2, pp. 29:1–29:35, Jun. 2018.
  • [14] S. Sridhara, F. Wirz, J. de Ruiter, C. Schutijser, M. Legner, and A. Perrig, “Global distributed secure mapping of network addresses,” in Proceedings of the ACM SIGCOMM Workshop on Technologies, Applications, and Uses of a Responsible Internet (TAURIN), 2021.
  • [15] F. Yavuz, J. Zhao, O. Yağan, and V. Gligor, “Designing secure and reliable wireless sensor networks under a pairwise key predistribution scheme,” in 2015 IEEE International Conference on Communications (ICC).   IEEE, 2015, pp. 6277–6283.
  • [16] R. Eletreby and O. Yağan, “Connectivity of inhomogeneous random K-out graphs,” IEEE Transactions on Information Theory, vol. 66, no. 11, pp. 7067–7080, 2020.
  • [17] M. Sood and O. Yağan, “On the size of the giant component in inhomogeneous random k-out graphs,” in 2020 59th IEEE Conference on Decision and Control (CDC), 2020, pp. 5592–5597.
  • [18] O. Yağan and A. M. Makowski, “Wireless sensor networks under the random pairwise key predistribution scheme: Can resiliency be achieved with small key rings?” IEEE/ACM Transactions on Networking, vol. 24, no. 6, pp. 3383–3396, 2016.
  • [19] M. Sood and O. Yağan, “Tight bounds for the probability of connectivity in random k-out graphs,” in ICC 2021-IEEE International Conference on Communications.   IEEE, 2021, pp. 1–6.
  • [20] R. Merris, Graph theory, ser. Wiley-Interscience series in discrete mathematics and optimization.   New York: John Wiley, 2001.
  • [21] H. Zhang and S. Sundaram, “Robustness of information diffusion algorithms to locally bounded adversaries,” in 2012 American Control Conference (ACC), 2012, pp. 5855–5861.
  • [22] H. J. LeBlanc, H. Zhang, X. Koutsoukos, and S. Sundaram, “Resilient asymptotic consensus in robust networks,” IEEE Journal on Selected Areas in Communications, vol. 31, no. 4, pp. 766–781, 2013.
  • [23] H. Zhang and S. Sundaram, “Robustness of complex networks with implications for consensus and contagion,” in Decision and Control (CDC), 2012 IEEE 51st Annual Conference on.   IEEE, 2012, pp. 3426–3432.
  • [24] E. C. Elumar and O. Yağan, “Robustness of random k-out graphs,” in 2021 60th IEEE Conference on Decision and Control (CDC), 2021, pp. 5526–5531.
  • [25] O. Yağan and A. M. Makowski, “On the scalability of the random pairwise key predistribution scheme: Gradual deployment and key ring sizes,” Performance Evaluation, vol. 70, no. 7, pp. 493 – 512, 2013.
  • [26] A. Frieze and M. Karoński, Introduction to random graphs.   Cambridge University Press, 2016.
  • [27] A. Mei, A. Panconesi, and J. Radhakrishnan, “Unassailable sensor networks,” in Proc. of the 4th International Conference on Security and Privacy in Communication Netowrks, ser. SecureComm ’08.   New York, NY, USA: ACM, 2008.
  • [28] R. Di Pietro, L. V. Mancini, A. Mei, A. Panconesi, and J. Radhakrishnan, “Redoubtable sensor networks,” ACM Transactions on Information and System Security (TISSEC), vol. 11, no. 3, pp. 1–22, 2008.
  • [29] H. Zhang, E. Fata, and S. Sundaram, “A notion of robustness in complex networks,” IEEE Transactions on Control of Network Systems, vol. 2, no. 3, pp. 310–320, 2015.
  • [30] P. Erdős and A. Rényi, “On the evolution of random graphs,” Publ. Math. Inst. Hung. Acad. Sci, vol. 5, no. 1, pp. 17–60, 1960.
  • [31] NIST Digital Library of Mathematical Functions,” http://dlmf.nist.gov/8.17, Release 1.1.2 of 2021-06-15, f. W. J. Olver, A. B. Olde Daalhuis, D. W. Lozier, B. I. Schneider, R. F. Boisvert, C. W. Clark, B. R. Miller, B. V. Saunders, H. S. Cohl, and M. A. McClain, eds. [Online]. Available: http://dlmf.nist.gov/
  • [32] W. Hoeffding, “On the Distribution of the Number of Successes in Independent Trials,” The Annals of Mathematical Statistics, vol. 27, no. 3, pp. 713 – 721, 1956. [Online]. Available: https://doi.org/10.1214/aoms/1177728178