This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Robust Lower Bounds for Graph Problems in the Blackboard Model of Communication

Christian Konrad Department of Computer Science, University of Bristol [email protected] Peter Robinson Department of Computer Science, City University of Hong Kong [email protected]  and  Viktor Zamaraev Department of Computer Science, University of Liverpool [email protected]
(2020)
Abstract.

We give lower bounds on the communication complexity of graph problems in the multi-party blackboard model. In this model, the edges of an nn-vertex input graph are partitioned among kk parties, who communicate solely by writing messages on a shared blackboard that is visible to every party. We show that any non-trivial graph problem on nn-vertex graphs has blackboard communication complexity Ω(n)\Omega(n) bits, even if the edges of the input graph are randomly assigned to the kk parties. We say that a graph problem is non-trivial if the output cannot be computed in a model where every party holds at most one edge and no communication is allowed. Our lower bound thus holds for essentially all key graph problems relevant to distributed computing, including Maximal Independent Set (MIS), Maximal Matching, (Δ+1\Delta+1)-coloring, and Dominating Set. In many cases, e.g., MIS, Maximal Matching, and (Δ+1)(\Delta+1)-coloring, our lower bounds are optimal, up to poly-logarithmic factors.

journalyear: 2020copyright: acmlicensedconference: ; ;

1. Introduction

The multi-party blackboard model of communication is a well-established model of distributed computation (Drucker et al., 2012; Phillips et al., 2012; Braverman and Oshman, 2015; Vempala et al., 2020). In this model, the input is split among kk parties, who collectively need to solve a joint problem by communicating with each other. However, as opposed to point-to-point message passing models where pairs of parties have private communication links, the parties communicate solely via a shared blackboard that is visible to all parties. The objective is to design communication protocols with minimal communication complexity, i.e., protocols that instruct the parties to write as few bits onto the shared blackboard as possible. The model is asynchronous, and the order in which the parties write onto the blackboard is solely determined by the communication protocol.

In this paper, we initiate the study of fundamental graph problems in the blackboard model. Given an input graph G=(V,E)G=(V,E) with n=|V|n=|V|, we consider an edge-partition model, where the edges of GG are partitioned among the kk parties either adversarially or uniformly at random. Our main result is a lower bound that applies to a large class of graph problems and that is tight (up to poly-logarithmic factors) in many cases:

Theorem 1 (simplified).
Every non-trivial graph problem has (randomized) blackboard communication complexity of Ω(n)\Omega(n) bits, even if the edges of the input graph are partitioned randomly among the parties.

Informally, a problem is non-trivial if it does not admit a deterministic protocol where every party holds at most one edge of the input graph and no communication occurs at all (see Section 3 for a precise definition). Intuitively, the knowledge of at most one edge is not enough for the parties to solve any interesting graph problem, and, as proved in this paper, most graph problems, including Minimum Vertex Cover, Minimum Dominating Set, Maximal/Maximum Matching, Maximal Independent Set, and (Δ+1)(\Delta+1)-coloring, are therefore non-trivial.

In many cases, our lower bound is tight up to poly-logarithmic factors, even if the edge partitioning is adversarial. For Maximal Matching, obtaining a blackboard communication protocol with communication cost O(nlogn)O(n\log n) is trivial. For others, such as Maximal Independent Set and (Δ+1)(\Delta+1)-coloring, protocols with communication cost O(npolylogn)O(n\operatorname{\mathrm{poly}}\log n) can be derived from known data streaming algorithms. We refer the reader to Table 1 for an overview of upper bounds that match our lower bounds up to polylogn\operatorname{\mathrm{poly}}\log n factors. These upper bounds will be discussed in Section 4.

Problem CC Upper Bound Reference
Maximal Matching O(nlogn)O(n\log n) trivial
(1ϵ)(1-\epsilon)-approx. Maximum Matching (general graphs) O(nlogn(1ϵ)O(1ϵ))O(n\log n(\frac{1}{\epsilon})^{O(\frac{1}{\epsilon})}) McGregor (McGregor, 2005)
(1ϵ)(1-\epsilon)-approx. Maximum Matching (bipartite graphs) O(n(logn+1ϵlog(1ϵ))1ϵ2)O(n(\log n+\frac{1}{\epsilon}\log(\frac{1}{\epsilon}))\frac{1}{\epsilon^{2}}) Assadi et al. (Sepehr Assadi, 2021)
Maximal Independent Set O(npolylogn)O(n\operatorname{\mathrm{poly}}\log n) Ghaffari et al. (Ghaffari et al., 2018)
(Δ+1\Delta+1)-coloring O(nlog4n)O(n\log^{4}n) Assadi et al. (Assadi et al., 2019)
Table 1. Matching (up to poly-logarithmic factors) upper bounds to our Ω(n)\Omega(n) lower bound.

1.1. Related Work

Multi-party Communication Models

Yao introduced the two-party communication complexity framework in 1979 (Yao, 1979), which was extended to multiple parties by Chandra et al. a few years later (Chandra et al., 1983). The various multi-party communication complexity models known today can be categorized by 1) how input data is partitioned among the parties; and 2) how the parties communicate with each other. Regarding 1), in the number-on-the-forehead model, every party knows the input of all other parties except their own, which was also the model considered by Chandra et al. This stands in contrast to the number-in-hand model, where every party knows their own input, which is the model studied in this paper. Regarding 2), in the message passing model, every pair of vertices has access to a private communication link that is not visible to other parties. This model is similar to the coordinator model, where every party solely communicates with a third party referred to as the coordinator. The two models are identical up to a O(logk)O(\log k) factor in the communication complexity (see, for example, (Phillips et al., 2012)). The blackboard model can be regarded as a variant of the coordinator model, where every message received by the coordinator is immediately visible to every other party. Writing on the blackboard can also be regarded as a message broadcast.

Multi-party Communication Complexity of Graph Problems

Various graph problems have been studied in the multi-party number-in-hand communication settings. Typically, the edges of the input graph are distributed among the participating parties, either with or without duplication, with the without duplication setting usually being easier to handle. In the message passing model, tight bounds are known for testing bipartiteness, cycle-freeness, connectivity, and degree computations (Woodruff and Zhang, 2017), approximate maximum matching and approximate maximum flow (Huang et al., 2020), and spanner computations (Fernandez et al., 2020). In contrast, graph problems have not been considered in the blackboard model to the best of our knowledge. However, the Maximum Matching problem has been considered in the simultaneous message model, which can be seen as a variant of the blackboard model, where all parties simultaneously send a single message to a coordinator who then outputs the result (Dobzinski et al., 2014; Alon et al., 2015; Konrad, 2015; Assadi et al., 2016b).

Further Results in the Blackboard Model

While graph problems have not yet been studied in the blackboard model, the model has nevertheless attracted significant attention in recent years. For example, tight bounds are known for Set-Disjointness (Braverman and Oshman, 2015), bit-wise XOR and logical AND computations (Phillips et al., 2012), optimization problems such as solving linear systems (Vempala et al., 2020), and distributed task allocation problems (Drucker et al., 2012). We point out that the blackboard model also abstracts aspects of real-world graph processing systems with shared memory such as (Low et al., 2012) and, as observed in (Braverman and Oshman, 2015), models single-hop wireless ad hoc networks where nodes communicate via broadcast.

1.2. Techniques

Consider a graph problem P and the uniform distribution λ\lambda over all tt-vertex graphs, for some constant tt. We denote a graph chosen from λ\lambda as gadget. Our hard input graph distribution is the product distribution μ=λn/t\mu=\lambda^{n/t}, i.e., graphs consisting of n/tn/t disjoint and independent gadgets chosen from λ\lambda.

Let 𝒫\mathcal{P} be a 𝖡𝖡(k)\mathsf{BB}(k) communication protocol for problem P. We measure the amount of information about the edges of the input graph that is necessarily revealed by the transcript of 𝒫\mathcal{P}, i.e., by the bits written on the blackboard. This quantity is usually referred to as the external information cost (see (Braverman, 2017) for an excellent exposition) of a protocol 𝒫\mathcal{P} and constitutes a lower bound on the communication cost of 𝒫\mathcal{P}.

At the core of our lower bound argument lies a direct sum argument. Towards a contradiction, assume that the external information cost of 𝒫\mathcal{P} is δn\delta\cdot n, for a sufficiently small constant δ1t\delta\ll\frac{1}{t}. We then prove that 𝒫\mathcal{P} can be used to obtain a protocol 𝒬\mathcal{Q} with external information cost tδt\cdot\delta that solves P on a single tt-vertex gadget, i.e., on inputs chosen from λ\lambda. We further argue that since tδt\cdot\delta is a very small constant, 𝒬\mathcal{Q} cannot depend much on the actual input of the problem. We then give a compression lemma, showing that there is a protocol 𝒬\mathcal{Q}^{\prime} that avoids communication altogether, while only marginally increasing the probability of error. Finally, to show that such a “silent” protocol 𝒬\mathcal{Q}^{\prime} cannot exist, we employ Yao’s lemma, which allows us to focus on deterministic protocols with distributional error, and then give simple problem-specific combinatorial arguments.

1.3. Outline

In Section 2, we define the 𝖡𝖡(k)\mathsf{BB}(k) model and provide the necessary context on information theory. Next, we present our main result, our lower bound for non-trivial graph problems in the blackboard model, in Section 3. In Section 4, we summarize how known algorithms can be adapted to yield optimal (up to poly-logarithmic factors) communication protocols in the blackboard model. Last, we conclude in Section 5.

2. Preliminaries and Computing Models

2.1. The Blackboard Model

In the (shared) blackboard model, denoted by 𝖡𝖡(k)\mathsf{BB}(k), we have kk parties that communicate by writing messages on a shared blackboard that is visible to all parties. The way the parties interact, and, in particular, the order in which the parties write onto the blackboard, is solely specified by the given communication protocol (and the current state of the blackboard), i.e., the model is therefore asynchronous. All parties have access to both private and public random coins.

In this paper, we study graph problems in the blackboard model. Given an input graph G=(V,E)G=(V,E) with n=|V|n=|V|, we consider an edge-partition model without duplications, where the edges of GG are partitioned among the kk parties according to a partition function Z:[V×V][k]Z:[V\times V]\rightarrow[k] that maps the set of all pairs of vertices of GG to the machines (since we consider undirected graphs, we assume that Z(u,v)=Z(v,u)Z(u,v)=Z(v,u), for every u,vVu,v\in V). This function does not reveal the set of edges of GG itself, however, it reveals that if an edge uvuv is contained in the input graph, then it is located at machine Z(u,v)Z(u,v). For our lower bounds, we assume that ZZ is chosen uniformly at random from the set of all symmetric mappings from [V×V][V\times V] to [k][k], and that ZZ is known by all parties. For our upper bounds, we assume that ZZ is chosen adversarially and ZZ is not given to the parties. Instead, each machine ii then only knows the edges that are assigned to it at the beginning of the algorithm, i.e., the set of edges {uvE:Z(u,v)=i}\{uv\in E\ :\ Z(u,v)=i\}.

For a given problem P, we say that a communication protocol 𝒫\mathcal{P} has error ϵ\epsilon if, for any input graph, the probability that the joint output of the parties is a correct solution for P is at least 1ϵ1-\epsilon. This probability is taken over all private and public coins, and the random selection of the partition function ZZ.111We adapt this definition accordingly when considering adversarially chosen partition functions for the upper bounds in Section 4. We consider the global output of 𝒫\mathcal{P} as the union of the local outputs of the machines. For example, for MIS, every party is required to output an independent set so that the union of the parties’ outputs constitutes an MIS in the input graph. We say that a protocol is silent if none of the parties write on the blackboard.

The transcript of a protocol 𝒫\mathcal{P}, denoted by Π(𝒫)\Pi(\mathcal{P}), is the entirety of the bits written onto the blackboard. The communication cost of a protocol 𝒫\mathcal{P} is the maximum length of Π(𝒫)\Pi(\mathcal{P}) in an execution of 𝒫\mathcal{P}, which we denote by |Π(𝒫)||\Pi(\mathcal{P})|. Then, the randomized ϵ\epsilon-error blackboard communication complexity of a problem P, denoted by CCϵ(P)\textsc{CC}_{\epsilon}(P) is the minimum cost of a protocol 𝒫\mathcal{P} that solves P with error at most ϵ\epsilon.

2.2. Tools From Information Theory

Let AA be a random variable distributed according to a distribution 𝒟\mathcal{D}. We denote the (Shannon) entropy of AA by H𝒟(A)H_{\mathcal{D}}(A), where the index 𝒟\mathcal{D} may be dropped if the distribution is clear from the context. The mutual information of two random variables A,BA,B distributed according to 𝒟\mathcal{D} is denoted by I𝒟[A:B]=H𝒟(A)H𝒟(AB)\operatorname*{\textbf{{I}}}_{\mathcal{D}}[A\ :\ B]=H_{\mathcal{D}}(A)-H_{\mathcal{D}}(A\ \mid\ B) (again, 𝒟\mathcal{D} may be dropped), where H𝒟(AB)H_{\mathcal{D}}(A\ \mid\ B) is the entropy of AA conditioned on BB. We frequently use the shorthand {c}\{c\} for the event {C=c}\{C\!=\!c\}, for a random variable CC, and we write EC\operatorname*{\textbf{{E}}}_{C} to denote the expected value operator where the expectation is taken over the range of CC. The conditional mutual information of AA and BB conditioned on random variable CC is defined as

I𝒟[A:BC]=EC[I𝒟C=c[A:BC=c]]=cPr[c]I𝒟C=c[A:Bc].\operatorname*{\textbf{{I}}}_{\mathcal{D}}[A:B\mid C]=\operatorname*{\textbf{{E}}}_{C}\left[\operatorname*{\textbf{{I}}}_{\mathcal{D}\mid C=c}[A:B\mid C\!=\!c]\right]=\sum_{c}\operatorname*{\textbf{{Pr}}}[c]\operatorname*{\textbf{{I}}}_{\mathcal{D}\mid C=c}[A:B\mid c].

For a more detailed introduction to information theory, we refer the reader to (Cover and Thomas, 2006).

We will use the following properties of mutual information:

  1. (1)

    Chain rule for mutual information. I[A:B,C]=I[A:B]+I[A:CB]\operatorname*{\textbf{{I}}}[A\ :\ B,C]=\operatorname*{\textbf{{I}}}[A\ :\ B]+\operatorname*{\textbf{{I}}}[A\ :\ C\ \mid\ B].

  2. (2)

    Independent events. For random variables A,B,CA,B,C and event EE: I[A:BC,E]=I[A:BC]\operatorname*{\textbf{{I}}}[A\ :\ B\mid C,E]=\operatorname*{\textbf{{I}}}[A\ :\ B\mid C], if EE is independent of A,B,CA,B,C.

  3. (3)

    Conditioning on independent variable. (see e.g. Claim 2.3. in (Assadi et al., 2016a) for a proof) Let A,B,C,DA,B,C,D be jointly distributed random variables so that AA and DD are independent conditioned on CC. Then, I[A:BC,D]I[A:BC].\operatorname*{\textbf{{I}}}[A\ :\ B\mid C,D]\geqslant\operatorname*{\textbf{{I}}}[A\ :\ B\mid C]\ .

Our lower bound proof relies on Pinsker’s Inequality, which relates the total variation distance to the Kullback-Leibler divergence of two distributions:

Definition 0 (Total Variation Distance, Kullback-Leibler Divergence).

Let XX be a discrete random variable and consider two probability distributions 𝐩(X)\mathbf{p}(X) and 𝐪(X)\mathbf{q}(X). The total variation distance between 𝐩(X)\mathbf{p}(X) and 𝐪(X)\mathbf{q}(X) is defined as

δTV(𝐩(X),𝐪(X))=12x|𝐩(x)𝐪(x)|,\delta_{TV}\left(\mathbf{p}(X),\mathbf{q}(X)\right)=\frac{1}{2}\sum_{x}\left|\mathbf{p}(x)-\mathbf{q}(x)\right|,

and the Kullback–Leibler divergence from 𝐪(X)\mathbf{q}(X) to 𝐩(X)\mathbf{p}(X) (measured in bits) is defined as

D[𝐩(X)𝐪(X)]=x𝐩(x)log𝐩(x)𝐪(x).\displaystyle\operatorname*{\textbf{{D}}}\left[\mathbf{p}(X)\parallel\mathbf{q}(X)\right]=\sum_{x}\mathbf{p}(x)\log\frac{\mathbf{p}(x)}{\mathbf{q}(x)}.
Theorem 2 (Pinsker’s Inequality (Pinsker, 1964)).

Let XX be a random variable. For any two probability distributions 𝐩(X)\mathbf{p}(X) and 𝐪(X)\mathbf{q}(X), it holds that

δTV(𝐩(X),𝐪(X))log22D[𝐩(X)𝐪(X)].\delta_{TV}\left(\mathbf{p}(X),\mathbf{q}(X)\right)\leqslant\sqrt{\frac{\log 2}{2}\operatorname*{\textbf{{D}}}\left[\mathbf{p}(X)\parallel\mathbf{q}(X)\right]}.

3. Lower Bound

3.1. Input Distributions

In our input distributions, we assume that the vertex set VV of our input graph G=(V,E)G=(V,E) is fixed and known to all parties. Furthermore, w.l.o.g. we assume that V=[|V|]V=[|V|]222For an integer \ell we write [][\ell] to denote the set {1,2,,}\{1,2,\dots,\ell\}.. We will now define a distribution ρN\rho_{N} of partition functions that map the set of potential edges V×VV\times V to the parties [k][k], where the subscript NN denotes the cardinality of the set of vertices, and two distributions λ\lambda and μ\mu over the edge set of our input graph. Since we need to specify both the partition function and the input graph as the input to a protocol in the 𝖡𝖡(k)\mathsf{BB}(k) model, the relevant distributions are therefore the product distributions λ×ρN\lambda\times\rho_{N} and μ×ρN\mu\times\rho_{N}.

Partition Function

We denote by ρN\rho_{N} the uniform distribution over all symmetric functions mapping V×VV\times V to [k][k], where |V|=N|V|=N, and we denote by ZρNZ\sim\rho_{N} the partition function associated with our input. In this section, we assume that ZZ is known to all parties.

Input Graph

We consider two input distributions, which we denote the single-gadget and the multi-gadget distributions. For a given gadget size tO(1)t\in O(1), let λ\lambda be the probability distribution over tt-vertex graphs where we sample each edge uniformly and independently with probability 12\tfrac{1}{2}. We call the resulting graph a gadget and say that λ\lambda is a single-gadget distribution. We define the multi-gadget distribution μ\mu to be the probability distribution where we sample n/tn/t gadgets independently from λ\lambda. More precisely, we assume that gadget ii is induced by vertices {(i1)t+1,,it}\{(i-1)t+1,\dots,it\}, and the edges interconnecting gadget ii are distributed according to λ\lambda. For simplicity, we assume that n/tn/t is an integer. To differentiate between the two distributions, we always denote the number of vertices of the input graph when chosen from the single-gadget distribution λ\lambda by tt, and when chosen from the multi-gadget distribution μ\mu by nn.

3.2. Non-trivial Graph Problems

The lower bounds proved in this paper hold for non-trivial problems, which are defined by means of good partitions:

Definition 0 (Good Partition).

Consider a single-gadget partition function zz. We say that zz is good, if every party receives at most one potential edge under zz.

Observe that this is only possible if the number of parties exceeds the number of potential edges, i.e., if k(t2)k\geqslant{t\choose 2}. In our setting, tt is a small constant and kk is sufficiently large. Almost all partitions are therefore good.

Definition 0 (Non-trivial Problem).

We say that a graph problem is non-trivial if there exists a natural number tt such that for every good partition zz there is no deterministic silent 𝖡𝖡(k)\mathsf{BB}(k) model protocol that solves the problem on every instance (G,z)(G,z), where GλG\sim\lambda.

For the parties to contribute to the global output of the protocol by only seeing at most one edge and without communication is impossible for most problems. The class of non-trivial problems is therefore large. In Section 3.8, we give formal proofs, demonstrating that many key problems considered in distributed computing are non-trivial.

3.3. Information Cost

We now define the (external) information cost of a protocol in the 𝖡𝖡(k)\mathsf{BB}(k) model:

Definition 0 ((External) information cost).

Let 𝒫\mathcal{P} be an ϵ\epsilon-error protocol in the 𝖡𝖡(k)\mathsf{BB}(k) model for a problem P. We define

𝖨𝖢𝗈𝗌𝗍μ×ρn(𝒫)=Iμ×ρn[E(G):Π(𝒫)Z,R𝒫],\operatorname*{\mathsf{ICost}}_{\mu\times\rho_{n}}(\mathcal{P})=\operatorname*{\textbf{{I}}}_{\mu\times\rho_{n}}[E(G):\Pi(\mathcal{P})\mid Z,R_{\mathcal{P}}],

where R𝒫R_{\mathcal{P}} are the public random coins used by 𝒫\mathcal{P} and ZZ is the partition function.

Informally, 𝖨𝖢𝗈𝗌𝗍μ×ρn(𝒫)\operatorname*{\mathsf{ICost}}_{\mu\times\rho_{n}}(\mathcal{P}) is the amount of information revealed about the input edge set by the transcript, given that the public randomness and the partition function are known. We define 𝖨𝖢𝗈𝗌𝗍λ×ρt(𝒫)\operatorname*{\mathsf{ICost}}_{\lambda\times\rho_{t}}(\mathcal{P}) in a similar way.

It is well-known that the external information cost of a protocol is a lower bound on its communication cost. For completeness, we give a proof in the following lemma:

Lemma 0.

CCϵ(P)min𝒬𝖨𝖢𝗈𝗌𝗍μ×ρn(𝒬),\textsc{CC}_{\epsilon}(P)\geqslant\min_{\mathcal{Q}}\operatorname*{\mathsf{ICost}}_{\mu\times\rho_{n}}(\mathcal{Q}), where 𝒬\mathcal{Q} ranges over all ϵ\epsilon-error randomized protocols for problem PP.

Proof.
min𝒬𝖨𝖢𝗈𝗌𝗍μ×ρn(𝒬)\displaystyle\min_{\mathcal{Q}}\operatorname*{\mathsf{ICost}}_{\mu\times\rho_{n}}(\mathcal{Q}) =min𝒬Iμ×ρn[E(G):Π(𝒫)Z,R𝒬]min𝒬Hμ×ρn[Π(𝒬)Z,R𝒬]min𝒬Hμ×ρn[Π(𝒬)]min𝒬|Π(𝒬)|\displaystyle=\min_{\mathcal{Q}}\operatorname*{\textbf{{I}}}_{\mu\times\rho_{n}}[E(G):\Pi(\mathcal{P})\mid Z,R_{\mathcal{Q}}]\leqslant\min_{\mathcal{Q}}\operatorname*{\textbf{{H}}}_{\mu\times\rho_{n}}[\Pi(\mathcal{Q})\mid Z,R_{\mathcal{Q}}]\leqslant\min_{\mathcal{Q}}\operatorname*{\textbf{{H}}}_{\mu\times\rho_{n}}[\Pi(\mathcal{Q})]\leqslant\min_{\mathcal{Q}}|\Pi(\mathcal{Q})|
CCϵ(P),\displaystyle\leqslant\textsc{CC}_{\epsilon}({P}),

where the second last step follows from Theorem 6.1 in (Rao and Yehudayoff, 2020). ∎

3.4. Direct Sum Argument

In this section, we show that the information cost of a protocol for the multi-gadget case is lower-bounded by the information cost of a protocol for the single-gadget case, multiplied by the number of gadgets:

Theorem 5.

Let 𝒫\mathcal{P} be an ϵ\epsilon-error randomized 𝖡𝖡(k)\mathsf{BB}(k) protocol. Then, there exists an ϵ\epsilon-error 𝖡𝖡(k)\mathsf{BB}(k) protocol 𝒬\mathcal{Q} such that:

(1) 𝖨𝖢𝗈𝗌𝗍λ×ρt(𝒬)tn𝖨𝖢𝗈𝗌𝗍μ×ρn(𝒫).\displaystyle\operatorname*{\mathsf{ICost}}_{\lambda\times\rho_{t}}(\mathcal{Q})\leqslant\frac{t}{n}\operatorname*{\mathsf{ICost}}_{\mu\times\rho_{n}}(\mathcal{P})\ .
Proof.

Given 𝒫\mathcal{P}, we construct protocol 𝒬\mathcal{Q} as follows. Let (H,Z)λ×ρt(H,Z)\sim\lambda\times\rho_{t} denote the input to 𝒬\mathcal{Q}. The parties construct input (G,Z)(G,Z^{\prime}) for 𝒫\mathcal{P} from (H,Z)(H,Z), as follows:

  1. (1)

    The parties use public randomness R𝒬R_{\mathcal{Q}} to sample a uniform random index I[n/t]I\in[n/t].

  2. (2)

    The parties use public randomness to sample a partition ZZ^{\prime} from ρn\rho_{n} conditioned on the partition of gadget II being equivalent to ZZ.

  3. (3)

    The parties set the single input gadget HH to be gadget II in the multi-gadget input GG.

  4. (4)

    Given ZZ^{\prime}, the parties know which potential edges of the gadgets different to gadget II are hosted locally. They then use private randomness to sample the existence of the individual edges.

  5. (5)

    The parties run protocol 𝒫\mathcal{P} on input (G,Z)(G,Z^{\prime}) and return the part of the output that concerns gadget II.

Observe that the random variables R𝒬,ZR_{\mathcal{Q}},Z are identical to I,R𝒫,ZI,R_{\mathcal{P}},Z^{\prime}, where R𝒫R_{\mathcal{P}} denotes the public coins used by protocol 𝒫\mathcal{P}. Observe further that the constructed input (G,Z)(G,Z^{\prime}) is distributed according to μ×ρn\mu\times\rho_{n}.

Denote by EIE_{I} the edges of gadget II of the multi-gadget input. We obtain:

𝖨𝖢𝗈𝗌𝗍λ×ρt(𝒬)\displaystyle\operatorname*{\mathsf{ICost}}_{\lambda\times\rho_{t}}(\mathcal{Q}) =Iλ×ρt[E(H):Π(𝒬)R𝒬,Z]\displaystyle=\operatorname*{\textbf{{I}}}_{\lambda\times\rho_{t}}[E(H):\Pi(\mathcal{Q})\mid R_{\mathcal{Q}},Z]
=Iμ×ρn[EI:Π(𝒫)R𝒫,I,Z]\displaystyle=\operatorname*{\textbf{{I}}}_{\mu\times\rho_{n}}[E_{I}:\Pi(\mathcal{P})\mid R_{\mathcal{P}},I,Z^{\prime}]
=EiI[Iμ×ρn[Ei:Π(𝒫)R𝒫,Z,I=i]]\displaystyle=\operatorname*{\textbf{{E}}}_{i\sim I}\left[\operatorname*{\textbf{{I}}}_{\mu\times\rho_{n}}[E_{i}:\Pi(\mathcal{P})\mid R_{\mathcal{P}},Z^{\prime},I=i]\right]
(2) tni[n/t]Iμ×ρn[Ei:Π(𝒫)R𝒫,E1,,Ei1,Z,I=i],\displaystyle\leqslant\frac{t}{n}\sum_{i\in[n/t]}\operatorname*{\textbf{{I}}}_{\mu\times\rho_{n}}[E_{i}:\Pi(\mathcal{P})\mid R_{\mathcal{P}},E_{1},\dots,E_{i-1},Z^{\prime},I=i],
where (2) follows from Property (3) in Section 2.2 since EiE_{i} and E1,,EiE_{1},\dots,E_{i} are independent. Moreover, we observe that EiE_{i} and Π(𝒫)\Pi(\mathcal{P}) are independent of the event {I=i}\{I\!=\!i\}, which yields
𝖨𝖢𝗈𝗌𝗍λ×ρt(𝒬)\displaystyle\operatorname*{\mathsf{ICost}}_{\lambda\times\rho_{t}}(\mathcal{Q}) tni[n/t]Iμ×ρn[Ei:Π(𝒫)R𝒫,E1,,Ei1,Z]\displaystyle\leqslant\frac{t}{n}\sum_{i\in[n/t]}\operatorname*{\textbf{{I}}}_{\mu\times\rho_{n}}[E_{i}:\Pi(\mathcal{P})\mid R_{\mathcal{P}},E_{1},\dots,E_{i-1},Z^{\prime}]
(3) tnIμ×ρn[E(G):Π(𝒫)R𝒫,Z]\displaystyle\leqslant\frac{t}{n}\operatorname*{\textbf{{I}}}_{\mu\times\rho_{n}}[E(G):\Pi(\mathcal{P})\mid R_{\mathcal{P}},Z^{\prime}]
=tn𝖨𝖢𝗈𝗌𝗍μ×ρn(𝒫).\displaystyle=\frac{t}{n}\operatorname*{\mathsf{ICost}}_{\mu\times\rho_{n}}(\mathcal{P}).

where we used the chain rule of mutual information (Property (1) in Section 2.2) in (3). ∎

3.5. Zero Communication Protocol

Next, we show that if a protocol 𝒬\mathcal{Q} solves single-gadget instances with small information cost, then there exists a silent protocol (i.e., a protocol that avoids all communication) with only slightly increased error.

Theorem 6.

For a problem P, let 𝒬\mathcal{Q} be an ϵ\epsilon-error randomized protocol in the 𝖡𝖡(k)\mathsf{BB}(k) model with

(4) 𝖨𝖢𝗈𝗌𝗍λ×ρt(𝒬)(2ϵ3)4.\displaystyle\operatorname*{\mathsf{ICost}}_{\lambda\times\rho_{t}}(\mathcal{Q})\leqslant\left(\tfrac{2\epsilon}{3}\right)^{4}\ .

Then, there exists a silent protocol in the 𝖡𝖡(k)\mathsf{BB}(k) model with error at most 4ϵ4\epsilon for problem P.

Proof.

Consider the protocol 𝒬\mathcal{Q} as stated in the premise of the theorem. As a first step, we leverage the assumption that 𝖨𝖢𝗈𝗌𝗍λ×ρt(𝒬)\operatorname*{\mathsf{ICost}}_{\lambda\times\rho_{t}}(\mathcal{Q}) is small to show that the messages sent by the algorithm are close to being independent of the edges of the sampled gadget HH.

Lemma 0.

Suppose we sample the edges E(H)λE(H)\sim\lambda of gadget HH and a partition function ZρtZ\sim\rho_{t}. Let RR be the public randomness used by 𝒬\mathcal{Q}. Let 𝐩(Π(𝒬)E(H),R,Z)\mathbf{p}(\Pi(\mathcal{Q})\mid E(H),R,Z) denote the probability distribution of Π(𝒬)\Pi(\mathcal{Q}) conditioned on E(H)E(H), RR, and ZZ, and define 𝐩(Π(𝒬)R,Z)\mathbf{p}(\Pi(\mathcal{Q})\mid R,Z) similarly. Then, with probability at least 12ϵ31-\tfrac{2\epsilon}{3}, it holds that

(5) δTV(𝐩(Π(𝒬)E(H),R,Z),𝐩(Π(𝒬)R,Z))2ϵ3.\displaystyle\delta_{TV}\Big{(}\mathbf{p}(\Pi(\mathcal{Q})\mid E(H),R,Z),\mathbf{p}(\Pi(\mathcal{Q})\mid R,Z)\Big{)}\leqslant\tfrac{2\epsilon}{3}\ .
Proof.

We will derive an upper bound on the expected total variation distance, where the expectation ranges over the input gadget HH, the public random string RR, and the partition function ZZ. In the following, we use the abbreviation E=E(H)E=E(H). By applying Pinsker’s inequality (see Theorem 2) and taking expectations on both sides, we get

EE,R,Z[δTV(𝐩(Π(𝒬)E,R,Z),𝐩(Π(𝒬)R,Z))]\displaystyle\operatorname*{\textbf{{E}}}_{E,R,Z}\left[\delta_{TV}\left(\mathbf{p}(\Pi(\mathcal{Q})\mid E,R,Z),\mathbf{p}(\Pi(\mathcal{Q})\mid R,Z)\right)\right] EE,R,Z[log22D[𝐩(Π(𝒬)E,R,Z)𝐩(Π(𝒬)R,Z)]]\displaystyle\leqslant\operatorname*{\textbf{{E}}}_{E,R,Z}\left[\sqrt{\frac{\log 2}{2}\operatorname*{\textbf{{D}}}\left[\mathbf{p}(\Pi(\mathcal{Q})\mid E,R,Z)\parallel\mathbf{p}(\Pi(\mathcal{Q})\mid R,Z)\right]}\right]
From Jensen’s inequality for concave functions and the fact that log221\frac{\log 2}{2}\leqslant 1, we get
EE,R,Z[δTV(𝐩(Π(𝒬)E,R,Z),𝐩(Π(𝒬)R,Z))]\displaystyle\operatorname*{\textbf{{E}}}_{E,R,Z}\left[\delta_{TV}\left(\mathbf{p}(\Pi(\mathcal{Q})\mid E,R,Z),\mathbf{p}(\Pi(\mathcal{Q})\mid R,Z)\right)\right] EE,R,Z[D[𝐩(Π(𝒬)E,R,Z)𝐩(Π(𝒬)R,Z)]].\displaystyle\leqslant\sqrt{\operatorname*{\textbf{{E}}}_{E,R,Z}\left[\operatorname*{\textbf{{D}}}\left[\mathbf{p}(\Pi(\mathcal{Q})\mid E,R,Z)\parallel\mathbf{p}(\Pi(\mathcal{Q})\mid R,Z)\right]\right].}

In the following derivation, π\pi ranges over all possible transcripts of 𝒬\mathcal{Q}. By using the definition of the Kullback-Leibler divergence (see Def. 1), we obtain that

EE,R,Z[δTV(𝐩(Π(𝒬)E,R,Z),𝐩(Π(𝒬)R,Z))]\displaystyle\operatorname*{\textbf{{E}}}_{E,R,Z}\left[\delta_{TV}\left(\mathbf{p}(\Pi(\mathcal{Q})\mid E,R,Z),\mathbf{p}(\Pi(\mathcal{Q})\mid R,Z)\right)\right] e,r,zPr[e,r,z]πPr[πe,r,z]log(Pr[πe,r,z]Pr[πr,z])\displaystyle\leqslant\sqrt{\sum_{e,r,z}\operatorname*{\textbf{{Pr}}}\left[e,r,z\right]\sum_{\pi}\operatorname*{\textbf{{Pr}}}\left[\pi\mid e,r,z\right]\log\left(\frac{\operatorname*{\textbf{{Pr}}}\left[\pi\mid e,r,z\right]}{\operatorname*{\textbf{{Pr}}}\left[\pi\mid r,z\right]}\right)}
=e,π,r,zPr[e,π,r,z]log(Pr[πe,r,z]Pr[er,z]Pr[r,z]Pr[er,z]Pr[πr,z]Pr[r,z])\displaystyle=\sqrt{\sum_{e,\pi,r,z}\operatorname*{\textbf{{Pr}}}\left[e,\pi,r,z\right]\log\left(\frac{\operatorname*{\textbf{{Pr}}}\left[\pi\mid e,r,z\right]\operatorname*{\textbf{{Pr}}}\left[e\mid r,z\right]\operatorname*{\textbf{{Pr}}}\left[r,z\right]}{\operatorname*{\textbf{{Pr}}}\left[e\mid r,z\right]\operatorname*{\textbf{{Pr}}}\left[\pi\mid r,z\right]\operatorname*{\textbf{{Pr}}}\left[r,z\right]}\right)}
=e,π,r,zPr[e,π,r,z]log(Pr[e,π,r,z]Pr[er,z]Pr[πr,z]Pr[r,z])\displaystyle=\sqrt{\sum_{e,\pi,r,z}\operatorname*{\textbf{{Pr}}}\left[e,\pi,r,z\right]\log\left(\frac{\operatorname*{\textbf{{Pr}}}\left[e,\pi,r,z\right]}{\operatorname*{\textbf{{Pr}}}\left[e\mid r,z\right]\operatorname*{\textbf{{Pr}}}\left[\pi\mid r,z\right]\operatorname*{\textbf{{Pr}}}\left[r,z\right]}\right)}
(6) =I[E(H):Π(𝒬)R,Z]\displaystyle=\sqrt{\operatorname*{\textbf{{I}}}[E(H):\Pi(\mathcal{Q})\mid R,Z]}
=𝖨𝖢𝗈𝗌𝗍λ×ρt(𝒬)\displaystyle=\sqrt{\operatorname*{\mathsf{ICost}}_{\lambda\times\rho_{t}}(\mathcal{Q})}
(2ϵ3)2,\displaystyle\leqslant\left(\tfrac{2\epsilon}{3}\right)^{2},

where we used the definition of conditional mutual information in (6), and the upper bound (4) in the last step. To translate the upper bound on the expected value to an upper bound on the probability that the total variation distance is large, we apply Markov’s inequality. That is, if we sample EE, RR, and ZZ from the respective distributions stated in the premise of the lemma, then it holds that

Pr[δTV(𝐩(Π(𝒬)E,R,Z),𝐩(Π(𝒬)R,Z))2ϵ3]\displaystyle\operatorname*{\textbf{{Pr}}}\left[\delta_{TV}\left(\mathbf{p}(\Pi(\mathcal{Q})\mid E,R,Z),\mathbf{p}(\Pi(\mathcal{Q})\mid R,Z)\right)\geqslant\tfrac{2\epsilon}{3}\right] 32ϵEE,R,Z[δTV(𝐩(Π(𝒬)E,R,Z),𝐩(Π(𝒬)R,Z))]\displaystyle\leqslant\frac{3}{2\epsilon}\operatorname*{\textbf{{E}}}_{E,R,Z}\left[\delta_{TV}\left(\mathbf{p}(\Pi(\mathcal{Q})\mid E,R,Z),\mathbf{p}(\Pi(\mathcal{Q})\mid R,Z)\right)\right]
2ϵ3,\displaystyle\leqslant\frac{2\epsilon}{3},

and the lemma follows. ∎

Equipped with Lemma 7, we are now ready to prove Theorem 6. The players use the even bits of their public randomness to sample RR, which is the public randomness of protocol 𝒬\mathcal{Q}, yielding a new protocol 𝒬R\mathcal{Q}_{R} that does not use public randomness but otherwise behaves the same way as 𝒬\mathcal{Q} given RR. Then, the players use the odd bits of their public randomness to sample a transcript Π\Pi from the distribution 𝐩(Π(𝒬R)Z)=𝐩(Π(𝒬)R,Z)\mathbf{p}(\Pi(\mathcal{Q}_{R})\mid Z)=\mathbf{p}(\Pi(\mathcal{Q})\mid R,Z). After these two steps, all players know RR and Π\Pi. Finally, each player PiP_{i} computes its output by applying the state transition function of 𝒬R\mathcal{Q}_{R} to ZZ, Π\Pi, its input assignment, and its private random bits. Thus, we have obtained a silent protocol 𝒮\mathcal{S}.

To prove the lemma, we need to obtain an upper-bound on the error probability of 𝒮\mathcal{S}. Clearly, 𝒮\mathcal{S} fails in all instances where 𝒬\mathcal{Q} fails. In fact, the only difference between computing the output of 𝒮\mathcal{S} compared to 𝒬\mathcal{Q} is that the transcript used in QQ is sampled from distribution 𝐩(Π(𝒬)E(H),R,Z)\mathbf{p}(\Pi(\mathcal{Q})\mid E(H),R,Z), whereas the one we use in 𝒮\mathcal{S} is sampled from 𝐩(Π(𝒬)R,Z)\mathbf{p}(\Pi(\mathcal{Q})\mid R,Z). Thus, the only additional error probability of protocol 𝒮\mathcal{S} is determined by the difference in the probability mass assigned to each transcript π\pi between the distributions 𝐩(Π(𝒬)E(H),R,Z)\mathbf{p}(\Pi(\mathcal{Q})\mid E(H),R,Z) and 𝐩(Π(𝒬)R,Z)\mathbf{p}(\Pi(\mathcal{Q})\mid R,Z). We obtain

Pr[𝒮 errs]\displaystyle\operatorname*{\textbf{{Pr}}}\left[\text{$\mathcal{S}$ errs}\right] =Pr[𝒮 errs𝒬 errs ]Pr[𝒬 errs]+Pr[𝒮 errs𝒬 correct]Pr[𝒬 correct]\displaystyle=\operatorname*{\textbf{{Pr}}}\left[\text{$\mathcal{S}$ errs}\mid\text{$\mathcal{Q}$ errs }\right]\operatorname*{\textbf{{Pr}}}\left[\text{$\mathcal{Q}$ errs}\right]+\operatorname*{\textbf{{Pr}}}\left[\text{$\mathcal{S}$ errs}\mid\text{$\mathcal{Q}$ correct}\right]\operatorname*{\textbf{{Pr}}}\left[\text{$\mathcal{Q}$ correct}\right]
(7) ϵ+Pr[𝒮 errs𝒬 correct].\displaystyle\leqslant\epsilon+\operatorname*{\textbf{{Pr}}}\left[\text{$\mathcal{S}$ errs}\mid\text{$\mathcal{Q}$ correct}\right].

Let {δTV2ϵ3}\{\delta_{TV}\leqslant\tfrac{2\epsilon}{3}\} be the event that δTV(𝐩(Π(𝒬)E(H),R,Z),𝐩(Π(𝒬)R,Z))2ϵ3\delta_{TV}\Big{(}\mathbf{p}(\Pi(\mathcal{Q})\mid E(H),R,Z),\mathbf{p}(\Pi(\mathcal{Q})\mid R,Z)\Big{)}\leqslant\tfrac{2\epsilon}{3}, and define event {δTV>2ϵ3}\{\delta_{TV}>\tfrac{2\epsilon}{3}\} similarly. To upper-bound the probability Pr[𝒮 errs𝒬 correct]\operatorname*{\textbf{{Pr}}}\left[\text{$\mathcal{S}$ errs}\mid\text{$\mathcal{Q}$ correct}\right], we condition on whether event {δTV2ϵ3}\{\delta_{TV}\leqslant\tfrac{2\epsilon}{3}\} holds. Continuing from (7), we get

Pr[𝒮 errs]\displaystyle\operatorname*{\textbf{{Pr}}}\left[\text{$\mathcal{S}$ errs}\right] ϵ+Pr[𝒮 errsδTV2ϵ3,𝒬 correct]Pr[δTV2ϵ3|𝒬 correct]\displaystyle\leqslant\epsilon+\operatorname*{\textbf{{Pr}}}\left[\text{$\mathcal{S}$ errs}\mid\delta_{TV}\leqslant\tfrac{2\epsilon}{3},\text{$\mathcal{Q}$ correct}\right]\operatorname*{\textbf{{Pr}}}\left[\delta_{TV}\leqslant\tfrac{2\epsilon}{3}\ \middle|\ \text{$\mathcal{Q}$ correct}\right]
+Pr[𝒮 errsδTV>2ϵ3,𝒬 correct]Pr[δTV>2ϵ3|𝒬 correct]\displaystyle\phantom{-----}+\operatorname*{\textbf{{Pr}}}\left[\text{$\mathcal{S}$ errs}\mid\delta_{TV}>\tfrac{2\epsilon}{3},\text{$\mathcal{Q}$ correct}\right]\operatorname*{\textbf{{Pr}}}\left[\delta_{TV}>\tfrac{2\epsilon}{3}\ \middle|\ \text{$\mathcal{Q}$ correct}\right]
(8) ϵ+Pr[𝒮 errsδTV2ϵ3,𝒬 correct]+Pr[δTV>2ϵ3|𝒬 correct].\displaystyle\leqslant\epsilon+\operatorname*{\textbf{{Pr}}}\left[\text{$\mathcal{S}$ errs}\mid\delta_{TV}\leqslant\tfrac{2\epsilon}{3},\text{$\mathcal{Q}$ correct}\right]+\operatorname*{\textbf{{Pr}}}\left[\delta_{TV}>\tfrac{2\epsilon}{3}\ \middle|\ \text{$\mathcal{Q}$ correct}\right].

Conditioned on 𝒬\mathcal{Q} being correct, we know that the additional error of 𝒮\mathcal{S} is at most

π|Pr[Π(𝒬)=πE(H),R,Z]Pr[Π(𝒬)=πR,Z]|2δTV(𝐩(Π(𝒬)E(H),R,Z),𝐩(Π(𝒬)R,Z)),\sum_{\pi}\big{|}\operatorname*{\textbf{{Pr}}}\left[\Pi(\mathcal{Q})\!=\!\pi\mid E(H),R,Z\right]-\operatorname*{\textbf{{Pr}}}\left[\Pi(\mathcal{Q})\!=\!\pi\mid R,Z\big{]}\right|\leqslant 2\delta_{TV}\left(\mathbf{p}(\Pi(\mathcal{Q})\mid E(H),R,Z),\mathbf{p}(\Pi(\mathcal{Q})\mid R,Z)\right),

which, conditioned on event {δTV2ϵ3}\{\delta_{TV}\leqslant\tfrac{2\epsilon}{3}\}, yields the bound Pr[𝒮 errsδTV2ϵ3,𝒬 correct]4ϵ3\operatorname*{\textbf{{Pr}}}\left[\text{$\mathcal{S}$ errs}\mid\delta_{TV}\leqslant\tfrac{2\epsilon}{3},\text{$\mathcal{Q}$ correct}\right]\leqslant\tfrac{4\epsilon}{3}.

To complete the proof, we will derive an upper bound on Pr[δTV>2ϵ3|𝒬 correct]\operatorname*{\textbf{{Pr}}}\left[\delta_{TV}>\tfrac{2\epsilon}{3}\ \middle|\ \text{$\mathcal{Q}$ correct}\right]. Observe that

Pr[δTV2ϵ3]\displaystyle\operatorname*{\textbf{{Pr}}}\left[\delta_{TV}\leqslant\tfrac{2\epsilon}{3}\right] Pr[δTV2ϵ3𝒬 correct]+Pr[𝒬 errs]\displaystyle\leqslant\operatorname*{\textbf{{Pr}}}\left[\delta_{TV}\leqslant\tfrac{2\epsilon}{3}\mid\text{$\mathcal{Q}$ correct}\right]+\operatorname*{\textbf{{Pr}}}\left[\text{$\mathcal{Q}$ errs}\right]
Pr[δTV2ϵ3𝒬 correct]+ϵ.\displaystyle\leqslant\operatorname*{\textbf{{Pr}}}\left[\delta_{TV}\leqslant\tfrac{2\epsilon}{3}\mid\text{$\mathcal{Q}$ correct}\right]+\epsilon.

By applying Lemma 7 to the left-hand side, we get

Pr[δTV>2ϵ3𝒬 correct]\displaystyle\operatorname*{\textbf{{Pr}}}\left[\delta_{TV}>\tfrac{2\epsilon}{3}\mid\text{$\mathcal{Q}$ correct}\right] 2ϵ3+ϵ.\displaystyle\leqslant\tfrac{2\epsilon}{3}+\epsilon.

Plugging these two bounds into the right-hand side of (8), we get Pr[𝒮 errs]2ϵ+6ϵ3=4ϵ.\operatorname*{\textbf{{Pr}}}\left[\text{$\mathcal{S}$ errs}\right]\leqslant 2\epsilon+\frac{6\epsilon}{3}=4\epsilon. This completes the proof of Theorem 6. ∎

3.6. From Randomized to Deterministic Protocols

In this section, we show that the existence of a randomized silent protocol for single-gadget instances with a sufficiently small error implies existence of a deterministic silent protocol for single-gadget instances that solves the problem on every input graph under some good partition.

First, we will argue that the probability that the partition function is good is at least 1/21/2:

Lemma 0.

Let t1t\geqslant 1 and k3t4k\geqslant 3t^{4}. Consider a single gadget input on tt vertices. Then, the probability that the partition function is good is at least 1/21/2.

Proof.

There are (t2)t2{t\choose 2}\leqslant t^{2} vertex pairs. The probability that they are all assigned to different parties is at least:

k(k1)(k2)(kt2+1)kt2(kt2+1k)t2=(1t21k)t2exp(2t2(t21)k)12,\frac{k\cdot(k-1)\cdot(k-2)\cdot\dots\cdot(k-t^{2}+1)}{k^{t^{2}}}\geqslant\left(\frac{k-t^{2}+1}{k}\right)^{t^{2}}=\left(1-\frac{t^{2}-1}{k}\right)^{t^{2}}\geqslant\exp\left(\frac{-2t^{2}(t^{2}-1)}{k}\right)\geqslant\frac{1}{2}\ ,

where in the penultimate inequality we used the fact that 1xexp(2x)1-x\geqslant\exp(-2x) for any x[0,1/2]x\in[0,1/2]. ∎

We now state and prove the main result of the section.

Lemma 0.

Let t1t\geqslant 1, k3t4k\geqslant 3t^{4}, and ϵ13(t2)\epsilon\leqslant\frac{1}{3{t\choose 2}}. Let 𝒫\mathcal{P} be a randomized silent protocol with error at most ϵ\epsilon (where the error is over the random partition ρt\rho_{t} and the random coin flips). Then, there exists a good partition zz and a deterministic silent protocol that succeeds on every input (H,z)(H,z), where HH is a tt-vertex graph.

Proof.

First, the assumption of the lemma together with Yao’s minimax principle (Yao, 1977) imply that there exists a deterministic silent protocol 𝒬\mathcal{Q} with distributional error at most ϵ\epsilon over the input distribution λ×ρt\lambda\times\rho_{t}.

Now, for a fixed partition function zz, observe that if 𝒬\mathcal{Q} is not correct on all tt-vertex input graphs, then its error conditioned on zz is at least 1(t2)\frac{1}{{t\choose 2}}. For the sake of a contradiction, suppose that there is no good partition such that 𝒬\mathcal{Q} succeeds on all inputs. Recall further that at least half of all partition functions are good ( Lemma 8). Then:

ϵ=Pr[𝒬 errs]\displaystyle\epsilon=\operatorname*{\textbf{{Pr}}}[\mathcal{Q}\text{ errs}] Pr[z good]Pr[𝒬 errsz good]121(t2),\displaystyle\geqslant\operatorname*{\textbf{{Pr}}}[z\text{ good}]\cdot\operatorname*{\textbf{{Pr}}}[\mathcal{Q}\text{ errs}\mid z\text{ good}]\geqslant\frac{1}{2}\frac{1}{{t\choose 2}}\ ,

contradicting the assumption that ϵ131(t2)\epsilon\leqslant\frac{1}{3}\frac{1}{{t\choose 2}}. ∎

3.7. Main Lower Bound Theorem

Theorem 1. P be a non-trivial graph problem. Then there exists δ(P)>0\delta(P)>0 and t=t(P)t=t(P) such that for any k3t4k\geqslant 3t^{4} and ϵδ(P)\epsilon\leqslant\delta(P) the randomized ϵ\epsilon-error blackboard communication complexity CCϵ(P)\textsc{CC}_{\epsilon}(P) of PP is Ω(n)\Omega(n), even if the edges of the input graph are partitioned randomly among the parties.

Proof.

Let t=t(P)t=t(P) be the minimum integer such that for every good partition zz from ρt\rho_{t} there is no deterministic silent 𝖡𝖡(k)\mathsf{BB}(k) model protocol that solves PP on every instance (H,z)(H,z), where HλH\sim\lambda (such a tt exists since P is non-trivial). Let δ(P)=112(t2)\delta(P)=\frac{1}{12{t\choose 2}} and ϵδ(P)\epsilon\leqslant\delta(P).

Suppose, towards a contradiction, that CCϵ(P)1t(2ϵ3)4n\textsc{CC}_{\epsilon}(P)\leqslant\frac{1}{t}\left(\frac{2\epsilon}{3}\right)^{4}n. Then, by Lemma 4, there exists a protocol 𝒫\mathcal{P} for PP with 𝖨𝖢𝗈𝗌𝗍μ×ρn(𝒫)1t(2ϵ3)4n\operatorname*{\mathsf{ICost}}_{\mu\times\rho_{n}}(\mathcal{P})\leqslant\frac{1}{t}\left(\frac{2\epsilon}{3}\right)^{4}n. By Theorem 5, this in turn implies the existence of a protocol 𝒬\mathcal{Q} for input instances from λ×ρt\lambda\times\rho_{t} such that

𝖨𝖢𝗈𝗌𝗍λ×ρt(𝒬)tn𝖨𝖢𝗈𝗌𝗍μ×ρn(𝒫)(2ϵ3)4.\operatorname*{\mathsf{ICost}}_{\lambda\times\rho_{t}}(\mathcal{Q})\leqslant\frac{t}{n}\operatorname*{\mathsf{ICost}}_{\mu\times\rho_{n}}(\mathcal{P})\leqslant\left(\frac{2\epsilon}{3}\right)^{4}.

Therefore, by Theorem 6, there exists a silent protocol \mathcal{R} for input instances λ×ρt\lambda\times\rho_{t} with error at most 4ϵ13(t2)4\epsilon\leqslant\frac{1}{3{t\choose 2}}. Consequently, by Lemma 9, there exists a good partition zz from ρt\rho_{t} and a deterministic silent protocol that succeeds on every input (G,z)(G,z), where GG is a tt-vertex graph. But this contradicts the choice of tt. ∎

3.8. Lower Bounds for Specific Problems

In this section we establish non-triviality of several fundamental graph problems. We point out that lower bounds are known for all of these problems in other distributed computing models such as the LOCAL model (see (Balliu et al., 2019; Kuhn et al., 2016)). However, these lower bounds do not directly translate to our setting since, in the 𝖡𝖡(k)\mathsf{BB}(k) model, a player who receives an edge as input does not necessarily need to produce an output for this particular edge or its vertices.

Lemma 0.

Maximal Matching is a non-trivial problem.

Proof.

For contradiction, suppose that for every natural tt there exist a good partition zz from ρt\rho_{t} and a protocol 𝒬\mathcal{Q} that solves Maximal Matching on every instance chosen from λ\lambda under partition zz. Let t3t\geqslant 3. First, observe that if the input graph is the empty graph, then no party can output an edge. This implies that if a party does not receive an edge then its output needs to be empty. Suppose next that the input graph consists of a single edge. Then, the party that hosts this edge needs to output this edge since otherwise the computed matching would not be maximal. This implies further that any party that holds an edge needs to output their edge. However, if two machines hold edges that share an endpoint then the output would not be a matching, a contradiction. ∎

Lemma 0.

Maximal Independent Set, Minimum Dominating Set, (Δ+1)(\Delta+1)-coloring, and Minimum Vertex Cover are non-trivial problems.

Proof.

Maximal Independent Set and Minimum Dominating Set. For contradiction, suppose that for every natural tt there exist a good partition zz from ρt\rho_{t} and a protocol 𝒬\mathcal{Q} that solves Maximal Independent Set (resp. Minimum Dominating Set) on every instance chosen from λ\lambda under partition zz. Let t4t\geqslant 4. First, observe that if the input graph is the empty graph, all vertices must be output. This implies that there is a function f:[t][k]f:[t]\rightarrow[k] such that for every i[t]i\in[t], if party f(i)f(i) hosts no edges, then its output contains vertex ii. Let a,b,c,da,b,c,d be distinct vertices of the input graph. Since there are six different pairs of distinct vertices from {a,b,c,d}\{a,b,c,d\}, at least one of these pairs, say (a,b)(a,b), is assigned by the partition zz to a party q{f(a),f(b),f(c),f(d)}q\not\in\{f(a),f(b),f(c),f(d)\}. But then on the input graph with a unique edge (a,b)(a,b), the output will contain both vertices aa and bb, which is not possible, as the output of 𝒬\mathcal{Q} should be an independent set (resp. a minimum dominating set).

(Δ+1)(\Delta+1)-coloring. The proof works similarly to the previous one. Indeed, for the empty input graph, the output should be a vertex coloring, where every vertex is assigned the same color, say color 1. This implies that there is a function f:[t][k]f:[t]\rightarrow[k] such that for every i[t]i\in[t], if party f(i)f(i) hosts no edges, then its output assigns color 1 to vertex ii. Then, following the same arguments as in the previous proof, we conclude that on an input graph with at least four vertices a,b,c,da,b,c,d and a unique edge (a,b)(a,b), the output coloring assigns color 1 to both vertices aa and bb, which is not a proper coloring, a contradiction.

Minimum Vertex Cover. Since for any graph G=(V,E)G=(V,E) a vertex set SVS\subseteq V is a minimum vertex cover if and only if the set VSV\setminus S is a maximum independent set, the existence of a deterministic silent protocol for Minimum Vertex Cover would imply the existence of one for Maximum Independent Set. Furthermore, as any maximum independent set is also a maximal independent set, the existence of a desired protocol for Maximum Independent Set would imply the existence of one for Maximal Independent Set, a contradiction. ∎

In the proof of non-triviality of Maximal Matching we used gadgets of size 3 and in the proofs of non-triviality of all other problems we used gadgets of size 4. Together with Theorem 1 this implies the following

Corollary 0.
  1. (1)

    For any ϵ136\epsilon\leqslant\frac{1}{36} and k243k\geqslant 243, the randomized ϵ\epsilon-error blackboard communication complexity of Maximal Matching is Ω(n)\Omega(n).

  2. (2)

    For any ϵ172\epsilon\leqslant\frac{1}{72} and k768k\geqslant 768, the randomized ϵ\epsilon-error blackboard communication complexity of each of the problems Maximal Independent Set, Minimum Dominating Set, (Δ+1)(\Delta+1)-coloring, and Minimum Vertex Cover is Ω(n)\Omega(n).

4. Upper Bounds

In this section, we assume that the partition function zz is arbitrary (potentially adversarially chosen), and each party ii only knows the edges allocated to them, i.e., the set of edges Ei={uvE:z(u,v)=i}E_{i}=\{uv\in E\ :\ z(u,v)=i\}. Thus, in contrast to our lower bounds, the error probability of a protocol is now only computed over the private and public coin flips.

4.1. Maximal Matching and Maximum Matching Approximation

Computing a maximal matching in the 𝖡𝖡\mathsf{BB}(k) model with communication cost O(nlogn)O(n\log n) is straightforward. The parties can simulate the Greedy matching algorithm, as follows: The blackboard serves as a variable MM that contains the current matching. Initially, the blackboard is empty and represents the empty set, i.e., MM\leftarrow\varnothing. The parties proceed in order. At party ii’s turn, party ii attempts to add each of its edges eEie\in E_{i} in any order to the blackboard if possible, i.e., if M{e}M\cup\{e\} is still a matching. Since any matching in an nn-vertex graph is of size at most n/2n/2, and each edge requires space O(logn)O(\log n) to be written onto the blackboard (e.g., by indicating the two endpoints of the edge), the communication cost is O(nlogn)O(n\log n).

Various algorithms are known for computing a (1ϵ)(1-\epsilon)-approximation to Maximum Matching by repeatedly computing maximal matchings in subgraphs of the input graph (e.g. (McGregor, 2005; Ahn and Guha, 2011; Eggert et al., 2012; Konrad, 2018; binti Khalil and Konrad, 2020; Sepehr Assadi, 2021)). In order to implement these algorithms in the 𝖡𝖡\mathsf{BB}(k) model, we need to make sure that the kk parties know which of their edges are contained in each of the respective subgraphs. If this can be achieved with communication cost O(s)O(s) per subgraph, then a 𝖡𝖡\mathsf{BB}(k) model implementation with cost O((nlogn+s)r)O((n\log n+s)\cdot r) can be obtained, where rr is the number of matchings computed.

The algorithm by McGregor (McGregor, 2005) follows this scheme and relies on the computation of 1ϵO(1ϵ)\frac{1}{\epsilon}^{O(\frac{1}{\epsilon})} maximal matchings in order to establish a (1ϵ)(1-\epsilon)-approximate matching in general graphs. Specifying the respective subgraphs with small communication cost is straightforward for this algorithm, requiring a cost of O(nlogn)O(n\log n) (details omitted), and we thus obtain a (1ϵ)(1-\epsilon)-approximation 𝖡𝖡\mathsf{BB}(k) model algorithm with communication cost O(nlogn1ϵO(1ϵ))O\left(n\log n\cdot\frac{1}{\epsilon}^{O(\frac{1}{\epsilon})}\right).

The dependency on ϵ\epsilon can be improved in the case of bipartite graphs. Assadi et al. (Sepehr Assadi, 2021) very recently gave an algorithm for Maximum Matching in bipartite graphs that solely relies on the computation of O(1ϵ2)O(\frac{1}{\epsilon^{2}}) maximal matchings in subgraphs of the input graph. Specifying the respective subgraphs, however, is slightly more involved, but can be done with communication cost O(n1ϵlog(1ϵ))O\left(n\frac{1}{\epsilon}\log(\frac{1}{\epsilon})\right). Taking this into account, a 𝖡𝖡\mathsf{BB}(k) model protocol with communication cost O((nlogn+n1ϵlog(1ϵ))1ϵ2)O\left(\left(n\log n+n\frac{1}{\epsilon}\log(\frac{1}{\epsilon})\right)\cdot\frac{1}{\epsilon^{2}}\right) can be obtained.

4.2. Maximal Independent Set

We now discuss how the random order Greedy MIS algorithm can be implemented in the 𝖡𝖡\mathsf{BB}(k) model with communication cost O(npolylogn)O(n\operatorname{\mathrm{poly}}\log n). A similar implementation in the massively parallel computation and the congested clique models was previously discussed by Ghaffari et al. (Ghaffari et al., 2018). The key idea of this implementation, i.e., a residual sparsity property of the random order Greedy algorithm, goes back to the multi-pass correlation clustering streaming algorithm by Ahn et al. (Ahn et al., 2015) and have since been used at multiple occasions (Konrad, 2018; Gamlath et al., 2019).

The random order Greedy MIS algorithm proceeds as follows: First, the algorithm identifies a uniform random permutation π:V[n]\pi:V\rightarrow[n], assigning each vertex a rank. Denote by viv_{i} the vertex with rank ii. The algorithm processes the vertices VV in the order specified by π\pi. When processing vertex viv_{i}, the algorithm attempts to insert viv_{i} into an initially empty independent set II if possible, i.e., if I{vi}I\cup\{v_{i}\} is an independent set.

To simulate this algorithm in the 𝖡𝖡\mathsf{BB}(k) model, the kk parties first agree on a uniform random permutation π\pi using public randomness. The simulation then proceeds in O(loglogn)O(\log\log n) phases, where in phase ii Greedy is simulated on vertices Vi:={vj:ki1j<ki}V_{i}:=\{v_{j}\ :\ k_{i-1}\leqslant j<k_{i}\}, for

ki=n112i.k_{i}=n^{1-\frac{1}{2^{i}}}\ .

Denote by IiI_{i} the independent set computed after phase ii (i.e., after having processed all vertices with ranks smaller than kik_{i}), and let I0={}I_{0}=\{\}. In each phase ii, the parties write the subgraph induced by vertices ViN(Ii1)V_{i}\setminus N(I_{i-1}) onto the blackboard, where N(Ii1)N(I_{i-1}) denotes the neighborhood of Ii1I_{i-1}. This information, together with Ii1I_{i-1}, then allows all parties to locally continue the simulation of Greedy on vertices ViV_{i}, resulting in independent set IiI_{i}. As proved by Ghaffari et al., the subgraph induced by vertices ViN(Ii1)V_{i}\setminus N(I_{i-1}) has O(npolylogn)O(n\operatorname{\mathrm{poly}}\log n) edges with high probability. Since the parties know Ii1I_{i-1} from the local simulations of the previous phase, they know which of their edges are contained in subgraph G[ViN(Ii1)]G[V_{i}\setminus N(I_{i-1})] and write all of these edges onto the blackboard.

After O(loglogn)O(\log\log n) phases, it can be seen that the graph induced by vertices with rank larger than kO(loglogn)k_{O(\log\log n)} has O(npolylogn)O(n\operatorname{\mathrm{poly}}\log n) edges. The parties then write this subgraph onto the blackboard, which allows all parties to complete their local simulations of Greedy.

The communication cost is dominated by the vertex-induced subgraphs written onto the blackboard. Each of these subgraphs contains O(npolylogn)O(n\operatorname{\mathrm{poly}}\log n) edges, and, since there are O(loglogn)O(\log\log n) phases, the total communication cost is O(npolylogn)O(n\operatorname{\mathrm{poly}}\log n).

4.3. (Δ+1)(\Delta+1)-coloring

We now discuss how the one-pass O(nlog3n)O(n\log^{3}n)-space streaming algorithm by Assadi et al. (Assadi et al., 2019) for (Δ+1)(\Delta+1)-coloring can be implemented in the 𝖡𝖡(k)\mathsf{BB}(k) model.

Assadi et al.’s streaming algorithm assumes that Δ\Delta is known prior to the processing of the stream. For each vertex vVv\in V, the algorithm first samples uniform random O(logn)O(\log n) colors Pv[Δ+1]P_{v}\subseteq[\Delta+1] from the color palette [Δ+1][\Delta+1]. Then, while processing the stream, the algorithm maintains the set of edges uvEuv\in E such that PuPvP_{u}\cap P_{v}\neq\emptyset, i.e., the palettes of uu and vv intersect. Denote this set of edges by EconflictE_{\text{conflict}}. Assadi et al. prove that |Econflict|=O(nlog2n)|E_{\text{conflict}}|=O(n\log^{2}n) w.h.p., and there exists a coloring χ:V[Δ+1]\chi:V\rightarrow[\Delta+1] of the conflict graph Gconflict=(V,Econflict)G_{\text{conflict}}=(V,E_{\text{conflict}}) such that every vertex vv is assigned a color from PvP_{v} w.h.p., i.e., χ(v)Pv\chi(v)\in P_{v}. It is easy to see that this coloring is also a valid coloring of the input graph GG. Since |Econflict|=O(nlog2n)|E_{\text{conflict}}|=O(n\log^{2}n), and space O(logn)O(\log n) is accounted for storing an edge, the space complexity of this algorithm is O(nlog3n)O(n\log^{3}n).

We implement this algorithm in the 𝖡𝖡\mathsf{BB}(k) model as follows. The parties maintain a guess Δ\Delta^{\prime} of Δ\Delta that is initially set to (n1)/2\lceil(n-1)/2\rceil and updated according to a binary search in the range [n1][n-1]. In each step of the binary search, our algorithm simulates Assadi et al.’s algorithm using the guess Δ\Delta^{\prime} instead of Δ\Delta. To this end, first, the parties use public randomness to determine the palettes PvP_{v} for every vertex vv. Knowing all the vertices’ palettes, every party ii is able to identify which of their edges are contained in EconflictE_{\text{conflict}}. Denote the conflict edges of party ii by EconflictiE_{\text{conflict}}^{i}. Then, every party writes EconflictiE_{\text{conflict}}^{i} to the blackboard. Next, the parties locally each compute the same coloring χΔ:V[Δ+1]\chi_{\Delta^{\prime}}:V\rightarrow[\Delta^{\prime}+1] of the conflict graph respecting the color palettes of the vertices, if such a coloring exists. If such a coloring exists, the parties conclude that ΔΔ\Delta\leqslant\Delta^{\prime}, and if such a coloring does not exist, the parties conclude that Δ>Δ\Delta>\Delta^{\prime}. The parties continue the binary search and output the coloring computed for the smallest value of Δ\Delta^{\prime}. This coloring is clearly a valid (Δ+1)(\Delta+1)-coloring.

Each iteration of the binary search requires writing the conflict graph onto the blackboard, which requires O(nlog3n)O(n\log^{3}n) bits. Since the binary search requires O(logn)O(\log n) rounds, this protocol has communication cost O(nlog4n)O(n\log^{4}n) bits.

5. Conclusion

In this paper, we gave a new lower bound technique that allowed us to prove Ω(n)\Omega(n) lower bounds on the communication complexity of most graph problems in the 𝖡𝖡(k)\mathsf{BB}(k) model, even if the edges are assigned uniformly at random to the parties. The strength of our approach is its wide applicability and the robustness under the random assignment of edges. We also showed that our lower bounds are tight, up to poly-logarithmic factors, for Maximal Independent Set, Maximal Matching, and (Δ+1)(\Delta+1)-coloring.

We conclude with pointing out a connection between the two-party communication setting and the 𝖡𝖡(k)\mathsf{BB}(k) model. It is not hard to see that a 𝖡𝖡(k)\mathsf{BB}(k) model protocol 𝒫\mathcal{P} can be implemented in the two-party communication setting with the same (or less) communication cost. Hence, lower bounds in the two-party communication model translate to the 𝖡𝖡(k)\mathsf{BB}(k) model. While some strong lower bounds in the two-party communication setting are known, such as the lower bound by Halldórsson et al. (Halldórsson et al., 2012) who showed that computing an α\alpha-approximation to Maximum Independent Set requires communication cost Ω~(n2/α2)\tilde{\Omega}(n^{2}/\alpha^{2}), these lower bounds do not hold under the random edge assignment assumption. This begs the following question: How can we prove strictly stronger lower bounds ω(n)\omega(n) for graph problems in the 𝖡𝖡(k)\mathsf{BB}(k) model if edges are assigned uniformly at random to the parties?

References

  • (1)
  • Ahn et al. (2015) Kook Jin Ahn, Graham Cormode, Sudipto Guha, Andrew McGregor, and Anthony Wirth. 2015. Correlation Clustering in Data Streams. In Proceedings of the 32nd International Conference on Machine Learning, ICML 2015, Lille, France, 6-11 July 2015 (JMLR Workshop and Conference Proceedings, Vol. 37), Francis R. Bach and David M. Blei (Eds.). JMLR.org, 2237–2246.
  • Ahn and Guha (2011) Kook Jin Ahn and Sudipto Guha. 2011. Linear Programming in the Semi-streaming Model with Application to the Maximum Matching Problem. In Automata, Languages and Programming - 38th International Colloquium, ICALP 2011, Zurich, Switzerland, July 4-8, 2011, Proceedings, Part II (Lecture Notes in Computer Science, Vol. 6756), Luca Aceto, Monika Henzinger, and Jirí Sgall (Eds.). Springer, 526–538.
  • Alon et al. (2015) Noga Alon, Noam Nisan, Ran Raz, and Omri Weinstein. 2015. Welfare Maximization with Limited Interaction. In Proceedings of the 2015 IEEE 56th Annual Symposium on Foundations of Computer Science (FOCS) (FOCS ’15). IEEE Computer Society, USA, 1499–1512. https://doi.org/10.1109/FOCS.2015.95
  • Assadi et al. (2019) Sepehr Assadi, Yu Chen, and Sanjeev Khanna. 2019. Sublinear Algorithms for (Δ\Delta + 1) Vertex Coloring. In Proceedings of the Thirtieth Annual ACM-SIAM Symposium on Discrete Algorithms, SODA 2019, San Diego, California, USA, January 6-9, 2019, Timothy M. Chan (Ed.). SIAM, 767–786. https://doi.org/10.1137/1.9781611975482.48
  • Assadi et al. (2016a) Sepehr Assadi, Sanjeev Khanna, and Yang Li. 2016a. Tight bounds for single-pass streaming complexity of the set cover problem. In Proceedings of the 48th Annual ACM SIGACT Symposium on Theory of Computing, STOC 2016, Cambridge, MA, USA, June 18-21, 2016, Daniel Wichs and Yishay Mansour (Eds.). ACM, 698–711.
  • Assadi et al. (2016b) Sepehr Assadi, Sanjeev Khanna, Yang Li, and Grigory Yaroslavtsev. 2016b. Maximum Matchings in Dynamic Graph Streams and the Simultaneous Communication Model. In Proceedings of the Twenty-Seventh Annual ACM-SIAM Symposium on Discrete Algorithms, SODA 2016, Arlington, VA, USA, January 10-12, 2016, Robert Krauthgamer (Ed.). SIAM, 1345–1364. https://doi.org/10.1137/1.9781611974331.ch93
  • Balliu et al. (2019) Alkida Balliu, Sebastian Brandt, Juho Hirvonen, Dennis Olivetti, Mikaël Rabie, and Jukka Suomela. 2019. Lower Bounds for Maximal Matchings and Maximal Independent Sets. In 60th IEEE Annual Symposium on Foundations of Computer Science, FOCS 2019, Baltimore, Maryland, USA, November 9-12, 2019. 481–497. https://doi.org/10.1109/FOCS.2019.00037
  • binti Khalil and Konrad (2020) Lidiya Khalidah binti Khalil and Christian Konrad. 2020. Constructing Large Matchings via Query Access to a Maximal Matching Oracle. In 40th IARCS Annual Conference on Foundations of Software Technology and Theoretical Computer Science, FSTTCS 2020, December 14-18, 2020, BITS Pilani, K K Birla Goa Campus, Goa, India (Virtual Conference) (LIPIcs, Vol. 182), Nitin Saxena and Sunil Simon (Eds.). Schloss Dagstuhl - Leibniz-Zentrum für Informatik, 26:1–26:15. https://doi.org/10.4230/LIPIcs.FSTTCS.2020.26
  • Braverman (2017) Mark Braverman. 2017. Interactive Information Complexity. SIAM Rev. 59, 4 (2017), 803–846. https://doi.org/10.1137/17M1139254
  • Braverman and Oshman (2015) Mark Braverman and Rotem Oshman. 2015. On information complexity in the broadcast model. In Proceedings of the 2015 ACM Symposium on Principles of Distributed Computing. 355–364.
  • Chandra et al. (1983) Ashok K. Chandra, Merrick L. Furst, and Richard J. Lipton. 1983. Multi-Party Protocols. In Proceedings of the Fifteenth Annual ACM Symposium on Theory of Computing (STOC ’83). Association for Computing Machinery, New York, NY, USA, 94–99. https://doi.org/10.1145/800061.808737
  • Cover and Thomas (2006) Thomas M. Cover and Joy A. Thomas. 2006. Elements of Information Theory (Wiley Series in Telecommunications and Signal Processing). Wiley-Interscience, USA.
  • Dobzinski et al. (2014) Shahar Dobzinski, Noam Nisan, and Sigal Oren. 2014. Economic Efficiency Requires Interaction. In Proceedings of the Forty-Sixth Annual ACM Symposium on Theory of Computing (New York, New York) (STOC ’14). Association for Computing Machinery, New York, NY, USA, 233–242. https://doi.org/10.1145/2591796.2591815
  • Drucker et al. (2012) Andrew Drucker, Fabian Kuhn, and Rotem Oshman. 2012. The Communication Complexity of Distributed Task Allocation (PODC ’12). Association for Computing Machinery, New York, NY, USA, 67–76. https://doi.org/10.1145/2332432.2332443
  • Eggert et al. (2012) Sebastian Eggert, Lasse Kliemann, Peter Munstermann, and Anand Srivastav. 2012. Bipartite Matching in the Semi-streaming Model. Algorithmica 63, 1-2 (2012), 490–508. https://doi.org/10.1007/s00453-011-9556-8
  • Fernandez et al. (2020) Manuel Fernandez, David P. Woodruff, and Taisuke Yasuda. 2020. Graph Spanners in the Message-Passing Model. In 11th Innovations in Theoretical Computer Science Conference, ITCS 2020, January 12-14, 2020, Seattle, Washington, USA (LIPIcs, Vol. 151), Thomas Vidick (Ed.). Schloss Dagstuhl - Leibniz-Zentrum für Informatik, 77:1–77:18. https://doi.org/10.4230/LIPIcs.ITCS.2020.77
  • Gamlath et al. (2019) Buddhima Gamlath, Sagar Kale, Slobodan Mitrovic, and Ola Svensson. 2019. Weighted Matchings via Unweighted Augmentations. In Proceedings of the 2019 ACM Symposium on Principles of Distributed Computing, PODC 2019, Toronto, ON, Canada, July 29 - August 2, 2019, Peter Robinson and Faith Ellen (Eds.). ACM, 491–500.
  • Ghaffari et al. (2018) Mohsen Ghaffari, Themis Gouleakis, Christian Konrad, Slobodan Mitrovic, and Ronitt Rubinfeld. 2018. Improved Massively Parallel Computation Algorithms for MIS, Matching, and Vertex Cover. In Proceedings of the 2018 ACM Symposium on Principles of Distributed Computing, PODC 2018, Egham, United Kingdom, July 23-27, 2018, Calvin Newport and Idit Keidar (Eds.). ACM, 129–138.
  • Halldórsson et al. (2012) Magnús M. Halldórsson, Xiaoming Sun, Mario Szegedy, and Chengu Wang. 2012. Streaming and Communication Complexity of Clique Approximation. In Automata, Languages, and Programming - 39th International Colloquium, ICALP 2012, Warwick, UK, July 9-13, 2012, Proceedings, Part I (Lecture Notes in Computer Science, Vol. 7391), Artur Czumaj, Kurt Mehlhorn, Andrew M. Pitts, and Roger Wattenhofer (Eds.). Springer, 449–460.
  • Huang et al. (2020) Zengfeng Huang, Bozidar Radunovic, Milan Vojnovic, and Qin Zhang. 2020. Communication complexity of approximate maximum matching in the message-passing model. Distributed Comput. 33, 6 (2020), 515–531. https://doi.org/10.1007/s00446-020-00371-6
  • Konrad (2015) Christian Konrad. 2015. Maximum Matching in Turnstile Streams. In Algorithms - ESA 2015 - 23rd Annual European Symposium, Patras, Greece, September 14-16, 2015, Proceedings (Lecture Notes in Computer Science, Vol. 9294), Nikhil Bansal and Irene Finocchi (Eds.). Springer, 840–852. https://doi.org/10.1007/978-3-662-48350-3_70
  • Konrad (2018) Christian Konrad. 2018. A Simple Augmentation Method for Matchings with Applications to Streaming Algorithms. In 43rd International Symposium on Mathematical Foundations of Computer Science, MFCS 2018, August 27-31, 2018, Liverpool, UK (LIPIcs, Vol. 117), Igor Potapov, Paul G. Spirakis, and James Worrell (Eds.). Schloss Dagstuhl - Leibniz-Zentrum für Informatik, 74:1–74:16.
  • Kuhn et al. (2016) Fabian Kuhn, Thomas Moscibroda, and Roger Wattenhofer. 2016. Local Computation: Lower and Upper Bounds. J. ACM 63, 2 (2016), 17:1–17:44. https://doi.org/10.1145/2742012
  • Low et al. (2012) Yucheng Low, Joseph Gonzalez, Aapo Kyrola, Danny Bickson, Carlos Guestrin, and Joseph M Hellerstein. 2012. Distributed graphlab: A framework for machine learning in the cloud. arXiv preprint arXiv:1204.6078 (2012).
  • McGregor (2005) Andrew McGregor. 2005. Finding Graph Matchings in Data Streams. In Approximation, Randomization and Combinatorial Optimization, Algorithms and Techniques, 8th International Workshop on Approximation Algorithms for Combinatorial Optimization Problems, APPROX 2005 and 9th InternationalWorkshop on Randomization and Computation, RANDOM 2005, Berkeley, CA, USA, August 22-24, 2005, Proceedings (Lecture Notes in Computer Science, Vol. 3624), Chandra Chekuri, Klaus Jansen, José D. P. Rolim, and Luca Trevisan (Eds.). Springer, 170–181.
  • Phillips et al. (2012) Jeff M Phillips, Elad Verbin, and Qin Zhang. 2012. Lower bounds for number-in-hand multiparty communication complexity, made easy. In Proceedings of the twenty-third annual ACM-SIAM symposium on Discrete Algorithms. SIAM, 486–501.
  • Pinsker (1964) Mark S Pinsker. 1964. Information and information stability of random variables and processes. Holden-Day.
  • Rao and Yehudayoff (2020) Anup Rao and Amir Yehudayoff. 2020. Communication Complexity: and Applications. Cambridge University Press.
  • Sepehr Assadi (2021) Robert E. Tarjan Sepehr Assadi, Cliff Liu. 2021. An Auction Algorithm for Bipartite Matching in Streaming and Massively Parallel Computation Models. In 4th Symposium on Simplicity in Algorithms, SOSA@SODA 2021.
  • Vempala et al. (2020) Santosh S. Vempala, Ruosong Wang, and David P. Woodruff. 2020. The Communication Complexity of Optimization. In Proceedings of the Thirty-First Annual ACM-SIAM Symposium on Discrete Algorithms (Salt Lake City, Utah) (SODA ’20). Society for Industrial and Applied Mathematics, USA, 1733–1752.
  • Woodruff and Zhang (2017) David P. Woodruff and Qin Zhang. 2017. When Distributed Computation is Communication Expensive. Distrib. Comput. 30, 5 (Oct. 2017), 309–323. https://doi.org/10.1007/s00446-014-0218-3
  • Yao (1977) Andrew Chi-Chin Yao. 1977. Probabilistic computations: Toward a unified measure of complexity. In 18th Annual Symposium on Foundations of Computer Science (sfcs 1977). IEEE Computer Society, 222–227.
  • Yao (1979) Andrew Chi-Chih Yao. 1979. Some Complexity Questions Related to Distributive Computing(Preliminary Report). In Proceedings of the Eleventh Annual ACM Symposium on Theory of Computing (Atlanta, Georgia, USA) (STOC ’79). Association for Computing Machinery, New York, NY, USA, 209–213. https://doi.org/10.1145/800135.804414