This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Algorithm-Level Confidentiality for Average Consensus on Time-Varying Directed Graphs

Huan Gao,  and Yongqiang Wang This paper has been accepted to IEEE Transactions on Network Science and Engineering as a regular paper. Please cite this paper as: Huan Gao and Yongqiang Wang, “Algorithm-Level Confidentiality for Average Consensus on Time-Varying Directed Graphs,” in IEEE Transactions on Network Science and Engineering, doi: 10.1109/TNSE.2022.3140274.Part of the results was presented at 2018 IEEE Conference on Communications and Network Security (CNS) [1]. The work was supported in part by the National Science Foundation under Grants ECCS-1912702 and CCF-2106293.Huan Gao was with the Department of Electrical and Computer Engineering, Clemson University, Clemson, SC 29634, USA. He is now with the School of Automation, Northwestern Polytechnical University, Xi’an 710129, China (email: [email protected]). The work was done when Huan Gao was with Clemson University.Yongqiang Wang is with the Department of Electrical and Computer Engineering, Clemson University, Clemson, SC 29634, USA (email: [email protected]).
Abstract

Average consensus plays a key role in distributed networks, with applications ranging from time synchronization, information fusion, load balancing, to decentralized control. Existing average consensus algorithms require individual agents to exchange explicit state values with their neighbors, which leads to the undesirable disclosure of sensitive information in the state. In this paper, we propose a novel average consensus algorithm for time-varying directed graphs that can protect the confidentiality of a participating agent against other participating agents. The algorithm injects randomness in interaction to obfuscate information on the algorithm-level and can ensure information-theoretic privacy without the assistance of any trusted third party or data aggregator. By leveraging the inherent robustness of consensus dynamics against random variations in interaction, our proposed algorithm can also guarantee the accuracy of average consensus. The algorithm is distinctly different from differential-privacy based average consensus approaches which enable confidentiality through compromising accuracy in obtained consensus value. Numerical simulations confirm the effectiveness and efficiency of our proposed approach.

Index Terms:
Average consensus, confidentiality, time-varying directed graphs.

I Introduction

Averaging consensus is an important tool in distributed computing. For a network of NN agents interacting on a graph, average consensus can enable the states of all agents to converge to the average of their initial values through local interactions between neighboring agents.

Recently, average consensus is finding increased applications in load balancing [2, 3], network synchronization [4], distributed information fusion [5, 6], and decentralized control [7, 8]. To ensure all agents converge to the average value of their initial values, conventional average consensus approaches require individual agents to exchange explicit state values with their neighbors. This results in the disclosure of sensitive state information, which is sometimes undesirable in terms of confidentiality. In fact, in many applications such as the smart grid, health-care or banking networks, confidentiality is crucial for promoting participation in collaboration since individual agents tend not to trade confidentiality for performance [9, 10, 11]. For instance, a group of people using average consensus to reach a common opinion may want to keep their individual opinions secret [12]. Another typical example is power systems in which multiple generators have to reach agreement on cost while maintaining their individual generation information confidential [13].

To achieve confidentiality in average consensus, recently results have started to emerge. A commonly used confidentiality mechanism is differential privacy from the database literature [14, 15, 16, 17, 18, 19, 20] (and its variants [21, 22]) which injects independent (and hence uncorrelated) noises directly to agents’ states in order to achieve confidentiality in average consensus. However, the use of independent noises on the states in these approaches prevents converging to the exact average value [23]. To improve consensus accuracy, which is crucial in cyber-physical systems and sensor networks, [24, 25, 26, 27, 28, 29, 30, 31] inject carefully calculated correlated additive noises to agents’ states, instead of independent (and hence uncorrelated) noises used in differential-privacy based approaches. (A similar approach was proposed in [32] to achieve maximum consensus.) However, these prior works only consider average consensus under balanced and static network topologies. Different from injecting noises to agents’ states in the aforementioned approaches, [33] employed carefully designed mask maps to protect the actual states. Observability based approaches have also been reported to protect the confidentiality of multi-agent consensus [34, 35, 36]. Its idea is to design the topology of interactions such that the observability from a compromised agent is minimized, which amounts to minimizing the ability of the compromised agent to infer the initial states of other agents. Recently, encryption based approaches have been proposed to protect the confidentiality by encrypting exchanged messages with the assistance of additive homomorphic encryption [37, 38, 39, 40], with the price of increasing computation and communication overhead. Another confidentiality approach was proposed in [41] where each agent’s confidentiality is protected by decomposing its state into two sub-states. However, [41] relies on undirected interactions and is inapplicable to time-varying directed graphs considered in this paper.

This paper addresses the confidentiality of average consensus under time-varying directed graphs that are not necessarily balanced. Since push-sum based average consensus approaches do not require balanced topologies, we build our confidential average consensus algorithm on the push-sum approach. More specifically, to protect the confidentiality of the initial states of participating agents, in the first several iterations, we let agents send random values instead of their actual states to obfuscate their initial values. Of course, to guarantee the accuracy of average consensus, we have to judiciously design the state-update rule such that the randomness added in the first several iterations does not affect the final convergence result. Different from approaches injecting correlated additive noises directly to agents’ states, our approach adds independent (and hence uncorrelated in time) randomness directly to the average consensus dynamics, which makes it applicable to time-varying directed graphs. Compared with our prior work in [40] which employs homomorphic encryption to preserve confidentiality and [41] which protects each agent’s confidentiality by decomposing its state into two sub-states, this paper proposes a different approach that enables confidentiality by judiciously adding randomness in interaction dynamics. More importantly, both [40] and [41] rely on undirected interactions and hence are inapplicable to time-varying directed graphs considered in this paper. Some of the results here were presented at the 2018 IEEE Conference on Communications and Network Security (CNS) [1]. Compared with the conference version, the journal version has the following significant differences: 1) the journal version extends the results for constant directed graphs in [1] to time-varying directed graphs; 2) the journal version provides formal and rigorous analysis of convergence rate that does not exist in the conference version; 3) the journal version allows multiple adversaries to collude to infer the sensitive value a target agent, which is not addressed in our conference version; and 4) the journal version revises and enhances the proposed confidential average consensus algorithm to guarantee information-theoretic privacy, which is stronger than the confidentiality achieved in the conference version.

II Preliminaries and Problem Formulation

II-A Graph Representation

We represent a network of NN agents as a sequence of time-varying directed graphs {𝒢(k)=(𝒱,(k))}\{\mathcal{G}(k)=(\mathcal{V},\,\mathcal{E}(k))\} where 𝒱={1,2,,N}\mathcal{V}=\{1,2,\ldots,N\} is the set of agents and k=0,1,k=0,1,\ldots is the time index. (k)𝒱×𝒱\mathcal{E}(k)\subset\mathcal{V}\times\mathcal{V} is the edge set at time kk, whose elements are such that (i,j)(k)(i,\,j)\in\mathcal{E}(k) holds if and only if there exists a directed edge from agent jj to agent ii at time kk, i.e., agent jj can send messages to agent ii at time kk. For notational convenience, we assume that there are no self edges, i.e., (i,i)(k)(i,\,i)\notin\mathcal{E}(k) for all kk and i𝒱i\in\mathcal{V}. At time kk, each edge (i,j)(k)(i,\,j)\in\mathcal{E}(k) has an associated weight, pij(k)>0p_{ij}(k)>0. The out-neighbor set of agent ii at time kk, which represents the set of agents that can receive messages from agent ii at time kk, is denoted as 𝒩iout(k)={j𝒱|(j,i)(k)}\mathcal{N}_{i}^{out}(k)=\{j\in\mathcal{V}\,|\,(j,\,i)\in\mathcal{E}(k)\}. Similarly, at time kk, the in-neighbor set of agent ii, which represents the set of agents that can send messages to agent ii at time kk, is denoted as 𝒩iin(k)={j𝒱|(i,j)(k)}\mathcal{N}_{i}^{in}(k)=\{j\in\mathcal{V}\,|\,(i,\,j)\in\mathcal{E}(k)\}. From the above definitions, it can be obtained that i𝒩jout(k)i\in\mathcal{N}_{j}^{out}(k) and j𝒩iin(k)j\in\mathcal{N}_{i}^{in}(k) are equivalent. Agent ii’s out-degree at time instant kk is represented by Diout(k)=|𝒩iout(k)|D_{i}^{out}(k)=|\mathcal{N}_{i}^{out}(k)| and its in-degree is represented by Diin(k)=|𝒩iin(k)|D_{i}^{in}(k)=|\mathcal{N}_{i}^{in}(k)|, where |𝒮||\mathcal{S}| is the cardinality of the set 𝒮\mathcal{S}.

At iteration kk, the incidence matrix 𝐂(k)=[cil(k)]N×E(k)\mathbf{C}(k)=[c_{il}(k)]_{N\times E(k)} for graph 𝒢(k)=(𝒱,(k))\mathcal{G}(k)=(\mathcal{V},\mathcal{E}(k)) is defined as

cil(k)={1if thel-th edge in(k)is(i,j)1if thel-th edge in(k)is(j,i)0otherwisec_{il}(k)=\left\{\begin{aligned} 1\quad&\textnormal{if the}\ l\textnormal{-th edge in}\ \mathcal{E}(k)\ \textnormal{is}\ (i,\,j)\\ -1\quad&\textnormal{if the}\ l\textnormal{-th edge in}\ \mathcal{E}(k)\ \textnormal{is}\ (j,\,i)\\ 0\quad&\textnormal{otherwise}\end{aligned}\right. (1)

where E(k)=|(k)|E(k)=|\mathcal{E}(k)| represents the number of edges in (k)\mathcal{E}(k). Note that the ll-th column of 𝐂(k)\mathbf{C}(k) is corresponding to the ll-th edge in (k)\mathcal{E}(k), and the sum of each column of 𝐂(k)\mathbf{C}(k) is 0, i.e., 𝟏T𝐂(k)=𝟎T\mathbf{1}^{T}\mathbf{C}(k)=\mathbf{0}^{T}.

For a sequence of time-varying directed graphs {𝒢(k)=(𝒱,(k))}\{\mathcal{G}(k)=(\mathcal{V},\,\mathcal{E}(k))\}, we define \mathcal{E}_{\infty} as the set of directed edges (i,j)(i,\,j) that exist for infinitely many time instants, i.e.,

={(i,j)|(i,j)(k)for infinitely many indicesk}\mathcal{E}_{\infty}=\big{\{}(i,\,j)\big{|}(i,\,j)\in\mathcal{E}(k)\ \text{for infinitely many indices}\ k\big{\}} (2)

We focus on time-varying directed graphs which satisfy the following assumptions:

Assumption 1

For a sequence of time-varying directed graphs {𝒢(k)=(𝒱,(k))}\{\mathcal{G}(k)=(\mathcal{V},\,\mathcal{E}(k))\}, for any i,j𝒱i,j\in\mathcal{V} with iji\neq j, there exists at least one directed path from ii to jj in (𝒱,)(\mathcal{V},\,\mathcal{E}_{\infty}), i.e., (𝒱,)(\mathcal{V},\,\mathcal{E}_{\infty}) is strongly connected.

Assumption 2

For a sequence of time-varying directed graphs {𝒢(k)=(𝒱,(k))}\{\mathcal{G}(k)=(\mathcal{V},\,\mathcal{E}(k))\}, there exists an integer T1T\geq 1 such that for every (i,j)(i,\,j)\in\mathcal{E}_{\infty}, agent jj directly communicates with agent ii at least once in every TT consecutive time instants. TT is called intercommunication interval bound.

Assumption 3

We assume that each agent ii has access to its out-degree Diout(k)D_{i}^{out}(k) at each iteration kk.

Remark 1

Assumption 3 is widely used in existing literature on time-varying directed graphs such as [42, 43, 44]. In fact, in many directed graphs, it is feasible for a node to know its out-neighbors. For example, in many safety-critical systems such as industrial control systems, the exchange of data occurs in a directed way due to unidirectional gateways (aka data diode) whereas control messages (a special type of messages used to configure network connections) can be exchanged in a bidirectional manner to establish connections [45].

II-B The Conventional Push-Sum

The conventional push-sum considers NN agents interacting on a constant directed graph 𝒢=(𝒱,)\mathcal{G}=(\mathcal{V},\,\mathcal{E}), with each agent having an initial state xi0x_{i}^{0} (i=1,2,,Ni=1,2,\ldots,N) [46, 47]. Represent the average value of all initial states as x¯0=j=1Nxj0/N\bar{x}^{0}={\sum_{j=1}^{N}x_{j}^{0}}/N. The conventional push-sum algorithm conducts two iterative computations simultaneously, and allows each agent to obtain the exact average of the initial values x¯0\bar{x}^{0} in an asymptotic way. This mechanism is summarized in Algorithm 0 below:

  Algorithm 0: The conventional push-sum algorithm

 

  1. 1.

    NN agents interact on a constant directed graph 𝒢=(𝒱,)\mathcal{G}=(\mathcal{V},\,\mathcal{E}). Each agent ii is initialized with si(0)=xi0s_{i}(0)=x_{i}^{0}, wi(0)=1w_{i}(0)=1, and πi(0)=si(0)/wi(0)\pi_{i}(0)=s_{i}(0)/w_{i}(0). The weight pijp_{ij} associated with the edge (i,j)(i,\,j)\in\mathcal{E} satisfies pij(0,1)p_{ij}\in(0,1) if j𝒩iin{i}j\in\mathcal{N}_{i}^{in}\cup\{i\} is true and pij=0p_{ij}=0 otherwise. For any given j=1,2,,Nj=1,2,\ldots,N, pijp_{ij} satisfies i=1Npij=1\sum_{i=1}^{N}p_{ij}=1.

  2. 2.

    At iteration step kk:

    1. (a)

      Agent ii calculates pjisi(k)p_{ji}s_{i}(k) and pjiwi(k)p_{ji}w_{i}(k), and sends both values to all of its out-neighbors j𝒩ioutj\in\mathcal{N}_{i}^{out}.

    2. (b)

      After receiving the values of pijsj(k)p_{ij}s_{j}(k) and pijwj(k)p_{ij}w_{j}(k) from all its in-neighbors j𝒩iinj\in\mathcal{N}_{i}^{in}, agent ii updates sis_{i} and wiw_{i} as follows:

      {si(k+1)=j𝒩iin{i}pijsj(k)wi(k+1)=j𝒩iin{i}pijwj(k)\left\{\begin{aligned} &\ s_{i}(k+1)=\sum_{j\in\mathcal{N}_{i}^{in}\cup\{i\}}p_{ij}s_{j}(k)\\ &\ w_{i}(k+1)=\sum_{j\in\mathcal{N}_{i}^{in}\cup\{i\}}p_{ij}w_{j}(k)\end{aligned}\right. (3)
    3. (c)

      Agent ii uses the ratio πi(k+1)=si(k+1)/wi(k+1)\pi_{i}(k+1)=s_{i}(k+1)/w_{i}(k+1) to estimate the average value x¯0=j=1Nxj0/N\bar{x}^{0}={\sum_{j=1}^{N}x_{j}^{0}}/N.

 

For the sake of notational simplicity, we rewrite (3) in the following more compact form:

{𝐬(k+1)=𝐏𝐬(k)𝐰(k+1)=𝐏𝐰(k)\left\{\begin{aligned} &\ \mathbf{s}(k+1)=\mathbf{P}\mathbf{s}(k)\\ &\ \mathbf{w}(k+1)=\mathbf{P}\mathbf{w}(k)\end{aligned}\right. (4)

where 𝐬(k)=[s1(k),s2(k),,sN(k)]T\mathbf{s}(k)=[s_{1}(k),s_{2}(k),\ldots,s_{N}(k)]^{T} and 𝐰(k)=[w1(k),w2(k),,wN(k)]T\mathbf{w}(k)=[w_{1}(k),w_{2}(k),\ldots,w_{N}(k)]^{T}, and 𝐏=[pij]\mathbf{P}=[p_{ij}]. From Algorithm 0, we have 𝐬(0)=[x10,x20,,xN0]T\mathbf{s}(0)=[x_{1}^{0},x_{2}^{0},\ldots,x_{N}^{0}]^{T} and 𝐰(0)=𝟏\mathbf{w}(0)=\mathbf{1}. We can also obtain that the matrix 𝐏\mathbf{P} is column-stochastic, i.e., i=1Npij=1\sum_{i=1}^{N}p_{ij}=1 holds for j=1,2,,Nj=1,2,\ldots,N.

At iteration step kk, each agent computes the ratio πi(k+1)=si(k+1)/wi(k+1)\pi_{i}(k+1)=s_{i}(k+1)/w_{i}(k+1) to estimate the average value x¯0=j=1Nxj0/N\bar{x}^{0}={\sum_{j=1}^{N}x_{j}^{0}}/N. Since 𝒢\mathcal{G} is assumed to be a strongly connected directed graph, 𝐏k\mathbf{P}^{k} will converge to a rank-11 matrix exponentially fast [48, 49]. Defining 𝐏\mathbf{P}^{\infty} as the limit of 𝐏k\mathbf{P}^{k} as kk\rightarrow\infty, we can obtain the form of 𝐏\mathbf{P}^{\infty} as 𝐏=𝐯𝟏T\mathbf{P}^{\infty}=\mathbf{v}\mathbf{1}^{T} where 𝐯=[v1,v2,,vN]T\mathbf{v}=[v_{1},v_{2},\ldots,v_{N}]^{T}. Using the facts 𝐬(k)=𝐏k𝐬(0)\mathbf{s}(k)=\mathbf{P}^{k}\mathbf{s}(0) and 𝐰(k)=𝐏k𝐰(0)\mathbf{w}(k)=\mathbf{P}^{k}\mathbf{w}(0), we can further have [50]:

πi()=si()wi()=[𝐏𝐬(0)]i[𝐏𝐰(0)]i=vij=1Nsj(0)vij=1Nwj(0)=x¯0\displaystyle\pi_{i}(\infty)=\frac{s_{i}(\infty)}{w_{i}(\infty)}=\frac{[\mathbf{P}^{\infty}\mathbf{s}(0)]_{i}}{[\mathbf{P}^{\infty}\mathbf{w}(0)]_{i}}=\frac{v_{i}\sum_{j=1}^{N}s_{j}(0)}{v_{i}\sum_{j=1}^{N}w_{j}(0)}=\bar{x}^{0} (5)

where [𝐏𝐬(0)]i[\mathbf{P}^{\infty}\mathbf{s}(0)]_{i} and [𝐏𝐰(0)]i[\mathbf{P}^{\infty}\mathbf{w}(0)]_{i} represent the ii-th element of vector 𝐏𝐬(0)\mathbf{P}^{\infty}\mathbf{s}(0) and vector 𝐏𝐰(0)\mathbf{P}^{\infty}\mathbf{w}(0), respectively. Hence, all estimates π1(k),π2(k),,πN(k)\pi_{1}(k),\pi_{2}(k),\ldots,\pi_{N}(k) will asymptotically converge to the average x¯0=j=1Nxj0/N\bar{x}^{0}={\sum_{j=1}^{N}x_{j}^{0}}/N.

II-C Problem Formulation

In this paper, we will address average consensus under time-varying directed graphs while protecting the confidentiality of participating agents against adversaries. To this end, we first present some assumptions and definitions.

Assumption 4

We assume that all agents’ initial states xi0x_{i}^{0} are bounded. Without loss of generality, the lower bound and upper bound are denoted as aa and bb, respectively. Both aa and bb are assumed known to all agents.

Remark 2

It is worth noting that although the bounds aa and bb are assumed known to all agents, this does not mean that the minimum and maximum of all agents’ initial states are known to all agents. In fact, aa (resp. bb) can be arbitrarily small (resp. large) compared with the actual minimum (resp. maximum) of agents’ initial states.

Definition 1

We define an honest-but-curious adversary as an agent who follows all protocol steps correctly but collects received messages in an attempt to infer the initial value of other participating agents.

Assumption 5

We assume that agents can collude, i.e., a set of honest-but-curious agents 𝒜\mathcal{A} can share information with each other to infer the initial value xi0x_{i}^{0} of a target agent i𝒜i\notin\mathcal{A}.

Definition 2

We define that confidentiality (privacy) of the initial value xi0x_{i}^{0} of agent ii is preserved if xi0x_{i}^{0} is indistinguishable from the viewpoint of honest-but-curious adversaries 𝒜\mathcal{A}. By “indistinguishable,” we mean that the probability distribution of information set accessible to 𝒜\mathcal{A} does not change when agent ii’s initial state xi0x_{i}^{0} is altered to any x~i0xi0\tilde{x}_{i}^{0}\neq x_{i}^{0} under the constraint that the sum of the initial states of all nodes not in 𝒜\mathcal{A} (i.e., j𝒱𝒜xj0\sum_{j\in\mathcal{V}\setminus\mathcal{A}}x_{j}^{0}) is unchanged.

Our definition of confidentiality requires perfect indistinguishability of a target agent’s different initial states from the viewpoint of honest-but-curious adversaries 𝒜\mathcal{A}, and, therefore, is more stringent than the confidentiality definition in [31, 32, 51, 52, 53] which defines confidentiality as the inability of an adversary to uniquely determine the sensitive value.

We next show that the conventional push-sum is not confidential. From (3) and (4), an honest-but-curious agent ii can receive pijsj(0)p_{ij}s_{j}(0) and pijwj(0)p_{ij}w_{j}(0) from its in-neighbor agent jj after the first iteration step k=0k=0. Then agent ii is able to uniquely determine xj0x_{j}^{0} by xj0=sj(0)=pijsj(0)pijwj(0)x_{j}^{0}=s_{j}(0)=\frac{p_{ij}s_{j}(0)}{p_{ij}w_{j}(0)} using the fact wj(0)=1w_{j}(0)=1. Therefore, an honest-but-curious agent can always infer the initial values of all its in-neighbors, and hence the conventional push-sum algorithm cannot provide confidentiality against honest-but-curious adversaries. It is worth noting that using a similar argument, we can also obtain that the conventional push-sum is not confidential even when the weight is allowed to be time-varying (e.g., [47].)

III The Confidentiality Algorithm and Performance Analysis

In this section, we will propose a confidential average consensus algorithm for time-varying directed graphs, and then provide rigorous analysis of its convergence rate and enabled strength of confidentiality.

III-A Confidential Average Consensus Algorithm

The analysis above reveals that using the same weight pijp_{ij} for both pijsj(0)p_{ij}s_{j}(0) and pijwj(0)p_{ij}w_{j}(0) discloses the initial state value. Motivated by this observation and the work in [29], here we introduce a novel confidential average consensus algorithm which injects randomness in the dynamics of interactions in iterations k=0,,Kk=0,\ldots,K. Note that here KK is a non-negative integer and is known to every agent. Its influence will be discussed in detail in Remark 11 and Remark 12.

  Algorithm 1: Confidential average consensus algorithm

 

  1. 1.

    NN agents interact on a sequence of time-varying directed graphs {𝒢(k)=(𝒱,(k))}\{\mathcal{G}(k)=(\mathcal{V},\,\mathcal{E}(k))\}. Each agent ii is initialized with si(0)=1N2+(N2)(xi0a)(ba)N2[1N2,N1N2]s_{i}(0)=\frac{1}{N^{2}}+\frac{(N-2)(x_{i}^{0}-a)}{(b-a)N^{2}}\in[\frac{1}{N^{2}},\,\frac{N-1}{N^{2}}], wi(0)=1w_{i}(0)=1, and πi(0)=baN2[N×frac(Nsi(0)wi(0))1]+a\pi_{i}(0)=\frac{b-a}{N-2}[N\times{\rm frac}(\frac{Ns_{i}(0)}{w_{i}(0)})-1]+a where the function frac(x)=xx{\rm frac}(x)=x-\lfloor x\rfloor denotes the fractional part of a real number xx (here x\lfloor x\rfloor represents the largest integer not greater than xx).

  2. 2.

    At iteration step kk:

    1. (a)

      Agent ii generates a set of random weights {pji(k)(ε, 1)|j𝒩iout(k){i}}\big{\{}p_{ji}(k)\in(\varepsilon,\,1)\,\big{|}\,j\in\mathcal{N}_{i}^{out}(k)\cup\{i\}\big{\}} with the sum of this set equal to 11, and sets Δwji(k)=pji(k)wi(k)\Delta w_{ji}(k)=p_{ji}(k)w_{i}(k) for j𝒩iout(k){i}j\in\mathcal{N}_{i}^{out}(k)\cup\{i\}.

    2. (b)

      If kKk\leq K, agent ii independently generates uniformly distributed values Δsji(k)[0, 1)\Delta s_{ji}(k)\in[0,\,1) for its out-neighbors j𝒩iout(k)j\in\mathcal{N}_{i}^{out}(k), and sets Δsii(k)=frac(si(k)j𝒩iout(k)Δsji(k))\Delta s_{ii}(k)={\rm frac}\big{(}s_{i}(k)-\sum_{j\in\mathcal{N}_{i}^{out}(k)}\Delta s_{ji}(k)\big{)}; otherwise, agent ii sets Δsji(k)=pji(k)si(k)\Delta s_{ji}(k)=p_{ji}(k)s_{i}(k) for j𝒩iout(k){i}j\in\mathcal{N}_{i}^{out}(k)\cup\{i\}.

    3. (c)

      Agent ii sends Δsji(k)\Delta s_{ji}(k) and Δwji(k)\Delta w_{ji}(k) to its out-neighbors j𝒩iout(k)j\in\mathcal{N}_{i}^{out}(k).

    4. (d)

      After receiving Δsij(k)\Delta s_{ij}(k) and Δwij(k)\Delta w_{ij}(k) from its in-neighbors j𝒩iin(k)j\in\mathcal{N}_{i}^{in}(k), agent ii updates sis_{i} and wiw_{i} as

      si(k+1)={frac(j𝒩iin(k){i}Δsij(k))forkKj𝒩iin(k){i}Δsij(k)forkK+1s_{i}(k+1)=\left\{\begin{aligned} &{\rm frac}\Big{(}\sum_{j\in\mathcal{N}_{i}^{in}(k)\cup\{i\}}\Delta s_{ij}(k)\Big{)}\ \textnormal{for}\,k\leq K\\ &\sum_{j\in\mathcal{N}_{i}^{in}(k)\cup\{i\}}\Delta s_{ij}(k)\quad\ \textnormal{for}\,k\geq K+1\\ \end{aligned}\right. (6)

      and

      wi(k+1)=j𝒩iin(k){i}Δwij(k)fork0\displaystyle w_{i}(k+1)=\sum_{j\in\mathcal{N}_{i}^{in}(k)\cup\{i\}}\Delta w_{ij}(k)\quad\textnormal{for}\ k\geq 0 (7)

      respectively.

    5. (e)

      Agent ii uses the ratio πi(k+1)=baN2[N×frac(Nsi(k+1)wi(k+1))1]+a\pi_{i}(k+1)=\frac{b-a}{N-2}[N\times{\rm frac}(\frac{Ns_{i}(k+1)}{w_{i}(k+1)})-1]+a to estimate the average value x¯0=j=1Nxj0/N\bar{x}^{0}={\sum_{j=1}^{N}x_{j}^{0}}/N.

 

Remark 3

Compared to the conventional confidentiality-violating push-sum algorithm which broadcasts messages, Algorithm 1 needs agent ii to send different random numbers to different out-neighbors in iterations kKk\leq K. This is a price of obtaining confidentiality without losing accuracy in the time-varying directed topology case.

Remark 4

The way of injecting randomness in Δsji(k)\Delta s_{ji}(k) is different in iterations kKk\leq K from k>Kk>K. In fact, in iterations kKk\leq K, Δsji(k)\Delta s_{ji}(k) can be nonzero even when si(k)s_{i}(k) is zero. This is crucial in enabling strong confidentiality as receiving Δsji(k)\Delta s_{ji}(k) of a value zero will not allow the recipient to infer information about si(k)s_{i}(k).

Remark 5

In Algorithm 1, the coupling weights are randomly chosen at every iteration. Compared with the commonly used deterministic setting in which the weights are set as pji(k)=1/(Diout(k)+1)p_{ji}(k)=1/(D_{i}^{out}(k)+1) for all j𝒩iout(k){i}j\in\mathcal{N}_{i}^{out}(k)\cup\{i\}, our setting is more general since it includes the commonly used setting as a special case by fixing pji(k)p_{ji}(k) to deterministic values. Furthermore, the random weights beyond step KK in Algorithm 1 can provide additional confidentiality protection for intermediate states si(k)s_{i}(k) after iteration KK. Given Δsji(k)=pji(k)si(k)\Delta s_{ji}(k)=p_{ji}(k)s_{i}(k) for kK+1k\geq K+1, we can see that using random weights pji(k)p_{ji}(k) makes the intermediate state si(k)s_{i}(k) more difficult to infer than using deterministic weights pji(k)p_{ji}(k).

Setting Δsji(k)\Delta s_{ji}(k), Δwji(k)\Delta w_{ji}(k), and pji(k)p_{ji}(k) to 0 for j𝒩iout(k){i}j\notin\mathcal{N}_{i}^{out}(k)\cup\{i\}, we can rewrite the update rules of 𝐬(k)\mathbf{s}(k) for kK+1k\geq K+1 and 𝐰(k)\mathbf{w}(k) for k0k\geq 0 as

{si(k+1)=j=1NΔsij(k)=j=1Npij(k)sj(k)forkK+1wi(k+1)=j=1NΔwij(k)=j=1Npij(k)wj(k)fork0\left\{\begin{aligned} &s_{i}(k+1)=\sum_{j=1}^{N}\Delta s_{ij}(k)=\sum_{j=1}^{N}p_{ij}(k)s_{j}(k)\ \,\textnormal{for}\ k\geq K+1\\ &w_{i}(k+1)=\sum_{j=1}^{N}\Delta w_{ij}(k)=\sum_{j=1}^{N}p_{ij}(k)w_{j}(k)\ \ \textnormal{for}\ k\geq 0\end{aligned}\right. (8)

Denoting 𝐬(k)\mathbf{s}(k), 𝐰(k)\mathbf{w}(k), and 𝐏(k)\mathbf{P}(k) as 𝐬(k)=[s1(k)sN(k)]T\mathbf{s}(k)=[s_{1}(k)\,\cdots\,s_{N}(k)]^{T}, 𝐰(k)=[w1(k)wN(k)]T\mathbf{w}(k)=[w_{1}(k)\,\cdots\,w_{N}(k)]^{T}, and 𝐏(k)=[pij(k)]N×N\mathbf{P}(k)=[p_{ij}(k)]_{N\times N}, we can further rewrite (8) into a matrix form

{𝐬(k+1)=𝐏(k)𝐬(k)forkK+1𝐰(k+1)=𝐏(k)𝐰(k)fork0\left\{\begin{aligned} &\ \mathbf{s}(k+1)=\mathbf{P}(k)\mathbf{s}(k)\quad\ \ \text{for}\ k\geq K+1\\ &\ \mathbf{w}(k+1)=\mathbf{P}(k)\mathbf{w}(k)\quad\text{for}\ k\geq 0\end{aligned}\right. (9)

For iteration k=0k=0, we have 𝐬(0)=[x10,x20,,xN0]T\mathbf{s}(0)=[x_{1}^{0},x_{2}^{0},\ldots,x_{N}^{0}]^{T} and 𝐰(0)=𝟏\mathbf{w}(0)=\mathbf{1}. From Algorithm 1, we know that 𝐏(k)\mathbf{P}(k) in (9) is time-varying and column-stochastic for k0k\geq 0.

Defining the transition matrix as follows

𝚽(k:t)=𝐏(k)𝐏(t)\displaystyle\mathbf{\Phi}(k:t)=\mathbf{P}(k)\cdots\mathbf{P}(t) (10)

for all kk and tt with ktk\geq t, where 𝚽(k:k)=𝐏(k)\mathbf{\Phi}(k:k)=\mathbf{P}(k), we can rewrite (9) as

{𝐬(k+1)=𝚽(k:K+1)𝐬(K+1)forkK+1𝐰(k+1)=𝚽(k:0)𝐰(0)fork0\left\{\begin{aligned} &\mathbf{s}(k+1)=\mathbf{\Phi}(k:K+1)\mathbf{s}(K+1)\quad\text{for}\ k\geq K+1\\ &\mathbf{w}(k+1)=\mathbf{\Phi}(k:0)\mathbf{w}(0)\qquad\text{for}\ k\geq 0\end{aligned}\right. (11)

III-B Convergence Analysis

Next we prove that Algorithm 1 can guarantee that the estimates of all agents converge to the exact average value of initial values. We will also analyze the rate of convergence of Algorithm 1. Using the convergence definition in [43] and [54], we define the rate of convergence to be at least γ(0, 1)\gamma\in(0,\,1) if there exists a positive constant value CC such that 𝝅(k)x¯0𝟏Cγk\big{\|}\boldsymbol{\pi}(k)-\bar{x}^{0}\mathbf{1}\big{\|}\leq C\gamma^{k} is true for all kk, where 𝝅(k)=[π1(k),,πN(k)]T\boldsymbol{\pi}(k)=[\pi_{1}(k),\ldots,\pi_{N}(k)]^{T} and x¯0=j=1Nxj0/N\bar{x}^{0}={\sum_{j=1}^{N}x_{j}^{0}}/N is the average value. Note that this definition means a smaller γ\gamma corresponding to a faster convergence. To analyze the convergence rate of Algorithm 1, we first introduce Lemma 1 below:

Lemma 1

For a network of NN agents represented by a sequence of time-varying directed graphs {𝒢(k)=(𝒱,(k))}\{\mathcal{G}(k)=(\mathcal{V},\,\mathcal{E}(k))\} which satisfy Assumptions 1, 2, and 3, under Algorithm 1, each agent ii has wi(k)εT(N1)w_{i}(k)\geq\varepsilon^{T(N-1)} for k1k\geq 1 where TT is defined in Assumption 2.

Proof: For k1k\geq 1, from (11) we have

𝐰(k)=𝚽(k1:0)𝟏\displaystyle\mathbf{w}(k)=\mathbf{\Phi}(k-1:0)\mathbf{1} (12)

Represent δ(k)\delta(k) as

δ(k)min1iNwi(k)=min1iN[𝚽(k1:0) 1]i\displaystyle\delta(k)\triangleq\min_{1\leq i\leq N}{w}_{i}(k)=\min_{1\leq i\leq N}[\mathbf{\Phi}(k-1:0)\,\mathbf{1}]_{i} (13)

for k1k\geq 1. To prove wi(k)εT(N1)w_{i}(k)\geq\varepsilon^{T(N-1)} for k1k\geq 1, it is sufficient to prove δ(k)εT(N1)\delta(k)\geq\varepsilon^{T(N-1)} for k1k\geq 1. We divide our proof into two parts: 1kT(N1)1\leq k\leq T(N-1) and kT(N1)+1k\geq T(N-1)+1.

Part 1: δ(k)εT(N1)\delta(k)\geq\varepsilon^{T(N-1)} for 1kT(N1)1\leq k\leq T(N-1). One can verify that the following relationship holds

[𝚽(k1:0)]ii\displaystyle[\mathbf{\Phi}(k-1:0)]_{ii} (14)
=\displaystyle= [𝐏(k1)𝐏(0)]ii\displaystyle[\mathbf{P}(k-1)\cdots\mathbf{P}(0)]_{ii}
\displaystyle\geq [𝐏(k1)]ii[𝐏(k2)𝐏(0)]ii\displaystyle[\mathbf{P}(k-1)]_{ii}\,[\mathbf{P}(k-2)\cdots\mathbf{P}(0)]_{ii}
\displaystyle\geq ε[𝚽(k2:0)]ii\displaystyle\varepsilon\,[\mathbf{\Phi}(k-2:0)]_{ii}

Given [𝚽(0:0)]ii=[𝐏(0)]iiε[\mathbf{\Phi}(0:0)]_{ii}=[\mathbf{P}(0)]_{ii}\geq\varepsilon, one can obtain [𝚽(k1:0)]iiεk[\mathbf{\Phi}(k-1:0)]_{ii}\geq\varepsilon^{k}. Therefore, it follows that

[𝚽(k1:0) 1]i\displaystyle\left[\mathbf{\Phi}(k-1:0)\,\mathbf{1}\right]_{i} [𝚽(k1:0)]ii\displaystyle\geq[\mathbf{\Phi}(k-1:0)]_{ii} (15)
εkεT(N1)\displaystyle\geq\varepsilon^{k}\geq\varepsilon^{T(N-1)}

is true for i=1,,Ni=1,\ldots,N and 1kT(N1)1\leq k\leq T(N-1), implying δ(k)εT(N1)\delta(k)\geq\varepsilon^{T(N-1)} for 1kT(N1)1\leq k\leq T(N-1).

Part 2: δ(k)εT(N1)\delta(k)\geq\varepsilon^{T(N-1)} for kT(N1)+1k\geq T(N-1)+1. Under Assumptions 1 and 2, and the requirements on weights pij(k)p_{ij}(k) in Algorithm 1, and following the arguments in Lemma 2 in [55], we can obtain

[𝚽(k1:kT(N1))]ijεT(N1)[\mathbf{\Phi}(k-1:k-T(N-1))]_{ij}\geq\varepsilon^{T(N-1)} (16)

for 1i,jN1\leq i,\,j\leq N. Since kT(N1)+1k\geq T(N-1)+1 holds and 𝐏(k)\mathbf{P}(k) is a column-stochastic matrix,

𝚽(kT(N1)1:0)=𝐏(kT(N1)1)𝐏(0)\mathbf{\Phi}(k-T(N-1)-1:0)=\mathbf{P}(k-T(N-1)-1)\cdots\mathbf{P}(0) (17)

should also be a column-stochastic matrix. Further using the fact 𝚽(k1:0)=𝚽(k1:kT(N1))𝚽(kT(N1)1:0)\mathbf{\Phi}(k-1:0)=\mathbf{\Phi}(k-1:k-T(N-1))\mathbf{\Phi}(k-T(N-1)-1:0) leads to [𝚽(k1:0)]ijεT(N1)[\mathbf{\Phi}(k-1:0)]_{ij}\geq\varepsilon^{T(N-1)} for 1i,jN1\leq i,\,j\leq N. Therefore, we have

[𝚽(k1:0) 1]iNεT(N1)εT(N1)[\mathbf{\Phi}(k-1:0)\,\mathbf{1}]_{i}\geq N\varepsilon^{T(N-1)}\geq\varepsilon^{T(N-1)} (18)

for i=1,,Ni=1,\ldots,N, meaning δ(k)εT(N1)\delta(k)\geq\varepsilon^{T(N-1)} for kT(N1)+1k\geq T(N-1)+1.

Based on δ(k)εT(N1)\delta(k)\geq\varepsilon^{T(N-1)} for k1k\geq 1, we can obtain wi(k)εT(N1)w_{i}(k)\geq\varepsilon^{T(N-1)} for k1k\geq 1. In summary, we always have wi(k)εT(N1)w_{i}(k)\geq\varepsilon^{T(N-1)} for k1k\geq 1. \blacksquare

Theorem 1

For a network of NN agents represented by a sequence of time-varying directed graphs {𝒢(k)=(𝒱,(k))}\{\mathcal{G}(k)=(\mathcal{V},\,\mathcal{E}(k))\} satisfying Assumptions 1, 2, 3, and 4, under Algorithm 1, the estimate πi(k)=baN2[N×frac(Nsi(k)wi(k))1]+a\pi_{i}(k)=\frac{b-a}{N-2}[N\times{\rm frac}(\frac{Ns_{i}(k)}{w_{i}(k)})-1]+a of each agent ii will converge to the average x¯0=j=1Nxj0/N\bar{x}^{0}={\sum_{j=1}^{N}x_{j}^{0}}/N. More specifically, the rate of convergence of Algorithm 1 is at least γ=(1εT(N1))1T(N1)(0, 1)\gamma=(1-\varepsilon^{T(N-1)})^{\frac{1}{T(N-1)}}\in(0,\,1), meaning that there exists a positive constant value CC satisfying 𝛑(k)x¯0𝟏Cγk\big{\|}\boldsymbol{\pi}(k)-\bar{x}^{0}\mathbf{1}\big{\|}\leq C\gamma^{k} for all kk.

Proof: From (6), we have

frac(i=1Nsi(k+1))\displaystyle{\rm frac}\Big{(}\sum_{i=1}^{N}s_{i}(k+1)\Big{)} (19)
=\displaystyle= frac(i=1Nfrac(j𝒩iin(k){i}Δsij(k)))\displaystyle{\rm frac}\bigg{(}\sum_{i=1}^{N}{\rm frac}\Big{(}\sum_{j\in\mathcal{N}_{i}^{in}(k)\cup\{i\}}\Delta s_{ij}(k)\Big{)}\bigg{)}
=\displaystyle= frac(i=1Nj𝒩iin(k)Δsij(k)+i=1NΔsii(k))\displaystyle{\rm frac}\bigg{(}\sum_{i=1}^{N}\sum_{j\in\mathcal{N}_{i}^{in}(k)}\Delta s_{ij}(k)+\sum_{i=1}^{N}\Delta s_{ii}(k)\bigg{)}
=\displaystyle= frac(i=1Nj𝒩iin(k)Δsij(k)\displaystyle{\rm frac}\bigg{(}\sum_{i=1}^{N}\sum_{j\in\mathcal{N}_{i}^{in}(k)}\Delta s_{ij}(k)
+i=1Nfrac(si(k)j𝒩iout(k)Δsji(k)))\displaystyle\qquad\quad+\sum_{i=1}^{N}{\rm frac}\Big{(}s_{i}(k)-\sum_{j\in\mathcal{N}_{i}^{out}(k)}\Delta s_{ji}(k)\Big{)}\bigg{)}
=\displaystyle= frac(i=1Nsi(k))\displaystyle{\rm frac}\Big{(}\sum_{i=1}^{N}s_{i}(k)\Big{)}

for kKk\leq K where in the derivation we used the property frac(x+y)=frac(x+frac(y))=frac(frac(x)+y)=frac(frac(x)+frac(y)){\rm frac}(x+y)={\rm frac}(x+{\rm frac}(y))={\rm frac}({\rm frac}(x)+y)={\rm frac}({\rm frac}(x)+{\rm frac}(y)) for any x,yx,\,y\in\mathbb{R}. According to Assumption 4, xi0[a,b]x_{i}^{0}\in[a,\,b] holds for each agent ii. Given si(0)=1N2+(N2)(xi0a)(ba)N2s_{i}(0)=\frac{1}{N^{2}}+\frac{(N-2)(x_{i}^{0}-a)}{(b-a)N^{2}}, one can obtain si(0)[1N2,N1N2]s_{i}(0)\in[\frac{1}{N^{2}},\,\frac{N-1}{N^{2}}], leading to

frac(i=1Nsi(0))=i=1Nsi(0)[1N,N1N](0, 1){\rm frac}\big{(}\sum_{i=1}^{N}s_{i}(0)\big{)}=\sum_{i=1}^{N}s_{i}(0)\in[\frac{1}{N},\,\frac{N-1}{N}]\subset(0,\,1) (20)

Further combining (19) and (20) yields

frac(i=1Nsi(K+1))==frac(i=1Nsi(0))=i=1Nsi(0){\rm frac}\big{(}\sum_{i=1}^{N}s_{i}(K+1)\big{)}=\cdots={\rm frac}\big{(}\sum_{i=1}^{N}s_{i}(0)\big{)}=\sum_{i=1}^{N}s_{i}(0) (21)

Since 𝐏(k)\mathbf{P}(k) is column stochastic, from (9) we have

𝟏T𝐰(k+1)=𝟏T𝐰(k)==𝟏T𝐰(0)=N\displaystyle\mathbf{1}^{T}\mathbf{w}(k+1)=\mathbf{1}^{T}\mathbf{w}(k)=\cdots=\mathbf{1}^{T}\mathbf{w}(0)=N (22)

for k0k\geq 0. Then we rewrite (11) as

{𝐬(K+l+1)=𝚽(K+l:K+1)𝐬(K+1)𝐰(K+l+1)=𝚽(K+l:K+1)𝐰(K+1)\left\{\begin{aligned} &\mathbf{s}(K+l+1)=\mathbf{\Phi}(K+l:K+1)\mathbf{s}(K+1)\\ &\mathbf{w}(K+l+1)=\mathbf{\Phi}(K+l:K+1)\mathbf{w}(K+1)\\ \end{aligned}\right. (23)

for l1l\geq 1. Under Assumptions 1 and 2, and the requirements on weights pij(k)p_{ij}(k) in Algorithm 1, following Proposition 1(b) in [56], we know that the transition matrix 𝚽(K+l:K+1)\mathbf{\Phi}(K+l:K+1) will converge to a stochastic vector 𝝋(K+l)\boldsymbol{\varphi}(K+l) with a geometric rate for all ii and jj, i.e., for all i,j=1,,Ni,j=1,\ldots,N and l1l\geq 1, we have

|[𝚽(K+l:K+1)]ijφi(K+l)|C0γl1\displaystyle\big{|}[\mathbf{\Phi}(K+l:K+1)]_{ij}-\varphi_{i}(K+l)\big{|}\leq C_{0}\gamma^{l-1} (24)

with C0=2(1+εT(N1))/(1εT(N1))C_{0}=2({1+\varepsilon^{-T(N-1)}})/({1-\varepsilon^{T(N-1)}}) and γ=(1εT(N1))1T(N1)\gamma=(1-\varepsilon^{T(N-1)})^{\frac{1}{T(N-1)}}. Defining 𝐌(K+l:K+1)\mathbf{M}(K+l:K+1) as

𝐌(K+l:K+1)𝚽(K+l:K+1)𝝋(K+l) 1T\displaystyle\mathbf{M}(K+l:K+1)\triangleq\mathbf{\Phi}(K+l:K+1)-\boldsymbol{\varphi}(K+l)\,\mathbf{1}^{T} (25)

we have

|[𝐌(K+l:K+1)]ij|C0γl1\displaystyle\left|\left[\mathbf{M}(K+l:K+1)\right]_{ij}\right|\leq C_{0}\gamma^{l-1} (26)

for all i,j=1,,Ni,j=1,\ldots,N and l1l\geq 1. Further combining (25) with (23) leads to

{𝐬(K+l+1)=𝐌(K+l:K+1)𝐬(K+1)+𝝋(K+l) 1T𝐬(K+1)𝐰(K+l+1)=𝐌(K+l:K+1)𝐰(K+1)+N𝝋(K+l)\left\{\begin{aligned} \mathbf{s}(K+l+1)=&\mathbf{M}(K+l:K+1)\mathbf{s}(K+1)\\ &+\,\boldsymbol{\varphi}(K+l)\,\mathbf{1}^{T}\mathbf{s}(K+1)\\ \mathbf{w}(K+l+1)=&\mathbf{M}(K+l:K+1)\mathbf{w}(K+1)\\ &+\,N\boldsymbol{\varphi}(K+l)\\ \end{aligned}\right. (27)

where in the derivation we used 𝟏T𝐰(K+1)=N\mathbf{1}^{T}\mathbf{w}(K+1)=N from (22). Then from (27), we have

si(K+l+1)wi(K+l+1)j=1Nsj(K+1)N\displaystyle\frac{s_{i}(K+l+1)}{w_{i}(K+l+1)}-\frac{\sum_{j=1}^{N}s_{j}(K+1)}{N} (28)
=\displaystyle= si(K+l+1)wi(K+l+1)𝟏T𝐬(K+1)N\displaystyle\frac{s_{i}(K+l+1)}{w_{i}(K+l+1)}-\frac{\mathbf{1}^{T}\mathbf{s}(K+1)}{N}
=\displaystyle= si(K+l+1)wi(K+l+1)𝟏T𝐬(K+1)wi(K+l+1)Nwi(K+l+1)\displaystyle\frac{s_{i}(K+l+1)}{w_{i}(K+l+1)}-\frac{\mathbf{1}^{T}\mathbf{s}(K+1)w_{i}(K+l+1)}{Nw_{i}(K+l+1)}
=\displaystyle= [𝐌(K+l:K+1)𝐬(K+1)]i+φi(K+l)𝟏T𝐬(K+1)wi(K+l+1)\displaystyle\frac{[\mathbf{M}(K+l:K+1)\mathbf{s}(K+1)]_{i}+\varphi_{i}(K+l)\mathbf{1}^{T}\mathbf{s}(K+1)}{w_{i}(K+l+1)}
𝟏T𝐬(K+1)[𝐌(K+l:K+1)𝐰(K+1)]iNwi(K+l+1)\displaystyle\,-\frac{\mathbf{1}^{T}\mathbf{s}(K+1)[\mathbf{M}(K+l:K+1)\mathbf{w}(K+1)]_{i}}{Nw_{i}(K+l+1)}
𝟏T𝐬(K+1)Nφi(K+l)Nwi(K+l+1)\displaystyle\,-\frac{\mathbf{1}^{T}\mathbf{s}(K+1)N\varphi_{i}(K+l)}{Nw_{i}(K+l+1)}
=\displaystyle= [𝐌(K+l:K+1)𝐬(K+1)]iwi(K+l+1)\displaystyle\frac{[\mathbf{M}(K+l:K+1)\mathbf{s}(K+1)]_{i}}{w_{i}(K+l+1)}
𝟏T𝐬(K+1)[𝐌(K+l:K+1)𝐰(K+1)]iNwi(K+l+1)\displaystyle\,-\frac{\mathbf{1}^{T}\mathbf{s}(K+1)[\mathbf{M}(K+l:K+1)\mathbf{w}(K+1)]_{i}}{Nw_{i}(K+l+1)}

Therefore, for i=1,,Ni=1,\ldots,N and l1l\geq 1, we can obtain

|Nsi(K+l+1)wi(K+l+1)j=1Nsj(K+1)|\displaystyle\big{|}N\frac{s_{i}(K+l+1)}{w_{i}(K+l+1)}-\sum_{j=1}^{N}s_{j}(K+1)\big{|} (29)
\displaystyle\leq N|[𝐌(K+l:K+1)𝐬(K+1)]i|wi(K+l+1)\displaystyle\frac{N\big{|}[\mathbf{M}(K+l:K+1)\mathbf{s}(K+1)]_{i}\big{|}}{w_{i}(K+l+1)}
+N|𝟏T𝐬(K+1)[𝐌(K+l:K+1)𝐰(K+1)]i|Nwi(K+l+1)\displaystyle\,+\frac{N\big{|}\mathbf{1}^{T}\mathbf{s}(K+1)[\mathbf{M}(K+l:K+1)\mathbf{w}(K+1)]_{i}\big{|}}{Nw_{i}(K+l+1)}
\displaystyle\leq NεT(N1)(maxj|[𝐌(K+l:K+1)]ij|)𝐬(K+1)1\displaystyle\frac{N}{\varepsilon^{T(N-1)}}\big{(}\max_{j}\big{|}[\mathbf{M}(K+l:K+1)]_{ij}\big{|}\big{)}\big{\|}\mathbf{s}(K+1)\big{\|}_{1}
+NεT(N1)|𝟏T𝐬(K+1)|(maxj|[𝐌(K+l:K+1)]ij|)\displaystyle\,+\frac{N}{\varepsilon^{T(N-1)}}\big{|}\mathbf{1}^{T}\mathbf{s}(K+1)\big{|}\big{(}\max_{j}\big{|}[\mathbf{M}(K+l:K+1)]_{ij}\big{|}\big{)}

where in the derivation we used wi(K+l+1)εT(N1)w_{i}(K+l+1)\geq\varepsilon^{T(N-1)} from Lemma 1 and 𝐰(K+1)1=i=1N|wi(K+1)|=𝟏T𝐰(K+1)=N\big{\|}\mathbf{w}(K+1)\big{\|}_{1}=\sum_{i=1}^{N}|w_{i}(K+1)|=\mathbf{1}^{T}\mathbf{w}(K+1)=N from (22). Further using the relationship |𝟏T𝐬(K+1)|𝐬(K+1)1\big{|}\mathbf{1}^{T}\mathbf{s}(K+1)\big{|}\leq\big{\|}\mathbf{s}(K+1)\big{\|}_{1} and (26) yields

|Nsi(k)wi(k)j=1Nsj(K+1)|C1γk\displaystyle\big{|}N\frac{s_{i}(k)}{w_{i}(k)}-\sum_{j=1}^{N}s_{j}(K+1)\big{|}\leq C_{1}\gamma^{k} (30)

for kK+2k\geq K+2 with C1C_{1} given by

C1=2NC0𝐬(K+1)1εT(N1)γK2\displaystyle C_{1}=2NC_{0}\big{\|}\mathbf{s}(K+1)\big{\|}_{1}\varepsilon^{-T(N-1)}\gamma^{-K-2} (31)

Given si(0)=1N2+(N2)(xi0a)(ba)N2s_{i}(0)=\frac{1}{N^{2}}+\frac{(N-2)(x_{i}^{0}-a)}{(b-a)N^{2}}, we have

x¯0\displaystyle\bar{x}^{0} =1Nj=1Nxj0=baN2(Nj=1Nsj(0)1)+a\displaystyle=\frac{1}{N}\sum_{j=1}^{N}x_{j}^{0}=\frac{b-a}{N-2}\Big{(}N\sum_{j=1}^{N}s_{j}(0)-1\Big{)}+a (32)
=baN2(N×frac(j=1Nsj(K+1))1)+a\displaystyle=\frac{b-a}{N-2}\Big{(}N\times{\rm frac}\big{(}\sum_{j=1}^{N}s_{j}(K+1)\big{)}-1\Big{)}+a

where in the derivation we used frac(j=1Nsj(K+1))=j=1Nsj(0){\rm frac}\big{(}\sum_{j=1}^{N}s_{j}(K+1)\big{)}=\sum_{j=1}^{N}s_{j}(0) from (21). Combining (32) with πi(k)=baN2[N×frac(Nsi(k)wi(k))1]+a\pi_{i}(k)=\frac{b-a}{N-2}[N\times{\rm frac}(\frac{Ns_{i}(k)}{w_{i}(k)})-1]+a leads to

πi(k)x¯0\displaystyle\pi_{i}(k)-\bar{x}^{0} (33)
=\displaystyle= baN2N(frac(Nsi(k)wi(k))frac(j=1Nsj(K+1)))\displaystyle\frac{b-a}{N-2}N\Big{(}{\rm frac}\big{(}\frac{Ns_{i}(k)}{w_{i}(k)}\big{)}-{\rm frac}\big{(}\sum_{j=1}^{N}s_{j}(K+1)\big{)}\Big{)}

From (20) and (21), one can obtain frac(j=1Nsj(K+1))=j=1Nsj(0)[1N,N1N]{\rm frac}\big{(}\sum_{j=1}^{N}s_{j}(K+1)\big{)}=\sum_{j=1}^{N}s_{j}(0)\in[\frac{1}{N},\,\frac{N-1}{N}]. Defining η\eta as ηj=1Nsj(0)[1N,N1N]\eta\triangleq\sum_{j=1}^{N}s_{j}(0)\in[\frac{1}{N},\,\frac{N-1}{N}], then there must exist an integer QQ such that j=1Nsj(K+1)=Q+η\sum_{j=1}^{N}s_{j}(K+1)=Q+\eta holds. From (30), we can see that there must exist a positive integer K1K+2K_{1}\geq K+2 such that

|Nsi(k)wi(k)j=1Nsj(K+1)|C1γk<1N\displaystyle\big{|}N\frac{s_{i}(k)}{w_{i}(k)}-\sum_{j=1}^{N}s_{j}(K+1)\big{|}\leq C_{1}\gamma^{k}<\frac{1}{N} (34)

holds for kK1k\geq K_{1}. Then it follows naturally one has

Q<Q+ηC1γkNsi(k)wi(k)Q+η+C1γk<Q+1\displaystyle Q<Q+\eta-C_{1}\gamma^{k}\leq N\frac{s_{i}(k)}{w_{i}(k)}\leq Q+\eta+C_{1}\gamma^{k}<Q+1 (35)

for kK1k\geq K_{1}, which leads to

0<ηC1γkfrac(Nsi(k)wi(k))η+C1γk<1\displaystyle 0<\eta-C_{1}\gamma^{k}\leq{\rm frac}\big{(}N\frac{s_{i}(k)}{w_{i}(k)}\big{)}\leq\eta+C_{1}\gamma^{k}<1 (36)

for kK1k\geq K_{1}. Thus, we have

|πi(k)x¯0|\displaystyle\big{|}\pi_{i}(k)-\bar{x}^{0}\big{|} =baN2N|frac(Nsi(k)wi(k))η|\displaystyle=\frac{b-a}{N-2}N\big{|}{\rm frac}\big{(}\frac{Ns_{i}(k)}{w_{i}(k)}\big{)}-\eta\big{|} (37)
baN2NC1γk\displaystyle\leq\frac{b-a}{N-2}NC_{1}\gamma^{k}

and further

𝝅(k)x¯0𝟏C2γk\displaystyle\big{\|}\boldsymbol{\pi}(k)-\bar{x}^{0}\mathbf{1}\big{\|}\leq C_{2}\gamma^{k} (38)

for kK1k\geq K_{1} with C2baN2NNC1C_{2}\triangleq\frac{b-a}{N-2}N\sqrt{N}C_{1}.

For kK11k\leq K_{1}-1, from (33), we have

|πi(k)x¯0|<baN2N\displaystyle|\pi_{i}(k)-\bar{x}^{0}|<\frac{b-a}{N-2}N (39)

which further implies

𝝅(k)x¯0𝟏<baN2NN\displaystyle\|\boldsymbol{\pi}(k)-\bar{x}^{0}\mathbf{1}\|<\frac{b-a}{N-2}N\sqrt{N} (40)

for kK11k\leq K_{1}-1. Defining CC as

Cmax{C2,baN2NNγk| 0kK11}\displaystyle C\triangleq\max\big{\{}C_{2},\frac{b-a}{N-2}N\sqrt{N}\gamma^{-k}\,\big{|}\,0\leq k\leq K_{1}-1\big{\}} (41)

one can obtain

𝝅(k)x¯0𝟏Cγk\displaystyle\big{\|}\boldsymbol{\pi}(k)-\bar{x}^{0}\mathbf{1}\big{\|}\leq C\gamma^{k} (42)

for all kk. Therefore, each agent ii will converge to the average value x¯0=j=1Nxj0/N\bar{x}^{0}={\sum_{j=1}^{N}x_{j}^{0}}/N with the rate of convergence of at least γ=(1εT(N1))1T(N1)(0, 1)\gamma=(1-\varepsilon^{T(N-1)})^{\frac{1}{T(N-1)}}\in(0,\,1). \blacksquare

From Theorem 1, we can see that a smaller γ\gamma means a faster convergence. Under the relationship γ=(1εT(N1))1T(N1)\gamma=(1-\varepsilon^{T(N-1)})^{\frac{1}{T(N-1)}}, to expedite the convergence, i.e., a smaller γ\gamma, it is sufficient to increase ε\varepsilon, which amounts to reducing the width of the range (ε, 1)(\varepsilon,\,1) for the random selection of pji(k)p_{ji}(k). Note that although a reduced range (ε, 1)(\varepsilon,\,1) enables an honest-but-curious adversary to obtain a better estimation of the range of agent ii’s intermediate states si(k)s_{i}(k) and wi(k)w_{i}(k) for kK+1k\geq K+1 from received pji(k)si(k)p_{ji}(k)s_{i}(k) and pji(k)wi(k)p_{ji}(k)w_{i}(k), it does not affect the confidentiality of agent ii’s initial state xi0x_{i}^{0}, as will be shown in the following subsection. It is also worth noting that to meet the requirement of randomly selecting weights in our algorithm, ε\varepsilon cannot be arbitrarily close to 11. In fact, ε\varepsilon must satisfy ε<1/maxi,k(Diout(k)+1)\varepsilon<1/\max_{i,k}({D_{i}^{out}(k)}+1). An easy way to select ε\varepsilon is to set 0<ε<1/N0<\varepsilon<1/N since Diout(k)N1D_{i}^{out}(k)\leq N-1 is true for all kk and i𝒱i\in\mathcal{V}.

Remark 6

Theorem 1 provides a detailed analysis of the rate of convergence under time-varying directed graphs, the results on which are sparse in the literature on Push-Sum under time-varying random weights.

III-C Confidentiality Analysis

Before presenting our main confidentiality result, we first introduce the following lemma.

Lemma 2

Consider a network of NN agents represented by a sequence of time-varying directed graphs {𝒢(k)=(𝒱,(k))}\{\mathcal{G}(k)=(\mathcal{V},\,\mathcal{E}(k))\} which satisfy Assumptions 1, 2, 3, 4, and 5. Under the proposed Algorithm 1, if at some time instant 0kK0\leq k^{*}\leq K, agent i𝒜i\notin\mathcal{A} has an in-neighbor or out-neighbor ll not belonging to 𝒜\mathcal{A}, then under s{Δsmn(k)|(m,n)(k),(m,n)(i,l),(m,n)(l,i),k=0,1,,K}\mathcal{I}_{s}^{*}\triangleq\{\Delta s_{mn}(k)\,|\,(m,\,n)\in\mathcal{E}(k),(m,\,n)\neq(i,\,l),(m,\,n)\neq(l,\,i),k=0,1,\ldots,K\}, the joint probability distributions of si(K+1)s_{i}(K+1) and sl(K+1)s_{l}(K+1) are identical for any two different initial states 𝐱0,𝐱~0[a,b]N\mathbf{x}^{0},\tilde{\mathbf{x}}^{0}\in[a,\,b]^{N} subject to xi0+xl0=x~i0+x~l0x_{i}^{0}+x_{l}^{0}=\tilde{x}_{i}^{0}+\tilde{x}_{l}^{0} and xj0=x~j0x_{j}^{0}=\tilde{x}_{j}^{0} for j𝒱{i,l}j\in\mathcal{V}\setminus\{i,l\}.

Proof: According to (6), we can rewrite the update rule of si(k)s_{i}(k) as follows

si(k+1)\displaystyle s_{i}(k+1) (43)
=\displaystyle= frac(j𝒩iin(k)Δsij(k)+si(k)j𝒩iout(k)Δsji(k))\displaystyle{\rm frac}\Big{(}\sum_{j\in\mathcal{N}_{i}^{in}(k)}\Delta s_{ij}(k)+s_{i}(k)-\sum_{j\in\mathcal{N}_{i}^{out}(k)}\Delta s_{ji}(k)\Big{)}

for kKk\leq K. We stack all variables Δsmn(k)\Delta s_{mn}(k) into a vector 𝚫𝐬(k)\boldsymbol{\Delta}\mathbf{s}(k) according to indices of edges in (k)\mathcal{E}(k), namely, the index of Δsmn(k)\Delta s_{mn}(k) is determined by the index of the edge (m,n)(m,\,n) in (k)\mathcal{E}(k). Then we can further rewrite (43) in the following more compact form:

𝐬(k+1)=frac(𝐬(k)+𝐂(k)𝚫𝐬(k))\displaystyle\mathbf{s}(k+1)={\rm frac}\big{(}\mathbf{s}(k)+\mathbf{C}(k)\boldsymbol{\Delta}\mathbf{s}(k)\big{)} (44)

for kKk\leq K, where 𝐂(k)\mathbf{C}(k) is the incidence matrix for graph 𝒢(k)\mathcal{G}(k) at iteration kk.

Define the subsets of agents \mathcal{H} and \mathcal{R} as ={i,l}\mathcal{H}=\{i,l\} and =𝒱(𝒜)\mathcal{R}=\mathcal{V}\setminus(\mathcal{H}\cup\mathcal{A}), respectively. Let N𝒜=|𝒜|N_{\mathcal{A}}=|\mathcal{A}| and N=||N_{\mathcal{R}}=|\mathcal{R}| represent the number of agents in 𝒜\mathcal{A} and \mathcal{R}, respectively. It is clear that \mathcal{H}, 𝒜\mathcal{A}, and \mathcal{R} are disjoint, and 𝒜=𝒱\mathcal{H}\cup\mathcal{A}\cup\mathcal{R}=\mathcal{V} holds. For graph 𝒢(k)=(𝒱,(k))\mathcal{G}(k)=(\mathcal{V},\mathcal{E}(k)), we define the subgraph 𝒢(k)\mathcal{G_{H}}(k) as 𝒢(k)=(,(k))\mathcal{G_{H}}(k)=(\mathcal{H},\mathcal{E_{H}}(k)) where (k)(k)\mathcal{E_{H}}(k)\subset\mathcal{E}(k) is the set of edges entirely within \mathcal{H}. The union of subgraphs 𝒢(k)\mathcal{G_{H}}(k) for iterations k=0,1,,Kk=0,1,\ldots,K is further denoted as 𝒢=k=0K𝒢(k)=(,)\mathcal{G}_{\mathcal{H}}^{*}=\bigcup\limits_{k=0}^{K}\mathcal{G_{H}}(k)=(\mathcal{H},\mathcal{E}_{\mathcal{H}}^{*}), where =k=0K(k)\mathcal{E}_{\mathcal{H}}^{*}=\bigcup\limits_{k=0}^{K}\mathcal{E_{H}}(k) is the union of edge sets (k)\mathcal{E_{H}}(k) for iterations k=0,1,,Kk=0,1,\ldots,K. Denote the edge set 𝒜(k)\mathcal{E_{A}}(k) as the collection of edges associated with all agents in 𝒜\mathcal{A} at iteration kk, i.e., 𝒜(k)={(m,n)|(m,n)(k),morn𝒜}\mathcal{E_{A}}(k)=\{(m,\,n)\,|\,(m,\,n)\in\mathcal{E}(k),m\ \textnormal{or}\ n\in\mathcal{A}\}. Then the set of remaining edges not belonging to (k)\mathcal{E_{H}}(k) or 𝒜(k)\mathcal{E_{A}}(k) can be denoted as (k)=(k)((k)𝒜(k))\mathcal{E_{R}}(k)=\mathcal{E}(k)\setminus(\mathcal{E_{H}}(k)\cup\mathcal{E_{A}}(k)), which is a collection of edges that are either entirely within \mathcal{R} or between \mathcal{R} and \mathcal{H}. Let E(k)=|(k)|E_{\mathcal{H}}(k)=|\mathcal{E_{H}}(k)|, E𝒜(k)=|𝒜(k)|E_{\mathcal{A}}(k)=|\mathcal{E_{A}}(k)|, and E(k)=|(k)|E_{\mathcal{R}}(k)=|\mathcal{E_{R}}(k)| represent the number of edges in (k)\mathcal{E_{H}}(k), 𝒜(k)\mathcal{E_{A}}(k), and (k)\mathcal{E_{R}}(k), respectively. Note that (k)\mathcal{E_{H}}(k), 𝒜(k)\mathcal{E_{A}}(k), and (k)\mathcal{E_{R}}(k) are disjoint, and we always have (k)𝒜(k)(k)=(k)\mathcal{E_{H}}(k)\cup\mathcal{E_{A}}(k)\cup\mathcal{E_{R}}(k)=\mathcal{E}(k). Without loss of generality, we can partition the state vector 𝐬(k)\mathbf{s}(k) as 𝐬(k)=[𝐬(k)T𝐬𝒜(k)T𝐬(k)T]T\mathbf{s}(k)=[\mathbf{s}_{\mathcal{H}}(k)^{T}\ \mathbf{s}_{\mathcal{A}}(k)^{T}\ \mathbf{s}_{\mathcal{R}}(k)^{T}]^{T} where 𝐬(k)\mathbf{s}_{\mathcal{H}}(k), 𝐬𝒜(k)\mathbf{s}_{\mathcal{A}}(k), and 𝐬(k)\mathbf{s}_{\mathcal{R}}(k) denote the state vectors of agents in \mathcal{H}, 𝒜\mathcal{A}, and \mathcal{R}, respectively. Then we can further rewrite the update rule of 𝐬(k)\mathbf{s}(k) for kKk\leq K as

[𝐬(k+1)𝐬𝒜(k+1)𝐬(k+1)]=frac([𝐬(k)𝐬𝒜(k)𝐬(k)]\displaystyle\begin{bmatrix}\mathbf{s}_{\mathcal{H}}(k+1)\\ \mathbf{s}_{\mathcal{A}}(k+1)\\ \mathbf{s}_{\mathcal{R}}(k+1)\end{bmatrix}={\rm frac}\Bigg{(}\begin{bmatrix}\mathbf{s}_{\mathcal{H}}(k)\\ \mathbf{s}_{\mathcal{A}}(k)\\ \mathbf{s}_{\mathcal{R}}(k)\end{bmatrix} (45)
+[𝐂(k)𝐂𝒜1(k)𝐂1(k)𝟎N𝒜×E(k)𝐂𝒜2(k)𝟎N𝒜×E(k)𝟎N×E(k)𝐂𝒜3(k)𝐂2(k)][𝚫𝐬(k)𝚫𝐬𝒜(k)𝚫𝐬(k)])\displaystyle\quad\ +\begin{bmatrix}\mathbf{C}_{\mathcal{H}}(k)&\mathbf{C}_{\mathcal{A}}^{1}(k)&\mathbf{C}_{\mathcal{R}}^{1}(k)\\ \mathbf{0}_{N_{\mathcal{A}}\times E_{\mathcal{H}}(k)}&\mathbf{C}_{\mathcal{A}}^{2}(k)&\mathbf{0}_{N_{\mathcal{A}}\times E_{\mathcal{R}}(k)}\\ \mathbf{0}_{N_{\mathcal{R}}\times E_{\mathcal{H}}(k)}&\mathbf{C}_{\mathcal{A}}^{3}(k)&\mathbf{C}_{\mathcal{R}}^{2}(k)\\ \end{bmatrix}\begin{bmatrix}\boldsymbol{\Delta}\mathbf{s}_{\mathcal{H}}(k)\\ \boldsymbol{\Delta}\mathbf{s}_{\mathcal{A}}(k)\\ \boldsymbol{\Delta}\mathbf{s}_{\mathcal{R}}(k)\end{bmatrix}\Bigg{)}

where 𝐂(k)\mathbf{C}_{\mathcal{H}}(k) is the incidence matrix for subgraph 𝒢(k)=(,(k))\mathcal{G_{H}}(k)=(\mathcal{H},\mathcal{E_{H}}(k)), and 𝚫𝐬(k)\boldsymbol{\Delta}\mathbf{s}_{\mathcal{H}}(k), 𝚫𝐬𝒜(k)\boldsymbol{\Delta}\mathbf{s}_{\mathcal{A}}(k), and 𝚫𝐬(k)\boldsymbol{\Delta}\mathbf{s}_{\mathcal{R}}(k) are vectors stacking Δsmn(k)\Delta s_{mn}(k) corresponding to edge sets (k)\mathcal{E_{H}}(k), 𝒜(k)\mathcal{E_{A}}(k), and (k)\mathcal{E_{R}}(k), respectively. It is obvious that 𝚫𝐬𝒜(k)\boldsymbol{\Delta}\mathbf{s}_{\mathcal{A}}(k) is completely known to agents in 𝒜\mathcal{A} since every element in 𝚫𝐬𝒜(k)\boldsymbol{\Delta}\mathbf{s}_{\mathcal{A}}(k) is either sent or received by the agents in 𝒜\mathcal{A}. From (45), we can further obtain

[𝐬(K+1)𝐬𝒜(K+1)𝐬(K+1)]=frac([𝐬(0)𝐬𝒜(0)𝐬(0)]\displaystyle\begin{bmatrix}\mathbf{s}_{\mathcal{H}}(K+1)\\ \mathbf{s}_{\mathcal{A}}(K+1)\\ \mathbf{s}_{\mathcal{R}}(K+1)\end{bmatrix}={\rm frac}\Bigg{(}\begin{bmatrix}\mathbf{s}_{\mathcal{H}}(0)\\ \mathbf{s}_{\mathcal{A}}(0)\\ \mathbf{s}_{\mathcal{R}}(0)\end{bmatrix} (46)
+[𝐂𝐂𝒜1𝐂1𝟎N𝒜×E𝐂𝒜2𝟎N𝒜×E𝟎N×E𝐂𝒜3𝐂2][𝚫𝐬𝚫𝐬𝒜𝚫𝐬])\displaystyle\qquad\qquad+\begin{bmatrix}\mathbf{C}_{\mathcal{H}}^{*}&\mathbf{C}_{\mathcal{A}}^{1*}&\mathbf{C}_{\mathcal{R}}^{1*}\\ \mathbf{0}_{N_{\mathcal{A}}\times E_{\mathcal{H}}^{*}}&\mathbf{C}_{\mathcal{A}}^{2*}&\mathbf{0}_{N_{\mathcal{A}}\times E_{\mathcal{R}}^{*}}\\ \mathbf{0}_{N_{\mathcal{R}}\times E_{\mathcal{H}}^{*}}&\mathbf{C}_{\mathcal{A}}^{3*}&\mathbf{C}_{\mathcal{R}}^{2*}\end{bmatrix}\begin{bmatrix}\boldsymbol{\Delta}\mathbf{s}_{\mathcal{H}}^{*}\\ \boldsymbol{\Delta}\mathbf{s}_{\mathcal{A}}^{*}\\ \boldsymbol{\Delta}\mathbf{s}_{\mathcal{R}}^{*}\end{bmatrix}\Bigg{)}

where E=k=0KE(k)E_{\mathcal{H}}^{*}=\sum_{k=0}^{K}E_{\mathcal{H}}(k), E=k=0KE(k)E_{\mathcal{R}}^{*}=\sum_{k=0}^{K}E_{\mathcal{R}}(k), and

𝐂\displaystyle\mathbf{C}_{\mathcal{H}}^{*} =[𝐂(0)𝐂(K)]\displaystyle=\begin{bmatrix}\mathbf{C}_{\mathcal{H}}(0)&\cdots&\mathbf{C}_{\mathcal{H}}(K)\\ \end{bmatrix} (47)
𝐂𝒜t\displaystyle\mathbf{C}_{\mathcal{A}}^{t*} =[𝐂𝒜t(0)𝐂𝒜t(K)],t=1,2,3\displaystyle=\begin{bmatrix}\mathbf{C}_{\mathcal{A}}^{t}(0)&\cdots&\mathbf{C}_{\mathcal{A}}^{t}(K)\\ \end{bmatrix},\ t=1,2,3
𝐂t\displaystyle\mathbf{C}_{\mathcal{R}}^{t*} =[𝐂t(0)𝐂t(K)],t=1,2\displaystyle=\begin{bmatrix}\mathbf{C}_{\mathcal{R}}^{t}(0)&\cdots&\mathbf{C}_{\mathcal{R}}^{t}(K)\\ \end{bmatrix},\ t=1,2
𝚫𝐬\displaystyle\boldsymbol{\Delta}\mathbf{s}_{\mathcal{H}}^{*} =[𝚫𝐬(0)T𝚫𝐬(K)T]T\displaystyle=\begin{bmatrix}\boldsymbol{\Delta}\mathbf{s}_{\mathcal{H}}(0)^{T}&\cdots&\boldsymbol{\Delta}\mathbf{s}_{\mathcal{H}}(K)^{T}\\ \end{bmatrix}^{T}
𝚫𝐬𝒜\displaystyle\boldsymbol{\Delta}\mathbf{s}_{\mathcal{A}}^{*} =[𝚫𝐬𝒜(0)T𝚫𝐬𝒜(K)T]T\displaystyle=\begin{bmatrix}\boldsymbol{\Delta}\mathbf{s}_{\mathcal{A}}(0)^{T}&\cdots&\boldsymbol{\Delta}\mathbf{s}_{\mathcal{A}}(K)^{T}\\ \end{bmatrix}^{T}
𝚫𝐬\displaystyle\boldsymbol{\Delta}\mathbf{s}_{\mathcal{R}}^{*} =[𝚫𝐬(0)T𝚫𝐬(K)T]T\displaystyle=\begin{bmatrix}\boldsymbol{\Delta}\mathbf{s}_{\mathcal{R}}(0)^{T}&\cdots&\boldsymbol{\Delta}\mathbf{s}_{\mathcal{R}}(K)^{T}\\ \end{bmatrix}^{T}

It is worth noting that 𝐂\mathbf{C}_{\mathcal{H}}^{*} is the incidence matrix for graph 𝒢=k=0K𝒢(k)\mathcal{G}_{\mathcal{H}}^{*}=\bigcup\limits_{k=0}^{K}\mathcal{G_{H}}(k). From (46), we have

𝐬(K+1)\displaystyle\mathbf{s}_{\mathcal{H}}(K+1) (48)
=\displaystyle= frac(𝐬(0)+𝐂𝚫𝐬+𝐂𝒜1𝚫𝐬𝒜+𝐂1𝚫𝐬)\displaystyle{\rm frac}\big{(}\mathbf{s}_{\mathcal{H}}(0)+\mathbf{C}_{\mathcal{H}}^{*}\boldsymbol{\Delta}\mathbf{s}_{\mathcal{H}}^{*}+\mathbf{C}_{\mathcal{A}}^{1*}\boldsymbol{\Delta}\mathbf{s}_{\mathcal{A}}^{*}+\mathbf{C}_{\mathcal{R}}^{1*}\boldsymbol{\Delta}\mathbf{s}_{\mathcal{R}}^{*}\big{)}

Denoting 𝐯=[v1v2]T\mathbf{v}=[v_{1}\,v_{2}]^{T} as

𝐯frac(𝐂𝚫𝐬)\displaystyle\mathbf{v}\triangleq{\rm frac}\big{(}\mathbf{C}_{\mathcal{H}}^{*}\boldsymbol{\Delta}\mathbf{s}_{\mathcal{H}}^{*}\big{)} (49)

next we prove that 𝐯\mathbf{v} is uniformly distributed in [0, 1)×[0, 1)[0,\,1)\times[0,\,1) subject to frac(v1+v2)=0{\rm frac}(v_{1}+v_{2})=0 if at some time instant 0kK0\leq k^{*}\leq K, agent ii has an in-neighbor or out-neighbor ll not belonging to 𝒜\mathcal{A}.

Given ={i,l}\mathcal{H}=\{i,l\}, if at some time instant 0kK0\leq k^{*}\leq K, agent ii has an in-neighbor or out-neighbor ll not belonging to 𝒜\mathcal{A}, the columns of 𝐂\mathbf{C}_{\mathcal{H}}^{*} are either [11]T[1\ -1]^{T} (ll is an in-neighbor of agent ii) or [1 1]T[-1\ 1]^{T} (ll is an out-neighbor of agent ii). Denote the jj-th column of 𝐂\mathbf{C}_{\mathcal{H}}^{*} as 𝐜j\mathbf{c}_{j}^{*}, then 𝐜j\mathbf{c}_{j}^{*} can be expressed as

𝐜j=rj𝐜1\displaystyle\mathbf{c}_{j}^{*}=r_{j}\mathbf{c}_{1}^{*} (50)

where rj{1,1}r_{j}\in\{1,-1\} is the corresponding coefficient. Defining 𝐫\mathbf{r} as 𝐫=[r1rE]\mathbf{r}=[r_{1}\cdots r_{E_{\mathcal{H}}^{*}}], one can obtain

𝐯\displaystyle\mathbf{v} =frac(𝐂𝚫𝐬)=frac(𝐜1𝐫𝚫𝐬)\displaystyle={\rm frac}\big{(}\mathbf{C}_{\mathcal{H}}^{*}\boldsymbol{\Delta}\mathbf{s}_{\mathcal{H}}^{*}\big{)}={\rm frac}\big{(}\mathbf{c}_{1}^{*}\mathbf{r}\boldsymbol{\Delta}\mathbf{s}_{\mathcal{H}}^{*}\big{)} (51)
=frac(𝐜1frac(𝐫𝚫𝐬))\displaystyle={\rm frac}\big{(}\mathbf{c}_{1}^{*}\cdot{\rm frac}(\mathbf{r}\boldsymbol{\Delta}\mathbf{s}_{\mathcal{H}}^{*})\big{)}

Since all the elements in 𝚫𝐬\boldsymbol{\Delta}\mathbf{s}_{\mathcal{H}}^{*} are independent of each other and uniformly distributed in [0, 1)[0,\,1), we have that frac(𝐫𝚫𝐬){\rm frac}(\mathbf{r}\boldsymbol{\Delta}\mathbf{s}_{\mathcal{H}}^{*}) is uniformly distributed in [0, 1)[0,\,1). As 𝐜1\mathbf{c}_{1}^{*} is either [11]T[1\ -1]^{T} or [1 1]T[-1\ 1]^{T}, we have that 𝐯\mathbf{v} is uniformly distributed in [0, 1)×[0, 1)[0,\,1)\times[0,\,1) subject to frac(v1+v2)=0{\rm frac}(v_{1}+v_{2})=0.

From (48), we have 𝐬(K+1)=frac(𝐬(0)+𝐯+𝐂𝒜1𝚫𝐬𝒜+𝐂1𝚫𝐬)\mathbf{s}_{\mathcal{H}}(K+1)={\rm frac}\big{(}\mathbf{s}_{\mathcal{H}}(0)+\mathbf{v}+\mathbf{C}_{\mathcal{A}}^{1*}\boldsymbol{\Delta}\mathbf{s}_{\mathcal{A}}^{*}+\mathbf{C}_{\mathcal{R}}^{1*}\boldsymbol{\Delta}\mathbf{s}_{\mathcal{R}}^{*}\big{)}. Further taking into account the fact that 𝐯\mathbf{v} is uniformly distributed in [0, 1)×[0, 1)[0,\,1)\times[0,\,1) subject to frac(v1+v2)=0{\rm frac}(v_{1}+v_{2})=0 if at some time instant 0kK0\leq k^{*}\leq K, agent ii has an in-neighbor or out-neighbor ll not belonging to 𝒜\mathcal{A}, we can obtain that 𝐬(K+1)\mathbf{s}_{\mathcal{H}}(K+1) is also uniformly distributed in [0, 1)×[0, 1)[0,\,1)\times[0,\,1) subject to frac(𝟏T𝐬(K+1))=frac(𝟏T𝐬(0)+𝟏T𝐂𝒜1𝚫𝐬𝒜+𝟏T𝐂1𝚫𝐬){\rm frac}(\mathbf{1}^{T}\mathbf{s}_{\mathcal{H}}(K+1))={\rm frac}(\mathbf{1}^{T}\mathbf{s}_{\mathcal{H}}(0)+\mathbf{1}^{T}\mathbf{C}_{\mathcal{A}}^{1*}\boldsymbol{\Delta}\mathbf{s}_{\mathcal{A}}^{*}+\mathbf{1}^{T}\mathbf{C}_{\mathcal{R}}^{1*}\boldsymbol{\Delta}\mathbf{s}_{\mathcal{R}}^{*}).

Now we analyze the probability distributions of 𝐬(K+1)\mathbf{s}_{\mathcal{H}}(K+1) under different initial conditions of agent ii. For any two different initial conditions 𝐱0,𝐱~0[a,b]N\mathbf{x}^{0},\,\tilde{\mathbf{x}}^{0}\in[a,\,b]^{N} subject to xi0+xl0=x~i0+x~l0x_{i}^{0}+x_{l}^{0}=\tilde{x}_{i}^{0}+\tilde{x}_{l}^{0} and xj0=x~j0x_{j}^{0}=\tilde{x}_{j}^{0} for j𝒱{i,l}j\in\mathcal{V}\setminus\{i,l\}, one can obtain si(0)+sl(0)=s~i(0)+s~l(0)s_{i}(0)+s_{l}(0)=\tilde{s}_{i}(0)+\tilde{s}_{l}(0). Note that all the elements in 𝚫𝐬𝒜\boldsymbol{\Delta}\mathbf{s}_{\mathcal{A}}^{*} and 𝚫𝐬\boldsymbol{\Delta}\mathbf{s}_{\mathcal{R}}^{*} belong to the set s={Δsmn(k)|(m,n)(k),(m,n)(i,l),(m,n)(l,i),k=0,1,,K}\mathcal{I}_{s}^{*}=\{\Delta s_{mn}(k)\,|\,(m,\,n)\in\mathcal{E}(k),(m,\,n)\neq(i,\,l),(m,\,n)\neq(l,\,i),k=0,1,\ldots,K\}. Thus, given s\mathcal{I}_{s}^{*}, both 𝐬(K+1)\mathbf{s}_{\mathcal{H}}(K+1) and 𝐬~(K+1)\tilde{\mathbf{s}}_{\mathcal{H}}(K+1) are uniformly distributed in [0, 1)×[0, 1)[0,\,1)\times[0,\,1) subject to frac(𝟏T𝐬(K+1))=frac(𝟏T𝐬(0)+𝟏T𝐂𝒜1𝚫𝐬𝒜+𝟏T𝐂1𝚫𝐬)=frac(𝟏T𝐬~(0)+𝟏T𝐂𝒜1𝚫𝐬𝒜+𝟏T𝐂1𝚫𝐬)=frac(𝟏T𝐬~(K+1)){\rm frac}(\mathbf{1}^{T}\mathbf{s}_{\mathcal{H}}(K+1))={\rm frac}(\mathbf{1}^{T}\mathbf{s}_{\mathcal{H}}(0)+\mathbf{1}^{T}\mathbf{C}_{\mathcal{A}}^{1*}\boldsymbol{\Delta}\mathbf{s}_{\mathcal{A}}^{*}+\mathbf{1}^{T}\mathbf{C}_{\mathcal{R}}^{1*}\boldsymbol{\Delta}\mathbf{s}_{\mathcal{R}}^{*})={\rm frac}(\mathbf{1}^{T}\tilde{\mathbf{s}}_{\mathcal{H}}(0)+\mathbf{1}^{T}\mathbf{C}_{\mathcal{A}}^{1*}\boldsymbol{\Delta}\mathbf{s}_{\mathcal{A}}^{*}+\mathbf{1}^{T}\mathbf{C}_{\mathcal{R}}^{1*}\boldsymbol{\Delta}\mathbf{s}_{\mathcal{R}}^{*})={\rm frac}(\mathbf{1}^{T}\tilde{\mathbf{s}}_{\mathcal{H}}(K+1)). Therefore, given s\mathcal{I}_{s}^{*}, the probability distributions of 𝐬(K+1)\mathbf{s}_{\mathcal{H}}(K+1) under different 𝐱0\mathbf{x}^{0} and 𝐱~0\tilde{\mathbf{x}}^{0} are identical when xj0=x~j0x_{j}^{0}=\tilde{x}_{j}^{0} for j𝒱{i,l}j\in\mathcal{V}\setminus\{i,l\} and xi0+xl0=x~i0+x~l0x_{i}^{0}+x_{l}^{0}=\tilde{x}_{i}^{0}+\tilde{x}_{l}^{0} hold for some agent ll that is an in-neighbor or out-neighbor of agent ii at some time instant 0kK0\leq k^{*}\leq K but does not belong to 𝒜\mathcal{A}. \blacksquare

Now we are in position to present our main confidentiality result.

Theorem 2

Consider a network of NN agents represented by a sequence of time-varying directed graphs {𝒢(k)=(𝒱,(k))}\{\mathcal{G}(k)=(\mathcal{V},\,\mathcal{E}(k))\} which satisfy Assumptions 1, 2, 3, 4, and 5. Under the proposed Algorithm 1, the confidentiality of agent i𝒜i\notin\mathcal{A} can be preserved against 𝒜\mathcal{A} if at some time instant 0kK0\leq k^{*}\leq K, agent ii has an in-neighbor or out-neighbor ll not belonging to 𝒜\mathcal{A}.

Proof: To show that the confidentiality of agent ii can be protected, we have to show that under different initial values of agent ii, the probability distributions are identical for honest-but-curious agents 𝒜\mathcal{A}’ information set. It suffices to prove that the probability distributions of information sets accessible to 𝒜\mathcal{A} are identical for all 𝐱0,𝐱~0[a,b]N\mathbf{x}^{0},\,\tilde{\mathbf{x}}^{0}\in[a,\,b]^{N} subject to xi0+xl0=x~i0+x~l0x_{i}^{0}+x_{l}^{0}=\tilde{x}_{i}^{0}+\tilde{x}_{l}^{0} and xj0=x~j0x_{j}^{0}=\tilde{x}_{j}^{0} for j𝒱{i,l}j\in\mathcal{V}\setminus\{i,l\} where agent ll is an in-neighbor or out-neighbor of agent ii at some time instant 0kK0\leq k^{*}\leq K and does not belong to 𝒜\mathcal{A}. Note that under such 𝐱0\mathbf{x}^{0} and 𝐱~0\tilde{\mathbf{x}}^{0}, the sum of the initial states of all nodes not in 𝒜\mathcal{A} keeps unchanged, i.e., j𝒱𝒜xj0=j𝒱𝒜x~j0\sum_{j\in\mathcal{V}\setminus\mathcal{A}}x_{j}^{0}=\sum_{j\in\mathcal{V}\setminus\mathcal{A}}\tilde{x}_{j}^{0}. Given si(0)=1N2+(N2)(xi0a)(ba)N2s_{i}(0)=\frac{1}{N^{2}}+\frac{(N-2)(x_{i}^{0}-a)}{(b-a)N^{2}}, this is equivalent to proving that the probability distributions of information sets accessible to 𝒜\mathcal{A} are identical for all 𝐬(0),𝐬~(0)[1N2,N1N2]N\mathbf{s}(0),\,\tilde{\mathbf{s}}(0)\in[\frac{1}{N^{2}},\,\frac{N-1}{N^{2}}]^{N} subject to si(0)+sl(0)=s~i(0)+s~l(0)s_{i}(0)+s_{l}(0)=\tilde{s}_{i}(0)+\tilde{s}_{l}(0) and sj(0)=s~j(0)s_{j}(0)=\tilde{s}_{j}(0) for j𝒱{i,l}j\in\mathcal{V}\setminus\{i,l\}. Denoting s(k)\mathcal{I}_{s}(k) and w(k)\mathcal{I}_{w}(k) as

s(k)\displaystyle\mathcal{I}_{s}(k) ={𝐬𝒜(k),𝚫𝐬𝒜(k)}\displaystyle=\{\mathbf{s}_{\mathcal{A}}(k),\boldsymbol{\Delta}\mathbf{s}_{\mathcal{A}}(k)\} (52)
w(k)\displaystyle\mathcal{I}_{w}(k) ={𝐰𝒜(k),𝚫𝐰𝒜(k)}\displaystyle=\{\mathbf{w}_{\mathcal{A}}(k),\boldsymbol{\Delta}\mathbf{w}_{\mathcal{A}}(k)\}

where 𝐬𝒜(k)\mathbf{s}_{\mathcal{A}}(k) and 𝐰𝒜(k)\mathbf{w}_{\mathcal{A}}(k) are state vectors of agents in 𝒜\mathcal{A}, and 𝚫𝐬𝒜(k)\boldsymbol{\Delta}\mathbf{s}_{\mathcal{A}}(k) and 𝚫𝐰𝒜(k)\boldsymbol{\Delta}\mathbf{w}_{\mathcal{A}}(k) are augmented vectors of Δsmn(k)\Delta s_{mn}(k) and Δwmn(k)\Delta w_{mn}(k) corresponding to edge set 𝒜(k)\mathcal{E_{A}}(k), respectively. We can see that agents in 𝒜\mathcal{A} have access to both s(k)\mathcal{I}_{s}(k) and w(k)\mathcal{I}_{w}(k) at each iteration kk. Further denote 1s\mathcal{I}_{1}^{s}, 1w\mathcal{I}_{1}^{w}, 2s\mathcal{I}_{2}^{s}, 2w\mathcal{I}_{2}^{w}, 1\mathcal{I}_{1}, and 2\mathcal{I}_{2} as

1s\displaystyle\mathcal{I}_{1}^{s} =k=0Ks(k)1w=k=0Kw(k){𝐰(0)}\displaystyle=\bigcup\limits_{k=0}^{K}\mathcal{I}_{s}(k)\qquad\quad\,\mathcal{I}_{1}^{w}=\bigcup\limits_{k=0}^{K}\mathcal{I}_{w}(k)\cup\{\mathbf{w}(0)\} (53)
2s\displaystyle\mathcal{I}_{2}^{s} =k=K+1s(k)2w=k=K+1w(k)\displaystyle=\bigcup\limits_{k=K+1}^{\infty}\mathcal{I}_{s}(k)\qquad\mathcal{I}_{2}^{w}=\bigcup\limits_{k=K+1}^{\infty}\mathcal{I}_{w}(k)
1\displaystyle\mathcal{I}_{1} =1s1w2=2s2w\displaystyle=\mathcal{I}_{1}^{s}\cup\mathcal{I}_{1}^{w}\qquad\quad\ \ \mathcal{I}_{2}=\mathcal{I}_{2}^{s}\cup\mathcal{I}_{2}^{w}

According to Algorithm 1, 12\mathcal{I}\triangleq\mathcal{I}_{1}\cup\mathcal{I}_{2} represents all the information accessible to 𝒜\mathcal{A}. We denote the conditional probability density of \mathcal{I} given 𝐬(0)\mathbf{s}(0) as f(|𝐬(0))f(\mathcal{I}\,|\,\mathbf{s}(0)). Thus, to show that the confidentiality of agent ii can be protected, it is equivalent to proving f(|𝐬(0))=f(|𝐬~(0))f(\mathcal{I}\,|\,\mathbf{s}(0))=f(\mathcal{I}\,|\,\tilde{\mathbf{s}}(0)) for all 𝐬(0),𝐬~(0)[1N2,N1N2]N\mathbf{s}(0),\,\tilde{\mathbf{s}}(0)\in[\frac{1}{N^{2}},\,\frac{N-1}{N^{2}}]^{N} subject to si(0)+sl(0)=s~i(0)+s~l(0)s_{i}(0)+s_{l}(0)=\tilde{s}_{i}(0)+\tilde{s}_{l}(0) and sj(0)=s~j(0)s_{j}(0)=\tilde{s}_{j}(0) for j𝒱{i,l}j\in\mathcal{V}\setminus\{i,l\}.

Since

f(|𝐬(0))=f(1,2|𝐬(0))=f(1|𝐬(0))f(2|1,𝐬(0))\displaystyle f(\mathcal{I}\,|\,\mathbf{s}(0))=f(\mathcal{I}_{1},\mathcal{I}_{2}\,|\,\mathbf{s}(0))=f(\mathcal{I}_{1}\,|\,\mathbf{s}(0))f(\mathcal{I}_{2}\,|\,\mathcal{I}_{1},\mathbf{s}(0)) (54)

holds, to show f(|𝐬(0))=f(|𝐬~(0))f(\mathcal{I}\,|\,\mathbf{s}(0))=f(\mathcal{I}\,|\,\tilde{\mathbf{s}}(0)), it suffices to prove that

f(1|𝐬(0))=f(1|𝐬~(0))\displaystyle f(\mathcal{I}_{1}\,|\,\mathbf{s}(0))=f(\mathcal{I}_{1}\,|\,\tilde{\mathbf{s}}(0)) (55)

and

f(2|1,𝐬(0))=f(2|1,𝐬~(0))\displaystyle f(\mathcal{I}_{2}\,|\,\mathcal{I}_{1},\mathbf{s}(0))=f(\mathcal{I}_{2}\,|\,\mathcal{I}_{1},\tilde{\mathbf{s}}(0)) (56)

hold for any 𝐬(0),𝐬~(0)[1N2,N1N2]N\mathbf{s}(0),\,\tilde{\mathbf{s}}(0)\in[\frac{1}{N^{2}},\,\frac{N-1}{N^{2}}]^{N} subject to si(0)+sl(0)=s~i(0)+s~l(0)s_{i}(0)+s_{l}(0)=\tilde{s}_{i}(0)+\tilde{s}_{l}(0) and sj(0)=s~j(0)s_{j}(0)=\tilde{s}_{j}(0) for j𝒱{i,l}j\in\mathcal{V}\setminus\{i,l\}.

We first show f(1|𝐬(0))=f(1|𝐬~(0)f(\mathcal{I}_{1}\,|\,\mathbf{s}(0))=f(\mathcal{I}_{1}\,|\,\tilde{\mathbf{s}}(0). Since 1w\mathcal{I}_{1}^{w} is independent of 1s\mathcal{I}_{1}^{s} and 𝐬(0)\mathbf{s}(0), one can obtain f(1|𝐬(0))=f(1s,1w|𝐬(0))=f(1w)f(1s|𝐬(0))f(\mathcal{I}_{1}\,|\,\mathbf{s}(0))=f(\mathcal{I}_{1}^{s},\mathcal{I}_{1}^{w}\,|\,\mathbf{s}(0))=f(\mathcal{I}_{1}^{w})f(\mathcal{I}_{1}^{s}\,|\,\mathbf{s}(0)). From (45), we can see that given 𝐬𝒜(0)\mathbf{s}_{\mathcal{A}}(0), 1s\mathcal{I}_{1}^{s} is independent of 𝐬(0)\mathbf{s}_{\mathcal{H}}(0) and 𝐬(0)\mathbf{s}_{\mathcal{R}}(0), where ={i,l}\mathcal{H}=\{i,l\} and =𝒱(𝒜)\mathcal{R}=\mathcal{V}\setminus(\mathcal{H}\cup\mathcal{A}). So f(1s|𝐬(0))=f(1s|𝐬𝒜(0),𝐬(0),𝐬(0))=f(1s|𝐬𝒜(0))f(\mathcal{I}_{1}^{s}\,|\,\mathbf{s}(0))=f(\mathcal{I}_{1}^{s}\,|\,\mathbf{s}_{\mathcal{A}}(0),\mathbf{s}_{\mathcal{H}}(0),\mathbf{s}_{\mathcal{R}}(0))=f(\mathcal{I}_{1}^{s}\,|\,\mathbf{s}_{\mathcal{A}}(0)) holds. Therefore, we have

f(1|𝐬(0))\displaystyle f(\mathcal{I}_{1}\,|\,\mathbf{s}(0)) =f(1w)f(1s|𝐬𝒜(0))=f(1w)f(1s|𝐬~𝒜(0))\displaystyle=f(\mathcal{I}_{1}^{w})f(\mathcal{I}_{1}^{s}\,|\,\mathbf{s}_{\mathcal{A}}(0))=f(\mathcal{I}_{1}^{w})f(\mathcal{I}_{1}^{s}\,|\,\tilde{\mathbf{s}}_{\mathcal{A}}(0)) (57)
=f(1|𝐬~(0))\displaystyle=f(\mathcal{I}_{1}\,|\,\tilde{\mathbf{s}}(0))

where we used 𝐬𝒜(0)=𝐬~𝒜(0)\mathbf{s}_{\mathcal{A}}(0)=\tilde{\mathbf{s}}_{\mathcal{A}}(0) in the derivation.

To show f(2|1,𝐬(0))=f(2|1,𝐬~(0))f(\mathcal{I}_{2}\,|\,\mathcal{I}_{1},\mathbf{s}(0))=f(\mathcal{I}_{2}\,|\,\mathcal{I}_{1},\tilde{\mathbf{s}}(0)), it suffices to prove

f(2,𝐬(K+1),𝐰(K+1)|1,𝐬(0))\displaystyle f(\mathcal{I}_{2},\mathbf{s}(K+1),\mathbf{w}(K+1)\,|\,\mathcal{I}_{1},\mathbf{s}(0)) (58)
=\displaystyle= f(2,𝐬(K+1),𝐰(K+1)|1,𝐬~(0))\displaystyle f(\mathcal{I}_{2},\mathbf{s}(K+1),\mathbf{w}(K+1)\,|\,\mathcal{I}_{1},\tilde{\mathbf{s}}(0))

Given 𝐬(K+1)\mathbf{s}(K+1) and 𝐰(K+1)\mathbf{w}(K+1), 2\mathcal{I}_{2} is independent of 1\mathcal{I}_{1}, which further leads to

f(2,𝐬(K+1),𝐰(K+1)|1,𝐬(0))\displaystyle f(\mathcal{I}_{2},\mathbf{s}(K+1),\mathbf{w}(K+1)\,|\,\mathcal{I}_{1},\mathbf{s}(0)) (59)
=\displaystyle= f(2|𝐬(K+1),𝐰(K+1),1,𝐬(0))\displaystyle f(\mathcal{I}_{2}\,|\,\mathbf{s}(K+1),\mathbf{w}(K+1),\mathcal{I}_{1},\mathbf{s}(0))
×f(𝐬(K+1),𝐰(K+1)|1,𝐬(0))\displaystyle\times f(\mathbf{s}(K+1),\mathbf{w}(K+1)\,|\,\mathcal{I}_{1},\mathbf{s}(0))
=\displaystyle= f(2|𝐬(K+1),𝐰(K+1))\displaystyle f(\mathcal{I}_{2}\,|\,\mathbf{s}(K+1),\mathbf{w}(K+1))
×f(𝐬(K+1),𝐰(K+1)|1,𝐬(0))\displaystyle\times f(\mathbf{s}(K+1),\mathbf{w}(K+1)\,|\,\mathcal{I}_{1},\mathbf{s}(0))
=\displaystyle= f(2|𝐬(K+1),𝐰(K+1))\displaystyle f(\mathcal{I}_{2}\,|\,\mathbf{s}(K+1),\mathbf{w}(K+1))
×f(𝐬(K+1)|𝐰(K+1),1,𝐬(0))f(𝐰(K+1)|1,𝐬(0))\displaystyle\times f(\mathbf{s}(K+1)\,|\,\mathbf{w}(K+1),\mathcal{I}_{1},\mathbf{s}(0))f(\mathbf{w}(K+1)\,|\,\mathcal{I}_{1},\mathbf{s}(0))

Further taking into account the facts that 1) 𝐬(K+1)\mathbf{s}(K+1) is conditionally independent of 𝐰(K+1)\mathbf{w}(K+1) and 1w\mathcal{I}_{1}^{w} given 1s\mathcal{I}_{1}^{s} and 𝐬(0)\mathbf{s}(0); and 2) 𝐰(K+1)\mathbf{w}(K+1) is conditionally independent of 1s\mathcal{I}_{1}^{s} and 𝐬(0)\mathbf{s}(0) given 1w\mathcal{I}_{1}^{w}, one can obtain

f(2,𝐬(K+1),𝐰(K+1)|1,𝐬(0))\displaystyle f(\mathcal{I}_{2},\mathbf{s}(K+1),\mathbf{w}(K+1)\,|\,\mathcal{I}_{1},\mathbf{s}(0)) (60)
=\displaystyle= f(2|𝐬(K+1),𝐰(K+1))f(𝐬(K+1)|1s,𝐬(0))\displaystyle f(\mathcal{I}_{2}\,|\,\mathbf{s}(K+1),\mathbf{w}(K+1))f(\mathbf{s}(K+1)\,|\,\mathcal{I}_{1}^{s},\mathbf{s}(0))
×f(𝐰(K+1)|1w)\displaystyle\times f(\mathbf{w}(K+1)\,|\,\mathcal{I}_{1}^{w})

Similarly, one can also obtain

f(2,𝐬(K+1),𝐰(K+1)|1,𝐬~(0))\displaystyle f(\mathcal{I}_{2},\mathbf{s}(K+1),\mathbf{w}(K+1)\,|\,\mathcal{I}_{1},\tilde{\mathbf{s}}(0)) (61)
=\displaystyle= f(2|𝐬(K+1),𝐰(K+1))f(𝐬(K+1)|1s,𝐬~(0))\displaystyle f(\mathcal{I}_{2}\,|\,\mathbf{s}(K+1),\mathbf{w}(K+1))f(\mathbf{s}(K+1)\,|\,\mathcal{I}_{1}^{s},\tilde{\mathbf{s}}(0))
×f(𝐰(K+1)|1w)\displaystyle\times f(\mathbf{w}(K+1)\,|\,\mathcal{I}_{1}^{w})

Therefore, to show f(2|1,𝐬(0))=f(2|1,𝐬~(0))f(\mathcal{I}_{2}\,|\,\mathcal{I}_{1},\mathbf{s}(0))=f(\mathcal{I}_{2}\,|\,\mathcal{I}_{1},\tilde{\mathbf{s}}(0)), it suffices to prove f(𝐬(K+1)|1s,𝐬(0))=f(𝐬(K+1)|1s,𝐬~(0))f(\mathbf{s}(K+1)\,|\,\mathcal{I}_{1}^{s},\mathbf{s}(0))=f(\mathbf{s}(K+1)\,|\,\mathcal{I}_{1}^{s},\tilde{\mathbf{s}}(0)). Denote s\mathcal{I}_{\mathcal{R}}^{s} as s{𝚫𝐬(k)|k=0,1,,K}\mathcal{I}_{\mathcal{R}}^{s}\triangleq\{\boldsymbol{\Delta}\mathbf{s}_{\mathcal{R}}(k)\,|\,k=0,1,\ldots,K\}. To prove f(𝐬(K+1)|1s,𝐬(0))=f(𝐬(K+1)|1s,𝐬~(0))f(\mathbf{s}(K+1)\,|\,\mathcal{I}_{1}^{s},\mathbf{s}(0))=f(\mathbf{s}(K+1)\,|\,\mathcal{I}_{1}^{s},\tilde{\mathbf{s}}(0)), we only need to prove

f(𝐬(K+1),s|1s,𝐬(0))=f(𝐬(K+1),s|1s,𝐬~(0))\displaystyle f(\mathbf{s}(K+1),\mathcal{I}_{\mathcal{R}}^{s}\,|\,\mathcal{I}_{1}^{s},\mathbf{s}(0))=f(\mathbf{s}(K+1),\mathcal{I}_{\mathcal{R}}^{s}\,|\,\mathcal{I}_{1}^{s},\tilde{\mathbf{s}}(0)) (62)

From (46), we can obtain the following facts: 1) 𝐬(K+1)\mathbf{s}_{\mathcal{H}}(K+1) is conditionally independent of 𝐬𝒜(K+1)\mathbf{s}_{\mathcal{A}}(K+1) and 𝐬(K+1)\mathbf{s}_{\mathcal{R}}(K+1) given s\mathcal{I}_{\mathcal{R}}^{s}, 1s\mathcal{I}_{1}^{s}, and 𝐬(0)\mathbf{s}(0); and 2) 𝐬𝒜(K+1)\mathbf{s}_{\mathcal{A}}(K+1) and 𝐬(K+1)\mathbf{s}_{\mathcal{R}}(K+1) are conditionally independent of 𝐬(0)\mathbf{s}_{\mathcal{H}}(0) given s\mathcal{I}_{\mathcal{R}}^{s}, 1s\mathcal{I}_{1}^{s}, 𝐬𝒜(0)\mathbf{s}_{\mathcal{A}}(0), and 𝐬(0)\mathbf{s}_{\mathcal{R}}(0). Taking into account these facts, we have

f(𝐬(K+1),s|1s,𝐬(0))\displaystyle f(\mathbf{s}(K+1),\mathcal{I}_{\mathcal{R}}^{s}\,|\,\mathcal{I}_{1}^{s},\mathbf{s}(0)) (63)
=\displaystyle= f(𝐬(K+1),𝐬𝒜(K+1),𝐬(K+1),s|1s,𝐬(0))\displaystyle f(\mathbf{s}_{\mathcal{H}}(K+1),\mathbf{s}_{\mathcal{A}}(K+1),\mathbf{s}_{\mathcal{R}}(K+1),\mathcal{I}_{\mathcal{R}}^{s}\,|\,\mathcal{I}_{1}^{s},\mathbf{s}(0))
=\displaystyle= f(𝐬(K+1)|𝐬𝒜(K+1),𝐬(K+1),s,1s,𝐬(0))\displaystyle f(\mathbf{s}_{\mathcal{H}}(K+1)\,|\,\mathbf{s}_{\mathcal{A}}(K+1),\mathbf{s}_{\mathcal{R}}(K+1),\mathcal{I}_{\mathcal{R}}^{s},\mathcal{I}_{1}^{s},\mathbf{s}(0))
×f(𝐬𝒜(K+1),𝐬(K+1)|s,1s,𝐬(0))f(s|1s,𝐬(0))\displaystyle\times f(\mathbf{s}_{\mathcal{A}}(K+1),\mathbf{s}_{\mathcal{R}}(K+1)\,|\,\mathcal{I}_{\mathcal{R}}^{s},\mathcal{I}_{1}^{s},\mathbf{s}(0))f(\mathcal{I}_{\mathcal{R}}^{s}\,|\,\mathcal{I}_{1}^{s},\mathbf{s}(0))
=\displaystyle= f(𝐬(K+1)|s,1s,𝐬(0))\displaystyle f(\mathbf{s}_{\mathcal{H}}(K+1)\,|\,\mathcal{I}_{\mathcal{R}}^{s},\mathcal{I}_{1}^{s},\mathbf{s}(0))
×f(𝐬𝒜(K+1),𝐬(K+1)|s,1s,𝐬𝒜(0),𝐬(0))f(s)\displaystyle\times f(\mathbf{s}_{\mathcal{A}}(K+1),\mathbf{s}_{\mathcal{R}}(K+1)\,|\,\mathcal{I}_{\mathcal{R}}^{s},\mathcal{I}_{1}^{s},\mathbf{s}_{\mathcal{A}}(0),\mathbf{s}_{\mathcal{R}}(0))f(\mathcal{I}_{\mathcal{R}}^{s})

where in the derivation we used the independence between s\mathcal{I}_{\mathcal{R}}^{s} and {1s,𝐬(0)}\{\mathcal{I}_{1}^{s},\mathbf{s}(0)\}. Similarly, one can obtain

f(𝐬(K+1),s|1s,𝐬~(0))\displaystyle f(\mathbf{s}(K+1),\mathcal{I}_{\mathcal{R}}^{s}\,|\,\mathcal{I}_{1}^{s},\tilde{\mathbf{s}}(0)) (64)
=\displaystyle= f(𝐬(K+1)|s,1s,𝐬~(0))\displaystyle f(\mathbf{s}_{\mathcal{H}}(K+1)\,|\,\mathcal{I}_{\mathcal{R}}^{s},\mathcal{I}_{1}^{s},\tilde{\mathbf{s}}(0))
×f(𝐬𝒜(K+1),𝐬(K+1)|s,1s,𝐬~𝒜(0),𝐬~(0))f(s)\displaystyle\times f(\mathbf{s}_{\mathcal{A}}(K+1),\mathbf{s}_{\mathcal{R}}(K+1)\,|\,\mathcal{I}_{\mathcal{R}}^{s},\mathcal{I}_{1}^{s},\tilde{\mathbf{s}}_{\mathcal{A}}(0),\tilde{\mathbf{s}}_{\mathcal{R}}(0))f(\mathcal{I}_{\mathcal{R}}^{s})

From Lemma 2, we have that if at some time instant 0kK0\leq k^{*}\leq K, agent ii has an in-neighbor or out-neighbor ll not belonging to 𝒜\mathcal{A}, f(𝐬(K+1)|s,𝐬(0))=f(𝐬(K+1)|s,𝐬~(0))f(\mathbf{s}_{\mathcal{H}}(K+1)\,|\,\mathcal{I}_{s}^{*},\mathbf{s}(0))=f(\mathbf{s}_{\mathcal{H}}(K+1)\,|\,\mathcal{I}_{s}^{*},\tilde{\mathbf{s}}(0)) holds, where s={Δsmn(k)|(m,n)(k),(m,n)(i,l),(m,n)(l,i),k=0,1,,K}={Δsmn(k)|(m,n)𝒜(k)(k),k=0,1,,K}\mathcal{I}_{s}^{*}=\{\Delta s_{mn}(k)\,|\,(m,\,n)\in\mathcal{E}(k),(m,\,n)\neq(i,\,l),(m,\,n)\neq(l,\,i),k=0,1,\ldots,K\}=\{\Delta s_{mn}(k)\,|\,(m,\,n)\in\mathcal{E_{A}}(k)\cup\mathcal{E_{R}}(k),k=0,1,\ldots,K\} is the collection of all elements Δsmn(k)\Delta s_{mn}(k) in 𝚫𝐬𝒜(k)\boldsymbol{\Delta}\mathbf{s}_{\mathcal{A}}(k) and 𝚫𝐬(k)\boldsymbol{\Delta}\mathbf{s}_{\mathcal{R}}(k) from iteration 0 to iteration KK. Given that 𝚫𝐬𝒜(k)1s\boldsymbol{\Delta}\mathbf{s}_{\mathcal{A}}(k)\in\mathcal{I}_{1}^{s} and 𝚫𝐬(k)s\boldsymbol{\Delta}\mathbf{s}_{\mathcal{R}}(k)\in\mathcal{I}_{\mathcal{R}}^{s} hold for k=0,1,,Kk=0,1,\ldots,K, we further have f(𝐬(K+1)|s,1s,𝐬(0))=f(𝐬(K+1)|s,1s,𝐬~(0)f(\mathbf{s}_{\mathcal{H}}(K+1)\,|\,\mathcal{I}_{\mathcal{R}}^{s},\mathcal{I}_{1}^{s},\mathbf{s}(0))=f(\mathbf{s}_{\mathcal{H}}(K+1)\,|\,\mathcal{I}_{\mathcal{R}}^{s},\mathcal{I}_{1}^{s},\tilde{\mathbf{s}}(0). Based on 𝐬𝒜(0)=𝐬~𝒜(0)\mathbf{s}_{\mathcal{A}}(0)=\tilde{\mathbf{s}}_{\mathcal{A}}(0) and 𝐬(0)=𝐬~(0)\mathbf{s}_{\mathcal{R}}(0)=\tilde{\mathbf{s}}_{\mathcal{R}}(0), we have f(𝐬(K+1),s|1s,𝐬(0))=f(𝐬(K+1),s|1s,𝐬~(0))f(\mathbf{s}(K+1),\mathcal{I}_{\mathcal{R}}^{s}\,|\,\mathcal{I}_{1}^{s},\mathbf{s}(0))=f(\mathbf{s}(K+1),\mathcal{I}_{\mathcal{R}}^{s}\,|\,\mathcal{I}_{1}^{s},\tilde{\mathbf{s}}(0)), implying f(2|1,𝐬(0))=f(2|1,𝐬~(0))f(\mathcal{I}_{2}\,|\,\mathcal{I}_{1},\mathbf{s}(0))=f(\mathcal{I}_{2}\,|\,\mathcal{I}_{1},\tilde{\mathbf{s}}(0)) if at some time instant 0kK0\leq k^{*}\leq K, agent ii has an in-neighbor or out-neighbor ll not belonging to 𝒜\mathcal{A}.

Therefore, we have f(|𝐬(0))=f(|𝐬~(0))f(\mathcal{I}\,|\,\mathbf{s}(0))=f(\mathcal{I}\,|\,\tilde{\mathbf{s}}(0)) for any 𝐬(0),𝐬~(0)[1N2,N1N2]N\mathbf{s}(0),\,\tilde{\mathbf{s}}(0)\in[\frac{1}{N^{2}},\,\frac{N-1}{N^{2}}]^{N} subject to si(0)+sl(0)=s~i(0)+s~l(0)s_{i}(0)+s_{l}(0)=\tilde{s}_{i}(0)+\tilde{s}_{l}(0) and sj(0)=s~j(0)s_{j}(0)=\tilde{s}_{j}(0) for j𝒱{i,l}j\in\mathcal{V}\setminus\{i,l\}, meaning that the confidentiality of agent ii can be preserved if at some time instant 0kK0\leq k^{*}\leq K, agent ii has an in-neighbor or out-neighbor ll that does not belong to 𝒜\mathcal{A}. \blacksquare

Remark 7

Compared with our previous work [1] which defines privacy as the positivity of probability that adversaries’ accessible information set being the same under two different initial states, in this work we significantly improved our confidential results by proving that the probability distributions of information sets are identical under different initial states, meaning that the initial states are perfectly indistinguishable from the viewpoint of adversaries.

Remark 8

It is worth noting that although the confidential approach in [30] looks similar to ours (both protocols employ random values in the first several iterations), they are in fact significantly different. More specifically, to guarantee the accuracy of average consensus, not only does the confidential protocol in [30] require pairwise local averaging of exchanged random values in the first several iterations, but it also needs to compensate the errors induced by the random values immediately after the first several iterations, which is equivalent to introducing time-correlated additive random noises to agents’ states. To the contrary, our confidential approach exchanges time-uncorrelated random values for iterations kKk\leq K. To ensure consensus accuracy, each agent ii uses a carefully-designed Δsii(k)\Delta s_{ii}(k) to compensate the errors induced by the random values at each iteration kKk\leq K. Therefore, the random values used in our approach are space-correlated. As shown in Theorem 2, the space-correlated randomness can make the probability distributions of information sets accessible to adversaries identical under different initial states and hence achieves information-theoretic privacy, which is stronger than the confidentiality achieved using time-correlated noises in [30].

Next we proceed to show that if the conditions in Theorem 2 are not met, then the confidentiality of agent ii may be breached.

Theorem 3

Consider a network of NN agents represented by a sequence of time-varying directed graphs {𝒢(k)=(𝒱,(k))}\{\mathcal{G}(k)=(\mathcal{V},\,\mathcal{E}(k))\} which satisfy Assumptions 1, 2, 3, 4, and 5. Under the proposed Algorithm 1, the confidentiality of agent i𝒜i\notin\mathcal{A} cannot be preserved against 𝒜\mathcal{A} if all in-neighbors and out-neighbors of agent ii belong to 𝒜\mathcal{A}, i.e., {𝒩iin(k)𝒩iout(k)|k0}𝒜\{\mathcal{N}_{i}^{in}(k)\cup\mathcal{N}_{i}^{out}(k)\,\big{|}\,k\geq 0\}\subset\mathcal{A}. In fact, when {𝒩iin(k)𝒩iout(k)|k0}𝒜\{\mathcal{N}_{i}^{in}(k)\cup\mathcal{N}_{i}^{out}(k)\,\big{|}\,k\geq 0\}\subset\mathcal{A} is true, the initial value xi0x_{i}^{0} of agent ii can be uniquely determined by honest-but-curious agents in 𝒜\mathcal{A}.

Proof: From (43) we have

frac(si(k+1)si(k))\displaystyle{\rm frac}\big{(}s_{i}(k+1)-s_{i}(k)\big{)} (65)
=\displaystyle= frac(n𝒩iin(k)Δsin(k)m𝒩iout(k)Δsmi(k))\displaystyle{\rm frac}\Big{(}\sum_{n\in\mathcal{N}_{i}^{in}(k)}\Delta s_{in}(k)-\sum_{m\in\mathcal{N}_{i}^{out}(k)}\Delta s_{mi}(k)\Big{)}

for kKk\leq K and further

frac(si(K+1)si(0))\displaystyle{\rm frac}\big{(}s_{i}(K+1)-s_{i}(0)\big{)} (66)
=\displaystyle= frac(k=0K[n𝒩iin(k)Δsin(k)m𝒩iout(k)Δsmi(k)])\displaystyle{\rm frac}\bigg{(}\sum_{k=0}^{K}\Big{[}\sum_{n\in\mathcal{N}_{i}^{in}(k)}\Delta s_{in}(k)-\sum_{m\in\mathcal{N}_{i}^{out}(k)}\Delta s_{mi}(k)\Big{]}\bigg{)}

Since si(0)[1N2,N1N2]s_{i}(0)\in[\frac{1}{N^{2}},\,\frac{N-1}{N^{2}}] holds, one can obtain

si(0)=\displaystyle s_{i}(0)= frac(si(K+1)\displaystyle{\rm frac}\bigg{(}s_{i}(K+1) (67)
k=0K[n𝒩iin(k)Δsin(k)m𝒩iout(k)Δsmi(k)])\displaystyle-\sum_{k=0}^{K}\Big{[}\sum_{n\in\mathcal{N}_{i}^{in}(k)}\Delta s_{in}(k)-\sum_{m\in\mathcal{N}_{i}^{out}(k)}\Delta s_{mi}(k)\Big{]}\bigg{)}

According to the requirements in Algorithm 1, we have Δsii(k)+m𝒩iout(k)Δsmi(k)=si(k)\Delta s_{ii}(k)+\sum_{m\in\mathcal{N}_{i}^{out}(k)}{\Delta s_{mi}(k)}=s_{i}(k) for kK+1k\geq K+1 and Δwii(k)+m𝒩iout(k)Δwmi(k)=wi(k)\Delta w_{ii}(k)+\sum_{m\in\mathcal{N}_{i}^{out}(k)}{\Delta w_{mi}(k)}=w_{i}(k) for k0k\geq 0. Plugging these relationships into (6) and (7), we can obtain

si(k+1)si(k)=\displaystyle s_{i}(k+1)-s_{i}(k)= (68)
n𝒩iin(k)Δsin(k)m𝒩iout(k)Δsmi(k)\displaystyle\qquad\sum\limits_{n\in\mathcal{N}_{i}^{in}(k)}{\Delta s_{in}(k)}-\sum\limits_{m\in\mathcal{N}_{i}^{out}(k)}{\Delta s_{mi}(k)}

for kK+1k\geq K+1 and

wi(k+1)wi(k)=\displaystyle w_{i}(k+1)-w_{i}(k)= (69)
n𝒩iin(k)Δwin(k)m𝒩iout(k)Δwmi(k)\displaystyle\qquad\sum\limits_{n\in\mathcal{N}_{i}^{in}(k)}{\Delta w_{in}(k)}-\sum\limits_{m\in\mathcal{N}_{i}^{out}(k)}{\Delta w_{mi}(k)}

for k0k\geq 0, which further lead to

si(k)si(K+1)=\displaystyle s_{i}(k)-s_{i}(K+1)= (70)
l=K+1k1[n𝒩iin(k)Δsin(l)m𝒩iout(k)Δsmi(l)]\displaystyle\qquad\sum\limits_{l=K+1}^{k-1}\Big{[}\sum\limits_{n\in\mathcal{N}_{i}^{in}(k)}{\Delta s_{in}(l)}-\sum\limits_{m\in\mathcal{N}_{i}^{out}(k)}{\Delta s_{mi}(l)}\Big{]}

for kK+1k\geq K+1 and

wi(k)wi(0)=\displaystyle w_{i}(k)-w_{i}(0)= (71)
l=0k1[n𝒩iin(k)Δwin(l)m𝒩iout(k)Δwmi(l)]\displaystyle\qquad\sum\limits_{l=0}^{k-1}\Big{[}\sum\limits_{n\in\mathcal{N}_{i}^{in}(k)}{\Delta w_{in}(l)}-\sum\limits_{m\in\mathcal{N}_{i}^{out}(k)}{\Delta w_{mi}(l)}\Big{]}

for k0k\geq 0.

Under Assumption 5, if {𝒩iout(k)𝒩iin(k)|k0}𝒜\{\mathcal{N}_{i}^{out}(k)\cup\mathcal{N}_{i}^{in}(k)\,\big{|}\,k\geq 0\}\subset\mathcal{A} is true, then all terms on the right-hand side of (70) and (71) are known to the honest-but-curious agents in 𝒜\mathcal{A}. Further taking into account wi(0)=1w_{i}(0)=1, we have that agents in 𝒜\mathcal{A} can uniquely determine wi(k)w_{i}(k) for all kk. Under Assumption 1 and {𝒩iout(k)𝒩iin(k)|k0}𝒜\{\mathcal{N}_{i}^{out}(k)\cup\mathcal{N}_{i}^{in}(k)\,\big{|}\,k\geq 0\}\subset\mathcal{A}, there must exist at least one agent j𝒜j\in\mathcal{A} such that (j,i)(j,\,i)\in\mathcal{E}_{\infty} is true. This is because otherwise graph (𝒱,)(\mathcal{V},\,\mathcal{E}_{\infty}) is not strongly connected, which does not satisfy Assumption 1. According to Assumption 2, agent ii directly communicates with agent j𝒜j\in\mathcal{A} at least once in every TT consecutive time instants. So there must exist kK+1k^{\prime}\geq K+1 at which agent ii directly communicates with agent jj, i.e., agent ii sends Δsji(k)\Delta s_{ji}(k^{\prime}) and Δwji(k)\Delta w_{ji}(k^{\prime}) to agent jj at iteration kk^{\prime}. As j𝒜j\in\mathcal{A}, every honest-but-curious agent in 𝒜\mathcal{A} has access to Δsji(k)\Delta s_{ji}(k^{\prime}) and Δwji(k)\Delta w_{ji}(k^{\prime}). So agents in 𝒜\mathcal{A} can easily infer si(k)s_{i}(k^{\prime}) by using the following relationship

si(k)=Δsji(k)Δwji(k)wi(k)=pji(k)si(k)pji(k)wi(k)wi(k)s_{i}(k^{\prime})=\frac{\Delta s_{ji}(k^{\prime})}{\Delta w_{ji}(k^{\prime})}w_{i}(k^{\prime})=\frac{p_{ji}(k^{\prime})s_{i}(k^{\prime})}{p_{ji}(k^{\prime})w_{i}(k^{\prime})}w_{i}(k^{\prime}) (72)

and then determine the value of si(K+1)s_{i}(K+1) using (70). Further making use of (67), agents in 𝒜\mathcal{A} can infer the value of si(0)s_{i}(0), and then uniquely determine the initial value of agent ii using xi0=baN2(N2si(0)1)+ax_{i}^{0}=\frac{b-a}{N-2}(N^{2}s_{i}(0)-1)+a. Therefore, the confidentiality of agent i𝒜i\notin\mathcal{A} cannot be preserved against 𝒜\mathcal{A} if all in-neighbors and out-neighbors of agent ii belong to 𝒜\mathcal{A}. \blacksquare

Remark 9

It is worth noting that in confidential average consensus, topology requirements such as the ones in Theorem 2 are widely used. In fact, to guarantee both accuracy and confidentiality, [31, 26, 27, 28, 29, 30, 33, 34, 35, 39, 40, 41] all rely on similar topology requirements.

Remark 10

Our algorithm can protect the confidentiality of an agent even when all its neighbors interact (at least one does not collude) with adversaries, which is not allowed in [27] and [31].

Remark 11

From the above analysis, we know that introducing randomness into interaction dynamics by each agent ii for kKk\leq K is key to protect confidentiality against honest-but-curious agents. It is worth noting that compared with the conventional push-sum approach which does not take confidentiality into consideration, the introduced randomness in our approach has no influence on the convergence rate γ\gamma. However, the randomness does delay the convergence process and hence leads to a trade-off between confidentiality preservation and convergence time. This is confirmed in our numerical simulations in Fig. 2, which shows that convergence only initiates after k=K+1k=K+1.

Remark 12

If an adversary can obtain side information, then a larger KK protects the confidentiality of more intermediate states si(k)s_{i}(k) for 1kK1\leq k\leq K. This is because for kK+1k\geq K+1, si(k)s_{i}(k) can be easily obtained by its out-neighbor jj due to the relationship si(k)=wi(k)Δsji(k)/Δwji(k)s_{i}(k)=w_{i}(k)\Delta s_{ji}(k)/\Delta w_{ji}(k) if side information about wi(k)w_{i}(k) is available to the adversary jj. Therefore, although a larger KK leads to more delay in the convergence process, as discussed in Remark 11, it can protect more intermediate states (si(k)s_{i}(k) for 1kK1\leq k\leq K) when an adversary can obtain side information. Of course, if side information is not of concern, a smaller KK is preferable to minimize the delay in the convergence process.

Remark 13

Our algorithm can be extended to preserve confidentiality against external eavesdroppers wiretapping all communication links without compromising algorithmic accuracy by patching partially homomorphic encryption. More specifically, using public-key cryptosystems (e.g., Paillier [57], RSA [58], and ElGamal [59]), each agent generates and floods its public key before the consensus iteration starts. Then in decentralized implementation, an agent encrypts its messages to be sent, which can be decrypted by a legitimate recipient without the help of any third party. Note that since public-key cryptosystems can only deal with integers, the final consensus result would be subject to a quantization error. However, as indicated in our previous work [40], the quantization error can be made arbitrarily small in implementation.

IV Results Validation

We conducted numerical simulations to verify the correctness and the effectiveness of our proposed approach.

We first evaluated our proposed Algorithm 1 under a network of N=5N=5 agents interacting on a time-varying directed graph. More specifically, we used the interaction graph in Fig. 1(a) when kk is even and Fig. 1(b) when kk is odd. It can be verified that this time-varying directed graph satisfies Assumptions 1 and 2. Parameter ε\varepsilon was set to 0.050.05. The initial values xi0x_{i}^{0} for i=1,,Ni=1,\ldots,N were randomly chosen from (50, 50)(-50,\,50). We used e(k)e(k) to measure the estimation error between the estimate πi(k)=si(k)/wi(k)\pi_{i}(k)=s_{i}(k)/w_{i}(k) and the true average value x¯0=j=1Nxj0/N\bar{x}^{0}={\sum_{j=1}^{N}x_{j}^{0}}/{N} at iteration kk, i.e.,

e(k)=𝝅(k)x¯0𝟏=(i=1N(πi(k)x¯0)2)1/2\displaystyle e(k)=\big{\|}\boldsymbol{\pi}(k)-\bar{x}^{0}\mathbf{1}\big{\|}=\big{(}\sum\limits_{i=1}^{N}(\pi_{i}(k)-\bar{x}^{0})^{2}\big{)}^{1/2} (73)

Three experiments were conducted with parameter KK being 1010, 2020, and 3030, respectively. The evolution of e(k)e(k) is shown in Fig. 2. It can be seen that e(k)e(k) approached 0, meaning that every agent converged to the average value x¯0=j=1Nxj0/N\bar{x}^{0}=\sum_{j=1}^{N}x_{j}^{0}/{N}. From Fig. 2, we can also see that Algorithm 1 did not start to converge in the first K+1K+1 iterations due to the introduced randomness, which confirms our analysis in Remark 11.

Refer to caption
Figure 1: A time-varying directed graph with 55 agents.
Refer to caption
Figure 2: The evolution of error e(k)e(k) under different KK.

We also evaluated the influence of parameter ε\varepsilon on the convergence rate γ\gamma. The interaction graph was the same as above. KK was set to 1010. The simulation results are given in Fig. 3 where the mean and variance of γ\gamma from 1,0001,000 runs of the algorithm are shown under different values of ε\varepsilon. Fig. 3 shows that as ε\varepsilon increases, the convergence rate γ\gamma decreases (i.e., the convergence speed increases), which confirms our analysis in Sec. III-B.

Refer to caption
Figure 3: The influence of ε\varepsilon on the convergence rate γ\gamma.

We then compared the proposed Algorithm 1 with existing data-obfuscation based approaches, more specifically, the differential-privacy based approach in [14], the decaying-noise approach in [27], and the finite-noise-sequence approach in [31]. Under the same setup as in the previous simulation, we chose the initial values as {10,15,20,25,30}\{10,15,20,25,30\}, which led to an average value 2020. We adopted the weight matrix 𝐖\mathbf{W} from [14], i.e., the ijij-th entry was wij=1/(|𝒩jout|+1)w_{ij}={1}/{(|\mathcal{N}_{j}^{out}|+1)} for i𝒩jout{j}i\in\mathcal{N}_{j}^{out}\cup\{j\} and wij=0w_{ij}=0 for i𝒩jout{j}i\notin\mathcal{N}_{j}^{out}\cup\{j\}. As the graph is directed and imbalanced, and does not meet the undirected or balanced assumption in [14, 31, 27], all three approaches failed to achieve average consensus, as shown in the numerical simulation results in Fig. 4, Fig. 5, and Fig. 6, respectively.

Refer to caption
Figure 4: The evolution of xi(k)x_{i}(k) under the approach in [14].
Refer to caption
Figure 5: The evolution of xi(k)x_{i}(k) under the approach in [27].
Refer to caption
Figure 6: The evolution of xi(k)x_{i}(k) under the approach in [31].

Finally, we conducted numerical simulations to verify the scalability of our proposed Algorithm 1 using a network of N=1,000N=1,000 agents. At every iteration kk, each agent ii was assumed to have three out-neighbors, i.e.,

𝒩iout(k)={{i¯+1,i+1¯+1,i+2¯+1}ifkis even{i2¯+1,i3¯+1,i4¯+1}ifkis odd\mathcal{N}_{i}^{out}(k)=\left\{\begin{aligned} &\left\{\overline{i}+1,\overline{i+1}+1,\overline{i+2}+1\right\}\qquad\textnormal{if}\ k\ \textnormal{is even}\\ &\left\{\overline{i-2}+1,\overline{i-3}+1,\overline{i-4}+1\right\}\ \textnormal{if}\ k\ \textnormal{is odd}\\ \end{aligned}\right. (74)

where the superscript “¯\bar{\quad}” represents modulo operation on NN, i.e., i¯imodN\overline{i}\triangleq i\mod N. The initial values xi0x_{i}^{0} for i=1,2,,Ni=1,2,\ldots,N were randomly chosen from (50, 50)(-50,\,50). ε\varepsilon and KK were set to 0.050.05 and 1010, respectively. The evolution of estimation error e(k)=𝝅(k)x¯0𝟏e(k)=\|\boldsymbol{\pi}(k)-\bar{x}^{0}\mathbf{1}\| is shown in Fig. 7. It can be seen that e(k)e(k) converged to 0, implying that our proposed algorithm can guarantee the convergence of all agents to the actual average value even when the network size is large.

Refer to caption
Figure 7: The evolution of error e(k)e(k) in a network of N=1,000N=1,000 agents.

V Conclusions

We proposed a confidential average consensus algorithm for time-varying directed graphs. In distinct difference from existing differential-privacy based approaches which enable confidentiality through compromising the accuracy of obtained consensus value, we leveraged the inherent robustness of average consensus to embed randomness in interaction dynamics, which guarantees confidentiality of participating agents without sacrificing the accuracy of average consensus. Finally, we provided numerical simulation results to confirm the effectiveness and efficiency of our proposed approach.

References

  • [1] H. Gao, C. Zhang, M. Ahmad, and Y. Q. Wang, “Privacy-preserving average consensus on directed graphs using push-sum,” in 2018 IEEE Conference on Communications and Network Security (CNS), 2018.
  • [2] J. E. Boillat, “Load balancing and poisson equation in a graph,” Concurrency and Computation: Practice and Experience, vol. 2, no. 4, pp. 289–313, 1990.
  • [3] G. Cybenko, “Dynamic load balancing for distributed memory multiprocessors,” J. Parallel Distrib. Comput., vol. 7, no. 2, pp. 279–301, 1989.
  • [4] N. A. Lynch, Distributed algorithms.   San Francisco, CA: Morgan Kaufmann Publishers, 1996.
  • [5] D. S. Scherber and H. C. Papadopoulos, “Locally constructed algorithms for distributed computations in ad-hoc networks,” in Proc. Information Processing Sensor Networks (IPSN), 2004, pp. 11–19.
  • [6] L. Xiao, S. Boyd, and S. Lall, “A scheme for robust distributed sensor fusion based on average consensus,” in Proc. Information Processing Sensor Networks (IPSN), 2005, pp. 63–70.
  • [7] R. Olfati-Saber and R. M. Murray, “Consensus problems in networks of agents with switching topology and time-delays,” IEEE Trans. Autom. Control, vol. 49, no. 9, pp. 1520–1533, 2004.
  • [8] W. Ren and R. W. Beard, “Consensus seeking in multiagent systems under dynamically changing interaction topologies,” IEEE Trans. Autom. Control, vol. 50, no. 5, pp. 655–661, 2005.
  • [9] O. L. Mangasarian, “Privacy-preserving linear programming,” Optimization Letters, vol. 5, no. 1, pp. 165–172, 2011.
  • [10] R. Hoenkamp, G. B. Huitema, and A. J. C. de Moor-van Vugt, “The neglected consumer: the case of the smart meter rollout in the netherlands,” Renew. Energy Law Policy Rev., pp. 269–282, 2011.
  • [11] Y. Lou, L. Yu, S. Wang, and P. Yi, “Privacy preservation in distributed subgradient optimization algorithms,” IEEE Trans. Cybern., vol. 48, no. 7, pp. 2154–2165, 2017.
  • [12] J. N. Tsitsiklis, “Problems in decentralized decision making and computation,” Ph.D. dissertation, 1984.
  • [13] Z. Zhang and M. Y. Chow, “Incremental cost consensus algorithm in a smart grid environment,” in 2011 IEEE Power and Energy Society General Meeting, 2011, pp. 1–6.
  • [14] Z. Huang, S. Mitra, and G. Dullerud, “Differentially private iterative synchronous consensus,” in Proceedings of the 2012 ACM workshop on Privacy in the electronic society, 2012, pp. 81–90.
  • [15] Z. Huang, S. Mitra, and N. Vaidya, “Differentially private distributed optimization,” in Proc. Int. Conf. Distrib. Comput. Netw., 2015, pp. 4:1–4:10.
  • [16] E. Nozari, P. Tallapragada, and J. Cortés, “Differentially private average consensus with optimal noise selection,” IFAC-PapersOnLine, vol. 48, no. 22, pp. 203–208, 2015.
  • [17] ——, “Differentially private average consensus: obstructions, trade-offs, and optimal algorithm design,” Automatica, vol. 81, pp. 221–231, 2017.
  • [18] V. Katewa, F. Pasqualetti, and V. Gupta, “On privacy vs. cooperation in multi-agent systems,” Int. J. Control, vol. 91, no. 7, pp. 1693–1707, 2018.
  • [19] L. Gao, S. Deng, W. Ren, and C. Hu, “Differentially private consensus with quantized communication,” IEEE Trans. Cybern., 2019.
  • [20] D. Ye, T. Zhu, W. Zhou, and S. Y. Philip, “Differentially private malicious agent avoidance in multiagent advising learning,” IEEE Trans. Cybern., 2019.
  • [21] M. Kefayati, M. S. Talebi, B. H. Khalaj, and H. R. Rabiee, “Secure consensus averaging in sensor networks using random offsets,” in Proc. IEEE Int. Conf. Telecommun. Malaysia Int. Conf. Commun., 2007, pp. 556–560.
  • [22] X. Wang, J. He, P. Cheng, and J. Chen, “Privacy preserving average consensus with different privacy guarantee,” in 2018 Annual American Control Conference (ACC).   IEEE, 2018, pp. 5189–5194.
  • [23] Y. Wang, Z. Huang, S. Mitra, and G. E. Dullerud, “Differential privacy in linear distributed control systems: Entropy minimizing mechanisms and performance tradeoffs,” IEEE Trans. Control Netw. Syst., vol. 4, no. 1, pp. 118–130, 2017.
  • [24] J. He, L. Cai, P. Cheng, J. Pan, and L. Shi, “Distributed privacy-preserving data aggregation against dishonest nodes in network systems,” IEEE Internet of Things Journal, vol. 6, no. 2, pp. 1462–1470, 2018.
  • [25] ——, “Consensus-based data-privacy preserving data aggregation,” IEEE Transactions on Automatic Control, vol. 64, no. 12, pp. 5222–5229, 2019.
  • [26] J. He, L. Cai, C. Zhao, P. Cheng, and X. Guan, “Privacy-preserving average consensus: privacy analysis and algorithm design,” IEEE Transactions on Signal and Information Processing over Networks, vol. 5, no. 1, pp. 127–138, 2018.
  • [27] Y. Mo and R. M. Murray, “Privacy preserving average consensus,” IEEE Trans. Autom. Control, vol. 62, no. 2, pp. 753–765, 2017.
  • [28] T. Charalambous, N. E. Manitara, and C. N. Hadjicostis, “Privacy-preserving average consensus over digraphs in the presence of time delays,” in 2019 57th Annual Allerton Conference on Communication, Control, and Computing (Allerton).   IEEE, 2019, pp. 238–245.
  • [29] N. Gupta, J. Kat, and N. Chopra, “Statistical privacy in distributed average consensus on bounded real inputs,” in 2019 American Control Conference (ACC).   IEEE, 2019, pp. 1836–1841.
  • [30] A. B. Pilet, D. Frey, and F. Taiani, “Robust privacy-preserving gossip averaging,” in International Symposium on Stabilizing, Safety, and Security of Distributed Systems.   Springer, 2019, pp. 38–52.
  • [31] N. E. Manitara and C. N. Hadjicostis, “Privacy-preserving asymptotic average consensus,” in 2013 European Control Conference, 2013, pp. 760–765.
  • [32] X. Duan, J. He, P. Cheng, Y. Mo, and J. Chen, “Privacy preserving maximum consensus,” in 2015 IEEE 54th Annual Conference on Decision and Control (CDC), 2015, pp. 4517–4522.
  • [33] C. Altafini, “A dynamical approach to privacy preserving average consensus,” in 2019 IEEE 58th Conference on Decision and Control (CDC).   IEEE, 2019, pp. 4501–4506.
  • [34] S. Pequito, S. Kar, S. Sundaram, and A. P. Aguiar, “Design of communication networks for distributed computation with privacy guarantees,” in 2014 IEEE 53rd Annual Conference on Decision and Control (CDC), 2014, pp. 1370–1376.
  • [35] I. D. Ridgley, R. A. Freeman, and K. M. Lynch, “Simple, private, and accurate distributed averaging,” in 2019 57th Annual Allerton Conference on Communication, Control, and Computing (Allerton).   IEEE, 2019, pp. 446–452.
  • [36] A. Alaeddini, K. Morgansen, and M. Mesbahi, “Adaptive communication networks with privacy guarantees,” in Proceedings of 2017 American Control Conference, 2017, pp. 4460–4465.
  • [37] M. Kishida, “Encrypted average consensus with quantized control law,” in 2018 IEEE Conference on Decision and Control (CDC).   IEEE, 2018, pp. 5850–5856.
  • [38] C. N. Hadjicostis and A. D. Dominguez-Garcia, “Privacy-preserving distributed averaging via homomorphically encrypted ratio consensus,” IEEE Trans. Autom. Control, 2020.
  • [39] W. Fang, M. Zamani, and Z. Chen, “Secure and privacy preserving consensus for second-order systems based on paillier encryption,” arXiv preprint arXiv:1805.01065, 2018.
  • [40] M. Ruan, H. Gao, and Y. Wang, “Secure and privacy-preserving consensus,” IEEE Trans. Autom. Control, vol. 64, no. 10, pp. 4035–4049, 2019.
  • [41] Y. Wang, “Privacy-preserving average consensus via state decomposition,” IEEE Trans. Autom. Control, vol. 64, no. 11, pp. 4711–4716, 2019.
  • [42] A. Nedić and A. Olshevsky, “Distributed optimization over time-varying directed graphs,” IEEE Transactions on Automatic Control, vol. 60, no. 3, pp. 601–615, 2014.
  • [43] A. Nedic, A. Olshevsky, and W. Shi, “Achieving geometric convergence for distributed optimization over time-varying graphs,” SIAM Journal on Optimization, vol. 27, no. 4, pp. 2597–2633, 2017.
  • [44] J. Zhu, C. Xu, J. Guan, and D. O. Wu, “Differentially private distributed online algorithms over time-varying directed networks,” IEEE Transactions on Signal and Information Processing over Networks, vol. 4, no. 1, pp. 4–17, 2018.
  • [45] A. Scott, “Tactical data diodes in industrial automation and control systems,” SANS Institute InfoSec Reading Room, 2015.
  • [46] D. Kempe, A. Dobra, and J. Gehrke, “Gossip-based computation of aggregate information,” in Proc. 44th IEEE Symp. Found. Comput. Sci., 2003, pp. 482–491.
  • [47] F. Bénézit, V. Blondel, P. Thiran, J. Tsitsiklis, and M. Vetterli, “Weighted gossip: Distributed averaging using non-doubly stochastic matrices,” in Proc. IEEE Int. Symp. Inf. Theory, 2010, pp. 1753–1757.
  • [48] E. Seneta, Non-negative Matrices and Markov Chains.   Springer, 1973.
  • [49] J. A. Fill, “Eigenvalue bounds on convergence to stationarity for nonreversible markov chains, with an application to the exclusion process,” Ann. Appl. Probab., pp. 62–87, 1991.
  • [50] C. N. Hadjicostis, A. D. Domínguez-García, and T. Charalambous, “Distributed averaging and balancing in network systems: with applications to coordination and control,” Foundations and Trends® in Systems and Control, vol. 5, no. 2-3, pp. 99–292, 2018.
  • [51] K. Liu, H. Kargupta, and J. Ryan, “Random projection-based multiplicative data perturbation for privacy preserving distributed data mining,” IEEE Trans. Knowl. Data Eng., vol. 18, no. 1, pp. 92–106, 2006.
  • [52] S. Han, W. K. Ng, L. Wan, and V. C. S. Lee, “Privacy-preserving gradient-descent methods,” IEEE Trans. Knowl. Data Eng., vol. 22, no. 6, pp. 884–899, 2010.
  • [53] N. Cao, C. Wang, M. Li, K. Ren, and W. Lou, “Privacy-preserving multi-keyword ranked search over encrypted cloud data,” IEEE Trans. Parallel Distrib. Syst., vol. 25, no. 1, pp. 222–233, 2014.
  • [54] A. Nedić, A. Olshevsky, and W. Shi, “A geometrically convergent method for distributed optimization over time-varying graphs,” in Proc. IEEE 55th Conf. Decis. Control, 2016, pp. 1023–1029.
  • [55] A. Nedic and A. Ozdaglar, “Distributed subgradient methods for multi-agent optimization,” IEEE Transactions on Automatic Control, vol. 54, no. 1, pp. 48–61, 2009.
  • [56] A. Nedić, A. Ozdaglar, and P. A. Parrilo, “Constrained consensus and optimization in multi-agent networks,” IEEE Trans. Autom. Control, vol. 55, no. 4, pp. 922–938, 2010.
  • [57] P. Paillier, “Public-key cryptosystems based on composite degree residuosity classes,” in International conference on the theory and applications of cryptographic techniques.   Springer, 1999, pp. 223–238.
  • [58] R. L. Rivest, A. Shamir, and L. Adleman, “A method for obtaining digital signatures and public-key cryptosystems,” Communications of the ACM, vol. 21, no. 2, pp. 120–126, 1978.
  • [59] T. ElGamal, “A public key cryptosystem and a signature scheme based on discrete logarithms,” IEEE transactions on information theory, vol. 31, no. 4, pp. 469–472, 1985.
[Uncaptioned image] Huan Gao was born in Shandong, China. He received the B.S. degree in automation and the M.Sc. degree in control theory and control engineering from Northwestern Polytechnical University, Xi’an, Shaanxi, China, in 2011 and 2015, respectively, and the Ph.D. degree in electrical engineering from Clemson University, Clemson, SC, USA, in 2020. He is currently an Associate Professor with the School of Automation, Northwestern Polytechnical University. His research interests include decentralized optimization, cooperative control, and privacy preservation in distributed systems.
[Uncaptioned image] Yongqiang Wang was born in Shandong, China. He received the B.S. degree in electrical engineering and automation, the B.S. degree in computer science and technology from Xi’an Jiaotong University, Xi’an, Shaanxi, China, in 2004, and the M.Sc. and Ph.D. degrees in control science and engineering from Tsinghua University, Beijing, China, in 2009. From 2007 to 2008, he was with the University of Duisburg-Essen, Germany, as a Visiting Student. He was a Project Scientist with the University of California at Santa Barbara, Santa Barbara. He is currently an Associate Professor with the Department of Electrical and Computer Engineering, Clemson University. His current research interests include distributed control, optimization, and learning, with emphasis on privacy protection. He currently serves as an associate editor for IEEE Transactions on Control of Network systems and IEEE Transactions on Automatic Control.