This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Distributed Estimation over Directed Graphs Resilient to Sensor Spoofing

Shamik Bhattacharyya    Kiran Rokade    and Rachel Kalpana Kalaimani This work has been partially supported by DST-INSPIRE Faculty Grant, Department of Science and Technology (DST), Govt. of India (ELE/16-17/333/DSTX/RACH)Shamik Bhattacharyya is with the Electrical Engineering Department, Indian Institute of Technology Madras, TN 600036, India (e-mail: [email protected]). Kiran Rokade is with the Electrical and Computer Engineering Department, Cornell University, Ithaca, NY 14850, USA (e-mail: [email protected]).Rachel K. Kalaimani is with the Electrical Engineering Department, Indian Institute of Technology Madras, TN 600036, India (e-mail: [email protected]).
Abstract

This paper addresses the problem of distributed estimation of an unknown dynamic parameter by a multi-agent system over a directed communication network in the presence of an adversarial attack on the agents’ sensors. The mode of attack of the adversaries is to corrupt the sensor measurements of some of the agents, while the communication and information processing capabilities of those agents remain unaffected. To ensure that all the agents, both normal as well as those under attack, are able to correctly estimate the parameter value, the Resilient Estimation through Weight Balancing (REWB) algorithm is introduced. The only condition required for the REWB algorithm to guarantee resilient estimation is that at any given point in time, less than half of the total number of agents are under attack. The paper discusses the development of the REWB algorithm using the concepts of weight balancing of directed graphs, and the consensus+innovations approach for linear estimation. Numerical simulations are presented to illustrate the performance of our algorithm over directed graphs under different conditions of adversarial attacks.

{IEEEkeywords}

Directed Graphs, Distributed Estimation, Resilient Consensus, Weight-balancing

1 Introduction

\IEEEPARstart

The advancement in wireless sensor networks (WSNs) has diversified their areas of application to agriculture [1], healthcare [2], and renewable energy [3] to name a few. As a result the scale and complexity of the networks is also on the rise [4], [5]. This necessitates the use of more distributed approaches to signal processing over WSNs, and distributed estimation is a key aspect of it. Distributed estimation is about determining a parameter of interest locally at each sensor node with cooperation between neighboring nodes [6]. The increase in areas of application of WSNs has in turn made them more vulnerable to adversarial attacks [7]. A major mode of such attacks are aimed to manipulate the normal functioning of the sensor nodes and thus disrupt the overall signal processing capability of the WSNs. Some commonly used threat models are Byzantine [8], malicious [9], sensor spoofing [10], etc. Hence the distributed estimation algorithms need to be resilient to adversarial attacks in order to be more effective.

Different consensus algorithms resilient to adversarial attacks appear in the literature such as Mean-Subsequence-Reduced algorithm [11] and Median Consensus Algorithm [12]. These algorithms ensure consensus only for the normal agents, while the agents under attack may have arbitrary values. We are interested in a resilient distributed estimation algorithm, that will ensure that both the normal agents and the agents under attack can reach consensus over the true value of the parameter to be estimated. The consensus+innovations approach illustrated in [13] uses the consensus framework to design resilient algorithms for linear estimation. The Constant weight Saturated Innovation Update (CSIU) algorithm [14] is one such resilient estimation algorithm which ensures that all the agents are able to estimate the parameter of interest, provided less than three-tenth of the total agents are under attack. This was further improved in [15], where the Saturated Innovation Update (SIU) algorithm ensures all the agents’ estimate converge to the desired parameter value provided the adversaries attack less than half of the total agents. Also in [14], a new term resilience index was used to provide a bound for the fraction of sensor nodes under attack.

Both the CSIU and SIU algorithms are designed on undirected graphs representing bidirectional communication links between the agents. In many practical scenarios, the power levels at which sensor nodes broadcast information or, their interference and noise patterns, differ from node to node [16], [17]. The communication between nodes in such cases is unidirectional which is aptly represented by a directed graph. Here we consider a time-invariant network topology with unidirectional communication links between agents. To the best of our knowledge, an extension of the consensus+innovations approach to directed graphs is non-existent in the literature, except for the recent work [18]. However, we observed that for the algorithm presented in [18], choosing appropriate parameters is not an easy task. In contrast, we propose an algorithm which guarantees convergence over a given range of parameter values. Also, unlike [18], where the set of adversarial agents and the unknown parameter are fixed, our proposed algorithm works even when the set of agents under attack and the unknown parameter changes with time.

The model of attack by the adversaries is designed on the idea of sensor spoofing [10] where the sensor readings of the agents under attack are corrupted through data falsification or false data injection. Note that such an attack on the agents is restricted to their sensors. In particular, the agents under attack can perform computations and communicate with their neighbours. Also the agents under attack by the adversaries are not known a-priori by the normal agents. Moreover we allow for a more general scenario where the adversaries may attack different agents over time. We present an algorithm, Resilient Estimation through Weight Balancing (REWB), which ensures that all agents asymptotically converge to the value to be estimated provided less than half of the total number of agents are affected by adversaries. The agents operate in a distributed manner using only the local information available to them. The main contribution of this paper is the proposed REWB algorithm which ensures that over a directed communication network, both the normal agents as well as the agents under attack asymptotically estimate the actual value of the unknown parameter in the presence of a sensor spoofing attack by the adversaries.

Technically, the contributions we make in this paper can be summarized as follows:

  • We propose a novel REWB algorithm that estimates an unknown time-varying parameter with a decaying bound on its variations in the presence of sensor spoofing attacks by simultaneously balancing the unbalanced directed communication network (Algorithm 1). The REWB algorithm brings together the weight-balancing and consensus+innovation approaches over relative time-scales to achieve this.

  • We show that the proposed REWB algorithm ensures convergence of each agent, both normal as well as those under attack, to the actual value of the unknown parameter provided less than half of the total agents are under attack at any given time (Theorem 1).

  • As an intermediate result, we provide an explicit rate of convergence of the Laplacian of an unbalanced weighted digraph to the Laplacian of the associated balanced digraph (Lemma 1).

Notations. \mathbb{R} denotes the set of real numbers, and N\mathbb{R}^{N} represents the NN-dimensional Euclidean space. For any set 𝒮\mathcal{S}, the cardinality of the set is denoted by |𝒮||\mathcal{S}|. 𝟏(1,1,,1)\mathbf{1}\coloneqq(1,1,\ldots,1) and 𝟎(0,0,,0)\mathbf{0}\coloneqq(0,0,\ldots,0), of appropriate dimensions. For a real-valued vector vv, vTv^{T} denotes the transpose of the vector, v||v|| denotes its l2l_{2}-norm and v||v||_{\infty} denotes its \infty-norm. Similarly for a real-valued matrix MM, MTM^{T} denotes the transpose of the matrix, and M||M|| denotes its spectral norm. Among the eigenvalues of MM, λ2(M)\lambda_{2}(M) represents the second lowest eigenvalue of MM in magnitude, while λmax(M)\lambda_{\text{max}}(M) denotes its largest eigenvalue in magnitude. For a real-valued vector vv, diag(v)(v) represents a diagonal matrix with vv as the main diagonal.

The rest of the paper is organised as follows. Section-2 discusses the details of the problem such as the inter-agent communication network, the threat model of the adversaries and the concept of resilience index. Section-3 starts with the development of the REWB algorithm using the weight-balancing approach, followed by the details of the algorithm, finally leading to our main result. Some numerical simulations are presented in Section-4 to validate the performance of the REWB algorithm. Finally the conclusions are presented in Section-5.

2 Problem Formulation

2.1 System Model

Consider a system of NN agents where each agent is equipped with sensing, computing and communication capabilities - it can record measurements using its sensor, can perform computations using its own data and the information received from its neighbouring agents, and can also share its data with the neighbours. The aim of each agent is to estimate an unknown parameter θ(t)M\theta^{*}(t)\in\mathbb{R}^{M} in a distributed manner even while some agents’ sensor measurements are corrupted by adversaries. The precise model of sensor measurement corruption will be described shortly.

The communication among the agents is modelled as a directed graph Γ=(𝒱,)\Gamma=(\mathcal{V},\mathcal{E}), where the vertex set 𝒱={1,2,,N}\mathcal{V}=\{1,2,\ldots,N\} represents the set of NN agents. The set of directed edges 𝒱×𝒱\mathcal{E}\subset\mathcal{V}\times\mathcal{V} represents the information exchange links between the agents, where (i,j)(i,j)\in\mathcal{E} if agent jj can send information to agent ii. A directed path from ii to jj is the sequence of directed edges (i,i1),(i1,i2),,(ip,j)(i,i_{1}),(i_{1},i_{2}),\ldots,(i_{p},j). The set of in-neighbours of agent jj is defined as 𝒩j={i𝒱:(j,i)}\mathcal{N}_{j}=\{i\in\mathcal{V}:(j,i)\in\mathcal{E}\}, and the corresponding in-degree is denoted as djin=|𝒩j|d_{j}^{\textrm{in}}=|\mathcal{N}_{j}|. The set of out-neighbours of agent-jj is defined as 𝒪j={i𝒱:(i,j)}\mathcal{O}_{j}=\{i\in\mathcal{V}:(i,j)\in\mathcal{E}\}, and the corresponding out-degree is denoted as djout=|𝒪j|d_{j}^{\text{out}}=|\mathcal{O}_{j}|. A corresponding diagonal matrix is defined as Dout=diag(d1out,,dNout)D^{\text{out}}=\text{diag}\big{(}d_{1}^{\text{out}},\ldots,d_{N}^{\text{out}}\big{)}. The adjacency matrix, AA is a square matrix of size N×NN\times N defined as A=[aij]A=[a_{ij}] where aij=1a_{ij}=1 if (i,j)(i,j)\in\mathcal{E}, and aij=0a_{ij}=0 otherwise. The Laplacian, LL is defined as LDoutAL\coloneqq D^{\text{out}}-A.

Definition 1 (Strongly Connected Graph )

A directed graph is said to be strongly connected if there exists a directed path between every pair of vertices in the graph.

The flow of information is such that each agent ii is able to receive information from its in-neighbours (𝒩i)(\mathcal{N}_{i}), and send its own data to its out-neighbours (𝒪i)(\mathcal{O}_{i}). So the information about any agent ii can be received by another agent jj either directly if a directed communication link exists between them, or indirectly via intermediate agent(s) provided the corresponding directed path exists. In order to ensure that the information about every agent ii reaches every other agent j,(ij;i,j𝒱)j,(i\neq j;i,j\in\mathcal{V}), we introduce the following assumption.

Assumption 1

The directed graph Γ\Gamma is Strongly Connected.

Now we proceed to model the effect of the adversaries, which attack the agents with a motive to disrupt the estimation process thus trying to prevent them from correctly estimating the value θ(t)\theta^{*}(t). At every time-step t0t\geq 0, the agents which are under attack by the adversaries are termed as the the set of Bad (or affected) agents, denoted as t\mathcal{B}_{t}. The remaining agents form the set of Good (or normal) agents, denoted as 𝒢t\mathcal{G}_{t}. The set of bad agents can vary with time, and are also not known a-priori to the set of good agents. So for each t0t\geq 0, the set 𝒱\mathcal{V} is partitioned into 𝒢t\mathcal{G}_{t} and t\mathcal{B}_{t}. Thus 𝒢tt=𝒱,\mathcal{G}_{t}\cup\mathcal{B}_{t}=\mathcal{V},\forall t0t\geq 0. The attack model of the adversary is sensor spoofing attacks. Here the adversary introduces spurious signals into the sensor readings non-invasively [10]. The corruption of sensor readings remains undetected by commonly used filters [19]. So even after nullifying the noise in sensor readings, the effect of the spoofing attack would still percolate into the measurements available to the agent. The agents use these sensor measurements to estimate the unknown parameter. In order to specifically highlight the effect of the adversary, we consider the sensor measurements available to the agents to be free of the effect of any measurement noise. The sensor measurements available to the agents under attack are arbitrary values manipulated by the adversary. Accordingly, we model the sensor measurements recorded by the agents as

for all i𝒢t , yi(t)=θ(t)for all it , yi(t)=θ(t)+ζi(t)\begin{split}\text{for all }i\in\mathcal{G}_{t}&\text{ , }y_{i}(t)=\theta^{*}(t)\\ \text{for all }i\in\mathcal{B}_{t}&\text{ , }y_{i}(t)=\theta^{*}(t)+\zeta_{i}(t)\end{split} (1)

where ζi(t)M\zeta_{i}(t)\in\mathbb{R}^{M} is a vector of arbitrary values reflecting the effect of the adversaries. In the above model there is no boundedness assumption or stochastic approximation considered for ζi(t)\zeta_{i}(t). This preserves the arbitrary nature of the data being manipulated by the adversary. So, if |t|=0|\mathcal{B}_{t}|=0 t0\forall t\geq 0, then yi(t)=θ(t)y_{i}(t)=\theta^{*}(t) i𝒱\forall i\in\mathcal{V}, and the estimation problem would be trivial as the sensor measurements directly provide the correct value of the parameter. Here we are interested in the non-trivial case where there exists some t0t\geq 0 such that |t|0|\mathcal{B}_{t}|\neq 0. This means some of the sensor measurements would be corrupted as yi(t)=θ(t)+ζi(t)y_{i}(t)=\theta^{*}(t)+\zeta_{i}(t) it\forall i\in\mathcal{B}_{t}. So each agent needs to perform some additional computations in order to estimate the true value of θ(t)\theta^{*}(t) in a distributed manner. It should be noted that under this threat model, the bad agents are still able to perform their computations as per design as well as communicate with their neighbours.

The unknown time-varying parameter that is to be estimated is some physical quantity which can be measured by a sensor. So we can safely assume its Euclidean norm to be bounded. Moreover, we also assume that the variations in the unknown parameter asymptotically decay with time.

Assumption 2

The Euclidean norm of the unknown vector quantity that is to be estimated lies within an upper bound known to each agent :

θ(t)Θ\|\theta^{*}(t)\|\leq\Theta (2)

Also, the Euclidean norm of the variation in the unknown vector quantity has a decaying bound :

θ(t+1)θ(t)1/(1+t)θ1\|\theta^{*}(t+1)-\theta^{*}(t)\|\leq 1/(1+t)^{\theta_{1}} (3)

As a consequence of (3), we have the time varying parameter θ(t)\theta^{*}(t) eventually converging to some constant value θ^\hat{\theta}. Specifically, θ(t)θ^ as t\theta^{*}(t)\to\hat{\theta}\text{ as }t\to\infty.

Remark : The above assumption focuses on a particular subset of dynamic parameter estimation. Note that this is a modest extension from the static parameter estimation case.

Now to estimate θ(t)\theta^{*}(t) in a distributed manner, for all t0t\geq 0 each agent ii maintains its own estimate of θ(t)\theta^{*}(t) denoted by xi(t)Mx_{i}(t)\in\mathbb{R}^{M}, also referred to as the state of agent ii. In order to update the state, each agent ii follows the discrete-time single integrator dynamics :

xi(t+1)=xi(t)+ui(t),t0x_{i}(t+1)=x_{i}(t)+u_{i}(t),t\geq 0 (4)

So at every time step, each agent ii performs the following steps in the given sequence :

  1. S1S1 -

    broadcasts its own estimate xi(t)x_{i}(t) to its out-neighbours 𝒪i\mathcal{O}_{i}

  2. S2S2 -

    receives the estimates from its corresponding in-neighbours : xj(t)x_{j}(t), j𝒩ij\in\mathcal{N}_{i}

  3. S3S3 -

    collects sensor measurement of θ(t)\theta^{*}(t) : yi(t)y_{i}(t)

  4. S4S4 -

    updates its own estimate following (4), where ui(t)=f(yi(t),{xj(t),j𝒩i})u_{i}(t)=f(y_{i}(t),\{x_{j}(t),j\in\mathcal{N}_{i}\}), and ff is defined later in Section 3.2.

In a distributed estimation problem with xi(t)x_{i}(t) as the state of agent ii and θ(t)\theta^{*}(t) as the parameter of interest, the aim is to achieve

xi(t)θ(t) as t , for all i𝒱x_{i}(t)\longrightarrow\theta^{*}(t)\text{ as }t\rightarrow\infty\text{ , for all }i\in\mathcal{V} (5)

For the resilient estimation problem considered here, the additional challenge is to achieve (5) even in the presence of adversaries attacking some of the agents. In order to quantify how resilient an algorithm is to the adversarial attacks, we use a measure called the Resilience Index [15]. The resilience index (s)(s) is an upper bound on the fraction of agents which are under attack by the adversaries at any time-step tt. So, s|t|Ns\geq\frac{|\mathcal{B}_{t}|}{N} for all t0t\geq 0, ss\in\mathbb{R}. Thus s=0s=0 would indicate the trivial case where bad agents are totally absent. Having s=1s=1 allows for the possibility of all the agents being under attack at any time-step tt.

In the sequel, we initially proceed to design an algorithm which provides us with a suitable value of ui(t)u_{i}(t) t0\forall t\geq 0, for all i𝒱i\in\mathcal{V} introduced in (4). Then we present our main result on how the newly designed algorithm, under the assumptions made so far, achieves (5).

3 Results

The aim of each agent in the multi-agent system under consideration, is to estimate an unknown static parameter in a distributed manner, as given in (5). The technique used for the distributed estimation of θ(t)\theta^{*}(t) is based on the consensus+innovations approach [13]. Based on this approach we proceed to design an algorithm such that the desired objective, xi(t)θ(t) as t , for all i𝒱x_{i}(t)\longrightarrow\theta^{*}(t)\text{ as }t\rightarrow\infty\text{ , for all }i\in\mathcal{V}, is achieved through fulfilling the following two smaller goals simultaneously as tt\rightarrow\infty:

  1. G1G1 :

    the state of each agent, xi(t)x_{i}(t), approaches the average of the states of all agents, x¯(t)(1/N)i=1Nxi(t)\bar{x}(t)\coloneqq(1/N)\sum_{i=1}^{N}x_{i}(t)

  2. G2G2 :

    x¯(t)\bar{x}(t) approaches the unknown value to be estimated, θ(t)\theta^{*}(t).

3.1 Weight Balancing

In a Multi-Agent System (MAS), the communication network is usually modelled as a graph, with the nodes of the graph representing the agents and the edges between the nodes representing the corresponding communication links between the agents. When the flow of information between agents is bi-directional, the model used is an undirected graph. On the other hand, when the flow of information between agents is unidirectional, a directed graph (or digraph) is required to model it. The directed edges of the digraph represent the unidirectional communication links while preserving their direction of information flow. A weighted graph has each of its edges assigned a real or integer value, referred to as edge weights. Unless specifically mentioned, the edge weights are taken as unity.

In case of an undirected graph, the sum of edge weights of the incoming edges is equal to the sum of the edge weights of the outgoing ones. But this balance in edge weights does not necessarily hold true in the case of a digraph. To overcome this imbalance, we need to find a suitable weight vector wNw\in\mathbb{R}^{N}, where each outgoing edge of agent ii is assigned the weight wiw_{i}. These weights are said to balance the graph if widiout=j𝒩iwjw_{i}d_{i}^{\text{out}}=\sum_{j\in\mathcal{N}_{i}}w_{j}. The notion of a balanced graph is formally defined below.

Definition 2 (Balanced Graph)

A graph Γ\Gamma of NN nodes is said to be balanced if there exists wNw\in\mathbb{R}^{N} such that

L𝟏=𝟎 , 𝟏TL=𝟎TL\mathbf{1}=\mathbf{0}\text{ , }\mathbf{1}^{T}L=\mathbf{0}^{T} (6)

where L(DoutA)diag(w)L\coloneqq(D^{\text{out}}-A)\text{diag}(w)

The weights (w1,w2,,wN)\big{(}w_{1},w_{2},\ldots,w_{N}\big{)} which balance a given digraph are called the balancing weights of the corresponding digraph [17]. Note that an undirected graph is inherently balanced with w=𝟏w=\mathbf{1} as the vector of balancing weights. On the other hand, for a strongly connected digraph the vector of balancing weights is non-trivial. Note that this vector of balancing weights is also unique to the given digraph, up to scaling [17]. For example, consider the strongly connected digraph shown in Fig.1. For this digraph the vector of balancing weights is w=[0.5,1.5,1]Tw=[0.5,1.5,1]^{T}, which is non-trivial and unique up to scaling.

Refer to caption

Figure 1: A directed graph of 3 nodes

The SIU algorithm proposed in [15] for resilient estimation does not work in general for directed graphs. This is later illustrated through a numerical example in Fig.5 in Section 4. We propose to use the idea of balancing weights described above to achieve resilient estimation over digraphs. For the directed graph Γ\Gamma, we use the following update rule, proposed in [17], to iteratively compute a set of balancing weights. Let wi(t)w_{i}(t)\in\mathbb{R} denote the weight at node ii at time-step tt. The initial set of weights assigned to the agents satisfy : wi(0)(1/dmaxout)2Φ+1w_{i}(0)\leq(1/d^{\text{out}}_{\text{max}})^{2\Phi+1}, where dmaxoutd^{\text{out}}_{\text{max}} represents the maximum out-degree and Φ\Phi represents the diameter of the concerned digraph [17]. Then for all t0t\geq 0,

wi(t+1)=12wi(t)+1diout(j𝒩i12wj(t)).w_{i}(t+1)=\frac{1}{2}w_{i}(t)+\frac{1}{d_{i}^{\text{out}}}\Big{(}\sum_{j\in\mathcal{N}_{i}}\frac{1}{2}w_{j}(t)\Big{)}. (7)

Let w(t)=(w1(t),w2(t),,wN(t))w(t)=\big{(}w_{1}(t),w_{2}(t),\ldots,w_{N}(t)\big{)} represent the vector of node-weights at time-step tt. Then the corresponding vector representation of (7) is given by :

w(t+1)=Pw(t)w(t+1)=Pw(t) (8)

where P0.5(I+(Dout)1A)P\coloneqq 0.5\big{(}I+(D^{\text{out}})^{-1}A\big{)}. So for the limiting case, limtw(t)=limtPtw(0)\lim_{t\rightarrow\infty}w(t)=\lim_{t\rightarrow\infty}P^{t}w(0). Now from Lemma 1 of [17], we know that limtPt\lim_{t\rightarrow\infty}P^{t} exists, and that the sequence {w(t)}t0\{w(t)\}_{t\geq 0} converges to the vector of balancing weights. So we define here the vector of weights which balances the digraph Γ\Gamma as

wlimtw(t)=limtPtw(0)w^{\infty}\coloneqq\lim_{t\rightarrow\infty}w(t)=\lim_{t\rightarrow\infty}P^{t}w(0) (9)

The time-varying weighted Laplacian matrix is represented as

L(t)=(DoutA)W(t), where W(t)=diag(w(t))L(t)=\big{(}D^{\text{out}}-A\big{)}W(t),\text{ where }W(t)=\text{diag}\big{(}w(t)\big{)} (10)

Then the Laplacian matrix for the limiting case can be defined using the result from (9) in (10) as

L(DoutA)W, where W=diag{w}L_{\infty}\coloneqq(D^{\text{out}}-A)W_{\infty},\text{ where }W_{\infty}=\text{diag}\{w^{\infty}\} (11)

Now as ww^{\infty} balances the digraph, LL_{\infty} satisfies the desired balancing condition expressed in (6). By definition of L(t)L(t) we have 𝟏TL(t)=𝟎T\mathbf{1}^{T}L(t)=\mathbf{0}^{T} for all t0t\geq 0. So to arrive at the desired balanced graph condition, we need L(t)𝟏=𝟎L(t)\mathbf{1}=\mathbf{0} which is eventually achieved with L(t)L(t) converging to LL_{\infty} as tt\rightarrow\infty. Next we state a lemma which provides an explicit rate for this convergence and additionally provides the rate of decay of L(t)𝟏L(t)\mathbf{1} to 𝟎\mathbf{0}.

Lemma 1

Given L(t)=(DoutA)W(t)L(t)=\big{(}D^{\text{out}}-A\big{)}W(t) and L=(DoutA)WL_{\infty}=(D^{\text{out}}-A)W_{\infty}, there exists constants C>0C>0 and η(0,1)\eta\in(0,1), such that L(t)DoutCηt\|L(t)-D^{\text{out}}\|\leq C\eta^{t} , L(t)𝟏Cηt\|L(t)\mathbf{1}\|\leq C\eta^{t} for all t0t\geq 0.

The proof of Lemma 1 is given in Appendix 6.

3.2 Algorithm

Now we introduce our algorithm, Resilient Estimation through Weight Balancing (REWB). It consists of two main update steps : one for the state of the agents, and the other for the node weights.
The updates performed by agent ii at time-step tt are :

  1. i)

    Updating the estimate

    xi(t+1)=(1β(t)wi(t)diout)xi(t)+β(t)(j𝒩iwj(t)xj(t))+α(t)ki(t)(yi(t)xi(t))\begin{split}x_{i}(t+1)=\big{(}1&-\beta(t)w_{i}(t)d_{i}^{\text{out}}\big{)}x_{i}(t)\\ &+\beta(t)\Big{(}\sum_{j\in\mathcal{N}_{i}}w_{j}(t)x_{j}(t)\Big{)}\\ &+\alpha(t)k_{i}(t)\big{(}y_{i}(t)-x_{i}(t)\big{)}\end{split} (12)
  2. ii)

    Updating the weight

    wi(t+1)=12wi(t)+1diout(j𝒩i12wj(t))w_{i}(t+1)=\frac{1}{2}w_{i}(t)+\frac{1}{d_{i}^{\text{out}}}\Big{(}\sum_{j\in\mathcal{N}_{i}}\frac{1}{2}w_{j}(t)\Big{)} (13)

The update law (12), used by agents to update their estimate of θ(t)\theta^{*}(t), is based upon the consensus+innovation approach. The first two terms, dealing with the agent’s own and neighbours’ estimates and the corresponding node-weights, constitute the consensus part of the update law. The third term, involving the measurements yi(t)y_{i}(t) and a scaling factor ki(t)k_{i}(t), constitute the innovation part. These two parts working simultaneously through the same update law help in achieving the smaller goals G1G1 and G2G2 mentioned before. The above update law uses step-size parameters β(t)\beta(t) and α(t)\alpha(t) to assign proper weightage to its consensus and innovation parts respectively. The parameters are defined as :

α(t)=α0(1+t)α1,β(t)=β0(1+t)β1\alpha(t)=\frac{\alpha_{0}}{(1+t)^{\alpha_{1}}},\beta(t)=\frac{\beta_{0}}{(1+t)^{\beta_{1}}} (14)

where 0<α01/(12s)0<\alpha_{0}\leq 1/(1-2s) , 0<β0<ψ0<\beta_{0}<\psi , 0<β1<α1<θ10<\beta_{1}<\alpha_{1}<\theta_{1}. The constant ψ\psi is defined as ψ2/(Ndmaxin(dmaxin+dmaxout))\psi\coloneqq 2/\big{(}Nd^{\text{in}}_{\text{max}}(d^{\text{in}}_{\text{max}}+d^{\text{out}}_{\text{max}})\big{)}. Note that β1<α1\beta_{1}<\alpha_{1} implies that, in the state update law (12), the weight of the innovation term decays faster than the weight of the consensus term.

The scaling factor, ki(t)k_{i}(t), is used in the innovation part in order to ensure that the effect of the adversaries on the state of an agent always remains bounded.

ki(t){1, if yi(t)xi(t)γ(t)γ(t)yi(t)xi(t), otherwise k_{i}(t)\coloneqq\begin{cases}1&,\text{ if }\|y_{i}(t)-x_{i}(t)\|\leq\gamma(t)\\ \frac{\gamma(t)}{\|y_{i}(t)-x_{i}(t)\|}&,\text{ otherwise }\end{cases} (15)

where γ(t)\gamma(t) is the output of a dynamical system defined as

γ(t)γ1(t)+γ2(t)\gamma(t)\coloneqq\gamma_{1}(t)+\gamma_{2}(t) (16)

The dynamics of γ1(t)\gamma_{1}(t) and γ2(t)\gamma_{2}(t) are defined as

γ1(t+1)\displaystyle\gamma_{1}(t+1) (1c1μ(t)+(1+N)α(t))γ1(t)\displaystyle\coloneqq\big{(}1-c_{1}\mu(t)+(1+\sqrt{N})\alpha(t)\big{)}\gamma_{1}(t)
+(1+N)α(t)γ2(t)+c2ηt\displaystyle\hskip 43.05542pt+(1+\sqrt{N})\alpha(t)\gamma_{2}(t)+c_{2}\eta^{t} (17)
γ2(t+1)\displaystyle\gamma_{2}(t+1) α(t)γ1(t)+(1α(t)(12s))γ2(t)\displaystyle\coloneqq\alpha(t)\gamma_{1}(t)+\big{(}1-\alpha(t)(1-2s)\big{)}\gamma_{2}(t)
+1/(1+t)θ1\displaystyle\hskip 86.11084pt+1/(1+t)^{\theta_{1}} (18)

where, μ(t)=μ0(t+1)μ1\mu(t)=\frac{\mu_{0}}{(t+1)^{\mu_{1}}}, μ0>0\mu_{0}>0, β1<μ1<α1\beta_{1}<\mu_{1}<\alpha_{1}, c1>0c_{1}>0, c2>0c_{2}>0, 0<η<10<\eta<1. The above time-varying system in two variables plays a crucial role in proving our main result. From the definition of ki(t)k_{i}(t) in (15), a corresponding diagonal matrix is defined as

K(t)diag(k1(t),k2(t),,kN(t))K(t)\coloneqq\text{diag}\big{(}k_{1}(t),k_{2}(t),\ldots,k_{N}(t)\big{)} (19)

Let x(t)=(x1T(t),x2T(t),,xNT(t))N×Mx(t)=\big{(}x_{1}^{T}(t),x_{2}^{T}(t),\ldots,x_{N}^{T}(t)\big{)}\in\mathbb{R}^{N\times M} represent the matrix whose rows are the state vectors of the agents at time-step tt. Also let y(t)=(y1T(t),y2T(t),,yNT(t))N×My(t)=\big{(}y_{1}^{T}(t),y_{2}^{T}(t),\ldots,y_{N}^{T}(t)\big{)}\in\mathbb{R}^{N\times M} represent the matrix whose rows are the sensor measurements of the agents at time-step tt. Now we summarise our REWB algorithm as follows :

Algorithm 1 REWB

Given : Graph Γ\Gamma, Θθ(t)\Theta\geq\|\theta^{*}(t)\|, Resilience index ss, and θ1\theta_{1}
Initialize : 0<α01/(12s)0<\alpha_{0}\leq 1/(1-2s), 0<β0<ψ0<\beta_{0}<\psi, x(0)=0x(0)=0, μ0<(λmβ0λM)β0/(2c1)\mu_{0}<(\lambda_{m}-\beta_{0}\lambda_{M})\beta_{0}/(2c_{1}), γ1(0)=0,γ2(0)=Θ\gamma_{1}(0)=0,\gamma_{2}(0)=\Theta , wi(0)(1dmaxout)2Φ+1w_{i}(0)\leq\big{(}\frac{1}{d^{\text{out}}_{\text{max}}}\big{)}^{2\Phi+1}
Choose : 0<β1<μ1<α1<θ10<\beta_{1}<\mu_{1}<\alpha_{1}<\theta_{1}
for t=0,1,t=0,1,\ldots do

  • record y(t)y(t)

  • exchange x(t)x(t) among neighbouring agents

  • update x(t)x(t) :
    x(t+1)=(Iβ(t)L(t))x(t)+α(t)K(t)(y(t)x(t))x(t+1)=\big{(}I-\beta(t)L(t)\big{)}x(t)+\alpha(t)K(t)\big{(}y(t)-x(t)\big{)}

  • update w(t)w(t) :
    w(t+1)=Pw(t)w(t+1)=Pw(t)

  • update γ(t)\gamma(t) : using equations (16), (17) & (18)

end for

3.3 Main Result

The following theorem states our main result on resilient distributed estimation using the REWB algorithm.

Theorem 1

Suppose Assumptions 1 and 2 hold, and the effect of the adversaries is modelled as in (1). Then the REWB algorithm ensures that the state of every agent, xi(t)x_{i}(t) converges to θ(t)\theta^{*}(t), provided s[0,12)s\in[0,\frac{1}{2}). In particular,

limt(t+1)δ1xi(t)θ(t)=0 , for all i𝒱\lim_{t\rightarrow\infty}(t+1)^{\delta_{1}}\|x_{i}(t)-\theta^{*}(t)\|=0\text{ , for all }i\in\mathcal{V} (20)

where 0δ1α1β10\leq\delta_{1}\leq\alpha_{1}-\beta_{1}

The proof of Theorem-1 is given in Appendix-8. Here we provide a remark on the above theorem.

Remark 1

From Theorem-1 it can be inferred that as long as less than half the total number of agents are under attack by the adversaries, the REWB algorithm ensures that each agent correctly estimates θ(t)\theta^{*}(t). Also note that all the agents, even the bad agents, achieve consensus and estimate θ(t)\theta^{*}(t) in a distributed manner.

Remark 2

As noted in Lemma 1, the dynamic weights w(t)w(t) converge to the balancing weights at an exponential rate (ηt\eta^{t}), whereas all time-varying signals in the dynamics of the state update rule converge at a polynomial rate (α(t)=α0/(1+t)α1\alpha(t)=\alpha_{0}/(1+t)^{\alpha_{1}} etc.). Thus, the weights converge faster, which are in turn used in the state update rule. This two time-scales approach facilitates convergence of the algorithm.

4 Simulation Results

We evaluate the performance of our proposed REWB algorithm through numerical simulations. A random network, consisting of 100 agents with directed edges, is generated where each possible edge has a probability of 0.50.5. It models the communication network among the agents. Each agent estimates a scalar time-varying parameter θ(t)=25+1/(t+1)\theta^{*}(t)=25+1/(t+1) with Θ=50\Theta=50 and θ1=1\theta_{1}=1. The required algorithm parameters are chosen as : α0=0.01,α1=0.075,β0=0.01,β1=0.01,μ0=0.025,μ1=0.025,c1=75,c2=75\alpha_{0}=0.01,\alpha_{1}=0.075,\beta_{0}=0.01,\beta_{1}=0.01,\mu_{0}=0.025,\mu_{1}=0.025,c_{1}=75,c_{2}=75 , and η=0.5\eta=0.5. The initial weights are chosen as wi(0)=0.1w_{i}(0)=0.1 i𝒱\forall i\in\mathcal{V}.

The noise term ζi(t)\zeta_{i}(t) models the effect of the adversaries on the sensor measurements of agent ii. For each bad agent iti\in\mathcal{B}_{t}, at every time step, ζi(t)\zeta_{i}(t) takes on a random value uniformly distributed between 0 and Θ-\Theta. Note that the REWB algorithm works for any other range also. We select the base resilience index to be s=0.405s=0.405, and correspondingly choose |t|=40|\mathcal{B}_{t}|=40. At first we consider two cases with respect to the set of agents under attack and observe the performance of the REWB algorithm. In Fig. 2(a), t\mathcal{B}_{t} has a fixed set of agents, while in Fig. 2(b), t\mathcal{B}_{t} is allowed to vary with time. Both the plots in Fig. 2 show the error in estimation of θ(t)\theta^{*}(t) by the agents, given by x(t)θ(t)𝟏||x(t)-\theta^{*}(t)\mathbf{1}||. From the proof of Theorem 1 we have |xi(t)θ(t)|γ(t)|x_{i}(t)-\theta^{*}(t)|\leq\gamma(t), i𝒱\forall i\in\mathcal{V}, t0\forall t\geq 0. Then for a set of NN agents, we have x(t)θ(t)𝟏Nγ(t)||x(t)-\theta^{*}(t)\mathbf{1}||\leq\sqrt{N}\gamma(t). Fig. 2 shows that, regardless of adversaries attacking a fixed or varying set of agents, the REWB algorithm ensures that the estimation error always remains bounded by Nγ(t)\sqrt{N}\gamma(t), and consequently dies down asymptotically.

Refer to caption
(a)
Refer to caption
(b)
Figure 2: Performance of REWB with adversarial attacks on 40 agents where the set of bad agents is (a) fixed, and (b) variable.

Next we use two different variations in operating conditions compared to the one used in Fig. 2(a) and observe their effect in the performance of the REWB algorithm in Fig. 3. For the plot in Fig. 3(a), the resilience index is decreased to s=0.255s=0.255 and correspondingly we choose |t|=25|\mathcal{B}_{t}|=25. As can be observed the estimation error dies down much faster with a decrease in ss. Next for the plot Fig. 3(b), we simulate an increase in the degree of manipulation done by the adversaries on the sensor measurements by increasing the noise level. We assign ζi(t)=5Θ\zeta_{i}(t)=5\Theta it\forall i\in\mathcal{B}_{t}, t0\forall t\geq 0. As is evident from Fig. 3(b), a high value of ζi(t)\zeta_{i}(t) is also quite efficiently handled by the REWB algorithm, with the estimation error remaining bounded by Nγ(t)\sqrt{N}\gamma(t) at all times and eventually converging to 0. From Fig. 2 and Fig. 3 it is evident that the REWB algorithm ensures that even the bad agents are able to eventually correctly estimate the true value of θ(t)\theta^{*}(t), along with the good agents. This is in accordance with the Remark stated in Section 3.3.

Refer to caption
(a)
Refer to caption
(b)
Figure 3: Performance of REWB with (a) decrease in resilience index, and (b) increased manipulation in sensor measurements by adversaries.

In Section 3.1, we mentioned that the SIU algorithm in [15] does not give convergence in general when applied over a directed network of agents. In Fig. 5, we compare the performance of our REWB algorithm with the SIU algorithm in estimating the value of a scalar constant parameter θ\theta^{*}\in\mathbb{R} over a directed network of 100 agents with s=0.405s=0.405. The two plots on the left show how the states of the agents behave with time, while the two plots on the right show the net estimation error. Fig. 5(a) shows how on applying the SIU algorithm, the states of the agents diverge away from each other and never achieve consensus, leading to a constant estimation error. On the other hand, Fig. 5(b) shows how our REWB algorithm not only ensures the agents reach consensus but they also correctly estimate the value of θ\theta^{*}. This is made possible by the introduction of the weight balancing idea while designing the REWB algorithm. The dynamics of the time-varying weights ensure that the weighted graph eventually approaches a balanced condition, and thus consensus is achieved.

Refer to caption
Refer to caption
(a)
Refer to caption
Refer to caption
(b)
Figure 5: Performance of (a) the SIU [15] algorithm and (b) the proposed REWB algorithm over directed graphs

5 Conclusion

In this paper we propose the Resilient Estimation through Weight Balancing (REWB) algorithm. It is a distributed estimation algorithm designed to work for a network of sensor nodes with directed communication links. The REWB algorithm is resilient to sensor spoofing type adversarial attacks while estimating an unknown time-varying parameter with a decaying bound on its variations. It ensures that along with the unaffected agents, even the agents under attack estimate the true value of the parameter in a distributed manner. The proposed algorithm is developed based on the consensus+innovation approach and uses the weight-balancing idea to ensure consensus over directed graph. Through numerical simulations it is shown that the proposed algorithm accurately estimates an unknown parameter under different attack conditions, provided less than half of the agents are under adversarial attack at any given point of time. Future direction of work is to consider other models of adversarial attacks.

\useRomanappendicesfalse\appendices

6

Proof 6.2 (Proof of Lemma 1).

By definition, PP is a primitive matrix with spectral radius 1 [17]. Then by properties of primitive matrices : limtPt\lim_{t\rightarrow\infty}P^{t} exists, and limtPt=P=uvT\lim_{t\rightarrow\infty}P^{t}=P_{\infty}=uv^{T}, where u,vu,v are the right and left eigen-vectors of PP corresponding to eigen-value 1, and vTu=1v^{T}u=1. Then from (7) and (9) we have :

w(t)w=(PtP)w(0)w(t)-w^{\infty}=(P^{t}-P_{\infty})w(0) (21)

Using the properties of PP discussed above we get (PP)t=PtP(P-P_{\infty})^{t}=P^{t}-P_{\infty}, for all t1t\geq 1. Then from (21) we have

w(t)w=(PtP)w(0)=(PP)tw(0)w(t)-w^{\infty}=(P^{t}-P_{\infty})w(0)=(P-P_{\infty})^{t}w(0)

So by Theorem 8.3 in [20], there exists c>0,η<1c>0,\eta<1 such that for all t0t\geq 0

w(t)wcηtw(0)\|w(t)-w^{\infty}\|\leq c\eta^{t}\|w(0)\| (22)

Using (10) and (11), and applying the properties of sub-multiplicativity of spectral norm we have

L(t)LDoutAW(t)W\|L(t)-L_{\infty}\|\leq\|D^{\text{out}}-A\|\|W(t)-W_{\infty}\| (23)

Now W(t),WW(t),W_{\infty} are diagonal matrices and for any diagonal matrix M=diag(m1,,mN)M=\text{diag}(m_{1},\ldots,m_{N}) we have Mmm||M||\leq||m||_{\infty}\leq||m||. Applying this to (23) and using the result from (22) we get

L(t)LCLηt\|L(t)-L_{\infty}\|\leq C_{L}\eta^{t} (24)

where CL=cDoutAw(0)>0C_{L}=c\|D^{\text{out}}-A\|\|w(0)\|>0.
From (10) using the sub-multiplicativity property of the norm and the result from (24) we get

L(t)𝟏=(L(t)L)𝟏NCLηt\|L(t)\mathbf{1}\|=\|(L(t)-L_{\infty})\mathbf{1}\|\leq\sqrt{N}C_{L}\eta^{t} (25)

By choosing CNCLC\geq\sqrt{N}C_{L} and applying to (24) and (25) we get :

L(t)LCηt , L(t)𝟏Cηt\|L(t)-L_{\infty}\|\leq C\eta^{t}\text{ , }\|L(t)\mathbf{1}\|\leq C\eta^{t}

7

Here we introduce some intermediate lemmas which will be useful in the proof of Theorem-1. At first, before proceeding to a time-varying system in two variables, we first analyse the dynamics of a scalar time-varying system. Consider a linear scalar time-varying system -

vt+1=(1r1(t))vt+r2(t)v_{t+1}=\big{(}1-r_{1}(t)\big{)}v_{t}+r_{2}(t) (26)

where

r1(t)=c1(1+t)δ1,r2(t)=c2(1+t)δ2r_{1}(t)=\frac{c_{1}}{(1+t)^{\delta_{1}}},r_{2}(t)=\frac{c_{2}}{(1+t)^{\delta_{2}}} (27)

where c1,c2,δ2c_{1},c_{2},\delta_{2} are positive constants, and 0δ110\leq\delta_{1}\leq 1.

The following result is based upon the results introduced in Lemma 25 in [21] and Lemma 3 in [15]. It provides a relation between δ1\delta_{1} and δ2\delta_{2} under which the dynamics of the scalar time-varying system in (26) is bounded. It also gives the condition under which the system dynamics converges to zero, and the corresponding rate of convergence.

Proposition 7.3.

Consider the system given in (26) where r1(t),r2(t)r_{1}(t),r_{2}(t) is given by (27). Then if δ1=δ2\delta_{1}=\delta_{2}, there exists B>0B>0, such that for sufficiently large non-negative integers j<tj<t ,

0k=jt1(l=k+1t1(1r1(l)))r2(k)B0\leq\sum_{k=j}^{t-1}\left(\prod_{l=k+1}^{t-1}\big{(}1-r_{1}(l)\big{)}\right)r_{2}(k)\leq B

Moreover the constant BB can be chosen independently of t,jt,j. Also, if δ2>δ1\delta_{2}>\delta_{1}, then for arbitrary fixed jj,

limtk=jt1(l=k+1t1(1r1(l)))r2(k)=0\lim_{t\rightarrow\infty}\sum_{k=j}^{t-1}\left(\prod_{l=k+1}^{t-1}\big{(}1-r_{1}(l)\big{)}\right)r_{2}(k)=0

and correspondingly limt(t+1)δ0vt=0\lim_{t\rightarrow\infty}(t+1)^{\delta_{0}}v_{t}=0 for all 0δ0<δ2δ10\leq\delta_{0}<\delta_{2}-\delta_{1}, and for all initial conditions v0v_{0}.

The following result provides the rate of convergence of a scalar system modified from (26).

Proposition 7.4 (Lemma 4 in [15]).

Consider the scalar time-varying linear system :

vt+1=(1c3r2(t)+c4r1(t))vt+c5r1(t)v_{t+1}=\big{(}1-c_{3}r_{2}(t)+c_{4}r_{1}(t)\big{)}v_{t}+c_{5}r_{1}(t) (28)

where r1(t),r2(t)r_{1}(t),r_{2}(t) are given by :

r1(t)=c1(1+t)δ1,r2(t)=c2(1+t)δ2r_{1}(t)=\frac{c_{1}}{(1+t)^{\delta_{1}}},r_{2}(t)=\frac{c_{2}}{(1+t)^{\delta_{2}}}

where c1,c2,,c5>0c_{1},c_{2},\ldots,c_{5}>0, and 0<δ2<δ1<10<\delta_{2}<\delta_{1}<1.
The system in (28) satisfies limt(t+1)δ0vt=0\lim_{t\rightarrow\infty}(t+1)^{\delta_{0}}v_{t}=0 for all 0δ0<δ1δ20\leq\delta_{0}<\delta_{1}-\delta_{2}, and for all initial conditions v0v_{0}.

Now by using the results above we introduce the following lemma which proves the convergence of γ1(t)\gamma_{1}(t) and γ2(t)\gamma_{2}(t) introduced in (17) and (18).

Lemma 7.5.

The system in (17) and (18) satisfies

limt(t+1)δ0γ1(t)=0\lim_{t\rightarrow\infty}(t+1)^{\delta_{0}}\gamma_{1}(t)=0 (29)
limt(t+1)δ0γ2(t)=0\lim_{t\rightarrow\infty}(t+1)^{\delta_{0}}\gamma_{2}(t)=0 (30)

where 0δ0<α1μ10\leq\delta_{0}<\alpha_{1}-\mu_{1}

Proof 7.6.

Step 1 : As α(t),μ(t)\alpha(t),\mu(t) are decreasing in tt, and α1>μ1\alpha_{1}>\mu_{1} , there exists a finite T>0T>0 such that for all t>Tt>T

01(12s)α(t)101c1μ(t)+(1+N)α(t)1\begin{split}0&\leq 1-(1-2s)\alpha(t)\leq 1\\ 0&\leq 1-c_{1}\mu(t)+(1+\sqrt{N})\alpha(t)\leq 1\end{split} (31)

From (18) we can express γ2(t)\gamma_{2}(t) as

γ2(t)=τ=Tt1(1(12s)α(τ))γ2(T)+τ=Tt1(j=τ+1t1(1(12s)α(j)))(α(τ)γ1(τ)+1(1+τ)θ1).\gamma_{2}(t)=\prod_{\tau=T}^{t-1}(1-(1-2s)\alpha(\tau))\gamma_{2}(T)+\\ \sum_{\tau=T}^{t-1}\left(\prod_{j=\tau+1}^{t-1}(1-(1-2s)\alpha(j))\right)\big{(}\alpha(\tau)\gamma_{1}(\tau)+\frac{1}{(1+\tau)^{\theta_{1}}}\big{)}. (32)

Let s(t):=τ=Tt1(j=τ+1t1(1(12s)α(j)))1(1+τ)θ1s(t):=\sum_{\tau=T}^{t-1}\left(\prod_{j=\tau+1}^{t-1}(1-(1-2s)\alpha(j))\right)\frac{1}{(1+\tau)^{\theta_{1}}}. Using the second part of Proposition 7.3, we obtain : s(t)0s(t)\rightarrow 0 as tt\rightarrow\infty. Hence, T>0\exists T>0 such that |s(t)|γ1(T)|s(t)|\leq\gamma_{1}(T) tT\forall t\geq T. Using this, along with (31), in (32) provides

|γ2(t)||γ2(T)|+σ1supl[T,t]|γ1(l)||\gamma_{2}(t)|\leq|\gamma_{2}(T)|+\sigma_{1}\sup_{l\in[T,t]}|\gamma_{1}(l)| (33)

for some constant σ1>0\sigma_{1}>0.

Step 2 : From (17) and (31) we have

|γ1(t+1)|(1c1μ(t)+(1+N)α(t))supl[T,t]|γ1(l)|+(1+N)α(t)|γ2(t)|+c2ηt.\begin{split}|\gamma_{1}(t+1)|&\leq\big{(}1-c_{1}\mu(t)+(1+\sqrt{N})\alpha(t)\big{)}\sup_{l\in[T,t]}|\gamma_{1}(l)|\\ &+(1+\sqrt{N})\alpha(t)|\gamma_{2}(t)|+c_{2}\eta^{t}.\end{split} (34)

Applying (33) we get

|γ1(t+1)|(1c1μ(t)+σ2α(t))supl[T,t]|γ1(l)|+σ3α(t)+c2ηt\begin{split}\therefore|\gamma_{1}(t+1)|&\leq\big{(}1-c_{1}\mu(t)+\sigma_{2}\alpha(t)\big{)}\sup_{l\in[T,t]}|\gamma_{1}(l)|\\ &+\sigma_{3}\alpha(t)+c_{2}\eta^{t}\end{split}

where σ2=(1+N)(1+σ1)\sigma_{2}=(1+\sqrt{N})(1+\sigma_{1}) and σ3=(1+N)|γ2(T)|\sigma_{3}=(1+\sqrt{N})|\gamma_{2}(T)|. Now as 0<η<10<\eta<1, there exists σ4>0\sigma_{4}>0 such that for all t>0t>0

σ3α(t)+c2ηt<σ4α(t)\sigma_{3}\alpha(t)+c_{2}\eta^{t}<\sigma_{4}\alpha(t) (35)
|γ1(t+1)|(1c1μ(t)+σ2α(t))supl[T,t]|γ1(l)|+σ4α(t)\therefore|\gamma_{1}(t+1)|\leq\big{(}1-c_{1}\mu(t)+\sigma_{2}\alpha(t)\big{)}\sup_{l\in[T,t]}|\gamma_{1}(l)|+\sigma_{4}\alpha(t)

We define a new system -

m(t+1)=max(m(t),(1c1μ(t)+σ2α(t))m(t)+σ4α(t))m(t+1)=\textrm{max}(m(t),\big{(}1-c_{1}\mu(t)+\sigma_{2}\alpha(t)\big{)}m(t)+\sigma_{4}\alpha(t)) (36)

for all t>Tt>T and initial condition m(T)=γ1(T)m(T)=\gamma_{1}(T). So by definition of m(t)m(t) we have :

m(t)supl[T,t]|γ1(l)|m(t)\geq\sup_{l\in[T,t]}|\gamma_{1}(l)| (37)

We define another new system :

m~(t+1)=(1c1μ(t)+σ2α(t))m~(t)+σ4α(t))\tilde{m}(t+1)=\big{(}1-c_{1}\mu(t)+\sigma_{2}\alpha(t)\big{)}\tilde{m}(t)+\sigma_{4}\alpha(t)) (38)

for all t>Tt>T and initial condition m~(T)=m(T)=γ1(T)\tilde{m}(T)=m(T)=\gamma_{1}(T). By definition m~(T)0\tilde{m}(T)\geq 0. Also for t>Tt>T, from (31) we have 1c1μ(t)+σ2α(t)01-c_{1}\mu(t)+\sigma_{2}\alpha(t)\geq 0. Then m~(t)0\tilde{m}(t)\geq 0 for all tTt\geq T. Now using Proposition 7.3 and (38) we have

limtm~(t)=0\lim_{t\rightarrow\infty}\tilde{m}(t)=0

Step 3 : By virtue of m~(t)\tilde{m}(t) being a non-negative sequence which converges to 0, there exists a time T1TT_{1}\geq T such that m~(T1+1)m~(T1)\tilde{m}(T_{1}+1)\leq\tilde{m}(T_{1}). We choose the smallest value among all such possible T1TT_{1}\geq T. Then from the definition of T1T_{1} we have m~(T)<m~(T+1)<<m~(T1)\tilde{m}(T)<\tilde{m}(T+1)<\ldots<\tilde{m}(T_{1}). So from (36), m(t)=m~(t)m(t)=\tilde{m}(t) for all t[T,T1]t\in[T,T_{1}].

m(t)m(T1) , for all t[T,T1]\therefore m(t)\leq m(T_{1})\text{ , for all }t\in[T,T_{1}] (39)

Also by definition of T1,m(t)T_{1},m(t) we have m(T1+1)=m(T1)m(T_{1}+1)=m(T_{1}).
Let for all tT1t\geq T_{1}

π(t)m(T1)(1c1μ(t)+σ2α(t))m(T1)σ4α(t)\pi(t)\coloneqq m(T_{1})-\big{(}1-c_{1}\mu(t)+\sigma_{2}\alpha(t)\big{)}m(T_{1})-\sigma_{4}\alpha(t)

By algebraic manipulation

π(t)=(σ5(t+1)μ1σ6(t+1)α1)m(T1)\pi(t)=\left(\frac{\sigma_{5}}{(t+1)^{\mu_{1}}}-\frac{\sigma_{6}}{(t+1)^{\alpha_{1}}}\right)m(T_{1})

where σ5=c1μ0>0\sigma_{5}=c_{1}\mu_{0}>0 , σ6=(σ2+σ4m(T1))α0>0\sigma_{6}=\left(\sigma_{2}+\frac{\sigma_{4}}{m(T_{1})}\right)\alpha_{0}>0.
Now m(T1)=0m(T_{1})=0, and since m(T1+1)=m(T1)m(T_{1}+1)=m(T_{1}) we have

π(T1)0T1(σ6σ5)1/(α1μ1)1\pi(T_{1})\geq 0\iff T_{1}\geq\left(\frac{\sigma_{6}}{\sigma_{5}}\right)^{1/(\alpha_{1}-\mu_{1})}-1

So we have π(t)0\pi(t)\geq 0 for all tT1t\geq T_{1}. Then using (36) we have

m(t)=m(T1) , for all tT1m(t)=m(T_{1})\text{ , for all }t\geq T_{1} (40)

Now combining the results from (37), (39) and (40) we get

supt0|γ1(t)|<\sup_{t\geq 0}|\gamma_{1}(t)|<\infty (41)

Step 4 : Let supt[T,t]|γ1(t)|=B1<\sup_{t\in[T,t]}|\gamma_{1}(t)|=B_{1}<\infty. Then from (33) we have

suptT|γ2(t)||γ2(T)|+σ1B1<\sup_{t\geq T}|\gamma_{2}(t)|\leq|\gamma_{2}(T)|+\sigma_{1}B_{1}<\infty (42)

As T<T<\infty, we have

supt[0,T]|γ2(t)|<\sup_{t\in[0,T]}|\gamma_{2}(t)|<\infty (43)

So combining (42) and (43) we have

supt0|γ2(t)|<\sup_{t\geq 0}|\gamma_{2}(t)|<\infty (44)

Step 5 : Let supt0|γ2(t)|=B2<\sup_{t\geq 0}|\gamma_{2}(t)|=B_{2}<\infty. Then for sufficiently large tt, from (34) we have

|γ1(t+1)||1c1μ(t)+(1+N)α(t)||γ1(t)|+(1+N)α(t)B2+c2ηt\begin{split}|\gamma_{1}(t+1)|&\leq|1-c_{1}\mu(t)+(1+\sqrt{N})\alpha(t)||\gamma_{1}(t)|\\ &+(1+\sqrt{N})\alpha(t)B_{2}+c_{2}\eta^{t}\end{split}

Now as 0<η<10<\eta<1, there exists Cη>0C_{\eta}>0 and Tη>0T_{\eta}>0 such that for all t>Tηt>T_{\eta}

(1+N)B2α(t)+c2ηt<Cηα(t)(1+\sqrt{N})B_{2}\alpha(t)+c_{2}\eta^{t}<C_{\eta}\alpha(t) (45)

For a suitable choice of Cη=σ7C_{\eta}=\sigma_{7}, (45) holds for all t>0t>0.

|γ1(t+1)||1c1μ(t)+(1+N)α(t)||γ1(t)|+σ7α(t)\therefore|\gamma_{1}(t+1)|\leq|1-c_{1}\mu(t)+(1+\sqrt{N})\alpha(t)||\gamma_{1}(t)|+\sigma_{7}\alpha(t) (46)

As (46) falls under the purview of Proposition 7.4, we can infer (29).

Step 6 : As a consequence of Proposition 7.4, there exists R1>0R_{1}>0 such that |γ1(t)|<R1/(t+1)δ0|\gamma_{1}(t)|<R_{1}/(t+1)^{\delta_{0}} for all 0δ0<α1μ10\leq\delta_{0}<\alpha_{1}-\mu_{1}. We choose δ0min{α1μ1,θ1α1}\delta_{0}\leq\min\{\alpha_{1}-\mu_{1},\theta_{1}-\alpha_{1}\}, which ensures (1+t)θ1(1+t)(α1+δ0)(1+t)^{-\theta_{1}}\leq(1+t)^{-(\alpha_{1}+\delta_{0})}. Thus for sufficiently large tt we have -

|γ2(t+1)|(1(12s)α(t))|γ2(t)|+α0R1(t+1)α1+δ0|\gamma_{2}(t+1)|\leq\big{(}1-(1-2s)\alpha(t)\big{)}|\gamma_{2}(t)|+\frac{\alpha_{0}R_{1}}{(t+1)^{\alpha_{1}+\delta_{0}}} (47)

As (47) falls under the purview of Proposition 7.3, we have

limt(t+1)δ0γ2(t)=0\lim_{t\rightarrow\infty}(t+1)^{\delta^{\prime}_{0}}\gamma_{2}(t)=0

for all 0δ0<δ00\leq\delta^{\prime}_{0}<\delta_{0}.
By making δ0\delta_{0} arbitrarily close to α1μ1\alpha_{1}-\mu_{1} we get (30).

Let us define a new matrix JJ as JI1N𝟏𝟏TJ\coloneqq I-\frac{1}{N}\mathbf{1}\mathbf{1}^{T}.
The following lemma provides a bound for Jβ(t)L\|J-\beta(t)L^{\infty}\|.

Lemma 7.7.

Given c1>0c_{1}>0, L=(DoutA)WL_{\infty}=\big{(}D^{\text{out}}-A\big{)}W_{\infty} where W=diag(w)W_{\infty}=\text{diag}\big{(}w^{\infty}\big{)}, β(t)=β0(1+t)β1\beta(t)=\frac{\beta_{0}}{(1+t)^{\beta_{1}}}, μ(t)=μ0(1+t)μ1\mu(t)=\frac{\mu_{0}}{(1+t)^{\mu_{1}}} where 0<β0<ψ0<\beta_{0}<\psi, μ0>0\mu_{0}>0 and 0<β1<μ1<10<\beta_{1}<\mu_{1}<1, there exists T>0T>0 such that Jβ(t)L1c1μ(t)<1\|J-\beta(t)L_{\infty}\|\leq 1-c_{1}\mu(t)<1 for all tTt\geq T.

Proof 7.8.

Using the property 𝟏TL=0\mathbf{1}^{T}L_{\infty}=0 we can write

Jβ(t)L2=λmax(Jβ(t)M2+β2(t)M3)=supxN,x=1xT(Jβ(t)M2+β2(t)M3)x\begin{split}&\|J-\beta(t)L_{\infty}\|^{2}=\lambda_{\text{max}}\left(J-\beta(t)M_{2}+\beta^{2}(t)M_{3}\right)\\ &=\sup_{x\in\mathbb{R}^{N},||x||=1}x^{T}\left(J-\beta(t)M_{2}+\beta^{2}(t)M_{3}\right)x\end{split} (48)

where M2=(LT+L)M_{2}=\big{(}L_{\infty}^{T}+L_{\infty}\big{)} and M3=LTLM_{3}=L_{\infty}^{T}L_{\infty}. Now by definition L𝟏=0L_{\infty}\mathbf{1}=0. So we have

M2𝟏=0 ; M3𝟏=0M_{2}\mathbf{1}=0\text{ ; }M_{3}\mathbf{1}=0 (49)

Also, M2M_{2} and M3M_{3} are the Laplacians of the corresponding graph. As the graphs are strongly connected, we can infer :

λ2(M2)>0 ; λ2(M3)>0\lambda_{2}\big{(}M_{2}\big{)}>0\text{ ; }\lambda_{2}\big{(}M_{3}\big{)}>0

i.e the 2nd lowest eigen-value of each of the Laplacians is strictly positive.
Let xspan{𝟏}x=α𝟏,αx\in\text{span}\{\mathbf{1}\}\equiv x=\alpha\mathbf{1},\alpha\in\mathbb{R}. Then using (49) we have

xT(I1N𝟏𝟏Tβ(t)M2+β2(t)M3)x=0x^{T}\left(I-\frac{1}{N}\mathbf{1}\mathbf{1}^{T}-\beta(t)M_{2}+\beta^{2}(t)M_{3}\right)x=0 (50)

Now suppose xspan{𝟏}xT𝟏=0x\in\text{span}\{\mathbf{1}^{\perp}\}\equiv x^{T}\mathbf{1}=0. Then we have

xT(I1N𝟏𝟏Tβ(t)M2+β2(t)M3)x(1(β(t)λmβ2(t)λM))xTx\begin{split}x^{T}\left(I-\frac{1}{N}\mathbf{1}\mathbf{1}^{T}-\beta(t)M_{2}+\beta^{2}(t)M_{3}\right)x\\ \leq\big{(}1-\big{(}\beta(t)\lambda_{m}-\beta^{2}(t)\lambda_{M}\big{)}\big{)}x^{T}x\end{split} (51)

where λm=λ2(M2)\lambda_{m}=\lambda_{2}(M_{2}) and λM=λmax(M3)\lambda_{M}=\lambda_{\text{max}}(M_{3}).
Having β0<ψ\beta_{0}<\psi and β1<1\beta_{1}<1 ensures that β(t)<λm/λM\beta(t)<\lambda_{m}/\lambda_{M}, and in turn I1N𝟏𝟏Tβ(t)L2<1\|I-\frac{1}{N}\mathbf{1}\mathbf{1}^{T}-\beta(t)L_{\infty}\|^{2}<1 t0\forall t\geq 0. We choose an ϵ\epsilon such that 0<ϵλmβ(t)λM0<\epsilon\leq\lambda_{m}-\beta(t)\lambda_{M}. Then for all t0t\geq 0

β(t)λmβ2(t)λMβ(t)ϵ>0\beta(t)\lambda_{m}-\beta^{2}(t)\lambda_{M}\geq\beta(t)\epsilon>0 (52)

Then from (48), (50), (51) and (52) we have

Jβ(t)L1β(t)ϵ<1\|J-\beta(t)L_{\infty}\|\leq\sqrt{1-\beta(t)\epsilon}<1 (53)

Now, as μ1>β1\mu_{1}>\beta_{1}, there exists time T>0T>0 such that for all t>Tt>T

1(1+t)μ1β1ϵβ02c1μ02c1μ(t)ϵβ(t)1ϵβ(t)12c1μ(t)+c12μ2(t)\begin{split}\frac{1}{(1+t)^{\mu_{1}-\beta_{1}}}\leq\frac{\epsilon\beta_{0}}{2c_{1}\mu_{0}}\implies 2c_{1}\mu(t)\leq\epsilon\beta(t)\\ \implies 1-\epsilon\beta(t)\leq 1-2c_{1}\mu(t)+c_{1}^{2}\mu^{2}(t)\end{split}
1ϵβ(t)1c1μ(t)\therefore\sqrt{1-\epsilon\beta(t)}\leq 1-c_{1}\mu(t) (54)

Also as c1>0c_{1}>0, we have 1c1μ(t)<11-c_{1}\mu(t)<1. Then from (53) and (54) we have

Jβ(t)L1c1μ(t)<1 , tT\|J-\beta(t)L_{\infty}\|\leq 1-c_{1}\mu(t)<1\text{ , }t\geq T

8

Proof 8.9 (Proof of Theorem 1).

Let x¯(t)1×M\bar{x}(t)\in\mathbb{R}^{1\times M} denote the average of the states of the agents : x¯(t)=(1/N)𝟏Tx(t)\bar{x}(t)=(1/N)\mathbf{1}^{T}x(t).

We define p(t)p(t) as the difference between the state of the agents and their average, and q(t)q(t) as the difference between the average value and the unknown parameter θ(t)\theta^{*}(t).

p(t)\displaystyle p(t) x(t)𝟏x¯(t),p(t)N×M\displaystyle\coloneqq x(t)-\mathbf{1}\bar{x}(t),p(t)\in\mathbb{R}^{N\times M} (55)
q(t)\displaystyle q(t) x¯T(t)θ(t),q(t)M.\displaystyle\coloneqq\bar{x}^{T}(t)-\theta^{*}(t),q(t)\in\mathbb{R}^{M}. (56)

Now the norm of the difference between the state of each agent and θ(t)\theta^{*}(t) can be upper bounded as

xi(t)θ(t)xi(t)x¯T(t)+x¯T(t)θ(t)xi(t)θ(t)p(t)+q(t) for all i𝒱.\begin{split}\|x_{i}(t)-\theta^{*}(t)\|&\leq\|x_{i}(t)-\bar{x}^{T}(t)\|+\|\bar{x}^{T}(t)-\theta^{*}(t)\|\\ \implies\|x_{i}(t)-\theta^{*}(t)\|&\leq\|p(t)\|+\|q(t)\|\text{ for all }i\in\mathcal{V}.\end{split} (57)

To show that the states converge to θ(t)\theta^{*}(t), we express xi(t)θ(t)γ(t)\|x_{i}(t)-\theta^{*}(t)\|\leq\gamma(t) and show that γ(t)\gamma(t) has a dynamics that converges to 0. In what follows, we firstly we use the method of induction to show that

p(t)γ1(t) and q(t)γ2(t) for all t0\|p(t)\|\leq\gamma_{1}(t)\text{ and }\|q(t)\|\leq\gamma_{2}(t)\text{ for all }t\geq 0 (58)

and in the process also define the dynamics of γ1(t)\gamma_{1}(t) and γ2(t)\gamma_{2}(t). After that we express these dynamics as a linear time-varying system which is asymptotically stable. From there, using (57), (16) and (58) we arrive at our desired result.

Dynamics of x(t)1x¯(t)x(t)-\mathbf{1}\bar{x}(t) :

p(t+1)=x(t+1)𝟏x¯(t+1)=Jx(t+1)p(t+1)=x(t+1)-\mathbf{1}\bar{x}(t+1)=Jx(t+1)

p(t+1)=M1+α(t)JK(t)(y(t)x(t))\implies p(t+1)=M_{1}+\alpha(t)JK(t)(y(t)-x(t)) (59)

where JI1N𝟏𝟏TJ\coloneqq I-\frac{1}{N}\mathbf{1}\mathbf{1}^{T}, and M1J(Iβ(t)L(t))x(t)M_{1}\coloneqq J(I-\beta(t)L(t))x(t). Expanding M1M_{1} and using 𝟏TL(t)=0\mathbf{1}^{T}L(t)=0 and J𝟏=0J\mathbf{1}=0, followed by applying norm and its properties of triangle-inequality and sub-multiplicativity we get

M1\displaystyle\|M_{1}\| (Jβ(t)L)p(t)+β(t)(L(t)𝟏q(t)\displaystyle\leq\|(J-\beta(t)L_{\infty})\|\|p(t)\|+\beta(t)\big{(}\|L(t)\mathbf{1}\|\|q(t)\|
+(L(t)L)p(t)+L(t)𝟏θ(t))\displaystyle+\|(L(t)-L_{\infty})\|\|p(t)\|+\|L(t)\mathbf{1}\|\|\theta^{*}(t)\|\big{)} (60)

Now applying Lemmas 1 and 7.7, and Assumption-2 in (8.9) :

M1(1c1μ(t))p(t)+Cβ(t)ηt(p(t)+q(t)+1+Θ).\|M_{1}\|\leq(1-c_{1}\mu(t))\|p(t)\|+C\beta(t)\eta^{t}\big{(}\|p(t)\|+\|q(t)\|+1+\Theta\big{)}. (61)

Applying norm to (59) and using (19), J=𝟏\|J\|=\mathbf{1} we get

p(t+1)\displaystyle\|p(t+1)\| (1c1μ(t)+Cβ(t)ηt)p(t)\displaystyle\leq(1-c_{1}\mu(t)+C\beta(t)\eta^{t})\|p(t)\| (62)
+Cβ(t)ηt(1+Θ+q(t))+Nα(t)γ(t).\displaystyle+C\beta(t)\eta^{t}\big{(}1+\Theta+\|q(t)\|\big{)}+\sqrt{N}\alpha(t)\gamma(t).

Dynamics of x¯T(t)θ(t)\bar{x}^{T}(t)-\theta^{*}(t) :

q(t+1)=x¯T(t+1)θ(t+1)=x¯T(t)θ(t)+α(t)N𝟏TK(t)(y(t)x(t))+θ(t+1)θ(t)Δθ(t+1)\begin{split}q(t&+1)=\bar{x}^{T}(t+1)-\theta^{*}(t+1)=\bar{x}^{T}(t)-\theta^{*}(t)\\ &+\frac{\alpha(t)}{N}\mathbf{1}^{T}K(t)(y(t)-x(t))+\underbrace{\theta^{*}(t+1)-\theta^{*}(t)}_{\Delta\theta^{*}(t+1)}\end{split} (63)

We define two diagonal matrices K𝒢(t),K(t)K_{\mathcal{G}}(t),K_{\mathcal{B}}(t) where [K𝒢(t)]iiki(t)[K_{\mathcal{G}}(t)]_{ii}\coloneqq k_{i}(t) if i𝒢ti\in\mathcal{G}_{t} and [K(t)]iiki(t)[K_{\mathcal{B}}(t)]_{ii}\coloneqq k_{i}(t) if iti\in\mathcal{B}_{t}, and rest of the entries are equal to 0.

K(t)+K𝒢(t)=K(t) for all t0\therefore K_{\mathcal{B}}(t)+K_{\mathcal{G}}(t)=K(t)\text{ for all }t\geq 0 (64)

Using (1) and (55) in (63), followed by applying the l2l_{2}-norm and its properties of triangle-inequality and sub-multiplicativity we get

q(t+1)1\displaystyle\|q(t+1)\|\leq\|1 α(t)Ni𝒢tki(t)q(t)+Δθ(t+1)\displaystyle-\frac{\alpha(t)}{N}\sum_{i\in\mathcal{G}_{t}}k_{i}(t)\|\|q(t)\|+\|\Delta\theta^{*}(t+1)\|
+α(t)Ni𝒢tki(t)(xi(t)x¯T(t))\displaystyle+\frac{\alpha(t)}{N}\|\sum_{i\in\mathcal{G}_{t}}k_{i}(t)(x_{i}(t)-\bar{x}^{T}(t))\|
+α(t)Nitki(t)(yi(t)xi(t))\displaystyle+\frac{\alpha(t)}{N}\|\sum_{i\in\mathcal{B}_{t}}k_{i}(t)(y_{i}(t)-x_{i}(t))\| (65)

Dynamics of γ1(t)\gamma_{1}(t) and γ2(t)\gamma_{2}(t) via method of Induction :
By the method of induction we wish to show (58), and in the process arrive at the dynamics of γ1(t)\gamma_{1}(t) and γ2(t)\gamma_{2}(t).
Step 1 : at t=0t=0, p(0)=0\|p(0)\|=0 as x(0)=𝟎x(0)=\mathbf{0}, and q(0)=Θ\|q(0)\|=\Theta as θ(t)<Θ\|\theta^{*}(t)\|<\Theta. Choosing γ1(0)=0\gamma_{1}(0)=0, γ2(0)=Θ\gamma_{2}(0)=\Theta we have

p(0)γ1(0) , q(0)γ2(0)\|p(0)\|\leq\gamma_{1}(0)\text{ , }\|q(0)\|\leq\gamma_{2}(0) (66)

Step 2 : for some t>0t>0 we assume that

p(t)γ1(t) , q(t)γ2(t)\|p(t)\|\leq\gamma_{1}(t)\text{ , }\|q(t)\|\leq\gamma_{2}(t) (67)

Step 3 : based on the assumption (67) from Step-2, we need to show that

p(t+1)γ1(t+1) , q(t+1)γ2(t+1)\|p(t+1)\|\leq\gamma_{1}(t+1)\text{ , }\|q(t+1)\|\leq\gamma_{2}(t+1) (68)

Applying (67) to (62) and using (16) we have

p(t\displaystyle\|p(t +1)(1c1μ(t)+Cβ(t)ηt+Nα(t))γ1(t)\displaystyle+1)\|\leq(1-c_{1}\mu(t)+C\beta(t)\eta^{t}+\sqrt{N}\alpha(t))\gamma_{1}(t)
+(Cβ(t)ηt+Nα(t))γ2(t)+C(1+Θ)β(t)ηt\displaystyle+(C\beta(t)\eta^{t}+\sqrt{N}\alpha(t))\gamma_{2}(t)+C(1+\Theta)\beta(t)\eta^{t}

Now as η<1\eta<1 , there exists c2>0c_{2}>0 and T>0T>0 such that for all t>Tt>T

C(1+Θ)β(t)ηtc2ηt , and Cβ(t)ηtα(t)C(1+\Theta)\beta(t)\eta^{t}\leq c_{2}\eta^{t}\text{ , and }C\beta(t)\eta^{t}\leq\alpha(t) (69)

By appropriate choice of β0<α0C\beta_{0}<\frac{\alpha_{0}}{C}, c2>C(1+Θ)β0c_{2}>C(1+\Theta)\beta_{0} and μ0<(λmβ0λM)β0/(2c1)\mu_{0}<(\lambda_{m}-\beta_{0}\lambda_{M})\beta_{0}/(2c_{1}), (69) and (54) holds for all t>0t>0

p(t+1)(1c1μ(t)+(1+N)α(t))γ1(t)+(1+N)α(t)γ2(t)+c2ηt\begin{split}\therefore\|p(t+1)\|\leq(1-c_{1}\mu(t)+(1+\sqrt{N})\alpha(t))\gamma_{1}(t)\\ +(1+\sqrt{N})\alpha(t)\gamma_{2}(t)+c_{2}\eta^{t}\end{split} (70)

We define the dynamics of γ1(t)\gamma_{1}(t) as -

γ1(t+1)(1c1μ(t)+(1+N)α(t))γ1(t)+(1+N)α(t)γ2(t)+c2ηt\begin{split}\gamma_{1}(t+1)\coloneqq(1-c_{1}\mu(t)+(1+\sqrt{N})\alpha(t))\gamma_{1}(t)\\ +(1+\sqrt{N})\alpha(t)\gamma_{2}(t)+c_{2}\eta^{t}\end{split} (71)

Using (16), (57), and (67)

ki(t)=1 for all i𝒢t [yi(t)=θ(t)i𝒢t]k_{i}(t)=1\text{ for all }i\in\mathcal{G}_{t}\text{ [}\because y_{i}(t)=\theta^{*}(t)\forall i\in\mathcal{G}_{t}] (72)

Now from (64), (8.9), (72) and further using (16), (3) and α0<1/(12s)\alpha_{0}<1/(1-2s), s<1/2s<1/2 we get

q(t+1)(1α(t)(12s))γ2(t)+α(t)γ1(t)+1/(1+t)θ1\|q(t+1)\|\leq(1-\alpha(t)(1-2s))\gamma_{2}(t)+\alpha(t)\gamma_{1}(t)+1/(1+t)^{\theta_{1}} (73)

We define the dynamics of γ2(t)\gamma_{2}(t) as

γ2(t+1)α(t)γ1(t)+(1α(t)(12s))γ2(t)+1/(1+t)θ1\gamma_{2}(t+1)\coloneqq\alpha(t)\gamma_{1}(t)+(1-\alpha(t)(1-2s))\gamma_{2}(t)+1/(1+t)^{\theta_{1}} (74)

Then from (70), (71) and (73), (74) we can infer (68). Thus from steps 1,2 and 3 we have (58).

Asymptotic stability of γ1(t)\gamma_{1}(t) and γ2(t)\gamma_{2}(t) :
Using Lemma 7.5 we can say that a linear time-varying system with state-variables γ1(t)\gamma_{1}(t) and γ2(t)\gamma_{2}(t), and state dynamics given by (71) and (74) respectively, is asymptotically stable, i.e
limt(t+1)δ0γ1(t)=0\lim_{t\rightarrow\infty}(t+1)^{\delta_{0}}\gamma_{1}(t)=0, limt(t+1)δ0γ2(t)=0\lim_{t\rightarrow\infty}(t+1)^{\delta_{0}}\gamma_{2}(t)=0.

9

For our REWB algorithm, the value of a constant ψ\psi, which is an upper bound to the parameter β0\beta_{0}, is defined as ψ:=2/(Ndmaxin(dmaxin+dmaxout))\psi:=2/(Nd^{\text{in}}_{\text{max}}(d^{\text{in}}_{\text{max}}+d^{\text{out}}_{\text{max}})). Here we provide a detailed proof of how we arrived at this value of ψ\psi.

We have 0<β0<λ2(LT+L)/λmax(LTL)0<\beta_{0}<\lambda_{2}(L_{\infty}^{T}+L_{\infty})/\lambda_{\text{max}}(L_{\infty}^{T}L_{\infty}). Now using Gershgorin’s Disk Theorem we can write :

λmax\displaystyle\lambda_{\text{max}} (LTL)2maxi[LTL]ii\displaystyle(L_{\infty}^{T}L_{\infty})\leq 2\max_{i}[L_{\infty}^{T}L_{\infty}]_{ii}
=2maxi([L]ii2+j𝒩i,ji[L]ij2)\displaystyle=2\max_{i}\big{(}[L_{\infty}]^{2}_{ii}+\sum_{j\in\mathcal{N}_{i},j\neq i}[L_{\infty}]^{2}_{ij}\big{)}
=2maxi((dioutwi)2+j𝒩i,ji(wj)2)\displaystyle=2\max_{i}\big{(}(d_{i}^{\text{out}}w_{i}^{\infty})^{2}+\sum_{j\in\mathcal{N}_{i},j\neq i}(w_{j}^{\infty})^{2}\big{)}

Now using dioutwi1d_{i}^{\text{out}}w_{i}^{\infty}\leq 1 from [17], and 1diout1\frac{1}{d_{i}^{\text{out}}}\leq 1, we have :

λmax(LTL)2maxi(1+j𝒩i,ji(1diout)2)2dmaxin\lambda_{\text{max}}(L_{\infty}^{T}L_{\infty})\leq 2\max_{i}\big{(}1+\sum_{j\in\mathcal{N}_{i},j\neq i}(\frac{1}{d_{i}^{\text{out}}})^{2}\big{)}\leq 2d_{\text{max}}^{\text{in}}

From the results in [22], we can say : λ2(LT+L)>4N(dmaxin+dmaxout)\lambda_{2}(L_{\infty}^{T}+L_{\infty})>\frac{4}{N(d^{\text{in}}_{\text{max}}+d^{\text{out}}_{\text{max}})}.

2Ndmaxin(dmaxin+dmaxout)<λ2(LT+L)λmax(LTL)\therefore\frac{2}{Nd^{\text{in}}_{\text{max}}(d^{\text{in}}_{\text{max}}+d^{\text{out}}_{\text{max}})}<\frac{\lambda_{2}(L_{\infty}^{T}+L_{\infty})}{\lambda_{\text{max}}(L_{\infty}^{T}L_{\infty})}

So defining ψ:=2/(Ndmaxin(dmaxin+dmaxout))\psi:=2/(Nd^{\text{in}}_{\text{max}}(d^{\text{in}}_{\text{max}}+d^{\text{out}}_{\text{max}})) and choosing any non-zero positive value of β0<ψ\beta_{0}<\psi satisfies the requirement of our REWB algorithm.

10

In the proof for Lemma 7.7 in Appendix 7, we use the fact that the second eigenvalues of the matrices M2M_{2} and M3M_{3} are non-zero, where M2=(LT+L)M_{2}=\big{(}L_{\infty}^{T}+L_{\infty}\big{)} and M3=LTLM_{3}=L_{\infty}^{T}L_{\infty}. Here we provide a reasoning for the same.

  • zero column sum : from (11) and (6) we have 𝟏TL=𝟎,L𝟏=𝟎\mathbf{1}^{T}L_{\infty}=\mathbf{0},L_{\infty}\mathbf{1}=\mathbf{0}. Using these, we get

    𝟏TM2=𝟏TLT+𝟏TL=𝟎\displaystyle\mathbf{1}^{T}M_{2}=\mathbf{1}^{T}L_{\infty}^{T}+\mathbf{1}^{T}L_{\infty}=\mathbf{0} ;𝟏TM3=𝟏TLTL=𝟎\displaystyle;\mathbf{1}^{T}M_{3}=\mathbf{1}^{T}L_{\infty}^{T}L_{\infty}=\mathbf{0}
  • positive diagonal elements :

    [M2]ii\displaystyle[M_{2}]_{ii} =[LT]ii+[L]ii=2[L]ii>0\displaystyle=[L_{\infty}^{T}]_{ii}+[L_{\infty}]_{ii}=2[L_{\infty}]_{ii}>0
    [M3]ii\displaystyle[M_{3}]_{ii} =[LT]i:[L]:i=j=1N[L]ji2>0\displaystyle=[L_{\infty}^{T}]_{i:}[L_{\infty}]_{:i}=\sum_{j=1}^{N}[L_{\infty}]_{ji}^{2}>0

    where [M]ii,[M]i:,[M]:i[M]_{ii},[M]_{i:},[M]_{:i} represent the (i,i)(i,i)-th element, ii-th row, and ii-th column of matrix MM respectively.

  • non-diagonal elements in M2M_{2} : [M2]ij=[L]ji+[L]ij[M_{2}]_{ij}=[L_{\infty}]_{ji}+[L_{\infty}]_{ij}.

  • non-diagonal elements in M3M_{3} :

    [M3]ij=[LT]i:[L]:j=k=1N[L]ki[L]kj[M_{3}]_{ij}=[L_{\infty}^{T}]_{i:}[L_{\infty}]_{:j}=\sum_{k=1}^{N}[L_{\infty}]_{ki}[L_{\infty}]_{kj}

From the expression of the entries of M2M_{2} and M3M_{3}, one can deduce that these matrices will have a nonzero entry in the ijijth position if the digraph has an edge between ii and jj. This shows that the connectivity of the graph corresponding to LL_{\infty} would be preserved in the new graph corresponding to M2M_{2} and M3M_{3}. Now, as M2M_{2} and M3M_{3} are valid Laplacians, and their corresponding graphs are connected, we can infer that the second eigenvalues of M2M_{2} and M3M_{3} are non-zero.

References

  • [1] Loubna Hamami and Bouchaib Nassereddine “Application of wireless sensor networks in the field of irrigation: A review” In Computers and Electronics in Agriculture 179, 2020, pp. 105782
  • [2] Carol Habib, Abdallah Makhoul, Rony Darazi and Raphaël Couturier “Health risk assessment and decision-making for patient monitoring and decision-support using Wireless Body Sensor Networks” In Information Fusion 47, 2019, pp. 10–22
  • [3] Etimad Fadel et al. “A survey on wireless sensor networks for smart grid” In Computer Communications 71, 2015, pp. 22–33
  • [4] I.. Akyildiz, Weilian Su, Y. Sankarasubramaniam and E. Cayirci “A survey on sensor networks” In IEEE Communications Magazine 40.8, 2002, pp. 102–114
  • [5] I.F. Akyildiz, W. Su, Y. Sankarasubramaniam and E. Cayirci “Wireless sensor networks: a survey” In Computer Networks 38.4, 2002, pp. 393–422
  • [6] Shaoming He, Hyo-Sang Shin, Shuoyuan Xu and Antonios Tsourdos “Distributed estimation over a low-cost sensor network: A Review of state-of-the-art” In Information Fusion 54, 2020, pp. 21–43
  • [7] Haowen Chan and A. Perrig “Security and privacy in sensor networks” In Computer 36.10, 2003, pp. 103–105
  • [8] Nitin H. Vaidya, Lewis Tseng and Guanfeng Liang “Iterative Approximate Byzantine Consensus in Arbitrary Directed Graphs” In Proceedings of the 2012 ACM Symposium on Principles of Distributed Computing, 2012, pp. 365–374
  • [9] Bo Chen, Li Yu, Daniel W.C. Ho and Wen-An Zhang “Resilient Consensus Through Event-Based Communication” In IEEE Transactions on Control of Network Systems 7.1, 2020, pp. 471–482
  • [10] Yasser Shoukry and Paulo Tabuada “Event-Triggered State Observers for Sparse Sensor Noise/Attacks” In IEEE Transactions on Automatic Control 61.8, 2016, pp. 2079–2091
  • [11] Seyed Mehran Dibaji and Hideaki Ishii “Resilient consensus of second-order agent networks: Asynchronous update rules with delays” In Automatica 81, 2017, pp. 123–132
  • [12] H. Zhang and S. Sundaram “A simple median-based resilient consensus algorithm” In 2012 50th Annual Allerton Conference on Communication, Control, and Computing (Allerton), 2012, pp. 1734–1741
  • [13] Soummya Kar and Jose M.. Moura “Consensus + Innovations Distributed Inference over Networks” In IEEE Signal Processing Magazine, 2013, pp. 99–109
  • [14] Yuan Chen, Soummya Kar and Jose M.. Moura “Resilient Distributed Estimation : Exponential Convergence under Sensor Attacks” In Proceedings for the IEEE Conference on Decision and Control 2018, 2018, pp. 7275–7282
  • [15] Yuan Chen, Soummya Kar and Jose M.. Moura “Resilient Distributed Estimation : Sensor Attacks” In IEEE Transactions on Automatic Control 64.9, 2019, pp. 3772–3779
  • [16] A.. Domínguez-García and C.. Hadjicostis “Distributed strategies for average consensus in directed graphs” In 2011 50th IEEE Conference on Decision and Control and European Control Conference, 2011, pp. 2124–2129
  • [17] A. Makhdoumi and A. Ozdaglar “Graph balancing for distributed subgradient methods over directed graphs” In 2015 54th IEEE Conference on Decision and Control (CDC), 2015, pp. 1364–1371
  • [18] Min Meng, Xiuxian Li and Gaoxi Xiao “Distributed Estimation Under Sensor Attacks: Linear and Nonlinear Measurement Models” In IEEE Transactions on Signal and Information Processing over Networks 7, 2021, pp. 156–165
  • [19] Anomadarshi Barua and Mohammad Abdullah Al Faruque “Special Session: Noninvasive Sensor-Spoofing Attacks on Embedded and Cyber-Physical Systems” In 2020 IEEE 38th International Conference on Computer Design (ICCD), 2020, pp. 45–48
  • [20] João P. Hespanha “Linear Systems Theory” Princeton University Press, 2018
  • [21] Soummya Kar, Jose M.. Moura and Kavita Ramanan “Distributed Parameter Estimation in Sensor Networks: Nonlinear Observation Models and Imperfect Communication” In IEEE Transactions on Information Theory 58.6, 2012, pp. 3575–3605
  • [22] Bojan Mohar “Eigenvalues, diameter, and mean distance in graphs” In Graphs and Combinatorics 7, 1991