This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Inferring Topology of Networked Dynamical Systems by Active Excitations

Yushan Li, Jianping He, Cailian Chen and Xinping Guan The authors are with Dept. of Automation, Shanghai Jiao Tong University, Key Laboratory of System Control and Information Processing, Ministry of Education of China, and Shanghai Engineering Research Center of Intelligent Control and Management, Shanghai, Chin. E-mail address: {yushan_li, jphe, cailianchen, xpguan}@sjtu.edu.cn.
Abstract

Topology inference for networked dynamical systems (NDSs) has received considerable attention in recent years. The majority of pioneering works have dealt with inferring the topology from abundant observations of NDSs, so as to approximate the real one asymptotically. Leveraging the characteristic that NDSs will react to various disturbances and the disturbance’s influence will consistently spread, this paper focuses on inferring the topology by a few active excitations. The key challenge is to distinguish different influences of system noises and excitations from the exhibited state deviations, where the influences will decay with time and the exciatation cannot be arbitrarily large. To practice, we propose a one-shot excitation based inference method to infer hh-hop neighbors of a node. The excitation conditions for accurate one-hop neighbor inference are first derived with probability guarantees. Then, we extend the results to hh-hop neighbor inference and multiple excitations cases, providing the explicit relationships between the inference accuracy and excitation magnitude. Specifically, the excitation based inference method is not only suitable for scenarios where abundant observations are unavailable, but also can be leveraged as auxiliary means to improve the accuracy of existing methods. Simulations are conducted to verify the analytical results.

I Introduction

Networked dynamical systems (NDSs) have been extensively used in numerous applications in the last decades, e.g., electric power systems [1], transportation systems [2], and multi-robot systems [3]. The topology of NDSs is fundamental to characterizing interactions between individual nodes and determines the system convergence. Inferring the topology from observations provides insightful interpretability about NDSs and associated task implementations, and has become a hotspot research topic.

In the literature, plenty of works have been developed to address the topology inference problem from different aspects [4]. For instance, in terms of static topology, [5, 6, 7, 8] focus on inferring the causality/dependency relationships between nodes, while [9, 10, 11] reconstruct the topology by finding the most suitable eigenvalues and eigenvectors from the sample covariance matrix. Considering the topology is time-varying by rules, available methods include graphical Lasso-based methods [12] and SEM models [13], which take the varying topology as a sequence of static topologies and infer them, respectively. In addition to the dynamic topology inference, many kernel-based methods are proposed to deal with cases with nonlinear system models [14, 15].

Despite the tremendous advances of the above works, almost all of the approaches are based on a large scale of observations over the systems. In other words, the feasibility lies in digging up the regularity of the dynamical evolution process from the observation sequences, which corresponds to the common intuition that more data make the interpretability better [16]. Unfortunately, when the observations over the NDS are very limited, the aforementioned methods cannot work well. For example, for a linear time-invariant NDS of nn nodes, at least (n+1)(n+1) groups of consecutive global observations are required to obtain a unique least square estimate of the topology matrix. When more observations are not allowable due to some practical limitations, directly inferring the topology from observations will be extremely difficult.

Inspired by the phenomenon that a thrown stone into water will cause waves, we are able to proactively inject inputs into the systems to excite corresponding reaction behaviors, i.e., the injected inputs on one node will spread to other neighbor nodes. Related examples include using Traceroute to probe the routing topology of the Internet [17], or utilizing inverters to probe the electric distribution network [18]. Therefore, it is possible to reveal the underlying topology of NDSs by investigating the relationships between the excitations and reactions [19, 20, 21]. This idea has motivated the study of this paper, where we aim to leverage a few active excitations to do the inference tasks. It is worth noting that if the excitations are allowed to be abundant, then the problem falls into the realm of typical system identification [22, 23], which is not the focus of this paper.

Few excitations indicate small inference costs but incur new challenges. On the one hand, the influence of the excitation is closely coupled with that of stochastic noises, making it hard to directly distinguish their difference. On the other hand, the spreading effect of the excitation will decay with time and the excitation cannot be arbitrarily large, limiting the scope and accuracy of the inferred topology. To address these issues, we introduce the probability measurement to infer the topology from a local node, and demonstrate how to determine whether the information flow between two nodes exists. The main contributions are summarized as follows.

  • We investigate the possibility of inferring the topology of NDSs by a few active excitations, taking both the process and measurement noises into account. Specifically, we utilize hypothesis test to establish criteria of how to determine the connections between nodes from the exhibited state deviations after excitations.

  • Considering the spreading effects of excitations in NDSs, we first propose one-shot excitation based method to infer one-hop neighbors of a single node. Then, we prove the critical excitation condition given tolerable misjudgment probability, providing reliable excitation design guidance.

  • Based on the one-hop inference procedures, we extend the theoretical analysis to multi-hop neighbor inference by one-shot excitation and multiple excitation cases, respectively. The relationship between inference accuracy and excitation magnitude is derived with probability guarantees. Simulations verify our theoretical results.

The proposed inference method by a few active excitations applies to situations where the observations about NDSs are not sufficient. It can also be leveraged as an auxiliary measure to enhance the accuracy of existing methods by large scales of observations, by treating the inferred results as the constraints in counterpart problem modeling. The remainder of this paper is organized as follows. In Section II, some preliminaries of NDSs and problem modeling are presented. The inference method and performance analysis are provided in Section III. Simulation results are shown in Section IV. Finally, Section V concludes the paper.

II Preliminaries and Problem Formulation

II-A Graph Basics and Notations

Let 𝒢=(𝒱,)\mathcal{G}=(\mathcal{V},\mathcal{E}) be a directed graph that models the networked system, where 𝒱={1,,n}\mathcal{V}=\{1,\cdots,n\} is the finite set of nodes and 𝒱×𝒱\mathcal{E}\subseteq\mathcal{V}\times\mathcal{V} is the set of interaction edges. An edge (i,j)(i,j)\in\mathcal{E} indicates that ii will use information from jj. The adjacency matrix A=[aij]n×nA=[a_{ij}]_{n\times n} of 𝒢\mathcal{G} is defined such that aij>0{a}_{ij}\!>\!0 if (i,j)(i,j) exists, and aij=0{a}_{ij}\!=\!0 otherwise. Denote 𝒩i={j𝒱:aij>0}{\mathcal{N}_{i}}=\{j\in\mathcal{V}:a_{ij}>0\} as the in-neighbor set of ii, and di=|𝒩i|d_{i}=\left|{\mathcal{N}_{i}}\right| as its in-degree. Throughout this paper, let 𝟎\bm{0} and 𝟏\bm{1} be all-zero and all-one matrices in compatible dimensions, ρmin(M)\rho_{\min}(M) and ρmax(M)\rho_{\max}(M) be the smallest and largest eigenvalues of the matrix MM, respectively.

II-B System Model

Consider the following networked dynamical model

xt\displaystyle x_{t} =Wxt1+θt1,\displaystyle=Wx_{t-1}+\theta_{t-1}, (1)
yt\displaystyle y_{t} =xt+υt,\displaystyle=x_{t}+\upsilon_{t},

where xtx_{t} and yty_{t} represents the system state and corresponding observation at time tt, W=[wij]n×nW=[w_{ij}]_{n\times n} is the interaction topology matrix related to the adjacent matrix AA, and θt\theta_{t} and υ\upsilon represent the process and observation noises, satisfying the following Gauss-Markov assumption.

Assumption 1.

θt\theta_{t} and υt\upsilon_{t} are i.i.d. Gaussian noises, subject to N(0,σθ2I){N}(0,\sigma^{2}_{\theta}I) and N(0,συ2I){N}(0,\sigma^{2}_{\upsilon}I), respectively. They are also independent of {xt}t=0t=t\{x_{t^{\prime}}\}_{t^{\prime}=0}^{t^{\prime}=t} and {yt}t=0t=t\{y_{t^{\prime}}\}_{t^{\prime}=0}^{t^{\prime}=t}.

Next, we characterize the stability of (1) by defining

𝒮a=\displaystyle\mathcal{S}_{a}= {Zn×n,ρmax(Z)<1},\displaystyle\{Z\in\mathbb{R}^{n\times n},\rho_{\max}(Z)<1\}, (2)
𝒮m=\displaystyle\mathcal{S}_{m}= {Zn×n,ρmax(Z)=1 and the geometric\displaystyle\{Z\in\mathbb{R}^{n\times n},\rho_{\max}(Z)=1~{}\text{~{}and the geometric}
multiplicity of eigenvalue 1 equals to one}.\displaystyle\text{multiplicity of eigenvalue 1 equals to one}\}.

Then, WW is called asymptotically stable if W𝒮aW\in\mathcal{S}_{a}, or marginally stable matrix if W𝒮mW\in\mathcal{S}_{m}. Concerning its setup, popular choices include the Laplacian and Metropolis rules [24], which are given by

wij\displaystyle w_{ij} ={γaij/max{di,i𝒱},by Laplacian rule,aij/max{di,dj},by Metropolis rule,\displaystyle=\left\{\begin{aligned} &\gamma a_{ij}/\max\{d_{i},i\in\mathcal{V}\},&&\text{by Laplacian rule},\\ &a_{ij}/\max\{d_{i},d_{j}\},&&\text{by Metropolis rule},\end{aligned}\right. (3)
wii\displaystyle w_{ii} =1jiwij.\displaystyle=1-\sum\nolimits_{j\neq i}w_{ij}. (4)

where the auxiliary parameter γ\gamma satisfies 0<γ10<\gamma\leq 1. Note that if WW is specified by either one of the two rules, then W𝒮mW\in\mathcal{S}_{m}. A typical matrix in 𝒮a\mathcal{S}_{a} can be directly obtained via multiplying (3) and (4) by a factor 0<α<10<\alpha<1, which is common in adaptive diffusion networks [25]. Based on (1), the observation yty_{t} can be recursively expanded as

yt=xt+υt=Wtx0+m=1tWm1θtm+υt.\displaystyle y_{t}=x_{t}+\upsilon_{t}=W^{t}x_{0}+\sum\nolimits_{m=1}^{t}W^{m-1}\theta_{t-m}+\upsilon_{t}. (5)

Considering different stabilities, it holds that

limtWt={𝟎,ifW𝒮aW,ifW𝒮m,\mathop{\lim}\limits_{t\to\infty}W^{t}=\left\{\begin{aligned} &\bm{0},~{}&&\text{if}~{}W\in\mathcal{S}_{a}\\ &W^{\infty},~{}&&\text{if}~{}W\in\mathcal{S}_{m},\end{aligned}\right. (6)

where W<\|W^{\infty}\|<\infty. Therefore, if W𝒮a𝒮mW\in\mathcal{S}_{a}\cup\mathcal{S}_{m}, yty_{t} is strictly bounded in the expectation sense.

II-C Inference Modeling and Problem of Interest

Since the information flow between nodes is specified by the topology of NDSs, we first define the hh-hop neighbor of a single node.

Definition 1 (hh-hop out-neighbor).

Node ii is a hh-hop out-neighbor of node jj if the minimal edge number of an acyclic path from jj to ii is hh, satisfying

l=1hailil+1=ai1i2ai2i3ai3i4aih1ihaihj>0,\prod\limits_{l=1}^{h}{{a_{{i_{l}}{i_{l+1}}}}}={a_{{i_{1}}{i_{2}}}}{a_{{i_{2}}{i_{3}}}}{a_{{i_{3}}{i_{4}}}}...{a_{{i_{h-1}}{i_{h}}}}{a_{{i_{h}}{j}}}>0, (7)

where node i1=i{i_{1}}=i and ih+1=j{i_{h+1}}=j. All the hh-hop out-neighbors are represented by the set 𝒩j,hout\mathcal{N}_{j,h}^{out}.

Note that when the topology is undirected, there is no need to differentiate the in/out-neighbors. If j𝒩j,houtj\in\mathcal{N}_{j,h}^{out}, node jj is also called the hh-hop in-neighbor of node ii. Unless otherwise specified, we mainly focus on the hh-hop out-neighbors of a node in the following. To present an explicit expression for 𝒩j,hout\mathcal{N}_{j,h}^{out}, we first define 𝒩j,he\mathcal{N}_{j,h}^{e} as the node set where all the nodes can reached from node jj within hh hops. Then, 𝒩j,hout\mathcal{N}_{j,h}^{out} is recursively formulated as

𝒩j,hout=𝒩j,he\{l=1h1𝒩j,lout}.\mathcal{N}_{j,h}^{out}=\mathcal{N}_{j,h}^{e}\backslash\left\{\mathop{\cup}\limits_{l=1}^{h-1}\mathcal{N}_{j,l}^{out}\right\}. (8)

When h=1h=1, 𝒩j,1out=𝒩j,1e\mathcal{N}_{j,1}^{out}=\mathcal{N}_{j,1}^{e}. The following assumption is made throughout this paper.

Assumption 2.

The topology matrix W𝒮a𝒮mW\in\mathcal{S}_{a}\cup\mathcal{S}_{m}, and the elements of WW are all non-negative. For all wij>0w_{ij}>0, there exists a lower bound w¯\underline{w} such that wijw¯>0w_{ij}\geq\underline{w}>0.

Finally, the problem of interest is formulated as follows. Consider that there are no sufficient observations of the NDS model (1) to support existing estimation or regression methods of inferring topology, e.g., the causality based estimator in [7]. Leveraging the characteristic that a NDS is easily subjected to various disturbances and exhibits state deviation, we aim to reduce the inference dependence on observation scales, and propose an active excitation based method to infer the topology from limited new observations. Mathematically, let etje_{t}^{j} be the excitation input on node jj at time tt, and y~t+hi\tilde{y}_{t+h}^{i} be the observation of ii at time (t+h)(t+h). Then, the goal of this paper is to find 𝒩j,hout\mathcal{N}_{j,h}^{out} from the limited observations {yt,y~t+h,h=1,,h}\{y_{t},\tilde{y}_{t+h^{\prime}},h^{\prime}=1,\cdots,h\}.

The above problem is very challenging, as the stochastic process and measurement noises will also affect {y~t+h,h=1,,h}\{\tilde{y}_{t+h^{\prime}},h^{\prime}=1,\cdots,h\} and even accumulate. We will address these issues from the following aspects.

  • To make the excitation’s impact differentiable, we resort to the tool of hypothesis testing to derive the conditions for excitation magnitude to guarantee arbitrary misjudgement probability of one-hop neighbor inference.

  • To overcome the decaying effects of excitations, we establish the probabilistic relationship between the one-shot excitation magnitude and the accuracy of hh-hop neighbor inference.

  • We further extend the analysis to multiple excitations cases and illustrate how to use excitations to improve the accuracy of existing topology inference methods.

III Excitation-based Inference Method

In this section, we first analyze the reaction behavior of a NDS under excitation inputs. Then, we focus on how to infer the one-hop neighbors by one-shot excitation and characterize the inference accuracy in probability. Finally, we discuss how to infer the hh-hop neighbors and multiple excitations cases.

III-A Observation Modeling Under excitation

Since only {yt}t=0T\{y_{t}\}_{t=0}^{T} are directly available, for every two adjacent observations, it follows that

yt\displaystyle y_{t} =W(yt1υt1)+θt1+υt\displaystyle=W(y_{t-1}-\upsilon_{t-1})+\theta_{t-1}+\upsilon_{t}
=Wyt1+ωt,\displaystyle=Wy_{t-1}+\omega_{t}, (9)

where ωt=Wυt1+θt1+υt\omega_{t}=-W\upsilon_{t-1}+\theta_{t-1}+\upsilon_{t}, satisfying N(0,συ2WW𝖳+συ2I+σθ2I)N(0,\sigma_{\upsilon}^{2}WW^{\mathsf{T}}+\sigma_{\upsilon}^{2}I+\sigma_{\theta}^{2}I). Besides, ωt\omega_{t} is independent of all {xt}t<t\{x_{t^{\prime}}\}_{t^{\prime}<t} and {θt}t<t1\{\theta_{t^{\prime}}\}_{t^{\prime}<t-1}. We point out that (III-A) only represents the quantitative relationship between adjacent observations, not a causal dynamical process.

Similar to (III-A), the observation at time t+ht+h can be recursively written as

yt+h=Γ(h)yt+vt+hΓ(h)vt+m=1hΓ(m1)θt+hm,\displaystyle y_{t+h}\!=\!\Gamma(h)y_{t}\!+\!v_{t+h}\!-\!\Gamma(h)v_{t}\!+\!\sum_{m=1}^{h}\Gamma(m\!-\!1)\theta_{t+h-m}, (10)

where Γ(h)=Wh\Gamma(h)=W^{h} is the hh-step translation matrix. For ease notation, let ωt,h=vt+hΓ(h)vt+m=1hΓ(m1)θt+hm\omega_{t,h}=v_{t+h}-\Gamma(h)v_{t}+\sum_{m=1}^{h}\Gamma(m-1)\theta_{t+h-m}. Then, the deviation between yt+hiy_{t+h}^{i} and ytiy_{t}^{i} is represented by

yt+hiyti\displaystyle y_{t+h}^{i}-y_{t}^{i} =[Γ(h)yt]iyti+ωt,hi,\displaystyle=[\Gamma(h)y_{t}]^{i}-y_{t}^{i}+\omega_{t,h}^{i}, (11)

where ωt,hiN(0,σω,h2(i))\omega_{t,h}^{i}\!\sim\!N(0,\sigma_{\omega,h}^{2}(i)) and σω,h2(i)\sigma_{\omega,h}^{2}(i) is given by

σω,h2(i)=(1+j=1nΓij2(h))συ2+(m=1hj=1nΓij2(m1))σθ2,\sigma_{\omega,h}^{2}(i)\!=\!\left(1\!+\!\sum\limits_{j=1}^{n}\Gamma_{ij}^{2}(h)\right)\sigma_{\upsilon}^{2}\!+\!\left(\sum\limits_{m=1}^{h}\sum\limits_{j=1}^{n}\Gamma_{ij}^{2}(m\!-\!1)\right)\sigma_{\theta}^{2}, (12)

which is obtained from the mutual independence of the process and observation noises.

Note under Assumption 2, it holds that Γ(h)𝟏𝟏\Gamma(h)\bm{1}\!\leq\!\bm{1} and j=1nΓij2(h)1\sum\nolimits_{j=1}^{n}\Gamma_{ij}^{2}(h)\!\leq\!1. Leveraging the two properties, one can induce that

|yt+hiyti|Δytmax,σω,h2(i)2συ2+hσθ2,\displaystyle|y_{t+h}^{i}-y_{t}^{i}|\leq\Delta y_{t}^{\max},~{}\sigma_{\omega,h}^{2}(i)\leq 2\sigma_{\upsilon}^{2}+h\sigma_{\theta}^{2}, (13)

where the deviation bound Δytmax\Delta y_{t}^{\max} is given by

Δytmax={max{|ytiytj|:i,j𝒱},ifW𝒮mmax{|yti|:i𝒱},ifW𝒮a.\Delta y_{t}^{\max}=\left\{\begin{aligned} &\max\{|y_{t}^{i}-y_{t}^{j}|:i,j\in\mathcal{V}\},&&\text{if}~{}W\in\mathcal{S}_{m}\\ &\max\{|y_{t}^{i}|:i\in\mathcal{V}\},&&\text{if}~{}W\in\mathcal{S}_{a}.\end{aligned}\right. (14)

It is worth noting that Δytmax\Delta y_{t}^{\max} will fluctuate around zero as tt increases in either case of 𝒮m\mathcal{S}_{m} and 𝒮a\mathcal{S}_{a}. For simplicity without loss of generality, consider node jj is injected with positive excitation input etj>0e_{t}^{j}>0 at time tt. Then, the observation deviation is given by

y~t,hi,Δ=y~t+hiyti{Δytmax+Γijetj+ωt,hi,ifΓij>0,Δytmax+ωt,hi,ifΓij=0.\tilde{y}_{t,h}^{i,\Delta}\!=\!\tilde{y}_{t+h}^{i}\!-\!y_{t}^{i}\!\leq\!\left\{\begin{aligned} &\Delta y_{t}^{\max}\!+\!\Gamma_{ij}e_{t}^{j}\!+\!\omega_{t,h}^{i},~{}\text{if}~{}\Gamma_{ij}>0,\\ &\Delta y_{t}^{\max}\!+\!\omega_{t,h}^{i},~{}\text{if}~{}\Gamma_{ij}=0.\end{aligned}\right. (15)

Note that the term Γijetj\Gamma_{ij}e_{t}^{j} in (15) represents the influence of the excitation input eje_{j} over ii after hh steps. Hereafter, we will drop the subscript tt in the variables if it does not cause confusion.

Remark 1.

The excitation put etje_{t}^{j} cannot be arbitrarily large due to the internal constraints in NDSs, otherwise one can easily infer the connections by a extremely large excitation input, which makes the inference trivial. Therefore, it is of greater necessity to investigate the relationships between the inference accuracy and excitation magnitude, providing available excitation guidance to obtain accurate inference results with probability guarantees.

III-B One-hop Neighbor Inference

After node jj is injected with excitation input etje_{t}^{j}, the one-step observation deviation of node ii is given by

y~t+1i,Δ=y~t+1iyti{Δytmax+wijetj+ωt+1i,ifwij>0,Δytmax+ωt+1i,ifwij=0.\tilde{y}_{t+1}^{i,\Delta}\!=\!\tilde{y}_{t+1}^{i}\!-\!y_{t}^{i}\!\leq\!\left\{\begin{aligned} &\Delta y_{t}^{\max}\!+\!w_{ij}e_{t}^{j}\!+\!\omega_{t+1}^{i},~{}\text{if}~{}w_{ij}>0,\\ &\Delta y_{t}^{\max}\!+\!\omega_{t+1}^{i},~{}\text{if}~{}w_{ij}=0.\end{aligned}\right. (16)

For legibility, we temporarily assume Δytmax=0\Delta y_{t}^{\max}=0 and take its influence into consideration after the analysis. Since y~t,1i,Δ\tilde{y}_{t,1}^{i,\Delta} is associated with the stochastic process and measurement noises, finding 𝒩jout\mathcal{N}_{j}^{out} can be modeled as a typical binary hypothesis testing. The null and alternative hypothesis are respectively defined as

{H0:i𝒩jout,H1:i𝒩jout.\left\{\begin{aligned} &H_{0}:i\notin\mathcal{N}_{j}^{out},\\ &H_{1}:i\in\mathcal{N}_{j}^{out}.\end{aligned}\right. (17)

Then, denote Pr{H0|y~t,1i,Δ}\Pr\{{H_{0}}|\tilde{y}_{t,1}^{i,\Delta}\} (Pr{H1|y~t,1i,Δ}\Pr\{{H_{1}}|\tilde{y}_{t,1}^{i,\Delta}\}) as the probability that H0H_{0} (H1H_{1}) holds given the observation y~t,1i,Δ\tilde{y}_{t,1}^{i,\Delta}. Then, we have the following decision criterion

{Pr{H1|y~t,1i,Δ}Pr{H0|y~t,1i,Δ}H1holds,Pr{H1|y~t,1i,Δ}<Pr{H0|y~t,1i,Δ}H0holds,\left\{\begin{aligned} &\Pr\{{H_{1}}|\tilde{y}_{t,1}^{i,\Delta}\}\geq\Pr\{{H_{0}}|\tilde{y}_{t,1}^{i,\Delta}\}~{}\Rightarrow~{}H_{1}~{}\text{holds},\\ &\Pr\{{H_{1}}|\tilde{y}_{t,1}^{i,\Delta}\}<\Pr\{{H_{0}}|\tilde{y}_{t,1}^{i,\Delta}\}~{}\Rightarrow~{}H_{0}~{}\text{holds},\end{aligned}\right. (18)

which is also called the maximum posterior probability criterion. However, it is possible that (18) is misjudged in the test, for example, H0H_{0} is true but H1H_{1} is decided (Type I Error) or H1H_{1} is true but H0H_{0} is decided (Type II Error). Accordingly, let Pr{D1|H0}\Pr\{D_{1}|H_{0}\} be the false alarm probability and Pr{D0|H1}\Pr\{D_{0}|H_{1}\} be the missed detection probability, respectively. Therefore, the overall misjudgement probability is given by

δe=Pr{D1|H0}+Pr{D0|H1}.\delta_{e}=\Pr\{D_{1}|H_{0}\}+\Pr\{D_{0}|H_{1}\}. (19)

Suppose the inference center has no prior information about H1H_{1} and H0H_{0}, i.e., Pr(H1)=Pr(H0)=0.5\Pr(H_{1})=\Pr(H_{0})=0.5. Under hypothesis testing (18), the following result presents the probabilistic relationship between the inference accuracy and the injected excitation magnitude.

Theorem 1 (Critical excitation for one-hop neighbors).

To ensure the misjudgement probability within a threshold δ¯e\bar{\delta}_{e}, the excitation eje_{j} should satisfy

|ej|22σω(i)wijerf1(1δ¯e),|e^{j}|\geq\frac{2\sqrt{2}\sigma_{\omega({i})}}{w_{ij}}\text{erf}^{-1}(1-\bar{\delta}_{e}), (20)

where the Gaussian error erf(z)=2π0zexp(r2)dr\text{erf}(z)=\frac{2}{\sqrt{\pi}}\int_{0}^{z}\exp{(-r^{2})}\mathrm{d}r and erf1()\text{erf}^{-1}(\cdot) is the reverse mapping of erf(z)\text{erf}(z).

Proof.

The proof consists of two steps. First, we prove the decision threshold z0z_{0} is given by z0=wijej2z_{0}=\frac{w_{ij}e^{j}}{2}. Then, we demonstrate the critical excitation magnitude under the z0z_{0}.

For simplicity without losing generality, we begin with the case where the excitation input ej>0e^{j}>0. Note that ωi\omega^{i} is a continuous random variable, the likelihood ratio lr(z)l_{r}(z) in the test is given by

lr(z)=fω(z|H1)fω(z|H0),l_{r}(z)=\frac{f_{\omega}(z|H_{1})}{f_{\omega}(z|H_{0})}, (21)

where fω()f_{\omega}(\cdot) is the probability density function of ωi\omega^{i}. Due to the prior probabilities Pr(H1)=Pr(H0)\Pr(H_{1})=\Pr(H_{0}), the decision threshold z0z_{0} satisfies

lr(z0)=fω(z0|H1)fω(z0|H0)=Pr{H1}Pr{H0}=1.l_{r}(z_{0})=\frac{f_{\omega}(z_{0}|H_{1})}{f_{\omega}(z_{0}|H_{0})}=\frac{\Pr\{H_{1}\}}{\Pr\{H_{0}\}}=1. (22)

Since wiN(0,σω2)w^{i}\sim N(0,\sigma^{2}_{\omega}), substituting fω(y)=12πσωexp(z22σω2)f_{\omega}(y)=\frac{1}{\sqrt{2\pi}\sigma_{\omega}}\exp{(-\frac{z^{2}}{2\sigma^{2}_{\omega}})} into (22), it yields that

lr(z0)=exp((z0wijej)22σω2)exp(z022σω2)=1.l_{r}(z_{0})=\frac{\exp{(-\frac{(z_{0}-w_{ij}e^{j})^{2}}{2\sigma^{2}_{\omega}})}}{\exp{(-\frac{z_{0}^{2}}{2\sigma^{2}_{\omega}})}}=1. (23)

It follows from (23) that z02(z0wijej)2=0z_{0}^{2}-(z_{0}-w_{ij}e^{j})^{2}=0, leading to

z0=wijej2.z_{0}=\frac{w_{ij}e^{j}}{2}. (24)

Next, by the definition of δe\delta_{e}, one has

δe=\displaystyle\delta_{e}= Pr{D1|H0}+Pr{D0|H1}=z0+12πσωexp(z22σω2)dz\displaystyle\Pr\{D_{1}|H_{0}\}+\Pr\{D_{0}|H_{1}\}=\int_{z_{0}}^{+\infty}\!\!\!\!\!\frac{1}{\sqrt{2\pi}\sigma_{\omega}}\exp{(-\frac{z^{2}}{2\sigma^{2}_{\omega}})}\mathrm{d}z
+z012πσωexp((zwijej)22σω2)dz.\displaystyle+\int_{-\infty}^{z_{0}}\!\frac{1}{\sqrt{2\pi}\sigma_{\omega}}\exp{(-\frac{(z-w_{ij}e^{j})^{2}}{2\sigma^{2}_{\omega}})}\mathrm{d}z. (25)

​​Substitute z=z+wijej2=z+z0z=z^{\prime}+\frac{w_{ij}e^{j}}{2}=z^{\prime}+z_{0} into (III-B), yielding

δe=\displaystyle\delta_{e}= Pr{D1|H0}+Pr{D0|H1}\displaystyle\Pr\{D_{1}|H_{0}\}+\Pr\{D_{0}|H_{1}\}
=\displaystyle= 0+12πσωexp((z+z0)22σω2)dz\displaystyle\int_{0}^{+\infty}\!\!\!\!\!\frac{1}{\sqrt{2\pi}\sigma_{\omega}}\exp{(-\frac{(z^{\prime}+z_{0})^{2}}{2\sigma^{2}_{\omega}})}\mathrm{d}z^{\prime}
+012πσωexp((zz0)22σω2)dz\displaystyle+\int_{-\infty}^{0}\!\frac{1}{\sqrt{2\pi}\sigma_{\omega}}\exp{(-\frac{(z^{\prime}-z_{0})^{2}}{2\sigma^{2}_{\omega}})}\mathrm{d}z^{\prime}
=\displaystyle= 20+12πσωexp((z+z0)22σω2)dz\displaystyle 2\int_{0}^{+\infty}\!\!\!\!\!\frac{1}{\sqrt{2\pi}\sigma_{\omega}}\exp{(-\frac{(z^{\prime}+z_{0})^{2}}{2\sigma^{2}_{\omega}})}\mathrm{d}z^{\prime}
=\displaystyle= 2z0+12πσωexp(z22σω2)dz.\displaystyle 2\int_{z_{0}}^{+\infty}\!\!\!\!\!\frac{1}{\sqrt{2\pi}\sigma_{\omega}}\exp{(-\frac{z^{2}}{2\sigma^{2}_{\omega}})}\mathrm{d}z. (26)

​​Note that z0+12πσωexp(z22σω2)dz=(1erf(z02σω))/2\int_{z_{0}}^{+\infty}\!\!\!\!\!\frac{1}{\sqrt{2\pi}\sigma_{\omega}}\exp{(-\frac{z^{2}}{2\sigma^{2}_{\omega}})}\mathrm{d}z=(1-\text{erf}(\frac{z_{0}}{\sqrt{2}\sigma_{\omega}}))/2, thus it yields that

δe=1erf(z02σω).\delta_{e}=1-\text{erf}(\frac{z_{0}}{\sqrt{2}\sigma_{\omega}}). (27)

Substituting z0=wijej2z_{0}=\frac{w_{ij}e^{j}}{2} and δe=δ¯e\delta_{e}=\bar{\delta}_{e} into (27), we obtain

ej=22σωwijerf1(1δ¯e).e^{j}=\frac{2\sqrt{2}\sigma_{\omega}}{w_{ij}}\text{erf}^{-1}(1-\bar{\delta}_{e}). (28)

The result is likewise when ej<0e^{j}<0 due to the symmetry of Gaussian distribution. By the monotone increasing property of erf(z)\text{erf}(z), to guarantee δeδ¯e\delta_{e}\leq\bar{\delta}_{e}, the excitation input must satisfy |ej|22σωwijerf1(1δ¯e)|e^{j}|\geq\frac{2\sqrt{2}\sigma_{\omega}}{w_{ij}}\text{erf}^{-1}(1-\bar{\delta}_{e}). The proof is completed. ∎

Theorem 1 gives the lower magnitude bound of the excitation input to guarantee the specified misjudgment probability in a single time. Given the excitation input eje^{j} satisfying (20), one has with probability at least (1δ¯e)(1-\bar{\delta}_{e}) to accurately discriminate whether i𝒩jouti\in\mathcal{N}_{j}^{out}. Note that the interaction weight wijw_{ij} is not priorly known in reality. Thus the decision threshold in theory, wijej2\frac{w_{ij}e^{j}}{2}, is unavailable. However, we can enable the hypothesis test by specifying the least interaction weight that one wishes to discriminate between two nodes.

To practice, since WFnWn\|W\|_{F}\leq\sqrt{n}\|W\|\leq\sqrt{n}, thus we have j=1nwij2n\sum\limits_{j=1}^{n}w_{ij}^{2}\leq n and

σω2(i)(1+n)συ2+σθ2=σ¯ω2.\sigma_{\omega}^{2}(i)\leq(1+n)\sigma_{\upsilon}^{2}+\sigma_{\theta}^{2}=\bar{\sigma}_{\omega}^{2}. (29)

Specifically, if WW is row-stochastic, then the upper bound σ¯ω2\bar{\sigma}_{\omega}^{2} can be further reduced to 2συ2+σθ22\sigma_{\upsilon}^{2}+\sigma_{\theta}^{2}. Next, suppose that one aims to judge whether i𝒩jouti\in\mathcal{N}_{j}^{out} such that wij>w¯w_{ij}>\underline{w}, where w¯\underline{w} is the weight lower bound. Given the desired error probability bound δ¯e\bar{\delta}_{e} and the excitation input eje^{j} such that ej=22σ¯ωw¯erf1(1δ¯e)e^{j}=\frac{2\sqrt{2}\bar{\sigma}_{\omega}}{\underline{w}}\text{erf}^{-1}(1-\bar{\delta}_{e}), then with probability at 1δ¯e1-\bar{\delta}_{e} one can discriminate whether i𝒩jouti\in\mathcal{N}_{j}^{out} by

{i𝒩jout,if|y~t+1i,Δ|Δytmax+w¯|ej|2,i𝒩jout,else,\left\{\begin{aligned} &i\in\mathcal{N}_{j}^{out},&&~{}\text{if}~{}|\tilde{y}^{i,\Delta}_{t+1}|\geq\Delta y_{t}^{\max}+\frac{\underline{w}|e^{j}|}{2},\\ &i\notin\mathcal{N}_{j}^{out},&&~{}\text{else},\end{aligned}\right. (30)

where the parameters in (30) are all computable or known. Applying (30) to all other node and one can obtain an estimated set of 𝒩jout\mathcal{N}_{j}^{out}. Note that although the observation deviation |(Wyt)iyti||(Wy_{t})^{i}-y_{t}^{i}| will affect the performance of excitation based topology inference, its influence is strictly bounded under Assumption 2. The whole procedures are summarized in Algorithm 1, where we use the lower bound w¯\underline{w} that is sufficient to guarantee the accuracy probability.

A direct result from Theorem 1 is lim|ej|δe=0\mathop{\lim}\limits_{|e^{j}|\to\infty}\delta_{e}=0, which corresponds to the common intuition. As long as the excitation input is large enough, the one-hop neighbors of jj can always be inferred. Under this situation, it is also very likely that the two-hop (even more) out-neighbors of the excited node can also be identified by just single excitation.

Algorithm 1 Excitation-based Topology Inference
0:  Observations yty_{t}, target excited node jj, desired lower bound of interaction weight w¯\underline{w}, upper bound σ¯ω\bar{\sigma}_{\omega}, and tolerant error probability δ¯e\bar{\delta}_{e}.
0:  Estimation of the one-hop out-neighbor of jj, 𝒩^jout\hat{\mathcal{N}}_{j}^{out}.
1:  Initialize 𝒩^jout=\hat{\mathcal{N}}_{j}^{out}=\emptyset.
2:  Calculate the critical excitation ej=22σ¯ωw¯erf1(1δ¯e)e^{j}=\frac{2\sqrt{2}\bar{\sigma}_{\omega}}{\underline{w}}\text{erf}^{-1}(1-\bar{\delta}_{e}).
3:  Excite node jj with eje^{j} and obtain the observation yt+1y_{t+1}.
4:  Compute Δytmax=max{|ytiytj|:i,j𝒱}\Delta y_{t}^{\max}=\max\{|y_{t}^{i}-y_{t}^{j}|:i,j\in\mathcal{V}\}.
5:  for i𝒱i\in\mathcal{V} do
6:     Compute the observation deviation y~t,1i,Δ=yt+1iyti\tilde{y}^{i,\Delta}_{t,1}=y_{t+1}^{i}-y_{t}^{i}.
7:     if y~t,1i,Δ>Δytmax+w¯ej2\tilde{y}^{i,\Delta}_{t,1}>\Delta y_{t}^{\max}+\frac{\underline{w}e^{j}}{2} then
8:        𝒩^jout=𝒩^jout{i}\hat{\mathcal{N}}_{j}^{out}=\hat{\mathcal{N}}_{j}^{out}\cup\{i\}.
9:     end if
10:  end for
11:  return  The one-hop neighbor set estimation 𝒩^jout\hat{\mathcal{N}}_{j}^{out}.

III-C Multi-hop Neighbor Inference

In this part, we will demonstrate how to identify multi-hop out-neighbors of a node by single excitation. Similar with the hypothesis (17), we first define the following hypothesis that tests whether i𝒩j,hei\in\mathcal{N}_{j,h}^{e}, i.e.,

{H0(h):i𝒩j,he,H1(h):i𝒩j,he.\left\{\begin{aligned} &H_{0}(h):i\notin\mathcal{N}_{j,h}^{e},\\ &H_{1}(h):i\in\mathcal{N}_{j,h}^{e}.\end{aligned}\right. (31)

Note that (31) is a test using the observation deviation y~t,hi,Δ\tilde{y}_{t,h}^{i,\Delta} to judge whether node ii is an out-neighbor of node ii within hh steps. Although it cannot infer the hh-hop neighbor directly, valuable information can still be extracted for the final inference. To begin with, we present the following result.

Lemma 1 (Critical excitation for neighbors within hh hops).

Under hypothesis test (31), to ensure the misjudgement probability for all the neighbor within h-hop is lower than δ¯e\bar{\delta}_{e}, the excitation eje_{j} should satisfy

|ej|22σω,hΓij(h)erf1(1δ¯e).|e^{j}|\geq\frac{2\sqrt{2}\sigma_{\omega,h}}{\Gamma_{ij}(h)}\text{erf}^{-1}(1-\bar{\delta}_{e}). (32)
Proof.

Directly focusing on the hh-step node response after the excitation input is injected on jj, Γ(h)\Gamma(h) becomes the equivalent topology that corresponds to the hh-step process. Based on Theorem 1, when |ej|22σω,hΓijerf1(1δ¯e)|e^{j}|\geq\frac{2\sqrt{2}\sigma_{\omega,h}}{\Gamma_{ij}}\text{erf}^{-1}(1-\bar{\delta}_{e}) ensures the misjudgement probability is no more than (1δ¯e)(1-\bar{\delta}_{e}), which completes the proof. ∎

Note that Lemma 1 only illustrates how to reduce the misjudgement probability of i𝒩j,hei\in\mathcal{N}_{j,h}^{e}, and does not provide information about whether i𝒩j,houti\in\mathcal{N}_{j,h}^{out}. A key insight is that if ii is decided not in 𝒩j,h1e\mathcal{N}_{j,h-1}^{e} but in 𝒩j,he\mathcal{N}_{j,h}^{e}, then it is very likely that i𝒩j,houti\in\mathcal{N}_{j,h}^{out} is true. Starting from this point, we utilize a single-time excitation input and do hh-rounds tests to achieve the inference goal. Two auxiliary functions are defined as

F0(z,ej)=zej2+12πσω,hexp(r22σω,h2)dr,\displaystyle F_{0}(z,e^{j})\!=\!\int_{\frac{ze^{j}}{2}}^{+\infty}\!\!\frac{1}{\sqrt{2\pi}\sigma_{\omega,h}}\exp{(-\frac{r^{2}}{2\sigma^{2}_{\omega,h}})}\mathrm{d}r, (33)
F1(z,ej)=zej2+12πσω,hexp((rzej)22σω,h2)dr,\displaystyle F_{1}(z,e^{j})\!=\!\int_{\frac{ze^{j}}{2}}^{+\infty}\!\!\frac{1}{\sqrt{2\pi}\sigma_{\omega,h}}\exp{(-\frac{(r-ze^{j})^{2}}{2\sigma^{2}_{\omega,h}})}\mathrm{d}r, (34)

where z[0,1]z\in[0,1]. Based on F0(z)F_{0}(z) and F1(z)F_{1}(z), the inference probability of multi-hop out-neighbors is presented as follows.

Theorem 2 (Lower probability bound of neighbor inference).

Given the maximum false alarm probability α\alpha of hypothesis test (31), if the single-time excitation input ejemj=22σerf1(12α)Γijmine^{j}\geq{e^{j}_{m}}=\frac{2\sqrt{2}\sigma\text{erf}^{-1}(1-2\alpha)}{\Gamma_{ij}^{\min}}, then we have

Pr{i𝒩j,hout}F1(Γijmin,emj)(2αF1(Γijmax,emj)),\displaystyle\!\!\Pr\{i\in\mathcal{N}_{j,h}^{out}\}\!\geq\!F_{1}(\Gamma_{ij}^{\min},e^{j}_{m})(2\!-\!\alpha\!-\!F_{1}(\Gamma_{ij}^{\max},e^{j}_{m})), (35)

where Γijmin\Gamma_{ij}^{\min} and Γijmax\Gamma_{ij}^{\max} are given by

{Γijmin=min{Γij(l),l=1,,h},Γijmax=max{Γij(l),l=1,,h}.\left\{\begin{aligned} &\Gamma_{ij}^{\min}=\min\{\Gamma_{ij}(l),l=1,\cdots,h\},\\ &\Gamma_{ij}^{\max}=\max\{\Gamma_{ij}(l),l=1,\cdots,h\}.\end{aligned}\right. (36)
Proof.

The proof consists of three steps. Denote the false alarm probability by δf(h)=Pr{D1(h)|H0(h)}\delta_{f}(h)=\Pr\{D_{1}(h)|H_{0}(h)\} and the missed detection probability by δm(h)=Pr{D0(h)|H1(h)}\delta_{m}(h)=\Pr\{D_{0}(h)|H_{1}(h)\}. We first prove the critical excitation magnitude for identifying the neighbors within hh-hops. Then, we find the lower and upper bounds of δf(l)\delta_{f}(l) and δd(l)\delta_{d}(l).

Based on the famous Neyman-Pearson rule, with a specified δf=α\delta_{f}=\alpha, one has

α=z0+12πσω,hexp(z22σω,h2)dz=(1erf(z02σω,h))2.\displaystyle\alpha\!=\!\int_{z_{0}}^{+\infty}\!\!\!\frac{1}{\sqrt{2\pi}\sigma_{\omega,h}}\exp{(-\frac{z^{2}}{2\sigma^{2}_{\omega,h}})}\mathrm{d}z\!=\!\frac{(1-\text{erf}(\frac{z_{0}}{\sqrt{2}\sigma_{\omega,h}}))}{2}. (37)

It follows from (37) that

z0=2σω,herf1(12α).\displaystyle z_{0}=\sqrt{2}\sigma_{\omega,h}\text{erf}^{-1}(1-2\alpha). (38)

Due to the prior probabilities Pr{H1}=Pr{H0}\Pr\{H_{1}\}=\Pr\{H_{0}\} and based on Lemma 1, z0=Γij(h)ej2z_{0}=\frac{\Gamma_{ij}(h)e^{j}}{2} also holds at hh-step response. Substituting it into (38), it yields that

ej=22σω,herf1(12α)Γij(h).\displaystyle e^{j}=\frac{2\sqrt{2}\sigma_{\omega,h}\text{erf}^{-1}(1-2\alpha)}{\Gamma_{ij}(h)}. (39)

Next, note that F0(z,ej)F_{0}(z,e^{j}) decreases with zejze^{j} increasing. If the excitation input is designed such that

emj=22σerf1(12α)min{Γij(h),h=1,,n},\displaystyle e^{j}_{m}=\frac{2\sqrt{2}\sigma\text{erf}^{-1}(1-2\alpha)}{\min\{\Gamma_{ij}(h),h=1,\cdots,n\}}, (40)

then one infers that

δf(l)=F0(Γij(l),emj)α,1lh.\displaystyle\delta_{f}(l)=F_{0}(\Gamma_{ij}(l),e^{j}_{m})\leq\alpha,~{}\forall 1\leq l\leq h. (41)

Meanwhile, recall the detection probability δd(h)=Pr{D1(h)|H1(h)}\delta_{d}(h)=\Pr\{D_{1}(h)|H_{1}(h)\} is calculated by

δd(h)=Γij(h)ej212πσω,hexp((zΓij(h)ej)22σω,h2)dz.\displaystyle\delta_{d}(h)\!=\!\int_{\frac{\Gamma_{ij}(h)e^{j}}{2}}^{\infty}\!\frac{1}{\sqrt{2\pi}\sigma_{\omega,h}}\exp{(-\frac{(z-\Gamma_{ij}(h)e^{j})^{2}}{2\sigma^{2}_{\omega,h}})}\mathrm{d}z. (42)

Since δd(h)\delta_{d}(h) increases with Γij(h)emaxj\Gamma_{ij}(h)e^{j}_{\max} increasing, one has

F1(Γijmin,emj)δd(h)F1(Γijmax,emj).\displaystyle F_{1}(\Gamma_{ij}^{\min},e^{j}_{m})\leq\delta_{d}(h)\leq F_{1}(\Gamma_{ij}^{\max},e^{j}_{m}). (43)

Finally, utilizing the Law of Total Probability, the probability that ii is decided as member of 𝒩j,hout\mathcal{N}_{j,h}^{out} is calculated by

Pr{i𝒩j,hout}=Pr{D1(h)|H1(h)}Pr{D0(h1)|H0(h1)}\displaystyle\!\!\!\Pr\{i\!\in\!\mathcal{N}_{j,h}^{out}\}\!=\!\Pr\{D_{1}(h)|H_{1}(h)\}\Pr\{D_{0}(h\!-\!1)|H_{0}(h\!-\!1)\}
+Pr{D1(h)|H1(h)}Pr{D0(h1)|H1(h1)}\displaystyle+\Pr\{D_{1}(h)|H_{1}(h)\}\Pr\{D_{0}(h\!-\!1)|H_{1}(h\!-\!1)\} . (44)

​​Substitute Pr{D0(h1)|H0(h1)}=1δf(h1)\Pr\{D_{0}(h-1)|H_{0}(h-1)\}=1-\delta_{f}(h-1) and Pr{D0(h1)|H1(h1)}=1δd(h1)\Pr\{D_{0}(h-1)|H_{1}(h-1)\}=1-\delta_{d}(h-1) into (III-C), and it yields that

Pr{i𝒩j,hout}=\displaystyle\!\!\!\!\Pr\{i\in\mathcal{N}_{j,h}^{out}\}= δd(h)(2δf(h1)δd(h1))\displaystyle\delta_{d}(h)(2-\delta_{f}(h-1)-\delta_{d}(h-1))
\displaystyle\geq F1(Γijmin,emj)(2αF1(Γijmax,emj)).\displaystyle F_{1}(\Gamma_{ij}^{\min},e^{j}_{m})(2\!-\!\alpha\!-\!F_{1}(\Gamma_{ij}^{\max},e^{j}_{m})). (45)

The proof is completed. ∎

Theorem 2 provides the lower probability bounds for Pr{i𝒩j,hout}\Pr\{i\in\mathcal{N}_{j,h}^{out}\} given the maximum false alarm probability α\alpha of the test (31). Note that the test (31) is implemented multiple rounds to infer the neighbor within hh^{\prime}-hop, h=1,,hh^{\prime}=1,\cdots,h, respectively. Therefore, a notable characteristic of the bounds by (35) is that they can be calculated recursively with just one-shot excitation input. The higher the hop number is, the lower the probability bound is. The practical application of this test is similar to (30) and omitted here.

III-D Extensions and Discussions

Theorem 1 and 2 illustrate the conditions and performances of using just one-time excitation. However, there are also situations where a large excitation input is not allowed in the network dynamics, making the methods not directly available. To overcome this deficiency, multi-excitation is a promising alternative to achieve the inference goal. In this part, we will briefly show how to address the issue.

Suppose node jj is excited mm times with the same excitation input eje^{j}, the inference center obtains the average observation deviation of mm rounds by y~m¯i,Δ=1ml=1myi,Δ(l)\tilde{y}^{i,\Delta}_{\bar{m}}=\frac{1}{m}\sum\limits_{l=1}^{m}y^{i,\Delta}{(l)}.

Corallary 1 (Upper bound of the misjudgement probability under multiple excitations).

Given excitation input ej>0e^{j}>0 and implement mm times of excitations, the misjudgement probability satisfies

δe(m)2q0ej2+12πσ/mexp(z22σ2/m)dz,\delta_{e}(m)\leq 2\int_{\frac{q_{0}e^{j}}{2}}^{+\infty}\frac{1}{\sqrt{2\pi}\sigma/\sqrt{m}}\exp{(-\frac{z^{2}}{2\sigma^{2}/m})}\mathrm{d}z, (46)

where q0=min{wij:j𝒱}q_{0}=\min\{w_{ij}:j\in\mathcal{V}\}.

Proof.

Based on the independent identically distributed characteristic of yi,Δ(l)y^{i,\Delta}{(l)}, yi,Δ(l)y^{i,\Delta}{(l)} is subject to N(0,σω2m)N(0,\frac{\sigma^{2}_{\omega}}{m}). Then, the misjudgement probability is calculated by

δe(m)=\displaystyle\delta_{e}(m)= Pr{D1|H0}+Pr{D0|H1}\displaystyle\Pr\{D_{1}|H_{0}\}+\Pr\{D_{0}|H_{1}\}
=\displaystyle= wijej2+12πσω/mexp(z22σω2/m)dz\displaystyle\int_{\frac{w_{ij}e^{j}}{2}}^{+\infty}\frac{1}{\sqrt{2\pi}\sigma_{\omega}/\sqrt{m}}\exp{(-\frac{z^{2}}{2\sigma^{2}_{\omega}/m})}\mathrm{d}z
+wijej212πσω/mexp((ywijej)22σω2/m)dz\displaystyle+\int_{-\infty}^{\frac{w_{ij}e^{j}}{2}}\!\frac{1}{\sqrt{2\pi}\sigma_{\omega}/\sqrt{m}}\exp{(-\frac{(y-w_{ij}e^{j})^{2}}{2\sigma^{2}_{\omega}/m})}\mathrm{d}z
=\displaystyle= 2wijej2+12πσω/mexp(z22σω2/m)dz\displaystyle 2\int_{\frac{w_{ij}e^{j}}{2}}^{+\infty}\frac{1}{\sqrt{2\pi}\sigma_{\omega}/\sqrt{m}}\exp{(-\frac{z^{2}}{2\sigma^{2}_{\omega}/m})}\mathrm{d}z
\displaystyle\leq 2q0ej2+12πσω/mexp(z22σω2/m)dz,\displaystyle 2\int_{\frac{q_{0}e^{j}}{2}}^{+\infty}\frac{1}{\sqrt{2\pi}\sigma_{\omega}/\sqrt{m}}\exp{(-\frac{z^{2}}{2\sigma^{2}_{\omega}/m})}\mathrm{d}z, (47)

which completes the proof. ∎

From Corallary 1, we have that the variance σω2/m\sigma^{2}_{\omega}/m and δe(m)\delta_{e}(m) will decrease as mm grows. Therefore, it follows that

limmδe(m)=0.\mathop{\lim}\limits_{m\to\infty}\delta_{e}(m)=0. (48)

Corollary 1 illustrates that even when the magnitude of the excitation input is constrained, the misjudgment probability can be significantly reduced by increasing the excitation times. Due to wijw_{ij} is not priorly known, we can relax the decision threshold as in (30). Given the maximum available excitation input emaxje^{j}_{\max} and specified weight threshold w¯ij\underline{w}_{ij}, one has with probability at least 1δ¯e,m1-\bar{\delta}_{e,m} to discriminate whether i𝒩jouti\in\mathcal{N}_{j}^{out} by the following multiple excitation testing

{i𝒩jout,if|y~m¯i,Δ|l=1mΔymax(l)m+w¯ijemaxj2,i𝒩jout,else,\!\left\{\begin{aligned} &i\in\mathcal{N}_{j}^{out},&&\text{if}~{}|\tilde{y}^{i,\Delta}_{\bar{m}}|\geq\frac{\sum\nolimits_{l=1}^{m}\Delta y^{\max}(l)}{m}+\frac{\underline{w}_{ij}e^{j}_{\max}}{2},\\ &i\notin\mathcal{N}_{j}^{out},&&\text{else},\end{aligned}\right. (49)

where δ¯e,m=2w¯ijemaxj2+12πσω/mexp(z22σω2/m)dz\bar{\delta}_{e,m}=2\int_{\frac{\underline{w}_{ij}e^{j}_{\max}}{2}}^{+\infty}\frac{1}{\sqrt{2\pi}\sigma_{\omega}/\sqrt{m}}\exp{(-\frac{z^{2}}{2\sigma^{2}_{\omega}/m})}\mathrm{d}z. A minor drawback of this method is that if the weight between two nodes is small and the excitation time is also limited, an existing edge may be regarded as not existing.

Refer to caption
(a) One-hop neighbor inference of a node using different excitation inputs.
Refer to caption
(b) Multi-hop neighbor inference of a node using the same excitation input.
Refer to caption
(c) Comparisons of the topology inference without and with the excitation based method.
Figure 1: Simulation results of the proposed excitation-based method. The experiment accuracy in (a)-(b) is obtained by implementing the hypothesis test 10001000 times and then computing the ratio of positive results.

Finally, we illustrate how to use the excitation method to prove the performance of existing inferences. Suppose the observer has gained the observations from 0 to tt moments. Traditionally, the inference problem can be formulated as solving the ordinary least square problem

W^=argWminm=1t+1ytWyt122.\hat{W}=\mathop{\arg}\limits_{W}\min\sum\nolimits_{m=1}^{t+1}\|y_{t}-Wy_{t-1}\|^{2}_{2}. (50)

By the excitation-based method, we can inject the excitation input eje^{j} on node jj at moment tt. Based on the observation y~t+1\tilde{y}_{t+1}, the results of the excitation based inference method are utilized to solve the following constrained least square problem

minW\displaystyle\mathop{\min}\limits_{W}~{} m=1tytWyt122\displaystyle\sum\nolimits_{m=1}^{t}\|y_{t}-Wy_{t-1}\|^{2}_{2} (51a)
s.t. Wij>0,if|y~t+1i,Δ|Δytmax+w¯ij|ej|2,\displaystyle W_{ij}>0,~{}\text{if}~{}|\tilde{y}^{i,\Delta}_{t+1}|\geq\Delta y_{t}^{\max}+\frac{\underline{w}_{ij}|e^{j}|}{2}, (51b)
Wij=0,if|y~t+1i,Δ|<Δytmax+w¯ij|ej|2.\displaystyle W_{ij}=0,~{}\text{if}~{}|\tilde{y}^{i,\Delta}_{t+1}|<\Delta y_{t}^{\max}+\frac{\underline{w}_{ij}|e^{j}|}{2}. (51c)

By solving problem (51), the final inferred global topology has smaller errors compared with that of (50).

Remark 2.

The key insight of improving the inference accuracy of (50) lies in that the topology is estimated in the independently row-by-row manner (i.e., solving W[i,:]W_{[i,:]}). Since the connections between jj and 𝒩jout\mathcal{N}_{j}^{out} constitute a column of WW, the explicit constraints (51b) and (51c) for wijw_{ij} reduce the uncertainty of all other elements in ii-th row of WW, thus making the global inference accuracy improved.

IV Numerical Simulations

In this section, we present numerical simulations to demonstrate the performance of the analytical results. First, we display the basic setup. Then, we conduct groups of experiments under different conditions, including the system stability and noise variance. Detailed analysis is also provided to demonstrate the performance of the proposed method.

The most critical components are the adjacent matrix AA and the interaction matrix WW. For the setting of interaction matrix WW, we randomly generate a directed topology structure with |𝒱|=20|\mathcal{V}|=20, and the weight of WW is designed by the Laplacian rule. To save space, we mainly present the results of the case W𝒮mW\in\mathcal{S}_{m} (the results of case W𝒮aW\in\mathcal{S}_{a} are likewise). For generality, the initial states of all agents are randomly selected from the interval [100,100][-100,100], and the variance of the process and observation noise satisfy σθ2=1\sigma_{\theta}^{2}=1 and συ2=1\sigma_{\upsilon}^{2}=1.

Now, we move on to verify the performance excitation-based method, as shown in Fig. 1. First, we excite a target node jj and wish to find its one-hop out-neighbor ii subject to wij0.4w_{ij}\geq 0.4. Given the lower probability bound δ¯e\bar{\delta}_{e}, the critical excitation input is calculated by |ej|=22σω(i)wijerf1(1δ¯e)|e^{j}|=\frac{2\sqrt{2}\sigma_{\omega({i})}}{w_{ij}}\text{erf}^{-1}(1-\bar{\delta}_{e}). We use the input to conduct the hypothesis test 10001000 times and compute the ratio of positive results. As one expects, considering the same one-hop neighbor connection to be inferred, larger excitation input ensures higher accuracy of the decision results, as shown in Fig. 1(a). Next, the multi-hop neighbor inference results are provided in Fig. 1(b). It is easy to see that given the maximum false alarm probability α\alpha and under the same excitation input, the accuracy for multi-hop neighbor inference will decrease as the hop number grows, which corresponds to the common intuition. The probability lower bound here is computed by (35). We note that the dashed lines in Fig. 1(a) and Fig. 1(b) are lower bounds of the accuracy in theory. Thus it makes sense that the actual accuracy in experiments is higher than that bound. The multiple excitation cases are likewise and are omitted here.

Finally, we provide the results of improving the inference performance of the causality based estimator in [7] to solve (50). Here we directly present the case W𝒮mW\in\mathcal{S}_{m} and consider the following two error indexes

ε1\displaystyle{\varepsilon_{1}} =(sign(W^)sign(W)0)/n2,\displaystyle=({{\|{\text{sign}({\hat{W}})-\text{sign}(W)\|}_{0}}})/n^{2}, (52)
ε2\displaystyle{\varepsilon_{2}} =(W^WFrob)/WFrob,\displaystyle=({\|{{\hat{W}}-W}\|_{\text{Frob}}})/{\|W\|_{\text{Frob}}}, (53)

which represents the structure and magnitude errors, respectively. As we can see from Fig. 1(c), with the same observations, the inference error is largely reduced by combining the excitation based inference results to solve (51), especially in terms of the structure error.

V Conclusions

In this paper, we investigated the topology inference problem of NDSs by using very few excitations. First, we introduced the definition of hh-hop neighbor and proposed the one-shot excitation based method. By utilizing the tool of hypothesis testing, we proved the magnitude condition of the excitation input with probability guarantees. Then, we extended the one-hop inference method to hh-hop neighbor and multiple excitations cases. The inference accuracy was rigorously analyzed. Finally, the performance study by simulations verified our performance analysis. The proposed inference method is helpful in scenarios of insufficient observations over NDSs, and can also be used as auxiliary means to improve the accuracy of existing methods. Future directions include extending the method to infer the specific values of the global topology, and making the few excitations cooperate to finish the inference task.

References

  • [1] G. Cavraro and V. Kekatos, “Graph algorithms for topology identification using power grid probing,” IEEE Control Systems Letters, vol. 2, no. 4, pp. 689–694, 2018.
  • [2] J. A. Deri and J. M. Moura, “New York City Taxi analysis with graph signal processing,” in IEEE Global Conference on Signal and Information Processing (GlobalSIP).   IEEE, 2016, pp. 1275–1279.
  • [3] Y. Li, J. He, and C. Lin, “Topology inference on partially observable mobile robotic networks under formation control,” in 2021 European Control Conference (ECC).   IEEE, 2021, pp. 497–502.
  • [4] I. Brugere, B. Gallagher, and T. Y. Berger-Wolf, “Network structure inference, a survey: Motivations, methods, and applications,” ACM Computing Surveys (CSUR), vol. 51, no. 2, pp. 1–39, 2018.
  • [5] J. Etesami and N. Kiyavash, “Measuring causal relationships in dynamical systems through recovery of functional dependencies,” IEEE Transactions on Signal and Information Processing over Networks, vol. 3, no. 4, pp. 650–659, 2017.
  • [6] A. Santos, V. Matta, and A. H. Sayed, “Local tomography of large networks under the low-observability regime,” IEEE Transactions on Information Theory, vol. 66, no. 1, pp. 587–613, 2020.
  • [7] Y. Li and J. He, “Topology inference for networked dynamical systems: A causality and correlation perspective,” in 60th IEEE Conference on Decision and Control (CDC).   IEEE, 2021, pp. 1218–1223.
  • [8] Q. Jiao, Y. Li, and J. He, “Topology inference for consensus-based cooperation under time-invariant latent input,” in 2021 IEEE 94th Vehicular Technology Conference (VTC2021-Fall), 2021, pp. 1–5.
  • [9] S. Segarra, A. G. Marques, G. Mateos, and A. Ribeiro, “Network topology inference from spectral templates,” IEEE Transactions on Signal and Information Processing over Networks, vol. 3, no. 3, pp. 467–483, 2017.
  • [10] S. Segarra, M. T. Schaub, and A. Jadbabaie, “Network inference from consensus dynamics,” in 56th IEEE Conference on Decision and Control (CDC).   IEEE, 2017, pp. 3212–3217.
  • [11] Y. Zhu, M. T. Schaub, A. Jadbabaie, and S. Segarra, “Network inference from consensus dynamics with unknown parameters,” IEEE Transactions on Signal and Information Processing over Networks, vol. 6, pp. 300–315, 2020.
  • [12] A. J. Gibberd and J. D. Nelson, “High dimensional changepoint detection with a dynamic graphical Lasso,” in 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).   IEEE, 2014, pp. 2684–2688.
  • [13] B. Baingana and G. B. Giannakis, “Tracking switched dynamic network topologies from information cascades,” IEEE Transactions on Signal Processing, vol. 65, no. 4, pp. 985–997, 2016.
  • [14] G. Karanikolas, G. B. Giannakis, K. Slavakis, and R. M. Leahy, “Multi-kernel based nonlinear models for connectivity identification of brain networks,” in 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).   IEEE, 2016, pp. 6315–6319.
  • [15] S. Wang, E. D. Herzog, I. Z. Kiss, W. J. Schwartz, G. Bloch, M. Sebek, D. Granados-Fuentes, L. Wang, and J.-S. Li, “Inferring dynamic topology for decoding spatiotemporal structures in complex heterogeneous networks,” Proceedings of the National Academy of Sciences, vol. 115, no. 37, pp. 9300–9305, 2018.
  • [16] X. Dong, D. Thanou, M. Rabbat, and P. Frossard, “Learning graphs from data: A signal representation perspective,” IEEE Signal Processing Magazine, vol. 36, no. 3, pp. 44–63, 2019.
  • [17] B. Holbert, S. Tati, S. Silvestri, T. F. La Porta, and A. Swami, “Network topology inference with partial information,” IEEE Transactions on Network and Service Management, vol. 12, no. 3, pp. 406–419, 2015.
  • [18] G. Cavraro and V. Kekatos, “Inverter probing for power distribution network topology processing,” IEEE Transactions on Control of Network Systems, vol. 6, no. 3, pp. 980–992, 2019.
  • [19] H. H. Weerts, A. G. Dankers, and P. M. Van den Hof, “Identifiability in dynamic network identification,” IFAC-PapersOnLine, vol. 48, no. 28, pp. 1409–1414, 2015.
  • [20] J. M. Hendrickx, M. Gevers, and A. S. Bazanella, “Identifiability of dynamical networks with partial node measurements,” IEEE Transactions on Automatic Control, vol. 64, no. 6, pp. 2240–2253, 2018.
  • [21] A. S. Bazanella, M. Gevers, and J. M. Hendrickx, “Network identification with partial excitation and measurement,” in 58th IEEE Conference on Decision and Control (CDC).   IEEE, 2019, pp. 5500–5506.
  • [22] M. Coutino, E. Isufi, T. Maehara, and G. Leus, “State-space network topology identification from partial observations,” IEEE Transactions on Signal and Information Processing over Networks, vol. 6, pp. 211–225, 2020.
  • [23] X. Mao, J. He, and C. Zhao, “An improved subspace identification method with variance minimization and input design,” in 2022 American Control Conference (ACC).   IEEE, to be published.
  • [24] A. H. Sayed et al., “Adaptation, learning, and optimization over networks,” Foundations and Trends® in Machine Learning, vol. 7, no. 4-5, pp. 311–801, 2014.
  • [25] V. Matta and A. H. Sayed, “Consistent tomography under partial observations over adaptive networks,” IEEE Transactions on Information Theory, vol. 65, no. 1, pp. 622–646, 2019.