This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Distributed Prediction-Correction Algorithms for Time-Varying Nash Equilibrium Tracking

Ziqin Chen [email protected]    Ji Ma [email protected]    Peng Yi [email protected]    Yiguang Hong [email protected] School of Electronics and Information Engineering, Tongji University, Shanghai, P. R. China School of Aerospace Engineering, Xiamen University, Xiamen, P. R. China
Abstract

This paper focuses on a time-varying Nash equilibrium trajectory tracking problem, that is applicable to a wide range of non-cooperative game applications arising in dynamic environments. To solve this problem, we propose a distributed prediction correction algorithm, termed DPCA, in which each player predicts future strategies based on previous observations and then exploits predictions to effectively track the NE trajectory by using one or multiple distributed gradient descent steps across a network. We rigorously demonstrate that the tracking sequence produced by the proposed algorithm is able to track the time-varying NE with a bounded error. We also show that the tracking error can be arbitrarily close to zero when the sampling period is small enough. Furthermore, we achieve linear convergence for the time-invariant Nash equilibrium seeking problem as a special case of our results. Finally, a numerical simulation of a multi-robot surveillance scenario verify the tracking performance and the prediction necessary for the proposed algorithm.

keywords:
Non-cooperative games, time varying nash equilibrium tracking, distributed prediction correction methods.
thanks: This work was sponsored by the National Key Research and Development Program of China under No. 2022YFA1004700, by the National Natural Science Foundation of China under Grant No. 62103343 and 62003239, by Shanghai Sailing Program under Grant Nos. 20YF1453000, , and partially by Shanghai Municipal Science and Technology Major Project No. 2021SHZDZX0100.

, , and thanks: Corresponding author.

1 Introduction

Game theory has been extensively used in numerous applications, including social networks Ghaderi \BBA Srikant (\APACyear2014), sensor networks Stankovic \BOthers. (\APACyear2011) and smart grids Zeng \BOthers. (\APACyear2019). The Nash equilibrium (NE), an important satisfactory solution concept in noncooperative games, emerges to be of both theoretical significance and practical relevance Başar \BBA Olsder (\APACyear1998). Thus, the NE seeking method is becoming a key instrument in solving noncooperative game problems.

Recently, with the penetration of multi-player systems, several distributed approaches to achieve the NE seeking have been proposed Gadjov \BBA Pavel (\APACyear2018); De Persis \BBA Grammatico (\APACyear2019); Zhu \BOthers. (\APACyear2020); Chen \BOthers. (\APACyear2022). Such approaches, however, require time-invariant cost functions and NEs. In fact, in a variety of dynamic real-world situations, such as real-time traffic networks, online auctions, and intermittent renewable generations, the cost functions and NEs are time-varying. As a result, both academia and industry begin to focus on the time-varying NE tracking problem, which is a natural extension of time-invariant NE-related tasks in dynamic environments Lu \BOthers. (\APACyear2020); Su \BOthers. (\APACyear2021).

A natural ideal for solving the NE tracking problem is to treat it as a sequence of static problems and solve them one by one using time-invariant NE seeking methods. However, these time-invariant methods necessitate iterative multiples or even infinite rounds for each static problem, and thus are not suitable for real-time tracking situations. To fix this drawback, a solvable approach is to predict the next NE at the current instant, and use the predicted variables as the initial value when solving the static problem at the next instant. Intuitively, only if the predicted model is properly constructed to bring the predicted variables close to the next NE, it is possible to achieve satisfactory tracking performance. This idea is called the prediction-correction technique, and it has been used in solving centralized time-varying optimization problems Simonetto \BOthers. (\APACyear2017). The concept of changing initial values for iterative algorithms also motivates ADMM techniques in distributed time-varying optimizations Liu \BOthers. (\APACyear2022), which are to approximately solve each static optimization problem and use its output as the initial value for solving the next one. So far, neither prediction-correction techniques nor ADMM techniques have been used in solving distributed NE tracking problems to our knowledge. Another approach to solving NE tracking problems is found in the continuous-time regime Ye \BBA Hu (\APACyear2015); Huang \BOthers. (\APACyear2022), which relies on ODE solutions and may not be implemented using digital computation. In this paper, we focus on discrete-time algorithms.

The objective of this paper is to solve a time-varying NE tracking problem in a distributed manner with a prediction correction algorithm (DPCA). Specifically, a prediction step is proposed to predict the evolution of the NE trajectory using historical information, followed by a correction step to correct predictions by solving a time-invariant NE seeking problem obtained at each time. The contributions are summarized as follows.

1) To the best of our knowledge, our work is the first to study a distributed discrete-time algorithm for a time-varying NE tracking with provable convergence.

2) We prove that the proposed DPCA can track time-varying NEs with a bounded error that is in inverse proportion to the number of correction steps. This result provides a quantitative analysis of the trade-off between tracking accuracy and algorithm iterations. Moreover, our proof demonstrates that the minimum number of correction steps can be one, thereby meeting the requirement of real-time tracking.

3) Furthermore, we prove that with a small enough sampling period, the tracking error can be arbitrarily close to zero. For a time-invariant NE seeking problem, a special case of our formulation, we show that the tracking error linearly converges to zero.

4) A multi-robot surveillance scenario is used to test the performance of our algorithm with only one correction step. In addition, we compare the DPCA to no-predictor algorithms in terms of tracking error. The result verifies the significance of the introduced prediction step in the NE tracking and convergence.

This paper is organized as follows. The distributed NE trajectory tracking problem is presented in Section 2. The detailed algorithm design is proposed in Section 3 and the main results are established in Section 4. The result verification is investigated in Section 5. Finally, concluding remarks are given in Section 6.

Notation: Let n\mathbb{R}^{n} be the nn-dimensional Euclidean space. For a scalar aa\in{\mathbb{R}}, a\lceil a\rceil is the smallest integer not smaller than aa. For xnx\in{\mathbb{R}^{n}}, denote its European norm by x\|x\| and its transpose by xTx^{T}. [xi]i[x_{i}]_{i\in{\mathcal{I}}} stacks the vector xix_{i} as a new column vector in the order of the index set \mathcal{I}. For matrices AA, the Kronecker product is producted by \otimes and the smallest eigenvalue is denoted by λmin(A)\lambda_{\min}(A). Denote by 0n,1nn0_{n},~{}1_{n}\in{\mathbb{R}^{n}}, and Inn×nI_{n}\in{\mathbb{R}^{n\times n}} the vectors of all zeros and ones, and the identical matrix, respectively. For a differentiable function J(x):nJ(x):\mathbb{R}^{n}\!\rightarrow\!\mathbb{R}, denote iJ(x)=J(x)xi\nabla_{i}J(x)=\frac{\partial J(x)}{\partial x_{i}} with respect to xix_{i}, where x=[xi]ix=[x_{i}]_{i\in\mathcal{I}}.

2 Problem Formulation

The time-varying game under consideration is denoted as Γt={𝒱,Jit,xi}\Gamma^{t}=\{\mathcal{V},J^{t}_{i},x_{i}\}, where 𝒱={1,,N}\mathcal{V}=\{1,\cdots,N\} denotes the set of NN players and Jit:NJ_{i}^{t}:\mathbb{R}^{N}\rightarrow\mathbb{R} is the private time-varying cost function of player ii. In the game Γt\Gamma^{t}, each player ii aims to determine his time-varying strategy xi(t)x_{i}(t) to track the optimal trajectory shown below.

xi(t)=argminxiJit(xi,xi),i𝒱,x_{i}^{*}(t)=\underset{x_{i}\in\mathbb{R}}{\arg\min}J_{i}^{t}\left(x_{i},x_{-i}\right),~{}i\in{\mathcal{V}}, (1)

where xiN1x_{-i}\in{\mathbb{R}^{N-1}} represents the joint strategies of all players except player ii. If for any t0t\geq 0, there is

Jit(xi(t),xi(t))=infxi(t)Jit(xi(t),xi(t)),\displaystyle J^{t}_{i}(x^{*}_{i}(t),x^{*}_{-i}(t))=\underset{x_{i}(t)\in{\mathbb{R}}}{\operatorname{inf}}J^{t}_{i}(x_{i}(t),x^{*}_{-i}(t)),

the time-varying vector x(t)=[xi(t)]i𝒱x^{*}(t)=[x_{i}^{*}(t)]_{i\in{\mathcal{V}}} is known as the NE trajectory. The NE trajectory tracking problem is a common issue in dynamic environments Huang \BOthers. (\APACyear2022). Take a practical example, consider a target protecting problem in +2\mathbb{R}_{+}^{2} by NN players in order to block NN intruders with locations pitp_{i}^{t} and protect a time-varying target btb_{t}. To protect the target to the greatest extent possible, it is preferable to keep the weighted center σ(x)=i=1Nxi(t)/N,x=[xi]i𝒱\sigma(x)\!=\!\sum_{i=1}^{N}x_{i}(t)/N,~{}x\!=\![x_{i}]_{i\in{\mathcal{V}}} of all players to track the target btb_{t}. In this case, each player is constantly adjusting his position to minimize the time varying Jit=xipit2+γ1xibt+γ2σ(x)bt2J_{i}^{t}\!=\!\|x_{i}\!-p_{i}^{t}\|^{2}+\gamma_{1}\|x_{i}\!-b_{t}\|+\gamma_{2}\|\sigma(x)\!-b_{t}\|^{2}, and the problem can be cast as a NE trajectory tracking problem (1).

To distributively solve the NE trajectory tracking problem, each player ii estimates all other players’ strategies for determining his tracking strategy over a communication graph. Assume that the underlying communication graph 𝒢(𝒱,A)\mathcal{G}(\mathcal{V},A) is undirected and connected. The adjacent matrix is defined as A=[aij]N×NA=[a_{ij}]_{N\times N}, if the communication link between ii and jj exists, aij>0a_{ij}>0; otherwise, aij=0a_{ij}=0. The corresponding Laplacian matrix is L=[lij]N×NL=[l_{ij}]_{N\times N} with lij=aijl_{ij}=-a_{ij} if iji\neq j, and lij=jiNaijl_{ij}=\sum_{j\neq i}^{N}a_{ij} otherwise. The neighbors set of player ii is defined as 𝒩i={j𝒱|aij>0}\mathcal{N}_{i}=\{j\in\mathcal{V}|a_{ij}>0\}. For an undirected and connected graph 𝒢\mathcal{G}, all eigenvalues of LL are real numbers and can be arranged by an ascending order 0=λ1<λN0=\lambda_{1}<\cdots\leq\lambda_{N}.

Some definitions about player ii’s estimated strategy are provided below. xi=[xii;xii]Nx^{i}=[x^{i}_{i};x^{i}_{-i}]\in{\mathbb{R}^{N}} represents that player ii’s estimates of all players’ strategies; xiix_{i}^{i}\in{\mathbb{R}} represents player ii’s estimate of his strategy, which is indeed his actual strategy, i.e., xii=xix_{i}^{i}=x_{i}; xiiN1x_{-i}^{i}\in{\mathbb{R}^{N-1}} represents player ii’s estimates on all players’ strategies but his own. By the estimated strategy vector xix^{i} and the sampling period h=tk+1tkh=t_{k+1}-t_{k}, each player ii treat the problem (1) as a sequence of static problems

minxiNJitk(xi),\displaystyle\min_{x^{i}\in\mathbb{R}^{N}}\quad J_{i}^{t_{k}}(x^{i}),
s.txi=xj,k,\displaystyle\quad\text{s.t}\quad~{}~{}~{}x^{i}=x^{j},~{}k\in{\mathbb{N}}, (2)

and then solve them one by one. However, the sampling period hh in the real-time tracking is unable to afford the time spent in exactly solving every static problems by using time-invariant NE seeking algorithms in De Persis \BBA Grammatico (\APACyear2019); Zhu \BOthers. (\APACyear2020); Chen \BOthers. (\APACyear2022). Therefore, how to design a dynamic algorithm to ensure good tracking performance with few iterative rounds during hh is the problem studied in this paper.

To proceed, we make some standard assumptions, which were also required in relevant game works  Ye \BBA Hu (\APACyear2015); Lu \BOthers. (\APACyear2020).

Assumption 1.

The first and second derivatives of the time-varying NE trajectory exist and are bounded, i.e., x˙(t)c1\|\dot{x}^{*}(t)\|\leq c_{1} and x¨(t)c2\|\ddot{x}^{*}(t)\|\leq c_{2}.

Assumption 2.

Define the preudogradient as Ft(x)=[iJit(x)]i𝒱:NNF^{t}(x)\!=\![\nabla_{i}J_{i}^{t}(x)]_{i\in{\mathcal{V}}}:\mathbb{R}^{N}\!\rightarrow\!\mathbb{R}^{N}. The following statements are required for any t0t\geq 0.
i) Ft(x)F^{t}(x) is μ\mu-strongly monotone with μ>0\mu>0, that is, for any x,yNx,y\in{\mathbb{R}^{N}},

Ft(x)Ft(y),xyμxy2.\langle F^{t}(x)-F^{t}(y),x-y\rangle\geq\mu\|x-y\|^{2}.

ii) There exist some constants θi0\theta_{i}\geq 0 such that for any fixed xiN1x_{-i}\in{\mathbb{R}^{N-1}}, we have that for all xi,yix_{i},y_{i}\in{\mathbb{R}},

iJit(xi,xi)iJit(yi,xi)θixiyi.\|\nabla_{i}J^{t}_{i}(x_{i},x_{-i})-\nabla_{i}J^{t}_{i}(y_{i},x_{-i})\|\leq\theta_{i}\|x_{i}-y_{i}\|.

iii) There exist some constants θi0\theta_{-i}\geq 0 such that for any fixed xix_{i}\in{\mathbb{R}}, we have that for all xi,yiN1,x_{-i},y_{-i}\in{\mathbb{R}^{N-1}},

iJit(xi,xi)iJit(xi,yi)θixiyi.\|\nabla_{-i}J^{t}_{i}(x_{i},x_{-i})\!-\!\nabla_{-i}J^{t}_{i}(x_{i},y_{-i})\|\!\leq\!\theta_{-i}\|x_{-i}\!-\!y_{-i}\|.
Remark 1.

Assumption 1 ensures that the variation of the time-varying NE is bounded, which is critical for bounded tracking errors. Assumption 2-i) guarantees the uniqueness of the NE trajectory. Assumptions 2-ii) and iii) correspond to the smooth of Jit(x)J_{i}^{t}(x) on xN.x\in\mathbb{R}^{N}.

The augmented preudogradient is further defined as 𝑭t(𝒙)=[iJit(xi)]i𝒱:N2N\bm{F}^{t}(\bm{x})\!\!=\!\![\nabla_{i}J^{t}_{i}(x^{i})]_{i\in{\mathcal{V}}}:\mathbb{R}^{N^{2}}\!\!\rightarrow\!\!\mathbb{R}^{N} with 𝒙=[xi]i𝒱N2\bm{x}\!=\![x^{i}]_{i\in{\mathcal{V}}}\!\in\!\mathbb{R}^{N^{2}}. Meanwhile, denote θ=maxi𝒱{(θi)2+(θi)2}\theta\!=\!\max_{i\in\mathcal{V}}\{\sqrt{(\theta_{i})^{2}+(\theta_{-i})^{2}}\}. Under Assumption 2, the θ\theta-Lipschitz continuity of 𝑭t(𝒙)\bm{F}^{t}(\bm{x}) is given in Lemma 1, whose proof is similar to that of Lemma 2 in Chen \BOthers. (\APACyear2022) and is thus omitted.

Lemma 1.

Under Assumption 2, the augmented preudogradient 𝐅t(𝐱)\bm{F}^{t}(\bm{x}) is θ\theta-Lipschitz continuous in 𝐱N2\bm{x}\in\mathbb{R}^{N^{2}}.

Algorithm 1 Distributed Prediction Correction.
Initialization: Initialize any x0iNx^{i}_{0}\in{\mathbb{R}^{N}}. Determine the number of correction steps τ+\tau\in{\mathbb{N}^{+}}, the constant q>0q>0, the stepsize α>0\alpha>0 and riNr_{i}\in{\mathbb{R}^{N}}, which is a vector with iith element being ε>0\varepsilon>0 and the rest being 0.
1: For times k=0,1,2k=0,1,2\cdots do,
2: Generate predicted variables using prior information xk+1|ki={xki,ifk<q,q+1qxki1qxkqi,otherwise.\displaystyle\vspace{-0.5em}\quad\quad x^{i}_{k+1|k}=\left\{\begin{aligned} &x^{i}_{k},~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}\text{if}~{}k<q,\\ &\frac{q+1}{q}x^{i}_{k}-\frac{1}{q}x^{i}_{k-q},~{}~{}~{}\text{otherwise}.\end{aligned}\right.\vspace{-0.5em} (3) 3: Update the local cost functions Jitk+1(x)J_{i}^{t_{k+1}}(x).
4: Initialize corrected variables x^k+1i,0=xk+1|ki.\hat{x}^{i,0}_{k+1}=x^{i}_{k+1|k}.
5: for s=0:τ1s=0:\tau-1 do
  Receive neighbor’s strategies x^k+1j,s\hat{x}^{j,s}_{k+1} and execute x^k+1i,s+1\displaystyle\quad\quad\hat{x}^{i,s+1}_{k+1} =x^k+1i,s+αi=1Naij(x^k+1j,sx^k+1i,s)\displaystyle=\hat{x}^{i,s}_{k+1}+\alpha\sum_{i=1}^{N}a_{ij}(\hat{x}^{j,s}_{k+1}-\hat{x}^{i,s}_{k+1}) (4) riiJitk+1(x^k+1i,s),\displaystyle\quad\quad\quad\quad\quad\quad\quad\quad-r_{i}\nabla_{i}J_{i}^{t_{k+1}}(\hat{x}^{i,s}_{k+1}),   end for.
6. Configure tracking variables xk+1i=x^k+1i,τx^{i}_{k+1}=\hat{x}^{i,\tau}_{k+1}.
7. end for

3 Algorithm Design

In this section, Algorithm 1 is designed for each player i,i𝒱i,~{}i\!\in\!{\mathcal{V}} to generate a sequence xki,kx^{i}_{k},~{}k\!\in\!{\mathbb{N}} to track time-varying NEs x(tk)x^{*}(t_{k}). Algorithm 1 involves two steps: prediction and correction, which are depicted below.

The prediction step design: We derive the prediction step (3) by observing the dynamic of time-varying NE. Due to Ft(x(t))=0NF^{t}(x^{*}(t))=0_{N}, the NE trajectory x(t)x^{*}(t) satisfies the following nonlinear dynamical system

x˙(t)=Ft(x(t)).\displaystyle\dot{x}^{*}(t)=F^{t}(x^{*}(t)). (5)

Then we employ two methods to discretely estimate (5) and generate auxiliary prediction sequences xk+1|kNx_{k+1|k}\in{\mathbb{R}^{N}}. Meanwhile, the prediction errors Δk=xk+1|kx(tk+1)\Delta_{k}=\|x_{k+1|k}-x^{*}(t_{k+1})\| of these two methods are compared.

i) The first method is to set the predicted variable as the NE x(tk)x^{*}(t_{k}) at the last time instant tkt_{k}, in other words,

xk+1|k=x(tk).\displaystyle x_{k+1|k}=x^{*}(t_{k}). (6)

In this case, the prediction error is measured as

Δk\displaystyle\Delta_{k} =x(tk)x(tk+1),\displaystyle=\|x^{*}(t_{k})-x^{*}(t_{k+1})\|,
=tktk+1x˙(t)𝑑τc1h.\displaystyle=\|\int_{t_{k}}^{t_{k+1}}\dot{x}^{*}(t)d\tau\|\leq c_{1}h. (7)

ii) The second method involves a qq-step interpolation method. In other words,

x¯k+1|k=q+1qx(tk)1qx(tkq),kq.\displaystyle\bar{x}_{k+1|k}=\frac{q+1}{q}x^{*}(t_{k})-\frac{1}{q}x^{*}(t_{k-q}),~{}k\geq q. (8)

Recalling the dynamic (5), the prediction error Δ¯k=x¯k+1|kx(tk+1)\bar{\Delta}_{k}=\|\bar{x}_{k+1|k}-x^{*}(t_{k+1})\| is then computed as follows,

Δ¯k\displaystyle\bar{\Delta}_{k} =1qx(tk)1qx(tkq)tktk+1Fτ(x)𝑑τ,\displaystyle=\left\|\frac{1}{q}x^{*}(t_{k})-\frac{1}{q}x^{*}(t_{k-q})-\int_{t_{k}}^{t_{k+1}}F^{\tau}(x^{*})d\tau\right\|, (9)
1qr=0q1tktk+1Fτ(x)Fτ(qr)h(x)𝑑τ.\displaystyle\leq\frac{1}{q}\sum_{r=0}^{q-1}\int_{t_{k}}^{t_{k+1}}\left\|F^{\tau}(x^{*})-F^{\tau-(q-r)h}(x^{*})\right\|d\tau.

Using the Mean Value Theorem and Assumption 1 yields

Fτ(x)Fτ(qr)h(x)\displaystyle\|F^{\tau}(x^{*})-F^{\tau-(q-r)h}(x^{*})\| =dFt(x)dt|t=ξ(qr)h,\displaystyle=\left\|\frac{dF^{t}(x^{*})}{dt}\Big{|}_{t=\xi}(q-r)h\right\|,

where ξ(τ(qr)h,τ)\xi\in(\tau-(q-r)h,\tau). Assumption 1 ensures that

Δ¯kc2(q+1)h22.\displaystyle\bar{\Delta}_{k}\leqslant\frac{c_{2}(q+1)h^{2}}{2}. (10)

If the sampling period hh is small enough, then Δ¯kΔk\bar{\Delta}_{k}\leq{\Delta}_{k}, which indicates that the second method outperforms the first. In fact, the sampling period h1h\ll 1 in the dynamic environment. Hence, it motivates the design of prediction in (3), while ensuring that predicted variables xk+1|kix^{i}_{k+1|k} are close to x(tk+1)x^{*}(t_{k+1}) at the time instant tk+1t_{k+1}.

Remark 2.

The qq step interpolation method ensure the flexibility of the prediction’s design. When q=1q=1, the interpolation method in (3) is related to two steps.

The correction step design: We employ a distributed gradient descent (DGD) algorithm (4) to correct errors produced in the prediction. In particular, the predicted variable xk+1|kix^{i}_{k+1|k} that is close to x(tk+1)x^{*}(t_{k+1}) is harnessed as the correction step’s initial value. It differs from the no-predictor DGD algorithm, which uses any initial state to exact update. In this sense, Algorithm 1 may require fewer iterative rounds τ\tau and take less computation time to achieve the same tracking performance as the no-predictor DGD algorithm.

4 Main Result

This section establishes the convergence of Algorithm 1. We first analyze the prediction error in Lemma 2, and then we examine the tracking error in Lemma 3. Following those, Theorem 1 rigorously characterizes the upper bound of the tracking error. Before we begin, we define stacked variables 𝒙k+1|k=[xk+1|ki]i𝒱\bm{x}_{k+1|k}\!=\![x_{k+1|k}^{i}]_{i\in{\mathcal{V}}}, 𝒙k=[xki]i𝒱\bm{x}_{k}\!=\![x^{i}_{k}]_{i\in{\mathcal{V}}} and 𝒙(tk)=1Nx(tk)\bm{x}^{*}(t_{k})\!=\!1_{N}\otimes x^{*}(t_{k}). The proofs of Lemmas 2 and 3 can be found in Appendix.

Lemma 2.

Suppose Assumptions 1 and 2 hold. For any kqk\geq q, the norm of prediction errors is upper bounded by

𝒙k+1|k𝒙(tk+1)q+1q𝒙k𝒙(tk)\displaystyle\|\bm{x}_{k+1|k}-\bm{x}^{*}(t_{k+1})\|\leq\frac{q+1}{q}\|\bm{x}_{k}-\bm{x}^{*}(t_{k})\|
+1q𝒙kq𝒙(tkq)+c2N(q+1)2h2.\displaystyle\quad\quad\quad+\frac{1}{q}\|\bm{x}_{k-q}-\bm{x}^{*}(t_{k-q})\|+\frac{c_{2}\sqrt{N}(q+1)}{2}h^{2}. (11)
Remark 3.

Lemma 2 gives the prediction error when kqk\geq q, but not when k<qk<q. The reason is that the prediction error analysis for the case k<qk<q is able to be included in the proof of Theorem 1.

Lemma 3.

Suppose Assumptions 1 and 2 hold. There exist ε\varepsilon and α\alpha satisfying

{α(0,λ22λN2),ε(0,min{αλ2μN4θ2+2θμN+μ2,μ2Nθ2,N4αλ2μ}),\begin{cases}\alpha\in(0,\frac{\lambda_{2}}{2\lambda_{N}^{2}}),\\ \varepsilon\in\left(0,\min\left\{\frac{\alpha\lambda_{2}\mu N}{4\theta^{2}+2\theta\mu N+\mu^{2}},\frac{\mu}{2N\theta^{2}},\frac{N}{4\alpha\lambda_{2}\mu}\right\}\right),\end{cases}

such that the following matrix is positive

Aα,ε=[εμNε2θ2εθNεθNαλ2α2λN2ε2θ2εθ],\displaystyle A_{\alpha,\varepsilon}=\left[\begin{matrix}\small\frac{\varepsilon\mu}{N}-\varepsilon^{2}\theta^{2}&-\frac{\varepsilon\theta}{\sqrt{N}}\\ -\frac{\varepsilon\theta}{\sqrt{N}}&\alpha\lambda_{2}-\alpha^{2}\lambda_{N}^{2}-\varepsilon^{2}\theta^{2}-\varepsilon\theta\end{matrix}\right], (12)

and ν=2λmin(Aα,ε)(0,1)\nu=2\lambda_{\min}(A_{\alpha,\varepsilon})\in(0,1). Then for any kk\in{\mathbb{N}}, the norm of correction errors is upper bounded by

𝒙k+1𝒙(tk+1)ρτ𝒙k+1|k𝒙(tk+1),\displaystyle\|\bm{x}_{k+1}-\bm{x}^{*}(t_{k+1})\|\leq\rho^{\tau}\|\bm{x}_{k+1|k}-\bm{x}^{*}(t_{k+1})\|, (13)

where the coefficient ρ=(1ν)12(0,1)\rho=(1-\nu)^{\frac{1}{2}}\in(0,1).

Remark 4.

Lemma 3 implies that the correction step (4) provides a discrete time approach to time-invariant distributed NE seeking problems in Gadjov \BBA Pavel (\APACyear2018); De Persis \BBA Grammatico (\APACyear2019); Zhu \BOthers. (\APACyear2020); Chen \BOthers. (\APACyear2022). Meanwhile, Lemma 3 proves the linear convergence rate O(ρτ)O(\rho^{\tau}) of this discrete time approach.

With Lemmas 2 and 3, we obtain our main theorem.

Theorem 1.

Suppose Assumptions 1 and 2 hold. For any positive integer q>2ρτ1ρτq>\lceil\frac{2\rho^{\tau}}{1-\rho^{\tau}}\rceil, there are positive constants γ(0,1)\gamma\in(0,1) and C1C_{1} satisfying

ρτ(q+1)q1γ+ρτq1γq+11,\displaystyle\frac{\rho^{\tau}(q+1)}{q}\frac{1}{\gamma}+\frac{\rho^{\tau}}{q}\frac{1}{\gamma^{q+1}}\leq 1, (14)
C1𝒙0𝒙(t0)+Nρτhc11ρτ.\displaystyle C_{1}\geq\|\bm{x}_{0}-\bm{x}^{*}(t_{0})\|+\frac{\sqrt{N}\rho^{\tau}hc_{1}}{1-\rho^{\tau}}. (15)

Then Algorithm 1 ensures the norm of tracking errors to be upper bounded by

𝒙k𝒙(tk)C1γk+C2h2,\|\bm{x}_{k}-\bm{x}^{*}(t_{k})\|\leq C_{1}\gamma^{k}+C_{2}h^{2}, (16)

where the positive constant C2=c2Nq(q+1)ρτ2q2ρτ(q+2)C_{2}=\frac{c_{2}\sqrt{N}q(q+1)\rho^{\tau}}{2q-2\rho^{\tau}(q+2)}.

{pf}

We prove that there must be a γ0(0,1)\gamma_{0}\in(0,1) such that (14) holds. Define a decaying function g(γ):g(\gamma):\mathbb{R}\rightarrow\mathbb{R} related to γ(0,+)\gamma\in(0,+\infty) as follows,

g(γ)=ρτ(q+1)q1γ+ρτq1γq+1.\displaystyle g(\gamma)=\frac{\rho^{\tau}(q+1)}{q}\frac{1}{\gamma}+\frac{\rho^{\tau}}{q}\frac{1}{\gamma^{q+1}}.

By q>2ρτ1ρτq>\lceil\frac{2\rho^{\tau}}{1-\rho^{\tau}}\rceil, we can easily obtain that

g(1)=ρτ(q+1)q+ρτq<1.\displaystyle g(1)=\frac{\rho^{\tau}(q+1)}{q}+\frac{\rho^{\tau}}{q}<1.

Thus, there is a constant γ0(0,1)\gamma_{0}\in(0,1) satisfying (14).

Next, we prove (16) by starting with the case k<qk<q, and then complete the proof for any kk\in{\mathbb{N}}.

For any k<qk<q, there is 𝒙k+1|k=𝒙k\bm{x}_{k+1|k}=\bm{x}_{k}. According to Lemma 3 and the triangle inequality, we have

𝒙k+1𝒙(tk+1)\displaystyle\quad\|\bm{x}_{k+1}-\bm{x}^{*}(t_{k+1})\|
ρτ𝒙k𝒙(tk)+ρτ𝒙(tk)𝒙(tk+1),\displaystyle\leq\rho^{\tau}\|\bm{x}_{k}-\bm{x}^{*}(t_{k})\|+\rho^{\tau}\|\bm{x}^{*}(t_{k})-\bm{x}^{*}(t_{k+1})\|,
(ρτ)k+1𝒙0𝒙(t0)+r=0k(ρτ)r+1\displaystyle\leq(\rho^{\tau})^{k+1}\|\bm{x}_{0}-\bm{x}^{*}(t_{0})\|+\sum_{r=0}^{k}(\rho^{\tau})^{r+1}
𝒙(tk+1r)𝒙(tkr).\displaystyle\quad\quad\quad\quad\quad\quad\quad\quad\quad\quad\|\bm{x}^{*}(t_{k+1-r})-\bm{x}^{*}(t_{k-r})\|. (17)

The upper bound of the last term of (17) is analyzed. Based on (7), we know that x(tk+1)x(tk)hc1\|x^{*}(t_{k+1})-x^{*}(t_{k})\|\leq hc_{1}. By the sequence summation, we rewrite (17) as

𝒙k+1𝒙(tk+1)\displaystyle\quad\|\bm{x}_{k+1}-\bm{x}^{*}(t_{k+1})\|
(ρτ)k+1𝒙0𝒙(t0)+Nρτ(1(ρτ)k+1)hc11ρτ.\displaystyle\leq(\rho^{\tau})^{k+1}\|\bm{x}_{0}-\bm{x}^{*}(t_{0})\|+\frac{\sqrt{N}\rho^{\tau}(1-(\rho^{\tau})^{k+1})hc_{1}}{1-\rho^{\tau}}.

By (14) with ρτ<γ<1\rho^{\tau}<\gamma<1, we have that for any k<qk<q,

𝒙k𝒙(tk)\displaystyle\|\bm{x}_{k}-\bm{x}^{*}(t_{k})\| γk𝒙0𝒙(t0)+Nρτhc11ρτC1γk.\displaystyle\leq\gamma^{k}\|\bm{x}_{0}-\bm{x}^{*}(t_{0})\|+\frac{\sqrt{N}\rho^{\tau}hc_{1}}{1-\rho^{\tau}}\leq C_{1}\gamma^{k}.

At last, we prove (16) for any kk\in{\mathbb{N}} via the mathematical induction. Assume that 𝒙k𝒙(tk)C1γk+C2h2\|\bm{x}_{k^{\prime}}-\bm{x}^{*}(t_{k^{\prime}})\|\leq C_{1}\gamma^{k^{\prime}}+C_{2}h^{2} for any kk,kqk^{\prime}\leq k,~{}k\geq q. We prove 𝒙k+1𝒙(tk+1)C1γk+1+C2h2.\|\bm{x}_{k^{\prime}+1}-\bm{x}^{*}(t_{k^{\prime}+1})\|\leq C_{1}\gamma^{k^{\prime}+1}+C_{2}h^{2}. By inputting (11) into (13), we have

𝒙k+1𝒙(tk+1)\displaystyle\quad\|\bm{x}_{k^{\prime}+1}-\bm{x}^{*}(t_{k^{\prime}+1})\|
ρτ(q+1)q𝒙k𝒙(tk)\displaystyle\leq\frac{\rho^{\tau}(q+1)}{q}\|\bm{x}_{k}-\bm{x}^{*}(t_{k})\|
+ρτq𝒙kq𝒙(tkq)+c2ρτN(q+1)2h2.\displaystyle\quad+\frac{\rho^{\tau}}{q}\|\bm{x}_{k-q}-\bm{x}^{*}(t_{k-q})\|+\frac{c_{2}\rho^{\tau}\sqrt{N}(q+1)}{2}h^{2}. (18)

Integrating the initial argument 𝒙k𝒙(tk)C1γk+C2h2\|\bm{x}_{k^{\prime}}-\bm{x}^{*}(t_{k^{\prime}})\|\!\leq\!C_{1}\gamma^{k^{\prime}}\!+\!C_{2}h^{2} for any kk,kqk^{\prime}\leq k,~{}k\geq q, the definition of C2C_{2} and the inequality (14) into (18), we obtain that for any kqk\geq q, 𝒙k+1𝒙(tk+1)C1γk+1+C2h2.\|\bm{x}_{k^{\prime}+1}-\bm{x}^{*}(t_{k^{\prime}+1})\|\leq C_{1}\gamma^{k^{\prime}+1}+C_{2}h^{2}. Combining 𝒙k𝒙(tk)C1γk\|\bm{x}_{k}-\bm{x}^{*}(t_{k})\|\leq C_{1}\gamma^{k} for any k[0,q)k\in[0,q), we conclude that for any kk\in{\mathbb{N}}, the inequality (16) holds. Theorem 1 shows that each sequence {xki}\{x^{i}_{k}\} approaches a neighborhood of the NE trajectory x(tk)x^{*}(t_{k}) as time passes. The exponentially convergence of Algorithm 1 is shown in the following corollary.

Corollary 1.

Under the same conditions of Theorem 1, the sequence {𝐱k}\{\bm{x}_{k}\} generated by Algorithm 1 converges to 𝐱(tk)\bm{x}^{*}(t_{k}) exponentially up to a bounded error as

lim supk𝒙k𝒙(tk)C2h2=O(h2).\displaystyle\limsup_{k\rightarrow\infty}\|\bm{x}_{k}-\bm{x}^{*}(t_{k})\|\leq C_{2}h^{2}=O(h^{2}).
Remark 5.

Corollary 1 describes the exponential convergence property of Algorithm 1. The convergence accuracy is determined by the values of hh and C2C_{2} that is in inverse proportion to the correction steps τ+\tau\in{\mathbb{N}^{+}}. Note that increasing the value of τ\tau improves tracking accuracy while necessitating much more computation time. Thus, the number of correction steps τ\tau can be adjusted to match the actual tracking bound. Furthermore, if the sampling period hh is small enough, then Algorithm 1 ensure that the sequence {𝐱k}\{\bm{x}_{k}\} is sufficiently close to the NE trajectory.

5 Example

A robotic surveillance scenario Carnevale \BOthers. (\APACyear2022) is used to validate the effectiveness of Algorithm 1, including a tracking error comparison between Algorithm 1 and the no-predictor algorithm. Consider five robots collaborating to protect a time-varying target from five intruders. Each intruder ii moves according to the following rule:

pit=pic+5[cos(t/20)sin(t/10)],\displaystyle p_{i}^{t}=p_{i}^{c}+5\begin{bmatrix}\cos(t/20)\\ \sin(t/10)\end{bmatrix}, (19)

where pic=[8+6(i1);8+6(i1)],i{1,,5}p_{i}^{c}=[8+6(i-1);8+6(i-1)],~{}i\in\{1,\cdots,5\}. The target moves via a similar law with (19), so we set it to be bt=i=15pit/10.b_{t}=\sum_{i=1}^{5}p_{i}^{t}/10. The dynamic protection strategy of each robot is determined by the following cost function

fit(xi,xi)\displaystyle f^{t}_{i}(x_{i},x_{-i}) =γ12xi(t)pit2+γ22xi(t)bt2\displaystyle=\frac{\gamma_{1}}{2}\|x_{i}(t)-p_{i}^{t}\|^{2}+\frac{\gamma_{2}}{2}\|x_{i}(t)-b_{t}\|^{2}
+γ310σ(x(t))bt2,\displaystyle\quad\quad\quad\quad\quad\quad+\frac{\gamma_{3}}{10}\|\sigma(x(t))-b_{t}\|^{2}, (20)

where xi(t)2x_{i}(t)\in{\mathbb{R}^{2}} denotes the robot’s position, γ1=γ2=0.1\gamma_{1}=\gamma_{2}=0.1 and γ3=0.875\gamma_{3}=0.875. The aggregative variable σ(x(t))=i=15xi(t)/5\sigma(x(t))=\sum_{i=1}^{5}x_{i}(t)/5. The network graph is described by a ring. The parameters for Algorithm 1 are set to q=400q=400, α=0.05\alpha=0.05, ε=0.01\varepsilon=0.01, h=0.1h=0.1 and τ=1\tau=1. The element of initial tracking sequences {x0i}10\{x_{0}^{i}\}\in{\mathbb{R}^{10}} is generated at random within an interval [5+5(i1),10+5(i1)][5+5(i-1),10+5(i-1)].

Fig. 1 shows that for any i{1,,5}i\in\{1,\cdots,5\}, every element of xki10x^{i}_{k}\in{\mathbb{R}^{10}} generated by Algorithm 1 can effectively reach consensus and track every element of the NE trajectory x(t)x^{*}(t). Fig. 2 depicts that how the network constructs a smarter formation that protects the time-varying target btb_{t} better. As shown in Fig. 3(a), the tracking error 𝒙k𝒙(tk)\|\bm{x}_{k}-\bm{x}^{*}(t_{k})\| converges to a constant term 7h2=0.077h^{2}=0.07. Fig. 3 (b) demonstrates the importance of the prediction step in the NE trajectory tracking.

Refer to caption
((a))
Refer to caption
((b))
Figure 1: Tracking performance of NE trajectory in terms of Algorithm 1. The solid lines represent x(t)=[xi(t)]i{1,,5}x^{*}(t)=[x_{i}^{*}(t)]_{i\in\{1,\cdots,5\}}, where xi(t)=[xi1(t);xi2(t)]x_{i}^{*}(t)=[x_{i1}^{*}(t);x_{i2}^{*}(t)]. The dotted lines represent xki(t)10x^{i}_{k}(t)\in{\mathbb{R}^{10}} in Algorithm 1.
Refer to caption
((a))
Refer to caption
((b))
Refer to caption
((c))
Refer to caption
((d))
Figure 2: The positions of robots, the time-varying intruders and targets at different time. The red star, circles and squares represent the target, robots and intruders, respectively.
Refer to caption
((a))
Refer to caption
((b))
Figure 3: The tracking error verification [cf. Theorem 1].

6 Conclusion

In this paper, we proposed a distributed prediction-correction algorithm (DPCA) to track the NE trajectory. By analyzing the coupling relationship between the prediction and the correction procedures, we rigorously proved that the tracking sequence generated by the DPCA can track the NE trajectory with a bounded error. This bound is proportional to the sampling period and inversely proportional to the number of correction steps. The results filled a gap in the study of NE tracking problems using discrete time algorithms. Moreover, the linear convergence rate for static games with time-invariant cost functions has been included in our results. The Future research of interests includes considering the communication efficient and privacy protected NE tracking algorithm.

7 Appendix

7.1 Proof of Lemma 2

In consideration of the error bound in (10), we prove Lemma 2. By the prediction step in (3), we obtain

{xk+1|ki=q+1qxki1qxkqi,kq,x(tk+1)=q+1qx(tk)1qx(tkq)Δ¯k.\begin{cases}&x^{i}_{k+1|k}=\frac{q+1}{q}x^{i}_{k}-\frac{1}{q}x^{i}_{k-q},~{}k\geq q,\\ &x^{*}(t_{k+1})=\frac{q+1}{q}x^{*}(t_{k})-\frac{1}{q}x^{*}(t_{k-q})-\bar{\Delta}_{k}.\end{cases} (21)

Then the prediction error can be computed as

xk+1|kix(tk+1)\displaystyle x^{i}_{k+1|k}-x^{*}(t_{k+1})
=q+1q(xkix(tk))1q(xkqix(tkq))+Δ¯k.\displaystyle\quad=\frac{q+1}{q}(x^{i}_{k}-x^{*}(t_{k}))-\frac{1}{q}(x^{i}_{k-q}-x^{*}(t_{k-q}))+\bar{\Delta}_{k}.

Stacking the above equality, considering the norm of the resulting expression and applying the triangle inequality, Lemma 2 with (11) holds.

7.2 Proof of Lemma 3

The proof Lemma 3 is divided into two parts: i) Prove that the matrix Aα,ϵA_{\alpha,\epsilon} is positive and the constant ν(0,1)\nu\in(0,1); ii) Prove the relationship between the tracking error and the prediction error in (13).

i) By 0<ε<min{αλ2μN4θ2+2θμN+μ2,μ2Nθ2,N4αλ2μ}0<\varepsilon<\min\{\frac{\alpha\lambda_{2}\mu N}{4\theta^{2}+2\theta\mu N+\mu^{2}},\frac{\mu}{2N\theta^{2}},\frac{N}{4\alpha\lambda_{2}\mu}\}, it yields

μ2N>εθ2andαλ22ε>2θ2μN+θ+μ2N>2θ2μN+θ+εθ2.\displaystyle\frac{\mu}{2N}>\varepsilon\theta^{2}~{}\text{and}~{}\frac{\alpha\lambda_{2}}{2\varepsilon}>\frac{2\theta^{2}}{\mu N}+\theta+\frac{\mu}{2N}>\frac{2\theta^{2}}{\mu N}+\theta+\varepsilon\theta^{2}.

Meanwhile, α(0,λ22λN2)\alpha\in(0,\frac{\lambda_{2}}{2\lambda^{2}_{N}}) ensures αλ22ε>α2λN2ε\frac{\alpha\lambda_{2}}{2\varepsilon}>\frac{\alpha^{2}\lambda_{N}^{2}}{\varepsilon}, and thus,

μ2N[αλ2εα2λN2εεθ2θ]θ2N2.\displaystyle\frac{\mu}{2N}\left[\frac{\alpha\lambda_{2}}{\varepsilon}-\frac{\alpha^{2}\lambda_{N}^{2}}{\varepsilon}-\varepsilon\theta^{2}-\theta\right]\geq\frac{\theta^{2}}{N^{2}}.

By using again εθ2μ2N\varepsilon\theta^{2}\leq\frac{\mu}{2N} and multiplying both side of the deriving results by ε2\varepsilon^{2}, we have

(εμNε2θ2)[αλ2α2λN2ε2θ2εθ]ε2θ2N2>0,\displaystyle(\frac{\varepsilon\mu}{N}-\varepsilon^{2}\theta^{2})\left[\alpha\lambda_{2}-\alpha^{2}\lambda_{N}^{2}-\varepsilon^{2}\theta^{2}-\varepsilon\theta\right]-\frac{\varepsilon^{2}\theta^{2}}{N^{2}}>0,

which is the determinant of Aα,εA_{\alpha,\varepsilon}. Thus, Aα,ε>0A_{\alpha,\varepsilon}>0. By further considering ε<N4αλ2μ\varepsilon<\frac{N}{4\alpha\lambda_{2}\mu}, it yields εμNαλ2<14\frac{\varepsilon\mu}{N}\alpha\lambda_{2}<\frac{1}{4} and

(εμNε2θ2)[αλ2α2λN2ε2θ2εθ]ε2θ2N2<14.\displaystyle(\frac{\varepsilon\mu}{N}-\varepsilon^{2}\theta^{2})\left[\alpha\lambda_{2}-\alpha^{2}\lambda_{N}^{2}-\varepsilon^{2}\theta^{2}-\varepsilon\theta\right]-\frac{\varepsilon^{2}\theta^{2}}{N^{2}}<\frac{1}{4}.

which indicates λmin(Aα,ε)<12\lambda_{\min}(A_{\alpha,\varepsilon})<\frac{1}{2}. Thus, ν(0,1)\nu\in(0,1).

ii) Defining 𝑳=LIN\bm{L}=L\otimes I_{N} and 𝒙^k+1s=[x^k+1i,s]i𝒱\hat{\bm{x}}^{s}_{k+1}=[\hat{x}^{i,s}_{k+1}]_{i\in\mathcal{V}}, the correction step (4) can be written as a stacked form

𝒙^k+1s+1𝒙^k+1s=α𝑳𝒙^k+1sR𝑭tk+1(𝒙^k+1s),\displaystyle\hat{\bm{x}}^{s+1}_{k+1}-\hat{\bm{x}}^{s}_{k+1}=-\alpha\bm{L}\hat{\bm{x}}^{s}_{k+1}-R\bm{F}^{t_{k+1}}(\hat{\bm{x}}^{s}_{k+1}), (22)

where R=diag(r1,,rN)R=\text{diag}(r_{1},\cdots,r_{N}). We first prove that at the equilibrium of system (22), x^k+1i\hat{x}^{i}_{k+1} reaches the NE of game Γtk+1\Gamma^{t_{k+1}}. If we set the left hand side of  (22) to zero, then

α𝑳𝒙^k+1+R𝑭tk+1(𝒙^k+1)=0N2.\alpha\bm{L}\hat{\bm{x}}^{*}_{k+1}+R\bm{F}^{t_{k+1}}(\hat{\bm{x}}^{*}_{k+1})=0_{N^{2}}. (23)

Premultiplying both sides by 1NTIN1^{T}_{N}\otimes I_{N} gives

0N=α(1NTIN)(LIN)𝒙^k+1\displaystyle 0_{N}=\alpha\left(1_{N}^{T}\otimes I_{N}\right)\left(L\otimes I_{N}\right)\hat{\bm{x}}^{*}_{k+1}
+(1NTIN)R𝑭tk+1(𝒙^k+1).\displaystyle\quad\quad\quad\quad\quad+\left(1_{N}^{T}\otimes I_{N}\right)R\bm{F}^{t_{k+1}}\left(\hat{\bm{x}}^{*}_{k+1}\right).

Using 1NTL=0NT1_{N}^{T}L=0^{T}_{N} yields 0N=(1NTIN)R𝑭tk+1(𝒙^k+1)0_{N}=\left(1_{N}^{T}\otimes I_{N}\right)R\bm{F}^{t_{k+1}}\left(\hat{\bm{x}}^{*}_{k+1}\right). By the definitions of RR and 𝑭tk+1\bm{F}^{t_{k+1}}, 𝑭tk+1(𝒙^k+1)=0N\bm{F}^{t_{k+1}}(\hat{\bm{x}}^{*}_{k+1})=0_{N}, which further implies that 𝑳𝒙^k+1=0N2\bm{L}\hat{\bm{x}}^{*}_{k+1}=0_{N^{2}} from (23). There exist some x(tk+1)Nx^{*}(t_{k+1})\in{\mathbb{R}^{N}} such that 𝒙^k+1=1Nx(tk+1)\hat{\bm{x}}^{*}_{k+1}=1_{N}\otimes x^{*}(t_{k+1}) by the property of LL under the undirected connected graph 𝒢\mathcal{G}. Thus, 𝑭tk+1(1Nx(tk+1))=0N\bm{F}^{t_{k+1}}(1_{N}\otimes x^{*}(t_{k+1}))=0_{N} and Ftk+1(x(tk+1))=0NF^{t_{k+1}}(x^{*}(t_{k+1}))=0_{N}. In other words, x(tk+1)x^{*}(t_{k+1}) is the NE of Γtk+1\Gamma^{t_{k+1}} and 𝒙^k+1=1Nx(tk+1)=𝒙(tk+1)\hat{\bm{x}}^{*}_{k+1}=1_{N}\otimes x^{*}(t_{k+1})=\bm{x}^{*}(t_{k+1}).

Next, we prove (13). Before doing it, some coordinate transformations are required. Define M=[M1;M2]M=[M_{1};M_{2}] with M1=1NNM_{1}=\frac{1_{N}}{\sqrt{N}} and M2N×N1M_{2}\in{\mathbb{R}^{N\times{N-1}}}, which satisfy M2TM1=0N1M_{2}^{T}M_{1}=0_{N-1}, M2TM2=IN1M_{2}^{T}M_{2}=I_{N-1}, and M2M2T=INM1M1TM_{2}M_{2}^{T}=I_{N}-M_{1}M_{1}^{T}. The unitary matrix MM can convert LL to a diagonal form MTLM=diag(0,λ2,,λN)M^{T}LM=\text{diag}(0,\lambda_{2},\cdots,\lambda_{N}). Denote 𝒙~k+1s=𝒙^k+1s𝒙(tk+1)\tilde{\bm{x}}^{s}_{k+1}=\hat{\bm{x}}^{s}_{k+1}-\bm{x}^{*}(t_{k+1}) and perform the coordinate transformation as

𝒙¯k+1s,1=(M1TIN)𝒙~k+1s,𝒙¯k+1s,2=(M2TIN)𝒙~k+1s.\displaystyle\bar{\bm{x}}^{s,1}_{k+1}=(M_{1}^{T}\otimes I_{N})\tilde{\bm{x}}^{s}_{k+1},~{}\bar{\bm{x}}^{s,2}_{k+1}=(M_{2}^{T}\otimes I_{N})\tilde{\bm{x}}^{s}_{k+1}. (24)

Then inputting (22) into (24), we obtain that

𝒙¯k+1s+1,1𝒙¯k+1s,1\displaystyle\bar{\bm{x}}^{s+1,1}_{k+1}-\bar{\bm{x}}^{s,1}_{k+1} =(M1TIN)RΘk+1s,\displaystyle=-\left(M_{1}^{T}\otimes I_{N}\right)R\Theta^{s}_{k+1},
𝒙¯k+1s+1,2𝒙¯k+1s,2\displaystyle\bar{\bm{x}}^{s+1,2}_{k+1}-\bar{\bm{x}}^{s,2}_{k+1} =α[(M2TLM2)IN]𝒙¯k+1s,2\displaystyle=-\alpha\left[\left(M_{2}^{T}LM_{2}\right)\otimes I_{N}\right]\bar{\bm{x}}^{s,2}_{k+1} (25)
(M2TIN)RΘk+1s,\displaystyle\quad\quad\quad\quad\quad-\left(M_{2}^{T}\otimes I_{N}\right)R\Theta^{s}_{k+1},

where Θk+1s=𝑭tk+1(𝒙^k+1s)𝑭tk+1(𝒙(tk+1))\Theta^{s}_{k+1}=\bm{F}^{t_{k+1}}\left(\hat{\bm{x}}^{s}_{k+1}\right)-\bm{F}^{t_{k+1}}\left(\bm{x}^{*}(t_{k+1})\right).

Construct a Lyapunov function as follows,

Vk+1s(𝒙¯k+1s,1,𝒙¯k+1s,2)=𝒙¯k+1s,12+𝒙¯k+1s,22.\displaystyle V^{s}_{k+1}(\bar{\bm{x}}^{s,1}_{k+1},\bar{\bm{x}}^{s,2}_{k+1})=\|\bar{\bm{x}}^{s,1}_{k+1}\|^{2}+\|\bar{\bm{x}}^{s,2}_{k+1}\|^{2}.

By the facts (a+b)T(a+b)bTb=2aTb+aTa(a+b)^{T}(a+b)-b^{T}b=2a^{T}b+a^{T}a for any vectors a,ba,b and 𝒙¯k+1s+1,1=𝒙¯k+1s+1,1𝒙¯k+1s,1+𝒙¯k+1s,1\bar{\bm{x}}^{s+1,1}_{k+1}=\bar{\bm{x}}^{s+1,1}_{k+1}-\bar{\bm{x}}^{s,1}_{k+1}+\bar{\bm{x}}^{s,1}_{k+1}, it yields

Vk+1s+1Vk+1s=[𝒙¯k+1s+1,1𝒙¯k+1s,1]T[𝒙¯k+1s+1,1𝒙¯k+1s,1]\displaystyle V^{s+1}_{k+1}-V^{s}_{k+1}=[\bar{\bm{x}}^{s+1,1}_{k+1}-\bar{\bm{x}}^{s,1}_{k+1}]^{T}[\bar{\bm{x}}^{s+1,1}_{k+1}-\bar{\bm{x}}^{s,1}_{k+1}]
+2[𝒙¯k+1s+1,1𝒙¯k+1s,1]T𝒙¯k+1s,1+2[𝒙¯k+1s+1,2𝒙¯k+1s,2]T𝒙¯k+1s,2\displaystyle\quad+2[\bar{\bm{x}}^{s+1,1}_{k+1}-\bar{\bm{x}}^{s,1}_{k+1}]^{T}\bar{\bm{x}}^{s,1}_{k+1}+2[\bar{\bm{x}}^{s+1,2}_{k+1}-\bar{\bm{x}}^{s,2}_{k+1}]^{T}\bar{\bm{x}}^{s,2}_{k+1}
+[𝒙¯k+1s+1,2𝒙¯k+1s,2]T[𝒙¯k+1s+1,2𝒙¯k+1s,2].\displaystyle\quad+[\bar{\bm{x}}^{s+1,2}_{k+1}-\bar{\bm{x}}^{s,2}_{k+1}]^{T}[\bar{\bm{x}}^{s+1,2}_{k+1}-\bar{\bm{x}}^{s,2}_{k+1}]. (26)

Now we give upper bounds of each term of the right side of (26) based on (25). The first term is written as

[𝒙¯k+1s+1,1𝒙¯k+1s,1]T[𝒙¯k+1s+1,1𝒙¯k+1s,1](M1TIN)RΘk+1s2.[\bar{\bm{x}}^{s+1,1}_{k+1}-\bar{\bm{x}}^{s,1}_{k+1}]^{T}[\bar{\bm{x}}^{s+1,1}_{k+1}-\bar{\bm{x}}^{s,1}_{k+1}]\leq\|\left(M_{1}^{T}\otimes I_{N}\right)R\Theta^{s}_{k+1}\|^{2}.

Observing λ2M2TLM2λN\lambda_{2}\leq\|M_{2}^{T}LM_{2}\|\leq\lambda_{N}, we have

[𝒙¯k+1s+1,2𝒙¯k+1s,2]T[𝒙¯k+1s+1,2𝒙¯k+1s,2]\displaystyle\quad[\bar{\bm{x}}^{s+1,2}_{k+1}-\bar{\bm{x}}^{s,2}_{k+1}]^{T}[\bar{\bm{x}}^{s+1,2}_{k+1}-\bar{\bm{x}}^{s,2}_{k+1}]
2α(M2TLM2IN)𝒙¯k+1s,22+2(M2TIN)RΘk+1s2,\displaystyle\leq 2\|\alpha(M_{2}^{T}LM_{2}\otimes I_{N})\bar{\bm{x}}^{s,2}_{k+1}\|^{2}\!+\!2\|\left(M_{2}^{T}\otimes I_{N}\right)R\Theta^{s}_{k+1}\|^{2},
2α2λN2𝒙¯k+1s,22+2(M2TIN)RΘk+1s2.\displaystyle\leq 2\alpha^{2}\lambda_{N}^{2}\|\bar{\bm{x}}^{s,2}_{k+1}\|^{2}\!+\!2\|\left(M_{2}^{T}\otimes I_{N}\right)R\Theta^{s}_{k+1}\|^{2}.

Applying the same argument, we have

2[𝒙¯k+1s+1,1𝒙¯k+1s,1]T𝒙¯k+1s,1=2(𝒙¯k+1s,1)T(M1TIN)RΘk+1s,\displaystyle 2[\bar{\bm{x}}^{s+1,1}_{k+1}\!-\!\bar{\bm{x}}^{s,1}_{k+1}]^{T}\bar{\bm{x}}^{s,1}_{k+1}\!=\!-2(\bar{\bm{x}}^{s,1}_{k+1})^{T}\left(M_{1}^{T}\otimes I_{N}\right)R\Theta^{s}_{k+1},
2[𝒙¯k+1s+1,2𝒙¯k+1s,2]T𝒙¯k+1s,22αλ2𝒙¯k+1s,222(𝒙¯k+1s,2)T\displaystyle 2[\bar{\bm{x}}^{s+1,2}_{k+1}-\bar{\bm{x}}^{s,2}_{k+1}]^{T}\bar{\bm{x}}^{s,2}_{k+1}\leq-2\alpha\lambda_{2}\|\bar{\bm{x}}^{s,2}_{k+1}\|^{2}-2(\bar{\bm{x}}^{s,2}_{k+1})^{T}
(M2TIN)RΘk+1s.\displaystyle\quad\quad\quad\quad\quad\quad\quad\quad\quad\quad\quad\quad\quad\left(M_{2}^{T}\otimes I_{N}\right)R\Theta^{s}_{k+1}. (27)

By inserting (27) into (26), applying the relationships

{𝒙~k+1s=(𝒙¯k+1s,1)T[M1TIN]+(𝒙¯k+1s,2)T[M2TIN],RΘk+1s2=(M1TIN)RΘk+1s2+(M2TIN)RΘk+1s2,\begin{cases}\tilde{\bm{x}}^{s}_{k+1}=(\bar{\bm{x}}^{s,1}_{k+1})^{T}[M_{1}^{T}\otimes I_{N}]+(\bar{\bm{x}}^{s,2}_{k+1})^{T}[M_{2}^{T}\otimes I_{N}],\\ \|R\Theta^{s}_{k+1}\|^{2}=\|\left(M_{1}^{T}\otimes I_{N}\right)R\Theta^{s}_{k+1}\|^{2}\\ \quad\quad\quad\quad\quad\quad\quad\quad+\|\left(M_{2}^{T}\otimes I_{N}\right)R\Theta^{s}_{k+1}\|^{2},\end{cases}

and M2=1\|M_{2}\|=1, it derives that

Vk+1s+1Vk+1s2(𝒙~k+1s)TRΘk+1s+2RΘk+1s2\displaystyle V^{s+1}_{k+1}-V^{s}_{k+1}\leq-2(\tilde{\bm{x}}^{s}_{k+1})^{T}R\Theta^{s}_{k+1}+2\|R\Theta^{s}_{k+1}\|^{2}
+2α2λN2𝒙¯k+1s,222αλ2𝒙¯k+1s,22.\displaystyle\quad\quad\quad\quad\quad\quad\quad+2\alpha^{2}\lambda_{N}^{2}\|\bar{\bm{x}}^{s,2}_{k+1}\|^{2}-2\alpha\lambda_{2}\|\bar{\bm{x}}^{s,2}_{k+1}\|^{2}. (28)

We split 𝒙~k+1s\tilde{\bm{x}}^{s}_{k+1} into 𝒙~k+1s,1+𝒙~k+1s,2\tilde{\bm{x}}^{s,1}_{k+1}+\tilde{\bm{x}}^{s,2}_{k+1} to estimate

(𝒙~k+1s)TRΘk+1s\displaystyle\quad-(\tilde{\bm{x}}^{s}_{k+1})^{T}R\Theta^{s}_{k+1}
=(𝒙~k+1s,1)TR[𝑭tk+1(𝒙~k+1s,1+𝒙(tk+1))𝑭tk+1(𝒙(tk+1))]\displaystyle=-(\tilde{\bm{x}}^{s,1}_{k+1})^{T}R[\bm{F}^{t_{k+1}}(\tilde{\bm{x}}^{s,1}_{k+1}\!\!+\!\bm{x}^{*}(t_{k+1}))\!\!-\!\bm{F}^{t_{k+1}}\left(\bm{x}^{*}(t_{k+1})\right)]
(𝒙~k+1s,2)TR[𝑭tk+1(𝒙~k+1s,1+𝒙(tk+1))𝑭tk+1(𝒙(tk+1))]\displaystyle\quad-(\tilde{\bm{x}}^{s,2}_{k+1})^{T}R[\bm{F}^{t_{k+1}}(\tilde{\bm{x}}^{s,1}_{k+1}\!\!+\!\bm{x}^{*}(t_{k+1}))\!\!-\!\bm{F}^{t_{k+1}}\left(\bm{x}^{*}(t_{k+1})\right)]
(𝒙~k+1s,1)TR[𝑭tk+1(𝒙~k+1s,1+𝒙~k+1s,2+𝒙(tk+1))\displaystyle\quad-(\tilde{\bm{x}}^{s,1}_{k+1})^{T}R[\bm{F}^{t_{k+1}}(\tilde{\bm{x}}^{s,1}_{k+1}+\tilde{\bm{x}}^{s,2}_{k+1}+\bm{x}^{*}(t_{k+1}))
𝑭tk+1(𝒙~k+1s,1+𝒙(tk+1))]\displaystyle\quad\quad\quad\quad\quad\quad\quad\quad\quad\quad\quad-\bm{F}^{t_{k+1}}(\tilde{\bm{x}}^{s,1}_{k+1}+\bm{x}^{*}(t_{k+1}))]
(𝒙~k+1s,2)TR[𝑭tk+1(𝒙~k+1s,1+𝒙~k+1s,2+𝒙(tk+1))\displaystyle\quad-(\tilde{\bm{x}}^{s,2}_{k+1})^{T}R[\bm{F}^{t_{k+1}}(\tilde{\bm{x}}^{s,1}_{k+1}+\tilde{\bm{x}}^{s,2}_{k+1}+\bm{x}^{*}(t_{k+1}))
𝑭tk+1(𝒙~k+1s,2+𝒙(tk+1))].\displaystyle\quad\quad\quad\quad\quad\quad\quad\quad\quad\quad\quad-\bm{F}^{t_{k+1}}(\tilde{\bm{x}}^{s,2}_{k+1}+\bm{x}^{*}(t_{k+1}))]. (29)

As for 𝑭tk+1(1Ny)=Ftk+1(y)\bm{F}^{t_{k+1}}(1_{N}\otimes y)=F^{t_{k+1}}(y) for any yNy\in{\mathbb{R}^{N}}, it follows by the strong monotonically of Ftk+1F^{t_{k+1}} that

(𝒙~k+1s,1)TR[𝑭tk+1(𝒙~k+1s,1+𝒙(tk+1))𝑭tk+1(𝒙(tk+1))]\displaystyle(\tilde{\bm{x}}^{s,1}_{k+1})^{T}R[\bm{F}^{t_{k+1}}(\tilde{\bm{x}}^{s,1}_{k+1}+\bm{x}^{*}(t_{k+1}))-\bm{F}^{t_{k+1}}\left(\bm{x}^{*}(t_{k+1})\right)]
=ε(𝒙¯k+1s,1)TN[Ftk+1((𝒙¯k+1s,1)TN+x(tk+1))\displaystyle=\frac{\varepsilon(\bar{\bm{x}}^{s,1}_{k+1})^{T}}{\sqrt{N}}[F^{t_{k+1}}(\frac{(\bar{\bm{x}}^{s,1}_{k+1})^{T}}{\sqrt{N}}+x^{*}(t_{k+1}))
Ftk+1(x(tk+1))]εμN𝒙¯s,1k+12,\displaystyle\quad\quad\quad\quad\quad\quad\quad-F^{t_{k+1}}(x^{*}(t_{k+1}))]\geq\frac{\varepsilon\mu}{N}\|\bar{\bm{x}}^{s,1}_{k+1}\|^{2}, (30)

where (1NTIN)R=IN(1_{N}^{T}\otimes I_{N})R=I_{N} and (𝒙~k+1s,1)TR=ε(𝒙¯k+1s,1)TN(\tilde{\bm{x}}^{s,1}_{k+1})^{T}R=\frac{\varepsilon(\bar{\bm{x}}^{s,1}_{k+1})^{T}}{\sqrt{N}} are used. Since RT𝒙~k+1s,2=ε𝒙~k+1s,2=ε𝒙¯k+1s,2\|R^{T}\tilde{\bm{x}}^{s,2}_{k+1}\|=\varepsilon\|\tilde{\bm{x}}^{s,2}_{k+1}\|=\varepsilon\|\bar{\bm{x}}^{s,2}_{k+1}\|,

(𝒙~k+1s,2)TR[𝑭tk+1(𝒙~k+1s,1+𝒙(tk+1))𝑭tk+1(𝒙(tk+1))]\displaystyle(\tilde{\bm{x}}^{s,2}_{k+1})^{T}R[\bm{F}^{t_{k+1}}(\tilde{\bm{x}}^{s,1}_{k+1}+\bm{x}^{*}(t_{k+1}))-\bm{F}^{t_{k+1}}\left(\bm{x}^{*}(t_{k+1})\right)]
εθtk+1N𝒙¯k+1s,1𝒙¯k+1s,2.\displaystyle\leq\frac{\varepsilon\theta^{t_{k+1}}}{\sqrt{N}}\|\bar{\bm{x}}^{s,1}_{k+1}\|\|\bar{\bm{x}}^{s,2}_{k+1}\|. (31)

Similar operations are used to obtain that

[𝒙~k+1s]TRΘk+1s\displaystyle-[\tilde{\bm{x}}^{s}_{k+1}]^{T}R\Theta^{s}_{k+1} 2εθtk+1N𝒙¯k+1s,1𝒙¯k+1s,2\displaystyle\leq\frac{2\varepsilon\theta^{t_{k+1}}}{\sqrt{N}}\|\bar{\bm{x}}^{s,1}_{k+1}\|\|\bar{\bm{x}}^{s,2}_{k+1}\|
+εθtk+1𝒙¯k+1s,22εμtk+1N𝒙¯k+1s,12.\displaystyle\quad\quad+\varepsilon\theta^{t_{k+1}}\|\bar{\bm{x}}^{s,2}_{k+1}\|^{2}\!-\!\frac{\varepsilon\mu^{t_{k+1}}}{N}\|\bar{\bm{x}}^{s,1}_{k+1}\|^{2}.

Then the difference of Vk+1V_{k+1} in (28) is reorganized as

Vk+1s+1Vk+1s\displaystyle\quad V^{s+1}_{k+1}-V^{s}_{k+1}
4εθtk+1N𝒙¯k+1s,1𝒙¯k+1s,22εμtk+1N𝒙¯k+1s,12+2RΘk+1s2\displaystyle\leq\frac{4\varepsilon\theta^{t_{k+1}}}{\sqrt{N}}\|\bar{\bm{x}}^{s,1}_{k+1}\|\|\bar{\bm{x}}^{s,2}_{k+1}\|\!-\!\frac{2\varepsilon\mu^{t_{k+1}}}{N}\|\bar{\bm{x}}^{s,1}_{k+1}\|^{2}\!+\!2\|R\Theta^{s}_{k+1}\|^{2}
+(2εθtk+1+2α2λN22αλ2)𝒙¯k+1s,22.\displaystyle\quad\quad\quad\quad\quad\quad+(2\varepsilon\theta^{t_{k+1}}+2\alpha^{2}\lambda_{N}^{2}-2\alpha\lambda_{2})\|\bar{\bm{x}}^{s,2}_{k+1}\|^{2}. (32)

We further bound RΘk+1s\|R\Theta^{s}_{k+1}\| by

RΘk+1s=R[𝑭tk+1(𝒙~k+1s,1+𝒙~k+1s,2+𝒙(tk+1))\displaystyle R\Theta^{s}_{k+1}=R[\bm{F}^{t_{k+1}}(\tilde{\bm{x}}^{s,1}_{k+1}+\tilde{\bm{x}}^{s,2}_{k+1}+\bm{x}^{*}(t_{k+1}))
𝑭tk+1(𝒙~k+1s,1+𝒙(tk+1))]+R[𝑭tk+1(𝒙~k+1s,1+𝒙(tk+1))\displaystyle~{}-\bm{F}^{t_{k+1}}(\tilde{\bm{x}}^{s,1}_{k+1}\!+\!\bm{x}^{*}(t_{k+1}))]\!+\!R[\bm{F}^{t_{k+1}}(\tilde{\bm{x}}^{s,1}_{k+1}\!+\!\bm{x}^{*}(t_{k+1}))
𝑭tk+1(𝒙)]εθtk+1𝒙¯s,2k+1+εθtk+1𝒙¯s,1k+1.\displaystyle~{}-\bm{F}^{t_{k+1}}\left(\bm{x}^{*}\right)]\leq\varepsilon\theta^{t_{k+1}}\|\bar{\bm{x}}^{s,2}_{k+1}\|+\varepsilon\theta^{t_{k+1}}\|\bar{\bm{x}}^{s,1}_{k+1}\|. (33)

To sum up, we conclude that

Vk+1s+1Vk+1s\displaystyle V^{s+1}_{k+1}\!-\!V^{s}_{k+1} 2[𝒙¯k+1s,1𝒙¯k+1s,2]Aα,ε[𝒙¯k+1s,1𝒙¯k+1s,2],\displaystyle\leq\!-2[\|\bar{\bm{x}}^{s,1}_{k+1}\|~{}~{}\|\bar{\bm{x}}^{s,2}_{k+1}\|]A_{\alpha,\varepsilon}\left[\begin{matrix}\|\bar{\bm{x}}^{s,1}_{k+1}\|\\ \|\bar{\bm{x}}^{s,2}_{k+1}\|\end{matrix}\right],
νVk+1s,\displaystyle\leq-\nu V^{s}_{k+1}, (34)

which implies that

𝒙^k+1s+1𝒙(tk+1)2(1ν)𝒙^k+1s𝒙(tk+1)2.\displaystyle\|\hat{\bm{x}}^{s+1}_{k+1}-\bm{x}^{*}(t_{k+1})\|^{2}\leq(1-\nu)\|\hat{\bm{x}}^{s}_{k+1}-\bm{x}^{*}(t_{k+1})\|^{2}. (35)

Since 𝒙^k+1s\hat{\bm{x}}^{s}_{k+1} is initialized by the predicted variable 𝒙k+1|k\bm{x}_{k+1|k} and the corrected variable 𝒙k+1=𝒙^k+1τ\bm{x}_{k+1}=\hat{\bm{x}}^{\tau}_{k+1}, considering the relation in (35) between two consecutive iterates of the sequence 𝒙^k+1s\hat{\bm{x}}^{s}_{k+1}, we obtain (13).

References

  • Başar \BBA Olsder (\APACyear1998) \APACinsertmetastarNE{APACrefauthors}Başar, T.\BCBT \BBA Olsder, G\BPBIJ.  \APACrefYear1998. \APACrefbtitleDynamic Noncooperative Game Theory Dynamic noncooperative game theory. \APACaddressPublisherSIAM. \PrintBackRefs\CurrentBib
  • Carnevale \BOthers. (\APACyear2022) \APACinsertmetastarfangzhen{APACrefauthors}Carnevale, G., Camisa, A.\BCBL \BBA Notarstefano, G.  \APACrefYearMonthDay2022. \BBOQ\APACrefatitleDistributed Online Aggregative Optimization for Dynamic Multi-Robot Coordination Distributed online aggregative optimization for dynamic multi-robot coordination.\BBCQ \APACjournalVolNumPagesIEEE Trans. Autom. Control1-8. {APACrefDOI} 10.1109/TAC.2022.3196627 \PrintBackRefs\CurrentBib
  • Chen \BOthers. (\APACyear2022) \APACinsertmetastarstaticNE4{APACrefauthors}Chen, Z., Ma, J., Liang, S.\BCBL \BBA Li, L.  \APACrefYearMonthDay2022. \BBOQ\APACrefatitleDistributed Nash equilibrium seeking under quantization communication Distributed nash equilibrium seeking under quantization communication.\BBCQ \APACjournalVolNumPagesAutomatica141110318. \PrintBackRefs\CurrentBib
  • De Persis \BBA Grammatico (\APACyear2019) \APACinsertmetastarstaticNE2{APACrefauthors}De Persis, C.\BCBT \BBA Grammatico, S.  \APACrefYearMonthDay2019. \BBOQ\APACrefatitleDistributed averaging integral Nash equilibrium seeking on networks Distributed averaging integral nash equilibrium seeking on networks.\BBCQ \APACjournalVolNumPagesAutomatica110108548. \PrintBackRefs\CurrentBib
  • Gadjov \BBA Pavel (\APACyear2018) \APACinsertmetastarstaticNE1{APACrefauthors}Gadjov, D.\BCBT \BBA Pavel, L.  \APACrefYearMonthDay2018. \BBOQ\APACrefatitleA passivity-based approach to Nash equilibrium seeking over networks A passivity-based approach to nash equilibrium seeking over networks.\BBCQ \APACjournalVolNumPagesIEEE Trans. Automat. Control6431077–1092. \PrintBackRefs\CurrentBib
  • Ghaderi \BBA Srikant (\APACyear2014) \APACinsertmetastarsocial{APACrefauthors}Ghaderi, J.\BCBT \BBA Srikant, R.  \APACrefYearMonthDay2014. \BBOQ\APACrefatitleOpinion dynamics in social networks with stubborn agents: Equilibrium and convergence rate Opinion dynamics in social networks with stubborn agents: Equilibrium and convergence rate.\BBCQ \APACjournalVolNumPagesAutomatica50123209–3215. \PrintBackRefs\CurrentBib
  • Huang \BOthers. (\APACyear2022) \APACinsertmetastarmulticluster{APACrefauthors}Huang, B., Yang, C., Meng, Z., Chen, F.\BCBL \BBA Ren, W.  \APACrefYearMonthDay2022. \BBOQ\APACrefatitleDistributed nonlinear placement for multicluster systems: A time-varying nash equilibrium-seeking approach Distributed nonlinear placement for multicluster systems: A time-varying nash equilibrium-seeking approach.\BBCQ \APACjournalVolNumPagesIEEE Trans. Cybern.521111614–11623. \PrintBackRefs\CurrentBib
  • Liu \BOthers. (\APACyear2022) \APACinsertmetastarADMM1{APACrefauthors}Liu, Y., Wu, G., Tian, Z.\BCBL \BBA Ling, Q.  \APACrefYearMonthDay2022. \BBOQ\APACrefatitleDQC-ADMM: Decentralized Dynamic ADMM With Quantized and Censored Communications Dqc-admm: Decentralized dynamic admm with quantized and censored communications.\BBCQ \APACjournalVolNumPagesIEEE Trans. Neural Netw. Learn. Syst.3383290-3304. {APACrefDOI} 10.1109/TNNLS.2021.3051638 \PrintBackRefs\CurrentBib
  • Lu \BOthers. (\APACyear2020) \APACinsertmetastaronline3{APACrefauthors}Lu, K., Li, G.\BCBL \BBA Wang, L.  \APACrefYearMonthDay2020. \BBOQ\APACrefatitleOnline distributed algorithms for seeking generalized Nash equilibria in dynamic environments Online distributed algorithms for seeking generalized nash equilibria in dynamic environments.\BBCQ \APACjournalVolNumPagesIEEE Trans. Autom. Control6652289–2296. \PrintBackRefs\CurrentBib
  • Simonetto \BOthers. (\APACyear2017) \APACinsertmetastarsimonetto2016{APACrefauthors}Simonetto, A., Koppel, A., Mokhtari, A., Leus, G.\BCBL \BBA Ribeiro, A.  \APACrefYearMonthDay2017. \BBOQ\APACrefatitleDecentralized prediction-correction methods for networked time-varying convex optimization Decentralized prediction-correction methods for networked time-varying convex optimization.\BBCQ \APACjournalVolNumPagesIEEE Trans. Autom. Control62115724–5738. \PrintBackRefs\CurrentBib
  • Stankovic \BOthers. (\APACyear2011) \APACinsertmetastarsensor{APACrefauthors}Stankovic, M\BPBIS., Johansson, K\BPBIH.\BCBL \BBA Stipanovic, D\BPBIM.  \APACrefYearMonthDay2011. \BBOQ\APACrefatitleDistributed seeking of Nash equilibria with applications to mobile sensor networks Distributed seeking of nash equilibria with applications to mobile sensor networks.\BBCQ \APACjournalVolNumPagesIEEE Trans. Automat. Control574904–919. \PrintBackRefs\CurrentBib
  • Su \BOthers. (\APACyear2021) \APACinsertmetastarsu2021online{APACrefauthors}Su, Y., Liu, F., Wang, Z., Mei, S.\BCBL \BBA Lu, Q.  \APACrefYearMonthDay2021. \BBOQ\APACrefatitleOnline distributed tracking of generalized Nash equilibrium on physical networks Online distributed tracking of generalized nash equilibrium on physical networks.\BBCQ \APACjournalVolNumPagesAutonomous Intelligent Systems111–12. \PrintBackRefs\CurrentBib
  • Ye \BBA Hu (\APACyear2015) \APACinsertmetastaryemaojiao{APACrefauthors}Ye, M.\BCBT \BBA Hu, G.  \APACrefYearMonthDay2015. \BBOQ\APACrefatitleDistributed seeking of time-varying Nash equilibrium for non-cooperative games Distributed seeking of time-varying nash equilibrium for non-cooperative games.\BBCQ \APACjournalVolNumPagesIEEE Trans. Autom. Control60113000–3005. \PrintBackRefs\CurrentBib
  • Zeng \BOthers. (\APACyear2019) \APACinsertmetastarsmart{APACrefauthors}Zeng, X., Chen, J., Liang, S.\BCBL \BBA Hong, Y.  \APACrefYearMonthDay2019. \BBOQ\APACrefatitleGeneralized Nash equilibrium seeking strategy for distributed nonsmooth multi-cluster game Generalized nash equilibrium seeking strategy for distributed nonsmooth multi-cluster game.\BBCQ \APACjournalVolNumPagesAutomatica10320–26. \PrintBackRefs\CurrentBib
  • Zhu \BOthers. (\APACyear2020) \APACinsertmetastarstaticNE3{APACrefauthors}Zhu, Y., Yu, W., Wen, G.\BCBL \BBA Chen, G.  \APACrefYearMonthDay2020. \BBOQ\APACrefatitleDistributed Nash equilibrium seeking in an aggregative game on a directed graph Distributed nash equilibrium seeking in an aggregative game on a directed graph.\BBCQ \APACjournalVolNumPagesIEEE Trans. Automat. Control6662746–2753. \PrintBackRefs\CurrentBib