This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

On Controllability and Persistency of Excitation in Data-Driven Control: Extensions of Willems’ Fundamental Lemma

Yue Yu, Shahriar Talebi, Henk J. van Waarde, Ufuk Topcu, Mehran Mesbahi, and Behçet Açıkmeşe Yue Yu and Ufuk Topcu are with the Department of Aerospace Engineering, University of Texas at Austin, Austin, TX 78712 USA. Shahriar Talebi, Mehran Mesbahi and Behçet Açıkmeşe are with the Department of Aeronautics and Astronautics, University of Washington, Seattle, WA 98195 USA. Henk J. van Waarde is with the Control Group, Department of Engineering, University of Cambridge, Trumpington Street, Cambridge CB2 1PZ, UK. (emails:[email protected], [email protected], [email protected], [email protected], [email protected], [email protected])
Abstract

Willems’ fundamental lemma asserts that all trajectories of a linear time-invariant system can be obtained from a finite number of measured ones, assuming that controllability and a persistency of excitation condition hold. We show that these two conditions can be relaxed. First, we prove that the controllability condition can be replaced by a condition on the controllable subspace, unobservable subspace, and a certain subspace associated with the measured trajectories. Second, we prove that the persistency of excitation requirement can be relaxed if the degree of a certain minimal polynomial is tightly bounded. Our results show that data-driven predictive control using online data is equivalent to model predictive control, even for uncontrollable systems. Moreover, our results significantly reduce the amount of data needed in identifying homogeneous multi-agent systems.

I Introduction

Willems’ fundamental lemma provides a data-based parameterization of trajectories generated by linear time invariant (LTI) systems [1]. In particular, consider the LTI system

xt+1=\displaystyle x_{t+1}= Axt+But,\displaystyle Ax_{t}+Bu_{t}, (1a)
yt=\displaystyle y_{t}= Cxt+Dut,\displaystyle Cx_{t}+Du_{t}, (1b)

where utm,xtn,ytpu_{t}\in\mathbb{R}^{m},x_{t}\in\mathbb{R}^{n},y_{t}\in\mathbb{R}^{p} denote the input, state and output of the system at discrete time tt, respectively. If system (1) is controllable, the lemma asserts that every length-LL input-output trajectory of system (1) is a linear combination of a finite number of measured ones. These measured trajectories can be extracted from one single trajectory with persistent excitation of order n+Ln+L [1], or multiple trajectories with collective persistent excitation of order n+Ln+L [2]. By parameterizing trajectories of the system (1) using measured data, the lemma has profound implications in system identification [3, 4, 5], and inspired a series of recent results including data-driven simulation [3, 6], output matching [7], control by interconnection [8], set-invariance control [9], linear quadratic regulation [10], and predictive control [11, 12, 13, 14, 15, 16, 17].

In the meantime, whether conditions of controllability and persistency of excitation in Willems’ fundamental lemma are necessary has not been investigated in depth. The current work aims to address this issue by answering the following two questions. First, without controllability, to what extent can the linear combinations of a finite number of measured input-output trajectories parameterize all possible ones? Second, can the order of persistency of excitation be reduced?

Recent results on system identification has shown that, assuming sufficient persistency of excitation, the linear combinations of a finite number of measured input-output trajectories contain any trajectory whose initial state is in the controllable subspace [18, 19]. As we show subsequently, such results only partially answer the first question above: trajectories with initial state outside the controllable subspace can also be contained in said linear combinations.

We answer the aforementioned questions by introducing extensions of Willems’ fundamental lemma. First, we show that, with sufficient persistency of excitation, any length-LL input-output trajectory whose initial state is in some subspace is a linear combination of a finite number of measured ones. Said subspace is the sum of the controllable subspace, unobservable subspace, and a certain subspace associated with the measured trajectories. Second, we show that the order of persistency of excitation required by Willems’ fundamental lemma can be reduced from n+Ln+L to δmin+L\delta_{\min}+L, where δmin\delta_{\min} is the degree of the minimal polynomial of the system matrix AA. Our first result completes those presented in [18, 19] by showing exactly which trajectories are parameterizable by a finite number of measured ones for an arbitrary LTI systems. Furthermore, this result shows that data-driven predictive control using online data is equivalent to model predictive control, not only for controllable systems, as shown in [12, 13], but also for uncontrollable ones. Our second result, compared with those in [2], can reduce the amount of data samples used in identifying homogeneous multi-agent systems by an order of magnitude.

The rest of the paper is organized as follows. We first prove an extended Willems’ fundamental lemma and discuss its implications in Section II. We provide ramifications of this extension for a representative set of applications in Section III before providing concluding remarks in Section IV.

Notation: We let \mathbb{R}, and \mathbb{N} and +\mathbb{N}_{+} denote the set of real numbers, non-negative integers, and positive integers, respectively. The image and right kernel of matrix MM is denoted by imM\operatorname{im}M and kerM\ker M, respectively. Let xM=xMx\left\lVert x\right\rVert_{M}=x^{\top}Mx for any matrix MM and vector xx. When applied to subspaces, we let ++ and ×\times denote the sum [20, p.2] and Cartesian product operation [20, p.370], respectively. We let \otimes denote the Kronecker product. Given a signal f:qf:\mathbb{N}\to\mathbb{R}^{q} and i,ji,j\in\mathbb{N} with iji\leq j, we denote f[i,j]=[fifi+1fj]f_{[i,j]}=\begin{bmatrix}f_{i}^{\top}&f_{i+1}^{\top}&\cdots&f_{j}^{\top}\end{bmatrix}^{\top}, and the Hankel matrix of depth dd (dji+1d\leq j-i+1) associated with f[i,j]f_{[i,j]} as

Hd(f[i,j])=[fifi+1fjd+1fi+1fi+2fjd+2fi+d1fi+dfj].H_{d}(f_{[i,j]})=\begin{bmatrix}f_{i}&f_{i+1}&\cdots&f_{j-d+1}\\ f_{i+1}&f_{i+2}&\cdots&f_{j-d+2}\\ \vdots&\vdots&&\vdots\\ f_{i+d-1}&f_{i+d}&\cdots&f_{j}\end{bmatrix}.

II Extensions of Willems’ fundamental lemma

In this section, we introduce extensions of Willems’ fundamental lemma. Throughout we let

(u[0,Ti1]i,x[0,Ti1]i,y[0,Ti1]i)(u^{i}_{[0,T^{i}-1]},x^{i}_{[0,T^{i}-1]},y^{i}_{[0,T^{i}-1]}) (2)

denote a length-TiT^{i} (Ti+T^{i}\in\mathbb{N}_{+}) input-state-output trajectory generated by system (1) for all i=1,2,,τi=1,2,\ldots,\tau, where τ+\tau\in\mathbb{N}_{+} is the total number of trajectories. We let

{u[0,Ti1]i}i=1τ,{x[0,Ti1]i}i=1τ,{y[0,Ti1]i}i=1τ,\{u^{i}_{[0,T^{i}-1]}\}_{i=1}^{\tau},\enskip\{x^{i}_{[0,T^{i}-1]}\}_{i=1}^{\tau},\enskip\{y^{i}_{[0,T^{i}-1]}\}_{i=1}^{\tau}, (3)

denote the set of input, state, and output trajectories, respectively. We will use the following subspaces,

=im[BABAn1B],\displaystyle\mathcal{R}=\operatorname{im}\begin{bmatrix}B&AB&\cdots&A^{n-1}B\end{bmatrix}, (4)
𝒪=ker[C(CA)(CAn1)],\displaystyle\mathcal{O}=\ker\begin{bmatrix}C^{\top}&(CA)^{\top}&\cdots&(CA^{n-1})^{\top}\end{bmatrix}^{\top},
𝒦[x01,x02,,x0τ]=im[X0AX0An1X0],\displaystyle\mathcal{K}[x_{0}^{1},x_{0}^{2},\ldots,x_{0}^{\tau}]=\operatorname{im}\begin{bmatrix}X_{0}&AX_{0}&\cdots&A^{n-1}X_{0}\end{bmatrix},

where X0=[x01x02x0τ]X_{0}=\begin{bmatrix}x_{0}^{1}&x_{0}^{2}&\cdots&x_{0}^{\tau}\end{bmatrix}, and x0ix_{0}^{i} is the first state in the trajectory x[0,Ti1]ix_{[0,T^{i}-1]}^{i} for all 1iτ1\leq i\leq\tau. In particular, \mathcal{R} is known as the controllable subspace (we say that system (1) is controllable if =n\mathcal{R}=\mathbb{R}^{n}), 𝒪\mathcal{O} is known as the unobservable subspace, and 𝒦[x01,x02,,x0τ]\mathcal{K}[x_{0}^{1},x_{0}^{2},\ldots,x_{0}^{\tau}] is the smallest AA-invariant subspace containing x01,x02,,x0τx_{0}^{1},x_{0}^{2},\ldots,x_{0}^{\tau}. Using the Cayley-Hamilton theorem, one can verify that (1) ensures

xt=Atx0+j=0t1Atj1Buj+𝒦[x0],x_{t}=\textstyle A^{t}x_{0}+\sum_{j=0}^{t-1}A^{t-j-1}Bu_{j}\in\mathcal{R}+\mathcal{K}[x_{0}], (5)

for all tt\in\mathbb{N}, where 𝒦[x0]\mathcal{K}[x_{0}] is defined similar to 𝒦[x01,x02,,x0τ]\mathcal{K}[x_{0}^{1},x_{0}^{2},\ldots,x_{0}^{\tau}]. We let δ+\delta\in\mathbb{N}_{+} be such that

δδmindegree of pmin(A),\delta\geq\delta_{\min}\coloneqq\text{degree of }p_{\min}(A), (6)

where pmin(A)p_{\min}(A) is the minimal polynomial of matrix AA [20, Def. 3.3.2], and, due to the Cayley-Hamilton theorem, δmin\delta_{\min} is upper bounded by nn. We will also use the following definitions that streamline our subsequent analysis.

Definition 1.

We say a length-LL input-output trajectory (u¯[0,L1],y¯[0,L1])(\widebar{u}_{[0,L-1]},\widebar{y}_{[0,L-1]}), with L+L\in\mathbb{N}_{+}, is parameterizable by {u[0,Ti1],y[0,Ti1]}i=1τ\{u_{[0,T^{i}-1]},y_{[0,T^{i}-1]}\}_{i=1}^{\tau} if there exists gi=1τ(TiL+1)g\in\mathbb{R}^{\sum_{i=1}^{\tau}(T^{i}-L+1)} such that

[u¯[0,L1]y¯[0,L1]]=[HL(u[0,T11]1)HL(u[0,Tτ1]τ)HL(y[0,T11]1)HL(y[0,Tτ1]τ)]g.\displaystyle\begin{bmatrix}\widebar{u}_{[0,L-1]}\\ \widebar{y}_{[0,L-1]}\end{bmatrix}=\begin{bmatrix}H_{L}(u^{1}_{[0,T^{1}-1]})&\cdots&H_{L}(u^{\tau}_{[0,T^{\tau}-1]})\\ H_{L}(y^{1}_{[0,T^{1}-1]})&\cdots&H_{L}(y^{\tau}_{[0,T^{\tau}-1]})\end{bmatrix}g. (7)

As an example, if τ=2\tau=2, L=2L=2, T1=3,T2=4T^{1}=3,T^{2}=4, then (7) assumes the form,

[u¯0u¯1\hdashline[2pt/2pt]y¯0y¯1]=[u01u11u_0^2u_1^2u22u11u21u_1^2u_2^2u32\hdashline[2pt/2pt]y01y11y_0^2y_1^2y22y11y21y_1^2y_2^2y32]g.\left[\begin{array}[]{c}\widebar{u}_{0}\\ \widebar{u}_{1}\\ \hdashline[2pt/2pt]\widebar{y}_{0}\\ \widebar{y}_{1}\end{array}\right]=\left[\begin{array}[]{cc;{2pt/2pt}ccc}u_{0}^{1}&u_{1}^{1}&u_0^2&u_1^2&u_{2}^{2}\\ u_{1}^{1}&u_{2}^{1}&u_1^2&u_2^2&u_{3}^{2}\\ \hdashline[2pt/2pt]y_{0}^{1}&y_{1}^{1}&y_0^2&y_1^2&y_{2}^{2}\\ y_{1}^{1}&y_{2}^{1}&y_1^2&y_2^2&y_{3}^{2}\end{array}\right]g.
Definition 2 (Collective persistent excitation [2]).

We say that {u[0,Ti1]}i=1τ\{u_{[0,T^{i}-1]}\}_{i=1}^{\tau} is collectively persistently exciting of order d+d\in\mathbb{N}_{+} if dTid\leq T^{i} for all i=1,2,,τi=1,2,\ldots,\tau, and the mosaic-Hankel matrix,

[Hd(u[0,T11]1)Hd(u[0,Tτ1]τ)],\begin{bmatrix}H_{d}(u^{1}_{[0,T^{1}-1]})&\cdots&H_{d}(u^{\tau}_{[0,T^{\tau}-1]})\end{bmatrix}, (8)

has full row rank.

If τ=1\tau=1, then Definition 2 reduces to the traditional notion of persistency of excitation [1].

The Willems’ fundamental lemma asserts that: if the system (1) is controllable and {u[0,Ti1]i}i=1τ\{u^{i}_{[0,T^{i}-1]}\}_{i=1}^{\tau} is collectively persistently exciting of order n+Ln+L, then (u¯[0,L1],y¯[0,L1])(\widebar{u}_{[0,L-1]},\widebar{y}_{[0,L-1]}) is parameterizable by {u[0,Ti1],y[0,Ti1]}i=1τ\{u_{[0,T^{i}-1]},y_{[0,T^{i}-1]}\}_{i=1}^{\tau} if and only if it is an input-output trajectory of the system (1) [1, 2]. As our main contribution, the following theorem shows that not only the controllability assumption can be relaxed, but also the required order of collective persistent excitation can be reduced from n+Ln+L to, at best, δmin+L\delta_{\min}+L.

Theorem 1.

Let {u[0,Ti1]i,x[0,Ti1]i,y[0,Ti1]i}i=1τ\{u^{i}_{[0,T^{i}-1]},x^{i}_{[0,T^{i}-1]},y^{i}_{[0,T^{i}-1]}\}_{i=1}^{\tau} be a set of input-state-output trajectories generated by the system (1). If {u[0,Ti1]}i=1τ\{u_{[0,T^{i}-1]}\}_{i=1}^{\tau} is collectively persistently exciting of order δ+L\delta+L where δ\delta satisfies (6), then

im[H1(x[0,T1L]1)H1(x[0,TτL]τ)HL(u[0,T11]1)HL(u[0,Tτ1]τ)]\displaystyle\operatorname{im}\begin{bmatrix}H_{1}(x^{1}_{[0,T^{1}-L]})&\cdots&H_{1}(x^{\tau}_{[0,T^{\tau}-L]})\\ H_{L}(u^{1}_{[0,T^{1}-1]})&\cdots&H_{L}(u^{\tau}_{[0,T^{\tau}-1]})\end{bmatrix} (9)
=(+𝒦[x01,x02,,x0τ])×mL.\displaystyle=(\mathcal{R}+\mathcal{K}[x^{1}_{0},x^{2}_{0},\ldots,x^{\tau}_{0}])\times\mathbb{R}^{mL}.

Further, (u¯[0,L1],y¯[0,L1])(\widebar{u}_{[0,L-1]},\widebar{y}_{[0,L-1]}) is parameterizable by {u[0,Ti1],y[0,Ti1]}i=1τ\{u_{[0,T^{i}-1]},y_{[0,T^{i}-1]}\}_{i=1}^{\tau} if and only if there exists a state trajectory x¯[0,L1]\widebar{x}_{[0,L-1]} with

x¯0+𝒪+𝒦[x01,x02,,x0τ],\widebar{x}_{0}\in\mathcal{R}+\mathcal{O}+\mathcal{K}[x_{0}^{1},x_{0}^{2},\ldots,x_{0}^{\tau}], (10)

such that (u¯[0,L1],x¯[0,L1],y¯[0,L1])(\widebar{u}_{[0,L-1]},\widebar{x}_{[0,L-1]},\widebar{y}_{[0,L-1]}) is an input-state-output trajectory generated by (1).

Proof.

See the Appendix. ∎

Remark 1.

The equality in (9) generalizes [18, Lem. 2] by proving stronger results using weaker assumptions. Particularly, the assumption of n+Ln+L order of persistent excitation in [18, Lem. 2] is reduced to δ+L\delta+L with δδmin\delta\geq\delta_{\min}, and the controllable subspace \mathcal{R} used in [18, Lem. 2] is extended to its superset +𝒦[x01,x02,,x0τ]\mathcal{R}+\mathcal{K}[x_{0}^{1},x_{0}^{2},\ldots,x_{0}^{\tau}].

Due to its dependence on state x¯0\widebar{x}_{0} and the unknown subspaces in (4), the condition in (10) is difficult to verify, especially if merely input-output trajectories are available. One approach to guarantee its satisfaction is to ensure the right hand side of (10) equals n\mathbb{R}^{n}, either by assuming controllability (i.e., =n\mathcal{R}=\mathbb{R}^{n}) as in [1] and [2], or requiring the input-state-output trajectories in (2) satisfy im[x01x02x0τ]=n\operatorname{im}\begin{bmatrix}x_{0}^{1}&x_{0}^{2}&\ldots&x_{0}^{\tau}\end{bmatrix}=\mathbb{R}^{n}. The following corollary, however, provides an alternative approach that requires neither controllability nor state measurements.

Corollary 1.

Let (u[0,K1],y[0,K1])(u_{[0,K-1]},y_{[0,K-1]}) be an input-output trajectory generated by the system (1), and u[0,T1]u_{[0,T-1]} be persistently exciting of order δ+L\delta+L, where δ\delta satisfies (6) and LTKL\leq T\leq K. Then (u[t,t+L1],y[t,t+L1])(u_{[t,t+L-1]},y_{[t,t+L-1]}) is parameterizable by (u[0,T1],y[0,T1])(u_{[0,T-1]},y_{[0,T-1]}) for all 0tKL0\leq t\leq K-L. Conversely, if (u¯[0,L1],y¯[0,L1])(\widebar{u}_{[0,L-1]},\widebar{y}_{[0,L-1]}) is parameterizable by (u[0,T1],y[0,T1])(u_{[0,T-1]},y_{[0,T-1]}), then it is an input-output trajectory generated by the system (1).

Proof.

See the Appendix. ∎

Corollary 1 shows that any length-LL segment of an input-output trajectory is parameterizable by its first length-TT segment, assuming sufficient persistency of excitation. This result is particularly useful in predictive control as we will show in Section III-A.

Another implication of Theorem 1 is that the order of persistent excitation required by trajectory parameterization only depends on the degree of the minimal polynomial of matrix AA in (1), instead of its dimension. In general, it is difficult to establish a bound on the degree of the minimal polynomial of an unknown matrix tighter than its dimension. However, the following corollary shows an exception; its usefulness will be illustrated later in Section III-B.

Corollary 2.

If there exists A¯n¯×n¯\widebar{A}\in\mathbb{R}^{\widebar{n}\times\widebar{n}} such that A=INA¯A=I_{N}\otimes\widebar{A}, where n¯,N+\widebar{n},N\in\mathbb{N}_{+}, then Theorem 1 holds with δ=n¯\delta=\widebar{n}.

The proof follows from Theorem 1 and the fact that the minimal polynomial of AA is the same as that of A¯\widebar{A}, which has degree at most n¯\widebar{n}.

III Applications

In this section, we provide two examples that illustrate the distinct implications of Theorem 1.

III-A Online data-enabled predictive control

Model predictive control (MPC) provides an effective control strategy for systems with physical and operational constraints [21, 22]. In particular, consider system (1) with output feedback. At each sampling time t+t\in\mathbb{N}_{+}, given the input-output measurements history (u[0,t1],y[0,t1])(u_{[0,t-1]},y_{[0,t-1]}), MPC solves the following optimization for input u¯t\widebar{u}_{t}

minimizex¯[tN,t+L1]u¯[t,t+L1]y¯[t,t+L1]k=tt+L1(y¯krkQ2+u¯kR2)subject tox¯k+1=Ax¯k+Buk,yk=Cx¯k+Duk,tNkt1,x¯k+1=Ax¯k+Bu¯k,y¯k=Cx¯k+Du¯k,tkt+L1,y¯k𝕐,u¯k𝕌,tkt+L1,\begin{array}[]{ll}\underset{\begin{subarray}{c}\widebar{x}_{[t-N,t+L-1]}\\ \widebar{u}_{[t,t+L-1]}\\ \widebar{y}_{[t,t+L-1]}\end{subarray}}{\mbox{minimize}}&\sum_{k=t}^{t+L-1}\big{(}\left\lVert\widebar{y}_{k}-r_{k}\right\rVert^{2}_{Q}+\left\lVert\widebar{u}_{k}\right\rVert^{2}_{R}\big{)}\\ \mbox{subject to}&\widebar{x}_{k+1}=A\widebar{x}_{k}+Bu_{k},\enskip y_{k}=C\widebar{x}_{k}+Du_{k},\\ &t-N\leq k\leq t-1,\\ &\widebar{x}_{k+1}=A\widebar{x}_{k}+B\widebar{u}_{k},\enskip\widebar{y}_{k}=C\widebar{x}_{k}+D\widebar{u}_{k},\\ &t\leq k\leq t+L-1,\\ &\widebar{y}_{k}\in\mathbb{Y},\enskip\widebar{u}_{k}\in\mathbb{U},\enskip t\leq k\leq t+L-1,\end{array} (11)

where NtN\leq t. The sets 𝕌m\mathbb{U}\subset\mathbb{R}^{m} and 𝕐p\mathbb{Y}\subset\mathbb{R}^{p} describe feasible inputs and outputs, respectively. Positive semi-definite matrices Qp×pQ\in\mathbb{R}^{p\times p} and Rm×mR\in\mathbb{R}^{m\times m}, together with reference output trajectory r[t,t+L1]pLr_{[t,t+L-1]}\in\mathbb{R}^{pL}, define the output tracking cost function. The idea is to find a length-(N+L)(N+L) input-output trajectory that agrees with the past length-NN measurements history (u[tN,t1],y[tN,t1])(u_{[t-N,t-1]},y_{[t-N,t-1]}) and optimizes the future length-LL trajectory (u¯[t,t+L1],y¯[t,t+L1])(\widebar{u}_{[t,t+L-1]},\widebar{y}_{[t,t+L-1]}). If system (1) is observable and NnN\geq n, the initial state of such a trajectory is uniquely determined [7, Lem. 1].

Recently, [12, 13] proposed a data-enabled predictive control (DeePC) that replaces optimization (11) with

minimizeg,u¯[t,t+L1]y¯[t,t+L1]k=tt+L1(y¯krkQ2+u¯kR2)subject to[u[tN,t1]u¯[t,t+L1]y[tN,t1]y¯[t,t+L1]]=[HN+L(u[0,T1])HN+L(y[0,T1])]g,u¯k𝕌,y¯k𝕐,tkt+L1,\begin{array}[]{ll}\underset{\begin{subarray}{c}g,\widebar{u}_{[t,t+L-1]}\\ \widebar{y}_{[t,t+L-1]}\end{subarray}}{\mbox{minimize}}&\sum_{k=t}^{t+L-1}\big{(}\left\lVert\widebar{y}_{k}-r_{k}\right\rVert^{2}_{Q}+\left\lVert\widebar{u}_{k}\right\rVert^{2}_{R}\big{)}\\ \mbox{subject to}&\begin{bmatrix}u_{[t-N,t-1]}\\ \widebar{u}_{[t,t+L-1]}\\ y_{[t-N,t-1]}\\ \widebar{y}_{[t,t+L-1]}\end{bmatrix}=\begin{bmatrix}H_{N+L}(u_{[0,T-1]})\\ H_{N+L}(y_{[0,T-1]})\end{bmatrix}g,\\ &\widebar{u}_{k}\in\mathbb{U},\enskip\widebar{y}_{k}\in\mathbb{Y},\enskip t\leq k\leq t+L-1,\end{array} (12)

where TtT\leq t. If system (1a) is controllable, Willems’ fundamental lemma guarantees that optimization (12) is equivalent to the one in (11), assuming u[0,T1]u_{[0,T-1]} is sufficiently persistently exciting.

Unfortunately, such guarantee does not exist if the system (1) is uncontrollable, and is non-trivial to verify even if it is controllable. Particularly, verifying the controllability of the unknown system (1) either requires (A,B)(A,B) having special zero patterns [23], or computation using (u[0,T1],y[0,T1])(u_{[0,T-1]},y_{[0,T-1]}) that scale cubically with TT [18].

Fortunately, Corollary 1 guarantees the equivalence between (11) and (12) by only assuming sufficient persistency of excitation on u[0,T1]u_{[0,T-1]}, regardless of the controllability of the system (1). First, if (u¯[t,t+L1],y¯[t,t+L1])(\widebar{u}_{[t,t+L-1]},\widebar{y}_{[t,t+L-1]}) satisfies the constraints in (11), then

([u[tN,t1]u¯[t,t+L1]],[y[tN,t1]y¯[t,t+L1]])\left(\begin{bmatrix}u_{[t-N,t-1]}\\ \widebar{u}_{[t,t+L-1]}\end{bmatrix},\begin{bmatrix}y_{[t-N,t-1]}\\ \widebar{y}_{[t,t+L-1]}\end{bmatrix}\right)

is a segment of an input-output trajectory generated by the system (1) that starts with (u[0,T1],y[0,T1])(u_{[0,T-1]},y_{[0,T-1]}). Hence Corollary 1 implies that (u¯[t,t+L1],y¯[t,t+L1])(\widebar{u}_{[t,t+L-1]},\widebar{y}_{[t,t+L-1]}) satisfies the constraints in (12). Second, Corollary 1 also implies that any (u¯[t,t+L1],y¯[t,t+L1])(\widebar{u}_{[t,t+L-1]},\widebar{y}_{[t,t+L-1]}) satisfying the constraints in (12) also satisfies those in (11). Therefore, the constraints of the optimizations in (11) and (12) are equivalent, and so are the optimizations themselves. Note that if we replace (u[0,T1],y[0,T1])(u_{[0,T-1]},y_{[0,T-1]}) in (12) with an arbitrary input-output trajectory generated by the system (1) offline, then the above equivalence may not hold. Hence we term optimization (12) the online DeePC problem.

To illustrate our results, we consider the system (1) where

A=[10.5000100000.90.50000.9],B=[0.1250.500],C=[10000100],D=[00].A=\begin{bmatrix}1&0.5&0&0\\ 0&1&0&0\\ 0&0&0.9&0.5\\ 0&0&0&0.9\end{bmatrix},B=\begin{bmatrix}0.125\\ 0.5\\ 0\\ 0\end{bmatrix},C=\begin{bmatrix}1&0\\ 0&0\\ 0&1\\ 0&0\end{bmatrix}^{\top},D=\begin{bmatrix}0\\ 0\end{bmatrix}. (13)

One can verify that the above (A,B)(A,B) is uncontrollable. We let N=4,L=5,T=25,K=80N=4,L=5,T=25,K=80 and generate online input-output trajectories (u[0,T1],y[0,T1])(u_{[0,T-1]},y_{[0,T-1]}), where utu_{t} is sampled uniformly from [0.04,0.04][-0.04,0.04] for all t=0,,T1t=0,\ldots,T-1 such that u[0,T1]u_{[0,T-1]} is persistently exciting of order n+L+Nn+L+N. At time t=T,T+1,,Kt=T,T+1,\ldots,K, given (u[0,t1],y[0,t1])(u_{[0,t-1]},y_{[0,t-1]}), we obtain the input u¯t\widebar{u}_{t} by solving (12), where Q=I2Q=I_{2}, R=0.5R=0.5, 𝕐=2\mathbb{Y}=\mathbb{R}^{2}, 𝕌=[1,1]\mathbb{U}=[-1,1], and rt=[30.1]r_{t}=\begin{bmatrix}-3&0.1\end{bmatrix}^{\top} for all tt\in\mathbb{N}. Fig. 1 shows that, although the system contains an uncontrollable output (right), online DeePC still ensures that the controllable output (left) tracks the reference value as desired.

Refer to caption
Figure 1: Output trajectory for Online DeePC. The blue and red curves denote y[0,T1]y_{[0,T-1]} and y[T,K1]y_{[T,K-1]}, respectively. The dashed lines denote the output reference values.

III-B Identification of homogeneous multi-agent systems

An important paradigm in formation control is the homogeneous multi-agent system, i.e., a network of NN agents with the same LTI dynamics [24, 25]. In such a system, each agent ii can measure the state of agent jj in a local coordinate system if (i,j)(i,j) is an edge of a directed graph 𝒢\mathcal{G}, which is composed of NN nodes and MM edges. The dynamics of this multi-agent system is given by (1) with

A=INA¯,B=INB¯,C=EIn¯,D=0Mn¯×Nm¯,A=I_{N}\otimes\widebar{A},B=I_{N}\otimes\widebar{B},C=E\otimes I_{\widebar{n}},D=0_{M\widebar{n}\times N\widebar{m}}, (14)

where A¯n¯×n¯\widebar{A}\in\mathbb{R}^{\widebar{n}\times\widebar{n}} and B¯n¯×m¯\widebar{B}\in\mathbb{R}^{\widebar{n}\times\widebar{m}} describe the input-state dynamics of an individual agent. Each row of matrix EM×NE\in\mathbb{R}^{M\times N} is indexed by a directed edge, i.e., an edge with a head and a tail, in graph 𝒢\mathcal{G}: the ii-th entry in each row is “11” if node ii is the head of the corresponding edge, “1-1” if it is the tail, and “0” otherwise. We assume that B¯\widebar{B} is a non-zero matrix and (A¯,B¯)(\widebar{A},\widebar{B}) is controllable.

If at least one non-zero entry in matrix EE is known, then system matrices in (14) can be computed using the following Markov parameters [26, Sec. 3.4.4]

Mk=\displaystyle M_{k}= CAk1B+D\displaystyle CA^{k-1}B+D (15)
=\displaystyle= (EIn¯)(INA¯)k1(INB¯)\displaystyle(E\otimes I_{\widebar{n}})(I_{N}\otimes\widebar{A})^{k-1}(I_{N}\otimes\widebar{B})
=\displaystyle= E(A¯k1B¯),k=1,2,,n¯+1.\displaystyle E\otimes(\widebar{A}^{k-1}\widebar{B}),\enskip\forall k=1,2,\ldots,\widebar{n}+1.

In particular, let (Mk)ij(M_{k})_{ij} denote the ijij-th n¯×m¯\widebar{n}\times\widebar{m} block of MkM_{k}. If we know Eij=1E_{ij}=1 (the case of ``1"``-1" is similar), then (15) implies (Mk)ij=A¯k1B¯(M_{k})_{ij}=\widebar{A}^{k-1}\widebar{B}. For example, if E=[11]E=\begin{bmatrix}1&-1\end{bmatrix}, then (15) says Mk=[A¯k1B¯A¯k1B¯]M_{k}=\begin{bmatrix}\widebar{A}^{k-1}\widebar{B}&-\widebar{A}^{k-1}\widebar{B}\end{bmatrix}. Hence given the Markov parameters (15) and that Eij=1E_{ij}=1, we know EklE_{kl} is “11” if (M1)kl=(M1)ij(M_{1})_{kl}=(M_{1})_{ij}, “1-1” if (M1)kl=(M1)ij(M_{1})_{kl}=-(M_{1})_{ij}, and “0” otherwise. Further, B¯=(M1)ij\widebar{B}=(M_{1})_{ij} and A¯\widebar{A} is the unique solution to the following linear equations111Since (A¯,B¯)(\widebar{A},\widebar{B}) is controllable, matrix [(M1)ij(Mn¯1)ij]=[B¯A¯n¯B¯]\begin{bmatrix}(M_{1})_{ij}&\cdots&(M_{\widebar{n}-1})_{ij}\end{bmatrix}=\begin{bmatrix}\widebar{B}&\cdots&\widebar{A}^{\widebar{n}}\widebar{B}\end{bmatrix} has full column rank and (16) has a unique solution.

A¯[(M1)ij(Mn¯)ij]=[(M2)ij(Mn¯+1)ij].\displaystyle\widebar{A}\begin{bmatrix}(M_{1})_{ij}&\cdots&(M_{\widebar{n}})_{ij}\end{bmatrix}=\begin{bmatrix}(M_{2})_{ij}&\cdots&(M_{\widebar{n}+1})_{ij}\end{bmatrix}. (16)

Therefore, given at least one non-zero entry in matrix EE, in order to compute the system matrices in (14), it suffices to know the Markov parameters (15). These parameters can be computed using Corollary 2 via a data-driven simulation procedure [3, 7] (see Appendix for details).

In numerical simulations, we consider the homogeneous multi-agent system used in [24, Example 3], discretized with a sampling time of 0.1s such that the system dynamics is given by (14) where

A¯=[0.99640.00260.00040.04600.00450.90370.01880.38340.00980.03390.93830.13020.00050.00170.09681.0067],B¯=[0.04450.01670.34070.72490.52780.42140.02680.0215].\widebar{A}=\begin{bmatrix}0.9964&0.0026&-0.0004&-0.0460\\ 0.0045&0.9037&-0.0188&-0.3834\\ 0.0098&0.0339&0.9383&0.1302\\ 0.0005&0.0017&0.0968&1.0067\end{bmatrix},\widebar{B}=\begin{bmatrix}0.0445&0.0167\\ 0.3407&-0.7249\\ -0.5278&0.4214\\ -0.0268&0.0215\end{bmatrix}.

In addition, we let E=[𝟏N1IN1]E=\begin{bmatrix}\mathbf{1}_{N-1}&-I_{N-1}\end{bmatrix}, where 𝟏N1N1\mathbf{1}_{N-1}\in\mathbb{R}^{N-1} is the vector of all 11’s.

We compare Corollary 2 against the results in [2, Thm. 2] in terms of the least amount of input-output data needed to compute matrices in (14). In particular, since matrix A¯\widebar{A} has spectrum radius 1.031.03, we use use input-output trajectories {u[0,T1]i,y[0,T1]i}i=1τ\{u^{i}_{[0,T-1]},y^{i}_{[0,T-1]}\}_{i=1}^{\tau} with relatively short length T=120T=120 to avoid numerical instability [2]. The entries in {u[0,T1]i}i=1τ\{u_{[0,T-1]}^{i}\}_{i=1}^{\tau} are sampled uniformly from [0.1,0.1][-0.1,0.1]. Using Corollary 2, the data-driven simulation procedure requires {u[0,T1]i}i=1τ\{u^{i}_{[0,T-1]}\}_{i=1}^{\tau} to be collectively persistently exciting of order (N+1)n¯+1(N+1)\widebar{n}+1. In other words, matrix (8) with d=(N+1)n¯+1d=(N+1)\widebar{n}+1 has full row rank, hence it must have at least as many columns as rows, i.e.,

τ((N+1)n¯+1)Nm¯T(N+1)n¯=8N2+10N1164N.\textstyle\tau\geq\frac{((N+1)\widebar{n}+1)N\widebar{m}}{T-(N+1)\widebar{n}}=\frac{8N^{2}+10N}{116-4N}.

In comparison, if we use [2, Thm. 2] instead of Corollary 2, we need {u[0,T1]i}i=1τ\{u^{i}_{[0,T-1]}\}_{i=1}^{\tau} to be collectively persistently exciting of order 2Nn+12Nn+1 (see [2, Sec. IV-A]). In other words, matrix (8) with d=2Nn+1d=2Nn+1 has full row rank, which implies

τ(2Nn¯+1)Nm¯T2Nn¯=16N2+2N1208N.\textstyle\tau\geq\frac{(2N\widebar{n}+1)N\widebar{m}}{T-2N\widebar{n}}=\frac{16N^{2}+2N}{120-8N}.

In Fig. 2, we show the minimum number of input-output trajectories required to compute matrices in (14) in numerical simulations. The results tightly match the aforementioned two lower bounds, and the number of trajectories required by Corollary 2 is one order of magnitude less than that of [2, Thm. 1] when N=14N=14.

Refer to caption
Figure 2: Minimum number of length T=120T=120 input-output trajectories required to identify system (14).

IV Conclusions

We introduced an extension of Willems’ fundamental lemma by relaxing the controllability and persistency of excitation assumptions. We demonstrate the usefulness of our results in the context of DeePC and identification of homogeneous multi-agent system. Future directions include generalizations to noisy data and nonlinear systems.

Appendix

Proof of Theorem 1

We start by proving the first statement using a double inclusion argument. Using an argument similar to (5), one can directly show that the left hand side of (9) is included in its right hand side. To show the other direction, we show that the left kernel of matrix

[H1(x[0,T1L]1)H1(x[0,TτL]τ)HL(u[0,T11]1)HL(u[0,TτL]τ)]\begin{bmatrix}H_{1}(x^{1}_{[0,T^{1}-L]})&\cdots&H_{1}(x^{\tau}_{[0,T^{\tau}-L]})\\ H_{L}(u^{1}_{[0,T^{1}-1]})&\cdots&H_{L}(u^{\tau}_{[0,T^{\tau}-L]})\end{bmatrix} (17)

is orthogonal to (+𝒦[x01,x02,,x0τ])×mL(\mathcal{R}+\mathcal{K}[x_{0}^{1},x_{0}^{2},\ldots,x_{0}^{\tau}])\times\mathbb{R}^{mL}. To this end, let

v=[ξη1η2ηL],v^{\top}=\begin{bmatrix}\xi^{\top}&\eta_{1}^{\top}&\eta_{2}^{\top}&\cdots\eta_{L}^{\top}\end{bmatrix}, (18)

be an arbitrary row vector in the left kernel of matrix (17), where ξn,η1,η2,,ηLm\xi\in\mathbb{R}^{n},\eta_{1},\eta_{2},\ldots,\eta_{L}\in\mathbb{R}^{m}. Since δδmin\delta\geq\delta_{\min}, using [20, Def. 3.3.2], we know there exists α0k,α1k,,αδ1,k\alpha_{0k},\alpha_{1k},\ldots,\alpha_{\delta-1,k}\in\mathbb{R} such that

Ak+j=0δ1αjkAj=0n×n,k=δ,δ+1,A^{k}+\textstyle\sum_{j=0}^{\delta-1}\alpha_{jk}A^{j}=0_{n\times n},\enskip\forall k=\delta,\delta+1,\ldots (19)

The above equation implies that AkB=j=0δ1αjkAjBA^{k}B=\textstyle-\sum_{j=0}^{\delta-1}\alpha_{jk}A^{j}B and Akx0i=j=0δ1αjkAjx0iA^{k}x_{0}^{i}=\textstyle-\sum_{j=0}^{\delta-1}\alpha_{jk}A^{j}x_{0}^{i} for all k=δ,,n1k=\delta,\ldots,n-1 and i=1,,τ.i=1,\ldots,\tau. Therefore in order to show the left kernel of matrix (17) is orthogonal to (+𝒦[x01,x02,,x0τ])×mL(\mathcal{R}+\mathcal{K}[x_{0}^{1},x_{0}^{2},\ldots,x_{0}^{\tau}])\times\mathbb{R}^{mL}, it suffices to show the following

η1=\displaystyle\eta_{1}^{\top}= η2==ηL=0m,\displaystyle\eta_{2}^{\top}=\cdots=\eta_{L}^{\top}=0_{m}^{\top}, (20a)
ξB=\displaystyle\xi^{\top}B= ξAB==ξAδ1B=0m,\displaystyle\xi^{\top}AB=\cdots=\xi^{\top}A^{\delta-1}B=0_{m}^{\top}, (20b)
ξx0i=\displaystyle\xi^{\top}x^{i}_{0}= ξAx0i==ξAδ1x0i=0.i=1,,τ.\displaystyle\xi^{\top}Ax^{i}_{0}=\cdots=\xi^{\top}A^{\delta-1}x^{i}_{0}=0.\enskip\forall i=1,\ldots,\tau. (20c)

In order to prove (20a), we let w0,w1,,wδn+m(δ+L)w_{0},w_{1},\ldots,w_{\delta}\in\mathbb{R}^{n+m(\delta+L)} be such that w0=[v0mδ]w_{0}=\begin{bmatrix}v^{\top}&0_{m\delta}^{\top}\end{bmatrix}^{\top} and wjw_{j} equals

[ξAjξAj1BξBη1ηL0m(δj)],\displaystyle\begin{bmatrix}\xi^{\top}A^{j}&\xi^{\top}A^{j-1}B&\cdots\xi^{\top}B&\eta_{1}^{\top}&\cdots&\eta_{L}^{\top}&0_{m(\delta-j)}^{\top}\end{bmatrix}^{\top},

for j=1,,δj=1,\ldots,\delta. Since vv^{\top} is in the left kernel of matrix (17), using (1) one can verify that w0,w1,,wδw_{0}^{\top},w_{1}^{\top},\ldots,w_{\delta}^{\top} are in the left kernel of

[H1(x[0,T1δL]1)H1(x[0,TτδL]τ)Hδ+L(u[0,T11]1)Hδ+L(u[0,Tτ1]τ)].\begin{bmatrix}H_{1}(x^{1}_{[0,T^{1}-\delta-L]})&\cdots&H_{1}(x^{\tau}_{[0,T^{\tau}-\delta-L]})\\ H_{\delta+L}(u^{1}_{[0,T^{1}-1]})&\cdots&H_{\delta+L}(u^{\tau}_{[0,T^{\tau}-1]})\end{bmatrix}. (21)

Let αδδ=1\alpha_{\delta\delta}=1 and k=δk=\delta in (19), we have 0n×n=j=0δαjδAj0_{n\times n}=\sum_{j=0}^{\delta}\alpha_{j\delta}A^{j}. Hence

j=0δαjδwj=[j=0δαjδξAjr]=[0nr],\textstyle\sum_{j=0}^{\delta}\alpha_{j\delta}w_{j}^{\top}=\begin{bmatrix}\sum_{j=0}^{\delta}\alpha_{j\delta}\xi^{\top}A^{j}&r^{\top}\end{bmatrix}=\begin{bmatrix}0_{n}&r^{\top}\end{bmatrix}, (22)

for some vector rm(δ+L)r\in\mathbb{R}^{m(\delta+L)}. Since row vectors w0,w1,,wδw_{0}^{\top},w_{1}^{\top},\ldots,w_{\delta}^{\top} are in the left kernel of matrix (21), equation (22) implies that rr^{\top} is in the left kernel of matrix

[Hδ+L(u[0,Ti1]1)Hδ+L(u[0,Ti1]τ)].\begin{bmatrix}H_{\delta+L}(u^{1}_{[0,T^{i}-1]})&\cdots&H_{\delta+L}(u^{\tau}_{[0,T^{i}-1]})\end{bmatrix}. (23)

Since {u[0,Ti1]i}i=1τ\{u_{[0,T^{i}-1]}^{i}\}_{i=1}^{\tau} is collectively persistently exciting of order δ+L\delta+L, matrix (23) has full row rank. Therefore

r=0m(δ+L).r=0_{m(\delta+L)}. (24)

Observe that the last mm entries of rr are given by αδδηL=ηL\alpha_{\delta\delta}\eta_{L}=\eta_{L}, hence equation (24) implies that ηL=0m\eta_{L}=0_{m}. Then the last 2m2m entries of rr are given by [αδδηL1+α(δ1)δηLαδδηL]\begin{bmatrix}\alpha_{\delta\delta}\eta_{L-1}^{\top}+\alpha_{(\delta-1)\delta}\eta_{L}^{\top}&\alpha_{\delta\delta}\eta_{L}^{\top}\end{bmatrix}^{\top}. Since ηL=0m\eta_{L}=0_{m} and αδδ=1\alpha_{\delta\delta}=1, equation (24) also implies that ηL1=0m\eta_{L-1}=0_{m}. By repeating similar induction we can prove that (20a) holds.

Next, since (20a) holds, the first mδm\delta entries in rr are

[j=1δαjδξAj1Bj=2δαjδξAj2BαδδξB].\displaystyle\begin{bmatrix}\sum\limits_{j=1}^{\delta}\alpha_{j\delta}\xi^{\top}A^{j-1}B&\sum\limits_{j=2}^{\delta}\alpha_{j\delta}\xi^{\top}A^{j-2}B&\cdots&\alpha_{\delta\delta}\xi^{\top}B\end{bmatrix}^{\top}.

By combining this with (24) we get that

0m=j=kδαjδξAjkB,k=1,,δ.\textstyle 0_{m}^{\top}=\sum_{j=k}^{\delta}\alpha_{j\delta}\xi^{\top}A^{j-k}B,\enskip\forall k=1,\ldots,\delta. (25)

Since αδδ=1\alpha_{\delta\delta}=1, considering k=δk=\delta in (25) implies that ξB=0m\xi^{\top}B=0_{m}. Substitute this back into (25) and considering k=δ1k=\delta-1 implies that ξAB=0m\xi^{\top}AB=0_{m}. By repeating a similar reasoning for k=δ2,δ3,,1k=\delta-2,\delta-3,\ldots,1 we can prove that (20b) holds.

Further, by using (19) and (20b) we can show that ξAkB=0m\xi^{\top}A^{k}B=0_{m} for all k1k\geq 1. Combining this together with the fact that row vector (18) is in the left kernel of matrix (17) and that (20a) also holds, it follows that,

0=\displaystyle 0= ξxki=ξ(Akx0i+j=0k1Akj1Buji)=ξAkx0i,\displaystyle\xi^{\top}x_{k}^{i}=\textstyle\xi^{\top}\big{(}A^{k}x^{i}_{0}+\sum_{j=0}^{k-1}A^{k-j-1}Bu^{i}_{j}\big{)}=\xi^{\top}A^{k}x^{i}_{0},
k=0,,TiL,i=1,,τ.\displaystyle\forall k=0,\ldots,T^{i}-L,\enskip i=1,\ldots,\tau.

Since Tiδ+LT^{i}\geq\delta+L for i=1,,τi=1,\ldots,\tau by assumption, we conclude that (20c) holds.

We now prove the second statement. Given the input, state, and output trajectories in (3), suppose (7) holds. Let

x¯0=\displaystyle\widebar{x}_{0}= [H1(x[0,T1L]1)H1(x[0,TτL]τ)]g,\displaystyle\begin{bmatrix}H_{1}(x^{1}_{[0,T^{1}-L]})&\cdots&H_{1}(x^{\tau}_{[0,T^{\tau}-L]})\end{bmatrix}g, (26)
x¯t+1=\displaystyle\widebar{x}_{t+1}= Ax¯t+Bu¯t,0tL2.\displaystyle A\widebar{x}_{t}+B\widebar{u}_{t},\enskip 0\leq t\leq L-2.

Then from (9) we know x¯0\widebar{x}_{0} satisfies (10), and one can verify that (u¯[0,L1],x¯[0,L1],y¯[0,L1])(\widebar{u}_{[0,L-1]},\widebar{x}_{[0,L-1]},\widebar{y}_{[0,L-1]}) is indeed an input-state-output trajectory of system (1).

Conversely, let (u¯[0,L1],x¯[0,L1],y¯[0,L1])(\widebar{u}_{[0,L-1]},\widebar{x}_{[0,L-1]},\widebar{y}_{[0,L-1]}) be an input-state-output trajectory of system (1) with x¯0+𝒪+𝒦[x01,x02,,x0τ]\widebar{x}_{0}\in\mathcal{R}+\mathcal{O}+\mathcal{K}[x_{0}^{1},x_{0}^{2},\ldots,x_{0}^{\tau}]. Then there exists

x¯0a+𝒦[x01,x02,,x0τ],x¯0b𝒪,\widebar{x}_{0}^{a}\in\mathcal{R}+\mathcal{K}[x_{0}^{1},x_{0}^{2},\ldots,x_{0}^{\tau}],\enskip\widebar{x}_{0}^{b}\in\mathcal{O}, (27)

such that x¯0=x¯0a+x¯0b\widebar{x}_{0}=\widebar{x}_{0}^{a}+\widebar{x}_{0}^{b} and

[u¯[0,L1]y¯[0,L1]]=[0IOLTL][x¯0a+x¯0bu¯[0,L1]]\begin{bmatrix}\widebar{u}_{[0,L-1]}\\ \widebar{y}_{[0,L-1]}\end{bmatrix}=\begin{bmatrix}0&I\\ O_{L}&T_{L}\end{bmatrix}\begin{bmatrix}\widebar{x}_{0}^{a}+\widebar{x}_{0}^{b}\\ \widebar{u}_{[0,L-1]}\end{bmatrix} (28)

where

TL=\displaystyle T_{L}= [D000CBD00CABCBD0CAL2BCAL3BCAL4BD],\displaystyle\begin{bmatrix}D&0&0&\cdots&0\\ CB&D&0&\cdots&0\\ CAB&CB&D&\cdots&0\\ \vdots&\vdots&\vdots&\ddots&\vdots\\ CA^{L-2}B&CA^{L-3}B&CA^{L-4}B&\cdots&D\end{bmatrix}, (29)
OL=\displaystyle O_{L}= [C(CA)(CA2)(CAL1)].\displaystyle\begin{bmatrix}C^{\top}&(CA)^{\top}&(CA^{2})^{\top}&\cdots&(CA^{L-1})^{\top}\end{bmatrix}^{\top}.

Further, using the Cayley-Hamilton theorem one can show that 𝒪kerOL\mathcal{O}\subset\ker O_{L} for any L+L\in\mathbb{N}_{+}. Hence (27) implies

OLx0b=0Lp.O_{L}x_{0}^{b}=0_{Lp}. (30)

Since x¯0a+𝒦[x01,x02,,x0τ]\widebar{x}_{0}^{a}\in\mathcal{R}+\mathcal{K}[x_{0}^{1},x_{0}^{2},\ldots,x_{0}^{\tau}], the first statement of Theorem 1 implies that there exists gi=1τ(TiL+1)g\in\mathbb{R}^{\sum_{i=1}^{\tau}(T^{i}-L+1)} such that

[x¯0au¯[0,L1]]=[H1(x[0,T1L]1)H1(x[0,TτL]τ)HL(u[0,T11]1)HL(u[0,Tτ1]τ)]g.\begin{bmatrix}\widebar{x}_{0}^{a}\\ \widebar{u}_{[0,L-1]}\end{bmatrix}=\begin{bmatrix}H_{1}(x^{1}_{[0,T^{1}-L]})&\cdots&H_{1}(x^{\tau}_{[0,T^{\tau}-L]})\\ H_{L}(u^{1}_{[0,T^{1}-1]})&\cdots&H_{L}(u^{\tau}_{[0,T^{\tau}-1]})\end{bmatrix}g. (31)

Notice that

[0IOLTL][H1(x[0,TiL]i)HL(u[0,Ti1]i)]=[HL(u[0,Ti1]i)HL(y[0,Ti1]i)].\displaystyle\begin{bmatrix}0&I\\ O_{L}&T_{L}\end{bmatrix}\begin{bmatrix}H_{1}(x^{i}_{[0,T^{i}-L]})\\ H_{L}(u^{i}_{[0,T^{i}-1]})\end{bmatrix}=\begin{bmatrix}H_{L}(u^{i}_{[0,T^{i}-1]})\\ H_{L}(y^{i}_{[0,T^{i}-1]})\end{bmatrix}. (32)

for i=1,,τi=1,\ldots,\tau. Substituting (30), (31) and (32) into (28) gives (7), thus completing the proof. ∎

Proof of Corollary 1

First, let x[0,K1]x_{[0,K-1]} be such that (u[0,K1],x[0,K1],y[0,K1])(u_{[0,K-1]},x_{[0,K-1]},y_{[0,K-1]}) is an input-state-output trajectory generated by the system (1). Then (5) holds for all 0tK10\leq t\leq K-1. Hence (u[t,t+L1],x[t,t+L1],y[t,t+L1])(u_{[t,t+L-1]},x_{[t,t+L-1]},y_{[t,t+L-1]}) is an input-state-output trajectory generated by system (1) that satisfies (5). From the second statement in Theorem 1 we know that (u[t,t+L1],y[t,t+L1])(u_{[t,t+L-1]},y_{[t,t+L-1]}) is parameterizable by (u[0,T1],y[0,T1])(u_{[0,T-1]},y_{[0,T-1]}). Second, if (u¯[0,L1],y¯[0,L1])(\widebar{u}_{[0,L-1]},\widebar{y}_{[0,L-1]}) is parameterizable by (u[0,T1],y[0,T1])(u_{[0,T-1]},y_{[0,T-1]}), then one can verify that it is indeed an input-output trajectory generated by the system (1) by constructing a length-LL state trajectory similar to the one in (26). ∎

Markov parameters computation in Section III-B

Let {u[0,Ti1]i,y[0,Ti1]i}i=1τ\{u^{i}_{[0,T^{i}-1]},y^{i}_{[0,T^{i}-1]}\}_{i=1}^{\tau} be input-output trajectories of the system described by (14), such that inputs {u[0,Ti1]i}i=1τ\{u^{i}_{[0,T^{i}-1]}\}_{i=1}^{\tau} are collectively persistently exciting of order (N+1)n¯+1(N+1)\widebar{n}+1. Let n=Nn¯,m=Nm¯,p=Mn¯n=N\widebar{n},m=N\widebar{m},p=M\widebar{n}. Since (A¯,B¯)(\widebar{A},\widebar{B}) in (14) is controllable, one can verify that +𝒦[x01,x02,,x0τ]=n\mathcal{R}+\mathcal{K}[x_{0}^{1},x_{0}^{2},\ldots,x_{0}^{\tau}]=\mathbb{R}^{n}, regardless of the values of x01,x02,,x0τx^{1}_{0},x_{0}^{2},\ldots,x^{\tau}_{0}. Using Corollary 2 we know that there exists a matrix Gk(i=1τ(Tin))×mG_{k}\in\mathbb{R}^{(\sum_{i=1}^{\tau}(T^{i}-n))\times m} such that,

[Hn+1(u[0,T11]1)Hn+1(u[0,Tτ1]τ)Hn+1(y[0,T11]1)Hn+1(y[0,Tτ1]τ)]Gk\displaystyle\begin{bmatrix}H_{n+1}(u^{1}_{[0,T^{1}-1]})&\cdots&H_{n+1}(u^{\tau}_{[0,T^{\tau}-1]})\\ H_{n+1}(y^{1}_{[0,T^{1}-1]})&\cdots&H_{n+1}(y^{\tau}_{[0,T^{\tau}-1]})\end{bmatrix}G_{k} (33)
=\displaystyle= [0m(nk)×mIm0(pn+kmkp)×mM0Mk]\displaystyle\begin{bmatrix}0^{\top}_{m(n-k)\times m}&I_{m}&0^{\top}_{(pn+km-kp)\times m}&M_{0}^{\top}&\cdots&M_{k}^{\top}\end{bmatrix}^{\top}

for all k=0,1,,n¯+1k=0,1,\ldots,\widebar{n}+1, where M0=0p×mM_{0}=0_{p\times m} and MkM_{k} is given by (15); also see [7, Sec. 4.5] for details. Next, given M0,,Mk1M_{0},\ldots,M_{k-1}, we can compute MkM_{k} by first solving the first m(n+1)+pnm(n+1)+pn equations in (33) for matrix GkG_{k}, then substituting the solution into the last pp equations in (33). By repeating this process for k=1,,n¯+1k=1,\ldots,\widebar{n}+1 we obtain Markov parameters (15). Using Kalman decomposition one can verify that Markov parameters of system (14) are the same as those of a reduced order controllable and observable system with state dimension less than nn. Hence matrix MkM_{k} obtained this way is unique; see [7, Prop. 1].

References

  • [1] J. C. Willems, P. Rapisarda, I. Markovsky, and B. L. De Moor, “A note on persistency of excitation,” Syst. Control Lett., vol. 54, no. 4, pp. 325–329, 2005.
  • [2] H. J. van Waarde, C. De Persis, M. K. Camlibel, and P. Tesi, “Willems’ fundamental lemma for state-space systems and its extension to multiple datasets,” IEEE Control Syst. Lett., vol. 4, no. 3, pp. 602–607, 2020.
  • [3] I. Markovsky, J. C. Willems, P. Rapisarda, and B. L. De Moor, “Algorithms for deterministic balanced subspace identification,” Automatica, vol. 41, no. 5, pp. 755–766, 2005.
  • [4] T. Katayama, Subspace methods for system identification.   Springer Science & Business Media, 2006.
  • [5] I. Markovsky, “From data to models,” in Low-Rank Approximation.   Springer, 2019, pp. 37–70.
  • [6] J. Berberich and F. Allgöwer, “A trajectory-based framework for data-driven system analysis and control,” in Eur. Control Conf.   IEEE, 2020, pp. 1365–1370.
  • [7] I. Markovsky and P. Rapisarda, “Data-driven simulation and control,” Int. J. Control, vol. 81, no. 12, pp. 1946–1959, 2008.
  • [8] T. Maupong and P. Rapisarda, “Data-driven control: A behavioral approach,” Syst. Control Lett., vol. 101, pp. 37–43, 2017.
  • [9] A. Bisoffi, C. De Persis, and P. Tesi, “Data-based guarantees of set invariance properties,” arXiv preprint arXiv:1911.12293[eecs.SY], 2019.
  • [10] C. De Persis and P. Tesi, “Formulas for data-driven control: Stabilization, optimality, and robustness,” IEEE Trans. Autom. Control, vol. 65, no. 3, pp. 909–924, 2019.
  • [11] L. Huang, J. Coulson, J. Lygeros, and F. Dörfler, “Data-enabled predictive control for grid-connected power converters,” in IEEE Conf. Decision Control.   IEEE, 2019, pp. 8130–8135.
  • [12] J. Coulson, J. Lygeros, and F. Dörfler, “Data-enabled predictive control: In the shallows of the deepc,” in Eur. Control Conf.   IEEE, 2019, pp. 307–312.
  • [13] D. Alpago, F. Dörfler, and J. Lygeros, “An extended Kalman filter for data-enabled predictive control,” IEEE Control Systems Letters, vol. 4, no. 4, pp. 994–999, 2020.
  • [14] J. Berberich, J. Koehler, M. A. Muller, and F. Allgower, “Data-driven model predictive control with stability and robustness guarantees,” IEEE Transactions on Automatic Control, pp. 1–1, 2020.
  • [15] A. Allibhoy and J. Cortés, “Data-based receding horizon control of linear network systems,” arXiv preprint arXiv:2003.09813, 2020.
  • [16] M. Yin, A. Iannelli, and R. S. Smith, “Maximum likelihood estimation in data-driven modeling and control,” arXiv preprint arXiv:2011.00925 [eess.SY], 2020.
  • [17] F. Fabiani and P. J. Goulart, “The optimal transport paradigm enables data compression in data-driven robust control,” arXiv preprint arXiv:2005.09393 [eess.SY], 2020.
  • [18] V. K. Mishra, I. Markovsky, and B. Grossmann, “Data-driven tests for controllability,” IEEE Control Syst. Lett., vol. 5, no. 2, pp. 517–522, 2020.
  • [19] I. Markovsky and F. Dörfler, “Identifiability in the behavioral setting,” Dept. ELEC, Vrije Universiteit Brussel, Tech. Rep., 2020.
  • [20] R. A. Horn and C. R. Johnson, Matrix analysis.   Cambridge university press, 2012.
  • [21] D. Q. Mayne, J. B. Rawlings, C. V. Rao, and P. O. Scokaert, “Constrained model predictive control: Stability and optimality,” Automatica, vol. 36, no. 6, pp. 789–814, 2000.
  • [22] D. Q. Mayne, “Model predictive control: Recent developments and future promise,” Automatica, vol. 50, no. 12, pp. 2967–2986, 2014.
  • [23] H. Mayeda and T. Yamada, “Strong structural controllability,” SIAM Journal on Control and Optimization, vol. 17, no. 1, pp. 123–138, 1979.
  • [24] G. Wen, Z. Duan, W. Ren, and G. Chen, “Distributed consensus of multi-agent systems with general linear node dynamics and intermittent communications,” Int. J. Robust. Nonlinear Control, vol. 24, no. 16, pp. 2438–2457, 2014.
  • [25] K.-K. Oh, M.-C. Park, and H.-S. Ahn, “A survey of multi-agent formation control,” Automatica, vol. 53, pp. 424–440, 2015.
  • [26] M. Verhaegen and V. Verdult, Filtering and system identification: a least squares approach.   Cambridge University Press, 2007.