This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Data-Driven Predictive Control With Adaptive Disturbance Attenuation For Constrained Systems

Nan Li [email protected]    Ilya Kolmanovsky [email protected]    Hong Chen [email protected] Department of Aerospace Engineering, Auburn University, Auburn, AL 36849, USA Department of Aerospace Engineering, University of Michigan, Ann Arbor, MI 48105, USA College of Electronic and Information Engineering, Tongji University, Shanghai, China
Abstract

In this paper, we propose a novel data-driven predictive control approach for systems subject to time-domain constraints. The approach combines the strengths of \mathcal{H}_{\infty} control for rejecting disturbances and MPC for handling constraints. In particular, the approach can dynamically adapt \mathcal{H}_{\infty} disturbance attenuation performance depending on measured system state and forecasted disturbance level to satisfy constraints. We establish theoretical properties of the approach including robust guarantees of closed-loop stability, disturbance attenuation, constraint satisfaction under noisy data, as well as sufficient conditions for recursive feasibility, and illustrate the approach with a numerical example.

keywords:
Data-Driven Control, \mathcal{H}_{\infty} Control, Model Predictive Control, Constraints, Linear Matrix Inequality

, ,

1 Introduction

In addition to stability, disturbance rejection and constraint satisfaction are important topics in control systems engineering, especially as modern systems often operate in complex and dynamic environments while increasingly stringent safety and performance requirements are imposed on them. The \mathcal{H}_{\infty} control is an effective approach to addressing the former – it aims to minimize the effect of disturbances on system outputs [1]. Since introduced in the early 1980s, \mathcal{H}_{\infty} control has been successfully implemented in aerospace, automotive, and many other sectors. However, time-domain constraints on state and control variables are not handled in conventional \mathcal{H}_{\infty} control. Meanwhile, Model Predictive Control (MPC) stands out from many other control approaches for its ability to explicitly and non-conservatively handle time-domain constraints [2]. Therefore, combing \mathcal{H}_{\infty} control and MPC to achieve desired disturbance rejection properties while satisfying constraints is appealing and has been studied in, e.g., [3, 4, 5, 6, 7, 8, 9, 10, 11, 12]. A major strength of such a combination is the ability to dynamically adapt \mathcal{H}_{\infty} disturbance attenuation performance (depending on measured/estimated system state and forecasted disturbance level) to satisfy constraints through solving an MPC optimization problem repeatedly in an online manner [3, 4].

Conventional \mathcal{H}_{\infty} control and MPC methods rely on parametric models of the system to be controlled. As engineered systems are becoming increasingly complex and cyber-physical, first-principle models are more difficult to obtain. Meanwhile, with the rapid advances in sensing, computation, and communication technologies, data is more readily available. This has spurred the development of data-driven methods for system modeling, analysis, and control. In particular, practitioners may favor an end-to-end solution that bypasses the intermediate steps of modeling and analysis and produces a controller with desired properties directly from measured data of system behavior. Therefore, it is beneficial to extend model-based methods for \mathcal{H}_{\infty} control, MPC, and the aforementioned combination to their data-driven counterparts.

Here we provide a brief review of existing data-driven control methods in the literature: Reinforcement Learning (RL) can be used to train optimal controllers from data [13, 14]. However, the majority of RL methods optimize the average behavior of the closed-loop system when it is subject to disturbances with certain statistics and do not provide a worst-case robustness guarantee. Furthermore, they typically handle constraints through penalties, hence making constraints soft. Integrating data-driven uncertainty estimation/system identification with(in) the MPC framework can lead to reliable usage of data for improving MPC performance while maintaining certain robustness guarantees especially for satisfying constraints [15, 16]. Along these lines, the Data-enabled Predictive Control (DeePC) is an emerging technique and has attracted increasing attention from researchers recently. The DeePC uses measured input-output data to create a non-parametric system model based on behavioral systems theory and uses the model to predict future trajectories [17]. It has demonstrated superior performance in various applications [18, 19, 20, 21]. However, DeePC entails higher computational cost than conventional MPC due to its high-dimensional, data-based non-parametric system model [22, 23]. Also, how to handle noisy data and provide certain robustness guarantees under noisy data remains to be an open question in DeePC [24, 25, 26]. Furthermore, to the best of our knowledge, there has been no previous work integrating \mathcal{H}_{\infty} control and MPC (or, designing an MPC that has an \mathcal{H}_{\infty}-type disturbance attenuation property) in the general data-driven MPC literature. An approach to synthesizing an \mathcal{H}_{\infty} controller using noisy data based on a matrix S-lemma was introduced in [27]. The approach reduces a data-driven \mathcal{H}_{\infty} control synthesis problem to a low-dimensional Linear Matrix Inequality (LMI) optimization problem, which is computationally tractable with state-of-the-art interior-point LMI solvers. However, time-domain constraints are not handled by the approach of [27], and moving-horizon, MPC-type implementation of the approach was not considered in [27]. More classical data-driven control methods also include self-tuning regulators [28] and iterative learning control [29]. A comprehensive survey of data-driven control methods can be found in [30].

In this paper, we fill the gap in the literature by proposing a novel data-driven control approach that combines the strengths of \mathcal{H}_{\infty} control for rejecting disturbances and MPC for handling constraints. Our approach can be viewed as a data-driven counterpart of the model-based moving-horizon \mathcal{H}_{\infty} control approach of [3, 4] and enjoys similar properties including dynamic adaptation of \mathcal{H}_{\infty} performance depending on measured system state and forecasted disturbance level for satisfying constraints. Specifically, the contributions include:

  1. 1.

    Our approach is the first data-driven MPC method in the literature that focuses on \mathcal{H}_{\infty}-type disturbance attenuation for systems with time-domain constraints.

  2. 2.

    We conduct a comprehensive analysis of the theoretical properties of our approach. The results include robust guarantees of closed-loop stability, disturbance attenuation, and constraint satisfaction under noisy data, as well as conditions for online problem recursive feasibility.

The paper is organized as follows: We describe the problem treated in this paper including key assumptions and preliminaries in Section 2. We develop an approach to synthesizing an \mathcal{H}_{\infty} controller that also enforces time-domain constraints for an unknown system using noisy trajectory data in Section 3. The development in this section has merit in its own right because it extends the data-driven \mathcal{H}_{\infty} control synthesis approach of [27] (which does not handle constraints) to the constrained case. It is also an essential building block of the data-driven MPC algorithm developed in Section 4. Then, Section 4 presents our proposed data-driven MPC algorithm and analyzes its theoretical properties. We illustrate the algorithm with a numerical example in Section 5. Finally, we conclude the paper in Section 6.

The notations used in this paper are mostly standard. We use n\mathbb{R}^{n} to denote the space of nn-dimensional real vectors, n×m\mathbb{R}^{n\times m} the space of nn-by-mm real matrices, and \mathbb{N} the set of natural numbers including zero. Given a vector xnx\in\mathbb{R}^{n}, we use x\|x\| to denote its Euclidean norm, i.e., x=xx\|x\|=\sqrt{x^{\top}x}. Given a matrix Mn×mM\in\mathbb{R}^{n\times m}, its kernel is the subspace of all xmx\in\mathbb{R}^{m} such that Mx=0Mx=0, i.e., ker(M)={xm:Mx=0}\text{ker}(M)=\{x\in\mathbb{R}^{m}:Mx=0\}. Given two symmetric matrices M,Nn×nM,N\in\mathbb{R}^{n\times n}, MNM\succ N means that MNM-N is positive definite, i.e., x(MN)x>0x^{\top}(M-N)x>0 for all non-zero xnx\in\mathbb{R}^{n}, and MNM\succeq N means that MNM-N is positive semidefinite, i.e., x(MN)x0x^{\top}(M-N)x\geq 0 for all xnx\in\mathbb{R}^{n}. Similarly, MNM\prec N and MNM\preceq N means that MNM-N is negative definite and negative semidefinite, respectively. For an optimization problem, by “solution” we mean a feasible solution, i.e., a set of values for the decision variables that satisfies all constraints, and by “optimal solution” we mean a feasible solution where the objective function (almost) reaches its maximum (or minimum) value. Because the optimization problems appearing in this paper are all convex problems, an “optimal solution” is globally optimal.

2 Problem Statement and Preliminaries

We consider the control of dynamic systems that can be represented by the following linear time-invariant model:

x(t+1)\displaystyle x(t+1) =Aox(t)+Bou(t)+w(t)\displaystyle=A_{\text{o}}x(t)+B_{\text{o}}u(t)+w(t) (1a)
y1(t)\displaystyle y_{1}(t) =C1x(t)+D1u(t)\displaystyle=C_{1}x(t)+D_{1}u(t) (1b)
y2(t)\displaystyle y_{2}(t) =C2x(t)+D2u(t)\displaystyle=C_{2}x(t)+D_{2}u(t) (1c)

where x(t)nx(t)\in\mathbb{R}^{n} denotes the system state at the discrete time tt\in\mathbb{N}, u(t)mu(t)\in\mathbb{R}^{m} denotes the control input, w(t)nw(t)\in\mathbb{R}^{n} denotes an unmeasured disturbance input, y1(t)p1y_{1}(t)\in\mathbb{R}^{p_{1}} and y2(t)p2y_{2}(t)\in\mathbb{R}^{p_{2}} are two outputs the roles of which are introduced below. The goal is to design a control algorithm that achieves the following three objectives:

  1. 1.

    Guaranteeing closed-loop stability;

  2. 2.

    Optimizing disturbance attenuation in terms of the \mathcal{H}_{\infty} performance from ww to output y1y_{1};

  3. 3.

    Enforcing the following constraints on output y2y_{2} at all times tt\in\mathbb{N}:

    y2v(t)y2v,max,v=1,2,,p2y_{2v}(t)\leq y_{2v,\max},\quad v=1,2,\dots,p_{2} (2)

    where y2v(t)y_{2v}(t) denotes the vvth entry of y2(t)y_{2}(t) and y2v,max0y_{2v,\max}\geq 0 for all v=1,2,,p2v=1,2,\dots,p_{2}. Constraints in this form can represent state/output variable bounds, control input limits, etc.

In addition to the standard assumptions of (Ao,Bo)(A_{\text{o}},B_{\text{o}}) being stabilizable and (C1,Ao)(C_{1},A_{\text{o}}) being detectable, we also assume that the system model (Ao,Bo)(A_{\text{o}},B_{\text{o}}) is unknown and only trajectory data are available. This calls for a data-driven control approach.

Assume we have data {(xj+,xj,uj)}j=1J\{(x^{+}_{j},x_{j},u_{j})\}_{j=1}^{J}, where (xj,uj)(x_{j},u_{j}) denotes a pair of previous state and control input values, xj+x^{+}_{j} denotes the corresponding next state value, and the subscript jj indicates the jjth data point. The data can be collected from a single or multiple trajectories. According to (1a), we have

xj+=Aoxj+Bouj+wjx^{+}_{j}=A_{\text{o}}x_{j}+B_{\text{o}}u_{j}+w_{j} (3)

where wjw_{j} is the disturbance input value at the time of collecting the data point (xj+,xj,uj)(x^{+}_{j},x_{j},u_{j}). Organizing data with the following matrices:

X+\displaystyle X^{+} =[x1+,x2+,,xJ+]\displaystyle=\left[x^{+}_{1},\,x^{+}_{2},\,\cdots,\,x^{+}_{J}\right] (4a)
X\displaystyle X =[x1,x2,,xJ]\displaystyle=\left[x_{1},\,x_{2},\,\cdots,\,x_{J}\right] (4b)
U\displaystyle U =[u1,u2,,uJ]\displaystyle=\left[u_{1},\,u_{2},\,\cdots,\,u_{J}\right] (4c)
W\displaystyle W =[w1,w2,,wJ]\displaystyle=\left[w_{1},\,w_{2},\,\cdots,\,w_{J}\right] (4d)

the relation (3) implies

W=X+AoXBoUW=X^{+}-A_{\text{o}}X-B_{\text{o}}U (5)

Because the disturbance input wjw_{j} is not measured, WW in (4d) and (5) is unknown. We make the following assumption about the data:

Assumption 1. The disturbance input values w1,w2,w_{1},w_{2}, ,wJ\dots,w_{J}, collected in WW, satisfy the following quadratic matrix inequality:

[IW][Φ11Φ12Φ12Φ22][IW]0\begin{bmatrix}I\\ W^{\top}\end{bmatrix}^{\top}\begin{bmatrix}\Phi_{11}&\Phi_{12}\\ \Phi_{12}^{\top}&\Phi_{22}\end{bmatrix}\begin{bmatrix}I\\ W^{\top}\end{bmatrix}\succeq 0 (6)

where Φ11=Φ11n×n\Phi_{11}=\Phi_{11}^{\top}\in\mathbb{R}^{n\times n}, Φ12n×J\Phi_{12}\in\mathbb{R}^{n\times J}, and Φ22=Φ220\Phi_{22}=\Phi_{22}^{\top}\prec 0 are known matrices.

When Φ12=0\Phi_{12}=0 and Φ22=I\Phi_{22}=-I, (6) reduces to j=1JwjwjΦ11\sum_{j=1}^{J}w_{j}w_{j}^{\top}\preceq\Phi_{11}, which has the interpretation that the total energy of the disturbance inputs in data is bounded by Φ11\Phi_{11}. A known norm bound on each individual disturbance wjw_{j}, wjε\|w_{j}\|\leq\varepsilon, implies a bound in the form of (6) with Φ11=(Jε2)I\Phi_{11}=(J\varepsilon^{2})I, Φ12=0\Phi_{12}=0, and Φ22=I\Phi_{22}=-I.

Plugging W=X+AXBUW=X^{+}-AX-BU into (6) and after some algebra, we obtain

[IAB][ΘXΦ12XΦ22(X+)XΦ22XUΦ12UΦ22(X+)UΦ22XUΦ22U]\displaystyle\begin{bmatrix}I\\ A^{\top}\\ B^{\top}\end{bmatrix}^{\!\!\top}\!\!\begin{bmatrix}\Theta&\star&\star\\ -X\Phi_{12}^{\top}-X\Phi_{22}(X^{+})^{\top}&X\Phi_{22}X^{\top}&\star\\ -U\Phi_{12}^{\top}-U\Phi_{22}(X^{+})^{\top}&U\Phi_{22}X^{\top}&U\Phi_{22}U^{\top}\end{bmatrix} [IAB]\displaystyle\!\!\begin{bmatrix}I\\ A^{\top}\\ B^{\top}\end{bmatrix}
0\displaystyle\!\!\!\!\!\!\!\!\!\!\!\!\succeq 0 (7)

where Θ=Φ11+X+Φ12+Φ12(X+)+X+Φ22(X+)\Theta=\Phi_{11}+X^{+}\Phi_{12}^{\top}+\Phi_{12}(X^{+})^{\top}+X^{+}\Phi_{22}(X^{+})^{\top}, and \star indicates the transpose of the related element below the diagonal. We let Σ\Sigma be the collection of (A,B)(A,B) satisfying (2):

Σ={(A,B):(2) is satisfied}\Sigma=\{(A,B):\text{\eqref{equ:7} is satisfied}\} (8)

According to (5), (Ao,Bo)Σ(A_{\text{o}},B_{\text{o}})\in\Sigma.

Under Assumption 1, we propose a control approach to achieving the three objectives below (1) for the unknown system (1). Our approach is based on the following matrix S-lemma developed in [27]:

Lemma 1 [27]. Let M,N(2n+m)×(2n+m)M,N\in\mathbb{R}^{(2n+m)\times(2n+m)} be symmetric and partitioned as follows:

M=[M11M12M12M22]N=[N11N12N12N22]M=\begin{bmatrix}M_{11}&M_{12}\\ M_{12}^{\top}&M_{22}\end{bmatrix}\quad N=\begin{bmatrix}N_{11}&N_{12}\\ N_{12}^{\top}&N_{22}\end{bmatrix} (9)

where M11,N11n×nM_{11},N_{11}\in\mathbb{R}^{n\times n}. Suppose M220M_{22}\preceq 0, N220N_{22}\preceq 0, ker(N22)ker(N12)\text{ker}(N_{22})\subseteq\text{ker}(N_{12}), and there exists Z¯(n+m)×n\bar{Z}\in\mathbb{R}^{(n+m)\times n} satisfying [IZ¯]N[IZ¯]0\begin{bmatrix}I\\ \bar{Z}\end{bmatrix}^{\top}N\begin{bmatrix}I\\ \bar{Z}\end{bmatrix}\succ 0. Then, we have that

[IZ]M[IZ]0 for all Z satisfying [IZ]N[IZ]0\begin{bmatrix}I\\ Z\end{bmatrix}^{\top}\!M\begin{bmatrix}I\\ Z\end{bmatrix}\succ 0\text{ for all $Z$ satisfying }\begin{bmatrix}I\\ Z\end{bmatrix}^{\top}\!N\begin{bmatrix}I\\ Z\end{bmatrix}\succeq 0 (10)

if and only if there exist α0\alpha\geq 0 and β>0\beta>0 such that

MαN[βI000]M-\alpha N\succeq\begin{bmatrix}\beta I&0\\ 0&0\end{bmatrix} (11)

Proof. See Theorem 13 of [27]. \blacksquare

3 Constrained \mathcal{H}_{\infty} Control

In this section, we consider a linear-feedback u(t)=Kx(t)u(t)=Kx(t) with a constant gain matrix Km×nK\in\mathbb{R}^{m\times n}. This controller yields the following closed-loop system:

x(t+1)\displaystyle x(t+1) =(Ao+BoK)x(t)+w(t)\displaystyle=\left(A_{\text{o}}+B_{\text{o}}K\right)x(t)+w(t) (12a)
y1(t)\displaystyle y_{1}(t) =(C1+D1K)x(t)\displaystyle=\left(C_{1}+D_{1}K\right)x(t) (12b)
y2(t)\displaystyle y_{2}(t) =(C2+D2K)x(t)\displaystyle=\left(C_{2}+D_{2}K\right)x(t) (12c)

The closed-loop transfer matrix from disturbance input ww to performance output y1y_{1} is given by

G1(z)=(C1+D1K)(zI(Ao+BoK))1G_{1}(z)=\left(C_{1}+D_{1}K\right)\left(zI-(A_{\text{o}}+B_{\text{o}}K)\right)^{-1} (13)

For a performance level γ>0\gamma>0, the closed-loop matrix Ac=Ao+BoKA_{\text{c}}=A_{\text{o}}+B_{\text{o}}K is Schur stable and the \mathcal{H}_{\infty} norm of G1(z)G_{1}(z) satisfies G1(z)<γ\|G_{1}(z)\|_{\mathcal{H}_{\infty}}<\gamma if and only if there exists a matrix P=P0P=P^{\top}\succ 0 satisfying

P(Ao+BoK)P(Ao+BoK)(C1+D1K)(C1+D1K)\displaystyle P-(A_{\text{o}}+B_{\text{o}}K)^{\top}P(A_{\text{o}}+B_{\text{o}}K)-(C_{1}+D_{1}K)^{\top}(C_{1}+D_{1}K)
(Ao+BoK)P(γ2IP)1P(Ao+BoK)0\displaystyle-(A_{\text{o}}+B_{\text{o}}K)^{\top}P(\gamma^{2}I-P)^{-1}P(A_{\text{o}}+B_{\text{o}}K)\succ 0 (14a)
γ2IP0\displaystyle\gamma^{2}I-P\succ 0 (14b)

See Theorem 2.2 of [31]. Now let Q=P1Q=P^{-1} (hence, Q=Q0Q=Q^{\top}\succ 0) and Y=KQY=KQ. With some algebra and Schur complement arguments, it can be shown that (14) is equivalent to

R=Q(C1Q+D1Y)(C1Q+D1Y)0\displaystyle R=Q-(C_{1}Q+D_{1}Y)^{\top}(C_{1}Q+D_{1}Y)\succ 0 (15a)
Qγ2I(AoQ+BoY)R1(AoQ+BoY)0\displaystyle Q-\gamma^{-2}I-(A_{\text{o}}Q+B_{\text{o}}Y)R^{-1}(A_{\text{o}}Q+B_{\text{o}}Y)^{\top}\succ 0 (15b)

Note that (15b) can be written as

[IAoBo][Qγ2I000QR1QQR1Y0YR1QYR1Y][IAoBo]0\begin{bmatrix}I\\ A_{\text{o}}^{\top}\\ B_{\text{o}}^{\top}\end{bmatrix}^{\!\top}\!\!\begin{bmatrix}Q-\gamma^{-2}I&0&0\\ 0&-QR^{-1}Q&-QR^{-1}Y^{\top}\\ 0&-YR^{-1}Q&-YR^{-1}Y^{\top}\end{bmatrix}\begin{bmatrix}I\\ A_{\text{o}}^{\top}\\ B_{\text{o}}^{\top}\end{bmatrix}\succ 0 (16)

Now consider the quadratic Lyapunov function V(x(t))=x(t)Px(t)V(x(t))=x(t)^{\top}Px(t). When (14) holds, we can derive the following dissipation inequality:

V(x(t+1))V(x(t))=x(t+1)Px(t+1)x(t)Px(t)\displaystyle V(x(t+1))-V(x(t))=x(t+1)^{\top}Px(t+1)-x(t)^{\top}Px(t)
=x(t)(AcPAcP)x(t)+w(t)Pw(t)\displaystyle=x(t)^{\top}(A_{\text{c}}^{\top}PA_{\text{c}}-P)x(t)+w(t)^{\top}Pw(t)
+x(t)AcPw(t)+w(t)PAcx(t)\displaystyle\,\,+x(t)^{\top}A_{\text{c}}^{\top}Pw(t)+w(t)^{\top}PA_{\text{c}}x(t)
x(t)(C1+D1K)(C1+D1K)x(t)+γ2w(t)w(t)\displaystyle\leq-x(t)^{\top}(C_{1}+D_{1}K)^{\top}(C_{1}+D_{1}K)x(t)+\gamma^{2}w(t)^{\top}w(t)
x(t)AcP(γ2IP)1PAcx(t)w(t)(γ2IP)w(t)\displaystyle\,\,-x(t)^{\top}A_{\text{c}}^{\top}P(\gamma^{2}I-P)^{-1}PA_{\text{c}}x(t)-w(t)^{\top}(\gamma^{2}I-P)w(t)
+x(t)AcPw(t)+w(t)PAcx(t)\displaystyle\,\,+x(t)^{\top}A_{\text{c}}^{\top}Pw(t)+w(t)^{\top}PA_{\text{c}}x(t)
=y1(t)y1(t)+γ2w(t)w(t)S\displaystyle=-y_{1}(t)^{\top}y_{1}(t)+\gamma^{2}w(t)^{\top}w(t)-S
y1(t)2+γ2w(t)2\displaystyle\leq-\|y_{1}(t)\|^{2}+\gamma^{2}\|w(t)\|^{2} (17)

where S=(γ2IP)1/2PAcx(t)(γ2IP)1/2w(t)2S=\|(\gamma^{2}I-P)^{-1/2}PA_{\text{c}}x(t)-(\gamma^{2}I-P)^{1/2}w(t)\|^{2} satisfies S0S\geq 0. We define, for r>0r>0, the ellipsoidal set

(P,r)={xn:V(x)r}\mathcal{E}(P,r)=\{x\in\mathbb{R}^{n}:V(x)\leq r\} (18)

and we arrive at the following result:

Lemma 2. Suppose (14) holds and the energy of the disturbance input is bounded as t=0w(t)2σ0\sum_{t=0}^{\infty}\|w(t)\|^{2}\leq\sigma_{0} for some σ00\sigma_{0}\geq 0. Then, for any r0x(0)Px(0)+γ2σ0r_{0}\geq x(0)^{\top}Px(0)+\gamma^{2}\sigma_{0}, (P,r0)\mathcal{E}(P,r_{0}) is an invariant set of (12), i.e., x(t)(P,r0)x(t)\in\mathcal{E}(P,r_{0}) for all tt\in\mathbb{N}.

Proof. Suppose (14) holds, t=0w(t)2σ0\sum_{t=0}^{\infty}\|w(t)\|^{2}\leq\sigma_{0}, and r0x(0)Px(0)+γ2σ0r_{0}\geq x(0)^{\top}Px(0)+\gamma^{2}\sigma_{0}. Then, the dissipation inequality (3) implies

V(x(t))V(x(0))i=0t1y1(i)2+γ2i=0t1w(i)2\displaystyle V(x(t))\leq V(x(0))-\sum_{i=0}^{t-1}\|y_{1}(i)\|^{2}+\gamma^{2}\sum_{i=0}^{t-1}\|w(i)\|^{2}
x(0)Px(0)+γ2i=0w(i)2r0\displaystyle\quad\quad\leq x(0)^{\top}Px(0)+\gamma^{2}\sum_{i=0}^{\infty}\|w(i)\|^{2}\leq r_{0} (19)

for all tt\in\mathbb{N}. This shows x(t)(P,r0)x(t)\in\mathcal{E}(P,r_{0}) for all tt\in\mathbb{N}. \blacksquare

We now consider the optimization problem (20) for designing the feedback gain K=YQ1K=YQ^{-1}. In (20), eve_{v} denotes the vvth standard basis vector, σ0>0\sigma_{0}>0 is a forecasted energy bound of the disturbance input, and r0>0r_{0}>0 is a design parameter to be tuned so that (20) is feasible. We note that (20) is an LMI (hence, convex) problem in the decision variables (η,Q,Y,α,β)(\eta,Q,Y,\alpha,\beta). A design of KK based on (20) has several desirable properties, stated in the following theorem:

 

maxη>0,Q=Q,Y,α0,β>0η\displaystyle\max_{\eta>0,Q=Q^{\top}\!,Y,\alpha\geq 0,\beta>0}\,\eta (20a)
s.t. [Q(η+β)I000000QYQ000C1Q+D1YI]α[ΘXΦ12XΦ22(X+)XΦ22XUΦ12UΦ22(X+)UΦ22XUΦ22U00000000 0]\displaystyle\begin{bmatrix}Q-(\eta+\beta)I&\star&\star&\star&\star\\ 0&0&\star&\star&\star\\ 0&0&0&\star&\star\\ 0&Q&Y^{\top}&Q&\star\\ 0&0&0&C_{1}Q+D_{1}Y&I\end{bmatrix}\succeq\alpha\begin{bmatrix}\Theta&\star&\star&\star&\,\star\\ -X\Phi_{12}^{\top}-X\Phi_{22}(X^{+})^{\top}&X\Phi_{22}X^{\top}&\star&\star&\,\star\\ -U\Phi_{12}^{\top}-U\Phi_{22}(X^{+})^{\top}&U\Phi_{22}X^{\top}&U\Phi_{22}U^{\top}&\star&\,\star\\ 0&0&0&0&\,\star\\ 0&0&0&0&\,0\end{bmatrix} (20b)
[QC1Q+D1YI]0(20c)[r0x(0)Q10σ01η]0(20d)[y2v,max2r01(C2Q+D2Y)evQ]0,v=1,2,,p2\displaystyle\begin{bmatrix}Q&\star\\ C_{1}Q+D_{1}Y&I\end{bmatrix}\succ 0\quad\,({\text{20c}})\quad\,\,\,\begin{bmatrix}r_{0}&\star&\star\\ x(0)&Q&\star\\ 1&0&\sigma_{0}^{-1}\eta\end{bmatrix}\succeq 0\quad\,({\text{20d}})\quad\,\,\,\begin{bmatrix}y_{2v,\max}^{2}\,r_{0}^{-1}&\star\\ (C_{2}Q+D_{2}Y)^{\top}e_{v}&Q\end{bmatrix}\succeq 0,\,\,v=1,2,\dots,p_{2} (20e)

where Θ=Φ11+X+Φ12+Φ12(X+)+X+Φ22(X+)\Theta=\Phi_{11}+X^{+}\Phi_{12}^{\top}+\Phi_{12}(X^{+})^{\top}+X^{+}\Phi_{22}(X^{+})^{\top}, and \star indicates the transpose of the related element below the diagonal.  

Theorem 1. Suppose

  1. (a)

    The offline data satisfy Assumption 1 and there exists (A,B)=(A¯,B¯)(A,B)=(\bar{A},\bar{B}) such that (2) is strictly positive-definite;

  2. (b)

    The online disturbance inputs satisfy the energy bound t=0w(t)2σ0\sum_{t=0}^{\infty}\|w(t)\|^{2}\leq\sigma_{0};

  3. (c)

    The tuple (η0,Q0,Y0,α0,β0)(\eta_{0},Q_{0},Y_{0},\alpha_{0},\beta_{0}) is a solution to (20).

Then, if we control the unknown system (1) using the linear-feedback u(t)=K0x(t)u(t)=K_{0}x(t), with K0=Y0Q01K_{0}=Y_{0}Q_{0}^{-1}, the closed-loop system has the following properties:

  1. (i)

    The closed-loop matrix Ac,0=Ao+BoK0A_{\text{c},0}=A_{\text{o}}+B_{\text{o}}K_{0} is Schur stable and the closed-loop \mathcal{H}_{\infty} norm from ww to y1y_{1} is less than γ0=η01/2\gamma_{0}=\eta_{0}^{-1/2};

  2. (ii)

    The constraints in (2) are satisfied at all times.

Proof. First, let

M=[M11M12M12M22]=[QηI000QR1QQR1Y0YR1QYR1Y]\displaystyle M=\left[\begin{array}[]{c|c}M_{11}&M_{12}\\ \hline\cr M_{12}^{\top}&M_{22}\end{array}\right]=\left[\begin{array}[]{c|cc}Q-\eta I&0&0\\ \hline\cr 0&-QR^{-1}Q&-QR^{-1}Y^{\top}\\ 0&-YR^{-1}Q&-YR^{-1}Y^{\top}\end{array}\right] (26)
N=[N11N12N12N22]\displaystyle N=\left[\begin{array}[]{c|c}N_{11}&N_{12}\\ \hline\cr N_{12}^{\top}&N_{22}\end{array}\right] (29)
=[ΘXΦ12XΦ22(X+)XΦ22XUΦ12UΦ22(X+)UΦ22XUΦ22U]\displaystyle=\left[\begin{array}[]{c|cc}\Theta&\star&\star\\ \hline\cr-X\Phi_{12}^{\top}-X\Phi_{22}(X^{+})^{\top}&X\Phi_{22}X^{\top}&\star\\ -U\Phi_{12}^{\top}-U\Phi_{22}(X^{+})^{\top}&U\Phi_{22}X^{\top}&U\Phi_{22}U^{\top}\end{array}\right] (33)

where R=Q(C1Q+D1Y)(C1Q+D1Y)R=Q-(C_{1}Q+D_{1}Y)^{\top}(C_{1}Q+D_{1}Y). Using a Schur complement argument, it can be shown that (20b) is equivalent to

MαN[βI000]M-\alpha N\succeq\begin{bmatrix}\beta I&0\\ 0&0\end{bmatrix} (34)

Meanwhile, for the MM and NN defined and partitioned as above, the following conditions can be verified: 1) M220M_{22}\preceq 0, using the fact that R0R\succ 0 due to the constraint (20c), 2) N220N_{22}\preceq 0, using Φ220\Phi_{22}\prec 0 according to Assumption 1, and 3) ker(N22)ker(N12)\text{ker}(N_{22})\subseteq\text{ker}(N_{12}). Now with Z=[A,B]Z=[A,B]^{\top} and Z¯=[A¯,B¯]\bar{Z}=[\bar{A},\bar{B}]^{\top} (with (A¯,B¯)(\bar{A},\bar{B}) given in assumption (a)), it can be seen that the assumptions of Lemma 1 are all satisfied. According to Lemma 1, (34) holds for some α0\alpha\geq 0 and β>0\beta>0 if and only if

[IAB]M[IAB]0 for all (A,B)Σ\begin{bmatrix}I\\ A^{\top}\\ B^{\top}\end{bmatrix}^{\top}M\begin{bmatrix}I\\ A^{\top}\\ B^{\top}\end{bmatrix}\succ 0\text{ for all $(A,B)\in\Sigma$} (35)

where Σ\Sigma is defined in (8). Recall that (Ao,Bo)Σ(A_{\text{o}},B_{\text{o}})\in\Sigma. Therefore, we have shown that if (η,Q=Q,Y,α,β)(\eta,Q=Q^{\top},Y,\alpha,\beta) satisfies (20b), (20c), and η>0,α0,β>0\eta>0,\alpha\geq 0,\beta>0, then (16) (hence, (15b)), with γ=η1/2\gamma=\eta^{-1/2}, necessarily holds. Note that (20c) also implies (15a) and Q0Q\succ 0. Now, if we let P=Q1P=Q^{-1} and K=YQ1K=YQ^{-1}, we arrive at (14) with γ=η1/2\gamma=\eta^{-1/2}. This proves part (i).

Now suppose (η,Q=Q,Y,α,β)(\eta,Q=Q^{\top},Y,\alpha,\beta) satisfies (20b)-(20e) and η>0,α0,β>0\eta>0,\alpha\geq 0,\beta>0. Above we have shown that, with P=Q1P=Q^{-1}, K=YQ1K=YQ^{-1}, and γ=η1/2\gamma=\eta^{-1/2}, (14) holds. Meanwhile, using a Schur complement argument, (20d) implies r0x(0)Q1x(0)+η1σ0=x(0)Px(0)+γ2σ0r_{0}\geq x(0)^{\top}Q^{-1}x(0)+\eta^{-1}\sigma_{0}=x(0)^{\top}Px(0)+\gamma^{2}\sigma_{0}. In this case, according to Lemma 2, t=0w(t)2σ0\sum_{t=0}^{\infty}\|w(t)\|^{2}\leq\sigma_{0} in assumption (b) implies that (P,r0)\mathcal{E}(P,r_{0}) is an invariant set of the closed-loop system, i.e., x(t)(P,r0)x(t)\in\mathcal{E}(P,r_{0}) for all tt\in\mathbb{N}. Recall that (P,r)\mathcal{E}(P,r) is the ellipsoidal set defined in (18). The support function of (P,r)\mathcal{E}(P,r) is h(P,r)(ζ)=rζP1ζh_{\mathcal{E}(P,r)}(\zeta)=\sqrt{r\,\zeta^{\top}P^{-1}\zeta}. Hence, x(t)(P,r0)x(t)\in\mathcal{E}(P,r_{0}) implies

y2v(t)=ev(C2+D2K)x(t)h(P,r0)((C2+D2K)ev)\displaystyle\!y_{2v}(t)=e_{v}^{\top}(C_{2}+D_{2}K)x(t)\leq h_{\mathcal{E}(P,r_{0})}((C_{2}+D_{2}K)^{\top}e_{v})
=r0ev(C2+D2K)P1(C2+D2K)ev\displaystyle=\sqrt{r_{0}\,e_{v}^{\top}(C_{2}+D_{2}K)P^{-1}(C_{2}+D_{2}K)^{\top}e_{v}}
=r0ev(C2Q+D2Y)Q1(C2Q+D2Y)ev\displaystyle=\sqrt{r_{0}\,e_{v}^{\top}(C_{2}Q+D_{2}Y)Q^{-1}(C_{2}Q+D_{2}Y)^{\top}e_{v}} (36)

Meanwhile, using a Schur complement argument, (20e) is equivalent to

y2v,max2r01ev(C2Q+D2Y)Q1(C2Q+D2Y)ev0\displaystyle y_{2v,\max}^{2}\,r_{0}^{-1}-e_{v}^{\top}(C_{2}Q+D_{2}Y)Q^{-1}(C_{2}Q+D_{2}Y)^{\top}e_{v}\geq 0
\displaystyle\iff
r0ev(C2Q+D2Y)Q1(C2Q+D2Y)evy2v,max\displaystyle\sqrt{r_{0}\,e_{v}^{\top}(C_{2}Q+D_{2}Y)Q^{-1}(C_{2}Q+D_{2}Y)^{\top}e_{v}}\leq y_{2v,\max} (37)

Combing (3) and (3), we obtain y2v(t)y2v,maxy_{2v}(t)\leq y_{2v,\max}. This completes the proof of part (ii). \blacksquare

Theorem 1 states the stability, disturbance attenuation, and constraint enforcement properties of the feedback gain KK for the unknown system (1) synthesized using its trajectory data according to (20). The existence of (A¯,B¯)(\bar{A},\bar{B}) such that (2) is strictly positive-definite in assumption (a) of Theorem 1 can be checked offline. We make the following two remarks about Theorem 1:

Remark 1. The \mathcal{H}_{\infty} norm from ww to y1y_{1} represents a level of disturbance attenuation of the closed-loop system. In particular, using (3) recursively we can obtain the following dissipation inequality that holds for all tt\in\mathbb{N}:

i=0ty1(i)2γ2i=0tw(i)2+x(0)Px(0)\sum_{i=0}^{t}\|y_{1}(i)\|^{2}\leq\gamma^{2}\sum_{i=0}^{t}\|w(i)\|^{2}+x(0)^{\top}Px(0) (38)

and this indicates that the ł2\l_{2}-gain from disturbance ww to output y1y_{1} is bounded by γ\gamma. Because (20) maximizes η>0\eta>0, which is equivalent to minimizing γ=η1/2\gamma=\eta^{-1/2}, the obtained controller seeks to maximize disturbance attenuation while satisfying the time-domain constraints in (2).

Remark 2. Differently from many other robust control formulations that assume set-bounded disturbances (i.e., w(t)𝕎w(t)\in\mathbb{W} for some known set bound 𝕎\mathbb{W} and for all tt\in\mathbb{N}), our formulation (20) enforces constraints for disturbance inputs that have bounded total energy where the energy can be distributed arbitrarily over time (i.e., t=0w(t)2σ0\sum_{t=0}^{\infty}\|w(t)\|^{2}\leq\sigma_{0}). This total energy model is particularly suitable for modeling transient disturbances such as wind gusts in aircraft flight control or wind turbine control, power outages in power systems, temporary actuator or sensor failures, etc. Such disturbances occur infrequently and typically last a short period of time (as compared to persistent disturbances), but they can have a significant magnitude. Meanwhile, predicting exactly when they will occur can be difficult or impossible. In such a case, on the one hand, our formulation based on a total energy disturbance model may lead to a less conservative solution (hence, better performance) than those assuming a set bound at all times; on the other hand, estimating/predicting an energy bound for such transient disturbance events using historic data and/or real-time information (such as weather conditions) is possible in many applications. In what follows we assume a mechanism that is able to forecast an energy bound for future disturbances is available. Treating set-bounded persistent disturbances is left for future research.

4 Moving-Horizon Control

We now consider a strategy for implementing the control developed in Section 3 in a moving-horizon manner: At each time tt\in\mathbb{N}, one solves (20) with the current system state as the initial condition x(0)x(0) in (20d) for a feedback gain KtK_{t} and uses the control u(t)=Ktx(t)u(t)=K_{t}x(t) over one time step. This way, the feedback gain becomes adaptive to system state and the control becomes nonlinear, possibly leading to improved performance while satisfying constraints. However, this simple implementation may fail to guarantee stability and disturbance attenuation, as discussed in [3, 4]. To recover a disturbance attenuation guarantee, following the strategy of [3, 4], we consider the following inequality:

x(t)Ptx(t)x(t)Pt1x(t)Δtx(t)^{\top}P_{t}x(t)-x(t)^{\top}P_{t-1}x(t)\leq\Delta_{t} (39)

where PtP_{t} is associated with the feedback gain KtK_{t} that is used at time tt and with an \mathcal{H}_{\infty} performance level of γt\gamma_{t} (i.e., the triple (Pt,Kt,γt)(P_{t},K_{t},\gamma_{t}) satisfies (14)), Pt1P_{t-1} is associated with Kt1K_{t-1} and γt1\gamma_{t-1}, and Δt\Delta_{t} keeps track of a previous dissipation level and is defined recursively according to

Δt=Δt1(x(t1)Pt1x(t1)x(t1)Pt2x(t1))\!\Delta_{t}\!=\!\Delta_{t-1}-\Big{(}x(t-1)^{\top}P_{t-1}x(t-1)-x(t-1)^{\top}P_{t-2}x(t-1)\Big{)} (40)

for t2t\geq 2 and Δ1=0\Delta_{1}=0. Note that the definition of Δt\Delta_{t} in (40) only uses state and PP-matrix information up to time t1t-1.

Lemma 3. Suppose (39) holds for all t1t\geq 1 where Δt\Delta_{t} is defined recursively according to (40). Then, the following dissipation inequality will be satisfied for all tt\in\mathbb{N}:

i=0ty1(i)2γ¯t2i=0tw(i)2+x(0)P0x(0)\sum_{i=0}^{t}\|y_{1}(i)\|^{2}\leq\bar{\gamma}_{t}^{2}\sum_{i=0}^{t}\|w(i)\|^{2}+x(0)^{\top}P_{0}x(0) (41)

where γ¯t=max{γ0,γ1,,γt}\bar{\gamma}_{t}=\max\{\gamma_{0},\gamma_{1},\dots,\gamma_{t}\}.

Proof. For each ii, because (Pi,Ki,γi)(P_{i},K_{i},\gamma_{i}) satisfies (14), similar to (3), we can derive the following inequality:

x(i+1)Pix(i+1)x(i)Pix(i)y1(i)2+γi2w(i)2x(i+1)^{\top}P_{i}x(i+1)-x(i)^{\top}P_{i}x(i)\leq-\|y_{1}(i)\|^{2}+\gamma_{i}^{2}\|w(i)\|^{2} (42)

Summing over i=0,1,,ti=0,1,\dots,t, we obtain

i=0ty1(i)2i=0tγi2w(i)2+x(0)P0x(0)+\displaystyle\!\sum_{i=0}^{t}\|y_{1}(i)\|^{2}\leq\sum_{i=0}^{t}\gamma_{i}^{2}\|w(i)\|^{2}+x(0)^{\top}P_{0}x(0)\,+\, (43)
i=1t(x(i)Pix(i)x(i)Pi1x(i))x(t+1)Ptx(t+1)\displaystyle\!\sum_{i=1}^{t}\!\Big{(}x(i)^{\top}P_{i}x(i)-x(i)^{\top}P_{i-1}x(i)\Big{)}\!-x(t+1)^{\top}P_{t}x(t+1)

The definition (40) yields the following closed-form expression for Δt\Delta_{t}:

Δt=i=1t1(x(i)Pix(i)x(i)Pi1x(i))\Delta_{t}=-\sum_{i=1}^{t-1}\Big{(}x(i)^{\top}P_{i}x(i)-x(i)^{\top}P_{i-1}x(i)\Big{)} (44)

Supposing (39) holds at tt, we have

i=1t(x(i)Pix(i)x(i)Pi1x(i))0\sum_{i=1}^{t}\Big{(}x(i)^{\top}P_{i}x(i)-x(i)^{\top}P_{i-1}x(i)\Big{)}\leq 0 (45)

Combining (43), (45), and γ¯t=maxi=0,,tγi\bar{\gamma}_{t}=\max_{i=0,\dots,t}\gamma_{i}, we obtain

i=0ty1(i)2\displaystyle\sum_{i=0}^{t}\|y_{1}(i)\|^{2} i=0tγi2w(i)2+x(0)P0x(0)\displaystyle\leq\sum_{i=0}^{t}\gamma_{i}^{2}\|w(i)\|^{2}+x(0)^{\top}P_{0}x(0)
x(t+1)Ptx(t+1)\displaystyle\quad\quad\quad-x(t+1)^{\top}P_{t}x(t+1)
γ¯t2i=0tw(i)2+x(0)P0x(0)\displaystyle\leq\bar{\gamma}_{t}^{2}\sum_{i=0}^{t}\|w(i)\|^{2}+x(0)^{\top}P_{0}x(0)\quad\blacksquare (46)

We now consider the following moving-horizon approach to constrained \mathcal{H}_{\infty} control for the unknown system (1): At each time tt\in\mathbb{N}, we solve the optimization problem (47) for designing the feedback gain Kt=YtQt1K_{t}=Y_{t}Q_{t}^{-1} and use the control u(t)=Ktx(t)u(t)=K_{t}x(t) over one time step. In particular, the constraint (47f) is excluded at the initial time t=0t=0 and included for t1t\geq 1. In (47), x(t)x(t) denotes the measured current system state, Pt1=Qt11P_{t-1}=Q_{t-1}^{-1} is from the previous time, Δt\Delta_{t} is defined according to (40), σt\sigma_{t} represents a forecasted bound on the total energy of present and future disturbances, i.e., i=tw(i)2σt\sum_{i=t}^{\infty}\|w(i)\|^{2}\leq\sigma_{t}, and rt>0r_{t}>0 is a design parameter, of which a design method is informed by Lemma 4 and Theorem 2. For a moving-horizon control algorithm, recursive feasibility (i.e., the online optimization problem being feasible at a given time implies the problem being feasible again at the next time) is a highly desirable property. Before we discuss the closed-loop properties of the proposed algorithm, the following lemma provides a recursive feasibility result:

 

maxη>0,Q=Q,Y,α0,β>0η\displaystyle\max_{\eta>0,Q=Q^{\top}\!,Y,\alpha\geq 0,\beta>0}\,\eta (47a)
s.t. [Q(η+β)I000000QYQ000C1Q+D1YI]α[ΘXΦ12XΦ22(X+)XΦ22XUΦ12UΦ22(X+)UΦ22XUΦ22U00000000 0]\displaystyle\begin{bmatrix}Q-(\eta+\beta)I&\star&\star&\star&\star\\ 0&0&\star&\star&\star\\ 0&0&0&\star&\star\\ 0&Q&Y^{\top}&Q&\star\\ 0&0&0&C_{1}Q+D_{1}Y&I\end{bmatrix}\succeq\alpha\begin{bmatrix}\Theta&\star&\star&\star&\,\star\\ -X\Phi_{12}^{\top}-X\Phi_{22}(X^{+})^{\top}&X\Phi_{22}X^{\top}&\star&\star&\,\star\\ -U\Phi_{12}^{\top}-U\Phi_{22}(X^{+})^{\top}&U\Phi_{22}X^{\top}&U\Phi_{22}U^{\top}&\star&\,\star\\ 0&0&0&0&\,\star\\ 0&0&0&0&\,0\end{bmatrix} (47b)
[QC1Q+D1YI]0(47c)[rtx(t)Q10σt1η]0\displaystyle\begin{bmatrix}Q&\star\\ C_{1}Q+D_{1}Y&I\end{bmatrix}\succ 0\quad\quad\quad\quad\quad\quad\quad\quad\,({\text{47c}})\quad\quad\quad\quad\quad\quad\quad\begin{bmatrix}r_{t}&\star&\star\\ x(t)&Q&\star\\ 1&0&\sigma_{t}^{-1}\eta\end{bmatrix}\succeq 0 (47d)
[y2v,max2rt1(C2Q+D2Y)evQ]0,v=1,2,,p2(47e)[x(t)Pt1x(t)+Δtx(t)Q]0\displaystyle\begin{bmatrix}y_{2v,\max}^{2}\,r_{t}^{-1}&\star\\ (C_{2}Q+D_{2}Y)^{\top}e_{v}&Q\end{bmatrix}\succeq 0,\,\,v=1,2,\dots,p_{2}\quad\quad({\text{47e}})\quad\quad\quad\,\begin{bmatrix}x(t)^{\top}P_{t-1}x(t)+\Delta_{t}&\star\\ x(t)&Q\end{bmatrix}\succeq 0 (47f)

where Θ=Φ11+X+Φ12+Φ12(X+)+X+Φ22(X+)\Theta=\Phi_{11}+X^{+}\Phi_{12}^{\top}+\Phi_{12}(X^{+})^{\top}+X^{+}\Phi_{22}(X^{+})^{\top}, \star indicates the transpose of the related element below the diagonal, and the constraint (47f) is excluded at t=0t=0 and included for t1t\geq 1.  

Lemma 4. Suppose (47) is feasible at time tt and (ηt,Qt,Yt,αt,βt)(\eta_{t},Q_{t},Y_{t},\alpha_{t},\beta_{t}) denotes a solution. At time t+1t+1, if the forecasted disturbance energy bound σt+1\sigma_{t+1} satisfies σt+1σtw(t)2\sigma_{t+1}\leq\sigma_{t}-\|w(t)\|^{2} and the parameter rt+1r_{t+1} is chosen to be rt+1=rtr_{t+1}=r_{t}, (47) will be feasible again. In particular, in this case, (ηt,Qt,Yt,αt,βt)(\eta_{t},Q_{t},Y_{t},\alpha_{t},\beta_{t}) remains to be a solution to (47) at time t+1t+1.

Proof. First, we note that the constraints (47b), (35c), and (35e) with rt+1=rtr_{t+1}=r_{t} do not change from tt to t+1t+1. The solution (ηt,Qt,Yt,αt,βt)(\eta_{t},Q_{t},Y_{t},\alpha_{t},\beta_{t}) satisfying (47b) and (35c) implies (14) with P=Qt1P=Q_{t}^{-1}, K=YtQt1K=Y_{t}Q_{t}^{-1}, and γ=ηt1/2\gamma=\eta_{t}^{-1/2}. In this case, similar to (3), the following inequality holds

x(t+1)Qt1x(t+1)x(t)Qt1x(t)\displaystyle x(t+1)^{\top}Q_{t}^{-1}x(t+1)-x(t)^{\top}Q_{t}^{-1}x(t)
y1(t)2+ηt1w(t)2\displaystyle\quad\leq-\|y_{1}(t)\|^{2}+\eta_{t}^{-1}\|w(t)\|^{2} (48)

Meanwhile, (ηt,Qt)(\eta_{t},Q_{t}) satisfying the constraint (47d) at time tt is equivalent to

rtx(t)Qt1x(t)+ηt1σtr_{t}\geq x(t)^{\top}Q_{t}^{-1}x(t)+\eta_{t}^{-1}\sigma_{t} (49)

If σt+1σtw(t)2\sigma_{t+1}\leq\sigma_{t}-\|w(t)\|^{2} and rt+1=rtr_{t+1}=r_{t}, together with (4) and (49), we can obtain

x(t+1)Qt1x(t+1)+ηt1σt+1\displaystyle x(t+1)^{\top}Q_{t}^{-1}x(t+1)+\eta_{t}^{-1}\sigma_{t+1}
x(t)Qt1x(t)+ηt1w(t)2+ηt1(σtw(t)2)\displaystyle\leq x(t)^{\top}Q_{t}^{-1}x(t)+\eta_{t}^{-1}\|w(t)\|^{2}+\eta_{t}^{-1}(\sigma_{t}-\|w(t)\|^{2})
=x(t)Qt1x(t)+ηt1σt\displaystyle=x(t)^{\top}Q_{t}^{-1}x(t)+\eta_{t}^{-1}\sigma_{t}
rt=rt+1\displaystyle\leq r_{t}=r_{t+1} (50)

which implies that (ηt,Qt)(\eta_{t},Q_{t}) also satisfies the constraint (47d) at time t+1t+1. Last, for t=0t=0, we have Δt+1=Δ1=0\Delta_{t+1}=\Delta_{1}=0; for t1t\geq 1, QtQ_{t} satisfying the constraint (47f) implies (39) with Pt=Qt1P_{t}=Q_{t}^{-1}. In the latter case, similar to (44)–(45), we have

Δt+1=i=1t(x(i)Pix(i)x(i)Pi1x(i))0\Delta_{t+1}=-\sum_{i=1}^{t}\Big{(}x(i)^{\top}P_{i}x(i)-x(i)^{\top}P_{i-1}x(i)\Big{)}\geq 0 (51)

At time t+1t+1, the constraint (47f) is equivalent to

x(t+1)Q1x(t+1)x(t+1)Ptx(t+1)Δt+1x(t+1)^{\top}Q^{-1}x(t+1)-x(t+1)^{\top}P_{t}x(t+1)\leq\Delta_{t+1} (52)

where Pt=Qt1P_{t}=Q_{t}^{-1}. Because Δt+10\Delta_{t+1}\geq 0, it is clear that Q=QtQ=Q_{t} satisfies (52) (hence, (47f)).

Therefore, we have shown that if σt+1σtw(t)2\sigma_{t+1}\leq\sigma_{t}-\|w(t)\|^{2} and rt+1=rtr_{t+1}=r_{t}, a solution at time tt, (ηt,Qt,Yt,αt,βt)(\eta_{t},Q_{t},Y_{t},\alpha_{t},\beta_{t}), still satisfies all constraints of (47) at time t+1t+1, i.e., (ηt,Qt,Yt,αt,βt)(\eta_{t},Q_{t},Y_{t},\alpha_{t},\beta_{t}) remains to be a solution to (47) at t+1t+1. This proves the result. \blacksquare

Remark 3. Lemma 4 represents a sufficient condition for the online optimization problem (47) to be recursively feasible. The assumption σt+1σtw(t)2\sigma_{t+1}\leq\sigma_{t}-\|w(t)\|^{2} is reasonable because σt\sigma_{t} bounds the total energy of disturbance inputs from time tt and σt+1\sigma_{t+1} bounds that from t+1t+1 – they differ by the energy of the disturbance input at time tt. Then, rt+1=rtr_{t+1}=r_{t} yields a simple strategy for setting the parameter rtr_{t} – at each time tt, rtr_{t} is set to its previous value. At time instants at which (47) is infeasible with rt=rt1r_{t}=r_{t-1} (due to errors in forecasted disturbance energy bound and violations of σt+1σtw(t)2\sigma_{t+1}\leq\sigma_{t}-\|w(t)\|^{2}), an alternative strategy that promotes feasibility is to include rtr_{t} or its inverse ρt=rt1\rho_{t}=r_{t}^{-1} as a decision variable optimized together with the other variables (η,Q,Y,α,β)(\eta,Q,Y,\alpha,\beta). In this case, (47) becomes a nonconvex problem with a single nonconvex variable rtr_{t} or ρt\rho_{t}, which can be solved using a branch-and-bound type algorithm with branching on rtr_{t} or ρt\rho_{t} [32].

We now analyze the closed-loop properties of the system under the proposed moving-horizon control approach. The properties are given in Theorem 2.

Theorem 2. Suppose

  1. (a)

    The offline data satisfy Assumption 1 and there exists (A,B)=(A¯,B¯)(A,B)=(\bar{A},\bar{B}) such that (2) is strictly positive-definite;

  2. (b)

    The online disturbance inputs satisfy i=tw(i)2σt\sum_{i=t}^{\infty}\|w(i)\|^{2}\leq\sigma_{t} for all tt\in\mathbb{N}, where σt\sigma_{t} are forecasted energy bounds and used in (47);

  3. (c)

    The online optimization problem (47) is feasible at all tt\in\mathbb{N} and (ηt,Qt,Yt,αt,βt)(\eta_{t},Q_{t},Y_{t},\alpha_{t},\beta_{t}) denotes a solution to (47) at each tt;

  4. (d)

    The solutions (ηt,Qt,Yt,αt,βt)(\eta_{t},Q_{t},Y_{t},\alpha_{t},\beta_{t}) have a common lower bound η¯>0\underline{\eta}>0 on their objective values, i.e., ηtη¯\eta_{t}\geq\underline{\eta} for all tt\in\mathbb{N}.

Then, if we control the unknown system (1) using u(t)=Ktx(t)u(t)=K_{t}x(t), with Kt=YtQt1K_{t}=Y_{t}Q_{t}^{-1}, at all tt\in\mathbb{N}, the closed-loop system has the following properties:

  1. (i)

    The system state x(t)x(t) converges to 0 as tt\to\infty;

  2. (ii)

    The following dissipation inequality is satisfied for all tt\in\mathbb{N}:

    i=0ty1(i)2γ¯2i=0tw(i)2+x(0)P0x(0)\sum_{i=0}^{t}\|y_{1}(i)\|^{2}\leq\bar{\gamma}^{2}\sum_{i=0}^{t}\|w(i)\|^{2}+x(0)^{\top}P_{0}x(0) (53)

    where γ¯=η¯1/2\bar{\gamma}=\underline{\eta}^{-1/2}, indicating that the ł2\l_{2}-gain from disturbance ww to output y1y_{1} is bounded by γ¯\bar{\gamma};

  3. (iii)

    The constraints in (2) are satisfied at all times.

Furthermore, suppose (a) and (b) hold and

  1. (e)

    Problem (47) is feasible at the initial time t=0t=0, the energy bounds σt\sigma_{t} satisfy σt+1σtw(t)2\sigma_{t+1}\leq\sigma_{t}-\|w(t)\|^{2} for all tt\in\mathbb{N}, and the parameters rtr_{t} are chosen as rt=r0r_{t}=r_{0} for all tt\in\mathbb{N};

  2. (f)

    The tuple (ηt,Qt,Yt,αt,βt)(\eta_{t},Q_{t},Y_{t},\alpha_{t},\beta_{t}) is a (not only feasible but also) optimal solution to (47) at each tt\in\mathbb{N}.

Then, the following results hold:

  1. (iv)

    The conditions (c) and (d) are necessarily satisfied;

  2. (v)

    The solutions have non-decreasing objective values, i.e., ηt+1ηt\eta_{t+1}\geq\eta_{t} for all tt\in\mathbb{N};

  3. (vi)

    The dissipation inequality (53) holds true with the gain of γ¯=η01/2\bar{\gamma}=\eta_{0}^{-1/2}.

Proof. We start with proving the inequality (53) in (ii). At each time tt\in\mathbb{N}, the solution (ηt,Qt,Yt,αt,βt)(\eta_{t},Q_{t},Y_{t},\alpha_{t},\beta_{t}) satisfies the constraints (47b) and (35c). Following similar steps as in the proof of Theorem 1, part (i), we can show that (14), with P=Qt1P=Q_{t}^{-1}, K=YtQt1K=Y_{t}Q_{t}^{-1}, and γ=ηt1/2\gamma=\eta_{t}^{-1/2}, holds. The constraint (47f) is equivalent to

x(t)Q1x(t)x(t)Pt1x(t)Δtx(t)^{\top}Q^{-1}x(t)-x(t)^{\top}P_{t-1}x(t)\leq\Delta_{t} (54)

Because (47f) (hence, (54)) is satisfied by QtQ_{t} at each t1t\geq 1, we have (39) for all t1t\geq 1. In this case, according to Lemma 3, (41) holds for all tt\in\mathbb{N}. Meanwhile, ηtη¯\eta_{t}\geq\underline{\eta} for all tt\in\mathbb{N} (assumption (d)) implies that γ¯tγ¯\bar{\gamma}_{t}\leq\bar{\gamma} for all tt, where γ¯t=maxi=0,,tγi\bar{\gamma}_{t}=\max_{i=0,\dots,t}\gamma_{i}, γi=ηi1/2\gamma_{i}=\eta_{i}^{-1/2}, and γ¯=η¯1/2\bar{\gamma}=\underline{\eta}^{-1/2}. The combination of (41) and γ¯tγ¯\bar{\gamma}_{t}\leq\bar{\gamma} leads to (53). This proves part (ii). Now, because (53) holds for all tt\in\mathbb{N}, as tt\to\infty, we obtain

i=0y1(i)2γ¯2i=0w(i)2+x(0)P0x(0)\sum_{i=0}^{\infty}\|y_{1}(i)\|^{2}\leq\bar{\gamma}^{2}\sum_{i=0}^{\infty}\|w(i)\|^{2}+x(0)^{\top}P_{0}x(0) (55)

Note that under assumption (b), the right-hand side of (55) is bounded by γ¯2σ0+x(0)P0x(0)\bar{\gamma}^{2}\sigma_{0}+x(0)^{\top}P_{0}x(0), and hence the series on the left-hand side converges according to the monotone convergence theorem. And (55) implies y1(t)0y_{1}(t)\to 0 as tt\to\infty. When (C1,Ao)(C_{1},A_{\text{o}}) is detectable, y1(t)0y_{1}(t)\to 0 implies x(t)0x(t)\to 0. This proves part (i). For part (iii), because (ηt,Qt,Yt)(\eta_{t},Q_{t},Y_{t}) satisfies (47d) and (35e), following similar steps as in the proof of Theorem 1, part (ii), it can be shown that x(t)(Qt1,rt)x(t)\in\mathcal{E}(Q_{t}^{-1},r_{t}) (due to (47d)) and this implies y2v(t)=ev(C2+D2Kt)x(t)y2v,maxy_{2v}(t)=e_{v}^{\top}(C_{2}+D_{2}K_{t})x(t)\leq y_{2v,\max} for all v=1,2,,p2v=1,2,\dots,p_{2} (due to (35e)).

Now suppose (a), (b), (e), and (f) hold. According to Lemma 4, when (47) is feasible at t=0t=0, σ1σ0w(0)2\sigma_{1}\leq\sigma_{0}-\|w(0)\|^{2}, and r1=r0r_{1}=r_{0}, an optimal solution (η0,Q0,Y0,α0,β0)(\eta_{0},Q_{0},Y_{0},\alpha_{0},\beta_{0}) to (47) at t=0t=0 remains to be a feasible solution to (47) at t=1t=1. In this case, (47) is feasible at t=1t=1 and an optimal solution (η1,Q1,Y1,α1,β1)(\eta_{1},Q_{1},Y_{1},\alpha_{1},\beta_{1}) has an objective value at least as large as that of the feasible solution (η0,Q0,Y0,α0,β0)(\eta_{0},Q_{0},Y_{0},\alpha_{0},\beta_{0}), i.e., η1η0\eta_{1}\geq\eta_{0}. Using the same argument recursively, we can conclude that (47) is feasible at all tt\in\mathbb{N} and ηt+1ηt\eta_{t+1}\geq\eta_{t} for all tt. The latter also implies that η0>0\eta_{0}>0 is a common lower bound for all ηt\eta_{t}, i.e., (d) is satisfied with η¯=η0\underline{\eta}=\eta_{0}. Accordingly, (53) holds with γ¯=η¯1/2=η01/2\bar{\gamma}=\underline{\eta}^{-1/2}=\eta_{0}^{-1/2}. This completes the proofs of parts (iv), (v), and (vi). \blacksquare

We make the following remark about Theorem 2:

Remark 4. Theorem 2 shows that our objectives of closed-loop stability, disturbance attenuation, and constraint enforcement stated in Section 2 are fulfilled by the proposed data-driven moving-horizon \mathcal{H}_{\infty} control based on (47). In particular, part (i) shows that the system state converges to zero asymptotically for any disturbance input signal that has bounded total energy. Parts (v) and (vi) show that, under assumption (e) (which is reasonable as discussed in Remark 3), the moving-horizon approach based on (47) achieves a disturbance attenuation performance at least as good as that achieved by a constant linear-feedback designed based on (20). In particular, the moving-horizon approach based on (47) is able to dynamically adapt the feedback gain KtK_{t} to measured/estimated system state x(t)x(t) and forecasted disturbance energy bound σt\sigma_{t} and hence has the potential to achieve improved disturbance attenuation performance while satisfying constraints. We will demonstrate this improvement with a numerical example in the following section.

5 Numerical Example

In this section, we use a numerical example to illustrate our proposed control approach. Consider a system in the form of (1) with the following parameters:

Ao\displaystyle A_{\text{o}} =[0.81470.91340.27850.90580.63240.54690.12700.09750.9575]Bo=[0.67870.75770.7431]\displaystyle=\begin{bmatrix}0.8147&0.9134&0.2785\\ 0.9058&0.6324&0.5469\\ 0.1270&0.0975&0.9575\end{bmatrix}\quad B_{\text{o}}=\begin{bmatrix}-0.6787\\ -0.7577\\ -0.7431\end{bmatrix} (56a)
C1\displaystyle C_{1} =[ 100]D1=  0\displaystyle=\,\begin{bmatrix}\,1&0&0\,\,\end{bmatrix}\quad\quad D_{1}=\,\,0 (56b)
C2\displaystyle C_{2} =[010000]D2=[01]ymax=[10.5]\displaystyle=\begin{bmatrix}0&1&0\\ 0&0&0\end{bmatrix}\quad\quad D_{2}=\begin{bmatrix}0\\ 1\end{bmatrix}\quad\quad y_{\max}=\begin{bmatrix}1\\ 0.5\end{bmatrix} (56c)

The parameters in (56b) mean that we want to minimize the effect of disturbance input w(t)w(t) on the first state x1(t)x_{1}(t), and the parameters in (56c) mean that the second state x2(t)x_{2}(t) and the control input u(t)u(t) should satisfy the constraints x2(t)1x_{2}(t)\leq 1 and u(t)0.5u(t)\leq 0.5. Assume (Ao,Bo)(A_{\text{o}},B_{\text{o}}) is unknown. We simulate the system and collect J=100J=100 data of (xj+,xj,uj)(x^{+}_{j},x_{j},u_{j}) with random disturbance input wjw_{j} that satisfies wjε=102\|w_{j}\|\leq\varepsilon=10^{-2}. Then, we construct the matrices in Assumption 1 as Φ11=(Jε2)I\Phi_{11}=(J\varepsilon^{2})I, Φ12=0\Phi_{12}=0, and Φ22=I\Phi_{22}=-I.

We implement three data-driven approaches using the same set of collected data {(xj+,xj,uj)}j=1J\{(x^{+}_{j},x_{j},u_{j})\}_{j=1}^{J} to control the system: 1) the data-driven \mathcal{H}_{\infty} control approach of [27] (which does not handle constraints), 2) the constrained \mathcal{H}_{\infty} control based on (20) (without moving-horizon implementation), and 3) the moving-horizon \mathcal{H}_{\infty} control based on (47).

We consider the initial condition x(0)=[0.95,0,0]x(0)=[0.95,0,0]^{\top}. For (20) and (47), to illustrate the results of Theorems 1 and 2, we use the parameters σ0=102\sigma_{0}=10^{-2}, σt+1=σtw(t)2\sigma_{t+1}=\sigma_{t}-\|w(t)\|^{2}, and rt=10r_{t}=10 for all tt\in\mathbb{N}.

The system is stabilized and x(t)x(t) converges to 0 under each of the three approaches. However, as shown in Fig. 1, the control using the approach of [27] violates the constraint u(t)0.5u(t)\leq 0.5. This is expected because the approach of [27] does not consider any constraints. In contrast, (20) and (47) both satisfy the constraints, illustrating the effectiveness of our proposed approach for handling constraints. We then compare the time history of γ\gamma using the constrained \mathcal{H}_{\infty} control based on (20) (without moving-horizon implementation) versus that using the moving-horizon \mathcal{H}_{\infty} control based on (47) in Fig. 2. Note that a smaller γ\gamma indicates a stronger attenuation of the effect of disturbance input w(t)w(t) on performance output y1(t)y_{1}(t). It can be seen that both approaches start with a large γ\gamma. This is because the control has to sacrifice some disturbance attenuation performance for satisfying the constraints on x2(t)x_{2}(t) and u(t)u(t). Because the approach of (20) uses a fixed gain, the disturbance attenuation performance level γ\gamma remains constant over time. In contrast, as the state x(t)x(t) and the control u(t)u(t) are getting farther away from the constraint boundaries, the moving-horizon approach based on (47) adjusts the gain and achieves a lower γ\gamma, illustrating the effectiveness of our proposed moving-horizon approach for improving performance while satisfying constraints.

Refer to caption
Figure 1: Control input time history.
Refer to caption
Figure 2: Disturbance attenuation level time history.

In this example, the optimization problem (47) is feasible at every time step, which is consistent with the recursive feasibility result of Lemma 4 and Theorem 2. The average computation time per step for solving (47) using the MATLAB-based LMI solver mincx on a MacBook Air (M1 CPU, 8 GB RAM) is 12.5 ms, indicating the computational feasibility of the approach.

6 Conclusions

In this paper, we proposed a novel data-driven moving-horizon control approach for constrained systems. The approach optimizes \mathcal{H}_{\infty}-type disturbance rejection while satisfying constraints. We established theoretical guarantees of the approach regarding closed-loop stability, disturbance attenuation, constraint satisfaction under noisy offline data, and online problem recursive feasibility. The effectiveness of the approach has been illustrated with a numerical example. Future work includes applying the proposed approach to practical control engineering problems.

References

  • [1] K. Zhou, J. C. Doyle, and K. Glover, Robust and optimal control. USA: Prentice Hall, 1996.
  • [2] J. M. Maciejowski, Predictive control: with constraints. USA: Prentice Hall, 2002.
  • [3] H. Chen and C. Scherer, “Disturbance attenuation with actuator constraints by moving horizon HH_{\infty} control,” IFAC Proceedings, vol. 37, no. 1, pp. 415–420, 2004.
  • [4] H. Chen and C. W. Scherer, “Moving horizon HH_{\infty} control with performance adaptation for constrained linear systems,” Automatica, vol. 42, no. 6, pp. 1033–1040, 2006.
  • [5] S.-M. Lee and J. H. Park, “Robust HH_{\infty} model predictive control for uncertain systems using relaxation matrices,” International Journal of Control, vol. 81, no. 4, pp. 641–650, 2008.
  • [6] P. E. Orukpe, “Towards a less conservative model predictive control based on mixed H2/HH_{2}/H_{\infty} control approach,” International Journal of Control, vol. 84, no. 5, pp. 998–1007, 2011.
  • [7] H. Huang, D. Li, and Y. Xi, “Mixed H2/HH_{2}/H_{\infty} robust model predictive control with saturated inputs,” International Journal of Systems Science, vol. 45, no. 12, pp. 2565–2575, 2014.
  • [8] M. Benallouch, G. Schutz, D. Fiorelli, and M. Boutayeb, “HH_{\infty} model predictive control for discrete-time switched linear systems with application to drinking water supply network,” Journal of Process Control, vol. 24, no. 6, pp. 924–938, 2014.
  • [9] Y. Song, X. Fang, and Q. Diao, “Mixed H2/HH_{2}/H_{\infty} distributed robust model predictive control for polytopic uncertain systems subject to actuator saturation and missing measurements,” International Journal of Systems Science, vol. 47, no. 4, pp. 777–790, 2016.
  • [10] Y. Song, Z. Wang, D. Ding, and G. Wei, “Robust H2/HH_{2}/H_{\infty} model predictive control for linear systems with polytopic uncertainties under weighted MEF-TOD protocol,” IEEE Transactions on Systems, Man, and Cybernetics: Systems, vol. 49, no. 7, pp. 1470–1481, 2017.
  • [11] Y. Zhang, C.-C. Lim, and F. Liu, “Robust mixed H2/HH_{2}/H_{\infty} model predictive control for Markov jump systems with partially uncertain transition probabilities,” Journal of the Franklin Institute, vol. 355, no. 8, pp. 3423–3437, 2018.
  • [12] A. Shokrollahi and S. Shamaghdari, “Robust HH_{\infty} model predictive control for constrained Lipschitz non-linear systems,” Journal of Process Control, vol. 104, pp. 101–111, 2021.
  • [13] F. L. Lewis, D. Vrabie, and K. G. Vamvoudakis, “Reinforcement learning and feedback control: Using natural decision methods to design optimal adaptive controllers,” IEEE Control Systems Magazine, vol. 32, no. 6, pp. 76–105, 2012.
  • [14] B. Recht, “A tour of reinforcement learning: The view from continuous control,” Annual Review of Control, Robotics, and Autonomous Systems, vol. 2, pp. 253–279, 2019.
  • [15] U. Rosolia and F. Borrelli, “Learning model predictive control for iterative tasks. A data-driven control framework,” IEEE Transactions on Automatic Control, vol. 63, no. 7, pp. 1883–1896, 2017.
  • [16] L. Hewing, K. P. Wabersich, M. Menner, and M. N. Zeilinger, “Learning-based model predictive control: Toward safe learning in control,” Annual Review of Control, Robotics, and Autonomous Systems, vol. 3, pp. 269–296, 2020.
  • [17] J. Coulson, J. Lygeros, and F. Dörfler, “Data-enabled predictive control: In the shallows of the DeePC,” in 18th European Control Conference, pp. 307–312, IEEE, 2019.
  • [18] E. Elokda, J. Coulson, P. N. Beuchat, J. Lygeros, and F. Dörfler, “Data-enabled predictive control for quadcopters,” International Journal of Robust and Nonlinear Control, vol. 31, no. 18, pp. 8916–8936, 2021.
  • [19] L. Huang, J. Coulson, J. Lygeros, and F. Dörfler, “Decentralized data-enabled predictive control for power system oscillation damping,” IEEE Transactions on Control Systems Technology, vol. 30, no. 3, pp. 1065–1077, 2021.
  • [20] V. Chinde, Y. Lin, and M. J. Ellis, “Data-enabled predictive control for building HVAC systems,” Journal of Dynamic Systems, Measurement, and Control, vol. 144, no. 8, p. 081001, 2022.
  • [21] N. Li, E. Taheri, I. Kolmanovsky, and D. Filev, “Minimum-time trajectory optimization with data-based models: A linear programming approach,” arXiv preprint arXiv:2312.05724, 2023.
  • [22] S. Baros, C.-Y. Chang, G. E. Colon-Reyes, and A. Bernstein, “Online data-enabled predictive control,” Automatica, vol. 138, p. 109926, 2022.
  • [23] L. Dai, T. Huang, R. Gao, Y. Zhang, and Y. Xia, “Cloud-based computational data-enabled predictive control,” IEEE Internet of Things Journal, vol. 9, no. 24, pp. 24949–24962, 2022.
  • [24] J. Berberich, J. Köhler, M. A. Müller, and F. Allgöwer, “Data-driven model predictive control with stability and robustness guarantees,” IEEE Transactions on Automatic Control, vol. 66, no. 4, pp. 1702–1717, 2020.
  • [25] J. Coulson, J. Lygeros, and F. Dörfler, “Distributionally robust chance constrained data-enabled predictive control,” IEEE Transactions on Automatic Control, vol. 67, no. 7, pp. 3289–3304, 2021.
  • [26] L. Huang, J. Zhen, J. Lygeros, and F. Dörfler, “Robust data-enabled predictive control: Tractable formulations and performance guarantees,” IEEE Transactions on Automatic Control, vol. 68, no. 5, pp. 3163–3170, 2023.
  • [27] H. J. van Waarde, M. K. Camlibel, and M. Mesbahi, “From noisy data to feedback controllers: Nonconservative design via a matrix S-lemma,” IEEE Transactions on Automatic Control, vol. 67, no. 1, pp. 162–175, 2020.
  • [28] K. J. Åström, U. Borisson, L. Ljung, and B. Wittenmark, “Theory and applications of self-tuning regulators,” Automatica, vol. 13, no. 5, pp. 457–476, 1977.
  • [29] D. A. Bristow, M. Tharayil, and A. G. Alleyne, “A survey of iterative learning control,” IEEE Control Systems Magazine, vol. 26, no. 3, pp. 96–114, 2006.
  • [30] Z.-S. Hou and Z. Wang, “From model-based control to data-driven control: Survey, classification and perspective,” Information Sciences, vol. 235, pp. 3–35, 2013.
  • [31] C. E. de Souza and L. Xie, “On the discrete-time bounded real lemma with application in the characterization of static state feedback HH_{\infty} controllers,” Systems & Control Letters, vol. 18, no. 1, pp. 61–71, 1992.
  • [32] H. Tuy, “On nonconvex optimization problems with separated nonconvex variables,” Journal of Global Optimization, vol. 2, pp. 133–144, 1992.