This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Adversarial Attacks to Direct Data-driven Control for Destabilization

Hampei Sasahara This work was supported by JSPS KAKENHI Grant Number 22K21272.H. Sasahara is with the Department of Systems and Control Engineering, School of Engineering, Tokyo Institute of Technology, Tokyo 152-8552, Japan [email protected]The MATLAB codes are available at http://hampei.net/codes/HS23_MATLAB_AdvAttacks.zip
Abstract

This study investigates the vulnerability of direct data-driven control to adversarial attacks in the form of a small but sophisticated perturbation added to the original data. The directed gradient sign method (DGSM) is developed as a specific attack method, based on the fast gradient sign method (FGSM), which has originally been considered in image classification. DGSM uses the gradient of the eigenvalues of the resulting closed-loop system and crafts a perturbation in the direction where the system becomes less stable. It is demonstrated that the system can be destabilized by the attack, even if the original closed-loop system with the clean data has a large margin of stability. To increase the robustness against the attack, regularization methods that have been developed to deal with random disturbances are considered. Their effectiveness is evaluated by numerical experiments using an inverted pendulum model.

I INTRODUCTION

Advances in computing power and an increase in available data have led to the success of data-driven methods in various applications, such as autonomous driving [1], communications [2], and games [3]. In the field of control theory, this success has sparked a trend towards direct data-driven control methods, which aim at designing controllers directly from data without the need for a system identification process [4, 5, 6]. One prominent scheme is the Willems’ fundamental lemma-based approach [7], which provides explicit control formulations and requires low computational complexity [8].

Meanwhile, in the context of image classification, it has been reported that a data-driven method using neural networks is susceptible to adversarial attacks [9, 10, 11]. Specifically, adding small perturbations to images that remain imperceptible to human vision system can change the prediction of the trained neural network classifier. This type of vulnerability has also been observed in different domains such as speech recognition [12] and reinforcement learning [13]. Influenced by those results, adversarial attacks and defenses have become a critical area of research on data-driven techniques.

Most work on control system security focuses on vulnerabilities of control systems themselves and defense techniques with explicit model knowledge against attacks exploiting these vulnerabilities, such as zero-dynamics attack analysis [14], observer-based attack detection [15], and moving target defense [16]. In addition, there have also been recent studies on data-driven approaches, such as data-driven stealthy attack design [17, 18] and data-driven attack detection [19]. However, the vulnerability of data-driven control algorithm has received less attention, and there is a need for dedicated techniques to address this issue.

The main objective of this study is to evaluate the robustness of direct data-driven control methods against adversarial attacks, and to provide insights on how to design secure and reliable data-driven controller design algorithms. The aim of the attacker is to disrupt the stability of the closed-loop system by making small modifications to the data. As the worst-case scenario, we first consider a powerful attacker who has complete knowledge of the system, the controller design algorithm, and the clean input and output data. Subsequently, we consider gray box attacks where we assume that the adversary has access to the model and the algorithm but not the data, and additionally may not know design parameters in the algorithm. Effectiveness of crafted perturbations without partial knowledge is known as the transferability property, which has been confirmed in the domain of computer vision [20] and reinforcement learning [13]. We observe that the data and parameter transferability property holds in direct data-driven control as well.

Our first contribution is to demonstrate the vulnerability of direct data-driven control. We introduce a specific attack, which we refer to as the directed gradient sign method (DGSM), based on the fast gradient sign method (FGSM), which has originally been developed for efficient computation of a severe adversarial perturbation in image classification [10]. The idea behind FGSM is to calculate the perturbation vector in the direction of the gradient of the cost function while limiting each element’s absolute value to a specified small constant. DGSM is an adaptation of this method, designed to destabilize the targeted control system. DGSM calculates the gradient of the eigenvalues of the resulting closed-loop system and determines the perturbation in the direction that makes the system less stable. Fig. 1 illustrates a demonstration of DGSM applied to a discrete-time linear system. It is shown that while the system can be stabilized by using clean data where the resulting eigenvalues are far from the unit circle it can be made unstable by a small but sophisticated perturbation.

Refer to caption
Figure 1: Demonstration of DGSM applied to a discrete-time linear system with three-dimensional input. The adversarial perturbation created by DGSM is added to the original signal, but the perturbed signal appears almost identical to the original one. Nevertheless, the resulting closed-loop system obtained through direct data-driven control with a regularizer becomes unstable due to the adversarial attack. Indeed, the eigenvalues of the closed-loop system with the clean data are {0.0177,0.0212,0.0275}\{-0.0177,0.0212,-0.0275\}, while those with the perturbed data are {0.1824,0.3613,1.0120}\{0.1824,0.3613,1.0120\}. The specific parameters of this instance are provided in Appendix. Note that the output signal is also perturbed but its illustration is omitted for clarity.

Second, we investigate defense methods using regularization. We consider two regularization approaches: the first is the certainty-equivalence regularization that links the direct data-driven control with the indirect one via system identification using the ordinary least-square estimation [21, 22]. The second is the robustness-inducing regularization that ensures robustness against noise [8]. We demonstrate that both approaches can improve robustness against adversarial attacks and compare their effectiveness.

Organization and Notation

The paper is organized as follows. Sec. II reviews key concepts of direct data-driven control based on the fundamental lemma and discusses a technique for generating adversarial perturbations used in image classification with neural networks. In Sec. III, we outline the attack scenario and present the adversarial method adapted for direct data-driven control that leads to destabilization. Sec. IV provides experimental evaluation to discuss the vulnerabilities of interest and the improvement in robustness through regularization. Finally, Sec. V concludes and summarizes the paper.

We denote the transpose of a matrix MM by M𝖳M^{\sf T}, the trace and the spectrum of a square matrix MM by tr(M){\rm tr}(M) and σ(M)\sigma(M), respectively, the maximum and minimum singular values of a matrix MM by σmax(M)\sigma_{\rm max}(M) and σmin(M),\sigma_{\rm min}(M), respectively, the max norm of a matrix MM by Mmax\|M\|_{\rm max}, the right inverse of a right-invertible matrix MM by MM^{\dagger}, the positive and negative (semi)definiteness of a Hermetian matrix MM by MM\succ (\succeq) 0 and MM\prec (\preceq) 0, respectively, and the component-wise sign function by sign(){\rm sign}(\cdot).

II PRELIMINARY

II-A Data-Driven Control based on Fundamental Lemma

We first review the direct data-driven control based on the Willems’ fundamental lemma [7]. Consider a discrete-time linear time-invariant system x(t+1)=Ax(t)+Bu(t)+d(t)x(t+1)=Ax(t)+Bu(t)+d(t) for tt\in\mathbb{N} where x(t)nx(t)\in\mathbb{R}^{n} is the state, u(t)mu(t)\in\mathbb{R}^{m} is the control input, and d(t)nd(t)\in\mathbb{R}^{n} is the exogenous disturbance. Assume that the pair (A,B)(A,B) is unknown to the controller designer but it is stabilizable. We consider the linear quadratic regulator (LQR) problem [23, Chap. 6], which has widely been studied as a benchmark problem. Specifically, design a static state-feedback control u(t)=Kx(t)u(t)=Kx(t) that minimizes the cost function J(K)=i=1nt=0{x(t)𝖳Qx(t)+u(t)𝖳Ru(t)}|x(0)=eiJ(K)=\sum_{i=1}^{n}\sum_{t=0}^{\infty}\{x(t)^{\sf T}Qx(t)+u(t)^{\sf T}Ru(t)\}|_{x(0)=e_{i}} with Q0Q\succeq 0 and R0R\succ 0 where eie_{i} is the iith canonical basis vector. It is known that the cost function can be rewritten as J(K)=tr(QP)+tr(K𝖳RKP)J(K)={\rm tr}(QP)+{\rm tr}(K^{\sf T}RKP) where PIP\succeq I is the controllability Gramian of the closed-loop system when A+BKA+BK is Schur.

The objective of direct data-driven control is to design the optimal feedback gain using data of input and output signals without explicit system identification. Assume that the time series U0:=[u(0)u(1)u(T1)]m×TU_{0}:=[u(0)\ u(1)\ \cdots u(T-1)]\in\mathbb{R}^{m\times T} and X:=[x(0)x(1)x(T1)x(T)]n×(T+1)X:=[x(0)\ x(1)\ \cdots x(T-1)\ x(T)]\in\mathbb{R}^{n\times(T+1)} are available. The first and last TT-long time series of XX are denoted by X0m×TX_{0}\in\mathbb{R}^{m\times T} and X1m×T,X_{1}\in\mathbb{R}^{m\times T}, respectively. Letting D0:=[d(0)d(1)d(T1)]m×T,D_{0}:=[d(0)\ d(1)\ \cdots\ d(T-1)]\in\mathbb{R}^{m\times T}, we have the relationship

X1D0=[BA]W0,X_{1}-D_{0}=[B\ A]W_{0},

where W0:=[U0𝖳X0𝖳]𝖳W_{0}:=[U_{0}^{\sf T}\ X_{0}^{\sf T}]^{\sf T}. We here assume that rankW0=n+m{\rm rank}\,W_{0}=n+m holds. This rank condition, which is generally necessary for data-driven LQR design [24], is satisfied if the input signal is persistently exciting in the noiseless case as shown by the Willems’ fundamental lemma [25].

The key idea of the approach laid out in [7] is to parameterize the controller using the available data by introducing a new variable GT×nG\in\mathbb{R}^{T\times n} with the relationship

[K𝖳I]𝖳=W0G.[K^{\sf T}\ I]^{\sf T}=W_{0}G. (1)

Then the closed-loop matrix can be parameterized directly by data matrices as A+BK=[BA]W0G=(X1D0)G.A+BK=[B\ A]W_{0}G=(X_{1}-D_{0})G. The LQR controller design can be formulated as

minP,K,Gtr(QP)+tr(K𝖳RKP)s.t.X1GPG𝖳X1𝖳P+I0PIand(1)\begin{array}[]{cl}\displaystyle{\min_{P,K,G}}&{\rm tr}(QP)+{\rm tr}(K^{\sf T}RKP)\\ {\rm s.t.}&X_{1}GPG^{\sf T}X_{1}^{\sf T}-P+I\preceq 0\\ &P\succeq I\ {\rm and}\ \eqref{eq:KI}\\ &\end{array} (2)

by disregarding the noise term.

However, it has been revealed that the formulation (2) is not robust to disturbance [21]. To enhance robustness against disturbance, a regularized formulation has been proposed:

minP,K,Gtr(QP)+tr(K𝖳RKP)+γΠGs.t.X1GPG𝖳X1𝖳P+I0PIand(1)\begin{array}[]{cl}\displaystyle{\min_{P,K,G}}&{\rm tr}(QP)+{\rm tr}(K^{\sf T}RKP)+\gamma\|\Pi G\|\\ {\rm s.t.}&X_{1}GPG^{\sf T}X_{1}^{\sf T}-P+I\preceq 0\\ &P\succeq I\ {\rm and}\ \eqref{eq:KI}\\ \end{array} (3)

with a constant γ0\gamma\geq 0 where Π:=IW0W0\Pi:=I-W_{0}^{\dagger}W_{0} and \|\cdot\| is any matrix norm. The regularizer γΠG\gamma\|\Pi G\| is referred to as certainty-equivalence regularization because it leads to the controller equivalent to the certainty-equivalence indirect data-driven LQR with least-square estimation of the system model when γ\gamma is sufficiently large [21]. Meanwhile, another regularization that can guarantee robustness has been proposed:

minP,K,Gtr(QP)+tr(K𝖳RKP)+ρtr(GPG𝖳)s.t.X1GPG𝖳X1𝖳P+I0PIand(1)\begin{array}[]{cl}\displaystyle{\min_{P,K,G}}&{\rm tr}(QP)+{\rm tr}(K^{\sf T}RKP)+\rho\,{\rm tr}(GPG^{\sf T})\\ {\rm s.t.}&X_{1}GPG^{\sf T}X_{1}^{\sf T}-P+I\preceq 0\\ &P\succeq I\ {\rm and}\ \eqref{eq:KI}\\ \end{array} (4)

with a constant ρ0\rho\geq 0. The regularizer ρtr(GPG𝖳)\rho\,{\rm tr}(GPG^{\sf T}) plays the role to reduce the size of the matrix GPG𝖳GPG^{\sf T} to achieve the actual stability requirement (X1D0)GPG𝖳(X1D0)𝖳P+I0(X_{1}-D_{0})GPG^{\sf T}(X_{1}-D_{0})^{\sf T}-P+I\preceq 0 using the constraint X1GPG𝖳X1𝖳P+I0X_{1}GPG^{\sf T}X_{1}^{\sf T}-P+I\preceq 0. We refer to the latter one as robustness-inducing regularization. For reformulation of (2), (3), and (4) into convex programs, see [26].

II-B Fast Gradient Sign Method

The fast gradient sign method (FGSM) is a method to efficiently compute an adversarial perturbation for a given image [10]. Let L(X,Y;θ)L(X,Y;\theta) be the loss function of the neural network where X𝒳X\in\mathcal{X} is the input image, Y𝒴Y\in\mathcal{Y} is the label, and θ\theta is the trained parameter, and let f:𝒳𝒴f:\mathcal{X}\to\mathcal{Y} be the trained classification model. The objective of the adversary is to cause misclassification by adding a small perturbation Δ𝒳\Delta\in\mathcal{X} such that f(X+Δ)f(X)f(X+\Delta)\neq f(X). Specifically, the max norm of the perturbation is restricted, i.e., Δmaxϵ\|\Delta\|_{\rm max}\leq\epsilon with a small constant ϵ>0\epsilon>0.

The core idea of FGSM is to choose a perturbation that locally maximizes the loss function. The linear approximation of the loss function with respect to Δ\Delta is given by

L(X+Δ,Y;θ)L(X,Y;θ)+k,(XL(X,Y;θ))kΔkL(X+\Delta,Y;\theta)\simeq L(X,Y;\theta)+\sum_{k,\ell}(\nabla_{X}L(X,Y;\theta))_{k\ell}\Delta_{k\ell} (5)

where the subscript ()k(\cdot)_{k\ell} denotes the (k,)(k,\ell) component. The right-hand side of (5) is maximized by choosing Δk=ϵsign(XL(X,Y;θ))k\Delta_{k\ell}=\epsilon\,{\rm sign}(\nabla_{X}L(X,Y;\theta))_{k\ell}, whose matrix form is given by

Δ=ϵsign(XL(X,Y;θ)).\Delta=\epsilon\,{\rm sign}(\nabla_{X}L(X,Y;\theta)).

FGSM creates a series of perturbations in the form increasing ϵ\epsilon until misclassification occurs. In the next section, we apply this idea to adversarial attacks on direct data-driven control for destabilization.

III ADVERSARIAL ATTACKS to DIRECT DATA-DRIVEN CONTROL

III-A Threat Model

This study considers the following threat model: The adversary can add a perturbation (ΔU,ΔX)(\Delta U,\Delta X) to the input and output data (U0,X)(U_{0},X). Additionally, the adversary knows the system model (A,B)(A,B), the data (U0,D0,X)(U_{0},D_{0},X), and the controller design algorithm. This scenario is depicted in Fig. 2. The controller K^\hat{K} is designed using the perturbed data (U^,X^):=(U0+ΔU,X+ΔX)(\hat{U},\hat{X}):=(U_{0}+\Delta U,X+\Delta X), which results in the closed-loop matrix A+BK^A+B\hat{K}. The attack objective is to destabilize the system by crafting a small perturbation such that the closed-loop matrix has an eigenvalue outside the unit circle.

Refer to caption
Figure 2: Threat model considered in this paper. The adversary is able to add a perturbation (ΔU,ΔX)(\Delta U,\Delta X) to the original input and output data (U0,X)(U_{0},X) with knowledge of the system model, the signals, and the controller design algorithm. The controller K^\hat{K} is designed using the perturbed data (U^,X^)(\hat{U},\hat{X}), which results in the closed-loop matrix A+BK^A+B\hat{K}.

Additionally, we consider gray-box attacks where the adversary has access to the system model (A,B)(A,B) and the controller design algorithm but not the data (U0,D0,X)(U_{0},D_{0},X), and additionally may not know the design parameters γ\gamma and ρ\rho. In this case, a reasonable attack strategy is to use hypothetical input U^\hat{U} and disturbance D^0\hat{D}_{0} and calculate the corresponding state trajectory X^\hat{X}. We refer to effectiveness of the attack without knowledge of the data as the transferability across data. Additionally, when the design parameters are unknown, hypothetical design parameters γ^\hat{\gamma} or ρ^\hat{\rho} are also used. We refer to the effectiveness in this scenario as the transferability across parameters. We numerically evaluate the transferability properties in Sec. IV.

III-B Directed Gradient Sign Method

We develop the directed gradient sign method (DGSM) to design a severe perturbation Δ:=(ΔU,ΔX)\Delta:=(\Delta U,\Delta X) that satisfies Δmaxϵ\|\Delta\|_{\rm max}\leq\epsilon with a small constant ϵ>0\epsilon>0. Let

Λ(U0,X,Δ):=σ(A+BK^)\Lambda(U_{0},X,\Delta):=\sigma(A+B\hat{K})

denote the eigenvalues of the closed-loop system with the direct data-driven control (3) or (4) using the perturbed data (U^,X^)(\hat{U},\hat{X}). The aim of the attack is to place some element of Λ(U0,X,Δ)\Lambda(U_{0},X,\Delta) outside the unit circle.

The core idea of DGSM is to choose a perturbation that locally shifts an eigenvalue in the less stable direction. We temporarily fix the eigenvalue of interest, denoted by λi(U0,X,Δ)\lambda_{i}(U_{0},X,\Delta), and denote its gradient with respect to Δ\Delta by Δλi(U0,X,Δ)\nabla_{\Delta}\lambda_{i}(U_{0},X,\Delta). The linear approximation of the eigenvalue with respect to Δ\Delta is given by

λi(U0,X,Δ)λi(U0,X,0)+k,Δλi(U0,X,Δ))Δk.\lambda_{i}(U_{0},X,\Delta)\simeq\lambda_{i}(U_{0},X,0)+\sum_{k,\ell}\nabla_{\Delta}\lambda_{i}(U_{0},X,\Delta))\Delta_{k\ell}. (6)

We choose Δk\Delta_{k\ell} such that the right-hand side of (6) moves closer to the unit circle. Specifically, DGSM crafts the perturbation

Δ=ϵsign(Πλi(Δλi(U0,X,Δ)))\Delta=\epsilon\,{\rm sign}(\Pi_{\lambda_{i}}(\nabla_{\Delta}\lambda_{i}(U_{0},X,\Delta)))

where Πλi:(m+n)×(2T+1)(m+n)×(2T+1)\Pi_{\lambda_{i}}:\mathbb{C}^{(m+n)\times(2T+1)}\to\mathbb{R}^{(m+n)\times(2T+1)} is defined by

Πλi(Z):=Re(λi)Re(Z)+Im(λi)Im(Z)\Pi_{\lambda_{i}}(Z):={\rm Re}\,(\lambda_{i}){\rm Re}\,(Z)+{\rm Im}\,(\lambda_{i}){\rm Im}\,(Z) (7)

with

Z:=Δλi(U0,X,Δ).Z:=\nabla_{\Delta}\lambda_{i}(U_{0},X,\Delta).

The role of the function Πλi\Pi_{\lambda_{i}} is illustrated in Fig. 3. Suppose that ZkZ_{k\ell} faces the direction of λi\lambda_{i}. More precisely, the angle between λi\lambda_{i} and ZkZ_{k\ell}, denoted by ϕ\phi, is less than π/2\pi/2, which leads to Πλi(Zk)>0\Pi_{\lambda_{i}}(Z_{k\ell})>0. We now suppose that the angle between λi\lambda_{i} and another element Zk~~Z_{\tilde{k}\tilde{\ell}}, denoted by ϕ~\tilde{\phi}, is greater than π/2\pi/2. Then we have Πλi(Zk~~)<0\Pi_{\lambda_{i}}(Z_{\tilde{k}\tilde{\ell}})<0. In both cases, owing to the function Πλi,\Pi_{\lambda_{i}}, the perturbed eigenvalue moves closer to the unit circle as depicted in the figure. By aggregating all components, the linear approximation of the perturbed eigenvalue λ^i\hat{\lambda}_{i} is given by

λ^iλi+ϵk,sign(Πλi(Zk))Zk,\textstyle{\hat{\lambda}_{i}\simeq\lambda_{i}+\epsilon\,\sum_{k,\ell}{\rm sign}(\Pi_{\lambda_{i}}(Z_{k\ell}))Z_{k\ell},}

which is expected to be placed outside the unit circle by increasing ϵ\epsilon.

Refer to caption
Figure 3: Role of the function Πλi\Pi_{\lambda_{i}} in (7). Since ZkZ_{k\ell} faces the direction of λi\lambda_{i}, the angle ϕ\phi between λi\lambda_{i} and ZkZ_{k\ell} is less than π/2\pi/2, which leads to Πλi(Zk)>0\Pi_{\lambda_{i}}(Z_{k\ell})>0. On the other hand, since ϕ~\tilde{\phi} between λi\lambda_{i} and Zk~~Z_{\tilde{k}\tilde{\ell}} is greater than π/2\pi/2, πλi(Zk~~)<0\pi_{\lambda_{i}}(Z_{\tilde{k}\tilde{\ell}})<0. As a result, in both cases, the perturbed eigenvalue moves closer to the unit circle.

DGSM performs the procedure above for every λi\lambda_{i} for i=1,,ni=1,\ldots,n increasing ϵ\epsilon until the resulting system is destabilized. Its algorithm is summarized in Algorithm 1, where {ϵk}\{\epsilon_{k}\} denotes possible candidates of the constant ϵ\epsilon in the ascending order. Algorithm 1 finds a perturbation Δ\Delta with the smallest ϵ\epsilon in {ϵk}\{\epsilon_{k}\} such that the resulting closed-loop system becomes unstable.

Algorithm 1 Directed Gradient Sign Method (DGSM)
0:  {ϵk},A,B,U0,X,γ,ρ\{\epsilon_{k}\},A,B,U_{0},X,\gamma,\rho
0:  Δ\Delta
1:  flag0{\rm flag}\leftarrow 0
2:  k0k\leftarrow 0
3:  while flag=0{\rm flag}=0 do
4:     kk+1k\leftarrow k+1
5:     for i=1,,ni=1,\ldots,n do
6:        Δϵksign(Πλi(Δλi(U0,X,Δ)))\Delta\leftarrow\epsilon_{k}{\rm sign}(\Pi_{\lambda_{i}}(\nabla_{\Delta}\lambda_{i}(U_{0},X,\Delta)))
7:        if |λi(U0,X,Δ)|>1|\lambda_{i}(U_{0},X,\Delta)|>1 then
8:           flag1{\rm flag}\leftarrow 1
9:           break
10:        end if
11:     end for
12:  end while
13:  return  Δ\Delta

IV NUMERICAL EXPERIMENTS

IV-A Experimental Setup

We evaluate our adversarial attacks through numerical experiments. We consider the inverted pendulum [27] with sampling period 0.010.01 whose system matrices are given by

A=[0.98440.04660.03470.03971.00090.00070.00040.02001.0000],B=[0.2500].A=\left[\begin{array}[]{ccc}0.9844&0.0466&0.0347\\ 0.0397&1.0009&0.0007\\ 0.0004&0.0200&1.0000\end{array}\right],\quad B=\left[\begin{array}[]{c}0.25\\ 0\\ 0\end{array}\right].

We set the weight matrices to Q=IQ=I and R=105IR=10^{-5}I. The input signal is randomly and independently generated by u(t)𝒩(0,1)u(t)\sim\mathcal{N}(0,1). We consider the disturbance-free case, i.e., d(t)=0d(t)=0. The time horizon is set to T=10T=10. The 22-induced norm is taken as the matrix norm in (3). The gradient Δ(λi(U0,X,Δ))\nabla_{\Delta}(\lambda_{i}(U_{0},X,\Delta)) is computed by the central difference approximation [28, Chapter 4].

IV-B Robustness Improvement by Regularization

We examine the improvement in robustness through regularization by comparing DGSM with a random attack where each element of Δ\Delta takes ϵ\epsilon or ϵ-\epsilon with equal probability. Let NallN_{\rm all} and NunstableN_{\rm unstable} denote the total number of samples and the number of the samples where the resulting closed-loop system is unstable, respectively. In addition, let ϵ¯\bar{\epsilon} denote the minimum ϵ\epsilon such that Nunstable/NallτN_{\rm unstable}/N_{\rm all}\geq\tau for a given threshold τ[0,1]\tau\in[0,1]. We set Nall=50N_{\rm all}=50 and τ=0.8\tau=0.8.

Fig. 4 depicts the curves of ϵ¯\bar{\epsilon} with varying γ\gamma for DGSM and the random attack when using the certainty-equivalence regularization (3). First, it is observed that the magnitude of the adversarial perturbation necessary for destabilization increases as the regularization parameter γ\gamma increases. This result implies that the regularization method originally proposed for coping with disturbance is also effective in improving robustness against adversarial attacks. Second, the necessary magnitude in DGSM is approximately 10%10\% of that in the random attack, which illustrates the significant impact of DGSM.

Refer to caption
Figure 4: Curves of ϵ¯\bar{\epsilon} with varying γ\gamma for DGSM and the random attack when using the certainty-equivalence regularization (3).

Fig. 5 depicts the curves of ϵ¯\bar{\epsilon} with varying ρ\rho when using the robustness-inducing regularization (4). This figure shows results similar to Fig. 4. Consequently, both regularization methods are effective for adversarial attacks.

Refer to caption
Figure 5: Curves of ϵ¯\bar{\epsilon} with varying ρ\rho when using the robustness-inducing regularization (4).

Next, we compare the effectiveness of the two regularization methods. We take γ=0.1\gamma=0.1 and ρ=105\rho=10^{-5} such that the resulting closed-loop performances J(K)J(K) are almost equal. Fig. 6 depicts Nunstable/NallN_{\rm unstable}/N_{\rm all} with the two regularized controller design methods (3) and (4) for varying ϵ\epsilon. It can be observed that the robustness-inducing regularization (4) always outperforms the certainty-equivalence regularization (3).

Refer to caption
Figure 6: Ratio of unstable instances Nunstable/NallN_{\rm unstable}/N_{\rm all} with the two regularized controller design methods (3) and (4) for varying ϵ\epsilon.

IV-C Transferability

We consider transferability across data where the data (U0,D0,X)(U_{0},D_{0},X) is unknown and DGSM uses a hypothetical input U^0\hat{U}_{0} whose elements are also randomly and independently generated by 𝒩(0,1)\mathcal{N}(0,1) and D^0=0\hat{D}_{0}=0. Fig. 7 depicts the curves of ϵ¯\bar{\epsilon} with varying γ\gamma for DGSM without knowledge of data and that with full knowledge when using the certainty-equivalence regularization (3). This figure shows that DGSM exhibits the transferability property across data.

Refer to caption
Figure 7: Curves of ϵ¯\bar{\epsilon} with varying γ\gamma for DGSM without knowledge of data and that with full knowledge.

Subsequently, we examine transferability across design parameters where the regularization parameter γ\gamma in addition to the data (U0,D0,X)(U_{0},D_{0},X) is unknown. We use γ=0.1\gamma=0.1 as a hypothetical parameter. Fig. 8 depicts the corresponding curves As in the transferability across data, the results confirm the transferability property across parameters.

Refer to caption
Figure 8: Curves of ϵ¯\bar{\epsilon} with varying γ\gamma for DGSM without knowledge of data and parameters and that with full knowledge.

IV-D Discussion

The regularization methods described by (3) and (4) provide a quantitative condition to ensure stability: The resulting closed-loop system with the certainty-equivalence regularization is stable when γ\gamma and the signal-to-noise ratio (SNR) defined by SNR:=σmin(W0)/σmax(D0){\rm SNR}:=\sigma_{\rm min}(W_{0})/\sigma_{\max}(D_{0}) are sufficiently large [26, Theorem 4.2]. That with the robustness-inducing regularization is stable when ρ\rho is sufficiently large and σmax(D0)\sigma_{\rm max}(D_{0}) is sufficiently small [8, Theorem 3]. One may expect that DGSM crafts a severe input perturbation such that its maximum singular value is large but its elements are small. However, for the single-input system, σmax(ΔU)=ϵT\sigma_{\rm max}(\Delta U)=\epsilon\sqrt{T} for any ΔU\Delta U whose elements take ϵ\epsilon or ϵ-\epsilon. This means that the input perturbations made by DGSM and the random attack have the same maximum singular value.

V CONCLUSION

This study has investigated the vulnerability of direct data-driven control, specifically focusing on the Willems’ fundamental lemma-based approach with two regularization methods, namely certainty-equivalence regularization and robustness-inducing regularization. To this end, a new method called DGSM, based on FGSM which has been originally been proposed for neural networks, has been introduced. It has been demonstrated that direct data-driven control can be vulnerable, i.e., the resulting closed-loop system can be destabilized by a small but sophisticated perturbation. Numerical experiments have indicated that strengthening regularization enhances robustness against adversarial attacks.

Future research should include further tests of the vulnerability with various types of data and systems under different operating conditions, a theoretical analysis of DGSM, and exploration of novel defense techniques for reliable direct data-driven control. For example, detection of adversarial perturbations [29] is a promising direction. Finally, for a more comprehensive understanding of the vulnerability, more sophisticated attacks should be considered.

The parameters in the simulation in Fig. 1 are as follows. The system is a marginally unstable Laplacian system considered in [21, 30]. Each element of the disturbance DD is randomly generated from 𝒩(0,d2)\mathcal{N}(0,d^{2}) with d=0.05d=0.05. The weight matrices are Q=IQ=I and R=103IR=10^{-3}I. The time horizon is set to T=15T=15. The magnitude of the adversarial perturbation is set to ϵ=0.16\epsilon=0.16. The controller is designed using the certainty-equivalence regularization (3) with a regularization parameter γ=103\gamma=10^{-3}.

References

  • [1] B. R. Kiran, I. Sobh, V. Talpaert, P. Mannion, A. A. A. Sallab, S. Yogamani, and P. Pérez, “Deep reinforcement learning for autonomous driving: A survey,” IEEE Trans. Intell. Transp. Syst., vol. 23, no. 6, pp. 4909–4926, 2022.
  • [2] N. C. Luong, D. T. Hoang, S. Gong, D. Niyato, P. Wang, Y.-C. Liang, and D. I. Kim, “Applications of deep reinforcement learning in communications and networking: A survey,” IEEE Commun. Surv. Tutor., vol. 21, no. 4, pp. 3133–3174, 2019.
  • [3] D. Silver, A. Huang, C. J. Maddison, A. Guez, L. Sifre, G. Van Den Driessche, J. Schrittwieser, I. Antonoglou, V. Panneershelvam, M. Lanctot et al., “Mastering the game of Go with deep neural networks and tree search,” Nature, vol. 529, no. 7587, pp. 484–489, 2016.
  • [4] M. C. Campi and S. M. Savaresi, “Direct nonlinear control design: The virtual reference feedback tuning (VRFT) approach,” IEEE Trans. Autom. Control, vol. 51, no. 1, pp. 14–27, 2006.
  • [5] F. L. Lewis and D. Liu, Reinforcement learning and approximate dynamic programming for feedback control.   John Wiley & Sons, 2013.
  • [6] H. Mohammadi, A. Zare, M. Soltanolkotabi, and M. R. Jovanović, “Convergence and sample complexity of gradient methods for the model-free linear–quadratic regulator problem,” IEEE Trans. Autom. Control, vol. 67, no. 5, pp. 2435–2450, 2021.
  • [7] C. De Persis and P. Tesi, “Formulas for data-driven control: Stabilization, optimality, and robustness,” IEEE Trans. Autom. Control, vol. 65, no. 3, pp. 909–924, 2019.
  • [8] ——, “Low-complexity learning of linear quadratic regulators from noisy data,” Automatica, vol. 128, no. 109548, 2021.
  • [9] J. Bruna, C. Szegedy, I. Sutskever, I. Goodfellow, W. Zaremba, R. Fergus, and D. Erhan, “Intriguing properties of neural networks,” in International Conference on Learning Representations, 2014.
  • [10] I. Goodfellow, J. Shlens, and C. Szegedy, “Explaining and harnessing adversarial examples,” in International Conference on Learning Representations, 2015.
  • [11] N. Akhtar and A. Mian, “Threat of adversarial attacks on deep learning in computer vision: A survey,” IEEE Access, vol. 6, pp. 14 410–14 430, 2018.
  • [12] N. Carlini, P. Mishra, T. Vaidya, Y. Zhang, M. Sherr, C. Shields, D. Wagner, and W. Zhou, “Hidden voice commands,” in 25th USENIX security symposium (USENIX security 16), 2016, pp. 513–530.
  • [13] S. Huang, N. Papernot, I. Goodfellow, Y. Duan, and P. Abbeel, “Adversarial attacks on neural network policies,” in International Conference on Learning Representations, 2017.
  • [14] F. Pasqualetti, F. Dörfler, and F. Bullo, “Control-theoretic methods for cyberphysical security: Geometric principles for optimal cross-layer resilient control systems,” IEEE Control Systems Magazine, vol. 35, no. 1, pp. 110–127, 2015.
  • [15] J. Giraldo et al., “A survey of physics-based attack detection in cyber-physical systems,” ACM Comput. Surv., vol. 51, no. 4, 2018.
  • [16] P. Griffioen, S. Weerakkody, and B. Sinopoli, “A moving target defense for securing cyber-physical systems,” IEEE Trans. Autom. Control, vol. 66, no. 5, pp. 2016–2031, 2020.
  • [17] R. Alisic and H. Sandberg, “Data-injection attacks using historical inputs and outputs,” in 2021 European Control Conference (ECC), 2021, pp. 1399–1405.
  • [18] R. Alisic, J. Kim, and H. Sandberg, “Model-free undetectable attacks on linear systems using LWE-based encryption,” IEEE Control Syst. Lett., 2023.
  • [19] V. Krishnan and F. Pasqualetti, “Data-driven attack detection for linear systems,” IEEE Control Syst. Lett., vol. 5, no. 2, pp. 671–676, 2020.
  • [20] C. Szegedy, W. Zaremba, I. Sutskever, J. Bruna, D. Erhan, I. Goodfellow, and R. Fergus, “Intriguing properties of neural networks,” in International Conference on Learning Representations, 2014.
  • [21] F. Dörfler, P. Tesi, and C. De Persis, “On the role of regularization in direct data-driven LQR control,” in 2022 IEEE 62nd Conference on Decision and Control (CDC), 2022, pp. 1091–1098.
  • [22] F. Dorfler, J. Coulson, and I. Markovsky, “Bridging direct & indirect data-driven control formulations via regularizations and relaxations,” IEEE Trans. Autom. Control, 2022.
  • [23] T. Chen and B. A. Francis, Optimal sampled-data control systems.   Springer, 2012.
  • [24] H. J. Van Waarde, J. Eising, H. L. Trentelman, and M. K. Camlibel, “Data informativity: A new perspective on data-driven analysis and control,” IEEE Trans. Autom. Control, vol. 65, no. 11, pp. 4753–4768, 2020.
  • [25] J. C. Willems, P. Rapisarda, I. Markovsky, and B. L. De Moor, “A note on persistency of excitation,” Systems & Control Letters, vol. 54, no. 4, pp. 325–329, 2005.
  • [26] F. Dörfler, P. Tesi, and C. De Persis, “On the certainty-equivalence approach to direct data-driven LQR design,” 2021, [Online]. Available: https://arxiv.org/pdf/2109.06643.pdf.
  • [27] P. Chalupa and V. Bobál, “Modelling and predictive control of inverted pendulum,” in 22nd European Conference on Modelling and Simulation, vol. 3, no. 6, 2008, pp. 531–537.
  • [28] R. L. Burden, J. D. Faires, and A. M. Burden, Numerical Analysis, 10th ed.   Cengage learning, 2015.
  • [29] J. H. Metzen, T. Genewein, V. Fischer, and B. Bischoff, “On detecting adversarial perturbations,” in International Conference on Learning Representations, 2017.
  • [30] S. Dean, H. Mania, N. Matni, B. Recht, and S. Tu, “On the sample complexity of the linear quadratic regulator,” Foundations of Computational Mathematics, vol. 20, no. 4, pp. 633–679, 2020.