This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Bayesian Differential Privacy for Linear Dynamical Systems

Genki Sugiura, Kaito Ito,  and Kenji Kashima The authors are with the Graduate School of Informatics, Kyoto University, Kyoto, Japan e-mail: [email protected].©\copyright 2021 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.
Abstract

Differential privacy is a privacy measure based on the difficulty of discriminating between similar input data. In differential privacy analysis, similar data usually implies that their distance does not exceed a predetermined threshold. It, consequently, does not take into account the difficulty of distinguishing data sets that are far apart, which often contain highly private information. This problem has been pointed out in the research on differential privacy for static data, and Bayesian differential privacy has been proposed, which provides a privacy protection level even for outlier data by utilizing the prior distribution of the data. In this study, we introduce this Bayesian differential privacy to dynamical systems, and provide privacy guarantees for distant input data pairs and reveal its fundamental property. For example, we design a mechanism that satisfies the desired level of privacy protection, which characterizes the trade-off between privacy and information utility.

Index Terms:
Control system security, differential privacy, stochastic system.

I Introduction

As the Internet-of-Things (IoT) and cloud computing are attracting more and more attention for their convenience, privacy protection and security have become key technologies in control systems. To cope with privacy threats, many privacy protection methods have been studied so far [1, 2, 3]. Among them, differential privacy [4] has been used to solve many privacy-related problems in areas such as smart grid [5], health management [6], and blockchain [7], because it can mathematically quantify privacy guarantees. Differential privacy has originally been applied to static data, but as shown in the example of power systems above, there is an urgent need to establish privacy protection techniques for dynamical systems. In recent years, the concept of differential privacy has been introduced to dynamical systems [8], and from the viewpoint of control systems theory, the relationship between privacy protection and the observability of systems has been clarified, and methods of controller design with privacy protection in mind have been studied [9, 10, 11].

Conventional differential privacy is a privacy measure based on the difficulty of distinguishing similar data, where xx and xx^{\prime} are regarded as being similar if |xx|c|x-x^{\prime}|\leq c for a prescribed c>0c>0. To put it conversely, there is no indistinguishability guarantee for xx and xx^{\prime} if |xx|>c|x-x^{\prime}|>c. This implies there is a risk of information leakage when there are outliers from normal data as pointed out in [12]. For example, unusual electricity consumption patterns may contain highly private information about the lifestyle. In the literature [13], a new concept called Bayesian Differential Privacy is developed for static data to solve this problem. Bayesian differential privacy considers the underlying probability distribution of the data, and attempts to guarantee privacy for data sets that are far apart.

In this study, we consider a prior distribution for the signal that we want to keep secret and introduce Bayesian differential privacy for linear dynamical systems. Similar to the conventional differential privacy cases [9], we consider a mechanism where stochastic noise is added to the output data. Note that applying a large noise increases the privacy protection level, but decreases the information usefulness [14]. In Theorem 7 below, a lower bound of noise scale to guarantee a prescribed Bayesian differential privacy level will be derived. Other properties including the relation to the conventional case are investigated based on this. The rest of this paper is organized as follows. In Section II, we introduce differential privacy for dynamical systems. In Section III, we propose Bayesian differential privacy for dynamical systems and derive a sufficient condition for added noise to achieve its privacy guarantee. In Section IV, considering the trade-off between privacy and information utility, we derive the Gaussian noise with the minimum energy while guaranteeing the Bayesian differential privacy. In Section V, the usefulness of Bayesian differential privacy is described via a numerical example. Some concluding remarks are given in Section VI.

Notations

The sets of real numbers and nonnegative integers are denoted by {\mathbb{R}} and +{\mathbb{Z}}_{+}, respectively. The imaginary unit is denoted by j{\rm j}. For vectors x1,,xmnx_{1},\dots,x_{m}\in{\mathbb{R}}^{n}, a collective vector [x1xm]nm[x_{1}^{\top}\cdots x_{m}^{\top}]^{\top}\in{\mathbb{R}}^{nm} is described by [x1;;xm][x_{1};\cdots;x_{m}] for the sake of simplicity of description. For a sequence u(t)n,t=0,1,,Tu(t)\in{\mathbb{R}}^{n},\ t=0,1,\ldots,T, a collective vector is denoted by UT[u(0);;u(T)](T+1)nU_{T}\coloneqq[u(0);\cdots;u(T)]\in{\mathbb{R}}^{(T+1)n} using a capital alphabet. For a square matrix An×nA\in{\mathbb{R}}^{n\times n}, its determinant is denoted by det(A)\mathrm{det}(A), and when its eigenvalues are real, its maximum and minimum eigenvalues are denoted by λmax(A)\lambda_{\max}(A) and λmin(A)\lambda_{\min}(A), respectively. We write A0A\succ 0 (resp. A0A\succeq 0) if AA is positive definite (resp. semidefinite). For A0A\succeq 0, the principal square root of AA is denoted by A1/2A^{1/2}. The identity matrix of size nn is denoted by InI_{n}. The subscript nn is omitted when it is clear from the context. The Euclidean norm of a vector xnx\in{\mathbb{R}}^{n} is denoted by |x||x|, and its weighted norm with A0A\succ 0 is denoted by |x|A(xAx)1/2|x|_{A}\coloneqq(x^{\top}Ax)^{1/2}. The indicator function of a set SnS\subset{\mathbb{R}}^{n} is denoted by 1S1_{S}, i.e., 1S(x)=11_{S}(x)=1 if xSx\in S, and 0, otherwise. For a topological space XX, the Borel algebra on XX is denoted by (X){\mathcal{B}}(X). Fix some complete probability space (Ω,,)(\Omega,{\mathcal{F}},{\mathbb{P}}), and let 𝔼{\mathbb{E}} be the expectation with respect to {\mathbb{P}}. For an n\mathbb{R}^{n}-valued random vector ww, w𝒩n(μ,Σ)w\sim{\mathcal{N}}_{n}(\mu,\Sigma) means that ww has a nondegenerate multivariate Gaussian distribution with mean μn\mu\in{\mathbb{R}}^{n} and covariance matrix Σ0\Sigma\succ 0. The so called 𝖰{\mathsf{Q}}-function is defined by 𝖰(c)12πcev22𝑑v{\mathsf{Q}}(c)\coloneqq\frac{1}{\sqrt{2\pi}}\int_{c}^{\infty}{\rm e}^{-\frac{v^{2}}{2}}dv, where 𝖰(c)<1/2{\mathsf{Q}}(c)<1/2 for c>0c>0, and 𝖱(ε,δ)(𝖰1(δ)+(𝖰1(δ))2+2ε)/2ε{\mathsf{R}}(\varepsilon,\delta)\coloneqq({\mathsf{Q}}^{-1}(\delta)+\sqrt{({\mathsf{Q}}^{-1}(\delta))^{2}+2\varepsilon})/2\varepsilon. The gamma function is denoted by Γ()\Gamma(\cdot). A random variable zz is said to have a χ2\chi^{2} distribution with kk degrees of freedom, denoted by zχk2z\sim\chi^{2}_{k}, if its distribution has the following probability density:

p(z;k)=zk/21ez/2 2k/2Γ(k/2),z0,k{1,2,}.\displaystyle p(z;k)=\frac{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{z}}^{k/2-1}\mathrm{e}^{-{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{z}}/2}}{\,2^{k/2}\Gamma(k/2)},\quad z\geq 0,\ k\in\{1,2,\ldots\}.

II Conventional differential Privacy for dynamical systems

In this section, we briefly overview fundamental results on differential privacy for dynamical systems. Consider the following discrete-time linear system:

{x(t+1)=Ax(t)+Bu(t),y(t)=Cx(t)+Du(t),\displaystyle\left\{\begin{array}[]{l}x(t+1)=Ax(t)+Bu(t),\\ y(t)=Cx(t)+Du(t),\end{array}\right. (3)

for t+t\in{\mathbb{Z}}_{+}, where x(t)nx(t)\in{\mathbb{R}}^{n}, u(t)mu(t)\in{\mathbb{R}}^{m}, and y(t)qy(t)\in{\mathbb{R}}^{q} denote the state, input, and output, respectively, and An×nA\in{\mathbb{R}}^{n\times n}, Bn×mB\in{\mathbb{R}}^{n\times m}, Cq×nC\in{\mathbb{R}}^{q\times n}, and Dq×mD\in{\mathbb{R}}^{q\times m}. For simplicity, we assume x(0)=0x(0)=0 and the information to be kept secret is the input sequence UTU_{T} up to a finite time T+T\in{\mathbb{Z}}_{+}. For (3), the output sequence YT(T+1)qY_{T}\in{\mathbb{R}}^{(T+1)q} is described by

YT=NTUT,\displaystyle Y_{T}=N_{T}U_{T}, (4)

where NT(T+1)q×(T+1)mN_{T}\in{\mathbb{R}}^{(T+1)q\times(T+1)m} is

NT\displaystyle N_{T} =𝒯T(A,B,C,D)\displaystyle={\cal T}_{T}(A,B,C,D) (5)
[D00CBDCABCBD0CAT1BCAT2BCBD].\displaystyle\coloneqq\left[\begin{array}[]{cccccc}D&0&\cdots&\cdots&0\\ CB&D&\ddots&&\vdots\\ CAB&CB&D&\ddots&\vdots\\ \vdots&\vdots&\ddots&\ddots&0\\ CA^{T-1}B&CA^{T-2}B&\cdots&CB&D\end{array}\right]. (11)

To proceed with differential privacy analysis, we consider the output yw(t)y(t)+w(t)y_{w}(t)\coloneqq y(t)+w(t) after adding the noise w(t)qw(t)\in{\mathbb{R}}^{q}; see Fig. 1. From (4), Yw,T(T+1)qY_{w,T}\in{\mathbb{R}}^{(T+1)q} can be described by

Yw,T=NTUT+WT,\displaystyle Y_{w,T}=N_{T}U_{T}+W_{T}, (12)

which defines a mapping :(T+1)m×Ω(UT,ω)Yw,T(T+1)q.{\mathcal{M}}\colon{\mathbb{R}}^{(T+1)m}\times\Omega\ni(U_{T},\omega)\mapsto Y_{w,T}\in{\mathbb{R}}^{(T+1)q}.

Refer to caption
Figure 1: Mechanism with output noise.

In differential privacy analysis, this mapping is called a mechanism.

Next, the definition of the differential privacy is given. We begin with the definition of the data similarity:

Definition 1.

Given a positive definite matrix K(T+1)m×(T+1)mK\in{\mathbb{R}}^{(T+1)m\times(T+1)m}, a pair of input data (UT,UT)(T+1)m×(T+1)m(U_{T},U^{\prime}_{T})\allowbreak\in{\mathbb{R}}^{(T+1)m}\times{\mathbb{R}}^{(T+1)m} is said to belong to the binary relation KK-adjacency if

|UTUT|K1.\displaystyle|U_{T}-U^{\prime}_{T}|_{K}\leq 1. (13)

The set of all pairs of the input data that are KK-adjacent is denoted by AdjK\mathrm{Adj}_{K}.

This KK-adjacency is an extension of cc-adjacency, which corresponds to K=I/c2K=I/c^{2}, for the 22-norm in previous work [9]. Next, we describe the definition of (K,ε,δ)(K,\varepsilon,\delta)-differential privacy in dynamical systems in the same way as for static data [9, Definition 2.4].

Definition 2 ((K,ε,δ)(K,\varepsilon,\delta)-differential privacy).

Given ε>0\varepsilon>0 and δ0\delta\geq 0, the mechanism (12) is said to be (K,ε,δ)(K,\varepsilon,\delta)-differentially private ((K,ε,δ)(K,\varepsilon,\delta)-DP) at a finite time instant T+T\in{\mathbb{Z}}_{+}, if

[NTUT+WT𝒮]\displaystyle{\mathbb{P}}[N_{T}U_{T}+W_{T}\in{\mathcal{S}}]
eε[NTUT+WT𝒮]+δ,𝒮((T+1)q)\displaystyle\leq{\rm e}^{\varepsilon}{\mathbb{P}}[N_{T}U^{\prime}_{T}+W_{T}\in{\mathcal{S}}]+\delta,\quad\forall{\mathcal{S}}\in{\mathcal{B}}\left({\mathbb{R}}^{(T+1)q}\right) (14)

for any (UT,UT)AdjK(U_{T},U^{\prime}_{T})\in{\rm Adj}_{K}.

Suppose that the output sequence Yw,TY_{w,T} and the state equation (3) are available for an attacker trying to estimate the value of the input sequence UTU_{T}. Differential privacy requires the output sequence statistics are close enough at least for adjacent data pairs. A sufficient condition for the mechanism induced by Gaussian noise to be (ε,δ)(\varepsilon,\delta)-differentially private under cc-adjacency is derived in [9, Theorem 2.6]. This result can be straightforwardly extended as follows:

Theorem 3.

The Gaussian mechanism (12) induced by WT𝒩(T+1)q(μw,Σw)W_{T}\sim{\mathcal{N}}_{(T+1)q}(\mu_{w},\Sigma_{w}) is (K,ε,δ)(K,\varepsilon,\delta)-differentially private at a finite time T+T\in{\mathbb{Z}}_{+} with ε>0\varepsilon>0 and 1/2>δ>01/2>\delta>0, if the covariance matrix Σw0\Sigma_{w}\succ 0 is chosen such that

λmax1/2(𝒪K,Σw,T)𝖱(ε,δ),\displaystyle\lambda_{\max}^{-1/2}\left({\mathcal{O}}_{K,\Sigma_{w},T}\right)\geq{\mathsf{R}}(\varepsilon,\delta), (15)

where

𝒪K,Σw,TK1/2NTΣw1NTK1/2.\displaystyle{\mathcal{O}}_{K,\Sigma_{w},T}\coloneqq K^{-1/2}N_{T}^{\top}\Sigma_{w}^{-1}N_{T}K^{-1/2}. (16)

Larger noise WTW_{T} and lower threshold for the adjacency in the sense of Σ\Sigma and KK, respectively, make the left-hand side of (15) larger. This implies that differential privacy is ensured for smaller ε\varepsilon and δ\delta because 𝖱{\mathsf{R}} is a decreasing function. Further insight of KK will be revealed in Corollary 8 below.

III Bayesian differential privacy for dynamical systems

III-A Formulation

In Definition 2, the difficulty of distinguishing data pairs whose KK-weighted distance is larger than a threshold 11 is not taken into account. Note that there is no design guideline for KK. In this section, we introduce Bayesian differential privacy for dynamical systems. To this end, we assume the following availability of the prior distribution of the data to be protected, and provide a privacy guarantee that takes into account the discrimination difficulty of data pairs based on the prior.

Assumption 4.

The input data to the mechanism, UTU_{T}, is an (T+1)m{\mathbb{R}}^{(T+1)m}-valued random variable with distribution UT{\mathbb{P}}_{U_{T}}. In addition, one can use the prior UT{\mathbb{P}}_{U_{T}} to design a mechanism.

The following is a typical example where a private input signal is a realization of a random variable.

Example 5.

Suppose that the input data u(t)u(t) to be protected is the reference for tracking control; see [14]. In many applications, tracking to the reference signal over specified frequency ranges is required. Such a control objective can be represented by filtering white noise ξ(t)\xi(t). To be more precise, we assume u(t)=r(t)u(t)=r(t) is generated by

xr(t+1)=Arxr(t)+Brξ(t),xr(0)=0,\displaystyle x_{r}(t+1)=A_{r}x_{r}(t)+B_{r}\xi(t),\ x_{r}(0)=0, (17)
r(t)=Crxr(t)+Drξ(t),\displaystyle r(t)=C_{r}x_{r}(t)+D_{r}\xi(t), (18)
ξ(t)𝒩(0,I),t{0,1,,T},\displaystyle\xi(t)\sim{\mathcal{N}}(0,I),\quad t\in\{0,1,\ldots,T\}, (19)

where Arl×l,Brl×m,Crm×l,Drm×mA_{r}\in{\mathbb{R}}^{l\times l},\ B_{r}\in{\mathbb{R}}^{l\times m},\ C_{r}\in{\mathbb{R}}^{m\times l},\ D_{r}\in{\mathbb{R}}^{m\times m}. The power spectrum of u(t)u(t) is characterized by the frequency transfer function

Gr(ejλ):=Cr(ejλIAr)1Br+Dr,λ[0,π).\displaystyle G_{r}({\rm e}^{{\rm j}\lambda}):=C_{r}({\rm e}^{{\rm j}\lambda}I-A_{r})^{-1}B_{r}+D_{r},\ \lambda\in[0,\pi). (20)

In this case,

r(t)={Drξ(0)(t=0),Drξ(t)+j=1tCrArj1Brξ(tj)(t1),\displaystyle r(t)=\begin{cases}D_{r}\xi(0)&(t=0),\\ D_{r}\xi(t)+\sum_{j=1}^{t}C_{r}A_{r}^{j-1}B_{r}\xi(t-j)&(t\geq 1),\end{cases} (21)

and the prior distribution is given by UT𝒩(0,ΣU)U_{T}\sim{\cal N}(0,\Sigma_{U}) with

ΣU:=ΞTΞT,\displaystyle\Sigma_{U}:=\Xi_{T}\Xi_{T}^{\top}, (22)
ΞT:=𝒯T(Ar,Br,Cr,Dr).\displaystyle\Xi_{T}:={\cal T}_{T}(A_{r},B_{r},C_{r},D_{r}). (23)

Note that the step reference signal whose value obeys r(t)r¯𝒩(0,Σs)r(t)\equiv\bar{r}\sim\mathcal{N}(0,\Sigma_{\rm s}) can be modeled by setting Ar=Cr=I,Br=Dr=0A_{r}=C_{r}=I,\ B_{r}=D_{r}=0 with xr(0)𝒩(0,Σs)x_{r}(0)\sim\mathcal{N}(0,\Sigma_{\rm s}) in (17), (18). This corresponds to the case where the initial state is the private information rather than the input sequence UTU_{T} as discussed in [9].

Then, based on the Bayesian differential privacy for static data [13], we define (UT,γ,ε,δ)({\mathbb{P}}_{U_{T}},\gamma,\varepsilon,\delta)-Bayesian differential privacy, which is an extension of differential privacy for dynamical systems.

Definition 6 ((UT,γ,ε,δ)({\mathbb{P}}_{U_{T}},\gamma,\varepsilon,\delta)-Bayesian differential privacy).

Assume that the random variables UT,UTU_{T},U^{\prime}_{T} are independent and both follow the distribution UT{\mathbb{P}}_{U_{T}}. Given 1γ0,ε>01\geq\gamma\geq 0,\ \varepsilon>0 and δ0\delta\geq 0, the mechanism (12) is said to be (UT,γ,ε,δ)({\mathbb{P}}_{U_{T}},\gamma,\varepsilon,\delta)-Bayesian differentially private ((UT,γ,ε,δ)({\mathbb{P}}_{U_{T}},\gamma,\varepsilon,\delta)-BDP) at a finite time instant T+T\in{\mathbb{Z}}_{+}, if

[[NTUT+WT𝒮UT]\displaystyle{\mathbb{P}}\Bigl{[}{\mathbb{P}}[N_{T}U_{T}+W_{T}\in{\mathcal{S}}\mid U_{T}]
eε[NTUT+WT𝒮UT]+δ]γ,𝒮((T+1)q).\displaystyle\leq\mathrm{e}^{\varepsilon}{\mathbb{P}}[N_{T}U^{\prime}_{T}+W_{T}\in{\mathcal{S}}\mid U^{\prime}_{T}]+\delta\Bigr{]}\geq\gamma,\quad\forall{\mathcal{S}}\in{\mathcal{B}}({\mathbb{R}}^{(T+1)q}). (24)

In (24), the outer (resp. inner) \mathbb{P} is taken with respect to (UT,UT)(U_{T},U^{\prime}_{T}) (resp. WTW_{T}). Roughly speaking, the definition of BDP is that the probability that the mechanism satisfies (K,ε,δ)(K,\varepsilon,\delta)-DP is greater than or equal to γ\gamma. Note that this definition places no direct restriction on the distance between a pair of input data UT,UTU_{T},U^{\prime}_{T}.

III-B Sufficient condition for noise scale

It is desirable that the added noise ww is small to retain the data usefulness; see e.g., Section V. The following theorem gives a sufficient condition for noise scale to guarantee (UT,γ,ε,δ)({\mathbb{P}}_{U_{T}},\gamma,\varepsilon,\delta)-Bayesian differential privacy.

Theorem 7.

Suppose that the prior distribution UT{\mathbb{P}}_{U_{T}} of UTU_{T} is 𝒩(T+1)m(0,Σ){\mathcal{N}}_{(T+1)m}(0,\Sigma). The Gaussian mechanism (12) induced by WT𝒩(T+1)q(μw,Σw)W_{T}\sim{\mathcal{N}}_{(T+1)q}(\mu_{w},\Sigma_{w}) is (UT,γ,ε,δ)({\mathbb{P}}_{U_{T}},\gamma,\varepsilon,\delta)-Bayesian differentially private at a finite time T+T\in{\mathbb{Z}}_{+} with 1γ01\geq\gamma\geq 0, ε>0\varepsilon>0, and 1/2>δ>01/2>\delta>0, if the covariance matrix Σw0\Sigma_{w}\succ 0 is chosen such that

λmax1/2(𝒪Σ,Σw,T)c(γ,T)𝖱(ε,δ)\displaystyle\lambda_{\max}^{-1/2}({\mathcal{O}}_{\Sigma,\Sigma_{w},T})\geq c(\gamma,T){\mathsf{R}}(\varepsilon,\delta) (25)

where 𝒪Σ,Σw,T{\mathcal{O}}_{\Sigma,\Sigma_{w},T} is defined by

𝒪Σ,Σw,T\displaystyle{\mathcal{O}}_{\Sigma,\Sigma_{w},T} Σ1/2NTΣw1NTΣ1/2\displaystyle\coloneqq\Sigma^{1/2}N_{T}^{\top}\Sigma_{w}^{-1}N_{T}\Sigma^{1/2} (26)

and c(γ,T)c(\gamma,T) is the unique c>0c>0 that satisfies

γ\displaystyle\gamma =0c2/212(T+1)m/2Γ((T+1)m/2)x(T+1)m/21ex/2𝑑x.\displaystyle=\int_{0}^{c^{2}/2}\dfrac{1}{2^{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{(T+1)m/2}}}\Gamma({\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{(T+1)m/2)}}}x^{{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{(T+1)m/2}}-1}\mathrm{e}^{-x/2}dx. (27)
Proof.

Using a similar argument as in the proof for [9, Theorem 2.6], for any fixed U,U(T+1)mU,U^{\prime}\in{\mathbb{R}}^{(T+1)m}, one has

[NTUT+WT𝒮UT=U]\displaystyle{\mathbb{P}}[N_{T}U_{T}+W_{T}\in{\mathcal{S}}\mid U_{T}=U]
eε[NTUT+WT𝒮UT=U]\displaystyle\leq{\rm e}^{\varepsilon}{\mathbb{P}}[N_{T}U^{\prime}_{T}+W_{T}\in{\mathcal{S}}\mid U^{\prime}_{T}=U^{\prime}]
+[Zεh12hUT=U,UT=U]\displaystyle+{\mathbb{P}}\left[Z\geq{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{\varepsilon h-\frac{1}{2h}}}\mid U_{T}=U,\ U^{\prime}_{T}=U^{\prime}\right]

where

h|YY|Σw11,YNTUT,YNTUT,\displaystyle h\coloneqq|Y-Y^{\prime}|_{\Sigma^{-1}_{w}}^{-1},\ Y\coloneqq N_{T}U_{T},\ Y^{\prime}\coloneqq N_{T}U^{\prime}_{T},

and Z𝒩(0,1)Z\sim{\mathcal{N}}(0,1). Then, the mechanism is (UT,γ,ε,δ)({\mathbb{P}}_{U_{T}},\gamma,\varepsilon,\delta)-Bayesian differentially private, if 𝖰(εh12h)δ{\mathsf{Q}}({\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{\varepsilon h-\frac{1}{2h}}})\leq\delta with probability at least γ\gamma, i.e.,

[h𝖱(ε,δ)]γ.\displaystyle{\mathbb{P}}[h\geq{\mathsf{R}}(\varepsilon,\delta)]\geq\gamma. (28)

The inequality (28) holds if (25) is satisfied. This is because

h2\displaystyle h^{-2} =|NT(UTUT)|Σw12\displaystyle=|N_{T}(U_{T}-U^{\prime}_{T})|^{2}_{\Sigma^{-1}_{w}}
=(UTUT)Σ1/2𝒪Σ,Σw,TΣ1/2(UTUT)\displaystyle=(U_{T}-U^{\prime}_{T})^{\top}\Sigma^{-1/2}{\mathcal{O}}_{\Sigma,\Sigma_{w},T}\Sigma^{-1/2}(U_{T}-U^{\prime}_{T})
|Σ1/2(UTUT)|2λmax(𝒪Σ,Σw,T),\displaystyle\leq|\Sigma^{-1/2}(U_{T}-U^{\prime}_{T})|^{2}\lambda_{\max}({\mathcal{O}}_{\Sigma,\Sigma_{w},T}),

and then, from the fact that |Σ1/2(UTUT)|2/2|\Sigma^{-1/2}(U_{T}-U^{\prime}_{T})|^{2}/2 follows χ2\chi^{2} distribution with (T+1)m(T+1)m degrees of freedom and the definition of c(γ,T)c(\gamma,T), |Σ1/2(UTUT)|2c(γ,T)2|\Sigma^{-1/2}(U_{T}-U^{\prime}_{T})|^{2}\leq c(\gamma,T)^{2} with probability γ\gamma. ∎

In order to clarify the connection between conventional and Bayesian DP, it is worthwhile comparing Theorems 3 and 7. Bayesian differential privacy with the prior distribution UT𝒩(T+1)m(0,Σ){\mathbb{P}}_{U_{T}}\sim{\mathcal{N}}_{(T+1)m}(0,\Sigma) corresponds to differential privacy with an adjacency weight KK.

Corollary 8.

Suppose that the prior distribution UT{\mathbb{P}}_{U_{T}} of UTU_{T} is 𝒩(T+1)m(0,Σ){\mathcal{N}}_{(T+1)m}(0,\Sigma). Let a finite time T+T\in{\mathbb{Z}}_{+} with 1γ01\geq\gamma\geq 0, ε>0\varepsilon>0, and 1/2>δ>01/2>\delta>0 be given. Then, the Gaussian mechanism (12) induced by WT𝒩(T+1)q(μw,Σw)W_{T}\sim{\mathcal{N}}_{(T+1)q}(\mu_{w},\Sigma_{w}) is (UT,γ,ε,δ)({\mathbb{P}}_{U_{T}},\gamma,\varepsilon,\delta)-Bayesian differentially private at the time TT, if the mechanism is (K,ε,δ)(K,\varepsilon,\delta)-differentially private at the time TT with

K:=Σ1/c(γ,T)2\displaystyle K:=\Sigma^{-1}/c(\gamma,T)^{2} (29)

with c(γ,T)c(\gamma,T) defined in Theorem 7.

Proof.

Let us define Yw,T:=NTUT+WT,Yw,T:=NTUT+WT.Y_{w,T}:=N_{T}U_{T}+W_{T},\ Y^{\prime}_{w,T}:=N_{T}U^{\prime}_{T}+W_{T}. If the mechanism is (K,ε,δ)(K,\varepsilon,\delta)-differentially private with KK defined in (29),

[[Yw,T𝒮UT]eε[Yw,T𝒮UT]+δ]\displaystyle{\mathbb{P}}\Bigl{[}{\mathbb{P}}[Y_{w,T}\in{\mathcal{S}}\mid U_{T}]\leq\mathrm{e}^{\varepsilon}{\mathbb{P}}[Y^{\prime}_{w,T}\in{\mathcal{S}}\mid U^{\prime}_{T}]+\delta\Bigr{]}
[[Yw,T𝒮UT]\displaystyle\geq{\mathbb{P}}\Bigl{[}{\mathbb{P}}[Y_{w,T}\in{\mathcal{S}}\mid U_{T}]\leq
eε[Yw,T𝒮UT]+δ||UTUT|Σ1c(γ,T)]\displaystyle\qquad\qquad\mathrm{e}^{\varepsilon}{\mathbb{P}}[Y^{\prime}_{w,T}\in{\mathcal{S}}\mid U^{\prime}_{T}]+\delta\ \big{|}\ |U_{T}-U^{\prime}_{T}|_{\Sigma^{-1}}\leq c(\gamma,T)\Bigr{]}
×[|UTUT|Σ1c(γ,T)].\displaystyle\times{\mathbb{P}}\bigl{[}|U_{T}-U^{\prime}_{T}|_{\Sigma^{-1}}\leq c(\gamma,T)\bigr{]}.

Note that it holds

[[Yw,T𝒮UT]\displaystyle\mathbb{P}\Bigl{[}\mathbb{P}[Y_{w,T}\in\mathcal{S}\mid U_{T}]\leq
eε[Yw,T𝒮UT]+δ||UTUT|Σ1c(γ,T)]=1,\displaystyle\qquad\mathrm{e}^{\varepsilon}\mathbb{P}[Y^{\prime}_{w,T}\in\mathcal{S}\mid U^{\prime}_{T}]+\delta\ \big{|}\ |U_{T}-U^{\prime}_{T}|_{\Sigma^{-1}}\leq c(\gamma,T)\Bigr{]}=1,

since the mechanism satisfies

[Yw,T𝒮UT]eε[Yw,T𝒮UT]+δ\mathbb{P}[Y_{w,T}\in\mathcal{S}\mid U_{T}]\leq\mathrm{e}^{\varepsilon}\mathbb{P}[Y^{\prime}_{w,T}\in\mathcal{S}\mid U^{\prime}_{T}]+\delta

whenever |UTUT|Σ1c(γ,T)|U_{T}-U^{\prime}_{T}|_{\Sigma^{-1}}\leq c(\gamma,T) (i.e. (UT,UT)AdjK(U_{T},U^{\prime}_{T})\in\mathrm{Adj}_{K}). Next, by definition of c(γ,T)c(\gamma,T), we have

[|UTUT|Σ1c(γ,T)]=γ.\mathbb{P}\bigl{[}|U_{T}-U^{\prime}_{T}|_{\Sigma^{-1}}\leq c(\gamma,T)\bigr{]}=\gamma.

Consequently, we obtain the desired result.

It should be emphasized that such a simple relation is obtained since the prior is Gaussian and the system is linear.

III-C Asymptotic analysis

For the conventional DP, it is known that when the system (3) is asymptotically stable, one can design a Gaussian noise which makes the induced mechanism differentially private for any time horizon TT [9, Corollary 2.9]. This is because, for an asymptotically stable system, the incremental gain from |UTUT||U_{T}-U^{\prime}_{T}| to |YTYT||Y_{T}-Y^{\prime}_{T}| is bounded by its HH^{\infty}-norm for any TT, and by the definition of DP, the distance |UTUT||U_{T}-U^{\prime}_{T}| is also bounded by a predetermined threshold. That is, even when the horizon of the data to be protected becomes longer, the distance between data sets where their indistinguishability is guaranteed is fixed.

On the other hand, for the proposed BDP, as TT becomes larger, |UTUT||U_{T}-U^{\prime}_{T}| tends to take larger values according to the prior UT{\mathbb{P}}_{U_{T}}. Consequently, to achieve BDP for a large time horizon TT, large noise is required. To see this from Theorem 7, c(γ,T)>0c(\gamma,T)>0 is plotted in Fig. 2 as a function of TT. As can be seen, as TT increases, c(γ,T)c(\gamma,T) becomes large, and therefore, from (25), the scale parameter Σw\Sigma_{w} of the noise is required to be large to guarantee BDP. This fact suggests that the privacy requirement of BDP (with fixed ε,γ,δ\varepsilon,\gamma,\delta) for the long (possibly infinite) horizon case is too strong. This issue will be resolved by an appropriate scaling with TT to quantify the long-time average privacy.

Refer to caption
Figure 2: Graph of c(γ,T)c(\gamma,T).

IV Design of mechanism

To motivate additional analysis in this section, let us consider a feedback interconnection of the plant 𝒫\cal P and the controller 𝒞\cal C:

𝒫:{xp(t+1)=Apxp(t)+Bpup(t),yp(t)=Cpxp(t),\displaystyle{\cal P}:\left\{\begin{array}[]{l}x_{p}(t+1)=A_{p}x_{p}(t)+B_{p}u_{p}(t),\\ y_{p}(t)=C_{p}x_{p}(t),\\ \end{array}\right.
𝒞:{xc(t+1)=Acxc(t)+Bce(t),up(t)=Ccxc(t),\displaystyle{\cal C}:\left\{\begin{array}[]{l}x_{c}(t+1)=A_{c}x_{c}(t)+B_{c}e(t),\\ u_{p}(t)=C_{c}x_{c}(t),\end{array}\right.

where the control objective is to make the tracking error e(t):=r(t)yp(t)e(t):=r(t)-y_{p}(t) small, and its private information is the reference signal r(t)r(t). The attacker, who can access to the output yp(t)y_{p}(t), attempts to estimate r(t)r(t). To prevent this inference, we add noise vv to rr, which leads to the following closed loop dynamics:

{x¯(t+1)=A¯x¯(t)+B¯(r(t)+v(t)),yp(t)=C¯x¯(t),e(t)=r(t)C¯x¯(t),\displaystyle\left\{\begin{array}[]{l}\bar{x}(t+1)=\bar{A}\bar{x}(t)+\bar{B}(r(t)+v(t)),\\ y_{p}(t)=\bar{C}\bar{x}(t),\\ e(t)=r(t)-\bar{C}\bar{x}(t),\\ \end{array}\right. (33)

where x¯(t):=[xp(t);xc(t)]\bar{x}(t):=[x_{p}(t);x_{c}(t)] and

A¯:=[ApBpCcBcCpAc],B¯:=[0Bc],C¯:=[Cp0].\displaystyle\bar{A}:=\begin{bmatrix}A_{p}&B_{p}C_{c}\\ {\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{-B_{c}C_{p}}}&A_{c}\end{bmatrix},\ \bar{B}:=\begin{bmatrix}0\\ {\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{B_{c}}}\end{bmatrix},\ \bar{C}:=\begin{bmatrix}C_{p}&0\end{bmatrix}.

Suppose that the distribution of VTV_{T} is given by 𝒩(0,Σv){\cal N}(0,\Sigma_{v}). Then, larger noise vv fluctuates ee more so that the variance of ET:=[e(0);;e(T)]E_{T}:=[e(0);\cdots;e(T)] is given by

ΘTΣvΘT\displaystyle\Theta_{T}\Sigma_{v}\Theta_{T}^{\top} (34)

with ΘT:=𝒯T(A¯,B¯,C¯,0)\Theta_{T}:={\cal T}_{T}(\bar{A},\bar{B},-\bar{C},0).

IV-A Minimization of the noise

The expression (34) motivates us to seek the Gaussian noise with the minimum energy among those satisfying the sufficient condition (25) for Bayesian differential privacy derived by Theorem 7. More specifically, we consider the following optimization problem with the covariance matrix of Gaussian noise Σw0\Sigma_{w}\succ 0 as the decision variable.

Problem 9.
minimizeΣw0\displaystyle\underset{\Sigma_{w}\succ 0}{\text{minimize}} Tr(Σw)\displaystyle{\rm Tr}(\Sigma_{w}) (35)
subject to λmax(Σ1/2NTΣw1NTΣ1/2)1c(γ,T)2𝖱(ε,δ)2.\displaystyle\lambda_{\max}(\Sigma^{1/2}N_{T}^{\top}\Sigma_{w}^{-1}N_{T}\Sigma^{1/2})\leq\frac{1}{c(\gamma,T)^{2}{\mathsf{R}}(\varepsilon,\delta)^{2}}. (36)

The constraint (36) is an inequality that is equivalent to the inequality (25). Under certain assumptions, this solution can be obtained as follows.

Theorem 10.

Assume that NTN_{T} is full row rank. The optimal solution to problem 9 is Σwc(γ,T)2𝖱(ε,δ)2NTΣNT\Sigma_{w}^{*}\coloneqq c(\gamma,T)^{2}{\mathsf{R}}(\varepsilon,\delta)^{2}N_{T}\Sigma N_{T}^{\top}.

Proof.

Denote 𝖭:=c(γ,T)𝖱(ε,δ)NTΣ1/2{\sf N}:=c(\gamma,T){\mathsf{R}}(\varepsilon,\delta)N_{T}\Sigma^{1/2} so that Σw=𝖭𝖭\Sigma_{w}^{*}={\sf N}{\sf N}^{\top}. Then, (36) is equivalent to I𝖭Σw1𝖭0I-{\sf N}^{\top}\Sigma_{w}^{-1}{\sf N}\succeq 0. By the Schur complement,

[Σw𝖭𝖭I]0.\displaystyle\begin{bmatrix}\Sigma_{w}&{\sf N}\\ {\sf N}^{\top}&I\end{bmatrix}\succeq 0. (37)

This implies ΣwΣw\Sigma_{w}\succeq\Sigma_{w}^{*}, and consequently Tr(Σw)Tr(Σw){\rm Tr}(\Sigma_{w})\geq{\rm Tr}(\Sigma_{w}^{*}). ∎

The obtained optimal solution Σw\Sigma_{w}^{*} is a constant multiple of the covariance matrix of the distribution of the output data YT=NTUTY_{T}=N_{T}U_{T} when the covariance matrix of the input data UTU_{T} is Σ\Sigma. This means that it is possible to efficiently conceal the input data from the output data by applying the noise having the same statistics (up to scaling) as the observed data.

IV-B Input noise mechanism

Refer to caption
Figure 3: Mechanism with input noise.

In this subsection, we study the case where noise is added to the input channel; see Fig. 3. Consider the following system with input noise:

{x(t+1)=Ax(t)+B(u(t)+v(t)),yv(t)=Cx(t)+D(u(t)+v(t)).\displaystyle\left\{\begin{array}[]{l}x(t+1)=Ax(t)+B(u(t)+v(t)),\\ y_{v}(t)=Cx(t)+D(u(t)+v(t)).\end{array}\right. (40)

As in the aforementioned section, we assume x(0)=0x(0)=0. The output sequence Yv,T=[yv(0);;yv(T)]Y_{v,T}=[y_{v}(0);\cdots;y_{v}(T)] can be described as

Yv,T=NTUT+NTVT.\displaystyle Y_{v,T}=N_{T}U_{T}+N_{T}V_{T}. (41)

For the system in (3), adding noise VTV_{T} to the input channel is equivalent to adding noise WT=NTVTW_{T}=N_{T}V_{T} to the output channel. For simplicity, we assume that NTN_{T} is square and nonsingular; this can be relaxed as in [9, Corollary 2.16]. From Theorem 7, we obtain the following corollary.

Corollary 11.

Suppose that the prior distribution UT{\mathbb{P}}_{U_{T}} of UTU_{T} is 𝒩(T+1)m(0,Σ){\mathcal{N}}_{(T+1)m}(0,\Sigma). The Gaussian mechanism (41) induced by VT𝒩(T+1)q(μv,Σv)V_{T}\sim{\mathcal{N}}_{(T+1)q}(\mu_{v},\Sigma_{v}) is (UT,γ,ε,δ)({\mathbb{P}}_{U_{T}},\gamma,\varepsilon,\delta)-Bayesian differentially private at a finite time T+T\in{\mathbb{Z}}_{+} with 1γ01\geq\gamma\geq 0, ε>0\varepsilon>0, and 1/2>δ>01/2>\delta>0, if the covariance matrix Σv0\Sigma_{v}\succ 0 is chosen such that

λmin1/2(Σ1/2ΣvΣ1/2)c(γ,T)𝖱(ε,δ)\displaystyle\lambda_{\min}^{1/2}(\Sigma^{-1/2}\Sigma_{v}\Sigma^{-1/2})\geq c(\gamma,T){\mathsf{R}}(\varepsilon,\delta) (42)

with c(γ,T)>0c(\gamma,T)>0 defined in Theorem 7.

Proof.

The desired result is a straightforward consequence of Theorem 7. ∎

In [9, Corollary 2.16], a sufficient condition for (ε,δ)(\varepsilon,\delta)-differential privacy in the sense of Definition 2 is given. The result concludes that the differential privacy level for the input noise mechanism does not depend on the system itself. Similarly, (42) does not depend on the system matrices in (3) either. The difference for the Bayesian case is that (42) depends on the covariance Σ\Sigma of the prior distribution of the signals to be protected.

Remark 12.

It is clear from Corollary 11 and Theorem 10 that the minimum energy Gaussian noise that satisfies the sufficient condition for privacy guarantee (42) for the input noise mechanism can be easily obtained by

Σv=c(γ,T)2𝖱(ε,δ)2Σ.\displaystyle\Sigma^{*}_{v}=c(\gamma,T)^{2}{\mathsf{R}}(\varepsilon,\delta)^{2}\Sigma. (43)

This characterization allows the natural interpretation that large noise is needed to protect large inputs; see also the next section.

V Numerical example

Consider the feedback system in Fig. 4, where the plant and controller in (33) are given by

Ap\displaystyle A_{p} =[1.20.510],Bp=[0.3,0],Cp=[0.2,0],\displaystyle=\left[\begin{array}[]{cc}1.2&-0.5\\ 1&0\end{array}\right],\ B_{p}=[-0.3,0]^{\top},\ C_{p}=[0.2,0], (46)
Ac\displaystyle A_{c} =[1100.1],Bc=[0,1],Cc=[1.5,0].\displaystyle=\left[\begin{array}[]{cc}1&1\\ 0&0.1\end{array}\right],\ B_{c}=[0,{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{-1}}]^{\top},\ C_{c}=[{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{1.5}},0]. (49)

The integral property of the controller enhances the low-frequency tracking performance. The Bode gain diagram is shown in Fig. 5. Suppose that the reference rr is the signals to be protected, and it is public information that its spectrum is concentrated over the frequency range below 3×102rad/s3\times 10^{-2}\ {\rm rad/s}. To represent this prior information, we took the frequency model for rr as in Example 5, which is set to be a lowpass filter (generated by lowpass(xi, 3e-2) in MATLAB). Recall that UT𝒩(0,ΣU)U_{T}\sim{\mathcal{N}}(0,\Sigma_{U}) with (22) and (23).

Refer to caption
Figure 4: Feedback system with input noise mechanism
Refer to caption
Figure 5: Bode gain diagram for the frequency domain reference model GrG_{r} and the transfer functions for the feedback system (33) with (46), (49).
Refer to caption
Figure 6: Reference signal r(t)r(t) and plant output yp(t)y_{p}(t) for three mechanisms.

We design input noise to make the system Bayesian differentially private for γ=0.5,T=100,ε=100,δ=0.1\gamma=0.5,\ T=100,\ \varepsilon=100,\ \delta=0.1. This leads to

c(γ,T)=14.1657,𝖱(ε,δ)=0.0774.c(\gamma,T)=14.1657,\ {\mathsf{R}}(\varepsilon,\delta)=0.0774. (50)

In what follows we compare the following three cases:

  • noise-free,

  • i.i.d. noise: VT𝒩(T+1)m(0,σiid2I){\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{V}}_{T}\sim{\mathcal{N}}_{(T+1)m}(0,\sigma_{\rm iid}^{2}I) with

    Tr(σiid2I)\displaystyle{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{{\rm Tr}(\sigma_{\rm iid}^{2}I)}} =c(γ,T)2𝖱(ε,δ)2λmax(ΣU)(T+1)m\displaystyle={\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{c(\gamma,T)^{2}{\mathsf{R}}(\varepsilon,\delta)^{2}\lambda_{\max}(\Sigma_{U})(T+1)m}}
    =121.604,\displaystyle={\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{121.604}}, (51)

    where σiid\sigma_{\rm iid} is the minimum σ\sigma satisfying (42), and

  • the minimum noise obtained in Theorem 10: VT𝒩(T+1)m(0,Σv{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{V}}_{T}\sim{\mathcal{N}}_{(T+1)m}(0,\Sigma_{v}^{*}), Σv:=c(γ,T)2𝖱(ε,δ)2ΣU\Sigma_{v}^{*}:=c(\gamma,T)^{2}{\mathsf{R}}(\varepsilon,\delta)^{2}\Sigma_{U} with

    Tr(Σv)=8.2574.\displaystyle{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{{\rm Tr}(\Sigma_{v}^{*})=8.2574}}. (52)

Fig. 6 shows the reference signal r(t)r(t) and plant output yp(t)y_{p}(t) for these three cases. It can be seen that the output error for the noise-free case is the smallest. This is because the (realized) trajectory of rr is fully utilized allowing for some possibility that information about rr may leak from ypy_{p}. On the other hand, the other two cases guarantee the same level of Bayesian differential privacy. Note that the error fluctuation is suppressed in the minimum noise case. Statistically, the output fluctuation caused by the added noise vv can be evaluated by (34). The value Tr(ΘTΣvΘT)/(c(γ,T)𝖱(ε,δ))2{\rm Tr}(\Theta_{T}\Sigma_{v}\Theta_{T}^{\top})/(c(\gamma,T){\mathsf{R}}(\varepsilon,\delta))^{2} is given by Tr(ΘTΣUΘT)=8.1998{\rm Tr}(\Theta_{T}\Sigma_{U}\Theta_{T}^{\top})={\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{8.1998}} for the minimum noise case, which is smaller than λmax(ΣU)Tr(ΘTΘT)=55.4202\lambda_{\max}(\Sigma_{U}){\rm Tr}(\Theta_{T}\Theta_{T}^{\top})={\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}{55.4202}} for the i.i.d. case.

The interpretation is as follows: The i.i.d. noise has uniform frequency distribution, which implies it adds more out-of-band noise than the minimum one. However, this out-of-band component does not contribute to the protection of rr since it is easily distinguished from rr thanks to its prior information in Fig. 5. Nevertheless, this out-of-band noise largely degrades the tracking performance.

Remark 13.

The out-of-band noise as in the i.i.d. case is effective when the prior distribution of the signals to be protected is not public information. That is, this noise can prevent the attacker from inferring the prior distribution e.g., via the empirical Bayes.

Remark 14.

Lastly, we would like to note that for the Gaussian prior 𝒩(0,ΣU)\mathcal{N}(0,\Sigma_{U}) in the numerical example, a reference signal UTU^{\prime}_{T} that deviates significantly from mean 0 in the sense of |0UT|ΣU1|0-U^{\prime}_{T}|_{\Sigma_{U}^{-1}} can be seen as an outlier. Therefore, out-of-band signals having large values |UT|ΣU1|U^{\prime}_{T}|_{\Sigma_{U}^{-1}} can be regarded as outliers. Bayesian differentially private mechanism provides privacy guarantees not only for in-band signals but also for out-of-band ones. In particular, the parameter γ\gamma for Bayesian differential privacy determines the extent to which privacy guarantees can be provided for out-of-band signals.

VI Conclusion

In this study, we introduced Bayesian differential privacy for linear dynamical systems using prior distributions of input data to provide privacy guarantees even for input data pairs with large differences, and gave sufficient conditions to achieve it. Furthermore, we derived the minimum energy Gaussian noise that satisfies the condition. As noticed in Subsection III-C, any finite noise cannot guarantee the Bayesian differential privacy for the infinite horizon case. This issue will be addressed in future work.

Acknowledgment

This work was supported in part by JSPS KAKENHI under Grant Number JP21H04875.

References

  • [1] L. Sweeney, “k-anonymity: A model for protecting privacy,” Int. J. Unc. Fuzz. Knowl. Based Syst., vol. 10, no. 5, pp. 557–570, Oct. 2002.
  • [2] A. Machanavajjhala, D. Kifer, J. Gehrke, and M. Venkitasubramaniam, “l-diversity: Privacy beyond k-anonymity,” ACM Trans. Knowl. Discov. Data, vol. 1, no. 1, pp. 3–es, 2007.
  • [3] N. Li, T. Li, and S. Venkatasubramanian, “t-closeness: Privacy beyond k-anonymity and l-diversity,” in Proc. 23rd IEEE Int. Conf. Data Eng., Apr. 2007, pp. 106–115.
  • [4] C. Dwork, F. McSherry, K. Nissim, and A. Smith, “Calibrating noise to sensitivity in private data analysis,” in Proc. 3rd Theory Cryptography Conf., 2006, pp. 265–284.
  • [5] H. Sandberg, G. Dán, and R. Thobaben, “Differentially private state estimation in distribution networks with smart meters,” in Proc. 54th IEEE Conf. Decis. Control, Dec. 2015, pp. 4492–4498.
  • [6] F. K. Dankar and K. El Emam, “The application of differential privacy to health data,” in Proc. Int. Conf. Extending Database Technol., Mar. 2012, pp. 158–166.
  • [7] M. Yang, A. Margheri, R. Hu, and V. Sassone, “Differentially private data sharing in a cloud federation with blockchain,” IEEE Cloud Comput., vol. 5, no. 6, pp. 69–79, Nov. 2018.
  • [8] J. Le Ny and G. J. Pappas, “Differentially private filtering,” IEEE Trans. Automat. Control, vol. 59, no. 2, pp. 341–354, Feb. 2014.
  • [9] Y. Kawano and M. Cao, “Design of privacy-preserving dynamic controllers,” IEEE Trans. Automat. Control, vol. 65, no. 9, pp. 3863–3878, Sep. 2020.
  • [10] J. Cortés, G. E. Dullerud, S. Han, J. Le Ny, S. Mitra, and G. J. Pappas, “Differential privacy in control and network systems,” in Proc. 55th IEEE Conf. Decis. Control, Dec. 2016, pp. 4252–4272.
  • [11] V. Katewa, F. Pasqualetti, and V. Gupta, “On privacy vs cooperation in multi-agent systems,” Int. J. of Control, vol. 91, no. 7, pp. 1693–1707, 2018.
  • [12] K. Ito, Y. Kawano, and K. Kashima, “Privacy protection with heavy-tailed noise for linear dynamical systems,” Automatica, vol. 131, p. 109732, 2021.
  • [13] A. Triastcyn and B. Faltings, “Bayesian differential privacy for machine learning,” in Proc. Int. Conf. Mach. Learn., Nov. 2020.
  • [14] Y. Kawano, K. Kashima, and M. Cao, “Modular control under privacy protection: Fundamental trade-offs,” Automatica, vol. 127, p. 109518, 2021.