This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Unexpected Information Leakage of Differential Privacy Due to Linear Property of Queries

Wen Huang,Shijie Zhou,Yongjian Liao Wen Huang, Shijie Zhou and Yongjian Liao is with the School of Information and Software Engineering,University of Electronic Science and Technology of China,Chengdu, Sichuan, China.
E-mail: [email protected] received October 17, 2020; revised October 17, 2020.
Abstract

The differential privacy is a widely accepted conception of privacy preservation and the Laplace mechanism is a famous instance of differential privacy mechanisms to deal with numerical data. In this paper, we find that the differential privacy does not take liner property of queries into account, resulting in unexpected information leakage. In specific, the linear property makes it possible to divide one query into two queries such as q(D)=q(D1)+q(D2)q(D)=q(D_{1})+q(D_{2}) if D=D1D2D=D_{1}\cup D_{2} and D1D2=D_{1}\cap D_{2}=\emptyset. If attackers try to obtain an answer of q(D)q(D), they not only can issue the query q(D)q(D), but also can issue the q(D1)q(D_{1}) and calculate the q(D2)q(D_{2}) by themselves as long as they know D2D_{2}. By different divisions of one query, attackers can obtain multiple different answers for the query from differential privacy mechanisms. However, from attackers’ perspective and from differential privacy mechanisms’ perspective, the totally consumed privacy budget is different if divisions are delicately designed. The difference leads to unexpected information leakage because the privacy budget is the key parameter to control the amount of legally released information from differential privacy mechanisms. In order to demonstrate the unexpected information leakage, we present a membership inference attacks against the Laplace mechanism. Specifically, we propose a method to obtain multiple independent identical distribution samples of linear query’s answer under constrains of differential privacy. The proposed method is based on linear property and some background knowledge of attackers. When the background knowledge is enough, the proposed method can obtain enough number of samples from differential privacy mechanisms such that the totally consumed privacy budget can be unreasonably large. Based on obtained samples, a hypothesis test method is used to determine whether a target record is in a target data set.

Index Terms:
Laplace mechanism, membership inference attacks, differential privacy,linear property

1 Introduction

The differential privacy is state of the art conception for privacy preservation because it formalizes a strong privacy guarantee by a solid mathematical foundation. That is, even if there is only one different record in two data sets, it is hard to tell one data set from another data set. In differential privacy, the privacy guarantee is quantified by probability with which attackers can tell one data set from another data set. The probability is controlled by a parameter called privacy budget.

One of most important reasons for popularity of differential privacy is that it makes no assumptions about the background knowledge available to attackers [5]. That is, even if attackers know all data records except one, attackers cannot get much information of the record which attackers don’t know. The amount of information which attackers can extract is strictly controlled by the privacy budget. However, intuitively, if attackers have more background knowledge, attackers have stronger ability to extract information. For example, if attackers know a target record related to a female, attackers can filter all records related to male, resulting in high probability to get target information.

In this paper, we explore the transformation from background knowledge to consuming extra privacy budget of differential privacy mechanisms. That is, how attackers can consume unexpected privacy budget by its background knowledge. We find out that the transformation is possible because differential privacy mechanisms do not take one property of query into account, namely linear property.

The linear property makes it possible to divide one query into two queries. For example, the counting query is one class of linear queries. For query count({b1,b2,b3})count(\{b_{1},b_{2},b_{3}\}), the next equality holds.

count({b1,b2,b3}\displaystyle count(\{b_{1},b_{2},b_{3}\}
=\displaystyle= count({b1,b2})+count({b3})\displaystyle count(\{b_{1},b_{2}\})+count(\{b_{3}\})
=\displaystyle= count({b1,b3})+count({b2})\displaystyle count(\{b_{1},b_{3}\})+count(\{b_{2}\})

Suppose that attackers aim at the answer of count({b1,b2,b3})count(\{b_{1},b_{2},b_{3}\}) and the consumed privacy budget is ϵ\epsilon when a differential privacy mechanism responds one query every time.

Attackers can issue two different queries, namely count({b1,b2})count(\{b_{1},b_{2}\}) and count({b1,b3})count(\{b_{1},b_{3}\}), to a differential privacy mechanism and obtain two responses. From the differential privacy mechanism’s perspective, it answers two different queries and each query consumes privacy budget ϵ\epsilon.

However, if {b2,b3}\{b_{2},b_{3}\} is background knowledge of attackers, attackers can get two answers for the same query. In specific, by adding up the first response which is an answer of count({b1,b2})count(\{b_{1},b_{2}\}) and count({b3})count(\{b_{3}\}) which is computed by attackers themselves, attackers can obtain an answer of count({b1,b2,b3}count(\{b_{1},b_{2},b_{3}\}. By adding up the second response and count({b2})count(\{b_{2}\}), attackers can obtain another answer of count({b1,b2,b3}count(\{b_{1},b_{2},b_{3}\}. So, from attackers’ perspective, the differential privacy mechanism answers the same query two times and consumed privacy budget for each time is ϵ\epsilon. That is, the totally consumed privacy budget is 2ϵ2*\epsilon for the same query.

The possibility to obtain multiple answers for one query leads to risk of leaking membership information of records. In specific, attackers obtain multiple answers of query count({b1,b2,b3})count(\{b_{1},b_{2},b_{3}\}) and they know that the record b2b_{2} and b3b_{3} are in the target data set. If the record b1b_{1} is in the target data set, the mean of obtained samples is 3. Otherwise, the mean of samples is 2. By estimating the mean of samples, attackers can determine whether the record b1b_{1} is in the target data set.

In a word, if the background knowledge of attackers is enough, attackers can consume unreasonably large privacy budget in total while differential privacy mechanisms don’t beware, resulting in risk of information leakage. The risk may be a big challenge for applications of differential privacy mechanisms because influenced linear query such as the counting query and the sum query are elemental queries which differential privacy mechanisms can handle well in previous researches.

In order to demonstrate the unexpected information leakage of differential privacy mechanisms, we take advantage of the above vulnerability to construct a membership inference attacks method against the Laplace mechanism. Our main contributions are as follows

(1) We point out that the liner property of queries is a source of information leakage for differential privacy. To the best of our knowledge, this paper is the first one to discuss information leakage caused by linear property of queries.

(2) A method is proposed to obtain multiple i.i.d.(independent identical distribution) samples for linear query’s answer under constrains of differential privacy. In general cases, multiple i.i.d. samples for query’s answer cannot be obtained directly in differential privacy. However, the linear query has two special properties. The first property is linear property and the second property is that the sensitivity can be calculated easily. Based on these two properties, we propose a method to obtain multiple i.i.d. samples for linear query’s answer.

(3) A membership inference attacks method against the Laplace mechanism is proposed. The goal of membership inference attacks is to determine whether a target record is in a target data set. However, the differential privacy claims that even if attackers know all records except one, attackers cannot extract information of the record they don’t know. Therefore, the membership inference attacks is an appropriate way to demonstrate the information leakage of differential privacy. In the proposed attacks method, the decision is made by hypothesis test method which is based on multiple i.i.d. samples obtained.

1.1 Related Work

The differential privacy is state of the art conception for privacy preservation [9]. Since the conception of differential privacy is proposed, there have been some analyses about its information leakage. There are mainly two lines of papers. The first line is related to assumption of data generation. In differential privacy, records in data sets are assumed to be independent. But the assumption is not always true in real application scenarios [17]. For example, Bob or one of his 9 immediate family members may have contracted a highly infectious disease. The whole family is infected together or none of them is infected with high probability [18]. The correlation among family members makes it easier for attackers to infer whether Bob is infected by querying ”how many people are infected by the disease in Bob’s family”. There are some papers aiming at fixing the correlation problem. Lv et al. research the problem in scenarios of big data [13]. Wu et al. research the problem from the perspective of game theory [19]. Zhang et al. research the problem in machine learning [22]. The key of fixing correlation problem is to find an appropriate quantity to quantify correlation among data records. Then added noise of differential privacy mechanisms is calibrated by the quantity. Chen et al. quantify degree of correlation by Gaussian correlation model [4]. Zhu et al. propose conception of correlated sensitivity to quantify degree of correlation [23]. A strength of the line is that the unexpected information leakage is not related to certain differential privacy mechanisms. A weakness is that the unexpected information leakage is only related to correlated records not for all records.

The second line is related to the extensional conception of sensitivity of differential privacy. There is a difficult trade-off in differential privacy mechanisms, namely trade-off between magnitude of noise and data utility [7] [10]. In order to reduce the magnitude of noise, various alternative conceptions for sensitivity are proposed such as elastic sensitivity [11] and record sensitivity [9]. Some of these proposed conceptions are weak and lead to unexpected information leakage. The local sensitivity is an example [14]. The goal of local sensitivity is to release data with database-based additive noise. That is, the noise magnitude is determined not only by the queries which are to be released, but also by the data set itself. However, the noise calibrated by local sensitivity is too small, resulting in information leakage. For example, let fmed(x)=median(x1,x2,,xn)f_{med}(x)=median(x_{1},x_{2},\dots,x_{n}), where xix_{i} are sorted real number from bound interval such as [0,Λ][0,\Lambda]. If the fmed(x)f_{med}(x) is released with noise magnitude proportional to local sensitivity, then the probability of receiving a non-zero answer is zero when x1==xm+1=0x_{1}=\dots=x_{m+1}=0,xm+2==xn=Λx_{m+2}=\dots=x_{n}=\Lambda. Whereas the probability of receiving a non-zero answer is non-negligible when x1==xm=0x_{1}=\dots=x_{m}=0,xm+1==xn=Λx_{m+1}=\dots=x_{n}=\Lambda. Where nn is odd and m=n+12m=\frac{n+1}{2}. The analyses for extensional conceptions of sensitivity are difficult because how big the noise magnitude needs to be is hard to quantify.

The goal of membership inference attacks is to determine that a given data record is in a target data set or not. The attack is firstly demonstrated by Homer [8]. Their paper has so much big influence on privacy community that the USA National Institute of Health (NIH) switches policies in case of membership information leakage in process of releasing statistical data. The membership inference attacks has been valued since then. Shokri et al. quantitatively investigate membership inference attacks against machine learning models [16]. Rahman et al. analyze membership inference attacks against deep learning model which is protected by differential privacy mechanisms [15]. Yeom et al. investigate the relationship between overfitting and membership inference attacks, finding that ”overfitting is sufficient to allow an attacker to perform membership inference attack” [21]. Backes et al. explore the membership inference attacks against miRNA expression data set [3]. Almadhoun et al. discuss membership inference attacks against differential privacy mechanisms [1]. They take advantage of dependence of data records. Almadhoun et al also analyze an application of membership inference against genomic data sets [2].

In brief, the first line is about the special records which are correlated with each other and the second line is about trade-off between noise magnitude and data utility. Different from the mentioned two lines, we will discuss unexpected information leakage from a novel angle. In this paper, from the view of query functions, we show that differential privacy mechanisms do not take liner property of queries into account, resulting in unexpected information leakage.

1.2 Organization

In the next section, background knowledge will be introduced briefly, including the conception of differential privacy, the linear property and the hypothesis test. In the section 3, how to obtain multiple answers for linear queries is presented. In the section 4, the proposed membership inference attacks method will be presented and an instance of attack will be given. In section 5, experimental results will be showed. At last, the conclusion will be claimed.

2 background

The introduction of background knowledge is divided into three parts including basic conceptions of differential privacy, the linear property and the hypothesis test method.

2.1 Differential Privacy

The differential privacy is the most popular conception of privacy preservation. It formalizes a strong privacy guarantee: even if attackers know all data records except one, attackers cannot get much information of the record which attackers don’t know [6] [12]. To that end, differential privacy mechanisms guarantee that its outputs are insensitive to any particular data record. That is, presence or absence of any record has limited influence on outputs of differential privacy mechanisms. The presence or absence of one record in a data set is captured by a conception of neighboring databases. Specifically, databases DD and DD^{\prime} are neighboring databases denoted by DDD\sim D^{\prime} if there is only one different record between DD and DD^{\prime} denoted by d(D,D)=1d(D,D^{\prime})=1.

Definition 1(differential privacy) Any random mechanism M:DnRdM:D^{n}\to R^{d} preserves ϵ\epsilon-DP(differential privacy) if for any neighboring databases DD and DD^{\prime} such that d(D,D)=1d(D,D^{\prime})=1 and for all sets of possible output SS:

P{M(D)S}eϵP{M(D)S}\displaystyle P\{M(D)\in S\}\leq e^{\epsilon}P\{M(D^{\prime})\in S\}

Here, the ϵ\epsilon is the privacy budget. The privacy guarantee is quantified by the privacy budget. The smaller the privacy budget is, the stronger the privacy guarantee is.

The Laplace mechanism is a famous and foundational mechanism in the field of differential privacy. It is to deal with numerical data. The Laplace mechanism is based on a conception of global sensitivity. The global sensitivity quantifies the max difference of query’s answer, which is caused by presence or absence of one record.

Definition 2 (global sensitivity) For a given database DD and query qq, the global sensitivity denoted by ΔD\Delta D is

ΔD=maxDD|q(D)q(D)|\displaystyle\Delta D=\max\limits_{D\sim D^{\prime}}|q(D)-q(D^{\prime})|

The Laplace mechanism covers the difference by noise drawn from Laplace distribution with location parameter μ=0\mu=0 and scale parameter b=ΔDϵb=\frac{\Delta D}{\epsilon}.

Definition 3 (Laplace mechanism) For given database DD, query qq and privacy budget ϵ\epsilon, the output of Laplace mechanism is

M(q,D,ϵ)=q(D)+Lap(0,ΔDϵ)\displaystyle M(q,D,\epsilon)=q(D)+Lap(0,\frac{\Delta D}{\epsilon})

Here, the Lap(0,ΔDϵ)Lap(0,\frac{\Delta D}{\epsilon}) represents noise drawn from a Laplace distribution with location parameter 0 and scale parameter ΔDϵ\frac{\Delta D}{\epsilon}.

Differential privacy mechanisms have many good properties which make it possible to build complex differential privacy mechanisms by basic block algorithms. Two of them are Sequential Composition Theorem and Parallel Composition Theorem.

Sequential Composition Theorem Let A1,A2,,AkA_{1},A_{2},\cdots,A_{k} be kk algorithms that satisfy ϵ1\epsilon_{1}-DP,ϵ2\epsilon_{2}-DP, \cdots,ϵk\epsilon_{k}-DP respectively. Publishing t=<t1,t2,,tk>t=<t_{1},t_{2},\cdots,t_{k}> satisfies (i=1kϵi)(\sum_{i=1}^{k}\epsilon_{i})-DP. Where, t1=A1(D),t2=A2(t1,D),t3=A3(<t1,t2,D>),,tk=Ak(<t1,t2,t3,,tk1>,D)t_{1}=A_{1}(D),t_{2}=A_{2}(t_{1},D),t_{3}=A_{3}(<t_{1},t_{2},D>),\cdots,t_{k}=A_{k}(<t_{1},t_{2},t_{3},\cdots,t_{k-1}>,D).

Parallel Composition Theorem Let A1,A2,,AkA_{1},A_{2},\cdots,A_{k} be kk algorithms that satisfy ϵ1\epsilon_{1}-DP,ϵ2\epsilon_{2}-DP, \dots,ϵk\epsilon_{k}-DP respectively. Publishing t=<t1,t2,,tk>t=<t_{1},t_{2},\cdots,t_{k}> satisfies (maxi[1,2k]ϵi)(max_{i\in[1,2\cdots k]}\epsilon_{i})-DP. Where, t1=A1(D1),t2=A2(t1,D2),t3=A3(<t1,t2,D3>),,tk=Ak(<t1,t2,t3,,tk1>,Dk)t_{1}=A_{1}(D_{1}),t_{2}=A_{2}(t_{1},D_{2}),t_{3}=A_{3}(<t_{1},t_{2},D_{3}>),\cdots,t_{k}=A_{k}(<t_{1},t_{2},t_{3},\cdots,t_{k-1}>,D_{k}) and DiDj=D_{i}\cap D_{j}=\emptyset for all iji\neq j,

2.2 Linear Property

The linear query is one kind of fundamental queries. It is used widely. For example, sum and counting query are popular linear queries. The linear query has a very good property, namely linear property.

Definition 4 (linear property) For linear query qq and data set DD such that D=D1D2D=D_{1}\cup D_{2} and =D1D2\varnothing=D_{1}\cap D_{2}, we have

q(D)=q(D1)+q(D2)\displaystyle q(D)=q(D_{1})+q(D_{2})

The counting query is one kind of linear queries so it satisfies the linear property.

2.3 Hypothesis Test

The hypothesis test is a statistic tool to determine whether fluctuation of experimental results is only due to randomness. Its main goal is to guarantee the confidence level of experimental conclusions. To that end, there are three steps in hypothesis test logically. Firstly, an assumption about experimental conclusions is made. The assumption needs to be contrary to conclusions which are achieved by experimental results. The assumption is called null hypothesis denoted by H0H_{0}. The opposite hypothesis of null hypothesis is called alternative hypothesis denoted by H1H_{1}. Secondly, a statistics is chosen. The statistics needs to be related to the assumption and its probability distribution function is known. Thirdly, the probability of experimental results is calculated through the known distribution. If the probability is smaller than a predetermined threshold, experimental conclusions are reliable. The procedure can guarantee that when the assumption which is contrary to experimental conclusions is right, the experimental results occur with extremely small probability. That is, the experimental results are not only due to randomness.

The concrete steps of hypothesis test method are clear. Firstly, a predetermined probability threshold needs to be chosen and it is called significance level. Secondly, the experimental results’ probability is calculated and the probability is called pp value. At last, a decision is made. In specific, if the pp value is greater than the predetermined significance level, alternative hypothesis is accepted. Otherwise, the null hypothesis is accepted.

The one-sample t-test is an useful hypothesis test method. It is used to determine whether samples come from a distribution with a certain mean.

3 multiple i.i.d. samples for linear query’s answer

This section presents how to obtain multiple i.i.d. samples for linear query’s answer under constrains of differential privacy. Firstly, the idea of proposed method will be introduced. Then, the proposed method will be presented.

Usually, multiple i.i.d. samples for a query’s answer cannot be obtained directly under constrains of differential privacy. Differential privacy mechanisms are random mechanisms so it can output different answers for the same query when the same query comes every time. However, in order to reduce consumption of the privacy budget, differential privacy mechanisms output the same answer for queries which have been answered before. To that end, there are some engineering methods available. For example, a table is maintained. The table records answered queries as well as answers of these queries. When a query comes, differential privacy mechanisms look up whether the query is in the table. If the query was answered before, the recorded answer will be returned so that differential privacy mechanisms do not need to compute the answer again.

The linear property makes it possible to divide a linear query into two parts. For example, the counting query is a linear query. There are four numbers in a set B={b1,b2,b3,b4}B=\{b_{1},b_{2},b_{3},b_{4}\}. The equality count(B)=count({b1,b2})+count({b3,b4})count(B)=count(\{b_{1},b_{2}\})+count(\{b_{3},b_{4}\}) holds. In addition, there are multiple divisions for one query. For instance, the equality count(B)=count({b1})+count({b2,b3,b4})count(B)=count(\{b_{1}\})+count(\{b_{2},b_{3},b_{4}\}) also holds.

The linear property makes it possible to obtain multiple different answers for one linear query under constrains of differential privacy. In specific, one linear query can be divided into two parts such as count(B)=count({b1,b2})+count({b3,b4})count(B)=count(\{b_{1},b_{2}\})+count(\{b_{3},b_{4}\}). If attackers want to obtain the answer of count(B)count(B) under the condition that all queries are only calculated by a differential privacy mechanism, attackers can issue the query count(B)count(B) to the differential privacy mechanism. In addition, in order to obtain another answer of count(B)count(B), attackers can also issue the query count({b1,b2})count(\{b_{1},b_{2}\}) to the differential privacy mechanism and calculate the query count({b3,b4})count(\{b_{3},b_{4}\}) by himself when attackers know the set {b3,b4}\{b_{3},b_{4}\}. In brief, by multiple divisions of a query, it is possible to obtain multiple answers for a linear query under constrains of differential privacy.

There is a necessary assumption for attackers to obtain multiple answers for a linear query. In specific, attackers need to know a set of records in the target data set which is protected by the differential privacy mechanism. The set of records can be regarded as background knowledge of attackers.

The assumption about the background knowledge in the proposed method is weaker than the assumption in differential privacy. The differential privacy claimed that even if attackers know all data records except one, the record which attackers don’t know can be protected [6] [12]. That is, the background knowledge of attackers could be all records except one in the differential privacy. In proposed method, the background knowledge is only a set of records instead of all records.

3.1 Proposed Method

Refer to caption

Fig 1: method to obtain multiple answers for the target query and membersip inference attacks mehtod

Attackers can issue query qs(D)q_{s}(D) to the differential privacy mechanism MM. The qs(D)q_{s}(D) means computing the query qq over the target data set DD under the condition ss which specifies the range of data records. In other words, the qs(D)q_{s}(D) is equal to q(DDs)q(D\cap D_{s}) where the DsD_{s} is the set of records which satisfy the condition ss. For example, ”count the number of students whose age is greater than 10 in a classroom”. The query qq is the counting query, the data set DD is the set of all students in the classroom and the condition ss is that ” the age is greater than 10”.

Suppose that the target record is denoted by xx and the background knowledge of attackers is a set of data records denoted by DknowD_{know}. The goal is to obtain multiple answers for a linear query q({x}Dknow)q(\{x\}\cup D_{know}). The ss represents the condition such that Ds={x}DknowD_{s}=\{x\}\cup D_{know}. Therefore, the target query can be expressed as qsq_{s} and its answer can be expressed as M(qs,D,ϵ)M(q_{s},D,\epsilon).

Attackers choose a subset DiDknowD_{i}\subset D_{know}, construct a query condition sis_{i} such that Dsi=Di{x}D_{s_{i}}=D_{i}\cup\{x\}, issue query qsi(D)q_{s_{i}}(D) to the differential privacy mechanism and obtain ai=M(qsi,D,ϵ)=qsi(D)+Lap(0,ϵ/ΔD)a_{i}=M(q_{s_{i}},D,\epsilon)=q_{s_{i}}(D)+Lap(0,\epsilon/\Delta D) as return. Let a^i=q(DknowDi)+ai\hat{a}_{i}=q(D_{know}\setminus D_{i})+a_{i}.

The data set DiD_{i} is a subset of DknowD_{know} and the DknowD_{know} is background knowledge of attackers so attackers can compute q(DknowDi)q(D_{know}\setminus D_{i}) locally by themselves. The a^i\hat{a}_{i} is an answer of the target query q({x}Dknow)q(\{x\}\cup D_{know}), which will be proved in next subsection.

Through another subset DjDknowD_{j}\subset D_{know}, attackers can get another answer of the target query. Therefore, attackers can get multiple answers of the target query. If DiDj=D_{i}\cap D_{j}=\emptyset for all iji\neq j, the totally consumed privacy budget is different from the differential privacy mechanism’s perspective and attackers’ perspective. The claim is proved in next subsection.

The whole idea of this paper is showed in Fig 1, including the method to obtain multiple answers for the target query and the membership inference attacks method which is presented in detail at next section.

method to obtain multiple answers of q({x}Dknow)q(\{x\}\cup D_{know})
0:    the background knowledge DknowD_{know};the differential privacy mechanism MM;a linear query qq;the total privacy budget ϵt\epsilon_{t};
0:    the mm answers a^1,a^2,,a^m\hat{a}_{1},\hat{a}_{2},\dots,\hat{a}_{m} ;
1:  let ϵ=ϵtm\epsilon=\frac{\epsilon_{t}}{m}
2:  for i=1i=1 to mm do
3:     choose a new subset DiDknowD_{i}\subset D_{know} such that DiDj=D_{i}\cap D_{j}=\emptyset for all jij\neq i ;
4:     construct a query condition sis_{i} such that Dsi=Di{x}D_{s_{i}}=D_{i}\cup\{x\};
5:     ai=M(qsi,D,ϵ)a_{i}=M(q_{s_{i}},D,\epsilon);
6:     a^i=ai+q(DknowDi)\hat{a}_{i}=a_{i}+q(D_{know}\setminus D_{i})
7:  end for
8:  return  a^1,a^2,,a^m\hat{a}_{1},\hat{a}_{2},\dots,\hat{a}_{m}

3.2 Correctness and Privacy

In this subsection, the proposed method is analyzed in terms of correctness and privacy respectively. In the theorem 1, we prove that linear queries’ multiple answers can be obtained. In theorem 2, we analyze the amount of background knowledge of attackers. Then, totally consumed privacy budget is analyzed from the two perspectives.

Theorem 1: a^1,a^2,,a^m\hat{a}_{1},\hat{a}_{2},\dots,\hat{a}_{m} are i.i.d. samples of the target linear queries’ answer calculated by differential privacy mechanisms over the target data set DD.

Proof:

For i\forall i, we have

a^i\displaystyle\hat{a}_{i} =\displaystyle= q(DknowDi)+ai\displaystyle q(D_{know}\setminus D_{i})+a_{i}
a^i\displaystyle\hat{a}_{i} =\displaystyle= q(DknowDi)+M(qsi,D,ϵ)\displaystyle q(D_{know}\setminus D_{i})+M(q_{s_{i}},D,\epsilon)
a^i\displaystyle\hat{a}_{i} =\displaystyle= q(DknowDi)+qsi(D)+Lap(0,ΔDϵ)\displaystyle q(D_{know}\setminus D_{i})+q_{s_{i}}(D)+Lap(0,\frac{\Delta D}{\epsilon})

We know that qsi(D)=q((Di{x})D)q_{s_{i}}(D)=q((D_{i}\cup\{x\})\cap D). And we also know

(Di{x})DDi{x}\displaystyle(D_{i}\cup\{x\})\cap D\ \subset\ D_{i}\cup\{x\}

And

(Di{x})(DknowDi)=\displaystyle(D_{i}\cup\{x\})\ \cap\ (D_{know}\setminus D_{i})=\emptyset

In addition, qq is linear query. We have

q(DknowDi)+qsi(D)\displaystyle q(D_{know}\setminus D_{i})+q_{s_{i}}(D)
=\displaystyle= q(DknowDi)+q((Di{x})D)\displaystyle q(D_{know}\setminus D_{i})+q((D_{i}\cup\{x\})\cap D)
=\displaystyle= q(Du)\displaystyle q(D_{u})

Here DuD_{u} is union set of DknowDiD_{know}\setminus D_{i} and (Di{x})D(D_{i}\cup\{x\})\cap D.

Du\displaystyle D_{u} =\displaystyle= (DknowDi)((Di{x})D)\displaystyle(D_{know}\setminus D_{i})\cup((D_{i}\cup\{x\})\cap D)
Du\displaystyle D_{u} =\displaystyle= ((DknowDi)D)((Di{x})D)\displaystyle((D_{know}\setminus D_{i})\cap D)\cup((D_{i}\cup\{x\})\cap D)
Du\displaystyle D_{u} =\displaystyle= ((DknowDi)(Di{x}))D)\displaystyle((D_{know}\setminus D_{i})\cup(D_{i}\cup\{x\}))\cap D)
Du\displaystyle D_{u} =\displaystyle= (Dknow{x})D)\displaystyle(D_{know}\cup\{x\})\cap D)

So, we have

q(Du)\displaystyle q(D_{u}) =\displaystyle= q((Dknow{x})D)\displaystyle q((D_{know}\cup\{x\})\cap D)
q(Du)\displaystyle q(D_{u}) =\displaystyle= qs(D)\displaystyle q_{s}(D)

So, we have

a^i\displaystyle\hat{a}_{i} =\displaystyle= q(DknowDi)+qsi(D)+Lap(0,ΔDϵ)\displaystyle q(D_{know}\setminus D_{i})+q_{s_{i}}(D)+Lap(0,\frac{\Delta D}{\epsilon})
a^i\displaystyle\hat{a}_{i} =\displaystyle= qs(D)+Lap(0,ΔDϵ)(1)\displaystyle q_{s}(D)+Lap(0,\frac{\Delta D}{\epsilon})\ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ \ (1)
a^i\displaystyle\hat{a}_{i} =\displaystyle= M(qs,D,ϵ)\displaystyle M(q_{s},D,\epsilon)

According to the equality (1), for all a^1,a^2,,a^m\hat{a}_{1},\hat{a}_{2},\dots,\hat{a}_{m}, the noise Lap(0,ΔDϵ)Lap(0,\frac{\Delta D}{\epsilon}) are i.i.d. samples of a Laplace distribution. So, the a^1,a^2,,a^m\hat{a}_{1},\hat{a}_{2},\dots,\hat{a}_{m} are i.i.d. samples of the target query’s answer. \hfill\Box

The background knowledge is a foundation of proposed method. The background knowledge is a set of known records in the proposed method, although it is captured by prior probability distribution in some papers such as [20]. It is easier to obtain some records of a target data set than to obtain a prior probability distribution of the target data set. For example, with respect to score of students, it is easier to collect some students’ score than to obtain a prior probability distribution of students’ score.

The number of records in DknowD_{know} is a way to quantify the amount of background knowledge of attackers.

Theorem 2: The number of records in DknowD_{know} needs to be greater than mm.

Proof:

As known before, a subset of DknowD_{know} can be used to generate an i.i.d. sample for the target query’s answer. The number of i.i.d. samples is mm. Therefore, the number of records in DknowD_{know} should guarantee that the number of possibly used subsets of DknowD_{know} is equal to or greater than mm.

In the proposed method, all possibly used subsets are disjoint with each other. Therefore, the number of possibly used subset is equal to the number of records in DknowD_{know} at most. In specific, each possibly used subset has only one record of DknowD_{know}.

In a word, the number of records in DknowD_{know} needs to be greater than mm.

\hfill\Box

The privacy budget is the key parameter for privacy guarantee in differential privacy mechanisms. The smaller the privacy budget is, the stronger the privacy guarantee is. One good property of differential privacy is that the totally consumed privacy budget can be calculated in a cumulative way.

Theorem 3: From attackers’ perspective, the totally consumed privacy budget is ϵt\epsilon_{t}.

Proof

For i\forall i, the consumed privacy budget of M(qsi,D,ϵ)M(q_{s_{i}},D,\epsilon) is ϵ=ϵtm\epsilon=\frac{\epsilon_{t}}{m}. By the theorem 1, attackers obtain mm answers for the target query from the differential privacy mechanism. According to the Sequential Composition Theorem, the totally consumed privacy budget is

i=1mϵ=mϵ=ϵt\displaystyle\sum_{i=1}^{m}\epsilon=m\epsilon=\epsilon_{t}

\hfill\Box

From the attackers’ perspective, the totally consumed privacy budget can be unreasonably large if there are enough records in the set of background knowledge. In specific, if attackers can consume ϵ\epsilon privacy budget to obtain one answer of the target query, then attackers can consume mϵm*\epsilon privacy budget to obtain mm different answers of the target query using the proposed method. So, if there are enough records in the set of background knowledge, attackers can consume unreasonably large privacy budget for the target query. However, from the differential privacy mechanism’s perspective, the totally consumed privacy budget is different.

Theorem 4: From the differential privacy mechanism’s perspective, when the target record xx is not in the target data set the totally consumed privacy budget is ϵ\epsilon. When the target record xx is in the target data set, the differential privacy mechanism responds mm different query with ϵ\epsilon privacy budget for each response, in resulting ϵt\epsilon_{t} in total.

Proof

Firstly, when the target record xx is not in the target data set DD, for i\forall\ i we have

qsi=q((Di{x})D)=q(DiD)=q(Di)\displaystyle q_{s_{i}}=q((D_{i}\cup\{x\})\cap D)=q(D_{i}\cap D)=q(D_{i})

The consumed privacy budget of qsiq_{s_{i}} is denoted by ϵi\epsilon_{i}. For 0<ij<m\forall\ 0<i\neq j<m, we know DiDj=D_{i}\cap D_{j}=\emptyset. According to the Parallel Composition Theorem, the totally consumed privacy budget is

max0<i<mϵi=ϵ\displaystyle\max\limits_{0<i<m}\epsilon_{i}=\epsilon

Secondly, when the target record xx is in the target data set DD, for i\forall\ i we have

qsi=q((Di{x})D)=q(Di{x})\displaystyle q_{s_{i}}=q((D_{i}\cup\{x\})\cap D)=q(D_{i}\cup\{x\})

For 0<ij<m\forall\ 0<i\neq j<m, we have

(Di{x})(Dj{x})={x}\displaystyle(D_{i}\cup\{x\})\cap(D_{j}\cup\{x\})=\{x\}\neq\emptyset

According to the Sequential Composition Theorem, the totally consumed privacy budget is

i=1mϵi=i=1mϵ=mϵ=ϵt\displaystyle\sum_{i=1}^{m}\epsilon_{i}=\sum_{i=1}^{m}\epsilon=m\epsilon=\epsilon_{t}

In a word, from the differential privacy mechanism’s perspective, it responds mm different query with ϵ\epsilon privacy budget for each response, in resulting ϵt\epsilon_{t} in total. That is, from the differential privacy mechanism’s perspective, the target query only consume privacy budget ϵ\epsilon. \hfill\Box

In a word, the totally consumed privacy budget is delicately analyzed from attackers’ perspective and differential privacy mechanism’s perspective in the theorem 3 and 4 respectively. The attackers can consume unreasonably large privacy budget while the differential privacy mechanism may not beware. Whether the differential privacy mechanism bewares the unreasonably large consumed privacy budget depends on weather the target record is in the target data set. That is, the presence or absence of the target record has key influence on behavior of the differential privacy mechanism. Thus, it is possible to infer whether the target record is in the target data set by observing outputs of differential privacy mechanism.

4 membership inference attacks method

In order to demonstrate the unexpected information leakage of differential privacy mechanisms due to linear property, a membership inference attacks method against the Laplace mechanism is constructed based on linear queries.

In this section, the main content is divided into five subsections. In the first subsection, the membership inference attacks model is introduced. In the second subsection, an analysis is given to discuss the security foundation of the Laplace mechanism. In the third subsection, reasons why counting query is an appropriate liner query to perform membership inference attacks is discussed. In the fourth subsection, the membership inference attacks method is presented. In the last subsection, the success rate of proposed attacks method is analyzed.

4.1 Membership Inference Attack model

The membership inference attacks is to determine whether a data record is in a data set. Specifically, there is a target data set with sensitive information and a differential privacy mechanism is used to protect the target data set. Attackers has black-box access to the differential privacy mechanism. Given a data record, attackers guess: whether the given record is in the target data set. The formal definition of membership inference attacks is as follows:

Definition 5 (membership inference attacks): Given black-box access to a differential privacy mechanism MM and a target record xx, attackers guess whether the target record is in the the target data set DD.

A(x,M)={1xD0xDA(x,M)=\left\{\begin{array}[]{rcl}1&&{x\in D}\\ 0&&{x\notin D}\end{array}\right.

Here, the AA represents membership inference attacks method. If the attacks method asserts that the target record is in the target data set, it outputs 1. Otherwise, it outputs 0.

There are three assumptions about the black-box access of the differential privacy mechanism.

(1) For the same query, the black-box access returns the same answer no matter how many times the query is issued. If attackers can get a new answer for the same query from the black-box access when attackers issue the query every time, then attackers can get multiple i.i.d. samples of the target query’s answer trivially. With multiple i.i.d. samples, it is trivial for attackers to infer sensitive information of records. Thus, this assumption is reasonable.

(2) For every different query, the black-box access returns an answer. The second assumption guarantees the utility of differential privacy mechanisms. If the black-box access refuses to answer queries which are issued first time, the utility of differential privacy mechanisms is damaged, resulting in useless mechanisms.

(3) There is a threshold of the totally consumed privacy budget. When the totally consumed privacy budget is greater than the threshold, the black-box access aborts. The consumed privacy budget is accumulative and when totally consumed privacy budget is greater than a certain limitation, leaking information is trivial.

The ability of attackers is limited. The black-box access means that attackers issue a query to a differential privacy mechanism and gets a result as return. Attackers cannot get any other information from the differential privacy mechanism.

4.2 Security Analysis of The Laplace Mechanism

The security of the Laplace mechanism is based on perturbation. The differential privacy requires that presence or absence of one record cannot be identified through mechanisms’ outputs. The Laplace mechanism satisfies the requirement by perturbation. Specifically, queries’ answer are perturbed by noise. The noise is drawn from a Laplace distribution whose variance is related to two factors. The first factor is strength of privacy guarantee. The second factor is maximum difference of query’s answer, which is caused by presence or absence of any one record. The maximum difference is captured by the conception of sensitivity which is denoted by ΔD\Delta D. The strength of privacy guarantee is captured by the conception of privacy budget which is denoted by ϵ\epsilon. The variance of noise distribution is related to ratio of the sensitivity to the privacy budget. Specifically, the variance is 2(ΔD/ϵ)22(\Delta D/\epsilon)^{2}.

The key idea behind the Laplace mechanism is: If the sensitivity is far less than fluctuation of added noise, it is hard to tell that variation of the Laplace mechanism’s outputs happens due to noise distribution’s fluctuation or due to presence or absence of a record. This idea can be described intuitively by formula ΔD2(ΔD/ϵ)2\Delta D\ll 2(\Delta D/\epsilon)^{2}.

The goal of covering the sensitivity can be achieved if the privacy budget is appropriately small. That is, for appropriately small privacy budget, the sensitivity could be much less than variance of noise distribution.

In a word, for the Laplace mechanism, the sensitivity is covered by fluctuation of noise. So, the security foundation of the Laplace mechanism can be described by formula ΔD2(ΔD/ϵ)2\Delta D\ll 2(\Delta D/\epsilon)^{2} intuitively. However, the formula may not hold for some queries, which we will discuss in next subsection.

The hypothesis test is an appropriate tool to construct membership inference attacks method against the Laplace mechanism. In the Laplace mechanism, the difference caused by presence or absence of one record is covered by the fluctuation of noise. However, the hypothesis test method is designed to determine that variation are just due to random noise or other reasons. In other words, the hypothesis test can be used to determine that variation of the Laplace mechanism’s outputs is just due to randomness of noise or due to presence or absence of a record.

4.3 Counting Query

The counting query is an appropriate query to perform membership inference attacks against the Laplace mechanism. In specific, it is acceptable for most cases that value of the privacy budget is from 0 to 1. Let’s check the inequality ΔD2(ΔD/ϵ)2\Delta D\ll 2(\Delta D/\epsilon)^{2}. For counting query, when the privacy budget is 1, the sensitivity is 1 and the variance of noise distribution is 2. The sensitivity is just a half of variance. When privacy budget is 0.5, the sensitivity is 1 and the variance is 8. So the sensitivity is one eighth of variance. In a word, the inequality ΔD2(ΔD/ϵ)2\Delta D\ll 2(\Delta D/\epsilon)^{2} may not hold for counting query when the privacy budget takes values which are generally considered to be an appropriate value for the privacy budget.

There are other three reasons why the target linear query should be the counting query. Firstly, the counting query makes the attacks method technically right. The counting query satisfies linear property and the sensitivity of counting query is always 1. These two properties make it possible to obtain multiple i.i.d. samples for the counting query’s answer. In addition, the hypothesis test method works on i.i.d. samples. Thus, the proposed membership inference attacks method is technically right due to unique properties of the counting query.

Secondly, the proposed attacks method is based on counting query so that it can work in a wide range. The counting query is very common. On the one hand, it is a common need to confirm the number of records which meet specific requirements, such as sqlsql statement ” select count(*) from table student where student.age \leq 20 ”. On the other hand, there are a lot of complex queries which are based on the counting query.

Thirdly, the difference of query’s answer, which is caused by presence or absence of one record, is most easily detected by hypothesis test method when the target query is the counting query. As mentioned before, the difference is smaller than or equal to the sensitivity and the sensitivity is far smaller than variance of noise distribution. Thus, it is hard to detect whether variation of the Laplace mechanism’s outputs is due to presence or absence of one record. However, when the target query is the counting query, the difference caused by presence or absence of any record is 1, reaching the value of sensitivity. So, the counting query makes it most easy to detect whether the variation of outputs is caused by presence or absence of one record.

4.4 Instance Attacks

This subsection presents how to perform membership inference attacks against the Laplace mechanism.

Attackers aim for determining whether a target record xx is in a target data set DD. The background knowledge of attackers is denoted by DknowD_{know} which is a subset of target data set DD. The target linear query is the counting query q({x}Dknow)q(\{x\}\cup D_{know}). The ss represents the condition such that Ds={x}DknowD_{s}=\{x\}\cup D_{know}. So, the target query can be expressed as qs(D)q_{s}(D) and its answer can be expressed as M(qs,D,ϵ)M(q_{s},D,\epsilon).

The membership inference attacks method consists of two subroutines. The first subroutine is to obtain multiple i.i.d. samples for the target query’s answer. The first subroutine is feasible due to the linear property of the counting query, which is proved in previous section. By the first subroutine, attackers can obtain mm i.i.d. samples of the target query’s answer and the samples are a^1,a^2,,a^m\hat{a}_{1},\hat{a}_{2},\dots,\hat{a}_{m}. When the target record xx is in the target data set DD,

qs(D)\displaystyle q_{s}(D) =\displaystyle= q(DsD)\displaystyle q(D_{s}\cap D)
=\displaystyle= q((Dknow{x})D)\displaystyle q((D_{know}\cup\{x\})\cap D)
=\displaystyle= q(Dknow{x})\displaystyle q(D_{know}\cup\{x\})

Therefore, the mean of samples is q(Dknow{x})q(D_{know}\cup\{x\}). When the target record xx is not in data set DD

qs(D)\displaystyle q_{s}(D) =\displaystyle= q(DsD)\displaystyle q(D_{s}\cap D)
=\displaystyle= q((Dknow{x})D)\displaystyle q((D_{know}\cup\{x\})\cap D)
=\displaystyle= q(Dknow)\displaystyle q(D_{know})

Therefore, the mean of samples is q(Dknow)q(D_{know}).

Attackers can determine whether the target record is in the target data set by comparing q(Dknow)q(D_{know}) with the mean of samples. The procedure of comparison is implemented by hypotheses test method in the second subroutine.

In the second subroutine, the hypothesis test method is used to determine whether the samples which are obtained in the first subroutine are drawn from a distribution with mean q(Dknow)q(D_{know}). If the mean of samples is determined to be q(Dknow)q(D_{know}), the assertion is that the target record is not in the target data set. Otherwise, the assertion is that the target record is in the target data set.

In the second subroutine, the key is to find an appropriate hypothesis test method. That is, it is key to find an appropriate statistics. The one-sample t-test is chosen. The TT statistics follows TT distribution and the TT statistics is as follows

T\displaystyle T =\displaystyle= X¯μS/m\displaystyle\frac{\bar{X}-\mu}{S/\sqrt{m}}

Here, the μ\mu is distribution mean and the mm is the number of samples. The X¯\bar{X} is the sample mean and the SS is sample standard variance. When the samples are denoted by a^1,a^2,,a^m\hat{a}_{1},\hat{a}_{2},\dots,\hat{a}_{m} we have

X¯=(a^1+a^2++a^m)/m\displaystyle\bar{X}=(\hat{a}_{1}+\hat{a}_{2}+\dots+\hat{a}_{m})/m

And,

S=(a^1X¯)2+(a^2X¯)2++(a^mX¯)2m\displaystyle S=\sqrt{\frac{(\hat{a}_{1}-\bar{X})^{2}+(\hat{a}_{2}-\bar{X})^{2}+\dots+(\hat{a}_{m}-\bar{X})^{2}}{m}}

The one-sample t-test is an appropriate hypothesis test method because it is designed to determine whether some samples are drawn from a distribution with a certain mean.

The significance level is 0.05. There are two widely accepted choices for significance level, namely 0.010.01 and 0.050.05. We desire that as long as difference is statistically significant, the alternative hypothesis could be accepted. So we choose 0.050.05 as value of significance level.

There is a table where pp value can be searched by the freedom and the value of statistic TT. The freedom is m1m-1 and the value of statistic TT can be calculated as above.

Membership Inference Attacks Method A
0:    the background knowledge DknowD_{know} of attackers;the target record xx;the Laplace mechanism MM;the counting query qq;total privacy budget ϵt\epsilon_{t};the number of samples mm;the significance level α=0.05\alpha=0.05;
0:    the xx is in data set DD or not;
1:  construct conditions ss such that Ds=Dknow{x}D_{s}=D_{know}\cup\{x\};
2:  let ϵ=ϵtm\epsilon=\frac{\epsilon_{t}}{m};
3:  for i=1i=1 to mm do
4:     a^i=M(qs,D,ϵ)\hat{a}_{i}=M(q_{s},D,\epsilon) by the first subroutine;
5:  end for
6:  make hypothesis: the null hypothesis is H0H_{0}: μ0=q(Dknow)\mu_{0}=q(D_{know}); the alternative hypothesis is H1H_{1}: μ0q(Dknow)\mu_{0}\neq q(D_{know});
7:  calculate X¯=a^1+a^2++a^mm\bar{X}=\frac{\hat{a}_{1}+\hat{a}_{2}+\dots+\hat{a}_{m}}{m};
8:  calculate S=(a^1X¯)2+(a^2X¯)2++(a^mX¯)2mS=\sqrt{\frac{(\hat{a}_{1}-\bar{X})^{2}+(\hat{a}_{2}-\bar{X})^{2}+\dots+(\hat{a}_{m}-\bar{X})^{2}}{m}};
9:  calculate T=X¯μ0SmT=\frac{\bar{X}-\mu_{0}}{\frac{S}{\sqrt{m}}};
10:  find pp according to TT and freedom m1m-1;if p<αp<\alpha, the assertion is xx\in D; if p>αp>\alpha, the assertion is xDx\notin D
11:  return  assertion

4.5 Success Rate Analysis

The success rate of proposed attacks method is tightly related to two factors including the amount of background knowledge of attackers and the threshold of privacy budget in black-box access. According to the two factors, there are four cases faced by attackers, as shown in Fig 2.

Refer to caption
Fig 2: cases faced by attackers

Suppose that the threshold of the privacy budget is ϵl\epsilon_{l}. When the totally consumed privacy budget is greater than the threshold from the differential privacy mechanism’s perspective, the black-box access of differential privacy mechanisms aborts. The ”enough background knowledge” means that attackers can construct as many as needed number of queries such that the totally consumed privacy budget is greater than the threshold, resulting in abortion of black-box access.

For the case 4, there is enough background knowledge available to attackers. That is, there is enough number of records in DknowD_{know}. In addition, the target record xx is in the target data set. By the theorem 4, from the differential privacy mechanism’s perspective, the totally consumed privacy budget is mϵm\epsilon. Because the background knowledge is enough, attackers can make the mm big enough such that mϵ>ϵlm\epsilon>\epsilon_{l}, result in abortion of the black-box.

For the case 3, there is enough number of records in DknowD_{know} and the target record xx is in not the target data set. By the theorem 4, from the differential privacy mechanism’s perspective, the totally consumed privacy budget is ϵ\epsilon no matter how many queries attackers issue. So the black-box access never aborts. By the theorem 3, from attackers’ perspective, the totally consumed privacy budget is mϵm\epsilon. Thus, attackers can issue as many as needed queries such that the mϵm\epsilon is unreasonably large, resulting in that attackers can find the absence of the target record trivially by the hypothesis test method.

For the case 2 and the case 1, the background knowledge is not enough. Thus, attackers cannot consume unreasonably large privacy budget or force the black-box access to abort. In these cases, the success rate is tightly related to the exact amount of background knowledge. For these two cases, we will give a delicate analysis about the success rate in theorem 5.

In brief, from attackers’ perspective, they issue as many as possible queries to consume as much as possible privacy budget. In specific, there are three possible results. Firstly, the black-box access aborts, which means that the totally consumed privacy budget is greater than the threshold and the target record is in the target data set. Secondly, the black-box access does not aborts and the totally consumed privacy budget is so big that attackers can find the absence of the target record. Thirdly, all records of attackers’ background knowledge are used and attackers obtain multiple answers of the target query. Attackers determine whether the target record is in the target data set by hypothesis test method. The success rate of using hypothesis test method is analyzed in theorem 5 at next.

assertion hypothesis H0H_{0} H1H_{1}
H0H_{0} P{accetpH0|H0isture}=1α=0.95P\{accetp\ H_{0}|H_{0}\ is\ ture\}=1-\alpha=0.95 P{accetpH0|H1isture}=δP\{accetp\ H_{0}|H_{1}\ is\ ture\}=\delta
H1H_{1} P{accetpH1|H0isture}=α=0.05P\{accetp\ H_{1}|H_{0}\ is\ ture\}=\alpha=0.05 P{accetpH1|H1isture}=1δP\{accetp\ H_{1}|H_{1}\ is\ ture\}=1-\delta
Table 1: success rate of hypothesis test

The hypothesis test method has two types error shown in Table 1. Based on the Table 1, the success rate of proposed attacks method can be calculated. The proof of theorem 5 is deferred to the appendix.

Theorem 5 The success rate RR of the proposed membership inference attacks method is

R=12(1.95T(m1,0.975)+(μ0μ1)mST(m1,0.975)+(μ0μ1)mSf(t)𝑑t)\displaystyle R=\frac{1}{2}(1.95-\int_{-T_{(m-1,0.975)}+\frac{(\mu_{0}-\mu_{1})*\sqrt{m}}{S}}^{T_{(m-1,0.975)}+\frac{(\mu_{0}-\mu_{1})*\sqrt{m}}{S}}f(t)dt)

\hfill\Box

Here, the T(m1,0.975)T_{(m-1,0.975)} is the value of TT statistic when the freedom of TT distribution is m1m-1 and the cumulative probability is 0.975. The f(t)f(t) is the probability density function of TT distribution with freedom m1m-1. The μ0\mu_{0} is the mean assumed in null hypothesis. The μ1\mu_{1} is the true mean of distribution when the null hypothesis is wrong. We have

μ0=q(Dknow)andμ1=q(Dknow{x})\displaystyle\mu_{0}=q(D_{know})\ \ and\ \ \mu_{1}=q(D_{know}\cup\{x\})

In above formula, μ0μ1\mu_{0}-\mu_{1} is equal to 1-1 because the query qq is the counting query. The mm is the number of obtained samples from the Laplace mechanism by black-box access. The SS is the standard deviation of obtained samples.

As mentioned before, the standard deviation of noise distribution is 2ΔD/ϵ\sqrt{2}\Delta D/\epsilon. In addition, for the counting query the sensitivity ΔD\Delta D is 1. The standard deviation of sample could be substituted with standard deviation of noise distribution. That is, the SS could be substituted with 2ΔD/ϵ\sqrt{2}\Delta D/\epsilon. Then, the RR is

R12(1.95T(m,0.975)ϵm2T(m,0.975)ϵm2f(t)𝑑t)\displaystyle R\approx\frac{1}{2}(1.95-\int_{-T_{(m,0.975)}-\frac{\epsilon\sqrt{m}}{\sqrt{2}}}^{T_{(m,0.975)}-\frac{\epsilon\sqrt{m}}{\sqrt{2}}}f(t)dt)

The ϵ\epsilon is the privacy budget consumed by each query. By the above formula, the RR increases when the number of obtained samples increases.

In terms of totally consumed privacy budget, the pattern is different. In specific, the totally consumed privacy budget ϵt\epsilon_{t} is equal to mϵm\epsilon. So, the RR is

R12(1.95T(m,0.975)ϵt/2mT(m,0.975)ϵt/2mf(t)𝑑t)\displaystyle R\approx\frac{1}{2}(1.95-\int_{-T_{(m,0.975)}-\epsilon_{t}/\sqrt{2m}}^{T_{(m,0.975)}-\epsilon_{t}/\sqrt{2m}}f(t)dt)

According to the above formula, the RR decreases when the number of obtained samples increases under the condition that the totally consumed privacy budget is predetermined.

By the theorem 2, the maximum number of samples is equal to the number of records in DknowD_{know} at most. Let rr be the number of records in DknowD_{know}. The above two formulas become

R12(1.95T(r,0.975)ϵr2T(r,0.975)ϵr2f(t)𝑑t)\displaystyle R\approx\frac{1}{2}(1.95-\int_{-T_{(r,0.975)}-\frac{\epsilon\sqrt{r}}{\sqrt{2}}}^{T_{(r,0.975)}-\frac{\epsilon\sqrt{r}}{\sqrt{2}}}f(t)dt)

And

R12(1.95T(r,0.975)ϵt/2rT(r,0.975)ϵt/2rf(t)𝑑t)\displaystyle R\approx\frac{1}{2}(1.95-\int_{-T_{(r,0.975)}-\epsilon_{t}/\sqrt{2r}}^{T_{(r,0.975)}-\epsilon_{t}/\sqrt{2r}}f(t)dt)

In above analyses, the consumed privacy budget is calculated from attackers’ perspective. However, there are two cases if the consumed privacy budget is calculated from differential privacy mechanisms’ perspective. Firstly, if the target record is in the target data set, the consumed privacy budget is the same with that of attackers’ perspective. Secondly, if the target record is not in the target data set, the consumed privacy budget is ϵ\epsilon which is the amount of privacy budget consumed by one query.

5 experiment

In this section, the detailed information of extensive experiments is given. The goal is to evaluate the success rate of proposed membership inference attacks method. As analyzed before, for the case 3 and 4, attackers can determine that the target record is in the target data set or not with probability 1. In this section, we evaluate the success rate for the case 1 and 2 where the background knowledge of attackers are not enough.

For the case 1 and 2, the success rate is related to two factors, namely consumed privacy budget and the number of obtained samples. In specific, the privacy budget is the key parameter to control the amount of information released by differential privacy mechanisms. The number of samples has important influence on performance of hypothesis test method. The intuition is consistent with the theoretical analysis in theorem 5. So we pay our attention to the two factors and investigate their influence on the success rate. All experiments are based on the counting query.

The privacy budget is unreasonably large if the value of privacy budget is 10 in real application scenarios. However, the key contribution of this paper is that the totally consumed privacy budget is different for delicately designed queries from attackers’ perspective and from differential privacy mechanisms’ perspective. As analyzed before, attackers could consume unreasonably large privacy budget while the differential privacy mechanism doesn’t beware it. This is the reason why we will choose a large value for totally consumed privacy budget to verify the success rate of proposed attacks method.

The number of samples is from 4 to 29. When the number of samples is less than 4, the success rate is not good enough. The reason is that the number of samples are so small that the standard deviation of samples has big error. For example, if there is only one sample, then the sample standard deviation is 0, resulting in huge error. According to experimental results, when the number of samples is less than 4, there are big fluctuation in experimental results. In order to show patterns of experimental results, experiments need to be repeated so many times if the number of samples is less than 4.

Refer to caption
(a)theoretical results
privacy budget is from 0.1 to 1
Refer to caption
(b)experiment results
privacy budget is from 0.1 to 1
Refer to caption
(c)theoretical results
privacy budget is from 0.1 to 10
Refer to caption
(d)experiment results
privacy budget is from 0.1 to 10
Fig 3: totally consumed privacy budget is a constant
Refer to caption
(a)theoretical results
privacy budget is from 0.01 to 0.1
Refer to caption
(b)experiment results
privacy budget is from 0.01 to 0.1
Refer to caption
(c)theoretical results
privacy budget is from 0.01 to 0.33
Refer to caption
(d)experiment results
privacy budget is from 0.01 to 0.33
Fig 4: consumed privacy budget of each query is a constant

In the first experiment, suppose that the totally consumed privacy budget is a constant to verify the theoretical analysis of the success rate. The privacy budget is from 0.1 to 1 and from 0.1 to 10 respectively. The number of samples is from 4 to 29. We repeat experiments and show average of these experimental results in Fig 3.

In Fig 3, the privacy budget is the totally consumed privacy budget by all samples. In specific, if the privacy budget is 1 and the number of samples is 10, consumed privacy budget of each sample is 0.1.

According to results in Fig 3, the experimental results are consistent with theoretical analysis. The success rate increases when the totally consumed privacy budget increases. If the totally consumed privacy becomes bigger, the privacy budget allocated to each sample becomes bigger. Thus, each sample is perturbed with smaller noise. So, the hypothesis test method has better performance. When the totally consumed privacy budget is constant, the success rate decreases as the number of samples increases. The reason is that when the number of samples is bigger the privacy budget of every sample is smaller, resulting in bigger noise of each sample. For example, totally consumed privacy budget is 1. When the number of samples is 5, the privacy budget for every sample is 0.2. While when the number of samples is 10, the privacy budget for every sample is 0.1.

In the second experiment, suppose that privacy budget of each sample is constant to verify theoretical analysis of the success rate. As in the first experiment, the range of the number of samples is from 4 to 29. The privacy budget is from 0.01 to 0.1 and from 0.01 to 0.33 respectively. We repeat experiments and show average of these experimental results in Fig 4.

In Fig 4, the privacy budget is the consumed privacy budget of each sample. In specific, if the privacy budget is 0.2 and the number of samples is 10, consumed privacy budget is 2 in total.

According to results in Fig 4, the experimental results are consistent with theoretical analysis. The success rate increases when the privacy budget of each sample increases. When the privacy budget is smaller, the added noise is bigger. It is hard to tell that variation of Laplace mechanism’s outputs is due to presence or absence of one record or due to fluctuation of added noise. So the success rate is low. When the privacy budget is bigger, the added noise is smaller so it is easier to tell the variation caused by presence or absence of one record from fluctuation of added noise.

The success rate increases when the number of samples increases. Intuitively speaking, when the privacy budget of each sample is constant, the amount of information of each sample is almost the same. So, when the number of samples increases, the amount of information released by differential privacy mechanisms increases, resulting in increase of the success rate.

In all above experiments, the totally consumed privacy budget is calculated from attackers’ perspective. Because the proposed method is a method to perform attacks, it is meaningful to calculate totally consumed privacy budget from attackers’ perspective.

From differential privacy mechanisms’ perspective, if the totally consumed privacy budget exceeds the threshold of the privacy budget, the black-box access aborts, resulting in that attackers know presence of the target record in the target data set. If the totally consumed privacy budget does not exceed the threshold , attackers can continually consume more privacy budget until the success rate is acceptable for attackers or all records in DknowD_{know} are used.

6 conclusion

In this paper, we find that the differential privacy does not take liner property of queries into account, resulting in unexpected information leakage. In specific, the totally consumed privacy budget of delicately designed queries may be different from attackers’ perspective and from differential privacy mechanisms’ perspective due to linear property. The difference leads to unexpected information leakage. In addition, we show how attackers can consume extra privacy budget by its background knowledge, which is against the old opinion that the amount of background knowledge almost has no influence on privacy guarantee of differential privacy.

In order to demonstrate the unexpected information leakage, we show a membership inference attacks against the Laplace mechanism. Specifically, based on linear property of queries, a method is proposed to obtain multiple i.i.d. samples of the linear query’s answer. Based on these samples, the hypothesis test method is used to determine whether a target record is in a target data set. Based on counting query, we perform extensive experiments to verify proposed membership inference attacks method. According to experimental results, the proposed membership inference attacks method works well.

References

  • [1] Nour Almadhoun, Erman Ayday, and Özgür Ulusoy. Differential privacy under dependent tuples?the case of genomic privacy. Bioinformatics, 36(6):1696–1703, 2020.
  • [2] Nour Almadhoun, Erman Ayday, and Özgür Ulusoy. Inference attacks against differentially private query results from genomic datasets including dependent tuples. Bioinformatics, 36(Supplement_1):i136–i145, 2020.
  • [3] Michael Backes, Pascal Berrang, Mathias Humbert, and Praveen Manoharan. Membership privacy in microrna-based studies. In Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security, CCS 2016, pages 319–330, New York, NY, USA, 2016. Association for Computing Machinery.
  • [4] J. Chen, H. Ma, D. Zhao, and L. Liu. Correlated differential privacy protection for mobile crowdsensing. IEEE Transactions on Big Data, pages 1–1, 2017.
  • [5] Damien Desfontaines and Balázs Pejó. Sok: Differential privacies. Proceedings on Privacy Enhancing Technologies, 2020(2):288–313, 2020.
  • [6] Weibei Fan, Jing He, Mengjiao Guo, Peng Li, Zhijie Han, and Ruchuan Wang. Privacy preserving classification on local differential privacy in data centers. Journal of Parallel and Distributed Computing, 135:70–82, 2020.
  • [7] Quan Geng, Wei Ding, Ruiqi Guo, and Sanjiv Kumar. Tight analysis of privacy and utility tradeoff in approximate differential privacy. In International Conference on Artificial Intelligence and Statistics, pages 89–99, 2020.
  • [8] Nils Homer, Szabolcs Szelinger, Margot Redman, David Duggan, Waibhav Tembe, Jill Muehling, John V Pearson, Dietrich A Stephan, Stanley F Nelson, and David W Craig. Resolving individuals contributing trace amounts of dna to highly complex mixtures using high-density snp genotyping microarrays. PLoS genetics, 4(8), 2008.
  • [9] Wen Huang, Shijie Zhou, Yongjian Liao, and Hongjie Chen. An efficient differential privacy logistic classification mechanism. IEEE Internet of Things Journal, 6(6):10620–10626, 2019.
  • [10] Wen Huang, Shijie Zhou, Yongjian Liao, and Ming Zhuo. Optimizing query times for multiple users scenario of differential privacy. IEEE Access, 7:183292–183299, 2019.
  • [11] Noah Johnson, Joseph P Near, and Dawn Song. Towards practical differential privacy for sql queries. Proceedings of the VLDB Endowment, 11(5):526–539, 2018.
  • [12] Yang Li, Dasen Yang, and Xianbiao Hu. A differential privacy-based privacy-preserving data publishing algorithm for transit smart card data. Transportation Research Part C: Emerging Technologies, 115:102634, 2020.
  • [13] Denglong Lv and Shibing Zhu. Correlated differential privacy protection for big data. In 2018 IEEE 32nd International Conference on Advanced Information Networking and Applications (AINA), pages 1011–1018. IEEE, 2018.
  • [14] Kobbi Nissim, Sofya Raskhodnikova, and Adam Smith. Smooth sensitivity and sampling in private data analysis. In Proceedings of the thirty-ninth annual ACM symposium on Theory of computing, pages 75–84, 2007.
  • [15] Md Atiqur Rahman, Tanzila Rahman, Robert Laganière, Noman Mohammed, and Yang Wang. Membership inference attack against differentially private deep learning model. Transactions on Data Privacy, 11(1):61–79, 2018.
  • [16] Reza Shokri, Marco Stronati, Congzheng Song, and Vitaly Shmatikov. Membership inference attacks against machine learning models. In 2017 IEEE Symposium on Security and Privacy (SP), pages 3–18. IEEE, 2017.
  • [17] Hao Wang, Zhengquan Xu, Shan Jia, Ying Xia, and Xu Zhang. Why current differential privacy schemes are inapplicable for correlated data publishing. World Wide Web, pages 1–23.
  • [18] Genqiang Wu, Xianyao Xia, and Yeping He. Extending differential privacy for treating dependent records via information theory. arXiv preprint arXiv, 1703, 2017.
  • [19] Xiaotong Wu, Taotao Wu, Maqbool Khan, Qiang Ni, and Wanchun Dou. Game theory based correlated privacy preserving analysis in big data. IEEE Transactions on Big Data, 2017.
  • [20] Bin Yang, Issei Sato, and Hiroshi Nakagawa. Bayesian differential privacy on correlated data. In Proceedings of the 2015 ACM SIGMOD international conference on Management of Data, pages 747–762. ACM, 2015.
  • [21] Samuel Yeom, Irene Giacomelli, Matt Fredrikson, and Somesh Jha. Privacy risk in machine learning: Analyzing the connection to overfitting. In 2018 IEEE 31st Computer Security Foundations Symposium (CSF), pages 268–282. IEEE, 2018.
  • [22] Tao Zhang, Tianqing Zhu, Ping Xiong, Huan Huo, Zahir Tari, and Wanlei Zhou. Correlated differential privacy: feature selection in machine learning. IEEE Transactions on Industrial Informatics, 16(3):2115–2124, 2019.
  • [23] Tianqing Zhu, Ping Xiong, Gang Li, and Wanlei Zhou. Correlated differential privacy: Hiding information in non-iid data set. IEEE Transactions on Information Forensics and Security, 10(2):229–242, 2014.

7 appendix

Theorem 5 The success rate RR of the proposed membership inference attacks is equal to

R=12(1.95T(m,0.975)+(μ0μ1)mST(m,0.975)+(μ0μ1)mSf(t)𝑑t)\displaystyle R=\frac{1}{2}(1.95-\int_{-T_{(m,0.975)}+\frac{(\mu_{0}-\mu_{1})*\sqrt{m}}{S}}^{T_{(m,0.975)}+\frac{(\mu_{0}-\mu_{1})*\sqrt{m}}{S}}f(t)dt)

Proof

As shown in table 1, there are two types of error of hypothesis test. According the Table 1, we have

R\displaystyle R =\displaystyle= (1α)+(1δ)(1α)+δ+α+(1δ)\displaystyle\frac{(1-\alpha)+(1-\delta)}{(1-\alpha)+\delta+\alpha+(1-\delta)}
=\displaystyle= 12(2αδ)\displaystyle\frac{1}{2}(2-\alpha-\delta)
=\displaystyle= 12(1.95δ)\displaystyle\frac{1}{2}(1.95-\delta)

For the α\alpha, we have

α\displaystyle\alpha
=\displaystyle= P{accetpH1|H0isture}\displaystyle P\{accetp\ H_{1}|H_{0}\ is\ ture\}
=\displaystyle= P{X¯μ0Sm<T|H0isture}\displaystyle P\{\frac{\bar{X}-\mu_{0}}{\frac{S}{\sqrt{m}}}<-T^{*}|H_{0}\ is\ ture\}
+\displaystyle+ P{T<X¯μ0Sm|H0isture}\displaystyle P\{T^{*}<\frac{\bar{X}-\mu_{0}}{\frac{S}{\sqrt{m}}}|H_{0}\ is\ ture\}
=\displaystyle= 2P{T<X¯μ0Sm|H0isture}\displaystyle 2*P\{T^{*}<\frac{\bar{X}-\mu_{0}}{\frac{S}{\sqrt{m}}}|H_{0}\ is\ ture\}
=\displaystyle= 2(1P{X¯μ0Sm<T|H0isture})\displaystyle 2*(1-P\{\frac{\bar{X}-\mu_{0}}{\frac{S}{\sqrt{m}}}<T^{*}|H_{0}\ is\ ture\})
=\displaystyle= 0.05\displaystyle 0.05

We have,

P{X¯μ0Sm<T|H0isture}=0.975\displaystyle P\{\frac{\bar{X}-\mu_{0}}{\frac{S}{\sqrt{m}}}<T^{*}|H_{0}\ is\ ture\}=0.975

Namely,

P{X¯μ0Sm<T}=0.975\displaystyle P\{\frac{\bar{X}-\mu_{0}}{\frac{S}{\sqrt{m}}}<T^{*}\}=0.975

The X¯μ0Sm\frac{\bar{X}-\mu_{0}}{\frac{S}{\sqrt{m}}} obeys TT distribution with freedom m1m-1. Let the TT^{*} be value of TT statistic with freedom m1m-1 and probability 0.975, namely T=T(m,0.975)T^{*}=T_{(m,0.975)}. In addition, let f(t)f(t) be the probability density function of TT statistic with freedom m1m-1 and probability 0.975. We have

0f(t)𝑑t+0Tf(t)𝑑t=0.975\displaystyle\int_{-\infty}^{0}f(t)dt+\int_{0}^{T^{*}}f(t)dt=0.975

So, we have

0Tf(t)𝑑t=0.475\displaystyle\int_{0}^{T^{*}}f(t)dt=0.475

For the δ\delta, we have

δ\displaystyle\delta
=\displaystyle= P{accetpH0|H1isture}\displaystyle P\{accetp\ H_{0}|H_{1}\ is\ ture\}
=\displaystyle= P{T<X¯μ0Sm<T|H1isture}\displaystyle P\{-T^{*}<\frac{\bar{X}-\mu_{0}}{\frac{S}{\sqrt{m}}}<T^{*}|H_{1}\ is\ ture\}

Suppose that the true mean is μ1\mu_{1}, we have

δ\displaystyle\delta
=\displaystyle= P{T+μ0μ1Sm<X¯μ1Sm<T+μ0μ1Sm|H1isture}\displaystyle P\{-T^{*}+\frac{\mu_{0}-\mu_{1}}{\frac{S}{\sqrt{m}}}<\frac{\bar{X}-\mu_{1}}{\frac{S}{\sqrt{m}}}<T^{*}+\frac{\mu_{0}-\mu_{1}}{\frac{S}{\sqrt{m}}}|H_{1}\ is\ ture\}
=\displaystyle= P{T+μ0μ1Sm<X¯μ1Sm<T+μ0μ1Sm}\displaystyle P\{-T^{*}+\frac{\mu_{0}-\mu_{1}}{\frac{S}{\sqrt{m}}}<\frac{\bar{X}-\mu_{1}}{\frac{S}{\sqrt{m}}}<T^{*}+\frac{\mu_{0}-\mu_{1}}{\frac{S}{\sqrt{m}}}\}

Denote μ0μ1Sm\frac{\mu_{0}-\mu_{1}}{\frac{S}{\sqrt{m}}} by T1T_{1}, we have

δ=T+T1T+T1f(t)𝑑t\displaystyle\delta=\int_{-T^{*}+T_{1}}^{T^{*}+T_{1}}f(t)dt

So, we have

R\displaystyle R =\displaystyle= 12(1.95δ)\displaystyle\frac{1}{2}(1.95-\delta)
=\displaystyle= 12(1.95T+T1T+T1f(t)𝑑t)\displaystyle\frac{1}{2}(1.95-\int_{-T^{*}+T_{1}}^{T^{*}+T_{1}}f(t)dt)
=\displaystyle= 12(1.95T(m,0.975)+(μ0μ1)mST(m,0.975)+(μ0μ1)mSf(t)𝑑t)\displaystyle\frac{1}{2}(1.95-\int_{-T_{(m,0.975)}+\frac{(\mu_{0}-\mu_{1})*\sqrt{m}}{S}}^{T_{(m,0.975)}+\frac{(\mu_{0}-\mu_{1})*\sqrt{m}}{S}}f(t)dt)

\hfill\Box