This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Variational Leakage: The Role of
Information Complexity in Privacy Leakage

Amir Ahooye Atashin [email protected] 0000-0002-6788-4002 University of GenevaGenevaSwitzerland Behrooz Razeghi [email protected] 0000-0001-9568-4166 University of GenevaGenevaSwitzerland Deniz Gündüz [email protected] 0000-0002-6788-4002 Imperial College LondonLondonUnited Kingdom  and  Slava Voloshynovskiy [email protected] 0000-0002-7725-395X University of GenevaGenevaSwitzerland
Abstract.

We study the role of information complexity in privacy leakage about an attribute of an adversary’s interest, which is not known a priori to the system designer. Considering the supervised representation learning setup and using neural networks to parameterize the variational bounds of information quantities, we study the impact of the following factors on the amount of information leakage: information complexity regularizer weight, latent space dimension, the cardinalities of the known utility and unknown sensitive attribute sets, the correlation between utility and sensitive attributes, and a potential bias in a sensitive attribute of adversary’s interest. We conduct extensive experiments on Colored-MNIST and CelebA datasets to evaluate the effect of information complexity on the amount of intrinsic leakage.

A repository of the proposed method implementation, Colored-MNIST dataset generator and the corresponding analysis is publicly available at:

https://github.com/BehroozRazeghi/Variational-Leakage

Information complexity, privacy, intrinsic leakage, statistical inference, information bottleneck

1. Introduction

Sensitive information sharing is a challenging problem in information systems. It is often handled by obfuscating the available information before sharing it with other parties. In (Makhdoumi et al., 2014), this problem has been formalized as the privacy funnel (PF) in an information theoretic framework. Given two correlated random variables 𝐒\mathbf{S} and 𝐗\mathbf{X} with a joint distribution P𝐒,𝐗P_{\mathbf{S},\mathbf{X}}\,, where 𝐗\mathbf{X} represents the available information and 𝐒\mathbf{S} the private latent variable, the goal of the PF model is to find a representation 𝐙\mathbf{Z} of 𝐗\mathbf{X} using a stochastic mapping P𝐙𝐗P_{\mathbf{Z}\mid\mathbf{X}} such that: (i) 𝐒𝐗𝐙\mathbf{S}\hbox{$\--$}\kern-1.5pt\hbox{$\circ$}\kern-1.5pt\hbox{$\--$}\mathbf{X}\hbox{$\--$}\kern-1.5pt\hbox{$\circ$}\kern-1.5pt\hbox{$\--$}\mathbf{Z} form a Markov chain; and (ii) representation 𝐙\mathbf{Z} is maximally informative about the useful data 𝐗\mathbf{X} (maximizing Shannon’s mutual information (MI) I(𝐗;𝐙)\mathrm{I}\left(\mathbf{X};\mathbf{Z}\right)) while being minimally informative about the sensitive data 𝐒\mathbf{S} (minimizing I(𝐒;𝐙)\mathrm{I}\left(\mathbf{S};\mathbf{Z}\right)). There have been many extensions of this model in the recent literature, e.g., (Makhdoumi et al., 2014; P. Calmon et al., 2015; Basciftci et al., 2016; Sreekumar and Gündüz, 2019; Hsu et al., 2019; Rassouli et al., 2019; Rassouli and Gündüz, 2019; Razeghi et al., 2020; Rassouli and Gündüz, 2021).

Refer to caption
Figure 1. The general setup.

In this paper, we will consider a delicate generalization of the PF model considered in (Basciftci et al., 2016; Rassouli and Gündüz, 2021), where the goal of the system designer is not to reveal the data that has available but another correlated utility variable. In particular, we assume that the data owner/user acquires some utility from the service provider based on the amount of information disclosed about a utility random variable 𝐔\mathbf{U} correlated with 𝐗\mathbf{X}, measured by I(𝐔;𝐙)\mathrm{I}\!\left(\mathbf{U};\mathbf{Z}\right). Therefore, considering Markov chain (𝐔,𝐒)𝐗𝐙\left(\mathbf{U},\mathbf{S}\right)\hbox{$\--$}\kern-1.5pt\hbox{$\circ$}\kern-1.5pt\hbox{$\--$}\mathbf{X}\hbox{$\--$}\kern-1.5pt\hbox{$\circ$}\kern-1.5pt\hbox{$\--$}\mathbf{Z}, the data owner’s aim is to share a representation 𝐙\mathbf{Z} of observed data 𝐗\mathbf{X}, through a stochastic mapping P𝐙𝐗P_{\mathbf{Z}\mid\mathbf{X}}, while preserving information about utility attribute 𝐔\mathbf{U} and obfuscate information about sensitive attribute 𝐒\mathbf{S} (see Fig. 1).

The implicit assumption in the PF model presented above and the related generative adversarial privacy framework (Huang et al., 2017; Tripathy et al., 2019) is to have pre-defined interests in the game between the ‘defender’ (data owner/user) and the ‘adversary’; that is, the data owner knows in advance what feature/ variable of the underlying data the adversary is interested in. Accordingly, the data release mechanism can be optimized/ tuned to minimize any inference the adversary can make about this specific random variable. However, this assumption is violated in most real-world scenarios. The attribute that the defender may assume as sensitive may not be the attribute of interest for the inferential adversary. As an example, for a given utility task at hand, the defender may try to restrict inference on gender recognition while the adversary is interested in inferring an individual’s identity or facial emotion. Inspired by (Issa et al., 2019), and in contrast to the above setups, we consider the scenario in which the adversary is curious about an attribute that is unknown to the system designer.

In particular, we argue that the information complexity of the representation measured by MI I(𝐗;𝐙)\mathrm{I}\!\left(\mathbf{X};\mathbf{Z}\right) can also limit the information leakage about the unknown sensitive variable. In this paper, obtaining the parameterized variational approximation of information quantities, we investigate the core idea of (Issa et al., 2019) in the supervised representation learning setup.

Notation: Throughout this paper, random vectors are denoted by capital bold letters (e.g., 𝐗\mathbf{X}), deterministic vectors are denoted by small bold letters (e.g., 𝐱\mathbf{x}), and alphabets (sets) are denoted by calligraphic fonts (e.g., 𝒳\mathcal{X}). We use the shorthand [N]\left[N\right] to denote the set {1,2,,N}\{1,2,\dots,N\}. H(P𝐗)𝔼P𝐗[logP𝐗]\mathrm{H}\left(P_{\mathbf{X}}\right)\!\coloneqq\mathbb{E}_{P_{\mathbf{X}}}\left[-\log P_{\mathbf{X}}\right] denotes the Shannon’s entropy, while H(P𝐗Q𝐗)𝔼P𝐗[logQ𝐗]\mathrm{H}\left(P_{\mathbf{X}}\|Q_{\mathbf{X}}\right)\coloneqq\mathbb{E}_{P_{\mathbf{X}}}\left[-\log Q_{\mathbf{X}}\right] denotes the cross-entropy of the distribution P𝐗P_{\mathbf{X}} relative to a distribution Q𝐗Q_{\mathbf{X}}. The relative entropy is defined as DKL(P𝐗Q𝐗)𝔼P𝐗[logP𝐗Q𝐗]\mathrm{D}_{\mathrm{KL}}\left(P_{\mathbf{X}}\|Q_{\mathbf{X}}\right)\coloneqq\mathbb{E}_{P_{\mathbf{X}}}\big{[}\log\frac{P_{\mathbf{X}}}{Q_{\mathbf{X}}}\big{]}. The conditional relative entropy is defined by:

DKL(P𝐙𝐗Q𝐙𝐗P𝐗)𝔼P𝐗[DKL(P𝐙𝐗=𝐱Q𝐙𝐗=𝐱)].\mathrm{D}_{\mathrm{KL}}\left(P_{\mathbf{Z}\mid\mathbf{X}}\|Q_{\mathbf{Z}\mid\mathbf{X}}\mid P_{\mathbf{X}}\right)\coloneqq\mathbb{E}_{P_{\mathbf{X}}}\left[\mathrm{D}_{\mathrm{KL}}\left(P_{\mathbf{Z}\mid\mathbf{X}=\mathbf{x}}\|Q_{\mathbf{Z}\mid\mathbf{X}=\mathbf{x}}\right)\right].

And the MI is defined by:

I(P𝐗;P𝐙𝐗)DKL(P𝐙𝐗P𝐙P𝐗)\mathrm{I}\left(P_{\mathbf{X}};P_{\mathbf{Z}\mid\mathbf{X}}\right)\coloneqq\mathrm{D}_{\mathrm{KL}}\left(P_{\mathbf{Z}\mid\mathbf{X}}\|P_{\mathbf{Z}}\mid P_{\mathbf{X}}\right)

We abuse notation to write H(𝐗)=H(P𝐗)\mathrm{H}\left(\mathbf{X}\right)=\mathrm{H}\left(P_{\mathbf{X}}\right) and I(𝐗;𝐙)=I(P𝐗;P𝐙𝐗)\mathrm{I}\left(\mathbf{X};\mathbf{Z}\right)=\mathrm{I}\left(P_{\mathbf{X}};P_{\mathbf{Z}\mid\mathbf{X}}\right) for random vectors 𝐗P𝐗\mathbf{X}\sim P_{\mathbf{X}} and 𝐙P𝐙\mathbf{Z}\sim P_{\mathbf{Z}}.

2. Problem Formulation

Given the observed data 𝐗\mathbf{X}, the data owner wishes to release a representation 𝐙\mathbf{Z} for a utility task 𝐔\mathbf{U}. Our aim is to investigate the potential statistical inference about a sensitive random attribute 𝐒\mathbf{S} from the released representation 𝐙\mathbf{Z}. The sensitive attribute 𝐒\mathbf{S} is possibly also correlated with 𝐔\mathbf{U} and 𝐗\mathbf{X}.

The objective is to obtain a stochastic map P𝐙𝐗:𝒳𝒵P_{\mathbf{Z}\mid\mathbf{X}}:\!\mathcal{X}\rightarrow\mathcal{Z} such that P𝐔𝐙P𝐔𝐗,𝐙𝒵,𝐔𝒰,𝐗𝒳P_{\mathbf{U}\mid\mathbf{Z}}\!\approx\!P_{\mathbf{U}\mid\mathbf{X}},\forall\,\mathbf{Z}\!\in\!\mathcal{Z},\forall\,\mathbf{U}\!\in\!\mathcal{U},\forall\,\mathbf{X}\!\in\!\mathcal{X}. This means that the posterior distribution of the utility attribute 𝐔\mathbf{U} is similar when conditioned on the released representation 𝐙\mathbf{Z} or on the original data 𝐗\mathbf{X}. Under logarithmic loss, one can measure the utility by Shannon’s MI (Makhdoumi et al., 2014; Tishby et al., 2000; Razeghi et al., 2020). The logarithmic loss function has been widely used in learning theory (Cesa-Bianchi and Lugosi, 2006), image processing (Andre et al., 2006), information bottleneck (Harremoës and Tishby, 2007), multi-terminal source coding (Courtade and Wesel, 2011), and PF (Makhdoumi et al., 2014).

Threat Model: We make minimal assumptions about the adversary’s goal, which can model a large family of potential adversaries. In particular, we have the following assumptions:

  • The distribution P𝐒𝐗P_{\mathbf{S}\mid\mathbf{X}} is unknown to the data user/owner. We only restrict attribute 𝐒\mathbf{S} to be discrete, which captures most scenarios of interest, e.g., a facial attribute, an identity, a political preference.

  • The adversary observes released representation 𝐙\mathbf{Z} and the Markov chain (𝐔,𝐒)𝐗𝐙(\mathbf{U},\mathbf{S})\hbox{$\--$}\kern-1.5pt\hbox{$\circ$}\kern-1.5pt\hbox{$\--$}\mathbf{X}\hbox{$\--$}\kern-1.5pt\hbox{$\circ$}\kern-1.5pt\hbox{$\--$}\mathbf{Z} holds.

  • We assume the adversary knows the mapping P𝐙𝐗P_{\mathbf{Z}\mid\mathbf{X}} designed by the data owner, i.e., the data release mechanism is public. Furthermore, the adversary may have access to a collection of the original dataset with the corresponding labels 𝐒\mathbf{S}.

Suppose that the sensitive attribute 𝐒𝒮\mathbf{S}\!\in\!\mathcal{S} has a uniform distribution over a discrete set 𝒮\mathcal{S}, where |𝒮|=2L<|\mathcal{S}|\!=\!2^{L}\!<\!\infty. If I(𝐒;𝐙)Lϵ\mathrm{I}(\mathbf{S};\mathbf{Z})\geq L-\epsilon, then equivalently H(𝐒𝐙)ϵ\mathrm{H}(\mathbf{S}\!\mid\!\mathbf{Z})\leq\epsilon. Also note that due to the Markov chain 𝐒𝐗𝐙\mathbf{S}\hbox{$\--$}\kern-1.5pt\hbox{$\circ$}\kern-1.5pt\hbox{$\--$}\mathbf{X}\hbox{$\--$}\kern-1.5pt\hbox{$\circ$}\kern-1.5pt\hbox{$\--$}\mathbf{Z}, we have I(𝐒;𝐙)=I(𝐗;𝐙)I(𝐗;𝐙𝐒)\mathrm{I}(\mathbf{S};\mathbf{Z})=\mathrm{I}(\mathbf{X};\mathbf{Z})-\mathrm{I}(\mathbf{X};\mathbf{Z}\mid\mathbf{S}). When 𝐒\mathbf{S} is not known a priori, the data owner has no control over I(𝐗;𝐙𝐒)\mathrm{I}(\mathbf{X};\mathbf{Z}\!\mid\!\mathbf{S}). On the other hand, I(𝐗;𝐙)\mathrm{I}(\mathbf{X};\mathbf{Z}) can be interpreted as the information complexity of the released representation, which plays a critical role in controlling the information leakage I(𝐒;𝐙)\mathrm{I}(\mathbf{S};\mathbf{Z}). Note also that a statistic 𝐙=f(𝐗)\mathbf{Z}\!=\!f\!\left(\mathbf{X}\right) induces a partition on the sample space 𝒳\mathcal{X}, where 𝐙\mathbf{Z} is sufficient statistic for 𝐔\mathbf{U} if and only if the assigned samples in each partition do not depend on 𝐔\mathbf{U}. Hence, intuitively, a larger |𝒰||\mathcal{U}| induces finer partitions on 𝒳\mathcal{X}, which could potentially lead to more leakage about the unknown random function 𝐒\mathbf{S} of 𝐗\mathbf{X}. This is the core concept of the notion of variational leakage, which we shortly address in our experiments.

Since the data owner does not know the particular sensitive variable of interest to the adversary, we argue that it instead aims to design P𝐙𝐗P_{\mathbf{Z}\mid\mathbf{X}} with the minimum (information) complexity and minimum utility loss. With the introduction of a Lagrange multiplier β[0,1]\beta\!\in\!\left[0,1\right], we can formulate the objective of the data owner by maximizing the associated Lagrangian functional:

(1) (P𝐙𝐗,β)=I(𝐔;𝐙)βI(𝐗;𝐙).\mathcal{L}\left(P_{\mathbf{Z}\mid\mathbf{X}},\beta\right)=\mathrm{I}\left(\mathbf{U};\mathbf{Z}\right)-\,\beta\,\mathrm{I}\left(\mathbf{X};\mathbf{Z}\right).

This is the well-known information bottleneck (IB) principle (Tishby et al., 2000), which formulates the problem of extracting, in the most succinct way, the relevant information from random variable 𝐗\mathbf{X} about the random variable of interest 𝐔\mathbf{U}. Given two correlated random variables 𝐔\mathbf{U} and 𝐗\mathbf{X} with joint distribution P𝐔,𝐗P_{\mathbf{U,X}}, the goal is to find a representation 𝐙\mathbf{Z} of 𝐗\mathbf{X} using a stochastic mapping P𝐙𝐗P_{\mathbf{Z}\mid\mathbf{X}} such that: (i) 𝐔𝐗𝐙\mathbf{U}\hbox{$\--$}\kern-1.5pt\hbox{$\circ$}\kern-1.5pt\hbox{$\--$}\mathbf{X}\hbox{$\--$}\kern-1.5pt\hbox{$\circ$}\kern-1.5pt\hbox{$\--$}\mathbf{Z}, and (ii) 𝐙\mathbf{Z} is maximally informative about 𝐔\mathbf{U} (maximizing I(𝐔;𝐙)\mathrm{I}\left(\mathbf{U};\mathbf{Z}\right)) and minimally informative about 𝐗\mathbf{X} (minimizing I(𝐗;𝐙)\mathrm{I}\left(\mathbf{X};\mathbf{Z}\right)).

Note that in the PF model, I(𝐗;𝐙)\mathrm{I}\left(\mathbf{X};\mathbf{Z}\right) measures the useful information, which is of the designer’s interest, while in the IB model, I(𝐔;𝐙)\mathrm{I}\left(\mathbf{U};\mathbf{Z}\right) measures the useful information. Hence, I(𝐗;𝐙𝐒)\mathrm{I}\left(\mathbf{X};\mathbf{Z}\mid\mathbf{S}\right) in PF quantifies the residual information, while I(𝐗;𝐙𝐔)\mathrm{I}\left(\mathbf{X};\mathbf{Z}\mid\mathbf{U}\right) in IB quantifies the redundant information.

In the sequel, we provide the parameterized variational approximation of information quantities, and then study the impact of the information complexity I(𝐗;𝐙)\mathrm{I}\left(\mathbf{X};\mathbf{Z}\right) on the information leakage for an unknown sensitive variable.

2.1. Variational Approximation of Information Measures

Let Q𝐔𝐙:𝒵𝒫(𝒰)Q_{\mathbf{U}\mid\mathbf{Z}}\!:\!\mathcal{Z}\!\rightarrow\!\mathcal{P}\!\left(\mathcal{U}\right), Q𝐒𝐙:𝒵𝒫(𝒮)Q_{\mathbf{S}\mid\mathbf{Z}}\!:\!\mathcal{Z}\!\rightarrow\!\mathcal{P}\!\left(\mathcal{S}\right), Q𝐙:𝒵𝒫(𝒵)Q_{\mathbf{Z}}\!:\!\mathcal{Z}\!\rightarrow\!\mathcal{P}\!\left(\mathcal{Z}\right) be variational approximations of the optimal utility decoder distribution P𝐔𝐙P_{\mathbf{U}\mid\mathbf{Z}}, adversary decoder distribution P𝐒𝐙P_{\mathbf{S}\mid\mathbf{Z}}, and latent space distribution P𝐙P_{\mathbf{Z}}, respectively. The common approach is to use deep neural networks (DNNs) to model/parameterized these distributions. Let Pϕ(𝐙𝐗)P_{\bm{\phi}}(\mathbf{Z}\!\!\mid\!\!\mathbf{X}) denote the family of encoding probability distributions P𝐙𝐗P_{\mathbf{Z}\mid\mathbf{X}} over 𝒵\mathcal{Z} for each element of space 𝒳\mathcal{X}, parameterized by the output of a DNN fϕf_{\bm{\phi}} with parameters ϕ\bm{\phi}. Analogously, let P𝜽(𝐔𝐙)P_{\bm{\theta}}(\mathbf{U}\!\!\mid\!\!\mathbf{Z}) and P𝝃(𝐒𝐙)P_{\bm{\xi}}\!\left(\mathbf{S}\!\mid\!\mathbf{Z}\right) denote the corresponding family of decoding probability distributions Q𝐔𝐙Q_{\mathbf{U}\mid\mathbf{Z}} and Q𝐒𝐙Q_{\mathbf{S}\mid\mathbf{Z}}, respectively, parameterized by the output of DNNs g𝜽g_{\bm{\theta}} and g𝝃g_{\bm{\xi}}. Let P𝖣(𝐗)=1Nn=1Nδ(𝐱𝐱n)P_{\mathsf{D}}\!\left(\mathbf{X}\right)\!=\!\frac{1}{N}\sum_{n=1}^{N}\!\delta(\mathbf{x}-\mathbf{x}_{n}), 𝐱n𝒳\mathbf{x}_{n}\in\mathcal{X} denote the empirical data distribution. In this case, Pϕ(𝐗,𝐙)=P𝖣(𝐗)Pϕ(𝐙𝐗)P_{\bm{\phi}}\!\left(\mathbf{X},\mathbf{Z}\right)\!=\!P_{\mathsf{D}}(\mathbf{X})P_{\bm{\phi}}\!\left(\mathbf{Z}\!\mid\!\mathbf{X}\right) denotes our joint inference data distribution, and Pϕ(𝐙)=𝔼P𝖣(𝐗)[Pϕ(𝐙𝐗)]P_{\bm{\phi}}(\mathbf{Z})\!=\!\mathbb{E}_{P_{\mathsf{D}}(\mathbf{X})}\left[P_{\bm{\phi}}(\mathbf{Z}\!\mid\!\mathbf{X})\right] denotes the learned aggregated posterior distribution over latent space 𝒵\mathcal{Z}.

Information Complexity: The information complexity can be decomposed as:

(2) I(𝐗;𝐙)\displaystyle\!\!\mathrm{I}\left(\mathbf{X};\mathbf{Z}\right)\!\!\!\! =\displaystyle= 𝔼P𝐗,𝐙[logP𝐗,𝐙P𝐗P𝐙]=𝔼P𝐗,𝐙[logP𝐙𝐗Q𝐙Q𝐙P𝐙]\displaystyle\!\!\!\!\mathbb{E}_{P_{\mathbf{X},\mathbf{Z}}}\Big{[}\log\frac{P_{\mathbf{X},\mathbf{Z}}}{P_{\mathbf{X}}P_{\mathbf{Z}}}\Big{]}\!=\!\mathbb{E}_{P_{\mathbf{X},\mathbf{Z}}}\Big{[}\log\frac{P_{\mathbf{Z}\mid\mathbf{X}}}{Q_{\mathbf{Z}}}\frac{Q_{\mathbf{Z}}}{P_{\mathbf{Z}}}\Big{]}
=\displaystyle= 𝔼P𝐗[DKL(P𝐙𝐗Q𝐙)]DKL(P𝐙Q𝐙).\displaystyle\!\!\!\mathbb{E}_{P_{\mathbf{X}}}\left[\mathrm{D}_{\mathrm{KL}}\left(P_{\mathbf{Z}\mid\mathbf{X}}\|Q_{\mathbf{Z}}\right)\right]\!-\!\mathrm{D}_{\mathrm{KL}}\left(P_{\mathbf{Z}}\|Q_{\mathbf{Z}}\right).

Where Q𝐙Q_{\mathbf{Z}} is the latent space’s prior.

Therefore, the parameterized variational approximation of information complexity (2) can be recast as:

(3) Iϕ(𝐗;𝐙)DKL(Pϕ(𝐙𝐗)Q𝐙P𝖣(𝐗))DKL(Pϕ(𝐙)Q𝐙).\displaystyle\mathrm{I}_{\bm{\phi}}\!\left(\mathbf{X};\mathbf{Z}\right)\coloneqq\mathrm{D}_{\mathrm{KL}}\!\left(P_{\bm{\phi}}(\mathbf{Z}\!\mid\!\mathbf{X})\,\|\,Q_{\mathbf{Z}}\mid P_{\mathsf{D}}(\mathbf{X})\right)-\mathrm{D}_{\mathrm{KL}}\!\left(P_{\bm{\phi}}(\mathbf{Z})\,\|\,Q_{\mathbf{Z}}\right).

The optimal prior Q𝐙Q^{\ast}_{\mathbf{Z}} minimizing the information complexity is Q𝐙(𝐳)=𝔼P𝖣(𝐗)[Pϕ(𝐙𝐗=𝐱)]Q^{\ast}_{\mathbf{Z}}(\mathbf{z})=\mathbb{E}_{P_{\mathsf{D}}(\mathbf{X})}\left[P_{\bm{\phi}}\left(\mathbf{Z}\mid\mathbf{X}=\mathbf{x}\right)\right]; however, it may potentially lead to over-fitting. A critical challenge is to guarantee that the learned aggregated posterior distribution Pϕ(𝐙)P_{\bm{\phi}}(\mathbf{Z}) conforms well to thd prior Q𝐙Q_{\mathbf{Z}} (Kingma et al., 2016; Rezende and Mohamed, 2015; Rosca et al., 2018; Tomczak and Welling, 2018; Bauer and Mnih, 2019). We can cope with this issue by employing a more expressive form for Q𝐙Q_{\mathbf{Z}}, which would allow us to provide a good fit of an arbitrary space for 𝒵\mathcal{Z}, at the expense of additional computational complexity.

Information Utility: The parameterized variational approximation of MI between the released representation 𝐙\mathbf{Z} and the utility attribute 𝐔\mathbf{U} can be recast as:

Iϕ,𝜽(𝐔;𝐙)𝔼P𝐔,𝐗[𝔼Pϕ(𝐙𝐗)[logP𝜽(𝐔𝐙)P𝐔P𝜽(𝐔)P𝜽(𝐔)]]=𝔼P𝐔,𝐗[𝔼Pϕ(𝐙𝐗)[logP𝜽(𝐔𝐙)]]𝔼P𝐔[logP𝐔P𝜽(𝐔)]+𝔼P𝐔[logP𝜽(𝐔)]=Hϕ,𝜽(𝐔𝐙)DKL(P𝐔P𝜽(𝐔))+H(P𝐔P𝜽(𝐔))Hϕ,𝜽(𝐔𝐙)PredictionFidelityDKL(P𝐔P𝜽(𝐔))DistributionDiscrepancyLoss,\begin{split}\mathrm{I}_{\bm{\phi},\bm{\theta}}&\!\left(\mathbf{U};\mathbf{Z}\right)\!\;\coloneqq\\ &\mathbb{E}_{P_{\mathbf{U},\mathbf{X}}}\Big{[}\mathbb{E}_{P_{\bm{\phi}}\left(\mathbf{Z}\mid\mathbf{X}\right)}\Big{[}\log\frac{P_{\bm{\theta}}\!\left(\mathbf{U}\!\mid\!\mathbf{Z}\right)}{P_{\mathbf{U}}}\cdot\frac{P_{\bm{\theta}}(\mathbf{U})}{P_{\bm{\theta}}(\mathbf{U})}\Big{]}\Big{]}=\\ &\mathbb{E}_{P_{\mathbf{U},\mathbf{X}}}\left[\mathbb{E}_{P_{\bm{\phi}}\left(\mathbf{Z}\mid\mathbf{X}\right)}\left[\log P_{\bm{\theta}}\!\left(\mathbf{U}\!\mid\!\mathbf{Z}\right)\right]\right]\\ &\qquad-\mathbb{E}_{P_{\mathbf{U}}}\big{[}\log\frac{P_{\mathbf{U}}}{P_{\bm{\theta}}(\mathbf{U})}\big{]}+\mathbb{E}_{P_{\mathbf{U}}}\left[\log P_{\bm{\theta}}(\mathbf{U})\right]\\ &=-\mathrm{H}_{\bm{\phi},\bm{\theta}}\left(\mathbf{U}\!\mid\!\mathbf{Z}\right)-\mathrm{D}_{\mathrm{KL}}\left(P_{\mathbf{U}}\|P_{\bm{\theta}}(\mathbf{U})\right)+\mathrm{H}\left(P_{\mathbf{U}}\|P_{\bm{\theta}}(\mathbf{U})\right)\\ &\geq\underbrace{-\mathrm{H}_{\bm{\phi},\bm{\theta}}\left(\mathbf{U}\!\mid\!\mathbf{Z}\right)}_{\mathrm{Prediction~{}Fidelity}}-\underbrace{\mathrm{D}_{\mathrm{KL}}\left(P_{\mathbf{U}}\,\|\,P_{\bm{\theta}}(\mathbf{U})\right)}_{\mathrm{Distribution~{}Discrepancy~{}Loss}},\end{split}

where Hϕ,𝜽(𝐔𝐙)=𝔼P𝐔,𝐗[𝔼Pϕ(𝐙𝐗)[logP𝜽(𝐔𝐙)]]\mathrm{H}_{\bm{\phi},\bm{\theta}}\!\left(\mathbf{U}\!\mid\!\mathbf{Z}\right)\!=\!-\mathbb{E}_{P_{\mathbf{U},\mathbf{X}}}\!\left[\mathbb{E}_{P_{\bm{\phi}}\left(\mathbf{Z}\mid\mathbf{X}\right)}\!\left[\log P_{\bm{\theta}}\!\left(\mathbf{U}\!\mid\!\mathbf{Z}\right)\right]\right] represents the parameterized decoder uncertainty, and in the last line we use the positivity of the cross-entropy H(P𝐔P𝜽(𝐔))\mathrm{H}\left(P_{\mathbf{U}}\|P_{\bm{\theta}}(\mathbf{U})\right).

3. Learning Model

System Designer. Given independent and identically distributed (i.i.d.) training samples {(𝐮n,𝐱n)}n=1N\{\left(\mathbf{u}_{n},\mathbf{x}_{n}\right)\}_{n=1}^{N} 𝒰×𝒳\subseteq\mathcal{U}\times\mathcal{X}, and using stochastic gradient descent (SGD)-type approach, DNNs fϕf_{\bm{\phi}}, g𝜽g_{\bm{\theta}}, D𝜼D_{\bm{\eta}}, and D𝝎D_{\bm{\omega}} are trained together to maximize a Monte-Carlo approximation of the deep variational IB functional over parameters ϕ\bm{\phi}, 𝜽\bm{\theta}, 𝜼\bm{\eta}, and 𝝎\bm{\omega} (Fig. 2). Backpropagation through random samples from the posterior distribution Pϕ(𝐙𝐗)P_{\bm{\phi}}(\mathbf{Z}\!\!\mid\!\!\mathbf{X})is required in our framework, which is a challenge since backpropagation cannot flow via random nodes; to overcome this hurdle, we apply the reparameterization approach (Kingma and Welling, 2014)).

The inferred posterior distribution is typically assumed to be a multi-variate Gaussian with a diagonal co-variance, i.e., Pϕ(𝐙𝐱)=𝒩(𝝁ϕ(𝐱),P_{\bm{\phi}}(\mathbf{Z}\!\mid\!\mathbf{x})=\mathcal{N}\big{(}\bm{\mu}_{\bm{\phi}}(\mathbf{x}), 𝖽𝗂𝖺𝗀(𝝈ϕ(𝐱)))\mathsf{diag}(\bm{\sigma}_{\bm{\phi}}(\mathbf{x}))\big{)}. Suppose 𝒵=d\mathcal{Z}=\mathbb{R}^{d}. We first sample a random variable 𝓔\bm{\mathcal{E}} i.i.d. from 𝒩(𝟎,𝐈d)\mathcal{N}\!\left(\bm{0},\mathbf{I}_{d}\right), then given data sample 𝐱𝒳\mathbf{x}\in\mathcal{X}, we generate the sample 𝐳=𝝁ϕ(𝐱)+𝝈ϕ(𝐱)𝜺\mathbf{z}=\bm{\mu}_{\bm{\phi}}(\mathbf{x})+\bm{\sigma}_{\bm{\phi}}(\mathbf{x})\odot\bm{\varepsilon}, where \odot is the element-wise (Hadamard) product. The latent space prior distribution is typically considered as a fixed dd-dimensional standard isotropic multi-variate Gaussian, i.e., Q𝐙=𝒩(𝟎,𝐈d)Q_{\mathbf{Z}}=\mathcal{N}\!\left(\bm{0},\mathbf{I}_{d}\right). For this simple choice, the information complexity upper bound

𝔼Pϕ(𝐗,𝐙)[logPϕ(𝐙𝐗)Q𝐙]=𝔼P𝖣(𝐗)[DKL(Pϕ(𝐙𝐗)Q𝐙)]\mathbb{E}_{P_{\bm{\phi}}(\mathbf{X},\mathbf{Z})}[\log\frac{P_{\bm{\phi}}\left(\mathbf{Z}\mid\mathbf{X}\right)}{Q_{\mathbf{Z}}}]=\mathbb{E}_{P_{\mathsf{D}}(\mathbf{X})}\left[\mathrm{D}_{\mathrm{KL}}\left(P_{\bm{\phi}}\!\left(\mathbf{Z}\!\mid\!\mathbf{X}\right)\|Q_{\mathbf{Z}}\right)\right]

has a closed-form expression, which reads as:

2DKL(Pϕ(𝐙𝐗=𝐱)Q𝐙)=𝝁ϕ(𝐱)22+d+i=1d(𝝈ϕ(𝐱)ilog𝝈ϕ(𝐱)i)2\,\mathrm{D}_{\mathrm{KL}}\left(P_{\bm{\phi}}\,\left(\mathbf{Z}\,\mid\,\mathbf{X}=\mathbf{x}\right)\|\right.\left.Q_{\mathbf{Z}}\right)={\|\bm{\mu}_{\bm{\phi}}(\mathbf{x})\|}_{2}^{2}+d+\sum_{i=1}^{d}(\bm{\sigma}_{\bm{\phi}}(\mathbf{x})_{i}-\log\bm{\sigma}_{\bm{\phi}}(\mathbf{x})_{i})

The KL\mathrm{KL}-divergences in (3) and (2.1) can be estimated using the density-ratio trick (Nguyen et al., 2010; Sugiyama et al., 2012), utilized in the GAN framework to directly match the data and generated model distributions. The trick is to express two distributions as conditional distributions, conditioned on a label C{0,1}C\in\{0,1\}, and reduce the task to binary classification. The key point is that we can estimate the KL-divergence by estimating the ratio of two distributions without modeling each distribution explicitly.

Consider DKL(Pϕ(𝐙)Q𝐙)=𝔼Pϕ(𝐙)[logPϕ(𝐙)Q𝐙]\mathrm{D}_{\mathrm{KL}}\!\left(P_{\bm{\phi}}(\mathbf{Z})\,\|\,Q_{\mathbf{Z}}\right)=\mathbb{E}_{P_{\bm{\phi}}(\mathbf{Z})}[\log\frac{P_{\bm{\phi}}(\mathbf{Z})}{Q_{\mathbf{Z}}}]. We now define ρ𝐙(𝐳c)\rho_{\mathbf{Z}}(\mathbf{z}\!\mid\!c) as ρ𝐙(𝐳c=1)=Pϕ(𝐙)\rho_{\mathbf{Z}}(\mathbf{z}\!\mid\!c\!=\!1)\!=\!P_{\bm{\phi}}(\mathbf{Z}), ρ𝐙(𝐳c=0)=Q𝐙\rho_{\mathbf{Z}}(\mathbf{z}\!\mid\!c\!=\!0)=Q_{\mathbf{Z}}. Suppose that a perfect binary classifier (discriminator) D𝜼(𝐳)D_{\bm{\eta}}(\mathbf{z}), with parameters 𝜼\bm{\eta}, is trained to associate the label c=1c=1 to samples from distribution Pϕ(𝐙)P_{\bm{\phi}}(\mathbf{Z}) and the label c=0c=0 to samples from Q𝐙Q_{\mathbf{Z}}. Using the Bayes’ rule and assuming that the marginal class probabilities are equal, i.e., ρ(c=1)=ρ(c=0)\rho(c=1)=\rho(c=0), the density ratio can be expressed as:

Pϕ(𝐙=𝐳)Q𝐙(𝐳)=ρ𝐙(𝐳c=1)ρ𝐙(𝐳c=0)=ρ𝐙(c=1𝐳)ρ𝐙(c=0𝐳)D𝜼(𝐳)1D𝜼(𝐳).\frac{P_{\bm{\phi}}(\mathbf{Z}=\mathbf{z})}{Q_{\mathbf{Z}}(\mathbf{z})}\!=\!\frac{\rho_{\mathbf{Z}}(\mathbf{z}\mid c=1)}{\rho_{\mathbf{Z}}(\mathbf{z}\mid c=0)}\!=\!\frac{\rho_{\mathbf{Z}}(c=1\mid\mathbf{z})}{\rho_{\mathbf{Z}}(c=0\mid\mathbf{z})}\!\approx\!\frac{D_{\bm{\eta}}(\mathbf{z})}{1-D_{\bm{\eta}}(\mathbf{z})}.

Therefore, given a trained discriminator D𝜼(𝐳)D_{\bm{\eta}}(\mathbf{z}) and MM i.i.d. samples {𝐳m}m=1M\{\mathbf{z}_{m}\}_{m=1}^{M} from Pϕ(𝐙)P_{\bm{\phi}}(\mathbf{Z}), we estimate DKL(Pϕ(𝐙)Q𝐙)\mathrm{D}_{\mathrm{KL}}\!\left(P_{\bm{\phi}}(\mathbf{Z})\,\|\,Q_{\mathbf{Z}}\right) as:

(4) DKL(Pϕ(𝐙)Q𝐙)1Mm=1MlogD𝜼(𝐳m)1D𝜼(𝐳m).\mathrm{D}_{\mathrm{KL}}\!\left(P_{\bm{\phi}}(\mathbf{Z})\,\|\,Q_{\mathbf{Z}}\right)\approx\frac{1}{M}\sum_{m=1}^{M}\log\frac{D_{\bm{\eta}}(\mathbf{z}_{m})}{1-D_{\bm{\eta}}(\mathbf{z}_{m})}.

Our model is trained using alternating block coordinate descend across five steps (See Algorithm 1).

Inferential Adversary: Given the publicly-known encoder ϕ\bm{\phi} and KK i.i.d. samples {(𝐬k,𝐳k)}k=1K𝒮×𝒵\{(\mathbf{s}_{k},\mathbf{z}_{k})\}_{k=1}^{K}\!\subseteq\mathcal{S}\times\mathcal{Z}, the adversary trains an inference network 𝝃\bm{\xi} to minimize H𝝃(𝐒𝐙)\mathrm{H}_{\bm{\xi}}(\mathbf{S}\!\mid\!\mathbf{Z}).

Refer to caption
Figure 2. The training and testing architecture. During training, the data user/owner trains the parameterized networks (ϕ,𝜽,𝜼,𝝎)\left(\bm{\phi},\bm{\theta},\bm{\eta},\bm{\omega}\right). During testing, only the encoder-decoder pair (ϕ,𝜽)(\bm{\phi},\bm{\theta}) is used. The adversary uses the publicly-known (fixed) encoder ϕ\bm{\phi} and a collection of the original dataset, and trains an inference network 𝝃\bm{\xi} to infer attribute 𝐒\mathbf{S} of his interest.
1:Inputs: Training Dataset: {(𝐮n,𝐱n)}n=1N\{\left(\mathbf{u}_{n},\mathbf{x}_{n}\right)\}_{n=1}^{N};     Hyper-Parameter: β\beta;
2:ϕ,𝜽,𝜼,𝝎\bm{\phi},\bm{\theta},\bm{\eta},\bm{\omega}\;\leftarrow Initialize Network Parameters
3:repeat(1) Train the Encoder and Utility Decoder (ϕ,𝜽)\left(\bm{\phi},\bm{\theta}\right)
4:     Sample a mini-batch {𝐮m,𝐱m}m=1MP𝖣(𝐗)P𝐔𝐗\{\mathbf{u}_{m},\mathbf{x}_{m}\}_{m=1}^{M}\sim P_{\mathsf{D}}(\mathbf{X})P_{\mathbf{U}\mid\mathbf{X}}
5:     Compute 𝐳mfϕ(𝐱m),m[M]\mathbf{z}_{m}\sim f_{\bm{\phi}}(\mathbf{x}_{m}),\forall m\in[M]
6:     Back-propagate loss: (ϕ,𝜽)=1Mm=1M(logP𝜽(𝐮m𝐳m)\mathcal{L}\!\left(\bm{\phi},\!\bm{\theta}\right)\!=\!-\frac{1}{M}\!\sum_{m=1}^{M}\!\!\big{(}\log P_{\bm{\theta}}(\mathbf{u}_{m}\!\mid\!\mathbf{z}_{m}) βDKL(Pϕ(𝐳m𝐱m)Q𝐙(𝐳m)))\qquad\qquad\qquad\quad-\beta\,\mathrm{D}_{\mathrm{KL}}\!\left(P_{\bm{\phi}}(\mathbf{z}_{m}\!\mid\!\mathbf{x}_{m})\|Q_{\mathbf{Z}}(\mathbf{z}_{m})\right)\!\big{)} (2) Train the Latent Space Discriminator 𝜼\bm{\eta}
7:     Sample {𝐱m}m=1MP𝖣(𝐗)\{\mathbf{x}_{m}\}_{m=1}^{M}\sim P_{\mathsf{D}}(\mathbf{X})
8:     Sample {𝐳~m}m=1MQ𝐙\{\mathbf{\widetilde{z}}_{m}\}_{m=1}^{M}\sim Q_{\mathbf{Z}}
9:     Compute 𝐳mfϕ(𝐱m),m[M]\mathbf{z}_{m}\sim f_{\bm{\phi}}(\mathbf{x}_{m}),\forall m\in[M]
10:     Back-propagate loss: (𝜼)=βMm=1M(logD𝜼(𝐳m)+log(1D𝜼(𝐳~m)))\mathcal{L}\!\left(\bm{\eta}\right)\!\!=\!\!-\frac{\beta}{M}\!\sum_{m=1}^{M}\!\!\big{(}\!\log\!D_{\bm{\eta}}(\mathbf{z}_{m})\!+\!\log\!\left(1\!-\!D_{\bm{\eta}}(\mathbf{\widetilde{z}}_{m})\right)) (3) Train the Encoder ϕ\bm{\phi} Adversarially
11:     Sample {𝐱m}m=1MP𝖣(𝐗)\{\mathbf{x}_{m}\}_{m=1}^{M}\sim P_{\mathsf{D}}(\mathbf{X})
12:     Compute 𝐳mfϕ(𝐱m),m[M]\mathbf{z}_{m}\sim f_{\bm{\phi}}(\mathbf{x}_{m}),\forall m\in[M]
13:     Back-propagate loss: (ϕ)=βMm=1MlogD𝜼(𝐳m)\mathcal{L}\!\left(\bm{\phi}\right)\!=\!\!\frac{\beta}{M}\!\sum_{m=1}^{M}\!\log D_{\bm{\eta}}(\mathbf{z}_{m})\! (4) Train the Attribute Class Discriminator 𝝎\bm{\omega}
14:     Sample {𝐮m}m=1MP𝐔\{\mathbf{u}_{m}\}_{m=1}^{M}\sim P_{\mathbf{U}}
15:     Sample {𝐳~m}m=1MQ𝐙\{\mathbf{\widetilde{z}}_{m}\}_{m=1}^{M}\sim Q_{\mathbf{Z}}
16:     Compute 𝐮~mg𝜽(𝐳~m),m[M]\mathbf{\widetilde{u}}_{m}\sim g_{\bm{\theta}}\left(\mathbf{\widetilde{z}}_{m}\right),\forall m\in[M]
17:     Back-propagate loss: (𝝎)=1Mm=1M(logD𝝎(𝐮m)+log(1D𝝎(𝐮~m)))\mathcal{L}\!\left(\bm{\omega}\right)\!\!=\!\!-\frac{1}{M}\!\sum_{m=1}^{M}\!\big{(}\!\log\!D_{\bm{\omega}}(\mathbf{u}_{m})\!+\!\log\!\left(1\!-\!D_{\bm{\omega}}(\mathbf{\widetilde{u}}_{m})\right)) (5) Train the Utility Decoder θ\bm{\theta} Adversarially
18:     Sample {𝐳~m}m=1MQ𝐙\{\mathbf{\widetilde{z}}_{m}\}_{m=1}^{M}\sim Q_{\mathbf{Z}}
19:     Compute 𝐮~mg𝜽(𝐳~m),m[M]\mathbf{\widetilde{u}}_{m}\sim g_{\bm{\theta}}\left(\mathbf{\widetilde{z}}_{m}\right),\forall m\in[M]
20:     Back-propagate loss: (𝝎)=1Mm=1Mlog(1D𝝎(𝐮~m))\mathcal{L}\!\left(\bm{\omega}\right)\!=\!\!\frac{1}{M}\sum_{m=1}^{M}\!\log\left(1\!-\!D_{\bm{\omega}}(\mathbf{\widetilde{u}}_{m})\right)
21:until Convergence
22:return ϕ,𝜽,𝜼,𝝎\bm{\phi},\bm{\theta},\bm{\eta},\bm{\omega}
Algorithm 1 Training Algorithm: Data Owner

4. Experiments

In this section, we show the impact of the following factors on the amount of leakage: (i) information complexity regularizer weight β(0,1]\beta\!\in\!(0,1], (ii) released representation dimension dzd_{\mathrm{z}}, (iii) cardinalities of the known utility and unknown sensitive attribute sets, (iv) correlation between the utility and sensitive attributes, and (v) potential bias in a sensitive attribute of adversary’s interest. We conduct experiments on the Colored-MNIST and large-scale CelebA datasets. The Colored-MNIST111Several papers have employed Colored-MNIST dataset; However, they are not unique, and researchers synthesized different versions based on their application. The innovative concept behind our version was influenced from the one used in (Rodríguez-Gálvez et al., 2020). is our modified version of MNIST (LeCun and Cortes, 2010), which is a collection of 70,00070,000 colored digits of size 28×2828\times 28. The digits are randomly colored with red, green, or blue based on two distributions, as explained in the caption of Fig. 4. The CelebA (Liu et al., 2015) dataset contains 202,599202,599 images of size 218×178218\times 178. We used TensorFlow 2.4.12.4.1 (Abadi et al., 2016) with Integrated Keras API. The method details and network architectures are provided in Appendix. A and Appendix B.

Refer to caption
(a)
Refer to caption
(b)
Refer to caption
(c)
Refer to caption
(d)
Refer to caption
(e)
Refer to caption
(f)
Refer to caption
(g)
Refer to caption
(h)
Figure 3. The results on CelebA dataset, considering isotropic Gaussian prior. (First Row): dz=64d_{\mathrm{z}}\!=\!64; (Second Row): dz=128d_{\mathrm{z}}\!=\!128; (Third Row): Estimated information leakage I(𝐒;𝐙)\mathrm{I}(\mathbf{S};\mathbf{Z}) using MINE; (Fourth Row): Estimated useful information I(𝐔;𝐙)\mathrm{I}(\mathbf{U};\mathbf{Z}) using MINE. (First Column): utility task is gender recognition (|𝒰|=2)(|\mathcal{U}|\!=\!2), adversary’s interest is heavy makeup (|𝒮|=2)(|\mathcal{S}|\!=\!2); (Second Column): utility task is emotion (smiling) recognition (|𝒰|=2)(|\mathcal{U}|\!=\!2), adversary’s interest is mouth slightly open (|𝒮|=2)(|\mathcal{S}|\!=\!2).

The first and second rows of Fig. 3 and Fig. 4 depict the trade-off among (i) information complexity, (ii) service provider’s accuracy on utility attribute 𝐔\mathbf{U}, and (iii) adversary’s accuracy on attribute 𝐒\mathbf{S}. The third row depicts the amount of information revealed about 𝐒\mathbf{S}, i.e., I(𝐒;𝐙)\mathrm{I}(\mathbf{S};\mathbf{Z}), for the scenarios considered in the first and second rows, which are estimated using MINE (Belghazi et al., 2018). The fourth row depicts the amount of released information about the utility attribute 𝐔\mathbf{U}, i.e., I(𝐔;𝐙)\mathrm{I}(\mathbf{U};\mathbf{Z}), corresponding to the considered scenarios in the first and second rows, also estimated using MINE. We consider different portions of the datasets available for training adversary’s network, denoted by the ‘data ratio’.

Refer to caption
(a)
Refer to caption
(b)
Refer to caption
(c)
Refer to caption
(d)
Refer to caption
(e)
Refer to caption
(f)
Refer to caption
(g)
Refer to caption
(h)
Refer to caption
(i)
Refer to caption
(j)
Refer to caption
(k)
Refer to caption
(l)
Figure 4. The results on Colored-MNIST dataset, considering isotropic Gaussian prior. (First Row): dz=8d_{\mathrm{z}}\!=\!8; (Second Row): dz=64d_{\mathrm{z}}\!=\!64; (Third Row): estimated information leakage I(𝐒;𝐙)\mathrm{I}(\mathbf{S};\mathbf{Z}) using MINE; (Fourth Row): estimated useful information I(𝐔;𝐙)\mathrm{I}(\mathbf{U};\mathbf{Z}) using MINE. (First Column): utility task is digit recognition (|𝒰|=10)(|\mathcal{U}|\!=\!10), while the adversary’s goal is the digit color (|𝒮|=3)(|\mathcal{S}|\!=\!3), setting PS(𝖱𝖾𝖽)=PS(𝖦𝗋𝖾𝖾𝗇)=PS(𝖡𝗅𝗎𝖾)=13P_{S}(\mathsf{Red})\!=\!P_{S}(\mathsf{Green})\!=\!P_{S}(\mathsf{Blue})\!=\!\frac{1}{3}; (Second Column): utility task is digit recognition (|𝒰|=10)(|\mathcal{U}|\!=\!10), while the adversary’s goal is the digit color, setting PS(𝖱𝖾𝖽)=12P_{S}(\mathsf{Red})\!=\!\frac{1}{2}, PS(𝖦𝗋𝖾𝖾𝗇)=16P_{S}(\mathsf{Green})\!=\!\frac{1}{6}, PS(𝖡𝗅𝗎𝖾)=13P_{S}(\mathsf{Blue})\!=\!\frac{1}{3}; (Third Column): utility task is digit color recognition (|𝒰|=3)(|\mathcal{U}|\!=\!3), while the adversary’s interest is the digit number (|𝒮|=10)(|\mathcal{S}|\!=\!10).

The experiments on CelebA consider the scenarios in which the attributes 𝐔\mathbf{U} and 𝐒\mathbf{S} are correlated, while |𝒰|=|𝒮|=2|\mathcal{U}|=|\mathcal{S}|=2. We provide utility accuracy curves for (i) training set, (ii) validation set, and (iii) test set. As we have argued, there is a direct relationship between information complexity and intrinsic information leakage. Note that, as β\beta increases, the information complexity is reduced, and we observe that this also results in a reduction in the information leakage. We also see that the leakage is further reduced when the dimension of the released representation 𝐙\mathbf{Z}, i.e., dzd_{\mathrm{z}}, is reduced. This forces the data owner to obtain a more succinct representation of the utility variable, removing any extra information.

In the Colored-MNIST experiments, provided that the model eliminates all the redundant information I(𝐗;𝐙𝐔)\mathrm{I}(\mathbf{X};\mathbf{Z}\!\mid\!\mathbf{U}) and leaves only the information about 𝐔\mathbf{U}, we expect the adversary’s performance to be close to ‘random guessing’ since the digit color is independent of its value. We investigate the impact of the cardinality of sets |𝒰||\mathcal{U}| and |𝒮||\mathcal{S}|, as well as possible biases in the distribution of 𝐒\mathbf{S}. The results show that it is possible to reach the same level of accuracy on the utility attribute 𝐔\mathbf{U}, while reducing the intrinsic leakage by increasing the regularizer weight β\beta, or equivalently, by reducing the information complexity Iϕ(𝐗;𝐙)\mathrm{I}_{\bm{\phi}}(\mathbf{X};\mathbf{Z}). An interesting possible scenario is to consider correlated attributes 𝐔\mathbf{U} and 𝐒\mathbf{S} with different cardinality sets 𝒰\mathcal{U} and 𝒮\mathcal{S}. For instance, utility task 𝐔\mathbf{U} is personal identification, while the adversary’s interest 𝐒\mathbf{S} is gender recognition.

5. Conclusion

We studied the variational leakage to address the amount of potential privacy leakage in a supervised representation learning setup. In contrast to the PF and generative adversarial privacy models, we consider the setup in which the adversary’s interest is not known a priori to the data owner. We study the role of information complexity in information leakage about an attribute of an adversary interest. This was addressed by approximating the information quantities using DNNs and experimentally evaluating the model on large-scale image databases. The proposed notion of variational leakage relates the amount of leakage to the minimal sufficient statistics.

References

  • (1)
  • Abadi et al. (2016) Martín Abadi et al. 2016. Tensorflow: A system for large-scale machine learning. In {\{USENIX}\} Symp. on Operating Sys. Design and Impl. ({\{OSDI}\}). 265–283.
  • Andre et al. (2006) Thomas Andre, Marc Antonini, Michel Barlaud, and Robert M Gray. 2006. Entropy-based distortion measure for image coding. In 2006 International Conference on Image Processing. IEEE, 1157–1160.
  • Basciftci et al. (2016) Yuksel Ozan Basciftci, Ye Wang, and Prakash Ishwar. 2016. On privacy-utility tradeoffs for constrained data release mechanisms. In 2016 Information Theory and Applications Workshop (ITA). IEEE, 1–6.
  • Bauer and Mnih (2019) Matthias Bauer and Andriy Mnih. 2019. Resampled priors for variational autoencoders. In The 22nd International Conference on Artificial Intelligence and Statistics. 66–75.
  • Belghazi et al. (2018) Mohamed Ishmael Belghazi, Aristide Baratin, Sai Rajeshwar, Sherjil Ozair, Yoshua Bengio, Aaron Courville, and Devon Hjelm. 2018. Mutual information neural estimation. In International Conference on Machine Learning. 531–540.
  • Cesa-Bianchi and Lugosi (2006) Nicolo Cesa-Bianchi and Gábor Lugosi. 2006. Prediction, learning, and games. Cambridge university press.
  • Courtade and Wesel (2011) Thomas A Courtade and Richard D Wesel. 2011. Multiterminal source coding with an entropy-based distortion measure. In 2011 IEEE International Symposium on Information Theory Proceedings. IEEE, 2040–2044.
  • Harremoës and Tishby (2007) Peter Harremoës and Naftali Tishby. 2007. The information bottleneck revisited or how to choose a good distortion measure. In 2007 IEEE International Symposium on Information Theory. IEEE, 566–570.
  • Hsu et al. (2019) Hsiang Hsu, Shahab Asoodeh, and Flavio P. Calmon. 2019. Obfuscation via Information Density Estimation. In Proceeding of the International Conference on Artificial Intelligence and Statistics (AISTATS).
  • Huang et al. (2017) Chong Huang, Peter Kairouz, Xiao Chen, Lalitha Sankar, and Ram Rajagopal. 2017. Context-aware generative adversarial privacy. Entropy 19, 12 (2017), 656.
  • Issa et al. (2019) Ibrahim Issa, Aaron B Wagner, and Sudeep Kamath. 2019. An operational approach to information leakage. IEEE Transactions on Information Theory 66, 3 (2019), 1625–1657.
  • Kingma and Ba (2014) D. P Kingma and J. Ba. 2014. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014).
  • Kingma et al. (2016) Durk P Kingma, Tim Salimans, Rafal Jozefowicz, Xi Chen, Ilya Sutskever, and Max Welling. 2016. Improved variational inference with inverse autoregressive flow. In Advances in neural information processing systems. 4743–4751.
  • Kingma and Welling (2014) Diederik P Kingma and Max Welling. 2014. Auto-encoding variational bayes. In International Conference on Learning Representations (ICLR).
  • LeCun and Cortes (2010) Yann LeCun and Corinna Cortes. 2010. MNIST handwritten digit database. http://yann.lecun.com/exdb/mnist/. (2010). http://yann.lecun.com/exdb/mnist/
  • Liu et al. (2015) Z. Liu, P. Luo, X. Wang, and X. Tang. 2015. Deep Learning Face Attributes in the Wild. In International Conference on Computer Vision (ICCV).
  • Makhdoumi et al. (2014) Ali Makhdoumi, Salman Salamatian, Nadia Fawaz, and Muriel Médard. 2014. From the information bottleneck to the privacy funnel. In 2014 IEEE Information Theory Workshop (ITW 2014). IEEE, 501–505.
  • Nguyen et al. (2010) XuanLong Nguyen, Martin J Wainwright, and Michael I Jordan. 2010. Estimating divergence functionals and the likelihood ratio by convex risk minimization. IEEE Transactions on Information Theory 56, 11 (2010), 5847–5861.
  • P. Calmon et al. (2015) Flavio P. Calmon, Ali Makhdoumi, and Muriel Médard. 2015. Fundamental limits of perfect privacy. In 2015 IEEE International Symposium on Information Theory (ISIT). IEEE, 1796–1800.
  • Rassouli and Gündüz (2019) Borzoo Rassouli and Deniz Gündüz. 2019. Optimal utility-privacy trade-off with total variation distance as a privacy measure. IEEE Transactions on Information Forensics and Security 15 (2019), 594–603.
  • Rassouli and Gündüz (2021) Borzoo Rassouli and Deniz Gündüz. 2021. On perfect privacy. In to appear in IEEE Journal on Selected Areas in Information Theory (JSAIT). IEEE.
  • Rassouli et al. (2019) Borzoo Rassouli, Fernando E Rosas, and Deniz Gündüz. 2019. Data Disclosure under Perfect Sample Privacy. IEEE Trans. on Inform. Forensics and Security (2019).
  • Razeghi et al. (2020) Behrooz Razeghi, Flavio P. Calmon, Deniz Gündüz, and Slava Voloshynovskiy. 2020. On Perfect Obfuscation: Local Information Geometry Analysis. In 2020 IEEE International Workshop on Information Forensics and Security (WIFS). 1–6.
  • Rezende and Mohamed (2015) Danilo Jimenez Rezende and Shakir Mohamed. 2015. Variational inference with normalizing flows. In Proceedings of the 32nd International Conference on Machine Learning. 1530–1538.
  • Rodríguez-Gálvez et al. (2020) Borja Rodríguez-Gálvez, Ragnar Thobaben, and Mikael Skoglund. 2020. A Variational Approach to Privacy and Fairness. arXiv preprint arXiv:2006.06332 (2020).
  • Rosca et al. (2018) Mihaela Rosca, Balaji Lakshminarayanan, and Shakir Mohamed. 2018. Distribution matching in variational inference. arXiv preprint arXiv:1802.06847 (2018).
  • Sreekumar and Gündüz (2019) Sreejith Sreekumar and Deniz Gündüz. 2019. Optimal Privacy-Utility Trade-off under a Rate Constraint. In 2019 IEEE International Symposium on Information Theory (ISIT). IEEE, 2159–2163.
  • Sugiyama et al. (2012) Masashi Sugiyama, Taiji Suzuki, and Takafumi Kanamori. 2012. Density-ratio matching under the Bregman divergence: A unified framework of density-ratio estimation. Annals of the Institute of Statistical Mathematics 64, 5 (2012), 1009–1044.
  • Tishby et al. (2000) Naftali Tishby, Fernando C Pereira, and William Bialek. 2000. The information bottleneck method. In IEEE Allerton.
  • Tomczak and Welling (2018) Jakub Tomczak and Max Welling. 2018. VAE with a VampPrior. In International Conference on Artificial Intelligence and Statistics. 1214–1223.
  • Tripathy et al. (2019) Ardhendu Tripathy, Ye Wang, and Prakash Ishwar. 2019. Privacy-preserving adversarial networks. In 57th Annual Allerton Conference on Communication, Control, and Computing (Allerton). IEEE, 495–505.

Appendices

Appendix A Training Details

All the experiments in the paper have been carried out with the following structure:

A.0.1. Pre-Training Phase

We utilize this phase to warm-up our model before running the main training Algorithm 1 for the Variational Leakage framework within all experiments. In the warm-up phase, we pre-trained encoder (fϕ)\left(f_{\bm{\phi}}\right) and utility-decoder (g𝜽)\left(g_{\bm{\theta}}\right) together for the few epochs via backpropagation (BP) with the Adam optimizer (Kingma and Ba, 2014). We found out the warm-up stage was helpful for faster convergence. Therefore, we initialize the encoder and the utility-decoder weights with the obtained values rather than random or zero initialization. For each experiment, the hyper-parameters of the learning algorithm in this phase were:

Experiment Dataset Learning Rate Max Iteration Batch Size
Colored-MNIST 0.005 50 1024
(both version)
CelebA 0.0005 100 512

A.0.2. Main Block-wise Training Phase

In contrast to the most DNNs training algorithms, each iteration only has one forward step through the network’s weights and then update weights via BP approach. Our training strategy is block-wise and consists of multiple blocks in the main algorithm loop. At each block, forward and backward steps have been done through the specific path in our model, and then corresponding parameters update based on the block’s output loss path.

Since it was not possible for us to use the Keras API’s default model training function, we implement Algorithm 1 from scratch in the Tensorflow. It is important to remember that we initialize all parameters to zero except for the (ϕ,𝜽)\left(\bm{\phi},\bm{\theta}\right) values which acquired in the previous stage. Furthermore, we set the learning rate of the block (1) in the Algorithm 1, five times larger than other blocks. The hyper-parameters of the Algorithm 1 for each experiment shown in the following table:

Experiment Dataset Learning Rate Max Iteration Batch Size
[blocks (2)-(5)]
Colored-MNIST 0.0001 500 2048
(both version)
CelebA 0.00001 500 1024

Appendix B Network Architectures

B.0.1. MI Estimation

For all experiments in this paper, we report estimation of MI between the released representation and sensitive attribute, i.e., I(𝐒;𝐙)\mathrm{I}\left(\mathbf{S};\mathbf{Z}\right), as well as the MI between the released representation and utility attribute, i.e., I(𝐔;𝐙)\mathrm{I}\left(\mathbf{U};\mathbf{Z}\right). To estimate MI, we employed the MINE model (Belghazi et al., 2018). The architecture of the model is depicted in Table. 1.

MINE    I(𝐔;𝐙)\mathrm{I}\left(\mathbf{U};\mathbf{Z}\right)
Input  𝐳dz\mathbf{z}\in\mathbb{R}^{d_{z}} Code; 𝐮|𝒰|\mathbf{u}\in\mathbb{R}^{|\mathcal{U}|}
x = Concatenate([z, u])
FC(100), ELU
FC(100), ELU
FC(100), ELU
FC(1)
Table 1. The architecture of the MINE network.

B.0.2. Colored-MNIST

In the Colored-MNIST experiment, we had two versions for data utility and privacy leakage evaluation. In the first version, we set the utility data to the class’ label of the input image and consider the color of the input image as sensitive data, and for the second one, we did vice versa. It is worth mentioning that both balanced and unbalanced Colored-MNIST datasets are applied with the same architecture given in Table 2.

Encoder fϕf_{\bm{\phi}}
Input  𝐱28×28×3\mathbf{x}\in\mathbb{R}^{28\times 28\times 3} Color Image
Conv(64,5,2), BN, LeakyReLU
Conv(128,5,2), BN, LeakyReLU
Flatten
FC(dz×4d_{z}\times 4), BN, Tanh
μ\mu: FC(dzd_{z}), σ\sigma: FC(dzd_{z})
zz=SamplingWithReparameterizationTrick[μ\mu,σ\sigma]
Utility Decoder g𝜽g_{\bm{\theta}}
Input  𝐳dz\mathbf{z}\in\mathbb{R}^{d_{z}} Code
FC(dz×4d_{z}\times 4), BN, LeakyReLU
FC(|𝒰||\mathcal{U}|), SOFTMAX
Latent Space Discriminator D𝜼D_{\bm{\eta}}
Input  𝐳dz\mathbf{z}\in\mathbb{R}^{d_{z}} Code
FC(512512), BN, LeakyReLU
FC(256256), BN, LeakyReLU
FC(11), Sigmoid
Utility Attribute Class Discriminator D𝝎D_{\bm{\omega}}
Input  𝐮|𝒰|\mathbf{u}\in\mathbb{R}^{|\mathcal{U}|}
FC(|𝒰|×8|\mathcal{U}|\times 8), BN, LeakyReLU
FC(|𝒰|×8|\mathcal{U}|\times 8), BN, LeakyReLU
FC(11), Sigmoid
Table 2. The architecture of DNNs used in for the Colored-MNIST experiments.

B.0.3. CelebA

In this experiment, we considered three scenarios for data utility and privacy leakage evaluation, as shown in Table 3. Note that all of the utility and sensitive attributes are binary. The architecture of the networks are presented in Table 4.

Scenario Number Utility Attribute Sensitive Attribute
1 Gender Heavy Makeup
2 Mouth Slightly Open Smiling
3 Gender Blond Hair
Table 3. Scenarios considered for CelebA experiments.
Encoder fϕf_{\bm{\phi}}
Input  𝐱64×64×3\mathbf{x}\in\mathbb{R}^{64\times 64\times 3} Color Image
Conv(16,3,2), BN, LeakyReLU
Conv(32,3,2), BN, LeakyReLU
Conv(64,3,2), BN, LeakyReLU
Conv(128,3,2), BN, LeakyReLU
Conv(256,3,2), BN, LeakyReLU
Flatten
FC(dz×4d_{z}\times 4), BN, Tanh
μ\mu: FC(dzd_{z}), σ\sigma: FC(dzd_{z})
zz=SamplingWithReparameterizationTrick[μ\mu,σ\sigma]
Utility Decoder g𝜽g_{\bm{\theta}}
Input  𝐳dz\mathbf{z}\in\mathbb{R}^{d_{z}} Code
FC(dzd_{z}), BN, LeakyReLU
FC(|𝒰||\mathcal{U}|), SOFTMAX
Latent Space Discriminator D𝜼D_{\bm{\eta}}
Input  𝐳dz\mathbf{z}\in\mathbb{R}^{d_{z}} Code
FC(512512), BN, LeakyReLU
FC(256256), BN, LeakyReLU
FC(11), Sigmoid
Utility Attribute Class Discriminator D𝝎D_{\bm{\omega}}
Input  𝐮|𝒰|\mathbf{u}\in\mathbb{R}^{|\mathcal{U}|}
FC(|𝒰|×4|\mathcal{U}|\times 4), BN, LeakyReLU
FC(|𝒰||\mathcal{U}|), BN, LeakyReLU
FC(11), Sigmoid
Table 4. The architecture of networks for the CelebA experiments.
Refer to caption
(a)
Refer to caption
(b)
Refer to caption
(c)
Figure 5. The three sub-modules that make up the main network.

Appendix C Implementation Overview

Fig. 6 demonstrates all sub networks of the proposed framework that attached together and their parameters learned in the Algorithm 1.

Refer to caption
Figure 6. Complete model structure in the training phase. The above model defines for Colored-MNIST dataset with the number of classes as utility attributes and digit colors as sensitive attributes. Also, the encoder output is set to 16 neurons. (The adversary model is not part of the data owner training algorithm)

However, we did not define and save our model in this form because of technical reasons to efficiently implement the training algorithm. As shown in Algorithm 1, the main loop consists of five blocks where only some networks are used in the forward phase, and mostly one of them would update their parameters via BP in each block. Therefore, we shattered the model into three sub-modules in the training stage for simplicity and performance. Fig. 5 shows the corresponding sub-modules of Fig. 6, which are used in our implementation. During training, all of the sub-module (a) parameters, call "autoencoder part" would update with BP after each forward step. For the (b) and (c) sub-modules, only the parameters of one network are updated when the corresponding error function values backpropagate, and we freeze the other networks parameters in the sub-module. For example, block (3) of Algorithm 1 is related to the (b) sub-module, but at the BP step, the latent space discriminator is frozen to prevent its parameters from updating. This procedure is vice versa for module (b) at block (2).

It should be mentioned that during our experiments, we found out that before running our main algorithm, it is beneficial to pre-train the autoencoder sub-module since we need to sample from the latent space, which uses in other parts of the main model during training. We justify this by mentioning that sampling meaningful data rather than random ones from latent variables from the beginning of learning helps the model to converge better and faster in comparison with starting Algorithm 1 with a randomly initiated autoencoder.