TableGAN-MCA: Evaluating Membership Collisions of GAN-Synthesized Tabular Data Releasing
Abstract.
Generative Adversarial Networks (GAN)-synthesized table publishing lets people privately learn insights without access to the private table. However, existing studies on Membership Inference (MI) Attacks show promising results on disclosing membership of training datasets of GAN-synthesized tables. Different from those works focusing on discovering membership of a given data point, in this paper, we propose a novel Membership Collision Attack against GANs (TableGAN-MCA), which allows an adversary given only synthetic entries randomly sampled from a black-box generator to recover partial GAN training data. Namely, a GAN-synthesized table immune to state-of-the-art MI attacks is vulnerable to the TableGAN-MCA. The success of TableGAN-MCA is boosted by an observation that GAN-synthesized tables potentially collide with the training data of the generator.
Our experimental evaluations on TableGAN-MCA have five main findings. First, TableGAN-MCA has a satisfying training data recovery rate on three commonly used real-world datasets against four generative models. Second, factors, including the size of GAN training data, GAN training epochs and the number of synthetic samples available to the adversary, are positively correlated to the success of TableGAN-MCA. Third, highly frequent data points have high risks of being recovered by TableGAN-MCA. Fourth, some unique data are exposed to unexpected high recovery risks in TableGAN-MCA, which may attribute to GAN’s generalization. Fifth, as expected, differential privacy, without the consideration of the correlations between features, does not show commendable mitigation effect against the TableGAN-MCA. Finally, we propose two mitigation methods and show promising privacy and utility trade-offs when protecting against TableGAN-MCA.
1. Introduction
Big data have emerged as valuable resources that allow companies, researchers and governments to enhance decision making, insight discovery and process optimization. However, sharing sensitive datasets without violating individual’s privacy is a long-standing challenge. For example, in , DeepMind was accused of an illegal acquisition of personal medical records of million patients for developing a kidney injuries diagnosing application (Online., 2017). To analyze those sensitive data in a privacy-preserving manner, ideally, we need a trusted third party that collects and processes raw data, and then releases a sanitized version of data trading off privacy and utility through web queries (see the paradigm shown in Fig. 1).

However, state-of-the-art solutions for releasing the sanitized data achieving trade-offs between utility and privacy are vulnerable to privacy inference attacks. For example, de-identification (removing unique identifiers for all data entries) is susceptible to linkage attacks (Narayanan and Shmatikov, 2008). Anonymization (Sweeney, 2002; Machanavajjhala et al., 2007; Li et al., 2007) suffers from background information attacks. Other synthetic dataset publishing mechanisms, such as NetMechanism (Blum et al., 2013), Iterative Construction (Gupta et al., 2012, 2013; Hardt and Rothblum, 2010; Hardt et al., 2012), are tailored for relatively small datasets (Dwork et al., 2014). More recently, Generative Networks, including Generative Adversarial Networks (GANs) (Goodfellow et al., 2014) and Variational Autoencoders (VAEs) (Kingma and Welling, 2013), produce synthetic data that achieve enhanced privacy and utility trade-offs. Such synthetic data conceal the detailed (privacy) of the raw data while keeping statistics similarity (Xu et al., 2019; Park et al., 2018). Nevertheless, recent works (Park et al., 2018; Hayes et al., 2020; Hilprecht et al., 2019; Chen et al., 2020; Stadler et al., [n.d.]) show the risk of membership disclosure (i.e., inferring whether a given data point belongs to the training dataset) against synthesized data by attacking generator APIs. They propose various Membership Inference Attacks (MIAs) against published generative models to disclose the membership information of training data.
To further explore the privacy disclosure risks of the GAN-synthesized tabular data, different from existing MIAs against generative models (Park et al., 2018; Hayes et al., 2020; Hilprecht et al., 2019; Chen et al., 2020; Stadler et al., [n.d.]), we propose a novel attack model, named Membership Collision Attack against GAN-synthesized Tables (TableGAN-MCA). Specifically, we reconstruct a proportion of actual training data from the published synthetic table with high confidence by inferring the membership collisions (substantiated in Section 3.1). Hence, TableGAN-MCA brings a novel privacy problem: training data exposure when analyzing published synthetic tabular data. In addition, TableGAN-MCA only queries a black-box generator (of the GAN) for synthetic data, which is similar to the most strict threat model introduced in the recent work - GAN-Leaks (Chen et al., 2020). We conceptualize the differences among recent works in Table 1.
Motivation. Our work is motivated by two observations in GAN-synthesized table (low-dimensional data) releasing.
-
•
Observation 1. Generated synthetic tables overlap with GAN’s training data (as the intersection illustrated in Fig. 2). For instance, in the Adult dataset, a synthetic dataset collides with the GAN’s training dataset by ( entries). Clearly, such an overlap brings severe privacy breaches if adversaries could locate the intersection. In the remainder of this paper, we call the overlap/intersection membership collision.
-
•
Observation 2. In the GAN-synthesized tabular data, membership collisions and data frequency are positively correlated (substantiated in Fig. 3). However, it is rare to trigger sample collisions in high-dimensional data, such as image synthesis, due to the curse of dimensionality. Thus, the distribution of tabular data with relatively small dimension brings additional privacy risks than that of image synthesis.
To perform the proposed TableGAN-MCA, we leverage shadow models (Shokri et al., 2017) to learn the patterns behind the collision (Observation 1) while taking the density of each synthetic data by counting its sample frequency in synthetic distribution (Observation 2) as additional feature when training the attack model. TableGAN-MCA shows promising results on commonly used real-world datasets, including Adult, Lawschool and Compas. For instance, TableGAN-MCA recovers , , of actual members released with the GAN-synthesized tabular data with approximately 80% confidence for Adult, Lawschool and Compas, respectively. Our results show that a well-trained GAN, robust to the MIAs proposed in (Park et al., 2018; Hayes et al., 2020; Chen et al., 2020), is still vulnerable to TableGAN-MCA.

In summary, our main contributions are as follows:
-
•
We propose a novel membership collision attack against GAN-synthesized tabular data publishing, named TableGAN-MCA, which can reinstate partial training data with high confidence. TableGAN-MCA exploits the weaknesses of GAN synthesis observed on low-dimensional data, i.e., GAN-synthesized data collide with its training data, and members (in the colliding member set) occur more frequently than non-members.
-
•
We extensively evaluate our proposed attacks on three commonly used real-world datasets, including Adult, Lawschool and Compas against four generative models, including TVAE (Xu et al., 2019), CTGAN (Xu et al., 2019), WGAN-GP (Gulrajani et al., 2017) and WGAN-WC (Arjovsky et al., 2017). Furthermore, we explore the factors that may impact the attack effectiveness, such as the size of GAN training data, GAN training epochs, GAN training data frequencies and the number of synthetic samples available to the attacker.
-
•
We discover that individuals in the training dataset have various risks of privacy leakage under TableGAN-MCA. Additionally, we show that GANs do not memorize those exposed data. Instead, when generalizing the distribution of the training data, GANs may increase or decrease the frequency of some individuals, and hence change their privacy risks.
-
•
We examine the effect of differential privacy (DP) to mitigate TableGAN-MCA. Our empirical results show that differential private generative model training achieves sub-optimal trade-offs against TableGAN-MCA. It is mainly due to the fact that TableGAN-MCA relies more on the common pattern of a distribution (like attribute correlations) which is not the focus of DP. In addition to DP, we propose two mitigation methods, naive defense and improved defense, that mitigate the effect of TableGAN-MCA.
2. Background of Generative Models
Generative Adversarial Networks (GANs) (Goodfellow et al., 2014) and its variants have made great achievements in generating high quality artificial data that mimic the real ones, by modeling the underlying data distribution. It is composed of two neural networks: a discriminator and a generator . It tries to minimize the distance between the real data distribution and the generated (artificial) data distribution by iteratively updating parameters of the networks.
The Wasserstein GAN (WGAN) (Arjovsky et al., 2017) applies Earth Mover (EM) distance under a K-Lipschitz constraint and achieves good performance in generating high fidelity samples. The loss function of the discriminator and the generator are as follows:
(1) |
(2) |
In this work, we use its weight clipping version (WGAN-WC) (Arjovsky et al., 2017), Gradient Penalty version (WGAN-GP) (Gulrajani et al., 2017) and CTGAN (state-of-the-art) (Xu et al., 2019). We also include TVAE from (Xu et al., 2019) for its comparable performance as CTGAN. Following (Xu et al., 2019), all three GANs uses recurrent networks in the generator. For categorical features, we use the gumble-softmax activation in the output of the generator. For numerical features, we use the sigmoid or the tanh activation in the output of the generator based on value range. The architecture and parameters of GANs are broken down in Appendix A.
3. Problem Formulation
In this section, we formulate our membership collisions problem, followed by the description of the threat model according to adversary’s goals, capabilities and background knowledge. We introduce all the notations used throughout the paper in Table 2.
Symbol | Description | Symbol | Description |
Private training dataset | Test dataset | ||
Released synthetic dataset | Shadow dataset | ||
Training data distribution | Generator oracle | ||
Prior Gaussian distribution | Indicator function | ||
Generated data distribution | A data point | ||
colliding member set | Adversary | ||
Number of synthetic copies | Attack classifier |
3.1. Membership Collision Problem
We let be a training set sampled from an implicit data distribution . Each private entry takes the form as , where represents the features and represents the class label. A data release mechanism GAN trains on the training set and outputs a well learned generator . Generator is a deterministic function that maps a prior distribution, i.e., Gaussian distribution , to the generated distribution that mimic real distribution . Then, a synthetic dataset is published and serves as a sanitized version of . We formalize the membership collisions as : a published synthetic datasets collide with its training set and result in a colliding member set . Notice that a data point result in . Similarly, a synthetic data point result in .
We aim to study how much an adversary increases its ability to assert whether a synthetic data point belongs to the colliding member set by estimating the generated distribution via the published synthetic dataset . Formally,
Definition 3.1 (Membership Collision Attack).
Given a synthetic dataset produced by a generative model that contains a colliding member set and an attack algorithm that outputs if it outputs the synthetic data , we say the generative model is subject to membership collision inference attack if there exists an entry such that
(3) |
where is a non-negligible value.
In this work, we consider that the prior advantage of the attacker is random guess, that is, . Thus, we evaluate the posterior advantage of the attacker thereafter.
3.2. Threat Model
In the context of GAN-synthesized data sharing, adversaries are external parties that wish to learn the statistics of the sensitive dataset by querying data owners or curators. In existing MIAs against GANs (Park et al., 2018; Hayes et al., 2020; Hilprecht et al., 2019; Chen et al., 2020), the adversary’s knowledge is: (1) having only limited synthetic data, (2) accessing a black-box generator API (unlimited synthetic data), (3) accessing a black-box generator plus a discriminator oracle, (4) accessing a white-box GAN. Our study focus on the most strict attack model: (1) and (2) (which is similar to the threat model in MC (Hilprecht et al., 2019) and “Full Black-box Generator” assumption in GAN-Leaks (Chen et al., 2020)). The attacker does not know the priori of the model’s structure, including meta-parameters, training data and any target data to infer membership. In TableGAN-MCA, the adversary’s goal is to recover the value of some members of the training set from the published synthetic datasets that may unintentionally contain colliding members. In this paper, we evaluate TableGAN-MCA under two threat models:
Attack model (1): accessible to limited synthetic data. We assume the adversary has one copy of synthetic dataset following , of size .
Attack model (2): accessible to unlimited synthetic data. We assume the adversary has ( is a positive integer) synthetic copies , each of which has size .

The x-axis represents all possibilities of and y-axis represents log of the number of eligible data points.
4. Membership Recovery Framework against GAN-synthesized Tables
In this section, we propose a membership indicator for inferring membership collisions from the statistics of the published table. Based on the membership indicator, we propose the TableGAN-MCA to recover the value of the training set of GAN-synthesized tables in the black-box setting.
4.1. Membership Indicator
The membership indicator is triggered by two observations. First, the released GAN-synthesized tables often overlap the training dataset of the GAN model. Second, such synthetic data points appearing frequently in the published GAN-synthesized data are more likely to be the colliding member of the training dataset. That is, . Fig. 3 depicts the observations from three datasets used in this paper, where we count the numbers of members and non-members, given numbers of appearance of the data points in the released synthetic tables. In Fig. 3 (left), approximately of synthetic data with a sampled frequency of more than three are colliding members. Conversely, almost unique synthetic data are non-colliding members in the Adult dataset. Thus, sample frequency is highly correlated with membership collisions and can be treated as an indicator to indicate membership. Formally, we estimated the membership indicator by the following equation.
(4) |
where an indicator function outputs if its argument is true, is the synthetic datasets available to the adversary following , of size .
To date, the adversary can launch a data reconstruction attack by setting a threshold for the value of a membership collisions indicator of Eq. (4), similar to (Chen et al., 2020). The adversary then claims that the synthetic data, having collisions indicators greater than a given threshold, are the recovered data. However, choosing an optimal threshold is a non-trivial task for an adversary without background knowledge about training data except the published synthetic data. To deal with it, we additionally leverage shadow model techniques (Shokri et al., 2017) to enhance the knowledge of adversaries to construct a robust TableGAN-MCA framework.
4.2. TableGAN-MCA

In a nutshell, TableGAN-MCA combines the membership collisions indicator and the shadow models (Shokri et al., 2017) to train an attack model to learn the relation between membership collisions (labels) and indicator values (features) in released GAN-synthesized tables. Fig. 4 depicts the framework of TableGAN-MCA and Alg. 1 shows the detailed implementation. Each step in Alg. 1 corresponds to the step index in Fig. 4. In summary, steps 2, 3, 4 and 5 train an attack classifier by giving synthetic data. Steps 1 and 6 infer membership collisions to recover training data.
In Steps and , represents estimated sample frequency following from Eq. 4. They are concatenated (“”) to (Step 1) and (Step 4) as an extra feature.
In Step 4, a label function is required to claim membership collisions in shadow datasets. For a shadow dataset such that , a membership collisions label for each data will be .
In Step , attack model outputs the predicted probability about whether a synthetic data is colliding member. Adversaries then expose a data set that with high prediction scores.
For attack model (unlimited synthetic data) such that , the adversary repeat the Step1 to Step 4 times and gets labeled shadow datasets such that each of them with size . Then the adversary concat (“”) all shadow datasets together to train the attack model.
Note that in the worst-case (to the adversary), where the intersection between the training set and the synthetic dataset could be empty, the adversary of TableGAN-MCA cannot recover anything from the private training data. To avoid such a case, we would discretize the synthetic dataset to generalize the range of each feature such that there is a non-empty intersection. In this way, we could (at least) recover coarse-grained information regarding the members within the training data. We show the details of the discretization operation in Section 5.1.
5. Evaluation
In this section, we first introduce the methods of tabular data synthesis, then introduce the evaluation metrics. Next we show the attack performance of TableGAN-MCA as well as the comparisons with recent works.
5.1. Dataset Synthesis
We perform experimental evaluations on three commonly used (Ruoss et al., 2020; Xu et al., 2019; Park et al., 2018; Barenstein, 2019; Chen et al., 2018) real-world tables, Adult (Ronny and Barry, 1996), Lawschool (Richard et al., [n.d.]) and Compas (Jeff et al., 2017).
Adult: The US Adult Census dataset is a repository of entries extracted from 1994 US Census dataset, where entries have complete information. After pre-processing, it remains numerical feature, categorical features and binary label.
Lawschool: This dataset comes from the Law School Admission Council’s National Longitudinal Bar Passage Study. It contains application records for different law schools with individuals. It has numerical features, categorical features and binary label.
Compas: COMPAS recidivism risk score and criminal history data is collected by ProPublica in 2016. After pre-processing, it remains entries with numerical features, categorical features and binary labels.
Note that unlike MIAs attacking classifiers that produce predicted labels with probability, generative models only output synthetic samples. The labels in generated datasets serve as an ordinary feature like other features when training attack models. Therefore, for simplicity, we use the three binary-labeled datasets in our experiments.
Tabular data synthesis. For training generative models, we apply Tabular Variational Autoencoder (TVAE) (Xu et al., 2019), CTGAN (Xu et al., 2019), WGAN-GP (Gulrajani et al., 2017) and WGAN-WC (Arjovsky et al., 2017) for their superior modeling quality in tabular synthesis. To facilitate data synthesis, we have the following additional data pre-processing. (1) We discretize imbalanced and sparse numerical values in given columns to categorical values. (2) We normalize numerical columns into or . (3) We one-hot encode all categorical features (4) We split the dataset into the training set (, records) and test set (, records) (see row 1 and row 2 in Table 3). The training set is used for dataset synthesis and the test set is used for examining the utility of the synthetic data.
Discretization in pre-processing. Features in tabular dataset are either categorical or numerical variables. Unlike pictures, some numerical columns are non-Gaussian distribution, that is, it either has long tails, sparse distribution or multiple modes. Generative models cannot model them well without appropriate pre-processing. To address this issue, we discretize the imbalanced and sparse numerical values to categorical values. In the experiments, such simple discretization in pre-processing exhibits decent performance in generating complex features while keeping original statistics. Note that discretization definitely makes some records of the original dataset share the same values (similar to -anonymity (Sweeney, 2002)). We show the uniqueness of the records after pre-processing in Table 3 (row three and four), where a large proportion of sensitive data points can still be uniquely identified before feeding into generative models.
Adult | Lawsch | Compas | |
of Train () | 31655 | 60215 | 3694 |
of Test () | 13567 | 25807 | 1584 |
5.2. Metrics
5.2.1. Data Utility Metrics
For data utility evaluation, we consider two measurements: machine learning efficacy (models trained on a synthetic dataset and the original dataset provide similar predictions) and distribution fitness (a synthetic dataset is statistically similar to its original dataset in all attributes).
For distribution fitness, we present -way marginals that are approximated by the Empirical Cumulative Distribution Function (ECDF) for each attribute. Having ECDFs of real and synthetic data, we compute attribute-wise Wasserstein distance, i.e., , where and are respective CDFs of real attribute and synthetic attribute (Ramdas et al., [n.d.]). We compare the expected value of ECDFs by .
5.2.2. Attack Performance Metrics
To evaluate the privacy of the released synthetic table, we consider membership collisions privacy, i.e., the TableGAN-MCA effect. We use precision and recall to evaluate the attack performance (following Shokri et al. (Shokri et al., 2017)), since the synthetic dataset that is used to inference has a skewed label distribution. Specifically, precision measures the probability of an entry inferred as a member is indeed the member of the training dataset, denoted as . Intuitively, it implies the confidence of the attacker in guessing positive membership. Recall measures the probability of a member is correctly inferred as a member by the attacker, denoted as . It reflects the percentage of positives exposed in the attack. In evaluation, we report precision and recall by Precision-Recall (PR) curve since it is more informative than ROC-curve under the case of skewed label distribution (Davis and Goadrich, 2006). A higher Area under the PR-curve (AUPRC) implies both higher precision and recall, and thus they are used to compare the attack efficacy.
In addition to the attack precision and recall, we also consider a recovery rate because it reflects what the proportion of training data are being exposed to TableGAN-MCA. Let be recovered training data sets of the attack algorithm . The recovery rate of is defined as below:
(5) |
Note that the recovery rate shares the same numerator as the recall of the attack model but the different denominator ( vs ).
Methods | DT | MLPC | Ada | LR | ||
Adult | Base | 85.39 | 84.11 | 86.28 | 84.74 | 0 |
TVAE | 79.24 | 77.53 | 78.72 | 80.2 | 0.0207 | |
CTGAN | 81.53 | 81.76 | 82.3 | 82.41 | 0.0266 | |
WGANWC | 82.74 | 83.62 | 84.16 | 83.96 | 0.0075 | |
WGANGP | 83.16 | 83.95 | 84.24 | 83.93 | 0.0039 | |
Lawschool | Base | 81.90 | 89.54 | 87.23 | 87.68 | 0 |
TVAE | 79.53 | 85.38 | 85.17 | 85.26 | ||
CTGAN | 76.35 | 80.91 | 81.02 | 81.37 | 0.0283 | |
WGANWC | 77.54 | 80.76 | 80.56 | 80.93 | 0.0073 | |
WGANGP | 80.14 | 86.10 | 86.02 | 86.79 | 0.0047 | |
Compas | Base | 69.89 | 70.58 | 71.65 | 71.46 | 0 |
TVAE | 64.42 | 68.07 | 64.33 | 68.33 | 0.0159 | |
CTGAN | 58.14 | 60.21 | 59.5 | 58.12 | 0.0373 | |
WGANWC | 65.34 | 66.5 | 65.06 | 68.46 | 0.0095 | |
WGANGP | 64.12 | 66.91 | 65.20 | 68.34 | 0.0179 |
5.3. Synthetic Data Utility
We evaluate machine learning efficacy of synthetic data generated by four generative models, CTGAN (Xu et al., 2019), TVAE (Xu et al., 2019), WGAN-GP and WGAN-WC vary four binary classifiers: DecisionTreeClassifier
, MLPClassifier
, AdaBoostClassifier
, and LogisticRegression
(Standard scikit-learn machine learning library, see middle columns in Table 4). We also compare ECDFs using the average of all attribute-wise Wasserstein distance (see the last column in Table 4). All numerical features are min-max scaled to and categorical features are one-hot encoded before feeding into the classifier. For “base”, we trained on the sensitive dataset that is used for data synthesis and test on the real test set (see Table 3). For synthetic data, we trained on a synthetic dataset of the same size as the sensitive dataset and test on the same real test set . To implement CTGAN and TVAE, we directly feed our pre-processed data into the module CTGANSynthesizer
and TVAESynthesizer
of the SDGym
(Carles, 2019) (published code for (Xu et al., 2019)).
According to Table 4, the synthetic dataset generated by WGAN-GP, WGAN-WC, CTGAN and TVAE can greatly restore the prediction ability of the model trained on original dataset. TVAE is least ideal than the others in the Adult dataset. CTGAN is least ideal than the others in the Compas dataset. We will use these learned generative models to perform TableGAN-MCA experiments later.
For marginal fitness, we depict an additional ECDF comparison between real and synthetic Adult, Lawschool and Compas datasets generated by WGAN-GP in Fig. 5. In our experiments, we depict ECDFs of continuous variables (i.e., age, isat) and more complex categorical variables (i.e., hours per week, priors count) since they are more difficult to fit. As can be seen in Fig. 5, the marginals of the synthetic dataset are almost indistinguishable from the original one, thus supporting any statistical queries.



5.4. Attack Performance
5.4.1. Performance Evaluation on TableGAN-MCA
In this section, we evaluate TableGAN-MCA of Alg. 1 on the Adult, Lawschool and Compas datasets. The training and inference data statistics of TableGAN-MCA are presented in Table 5, where positive percentage implies the membership collisions proportion. Both target models and shadow models are WGAN-GP. The attack model is trained on the shadow dataset and tested on the synthetic dataset .
Adult | Lawsch | Compas | |
(Train) | |||
(Inference) | |||
TableGAN-MCA provides a promising attack against the GAN-synthesized tables. We report the PR-curve of the attack model in Fig. 6 when , i.e., . In Fig. 6, PR-curve reflects the trade-off between precision and recall for different probability thresholds . Particularly, after providing the inference data (the released GAN-synthesized tables) to the TableGAN-MCA attack model, we receive a set of probabilities for each record of the test data that predicts whether a record is a member.
As illustrated in Fig. 6, we find that by setting a suitable threshold , the adversary can expose approximate colliding members with confidence over , and for the Adult, Lawschool and Compas datasets, respectively. This means that the adversary significantly increases its ability to assert that these entities are members. Furthermore, when setting confidence to , we have , , positive percentages being exposed, which correspond to , and individual’s sensitive entries in the Adult, Lawschool and Compas datasets, respectively. According to Fig. 6, we list TableGAN-MCA’s recovery rates (Eq.(5)) with different precision configurations in Table 6.

Datasets | () | Precision | Recall | ||
Adult | |||||
Lawsch | |||||
Compas | |||||
Adversary’s knowledge enhances the attack performance of TableGAN-MCA. Fig. 7 reports the PR-curve and AUPRC of TableGAN-MCA when . That is, the adversary has multiple copies of the released synthetic data. In particular, when , the adversary trains independent shadow GAN for each one of and finally obtains a shadow dataset times the size of . In Fig. 7, the performance of TableGAN-MCA is greatly improved by increasing the number of synthetic copies . We also show the PR-curve comparison for the three datasets in Fig. 7. That is, given recall, 10 copies boost the precision from () to (). Given precision, the recall is boosted from () to (). We conclude that more copies of the synthetic data allow the adversary to make better approximation to generated distribution , and thus generate more informative labeled shadow samples to train the attack model.




TableGAN-MCA achieves commendable attack performance even with fewer synthetic queries. When (an adversary queries the target Generator times), TableGAN-MCA achieves AUPRC ( is 0.6697), and recovers 1,409 data points ( is 1565) under precision in the Adult dataset. More details are shown in Appendix B.
The generation quality of the target/victim model positively impacts attack performance. Fig. 8(a) depicts TableGAN-MCA’s performance on four different target/victim models: WGAN-GP, WGAN-WC, CTGAN and TVAE. The shadow model in use is exactly the same as target models. Combining the results of Fig. 8(a) and Table 4, we conclude that target/victim generators with high generation quality often attain high attack performance. For instance, TVAE with the lowest prediction accuracy score in prediction accuracy (see Table 4) also achieves unsatisfactory performance in TableGAN-MCA on the adult dataset. This echoes what CTGAN performs in the Compas dataset. Additionally, we observe that attack performance of TVAE is more sensitive to its generation quality.
The type of shadow models has limited impact on attack performance. Note that the adversary may have no knowledge about the structures and parameters of target/victim generative models. Fig. 8(b) compares the attack performance by using four different shadow models (WGAN-GP, WGAN-WC, CTGAN and TVAE) to attack target WGAN-GP. As can be seen, various shadow model attacks (“wganwc”, “ctgan”, “tvae”) work as well as the identical shadow model attack (“wgangp”). TVAE shadow models perform worst, in large part due to its poor learning ability in the Adult dataset.


The success of TableGAN-MCA is mainly due to the observed collision, the membership collisions indicator and the shadow model in use. In particular, the collision between synthetic data and training set provides the opportunity for recovering training data. The membership collisions indicator, which captures the statistical patterns behind colliding members, guarantees more accurate and informative features for training the attack model. The shadow model in use provides enough labeled data to train the attack model so as to learn from the statistical patterns of the colliding members.
Attack scalability. The key to the success of the TableGAN-MCA is the possibility of collisions between raw training datasets and synthetic datasets. For real-world tabular data, attributes usually have a finite domain range. Hence, the dataset dimension indicates its overall domain range. Namely, low-dimensional tables are more likely to incur sample collisions when the generator creates those synthetic tables. Therefore, TableGAN-MCA discovers additional privacy risks – membership disclosure via collision attacks – for low-dimensional data. TableGAN-MCA potentially fits high-dimensional data if adversaries reduce the data granularity by generalizing attributes. The TableGAN-MCA works since the synthetic data would have a higher chance to collide with the training datasets. In our experiments, TableGAN-MCA achieves 0.871 AUPRC by bucketizing the ”age” attribute in synthetic Adult datasets into 10 bins (no-bucketization baseline: 0.668).
5.4.2. Comparisons between TableGAN-MCA and existing MIAs
Firstly, TableGAN-MCA recovers member data points from the GAN-synthesized tables previously assumed to be resilient to table-GAN (Park et al., 2018). We evaluate the performance of table-GAN against the same WGAN-GP that used in TableGAN-MCA evaluation. Notice that we test their MIAs directly on the target discriminator instead of the shadow discriminator due to the fact that if the target discriminator fails, the shadow model will perform even worse. We report the accuracy of membership prediction (member/non-member) of table-GAN, which are , and for Adult, Lawschool and Compas datasets, respectively ( is the baseline of random guess). Taken altogether the experiment results in Fig. 6, we conclude that GAN APIs with a black-box access assumed to be resilient to table-GAN (targeting on a discriminative model) (Park et al., 2018) may still disclose partial sensitive training information under TableGAN-MCA.
Secondly, the MIAs proposed in GAN-Leaks and LOGAN cannot disclose membership collisions. Note that the existing MIAs against GANs may work in the MCA scenario. Thus, we perform additional experiments to infer membership collisions of each synthetic data point using their methods. In particular, we evaluate LOGAN (black-box attack with no auxiliary knowledge) and GAN-Leaks (full black-box generator attacks) under threat model (1) (given one copy synthetic data, see details in Section 3.2) and report the result in Table 7. MC and table-GAN are not included in this experiment. The reason is two-fold. First, the distance function of MC is not directly applicable to non-image datasets. Second, table-GAN requires predicted probability vectors of the target discriminator, which is not permitted in our threat model. Note that the synthetic dataset has imbalanced membership collisions labels (Row 1 in Table 7) that are different from Shokri’s shadow model MIA (Shokri et al., 2017) (random observation with real members) since the number of colliding data points (members) is usually unequal to non-colliding ones (non-members).
We observed that the results in GAN-Leaks are close to the random guess baseline. This is due to the reconstruction loss for all synthetic data regardless of membership collisions (the optimal reconstruction of a synthetic data x is itself). Furthermore, LOGAN did not show convincing inference results since it never learns the intersection between the synthetic data and the private training data. In comparison, TableGAN-MCA learns such an intersection (by which we recover partial training data) through the intersections of the published synthetic data (by mimicking the private training data) and shadow (synthetic) data (by mimicking the original synthetic data).
In summary, the MIA classifiers that identify membership fail to identify those membership collisions since the decision boundaries of our attack classifier is different from those of MIAs against GANs.
Adult | Lawschool | Compas | |
Base | |||
LOGAN | |||
GAN-Leaks | |||
Proposed |
6. TableGAN-MCA Analysis
In this section, we discuss the factors that may impact the attack performance of TableGAN-MCA from the following aspects, such as GAN training set size, GAN training epochs and GAN training data frequencies. We choose WGAN-GP as targets as well as shadow model for its superior modeling quality and stability in TableGAN-MCA experiments.
6.1. GAN Training Set Size
The size of the training dataset for a GAN model positively impacts the attack performance. Fig. 9 depicts the positive impact of training dataset size on prediction accuracy and AUPRC of TableGAN-MCA, where in x-axis indicates the full size of a given dataset, . Especially, when the size of the training dataset is less than of the full dataset, increasing the size has a significant impact on the attack performance. The intuition behind the experimental results is two-fold. First, less training data decrease the number of colliding members (positives) in fixed amount of synthetic datasets thus decreases the attack effect. Second, GAN learns a less accurate data distribution if trained on a smaller dataset. Synthetic data generated by such a distribution contain less information than the original training data hence hard for the adversary to learn the statistical patterns of the members/non-members. Note that our results do not conflict with (Hayes et al., 2020; Chen et al., 2020; Lin et al., 2021) since we use different measurements (PR space vs ROC space) that focus on different domains (Davis and Goadrich, 2006). Additionally, our attack target (test data) is also different. We aim to recover the colliding member data from the released synthetic dataset whereas they aim to infer the membership of a random target data point, and thus we learn different decision boundaries.



6.2. GAN Training Epochs
Epochs impact the attack performance of TableGAN-MCA by impacting the knowledge learned by GAN models. We study the attack performance on different training stages by setting different epochs in Fig. 10, where we report the attack prediction accuracy and attack AUPRC.
As seen from Fig. 10, we find that the membership leakage starts at the very beginning of the training epoch, even before the GAN reaches the Nash equilibrium. Interestingly, in Adult and Compas, the attack effect seems to slightly decrease when we set a larger epochs for training GAN models. Since TableGAN-MCA tends to recover the data with high appearance frequency (recall Fig. 3), we conclude that with increasing epochs, GAN models learn more about the training data distribution; hence, the released synthetic data contain more information, which enhances the attack performance. However, once the GAN models learn the details of the data distribution, such details about the distribution would dilute the frequency of those data supposed to have high frequency. The attack performance of TableGAN-MCA is then potentially dropped.



6.3. Training Data Frequencies
Training data frequencies are positively correlated with training data recovery probabilities by TableGAN-MCA. We first compute the recovery possibility and appearance frequency for each data point. We then plot the recovery possibility over the values of data points frequency in Fig. 11. For each dataset, we set two precision-scores of TableGAN-MCA and plot the training data frequency-recovery rate curves. Overall, highly frequent training data are more susceptible to TableGAN-MCA. For instance, when attacking Adult datasets with precision, training data with appearance more than three times are recovered by TableGAN-MCA whereas only of unique training data () are recovered by TableGAN-MCA.
For highly frequent training data, GANs inevitably learn and output these common representations frequently; thus it is easy to recover such highly frequent data by TableGAN-MCA. The re-identification threats of these data caused by TableGAN-MCA are limited since each of them correspond to several individuals and lack of uniqueness.
Unique training data, on the other hand, have more risks for being linked to specific people once recovered by TableGAN-MCA. Therefore, it deserves further exploration for the reason of being exposed.

Generalization of GAN models may accidentally increase the appearance of some unique data points in the synthetic data, therefore increasing their probability to be recovered by TableGAN-MCA. Since the TableGAN-MCA is based on data density in modeled distribution , for a recovered unique training data point , we study how TableGAN-MCA is impacted by the difference between data density of in the training distribution and that of the modeled distribution .
According to our experiments, we discover that some unique training data (, ) have unexpected high exposure in modeled distribution . For example, in Fig. 12, we illustrate the average counts (from synthetic datasets following modeled distribution ) of five data points that appear in the training dataset only once. As we can see, these five data points have higher counts than what they have in the training dataset (). Such an observation indicates that the generator of GAN models unfairly increases the probability of exposure of some data points under TableGAN-MCA. We also find that such an observation is not rare. For instance, according to the statistics in Fig. 12 (Adult), roughly ( of the training dataset) unique entries at least double their exposure; roughly ( of the training dataset) unique entries at least triple their exposure.



Next, we explore the factors that potentially trigger our observations by a set of experiments inspired by unintended memorization (Carlini et al., 2019). Specifically, unintended memorization identifies the impact of the presence of one training input on the modeled distribution learned by GAN. Note that this experiment resembles the definition of differential privacy (DP) (Dwork et al., 2006b). DP is more generic and rigorous as it measures that the probability difference varies all possible functions and all data points, which is computationally infeasible in our measurements. In this case, we narrow down the design by observing the difference between a sample density in two generated distributions trained on neighboring training sets.
Let be the sensitive training set, be a target data point and be the neighboring dataset such that the Hamming distance . Let be a learned generator trained on and be a generator trained on . We measure the difference between the probability of producing a synthetic data point with (prior) and without (posterior) the input data point .
(6) |
Following a recent work (Carlini et al., 2019), GAN models do not memorize a data point if it does not exist in the training dataset . Thus, if Eq. (6) approaches , we say that the target data is unlikely to be memorized by the GAN. The pseudo-code of the experiment is presented in Alg. 2. In the experiment, we use different GANs () and some of target data to estimate Eq. (6). We report the experimental results of five target data () in Fig. 13.

From Fig. 13, we choose the same samples (data points) as in Fig. 12 to compare how prior (with a target ) and posterior densities (without target ) differ in modeled distribution. We find that the presence of the target entry has limited influence on its frequency in modeled distribution . Even if some data point is absent in the training set, its probability density in synthetic distribution is still high, e.g., in the Adult dataset. This is perhaps because GAN’s generalization smooths the sudden change that happened in the probability space of the training set. For instance, the density of the a target point in may be lower than the surrounding points, whereas the GAN smooths such sudden changes in the probability space, and thus it is unintended to increase its probability of exposure. In another aspect, such a rough probability space in real distribution may be attributed to insufficient sampling or unbalanced sampling. As such, cautious data collection may have positive impact in mitigating such influence. Understanding this complicated phenomenon with more explicit proof is our future work. Currently, we summarize that the unique training data recovered by TableGAN-MCA is mainly due to the GAN’s generalization rather than the unintended memorization. This result implies that mitigating the attack effect of TableGAN-MCA may inevitably compromise the availability of released synthetic datasets, since GAN generalization is closely related to its generation ability, which potentially impacts the quality of generated data.
7. Mitigation
In this section, we evaluate the mitigation effects of differential privacy and two customized defense methods against TableGAN-MCA.
7.1. Differentially Private WGAN-WC
Differentially Private WGAN (DP-GAN) only has acceptable trade-offs for larger privacy budgets, and may hardly eliminates TableGAN-MCA without compromise synthetic data utility. Differential privacy (Dwork et al., 2006b) provides a quantified solution to output randomized answers. In this work, we apply a standard approach of differentially private iterative training procedure (DP-SGD, short for DP stochastic gradient descent) (Abadi et al., 2016; McMahan et al., 2018) to the GAN to train a ()-differentially private generator oracle. Otherwise, since DP-SGD perturbs the training process of discriminative models, such mitigation may achieve sub-optimal trade-offs between membership collision privacy and synthetic data utility. In the experiments, we implement the DP framework according to (McMahan et al., 2018) and account the privacy budget ) using RDP accountant released in Tensorflow/Privacy project. Note that WGAN-GP, TVAE and CTGAN do not have DP versions, and thus we study the DP version of WGAN-WC. The generation quality and TableGAN-MCA effect of non-private baseline are shown in Fig. 14 followed by Table 4 and Fig. 8(a).
To implement DP-WGAN, we train a differentially private discriminator. The generator is differentially private because of the post-processing (Dwork et al., 2014). We add calibrated noise into each gradient of the discriminator during training. The accumulation of multiple Gaussian noise addition (Dwork et al., 2006a) relies on privacy accountant techniques (Abadi et al., 2016) and Rényi differential privacy (Mironov, 2017). We provide DP-related hyper-parameters in Table 8, Appendix C.1.
We provide the experimental results of the machine learning utility and TableGAN-MCA effect when sharing differentially private synthetic data in Fig. 14. The shadow GANs in use are non private WGAN-WC. The privacy budget measures the amount of privacy leakage and a smaller value means more privacy-preserved. denotes the probability of violating -DP, which is set to . As can be seen from Fig. 14, the DP method has some positive effect in defending against the TableGAN-MCA. For Adult datasets, when privacy budget , the attack AUPRC decreases by and model’s predicted accuracy decreases by in comparison to the no-DP baseline (see dash dots in Fig. 14). For the Compas dataset, when privacy budget , the attack AUPRC decreases by and model’s predicted accuracy decreases by in comparison to the no-DP baseline. We also depict the ECDF comparison between the original training data and differentially private synthetic data for each marginal to show marginal fitness compromise in Fig. 18 (Appendix C.1). It is not surprising that DP-WGAN achieves sub-optimal trade-offs when protecting against TableGAN-MCA, since the memorization experiment shows that the presence of individuals does not significantly affect the generated distribution. The membership collisions information that we intend to infer is perhaps highly correlated to population statistics (attributes correlation), which will be preserved even under DP training.






7.2. Customized Defense
7.2.1. Remove Colliding Members
Removing colliding members protects against TableGAN-MCA but it reduces the distribution fitness. The straightforward solution against TableGAN-MCA is to manually remove colliding members from the sampled synthetic dataset and share a cleaned version to the analysts (customers). The whole process is denoted as the “naive defense” (last steps in Alg. 3). We acknowledge the cleaned version can decrease the utility of original synthetic data, especially for distribution fitness. For example, we present the ECDF comparison of synthetic datasets generated by the naive defense (Fig. 15(b)) and no-defense (Fig. 15(a)). We show that the naive defense exhibits decreased marginal fitness compared with no-defense baseline. More ECDFs can be found in Figs. 19(a), 20(a), and 21(a) (Appendix C.2).
7.2.2. GAN-constrained Training
We propose a GAN-constrained training technique, to further improve synthetic data utility while protecting against TableGAN-MCA. This strategy is denoted as an “improved defense”. Simply put, we motivate GANs to generate a synthetic dataset that is disjoint with the training set while minimizing the distance between training data and generated data , which is
(7) |
where denotes a distance metric. Since the discriminator of the WGAN minimizes the Wasserstein distance, we additionally add a constraint during training to force each sampled batch of the generator to be disjoint with . To do so, we remove the intersection between the sampled batch and the training set every iteration before computing the loss function (see Alg. 3). Thus, WGAN automatically searches for the best substitution for such colliding samples at training.
7.2.3. Naive and Improved Defenses Evaluation
The improved defense in large part achieves superior trade-offs than the naive defense, and is almost comparable to the no-defense baseline. We evaluate synthetic data utility of the naive defense, the improved defense and the no-defense (baseline) on WGAN-GP. Note that the baseline is vulnerable to TableGAN-MCA while naive and improved defenses protect against it. We evaluate machine learning efficacy in Fig. 16(a) and marginal fitness in Fig. 16(b).
In Fig. 16(a), we train machine learning models (Logistic Regression Classifier) on synthetic data sampled from the naive defense, the improved defense and the no-defense generator and predict on the real test data. Fig. 16(a) shows that synthetic data generated by naive and improved defenses achieve satisfying prediction accuracy on the Adult and Lawschool datasets. In the Compas dataset, mitigation methods decrease the prediction accuracy compared to the no-defense baseline.
In Fig. 16(b), we compare ECDFs using (recall Section 5.2.1). The lower score implies better marginal fitness. The experimental result shows that the improved defense outperforms the naive defense, and is on par with the no-defense baseline. The improved defense succeeds in compensating the statistical deviation caused by the naive defense (see Fig. 15(c)). More ECDFs of the naive defense and the improved defense are shown in Figs. 19, 20, and 21.
In summary, both naive and improved defenses protect against TableGAN-MCA and in part preserve learning ability of released synthetic data. Moreover, the improved defense achieves better marginal fitness than the naive defense. Despite the potentially effective mitigation, TableGAN-MCA still remains a threat since the proposed defenses achieve sub-optimal privacy-utility trade-offs, eg, reduced synthetic data diversity, under-performance for tiny-domain datasets (see Compas datasets for details).
8. Related Work
Membership privacy is the existence of individuals (Li et al., 2013; Rahman et al., [n.d.]). Existing studies show membership disclosure on discriminative machine learning models, e.g., classifiers (Shokri et al., 2017; Yeom et al., 2018; Long et al., 2018; Salem et al., 2019; Nasr et al., 2019; Sablayrolles et al., 2019) and generative machine learning models, e.g., Generative Adversarial Networks (Park et al., 2018; Hayes et al., 2020; Hilprecht et al., 2019; Chen et al., 2020). In the discriminative settings, an adversary infers whether a specific data point is used to train a target model by querying classifier APIs and using predicted probability vectors, labels, logits, etc., to train attack models. For instance, Shokri’s shadow model (Shokri et al., 2017) infers membership against overfitted multi-class classifiers by training an attack model with labeled synthetic data, which mimic the private training data. Subsequent works further relax the adversary’s background knowledge (Salem et al., 2019) by extending attacks to the white-box (Nasr et al., 2019) and the label-only settings (Li and Zhang, 2020; Choo et al., 2020).


In the track of inferring membership against the generative models, there are several successful approaches, such as, table-GAN (Park et al., 2018), LOGAN (Hayes et al., 2020), MC (Hilprecht et al., 2019) and GAN-leaks (Chen et al., 2020). Note that some of these approaches is originally proposed against image data; however, they are possibly extendable to attack tabular data. That is, they are all related to this study. Hence, we briefly summarize these methods in this section. The conceptual comparisons are shown in Table 1. LOGAN (Hayes et al., 2020) and table-GAN (Park et al., 2018) leverage the output of the overfitting discriminator to train an attack model, which is a variant of Shokri et al. (Shokri et al., 2017) in the context of GAN synthesis. However, their attacks require the predicted probability vector of the target discriminator at the inference phase (see column in Table 1). In our experiment, we have already shown that a GAN resilient to their attacks may still expose training data to TableGAN-MCA. MC (Hilprecht et al., 2019) and GAN-leaks (Chen et al., 2020) extract a customized membership indicator of an overfitting generator to train an attack model. We share a similar theoretical bases with theirs, that is, the modeled distribution of the generator behaves differently on training input versus the non-training one. However, our attack further recovers partial training data by inferring membership of published synthetic data, which is out of their scope (see Columns and in Table 7). In this work, we empirically show that the membership inference classifier cannot be directly used to identify membership collisions in our attack model (see Table 7). Compared to those works, we propose a novel attack model, TableGAN-MCA, that exposes partial training data by exploiting the weakness of tabular data synthesis. Even though we share similar ideas with MIAs in generative setting, the attack model of TableGAN-MCA learns different decision boundaries. According to the experimental results, the success of the proposed attack relies more on population knowledge than individual presence, which is different from MIAs.
9. Conclusion
GAN-synthesized table releasing provides unprecedented opportunities for private data sharing that aims to study the regular pattern of population. In this work, we propose a novel membership collision attack, TableGAN-MCA, against the GAN-synthesized table. Our comprehensive experiments over the real-world datasets conclude some important findings. TableGAN-MCA achieves high recovering rate against the private training data from the published GAN-synthesized tables. Our in-depth studies suggest that the target model, training data size, training epochs and training data frequencies impact the attack performance of TableGAN-MCA. We further conclude that the training data leakage is mainly related to the published population statistics (attributes correlations), rather than the model memorization. To mitigate the effect of TableGAN-MCA, we find that differential privacy (applying DP-WGAN) does not show a satisfying result mainly due to the correlations between training data features. Based on our understanding on TableGAN-MCA, we propose two mitigation approaches, which substitute the published colliding members with similar non-private data entries. We hope that the concept of membership collisions defined and the attack methodology developed in this paper could inform the privacy community of such new potential leakage of data synthesis.
Acknowledgements.
The authors, affiliated with Southeast University, were partially supported by Jiangsu Provincial Key Laboratory of Network and Information Security (No. BM2003201). Minhui Xue was, in part, supported by the Australian Research Council (ARC) Discovery Project (DP210102670). Aiqun Hu and Minhui Xue are the corresponding authors of this paper.References
- (1)
- Abadi et al. (2016) Martin Abadi, Andy Chu, Ian Goodfellow, H. Brendan McMahan, Ilya Mironov, Kunal Talwar, and Li Zhang. 2016. Deep Learning with Differential Privacy. In Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security. ACM, Vienna Austria, 308–318.
- Arjovsky et al. (2017) Martin Arjovsky, Soumith Chintala, and Léon Bottou. 2017. Wasserstein Generative Adversarial Networks. In Proceedings of Machine Learning Research, Vol. 70. PMLR, Sydney, Australia, 214–223.
- Barenstein (2019) Matias Barenstein. 2019. ProPublica’s COMPAS Data Revisited. arXiv:1906.04711 [cs, econ, q-fin, stat] (July 2019). http://arxiv.org/abs/1906.04711
- Blum et al. (2013) Avrim Blum, Katrina Ligett, and Aaron Roth. 2013. A learning theory approach to noninteractive database privacy. Journal of the ACM (JACM) 60, 2 (2013), 1–25.
- Carles (2019) Sala Carles. 2019. SDGym Project. https://github.com/sdv-dev/SDGym
- Carlini et al. (2019) Nicholas Carlini, Chang Liu, Úlfar Erlingsson, Jernej Kos, and Dawn Song. 2019. The Secret Sharer: Evaluating and Testing Unintended Memorization in Neural Networks. In 28th USENIX Security Symposium (USENIX Security 19). USENIX Association, Santa Clara, CA, 267–284.
- Chen et al. (2020) Dingfan Chen, Ning Yu, Yang Zhang, and Mario Fritz. 2020. GAN-Leaks: A Taxonomy of Membership Inference Attacks against Generative Models. In Proceedings of the 2020 ACM SIGSAC Conference on Computer and Communications Security (CCS 2020). ACM, New York, NY, USA.
- Chen et al. (2018) Qingrong Chen, Chong Xiang, Minhui Xue, Bo Li, Nikita Borisov, Dali Kaarfar, and Haojin Zhu. 2018. Differentially private data generative models. arXiv preprint arXiv:1812.02274 (2018).
- Choo et al. (2020) Christopher A. Choquette Choo, Florian Tramer, Nicholas Carlini, and Nicolas Papernot. 2020. Label-Only Membership Inference Attacks. arXiv:2007.14321 [cs.CR]
- Davis and Goadrich (2006) Jesse Davis and Mark Goadrich. 2006. The relationship between Precision-Recall and ROC curves. In Proceedings of the 23rd international conference on Machine learning - ICML ’06. ACM Press, Pittsburgh, Pennsylvania, 233–240.
- Dwork et al. (2006a) Cynthia Dwork, Krishnaram Kenthapadi, Frank McSherry, Ilya Mironov, and Moni Naor. 2006a. Our Data, Ourselves: Privacy Via Distributed Noise Generation. In Advances in Cryptology - EUROCRYPT 2006, Vol. 4004. Springer Berlin Heidelberg, Berlin, Heidelberg, 486–503.
- Dwork et al. (2006b) Cynthia Dwork, Frank McSherry, Kobbi Nissim, and Adam Smith. 2006b. Calibrating noise to sensitivity in private data analysis. In Theory of cryptography conference. Springer, 265–284.
- Dwork et al. (2014) Cynthia Dwork, Aaron Roth, et al. 2014. The algorithmic foundations of differential privacy. Foundations and Trends® in Theoretical Computer Science 9, 3–4 (2014), 211–407.
- Goodfellow et al. (2014) Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. 2014. Generative adversarial nets. In Advances in neural information processing systems. 2672–2680.
- Gulrajani et al. (2017) Ishaan Gulrajani, Faruk Ahmed, Martin Arjovsky, Vincent Dumoulin, and Aaron C Courville. 2017. Improved Training of Wasserstein GANs. In Advances in Neural Information Processing Systems (NDSS). 5767–5777.
- Gupta et al. (2013) Anupam Gupta, Moritz Hardt, Aaron Roth, and Jonathan Ullman. 2013. Privately releasing conjunctions and the statistical query barrier. SIAM J. Comput. 42, 4 (2013), 1494–1520.
- Gupta et al. (2012) Anupam Gupta, Aaron Roth, and Jonathan Ullman. 2012. Iterative Constructions and Private Data Release. In Theory of Cryptography, Vol. 7194. Springer Berlin Heidelberg, Berlin, Heidelberg, 339–356.
- Hardt et al. (2012) Moritz Hardt, Katrina Ligett, and Frank Mcsherry. 2012. A Simple and Practical Algorithm for Differentially Private Data Release. In Advances in Neural Information Processing Systems 25. 2339–2347.
- Hardt and Rothblum (2010) Moritz Hardt and Guy N. Rothblum. 2010. A Multiplicative Weights Mechanism for Privacy-Preserving Data Analysis. In 2010 IEEE 51st Annual Symposium on Foundations of Computer Science (Las Vegas, NV, USA). 61–70.
- Hayes et al. (2020) Jamie Hayes, Luca Melis, George Danezis, and Emiliano De Cristofaro. 2020. LOGAN: Membership Inference Attacks Against Generative Models. In Proceedings on Privacy Enhancing Technologies, Vol. 2019. 133–152.
- Hilprecht et al. (2019) Benjamin Hilprecht, Martin Härterich, and Daniel Bernau. 2019. Monte Carlo and Reconstruction Membership Inference Attacks against Generative Models. In Proceedings on Privacy Enhancing Technologies, Vol. 2019. 232–249.
- Jeff et al. (2017) Larson Jeff, Roswell Marjorie, and Atlidakis Vaggelis. 2017. Compas Dataset. https://github.com/propublica/compas-analysis
- Kingma and Welling (2013) Diederik P Kingma and Max Welling. 2013. Auto-encoding variational bayes. arXiv preprint arXiv:1312.6114 (2013).
- Li et al. (2007) N. Li, T. Li, and S. Venkatasubramanian. 2007. t-Closeness: Privacy Beyond k-Anonymity and l-Diversity. In 2007 IEEE 23rd International Conference on Data Engineering. 106–115.
- Li et al. (2013) Ninghui Li, Wahbeh Qardaji, Dong Su, Yi Wu, and Weining Yang. 2013. Membership privacy: a unifying framework for privacy definitions. In Proceedings of the 2013 ACM SIGSAC conference on Computer & communications security - CCS ’13. Berlin, Germany, 889–900.
- Li and Zhang (2020) Zheng Li and Yang Zhang. 2020. Label-Leaks: Membership Inference Attack with Label. arXiv:2007.15528 [cs, stat] (July 2020).
- Lin et al. (2021) Zinan Lin, Vyas Sekar, and Giulia Fanti. 2021. On the Privacy Properties of GAN-generated Samples. In Proceedings of The 24th International Conference on Artificial Intelligence and Statistics (Proceedings of Machine Learning Research, Vol. 130), Arindam Banerjee and Kenji Fukumizu (Eds.). PMLR, 1522–1530. http://proceedings.mlr.press/v130/lin21b.html
- Long et al. (2018) Yunhui Long, Vincent Bindschaedler, Lei Wang, Diyue Bu, Xiaofeng Wang, Haixu Tang, Carl A. Gunter, and Kai Chen. 2018. Understanding Membership Inferences on Well-Generalized Learning Models. arXiv:1802.04889 [cs.CR]
- Machanavajjhala et al. (2007) Ashwin Machanavajjhala, Daniel Kifer, Johannes Gehrke, and Muthuramakrishnan Venkitasubramaniam. 2007. L -diversity: Privacy beyond k -anonymity. ACM Trans. Knowl. Discov. Data 1, 1 (March 2007), 3–es.
- McMahan et al. (2018) H Brendan McMahan, Galen Andrew, Ulfar Erlingsson, Steve Chien, Ilya Mironov, Nicolas Papernot, and Peter Kairouz. 2018. A general approach to adding differential privacy to iterative training procedures. arXiv preprint arXiv:1812.06210 (2018).
- Mironov (2017) Ilya Mironov. 2017. Rényi differential privacy. In 2017 IEEE 30th Computer Security Foundations Symposium (CSF). IEEE, 263–275.
- Narayanan and Shmatikov (2008) Arvind Narayanan and Vitaly Shmatikov. 2008. Robust De-Anonymization of Large Sparse Datasets. In Proceedings of the 2008 IEEE Symposium on Security and Privacy (SP ’08). IEEE Computer Society, USA, 111–125.
- Nasr et al. (2019) Milad Nasr, Reza Shokri, and Amir Houmansadr. 2019. Comprehensive privacy analysis of deep learning: Passive and active white-box inference attacks against centralized and federated learning. In 2019 IEEE Symposium on Security and Privacy (SP). IEEE, 739–753.
- Online. (2017) The Guardian Online. 2017. The Guardian view on Google’s NHS grab: legally inappropriate. https://www.theguardian.com/commentisfree/2017/may/17/the-guardian-view-on-googles-nhs-grab-legally-inappropriate
- Park et al. (2018) Noseong Park, Mahmoud Mohammadi, Kshitij Gorde, Sushil Jajodia, Hongkyu Park, and Youngmin Kim. 2018. Data synthesis based on generative adversarial networks. Proceedings of the VLDB Endowment 11, 10 (2018), 1071–1083.
- Rahman et al. ([n.d.]) Atiqur Rahman, Tanzila Rahman, Robert Laganiere, Noman Mohammed, and Yang Wang. [n.d.]. Membership Inference Attack against Differentially Private Deep Learning Model. ([n. d.]), 19.
- Ramdas et al. ([n.d.]) Aaditya Ramdas, Nicolas Garcia, and Marco Cuturi. [n.d.]. On Wasserstein Two Sample Testing and Related Families of Nonparametric Tests. ([n. d.]). http://arxiv.org/abs/1509.02237
- Richard et al. ([n.d.]) Sander Richard, Knaplund Kris, , and Winter Kit. [n.d.]. Law School Admissions. http://www.seaphe.org/databases/FOIA/lawschs1_1.dta
- Ronny and Barry (1996) Kohavi Ronny and Becker Barry. 1996. UCI Machine Learning Repository. https://archive.ics.uci.edu/ml/datasets/Adult
- Ruoss et al. (2020) Anian Ruoss, Mislav Balunović, Marc Fischer, and Martin Vechev. 2020. Learning Certified Individually Fair Representations. arXiv:2002.10312 [cs, stat] (Feb. 2020). http://arxiv.org/abs/2002.10312 arXiv: 2002.10312.
- Sablayrolles et al. (2019) Alexandre Sablayrolles, Matthijs Douze, Cordelia Schmid, Yann Ollivier, and Herve Jegou. 2019. White-box vs Black-box: Bayes Optimal Strategies for Membership Inference. In Proceedings of Machine Learning Research (PMLR), Vol. 97. 5558–5567.
- Salem et al. (2019) Ahmed Salem, Yang Zhang, Mathias Humbert, Pascal Berrang, Mario Fritz, and Michael Backes. 2019. ML-Leaks: Model and Data Independent Membership Inference Attacks and Defenses on Machine Learning Models. In Proceedings of the 2019 Network and Distributed Systems Security Symposium.
- Shokri et al. (2017) Reza Shokri, Marco Stronati, Congzheng Song, and Vitaly Shmatikov. 2017. Membership inference attacks against machine learning models. In 2017 IEEE Symposium on Security and Privacy (SP). IEEE, 3–18.
- Stadler et al. ([n.d.]) Theresa Stadler, Bristena Oprisanu, and Carmela Troncoso. [n.d.]. Synthetic Data – A Privacy Mirage. ([n. d.]).
- Sweeney (2002) Latanya Sweeney. 2002. k-ANONYMITY: A MODEL FOR PROTECTING PRIVACY. Int. J. Unc. Fuzz. Knowl. Based Syst. 10, 05 (Oct. 2002), 557–570.
- Xu et al. (2019) Lei Xu, Maria Skoularidou, Alfredo Cuesta-Infante, and Kalyan Veeramachaneni. 2019. Modeling tabular data using conditional gan. In Advances in Neural Information Processing Systems (NISP). 7333–7343.
- Yeom et al. (2018) Samuel Yeom, Irene Giacomelli, Matt Fredrikson, and Somesh Jha. 2018. Privacy risk in machine learning: Analyzing the connection to overfitting. In 2018 IEEE 31st Computer Security Foundations Symposium (CSF). IEEE, 268–282.
Appendix
Appendix A Network Structure and Parameters
WGAN-GP, WGAN-WC shares the same network architecture. We set the Generator as Recurrent Neural Networks (RNNs). According to our experiments, the RNN has a positive effect on stabilizing the generator’s outputs. Eq. (8) represents the Generator networks and Eq. (9) represents the Discriminator networks.
(8) |
(9) |
For TVAE and CTGAN, we applies the module CTGANSynthesizer
and TVAESynthesizer
of the SDGym
(Carles, 2019). Thus, the structures and hyper-parameters are exactly same as the originals’ (Xu et al., 2019).
Hyper-parameters. For Adult and Lawschool datasets, we train epochs and set batch size to . For Compas dataset, we train epochs and set batch size to . Since the Compas dataset is much smaller than others, we find that less iterations could incur under-fitting. Additionally, balancing the number of D and G training sessions also helps to converge faster.
Appendix B Number of Synthetic Queries
We thoroughly discuss how the number of synthetic queries influences the attack performance and corresponding attack tricks.
B.1. Limited Synthetic Queries
Many target model prediction APIs (MLaaS) implement a pay-per-query business model. Hence, reducing the number of synthetic queries saves the cost of performing TableGAN-MCA. However, a smaller synthetic dataset, having less membership collisions with the training dataset, decays the attack performance. To tackle this problem, we propose an approach that uses shadow data to fill up the synthetic data to match the size of the training set. That is, the adversaries obtain a synthetic dataset of size by querying the target Generator. The adversaries then generate the shadow dataset of size . After that, the TableGAN-MCA adversaries attack instead of the original .
We show the impact of a small on TableGAN-MCA in Fig. 17. We find that few synthetic queries also yield decent attack performance. This resonates with the memorization experiment that the success of TableGAN-MCA is contingent more on basic data patterns.

B.2. Unlimited Synthetic Queries
The TableGAN-MCA adversary continues to expose more training data when increasing the number of synthetic queries. In Fig. 7, we evaluate the TableGAN-MCA up to , which is not the ceiling of TableGAN-MCA capabilities. Due to computational constraints, we are limited to performing the attack up to and observe that the number of exposed training data of TableGAN-MCA is still increasing. This leaves open an interesting problem of whether the adversary could reconstruct the whole training dataset with unlimited queries.
Appendix C Mitigation
C.1. DP-WGAN
Datasets | Sampling rate | ||
Adult | 500/31655 | ||
500/31655 | |||
500/31655 | |||
500/31655 | |||
500/31655 | |||
500/31655 | |||
Lawschool | 500/43011 | ||
500/43011 | |||
500/43011 | |||
500/43011 | |||
500/43011 | |||
500/43011 | |||
Compas | 100/3694 | ||
100/3694 | |||
100/3694 | |||
100/3694 | |||
100/3694 |
We show the ECDFs of marginals for -DP synthesized data in Fig. 18. Smaller training data usually gains less satisfactory generation quality under DP training with a similar privacy budget.



C.2. Naive and Improved Defenses
We show additional ECDFs of marginals for “Remove Colliding Members” mitigation and “GAN-constrained Training” mitigation in Figs. 19, 20, and 21.





