This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

SocialGuard: An Adversarial Example Based Privacy-Preserving Technique for Social Images

Mingfu Xue Shichang Sun Zhiyu Wu Can He Jian Wang Weiqiang Liu College of Computer Science and Technology, Nanjing University of Aeronautics and Astronautics, Nanjing, China College of Science, Nanjing University of Aeronautics and Astronautics, Nanjing, China College of Electronic and Information Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing, China
Abstract

The popularity of various social platforms has prompted more people to share their routine photos online. However, undesirable privacy leakages occur due to such online photo sharing behaviors. Advanced deep neural network (DNN) based object detectors can easily steal users’ personal information exposed in shared photos. In this paper, we propose a novel adversarial example based privacy-preserving technique for social images against object detectors based privacy stealing. Specifically, we develop an Object Disappearance Algorithm to craft two kinds of adversarial social images. One can hide all objects in the social images from being detected by an object detector, and the other can make the customized sensitive objects be incorrectly classified by the object detector. The Object Disappearance Algorithm constructs perturbation on a clean social image. After being injected with the perturbation, the social image can easily fool the object detector, while its visual quality will not be degraded. We use two metrics, privacy-preserving success rate and privacy leakage rate, to evaluate the effectiveness of the proposed method. Experimental results show that, the proposed method can effectively protect the privacy of social images. The privacy-preserving success rates of the proposed method on MS-COCO and PASCAL VOC 2007 datasets are high up to 96.1% and 99.3%, respectively, and the privacy leakage rates on these two datasets are as low as 0.57% and 0.07%, respectively. In addition, compared with existing image processing methods (low brightness, noise, blur, mosaic and JPEG compression), the proposed method can achieve much better performance in privacy protection and image visual quality maintenance.

keywords:
Artificial intelligence security, Privacy protection, Social photos, Adversarial examples, Object detectors
journal: arXiv

1 Introduction

In recent years, artificial intelligence (AI) techniques have been widely applied in a variety of tasks. Although bringing many conveniences, the abuse of AI techniques will also cause a series of security problems, such as social privacy disclosure. With the popularity of social platforms, people are keen to share their photos online. However, malicious hackers or entities can leverage the deep neural network (DNN) based object detectors (or image classifiers) to extract valuable private information from those uploaded social photos. In this way, they can analyze and explore users’ preferences, and push the accurate and targeted commercial advertisements [1]. Moreover, the leakage of sensitive information (such as identity) can even cause the users to suffer from property losses. For example, Echizen [2] indicated that the fingerprint might be stolen from the shared photos that contain the “V” gesture.

So far, there are several defense strategies to protect the privacy of the shared social photos. The most direct and common countermeasure is to set user access control [3, 4, 5], where the strangers are not permitted to browse private personal photos. In addition, the image encryption based techniques [6, 7] are also effective for protecting sensitive data on shared photos. However, the above two methods will either reduce the users’ social experience (access control), or will require huge computation overhead (image encryption). Therefore, the image processing methods, such as blur, noise and occlusion, have been widely used to protect the privacy contents in images. However, these image processing methods will greatly affect the visual qualities of the photos. Moreover, they are ineffective to resist the detection of those advanced DNN based object detectors [8, 9].

In this work, we propose a novel adversarial example based method to protect social privacy from being stolen. The key idea is that, the DNN based object detectors are sensitive to small perturbations [8, 10, 11], i.e., adversarial examples. By constructing adversarial version of these uploaded social photos, i.e., adding some visual invisible perturbations into the photos, the DNN based object detectors can be successfully fooled. Consequently, the object detectors will make incorrect predictions, or even completely unaware of the existence of the objects in a photo, thus the private information in the social photos can be protected.

The main contributions of this paper are as follows:

  • 1.

    Protect the social privacy effectively without affecting the visual qualities of the photos. We propose a method based on adversarial examples to protect the privacy in social photos, which can make Faster R-CNN [12] detector fails to predict any bounding box in a social image. Experimental evaluations on two datasets (MS-COCO [13] and PASCAL VOC 2007 [14]) show that, the privacy-preserving success rate of the proposed method is high up to 96.1% and 99.3%, respectively. Compared with existing social privacy protection methods, the proposed method can effectively protect image contents by fooling DNN based object detectors, without affecting the visual qualities of the social photos.

  • 2.

    Supporting customized privacy setting. Some objects (such as person or commercialized objects) in social images contain personal information and thus can be considered as sensitive objects. The proposed method can not only hide all the objects from being detected, but can also make the customized sensitive objects in a social image be incorrectly predicted. The user can customize the sensitive objects that he wants to keep private. After the social image is processed by the proposed method, Faster R-CNN will identify the customized sensitive objects incorrectly.

  • 3.

    Use the attack method as a defense. Adversarial example used to be a malicious attack method against DNN models. In this paper, we are doing the opposite by using the attack as a defense technique for resisting the detection of DNN based object detectors, so as to protect the privacy information in shared social photos from being disclosed.

The rest of this paper is organized as follows. Section 2 describes the related work on adversarial example attacks, the workflow of Faster R-CNN, and the existing social privacy protection works against DNN based privacy stealing. Section 3 elaborates the proposed method. Section 4 presents the experimental results on two standard image datasets. This paper is concluded in Section 5.

2 Related work

In this section, the adversarial example attacks, the workflow of Faster R-CNN, and the existing four social privacy protection works against DNN based privacy stealing, are reviewed.

2.1 Adversarial example attacks

Some researches have indicated that DNN models are vulnerable to adversarial examples, and a range of adversarial example generation methods have been proposed. Szegedy et al. [15] first proposed adversarial example attacks against the DNN models, and developed the L-BFGS algorithm to construct the adversarial perturbations. By adding the generated perturbations into a clean image, the target DNN model will output the incorrect classification result. Then, Goodfellow et al. [16] proposed the Fast Gradient Sign Method (FGSM) to explain the linear behavior of adversarial perturbations, which is effective to accelerate the generation of adversarial examples. Carlini and Wagner [17] constructed the high-confidence adversarial examples with three different distance metrics (i.e., L0{L_{0}} distance, L2{L_{2}} distance, and L{L_{\infty}} distance), and their generated adversarial examples could successfully fool the distilled neural networks. Chen et al. [18] developed the ShapeShifter attack to craft physical adversarial examples to fool road sign object detectors. By applying the Expectation over Transformation technique [19, 20], the ShapeShifter is robust to the changes of different distances and angles.

2.2 Workflow of Faster R-CNN

The workflow of the Faster R-CNN detector includes two stages [12]: region proposal and box classification. In the first stage, the features in an image are extracted, and the region proposal network (RPN) is used to determine where an object may exist and mark the region proposal. In the second stage, the corresponding objects in the marked region proposals are classified by DNN based classifier. Meanwhile, the bounding box regression [21] is performed to obtain the accurate location of each object in the image.

2.3 Social privacy protection against DNN based privacy stealing

We proposed the idea of this paper in 2019 and applied for a Chinese patent [22]. Besides, few exploratory studies have been conducted on the protection of image privacy against DNN based privacy stealing. Li et al. [23] presented the Private Fast Gradient Sign Method (P-FGSM) to prevent the scene (e.g., hospital, church) in social images from being detected by DNN model. They added imperceptible perturbation to the social images to generate the adversarial social images, in which the scene will be incorrectly classified by DNN classifier. Shen et al. [24] proposed an algorithm to protect the sensitive information in social images against DNN based privacy stealing and keep the changes of the social images imperceptible. They also conducted a human subjective evaluation experiment to evaluate the factors that influence the visual sensitivity of human [24]. The difference between our work and [24] is that, [24] focuses on preventing sensitive information from being detected by DNN based classifier, while our work focuses on preventing sensitive objects from being detected by object detector. Shan et al. [25] proposed a Fawkes system to protect personal privacy against unauthorized face recognition models. Specifically, they add pixel-level perturbations to user’s photos before uploading them to the Internet [25]. The functionality of unauthorized facial recognition models trained on these photos with perturbations will be deteriorated seriously. The differences between our work and [25] are that: (i) The work [25] protects social privacy against facial recognition models, while our method protect social privacy against object detectors. (ii) The work [25] adds adversarial images to the training set of unauthorized facial recognition models to influence the training phase of these models (which is a strong assumption), while our work focuses on the testing phase of DNN detectors.

The above three related works focus on privacy protection against the image classifiers. Compared with image classifiers, attacking or deceiving object detectors is more challenging [26, 27, 8]. The reasons are as follows. First, an object detector is more complex, which can detect multiple objects at once. Second, the object detector can infer the true objects by the obtained information from the image background. To the best of the authors’ knowledge, the only social privacy protection work against object detectors is Liu et al. [8]. Liu et al. [8] constructed the adversarial examples with a stealth algorithm, which made all objects in the image invisible to the object detectors, thus protecting the image privacy from being detected. Compared to the work [8], our work has the following differences: (i) The mechanisms are different. The work [8] focuses on the region proposal stage (the first stage) of Faster R-CNN, while our method focuses on the classification stage (the second stage) of Faster R-CNN. Moreover, our method leverages the classification boundary of object detector to craft perturbations. Specifically, the position of the feature point of each bounding box is moved and crossed the classification boundary, which make object detector cannot find any sensitive object and abandon the corresponding bounding boxes, while the stealth algorithm in [8] performs a backpropagation on the loss function to create perturbations and suppress the generation of bounding boxes. (ii) The effects are different. The work [8] can hide all the objects in a social image from being detected, while our method can not only achieve this, but can also make the customized sensitive objects in a social image be misclassified as other classes.

3 Proposed adversarial example based privacy-preserving technique

In this section, the proposed adversarial example based social privacy-preserving technique is elaborated. First, the privacy protection goals for social photos are described. Second, the overall flow of the proposed method is presented. Third, the Object Disappearance Algorithm is elaborated. The proposed method can generate two kinds of adversarial examples. One can make all the objects in a social image invisible, and the other can make the customized sensitive objects in a social image be incorrectly predicted.

3.1 Privacy goal

DNN based object detectors perform excellently in various computer vision tasks. The deep learning based object detectors can be divided into two categories: (i) one-stage object detectors, such as You Only Look Once (YOLO) [28], and Single Shot Multibox Detector (SSD) [29]; (ii) two-stage object detectors, such as Faster R-CNN [12], and Region-based Fully Convolutional Networks (R-FCN) [30]. Existing research [31] has indicated that, compared to SSD and R-FCN, the Faster R-CNN can achieve more superior performance on the trade-off between speed and detection accuracy. Besides, generally, Faster R-CNN has higher detection accuracy than that of YOLO detector. Therefore, in this paper, the proposed method is targeting at protecting image privacy from being detected by Faster R-CNN detectors.

The privacy goal of the proposed method includes two aspects: one is that Faster R-CNN fails to predict any bounding box in an image, which will be discussed in Section 3.3, and the other is that Faster R-CNN can predict some bounding boxes in an image but classifies the sensitive objects incorrectly, which will be discussed in Section 3.4. If one of these two aspects is met, it is considered that the social privacy is successfully protected. To achieve the first aspect, we need to make sure that for each bounding box, the score of each bounding box is lower than the threshold, which can be formalized as:

max[Pk(x)]<T(k=1,2,,K)\max[P_{k}(x^{\prime})]<T\;\;(k=1,2,...,K) (1)

where xx^{\prime} is an adversarial social image. KK is the number of classes that an object detector can recognize (K=80K=80 in the experiment), and kk denotes the k-th category. PkP_{k} is the score of class kk of the bounding box. TT is the detection threshold of the system.

For an object detector, it outputs bounding boxes with high scores (higher than the threshold), and discards the bounding boxes with low scores (lower than the threshold). Equation (1) indicates that the scores predicted by Faster R-CNN for all the bounding boxes are all lower than the threshold. Therefore, the system will abandon all bounding boxes and detect nothing in the adversarial social image.

The proposed adversarial example based method can make all objects in a social image undetectable by Faster R-CNN (the first aspect), which will be discussed in Section 3.3. In addition, the proposed method can also make the customized sensitive objects incorrectly classified as another object (the second aspect), which will be discussed in Section 3.4.

3.2 Overall flow

The overall flow of the proposed method is shown in Fig. 1. The method takes users’ photos as the inputs, and outputs the adversarial examples to deceive the Faster R-CNN object detector. First, pre-detection is performed on the users’ photos to obtain the detected objects. Then, the classes that do not belong to those correctly detected classes are selected as non-sensitive classes (defined in Section 3.3). Second, the region proposals r1,r2,,rm{r_{1}},{r_{2}},...,{r_{m}} are obtained in the classification stage of Faster R-CNN. Third, the adversarial attack is implemented at the second stage (i.e., the classification stage) of Faster R-CNN by iteratively constructing the perturbation in the image. The perturbation is iteratively constructed until that, for every bounding box, the score of each object is lower than the threshold, which will cause Faster R-CNN to discard all the bounding boxes. In this way, the generated adversarial social image can successfully deceive the Faster R-CNN detector when it is uploaded to the social platforms.

Refer to caption
Figure 1: The overall flow of the proposed adversarial example based privacy-preserving technique.

3.3 Object Disappearance Algorithm

We propose an Object Disappearance Algorithm which is presented in Algorithm 1, to generate adversarial examples and prevent Faster R-CNN from predicting any bounding box (the first aspect as discussed in Section 3.1). The proposed method focuses on attacking the classification network at the second stage (i.e., the classification stage) of Faster R-CNN. The adversarial attack includes two steps. First, the pre-detection on the original image xx is performed to obtain the predicted classes, including the correct and incorrect prediction results. We define the correctly predicted classes as the sensitive classes that need to be protected. In this paper, for an unprocessed social image, the sensitive objects are the objects that appear in the image and can be detected by object detector, and the non-sensitive objects are the objects that do not appear in the image. For different social images, the sensitive objects and the non-sensitive objects are different. Then, the non-sensitive class set Y={y1,y2,,yKd}{Y^{\prime}}=\{{y^{\prime}_{1}},{y^{\prime}_{2}},...,{y^{\prime}_{K-d}}\} is selected, where KK is the number of classes that an object detector can recognize, and dd is the number of classes that are detected in the original image. y1,y2,,yKd{y^{\prime}_{1}},{y^{\prime}_{2}},...,{y^{\prime}_{K-d}} do not belong to these correctly detected classes and are the classes of objects that do not appear in the original image xx. Second, the region proposals r1,r2,,rm{r_{1}},{r_{2}},...,{r_{m}} are obtained in the classification stage of Faster R-CNN, and the perturbations are constructed iteratively in the original image xx with the following formula:

argminxi\displaystyle\mathop{\arg\min}\limits_{x{{}^{\prime}_{i}}}{\kern 1.0pt} 1mrjL(F(x)i,y)t\displaystyle\frac{1}{m}\sum\limits_{{r_{j}}}{L(F{{(x{{}^{\prime}_{i}})}},y{{}^{\prime}_{t}})} (2)

where xix{{}^{\prime}_{i}} is an adversarial example generated in the ii-th iteration. rjr_{j} represents the j-th predicted bounding box, and the total number of bounding boxes is mm. yt{y^{\prime}_{t}} is the label of the non-sensitive class and ytY{y^{\prime}_{t}}\in{Y^{\prime}}. yty^{\prime}_{t} is not a fixed class label and is sequentially selected from the set Y{Y^{\prime}} in each iteration. FF represents the classification network of Faster R-CNN, and LL is the loss function that computes the difference between the output F(x)iF(x{{}^{\prime}_{i}}) of the classification network and the non-sensitive class label yty^{\prime}_{t}.

Algorithm 1 Object Disappearance Algorithm.
1:original image xx; perturbation constraint ε\varepsilon; detection threshold TT; maximum number of iterations II;
2:adversarial example xx^{\prime};
3:Initialize x=0xx{{}^{\prime}_{0}}=x; i=1i=1;
4:Conduct a pre-detection on xx to obtain the classified objects;
5:Select the non-sensitive class set YY{{}^{\prime}};
6:Obtain the region proposals r1,r2,,rm{r_{1}},{r_{2}},...,{r_{m}} in the classification stage of Faster R-CNN.
7:while i<Ii<I do
8:     Add perturbation to the original image xx and get the adversarial
9:   example xix{{}^{\prime}_{i}};
10:     Clamp(xix{{}^{\prime}_{i}}, xi1εx{{}^{\prime}_{i-1}}-\varepsilon, x+i1εx{{}^{\prime}_{i-1}}+\varepsilon);
11:     Use the classification network to predict the class of objects for all
12:   the bounding boxes in xix{{}^{\prime}_{i}};
13:     Get the scores SS of each class for all the bounding boxes;
14:     smax=max(S){s_{max}}=\max({S});
15:     if smax<T{s_{max}}<T then
16:         x=xix^{\prime}=x{{}^{\prime}_{i}};
17:         break;
18:     end if
19:     i=i+1i=i+1;
20:end while
21:return xx^{\prime};

In each iteration, we sequentially select the non-sensitive class yt{y^{\prime}_{t}} from the set Y{Y^{\prime}}. Then, we construct the adversarial perturbation with formula (2) and utilize the constraint ||xix|i1|<ε||x{{}^{\prime}_{i}}-x{{}^{\prime}_{i-1}}||<\varepsilon to make the perturbation visually invisible, where ε\varepsilon restricts the intensity of the generated perturbations. After the adversarial example xix{{}^{\prime}_{i}} for the ii-th iteration is generated, the classification network FF is used to predict the classes of objects for all the bounding boxes r1,r2,,rm{r_{1}},{r_{2}},...,{r_{m}}. Subsequently, the scores S{S} of each class for all the bounding boxes r1,r2,,rm{r_{1}},{r_{2}},...,{r_{m}} will be obtained. SS can be considered as a matrix as follows:

S=(s11s12s1Ks21s22s2Ksm1sm2smK)S=\left({\begin{array}[]{*{20}{c}}{{s_{11}}}&{{s_{12}}}&\cdots&{{s_{1K}}}\\ {{s_{21}}}&{{s_{22}}}&\cdots&{{s_{2K}}}\\ \vdots&\vdots&\ddots&\vdots\\ {{s_{m1}}}&{{s_{m2}}}&\cdots&{{s_{mK}}}\end{array}}\right) (3)

where sijs_{ij} is the score of class jj for the ii-th bounding box (i=1,2,,mi=1,2,...,m, j=1,2,,Kj=1,2,...,K). The maximum score smax{s_{max}} is then calculated, where smax{s_{max}} represents the maximum score of all classes for all the bounding boxes r1,r2,,rm{r_{1}},{r_{2}},...,{r_{m}} in the generated adversarial example xix{{}^{\prime}_{i}}. We denote the maximum number of iterations as II. The iteration process is terminated when the iteration number reaches II or smax{s_{max}} is lower than the threshold, which means that there is no bounding box that can be detected by Faster R-CNN.

Here, we explain the reason why the above algorithm can cause all objects in the generated adversarial example invisible. For sensitive classes in an image, adding perturbation with Equation (2) can change the classification boundary of sensitive classes and reduce the scores of them. Thus, the scores of sensitive objects are below the detection threshold. For non-sensitive classes, the multi-classes perturbing method restraints the increase of non-sensitive objects’ scores with insufficient iterations. As a result, the scores of non-sensitive objects are also below the detection threshold. As a result, the score of each class for any bounding boxes will be lower than the threshold, which makes Faster R-CNN discard all the bounding boxes and detect nothing in the adversarial social image.

3.4 Customized privacy setting

In this section, we describe another method (denoted as MsenM_{sen}) of generating adversarial examples which can protect the customized sensitive objects in the social images. In a social photo, some objects are usually sensitive and private, such as: (i) Person. The person in a social photo may involve users’ personal information. For example, a common V-shaped gesture photo may contain the fingerprint information of an user. (ii) Commercialized objects. Some commercial objects (such as handbags, sports balls, and books) may reveal users’ interests. The leakage of these commercialized objects may bring unnecessary advertising to users. To avoid the disclosure of the sensitive objects in the shared social photos, the proposed method can also provide users with privacy settings, where users can customize the sensitive objects that he wants to protect and the specific class ynony_{non} that he wants the sensitive objects to be misclassified as. The proposed method can achieve this purpose by iteratively constructing perturbations on social images with the following formula:

argminxi\displaystyle\mathop{\arg\min}\limits_{x{{}^{\prime}_{i}}}{\kern 1.0pt} 1mrjL(F(x)i,y)non\displaystyle\frac{1}{m}\sum\limits_{{r_{j}}}{L(F{{(x{{}^{\prime}_{i}})}},y{{}_{non}})} (4)

The perturbation is iteratively constructed until that, for every bounding box, the scores of the customized sensitive objects are lower than the threshold. MsenM_{sen} can cause the score of the class ynony_{non} to be higher than the threshold, which will make Faster R-CNN classify the customized sensitive objects as the class ynony_{non}. As a result, Faster R-CNN cannot detect the customized sensitive object and will present the bounding boxes with wrong class (the non-sensitive class ynony_{non}).

We denote the method introduced in Section 3.3 as MallM_{all}, and the method in this section as MsenM_{sen}. We use AEallAE_{all} and AEsenAE_{sen} to represent the adversarial examples generated by MallM_{all} and MsenM_{sen}, respectively. The difference between the workflows of MallM_{all} and MsenM_{sen} is shown in Fig. 2. MallM_{all} performs pre-detection on the original social image to obtain the non-sensitive classes and utilizes multiple non-sensitive classes to generate adversarial example, while MsenM_{sen} does not perform pre-detection and utilizes a fixed non-sensitive class ynony_{non} to generate adversarial example.

Refer to caption
Figure 2: The difference between the workflows of MallM_{all} and MsenM_{sen}.

In a word, this paper provides two adversarial example based method (MallM_{all} and MsenM_{sen}) to protect the privacy of social photos. MallM_{all} hides all objects in a social image from being detected by Faster R-CNN, while MsenM_{sen} is used to protect the sensitive objects customized by the users.

4 Experimental results

In this section, we evaluate the proposed method. First, the experimental settings, the experimental datasets, and the two evaluation metrics, are described in Section 4.1. Second, we demonstrate the effectiveness of the proposed method in terms of two metrics: the privacy-preserving success rate (Section 4.2) and the privacy leakage rate (Section 4.3). Then, we explore the impacts of two parameters (perturbation constraint ε\varepsilon and detection threshold TT) on the proposed method in Section 4.4. Finally, we compare the proposed method with the existing image processing methods in Section 4.5.

4.1 Experimental setup

In our experiments, we evaluate the performance of the proposed method on the Faster R-CNN [12] Inception v2 [32] model. The model has been pre-trained on the Microsoft Common Objects in Context (MS-COCO) dataset [13] with Tensorflow [33] platform. We conduct the experiment on MS-COCO dataset [13] and PASCAL VOC 2007 dataset [14]. The MS-COCO dataset has more than 200,000 images for object detection. It contains 80 object categories including people, vehicles, animals, sports goods, and other common items [13]. The PASCAL VOC 2007 dataset [14] has 20 categories, most of which are the same as the categories in MS-COCO dataset. We randomly select 1,000 images from MS-COCO dataset [13], and 1,000 images from PASCAL VOC 2007 dataset [14], respectively, to evaluate the proposed method.

We use two metrics, named privacy-preserving success rate and privacy leakage rate, to evaluate the proposed method on protecting the privacy of social images. As mentioned in Section 3.3 and Section 3.4, the proposed social privacy-preserving method can generate two kinds of adversarial examples. As discussed in Section 3.4, we use AEallAE_{all} to represent the adversarial examples that aim to make all objects invisible, and AEsenAE_{sen} to represent the adversarial examples that aim to hide the customized sensitive object. The privacy-preserving success rate (RallR_{all}) for adversarial examples AEallAE_{all} can be formalized as:

Rall=NallN{R_{all}}=\frac{N_{all}}{N} (5)

where NallN_{all} is the number of adversarial social images AEallAE_{all} in which Faster R-CNN cannot present any bounding boxes. NN is the number of all the adversarial social images. The privacy-preserving success rate (RsenR_{sen}) for adversarial examples AEsenAE_{sen} can be formalized as:

Rsen=NsenN{R_{sen}}=\frac{N_{sen}}{N} (6)

where NsenN_{sen} is the number of adversarial examples AEsenAE_{sen} in which Faster R-CNN classifies the customized sensitive class incorrectly.

To further explore the effectiveness of the proposed method on preventing the privacy leakage, we use a metric named privacy leakage rate. The privacy leakage rate represents how many sensitive contents are detected by Faster R-CNN in social images. The privacy leakage rate PallP_{all} for the adversarial social images that aim to hide all the objects (AEallAE_{all}) can be calculated as follows:

Pall=AOP_{all}=\frac{A}{O} (7)

where AA is the number of bounding boxes that are detected in all the adversarial social images AEallAE_{all}, including the correct and incorrect detection results. OO is the number of bounding boxes that are detected in all the original social images, including the correct and incorrect detection results. The privacy leakage rate PsenP_{sen} for the adversarial social images that aim to hide the customized sensitive objects (AEsenAE_{sen}) can be expressed as:

Psen=AsenOsenP_{sen}=\frac{A_{sen}}{O_{sen}} (8)

where AsenA_{sen} is the number of bounding boxes of the customized sensitive classes that are correctly detected in the adversarial social images AEsenAE_{sen}. OsenO_{sen} is the number of bounding boxes of customized sensitive classes that are correctly detected in the original social images.

4.2 Privacy-preserving success rate

In this section, we use the privacy-preserving success rate to evaluate the proposed method. Fig. 3 shows four examples to illustrate the effectiveness of the proposed method. The four examples consist of the original images, the detection results of the original images, the detection results of adversarial examples AEallAE_{all}, and the detection results of adversarial examples AEsenAE_{sen}, respectively. Faster R-CNN can easily recognize all the sensitive objects if the shared images are unprocessed, as shown in Fig. 3 and Fig. 3. Specifically, all the persons are accurately detected in Fig. 3. However, Faster R-CNN detects nothing in the adversarial examples AEallAE_{all}, as shown in Fig. 3. Meanwhile, it is hard for humans to observe the differences between adversarial examples AEallAE_{all} and the original images. In Fig. 3, all the persons in adversarial examples AEsenAE_{sen} are misclassified as other objects. The above results indicate that, the proposed method can protect the social privacy effectively without affecting the visual effect of social images.

Refer to caption
Refer to caption
Refer to caption
Refer to caption
Figure 3: Image examples. (a) Original images; (b) the detection results of original images; (c) the detection results of AEallAE_{all}; (d) the detection results of AEsenAE_{sen}.

The privacy-preserving success rates of the proposed method are shown in TABLE 1. In the experiment, we select 1,000 images in each dataset and test the privacy-preserving success rates for the adversarial social image AEallAE_{all} and AEsenAE_{sen}, respectively. The success rates of AEallAE_{all} on MS-COCO and PASCAL VOC 2007 datasets are 96.1% and 99.3%, respectively. The success rates of AEsenAE_{sen} on the two datasets are 96.5% and 99.4%, respectively. The privacy-preserving success rates of AEallAE_{all} and AEsenAE_{sen} are very close and both reach a very high level. The results indicate that the proposed method performs well on protecting the privacy of social photos.

Table 1: Privacy-preserving success rates of the proposed method on MS-COCO and PASCAL VOC 2007 datasets. AEallAE_{all} represents the adversarial examples that aim to protect all objects. AEsenAE_{sen} represents the adversarial examples that aim to protect the customized sensitive object.
Dataset Privacy-preserving success rate
AEallAE_{all} AEsenAE_{sen}
MS-COCO [13] 96.1% 96.5%
PASCAL VOC 2007 [14] 99.3% 99.4%

4.3 Privacy leakage rate

In this section, we use privacy leakage rate to evaluate the proposed method. We calculate the privacy leakage rates for AEallAE_{all} and AEsenAE_{sen} on MS-COCO and PASCAL VOC 2007 datasets, respectively. As shown in TABLE 2, the privacy leakage rates for adversarial social images AEallAE_{all} on the two datasets are 0.57% and 0.07%, respectively, and the privacy leakage rates for adversarial social images AEsenAE_{sen} on the two datasets are 2.23% and 0.18%, respectively. The privacy leakage rates of AEallAE_{all} and AEsenAE_{sen} both reach a very low level. The results indicate that, our method can protect the sensitive objects from being detected by Faster R-CNN.

Table 2: Privacy leakage rates of two kinds of adversarial examples on MS-COCO and PASCAL VOC datasets.
Dataset Privacy leakage rate
AEallAE_{all} AEsenAE_{sen}
MS-COCO [13] 0.57% 2.23%
PASCAL VOC 2007 [14] 0.07% 0.18%

Considering that our method may have different privacy protection effects for different sensitive objects, we select 9 different sensitive objects and calculate the privacy leakage rates of these objects, respectively. The 9 classes are person, elephant, airplane, laptop, traffic light, book, sports ball, parking meter, and cell phone. In the experiment, we select 300 images from MS-COCO dataset [13] for each class. Fig. 4 shows the privacy leakage rates of the nine sensitive objects in the adversarial social images AEsenAE_{sen}. The result indicates that, the proposed method can effectively prevent the privacy information in the social image from being detected. The privacy leakage rates of the nine sensitive classes are all in a very low level (lower than 2%). Moreover, the privacy leakage rate of the elephant class is as low as 0.26%.

Refer to caption
Figure 4: Privacy leakage rates of AEsenAE_{sen} on nine sensitive object classes.

4.4 Parameter discussion

In the proposed method, there are two parameters that may affect the performance: the perturbation constraint ε\varepsilon and the detection threshold TT. The perturbation constraint ε\varepsilon constrains the intensity of the perturbation, which makes the difference between the original image and the adversarial example difficult to be noticed. The detection threshold TT determines the number of bounding boxes that are presented by Faster R-CNN.

We set the values of the perturbation constraint ε\varepsilon to be 1/2551/255\sim10/25510/255 because these values of ε\varepsilon are small enough and can make the generated perturbations difficult to be noticed. The privacy-preserving success rate Rsen{R}_{sen} and the privacy leakage rate Psen{P}_{sen} of adversarial social images AEsenAE_{sen} under different ε\varepsilon settings are shown in Fig. 5. It can be seen that when ε<3/255\varepsilon<3/255, both the privacy-preserving success rate Rsen{R}_{sen} and the privacy leakage rate Psen{P}_{sen} change drastically as ε\varepsilon increases, and Rsen{R}_{sen} and Psen{P}_{sen} tend to stabilize when ε>3/255\varepsilon>3/255. Moreover, when ε=3/255\varepsilon=3/255, the value of Rsen{R}_{sen} already reaches a high level (96.5%) and the value of Psen{P}_{sen} is only 2.52%. Note that, the larger the value of ε\varepsilon is, the larger the perturbation strength to an image. A large perturbation strength may affect the visual quality of the perturbed image. Therefore, ε=3/255\varepsilon=3/255 has the optimal performance when considering the trade-off between privacy preserving effect and perturbation strength.

We also study the relationship between ε\varepsilon and the image quality. Structural Similarity (SSIM) [34] and Peak Signal to Noise Ratio (PSNR) are two quantitative metrics to reflect the visual quality of the processed images. PSNR describes the pixel level difference between two images, while SSIM [34] shows structural changes of a processed image. The higher the value of PSNR and SSIM [34], the more similar between the original image and the modified image. Fig. 6 presents the PSNR and SSIM of the modified images AEsenAE_{sen} under different settings of perturbation constraint ε\varepsilon. Both PSNR and SSIM show downward trends with the increase of ε\varepsilon. In other words, the smaller the value of ε\varepsilon, the better the visual quality of the image. Therefore, the value of ε\varepsilon should be as small as possible to maintain a good visual quality for the adversarial example. In conclusion, ε=3/255\varepsilon=3/255 is supposed to be the optimal one when considering the trade-off between image quality and privacy leakage rate.

Refer to caption
Figure 5: The privacy-preserving success rate and the privacy leakage rate of adversarial social images AEsenAE_{sen} under different perturbation constraint ε\varepsilon.
Refer to caption
Figure 6: PSNR and SSIM between original images and the adversarial examples AEsenAE_{sen} under different perturbation constraint ε\varepsilon.

We further explore the impact of the detection threshold TT on the privacy-preserving success rate and privacy leakage rate. The privacy-preserving success rates of AEallAE_{all} and AEsenAE_{sen} under different TT are shown in Fig. 7. It can be seen that the success rate of AEallAE_{all} increases rapidly as TT increases when TT is in [0.2,0.3][0.2,0.3], and the success rate of AEsenAE_{sen} is always in a high level. When TT reaches 0.30.3, the privacy-preserving success rates for AEallAE_{all} and AEsenAE_{sen} are 96.1% and 96.5%, respectively. When T=0.4T=0.4, the privacy-preserving success rates for AEallAE_{all} and AEsenAE_{sen} are 99.5% and 99.6%, respectively. Generally, for most object detectors, the detection threshold is set to be higher than 0.50.5 in order to achieve high detection accuracy. This indicates that, the proposed method has a good performance in terms of privacy preserving when detection threshold TT is set to be greater or equal to 0.30.3.

Refer to caption
Figure 7: Privacy-preserving success rate of adversarial examples AEsenAE_{sen} under different detection threshold TT.

In addition, we also calculate the privacy leakage rate of the proposed method under different settings of the detection threshold TT, as shown in TABLE 3. It is shown that the privacy leakage rate decreases with the increase of TT. When T=0.4T=0.4, the privacy leakage rates of AEallAE_{all} and AEsenAE_{sen} are 0.06% and 0.42%, respectively, which indicate that the proposed method has a good performance in preventing privacy leakage. Generally, for most object detectors, the detection threshold is set to be higher than 0.5 in order to achieve high detection accuracy. The result indicates that, the proposed method can effectively reduce the scores of most bounding boxes to below 0.3, which will lead these bounding boxes to be discarded.

Table 3: Privacy leakage rate of adversarial examples under different detection threshold TT.
Detection threshold 0.2 0.24 0.28 0.32 0.36 0.4
Privacy leakage rate AEallAE_{all} 8.56% 3.84% 0.86% 0.27% 0.12% 0.06%
AEsenAE_{sen} 4.53% 3.36% 2.61% 1.03% 0.50% 0.42%

4.5 Comparison with traditional image processing methods

In this section, we compare the proposed method with the traditional image processing methods. The traditional image processing methods includes low brightness, noise, mosaic, blur, and JPEG (Joint Photographic Experts Group) compression [9, 8].

We select 500 images from MS-COCO dataset to conduct the experiment. Fig. 8 shows an example image processed by different methods and the detection results. As shown in Fig. 8(a), if the image is unprocessed, Faster R-CNN is able to accurately identify all the objects. As shown in Fig. 8(b), although low-brightness method seriously degrades the visual quality of the social image, Faster R-CNN can still identify all the objects correctly. As shown in Fig. 8(c)\simFig. 8(f), the persons are all correctly detected in the images processed by blur, mosaic, noise and JPEG compression method, which indicate that these methods are all ineffective to protect the sensitive objects. As shown in Fig. 8(g) and Fig. 8(h), Faster R-CNN detects nothing in AEallAE_{all} and recognizes person as dining table in AEsenAE_{sen}, which indicate that the proposed adversarial example based method can successfully deceive Faster R-CNN. The reason why the proposed method can resist the detection is that, the feature of a bounding box extracted by Faster R-CNN can be considered as a point in the hyperplane. The proposed method can change the position of the feature point and makes it cross the classification boundary, which will cause the bounding box to be misclassified or be discarded [8].

Refer to caption
(a) Original image
Refer to caption
(b) Low brightness
Refer to caption
(c) Blur
Refer to caption
(d) Mosaic
Refer to caption
(e) Noise
Refer to caption
(f) JPEG Compression
Refer to caption
(g) Adversarial examples AEallAE_{all}
Refer to caption
(h) Adversarial examples AEsenAE_{sen}
Figure 8: The image example processed by different methods and the detection results from Faster R-CNN.

TABLE 4 presents the comparison between the proposed method and the above five image processing methods (low brightness, noise, mosaic, blur, and JPEG compression [9, 8]) in terms of privacy-preserving success rate, privacy leakage rate, PSNR, and SSIM. The results indicate that the proposed method can effectively protect the privacy of social images as the privacy-preserving success rates of AEallAE_{all} and AEsenAE_{sen} are 96.2% and 96.6%, respectively, and the privacy leakage rates of AEallAE_{all} and AEsenAE_{sen} are 0.54% and 2.20%, respectively. The five traditional image processing methods all fail to protect the social privacy as the privacy-preserving success rates of the five methods are all lower than 5.4%. Compared with the five traditional image processing methods, the proposed method has almost no influence on the visual quality of social images as the values of PSNR and SSIM of the proposed method are significantly higher than that of the five traditional image processing methods. Especially, the values of PSNR and SSIM of the proposed AEsenAE_{sen} are 46.91 and 0.992, which are much higher than that of other methods.

Table 4: Comparison between the proposed adversarial example based method and other five image processing methods.
Methods Privacy-preserving success rate Privacy leakage rate PSNR SSIM
Low brightness 0.6% 100% 8.46 0.252
Blur (Gaussian) 0.4% 100% 24.08 0.676
Mosaic 2.2% 78.19% 23.87 0.683
Noise 0.8% 95.55% 22.68 0.437
JPEG compression 5.4% 57.32% 24.96 0.687
The proposed AEallAE_{all} 96.2% 0.54% 27.16 0.751
The proposed AEsenAE_{sen} 96.6% 2.20% 46.91 0.992

5 Conclusion

The malicious attackers can obtain private and sensitive information from uploaded social photos using object detectors. In this paper, we propose an Object Disappearance Algorithm to generate adversarial social images and prevent the object detectors from detecting privacy information. The proposed method injects imperceptible perturbations to the social images before they are uploaded online. Specifically, the proposed method generates two kinds of adversarial social images. One can hide all the objects from being detected by Faster R-CNN, and the other can lead Faster R-CNN to misclassify the customized objects. We use two metrics, named privacy-preserving success rate and privacy leakage rate, to evaluate the effectiveness of the proposed method, and use PSNR and SSIM to evaluate the impact of the proposed method on the visual quality of social images. Experimental results show that, compared with traditional image processing methods (low brightness, blur, mosaic, noise, and JPEG compression), the proposed method can effectively resist the detection of Faster R-CNN, while has the minimal influence on the visual quality of social images. The proposed method iteratively constructs perturbation to generate adversarial social images. In the future work, we will study the acceleration method for adversarial examples generation to reduce the overhead.

References

  • Hardt and Nath [2012] Hardt, M., Nath, S.. Privacy-aware personalization for mobile advertising. In: Proceedings of the ACM Conference on Computer and Communications Security, 2012, pp. 662–673.
  • onl [2017] Japan researchers warn of fingerprint theft from ‘peace’ sign. 2017. URL: https://phys.org/news/2017-01-japan-fingerprint-theft-peace.html.
  • Klemperer et al. [2012] Klemperer, P.F., Liang, Y., Mazurek, M.L., Sleeper, M., Ur, B., Bauer, L., et al. Tag, you can see it! Using tags for access control in photo sharing. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, 2012, pp. 377–386.
  • Vishwamitra et al. [2017] Vishwamitra, N., Li, Y., Wang, K., Hu, H., Caine, K., Ahn, G.. Towards PII-based multiparty access control for photo sharing in online social networks. In: Proceedings of the 22nd ACM on Symposium on Access Control Models and Technologies, 2017, pp. 155–166.
  • Li et al. [2019a] Li, F., Sun, Z., Li, A., Niu, B., Li, H., Cao, G.. Hideme: Privacy-preserving photo sharing on social networks. In: Proceedings of the IEEE Conference on Computer Communications, 2019a, pp. 154–162.
  • Sun et al. [2016] Sun, W., Zhou, J., Lyu, R., Zhu, S.. Processing-aware privacy-preserving photo sharing over online social networks. In: Proceedings of the 24th ACM Conference on Multimedia Conference, 2016, pp. 581–585.
  • Abdulla and Bakiras [2019] Abdulla, A.K., Bakiras, S.. HITC: Data privacy in online social networks with fine-grained access control. In: Proceedings of the 24th ACM Symposium on Access Control Models and Technologies, 2019, pp. 123–134.
  • Liu et al. [2017] Liu, Y., Zhang, W., Yu, N.. Protecting privacy in shared photos via adversarial examples based stealth. Security and Communication Networks 2017. 2017:1–15.
  • Wilber et al. [2016] Wilber, M.J., Shmatikov, V., Belongie, S.. Can we still avoid automatic face detection. In: Proceedings of the IEEE Winter Conference on Applications of Computer Vision, 2016, pp. 1–9.
  • Papernot et al. [2018] Papernot, N., McDaniel, P., Sinha, A., Wellman, M.. SoK: Security and privacy in machine learning. In: Proceedings of the IEEE European Symposium on Security and Privacy, 2018, pp. 399–414.
  • Akhtar and Mian [2018] Akhtar, N., Mian, A.. Threat of adversarial attacks on deep learning in computer vision: A survey. IEEE Access 2018. 6:14410–14430.
  • Ren et al. [2017] Ren, S., He, K., Girshick, R.B., Sun, J.. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Transactions on Pattern Analysis and Machine Intelligence 2017. 39(6):1137–1149.
  • Lin et al. [2014] Lin, T., Maire, M., Belongie, S.J., Hays, J., Perona, P., Ramanan, D., et al. Microsoft COCO: Common objects in context. In: Proceedings of the European Conference on Computer Vision, 2014, pp. 740–755.
  • Everingham et al. [2010] Everingham, M., Gool, L.V., Williams, C.K.I., Winn, J.M., Zisserman, A.. The pascal visual object classes (VOC) challenge. International Journal of Computer Vision 2010. 88(2):303–338.
  • Szegedy et al. [2014] Szegedy, C., Zaremba, W., Sutskever, I., Bruna, J., Erhan, D., Goodfellow, I.J., et al. Intriguing properties of neural networks. In: Proceedings of the 2nd International Conference on Learning Representations, 2014, pp. 1–10.
  • Goodfellow et al. [2015] Goodfellow, I.J., Shlens, J., Szegedy, C.. Explaining and harnessing adversarial examples. In: Proceedings of the 3rd International Conference on Learning Representations, 2015, pp. 1–11.
  • Carlini and Wagner [2017] Carlini, N., Wagner, D.. Towards evaluating the robustness of neural networks. In: Proceedings of the IEEE Symposium on Security and Privacy, 2017, pp. 39–57.
  • Chen et al. [2018] Chen, S., Cornelius, C., Martin, J., Chau, D.H.P.. Shapeshifter: Robust physical adversarial attack on faster R-CNN object detector. In: Proceedings of the Joint European Conference on Machine Learning and Knowledge Discovery in Databases, 2018, pp. 52–68.
  • Brown et al. [2017] Brown, T.B., Mané, D., Roy, A., Abadi, M., Gilmer, J.. Adversarial patch. arXiv:1712.09665 2017.
  • Athalye et al. [2018] Athalye, A., Engstrom, L., Ilyas, A., Kwok, K.. Synthesizing robust adversarial examples. In: Proceedings of the 35th International Conference on Machine Learning, 2018, pp. 284–293.
  • Girshick [2015] Girshick, R.. Fast R-CNN. In: Proceedings of the IEEE International Conference on Computer Vision, 2015, pp. 1440–1448.
  • Xue et al. [2019] Xue, M., Yuan, C., Sun, S., Wu, Z.. A privacy protection method and device for social photos against yolo object detector. 2019. Chinese Patent, Patent No. 201911346202.0, Filed December, 2019.
  • Li et al. [2019b] Li, C.Y., Shamsabadi, A.S., Sanchez-Matilla, R., Mazzon, R., Cavallaro, A.. Scene privacy protection. In: Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, 2019b, pp. 2502–2506.
  • Shen et al. [2019] Shen, Z., Fan, S., Wong, Y., Ng, T., Kankanhalli, M.S.. Human-imperceptible privacy protection against machines. In: Proceedings of the 27th ACM International Conference on Multimedia, 2019, pp. 1119–1128.
  • Shan et al. [2020] Shan, S., Wenger, E., Zhang, J., Li, H., Zheng, H., Zhao, B.Y.. Fawkes: Protecting personal privacy against unauthorized deep learning models. arXiv:2002.08327 2020.
  • Zhao et al. [2018] Zhao, Y., Zhu, H., Liang, R., Shen, Q., Zhang, S., Chen, K.. Seeing isn’t believing: Practical adversarial attack against object detectors. arXiv:1812.10217 2018.
  • Song et al. [2018] Song, D., Eykholt, K., Evtimov, I., Fernandes, E., Li, B., Rahmati, A., et al. Physical adversarial examples for object detectors. In: Proceedings of the 12th USENIX Workshop on Offensive Technologies, 2018, pp. 1–10.
  • Redmon et al. [2016] Redmon, J., Divvala, S.K., Girshick, R., Farhadi, A.. You only look once: Unified, real-time object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 779–788.
  • Liu et al. [2016] Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S.E., Fu, C., et al. SSD: Single shot multibox detector. In: Proceedings of the European Conference on Computer Vision, 2016, pp. 21–37.
  • Dai et al. [2016] Dai, J., Li, Y., He, K., Sun, J.. R-FCN: Object detection via region-based fully convolutional networks. In: Proceedings of the 30th International Conference on Neural Information Processing Systems, 2016, pp. 379–387.
  • Huang et al. [2017] Huang, J., Rathod, V., Sun, C., Zhu, M., Korattikara, A., Fathi, A., et al. Speed/accuracy trade-offs for modern convolutional object detectors. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017, pp. 3296–3297.
  • Szegedy et al. [2016] Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., Wojna, Z.. Rethinking the inception architecture for computer vision. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 2818–2826.
  • Abadi et al. [2016] Abadi, M., Agarwal, A., Barham, P., Brevdo, E., Chen, Z., Citro, C., et al. Tensorflow: Large-scale machine learning on heterogeneous distributed systems. arXiv:1603.04467 2016.
  • Wang et al. [2004] Wang, Z., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P.. Image quality assessment: From error visibility to structural similarity. IEEE Transactions on Image Processing 2004. 13(4):600–612.