This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Federated Learning without Revealing the Decision Boundaries

Roozbeh Yousefzadeh
Yale University and VA Connecticut Healthcare System
[email protected]
Abstract

We consider the recent privacy preserving methods that train the models not on original images, but on mixed images that look like noise and hard to trace back to the original images. We explain that those mixed images will be samples on the decision boundaries of the trained model, and although such methods successfully hide the contents of images from the entity in charge of federated learning, they provide crucial information to that entity about the decision boundaries of the trained model. Once the entity has exact samples on the decision boundaries of the model, they may use it for effective adversarial attacks on the model during training and/or afterwards. If we have to hide our images from that entity, how can we trust them with the decision boundaries of our model? As a remedy, we propose a method to encrypt the images, and have a decryption module hidden inside the model. The entity in charge of federated learning will only have access to a set of complex-valued coefficients, but the model will first decrypt the images and then put them through the convolutional layers. This way, the entity will not see the training images and they will not know the location of the decision boundaries of the model.

1 Introduction

Our focus in this paper is on privacy methods for federated learning and we specifically consider the methods that hide the contents of images from the entity in charge of federated training, denoted by 𝒯\mathcal{T}. In this setting, 𝒯\mathcal{T} is in charge of training a model, but training images are owned by several entities that are not willing to share their images with each other and with 𝒯\mathcal{T}. Owners are willing to participate in training the model, only if their images are kept private. For example, image owners can be medical institutions that have invested in analyzing and labeling medical images of patients. They may be willing to participate in training an image classification model while keeping their images private (Li et al., 2020).

Many studies have considered this problem and different methods have been proposed (Yang et al., 2019). For example, some methods propose that image owners compute the gradient of loss function of the model on their images and provide the gradients to 𝒯\mathcal{T} instead of providing the images. Several issues may arise in this setting, as the model parameters will change after implementing each gradient batch, and communicating the model parameters to image owners after each update can be quite expensive (Konečnỳ et al., 2016).

Another line of research mixes the training images so that they look like noise and then provides those mixed images to 𝒯\mathcal{T} for training. The method proposed by Huang et al. (2020) mixes the images so that they are provably hard to be traced back to original images, and labels the mixed images such that 𝒯\mathcal{T} can achieve relatively high accuracy when it trains the model on those mixed images. This is the line of research we are considering in this paper. We explain that those mixed images will be exact points on the decision boundaries of the trained model, and although 𝒯\mathcal{T} does not have access to original training images, it will know where the decision boundaries of the model will be. If we cannot trust 𝒯\mathcal{T} with our training images, how can we trust it with the decision boundaries of our model?

In the following, we first discuss the decision boundaries of deep learning models. We then explain the limitations of privacy achieved by training methods such as Huang et al. (2020), and finally, we propose the outline of a new method to preserve the privacy of training images without revealing the decision boundaries of the model.

2 Deep learning models and their decision boundaries

Deep learning models for image classification are classification functions that partition their domain and assign a class to each partition. Partitions are defined by decision boundaries and so is the model. In concept, this can be viewed the same as a linear regression model. A linear regression model partitions its domain with a hyper-plane where the hyper-plane is its decision boundary. Because the pixel space is bounded and usually normalized, the domain of an image classification model can be considered bounded, too, usually a hyper-cube where each dimension ranges between 0 and 1. The decision boundaries of deep learning models are geometrically complex and far from being a hyper-plane (Fawzi et al., 2018), and this complexity is what makes these models powerful classification functions.

As a result, these models/functions are considered black boxes and there are many lingering questions about them (Strang, 2019). The adversarial vulnerability of these models can be explained by the closeness of samples to decision boundaries, but it is not clear why decision boundaries are so close to samples (Shafahi et al., 2019). While many studies try to exploit the decision boundaries of the models for adversarial purposes (Tramer et al., 2020), another set of studies are trying to push the decision boundaries away from samples in order to make the models less vulnerable (Cohen et al., 2019). Studies on generalization of these models are also concerned about the decision boundaries (Elsayed et al., 2018; Jiang et al., 2019).

The training process of image classification models can be viewed as defining the location of decision boundaries in the domain. Recently, there have been some studies that train the models in interesting ways other than just minimizing a loss function. One of the most interesting of these methods is by Zhang et al. (2018). They use conventional training regimes and cross-entropy loss, but instead of training the model on original training images, they mix the images (with different labels) by averaging them pairwise in the pixel space and averaging their labels, as well. This way, they train the models on the mixed images and it turns out to be very effective for better generalization and more robustness of models against adversarial attacks and memorization. Although their study is not focused on the decision boundaries, their training method directly defines the location of decision boundaries in between training samples because the labels they use for mixed images are actually flip labels between classes.

Later, several studies leveraged the mixing aspect of this method for federated learning (Fu et al., 2019; Boulemtafes et al., 2020). Building on that line of research, Huang et al. (2020) proposed a method called InstaHide, where the entity in charge of training the model has access to certain mixed images that cannot be easily traced back to original images in the private training sets. Again, when the model learns such images, it learns where the location of its decision boundaries should be.

3 The extent of privacy we can achieve

The method proposed by Huang et al. (2020) successfully hides the original images from 𝒯\mathcal{T}. In fact, images used for training do not look anything similar to the original images and it is indeed hard to trace them back to original images. But to what extent does their method help us with preserving privacy?

The images used for training have flip labels which exactly define where the location of decision boundaries of the model should be. It follows that the entity that trains the model has direct access to samples on the decision boundaries of the model, because each training sample is a point on a decision boundary. Hence, by putting the training samples together, that entity can acquire a good picture of the partitions of each class defined in the domain.

If we have to hide the contents of original training samples from the entity in charge of training, should not we be concerned about the entity having access to the decision boundaries of the model? What is the extent of the privacy preserved by hiding the contents of images but giving out samples on the decision boundaries of our model?

From the adversarial attacks literature, we know that when an entity knows the decision boundaries of our models, they will know the functional behavior of it and they can affect that adversarially, during training and also after the training.

4 Decryption module inside the model

As a remedy, here we propose a method to encrypt the images with shearlet systems at the origin, so that only shearlet coefficients are shared with entity 𝒯\mathcal{T}. There will be a decryption module inside the model that would decrypt the images before putting them through the convolutional layers. After the model is trained, one can remove the decryption module from the model and directly feed images to its convolutional layers.111If desired, decryption module can still remain inside the model for the testing time as well. This way, the model is trained on images, but training images are kept hidden from 𝒯\mathcal{T}. Moreover, 𝒯\mathcal{T} will have no immediate knowledge about the decision boundaries of the trained model.

In the following, we will briefly explain the shearlet systems which are recently used for image encryption. Using a complicated function to build a shearlet system and using truncated filter boundaries will lead to complex-valued shearlet coefficients, and it would be hard to trace those coefficients back to the actual images without having the shearlet system.

4.1 Background on Shealets

Shearlets are a multi-scale framework for efficient processing and representation of multidimensional data, especially images. Shearlets were built on the scientific knowledge of wavelets in order to address the limitations of wavelets in handling anisotropic features, e.g., edge features in images (Labate et al., 2005).

Over the decades, special tools have been developed to process images (along with other forms of data and functional representations) and to extract useful features from them. Such tools include the Fourier transform, wavelets, shearlets, etc. In fact, wavelets were developed building on the knowledge of Fourier transform in the context of image and signal processing. For example, Daubechies (1990) showed that wavelets perform better than windowed Fourier transform on visual signals, because wavelets handle the frequencies in a nonlinear way. Although wavelets have proven to be more capable than the Fourier transform, they still have shortcomings when it comes to analyzing edge features, because they can only handle point-wise singularities. Shearlets are well-equipped in this respect because, as their name suggests, they extract anisotropic features by deploying a shear matrix. Interaction between the shear matrix and appropriate scaling allows them to operate on a plane in a rotation-like manner, capturing edge features with various shapes (Kutyniok & Labate, 2012).

This makes the shearlets more capable in image processing and at the same time, it makes them computationally more sophisticated and therefore, more suitable for encryption. Wavelets have been used for image encryption in a variety of settings (Chen & Zhao, 2005; 2008; Luo et al., 2015), which hints that shearlets can be even more effective in hiding the contents of images, as already used by Kumar et al. (2018); Chen et al. (2020).

4.2 Shearlet Systems

In order to decompose images with shearlets, we first need to create a shearlet system. A shearlet system is composed of three operations: Dilation, Shearing, and Translation. Here, we briefly review the definition provided by Andrade-Loarca et al. (2020).

Let Aa,a:=\{0}A_{a},a\in\mathbbm{R}:=\mathbbm{R}\backslash\{0\} be a periodic scaling matrix and Ss,sS_{s},s\in\mathbbm{R} be a shearing matrix given by

Aa=(a00|a|1/2)andSs=(1s01).A_{a}=\begin{pmatrix}a&0\\ 0&|a|^{1/2}\end{pmatrix}\;\text{and}\;S_{s}=\begin{pmatrix}1&s\\ 0&1\end{pmatrix}.

For B2×2B\in\mathbbm{R}^{2\times 2}, the dilation operator is defined by

DB:L2(2)L2(2),(DBf)(x)|det(B)|1/2f(M1x)D_{B}:L^{2}(\mathbbm{R}^{2})\rightarrow L^{2}(\mathbbm{R}^{2})\quad,\quad(D_{B}f)(x)\mapsto|\det(B)|^{-1/2}f(M^{-1}x)

Plugging AaA_{a} and SsS_{s} as BB in the dilation operator yields the scaling and shearing operators, respectively. The last ingredient is the translation operator, defined for Tt,t2T_{t},t\in\mathbbm{R}^{2} as

Tt:L2(2)L2(2),(Ttf)(x)f(xt)T_{t}:L^{2}(\mathbbm{R}^{2})\rightarrow L^{2}(\mathbbm{R}^{2})\quad,\quad(T_{t}f)(x)\mapsto f(x-t)

Then, for a given function ψL2(2)\psi\in L^{2}(\mathbbm{R}^{2}) the continuous shearlet system 𝒮(ψ)\mathcal{SH}(\psi) is defined by

𝒮(ψ)={ψa,s,t:=TtDSsDAaψ=a3/4ψ(Aa1Ss1(.t)):a,s,t2}.\mathcal{SH}(\psi)=\{\psi_{a,s,t}:=T_{t}D_{S_{s}}D_{A_{a}}\psi=a^{-3/4}\psi(A_{a}^{-1}S_{s}^{-1}(.-t)):a\in\mathbbm{R}^{*},s\in\mathbbm{R},t\in\mathbbm{R}^{2}\}. (1)

There are certain admissibility requirements for ψ\psi. The ψ\psi would actually be the key that would passed to the hidden module in the model and also shared with image owners. Both the hidden module and the image owner will first construct the shearlet system based on the given ψ\psi, and then use the 𝒮(ψ)\mathcal{SH}(\psi) to decompose or reconstruct the images, as we explain in the next section.

4.3 Processing images with shearlets

To decompose image xx with shearlet system 𝒮(ψ)\mathcal{SH}(\psi), we use

[c]=shdec(x,𝒮(ψ)),[c]=\mathrm{shdec}(x,\mathcal{SH}(\psi)), (2)

This operation is reversible, meaning that given a shearlet system and a set of shearlet coefficients, we can reconstruct an image

[x]=shrec(c,𝒮(ψ)).[x]=\mathrm{shrec}(c,\mathcal{SH}(\psi)). (3)

The first function will be used by image owners to decompose their images, and the second function will be hidden inside the model to reconstruct the images before putting them through the convolutional layers.

Generally, shearlet coefficients can be real-valued or complex. We suggest using truncated boundary conditions which would lead to complex shearlet coefficients, making it harder for the adversary to decrypt the images.

5 Conclusion

In this paper, we reviewed the limitations of privacy preserving methods that mix training images in order to hide them from the entity in charge of federated training. We explained that although such methods hide the contents of training images, they provide exact samples on the decision boundaries of the trained to that same entity. If the entity cannot be trusted with our training images, how can we trust it with the decision boundaries of our models? As a remedy, we proposed using shearlets to encrypt the images while transmitting them in the federated learning system, and then having a decryption module hidden inside the model in order to reconstruct the images before feeding them through the convolutional layers. In this setting, not only we have hidden the training images, we have also hidden the location of the decision boundaries.

6 Future work

We shall focus on the choice of ψ\psi for creating the shearlet system and rigorously show that inferring the images from their complex-valued shearlet coefficients will be hard. There has been some studies on this front in the literature which we aim to expand on. It can be particularly useful to study the shearlet encryption in the Shamir’s secret sharing scheme (Yang et al., 2004). This has been studied to some extent for wavelets (Ulutas et al., 2011), but shearlet systems can be more complicated than wavelets.

Acknowledgments

R.Y. is supported by a fellowship from the Department of Veterans Affairs. The views expressed in this manuscript are those of the author and do not necessarily reflect the position or policy of the Department of Veterans Affairs or the United States government.

References

  • Andrade-Loarca et al. (2020) Héctor Andrade-Loarca, Gitta Kutyniok, and Ozan Öktem. Shearlets as feature extractor for semantic edge detection: the model-based and data-driven realm. Proceedings of the Royal Society A, 476(2243):20190841, 2020.
  • Boulemtafes et al. (2020) Amine Boulemtafes, Abdelouahid Derhab, and Yacine Challal. A review of privacy-preserving techniques for deep learning. Neurocomputing, 384:21–45, 2020.
  • Chen & Zhao (2005) Linfei Chen and Daomu Zhao. Optical image encryption based on fractional wavelet transform. Optics Communications, 254(4-6):361–367, 2005.
  • Chen & Zhao (2008) Linfei Chen and Daomu Zhao. Image encryption with fractional wavelet packet method. Optik, 119(6):286–291, 2008.
  • Chen et al. (2020) Mingming Chen, Guangbiao Ma, Chen Tang, and Zhenkun Lei. Generalized optical encryption framework based on shearlets for medical image. Optics and Lasers in Engineering, 128:106026, 2020.
  • Cohen et al. (2019) Jeremy Cohen, Elan Rosenfeld, and Zico Kolter. Certified adversarial robustness via randomized smoothing. In International Conference on Machine Learning, pp. 1310–1320, 2019.
  • Daubechies (1990) Ingrid Daubechies. The wavelet transform, time-frequency localization and signal analysis. IEEE Transactions on Information Theory, 36(5):961–1005, 1990.
  • Elsayed et al. (2018) Gamaleldin Elsayed, Dilip Krishnan, Hossein Mobahi, Kevin Regan, and Samy Bengio. Large margin deep networks for classification. In Advances in Neural Information Processing Systems, pp. 842–852, 2018.
  • Fawzi et al. (2018) Alhussein Fawzi, Seyed-Mohsen Moosavi-Dezfooli, Pascal Frossard, and Stefano Soatto. Empirical study of the topology and geometry of deep networks. In IEEE Conference on Computer Vision and Pattern Recognition, pp.  3762–3770, 2018.
  • Fu et al. (2019) Yingwei Fu, Huaimin Wang, Kele Xu, Haibo Mi, and Yijie Wang. Mixup based privacy preserving mixed collaboration learning. In IEEE International Conference on Service-Oriented System Engineering, pp.  275–2755, 2019.
  • Huang et al. (2020) Yangsibo Huang, Zhao Song, Kai Li, and Sanjeev Arora. Instahide: Instance-hiding schemes for private distributed learning. In International Conference on Machine Learning, pp. 4507–4518, 2020.
  • Jiang et al. (2019) Yiding Jiang, Dilip Krishnan, Hossein Mobahi, and Samy Bengio. Predicting the generalization gap in deep networks with margin distributions. In International Conference on Learning Representations, 2019.
  • Konečnỳ et al. (2016) Jakub Konečnỳ, H Brendan McMahan, Felix X Yu, Peter Richtárik, Ananda Theertha Suresh, and Dave Bacon. Federated learning: Strategies for improving communication efficiency. arXiv preprint arXiv:1610.05492, 2016.
  • Kumar et al. (2018) Ravi Kumar, Basanta Bhaduri, and Bryan Hennelly. Qr code-based non-linear image encryption using shearlet transform and spiral phase transform. Journal of Modern Optics, 65(3):321–330, 2018.
  • Kutyniok & Labate (2012) Gitta Kutyniok and Demetrio Labate. Shearlets: Multiscale analysis for multivariate data. Springer Science & Business Media, 2012.
  • Labate et al. (2005) Demetrio Labate, Wang-Q Lim, Gitta Kutyniok, and Guido Weiss. Sparse multidimensional representation using shearlets. In Wavelets XI, volume 5914, pp.  59140U. International Society for Optics and Photonics, 2005.
  • Li et al. (2020) Tian Li, Anit Kumar Sahu, Ameet Talwalkar, and Virginia Smith. Federated learning: Challenges, methods, and future directions. IEEE Signal Processing Magazine, 37(3):50–60, 2020.
  • Luo et al. (2015) Yuling Luo, Minghui Du, and Junxiu Liu. A symmetrical image encryption scheme in wavelet and time domain. Communications in Nonlinear Science and Numerical Simulation, 20(2):447–460, 2015.
  • Shafahi et al. (2019) Ali Shafahi, W Ronny Huang, Christoph Studer, Soheil Feizi, and Tom Goldstein. Are adversarial examples inevitable? In International Conference on Learning Representations, 2019.
  • Strang (2019) Gilbert Strang. Linear Algebra and Learning from Data. Wellesley-Cambridge Press, 2019.
  • Tramer et al. (2020) Florian Tramer, Nicholas Carlini, Wieland Brendel, and Aleksander Madry. On adaptive attacks to adversarial example defenses. In Advances in Neural Information Processing Systems, pp. 1633–1645, 2020.
  • Ulutas et al. (2011) Mustafa Ulutas, Güzin Ulutas, and Vasif V Nabiyev. Medical image security and EPR hiding using Shamir’s secret sharing scheme. Journal of Systems and Software, 84(3):341–353, 2011.
  • Yang et al. (2004) Chou-Chen Yang, Ting-Yi Chang, and Min-Shiang Hwang. A (t, n) multi-secret sharing scheme. Applied Mathematics and Computation, 151(2):483–490, 2004.
  • Yang et al. (2019) Qiang Yang, Yang Liu, Tianjian Chen, and Yongxin Tong. Federated machine learning: Concept and applications. ACM Transactions on Intelligent Systems and Technology, 10(2):1–19, 2019.
  • Zhang et al. (2018) Hongyi Zhang, Moustapha Cisse, Yann N Dauphin, and David Lopez-Paz. mixup: Beyond empirical risk minimization. In International Conference on Learning Representations, 2018.