This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Soft-Label Anonymous Gastric X-ray Image Distillation

Abstract

This paper presents a soft-label anonymous gastric X-ray image distillation method based on a gradient descent approach. The sharing of medical data is demanded to construct high-accuracy computer-aided diagnosis (CAD) systems. However, the large size of the medical dataset and privacy protection are remaining problems in medical data sharing, which hindered the research of CAD systems. The idea of our distillation method is to extract the valid information of the medical dataset and generate a tiny distilled dataset that has a different data distribution. Different from model distillation, our method aims to find the optimal distilled images, distilled labels and the optimized learning rate. Experimental results show that the proposed method can not only effectively compress the medical dataset but also anonymize medical images to protect the patient’s private information. The proposed approach can improve the efficiency and security of medical data sharing.

Index Terms—  Medical image distillation, medical data sharing, anonymization, gastric X-ray images.

1 Introduction

Deep learning, in particular, deep convolutional neural networks (DCNNs), have been popular with many areas in computer vision [1]. Especially, in the field of medical image analysis, DCNN-based CAD systems are often used as an auxiliary diagnosis of diseases [2, 3]. The sharing of medical data is a primary method for building high-accuracy CAD systems [4]. However, there are still two main problems in medical data sharing. Firstly, with the increase in the number of medical equipment, the amount of medical image data grows exponentially, resulting in large size of the medical dataset [5]. Therefore, the sharing of medical image data is inefficient. It is necessary to extract valid data to reduce the size of the medical dataset. Secondly, because the medical images contain the patient's private information, there is still an enormous controversy over the use of these medical data [6]. It is difficult to share the medical image data of many diseases, which hindered the research and development of CAD systems.

To solve the problem of reducing the size of the dataset, many researchers have proposed unique solutions. For example, many instance selection and dataset pruning methods aim to select a subset of the entire training dataset and achieve performance comparable to the original full dataset, which can reduce the size of the dataset [7, 8]. Similarly, Campbell etalet\,al. proposed a Bayesian core-set construction method that can find valid data in the original dataset via the greedy iterative geodesic ascent [9]. Also, Sener etalet\,al. proposed a core-set construction approach based on active learning [10]. Active learning reduces the amount of data that need to be labeled by only labeling data that are difficult to classify [11, 12]. Although these previous approaches have made some progresses in solving the problem of reducing the size of standard image dataset, due to the high complexity of medical images and the need for professional knowledge of labeling, they still cannot meet the requirements for medical dataset.

Privacy protection problem has always been the main obstacle in medical data sharing [13, 14]. Researchers have also tackled the problem of hiding personal information. For example, some methods for removal of identifiers have been proposed to protect the privacy of patients [15]. These methods contain stripping identifier and one-way hashing algorithm. The stripping identifier approach is an anonymization algorithm which can remove all patient identifiers in a record. And the one-way hashing algorithm transforms a patient record string into another string, such that operations on the hash value cannot calculate the original patient record. In recent years, with the rise of big data, some ways to use cloud computing platforms to share medical data securely have been proposed. For example, Yang etalet\,al. proposed a hybrid solution method of privacy-preserving medical data sharing based on cloud computing [16]. Also, Fabian etalet\,al. proposed a method to realize the safe sharing of medical data with semi-trust cloud computing environments [17]. Since it is challenging to ensure the validity of anonymized medical images, all of these methods do not pay attention to the anonymization of medical image data itself.

In this paper, we propose a novel method that can not only effectively compress the medical dataset but also anonymize medical images to protect the patient's private information [18, 19]. Since the researches of gastric X-ray images have both of the above problems, and we have previously proposed methods for anonymous gastritis image generation and automatic detection of gastritis [20, 21], we also focus on gastric X-ray images in this research. Considering the gastric X-ray images with high resolutions can lead to expensive computing cost, we divide them into patches that have three categories. To maximize the compression of the dataset and distillation of the valid information, we distill each class into one image for training. Then we use the distilled images to estimate the labels of full gastric X-ray images. Experimental results show that our distillation method achieved competitive classification accuracy using a tiny distilled dataset. Furthermore, the sharing of medical image data can become more efficient and safer because the distilled images do not contain private information.

Our contributions are summarized as follows:

  • We propose a novel method for anonymous medical image distillation, which can improve the efficiency and security of medical image data sharing.

  • We realize high classification performance by distilling each class into one image for training.

2 Anonymous Gastric X-ray image distillation

This section shows the details of the soft-label anonymous gastric X-ray image distillation method. In subsection 2.1, we preprocess the training dataset by dividing the gastric images into patches. In subsection 2.2, we demonstrate the entire flow of the anonymous gastric image distillation algorithm. In subsection 2.3, we explain how to estimate the labels of full gastric X-ray images.

2.1 Patch-based Gastric X-ray Image Labeling

This subsection shows how we preprocess the training dataset. The gastric X-ray images in our dataset have high resolutions, e.g.e.g., 2,048 ×\times 2,048 pixels. In practical applications, the high-resolution images can lead to expensive computing cost. Therefore, we divide each gastric X-ray image into H×WH\times W patches (HH and WW respectively denote the number of patches in the vertical direction and the horizontal direction) and manually label these patch-based gastric X-ray images into the following three categories:

  • \mathcal{I}: patches outside of the stomach (irrelevant),

  • 𝒩\mathcal{N}: patches extracted from non-gastritis X-ray images (negative) inside of the stomach,

  • 𝒫\mathcal{P}: patches extracted from gastritis X-ray images (positive) inside of the stomach.

Figures 1 and 2 show the full gastric X-ray images and patch-based gastric X-ray images, respectively.

Algorithm 1 Anonymous Gastric Image Distillation
0:  θ0\theta_{0}: the random initial weights; MM: the number of distilled images; EE: distill epochs; II: distill steps;       α\alpha: learning rate; KK: batch size; TT: training steps; 𝐲~0\tilde{\mathbf{y}}_{0}: initial value for 𝐲~\tilde{\mathbf{y}}; α~0\tilde{\alpha}_{0}: initial value for α~\tilde{\alpha}
0:  𝐱~\tilde{\mathbf{x}}: distilled images; 𝐲~\tilde{\mathbf{y}}: distilled labels; α~\tilde{\alpha}: optimized learning rate
1:  Initialize 𝐱~\tilde{\mathbf{x}} = {x~m}m=1M\left\{\tilde{x}_{m}\right\}_{m=1}^{M} randomly, 𝐲~\tilde{\mathbf{y}} = {y~m}m=1M𝐲~0\left\{\tilde{y}_{m}\right\}_{m=1}^{M}\leftarrow\tilde{\mathbf{y}}_{0}, α~α~0\tilde{\alpha}\leftarrow\tilde{\alpha}_{0}
2:  for each training step t=1t=1 to TT do
3:     Get a minibatch of training data (𝐱t,𝐲t)\left(\mathbf{x}_{t},\mathbf{y}_{t}\right) = {xt,k,yt,k}k=1K\left\{x_{t,k},y_{t,k}\right\}_{k=1}^{K}
4:     Get the random initial weights θ0\theta_{0}
5:     for each distilling epoch e=1e=1 to EE do
6:        for each distilling step i=0i=0 to I1I-1 do
7:           Compute updated weights with gradient descent: θi+1θiα~θi(𝐱~,𝐲~,θi)\theta_{i+1}\leftarrow\theta_{i}-\tilde{\alpha}\,\nabla_{\theta_{i}}\ell\left(\tilde{\mathbf{x}},\tilde{\mathbf{y}},\theta_{i}\right)
8:           Evaluate the objective function on the minibatch of training data: =(𝐱t,𝐲t,θi+1)\mathcal{L}=\ell\left(\mathbf{x}_{t},\mathbf{y}_{t},\theta_{i+1}\right)
9:           Update distilled data: 𝐱~𝐱~α𝐱~\tilde{\mathbf{x}}\leftarrow\tilde{\mathbf{x}}-\alpha\nabla_{\tilde{\mathbf{x}}}\mathcal{L}, 𝐲~𝐲~α𝐲~\tilde{\mathbf{y}}\leftarrow\tilde{\mathbf{y}}-\alpha\nabla_{\tilde{\mathbf{y}}}\mathcal{L}, and α~α~αα~\tilde{\mathbf{\alpha}}\leftarrow\tilde{\mathbf{\alpha}}-\alpha\nabla_{\tilde{\mathbf{\alpha}}}\mathcal{L}
10:        end for
11:     end for
12:  end for

2.2 Anonymous Gastric Image Distillation

In this subsection, we explain our anonymous gastric image distillation algorithm. When we have the patch-based gastric training dataset (𝐱,𝐲)\left(\mathbf{x,y}\right) = {xg,yg}g=1G\left\{x_{g},y_{g}\right\}_{g=1}^{G}, where GG denotes the number of training images, xgx_{g} and ygy_{g} denote the image and the corresponding label, respectively. We parameterize a DCNN model as θ\theta, and let the twice-differentiable loss function (𝐱,𝐲,θ)\ell\left(\mathbf{x,y},\theta\right) denotes the loss of this network on the entire training dataset (𝐱,𝐲)\left(\mathbf{x,y}\right). Considering the gastritis patches and non-gastritis patches may have common features, we let the distilled images 𝐱~\tilde{\mathbf{x}} have soft labels 𝐲~\tilde{\mathbf{y}}, which contain \mathcal{I}, 𝒩\mathcal{N} and 𝒫\mathcal{P} [22]. In our gastric image distillation method, we distill the valid information of the entire training dataset to a tiny distilled dataset (𝐱~,𝐲~)\left(\tilde{\mathbf{x}},\tilde{\mathbf{y}}\right) = {x~m,y~m}m=1M\left\{\tilde{x}_{m},\tilde{y}_{m}\right\}_{m=1}^{M} with MGM\ll G. For each distilling step ii, the weights are updated as follows:

θi+1θiα~θi(𝐱~,𝐲~,θi),\theta_{i+1}\leftarrow\theta_{i}-\tilde{\alpha}\nabla_{\theta_{i}}\ell\left(\tilde{\mathbf{x}},\tilde{\mathbf{y}},\theta_{i}\right), (1)

where (𝐱~,𝐲~)\left(\tilde{\mathbf{x}},\tilde{\mathbf{y}}\right) denotes the distilled dataset, and α~\tilde{\alpha} denotes the optimized learning rate.

Refer to caption
Refer to caption
Fig. 1: Examples of full gastric X-ray images: (a) a sample of non-gastritis image, (b) a sample of gastritis image.
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Refer to caption
Fig. 2: Examples of patch-based gastric X-ray images: (a) irrelevant patches in \mathcal{I}, (b) non-gastritis patches in 𝒩\mathcal{N}, (c) gastritis patches in 𝒫\mathcal{P}.

With the derived new weights θi+1\theta_{i+1} of the DCNN model, we evaluate it on the entire training dataset (𝐱,𝐲)\left(\mathbf{x,y}\right). Our goal is to find the optimal distilled images 𝐱~\tilde{\mathbf{x}}^{*}, distilled labels 𝐲~\tilde{\mathbf{y}}^{*} and the optimized learning rate α~\tilde{\mathbf{\alpha}}^{*}. The objective function is defined as:

𝐱~,𝐲~,α~=argmin(𝐱~,𝐲~,α~;θi),=argmin(𝐱,𝐲,θi+1),=argmin(𝐱,𝐲,θiα~θi(𝐱~,𝐲~,θi)),\begin{split}\tilde{\mathbf{x}}^{*},\tilde{\mathbf{y}}^{*},\tilde{\mathbf{\alpha}}^{*}&=\mathrm{arg\,min}\,\mathcal{L}\left(\tilde{\mathbf{x}},\tilde{\mathbf{y}},\tilde{\alpha};\theta_{i}\right),\\ &=\mathrm{arg\,min}\,\ell\left(\mathbf{x,y,}\,\theta_{i+1}\right),\\ &=\mathrm{arg\,min}\,\ell\left(\mathbf{x,y,}\,\theta_{i}-\tilde{\alpha}\nabla_{\theta_{i}}\ell\left(\tilde{\mathbf{x}},\tilde{\mathbf{y}},\theta_{i}\right)\right),\end{split} (2)

where (𝐱~,𝐲~,θi)\ell\left(\tilde{\mathbf{x}},\tilde{\mathbf{y}},\theta_{i}\right) is twice-differentiable, and (𝐱~,𝐲~,α~;θi)\mathcal{L}\left(\tilde{\mathbf{x}},\tilde{\mathbf{y}},\tilde{\alpha};\theta_{i}\right) is differentiable.

To obtain the optimal distilled images, distilled labels and the optimized learning rate, we update the distilled images 𝐱~\tilde{\mathbf{x}}, distilled labels 𝐲~\tilde{\mathbf{y}} and the optimized learning rate α~\tilde{\mathbf{\alpha}} at each distilling step with gradient descent as follows:

𝐱~𝐱~α𝐱~,𝐲~𝐲~α𝐲~,α~α~αα~,\begin{split}&\tilde{\mathbf{x}}\leftarrow\tilde{\mathbf{x}}-\alpha\nabla_{\tilde{\mathbf{x}}}\mathcal{L},\\ &\tilde{\mathbf{y}}\leftarrow\tilde{\mathbf{y}}-\alpha\nabla_{\tilde{\mathbf{y}}}\mathcal{L},\\ &\tilde{\mathbf{\alpha}}\leftarrow\tilde{\mathbf{\alpha}}-\alpha\nabla_{\tilde{\mathbf{\alpha}}}\mathcal{L},\end{split} (3)

where 𝐱~\nabla_{\tilde{\mathbf{x}}}\mathcal{L}, 𝐲~\nabla_{\tilde{\mathbf{y}}}\mathcal{L} and α~\nabla_{\tilde{\mathbf{\alpha}}}\mathcal{L} respectively denote the gradient of \mathcal{L} based on 𝐱~\tilde{\mathbf{x}}, 𝐲~\tilde{\mathbf{y}} and α~\tilde{\mathbf{\alpha}}, α\alpha denotes the learning rate.

Then we demonstrate the details of the anonymous gastric image distillation algorithm. Firstly, we show our input and output settings. Let θ0\theta_{0} denotes the random initial weights of the DCNN model, MM, EE and II respectively denote the number of distilled images, distill epochs and distill steps. Also, let α\alpha, KK and TT are respectively denote learning rate, batch size and training steps. In addition, 𝐲~0\tilde{\mathbf{y}}_{0} is the initial value of 𝐲~\tilde{\mathbf{y}}, and α~0\tilde{\alpha}_{0} is the initial value of α~\tilde{\alpha}. And we obtain the distilled images 𝐱~\tilde{\mathbf{x}}, distilled labels 𝐲~\tilde{\mathbf{y}} and the optimized learning rate α~\tilde{\alpha} on the output.

Next, we explain the training process of the algorithm. First, we randomly initialize the distilled images 𝐱~\tilde{\mathbf{x}}. Then we initialize the distilled labels 𝐲~\tilde{\mathbf{y}} and the optimized learning rate α~\tilde{\alpha} with 𝐲~0\tilde{\mathbf{y}}_{0} and α~0\tilde{\alpha}_{0}, respectively. At each training step tt, we get a minibatch of training data (𝐱t,𝐲t)\left(\mathbf{x}_{t},\mathbf{y}_{t}\right) whose size is KK. Then we get the random initial weights θ0\theta_{0} and begin the distilling process. We repeat the distilling process for EE times. At each distilling step ii, we compute the updated weights with Eq. (1). Then we evaluate the objective function on the minibatch of training data with Eq. (2). Finally, we update the distilled data 𝐱~\tilde{\mathbf{x}}, 𝐲~\tilde{\mathbf{y}}, and α~\tilde{\alpha} with Eq. (3) based on gradient descent [23, 24]. Algorithm 1 shows the entire flow of our proposed approach.

2.3 Full Gastric X-ray Image Classification

In this subsection, we explain how to estimate the label of a full gastric X-ray image based on patches. First, when we have a test gastric image, we divide it into H×WH\times W patches. Then we put the divided patches into the trained DCNN model, and we can obtain the predicted labels of these patches. Next, we respectively calculate the number of patches whose predicted labels are 𝒩\mathcal{N} and 𝒫\mathcal{P}. The \mathcal{I} patches that extracted from outside of the stomach are not related to the gastritis/non-gastritis prediction, and hence we do not take into account them to the probability calculation. Finally, we estimate the label of a full gastric X-ray image as follows:

ytest={1ifNum(𝒫)Num(𝒫)+Num(𝒩)ϵ0otherwise,y^{\mathrm{test}}=\left\{\begin{matrix}1&\mathrm{if}\,\frac{\mathrm{Num}(\mathcal{P})}{\mathrm{Num}(\mathcal{P})+\mathrm{Num}(\mathcal{N})}\geq\epsilon\hfill\\ 0&\mathrm{otherwise}\hfill\end{matrix}\right., (4)

where ϵ\epsilon is a threshold, Num(𝒫)\mathrm{Num}(\mathcal{P}) denotes the number of gastritis patches, and Num(𝒩)\mathrm{Num}(\mathcal{N}) denotes the number of non-gastritis patches. Note that if ytest=1y^{\mathrm{test}}=1, the estimation result of a full gastric X-ray image is gastritis, and the estimation result is non-gastritis otherwise.

3 Experimental Results

In this section, we verify the effectiveness of our distillation method with experimental results. In subsection 3.1, we show the experimental settings of our method. In subsection 3.2, we evaluate the performance of our method by classifying the full gastric X-ray images.

3.1 Experimental Settings

This subsection shows the experimental settings of our research. The dataset used in our study contains 815 patients' (240 gastritis and 575 non-gastritis) gastric X-ray images. Each image has a ground truth (gastritis/non-gastritis), which was determined by patient diagnosis results of endoscopic examination and X-ray inspection. All of the gastric X-ray images are gray-scale and high-resolution. The training dataset contains 200 patients' (100 gastritis and 100 non-gastritis) images. Also, the rest of the patients' (140 gastritis and 475 non-gastritis) images are included in the test dataset. In the data preprocessing stage, we divided the images into patches (299 ×\times 299 pixels), i.e.i.e., HH = WW = 35, where the sliding interval was set to 50 pixels. Besides, the patches extracted from the training dataset were labeled as \mathcal{I}, 𝒩\mathcal{N} and 𝒫\mathcal{P} by a radiological technologist. Note that if the regions inside of the stomach were less than 1%\% in a patch, it was labeled as \mathcal{I}. In addition, if the regions inside of the stomach were more than 85%\% in a patch, it was labeled as 𝒩\mathcal{N} or 𝒫\mathcal{P}. And we discarded the rest of the patches in the training dataset. As a result, we obtained \mathcal{I}, 𝒩\mathcal{N} and 𝒫\mathcal{P} whose number of patches were 48,385, 42,785 and 45,127, respectively.

In the training phase, we loaded the patch-based training dataset (\mathcal{I}, 𝒩\mathcal{N} and 𝒫\mathcal{P}) into a DCNN model. Since the distilling process involves the calculation of a quadratic gradient, a complicated DCNN model can lead to inefficiency. Hence, we constructed the ResNet18 [25] model with PyTorch framework [26, 27], and the loss \ell was cross entropy loss. In our experiments, we set the number of distilled images of each category to 1, i.e.i.e., MM = 3. Besides, the distill epochs and steps were set to 3, i.e.i.e., EE = II = 3 (total 9 distill steps). And we initialized the soft labels with one-hot values of the original labels, which tends to have higher performance compared to the random initialization. The hard-label distillation method had the same settings as soft-label distillation, except for the labels being fixed [28]. As a result, the training process distilled each class into one image and saved the distilled images, distilled labels and the optimized learning rate. We performed 400 epochs in both of soft-label distillation and hard-label distillation and saved the results after every ten epochs for testing and evaluation. It is difficult for a DCNN model to learn from only several images, so we respectively selected 1,000, 2,000 and 3,000 images per category from the patch-based training dataset and trained three ResNet18 models until convergence.

In the test phase, we performed two experiments (Ex. I and Ex. II) to show the validity of our proposed method quantitatively. Firstly, in Ex. I, we selected the distilled soft-label images that have the best classification performance on the patch-based training dataset and evaluated the performance on the full X-ray images of the test dataset. Specifically, we evaluated the full gastric X-ray images classification performance of the ResNet18 models trained on the three random subsets. Secondly, in Ex. II, we used the hard-label distillation as a comparative method. As in Ex. I, we selected the best distilled hard-label images, which have the highest classification accuracy on the patch-based training dataset. And we utilized the distilled hard-label images to evaluate the classification performance on the full gastric X-ray images. Note that we set the threshold ϵ\epsilon to 0.4 in both of Exs. I and II, which tends to have an excellent classification performance. We utilized the following sensitivity (Sen), specificity (Spe) and harmonic mean (HM) of Sen and Spe as evaluation indexes:

Sen=TPTP+FN,\mathrm{Sen}=\frac{\mathrm{TP}}{\mathrm{TP+FN}}, (5)
Spe=TNTN+FP,\mathrm{Spe}=\frac{\mathrm{TN}}{\mathrm{TN+FP}}, (6)
HM=2×Sen×SpeSen+Spe,\mathrm{HM}=\frac{\mathrm{2\times Sen\times Spe}}{\mathrm{Sen+Spe}}, (7)

where TP, FN, TN and FP denote the number of true positive, false negative, true negative and false positive, respectively.

Table 1: Results of proposed method compared with ResNet18 trained on the random subsets in Ex. I.
Method Sen Spe HM
Proposed Method 0.886 0.869 0.877
ResNet18 (3000) 0.814 0.832 0.823
ResNet18 (2000) 0.907 0.760 0.827
ResNet18 (1000) 0.914 0.669 0.773
Table 2: Results of proposed method compared with hard-label distillation in Ex. II.
Method Sen Spe HM
Proposed Method 0.886 0.869 0.877
Hard-Label Distillation 0.829 0.884 0.856
Refer to caption
(a) distilled hard-label images at distill step 9
Refer to caption
(b) distilled soft-label images at distill step 9
Fig. 3: Examples of distilled images. I: the distilled image of label \mathcal{I}, N: the distilled image of label 𝒩\mathcal{N}, P: the distilled image of label 𝒫{\mathcal{P}}, LR: the optimized learning rate.

3.2 Results and Discussion

The experimental results are shown in Tables 1 and 2. Table 1 shows the full gastric X-ray images classification performance of our proposed method and ResNet18 trained on the random subsets. The ResNet18 model trained with 3,000 images per category (total 9,000 images) has an HM score of 0.823. On the other hand, our soft-label distillation method that distilled each class into only one image for training has a higher HM score of 0.877. We can see that the proposed method realizes competitive classification accuracy with a tiny distilled dataset. Table 2 shows the full gastric X-ray images classification performance of our soft-label distillation and the hard-label distillation method. The HM scores of both the two approaches exceed 0.85, but our soft-label distillation outperforms the hard-label distillation with a higher score. Furthermore, we can see that the classification performance of our method becomes more stable because of the balance of Sen score and Spe score. It means that the distilled images with soft-label can lead to better distillation results. Experimental results clearly showed the validity of our purposed soft-label distillation method.

Figure 3 shows examples of distilled hard-label images and soft-label images used in our experiments. With the distillation methods, the information of gastritis/non-gastritis patch images was extracted and merged into only one image. Note that I (base), N (base) and P (base) respectively denote the image belongs to label \mathcal{I}, 𝒩\mathcal{N} and 𝒫\mathcal{P} with the highest probability. From Fig. 3, we can see that the features of gastritis/non-gastritis patches cannot be distinguished, in other words, the gastric images were anonymized completely. Hence, the distilled images have no private information of patients. Figure 3 clearly showed that our proposed method can effectively compress and anonymize the medical image data.

4 CONCLUSION

In this paper, we have proposed a soft-label anonymous gastric image distillation method. The proposed method realizes high classification performance by distilling each class into one image for training. Furthermore, the sharing of medical image data with our approach can become more efficient and safer because the distilled images are anonymous.

References

  • [1] Yann LeCun, Yoshua Bengio, and Geoffrey Hinton, ``Deep learning,'' Nature, vol. 521, pp. 436–444, 2015.
  • [2] Geert Litjens, Thijs Kooi, Babak Ehteshami Bejnordi, Arnaud Arindra Adiyoso Setio, Francesco Ciompi, Mohsen Ghafoorian, Jeroen Awm Van Der Laak, Bram Van Ginneken, and Clara I Sánchez, ``A survey on deep learning in medical image analysis,'' Medical Image Analysis, vol. 42, pp. 60–88, 2017.
  • [3] Oliver Faust, Yuki Hagiwara, Tan Jen Hong, Oh Shu Lih, and U Rajendra Acharya, ``Deep learning for healthcare applications based on physiological signals: A review,'' Computer Methods and Programs in Biomedicine, vol. 161, pp. 1–13, 2018.
  • [4] Elissa R Weitzman, Liljana Kaci, and Kenneth D Mandl, ``Sharing medical data for health research: The early personal health record experience,'' Journal of Medical Internet Research, vol. 12, no. 2, pp. e14, 2010.
  • [5] Sandra Goldbeck-Wood, Alexander Dorozynski, Leiv Gunnar Lie, Masaya Yamauchi, Christopher Zinn, Deborah Josefson, and Miranda Ingram, ``Complementary medicine is booming worldwide,'' British Medical Journal, vol. 313, no. 7050, pp. 131–134, 1996.
  • [6] Pareek Narendra K. and Patidar Vinod, ``Medical image protection using genetic algorithm operations,'' Soft Computing, vol. 20, no. 2, pp. 763–772, 2016.
  • [7] J Arturo Olvera-López, J Ariel Carrasco-Ochoa, J Francisco Martínez-Trinidad, and Josef Kittler, ``A review of instance selection methods,'' Artificial Intelligence Review, vol. 34, no. 2, pp. 133–143, 2010.
  • [8] Olivier Bachem, Mario Lucic, and Andreas Krause, ``Practical coreset constructions for machine learning,'' arXiv preprint arXiv:1703.06476, 2017.
  • [9] Trevor Campbell and Tamara Broderick, ``Bayesian coreset construction via greedy iterative geodesic ascent,'' in Proceedings of the International Conference on Machine Learning (ICML), 2018, pp. 698–706.
  • [10] Ozan Sener and Silvio Savarese, ``Active learning for convolutional neural networks: A core-set approach,'' in Proceedings of the International Conference on Learning Representations (ICLR), 2018.
  • [11] David A Cohn, Zoubin Ghahramani, and Michael I Jordan, ``Active learning with statistical models,'' Journal of Artificial Intelligence Research, vol. 4, pp. 129–145, 1996.
  • [12] Simon Tong and Edward Chang, ``Support vector machine active learning for image retrieval,'' in Proceedings of the ACM International Conference on Multimedia (ACM MM), 2001, pp. 107–118.
  • [13] Deven McGraw, James X Dempsey, Leslie Harris, and Janlori Goldman, ``Privacy as an enabler, not an impediment: building trust into health information exchange,'' Health Affairs, vol. 28, no. 2, pp. 416–427, 2009.
  • [14] Bradley Malin, David Karp, and Richard H Scheuermann, ``Technical and policy approaches to balancing patient privacy and data sharing in clinical and translational research,'' Journal of Investigative Medicine, vol. 58, no. 1, pp. 11–18, 2010.
  • [15] Jules J Berman, ``Confidentiality issues for medical data miners,'' Artificial Intelligence in Medicine, vol. 26, no. 1-2, pp. 25–36, 2002.
  • [16] Ji-Jiang Yang, Jian-Qiang Li, and Yu Niu, ``A hybrid solution for privacy preserving medical data sharing in the cloud environment,'' Future Generation Computer Systems, vol. 43-44, pp. 74–86, 2015.
  • [17] Benjamin Fabian, Tatiana Ermakova, and Philipp Junghanns, ``Collaborative and secure sharing of healthcare data in multi-clouds,'' Information Systems, vol. 48, pp. 132–150, 2015.
  • [18] Yoshua Bengio, ``Gradient-based optimization of hyperparameters,'' Neural Computation, vol. 12, no. 8, pp. 1889–1900, 2000.
  • [19] Dougal Maclaurin, David Duvenaud, and Ryan Adams, ``Gradient-based hyperparameter optimization through reversible learning,'' in Proceedings of the International Conference on Machine Learning (ICML), 2015, pp. 2113–2122.
  • [20] Ren Togo, Kenta Ishihara, Takahiro Ogawa, and Miki Haseyama, ``Anonymous gastritis image generation via adversarial learning from gastric x-ray images,'' in Proceedings of the International Conference on Image Processing (ICIP), 2018, pp. 2082–2086.
  • [21] Misaki Kanai, Ren Togo, Takahiro Ogawa, and Miki Haseyama, ``Gastritis detection from gastric x-ray images via fine-tuning of patch-based deep convolutional neural network,'' in Proceedings of the International Conference on Image Processing (ICIP), 2019, pp. 1371–1375.
  • [22] Ilia Sucholutsky and Matthias Schonlau, ``Soft-label dataset distillation and text dataset distillation,'' arXiv preprint arXiv:1910.02551, 2019.
  • [23] James S Bergstra, Rémi Bardenet, Yoshua Bengio, and Balázs Kégl, ``Algorithms for hyper-parameter optimization,'' in Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), 2011, pp. 2546–2554.
  • [24] Chelsea Finn, Pieter Abbeel, and Sergey Levine, ``Model-agnostic meta-learning for fast adaptation of deep networks,'' in Proceedings of the International Conference on Machine Learning (ICML), 2017, pp. 1126–1135.
  • [25] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun, ``Deep residual learning for image recognition,'' in Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), 2016, pp. 770–778.
  • [26] Adam Paszke, Sam Gross, Soumith Chintala, Gregory Chanan, Edward Yang, Zachary DeVito, Zeming Lin, Alban Desmaison, Luca Antiga, and Adam Lerer, ``Automatic differentiation in pytorch,'' in Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), Workshop, 2017.
  • [27] Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, et al., ``Pytorch: An imperative style, high-performance deep learning library,'' in Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), 2019, pp. 8024–8035.
  • [28] Tongzhou Wang, Jun-Yan Zhu, Antonio Torralba, and Alexei A. Efros, ``Dataset distillation,'' arXiv preprint arXiv:1811.10959, 2018.