This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Sparse Network Inversion for Key Instance Detection in Multiple Instance Learning

Beomjo Shin, Junsu Cho, Hwanjo Yu§ Department of CSE
POSTECH, Korea
Email: {sbj9180, junsu7463, hwanjoyu}@postech.ac.kr
   Seungjin Choi Inference Lab
BARO AI, Korea
Email: [email protected]
Abstract

Multiple Instance Learning (MIL) involves predicting a single label for a bag of instances, given positive or negative labels at bag-level, without accessing to label for each instance in the training phase. Since a positive bag contains both positive and negative instances, it is often required to detect positive instances (key instances) when a set of instances is categorized as a positive bag. The attention-based deep MIL model is a recent advance in both bag-level classification and key instance detection (KID). However, if the positive and negative instances in a positive bag are not clearly distinguishable, the attention-based deep MIL model has limited KID performance as the attention scores are skewed to few positive instances. In this paper, we present a method to improve the attention-based deep MIL model in the task of KID. The main idea is to use the neural network inversion to find which instances made contribution to the bag-level prediction produced by the trained MIL model. Moreover, we incorporate a sparseness constraint into the neural network inversion, leading to the sparse network inversion which is solved by the proximal gradient method. Numerical experiments on an MNIST-based image MIL dataset and two real-world histopathology datasets verify the validity of our method, demonstrating the KID performance is significantly improved while the performance of bag-level prediction is maintained.

§§footnotetext: Corresponding author

I Introduction

Machine learning has successfully solved typical classification problems where a class label is assigned to each data instance. However, in many real-world classification problems, multiple instances are observed and only a general class label for these instances are given [1]. Such classification problems are called multiple instance learning (MIL) problem [2, 3]. MIL problems are commonly found in read-world scenarios, such as medical field [4]. For example, a medical image can consist of several image patches, each of which correspond to benign or malignant. The label of the image is determined by the presence of the malignant patch. These MIL problems are necessary because labeling every data instance requires a lot of costs (labor and time) from experts [4].

MIL has traditionally assumed that the bag-level label is positive if at least one of the instances in a bag is a positive instance, otherwise negative, and there are no dependency and no ordering between instances in a bag [5]. Under the assumption, MIL learns the relationship between a set of instances called bag and bag-level label. In addition to learning the mapping between bag and bag-level label, MIL also aims to find the positive instances called key instances that trigger the bag-level label without accessing instance-level label [6]. These two tasks of the MIL are called bag-level classification and key instance detection (KID), respectively.

The conventional methods for bag-level classification and KID in MIL are based on instance-space paradigm [7] that trains an instance-level classifier [8, 9, 10]. The methods have performance limits in bag-level classification. The reason is that the instance-level classifier is not sufficiently trained as the data for MIL do not contain instance-level labels [1]. To solve this problem, an attention-based deep MIL model is proposed [1], which is based on embedding-space paradigm [7]. The attention-based deep MIL model transforms the instances in a bag to instance embeddings, and then makes a bag embedding by aggregating the instance embeddings through attention-based pooling. Finally, the MIL model calculates a bag-level score by using a bag-level classifier that takes the bag embedding as input. This kind of MIL model has better bag-level classification performance than the methods based on instance-space paradigm because the model utilizes the bag-level data and label to train the bag-level classifier, not the instance-level classifier. Moreover, the attention-based deep MIL model can perform KID by using the attention scores of the instances in a bag when the bag-level label is positive.

Despite the superiority of the bag-level classification performance of the attention-based deep MIL model, the model has performance limits in terms of KID. The reason is that the model cannot learn every key instance in detail as the model uses only the bag-level data and bag-level label without accessing instance-level data and instance-level label. Thus, the attention-based deep MIL model focuses only on few distinguishable key instances that trigger a bag-level label, and then the indistinguishable key instances are treated as negative instances by the model.

In the paper, we propose a novel method to improve KID performance of the attention-based deep MIL model while maintaining the bag-level classification performance. Our method consists of two modules: (1) a trained MIL module to make a bag-level prediction; (2) neural network inversion with a sparseness constraint (sparse network inversion) module. For the first module, an attention-based deep MIL model is trained by optimizing negative log-likelihood function for bag-level label. In the second module, whenever the trained MIL model classifies the input bag as positive, the sparse network inversion changes the values of instances in the bag to fit the criterion of the MIL model. In addition to optimizing the bag to fit the criterion of the trained MIL model, parts of the optimized bag that do not affect the criterion are weakened by the effect of a sparseness constraint. That is, as a result of sparse network inversion, the parts of the optimized bag that make the bag positive are strengthened, and the other parts are weakened. Therefore, if the optimized bag is put to the trained MIL model then, the attention scores corresponding to the key instances become larger than before applying sparse network inversion to the bag. As a result, our method relieves the problem of the attention-based deep MIL model by removing the constraint that input data cannot be changed, so the KID performance of the attention-based deep MIL model is significantly improved. In the experiments, we show that our method significantly improves the KID performance with the superiority of bag-level classification performance in the attention-based deep MIL model for an MNIST-based image MIL dataset and two real-world histopathology datasets.

II Related Work

II-1 Multiple Instance Learning and Key Instance Detection

MIL is a type of typical weakly supervised learning. Since MIL is relatively free from the cost of labeling data compared to standard supervised learning, the methods for classification of multiple instances have been actively studied after it was first proposed for Drug Activity Prediction in 1997 [2, 8, 1]. KID is a task of MIL, which detects key instances in a positive bag [6]. KID is more challenging than bag-level classification [6], as the MIL models are trained only with bag-level labels without accessing instance-level labels. Despite its difficulty, KID is considered as an important task and many methods for KID have been developed [6, 11, 1] as the bag-level decision of the MIL model can be interpreted with the key instances that make the bag-level label positive. The interpretation of bag-level decision is important in practical applications. For example, in the computational pathology, providing both a final diagnosis and the basis of the diagnosis is more useful than providing only diagnosis about the image [1].

The conventional methods for bag-level classification and KID include the MIL methods based on instance-space paradigm and Voting Framework (VF) solution [8, 11, 6]. The MIL methods based on instance-space paradigm train an instance-level classifier to calculate the score of each instance in a bag for instance-level classification. Then, the responses of the instance-level classifier are aggregated for bag-level classification. These methods have a worse bag-level classification performance than the models with embedding-space paradigm, as the instance-level classifier is trained only with the bag-level data and label [1]. The VF solution is a method of using neighborhood relations among instances. Therefore, this method can be effective only when a proper neighborhood relationship among instances are obtained [6], whereas the method is difficult to apply in many real situations. In order to solve these problem, the methods that apply interpretable attention-based pooling or dynamic pooling to embedding-space paradigm have recently been proposed [1, 12]. These methods use the bag embedding to classify a bag. Since only bag-level information is used for training the MIL model without accessing the instance-level information, these methods have better bag-level classification performance than the MIL models based on instance-space paradigm. However, the MIL model using dynamic pooling is not suitable for situations where the instances in a bag are independent as the model utilizes the contextual information among the instances in the bag. In the case of the MIL model using attention-based pooling, the model has performance limits in terms of KID as the MIL model cannot learn every key instance in detail.

Refer to caption
Figure 1: The flow diagram of the proposed method.

II-2 Neural Network Inversion

Neural network inversion is a method that optimizes the input to find an input pattern that fits the criterion of the trained neural network model [13]. Neural network inversion consists of two steps: the first step is to train the neural network to optimize its criterion, and the second step is to optimize the input variables of the trained neural network [14, 13]. Neural network inversion is used to find optimal input(s) in various industrial fields, such as a plant or a device for a heating, ventilation, and air conditioning (HVAC) system of a building [14, 15].

For bag-level classification, we use the attention-based deep MIL model. For KID, to solve the problem of the attention-based deep MIL model, we apply a neural network inversion with a sparseness constraint that updates the instances in a positive bag. That is, our method removes the constraint on data, so that the optimized instances can be changed to fit the criterion of the model, and the model can easily find out the key instances by using the optimized instances. To the best of our knowledge, our method is the first attempt that applies neural network inversion to the MIL model.

III The Proposed Method

Suppose that we are given a set of images {Xn|n=1,,N}\{X_{n}\,|\,n=1,\ldots,N\}, each of which contains KK image patches. That is, an image XnX_{n} is treated as a bag of KK instances. The number of patches, KK, can vary depending on images. Each image in the training set is associated with a label Yn{0,1}Y_{n}\in\{0,1\} (Yn=1Y_{n}=1 indicates that XnX_{n} is a positive bag) at the bag level, without the knowledge of an instance-level label Yn,kY_{n,k}.

The flow diagram of our method is shown in Fig. 1, where an input image XnX_{n} is applied to the method which constitutes two modules: (1) an already-trained MIL module to yield a bag-level prediction; (2) a sparse network inversion module. We use the attention-based deep MIL model [1] that is already trained using the training set. Our contribution lies at the sparse network inversion to refine the input image, finding instances that make contributions to the bag-level prediction Y^n\widehat{Y}_{n} made by the trained MIL module. The MIL module is first trained using the training set. Then, whenever Y^n=1\widehat{Y}_{n}=1, the sparse network inversion is applied to refine the input image, leading to X^n\widehat{X}_{n}. The refined input image X^n\widehat{X}_{n} is applied to the MIL module again to recompute the attention scores that are used to finally detect key instances.

III-A Problem Formulation

MIL deals with a set of bags {Xn|n=1,,N}\{X_{n}\,|\,n=1,\ldots,N\}, where NN is the number of bags. A bag XnX_{n} is associated with multiple instances {Xn,kIRD|k=1,,K}\{X_{n,k}\in{\rm I\!R}^{D}\,|\,k=1,\ldots,K\}, where KK is the number of instances (KK can vary for different bags), and there are no dependency and no ordering between KK instances in a bag. Each instance in the bag XnX_{n} is associated with an instance-level label {Yn,k{0,1}|k=1,,K}\{Y_{n,k}\in\{0,1\}\,|\,k=1,\ldots,K\}, and the bag XnX_{n} is associated with the bag-level label Yn=OR(Yn,1,,Yn,K)Y_{n}=OR(Y_{n,1},\ldots,Y_{n,K}). That is, the bag-level label is positive if at least one of the instances in the bag is a positive instance, otherwise negative. Under the MIL assumption, an attention-based deep MIL model computes a bag probability PnP_{n} and attention scores ana_{n} for the nthn^{\text{th}} bag XnX_{n}, where ana_{n} is a set of attention scores of KK instances in the nthn^{\text{th}} bag, i.e., an={an,1,,an,K}a_{n}=\{a_{n,1},\ldots,a_{n,K}\}. For the nthn^{\text{th}} bag XnX_{n}, the bag-level prediction Y^n\widehat{Y}_{n} is computed by applying a threshold to bag probability PnP_{n}. In the case of KID, whenever Y^n=1\widehat{Y}_{n}=1, for the kthk^{\text{th}} instance Xn,kX_{n,k} in the nthn^{\text{th}} bag, an instance-level prediction Y^n,k\widehat{Y}_{n,k} is computed by applying a threshold to the normalized attention score a^n,k\widehat{a}_{n,k} after Min-Max normalization is applied to the attention score an,ka_{n,k}.

III-B Attention-based deep MIL model

We build the attention-based deep MIL model that uses attention-based pooling on embedding-space paradigm as shown in Fig. 1, and we train the model by optimizing negative log-likelihood function. There are two reasons we use this structure for the MIL model. First, embedding-space paradigm has better bag-level classification performance than instance-space paradigm. The second reason is that attention-based pooling allows embedding-space paradigm to detect key instances because the magnitude of the attention score corresponding to an instance in a positive bag tells how likely it is to be a key instance.

As shown in the Fig. 1, the attention-based deep MIL model transforms the instances in the bag XnX_{n} to instance embeddings, and then the model makes a bag embedding by aggregating the instance embeddings through attention-based pooling. Finally, the model calculates a bag probability PnP_{n} by putting the bag embedding to a bag-level classifier. For the MIL model, if we assume that Gn={gn,1,,gn,K}G_{n}=\{g_{n,1},\ldots,g_{n,K}\} is a set of the instance embeddings of the bag XnX_{n}, where gn,kg_{n,k} is the instance embedding of the kthk^{\text{th}} instance in the nthn^{\text{th}} bag and gn,kg_{n,k}\in I​R1×N, an attention score for the kthk^{\text{th}} instance in the nthn^{\text{th}} bag can be written as follows [1]:

an,k=exp{wtanh(Vgn,k)}j=1Kexp{wtanh(Vgn,j)}a_{n,k}=\frac{\exp\left\{\text{w}^{\top}\text{tanh}(\text{V}g_{n,k}^{\top})\right\}}{\sum_{j=1}^{K}\exp\left\{\text{w}^{\top}\text{tanh}(\text{V}g_{n,j}^{\top})\right\}} (1)

where an,ka_{n,k} is an attention score for the kthk^{\text{th}} instance in the nthn^{\text{th}} bag, w\text{w}\inI​RM×1 and V\text{V}\inI​RM×N are fully-connected layer’s parameters. The bag embedding for the nthn^{\text{th}} bag based on attention-based pooling can be written as follows [1]:

zn=j=1Kan,jgn,jz_{n}=\sum_{j=1}^{K}a_{n,j}g_{n,j} (2)

where znz_{n} is the bag embedding for the nthn^{\text{th}} bag. The bag embedding znz_{n} is put to a bag-level classifier H()H\left(\cdot\right), and then the bag probability is calculated by using the result of bag-level classifier H()H\left(\cdot\right). This process can be written as follows:

Pn=sigmoid(H(zn))P_{n}=\text{sigmoid}\left(H\left(z_{n}\right)\right) (3)

where PnP_{n} is the bag probability for the nthn^{\text{th}} bag. For the bag-level classifier H()H\left(\cdot\right), we use a single fully-connected layer, which can learn the interactions within the bag embedding. To make the bag-level prediction Y^n\widehat{Y}_{n}, we apply a threshold to PnP_{n}. In the case of KID, whenever Y^n=1\widehat{Y}_{n}=1, we apply Min-Max normalization to the attention score an,ka_{n,k} of kthk^{\text{th}} instance in the bag nthn^{\text{th}} bag. This process can be written as follows:

a^n,k=an,kmin(an)max(an)min(an)\widehat{a}_{n,k}=\frac{a_{n,k}-\min(a_{n})}{\max(a_{n})-\min(a_{n})} (4)

where a^n,k\widehat{a}_{n,k} is the normalized attention score of an,ka_{n,k} and ana_{n} is a set of attention scores of KK instances in the nthn^{\text{th}} bag, i.e., an={an,1,,an,K}a_{n}=\{a_{n,1},\ldots,a_{n,K}\}. Then, we make instance-level prediction Y^n,k\widehat{Y}_{n,k} by applying a threshold to a^n,k\widehat{a}_{n,k}. However, since the attention-based deep MIL model has limits in KID performance, we perform KID by recomputing the instance-level prediction Y^n,k\widehat{Y}_{n,k} of the optimized bag X^n\widehat{X}_{n} after optimizing the bag XnX_{n} to fit the criterion of the model by applying sparse network inversion.

III-C Neural Network Inversion

Neural network inversion is a method that optimizes the input to fit the criterion of the neural network model. To apply neural network inversion to an attention-based deep MIL model, we initialize the input as the values of the input image. Since we use negative log-likelihood as objective function when training the attention-based deep MIL model, we optimize the bag XnX_{n} by fitting the same criterion of the attention-based deep MIL model. That is, whenever the bag-level prediction Y^n\widehat{Y}_{n} for the bag XnX_{n} is positive, we apply neural network inversion to the bag XnX_{n}. Then, the parts of the optimized bag Xn^\widehat{X_{n}} that make the bag positive are strengthened, and the other parts are weakened. In other words, if the optimized bag Xn^\widehat{X_{n}} is put to the trained MIL model again, the attention scores corresponding to the key instances become larger. Formally, the objective function of neural network inversion can be written as follows:

l(Xn)=Y^nlogPn(1Y^n)log(1Pn)l(X_{n})=-\widehat{Y}_{n}\log P_{n}-\left(1-\widehat{Y}_{n}\right)\log(1-P_{n})\ (5)

where Y^n\widehat{Y}_{n} is bag-level prediction for the bag XnX_{n}. The reason we use Y^n\widehat{Y}_{n}, not YnY_{n}, is that we cannot access the true bag-level label YnY_{n} during test. To update the bag XnX_{n} by optimizing the objective function of neural network inversion, we use Stochastic Gradient Descent (SGD) optimization algorithm [16]. Each time the instances of the bag XnX_{n} is updated, we set the range of the values of instances from 0 to 11 as each instance is image data with a value between 0 and 11.

III-D Sparse Network Inversion

Sparse network inversion is a kind of neural network inversion that uses a sparseness constraint. That is, sparse network inversion updates the bag XnX_{n} by optimizing the objective function of neural network inversion with a sparseness constraint. We incorporate a sparseness constraint into neural network inversion to give regularization effect when we update the bag XnX_{n} and to further weaken the parts of the bag XnX_{n} that do not affect the criterion of the trained MIL model. Thus, if we put the optimized bag X^n\widehat{X}_{n} to the attention-based deep MIL model, we can detect more key instances than when using only attention-based deep MIL model or applying neural network inversion. The objective function of sparse network inversion can be written as follows:

L(Xn)=l(Xn)+λXn1L(X_{n})=l(X_{n})+\lambda\left\lVert X_{n}\right\rVert_{\text{1}} (6)

where l(Xn)l(X_{n}) is the objective function of neural network inversion and λ\lambda is a sparseness coefficient. Since the objective function of sparse network inversion involves a sparseness constraint for bag XnX_{n}, in order to optimize the bag XnX_{n}, we use proximal gradient method, which is an extension of typical gradient algorithm [17]. The method is used due to its simplicity and adequateness for solving data with high dimension [17, 18]. In the proximal gradient method, soft-thresholding function is used as proximal operator for a sparseness constraint. The soft-thresholding function can be written as follows [17]:

sλ(x)={xλif x>λ0if |x|λx+λif x<λs_{\lambda}(x)=\begin{cases}x-\lambda&\mbox{if }x>\lambda\\ 0&\mbox{if }\left|x\right|\leq\lambda\\ x+\lambda&\mbox{if }x<-\lambda\end{cases} (7)

where λ\lambda is the sparseness coefficient. We can write the soft-thresholding function compactly as follows:

sλ(x)=max(|x|λ,0) sign(x)s_{\lambda}(x)=\max(\left|x\right|-\lambda,0)\mbox{ sign($x$)} (8)

where sign(xx) means the sign of xx. Formally, when we use proximal gradient method for the bag XnX_{n}, update equation of sparse network inversion can be written as follows:

Xnt+1=sλ(Xntηl(Xnt))X_{n}^{t+1}=s_{\lambda}(X_{n}^{t}-\eta\nabla l(X_{n}^{t})) (9)

where sλ()s_{\lambda}(\cdot) is the soft-thresholding function, η\eta is the learning rate, and l()l(\cdot) is the objective function of neural network inversion. Since instances of the bag XnX_{n} is image data with a value between 0 and 11, each time each instance of the bag XnX_{n} is updated, we set the range of the instances in the bag from 0 to 11.

IV Experiments

In this section, we introduce an experimental setting, datasets, implementation details, and performance results and analyses of our proposed method. Through the section, we aim to show that our method improves the performance in terms of KID with the superiority of the bag-level classification performance of the attention-based deep MIL model [1]. We evaluated our method on various MIL datasets including an MNIST-based image MIL dataset, and two real-world histopathology datasets (COLON CANCER [19], BREAST CANCER [20]). Since our method is the same as the attention-based deep MIL model for the bag-level classification, to show the superiority of our proposed method in bag-level classification performance, we compare the attention-based deep MIL model with the MIL models based on instance-space paradigm. Since KID detects the positive instances that triggered a bag-level label in a positive bag, KID performance becomes meaningful only if the MIL model can classify the bags properly. Thus, we compare our proposed method with the attention-based deep MIL model and the MIL model that uses max pooling on instance-space paradigm, which showed best bag-level classification performance among the MIL models based on instance-space paradigm in our experiments.

TABLE I: Percentage of bags on various MIL datasets. Positive (Negative) presents the percentage of positive (negative) bags in each dataset.
Dataset Positive Negative
MNIST-based image MIL 50% 50%
COLON CANCER 51% 49%
BREAST CANCER 44.8% 55.2%
TABLE II: Percentage of instances in positive bags on various MIL datasets. Positive (Negative) presents the percentage of positive (negative) instances in each dataset.
Dataset Positive Negative
MNIST-based image MIL 10.2% 89.8%
COLON CANCER 53.5% 46.5%
BREAST CANCER 8.5% 91.5%

IV-A Experimental Setting

For all experiments, we use the same model architecture and optimizer that showed the high classification performance for each dataset as used in the precedent work [1]. To construct the attention-based deep MIL model, we apply attention-based pooling before the last layer of the deep neural network model. We call the attention-based deep MIL model Att, the method that applies neural network inversion to the attention-based deep MIL model Att+inv, and the method that applies sparse network inversion to the attention-based deep MIL model Att+sparse. In the MIL models based on instance-space paradigm, the models compute instance scores and aggregate the scores by max pooling or mean pooling. To construct the model, we applied max pooling or mean pooling after the last layer of the deep neural network model. We call these MIL models Inst+max, Inst+mean.

To do a fair evaluation, we use 10-fold cross-validation, and each experiment was performed five times independently. In the case of the MNIST-based image MIL dataset, we divided the MNIST dataset into an MNIST train dataset and an MNIST test dataset. Then, we created the MNIST-based image MIL train dataset and the MNIST-based image MIL test dataset by doing sampling from an MNIST train dataset and an MNIST test dataset independently.

To compare the bag-level classification performance, we use accuracy as a metric because the number of positive and negative bags on various MIL datasets is balanced as shown in Table 1. In the case of KID performance, since the number of positive and negative instances in positive bags on various MIL datasets are imbalanced as shown in Table 2, we use F1 measure as used in the precedent work [6, 21], as F1 measure is known to be insensitive to imbalance of the labels in data [22].

IV-B Datasets

IV-B1 MNIST-based image MIL dataset

To intuitively show that our method improves the KID performance while maintaining the bag-level classification performance of the attention-based deep MIL model, we created an MNIST-based image MIL dataset using the MNIST dataset. Each image of the MNIST-based image MIL dataset is shown in Fig. 3-(a). To make each data of the MNIST-based image MIL dataset that looks like an image as shown in Fig. 3-(a), we randomly extracted 400 images from the MNIST dataset and arranged them by 20×2020\times 20. That is, each data in the MIL dataset consists of 400 MNIST images as shown in Fig. 3-(a), where each MNIST image in a white box is an instance in a bag, and a set of 400 MNIST images is the bag. The label of each bag is assigned as a positive label if one image or more images of 400 MNIST images correspond to number 9, otherwise a negative label. The reason we chose the number 9 as the basis for making the bag label is that using the number 9 makes the problem difficult as the number 9 is easily confused with other numbers [1].

IV-B2 Histopathology datasets

To verify the superiority of our method, we conducted an experiment with weakly labeled real-world histopathology datasets including colon cancer dataset (COLON CANCER) [19] and breast cancer dataset (BREAST CANCER) [20].

COLON CANCER consists of 100 H&E stained histology images. Each image consists of 500×\times500 pixels. In the dataset, there are 29,756 nuclei that were marked. Among these nuclei, 22,444 nuclei have an associated class label, i.e., epithelial, inflammatory, fibroblast, and miscellaneous. To use this dataset in MIL, as shown in Fig. 4-(a), we created the bag with 27×\times27 patches as used in the precedent work [1]. Each of the bags is labeled as positive if it contains one or more nuclei associated with the epithelial class, otherwise negative [1, 19]. The reason we choose epithelial cells as the key instances that trigger bag-level label is that colon cancer is closely related to epithelial cells [23].

BREAST CANCER consists 58 H&E images. Each image consists of 896×\times768 pixels. To use this dataset in MIL, we made the bags with 32×\times32 patches as used in precedent work [1]. When we make bags, if each patch contains 75%\% or more white pixels, we discarded the patch. Each of the bags is labeled as positive if it contains one or more cancer cells, otherwise negative [1, 20].

IV-C Implementation details

To experiment with these datasets, we used an existing model architectures: Lenet 5 model [24] for MNIST-based image MIL dataset; the proposed model [19] for two real-world histopathology datasets. When constructing the models, ReLU is used as activation function in all layers except for the two layers that is used for attention-based pooling, where tanh is used as the activation function [1]. To optimize the MIL models, Adam optimization algorithm is used [25]. For the MIL models on MNIST-based image MIL dataset, beta values of 0.9 and 0.999, learning rate of 0.0005, and weight decay of 0.0001 are used [1]. In the two histopathology datasets, beta values of 0.9 and 0.999, learning rate of 0.0001, weight decay of 0.0005 are used [1]. When optimizing the models, we used 100 epochs and we stopped optimizing models based on the lowest validation error. Also, in the case of two histopathology datasets, to prevent overfitting due to the small amount of data, we performed data augmentation by randomly rotating and flipping the data. In order to improve KID performance of the attention-based deep MIL model, when we optimize data by applying neural network inversion, we updated each data 1,000 times by using SGD optimization algorithm [16] where learning rate is 0.001, momentum coefficient is 0.9. When we applied sparse network inversion, we used proximal gradient method. For MNIST-based image MIL dataset, we used 0.0001 as learning rate and we repeated update 200 times with varying the sparseness constraint coefficient from 0.0005 to 0.003. In COLON CANCER, the learning is 0.00001 and each data is updated 200 times with varying the sparseness constraint coefficient from 0.0004 to 0.0024. In the case of BREAST CANCER, 0.0001 is used as learning rate and each data is updated 100 times with varying the sparseness constraint coefficient from 0.003 to 0.006. For KID, we set the threshold to the optimal value among [0.1, 0.2, 0.3, 0.4, 0.5] for each model and dataset.

TABLE III: Bag-level classification accuracy on various MIL datasets. All results are averages of 5 times running and (±\pm) is a standard error of the mean.
Method MNIST-based image MIL COLON CANCER BREAST CANCER
Inst+max 0.804 ±\pm 0.232 0.868 ±\pm 0.025 0.536 ±\pm 0.062
Inst+mean 0.708 ±\pm 0.095 0.798 ±\pm 0.023 0.612 ±\pm 0.038
Attention 0.996 ±\pm 0.008 0.909 ±\pm 0.02 0.718 ±\pm 0.054
Refer to caption
(a) F1 measure for an MNIST-based image MIL dataset.
Refer to caption
(b) F1 measure for COLON CANCER.
Refer to caption
(c) F1 measure for BREAST CANCER.
Figure 2: F1 measure for MIL datasets.

IV-D Results and analyses

Refer to caption
(a) All patches of an image.
Refer to caption
(b) All refined patches.
Refer to caption
(c) Heatmap of an attention-based deep MIL model.
Refer to caption
(d) Heatmap of a sparse network inversion.
Refer to caption
(e) Visualization result of an attention-based deep MIL model.
Refer to caption
(f) Visualization result of sparse network inversion.
Refer to caption
(g) Ground truth.
Figure 3: Visualization results of an image in MNIST-based image MIL dataset.
Refer to caption
(a) All patches of an image.
Refer to caption
(b) All refined patches.
Refer to caption
(c) Heatmap of an attention-based deep MIL model.
Refer to caption
(d) Heatmap of a sparse network inversion.
Refer to caption
(e) Visualization result of an attention-based deep MIL model.
Refer to caption
(f) Visualization result of sparse network inversion.
Refer to caption
(g) Ground truth.
Figure 4: Visualization results of an image in COLON CANCER.

The bag-level classification results and KID results on all datasets are in Table 3 and Fig. 2, respectively. For bag-level classification performance, the attention-based deep MIL model outperforms the other MIL models which are based on instance-space paradigm for all datasets. On the other hand, in the KID performance, the attention-based deep MIL model shows a worse performance than the other model based on instance-space paradigm for MNIST-based image MIL dataset and COLON CANCER. In the case of BREAST CANCER, although the attention-based deep MIL model has better performance than the other model based on instance-space paradigm, the absolute performance is not optimal. Thus, despite the superior bag-level classification performance of the attention-based deep MIL model, there is a problem in using an attention-based deep MIL model for KID. However, if we apply neural network inversion or sparse network inversion to data in the attention-based deep MIL model, since these methods relieve the problem that attention-based deep MIL model focuses only on few distinguishable key instances by removing the constraint that data cannot be changed, the KID performance is improved. Especially, in the case of sparse network inversion, the KID performance is significantly improved as the effects of regularization and sparseness have a positive effect on the result.

To intuitively show that our method can provide better interpretable results than the attention-based deep MIL model, we provide a visualization of the results in Fig. 3 and Fig. 4, where Fig. 3 and Fig. 4 are related to MNIST-based image MIL dataset and COLON CANCER, respectively. Each figure consists of seven images: (a) all patches of an image in the each dataset; (b) all refined patches that are the result of sparse network inversion; (c) heatmap of the attention-based deep MIL model: every patch from (a) multiplied by its corresponding Min-Max normalized attention score of every patch from (a); (d) heatmap of the sparse network inversion: every patch from (a) multiplied by its corresponding Min-Max normalized attention score of every patch from (b); (e) visualization result of the attention-based deep MIL model: every patch from (a) multiplied by its corresponding value that applies threshold to Min-Max normalized attention score of every patch from (a); (f) visualization result of sparse network inversion: every patch from (a) multiplied by its corresponding value that applies threshold to Min-Max normalized attention score of every patch from (b); (g) ground truth of the patches in instance-level. As shown in the Fig. 3-(c) and Fig. 4-(c), Min-Max normalized attention scores of the attention-based deep MIL model are skewed to few distinguishable key instances. However, if we apply our method to the patches, Min-Max normalized attention scores are scattered around more key instances than before applying sparse network inversion to the patches as shown in the Fig. 3-(d) and Fig. 4-(d). By applying threshold to the normalized attention scores of the attention-based deep MIL model and sparse network inversion, we can perform KID as shown in the Fig. 3-(e), Fig. 3-(f), Fig. 4-(e), and Fig. 4-(f). From these results, although we use only bag-level data and bag-level label during training, we can confirm that the attention-based deep MIL model finds key instances in a positive bag. However, if we just use the attention-based deep MIL model as shown in Fig. 3-(e) and Fig. 4-(e), the model finds only few key instances. On the other hand, as shown in Fig. 3-(f) and Fig. 4-(f), our method allows the model to find more key instances than before applying our method.

From these experiments on the MNIST-based image MIL dataset and two real-world histopathology datasets, we can see that our method significantly improves the KID performance while the bag-level classification performance is maintained. This means that applying sparse network inversion to the attention-based deep MIL model helps improve the KID performance of the attention-based deep MIL model.

V Conclusion

In this paper, we propose a method to maintain the performance of bag-level classification and to improve KID performance of an attention-based deep MIL model. Our method applies sparse network inversion to the MIL model. In experiments, we measured the accuracy of bag-level classification and F1 measure of instance-level classification for our proposed method for an MNIST-based image MIL dataset and two histopathology datasets, and our proposed method significantly improved the KID performance of the attention-based deep MIL model. The bag-level decision of the MIL model can be interpreted with the key instances. Therefore, our research will be useful in area such as medicine where the interpretation of the model’s behavior can be important.

Acknowledgment

This work was supported by Institute for Information & communications Technology Promotion(IITP) grant funded by the Korea government(MSIT) (IITP-2018-0-00584), and Samsung Electronics Co., Ltd.

References

  • [1] M. Ilse, J. M. Tomczak, and M. Welling, “Attention-based deep multiple instance learning,” in ICML, 2018.
  • [2] T. G. Dietterich, R. H. Lathrop, and T. Lozano-Perez, “Solving the multiple instance problem with axis-parallel rectangles,” Artificial Intelligence, vol. 89, no. 1-2, pp. 31–71, 1997. [Online]. Available: http://lis.csail.mit.edu/pubs/tlp/multiple-instance-aij.pdf
  • [3] O. Maron and T. Lozano-Perez, “A framework for multiple instance learning,” in Advances in Neural Information Processing Systems 10.   Cambridge, MA: MIT Press, 1998. [Online]. Available: http://lis.csail.mit.edu/pubs/tlp/maron98framework.pdf
  • [4] G. Quellec, G. Cazuguel, B. Cochener, and M. Lamard, “Multiple-instance learning for medical image and video analysis,” IEEE Reviews in Biomedical Engineering, vol. PP, 01 2017.
  • [5] Z.-H. Zhou, Y.-Y. Sun, and Y.-F. Li, “Multi-instance learning by treating instances as non-i.i.d. samples,” Proceedings of the 26th International Conference On Machine Learning, ICML 2009, 07 2008.
  • [6] G. Liu, J. Wu, and Z.-H. Zhou, “Key instance detection in multi-instance learning,” in ACML, 2012.
  • [7] J. Amores, “Multiple instance classification: Review, taxonomy and comparative study,” Artificial Intelligence, vol. 201, p. 81–105, 08 2013.
  • [8] S. Andrews, I. Tsochantaridis, and T. Hofmann, “Support vector machines for multiple-instance learning,” Advances in Neural Information Processing Systems, vol. 15, pp. 561–568, 01 2002.
  • [9] X. Wang, Y. Yan, P. Tang, X. Bai, and W. Liu, “Revisiting multiple instance neural networks,” Pattern Recognition, vol. 74, pp. 15 – 24, 2018. [Online]. Available: http://www.sciencedirect.com/science/article/pii/S0031320317303382
  • [10] J. J. B. Ramón and L. D. Raedt, “Multi instance neural networks,” in ICML 2000, 2000.
  • [11] W. Wang, Y. Ning, H. Rangwala, and N. Ramakrishnan, “A multiple instance learning framework for identifying key sentences and detecting events,” in CIKM, 2016.
  • [12] Y. Yan, X. Wang, X. Guo, J. Fang, W. Liu, and J. Huang, “Deep multi-instance learning with dynamic pooling,” in ACML, 2018.
  • [13] J. Kindermann and A. Linden, “Inversion of neural networks by gradient descent,” Parallel Computing, vol. 14, pp. 277–286, 1990.
  • [14] D. A. Hoskins, J. N. Hwang, and J. Vagners, “Iterative inversion of neural networks and its application to adaptive control,” Trans. Neur. Netw., vol. 3, no. 2, pp. 292–301, Mar. 1992. [Online]. Available: https://doi.org/10.1109/72.125870
  • [15] S. Jang, Y.-E. Jang, Y.-J. Kim, and H. Yu, “Input initialization for inversion of neural networks using k-nearest neighbor approach,” Information Sciences, vol. 519, pp. 229 – 242, 2020. [Online]. Available: http://www.sciencedirect.com/science/article/pii/S0020025520300426
  • [16] S. Ruder, “An overview of gradient descent optimization algorithms,” arXiv preprint arXiv:1609.04747, 2016.
  • [17] A. Beck and M. Teboulle, “A fast iterative shrinkage-thresholding algorithm for linear inverse problems,” SIAM J. Imaging Sciences, vol. 2, pp. 183–202, 01 2009.
  • [18] K. Bredies and D. Lorenz, “Linear convergence of iterative soft-thresholding,” J. Fourier Anal. Appl., vol. 14, 09 2007.
  • [19] K. Sirinukunwattana, S. e. A. Raza, Y. Tsang, D. Snead, I. Cree, and N. Rajpoot, “Locality sensitive deep learning for detection and classification of nuclei in routine colon cancer histology images,” IEEE Transactions on Medical Imaging, vol. 35, pp. 1–1, 02 2016.
  • [20] E. D. Gelasca, J. Byun, B. Obara, and B. S. Manjunath, “Evaluation and benchmark for biological image segmentation,” 2008 15th IEEE International Conference on Image Processing, pp. 1816–1819, 2008.
  • [21] M. Kandemir and F. Hamprecht, “Instance label prediction by dirichlet process multiple instance learning,” Uncertainty in Artificial Intelligence - Proceedings of the 30th Conference, UAI 2014, pp. 380–389, 01 2014.
  • [22] M. Bekkar, H. Djema, and T. Alitouche, “Evaluation measures for models assessment over imbalanced data sets,” Journal of Information Engineering and Applications, vol. 3, pp. 27–38, 01 2013.
  • [23] L. Ricci-Vitiani, D. Lombardi, and E. Pilozzi, “Identification and expansion of human colon-cancer-initiating cells,” Nature, vol. 445, pp. 111–115, 01 2007.
  • [24] Y. Lecun, L. Bottou, Y. Bengio, and P. Haffner, “Gradient-based learning applied to document recognition,” Proceedings of the IEEE, vol. 86, pp. 2278 – 2324, 12 1998.
  • [25] D. Kingma and J. Ba, “Adam: A method for stochastic optimization,” International Conference on Learning Representations, 12 2014.