∎
471, Nagaodai-cho, Sakae-ku, Yokohama-city, Kanagawa, 244-8533, Japan.
Tel.: +81-45-853-8584
Fax: +81-45-853-8585
22email: [email protected]
1 Nikon Corporation, Research & Development Division
2 RIKEN Center for Advanced Intelligence Project
3 The University of Tokyo, Graduate School of Frontier Sciences
Binary classification with ambiguous training data
Abstract
In supervised learning, we often face with ambiguous (A) samples that are difficult to label even by domain experts. In this paper, we consider a binary classification problem in the presence of such A samples. This problem is substantially different from semi-supervised learning since unlabeled samples are not necessarily difficult samples. Also, it is different from 3-class classification with the positive (P), negative (N), and A classes since we do not want to classify test samples into the A class. Our proposed method extends binary classification with reject option, which trains a classifier and a rejector simultaneously using P and N samples based on the 0-1- loss with rejection cost . More specifically, we propose to train a classifier and a rejector under the 0-1-- loss using P, N, and A samples, where is the misclassification penalty for ambiguous samples. In our practical implementation, we use a convex upper bound of the 0-1-- loss for computational tractability. Numerical experiments demonstrate that our method can successfully utilize the additional information brought by such A training data.
Keywords:
Ambiguous samples Classification with reject option Binary classification1 Introduction
Supervised learning has been successfully deployed in various real-world applications, such as medical diagnosis (Bar et al., 2015; Wang et al., 2016; Esteva et al., 2017) and manufacturing systems (Park et al., 2016; Ren et al., 2017). However, when the amount of labeled data is limited, current supervised learning methods still do not work reliably (Pesapane et al., 2018).
To efficiently obtain labeled data, domain knowledge has been used in many application areas (Ren et al., 2017; Cruciani et al., 2018; Konishi et al., 2019; Bejnordi et al., 2017). However, as some studies have pointed out (Wagner et al., 2005; Li et al., 2016; Shahriyar et al., 2018), there are often ambiguous samples that are substantially difficult to label even by domain experts.
The goal of this paper is to propose a novel classification method that can handle such ambiguous data. More specifically, we consider a binary classification problem where, in addition to positive (P) and negative (N) samples, ambiguous (A) samples are available for training a classifier.
Naively, we may consider employing 3-class classification methods for the P, N, and A classes. However, since we classify test samples only in the P or N class, not in the A class, naive 3-class methods cannot be directly used in our problem. Moreover, they cannot utilize the information that the A class exists between the P and N classes. Another related approach is classification with reject option (Bartlett and Wegkamp, 2008; Cortes et al., 2016), where ambiguous test samples are not classified into the P or N classes, but rejected. However, classification methods with reject option do not consider ambiguous samples in the training phase and thus they cannot be employed in the current scenario.
Semi-supervised learning may also be related to the current problem, where unlabeled data is used for training a classifier in addition to P and N data (Odena, 2016; Sakai et al., 2017). In semi-supervised learning, unlabeled samples are P and N samples that have not yet been labeled and they are not necessarily difficult samples to be labeled. On the other hand, in our target problem of classification with ambiguous data, ambiguous data are typically distributed in the intersection of the P and N classes. Thus, since the problem setups are intrinsically different, merely using semi-supervised learning methods in the current problem may not be promising.
Classification with imperfect labeling (Cannings et al., 2020) allows incorrect labels in training data, so it can be useful to deal with a dataset where annotators forcibly give positive or negative labels to all samples. Open-set classification (Scheirer et al., 2012) detects samples classified into none of the training classes and classifies them into the “unknown” class, so we can apply it when we deal with a dataset where annotators skip labeling hard-to-label samples and we classify test samples into the P, N or A class. However, those two approaches are still different from our problem setting in terms of labels in input and output data. Table 1 summerizes the problem setting of different approaches.
|
|
|
|
|||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
|
|
|
|
|||||||||||
|
|
|
|
|||||||||||
|
|
|
|
|||||||||||
|
|
|
|
|||||||||||
|
|
|
|
|||||||||||
|
|
|
|
|||||||||||
|
|
|
|
To effectively solve the problem of classification with ambiguous data, we propose to extend classification with reject option that trains a classifier and a rejector simultaneously using P and N samples based on the 0-1- loss with rejection cost (Cortes et al., 2016). Our proposed method trains a classifier and a rejector under the 0-1-- loss using P, N, and A samples, where is the misclassification penalty for ambiguous samples. Then, in the test phase, we use the trained classifier to assign P or N labels to test samples. However, in the same way as the 0-1- loss, directly performing optimization with the 0-1-- loss is cumbersome due to its discrete nature. To cope with this problem, we introduce a convex upper bound of the 0-1-- loss and use it in our practical implementation. Through experiments, we demonstrate that the proposed method can improve the test classification accuracy by utilizing A samples in the training phase. We also consider a simple heuristic that we randomly relabel ambiguous samples into the positive or negative class and apply classification with reject option. We show that the heuristic is essentially equivalent to a special case of the proposed method, and thus it can be an easy-to-implement alternative to the proposed method.
The rest of this paper is organized as follows. We briefly review supervised learning in Section 2. Then, we define our problem setting and describe the details of our proposed method in Section 3. In Section 4, we experimentally evaluate its performance. Finally, in Section 5, we summarize our contributions and describe future works.
2 Supervised classification
In this section, we first define the standard supervised classification problem and then review its standard solution.
2.1 Formulation
Let be an input point and denote a binary label, which corresponds to the positive and negative classes, respectively. Suppose that we are given a set of positive and negative samples drawn independently from the probability distribution with density defined on . Let denote a discriminant function, with which a class label is predicted for test input point as .
The goal is to train the discriminant function so that the expected misclassification rate is minimized. Let us define the 0-1 loss as
(1) |
where is the indicator function that takes 1 if statement is true and 0 otherwise. Then, we can express this problem as
where denotes the optimal discriminant function and denotes the expectation over . In practice, since we do not know the true density , we usually use the empirical distribution to approximate the expectation:
(2) |
2.2 Support vector machine (SVM)
Because optimization with is computationally intractable, it is not practical to optimize the empirical risk directly. To overcome this problem, the hinge loss, an upper bound of called the hinge loss, defined by
(3) |
has been introduced as its surrogate. Since the hinge loss is convex, optimization can be reduced to a convex program. Further introducing the L2 regularization, basis functions , and slack variables with ⊤ being the transpose, the following quadratic program can be obtained as a dual optimization problem:
(6) | |||||
where are the coefficients of the discriminant function, is the L2 regularization parameter, and is the value of the discriminant function at sample point given by . The resulting discriminant function is given by .
3 Classification with ambiguous data
In this section, we formulate our target problem called classification with ambiguous data (CAD) and propose a new method for solving the CAD.
3.1 Formulation
We consider three class labels, i.e,. positive, ambiguous, and negative: . Suppose that we are given a set of positive, ambiguous, and negative samples drawn independently from the probability distribution with density defined on . Our goal is still to learn a discriminant function that classifies test samples into either the positive or negative class (not in the ambiguous class). Our key question in this scenario is if we can utilize the ambiguous training data to improve the classification accuracy of the discriminant function.
In this section, we develop a new method based on a method of classification with reject option (CRO) (Cortes et al., 2016). For this reason, before deriving the new method, we first review the CRO method.
3.2 Classification with reject option by SVM (CRO-SVM)
Cortes et al. (2016) introduced a rejection function to identify the regions with high risk for misclassification, in addition to discriminating the positive and negative classes. When the rejection function takes a positive value, the corresponding sample is accepted and is classified into the positive or negative class by classifier ; otherwise, the sample is rejected and is not classified. When a sample is rejected, the rejection cost is incurred, which trades off the risk of misclassification and the cost of rejection. To realize this idea, the 0-1-c loss was introduced:
(7) |
When , all samples are rejected because the loss function does not incur any cost. On the other hand, when , no samples are rejected because the expectation of the 0-1 loss is less than ; in that case, the 0-1- loss is reduced to the 0-1 loss. Therefore, effectively, we only consider such that . The rejection function and the discriminant function are simultaneously learned from training data.
Similarly to the 0-1 loss, the 0-1- loss has discrete nature and thus its direct optimization is computationally intractable. To avoid the discontinuity, the following surrogate loss called the max-hinge (MH) loss was introduced:
(8) |
where are the hyperparameters to control the shape of the surrogate loss.
In the same manner as the original SVM, introducing the L2 regularization, basis functions, and slack variables yields the following quadratic program:
(12) | |||||
where are the coefficients of the discriminant function, are the coefficients of the rejection function, are the L2 regularization parameters, and and denote the values of the discriminant function and rejection function at sample point given by and , respectively. The resulting discriminant function and rejection function are given by and .
We refer to this method as CRO-SVM.
3.3 Proposed method: classification with ambiguous data by SVM (CAD-SVM)
To handle ambiguous training data in the SVM formulation, we extend the 0-1- loss to the 0-1-c-d loss defined as
(13) |
Table 3 and Table 3 compare the behavior of the 0-1- loss and the 0-1-- loss. For positive and negative samples, the 0-1-- loss behaves the same as the 0-1- loss. On the other hand, for ambiguous samples, the 0-1-- loss incurs penalty when they are classified into the positive or negative class. Therefore, ambiguous samples tend to be classified into the ambiguous class if we employ the 0-1-- loss. Compared to the CRO formulation, where a rejector cannot be learned explicitly from positive and negative samples, the CAD utilizes ambiguous samples to learn a rejector explicitly.
The above discussion may mislead us as if we are just solving a 3-class problem with the positive, ambiguous, and negative classes. However, we do not classify test samples into the ambiguous class, but only into the positive and negative classes. To solve the CAD problem, we utilize a binary discriminant function and a rejection function , as in the CRO formulation reviewed above. More specifically, we train and under the 0-1-- loss, and we only use in the test phase to classify test samples into the positive and negative classes. Thanks to the interplay between and in the 0-1-- loss, we can utilize ambiguous data to train through .
|
|
|
|||||||||
---|---|---|---|---|---|---|---|---|---|---|---|
Positive | 0 | 1 | |||||||||
Negative | 1 | 0 |
|
|
|
|||||||||
---|---|---|---|---|---|---|---|---|---|---|---|
Positive | 0 | 1 | |||||||||
Ambiguous | 0 | ||||||||||
Negative | 1 | 0 |
Similarly to the 0-1- loss, we consider the following convex upper bound of the 0-1-- loss called the max-hinge-ambiguous (MHA) loss as a surrogate to avoid its discrete nature:
(14) | |||||
where is the hyperparameter to control the shape of the surrogate loss. See Figure 1 for its visualization.

Then, in the same way as the CRO-SVM, we have the following quadratic program:
(18) | |||||
This is our proposed method called CAD-SVM. The computational complexity of the CAD-SVM depends on implementation of the quadratic program. It naively costs , but if we use a fixed number of basis functions, the complexity reduces to .
The MHA loss depends on the choice of hyperparameters . To find good hyperparameter values, let us analyze the 0-1-- loss first.
For each , let , , and , where . Then the following lemma shows how and are related to and for the optimal classifier and rejector (its proof is available in Appendix A):
Lemma 1
For each , let
(19) |
Then
(20) |
Figure 2 illustrates the above results. This shows that when (or ) is large (i.e., imbalanced classification), the rejector accepts the sample and the classifier classifies that sample into the positive (or negative) class. On the other hand, when both and are not so large, the rejector rejects the sample.

Next, based on the above lemma, we have the following theorem for the MHA loss (its proof is given in Appendix B):
Theorem 3.1
For each , let
(21) |
Then, for
(22) |
the signs of match those of .
Based on the above theorem, we use Eq. (22) as hyperparameter values in our experiments in the next section and demonstrate that they work well in practice. Nevertheless, given that the above theorem is valid only for the optimal solutions, we may cross-validate better hyperparameter values around Eq. (22) to further improve the classification performance. Note that Eq. (22) does not include .
4 Numerical experiments
In this section, we report experimental results.
4.1 Datasets
For experiments, we use a toy dataset, a public dataset, and an in-house dataset.
4.1.1 Toy dataset
To understand the behavior of our method and related methods, we created a toy classification problem and applied the methods to it. The problem contains three regions: the positive, negative, and mixed regions (see Figure 3). The positive and negative regions are clearly separable, whereas the mixed region has no good discriminant function. For this problem, we want to clearly discriminate the positive and negative regions, with the influence of the mixed region avoided. We also studied how the proportion of ambiguous samples influences the results by changing the proportion of ambiguous samples in the mixed region . The number of all samples is 400, the total number of positive and negative samples in the separable regions is , and the total number of positive and negative samples in the mixed region is . Therefore, the expected maximum accuracy is .

4.1.2 Public datasets (PD1, PD2, PD3)
We processed a regression dataset, the Boston Housing Dataset (Harrison and Rubinfeld, 1978), to convert it to a classification dataset with ambiguous data. The original dataset consists of 13 features associated with the average house prices for 506 districts, where denotes the sample number. We annotated all the samples to be positive, negative, and ambiguous according to the following procedure:
- PD1 (P/N/A separable):
-
Simply, the samples with were labeled as positive, the samples with were labeled as negative, and the other samples were labeled as ambiguous. The numbers of samples were 190, 173, and 143 for the positive, negative, and ambiguous classes, respectively.
- PD2 (P/N/A mixed):
-
The samples with and the samples with were still labeled as positive and negative, respectively, but the remaining 143 samples were randomly labeled as positive, negative, or ambiguous.
- PD3 (separable and mixed):
-
We considered a hyperplane in the feature space and divided all the samples into two parts, the mixed part and the separable part . The coefficients of the hyperplane are selected so that the averages of over the samples in both parts were approximately matched. For the mixed part, each sample is randomly labeled as positive, negative or ambiguous, whereas for the separable part, the samples with were labeled as positive and the others were labeled as negative. The numbers of samples in the separable region were 83, 0, and 87, and those in the mixed region were 114, 107, and 115 for the positive, negative, and ambiguous classes, respectively.
4.1.3 In-house dataset from a cell culture process (ID)
As a real-world application, we prepared an in-house cell culture dataset. This dataset contained 124 fields of view (FOV). For each FOV, 2 images were acquired: one was in the middle, and the other was at the end of the culturing process. Each middle image was analyzed by the image processing software, CL-Quant (Alworth et al., 2010), and converted to 8 morphological features such as the average brightness and average area of cells.
Each final image was annotated by experts. If the cells in the image overall looked healthy/damaged, the image was labeled as positive/negative. However, some images contained both healthy and damaged cells, and they were labeled as ambiguous. The numbers of samples were 41, 59, and 24 for positive, negative, and ambiguous, respectively.
Our motivation was to predict the final state of each FOV annotated by the experts, using morphological features obtained in the middle of the culturing process. If we could predict it accurately, the culturing cost would be saved by aborting the culturing process where the cells would be damaged while keeping the healthy cells cultured. Therefore, we should focus on reducing misclassification between the positive and negative classes and we hope that information from ambiguous samples could be useful to improve the prediction accuracy. That was why test samples did not have the ambiguous label and were not classified into the ambiguous class in our scenario. Though ambiguous samples may occur in the test phase in the actual curturing process, we did not care which class they were classified into.
In the same manner as the PD1, PD2, and PD3 datasets, this dataset was also visualized by the PCA and the LPP in Figure 4 (d) and (h), respectively. They show that ambiguous points are roughly located between positive and negative points.

4.2 Experimental settings
Using the above datasets, we compared the classification performance of the SVM, the SVM-RL (random label), the LapSVM (Belkin et al., 2006), the two-step SVM, the CRO-SVM, the CRO-SVM-RL, and the CAD-SVM. The SVM-RL employs the SVM algorithm, but ambiguous samples are randomly relabeled as positive or negative, which effectively utilizes information of the ambiguous samples. The LapSVM is a semi-supervised learning method based on the SVM, which employs a regularization term defined by the graph Laplacian. Ambiguous samples are treated as unlabeled samples in the LapSVM. The two-step SVM is the method which learns the rejection function and the discriminant function sequentially—first the rejection function is learned to judges whether the sample is ambiguous or not, and then the discriminant function is learned only using the samples which are not rejected by the rejection function. When the rejection function is learned, class weights and are applied to non-ambiguous (i.e., positive and negative) and ambiguous samples, respectively. The CRO-SVM-RL is the CRO-SVM with random labels for ambiguous samples in the same manner as SVM-RL.
For each method, 500 test runs were performed by changing the training and test datasets which were randomly divided from the original dataset. The dividing ratio of the training and test dataset was 4:1 for the ID dataset and 1:2 for the other datasets. For each test run, 5-fold cross validation was performed to determine the parameters below. For validation and in the test phase, only positive and negative samples were applied to the discriminant function, thus we were able to evaluate the binary classification accuracy.
Note that our goal was to maximize the binary classification accuracy in the test phase, and that was equivalent to minimizing the expected 0-1 loss. Though we trained the models using various loss functions such as the hinge loss, the MH loss, and the MHA loss, the 0-1 loss was minimized by cross validation.
In total, we had 10 hyperparameters , where were the L2 regularization parameters, was the width of the Gaussian radial basis function in the basis functions , was the hyperparameter of the weight matrix of the graph Laplacian as (only in the LapSVM), was the coefficient of the graph Laplacian regularization (only in the LapSVM), were the hyperparameters of the 0-1- loss (in the CRO-SVM and CRO-SVM-RL), the 0-1-- loss (in the CAD-SVM) or the class weights (in the two-step SVM), and were the hyperparameters of the MH and MHA loss functions (in the CRO-SVM, CRO-SVM-RL and CAD-SVM). We used 5-fold cross validation in terms of the classification accuracy to choose the hyperparameters from , , , , and . The other hyperparameters were determined by Eq. (22). Quadratic programming problems were solved using cvxopt (Andersen et al., 2018).
4.3 Results
Figure 5 shows an example of the discriminant results for the toy dataset with . The upper half of the domain was the mixed region of positive and negative samples and samples in that region were expected to be classified into the ambiguous class. On the other hand, the lower half of the domain was perfectly separable into the positive and negative regions and samples in those regions were expected to be discriminated accurately (see Figure 3). The SVM made some misclassifications in the lower half of the domain, which were caused by the mixed region. The SVM-RL and LapSVM could not reduce the number of misclassifications, though they utilized information of ambiguous samples. This suggests that ambiguous samples should not be simply relabeled to positive or negative samples or should not be treated as unlabeled samples. The two-step SVM could not also reduce the number of misclassifications and it could not learn the ambiguous region. Thus, it would be disadvantageous to learn the ambiguous region by combining positive and negative classes. The CRO-SVM and CRO-SVM-RL successfully learned the ambiguous region but that did not lead to learning more accurate discriminant functions. The CAD-SVM also successfully learned the ambiguous region and then it was able to learn a more accurate discriminant function. Overall, the CAD-SVM was shown to be the most appropriate method in this toy experiment.

Tables 4 shows the test accuracy of the toy dataset by changing the parameters which are the propotion of ambiguous samples in the mixed region. When the proportion of the ambiguous samples was small, the proposed method did not work effectively since the number of ambiguous samples was too small. On the other hand, when the proportion of the ambiguous samples was large, it also did not show better performance. This was because, under this condition, only small numbers of positive and negative samples existed in the mixed region. Therefore, methods which do not utilize ambiguous samples showed reasonably good enough performance. However, when the proportion of ambiguous samples was , the CAD-SVM achieved a better score than the other methods.
0.1 | 0.3 | 0.5 | 0.7 | 0.9 | |
---|---|---|---|---|---|
SVM | |||||
SVM-RL | |||||
LapSVM | |||||
Two-step SVM | |||||
CRO-SVM | |||||
CRO-SVM-RL | |||||
CAD-SVM |
Table 5 summarizes the test accuracy of each method for the PD1, PD2, PD3, and ID datasets, respectively. For the PD1 dataset, the CAD-SVM was not superior to other methods since this condition was similar to the toy dataset with higher . For the PD2 dataset, though the dataset had a mixed region, the CAD-SVM also did not show good performance. However, for the PD3 dataset, which had a mixed region and a separable region, the CAD-SVM achieved the best performance among the compared methods. It is suggested that our proposed method, the CAD-SVM, works effectively when the dataset has both of a mixed region and a separable boundary between the positive and negative classes. The CAD-SVM utilizes ambiguous samples to learn a rejector which rejects mixed regions, thus it would be able to focus on learning a separable region. The CRO-SVM and the CRO-SVM-RL also learn a rejector from positive and negative samples, but the CAD-SVM would be superior since the CAD-SVM can explicitly learn a rejector using ambiguous samples. For the ID dataset, the CAD-SVM performed better than the other method, except for the CRO-SVM-RL. Overall, ambiguous samples can improve the binary classification accuracy under some conditions, and the CAD-SVM is one of the solutions that can utilize such ambiguous samples.
PD1 | PD2 | PD3 | ID | |
---|---|---|---|---|
SVM | ||||
SVM-RL | ||||
LapSVM | ||||
Two-step SVM | ||||
CRO-SVM | ||||
CRO-SVM-RL | ||||
CAD-SVM |
4.4 Discussions
Through all the experiments, the CRO-SVM-RL gave as good performance as the CAD-SVM. As detailed in Appendix C, we can show the following relations between the CRO-SVM-RL and the CAD-SVM:
-
1.
The 0-1- loss function for the randomly labeled (RL) dataset, in which we randomly relabeled ambiguous samples as positive or negative, reduces to the 0-1-- loss with .
-
2.
For the RL dataset, the MH loss can be regarded as a surrogate loss of the 0-1-- loss with , and it is calibrated under the conditions of Eq. (22).
Thus, though it is a simple heuristic, the CRO-SVM-RL is essentially equivalent to the CAD-SVM except for that the hyperparameter is fixed to . In practice, the CRO-SVM-RL is easier to implement and thus it may be used as an alternative to the CAD-SVM. However, we note that the CRO-SVM-RL alone does not provide rich theoretical insights that we have shown in Section 3.
Our goal for each experiment was minimizing the expected 0-1 loss in the test phase. Therefore, the SVM was naively an appropriate method since the hinge loss was calibrated to the 0-1 loss. Nevertheless, though the CRO-SVM-RL and the CAD-SVM minimized a surrogate of the 0-1-- loss in the training phase, they were able to achieve the better accuracy than the SVM in the test phase. It is suggested that the providing a reject option and incorporating ambiguous training samples could work as a kind of regularization, but further studies will be needed to clarify its mathematical properties.
We note that ambiguous samples are intrinsically hard-to-label samples even by experts, so they usually contain little information for classifying positive and negative samples, and thus we cannot expect large improvement to the binary classification accuracy. Nevertheless, our proposed method achieved statistically significant improvements for some cases.
5 Conclusion
In this study, we aimed to reduce the labeling cost and improve the classification accuracy by allowing labelers to give “ambiguous” labels for difficult samples. We extended a method of classification with reject option and proposed a novel classification method named the CAD-SVM that uses the 0-1-- loss. We derived a surrogate loss for the 0-1-- loss, which allowed us to convert the optimization problem into a convex quadratic program. We carried out numerical experiments and showed that ambiguous labels can be effectively used to improve the classification accuracy. We also showed that the CRO-SVM-RL, in which we randomly relabeled ambiguous samples to be positive or negative and applied classification with reject option, can be a practical alternative to the proposed method since it is essentially equivalent to the proposed method.
Though our proposed method was based on the SVM, it would be more useful if it can be applied to other models especially deep neural networks. However, further experimental studies will be needed to confirm if a naive application of the proposed MHA loss works well in practice. Indeed, it is known that changing models can cause other problems such as overfitting (Kiryo et al., 2017). Moreover, for deep neural networks, though we usually use the softmax cross entropy as the loss function, even the 0-1- loss function has not been extended to the softmax cross entropy. So, analyzing the influence of changing loss functions is also an important issue to be further investigated.
In addition to the experimental analysis with more complex models, our future study will conduct theoretical analysis of the proposed method such as statistical consistency and the rate of convergence. Extending the proposed loss function to semi-supervised problems, imperfect labeling problems or multi-class problems is also a promising direction to be pursued.
Acknowledgements.
The authors would like to thank Yasujiro Kiyota and Momotaro Ishikawa for providing the in-house dataset.References
- Alworth et al. (2010) Alworth SV, Watanabe H, Lee JS (2010) Teachable, high-content analytics for live-cell, phase contrast movies. Journal of biomolecular screening 15(8):968–977
- Andersen et al. (2018) Andersen MS, Dahl J, Vandenberghe L (2018) Cvxopt: A python package for convex optimization. URL http://cvxopt.org/
- Bar et al. (2015) Bar Y, Diamant I, Wolf L, Lieberman S, Konen E, Greenspan H (2015) Chest pathology detection using deep learning with non-medical training. In: 2015 IEEE 12th international symposium on biomedical imaging (ISBI), IEEE, pp 294–297
- Bartlett and Wegkamp (2008) Bartlett PL, Wegkamp MH (2008) Classification with a reject option using a hinge loss. Journal of Machine Learning Research 9(Aug):1823–1840
- Bejnordi et al. (2017) Bejnordi BE, Veta M, Van Diest PJ, Van Ginneken B, Karssemeijer N, Litjens G, Van Der Laak JA, Hermsen M, Manson QF, Balkenhol M, et al. (2017) Diagnostic assessment of deep learning algorithms for detection of lymph node metastases in women with breast cancer. Jama 318(22):2199–2210
- Belkin et al. (2006) Belkin M, Niyogi P, Sindhwani V (2006) Manifold regularization: A geometric framework for learning from labeled and unlabeled examples. Journal of machine learning research 7(Nov):2399–2434
- Cannings et al. (2020) Cannings TI, Fan Y, Samworth RJ (2020) Classification with imperfect training labels. Biometrika 107(2):311–330
- Cortes et al. (2016) Cortes C, DeSalvo G, Mohri M (2016) Learning with rejection. In: International Conference on Algorithmic Learning Theory, Springer, pp 67–82
- Cruciani et al. (2018) Cruciani F, Cleland I, Nugent C, McCullagh P, Synnes K, Hallberg J (2018) Automatic annotation for human activity recognition in free living using a smartphone. Sensors 18(7):2203
- Esteva et al. (2017) Esteva A, Kuprel B, Novoa RA, Ko J, Swetter SM, Blau HM, Thrun S (2017) Dermatologist-level classification of skin cancer with deep neural networks. Nature 542(7639):115–118
- Harrison and Rubinfeld (1978) Harrison D, Rubinfeld DL (1978) Hedonic housing prices and the demand for clean air. Journal of Environmental Economics and Management 5(1):81 – 102
- He and Niyogi (2004) He X, Niyogi P (2004) Locality preserving projections. In: Advances in Neural Information Processing Systems 16, MIT Press, Cambridge, MA, USA, pp 153–160
- Kiryo et al. (2017) Kiryo R, Niu G, Du Plessis MC, Sugiyama M (2017) Positive-unlabeled learning with non-negative risk estimator. In: Advances in neural information processing systems, pp 1675–1685
- Konishi et al. (2019) Konishi K, Mimura M, Nonaka T, Sase I, Nishioka H, Suga M (2019) Practical method of cell segmentation in electron microscope image stack using deep convolutional neural network. Microscopy 68(4):338–341
- Li et al. (2016) Li Y, Wu B, Ghanem B, Zhao Y, Yao H, Ji Q (2016) Facial action unit recognition under incomplete data based on multi-label learning with missing labels. Pattern Recognition 60:890–900
- Odena (2016) Odena A (2016) Semi-supervised learning with generative adversarial networks. arXiv preprint arXiv:160601583
- Park et al. (2016) Park JK, Kwon BK, Park JH, Kang DJ (2016) Machine learning-based imaging system for surface defect inspection. International Journal of Precision Engineering and Manufacturing-Green Technology 3(3):303–310
- Pesapane et al. (2018) Pesapane F, Codari M, Sardanelli F (2018) Artificial intelligence in medical imaging: threat or opportunity? radiologists again at the forefront of innovation in medicine. European radiology experimental 2(1):35
- Ren et al. (2017) Ren R, Hung T, Tan KC (2017) A generic deep-learning-based approach for automated surface inspection. IEEE transactions on cybernetics 48(3):929–940
- Sakai et al. (2017) Sakai T, du Plessis MC, Niu G, Sugiyama M (2017) Semi-supervised classification based on classification from positive and unlabeled data. In: Proceedings of the 34th International Conference on Machine Learning-Volume 70, JMLR. org, pp 2998–3006
- Scheirer et al. (2012) Scheirer WJ, de Rezende Rocha A, Sapkota A, Boult TE (2012) Toward open set recognition. IEEE transactions on pattern analysis and machine intelligence 35(7):1757–1772
- Shahriyar et al. (2018) Shahriyar SA, Alam KMR, Roy SS, Morimoto Y (2018) An approach for multi label image classification using single label convolutional neural network. In: 2018 21st international conference of computer and information technology (ICCIT), IEEE, pp 1–6
- Sugiyama (2015) Sugiyama M (2015) Introduction to Statistical Machine Learning. Morgan Kaufmann, Amsterdam, The Netherlands
- Vapnik (1995) Vapnik VN (1995) The Nature of Statistical Learning Theory. Springer-Verlag, Berlin, Germany
- Wagner et al. (2005) Wagner J, Kim J, André E (2005) From physiological signals to emotions: Implementing and comparing selected methods for feature extraction and classification. In: 2005 IEEE international conference on multimedia and expo, IEEE, pp 940–943
- Wang et al. (2016) Wang D, Khosla A, Gargeya R, Irshad H, Beck AH (2016) Deep learning for identifying metastatic breast cancer. arXiv preprint arXiv:160605718
Appendix A Proof of Lemma 1
Appendix B Proof of Theorem 3.1
At the condition of Eq. (22), the expectation value of the loss is
(31) | |||||
To find the minimum of the expectation value, we derive a linear programming problem considering as independent variables. As shown in Figure 6, we can calculate the boundary conditions subject to as,
(32) |

Thus, we determine the minimizers of subject to .
(33) | |||
(37) |
∎
Appendix C Relation between the CRO-SVM-RL and the CAD-SVM
For a dataset which contains positive, negative and ambiguous samples, we define a RL (randomly labeled) dataset as a dataset in which ambiguous samples are randomly relabeled into the positive or negative classes.
Theorem C.1
The risk of the 0-1- loss function for the RL dataset is equal to the risk of the 0-1-- loss with for the original dataset.
Proof
Let a relabeled label be which satisfies
(38) |
Thus, we can calculate the risk of the 0-1- loss for the relabeled label as
(39) | |||||
Therefore, the risk is equal to that of the 0-1-- loss for the original label . Note that the second term in the right-hand side is constant with respect to and . ∎
Theorem C.2
The MH loss for the RL dataset is convex and an upper bound of the 0-1-- loss with for the original dataset.
Proof
For the ambiguous label, we can calculate the expectation value of the MH loss function over the relabeling process as
(40) | |||||
Since is convex, the expectation is also convex. For positive and negative labels, it can be calculated in the same manner. ∎
Theorem C.3
The expectation of the risk of the MH loss for the RL dataset is calibrated to the 0-1-- loss with for the original dataset under the conditions of Eq. (22).
Proof
We minimize the expectation of the MH loss with respect to for the RL dataset as
(44) | |||||
Then, we calculate the minimizers as
(48) | |||||
The derived minimizers are consistent with Lemma 1 with . ∎