A Second-Order Approach to Learning with Instance-Dependent Label Noise
Abstract
The presence of label noise often misleads the training of deep neural networks. Departing from the recent literature which largely assumes the label noise rate is only determined by the true label class, the errors in human-annotated labels are more likely to be dependent on the difficulty levels of tasks, resulting in settings with instance-dependent label noise. We first provide evidences that the heterogeneous instance-dependent label noise is effectively down-weighting the examples with higher noise rates in a non-uniform way and thus causes imbalances, rendering the strategy of directly applying methods for class-dependent label noise questionable. Built on a recent work peer loss [24], we then propose and study the potentials of a second-order approach that leverages the estimation of several covariance terms defined between the instance-dependent noise rates and the Bayes optimal label. We show that this set of second-order statistics successfully captures the induced imbalances. We further proceed to show that with the help of the estimated second-order statistics, we identify a new loss function whose expected risk of a classifier under instance-dependent label noise is equivalent to a new problem with only class-dependent label noise. This fact allows us to apply existing solutions to handle this better-studied setting. We provide an efficient procedure to estimate these second-order statistics without accessing either ground truth labels or prior knowledge of the noise rates. Experiments on CIFAR10 and CIFAR100 with synthetic instance-dependent label noise and Clothing1M with real-world human label noise verify our approach. Our implementation is available at https://github.com/UCSC-REAL/CAL.
1 Introduction
Deep neural networks (DNNs) are powerful in revealing and fitting the relationship between feature and label when a sufficiently large dataset is given. However, the label usually requires costly human efforts for accurate annotations. With limited budgets/efforts, the resulting dataset would be noisy, and the existence of label noise may mislead DNNs to learn or memorize wrong correlations [10, 11, 35, 38, 47]. To make it worse, the label noise embedded in human annotations is often instance-dependent, e.g., some difficult examples are more prone to be mislabeled [34]. This hidden and imbalanced distribution of noise often has a detrimental effect on the training outcome [15, 23]. It remains an important and challenging task to learn with instance-dependent label noise.
Theory-supported works addressing instance-dependent label noise mostly rely on loss correction, which requires estimating noise rates [40]. Recent work has also considered the possibility of removing the dependency on estimating noise rates [5]. The proposed solution uses a properly specified regularizer to eliminate the effect of instance-dependent label noise. The common theme of the above methods is the focus on learning the underlying clean distribution by using certain forms of first-order statistics of model predictions. In this paper, we propose a second-order approach with the assistance of additional second-order statistics and explore how this information can improve the robustness of learning with instance-dependent label noise. Our main contributions summarize as follows.
-
1.
Departing from recent works [5, 24, 27, 29, 32, 40, 42] which primarily rely on the first-order statistics (i.e. expectation of the models’ predictions) to improve the robustness of loss functions, we propose a novel second-order approach and emphasize the importance of using second-order statistics (i.e. several covariance terms) when dealing with instance-dependent label noise.
-
2.
With the perfect knowledge of the covariance terms defined above, we identify a new loss function that transforms the expected risk of a classifier under instance-dependent label noise to a risk with only class-dependent label noise, which is an easier case and can be handled well by existing solutions. Based on peer loss [24], we further show the expected risk of class-dependent noise is equivalent to an affine transformation of the expected risk under the Bayes optimal distribution. Therefore we establish that our new loss function for Covariance-Assisted Learning (CAL) will induce the same minimizer as if we can access the clean Bayes optimal labels.
-
3.
We show how the second-order statistics can be estimated efficiently using existing sample selection techniques. For a more realistic case where the covariance terms cannot be perfectly estimated, we prove the worst-case performance guarantee of our solution.
-
4.
In addition to the theoretical guarantees, the performance of the proposed second-order approach is tested on the CIFAR10 and CIFAR100 datasets with synthetic instance-dependent label noise and the Clothing1M dataset with real-world human label noise.
1.1 Related Works
Below we review the most relevant literature.
Bounded loss functions Label noise encodes a different relation between features and labels. A line of literature treats the noisy labels as outliers. However, the convex loss functions are shown to be prone to mistakes when outliers exist [25]. To handle this setting, the cross-entropy (CE) loss can be generalized by introducing temperatures to logarithm functions and exponential functions [1, 2, 50]. Noting the CE loss grows explosively when the prediction approaches zero, some solutions focus on designing bounded loss functions [8, 9, 31, 36]. These methods focus on the numerical property of loss functions, and most of them do not discuss the type of label noise under treatment.
Learning clean distributions To be noise-tolerant [26], it is necessary to understand the effect of label noise statistically. With the class-dependent assumption, the loss can be corrected/reweighted when the noise transition is available, which can be estimated by discovering anchor points [22, 29, 39], exploiting clusterability [51], regularizing total variation [49], or minimizing volume of [20]. The loss correction/reweighting methods rely closely on the quality of the estimated noise transition matrix. To make it more robust, an additive slack variable [41] or a multiplicative dual [45] can be used for revision. Directly extending these loss correction methods to instance-dependent label noise is prohibitive since the transition matrix will become a function of feature and the number of parameters to be estimated is proportional to the number of training instances. Recent follow-up works often introduce extra assumption [40] or measure [4]. Statistically, the loss correction approach is learning the underlying clean distribution if a perfect is applied. When the class-dependent noise rate is known, surrogate loss [27], an unbiased loss function targeting on binary classifications, also learns the clean distribution. Additionally, the symmetric cross-entropy loss [36], an information-based loss [43], a correlated agreement (CA) based loss peer loss [24], and its adaptation for encouraging confident predictions [5] are proposed to learn the underlying clean distribution without knowing the noise transition matrix.
2 Preliminaries
This paper targets on a classification problem given a set of training examples with Instance-Dependent label Noise (IDN) denoted by , where is the set of indices. The corresponding noisy data distribution is denoted by . Examples are drawn according to random variables . Our goal is to design a learning mechanism that is guaranteed to be robust when learning with only accessing . Before proceeding, we summarize important definitions as follows.
Clean distribution Each noisy example corresponds to a clean example , which contains one unobservable ground-truth label, a.k.a. clean label. Denote by the clean distribution. Clean examples are drawn from random variables .
Bayes optimal distribution Denote by the Bayes optimal label given feature , that is: The distribution of is denoted by . Note the Bayes optimal distribution is different from the clean distribution when . Due to the fact that the information encoded between features and labels is corrupted by label noise, and both clean labels and Bayes optimal labels are unobservable, inferring the Bayes optimal distribution from the noisy dataset is a non-trivial task. Notably there exist two approaches [5, 6] that provide guarantees on constructing the Bayes optimal dataset. We would like to remind the readers that the noisy label , clean label , and Bayes optimal label for the same feature may disagree with each other.
Most of our developed approaches will focus on dealing with the Bayes optimal distribution . By referring to , as we shall see later, we are allowed to estimate the second-order statistics defined w.r.t. .
Noise transition matrix Traditionally, the noise transition matrix is defined based on the relationship between clean distributions and noisy distributions [5, 24, 30, 40]. In recent literature [6], the Bayes optimal label (a.k.a. distilled label in [6]) also plays a significant role. In the image classification tasks where the performance is measured by the clean test accuracy, predicting the Bayes optimal label achieves the best performance. This fact motivates us to define a new noise transition matrix based on the Bayes optimal label as follows:
where denotes the -th element of the matrix . Its expectation is defined as , with the -th element being .
Other notations Let and 111We focus on the closed-set label noise, i.e. , , and share the same space . be the space of feature and label , respectively. The classification task aims to identify a classifier that maps to accurately. One common approach is minimizing the empirical risk using DNNs with respect to the cross-entropy (CE) loss defined as: where denotes the -th component of and is the number of classes. Let be the indicator function taking value when the specified condition is satisfied and otherwise. Define the 0-1 loss as Define the Bayes optimal classifier as Noting the CE loss is classification-calibrated [3], given enough clean data, the Bayes optimal classifier can be learned using the CE loss:
Goal Different from the goals in surrogate loss [27], [43], peer loss [24], and CORES2 [5], which focus on recovering the performance of learning on clean distributions, we aim to learn a classifier from the noisy distribution which also minimizes . Note holds for the Bayes optimal classifier . Thus, in the sense of searching for the Bayes optimal classifier, our goals are aligned with the ones focusing on the clean distribution.
3 Insufficiency of First-Order Statistics
Peer loss [24] and its inspired confidence regularizer [5] are two recently introduced robust losses that operate without the knowledge of noise transition matrices, which presents them as preferred solutions for more complex noise settings. In this section, we will first review the usages of first-order statistics in peer loss and the confidence regularizer (Section 3.1), and then analyze the insufficiency of using only the first-order statistics when handling the challenging IDN (Section 3.2). Besides, we will anatomize the down-weighting effect of IDN and provide intuitions for how to make IDN easier to handle (Section 3.3).
We formalize our arguments using peer loss, primarily due to 1) its clean analytical form, and 2) that our later proposed solution will be built on peer loss too. Despite the focus on peer loss, we believe these observations are generally true when other existing training approaches meet IDN.
For ease of presentation, the following analyses focus on binary cases (with classes ). Note the class should be mapped to class following the notations in Section 2. For a clear comparison with previous works, we follow the notation in [24] and use class to represent classes when . The error rates in are then denoted as , . Most of the discussions generalize to the multi-class setting.
3.1 Using First-Order Statistics in Peer Loss
It has been proposed and proved in peer loss [24] and CORES2 [5] that the learning could be robust to label noise by considering some first-order statistics related to the model predictions. For each example , peer loss [24] has the following form:
where and are two randomly sampled peer samples for . The first-order statistics related to model predictions characterized by the peer term are further extended to a confidence regularizer in CORES2 [5]:
where is a hyperparameter controlling the ability of regularizer, and is the marginal distribution of given dataset . Although it has been shown in [5] that learning with an appropriate would be robust to instance-dependent label noise theoretically, in real experiments, converging to the guaranteed optimum by solving a highly non-convex problem is difficult.
3.2 Peer Loss with IDN
Now we analyze the possible performance degradation of using the binary peer loss function proposed in [24] to handle IDN. Denote by
the optimal classifier learned by minimizing 0-1 peer loss, where represents with 0-1 loss (could also be generalized for with 0-1 loss). Let . With a bounded variance in the error rates, supposing
the worst-case performance bound for using pure peer loss is provided in Theorem 1 and proved in Appendix B.1.
Theorem 1 (Performance of peer loss).
With the peer loss function proposed in [24], we have
Theorem 1 shows the ratio of wrong predictions given by includes two components. The former term is directly caused by IDN, indicating the error is increasing when the instance-dependent noise rates have larger mean (larger ) and larger variation (larger ). The latter term shows possible errors induced by an unbalanced . Theorem 1 generalizes peer loss where , i.e., the error rates are homogeneous across data instances, and there is no need to consider any second-order statistics that involve the distribution of noise rates.
3.3 Down-weighting Effect of IDN
We further discuss motivations and intuitions by studying how IDN affects the training differently from the class-dependent one. Intuitively, a high noise rate reduces the informativeness of a particular example , therefore “down-weighting" its contribution to training. We now analytically show this under peer loss.
As a building block, the invariant property (in terms of the clean distribution ) originally discovered by peer loss on class-dependent label noise is first adapted for the Bayes optimal distribution . Define and . Focusing on a particular class-dependent , we provide Lemma 1 and its proof in Appendix A.1.
Lemma 1 (Invariant property of peer loss [24]).
Peer loss is invariant to class-dependent label noise:
(1) |
Then we discuss the effect of IDN. Without loss of generality, consider a case where noisy examples are drawn from two noisy distributions and , and the noise rate of is higher than , i.e. , where . Assume a particular setting of IDN that the noise is class-dependent (but not instance-dependent) only within each distribution, and different between two distributions, i.e. part-dependent [40]. Let and be the Bayes optimal distribution related to and . For simplicity, we write as . Then and . Note and Then we have the following equality:
where
indicates down-weighting examples drawn from (compared to the class-dependent label noise).
What can we learn from this observation? First, we show the peer loss is already down weighting the importance of the more noisy examples. However, simply dropping examples with potentially high-level noise might lead the classifier to learn a biased distribution. Moreover, subjectively confusing examples are more prone to be mislabeled and critical for accurate predictions [34], thus need to be carefully addressed. Our second observation is that if we find a way to compensate for the “imbalances” caused by the down-weighting effects shown above, the challenging instance-dependent label noise could be transformed into a class-dependent one, which existing techniques can then handle. More specifically, the above result shows the down-weighting effect is characterized by , implying only using the first-order statistics of model predictions without considering the distributions of the noise transition matrix is insufficient to capture the complexity of the learning task. However, accurately estimating is prohibitive since the number of parameters to be estimated is almost at the order of – recall is the number of training examples and is the number of classes. Even though we can roughly estimate , applying element-wise correction relying on the estimated may accumulate errors. Therefore, to achieve the transformation from the instance-dependent to the easier class-dependent, we need to resort to other statistical properties of .
4 Covariance-Assisted Learning (CAL)
From the analyses in Section 3.3, we know the instance-dependent label noise will “automatically” assign different weights to examples with different noise rates, thus cause imbalances. When the optimal solution does not change under such down-weighting effects, the first-order statistics based on peer loss [5, 24] work well. However, for a more robust and general solution, using additional information to “balance” the effective weights of different examples is necessary. Although the Bayes optimal distribution is not accessible in real experiments, we first assume its existence for theoretical analyses in the ideal case, then we will discuss the gap to this optimal solution when we can only use a proxy that can be constructed efficiently.
4.1 Extracting Covariance from IDN
Again consider an instance-dependent noisy distribution with binary classes where . Define the following two random variables to facilitate analyses:
Recall and . Let be the covariance between random variables and w.r.t. the distribution . The exact effects of IDN on peer loss functions are revealed in Theorem 2 and proved in Appendix B.2.
Theorem 2 (Decoupling binary IDN).
In binary classifications, the expected peer loss with IDN writes as:
(2) |
Theorem 2 effectively divides the instance-dependent label noise into two parts. As shown in Eq. (2), the first line is the same as Eq. (1) in Lemma 1, indicating the average effect of instance-dependent label noise can be treated as a class-dependent one with parameters . The additional two covariance terms in the second and the third lines of Eq. (2) characterize the additional contribution of examples due to their differences in the label noise rates. The covariance terms will become larger for a setting with more diverse noise rates, capturing a more heterogeneous and uncertain learning environment. Interested readers are also referred to the high-level intuitions for using covariance terms at the end of Section 3.3.
We now briefly discuss one extension of Theorem 2 to a -class classification task. Following the assumption adopted in [24], we consider a particular setting of IDN whose the expected transition matrix satisfies . Denote by . Corollary 1 decouples the effects of IDN in multi-class cases and is proved in Appendix C.1.
Corollary 1 (Decoupling multi-class IDN).
In multi-class classifications, when the expected transition matrix satisfies , the expected peer loss with IDN writes as:
where is the marginal distribution of and is the conditional distribution of given .
4.2 Using Second-Order Statistics
Inspired by Theorem 2, if is available, we can subtract two covariance terms and make peer loss invariant to IDN. Specifically, define
We have the following optimality guarantee and its proof is deferred to Appendix B.3.
Theorem 3.
For a -class classification problem, a general loss function for our Covariance-Assisted Learning (CAL) approach is given by
(3) |
Eq. (3) shows the Bayes optimal distribution is critical in implementing the proposed covariance terms. However, cannot be obtained trivially, and only imperfect proxy constructions of the dataset (denoted by ) could be expected. Detailed constructions of are deferred to Section 4.2.1.
Advantages of using covariance terms There are several advantages of using the proposed covariance terms. Unlike directly correcting labels according to , the proposed covariance term can be viewed as a “soft” correction that maintains the information encoded in both original noisy labels and the estimated Bayes optimal labels. Keeping both information is beneficial as suggested in [13]. Moreover, compared to the direct loss correction approaches [30, 40, 41], we keep the original learning objective and apply “correction” using an additional term. Our method is more robust in practice compared to these direct end-to-end loss correction approaches due to two reasons: 1) The covariance term summarizes the impact of the complex noise using an average term, indicating that our approach is less sensitive to the estimation precision of an individual example; 2) As will be shown in Section 4.3, the proposed method is tolerant with accessing an imperfect .
Estimating the covariance terms relies on samples drawn from distribution . Thus, we need to construct a dataset , which is similar or unbiased w.r.t. . We will first show the algorithm for constructing , then provide details for DNN implementations.
4.2.1 Constructing
To achieve unbiased estimates of the variance terms, the high-level intuition for constructing is determining whether the label of each example in is Bayes optimal or not by comparing the likelihood, confidence, or loss of classifying the (noisy) label to some thresholds. There are several methods for constructing : distillation [6], searching to exploit [44], and sample sieve [5]. If the model does not overfit the label noise and learns the noisy distribution, both methods in [6] and [44] work well. However, for the challenging instance-dependent label noise, overfitting occurs easily thus techniques to avoid overfitting are necessary. In this paper, we primarily adapt the sample sieve proposed in [5], which uses a confidence regularizer to avoid overfitting, to construct . Specifically, as shown in [5], in each epoch , the regularized loss for each example is adjusted by the parameter , which can be calculated based on model predictions in linear time. In the ideal cases assumed in [5], any example with a positive adjusted loss is corrupted (with a wrong label).
We summarized the corresponding procedures in Algorithm 1, where the critical thresholds for comparing losses are denoted by and . At Line 1, if the loss adjusted by is small enough (smaller than the threshold ), we assume is the Bayes optimal label. Accordingly, at Line 1, if the adjusted loss is too large (larger than the threshold ), we treat as a corrupted one and assume the class with maximum predicted probability to be Bayes optimal one. For the examples with moderate adjusted loss, we drop it as indicated in Line 1. In ideal cases with infinite model capacity and sufficiently many examples (as assumed in [5]), we can set thresholds to guarantee a separation of clean and corrupted examples, thus will be an unbiased proxy to 222In the ideal case as assumed in Corollary 1 of [5], we have .. However, in real experiments, when both the model capacity and the number of examples are limited, we may need to tune and to obtain a high-quality construction of . In this paper, we set to ensure and reduce the effort to tuning both thresholds simultaneously.
Note that using to estimate the covariance terms could be made theoretically more rigorous by applying appropriate re-weighting techniques [6, 7, 14]. See Appendix D.1 for more discussions and corresponding guarantees. We omit the details here due to the space limit. Nonetheless, our approach is tolerant of an imperfect , which will be shown theoretically in Section 4.3.
4.2.2 Implementations
For implementations with deep neural network solutions, we need to estimate the transition matrix relying on and estimate the covariance terms along with stochastic gradient descent (SGD) updates.
Covariance Estimation in SGD As required in (3), with a particular , each computation for requires only one time check of the associated noisy label as follows:
(4) |
When is unbiased w.r.t. , the estimation in (4) is also unbiased because
Noting the covariance can be estimated empirically as
For each batch of data, the above estimation has complexities in computation and space. To reduce both complexities, with the cost of the estimation quality, we use examples to estimate the covariance in each batch, where is the set of sample indices of batch-. Per sample wise, Eq. (3) can be transformed to
With the above implementation, the estimation is done locally for each point in complexity.
4.3 CAL with Imperfect Covariance Estimates
As mentioned earlier, cannot be perfectly obtained in practice. Thus, there is a performance gap between the ideal case (with perfect knowledge of ) and the actually achieved one. We now analyze the effect of imperfect covariance terms (Theorem 4).
Denote the imperfect covariance estimates by , where is the expected ratio (a.k.a. probability) of correct examples in : With , the minimizer of the 0-1 CAL loss is given by:
Theorem 4 reports the error bound produced by . See Appendix B.4 for the proof.
Theorem 4 (Imperfect Covariance).
With , when , we have
Theorem 4 shows the quality of controls the scale of the worst-case error upper-bound. Compared with Theorem 1 where no covariance term is used, we know the covariance terms will always be helpful when . That is, the training with the assistance of covariance terms will achieve better (worst-case) accuracy on the Bayes optimal distribution when the construction is better than a dataset that includes each instance in randomly with 50% chance.
5 Experiments
We now present our experiment setups and results.
5.1 General Experiment Settings
Datasets and models The advantage of introducing our second-order approach is evaluated on three benchmark datasets: CIFAR10, CIFAR100 [17] and Clothing1M [42]. Following the convention from [5, 43], we use ResNet34 for CIFAR10 and CIFAR100 and ResNet50 for Clothing1M. Noting the expected peer term (a.k.a. confidence regularizer (CR) as implemented in [5]) is more stable and converges faster than the one with peer samples, we train with . It also enables a fair ablation study since is constructed relying on [5]. For numerical stability, we use a cut-off version of the cross-entropy loss . Specifically, we use for the traditional cross-entropy term, use for the CR term, and the covariance term. All the experiments use a momentum of . The weight decay is set as for CIFAR experiments and for Clothing1M.
Noise type For CIFAR datasets, the instance-dependent label noise is generated following the method from [5, 40]. The basic idea is randomly generating one vector for each class ( vectors in total) and project each incoming feature onto these vectors. The label noise is added by jointly considering the clean label and the projection results. See Appendix D.2 for details. In expectation, the noise rate is the overall ratio of examples with a wrong label in the entire dataset. For the Clothing1M dataset, we train on 1 million noisy training examples that encode the real-world human noise.
5.2 Baselines
We compare our method with several related works, where the cross-entropy loss is tested as a common baseline. Additionally, the generalized cross-entropy [50] is compared as a generalization of mean absolute error and cross-entropy designed for label noise. Popular loss correction based methods [29, 40, 41], sample selection based methods [5, 12, 37, 46], and noise-robust loss functions [24, 43] are also chosen for comparisons. All the compared methods adopt similar data augmentations, including standard random crop, random flip, and normalization. Note the recent work on part-dependent label noise [40] did not apply random crop and flip on the CIFAR dataset. For a fair comparison with [40], we remove the corresponding data augmentations from our approach and defer the comparison to Appendix D.3. The semi-supervised learning based methods with extra feature-extraction and data augmentations are not included. All the CIFAR experiments are repeated times with independently synthesized IDN. The highest accuracies on the clean testing dataset are averaged over trials to show the best generalization ability of each method.
5.3 Performance Comparisons
5.3.1 CIFAR
Method | Inst. CIFAR10 | Inst. CIFAR100 | ||||
---|---|---|---|---|---|---|
CE (Standard) | 85.450.57 | 76.231.54 | 59.751.30 | 57.791.25 | 41.150.83 | 25.681.55 |
Forward [29] | 87.221.60 | 79.372.72 | 66.564.90 | 58.191.37 | 42.801.01 | 27.913.35 |
[43] | 88.570.60 | 82.821.49 | 69.941.31 | 57.901.21 | 42.700.92 | 26.962.08 |
[50] | 85.810.83 | 74.661.12 | 60.763.08 | 57.030.27 | 39.811.18 | 24.872.46 |
Co-teaching [12] | 88.870.24 | 73.001.24 | 62.511.98 | 43.300.39 | 23.210.57 | 12.580.51 |
Co-teaching+ [46] | 89.800.28 | 73.781.39 | 59.226.34 | 41.710.78 | 24.450.71 | 12.580.51 |
JoCoR [37] | 88.780.15 | 71.643.09 | 63.461.58 | 43.661.32 | 23.950.44 | 13.160.91 |
Reweight-R [41] | 90.040.46 | 84.112.47 | 72.182.47 | 58.000.36 | 43.838.42 | 36.079.73 |
Peer Loss [24] | 89.120.76 | 83.260.42 | 74.531.22 | 61.160.64 | 47.231.23 | 31.712.06 |
CORES2 [5] | 91.140.46 | 83.671.29 | 77.682.24 | 66.470.45 | 58.991.49 | 38.553.25 |
CAL | 92.010.75 | 84.961.25 | 79.822.56 | 69.110.46 | 63.171.40 | 43.583.30 |
In experiments on CIFAR datasets, we use a batch size of , an initial learning rate of , and reduce it by a factor of at epoch .
Construct To construct , we update the DNN for epochs by minimizing (without dynamic sample sieve) and apply Algorithm 1 with .333Theoretically, we have if both the CE term and the CR term use a log loss without cut-off (). Current setting works well (not the best) for CIFAR experiments empirically. For a numerically stable solution, we use the square root of the noise prior for the CR term in as The hyperparameter is set to for CIFAR10 and for CIFAR100.
Train with CAL With an estimate of , we re-train the model epochs. The hyper-parameter is set to for CIFAR10 and for CIFAR100. Note the hyperparameters (, , ) can be better set if a clean validation set is available.
Performance Table 1 compares the means and standard deviations of test accuracies on the clean test dataset when the model is trained with synthesized instance-dependent label noise in different levels. All the compared methods use ResNet34 as the backbone. On CIFAR10, with a low-level label noise (), all the compared methods perform well and achieve higher average test accuracies than the standard CE loss. When the overall noise rates increase to high, most of the methods suffer from severe performance degradation while CAL still achieves the best performance. There are similar observations on CIFAR100. By comparing CAL with CORES2, we conclude that the adopted second-order statistics do work well and bring non-trivial performance improvement. Besides, on the CIFAR100 dataset with and , we observe Reweight-R [41] has a large standard deviation and a relatively high mean, indicating it may perform as well as or even better than CAL in some trials. It also shows the potential of using a revised transition matrix [41] in severe and challenging instance-dependent label noise settings.
5.3.2 Clothing1M
For Clothing1M, we first train the model following the settings in [5] and construct with the best model. Noting the overall accuracy of noisy labels in Clothing1M is about [42], we set an appropriate such that of training examples satisfying . With , we sample a class-balanced dataset by randomly choosing noisy examples for each class and continue training the model with and an initial learning rate of for epochs. Other parameters are set following [5]. See Appendix D.4 for more detailed experimental settings. Table 2 shows CAL performs well in the real-world human noise.
5.4 Ablation Study
Table 3 shows either the covariance term or the peer term can work well individually and significantly improve the performance when they work jointly. Comparing the first row with the second row, we find the second-order statistics can work well (except for ) even without the peer (CR) term. In row 4, we show the performance at epoch since the second-order statistics are estimated relying on the model prediction at this epoch. By comparing row 4 with row 5, we know the second-order statistics indeed lead to non-trivial improvement in the performance. Even though the covariance term individually can only achieve an accuracy of when , it can still contribute more than of the performance improvement (from to ) when it is implemented with the peer term. This observation shows the robustness of CAL.
row # | Cov. | Peer | Epoch | |||
---|---|---|---|---|---|---|
1 | ✗ | ✗ | Best | 90.47 | 82.56 | 64.65 |
2 | ✓ | ✗ | Best | 92.10 | 78.49 | 73.55 |
3 | ✗ | ✓ | Best | 91.85 | 84.41 | 78.74 |
4 | ✗ | ✓ | Fixed@65 | 90.73 | 82.76 | 77.70 |
5 | ✓ | ✓ | Best | 92.69 | 85.55 | 81.54 |
6 Conclusions
This paper has proposed a second-order approach to transforming the challenging instance-dependent label noise into a class-dependent one such that existing methods targeting the class-dependent label noise could be implemented. Currently, the necessary information for the covariance term is estimated based on a sample selection method. Future directions of this work include extensions to other methods for estimating the covariance terms accurately. We are also interested in exploring the combination of second-order information with other robust learning techniques.
Acknowledgements This research is supported in part by National Science Foundation (NSF) under grant IIS-2007951, and in part by Australian Research Council Projects, i.e., DE-190101473.
References
- [1] Ehsan Amid, Manfred KK Warmuth, Rohan Anil, and Tomer Koren. Robust bi-tempered logistic loss based on bregman divergences. In Advances in Neural Information Processing Systems, pages 14987–14996, 2019.
- [2] Ehsan Amid, Manfred K Warmuth, and Sriram Srinivasan. Two-temperature logistic regression based on the tsallis divergence. In The 22nd International Conference on Artificial Intelligence and Statistics, pages 2388–2396. PMLR, 2019.
- [3] Peter L Bartlett, Michael I Jordan, and Jon D McAuliffe. Convexity, classification, and risk bounds. Journal of the American Statistical Association, 101(473):138–156, 2006.
- [4] Antonin Berthon, Bo Han, Gang Niu, Tongliang Liu, and Masashi Sugiyama. Confidence scores make instance-dependent label-noise learning possible. arXiv preprint arXiv:2001.03772, 2020.
- [5] Hao Cheng, Zhaowei Zhu, Xingyu Li, Yifei Gong, Xing Sun, and Yang Liu. Learning with instance-dependent label noise: A sample sieve approach. In International Conference on Learning Representations, 2021.
- [6] Jiacheng Cheng, Tongliang Liu, Kotagiri Ramamohanarao, and Dacheng Tao. Learning with bounded instance-and label-dependent label noise. In Proceedings of the 37th International Conference on Machine Learning, ICML ’20, 2020.
- [7] Tongtong Fang, Nan Lu, Gang Niu, and Masashi Sugiyama. Rethinking importance weighting for deep learning under distribution shift. In H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, and H. Lin, editors, Advances in Neural Information Processing Systems, volume 33, pages 11996–12007. Curran Associates, Inc., 2020.
- [8] Aritra Ghosh, Himanshu Kumar, and PS Sastry. Robust loss functions under label noise for deep neural networks. In Thirty-First AAAI Conference on Artificial Intelligence, 2017.
- [9] Maoguo Gong, Hao Li, Deyu Meng, Qiguang Miao, and Jia Liu. Decomposition-based evolutionary multiobjective optimization to self-paced learning. IEEE Transactions on Evolutionary Computation, 23(2):288–302, 2018.
- [10] Bo Han, Gang Niu, Xingrui Yu, Quanming Yao, Miao Xu, Ivor Tsang, and Masashi Sugiyama. Sigua: Forgetting may make learning with noisy labels more robust. In International Conference on Machine Learning, pages 4006–4016. PMLR, 2020.
- [11] Bo Han, Quanming Yao, Tongliang Liu, Gang Niu, Ivor W Tsang, James T Kwok, and Masashi Sugiyama. A survey of label-noise representation learning: Past, present and future. arXiv preprint arXiv:2011.04406, 2020.
- [12] Bo Han, Quanming Yao, Xingrui Yu, Gang Niu, Miao Xu, Weihua Hu, Ivor Tsang, and Masashi Sugiyama. Co-teaching: Robust training of deep neural networks with extremely noisy labels. In Advances in neural information processing systems, pages 8527–8537, 2018.
- [13] Jiangfan Han, Ping Luo, and Xiaogang Wang. Deep self-learning from noisy labels. In Proceedings of the IEEE International Conference on Computer Vision, pages 5138–5147, 2019.
- [14] Jiayuan Huang, Arthur Gretton, Karsten Borgwardt, Bernhard Schölkopf, and Alex J Smola. Correcting sample selection bias by unlabeled data. In Advances in neural information processing systems, pages 601–608, 2007.
- [15] Lu Jiang, Di Huang, Mason Liu, and Weilong Yang. Beyond synthetic noise: Deep learning on controlled noisy labels. In Proceedings of the 37th International Conference on Machine Learning, volume 119, pages 4804–4815. PMLR, 13–18 Jul 2020.
- [16] Lu Jiang, Zhengyuan Zhou, Thomas Leung, Li-Jia Li, and Li Fei-Fei. Mentornet: Learning data-driven curriculum for very deep neural networks on corrupted labels. In International Conference on Machine Learning, pages 2304–2313. PMLR, 2018.
- [17] Alex Krizhevsky, Geoffrey Hinton, et al. Learning multiple layers of features from tiny images. Technical report, Citeseer, 2009.
- [18] Kuang-Huei Lee, Xiaodong He, Lei Zhang, and Linjun Yang. Cleannet: Transfer learning for scalable image classifier training with label noise. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 5447–5456, 2018.
- [19] Junnan Li, Richard Socher, and Steven C.H. Hoi. Dividemix: Learning with noisy labels as semi-supervised learning. In International Conference on Learning Representations, 2020.
- [20] Xuefeng Li, Tongliang Liu, Bo Han, Gang Niu, and Masashi Sugiyama. Provably end-to-end label-noise learning without anchor points. arXiv preprint arXiv:2102.02400, 2021.
- [21] Yuncheng Li, Jianchao Yang, Yale Song, Liangliang Cao, Jiebo Luo, and Li-Jia Li. Learning from noisy labels with distillation. In Proceedings of the IEEE International Conference on Computer Vision, pages 1910–1918, 2017.
- [22] Tongliang Liu and Dacheng Tao. Classification with noisy labels by importance reweighting. IEEE Transactions on pattern analysis and machine intelligence, 38(3):447–461, 2015.
- [23] Yang Liu. The importance of understanding instance-level noisy labels, 2021.
- [24] Yang Liu and Hongyi Guo. Peer loss functions: Learning from noisy labels without knowing noise rates. In Proceedings of the 37th International Conference on Machine Learning, ICML ’20, 2020.
- [25] Philip M Long and Rocco A Servedio. Random classification noise defeats all convex potential boosters. Machine learning, 78(3):287–304, 2010.
- [26] Naresh Manwani and PS Sastry. Noise tolerance under risk minimization. IEEE transactions on cybernetics, 43(3):1146–1151, 2013.
- [27] Nagarajan Natarajan, Inderjit S Dhillon, Pradeep K Ravikumar, and Ambuj Tewari. Learning with noisy labels. In Advances in neural information processing systems, pages 1196–1204, 2013.
- [28] Duc Tam Nguyen, Chaithanya Kumar Mummadi, Thi Phuong Nhung Ngo, Thi Hoai Phuong Nguyen, Laura Beggel, and Thomas Brox. Self: Learning to filter noisy labels with self-ensembling. In International Conference on Learning Representations, 2020.
- [29] Giorgio Patrini, Alessandro Rozza, Aditya Krishna Menon, Richard Nock, and Lizhen Qu. Making deep neural networks robust to label noise: A loss correction approach. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 1944–1952, 2017.
- [30] Giorgio Patrini, Alessandro Rozza, Aditya Krishna Menon, Richard Nock, and Lizhen Qu. Making deep neural networks robust to label noise: A loss correction approach. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), July 2017.
- [31] Jun Shu, Qian Zhao, Keyu Chen, Zongben Xu, and Deyu Meng. Learning adaptive loss for robust learning with noisy labels. arXiv preprint arXiv:2002.06482, 2020.
- [32] Arash Vahdat. Toward robustness against label noise in training deep discriminative neural networks. In Advances in Neural Information Processing Systems, pages 5596–5605, 2017.
- [33] Andreas Veit, Neil Alldrin, Gal Chechik, Ivan Krasin, Abhinav Gupta, and Serge Belongie. Learning from noisy large-scale datasets with minimal supervision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 839–847, 2017.
- [34] Qizhou Wang, Bo Han, Tongliang Liu, Gang Niu, Jian Yang, and Chen Gong. Tackling instance-dependent label noise via a universal probabilistic model. arXiv preprint arXiv:2101.05467, 2021.
- [35] Qizhou Wang, Jiangchao Yao, Chen Gong, Tongliang Liu, Mingming Gong, Hongxia Yang, and Bo Han. Learning with group noise. arXiv preprint arXiv:2103.09468, 2021.
- [36] Yisen Wang, Xingjun Ma, Zaiyi Chen, Yuan Luo, Jinfeng Yi, and James Bailey. Symmetric cross entropy for robust learning with noisy labels. In Proceedings of the IEEE International Conference on Computer Vision, pages 322–330, 2019.
- [37] Hongxin Wei, Lei Feng, Xiangyu Chen, and Bo An. Combating noisy labels by agreement: A joint training method with co-regularization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 13726–13735, 2020.
- [38] Xiaobo Xia, Tongliang Liu, Bo Han, Chen Gong, Nannan Wang, Zongyuan Ge, and Yi Chang. Robust early-learning: Hindering the memorization of noisy labels. In International Conference on Learning Representations, 2021.
- [39] Xiaobo Xia, Tongliang Liu, Bo Han, Nannan Wang, Jiankang Deng, Jiatong Li, and Yinian Mao. Extended T: Learning with mixed closed-set and open-set noisy labels. arXiv preprint arXiv:2012.00932, 2020.
- [40] Xiaobo Xia, Tongliang Liu, Bo Han, Nannan Wang, Mingming Gong, Haifeng Liu, Gang Niu, Dacheng Tao, and Masashi Sugiyama. Part-dependent label noise: Towards instance-dependent label noise. In Advances in Neural Information Processing Systems, volume 33, pages 7597–7610, 2020.
- [41] Xiaobo Xia, Tongliang Liu, Nannan Wang, Bo Han, Chen Gong, Gang Niu, and Masashi Sugiyama. Are anchor points really indispensable in label-noise learning? In Advances in Neural Information Processing Systems, pages 6838–6849, 2019.
- [42] Tong Xiao, Tian Xia, Yi Yang, Chang Huang, and Xiaogang Wang. Learning from massive noisy labeled data for image classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 2691–2699, 2015.
- [43] Yilun Xu, Peng Cao, Yuqing Kong, and Yizhou Wang. L_dmi: A novel information-theoretic loss function for training deep nets robust to label noise. In Advances in Neural Information Processing Systems, volume 32, 2019.
- [44] Quanming Yao, Hansi Yang, Bo Han, Gang Niu, and James T Kwok. Searching to exploit memorization effect in learning with noisy labels. In Proceedings of the 37th International Conference on Machine Learning, ICML ’20, 2020.
- [45] Yu Yao, Tongliang Liu, Bo Han, Mingming Gong, Jiankang Deng, Gang Niu, and Masashi Sugiyama. Dual T: Reducing estimation error for transition matrix in label-noise learning. In Advances in Neural Information Processing Systems, volume 33, pages 7260–7271, 2020.
- [46] Xingrui Yu, Bo Han, Jiangchao Yao, Gang Niu, Ivor W Tsang, and Masashi Sugiyama. How does disagreement help generalization against label corruption? arXiv preprint arXiv:1901.04215, 2019.
- [47] Chiyuan Zhang, Samy Bengio, Moritz Hardt, Benjamin Recht, and Oriol Vinyals. Understanding deep learning requires rethinking generalization. arXiv preprint arXiv:1611.03530, 2016.
- [48] Hongyi Zhang, Moustapha Cisse, Yann N. Dauphin, and David Lopez-Paz. mixup: Beyond empirical risk minimization. In International Conference on Learning Representations, 2018.
- [49] Yivan Zhang, Gang Niu, and Masashi Sugiyama. Learning noise transition matrix from only noisy labels via total variation regularization. arXiv preprint arXiv:2102.02414, 2021.
- [50] Zhilu Zhang and Mert Sabuncu. Generalized cross entropy loss for training deep neural networks with noisy labels. In Advances in neural information processing systems, pages 8778–8788, 2018.
- [51] Zhaowei Zhu, Yiwen Song, and Yang Liu. Clusterability as an alternative to anchor points when learning with noisy labels. arXiv preprint arXiv:2102.05291, 2021.
Appendix
Appendix A Proof for Lemmas
A.1 Proof for Lemma 1
A.2 Proof for Lemma 2
Peer Loss on the Bayes Optimal Distribution
Recall our goal is to learn a classifier from the noisy distribution which also minimizes the loss on the corresponding Bayes optimal distribution , i.e. . Before considering the case with label noise, we need to prove peer loss functions induce the Bayes optimal classifier when minimizing the 0-1 loss on as in Lemma 2.
Lemma 2.
Given the Bayes optimal distribution , the optimal peer classifier defined below:
also minimizes .
See the proof below. It has been shown in [24] that Lemma 2 holds for the clean distribution when the clean dataset is class-balanced, i.e. . For the Bayes optimal distribution , as shown in Lemma 2, there is no requirement for the prior .
Proof.
Recall is the Bayes optimal label defined as
We need to prove that the “optimal peer classifier" defined below:
is the same as the Bayes optimal classifier . To see this, suppose the claim is wrong. Denote by (notations and are defined only for this proof):
and denote by . Then
contradicting the optimality of . Thus our claim is proved. ∎
Appendix B Proof for Theorems
B.1 Proof for Theorem 1
B.2 Proof for Theorem 2
Proof.
The covariance in this proof is taken over the Bayes optimal distribution . Recall
and
We first have the following equality:
Term-B can be transformed to:
Similarly, Term-D turns to
Define two random variables
Then Term-A becomes
Similarly, Term-B can be further transformed to
Combining the above results, we have
∎
B.3 Proof for Theorem 3
B.4 Proof for Theorem 4
Proof.
Recall is the expected ratio (a.k.a. probability) of correct examples in , i.e. With , the classifier learned by minimizing the 0-1 CAL loss is
Note
Similarly,
When , and have the same feature set, we have
Therefore,
The rest of the proof can be accomplished by following the proof of Theorem 1. ∎
Appendix C Proof for Corollaries
C.1 Proof for Corollary 1
Appendix D More Discussions
D.1 Setting Thresholds and
In a high level, there are two strategies for setting and : 1) and 2) .
Strategy-1: :
This strategy may provide a higher ratio of true Bayes optimal labels among feasible examples in since some ambiguous examples are dropped. However, dropping examples changes the distribution of (as well as the distribution of the unobservable ), a.k.a. covariate shift [14, 6]. Importance re-weighting with weight is necessary for correcting the covariate shift, i.e. the weight of each feasible example should be changed from to . Let and be the marginal distributions of and on . With a particular kernel , the optimization problem is:
(6) |
The optimal solution is supposed to be . Note the selection of kernel is non-trivial, especially for complicated features [7] in DNN solutions. Using this strategy, with appropriate and such that all the examples in are Bayes optimal, the covariance could be guaranteed to be optimal when each example in is re-weighted by .
Strategy-2: :
Compared with Strategy-1, we effectively lose one degree of freedom for getting a better . However, this is not entirely harmful since and have the same feature set, indicating estimating is no longer necessary and is an optimal solution for (6) with this strategy.
Strategy selection
When we can get a high-quality by fine-tuning and or is already provided from other sources, we may solve the optimization problem in (6) to find the optimal weight . However, considering the fact that estimating introduces extra computation and potentially extra errors, we focus on Strategy-2 in this paper. Using Strategy-2 also reduces the effort on tuning hyperparameters. Besides, the proposed CAL loss is tolerant of an imperfect (shown theoretically in Section 4.3).
D.2 Generation of Instance-Dependent Label Noise
Pseudo codes for generate instance-based label noise are provided in Algorithm 2. This algorithm follows the state-of-the-art method [40]. Define the overall noise rate as .
Note Algorithm 2 cannot ensure when . To generate an informative dataset, we set as the upper bound of and distribute the remaining probability to other classes.
D.3 Performance without Data Augmentations
For a fair comparison with the recent work on instance-dependent label noise [40], we adopt the same data augmentations as [40] and re-produce their results using the same noise file as we employed in Table 1. Each noise rate is tested times with a different generation matrix (defined in Algorithm 2). Table 4 shows the advantages of our second-order approach.
Method | ||
---|---|---|
PTD-R-V[40] | ||
CAL |
D.4 More Implementation Details on Clothing1M
Construct
We first train the network for 120 epochs on 1 million noisy training images using the method in [5]. The batch-size is set to 32. The initial learning rate is set as 0.01 and reduced by a factor of 10 at 30, 60, 90 epochs. We sample 1000 mini-batches from the training data for each epoch while ensuring the (noisy) labels are balanced. Mixup [48] is adopted for data augmentations. Hyperparameter is set to 0 at first 80 epochs, and linearly increased to 0.4 for next 20 epochs and kept as 0.4 for the rest of the epochs. We construct with the best model.
Train with CAL
We change the loss to the CAL loss after getting and continue training the model (without mixup) with an initial learning rate of for 120 epochs (reduced by a factor of 10 at 30, 60, 90 epochs). We also tested re-train the model with and get an accuracy of . A randomly-collected balanced dataset with noisy examples in each class is employed in training with CAL. Examples that are not in this balanced dataset are removed from for ease of implementation.