22email: {2030809, zhaocairong}@tongji.edu.cn
33institutetext: Key Laboratory of Image Processing and Intelligent Control, Ministry of Education,
School of Artificial Intelligence and Automation,
Huazhong University of Science and Technology, China
33email: {cunlin_wu,Yang_Xiao}@hust.edu.cn
Learning from Noisy Labels with Coarse-to-Fine Sample Credibility Modeling
Abstract
Training deep neural network (DNN) with noisy labels is practically challenging since inaccurate labels severely degrade the generalization ability of DNN. Previous efforts tend to handle part or full data in a unified denoising flow via identifying noisy data with a coarse small-loss criterion to mitigate the interference from noisy labels, ignoring the fact that the difficulties of noisy samples are different, thus a rigid and unified data selection pipeline cannot tackle this problem well . In this paper, we first propose a coarse-to-fine robust learning method called CREMA, to handle noisy data in a divide-and-conquer manner. In coarse-level, clean and noisy sets are firstly separated in terms of credibility in a statistical sense. Since it is practically impossible to categorize all noisy samples correctly, we further process them in a fine-grained manner via modeling the credibility of each sample. Specifically, for the clean set, we deliberately design a memory-based modulation scheme to dynamically adjust the contribution of each sample in terms of its historical credibility sequence during training, thus alleviating the effect from noisy samples incorrectly grouped into the clean set. Meanwhile, for samples categorized into the noisy set, a selective label update strategy is proposed to correct noisy labels while mitigating the problem of correction error. Extensive experiments are conducted on benchmarks of different modality, including image classification (CIFAR, Clothing1M etc) and text recognition (IMDB), with either synthetic or natural semantic noises, demonstrating the superiority and generality of CREMA.
Keywords:
robust learning, label noise, divide-and-conquer†Yabiao Wang is corresponding author ([email protected]).

(a)

(b)

(c)
1 Introduction
Deep learning has achieved significant progress in the recognition of multimedia signals (e.g. images, text, speeches). The key to its success is the availability of large-scale datasets with reliable manual annotations. Collecting such datasets, however, is time-consuming and expensive. Some alternative ways to obtain labeled data, such as web crawling [46], inevitably yield samples with noisy labels, which are not appropriate to be directly utilized to train DNN since these complex models can easily over-fitting (i.e., memorizing) noisy labels [2, 50].
To handle this problem, classical Learning with Noisy Label (LNL) approaches focus on either identifying and dropping noisy samples (i.e., sample selection) [10, 14, 49, 45] or adjusting the objective term of each sample during training (i.e., loss adjustment) [29, 48, 37]. The former usually make use of small-loss trick to select clean samples, and then take them to update DNNs. However, the procedure of sample selection cannot guarantee that the selected clean samples are completely clean. In contrast, as indicated in Fig. 1, division relied on statistic metrics can still involve some hard noisy samples in the training set, which will be treated equally as other normal samples in the following training stages. Thus the negative impact brought by wrongly grouped noisy samples can still confuse the optimization process and lower the test performance of DNNs [49]. On the other hand, the latter schemes reweight loss values or update labels by estimating the confidence on how clean a sample is. Typical methods include loss correction via an estimated noise transition matrix [29, 8, 12]. However, estimating an accurate noise transition matrix is practically challenging. Recently, there are approaches directly correcting the labels of all training samples [37, 48]. However, we empirically find that unconstrained label correction in full data can do harm to clean samples and reversely hinder the model performance.
Towards the problems above, we propose a simple but effective method called CREMA (Coarse-to-fine sample cREdibility Modeling and Adaptive loss reweighting), which adaptively reduces the impact of noisy labels via modeling the credibility (i.e., quality) of each sample. In the coarse-level, with the estimated sample credibility by simple statistic metrics, clean and noisy samples can be roughly separated and handled in a divide-and-conquer manner. Since it is practically impossible to separate these samples perfectly, for the selected clean samples, we take their historical credibility sequences to adjust the contribution of each sample to their objective, thus mitigating the negative impact of hard noisy samples (i.e, noisy samples incorrectly grouped into the clean set) in a fine-grained manner. As for the separated noisy samples, some of them are actually clean (i.e., hard clean samples) and can be helpful for model training. Thus instead of discarding them as previous sample selection methods [10, 45], we make use of them via a selective label correction scheme.
The insight behind CREMA is from the observation on the loss value during training on noisy data (illustrated in Fig 1), it can be found that clean and noisy samples manifest distinctive statistical properties during training, where clean samples yield relatively smaller loss value [33] and more consistent prediction. Hence these statistical features can be utilized to coarsely model the sample credibility. However, Fig 1 also shows that the full data can not be perfectly separated by simple statistical metrics. This inspires us to adaptively cope with noises of different difficulty levels with more fine-grained design. For easily recognized noisy samples, we can directly apply certain label correction schemes while avoiding erroneous correction on normal samples. For samples that fall into the confusing area and hybrid with clean ones, since the coarsely estimated credibility in the current epoch is not informative enough to identify noisy samples, CREMA applies a fine-grained likelihood estimator of noisy samples by resorting the historical sequence of sample credibility. This is achieved by maintaining a historical memory bank along with the training process and estimating the likelihood function through a consistency metric and assumption of markov property of the sequence.
CREMA is built upon a classic co-training framework [45, 10]. The fine-grained sample credibility estimated by one network is used to adjust the loss term of credible samples for the other network. Extensive experiments are conducted on benchmarks of different modality, including image classification (CIFAR, MNIST, Clothing1M etc) and text recognition (IMDB), with either synthetic or natural semantic noises, demonstrating the superiority and generality of the proposed method. In a nutshell, the key contributions of this paper include:
CREMA: a novel LNL algorithm that combats noisy labels via coarse-to-fine sample credibility modeling. In coarse-level, clean and noisy sets are roughly separated and handled respectively, in the spirit of the idea of divide-and-conquer. Easily recognized noisy samples are handled via a selective label update strategy;
In CREMA, likelihood estimation of historical credibility sequence is proposed to help identify hard noisy samples, which naturally plays as the dynamical weight to modulate loss term of each training sample in a fine-grained manner;
CREMA is evaluated on six synthetic and real-world noisy datasets with different modality, noise type and strength. Extensive ablation studies and qualitative analysis are provided to verify the effectiveness of each component.

2 Related Works
The existing LNL approaches can be mainly categorized into three groups: loss adjustment, label correction, and noisy sample detection. Next, we will introduce and discuss existing works for training DNN with noisy labels.
Loss Adjustment. Adjusting the loss values of all training samples is able to reduce the negative impact of noisy labels. To do this, many approaches seek to robust loss functions, such as Robust MAE [7], generalized cross entropy [55], symmetric cross entropy [44], Improved MAE [42] and curriculum loss [23]. Rather than treat all samples equally, some methods rectify the loss of each sample through estimating the label transition matrix [12, 29, 8, 9, 46, 19, 52] or imposing different importance on each sample to formulate a weighted training procedure [40, 22, 4, 43]. The noise transition matrix, however, is relatively hard to be estimated and many approaches [12, 39, 21, 14, 20, 6, 5, 34, 35, 56] often make assumptions that a small clean-labeled dataset exists. In real-world scenarios, such condition is not always fulfilled, thus limiting the applications of these approaches.
Label correction. Label correction methods seek to refurbish the ground-truth of noisy samples, thus preventing DNN overfits to false labels. The most common ways to obtain the updated label include bootstrapping (i.e., a convex combination of the noisy label and the DNN prediction) [33, 11, 1, 41] and label replacing [37, 48, 36, 54]. One critical problem of label correction methods is to define the confidence of each label being clean, that is, samples with high clean probability should keep their labels almost unchanged, and vice versa. Previous solutions including cross-validation [33], fitting a two-component mixture model [1], local intrinsic dimensionality measurement [13, 24] and leveraging the prediction consistency of DNN models [36, 41]. However, updating the labels of all training sets is challenging, and well-designed regularization terms are important to prevent DNN from falling into trivial solutions [37, 48].
Noisy sample detection. One common knowledge used to discover noisy samples is the memorization effects (i.e., DNN fits clean samples first and then noisy ones). As a result, after a warm-up training stage with all noisy samples, DNN is able to identify the clean samples by taking the small-loss ones. The small-loss trick is exploited by many sample selection methods [10, 49, 26, 14, 45, 28, 36, 47]. After separating the noisy samples from the clean ones, Co-teaching [10] and corresponding variants [10, 49, 45, 47] update two parallel network parameters with the clean samples and abandoned the noisy ones. The idea of training two deep networks simultaneously is effective to avoid the confirmation bias problem (i.e., a model would accumulate its error through the self-training process) [10, 14, 16]. Other noisy measurement metrics such as Area Under the Margin (AUM) [32] is also proposed to better distinguish and remove noisy samples, which hypothesizes that noisy samples will in expectation have smaller AUM than clean samples. However, discarding the noisy samples means that valuable data may be lost, which leads to slow convergence of DNN models [4]. Instead, there are methods that utilize both clean and noisy samples to formulate a semi-supervised learning problem, by discarding only the labels of identified noisy samples. Thus naturally converting LNL problem into a semi-supervised learning one, for which powerful semi-supervised learning methods can be leveraged to boost performance [16, 57, 51, 3].
Hybrid. There are also researches taking two or more techniques above into account to boost performance of robust learning. For example, RoCL [58] and SELFIE [36] propose to dynamically discover informative (refurbished) samples and correct their labels with model predictions. Accordingly, CREMA belongs to hybrid group and it differs from existing methods in (1) it adaptively cope with noises of different difficulty levels via estimating the sample credibility in a coarse-to-fine fashion, easy and hard noisy samples are handled in a divide-and-conquer strategy; (2) it estimates likelihood of historical sample credibility sequence to dynamically modulate loss term of hard noisy samples; (3) it explores a selective label correction scheme to deal with hard clean samples while mitigating the correction error; (4) it is end-to-end trainable and does not require extra computation or any modification to the model.

(a) Global label learning.

(b) Selective label learning in CREMA.
3 Method
In this section, we introduce CREMA, an end-to-end approach for LNL problem. The technical pipeline of the approach is shown in Fig. 2. The training process is built upon a classic co-training framework [10, 45] to avoid confirmation bias and separate credible samples (mostly clean) and noisy samples (mostly noisy) via per-sample loss value in the coarse stage. The separated samples are handled by the succeeding fine-grained processes in a divide-and-conquer manner.
3.1 Coarse level separation
Formally, for multi-class classification with noisy label problem, let denote the training data, where is data sample and is the one-hot label over classes. denotes sample feature extracted by DNN model. With the loss from the DNN model, clean set , and noisy set are separated via the widely used low-loss criterion [10, 45],
(1) | ||||
where is the threshold, determined by a dynamic memory rate . Which is set for DNN to gradually distinguish data with highest loss value as noisy samples while keeping other samples as the clean set. The loss value simply serves as credibility of each sample in a coarse level. However, as illustrated in Fig. 1, this simple separation criterion can not strictly eliminate noisy samples. Hence we choose to handle them respectively, where is exploited to update DNN parameters via fine-grained sample credibility guided loss adjustment (Sec. 3.2), and is leveraged via another label learning scheme (Sec. 3.3).
3.2 Fine-grained sequential credibility modeling
Sequential credibility analysis. Previous works prefer to assess the data reliability purely based on its statistical property on a single point of time (e.g. the loss value in current epoch) during training process, i.e. they regard the credibility of the -th data sample proportional to the joint distribution or likelihood of its data-label pair,
(2) |
However, as shown in Fig. 1, the training curve of normal and noisy samples usually yield different statistic information, where noisy samples usually have relatively larger loss values and poorer prediction consistency compared with clean ones, therefore the historical record of data training is also informative enough to help distinguish noisy and clean data.
This observation inspires us to estimate the data credibility in a sequential manner. To be specific, we define a sequence with length as:
(3) |
Eq (3) illustrates a sliding window covering the feature snapshot of data from previous epochs to current time point, where denotes the model parameters at the -th epoch, and we model the data credibility with the likelihood and consistency of these historical sequences,
(4) |
The Eq (4) can be decoupled into two items, where measures the stability of training sequence given its label , while denotes log-likelihood of sequence generated from the -th data observation of neural network training process. To estimate the sequential log-likelihood, we further assume that the observation in sequence conforms to a certain markov property as:
(5) |
The assumption in Eq (5) is reasonable since in most iterative learning algorithm like SGD, the data feature distribution is only decided by the last observation and its label. With this assumption, we can further derive the likelihood as:
(6) | ||||
With Eq (6), we can represent the sequential likelihood as the summation of the conditional likelihood of data observation at each adjacent epochs within a sliding window of length . In the implementation, we can apply a normalized mixture model like GMM [16] or BMM [1] as estimator to estimate the conditional probability in Eq (6) via modeling the sample-wise loss value distribution. Meanwhile, with the conditional probability estimation, the stability measurement is further designed as a modulator to suppress loss on training sequence with intense fluctuation,
(7) |
(8) |
Adaptively loss adjustment. The sequential likelihood and stability measurement reflects how confident of the sample being clean. With the estimated credibility we reweight loss to update DNN as:
(9) |
Where is the objective function. is the sample credibility and it modulates the contribution of each sample through gradient descending algorithm. Note that Eq (9) is only applied on clean set , in this way, the negative impact of hard noisy samples within can be mitigated.
Objective function. Inspired by the design of symmetric cross entropy (SCE) function [44], a symmetric JS-divergence function with a co-regularization term is leveraged in CREMA as:
(10) | ||||
Where is the softmax function, and are features extracted by two models. The reason we choose JS-divergence [27] instead of cross entropy (CE) as loss function is that CE tends to over-fit noisy samples as these samples contribute relatively large gradient values during training. While JS-divergence mitigates this problem via using predictions of the current model as supervising signals as well. Since for noisy samples, DNN predictions are usually more reliable than its label. Following previous work [37], a prior label distribution term and a negative entropy term are included to regularize training and further alleviate the over-fitting problem.
3.3 Selective label distribution learning
Following the divide-and-conquer idea, we attempt to leverage the separated noisy samples as well. Some hard clean samples are blended with the separated noisy ones. Thus instead of discarding these wrongly labeled data as in most sample detection methods [10, 45], we resort to label correction approaches [37, 48] to exploit them with gradually corrected labels and further boost performance.
Specifically, labels of are treated as extra parameters and updated through back-propagation to optimize a certain objective, this means both the network parameters and labels are updated simultaneously during the training process, where original one-hot labels will turn into a soft label distribution after updating. Formally, is updated as . Where is learning rate and is the objective to supervise the label correction process as .
Empirical insight. In our experiments. we find that global label learning strategy (i.e., correcting labels of all training data) suffers from correction error in clean data. This can be observed from Fig. 3 (a), large gradient value will also be imposed on lots of correctly-labeled samples. Consequently, labels for these clean samples are unnecessarily updated. Compared with global label correction manners, we choose to only update the separated noisy samples (mostly noisy). As shown in Fig. 3 (b), the proposed selective label correction strategy focuses more on learning noisy labels. The number of correctly-labeled samples with large gradient value is way less than a global correction scheme. Indicating that the selective label correction scheme can mitigate the problem of correction error. Experiments in Sec 4.3 also quantitatively verify the effectiveness of the selective label learning strategy over the global label learning manner.
Putting this all together, Algorithm 1 delineates the proposed CREMA in detail. In a nutshell, CREMA is built on a divide-and-conquer framework. Firstly, clean set and noisy set are separated based on the low-loss criterion [10]. For , we compute the likelihood of historical credibility sequence, which helps to adaptively modulate the loss term of each training sample. As for , a selective label correction scheme is leveraged to update label distribution and model parameters simultaneously. After each training epoch, memory sequence and are updated with the feature snapshot of the most current epoch.
4 Experiments
4.1 Datasets and Implementation Details
Datasets. To validate the effectiveness of the proposed method, we experimentally investigate on four synthetic noisy datasets, i.e., IMDB [25], MNIST, CIFAR-10, CIFAR-100 [15] and two real-world label noise datasets, i.e., Clothing1M [46], and Animal10N [36]. IMDB [25] is a collection of highly polarized movie reviews (positive/negative). It consists of 25,000 training samples and 25,000 samples for testing. The task is formalized as a binary classification task to decide the polarity of sentiment for a review. MNIST consists of 70,000 images of size for 10 classes, in which 60,000 images for training and the left 10,000 images for testing. Both CIFAR-10 and CIFAR100 contain 50,000 training images and 10,000 testing images of size . Differently, the former has 10 classes, while CIFAR-100 has 100 classes. For the Clothing1M, it is a large-scale real-world noisy dataset which is collected from multiple online shopping websites. It contains 1 million training images and clean training subsets (47K for training, 14K for validation and 10K for test) with 14 classes. The noise rate for this dataset is around 38.5%. Animal-10N contains 55,000 human-labeled online images for 10 confusing animals. It includes approximately 8% noisy-labeled samples. Following the settings in previous works [36, 54], 50,000 images are exploited as a training set while the left for testing.
Implementation Details. For the three synthetic image classification noisy datasets, MNIST, CIFAR-10 and CIFAR-100, we follow the setting in previous works [10, 49, 45], experiments with three kinds of noise types are considered, i.e., symmetric noise (uniformly random), asymmetric noise, and pairflip noise. For the IMDB text classification dataset, we tokenize each sentence and the word embeddings have dimension 10,000. Symmetric noises with different noise level are tested. Specifically, symmetric noise is generated by replacing labels in each class with labels of other classes uniformly. Asymmetric noise simulates fine-grained classification (for example, lynx and cat in Animal-10N [36] with noisy labels, where labels are corrupted to a set of similar classes. Pairflip noise is generated by flipping each class to its adjacent class. Varying noise rates are conducted to fully evaluate the proposed method, where for symmetric label noise, we set on image datasets, on text dataset, for asymmetric noise and for pairflip label noise. For real-world noisy Clothing1M dataset, following [48, 54], we do not use the 50K clean data, and a randomly sampled pseudo-balanced subset includes about 260K images is leveraged as training data.
For the network structure, a 2-layer bi-directional LSTM network is adopted for IMDB. It is of 128 embedding size and 128 hidden size. A 9-layer CNN with Leaky-ReLU activation function [10] is used for MNIST, CIFAR-10, and CIFAR-100, while ResNet-50 is adopted for Clothing1M and Animal-10N datasets. The batch size is set as 64 for all the datasets. For fair comparisons, we train our model for 200 epochs in total and choose the average test accuracy of last 10 epochs as the final result in three image synthetic noisy datasets. For IMDB dataset, we set total training epochs as 100 and also test the accuracy of last ten epochs. Total training epochs for Clothing1M and Animal-10N are 80 and 150 respectively. Additionally, all the methods are implemented in PyTorch and run on NVIDIA Tesla V100 GPUs. Moreover, we use Adam optimizer for all the experiments and set the initial learning rate as 0.001, then it is degraded by a factor of every epochs for Clothing1M and epochs for Animal-10N. The two classifiers in our methods are two networks with the same structure but different initialization parameters. Following [10], is linearly decreased along with training until reach a lower bound value , for Clothing1M and Animal-10N datasets, we empirically set lower bound as 0.8 and 0.92 respectively.
Noise rates | Standard | PENCIL | Co-teaching | Co-teaching+ | JoCoR | CREMA (ours) |
---|---|---|---|---|---|---|
Symmetry-20% | ||||||
Symmetry-50% | ||||||
Symmetry-80% | ||||||
Asymmetry-40% | ||||||
Pairflip-40% | ||||||
Pairflip-45% |
Noise rates | Standard | PENCIL | Co-teaching | Co-teaching+ | JoCoR | CREMA (ours) |
---|---|---|---|---|---|---|
Symmetry-20% | ||||||
Symmetry-50% | ||||||
Symmetry-80% | ||||||
Asymmetric-40% | ||||||
Pairflip-40% | ||||||
Pairflip-45% |
Noise rates | Standard | PENCIL | Co-teaching | Co-teaching+ | JoCoR | CREMA (ours) |
---|---|---|---|---|---|---|
Symmetry-20% | ||||||
Symmetry-50% | ||||||
Symmetry-80% | ||||||
Asymmetric-40% | ||||||
Pairflip-40% | ||||||
Pairflip-45% |
Method | Category | Accuracy | ||
---|---|---|---|---|
LA | LC | ND | ||
Cross-Entropy | 69.21 | |||
GCE [55] | ✓ | 69.75 | ||
IMAE [42] | ✓ | 73.20 | ||
SCE [44] | ✓ | 71.02 | ||
DM [43] | ✓ | 73.30 | ||
F-correction [29] | ✓ | 69.84 | ||
M-correction [1] | ✓ | 71.00 | ||
Masking [9] | ✓ | 71.10 | ||
Joint-Optim [37] | ✓ | 72.23 | ||
Meta-Cleaner [53] | ✓ | 72.50 | ||
Meta-Learning [17] | ✓ | 73.47 | ||
PENCIL [48] | ✓ | 73.49 | ||
PLC [54] | ✓ | 74.02 | ||
Self-Learning [11] | ✓ | 74.45 | ||
ProSelfLC [41] | ✓ | 73.40 | ||
Co-teaching [10] | ✓ | 70.15 | ||
JoCoR [45] | ✓ | 70.30 | ||
C2D [57] | ✓ | 74.30 | ||
DivideMix† [16] | ✓ | 74.48 | ||
CREMA (Ours) | ✓ | ✓ | ✓ | 74.53 |
4.2 Comparison with state-of-the-art methods
Results on synthetic noisy datasets. Table 1, Table 2, Table 3 and Table 6 show the detailed results of the proposed CREMA and other methods in multiple synthetic noisy cases on four widely used datasets, i.e., MNIST, CIFAR-10, CIFAR-100 IMDB. Specifically, four state-of-the-art LNL methods that are highly related to our work are chosen for comparison: PENCIL [48], Co-teaching [10], Co-teaching+ [49], JoCoR [45]. Standard DNN training with cross entropy is also included as baseline. All the results of these methods are reproduced with their public code and suggested hyper-parameters for fair comparison. From these tables, we can observe that most of these methods show better performance than Standard in the most natural Symmetry-20% case, except for PENCIL on IMDB dataset, which verifies their robustness. Among them, JoCoR performs much better over other compared methods. However, when it comes to Pairflip-40% and Pairflip-45% noisy cases on image datasets, their performance drops significantly. On the contrary, the proposed CREMA can achieve consistent improvements over other methods on four benchmarks across various noise settings. In the Pairflip-40% and Pairflip-45% cases, the proposed method outperforms other baselines by a large margin. Specifically, CREMA can achieve 16.44% and 25.26% improvement in accuracy over JoCoR on CIFAR-10. When dealing with extremely noisy scenario, e.g. Symmetry-80%, CREMA can also perform generally better than other compared methods.The result demonstrates the superiority and generality of the proposed robust learning method across various types and levels of label noise on multimedia (i.e., image and text) datasets.
Results on real-world noisy datasets. Experiments on real-world noisy datasets Clothing1M [46], Animal-10N [36] are also conducted to verify the effectiveness of the proposed method. The baseline methods are chosen from recently proposed LNL methods. Specifically, several loss adjustment methods, including GCE [55], IMAE [42], SCE [44], DM [43], F-correction [29], M-correction [1], Masking [9], ActiveBias [4], label correction methods, including Joint-Optim [37], PENCIL [48], Self-Learning [11], PLC [54], ProSelfLC [41] noisy sample detection approaches, including Co-teaching [10], JoCoR [45], C2D [57], DivideMix [16] and a hybrid method SELFIE [36] are compared with the proposed method. Table 6 and Table 6 show results on two real-world noisy datasets respectively. On the large-scale Clothing1M dataset, CREMA outperforms all compared methods. Note that CREMA follows the standard DNN training procedure, and is similar to other co-training methods [10, 49, 45] in terms of training time since the time cost for sample credibility modeling is negligible compared with DNN update. It is worth noting that the proposed method outperforms these co-teaching methods by a large margin. Indicating that the discarded samples by co-teaching methods are actually valuable, and CREMA well utilized all training samples. The best test accuracy is achieved by CREMA among the compared methods in Animal-10N as well. The results indicate that the proposed method can work well on high noise level (i.e., Clothing1M) and fine-grained (i.e., Animal-10N) real-world noisy datasets.
Method | Test Accuracy (%) |
---|---|
Baseline | 72.81 |
+ Selective label update | 73.25 |
+ Sequential likelihood | 74.00 |
+ Stability measurement | 74.53 |
Estimator | Test Accuracy (%) |
---|---|
BMM | 74.09 |
GMM | 74.53 |
Length of sequence | 1 | 2 | 3 | 4 | 5 | 6 |
---|---|---|---|---|---|---|
Test Accuracy (%) | 73.25 | 73.99 | 74.53 | 74.40 | 74.27 | 73.96 |
4.3 Ablation studies
Component Analysis. CREMA contains several important components, including selective label learning strategy, sequential likelihood and stability measurement . To verify the effectiveness of each component, we conduct experiments on large-scale noisy dataset Clothing1M. The baseline method is built upon a simple co-teaching framework [45] combined with global label correction schemes (as in [48]), without the credibility guided loss adjustment strategy. The results are shown in Table 8, we can see that, conform to the observation in Fig. 3, the proposed selective label learning strategy achieves better results compared with the global correction counterpart. The sequential likelihood and stability measurement further boost the model performance with 0.75% and 0.53% accuracy gain, this indicates that the proposed sequential sample credibility modeling can effectively combat hard noisy samples mixed with clean ones. With all the three key components above, CREMA can achieve 74.53% test accuracy on Clothing1M.
Length of sequence . We also conduct experiments to investigate how the length of sequence affects the performance. Fig 4(a) shows results on Clothing1M with various values (in {1, 2, 3, 4, 5, 6}) of . It can be observed that increasing the length of sequence helps achieve higher accuracy at first but turn poor after hitting the peak value. Intuitively, when no temporal information is provided when , CREMA can not utilize consistency metric to identify hard noisy samples that blended with clean ones, thus leading to an inferior result. When is larger than 4, we also notice that performance degrades, this is probably due to unreliable model inside the very long sequence can harm sample credibility modeling and reversely hinder the final result.
Effect of different estimators. The probabilistic model plays the role of estimating the conditional probability in Eq (6). We compare two different estimators, Gaussian Mixture Model (GMM) [31] and Beta Mixture Model (BMM) [1] on Clothing1M. Table 8 shows the results. We can see that GMM obtains a relatively higher test accuracy, but BMM can also achieve good results (74.09%) as well. This indicates that the choice of normalized mixture model is not sensitive to the final result.
Reliability of the estimated sample credibility. Sample credibility plays the role of dynamical weight to modulate the loss term of each training sample, as in Eq (9). In Fig. 4 we visualize the empirical PDF of the learned credibility weight of all training samples between (b) the sequential estimation manner within CREMA and (a) its non-sequential counterpart (i.e., ) under two different noisy settings. It can be observed that the overall credibility weight of training samples is distinguishable for clean and noisy data in (b). Specifically, clean samples possess larger weight, thus contributing more gradients during the process of DNN training. Noisy samples are assigned with a relatively small weight to alleviate their negative impact. However, the non-sequential weights yield significantly more samples that cannot be correctly-separated via a fixed threshold. Indicating that the proposed fine-grained sequential credibility estimation is more effective for reliable sample weight modeling.

5 Conclusion
In this paper, we propose a novel end-to-end robust learning method, called CREMA. Towards the problem that previous works lack the consideration of intrinsic difference among difficulties of noisy samples. We follow the idea of divide-and-conquer that separates clean and noisy samples via estimating the credibility of each training sample. Two branches are designed to handle the imperfectly separated sample sets respectively. For easily recognizable noisy samples, we apply a selective label correction scheme avoiding erroneous label updates on clean samples. For hard noisy samples blended with clean ones, likelihood estimation of historical credibility sequence adaptively modulates the loss term of each sample during training. Extensive experiments conducted on several synthetic and real-world noisy datasets verify the superiority of the proposed method.
References
- [1] Arazo, E., Ortego, D., Albert, P., O’Connor, N.E., McGuinness, K.: Unsupervised label noise modeling and loss correction. In: Proc. International Conference on Machine Learning (ICML) (2019)
- [2] Arpit, D., Jastrzkebski, S., Ballas, N., Krueger, D., Bengio, E., Kanwal, M.S., Maharaj, T., Fischer, A., Courville, A., Bengio, Y., et al.: A closer look at memorization in deep networks. In: Proc. International Conference on Machine Learning (ICML). pp. 233–242 (2017)
- [3] Berthelot, D., Carlini, N., Goodfellow, I.J., Papernot, N., Oliver, A., Raffel, C.: Mixmatch: A holistic approach to semi-supervised learning. Proc. Advances in Neural Information Processing Systems (NeurIPS) (2019)
- [4] Chang, H.S., Learned-Miller, E., McCallum, A.: Active Bias: Training more accurate neural networks by emphasizing high variance samples. In: Proc. Advances in Neural Information Processing Systems (NeurIPS). pp. 1002–1012 (2017)
- [5] Dehghani, M., Severyn, A., Rothe, S., Kamps, J.: Avoiding your teacher’s mistakes: Training neural networks with controlled weak supervision. arXiv preprint arXiv:1711.00313 (2017)
- [6] Dehghani, M., Severyn, A., Rothe, S., Kamps, J.: Learning to learn from weak supervision by full supervision. In: Proc. Advances in Neural Information Processing Systems Workshop (NeurIPSW) (2017)
- [7] Ghosh, A., Kumar, H., Sastry, P.: Robust loss functions under label noise for deep neural networks. In: Proc. Association for the Advancement of Artificial Intelligence (AAAI) (2017)
- [8] Goldberger, J., Ben-Reuven, E.: Training deep neural-networks using a noise adaptation layer. In: Proc. International Conference on Learning Representations (ICLR) (2017)
- [9] Han, B., Yao, J., Niu, G., Zhou, M., Tsang, I., Zhang, Y., Sugiyama, M.: Masking: A new perspective of noisy supervision. In: Proc. Advances in Neural Information Processing Systems (NeurIPS). pp. 5836–5846 (2018)
- [10] Han, B., Yao, Q., Yu, X., Niu, G., Xu, M., Hu, W., Tsang, I., Sugiyama, M.: Co-teaching: Robust training of deep neural networks with extremely noisy labels. In: Proc. Advances in Neural Information Processing Systems (NeurIPS). pp. 8527–8537 (2018)
- [11] Han, J., Luo, P., Wang, X.: Deep self-learning from noisy labels. In: Proc. IEEE International Conference on Computer Vision (ICCV). pp. 5138–5147 (2019)
- [12] Hendrycks, D., Mazeika, M., Wilson, D., Gimpel, K.: Using trusted data to train deep networks on labels corrupted by severe noise. In: Proc. Advances in Neural Information Processing Systems (NeurIPS). pp. 10456–10465 (2018)
- [13] Houle, M.E.: Local intrinsic dimensionality I: An extreme-value-theoretic foundation for similarity applications. In: Proc. International Conference on Similarity Search and Applications (SISAP). pp. 64–79 (2017)
- [14] Jiang, L., Zhou, Z., Leung, T., Li, L.J., Fei-Fei, L.: MentorNet: Learning data-driven curriculum for very deep neural networks on corrupted labels. In: Proc. International Conference on Machine Learning (ICML) (2018)
- [15] Krizhevsky, A., Hinton, G., et al.: Learning multiple layers of features from tiny images (2009)
- [16] Li, J., Socher, R., Hoi, S.C.: Dividemix: Learning with noisy labels as semi-supervised learning. arXiv preprint arXiv:2002.07394 (2020)
- [17] Li, J., Wong, Y., Zhao, Q., Kankanhalli, M.: Learning to learn from noisy labeled data. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR). pp. 5046–5054 (2019)
- [18] Li, J., Xiong, C., Hoi, S.C.: Learning from noisy data with robust representation learning. In: Proc. IEEE International Conference on Computer Vision (ICCV). pp. 9485–9494 (2021)
- [19] Li, X., Liu, T., Han, B., Niu, G., Sugiyama, M.: Provably end-to-end label-noise learning without anchor points. Proc. International Conference on Machine Learning (ICML) (2021)
- [20] Li, Y., Yang, J., Song, Y., Cao, L., Luo, J., Li, L.J.: Learning from noisy labels with distillation. In: Proc. IEEE International Conference on Computer Vision (ICCV). pp. 1910–1918 (2017)
- [21] Litany, O., Freedman, D.: Soseleto: A unified approach to transfer learning and training with noisy labels. arXiv preprint arXiv:1805.09622 (2018)
- [22] Liu, T., Tao, D.: Classification with noisy labels by importance reweighting. IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI) 38(3), 447–461 (2015)
- [23] Lyu, Y., Tsang, I.W.: Curriculum loss: Robust learning and generalization against label corruption. In: Proc. International Conference on Learning Representations (ICLR) (2020)
- [24] Ma, X., Wang, Y., Houle, M.E., Zhou, S., Erfani, S.M., Xia, S.T., Wijewickrema, S., Bailey, J.: Dimensionality-driven learning with noisy labels. In: Proc. International Conference on Machine Learning (ICML) (2018)
- [25] Maas, A., Daly, R.E., Pham, P.T., Huang, D., Ng, A.Y., Potts, C.: Learning word vectors for sentiment analysis. In: Proceedings of the 49th annual meeting of the association for computational linguistics: Human language technologies. pp. 142–150 (2011)
- [26] Malach, E., Shalev-Shwartz, S.: Decoupling” when to update” from” how to update”. In: Proc. Advances in Neural Information Processing Systems (NeurIPS). pp. 960–970 (2017)
- [27] Manning, C., Schutze, H.: Foundations of statistical natural language processing. MIT press (1999)
- [28] Nguyen, D.T., Mummadi, C.K., Ngo, T.P.N., Nguyen, T.H.P., Beggel, L., Brox, T.: SELF: Learning to filter noisy labels with self-ensembling. In: Proc. International Conference on Learning Representations (ICLR) (2020)
- [29] Patrini, G., Rozza, A., Krishna Menon, A., Nock, R., Qu, L.: Making deep neural networks robust to label noise: A loss correction approach. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR). pp. 1944–1952 (2017)
- [30] Pereyra, G., Tucker, G., Chorowski, J., Kaiser, L., Hinton, G.E.: Regularizing neural networks by penalizing confident output distributions. ArXiv abs/1701.06548 (2017)
- [31] Permuter, H., Francos, J.M., Jermyn, I.: A study of gaussian mixture models of color and texture features for image classification and segmentation. Pattern Recognition (PR). 39, 695–706 (2006)
- [32] Pleiss, G., Zhang, T., Elenberg, E.R., Weinberger, K.Q.: Identifying mislabeled data using the area under the margin ranking. Proc. Advances in Neural Information Processing Systems (NeurIPS) (2020)
- [33] Reed, S., Lee, H., Anguelov, D., Szegedy, C., Erhan, D., Rabinovich, A.: Training deep neural networks on noisy labels with bootstrapping. In: Proc. International Conference on Learning Representations (ICLR) (2015)
- [34] Ren, M., Zeng, W., Yang, B., Urtasun, R.: Learning to reweight examples for robust deep learning. In: Proc. International Conference on Machine Learning (ICML) (2018)
- [35] Shu, J., Xie, Q., Yi, L., Zhao, Q., Zhou, S., Xu, Z., Meng, D.: Meta-Weight-Net: Learning an explicit mapping for sample weighting. In: Proc. Advances in Neural Information Processing Systems (NeurIPS). pp. 1917–1928 (2019)
- [36] Song, H., Kim, M., Lee, J.G.: SELFIE: Refurbishing unclean samples for robust deep learning. In: Proc. International Conference on Machine Learning (ICML). pp. 5907–5915 (2019)
- [37] Tanaka, D., Ikami, D., Yamasaki, T., Aizawa, K.: Joint optimization framework for learning with noisy labels. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR). pp. 5552–5560 (2018)
- [38] Tarvainen, A., Valpola, H.: Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. In: Proc. Advances in Neural Information Processing Systems (NeurIPS) (2017)
- [39] Veit, A., Alldrin, N., Chechik, G., Krasin, I., Gupta, A., Belongie, S.: Learning from noisy large-scale datasets with minimal supervision. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (July 2017)
- [40] Wang, R., Liu, T., Tao, D.: Multiclass learning with partially corrupted labels. IEEE Transactions on Neural Networks and Learning Systems 29(6), 2568–2580 (2017)
- [41] Wang, X., Hua, Y., Kodirov, E., Clifton, D.A., Robertson, N.M.: Proselflc: Progressive self label correction for training robust deep neural networks. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR). pp. 752–761 (2021)
- [42] Wang, X., Hua, Y., Kodirov, E., Robertson, N.M.: Imae for noise-robust learning: Mean absolute error does not treat examples equally and gradient magnitude’s variance matters. arXiv preprint arXiv:1903.12141 (2019)
- [43] Wang, X., Kodirov, E., Hua, Y., Robertson, N.M.: Derivative manipulation for general example weighting. arXiv preprint arXiv:1905.11233 (2019)
- [44] Wang, Y., Ma, X., Chen, Z., Luo, Y., Yi, J., Bailey, J.: Symmetric cross entropy for robust learning with noisy labels. In: Proc. IEEE International Conference on Computer Vision (ICCV). pp. 322–330 (2019)
- [45] Wei, H., Feng, L., Chen, X., An, B.: Combating noisy labels by agreement: A joint training method with co-regularization. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR). pp. 13726–13735 (2020)
- [46] Xiao, T., Xia, T., Yang, Y., Huang, C., Wang, X.: Learning from massive noisy labeled data for image classification. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR). pp. 2691–2699 (2015)
- [47] Yao, Q., Yang, H., Han, B., Niu, G., Kwok, J.: Searching to exploit memorization effect in learning with noisy labels. In: Proc. International Conference on Machine Learning (ICML) (2020)
- [48] Yi, K., Wu, J.: Probabilistic end-to-end noise correction for learning with noisy labels. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR). pp. 7017–7025 (2019)
- [49] Yu, X., Han, B., Yao, J., Niu, G., Tsang, I.W., Sugiyama, M.: How does disagreement help generalization against label corruption? In: Proc. International Conference on Machine Learning (ICML) (2019)
- [50] Zhang, C., Bengio, S., Hardt, M., Recht, B., Vinyals, O.: Understanding deep learning requires rethinking generalization. In: Proc. International Conference on Learning Representations (ICLR) (2017)
- [51] Zhang, H., Cissé, M., Dauphin, Y.N., Lopez-Paz, D.: mixup: Beyond empirical risk minimization. In: Proc. International Conference on Learning Representations (ICLR) (2018)
- [52] Zhang, M., Lee, J., Agarwal, S.: Learning from noisy labels with no change to the training process. In: Proc. International Conference on Machine Learning (ICML). pp. 12468–12478 (2021)
- [53] Zhang, W., Wang, Y., Qiao, Y.: Metacleaner: Learning to hallucinate clean representations for noisy-labeled visual recognition. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR). pp. 7365–7374 (2019)
- [54] Zhang, Y., Zheng, S., Wu, P., Goswami, M., Chen, C.: Learning with feature-dependent label noise: A progressive approach. In: Proc. International Conference on Learning Representations (ICLR) (2021)
- [55] Zhang, Z., Sabuncu, M.: Generalized cross entropy loss for training deep neural networks with noisy labels. In: Proc. Advances in Neural Information Processing Systems (NeurIPS). pp. 8778–8788 (2018)
- [56] Zhang, Z., Zhang, H., Arik, S.O., Lee, H., Pfister, T.: Distilling effective supervision from severe label noise. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR). pp. 9294–9303 (2020)
- [57] Zheltonozhskii, E., Baskin, C., Mendelson, A., Bronstein, A.M., Litany, O.: Contrast to divide: Self-supervised pre-training for learning with noisy labels. arXiv preprint arXiv:2103.13646 (2021)
- [58] Zhou, T., Wang, S., Bilmes, J.: Robust curriculum learning: From clean label detection to noisy label self-correction. In: Proc. International Conference on Learning Representations (ICLR) (2021)