This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Ensemble Learning with Manifold-Based Data Splitting for Noisy Label Correction

Hao-Chiang Shao,  Hsin-Chieh Wang, Weng-Tai Su, and Chia-Wen Lin Manuscript received February 28, 2021; revised ** **, ****. This work was supported in part by Ministry of Science and Technology, Taiwan, under Grant 110-2634-F-007-015-. (Corresponding author: Chia-Wen Lin)H.-C. Shao is with the Department of Statistics and Information Science, Fu Jen Catholic University, New Taipei City, Taiwan. (email: [email protected])H.-C. Wang and W.-T. Su are with the Department of Electrical Engineering, Nation Tsing Hua University, Hsinchu, Taiwan. (email: [email protected]; [email protected] )C.-W. Lin is with the Department of Electrical Engineering and the Institute of Communication Engineering, Nation Tsing Hua University, Hsinchu, Taiwan. (email: [email protected])
Abstract

Label noise in training data can significantly degrade a model’s generalization performance for supervised learning tasks. Here we focus on the problem that noisy labels are primarily mislabeled samples, which tend to be concentrated near decision boundaries, rather than uniformly distributed, and whose features should be equivocal. To address the problem, we propose an ensemble learning method to correct noisy labels by exploiting the local structures of feature manifolds. Different from typical ensemble strategies that increase the prediction diversity among sub-models via certain loss terms, our method trains sub-models on disjoint subsets, each being a union of the nearest-neighbors of randomly selected seed samples on the data manifold. As a result, each sub-model can learn a coarse representation of the data manifold along with a corresponding graph. Moreover, only a limited number of sub-models will be affected by locally-concentrated noisy labels. The constructed graphs are used to suggest a series of label correction candidates, and accordingly, our method derives label correction results by voting down inconsistent suggestions. Our experiments on real-world noisy label datasets demonstrate the superiority of the proposed method over existing state-of-the-arts.

Index Terms:
Ensemble Learning, Label Noise, Data Splitting, Graph Representation, Label Correction.

I Introduction

Learning from noisy data is a vital issue in representation learning for two reasons generally. First of all, as revealed in [1], for a classification model Φ(,θ)\Phi(\cdot,\theta) learned by optimizing its parameter θ\theta via the cross-entropy loss on the predicted label, its expected error is affected by i) the label noise, ii) the selected learning model Φ\Phi, and iii) the model parameter θ\theta learned in the training process. Then, when the learning model Φ\Phi—suitable or not—is determined, the influence caused by both Φ\Phi and θ\theta is fixed. Hence, the only factor that may downgrade the model performance becomes the label noise of the ground-truth label yiy_{i} of each sample xix_{i}. Second, because it is expensive and time-consuming to collect data with reliably clean annotations, in real-world applications people usually need to exploit low-cost open datasets, e.g., those collected by search engines and labeled by quick annotators or crowd-sourcing, for training purposes. These low-cost datasets usually contain a certain amount of noisy labels and thus are not as reliable as professionally-labelled datasets like COCO [2] and ImageNet [3]. Consequently, how to learn from noisy data, including how to correct noisy data, has become a crucial task in real-world applications.

Refer to caption
Figure 1: Illustration of Confusing label noise versus random label noise. Left: t-SNE map of instance features of Clothing1M. Right: t-SNE map of random perturbed CIFAR-10. In real-world applications, noisy labels are locally concentrated near decision borders, as indicated by arrows (b) and (c), and random noises shown in cIFAR-10 simulations are rare in real cases, as indicated by arrow (a).

Label noise can be roughly divided into two types: random label noise and confusing label noise, as illustrated in Fig, 1. The former typically involves mismatched descriptions or tags usually due to the negligence of an annotator. For this type of error, not only a label error may occur randomly in the sample space, but also the erroneous label is often of another irrelevant random class. In contrast, the latter usually occurs when a to-be-labeled sample contains confusing content or equivocal features and is the main cause of noisy labels in real-world applications. Confusing label noise often occurs on data samples lying near the decision boundaries, and such noisy labels should be corrected as one neighboring category to the current one in the feature space.

Most existing methods focused on correcting random label noises and reported promising performance on simulated datasets like noisy MNIST [4] and noisy CIFAR-10 [5] obtained by randomly altering data labels. Primarily, there are two sorts of strategies for solving the noisy label correction problem: transition matrix-based approaches [6, 7, 8] and class-prototype-based ones [9, 10]. All these methods learn from a given noisy dataset and correct noisy labels according to the distribution of labels of instances within a local neighborhood. As a result, they cannot well cope with real-world confusing label noise robustly because a confusing label is expected to lie near the decision boundary between categories and may be surrounded by other instances with incorrect confusing labels. Moreover, these methods may also give a wrong suggestion to a clean instance once it is located in the transition area between two akin classes in the feature space.

In contrast, some other methods resort to noise-robust loss functions to fight against label noise [11, 12, 13, 14]. Such noise-robust loss functions operate based on the assumption that whether an instance is mislabeled is irrelevant to its content. That is to say, the distribution of noisy labels is still assumed to be independent of that of data instances. Because this assumption is only valid in the case of random label noise, these methods usually cannot perform well on real-world data involving confusing label noise.

Still, a few label correction methods are ensemble learning-based approaches [15, 16, 17]. By taking a majority decision on predictions of multiple sub-models, ensemble learning approaches can improve the robustness of predictions. Some ensemble learning approaches are based on semi-supervised learning strategies, aiming to increase the freedom of predictions via suitable data augmentation schemes [18, 19]. Nevertheless, common data augmentation schemes cannot synthesize data distributing far more differently from the original dataset, especially for those with confusing noise labels. Therefore, these approaches usually need to collaborate with additional loss functions to guarantee the correctness of soft-labels predicted by different sub-networks. Although these methods were designed to increase the uncertainty of sub-networks’ predicted labels on noisy samples, they may also decrease the confidence of sub-networks’ predictions on clean instances. Hence, corrections suggested by these methods are usually considered as soft-labels for training another student model. A primary drawback of these methods is that they did not take the within-neighborhood label consistency into account, and thus the obtained soft-labels may still be noisy.

To address the above problems with confusing label noises, we propose an ensemble-based label correction algorithm by exploiting the local structures of data manifolds. As illustrated in Fig. 2, our noisy label correction scheme involves three iterative phases: i) kk-NN based data splitting, ii) multi-graph label propagation, and iii) confidence-guided label correction. . Because confusing label noises tend to be located densely near decision boundaries, they may destroy the label smoothness within some “inter-class transient bands” on the graph locally. By partitioning the source noisy dataset into disjoint subsets using our kk-NN splitting scheme, each noisy label, along with its kk-nearest-neighbors, will usually affect only a minority of the ensemble branches. As a result, each ensemble branch generates a graph that holds its own noisy local manifold structures so that such singulars can be treated as outliers during the majority decision process. To this end, we train the ensemble branches on the corresponding disjoint subsets independently. Through this design, each sub-network can learn not only a coarse global representation of the data manifold, but also different local manifold structures. We then derive label correction suggestions for each sample based on the predictions of the sample’s nearest-neighbors in individual disjoint subsets via the corresponding ensemble branches. Finally, our method suggests final label corrections by ruling out inconsistent suggestions derived according to graphs accessed by ensemble branches. Extensive experiments show the superiority of our method over existing state-of-the-arts.

The novelty of the proposed method is threefold.
\bullet We propose a novel iterative data splitting method to split training samples into disjoint subsets, each preserving some local manifold structure of source data while representing a coarse global approximation. This design allows the influence of mislabeled instances to be limited to a minority of ensemble branches.
\bullet Our design contains a novel noisy-label branch that can stably provide a correct suggestion for within-class clean labels. Hence, this branch can boost the accuracy of label correction result, especially for datasets primarily containing confusing label noises.
\bullet We adopt multi-graph label propagation, rather than a simple nearest-neighbor strategy, to derive label correction suggestions via multiple graph representations characterizing similar data manifolds. Hence, our method can take advantages of both nearest neighbors and manifold structures with the aid of graphs.

The rest of this paper is organized as follows. Some most relevant works are surveyed in Sec. II. Sec. III presents the proposed schemes for data splitting, multi-graph label propagation, and confidence-guided label correction. In Sec. IV, experimental results on public datasets with noise labels are demonstrated. Finally, conclusions are drawn in Sec. V.

II Related Work

II-A Loss Functions

To solve noisy label correction problem, several methods were developed by re-designing loss terms so that the learning system itself can re-weight the importance of training instances. For example, Wang et al. [20] employed an importance re-weighting strategy that enables the training on noisy data to reflect the results of noise free environment. Arazo et al. [21] evaluated their entropy-based per-sample loss for label correction by controlling the confidence of training sample via two weights. Still, some other losses, e.g., mean absolute error (MAE), generalized cross entropy (GCE), and reverse cross entropy (RCE), were proposed to avoid a biased learning system, the gradient saturation problem, or the overfitting/under-learning problem in the presence of noisy labels [22, 12, 23]. Nevertheless, although these methods prevent a learning system from fitting noisy samples, they may still be impracticable to complicated datasets due to the ignorance of hard samples.

II-B Label Correction

Most recent label correction methods work based on i) a noise transition matrix, or ii) class prototypes. As for the former sort of methods, they are based on an instance-independent assumption to derive the transition probability, which depicts how possible a clean label flipping into a noisy one, independent to any type of data structure. However, this assumption cannot prevent the generation of incorrect labels, nor can it prevent the transition matrix from learning noisy labels. Therefore, Xia et al. [8] proposed a transition revision method to address this issue, but their method is still limited by the fairness of the estimated initial transition matrix. Moreover, “class prototype” methods learn to represent each class via some class prototypes based on which data samples can receive corrected label information from their neighborhood [9, 10]. Nevertheless, these methods learn prototypes from a noisy dataset, so the quality and the accuracy of prototypes are still affected by the original noise distribution. Both these two types of methods learn from noisy data. Because there is no effective mechanism to verify the corrected labels suggested by the transition matrix or the class prototypes, the resulting pseudo-label may be incorrect and generate a worse correction suggestion if the class prototypes or the transition matrix overfits noisy training labels.

II-C Ensemble learning

Some other methods exploit i) ensemble learning that feeds stochastically-augmented inputs to a number of parallel sub-models to derive the corresponding prediction jointly with ii) an averaging method to derive label correction suggestions [16, 19]. These methods are based on the inconsistency among information learned by different sub-networks. For example, MLNT [15] randomly changes the labels of each sub-network’s training data to enhance the difference learned by each subnet, and DivideMix [16] uses an additional regularization term to increase the prediction diversity among different sub-models. However, making sub-models learn to be different from each other cannot ensure the correctness of any sub-model itself, and they may be worse than they are supposed to be due to an unmatched noise distribution. For example, Yan et al.’s method is a method works only on uniformly distributed label noises [24]. In real-world applications, label noise usually distributes in some specific way rather than uniformly. Consequently, it is still intractable for simple ensemble learning approaches to filter out noisy labels.

II-D Label Propagation

Label propagation is a graph-based semi-supervised method for generating pseudo-labels of unlabeled data based on given anchors [25, 26, 27]. Compared with the nearest-neighbor strategy that does not take the data manifold into account, label propagation methods can propagate labels from labeled anchors to unlabeled samples through a graph characterizing the manifold of the noisy dataset. For example, the method in [28] uses an affinity matrix to derive an approximate kk-NN graph, through which it performs label propagation for a large-scale diffusion task. Iscen et al. [26] proposed a transductive label propagation method to make label predictions by using i) a nearest-neighbor graph, ii) conjugate gradient based label propagation, and iii) a loss weighted by label certainty. Moreover, Bontonou et al. [29] proposed a graph smoothness loss, maximizing the distance between samples of different classes, to derive better feature and manifold representations for classification purpose. This loss increases the robustness of a learning model against noisy training inputs, though its performance is still similar to that of cross-entropy because only distances between instances in different classes are considered.

III Proposed Method

III-A Overview

Refer to caption
Figure 2: Framework of proposed ensemble-based label correction scheme. The proposed method has primarily three branches, namely i) noisy label branch controlled by n\mathcal{L}_{n}, ii) corrected label branch controlled by pseudo\mathcal{L}_{pseudo}, and iii) ensemble branches controlled by G\mathcal{L}_{G}. Our design focuses on the ensemble branches implemented in three phases. The noisy label branch is used to perform feature embedding and also used to regulate certainty weight jointly with the corrected label branch. Only the corrected label branch is used while deploying. The three phases of our ensemble branches are detailed in Sections III-BIII-D. Note that y(n)y(n) denotes the original label given by the noisy dataset.

Fig. 2 illustrates the framework of the proposed method for label correction. Let di=(xi,yi)𝒟d_{i}=(x_{i},y_{i})\in\mathcal{D} denote a sample-label pair from a data collection (𝒳,𝒴)(\mathcal{X},\mathcal{Y}), where 𝒟={di}\mathcal{D}=\{d_{i}\} is the source noisy dataset. Our ensemble learning scheme aims to split 𝒟\mathcal{D} into MM disjoint subsets, namely 𝒮m\mathcal{S}_{m} for m=1,,Mm=1,\dots,M, for training MM ensemble branches. During the training, our method first trains MM classifiers Φm\Phi_{m} with parameter θm\theta_{m}, through the MM ensemble branches, independently on the MM disjoint subsets 𝒮m\mathcal{S}_{m} derived by our data splitting strategy. Second, by feeding all training samples xix_{i} into each classifier, we extract totally MM different feature vector sets, each of which can be used to derive one graph representation of the data manifold of training dataset 𝒳={xi}\mathcal{X}=\{x_{i}\}. Third, for given the mm-th partial-label set (xjm,yjm)𝒮m(x^{m}_{j},y^{m}_{j})\in\mathcal{S}_{m} where only labels yjmy^{m}_{j} belonging to 𝒮m\mathcal{S}_{m} are available, we construct its graph representation and use the graph to predict the label correction suggestions through label propagation for all data samples in the remaining M1M-1 partial-label sets. As a result, we use the MM partial-label sets to predict totally M×MM\times M label correction suggestions based on the MM graphs derived in the second step. Fourth, based on the corrected labels y^lm𝒴^m\hat{y}^{m}_{l}\in\hat{\mathcal{Y}}_{m} and the MM sample-correction sets {(xlm,y^lm)|xlm𝒮m}\{(x^{m}_{l},\hat{y}^{m}_{l})\,|\,\forall x^{m}_{l}\in\mathcal{S}_{m}\} obtained in the ll-th training epoch, we generate another M×MM\times M label propagation suggestions. Finally, we derive the most likely correction result 𝒴¯={y¯i}\bar{\mathcal{Y}}=\{\bar{y}_{i}\} from the total 2×M22\times M^{2} label correction suggestions via a majority decision.

Specifically, based on the assumption that real-world noisy labels (e.g., obtained via crowd-sourcing) tend to be locally concentrated near decision borders, we devise a data splitting strategy to partition the noisy source dataset 𝒟\mathcal{D} into MM disjoint subsets 𝒮m\mathcal{S}_{m}, each being the training set for one ensemble branch. In this way, SmS_{m} is expected to retain the shapes of kk randomly-selected local neighborhoods, no matter noisy or not, as well as ensemble a random subset of xi𝒳x_{i}\in\mathcal{X}. Therefore, by taking a majority decision on the approximate manifolds learned by the MM sub-networks, the local influences brought by noisy samples can be mitigated. Finally, the noisy label branch in our design is trained on the whole original dataset 𝒟\mathcal{D} to derive the feature vectors of all xix_{i}. Meanwhile, the corrected label branch is trained based on corrected labels, i.e., (𝒳,𝒴¯)(\mathcal{X},\bar{\mathcal{Y}}), obtained after the l-th training epoch, and it is the only branch used for deployment.

Our key idea is to restrict the influence of any group of locally-concentrated noisy labels to only a minority of sub-networks via our local-patch-based data slitting strategy. As a result, most ensemble branches can learn their own relatively correct approximations of the data manifold around the noisy local patch, so that a suitable label correction result can be yielded by majority decision, accordingly. Also, no matter label noise is locally-concentrated or uniformly distributed in the data space, it can be effectively mitigated by the proposed ensemble method. Overall, each training epoch of our method can be divided into three phases, as will be elaborated in Sec. III-BIII-D.

III-B Phase-1: Data Splitting/Re-splitting

The first phase of each training epoch is to randomly scatter per local neighborhood of data points on the source data manifold, including noisy labels, into disjoint subsets. Specifically, as illustrated in Fig. 3, for each selected xix_{i}, we pick only its kk nearest-neighbors of the same class to form a local patch around xix_{i} with k=Nc/(M×B)k=N_{c}/(M\times B), where BB is a predefined hyper-parameter, and NcN_{c} denotes the cardinality of the class that xix_{i} belongs to. Therefore, because each disjoint subset 𝒮m\mathcal{S}_{m} holds different local feature relationships of each class, we can confine the negative influence of each locally-concentrated noisy neighborhood to primarily one 𝒮m\mathcal{S}_{m} and one ensemble branch by packaging every local kk-nearest neighbors. This is the main advantage of the proposed method. As a result, on one hand, when a noisy local neighborhood only contaminates a minority of ensemble branches, our method can vote down the negative influence of noisy labels by a majority decision. This case leads to a performance leap with our method: the performance upper-bound. On the other hand, in the extreme case that all noisy labels are uniformly distributed globally, our data splitting strategy will not alter the noise distribution so that our ensemble model leads to a as good performance as typical random-selection schemes: the baseline performance. In sum, our method can be expected to provide a performance in between the upper-bound and the baseline. Since real-word label noise distribution tends to be concentrated on decision boundaries, our method can usually lead to performance improvements.

The proposed data splitting method is composed of the following steps:
Initialization: Determine the hyper-parameters, including the number of ensemble branches MM and the number of data packages per class per branch BB.
kk-NN Packaging: As illustrated in Fig. 3, for each class, we separate its data into MBM\cdot B packages, each containing a randomly selected seed sample and its kk-NN in the feature domain with k=NcMBk=\frac{N_{c}}{M\cdot B}, where NcN_{c} denotes the number of samples belonging to class cc. Note that we adopt “random sampling without replacement” to guarantee that each sample and its kk-NN can be assigned into only one package.
Splitting: For each class, we randomly assign the MBM\cdot B packages into MM disjoint subsets. Each subset 𝒮m\mathcal{S}_{m} therefore stands for a coarse global approximation of the source data manifold but holds fine local manifold structures of different places independently.

Refer to caption

(a)
Refer to caption
(b)

Figure 3: Conceptual illustration of our data splitting strategy. For each randomly selected seed (highlighted in yellow), we pick up only its kk-NN of the same class (highlighted in white) to form a package, and k=NcMB˙k=\frac{N_{c}}{M\dot{B}} with NcN_{c} denoting the number of instances of class-cc. This figure illustrates that each subset can be regarded as a random subset of the whole dataset if B=1B=1. (a) An example of splitting “crosses” with k=7k=7. (b) An example of splitting “triangles” with k=5k=5. Note that dashed samples are mis-annotated ones.

Moreover, we initialize each training epoch with the data splitting process, and we name this strategy re-splitting. Re-splitting enables each resemble branch to learn a different coarse approximation of the data manifold to prevent the resemble branches from overfitting and being biased to the same training data. Because 𝒮m\mathcal{S}_{m} is changed for each epoch, we need to initialize the model parameter θml\theta^{l}_{m} for each ensemble branch to guarantee the fast convergence of the ll-th training epoch by

θml=(1αm)θpseudol1+αmθnoisyl1,\theta_{m}^{l}=(1-\alpha_{m})\theta_{\mbox{pseudo}}^{l-1}+\alpha_{m}\theta_{\mbox{noisy}}^{l-1}\mbox{,} (1)

where αml\alpha_{m}^{l} is a random value within [0,1][0,1]; and, θpseudol1\theta_{\mbox{pseudo}}^{l-1} and θnoisyl1\theta_{\mbox{noisy}}^{l-1} are respectively the model parameters, obtained after the (l1)(l-1)-th training epoch, of the corrected-label branch and that of the noisy-label branch.

III-C Phase-2: Multi-Graph Label Propagation

We then construct the mm-th graph representing the data manifold based on the features fim=f(xi,θm)f^{m}_{i}=f(x_{i},\theta_{m}) extracted by the mm-th model. Then, for each graph, we can produce M+MM+M label correction suggestions by taking i) the original sample-label pair (xjm,yjm)𝒮m(x_{j}^{m},y_{j}^{m})\in\mathcal{S}_{m} and ii) the sample-correction pair (xjm,y^jm,l)(x_{j}^{m},\hat{y}_{j}^{m,l}) obtained in the ll-th training epoch as different starting partial-labels, where y¯jm,l\bar{y}_{j}^{m,l} denotes the jj-th sample’s label correction suggestion given by the mm-th ensemble branch in the ll-th training epoch. As a result, our method can hence offer totally M(M+M)=2M2M\cdot(M+M)=2M^{2} sets of label correction suggestions.

Our multi-graph label propagation strategy works based on the features extracted by classification models Φm(;θm)\Phi_{m}(\cdot;\theta_{m}) of the MM ensemble branches. The first step is to build MM normalized weighted undirectional adjacency matrices 𝒲m\mathcal{W}_{m}, each representing the mm-th graph constructed based on fimf_{i}^{m}, for the label propagation process described in (4), and 𝒲m\mathcal{W}_{m} is required to be symmetric [25]. The matrix 𝒲m\mathcal{W}_{m} is defined as

𝒲m\displaystyle\mathcal{W}_{m} =\displaystyle= Dm12(Am+AmT)Dm12,\displaystyle D_{m}^{-\frac{1}{2}}(A_{m}+A^{T}_{m})D_{m}^{-\frac{1}{2}}\mbox{,} (2)

where DmD_{m} is a diagonal matrix used to normalize (Am+AmT)(A_{m}+A^{T}_{m}), and AmA_{m} is a directional adjacency matrix whose (ss,tt)-th entry is defined as follows:

Am(s,t)={sim(fsm,ftm)γ, for fsm𝒩(ftm),st0, otherwiseA_{m}(s,t)=\bigg{\{}\begin{array}[]{ll}sim(f^{m}_{s},f^{m}_{t})^{\gamma}&\mbox{, for }f^{m}_{s}\in\mathcal{N}(f^{m}_{t}),s\neq t\\ 0&\mbox{, otherwise}\end{array} (3)

where fsmf^{m}_{s} is the feature of the ss-th sample extracted by the model of the mm-th branch, 𝒩(ftm)\mathcal{N}(f^{m}_{t}) denotes the collection of the kk nearest neighbors of ftmf^{m}_{t}, sim(,)sim(\cdot,\cdot) denotes the cosine similarity metric ranging in [0,1][0,1], and γ\gamma is a predefined hyper-parameter. Note that AmA_{m} is asymmetric.

Next, we propagate the label information of i) the sample-label pairs (xjm,yjm)(x^{m}_{j},y^{m}_{j}) in the mm-th disjoint subset 𝒮m\mathcal{S}_{m} and ii) sample-correct pairs (xjm,y¯jm,l)(x^{m}_{j},\bar{y}^{m,l}_{j}) in turn through each of the MM graphs to obtain totally 2M22M^{2} label correction suggestions, i.e., partial-labels, for each data sample xix_{i}. We adopt such a strategy because of the following two concerns. First, while using only original noisy labels yjmy^{m}_{j} to derive the label propagation result, the obtained labels are probably contaminated, so this setting surely limits the accuracy of obtained labels. Second, while using only corrected labels y^jm,l\hat{y}^{m,l}_{j}, the propagation result may overfit the corrected labels, thereby misleading the label correction process. Therefore, we exploit both the original noisy labels yjmy^{m}_{j} and corrected labels y^jm,l\hat{y}^{m,l}_{j} in our label propagation process.

Let Zm,jZ_{m,j} denote the label correction suggestion, the label propagation process [25] is conducted by

Zm,j=(𝐈α𝒲m)1Yj,Z_{m,j}=(\mathbf{I}-\alpha\mathcal{W}_{m})^{-1}Y_{j}\mbox{,} (4)

where YjY_{j} is an N×C×2N\times C\times 2 matrix that records the reference labels associating the data samples belonging to the jj-th disjoint subset 𝒮j\mathcal{S}_{j}, 𝒲m\mathcal{W}_{m} is the normalized weighted undirectional adjacency matrix defined in (2), and NN and CC denote respectively the cardinality of the original dataset 𝒳\mathcal{X} and the cardinality of 𝒴\mathcal{Y}, i.e., the number of total classes. Note that the third dimension of YjY_{j} contains two partial-labels: the original noisy one yjmy^{m}_{j} and the corrected one y^jm,l\hat{y}^{m,l}_{j}.

Finally, the partial-label suggested by the mm-th ensemble branch based on the label information of 𝒮j\mathcal{S}_{j} becomes

Y^m,j(n,:)=argmaxcZm,j(n,c,:),\hat{Y}_{m,j}(n,:)=\mathop{\arg\!\max_{c}}Z_{m,j}(n,c,:)\mbox{,} (5)

where every Zm,jZ_{m,j} is an N×C×2N\times C\times 2 matrix, and Zm,j(n,c,)Z_{m,j}(n,c,\cdot) indicates how possible the nn-th sample belongs to the cc-th class. After obtaining 2M22M^{2} different label suggestions Y^m,j\hat{Y}_{m,j}, it turns to Phase-3 to derive the final correction result.

III-D Phase-3: Confidence-Guided Label Correction

We select the most frequent category among all label suggestions Y^m,j(n,:)\hat{Y}_{m,j}(n,:) via a majority decision as the final label correction result y¯(n)\bar{y}(n).

To guide the label correction, we propose a normalized average confidence level to devise our loss function. The normalized average confidence level ω¯(n)\bar{\omega}(n) of the final label correction result y¯(n)\bar{y}(n) is defined as

ω¯(n)=ω^(n)min{ω^(n)}max{ω^(n)}min{ω^(n)},\bar{\omega}(n)=\frac{\hat{\omega}(n)-\min\{\hat{\omega}(n)\}}{\max\{\hat{\omega}(n)\}-\min\{\hat{\omega}(n)\}}\mbox{,} (6)

where

ω^(n)=12M2(m,j,q)Ωnωm,j(n,q),\hat{\omega}(n)=\frac{1}{2M^{2}}\sum_{(m,j,q)\in\Omega_{n}}\omega_{m,j}(n,q)\mbox{,} (7)

where Ωn={(m,j,q)y¯(n)=Y^m,j(n,q)}\Omega_{n}=\{(m,j,q)\mid\bar{y}(n)=\hat{Y}_{m,j}(n,q)\} means that only those samples, whose label correction suggestions Y^m,j(n,q)\hat{Y}_{m,j}(n,q) are identical to the final corrected label y¯(n)\bar{y}(n), are used to calculate the normalized average confidence level. Therefore, ω¯\bar{\omega} can be maximized when i) the predictions have a high confidence level, which implies a relatively low entropy, or ii) the nn-th corrected labels suggested by all graphs are of the same category.

Note that the confidence level ωm,j(n)\omega_{m,j}(n) of the label propagation result Zm,jZ_{m,j} can be assessed based on the certainty weight described in [26]:

ωm,j(n,:)=1H(Z¯m,j(n,c,:))log(C),\omega_{m,j}(n,:)=1-\frac{H(\bar{Z}_{m,j}(n,c,:))}{\log(C)}\mbox{,} (8)

where

Z¯m,j(n,c,:)\displaystyle\bar{Z}_{m,j}(n,c,:) =\displaystyle= Zm,j(n,c,:)kZm,j(n,k,:),\displaystyle\frac{Z_{m,j}(n,c,:)}{\sum_{k}Z_{m,j}(n,k,:)}\mbox{,}
H(Z¯m,j(n,c,:))\displaystyle H(\bar{Z}_{m,j}(n,c,:)) =\displaystyle= c=1CZ¯m,j(n,c,:)log(Z¯m,j(n,c,:)),\displaystyle-\sum_{c=1}^{C}\bar{Z}_{m,j}(n,c,:)\log(\bar{Z}_{m,j}(n,c,:))\mbox{,}

where Z¯m,j\bar{Z}_{m,j} records the occurrence frequency of each label in the label propagation result matrix Zm,jZ_{m,j}, and H()H(\cdot) denotes the entropy. Therefore, a larger ωm,j(n,q)\omega_{m,j}(n,q) implies a label Y^m,j(n,q)\hat{Y}_{m,j}(n,q) with higher confidence (i.e., lower uncertainty).

After normalizing the final weights described in (6), we use i) the corrected labels and ii) the weighted cross entropy loss to train the next-epoch ensemble branches and the next-epoch corrected label branch. The loss functions used to train our ensemble model are detailed in the next subsection.

III-E Loss Functions

We adopt the following weighted cross-entropy to measure the loss of corrected (pseudo) labels branch:

pseudo=nω¯(n)y¯(n)logppseudo(n,y¯(n)),\mathcal{L}_{\mbox{pseudo}}=-\sum_{n}\bar{\omega}(n)\bar{y}(n)\log p_{\mbox{pseudo}}(n,\bar{y}(n))\mbox{,} (9)

where y¯(n)\bar{y}(n) is the pseudo label of x(n)x(n) derived from our label correction process in the current training epoch, and ppseudo(n,y¯(n))p_{\mbox{pseudo}}(n,\bar{y}(n)) denotes the probability, predicted by the softmax layer of the corrected (pseudo) label branch, of x(n)x(n) belonging to class y¯(n)\bar{y}(n).

Then, the noisy label branch is trained via the weighted cross-entropy loss, i.e.,

noisy=nnoisy(n),\mathcal{L}_{\mbox{noisy}}=-\sum_{n}\ell_{\mbox{noisy}}(n)\mbox{,} (10)

where

noisy(n)\displaystyle\ell_{\mbox{noisy}}(n) =\displaystyle= {ω¯(n)y(n)logpnoisy(n,c), if c=y¯(n)(1ω¯(n))y(n)logpnoisy(n,c), otherwise,\displaystyle\bigg{\{}\begin{array}[]{ll}\bar{\omega}(n)y(n)\log p_{\mbox{noisy}}(n,c)&\mbox{, if }c=\bar{y}(n)\\ (1-\bar{\omega}(n))y(n)\log p_{\mbox{noisy}}(n,c)&\mbox{, otherwise,}\end{array}

where pnoisyp_{\mbox{noisy}} denotes the probability predicted by the softmax layer of the noisy label branch. This loss means that if the n-th sample is corrected, we give it a complementary confidence level 1w(n)1-w(n) to highlight its low confidence.

Moreover, because the MM ensemble branches are mainly designed to construct the MM graphs rather than predicting the classification probability of each sample, we exploit a graph smoothness loss modified from [29] to derive a set of graph embedding via which the inter-class distances are maximized, similar to the concept described in [30]. This loss uses radial basis functions to measure the weighted distance between the features of any two samples of different classes. Because the two samples may come from different 𝒮m\mathcal{S}_{m}, this loss can lead to a large enough between-class margin on all graphs. That is,

G=(s,t)ω¯sω¯texp{α𝐩s(s)𝐩t(t)2},\mathcal{L}_{G}=\sum_{(s,t)}\sqrt{\bar{\omega}_{s}\bar{\omega}_{t}}\exp\big{\{}-\alpha\|\mathbf{p}^{\mathcal{I}(s)}_{s}-\mathbf{p}^{\mathcal{I}(t)}_{t}\|_{2}\big{\}}\mbox{,} (13)

where (s,t)Ω={(s,t)y¯(s)y¯(t)}(s,t)\in\Omega=\{(s,t)\mid\bar{y}(s)\neq\bar{y}(t)\}, (s)\mathcal{I}(s) is an indicator function that returns the index of the disjoint subset containing the ss-th sample, and 𝐩s(s)\mathbf{p}^{\mathcal{I}(s)}_{s} denotes the ss-th sample probability vector generated by the softmax layer of the subnetwork trained on the (s)\mathcal{I}(s)-th subset. This loss forces the distance between two nodes of different classes on the graph to be large.

G\mathcal{L}_{G} has an additional weighting factor w¯sw¯t\sqrt{\bar{w}_{s}\bar{w}_{t}} which gives an instance pair with a lower confidence level a lower weight to alleviate possible negative effects of instances with lower confidence levels. G\mathcal{L}_{G} is different from typical entropy loss and triplet loss [31] designs because G\mathcal{L}_{G} aims to map instances of different classes into feature clusters far away from each other. On the contrary, both cross-entropy loss and triplet loss encourage the network to map instances of the same class into the similar features. Consequently, G\mathcal{L}_{G} can assist our model to produce distinguishable features for label correction tasks. Finally, the total loss of the proposed model is

total=G+pseudo+n.\mathcal{L}_{\mbox{total}}=\mathcal{L}_{G}+\mathcal{L}_{\mbox{pseudo}}+\mathcal{L}_{n}\mbox{.} (14)

IV Experimental Results

IV-A Datasets

We evaluate the proposed method on i) simulated noisy datasets based on CIFAR-10/100 and ii) two other real-world noisy datasets: Clothing1M [32] and Food101-N [33]. The properties of Clothing1M and Food101-N are described below.

First, Clothing1M contains images with 14 categories of fashion clothes. The data in Clothing1M are collected from the Internet, and their category labels are generated based on the surrounding texts, which leads to a label noise rate of 38.46%. Clothing1M also provides a small clean set with labels corrected by human annotators. The numbers of images in Clothing1M’s label-corrected training, validation, testing sets, and the noisy labeled set are about 47.5k, 14.3k, 10.5k, and 1M, respectively. In the four sets, only the 1M noisy labeled set contains noisy labels. We use the noisy labeled set to train our model to evaluate its performance, and then we conduct the ablation study jointly on the noisy labeled set and the human-annotated clean testing set.

Second, Food101-N contains about 310K food images of 101 categories. The images in Food101-N are collected from the Internet with a noisy label rate of 20%. To evaluate the performance of our method, we apply it on Food101-N and derive a label correction result first, and then we train a classifier on the label-corrected dataset and conduct testing on Food-101’s testing set [34]. Note that Food-101’s testing set contains only 55k images with clean labels.

IV-B Experiment Settings

We summarize our experiment settings as the following four points. First, after the data splitting phase of each training epoch, we resize training images xim𝒮mx^{m}_{i}\in\mathcal{S}_{m} to 256×256256\times 256, and then cropped them randomly into 224×224224\times 224 patches. Hence, training samples for each training epoch are different variants of the source data, which improves the diversity of training samples for each epoch. Second, the optimizer is SGD, and the momentum value is 0.90.9. The initial learning rate is 0.010.01, and it is reduced by a factor of 110\frac{1}{10} every 55 epochs. The maximal training epoch is set to be 1515, and the batch size is 6464. We also adopt ridge (L2) regularization that is weighted by 5×1035\times 10^{-3}. Third, the number of branches is M=5M=5, and the number of data packages per class in each branch is B=4B=4. Fourth, the features of all training samples are extracted initially by a ResNet-50 [35] pretrained on ImageNet. After the first training epoch, the features used for our data (re-)splitting phase are derived by the latest network model trained in the noisy label branch.

IV-C Experiments on Clothing1M

We evaluate our model’s performance on Clothing1M through two different scenarios. The first scenario is to train our model and other state-of-the-arts on only noisy datasets without any additional supervision. In the second scenario, all models are trained on the same hybrid set consisting of i) the noisy set and ii) the clean set consisting of 50K human-annotated images, and then the obtained models are further fine-tuned on the 50K-image clean dataset via the cross-entropy loss. The results of the first and the second parts are shown respectively in Tables I and II. Note that in the second scenario, the labels of the clean dataset are kept unchanged, and their certainty weight are set to be 11. Table I demonstrates that our method achieves state-of-the-art performance, i.e., 75.17% accuracy, compared with the other methods. Meanwhile, in the second scenario, our method still outperforms all compared state-of-the-arts without any fine-tuning and is comparable to Self-Learning [10] after fine-tuning with the 50k clean set.

t Method Acc.(%) CrossEntropy 69.21 MLNT [15] 73.47 T-Revision [8] 74.18 Self-Learning [10] 74.45 MetaCleaner [36] 72.50 DivideMix [16] 74.76 NoiseRank [9] 73.82 ELR [37] 74.81 Ours 75.17

TABLE I: Classification accuracy of different methods on the noisy set of Clothing1M

h Method Acc.(%) Fine-tuning CrossEntropy 75.67 x DataCoef [38] 77.21 x Ours 78.01 x CrossEntropy 80.32 o Forward [6] 80.38 o MetaCleaner [36] 80.78 o CleanNet [33] 79.90 o Self-Learning [10] 81.16 o Ours 81.13 o

TABLE II: Classification accuracy of different methods on a hybrid set of the noisy set and the clean set, containing 50K images, of Clothing1M

IV-D Experiments on Food101-N

The classifier, trained on the data corrected by our method, achieves an second best accuracy of 85.13% on Food101-N, comparable to the best performance of 85.79% achieved by NoiseRank [9], as demonstrated in Table III. Although the noise rate of Food101-N is about 20%, much lower than that of Clothing1M, the noisy labels in Food101-N are much more complicated. Incorrect labels in Food101-N involve three different types: i) photos with confusing contents, ii) multi-label images, e.g. photos with multiple foods, and iii) anomalies, e.g. a shopfront photo labeled as “fish and chips”. Fig. 4 shows four samples of the second and the third types of mis-annotations. Because our method is designed for confusing label noises, i.e., only one of the three types of mis-annotations in Food101-N, the performance of our method is just comparable with typical data cleaning methods like NoiseRank [9]. Nevertheless, data cleaning methods tend to remove hard samples simultaneously, which leads to irreversible information loss. Consequently, our method works much better than NoiseRank on Clothing1M, a dataset containing no anomalies, but not so good as NoiseRank in fighting against multi-label samples and anomalies in Food101-N.

Refer to caption
Figure 4: Samples of three multi-label images and one anomaly in Food101-N.
TABLE III: Classification accuracy of different methods on Food101-N dataset with additional 50k clean set.
Method Acc.(%)
CrossEntropy 81.97
CleanNet [33] 83.95
MetaCleaner [36] 85.05
Self-Learning [10] 85.11
NoiseRank [9] 85.79
Ours 85.13

IV-E Analyses and Discussions

In this subsection, we discuss the effect of each component in our method, including i) the number of local manifolds BB, ii) the number of ensemble branches MM, iii) the settings of G\mathcal{L}_{G}, and iv) the re-splitting process.

\bullet Number of local manifolds BB
Fig. 5 shows that when M=5M=5, our method achieves the best accuracy 75.17% on the noisy set of Clothing1M when B=4B=4. In addition, B=1B=1 and B=2,048B=2,048 are two extreme conditions,. The former represents that each disjoint subset 𝒮m\mathcal{S}_{m} contains one large local patch, and therefore it is impossible for each subset to approximate the global data manifold. On the contrary, the latter denotes a strategy similar to random sampling, implying that each subset can coarsely approximate the whole data manifold yet loses the ability to retain the local information. When B=2,048B=2,048, the proposed ensemble learning strategy has no advantage in coping with noisy labels because data in all disjoint subsets follow the same distribution.

\bullet Number of Ensemble Branches MM
The number of ensemble branches, MM, is also a key hyper-parameter in our method. To evaluate the impact of MM, we conduct experiments by fixing B=4B=4. As shown in Fig. 6(a), the ensemble branches are beneficial for estimating data manifold because the cosine similarities in the adjacency matrices of derived graphs roughly increases with the number of ensemble branches. However, Fig. 6(b) shows that the testing accuracy tends to be saturated when M5M\geq 5. This implies that when M5M\geq 5, some ensemble branches may learn nearly the same coarse approximation of data manifold, and such information might be redundant thus cannot further improve the generalization performance. Therefore, we set B=4B=4 by default.

Moreover, Fig. 6 provides another interpretation. By checking the average cosine similarities in an adjacency matrices derived by different ensemble branches during training, we find that the final ensemble model performs the best when the average cosine similarity is in between 0.70.7 and 0.80.8. When the average cosine similarity reaches 0.90.9, the model becomes overfit because the testing accuracy of the ensemble model decreases, slightly though.

\bullet Effects of different combinations of (M,B)(M,B)
Fixing the number of local patches per class, i.e., MBM\cdot B, we here discuss the influence brought by different combinations of MM and BB. The experiments demonstrated in Fig. 7 are derived by setting MB=40M\cdot B=40 and MB=20M\cdot B=20. For M×B=20M\times B=20, we conduct experiments on (M,B)=(2,10)(M,B)=(2,10), (4,5)(4,5), (5,4)(5,4), and (10,2)(10,2). As for M×B=40M\times B=40, we examined (M,B)=(2,20)(M,B)=(2,20), (4,10)(4,10), (5,4)(5,4), and (10,2)(10,2) in turn. Based on these results, we have two observations. First, (M,B)=(4,5)(M,B)=(4,5) achieves the best accuracy of 75.15%75.15\%, whereas (M,B)=(8,5)(M,B)=(8,5) reaches about 75.05%75.05\%. Second, MM is far more critical than BB, and a suitable MM should range in between 4 and 10. This is because once a too large MM is used, the source noisy data will be divided into too small disjoint subsets, and such small subsets cannot retain the shape of the entire data manifold.

Refer to caption
Figure 5: Test accuracy (%) for different numbers of local manifolds (BB) with M=5M=5.
Refer to caption
Figure 6: (a) Average cosine similarity in each adjacency matrix with epoch from 1 to 15 in training process. (b) Test accuracy (%) with epoch from 1 to 15 in training process. The line denote different setting of MM and all the model use same B=4B=4.
Refer to caption
Figure 7: (a) Test accuracy (%) with different combinations of [M,B][M,B] and MB=20MB=20. (b) Test accuracy (%) with different combinations of [M,B][M,B] and MB=40MB=40.

\bullet Weighted Graph Smoothness Loss
Our G\mathcal{L}_{G} described in (13) is measured by the L2L_{2} distance between the probability vectors 𝐩s(s)\mathbf{p}_{s}^{\mathcal{I}(s)} produced by the softmax layer of different ensemble branches. This design encourages different sub-networks to produce the same classification results instead of similar feature vectors. To verify the effectiveness, we run experiments via JS-divergence, KL-divergence, and feature vectors for comparisons, as shown in Table IV. First, the experiments show that G\mathcal{L}_{G} is more effective than cross-entropy loss because cross entropy tends to make all instances in the same class have features of the same kind but G\mathcal{L}_{G} does not. Second, by simply using the L2L_{2} distance between probability vectors, i.e., GS+pp+L2, rather than the L2L_{2} distance between feature vectors, i.e., GS+ff+L2, the accuracy of our model can be boosted from 74.47% to 75.15%. Third, no matter L2L_{2} distance, KL-divergence, or JS-divergence is adopted, the softmax output can consistently boost the model accuracy up to 75.17%. This proves the effectiveness of G\mathcal{L}_{G} and the usage of softmax output. Note that wGS denotes weighted graph smoothness loss, pp probability vector, and ff feature vector.

TABLE IV: Classification accuracy of different GRAPH SMOOTHNESS LOSS (GS) on Clothing1M.
Method Acc.(%)
CrossEntropy 74.08
GS + Feature + L2 74.47
GS + Softmax + L2 75.15
Weighted GS + Softmax + L2 75.17
Weighted GS + Softmax + KL 74.92
Weighted GS + Softmax + JS 74.98

\bullet Data Re-splitting
Our experiment shows that the data re-splitting process described in Sec. III-B can on average increase the accuracy by about 0.04%. Specifically, without data re-splitting the accuracy is on average 75.13%, whereas the accuracy on average increases to 75.17% with the aid of data re-splitting. This experiment supports the idea that re-splitting can assist ensemble branches to learn different combination of local patches and also prevent them from overfitting.

We validate the data re-splitting process in the following two steps. First, we extract the feature vectors of images in the to-be-corrected noisy dataset by using a ResNet-50 pretrained on Clothing1M via cross entropy. Second, with the obtained features, we split the to-be-corrected noisy dataset into MM disjoint subsets and then apply our method for label correction without data re-splitting.

\bullet Certainty weight

Fig. 8 illustrates how certainty weight varies with training epochs. The certainty weight described in (6) represents the probability of the correctness of a corrected label. These three figures jointly show that our system design is valid because the number of samples with larger certainty weights increases with the number of training epochs.

Refer to caption
Figure 8: Distribution of corrected labels and certainty weights on Cifar-10 with 80% symmetric noise at (a) the 10-th epoch, (b) the 100-th epoch, and (c) the 200-th epoch. Note that the horizontal axis denotes the value of certainty weight, and the vertical axis represents the data amount.

IV-F Ablation Study

In this section, we examine our system architecture design and clarify the effects brought by different combinations of loss terms and system components for all variants of the proposed method. We first examine the performance of each of our model variants on CIFAR-10 with man-made symmetric noises. Fig. 9 shows the t-SNE maps of datasets used for this part of ablation study.

Refer to caption
Figure 9: t-SNE maps of clean training datasets (the top row) and to-be-corrected noisy datasets (the bottom row). Left: Clothing1M with 38.36% noise rate. Middle: CIFAR-10 with 50% symmetric noise. Right: CIFAR-10 with 40% symmetric noise.

In this experiment set, what we examine include i) the necessity of noisy label branch in our default architecture, ii) the impact of Mixup [39] used for data augmentation, and iii) the effects brought by different loss terms used in the ensemble branches. Mixup is a procedure developed for data augmentation for semi-supervised learning, that enables the obtained model to produce linear prediction results between any pair of samples. Hence, it can yield smooth and robust pseudo labels and also avoid overfitting problems. We summarize the comparison among different loss term designs as follows:

\bullet Ensemble branch:
Although the graph smoothness loss described in (13) is beneficial for deriving a better adjacency matrix, this loss may overfit wrongly-labelled samples easily. Hence, in our default setting, we measure this loss only on those samples whose labels remain unchanged after our correction procedure. However, such setting may reduces the effectiveness of this loss, because, when the loss rate is too high, the amount of uncorrected samples participating in the computation of the graph smoothness loss would be too small. As a result, we measure the weighted sum of i) the MAE (mean absolute error) en_mae\mathcal{L}_{en\_mae} between 11 and the classification probability of corrected samples, and ii) the cross entropy en_ce\mathcal{L}_{en\_ce} of uncorrected samples as an alternative of the loss term for ensemble branches. We then compare this alternative design with our default setting G\mathcal{L}_{G} described in (13).
\bullet Corrected label branch:
Similarly, we adopt the sum of i) MAE cor_mae\mathcal{L}_{cor\_mae} and ii) cross entropy cor_ce\mathcal{L}_{cor\_ce} as an alternative loss design for the corrected label branch. We use the soft-label, which is estimated as a convex combination, weighted by the certainty weight ω¯\bar{\omega} depicted in (6), of the prediction result derived by the corrected label branch and that derived the ensemble branches, to compute both the cor_mae\mathcal{L}_{cor\_mae} and the cor_ce\mathcal{L}_{cor\_ce}. A larger ω¯\bar{\omega} will produce a soft-label closer to the label suggested by the corrected label branch. In addition, the weighting factors for loss terms used in our ablation study are listed in Table V.

TABLE V: Noise rate versus weighting factors for alternative loss terms used in our ablation study. (This table needs re-layout.)
Cifar-10
Noise Rate 20 50 80 90 asym 40%
en_ce\mathcal{L}_{en\_ce} 1 1 1 1 1
en_mae\mathcal{L}_{en\_mae} 0 0 1 1 0
cor_ce\mathcal{L}_{cor\_ce} 1 1 1 1 1
cor_mse\mathcal{L}_{cor\_mse} 0 1 10 10 1
Cifar-100 Clothing1M
Noise Rate 20 50 80 90
en_ce\mathcal{L}_{en\_ce} 2.5 2.5 2.5 2.5 1
en_mae\mathcal{L}_{en\_mae} 0 1 1 1 1
cor_ce\mathcal{L}_{cor\_ce} 1 1 1 1 1
cor_mse\mathcal{L}_{cor\_mse} 10 40 80 80 1
TABLE VI: Test Accuracy (%) of experiments on Clothing1M. Each row corresponds a different system architecture.
Acc (%)
Ours (default) 75.17
Ours, w/o noisy branch 73.91
Ours + Mixup, w/o noisy branch 73.87
Ours + Mixup + Soft Label, w/o noisy branch 73.90

Finally, Tables VII, VIII, and VI show that by removing the noisy label branch, by changing the loss functions, or by using soft-labels, the proposed system can become suitable for correcting CIFAR-10/CIFAR-100 datasets with man-made random label noises. However, such modifications degrade our system performance for real-world applications, e.g. Clothing1M, whose data are contaminated by confusing label noises, as shown in Fig. VI. Hence, we finally choose to adopt our default architecture and loss settings because this combination is most beneficial for real-world applications.

TABLE VII: Test Accuracy (%) of experiments on Cifar-10 with symmetric and asymmetric noises. For symmetric noises, the note rate was set to be 20%, 50%, 80%, and 90% in turn, whereas the noise rate of the experiment on asymmetric noise is fixed to be 40%. Note that values in Last rows are the mean test accuracy values derived by models of last 10 training epochs.
Noise Level (%) 20 50 80 90 asym
Cross Entropy Best 86.8 79.4 62.9 42.7 85.0
Last 82.7 57.9 26.1 16.8 72.3
Mixup [39] Best 95.6 87.1 71.6 52.2
Last 92.3 77.6 46.7 43.9
P-correction [40] Best 92.4 89.1 77.5 58.9 88.5
Last 92.0 88.7 76.5 58.2 88.3
Meta-Learning [15] Best 92.9 89.3 77.4 58.7 89.2
Last 92.0 88.8 76.1 58.3 88.6
M-correction [21] Best 94.0 92.0 86.8 69.1 87.4
Last 93.8 91.9 86.8 68.7 86.3
DivideMix [16] Best 96.1 94.6 93.2 76.0 93.4
Last 95.7 94.4 92.9 75.4 92.2
Ours Best 90.1 81.8 64.4 46.2 88.6
(default) Last 88.8 81.6 56.1 39.8 88.4
Ours Best 93.3 91.1 77.1 52.2 91.2
w/o noisy branch Last 93.1 90.8 76.5 45.6 90.9
Ours w/o noisy branch Best 93.9 91.8 81.3 61.6 92.1
+ Mixup Last 93.4 91.2 80.4 59.8 91.8
Ours w/o noisy branch Best 94.2 93.9 92.5 76.2 92.3
+ Mixup + Soft Label Last 94.0 93.7 92.3 73.6 92.1
TABLE VIII: Test Accuracy (%) of experiments on Cifar-100 with symmetric noise. Note that 20, 50, 80, and 90 denote noise rates; and, values in Last rows are the mean test accuracy values derived by models of last 10 training epochs.
Noise Level (%) 20 50 80 90
Cross Entropy Best 62.0 46.7 19.9 10.1
Last 61.8 37.3 8.8 3.5
Mixup [39] Best 67.8 57.3 30.8 14.6
Last 66.0 46.6 17.6 8.1
P-correction [40] Best 69.4 57.5 31.1 15.3
Last 68.1 56.4 20.7 8.8
Meta-Learning [15] Best 68.5 59.2 42.4 19.5
Last 67.7 58.0 40.1 14.3
M-correction [21] Best 73.9 66.1 48.2 24.3
Last 73.4 65.4 47.6 20.5
DivideMix [16] Best 77.3 74.6 60.2 31.5
Last 76.9 74.2 59.6 31.0
Ours Best 65.6 50.3 23.1 10.4
(default) Last 65.0 49.8 21.6 8.2
Ours Best 74.1 65.9 40.8 15.2
w/o noisy branch Last 73.8 65.5 40.3 13.2
Ours w/o noisy branch Best 76.1 70.2 45.3 21.8
+ Mixup Last 75.9 69.8 44.6 19.9
Ours w/o noisy branch Best 76.6 73.9 57.2 28.6
+ Mixup + Soft Label Last 76.4 73.1 56.9 28.4

V Conclusion

We proposed an ensemble learning method for noisy label correction by using local patches of feature manifold. Based on the facts that most real-world noisy labels come from confusing label noises, we have proposed a nearest-neighbor-based data splitting scheme to tackle noisy labels locally-concentrated near decision boundaries by capturing the local structures of data manifolds . Hence, each sub-model can learn a coarse representation of the feature manifold, and only a few sub-models will be affected by locally-dense noisy labels. We have also proposed a multi-graph label propagation algorithm to derive reasonable label correction suggestions, and a confidence-guided label correction mechanism to determine proper label from the correction suggestions via majority decision. Extensive experiments and in-depth analyses on real-world datasets evidently demonstrate the superiority of the proposed method over existing noisy label correction strategies.

References

  • [1] B. Settles, “Active learning literature survey,” University of Wisconsin-Madison Department of Computer Sciences, Tech. Rep., 2009.
  • [2] T.-Y. Lin, M. Maire, S. Belongie, J. Hays, P. Perona, D. Ramanan, P. Dollár, and C. L. Zitnick, “Microsoft coco: Common objects in context,” in Proc. European Conf. Comput. Vis., 2014, pp. 740–755.
  • [3] J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, and L. Fei-Fei, “Imagenet: A large-scale hierarchical image database,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., 2009, pp. 248–255.
  • [4] Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner, “Gradient-based learning applied to document recognition,” Proc. IEEE, vol. 86, no. 11, pp. 2278–2324, 1998.
  • [5] A. Krizhevsky, Learning Multiple Layers of Features from Tiny Images.   Master Thesis, Dept. Comput. Sci., Univ. Toronto, 2009.
  • [6] G. Patrini, A. Rozza, A. Krishna Menon, R. Nock, and L. Qu, “Making deep neural networks robust to label noise: A loss correction approach,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., 2017, pp. 1944–1952.
  • [7] D. Hendrycks, M. Mazeika, D. Wilson, and K. Gimpel, “Using trusted data to train deep networks on labels corrupted by severe noise,” in Proc. Adv. Neural Inf. Process. Syst., 2018, pp. 10 456–10 465.
  • [8] X. Xia, T. Liu, N. Wang, B. Han, C. Gong, G. Niu, and M. Sugiyama, “Are anchor points really indispensable in label-noise learning?” in Proc. Adv. Neural Inf. Process. Syst., 2019, pp. 6838–6849.
  • [9] K. Sharma, P. Donmez, E. Luo, Y. Liu, and I. Z. Yalniz, “Noiserank: Unsupervised label noise reduction with dependence models,” arXiv preprint arXiv:2003.06729, 2020.
  • [10] J. Han, P. Luo, and X. Wang, “Deep self-learning from noisy labels,” in Proc. IEEE/CVF Int. Conf. Comput. Vis., 2019, pp. 5138–5147.
  • [11] D. Tanaka, D. Ikami, T. Yamasaki, and K. Aizawa, “Joint optimization framework for learning with noisy labels,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., 2018, pp. 5552–5560.
  • [12] Z. Zhang and M. Sabuncu, “Generalized cross entropy loss for training deep neural networks with noisy labels,” in Proc. Adv. Neural Inf. Process. Syst., 2018, pp. 8778–8788.
  • [13] H. Wei, L. Feng, X. Chen, and B. An, “Combating noisy labels by agreement: A joint training method with co-regularization,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., 2020, pp. 13 726–13 735.
  • [14] F. Chen, R. Ji, J. Su, D. Cao, and Y. Gao, “Predicting microblog sentiments via weakly supervised multimodal deep learning,” IEEE Trans. Multimedia, vol. 20, pp. 997–1007, 2017.
  • [15] J. Li, Y. Wong, Q. Zhao, and M. S. Kankanhalli, “Learning to learn from noisy labeled data,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., 2019, pp. 5051–5059.
  • [16] J. Li, R. Socher, and S. C. Hoi, “Dividemix: Learning with noisy labels as semi-supervised learning,” arXiv preprint arXiv:2002.07394, 2020.
  • [17] D. T. Nguyen, C. K. Mummadi, T. P. N. Ngo, T. H. P. Nguyen, L. Beggel, and T. Brox, “Self: Learning to filter noisy labels with self-ensembling,” arXiv preprint arXiv:1910.01842, 2019.
  • [18] S. Laine and T. Aila, “Temporal ensembling for semi-supervised learning,” arXiv preprint arXiv:1610.02242, 2016.
  • [19] A. Tarvainen and H. Valpola, “Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results,” in Proc. Adv. Neural Inf. Process. Syst., 2017, pp. 1195–1204.
  • [20] R. Wang, T. Liu, and D. Tao, “Multiclass learning with partially corrupted labels,” IEEE Trans. Neural Netw. Learn. Syst., vol. 29, no. 6, pp. 2568–2580, 2017.
  • [21] E. Arazo, D. Ortego, P. Albert, N. O’Connor, and K. McGuinness, “Unsupervised label noise modeling and loss correction,” in Proc. Int. Conf. Mach. Learn., 2019, pp. 312–321.
  • [22] A. Ghosh, H. Kumar, and P. Sastry, “Robust loss functions under label noise for deep neural networks,” arXiv preprint arXiv:1712.09482, 2017.
  • [23] Y. Wang, X. Ma, Z. Chen, Y. Luo, J. Yi, and J. Bailey, “Symmetric cross entropy for robust learning with noisy labels,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., 2019, pp. 322–330.
  • [24] Y. Yan, Z. Xu, I. W. Tsang, G. Long, and Y. Yang, “Robust semi-supervised learning through label aggregation,” in Proc. AAAI Conf. Artificial Intell., 2016.
  • [25] D. Zhou, O. Bousquet, T. N. Lal, J. Weston, and B. Schölkopf, “Learning with local and global consistency,” in Proc. Adv. Neural Inf. Process. Syst., 2004, pp. 321–328.
  • [26] A. Iscen, G. Tolias, Y. Avrithis, and O. Chum, “Label propagation for deep semi-supervised learning,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., 2019, pp. 5070–5079.
  • [27] A. Erdem and M. Pelillo, “Graph transduction as a noncooperative game,” Neural Comput., vol. 24, no. 3, pp. 700–723, 2012.
  • [28] M. Douze, A. Szlam, B. Hariharan, and H. Jégou, “Low-shot learning with large-scale diffusion,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., 2018, pp. 3349–3358.
  • [29] M. Bontonou, C. Lassance, G. B. Hacene, V. Gripon, J. Tang, and A. Ortega, “Introducing graph smoothness loss for training deep learning architectures,” arXiv preprint arXiv:1905.00301, 2019.
  • [30] R. Wang, X.-J. Wu, and J. Kittler, “Graph embedding multi-kernel metric learning for image set classification with grassmannian manifold-valued features,” IEEE Trans. Multimedia, vol. 23, pp. 228–242, 2020.
  • [31] F. Schroff, D. Kalenichenko, and J. Philbin, “Facenet: A unified embedding for face recognition and clustering,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., 2015, pp. 815–823.
  • [32] T. Xiao, T. Xia, Y. Yang, C. Huang, and X. Wang, “Learning from massive noisy labeled data for image classification,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., 2015, pp. 2691–2699.
  • [33] K.-H. Lee, X. He, L. Zhang, and L. Yang, “Cleannet: Transfer learning for scalable image classifier training with label noise,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., 2018, pp. 5447–5456.
  • [34] L. Bossard, M. Guillaumin, and L. Van Gool, “Food-101–mining discriminative components with random forests,” in Proc. European Conf. Comput. Vis.   Springer, 2014, pp. 446–461.
  • [35] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., 2016, pp. 770–778.
  • [36] W. Zhang, Y. Wang, and Y. Qiao, “Metacleaner: Learning to hallucinate clean representations for noisy-labeled visual recognition,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., 2019, pp. 7373–7382.
  • [37] S. Liu, J. Niles-Weed, N. Razavian, and C. Fernandez-Granda, “Early-learning regularization prevents memorization of noisy labels,” arXiv preprint arXiv:2007.00151, 2020.
  • [38] Z. Zhang, H. Zhang, S. O. Arik, H. Lee, and T. Pfister, “Distilling effective supervision from severe label noise,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., 2020, pp. 9294–9303.
  • [39] H. Zhang, M. Cisse, Y. N. Dauphin, and D. Lopez-Paz, “mixup: Beyond empirical risk minimization,” arXiv preprint arXiv:1710.09412, 2017.
  • [40] K. Yi and J. Wu, “Probabilistic end-to-end noise correction for learning with noisy labels,” in Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., 2019, pp. 7017–7025.