Federated Learning with Instance-Dependent Noisy Label
Abstract
Federated learning (FL) with noisy labels poses a significant challenge. Existing methods designed for handling noisy labels in centralized learning tend to lose their effectiveness in the FL setting, mainly due to the small dataset size and the heterogeneity of client data. While some attempts have been made to tackle FL with noisy labels, they primarily focused on scenarios involving class-conditional noise. In this paper, we study the more challenging and practical issue of instance-dependent noise (IDN) in FL. We introduce a novel algorithm called FedBeat (Federated Learning with Bayesian Ensemble-Assisted Transition Matrix Estimation). FedBeat aims to build a global statistically consistent classifier using the IDN transition matrix (IDNTM), which encompasses three synergistic steps: (1) A federated data extraction step that constructs a weak global model and extracts high-confidence data using a Bayesian model ensemble method. (2) A federated transition matrix estimation step in which clients collaboratively train an IDNTM estimation network based on the extracted data. (3) A federated classifier correction step that enhances the global model’s performance by training it using a loss function tailored for noisy labels, leveraging the IDNTM. Experiments conducted on CIFAR-10 and SVHN verify that the proposed method significantly outperforms state-of-the-art methods.
Index Terms— Federated Learning, Instance-Dependent Noise
1 Introduction
Advancements in mobile edge devices have led to the collection of massive datasets, sparking significant interest in utilizing this data for training deep learning models. However, leveraging such mobile device data for deep learning poses a complex challenge due to privacy concerns. To address this issue, Federated Learning (FL) has emerged as a promising research area that enables clients with decentralized data to collaborate and train a shared model under the guidance of a central server. Introduced by [1], FedAvg has been instrumental in reducing server-client communication while preserving data privacy through local stochastic gradient descent. Various researchers have extended the FedAvg framework to other real-world scenarios. For instance, some studies [2, 3, 4] address the challenge of heterogeneous client data, while others [5, 6, 7] focus on mitigating the effects of partial client participation.

While significant progress has been made in adapting FL to realistic settings, the issue of label quality associated with data collected by edge devices has often been overlooked. Unlike traditional centralized learning, where expert-provided labels are available for each data sample, ensuring label quality in FL is considerably more challenging. Many users lack expertise, making it unreasonable to expect all data samples to have ground truth labels. Consequently, addressing the problem of noisy labels in FL is of paramount importance.
In centralized settings, extensive research has been conducted on learning with noisy data. Some researchers [8, 9, 10] have proposed noise-cleaning approaches that aim to detect noisy samples and train with clean ones. Others [11, 12, 13, 14] have estimated transition matrices to build statistically consistent classifiers, reducing the impact of noisy labels by inferring clean distributions based on the transition matrix and noisy class-posterior probabilities. However, unlike the sufficient dataset size in centralized learning, the dataset stored on each client in FL is relatively small, making it impossible for individual clients to train high-performance models without collaboration. Moreover, the heterogeneity of clients’ data distributions complicates the collaboration process. Only a few studies [15, 16, 17] have attempted to address FL with noisy labels. However, these methods primarily focus on class-conditional noise (CCN), where each instance from one class has a fixed probability of being assigned to another. In reality, noisy labels are often instance-based, indicating that a data sample is more likely to be mislabeled due to its content rather than the class label it belongs to. This form of noise is commonly referred to as instance-dependent noise (IDN).
In this paper, we focus on addressing IDN in the context of FL, as it is more relevant for handling real-world noisy scenarios [11] and a previously unexplored problem. Our key idea relies on constructing an instance-dependent noise transition matrix (IDNTM) estimation network to establish the correspondence between the clean label and the noisy label, which subsequently enables the training of a statistically consistent classifier. Training such an IDNTM estimation network would typically be straightforward if we had access to a dataset containing both noisy and clean labels. However, since the training datasets of the clients only comprise noisy labels, training the IDNTM estimation network becomes a challenging task, particularly in the FL setting. To tackle this challenge, we introduce a new algorithm called FedBeat, which stands for Federated Learning with Bayesian ensemble-assisted transition matrix estimation. FedBeat encompasses three synergistic steps, as illustrates in Figure 1:
1. Federated Data Extraction: This step is used to extract a dataset with high-confidence pseudo-labels to prepare for the training of the IDNTM estimation network. To this end, a weak global classification model is first trained on the noisy labels, which is then used to generate pseudo-labels for the training data samples. To increase the amount of high-confidence pseudo-labels, a Bayesian model ensemble approach is employed to generate pseudo-labels.
2. Federated Transition Matrix Estimation: The extracted dataset with both high-confidence pseudo-label and noisy labels are used to train the IDNTM estimation network in federated way. When presented with a new data sample, the IDNTM estimation network generates the corresponding IDNTM, where elements represent the transition probabilities from ground-truth labels to noisy labels.
3. Federated Classifier Correction: With the IDNTM estimation network, we make a correction to the weak global classifier for improvement. In this step, however, the loss function incorporates the output of the IDNTM on the predicted labels, and measures the statistical difference between the actual noisy label and this output.
2 Problem Formulation
2.1 Learning Objective
We start with the standard -class classification problem in FL, involving clients and a central server. For each client , it contains the local dataset where is denoted the size of dataset, i.e. , and . Each client has a local model and a model output , which is the class probability distribution (a.k.a., confidence score vector). The global objective is defined as follows: , where is the total dataset size. Denote as the one-hot form of , thus is the local objective function of client , in which represents the loss function, and is the -th sample of client .
2.2 Noise Model
In the classic FL setting, it is assumed that training samples in each client are labeled with ground-truth. However, this assumption is overly idealistic in real world. In this paper, to capture label noise in real-world data, we consider the IDN model, which assumes that the noise rate is not constant and varies based on each sample’s features, rather than varying on each class. Let be the overall distribution of pairwise random variables , where denotes the variables of all training samples stored at clients, and represents the variable of corresponding labels. It can be treated as if each data sample in each client is independently drawn according to .
Correspondingly, define as the overall distribution of pairwise random variables , where represents the variable of corresponding noisy labels. The actual dataset held by each client is denoted as , which depends on the noisy distribution . To establish the bridge between the clean distribution and the noisy distribution and build a statistically consistent classifier, the IDNTM (i.e. ) is introduced. Here, we define as the -th element of IDNTM, representing the transition probability of a sample with a clean label transitioning to a noisy label . Thus the noisy class-posterior probability can be inferred by the IDNTM and the clean class-posterior probability as follows:
(1) |
3 Proposed Method
To address FL with instance-dependent noisy labels, we propose FedBeat, which comprises three synergistic steps as follows.
3.1 Federated Data Extraction
As mentioned earlier, we need an IDNTM for each instance transiting from a noisy label distribution to a clean one. Typically, given a subset of instances that have both noisy and clean labels, training an IDNTM estimation network is a relatively straightforward task. However, in an FL context, clients face challenges in accurately labeling each instance, rendering it infeasible to directly build an estimation network for the IDNTM.
Therefore, our first step primarily focuses on training a weak global model from the noisy distribution by incorporating the noisy labels during training to help isolate high-confidence clean data. This initial training lasts for a total of rounds and each client performs local steps during each round. In each round, FedBeat requires the server to disseminate the global model to all clients, which will then fully participate in training and uploading their local models. The server then aggregates the received local models to produce an updated global model by using approaches such as FedAvg [1] or FedProx [4]. We denote the local model uploaded by client during the final communication round as , and the global model as .
Once the weak global model is obtained, it is returned to the clients to generate peudo-labels on the training data samples. However, rather than returning only , we employ a Bayesian model ensemble [18] approach to boost the classification accuracy, thereby extracting a greater amount of high-confidence data. Particularly, after denoting the average model , the server then calculates the standard deviation of the local models:
(2) |
The standard deviation is also returned to each client. Here the square operation is element-wise. With both and , each client samples ensemble models . It is important to note that the ensemble models can vary among clients. Such an operation is to derive smoother predictive distributions and minimize the risk of overconfidence. Specifically, each client computes the average outputs (i.e., the average confidence score vector) for each instance using its ensemble models as follows:
(3) |
where . Let be the -th elements of , then the pseudo-label of is generated according to , with its associated confidence score being . Using the predetermined confidence threshold value (a hyper-parameter), client filters its local dataset to construct a extracted dataset containing feature inputs (e.g., images), noisy labels, and pseudo-labels:
(4) |
where we denote the size of the extracted dataset as .
With a proper choice of , the pseudo-label accuracy is high (see Table 5). However, the number of extracted instances can be small compared to the overall dataset size. This limited number presents two challenges: First, training an accurate classification model solely with these instances is impractical. Second, the limited number of filtered instances per client hinders the individual training of an effective IDNTM.
3.2 Federated Transition Matrix Estimation
After extracting data with high-confidence pseudo-labels, an IDNTM estimation network is trained using the data extracted on each client. When provided with an input example , the estimation network generates the IDNTM , which contains the probability of transitioning from the clean distribution to the noisy distribution. The training is conducted for rounds, and at each round , a subset of clients is selected to participate in the training. For each selected client , the current version of the estimation network, , is first received, and it is then set as the local network . The client then performs steps of local training based on a loss function .
(5) |
where denotes the local step, the subscript denotes the -th element of the vector which is usually corresponding to class , is the estimated IDNTM for the example generated by the local network . Here means we pick the -th element from the -dimensional column vector transposed by the row vector. It is noteworthy that both the noisy label and the pseudo-label are hard labels, and in the one-hot form of and respectively. can be viewed as a traditional cross-entropy loss function which treats the noisy label as the ground-truth and as the predicted output. Such a loss function ensures that the estimated IDNTM, which is generated based on the features of each instance, can successfully map the pseudo-label to the noisy label. Once the selected client completes their local updates, both the updated parameter and the size of their extracted dataset are sent back to the server for further processing. The server then updates the global as follows:
(6) |
where is the sum of the number of all extracted examples. Note that it is crucial to upload for aggregation on the server instead of using directly as in the federated data extraction step. Though the ratio of aggregation weights among the clients for this process is similar to the previous step in most of IID settings, IID with a high noise rate or non-IID case either causes a huge imbalance in the amount of data extracted. Upon completing rounds, we obtain the final IDNTM estimation network , denoted as , which generates the IDNTM that approximates the ground-truth transition matrix.
3.3 Federated Classifier Correction
Once the estimation network has been trained, it can be employed to correct the weak global classification model derived from the data extraction step. Since is trained on the noisy dataset , simply continuing to train it in the same way as before would result in overfitting for low noise rates or difficulty in converging for high noise rates. Furthermore, the output distribution would still be based on the noisy dataset.
To obtain a strong classification model, we perform federated classifier correction by conducting communication rounds for global model aggregation, with each round consisting of local steps for fine-tuning the local models similar to the data extraction step. Thus, to minimize the impact of noisy labels, we use the IDNTM estimation network to map the original predictions to noisy outputs. We define the classifier correction loss function as follows:
(7) |
where represents the local model on client during the -th epoch of round , is the adjusted prediction obtained by mapping the original prediction through the IDNTM estimated by the network .
After calculating the loss function for all clients, we aggregate the local models using a global aggregation procedure, similar to the one employed in the federated data extraction step:
(8) |
We argue that the federated classifier correction step works by utilizing the IDNTM in reverse. Specifically, the IDNTM produced by adjusts the clean label distribution to approximate the noisy label distribution. Therefore, we can adjust the output distribution in reverse to approximate a clean label distribution by ensuring that the adjusted output closely matches the noisy label distribution.
4 Experiments
4.1 Experiment Setup
Baseline Methods. Given that no previous work has considered IDN in the FL setting, we incorporate several baseline methods to demonstrate the superior performance of our proposed method. We consider three types of baselines. First, we examine some traditional FL methods that learn under the assumption of clean labels (e.g. FedAvg [1] and FedProx [4]). Then, we look at a scenario where each client individually applies the state-of-the-art IDN method developed in centralized learning (BLTM [12]). Lastly, we consider FedCorr [15] which is a state-of-the-art FL approach solving CCN.
Implementation Details. Our experiments encompass a range of noise levels (IDN rates: 30%, and 50%) across the CIFAR-10 and SVHN datasets. Specific details vary based on the dataset: CIFAR-10:We consider 10 clients, each having 5000 balanced training data samples in the IID setting. In the non-IID scenario, the sample count is scaled down to 60% of the original. SVHN: In the IID setting, we conisder 10 clients, where each client has 4659 balanced training data samples. In the non-IID setting, the number of samples is reduced to 50% of the original. We use the Dirichlet distribution () to simulate the non-IID case among clients.
For the initial step, we employ ResNet-34 [19] for CIFAR-10 and ResNet-18 [19] for SVHN as the backbone model with an initial learning rate of 0.01. We warm-up the classification network for a total of 15 rounds, with 10 clients participating in each round, and each client performing 2 local epochs. We then set the number of model ensembles, denoted by , to 10 and the threshold to 0.65. For the second step, which involves training the estimation network, we utilize ResNet-34 for CIFAR-10 and ResNet-18 for SVHN with a 100-dimensional output as the backbone model with an initial learning rate of 0.01. We conduct rounds, set to 85, with 10 clients participating in each round. For the final step, we use the model obtained from the final round of the first step as the backbone model with an initial learning rate of 0.0005. We conduct rounds, set to 50, with 10 clients participating in each round.
IID | non-IID () | |||
---|---|---|---|---|
IDN-30% | IDN-50% | IDN-30% | IDN-50% | |
FedAvg | 73.10 0.97 | 61.99 1.82 | 61.41 3.26 | 47.64 1.56 |
FedProx | 71.97 1.16 | 58.66 0.96 | 61.57 1.65 | 47.74 0.95 |
BLTM-local | 45.75 0.55 | 36.25 0.58 | 57.64 2.01 | 49.13 1.27 |
FedCorr | 65.90 1.50 | 54.41 0.89 | 62.23 2.34 | 50.46 2.29 |
FedBeat(ours) | 81.58 0.24 | 74.51 2.71 | 72.61 0.31 | 58.44 3.53 |
IID | non-IID () | |||
---|---|---|---|---|
IDN-30% | IDN-50% | IDN-30% | IDN-50% | |
FedAvg | 88.35 0.91 | 74.24 0.24 | 83.79 0.47 | 61.86 2.98 |
FedProx | 89.46 0.23 | 71.81 2.34 | 84.64 0.21 | 64.06 1.13 |
BLTM-local | 71.14 1.39 | 52.26 0.67 | 70.44 1.34 | 57.13 2.78 |
FedCorr | 86.18 3.52 | 70.56 4.63 | 81.20 1.61 | 59.37 3.15 |
FedBeat(ours) | 94.59 0.25 | 87.97 2.90 | 92.59 0.40 | 75.26 2.89 |
4.2 Comparison with Baseline Methods
We first evaluate FedBeat against the baseline methods at different noise levels under the IID setting. The results in Tables 1 and 2 from both datasets indicate that our method significantly surpasses the performance of the baselines. Specifically, on CIFAR-10, FedBeat outperforms the baseline methods by at least in the IDN- scenario and the performance gap widens to in the IDN- scenario. Similar observations can be made from the results on the SVHN dataset. Note that BLTM-local exhibits the worst test accuracy performance under various noise levels. This highlights that directly applying IDN methods developed for centralized learning without suitable cooperation may not yield the desired results in the FL setting. The lackluster performance of FedCorr and BLTM-local emphasizes the effectiveness of our proposed method in FL.
We also perform experiments under the non-IID setting. The results in Tables 1 and 2 show FedBeat, consistently outperforms the baseline methods. Except for BLTM-local, all methods demonstrate a drop in test accuracy on both datasets compared to their performance under IID settings. The stability of BLTM-local’s performance can be attributed to its design, as it does not involve any inter-client cooperation.
Further, we conducted additional non-IID experiments with varying Dirichlet parameters, while keeping the IDN rate fixed at 30%. These results, presented in Table 3, validate that our method consistently performs best under different degrees of non-IID settings. Specifically, FedBeat outperforms all baselines by at least 9.89% with , and by at least 7.49% with .
FedAvg | 81.46 1.66 | 83.79 0.47 | 84.98 0.72 |
---|---|---|---|
FedProx | 82.23 1.42 | 84.64 0.21 | 84.18 1.01 |
BLTM-local | 73.43 0.93 | 70.44 1.34 | 60.08 0.68 |
FedCorr | 76.42 1.88 | 81.20 1.61 | 82.61 2.87 |
FedBeat(ours) | 92.12 0.49 | 92.59 0.40 | 92.47 0.31 |
IDN-30% | IDN-50% | |
---|---|---|
w/o ensemble | 92.57% / 1713 / 90.15 0.30 | 74.24% / 875 / 71.12 0.84 |
w/ ensemble | 96.01% / 1736 / 92.59 0.40 | 83.75% / 809 / 75.26 2.89 |
IID | non-IID () | |||
---|---|---|---|---|
IDN-30% | IDN-50% | IDN-30% | IDN-50% | |
95.49% / 4112 | 78.81% / 3917 | 92.28% / 2008 | 70.05% / 1876 | |
97.73% / 3643 | 90.56% / 1800 | 96.01% / 1736 | 83.75% / 809 | |
99.21% / 2362 | 94.77% / 417 | 97.79% / 1303 | 85.56% / 250 |
4.3 Impact of Model Ensemble
We compare our method with an ablated variant where the first step uses only the average of local models to extract instances with clean pseudo labels. We set the same extraction threshold and ablated variant to isolate the differences between them. As shown in Table 4, the results (the accuracy of pseudo labeling / the number of extracted data / the final classification accuracy) demonstrate the use of Bayesian model ensembles allows us to filter a similar number of instances with a higher accuracy. This improved accuracy in pseudo-labeling aids the second step in training a more reliable IDNTM estimation network, thereby enhancing the final model’s performance.
4.4 Impact of Threshold
We further examine the impact of threshold . We present the results (the accuracy of pseudo labeling / the number of extracted data) under different values in Table 5. If is set too high, the requirement for confidence scores to pass the threshold and be included in the dataset becomes stringent. This could result in too few samples being filtered into the dataset, despite a high correct rate of pseudo-labeling. Conversely, setting a low can result in an excessive number of instances with incorrect pseudo-labels being included in the extracted dataset. Both an excessively small dataset (even with a high accuracy) and a low correct rate (even with a large dataset) can impair the learning performance of the transition network. Therefore, we suggest using a moderate threshold , which ensures a sufficient number of accurate extracted data.
5 Conclusion
In this paper, we studied FL with IDN labels. We introduced a novel algorithm called FedBeat, whose key idea relies on training an IDNTM estimation network to generate the IDNTM that maps ground-truth labels to their corresponding noisy labels. By integrating the IDNTM into the federated training procedure, we aim to enhance the alignment between the ground-truth labels and the model’s predicted outputs, ultimately resulting in a global statistically consistent classifier and improved model accuracy. We evaluated the effectiveness of FedBeat through extensive experiments, demonstrating its superiority over state-of-the-art FL algorithms.
References
- [1] Brendan McMahan, Eider Moore, Daniel Ramage, Seth Hampson, and Blaise Aguera y Arcas, “Communication-efficient learning of deep networks from decentralized data,” in Artificial intelligence and statistics. PMLR, 2017, pp. 1273–1282.
- [2] Yue Zhao, Meng Li, Liangzhen Lai, Naveen Suda, Damon Civin, and Vikas Chandra, “Federated learning with non-iid data,” arXiv preprint arXiv:1806.00582, 2018.
- [3] Felix Sattler, Simon Wiedemann, Klaus-Robert Müller, and Wojciech Samek, “Robust and communication-efficient federated learning from non-iid data,” IEEE transactions on neural networks and learning systems, vol. 31, no. 9, pp. 3400–3413, 2019.
- [4] Tian Li, Anit Kumar Sahu, Manzil Zaheer, Maziar Sanjabi, Ameet Talwalkar, and Virginia Smith, “Federated optimization in heterogeneous networks,” Proceedings of Machine learning and systems, vol. 2, pp. 429–450, 2020.
- [5] Yae Jee Cho, Jianyu Wang, and Gauri Joshi, “Client selection in federated learning: Convergence analysis and power-of-choice selection strategies,” arXiv preprint arXiv:2010.01243, 2020.
- [6] Sai Praneeth Karimireddy, Satyen Kale, Mehryar Mohri, Sashank Reddi, Sebastian Stich, and Ananda Theertha Suresh, “Scaffold: Stochastic controlled averaging for federated learning,” in International Conference on Machine Learning. PMLR, 2020, pp. 5132–5143.
- [7] Shiqiang Wang and Mingyue Ji, “A unified analysis of federated learning with arbitrary client participation,” arXiv preprint arXiv:2205.13648, 2022.
- [8] Julián Luengo, Seong-O Shim, Saleh Alshomrani, Abdulrahman Altalhi, and Francisco Herrera, “Cnc-nos: Class noise cleaning by ensemble filtering and noise scoring,” Knowledge-Based Systems, vol. 140, pp. 27–49, 2018.
- [9] Hongxin Wei, Lei Feng, Xiangyu Chen, and Bo An, “Combating noisy labels by agreement: A joint training method with co-regularization,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2020, pp. 13726–13735.
- [10] Junnan Li, Yongkang Wong, Qi Zhao, and Mohan S Kankanhalli, “Learning to learn from noisy labeled data,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 5051–5059.
- [11] De Cheng, Tongliang Liu, Yixiong Ning, Nannan Wang, Bo Han, Gang Niu, Xinbo Gao, and Masashi Sugiyama, “Instance-dependent label-noise learning with manifold-regularized transition matrix estimation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 16630–16639.
- [12] Shuo Yang, Erkun Yang, Bo Han, Yang Liu, Min Xu, Gang Niu, and Tongliang Liu, “Estimating instance-dependent bayes-label transition matrix using a deep neural network,” in International Conference on Machine Learning. PMLR, 2022, pp. 25302–25312.
- [13] Xiaobo Xia, Tongliang Liu, Bo Han, Nannan Wang, Mingming Gong, Haifeng Liu, Gang Niu, Dacheng Tao, and Masashi Sugiyama, “Part-dependent label noise: Towards instance-dependent label noise,” Advances in Neural Information Processing Systems, vol. 33, pp. 7597–7610, 2020.
- [14] Pengfei Chen, Junjie Ye, Guangyong Chen, Jingwei Zhao, and Pheng-Ann Heng, “Beyond class-conditional assumption: A primary attempt to combat instance-dependent label noise,” in Proceedings of the AAAI Conference on Artificial Intelligence, 2021, vol. 35, pp. 11442–11450.
- [15] Jingyi Xu, Zihan Chen, Tony QS Quek, and Kai Fong Ernest Chong, “Fedcorr: Multi-stage federated learning for label noise correction,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 10184–10193.
- [16] Seunghan Yang, Hyoungseob Park, Junyoung Byun, and Changick Kim, “Robust federated learning with noisy labels,” IEEE Intelligent Systems, vol. 37, no. 2, pp. 35–43, 2022.
- [17] Vasileios Tsouvalas, Aaqib Saeed, Tanir Özçelebi, and Nirvana Meratnia, “Federated learning with noisy labels: Achieving generalization in the face of label noise,” in First Workshop on Interpolation Regularizers and Beyond at NeurIPS 2022, 2022.
- [18] Hong-You Chen and Wei-Lun Chao, “Fedbe: Making bayesian model ensemble applicable to federated learning,” arXiv preprint arXiv:2009.01974, 2020.
- [19] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770–778.