Semi-Supervised Federated Learning with non-IID Data: Algorithm and System Design
Abstract
Federated Learning (FL) allows edge devices (or clients) to keep data locally while simultaneously training a shared high-quality global model. However, current research is generally based on an assumption that the training data of local clients have ground-truth. Furthermore, FL faces the challenge of statistical heterogeneity, i.e., the distribution of the client’s local training data is non-independent identically distributed (non-IID). In this paper, we present a robust semi-supervised FL system design, where the system aims to solve the problem of data availability and non-IID in FL. In particular, this paper focuses on studying the labels-at-server scenario where there is only a limited amount of labeled data on the server and only unlabeled data on the clients. In our system design, we propose a novel method to tackle the problems, which we refer to as Federated Mixing (). improves the naive combination of FL and semi-supervised learning methods and designs parameter decomposition strategies for disjointed learning of labeled, unlabeled data, and global models. To alleviate the non-IID problem, we propose a novel aggregation rule based on the frequency of the client’s participation in training, namely the aggregation algorithm, which can adjust the weight of the corresponding local model according to this frequency. Extensive evaluations conducted on CIFAR-10 dataset show that the performance of our proposed method is significantly better than those of the current baseline. It is worth noting that our system is robust to different non-IID levels of client data.
Index Terms:
Federated Learning, Semi-supervised Learning, non-IID, Aggregation AlgorithmI Introduction
Federated Learning (FL) [1, 2] is a distributed machine learning paradigm that allows multiple edge devices (or clients) to cooperatively train a shared global model [3, 4, 5]. The most obvious difference between FL and traditional distributed machine learning is that clients can privately access local training data without sharing data with cloud centers [6, 7, 8]. However, the current mainstream work is based on an unrealistic assumption: the training data of local clients have ground-truth [9]. In our daily lives, it is not common for each client to have rich labeled data. For example, in the early stage of COVID-19 epidemic, community hospitals without enough labeled data may not be able to train a high-precision pathophoresis prediction model. On the other hand, in most cases, putting together a properly labeled dataset for a given FL task is a time-consuming, expensive, and complicated endeavor [9]. Therefore, it is challenging to train a high-quality global model in the real scenario of a lack of labeled data.
In the face of the above challenges, recent works [9, 10, 11, 12, 13, 14] study how to design a semi-supervised FL (SSFL) system that can efficiently integrate semi-supervised learning into FL techniques. For example, Jeong et al. in [12] proposed an SSFL system with a new inter-client consistency loss to achieve this goal. In fact, consistency regularization technique is widely used in semi-supervised learning which keeps the same output that the same data inject two different noises [15, 16]. Furthermore, pseudo-label methods are important for SSFL, where they mainly utilize pseudo-labels whose predicted value is higher than the confidence threshold to achieve high-precision SSFL [9, 10, 17]. However, it is worth noting that there still remain gaps when deploying SSFL in practice.
First, traditional SSFL methods generally introduce semi-supervised techniques (such as consistency loss and pseudo-label) directly into the FL system, which ignores the implicit contribution between iterative updates of the global model. Previous work only focused on how to set pseudo-labels or how to decompose the parameters of labeled and unlabeled data for disjoint learning. In this way, the learned global model will be biased towards labeled data (supervised model) or unlabeled data (unsupervised model) instead of the global model [12]. This implies that we need to observe the implicit effects between iterations of the global model at a fine-grained level.
Second, the non-independent identically distributed (non-IID) of data between clients has always been a key and challenging issue in FL. The reason is that there are too many differences in data distribution, features, and the number of labels between clients, which is not conducive to the convergence of the global model. Currently, many efforts have effectively alleviated the non-IID problem, such as [18] utilized local batch normalization to alleviate the feature shift before average aggregating local models. However, methods such as these add additional computational and communication overhead to the server or client.
In this paper, to address the first issue, we propose the Federated Mixing () algorithm, which performs parameter decomposition of disjointed learning for supervised model (learned on labeled data), unsupervised model (learned on unlabeled data), and global model. In particular, this algorithm analyzes the implicit effects between iterations of the global model in a fine-grained manner. To address the second issue, we propose a novel aggregation rule called Federated Frequency (), which dynamically adjusts the weight of the corresponding local model by recording the training frequency of the client to alleviate the non-IID problem. Furthermore, we introduce the Dirchlet distribution function to simulate the different non-IID level scenario in our experiment. The main contributions of this paper are as follows:
-
•
We present a robust Semi-supervised Federated Learning system design, where the system aims to solve the problem of data availability and non-IID in FL. In our system design, we propose algorithm to improve the naive combination of FL and semi-supervised learning methods.
-
•
We propose a novel aggregation rule called , which dynamically adjusts the weight of the corresponding local model by recording the training frequency of the client to alleviate the non-IID problem.
-
•
We conduct extensive evaluations on CIFAR-10 dataset, which show that the performance of our designed system is 3% higher than the baseline.
II related work
II-A Semi-supervised Federated Learning
Semi-supervised federated learning attempts to use semi-supervised learning techniques [19, 20, 21, 22, 23] to further improve the performance of the FL model in scenarios where there is unlabeled data on the client side [11]. For example, Long et al. in [10] proposed a semi-supervised federated learning (SSFL) system, FedSemi, which unifies the consistency-based semi-supervised learning model [24], dual model [15], and average teacher model [25] to achieve SSFL. The DS-FL system [26] was proposed to solve the communication overhead problem in SSFL. Reference [27] proposes a method to study the distribution of non-IID data, which introduces a probability distance metric to evaluate the difference in client data distribution in SSFL. Different from the literature [9, 10, 26], in this paper, we focus on labels-at-server scenario and also solve the problem of data availability and data heterogeneity in SSFL.
II-B Robust Federated Learning
If the local data set distribution of each client is inconsistent (i.e., non-IID problem) [28, 29, 30, 31, 18, 32, 33], the local objective loss function of the client will be inconsistent with the global objective[34]. In particular, when the model of the local client is updated larger, such a difference will be more obvious. Therefore, we need to design some robust FL systems to solve the above problems. Some studies try to design a robust federated learning algorithm to solve the non-IID problem. For example, FedProx [6] limits the distance between the local model and the global model by introducing an additional regularization term in the local target function to limiting the size of the local model update. However, this method has a disadvantage that each client needs to individually adjust the local regularization term to obtain good model performance. FedNova [35] improved FedAvg in the aggregation phase, which normalized and scaled the model update according to the local training batch of the client. Although previous studies have alleviated the problem of non-IID to some extent, they only evaluated the data distribution at specific non-IID levels and lacked extensive experimental verification for different non-IID scenarios. Therefore, we propose a more comprehensive data distribution and data partition strategy, i.e., we introduce the Dirichlet distribution function to simulate different non-IID levels of client data.
III Preliminaries
III-A Federated Learning
Federated learning solves the problem of data island on the premise of privacy protection. In particular, the FL is a distributed machine learning framework, which requires clients to hold data locally, where these clients coordinate to train a shared global model . In FL, there is a server and clients, each of which holds an IID or non-IID datasets . Specifically, for a training sample on the client side, let be the loss function at the client, where denotes the model’s trainable parameters. Therefore, we let be the loss function at the server. Thus, FL needs to optimize the following objective function at the server:
(1) |
where indicates the relative influence of -th client on the global model. In FL, to minimize the above objective function, the server and clients execute the following steps:
-
•
Step 1, Initialization: The server sends the initialized global model to the selected clients.
-
•
Step 2, Local training: The client uses the local optimizer (e.g., SGD, Adam) on the local dataset to train the received initialization model. Then, each client uploads the local model to the server.
-
•
Step 3, Aggregation: The server collects and uses a certain algorithm (e.g., FedAvg [1]) to aggregate the model updates uploaded by these clients to obtain a new global model, i.e., = + . Then, the server sends the updated global model to all selected clients.
Note that FL repeats the above steps until the global model converges.
III-B Semi-supervised Learning
In the real world, (e.g., financial and medical fields), unlabeled data is easy to gain, while labeled data is often difficult to obtain. Meanwhile, annotating data requires a lot of manpower and material resources. To this end, the researchers proposed a machine learning paradigm namely Semi-supervised Learning [36, 37] to learn a high-precision model on a mixed dataset (part of the data is labeled, and some of the data is unlabeled). Thus, in recent years, semi-supervised learning has become a hot research direction in the field of deep learning. In this section, we introduce a basic assumption and two methods of semi-supervised learning.
Assumption 1
In machine learning, there is a basic assumption that if the feature of two unlabeled samples and are similar, the corresponding model prediction results and are the same [38], i.e., , where is the prediction function.
According to the above assumption, we adopt two common semi-supervised learning methods as follows:
Consistency Regularization: The main idea of this method is that the model prediction results should be the same whether noise is added or not on an unlabeled training sample [15, 36]. We generally use data augmentation (such as image flipping and shifting) methods to add noise to increase the diversity of the dataset. Specifically, for an unlabeled sample in unlabeled dataset and its perturbation form , our goal is to minimize the distance , where is the output of sample on model . The common distance measurement method is Kullback-Leiber (KL) divergence. Thus, we can calculate the consistency loss as follows:
(2) |
where is the total number of unlabeled samples and indicates the model output of unlabeled sample .
Pseudo-label: The pseudo-label method [24] is to utilize some labeled samples to train a model that to set the pseudo labels for unlabeled samples. Previous work generally used sharpening [39] and argmax [24] methods to set pseudo-label, where the former make the distribution of model output extreme of unlabeled samples and the latter change the model output to one-hot of unlabeled samples.

IV Problem Definition
In the SSFL system, there are two essential scenarios of SSFL based on the location of the labeled data. The first scenario considers a conventional case where clients have both labeled and unlabeled data (labels-at-client), and the second scenario considers a more challenging case, where the labeled data is only available at the server (labels-at-server). In particular, in this paper, we consider only the labels-at-server scenario. Next, we give the definition of the problem studied in this paper as follows:
Labels-at-server Scenario: In SSFL, we assume that there is a server and clients, where the server holds a labeled dataset and each client holds a local unlabeled dataset . Thus, in this scenario, for unlabeled training sample , let be the loss function at the client side:
(3) | ||||
where is the number of unlabeled samples, is the data augmentation function (e.g., flip and shift of the unlabeled samples), is the pseudo label of unlabeled sample , and indicates the output of unlabeled sample on model of the -th client. For labeled sample , let be the loss function at the server side:
(4) |
where is the number of labeled samples, and indicates the output of labeled sample on model .
Therefore, the objective function of this scenario in SSFL system is to minimize the following loss function:
(5) |
Note that the whole learning process is similar to the traditional FL system, except that the server not only aggregates the client model parameters but also trains the model with labeled data.
V Algorithm and System Design
V-A Semi-supervised Federated Learning System Design
In our system setting, the server holds a labeled dataset , where indicates the number of labeled samples. For clients, we assume that -th client holds a local unlabeled dataset , where denotes the number of unlabeled samples in the local client. Similar to the traditional FL system, the server and clients in the SSFL are cooperative to train a high-performance global model . The goal of previous work is to optimize the objective function mentioned above, i.e., Equation (5). However, they ignore the implicit contribution between iterations of the global model, which results in the learned global model not being optimal. Inspired by the above facts, we propose an SSFL algorithm called that focuses on the implicit contributions between iterations of the global model in a fine-grained manner. We define the supervised model trained on the labeled dataset as , the unsupervised model trained on the unlabeled dataset as , and the aggregated global model as . Specifically, we design a strategy that assigns three weights and to the unsupervised model , supervised model , and the previous round of global model, respectively. The designed algorithm can capture the implicit relationship between each iteration of the global model in a fine-grained manner. Thus, the steps of our proposed algorithm are as follows:
-
•
Step 1, Initialization: The server randomly selects a certain proportion of () clients from all local clients to send the initialized global model . Note that the global model also remains on the server-side.
-
•
Step 2, Server Training: Unlike FL, in our SSFL system, the server not only aggregates the model uploaded by the clients, but also trains the supervised model (i.e., ) on the labeled dataset . Thus, the server uses the local optimizer on the labeled dataset to train the supervised model . The minimization of the objective function is defined as follows:
(6) where is the hyperparameter, and are from labeled dataset , and means the output of labeled samples on supervised model at -th training round.
-
•
Step 3, Local Training: The -th client utilizes the local unlabeled data to train the received global model (i.e., ) and then obtains the unsupervised model . Thus, we define the following objective function:
(7) where and are hyperparameters to control the ratio between the loss terms, is the unsupervised model of the -th client at -th training round, is from unlabeled dataset , is the form of perturbation, i.e., is the shift augmentation, is the flip augmentation, is a penalty term that aims to let the -th client unsupervised model learn the knowledge of the supervised model (note that is inferred from Equation (9)), and is pseudo label obtained by using our proposed argmax method. The argmax method is defined as follows:
(8) where is a function that can output the maximum probability that unlabeled data belongs to a certain class, is the one-hot function that can change the numerical value to 1, represents the number of unlabeled data after data augmentation, and is from unlabeled dataset . Specifically, we discard low-confident predictions below confidence threshold when generating pseudo-labels.
Figure 2: Overview of the proposed argmax method. -
•
Step 4, Aggregation: The server uses the proposed (see Section IV-B) aggregation algorithm to aggregate the unsupervised models uploaded by the clients to obatin the global unsupervised model, i.e., where is the unsupervised model of the -th client at -th training round and is the weight of the -th client. The server then aggregates the global unsupervised model , the supervised model , and the global model from the previous round to obtain a new global model :
(9) where , , and are the corresponding weights of the three models and .
Repeat all the above steps until the global model converges. The proposed algorithm is shown in Algorithm 1.
V-B Aggregation Algorithm
In this section, we present the designed aggregation algorithm, which can dynamically adjust the weight of the corresponding local model according to the training frequency of the client to alleviate the non-IID problem. We observe that the parameter distribution of the global model will be biased towards clients that often participate in federated training, which is obviously not friendly to the robustness of the global model. Therefore, our insight is to reduce the influence of clients with high training frequency on the global model to improve the robustness of the model. Thus, the formal expression of the aggregation algorithm is as follows:
(10) |
where is the sample proportion of the server, is the total number of clients, , is the number of times that the -th client has been trained up to the -th round, and denotes the set of clients selected by the server in round . Then, for the client, the aggregation rule is expressed as follows:
V-C Dirchlet Data Distribution Function
To better evaluate the robustness of the designed system to non-IID data, in this paper, we introduce the Dirchlet distribution function [40, 41], which is a popular non-IID function, to adjust the non-IID level of the local client data. Specifically, we generate data distributions of different non-IID levels by adjusting the parameter (i.e., ) of the Dirchlet distribution function. We assume that the local dataset of -th client has classes, and thus, the definition of Dirichlet distribution function is as follows:
(11) |
(12) |
where is a set of samples randomly selected from the Dirichlet function, i.e., and , are the parameters of the Dirichlet distribution function (where ), and denotes the proportion of the -th class data in all data of the client. In particular, the smaller the , the higher the non-IID level of the data distribution of each client; otherwise, the data distribution of the client tends to the IID setting. Therefore, we adjust the parameters of the Dirichlet distribution function to simulate the different non-IID levels of the client’s local dataset. For example, as shown in Fig. 3, we demonstrate the data distribution when .



VI Experiment
In the Labels-at-Server scenario, we compared our method and the baseline [12] on the CIFAR-10 dataset. Furthermore, we simulate the federated learning setup (one server and clients) on a commodity machine with Intel(R) Core(TM) i9-9900K CPU @ 3.60GHz and NVIDIA GeForce RTX 2080Ti GPU.
VI-A Experiment Setup
Dataset 1) CIFAR-10 dataset under IID setting: We use the CIFAR-10 dataset including 56,000 training samples and 2,000 test samples as the validation dataset in our experiment. The training set includes 55,000 unlabeled samples and 1,000 labeled samples, where the former is used to train the unsupervised model at the local and the latter is used to train the supervised model at the server. The unlabeled samples are equally distributed to 100 clients at a ratio of , of which there are 550 samples for each client (i.e., 55 for each class, and 10 classes in total). Similarly, the labeled samples have a total of 1,000 and contain 10 classes on the server, of which there are 100 samples in each class. Meanwhile, we set a participation rate of clients, i.e., 5 clients are randomly selected for training in each round.
Dataset 2) CIFAR-10 dataset under non-IID setting: Our setting is similar to the above IID setting, except that the Dirchlet distribution function is introduced to adjust the non-IID level of the local client data. Specifically, we generate data distributions of different non-IID levels by adjusting the parameter (i.e., ) of the Dirchlet distribution function. Meanwhile, we simulate quantity imbalance and class imbalance the local client data. In particular, we make each client hold a different number of training samples. For example, some customers have 580 samples, while some customers have less than 50 samples. Second, we make the client hold a different sample size for each category in the data. For example, some clients have ten types of data, and some clients have less than two types of data.
Baseline and training details: Our baseline is [12] naively using unsupervised model and supervised model parameter decomposition strategy, i.e., . In the training process, both our model and baseline use Stochastic Gradient Descent (SGD) to optimize the ResNet-9 neural network with initial learning rate . We set training round , the number of labeled samples on sever is , local client training epoch and mini-batch size , the server training epoch and mini-batch size . Second, we set the data augmentation number in the argmax method .
VI-B Experiment Results
As shown in Fig. 4, under IID and non-IID settings, our method is better than baseline under each different aggregation method settings. For example, under a non-IID setting, the convergence accuracy of our method is 47.5% about 3% higher than that of the baseline. In particular, the accuracy of our method increases faster and more stable in the early stage of model training. The reason is that: (1) The focuses on the implicit contributions between iterations of the global model in a fine-grained manner, while the only naively uses model parameter decomposition. (2) Frequency-based aggregation method is more suitable for non-IID settings. Notably, the only requires the server to give appropriate weights in the aggregation process according to the training frequency of each client, which does not bring additional computational overhead to the server and local clients.






Fig. 5 shows the performance comparison of the proposed method under different hyperparameter settings. To be specific, the three hyperparameters are the weights of the global unsupervised model, the supervised model, and the previous round of global model. From Fig. 5, we can find that under the non-IID setting, as decreases, the accuracy curve of the proposed method becomes unstable. The reason for this phenomenon is that: with the decrease of global unsupervised model weight, is losing its effect. In particular, when , , and , our global model is the aggregation of the optimal weights of the three models. In addition, we find that the proposed method is easy to achieve better performance when these three parameters are relatively close under the IID setting, as shown in Fig. 5.
Fig. 6 shows the performance comparison of the proposed method on different non-IID levels of client data. In this experiment, we let denote the highest non-IID level of the client data. In this case, as the value of increases, the local client data distribution is closer to the IID setting. It can be seen from Fig. 6 that for different non-IID levels, our method all can achieve stable accuracy. Meanwhile, the model convergence accuracy under settings does not differ by more than 1%. Therefore, our method is not sensitive to the different levels of client data distribution, i.e., it is robust to different types of data distribution settings.
Fig. 6 shows the performance comparison of the proposed method in the case of different numbers of labeled samples at the server. Obviously, the converged accuracy of our method is 47% with 800 labeled samples, which is 2% higher than . However, when the number of labeled samples is reduced to 700, the accuracy of our model decreases greatly. Therefore, we regard as the best setting for our method.
VI-C Analysis
In this section, we further analyze the advantages of compared to in labels-at-server scenario.
1) The performance of training model under the CIFAR-10 dataset is better than . This is due to simply uses the strategy of parameter decomposition of unsupervised model and supervised model in the training process, i.e., . In this way, the learned global model will be biased towards unlabeled data (unsupervised model) or labeled data (supervised model) instead of the overall data. Thus, in order to avoid the drift problem of the global model, adds the global model from the previous round to the model parameter aggregation, i.e., . Meanwhile, we conducted a sensitivity experiment of model performance to different hyperparameter weights to find the optimal weight combination.
2) is robust to different levels of non-IID data. In our experiment, we introduced the Dirichlet distribution function to simulate the local client non-IID data in FL. In details, we generate data distributions of different non-IID levels by adjusting the parameters of the Dirchlet distribution function, i.e., respectively correspond to different levels of non-IID. The results show that the performance difference of our model does not exceed 1% under different levels of non-IID settings. uses a pseudo-random method to generate the non-IID data distribution of each client. However, in reality there is no such distribution, which will cause the model to lose its robustness.
VII conclusion
In this paper, we studied the labels-at-server scenario and addressed the problem of data availability and non-IID in FL. To solve the first problem, we designed a robust SSFL system that uses the algorithm to achieve high-precision semi-supervised learning. To tackle the non-IID problem, we propose a novel aggregation algorithm , which effectively achieves the stable performance of the global model in the training process without adding additional computational overhead. Through experimental verification, our robust SSFL system is significantly better than the baseline in performance. In future work, we will further improve the algorithm to maximize the use of unlabeled data. Furthermore, we will continue to strengthen the theory of SSFL so that it can be better applied in real-world scenarios.
Acknowledgment
This work is supported by Heilongjiang Provincial Natural Science Foundation of China (Grant No. LH2020F044), the 2019–“Chunhui Plan” Cooperative Scientific Research Project of Ministry of Education of China (Grant No. HLJ2019015), the Fundamental Research Funds for Heilongjiang Universities, China (Grant No. 2020-KYYWF-1014), and also supported by NSFC under grant No. 6210071210.
References
- [1] B. McMahan, E. Moore, D. Ramage, S. Hampson, and B. A. y Arcas, “Communication-efficient learning of deep networks from decentralized data,” in Artificial Intelligence and Statistics. PMLR, 2017, pp. 1273–1282.
- [2] J. Kang, Z. Xiong, D. Niyato, S. Xie, and J. Zhang, “Incentive mechanism for reliable federated learning: A joint optimization approach to combining reputation and contract theory,” IEEE Internet of Things Journal, vol. 6, no. 6, pp. 10 700–10 714, 2019.
- [3] Y. Liu, J. James, J. Kang, D. Niyato, and S. Zhang, “Privacy-preserving traffic flow prediction: A federated learning approach,” IEEE Internet of Things Journal, vol. 7, no. 8, pp. 7751–7763, 2020.
- [4] W. Y. B. Lim, Z. Xiong, J. Kang, D. Niyato, C. S. Leung, C. Miao, and S. Shen, “When information freshness meets service latency in federated learning: A task-aware incentive scheme for smart industries,” IEEE Transactions on Industrial Informatics, 2020.
- [5] Y. Liu, S. Garg, J. Nie, Y. Zhang, Z. Xiong, J. Kang, and M. S. Hossain, “Deep anomaly detection for time-series data in industrial iot: a communication-efficient on-device federated learning approach,” IEEE Internet of Things Journal, vol. 8, no. 8, pp. 6348–6358, 2020.
- [6] T. Li, A. K. Sahu, M. Zaheer, M. Sanjabi, A. Talwalkar, and V. Smith, “Federated optimization in heterogeneous networks,” arXiv preprint arXiv:1812.06127, 2018.
- [7] Y. Liu, X. Yuan, Z. Xiong, J. Kang, X. Wang, and D. Niyato, “Federated learning for 6g communications: Challenges, methods, and future directions,” China Communications, vol. 17, no. 9, pp. 105–118, 2020.
- [8] J. Kang, Z. Xiong, D. Niyato, Y. Zou, Y. Zhang, and M. Guizani, “Reliable federated learning for mobile networks,” IEEE Wireless Communications, vol. 27, no. 2, pp. 72–80, 2020.
- [9] Y. Liu, X. Yuan, R. Zhao, Y. Zheng, and Y. Zheng, “Rc-ssfl: Towards robust and communication-efficient semi-supervised federated learning system,” arXiv preprint arXiv:2012.04432, 2020.
- [10] Z. Long, L. Che, Y. Wang, M. Ye, J. Luo, J. Wu, H. Xiao, and F. Ma, “Fedsemi: An adaptive federated semi-supervised learning framework,” arXiv preprint arXiv:2012.03292, 2020.
- [11] Y. Jin, X. Wei, Y. Liu, and Q. Yang, “Towards utilizing unlabeled data in federated learning: A survey and prospective,” arXiv e-prints, pp. arXiv–2002, 2020.
- [12] W. Jeong, J. Yoon, E. Yang, and S. J. Hwang, “Federated semi-supervised learning with inter-client consistency & disjoint learning,” in International Conference on Learning Representations, 2020.
- [13] Y. Zhu, Y. Liu, J. James, and X. Yuan, “Semi-supervised federated learning for travel mode identification from gps trajectories,” IEEE Transactions on Intelligent Transportation Systems, 2021.
- [14] B. Wang, A. Li, H. Li, and Y. Chen, “Graphfl: A federated learning framework for semi-supervised node classification on graphs,” arXiv preprint arXiv:2012.04187, 2020.
- [15] L. Samuli and A. Timo, “Temporal ensembling for semi-supervised learning,” in International Conference on Learning Representations (ICLR), vol. 4, no. 5, 2017, p. 6.
- [16] S. Park, J. Park, S.-J. Shin, and I.-C. Moon, “Adversarial dropout for supervised and semi-supervised learning,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 32, no. 1, 2018.
- [17] K. Sohn, D. Berthelot, N. Carlini, Z. Zhang, H. Zhang, C. A. Raffel, E. D. Cubuk, A. Kurakin, and C.-L. Li, “Fixmatch: Simplifying semi-supervised learning with consistency and confidence,” Advances in Neural Information Processing Systems, vol. 33, 2020.
- [18] X. Li, M. JIANG, X. Zhang, M. Kamp, and Q. Dou, “Fedbn: Federated learning on non-iid features via local batch normalization,” in International Conference on Learning Representations, 2020.
- [19] X. Zhu and A. B. Goldberg, “Introduction to semi-supervised learning,” Synthesis lectures on artificial intelligence and machine learning, vol. 3, no. 1, pp. 1–130, 2009.
- [20] O. Chapelle, B. Scholkopf, and A. Zien, “Semi-supervised learning (chapelle, o. et al., eds.; 2006)[book reviews],” IEEE Transactions on Neural Networks, vol. 20, no. 3, pp. 542–542, 2009.
- [21] D. P. Kingma, S. Mohamed, D. J. Rezende, and M. Welling, “Semi-supervised learning with deep generative models,” in Advances in neural information processing systems, 2014, pp. 3581–3589.
- [22] X. Zhai, A. Oliver, A. Kolesnikov, and L. Beyer, “S4l: Self-supervised semi-supervised learning,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2019, pp. 1476–1485.
- [23] P. K. Mallapragada, R. Jin, A. K. Jain, and Y. Liu, “Semiboost: Boosting for semi-supervised learning,” IEEE transactions on pattern analysis and machine intelligence, vol. 31, no. 11, pp. 2000–2014, 2008.
- [24] D.-H. Lee et al., “Pseudo-label: The simple and efficient semi-supervised learning method for deep neural networks,” in Workshop on challenges in representation learning, ICML, vol. 3, no. 2, 2013.
- [25] A. Tarvainen and H. Valpola, “Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results,” in Proceedings of the 31st International Conference on Neural Information Processing Systems, 2017, pp. 1195–1204.
- [26] S. Itahara, T. Nishio, Y. Koda, M. Morikura, and K. Yamamoto, “Distillation-based semi-supervised federated learning for communication-efficient collaborative training with non-iid private data,” IEEE Transactions on Mobile Computing, pp. 1–1, 2021.
- [27] Z. Zhang, Y. Yang, Z. Yao, Y. Yan, J. E. Gonzalez, and M. W. Mahoney, “Improving semi-supervised federated learning by reducing the gradient diversity of models,” arXiv preprint arXiv:2008.11364, 2020.
- [28] X. Li, K. Huang, W. Yang, S. Wang, and Z. Zhang, “On the convergence of fedavg on non-iid data,” arXiv preprint arXiv:1907.02189, 2019.
- [29] Y. Zhao, M. Li, L. Lai, N. Suda, D. Civin, and V. Chandra, “Federated learning with non-iid data,” arXiv preprint arXiv:1806.00582, 2018.
- [30] F. Sattler, S. Wiedemann, K.-R. Müller, and W. Samek, “Robust and communication-efficient federated learning from non-iid data,” IEEE transactions on neural networks and learning systems, vol. 31, no. 9, pp. 3400–3413, 2019.
- [31] C. Briggs, Z. Fan, and P. Andras, “Federated learning with hierarchical clustering of local updates to improve training on non-iid data,” in 2020 International Joint Conference on Neural Networks (IJCNN). IEEE, 2020, pp. 1–9.
- [32] H. Wang, Z. Kaplan, D. Niu, and B. Li, “Optimizing federated learning on non-iid data with reinforcement learning,” in IEEE INFOCOM 2020-IEEE Conference on Computer Communications. IEEE, 2020, pp. 1698–1707.
- [33] Y. Chen, Y. Ning, M. Slawski, and H. Rangwala, “Asynchronous online federated learning for edge devices with non-iid data,” in 2020 IEEE International Conference on Big Data (Big Data). IEEE, 2020, pp. 15–24.
- [34] Q. Li, Y. Diao, Q. Chen, and B. He, “Federated learning on non-iid data silos: An experimental study,” arXiv preprint arXiv:2102.02079, 2021.
- [35] J. Wang, Q. Liu, H. Liang, G. Joshi, and H. V. Poor, “Tackling the objective inconsistency problem in heterogeneous federated optimization,” Advances in Neural Information Processing Systems, vol. 33, 2020.
- [36] M. Sajjadi, M. Javanmardi, and T. Tasdizen, “Regularization with stochastic transformations and perturbations for deep semi-supervised learning,” arXiv preprint arXiv:1606.04586, 2016.
- [37] Q. Xie, Z. Dai, E. Hovy, T. Luong, and Q. Le, “Unsupervised data augmentation for consistency training,” Advances in Neural Information Processing Systems, vol. 33, 2020.
- [38] X. Yang, Z. Song, I. King, and Z. Xu, “A survey on deep semi-supervised learning,” arXiv preprint arXiv:2103.00550, 2021.
- [39] D. Berthelot, N. Carlini, I. Goodfellow, N. Papernot, A. Oliver, and C. A. Raffel, “Mixmatch: A holistic approach to semi-supervised learning,” Advances in Neural Information Processing Systems, vol. 32, 2019.
- [40] M. Yurochkin, M. Agarwal, S. Ghosh, K. Greenewald, N. Hoang, and Y. Khazaeni, “Bayesian nonparametric federated learning of neural networks,” in International Conference on Machine Learning. PMLR, 2019, pp. 7252–7261.
- [41] T.-M. H. Hsu, H. Qi, and M. Brown, “Measuring the effects of non-identical data distribution for federated visual classification,” arXiv preprint arXiv:1909.06335, 2019.