This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Price of Stability in Quality-Aware Federated Learning

Yizhou Yan, Xinyu Tang, Chao Huang, and Ming Tang This work was supported in part by Guangdong Basic and Applied Basic Research Foundation under Grant 2023A1515012819 and the National Natural Science Foundation of China under Grant 62202214. (Corresponding author: Chao Huang) Both authors contributed equally to this research. Department of Computer Science and Engineering, Southern University of Science and Technology
Department of Computer Science, The University of California, Davis
Abstract

Federated Learning (FL) is a distributed machine learning scheme that enables clients to train a shared global model without exchanging local data. The presence of label noise can severely degrade the FL performance, and some existing studies have focused on algorithm design for label denoising. However, they ignored the important issue that clients may not apply costly label denoising strategies due to them being self-interested and having heterogeneous valuations on the FL performance. To fill this gap, we model the clients’ interactions as a novel label denoising game and characterize its equilibrium. We also analyze the price of stability, which quantifies the difference in the system performance (e.g., global model accuracy, social welfare) between the equilibrium outcome and the socially optimal solution. We prove that the equilibrium outcome always leads to a lower global model accuracy than the socially optimal solution does. We further design an efficient algorithm to compute the socially optimal solution. Numerical experiments on MNIST dataset show that the price of stability increases as the clients’ data become noisier, calling for an effective incentive mechanism.

publicationid: pubid: 979-8-3503-1090-0/23/$31.00 © 2023 IEEE

I Introduction

Federated learning (FL) is an emerging machine learning paradigm that enables distributed clients (e.g., devices, organizations) to collaboratively train a global model while keeping their data local [1]. This approach offers several advantages over traditional centralized machine learning, such as improved data privacy and the ability to leverage distributed computational resources. Thus, FL is a promising candidate approach to improving the performance of networking systems. This can be achieved by letting network entities (e.g., devices, operators) act as clients and cooperatively learn system characteristics (e.g., user behaviors, network environment) for prediction and adaptive control.

An FL process typically follows an iterative procedure that consists of several rounds of communication between the clients and a trusted central server. In each round, the server sends the current global model to the clients. These clients then use their local data to compute model updates, e.g., gradients or parameter changes, and send these updates back to the server. The server updates the global model by aggregating the received updates (e.g., weighted averaging). This process repeats until the global model converges [2].

The performance of the global model crucially depend on data heterogeneity and data quality among clients:

  • Data heterogeneity: It pertains to the differences in data distribution among clients. It has been shown to severely compromise the FL convergence and generalization. Existing studies have concentrated on addressing data heterogeneity by using techniques such as regularization and personalization [3].

  • Data quality: It refers to the conditions of the training data such as the noise rate of labels. Label noise can arise from various sources, e.g., mislabeling, human errors during data collection and annotation. Take user mobility traces as an example. Indoor mobile devices connecting to WiFi tend to have better localization accuracy than those devices without WiFi connection do. This leads to a training dataset with higher quality (i.e., fewer noisy/wrong labels). Label noise is a common challenge in real-world data [4], but it is under-explored in FL and hence the focus of our paper.

Some recent studies have focused on the label noise detection and label denoising (i.e., correct the wrong labels) [4, 5, 6]. For example, Xu et. al in [5] proposed FedCorr, a multi-stage framework that dynamically identifies and corrects incorrect labels based on per-sample losses. Zeng at. al in [6] proposed a consensus-based approach to detect label noise. These recent advances have presented FL clients with the opportunity to provide higher quality data, subsequently improving model performance. However, it is important to recognize that in practice, clients (e.g., devices, companies) are self-interested and heterogeneous in terms of their valuations on model performance [7]. Meanwhile, denoising is a resource-intensive process, often requiring significant computing power and/or manpower [5]. Given these constraints, it is imperative for clients to optimize their denoising strategies, striking a balance between the expected improvement in model performance and the associated denoising costs. This motivates our first key question:

Question 1.

How will heterogeneous self-interested clients decide their label denoising strategies?

To answer Question 1, we model the interactions among heterogeneous clients as a non-cooperative game, in which each client decides the denoising strategy to maximize a tradeoff between the global model performance and the denoising cost. Solving (and even formulating) the game is challenging, as it is unclear how label noise affects FL. To this end, we consider two types of label noise below [8].

  • Random flipping: It refers to a case where a label is randomly flipped to another class. Random flipping can model where the annotators have limited expertise or when the annotation process is subject to random errors.

  • Instance-dependent noise: Unlike random flipping, instance-dependent noises are not uniformly distributed across the dataset. Instead, the probability of mislabeling depends on the characteristics of the instances themselves. This type of noise can model scenarios where some instances are inherently more difficult to annotate correctly or are more prone to ambiguity.

We conduct experiments to evaluate how these two noise types affect the global model performance, which helps with the game formulation. We characterize the game equilibrium and propose an efficient algorithm to compute it.

It is known that in many situations, the game equilibrium can be inefficient, leading to suboptimal outcomes for the entire system. This inefficiency arises because clients aim to maximize their individual payoffs, often without considering the collective welfare of the system [9]. This motivates the study of the price of stability (PoS), which measures the efficiency loss resulting from non-cooperative behavior compared to a socially optimal solution.

Question 2.

What is the PoS in quality-aware FL?

Answering Question 2 is not easy either in the context of quality-aware FL. Specifically, due to the coupling of the clients’ denoising decisions and their multi-dimensional heterogeneity (i.e., valuations and data quality/noise rate), it is difficult to determine the equilibrium outcome and socially optimal solution in closed-from. Nonetheless, we prove the performance degradation of the equilibrium outcome in terms of the global model accuracy. Moreover, we propose an efficient algorithm to compute the socially optimal solution and empirically evaluate the PoS.

The key contributions of this paper are summarized below.

  • To our best knowledge, we present the first work that studies the PoS in quality-aware FL. Our study facilitates understanding the extent to which the system’s performance degrades due to non-cooperative behavior and helps identify opportunities to improve the overall efficiency in practical FL systems.

  • We derive the socially optimal solution and equilibrium outcome of quality-aware FL and prove their uniqueness. We prove that the equilibrium outcome always leads to a higher average noise rate and a lower global model accuracy than the socially optimal solution does. We further design an efficient algorithm to obtain the socially optimal solution for empirical PoS evaluation.

  • We conduct case studies to investigate two types of label noise: random flipping and instance-dependent noise. Numerical experiments show that random flipping is more harmful than instance-dependent noise to the global model accuracy. Moreover, we find that PoS increases as the clients’ data become noisier, calling for an effective incentive mechanism.

II System Model

We first present the quality-aware FL model and characterize the clients’ payoff. Then, we formulate the social welfare maximization problem and the non-cooperative game.

II-A Quality-Aware FL Model

We consider NN clients, denoted by 𝒩={1,2,,N}\mathcal{N}=\{1,2,\ldots,N\}. Each client n𝒩n\in\mathcal{N} has a local dataset 𝒟n\mathcal{D}_{n} with dnd_{n} data samples. We focus on supervised learning tasks in FL, where each data sample consists of a set of features and a label. These clients cooperate in training a global model that maps from feature to label with their local datasets. We consider a scenario where the utility of each client relies on the accuracy of the global model, e.g., network operators can use the global model for product improvement to earn revenues.

There may exist label noise/errors in clients’ datasets, i.e., the labels of some data samples may be incorrect. Let ϵn\epsilon_{n} denote the noise rate of client nn’s dataset, i.e., the fraction of client nn’s data samples that have incorrect labels. Suppose clients are able to rectify incorrect labels with some costs [5]. Let decision variable xn[0,ϵn]x_{n}\in[0,\epsilon_{n}] denote the noise rate of client nn’s dataset after label correction, and 𝒙(xn,n𝒩)\bm{x}\triangleq\left(x_{n},n\in\mathcal{N}\right). Then, the average noise rate after correction among all clients is given by

x¯(𝒙)=n𝒩dnxnn𝒩dn.\bar{x}(\bm{x})=\frac{\sum_{n\in\mathcal{N}}d_{n}x_{n}}{\sum_{n^{\prime}\in\mathcal{N}}d_{n^{\prime}}}. (1)

After the FL process, the global model has an accuracy of

A(𝒙)=g(x¯(𝒙)),A\left(\bm{x}\right)=g\left(\bar{x}(\bm{x})\right), (2)

where g()g(\cdot) is a concave and decreasing function. Such assumptions on g()g(\cdot) are supported by the experimental results under random flipping and instance-dependent noise in Section V. Intuitively, the average noise rate (rather than the noise rate of any arbitrary client) affects the accuracy of the global model. When the average noise rate is higher, the accuracy is lower. Meanwhile, the marginal decrease of the accuracy is increasing in the average noise rate.

From a social perspective, correcting labels can improve the global model accuracy. From a client’s perspective, it incurs additional costs. Thus, the clients’ individual interests may conflict with the goal of social welfare maximization.

II-B Payoff of the Clients

Given decision variable 𝒙\bm{x} and accuracy A(𝒙)A\left(\bm{x}\right), the utility of client n𝒩n\in\mathcal{N} is given by

Un(A(𝒙))=wnA(𝒙),U_{n}\left(A\left(\bm{x}\right)\right)=w_{n}A\left(\bm{x}\right), (3)

where wn>0w_{n}>0 is the unit revenue of client nn. For example, when clients are network operators, wnw_{n} is the revenue that the operator can earn per accuracy improvement [7, 10].

Given decision xnx_{n}, client nn has a cost for label correction:

Cn(xn)=dn(f(xn)f(ϵn)).C_{n}\left(x_{n}\right)=d_{n}\left(f\left(x_{n}\right)-f\left(\epsilon_{n}\right)\right). (4)

We assume that f()f(\cdot) is a strictly convex and decreasing function. Intuitively, the label correction cost is proportional to the size of the dataset dnd_{n}. Meanwhile, as the noise rate of the dataset decreases, the cost of reducing the same noise rate of the dataset increases. If the dataset has not been corrected (i.e., xn=ϵnx_{n}=\epsilon_{n}), the cost is zero. We show a specific formulation of f()f(\cdot) with practical insights in Section V.

The payoff function of a client n𝒩n\in\mathcal{N} is defined as the difference between its utility and cost:

Pn(𝒙)=Un(A(𝒙))Cn(xn).P_{n}\left(\bm{x}\right)=U_{n}\left(A\left(\bm{x}\right)\right)-C_{n}\left(x_{n}\right). (5)

II-C Social Welfare Maximization and Game Formulation

From a social perspective, clients should optimize the average noise rate after correction to maximize the social welfare. The socially optimal solution 𝒙swm\bm{x}^{\textsc{swm}} is given by

𝒙swmarg\displaystyle\bm{x}^{\textsc{swm}}\triangleq\arg~{} maximize𝒙\displaystyle\underset{\bm{x}}{\text{maximize}} SW(𝒙)n𝒩Pn(𝒙)\displaystyle\textstyle\text{SW}(\bm{x})\triangleq\sum_{n\in\mathcal{N}}P_{n}\left(\bm{x}\right) (6a)
subject to xn[0,ϵn],n𝒩.\displaystyle x_{n}\in[0,\epsilon_{n}],~{}n\in\mathcal{N}. (6b)

However, clients are usually self-interested. The interaction among clients is modeled as a non-cooperative game.

Game 1 (Label Denoising Game).
  • Players: clients n𝒩n\in\mathcal{N};

  • Strategy: xn[0,ϵn]x_{n}\in[0,\epsilon_{n}] for each client n𝒩n\in\mathcal{N};

  • Payoff function: Pn(𝒙)P_{n}\left(\bm{x}\right) for each client n𝒩n\in\mathcal{N}.

The stable state of the quality-aware FL system with self-interested clients corresponds to the situation under Nash equilibrium (NE) 𝒙ne\bm{x}^{\textsc{ne}} of Game 1, i.e., the strategy profile such that no client has the incentive to deviate from. Let 𝒙n(xi,i𝒩{n})\bm{x}_{-n}\triangleq(x_{i},i\in\mathcal{N}\setminus\{n\}) denote the decisions of all clients except client n𝒩n\in\mathcal{N}. For presentation simplicity, we use notations Pn(𝒙)P_{n}(\bm{x}) and Pn(xn,𝒙n)P_{n}(x_{n},\bm{x}_{-n}) interchangeably.

Definition 1 (NE).

An NE of Game 1 is an 𝐱ne\bm{x}^{\textsc{ne}} satisfying

Pn(xnne,𝒙nne)Pn(xn,𝒙nne),xn[0,ϵn],n𝒩.P_{n}\left(x^{\textsc{ne}}_{n},\bm{x}^{\textsc{ne}}_{-n}\right)\geq P_{n}\left(x_{n},\bm{x}^{\textsc{ne}}_{-n}\right),~{}x_{n}\in[0,\epsilon_{n}],n\in\mathcal{N}. (7)

We aim to analyze the label denoising strategies of clients and determine the PoS of quality-aware FL systems.

III Theoretical Analysis

In this section, we first derive the socially optimal solution and the associated necessary condition, which provides practical insights and algorithm design guidance. Then, we derive the NE of Game 1. Finally, we characterize the PoS in model accuracy, which is challenging to analyze due to the coupled decision variables and client heterogeneity.

III-A Socially Optimal Solution

Since problem (6) is a convex programming problem, we determine its optimal solution by checking the KKT conditions. Specifically, let hnswm(xn)(i𝒩Pi(𝒙))/xnh^{\textsc{swm}}_{n}(x_{n})\triangleq{\partial(\sum_{i\in\mathcal{N}}P_{i}\left(\bm{x}\right))}/{\partial x_{n}} denote the partial derivative of the social welfare with respect to xnx_{n}. Let w¯i𝒩wi/N\bar{w}\triangleq\sum_{i\in\mathcal{N}}w_{i}/N and d¯i𝒩di/N\bar{d}\triangleq\sum_{i\in\mathcal{N}}d_{i}/N. Thus,

hnswm(xn)=(w¯d¯g(x¯(𝒙))x¯(𝒙)f(xn)xn)dn,h^{\textsc{swm}}_{n}\left(x_{n}\right)=\Big{(}\frac{\bar{w}}{\bar{d}}\frac{\partial g\left(\bar{x}(\bm{x})\right)}{\partial\bar{x}(\bm{x})}-\frac{\partial f(x_{n})}{\partial x_{n}}\Big{)}d_{n}, (8)

which is strictly decreasing in xnx_{n} due to the strict convexity of f()f(\cdot) and the concavity of g()g(\cdot). We use xnswmx^{\circ\textsc{swm}}_{n} to denote the unique zero point of hnswm(xn)h^{\textsc{swm}}_{n}(x_{n}). The optimal solution to problem (6) is given as follows.

Lemma 1 (Optimal Solution).

The optimal solution xnswmx^{\textsc{swm}}_{n} for n𝒩n\in\mathcal{N} to problem (6) is given by

xnswm={ϵn,ifhnswm(ϵn)>0,0,ifhnswm(0)<0,xnswm,otherwise.x^{\textsc{swm}}_{n}=\left\{\begin{array}[]{ll}\epsilon_{n},&\rm{if}\;\text{$h^{\textsc{swm}}_{n}\left(\epsilon_{n}\right)>0$},\\ 0,&\rm{if}\;\text{$h^{\textsc{swm}}_{n}\left(0\right)<0$},\\ x^{\circ\textsc{swm}}_{n},&\rm{otherwise}.\end{array}\right. (9)

Further, xnswmx^{\textsc{swm}}_{n} decreases in w¯\bar{w} and increases in d¯\bar{d}. Given a fixed value of d¯\bar{d}, xnswmx^{\textsc{swm}}_{n} is independent of dnd_{n}.

Lemma 1 can be proven by analyzing the impact of system parameters w¯\bar{w}, d¯\bar{d}, and dnd_{n} on the zero point of hnswm(xn)h^{\textsc{swm}}_{n}(x_{n}) in (8). The detailed proof is given in [11].

Lemma 1 shows that if the average unit revenue of the clients (i.e., w¯\bar{w}) is larger, then each client nn should choose a smaller noise rate xnx_{n} (i.e., a higher data quality) to maximize the social welfare. Moreover, if the average number of data samples is smaller, then a smaller noise rate xnx_{n} should be chosen by the clients in order to maintain sufficient data samples with correct labels. However, given a fixed average number of data samples, the increase in the number of data samples of a client does not affect the optimal solution.

According to Lemma 1, we further derive the necessary condition for the optimal solution 𝒙swm\bm{x}^{\textsc{swm}}. The condition can provide additional insights and will be used for the algorithm design in Section IV-A. We first present a lemma to support our analysis, with the proof given in [11].

Lemma 2.

If l()l(\cdot) is convex, then for any a<b<c<da<b<c<d,

l(b)l(a)/(ba)l(d)l(c)/(dc).{l\left(b\right)-l\left(a\right)}/{(b-a)}\leq{l\left(d\right)-l\left(c\right)}/{(d-c)}. (10)

If l()l(\cdot) is strictly convex, then the strict inequality holds in (10). If l()l(\cdot) is concave or strictly concave, the inequality operator in (10) should be replaced by \geq or >>, respectively.

Now, we can derive the necessary condition for 𝒙swm\bm{x}^{\textsc{swm}}.

Proposition 1 (Necessary Condition for Optimal Solution).

An 𝐱swm\bm{x}^{\textsc{swm}} is an optimal solution to problem (6) only if xnswm=min{maxi𝒩xiswm,ϵn}x^{\textsc{swm}}_{n}=\min\{\max_{i\in\mathcal{N}}x^{\textsc{swm}}_{i},\epsilon_{n}\} for any n𝒩n\in\mathcal{N}. That is, there exists a threshold θ0\theta\geq 0 such that

xnswm={ϵn,ifϵn<θ,θ,otherwise.x^{\textsc{swm}}_{n}=\left\{\begin{array}[]{ll}\epsilon_{n},&\rm{if}~{}\text{$\epsilon_{n}<\theta$},\\ \theta,&\rm{otherwise}.\end{array}\right. (11)
Proof.

We prove by contradiction. Suppose there exists an xnx_{n} that satisfies xn<ϵnx_{n}<\epsilon_{n} and xn<xmx_{n}<x_{m} for an m𝒩m\in\mathcal{N} in solution 𝒙\bm{x}. Then, the social welfare can be increased by replacing xnx_{n} and xmx_{m} with xnx_{n}^{\prime} and xmx_{m}^{\prime}, respectively. Here, xn=xn+δ/dnx_{n}^{\prime}=x_{n}+{\delta}/{d_{n}}, xm=bδ/dmx_{m}^{\prime}=b-{\delta}/{d_{m}} for a small positive δ\delta that ensures xn<ϵnx_{n}^{\prime}<\epsilon_{n} and xn<xmx_{n}^{\prime}<x_{m}^{\prime}. The social welfare is increased because the model accuracy remains unchanged but the total cost is reduced, since dn(f(xn)f(xn))dm(f(xm)f(xm))d_{n}\left(f\left(x_{n}\right)-f\left(x_{n}^{\prime}\right)\right)-d_{m}\left(f\left(x_{m}^{\prime}\right)-f\left(x_{m}\right)\right) is greater than zero based on Lemma 2. Thus, 𝒙\bm{x} cannot be the socially optimal solution. ∎

Based on Proposition 1, to maximize social welfare, clients need to give priority to reducing the noise rate of those clients’ datasets with the highest noise rate. Meanwhile, Proposition 1 inspires the efficient algorithm design for deriving the optimal solution in Section IV-A.

III-B Nash Equilibrium of Label Denoising Game

We derive the NE of Game 1 based on the best response of each client n𝒩n\in\mathcal{N}. Specifically, let hnne(xn)Pn(𝒙)/xnh^{\textsc{ne}}_{n}(x_{n})\triangleq{\partial P_{n}(\bm{x})}/{\partial x_{n}} denote the partial derivative of the payoff function of client nn with respect to xnx_{n}. Thus,

hnne(xn)=(wnNd¯g(x¯(𝒙))x¯(𝒙)f(xn)xn)dn,h^{\textsc{ne}}_{n}\left(x_{n}\right)=\left(\frac{w_{n}}{N\bar{d}}\frac{\partial g\left(\bar{x}(\bm{x})\right)}{\partial\bar{x}(\bm{x})}-\frac{\partial f(x_{n})}{\partial x_{n}}\right)d_{n}, (12)

which is strictly decreasing in xnx_{n} due to the strict convexity of f()f(\cdot) and the concavity of g()g(\cdot). Let xnnex^{\circ\textsc{ne}}_{n} denote the zero point of hnne(xn)h^{\textsc{ne}}_{n}\left(x_{n}\right). The NE is determined as follows.

Proposition 2 (NE).

The unique NE of Game 1 is

xnne={ϵn,ifhnne(ϵn)>0,0,ifhnne(0)<0,xnne,otherwise.x^{\textsc{ne}}_{n}=\left\{\begin{array}[]{ll}\epsilon_{n},&\rm{if}\;\text{$h^{\textsc{ne}}_{n}\left(\epsilon_{n}\right)>0$},\\ 0,&\rm{if}\;\text{$h^{\textsc{ne}}_{n}\left(0\right)<0$},\\ x^{\circ\textsc{ne}}_{n},&\rm{otherwise}.\end{array}\right. (13)

As either wnw_{n} increases or d¯\bar{d} decreases, xnnex^{\textsc{ne}}_{n} decreases.

Proposition 2 is proven based on the best response of the clients and the impact of system parameters on the zero point of hnne(xn)h^{\textsc{ne}}_{n}\left(x_{n}\right). The proof of uniqueness is given in an online technical report [11]. In contrast with the optimal solution in Lemma 1, client nn cares about its own unit revenue, i.e., it reduces its noise rate under a larger wnw_{n}. In addition, similar to the optimal solution, if the average number of data samples decreases, client nn has the incentive to reduce its noise rate to improve the accuracy of the global model.

III-C Price of Stability in Model Accuracy

We characterize the PoS111We use the term “PoS” to refer to the performance reduction due to the self-interested clients. In Section V, we will follow the conventional mathematical definition of PoS and define it as the ratio of the maximum social welfare to the social welfare under the unique NE in our work. by showing that when compared wtih the optimal solution to problem (6), the NE of Game 1 leads to a higher average noise rate and hence a lower global model accuracy. Note that determining such an analytical result is not straightforward, because the socially optimal solution is not the solution that maximizes the model accuracy due to the fact that higher accuracy is at the expense of higher correction costs for the clients. Specifically,

Lemma 3 (PoS in Label Denoising).

The NE of Game 1 leads to a higher average noise rate than the optimal solution to problem (6), i.e., x¯(𝐱ne)x¯(𝐱swm)\bar{x}(\bm{x}^{\textsc{ne}})\geq\bar{x}(\bm{x}^{\textsc{swm}}).

Proof.

We prove this by contradiction. Suppose x¯(𝒙ne)<x¯(𝒙swm)\bar{x}(\bm{x}^{\textsc{ne}})<\bar{x}(\bm{x}^{\textsc{swm}}). Thus, there exists an n𝒩n\in\mathcal{N} such that xnne<xnswmx^{\textsc{ne}}_{n}<x^{\textsc{swm}}_{n}. Then, there exists a small positive δ\delta that ensures x¯(𝒙ne)+dnδ/(Nd¯)<x¯(𝒙swm)dnδ/(Nd¯)\bar{x}(\bm{x}^{\textsc{ne}})+{d_{n}\delta}/(N\bar{d})<\bar{x}(\bm{x}^{\textsc{swm}})-{d_{n}\delta}/(N\bar{d}) and xnne+δ<xnswmδx^{\textsc{ne}}_{n}+\delta<x^{\textsc{swm}}_{n}-\delta. Based on the definition of 𝒙ne\bm{x}^{\textsc{ne}} and 𝒙swm\bm{x}^{\textsc{swm}}, wn(g(x¯(𝒙ne))g(x¯(𝒙ne)+dnδ/(Nd¯)))dn(f(xnne)f(xnne+δ))w_{n}(g(\bar{x}(\bm{x}^{\textsc{ne}}))-g(\bar{x}(\bm{x}^{\textsc{ne}})+{d_{n}\delta}/{(N\bar{d})}))\geq d_{n}(f(x^{\textsc{ne}}_{n})-f(x^{\textsc{ne}}_{n}+\delta)) and Nw¯(g(x¯(𝒙swm)dnδ/(Nd¯))g(x¯(𝒙swm)))dn(f(𝒙swmδ)f(𝒙swm))N\bar{w}(g(\bar{x}(\bm{x}^{\textsc{swm}})-{d_{n}\delta}/{(N\bar{d})})-g(\bar{x}(\bm{x}^{\textsc{swm}})))\leq d_{n}(f(\bm{x}^{\textsc{swm}}-\delta)-f(\bm{x}^{\textsc{swm}})), respectively. However, according to Lemma 2, g(x¯(𝒙swm)dnδ/(Nd¯))g(x¯(𝒙swm))g(x¯(𝒙ne))g(x¯(𝒙ne)+dnδ/(Nd¯))g(\bar{x}(\bm{x}^{\textsc{swm}})-{d_{n}\delta}/{(N\bar{d})})-g(\bar{x}(\bm{x}^{\textsc{swm}}))\geq g(\bar{x}(\bm{x}^{\textsc{ne}}))-g(\bar{x}(\bm{x}^{\textsc{ne}})+{d_{n}\delta}/{(N\bar{d})}) and f(xnne)f(xnne+δ)>f(xnswmδ)f(xnswm)f(x^{\textsc{ne}}_{n})-f(x^{\textsc{ne}}_{n}+\delta)>f(x^{\textsc{swm}}_{n}-\delta)-f(x^{\textsc{swm}}_{n}). This leads to a contradiction. ∎

Based on Lemma 3, we derive the main result on accuracy.

Theorem 1 (PoS in Model Accuracy).

The socially optimal solution always leads to a higher accuracy of the global model than the NE does, i.e., A(𝐱swm)A(𝐱ne)A(\bm{x}^{\textsc{swm}})\geq A(\bm{x}^{\textsc{ne}}).

Theorem 1 implies that, when compared with maximizing the social welfare, clients always obtain a global model with lower accuracy by considering their own interests at NE.

Moreover, we present the PoS in social welfare for motivating our experiments in Section V. This result is obtained based on the definition of social welfare maximization.

Lemma 4 (PoS in Social Welfare).

The socially optimal solution always leads to a higher social welfare than NE, i.e., SW(𝐱swm)SW(𝐱ne)\text{SW}(\bm{x}^{\textsc{swm}})\geq\text{SW}(\bm{x}^{\textsc{ne}}).

IV Algorithm Design

We propose algorithms for determining the optimal solution to problem (6) and NE, respectively. These algorithms are applied for the case study in Section V. Designing a low-complexity algorithm for deriving the optimal solution is nontrivial due to the multi-variate decisions. Furthermore, we propose a best response algorithm to derive the NE.

IV-A Social Welfare Maximization Algorithm

Based on Lemma 1, we can determine the optimal solution to problem (6) by finding the threshold θ\theta. Let xn(θ)min(θ,ϵn)x_{n}(\theta)\triangleq\min\left(\theta,\epsilon_{n}\right), and 𝒙(θ)(xn(θ),n𝒩)\bm{x}(\theta)\triangleq(x_{n}(\theta),n\in\mathcal{N}). Define a function H(θ)SW(𝒙(θ))H\left(\theta\right)\triangleq\text{SW}\left(\bm{x}(\theta)\right). Then, we formulate a problem with respect to θ\theta as follows:

maximize𝜃\displaystyle\underset{\theta}{\text{maximize}} H(θ)\displaystyle~{}~{}~{}H\left(\theta\right) (14a)
subject to θ[0,maxn𝒩ϵn].\displaystyle~{}~{}~{}\theta\in[0,\max_{n\in\mathcal{N}}\epsilon_{n}]. (14b)

Given the optimal solution θswm\theta^{\textsc{swm}} to problem (14), we can obtain the optimal solution to problem (6) based on (11).

Algorithm 1 Social Welfare Maximization Algorithm
1:θswm\theta^{\textsc{swm}} \leftarrow determine the value of θ[0,maxn𝒩ϵn]\textstyle\theta\in[0,\max_{n\in\mathcal{N}}\epsilon_{n}] that maximizes H(θ)H(\theta) using ternary search;
2:xnswmx^{\textsc{swm}}_{n} \leftarrow given θswm\theta^{\textsc{swm}}, compute xnx_{n} using (11) for n𝒩n\in\mathcal{N};

To solve problem (14), we derive the property of H(θ)H\left(\theta\right).

Lemma 5 (Function Property).

H(θ)H\left(\theta\right) is a unimodal function, i.e., H(θ)H\left(\theta\right) is monotonically increasing up to a certain value of θ\theta and then monotonically decreasing.

The proof is given in the online technical report [11]. Based on Lemma 5, since H(θ)H(\theta) is a unimodal function with respect to a single-dimensional variable θ\theta, we can find its maximum point by ternary search, based on which we can determine the optimal solution to problem (6) using Proposition 1. We present the algorithm in Algorithm 1. The time complexity for obtaining optimal θ\theta is 𝒪(Nlog(1/μ))\mathcal{O}(N\log({1}/{\mu})), where μ\mu denotes a pre-defined threshold that bounds the gap between algorithm output and ground truth values of θ\theta.

IV-B Best Response Algorithm for NE

According to (12), we propose a best response algorithm in Algorithm 2 to find the NE of Game 1. Sepcifically, given the decisions of other clients, each client alternatively updates its noise rate to maximize its payoff. The algorithm terminates until convergence, i.e., the decision update gap of every client is smaller than a predefined Convg_Thresh.

Algorithm 2 Best Response Algorithm
xnϵn,n𝒩x_{n}\leftarrow\epsilon_{n},n\in\mathcal{N}; n0n\leftarrow 0; Convg_Set\text{Convg\_Set}\leftarrow\emptyset;
while |Convg_Set|<N|\text{Convg\_Set}|<N do
     nmod(n,N)+1n\leftarrow\text{mod}(n,N)+1; xnx_{n}^{\prime}\leftarrow zero point of hnne(xn)h^{\textsc{ne}}_{n}(x_{n});
     if |xnxn|>Convg_Thresh|x_{n}-x_{n}^{\prime}|>\text{Convg\_Thresh} then
         xnxnx_{n}\leftarrow x_{n}^{\prime}; Convg_Set\text{Convg\_Set}\leftarrow\emptyset;
     end if
     Convg_SetConvg_Set{n}\text{Convg\_Set}\leftarrow\text{Convg\_Set}\cup\{n\};
end while

V Case Study with Different Noise Types

We present case studies to validate our theoretical analysis and draw new insights. We consider N=5N=5 clients, e.g., a cross-silo FL scenario where organizations cooperate for training. The clients perform FedAvg algorithm over MNIST dataset using CNN. Each client has dn=1.2×104d_{n}=1.2\times 10^{4} data samples. We study two types of label noise [8]:

  • Random flipping: A client nn’s label flips to an random incorrect label with probability ϵn\epsilon_{n}. It models the case where data collection is subject to random errors.

  • Instance dependent: Each data sample is relabeled based on the inference result of a pre-trained CNN model. This approach simulates the human annotation process, and the noise rates can be adjusted by changing the accuracy of the pre-trained CNN model. For example, if a CNN model has an accuracy 0.80.8, then the error rate of a dataset whose labels are generated from the CNN model’s inference results is around 0.20.2.

Refer to caption
Refer to caption

(a)           (b)

Figure 1: Model accuracy under different noise rates: (a) random flipping; (b) instance dependency.

V-A Impact of Label Noise on FL

We first study how label noise affects FL in terms of convergence and generalization. Fig. 1 plots how the test accuracy of the global model changes with the training rounds (i.e., convergence). We observe that for both random flipping and instance-dependent noises, the existence of label noise slows FL convergence. Moreover, under random flipping, a large noise rate (e.g., 0.20.2) causes overfitting. However, overfitting is not obvious under instance-dependent.

Refer to caption
Figure 2: Model accuracy under random flipping and instance dependency.

Fig. 2 plots how the test accuracy changes with the noise rate at round 5050 (i.e., generalization). We observe that the global model performance decreases in the error rate. Interestingly, we observe that instance-dependent noise is less harmful to the model accuracy. Furthermore, we use curve fitting to simulate the impact of noise rate on global model accuracy and obtain the results below:

  • Random flipping: The accuracy degradation across noise rate ϵ\epsilon fits a linear function. In addition, Table I shows that given the same average noise rate, the heterogeneous noise rates of the clients do not have a significant impact on the accuracy. Based on the above observation, we adopt a linear formulation of g(x¯(𝒙))g(\bar{x}(\bm{x})) with respect to x¯(𝒙)\bar{x}(\bm{x}) for random flipping:

    g(x¯(𝒙))=κx¯(𝒙)+g0,g\left(\bar{x}(\bm{x})\right)=\kappa\bar{x}(\bm{x})+g_{0}, (15)

    where κ<0\kappa<0 and g0(0,1]g_{0}\in(0,1] correspond to the accuracy decreasing per unit of noise rate and the model accuracy under no label error, respectively.

  • Instance dependent: The accuracy degradation across noise rate ϵ\epsilon approximately fits a quadratic function:

    g(x¯(𝒙))=g2(x¯(𝒙))2+g1x¯(𝒙)+g0,g\left(\bar{x}(\bm{x})\right)=g_{2}\left(\bar{x}(\bm{x})\right)^{2}+g_{1}\bar{x}(\bm{x})+g_{0}, (16)

    where g2<0g_{2}<0, g1<0g_{1}<0 and g0(0,1]g_{0}\in(0,1].

Note that the two fitted functions validate our assumption on g()g(\cdot), which is a concave and decreasing function. In what follows, we will use the fitted functions to evaluate the PoS.

V-B Impact of Label Noise on PoS

To study the PoS, we use the following function to help model the denoising cost (see (4)):

f(x)=clnx,f(x)=c\ln{x}, (17)

where c<0c<0. Formulation (17) is determined based on an error correction method called FedCorr [5].222In each round of training, each client rectifies the data samples with incorrect labels using the recent training models. Assuming that with the model obtained in each training round, a fixed fraction of noisy data samples can be found, then the correction cost is given by (17). Let ϕ(μ,σ,a,b)\phi(\mu,\sigma,a,b) denote a truncated normal distribution with mean μ\mu, variance σ\sigma, and truncated range [a,b][a,b]. We consider c=1c=-1,333Note that the relative values of cc and wnw_{n} (rather than their absolute values) affect simulation results. We choose a unit cost for simplicity. wnϕ(106,104,5×105,1.5×106)w_{n}\sim\phi(10^{6},10^{4},5\times 10^{5},1.5\times 10^{6}), ϵn{0,0.02,,0.2}\epsilon_{n}\in\{0,0.02,\cdots,0.2\} for n𝒩n\in\mathcal{N}. Each experiment is repeated for 100 runs, and we show the averaged results. Algorithms 1 and 2 are used to compute the socially optimal solution and NE, respectively.

TABLE I: Model accuracy under heterogeneous noise rates.
Noise Rates (ϵn,n𝒩)(\epsilon_{n},n\in\mathcal{N}) Average Variance
Accuracy Across Runs
(0.1, 0.1, 0.1, 0.1, 0.1) 0.888 0.00044
(0.05, 0.05, 0.1, 0.15, 0.15) 0.889 0.00066
(0.04, 0.08, 0.1, 0.12, 0.16) 0.889 0.0005

In Fig. 3, we compare the socially optimal solution (denoted by “SWM”) and NE in terms of model accuracy and the PoS in social welfare (i.e., the social welfare under SWM divided by that under NE). We observe that for both random flipping and instance-dependent noises, the model accuracy and social welfare under “SWM” are higher than that under NE.444In Figs. 3(c) and 3(d), a PoS that is larger than one implies a higher social welfare under SWM than that under NE. This observation is consistent with Theorem 1 and Lemma 4. Importantly, in Fig. 3(b), the PoS in social welfare is generally larger under higher noise rate. That is, the non-cooperative behavior of clients imposes a higher efficiency loss when they have noiser data. This calls for an effective incentive mechanism to reduce the gap between socially optimal solutions and the equilibrium outcomes. Perhaps counter-intuitively, in Fig. 3(d), the impact of unit revenue ww on the PoS relies on label noise. Under instance dependent, a larger unit revenue leads to a higher necessity for motivating cooperative behaviors among clients.

Refer to caption
Refer to caption

(a)           (b)
Refer to caption Refer to caption
    (c)           (d)

Figure 3: (a) Model accuracy and (b) PoS under different values of ϵ\epsilon with ϵn=ϵ\epsilon_{n}=\epsilon for n𝒩n\in\mathcal{N}; (c) model accuracy and (d) PoS under different values of ww with wn=ww_{n}=w for n𝒩n\in\mathcal{N}.

VI Conclusion

This paper presents a first study on the PoS in quality-aware federated learning. We formulate the clients’ interactions as a novel label denoising game and characterize its equilibrium. We further derive the socially optimal solution, and find that the clients’ non-cooperative behavior at NE always lead to a worse global model. Numerical results show that as the clients’ data become noisier, the efficiency loss increases. This motivates the future work for an effective incentive mechanism as a remedy.

References

  • [1] P. Kairouz, H. B. McMahan, B. Avent, A. Bellet, M. Bennis, A. N. Bhagoji, K. Bonawitz, Z. Charles, G. Cormode, R. Cummings et al., “Advances and open problems in federated learning,” Foundations and Trends® in Machine Learning, vol. 14, no. 1–2, pp. 1–210, 2021.
  • [2] S. Ke, C. Huang, and X. Liu, “Quantifying the impact of label noise on federated learning,” Proc. AAAI Workshop on Representation Learning for Responsible Human-Centric AI, 2023.
  • [3] A. Z. Tan, H. Yu, L. Cui, and Q. Yang, “Towards personalized federated learning,” IEEE Transactions on Neural Networks and Learning Systems, 2022.
  • [4] X. Fang and M. Ye, “Robust federated learning with noisy and heterogeneous clients,” in Proc. IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022.
  • [5] J. Xu, Z. Chen, T. Q. Quek, and K. F. E. Chong, “Fedcorr: Multi-stage federated learning for label noise correction,” in Proc. IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022.
  • [6] B. Zeng, X. Yang, Y. Chen, H. Yu, and Y. Zhang, “CLC: A consensus-based label correction approach in federated learning,” ACM Transactions on Intelligent Systems and Technology, vol. 13, no. 5, pp. 1–23, 2022.
  • [7] M. Tang and V. W. Wong, “An incentive mechanism for cross-silo federated learning: A public goods perspective,” in Proc. IEEE Conference on Computer Communications, 2021.
  • [8] K. Gu, X. Masotto, V. Bachani, B. Lakshminarayanan, J. Nikodem, and D. Yin, “An instance-dependent simulation framework for learning with label noise,” Machine Learning, Jun. 2022.
  • [9] K. Donahue and J. Kleinberg, “Optimality and stability in federated learning: A game-theoretic approach,” Advances in Neural Information Processing Systems, vol. 34, pp. 1287–1298, 2021.
  • [10] N. Zhang, Q. Ma, and X. Chen, “Enabling long-term cooperation in cross-silo federated learning: A repeated game perspective,” IEEE Transactions on Mobile Computing, 2022.
  • [11] Y. Yan, X. Tang, C. Huang, and M. Tang, “Online technical report,” https://www.dropbox.com/s/kgfla39bm6gpnbe/OnlineTechnicalReport.pdf?dl=0.