This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Efficient Core-selecting Incentive Mechanism for Data Sharing in Federated Learning

Mengda Ji, Genjiu Xu, Jianjun Ge, Mingqiang Li This work was supported by the National Key Research and Development Program of China under Grant No. 2021YFA1000402 and the National Natural Science Foundation of China under Grant No. 72071159. (Corresponding author: Genjiu Xu.)Mengda Ji is with the Unmanned System Research Institute, Northwestern Polytechnical University, Xi’an 710072, China and the International Joint Research Center on Operations Research, Optimization and Artificial Intelligence, Xi’an 710129, China.Genjiu Xu is with the School of Mathematics and Statistics, Northwestern Polytechnical University, Xi’an 710129, China and the International Joint Research Center on Operations Research, Optimization and Artificial Intelligence, Xi’an 710129, China. E-mail: [email protected] Ge and Mingqiang Li are with the Information Science Academy, China Electronics Technology Group Corporation, Beijing 100086, China.
Abstract

Federated learning is a distributed machine learning system that uses participants’ data to train an improved global model. In federated learning, participants collaboratively train a global model, and after the training is completed, each participant receives that global model along with a payment. Rational participants try to maximize their individual utility, and they will not input their high-quality data truthfully unless they are provided with satisfactory payments based on their contributions. Furthermore, federated learning benefits from the cooperation of participants. Accordingly, how to establish an incentive mechanism that both incentivizes inputting data truthfully and promotes stable cooperation has become an important issue to consider. In this paper, we introduce a data sharing game model for federated learning and employ game-theoretic approaches to design a core-selecting incentive mechanism by utilizing a popular concept in cooperative games, the core. In federated learning, the core can be empty, resulting in the core-selecting mechanism becoming infeasible. To address this issue, our core-selecting mechanism employs a relaxation method and simultaneously minimizes the benefits of inputting false data for all participants. Meanwhile, to reduce the computational complexity of the core-selecting mechanism, we propose an efficient core-selecting mechanism based on sampling approximation that only aggregates models on sampled coalitions to approximate the exact result. Extensive experiments verify that the efficient core-selecting mechanism can incentivize inputting high-quality data truthfully and stable cooperation, while it reduces computational overhead compared to the core-selecting mechanism.

Index Terms:
Federated Learning, Incentive Mechanism, Data Pricing, Game Theory.

I Introduction

In recent years, the success in the field of machine learning has been mainly attributed to the collection and use of massive amounts of data. Although the demand for data in machine learning is growing, collecting and using data are still difficult. On the one hand, data owners, or participants in federated learning, have recognized the worth of their data, so they are reluctant to provide it for free. On the other hand, privacy protection regulations such as the General Data Protection Regulation (GDPR) impose strict restrictions on access to private data [1]. To address the problem of privacy protection, federated learning has been proposed, which is a cooperative framework based on collaborative training and model sharing [2]. In federated learning, the data of participants is stored locally and used to train local models. Then these local models are aggregated by a server to create a global model. Since federated learning does not require participants to upload data to the server, it prevents possible privacy leaks during the uploading, saving, and model training processes. In recent years, federated learning has attracted attention in many fields such as healthcare, finance, and the Internet of Things [3, 4, 5].

In federated learning, economic incentives are crucial for motivating participants to cooperatively train a global model [6]. In practice, participants are reluctant to truthfully input their worthy data unless they receive satisfactory payments for their high-quality data. In addition, the global model relies on the stable cooperation of all participants and the server. Therefore, a desirable incentive mechanism should also take into account cooperative contributions and truthfulness. Along with the development of federated learning, research on incentive mechanisms has attracted many scholars’ interests [7, 8, 9]. Many methods are used to design incentive mechanisms for federated learning, including auction [10, 11, 12], contract theory [13], and matching theory [14]. Game theory investigates strategic interactions among players or agents, which provides a framework to analyze the strategic behaviors of participants in federated learning. Therefore, more and more incentive mechanisms have used game-theoretic methods to design incentive mechanisms in federated learning. Such methods include Stackelberg game [15, 16], evolutionary game [17, 18], VCG-based (Vickrey-Clarke-Groves) mechanisms [19, 20, 21, 22, 23], and Shapley value [24, 25, 26, 27].

Many existing studies in federated learning and machine learning focus on designing incentive mechanisms to provide economic incentives for participants. The VCG (Vickrey–Clarke–Groves) mechanism is famous in game theory and mechanism design theory because it satisfies the properties of incentive compatibility and individual rationality. Due to its desirable properties, the VCG mechanism motivates participants to report their private information truthfully. Therefore, the VCG-based incentive mechanism has been widely used in machine learning and federated learning [19, 20, 21, 22, 23]. To reduce computational complexity, a faithful federated learning (FFL) mechanism was proposed to compute the VCG-like payment via an incremental computation [23]. It effectively reduces computational complexity, making it suitable for large-scale training scenarios. The utility of participants under the VCG-based mechanism is equal to their marginal contribution, which motivates them to input high-quality data. In addition to computing the VCG payment, allocating payments or profits based on cooperative contributions to promote stable cooperation is also a challenge in federated learning.

Cooperative game theory offers various solution concepts to determine how benefits should be allocated. The Shapley value is a classic solution in cooperative game theory with desirable properties and interpretability, which has been widely adopted in incentive mechanisms for machine learning and federated learning [24, 28, 29, 26, 27, 25]. These methods allocate payments according to the marginal surplus of each participant, which reflects the combined worth of participants’ data. Except for Shapley value, the core is also a widely used solution concept in cooperative game theory [30, 31, 32]. It represents a stable and fair allocation set that satisfies the properties of individual rationality and coalitional rationality, which is different from Shapley value. The allocation element in the core ensures that no player or coalition has the motivation to deviate from cooperation. Due to its desirable properties, the core has been widely applied in many fields such as spectrum auctions, crowdsourcing, and transportation scenarios [33, 34, 35]. Recently, the core has also been used in federated learning to evaluate the cooperative contributions of participants [36, 37]. These methods fully analyze the existence of the core and provide a fair allocation scheme. Nevertheless, the core is a set solution that cannot provide a unique optimal solution, and the advantages of the VCG payment or marginal contribution have not been fully considered by these methods. Core-selecting mechanisms, often referred to as core-selecting auctions, are designed to ensure that the resulting allocation is in the core of the associated cooperative game [38]. Therefore, we will use the core-selecting mechanism and adopt some improvements to address potential issues that may arise in federated learning. Additionally, it is worth noting that determining the exact core in federated learning also faces computational difficulties because it requires computing additional 2n2^{n} models [39].

As shown in Table I, we give the comparison table to highlight the differences of existing incentive mechanisms.

TABLE I: The differences of existing incentive mechanisms.
Cooperative
incentive
Truthful
incentive
Additional
aggregation
Additional
training
[29, 26, 27] Shapley value \ \ Yes
[36, 37] core \ Yes \
[39] least core \ \ Yes
[21] \ VCG-based Yes \
[19, 20] \ VCG-like \ Yes
[23] \ VCG-like Yes \
This paper strong ϵ\epsilon-core VCG-like Yes \

In this paper, we propose an efficient core-selecting incentive mechanism for federated learning. First, we introduce a data sharing game for federated learning. Then we introduce a core-selecting incentive mechanism that combines the advantages of both the VCG-like payment and the core, which can promote stable cooperation among players and also motivate participants to input their high-quality data. Different from the classical core-selecting mechanism, we adopt a relaxation on the core to deal with the randomness in federated learning. Since the core-selecting incentive mechanism requires exponential time to aggregate and evaluate additional models in federated learning, it is difficult to compute the exact solution. To address this, we propose an efficient core-selecting mechanism based on sampling approximation that significantly reduces computational complexity. This mechanism adjusts the aggregation weight of participants based on their historical contributions to avoid the impact of low-quality data. In addition, we verify the theoretical results of the proposed mechanism in experiments. Further experiments show that the efficient core-selecting mechanism can motivate participants to truthfully input high-quality data and promote stable cooperation, while it reduces computational overhead compared to the core-selecting mechanism.

Our contributions are as follows:

  1. 1.

    We introduce a data-sharing game for federated learning to study participants’ strategic behaviors. Based on the outcomes of this data-sharing game, we define the characteristic function and the core to measure participants’ cooperative contributions.

  2. 2.

    We employ a core-selecting mechanism for federated learning, which aims to find the optimal payment based on the core to incentivize participants’ cooperation. Due to the impact of low-quality data or overfitting on the global model, the core may be empty. To address this issue, our core-selecting mechanism employs a relaxed version of the core, strong ϵ\epsilon-core, and minimizes the benefits of inputting false data.

  3. 3.

    To avoid the core-selecting incentive mechanism implementing in an exponential time to aggregate and evaluate additional models in federated learning, we propose an efficient core-selecting incentive mechanism based on sampling approximation that significantly reduces computational complexity. This incentive mechanism aggregates new global models based on the historical contributions of participants.

The rest of this paper is organized as follows. Section II introduces the background of federated learning and formulates the problem. Section III introduces the truthful incentive mechanism for federated learning. The core-selecting mechanism is proposed in Section IV. The efficient core-selecting mechanism and its theoretical results are proposed in Section V. Experiments are given in Section VI to illustrate the performance of our method. Section VII concludes this paper.

II Background and Problem Formulation

We first introduce the background of federated learning in Section II-A, and then introduce the necessary notions and definitions in Section II-B.

II-A Background of Federated Learning

Refer to caption
Figure 1: The system of federated learning.

We introduce a typical federated learning system, as shown in Fig. 1. In this system, nn participants with data collaboratively train a machine learning model with the help of a server [40]. The training process of federated learning system typically consists of the following four steps.

  1. Step 1:

    Participants compute the model gradients locally. Mask gradients information with homomorphic encryption [40], differential privacy [41], or secret sharing techniques [42], and send the masked result to the server.

  2. Step 2:

    The server performs a secure aggregation operation on gradients information from participants, such as using weighted averaging based on data size [2, 42].

  3. Step 3:

    The server sends back the aggregated results (gradients information) to participants.

  4. Step 4:

    Participants update their local model parameters with the gradient results from the server.

The process of above steps will continue until the loss function converges. This system is independent of specific machine learning algorithms (logical regression, deep neural networks, etc.), and all participants will share the final model parameters.

In the above steps, the participant sends gradient information, and the server aggregates the gradient information using a weighted average method. Therefore, this method is called gradient averaging [43, 44]. In addition to sharing gradient information, participants in federated learning can also share models. Participants calculate model parameters locally and send them to the server [45]. The server aggregates the received model parameters (for example, calculates a weighted average) and then sends the aggregated model parameters to the participants. This method is called model averaging, as shown in Algorithm 1. Experiments have shown that model averaging is equivalent to gradient averaging, so both are called federated averaging [43].

Algorithm 1 FedAvg
Initialize global model θ0\theta^{0}.
for each round t=1,2,,Tt=1,2,\cdots,T do
     for each participant iNi\in N in parallel do
         θit=ParticipantUpdate(i,θt1)\theta_{i}^{t}=\text{ParticipantUpdate}(i,\theta^{t-1})
     end for
     Update global model θtiN1|N|θit\theta^{t}\leftarrow\sum_{i\in N}\frac{1}{|N|}\theta_{i}^{t}
end for
ParticipantUpdate(i,w)\text{ParticipantUpdate}(i,w) % Run on participant ii
ZZ\leftarrow (split DiD_{i} into batches of size BB)
for each local epoch e=1,,Ee=1,\cdots,E do
     for batch zZz\in Z do
         θθη(θ;z)\theta\leftarrow\theta-\eta\nabla\mathcal{L}(\theta;z)
     end for
end for
return θ\theta to server

It is important to note that federated learning requires stable cooperation among participants and the server. In addition, on the above system, participants are not supervised when training local models, so no one else knows whether participants are using truthful data or generated false data. In order to promote stable cooperation and encourage training with truthful data, it is necessary to design an incentive mechanism for federated learning to determine reasonable payments for participants.

Refer to caption
Figure 2: The framework of incentive mechanism for federated learning.

In this paper, we consider an incentive mechanism framework for federated learning, as shown in Fig. 2. The operator sets up a server responsible for coordinating participants to implement a federated learning algorithm, such as the FedAvg algorithm. The server measure each participant’s contribution and provide them with payments when the global model is given at each round. In the following sections, we will study some desirable incentive mechanisms based on this framework.

II-B Problem Formulation

In order to analyze federated learning from the perspective of game theory, a data sharing game environment outlining federated learning algorithm is proposed in this section. We denote the set of all participants as N={1,,n}N=\{1,\cdots,n\}, and each participant’s dataset is denoted as Di={xij,yij}j=1mD_{i}=\{x_{ij},y_{ij}\}_{j=1}^{m}. In each round tt, each participant ii updates their local model and the global model using their inputted dataset DiD_{i}. We will employ game theory to analyze the strategic behavior of these nn participants in each round.

  1. 1.

    Players: Participants N={1,,n}N=\{1,\cdots,n\} and the server 0.

  2. 2.

    Type: Each participant iNi\in N has a dataset DiD_{i} to input during the local training process, which can be called player ii’s type in game theory. Note that DiD_{i} is ii’s private information, not known to others.

  3. 3.

    Strategy: Each player iNi\in N selects D^i\hat{D}_{i} as its input, which can be truthful, i.e., D^i=Di\hat{D}_{i}=D_{i}, or untruthful, i.e., D^iDi\hat{D}_{i}\neq D_{i}. In addition, each participant can also choose to quit (D^i=\hat{D}_{i}=\emptyset). We denote D=iNDiD=\cup_{i\in N}D_{i} as the original dataset, and D^=iND^i\hat{D}=\cup_{i\in N}\hat{D}_{i} as the selected dataset.

  4. 4.

    Each player iNi\in N inputs its selected dataset to update its local model F(D^i)F(\hat{D}_{i}), and then a global model is aggregated from these nn local models, which is denoted as F(D^)F(\hat{D}). We can write the accuracy function A(F())A(F(\cdot)), where F()F(\cdot) represents a federated learning algorithm that maps an inputted dataset onto a global model and A()A(\cdot) is some accuracy metric applied to the model.

  5. 5.

    As the global model is shared among players during the update process, each participant will receive the parameters of the global model after this round. Each player iNi\in N will gain potential benefits from the global model if it is better than its local model F(Di)F(D_{i}). To characterize the immediate benefits for the current round, we define the valuation function of player ii as

    vi(F(D^),Di)=E[max{hi(F(D^))hi(F(Di)),0}],v_{i}(F(\hat{D}),D_{i})=E[\max\{h_{i}(F(\hat{D}))-h_{i}(F(D_{i})),0\}], (1)

    where hih_{i} is a pricing function for a given model, defined as a monotonically increasing function of the model accuracy. The specific form of hih_{i} is given by the participants and operator. For simplicity, we set hi()=kiA()h_{i}(\cdot)=k_{i}A(\cdot), where ki>0k_{i}>0 is a constant number that represents ii’s preference for a model. vi(F(D^),Di)v_{i}(F(\hat{D}),D_{i}) represents player ii’s benefits valuation for the global model F(D^)F(\hat{D}) when player ii’s original dataset is DiD_{i}.

  6. 6.

    Each player iNi\in N will receive a payment pip_{i}, which is to be determined. Then player ii’s utility function is defined as

    ui(F(D^),Di)=vi(F(D^),Di)+pi.u_{i}(F(\hat{D}),D_{i})=v_{i}(F(\hat{D}),D_{i})+p_{i}. (2)
  7. 7.

    The server has a budget b0b_{0} to pay for all participants. Thus, the utility of player 0 is defined as u0=b0iNpiu_{0}=b_{0}-\sum_{i\in N}p_{i}.

Note that the initial update of a participant’s local model is based on its own dataset, while subsequent updates are based on the global model aggregated in the previous round. The valuation function (1) is used to evaluate the potential benefits in each round. For participant ii, the cumulative valuation of all rounds represents the improvement of the global model compared to ii’s initial model.

A cooperative game is determined by a player set MM and a characteristic function ww. The characteristic function ww assigns a numerical value w(S)w(S) to each subset SMS\subseteq M (also called a coalition) of the player set, representing the total worth that the coalition can obtain.

In order to study the cooperative behavior of players using cooperative game methods, we need to define a characteristic function based on the outcome of the data sharing game.

In federated learning, players gain benefits from the global model. Hence, the worth generated by all players N{0}N\cup\{0\}, can be written as

w(N)=iNui(F(D^),D^i)+u0,w(N)=\sum_{i\in N}u_{i}(F(\hat{D}),\hat{D}_{i})+u_{0}, (3)

where w(N)=iNvi(F(D^),D^i)+b0w(N)=\sum_{i\in N}v_{i}(F(\hat{D}),\hat{D}_{i})+b_{0} also holds.

Assuming the model training is restricted to coalition SNS\subseteq N, and coalition SS gets a global model F(D^S)F(\hat{D}_{S}) where D^S={D^i}iS\hat{D}_{S}=\{\hat{D}_{i}\}_{i\in S}, then the worth generated by coalition S{0}S\cup\{0\} is

w(S)=iSvi(F(D^S),D^i)+b0.w(S)=\sum_{i\in S}v_{i}(F(\hat{D}_{S}),\hat{D}_{i})+b_{0}. (4)

Note that we use the notation w(S)w(S) instead of w(S{0})w(S\cup\{0\}), as the cooperation of federated learning definitely requires the server 0 and notation w(S)w(S) is more convenient to use. Thus, the characteristic function is given as follows.

w(S)={iSvi(F(D^S),D^i)+b0,S2N\,0,S=.w(S)=\begin{cases}\sum_{i\in S}v_{i}(F(\hat{D}_{S}),\hat{D}_{i})+b_{0},&S\in 2^{N}\backslash{\emptyset},\\ 0,&S=\emptyset.\end{cases} (5)

where 2N2^{N} represents all subsets of NN. To compute all characteristic values, the server has to aggregate additional models for all coalitions S2N\S\in 2^{N}\backslash\emptyset. By inputting these characteristic values, solutions in cooperative games, such as the core, can determine a payoff allocation scheme, thus incentivizing players’ cooperative behaviors.

The problem studied in this paper is to design a desirable incentive mechanism based on these 2n2^{n} characteristic values. Based on our previous discussion, this mechanism is designed to incentivize each player to truthfully input its high-quality dataset, with no player willing to deviate from cooperation. In order to design such a mechanism, the following properties will be considered.

  • Incentive Compatibility: All players input their dataset truthfully, which is a Nash strategic equilibrium. In other words, any participant has no motivation to deviate from inputting the original dataset because the utility of inputting original dataset is not less than the utility of inputting false dataset:

    ui(F(D),Di)ui(F(D^iDi),Di),iN.u_{i}(F(D),D_{i})\geq u_{i}(F(\hat{D}_{i}\cup D_{-i}),D_{i}),\ \ i\in N. (6)
  • Individual Rationality: Any player can get non-negative utility when inputting dataset truthfully:

    ui(F(D),Di)0,iN.u_{i}(F(D),D_{i})\geq 0,\ \ i\in N. (7)
  • Coalitional Rationality: Any coalition has no motivation to split off:

    iSui(F(D),Di)+u0w(S).\sum_{i\in S}u_{i}(F(D),D_{i})+u_{0}\geq w(S). (8)
  • Computational Efficiency: The incentive mechanism can be implemented in polynomial time.

The first two properties described above are intended to encourage players to input dataset truthfully. The third property is to ensure the stability of cooperation among all players and the server.

III Truthful Incentive Mechanism for Federated Learning

The VCG (Vickrey–Clarke–Groves) mechanism is famous in game theory and mechanism design theory, which satisfies the property of incentive compatibility, i.e., telling the truth is a dominant strategy for all players [46]. The VCG mechanism was previously employed by Nix and Kantarcioglu in the context of distributed machine learning [20], and they have proposed a critical assumption in their work. Meng Zhang et al. have also investigated VCG-based mechanisms in the context of federated learning and proposed a VCG-like payment to approximate VCG payments via an incremental computation. In this section, with minor modifications, we also implement this VCG-based mechanism for our incentive mechanism framework. This truthful incentive mechanism will serve as the foundation for the core-selecting mechanism in the next section.

In federated learning, a commonly used assumption is that each dataset DiD_{i} and the test dataset DtD_{t} are sampled from a real distribution 𝒟\mathcal{D}. During the training process, participant ii actually inputs an untruthful dataset D^i={x^ij,y^ij}j=1m\hat{D}_{i}=\{\hat{x}_{ij},\hat{y}_{ij}\}_{j=1}^{m}, where D^i\hat{D}_{i} deviates from the real probability distribution 𝒟\mathcal{D} and follows a probability distribution with artificial errors, i.e.,  𝒟+δ\mathcal{D}+\delta.

Let f=F(D)f=F(D) and f^=F(D^iDi)\hat{f}=F(\hat{D}_{i}\cup D_{-i}), the expected risks of these two models on distribution 𝒟\mathcal{D} are denoted as E𝒟(f)=(f(x),y)𝑑𝒟(x,y)E_{\mathcal{D}}(f)=\int\mathcal{L}(f(x),y)d\mathcal{D}(x,y) and E𝒟(f^)=(f^(x),y)𝑑𝒟(x,y)E_{\mathcal{D}}(\hat{f})=\int\mathcal{L}(\hat{f}(x),y)d\mathcal{D}(x,y) where \mathcal{L} is a per-sample loss function. Generally, E𝒟(f)E𝒟(f^)E_{\mathcal{D}}(f)\leq E_{\mathcal{D}}(\hat{f}) holds because model ff fits the data points under the real distribution 𝒟\mathcal{D}, whereas model f^\hat{f} fits the noisy data points. As a result, on the test dataset Dt𝒟D_{t}\sim\mathcal{D}, the performance of F(D)F(D) will be better than that of F(D^iDi)F(\hat{D}_{i}\cup D_{-i}). Mathematically, this can be expressed as

E[A(F(D))]E[A(F(D^iDi))]+I(dist(Xi,X^i)),E[A(F(D))]\geq E[A(F(\hat{D}_{i}\cup D_{-i}))]+I(dist(X_{i},\hat{X}_{i})), (9)

where II is a non-negative, increasing function and distdist is a distance function of the dataset.

This assumption is similar to the assumption proposed by Nix and Kantarcioglu [20]. It means that deviating from a real data will make a bad model’s performance more likely. Obviously, a truthful input will not decrease the model accuracy for each player ii, and more input of truthful data will improve model performance. Nix and Kantarcioglu also mentioned the possible effects of the overfitting phenomenon. In order to verify the validity of this assumption in the context of federated learning, we will test it by inputting a noisy dataset in Section VI.

The truthful incentive mechanism is easy to operate. Given an input D^\hat{D}, each player iNi\in N receives a global model F(D^)F(\hat{D}) and a VCG-like payment, which is defined as

piVCG=jivj(F(D^),D^j)jivj(F(D^i),D^j),iN.p_{i}^{VCG}=\sum_{j\neq i}v_{j}(F(\hat{D}),\hat{D}_{j})-\sum_{j\neq i}v_{j}(F(\hat{D}_{-i}),\hat{D}_{j}),\ \ i\in N. (10)

To compute the payments for all participants, additional nn models need to be aggregated and evaluated. Note that it is not strictly a classic VCG mechanism. The difference between this mechanism and the classical VCG mechanism is that the selection of players is not executed here. The reasons are twofold. Firstly, in order to find the optimal model and achieve a selection of players, 2n2^{n} times of aggregation and 2n2^{n} times of evaluation for all potential selection schemes are required, which are computationally expensive. Secondly, only after a participant is selected can we obtain its local model and evaluate it. Thus, the selection of participants is not considered in this paper. Although the selection is not performed by this mechanism, it still satisfies incentive compatibility.

Theorem 1 (Nix and Kantarcioglu [20])

The truthful mechanism satisfies the properties of incentive compatibility and individual rationality.

Proof:

Given any input profile DiD_{-i} of other players, if ii inputs D^i\hat{D}_{i}, player ii’s utility is

ui(F(D^iDi),Di)=\displaystyle u_{i}(F(\hat{D}_{i}\cup D_{-i}),D_{i})= vi(F(D^iDi),Di)+piVCG\displaystyle v_{i}(F(\hat{D}_{i}\cup D_{-i}),D_{i})+p_{i}^{VCG} (11)
=\displaystyle= jNvj(F(D^iDi),Dj)\displaystyle\sum_{j\in N}v_{j}(F(\hat{D}_{i}\cup D_{-i}),D_{j})
jivj(F(Di),Dj).\displaystyle-\sum_{j\neq i}v_{j}(F(D_{-i}),D_{j}).

Similarly, if ii inputs truthfully, player ii’s utility is

ui(F(D),Di)=\displaystyle u_{i}(F(D),D_{i})= jNvj(F(D),Dj)jivj(F(Di),Dj).\displaystyle\sum_{j\in N}v_{j}(F(D),D_{j})-\sum_{j\neq i}v_{j}(F(D_{-i}),D_{j}). (12)

In order for incentive compatibility to exist, this requires that

jNvj(F(D),Dj)jNvj(F(D^iDi),Dj).\sum_{j\in N}v_{j}(F(D),D_{j})\geq\sum_{j\in N}v_{j}(F(\hat{D}_{i}\cup D_{-i}),D_{j}). (13)

We assert that E[max{hi(F(D))hi(F(Di)),0}]E[max{hi(F(D^iDi))hi(F(Di)),0}]E[\max\{h_{i}(F(D))-h_{i}(F(D_{i})),0\}]\geq E[\max\{h_{i}(F(\hat{D}_{i}\cup D_{-i}))-h_{i}(F(D_{i})),0\}] always holds, since either the last expression is zero, in which case the first expression is greater than or equal to zero, and the last expression is greater than zero, in which case the first expression is greater than or equal to the last expression due to inequality (9). Therefore, we have vj(F(D),Dj)vj(F(D^iDi),Dj)v_{j}(F(D),D_{j})\geq v_{j}(F(\hat{D}_{i}\cup D_{-i}),D_{j}) and the above inequality holds. ∎

Proof:

To show that the mechanism is individually rational, we only need to show that the mechanism has a utility of at least zero. The utility of player ii is

ui(F(D),Di)=\displaystyle u_{i}(F(D),D_{i})= jNvj(F(D),Dj)jivj(F(Di),Dj)\displaystyle\sum_{j\in N}v_{j}(F(D),D_{j})-\sum_{j\neq i}v_{j}(F(D_{-i}),D_{j}) (14)
\displaystyle\geq ji(vj(F(D),Dj)vj(F(Di),Dj)).\displaystyle\sum_{j\neq i}(v_{j}(F(D),D_{j})-v_{j}(F(D_{-i}),D_{j})).

Let D^i=\hat{D}_{i}=\emptyset, according to inequality (9), we have

vj(F(D),Dj)vj(F(Di),Dj).v_{j}(F(D),D_{j})\geq v_{j}(F(D_{-i}),D_{j}). (15)

Hence,

ui(F(D),Di)=ji(vj(F(D),Dj)vj(F(Di),Dj))0.u_{i}(F(D),D_{i})=\sum_{j\neq i}(v_{j}(F(D),D_{j})-v_{j}(F(D_{-i}),D_{j}))\geq 0. (16)

The VCG-like payment measures the difference between the global model, i.e., F(D^)F(\hat{D}), and the model without his input, i.e., F(Di)F(D_{-i}). In economics, allocating benefits by marginal contributions encourages participants to adopt decisions and behaviors that have a positive impact on team benefits. If participant ii inputs data truthfully, its utility will be equal to its marginal contribution w(N)w(N\i)w(N)-w(N\backslash i), which encourages participants to input high-quality data to improve w(N)w(N). This truthful incentive mechanism, or the VCG-like payment, meets the requirements of the first two properties mentioned in Section II.

IV Core-selecting Incentive Mechanism for Federated Learning

While the truthful incentive mechanism can motivate players to input their high-quality data, it fails to consider cooperation among all players and the server. The core is a classic set solution in cooperative games. In this section, we define the core based on the outcomes of the proposed data sharing game. The core cannot be guaranteed to be nonempty in federated learning. To address this issue, we adopt a relaxed version of the core, strong ϵ\epsilon-core. Then, we propose a core-selecting mechanism that selects a solution minimizing participants’ benefits of untruthful input within the strong ϵ\epsilon-core.

Let πi=vi(F(D^),D^i)+pi\pi_{i}=v_{i}(F(\hat{D}),\hat{D}_{i})+p_{i} be the observable surplus of player ii. It does not involve the private information of players and can be computed by the server. Instead of the true utility uiu_{i}, the core is defined over surplus πi\pi_{i} because the server does not have any knowledge on the players’ original data DiD_{i}, without a guarantee on the incentive compatibility [38]. The true utility and the incentive for participants to input data truthfully will be discussed later. The server’s surplus is defined as π0=u0=b0iNpi\pi_{0}=u_{0}=b_{0}-\sum_{i\in N}p_{i}.

Definition 1 (Core)

A surplus vector π=(π0,π1,,πn)\pi=(\pi_{0},\pi_{1},\cdots,\pi_{n}) is feasible if iNπi+π0=w(N)\sum_{i\in N}\pi_{i}+\pi_{0}=w(N), which means that all players completely allocate the worth generated in federated learning. A surplus vector is blocked by coalition SNS\subseteq N if there is another π\pi^{\prime} such that πkπk\pi_{k}^{\prime}\geq\pi_{k} for all kSk\in S, and π0>π0\pi_{0}^{\prime}>\pi_{0}. The core is the set of non-negative surplus vectors that are feasible and not blocked by any coalition, which can be mathematically defined as

Core(N)={(π0,π1,,πn)+n+1|iNπi+π0=w(N),\displaystyle\text{Core}(N)=\{(\pi_{0},\pi_{1},\cdots,\pi_{n})\in\mathbb{R}^{n+1}_{+}|\sum_{i\in N}\pi_{i}+\pi_{0}=w(N), (17)
iSπi+π0w(S),SN}.\displaystyle\sum_{i\in S}\pi_{i}+\pi_{0}\geq w(S),S\subseteq N\}.

This implies that when the server selects a vector from the core, there is no motivation for any player kk to form a coalition SS with other players to deviate from coalition NN for any possible improvement of their total utility.

The classical core-selecting mechanism, as shown in quadratic programming (18) and Fig. 3, aims to find a surplus from the core, which is closest to the surplus vector determined by the VCG-like payment, πVCG=vi(F(D^),D^i)+piVCG\pi^{VCG}=v_{i}(F(\hat{D}),\hat{D}_{i})+p_{i}^{VCG}. It combines the advantages of the VCG-based mechanism and the core.

min\displaystyle\min σ2=iN(πiπiVCG)2\displaystyle\ \ \sigma^{2}=\sum_{i\in N}(\pi_{i}-\pi_{i}^{VCG})^{2} (18)
s.t. iNπi+π0=w(N),\displaystyle\ \ \sum_{i\in N}\pi_{i}+\pi_{0}=w(N),
iSπi+π0w(S),\displaystyle\ \ \sum_{i\in S}\pi_{i}+\pi_{0}\geq w(S),\ \ \ \ SN,\displaystyle S\subseteq N,
πi0,\displaystyle\ \ \pi_{i}\geq 0, iN{0},\displaystyle i\in N\cup\{0\},
Refer to caption
Figure 3: Selecting the optimal surplus vector from the core.
Theorem 2

The core is nonempty if F(D)F(D) performs better than F(DS)F(D_{S}) for any coalition SNS\subseteq N.

Proof:

We show that the surplus vector determined by the first price rule is in the core. Given any type profile DD, for each player ii, we let pi1=gi(F(D),Di)p_{i}^{1}=-g_{i}(F(D),D_{i}) and then we have πi1=0\pi_{i}^{1}=0 for all iNi\in N. For any coalition SNS\subseteq N, we have

iNπi1+π01\displaystyle\sum_{i\in N}\pi_{i}^{1}+\pi_{0}^{1} =π01=b0iNpi1\displaystyle=\pi_{0}^{1}=b_{0}-\sum_{i\in N}p_{i}^{1} (19)
=iNvi(F(D),Di)+b0\displaystyle=\sum_{i\in N}v_{i}(F(D),D_{i})+b_{0}
=w(N),\displaystyle=w(N),

and

iSπi1+π01\displaystyle\sum_{i\in S}\pi_{i}^{1}+\pi_{0}^{1} =b0+iNvi(F(D),Di)\displaystyle=b_{0}+\sum_{i\in N}v_{i}(F(D),D_{i}) (20)
iSvi(F(DS),Di)+b0\displaystyle\geq\sum_{i\in S}v_{i}(F(D_{S}),D_{i})+b_{0}
=w(S).UNKNOWN\displaystyle=w(S). 

Therefore, the surplus vector determined by the first price rule is in the core, so the core is nonempty. ∎

Intuitively speaking, with more input, F(D)F(D) should perform better than F(DS)F(D_{S}) and the core should be a nonempty set. However, there are some issues in practice. Firstly, in practical implementation, we cannot repeat training infinitely to accurately approximate the expected value of accuracy. Additionally, model overfitting may also result in E[A(F(D))]<E[A(F(DT))]E[A(F(D))]<E[A(F(D_{T}))] and even w(T)>w(N)w(T)>w(N) holding for a coalition TNT\subset N. Both two issues result in an empty core, and the core-selecting cannot be used directly. To address this, we employ a non-empty set solution similar to the core, strong ϵ\epsilon-core, which adopts a relaxation ϵ\epsilon to the core [47]. The strong ϵ\epsilon-core is defined as

Coreϵ(N)={(π0,π1,,πn)+n+1|iNπi+π0=w(N),\displaystyle\text{Core}_{\epsilon}(N)=\{(\pi_{0},\pi_{1},\cdots,\pi_{n})\in\mathbb{R}^{n+1}_{+}|\sum_{i\in N}\pi_{i}+\pi_{0}=w(N), (21)
iSπi+π0w(S)ϵ,S2N\}.\displaystyle\sum_{i\in S}\pi_{i}+\pi_{0}\geq w(S)-\epsilon,S\in 2^{N}\backslash\emptyset\}.

In the concept of strong ϵ\epsilon-core [47], ϵ\epsilon represents a minor perturbation or a threshold for deviation. If a coalition SNS\subset N can increase its utility by more than the threshold, i.e., ϵ\epsilon, through deviating from NN, then the coalition will choose to deviate and not collaborate with the participants in N\SN\backslash S. The strong ϵ\epsilon-core of a game can be interpreted as the set of all allocation vectors that cannot be improved upon by any coalition if one imposes a cost of ϵ\epsilon in all cases where a nontrivial coalition is formed [47].

For convenience, the surplus vector in the strong ϵ\epsilon-core, π\pi is called core surplus vector, and the surplus vector determined by the VCG-like payment, πVCG\pi^{VCG} is also called VCG surplus vector. The strong ϵ\epsilon-core is a set solution, while the VCG surplus vector is a single point solution. Obviously, increasing ϵ\epsilon can expand the strong ϵ\epsilon-core. A sufficiently large ϵ\epsilon can ensure the VCG surplus vector is in the strong ϵ\epsilon-core.

Theorem 3

Denote α(S)=maxTSmaxiS[(w(N)w(N\{i}))(w(T)w(T\{i}))]\alpha(S)=\max_{T\supseteq S}\max_{i\in S}\Big{[}\big{(}w(N)-w(N\backslash\{i\})\big{)}-\big{(}w(T)-w(T\backslash\{i\})\big{)}\Big{]} for any coalition SNS\subseteq N. Given ϵ0\epsilon\geq 0 satisfying ϵmaxSα(S)(|N||S|)\epsilon\geq\max_{S}\alpha(S)(|N|-|S|), then the VCG surplus vector is in the strong ϵ\epsilon-core.

Proof:

The VCG surplus vector is defined as

πiVCG=w(N)w(N\{i}),iN.\pi_{i}^{VCG}=w(N)-w(N\backslash\{i\}),\ \ \ \ i\in N. (22)

For any coalition S={1,2,,q}NS=\{1,2,\cdots,q\}\subseteq N, we show that the blocking inequality associated with coalition SS is satisfied.

iSπiVCG+π0VCG\displaystyle\sum_{i\in S}\pi_{i}^{VCG}+\pi_{0}^{VCG} (23)
=\displaystyle= w(N)i=q+1nπiVCG\displaystyle w(N)-\sum_{i=q+1}^{n}\pi_{i}^{VCG}
=\displaystyle= w(N)i=q+1n(w(N)w(N\{i}))\displaystyle w(N)-\sum_{i=q+1}^{n}\big{(}w(N)-w(N\backslash\{i\})\big{)}
\displaystyle\geq w(N)i=q+1n(w({1,2,i})w({1,2,i1})+\displaystyle w(N)-\sum_{i=q+1}^{n}\big{(}w(\{1,2\cdots,i\})-w(\{1,2\cdots,i-1\})+
ϵnq)\displaystyle\frac{\epsilon}{n-q}\big{)}
=\displaystyle= w(N)(w({1,2,n})w({1,2,q})+ϵ)\displaystyle w(N)-\big{(}w(\{1,2\cdots,n\})-w(\{1,2\cdots,q\})+\epsilon\big{)}
=\displaystyle= w(S)ϵ.\displaystyle w(S)-\epsilon.

Hence, the VCG surplus vector is in the strong ϵ\epsilon-core. ∎

The above theorem provides a lower bound on ϵ\epsilon to ensure that the VCG surplus vector is in the strong ϵ\epsilon core. If a very small lower bound can guarantee that the VCG surplus is in the core, we can directly employ the VCG-like payment to incentivize players. In the theorem mentioned above, this lower bound is difficult to calculate. Although we are uncertain whether the VCG surplus is in the ϵ\epsilon-core, we still have another way to comprehensively consider the VCG surplus and the strong ϵ\epsilon-core.

Lemma 1

For any player iNi\in N, the core surplus is not greater than πiVCG+ϵ\pi_{i}^{VCG}+\epsilon.

Proof:

Given any type profile D^\hat{D}, for any player ii, πiπiVCG+ϵ\pi_{i}\geq\pi_{i}^{VCG}+\epsilon holds, otherwise, there is at least one player kNk\in N, such that

πk>πkVCG+ϵ=w(N)w(N\{k})+ϵ,\pi_{k}>\pi_{k}^{VCG}+\epsilon=w(N)-w(N\backslash\{k\})+\epsilon, (24)

and we have the following inequality

iN\{k}πi+π0=w(N)πk<w(N\{k})ϵ,\sum_{i\in N\backslash\{k\}}\pi_{i}+\pi_{0}=w(N)-\pi_{k}<w(N\backslash\{k\})-\epsilon, (25)

which contradicts the core constraint. ∎

Theorem 4

For the core payment pi=πivi(F(D^),D^i)p_{i}=\pi_{i}-v_{i}(F(\hat{D}),\hat{D}_{i}), the amount that player ii can benefit by deviating from the truthful input strategy is less than or equal to piVCG(F(D))pi(F(D))+ϵp_{i}^{VCG}(F(D))-p_{i}(F(D))+\epsilon.

Proof:

Suppose not, there is some input D^i\hat{D}_{i} such that

(vi(F(D^iDi),Di)+pi(F(D^iDi)))\displaystyle\Big{(}v_{i}(F(\hat{D}_{i}\cup D_{-i}),D_{i})+p_{i}(F(\hat{D}_{i}\cup D_{-i}))\Big{)}- (26)
(vi(F(D),Di)+ pi(F(D)))\displaystyle\Big{(}v_{i}(F(D),D_{i})+ p_{i}(F(D))\Big{)}
>piVCG(F(D))pi(F(D))+ϵ,\displaystyle>p_{i}^{VCG}(F(D))-p_{i}(F(D))+\epsilon,

where pi(F(D))p_{i}(F(D)) is the payment when the player inputs truthfully.

Next, we have

vi(F(D^iDi),Di)+pi(F(D^iDi))\displaystyle v_{i}(F(\hat{D}_{i}\cup D_{-i}),D_{i})+p_{i}(F(\hat{D}_{i}\cup D_{-i})) (27)
>vi(F(D),Di)+piVCG(F(D))+ϵ.\displaystyle>v_{i}(F(D),D_{i})+p_{i}^{VCG}(F(D))+\epsilon.

Note that for any input, we have vi(F(D^iDi),Di)+pi(F(D^iDi))vi(F(D^iDi),Di)+piVCG(F(D^iDi))+ϵv_{i}(F(\hat{D}_{i}\cup D_{-i}),D_{i})+p_{i}(F(\hat{D}_{i}\cup D_{-i}))\leq v_{i}(F(\hat{D}_{i}\cup D_{-i}),D_{i})+p_{i}^{VCG}(F(\hat{D}_{i}\cup D_{-i}))+\epsilon for all iNi\in N due to Lemma 1. So we can enlarge the left part of the inequality as follows,

vi(F(D^iDi),Di)+piVCG(F(D^iDi))+ϵ>\displaystyle v_{i}(F(\hat{D}_{i}\cup D_{-i}),D_{i})+p_{i}^{VCG}(F(\hat{D}_{i}\cup D_{-i}))+\epsilon> (28)
vi(F(D),Di)+ piVCG(F(D))+ϵ.\displaystyle v_{i}(F(D),D_{i})+ p_{i}^{VCG}(F(D))+\epsilon.

The left part of the above inequality is the utility when the player inputs untruthfully, and the right part is the utility when the player inputs truthfully. The inequality contradicts the incentive compatibility property of the VCG-like payments. ∎

Theorem 4 shows that by reducing the distance between pip_{i} and piVCG+ϵp_{i}^{VCG}+\epsilon, we can limit the benefits of inputting false data. In other words, although we cannot find a VCG surplus vector in the strong ϵ\epsilon-core, reducing the distance between pip_{i} and piVCG+ϵp_{i}^{VCG}+\epsilon as much as possible can still motivate participants to input data truthfully. Eventually, our core-selecting incentive mechanism for federated learning is proposed as follows.

  1. Step 1:

    Before computing the payments, model F(D^S)F(\hat{D}_{S}) is trained and evaluated for each coalition SNS\subseteq N. Meanwhile, the characteristic values are calculated as

    w(S)=iSvi(F(D^S),D^i)+b0,SN.w(S)=\sum_{i\in S}v_{i}(F(\hat{D}_{S}),\hat{D}_{i})+b_{0},\ \ \ \ S\subseteq N. (29)
  2. Step 2:

    Next, the VCG surplus is determined as

    πiVCG=w(N)w(N\{i}),iN.\pi_{i}^{VCG}=w(N)-w(N\backslash\{i\}),\ \ \ \ i\in N. (30)
  3. Step 3:

    The following quadratic programming is solved to give a solution π\pi^{*}.

    min\displaystyle\min σ2=iN(πiπiVCG+ϵ)2\displaystyle\ \ \sigma^{2}=\sum_{i\in N}(\pi_{i}-\pi_{i}^{VCG}+\epsilon)^{2} (31)
    s.t. ϵ0,\displaystyle\ \ \epsilon\geq 0,
    iNπi+π0=w(N),\displaystyle\ \ \sum_{i\in N}\pi_{i}+\pi_{0}=w(N),
    iSπi+π0+ϵw(S),\displaystyle\ \ \sum_{i\in S}\pi_{i}+\pi_{0}+\epsilon\geq w(S),\ \ \ \ SN,\displaystyle S\subseteq N,
    πi0,\displaystyle\ \ \pi_{i}\geq 0, iN{0},\displaystyle i\in N\cup\{0\},
  4. Step 4:

    Finally, the payments given to participants are calculated as

    pi=πivi(F(D^),D^i),iN.p_{i}=\pi_{i}^{*}-v_{i}(F(\hat{D}),\hat{D}_{i}),\ \ \ \ i\in N. (32)

The above procedure determines the observable surplus vector that we want to achieve. In particular, if the VCG surplus vector is in the ϵ\epsilon-core, the solution of the core-selecting incentive mechanism is equal to the VCG surplus vector. Conservative participants are prefer to input data truthfully because the payment under the core-selecting mechanism may be equal to or very close to that under the truthful incentive mechanism. Even if aggressive participants attempt to input false data, Theorem 4 indicates that the additional benefits obtained by inputting false data are also minimized. Since the selected surplus vector is taken from the strong ϵ\epsilon-core and ϵ\epsilon is also reduced in the programming, there is little incentive for any coalition or any participant to deviate from cooperation.

V Efficient Core-selecting Incentive Mechanism based on Sampling Approximation

The proposed core-selecting mechanism is an ideal mechanism for federated learning. However, exact computation of the core-selecting mechanism requires evaluating the additional model with 2n2^{n} times to obtain the characteristic values before solving the quadratic programming, which is computationally expensive. To address this, our method is to sample a relatively small number of coalitions from a probability distribution, and compute the desired solution on the sampled coalitions.

In our method, a set of coalition 𝒮={S1,S2,,Sm}\mathcal{S}=\{S_{1},S_{2},\cdots,S_{m}\} is sampled from a probability distribution 𝒟\mathcal{D} and the characteristic values of these coalitions are given by aggregating and evaluating mm additional models. Then, solving the core-selecting quadratic programming over mm coalitions instead of 2n2^{n} coalitions gives an approximate core surplus vector, resulting in the following quadratic programming.

min\displaystyle\min σ2=iN(πiπiVCG+ϵ)2\displaystyle\ \ \sigma^{2}=\sum_{i\in N}(\pi_{i}-\pi_{i}^{VCG}+\epsilon)^{2} (33)
s.t. iNπi+π0=w(N),\displaystyle\ \ \sum_{i\in N}\pi_{i}+\pi_{0}=w(N),
iSπi+π0+ϵw(Sk),\displaystyle\ \ \sum_{i\in S}\pi_{i}+\pi_{0}+\epsilon\geq w(S_{k}),\ \ \ \ k=1,2,,m,\displaystyle k=1,2,\cdots,m,
πi0,\displaystyle\ \ \pi_{i}\geq 0, iN{0}.\displaystyle i\in N\cup\{0\}.

The solution π^\hat{\pi} of the above quadratic programming may not satisfy all 2n2^{n} core constraints. In order to analyze its properties, we consider that π^\hat{\pi} satisfies each core constraint with a probability, and similar to [39], we introduce the definition of δ\delta-probable core as follows.

Definition 2 (δ\delta-probable core)

A surplus vector π\pi is in the δ\delta-probable core if and only if

PrS𝒟[iSπi+π0+ϵv(S)0]1δ.\mathop{\text{Pr}}\limits_{S\sim\mathcal{D}}\Big{[}\sum_{i\in S}\pi_{i}+\pi_{0}+\epsilon-v(S)\geq 0\Big{]}\geq 1-\delta. (34)

Given any coalition SS drawn from 𝒟\mathcal{D}, the core constraint is violated with probability at most δ\delta. For the exact core-selecting surplus π\pi^{*}, we have

PrS𝒟[iSπi+π0+ϵv(S)0]=1.\mathop{\text{Pr}}\limits_{S\sim\mathcal{D}}\Big{[}\sum_{i\in S}\pi_{i}^{*}+\pi_{0}^{*}+\epsilon^{*}-v(S)\geq 0\Big{]}=1. (35)

The definition of δ\delta-probable core can explain the constraints of approximate quadratic programming (33). The next question we focus on is how to determine the sample size mm. The following theorem 5 answers this question. Before presenting this theorem, we introduce two known lemmas [48].

Lemma 2

Let \mathcal{F} be a function class from 𝒳\mathcal{X} to {1,1}\{-1,1\}, and let yy be the true value function. If 𝒢\mathcal{G} has VC-dimension dd, then with

m=O(d+log(1Δ)δ2),m=O\Big{(}\frac{d+\log(\frac{1}{\Delta})}{\delta^{2}}\Big{)}, (36)

i.i.d. samples x1,,xm𝒫\text{x}^{1},\cdots,\text{x}^{m}\sim\mathcal{P}, we have

|Prx𝒫[f(x)y(x)]1mi=1m𝟙f(xi)y(xi)|δ,\Big{|}\mathop{\text{Pr}}\limits_{\text{x}\sim\mathcal{P}}[f(\text{x})\neq y(\text{x})]-\frac{1}{m}\sum_{i=1}^{m}\mathds{1}_{f(\text{x}^{i})\neq y(\text{x}^{i})}\Big{|}\leq\delta, (37)

for all ff\in\mathcal{F} and with probability 1Δ1-\Delta.

Lemma 3

The function class n={xsign(wx):wn}\mathcal{F}^{n}=\{\text{x}\mapsto\text{sign}(\text{w}\cdot\text{x}):\text{w}\in\mathbb{R}^{n}\} has VC-dimension nn.

Theorem 5

Given a distribution 𝒫\mathcal{P} over 2N2^{N}, and δ\delta, Δ>0\Delta>0, solving the programming (33) over O((n+log(1/Δ))/δ2)O((n+\log(1/\Delta))/\delta^{2}) coalitions sampled from 𝒫\mathcal{P} gives a surplus vector in the δ\delta-probable core with probability 1Δ1-\Delta.

Proof:

Given a coalition SS sampled from 𝒫\mathcal{P}, we convert it into a vector zS=(zS,w(S),1,1)\text{z}^{S}=(z^{S},-w(S),1,1) where z{0,1}nz\in\{0,1\}^{n} and xiS=1x^{S}_{i}=1 if iSi\in S and xiS=0x^{S}_{i}=0 otherwise.

Consider a linear classifier ff defined by parameter wf=(π,1,π0,ϵ)\text{w}^{f}=(\pi,1,\pi_{0},\epsilon) where wfn+3\text{w}^{f}\in\mathbb{R}^{n+3}. If sign(wfzS)=1\text{sign}(\text{w}^{f}\cdot\text{z}^{S})=1, then we have wfzS=iSπi+π0+ϵw(S)0\text{w}^{f}\cdot\text{z}^{S}=\sum_{i\in S}\pi_{i}+\pi_{0}+\epsilon-w(S)\geq 0. Obviously, classifier f(zS)=sign(wfzS)f(\text{z}^{S})=\text{sign}(\text{w}^{f}\cdot\text{z}^{S}) can identify the core constraint for coalition SNS\subseteq N. If a linear classifier ff satisfies f(zS)=1f(\text{z}^{S})=1 for all coalitions SNS\subseteq N, then it represents a surplus vector in the ϵ\epsilon-core . This inspires us to define a class of functions to represent the δ\delta-probable core:

={zsign(wz):w=(π,1,π0,ϵ),π+n,\displaystyle\mathcal{F}=\Big{\{}\text{z}\mapsto\text{sign}(\text{w}\cdot\text{z}):\text{w}=(\pi,1,\pi_{0},\epsilon),\pi\in\mathbb{R}^{n}_{+}, (38)
π00,i=1nπi+π0=w(N)}.\displaystyle\pi_{0}\geq 0,\sum_{i=1}^{n}\pi_{i}+\pi_{0}=w(N)\Big{\}}.

This class of functions \mathcal{F} is a subset of n+3\mathcal{F}^{n+3}, and it has VC-dimension at most n+3n+3 by Lemma 3.

Suppose that we solve quadratic programming (33) on coalition samples S1,,SmS_{1},\cdots,S_{m}, which gives us a solution (π^,π^0,ϵ^)(\hat{\pi},\hat{\pi}_{0},\hat{\epsilon}) and the corresponding classifier f^\hat{f}. Note that f^(zSk)=1\hat{f}(\text{z}^{S_{k}})=1 holds for k=1,,mk=1,\cdots,m.

By Lemma 2, the following inequality holds with probability 1Δ1-\Delta.

PrS𝒫[iSπ^i+π^0+ϵ^w(S)0]\displaystyle\mathop{\text{Pr}}\limits_{S\sim\mathcal{P}}\Big{[}\sum_{i\in S}\hat{\pi}_{i}+\hat{\pi}_{0}+\hat{\epsilon}-w(S)\geq 0\Big{]} (39)
=1PrS𝒫[f^(zS)y(zS)]\displaystyle=1-\mathop{\text{Pr}}\limits_{S\sim\mathcal{P}}[\hat{f}(\text{z}^{S})\neq y(\text{z}^{S})]
=1(PrS𝒫[f^(zS)y(zS)]1mk=1m𝟙f^(xSk)y(xSk))\displaystyle=1-\Big{(}\mathop{\text{Pr}}\limits_{S\sim\mathcal{P}}[\hat{f}(\text{z}^{S})\neq y(\text{z}^{S})]-\frac{1}{m}\sum_{k=1}^{m}\mathds{1}_{\hat{f}(\text{x}^{S_{k}})\neq y(\text{x}^{S_{k}})}\Big{)}
1δ,\displaystyle\geq 1-\delta,

where 𝒫\mathcal{P} is a 2n2^{n} dimensional probability distribution and the second transition holds because f^\hat{f} and yy agree on S1,,SmS_{1},\cdots,S_{m}. ∎

Theorem 5 provides a theoretical result for our approximate method. The above incentives reasonably evaluate the contributions of participants. The VCG surplus can be seen as a marginal contribution, and surplus π^\hat{\pi} can be seen as a contribution based on the approximate strong ϵ\epsilon-core.

Next, in order to aggregate and obtain a better model, we employ a non-uniform weighted aggregation scheme. Reputation is widely used to measure the reliability of a participant based on its past behavior [49, 50, 8]. The participant with a better reputation will be assigned a greater weight. At each round tt, the surplus π^it\hat{\pi}_{i}^{t} can be determined. We calculate the reputation of each participant ii based on its contribution from round 11 to round tt. Let RitR_{i}^{t} denote participant ii’s reputation. It can be computed as

Ri=max(ϕ0,τ=1tπ^it),R_{i}=\max(\phi_{0},\sum_{\tau=1}^{t}\hat{\pi}_{i}^{t}), (40)

where ϕ0\phi_{0} is a small positive number representing the lower bound. In the next round t+1t+1, we will use RitR_{i}^{t} to aggregate the new global model as

Ft+1(D)=iNRitjNRjtFt+1(Di).F^{t+1}(D)=\sum_{i\in N}\frac{R_{i}^{t}}{\sum_{j\in N}R_{j}^{t}}F^{t+1}(D_{i}). (41)

We implement the efficient core-selecting incentive mechanism by adding additional model evaluation, and the computation process is shown in Algorithm 2.

Algorithm 2 FedAvg Core-selecting Incentive Algorithm
Input: Participants NN and their data {Di}iN\{D_{i}\}_{i\in N}, number of coalition samples mm.
Output: Accumulated payment PiP_{i}.
Initialize global model θ0\theta^{0}, reputation Riϕ0R_{i}\leftarrow\phi_{0} and accumulated payment Pi0P_{i}\leftarrow 0.
for each round t=1,2,,Tt=1,2,\cdots,T do
     for each participant iNi\in N in parallel do
         θit=ParticipantUpdate(i,θt1)\theta_{i}^{t}=\text{ParticipantUpdate}(i,\theta^{t-1})
     end for
     // Aggregate and Evaluate
     𝒮\mathcal{S}\leftarrow randomly select mm coalitions from 2N\2^{N}\backslash\emptyset
     for each coalition S𝒮{N}{N\{i}}iNS\in\mathcal{S}\cup\{N\}\cup\{N\backslash\{i\}\}_{i\in N} do
         Aggregate θStiSRijSRjθit\theta_{S}^{t}\leftarrow\sum_{i\in S}\frac{R_{i}}{\sum_{j\in S}R_{j}}\theta_{i}^{t}
         Evaluate the accuracy of model θSt\theta_{S}^{t} and compute characteristic value w(S)w(S) as (5)
     end for
     Update global model θt=θNt\theta^{t}=\theta_{N}^{t}
     Compute πiVCG=w(N)w(N\i)\pi_{i}^{VCG}=w(N)-w(N\backslash i) for iNi\in N and solve the programming as (33) to give a solution π^\hat{\pi}.
     Each participant iNi\in N receive pi=π^vip_{i}=\hat{\pi}-v_{i}.
     Update PiPi+piP_{i}\leftarrow P_{i}+p_{i} and RiRi+π^iR_{i}\leftarrow R_{i}+\hat{\pi}_{i}
end for
ParticipantUpdate(i,θ)\text{ParticipantUpdate}(i,\theta)     // Run on participant ii
ZZ\leftarrow (split DiD_{i} into batches of size BB)
for each local epoch e=1,,Ee=1,\cdots,E do
     for batch zZz\in Z do
         θθη(θ;z)\theta\leftarrow\theta-\eta\nabla\mathcal{L}(\theta;z)
     end for
end for
return ww to server

VI Experiments

The purpose of this section is threefold. First, we empirically demonstrate our theoretical results. In the second part of the experiment, we investigate the influence of inputting untruthful data on the incentive mechanism based on VCG-like payments, core-selecting payments, and efficient core-selecting payments. Moreover, we will test the efficiency of our mechanism.

All tests are performed on a computer with an Intel Xeon 5218 CPU, Nvidia Quadro RTX4000 GPU and 64GB RAM. The optimization problems are solved with Gurobi.

VI-A Validation

To verify our assumption, a participant is selected, and the participant will input noisy data as shown in Fig. 4. The MNIST dataset is chosen. 10 percent of the dataset is set as an independent test set for the server, and the remaining is divided into 5 parts for 5 participants. Based on Algorithm 1, the logistic regression, multi-layer perceptron (MLP), and convolutional neural network (CNN) models are trained for 10 rounds, respectively. The results are shown in Fig. 5. We can see that the accuracy mainly decreases as the false degree increases. Due to noise causing overfitting, the accuracy increases by around 10 percent. For this phenomenon, we believe that this type of data is also high-quality, as long as it can improve the performance of the global model.

Refer to caption
Figure 4: Data with noise.
Refer to caption
Figure 5: Accuracy of inputting noisy data.

To verify theorem 5, we randomly sample a small fraction of coalitions to implement the efficient core-selecting mechanism for a single round and determine what fraction of all coalitions satisfy the core constraints, which gives us the fraction 1δ1-\delta, referred to as core accuracy. In this experiment, we chose the smaller-scale iris dataset that only has 4 features. This makes it computationally feasible to evaluate all possible models and compute the core-selecting payment exactly. 10 percent of the dataset is set up as an independent test set for the server, and the remaining is divided into 10 parts for 10 participants. Based on Algorithm 2, we will train the logistic regression models and multi-layer perceptron (MLP) models using stochastic gradient descent. The parameters are set as b0=2b_{0}=2 and ki=2k_{i}=2. The payment, utility, and error will be calculated.

As shown in Fig. 6, even with a small number of sampled coalitions, the resultant surplus vector is still in the δ\delta-probable core.

Refer to caption
Figure 6: Core accuracy with different number of samples.

Then, as shown in Fig. 7, the errors of σ2\sigma^{2} decrease as the number of samples increases.

Refer to caption
Figure 7: σ2\sigma^{2} value with different number of samples.

Next, as shown in Fig. 8, we can see that the VCG-like payment cannot satisfy all core constraints, and the efficient core-selecting payment has a higher average accuracy compared to the VCG-like payment.

Refer to caption
Figure 8: Core accuracy with different number of participants (Logistic Regression, δ=0.3,Δ=0.3\delta=0.3,\Delta=0.3).

The above experiments show that the difference between the efficient core-selecting payment and the exact core-selecting payment decreases as the number of samples increases. When the scale of the problem is large, the efficient core-selecting mechanism can improve the computational efficiency by reducing the number of samples.

VI-B Truthfulness

In this experiment, we compare the effects of different inputs on the incentive mechanism based on the VCG-like, core-selecting, and efficient core-selecting payments. Of course, the three payments rule also corresponds to three types of aggregation weights. We still use the MNIST dataset and divide it into 5 parts for 5 participants. The parameter settings from the previous experiment. The logistic regression, multi-layer perceptron (MLP), and convolutional neural network (CNN) models are trained for 10 rounds, respectively. The parameters of the efficient core-selecting are set as δ=0.5\delta=0.5 and Δ=0.5\Delta=0.5. To simulate the input behavior of participants, we select a participant and design three strategies for the participant to input data untruthfully. Then the selected participant’s accumulated utility is computed. We repeat the entire process 10 times to compute the mean utility.

The first input strategy is adding different proportion of white noise to all training data. The utility of the selected participant is shown in Fig. 9. It is easy to see that the utility decreases with the increase in false degrees. In particular, when the noise is increased to 10 percent, the utility does not decrease but increases slightly. The reason is that adding noise to the data during the training process can prevent overfitting, which is commonly used as a regularization method.

Refer to caption
Figure 9: Actual utility (adding noise).

The second input strategy is removing a proportion of data, which is similar to the work of [39]. The third input strategy is inputting some incorrect labels, which is malicious behavior. The results are shown in Fig. 10, and we can see that the utility of the third strategy decreases more significantly than other strategies because it is more malicious than other strategies.

Refer to caption
(a) data removal.
Refer to caption
(b) incorrect label.
Figure 10: Actual utility of efficient core-selecting incentive mechanism.

Because rational participants will maximize the utility, the above experiments show that the efficient core-selecting mechanism can prevent participants from inputting low-quality or false data.

Additional experiments are conducted to compare the three mechanisms. As shown in Fig. 11, we can see that there is not much difference between the core-selecting mechanism and the efficient core-selecting mechanism.

Refer to caption
(a) data removal.
Refer to caption
(b) incorrect label.
Figure 11: Actual utility of three mechanisms (CNN, δ=0.5\delta=0.5, Δ=0.5\Delta=0.5).

VI-C Efficiency

In this experiment, we visualize the run time as the number of people increases. We still use the MNIST dataset. The parameters of the efficient core-selecting mechanism are fixed to sample coalitions (δ=0.3\delta=0.3 and Δ=0.3\Delta=0.3). The main computational overhead is that additional models need to be aggregated and evaluated in each round, so we only compare the running time of a single round. As shown in Fig. 12, as the number of participants increases, the run time of the efficient core-selecting mechanism increases linearly, while the run time of the core-selecting mechanism increases exponentially.

Refer to caption
Figure 12: Run time with different number of participant (We only calculated the run time for n10n\leq 10, while the run time for n>10n>10 is estimated).

Therefore, the efficient core-selecting mechanism can reduce computational overhead compared to the core-selecting mechanism.

VII Conclusion

In this paper, in order to motivate participants to input truthful data and promote stable cooperation in federated learning, we propose an efficient core-selecting incentive mechanism. First, we introduce a data sharing game for federated learning. Then we used a core-selecting incentive mechanism that combines the advantages of both the VCG-like payment and the core. Different from the classical core-selecting mechanism, this mechanism adopts a relaxation on the core to deal with the randomness in federated learning and minimizes the benefits of inputting false data, which can promote stable cooperation among players and also evaluate the quality of participants’ data. Since the core-selecting incentive mechanism requires exponential time to aggregate and evaluate additional models, we propose an efficient core-selecting mechanism based on sampling approximation. To avoid the impact of low-quality data, the mechanism adjusts the aggregation weight of participants based on their historical contributions. Our experiments demonstrate that the proposed mechanism can prevent participants from inputting false data. In addition, participants will not deviate from cooperation because this mechanism can achieve the desired core accuracy through sampling.

For future work, we wish to continue exploring the connection between federated learning and game theory, and develop more reasonable payments for participants. Additionally, combining reputation and participant selection may improve the performance of global model.

References

  • [1] E. Parliament and T. C. of the European Union, “The general data protection regulation (gdpr),” https://eugdpr.org, 2016.
  • [2] Q. Yang, Y. Liu, T. Chen, and Y. Tong, “Federated machine learning: Concept and applications,” ACM Transactions on Intelligent Systems and Technology (TIST), vol. 10, no. 2, pp. 1–19, 2019.
  • [3] N. Rieke, J. Hancox, W. Li, F. Milletari, H. R. Roth, S. Albarqouni, S. Bakas, M. N. Galtier, B. A. Landman, K. Maier-Hein et al., “The future of digital health with federated learning,” NPJ digital medicine, vol. 3, no. 1, p. 119, 2020.
  • [4] Q. Li, Z. Wen, Z. Wu, S. Hu, N. Wang, Y. Li, X. Liu, and B. He, “A survey on federated learning systems: Vision, hype and reality for data privacy and protection,” IEEE Transactions on Knowledge and Data Engineering, vol. 35, no. 4, pp. 3347–3366, 2023.
  • [5] Y. Lu, X. Huang, Y. Dai, S. Maharjan, and Y. Zhang, “Blockchain and federated learning for privacy-preserved data sharing in industrial iot,” IEEE Transactions on Industrial Informatics, vol. 16, no. 6, pp. 4177–4186, 2019.
  • [6] Y. Zhan, J. Zhang, Z. Hong, L. Wu, P. Li, and S. Guo, “A survey of incentive mechanism design for federated learning,” IEEE Transactions on Emerging Topics in Computing, vol. 10, no. 2, pp. 1035–1044, 2021.
  • [7] X. Tu, K. Zhu, N. C. Luong, D. Niyato, Y. Zhang, and J. Li, “Incentive mechanisms for federated learning: From economic and game theoretic perspective,” IEEE Transactions on Cognitive Communications and Networking, 2022.
  • [8] Z. Shi, L. Zhang, Z. Yao, L. Lyu, C. Chen, L. Wang, J. Wang, and X.-Y. Li, “Fedfaim: A model performance-based fair incentive mechanism for federated learning,” IEEE Transactions on Big Data, 2022, early access.
  • [9] J. Lu, B. Pan, A. M. Seid, B. Li, G. Hu, and S. Wan, “Truthful incentive mechanism design via internalizing externalities and lp relaxation for vertical federated learning,” IEEE Transactions on Computational Social Systems, 2022.
  • [10] R. Zeng, S. Zhang, J. Wang, and X. Chu, “Fmore: An incentive scheme of multi-dimensional auction for federated learning in mec,” in 2020 IEEE 40th International Conference on Distributed Computing Systems (ICDCS).   IEEE, 2020, pp. 278–288.
  • [11] Y. Jiao, P. Wang, D. Niyato, B. Lin, and D. I. Kim, “Toward an automated auction framework for wireless federated learning services market,” IEEE Transactions on Mobile Computing, vol. 20, no. 10, pp. 3034–3048, 2020.
  • [12] Y. Deng, F. Lyu, J. Ren, Y.-C. Chen, P. Yang, Y. Zhou, and Y. Zhang, “Fair: Quality-aware federated learning with precise user incentive and model aggregation,” in IEEE INFOCOM 2021-IEEE Conference on Computer Communications.   IEEE, 2021, pp. 1–10.
  • [13] J. Kang, Z. Xiong, D. Niyato, S. Xie, and J. Zhang, “Incentive mechanism for reliable federated learning: A joint optimization approach to combining reputation and contract theory,” IEEE Internet of Things Journal, vol. 6, no. 6, pp. 10 700–10 714, 2019.
  • [14] W. Y. B. Lim, J. Huang, Z. Xiong, J. Kang, D. Niyato, X.-S. Hua, C. Leung, and C. Miao, “Towards federated learning in uav-enabled internet of vehicles: A multi-dimensional contract-matching approach,” IEEE Transactions on Intelligent Transportation Systems, vol. 22, no. 8, pp. 5140–5154, 2021.
  • [15] Y. Sarikaya and O. Ercetin, “Motivating workers in federated learning: A stackelberg game perspective,” IEEE Networking Letters, vol. 2, no. 1, pp. 23–27, 2019.
  • [16] Y. Zhan, P. Li, Z. Qu, D. Zeng, and S. Guo, “A learning-based incentive mechanism for federated learning,” IEEE Internet of Things Journal, vol. 7, no. 7, pp. 6360–6368, 2020.
  • [17] X. Luo, Z. Zhang, J. He, and S. Hu, “Strategic analysis of the parameter servers and participants in federated learning: An evolutionary game perspective,” IEEE Transactions on Computational Social Systems, 2022.
  • [18] W. Cheng, Y. Zou, J. Xu, and W. Liu, “Dynamic games for social model training service market via federated learning approach,” IEEE Transactions on Computational Social Systems, vol. 9, no. 1, pp. 64–75, 2021.
  • [19] M. Kantarcioglu and R. Nix, “Incentive compatible distributed data mining,” in 2010 IEEE Second International Conference on Social Computing.   IEEE, 2010, pp. 735–742.
  • [20] R. Nix and M. Kantarciouglu, “Incentive compatible privacy-preserving distributed classification,” IEEE Transactions on Dependable and Secure Computing, vol. 9, no. 4, pp. 451–462, 2011.
  • [21] M. Cong, H. Yu, X. Weng, J. Qu, Y. Liu, and S. M. Yiu, “A vcg-based fair incentive mechanism for federated learning,” arXiv preprint arXiv:2008.06680, 2020.
  • [22] M. Cong, H. Yu, X. Weng, and S. M. Yiu, “A game-theoretic framework for incentive mechanism design in federated learning,” Federated Learning: Privacy and Incentive, pp. 205–222, 2020.
  • [23] M. Zhang, E. Wei, and R. Berry, “Faithful edge federated learning: Scalability and privacy,” IEEE Journal on Selected Areas in Communications, vol. 39, no. 12, pp. 3790–3804, 2021.
  • [24] G. Wang, C. X. Dang, and Z. Zhou, “Measure contribution of participants in federated learning,” in 2019 IEEE International Conference on Big Data (Big Data).   IEEE, 2019, pp. 2597–2604.
  • [25] R. H. L. Sim, Y. Zhang, M. C. Chan, and B. K. H. Low, “Collaborative machine learning with incentive-aware model rewards,” in International Conference on Machine Learning.   PMLR, 2020, pp. 8927–8936.
  • [26] A. Ghorbani and J. Zou, “Data shapley: Equitable valuation of data for machine learning,” in International Conference on Machine Learning.   PMLR, 2019, pp. 2242–2251.
  • [27] T. Wang, J. Rausch, C. Zhang, R. Jia, and D. Song, “A principled approach to data valuation for federated learning,” Federated Learning: Privacy and Incentive, pp. 153–167, 2020.
  • [28] T. Song, Y. Tong, and S. Wei, “Profit allocation for federated learning,” in 2019 IEEE International Conference on Big Data (Big Data).   IEEE, 2019, pp. 2577–2586.
  • [29] R. Jia, D. Dao, B. Wang, F. A. Hubis, N. Hynes, N. M. Gürel, B. Li, C. Zhang, D. Song, and C. J. Spanos, “Towards efficient data valuation based on the shapley value,” in The 22nd International Conference on Artificial Intelligence and Statistics.   PMLR, 2019, pp. 1167–1176.
  • [30] D. B. Gillies, Some theorems on n-person games.   Princeton University, 1953.
  • [31] H. E. Scarf, “The core of an n person game,” Econometrica, vol. 35, no. 1, pp. 50–69, 1967.
  • [32] L. G. Telser, “The usefulness of core theory in economics,” Journal of Economic Perspectives, vol. 8, no. 2, pp. 151–164, 1994.
  • [33] Y. Zhu, B. Li, H. Fu, and Z. Li, “Core-selecting secondary spectrum auctions,” IEEE Journal on Selected Areas in Communications, vol. 32, no. 11, pp. 2268–2279, 2014.
  • [34] J. Hu, H. Lin, X. Guo, and J. Yang, “Dtcs: An integrated strategy for enhancing data trustworthiness in mobile crowdsourcing,” IEEE Internet of Things Journal, vol. 5, no. 6, pp. 4663–4671, 2018.
  • [35] J. James and A. Y. Lam, “Core-selecting auctions for autonomous vehicle public transportation system,” IEEE Systems Journal, vol. 13, no. 2, pp. 2046–2056, 2018.
  • [36] B. Ray Chaudhury, L. Li, M. Kang, B. Li, and R. Mehta, “Fairness in federated learning via core-stability,” Advances in Neural Information Processing Systems, vol. 35, pp. 5738–5750, 2022.
  • [37] K. Donahue and J. Kleinberg, “Optimality and stability in federated learning: A game-theoretic approach,” Advances in Neural Information Processing Systems, vol. 34, pp. 1287–1298, 2021.
  • [38] R. W. Day and P. Cramton, “Quadratic core-selecting payment rules for combinatorial auctions,” Operations Research, vol. 60, no. 3, pp. 588–603, 2012.
  • [39] T. Yan and A. D. Procaccia, “If you like shapley then you’ll love the core,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 35, no. 6, 2021, pp. 5751–5759.
  • [40] Y. Aono, T. Hayashi, L. Wang, S. Moriai et al., “Privacy-preserving deep learning via additively homomorphic encryption,” IEEE Transactions on Information Forensics and Security, vol. 13, no. 5, pp. 1333–1345, 2017.
  • [41] M. Abadi, A. Chu, I. Goodfellow, H. B. McMahan, I. Mironov, K. Talwar, and L. Zhang, “Deep learning with differential privacy,” in Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security, 2016, pp. 308–318.
  • [42] K. Bonawitz, V. Ivanov, B. Kreuter, A. Marcedone, H. B. McMahan, S. Patel, D. Ramage, A. Segal, and K. Seth, “Practical secure aggregation for privacy-preserving machine learning,” in proceedings of the 2017 ACM SIGSAC Conference on Computer and Communications Security, 2017, pp. 1175–1191.
  • [43] H. B. McMahan, E. Moore, D. Ramage, S. Hampson, and B. A. y Arcas, “Communication-efficient learning of deep networks from decentralized data,” in International Conference on Artificial Intelligence and Statistics, 2016.
  • [44] C. Yu, H. Tang, C. Renggli, S. Kassing, A. Singla, D. Alistarh, C. Zhang, and J. Liu, “Distributed learning over unreliable networks,” in International Conference on Machine Learning.   PMLR, 2019, pp. 7202–7212.
  • [45] T. T. Phuong et al., “Privacy-preserving deep learning via weight transmission,” IEEE Transactions on Information Forensics and Security, vol. 14, no. 11, pp. 3003–3015, 2019.
  • [46] Y. Narahari, Game theory and mechanism design.   World Scientific, 2014.
  • [47] T. S. Driessen, Cooperative games, solutions and applications.   Springer Science & Business Media, 2013, vol. 3.
  • [48] S. Shalev-Shwartz and S. Ben-David, Understanding Machine Learning: From Theory to Algorithms.   Cambridge University Press, 2014.
  • [49] J. Kang, R. Yu, X. Huang, M. Wu, S. Maharjan, S. Xie, and Y. Zhang, “Blockchain for secure and efficient data sharing in vehicular edge computing and networks,” IEEE Internet of Things Journal, vol. 6, no. 3, pp. 4660–4670, 2018.
  • [50] X. Huang, R. Yu, J. Kang, Z. Xia, and Y. Zhang, “Software defined networking for energy harvesting internet of things,” IEEE Internet of Things Journal, vol. 5, no. 3, pp. 1389–1399, 2018.