This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

BF-Meta: Secure Blockchain-enhanced Privacy-preserving Federated Learning for Metaverse

Wenbo Liu The Department of Electrical
and Electronic Engineering
The University of Hong Kong
Hong Kong, China
[email protected]
   Handi Chen The Department of Electrical
and Electronic Engineering
The University of Hong Kong
Hong Kong, China
[email protected]
   Edith C.H. Ngai The Department of Electrical
and Electronic Engineering
The University of Hong Kong
Hong Kong, China
[email protected]
Abstract

The metaverse, emerging as a revolutionary platform for social and economic activities, provides various virtual services while posing security and privacy challenges. Wearable devices serve as bridges between the real world and the metaverse. To provide intelligent services without revealing users’ privacy in the metaverse, leveraging federated learning (FL) to train models on local wearable devices is a promising solution. However, centralized model aggregation in traditional FL may suffer from external attacks, resulting in a single point of failure. Furthermore, the absence of incentive mechanisms may weaken users’ participation during FL training, leading to degraded performance of the trained model and reduced quality of intelligent services. In this paper, we propose BF-Meta, a secure blockchain-empowered FL framework with decentralized model aggregation, to mitigate the negative influence of malicious users and provide secure virtual services in the metaverse. In addition, we design an incentive mechanism to give feedback to users based on their behaviors. Experiments conducted on five datasets demonstrate the effectiveness and applicability of BF-Meta.

Index Terms:
metaverse, blockchain, incentive mechanism, decentralized aggregation.
footnotetext: *Corresponding author: Edith C.H. Ngai.

I Introduction

The metaverse integrates virtual reality [1], artificial intelligence [2], 3D constructions [3], and other advanced digital techniques to construct a worldwide virtual community. The development of advanced applications for the metaverse, such as virtual games, virtual socialization, and remote healthcare, is transforming the way people engage with the world. Leveraging machine learning to provide intelligent personalization and optimize feedback significantly improves the quality of services (QoS) in metaverse applications. Nevertheless, the training of machine learning models relies heavily on collecting user data, which introduces significant challenges for users’ privacy and security [4]. Federated learning (FL) provides a distributed training scheme for intelligent services in the metaverse while preserving personal privacy [5].

However, assuming all participants are trustworthy in FL is unrealistic. Some malicious users may exploit the training process for personal gain or disrupt the system training performance [6, 7]. Additionally, when users enjoy intelligent services in the metaverse, wearable devices continuously collect personal information to train local models. The curious server can easily infer training data distribution and obtain users’ personal information with multiple updated local models. Furthermore, once the server is attacked, the risk of privacy leakage increases dramatically, potentially compromising the QoS and leading to system failure. Applying blockchain to FL can mitigate multiple attacks during model training while ensuring secure data storage [8, 9]. Through the design of incentive and consensus mechanisms, blockchains also encourage and monitor users to participate in FL honestly [10, 11, 12] to ensure the QoS in the metaverse.

Early blockchain-empowered FL systems are designed to record users’ credits in blocks. Tian et al. [13] propose a credit investigation solution to identify potentially malicious users by establishing a real-virtual combined credit system. The authors implement a real-name registration mechanism to transfer credits from the real world, incentivizing honest participation in FL. Kang et al. [14] apply blockchain with a designed incentive mechanism in FL to encourage users’ participation. Gupta et al. [15] investigate the deployment of blockchain for user authentication and the documentation of illegal behaviors in the metaverse. The authors propose recording illegal behaviors on the blockchain and adjusting user reputation based on these records. However, these studies lack a mechanism to verify uploaded models before aggregation to resist poisoned models and ensure the robustness of training, which is significant for models’ performance. Li et al. [16] propose a verification mechanism to ensure the performance of local model parameters. However, this verification process is time-consuming. A more efficient detection method needs to be developed for time-sensitive services in the metaverse.

To address these challenges, this study proposes BF-Meta, a novel blockchain-empowered FL framework with decentralized model aggregation, to detect malicious models before aggregation. Furthermore, we introduce a reputation-based incentive mechanism to encourage users’ honest updates during training and enhance the security of services in the metaverse. The main contributions are listed as follows:

  • We propose BF-Meta, a secure blockchain-empowered FL framework to collect data from physical wearable devices and provide services in the metaverse. A consensus mechanism in BF-Meta is designed to verify FL models and provide secure model storage to ensure the security of training even with potential malicious clients.

  • A reputation-based incentive mechanism is designed to monitor users’ behaviors, detect abnormal behaviors, and exclude malicious users during FL training.

  • We demonstrate the generalisability of BF-Meta and the stability of the framework’s performance on five datasets. The sufficient experimental results prove the effectiveness of our proposed framework.

The rest of the paper is organized as follows: Section II presents the blockchain-empowered FL framework for the metaverse. Section III details the design of the incentive mechanism. Section IV analyzes the security and experiment performance of BF-Meta. Section V concludes the paper.

II Blockchain-empowered FL Framework for metaverse

In this section, we describe the system architecture and workflow of the proposed BF-Meta.

II-A System Overview

Refer to caption
Figure 1: System overview of BF-Meta.

Fig. 1 shows the architecture of BF-Meta consisting of six layers from bottom to top: hardware, data, network, consensus, incentive, and application layers.

II-A1 Hardware Layer

The hardware layer includes multiple wearable devices and a central server. Users’ data are collected when using wearable devices and for local training. The server is responsible for verifying and aggregating local models to provide applications.

II-A2 Data Layer

In the data layer, the stored data includes on-chain and off-chain storage. The collected user information, local models, and blockchains are stored on the respective users’ wearable devices. The central server stores models from all clients for verification and aggregation and synchronizes them on the blockchain. The encrypted hash values of the trained model parameters are uploaded and stored on the blockchain along with a digital signature.

II-A3 Network Layer

In the network layer, peer-to-peer (P2P) communication is leveraged to transmit models between devices. Each update of the blockchain is synchronized through broadcasting.

II-A4 Consensus Layer

Consensus mechanisms prevent any single party or a small group of parties from controlling the entire system. To ensure data interaction security, miners in the consensus employ proof of worklayer. We also utilize an accuracy-based verification mechanism to enhance proof of work consensus and prevent poisoning.

II-A5 Incentive Layer

The incentive layer encourages clients to invest their computation resources and monitor their behaviors during training. BF-Meta utilizes reputation in the metaverse to reward and punish clients according to their behaviors.

II-A6 Application Layer

After completing the training, the models in BF-Meta can be utilized to provide various services in the metaverse, such as e-commerce, online games, advertising recommendations, and virtual socialization.

II-B Threat Models

In the metaverse, communities may involve thousands of users worldwide. We consider that all users are semi-trusted and may upload stale parameters without training to reduce the computation cost for personal gains. These users are viewed as “lazy” users. Repeated stale parameters may reduce the accuracy of the global model and hinder training convergence. Therefore, it is necessary to detect lazy participants to eliminate low-quality model parameters from model aggregation. To avoid the detection of hash values stored in blocks, the malicious clients may poison the model by manipulating historical hash values.

II-C Latency Calculation

The latency of BF-Meta consists of two parts: system latency tct_{c} and blockchain latency tbt_{b}. The system latency tct_{c} represents the total time consumption of FL until convergence, consisting of transmission latency between a client and the server (steps 2, 3, and 8 in Fig. 2), which may be affected by internet conditions and device hardware status. Therefore, the system latency can be represented as:

tc=n(tfl+tcs+tsc),t_{c}=n(t_{fl}+t_{c\rightarrow s}+t_{s\rightarrow c}), (1)

where tflt_{fl}, tcst_{c\rightarrow s} and tsct_{s\rightarrow c} represent the latency of FL training, the transmission latency from client to server, and the transmission latency from server to client in one communication rounds, respectively. nn indicates the number of communication rounds required for convergence.

Refer to caption
Figure 2: The workflow of BF-Meta. The left and right parts indicate the workflow of the client and the server sides, respectively.

The blockchain latency tbt_{b} denotes the time consumption related to blockchain verification, which includes the latency of block generation tbgt_{bg} (step 4 in Fig. 2), block consensus tbvt_{bv} (steps 5 and 7 in Fig. 2), and blockchain synchronous tbst_{bs} (step 8 in Fig. 2). The latency can be calculated by equation (2) as follows:

tb=tbg+tbv+tbs.t_{b}=t_{bg}+t_{bv}+t_{bs}. (2)

Herein, the latency of block consensus may be impacted by the difficulty of the consensus mechanism.

II-D Secure FL Training in BF-Meta

1 /* Update(k,ωkt)Update(k,\omega_{k}^{t}) on client kk*/
2 Collect training datasets from wearable devices;
3 Synchronous updated blockchain;
4 Obtain the global model;
5 foreach epoch ee do
6       foreach batch bb do
7             ωktωkt1ηδ(ωkt1;b)\omega_{k}^{t}\xleftarrow{}\omega_{k}^{t-1}-\eta\nabla\delta(\omega_{k}^{t-1};b);
8            
9       end foreach
10      
11 end foreach
12return the local model to the server for verification.
13 /* Verify(ωkt)Verify(\omega_{k}^{t}) on server */
14 Calculate the hash value of ωkt\omega_{k}^{t} (Hash(ωkt)Hash(\omega_{k}^{t})); if Hash(ωkt)Hash(\omega_{k}^{t}) is not duplicated then
15       βk=0\beta_{k}=0;
16       return send validated model ωkt\omega_{k}^{t} to miner;
17      
18 end if
19else βk=1\beta_{k}=1;
20 return error message to client kk.
21 /* Block storing Hashb(ωk)Hash^{b}(\omega_{k}) will be uploaded to the blockchain via proof of work. */
22 /* Aggregate(𝒲)Aggregate(\mathcal{W}) on server */
23 Synchronous updated blockchain;
24 foreach model ωkt\omega_{k}^{t} in model list 𝒲\mathcal{W} do
25       if Hashb(ωkt)=Hash(ωkt)Hash^{b}(\omega_{k}^{t})=Hash(\omega_{k}^{t})  then
26             αk=0\alpha_{k}=0;
27            
28       end if
29      
30 end foreach
31else αk=1\alpha_{k}=1;
32 ωt+1k=1K(1αk)(1βk)nknωkt+1\omega^{t+1}\xleftarrow{}\sum_{k=1}^{K}(1-\alpha_{k})(1-\beta_{k})\frac{n_{k}}{n}\omega^{t+1}_{k};
33 Send model ωkt+1\omega^{t+1}_{k} to miner for uploading to blockchain;
return ωkt+1\omega^{t+1}_{k} to all clients.
Algorithm 1 Training process in BF-Meta.

The operations of this blockchain-empowered FL system can be divided into eight main steps, as shown in Fig. 2. The detailed procedure is explained as follows:

  1. 1.

    Users’ historical records and preference information are collected as a private training dataset.

  2. 2.

    Users train the local models based on the private dataset with the received models from the last synchronization. If it is the first round, the initial global model is used.

  3. 3.

    Clients send their trained local model to the miner deployed with blockchain service. The transmitted data are encrypted by the hash function. Miners firstly verify the received blocks with the proof of work consensus mechanism by calculating a Nonce value that makes the hash value of the new block less than a target value and then upload the new block to the blockchain in sequence.

  4. 4.

    The server verifies the validity of the received blocks by traversing and comparing the hash values stored in the blockchain.

  5. 5.

    After verifying, qualified local models are downloaded to the server for model aggregation

  6. 6.

    The server aggregates local models to generate the global model for FL.

  7. 7.

    The updated global model is validated and uploaded to the blockchain.

  8. 8.

    The latest status of the blockchain is synchronized to all participants. By downloading updated models from the latest blockchain, clients can update their local models for the next round of training.

The step numbers correspond to numbers in Fig. 2, repeating steps 2-8 until the model converges.

III Reputation-based Incentive Mechanism

To ensure the performance of training, an incentive mechanism plays a crucial role in ensuring all clients are motivated to participate in training honestly. The incentive mechanism involves reward and punishment schemes to monitor clients’ behaviors. This section explains the design of client monitoring and incentive schemes in the proposed BF-Meta.

III-A Client Monitoring Scheme

To detect and record malicious behaviors in FL, a blockchain-enabled incentive mechanism is leveraged to assist the server in model selection for aggregation. In BF-Meta, each client is assigned a unique address for recording and identifying both clients and their respective models. Once the local parameters are received, the server traverses the blockchain to retrieve the latest model parameters uploaded by the same client from the blockchain with the unique address. After that, the server compares the hash value of the received models with the hash value of the previous models. If they match, the client will be detected as a lazy client and added to a blacklist.

To avoid false positive detections of lazy clients, a threshold for model accuracy is introduced. The threshold is set as the average accuracy of all received models in each aggregation. Upon receiving the local models, the server uses a partial test dataset to evaluate their reliability. If the assessed accuracy is lower than the accuracy threshold, the server logs this result and initiates a historical query on the blockchain to investigate potential causes. Any lazy and low-accuracy models are categorized as “low-quality” and will be excluded from model aggregation. Furthermore, the server computes the hash value of the received model and queries the blockchain to compare it with the historical hash values stored to detect data falsification. Once data falsification or repeated model is detected, the data falsification indicator (α\alpha) or lazy client indicator (β\beta) will be set as 1, respectively. The corresponding punishment will be reflected in the client’s reputation, as mentioned in Section IIIB.

III-B Incentive Mechanism

This section introduces the designed incentive mechanism in BF-Meta to motivate clients by detecting and punishing malicious behaviors through the client monitoring scheme and rewarding honest clients accordingly. A typical incentive mechanism [17] is to apply all the past performances to determine the assessment weights. From a long-term perspective, the historical performance of all clients in the vast metaverse may exhibit similarities due to analogous variations in model accuracy over time. Applying all records to determine weights can be challenging, as it may not sufficiently increase the gaps among clients’ reputations and effectively encourage their participation in FL. Therefore, we introduce a time-varying factor for the reputation mechanism, where the performance of the last model aggregation will directly determine the current reputation.

Specifically, we define two reputation factors to evaluate the weights of received models: model quality and data quantity. These two factors denote the accuracy of the local model and the size of the training datasets, respectively. The server sampled the test dataset to test the model’s accuracy and evaluate the model’s quality. The server transforms the test accuracy into data quality weight (ωk\omega_{k}). The model quality factor is calculated by:

ωk=θkθminθmaxθmin,\omega_{k}=\frac{\theta_{k}-\theta_{min}}{\theta_{max}-\theta_{min}}, (3)

where ωk\omega_{k} is the model quality factor, and θk\theta_{k} denotes the test accuracy of client kk. θmin\theta_{min} and θmax\theta_{max} represent the minimum and maximum values of test accuracy among clients selected for aggregation.

Data quantity factor (φk\varphi_{k}) is calculated depending on the size of data the client collects and utilizes. The data quantity factor can be obtained by:

φk=ŁkŁminŁmaxŁmin,\varphi_{k}=\frac{\L_{k}-\L_{min}}{\L_{max}-\L_{min}}, (4)

where φk\varphi_{k} denotes the data quantity factor, and Łk\L_{k} denotes the data size of client k. Łmin\L_{min} and Łmax\L_{max} represent the minimum and maximum values of data size among clients in FL, respectively.

Based on these two factors, the incentive mechanism is operated based on the basic reputation (RbasicR_{basic}), data quality (RqualityR_{quality}), and the data quantity (RquantityR_{quantity}) to calculate the rewarded or punished reputation in a round of model aggregation. The reputation is calculated as follows:

Rkt=(1βk)(1αk2)(Rkt1+Rbasic+Rquantityωkt+Rqualityφkt1),\displaystyle\begin{split}R_{k}^{t}=&(1-\beta_{k})(1-\frac{\alpha_{k}}{2})(R_{k}^{t-1}+R_{basic}+R_{quantity}\omega_{k}^{t}+\\ &R_{quality}\varphi_{k}^{t-1}),\end{split} (5)

where ωkt\omega_{k}^{t} and φkt1\varphi_{k}^{t-1} are factors to indicate the data quantity and model quality, respectively.

In BF-Meta, the selection of clients for the following round aggregation is closely tied to their reputation value. Specifically, normalization is employed to calculate probabilities based on reputation values. Reputation values are transformed into probabilities (010\sim 1). Clients with higher reputations have a higher probability of being selected for aggregation in the following round of training. Two reputation factors, model quality and data quantity, determine the reputation feedbacked to clients in each round. The reputation updates will be repeated until the training is finished.

IV Security Analysis and Experimental Study

In this section, we first discuss several implementation cases. Then, we evaluate the performance of BF-Meta by security analysis and experimental study.

IV-A Implementation Discussion

BF-Meta can be implemented on the metaverse to provide intelligent services.

IV-A1 Recommendation System

Metaverse is a virtual mapping of the real world. Personalized recommendation services are necessary to attract and retain users. To achieve this target, the service providers need to collect users’ browsing history and behaviors to train a model for offering personalized service recommendations. BF-Meta, with the incentive mechanism, can be implemented as a secure distributed training framework to train local models for personalized service without revealing data privacy and encourage operators to share their user information.

IV-A2 Credit Investigation

Credit investigation is crucial for assessing the risk level of users in the metaverse by reviewing their credit information from the real world. In this case, BF-Meta can be leveraged to securely train an evaluating model to evaluate the risk of user transactions to guarantee the security of the metaverse. Security of BF-Meta can effectively protect the users’ information and training process data from falsification and attack.

IV-A3 Decentralized Social Platform

BF-Meta can also be applied to the architecture of social platforms in the metaverse. By collecting users’ historical behaviors and information, operators can train local models for like-minded people’s recommendations. Users with similar models can be recommended as friends.

IV-B Security Analysis

BF-Meta can resist several attacks. We analyze the security from the following attacks.

IV-B1 Model Poisoning Attacks

In FL, model poisoning attacks aim to corrupt the shared model by falsifying the transmitting model parameter or continuously uploading untrained model parameters. With BF-Meta, every model update is verified by comparing the hash values between the received model and the previous hash value stored on the blockchain of the same client so that the repeatedly transmitted model can be detected and that client will be blacklisted. Moreover, if a malicious client poisons his model when updating. The accuracy of his model will be lower than that of other honest clients’. Once a low-quality model is detected, the reputation penalty will reduce its probability of being selected in the following rounds, thereby forfeiting the model reward.

IV-B2 Sybil Attacks

In a Sybil attack, an attacker creates multiple fake identities to gain a disproportionate influence over the network. BF-Meta can validate each participant’s identity when receiving the model. By comparing the signature with the authorization list of the blockchain, the timestamp of model uploading, and model quality, the server will identify the fake identity. If the signature of the attacker is not in the authorization list of the blockchain, it will be detected as a malicious node and blacklisted.

IV-B3 Replay Attacks

In a replay attack, an attacker intercepts a valid data transmission and retransmits it to create unauthorized effects. In this case, the speed and efficiency of training can be reduced by channel blocking and training obstruction. BF-Meta’s use of timestamps and sequence numbers in transaction records helps prevent replay attacks by ensuring that each transaction is uniquely identified and cannot be reused. When receiving model parameters, BF-Meta compares the hash values and timestamps between the current received model and the previous ones of the same client. Once the replay attacks are detected, the client will be blacklisted with its reward forfeited.

IV-C Experimental Study

IV-C1 Experiment Settings

The experiments are built with Python 3.8.0 and Pytorch running on CPU 12th Gen Intel(R) Core (TM) i7-12700H to establish BF-Meta and use Web3 to achieve the interaction between FL training and the blockchain system. The settings of the experiment are listed in Table I. We consider a metaverse platform with K=30K=30 users participating in FL with private keys. A selection rate η\eta is introduced to represent the chosen clients for model aggregation. Moreover, we set the target accuracy equal to 80%80\% as the criterion for the termination of FL training. Based on the clients’ behaviors, participants are divided into two categories: normal clients and lazy clients. Normal clients participate in training honestly, while lazy clients become “free-riders” in FL as mentioned in the threat models. We implement BF-Meta by Pytorch and a blockchain platform named Ganache based on five public datasets. The datasets and the corresponding models are detailed as follows:

  • MovieLens: we use a deep neural network (DNN) model with 5 hidden layers, and the numbers of neurons in each layer are 20, 30, 50, 30, and 20, respectively;

  • Wisconsin Breast Cancer Dataset: we use a DNN model with 6 hidden layers, and the numbers of neurons in each layer are 10, 30, 40, 40, 30, and 10, respectively;

  • Wheat Seed Dataset: we use DNN models with 4 hidden layers, and the numbers of neurons in each layer are 10, 30, 30, and 10, respectively;

  • Pima Indian Diabetes Dataset: we use a DNN model with 6 hidden layers, and the numbers of neurons in each layer are 20, 30, 50, 50, 30, and 20, respectively;

  • MNIST: we use a CNN model with 4 convolutional layers, 4 pooling layers, and 3 full-connecting layers whose neuron numbers in each are 256, 128, and 10, respectively.

TABLE I: Default experimental settings.
Parameters Setting
Target accuracy (Acc) 80%
Total clients in FL (K) 30
Malicious rate [3%, 10%, 30%, 50%]
Client rate for aggregation 50%

IV-C2 Experimental Evaluation

TABLE II: Latency over different training datasets.
Datasets Accuracy Average training latency Average blockchain latency Total latency
MovieLens 81.20 % 35.5 s 3.027 s 11min 24 s
Wisconsin Breast Cancer Dataset 80.74 % 40.43 s 2.591 s 12 min 11 s
Wheat Seed Data 81.63 % 20.4 s 2.654 s 6 min 9 s
Pima Indian Diabetes Dataset 80.17 % 47.71 s 2.431 s 12 min 30s
MNIST 83.72 % 1 min 46s 3.254 s 36 min 25s

We simulate the latency on the mentioned datasets during FL training to evaluate the stability of BF-Meta and analyze the relationship between the accuracy and malicious user ratio in clients to evaluate the performance of the detection function.

Table II shows the latency of BF-Meta on different datasets. The target accuracy is set as 80%. We evaluate the system, blockchain, and total latency as explained in section II. Fig. 3 shows the variation trends of the aggregated model in each dataset in terms of model aggregation epoch. Even if there are obvious differentiations in total latency and system latency among these datasets, average blockchain latency remains stable. In Table II, both total latency and average system latency experiments on each dataset are different, distributing from 6 min to 36 min in total latency and 20 seconds to nearly 2 min in system latency. The difference is caused by different model structures used for distinct datasets. An obvious difference occurs between MNIST, a graphic dataset, and other digital datasets because the convolutional neural network (CNN) requires more time to load, process images, and calculate gradients, while the time consumption for the same operations in the feed-forward neural network is lower. However, the latency of BF-Meta on different datasets are similar since the blockchain latency is only influenced by the efficiency of the blockchain committee, miners, and the network condition in the FL training process. Table II and Fig 3 demonstrate that the BF-Meta we proposed has excellent generalization capabilities across different training tasks.

Refer to caption
Figure 3: Accuracy of BF-Meta on five datasets.
Refer to caption
(a) FedAvg.
Refer to caption
(b) BF-Meta.
Figure 4: Accuracy over epochs of FedAvg and BF-Meta.
Refer to caption
(a) 3% malicious clients.
Refer to caption
(b) 10% malicious clients
Refer to caption
(c) 30% malicious clients.
Refer to caption
(d) 50% malicious clients
Figure 5: Comparison of accuracy of FedAvg and BF-Meta with distinct rates of malicious users.

Fig. 4 shows the comparison between the model accuracy of FedAvg and BF-Meta with distinct malicious client rates in 30 participants and a fixed selection rate of 50% in FL. In Fig. 4(a) and Fig. 4(b), as the rate of malicious clients increases, fluctuations in the aggregated model accuracy intensify in the absence of lazy client detection. Conversely, the accuracy fluctuation of BF-Meta is reduced. The reason is that client selection during model aggregation is completely random. It means that each client has the same possibility to be selected in model aggregation in traditional FL. The aggregated model accuracy will be impacted when malicious clients are selected for aggregation. Moreover, more malicious clients are likely to be selected in model aggregation with the increasing number of malicious clients amount, so the fluctuation in 50% malicious clients is more dramatic than that in other rates.

To detail the difference, Fig 5 illustrates the accuracy comparison of distinct malicious client rates. As shown in Fig. 5(a) and Fig. 5(b), the accuracy of FedAvg is slightly lower than that of BF-Meta. As shown in Fig. 5(c) and Fig. 5(d), with the increasing malicious rate, the superiority in accuracy becomes more obvious. In Fig. 5(c), the accuracy of BF-Meta is 12% higher than FedAvg since BF-Meta detects and defends against attacks by lazy clients. As shown in Fig. 5(d), the accuracy gap between BF-Meta and FedAvg is much more obvious. The model accuracy of BF-Meta increases stably along the training epoch, while FedAvg cannot converge at all due to lazy clients. The experimental results demonstrate that malicious clients occupy a large proportion of clients in FL so that more bad parameters are uploaded in model aggregation. For BF-Meta with the client monitoring scheme, malicious clients are blacklisted once they upload the repeated parameters. Correspondingly, the reputation will be zeroed until the end of FL. For epoch 5 in Fig. 5(d), the decrease in accuracy is caused by the selection of malicious clients in model aggregation. Once lazy clients are detected and blacklisted, the model accuracy increases again. These experimental results demonstrate the client monitoring scheme in BF-Meta can effectively avoid the influence of malicious clients and improve accuracy.

Refer to caption
(a) Initial reputation.
Refer to caption
(b) Reputation after one round.
Refer to caption
(c) Reputation after two rounds.
Figure 6: Variations in client reputation over rounds. Herein, the blue bars represent normal clients while the red bar represents the malicious client (client index = 3).

Fig. 6 illustrates the variation in clients’ reputations to evaluate the incentive mechanism. In Fig. 6(a), at the beginning of FL, all the clients are given the same reputations. Each client has an equal probability of being selected for model aggregation, even if the malicious client is marked in red (index 3). Fig. 6(b) shows the reputations of attack detection of malicious clients after one round. When detecting the inequality of the hash value of the malicious client (index 3), the incentive mechanism sets its reputation to zero so that the malicious client (index 3) will not be selected at the next model aggregation. In the meantime, the reputation of the normal clients selected in model aggregation increases, so that these clients are more likely to be selected in the next model aggregation than others. In this case, the malicious model parameters can be rejected from model aggregation. Fig. 6(c) indicates the reputations after two rounds, where the reputation of the malicious client (index 3) is much lower than others. The minor increase in malicious client (index 3) reputation is caused by the basic reputation in FL. However, the probability of the malicious client (index 3) being selected is much less than the others.

V Conclusion

In this paper, we propose BF-Meta, a secure blockchain-empowered FL framework for the metaverse that incorporates incentive and malicious client monitoring schemes. It can efficiently resist model poisoning attacks and detect lazy clients, thereby ensuring the security of FL training tasks. Furthermore, we introduce a reputation-based incentive mechanism to encourage users to participate honestly in FL training by providing feedback on their behaviors. Overall, BF-Meta provides a secure framework to support intelligent services in the metaverse without revealing private user information.

Acknowledgement

This research was supported by Meta AR/VR Policy Research Fund, HKU-SCF FinTech Academy, and Hong Kong General Research Funds (grant no. 17203320, 17209822).

References

  • [1] I. Wohlgenannt, A. Simons, and S. Stieglitz, “Virtual reality,” Business & Information Systems Engineering, vol. 62, pp. 455–461, 2020.
  • [2] T. Huynh-The, Q.-V. Pham, X.-Q. Pham, T. T. Nguyen, Z. Han, and D.-S. Kim, “Artificial intelligence for the metaverse: A survey,” Engineering Applications of Artificial Intelligence, vol. 117, p. 105581, 2023.
  • [3] X. Wu, P. Dai, W. Deng, H. Chen, Y. Wu, Y.-P. Cao, Y. Shan, and X. Qi, “Cl-nerf: continual learning of neural radiance fields for evolving scene representation,” Advances in Neural Information Processing Systems, vol. 36, 2024.
  • [4] Y. Otoum, N. Gottimukkala, N. Kumar, and A. Nayak, “Machine learning in metaverse security: Current solutions and future challenges,” ACM Computing Surveys, vol. 56, no. 8, pp. 1–36, 2024.
  • [5] C. Zhang, Y. Xie, H. Bai, B. Yu, W. Li, and Y. Gao, “A survey on federated learning,” Knowledge-Based Systems, vol. 216, p. 106775, 2021.
  • [6] S. Li, E. Ngai, F. Ye, and T. Voigt, “Auto-weighted robust federated learning with corrupted data sources,” vol. 13, no. 5, 2022.
  • [7] S. Li, E. Ngai, and T. Voigt, “An experimental study of byzantine-robust aggregation schemes in federated learning,” IEEE Transactions on Big Data, pp. 1–13, 2023.
  • [8] M. Xu, Y. Guo, C. Liu, Q. Hu, D. Yu, Z. Xiong, D. Niyato, and X. Cheng, “Exploring blockchain technology through a modular lens: A survey,” ACM Computing Surveys, vol. 56, no. 9, pp. 1–39, 2024.
  • [9] A. B. Mahammad and R. Kumar, “Scalable and security framework to secure and maintain healthcare data using blockchain technology,” in 2023 International Conference on Computational Intelligence and Sustainable Engineering Solutions (CISES), pp. 417–423, IEEE, 2023.
  • [10] Z. Zheng, S. Xie, H.-N. Dai, X. Chen, and H. Wang, “Blockchain challenges and opportunities: A survey,” International journal of web and grid services, vol. 14, no. 4, pp. 352–375, 2018.
  • [11] Z. Ning, H. Chen, X. Wang, S. Wang, and L. Guo, “Blockchain-enabled electrical fault inspection and secure transmission in 5g smart grids,” IEEE Journal of Selected Topics in Signal Processing, vol. 16, no. 1, pp. 82–96, 2022.
  • [12] J. Deng, H. Chen, Y. Chan, and E. Ngai, “M2f: Multi-centered fairness-aware federated learning framework,” in 19th EAI International Conference on Heterogeneous Networking for Quality, Reliability, Security and Robustness, 2023.
  • [13] X. Tian, J. Zhou, Q. Chen, and Z. Xiao, “A blockchain-based solution of credit investigation in metaverse,” in 2023 2nd International Conference on Artificial Intelligence and Computer Information Technology (AICIT), pp. 1–4, IEEE, 2023.
  • [14] J. Kang, D. Ye, J. Nie, J. Xiao, X. Deng, S. Wang, Z. Xiong, R. Yu, and D. Niyato, “Blockchain-based federated learning for industrial metaverses: Incentive scheme with optimal aoi,” in 2022 IEEE International Conference on Blockchain (Blockchain), pp. 71–78, IEEE, 2022.
  • [15] A. Gupta, H. U. Khan, S. Nazir, M. Shafiq, and M. Shabaz, “Metaverse security: Issues, challenges and a viable zta model,” Electronics, vol. 12, no. 2, p. 391, 2023.
  • [16] Y. Li, C. Chen, N. Liu, H. Huang, Z. Zheng, and Q. Yan, “A blockchain-based decentralized federated learning framework with committee consensus,” IEEE Network, vol. 35, no. 1, pp. 234–241, 2020.
  • [17] Y. Zhao, J. Zhao, L. Jiang, R. Tan, and D. Niyato, “Mobile edge computing, blockchain and reputation-based crowdsourcing iot federated learning: A secure, decentralized and privacy-preserving system,” arXiv preprint arXiv:1906.10893, pp. 2327–4662, 2019.