Split Learning without Local Weight Sharing
To Enhance Client-side Data Privacy
Abstract
Split learning (SL) aims to protect user data privacy by distributing deep models between the client-server and keeping private data locally. In SL training with multiple clients, the local model weights are shared among the clients for local model update. This paper first reveals data privacy leakage exacerbated from local weight sharing among the clients in SL through model inversion attacks. Then, to reduce the data privacy leakage issue, we propose and analyze privacy-enhanced SL (P-SL) (or SL without local weight sharing). We further propose parallelized P-SL to expedite the training process by duplicating multiple server-side model instances without compromising accuracy. Finally, we explore P-SL with late participating clients and devise a server-side cache-based training method to address the forgetting phenomenon in SL when late clients join. The experiment results demonstrate that P-SL helps reduce up to of client-side data leakage, which essentially achieves a better privacy-accuracy trade-off than the current trend by using differential privacy mechanisms. Moreover, P-SL and its cache-based version achieve comparable accuracy to baseline SL under various data distributions, while incurring lower costs for computation and communication. Additionally, caching-based training in P-SL mitigates the negative effect of forgetting, stabilizes the learning, and enables practical and low-complexity training in a dynamic environment with late-arriving clients.
Index Terms:
Distributed training, split learning, privacy preservation, model parallelism, neural networks.I Introduction
Deep learning (DL), influenced by the rapid growth of data, is becoming increasingly important in our daily lives. However, the privacy of data used in the model needs to be protected as required by various privacy regulations [1]. Split learning (SL) [2, 3, 4, 5] is one popular collaborative learning technique that aims to protect user privacy by enabling model training without exposing users’ raw private data. In a simple vanilla setting, SL divides a deep model into two parts deployed between a client (data owner) and a server (computing service), where only smashed data (local model part’s output after feeding raw data) is exposed for collaborative training with the server part [2]. Compared to federated learning (FL) [6], SL is suitable for DL applications on resource-constrained devices (e.g., IoT, mobile) because the clients only need to run the first few layers of the deep model, while the server handles the rest, which involves the most costly computations. With the growing availability of different sources of data, SL has been extended to process learning on multiple clients [2, 7, 8, 9]. In [10, 11], the authors conduct a comprehensive evaluation of SL across various scenarios, ranging from a low to a high number of clients, balanced to imbalanced and extreme data distributions, etc., to provide a thorough insight into SL.

Regarding SL on multiple data sources, clients typically share their local weights with others to aggregate the learned knowledge from different data sources. This can be done by sequentially passing weights to the next client [2] or by averaging all local weights at the client side [9]. In these settings, it is assumed that only the server is semi-trustworthy (honest-but-curious [12]) while all clients trust each other. However, if a client is malicious and colludes with the server, sharing local weights can lead to potential data leakage. Fig. 1 illustrates an example of data leakage in the original SL [2] with two clients, and . In this scenario, acts as the victim, while is an adversary capable of colluding with the server. employs the model inversion attack [13] to train a decoder [13, 14] using ’s shared local weights. This decoder is then utilized by to reconstruct raw data given ’s smashed data, which becomes exposed during training or inference as the nature of SL. The smashed data can be acquired by through collusion with the server or eavesdropping on the communication of . Furthermore, a decoder trained on ’s local model could be utilized to attack the subsequent client, which receives ’s local weights for model updates (e.g., if available). This situation exemplifies the white-box setting of the model inversion attack in [13], where the target (local) model is publicly accessible to the nearby/adjacent adversaries due to local weight sharing. Further details about the white-box model inversion attack and its high efficiency can be found in [13, 15]. To address these privacy concerns, we raise the research question (RQ): How to develop novel effective SL-based training methods to minimize data leakage in multi-client SL? In response to this question, we propose a privacy-enhanced SL (P-SL) scheme, which fundamentally obviate sharing local weights among clients during training. The proposed P-SL not only preserves client-server privacy of the original SL but also enhances data privacy at the client side. To the best of our knowledge, this work is the first to identify data leakage in SL exacerbated by the default local weight sharing and investigate P-SL performance under various data distributions.
Furthermore, in SL, apart from the issue of data leakage among clients, ensuring the commitment of all clients to participate in the training process simultaneously poses a significant challenge [11]. Due to various network, energy, and resource constraints, some devices may not be active throughout the entire training process or may join the training process at a later stage, after collaborative training has already concluded. Handling the training of new clients who join after the initial training, referred to as newcomer training, presents a challenge. This raises another RQ, How to ensure stable, low complexity, and high accuracy P-SL in dynamic environments where additional clients join later? As the first training cycle has already been completed, the learning of new clients can deteriorate the knowledge learned by existing clients, leading to the phenomenon of forgetting [16] as recognized in [11]. To overcome this challenge, we devise a cache-based approach to address the forgetting phenomenon, thus enhancing the learning performance of P-SL. In summary, the contributions of this paper are as follows:
-
•
We consider a stronger adversarial threat model, where all participants (both clients and the server) are assumed to be honest-but-curious. Based on this threat model, we reveal the issue of client-side data leakage in the original SL and its variants through the lens of model inversion attacks.
-
•
To address the privacy concerns under the defined threat model, we propose P-SL, which significantly reduces data leakage by up to compared to the original SL. In P-SL, clients no longer share their local weights but can still collaborate with the server to leverage their local knowledge and improve training effectiveness.
-
•
We conduct a comprehensive empirical evaluation using various datasets with different data distributions to demonstrate that the learning performance of P-SL is comparable to that of the original SL. Additionally, we propose a parallelized version of P-SL that enables simultaneous learning by clients (clients training is performed sequentially in the original SL), reducing the training time without sacrificing accuracy.
-
•
To tackle the forgetting phenomenon experienced by existing clients when new clients join the training process, we propose a server-side caching approach for P-SL. The approach allows the training of newly arriving clients without the need for retraining existing clients, thereby reducing training overhead. Our experiment results highlight the advantages of caching in SL, particularly in dynamic training environments.
The rest of this paper is structured as follows: Section II provides background information on SL and its variants, different local data distributions, and the current state of research on privacy preservation in SL. Section III presents the identified threat model underlying our proposed P-SL approach. In Section IV, we explore the parallelization of P-SL and propose the cache-based approach, which handles newly arriving clients to ensure the reliability of P-SL. Section V presents the experiment results in terms of accuracy and the privacy preservation capability of the proposal, followed by the conclusion and directions for future research in Section VI.
II Background
This section presents background information on SL with its variants for multiple clients, distribution of user data, and current research on privacy preservation for SL.
II-A Vanilla split learning
A deep model, denoted as , is a hypothesis that maps an input to an output . Model training involves finding the parameters (weights) that accurately capture the relationship between and . In order to ensure user data privacy during model training, SL [2] divides the layers of the deep model into multiple parts. In a simple vanilla setting, the model is split into two components: . The localized part contains the initial layers, while the remaining part is deployed on the server, which is typically the most computationally intensive component. During training, the client performs forward propagation on its local data batch and sends the resulting output (referred to as intermediate data or smashed data) along with the corresponding ground-truth labels to the server. The server continues forward propagation on the received smashed data to compute the loss between and . The gradients of the loss function are then backpropagated at the server until the split layer, at which point the deep model is cut/split. The split layer’s gradients are sent back to the client, where the remaining backpropagation is performed locally, all the way to the first layer. Based on the computed gradients, both the client and the server update their respective weights, and . This process, known as simple vanilla SL, serves as the core mechanism for many other variants, including SL with multiple clients and our proposed approach.
II-B SL with multiple clients
SL can be extended to train a deep model on clients. The deep model is also split into two parts, , where is distributed to all clients ( to client ), and is deployed on the central server. The training procedure involves utilizing data from multiple clients in a round-robin fashion. In this setting, when the training process of client is completed, client receives the weights of to initialize its own weights . Then, client continues training on its own data, collaborating with the server following the vanilla setting. Once the training is finished, client shares its trained weights, , with the next client [2]. This process continues until the last client completes training, and the weights trained by client are the model weights that are passed back to all clients for inference.
The model training process in SL is typically performed sequentially among the clients, which can introduce significant latency. To address this issue, several studies have focused on improving the training speed. In [8], the authors set up the mini-batch of each client proportional to its local data size, allowing for parallel processing of the training model. All clients are initialized with the exact weights, and after each iteration, the gradients are averaged before being updated on the clients. This synchronization strategy ensures that all clients have the same model weights during training.
SplitFed learning (SFL) [9] is an innovative approach that combines the strengths of FL and SL. In SFL, clients perform forward propagation in parallel on their respective data and send the smashed data to a central server. Upon receiving the gradients from the server, the clients execute the backpropagation step and then send the updated weights to a Fed server. The Fed server aggregates the weight updates using an averaging function () and disseminates a single update to all clients. Similar to [8], in SFL, after each global epoch, clients synchronize their models with identical weights, which also renders them susceptible to model inversion attacks in a white-box setting such as an adversarial client which possesses the same model weights as the victims.
II-C Privacy-enhancing SL approaches
Critical privacy vulnerabilities of SL are based on the fact that a neural network is naturally predisposed to be functionally inverted [17]. That is, the smashed data exposed by clients may be exploited to recover the raw input data. Therefore, privacy protection techniques in SL typically aim to minimize data leakage from the smashed data. For example, the noise defense approach [18, 19] applies additive Laplacian noise to the smashed data before sending it to the server. By introducing noise, the target model is no longer a one-to-one function, making it harder for an attacker to learn the mapping from smashed to raw data. Another method involves adding latent noise through binarization [20] to reduce the correlation between smashed and input data. However, these mechanisms require efforts to mitigate the impact of noise perturbation on model accuracy [21].
The work [22] aims to reduce raw data leakage by adding an additional distance correlation-based loss term to the loss function. This distance correlation loss minimizes the correlation between the raw and smashed data, ensuring that the smashed data contains minimal information for reconstructing the raw data while still being valuable for achieving model utility. In [20], the authors extend this idea by suggesting that the additional loss term can be any leakage metric, not limited to distance correlation. However, the application of an extra loss term may still result in privacy leakage because the smashed data exposes too much information to be adequately protected by a single leakage metric in the loss function [17]. To overcome this limitation, the authors in [23] propose client-based privacy protection, which employs two different loss functions computed on the client and server sides. In line with this approach, [24] designs a framework consisting of two steps: a pre-training step that establishes a feature extractor with strong model-inversion resistance, and a resistance transfer step that initializes the client-side models using the feature extractor. This framework requires sufficient computational resources for pre-training on a source task and may be vulnerable during the early training epochs. Similarly, [25] proposes a learnable pruning filter for selectively removing channels in the latent representation space at the split layer (smashed data) to prevent various state-of-the-art reconstruction attacks. However, the pruned network in [25] may not be suitable for deployment on low-end clients due to the requirement of a large amount of computing power. To preserve both data privacy and label information, [26] employs sample diversity to mix the smashed data of client-side models and create obfuscated labels before transmitting them from clients to the server. The mixed smashed data maintains a low distance correlation with the raw data, thereby preventing private data from being individually reconstructed. However, it should be noted that this mixing technique does not effectively reduce data leakage as intended when performing inference on a single data sample.
In multi-head SL (MHSL) [27, 28], the authors explore the feasibility of SFL without requiring client-side synchronization. The objective is to reduce the extra communication and computation overhead at the client side using synchronization. The study further considers the case which the server gains information of raw data through the clients’ smashed data, but the evaluation does not determine which approach, MHSL or SFL, results in less information leakage. Moreover, the analysis solely focuses on leakage from visualizing smashed data, which exhibits limited quality as demonstrated in [20]. Federated Split Learning (FSL) in [23] is another study that considers not sharing clients’ neural network weights to speed up the training time. The authors propose setting up the same number of clients and edge servers for parallelizing the training with the aim of reducing learning latency. In contrast, our approach prioritizes privacy and we design the non-local-weight-sharing mechanism to address the privacy attacks under an extended threat model. Additionally, we evaluate our scheme across multiple scenarios and data distributions.
II-D SL under diverse data distributions
In general, data is often distributed among clients in an imbalanced manner. For example, some sensors may be more active than others, resulting more data from certain sources. Similarly, in domains like healthcare, larger institutions tend to have more patient data available [11]. In a classification task, under balanced data distribution, each client holds samples from all classes in similar quantities. However, in an imbalanced data distribution, each client still possesses samples from all classes, but the total number of samples for each class is imbalanced. It is important to note that the ratio of samples between classes at each client remains similar to the overall dataset ratio. In the study from [11], the authors investigate three different data distributions for user data: balanced, imbalanced, and non-IID (non-independent identically distributed). Their findings reveal that SL performs well (compared to FL) under both balanced and imbalanced data while being highly sensitive to non-IID data. Therefore, this study mainly focuses on investigating and evaluating SL specifically under balanced and imbalanced data settings.
III Privacy-enhanced split learning
We define a threat model as the underlying context for the proposed P-SL and the analysis of data leakage.
III-A Threat model
In traditional SL, only the server is assumed to be honest-but-curious [29], meaning it follows the training procedure but may have a curiosity about the raw data from clients. This threat model is assumed in most of the literature on privacy protection techniques in SL. In our study, we extend this assumption to include honest-but-curious clients as well. To the best of our knowledge, our work is the first to consider privacy attacks from both honest-but-curious clients and server in SL.
Model Inversion Attack. Under this identified threat model, any participant in collaborative learning can utilize model inversion attacks to reveal the private data of a target user (client). The model inversion attack consists of three phases: 1) gathering/generating a training set; 2) training a decoder (inverse network [15]) with the data; and 3) recovering raw data from smashed data using the decoder. The attacker can be any adversarial client or the server, as outlined in the following scenarios, with the assumption of client-server collusion for sharing smashed data or querying the local model. Note that due to local weight sharing, all clients in SL have similar local model.
-
•
Attacker as a client: An adversarial client trains the decoder on data generated using its raw data and its local model. Subsequently, the raw data of victim clients can be reconstructed from the smashed data received from the server.
-
•
Attacker as the server: The curious server generates training data by (black-box) querying the adversarial client with a bag of samples (of the same type as raw data [15]). This data is then used to train the decoder, which can be employed to reconstruct raw data from any client’s smashed data.
We define the leakage of user data as the disparity between the reconstructed data and the original private raw data of a client. The quality of the reconstruction can be evaluated using various metrics such as mean squared error (MSE), structural similarity index measure (SSIM), peak signal-to-noise ratio (PNSR), Kullback–Leibler divergence, and so on [13, 30, 20, 31]. In our work, we utilize SSIM as the main metric for measuring data leakage.
III-B The proposed algorithm

To lessen the adverse effects of model inversion attacks, we propose a non-local-weight-sharing method at the client side for SL. The proposed P-SL algorithm is presented in Alg. 1, followed by computation analysis.
Fig. 2 illustrates the architecture of P-SL, highlighting its differences from SL and SFL. The proposed P-SL is based on multi-client SL, where multiple clients connect to a central server, and there is no communication among themselves (such as sharing snapshots [2] - local weights) or the use of a Fed server (for local model aggregation [9]). Alg. 1 presents the collaborative training procedure between clients and the server in the proposed P-SL. In the initial phase, clients and the server receive their corresponding parts, and , from a split model, . Then, they initialize their model weights, and . During a global epoch, following a round-robin manner, each client starts its training with the server, following the simple vanilla SL procedure, which is demonstrated by the inner while loop (lines ). Note that the box (lines ) represents the operations executed at the server, and the transmission of data between clients and the server (e.g., smashed data, labels, gradients, etc.) is done via a network connection. Once the training of clients is completed, we have different local models combined with the server model to form distinct deep models (i.e. where ). After training, each client performs inference on its live data using its local private model in combination with the shared server model. In contrast to SL and SFL, P-SL maintains the client-server collaboration but prohibits weight exchanges among clients, thereby reducing local computation, which will be examined in the following.
III-C Computation analysis
The convergence guarantee of pure SL on homogeneous data is straightforward and can be reduced to the case of SGD [32]. However, the non-local-weight-sharing of P-SL introduces additional challenges due to the presence of multiple different local models that can be individually combined with the server model, denoted as for training. Let us consider as the primary deep model being trained on the smashed data of all clients. If the training (smashed) data remains stable (with minimal changes), the convergence of P-SL can be simplified to the SGD case. Therefore, the convergence of P-SL relies on the convergence of each client. After a few training rounds, when a client’s local model has converged, indicating small local weight updates, the corresponding smashed data from that client also stabilizes. Stable training data facilitates the convergence of the server-side model’s training. Additionally, the learning of the server-side model is influenced by the size of the training data, specifically the size of the smashed data. A larger smashed data size leads to a more extensive training dataset, resulting in a more complex server-side model (e.g., more layers) required to learn and memorize. This insight aligns with the experiment findings in [27, 28], where the authors evaluate model accuracy with different cutting points (which determine the division of the deep model for deployment on clients and the server, respectively). Performance tends to degrade when the client model is thicker (possessing more local layers) because it requires more time to converge, and the corresponding smashed data size also increases. However, for low-end devices, it is preferable to have thin client models, which will facilitate the learning performance of P-SL, as previously discussed.
Scheme | Computation | Communication |
---|---|---|
SL | ||
SFL | ||
P-SL |
To analyze total computation and communication costs of P-SL and provide a comparison to SL and SFL, we assume a balanced data distribution for simplicity. Let us consider the following variables: as the number of clients, as the total number of dataset items, as the size of the split layer (the last layer of ), as the computation cost for processing one forward and backward propagation on with one data item, as the cost for updating a client’s local weights from the received weights from the previous client or the Fed server, and as the size of the local model . Table I demonstrates that the formulated computation and communication costs at the client side in P-SL are lower than SL and SFL, respectively, due to the absence of local weight sharing. The reduction in costs depends on the size of the local model and is independent on data distribution. The factor of in the communication costs represents the uploading of smashed data and the downloading of corresponding gradients (), or the uploading and downloading of local models () at the client side.
IV P-SL for scalable and dynamic environments
In this section, we further investigate and propose approaches to enhance the performance of P-SL in a dynamic environment where multiple server instances exist or newly participating clients arrive.
IV-A Parallelizing P-SL with multiple server instances

In the proposed P-SL, each client conducts collaborative training with the server separately. This results in high latency and large idle time at the client side, as only one client is active during training. Therefore, we can process clients’ training simultaneously if multiple server instances are available. In Fig. 3, paralleling P-SL follows the steps below in each round:
-
1.
Setup phase: During this phase, all clients receive the same model , and the server starts with model . The server sets up a pool of instances (in this example, there are two instances).
-
2.
Client computation: Clients connect to the server and are associated with available instances. Then, they perform forward propagation on their local models using their local data in parallel and independently. After this, they send their smashed data to the server.
-
3.
Server computation: The corresponding server instances perform forward-backward operations on the received smashed data from the clients and send back the computed gradients.
-
4.
Client-server collaboration: The collaborative training between a client and the corresponding server instance is indicated by label ①. Upon completing the training, resulting in a pair client-server model, , the server instance becomes available and waits in the pool for the next client to connect (label ③).
-
5.
Server model aggregation: When a server instance becomes available after training, a snapshot of the server model weights, , is recorded (label ②). After a certain period of time or a predetermined number of snapshots is recorded, the server aggregates (using the function) all snapshots to form a new version of the server model weights, . Then all server instances, , update their weights to the new aggregated weights, , for the next round of training.
It should be noted that the aggregation of the server models, , is performed asynchronously, and the degree of parallelization depends on the number of server instances. The parallelization of P-SL differs from the client-side parallelization in SFL, as it does not require the Fed server for local model aggregation.
IV-B P-SL with newly participating clients
IV-B1 A case study
In practice, setting up training when all clients simultaneously participate presents challenges due to the unstable nature of IoT/mobile devices. While previous studies, such as [11], examine offline clients’ participation during training, there is limited research on the scenario where a new client with its data wants to join the training to benefit from the knowledge acquired by existing clients. To address this real-world situation, we conduct experiments involving clients, where we initially allow clients () to collaboratively learn their models using P-SL, referred to as the first training phase. Subsequently, and join the training at a later stage, which we refer to as the second training phase. Both and possess their own data and aim learn their models while leveraging knowledge from the other clients’ data. Two possible solutions can be considered for the second training phase: 1. Training all clients, which would impose additional overhead on the existing clients; and 2. Training only the new arriving clients, reducing training complexity. While a hybrid approach, involving training new clients for a few epochs and then training all clients together, is also possible, we focus on the extreme cases (training all or training new) to study the impact of introducing new information to existing knowledge.
on the Fashion dataset.
Client | ||||||
---|---|---|---|---|---|---|
Training stage | Balanced data distribution | |||||
w. 4 clients | ||||||
w. ALL clients | ||||||
w. NEW clients | ||||||
Training stage | Imbalanced data distribution | |||||
w. 4 clients | ||||||
w. ALL clients | ||||||
w. NEW clients |
We conduct experiments on both the Fashion and CIFAR10 datasets to investigate the scenarios involving new clients joining the training process. The detailed settings are deferred to the experiment evaluation section. Table II presents the accuracy results of the first training phase (without and ), the second training phase with solution one (training all clients), and solution two (training new clients only). Note that all training are performed using P-SL. From the obtained results, we observe that training all clients helps new clients learn their deep models while slightly improving the accuracy of existing clients (e.g., and ) due to the reinforcement learning from the newcomers’ data. Conversely, training only the new clients leads to the server forgetting the knowledge learned from the existing clients, thereby reducing the learning performance of the new joining clients ( and ). Additionally, the accuracy of the existing clients also decreases due to the updating of the server model during training with the new clients. Similar effect can be observed in Fig. 4, which visualizes the results on CIFAR10. After the second training phase with the new clients only, the accuracy of the existing clients dropped by with both balanced and imbalanced data, indicating the phenomenon of forgetting in deep learning. Therefore, training all clients when newcomers join is a suitable approach to maintain the benefits of collaborative learning. However, retraining the existing clients incurs network and computation overhead, which is a limitation for low-end devices.


In summary, our experiments on P-SL involving clients joining after the initial learning phase demonstrate that retraining the entire network is beneficial for newcomers and enhances the performance of existing clients. However, this approach also incurs additional costs for the existing clients, which can be a disadvantage, particularly for low-end devices. In this context, it is preferable that existing clients do not need to retrain, which in turn reduces accuracy due to the forgetting phenomenon.
IV-B2 Cached-based P-SL algorithm
To address the issue of forgetting, we propose an enhanced method for training only newcomers to reduce the additional cost of retraining while preserving the knowledge acquired by existing clients. Our approach involves caching the smashed data sent from clients to the server during training, which enhances the learning process of the server model. By caching the data, the server can review knowledge while incorporating new information, mitigating the catastrophic forgetting phenomenon that can occur when a model is serially trained among clients. To incorporate caching into the server part of P-SL, we modify the execution as depicted in the box from line to line in Alg. 1. This modification enables the server to cache smashed data from all clients. Subsequently, this cached data can be combined with incoming smashed data during the training of the next client, allowing the server to ‘review’ previous knowledge. The specific details of this modification are presented in Alg. 2.
For each iteration of client ’s training, upon receiving the smashed data, , and the corresponding labels, , the server stores them in a cache pool (line ). Before performing forward propagation, the server randomly selects cached data from the cache pool and concatenates it with the incoming smashed data and labels from client to form as shown in lines . Subsequently, the server proceeds with the forward and backward passes using the concatenated data as usual (lines ). Let denote the loss function used to measure the discrepancy between the ground-truth labels and the model’s predicted outputs. The gradients at the server’s last layer are computed as follows:
It is important to note that the computed gradients for the split layer have the size of the concatenated data instead of the size of . Therefore, the server needs to slice the gradients to fit the size of before sending them to the client (line ). The execution in clients remains the same as in P-SL, as shown in Alg. 1. From the above equation, the gradients are computed not only based on the errors from training with ’s data but also from the other clients’ data (cached smashed data and labels). Consequently, by updating using these gradients, the server can simultaneously learn new knowledge from ’s data and review knowledge previously acquired from other clients’ data.
IV-B3 Computation and privacy analysis
In cache-based P-SL, we have made modifications only to the server’s procedure, ensuring that the cost at the client side remains the same as in P-SL. The additional costs incurred at the server, such as storing cached data and processing concatenation, are considered acceptable because the server is assumed to possess sufficient computing resources to serve multiple clients. Moreover, we have the flexibility to control the size of the cached data, allowing us to adjust the server’s performance accordingly. As a result, cache-based P-SL does not increase the cost at the client side, thereby preserving the benefits when applied to IoT/mobile environments.
In terms of data privacy, P-SL already safeguards against sharing local weights among clients, thereby reducing the risk of model inversion attacks by a malicious client. Additionally, the caching approach in cache-based P-SL does not violate any privacy concerns, as the cached data is public by default in SL, and clients willingly to share it with the server to derive the utility of the learning process. In summary, cache-based P-SL does not increase the cost at the client side or compromise the privacy of local private data. However, there is an additional overhead in terms of computing and storage resources at the server, which is more feasible to handle compared to low-end devices at the client side. To comprehensively evaluate the learning performance of the proposed scheme, we conduct experiments and present the results in the following section.
V Experiment evaluation
To conduct experiments to evaluate the performance of the proposed P-SL, we consider classification tasks on small-scale image datasets using deep models based on 2D-CNN. Table III summarizes of the selected datasets along with the corresponding Very Deep Convolutional Networks (VGG) [33]-based deep models. The deep models are divided into two parts: the first two convolutions are deployed at the clients, while the rest of the model resides on the server.
Dataset | Input size | Samples | Deep model architecture | |
---|---|---|---|---|
Client side | Server side | |||
Fashion | 2conv | 4conv+1dense | ||
CIFAR10 | 2conv | 8conv+1dense |
In these evaluations, we select two datasets: Fashion [34] and CIFAR10 [35], both of which consists of 10 classes and have separate training and testing sets. We distribute a total of samples from the train set among clients and evaluate the learning performance of each client using the same testing set. To simulate an imbalanced data distribution, we assign a varying number of samples to each client following a half bell curve of the standard normal distribution. Table IV provides the details of the total number of data samples allocated to each client for the case of clients.
Client index | ||||||
---|---|---|---|---|---|---|
Splitting ratio | ||||||
No. of samples |
V-A P-SL training accuracy
Using the selected datasets and local data distributions described in the previous section, we implement P-SL with clients and a central server. After training, we measure the learning performance of each client when performing inference on a test set collaboratively with the server (). We compare the results with multiple SL, referred to as SL, where we set up different SL processes between client-server pairs (). The key difference between P-SL and SL is the collaboration at the server side. In SL, there are distinct server instances, each associated with a different client, and they do not collaboratively aggregate server models. Conversely, in P-SL, we utilize only one server instance, allowing the server model to learn from the smashed data of the entire dataset. We also include the results of SL and SFL for benchmarking reference in Table V.
Dataset | Balanced data | Imbalanced data | ||
---|---|---|---|---|
SL | SFL | SL | SFL | |
Fashion | ||||
CIFAR10 |
Client | ||||||
---|---|---|---|---|---|---|
Scheme | With balanced data | |||||
SL | ||||||
P-SL | ||||||
Scheme | With imbalanced data | |||||
SL | ||||||
P-SL |
The training accuracy of each client with the Fashion dataset is presented in Table VI, while the results with the CIFAR10 dataset are visualized in Fig. 5. With SL, the accuracy of each client depends on the number of data samples held by that client, as expected. Therefore, under balanced data, these clients have similar accuracy (around with Fashion), while their accuracy ranges from lower ( with ) to higher ( with ) values with imbalanced data (see Table VI). We can visually observe similar results in Fig. 5, which shows the results using CIFAR10, a more complex and difficult dataset that leads to a higher accuracy difference between clients with fewer and more data samples.


In the proposed P-SL, despite the separate training of local models, the learning performance is better than SL attributing to the shared server model, which aggregates knowledge from all clients. P-SL achieves a higher accuracy with the Fashion dataset and a higher accuracy with the CIFAR10 dataset compared to SL under a balanced data distribution. Under an imbalanced distribution, the results are even more impressive as we observe significant accuracy improvements for clients with fewer data, such as , , etc. (see Fig. 5b). This demonstrates the benefit of P-SL in collaborative learning, even without weight sharing among clients. We also compare our results with SL and SFL, which achieve state-of-the-art collaborative learning performance. By sharing local models among clients, knowledge is aggregated at both the client and server sides, resulting in higher accuracy for SL and SFL compared to P-SL, which only aggregates knowledge at the server. In summary, our experiments demonstrate that P-SL, without local weight sharing, still benefits collaborative learning between multiple clients and a central server. Under imbalanced data distribution, clients with less data can learn more by training with clients with more data.
In our experiments, we follow a fixed training order for the clients, starting with , followed by , and so on up to . This fixed training order may have an impact on the accuracy performance since the learning process can be influenced by the presence of more or less data samples during training, especially under an imbalanced data distribution. However, a detailed investigation of the training order and its effects is deferred to the next section for further analysis and discussion.
V-B Privacy preservation at client side with P-SL
We conduct experiments to evaluate the privacy preservation of P-SL. During the training process, we employ model inversion attacks to reconstruct the raw data of all clients using the smashed data that clients send to the server. Specifically, these experiments are conducted with six clients under a balanced data distribution. Whereas all clients are engaged in training, we train a decoder using the local weights and data of one client, which in our case is client (representing a malicious client who is curious about the data of other clients). The trained decoder is used to reconstruct raw data from the smashed data that any client sent to the server. Based on this experiment setup, we evaluate the privacy preservation by measuring the amount of data leakage from clients’ raw data. Data leakage is quantified using the SSIM [36], which is a perceptual metric that assesses image quality degradation. SSIM provides a measure of similarity between the raw and reconstructed images and has been commonly used in previous works to evaluate data leakage, such as in [13, 31, 37]. Unlike other metrics such as MSE or PSNR, SSIM offers a more intuitive and interpretable metric, where values range between and . A value of indicates the least similarity, while a value of represents the highest similarity, indicating the most leakage.

Fig. 6 demonstrates the reconstructed images from other clients assuming that is malicious. The decoder is trained using ’s local model and its raw data, resulting in clear reconstructions from ’s smashed data. However, the quality of reconstruction significantly drops when applying the decoder to the smashed data of other clients. The reconstructed images from and in Fig. 6 are vague and contain a high level of noise compared to the raw images. The numerical results presented in Table VII reveal that the reconstruction quality (SSIM value) of all clients (except ) in P-SL is only around . On the other hand, with SL and SFL, as partially visualized in Fig. 1, can almost entirely reconstruct the raw data of all other clients, with SSIM values exceeding . This is similar to self-reconstructing its own data.
and SFL when is the attacker.
Scheme | ||||||
---|---|---|---|---|---|---|
SL | ||||||
SFL | ||||||
P-SL | ||||||
P-SL* | 4-4 | 1-2 | 1-2 | 2-2 | 1-2 | 2-2 |
*: Data leakage is measured using the MSE metric. |
The last row of Table VII presents the leakage of P-SL measured by MSE between the raw and reconstructed data. The results show that the errors in reconstructing data for clients to are more than higher than the reconstruction error for . While a value of in the MSE metric indicates full leakage (no reconstruction error), there is no upper bound for non-leak or partial leak scenarios. Additionally, the correlation of leakage among clients does not exhibit a clear and meaningful trend when using MSE, unlike when using SSIM. Therefore, we primarily present the leakage measure using the SSIM metric. In [31], the authors made a similar observation and suggested using SSIM to measure leakage instead of MSE and its related PSNR.
with different attackers.
Attacker | ||||||
---|---|---|---|---|---|---|
We also conduct experiments with imbalanced data and obtain similar results, as presented in Table VIII. Based on the experiment results, we can conclude that P-SL outperforms SL and SFL in preserving data privacy at the client side. However, it is important to note that the SSIM values between reconstructed and raw images in P-SL are still significantly different from , indicating the presence of leakage. This leakage can be attributed to the query-free attack described in [13], where the attacker does not require knowledge of the target model or the ability to query it. The only assumption for this type of attack is that the attacker and the victim share the same data distribution. In our experiments, the data is uniformly distributed among all clients, regardless of the number of samples they have. Therefore, the local model of the attacker acts as the shadow model for the query-free attack, leading to partial data leakage. Furthermore, an interesting observation from Table VIII is that attackers with more data (e.g. ) can reconstruct higher quality images compared to the ones with fewer data (e.g., ).
Experiments on 1D time-series data. Most of the work on SL in the literature typically assumes deep neural network models such as 2D CNN; however, sequential/time-series data are also common in the IoT environment. Therefore, following [18], we conduct our experiments on a 1D time-series dataset to further evaluate the efficiency of our proposed scheme. We set up a two-layer 1D CNN model to classify five heart diseases on the ECG dataset provided by the authors. All other configurations exactly follow [18], except that we split and deploy the first convolutional layer to six clients. After running for epochs, we successfully reproduce the model training accuracy result, which is when using SL. Regarding P-SL, we achieve the comparable model accuracy, which is . To evaluate the privacy preservation of P-SL, we reconstruct the input data from smashed data and compare the similarity to the raw input using distance correlation (DC) and dynamic time warping (DTW) in alignment with [18].

Fig. 7 demonstrates the reconstructed results from other clients assuming that is malicious, regarding experiments on the ECG dataset. Similar to Fig. 6, the decoder trained on ’s local model and data can effectively reconstruct raw input from ’s smashed data, with the similarity quantified as , , and in terms of DTW, DC, and MSE metrics, respectively. However, the reconstruction quality significantly drops when applied to the smashed data of other clients (i.e. and ). The mean similarity is , , and , as measured by DTW, DC, and MSE, respectively. Note that, lower DTW, higher DC, and lower MSE values indicate higher similarity between raw data and the reconstructed results. The experiment results with 1D time-series data further demonstrate the effectiveness of our P-SL in mitigating privacy attacks at the client side of SL.
Comparison to related works. Noise-based protection techniques, such as differential privacy (DP) [38], are commonly used to ensure guaranteed privacy for user’s private data. Recently, various approaches [18, 39, 31, 20] have been proposed to apply local DP to protect user data privacy in SL. This makes DP a competitive approach when considering the aforementioned defined threat model. To compare DP to P-SL, we utilize the Laplace mechanism, as described in [31]. However, it is important to note that there exists a trade-off between accuracy and privacy, as increasing the amount of added noise for higher privacy leads to a reduction in model accuracy. Furthermore, we also experiment with DISCO [25] and ResSFL [24], which are two state-of-the-art approaches aimed at tackling model inversion (MI) attacks by training the client model to be MI-resistant feature extractors. Both of these methods witness accuracy-privacy trade-offs by managing the contribution of resistance factor during training. We quantitatively define privacy as the dissimilarity, represented by SSIM, between raw input data and the reconstructed results.

The experiments are conducted on the Fashion dataset with varying noise levels, pruning ratios, and regularization strength of DP, DISCO, and ResSFL, respectively. The results are shown in Fig. 8, where we can observe the trade-offs of DP and DISCO (lines) in sacrificing accuracy for improved privacy. On the other hand, our P-SL does not provide a method for controlling the trade-off; however, its result (dot) stays above the lines, indicating that P-SL preserves more privacy (larger dissimilarity in reconstruction) with less accuracy sacrificed. According to the experiments, P-SL achieves privacy similar to applying DP with , while maintaining an accuracy comparable to utilizing DP with . Regarding ResSFL, a simulated attacker is required in training the local model to resist model inversion attacks. However, simulating a weak attacker does not help defend against a strong real attacker, as we can achieve high-quality reconstructed results with a more complex decoder than the one used in training. Furthermore, requiring additional model training (adversarial networks in DISCO and ResSFL) would lead to extra costs for resource-constrained clients in SL. Comparing to other related works, even though lacking flexibility in managing accuracy sacrificed, our P-SL achieves better accuracy-privacy trade-offs with less processing overhead at the client side. Following, we continue to present further evaluation of P-SL such as in dynamic environments.
V-C P-SL in dynamic environments
V-C1 With multiple server instances
Table IX presents the experiment results of parallelizing P-SL with clients and server instances. Each client is randomly associated with a server instance to perform the training. With two servers available, two groups of clients can process training in parallel, theoretically speeding up the training by a factor of two. Based on the reported results, it is evident that parallelized P-SL achieves similar results to sequential P-SL, where we sequentially train each client with a single server. Therefore, parallelized P-SL can be considered ‘scalable’, as it speeds up the training without compromising the model’s accuracy.
with two server instances.
Client | ||||||
---|---|---|---|---|---|---|
Data dist. | Fashion dataset | |||||
Balance | ||||||
Imbalance | ||||||
Data dist. | CIFAR10 dataset | |||||
Balance | ||||||
Imbalance |
We further evaluate the scalability of P-SL by setting up experiments with clients connecting to a varying number of server instances. The experiments are conducted on the Fashion dataset, which is distributed evenly among the clients. The results are presented in Table X, where we report the mean training accuracy of the clients corresponding to the number of server instances set up. Note that the number of server instances corresponds to the speedup in training time. The results exhibit a stable trend in model training accuracy regardless of the number of server instances. Therefore, depending on the resources available at the server side, we can dynamically speed up the training time of P-SL.
Num. of server instances | 2 | 4 | 5 | 10 | 20 |
---|---|---|---|---|---|
Mean accuracy of clients |
V-C2 With late joining clients
By leveraging cached data during the training of new clients, cache-based P-SL facilitates the review of previous knowledge, resulting in more stable performance and higher accuracy for both new and existing clients. Fig. 9 illustrates the learning performance of newcomers ( and ) and existing clients (, , , and ) using P-SL (left column) and cached-based P-SL (right column) on the Fashion and CIFAR10 datasets. Using caching, learning with only newcomers in P-SL becomes more stable, and the accuracy achieved is comparable to that of retraining the entire network. Therefore, we can train newcomers exclusively using cache-based P-SL, thereby saving the additional costs associated with full retraining while experiencing only a slight reduction in the accuracy of existing clients.
![]() ![]() |
![]() ![]() |
![]() ![]() |
![]() ![]() |
V-C3 Order of clients in training
In the previous section, we discussed the impact of the order in which clients participate in the training on the final accuracy, particularly in scenarios with an imbalanced data distribution where some clients have more data than others. We conduct experiments with P-SL, where for each epoch, we randomly select the order of clients to participate in the training with the server. We compare the learning performance to the fixed order, which starts from and ends at each epoch, to assess the effect of client order. The experiment results reveal no significant difference between training with a fixed or random order under a balanced data distribution. Due to the similar quantity and distribution of data, the learning performance of the server with is also similar to that of any other client. However, under an imbalanced data distribution, the learning performance of the server with clients with more data differs from those with fewer data. We plot the learning performance of P-SL in Fig. 10a, where it can be observed that the achieved accuracy is not stable. However, when using cache-based P-SL (shown in Fig. 10b), the caching approach demonstrates its effectiveness in stabilizing the learning curve.


V-C4 Training with Non-IID data
As the training data collected by the individual clients based on their local environment and usage patterns, it is practical to assume that it is non-IID distributed, e.g., data collected from a single person can only be obtained [40]. We continue to evaluate the learning performance of our proposed method in a non-IID setting, which simulates the worst-case statistical heterogeneity of local data. Following the non-IID setting described in [11], we distribute data from only half of the classes ( out of ) to each client. The accuracy results obtained after the full training process of SL, SFL, and P-SL with and without caching are reported in Table XI. The results demonstrate that all the evaluated schemes are sensitive to non-IID data, as the learning objectives of each client diverge when the training data is heterogeneous [11]. These findings align with those in [10, 11], where SL performs worse than SFL in non-IID settings. Our proposed P-SL is particularly sensitive to non-IID data due to the absence of client-side synchronization. Unfortunately, although caching helps stabilize the training loss, it only slightly improves learning accuracy, limiting the applicability of P-SL in non-IID settings. However, caching supports the learning of SL and SFL in non-IID data scenarios, enabling them to achieve comparable results to those obtained in IID data settings (both balanced and imbalanced). This advancement promotes the development of SL in non-IID data environments, an area that has received limited study thus far.
Fashion dataset | CIFAR10 dataset | |||||
Scheme | P-SL | SL | SFL | P-SL | SL | SFL |
Without caching | ||||||
With caching |
Based on the above experiments, we can conclude that caching plays a crucial role in stabilizing and maintaining the learning performance of P-SL. As for parallelization to speed up training, this caching approach can be extended to the server with multiple instances that share a cache pool. Furthermore, the strategy for caching, such as determining which and how much data to cache, will be a topic for future research.
VI Conclusion
This paper addresses the issue of data leakage in traditional SL and its variants that arise from the sharing of local weights during client training. We propose and analyze a variant called SL without local weight sharing, P-SL, to enhance the privacy preservation of user data. The experiment results across various data distributions demonstrate that P-SL enables collaborative learning from distributed clients while reducing data leakage at the client side by half and maintaining comparable accuracy to SL and SFL. Furthermore, P-SL can be parallelized to expedite client training without sacrificing model accuracy. We also investigate P-SL in a dynamic environment where new clients join the training, which can impact the existing clients due to the forgetting phenomenon. To address this, we propose a server-caching mechanism for P-SL, which facilitates the review of learned knowledge during training with newcomers. The experimental results show that cache-based P-SL stabilizes the learning performance and allows for training only the late-arriving clients, reducing client-side overhead while mitigating the server-side forgetting issue. In conclusion, this paper presents P-SL as an effective approach for preserving user data privacy in collaboratively distributed learning, particularly for IoT/mobile devices in real-world dynamic environments. Future research directions include integrating with other privacy-preserving techniques, evaluating on other domains/tasks, and exploring caching strategies to enhance the proposal further.
References
- [1] X. Liu, L. Xie, Y. Wang, J. Zou, J. Xiong, Z. Ying, and A. V. Vasilakos, “Privacy and Security Issues in Deep Learning: A Survey,” IEEE Access, vol. 9, pp. 4566–4593, 2021.
- [2] O. Gupta and R. Raskar, “Distributed learning of deep neural network over multiple agents,” Journal of Network and Computer Applications, vol. 116, pp. 1–8, 2018.
- [3] P. Vepakomma, O. Gupta, T. Swedish, and R. Raskar, “Split learning for health: Distributed deep learning without sharing raw patient data,” 2018. [Online]. Available: https://arxiv.org/abs/1812.00564
- [4] M. G. Poirot, P. Vepakomma, K. Chang, J. Kalpathy-Cramer, R. Gupta, and R. Raskar, “Split Learning for collaborative deep learning in healthcare,” 2019.
- [5] P. Vepakomma and R. Raskar, Split Learning: A Resource Efficient Model and Data Parallel Approach for Distributed Deep Learning. Cham: Springer International Publishing, 2022, pp. 439–451.
- [6] Q. Yang, Y. Liu, T. Chen, and Y. Tong, “Federated Machine Learning: Concept and Applications,” ACM Trans. Intell. Syst. Technol., vol. 10, no. 2, 2019.
- [7] V. Turina, Z. Zhang, F. Esposito, and I. Matta, Combining Split and Federated Architectures for Efficiency and Privacy in Deep Learning, 2020, pp. 562–563.
- [8] J. Jeon and J. Kim, “Privacy-Sensitive Parallel Split Learning,” in 2020 International Conference on Information Networking, 2020, pp. 7–9.
- [9] C. Thapa, P. C. Mahawaga Arachchige, S. Camtepe, and L. Sun, “SplitFed: When Federated Learning Meets Split Learning,” Proceedings of the AAAI Conference on Artificial Intelligence, vol. 36, no. 8, pp. 8485–8493, 2022.
- [10] Y. Gao, M. Kim, S. Abuadbba, Y. Kim, C. Thapa, K. Kim, S. A. Camtep, H. Kim, and S. Nepal, “End-to-End Evaluation of Federated Learning and Split Learning for Internet of Things,” in 2020 International Symposium on Reliable Distributed Systems, 2020, pp. 91–100.
- [11] Y. Gao, M. Kim, C. Thapa, A. Abuadbba, Z. Zhang, S. Camtepe, H. Kim, and S. Nepal, “Evaluation and Optimization of Distributed Machine Learning Techniques for Internet of Things,” IEEE Transactions on Computers, vol. 71, no. 10, pp. 2538–2552, 2022.
- [12] A. J. Paverd and A. C. Martin, “Modelling and automatically analysing privacy properties for honest-but-curious adversaries,” Tech. Rep., 2014.
- [13] Z. He, T. Zhang, and R. B. Lee, “Model Inversion Attacks against Collaborative Inference,” in Proceedings of the 35th Annual Computer Security Applications Conference, 2019, pp. 148–162.
- [14] D. P. Kingma and M. Welling, An Introduction to Variational Autoencoders, 2019, vol. 12.
- [15] Z. He, T. Zhang, and R. B. Lee, “Attacking and Protecting Data Privacy in Edge-Cloud Collaborative Inference Systems,” IEEE Internet of Things Journal, vol. 8, no. 12, pp. 9706–9716, 2021.
- [16] R. M. French, “Catastrophic forgetting in connectionist networks,” Trends in Cognitive Sciences, vol. 3, no. 4, pp. 128–135, 1999.
- [17] Q. Duan, S. Hu, R. Deng, and Z. Lu, “Combined Federated and Split Learning in Edge Computing for Ubiquitous Intelligence in Internet of Things: State-of-the-Art and Future Directions,” Sensors, vol. 22, no. 16, 2022.
- [18] S. Abuadbba, K. Kim, M. Kim, C. Thapa, S. A. Camtepe, Y. Gao, H. Kim, and S. Nepal, “Can We Use Split Learning on 1D CNN Models for Privacy Preserving Training?” in Proceedings of the 15th ACM Asia Conference on Computer and Communications Security, 2020, pp. 305–318.
- [19] T. Titcombe, A. J. Hall, P. Papadopoulos, and D. Romanini, “Practical Defences Against Model Inversion Attacks for Split Neural Networks,” 2021.
- [20] N. D. Pham, A. Abuadbba, Y. Gao, K. T. Phan, and N. Chilamkurti, “Binarizing Split Learning for Data Privacy Enhancement and Computation Reduction,” Trans. Info. For. Sec., vol. 18, pp. 3088–3100, 2023.
- [21] J. Wang, J. Zhang, W. Bao, X. Zhu, B. Cao, and P. S. Yu, “Not Just Privacy: Improving Performance of Private Deep Learning in Mobile Cloud,” in Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 2018, pp. 2407–2416.
- [22] P. Vepakomma, A. Singh, O. Gupta, and R. Raskar, “NoPeek: Information leakage reduction to share activations in distributed deep learning,” in 2020 International Conference on Data Mining Workshops, 2020, pp. 933–942.
- [23] V. Turina, Z. Zhang, F. Esposito, and I. Matta, “Federated or Split? A Performance and Privacy Analysis of Hybrid Split and Federated Learning Architectures,” in 2021 IEEE 14th International Conference on Cloud Computing, 2021, pp. 250–260.
- [24] J. Li, A. S. Rakin, X. Chen, Z. He, D. Fan, and C. Chakrabarti, “ResSFL: A Resistance Transfer Framework for Defending Model Inversion Attack in Split Federated Learning,” in 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 10 184–10 192.
- [25] A. Singh, A. Chopra, E. Garza, E. Zhang, P. Vepakomma, V. Sharma, and R. Raskar, “DISCO: Dynamic and Invariant Sensitive Channel Obfuscation for deep neural networks,” in 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 12 120–12 130.
- [26] D. Xiao, C. Yang, and W. Wu, “Mixing Activations and Labels in Distributed Training for Split Learning,” IEEE Transactions on Parallel and Distributed Systems, vol. 33, no. 11, pp. 3165–3177, 2022.
- [27] P. Joshi, C. Thapa, S. Camtepe, M. Hasanuzzamana, T. Scully, and H. Afli, “Splitfed learning without client-side synchronization: Analyzing client-side split network portion size to overall performance,” in Proceedings of the 7th Collaborative European Research Conference, vol. 2348, 2021, pp. 1–7.
- [28] P. Joshi, C. Thapa, S. Camtepe, M. Hasanuzzaman, T. Scully, and H. Afli, “Performance and Information Leakage in Splitfed Learning and Multi-Head Split Learning in Healthcare Data and Beyond,” Methods and Protocols, vol. 5, no. 4, 2022.
- [29] A. J. Paverd and A. C. Martin, “Modelling and Automatically Analysing Privacy Properties for Honest-but-Curious Adversaries,” Tech. Rep., 2014.
- [30] H. Dong, C. Wu, Z. Wei, and Y. Guo, “Dropping Activation Outputs With Localized First-Layer Deep Network for Enhancing User Privacy and Data Security,” IEEE Transactions on Information Forensics and Security, vol. 13, no. 3, pp. 662–670, 2018.
- [31] J. Ryu, Y. Zheng, Y. Gao, A. Abuadbba, J. Kim, D. Won, S. Nepal, H. Kim, and C. Wang, “Can differential privacy practically protect collaborative deep learning inference for IoT?” Wireless Networks, 2022.
- [32] Y. Li and X. Lyu, “Convergence Analysis of Sequencial Split Learning on Heterogeneous Data,” 2023.
- [33] K. Simonyan and A. Zisserman, “Very Deep Convolutional Networks for Large-Scale Image Recognition,” 2014.
- [34] H. Xiao, K. Rasul, and R. Vollgraf, “Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms,” 2017.
- [35] A. Krizhevsky, “Learning multiple layers of features from tiny images,” Tech. Rep., 2009.
- [36] Z. Wang, A. Bovik, H. Sheikh, and E. Simoncelli, “Image quality assessment: from error visibility to structural similarity,” IEEE Transactions on Image Processing, vol. 13, no. 4, pp. 600–612, 2004.
- [37] H. R. Roth, A. Hatamizadeh, Z. Xu, C. Zhao, W. Li, A. Myronenko, and D. Xu, Split-U-Net: Preventing Data Leakage in Split Learning for Collaborative Multi-modal Brain Tumor Segmentation. Springer Nature Switzerland, 2022, pp. 47–57.
- [38] C. Dwork, Differential Privacy. Boston, MA: Springer US, 2011, pp. 338–340.
- [39] P. C. Mahawaga Arachchige, P. Bertok, I. Khalil, D. Liu, S. Camtepe, and M. Atiquzzaman, “Local Differential Privacy for Deep Learning,” IEEE Internet of Things Journal, vol. 7, no. 7, pp. 5827–5842, 2020.
- [40] Y. Zhao, M. Li, L. Lai, N. Suda, D. Civin, and V. Chandra, “Federated Learning with Non-IID Data,” 2018.