Mobility-Aware Routing and Caching in Small Cell Networks using Federated Learning
Abstract
We consider a service cost minimization problem for resource-constrained small-cell networks with caching, where the challenge mainly stems from (i) the insufficient backhaul capacity and limited network bandwidth and (ii) the limited storing capacity of small-cell base stations (SBSs). Besides, the optimization problem is NP-hard since both the users’ mobility patterns and content preferences are unknown. In this paper, we develop a novel mobility-aware joint routing and caching strategy to address the challenges. The designed framework divides the entire geographical area into small sections containing one SBS and several MUs. Based on the concept of one-stop-shop (OSS), we propose a federated routing and popularity learning (FRPL) approach in which the SBSs cooperatively learn the routing and preference of their respective MUs and make a caching decision. The FRPL method completes multiple tasks in one shot, thus reducing the average processing time per global aggregation of learning. By exploiting the outcomes of FRPL together with the estimated service edge of SBSs, the proposed cache placement solution greedily approximates the minimizer of the challenging service cost optimization problem. Theoretical and numerical analyses show the effectiveness of our proposed approaches.
Index Terms: Caching, federated learning, routing, mobility patterns, small cell network, one-stop-shop, multiple tasks.
I Introduction
Wireless caching is a promising concept to reduce the peak traffic and the backhaul load, particularly for video content delivery [2, 3, 4]. The caching concept originates from the content reuse property of video streaming, i.e., users are likely to request the same video content. Thus saving popular content at the local small-cell base stations (SBSs) during the off-peak time or pushing them at user devices directly through broadcasting improves the network’s throughput and coverage performance. Besides, it enhances the users’ quality of experience (QoE) [5, 6].
Realizing the potential of wireless caching necessitates optimizing several parameters. Under lack of sufficient prior knowledge, one can take advantage of a recently-emerged concept, namely federated learning (FL). FL enables several partners to jointly learn the parameters of a specific model in a distributed manner, i.e., without requiring any data exchange [7]. Thus, using FL strongly reduces the amount of data uploaded via the wireless uplink channel. Besides, FL maintains the benefits of cognitive- and swift reaction to the mobility and other characteristics of cellular networks, as well as preserving personal data privacy [8]. Besides, the FL concept is a widespread tool for wireless edge network optimization [8, 9, 10, 11, 12], due to the following three reasons: (i) The ever-increasing number of devices in the internet of things (IoT) generate an enormous amount of data at the network edge. That entails an enabling technology and efficient approaches for sensitive data management and utilization in wireless edge network; (ii) With the computation- and storage constraints of increasingly complex edge networks, conventional network optimization approaches built on static models perform poorly in modeling dynamic networks; (iii) Conventional machine learning-based approaches can optimize decision-making through interactions with the dynamic environment. However, such methods require user data as input, which might be sensitive or inaccessible in nature due to regulatory constraints. Thus FL is an enabling technology for resource-constrained network optimization (concerning computation and storage resources), and for bringing intelligence to the wireless edge network.
Reference [13] formulates a joint link scheduling and power allocation problem for (device-to-device) D2D-assisted wireless caching network. The authors then decompose the original problem into two subproblems and alternately solve subproblems by using the convex optimization method [14]. In [15], Want et al. investigate the NP-hard cache placement problem for D2D-assisted network. The authors propose a dynamic programming algorithm that greedily assigns the contents to the empty caches. In [16], the authors investigate an energy-efficient transmission mode selection problem for D2D-assisted small-cell networks. By relaxing the hard problem and using the bandit theory [17], they achieve the maximum network utility. Furthermore, reference [18] studies a joint transmission and caching policy optimization problem for D2D-assisted small-cell networks, given that the users’ demands are known a prior. Wireless caching in small-cell networks is also studied in [19, 20, 21, 22]. Reference [19] develops a context-aware proactive caching approach for small-cell networks. Such an approach and its variant in [20] learn the context-specific content popularity online. In [21], the authors introduce a learning-based approach to caching in heterogeneous small-cell networks. Besides, [22] investigates a joint femto-caching and power control problem in small-cell networks without knowing any prior information on file popularity.
References [23, 24, 25] study caching based on network coding (coded-multicast). [23] investigates the caching policy optimization based on multicast transmissions. The results reveal that properly combining caching and multicast transmission reduces energy costs. By jointly optimizing the cache placement and multicast transmission, [24] proposes a novel coded caching approach that can guarantee a global caching gain. In [25], the authors develop an optimal femto-cache placement strategy that maximizes the number of network coded file downloads from femto-caches, thereby significantly reducing the macro-cell base station (MBS) bandwidth usage.
The State-Of-The Art Research of Caching in Small-Cell Networks
Ref. | Optimization Problem | Prior Information | Resource Constrained | Merit |
---|---|---|---|---|
[15] | Transmission, Caching policy | Popularity | None | Low energy and economical cost |
[16] | Cache placement | User context | Yes | High cache hit ratio |
[18] | Cache placement | User demand | None | A lower bound on the training time |
[22] | Transmission, Caching policy | Popularity | None | Low energy cost |
[25] | Cache placement | Popularity | None | Low computation complexity |
[26] | Energy efficiency, Cache placement | Popularity, Perfect CSI | None | High cache hit ratio |
[27] | Caching utility | Popularity, Mobility data | None | High caching utility |
Our work | Routing, Cache placement | None | Yes | See Section I-A |
Mobility-aware caching exploits users’ mobility statistics for allocating caching resources [26, 27, 28, 29]. Reference [26], e.g., proposes a mobility-aware cache placement strategy that maximizes the data offloading ratio in D2D-assisted networks, given precise prior information of file popularity. Similarly, [27] assumes that each SBS has the global channel state information (CSI) and file popularity. Based on this assumption, the authors develop a green mobility-aware caching model to jointly optimize the cache placement and power allocation among SBSs and mobile devices. Besides, [28] explores the mobility-aware content caching problem for small-cell networks to maximize the caching gain given mobile users’ (MUs’) preferences and MUs’ mobility patterns. Using prior knowledge about the MUs’ locations at a central processor, [29] proposes a content cooperative caching approach that maximizes the sum mean opinion score (MOS) of MUs. Table I summarizes the state-of-the-art research in caching for resource-constrained small-cell network.
I-A Summary of Contributions
In the majority of references mentioned above, the popularity profile of content files is either known a prior or modeled as a Zipf-like distribution [30][31]. However, such an assumption is unrealistic in practice, particularly for the dense small-cell networks: (i) The file popularity is often non-stationary due to the dynamic nature of contents; (ii) Perfect knowledge of file popularity is technically difficult to guarantee, especially for intensive instantaneous demands; (iii) File popularity is an outcome of the decisions of a dense population, and very costly to predict [4]. In such a situation, developing a low-cost and effective learning-based approach for estimating the files’ popularity becomes imperative. Besides, to our best knowledge, no research has considered joint routing and cache placement optimization for resource-constrained small-cell networks so far.
Against this background, we focus on developing mobility-aware routing and caching strategies for dynamic small-cell networks under uncertainty. To jointly optimize the routing and cache placement, we first formulate the network cost minimization problem, which is an NP-hard integer programming (NIP) problem. We propose a federated routing and popularity learning (FRPL) approach by which the SBSs learn the popularity of files as a function of the non-stationary pedestrian- and request densities. To avoid unnecessary data retransmission via the backhaul, we develop a novel content transmission protocol that improves the cache-efficiency (CE) performance of SBSs. Besides, we optimize the cache placement using an algorithm that greedily approximates the minimizer of the NIP problem. The contributions of this paper are as follows:
-
•
Motivated by the notion of the one-stop-shop (OSS), we propose an FRPL approach that enables SBSs to complete multiple tasks simultaneously, thus reducing the processing time. That is crucial as the time slot within one global aggregation is limited.
-
•
The proposed FRPL method jointly learns the routing and content popularity considering the influence of both the users’ mobility and the SBSs’ geographical location. Therefore, it is robust against even in a highly dynamic and non-stationary network.
-
•
We design a cache placement approach that exploits the outcome of FRPL method combined with a novel transmission protocol to minimize the network’s content delivery cost by optimizing the cached content and delivery strategy.
-
•
Numerical results show that our cache placement policy guarantees a higher cache hit rate for SBSs compared to the existing schemes, although the MUs’ demand density is a priori unknown. Besides, the proposed greedy cache placement approach reduces the network cost compared to the state-of-the-art research. We also show the effectiveness of the proposed FRPL approach.
In Section II, we present the network model, FL model, and content request model. We then formulate the associated resource-constrained network optimization problem. In Section III, we propose an FRPL approach for multi-task learning. The analysis of the proposed model and solution appear in Section IV. Section V includes the numerical analysis. Section VI concludes the paper.
List of Most Important Variables
Symbols | Definitions/Explanations |
---|---|
The number of the samples collected by a participant . | |
The local FL model hyperparameter. | |
The update of the global FL model. | |
The probability that a request for file at time , and . | |
The binary caching strategy of file at SBS at time , and . | |
The pedestrian density at time at the site of SBS . | |
Expected request density of a specific file at time . | |
Cache capacity of SBS . | |
Cost incurred by caching file at each SBS. | |
; ; | Costs incurred by retrieving file from backhaul to MBS, from MBS to SBS , or directly from SBS . |
Normalized costs incurred by retrieving file from MBS to SBS . | |
; | Loss function for training local FL model hyperparameters () w.r.t. the pedestrian (request)-density prediction task. |
The final location of the centroid of cluster . | |
; | Statistical function of pedestrian density for those who might come (leave) cell from (to) neighboring cell . |
The gradient function of w.r.t. the global FL model . | |
The gradient function of w.r.t. the global FL model . | |
The serving edge consisting of a set of MUs of which SBS will serve at time . |
Notations: Bold lower case letters represent vectors, whereas bold upper case letters denote matrices. , , and denote transpose, -norm, and -norm operations, respectively. Besides, is used to denote the statistical expectation operation. represents the gradient of a function. Table II summarizes the frequently-used variables in the order of appearance in the paper.
II System Model and Problem Formulation
II-A Network Model
We consider a network consisting of SBSs and one MBS. The network serves MUs simultaneously. Due to the dense deployment, the coverage areas of the SBSs and the MBS overlap; Hence, at each time, an MU might be in the communication range of multiple entities that can contribute to learning and service delivery (contributors or participants). Before proceeding, we present a definition.
Definition 1 (Global Aggregation):
A global aggregation at period refers to the required time to perform the cascaded FRPL and cache placement approaches until convergence, i.e., completing the involved tasks.

Throughout the paper, we use two time scales: (i) . It refers to a learning round (period) in the FRPL method, where is the required time until the global loss function converges; and (ii) , which corresponds every instance of the global aggregation. Thus, each global aggregation consists of several learning rounds. The proposed cache placement approach takes several iterations; however, at one time instant (learning period), both the routing strategy and cache placement remain fixed.
FL is the core of the proposed FRPL methodology. The MBS and SBSs collaboratively learn a shared model while each entity retains its training data at its side. The FL model trained at the participant’s side is the local model. The MBS integrates the local models and generates the global model, which improves the local model of each participant. Fig. 1 shows the workflow and details follow.
Consider participants labeled as . We denote the input data of participant , by , where indicates the number of the samples. Thus is the entry vector to train the local FL model. The output vector is . The vector captures the model parameters of the local FL model. Under a linear model assumption, we have , where is measurement noise typically approximated as Gaussian i.i.d. samples. In the standard gradient descent (SGD) methods [32], [33], training and update procedures of the local FL model aims at finding optimal parameters that minimize the squared error cost function . We define a vector to capture the parameters related to the global FL model. The update of the global model therefore is given by . Section III provides more details of the FRPL framework.
II-B Content Request Model
We consider a content library with contents. The size of each file is . The MUs request file randomly and independently with probability so that is the request probability vector. Traditionally, upon requesting a file within the time deadline , an MU obtains it through random caching, local SBS caching, or MBS caching [2]. Nevertheless, such naive caching mechanisms barely guarantee a high cache efficiency for the following reasons: (i) They largely neglect the limited storing capacity and the finite bandwidth; (ii) They often retransmit the data unnecessarily. Therefore, we decompose the cache domain of an MU’s required contents into three categories:
-
•
Local Caching: The MU first checks the local SBS. If the required content is cached, then the MU obtains it directly from there within the time deadline . Otherwise, the MU receives it from one of the following sources:
-
•
Intra-Cell Caching: If the local SBS does not have the required content, it can fetch it from other SBSs in the intra-cell domain upon availability so that the MU is served within the deadline .
-
•
Inter-Cell Caching: If no SBS in the intra-cell domain has the required content, the local SBS fetches it via the backhaul link, from an external SBS deployed in the other overlapped cells or, in the worst case, from the MBS.
II-C Problem Statement
We focus on the following challenges: (i) When MUs migrate into one area, proper retrieving of several requested files over a backhaul while minimizing the cost becomes complicated; (ii) The MUs’ preferences affect the cache efficiency of SBSs. Moreover, distributed caching might result in duplicate files in a small area; (iii) Joint learning and cache placement, also service provisioning, can yield long delays.
Let be a binary caching strategy of the required contents in SBS, where indicates the th SBS stores the file , and indicates otherwise. Let denote the SBS’s aggregated cost as a result of caching, where is the cost of caching content . The cost mainly depends on the file size . The expected cost for retrieving content from SBS within a global aggregation period can be written as
(1) |
where represents the pedestrian density at time slot at the site of SBS . Besides, indicates the expected request density (i.e., the number of requests per time slot) of a specific file at time . Finally, corresponds to the worst case cost, i.e., when fetching file from the MBS backhaul. is a constant cost for retrieving via an MBS backhaul, denotes the cost of retrieving for SBS from MBS backhaul, and refers to the cost incurred by retrieving for MUs from the SBS . Besides, is the expected cost for retrieving from SBS , if available. Minimizing the aggregated cost in (1) over the SBSs, i.e., , yields the optimal routing strategy for each file. Such content routing strategy stipulates that retrieving file through the desired SBS leads to a least aggregated cost.
The expected cost in (1) is valid for the costs for every file request, regardless of the file being cached or not; i.e., it can serve the general transmission cost function for small-cell networks. Besides, it is an upper bound for the transmission cost incurred for retrieving file from SBS in a global aggregation period. Note that, in our mobility-aware scenario, a multicase-based caching policy is complex to implement. Indeed, most of the papers in that direction model the requests by a Poisson point process (PPP) across SBSs [23]; however, such assumption is not valid in our case, as the mobility pattern, and, thus, the user-SBS association, is dynamic and unknown.
The optimal mobility-aware routing strategy and cache placement is equivalent to minimizing the network cost per global aggregation. Formally,
(2a) | |||
(2b) | |||
(2c) |
where Constraint (2b) means that the total size of cached files cannot exceed the cache capacity of SBS .
Problem (2) is an integer programming problem that is NP-hard. Moreover, the objective function is not available since it involves unknown popularity and pedestrian density . In particular, there exists possible caching strategy matrices , implying an exponential growth in complexity as a function of and . Therefore, it is essential to develop an efficient approach to solve the problem (2) while maintaining a low delay.
III Federated Routing and Popularity Learning
Based on the OSS concept, we propose an FRPL approach to learn the pedestrian- and request density while ensuring a fast model aggregation. It consists of three major steps:
Geographical Location Division: We divide the entire area uniformly into small parts. Each area includes an SBS and a set of connected MUs at time slot , denoted by . Usually, MUs in have the same network cell ID at this moment.
Dual-Task Execution: The SBS of each sub-area aims at learning the pedestrian- and files’ request density by exploiting the location- and request information. Details follow.
-
•
TASK 1 (Pedestrian Density Prediction): To predict the pedestrian density of a cell , , the corresponding SBS derives the following statistics for the set of MUs at areas: (i) The number of pedestrian clusters in the transition region of neighboring cells that have as the predicted next cell [34]; (ii) The number of pedestrians already transited to ; (iii) The number of pedestrians that probably leave cell .
To search the neighboring areas to find the candidates of pedestrian clusters, the SBS uses the -means clustering algorithm [35] with the loss function given by , where is the centroid of all objects assigned to ’s class.
In words, the pedestrian clustering minimizes the sum of squared errors between data points and their respective centroids until reaching a stationary centroid.
To obtain the number of pedestrian clusters that are approaching the desired cell , the SBS uses the following detection criterion:(3) where is the location of the SBS , and is the final location of the centroid of cluster with . We use to denote the statistical function of density associated with the pedestrians that might transit to cell from neighboring cell . Besides, is the statistical function of density associated with the pedestrians might leave cell to the neighboring cell . Thus, the pedestrian density of cell from the neighboring cell yields . Moreover, refers to the density function associated with the pedestrians already moved to cell . Then the pedestrian density of the desired cell yields
(4) given that their respective centroids of clusters satisfy the detection criterion in (3) accordingly. Fig. 2 illustrates the procedure described above.
Figure 2: Prediction of in a dense network where moving pedestrians (purple dots) that might move to the desired cell (crosses) are clustered, whereas moving pedestrians (blue dots) that might leave the current cell are clustered at time slot . -
•
TASK 2 (Request Density Prediction): The SBS first searches the neighboring areas to find those MUs falling into , and to observe the demographic information about those associated users and their ratings. Naturally, the files with high ratings have more request density.
Using the files’ rating together with the users’ features, the SBS learns the request density of every file , , by minimizing the least squared error between the estimated request density and the actual one. By exploring the linear regression model to predict , the SBS uses the following loss function
(5) where is a regularization hyperparameter. The popularity of file is . Thus, the expected request density is given by
(6) Later in Section IV, the SBS uses this metric for cache placement.
Fast Model Aggregation: In standard SGD methods [32], [33], the unknown model is iteratively estimated by computing at each epoch, evaluating a gradient associated to the loss function
(7) |
with being the indices of prediction tasks.
Thus the training procedure of the FRPL finds the optimal parameters that minimize the sum of squared error cost function, i.e., in (7). For the linear regression FRPL in the request density prediction task, the model averaging constraint guarantees that SBSs and MBS share the same FL model after the convergence.
In the framework of FRPL, SBSs send the local models to the MBS.111For each global aggregation, all participants update their parameters to facilitate the highest running efficiency of FRPL [9]. At time slot , the global model, denoted by , is given by222We denote the local and global model parameters with time-scale index , to elaborate the procedure of aggregating the local and global FL models at each epoch.
(8) |
where indicates the number of samples collected by the local model w.r.t. the task . Besides, is the total number of samples collected by all local models w.r.t. the task .
Note that indicates the total number of training data points. The MBS then sends back to the SBSs.
After receiving from the MBS, the SBSs use the SGD method to update the local models . The update of local model follows as [32]
(9) |
w.r.t. the task , where denotes the learning rate. Moreover, corresponds to the gradient function of w.r.t. . For simplicity, we hereby define as the squared error cost function of the global model.
After receiving the updated local models, the MBS also updates the global model as [36]
(10) |
where is the gradient function of w.r.t. . Besides, is given by
(11) |
Algorithm 1 summarizes the described stages.
For each local FL model, the respective SBS searches the neighboring areas to find the candidates of pedestrian clusters. |
Calculate the gradient of the loss function . |
Monitor the statistic of clusters and estimate the number of clusters by following the criterion in (3). |
Update the local FL model and the centroids of clusters. |
Estimate the pedestrian density |
Search the neighboring areas and observe the statistical information in terms of users’ demographic information as well as their ratings. |
Pre-process data by merging the users’ demographic information and rating information. |
Learn the request density of each file , by means of minimizing the least squared error between the estimated and actual request density. |
Update the local FL model . |
Estimate the expected request density |
The distributed SBSs invoke the SGD method to update the local FL models , , to MBS to aggregate the local FL models. |
Update the gradient function of global FL model by (11). |
Finally, we note that the dimension of the raw pedestrians’ data (including geographic location, date, time, sojourn time) affects the outcome of -means clustering algorithm [35]. We therefore state the following proposition.
Proposition 1:
The expected request density has the following properties:
-
•
When the datasets utilized for user clustering in transition region of neighboring cells are insufficient (empty or very few data points), the associated can be lower-bounded as
(12) -
•
If the datasets’ dimension is larger than , the associated can be upper-bounded as
(13) where indicates the maximum number of non-empty clusters.
Proof:
See Appendix A. ∎
IV Cache Placement Policy
For each SBS , let denote the serving edge, i.e., the set of MUs that it severs at time at both the intra- and inter-cell domains, as shown in Fig. 3. Obviously, the is time-variant and depends on the mobility of users.
For each global aggregation, every SBS attempts to serve as many MUs as possible across its service edge to reduce the aggregated cost of retrieving contents across SBSs, i.e., in (2). Thus, for simplicity, we assume that at each global aggregation, the maximum number of MUs that SBS serves in its service edge equals to the maximum pedestrian density that FRPL method predicts.333By Proposition 1, at time , the pedestrian density of each cell can be upper bounded by , where corresponds to the maximum number of non-empty clusters; That is, the maximum number of MUs that SBS can serve is equivalent to the upper bound of the pedestrian density of this cell.

IV-A The Algorithm
After learning and using the FRPL method, each SBS knows also . Then the optimal cache placement problem is equivalent to minimizing the network cost function given by
(14) |
Thus the optimization problem follows as
(15a) | |||
(15b) | |||
(15c) |
Problem (15) is an NP-hard integer programming. Given cache placement policy , the content retrieval policy is determined based on our content transmission protocol introduced in Section II-C. To solve the problem with low complexity, we develop Algorithm 2, which greedily places the files to a cache and provides an approximate solution.
Let be the total size of files that are already cached at SBS during each iteration of our algorithm. Besides, denotes the set of pairs , where SBS does not store file yet despite having the sufficient cache capacity. At iteration , the algorithm conducts the cache placement by picking the pair with the lowest , provided that such cost is lower than the one in the previous iteration. Note that if the algorithm starts with the pair that results in the lowest cost , then the relative network cost caused by SBS at each iteration remains fixed. Formally,
(16) |
Then, if , the algorithm updates and ; Otherwise, when implies that the cache of SBS is full, the algorithm excludes all pairs from , and stops the cache placement for SBS . The algorithm terminates when all the caches are full. Algorithm 2 summarizes the described greedy cache placement.
Pick the pair with the lowest network cost according to (16). |
IV-B Discussion
Let be a binary variable that indicates whether SBS caches a file requested by MU . We observe that the approximation ratio of our proposed cache policy is proportional to the number of MUs that an SBS serves at each global aggregation [37] when neglecting the normalized cost for retrieving files from MBS.444Numerical analyses show that the normalized cost parameter for retrieving files from MBS backhaul has only a limited influence on the network cost minimization. To support this conclusion, we present the following results.
Proposition 2:
For each global aggregation period, the normalized cost caused by caching at SBSs is equal to the aggregated costs for caching request contents over the serving edges. Then we have
(17) |
where is the estimated service edge of SBS at time .
Proof:
See Appendix B. ∎
Theorem 1:
When the normalized cost for retrieving files from MBS backhaul is neglected, i.e., for = 0, our proposed cache placement policy in Algorithm 2 is a polynomial-time -approximation algorithm with being the maximum number of MUs that an SBS serves at its service edge at each time.
Proof:
See Appendix C. ∎
IV-C Complexity Analysis
Based on the optimization criterion given by (16), the computational complexity of cache placement using Algorithm 2 at SBS is polynomial time , with being the required number of iterations to solve problem (16). As is the set of all possible (SBS, file) pairs that impose the lowest to fill the cache of SBS , the total computational complexity of the cache placement policy yields . Hence the proposed policy has a linear complexity in the number of (SBS, file) pair.
Parameters | Value |
---|---|
Global aggregation periods (rounds) | |
The volume of content library | |
The total number of MUs | 6040 |
The learning rate | |
The hyperparameter for regularization, i.e., | [1, 10] |
The size of each content , i.e., | 1MB |
Cache size of SBS , i.e., | [50MB:500MB] |
Probability of randomly selecting contents, i.e., | [0.01, 0.1, 0.8] |
Power cost of caching content , i.e., | 1.5 mW |
Power cost of retrieving via MBS, i.e., | 13 mW |
Power cost of retrieving for SBS from MBS, i.e., | 370 mW |
Power cost of retrieving via SBS , i.e., | 180 mW |
V Simulation Results and Analysis
V-A Datasets and System Parameter Setups
We use a real-world dataset, namely, MovieLens 1M [38], to evaluate our proposed learning and caching strategies. Similar to [19], we assume that the movie rating process in the datasets is a streaming request. The MovieLens 1M datasets consist of (i) the users’ demographic information given in the format of <user ID*, age (in 7 categories), gender (in 2 categories), occupation (in 21 categories), Zip-code>; (ii) the rating information given in the format of <user ID*, movie ID, rating (in 5 categories), timestamp>.
To predict the request density, in the experiments, we merge the data of users’ demographic- and rating information based on the label of the user ID*. Against this join-type dataset, we pick the attributes of <movie ID, user ID*, their gender, age, Zip-code, and occupation> as the context dimensions. Table III gathers the most important parameters of our experiments.
V-B Performance Evaluation
We evaluate the average caching efficiency (the average ratio of cache hits within one global aggregation period compared to the total requests) of our proposals in comparison with the following cache placement approaches:


-
•
Optimal with Full Information: This scheme has full information about the users’ demands; thus, it has the potential of providing the best caching performance;
-
•
-Greedy: The policy selects a set of files uniformly at random with probability , while with probability , it selects the files with the highest estimated popularity so far;
-
•
Least-Recently-Used (LRU): It fetches the requested files from the potential contributor (i.e., MBS backhaul) and caches it when a cache miss event occurs. If an item exceeds the cache capacity of SBS, it removes the file in the cache that has been least recently used;
-
•
Random: This scheme selects a random set of files to cache in each time slot.
In Fig. 4, we compare the performance of different caching approaches using the same dataset, so that the file set remains constant. The figure shows that our proposed algorithm achieves a gain significantly higher than that of the random- and -greedy approaches under different . The gain grows as the cache size increases. The reason is that the proposed algorithm greedily places the cache to SBSs after learning the expected request densities of files by Algorithm 1.
Besides, our proposed approach exhibits superior performance compared to the LRU approach for large storage space. Also, by comparison between the proposed and optimal approaches, we deduce that the proposed algorithm provides a near-optimal solution while pursuing a minimum network cost.

(a)

(b)
Fig. 5 shows the cumulative request density of the cached files up to time slot with MB. The cumulative request density is the sum of request densities of desired files at each round. The figure shows that the proposed algorithm achieves larger cumulative request density than -greedy and random caching methods. In particular, at time slot , the cumulative request density achieved by the proposed greedy algorithm is , , , , and times the cumulative request density achieved by LRU, -greedy (for ), and random placement, respectively.
In Fig. 6, we evaluate the network cost of the proposed FRPL- and local caching methods. It can be seen from Fig. 6(a) that the cost of our cache placement solution is less than that of local caching. The figure also shows that, compared to the local caching, the FRPL approach requires fewer iterations to arrive at a given level of network cost. This is because the FRPL approach accurately learns the expected request density from the context of observed files’ rating and users’ features. Fig. 6(b) depicts the network cost of the proposed FRPL approach ( = 10) for distinct cache sizes. We assume that the caching entity (i.e., the SBS) updates its cache content every 4-hours. Fig. 6(b) then shows that expanding the cache from 5 to 32.5 reduces the network cost of the FRPL approach. The reason is that enlarging the cache size expands the solution space of problem (15).

(a)

(b)

(a)

(b)

(c)
Fig. 7(a) and Fig. 7(b) show the network cost of different placement methods as a function of the cache size ( of file library) and the transmission cost ratio , respectively. Fig. 7(a) shows that the proposed algorithm reduces the network cost compared to -greedy (), and local caching approaches for cache sizes varying from 10 to 50. The result in Fig. 7(a) is consistent with that in Fig. 6(a). Besides, Fig. 7(b) shows that the transmission cost ratio dramatically influences the cost performance of different methods. In particular, when increasing from 1 to 5, the cost of different placement approaches falls abruptly. Indeed, the parameter of normalized cost of file retrieval from the SBSs affects the performance more than that from the MBS over a backhaul.
In Fig. 8(a)-8(c), we evaluate the normalized network cost versus the average caching efficiency as a function of the transmission cost ratio and the cost of caching content for different placement approaches. We utilize movie rating data collected from MovieLens 1M [38] associated with distinct cache content update intervals. To model the user demand process, we synthesize two datasets with cache content update intervals equal to 15 hours (Fig. 8(a)) and 20 hours (Fig. 8(b) and Fig. 8(c)). Fig. 8(a) shows that: (i) with the decreasing cost of caching content , i.e., , the proposed placement approach reduces the normalized network cost; (ii) with the increasing transmission cost ratio and for a given , our approach improves the cost performance. Besides, Fig. 8(b) shows that when facing ultra-high request traffic, our approach reduces the normalized network cost. Finally, Fig. 8(c) reveals that the proposed placement approach reduces the network cost compared to the -greedy () and local caching methods. The reason is that our proposal accurately predicts the expected request density of each file particularly for high-dimensional datasets.
VI Concluding Remarks and Future Directions
We developed mobility-aware routing and caching strategies for resource-constrained small-cell networks in an FL framework. To jointly optimize the routing and cache placement, we first formulated the NP-hard network cost minimization problem. To address the complexity issue, we first proposed an approach by which the SBSs learn the pedestrian density and request density. Afterward, based on the predictions, a greedy algorithm optimizes the cache placement and minimizes the network cost. Numerical results established that our proposed method yields a near-optimal solution.
Potential future works include the extension of our proposal in the following directions:
(i) Integrating energy-efficient real-time routing architectures using machine learning methods to reduce the routing overhead [39] and discovery latency [40] in mobile edge networks.
(ii) Allowing for time-variant popularity (i.e., dynamic content library), which renders learning, prediction, and energy-efficient network performance challenging.
(iii) Finally, it is essential to investigate a proper offloading of the learning tasks to potential participants, under constraints such as spectrum scarcity, transmission latency, and the like.
Appendix A Proof of Proposition 1.
-means clustering performs poor on datasets with dimensional real space . As a result, the -means clustering often converges with at least one or more clusters with either empty or very few data points. In this case, to facilitate fast model aggregation, we discard the empty clusters or summarize very few data points when performing the pedestrian density prediction. That results in non-empty clusters. Cluster filtration is essential, as a modest value of enables solutions that summarize the underlying data better [35].
Thus, when the datasets for user clustering in transition regions of neighboring cells are scarce, the estimated pedestrian density reaches trough of . That results in a lower bound . In contrast, when the number of dimensions of datasets is larger than ,
the estimated pedestrian density reaches at a peak value of , resulting in an upper bound
, which completes the proof.
Appendix B Proof of Proposition 2.
The normalized caching cost at SBSs is given by
(18) |
Based on the definition of , can be rewritten as
(19) |
Then, by the definition of , the aggregated cost for caching the request contents at the serving edge equals
(20) |
When no user requests content that is in the cache of SBS , the involved normalized cost and also the the aggregated cost at the serving edge, i.e., , become zero.
If that event occurs, one concludes that the normalized cost in (19) equals the aggregated cost of MUs that fall into the edge , i.e., in (20).
When some MUs request a content that is not in the cache, the proposed placement policy selects a unique pair from associated with the lowest . This means that . In this case, we have . Meanwhile, . That is because a requested content that falls into the set will be cached at SBS deployed at either the intra-cell domain or the inter-cell domain, so that . Therefore the equality of the normalized cost in (19), i.e., traversing all the candidate pairs of the set , equaling to the aggregated cost in (20) holds, when an event of requesting an uncached content occurs.
Based on the discussion above we have
(21) |
This completes the proof of Proposition 2.
Appendix C Proof of Theorem 1.
The network cost minimization problem is given by
(22) |
Let and be two distinct subsets of , i.e., , , and . When keeping in the summation only a subset of the terms and all the terms are positive [23], i.e., , , , , , and , the following inequalities hold, i.e., , , and . Thus, the network cost in (22) can be rewritten by
(23) |
where inequality holds if one neglects the normalized cost for retrieving files from MBS, i.e., .
Based on Proposition 2, we arrive at the following equivalent transformations against the normalized cost for caching contents in (18) and the normalized cost for retrieving contents in (22) in the form of
(24) |
(25) |
Substituting (24) and (25) back into (23) yields (26), shown at the top of this page.
Based on the facts that , we therefore can arrive at that . Substituting this lower bound approximation into (26) results in (27), shown at the top of this page.
(26) |
(27) |
Next, we let be the optimal cache placement solution to the optimization problem (16). In addition, define as an optimal binary variable deciding whether file requested by MU is selected to be cached. Thus we have , and . Besides, by following the properties in Proposition 2, we have that with being the required number of iterations of Algorithm 2. In addition, let be the network cost of small-cell networks given that = 0. Eventually, we can arrive at
(28) |
By following (28), we conclude that where . Based on the above observations, we thereby deduce that when = 0, our proposed cache placement policy in Algorithm 2 indeed is of a polynomial-time -approximation algorithm. This completes the proof of Theorem 1.
References
- [1] Y. Cao, S. Maghsudi, and T. Ohtsuki, “Mobility-aware routing and caching: A federated learning assisted approach,” in ICC 2021 - IEEE International Conference on Communications, 2021, pp. 1–6.
- [2] H. Ahlehagh and S. Dey, “Video-aware scheduling and caching in the radio access network,” IEEE/ACM Trans. Netw., vol. 22, no. 5, pp. 1444–1462, Oct. 2014.
- [3] K. Shanmugam, N. Golrezaei, A. G. Dimakis, A. F. Molisch, and G. Caire, “Femtocaching: Wireless content delivery through distributed caching helpers,” IEEE Trans. Inf. Theory, vol. 59, no. 12, pp. 8402–8413, Dec. 2013.
- [4] S. Maghsudi and M. van der Schaar, “A non-stationary bandit-learning approach to energy-efficient femto-caching with rateless-coded transmission,” IEEE Trans. Wireless Commun., vol. 19, no. 7, pp. 5040–5056, Jul. 2020.
- [5] S. Maghsudi and E. Hossain, “Distributed user association in energy harvesting small cell networks: A probabilistic bandit model,” IEEE Trans. Wireless Commun., vol. 16, no. 3, pp. 1549–1563, Mar. 2017.
- [6] S. Li, J. Xu, M. van der Schaar, and W. Li, “Trend-aware video caching through online learning,” IEEE Trans. Multimedia, vol. 18, no. 12, pp. 2503–2516, Dec. 2016.
- [7] J. Konečný, H. McMahan, F. Yu, P. Richtárik, A. Suresh, and D. Bacon, “Federated learning: Strategies for improving communication efficiency,” arXiv preprint arXiv:1610.05492, Oct. 2017.
- [8] W. Y. B. Lim, N. C. Luong, D. T. Hoang, Y. Jiao, Y.-C. Liang, Q. Yang, D. Niyato, and C. Miao, “Federated learning in mobile edge networks: A comprehensive survey,” arXiv preprint arXiv:1909.11875, 2019.
- [9] H. H. Yang, Z. Liu, T. Q. S. Quek, and H. V. Poor, “Scheduling policies for federated learning in wireless networks,” IEEE Trans. Commun., vol. 68, no. 1, pp. 317–333, Jan. 2020.
- [10] G. Zhu, Y. Wang, and K. Huang, “Low-latency broadband analog aggregation for federated edge learning,” arXiv preprint arXiv:1812.11494, 2018.
- [11] T. Li, A. K. Sahu, A. Talwalkar, and V. Smith, “Federated learning: Challenges, methods, and future directions,” IEEE Signal Proces. Mag., vol. 37, no. 3, pp. 50–60, May 2020.
- [12] M. Chen, H. V. Poor, W. Saad, and S. Cui, “Convergence time optimization for federated learning over wireless networks,” IEEE Trans. Wireless Commun., vol. 20, no. 4, pp. 2457–2471, Apr. 2021.
- [13] L. Zhang, M. Xiao, G. Wu, and S. Li, “Efficient scheduling and power allocation for D2D-assisted wireless caching networks,” IEEE Trans. Commun., vol. 64, no. 6, pp. 2438–2452, Jun. 2016.
- [14] S. Boyd and L. Vandenberghe, Convex optimization. UK: Cambridge Uni. Press, 2004.
- [15] R. Wang, J. Zhang, S. H. Song, and K. B. Letaief, “Exploiting mobility in cache-assisted D2D networks: Performance analysis and optimization,” IEEE Trans. Wireless Commun., vol. 17, no. 8, pp. 5592–5605, Aug. 2018.
- [16] S. Maghsudi and D. Niyato, “On transmission mode selection in D2D-enhanced small cell networks,” IEEE Wireless Commun. Lett., vol. 6, no. 5, pp. 618–621, Oct. 2017.
- [17] S. Maghsudi and S. Stańczak, “Transmission mode selection for network-assisted device to device communication: A levy-bandit approach,” in 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Jul. 2014, pp. 7009–7013.
- [18] M. Gregori, J. Gómez-Vilardebó, J. Matamoros, and D. Gündüz, “Wireless content caching for small cell and D2D networks,” IEEE J. Sel. Areas Commun., vol. 34, no. 5, pp. 1222–1234, May 2016.
- [19] S. Müller, O. Atan, M. van der Schaar, and A. Klein, “Context-aware proactive content caching with service differentiation in wireless networks,” IEEE Trans. Wireless Commun., vol. 16, no. 2, pp. 1024–1036, Feb. 2017.
- [20] S. Müller, O. Atan, M. van der Schaar, and A. Klein, “Smart caching in wireless small cell networks via contextual multi-armed bandits,” in 2016 IEEE International Conference on Communications (ICC), Jul. 2016, pp. 1–7.
- [21] B. N. Bharath, K. G. Nagananda, and H. V. Poor, “A learning-based approach to caching in heterogenous small cell networks,” IEEE Trans. Commun., vol. 64, no. 4, pp. 1674–1686, Apr. 2016.
- [22] S. Maghsudi and M. van der Schaar, “A bandit learning approach to energy-efficient femto-caching under uncertainty,” in 2019 IEEE Global Communications Conference (GLOBECOM), Dec. 2019, pp. 1–6.
- [23] K. Poularakis, G. Iosifidis, V. Sourlas, and L. Tassiulas, “Exploiting caching and multicast for 5G wireless networks,” IEEE Trans. Wireless Commun., vol. 15, no. 4, pp. 2995–3007, Apr. 2016.
- [24] M. A. Maddah-Ali and U. Niesen, “Fundamental limits of caching,” IEEE Trans. Inf. Theory, vol. 60, no. 5, pp. 2856–2867, May 2014.
- [25] Y. N. Shnaiwer, S. Sorour, N. Aboutorab, P. Sadeghi, and T. Y. Al-Naffouri, “Network-coded content delivery in femtocaching-assisted cellular networks,” in 2015 IEEE Global Communications Conference (GLOBECOM), Dec. 2015, pp. 1–6.
- [26] R. Wang, J. Zhang, S. H. Song, and K. B. Letaief, “Mobility-aware caching in D2D networks,” IEEE Trans. Wireless Commun., vol. 16, no. 8, pp. 5001–5015, Aug. 2017.
- [27] M. Chen, Y. Hao, L. Hu, K. Huang, and V. K. N. Lau, “Green and mobility-aware caching in 5G networks,” IEEE Trans. Wireless Commun., vol. 16, no. 12, pp. 8347–8361, Dec. 2017.
- [28] Y. Guan, Y. Xiao, H. Feng, C. Shen, and L. J. Cimini, “Mobicacher: Mobility-aware content caching in small-cell networks,” in 2014 IEEE Global Communications Conference, Dec. 2014, pp. 4537–4542.
- [29] Z. Yang, Y. Liu, Y. Chen, and L. Jiao, “Learning automata based Q-learning for content placement in cooperative caching,” IEEE Trans. Commun., vol. 68, no. 6, pp. 3667–3680, Jun. 2020.
- [30] L. Breslau, P. Cao, L. Fan, G. Phillips, and S. Shenker, “Web caching and Zipf-like distributions: Evidence and implications,” in IEEE INFOCOM’99, vol. 1, Mar. 1999, pp. 126–134.
- [31] M. Hefeeda and O. Saleh, “Traffic modeling and proportional partial caching for peer-to-peer systems,” IEEE/ACM Trans. Netw., vol. 16, no. 6, pp. 1447–1460, Dec. 2008.
- [32] J. Konečný, H. McMahan, D. Ramage, and P. Richtárik, “Federated optimization: Distributed machine learning for on-device intelligence,” arXiv preprint arXiv:1610.02527, Oct. 2016.
- [33] B. McMahan, E. Moore, D. Ramage, S. Hampson, and B. A. y Arcas, “Communication-efficient learning of deep networks from decentralized data,” in Artificial intelligence and statistics, 2017, pp. 1273–1282.
- [34] N. P. Kuruvatti, A. Klein, and H. D. Schotten, “Prediction of dynamic crowd formation in cellular networks for activating small cells,” in 2015 IEEE 81st Vehicular Technology Conference, Jul. 2015, pp. 1–5.
- [35] K. Bennett, P. Bradley, and A. Demiriz, “Constrained K-means clustering,” in MSR-TR-2000-65, May 2000, pp. 1–9.
- [36] M. Chen, Z. Yang, W. Saad, C. Yin, H. V. Poor, and S. Cui, “Performance optimization of federated learning over wireless networks,” in 2019 IEEE Global Communications Conference (GLOBECOM), Dec. 2019, pp. 1–6.
- [37] H. Kellerer, U. Pferschy, and D. Pisinger, Knapsack Problems. Berlin, Germany: Springer-Verlag, 2004.
- [38] F. M. Harper and J. A. Konstan, “The movielens datasets: History and context,” ACM Transactions on Interactive Intelligent Systems, vol. 5, no. 4, Dec. 2015.
- [39] B. Tavli and W. Heinzelman, “Energy-efficient real-time multicast routing in mobile ad hoc networks,” IEEE Trans. Comput., vol. 60, no. 5, pp. 707–722, May 2011.
- [40] T. Imielinski and H. F. Korth, Mobile computing. Springer Science & Business Media, 1996, vol. 353.