EdgeMap: CrowdSourcing High Definition Map in Automotive Edge Computing
Abstract
High definition (HD) map needs to be updated frequently to capture road changes, which is constrained by limited specialized collection vehicles. To maintain an up-to-date map, we explore crowdsourcing data from connected vehicles. Updating the map collaboratively is, however, challenging under constrained transmission and computation resources in dynamic networks. In this paper, we propose EdgeMap, a crowdsourcing HD map to minimize the usage of network resources while maintaining the latency requirements. We design a DATE algorithm to adaptively offload vehicular data on a small time scale and reserve network resources on a large time scale, by leveraging the multi-agent deep reinforcement learning and Gaussian process regression. We evaluate the performance of EdgeMap with extensive network simulations in a time-driven end-to-end simulator. The results show that EdgeMap reduces more than 30% resource usage as compared to state-of-the-art solutions.
Index Terms:
HD Map, Automotive Edge Computing, Multi-Agent Deep Reinforcement LearningI Introduction
The high definition (HD) map is an essential building block for realizing autonomous driving (AV), as well as advanced driving assistance system (ADAS). It is developed to provide the most accurate presentation of the roads with extremely high precision, e.g., centimeter-level, which includes various information such as traffic signs and 3D point cloud. The HD map is built with diverse sensors, e.g., cameras, LiDAR, GPS, IMU, and radars, by using different techniques such as simultaneous localization and mapping (SLAM) algorithms.
To achieve precise localization of vehicles, the HD map needs to incorporate up-to-date road information such as constructions, accidents, and temporary maintenance. However, this can be challenging because updating the HD map usually requires specialized collection vehicles to traverse each road, collect and process information. Connected vehicles (CV) [1], which connects vehicles with wireless technologies, has been exploited to collaboratively build and update the HD map. CarMap [2] proposed a lean presentation of SLAM feature maps and substantially decreases the needed bandwidth of map updates. However, it requires all the computation of the SLAM, e.g., feature extraction and local mapping, to be executed on connected vehicles, which could pose an extra computation burden for vehicles.
Multi-access edge computing (MEC) [3] emerges recently to deploy distributed computation, storage, and networking resources, e.g., edge servers, in close proximity of users, e.g., at the network edge. By leveraging the capability of MEC, the SLAM computation for building the HD map can be offloaded to the edge server for the acceleration of computations. Nevertheless, transmitting the massive raw data generated by real-time sensors, e.g., RGB-D images, demands substantial radio resource usage and may overwhelm the radio transmission under a large number of connected vehicles. Thus, there is a pushing need to design an HD map that uses minimal resources while maintaining up-to-date information.
In this paper, we propose an HD map, named EdgeMap, which crowdsources the data from connected vehicles in automotive edge computing. We design EdgeMap to use the minimum network resources, i.e., radio bandwidth in the RAN and computation resource in the edge server, while satisfying the latency requirements. We formulate the vehicular offloading and resource reservation problem to seek the optimal offloading decision and resource usage. We design a novel DATE algorithm to decouple the problem into the vehicular offloading and resource reservation subproblem in a small and large time scale, respectively. On the one hand, we design an asynchronous multi-agent deep reinforcement learning (aMARL) method to intelligently offload vehicular data based on local network state and partial global state in a distributed manner. On the other hand, we design a novel learning-based gradient descent method based on Gaussian process regression (GPR) to online learn and update the resource reservation gradually.
The contributions of this paper are summarized as follows:
-
•
We design a new HD map, EdgeMap, by crowdsourcing data from connected vehicles in automotive edge computing.
-
•
We propose a novel DATE algorithm to minimize the resource usage of EdgeMap while maintaining the latency requirements by managing the vehicular offloading and resource reservation.
-
•
We evaluate the performance of EdgeMap via extensive network simulations in a time-driven end-to-end simulator. The results show that EdgeMap significantly outperforms state-of-the-art solutions.
II System Overview
Fig. 1 shows the overview of EdgeMap. In the data plane, the feature extraction module extracts the features from sensor images based on the ORB algorithm [4]. The features are used for the relocalization of vehicles locally based on the local map and further processed to build the global map based on the SLAM technique [4]. The SLAM computation can be partially executed in both the vehicle and the edge server, where the split ratio of the computation is determined by the vehicular offloading agent. The intermediate data generated by the computation in the vehicle is transmitted to the edge server, and the remaining computation is executed in the edge server. The global map is built and updated after the computation of a connected vehicle is completed. Then, the new updates in the global map are broadcast to all connected vehicles.
In the control plane, the two main modules, i.e., the vehicular offloading agent in each vehicle and the resource reservation controller, are designed to minimize resource usage while maintaining the latency requirements. The vehicular offloading agent is designed based on multi-agent DRL technique to observe local and global network states prepared by the state collector module, e.g., radio channel quality, vehicle mobility, and server workload, and optimize the split ratio of every vehicular offloading in a small time scale (e.g., subseconds). The resource reservation controller determines the reserved transmission and computation resources from the network operator, which is designed based on Gaussian process regression and runs on a large time scale (e.g., minutes).
III System Model and The Problem
We consider a mobile network with a cellular base station (BS) and an edge computing server111This system model can be easily extended to support multiple BSs and edge servers.. There are a set of vehicles, denoted as , that connect to the BS wirelessly. We illustrate the timeline of EdgeMap in Fig. 2, where two connected vehicles offload the computation to the edge server asynchronously and the resource reservation is configured every time interval. Each vehicular offloading is composed of four stages, i.e., local processing, uplink transmission, edge processing and downlink broadcast. Without loss of generality, we consider the computation can be split222The continuous computation splitting model can be easily extended to support different schemes such as discrete splitting. with any ratios ranging from 0 to 1. The split ratio of th vehicle at the time slot , denoted as , can be determined as the last offloading is completed. For example, if the split ratio is 0.2, it means there are 20% computation executed in the vehicle and the remaining 80% computation are processed in the edge server. Meanwhile, the intermediate data after the 20% computation are transmitted to the edge server for remaining computation. The updates of EdgeMap, e.g., new point clouds, are broadcasted to all connected vehicles via downlink broadcast. Since all connected vehicles are allowed to start their offloadings asynchronously, the vehicular offloadings may overlap with each other, e.g., overlaps occur in both radio transmission and edge processing as shown in Fig. 2.
To accomplish EdgeMap, the service provider requests network resource reservations in multiple technical domains from the network operator333The isolated resources can be provisioned by the network operator using the network slicing technique [5].. We consider there are three kinds of resources denoted as , i.e., uplink and downlink radio bandwidth in radio access networks, computing resource in the edge server, i.e., denotes the reserved th resource at the time . For the sake of simplicity, we denote the set of split ratios at the time slot as , and the set of resource reservations at the time as .
The EdgeMap is designed to provide up-to-date information about the transportation system, e.g., the latest point clouds. Thus, we define the latency of the th vehicle’s offloading started at the time slot as
(1) |
where is function with respect to both the split ratios and the resource reservations under the network state . This function is complicated by the asynchronous vehicular offloadings and their overlappings in different stages under high-dim network states. As a result, this function cannot be accurately represented by using mathematical expressions444Although there are several approximated latency models, they did not consider the overlapping scenarios and obtain poor performances (see Sec. V)..
Problem. We aim to minimize the total resource usage of EdgeMap while satisfying the latency requirements. Therefore, given a time period , e.g., 1 hour, we formulate the vehicular offloading and resource reservation problem as
(2) | ||||
(3) | ||||
(4) | ||||
(5) |
where is the weight factor for the th resource. The objective function in Eq. 2 is defined to minimize the total resource usage and the constraints in Eq. 3 assure the latency requirements of vehicular offloadings to be satisfied at any time slots. The constraints in Eq. 4 and Eq. 5 limit the lower and upper bound of the optimization variables, i.e., .
Challenges. The problem is difficult to be solved in three aspects. First, the latency function in Eq. 1 is unknown, which makes the problem to be a black-box optimization problem. As a result, model-based algorithms, e.g., linear programming and convex optimization [6], can not be used to solve the problem. Second, the problem is Markovian, where the decision of vehicular offloading affects not only the current latency of vehicles but also the future network state such as server workloads. As a result, the conventional black-box optimization that captures no time-correlated transitions, fails to address this problem. Third, the coupling optimization variables are heterogeneous in terms of time scales, where the asynchronous vehicular offloading decision of a vehicle can only be made upon its last offloading completion, and the resource reservation happens periodically.
IV The DATE Algorithm
In this section, we design a distributed adaptive offloading and resource reservation (DATE) algorithm to effectively solve the problem . First, we decouple the problem into two subproblems in different time scales, i.e., a vehicular offloading subproblem to determine the split ratio of vehicular offloadings and a resource reservation subproblem to optimize the reservation of network resources. Second, we address the vehicular offloading subproblem by proposing an asynchronous multi-agent deep reinforcement learning (aMARL) method based on proximal policy optimization (PPO) [7]. Third, we solve the resource reservation subproblem by proposing a learning-based gradient descent method based on Gaussian process regression (GPR).
IV-A Problem Decoupling
We decouple the problem into two subproblems, which is based on the observation that the resources are reserved in a large time scale (e.g., minutes) and the vehicular offloadings are determined in a small time scale (e.g., subseconds). In other words, when we optimize the split ratios for vehicular offloadings, the resource reservations are static.
On the one hand, we build the vehicle offloading subproblem under the static resource reservations as
(6) |
where we rewrite the original latency function in Eq. 1 into .
On the other hand, we build the resource reservation subproblem with the static vehicular offloading policy as
(7) | ||||
(8) | ||||
(9) |
where is the maximum latency among all vehicle offloadings. Here, we reduce the original latency function in Eq. 1 into , where the split ratios and states in the small time scale are omitted.
IV-B The Vehicular Offloading Subproblem
The vehicular offloading subproblem is challenging to solve due to the unknown latency function of vehicle offloadings. We address this problem by leveraging the multi-agent deep reinforcement learning (MARL) technique that has been increasingly applied to solve networking problems and demonstrated promising performance improvements.
IV-B1 MARL Basis
We consider a general multi-agent reinforcement learning setting, where multiple agents interact with an environment asynchronously in discrete decision epochs. At each decision time slot , the agents observe their local states , take actions based on their policies independently. These agents receive reward from the environment, and then the environment transits to the next state based on the actions taken by all agents. The objective is to find policies for all agents to map local states to actions, that maximizes the discounted cumulative reward , where is the discounted factor.
IV-B2 MARL Design
To apply the MARL for solving the vehicular offloading problem, we define the state, action space and reward function as follows.
State Space. We design the state space to incorporate the informative local vehicular status and comprehensive system status. The first part is , which are the CPU frequency, RAM, speed, and radio quality of the vehicle at the time slot , respectively. These metrics can provide information about hardware specifications and the mobility of the vehicle. The second part is , which are the server workload at the time slot and current resource reservations, respectively. These metrics indicate the workload of the edge server and available network resources for the system. Therefore, the state space can be expressed as
(10) |
Action Space. We design the action space to allow the vehicle to adaptively offload its computation to the edge server. The action space is expressed as
(11) |
for the th vehicle at the time slot .
Reward. We design the reward function to guide the policy update of the agents to minimize the latency of vehicular offloadings as shown in Eq. 6. The reward function is expressed as
(12) |
The objective is to derive the policies parameterized by deep neural networks that solve the vehicular offloading problem in Eq. 6. To this end, we design all the agents to share the identical policy network , i.e., parameter sharing [8]. As shown in Fig. 3, we create an agent in each vehicle to determine the split ratio of its vehicular offloadings. Here, we leverage the actor-critic architecture to achieve robust learning performances, i.e., an actor for taking the action, and a critic for guiding the update of the actor. To update the policy network, we centralize the experience from all the agents periodically, calculate the update of the policy network globally, and enforce the update for all the agents accordingly.
IV-B3 MARL Training
We design the training method of policy networks based on the proximal policy optimization (PPO) [7]. Denote as the policy with parameters , the vanilla policy gradient training method updates the actor policy network at step according to
(13) |
where is the step size (i.e., the learning rate), is the expected finite-horizon return of the policy and
(14) |
where is the advantage function of the current policy, and and are the action-value function and the value function, respectively. However, since this update is unconstrained, a single update step might lead to significant performances difference from the previous policy, and thus can lead to the collapse of policy performance.
We address this issue, based on PPO [7], by updating the actor policy network at step via
(15) |
Here, the surrogate objective function is expressed as
(16) |
where and is a hyperparameter to control the difference between the updated policy and the previous policy. In this way, the update steps with too large policy differences are prevented and the smooth improvement on policy performance is assured.

IV-C The Resource Reservation Subproblem
The resource reservation subproblem is difficult to solve because the latency function is unknown, even if under the static vehicular offloading policies. We solve this problem in two steps. First, we use the Gaussian process regression (GPR) to regress the unknown latency function and define expected gradients accordingly. Second, we design a learning-based gradient descent method to gradually update the resource reservation based on the expected gradients.
Gaussian Process Regression. Given a resource reservation , we can obtain the noisy observation under the static vehicular offloading policy , where is the Gaussian noise with a zero mean and variance. Denote and as the set of resource reservations and the corresponding observed latencies, respectively. The posterior distribution is expressed as
(17) |
where the observations denote . We use the Gaussian process (GP) to model the prior distribution [9]. Here, can be described as where and are the mean function and covariance function, respectively.
Given a resource allocation , the posterior distribution can be derived as
(18) |
where , and
(19) |
where and .
Expected Gradient. Based on the posterior distribution, we define the expected gradient [10] under the resource reservation as
(20) |
where and are an identity matrix and small positive constant, respectively.
Then, we solve the resource reservation subproblem by using the subgradient descent method [6] based on the expected gradients. The resource reservation at the time interval is updated according to
(21) | ||||
(22) |
where are appropriate step sizes, is the multiplier and .
IV-D The DATE Algorithm
We summarize the distributed adaptive offloading and resource reservation (DATE) algorithm as follows. First, we train the policy networks under the varying resource reservations. As the time to reserve resources, we centralize the transitions of all agents and used them to update the policy networks. After the training converges, we can obtain the optimal policy network . Second, we train the Gaussian process regression model with the optimal policy network. We use the optimal policy network to rollout the vehicular offloading until the time to reserve resources. The latency is calculated and the resource-latency pair is stored in the GPR training dataset. The GPR model is updated as new pairs are added in the training dataset. The resource reservation and the multiplier are also updated according to Eq. 21 and Eq. 22. As the algorithm completes, we obtain the optimal policy network and the well-trained GPR model.
V Performance Evaluation
In this section, we evaluate the performance of EdgeMap with extensive network simulations.
End-to-End Simulator. We build a time-driven network simulator, which is composed of multiple vehicle computation modules, a wireless transmission module, and a server computation module as depicted in Fig. 4. For each simulation time, e.g., 1ms, the tasks in either vehicle and server computation modules decrease their remaining time. Meanwhile, the tasks in either uplink and broadcast transmission modules update the remaining buffer size. In particular, when the vehicular offloading decision is determined, a task is created on this vehicle’s computation module. The task describes the remaining vehicle computation time, data size of uplink and broadcast transmission packet, and server computation time, where these data are sampled from experimental measurements. The simulation of server computation is based on multiple parallel FIFO (First-In-First-Out) service queues, where a default min-load scheduling scheme is applied to choose the queue for serving the incoming task. Besides, the higher computation capacity of the server reduces the computation time of the task in each service queue.
The wireless transmission module is developed based on an open-source 5G simulator [11], where we use the urban micro (UMi - Street Canyon) channel model recommended in ETSI TR 138.901 [12]. The maximum wireless bandwidth for uplink and downlink channels are 5MHz and the bandwidth is equally shared by all the vehicles if they have traffic to transmit (non-empty buffer). Thus, the transmission of tasks are simulated by calculating their wireless data rates555The mobility of vehicles and the location of the base station are obtained from a trace, which is collected from an intersection scenario built in Unity3d [13]. and decreasing their remaining uplink/downlink data sizes. A task is sent to the next simulation module only if it is completed in the last module, e.g., zero remaining computation time or transmission data size. Each vehicle starts the next vehicular offloading as the last offloading completes.
DRL agents. We implement multiple DRL agents by using PyTorch 1.9.1. We use 3-layer fully-connected neural network in both actor and critic networks, where there are 128 neurons in both layers with Leaky Recifier [14] activation functions, and adopt [14] as the activation function at the output layer of the actor. On training the DRL agents, we use the following hyper-parameters. The learning rates of actor and critic networks are 1e-4 and 3e-4, respectively. The maximum episode length is 100, an epoch includes 4000 transitions, and and . We add the decaying Gaussian noise on actions during the training phase for balancing the exploitation and exploration.

We use scikit-learn v1.0.0 [15] to build the Gaussian process regression (GPR) model. The resource reservation can be configured every 100 vehicular offloadings.






Parameters. To obtain the transmission and computation demands of vehicular offloadings, we conduct the experiments with ORB-SLAM3 [4] on the Mono mode. We use the EuRoC dataset with the scenarios of Machine Hall 01-05. We measure the data size of images, which have 353.5KB mean and 22.7KB std. We profile the ORB-SLAM3 computation latency with both a Desktop (AMD Ryzen 3600 3.8GHz, 32GB RAM) as the edge server and a Laptop (Intel i7-6500U 2.5GHz, 8GB RAM) as a vehicle. The computation latency in the Desktop and the Laptop are 286.54ms mean and 68.89ms std, and 609.27ms mean and 165.44ms std, respectively. Here the computation latency includes the tracking and local mapping, and excludes the global mapping. To determine the intermediate data size and computation under different split ratios, we linearly interpolate the data size for radio transmission and computation latency in the edge server. The weight factors for all kinds of resources are .
Comparison. We compare EdgeMap with the following algorithms:
-
•
Baseline: The Baseline offloads all the images to the edge server, i.e., the split ratio is always 0. The resource reservation is determined by decreasing from the maximum resource usage until the latency requirements are met.
- •
-
•
VirtualEdge: The VirtualEdge [10] offloads all the images to the edge server, uses GPR to regress the maximum latency function and updates the resource reservation with a proximal gradient descent method.
Convergence. Fig. 5 shows the average latency of vehicular offloadings achieved by EdgeMap during the training phase of aMARL agents. It can be observed that the aMARL agents gradually learn and reduce the latency of vehicular offloadings in 80 epochs. As compared to the initial performance at the epoch 0, i.e., 475.4ms average latency, the final trained aMARL agents achieve 352.8ms average latency, i.e., 25.8% latency reduction. This validates the effectiveness of the DATE algorithm on training multiple asynchronous agents simultaneously. Fig. 6 shows the overall convergence of EdgeMap, where the resource reservations are updated with the trained aMARL agents. In the first 40 iterations, we allow random exploration for initializing the GPR model. After that, the latency performance converges to the latency requirements (i.e., 500ms), along with the usage decrease on different resources. Eventually, EdgeMap uses 6.2% uplink radio bandwidth (0.31 MHz), 2% downlink radio bandwidth (0.1 MHz) and 19.4% computation resource (1.94 server capacity). This result justifies EdgeMap can achieve a significant reduction of resource usage while maintaining the given latency requirements.
Performance. Fig. 7 shows both latency and resource performance of different algorithms. It can be seen that Baseline and DCOR have similar performance, which satisfy the latency requirement and use nearly 42% resources on average. EdgeMap uses 9.2% average resource usage and obtains the average latency of 502.8 ms, in other words, it can reduce nearly 33% average resource usage as compared to Baseline and DCOR. Fig. 8 shows the cumulative probability of vehicular offloading latency under different methods. This result is obtained by randomly generating resource reservations and measuring the latency of each vehicular offloading. It can be observed that EdgeMap achieves the lowest latency (338.5ms mean) and the local process method gets the highest latency (716.0ms mean). This result validates that EdgeMap can intelligently optimize the adaptive vehicular offloading under different resource reservations.
Scalability. Fig. 9 shows the average resource usage of algorithms under different number of vehicles in the network. It can be seen that EdgeMap can always obtain the lowest average resource usage as compared to other algorithms, and the performance gap becomes larger as the number of vehicles increases. This result indicates EdgeMap can scale to manage vehicular offloading and resource reservation in large-scale networks with more vehicles. Fig. 10 shows the latency of vehicular offloading obtained by EdgeMap in different stages, i.e., local processing, uplink transmission, queuing and computing in the edge server, and the downlink broadcast. It can be seen that vehicular offloadings spend a major of time in the local processing stage on vehicles. This is because EdgeMap aims to minimize the resource usage, and the computing capacity on vehicles are exploited as much as possible to decrease the transmission data size and remaining computation at the edge server. For example, EdgeMap may need minimal resources if the local processing alone can satisfy the latency requirements.
VI Related Work
This work relates to the computation offloading and resource allocation in edge computing, which has been extensively studied with both model-based and model-free approaches. The model-based approaches [17, 16, 18] formulate the problem with mathematical expressions and derive the algorithms based on linear and non-linear, and convex optimization. For example, Zhao et. al proposed a distributed computation offloading and resource allocation (DCOR) algorithm to alternatively solve the subproblem of computation offloading and resource allocation. However, these approaches fail to accurately represent the problem due to the complex network dynamics, e.g., overlapping of computation offloadings. The mode-free approaches [19, 20, 21, 22] exploited the advanced machine learning techniques, e.g., DRL, to handle the high-dim network state and dynamics. For instance, Chen et. al proposed a temporal attentional deterministic policy gradient (TADPG) to improve task completion time and the average energy consumption by optimizing the partial task offloading. However, these approaches overlook the different time scales of computation offloading (in subseconds) and resource reservation (in minutes), and thus cannot be implemented in real operational networks.
VII Conclusion
In this paper, we have presented EdgeMap, a crowdsourcing HD map in automotive edge computing. We have formulated the vehicular offloading and resource reservation problem and proposed the DATE algorithm to solve it effectively. The performance of EdgeMap has been validated with extensive network simulations. The results have shown that EdgeMap significantly outperforms the state-of-the-art solutions in terms of performance and scalability.
References
- [1] J. Contreras-Castillo, S. Zeadally, and J. A. Guerrero-Ibañez, “Internet of vehicles: architecture, protocols, and security,” IEEE internet of things Journal, vol. 5, no. 5, pp. 3701–3709, 2017.
- [2] F. Ahmad et al., “Carmap: Fast 3d feature map updates for automobiles,” in 17th USENIX Symposium on Networked Systems Design and Implementation (NSDI 20), 2020, pp. 1063–1081.
- [3] S. Liu, L. Liu, J. Tang, B. Yu, Y. Wang, and W. Shi, “Edge computing for autonomous driving: Opportunities and challenges,” Proceedings of the IEEE, vol. 107, no. 8, pp. 1697–1716, 2019.
- [4] C. Campos et al., “Orb-slam3: An accurate open-source library for visual, visual–inertial, and multimap slam,” IEEE Transactions on Robotics, 2021.
- [5] Q. Liu, T. Han, and E. Moges, “Edgeslice: Slicing wireless edge computing network with decentralized deep reinforcement learning,” in 2020 IEEE ICDCS. IEEE, 2020, pp. 234–244.
- [6] S. Boyd and L. Vandenberghe, Convex Optimization. Cambridge university press, 2004.
- [7] J. Schulman, F. Wolski, P. Dhariwal, A. Radford, and O. Klimov, “Proximal policy optimization algorithms,” arXiv preprint arXiv:1707.06347, 2017.
- [8] J. K. Terry, N. Grammel, A. Hari, L. Santos, and B. Black, “Revisiting parameter sharing in multi-agent deep reinforcement learning,” arXiv preprint arXiv:2005.13625, 2020.
- [9] C. E. Rasmussen, “Gaussian processes in machine learning,” in Summer school on machine learning. Springer, 2003, pp. 63–71.
- [10] Q. Liu and T. Han, “Virtualedge: Multi-domain resource orchestration and virtualization in cellular edge computing,” in 2019 IEEE 39th International Conference on Distributed Computing Systems (ICDCS). IEEE, 2019, pp. 1051–1060.
- [11] E. J. Oughton, K. Katsaros, F. Entezami, D. Kaleshi, and J. Crowcroft, “An open-source techno-economic assessment framework for 5G deployment,” IEEE Access, vol. 7, pp. 155 930–155 940, 2019.
- [12] G. T. . R. 14, Study on channel model for frequencies from 0.5 to 100 GHz. 3GPP, 2018.
- [13] Q. Liu, T. Han, J. L. Xie, and B. Kim, “Livemap: Real-time dynamic map in automotive edge computing,” in IEEE INFOCOM 2021-IEEE Conference on Computer Communications. IEEE, 2021, pp. 1–10.
- [14] I. Goodfellow et al., Deep learning. MIT press, 2016.
- [15] F. Pedregosa et al., “Scikit-learn: Machine learning in Python,” Journal of Machine Learning Research, vol. 12, pp. 2825–2830, 2011.
- [16] J. Zhao, Q. Li, Y. Gong, and K. Zhang, “Computation offloading and resource allocation for cloud assisted mobile edge computing in vehicular networks,” IEEE Transactions on Vehicular Technology, vol. 68, no. 8, pp. 7944–7956, 2019.
- [17] C. Wang, C. Liang, F. R. Yu, Q. Chen, and L. Tang, “Computation offloading and resource allocation in wireless cellular networks with mobile edge computing,” IEEE Transactions on Wireless Communications, vol. 16, no. 8, pp. 4924–4938, 2017.
- [18] J. Zhang et al., “Joint computation offloading and resource allocation optimization in heterogeneous networks with mobile edge computing,” IEEE Access, vol. 6, pp. 19 324–19 337, 2018.
- [19] J. Li et al., “Deep reinforcement learning based computation offloading and resource allocation for mec,” in 2018 IEEE Wireless Communications and Networking Conference (WCNC). IEEE, 2018, pp. 1–6.
- [20] Y. Liu et al., “Deep reinforcement learning for offloading and resource allocation in vehicle edge computing and networks,” IEEE Transactions on Vehicular Technology, vol. 68, no. 11, pp. 11 158–11 168, 2019.
- [21] J. Chen et al., “A drl agent for jointly optimizing computation offloading and resource allocation in mec,” IEEE Internet of Things Journal, 2021.
- [22] S. Nath et al., “Multi-user multi-channel computation offloading and resource allocation for mobile edge computing,” in IEEE International Conference on Communications (ICC). IEEE, 2020, pp. 1–6.