This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Wireless and Service Allocation for Mobile Computation Offloading with Task Deadlines

Hong Chen1, Terence D. Todd1, Dongmei Zhao1 and George Karakostas2
This work has been submitted to the IEEE for possible publication. Copyright may be transferred without notice, after which this version may no longer be accessible. 1Department of Electrical and Computer Engineering
2Department of Computing and Software
McMaster University
Hamilton, Ontario, CANADA
Email: {chenh151,todd,dzhao,karakos}@mcmaster.ca
Abstract

In mobile computation offloading (MCO), mobile devices (MDs) can choose to either execute tasks locally or to have them executed on a remote edge server (ES). This paper addresses the problem of assigning both the wireless communication bandwidth needed, along with the ES capacity that is used for the task execution, so that task completion time constraints are satisfied. The objective is to obtain these allocations so that the average power consumption of the mobile devices is minimized, subject to a cost budget constraint. The paper includes contributions for both soft and hard task completion deadline constraints. The problems are first formulated as mixed integer nonlinear programs (MINLPs). Approximate solutions are then obtained by decomposing the problems into a collection of convex subproblems that can be efficiently solved. Results are presented that demonstrate the quality of the proposed solutions, which can achieve near optimum performance over a wide range of system parameters.

Index Terms:
Edge computing, mobile computation offloading, soft and hard task completion deadlines, cost budget constraints, power efficiency.

I Introduction

Mobile computation offloading (MCO) can be used to improve mobile device (MD) performance by running computational tasks on a remote cloud server rather than executing them locally [1, 2, 3]. Since the energy needed for task execution is incurred by the cloud server, a reduction in mobile device energy consumption can often be obtained [4, 5, 6, 7, 8, 9, 10]. During MCO, wireless communications is used by the MD to communicate with the cloud server. This interaction incurs MD energy use that would not otherwise exist if the task were executed at the MD. MCO also incurs added latency due to the time needed for the MD to interact with the cloud server [11, 12]. An edge server (ES) located close to the network base stations is typically used to reduce this delay by providing high interconnection bandwidth between the base station (BS) and the ES [13].

The question of whether a given task should be offloaded has been studied extensively [14, 15, 16, 17, 18, 19, 20, 21, 22, 23]. It is clear from this work that in order to obtain good performance, the offloading decisions should incorporate both the limited edge server computational capacity [21, 22, 23], and the temporal evolution of the system during the computation offload. This includes the queueing behaviour experienced by offloaded tasks awaiting execution at the ES [18, 19, 20]. Prior work has also considered the question of how best to configure system resources so that MCO is best accommodated [14, 18, 19, 20, 24, 25]. These are the issues that are considered in our paper and involve the tradeoffs between wireless communication and edge server capacity assignment and how these affect the delay performance experienced by the MDs.

The wireless and execution capacity assignment problem in MCO can be informally stated as follows. A network leaseholder (NL) purchases both wireless channel capacity and edge server execution services, subject to a cost budget constraint. The leased resources are then used to provide MCO to a large set of mobile devices [26]. When an MD generates a task for execution, there is an associated deadline, which gives the time by which task execution should be completed with a high degree of certainty [27]. The objective is to find a joint wireless and ES resource assignment that minimizes the mean MD power consumption subject to the budget constraint and constraints on the task completion times. Note that this problem is different than that of network slice creation [28]. In this case, the NL simply purchases services from the network owner (NO), who prices the cost of unit wireless channel and computational resources. Due to the edge server placement, we consider the case where the dominant latencies are that of wireless access and edge server execution [13].

The paper is novel in that it includes formulations for both soft and hard task completion time deadlines. In the soft deadline case, the wireless and edge server capacity assignments are designed so that the probability of task completion time deadline violation is upper bounded. In the hard deadline case, task execution deadlines must always be respected, which is accomplished by including concurrent local execution (CLE) [29] into the problem formulation. In CLE, local execution of the task may be initiated while offloading is ongoing, so that the task completion time deadline is always met.

The inclusion of task deadline constraints significantly increases the difficulty of the problem compared to that of prior work with no completion time requirements or that uses a mean delay criterion [30, 31]. In order to obtain solutions to the problem, a queuing model is used to obtain the delay distribution experienced by tasks that are offloaded to the ES [32, 31]. This model is incorporated into the resulting optimization problems, which are formulated as mixed integer nonlinear programming problems (MINLPs) that are computationally hard to solve exactly. Approximate solutions are obtained by decomposing the non-convex non-linear formulation into a collection of convex subproblems that can be solved efficiently, and then picking the best of these solutions.

A variety of results are presented that characterize the tradeoffs between task deadline violation, average MD power consumption and the cost budget. Our results show the quality of the proposed solutions, which can achieve close-to-optimum performance for a wide range of system parameters. The results also show that with CLE, the proposed solution not only guarantees respecting all hard task completion deadlines, but does so with only slightly higher MD power consumption when compared to the soft task completion deadlines solution with a small deadline violation probability. On the other hand, we show that there is an apparent trade-off in the case of soft task completion deadlines between the average power consumption and the deadline violation probability. Namely, the average MD power consumption of our solution is significantly reduced when a higher deadline violation probability is tolerable.

The main contributions of the paper are summarized below.

  • This paper addresses the problem of assigning computational and wireless channel resources for MCO, subject to task execution completion time deadlines. The work is the first that generates joint resource assignments for both soft and hard task deadlines using very general system modelling assumptions compared to prior work. The soft deadline case aims to create assignments so that the probability of task completion time deadline violation is upper bounded. In the hard deadline case, the paper is also unique in that it creates resource assignments where task completion time deadlines are always satisfied. This is done by incorporating CLE into the problem formulations. For this reason, this is the first paper that obtains system resource assignments for MCO that ensure that task completion time deadlines are always satisfied.

  • Modeling both soft and hard job completion time targets significantly increases the difficulty of the problem compared to prior work with no completion time requirements or that uses a mean delay criterion [30] [31]. In both deadline cases, the paper addresses this by incorporating an ES queueing system into the problem formulation that models the delay distribution experienced by arriving tasks. The assignment problem is addressed by inverting the estimated probability density function (PDF) of the task completion time and incorporating it into the optimizations. These resource assignments are obtained under very general modeling assumptions, where the wireless channels are modeled as arbitrary base station specific sets of Markov processes and task execution times have a general probability distribution.

  • The problems are first formulated as MINLPs, with integral decision variables for the number of wireless channels reserved, and a continuous decision variable for the portion of ES reserved. Even the relaxations of these MINLPs are difficult to solve, since they are non-convex. Hence, instead of following the common practice of solving the relaxation and rounding the fractional solution, we observe that the discretization of the continuous variable and the replacement of the discrete channel variables by approximate functions of the continuous blocking probabilities, allows us to break the original non-convex MINLPs into collections of convex subproblems, that can be solved efficiently. Our solutions are approximate, and their accuracy depends on both the discretization granularity and the approximation functions used for blocking probabilities. On the other hand, they are based on very general assumptions, i.e., the existence of convex upper bound approximations of the inversion of blocking probabilities. The more restricted the system model is, the better these approximations are.

The remainder of the paper is organized as follows. In Section II the prior work most related to our paper is reviewed. The system model and problem formulation is then described in Section III. In Section III-A, the general design problem is first considered assuming soft task completion time deadlines, where the probability of deadline violation is bounded. Following this, in Section III-B a formulation is described when task completion times are subject to hard deadlines. The problem formulations in both cases are non-convex and difficult to deal with directly using conventional optimization approaches. In Section IV, approximation solutions are proposed where the original problems are decomposed into convex subproblems that can be efficiently solved. Both the soft and hard deadline cases are considered in Sections IV-A and IV-B. Section V then introduces some common system assumptions used in the remainder of the paper when solving the optimizations. Both the soft and hard deadline cases are then treated in detail in Sections V-A and V-B. In Section VI simulation results that demonstrate the proposed designs are given. Both the single class and multiple classes of tasks cases are considered in Sections VI-A and VI-B. Finally, we present our conclusions of the work in Section VII.

II Related Work

A large amount of prior MCO work considers the problem based on system state inputs sampled at task generation times, i.e., the models assume that the system is static throughout the offload period [14, 18, 19, 20, 24, 25, 21, 22, 15, 17, 33, 34, 23]. As in our paper, task offloading decisions become more complex when the MD interacts with the network over wireless channels that may change randomly during the offload. Reference [32] studies a distributed computation offloading problem with delay constraints using stochastic communication channels but does not take into account the energy consumption incurred during task offloading. The work in [30] uses a Markov decision process that analyzes the mean task delay and the average system throughput. Unlike our paper, a throughput maximization problem is formulated with constraints on the average task delay, rather than using the delay distribution. In [31], task offloading is modeled as a game using a network of queues to obtain the end-to-end delay. The problem is transformed into one with a generalized Nash equilibrium solution that captures the conflicting interests in resource allocation among mobile network operators and computing resource providers. In references [30] and [31] the average delay is considered rather than the stringent types of soft and hard delay constraints considered in our paper. Reference [35] considers task offloading with statistical QoS guarantees (i.e., tasks are allowed to complete before a given deadline with a probability above a given threshold) to maximize the MD energy efficiency. The energy efficiency is defined as the ratio of the overall executed (transmitted) bits of tasks to the total energy consumption of the MDs. Statistical computation and transmission models are introduced to quantify the correlation between the statistical quality of service (QoS) guarantee and task offloading process. Unlike the models used in [31] and [35], our paper uses a task offloading and resource allocation formulation that uses very general system model assumptions, including base station specific sets of Markov processes for channel modelling.

TABLE I: Related Work Summary
References Joint channel and computation resource assignment Soft task deadlines Hard task deadlines Resource expense Temporal evolution
[17][21][22][23] \checkmark
[24][36]
[26] \checkmark \checkmark
[30][31]
[32] \checkmark
[35]
Our paper

Reducing both mobile energy consumption and task execution time is a common objective in mobile computation offloading. The work in [36] investigates a latency minimization problem in a multi-user time-division multiple access system with joint communication and computation resource allocation. Our paper, instead, uses a soft task deadline criterion based on modelling the distributions of both upload and execution time delays. Hard completion time constraints are considered in references [17, 21, 22, 33, 32, 23]. However, unlike our work, they consider the hard completion time requirement as a constraint in the problem formulation. For this reason, if the provided network resources or the MD transmit power are insufficient, the hard completion time constraints may not be satisfied. In our work, we avoid this infeasibility by applying CLE that ensures that hard completion time constraints are always satisfied. A benefit from integrating CLE into the problem formulation is that we no longer require the hard completion time constraints in the problem formulation. The objective in [21] is to minimize the energy consumption of the entire system, and in references [22] and [33], the objective is to minimize the total energy consumption of all MDs. Instead of satisfying delay constraints, the work in [24, 25, 34] optimize a utility function that is a weighted sum of task completion time and energy consumption. Unlike the above work, two different kinds of delay constraints are introduced in our paper, i.e., soft deadlines captured by the statistics of the completion time of the tasks and hard deadlines that are always satisfied by CLE.

Prior work has considered the optimization of wireless network and computational server resources to improve MCO performance [14, 18, 19, 20, 24, 25]. In particular, offloading decisions and base station associations are optimized with transmission power and channel assignments in a cellular network to minimize the total energy consumption of all MDs, subject to task’s latency constraints [17]. Reference [21] studies the problem of task offloading and channel resource allocation for ultra-dense networks and minimizes the total energy consumption of the system with a limited delay tolerance. The work in [22] studies MCO by considering application latency fairness and minimizes MD energy consumption by jointly optimizing the offloading ratio, channel assignments, and channel time allocations. Reference [23] investigates the power minimization problem for meeting the service delay requirements in multi-cell multi-user mobile edge computing networks. Channel assignment and power allocation problems are considered jointly. The work in [26] studies the joint resource management of link scheduling, channel assignment and power control for device-to-device communication assisted multi-tier fog computing with the objective of maximizing the network operator profit under deadline requirements. It considers the service charge collected from all end users, total expense in renting third-party fog nodes, and the energy cost of the ES. All of this work [21, 22, 17, 23, 26] optimizes radio resources and offloading decisions without considering edge server computational capability. The work in [24] investigates relay-assisted computation offloading to minimize the weighted sum of task execution delay and the energy consumption by jointly optimizing the offloading ratio, bandwidth allocation, processor speeds, and transmit power.

Table I summarizes the work described above that is most related to our paper, and compares it to this paper on five key properties: {LaTeXdescription}

The column denotes work where both channel and computation resource assignments are jointly generated. Our work differs from the rest in that we assign aggregate channel resources from the network operator to each base station so that it can support its associated mobile device population, i.e., we do not allocate channel and computation resources of each BS and ES to individual MDs.

The work selected in this column considers some form of soft (i.e., statistical) task deadlines. However, the models we use in this paper are quite different with more general underlying assumptions. Since our soft deadline model aims to set bounds on the probability of task deadline violation, we model the complete delay distribution experienced by executed tasks. This includes the base station channel delay (which is modeled by base station specific Markov processes) and the queueing delay experienced at the ES, where execution times can have a general distribution.

Although there is other work selected in this column, a significant difference exists compared with our paper, which we have already discussed above. Namely, our work can always satisfy all hard task deadlines by incorporating the CLE mechanism into the modeled system. The related work, instead, considers the existence of hard deadlines as a problem constraint that may result in problem infeasibility, which can never happen in our case.

This column denotes work where the resources provided to the MDs are charged by a third-party (e.g., network operator). The work selected considers computational resource expense but not on the wireless base station side. A network profit maximization problem is studied where an expense budget is not considered, unlike the case in our work.

Temporal evolution means that the offload periods may include stochastic changes to the wireless channels and the ES, so that this information must be modeled in the problem formulation, as in our paper. The randomness modeled in the selected work has different underlying assumptions compared to our paper.

III System Model and Problem Formulation

As shown in Fig. 1, we consider a network that consists of NN BSs that are owned and operated by a NO. The set of BSs is denoted by 𝒩={1,2,,N}\mathcal{N}=\{1,2,\ldots,N\} and indexed by n𝒩n\in\mathcal{N}. The network also contains an ES. Tasks generated by an MD can be offloaded through the wireless network and executed on the ES.

The NO permits a NL to rent wireless communication and ES computational capacity that the NL can use for mobile computation offloading for its MDs. When this is done, for each BS nn, there are up to KnK_{n} available channels that can be selected by the NL. The cost of renting a channel from BS nn is set by the NO to αn\alpha_{n}. When a channel is included in the agreement, the NO agrees to provision its network so that sufficient resources are available to allow the traffic generated on the channel to be carried to the ES with an acceptable delay with a high degree of certainty. Since the ES is located at the edge of the network, we focus on the dominant sources of delay, i.e., wireless access at the BSs and task execution at the ES [13].

Refer to caption
Figure 1: System Model

In order to use the computing resources at the ES, the NL must also lease CPU resources at the ES. The cost (based on the number of CPU cycles per second) for leasing on the CPU resource is denoted by β\beta. The maximum available CPU speed for rental is fCf^{\rm{C}} CPU cycles per second.

When an agreement is made between the NO and NL, xnx_{n} is defined as the number of channels from BS nn that are included, and y[0,1]y\in\left[{0,1}\right] is defined as the fraction of maximum CPU speed at the ES that is included, i.e., the CPU speed available for the NL will be yfCyf^{\rm{C}}. It is assumed that the NL has a cost budget, denoted by BmaxB^{\text{max}}. Accordingly, the total rent must satisfy the following constraint:

n=1Nαnxn+βyfCBmax.\textstyle\sum_{n=1}^{N}{{\alpha_{n}}{x_{n}}}+\beta y{f^{\rm{C}}}\leq{B^{\max}}. (1)

There are JJ classes of tasks generated by the MDs, which may need to be offloaded to the ES. Let 𝒥={1,2,,J}\mathcal{J}=\{1,2,\ldots,J\} be the set of task classes. The class jj of a task is defined by parameters sjs_{j}, qjq_{j}, and djd_{j}, where sjs_{j} is the input data size in bits, qjq_{j} is the computation load in number of CPU cycles, and djd_{j} is the deadline of the task in seconds. In what follows, d~j=dj/τ\tilde{d}_{j}=\left\lfloor{d_{j}/\tau}\right\rfloor is the task deadline rounded down to time slots of the same duration τ\tau as the wireless transmission time slots (see below). The probability of a task generated by an MD belonging to class jj is denoted by PjCP_{j}^{\rm{C}}; we assume that this probability is known, e.g., by observing the past history of offloading requests.

Our objective is to create a NO/NL contract for MCO. In MCO, tasks generated by an MD can be executed either locally (at the MD itself) or offloaded through the network and executed on the ES. We focus on two goals, each depending on how hard the task deadline constraint is. Our first goal is to accomplish this so that the mean mobile power consumption is minimized subject to the cost budget constraint and such that the probability that task execution deadline violation is bounded, i.e., the deadline constraints can be violated, albeit rarely. Our second goal is to create a power-efficient, budget-respecting assignment which respects all task deadlines, i.e., deadline constraints are hard; for that purpose we will employ CLE [29].

We model the wireless channels between the MDs and the BSs as discrete-time Markov processes. It is assumed that there are InI_{n} channel models for BS nn, which are a function of the radio propagation environment that the MDs experience at that BS. n={1,2,,In}\mathcal{I}_{n}=\{1,2,\ldots,I_{n}\} is the set of all wireless channel models in BS nn. For each of the channel models, the Markovian transition probabilities are defined in the usual way, i.e., given the channel state in the current time slot, there is a probability associated to its transition to another state in the next time slot. The time slot duration is defined to be τ\tau seconds. A class jj task, offloaded to BS nn by the MD, encounters channel model kk with probability Pn,j,kGP^{\rm{G}}_{n,j,k}; as with task generation probabilities PjCP_{j}^{\rm{C}} above, we assume that this probability is also known, e.g., by observing the past history of offloading requests.

To obtain the design, the decision to offload the execution of a task is made using a local execute on blocking (LEB) mechanism as follows. When an MD in BS nn generates a class jj task, the MD offloads the task if at least one of the xnx_{n} channels is available for immediate use. Otherwise, the MD executes the task locally. When a channel is available, the MD begins the offload by uploading the sjs_{j} task bits needed for execution on the ES. The LEB mechanism is useful in that either local execution or remote offloading is initiated immediately at task release time, which may be advantageous when task deadlines are tight. It also provides a simple mechanism for assessing when the current level of local congestion is high, which would suggest that local execution is beneficial.

Tasks arrive at BS nn according to a stationary process with average arrival rate λn\lambda_{n} tasks per second. According to the LEB mechanism, a new task is blocked from BS channel access if all the xnx_{n} channels are busy with uploading other tasks. We denote the task blocking probability at BS nn by PBn(xn)P_{{\rm{B}}n}(x_{n}), which is a function of xnx_{n}. For the sake of notation simplicity, we use PBnP_{{\rm{B}}n} in the rest of the paper. Let pLp^{\text{L}} be the power needed in the MD to process tasks. When a class jj task is blocked from offloading and executed locally, the local execution time is given as Lj=qj/fL_{j}={q_{j}}/f, where ff is the MD’s execution speed in number of CPU cycles per time slot111LjL_{j} is normally measured in CPU cycles, but in order to apply CLE and to simplify the system, we round it up to a multiple of τ\tau.. Define L¯\bar{L} as the average local execution time of tasks. Since the task blocking is caused by channel access, which is the same for all task classes, we have L¯=j=1JPjCLj\bar{L}=\sum_{j=1}^{J}P_{j}^{\rm{C}}L_{j}. The average energy consumption for executing a task locally is given by pLL¯p^{\rm{L}}{\bar{L}}. Consider all the tasks that are generated in BS nn and blocked from offloading in one second, then the mean energy for executing these tasks locally is

EnL(xn)=PBnλnpLL¯,E_{n}^{\rm{L}}(x_{n})={{P_{{\rm{B}}n}}{{\lambda_{n}}{p^{\rm{L}}}{\bar{L}}}}, (2)

which is the average power consumption of the MDs.

The wireless upload transmission time tn,j,kWt_{n,j,k}^{\text{W}} of a jjth class task in BS nn when the wireless channel model is kk, is measured in time slots. The mean wireless upload transmission time t¯n,j,kW\bar{t}^{\rm{W}}_{n,j,k} for jjth class tasks in BS nn according to channel model kk can be calculated, since Pr[tn,j,kW=l]\Pr[{t_{n,j,k}^{\rm{W}}=l}] can be computed for all ll from channel model kk. Moreover, the mean wireless transmission time t¯nW\bar{t}_{n}^{\rm{W}} for BS nn is

t¯nW=j=1Jk=1InPjCPn,j,kGt¯n,j,kW.\bar{t}_{n}^{\rm{W}}=\sum\limits_{j=1}^{J}\sum\limits_{k=1}^{{I_{n}}}{P_{j}^{\rm{C}}}P^{\rm{G}}_{n,j,k}\bar{t}_{n,j,k}^{\rm{W}}. (3)

Under the stated assumptions, the aggregate mean task arrival rate λ\lambda at the ES is given by

λ=n=1N(1PBn)λn.\lambda=\textstyle\sum_{n=1}^{N}{({1-{P_{{\rm{B}}n}}}){{\lambda_{n}}}}. (4)

As is normally the case for stability in a single server queueing system, the following constraint must always be satisfied,

λ<μC,\lambda<{\mu^{\rm{C}}}, (5)

where μC{\mu^{\rm{C}}} denotes the mean service rate at the ES, i.e, μC=yfC/j=1JPjCqj{\mu^{\rm{C}}}=yf^{\rm{C}}/\sum_{j=1}^{J}P_{j}^{\rm{C}}q_{j}. As will become clear later, we can relax this constraint to λyfC/j=1JPjCqj\lambda\leq yf^{\rm{C}}/\sum_{j=1}^{J}P_{j}^{\rm{C}}q_{j} without affecting our proposed solutions.

Let tn,j,kCt_{n,j,k}^{\rm{C}} be the delay (including both queueing and execution time) experienced by a jjth class task from BS nn at the ES, under wireless channel model kk. It takes continuous values, and Pr[tn,j,kCt]\Pr[{{t_{n,j,k}^{\rm{C}}}\leq t}], for any t0t\geq 0, is a function of λ\lambda and μC\mu^{\rm{C}}. In what follows, t~n,j,kC{\tilde{t}}_{n,j,k}^{\rm{C}} is the discretization of tn,j,kCt_{n,j,k}^{\rm{C}}, measured in time slots; its distribution is calculated by

Pr[t~n,j,kC=b]=Pr[tn,j,kCbτ]Pr[tn,j,kC(b1)τ]\displaystyle\Pr[{{{\tilde{t}}_{n,j,k}^{\rm{C}}}=b}]=\Pr[{{t_{n,j,k}^{\rm{C}}}\leq b\tau}]-\Pr[{{t_{n,j,k}^{\rm{C}}}\leq(b-1)\tau}] (6)

for any number of time slots b0b\geq 0. Table II lists the related notation and their associated meanings.

TABLE II: Summary of Notation
Notation Definition Units
𝒩\mathcal{N} Set of BSs, |𝒩|=N|\mathcal{N}|=N
𝒥\mathcal{J} Set of task classes, |𝒥|=J|\mathcal{J}|=J
n\mathcal{I}_{n} Set of channel models of BS nn, |n|=In|\mathcal{I}_{n}|=I_{n}
KnK_{n} Number of available channels in BS nn
fCf^{\rm{C}} Maximum available ES capacity CPU cycles/sec
αn\alpha_{n} Unit price of wireless channels from BS nn $ per channel
β\beta Unit price of ES capacity $ per bps
xnx_{n} Number of channels from BS nn
yy Fraction of maximum ES capacity
BmaxB^{\max} Cost budget $
sjs_{j} Data size of a task in class jj bits
qjq_{j} Computation load of a task in class jj CPU cycles
djd_{j} Deadline of a task in class jj sec
d~j\tilde{d}_{j} Discretized deadline of a task in class jj Time slots
PjCP_{j}^{\rm{C}} Probability of a task belonging to class jj
Pn,j,kGP^{\rm{G}}_{n,j,k} Probability of a class jj task in BS nn with channel model kk
PBnP_{{\rm{B}}n} Blocking probability in BS nn
μC\mu^{\rm{C}} Mean service rate at the ES Tasks/sec
λn\lambda_{n} Average task arrival rate in BS nn Tasks/sec
λ\lambda Aggregate average task arrival rate at ES Tasks/sec
τ\tau Time slot sec
tn,j,kWt_{n,j,k}^{\text{W}} Wireless transmission time of a jjth class task in BS nn with channel model kk Time slots
t¯n,j,kW\bar{t}_{n,j,k}^{\text{W}} Mean wireless transmission time of a jjth class task in BS nn with channel model kk Time slots
t¯nW\bar{t}_{n}^{\text{W}} Mean task uploading transmission time in BS nn Time slots
tn,j,kCt_{n,j,k}^{\rm{C}} Execution time at ES for class jj tasks from BS nn with channel model kk sec
t~n,j,kC{\tilde{t}}_{n,j,k}^{\rm{C}} Discretized value of tn,j,kCt_{n,j,k}^{\rm{C}} Time slots
tjLt_{j}^{\rm{L}} Latest feasible starting time for local execution Time slots
εj\varepsilon_{j} Tolerable probability a class jj task exceeds deadline
pLp^{\rm{L}} Local energy consumption per time slot Joules
pTp^{\rm{T}} Wireless transmission energy per time slot Joules
EnTE_{n}^{\rm{T}} Average MD power consumption for uploading tasks in BS nn Watts
EnCE_{n}^{\rm{C}} Average MD power consumption for uploading and executing tasks in BS nn Watts

III-A Problem Formulation with Soft Deadlines

We consider the distribution of total delay for an offloaded task, which is the sum of the data upload delay tn,j,kWt_{n,j,k}^{\rm{W}} and the task execution at ES delay tn,j,kCt_{n,j,k}^{\rm{C}}, for BS nn, task class jj, and channel model kk. Note that both delays are random variables. As mentioned earlier, the data transmission delay from the BS to the ES is negligible. In addition, in this paper we consider the case of a very small amount of data returned once the execution is completed, and, therefore, we consider only uploading delays between MD and BS.

Following common practice (e.g., [27]) in modelling soft deadlines along the lines of QoS requirements, a jjth class task in BS nn under wireless channel model kk, must have a total delay satisfying

Pr[tn,j,kW+tn,j,kCdj]1εj,{\rm{Pr}}[{t_{n,j,k}^{\rm{W}}+t_{n,j,k}^{\rm{C}}\leq{d_{j}}}]\geq 1-{\varepsilon_{j}}, (7)

where 0<εj10<\varepsilon_{j}\leq 1 is the (given) tolerated probability the completion time of a class jj task exceeds its deadline.222The case εj=0\varepsilon_{j}=0 corresponds to the case of hard deadlines, and will be dealt with in the next section. Note that tn,j,kWt_{n,j,k}^{\rm{W}} takes discrete values (number of time slots), tn,j,kCt_{n,j,k}^{\rm{C}} takes discrete values (number of CPU cycle periods), while djd_{j} is continuous (in seconds), so (7) assumes that all quantities are first converted to secs. Its LHS is a function of xn,yx_{n},y.

The joint probability distribution of total delay is

Pr[tn,j,kW+tn,j,kCdj]=l=1lmaxPr[tn,j,kW=l]Pr[tn,j,kCdjlτ],\Pr[{t_{n,j,k}^{\rm{W}}+t_{n,j,k}^{\rm{C}}\leq{d_{j}}}]=\\ \sum_{l=1}^{l_{\max}}{\Pr[{t_{n,j,k}^{\rm{W}}=l}]}\Pr[{t_{n,j,k}^{\rm{C}}\leq{d_{j}}-l\tau}], (8)

where lmax=(djqj/yfC)/τl_{\max}=\lfloor(d_{j}-{q_{j}}/{yf^{\rm C}})/{\tau}\rfloor is the maximum value that ll can take, since qj/yfC{q_{j}}/{yf^{\rm C}} is the execution time at the ES without queueing.

The average power consumption of MDs in BS nn to upload tasks that are granted channels for offloading is

EnT(xn)=(1PBn)λnpTt¯nW,E_{n}^{\rm{T}}(x_{n})={({1-{P_{{\rm{B}}n}}}){{\lambda_{n}}}{p^{\rm{T}}}\bar{t}_{n}^{\rm{W}}}, (9)

where pT{p^{\rm{T}}} is the transmission energy per time slot used by the MD for uploading the task bits. Therefore, the expected average power consumption of the MDs for uploading and executing tasks arriving at BS nn is EnL(xn)+EnT(xn)E_{n}^{\rm{L}}(x_{n})+E_{n}^{\rm{T}}(x_{n}).

Our objective is to create an allocation that minimizes EnL(xn)+EnT(xn)E_{n}^{\rm{L}}(x_{n})+E_{n}^{\rm{T}}(x_{n}) under the cost budget and deadline constraints (1) and (7). The problem can be formulated as follows:

min𝕩,yn=1N[EnL(xn)+EnT\displaystyle\mathop{\min}_{\mathbb{x},y}\sum\limits_{n=1}^{N}[E_{n}^{\rm{L}}(x_{n})+E_{n}^{\rm{T}} (xn)] s.t.\displaystyle(x_{n})]\text{ s.t.} (10)
n=1Nαnxn+βfCy\displaystyle\sum\limits_{n=1}^{N}{{\alpha_{n}}{x_{n}}}+\beta{f^{\rm{C}}}y Bmax\displaystyle\leq{B^{\max}} (11)
Pr[tn,j,kW+tn,j,kCdj]\displaystyle{\rm{Pr}}[{t_{n,j,k}^{\rm{W}}+t_{n,j,k}^{\rm{C}}\leq{d_{j}}}] 1εj,\displaystyle\geq 1-{\varepsilon_{j}}, n,j,k\displaystyle\forall n,j,k (12)
(fC/j=1JPjCqj)y\displaystyle(f^{\rm{C}}/\sum\limits_{j=1}^{J}P_{j}^{\rm{C}}q_{j})y λ\displaystyle\geq\lambda (13)
xn\displaystyle x_{n} {0,1,,Kn},\displaystyle\in\{0,1,\ldots,K_{n}\},\ \ \ n𝒩\displaystyle\forall n\in\mathcal{N} (14)
0y\displaystyle 0\leq y 1.\displaystyle\leq 1. (15)

Constraints (11) and (12) are constraints (1) and (7). Constraint (13) is the (relaxed) queue stability requirement for ES; it is equivalent to (5), since equality leads to infinite mean queueing delay, which is never optimal. The optimization problem (10)-(15) is a mixed integer nonlinear programming (MINLP) problem. Constraint (14) ensures that the number of channels assigned does not exceed the maximum number available in each BS. Even the fractional relaxation of MINLP problem (10)-(15) is non-convex, due to its objective and constraints (12), and, as a result, it is computationally inefficient to solve it exactly. Hence we are going to propose approximate solutions for it.

III-B Problem Formulation with Hard Deadlines

For the case of hard deadline constraints, i.e., when the task deadline must be respected, we employ CLE [29]. In CLE, local execution of the task may be initiated while offloading is ongoing, so that the task deadline is always met, even if offloading fails to finish in time due to the stochastic nature of the wireless channels. Guaranteeing task completion before its deadline may incur additional costs (due to potentially simultaneous local and remote execution of the same task).

When CLE is employed, and in order to ensure that the local execution of a task from class jj finishes by its deadline, the latest feasible starting time for local execution is

tjL=d~jLj+1.t_{j}^{\rm{L}}=\tilde{d}_{j}-L_{j}+1. (16)

The expected wireless transmission power is still given by (9). However, due to the overlap of offloading and local execution because of CLE, there is an extra mean power consumption due to a (potential) overlap with local execution. This expected overlap power consumption is

En,j,kO(xn,y)=(1PBn)λn\displaystyle E_{n,j,k}^{O}(x_{n},y)=(1-P_{\rm{B}n})\lambda_{n}
t=tjLd~jl=1tqjyfCτPr[tn,j,kW=l]Pr[t~n,j,kC=tl]pL(ttjL+1),\displaystyle\cdot\sum\limits_{t=t_{j}^{\rm{L}}}^{{{\tilde{d}}_{j}}}\!\!\sum\limits_{l=1}^{t-\lceil{\frac{q_{j}}{y{f^{\rm C}}\tau}}\rceil}\!\!\!\!\Pr[t_{n,j,k}^{\rm{W}}=l]\Pr[{{{\tilde{t}}_{n,j,k}^{\rm{C}}}=t-l}]\cdot{p^{\rm{L}}}({t-t_{j}^{\rm{L}}+1}), (17)

where tt is the number of time slots needed to complete the offloaded task, and (ttjL+1)({t-t_{j}^{\rm{L}}+1}) is the offloading and local execution overlap. Note that Pr[t~n,j,kC=tl]\Pr[{{{\tilde{t}}_{n,j,k}^{\rm{C}}}=t-l}] is a function of xnx_{n} and yy.

In case the task offloading goes beyond the finish of the local execution of a task, there is an extra power consumption incurred, whose expected value is

En,j,kB(xn,y)=(1PBn)λn\displaystyle E_{n,j,k}^{B}(x_{n},y)=(1-P_{\rm{B}n})\lambda_{n}
t=d~j+1+l=1tqjyfCτPr[tn,j,kW=l]Pr[t~n,j,kC=tl]pLLj.\displaystyle\cdot\sum\limits_{t={{\tilde{d}}_{j}}+1}^{+\infty}\!\!\!\!{\sum\limits_{l=1}^{t-\lceil{\frac{q_{j}}{y{f^{\rm C}}\tau}}\rceil}{\Pr[{t_{n,j,k}^{\rm{W}}=l}]\Pr[{{{\tilde{t}}_{n,j,k}^{\rm{C}}}=t-l}]}}{p^{\rm{L}}}{L_{j}}. (18)

Hence, the expected power consumption of MDs for offloaded tasks in BS nn in one second is

EnC(xn,y)=EnT(xn)\displaystyle E_{n}^{\rm{C}}(x_{n},y)=E_{n}^{T}(x_{n})
+j=1Jk=1InPjCPn,j,kG[En,j,kO(xn,y)+En,j,kB(xn,y)],\displaystyle+\sum\limits_{j=1}^{J}\sum\limits_{k=1}^{{I_{n}}}{P_{j}^{\rm{C}}}P^{\rm{G}}_{n,j,k}[E_{n,j,k}^{O}(x_{n},y)+E_{n,j,k}^{B}(x_{n},y)], (19)

and the expected power consumption of MDs for tasks arriving at BS nn in one second is EnL(xn)+EnC(xn,y)E_{n}^{\rm{L}}(x_{n})+E_{n}^{\rm{C}}(x_{n},y).

As before, our objective is to minimize the total expected power consumption of the MDs for uploading and executing the tasks that are generated in one second, but now subject to hard deadline constraints. The problem is formulated as follows:

min𝕩,yn=1N[EnL(xn)+EnC\displaystyle\mathop{\min}\limits_{\mathbb{x},y}\sum\limits_{n=1}^{N}[E_{n}^{\rm{L}}(x_{n})+E_{n}^{\rm{C}} (xn,y)] s.t.\displaystyle(x_{n},y)]\text{ s.t.} (20)
n=1Nαnxn+βfCy\displaystyle\sum\limits_{n=1}^{N}{{\alpha_{n}}{x_{n}}}+\beta{f^{\rm{C}}}y Bmax\displaystyle\leq{B^{\max}} (21)
(fC/j=1JPjCqj)y\displaystyle(f^{\rm{C}}/\sum\limits_{j=1}^{J}P_{j}^{\rm{C}}q_{j})y λ\displaystyle\geq\lambda (22)
xn\displaystyle x_{n} {0,1,,Kn},n𝒩\displaystyle\in\{0,1,\ldots,K_{n}\},\ \ \ \ \forall n\in\mathcal{N} (23)
0y\displaystyle 0\leq y 1.\displaystyle\leq 1. (24)

IV General Approximate Allocation Solutions

In this section, we propose approximate solutions for optimization problems (10)-(15) and (20)-(24), by decomposing them into convex optimization subproblems which can be solved efficiently.

IV-A Approximate Solution for Soft Deadlines

In this subsection, we propose an approximate solution for the optimization problem (10)-(15) by decomposing it into several convex subproblems that can be solved efficiently, solve them, and then keep the best solution. More specifically, we discretize variable y[0,1]y\in[0,1] by breaking [0,1][0,1] into YY equal segments, so that yy takes values ya=a/Yy_{a}=a/Y, for a=0,1,,Ya=0,1,\ldots,Y. With yy fixed, we show that the relaxation of (10)-(15) can be approximated by a convex optimization problem, which can be solved in polynomial time. The resulting (fractional) xnx_{n}’s are then rounded to integer values (and this is another source of suboptimality for our solution method). After solving the resulting Y+1Y+1 problems, we output the minimum solution x,yx^{*},y^{*}. Obviously, the quality of the approximation depends on the discretization parameter YY.

We consider the relaxed version of problem (10)-(15), i.e., constraint (14) has been replaced by xn0,nx_{n}\geq 0,\forall n. With yy fixed, we show that the non-convex problem (10)-(15) can be transformed into an equivalent convex optimization problem with the PBnP_{{\rm{B}}n}’s as the decision variables. First, we concentrate on constraints (12), (13). Note that Pr[tn,j,kW+tn,j,kCdj]{\rm{Pr}}[{t_{n,j,k}^{\rm{W}}+t_{n,j,k}^{\rm{C}}\leq{d_{j}}}] is a monotonically decreasing function of the aggregate mean task arrival rate λ\lambda. Hence, by binary search in the range [0,yfC/j=1JPjCqj][0,yf^{\rm{C}}/\sum_{j=1}^{J}P_{j}^{\rm{C}}q_{j}], we can approximate within any desired accuracy the maximum possible value of λ\lambda that satisfies constraints (12) for all n,j,kn,j,k. Let λ\lambda^{*} be this maximum value (note that λ<μC\lambda^{*}<\mu^{\rm{C}}, so stability is ensured). Using (4), constraints (12), (13) can be replaced by constraint

n=1N(1PBn)λnλ.\displaystyle\sum_{n=1}^{N}{({1-{P_{{\rm{B}}n}}}){{\lambda_{n}}}}\leq{\lambda^{*}}. (25)

Next, we note that the blocking probability PBnP_{{\rm{B}}n} is monotonically decreasing in xnx_{n}; let PBnminP_{{\rm{B}}n}^{\text{min}} be the blocking probability when xn=Knx_{n}=K_{n}. Then constraints (14) can be replaced by the equivalent constraints

PBnminPBn1,n𝒩.\displaystyle P_{{\rm{B}}n}^{\text{min}}\leq P_{{\rm{B}}n}\leq 1,\ \forall n\in\mathcal{N}. (26)

Constraint (11) is the only remaining constraint with an explicit dependence on the xnx_{n}’s. Since PBnP_{{\rm{B}}n} is a function of xnx_{n}, one could potentially use its inverse to replace xnx_{n} with a function of PBnP_{{\rm{B}}n}. However, such an inversion function may not exist explicitly (and even if it does, it may be non-convex). In its stead, we can use a convex upper bound approximation FF of the inversion of blocking probability, so that

xnF(PBn),n𝒩.\displaystyle x_{n}\leq F({P_{{\rm{B}}n}}),\ \forall n\in\mathcal{N}. (27)

Hence, the new convex optimization problem that approximates the original one when yy is fixed, is the following:

minBn=1N[EnL(PBn)+EnT\displaystyle\mathop{\min\ }\limits_{\mathbb{P}_{\rm{B}}}\sum\limits_{n=1}^{N}[E_{n}^{\rm{L}}(P_{\rm{B}n})+E_{n}^{\rm{T}} (PBn)] s.t.\displaystyle(P_{\rm{B}n})]\text{ s.t.} (28)
n=1NαnF(PBn)\displaystyle\sum\limits_{n=1}^{N}{{\alpha_{n}}F({P_{{\rm{B}}n}})} BmaxβfCy\displaystyle\leq{B^{\max}}-\beta f^{\rm{C}}y\ (29)
n=1N(1PBn)λn\displaystyle\sum\limits_{n=1}^{N}{({1-{P_{{\rm{B}}n}}}){{\lambda_{n}}}} λ\displaystyle\leq{\lambda^{*}} (30)
PBnminPBn\displaystyle P_{{\rm{B}}n}^{\text{min}}\leq P_{{\rm{B}}n} 1,\displaystyle\leq 1, n𝒩.\displaystyle\forall n\in\mathcal{N}. (31)

After solving (28)-(31) and obtaining the PBnP_{{\rm{B}}n}’s, we can compute the largest integral xnx^{*}_{n} which achieves a blocking probability equal to or bigger than PBnP_{{\rm{B}}n}, for all n𝒩n\in\mathcal{N}.

Algorithm 1 General Case Approximation for Soft Deadlines (GCASD)
0:  λn,pT,pL,αn,Kn,β,fC,Y,sj,dj,qj,PjC,Pn,j,kG\lambda_{n},p^{\rm{T}},p^{\rm{L}},\alpha_{n},K_{n},\beta,f^{\rm{C}},Y,s_{j},d_{j},q_{j},P_{j}^{\rm{C}},P^{\rm{G}}_{n,j,k}, PDFs of tW,tCt^{\rm{W}},t^{\rm{C}}
1:  cost=cost^{*}=\infty
2:  for all a=0,,Ya=0,\ldots,Y do
3:     y=a/Yy=a/Y
4:     Obtain λ\lambda^{*}, the upper bound of λ\lambda, by binary search in [0,μC][0,\mu^{\rm C}]
5:     [PB,cost]=[P_{\rm{B}},cost]= [solution, objective] of (28)-(31)
6:     xint=x_{int}= max integral xx with blocking probabilities PB\geq P_{\rm{B}}
7:     if cost<costcost<cost^{*} then
8:        x=xint;y=y;cost=costx^{*}=x_{int};y^{*}=y;cost^{*}=cost
9:     end if
10:  end for
11:  return  x,yx^{*},y^{*}

Complexity Analysis: Algorithm GCASD (cf. Algorithm 1) codifies the solution method described above. Problem (28)-(31) is convex, and can be solved in time O()O({\cal L}), for a polynomial \cal L. Line 6 takes time O(NlogKmax)O(N\log K_{max}) (recall that there are NN BSs, and KmaxK_{max} is the largest KnK_{n}). Hence Algorithm 1 has a running time of O(Y(+logμCϵ+NlogKmax))O(Y({\cal L}+\log\frac{\mu^{\rm C}}{\epsilon}+N\log K_{max})), where YY is the granularity of yy, and O(logμCϵ)O(\log\frac{\mu^{\rm C}}{\epsilon}) is the binary search cost of line 4 of the algorithm, in order to get a λ\lambda^{*} within ϵ\epsilon of the optimal.

IV-B Approximate Solution for Hard Deadlines

In this subsection, we use a similar approach in order to solve (20)-(24). Here we decompose the original problem into several subproblems by discretizing both variable yy as before, and λ\lambda. Then, for every possible (fixed) pair (y,λ)(y,\lambda), the non-convex problem (20)-(24) can be transformed into a convex optimization problem with PBnP_{{\rm{B}}n} as its decision variables, which can be solved in polynomial time. By calculating the pair (y,λ)(y^{*},\lambda^{*}) whose subproblem achieves minimum average power consumption, integer values xnx_{n}^{*} for the original optimization problem are obtained from PBnP_{{\rm{B}}n}^{*}.

In more detail, we discretize y[0,1]y\in[0,1] by breaking [0,1][0,1] into YY equal segments, and then we discretize λ[0,yfC/j=1JPjCqj]\lambda\in[0,yf^{\rm{C}}/\textstyle\sum_{j=1}^{J}P_{j}^{\rm{C}}q_{j}] by breaking interval [0,yfC/j=1JPjCqj][0,yf^{\rm{C}}/\textstyle\sum_{j=1}^{J}P_{j}^{\rm{C}}q_{j}] into Λ\Lambda equal segments. At iteration (m,i)(m,i) of this discretization, y=y(m)y=y^{(m)} and λ=λ(i)\lambda=\lambda^{(i)} are fixed. Then Pr[t~n,j,kC=tl]\Pr[{{{\tilde{t}}_{n,j,k}^{\rm{C}}}=t-l}] can be calculated directly for any tt and ll, and the original optimization problem (20)-(24) becomes

min𝕩n=1N[EnL(xn)+EnC\displaystyle\mathop{\min}\limits_{\mathbb{x}}\sum\limits_{n=1}^{N}[E_{n}^{\rm{L}}(x_{n})+E_{n}^{\rm{C}} (xn)] s.t.\displaystyle(x_{n})]\text{ s.t.} (32)
n=1Nαnxn\displaystyle\sum\limits_{n=1}^{N}{{\alpha_{n}}{x_{n}}} BmaxβfCy(m)\displaystyle\leq{B^{\max}}-\beta f^{C}y^{(m)}\ \ (33)
n=1N(1PBn)λn\displaystyle\sum\limits_{n=1}^{N}{({1-{P_{{\rm{B}}n}}}){{\lambda_{n}}}} λ(i)\displaystyle\leq\lambda^{(i)} (34)
xn\displaystyle x_{n} {0,1,,Kn},\displaystyle\in\{0,1,\ldots,K_{n}\}, n𝒩\displaystyle\forall n\in\mathcal{N} (35)

This is still a non-convex non-linear integer program, which cannot be solved efficiently. As in Section III, and by using (26)-(27), it becomes

minBn=1N[EnL(PBn)+\displaystyle\mathop{\min}\limits_{\mathbb{P}_{\rm{B}}}\sum\limits_{n=1}^{N}[E_{n}^{\rm{L}}(P_{\rm{B}n})+ EnC(PBn)] s.t.\displaystyle E_{n}^{\rm{C}}(P_{\rm{B}n})]\text{ s.t.} (36)
n=1NαnF(PBn)\displaystyle\sum\limits_{n=1}^{N}{{\alpha_{n}}F({P_{{\rm{B}}n}})} BmaxβfCy(m)\displaystyle\leq{B^{\max}}-\beta f^{C}y^{(m)}\ \ (37)
n=1N(1PBn)λn\displaystyle\sum\limits_{n=1}^{N}{({1-{P_{{\rm{B}}n}}}){{\lambda_{n}}}} λ(i)\displaystyle\leq\lambda^{(i)} (38)
PBnminPBn\displaystyle P_{{\rm{B}}n}^{\text{min}}\leq P_{{\rm{B}}n} 1,\displaystyle\leq 1, n𝒩.\displaystyle\forall n\in\mathcal{N}. (39)

Problem (36)-(39) is a convex program and can be solved efficiently. Hence, we can obtain the optimal blocking probabilities PBnP_{{\rm{B}}n}^{*}, corresponding to a pair (y(m),λ(i))(y^{(m)},\lambda^{(i)}). We can compute the largest integral xnx^{*}_{n} which achieves blocking probabilities no smaller than PBnP_{{\rm{B}}n}^{*}, for all n𝒩n\in\mathcal{N}, by using binary search based on the fact that the PBnP_{Bn}’s are decreasing functions of the xnx_{n}’s. After collecting the solutions for all iterations (m,i)(m,i), we output the minimum cost one 𝕩,y\mathbb{x}^{*},y^{*}.

Algorithm 2 General Case Approximation for Hard Deadlines (GCAHD)
0:  λn,pT,pL,αn,Kn,β,fC,Y,sj,dj,qj,Λ,PjC,Pn,j,kG\lambda_{n},p^{\rm{T}},p^{\rm{L}},\alpha_{n},K_{n},\beta,f^{\rm{C}},Y,s_{j},d_{j},q_{j},\Lambda,P_{j}^{\rm{C}},P^{\rm{G}}_{n,j,k}, PDFs of tW,tCt^{\rm{W}},t^{\rm{C}}
1:  cost=,y=0,λ=0cost^{*}=\infty,y=0,\lambda=0
2:  while y1y\leq 1 do
3:     while λyfC/j=1JPjCqj\lambda\leq yf^{\rm{C}}/\textstyle\sum_{j=1}^{J}P_{j}^{\rm{C}}q_{j} do
4:        [PB,cost]=[P_{\rm{B}},cost]= [solution, objective] of (36)-(39)
5:        xint=x_{int}= max integral xx with blocking probabilities PB\geq P_{\rm{B}}
6:        if cost<costcost<cost^{*} then
7:           x=xint;y=y;cost=costx^{*}=x_{int};y^{*}=y;cost^{*}=cost
8:        end if
9:        λ=λ+yfC/j=1JPjCqjΛ\lambda=\lambda+\frac{yf^{\rm{C}}/\textstyle\sum_{j=1}^{J}P_{j}^{\rm{C}}q_{j}}{\Lambda}
10:     end while
11:     y=y+1Yy=y+\frac{1}{Y}
12:  end while
13:  return  x,yx^{*},y^{*}

Complexity Analysis: Algorithm GCAHD (cf. Algorithm 2) codifies the solution method described above. Problem (36)-(39) is convex, and can be solved (line 4) in time O()O({\cal L}), for a polynomial \cal L. Line 5 takes time O(NlogKmax)O(N\log K_{max}) (recall that there are NN BSs, and KmaxK_{max} is the largest KnK_{n}). Hence Algorithm 2 has a running time of O(YΛ(+NlogKmax))O(Y\Lambda({\cal L}+N\log K_{max})), where YY and Λ\Lambda are the granularity of yy and λ\lambda respectively.

V Task Arrival and Offloading Assumptions

In the remainder of this paper, we assume that tasks arrive from the MDs at BS nn according to a Poisson process with mean arrival rate λn\lambda_{n}. The Poisson process assumption is commonly made in this type of situation, since the number of mobile devices in a given coverage area is typically quite large, each contributing to a small fraction of the total load [37]. In this case, we can invoke the insensitivity property of the Erlang B formula, to compute the probability of blocking at each BS [38]. Note that, typically, the Erlang B result is derived using the M/M/N/NM/M/N/N Markovian queue, which assumes exponentially distributed channel upload (i.e., service) times [39]. Due to insensitivity, the result holds for any service time distribution with the same mean. Therefore, the blocking probability for a task arriving at BS nn is

PBn=(λnμnW)xn1xn![r=0xn(λnμnW)r1r!]1{P_{{\rm{B}}n}}={\left({\frac{{{\lambda_{n}}}}{\mu_{n}^{\text{W}}}}\right)^{{x_{n}}}}\frac{1}{{{x_{n}}!}}{\left[{\sum\limits_{r=0}^{{x_{n}}}{{{\left({\frac{{{\lambda_{n}}}}{\mu_{n}^{\text{W}}}}\right)}^{r}}\frac{1}{{r!}}}}\right]^{-1}} (40)

where μnW\mu_{n}^{\text{W}} denotes the mean service rate, which can be calculated by μnW=1/t¯nW\mu_{n}^{\text{W}}=1/{{\bar{t}}_{n}^{\rm{W}}}. Function (40) is convex in xnx_{n} [40].

Note that due to the Poisson process task arrival assumption, the channel state sampled by arriving tasks is given by the steady-state equilibrium probability distribution of the Markovian channel at that MD. This follows from the PASTA rule [41].

We assume that the aggregate task arrival process at ES is Poisson [42], and, therefore, arriving tasks sample the asymptotic equilibrium state distribution of ES. This approximation is justified due to the mixing of arrivals at ES from BSs operating independently. In this case, ES can be modeled as an M/G/1M/G/1 queue, whose waiting time is given by the random variable wCw^{\rm{C}}. Given λ\lambda and knowledge of the data upload distribution, the distribution of wCw^{\rm{C}} can be obtained by numerical inversion of the probability generating function of system waiting time for M/G/1M/G/1 [37]. In this case, the execution time of a task at the ES depends only on which class it belongs to, i.e., tn,j,kC=tjCt_{n,j,k}^{\rm{C}}=t_{j}^{\rm{C}}, for all nn and kk, and tjC=wC+qj/yfCt_{j}^{\rm{C}}=w^{\rm{C}}+q_{j}/yf^{\rm{C}}. Thus, Pr[tn,j,kW+tjCdj]{\rm{Pr}}[{t_{n,j,k}^{\rm{W}}+t_{j}^{\rm{C}}\leq{d_{j}}}] can be easily obtained.

When applying algorithms GCASD (Algorithm 1) and GCAHD (Algorithm 2) in this case, the upper bound FF used in problem (28)-(31) and (36)-(39) becomes [43]:

xnλnμnW(1PBn)+1PBn,n.x_{n}\leq\frac{\lambda_{n}}{\mu_{n}^{\rm{W}}}(1-P_{{\rm{B}}n})+\frac{1}{P_{{\rm{B}}n}},\ \forall n. (41)

V-A Approximation with Soft Deadlines

In this case, problem (28)-(31) becomes:

minBn=1N[EnL(PBn)+EnT(PBn)]\displaystyle\mathop{\min\ }\limits_{\mathbb{P}_{\rm{B}}}\sum\limits_{n=1}^{N}[E_{n}^{\rm{L}}(P_{{\rm{B}}n})+E_{n}^{\rm{T}}(P_{{\rm{B}}n})] s.t. (42)
n=1Nαn(λnμnW(1PBn)+1PBn)\displaystyle\sum\limits_{n=1}^{N}{\alpha_{n}}({\frac{{{\lambda_{n}}}}{{{\mu_{n}^{\rm{W}}}}}({1-{P_{{\rm{B}}n}}})+\frac{1}{{{P_{{\rm{B}}n}}}}}) Bmax\displaystyle\leq{B^{\max}}- βfCy\displaystyle\beta{f^{\rm{C}}}y (43)
n=1N(1PBn)λn\displaystyle\sum\limits_{n=1}^{N}{(1-{P_{{\rm{B}}n}})}{\lambda_{n}} λ\displaystyle\leq{\lambda^{*}} (44)
PBnminPBn\displaystyle P_{{\rm{B}}n}^{\text{min}}\leq P_{{\rm{B}}n} 1,\displaystyle\leq 1, n𝒩.\displaystyle\forall n\in\mathcal{N}. (45)

Problem (42)-(45) is convex, and can be solved in polynomial time. Hence Algorithm 1 can be implemented efficiently.

V-B Approximation with Hard Deadlines

In this case, problem (36)-(39) becomes

minBn=1N[EnL(PBn)+EnC(PBn)]\displaystyle\mathop{\min}\limits_{\mathbb{P}_{\rm{B}}}\sum\limits_{n=1}^{N}[E_{n}^{\rm{L}}(P_{{\rm{B}}n})+E_{n}^{\rm{C}}(P_{{\rm{B}}n})] s.t. (46)
n=1Nαn(λnμnW(1PBn)+1PBn)\displaystyle\sum\limits_{n=1}^{N}{{\alpha_{n}}({\frac{{{\lambda_{n}}}}{{{\mu_{n}^{\rm{W}}}}}({1-{P_{{\rm{B}}n}}})+\frac{1}{{{P_{{\rm{B}}n}}}}})} Bmax\displaystyle\leq{B^{\max}}- βfCy(m)\displaystyle\beta f^{C}y^{(m)} (47)
n=1N(1PBn)λn\displaystyle\sum\limits_{n=1}^{N}{({1-{P_{{\rm{B}}n}}}){{\lambda_{n}}}} λ(i)\displaystyle\leq\lambda^{(i)} (48)
PBnminPBn\displaystyle P_{{\rm{B}}n}^{\text{min}}\leq P_{{\rm{B}}n} 1,\displaystyle\leq 1, n𝒩.\displaystyle\forall n\in\mathcal{N}. (49)

Problem (46)-(49) is convex, and can be solved efficiently.

VI Simulation Results

In this section, we present simulation results to demonstrate the performance of our proposed algorithms GCASD (Algorithm 1) and GCAHD (Algorithm 2). We adopt the two-state Gilbert-Elliot channel model [44], i.e., the channel states change by following a Markov chain with two states, “Good” (G) and “Bad” (B). This model is commonly used to characterize the effects of burst noise in wireless channels, where the channel can abruptly transition between good and bad conditions [45]. The Gilbert-Elliot channel is a difficult one for computation offloading algorithms to deal with compared to those where there is much more correlation in the channel quality as the offloading progresses. Let BgB_{g} and BbB_{b}, respectively, be the data transmission rate when the channel is in the G and B states. We consider that all channels have the same BgB_{g} and BbB_{b} values but differ in their state transition probabilities that result in different propagation models. The transition probabilities for propagation model kk in BS nn are denoted as Pn,kGG,Pn,kGB,Pn,kBG{{P_{n,k}^{\rm{GG}}}},{{P_{n,k}^{\rm{GB}}}},~{}{{P_{n,k}^{\rm{BG}}}}, and Pn,kBB{{P_{n,k}^{\rm{BB}}}}. In each time slot, the channel state Markov chain transitions in accordance with these probabilities. Denote πn,kG\pi_{n,k}^{\rm{G}} and πn,kB\pi_{n,k}^{\rm{B}}, respectively, as the stationary probabilities of a channel in BS nn for propagation model kk being in the G and B states. Two sets of simulations are performed with set 1 for single class of tasks and set 2 for multiple classes of tasks. Default parameters used in the simulations are summarized in Table III. The parameter settings that we use were taken from the references [26, 23] and [32]. These references summarize parameter settings for various types of applications including those that are inherently delay sensitive, such as gaming, face recognition and healthcare use. We intentionally use a wide range of parameter values based on the referenced ranges so that we can make conclusions that apply in general settings.

VI-A Simulation set 1: single class of tasks

In this subsection, we will assume that all the tasks generated at the MDs have the same data size ss and same computation load qq, i.e., sj=ss_{j}=s and qj=qq_{j}=q for all jj. When the channel is in the G state, the transmission rate of the wireless channel allows a task to be uploaded within one time slot; while when the channel is in the B state, the data transmission rate is zero. Since there is only one class of the tasks, subscript jj can be dropped from the notation.

Let tn,kWt_{n,k}^{\text{W}} be the time needed for uploading a task in BS nn with channel model kk. The probability that one task in BS nn with channel model kk can be uploaded in ll time slots is given as follows

Pr[tn,kW=l]={πn,kG,whenl=1πn,kBPn,kBBl2Pn,kBG,whenl2\Pr[{{t_{n,k}^{\rm{W}}}=l}]=\left\{{\begin{array}[]{ll}\pi_{n,k}^{\rm{G}},&\mbox{when}\ l=1\\ \pi_{n,k}^{\rm{B}}{P_{n,k}^{\rm{BB}}}^{l-2}{{P_{n,k}^{\rm{BG}}}},&\mbox{when}\ l\geq 2\end{array}}\right. (50)

The mean wireless transmission time of a task in BS nn uploaded through a channel with propagation model kk can be calculated as follows

t¯n,kW=l=1lPr[tn,kW=l]=1+Pn,kGBPn,kBG2+Pn,kGBPn,kBG.{{\bar{t}}_{n,k}^{\rm{W}}}=\sum\limits_{l=1}^{\infty}l\Pr[{{t_{n,k}^{\rm{W}}}=l}]=1+\frac{P_{n,k}^{\rm{GB}}}{{{P_{n,k}^{\rm{BG}}}^{2}+{P_{n,k}^{\rm{GB}}}{P_{n,k}^{\rm{BG}}}}}. (51)

Based on this, the mean wireless transmission time of the tasks in BS nn is t¯nW=k=1InPn,kGt¯n,kW{{\bar{t}}_{n}^{\rm{W}}}=\textstyle\sum_{k=1}^{{I_{n}}}P^{\rm{G}}_{n,k}{{\bar{t}}_{n,k}^{\rm{W}}}, where Pn,kGP^{\rm{G}}_{n,k} is the probability that a task in BS nn is uploaded through a channel with propagation model kk.

With a single class of tasks, the ES server becomes an M/D/1M/D/1 queueing system, tn,j,kC=tCt_{n,j,k}^{\rm{C}}=t^{\rm{C}} for all nn, jj and kk, and the distribution of delay is given by [46]

Pr[tCt^]=(1λμC)z=0t^μC[λ(zμCt^)]zz!eλ(zμCt^)\Pr[{{t^{\rm{C}}}\leq\hat{t}}]=\left({1-\frac{\lambda}{{{\mu^{\rm{C}}}}}}\right)\sum\limits_{z=0}^{\lfloor{\hat{t}{\mu^{\rm{C}}}}\rfloor}{\frac{{{{[{{\lambda}({\frac{z}{{{\mu^{\rm{C}}}}}-\hat{t}})}]}^{z}}}}{{z!}}}{e^{-\lambda({\frac{z}{{{\mu^{\rm{C}}}}}-\hat{t}})}} (52)

where μC=yfC/q\mu^{\rm{C}}={y{f^{\rm{C}}}}/{q}.

For comparison, we also run a discrete event simulation (DES) of the system using the xnx_{n}’s and yy solutions obtained from the proposed algorithms to validate our model assumptions, and these solutions are denoted as DESSD and DESHD, respectively, for the soft deadline (SD) and hard deadline (HD) cases. In addition, we simulate a DES-based OPT scheme for each proposed algorithm as follows. For GCASD, we first obtain all the possible combinations of xnx_{n}’s under constraint (14); for a given combination of xnx_{n}’s, we can obtain the solution of yy based on (11) and (15), and then check if constraint (13) is satisfied based on the current set of xnx_{n}’s and yy. If not, we go to the next set of xnx_{n}’s and repeat this procedure. If it is satisfied, we use this set of xnx_{n}’s and yy to run the DES for the system, and then check if (12) is satisfied. If not, we proceed to the next combination of xnx_{n}’s and repeat the above procedure. If the constraints are satisfied, we save the obtained average power. After going through all the possible combinations of xnx_{n}’s, we obtain the minimum average power and the corresponding xnx_{n}’s and yy. For GCAHD, we first obtain all the possible combinations of xnx_{n}’s under constraint (23); for a given combination of xnx_{n}’s, we can obtain the solution of yy based on (21) and (24), and then check if constraint (22) is satisfied based on the current set of xnx_{n}’s and yy. If not, we go to the next set of xnx_{n}’s and repeat this procedure. If it is satisfied, we use this set of xnx_{n}’s and yy to run the DES for the system. Then, we save the obtained mean power consumption. After going through all the possible combinations of xnx_{n}’s, we obtain the minimum average power and the corresponding xnx_{n}’s and yy.

TABLE III: Default Parameters
Parameter Value in set 1 Value in set 2
τ\tau 1 s
pLp^{\text{L}} 250 mW
pTp^{\text{T}} 2.5 mW
λn\lambda_{n} 11, 13, 15 tasks/s
KnK_{n} 15, 15, 20
αn\alpha_{n} 1, 1, 1 $
β\beta 0.3×1060.3\times 10^{-6} $ 0.25×1060.25\times 10^{-6} $
fCf^{\rm{C}} 75M cycles/s 200M cycles/s
ff 1M cycles/s 2M cycles/s
BmaxB^{\max} 140 $ 90 $
Bg,BbBg,B_{b} 2M, 0 bits per time slot 5M, 1M bits per time slot
sjs_{j} 2M bits 5M, 10M, 15M bits
djd_{j} 4 s 6, 11, 16 s
qjq_{j} 3M CPU cycles 10M, 20M, 30M CPU cycles
Refer to caption
(a) Soft deadlines
Refer to caption
(b) Hard deadlines
Figure 2: Average power consumption versus cost budget (Single class of tasks)
Refer to caption
(a) Soft deadlines
Refer to caption
(b) Hard deadlines
Figure 3: Average power consumption versus mean arrival rate (Single class of tasks)

In the simulation, we consider a cellular network consisting of 3 BSs. There are two propagation models at each BS with transition probabilities Pn,1GG=0.9{P_{n,1}^{\rm{GG}}}=0.9, Pn,2GG=0.7{P_{n,2}^{\rm{GG}}}=0.7, Pn,1BB=0.1{P_{n,1}^{\rm{BB}}}=0.1, and Pn,2BB=0.3{P_{n,2}^{\rm{BB}}}=0.3 for n=1,2,3n=1,2,3. The probabilities of the different channel models in BS 1 are P1,1G=0.8{P^{\rm{G}}_{1,1}}=0.8 and P1,2G=0.2{P^{\rm{G}}_{1,2}}=0.2; and those in BSs 2 and 3 are P2,1G=0.5{P^{\rm{G}}_{2,1}}=0.5, P2,2G=0.5{P^{\rm{G}}_{2,2}}=0.5, P3,1G=0.2{P^{\rm{G}}_{3,1}}=0.2, and P3,2G=0.8{P^{\rm{G}}_{3,2}}=0.8.

Refer to caption
(a) Soft deadlines
Refer to caption
(b) Hard deadlines
Figure 4: Average power consumption versus available ES capacity (Single class of tasks)

Figs. 2(a) and 2(b) show the average power consumption of MDs versus BmaxB^{\max} for the SD and HD cases, respectively. In Fig. 2(a), when the tolerable violation of latency ε\varepsilon is 1%, the average power consumption of MDs is a constant for all the solutions. This is because all the tasks are executed locally regardless of the cost budget, since the tight delay constraints cannot be satisfied if a task is offloaded. When ε\varepsilon is 3% or 5%, some tasks are allowed to be offloaded, and the average power consumption of the MDs decreases with BmaxB^{\max} for all the solutions. This happens since, when the cost budget is small, the optimization is constrained by the cost budget, which limits the number of offloaded tasks; and with the increase of BmaxB^{\max}, more channel and ES resource is available, leading to more MDs offloading their tasks. When BmaxB^{\max} is large, the budget constraint is loose, and the task offloading completion is mainly affected by the changing wireless transmission conditions. Fig. 2(a) also shows that the average MD power consumption decreases with ε\varepsilon for all the solutions, since larger ε\varepsilon makes it easier to meet the latency constraint through offloading, which results in more offloaded tasks and saves power in the MDs.

By comparing the average MD power consumption for ε=3%\varepsilon=3\% and ε=5%\varepsilon=5\% in Fig. 2(a), it is seen that the gap is small when the cost budget is small. The gap then increases as the cost budget increases, and finally becomes constant when the cost budget is sufficiently large. When the cost budget is low, the number of channels is small, which forces most tasks to be executed locally, regardless of the value of ε\varepsilon. As the cost budget increases, more channels are available, and the offloading decisions are determined by both ε\varepsilon and the available channel resources. When the cost budget is sufficiently high, the offloading decisions are mainly determined by the value of ε\varepsilon. The figure also shows that the average MD power consumption using GCASD is almost the same as using DESSD, which validates the model and approximations used in designing GCASD. The performance of GCASD is also close to DESSD-based OPT, which further shows good performance of the former.

By comparing Figs. 2(b) and 2(a), it can be seen that the average MD power consumption for the HD case is slightly higher than that for the SD case with ε=3%\varepsilon=3\% and much lower than that for the SD case with ε=1%\varepsilon=1\%. For the SD case, when ε=1%\varepsilon=1\%, the tight (soft) delay constraint forces all the tasks to be executed locally, which results in the highest average power consumption of the MDs; and the power consumption decreases as ε\varepsilon increases and more tasks are allowed to be offloaded. Without having to use CLE, the SD solutions result in lower average MD power consumption than the corresponding HD solutions. However, this is at a price that up to ε\varepsilon of the tasks do not meet their completion deadlines. On the other hand, using CLE in the GCAHD only incur slightly higher power consumption of the MDs compared to GCASD when ε=3%\varepsilon=3\% For the HD case, the total average power consumption of the MDs decreases with BmaxB^{\max} when BmaxB^{\max} is small and becomes a constant when BmaxB^{\max} becomes larger for all schemes, which is the same as that of the SD case with ε=3%\varepsilon=3\% and 5%5\%.

Figs. 3(a) and 3(b) show the average power consumption versus λn\lambda_{n} (same for all BSs) for the SD and HD cases, respectively. The figures show that the power consumption increases linearly with λn\lambda_{n} for all schemes, since both the local execution power and the uploading transmission power are proportional to the mean task arrival rate. The average MD power consumption using GCAHD is close to that using GCASD with ε=3%\varepsilon=3\% but much lower than that using GCASD with ε=1%\varepsilon=1\%. This demonstrates that the use of CLE in GCAHD is minimized, while always ensuring the HD of the tasks. Fig. 3(a) shows that the performance of GCASD is very close to DESSD and DESSD-based OPT; and Fig. 3(b) shows that the performance of GCAHD is very close to DESHD and DESHD-based OPT. These observations are consistent with the ones from Figs. 2(a) and 2(b). This further demonstrates the good performance of GCASD and GCAHD and validates the model and approximations used in designing the proposed algorithms.

Figs. 4(a) and 4(b) show the average power consumption of the MDs versus fCf^{\rm{C}}, which is the ES capacity, for the SD and HD cases, respectively. For the SD case with ε=1%\varepsilon=1\%, all tasks are executed locally; and when ε=3%\varepsilon=3\% and 5%5\%, offloading is possible for some tasks, and the number of tasks that can be offloaded increases with the ES capacity, resulting in lower power consumption of the MDs. As the ES capacity is sufficiently high, the average power consumption of MDs becomes a constant, since the offloading decisions are determined by the cost budget which limits the number of wireless channels for uploading tasks. Note that the slight increase in average power consumption when fCf^{\rm{C}} is between 60 and 80 is caused by the discretization errors of variable yy in algorithms 1 and 2. Increasing the YY values in the algorithms helps reduce the discretization errors but significantly increase the amount of time for running the simulations. Comparing the average power consumption of the HD and the SD cases shown in Figs. 4(a) and 4(b), we have consistent observations as in previous figures.

VI-B Simulation set 2: multiple classes of tasks

In this subsection, tasks have multiple classes. The two-state Gilbert-Elliot channels are considered. Let BgB_{g} and BbB_{b}, respectively, be the data transmission rates when a channel is in the G and B states. Given the channel state transision probabilities, the distribution of wireless transmission time tn,j,kWt^{\rm W}_{n,j,k} for uploading a class jj task in BS nn through a channel with propagation model kk can be calculated from [29].

At the ES, the system of serving the uploaded tasks becomes an M/G/1M/G/1 queueing system. Let BB be a random variable representing the execution time of the tasks. We have Pr[B=qjyfC]=PjC\Pr[B=\frac{q_{j}}{yf^{\rm C}}]=P^{\rm C}_{j}, then the probability density function of BB can be written as

fB(b~)\displaystyle f_{B}(\tilde{b}) =j=1JPr[B=qjyfC]δ(b~qjyfC)\displaystyle=\sum_{j=1}^{J}\Pr\left[B=\frac{q_{j}}{yf^{\rm C}}\right]\delta\left(\tilde{b}-\frac{q_{j}}{yf^{\rm C}}\right)
=j=1JPjCδ(b~qjyfC),\displaystyle=\sum_{j=1}^{J}P^{\rm C}_{j}\delta\left(\tilde{b}-\frac{q_{j}}{yf^{\rm C}}\right), (53)

and the Laplace-Stieltjes transform of fB(b~)f_{B}(\tilde{b}) is given by

g(s)=j=1JPjCeqjyfCs.g(s)=\sum_{j=1}^{J}{P^{\rm C}_{j}}e^{-\frac{q_{j}}{yf^{\rm C}}s}. (54)

The Laplace-Stieltjes transform of the probability density function of queuing time wCw^{\rm{C}} is given by the Pollaczek-Khinchine transform [37] as

W(s)=(1λb¯)ssλ(1g(s)),W^{*}(s)=\frac{(1-\lambda\bar{b})s}{s-\lambda(1-g(s))}, (55)

where b¯\bar{b} is the mean of BB. The distribution of wCw^{\rm{C}} can be obtained by numerical inversion of (55).

In the simulation, we consider a cellular network consisting of 3 BSs, 3 task classes, and 2 channel propagation models. The channel state transition probabilities are Pn,1GG=0.9{P_{n,1}^{\rm{GG}}}=0.9, Pn,1BB=0.1{P_{n,1}^{\rm{BB}}}=0.1, Pn,2GG=0.6{P_{n,2}^{\rm{GG}}}=0.6, and Pn,1BB=0.4{P_{n,1}^{\rm{BB}}}=0.4 for n=1,2,3n=1,2,3. The probabilities of accessing channels with different propagation models in BS 1 are P1,1G=0.8{P^{\rm{G}}_{1,1}}=0.8 and P1,2G=0.2{P^{\rm{G}}_{1,2}}=0.2; those in BSs 2 and 3 are P2,1G=0.5{P^{\rm{G}}_{2,1}}=0.5, P2,2G=0.5{P^{\rm{G}}_{2,2}}=0.5, P3,1G=0.2{P^{\rm{G}}_{3,1}}=0.2, and P3,2G=0.8{P^{\rm{G}}_{3,2}}=0.8. The probabilities of a task belonging to different classes are P1C=0.6P^{\rm C}_{1}=0.6, P2C=0.3P^{\rm C}_{2}=0.3, and P3C=0.1P^{\rm C}_{3}=0.1.

Refer to caption
(a) Soft deadlines
Refer to caption
(b) Hard deadlines
Figure 5: Average power consumption versus cost budget (Multiple classes of tasks)

Figs. 5(a) and 5(b) show the average power consumption of MDs versus BmaxB^{\max} for the SD and HD cases, respectively. In Fig. 5(a), when ε\varepsilon is 0.5%, all the tasks are executed locally regardless of the cost budget, since offloading cannot satisfy the tight delay constraints. When ε\varepsilon is 1% or 6%, the average power consumption of MDs decreases with BmaxB^{\max} and then becomes a constant. By comparing the power consumption of the MD in the SD and HD cases, we can see that the average power consumption of MDs for the HD case is slightly higher than that for the SD case with ε=1%\varepsilon=1\% and much lower than that for the SD case with ε=0.5%\varepsilon=0.5\%. Figs. 6(a) and 6(b) show the total average power consumption of the MDs versus fCf^{\rm{C}}. All the results show that our GCASD and GCAHD solutions achieve the average power consumption performance that is very close to DES-based OPT, and the observations in the multi-class simulations are consistent with the single-class simulations.

Refer to caption
(a) Soft deadlines
Refer to caption
(b) Hard deadlines
Figure 6: Average power consumption versus available ES capacity (Multiple classes of tasks)

VII Conclusions

This paper has studied joint wireless network and task service allocation for mobile computation offloading. The objective is to minimize the total average power consumption of MDs for completing the arriving tasks, while satisfying the delay constraints of tasks and the cost budget of the network customer. The formulations presented included both soft and hard task completion time deadlines. The designs were formulated as MINLPs and approximate solutions were obtained by decomposing the formulations into convex subproblems. Simulation results were presented that characterize the performance of the system and show various tradeoffs between task deadline violation, average mobile device power consumption and the cost budget. Results were presented that demonstrate the quality of the proposed solutions, which can achieve close-to-optimum performance over a wide range of system parameters. The optimum allocation were obtained by doing exhaustive search-based discrete event simulations for assigning the wireless channels from each BSs and ES capacity.

References

  • [1] T. H. Noor, S. Zeadally, A. Alfazi, and Q. Z. Sheng, “Mobile cloud computing: Challenges and future research directions,” Journal of Network and Computer Applications, vol. 115, pp. 70–85, 2018.
  • [2] Y. Kwon, H. Yi, D. Kwon, S. Yang, Y. Cho, and Y. Paek, “Precise execution offloading for applications with dynamic behavior in mobile cloud computing,” Pervasive and Mobile Computing, vol. 27, pp. 58–74, 2016.
  • [3] H. Ba, W. Heinzelman, C.-A. Janssen, and J. Shi, “Mobile computing-a green computing resource,” in 2013 IEEE Wireless Communications and Networking Conference (WCNC).   IEEE, 2013, pp. 4451–4456.
  • [4] Z. Gu, R. Takahashi, and Y. Fukazawa, “Real-time resources allocation framework for multi-task offloading in mobile cloud computing,” in 2019 International Conference on Computer, Information and Telecommunication Systems (CITS).   IEEE, 2019, pp. 1–5.
  • [5] J. Zhang, X. Hu, Z. Ning, E. C.-H. Ngai, L. Zhou, J. Wei, J. Cheng, and B. Hu, “Energy-latency tradeoff for energy-aware offloading in mobile edge computing networks,” IEEE Internet of Things Journal, vol. 5, no. 4, pp. 2633–2645, 2017.
  • [6] G. Huerta-Canepa and D. Lee, “A virtual cloud computing provider for mobile devices,” in Proceedings of the 1st ACM Workshop on Mobile Cloud Computing Services: Social Networks and Beyond, June 2010, p. 6.
  • [7] B.-G. Chun, S. Ihm, P. Maniatis, M. Naik, and A. Patti, “CloneCloud: Elastic Execution between Mobile Device and Cloud,” in Proceedings of the Sixth Conference on Computer Systems, ser. EuroSys ’11.   New York, NY, USA: ACM, 2011, pp. 301–314. [Online]. Available: http://doi.acm.org/10.1145/1966445.1966473
  • [8] Y. Shi, S. Chen, and X. Xu, “Maga: A mobility-aware computation offloading decision for distributed mobile cloud computing,” IEEE Internet of Things Journal, vol. 5, no. 1, pp. 164–174, Feb 2018.
  • [9] S. Zhou, Y. Sun, Z. Jiang, and Z. Niu, “Exploiting moving intelligence: Delay-optimized computation offloading in vehicular fog networks,” IEEE Communications Magazine, vol. 57, no. 5, pp. 49–55, May 2019.
  • [10] D. Mazza, D. Tarchi, and G. E. Corazza, “A unified urban mobile cloud computing offloading mechanism for smart cities,” IEEE Communications Magazine, vol. 55, no. 3, pp. 30–37, March 2017.
  • [11] H. A. Alameddine, S. Sharafeddine, S. Sebbah, S. Ayoubi, and C. Assi, “Dynamic task offloading and scheduling for low-latency iot services in multi-access edge computing,” IEEE Journal on Selected Areas in Communications, vol. 37, no. 3, pp. 668–682, 2019.
  • [12] J. Liu, J. Ren, Y. Zhang, X. Peng, Y. Zhang, and Y. Yang, “Efficient dependent task offloading for multiple applications in mec-cloud system,” IEEE Transactions on Mobile Computing, pp. 1–1, 2021.
  • [13] Huawei Inc., “5g network architecture - a high-level perspective,”
    https://www.huawei.com/en/technology-insights/industry-insights/outlook/mobile-broadband/insights-reports/5g-network-architecture, 2016.
  • [14] O. Muñoz, A. Pascual-Iserte, and J. Vidal, “Optimization of Radio and Computational Resources for Energy Efficiency in Latency-Constrained Application Offloading,” IEEE Transactions on Vehicular Technology, vol. 64, no. 10, pp. 4738–4755, October 2015.
  • [15] B. Dab, N. Aitsaadi, and R. Langar, “Joint optimization of offloading and resource allocation scheme for mobile edge computing,” in 2019 IEEE Wireless Communications and Networking Conference (WCNC), 2019, pp. 1–7.
  • [16] M. Sheng, Y. Wang, X. Wang, and J. Li, “Energy-efficient multiuser partial computation offloading with collaboration of terminals, radio access network, and edge server,” IEEE Transactions on Communications, vol. 68, no. 3, pp. 1524–1537, 2020.
  • [17] H. Chen, D. Zhao, Q. Chen, and R. Chai, “Joint computation offloading and radio resource allocations in small-cell wireless cellular networks,” IEEE Transactions on Green Communications and Networking, vol. 4, no. 3, pp. 745–758, 2020.
  • [18] J. Du, L. Zhao, J. Feng, and X. Chu, “Computation offloading and resource allocation in mixed fog/cloud computing systems with min-max fairness guarantee,” IEEE Transactions on Communications, vol. 66, no. 4, pp. 1594–1608, April 2018.
  • [19] X. Yang, X. Yu, H. Huang, and H. Zhu, “Energy efficiency based joint computation offloading and resource allocation in multi-access mec systems,” IEEE Access, vol. 7, pp. 117 054–117 062, 2019.
  • [20] J. Zhang, W. Xia, F. Yan, and L. Shen, “Joint computation offloading and resource allocation optimization in heterogeneous networks with mobile edge computing,” IEEE Access, vol. 6, pp. 19 324–19 337, 2018.
  • [21] X. Chen, Z. Liu, Y. Chen, and Z. Li, “Mobile edge computing based task offloading and resource allocation in 5g ultra-dense networks,” IEEE Access, vol. 7, pp. 184 172–184 182, 2019.
  • [22] S. Mu, Z. Zhong, and D. Zhao, “Energy-efficient and delay-fair mobile computation offloading,” IEEE Transactions on Vehicular Technology, vol. 69, no. 12, pp. 15 746–15 759, 2020.
  • [23] M. Masoudi and C. Cavdar, “Device vs edge computing for mobile services: Delay-aware decision making to minimize power consumption,” IEEE Transactions on Mobile Computing, vol. 20, no. 12, pp. 3324–3337, 2021.
  • [24] X. Chen, Y. Cai, Q. Shi, M. Zhao, B. Champagne, and L. Hanzo, “Efficient resource allocation for relay-assisted computation offloading in mobile-edge computing,” IEEE Internet of Things Journal, vol. 7, no. 3, pp. 2452–2468, 2020.
  • [25] S. Nath, Y. Li, J. Wu, and P. Fan, “Multi-user multi-channel computation offloading and resource allocation for mobile edge computing,” in ICC 2020 - 2020 IEEE International Conference on Communications (ICC), 2020, pp. 1–6.
  • [26] C. Yi, S. Huang, and J. Cai, “Joint resource allocation for device-to-device communication assisted fog computing,” IEEE Transactions on Mobile Computing, vol. 20, no. 3, pp. 1076–1091, 2021.
  • [27] H. Park, Y. Jin, J. Yoon, and Y. Yi, “On the economic effects of user-oriented delayed wi-fi offloading,” IEEE Transactions on Wireless Communications, vol. 15, no. 4, p. 2684–2697, 2016.
  • [28] L. Cominardi, T. Deiss, M. Filippou, V. Sciancalepore, F. Giust, and D. Sabella, “Mec support for network slicing: Status and limitations from a standardization viewpoint,” IEEE Communications Standards Magazine, vol. 4, no. 2, pp. 22–30, 2020.
  • [29] A. Hekmati, P. Teymoori, T. D. Todd, D. Zhao, and G. Karakostas, “Optimal mobile computation offloading with hard deadline constraints,” IEEE Transactions on Mobile Computing, vol. 19, no. 9, pp. 2160–2173, 2020.
  • [30] Y. Deng, Z. Chen, and X. Chen, “Resource allocation for multi-user mobile-edge computing systems with delay constraints,” in GLOBECOM 2020 - 2020 IEEE Global Communications Conference, 2020, pp. 1–6.
  • [31] C. W. Zaw, N. H. Tran, Z. Han, and C. S. Hong, “Radio and computing resource allocation in co-located edge computing: A generalized nash equilibrium model,” IEEE Transactions on Mobile Computing, pp. 1–1, 2021.
  • [32] S. Yue, J. Ren, N. Qiao, Y. Zhang, H. Jiang, Y. Zhang, and Y. Yang, “Todg: Distributed task offloading with delay guarantees for edge computing,” IEEE Transactions on Parallel and Distributed Systems, vol. 33, no. 7, pp. 1650–1665, 2022.
  • [33] Y. Geng, Y. Yang, and G. Cao, “Energy-efficient computation offloading for multicore-based mobile devices,” in IEEE INFOCOM 2018 - IEEE Conference on Computer Communications, 2018, pp. 46–54.
  • [34] M.-H. Chen, B. Liang, and M. Dong, “Multi-user multi-task offloading and resource allocation in mobile cloud systems,” IEEE Transactions on Wireless Communications, vol. 17, no. 10, pp. 6790–6805, 2018.
  • [35] Q. Li, S. Wang, A. Zhou, X. Ma, F. Yang, and A. X. Liu, “Qos driven task offloading with statistical guarantee in mobile edge computing,” IEEE Transactions on Mobile Computing, vol. 21, no. 1, pp. 278–290, 2022.
  • [36] J. Ren, G. Yu, Y. Cai, and Y. He, “Latency optimization for resource allocation in mobile-edge computation offloading,” IEEE Transactions on Wireless Communications, vol. 17, no. 8, pp. 5506–5519, 2018.
  • [37] A. Y. Khintchine, “Mathematical theory of a stationary queue,” Matematicheskii Sbornik, vol. 39, no. 4, pp. 73–84, 1932.
  • [38] D. Y. Burman, “Insensitivity in queueing systems,” Advances in Applied Probability, vol. 13, no. 4, pp. 846–859, 1981.
  • [39] D. J. Daley and L. D. Servi, “Idle and busy periods in stable m/m/k queues,” Journal of Applied Probability, vol. 35, no. 4, pp. 950–962, 1998.
  • [40] E. J. Messerli, “B.s.t.j. brief: Proof of a convexity property of the erlang b formula,” The Bell System Technical Journal, vol. 51, no. 4, pp. 951–953, 1972.
  • [41] R. W. Wolff, “Poisson arrivals see time averages,” Operations Research, vol. 30, no. 2, pp. 223–414, 1982.
  • [42] D. N. Shanbhag and D. G. Tambouratzis, “Erlang’s formula and some results on the departure process for a loss system,” Journal of Applied Probability, vol. 10, no. 1, pp. 233–240, 1973.
  • [43] S. A. Berezner, A. E. Krzesinski, and P. G. Taylor, “On the inverse of erlang’s function,” Journal of Applied Probability, vol. 35, no. 1, pp. 246–252, 1998.
  • [44] E. N. Gilbert, “Capacity of a burst-noise channel,” The Bell System Technical Journal, vol. 39, no. 5, pp. 1253–1265, 1960.
  • [45] T. Blazek and C. F. Mecklenbräuker, “Measurement-based burst-error performance modeling for cooperative intelligent transport systems,” IEEE Transactions on Intelligent Transportation Systems, no. 99, pp. 1–10, 2018.
  • [46] G. J. Franx, “A simple solution for the m/d/1 waiting time distribution,” Operations Research Letters, vol. 29, no. 5, pp. 221–229, 2001.