This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

A Non-Stationary Bandit-Learning Approach to Energy-Efficient Femto-Caching with Rateless-Coded Transmission

Setareh Maghsudi and Mihaela van der Schaar
S. Maghsudi is with the Electrical Engineering and Computer Science Department, Technical University of Berlin, 10623 Berlin, Germany (e-mail:[email protected]). M. van der Schaar is with the Faculty of Engineering, University of California at Los Angeles, United States and the Faculty of Engineering, University of Cambridge, United Kingdom (e-mail:[email protected]). A part of this paper appears at the 2019 IEEE Global Communications Conference [1].
Abstract

The ever-increasing demand for media streaming together with limited backhaul capacity renders developing efficient file-delivery methods imperative. One such method is femto-caching, which, despite its great potential, imposes several challenges such as efficient resource management. We study a resource allocation problem for joint caching and transmission in small cell networks, where the system operates in two consecutive phases: (i) cache placement, and (ii) joint file- and transmit power selection followed by broadcasting. We define the utility of every small base station in terms of the number of successful reconstructions per unit of transmission power. We then formulate the problem as to select a file from the cache together with a transmission power level for every broadcast round so that the accumulated utility over the horizon is maximized. The former problem boils down to a stochastic knapsack problem, and we cast the latter as a multi-armed bandit problem. We develop a solution to each problem and provide theoretical and numerical evaluations. In contrast to the state-of-the-art research, the proposed approach is especially suitable for networks with time-variant statistical properties. Moreover, it is applicable and operates well even when no initial information about the statistical characteristics of the random parameters such as file popularity and channel quality is available.

Keywords: Change detection, femto-caching, mortal and piece-wise stationary multi-armed bandits, rateless coding, stochastic knapsack problem.

I Introduction

To cope with the ever-increasing demand for mobile services, future wireless networks deploy dense small cells to underlay the legacy macro cellular networks [2], where small base stations (SBS) connect to the core network via a backhaul link. This concept takes advantage of low-power short-range base stations that potentially offload the macrocell traffic [3], improve the cell edge performance, and increase the uplink capacity. Moreover, transmission over short-length links enhances power efficiency and mitigates security concerns. However, these advantages come at some cost, and the system designers face several challenges to realize the concept of dense small cell networks. On the one hand, a massive number of devices sharing limited wireless resources in combination with a dense deployment of SBSs results in an excessive cost of information acquisition and computations, rendering centralized control mechanisms infeasible. On the other hand, in a dense small cell network, not all SBSs can access a power grid. Consequently, the SBSs store or harvest the required energy for providing wireless services. Naturally, the SBSs must consume energy efficiently to reduce the cost and the requirement of frequent batteries recharge, also to remain environment-friendly.

Besides, the significant growth in the demand for media streaming in combination with limited backhaul capacity renders developing efficient file-delivery approaches imperative [4]. The mobile data traffic caused by on-demand multimedia transmission exhibits the asynchronous content reuse property [5]; that is, the users request a few popular files at different times, but at relatively small intervals. The concept of wireless caching takes advantage of this property: instead of frequently fetching the popular files from the core network, different network entities store and re-transmit the files upon demand. The three major strategies for wireless caching are as follows: (i) femto-caching, (ii) D2D-caching, and (iii) coded-caching (or, coded-multicast) [6]. In femto-caching, SBSs save popular files to reduce the dependency on the high-speed backhaul. In D2D-caching, small devices store different files locally and provide each other with the files on-demand. Thus, in D2D-caching, the nearby devices share the burdens. The third method combines caching of files on the users’ devices with the common multicast transmission of network-coded data. Despite the great potential to gain resource efficiency while improving the users’ satisfaction levels, there are various challenges associated with caching. A challenging issue is the cache placement problem, which refers to selecting a subset of available files to store.

I-A Related Works

In the past few years, different methods and aspects of wireless caching have attracted a great deal of attention from the research community. In the following, we briefly review the cutting-edge research. In [6], the authors provide an overview of the basics as well as the state-of-the-art of D2D-caching and compare various methods from a practical point of view. In [7], Zhang et al. investigate the problem of joint scheduling and power allocation in D2D-caching. They decompose the problem into three sub-problems and solve each sub-problem in a centralized manner using convex optimization. D2D-caching is also the topic of [8]. Assuming that the knowledge of the network’s topology and files’ popularity is available, the authors use stochastic geometry to design communication protocols. Probabilistic cache placement for D2D-caching is discussed in [9]. Similar to [8], the authors assume that a central controller knows the network’s topology and files’ popularity. Based on this assumption, they propose a solution to the formulated problem. In [10], the authors consider the bandwidth allocation problem in proxy-caching. They propose a centralized solution based on auction theory that requires information about the files’ popularity. Similarly, [11] proposes a centralized heuristics for joint cache placement and server selection problem in proxy-caching. The developed solution necessitates knowledge of files’ popularity. Physical layer caching in wireless ad-hoc networks is the focus of [12]. The authors propose a caching scheme that potentially enhances the capacity of wireless ad-hoc networks by changing the underlying topology. They also analyze asymptotic scaling laws. Caching methods based on network coding (coded-multicast) are investigated in [13] and [14], as examples. The former considers femto-caching in conjunction with opportunistic network coding and derives the optimal cache placement. The latter introduces the concept of global caching gain against the conventional local caching gain. The authors suggest a coded-caching scheme that exploits both gains. The focus of this paper is on femto-caching; thus, in what follows, we confine our attention to research works that study this caching method.

Reference [15] introduces the femto-caching comprehensively. In [16], the authors formulate the cache placement as an integer programming and solve it by developing a centralized method. Assuming that the files’ popularity is known, [17] studies offline D2D- and femto-caching. In both scenarios, they investigate the joint design of the transmission and caching policies by formulating a continuous-time optimization problem. The topic of [18] is dynamic femto-caching for mobile users in a network with a time-varying topology. It proposes two algorithms for sub-optimal cache placement. One of the algorithms is centralized. The other one, despite being decentralized, requires knowledge of the popularity profile, current network topology, and the cache of other SBSs. Resource allocation for a small cell caching system is investigated in [19], by using stochastic geometry to perform the analysis. In [20], the authors investigate the trade-off between two crucial performance metrics of caching systems, namely, delivery delay and power consumption. They formulate a joint optimization problem involving power control, user association, and file placement. They solve the problem by decomposing it into two smaller problems. The focus of [21] is on energy efficiency in cache-enabled small cell networks. They discuss different factors such as interference level, backhaul capacity, content popularity, and cache capacity concerning their impact on energy efficiency. They then investigate the required conditions under which caching contributes to energy efficiency.

While a great majority of research studies assume that SBSs are provided with some prior knowledge, for instance, about the files’ popularity, there are only a few papers that do not rely on such assumptions when approaching different aspects of wireless caching. For example, [22] defines the offloading time as the time overhead incurred due to the unavailability of the file requested by a typical user. Then, for an unknown popularity profile, a learning approach is proposed to minimize the offloading time. As another example, in [23], the authors study the learning of a distributed caching scheme in small cell networks. Reference [24] develops a learning-based cache placement scheme in small cell networks, where the SBS observes only the number of requests for the cached files (i.e., the cache hits), and not the requests for the other files (i.e., cache misses). Using the features of the requested content, the SBS learns which files to cache to maximize the cache hits in the future. In [25], the authors use an online approach to learn the files’ popularity. Learning-based cache placement is also studied in [26] and [27]. Although the aforementioned papers deal with unknown popularity, the developed methods do not include the power consumption, the time-variations in the files’ popularity, and the possible gain of broadcast transmission. Table I summarizes the contribution of some research papers.

TABLE I: Wireless Femto-caching in Literature
Reference Problem Approach Information Concentration
[16] Cache placement Convex optimization Local CSI, Popularity Centralized
[28] Cache placement, Multicast Randomized rounding Global CSI, Popularity Centralized
[29] Cache placement, Overlapping helpers Heuristics Popularity Centralized
[17] Transmission and caching policy Optimization Popularity Centralized
[18] Cache placement Heuristics Popularity, Topology Centralized
[19] Resource allocation Stochastic geometry Popularity, Topology Distributed
[30] Energy efficiency Heuristics Popularity, Topology Centralized
Our Work Cache placement, Resource allocation MAB, Statistics None Distributed

I-B Our Contribution

A large body of cutting-edge research considers popularity as a known Zipf-like distribution. Such an assumption is, however, unrealistic due to the following reasons: (i) To a large extent, file popularity is dynamic and often vanishes over time; (ii) In general, the popularity of a file is an outcome of the decisions of a large population. Even if possible, such a variable is very costly to predict, especially due to massive production at a rapid pace.

Moreover, the great majority of current research assumes that SBSs have access to an unlimited power resource; nevertheless, in dense small cell networks, many SBSs rely on a limited power supply. The scarcity of supply renders the power efficiency crucial. Therefore, it is essential to serve as many users as possible with a specific amount of power. This goal is difficult to achieve particularly in dense small cell networks, where due to a large number of users, acquiring global channel state information (CSI) at SBSs is remarkably expensive. In addition to efficiency, sophisticated power control has a significant impact on mitigating transmission impairments such as interference; nonetheless, most previous works exclude such physical-layer issues from the caching problem, disregarding the fact that the transmission policy of SBSs impacts the realized benefits of caching. More precisely, such exclusion might prevent developing optimal caching strategies: On the one hand, transmission power impacts the coverage area of every SBS and the number of successful downloads of a cached file. On the other hand, due to the dense deployment, even SBSs close to each other share the spectrum resources, so that the inter-cell interference in every small cell depends on the transmission power of the neighboring SBSs.

We enhance the state-of-the-art as summarized in the following. Firstly, in comparison to the existing research, we employ a more general and realistic system model. In particular,

  • We take the limited file and power storage capacities of SBSs into account.

  • In our setting, the SBSs do not have any prior information about the statistical characteristics of the files’ popularity or channel quality.

  • We consider a dynamic network model where the number of users associated with every SBS is a random variable.

Every SBS uses rateless coding to broadcast a file. When exploiting rateless codes, the receiver accumulates mutual information to decode the message [31]. While the required time for decoding different blocks varies with channel conditions, rateless coding can guarantee zero outage probability, which makes it suitable for applications such as delay-constrained multimedia transmission over fading channels [32], [33]. We analyze the outage probability when rateless coding is applied. Afterward, we formulate the joint cache placement and power control problem in cache-enabled small cell networks. To deal with the NP-hardness of the formulated problem, we exploit the fact that caching systems usually work in two consecutive phases: (i) placement phase, and (ii) delivery phase [27], [13]. Therefore we decompose the challenge into two problems: (i) cache placement, and (ii) broadcast-file and -power selection, as described below:

  • To efficiently solve the online cache placement problem, we concentrate on the dynamics of file popularity, noting the fact that the popularity of every file is time-varying. We propose an algorithmic solution using a stochastic knapsack problem in combination with time-series analysis and statistical methods such as maximum likelihood estimation and change detection. We design the procedure and select the parameters carefully to achieve a good performance in terms of detecting variations in file popularity and fast reaction to changes.

  • In the delivery phase, to efficiently exploit the limited power supply at every SBS, we cast the selection of the optimal (file, transmit power) pair as a multi-armed bandit (MAB) problem with mortal arms. We adjust the algorithm and parameters so that it exhibits low regret, implying efficient performance in terms of quality of service (QoS) satisfaction and energy consumption.

We further establish the applicability of our approach through extensive numerical analysis.

I-C Paper Organization

In Section II, we introduce the network model. Afterward, in Section III, we design and analyze our rateless-coded transmission protocol. Section IV includes the description and analysis of the caching and files’ popularity models. In Section V, we formulate the joint cache placement and power control problem, which is then decomposed into two problems: (i) cache placement in the placement phase, and (ii) joint file- and transmit power selection in the delivery phase. In Section VI, we solve the former problem by developing an algorithm based on the stochastic knapsack problem and some statistical methods. In Section VII, we cast and solve the latter problem as a multi-armed bandit game. We dedicate Section VIII to the numerical analysis. Section IX concludes the paper.

I-D Notation

Throughout the paper, random variables are shown by uppercase letters while any realization is shown by lowercase. For example, XX and xx represent a random variable and its realization, respectively. The probability density or mass function (pdf) and the cumulative density function (cdf) of a random variable XX are shown by fX(x)f_{X}(x) and FX(x)F_{X}(x), respectively. By [x]\mathbb{P}\left[x\right] we denote the probability of some event xx. Moreover, 𝔼[X]\mathbb{E}\left[X\right] represents the expected value of XX. We show a set and its cardinality by a unique letter, and distinguish them by using calligraphic and italic fonts, such as 𝒜\mathcal{A} and AA, respectively. For any variable aa, a^\hat{a} represents its estimated value, which is calculated, for example, using maximum likelihood estimation. Matrices are shown by bold uppercase letters, for instance 𝐀\mathbf{A}; thus [𝐀]N×M\left[\mathbf{A}\right]_{N\times M} denotes a matrix with NN rows and MM columns. Moreover, 𝐀l\mathbf{A}_{l} denotes the ll-th row of matrix 𝐀\mathbf{A}. [𝐀]l,m[\mathbf{A}]_{l,m} stands for the element of matrix 𝐀\mathbf{A} located at ll-th row and mm-th column. Unit vectors are shown by bold lowercase letters, for example, a. The Hadamard product of two matrices 𝐀\mathbf{A} and 𝐁\mathbf{B} is shown as 𝐀𝐁\mathbf{A}{\circ}\mathbf{B}. By 𝐀=𝐁𝟏𝟐\mathbf{A}=\mathbf{B^{{}^{\circ\frac{1}{2}}}}, we mean that matrix 𝐀\mathbf{A} is the element-wise square root of matrix 𝐁\mathbf{B}. Table II summarizes the variables that appear frequently in this paper.

TABLE II: List of Frequently Used Variables (ordered as appear in the paper)
Symbol Description
\mathcal{M} Set of SBSs (Deterministically deployed)
dd Small cell radius
𝒫={pmin,,pmax}\mathcal{P}=\{p_{\min},...,p_{\max}\} Set of power levels for SBSs
pmp_{m} Transmit power of SBS mm
𝒩Poi(λ)\mathcal{N}\sim\textup{Poi}(\lambda) Set of users (Poisson-distributed)
HnmRay(12βnm)H_{nm}\sim\textup{Ray}(\frac{1}{\sqrt{2\beta_{nm}}}) Channel coefficient of link mnm\to n (Rayleigh-distributed)
dnmd_{nm} Distance between SBS mm and user nn
𝒳m𝒩\mathcal{X}_{m}\subseteq\mathcal{N} Set of users assigned to SBS mm (Distribution by (1))
\mathcal{F} Set of files to be potentially cached
LfL_{f} Number of data blocks of file ff
LfL^{\prime}_{f} Minimum required number of successfully-received packets to recover the file with probability δ\delta
δ\delta Probability of successful file recovery if at least LfL^{\prime}_{f} packets are successfully received
Dm,fD_{m,f} Maximum number of rateless-coded packets produced by SBS mm to broadcast file ff
RnmR_{nm} SINR of user nn assigned to SBS mm (Expression and distribution by (2) and (3))
OnmO_{nm} Outage probability of the link mnm\to n
un,minu_{n,\min} Minimum required transmission rate of user nn (Deterministic)
Pnm,f(s)P_{nm,f}^{(s)} Probability of successful recovery of file ff transmitted by SBS mm at user nn
Wnm,f,pW_{nm,f,p} Time required by user nn to recover file ff transmitted by SBS mm at power pp
W(Xm),f,pW_{(X_{m}),f,p} Maximum order statistic of XmX_{m} random variables Wnm,f,pW_{nm,f,p}, n𝒳mn\in\mathcal{X}_{m} (Distribution by (12))
Em,f,pE_{m,f,p} Total energy spent by SBS mm for broadcasting file ff at power pp
CmC_{m}; SfS_{f} Cache size of SBS mm; Size of file ff in bits
Qf,m,tQ_{f,m,t} Number of requests for file ff from SBS mm at time instant tt (Distribution by (19))
𝒜θ\mathcal{A}_{\theta}\subseteq\mathcal{F} Set of alive files at broadcast round θ\theta
fm,θf_{m,\theta}; pm,θp_{m,\theta} The broadcast file and transmit power of SBS mm at round θ\theta
Em,θ(f,p)E_{m,\theta}(f,p) Energy spent by SBS mm at round θ\theta
Km,θ(f,p)K_{m,\theta}(f,p) Number of users that recover the file successfully at broadcast round θ\theta
gm,θ(f,p)g_{m,\theta}(f,p) Utility of SBS mm at round θ\theta
Γjj\Gamma_{j-j^{\prime}} Log-likelihood ratio for observations from time jj up to time jj^{\prime}
tat_{a}; tct_{c} Alarm time, Change time
m,θ=m,θ𝒫\mathcal{I}_{m,\theta}=\mathcal{L}_{m,\theta}\otimes\mathcal{P} Action set of SBS mm at broadcast round θ\theta
im,θ=(fm,θ,pm,θ)i_{m,\theta}=(f_{m,\theta},p_{m,\theta}) Action of SBS mm at broadcast round θ\theta
im,θi_{m,\theta}^{*}; gm,θg_{m,\theta}^{*} Optimal action and Maximum achievable reward of SBS mm at broadcast round θ\theta
dm,θd_{m,\theta} Regret of SBS mm at broadcast round θ\theta
νi,θ\nu_{i,\theta} The index of arm ii at broadcast round θ\theta
Vi,θV_{i,\theta} Number of times arm ii is played up to round θ\theta
g¯i,θ\bar{g}_{i,\theta} Average reward of arm ii up to round θ\theta
qf,m,θ(T)q_{f,m,\theta}^{(T)} Number of requests for file ff from SBS mm at broadcast round θ\theta

II Network Model

We study the downlink operation in an underlay cache-enabled small cell network with a set \mathcal{M} of MM SBSs. The service area consists of multiple grids with a radius of dd. The deployment of SBSs is deterministic such that there exists only one SBS in every area with a radius of dd. All SBSs are connected to a macro base station (MBS), which has access to the core network. Each SBS mm\in\mathcal{M} selects its transmission power from a finite set of power-levels 𝒫\mathcal{P} including PP elements belonging to [pmin,pmax]\left[p_{\min},p_{\max}\right]. As it is conventional in underlay networks, SBSs use the MBS’s licensed channels if they are idle; thus, provided that the sensing is perfect, there is no interference incurred by- or to the MBS. The SBSs, however, might cause inter-cell interference to each other as they do not coordinate before transmissions.

There exists a random set 𝒩\mathcal{N} of NN users in the network. Users are quasi-static and their geographical arrangement follows a spatial homogeneous Poisson point process (HPPP) with density λ\lambda, which is unknown to the SBSs. We consider a frequency non-selective fading channel model. For a link connecting user n𝒩n\in\mathcal{N} and SBS mm\in\mathcal{M}, we denote the random channel coefficient by HnmH_{nm}. The channel coefficient follows a Rayleigh distribution with unknown parameter 12βnm\frac{1}{\sqrt{2\beta_{nm}}}. Therefore the channel gain |Hnm|2\left|H_{nm}\right|^{2} is exponentially-distributed with unknown parameter βnm\beta_{nm}. The MBS, the SBSs, and the users do not have any prior knowledge, even statistical, about the channel quality.111Rayleigh fading is the most reasonable and widely-applied statistical model for propagation in an urban environment. Therefore, we use the Rayleigh distribution to develop the mathematical formulation and to enable the analysis. Note that it does not contradict our claim of the absence of prior statistical information at SBSs since we do not allow knowledge about the statistical characteristics of the distribution such as the expected value. It is worth noting that the analysis can be performed for any other fading model by following the same lines. Besides, the proposed joint cache placement and resource allocation method is applicable even if there is no mathematical model and analysis. Throughout the paper, we regard interference as noise. For the simplicity of notation and calculus, we neglect the large-scale fading (path loss), assuming that there is an inner power-control loop that compensates the path loss. This simplification does not affect our analysis, and the entire analysis can be performed in the same way also including the path loss.

Let dnmd_{nm} denote the distance between SBS mm\in\mathcal{M} and user n𝒩n\in\mathcal{N}. Then user nn is associated with SBS mm if dnm=minimum{dn1,,dnM}d_{nm}=\textup{minimum}\{d_{n1},...,d_{nM}\}. In words, every user is assigned to the nearest SBS. The distance-based user association is a widely applied association method (see [34] and the references therein) since it is simple and practical. The set of users assigned to every SBS mm\in\mathcal{M} is denoted by 𝒳m\mathcal{X}_{m}. Recall that (i) users are distributed according to an HPPP with density λ\lambda, and (ii) there is only one SBS in the area of radius dd. Consequently, the number of users assigned to SBS mm\in\mathcal{M} follows the distribution

fXm(x)=(λπd2)xx!eλπd2.f_{X_{m}}(x)=\frac{\left(\lambda\pi d^{2}\right)^{x}}{x!}e^{-\lambda\pi d^{2}}. (1)

It is worth mentioning that in ultra-dense small cell networks, a user might be located inside the intersection of the coverage areas of multiple SBSs. Hence, if the SBSs transmit the same file with specific coding constraints, the user benefits from multiple useful signals for rapid file reconstruction. However, we consider a scenario where each user is associated with only one SBS and receives recoverable signals only from that SBS. Therefore, each SBS decides on the files to cache independent of others. Also, note that the channel impairments such as interference can sometimes yield a packet delivery failure for some of the users assigned to an SBS.

III Transmission Model

Let \mathcal{F} be the finite set of FF files to cache. Every file ff\in\mathcal{F} consists of LfL_{f} data blocks of the same size.222Depending on the applied standard for wireless transmission, the size of each data block might change from a few Kilobytes (e.g., 1.5 Kilobytes for IPv4) and dozens of Kilobytes (e.g., 64 Kilobytes for IPv6). The exact size of the data block does not play a role in our proposed caching and transmission scheme since, if required, a data block of any size can be further segmented to blocks with a different size (often smaller) that is appropriate for rateless coding and transmission. Naturally, the receiver then re-integrates the blocks to each other as necessary to rebuild the original one. All SBSs apply rateless codes to transmit any file ff\in\mathcal{F}, as described below.

Every SBS mm\in\mathcal{M} first encodes the message at the packet-level using Luby-Transform (LT) rateless code [35]. LT codes are the first practical realization of fountain codes that allow successful recovery of any message with very small overhead. In LT codes, every coded packet is the exclusive-or of aa neighbors. The number aa is called the degree. It is selected according to some specific probability distribution. After being appended with a cyclic redundancy check (CRC) sequence, each LT-coded packet is encoded by a physical-layer channel code of rate RR bits/channel-use and is transmitted over the channel in a time slot. The receiver attempts to decode the packets at the physical layer first. If the CRC is incorrect, the corresponding packet is discarded. We assume that CRC bits are sufficient so that every error can be detected.333The overhead of CRC depends on the type being used, or, in other words, the number of CRC checksum bits. The checksum usually consists of 8, 16, or 32 bits in length and is appended to the message. In general, taking both overhead and error detection performance into account, CRC is considered superior to many other error detection codes such as parity check. The accumulation of correctly received packets is used by the decoder to recover the source file. Assume that the receiver desires to successfully reconstruct the message with probability δ\delta. Then it requires to accumulate (at least) any Lf=Lf+νf(Lf,δ)L^{\prime}_{f}=L_{f}+\nu_{f}(L_{f},\delta) LT-coded packets, provided that the number aa (i.e., the degree) is selected appropriately [35]. Here, νf\nu_{f} is the discrete random decoding overhead, rendering LfL^{\prime}_{f} a random variable as well; nonetheless, for point to point transmission, it can be deterministic. A common value is νf=0.05Lf\nu_{f}=0.05L_{f} so that Lf=1.05LfL^{\prime}_{f}=1.05L_{f} [36]. The receiver then recovers the original message using, for instance, the belief propagation algorithm. It then informs the SBS via an acknowledgment (Ack) packet.444To decode the rateless-coded transmission, a receiver requires to possess channel state information. In our setting, each user receives signal only from one SBS, whereas the SBS transmits to multiple users; as such, it is plausible to assume that the receiver can obtain the CSI whereas the SBS incurs high cost for CSI acquisition. Also, some other articles such as [37] suggest utilizing rateless codes to overcome the channel uncertainty at the transmitter. Moreover, some researchers address the problem of decoding rateless-coded data under imperfect CSI at the receiver. For example, in [38], the authors discuss rateless codes for quasi-static fading channels in conjunction with imperfect CSI at the receiver. If the time and energy are unconstrained, rateless coding results in zero file-reconstruction failure also in fading channels. However, in a realistic scenario, either the energy is limited or the application is delay-sensitive. For example, in [33], the authors apply rateless codes to wireless streaming with a sequence of playback deadlines. We also take such restrictions into account by assuming the following: In every broadcast round, SBS mm\in\mathcal{M} generates at most Dm,fLfD_{m,f}\geq L^{\prime}_{f} rateless-coded packets to broadcast any file ff\in\mathcal{F}. Every SBS mm selects Dm,fD_{m,f} as a deterministic value. In essence, Dm,fD_{m,f} is a deadline for the users interested in file ff\in\mathcal{F} to accumulate LfL^{\prime}_{f} rateless-coded packets while SBS mm is broadcasting file ff. The users with a higher channel quality are more likely to succeed in accumulating LfL^{\prime}_{f} rateless-coded packets out of Dm,fD_{m,f} transmitted ones. Naturally, Dm,fD_{m,f} shall be larger for larger files that consist of more packets. By selecting Dm,fD_{m,f}, SBS mm balances the trade-off between the energy consumption for packet transmission on the one hand and the number of successful file reconstructions at the users’ side on the other hand. Thus, the SBS stops transmitting as soon as one of the following events occurs:

  • All assigned users that have requested file ff recover it.

  • Dm,fD_{m,f} data packets are transmitted.

Let user n𝒩n\in\mathcal{N} be assigned to some SBS mm\in\mathcal{M}, and the SBS transmits at some power level pm𝒫p_{m}\in\mathcal{P}. Then in the signal-to-interference-plus-noise ratio (SINR) yields

Rnm(pm)=pmhnm2i,impihni2+p0,R_{nm}(p_{m})=\frac{p_{m}h_{nm}^{2}}{\sum_{i\in\mathcal{M},i\neq m}p_{i}h^{2}_{ni}+p_{0}}, (2)

where p0p_{0} is the power of additive white Gaussian noise. Note that RnmR_{nm} attains a lower-bound in the worst-case scenario, where every SBS j,jmj\in\mathcal{M},j\neq m, transmits with the maximum power, i.e., pj=pmaxp_{j}=p_{\max}. The lower-bound, Rnm,minR_{nm,\min}, is useful when SBS mm is not aware of the transmit power of other SBSs. The pdf of Rnm(pm)R_{nm}(p_{m}), denoted by fR(r)f_{R}(r), is given by Proposition 1.

Proposition 1.

The pdf of SINR, RnmR_{nm}, is given by

fR(r)=\displaystyle f_{R}(r)= (3)
i,im\displaystyle\sum_{i\in\mathcal{M},i\neq m} βnmβniAniexp(βnmp0r)(βnmr+βni)p0+1(βnmr+βni)2,\displaystyle\beta^{\prime}_{nm}\beta^{\prime}_{ni}A_{ni}\exp(-\beta^{\prime}_{nm}p_{0}r)\frac{(\beta^{\prime}_{nm}r+\beta^{\prime}_{ni})p_{0}+1}{(\beta^{\prime}_{nm}r+\beta^{\prime}_{ni})^{2}},

where for all ii\in\mathcal{M} we define

βni=βnipi,\beta^{\prime}_{ni}=\frac{\beta_{ni}}{p_{i}}, (4)

and

Ani=l,li,mβnlβnlβni.A_{ni}=\prod_{l\in\mathcal{M},l\neq i,m}\frac{\beta^{\prime}_{nl}}{\beta^{\prime}_{nl}-\beta^{\prime}_{ni}}. (5)
Proof:

See Appendix X-A. ∎

For a sufficiently long block length, the outage probability is a good approximation for the packet error probability. Let un,minu_{n,\min} denote the minimum transmission rate required by user n𝒩n\in\mathcal{N}. The outage probability yields

Onm(p)=[log(1+Rnm(p))<un,min]=FRnm(eun,min1).O_{nm}(p)=\mathbb{P}\left[\log\left(1+R_{nm}(p)\right)<u_{n,\min}\right]=F_{R_{nm}}\left(e^{u_{n,\min}}-1\right). (6)

where FRnm(r)=0rfR(r)𝑑rF_{R_{nm}}(r)=\int_{0}^{r}f_{R}(r)dr. In the general case, the solution of this integral at every desired point can be calculated numerically. For special cases, one can also calculate the closed-form solution. For example, for the interference-limited region where the noise power is negligible compared to the interference power, by using (3), we can write (6) as

Onm(p)=i,imAniβniβnmrnrn+βniβnm,O_{nm}(p)=\sum_{i\in\mathcal{M},i\neq m}A_{ni}\frac{\beta^{\prime}_{ni}}{\beta^{\prime}_{nm}}\frac{r_{n}}{r_{n}+\frac{\beta^{\prime}_{ni}}{\beta^{\prime}_{nm}}}, (7)

where rn=eun,min1r_{n}=e^{u_{n,\min}}-1.555Depending on some factors such as the density of small base stations, small cell networks operate in different regions [39]. Naturally, the user experiences the maximum outage probability at Rnm,minR_{nm,\min}. A receiver can reconstruct a file ff\in\mathcal{F} with probability δ\delta if it successfully receives at least LfL^{\prime}_{f} rateless-coded packets out of Dm,fD_{m,f} packets transmitted by SBS mm. Hence, for a user n𝒩n\in\mathcal{N}, the probability of successful decoding (file reconstruction) yields

Pnm,f,p(s)=δ(1l=0Lf1(Dm,fl)(1Onm(p))lOnmDm,fl(p)).P_{nm,f,p}^{(s)}=\delta\left(1-\sum_{l=0}^{L^{\prime}_{f}-1}\binom{D_{m,f}}{l}(1-O_{nm}(p))^{l}O_{nm}^{D_{m,f}-l}(p)\right). (8)

The quality of the wireless channel is random. Therefore, for each user, the required time until receiving enough packets for file reconstruction is a random variable. Moreover, when broadcasting a file ff\in\mathcal{F}, each SBS mm\in\mathcal{M} transmits only a limited number of rateless-coded packets. Such restriction, in combination with random channel quality, renders the number of successful file reconstructions at the users’ side as random. Consequently, the total power required by an SBS to transmit a file is also a random variable. In the following sections, we analyze these random variables.

III-A Characterization of Energy Consumption

Let 𝒬f,m𝒳m\mathcal{Q}_{f,m}\subseteq\mathcal{X}_{m} be the set of users that are associated with SBS mm\in\mathcal{M} and are interested in receiving a file ff\in\mathcal{F}. Moreover, Wnm,f,pW_{nm,f,p} denotes the number of rateless-coded packets transmitted by SBS mm with power pp until some user n𝒬f,mn\in\mathcal{Q}_{f,m} accumulates LfL^{\prime}_{f} packets. Then Wnm,f,pW_{nm,f,p} is a random variable following a negative binomial distribution, with the pdf

fWnm,f,p(w)=(w1wLf)(1Onm(p))LfOnmwLf(p).\displaystyle f_{W_{nm,f,p}}(w)=\binom{w-1}{w-L^{\prime}_{f}}\left(1-O_{nm}(p)\right)^{L^{\prime}_{f}}O_{nm}^{w-L^{\prime}_{f}}(p). (9)

Thus the cdf of Wnm,f,pW_{nm,f,p} yields

FWnm,f,p(w)\displaystyle F_{W_{nm,f,p}}(w) =w=0w(w1wLf)(1Onm(p))LfOnmwLf(p)\displaystyle=\sum_{w=0}^{w}\binom{w-1}{w-L^{\prime}_{f}}\left(1-O_{nm}(p)\right)^{L^{\prime}_{f}}O_{nm}^{w-L^{\prime}_{f}}(p) (10)
=(1IOnm(p)(wLf+1,Lf)),\displaystyle=\left(1-I_{O_{nm}(p)}\left(w-L^{\prime}_{f}+1,L^{\prime}_{f}\right)\right),

where I{}I_{\{\cdot\}} is the regularized incomplete beta function. Assume that SBS mm\in\mathcal{M} continuously transmits rateless-coded packets until every user n𝒬f,mn\in\mathcal{Q}_{f,m} receives LfL^{\prime}_{f} packets successfully, i.e., there is no limit as Dm,fD_{m,f}. The number of rateless-coded packets required for that event is the maximum order statistic of the random variables Wnm,f,pW_{nm,f,p}, n𝒬f,mn\in\mathcal{Q}_{f,m}. For a deterministic value AA, the cdf of the maximum order statistic of AA independent and non-identically distributed (i.ni.d) random variables, denoted by WaW_{a}, a=1,,Aa=1,...,A, is given by [40]

FW(A)(w)=Γjs=1jFrs(w)s=j+1A(1Frs(w)),F_{W_{(A)}}(w)=\sum_{\Gamma_{j}}\prod_{s=1}^{j}F_{rs}(w)\prod_{s=j+1}^{A}\left(1-F_{rs}(w)\right), (11)

where the summation Γj\Gamma_{j} extends over all permutations (1,,rA)\left(1,...,r_{A}\right) of 1,,A1,...,A random variables for which 1<<rj1<\cdots<r_{j} and rj+1<<Ar_{j}+1<\cdots<A. In our formulation, however, Qf,mQ_{f,m} (corresponding to AA here) is itself a random variable following a distribution given by (1); thus at every broadcast round, the largest reconstruction time, denoted by W(Qf,m)W_{(Q_{f,m})}, has a pdf as

fW(Qf,m),f,p(w)=\displaystyle f_{W_{(Q_{f,m}),f,p}}(w)= (12)
q=0,1,[Qf,m=q]\displaystyle\sum_{q=0,1,...}\mathbb{P}\left[Q_{f,m}=q\right] (FW(),f,p(w)FW(),f,p(w1)|Qf,m=q),\displaystyle\left(F_{W_{(\cdot),f,p}}(w)-F_{W_{(\cdot),f,p}}(w-1)|Q_{f,m}=q\right),

where FW(Qf,m),f,p(w)F_{W_{(Q_{f,m}),f,p}}(w) is concluded by substituting (10) in (11).

In practice, time and energy are constrained. Thus, every SBS mm\in\mathcal{M} transmits at most Dm,fD_{m,f} rateless-coded packets of any file ff\in\mathcal{F}, unless all users n𝒬f,mn\in\mathcal{Q}_{f,m} send the Ack signal to the SBS mm before Dm,fD_{m,f}, implying that no further transmission is required. Using (12), we calculate the probability of this event as

[W(Qf,m),f,p<Dm,f]\displaystyle\mathbb{P}\left[W_{(Q_{f,m}),f,p}<D_{m,f}\right] =w=0Dm,f1fW(Qf,m),f,p(w)\displaystyle=\sum_{w=0}^{D_{m,f}-1}f_{W_{(Q_{f,m}),f,p}}(w) (13)
=FW(Qf,m),f,p(Dm,f1).\displaystyle=F_{W_{(Q_{f,m}),f,p}}(D_{m,f}-1).

Therefore, the duration of a broadcast round θ=1,2,\theta=1,2,... in which an SBS mm\in\mathcal{M} transmits a file ff\in\mathcal{F} with power p𝒫p\in\mathcal{P} is a random variable defined as

Tθ(m,f,p)\displaystyle T_{\theta}(m,f,p) (14)
={Dm,fwith probability1FW(Qf,m),f,p(Dm,f1)w<Dm,fwith probabilityfW(Qf,m),f,p(w).\displaystyle=\begin{cases}D_{m,f}&\text{with probability}~{}1-F_{W_{(Q_{f,m}),f,p}}(D_{m,f}-1)\\ w<D_{m,f}&\text{with probability}~{}f_{W_{(Q_{f,m}),f,p}}(w).\end{cases}

In (14), the first case corresponds to transmitting Dm,fD_{m,f} packets, whereas the second case occurs if the transmission ends sooner than that. By Tθ=jT_{\theta=j}, we denote the duration of some θ=j\theta=j. The total energy spent by SBS mm\in\mathcal{M} is a random variable defined as

Em,f,p=pTθ(m,f,p).E_{m,f,p}=pT_{\theta}(m,f,p). (15)

Note that the maximum total energy is given by

Em,f,p(max)=Dm,fp.E_{m,f,p}^{(\max)}=D_{m,f}p. (16)

III-B Characterization of the Number of File Reconstructions

Let 𝒦f,m,p𝒬f,m\mathcal{K}_{f,m,p}\subseteq\mathcal{Q}_{f,m} denote the set of interested users that recover the file successfully, i.e.,

𝒦f,m,p={x𝒬f,m|Wxm,f,pDm,f}.\mathcal{K}_{f,m,p}=\left\{x\in\mathcal{Q}_{f,m}|W_{xm,f,p}\leq D_{m,f}\right\}. (17)

Since the channel quality is random, Kf,m,pK_{f,m,p} is a random variable. As discussed in Section III, one of the following events occurs:

  • All of the Qf,mQ_{f,m} requesting users reconstruct the broadcast file ff by some time w<Dm,fw<D_{m,f}, or

  • A subset of users 𝒦f,m,p𝒬f,m\mathcal{K}_{f,m,p}\subseteq\mathcal{Q}_{f,m}, where Kf,m,pQf,mK_{f,m,p}\leq Q_{f,m}, reconstruct the file by the time Dm,fD_{m,f}.

The first event implies that the maximum order statistic of Qf,mQ_{f,m} random variables is equal to some w<Dm,fw<D_{m,f}, which happens with probability fW(Qf,m),f,p(w)f_{W_{(Q_{f,m}),f,p}}(w). In the second event, the Kf,m,pK_{f,m,p}-th order statistic of Qf,mQ_{f,m} random variables is equal to Dm,fD_{m,f}, which happens with probability fW(k),f,p(Dm,f)f_{W_{(k),f,p}}(D_{m,f}). Based on the argument above, Kf,m,pK_{f,m,p} can be characterized as

Kf,m,p={Qf,mwith probabilityFW(Qf,m),f,p(Dm,f)k<Qf,mwith probabilityfW(k),f,p(Dm,f)K_{f,m,p}=\begin{cases}Q_{f,m}&\text{with probability}~{}F_{W_{(Q_{f,m}),f,p}}(D_{m,f})\\ k<Q_{f,m}&\text{with probability}~{}f_{W_{(k),f,p}}(D_{m,f})\end{cases} (18)

IV File Popularity and Caching Model

Every SBS mm\in\mathcal{M} has a finite cache size, denoted by CmC_{m}. Moreover, every file ff\in\mathcal{F} has a finite size Sf=LfBS_{f}=L_{f}B. Files are labeled in a way that S1S2SFS_{1}\leq S_{2}\leq...\leq S_{F}. To gain access to a file that is not already stored in its cache, every SBS has to fetch it from the MBS. To avoid the energy-cost of fetching and to reduce the backhaul traffic, also to improve the users’ satisfaction level through reducing the delay, the SBSs try to avoid repetitive fetching, by storing the most popular files in their cache. The vast majority of previous works models the file popularity by a Zipf-like distribution with a time-invariant skewness parameter. However, such a model stands in contrast to the regular experience where the popularity of most data files changes over time. For example, as a familiar pattern, the number of requests for a video clip tends to increase initially, remain static for a while and then decrease. Therefore, in our work, we model the file popularity as a piece-wise stationary random variable. Details follow.

For an SBS mm\in\mathcal{M} that serves XmX_{m} users, the number of requests for a file ff\in\mathcal{F} follows a Poisson distribution with parameter μf,tXm>0\mu_{f,t}X_{m}>0; that is, the total number of requests for each file is time-variant and depends on the number of users in the small cell. Every user can request multiple files. The SBSs have no prior information about the popularity of each file.666Traditionally, Poisson distribution describes the random arrival process, e.g., user arrival or service demand in a telecommunication network. Therefore, here we use it to model the request arrival for each file, although we consider the general setting in which the parameter of this distribution, i.e., the intensity is unknown. Moreover, it should be emphasized that based on the intensity, Poisson distribution produces values in 𝒩0\mathcal{N}_{0}. Let Qf,m,tQ_{f,m,t} be the total number of requests for file ff\in\mathcal{F} submitted to SBS mm\in\mathcal{M} at time instant tt. Then

fQf,m,t(q)=\displaystyle f_{Q_{f,m,t}}(q)= x=0,1,[Xm=x][Qf,m,t=q|Xm=x]\displaystyle\sum_{x=0,1,...}\mathbb{P}\left[X_{m}=x\right]\mathbb{P}\left[Q_{f,m,t}=q|X_{m}=x\right] (19)
=\displaystyle= x=0,1,(λπd2)xx!(xμf,t)qq!eλπd2xμf,t,\displaystyle\sum_{x=0,1,...}\frac{\left(\lambda\pi d^{2}\right)^{x}}{x!}\frac{\left(x\mu_{f,t}\right)^{q}}{q!}e^{-\lambda\pi d^{2}-x\mu_{f,t}},

where the equality follows by (1) and the definition of Poisson distribution.

Assumption 1.

The average file popularity is piece-wise constant so that μf,t\mu_{f,t} is a step function that remains fixed over intervals of unknown length and suffers ruptures at change points. The values of μf,t\mu_{f,t} before and after change times, as well as the change points, are unknown a priori.

The time-variations in the files’ popularity correspond to the notion of concept drift in machine learning [41], which describes the changes in the statistical characteristics or distribution of the problem instances over time. For example, [42] considers this notion. In essence, big data applications are often characterized by concept drift, in which trending topics change over time. Moreover, [43] analyzes and illustrates some examples of concept drift in video streaming. Through data analysis, the authors show that the popularity of a video decreases by moving away from the posting date, or in other words, as the content’s age increases.

If ψ\psi and ψ\psi^{\prime} are intensities before and after the change, we have |ψψ|>C\left|\psi^{\prime}-\psi\right|>C, with CC being a sufficiently large constant. Note that the changes in the popularity can be continuous; however, by assuming a step-wise model, we neglect small changes in popularity and tune our algorithm to detect abrupt changes rather than small ones. In such a setting, the change detection performs better, since the tradeoff between a false alarm and misdetection can be balanced easier. Moreover, the complexity of caching reduces as well. More precisely, if even small changes are taken into account, the cache placement problem (knapsack problem) has to be solved unnecessarily very often, which results in excessive complexity. Based on this analytical model, in the following, we define the file type.

Definition 1 (File Type).

At any time instant tt, a file ff\in\mathcal{F} is alive (i.e., still popular) if μf,t>αm\mu_{f,t}>\alpha_{m}. Otherwise, it is dead; that is, it is not popular anymore.

We use 𝒜m,θ\mathcal{A}_{m,\theta}\subseteq\mathcal{F} to denote the set of files that are considered to be alive by SBS mm\in\mathcal{M}, at the beginning of a broadcast round θ\theta. We assume that there always exists at least one alive file, i.e., Am,θ1A_{m,\theta}\geq 1 for all θ=1,2,\theta=1,2,.... Naturally, as we later explain in Section VI, every SBS selects the set of files for storage from the set of alive files, meaning that m,θ𝒜m,θ\mathcal{L}_{m,\theta}\subseteq\mathcal{A}_{m,\theta}. Thus, αm\alpha_{m} is selected by SBS mm\in\mathcal{M} based on its cache size CmC_{m}. Intuitively, an SBS with larger storage capacity would have smaller αm\alpha_{m}, implying that it is able to store also some files with low popularity.

V Problem Formulation

Let fm,θ𝒜θf_{m,\theta}\in\mathcal{A}_{\theta} and pm,θ𝒫p_{m,\theta}\in\mathcal{P} denote the broadcast file and transmit power of SBS mm at broadcast round θ\theta. Then, Em,θ(fm,θ,pm,θ):=Em,fm,θ,pm,θE_{m,\theta}\left(f_{m,\theta},p_{m,\theta}\right):=E_{m,f_{m,\theta},p_{m,\theta}} is the total energy spent during the broadcast round θ\theta. Moreover, let Km,θ(fm,θ,pm,θ):=Km,fm,θ,pm,θK_{m,\theta}\left(f_{m,\theta},p_{m,\theta}\right):=K_{m,f_{m,\theta},p_{m,\theta}} be the number of successful file deliveries. We define the utility of SBS mm at round θ\theta as

gm,θ(fm,θ,pm,θ)=Km,θ(fm,θ,pm,θ)Em,θ(fm,θ,pm,θ),g_{m,\theta}(f_{m,\theta},p_{m,\theta})=\frac{K_{m,\theta}\left(f_{m,\theta},p_{m,\theta}\right)}{E_{m,\theta}\left(f_{m,\theta},p_{m,\theta}\right)}, (20)

i.e., in terms of the number of successful reconstructions per energy consumption. Ideally, every SBS performs the cache placement and power allocation jointly. That is, it selects a file to cache, and at the same time it assigns some power to broadcast that file (in future based on the users’ request). Then the optimization problem of every SBS mm\in\mathcal{M} can be formulated as follows:

maximizefm,θm,θ,pm,θ𝒫θ=1,2,gm,θ(fm,θ,pm,θ),\begin{matrix}\underset{f_{m,\theta}\in\mathcal{L}_{m,\theta}\subseteq\mathcal{F},p_{m,\theta}\in\mathcal{P}}{\textup{maximize}}&\sum_{\theta=1,2,...}g_{m,\theta}(f_{m,\theta},p_{m,\theta}),\end{matrix} (21)

subject to the following constraint

fm,θSfCm\sum_{f\in\mathcal{L}_{m,\theta}}S_{f}\leq C_{m} (22)

for all θ=1,2,\theta=1,2,.... Solving the optimization problem in (21) is however not feasible since by the following reasons the objective function is not available: (i) The files’ popularity, the network’s structure (including number of users), and the CSI are not known a priori; (ii) Each SBS is only aware of its own transmit power level as well as the file being broadcast, and not those of other SBSs. In such a scenario, a natural solution would be to incorporate online learning methods that maximize the expected utility, that is,

maximizefm,θm,θ,pm,θ𝒫𝔼[θ=1,2,gm,θ(fm,θ,pm,θ)],\begin{matrix}\underset{f_{m,\theta}\in\mathcal{L}_{m,\theta}\subseteq\mathcal{F},p_{m,\theta}\in\mathcal{P}}{\textup{maximize}}&\mathbb{E}\left[\sum_{\theta=1,2,...}g_{m,\theta}(f_{m,\theta},p_{m,\theta})\right],\end{matrix} (23)

subject to (22), where the expectation is taken with respect to random utility, as well as any other possible randomness in the decision-making strategy.

In general, efficient learning of the optimal decision involves finding a balance between gathering information about every possible strategy (exploration) on one hand and accumulating utility (exploitation) on the other hand. In other words, it is desired to learn as fast and efficient as possible, to minimize the cost of information shortage. In this sense, the problem formulated in (23) is inefficient to solve due to the following reasons: (i) Each (file, power level) pair, i.e., (f,p)(f,p), f,p𝒫f\in\mathcal{F},p\in\mathcal{P} is defined as an action (strategy), yielding FPFP actions in total; thus, from the computational point of view, such formulation imposes excessive complexity due to a large number of actions; (ii) To estimate the potential gain of each action, every file ff\in\mathcal{F} has to be fetched from the core network and transmitted at various power levels. It is evident that such exploration is very costly in terms of power, not only since fetching is power-consuming, but also since a substantial amount of limited power is wasted during exploration. Moreover, frequent fetching by densely deployed SBSs yields large backhaul traffic; (iii) The dynamic file popularity is not taken into account.

Based on the discussion above and in agreement with many previous works such as [27] and [13], we consider the caching system to work in two consecutive phases: (i) cache placement phase, and (ii) delivery phase. Accordingly, we decompose the cache placement and power control problem into two sub-problems listed below, which are solved repeatedly and in parallel to converge to the optimal solution:

  • Cache Placement: During every broadcast round θ=1,2,\theta=1,2,..., every SBS mm\in\mathcal{M} observes the requests for every file ff\in\mathcal{F}. Then, at the end of each broadcast round, the SBS updates its cache, if necessary. Later in Section VI, we explain that the necessity is defied based on the variations in the files’ popularity.

  • Joint Broadcast File Selection and Power Allocation: At every broadcast round θ=1,2,\theta=1,2,..., SBS mm\in\mathcal{M} selects a file fm,θm,θf_{m,\theta}\in\mathcal{L}_{m,\theta} to be transmitted, together with a power pm,θ𝒫p_{m,\theta}\in\mathcal{P} for broadcast.

A summary of cache placement and delivery protocol is provided in Algorithm 1. The approach is also illustrated in Fig. 1. The two problems are described in detail and are solved in Section VI and Section VII, respectively. In Section VII-A, we discuss the effect of decoupling the cache placement and power control problems on the performance.777Note that the frequency of updating the cache (if necessary at all) is not necessarily similar to that of broadcasting. In other words, an SBS can decide to update the cache in some specific intervals, for example, in every KK rounds of broadcasting. Our proposed algorithm is also applicable in this case.

Algorithm 1 Summary of Caching and Transmission Protocol
1:  for θ=1,2,\theta=1,2,... do
2:     As described in Algorithm 2 (Section VI), update the cache m,θ\mathcal{L}_{m,\theta} if there is a change in files’ popularity.
3:     Select a file fm,θm,θf_{m,\theta}\in\mathcal{L}_{m,\theta} together with a transmission power level pm,θ𝒫p_{m,\theta}\in\mathcal{P} for broadcast, as described in Algorithm 3 (Section VII).
4:  end for
Refer to caption
Figure 1: The work flow and the timing structure for joint cache placement, broadcast file selection, and power control.

VI Cache Placement

As described in Section V, we perform the cache placement based on file popularity, or in other words, the requests. By (23), at every broadcast round, a file is selected from the cache; this implies that the dynamic cache placement is performed, if necessary, at broadcast (delivery) rounds and not at every single transmission round. Let μf,θ\mu_{f,\theta} be the popularity of file ff\in\mathcal{F} at the beginning of the broadcast round θ\theta. For an SBS mm\in\mathcal{M}, the cache placement problem is stated as follows: Given the set of files \mathcal{F}, where each file ff\in\mathcal{F} has a size SfS_{f} and a value μf,θ\mu_{f,\theta}, desired is to decide whether to include the file in cache (m,θ\mathcal{L}_{m,\theta}) or not, so that the total size of selected files does not exceed the cache capacity CmC_{m}, while the total value is maximized. Note that at every cache updating round, the saved files are selected from the set of alive files. Such a strategy reduces the complexity without affecting the performance adversely. Formally, for θ=1,2,\theta=1,2,..., the optimization problem yields

maximizefm,θμf,θxfs.t.m,θ𝒜m,θ,fm,θSfxfCm,xf{0,1}.\begin{matrix}\textup{maximize}&\sum_{f\in\mathcal{L}_{m,\theta}}\mu_{f,\theta}x_{f}\\ \textup{s.t.}&\mathcal{L}_{m,\theta}\subseteq\mathcal{A}_{m,\theta},\\ &\sum_{f\in\mathcal{L}_{m,\theta}}S_{f}x_{f}\leq C_{m},\\ &x_{f}\in\{0,1\}.\\ \end{matrix} (24)

Considering each file as an item whose value is defined in terms of popularity, the cache placement problem is a 0-1 knapsack problem, which is a combinatorial optimization problem. Note that in formulating the knapsack problem, the value of every item (file) ff\in\mathcal{F} can be also defined in terms of the number of requests; Nonetheless, according to our model, the number of requests is stochastic, which renders the knapsack problem stochastic; Hence, to simplify the problem, we use the intensity of requests for each file ff, μf,θ\mu_{f,\theta}, as the item’s value, keeping in mind that μf,θ\mu_{f,\theta} is deterministic, despite changing over time.

Being under intensive investigation, a variety of efficient algorithmic solutions are developed for the knapsack problem [44]. The cache placement problem stated by (24) is however more challenging than the conventional knapsack problem. In particular, according to our system model (see Section IV), (i) the SBSs do not have any prior knowledge on files popularity indicator μf,θ\mu_{f,\theta}, and (ii) files’ popularity is time-varying so that (24) has to be solved sequentially. Against such difficulties, we notice that at every time t=1,2,t=1,2,..., every SBS mm\in\mathcal{M} receives the requests for every file ff\in\mathcal{F}. This side-information that every SBS obtains at almost no cost can be used to estimate the file popularity; i.e., to predict how many users would attempt to recover the file if it is broadcast. Let μ^f,θ1\hat{\mu}_{f,\theta-1} be the estimated value of μf,θ1\mu_{f,\theta-1} at the end of broadcast round θ1\theta-1.888It is worth noting that in some previous works, for example, in [27] and [24], the authors assume the following: Every SBS observes the cache hits, i.e., the number of requests for the files in the cache, whereas the number of cache drops, i.e., the number of requests for the files that do not belong to the cache, is not observable. However, such an assumption is unnecessary in many cases, where the users submit the requests for files to the corresponding SBS, while not knowing the cache status of the SBS. Even in the setting where the requests are submitted to some entity other than the SBS, such information can be transferred to the SBS with negligible cost. To update the cache for the broadcast round θ\theta, the SBS uses the historical data. Hence the optimization problem (24) is restated as

maximizefm,θμ^f,θ1xfs.t.m,θ𝒜^m,θ1,fm,θSfxfCm,xf{0,1}.\begin{matrix}\textup{maximize}&\sum_{f\in\mathcal{L}_{m,\theta}}\hat{\mu}_{f,\theta-1}x_{f}\\ \textup{s.t.}&\mathcal{L}_{m,\theta}\subseteq\hat{\mathcal{A}}_{m,\theta-1},\\ &\sum_{f\in\mathcal{L}_{m,\theta}}S_{f}x_{f}\leq C_{m},\\ &x_{f}\in\{0,1\}.\\ \end{matrix} (25)

For cache placement, we use the SBS’s observations in combination with some statistical tests, as described below.

VI-A Maximum Likelihood Estimation

Consider a time period θ\theta of length TT, with its time instances being labeled as 1,,T1,...,T. Let 𝐐f,m,1:T=(qf,m,1,,qf,m,T)\mathbf{Q}_{f,m,1:T}=\left(q_{f,m,1},...,q_{f,m,T}\right) be a set of samples of the popularity of file ff\in\mathcal{F}, where qf,m,tq_{f,m,t} shows the number of requests for file ff submitted to SBS mm at time j=1,,Tj=1,...,T. Assume that the popularity is stationary during the time interval [1,T]\left[1,T\right]. Then the maximum likelihood estimate of μf,θ\mu_{f,\theta} yields

μ^f,θ=1Tt=1Tqf,m,t.\hat{\mu}_{f,\theta}=\frac{1}{T}\sum_{t=1}^{T}q_{f,m,t}. (26)

VI-B Change Point Detection

As described before, for any file ff\in\mathcal{F}, the number of requests qf,m,tq_{f,m,t} is a random variable following a Poisson distribution with parameter μf,t\mu_{f,t}. Therefore, the sequence of requests qf,m,tq_{f,m,t}, t=1,2,t=1,2,..., are i.ni.d random variables. Assume that the value of the parameter is μf,t=ψ0\mu_{f,t}=\psi_{0} initially at some reference time tt, and at some unknown point of time, tct_{c}, its value changes to μf,tc=ψ1\mu_{f,t_{c}}=\psi_{1}. Change point detection basically detects the change and estimates the change time. Generally, it is assumed that ψ0\psi_{0} is known, while ψ1\psi_{1} might be known or unknown; in the latter case, the change magnitude |ψ1ψ0|\left|\psi_{1}-\psi_{0}\right| or basically ψ1\psi_{1} can be estimated as well. Formally, after each sampling, change point detection tests the following hypotheses about the parameter μ\mu:

H0:\displaystyle H_{0}: μ=ψ0\displaystyle\hskip 5.0pt\mu=\psi_{0} (27)
H1:\displaystyle H_{1}: {μψ0ifψ1is unknownμ=ψ1ifψ1is known|μψ0|>Cif changes are larger thanC\displaystyle\begin{cases}\hskip 5.0pt\mu\neq\psi_{0}&\text{if}\hskip 3.0pt\psi_{1}\hskip 3.0pt\text{is unknown}\\ \hskip 5.0pt\mu=\psi_{1}&\text{if}\hskip 3.0pt\psi_{1}\hskip 3.0pt\text{is known}\\ \hskip 5.0pt\left|\mu-\psi_{0}\right|>C&\text{if changes are larger than}\hskip 3.0ptC\end{cases}

In this paper, we utilize the generalized likelihood ratio (GLR) test for change detection [45], as described in the following.999In addition to the GLR test, there exist several other methods for change-point detection. This includes Page-Hinkley test (P-H) [46], cumulative sum control chart (CUSUM) [47], sequential probability ratio test (CPRT), and the like. In this paper, we opt to use the GLR test since it applies to a wide range of probability distributions for the random variable under analysis. Moreover, it is suitable when the parameter’s value before the change point as well as the change magnitude are unknown. For unknown ψ1\psi_{1}, the log-likelihood ratio for the observations from time jj up to time jj^{\prime} is given by

Γjj(ψ1)=t=jjln(f(qf,m,t;ψ1)f(qf,m,t;ψ0)),\Gamma_{j-j^{\prime}}(\psi_{1})=\sum_{t=j}^{j^{\prime}}\ln\left(\frac{f(q_{f,m,t};\psi_{1})}{f(q_{f,m,t};\psi_{0})}\right), (28)

where f(x;μ)f(x;\mu) is the probability of xx which depends on the scalar parameter μ\mu. For Poisson distribution, (28) yields

Γjj(ψ1)=t=jjln(eψ0ψ0qf,m,teψ1ψ1qf,m,t).\Gamma_{j-j^{\prime}}(\psi_{1})=\sum_{t=j}^{j^{\prime}}\ln\left(\frac{e^{\psi_{0}}\psi_{0}^{q_{f,m,t}}}{e^{\psi_{1}}\psi_{1}^{q_{f,m,t}}}\right). (29)

In (29), two unknown variables are present. The first one is the change time and the second one is the change magnitude (value of ψ1\psi_{1}). These values can be calculated by using the maximum likelihood estimation; that is, we perform the following double maximization [45]:

γj=max1jjsupψ1:|ψ1ψ0|C>0Γjj(ψ1).\begin{matrix}\gamma_{j^{\prime}}=\underset{1\leq j\leq j^{\prime}}{\max}&\underset{\psi_{1}:\left|\psi_{1}-\psi_{0}\right|\geq C>0}{\sup}~{}\Gamma_{j-j^{\prime}}(\psi_{1}).\end{matrix} (30)

Then the alarm time, tat_{a}, is defined as

ta=min{j:γjh},t_{a}=\min\left\{j^{\prime}:\gamma_{j^{\prime}}\geq h\right\}, (31)

where h>0h>0 is an appropriately-selected constant.101010The constant hh addresses the tradeoff between false alarm and detection delay, and should be determined so that a balance is achieved. Then the change point, tct_{c}, and the mean value after the change, ψ1\psi_{1}, are estimated as [45]

(t^c,ψ^1)=argmax1jtasupψ1:|ψ1ψ0|C>0Γjj(ψ1).(\hat{t}_{c},\hat{\psi}_{1})=\underset{1\leq j\leq t_{a}}{\arg\max}~{}\underset{\psi_{1}:\left|\psi_{1}-\psi_{0}\right|\geq C>0}{\sup}~{}\Gamma_{j-j^{\prime}}(\psi_{1}). (32)

VI-C The Algorithm

During the initialization period of length TT, the SBS gathers observations of the requests for each file ff\in\mathcal{F} as statistical samples. Using this sample set and the maximum likelihood estimation described in Section VI-A, the SBS estimates an initial popularity for every file ff\in\mathcal{F}, denoted by μ^f,θ=0\hat{\mu}_{f,\theta=0}. The SBS then solves the knapsack problem (25) to initialize the cache for θ=1\theta=1. In the course of each broadcast round θ=1,2,\theta=1,2,..., and at every transmission trial, the SBS observes the requests for every file, which serve as new samples of popularity.111111Note that similar to the Ack signals, requests can be submitted to each SBS also while transmitting, for instance through a control channel. The new samples are used to detect any changes in μ^f,t\hat{\mu}_{f,t}, and to estimate the change magnitude (i.e., the value of μ^f,t\hat{\mu}_{f,t} after the change), as explained in Section VI-B. At the end of each broadcast round θ\theta, the final value of μ^f,t\hat{\mu}_{f,t} represents μ^f,θ\hat{\mu}_{f,\theta}; Formally, μ^f,θ=μ^f,Tθ1\hat{\mu}_{f,\theta}=\hat{\mu}_{f,T_{\theta-1}}. Afterward, if necessary (i.e., upon the occurrence of some change), the SBS again solves the knapsack problem (25) and updates the cache. Algorithm 2 provides a summary of the procedure.

Algorithm 2 Cache Placement
1:  Initialization:
  • Select the length of initial observation TT.

  • For every file ff\in\mathcal{F}, the request vector 𝐐f,m,1:T=(qf,m,1,,qf,m,T)\mathbf{Q}_{f,m,1:T}=(q_{f,m,1},...,q_{f,m,T}) is initialized with zero.

  • Let the initial set of alive files 𝒜0=\mathcal{A}_{0}=\varnothing. Moreover, let the total run time so far be Tp=0T_{p}=0.

  • Let qf,m,θ(T)=0q_{f,m,\theta}^{(T)}=0 denote the total number of requests for each file ff\in\mathcal{F} during broadcast round θ\theta.

  • Select the threshold for the popularity index to determine alive files, i.e., αm\alpha_{m}.

2:  for t=1,,Tt=1,...,T do
3:     For every file ff\in\mathcal{F}, observe the number of requests at time tt, i.e., qf,m,tq_{f,m,t}.
4:     Update the total number of requests for every file ff\in\mathcal{F} as qf,m,0(T)=qf,m,0(T)+qf,m,tq_{f,m,0}^{(T)}=q_{f,m,0}^{(T)}+q_{f,m,t}.
5:  end for
6:  for ff\in\mathcal{F} do
7:     Calculate μ^f,θ=0\hat{\mu}_{f,\theta=0}, by using the sample set 𝐐f,m,1:T\mathbf{Q}_{f,m,1:T} and the maximum likelihood estimate (26).
8:     If μ^f,θ=0>αm\hat{\mu}_{f,\theta=0}>\alpha_{m}, include file ff in 𝒜^0\hat{\mathcal{A}}_{0}.
9:  end for
10:  Solve the knapsack problem (25) by using the estimated values of μ^f,θ=0\hat{\mu}_{f,\theta=0}, f𝒜^0f\in\hat{\mathcal{A}}_{0}, in order to initialize the cache (m,1\mathcal{L}_{m,1}).
11:  for broadcast rounds θ=1,2,3,\theta=1,2,3,... do
12:     Initialization:
  • Let TθT_{\theta} denote the length of broadcast round θ\theta (unknown a priori, see Algorithm 3).

  • For every file ff\in\mathcal{F}: (i) The request vector 𝐐f,m,1:Tθ\mathbf{Q}_{f,m,1:T_{\theta}} is initialized with zero vector; (ii) Let the change detection flag 𝕀c,f=0\mathbb{I}_{c,f}=0; (iii) Let the change time tc,f=0t_{c,f}=0.

  • Let the estimated set of alive files 𝒜^θ=\hat{\mathcal{A}}_{\theta}=\varnothing.

13:     for t=1+Tp,2+Tp,t=1+T_{p},2+T_{p},... do
14:        for ff\in\mathcal{F} do
15:           Observe the number of requests at time tt, qf,m,tq_{f,m,t}.
16:           Update the total number of requests for every file ff\in\mathcal{F} as qf,m,θ(T)=qf,m,θ(T)+qf,m,tq_{f,m,\theta}^{(T)}=q_{f,m,\theta}^{(T)}+q_{f,m,t}.
17:           Update the sample average as
q¯f,m,t=ttc,f1ttc,fq¯f,m,t1+1ttc,fqf,m,t.\bar{q}_{f,m,t}=\frac{t-t_{c,f}-1}{t-t_{c,f}}\bar{q}_{f,m,t-1}+\frac{1}{t-t_{c,f}}q_{f,m,t}. (33)
18:           Let ψ0=q¯f,m,t\psi_{0}=\bar{q}_{f,m,t} and perform the change point detection by using GLR test described by (30) and (31).
19:           if a change is detected then
20:              Find ψ^1\hat{\psi}_{1}, t^c\hat{t}_{c} using (32).
21:              Let μ^f,t=ψ^1\hat{\mu}_{f,t}=\hat{\psi}_{1} and tc,f=t^ct_{c,f}=\hat{t}_{c}.
22:              Let q¯f,m,t=μ^f,t\bar{q}_{f,m,t}=\hat{\mu}_{f,t} and 𝕀c,f=1\mathbb{I}_{c,f}=1.
23:           else
24:              Let μ^f,t=q¯f,m,t\hat{\mu}_{f,t}=\bar{q}_{f,m,t}.
25:           end if
26:        end for
27:        Tθ=Tθ+1T_{\theta}=T_{\theta}+1.
28:     end for
29:     For ff\in\mathcal{F}, let μ^f,θ=μ^f,t\hat{\mu}_{f,\theta}=\hat{\mu}_{f,t}.
30:     if f:𝕀c,f=1\exists\hskip 1.0ptf\in\mathcal{F}:\mathbb{I}_{c,f}=1 then
31:        For ff\in\mathcal{F}, if μ^f,t>αm\hat{\mu}_{f,t}>\alpha_{m}, include ff in 𝒜^θ\hat{\mathcal{A}}_{\theta}.
32:        Solve the knapsack problem (25) by using μ^f,θ\hat{\mu}_{f,\theta}, f𝒜^θf\in\hat{\mathcal{A}}_{\theta}, to update the cache (find m,θ+1\mathcal{L}_{m,\theta+1}).
33:     else
34:        m,θ+1=m,θ\mathcal{L}_{m,\theta+1}=\mathcal{L}_{m,\theta}.
35:     end if
36:     Let Tp=Tp+TθT_{p}=T_{p}+T_{\theta}.
37:  end for

It is worth mentioning that our proposed scheme to update the cache is event-triggered. More precisely, detecting a change in files’ popularity results in an attempt to re-optimize the cache, i.e., to solve the knapsack problem (25). Based on the magnitude of the change(s), the re-optimization might yield an urge to fetch some new files from the core network via the backhaul link. Although an SBS incurs some cost for fetching new files, for example, to reimburse the resulted backhaul traffic, the situation is inevitable when the popularity of files is time-variant; Nevertheless, every SBS (or an MBS controlling the backhaul link) can reduce the burden of fetching by adapting its responsiveness to the changes in files’ popularity. In other words, the SBS might decide to react only to a change whose magnitude is larger than some threshold value of CC. Naturally, larger values of CC result in less sensitivity and hence the cache is updated less frequently. In the limit case, i.e., for very large values of CC, our approach becomes similar to conventional caching schemes when the cache update is performed at specific time-intervals (time-organized rather than event-based). Finally, it should be noted that the knapsack problem (25) regards the size of every file as a cost factor; as fetching the large files naturally yields more backhaul traffic, one can conclude that the backhaul limitation is also taken into consideration in the optimization problem for cache placement.

VII Broadcast-File Selection and Power Control

Multi-armed bandit is a class of sequential optimization problems with incomplete information. The seminal problem involves an agent that is given a finite set of arms (actions, interchangeably), each producing some finite reward upon being pulled. Given no prior information, the agent selects arms sequentially, one at every round, and observes the reward of the played arm. Provided with this limited feedback, the agent aims at satisfying some optimality condition which is usually designed based on the problem’s specific characteristics, such as the random nature of the reward generating process, duration of arms’ availability, and so on. As a result of a lack of information, at each trial, the player may choose some inferior arm in terms of average reward, thus experiencing some regret. The regret is the difference between the reward that would have been achieved had the agent selected the best arm and the actually achieved reward.

In this paper, we model the joint file and power level selection as a multi-armed bandit problem, where every (file, power level) pair is regarded as one action. At every broadcast round θ=1,2,\theta=1,2,..., a file fm,θf_{m,\theta} is selected from the cache m,θ\mathcal{L}_{m,\theta}. The transmit power pm,θp_{m,\theta} is selected from the set of power levels 𝒫\mathcal{P}. Thus the agent has a set of actions m,θ=m,θ𝒫\mathcal{I}_{m,\theta}=\mathcal{L}_{m,\theta}\bigotimes\mathcal{P}, which includes Im,θ=Lm,θPI_{m,\theta}=L_{m,\theta}P elements. For simplicity, at every broadcast round θ\theta, we denote the action as im,θ=(fm,θ,pm,θ)i_{m,\theta}=\left(f_{m,\theta},p_{m,\theta}\right). The utility of this action is then given by (20). Ideally, every SBS desires to select the broadcast file and power level so as to maximize its expected accumulated utility through the entire horizon. Formally,

maximizeim,θm,θ𝔼[θ=1,2,gm,θ(im,θ)].\begin{matrix}\underset{i_{m,\theta}\in\mathcal{I}_{m,\theta}}{\textup{maximize}}&\mathbb{E}\left[\sum_{\theta=1,2,...}g_{m,\theta}\left(i_{m,\theta}\right)\right].\end{matrix} (34)

However, due to the lack of prior information and also by the time-varying nature of the problem, (34) is infeasible. As a result, every SBS would opt for a less ambitious goal, as we describe below.

At every broadcast round θ=1,2,\theta=1,2,..., we define the optimal action as

im,θ:=(fm,θ,pm,θ):=argmaxfm,θ,p𝒫gm,θ(f,p),\begin{matrix}i_{m,\theta}^{*}:=\left(f^{*}_{m,\theta},p^{*}_{m,\theta}\right):=\underset{f\in\mathcal{L}_{m,\theta},p\in\mathcal{P}}{\arg\max}&g_{m,\theta}(f,p),\end{matrix} (35)

which results in a reward

gm,θ(im,θ)=gm,θ(fm,θ,pm,θ):=maxfm,θ,p𝒫gm,θ(f,p).g^{*}_{m,\theta}\left(i^{*}_{m,\theta}\right)=g^{*}_{m,\theta}\left(f^{*}_{m,\theta},p^{*}_{m,\theta}\right):=\underset{f\in\mathcal{L}_{m,\theta},p\in\mathcal{P}}{\max}g_{m,\theta}(f,p). (36)

The regret of playing action im,θi_{m,\theta} at round θ\theta is given by

dm,θ=gm,θ(im,θ)gm,θ(im,θ).d_{m,\theta}=g^{*}_{m,\theta}\left(i^{*}_{m,\theta}\right)-g_{m,\theta}(i_{m,\theta}). (37)

Let 𝒮\mathcal{S} be the set of selection policies. Moreover, assume that SBS mm\in\mathcal{M} uses some policy σ𝒮\sigma\in\mathcal{S} to select some action im,θ(σ)i_{m,\theta}^{(\sigma)} at successive rounds θ=1,2,\theta=1,2,.... Also, let dm,θ(σ)d_{m,\theta}^{(\sigma)} denote the regret caused by using policy σ\sigma at round θ\theta. Then the goal of every SBS mm\in\mathcal{M} is to minimize its expected accumulated regret over the entire transmission horizon by using an appropriate decision-making strategy. Formally,

minimizeσ𝒮𝔼σ[θ=1,2,dm,θ(σ)(im,θ(σ))].\begin{matrix}\underset{\sigma\in\mathcal{S}}{\textup{minimize}}&\mathbb{E}_{\sigma}\left[\sum_{\theta=1,2,...}d_{m,\theta}^{(\sigma)}\left(i_{m,\theta}^{(\sigma)}\right)\right].\end{matrix} (38)

Recall that after some broadcast rounds, the cache is updated based on the new estimations of the time-varying files’ popularity. Consequently, the bandit’s action set might vary over the broadcast rounds as well, since the cache, together with the set of power levels, determines the set of available actions; In other words, while some action im,θ=(f,p)i_{m,\theta}=\left(f,p\right) might be included in the action set for some broadcast round θ\theta, there is a chance that it does not exist anymore in round θ>θ\theta^{\prime}>\theta, simply since the file ff is omitted from the cache as a result of losing its popularity. This type of bandit problem is referred to as multi-armed bandit with mortal arms, i.e., a setting in which some arms might disappear and some other arms might appear with time. To solve the selection problem, we use the algorithm summarized in Algorithm 3, which is an adapted version of a policy that appears in [48].

The algorithm is based on the upper-confidence bound policy (UCB) [49]. Basically, in UCB, at every selection round, an index is calculated for every action, which corresponds to an upper-bound of a confidence interval for the expected reward of that arm. At every round, the index of an arm depends on the number of rounds that specific arm is selected so far, together with its achieved rewards. Formally, assume that the action space is not time-varying and let ViV_{i} be the number of times that an action ii\in\mathcal{I} is played so far, i.e.,

Vi,θ=s=1θ𝕀{is(UCB)=i},V_{i,\theta}=\sum_{s=1}^{\theta}\mathbb{I}_{\left\{i_{s}^{(\textup{UCB})}=i\right\}}, (39)

where 𝕀{x=y}\mathbb{I}_{\left\{x=y\right\}} is the indicator function that returns one if the condition x=yx=y holds and zero otherwise. Then, the average reward from pulling any action iθi\in\mathcal{I}_{\theta} yields

g¯i,θ=1Vi,θs=1θgi,s𝕀{is(UCB)=i}.\bar{g}_{i,\theta}=\frac{1}{V_{i,\theta}}\sum_{s=1}^{\theta}g_{i,s}\mathbb{I}_{\left\{i_{s}^{(\textup{UCB})}=i\right\}}. (40)

The UCB policy associates an index νi,θ\nu_{i,\theta} to each action ii\in\mathcal{I} at round θ\theta [49]

νi,θ=g¯i,θ+βζlog(θ)Vi,θ,\nu_{i,\theta}=\bar{g}_{i,\theta}+\beta\sqrt{\frac{\zeta\log(\theta)}{V_{i,\theta}}}, (41)

with β\beta and ζ\zeta being appropriate constants. The second term on the right-hand side of (41) is a padding function that guarantees enough exploration. The action with the largest index is selected to be played in the next round, i.e.,

iθ+1=(fθ+1,pθ+1)=argmaxiθνi,θ.i_{\theta+1}=\left(f_{\theta+1},p_{\theta+1}\right)=\underset{i\in\mathcal{I}_{\theta}}{\arg\max}~{}\nu_{i,\theta}. (42)

The time-variation of the action set and the mortality of arms render the described UCB policy insufficient; Hence, based on [48], we introduce the following adaptation into the procedure:

  • If the cache is changed, play each new action (i.e., broadcast the new files with the available power levels);

  • Set Vi,θ=1V_{i,\theta}=1 for all actions (old and new) but keep the average reward of the old files;

  • Continue by the UCB procedure.

The following proposition states the regret bound of the decision-making policy described in Algorithm 3.

Proposition 2.

The expected regret RθR_{\theta} of Algorithm 3 is O(Blog(θ))O\left(B\log(\theta)\right), where BB is the number of times a new file has appeared in the cache during the broadcast rounds 1,,θ1,...,\theta.

Proof:

See Appendix X-B. ∎

Algorithm 3 Broadcast File Selection and Power Control
1:  Recall that qf,m,θ1(T)q_{f,m,\theta-1}^{(T)} denotes the total number of requests for each file ff\in\mathcal{F} during broadcast round θ1\theta-1.
2:  Recall that m,θ\mathcal{L}_{m,\theta} is the cache before round θ\theta starts, which is returned by Algorithm 2.
3:  for θ=1,2,3,\theta=1,2,3,... do
4:     Let Tθ=0T_{\theta}=0 denote the length of broadcast round θ\theta in terms of transmission trials (unknown a priori).
5:     for all im,θi\in\mathcal{I}_{m,\theta} do
6:        Calculate the file index, νi,m,θ1\nu_{i,m,\theta-1}, using (41).
7:     end for
8:     Let im,θ=(fm,θ,pm,θ)=argmaxim,θνi,m,θ1i_{m,\theta}=\left(f_{m,\theta},p_{m,\theta}\right)=\underset{i\in\mathcal{I}_{m,\theta}}{\arg\max}~{}\nu_{i,m,\theta-1}.
9:     repeat
10:        Send a rateless-coded packet.
11:        If any Ack signal is received, increase the number of successful file recoveries by one; i.e., let Km,θ=Km,θ+1K_{m,\theta}=K_{m,\theta}+1.
12:        Let Tθ=Tθ+1T_{\theta}=T_{\theta}+1.
13:     until Km,θ=qf,m,θ1(T)K_{m,\theta}=q_{f,m,\theta-1}^{(T)} (All interested users recover the file.) or Tθ=Dm,fm,θT_{\theta}=D_{m,f_{m,\theta}} (Maximum possible number of rateless-coded packets are transmitted.).
14:     Calculate the total energy spent during the current broadcast round by (15).
15:     Calculate the total gain of the selection using (20).
16:  end for

VII-A Discussion on Optimality and Complexity

As the final remark, we discuss the effect of decoupling the cache placement and power control. As described in Section V, to solve the joint cache placement and the pair (broadcast file, power level) selection, we decompose the cache placement and power control problems. A decoupling of the optimization of the cache placement and the delivery phases reduces the complexity significantly and boosts the power efficiency. Despite these benefits and although such an approach is used also by some previous works, such decomposition might raise some concerns with respect to the optimality of the solution in terms of utility. Naturally, approximation and probabilistic analysis introduce a possible reduction in optimality which cannot be avoided; however, the following proposition describes the conditions under which the decomposition does not affect the optimality.

Proposition 3.

A decoupling of the optimization of cache placement and delivery phases does not introduce any adverse effect on the optimality in terms of utility if

maxfm,θ,p𝒫𝔼[gm,θ(f,p)]maxfm,θ𝔼[Qf,m,θ]pminLf.\underset{f\in\mathcal{L}_{m,\theta},p\in\mathcal{P}}{\max}\mathbb{E}\left[g_{m,\theta}(f,p)\right]\geq\underset{f\in\mathcal{F}-\mathcal{L}_{m,\theta}}{\max}\frac{\mathbb{E}\left[Q_{f,m,\theta}\right]}{p_{\min}L^{\prime}_{f}}. (43)
Proof:

See Appendix X-C. ∎

Computationally, the proposed scheme consists of three main blocks: (i) The knapsack problem (25), which is an integer programming that accepts a fully polynomial-time approximation scheme in the number of alive files (AA) and the cache capacity CC [50]; (ii) The change detection, which is performed through maximum likelihood estimation that has a polynomial complexity in terms of the samples [51]; and (iii) The multi-armed bandit problem, whose solution involves calculating the arms’ indexes and finding the largest index at every round, which, similar to the other blocks, has linear complexity in the number of actions, i.e., (file, power) pairs [52].

VIII Numerical Results

We consider a small cell with one SBS where the users dynamically join and leave the network. At each time, the randomly-distributed users are drawn from a Poisson process with a density of λ=38\lambda=38. As an example, a snapshot is depicted in Fig. 2. The SBS transmits using a power selected from the set 𝒫={1,2,4}\mathcal{P}=\{1,2,4\}. The SBS has a cache capacity of C=15C=15. There are F=10F=10 files to be cached potentially. There are two change points for popularity, t=1.5×103t=1.5\times 10^{3} and t=3×103t=3\times 10^{3}. The files’ characteristics, including the size, popularity density before the change and popularity density after the change points, are summarized in Table III. As an example, for File A, File B and File I, i.e., two files whose popularity vanishes and one file whose popularity increases, the requests are shown in Fig. 3 as a function of time.

TABLE III: Files Characteristics
File’s Label A B C D E F G H I J
Size 1 1 2 5 6 3 5 4 3 7
Popularity Before Change 5 6 3 4 6 0.1 1 4 7 5
Popularity After First Change 5 0.1 3 4 6 0.1 1 4 7 5
Popularity After Second Change 0.1 0.1 3 4 6 0.1 1 4 12 5
Refer to caption
Figure 2: An exemplary small cell model. The values on both axes are normalized by 100 meters.
Refer to caption
Figure 3: Time-variations in the number of requests.

The set of alive files, as well as the cache, are shown in Fig. 4, for both before and after the popularity change. From the figure, we can conclude the following: After a change occurs, the proposed cache placement method detects the change; Consequently, it replaces the file with the lost popularity with some popular file. By doing this, a cache update is triggered to adapt the cache to the occurred change.

Refer to caption
Figure 4: Set of alive files and the cache, before and after the change points (popularity loss). In each sub-figure, the alive or selected files are marked by the blue line.

Fig. 5 shows the utility of the proposed approach (mortal bandits with change-point detection), compared to the following methods: (i) Optimal: Given the full statistical information, the best option is selected using the exhaustive search; (ii) Greedy: After some initialization rounds which take place in the round-robin manner, the best arm so far is selected for the rest of the horizon; (iii) Greedy with Fixed Exploration: At every round, with probability 1ϵ1-\epsilon the best arm so far, is selected and with probability ϵ\epsilon an arm uniformly at random. The value of ϵ\epsilon remains fixed over the entire horizon, which implies that the probability of exploration does not change; (iv) Greedy with Decreasing Exploration: This method is similar to the previous one, except that ϵ\epsilon tends to zero with rate 1/t1/t, meaning that the probability of exploration decreases in time. From the figure, it is clear that the proposed approach detects the changes with an acceptable delay. Moreover, it adapts the set of alive files and following that, the cache. The algorithm then finds the optimal (file, power level) pair so that the average utility is almost equal to that of the optimal solution. Although the optimal solution performs slightly better than bandit in terms of the average utility, it imposes excessive computational complexity and cost for information acquisition. Moreover, Fig. 6 shows the actions taken by the proposed algorithm (bandit) compared to those taken by the exhaustive search (optimal) approach. Most of the rounds in which the bandit algorithm does not select the optimal action belong to a short time interval right after the change, i.e., before the algorithm detects the change and adjusts the action.

Refer to caption
Figure 5: The convergence of the average utility (in terms of reward/cost with unit 1/Joule) to the optimal value.

Finally, we discuss the applicability and performance of rateless coding in delay-sensitive transmission, after SBS mm\in\mathcal{M} selects a file and a power level for transmission. Consider a large file with video content including SfS_{f} segments, where each segment consists of LfL_{f} data blocks. Due to the large size of the file ff, for video transmission, the video-player does not wait for the entire video, i.e., all the Sf×LfS_{f}\times L_{f} blocks to arrive before playback; i.e., each segment i=1,2,,Sfi=1,2,...,S_{f} has some playback deadline Di=i×Dm,f/LfD_{i}=i\times D_{m,f}/L_{f}. The blocks belonging to each segment are transmitted using rateless coding over the fading channel where we regard interference as noise. Therefore, for each segment i=1,2,,Sfi=1,2,...,S_{f} at least LfL^{\prime}_{f} packets shall be arrived by the deadline, otherwise an outage occurs. In [53] and [54], the authors show that the outage performance has a strong relationship with the SNR or SINR (or, the transmission power pp and the channel quality) as well as the playback deadline (or, indirectly Dm,fD_{m,f}). More precisely, a transition phase can be observed after which the outage of rateless transmission reduces abruptly to zero. Fig. 7 illustrates this effect, where we simulate the outage probability for delay-sensitive transmission of a video content with Sf=1000S_{f}=1000 segments through a fading channel as a function of SINR. Each segment consists of Lf=2L_{f}=2 data blocks. Selecting Dm,f=1000DD_{m,f}=1000D with D>1D>1, any segment i=1,,Sfi=1,...,S_{f} should be decoded correctly after at most Di=i×DD_{i}=i\times D channel uses (transmissions rounds), otherwise an outage occurs. From the figure, one can conclude that for different levels of delay-tolerance, different signal strengths are necessary to ensure zero outage.

Refer to caption
Figure 6: The frequency of selecting each action by the proposed method compared to the optimal scheme. The actions with very low selection frequency are not shown.
Refer to caption
Figure 7: The relation between SINR (the strength of desired and undesired signals), playback deadline (delay tolerance in seconds), and outage (QoS performance).

IX Summary and Conclusion

We considered a femto-caching scenario in a dynamic small cell network and studied a joint cache placement, broadcast file selection, and power control problem. We developed a solution for the formulated problem based on combinatorial optimization and multi-armed bandit theory. Theoretical and numerical results show that, in contrast to the conventional strategies, the proposed approach performs well under uncertainty, i.e., when the information about several variables such as files’ popularity and channel characteristics is not available a priori. Moreover, it successfully adapts to time-varying popularity. The exact level of performance improvement depends on several factors such as the speed of variations in the files’ popularity. Future research directions include an extension of the model and solution to the game-theoretical case where every user is served by multiple SBSs, implying that the cache placement and transmission can be done more efficiently.

X Appendix

X-A Proof of Proposition 1

Let user n𝒩n\in\mathcal{N} be associated to SBS mm\in\mathcal{M}. There is one desired signal, say XX, which is exponentially-distributed with parameter βnm=βnm/pm\beta^{\prime}_{nm}=\beta_{nm}/p_{m}. Moreover, there are M1M-1 interference signals ii, where i{m}i\in\mathcal{M}-\{m\}. Hence the total interference power yields Y=i,imMYniY=\sum_{i\in\mathcal{M},i\neq m}^{M}Y_{ni}. Since YniY_{ni} is an exponentially-distributed random variable with parameter βni=βni/pi\beta^{\prime}_{ni}=\beta_{ni}/p_{i}, YY is the sum of i.ni.d exponential random variables. The pdf of YY is given by [55]

fY(y)=i,imMAniβniexp(βniy),f_{Y}(y)=\sum_{i\in\mathcal{M},i\neq m}^{M}A_{ni}\beta^{\prime}_{ni}\exp(-\beta^{\prime}_{ni}y), (44)

with AniA_{ni} given by (5). Let Z=Y+p0Z=Y+p_{0}. Then by using the relation F(Zz)=F(Yzp0)F(Z\leq z)=F(Y\leq z-p_{0}), the pdf of ZZ can be calculated as

fZ(z)\displaystyle f_{Z}(z) =ddzy=0zp0fY(y)𝑑y\displaystyle=\frac{d}{dz}\int_{y=0}^{z-p_{0}}f_{Y}(y)dy (45)
=i,imMAniβniexp(βni(zp0))𝕀{zp0}.\displaystyle=\sum_{i\in\mathcal{M},i\neq m}^{M}A_{ni}\beta^{\prime}_{ni}\exp(-\beta^{\prime}_{ni}(z-p_{0}))\mathbb{I}_{\left\{z\geq p_{0}\right\}.}

For R=XZR=\frac{X}{Z}, we have fR(r)=0zfX(rz)fZ(z)𝑑zf_{R}(r)=\int_{0}^{\infty}zf_{X}(rz)f_{Z}(z)dz. Therefore the proposition follows by (44).

X-B Proof of Proposition 2

We follow the same line as in [48]. As long as the cache does not change, the algorithm behaves like the standard UCB policy, which has a regret equal to O(log(θ))O\left(\log(\theta)\right). Removing a file from the cache does not affect this regret bound. Including a new file in the cache is equivalent to introducing PP new actions, since each action is defined as a (file, power level) pair. However, all of these PP new actions remain in the action set as long as the file ff remains in the cache, and disappear afterward. For θ\theta broadcast rounds, the maximum number of rounds a file remains in the cache is θ\theta; In other words, if a file remains in the cache for some duration θθ\theta^{\prime}\leq\theta, the incurred regret is O(log(θ))O(log(θ))O\left(\log(\theta^{\prime})\right)\leq O\left(\log(\theta)\right). Therefore, the regret is bounded as O(Blog(θ))O\left(B\log(\theta)\right).

X-C Proof of Proposition 3

By decoupling the cache placement and delivery phases, the set of files \mathcal{F} is divided into two parts: (i) The files that are stored in the cache, i.e., all fm,θf\in\mathcal{L}_{m,\theta}; and (ii) The files that are not stored in the cache, i.e., all fm,θf\in\mathcal{F}-\mathcal{L}_{m,\theta}. Assuming that the distance between the estimated file’s popularity μ^f\hat{\mu}_{f} and the true file popularity μf\mu_{f} tends to zero, solving the knapsack problem (25) returns the optimal cache placement. Also, by Proposition 2, the bandit strategy finds the most efficient pair (broadcast file, power level) that maximizes the expected utility in terms of (20). Therefore, a sub-optimality occurs only if one of the files that are not included in the cache produces an expected utility larger than that of the pair found by the MAB policy. Formally, to guarantee the optimality, one needs to have

maxfm,θ,p𝒫𝔼[gm,θ(f,p)]𝔼[gm,θ(f,p)]\underset{f\in\mathcal{L}_{m,\theta},p\in\mathcal{P}}{\max}\mathbb{E}\left[g_{m,\theta}(f,p)\right]\geq\mathbb{E}\left[g_{m,\theta}(f,p)\right] (46)

for all fm,θf\in\mathcal{F}-\mathcal{L}_{m,\theta} and p𝒫p\in\mathcal{P}. The right-hand side of (46) can be upper-bounded as

𝔼[gm,θ(f,p)],fm,θ,p𝒫\displaystyle\begin{matrix}\mathbb{E}\left[g_{m,\theta}(f,p)\right],\forall f\in\mathcal{F}-\mathcal{L}_{m,\theta},p\in\mathcal{P}\end{matrix} (47)
maxfm,θ,p𝒫𝔼[gm,θ(f,p)]𝔼[Km,θ(fθ,pθ)]𝔼[Em,θ(fθ,pθ)]\displaystyle\geq\underset{f\in\mathcal{F}-\mathcal{L}_{m,\theta},p\in\mathcal{P}}{\max}\mathbb{E}\left[g_{m,\theta}(f,p)\right]\geq\frac{\mathbb{E}\left[K_{m,\theta}(f_{\theta},p_{\theta})\right]}{\mathbb{E}\left[E_{m,\theta}(f_{\theta},p_{\theta})\right]}
𝔼[Qf,m,θ]pminLf=μf,θpminLf,\displaystyle\geq\frac{\mathbb{E}\left[Q_{f,m,\theta}\right]}{p_{\min}L^{\prime}_{f}}=\frac{\mu_{f,\theta}}{p_{\min}L^{\prime}_{f}},

where the second inequality follows from the approximation 𝔼[KE]𝔼[K]𝔼[E][K,E](𝔼[E])2+𝕍[E]𝔼[K](𝔼[E])3\mathbb{E}\left[\frac{K}{E}\right]\approx\frac{\mathbb{E}\left[K\right]}{\mathbb{E}\left[E\right]}-\frac{\mathbb{C}\left[K,E\right]}{\left(\mathbb{E}\left[E\right]\right)^{2}}+\frac{\mathbb{V}\left[E\right]\mathbb{E}\left[K\right]}{\left(\mathbb{E}\left[E\right]\right)^{3}} with \mathbb{C} and 𝕍\mathbb{V} denoting the covariance and variance respectively, together with the fact that Km,θK_{m,\theta} and Em,θE_{m,\theta} are independent random variables with positive expected values. The third inequality follows from the prerequisite of decoding a rateless-coded message (see Section III) and transmission with minimum power level.

References

  • [1] S. Maghsudi and M. van der Schaar, “A bandit learning approach to energy-efficient femto-caching under uncertainty,” in IEEE Global Communications Conference, 2019, pp. 1–6.
  • [2] S. Maghsudi and D. Niyato, “On transmission mode selection in D2D-enhanced small cell networks,” IEEE Wireless Communications Letters, vol. 6, no. 5, pp. 618–621, October 2017.
  • [3] S. Maghsudi and D. Niyato, “On power-efficient planning in dynamic small cell networks,” IEEE Wireless Communications Letters, vol. 7, no. 3, pp. 304–307, June 2018.
  • [4] M. Van der Schaar, D. S. Turaga, and R. Wong, “Classification-based system for cross-layer optimized wireless video transmission,” IEEE Transactions on Multimedia, vol. 8, no. 5, pp. 1082–1095, Oct 2006.
  • [5] M. Zink, K. Suh, Y. Gu, and J. Kurose, “Characteristics of YouTube network traffic at a campus network - measurements, models, and implications,” Computer Networks, vol. 53, no. 4, pp. 501–514, March 2009.
  • [6] M. Ji, G. Caire, and A. F. Molisch, “Wireless device-to-device caching networks: Basic principles and system performance,” IEEE Journal on Selected Areas in Communications, vol. 34, no. 1, pp. 176–189, Jan 2016.
  • [7] L. Zhang, M. Xiao, G. Wu, and S. Li, “Efficient scheduling and power allocation for D2D-assisted wireless caching networks,” IEEE Transactions on Communications, vol. 64, no. 6, pp. 2438 – 2452, June 2016.
  • [8] A. Altieri, P. Piantanida, L. R. Vega, and C. G. Galarza, “A stochastic geometry approach to distributed caching in large wireless networks,” in International Symposium on Wireless Communications Systems, August 2014, pp. 863–867.
  • [9] H. J. Kang and C. G. Kang, “Mobile device-to-device (D2D) content delivery networking: A design and optimization framework,” Journal of Communications and Networks, vol. 16, no. 5, pp. 568–577, October 2014.
  • [10] J. Dai, F. Liu, B. Li, and J. Liu, “Collaborative caching in wireless video streaming through resource auctions,” IEEE Journal on Selected Areas in Communications, vol. 30, no. 2, pp. 458–466, February 2012.
  • [11] Q. Zhang, Z. Xiang, W. Zhu, and L. Gao, “Cost-based cache replacement and server selection for multimedia proxy across wireless internet,” IEEE Transactions on Multimedia, vol. 6, no. 4, pp. 587–598, Aug 2004.
  • [12] A. Liu and V. K. N. Lau, “Asymptotic scaling laws of wireless Ad Hoc network with physical layer caching,” IEEE Transactions on Wireless Communications, vol. 15, no. 3, pp. 1657–1664, March 2016.
  • [13] M. A. Maddah-Ali and U. Niesen, “Fundamental limits of caching,” IEEE Transactions on Information Theory, vol. 60, no. 5, pp. 2856–2867, May 2014.
  • [14] Y. N. Shnaiwer, S. Sorour, N. Aboutorab, P. Sadeghi, and T. Y. Al-Naffouri, “Network-coded content delivery in femtocaching-assisted cellular networks,” in IEEE Global Communications Conference, December 2015, pp. 1–6.
  • [15] N. Golrezaei, K. Shanmugam, A. G. Dimakis, A. F. Molisch, and G. Caire, “Femtocaching: Wireless video content delivery through distributed caching helpers,” in IEEE International Conference on Computer Communications, March 2012, pp. 1107–1115.
  • [16] X. Peng, J. C. Shen, J. Zhang, and K. B. Letaief, “Backhaul-aware caching placement for wireless networks,” in IEEE Global Communications Conference, December 2015, pp. 1–6.
  • [17] M. Gregori, J. Gomez-Vilardebo, J. Matamoros, and D. Gunduz, “Wireless content caching for small cell and D2D networks,” IEEE Journal on Selected Areas in Communications, vol. 34, no. 5, pp. 1222–1234, May 2016.
  • [18] T. Wang, L. Song, and Z. Han, “Dynamic femtocaching for mobile users,” in IEEE Wireless Communications and Networking Conference, March 2015, pp. 861–865.
  • [19] J. Li, W. Chen, M. Xiao, F. Shu, and X. Liu, “Efficient video pricing and caching in heterogeneous networks,” IEEE Transactions on Vehicular Technology, vol. 65, no. 10, pp. 8744–8751, October 2016.
  • [20] H. Wu and H. Lu, “Delay and power tradeoff with consideration of caching capabilities in dense wireless networks,” IEEE Transactions on Wireless Communications, vol. 18, no. 10, pp. 5011–5025, October 2019.
  • [21] D. Liu and C. Yang, “Energy efficiency of downlink networks with caching at base stations,” IEEE Journal on Selected Areas in Communications, vol. 34, no. 4, pp. 907–922, April 2016.
  • [22] B. N. Bharath, K. G. Nagananda, and H. V. Poor, “A learning-based approach to caching in heterogenous small cell networks,” IEEE Transactions on Communications, vol. 64, no. 4, pp. 1674–1686, April 2016.
  • [23] A. Sengupta, S. Amuru, R. Tandon, R. M. Buehrer, and T. C. Clancy, “Learning distributed caching strategies in small cell networks,” in International Symposium on Wireless Communications Systems, August 2014, pp. 917–921.
  • [24] S. Muller, O. Atan, M. van der Schaar, and A. Klein, “Context-aware proactive content caching with service differentiation in wireless networks,” IEEE Transactions on Wireless Communications, vol. 16, no. 2, pp. 1024–1036, February 2017.
  • [25] S. Li, J. Xu, M. van der Schaar, and W. Li, “Trend-aware video caching through online learning,” IEEE Transactions on Multimedia, vol. 18, no. 12, pp. 2503–2516, December 2016.
  • [26] P. Blasco and D. Gunduz, “Learning-based optimization of cache content in a small cell base station,” in IEEE International Conference on Communications, June 2014, pp. 1897–1903.
  • [27] S. Muller, O. Atan, M. van der Schaar, and A. Klein, “Smart caching in wireless small cell networks via contextual multi-armed bandits,” in IEEE International Conference on Communications, May 2016, pp. 1–7.
  • [28] K. Poularakis, G. Iosifidis, V. Sourlas, and L. Tassiulas, “Exploiting caching and multicast for 5G wireless networks,” IEEE Transactions on Wireless Communications, vol. 15, no. 4, pp. 2995–3007, April 2016.
  • [29] J. N. Shim, B. Y. Min, K. Kim, J. Jang, and D. K. Kim, “Advanced femto-caching file placement technique for overlapped helper coverage,” in Vehicular Technology Conference, May 2014, pp. 1–5.
  • [30] C. Yang, Z. Chen, Y. Yao, B. Xia, and H. Liu, “Energy efficiency in wireless cooperative caching networks,” in IEEE International Conference on Communications, June 2014, pp. 4975–4980.
  • [31] S. Maghsudi and S. Stanczak, “A delay-constrained rateless coded incremental relaying protocol for two-hop transmission,” in IEEE Wireless Communications and Networking Conference, April 2012, pp. 168–172.
  • [32] S. Maghsudi and S. Stanczak, “On network-coded rateless transmission: Protocol design, clustering and cooperator assignment,” in International Symposium on Wireless Communication Systems, August 2012, pp. 306–310.
  • [33] J. Castura, Y. Mao, and S. Draper, “On rateless coding over fading channels with delay constraints,” in IEEE International Symposium on Information Theory, July 2006, pp. 1124–1128.
  • [34] S. Maghsudi and E. Hossain, “Distributed user association in energy harvesting small cell networks: A probabilistic bandit model,” IEEE Transactions on Wireless Communications, vol. 16, no. 3, pp. 1549–1563, March 2017.
  • [35] M. Luby, “LT codes,” in IEEE Symposium on Foundations of Computer Science, November 2002, pp. 271–280.
  • [36] D. Willkomm, J. Gross, and A. Wolisz, “Reliable link maintenance in cognitive radio systems,” in IEEE International Symposium on New Frontiers in Dynamic Spectrum Access Networks, November 2005, pp. 371–378.
  • [37] Y. Sun, C. E. Koksal, S. Lee, and N. B. Shroff, “Network control without CSI using rateless codes for downlink cellular systems,” in IEEE International Conference on Computer Communications, April 2013, pp. 1016–1024.
  • [38] A. Venkiah, P. Piantanida, C. Poullia, P. Duhamel, and D. Declercq, “Rateless coding for quasi-static fading channels using channel estimation accuracy,” in IEEE International Symposium on Information Theory, July 2008, pp. 2257–2261.
  • [39] B. Yang, G. Mao, M. Ding, X. Ge, and X. Tao, “Dense small cell networks: From noise-limited to dense interference-limited,” IEEE Transactions on Vehicular Technology, vol. 67, no. 5, pp. 4262–4277, May 2018.
  • [40] H. A. David and H. N. Nagaraja, Order Statistics, John Wiley & Sons, 2003.
  • [41] L. L. Minku, A. P. White, and X. Yao, “The impact of diversity on online ensemble learning in the presence of concept drift,” IEEE Transactions on Knowledge and Data Engineering, vol. 22, no. 5, pp. 730–742, May 2010.
  • [42] C. Tekin and M. van der Schaar, “Distributed online big data classification using context information,” in Annual Allerton Conference on Communication, Control, and Computing, 2013, pp. 1435–1442.
  • [43] G. Tyson, Y. Elkhatib, N. Sastry, and S. Uhlig, “Measurements and analysis of a major adult video portal,” ACM Transactions on Multimedia Computing Communications and Applications, vol. 12, no. 2, pp. 1–25, Jan 2016.
  • [44] S. Maghsudi and E. Hossain, “Distributed user association in energy harvesting small cell networks: An exchange economy with uncertainty,” IEEE Transactions on Green Communications and Networking, vol. 1, no. 3, pp. 294–108, September 2017.
  • [45] M. Basseville and I.V. Nikiforov, Detection of Abrupt Changes:Theory and Application, Prentice-Hall, 1993.
  • [46] E. S. Page, “Continuous inspection schemes,” Biometrika, vol. 41, no. 1-2, pp. 100–115, June 1954.
  • [47] G. A. Barnard, “Control charts and stochastic processes,” Journal of the Royal Statistical Society, Series B (Methodological), vol. 21, no. 2, pp. 239–271, June 1959.
  • [48] Z. Bnaya, R. Puzis, R. Stern, and A. Felner, “Volatile multi-armed bandits for guaranteed targeted social crawling,” in Proceedings of the AAAI Conference on Late-Breaking Developments in the Field of Artificial Intelligence, June 2013, pp. 8–10.
  • [49] P. Auer, N. Cesa-Bianchi, and P. Fischer, “Finite-time analysis of the multiarmed bandit problem,” Machine Learning, vol. 47, no. 2-3, pp. 235–256, May 2002.
  • [50] H. Kellerer, U. Pferschy, and D. Pisinger, Knapsack Problems, Springer, 2004.
  • [51] D. Ciuonzo, P. S. Rossi, and P. Willett, “Generalized Rao test for decentralized detection of an uncooperative target,” IEEE Signal Processing Letters, vol. 24, no. 5, pp. 678–682, May 2017.
  • [52] D. E. Knuth, The Art of Computer Programming, Volume 3: Sorting and Searching, Addison Wesley Longman Publishing Co., 2nd edition, 1998.
  • [53] J. Castura and Y. Mao, “Rateless coding over fading channels,” IEEE Communications Letters, vol. 10, no. 1, pp. 46–48, 2006.
  • [54] S. Maghsudi and S. Stanczak, “On channel selection for energy-constrained rateless-coded D2D communications,” in European Signal Processing Conference, Aug 2015, pp. 1028–1032.
  • [55] Y. Yao and A. U. H. Sheikh, “Outage probability analysis for microcell mobile radio systems with cochannel interferers in Rician/Rayleigh fading environment,” Electronics Letters, vol. 26, no. 13, pp. 864–866, June 1990.