Better and Simpler Learning-Augmented Online Caching
Abstract.
Lykouris and Vassilvitskii (ICML 2018) introduce a model of online caching with machine-learned advice, where each page request additionally comes with a prediction of when that page will next be requested. In this model, a natural goal is to design algorithms that (1) perform well when the advice is accurate and (2) remain robust in the worst case a la traditional competitive analysis. Lykouris and Vassilvitskii give such an algorithm by adapting the Marker algorithm to the learning-augmented setting. In a recent work, Rohatgi (SODA 2020) improves on their result with an approach also inspired by randomized marking. We continue the study of this problem, but with a somewhat different approach: We consider combining the BlindOracle algorithm, which just naïvely follows the predictions, with an optimal competitive algorithm for online caching in a black-box manner. The resulting algorithm outperforms all existing approaches while being significantly simpler. Moreover, we show that combining BlindOracle with LRU is in fact optimal among deterministic algorithms for this problem.
1. Introduction
Traditionally, the study of online algorithms focuses on worst-case robustness, with algorithms providing the same competitive guarantee against the offline optimal over all inputs. In recent years, however, there has been a surge of interest in studying online algorithms in the presence of structured inputs [MV17, LV18, PSK18, GP19, KPS+19, KPS+20, Roh20, LLM+20, Mit20]. A principal motivation for these works is the philosophy of beyond worst-case analysis [KP00, Rou19]: Many practical settings have inputs that follow restricted patterns, making classical worst-case competitive analysis too pessimistic to inform practice. In particular, the worst-case examples classical competitive analyses guard against often do not materialize. Furthermore, algorithms designed with the worst case in mind can be hamstrung by these considerations, ending up unnatural and losing performance on “nice” inputs.
Learning-augmented online algorithms, introduced by [LV18, PSK18], is a beyond worst-case framework motivated by the powerful predictive abilities of modern machine learning. The structure of the input is assumed to come in the form of a machine-learned predictor that provides predictions of future inputs. A concern with this setup may be that machine learning models typically have few worst-case guarantees. Nonetheless, with learning-augmented algorithms, we want the best of both worlds: Given a predictor, the objective is to design algorithms that (1) perform well in the optimistic scenario, where the predictor has low error, and (2) remain robust in the classical worst-case sense, when the predictor can be arbitrarily bad. That is, we would like our algorithm to be -competitive against the offline optimal on all inputs, where is a function of the predictor error such that for some . Such an algorithm is said to be -robust.
The focus of our work is learning-augmented online caching [LV18, Roh20]. In the online caching (a.k.a. online paging) problem, one seeks to maintain a cache of size online while serving requests for pages that may or may not be in the cache. For simplicity, we assume that pages must always be served from the cache and that bringing a page into the cache has unit cost. (In particular, if the cache is full, bringing a page into the cache requires also evicting a page already in the cache.) Thus, we seek to minimize the number of requests for which the page is not already in the cache, i.e., the number of cache misses. This is a classical online problem that has been the subject of extensive study over the past several decades (see [BE98] for an overview). From the worst-case perspective, this problem is well-understood for not only the version stated above [ST85, FKL+91, ACN00], but also for weighted generalizations [BBN12, BBN12a].
Online caching in the learning-augmented context was first considered by [LV18]. They introduce a model of prediction where the predictor, upon the arrival of each page, predicts the next time that this page will be requested. They show that the BlindOracle algorithm, which follows the predictor naïvely and evicts the page with the latest predicted arrival time, can have unbounded competitive ratio (i.e., is non-robust). They then give a different algorithm, PredictiveMarker, based on the Marker algorithm of [FKL+91], that achieves a competitive ratio of
where is the error of the predictor and is the cost of the offline optimal. (See Section 2.1 for precise definitions.) In [Roh20], Rohatgi introduces the LNonMarker algorithm, which is also based on randomized marking (but eschews the framework somewhat), and shows that this algorithm obtains a competitive ratio of
This bound is obtained by first constructing a non-robust algorithm and then using a black-box combination technique discussed in [LV18] to combine this non-robust algorithm with the Marker algorithm. Rohatgi also provides a lower bound of
for the competitive ratio of any learning-augmented online algorithm for caching in terms of , , and .
1.1. Our Contribution
We show that the strikingly simple approach of combining BlindOracle with an -competitive online caching algorithm (e.g., Marker) in a black-box fashion obtains a competitive ratio bound of
improving on that of LNonMarker. Thus, although BlindOracle is non-robust [LV18], we show that it should not be abandoned entirely. In fact, it has excellent performance when is small. So when this algorithm is combined with a -competitive algorithm, we start seeing performance improvement over the robust algorithm starting at .
And although our improvement in the competitive ratio is slight, previous approaches for learning-augmented online caching [LV18, Roh20] have relied on much more intricate constructions based on randomized marking. We therefore believe that our simple approach may yield better practical performance and may generalize more readily to other learning-augmented settings. (Indeed, we note that the deterministic combining algorithm is particularly simple: Just “follow” the algorithm with the better performance thus far.)
We also give precise bounds on the constant factors in the competitive ratios that we obtain. [FKL+91, BB00] provide optimal bounds for combining online algorithms online in a black-box manner, with better constants than the approach discussed in [LV18] and applied in [Roh20]. By composing a careful competitive analysis of BlindOracle with these “combiners,” we obtain constants in the competitive ratio that are lower than those of previous work.
Finally, we show that combining BlindOracle with a -competitive deterministic algorithm (e.g., LRU [ST85]) is the best one could hope to do among deterministic algorithms for learning-augmented online caching. In particular, we show that a linear dependence on in the competitive ratio is necessary. Therefore, if a logarithmic dependence on is to be achieved, as in Rohatgi’s lower bound, then randomization is needed, (perhaps surprisingly) even in the regime where is bounded.
Stated formally, our main result analyzing BlindOracle is the following:
Theorem 1.1 (restate=maintheorem,label=theorem:main).
For learning-augmented online caching, BlindOracle obtains a competitive ratio of
where is the loss incurred by the predictor and is the offline optimal cost. (For precise definitions, see Section 2.1.)
Plugging this bound into the results of [FKL+91, BB00] for competitively combining online algorithms online (see Section 2.2) yields the following corollaries:
Corollary 1.2 (restate=firstcorollary,label=corollary:det).
There exists a deterministic algorithm for learning-augmented online caching that achieves a competitive ratio of
Corollary 1.3 (restate=secondcorollary,label=corollary:rand).
There exists a randomized algorithm for learning-augmented online caching that achieves a competitive ratio of
for any .111The trade-off in and the additional cost is additive; thus, it does not factor into the competitive ratio. (Here, is the -th harmonic number.)
Finally, we state our matching lower bound for LABEL:corollary:det on deterministic algorithms for learning-augmented online caching:
Theorem 1.4 (restate=lowerboundtheorem,label=theorem:lower).
The competitive ratio bound for any deterministic learning-augmented online caching algorithm must be at least
1.2. Related Work
In addition to the predecessor works by [LV18, Roh20] on learning-augmented online caching, there have been several other recent papers in the space of learning-augmented online algorithms: [MV17] study repeated posted-price auctions, [PSK18, GP19] study the ski rental problem, and [PSK18, LLM+20, Mit20] study online scheduling. Of these, the scheduling algorithm of [PSK18] is the most similar in spirit to this present work: Both algorithms are based on combining a naïve and optimistic algorithm with a robust algorithm.
Other threads of research falling under beyond worst-case online algorithms include work on combining multiple algorithms with different performance characteristics [FKL+91, BB00, MNS12, GP19], designing online algorithms with distributional assumptions (e.g., stochasticity) on the input [KP00, BS12, MGZ12], and semi-online algorithms, where the input is assumed to have a predictable offline component and an adversarial online component [KPS+19, KPS+20].
The idea of learning-augmentation has also been explored in many other algorithmic and data structural settings in recent years. These include learned indices [KBC+18], bloom filters [Mit18], frequency estimation in streams [HIK+19], and nearest neighbor search [DIR+19], among others.
Finally, advice for online algorithms has also been considered with a more complexity theoretic spirit through the study of advice complexity of online algorithms; see the survey of [BFK+17].
1.3. Recent Developments
Recently, in work done independently of and concurrently with this paper, [ACE+20] also study a BlindOracle-like algorithm, which they term FollowThePrediction, in the more general setting of learning-augmented metrical task systems; they also use the “combiner” of [BB00] to make this algorithm robust. However, their prediction model, when specialized to online caching, is incomparable to that of [LV18] (which we follow).222Namely, their algorithms expect predictions to be in a different form: They expect predictions to be cache states (i.e., the set of pages in the cache at time ) rather than next arrival times of pages. Moreover, there exist sequences of “corresponding” inputs for each of these two models such that the predictor error approaches infinity in one model while remaining constant in the other. Thus, the theoretical results proved in these two models do not imply each other.
2. Preliminaries
2.1. Setup and Notation
In the online caching problem, we receive a sequence of page requests online, and our goal is to serve these requests using a cache of size while minimizing cost. In this problem, pages must be served from the cache and can be served at no cost; however, evicting a page from the cache has unit cost.333Note that this is equivalent to the “standard” version, where each cache miss has unit cost, up to a constant of .
We will establish competitive bounds comparing the performance of two online caching algorithms and . More precisely, we will show bounds of the form
where and are the costs of and , respectively, as measured in number of evictions made while serving a sequence of page requests. We will also compare our costs to the offline optimal algorithm , whose cost is the minimum possible cost of serving request sequence . We will omit the argument when the context is clear (i.e., just writing to represent ).
In our analysis, we use and to denote the cache states of and , respectively, just before the -th request. Formally, and are subsets of of size at most , containing for each cached page the index at which it was last served. That is, when serving the -th request, we remove some old request index from the cache and insert . Thus, if is such that , this operation is free; otherwise, it has unit cost. In the sequel, we will also refer to these indices as page requests.
In the learning-augmented online caching problem, the -th page request comes with a prediction for the next time page is requested. That is, at the time of the -th request, our algorithm receives the pair . Let be the tuple of all predictions. To define a notion of loss, let denote for each the next time page is actually requested, with if page is never requested again. The loss is then defined to be
We will omit arguments to if the context is clear. Note that if , then the offline optimal can be obtained, as the optimal algorithm always evicts the page that is next requested furthest into the future.
In stating our bounds, the essential quantity is often . To make this clear, we take and state our bounds in terms of in the sequel.
2.1.1. Inversions
Call a pair of page requests an inversion if but . Let denote the total number of inversions between the pair of sequences and . And as above, we will omit the arguments to when the context is clear.
2.1.2. BlindOracle
We now formally define the BlindOracle algorithm as follows: For each page request, if the requested page is already in the cache, do nothing. Otherwise, evict the page request whose predicted next arrival time is furthest away among all , with ties broken consistently (e.g., by always evicting the least recently used page with maximal ).
2.2. Combining Online Algorithms Competitively
In this section, we state some classical bounds on competitively combining online algorithms, due to [FKL+91] and [BB00]. This type of “black-box” combination was also considered by [LV18], but their approach has a worst constant than that of [FKL+91]. We also note that results of a similar flavor are proven by [PSK18, MNS12], but for other online problems.
The question of combining multiple online algorithms while remaining competitive against each was first considered in the seminal paper of [FKL+91]. They consider combining online algorithms for the online caching problem into a single algorithm such that is -competitive against for each . They show that such an is achievable if and only if
We will need only the special case of and , which we state below:
Theorem 2.1 ([FKL+91], special case).
Given any two algorithms and for the online caching problem, there exists an algorithm such that
Moreover, if and are deterministic, so is .
Indeed, we note that that this can be done deterministically with a “follow-the-leader” approach, in which we simulate both algorithms and at each step evict any page that is not in the cache of the better performing algorithm (as measured by total number of evictions after serving the current request).
[BB00] show that one can obtain a better approximation factor using a randomized scheme, namely multiplicative weights.444The result of [BB00] in fact holds for general metrical task systems. That is, at each point in time, the probability that the combined algorithm is following one of the algorithms is given by a probability distribution over the algorithms governed by the multiplicative weights update rule. For , their result can be stated as follows:
Theorem 2.2 ([BB00], special case).
Given any two algorithms and for the online caching problem and any , , there exists an algorithm such that
Remark.
Although we do not state the versions of these results for , one could imagine that they can be useful wishes to combine multiple machine-learned predictors.
2.3. From Loss to Inversions
We now state a lemma of [Roh20] that relates loss to the number of inversions, letting us lower bound the loss by lower bounding the number of inversions . Thus, instead of reasoning in terms of loss, we will reason in terms of inversions.
Lemma 2.3 ([Roh20]).
For any and , .
With this lemma, it suffices (up to a factor of ) to give our competitive ratio upper bounds in terms of the number of inversions .
3. A First Analysis of BlindOracle
In this section, we give a first analysis of BlindOracle, showing that it gets very good performance when the ratio is very small. In particular, our analysis shows that as , the competitive ratio achieved approaches .
Let be the offline optimal algorithm (i.e., such that ). Let be BlindOracle. Note that we can think of each of , , and as functions of the time , i.e., they are the cost of , the cost of , and the number of inversions, respectively, on the prefix consisting of the first requests.555This indexing is to be consistent with the definitions of and . We use the operator to denote the change (in a function of ) from time to time . For example, if evicts an element upon the -th request.
In our analysis, we maintain a matching between and at all times . Call a matching valid if it consists only of pairs such that the next arrival of is no later than the next arrival of . Indeed, our matching will be valid throughout the execution of the algorithm.
We now proceed with a potential function analysis, taking our potential (as a function of , , and ) to be the number of unmatched pages in . For notational simplicity, we will simply denote by . Given this setup, we show:
Proposition 3.1.
There exists a valid matching such that
Proof.
We induct on the length of the input and perform a case analysis to show that we can maintain a valid matching such that at each time step, the right-hand side increases at least as much as the left-hand side, i.e., .
For our base case, note that , so we may take to be the identity matching.
Now, upon a request at time , we update according to the following cases (and with the consequences listed for each case):
-
(1)
The requested page is in both and .
-
(a)
The cached pages are matched to each other.
-
•
Do nothing.
-
•
-
(b)
Otherwise:
-
(i)
Both cached pages are matched.
-
•
Remove the pairs and from .
-
•
Add the pairs and to .
-
•
As a result:
-
–
.
-
–
-
•
-
(ii)
Otherwise:
-
•
Remove any pairs involving from . (There is at most one such pair.)
-
•
Add the pair to .
-
•
As a result:
-
–
.
-
–
-
•
-
(i)
-
(a)
-
(2)
The requested page is in only.
-
•
Remove any pairs involving the evicted page from . (There is at most one such pair.)
-
•
Remove any pairs involving the requested page from . (There is at most one such pair.)
-
•
Add the pair to .
-
•
As a result:
-
–
.
-
–
.
-
–
-
•
-
(3)
The requested page is in only.
-
(a)
The evicted page is unmatched.
-
•
Add the pair to . (The arriving page cannot be in any valid matching.)
-
•
As a result:
-
–
.
-
–
.
-
–
-
•
-
(b)
The evicted page is matched.
-
(i)
arrives later than all unmatched pages in .
-
•
Remove the pair involving the evicted page from .
-
•
Add the pair to , where is any unmatched page.
-
•
Add the pair to . (The arriving page cannot be in any valid matching.)
-
•
As a result:
-
–
.
-
–
.
-
–
-
•
-
(ii)
There is an unmatched page arriving later than .
-
•
Remove the pair involving the evicted page from .
-
•
Add the pair to . (The arriving page cannot be in any valid matching.)
-
•
As a result:
-
–
.
-
–
.
-
–
, as there is an inversion between and . (Note that we do not count this inversion ever again, as gets evicted.)
-
–
-
•
-
(i)
-
(a)
-
(4)
The requested page is in neither nor .
-
(a)
evicts an unmatched page .
-
(i)
evicts an unmatched page .
-
•
Add the pair to .
-
•
As a result:
-
–
.
-
–
.
-
–
.
-
–
-
•
-
(ii)
evicts a matched page .
-
•
Remove the pair involving from .
-
•
Add the pair to .
-
•
As a result:
-
–
.
-
–
.
-
–
-
•
-
(i)
-
(b)
evicts a matched page .
-
(i)
evicts an unmatched page .
-
•
Remove the pair involving from .
-
•
Add the pair to .
-
•
As a result:
-
–
.
-
–
.
-
–
-
•
-
(ii)
evicts a matched page .
-
•
Remove the pair involving from .
-
•
Remove the pair involving from .
-
•
Add the pair to .
-
•
As a result:
-
–
.
-
–
.
-
–
Note that either arrives after , in which case we can add to and , or the pair forms an inversion, in which case and . (As before, since is getting evicted, we will not count this pair twice.)
-
–
-
•
-
(i)
-
(a)
It is not hard to verify that the change in the left-hand side of the bound is no more than the change in the right-hand side in each of the cases listed above, from which the proposition follows. ∎
Proposition 3.2.
The competitive ratio of algorithm is at most .
Proof.
Note that is bounded below by the number of inversions of by Lemma 2.3. By Proposition 3.1, , so . ∎
4. A More Careful Analysis
In this section, we give an asymptotically better (in ) bound for the performance of BlindOracle. A more careful analysis is needed to show an upper bound with a coefficient on the ratio . We use the same high-level approach for the proof as before, but with a more complicated potential function. Again, is the offline optimal algorithm and is the BlindOracle algorithm, and also as before, we use to denote change (in functions of ) from request to request .
We maintain in this proof a matching over pairs of page requests such that for each time step . Our potential function will be a function of , , and . For notational simplicity, we will simply denote by .
Given , , and at time , define to be the number of that are unmatched. Define to be the number of such that . In other words, counts how many page requests in are not matched to the same page request in . Let be the number of pages in predicted to appear no later than , with tie-breaking done in a consistent manner (e.g., by the last time the page was requested). Next, define
where and . Finally, we take
as our overall potential function.
Proposition 4.1.
For any input , there exists a matching consisting only of pairs satisfying such that
Proof.
We again induct on the length of the input, and we again perform a case analysis to show that we can maintain a matching consisting only of pairs satisfying such that at each time step, the right-hand side increases at least as much as the left-hand side.
For our analysis, we split the serving of each page request into two phases:
-
(1)
Matching. Update so that the page requests in and that are to be removed are unmatched. (Note that page requests are removed either because the corresponding page was requested again or because the corresponding page was evicted.)
-
(2)
Updating. Replace a page request from each of and with the new request and insert the new page request pair into .
We first analyze how updating affects the potential . This operation always decreases and each by , since we remove an unmatched pair. Next, for , observe that for a matched pair , the difference increases on the -th request only if there exists a such that and . In this case, the pair also forms an inversion. Any inversion is counted at most once this way because is evicted from . Thus, we have .
We now analyze the matching phase with a case analysis:
-
(1)
The requested page is in both and .
-
•
The previous page requests for in and are matched to each other, so we can just unmatch them.
-
•
As a result:
-
–
.
-
–
.
-
–
, since the pages were matched to each other.
-
–
-
•
-
(2)
The requested page is in only.
-
(a)
The previous request for the requested page is matched as and the page request evicted by is matched as .
-
•
Unmatch and and then match . The latter is okay since . Note that .
-
•
As a result:
-
–
.
-
–
.
-
–
.
-
–
, since and .
-
–
, since the arrival of also generates inversions of the form for all such that .
-
–
-
•
-
(b)
The previous request for the requested page is matched as and the page request evicted by is unmatched.
-
•
Unmatch . Note that .
-
•
As a result:
-
–
.
-
–
.
-
–
.
-
–
, since and .
-
–
, since the arrival of also generates inversions of the form for all such that .
-
–
-
•
-
(c)
The previous request for the requested page is unmatched and the page request evicted by is matched as .
-
•
Unmatch and match for an arbitrary unmatched . Doing so is okay because .
-
•
As a result:
-
–
.
-
–
.
-
–
.
-
–
, since and .
-
–
-
•
-
(d)
The previous request for the requested page is unmatched and the page request evicted by is unmatched.
-
•
Do nothing.
-
•
As a result:
-
–
.
-
–
-
•
-
(a)
-
(3)
The requested page is in only.
-
(a)
The previous request for the requested page is matched as and the page request evicted by is matched as .
-
•
Unmatch and . Note that .
-
•
As a result:
-
–
.
-
–
.
-
–
If , then and ; otherwise, , in which case and .
-
–
Moreover, .
-
–
-
•
-
(b)
The previous request for the requested page is matched as and the page request evicted by is unmatched.
-
•
Unmatch . Note that .
-
•
As a result:
-
–
.
-
–
.
-
–
.
-
–
, since .
-
–
-
•
-
(c)
The previous request for the requested page is unmatched and the page request evicted by is matched as .
-
•
Unmatch .
-
•
As a result:
-
–
.
-
–
.
-
–
If , then and ; otherwise, , in which case and .
-
–
-
•
-
(d)
The previous request for the requested page is unmatched and the page request evicted by is unmatched.
-
•
Do nothing.
-
•
As a result:
-
–
.
-
–
-
•
-
(a)
-
(4)
The requested page is in neither nor .
-
(a)
The previous request evicted by is matched as and the page request evicted by is matched as .
-
•
Unmatch and and then match . The latter is okay because .
-
•
As a result:
-
–
.
-
–
.
-
–
.
-
–
.
-
–
, since and .
-
–
-
•
-
(b)
The previous request evicted by is matched as and the page request evicted by is unmatched.
-
•
Unmatch .
-
•
As a result:
-
–
.
-
–
.
-
–
.
-
–
If , then and ; otherwise, , in which case and .
-
–
-
•
-
(c)
The previous request evicted by is unmatched and the page request evicted by is matched as .
-
•
Unmatch and match .
-
•
As a result:
-
–
.
-
–
.
-
–
.
-
–
.
-
–
, since and .
-
–
-
•
-
(d)
The previous request evicted by is unmatched and the page request evicted by is unmatched.
-
•
Do nothing.
-
•
As a result:
-
–
.
-
–
.
-
–
-
•
-
(a)
Since and each decrease by in the updating phase, we have in extra potential that we can use to pay for costs in the matching phase. Indeed, one can verify that this is sufficient for all of the cases described above—the tight cases are 2(a), 2(b), 2(c), 3(a), and 4(a). Thus, the proposition follows. ∎
Proposition 4.2.
The competitive ratio of algorithm is at most .
Proof.
Compose Proposition 4.1 with Lemma 2.3. ∎
Remark.
This analysis of BlindOracle is tight in the constant term—we can make arbitrarily small while having a competitive ratio of . (However, for very small , the bound of the previous section is better—in particular, if .)
5. Proofs of Upper Bounds
We are now ready to prove the results stated in Section 1: \maintheorem*
Proof.
From the analysis of the previous two sections, the desired bound immediately follows from taking the minimum of the bounds in Propositions 3.2, LABEL: and 4.2. ∎
*
Proof.
Combine BlindOracle with LRU using the “combiner” from Theorem 2.1, with the performance of BlindOracle being bounded by LABEL:theorem:main. ∎
*
Proof.
Like in the proof above, combine BlindOracle with algorithm Equitable of [ACN00]666We use Equitable because it achieves the optimal worst-case competitive ratio of for online caching; Marker has a competitive ratio of [ACN00]., this time using the “combiner” from Theorem 2.2. ∎
6. Deterministic Lower Bound
We now show that combining BlindOracle with LRU gets an optimal competitive ratio bound (in terms of , , and ) among all deterministic algorithms for learning-augmented online caching by proving LABEL:theorem:lower:
*
Proof.
Let be any deterministic algorithm for learning-augmented online caching.
We show there exists a family of inputs with ranging from to and arbitrarily large such , for some constant . That is, for ranging from to , we will show that we can make this inequality hold as with fixed. Dividing through by then implies the theorem.
We now construct such inputs . First, fix . Let be distinct pages. We make the following sequence of requests, which we call a phase:
-
(1)
Repeat times the following:
-
(a)
Make requests to in order, predicting each page to next appear requests from now except during the last iteration, where we predict each page to next appear requests from now.
-
(a)
-
(2)
Make a request to and predict that it will next appear pages from now.
-
(3)
For :
-
(a)
Request the page evicted by in during the previous request, if it exists. Otherwise, request an arbitrary page. For each page, provide the same prediction as the last time this page was requested.
-
(a)
We repeat the above as many times as needed.
In a single phase, observe that makes at most two evictions—once to evict and once upon the arrival of . On the other hand, I claim makes at least evictions. First, if the cache of after (1) does not consist of , then must have incurred cost at least during (1). Thus, we may assume that ’s cache consists of after (1). If so, has to evict a page for each of the remaining requests in the phase, as the arrival of forces an eviction and by induction, each arrival of (3) forces an eviction. Finally, observe that all the predictions are accurate except those for pages arriving in (3), in which case they are off by at most . Thus, over a single phase, . Putting all of these observations together, we get , with .
To make arbitrarily large, note that we can simply repeat the above phase multiple times in sequence; the same analysis holds, with all the values scaling linearly. Hence we have for arbitrarily large over the desired range of , so the theorem follows. ∎
Acknowledgments
I would like to thank Jelani Nelson for advising this project and Bailey Flanigan for providing many helpful references.
References
- [ACE+20] Antonios Antoniadis et al. “Online metric algorithms with untrusted predictions” In CoRR abs/2003.02144, 2020
- [ACN00] Dimitris Achlioptas, Marek Chrobak and John Noga “Competitive analysis of randomized paging algorithms” In Theor. Comput. Sci. 234.1-2, 2000, pp. 203–218
- [BB00] Avrim Blum and Carl Burch “On-line Learning and the Metrical Task System Problem” In Mach. Learn. 39.1, 2000, pp. 35–58
- [BBN12] Nikhil Bansal, Niv Buchbinder and Joseph Naor “A Primal-Dual Randomized Algorithm for Weighted Paging” In J. ACM 59.4, 2012, pp. 19:1–19:24
- [BBN12a] Nikhil Bansal, Niv Buchbinder and Joseph Naor “Randomized Competitive Algorithms for Generalized Caching” In SIAM J. Comput. 41.2, 2012, pp. 391–414
- [BE98] Allan Borodin and Ran El-Yaniv “Online computation and competitive analysis” Cambridge University Press, 1998
- [BFK+17] Joan Boyar et al. “Online Algorithms with Advice: A Survey” In ACM Comput. Surv. 50.2, 2017, pp. 19:1–19:34
- [BS12] Sébastien Bubeck and Aleksandrs Slivkins “The Best of Both Worlds: Stochastic and Adversarial Bandits” In COLT 2012 - The 25th Annual Conference on Learning Theory, June 25-27, 2012, Edinburgh, Scotland, 2012, pp. 42.1–42.23
- [DIR+19] Yihe Dong, Piotr Indyk, Ilya P. Razenshteyn and Tal Wagner “Learning Sublinear-Time Indexing for Nearest Neighbor Search” In CoRR abs/1901.08544, 2019
- [FKL+91] Amos Fiat et al. “Competitive Paging Algorithms” In J. Algorithms 12.4, 1991, pp. 685–699
- [GP19] Sreenivas Gollapudi and Debmalya Panigrahi “Online Algorithms for Rent-Or-Buy with Expert Advice” In Proceedings of the 36th International Conference on Machine Learning, ICML 2019, 9-15 June 2019, Long Beach, California, USA, 2019, pp. 2319–2327
- [HIK+19] Chen-Yu Hsu, Piotr Indyk, Dina Katabi and Ali Vakilian “Learning-Based Frequency Estimation Algorithms” In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019, 2019
- [KBC+18] Tim Kraska et al. “The Case for Learned Index Structures” In Proceedings of the 2018 International Conference on Management of Data, SIGMOD Conference 2018, Houston, TX, USA, June 10-15, 2018, 2018, pp. 489–504
- [KP00] Elias Koutsoupias and Christos H. Papadimitriou “Beyond Competitive Analysis” In SIAM J. Comput. 30.1, 2000, pp. 300–317
- [KPS+19] Ravi Kumar et al. “Semi-Online Bipartite Matching” In 10th Innovations in Theoretical Computer Science Conference, ITCS 2019, January 10-12, 2019, San Diego, California, USA, 2019, pp. 50:1–50:20
- [KPS+20] Ravi Kumar, Manish Purohit, Zoya Svitkina and Erik Vee “Interleaved Caching with Access Graphs” In Proceedings of the 2020 ACM-SIAM Symposium on Discrete Algorithms, SODA 2020, Salt Lake City, UT, USA, January 5-8, 2020, 2020, pp. 1846–1858
- [LLM+20] Silvio Lattanzi, Thomas Lavastida, Benjamin Moseley and Sergei Vassilvitskii “Online Scheduling via Learned Weights” In Proceedings of the 2020 ACM-SIAM Symposium on Discrete Algorithms, SODA 2020, Salt Lake City, UT, USA, January 5-8, 2020, 2020, pp. 1859–1877
- [LV18] Thodoris Lykouris and Sergei Vassilvitskii “Competitive Caching with Machine Learned Advice” In Proceedings of the 35th International Conference on Machine Learning, ICML 2018, Stockholmsmässan, Stockholm, Sweden, July 10-15, 2018, 2018, pp. 3302–3311
- [MGZ12] Vahab S. Mirrokni, Shayan Oveis Gharan and Morteza Zadimoghaddam “Simultaneous approximations for adversarial and stochastic online budgeted allocation” In Proceedings of the Twenty-Third Annual ACM-SIAM Symposium on Discrete Algorithms, SODA 2012, Kyoto, Japan, January 17-19, 2012, 2012, pp. 1690–1701
- [Mit18] Michael Mitzenmacher “A Model for Learned Bloom Filters and Optimizing by Sandwiching” In Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, NeurIPS 2018, 3-8 December 2018, Montréal, Canada, 2018, pp. 462–471
- [Mit20] Michael Mitzenmacher “Scheduling with Predictions and the Price of Misprediction” In 11th Innovations in Theoretical Computer Science Conference, ITCS 2020, January 12-14, 2020, Seattle, Washington, USA, 2020, pp. 14:1–14:18
- [MNS12] Mohammad Mahdian, Hamid Nazerzadeh and Amin Saberi “Online Optimization with Uncertain Information” In ACM Trans. Algorithms 8.1, 2012, pp. 2:1–2:29
- [MV17] Andres Muñoz Medina and Sergei Vassilvitskii “Revenue Optimization with Approximate Bid Predictions” In Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, 4-9 December 2017, Long Beach, CA, USA, 2017, pp. 1858–1866
- [PSK18] Manish Purohit, Zoya Svitkina and Ravi Kumar “Improving Online Algorithms via ML Predictions” In Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, NeurIPS 2018, 3-8 December 2018, Montréal, Canada, 2018, pp. 9684–9693
- [Roh20] Dhruv Rohatgi “Near-Optimal Bounds for Online Caching with Machine Learned Advice” In Proceedings of the 2020 ACM-SIAM Symposium on Discrete Algorithms, SODA 2020, Salt Lake City, UT, USA, January 5-8, 2020, 2020, pp. 1834–1845
- [Rou19] Tim Roughgarden “Beyond worst-case analysis” In Commun. ACM 62.3, 2019, pp. 88–96
- [ST85] Daniel Dominic Sleator and Robert Endre Tarjan “Amortized Efficiency of List Update and Paging Rules” In Commun. ACM 28.2, 1985, pp. 202–208