2024
\jmlrproceedings
\coltauthor\NameNobuaki Kikkawa \Email[email protected]
\NameHiroshi Ohno \Email[email protected]
\addrToyota Central R&D Labs., Inc., 41-1, Yokomichi, Nagakute, Aichi 480-1192, Japan
\DeclareMathOperator*\argmaxargmax
\DeclareMathOperator*\KLKL
\DeclareMathOperator*\erfcerfc
\DeclareMathOperator*\ierfcierfc
Unified theory of upper confidence bound policies for bandit problems targeting total reward, maximal reward, and more
Abstract
The upper confidence bound (UCB) policy is recognized as an order-optimal solution for the classical total-reward bandit problem. While similar UCB-based approaches have been applied to the max bandit problem, which aims to maximize the cumulative maximal reward, their order optimality remains unclear. In this study, we clarify the unified conditions under which the UCB policy achieves the order optimality in both total-reward and max bandit problems. A key concept of our theory is the oracle quantity, which identifies the best arm by its highest value. This allows a unified definition of the UCB policy as pulling the arm with the highest UCB of the oracle quantity. Additionally, under this setting, optimality analysis can be conducted by replacing traditional regret with the number of failures as a core measure. One consequence of our analysis is that the confidence interval of the oracle quantity must narrow appropriately as trials increase to ensure the order optimality of UCB policies. From this consequence, we prove that the previously proposed MaxSearch algorithm satisfies this condition and is an order-optimal policy for the max bandit problem. We also demonstrate that new bandit problems and their order-optimal UCB algorithms can be systematically derived by providing the appropriate oracle quantity and its confidence interval. Building on this, we propose PIUCB algorithms, which aim to pull the arm with the highest probability of improvement (PI). These algorithms can be applied to the max bandit problem in practice and perform comparably or better than the MaxSearch algorithm in toy examples. This suggests that our theory has the potential to generate new policies tailored to specific oracle quantities.
keywords:
bandit problem, UCB policy, max bandit problem, probability of improvement, order optimality, oracle quantity1 Introduction
The bandit problem (robbins1952some), which aims to pull the best arm based on past observations, is a fundamental issue in decision making and machine learning. Despite its simplicity, it serves as a theoretical foundation for real-world applications, such as Bayesian optimization (srinivas2009gaussian) and Monte Carlo Tree Search (MCTS) (kocsis2006bandit; browne2012survey). The scope of the bandit problem extends beyond classical total reward maximization, encompassing a variety of problem settings. These include settings that aim to maximize discounted total rewards (garivier2011upper), maximize a single best reward (cicirello2005max), or identify the optimal arm (audibert2010best). The optimal policy also varies depending on differences in the reward distributions for each arm (auer2002finite; bubeck2013bandits), as well as whether the rewards are provided stochastically or adversarially (auer2002finite; auer2002nonstochastic; bubeck2012regret). Many other policies have been proposed to address these varying conditions (agrawal1995continuum; li2010contextual; yue2012k; besbes2014stochastic). Among these policies, the upper confidence bound (UCB) policy (auer2002finite; audibert2010best; garivier2011upper; kikkawa2024materials), which utilizes confidence bounds of rewards, has proven its effectiveness under various conditions in stochastic settings.
Although UCB policies have proven generally effective in stochastic bandit problems, their validity is often evaluated case by case, depending on specific contexts and conditions. As a result, when considering variant settings, it becomes unclear what is permissible and what is not, based on existing knowledge derived from standard settings. This lack of clarity poses a significant obstacle to applying UCB policies in the variant settings. By establishing unified conditions for the effectiveness of UCB policies, we aim to provide a theoretical guideline for the application of UCB policies across diverse problem settings. As a step toward a unified theory, this study provides a comprehensive proof of the effectiveness of the UCB policy by focusing on two stochastic bandit problems: the classic total-reward bandit problem, which aims to maximize cumulative total rewards (robbins1952some), and the max bandit problem (also known as the max -armed bandit problem or extreme bandit problem), which seeks to maximize the single best reward (cicirello2005max; carpentier2014extreme; david2016pac). The key in this proof is the introduction of the oracle quantity , which addresses the differences in the objectives of the bandit problems. This approach proves the effectiveness of the UCB policy not only for the total-reward and max bandit problems but also for bandit problems maximizing other targets. This flexibility of the target definition can improve practical exploration efficiency, tailoring the policy to each target. As an example, using the probability of improvement (PI) as the oracle quantity, we constructed a PIUCB policy, which aims to maximize PI from the cumulative best reward. This policy can effectively be used as one of the max bandit algorithms, which performance is at least comparable to the MaxSearch algorithm, a UCB algorithm proposed by kikkawa2024materials for the max bandit problem.
This paper is organized as follows: First, the Related Work section reviews previous research on the fundamentals of bandit problems and UCB policies. Next, the Contributions section outlines the key achievements of this study. The Preliminaries section introduces the formal definition of the bandit problems, where we employ the oracle quantity to provide a unified representation of the different targets. This approach offers a consistent framework for understanding various bandit problem settings. In the Key Claims section and the Proofs section, we present and prove a theorem and related propositions that demonstrate the requirements for establishing order optimality of the UCB policy. This theorem serves as a guide for developing UCB policies tailored to each target setting. Following this, the Discussions section confirms that some total-reward UCB and MaxSearch algorithms are order-optimal using our theorem. In addition, we construct an order-optimal PIUCB policy from our theorem using PI as the oracle quantity. We also highlight the relation between our results and the previous research on regret lower bounds (lai1985asymptotically; burnetas1996optimal; garivier2019explore). Finally, in the Summary section, we conclude this paper and discuss future prospects for applying our theorem across a wide range of scenarios.
2 Related Work
The classical total-reward bandit problem was first formulated by robbins1952some. About thirty years later, lai1985asymptotically proved that any consistent policy must pull suboptimal arms at the order of in a stochastic setting. This results established a lower bound on regret for the classical total-reward bandit problem. This lower bound was later extended for general reward distributions by burnetas1996optimal. Furthermore, garivier2019explore proposed a fundamental inequality to derive regret lower bounds in various settings.
UCB-based policies are well-known for achieving the regret that matches the order of the lower bound described above in the classical bandit problems (lai1985asymptotically; agrawal1995sample; auer2002finite). Notably, auer2002finite proposed a simple policy, and since then, the effectiveness of UCB policies has been proven individually across various contexts and conditions (audibert2010best; garivier2011upper; bubeck2013bandits), For example, the UCB policies have been shown to be order-optimal when the reward distributions have heavy tails (bubeck2013bandits).
While many bandit problems focus on the sum or expectation of rewards, a distinct class of problems, known as the max bandit problems, aims to maximize the single best reward (cicirello2005max). In these problems, regret defined in a straightforward manner approaches zero for most policies involving random selection (nishihara2016no; kikkawa2024materials). This poses significant obstacles in proving order optimality of the algorithms. However, kikkawa2024materials recently demonstrated that UCB policies can be constructed by focusing on the number of mistaken choices and employing a carefully designed oracle. Their algorithm, called MaxSearch, practically outperforms other max bandit algorithms (streeter2006simple; achab2017max), as well as the classical UCB (auer2002finite) and the best arm identification algorithms (audibert2010best). Nevertheless, the proofs for the regret lower bound and order optimality of the max bandit problem remain open problems.
In the context of Bayesian optimization, PI appears as an acquisition function alongside expected improvement (EI) and UCB (kushner1964new; jones1998efficient; srinivas2009gaussian). Although it is rare for PI to be used in the context of bandit problem, there are instances of its use in early research on the max bandit problem (streeter2006simple). One of the challenges in using PI in the bandit problem is ensuring its theoretical authenticity, and our study justifies it by using PI as the oracle quantity.
3 Contributions
Our contributions are as follows:
-
•
We present a unified representation of the bandit problem using the oracle quantity and the number of mistaken choices (hereafter referred to as failures) instead of regret.
-
•
We generalize the proof of the lower and upper bounds on failures for any consistent policy and the UCB policy, respectively, within the present framework.
-
•
We clarify the conditions under which the UCB policy becomes order-optimal in our framework, and providing the first proof of the order optimality of the MaxSearch algorithm for the max bandit problem.
-
•
We propose the PIUCB algorithm, which uses UCB of PI as the selection index, from our framework. This algorithm is at least comparable to the existing state-of-the-art max bandit algorithms, and in some cases, outperforms them.
4 Preliminaries
In our framework, bandit problems are generalized by the oracle quantity as the following definitions:
Definition 4.1 (bandit problem).
The -armed bandit game is a game where a player repeatedly pulls one of the arms over times. When a player pulls the -th arm at the -th round, they receive a reward from an unknown reward distribution , where is a parameter set for the -th arm’s distribution. The sequence of a tuple of pulling arm indices and rewards over rounds is denoted , where . The player’s goal is to pull the optimal arm at round , where is an oracle quantity. The -armed bandit problem, often referred to as the bandit problem, is to find a policy that pulls the optimal arm most frequently, where is a sequence of random variables sampled from the uniform distribution , used by the policy at each round .
Remark 4.2.
If or simply , where , then Definition 4.1 corresponds to the classical total-reward bandit problem.
Remark 4.3.
Definition 4.4 (failures).
A failure occurs when the player pulls the non-optimal arm . The number of failures, or simply called failures, until round is defined as
(1) |
where is the indicating function of the event .
Remark 4.5.
The expected number of failures, , after rounds is analogous to the expected regret, , in our framework. This approach avoids difficulties in defining regret in the max bandit problem, where at for most policies, which causes fatal problems in theoretical analysis (nishihara2016no; kikkawa2024materials). In the classical bandit settings, and are equivalent up to a constant factor because , where , and the arm indices and are the worst and sub-optimal arms, respectively.
Definition 4.6 (consistency).
A policy is said to be “consistent” if the failures satisfies , where .
Remark 4.7.
Under a “consistent” policy, the number of successful pulls satisfies , where . In other words, with infinite pulls, a consistent policy always pulls the best arm mostly. Conversely, an inconsistent policy may fail indefinitely.
Definition 4.8 (UCB policy).
Assume that the oracle quantity satisfies
(2) |
with a confidence level , where LCB is an abbreviation for lower confidence bound. A policy that pulls at each round is called a UCB policy in general.
Remark 4.9.
In the classical bandit problem, the UCB of under the sub-gaussian assumption for yields the well-known selection index (auer2002finite),
(3) |
where , , and are the sample mean, number of the -th arm pulling, and a positive constant, respectively.
Our definitions are based solely on . Consequently, the claims and algorithms in the following sections are concretely defined by specifying the form of .
5 Key Claims
Under the above definitions, we claim the following.
Theorem 5.1 (order optimality of UCB policy).
The UCB policy gives almost surely for any oracle quantity that satisfies the following conditions.
-
a)
and exist at each round , and the arms with their maxima are uniquely determined almost surely.
-
b)
There exists such that the confidence interval
(4) with , where .
-
c)
There exists a modified parameter set such that
(5) and
(6) when for all , where indicates that the distribution is absolutely continuous with respect to the distribution .
Remark 5.2.
The proof of Theorem 5.1 is almost the same as the corresponding proofs for the classical bandit problem (lai1985asymptotically; auer2002finite; bartlett2014learning; kikkawa2024materials). This proof consists of the proofs of the following two propositions.
Proposition 5.3 (upper bound of failures).
The UCB policy guarantees for any oracle quantity that satisfies conditions (a) and (b) in Theorem 5.1.
Proposition 5.4 (lower bound of failures).
For any consistent policy and oracle quantity, the failures almost surely if conditions (a) and (c) in Theorem 5.1 are satisfied.
6 Proofs
6.1 Proof for Proposition 5.3
Assume the confidence bounds are given by Equation (2) in condition (b) with a confidence level of . Then, the number of times any of the violates these confidence bounds under the UCB policy until round is given by
(7) |
under the assumption of in condition (b). In all other cases, where the confidence bounds of Equation (2) hold for all , the UCB policy pulls a non-optimal arm when at time . Then,
(8) |
must hold in this case. Thus, if Equation (4) holds, inequality (8) implies . Then, . Therefore, combining it with Equation (7), Proposition 5.3 is established since all possible cases are covered.
Remark 6.1.
The improvement of this proof over the previous studies (auer2002finite; kikkawa2024materials) lies in focusing on the width of the confidence interval and identifying the conditions under which holds, without providing an explicit expression for the confidence bounds.
6.2 Proof for Proposition 5.4
From Definition 4.1, the probability of an event at time under the original and modified distributions in condition (c) can be written as:
(9) |
(10) |
for any which can uniquely determine . These probabilities are related to each other, as shown in the following lemma.
Lemma 6.2.
Assume for all . Then,
(11) |
holds almost surely as when setting . The factor is given as
(12) |
for , where and is the Kullback-Leibler (KL) divergence of from .
Proof 6.3.
From the definitions
(13) | ||||
(14) | ||||
(15) |
for any , where . Here,
(16) |
as , as a result of the law of large numbers under the assumption of the absolute continuity. Thus, holds almost surely as , since satisfies Equation (12). Because and when , for .
Using Lemma 6.2 and the Markov inequality,