Contextual Bandit Learning with Predictable Rewards

Similar documents
New Algorithms for Contextual Bandits

Reducing contextual bandits to supervised learning

Lecture 4: Lower Bounds (ending); Thompson Sampling

New bounds on the price of bandit feedback for mistake-bounded online multiclass learning

Learning with Exploration

On the Complexity of Best Arm Identification in Multi-Armed Bandit Models

Lecture 3: Lower Bounds for Bandit Algorithms

Online Learning and Sequential Decision Making

Sparse Linear Contextual Bandits via Relevance Vector Machines

Improved Algorithms for Linear Stochastic Bandits

Slides at:

Bandit Convex Optimization: T Regret in One Dimension

Csaba Szepesvári 1. University of Alberta. Machine Learning Summer School, Ile de Re, France, 2008

The Multi-Armed Bandit Problem

Learning for Contextual Bandits

Advanced Machine Learning

A Time and Space Efficient Algorithm for Contextual Linear Bandits

COS 402 Machine Learning and Artificial Intelligence Fall Lecture 22. Exploration & Exploitation in Reinforcement Learning: MAB, UCB, Exp3

Contextual semibandits via supervised learning oracles

Lecture 10 : Contextual Bandits

Explore no more: Improved high-probability regret bounds for non-stochastic bandits

Multi-armed Bandits in the Presence of Side Observations in Social Networks

Regret Analysis of Stochastic and Nonstochastic Multi-armed Bandit Problems, Part I. Sébastien Bubeck Theory Group

THE first formalization of the multi-armed bandit problem

Explore no more: Improved high-probability regret bounds for non-stochastic bandits

Multi-armed bandit models: a tutorial

Bandits with Delayed, Aggregated Anonymous Feedback

On the Generalization Ability of Online Strongly Convex Programming Algorithms

The information complexity of sequential resource allocation

Evaluation of multi armed bandit algorithms and empirical algorithm

Lecture 5: Regret Bounds for Thompson Sampling

Practical Agnostic Active Learning

Optimal and Adaptive Online Learning

Taming the Monster: A Fast and Simple Algorithm for Contextual Bandits

Introduction to Algorithms / Algorithms I Lecturer: Michael Dinitz Topic: Intro to Learning Theory Date: 12/8/16

Combinatorial Multi-Armed Bandit: General Framework, Results and Applications

Grundlagen der Künstlichen Intelligenz

The Multi-Armed Bandit Problem

EASINESS IN BANDITS. Gergely Neu. Pompeu Fabra University

Online Regret Bounds for Markov Decision Processes with Deterministic Transitions

Bandit models: a tutorial

Informational Confidence Bounds for Self-Normalized Averages and Applications

Two optimization problems in a stochastic bandit model

Online Learning and Online Convex Optimization

On the Complexity of Best Arm Identification with Fixed Confidence

Machine Learning Theory (CS 6783)

Analysis of Thompson Sampling for the multi-armed bandit problem

Administration. CSCI567 Machine Learning (Fall 2018) Outline. Outline. HW5 is available, due on 11/18. Practice final will also be available soon.

Bayesian and Frequentist Methods in Bandit Models

Tutorial: PART 2. Online Convex Optimization, A Game- Theoretic Approach to Learning

Lecture 19: UCB Algorithm and Adversarial Bandit Problem. Announcements Review on stochastic multi-armed bandit problem

Exploration Scavenging

Beating the Hold-Out: Bounds for K-fold and Progressive Cross-Validation

FORMULATION OF THE LEARNING PROBLEM

Stat 260/CS Learning in Sequential Decision Problems. Peter Bartlett

Better Algorithms for Selective Sampling

Hybrid Machine Learning Algorithms

Anytime optimal algorithms in stochastic multi-armed bandits

Agnostic Online learnability

Understanding Generalization Error: Bounds and Decompositions

On Minimaxity of Follow the Leader Strategy in the Stochastic Setting

Full-information Online Learning

Reinforcement Learning

On the Complexity of Best Arm Identification with Fixed Confidence

Generalization bounds

Bandit Algorithms. Zhifeng Wang ... Department of Statistics Florida State University

Adaptive Online Gradient Descent

Discover Relevant Sources : A Multi-Armed Bandit Approach

Bregman Divergence and Mirror Descent

Advanced Topics in Machine Learning and Algorithmic Game Theory Fall semester, 2011/12

Learning from Logged Implicit Exploration Data

Revisiting the Exploration-Exploitation Tradeoff in Bandit Models

arxiv: v3 [cs.lg] 30 Jun 2012

Stratégies bayésiennes et fréquentistes dans un modèle de bandit

Learning Theory. Ingo Steinwart University of Stuttgart. September 4, 2013

Reinforcement Learning

The No-Regret Framework for Online Learning

Gambling in a rigged casino: The adversarial multi-armed bandit problem

Generalization theory

CS261: A Second Course in Algorithms Lecture #11: Online Learning and the Multiplicative Weights Algorithm

Lecture 7 Introduction to Statistical Decision Theory

Experts in a Markov Decision Process

arxiv: v1 [cs.lg] 1 Jun 2016

On Bayesian bandit algorithms

PAC Subset Selection in Stochastic Multi-armed Bandits

The information complexity of best-arm identification

Yevgeny Seldin. University of Copenhagen

The sample complexity of agnostic learning with deterministic labels

Online Sparse Linear Regression

Minimax Concave Penalized Multi-Armed Bandit Model with High-Dimensional Convariates

Lecture Learning infinite hypothesis class via VC-dimension and Rademacher complexity;

Learning to play K-armed bandit problems

Beat the Mean Bandit

Exponential Weights on the Hypercube in Polynomial Time

Bandits, Query Learning, and the Haystack Dimension

Regret Analysis of Stochastic and Nonstochastic Multi-armed Bandit Problems. Sébastien Bubeck Theory Group

A. Notation. Attraction probability of item d. (d) Highest attraction probability, (1) A

Doubly Robust Policy Evaluation and Learning

Exploration. 2015/10/12 John Schulman

Bandits for Online Optimization

Transcription:

Contetual Bandit Learning with Predictable Rewards Alekh Agarwal alekh@cs.berkeley.edu Miroslav Dudík mdudik@yahoo-inc.com Satyen Kale sckale@us.ibm.com John Langford jl@yahoo-inc.com Robert. Schapire schapire@cs.princeton.edu Abstract Contetual bandit learning is a reinforcement learning problem where the learner repeatedly receives a set of features (contet), takes an action and receives a reward based on the action and contet. We consider this problem under a realizability assumption: there eists a function in a (known) function class, always capable of predicting the epected reward, given the action and contet. Under this assumption, we show three things. We present a new algorithm Regressor limination with a regret similar to the agnostic setting (i.e. in the absence of realizability assumption). We prove a new lower bound showing no algorithm can achieve superior performance in the worst case even with the realizability assumption. However, we do show that for any set of policies (mapping contets to actions), there is a distribution over rewards (given contet) such that our new algorithm has constant regret unlike the previous approaches. Introduction We are interested in the online contetual bandit setting, where on each round we first see a contet X, based on which we choose an action a A, and then observe a reward r. This formalizes several natural scenarios. For eample, a common task at major internet engines is to display the best ad from a pool of options given some contet such as information about the user, the page visited, the search query issued etc. The action set consists of the candidate ads and the reward is typically binary based on whether the user clicked the displayed ad or not. Another Appearing in Proceedings of the 5 th International Conference on Artificial Intelligence and Statistics (AISTATS) 202, La Palma, Canary Islands. Volume XX of JMLR: W&CP XX. Copyright 202 by the authors. natural application is the design of clinical trials in the medical domain. In this case, the actions are the treatment options being compared, the contet is the patient s medical record and reward is based on whether the recommended treatment is a success or not. Our goal in this setting is to compete with a particular set of policies, which are deterministic rules specifying which action to choose in each contet. We note that this setting includes as special cases the classical K-armed bandit problem (Lai and Robbins, 985) and associative reinforcement learning with linear reward functions (Auer, 2003; Chu et al., 20). The performance of algorithms in this setting is typically measured by the regret, which is the difference between the cumulative reward of the best policy and the algorithm. For the setting with an arbitrary set of policies, the achieved regret guarantee is O( KT ln(n/δ)) where K is the number of actions, T is the number of rounds, N is the number of policies and δ is the probability of failing to achieve the regret (Beygelzimer et al., 200; Dudík et al., 20). While this bound has a desirably small dependence on the parameters T, N, the scaling with respect to K is often too big to be meaningful. For instance, the number of ads under consideration can be huge, and a rapid scaling with the number of alternatives in a clinical trial is clearly undesirable. Unfortunately, the dependence on K is unavoidable as proved by eisting lower bounds (Auer et al., 2003). Large literature on linear bandits manages to avoid this dependence on K by making additional assumptions. For eample, Auer (2003) and Chu et al. (20) consider the setting where the contet consists of feature vectors a R d describing each action, and the epected reward function (given a contet and action a) has the form w T a for some fied vector w R d. Dani et al. (2008) consider a continuous action space with a R d, without contets, with a linear epected reward w T a, which is generalized by Filippi et al. (200) to σ(w T a) with a known Lipschitzcontinuous link function σ. A striking aspect of the linear and generalized linear setting is that while the regret grows 9

Contetual Bandit Learning with Predictable Rewards rapidly with the dimension d, it grows either only gently with the number of actions K (poly-logarithmic for Auer, 2003), or is independent of K (Dani et al., 2008; Filippi et al., 200). In this paper, we investigate whether a weaker dependence on the number of actions is possible in more general settings. Specifically, we omit the linearity assumption while keeping the realizability i.e., we still assume that the epected reward can be perfectly modeled, but do not require this to be a linear or a generalized linear model. We consider an arbitrary class F of functions f : (X, A) 0, that map a contet and an action to a real number. We interpret f(, a) as a predicted epected reward of the action a on contet and refer to functions in F as regressors. For eample, in display advertising, the contet is a vector of features derived from the tet and metadata of the webpage and information about the user. The action corresponds to the ad, also described by a set of features. Additional features might be used to model interaction between the ad and the contet. A typical regressor for this problem is a generalized linear model with a logistic link, modeling the probability of a click. The set of regressors F induces a natural set of policies Π F containing maps π f : X A defined as π f () = argma a f(, a). We make the assumption that the epected reward for a contet and action a equals f (, a) for some unknown function f F. The question we address in this paper is: Does this realizability assumption allow us to learn faster? We show that for an arbitrary function class, the answer to the above question is no. The K dependence in regret is in general unavoidable even with the realizability assumption. Thus, the structure of linearity or controlled non-linearity was quite important in the past works. Given this answer, a natural question is whether it is at least possible to do better in various special cases. To answer this, we create a new natural algorithm, Regressor limination (R), which takes advantage of realizability. Structurally, the algorithm is similar to Policy limination (P) of Dudík et al. (20), designed for the agnostic case (i.e, the general case without realizability assumption). While P proceeds by eliminating poorly performing policies, R proceeds by eliminating poorly predicting regressors. However, realizability assumption allows much more aggressive elimination strategy, different from the strategy used in P. The analysis of this elimination strategy is the key technical contribution of this paper. 20 The general regret guarantee for Regressor limination is O( KT ln(nt/δ)), similar to the agnostic case. However, we also show that for all sets of policies Π there eists a set of regressors F such that Π = Π F and the regret of Regressor limination is O(ln(N/δ)), i.e., independent of the number of rounds and actions. At the first sight, this seems to contradict our worst-case lower bound. This apparent parado is due to the fact that the same set of policies can be generated by two very different sets of regressors. Some regressor sets allow better discrimination of the true reward function, whereas some regressor sets will lead to the worst-case guarantee. The remainder of the paper is organized as follows. In the net section we formalize our setting and assumptions. Section 3 provides our algorithm which is analyzed in Section 4. In Section 5 we present the worst-case lower bound, and in Section 6, we show an improved dependence on K in favorable cases. Our algorithm assumes the eact knowledge of the distribution over contets (but not over rewards). In Section 7 we sketch how this assumption can be removed. Another major assumption is the finiteness of the set of regressors F. This assumption is more difficult to remove, as we discuss in Section 8. 2 Problem Setup We assume that the interaction between the learner and nature happens over T rounds. At each round t, nature picks a contet t X and a reward function r t : A 0, sampled i.i.d. in each round, according to a fied distribution D(, r). We assume that D() is known (this assumption is removed in Section 7), but D(r ) is unknown. The learner observes t, picks an action a t A, and observes the reward for the action r t (a t ). We are given a function class F : X A 0, with F = N, where F is the cardinality of F. We assume that F contains a perfect predictor of the epected reward: Assumption (Realizability). There eists a function f F such that r r(a) = f (, a) for all X, a A. We recall as before that the regressor class F induces the policy class Π F containing maps π f : X A defined by f F as π f () = argma a f(, a). The performance of an algorithm is measured by its epected regret relative to the best fied policy: regret T = sup π f Π F T f ( t, π f ( t ) ) f ( t, a t ). t= By definition of π f, this is equivalent to regret T = T f ( t, π f ( t ) ) f ( t, a t ). t= 3 Algorithm Our algorithm, Regressor limination, maintains a set of regressors that accurately predict the observed rewards. In each round, it chooses an action that sufficiently eplores among the actions represented in the current set of regressors (Steps 2). After observing the reward (Step 3), the inaccurate regressors are eliminated (Step 4).

Agarwal, Dudík, Kale, Langford and Schapire Sufficient eploration is achieved by solving the conve optimization problem in Step. We construct a distribution P t over current regressors, and then act by first sampling a regressor f P t and then choosing an action according to π f. Similarly to the Policy limination algorithm of Dudík et al. (20), we seek a distribution P t such that the inverse probability of choosing an action that agrees with any policy in the current set is in epectation bounded from above. Informally, this guarantees that actions of any of the current policies are chosen with sufficient probabilities. Using this construction we relate the accuracy of regressors to the regret of the algorithm (Lemma 4.3). A priori, it is not clear whether the constraint (3.) is even feasible. We prove feasibility by a similar argument as in Dudík et al. (20) (see Lemma A. in Appendi A). Compared with Dudík et al. (20) we are able to obtain tighter constraints by doing a more careful analysis. Our elimination step (Step 4) is significantly tighter than a similar step in Dudík et al. (20): we eliminate regressors according to a very strict O(/t) bound on the suboptimality of the least squares error. Under the realizability assumption, this stringent constraint will not discard the optimal regressor accidentally, as we show in the net section. This is the key novel technical contribution of this work. Replacing D() in the Regressor limination algorithm with the empirical distribution over observed contets is straightforward, as was done in Dudík et al. (20), and is discussed further in Section 7. 4 Regret Analysis Here we prove an upper bound on the regret of Regressor limination. The proved bound is no better than the one for eisting agnostic algorithms. This is necessary, as we will see in Section 5, where we prove a matching lower bound. Theorem 4.. For all sets of regressors F with F = N and all distributions D(, r), with probability δ, the regret of Regressor limination is O( KT ln(nt/δ)). Proof. By Lemma 4. (proved below), in round t if we sample an action by sampling f from P t and choosing π f ( t ), then the epected regret is O( K ln(nt/δ)/t) with probability at least δ/2t 2. The ecess regret for sampling a uniform random action is at most µ T per round. Summing up over all the T rounds and taking a union bound, the total epected regret is O ( KT ln(nt/δ) ) with probability at least δ. Further, the net regret is a martingale; hence the Azuma-Hoeffding inequality with range 0, applies. So with probability at least δ we have a regret of O ( KT ln(nt/δ) + T ln(/δ) ) = O ( KT ln(nt/δ) ). 2 Algorithm Regressor limination Input: a set of reward predictors F = {f : (X, A) 0, } distribution D over contets, confidence parameter δ. Notation: π f () := argma a f(, a ). ˆR t (f) := t t t = (f( t, a t ) r t (a t ))2. For F F, define A(F, ) := {a A : π f () = a for some f F } µ := min{/2k, / T }. For a distribution P on F F, define conditional distribution P ( ) on A as: w.p. ( µ), sample f P and return π f (), and w.p. µ, return a uniform random a A(F, ). δ t = δ/2nt 3 log 2 (t), for t =, 2,..., T. Algorithm: F 0 F For t =, 2,..., T :. Find distribution P t on F t such that f F t : P t(π A(Ft, ) (3.) f () ) 2. Observe t and sample action a t from P t( t ). 3. Observe r t (a t ). 4. Set { F t = f F t : ˆRt (f) < min (f ) + 8 ln(/δ } t) f F t ˆRt t Lemma 4.. With probability at least δ t Nt log 2 (t) δ/2t 2, we have:. f F t. 2. For any f F t, r(π f ()) r(π f ()),r 200K ln(/δt ). t Proof. Fi an arbitrary function f F. For every round t, define the random variable Y t = (f( t, a t ) r t (a t )) 2 (f ( t, a t ) r t (a t )) 2. Here, t is drawn from the unknown data distribution D, r t is drawn from the reward distribution conditioned on t, and a t is drawn from P t (which is defined conditioned on the choice of t and is independent of r t ). Note that this random variable is well-defined for all functions f F, not just the ones in F t.

Contetual Bandit Learning with Predictable Rewards Let t and Var t denote the epectation and variance conditioned on all the randomness up to round t. Using a form of Freedman s inequality from Bartlett et al. (2008) (see Lemma B.) and noting that Y t, we get that with probability at least δ t log 2 (t), we have t Y t t = t = Y t 4 t Var t Y t ln(/δ t ) + 2 ln(/δ t ). t = From Lemma 4.2, we see that Var t Y t 4 t Y t so t Y t t = t = Y t 8 t t Y t ln(/δ t ) + 2 ln(/δ t ). t = t For notational convenience, define X = t = t Y t, Z = t t = Y t, and C = ln(/δ t ). The above inequality is equivalent to: X 2 Z 8CX + 2C 2 (X 4C) 2 Z 8C 2. This gives Z 8C 2. Since Z = t( ˆR t (f) ˆR t (f )), we get that ˆR t (f ) ˆR t (f) + 8C2. t By a union bound, with probability at least δ t Nt log 2 (t), for all f F and all rounds t t, we have ˆR t (f ) ˆR t (f) + 8 ln(/δ t) t and so f is not eliminated in any elimination step and remains in F t. Furthermore, suppose f is also not eliminated and survives in F t. Then we must have ˆR t (f) ˆR t (f ) 8C 2 /t, or in other words, Z 8C 2. Thus, (X 4C) 2 36C 2, which implies that X 2 00C 2, and hence: t Y t 00 ln(/δ t ). (4.) t = By Lemma 4.3 and since P t is measurable with respect to the past sigma field up to time t, for all t t we have r(π f ()) r(π f ()) 2 2K t Y t.,r t,r t,a t Summing up over all t t, and using (4.) along with Jensen s inequality we get that 200K r(π ln(/δt ) f ()) r(π f ()).,r t 22 Lemma 4.2. Fi a function f F. Suppose we sample, r from the data distribution D, and an action a from an arbitrary distribution such that r and a are conditionally independent given. Define the random variable Y = (f(, a) r(a)) 2 (f (, a) r(a)) 2. Then we have Y = (f(, a) f (, a)) 2,r,a,a Var Y 4 Y.,r,a,r,a Proof. Using shorthands f a for f(, a) and r a for r(a), we can rearrange the definition of Y as Hence, we have Y = (f a f a)(f a + f a 2r a ). (4.2) Y = (f a fa)(f a + fa 2r a ),r,a,r,a =,a r (f a f a)(f a + f a 2r a ) ( ) = (f a f,a a) f a + fa 2 r a r =,a (fa f a) 2, proving the first part of the lemma. From (4.2), noting that f a, f a, r a are between 0 and, we obtain Y 2 (f a f a) 2 (f a + f a 2r a ) 2 4(f a f a) 2, yielding the second part of the lemma: Var Y Y 2 4 (fa f,r,a,r,a,r,a a) 2 = 4,r,a Y. Net we show how the random variable Y defined in Lemma 4.2 relates to the regret in a single round: Lemma 4.3. In the setup of Lemma 4.2, assume further that the action a is sampled from a conditional distribution p( ) which satisfies the following constraint, for f = f and f = f : K. (4.3) p(π f () ) Then we have r ( π f () ) r ( π f () ) 2,r 2K,r,a Y. This lemma is essentially a refined form of theorem 6. in Beygelzimer and Langford (2009) which analyzes the regression approach to learning in contetual bandit settings.

Agarwal, Dudík, Kale, Langford and Schapire Proof. Throughout, we continue using the shorthand f a for f(, a). Given a contet, let ã = π f () and a = π f (). Define the random variable = r r ( π f () ) r ( π f () ) = f a f ã. Note that 0 because f prefers a over ã for contet. Also we have f ã f a since f prefers ã over a for contet. Thus, f ã f ã + f a f a. (4.4) As in proof of Lemma 4.2, Y = (fa fa) 2 r,a a p(ã )(f ã f ã) 2 +p(a )(f a f a )2 p(ã )p(a ) p(ã ) + p(a ) 2. (4.5) The last inequality follows by first applying the chain a 2 + by 2 = ab( + y)2 + (a by) 2 a + b ab ( + y)2 a + b (valid for a, b > 0), and then applying inequality (4.4). For convenience, define Q = p(ã )p(a ) p(ã ) + p(a ), i.e., = Q p(ã ) + p(a ). Now, since p satisfies the constraint (4.3) for f = f and f = f, we conclude that = + Q p(ã ) p(a 2K. (4.6) ) We now have 2 = Q Q 2 Q 2 Q 2K Y,,r,a where the first inequality follows from the Cauchy- Schwarz inequality and the second from the inequalities (4.5) and (4.6). 5 Lower bound Here we prove a lower bound showing that the realizability assumption is not enough in general to eliminate a dependence on the number of actions K. The structure of this proof is similar to an earlier lower bound (Auer et al., 2003) differing in two ways: it applies to regressors of the sort we consider, and we work N, the number of regressors, into the lower bound. Since for every policy there eists a regressor with argma on that regressor realizing the policy, this lower bound also applies to policy based algorithms. 23 Theorem 5.. For every N and K such that ln N/ ln K T, and every algorithm A, there eists a function class F of cardinality at most N and a distribution D(, r) for which the realizability assumption holds, but the epected regret of A is Ω( KT ln N/ ln K). Proof. Instead of directly selecting F and D for which the epected regret of A is Ω( KT ln N/ ln K), we create a distribution over instances (F, D) and show that the epected regret of A is Ω( KT ln N/ ln K) when the epectation is taken also over our choice of the instance. This will immediately yield a statement of the theorem, since the algorithm must suffer at least this amount of regret on one of the instances. The proof proceeds via a reduction to the construction used in the lower bound of Theorem 5. of Auer et al. (2003). We will use M different contets for a suitable number M. To define the regressor class F, we begin with the policy class G consisting of all the K M mappings of the form g : X A, where X = {, 2,..., M} and A = {, 2,..., K}. We require M to be the largest integer such that K M N, i.e., M = ln N/ ln K. ach mapping g G defines a regressor f g F as follows: { /2 + ɛ if a = f g (, a) = /2 otherwise. The rewards are generated by picking a function f F uniformly at random at the beginning. quivalently, we choose a mapping g that independently maps each contet X to a random action a A, and set f = f g. In each round t, a contet t is picked uniformly from X. For any action a, a reward r t (a) is generated as a {0, } Bernoulli trial with probability of being equal to f(, a). Now fi a contet X. We condition on all of the randomness of the algorithm A, the choices of the contets t for t =, 2,..., T, and the values of g( ) for. Thus the only randomness left is in the choice of and the realization of the rewards in each round. Let P denote the reward distribution where the rewards of any action a for contet are chosen to be {0, } uniformly at random (the rewards for other contets are still chosen according to f(, a), however), and let denote the epectation under P. Let T be the rounds t where the contet t is. Now fi an action a A and let S a be a random variable denoting the number of rounds t T when A chooses a t = a. Note that conditioned on = a, the random variable S a counts the number of rounds in T that A chooses the optimal action a. We use a corollary of Lemma A. in Auer et al. (2003): Corollary 5. (Auer et al., 2003). Conditioned on the choices of the contets t for t =, 2,..., T, and the val-

Contetual Bandit Learning with Predictable Rewards ues of g( ) for, we have S a = a S a + T 2ɛ 2 S a. The proof uses the fact that when = a, rewards chosen using P are identical to those from the true distribution ecept for the rounds when A chooses the action a. Thus, if N is a random variable that counts the number the rounds in T that A chooses the optimal action for (without conditioning on ), we have N = S S + T S + T by Jensen s inequality. Now note that S = = t T 2ɛ 2 S 2ɛ 2 S, t T {a t = } {a t = } t T = = T K K. The third equality follows because is independent of the choices of the contets t for t =, 2,..., T, and g( ) for, and its distribution is uniform on A. Thus N T K + T 2ɛ 2 T K. Since in the rounds in T \ N, the algorithm A suffers an epected regret of ɛ, the epected ( regret of A over all the rounds in T is at least Ω ɛ T ɛ2 K T ). 3/2 Note that this lower bound is independent of the choice of g( ) for. Thus, we can remove the conditioning on g( ) for and conclude that only conditioned on the choices of the contets t for t =, 2,..., T, the epected regret ( of the algorithm over all the rounds in T is at least Ω ɛ T ɛ2 K T ). 3/2 Summing up over all, and removing the conditioning on the choices of the contets t for t =, 2,..., T by taking an epectation, we get the following lower bound on the epected regret of A: Ω ( X (ɛ T ɛ2 K T 3/2 ) ). Note that T is distributed as Binomial(T, /M). Thus, 24 T = T/M. Furthermore, by Jensen s inequality T 3/2 T 3 ( ) /2 T 3T (T ) T (T )(T 2) = + M M 2 + M 3 5T 3/2 M, 3/2 as long as M T. Plugging these bounds in, the lower bound on the epected regret becomes ) Ω (ɛt ɛ2 T 3/2. KM Choosing ɛ = Θ ( KM/T ), we get that the epected regret of A is lower bounded by Ω( KMT ) = Ω( KT ln N/ ln K). 6 Analysis of nontriviality Since the worst-case regret bound of our new algorithm is the same as for agnostic algorithms, a skeptic could conclude that there is no power in the realizability assumption. Here, we show that in some cases, realizability assumption can be very powerful in reducing regret. Theorem 6.. For any algorithm A working with a set of policies (rather than regressors), there eists a set of regressors F and a distribution D satisfying the realizability assumption such that the regret of A using the set Π F is Ω( T K ln N), but the epected regret of Regressor limination using F is at most O ( ln(n/δ) ). Proof. Let F be the set of functions and D the data distribution that achieve the lower bound of Theorem 5. for the algorithm A. Using Lemma 6. (see below), there eists a set of functions F such that Π F = Π F and the epected regret of Regressor limination using F is at most O ( ln(n/δ) ). This set of functions F and distribution D satisfy the requirements of the theorem. Lemma 6.. For any distribution D and a set of policies Π containing the optimal policy, there eists a set of functions F satisfying the realizability assumption, such that Π = Π F and the regret of regressor elimination using F is at most O ( ln(n/δ) ). Proof. The idea is to build a set of functions F such that Π = Π F, and for the optimal policy π the corresponding function f eactly gives the epected rewards for each contet and a, but for any other policy π the corresponding function f gives a terrible estimate, allowing regressor elimination to eliminate them quickly. The construction is as follows. For π, we define the function f as f (, a) =,rr(a). By optimality of π,

Agarwal, Dudík, Kale, Langford and Schapire π f = π. For every other policy π we construct an f such that π = π f but for which f(, a) is a very bad estimate of,rr(a) for all actions a. Fi and consider two cases: the first is that r r(π()) > 0.75 and the other is that r r(π()) 0.75. In the first case, we let f(, π()) = 0.5. In the second case we let f(, π()) =.0. Now consider each other action a in turn. If r r(a ) > 0.25 then we let f(, a ) = 0, and if r r(a ) 0.25 we let f(, a ) = 0.5. The regressor elimination algorithm eliminates regressor with a too-large squared loss regret. Now fi any policy π π, and the corresponding f, define, as in the proof of Lemma 4., the random variable Y t = (f( t, a t ) r t (a t )) 2 (f ( t, a t ) r t (a t )) 2. Note that t Y t = (f( t, a t ) f ( t, a t )) 2 t,a t 20, (6.) since for all (, a), (f(, a) f (, a)) 2 20 by construction. This shows that the epected regret is significant. Now suppose f is not eliminated and remains in F t. Then by equation 4. we get: t 20 t Y t 00 ln(/δ t ). t = The above bound holds with probability δ t Nt log 2 (t) uniformly for all f F t. Using the choice of δ t = δ/2nt 3 log 2 (t), we note that the bound fails to hold when t > 0 6 ln(n/δ). Thus, within 0 6 ln(n/δ) rounds all suboptimal regressors are eliminated, and the algorithm suffers no regret thereafter. Since the rewards are bounded in 0,, the total regret in the first 0 6 ln(n/δ) rounds can be at most 0 6 ln(n/δ), giving us the desired bound. 7 Removing the dependence on D While Algorithm is conceptually simple and enjoys nice theoretical guarantees, it has a serious drawback that it depends on the distribution D from which the contets t s are drawn in order to specify the constraint (3.). A similar issue was faced in the earlier work of Dudík et al. (20), where they replace the epectation under D with a sample average over the contets observed. We now discuss a similar modification for Algorithm and give a sketch of the regret analysis. The key change in Algorithm is to replace the constraint (3.) with the sample version. Let H t = {, 2,..., t }, and denote by H t the act of selecting a contet from H t uniformly at random. Now we 25 pick a distribution P t on F t such that f F t : Ht P t(π f () ) A(Ft, ) Ht (7.) Since Lemma A. applies to any distribution on the contets, in particular, the uniform distribution on H t, this constraint is still feasible. To justify this sample based approimation, we appeal to Theorem 6 of Dudík et al. (20) which shows that for any ɛ (0, ) and t 6K ln(8kn/δ), with probability at least δ D P t(π f () ) ( + ɛ) Ht P t(π + 7500 f () ) ɛ 3 K. Using quation (7.), since A(F t, t ) K, we get D P t(π f () ) 7525K, using ɛ = 0.999. The remaining analysis of the algorithm remains the same as before, ecept we now apply Lemma 4.3 with a worse constant in the condition (4.3). 8 Conclusion The included results gives us a basic understanding of the realizable assumption setting: it can, but does not necessarily, improve our ability to learn. We did not address computational compleity in this paper. There are some reasons to be hopeful however. Due to the structure of the realizability assumption, an eliminated regressor continues to have an increasingly poor regret over time, implying that it may be possible to avoid the elimination step and simply restrict the set of regressors we care about when constructing a distribution. A basic question then is: can we make the formation of this distribution computationally tractable? Another question for future research is the etension to infinite function classes. One would epect that this just involves replacing the log cardinality with something like a metric entropy or Rademacher compleity of F. This is not completely immediate since we are dealing with martingales, and direct application of covering arguments seems to yield a suboptimal O(/ t) rate in Lemma 4.. tending the variance based bound coming from Freedman s inequality from a single martingale to a supremum over function classes would need a Talagrand-style concentration inequality for martingales which is not available in the literature to the best of our knowledge. Understanding this issue better is an interesting topic for future work.

Contetual Bandit Learning with Predictable Rewards Acknowledgements This research was done while AA, SK and RS were visiting Yahoo!. References P. Auer, N. Cesa-Bianchi, Y. Freund, and R.. Schapire. The nonstochastic multiarmed bandit problem. SIAM J. Comput., 32():48 77, 2003. Peter Auer. Using confidence bounds for eploitationeploration trade-offs. J. Mach. Learn. Res., 3:397 422, March 2003. P. L. Bartlett, V. Dani, T. P. Hayes, S. Kakade, A. Rakhlin, and A. Tewari. High-probability regret bounds for bandit online linear optimization. In COLT, 2008. A. Beygelzimer and J. Langford. The offset tree for learning with partial labels. In KDD, 2009. A. Beygelzimer, J. Langford, L. Li, L. Reyzin, and R.. Schapire. An optimal high probability algorithm for the contetual bandit problem. CoRR, 200. URL http: //ariv.org/abs/002.4058. W. Chu, L. Li, L. Reyzin, and R. Schapire. Contetual bandits with linear payoff functions. In AISTATS, 20. V. Dani, T. P. Hayes, and S. M. Kakade. Stochastic linear optimization under bandit feedback. In Proceedings of the 2st Annual Conference on Learning Theory, 2008. M. Dudík, D. Hsu, S. Kale, N. Karampatziakis, J. Langford, L. Reyzin, and T. Zhang. fficient optimal learning for contetual bandits. In UAI, 20. S. Filippi, O. Cappé, A. Garivier, and Cs. Szepesvári. Parametric bandits: The generalized linear case. In NIPS, 200. T. L. Lai and Herbert Robbins. Asymptotically efficient adaptive allocation rules. Advances in Applied Mathematics, 6:4 22, 985. Maurice Sion. On general minima theorems. Pacific J. Math., 8():7 76, 958. A Feasibility Lemma A.. There eists a distribution P t on F t satisfying the constraint (3.). Proof. Let t refer to the space of all distributions on F t. We observe that t is a conve, compact set. For a distribution Q t, define the conditional distribution Q( ) on A as sample f Q, and return π f (). Note that Q (a ) = ( µ) Q(a ) + µ/k, where K := A(F t, ) for notational convenience. The feasibility of constraint (3.) can be written as min ma P t t f F t P t(π A(F t, ). f () ) 26 The LHS is equal to min ma P t t Q t f F t Q(f) P t(π, f () ) where we recall that P t is the distribution induced on A by P t as before. The function Q(f) P t(π f () ) f F t is linear (and hence concave) in Q and conve in P t. Applying Sion s Minima Theorem (stated below as Theorem A.), we see that the LHS is equal to Q(f) ma min Q t P t t P t(π f () ) ma Q t = ma Q t = ma Q t = ma Q t f F t f F t Q(f) Q (π f () ) a A(F t,) f F t :π f ()=a a A(F t,) µ ma Q t K. Q(a ) Q (a ) a A(F t,) Q(f) Q (a ) µ K Q (a ) The last inequality uses the fact that for any distribution P on {, 2,..., K}, K i= /P (i) is minimized when all P (i) equal /K. Hence the constraint is always feasible. Theorem A. (see Theorem 3.4 of Sion, 958). Let U and V be compact and conve sets, and φ : U V R a function which for all v V is conve and continuous in u and for all u U is concave and continuous in v. Then B min ma u U v V φ(u, v) = ma min v V u U Freedman-style Inequality φ(u, v). Lemma B. (see Bartlett et al., 2008). Suppose X, X 2,..., X T is a martingale difference sequence with X t b for all t. Let V = T t= Var tx t be the sum of conditional variances. Then for any δ < /e 2, with probability at least log 2 (T )δ we have T t= X t 4 V ln(/δ) + 2b ln(/δ).