Introduction to Bandit Algorithms. Introduction to Bandit Algorithms
|
|
- Ella Farmer
- 6 years ago
- Views:
Transcription
1
2 Stochastic K-Arm Bandit Problem Formulation Consider K arms (actions) each correspond to an unknown distribution {ν k } K k=1 with values bounded in [0, 1]. At each time t, the agent pulls an arm I t {1,..., K} and observes a reward x t ν It (i.i.d. sample from ν It ). The objective is to maximize the expected sum of rewards. Notations mean of each arm: mean of the best arm: µ k = E X νk [X] µ = max k µ k
3 Stochastic K-Arm Bandit To evaluate the performance of a strategy Cumulative Regret R n = nµ x t Objective: find a strategy with small expected cumulative regret E [ R n ] Note: E [ ] R n = nµ E [ n ] [ K µ It = E T k (n)(µ µ k ) ] = E [ K ] T k (n) k, k=1 k=1 T k (n) = n 1{I t = k} is the total number of times that arm k has been pulled up to time n k = µ µ k is the gap between the optimal arm and arm k
4 UCB Strategy Upper confidence bound (UCB) strategy selects an arm at time t that I t = arg max B t,tk (t 1)(k), B t,s (k) = ˆµ k,s + k s ˆµ k,s = 1 s s i=1 x k,i is the empirical mean of arm k at time s UCB is a strategy based on the optimism in the face of uncertainty" principle B t,tk(t 1)(k) represents an upper-bound of µ k with high probability
5 Chernoff-Hoeffding Inequality Proposition Let X i [a i, b i ] be independent random variables with µ i = E[X i ]. Then we have ( ) ( 2ɛ 2 ) P X i µ i ɛ 2 exp n i=1 (b i a i ) 2 i=1
6 Proof of Chernoff-Hoeffding Inequality P ( n X i µ i ɛ ) = P ( exp(s X i µ i ) exp(sɛ) ) (a) exp( sɛ)e [ exp(s i=1 (b) = exp( sɛ) i=1 n i=1 E [ exp ( s(x i µ i ) )] (c) exp( sɛ) = exp ( sɛ + s 2 (b i a i ) 2 /8 ) i=1 X i µ i ) ] i=1 n exp ( s 2 (b i a i ) 2 /8 ) i=1 By selecting s = 4ɛ/ n i=1 (b i a i ) 2, we obtain P ( n i=1 X i µ i ɛ ) ( 2ɛ exp 2 ni=1 ). (b i a i ) 2 The result follows by repeating the same procedure for P ( n i=1 X i µ i ɛ ). (a) by Markov inequality, (b) by independence of the random variables, and (c) by Hoeffding inequality.
7 Chernoff-Hoeffding Inequality & UCB Strategy Using Chernoff-Hoeffding inequality, we have P ( 1 s ) ( s X i µ ɛ exp( 2sɛ 2 1 ), P s i=1 ) s X i µ ɛ exp( 2sɛ 2 ) (1) i=1 if we set ɛ = s, Eq. 1 may be written as ( ) P ˆµ k,s + µ k exp( 4 log t) = t 4 s ( ) P ˆµ k,s µ k exp( 4 log t) = t 4 s (2)
8 UCB Regret Bound Proposition Each sub-optimal arm k is pulled on average at most E[T k (n)] 8 log n 2 k + π2 3 times. As a result, the expected cumulative regret of UCB is bounded as E[R n ] = k k E[T k (n)] 8 k, k >0 log n k + Kπ2 3 The expected cumulative regret of UCB is logarithmic in n
9 Proof of UCB Regret Bound (I) Gap of arm k: k = µ µ k Idea of the proof 1 To show that each sub-optimal arm k cannot be pulled more than 8 log n plus a small constant value 2 k or 2 To show that the event that a sub-optimal arm k is pulled at a time t at which T k (t 1) 8 log t, has a small probability 2 k
10 Proof of UCB Regret Bound (II) Using Eq. 2, we may write s, t, k µ k ˆµ k,s w.p. 1 t 4 (3) s µ k + ˆµ k,s w.p. 1 t 4 (4) s If a sub-optimal arm k is pulled at time t, we know that B t,tk (t 1)(k) B t,tk (t 1)(k ) ˆµ k,tk (t 1) + T k (t 1) ˆµ k,t k (t 1) + T k (t 1) (5) If Eqs. 3 and 4 hold, then by upper-bounding the left-hand-side of Eq. 5 and lower-bounding the right-hand-side of Eq. 5, we have µ k + 2 T k (t 1) µ T k (t 1) 8 log t 2 k
11 Proof of UCB Regret Bound (III) For any integer u, we may write T k (n) u + 1 { I t = k; T k (t 1) u } t=u+1 u + 1 { s : u < s t, s : 1 s t, B t,s(k) B t,s (k ) } t=u+1 u + t t=u+1 s=u+1 s =1 We set u = 8 log n 2, therefore we have k t 1 { B t,s(k) B t,s (k ) } (6) s : u < s t s > u = 8 log n 2 k 8 log t 2 k µ k + 2 < µ (7) s It is easy to show that the event { B t,s(k) B t,s (k ) } and Eq. 7 imply that either µ s > ˆµ k,s or ˆµ k,s > µ k + s
12 Proof of UCB Regret Bound (IV) So Eq. 6 may be rewritten as T k (n) 8 log n 2 k [ { } { }] t t + 1 µ s > ˆµ k,s + 1 ˆµ k,s > µ k + s t=u+1 s=u+1 s =1 (8) Taking the expectation from both sides of Eq. 8, we obtain E [ T k (n) ] [ { 8 log n } t t 2 + P µ k s > ˆµ k,s t=u+1 s=u+1 s =1 { } ] t t + P ˆµ k,s > µ k + s s=u+1 s =1 = 8 log n [ t t t t 2 + t 4 + t 4] k t=u+1 s=u+1 s =1 s=u+1 s =1 8 log n 2 k + t=u+1 [ 2 t s=1 t 3] 8 log n 2 k + 2 t 2 8 log n 2 k + π2 3
13 UCB Regret Bound Proposition We have the following uniform bound on the expected cumulative regret E[R n ] 8Kn ( log n + π2 ) 3 Proof: Using Cauchy-Schwarz, we may write E[R n] = k k E [ T k (n) ] = k k E [ T k (n) ] E [ T k (n) ] 2 k E[ T k (n) ] E [ T k (n) ] k 8Kn ( log n + π2 3 ) k
14 UCB Regret: Lower-Bound Asymptotic lower-bound (for a rich class of distributions) [Lai & Robbins, 1985] of the form E [ T k (n) ] lim sup n log n 1 KL(ν k ν ) where KL(ν ν ) = dν log(dν/dν ), and thus, E[R n ] = Ω(log n) Non-asymptotic minimax lower-bound inf sup R n = Ω( nk) alg. prob.
15 Adversarial Bandit Setting: The rewards are not necessarily i.i.d., they are arbitrarily selected by an adversary. In this case, we cannot hope to perform as well as if we have known the rewards in advance, because the adversary does not want to be revealed. Therefore, we compare the performance of our algorithm with the performance obtained by a class of strategies, and try to be almost as good as the best strategy in that class. We may consider two types of problems according to the information received at each round: 1) full information, and 2) bandit or partial information. At each time t = 1,..., n the adversary selects rewards x t(1),..., x t(k) (the reward values are in [0, 1]) without revealing them to the player the player selects an arm I t Full Information: the player observes the rewards of all the arms: x t(k), k = 1,..., K Bandit Information: the player observes the reward of the selected arm x t(i t)
16 Adversarial Bandit If we compare the performance of our algorithm with the class of constant strategies, the regret vs. the constant strategy k is defined as R n(k) = x t(k) x t(i t). The player s strategy can be stochastic, in which case we consider the expected regret of the best constant strategy R n = max E[ R ] n(k), 1 k K and the goal is to design an algorithm that is good for all reward sequences generated by the adversary, that is sup x 1,...,x n R n be small.
17 Adversarial Bandit - Full Information Exponentially Weighted Forecaster (EWF) Initialization: w 1 (k) = 1 for all k = 1,..., K At each time t = 1,..., n: the player selects an arm I t p t, where p t (k) = w t(k) K i=1 w t(i) with w t (k) = e η t 1 s=1 xs(k), η is the parameter of the algorithm.
18 Exponentially Weighted Forecaster (EWF) Proposition (EWF Regret Bound) Let η 1. Then the regret of EWF is bounded as R n log K η + ηn 8. 8 log K If we select η = n, the regret of EWF is bounded as n log K R n. 2 Remarks: The lower bound for this problem is of the same order R n = Ω( n log K). The number of arms appears logarithmic in the regret bound. It is possible to have an anytime algorithm (the horizon n is not ( ) necessarily known) by setting η t = O. log K t
19 EWF Regret Bound Proof: Let W t = K k=1 wt(k), then we may write Therefore, W t+1 K = W t k=1 w t(k) e ηxt(k) W t = k=1 [ ( = e ηe[xt(k)] E I pt e η x t(i) E J pt [x t(j)] e ηe[xt(i)] e η2 /8 Now for any k = 1,..., K, we may write log W n+1 W 1 K p t(k) e ηxt(k) [ = E I pt e ηx ] t(i) )] log W n+1 ηe [ n x ] t(i) + nη2 W 1 8. k=1 using Hoeffding s inequality K = log e η n x t(k) log K η x t(k) log K. So for any k = 1,..., K, we have E[R n(k)] = x t(k) E [ n x ] t(i t) log K η + nη 8.
20 Adversarial Bandit - Bandit Information Exploration-Exploitation using Exponential Weights (EXP3) Initialization: w 1 (k) = 1 for all k = 1,..., K At each time t = 1,..., n: the player selects an arm I t p t, where w t (k) p t (k) = (1 β) K i=1 w t(i) + β K with w t (k) = e η t 1 s=1 xs(k), where x s (k) = xs(k) p s(k) 1{I s = k}. η > 0 and β > 0 are the parameters of the algorithm.
21 EXP3 Proposition (EXP3 Regret Bound) Let η 1 and β = ηk. Then the regret of EXP3 is bounded as R n log K η + (e 1)ηnK. If we select η = log K (e 1)nK, the regret of EXP3 is bounded as R n 2.63 nk log K. Remarks: The lower bound for this problem is R n 1 20 nk.
22 EXP3 Regret Bound I Proof: Let W t = K k=1 wt(k). Note that E I s p s [ x s(k)] = K xs(k) i=1 ps(i) 1{i = k} = xs(k) and p s(k) E Is ps [ x s(i s)] = K xs(i) i=1 ps(i) K. Therefore, we have p s(i) W t+1 K = W t k=1 (a) K k= β w t(k) e η xt(k) W t = K k=1 p t(k) β/k 1 β e η xt(k) p t(k) β/k ( 1 + η xt(k) + (e 2)η 2 x t(k) 2) 1 β K p ( t(k) η x t(k) + (e 2)η 2 x t(k) 2) k=1 (a) This is because η x t(k) ηk/β and e x 1 + x + (e 2)x 2 for x 1.
23 EXP3 Regret Bound II Proof: log W t+1 W t 1 1 β K p ( t(k) η x t(k) + (e 2)η 2 x t(k) 2) k=1 so, we have log W n+1 W β k=1 K p ( t(k) η x t(k) + (e 2)η 2 x t(k) 2). For any k = 1,..., K, we also have log W n+1 W 1 K = log e η n x t(k) log K η x t(k) log K. k=1
24 EXP3 Regret Bound III Proof: By taking the expectation, for any k = 1,..., K, we may write E [ (1 β) x t(k) ] K p t(i) x t(i) i=1 [ ] x t(k) E x t(i t) (1 β) log K η βn + log K η E[R n(k)] log K η + (e 2)ηE + (e 2)ηnK + (e 1)ηnK [ k=1 ] K p t(k) x t(k) 2
The Multi-Arm Bandit Framework
The Multi-Arm Bandit Framework A. LAZARIC (SequeL Team @INRIA-Lille) ENS Cachan - Master 2 MVA SequeL INRIA Lille MVA-RL Course In This Lecture A. LAZARIC Reinforcement Learning Algorithms Oct 29th, 2013-2/94
More informationBandit models: a tutorial
Gdt COS, December 3rd, 2015 Multi-Armed Bandit model: general setting K arms: for a {1,..., K}, (X a,t ) t N is a stochastic process. (unknown distributions) Bandit game: a each round t, an agent chooses
More informationIntroduction to Reinforcement Learning Part 3: Exploration for decision making, Application to games, optimization, and planning
Introduction to Reinforcement Learning Part 3: Exploration for decision making, Application to games, optimization, and planning Rémi Munos SequeL project: Sequential Learning http://researchers.lille.inria.fr/
More informationMulti-armed bandit models: a tutorial
Multi-armed bandit models: a tutorial CERMICS seminar, March 30th, 2016 Multi-Armed Bandit model: general setting K arms: for a {1,..., K}, (X a,t ) t N is a stochastic process. (unknown distributions)
More informationAdvanced Machine Learning
Advanced Machine Learning Bandit Problems MEHRYAR MOHRI MOHRI@ COURANT INSTITUTE & GOOGLE RESEARCH. Multi-Armed Bandit Problem Problem: which arm of a K-slot machine should a gambler pull to maximize his
More informationIntroduction to Reinforcement Learning Part 3: Exploration for sequential decision making
Introduction to Reinforcement Learning Part 3: Exploration for sequential decision making Rémi Munos SequeL project: Sequential Learning http://researchers.lille.inria.fr/ munos/ INRIA Lille - Nord Europe
More informationRegret Analysis of Stochastic and Nonstochastic Multi-armed Bandit Problems, Part I. Sébastien Bubeck Theory Group
Regret Analysis of Stochastic and Nonstochastic Multi-armed Bandit Problems, Part I Sébastien Bubeck Theory Group i.i.d. multi-armed bandit, Robbins [1952] i.i.d. multi-armed bandit, Robbins [1952] Known
More informationTwo optimization problems in a stochastic bandit model
Two optimization problems in a stochastic bandit model Emilie Kaufmann joint work with Olivier Cappé, Aurélien Garivier and Shivaram Kalyanakrishnan Journées MAS 204, Toulouse Outline From stochastic optimization
More informationThe Multi-Armed Bandit Problem
The Multi-Armed Bandit Problem Electrical and Computer Engineering December 7, 2013 Outline 1 2 Mathematical 3 Algorithm Upper Confidence Bound Algorithm A/B Testing Exploration vs. Exploitation Scientist
More informationIntroduction to Reinforcement Learning and multi-armed bandits
Introduction to Reinforcement Learning and multi-armed bandits Rémi Munos INRIA Lille - Nord Europe Currently on leave at MSR-NE http://researchers.lille.inria.fr/ munos/ NETADIS Summer School 2013, Hillerod,
More informationStochastic bandits: Explore-First and UCB
CSE599s, Spring 2014, Online Learning Lecture 15-2/19/2014 Stochastic bandits: Explore-First and UCB Lecturer: Brendan McMahan or Ofer Dekel Scribe: Javad Hosseini In this lecture, we like to answer this
More informationMulti-Armed Bandits. Credit: David Silver. Google DeepMind. Presenter: Tianlu Wang
Multi-Armed Bandits Credit: David Silver Google DeepMind Presenter: Tianlu Wang Credit: David Silver (DeepMind) Multi-Armed Bandits Presenter: Tianlu Wang 1 / 27 Outline 1 Introduction Exploration vs.
More informationReinforcement Learning
Reinforcement Learning Lecture 5: Bandit optimisation Alexandre Proutiere, Sadegh Talebi, Jungseul Ok KTH, The Royal Institute of Technology Objectives of this lecture Introduce bandit optimisation: the
More information1 MDP Value Iteration Algorithm
CS 0. - Active Learning Problem Set Handed out: 4 Jan 009 Due: 9 Jan 009 MDP Value Iteration Algorithm. Implement the value iteration algorithm given in the lecture. That is, solve Bellman s equation using
More informationOn Bayesian bandit algorithms
On Bayesian bandit algorithms Emilie Kaufmann joint work with Olivier Cappé, Aurélien Garivier, Nathaniel Korda and Rémi Munos July 1st, 2012 Emilie Kaufmann (Telecom ParisTech) On Bayesian bandit algorithms
More informationOn the Complexity of Best Arm Identification with Fixed Confidence
On the Complexity of Best Arm Identification with Fixed Confidence Discrete Optimization with Noise Aurélien Garivier, Emilie Kaufmann COLT, June 23 th 2016, New York Institut de Mathématiques de Toulouse
More informationLecture 19: UCB Algorithm and Adversarial Bandit Problem. Announcements Review on stochastic multi-armed bandit problem
Lecture 9: UCB Algorithm and Adversarial Bandit Problem EECS598: Prediction and Learning: It s Only a Game Fall 03 Lecture 9: UCB Algorithm and Adversarial Bandit Problem Prof. Jacob Abernethy Scribe:
More informationBandit Algorithms. Zhifeng Wang ... Department of Statistics Florida State University
Bandit Algorithms Zhifeng Wang Department of Statistics Florida State University Outline Multi-Armed Bandits (MAB) Exploration-First Epsilon-Greedy Softmax UCB Thompson Sampling Adversarial Bandits Exp3
More informationStat 260/CS Learning in Sequential Decision Problems. Peter Bartlett
Stat 260/CS 294-102. Learning in Sequential Decision Problems. Peter Bartlett 1. Multi-armed bandit algorithms. Concentration inequalities. P(X ǫ) exp( ψ (ǫ))). Cumulant generating function bounds. Hoeffding
More informationBayesian and Frequentist Methods in Bandit Models
Bayesian and Frequentist Methods in Bandit Models Emilie Kaufmann, Telecom ParisTech Bayes In Paris, ENSAE, October 24th, 2013 Emilie Kaufmann (Telecom ParisTech) Bayesian and Frequentist Bandits BIP,
More informationCOS 402 Machine Learning and Artificial Intelligence Fall Lecture 22. Exploration & Exploitation in Reinforcement Learning: MAB, UCB, Exp3
COS 402 Machine Learning and Artificial Intelligence Fall 2016 Lecture 22 Exploration & Exploitation in Reinforcement Learning: MAB, UCB, Exp3 How to balance exploration and exploitation in reinforcement
More informationStat 260/CS Learning in Sequential Decision Problems. Peter Bartlett
Stat 260/CS 294-102. Learning in Sequential Decision Problems. Peter Bartlett 1. Adversarial bandits Definition: sequential game. Lower bounds on regret from the stochastic case. Exp3: exponential weights
More informationStratégies bayésiennes et fréquentistes dans un modèle de bandit
Stratégies bayésiennes et fréquentistes dans un modèle de bandit thèse effectuée à Telecom ParisTech, co-dirigée par Olivier Cappé, Aurélien Garivier et Rémi Munos Journées MAS, Grenoble, 30 août 2016
More informationRevisiting the Exploration-Exploitation Tradeoff in Bandit Models
Revisiting the Exploration-Exploitation Tradeoff in Bandit Models joint work with Aurélien Garivier (IMT, Toulouse) and Tor Lattimore (University of Alberta) Workshop on Optimization and Decision-Making
More informationRobustness of Anytime Bandit Policies
Robustness of Anytime Bandit Policies Antoine Salomon, Jean-Yves Audibert To cite this version: Antoine Salomon, Jean-Yves Audibert. Robustness of Anytime Bandit Policies. 011. HAL Id:
More informationOnline Learning and Sequential Decision Making
Online Learning and Sequential Decision Making Emilie Kaufmann CNRS & CRIStAL, Inria SequeL, emilie.kaufmann@univ-lille.fr Research School, ENS Lyon, Novembre 12-13th 2018 Emilie Kaufmann Sequential Decision
More informationThe information complexity of best-arm identification
The information complexity of best-arm identification Emilie Kaufmann, joint work with Olivier Cappé and Aurélien Garivier MAB workshop, Lancaster, January th, 206 Context: the multi-armed bandit model
More informationOn the Complexity of Best Arm Identification in Multi-Armed Bandit Models
On the Complexity of Best Arm Identification in Multi-Armed Bandit Models Aurélien Garivier Institut de Mathématiques de Toulouse Information Theory, Learning and Big Data Simons Institute, Berkeley, March
More informationThe information complexity of sequential resource allocation
The information complexity of sequential resource allocation Emilie Kaufmann, joint work with Olivier Cappé, Aurélien Garivier and Shivaram Kalyanakrishan SMILE Seminar, ENS, June 8th, 205 Sequential allocation
More informationSubsampling, Concentration and Multi-armed bandits
Subsampling, Concentration and Multi-armed bandits Odalric-Ambrym Maillard, R. Bardenet, S. Mannor, A. Baransi, N. Galichet, J. Pineau, A. Durand Toulouse, November 09, 2015 O-A. Maillard Subsampling and
More informationBandit Algorithms. Tor Lattimore & Csaba Szepesvári
Bandit Algorithms Tor Lattimore & Csaba Szepesvári Bandits Time 1 2 3 4 5 6 7 8 9 10 11 12 Left arm $1 $0 $1 $1 $0 Right arm $1 $0 Five rounds to go. Which arm would you play next? Overview What are bandits,
More informationTwo generic principles in modern bandits: the optimistic principle and Thompson sampling
Two generic principles in modern bandits: the optimistic principle and Thompson sampling Rémi Munos INRIA Lille, France CSML Lunch Seminars, September 12, 2014 Outline Two principles: The optimistic principle
More informationBandits : optimality in exponential families
Bandits : optimality in exponential families Odalric-Ambrym Maillard IHES, January 2016 Odalric-Ambrym Maillard Bandits 1 / 40 Introduction 1 Stochastic multi-armed bandits 2 Boundary crossing probabilities
More informationONLINE ADVERTISEMENTS AND MULTI-ARMED BANDITS CHONG JIANG DISSERTATION
c 2015 Chong Jiang ONLINE ADVERTISEMENTS AND MULTI-ARMED BANDITS BY CHONG JIANG DISSERTATION Submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy in Electrical and
More informationFrom Bandits to Monte-Carlo Tree Search: The Optimistic Principle Applied to Optimization and Planning
From Bandits to Monte-Carlo Tree Search: The Optimistic Principle Applied to Optimization and Planning Rémi Munos To cite this version: Rémi Munos. From Bandits to Monte-Carlo Tree Search: The Optimistic
More informationOnline Learning: Bandit Setting
Online Learning: Bandit Setting Daniel asabi Summer 04 Last Update: October 0, 06 Introduction [TODO Bandits. Stocastic setting Suppose tere exists unknown distributions ν,..., ν, suc tat te loss at eac
More informationThe multi armed-bandit problem
The multi armed-bandit problem (with covariates if we have time) Vianney Perchet & Philippe Rigollet LPMA Université Paris Diderot ORFE Princeton University Algorithms and Dynamics for Games and Optimization
More informationOnline Learning Schemes for Power Allocation in Energy Harvesting Communications
Online Learning Schemes for Power Allocation in Energy Harvesting Communications Pranav Sakulkar and Bhaskar Krishnamachari Ming Hsieh Department of Electrical Engineering Viterbi School of Engineering
More informationTHE first formalization of the multi-armed bandit problem
EDIC RESEARCH PROPOSAL 1 Multi-armed Bandits in a Network Farnood Salehi I&C, EPFL Abstract The multi-armed bandit problem is a sequential decision problem in which we have several options (arms). We can
More informationAn Asymptotically Optimal Algorithm for the Max k-armed Bandit Problem
An Asymptotically Optimal Algorithm for the Max k-armed Bandit Problem Matthew J. Streeter February 27, 2006 CMU-CS-06-110 Stephen F. Smith School of Computer Science Carnegie Mellon University Pittsburgh,
More informationBandits and Exploration: How do we (optimally) gather information? Sham M. Kakade
Bandits and Exploration: How do we (optimally) gather information? Sham M. Kakade Machine Learning for Big Data CSE547/STAT548 University of Washington S. M. Kakade (UW) Optimization for Big data 1 / 22
More informationBandit View on Continuous Stochastic Optimization
Bandit View on Continuous Stochastic Optimization Sébastien Bubeck 1 joint work with Rémi Munos 1 & Gilles Stoltz 2 & Csaba Szepesvari 3 1 INRIA Lille, SequeL team 2 CNRS/ENS/HEC 3 University of Alberta
More informationAnytime optimal algorithms in stochastic multi-armed bandits
Rémy Degenne LPMA, Université Paris Diderot Vianney Perchet CREST, ENSAE REMYDEGENNE@MATHUNIV-PARIS-DIDEROTFR VIANNEYPERCHET@NORMALESUPORG Abstract We introduce an anytime algorithm for stochastic multi-armed
More informationPure Exploration in Finitely Armed and Continuous Armed Bandits
Pure Exploration in Finitely Armed and Continuous Armed Bandits Sébastien Bubeck INRIA Lille Nord Europe, SequeL project, 40 avenue Halley, 59650 Villeneuve d Ascq, France Rémi Munos INRIA Lille Nord Europe,
More informationOrdinal optimization - Empirical large deviations rate estimators, and multi-armed bandit methods
Ordinal optimization - Empirical large deviations rate estimators, and multi-armed bandit methods Sandeep Juneja Tata Institute of Fundamental Research Mumbai, India joint work with Peter Glynn Applied
More informationMinimax strategy for Stratified Sampling for Monte Carlo
Journal of Machine Learning Research 1 2000 1-48 Submitted 4/00; Published 10/00 Minimax strategy for Stratified Sampling for Monte Carlo Alexandra Carpentier INRIA Lille - Nord Europe 40, avenue Halley
More informationReinforcement Learning
Reinforcement Learning Lecture 6: RL algorithms 2.0 Alexandre Proutiere, Sadegh Talebi, Jungseul Ok KTH, The Royal Institute of Technology Objectives of this lecture Present and analyse two online algorithms
More informationCsaba Szepesvári 1. University of Alberta. Machine Learning Summer School, Ile de Re, France, 2008
LEARNING THEORY OF OPTIMAL DECISION MAKING PART I: ON-LINE LEARNING IN STOCHASTIC ENVIRONMENTS Csaba Szepesvári 1 1 Department of Computing Science University of Alberta Machine Learning Summer School,
More informationThe Multi-Armed Bandit Problem
Università degli Studi di Milano The bandit problem [Robbins, 1952]... K slot machines Rewards X i,1, X i,2,... of machine i are i.i.d. [0, 1]-valued random variables An allocation policy prescribes which
More informationOn the Complexity of Best Arm Identification with Fixed Confidence
On the Complexity of Best Arm Identification with Fixed Confidence Discrete Optimization with Noise Aurélien Garivier, joint work with Emilie Kaufmann CNRS, CRIStAL) to be presented at COLT 16, New York
More informationAdaptive Learning with Unknown Information Flows
Adaptive Learning with Unknown Information Flows Yonatan Gur Stanford University Ahmadreza Momeni Stanford University June 8, 018 Abstract An agent facing sequential decisions that are characterized by
More informationAn Estimation Based Allocation Rule with Super-linear Regret and Finite Lock-on Time for Time-dependent Multi-armed Bandit Processes
An Estimation Based Allocation Rule with Super-linear Regret and Finite Lock-on Time for Time-dependent Multi-armed Bandit Processes Prokopis C. Prokopiou, Peter E. Caines, and Aditya Mahajan McGill University
More informationDynamic resource allocation: Bandit problems and extensions
Dynamic resource allocation: Bandit problems and extensions Aurélien Garivier Institut de Mathématiques de Toulouse MAD Seminar, Université Toulouse 1 October 3rd, 2014 The Bandit Model Roadmap 1 The Bandit
More informationRegret lower bounds and extended Upper Confidence Bounds policies in stochastic multi-armed bandit problem
Journal of Machine Learning Research 1 01) 1-48 Submitted 4/00; Published 10/00 Regret lower bounds and extended Upper Confidence Bounds policies in stochastic multi-armed bandit problem Antoine Salomon
More informationEASINESS IN BANDITS. Gergely Neu. Pompeu Fabra University
EASINESS IN BANDITS Gergely Neu Pompeu Fabra University EASINESS IN BANDITS Gergely Neu Pompeu Fabra University THE BANDIT PROBLEM Play for T rounds attempting to maximize rewards THE BANDIT PROBLEM Play
More informationMechanisms with Learning for Stochastic Multi-armed Bandit Problems
Mechanisms with Learning for Stochastic Multi-armed Bandit Problems Shweta Jain 1, Satyanath Bhat 1, Ganesh Ghalme 1, Divya Padmanabhan 1, and Y. Narahari 1 Department of Computer Science and Automation,
More informationStat 260/CS Learning in Sequential Decision Problems.
Stat 260/CS 294-102. Learning in Sequential Decision Problems. Peter Bartlett 1. Multi-armed bandit algorithms. Exponential families. Cumulant generating function. KL-divergence. KL-UCB for an exponential
More informationMulti-Armed Bandit: Learning in Dynamic Systems with Unknown Models
c Qing Zhao, UC Davis. Talk at Xidian Univ., September, 2011. 1 Multi-Armed Bandit: Learning in Dynamic Systems with Unknown Models Qing Zhao Department of Electrical and Computer Engineering University
More informationComplexity of stochastic branch and bound methods for belief tree search in Bayesian reinforcement learning
Complexity of stochastic branch and bound methods for belief tree search in Bayesian reinforcement learning Christos Dimitrakakis Informatics Institute, University of Amsterdam, Amsterdam, The Netherlands
More informationNew Algorithms for Contextual Bandits
New Algorithms for Contextual Bandits Lev Reyzin Georgia Institute of Technology Work done at Yahoo! 1 S A. Beygelzimer, J. Langford, L. Li, L. Reyzin, R.E. Schapire Contextual Bandit Algorithms with Supervised
More informationOnline learning with feedback graphs and switching costs
Online learning with feedback graphs and switching costs A Proof of Theorem Proof. Without loss of generality let the independent sequence set I(G :T ) formed of actions (or arms ) from to. Given the sequence
More informationDeviations of stochastic bandit regret
Deviations of stochastic bandit regret Antoine Salomon 1 and Jean-Yves Audibert 1,2 1 Imagine École des Ponts ParisTech Université Paris Est salomona@imagine.enpc.fr audibert@imagine.enpc.fr 2 Sierra,
More informationOnline Learning with Gaussian Payoffs and Side Observations
Online Learning with Gaussian Payoffs and Side Observations Yifan Wu 1 András György 2 Csaba Szepesvári 1 1 Department of Computing Science University of Alberta 2 Department of Electrical and Electronic
More informationAn Optimal Bidimensional Multi Armed Bandit Auction for Multi unit Procurement
An Optimal Bidimensional Multi Armed Bandit Auction for Multi unit Procurement Satyanath Bhat Joint work with: Shweta Jain, Sujit Gujar, Y. Narahari Department of Computer Science and Automation, Indian
More informationOrdinal Optimization and Multi Armed Bandit Techniques
Ordinal Optimization and Multi Armed Bandit Techniques Sandeep Juneja. with Peter Glynn September 10, 2014 The ordinal optimization problem Determining the best of d alternative designs for a system, on
More informationMultiple Identifications in Multi-Armed Bandits
Multiple Identifications in Multi-Armed Bandits arxiv:05.38v [cs.lg] 4 May 0 Sébastien Bubeck Department of Operations Research and Financial Engineering, Princeton University sbubeck@princeton.edu Tengyao
More information1 A Support Vector Machine without Support Vectors
CS/CNS/EE 53 Advanced Topics in Machine Learning Problem Set 1 Handed out: 15 Jan 010 Due: 01 Feb 010 1 A Support Vector Machine without Support Vectors In this question, you ll be implementing an online
More informationUpper-Confidence-Bound Algorithms for Active Learning in Multi-Armed Bandits
Upper-Confidence-Bound Algorithms for Active Learning in Multi-Armed Bandits Alexandra Carpentier 1, Alessandro Lazaric 1, Mohammad Ghavamzadeh 1, Rémi Munos 1, and Peter Auer 2 1 INRIA Lille - Nord Europe,
More informationPrediction and Playing Games
Prediction and Playing Games Vineel Pratap vineel@eng.ucsd.edu February 20, 204 Chapter 7 : Prediction, Learning and Games - Cesa Binachi & Lugosi K-Person Normal Form Games Each player k (k =,..., K)
More informationGrundlagen der Künstlichen Intelligenz
Grundlagen der Künstlichen Intelligenz Uncertainty & Probabilities & Bandits Daniel Hennes 16.11.2017 (WS 2017/18) University Stuttgart - IPVS - Machine Learning & Robotics 1 Today Uncertainty Probability
More informationBest Arm Identification in Multi-Armed Bandits
Best Arm Identification in Multi-Armed Bandits Jean-Yves Audibert Imagine, Université Paris Est & Willow, CNRS/ENS/INRIA, Paris, France audibert@certisenpcfr Sébastien Bubeck, Rémi Munos SequeL Project,
More informationExploration and exploitation of scratch games
Mach Learn (2013) 92:377 401 DOI 10.1007/s10994-013-5359-2 Exploration and exploitation of scratch games Raphaël Féraud Tanguy Urvoy Received: 10 January 2013 / Accepted: 12 April 2013 / Published online:
More informationMulti-Armed Bandit Formulations for Identification and Control
Multi-Armed Bandit Formulations for Identification and Control Cristian R. Rojas Joint work with Matías I. Müller and Alexandre Proutiere KTH Royal Institute of Technology, Sweden ERNSI, September 24-27,
More informationAn Experimental Evaluation of High-Dimensional Multi-Armed Bandits
An Experimental Evaluation of High-Dimensional Multi-Armed Bandits Naoki Egami Romain Ferrali Kosuke Imai Princeton University Talk at Political Data Science Conference Washington University, St. Louis
More informationReinforcement Learning
Reinforcement Learning Markov decision process & Dynamic programming Evaluative feedback, value function, Bellman equation, optimality, Markov property, Markov decision process, dynamic programming, value
More informationarxiv: v1 [cs.lg] 25 Jul 2018
Variational Bayesian Reinforcement Learning with Regret Bounds Brendan O Donoghue DeepMind bodonoghue@google.com arxiv:1807.09647v1 [cs.lg] 25 Jul 2018 July 25, 2018 Abstract We consider the exploration-exploitation
More informationFinite-time Analysis of the Multiarmed Bandit Problem*
Machine Learning, 47, 35 56, 00 c 00 Kluwer Academic Publishers. Manufactured in The Netherlands. Finite-time Analysis of the Multiarmed Bandit Problem* PETER AUER University of Technology Graz, A-8010
More informationRegret Analysis of Stochastic and Nonstochastic Multi-armed Bandit Problems. Sébastien Bubeck Theory Group
Regret Analysis of Stochastic and Nonstochastic Multi-armed Bandit Problems Sébastien Bubeck Theory Group Part 1: i.i.d., adversarial, and Bayesian bandit models i.i.d. multi-armed bandit, Robbins [1952]
More informationMulti-armed Bandits in the Presence of Side Observations in Social Networks
52nd IEEE Conference on Decision and Control December 0-3, 203. Florence, Italy Multi-armed Bandits in the Presence of Side Observations in Social Networks Swapna Buccapatnam, Atilla Eryilmaz, and Ness
More informationCOMPSCI 240: Reasoning Under Uncertainty
COMPSCI 240: Reasoning Under Uncertainty Andrew Lan and Nic Herndon University of Massachusetts at Amherst Spring 2019 Lecture 20: Central limit theorem & The strong law of large numbers Markov and Chebyshev
More informationReinforcement Learning
Reinforcement Learning Lecture 3: RL problems, sample complexity and regret Alexandre Proutiere, Sadegh Talebi, Jungseul Ok KTH, The Royal Institute of Technology Objectives of this lecture Introduce the
More informationAdministration. CSCI567 Machine Learning (Fall 2018) Outline. Outline. HW5 is available, due on 11/18. Practice final will also be available soon.
Administration CSCI567 Machine Learning Fall 2018 Prof. Haipeng Luo U of Southern California Nov 7, 2018 HW5 is available, due on 11/18. Practice final will also be available soon. Remaining weeks: 11/14,
More informationarxiv: v2 [cs.lg] 3 Nov 2012
arxiv:1204.5721v2 [cs.lg] 3 Nov 2012 Regret Analysis of Stochastic and Nonstochastic Multi-armed Bandit Problems Sébastien Bubeck 1 and Nicolò Cesa-Bianchi 2 1 Department of Operations Research and Financial
More informationInformational Confidence Bounds for Self-Normalized Averages and Applications
Informational Confidence Bounds for Self-Normalized Averages and Applications Aurélien Garivier Institut de Mathématiques de Toulouse - Université Paul Sabatier Thursday, September 12th 2013 Context Tree
More informationRegional Multi-Armed Bandits
School of Information Science and Technology University of Science and Technology of China {wzy43, zrd7}@mail.ustc.edu.cn, congshen@ustc.edu.cn Abstract We consider a variant of the classic multiarmed
More informationKULLBACK-LEIBLER UPPER CONFIDENCE BOUNDS FOR OPTIMAL SEQUENTIAL ALLOCATION
Submitted to the Annals of Statistics arxiv: math.pr/0000000 KULLBACK-LEIBLER UPPER CONFIDENCE BOUNDS FOR OPTIMAL SEQUENTIAL ALLOCATION By Olivier Cappé 1, Aurélien Garivier 2, Odalric-Ambrym Maillard
More informationBasics of reinforcement learning
Basics of reinforcement learning Lucian Buşoniu TMLSS, 20 July 2018 Main idea of reinforcement learning (RL) Learn a sequential decision policy to optimize the cumulative performance of an unknown system
More informationExploiting Correlation in Finite-Armed Structured Bandits
Exploiting Correlation in Finite-Armed Structured Bandits Samarth Gupta Carnegie Mellon University Pittsburgh, PA 1513 Gauri Joshi Carnegie Mellon University Pittsburgh, PA 1513 Osman Yağan Carnegie Mellon
More informationAlireza Shafaei. Machine Learning Reading Group The University of British Columbia Summer 2017
s s Machine Learning Reading Group The University of British Columbia Summer 2017 (OCO) Convex 1/29 Outline (OCO) Convex Stochastic Bernoulli s (OCO) Convex 2/29 At each iteration t, the player chooses
More informationMulti armed bandit problem: some insights
Multi armed bandit problem: some insights July 4, 20 Introduction Multi Armed Bandit problems have been widely studied in the context of sequential analysis. The application areas include clinical trials,
More informationLogarithmic Online Regret Bounds for Undiscounted Reinforcement Learning
Logarithmic Online Regret Bounds for Undiscounted Reinforcement Learning Peter Auer Ronald Ortner University of Leoben, Franz-Josef-Strasse 18, 8700 Leoben, Austria auer,rortner}@unileoben.ac.at Abstract
More informationOnline learning CMPUT 654. October 20, 2011
Online learning CMPUT 654 Gábor Bartók Dávid Pál Csaba Szepesvári István Szita October 20, 2011 Contents 1 Shooting Game 4 1.1 Exercises...................................... 6 2 Weighted Majority Algorithm
More informationAnnealing-Pareto Multi-Objective Multi-Armed Bandit Algorithm
Annealing-Pareto Multi-Objective Multi-Armed Bandit Algorithm Saba Q. Yahyaa, Madalina M. Drugan and Bernard Manderick Vrije Universiteit Brussel, Department of Computer Science, Pleinlaan 2, 1050 Brussels,
More informationComputational Oracle Inequalities for Large Scale Model Selection Problems
for Large Scale Model Selection Problems University of California at Berkeley Queensland University of Technology ETH Zürich, September 2011 Joint work with Alekh Agarwal, John Duchi and Clément Levrard.
More informationUpper-Confidence-Bound Algorithms for Active Learning in Multi-armed Bandits
Upper-Confidence-Bound Algorithms for Active Learning in Multi-armed Bandits Alexandra Carpentier 1, Alessandro Lazaric 1, Mohammad Ghavamzadeh 1, Rémi Munos 1, and Peter Auer 2 1 INRIA Lille - Nord Europe,
More informationOn Robust Arm-Acquiring Bandit Problems
On Robust Arm-Acquiring Bandit Problems Shiqing Yu Faculty Mentor: Xiang Yu July 20, 2014 Abstract In the classical multi-armed bandit problem, at each stage, the player has to choose one from N given
More informationMulti-armed Bandit with Additional Observations
Multi-armed Bandit with Additional Observations DOGGYU YU, aver Corporation ALEXADRE PROUTIERE, KTH SUMYEOG AH, JIWOO SHI, and YUG YI, KAIST We study multi-armed bandit (MAB) problems with additional observations,
More informationExponential Moving Average Based Multiagent Reinforcement Learning Algorithms
Exponential Moving Average Based Multiagent Reinforcement Learning Algorithms Mostafa D. Awheda Department of Systems and Computer Engineering Carleton University Ottawa, Canada KS 5B6 Email: mawheda@sce.carleton.ca
More informationLecture 4: Lower Bounds (ending); Thompson Sampling
CMSC 858G: Bandits, Experts and Games 09/12/16 Lecture 4: Lower Bounds (ending); Thompson Sampling Instructor: Alex Slivkins Scribed by: Guowei Sun,Cheng Jie 1 Lower bounds on regret (ending) Recap from
More informationarxiv: v1 [cs.lg] 21 Sep 2018
SIC - MMAB: Synchronisation Involves Communication in Multiplayer Multi-Armed Bandits arxiv:809.085v [cs.lg] 2 Sep 208 Etienne Boursier and Vianney Perchet,2 CMLA, ENS Cachan, CNRS, Université Paris-Saclay,
More information