Contextual Combinatorial Bandit and its Application on Diversified Online Recommendation

Similar documents
Sparse Linear Contextual Bandits via Relevance Vector Machines

Bandit Algorithms. Zhifeng Wang ... Department of Statistics Florida State University

Evaluation of multi armed bandit algorithms and empirical algorithm

Combinatorial Multi-Armed Bandit: General Framework, Results and Applications

Online Learning and Sequential Decision Making

Sequential Multi-armed Bandits

Combinatorial Multi-Armed Bandit and Its Extension to Probabilistically Triggered Arms

arxiv: v1 [cs.lg] 7 Sep 2018

Introducing strategic measure actions in multi-armed bandits

Improved Algorithms for Linear Stochastic Bandits

EFFICIENT ALGORITHMS FOR LINEAR POLYHEDRAL BANDITS. Manjesh K. Hanawal Amir Leshem Venkatesh Saligrama

Tsinghua Machine Learning Guest Lecture, June 9,

Stochastic Contextual Bandits with Known. Reward Functions

Two generic principles in modern bandits: the optimistic principle and Thompson sampling

Spectral Bandits for Smooth Graph Functions with Applications in Recommender Systems

The Multi-Armed Bandit Problem

Transferable Contextual Bandit for Cross-Domain Recommendation

Profile-Based Bandit with Unknown Profiles

Large-scale Information Processing, Summer Recommender Systems (part 2)

THE first formalization of the multi-armed bandit problem

Analysis of Thompson Sampling for the multi-armed bandit problem

Tutorial: PART 2. Online Convex Optimization, A Game- Theoretic Approach to Learning

Multi-armed bandit models: a tutorial

Bandit models: a tutorial

New Algorithms for Contextual Bandits

Bayesian Contextual Multi-armed Bandits

Impact of Data Characteristics on Recommender Systems Performance

Lecture 19: UCB Algorithm and Adversarial Bandit Problem. Announcements Review on stochastic multi-armed bandit problem

Multi-armed Bandits in the Presence of Side Observations in Social Networks

New bounds on the price of bandit feedback for mistake-bounded online multiclass learning

Collaborative Filtering. Radek Pelánek

Reinforcement Learning

Contextual semibandits via supervised learning oracles

Crowd-Learning: Improving the Quality of Crowdsourcing Using Sequential Learning

Reducing contextual bandits to supervised learning

A Time and Space Efficient Algorithm for Contextual Linear Bandits

COMBINATORIAL MULTI-ARMED BANDIT PROBLEM WITH PROBABILISTICALLY TRIGGERED ARMS: A CASE WITH BOUNDED REGRET

Learning to play K-armed bandit problems

Contextual Combinatorial Cascading Bandits

Bandits for Online Optimization

A Modified PMF Model Incorporating Implicit Item Associations

A Framework for Recommending Relevant and Diverse Items

Stratégies bayésiennes et fréquentistes dans un modèle de bandit

Content-based Recommendation

Adaptive Shortest-Path Routing under Unknown and Stochastically Varying Link States

An Estimation Based Allocation Rule with Super-linear Regret and Finite Lock-on Time for Time-dependent Multi-armed Bandit Processes

Discover Relevant Sources : A Multi-Armed Bandit Approach

Learning Algorithms for Minimizing Queue Length Regret

Hybrid Machine Learning Algorithms

Contextual Bandits in A Collaborative Environment

When Gaussian Processes Meet Combinatorial Bandits: GCB

Thompson Sampling for the MNL-Bandit

Learning symmetric non-monotone submodular functions

Performance and Convergence of Multi-user Online Learning

Grundlagen der Künstlichen Intelligenz

Multi-Attribute Bayesian Optimization under Utility Uncertainty

Advanced Machine Learning

COS 402 Machine Learning and Artificial Intelligence Fall Lecture 22. Exploration & Exploitation in Reinforcement Learning: MAB, UCB, Exp3

Lecture 10 : Contextual Bandits

Administration. CSCI567 Machine Learning (Fall 2018) Outline. Outline. HW5 is available, due on 11/18. Practice final will also be available soon.

On the Complexity of Best Arm Identification in Multi-Armed Bandit Models

Discovering Unknown Unknowns of Predictive Models

A UCB-Like Strategy of Collaborative Filtering

Learning Exploration/Exploitation Strategies for Single Trajectory Reinforcement Learning

arxiv: v1 [cs.lg] 12 Sep 2017

Reward Maximization Under Uncertainty: Leveraging Side-Observations on Networks

Bandit Algorithms. Tor Lattimore & Csaba Szepesvári

* Matrix Factorization and Recommendation Systems

arxiv: v6 [cs.lg] 29 Mar 2016

1 [15 points] Frequent Itemsets Generation With Map-Reduce

Lecture 4: Lower Bounds (ending); Thompson Sampling

Multi-task Linear Bandits

The Multi-Arm Bandit Framework

Advanced Topics in Machine Learning and Algorithmic Game Theory Fall semester, 2011/12

Regret Analysis of Stochastic and Nonstochastic Multi-armed Bandit Problems, Part I. Sébastien Bubeck Theory Group

Probabilistic Partial User Model Similarity for Collaborative Filtering

1 Overview. 2 Learning from Experts. 2.1 Defining a meaningful benchmark. AM 221: Advanced Optimization Spring 2016

arxiv: v2 [cs.ir] 14 May 2018

Tor Lattimore & Csaba Szepesvári

Asymmetric Correlation Regularized Matrix Factorization for Web Service Recommendation

1 A Support Vector Machine without Support Vectors

Data Mining Techniques

Basics of reinforcement learning

Bandits and Recommender Systems

Yevgeny Seldin. University of Copenhagen

The information complexity of sequential resource allocation

On the Optimality of Myopic Sensing. in Multi-channel Opportunistic Access: the Case of Sensing Multiple Channels

Stat 260/CS Learning in Sequential Decision Problems. Peter Bartlett

Estimation Considerations in Contextual Bandits

The geometry of Gaussian processes and Bayesian optimization. Contal CMLA, ENS Cachan

arxiv: v4 [cs.lg] 22 Jul 2014

Combinatorial Network Optimization With Unknown Variables: Multi-Armed Bandits With Linear Rewards and Individual Observations

Active Learning and Optimized Information Gathering

An Information-Theoretic Analysis of Thompson Sampling

Introduction to Bandit Algorithms. Introduction to Bandit Algorithms

Optimal and Adaptive Online Learning

The Knowledge Gradient for Sequential Decision Making with Stochastic Binary Feedbacks

Complex Bandit Problems and Thompson Sampling

Offline Evaluation of Ranking Policies with Click Models

An Optimal Bidimensional Multi Armed Bandit Auction for Multi unit Procurement

Transcription:

Contextual Combinatorial Bandit and its Application on Diversified Online Recommendation Lijing Qin Shouyuan Chen Xiaoyan Zhu Abstract Recommender systems are faced with new challenges that are beyond traditional techniques. For example, most traditional techniques are based on similarity or overlap among existing data, however, there may not exist sufficient historical records for some new users to predict their preference, or users can hold diverse interest, but the similarity based methods may probably over-narrow it. To address the above challenges, we develop a principled approach called contextual combinatorial bandit in which a learning algorithm can dynamically identify diverse items that interest a new user. Specifically, each item is represented as a feature vector, and each user is represented as an unknown preference vector. On each of n rounds, the bandit algorithm sequentially selects a set of items according to the item-selection strategy that balances exploration and exploitation, and collects the user feedback on these selected items. A reward function is further designed to measure the quality e.g. relevance or diversity of the selected set based on observed feedback, and the goal of the algorithm is to maximize the total rewards of n rounds. The reward function only needs to satisfy two mild assumptions that is general enough to accommodate a large class of nonlinear functions. To solve this bandit problem, we provide algorithm that achieves Õ n regret after playing n rounds. Experiments conducted on real-wold movie recommendation dataset demonstrate that our approach can effectively address the above challenges and hence improve the performance of recommendation task. State Key Laboratory of Intelligent Technology and Systems, Tsinghua National Laboratory for Information Science and Technology. Dept. of Computer Science and Technology, Tsinghua University, Beijing, China. Emails: qinlj09@mail.tsinghua.edu.cn, zxy-dcs@tsinghua.edu.cn. Dept. of Computer Science and Engineering, The Chinese University of Hong Kong, Shatin, N.T., Hong Kong. Email: sychen@cse.cuhk.edu.hk. 1 Introduction The multi-armed bandit MAB problem has been extensively studied in statistics and gained much popularity in the community of machine learning recently. It can be formulated as a sequential decision-making problem, where on each of n rounds a decision maker is presented with the choice of taking one of m arms or actions, each having an unknown distribution of reward. The goal of the decision maker is to maximize the total expected rewards over the course of n rounds. When each arm is represented by a feature vector that can be observed by the decision maker, the problem is known as contextual bandit problem, which is recently used to develop recommender systems that adapt to user feedback. User feedback is one kind of increasingly important source for online applications e.g. Netflix project, Google news, Amazon whose domains expand rapidly. Lots of new users don t have sufficient historical records, and hence are beyond the traditional recommendation technologies that anticipate a user s interest according to his/her past activities. Take movie recommendation for example, in contextual bandit setting, given a new user, we can repeatedly provide the user with one movie and collect his/her rating in multiple rounds. The algorithm helps decide which movie to recommend in the next round given the ratings in the previous rounds, i.e. whether we should try some new movies exploration or we should stick on the movies that the user has given high ratings so far exploitation. But in real-world scenario, recommender systems actually provide each user with a set of movies, rather than individual one. In this setting, not one simple arm but a set of arms called a super arm are played together on each round. The reward of a set of movies should not simply be the sum of ratings of each individual movie in this set. For example, we need a metric to qualify the diversity of the recommendation set to avoid redundant or over-specified recommendation lists. In this case, one can probably design a diversity promoting set function as the reward function of a super arm. The above problem can be described as a diversity promot-

ing exploration/exploitation problem. Motivated by the above challenges, we develop a principled approach called contextual combinatorial bandit to help predict user preference in recommender system. Specifically, movies are represented as feature vectors that can be regarded as contextual arms, and meanwhile users are represented as unknown preference variables, such that for users without enough historical records, a learning algorithm explores and exploits his/her preference by sequentially selecting sets of movies according to a diversity promoting reward function. Note that the contextual combinatorial bandit is a general framework, and except for the above example setting, it can also be used in other applications by defining different reward functions, as long as the functions satisfy two mild assumptions that we will discuss in section 3. In summary, the main contributions of this paper can be listed as follows: i We propose a general framework called contextual combinatorial bandit that can be used to address real-world recommendation challenges, i.e., user interest is unknown and meanwhile diverse. ii We develop C 2 UCB, an efficient algorithm for contextual combinatorial bandit, for which we show a Õ n regret bound after playing n rounds 1. iii We apply the contextual combinatorial bandit approach on online diversified movie recommendation application. As evaluation, we conduct experiments on well-known movie recommendation dataset. The experiment results demonstrate that our approach significantly improves the performance of recommendation task. 2 Related Work Most traditional recommendation techniques focus on learning user preference according to users historical records [12, 23, 6]. However, recent studies show that historical records may not well represent user interest [24, 25]. On one hand, some users may not provide sufficient records, in which case it is crucial to predict user preference dynamically according to user feedback [16, 17]. On the other hand, users can hold diverse interest, and thus recommendation techniques should not only aim at increasing relevance, but also consider improving diversity of recommended results [25, 22]. To recommend items to users without sufficient historical records, several studies formulate this task as a multi-armed bandit problem [14, 15, 5]. Multi-armed bandit is a well-studied topic in the fields of statistics and machine learning cf. [4, 2]. In traditional noncontextual bandit problem, the learner cannot access the features of arms, and the rewards of different arms are independent. In this setting, the upper confidence bound UCB algorithm is proven to be theoretically optimal [13, 3]. However, without using arm features, the performance of UCB algorithm is quite limited in many practical scenarios, especially when there are a large number of arms [14]. On the other hand, contextual bandit problem considers the case where the learner can observe the features of arms. Consequently, the learner can use these observations to infer the rewards of other unseen arms and improve the performance over time. Notably, Auer et al. [3] considered the contextual bandit problem and developed LinRel algorithm that achieved an Õ n regret bound after playing n rounds. Later, Li et al.[14] proposed LinUCB algorithm, which improves the practical performance of LinRel algorithm while enjoys similar regret bound [8]. They applied Lin- UCB algorithm on a personalized news recommendation task and demonstrated good performance [14]. In the settings of both non-contextual and contextual bandits, the learner is allowed to play one single arm on each round, i.e., recommend one item each time. However, recommending a single item on each round may not satisfy a user s diverse interest. Recently, several work generalized the classical non-contextual bandits to combinatorial bandits [10, 11, 7], where the learner can play a set of arms, which is termed as a super arm, on each round. However, as generalizations of non-contextual bandits, these work did not use arm features. Hence, their performance can be suboptimal in many recommendation tasks, particularly when the number of arms is large. Though our method inherits some concepts e.g. super arm from non-contextual combinatorial bandit, both problem formulation and regret analysis are quite different, which are actually our main contributions. Yue and Guestrin [26] proposed a linear submodular bandit approach for diversified retrieval. Their approach placed a strong restriction on user behavior. In particular, they assumed that user can only scan the items one by one in top-down fashion. In contrast, our framework has no limitation on user behavior. In addition, their framework is specifically designed for a certain type of submodular reward functions, while our approach allows a much larger class of reward functions. 1 Õ is variant of big O notation that ignores logarithmic factors.

3 Contextual Combinatorial Bandit In this section, we formulate the contextual combinatorial bandit problem. Let n be the number of rounds and m be the number of arms. Let S t 2 [m] be the set of all possible subsets of arms on round t. We call each set of arms S t S t a super arm. On each round t [n], a learner observes m feature vectors {x t 1,..., x t m} R d corresponding to m arms. Then, the learner is asked to choose one super arm S t S t to play. Once a super arm S t S t is played, the learner observes the scores of arms in {r t i} i St and receives a reward R t S t. For each arm i [m], its score r t i is assumed to be 3.1 r t i = θ T x t i + ɛ t i, where θ is a parameter unknown to the learner and the noise ɛ t i is a zero-mean random variable. On the other hand, the reward R t S t measures the quality of the super arm S t and its definition will be specified later. The goal of the learner [ is to maximize the expected ] cumulative reward E t [n] R ts t over n rounds. The reward R t S t on round t is an application dependent function which measures the quality of recommended set of arms S t [m]. The reward can simply be the sum of the scores of arms in S t, i.e. R t S t = r t i. However, our framework also allows other more complicated non-linear rewards. For example, in addition to the sum of scores of arms, the reward R t S t may also consider the diversity of arms in S t, which can be defined as a non-linear function of features of arms. Specifically, we consider the case where the expected reward E [R t S t ] is a function of three variables: super arm S t, feature vectors of arms X t {x t i} i [m] and expected scores r t {θ T x t i} i [m] associated with the arms. Formally, we denote the expected reward of playing S t as E [R t S t ] = f r S t. By choosing different types of expected reward f r,x, our framework covers both linear and non-linear rewards. Finally, in order to carry out our analysis, the expected reward f r,x is required to satisfy the following assumptions. Monotonicity The expected reward f r,x S is monotone non-decreasing with respect to the score vector r. Formally, for any set of feature vectors of arms X and super arm S, if ri r i for all i [m], we have f r,x S f r,xs. Lipschitz continuity The expected reward f r,x S is Lipschitz continuous with respect to the score vector r restricted on the arms in S. In particular, there exists a universal constant C > 0 such that, for any two score vectors r and r, we have f r,x S f r,xs C i S [ri r i] 2. Our framework does not require the player to have direct knowledge on how the reward function f r,x S is defined. Alternatively, we assume that the player has access to an oracle O S r, X, which takes the expected scores r and arms X as input, and returns the solution of the maximization problem arg max S S f r,x S. Since the maximization problems of many reward functions f r,x of practical interest are NP-hard, our framework allows the oracle to produce an approximate solution to the problem. More precisely, an oracle O S r, X is called α-approximation oracle for some α 1, if given input r and X, the oracle always returns a super arm S = O S r, X S satisfying f r,x S αopt r,x, where opt r,x = max S S f r,x S is the optimal value of the reward function. Under this setting, when α = 1, the α-approximation oracle is exact and always produces the optimal solution. Recall that the goal of the learner is to maximize its cumulative reward without knowing θ. Clearly, with the knowledge of θ, the optimal strategy is to choose S t = arg max St S t f rt,x t S t on round t. Hence, it is natural to evaluate a learner relative to this optimal strategy and the difference of the learner s total reward and the total reward of the optimal strategy is called regret. However, if a learner only has accesses to an α- approximation oracle for some α < 1, such evaluation would be unfair. Hence, in this paper, we use the notion of α-regret which compares the learner s strategy with α-fraction of the optimal rewards on round t. Formally, the α-regret on round t can be written as 3.2 Reg α t = αopt rt,x t f rt,x t S t, and we are interested in designing an algorithm whose total α-regret T Regα t is as small as possible. 4 Algorithm and α-regret Analysis In this section, we present contextual combinatorial upper confidence bound algorithm C 2 UCB. C 2 UCB is a general and efficient algorithm for the contextual combinatorial bandit problem. The basic idea of C 2 UCB is to maintain a confidence set for the true parameter θ. For each round t, the confidence set is constructed from feature vectors X 1,..., X and observed scores of selected arms {r 1 i} i S1,..., {r i} i S from previous rounds. As we will see later Theorem 4.2, our construction of the confidence sets ensures that the true parameter θ lies in the confidence set with high probability. Using this confidence set of parameter θ

and feature vectors of arms X t, the algorithm can efficiently compute an upper confidence bound for each score ˆr t = {ˆr t 1,..., ˆr t m}. The upper confidence bounds ˆr t and feature vectors of arms X t are given to the oracle as input. Then, the algorithm plays the super arm returned by the oracle and uses the observed scores to adjust the confidence sets. The pseudocode of the algorithm is listed in Algorithm 1. The algorithm has time complexity Ond 3 + md + h, where h denotes the time complexity of the oracle. Algorithm 1 C 2 UCB 1: input: λ, α 1,..., α n 2: Initialize V 0 λi d d, b 0 0 d 3: for t 1,..., n do 4: ˆθt V 1 b 5: for i 1,..., m do 6: r t i ˆθ t T x t i 7: ˆr t i r t i + α t x t i T Vt 1 x t i 8: end for 9: S t O St ˆr t, X t 10: Play super arm S t and observe {r t i} i St 11: V t V + x t ix t i T 12: b t b + r t ix t i 13: end for We now state our main theoretical result, a bound on the α-regret of Algorithm 1 when run with an α- approximation oracle. To carry out our analysis, we will need to assume that the l 2 -norms of parameter θ and feature vectors of arms X t are bounded. Using this assumption together with the monotonicity and Lipschitz continuity properties of the expected reward function f r,x, the following theorem states that the α-regret of Algorithm 1 is at most Od logn n + nd logn/δ, or Õ n if one ignores logarithmic factors and regards the dimensionality of the parameter d as a constant. Theorem 4.1. α-regret of the Algorithm 1. Without loss of generality, assume that θ 2 S, x t i 2 1 and r t i [0, 1] for all t 0 and i [m]. Given 0 < δ < 1, set α t = d log + λ 1/2 S. Then, 1+tm/λ δ with probability at least 1 δ, the total α-regret of C 2 UCB algorithm satisfies n Reg α t C 64nd log1 + nm/dλ λs + 2 log1/δ + d log1 + nm/λd, for any n 0. Note that the requirements x t i 2 1 and r t i [0, 1] can be satisfied through proper rescaling on x t i and θ. 4.1 Proof of Theorem 4.1 We begin with restating a concentration result from Abbasi-Yadkori et al., [1]. This result states that the true parameter θ lies within an ellipsoid centered at ˆθ t simultaneously for all t [n] with high probability. Theorem 4.2. [1, Theorem 2] Suppose the observed scores r t i are bounded in [0, 1]. Assume that θ 2 S and x t i 2 1 for all t 0 and i [m]. Define V t = V + n x t ix t i T and set V = λi. Then, with probability at least 1 δ, for all round t 0, the estimate ˆθ t satisfies 2 ˆθ t θ d log V 1 + tm/λ δ + λ 1/2 S. The proof of Theorem 4.2 is based on the theory of self-normalized processes. For an introduction to this theory, we refer interested readers to [21, 9]. Next, using Theorem 4.2, we show that with high probability, the upper confidence bounds of scores ˆr t also do not deviate far from the true value of scores r t for each round t [n]. Lemma 4.1. If we set α t = d log 1+tm/λ δ + λ 1/2 S, with probability at least 1 δ, we have 0 ˆr t i rt i 2α t x t i V 1, holds simultaneously for any round t 0 and any arm i [m]. Proof. By Theorem 4.2, the random event ˆθ t θ d log V 1 + tm/λ δ + λ 1/2 S holds for all t [n] simultaneously with probability at least 1 δ. Now assume the above random event happens, by the 2 We denote a M a T Ma, where a is a vector and M is a positive definite matrix.

definition of ˆr t i, we have ˆr t i rt i = ˆθ t T x t i + α t x t i V 1 θ T x t i ˆθ t θ T x t i + α t x t i V 1 ˆθ t θ x t i V 1 + α t x t i V 1 V 2α t x t i V 1. On the other hand, we have 1/λ 1/k, we obtain n x t i 2 V 1 2 n log 1 + x t i 2 V 1 = 2 log det V n 2 log det V. We remain to bound log det V n. Since x n i 2 1 and S i k for all i [n], the trace of V n can be bounded by tracev n tracev+nk. Apply the Determinant- Trace Inequality [1, Lemma 10], we have ˆr t i r i = ˆθ t T x t i + α t x t i V 1 θ T x t i = ˆθ t θ T x t i + α t x t i V 1 ˆθ t θ x t i V 1 + α t x t i V 1 V α t x t i V 1 + α t x t i V 1 = 0. To prove our main result Theorem 4.1, we need the following technical lemma. Lemma 4.2. Let V R d d be a positive definite matrix. For all t = 1, 2,..., let S t be a subset of [m] of size less than or equal to k and define V n = V + n x t ix t i T. Then, if λ k and x t i 2 1 for all t and i, we have n x t i 2 V 1 Proof. We have 2 log det V n log det V 2d logtracev + nk/d 2 log det V. log detv n d logtracev + nk/d. Based on Lemma 4.1, Lemma 4.2 and the two assumptions on the expected reward, we are now ready to prove our main theorem. Proof. Theorem 4.1 By Lemma 4.1, we have ˆr t i rt i holds simultaneously for all t [n] and i [m] with probability at least 1 δ. Now, assume that this random event holds and apply the monotonicity property of the expected reward, for any super arm S S t, we have f ˆrt,X t S f r S. Let S t S t be the super arm returned by the oracle S t = O St ˆr t, X t on round t. We now show that f ˆrt,X t S t αopt r. To see this, we denote St = arg max S St f r S as the maximizer of f r and Ŝ t as the optimal solution of arg max S St f ˆrt,X t S. Then, we have f ˆrt,X t S t αopt ˆrt,X t = αf ˆrt,X t Ŝt αf ˆrt,X t S t αf r S t = αopt r, detv n where we have used the definition of α-approximation = det V n 1 + oracle and the optimality of Ŝt. x n ix n i T Now, we can bound α-regret at round t as follows, i S n = detv n 1 det I + V 1/2 n 1 x niv 1/2 n 1 x ni T Reg α t = αopt r f r S t i S n f ˆrt,X t S t f r = detv n 1 det I + S t x n i 2 C ˆr V 1 t i rt i 2 i S n n = detv 1 + x t i 2 V 1. C 4αt 2 x t i 2 V 1, i S t Now, using the fact that u 2 log1 + u for any u [0, 1] and that x t i 2 V 1 x t i 2 /λ min V where the second inequality follows from the Lipschitz continuity property of the expected reward f r,x.

Therefore, with probability at least 1 δ, for all n 0, n Reg α t n n Reg a t 2 n C8n 4αt 2 x t i 2 V 1 n Cα n 32n x t i 2 V 1 Cα n 32n 2d logλ + nm/d 2d log λ = C 64nd log1 + nm/dλ d log 1 + nm/λ/δ + λs, where the last inequality follows from Lemma 4.2, the fact that S t m for all t and that V = λi. 5 Application: Online Diversified Movie Set Recommendation In this section, we consider an application with substantial practical interest: diversified movie recommendation. In this application, the recommender system recommends sets of movies, rather than individual ones. In addition, the recommended movies should be diversified such that the coverage of information that interests users is maximized. Furthermore, the recommender system need to use the user s feedback to improve its performance for future recommendations. This application can be naturally formulated as a contextual combinatorial problem as follows. Suppose, on each round t, there are m available movies and each movie is represented as a feature vector x t i R d. We can view the m movies as m arms and regard the feature vectors of movies as the feature vectors associated with arms. Then, the parameter θ R d corresponds to the user s unknown preference and the scores r t i are the ratings given by the user. On each round t, the system need to recommend a set of exactly k movies. This cardinality constraint is equivalent to assign the set of allowed super arms S t = {S S 2 [m] and S = k} to be the set of all subsets of size k for all t 0. Next, we define the expected reward f r,x S of a super arm S and construct an α-approximation oracle that associates to the expected reward. The definition of reward of super arm S should reflect both relevance and diversity of the set of movies in the super arm. In this paper, we consider the following definition of reward which is proposed recently by Qin and Zhu [22], 5.3 f r,x S = i S ri + λhs, X, where hs, X = 1 2 S log2πe+ 1 2 log detxst XS+ σ 2 I is called entropy regularizer since it quantifies the posterior uncertainty of ratings of movies in the set S. Here, the matrix XS R d S denotes a submatrix of X that consists of columns indexed by S and σ 2 is a smoothing parameter. This definition of entropy regularizer is derived as the differential entropy of ratings based on the Probabilistic Matrix Factorization PMF model. The derivation is omitted here due to space constraint and we refer interested readers to [22] for details. Finally, the parameter λ of Eq. 5.3 is a regularization constant which trades-off between relevance and diversity. As shown in [22], this definition of reward has several desirable properties. First, the value of entropy regularizer hs, X is maximized if the feature vectors of movies in S are orthogonal most dissimlar and is minimized when the feature vectors are linearly dependent most similar. This property captures the intuition of the diversity of a set of feature vectors. Second, the function f r,x is submodular and monotone for σ 2 0.0586. Consequently, there exists efficient approximation algorithms with rigorous guarantees to solve the combinatorial maximization problem of finding the super arm with the highest expected reward, which can be formulated as arg max S =k f r,x S. In particular, a simple greedy algorithm is guaranteed to find the super arm with reward larger than 1 1/eOP T, where OP T is the reward of the best super arm [20, 22]. By definition, this algorithm can be employed as a valid 1 1/e-approximation oracle in the contextual combinatorial bandit framework. The detailed implementation of the greedy algorithm as an 1 1/e-approximation oracle O div r, X is shown in Algorithm 2. The time complexity of Algorithm 2 is Ok 4, which is acceptable in most applications where k is a small constant ranging from 5 to 20. Now, we can plug this oracle O div k r, X in C 2 UCB to construct an algorithm for the online diversified movie recommendation application. This can be done by simply changing Line 9 of Algorithm 1 to S t O div k ˆr t, X t. We denote the resulting algorithm as C 2 UCB div, whose total time complexity is Ond 3 + md + k 4. To rigorously establish the theoretical guarantees for C 2 UCB div algorithm, we remain to verify whether the k

expected reward f r,x defined in Eq 5.3 satisfies the monotonicity and Lipschitz continuity properties, which are required by Theorem 4.1. The monotonicity property is straightforward since f r,x depends on r only through i S ri, which is clearly monotone with respect to r. On the other hand, for any set S of size k and any collection of feature vectors X, we have f r,x S f r,xs = ri r i i S k ri r i 2. i S Hence, the expected reward f r,x satisfies the Lipschitz continuity property with Lipschitz constant C = k. Therefore, by Theorem 4.1, we can immediate obtain the 1 1/e-regret bound of recommending diversified movie sets using C 2 UCB div as k 64nd log1 + nm/dλ λs + 2 log1/δ + d log1 + nm/λd. Algorithm 2 O div k r, X: a 1 1/e-approximation oracle for diversified movie set recommendation 1: input: scores r R m, feature vectors X R d m 2: S, C σ 2 3: for j = 1 to k do 4: for i [m]\s do 5: Σ is xi T XS 6: δ R i r i 7: δ g i 1 2 log2πeσ2 + Σ is CΣ T is 8: end for 9: i arg max δ R i + λδ g i i [m]\s 10: S S {i } 11: C XS T XS + σ 2 I 1 12: end for 13: return S 6 Experiments 6.1 Experiment Setup We conduct experiments on the MovieLens dataset, which is a public dataset consisting 1,000,029 ratings for 3900 movies by 6040 users of online movie recommendation service [19]. Each element of the dataset is represented by a tuple t i,j = u i, v j, r i,j, where u i denotes userid, v j denotes movieid, and r i,j which is an integer score between 1 and 5 denotes the rating of user i for movie j higher score indicates higher preference. We split the dataset into training and test set as follows. We construct the test set by randomly selecting 300 users such that each selected user has at least 100 ratings. The remaining 5740 users and their ratings belong to the training set. Then, we apply a rank-d probabilistic matrix factorization PMF algorithm on the training data to learn the feature vectors of movies each feature vector is d-dimensional. These feature vectors will be used later by the bandit algorithms as the feature vectors of arms. Baselines. We compare our combinatorial contextual bandit algorithm with the following baselines. k-linucb algorithm. LinUCB [14] is a contextual bandit algorithm which recommends exactly one arm at each time. To recommend a set of k movies, we repeat LinUCB algorithm k times on each round. By sequentially removing recommended arms, we ensure the k arms returned by LinUCB are distinct on each round. Finally, we highlight that the resulting bandit algorithm can be regarded as a combinatorial contextual bandit with linear expected reward function f r,x S = i S ri. Therefore, the major difference between k-linucb algorithm and our C 2 UCB div algorithm, which uses a reward function defined in Eq. 5.3, lies in that our algorithm optimizes the diversity of arms in set S t. Warm-start diversified movie recommendation. We denote this baseline as warm-start for short. For each user u in test set, we randomly select η ratings to train an user preference vector using PMF model. We call the parameter η as warm-start offset. With the estimated preference vector, one can repeatedly recommend sets of diverse recommendation results by maximizing the reward function 5.3. Note that this method cannot dynamically adapts to user s feedback, and thus each round is independent with the others. Metric. We use precision to measure the quality of recommended movie sets over n rounds. Specifically, for each user u in the test test, we define the set of preferred movies L u as the set of movies which user u assigned a rating of 4 or 5. Intuitively, a good movie set recommendation algorithm should recommend movie sets which cover a large fraction of preferred movies. Formally, on round t, suppose the recommendation algorithm recommends a set of movies S t. The precision p t,u of user u on round t is defined as p t,u = S t L u. S t

Then, the average precision of P t of all test users up to round t is given by P t = 1 t U u U t p i,u. i=1 Precision 0.91 0.85 0.8 0.75 0.7 0.65 0.6 C 2 UCB div warm-start all warm-start 2k k-linucb 0.91 0.85 0.8 0.75 0.7 0.65 0.6 C 2 UCB div warm-start all warm-start 2k k-linucb Note that we do not aim at predicting the ratings of movies, but to provide more satisfying recommendation lists. Hence, precision is a more appropriate metric rather than the root mean square error RMSE. Actually, our algorithm as well as baselines essentially used an l 2 -regularized linear regression method to predict movie ratings based on existing observations. This is equivalent to the rating prediction methods used by many matrix factorization algorithms, which are shown to have low RMSEs [18]. Moreover, we cannot use regret as a metric either, because the definitions of regrets vary greatly for different bandit algorithms. 6.2 Experiment Results We consider recommending different number of movies to each user on each round, i.e., the size of super arm k takes values in {5, 10, 15, 20}. For each k, we set the exploration parameter α t = 1.0. The parameters of entropy regularizer are set to be λ = 0.5 and σ = 1.0. For the warm-start baseline that allows an offlineestimated user preference, we consider two cases where η takes different values. In one case that we denote as warm-start 2k, η = k 2 which indicates the method can access ratings of two rounds. In the other case that we denote as warm-start all, η equals the total amount of observations, which indeed corresponds to the best solution, i.e., all observations are used to train user preference. The results are shown in Figure 1 and Table 1, where our approach is denoted as C 2 UCB div. We can see that, in all cases, the warm-start 2k baseline outperforms bandit algorithms on earlier rounds, which is reasonable since the warm-start 2k baseline is provided warm-start observations to learn the user preference. But when more user feedbacks are available, bandit algorithms improve performance by dynamically adapting to user feedbacks. Near the end of 10 rounds, C 2 UCB div can achieve a result that is comparable to warm-start all. Compared to k-linucb, our method C 2 UCB div finds a better match between recommended movies and user interest i.e., the movies liked by each given user, and thus improves the overall performance. Furthermore, when k is larger, C 2 UCB div algorithm obtains larger performance gain over k-linucb algorithm. The experiment results indicate that our method helps Precision 0.52 0.52 1 2 3 4 5 6 7 8 9 10 1 2 3 4 5 6 7 8 9 round round a k=5 b k=10 0.91 0.85 0.8 0.75 0.7 0.65 0.6 C 2 UCB div warm-start all warm-start 2k k-linucb 0.52 0.52 1 2 3 4 5 6 7 8 9 10 1 2 3 4 5 6 7 8 9 round round c k=15 d k=20 0.91 0.85 0.8 0.75 0.7 0.65 0.6 C 2 UCB div warm-start all warm-start 2k 10 k-linucb Figure 1: Experiment results comparing C 2 UCB div with k-linucb, warm-start 2t and warm-start all on different choices of k. t = 5 t = 10 warm-start KL CC KL CC η = 2k η = all k = 5 0.785 0.810 0.831 0.861 0.763 0.884 k = 10 0.779 0.806 0.814 0.851 0.745 0.862 k = 15 0.770 0.793 0.803 0.836 0.720 0.841 k = 20 0.762 0.784 0.791 0.822 0.692 0.824 Table 1: Precision values of competing algorithms. CC: C 2 UCB div algorithm. KL: k-linucb algorithm. uncover users diverse interest by using a non-linear diversity promoting reward function. 7 Conclusion We presented a general framework called contextual combinatorial bandit that accommodates combinatorial nature of contextual arms. We developed an efficient algorithm C 2 UCB for contextual combinatorial bandit and provide a rigorously regret analysis. We further applied this framework on online diversified movie recommendation task, and developed a specific algorithm C 2 UCB div for this application. Experiments on public MovieLens dataset demonstrate that our approach helps explore and exploit users diverse preference, and hence improves the performance of recommendation task. Acknowledgments This work is supported by China National Basic Research Program No. 2012CB316301, China National Basic Research Program No. 2013CB329403, and the Chinese Natural Science Foundation important project under grant No.61332007. 10

References [1] Y. Abbasi-Yadkori, C. Szepesvári, and D. Tax, Improved algorithms for linear stochastic bandits, in Advances in Neural Information Processing Systems, 2011, pp. 2312 2320. [2] V. Anantharam, P. Varaiya, and J. Walrand, Asymptotically efficient allocation rules for the multiarmed bandit problem with multiple plays-part i: Iid rewards; part ii: Markoveian rewards, Automatic Control, IEEE Transactions on, 32 1987, pp. 968 982. [3] P. Auer, N. Cesa-Bianchi, and P. Fischer, Finitetime analysis of the multiarmed bandit problem, Machine learning, 47 2002, pp. 235 256. [4] D. M. Berry, Bandit problems, 1985. [5] D. Bouneffouf, A. Bouzeghoub, and A. L. Gançarski, A contextual-bandit algorithm for mobile context-aware recommender system, in Neural Information Processing, Springer, 2012, pp. 324 331. [6] R. Burke, Hybrid recommender systems: Survey and experiments, User modeling and user-adapted interaction, 12 2002, pp. 331 370. [7] W. Chen, Y. Wang, and Y. Yuan, Combinatorial multi-armed bandit: General framework and applications, in Proceedings of the 30th International Conference on Machine Learning ICML-13, 2013, pp. 151 159. [8] W. Chu, L. Li, L. Reyzin, and R. E. Schapire, Contextual bandits with linear payoff functions, in International Conference on Artificial Intelligence and Statistics, 2011, pp. 208 214. [9] V. H. de la Peña, M. J. Klass, and T. L. Lai, Selfnormalized processes: exponential inequalities, moment bounds and iterated logarithm laws, Annals of probability, 2004, pp. 1902 1933. [10] Y. Gai, B. Krishnamachari, and R. Jain, Learning multiuser channel allocations in cognitive radio networks: A combinatorial multi-armed bandit formulation, in New Frontiers in Dynamic Spectrum, 2010 IEEE Symposium on, IEEE, 2010, pp. 1 9. [11] Y. Gai, B. Krishnamachari, and R. Jain, Combinatorial network optimization with unknown variables: Multi-armed bandits with linear rewards and individual observations, IEEE/ACM Transactions on Networking TON, 20 2012, pp. 1466 1478. [12] Y. Koren, R. Bell, and C. Volinsky, Matrix factorization techniques for recommender systems, Computer, 42 2009, pp. 30 37. [13] T. L. Lai and H. Robbins, Asymptotically efficient adaptive allocation rules, Advances in applied mathematics, 6 1985, pp. 4 22. [14] L. Li, W. Chu, J. Langford, and R. E. Schapire, A contextual-bandit approach to personalized news article recommendation, in Proceedings of the 19th international conference on World wide web, ACM, 2010, pp. 661 670. [15] L. Li, W. Chu, J. Langford, and X. Wang, Unbiased offline evaluation of contextual-bandit-based news article recommendation algorithms, in Proceedings of the fourth ACM international conference on Web search and data mining, ACM, 2011, pp. 297 306. [16] P. Massa and P. Avesani, Trust-aware recommender systems, in Proceedings of the 2007 ACM conference on Recommender systems, ACM, 2007, pp. 17 24. [17] S. E. Middleton, N. R. Shadbolt, and D. C. De Roure, Ontological user profiling in recommender systems, ACM Transactions on Information Systems TOIS, 22 2004, pp. 54 88. [18] A. Mnih and R. Salakhutdinov, Probabilistic matrix factorization, in Advances in neural information processing systems, 2007, pp. 1257 1264. [19] MovieLens dataset, in http://movielens.org. [20] G. Nemhauser, L. Wolsey, and M. Fisher, An analysis of approximations for maximizing submodular set functionsi, Mathematical Programming, 14 1978, pp. 265 294. [21] V. H. Peña, H. Víctor, T. L. Lai, and Q.-M. Shao, Self-Normalized Processes: Limit Theory and Statistical Applications, Springer, 2009. [22] L. Qin and X. Zhu, Promoting diversity in recommendation by entropy regularizer, in Proceedings of the Twenty-third international joint conference on Artificial Intelligence, AAAI Press, 2013, pp. 2698 2704. [23] B. Sarwar, G. Karypis, J. Konstan, and J. Riedl, Item-based collaborative filtering recommendation algorithms, in Proceedings of the 10th international conference on World Wide Web, ACM, 2001, pp. 285 295. [24] A. I. Schein, A. Popescul, L. H. Ungar, and D. M. Pennock, Methods and metrics for cold-start recommendations, in Proceedings of the 25th annual international ACM SIGIR conference on Research and development in information retrieval, ACM, 2002, pp. 253 260. [25] C. Yu, L. Lakshmanan, and S. Amer-Yahia, It takes variety to make a world: diversification in recommender systems, in Proceedings of the 12th international conference on extending database technology: Advances in database technology, ACM, 2009, pp. 368 378. [26] Y. Yue and C. Guestrin, Linear submodular bandits and their application to diversified retrieval, in Advances in Neural Information Processing Systems, 2011, pp. 2483 2491.