Graphs in Machine Learning

Size: px
Start display at page:

Download "Graphs in Machine Learning"

Transcription

1 Graphs in Machine Learning Michal Valko Inria Lille - Nord Europe, France Partially based on material by: Toma s Koca k November 23, 2015 MVA 2015/2016

2 Last Lecture Examples of applications of online SSL Analysis of online SSL SSL Learnability When does graph-based SSL provably help? Scaling harmonic functions to millions of samples Michal Valko Graphs in Machine Learning SequeL - 2/40

3 Previous Lab Session by Daniele Calandriello Content Semi-supervised learning Graph quantization Online face recognizer Short written report Questions to piazza Deadline: ~calandri/teaching.html Michal Valko Graphs in Machine Learning SequeL - 3/40

4 This Lecture Online decision-making on graphs Graph bandits Smoothness of rewards (preferences) on a given graph Observability graphs Exploiting side information Michal Valko Graphs in Machine Learning SequeL - 4/40

5 Final Class projects detailed description on the class website preferred option: you come up with the topic theory/implementation/review or a combination one or two people per project (exceptionally three) grade 60%: report + short presentation of the team deadlines strongly recommended DL for taking projects hard DL for taking projects submission of the project report (or later) - project presentation list of suggested topics on piazza Michal Valko Graphs in Machine Learning SequeL - 5/40

6 Online Decision Making on Graphs Michal Valko Graphs in Machine Learning SequeL - 6/40

7 Online Decision Making on Graphs: Smoothness Sequential decision making in structured settings we are asked to pick a node (or a few nodes) in a graph the graph encodes some structural property of the setting goal: maximize the sum of the outcomes application: recommender systems First application: Exploiting smoothness fixed graph iid outcomes neighboring nodes have similar outcomes Michal Valko Graphs in Machine Learning SequeL - 7/40

8 Online Decision Making on Graphs Movie recommendation: (in each time step) Recommend movies to a single user. Good prediction after a few steps (T N). Goal: Maximize overall reward (sum of ratings). Assumptions: Unknown reward function f : V (G) R. Function f is smooth on a graph. Neighboring movies similar preferences. Similar preferences neighboring movies. The Shawshank Redemption (1994) The Godfather (1972) The Godfather: Part II (1974) The Dark Knight (2008) Pulp Fiction (1994) The Good, the Bad and the Ugly (1966) Schindler s List (1993) 12 Angry Men (1957) The Lord of the Rings: The Return of the King (2003) Fight Club (1999) The Lord of the Rings: The Fellowship of the Ring (2001) Star Wars: Episode V - The Empire Strikes Back (1980) Inception (2010) Forrest Gump (1994) One Flew Over the Cuckoo s Nest (1975) The Lord of the Rings: The Two Towers (2002) Goodfellas (1990) Star Wars: Episode IV - A New Hope (1977) The Matrix (1999) Seven Samurai (1954) City of God (2002) Se7en (1995) The Usual Suspects (1995) The Silence of the Lambs (1991) Once Upon a Time in the West (1968) It s a Wonderful Life (1946) Léon: The Professional (1994) Casablanca (1942) Life Is Beautiful (1997) Raiders of the Lost Ark (1981) American History X (1998) Psycho (1960) Rear Window (1954) City Lights (1931) Saving Private Ryan (1998) Spirited Away (2001) The Intouchables (2011) Memento (2000) Terminator 2: Judgment Day (1991) Modern Times (1936) 1 0 Michal Valko Graphs in Machine Learning SequeL - 8/40

9 Let s be lazy: Ignore the structure! This is an multi-armed bandit problem! The performance depends on the number of movies (N arms). NT Worst case regret (to the best fixed strategy) RT = O What is N for imdb.com? 3,538,545 Michal Valko Graphs in Machine Learning SequeL - 9/40

10 Let s be lazy: Ignore the structure! Another problem of the typical bandits strategies for recommendation? If there is no information shared, we need to try all of the options! UCB/MOSS and likely TS start with pulling each of the arms once This is a problem both algorithmically and theoretically.... Watch all the movies and then I tell you which one you like.... What do we need for movie recommendation? An algorithm useful in the case T N! Exploiting the structure is a must! Michal Valko Graphs in Machine Learning SequeL - 10/40

11 Recap: Smooth graph functions f = (f 1,..., f N ) T : Vector of function values. Let L = QΛQ T be the eigendecomposition of the Laplacian. Diagonal matrix Λ whose diagonal entries are eigenvalues of L. Columns of Q are eigenvectors of L. Columns of Q form a basis. α : Unique vector such that Qα = f Note: Q T f = α S G (f) = f T Lf = f T QΛQ T f = α T Λα = α 2 Λ = N λ i (αi ) 2 i=1 Smoothness and regularization: Small value of (a) S G (f) (b) Λ norm of α (c) α i for large λ i Michal Valko Graphs in Machine Learning SequeL - 11/40

12 Smooth graph functions: Flixster eigenvectors 0.2 Eigenvector Eigenvector Eigenvector Eigenvector Eigenvector Eigenvector Eigenvector 7 Eigenvector 8 Eigenvector 9 Eigenvectors 0.2 from the Flixster 0.2 data corresponding 0.2 to the smallest few eigenvalues of the graph Laplacian projected onto the first principal 0 component 0of data. Colors indicate 0 the values Michal 0 Valko 1 Graphs 1 in Machine0 Learning SequeL 1-12/40

13 Online Learning Setting - Bandit Problem Learning setting for a bandit algorithm π In each time t step choose a node π(t). the π(t)-th row x π(t) of the matrix Q corresponds to the arm π(t). Obtain noisy reward r t = xπ(t) T α + ε t. Note: xπ(t)α T = f π(t) ε t is R-sub-Gaussian noise. ξ R, E[e ξεt ] exp ( ξ 2 R 2 /2 ) Minimize cumulative regret R T = T max (xaα T ) a T xπ(t) T α. t=1 What is a good result? Can t we just use linear bandits? Michal Valko Graphs in Machine Learning SequeL - 13/40

14 Online Decision Making on Graphs: Smoothness Linear bandit algorithms LinUCB (Li et al., 2010) Regret bound D T ln T LinearTS (Agrawal and Goyal, 2013) Regret bound D T ln N Note: D is ambient dimension, in our case N, length of x i. Number of actions, e.g., all possible movies HUGE! Spectral bandit algorithms SpectralUCB (Valko et al., ICML 2014) Regret bound d T ln T Operations per step: D 2 N SpectralTS (Kocák et al., AAAI 2014) Regret bound d T ln N Operations per step: D 2 + DN Note: d is effective dimension, usually much smaller than D. Michal Valko Graphs in Machine Learning SequeL - 14/40

15 Effective dimension Effective dimension: Largest d such that (d 1)λ d T log(1 + T /λ). Function of time horizon and graph properties λ i : i-th smallest eigenvalue of Λ. λ: Regularization parameter of the algorithm. Properties: d is small when the coefficients λ i grow rapidly above time. d is related to the number of non-negligible dimensions. Usually d is much smaller than D in real world graphs. Can be computed beforehand. Michal Valko Graphs in Machine Learning SequeL - 15/40

16 Effective dimension vs. Ambient dimension 7 Barabasi Albert graph N= Flixster graph: N= effective dimenstion effective dimenstion time T d D time T Note: In our setting T < N = D. Michal Valko Graphs in Machine Learning SequeL - 16/40

17 UCB-style algorithms: Estimate Expected reward Michal Valko Graphs in Machine Learning SequeL - 17/40

18 UCB-style algorithms: Sample Expected reward Michal Valko Graphs in Machine Learning SequeL - 17/40

19 UCB-style algorithms: Estimate... Expected reward Michal Valko Graphs in Machine Learning SequeL - 17/40

20 SpectralUCB Given a vector of weights α, we define its Λ norm as α Λ = N λ k αk 2 = α T Λα, k=1 and fit the ratings r v with a (regularized) least-squares estimate ( t ) α t = arg min α [ x v, α r v ] 2 + α 2 Λ. v=1 α Λ is a penalty for non-smooth combinations of eigenvectors. Michal Valko Graphs in Machine Learning SequeL - 18/40

21 SpectralUCB 1: Input: 2: N, T, {Λ L, Q}, λ, δ, R, C 3: Run: 4: Λ Λ L + λi 5: d max{d : (d 1)λ d T / ln(1 + T /λ)} 6: for t = 1 to T do 7: Update the basis coefficients α: 8: X t [x π(1),..., x π(t 1) ] T 9: r [r 1,..., r t 1] T 10: V t X tx T t + Λ 11: 12: α t V 1 t X T t r c t 2R d ln(1 + t/λ) + 2 ln(1/δ) + C 13: π(t) arg max a (xa T α + c t x a V 1 t 14: Observe the reward r t 15: end for ) Michal Valko Graphs in Machine Learning SequeL - 19/40

22 SpectralUCB: Synthetic experiment cumulative regret Barabasi Albert N=250, basis size=3, effective d=1 SpectralEliminator SpectralUCB LinUCB time T Michal Valko Graphs in Machine Learning SequeL - 20/40

23 SpectralUCB: Movie data experiments cumulative regret Movielens: Cumulative regret for randomly sampled users. T = 100 SpectralUCB LinUCB cumulative regret Flixster: Cumulative regret for randomly sampled users. T = 100 SpectralUCB LinUCB Michal Valko Graphs in Machine Learning SequeL - 21/40

24 SpectralUCB: Regret Bound d: Effective dimension. λ: Minimal eigenvalue of Λ = Λ L + λi. C: Smoothness upper bound, α Λ C. x T i α [ 1, 1] for all i. The cumulative regret R T of SpectralUCB is with probability 1 δ bounded as R T ( 8R d ln λ + T λ + 2 ln 1 δ + 4C + 4 ) dt ln λ + T λ. R T d T ln T Michal Valko Graphs in Machine Learning SequeL - 22/40

25 SpectralUCB: Regret Bound Derivation of the confidence ellipsoid for α with probability 1 δ. Using analysis of OFUL (Abbasi-Yadkori et al., 2011) x T ( α α ) x V 1 t ( ( ) ) R 2 ln Vt 1/2 + C δ Λ 1/2 Regret in one time step: r t = x α T xπ(t) T α 2c t x π(t) V 1 t Cumulative regret: R T = T r t T T rt 2 2( c T + 1) t=1 t=1 Upperbound for ln( V t / Λ ) ln V t Λ ln V T Λ 2T ln V T Λ ( ) λ + T 2d ln λ Michal Valko Graphs in Machine Learning SequeL - 23/40

26 SpectralUCB: Regret Bound Sylvester s determinant theorem: A + xx T = A I + A 1 xx T = A (1 + x T A 1 x) Goal: Upperbound determinant A + xx T for x 2 1 Upperbound x T A 1 x x T A 1 x = x T QΛ 1 Q T x = y T Λ 1 y = N i=1 λ 1 i yi 2 y 2 1. y is a canonical vector. x = Qy is an eigenvector of A. Michal Valko Graphs in Machine Learning SequeL - 24/40

27 SpectralUCB: Regret Bound Corollary: Determinant V T of V T = Λ + T t=1 x tx T t is maximized when all x t are aligned with axes. V T max ti =T (λi + t i ) ln V T Λ ln V T Λ ( max ln 1 + t i ti =T d i=1 ( ln 1 + T ) λ ( d ln 1 + T λ ( 2d ln 1 + T λ + λ i ) + T ) ) N i=d+1 λ d+1 ( ln 1 + t ) i λ d+1 Michal Valko Graphs in Machine Learning SequeL - 25/40

28 SpectralUCB: Improving the running time Reduced basis: We only need first few eigenvectors. Getting J eigenvectors: O(Jm log m) time for m edges Computationally less expensive, comparable performance. cumulative regret J = 20 J = 200 J = 2000 computational time in seconds time T 10 0 J = 20 J=200 J=2000 Michal Valko Graphs in Machine Learning SequeL - 26/40

29 SpectralUCB: How to make it even faster? UCB-style algorithms need to (re)-compute UCBs every t Can be a problem for large set of arms D 2 N N 3 Optimistic (UCB) approach vs. Thompson Sampling Play the arm maximizing probability of being the best Sample α from the distribution N ( α, v 2 V 1 ) Play arm which maximizes x T α and observe reward Compute posterior distribution according to reward received Only requires D 2 + DN N 2 per step update Michal Valko Graphs in Machine Learning SequeL - 27/40

30 Thomson Sampling: Estimate α ˆα Michal Valko Graphs in Machine Learning SequeL - 28/40

31 Thomson Sampling: Sample α α ˆα Michal Valko Graphs in Machine Learning SequeL - 28/40

32 Thomson Sampling: Estimate ˆα α Michal Valko Graphs in Machine Learning SequeL - 28/40

33 Thomson Sampling: Sample ˆα α α Michal Valko Graphs in Machine Learning SequeL - 28/40

34 Thomson Sampling: Estimate... ˆα α Michal Valko Graphs in Machine Learning SequeL - 28/40

35 SpectralTS for Graphs 1: Input: 2: N, T, {Λ L, Q}, λ, δ, R, C 3: Initialization: 4: v = R 6d log((λ + T )/δλ) + C 5: α = 0 N 6: f = 0 N 7: V = Λ L + λi N 8: Run: 9: for t = 1 to T do 10: Sample α N ( α, v 2 V 1 ) 11: π(t) arg max a xa T α 12: Observe a noisy reward r(t) = xπ(t)α T + ε t 13: f f + x π(t) r(t) 14: Update V V + x π(t) x T π(t) 15: Update α V 1 f 16: end for Michal Valko Graphs in Machine Learning SequeL - 29/40

36 SpectralTS: Regret bound d: Effective dimension. λ: Minimal eigenvalue of Λ = Λ L + λi. C: Smoothness upper bound, α Λ C. x T i α [ 1, 1] for all i. The cumulative regret R T of SpectralTS is with probability 1 δ bounded as R T 11g 4 + 4λ dt log λ + T + 1 p λ λ T + g ( ) T log 2 p λ δ, where p = 1/(4e π) and g = ( ( λ + T 4 log TN R 6d log δλ ) ) + C + R 2d log ( (λ + T )T 2 δλ ) + C. R T d T log N Michal Valko Graphs in Machine Learning SequeL - 30/40

37 SpectralTS: Analysis sketch Divide arms into two groups i = x T α x T i α g x i V 1 t i = x T α x T i α > g x i V 1 t Saturated arm arm i is unsaturated arm i is saturated Small standard deviation accurate regret estimate. High regret on playing the arm Low probability of picking Unsaturated arm Low regret bounded by a factor of standard deviation High probability of picking Michal Valko Graphs in Machine Learning SequeL - 31/40

38 SpectralTS: Analysis sketch Confidence ellipsoid for estimate µ of µ (with probability 1 δ/t 2 ) Using analysis of OFUL algorithm (Abbasi-Yadkori et al., 2011) ( ) ) (λ + T )T xi T α xi (R T α 2 2 d log + C x i δλ V 1 = l x i t V 1 t The key result coming from spectral properties of V t. log Vt Λ ( 2d log 1 + T ) λ Concentration of sample α around mean α (with probability 1 1/T 2 ) Using concentration inequality for Gaussian random variable. ( ( λ + T xi T α xi T α R 6d log δλ ) ) + C x i V 1 t 4 log(tn) = v xi V 1 4 log(tn) t Michal Valko Graphs in Machine Learning SequeL - 32/40

39 SpectralTS: Analysis sketch Define regret (t) = regret(t) 1{ xi T α(t) xt i α l x i V 1} t regret (t) 11g p x a(t) V t T 2 Super-martingale (i.e. E[Y t Y t 1 F t 1 ] 0) X t = regret (t) 11g p x a(t) V 1 t t Y t = X w. w=1 1 T 2 (Y t ; t = 0,..., T ) is a super-martingale process w.r.t. history F t. Azuma-Hoeffding inequality for super-martingales, w.p. 1 δ/2: T regret (t) 11g T x a(t) p V t T + g ( ) T ln 2 p λ δ t=1 t=1 Michal Valko Graphs in Machine Learning SequeL - 33/40

40 Spectral Bandits: Synthetic experiment cumulative regret Barabasi Albert N=250, basis size=3, effective d=1 SpectralTS LinearTS SpectralUCB LinUCB computational time in seconds time T SpectralTS LinearTS SpectralUCB LinUCB Michal Valko Graphs in Machine Learning SequeL - 34/40

41 Spectral Bandits: Real world experiment MovieLens dataset of 6k users who rated one million movies. cumulative regret Movielens data N=2019, average of 10 users, T=200, d = 5 SpectralUCB LinUCB SpectralTS LinearTS computational time in seconds time t SpectralTS LinearTS SpectralUCB LinUCB Michal Valko Graphs in Machine Learning SequeL - 35/40

42 Spectral Bandits Summary Spectral bandit setting (smooth graph functions). SpectralUCB Regret bound RT = Õ ( d ) T ln T SpectralTS ( Regret bound RT = Õ d ) T ln N Computationally more efficient. SpectralEliminator ( dt ) Regret bound RT = Õ ln T Better upper, empirically does not seem to work well (yet) Bounds scale with effective dimension d D. Michal Valko Graphs in Machine Learning SequeL - 36/40

43 SpectralEliminator: Pseudocode Input: N : the number of nodes, T : the number of pulls {Λ L, Q} spectral basis of L λ : regularization parameter β, {t j } J j parameters of the elimination and phases A 1 {x 1,..., x K }. for j = 1 to J do V tj γλ L + λi for t = t j to min(t j+1 1, T ) do Play x t A j with the largest width to observe r t : x t arg max x Aj x V 1 t V t+1 V t + x t xt T end for Eliminate the arms that are not promising: α t V 1 t { [x tj,..., x t ][r tj,..., r t ] T A j+1 x A j, α t, x + x V 1 t end for ]} β max x Aj [ α t, x x V 1β t Michal Valko Graphs in Machine Learning SequeL - 37/40

44 SpectralEliminator: Analysis SpectralEliminator Divide time into sets (t 1 = 1 t 2... ) to introduce independence for Azuma-Hoeffding inequality and observe R T J j=0 (t j+1 t j ) [ x x t, α j + ( x V 1 j Bound x x t, α j for each phase No bad arms: x x t, α j ( x V 1 j By algorithm: x 2 V 1 j t j s=t j 1 +1 min (1, x s 2 V 1 s 1 1 tj t j t j 1 ) log V j Λ + x t V 1)β j s=t j 1 +1 x s 2 V 1 + x t V 1)β ] j s 1 Michal Valko Graphs in Machine Learning SequeL - 38/40

45 Spectral Bandits: Is it possible to do better? Is d a good quantity that embodies the difficulty? Lower bound! For any d, we construct a graph that for any reasonable algorithm, the regret is at least Ω( dt ). How? By reduction to d-arm bandits problem. K MT K MT K MT weights ε K MT K MT K MT Michal Valko Graphs in Machine Learning SequeL - 39/40

46 SequeL Inria Lille MVA 2015/2016 Michal Valko sequel.lille.inria.fr

Graphs in Machine Learning

Graphs in Machine Learning Graphs in Machine Learning Michal Valko INRIA Lille - Nord Europe, France Partially based on material by: Rob Fergus, Tomáš Kocák March 17, 2015 MVA 2014/2015 Last Lecture Analysis of online SSL Analysis

More information

Spectral Bandits for Smooth Graph Functions with Applications in Recommender Systems

Spectral Bandits for Smooth Graph Functions with Applications in Recommender Systems Spectral Bandits for Smooth Graph Functions with Applications in Recommender Systems Tomáš Kocák SequeL team INRIA Lille France Michal Valko SequeL team INRIA Lille France Rémi Munos SequeL team, INRIA

More information

Two generic principles in modern bandits: the optimistic principle and Thompson sampling

Two generic principles in modern bandits: the optimistic principle and Thompson sampling Two generic principles in modern bandits: the optimistic principle and Thompson sampling Rémi Munos INRIA Lille, France CSML Lunch Seminars, September 12, 2014 Outline Two principles: The optimistic principle

More information

Multi-task Linear Bandits

Multi-task Linear Bandits Multi-task Linear Bandits Marta Soare Alessandro Lazaric Ouais Alsharif Joelle Pineau INRIA Lille - Nord Europe INRIA Lille - Nord Europe McGill University & Google McGill University NIPS2014 Workshop

More information

Online learning with noisy side observations

Online learning with noisy side observations Online learning with noisy side observations Tomáš Kocák Gergely Neu Michal Valko Inria Lille - Nord Europe, France DTIC, Universitat Pompeu Fabra, Barcelona, Spain Inria Lille - Nord Europe, France SequeL

More information

Efficient learning by implicit exploration in bandit problems with side observations

Efficient learning by implicit exploration in bandit problems with side observations Efficient learning by implicit exploration in bandit problems with side observations Tomáš Kocák, Gergely Neu, Michal Valko, Rémi Munos SequeL team, INRIA Lille - Nord Europe, France SequeL INRIA Lille

More information

Graphs in Machine Learning

Graphs in Machine Learning Graphs in Machine Learning Michal Valko Inria Lille - Nord Europe, France TA: Pierre Perrault Partially based on material by: Mikhail Belkin, Jerry Zhu, Olivier Chapelle, Branislav Kveton October 30, 2017

More information

Bandit Algorithms. Zhifeng Wang ... Department of Statistics Florida State University

Bandit Algorithms. Zhifeng Wang ... Department of Statistics Florida State University Bandit Algorithms Zhifeng Wang Department of Statistics Florida State University Outline Multi-Armed Bandits (MAB) Exploration-First Epsilon-Greedy Softmax UCB Thompson Sampling Adversarial Bandits Exp3

More information

Graphs in Machine Learning

Graphs in Machine Learning Graphs in Machine Learning Michal Valko INRIA Lille - Nord Europe, France Partially based on material by: Ulrike von Luxburg, Gary Miller, Doyle & Schnell, Daniel Spielman January 27, 2015 MVA 2014/2015

More information

The Multi-Arm Bandit Framework

The Multi-Arm Bandit Framework The Multi-Arm Bandit Framework A. LAZARIC (SequeL Team @INRIA-Lille) ENS Cachan - Master 2 MVA SequeL INRIA Lille MVA-RL Course In This Lecture A. LAZARIC Reinforcement Learning Algorithms Oct 29th, 2013-2/94

More information

Sparse Linear Contextual Bandits via Relevance Vector Machines

Sparse Linear Contextual Bandits via Relevance Vector Machines Sparse Linear Contextual Bandits via Relevance Vector Machines Davis Gilton and Rebecca Willett Electrical and Computer Engineering University of Wisconsin-Madison Madison, WI 53706 Email: gilton@wisc.edu,

More information

Bandits on graphs and structures

Bandits on graphs and structures Bandits on graphs and structures Michal Valko To cite this version: Michal Valko. Bandits on graphs and structures. Machine Learning [stat.ml]. École normale supérieure de Cachan - ENS Cachan, 2016.

More information

Knowledge Compilation and Machine Learning A New Frontier Adnan Darwiche Computer Science Department UCLA with Arthur Choi and Guy Van den Broeck

Knowledge Compilation and Machine Learning A New Frontier Adnan Darwiche Computer Science Department UCLA with Arthur Choi and Guy Van den Broeck Knowledge Compilation and Machine Learning A New Frontier Adnan Darwiche Computer Science Department UCLA with Arthur Choi and Guy Van den Broeck The Problem Logic (L) Knowledge Representation (K) Probability

More information

Graphs in Machine Learning

Graphs in Machine Learning Graphs in Machine Learning Michal Valko Inria Lille - Nord Europe, France TA: Pierre Perrault Partially based on material by: Ulrike von Luxburg, Gary Miller, Doyle & Schnell, Daniel Spielman October 10th,

More information

COS 402 Machine Learning and Artificial Intelligence Fall Lecture 22. Exploration & Exploitation in Reinforcement Learning: MAB, UCB, Exp3

COS 402 Machine Learning and Artificial Intelligence Fall Lecture 22. Exploration & Exploitation in Reinforcement Learning: MAB, UCB, Exp3 COS 402 Machine Learning and Artificial Intelligence Fall 2016 Lecture 22 Exploration & Exploitation in Reinforcement Learning: MAB, UCB, Exp3 How to balance exploration and exploitation in reinforcement

More information

Graphs in Machine Learning

Graphs in Machine Learning Graphs in Machine Learning Michal Valko Inria Lille - Nord Europe, France TA: Pierre Perrault Partially based on material by: Ulrike von Luxburg, Gary Miller, Mikhail Belkin October 16th, 2017 MVA 2017/2018

More information

Profile-Based Bandit with Unknown Profiles

Profile-Based Bandit with Unknown Profiles Journal of Machine Learning Research 9 (208) -40 Submitted /7; Revised 6/8; Published 9/8 Profile-Based Bandit with Unknown Profiles Sylvain Lamprier sylvain.lamprier@lip6.fr Sorbonne Universités, UPMC

More information

Graphs in Machine Learning

Graphs in Machine Learning Graphs in Machine Learning Michal Valko INRIA Lille - Nord Europe, France Partially based on material by: Mikhail Belkin, Jerry Zhu, Olivier Chapelle, Branislav Kveton February 10, 2015 MVA 2014/2015 Previous

More information

Online Learning and Sequential Decision Making

Online Learning and Sequential Decision Making Online Learning and Sequential Decision Making Emilie Kaufmann CNRS & CRIStAL, Inria SequeL, emilie.kaufmann@univ-lille.fr Research School, ENS Lyon, Novembre 12-13th 2018 Emilie Kaufmann Sequential Decision

More information

Large-scale Information Processing, Summer Recommender Systems (part 2)

Large-scale Information Processing, Summer Recommender Systems (part 2) Large-scale Information Processing, Summer 2015 5 th Exercise Recommender Systems (part 2) Emmanouil Tzouridis tzouridis@kma.informatik.tu-darmstadt.de Knowledge Mining & Assessment SVM question When a

More information

Bandit View on Continuous Stochastic Optimization

Bandit View on Continuous Stochastic Optimization Bandit View on Continuous Stochastic Optimization Sébastien Bubeck 1 joint work with Rémi Munos 1 & Gilles Stoltz 2 & Csaba Szepesvari 3 1 INRIA Lille, SequeL team 2 CNRS/ENS/HEC 3 University of Alberta

More information

Stratégies bayésiennes et fréquentistes dans un modèle de bandit

Stratégies bayésiennes et fréquentistes dans un modèle de bandit Stratégies bayésiennes et fréquentistes dans un modèle de bandit thèse effectuée à Telecom ParisTech, co-dirigée par Olivier Cappé, Aurélien Garivier et Rémi Munos Journées MAS, Grenoble, 30 août 2016

More information

Manifold Regularization

Manifold Regularization 9.520: Statistical Learning Theory and Applications arch 3rd, 200 anifold Regularization Lecturer: Lorenzo Rosasco Scribe: Hooyoung Chung Introduction In this lecture we introduce a class of learning algorithms,

More information

Regret Analysis of Stochastic and Nonstochastic Multi-armed Bandit Problems, Part I. Sébastien Bubeck Theory Group

Regret Analysis of Stochastic and Nonstochastic Multi-armed Bandit Problems, Part I. Sébastien Bubeck Theory Group Regret Analysis of Stochastic and Nonstochastic Multi-armed Bandit Problems, Part I Sébastien Bubeck Theory Group i.i.d. multi-armed bandit, Robbins [1952] i.i.d. multi-armed bandit, Robbins [1952] Known

More information

Machine Learning Theory (CS 6783)

Machine Learning Theory (CS 6783) Machine Learning Theory (CS 6783) Tu-Th 1:25 to 2:40 PM Kimball, B-11 Instructor : Karthik Sridharan ABOUT THE COURSE No exams! 5 assignments that count towards your grades (55%) One term project (40%)

More information

COMS 4721: Machine Learning for Data Science Lecture 20, 4/11/2017

COMS 4721: Machine Learning for Data Science Lecture 20, 4/11/2017 COMS 4721: Machine Learning for Data Science Lecture 20, 4/11/2017 Prof. John Paisley Department of Electrical Engineering & Data Science Institute Columbia University SEQUENTIAL DATA So far, when thinking

More information

Introduction to Bandit Algorithms. Introduction to Bandit Algorithms

Introduction to Bandit Algorithms. Introduction to Bandit Algorithms Stochastic K-Arm Bandit Problem Formulation Consider K arms (actions) each correspond to an unknown distribution {ν k } K k=1 with values bounded in [0, 1]. At each time t, the agent pulls an arm I t {1,...,

More information

Machine Learning Theory (CS 6783)

Machine Learning Theory (CS 6783) Machine Learning Theory (CS 6783) Tu-Th 1:25 to 2:40 PM Hollister, 306 Instructor : Karthik Sridharan ABOUT THE COURSE No exams! 5 assignments that count towards your grades (55%) One term project (40%)

More information

8.1 Concentration inequality for Gaussian random matrix (cont d)

8.1 Concentration inequality for Gaussian random matrix (cont d) MGMT 69: Topics in High-dimensional Data Analysis Falll 26 Lecture 8: Spectral clustering and Laplacian matrices Lecturer: Jiaming Xu Scribe: Hyun-Ju Oh and Taotao He, October 4, 26 Outline Concentration

More information

Administration. CSCI567 Machine Learning (Fall 2018) Outline. Outline. HW5 is available, due on 11/18. Practice final will also be available soon.

Administration. CSCI567 Machine Learning (Fall 2018) Outline. Outline. HW5 is available, due on 11/18. Practice final will also be available soon. Administration CSCI567 Machine Learning Fall 2018 Prof. Haipeng Luo U of Southern California Nov 7, 2018 HW5 is available, due on 11/18. Practice final will also be available soon. Remaining weeks: 11/14,

More information

Principal Component Analysis

Principal Component Analysis B: Chapter 1 HTF: Chapter 1.5 Principal Component Analysis Barnabás Póczos University of Alberta Nov, 009 Contents Motivation PCA algorithms Applications Face recognition Facial expression recognition

More information

Bandit models: a tutorial

Bandit models: a tutorial Gdt COS, December 3rd, 2015 Multi-Armed Bandit model: general setting K arms: for a {1,..., K}, (X a,t ) t N is a stochastic process. (unknown distributions) Bandit game: a each round t, an agent chooses

More information

The Multi-Armed Bandit Problem

The Multi-Armed Bandit Problem The Multi-Armed Bandit Problem Electrical and Computer Engineering December 7, 2013 Outline 1 2 Mathematical 3 Algorithm Upper Confidence Bound Algorithm A/B Testing Exploration vs. Exploitation Scientist

More information

The geometry of Gaussian processes and Bayesian optimization. Contal CMLA, ENS Cachan

The geometry of Gaussian processes and Bayesian optimization. Contal CMLA, ENS Cachan The geometry of Gaussian processes and Bayesian optimization. Contal CMLA, ENS Cachan Background: Global Optimization and Gaussian Processes The Geometry of Gaussian Processes and the Chaining Trick Algorithm

More information

Lecture 5: Regret Bounds for Thompson Sampling

Lecture 5: Regret Bounds for Thompson Sampling CMSC 858G: Bandits, Experts and Games 09/2/6 Lecture 5: Regret Bounds for Thompson Sampling Instructor: Alex Slivkins Scribed by: Yancy Liao Regret Bounds for Thompson Sampling For each round t, we defined

More information

Evaluation of multi armed bandit algorithms and empirical algorithm

Evaluation of multi armed bandit algorithms and empirical algorithm Acta Technica 62, No. 2B/2017, 639 656 c 2017 Institute of Thermomechanics CAS, v.v.i. Evaluation of multi armed bandit algorithms and empirical algorithm Zhang Hong 2,3, Cao Xiushan 1, Pu Qiumei 1,4 Abstract.

More information

1 MDP Value Iteration Algorithm

1 MDP Value Iteration Algorithm CS 0. - Active Learning Problem Set Handed out: 4 Jan 009 Due: 9 Jan 009 MDP Value Iteration Algorithm. Implement the value iteration algorithm given in the lecture. That is, solve Bellman s equation using

More information

Regret Analysis of Stochastic and Nonstochastic Multi-armed Bandit Problems. Sébastien Bubeck Theory Group

Regret Analysis of Stochastic and Nonstochastic Multi-armed Bandit Problems. Sébastien Bubeck Theory Group Regret Analysis of Stochastic and Nonstochastic Multi-armed Bandit Problems Sébastien Bubeck Theory Group Part 1: i.i.d., adversarial, and Bayesian bandit models i.i.d. multi-armed bandit, Robbins [1952]

More information

Subsampling, Concentration and Multi-armed bandits

Subsampling, Concentration and Multi-armed bandits Subsampling, Concentration and Multi-armed bandits Odalric-Ambrym Maillard, R. Bardenet, S. Mannor, A. Baransi, N. Galichet, J. Pineau, A. Durand Toulouse, November 09, 2015 O-A. Maillard Subsampling and

More information

PCA, Kernel PCA, ICA

PCA, Kernel PCA, ICA PCA, Kernel PCA, ICA Learning Representations. Dimensionality Reduction. Maria-Florina Balcan 04/08/2015 Big & High-Dimensional Data High-Dimensions = Lot of Features Document classification Features per

More information

The information complexity of sequential resource allocation

The information complexity of sequential resource allocation The information complexity of sequential resource allocation Emilie Kaufmann, joint work with Olivier Cappé, Aurélien Garivier and Shivaram Kalyanakrishan SMILE Seminar, ENS, June 8th, 205 Sequential allocation

More information

Basics of reinforcement learning

Basics of reinforcement learning Basics of reinforcement learning Lucian Buşoniu TMLSS, 20 July 2018 Main idea of reinforcement learning (RL) Learn a sequential decision policy to optimize the cumulative performance of an unknown system

More information

Estimation Considerations in Contextual Bandits

Estimation Considerations in Contextual Bandits Estimation Considerations in Contextual Bandits Maria Dimakopoulou Zhengyuan Zhou Susan Athey Guido Imbens arxiv:1711.07077v4 [stat.ml] 16 Dec 2018 Abstract Contextual bandit algorithms are sensitive to

More information

Multi-armed bandit models: a tutorial

Multi-armed bandit models: a tutorial Multi-armed bandit models: a tutorial CERMICS seminar, March 30th, 2016 Multi-Armed Bandit model: general setting K arms: for a {1,..., K}, (X a,t ) t N is a stochastic process. (unknown distributions)

More information

The multi armed-bandit problem

The multi armed-bandit problem The multi armed-bandit problem (with covariates if we have time) Vianney Perchet & Philippe Rigollet LPMA Université Paris Diderot ORFE Princeton University Algorithms and Dynamics for Games and Optimization

More information

Data Analysis and Manifold Learning Lecture 3: Graphs, Graph Matrices, and Graph Embeddings

Data Analysis and Manifold Learning Lecture 3: Graphs, Graph Matrices, and Graph Embeddings Data Analysis and Manifold Learning Lecture 3: Graphs, Graph Matrices, and Graph Embeddings Radu Horaud INRIA Grenoble Rhone-Alpes, France Radu.Horaud@inrialpes.fr http://perception.inrialpes.fr/ Outline

More information

Bandits and Exploration: How do we (optimally) gather information? Sham M. Kakade

Bandits and Exploration: How do we (optimally) gather information? Sham M. Kakade Bandits and Exploration: How do we (optimally) gather information? Sham M. Kakade Machine Learning for Big Data CSE547/STAT548 University of Washington S. M. Kakade (UW) Optimization for Big data 1 / 22

More information

THE first formalization of the multi-armed bandit problem

THE first formalization of the multi-armed bandit problem EDIC RESEARCH PROPOSAL 1 Multi-armed Bandits in a Network Farnood Salehi I&C, EPFL Abstract The multi-armed bandit problem is a sequential decision problem in which we have several options (arms). We can

More information

Exploration. 2015/10/12 John Schulman

Exploration. 2015/10/12 John Schulman Exploration 2015/10/12 John Schulman What is the exploration problem? Given a long-lived agent (or long-running learning algorithm), how to balance exploration and exploitation to maximize long-term rewards

More information

Tractable Learning in Structured Probability Spaces

Tractable Learning in Structured Probability Spaces Tractable Learning in Structured Probability Spaces Guy Van den Broeck DTAI Seminar - KU Leuven Dec 20, 2016 Structured probability spaces? Running Example Courses: Logic (L) Knowledge Representation (K)

More information

Jeffrey D. Ullman Stanford University

Jeffrey D. Ullman Stanford University Jeffrey D. Ullman Stanford University 2 Often, our data can be represented by an m-by-n matrix. And this matrix can be closely approximated by the product of two matrices that share a small common dimension

More information

Pack only the essentials: distributed sequential sampling for adaptive kernel DL

Pack only the essentials: distributed sequential sampling for adaptive kernel DL Pack only the essentials: distributed sequential sampling for adaptive kernel DL with Daniele Calandriello and Alessandro Lazaric SequeL team, Inria Lille - Nord Europe, France appeared in AISTATS 2017

More information

Tractable Learning in Structured Probability Spaces

Tractable Learning in Structured Probability Spaces Tractable Learning in Structured Probability Spaces Guy Van den Broeck UCLA Stats Seminar Jan 17, 2017 Outline 1. Structured probability spaces? 2. Specification language Logic 3. Deep architecture Logic

More information

Confidence intervals for the mixing time of a reversible Markov chain from a single sample path

Confidence intervals for the mixing time of a reversible Markov chain from a single sample path Confidence intervals for the mixing time of a reversible Markov chain from a single sample path Daniel Hsu Aryeh Kontorovich Csaba Szepesvári Columbia University, Ben-Gurion University, University of Alberta

More information

Machine Learning - MT Clustering

Machine Learning - MT Clustering Machine Learning - MT 2016 15. Clustering Varun Kanade University of Oxford November 28, 2016 Announcements No new practical this week All practicals must be signed off in sessions this week Firm Deadline:

More information

Lecture 19: UCB Algorithm and Adversarial Bandit Problem. Announcements Review on stochastic multi-armed bandit problem

Lecture 19: UCB Algorithm and Adversarial Bandit Problem. Announcements Review on stochastic multi-armed bandit problem Lecture 9: UCB Algorithm and Adversarial Bandit Problem EECS598: Prediction and Learning: It s Only a Game Fall 03 Lecture 9: UCB Algorithm and Adversarial Bandit Problem Prof. Jacob Abernethy Scribe:

More information

Bandits for Online Optimization

Bandits for Online Optimization Bandits for Online Optimization Nicolò Cesa-Bianchi Università degli Studi di Milano N. Cesa-Bianchi (UNIMI) Bandits for Online Optimization 1 / 16 The multiarmed bandit problem... K slot machines Each

More information

On the Complexity of Best Arm Identification in Multi-Armed Bandit Models

On the Complexity of Best Arm Identification in Multi-Armed Bandit Models On the Complexity of Best Arm Identification in Multi-Armed Bandit Models Aurélien Garivier Institut de Mathématiques de Toulouse Information Theory, Learning and Big Data Simons Institute, Berkeley, March

More information

Approximate Dynamic Programming

Approximate Dynamic Programming Approximate Dynamic Programming A. LAZARIC (SequeL Team @INRIA-Lille) ENS Cachan - Master 2 MVA SequeL INRIA Lille MVA-RL Course Approximate Dynamic Programming (a.k.a. Batch Reinforcement Learning) A.

More information

Lecture 24: Principal Component Analysis. Aykut Erdem May 2016 Hacettepe University

Lecture 24: Principal Component Analysis. Aykut Erdem May 2016 Hacettepe University Lecture 4: Principal Component Analysis Aykut Erdem May 016 Hacettepe University This week Motivation PCA algorithms Applications PCA shortcomings Autoencoders Kernel PCA PCA Applications Data Visualization

More information

Bandits : optimality in exponential families

Bandits : optimality in exponential families Bandits : optimality in exponential families Odalric-Ambrym Maillard IHES, January 2016 Odalric-Ambrym Maillard Bandits 1 / 40 Introduction 1 Stochastic multi-armed bandits 2 Boundary crossing probabilities

More information

U.C. Berkeley Better-than-Worst-Case Analysis Handout 3 Luca Trevisan May 24, 2018

U.C. Berkeley Better-than-Worst-Case Analysis Handout 3 Luca Trevisan May 24, 2018 U.C. Berkeley Better-than-Worst-Case Analysis Handout 3 Luca Trevisan May 24, 2018 Lecture 3 In which we show how to find a planted clique in a random graph. 1 Finding a Planted Clique We will analyze

More information

1 A Support Vector Machine without Support Vectors

1 A Support Vector Machine without Support Vectors CS/CNS/EE 53 Advanced Topics in Machine Learning Problem Set 1 Handed out: 15 Jan 010 Due: 01 Feb 010 1 A Support Vector Machine without Support Vectors In this question, you ll be implementing an online

More information

L11: Pattern recognition principles

L11: Pattern recognition principles L11: Pattern recognition principles Bayesian decision theory Statistical classifiers Dimensionality reduction Clustering This lecture is partly based on [Huang, Acero and Hon, 2001, ch. 4] Introduction

More information

Combinatorial Cascading Bandits

Combinatorial Cascading Bandits Combinatorial Cascading Bandits Branislav Kveton Adobe Research San Jose, CA kveton@adobe.com Zheng Wen Yahoo Labs Sunnyvale, CA zhengwen@yahoo-inc.com Azin Ashkan Technicolor Research Los Altos, CA azin.ashkan@technicolor.com

More information

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra. DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1

More information

Active Learning and Optimized Information Gathering

Active Learning and Optimized Information Gathering Active Learning and Optimized Information Gathering Lecture 7 Learning Theory CS 101.2 Andreas Krause Announcements Project proposal: Due tomorrow 1/27 Homework 1: Due Thursday 1/29 Any time is ok. Office

More information

Rank and Rating Aggregation. Amy Langville Mathematics Department College of Charleston

Rank and Rating Aggregation. Amy Langville Mathematics Department College of Charleston Rank and Rating Aggregation Amy Langville Mathematics Department College of Charleston Hamilton Institute 8/6/2008 Collaborators Carl Meyer, N. C. State University College of Charleston (current and former

More information

Stochastic bandits: Explore-First and UCB

Stochastic bandits: Explore-First and UCB CSE599s, Spring 2014, Online Learning Lecture 15-2/19/2014 Stochastic bandits: Explore-First and UCB Lecturer: Brendan McMahan or Ofer Dekel Scribe: Javad Hosseini In this lecture, we like to answer this

More information

Alireza Shafaei. Machine Learning Reading Group The University of British Columbia Summer 2017

Alireza Shafaei. Machine Learning Reading Group The University of British Columbia Summer 2017 s s Machine Learning Reading Group The University of British Columbia Summer 2017 (OCO) Convex 1/29 Outline (OCO) Convex Stochastic Bernoulli s (OCO) Convex 2/29 At each iteration t, the player chooses

More information

Grundlagen der Künstlichen Intelligenz

Grundlagen der Künstlichen Intelligenz Grundlagen der Künstlichen Intelligenz Uncertainty & Probabilities & Bandits Daniel Hennes 16.11.2017 (WS 2017/18) University Stuttgart - IPVS - Machine Learning & Robotics 1 Today Uncertainty Probability

More information

Stat 260/CS Learning in Sequential Decision Problems. Peter Bartlett

Stat 260/CS Learning in Sequential Decision Problems. Peter Bartlett Stat 260/CS 294-102. Learning in Sequential Decision Problems. Peter Bartlett 1. Thompson sampling Bernoulli strategy Regret bounds Extensions the flexibility of Bayesian strategies 1 Bayesian bandit strategies

More information

The Multi-Armed Bandit Problem

The Multi-Armed Bandit Problem Università degli Studi di Milano The bandit problem [Robbins, 1952]... K slot machines Rewards X i,1, X i,2,... of machine i are i.i.d. [0, 1]-valued random variables An allocation policy prescribes which

More information

Models of collective inference

Models of collective inference Models of collective inference Laurent Massoulié (Microsoft Research-Inria Joint Centre) Mesrob I. Ohannessian (University of California, San Diego) Alexandre Proutière (KTH Royal Institute of Technology)

More information

Marks. bonus points. } Assignment 1: Should be out this weekend. } Mid-term: Before the last lecture. } Mid-term deferred exam:

Marks. bonus points. } Assignment 1: Should be out this weekend. } Mid-term: Before the last lecture. } Mid-term deferred exam: Marks } Assignment 1: Should be out this weekend } All are marked, I m trying to tally them and perhaps add bonus points } Mid-term: Before the last lecture } Mid-term deferred exam: } This Saturday, 9am-10.30am,

More information

Lecture 4: Lower Bounds (ending); Thompson Sampling

Lecture 4: Lower Bounds (ending); Thompson Sampling CMSC 858G: Bandits, Experts and Games 09/12/16 Lecture 4: Lower Bounds (ending); Thompson Sampling Instructor: Alex Slivkins Scribed by: Guowei Sun,Cheng Jie 1 Lower bounds on regret (ending) Recap from

More information

Contextual Bandits in A Collaborative Environment

Contextual Bandits in A Collaborative Environment Contextual Bandits in A Collaborative Environment Qingyun Wu 1, Huazheng Wang 1, Quanquan Gu 2, Hongning Wang 1 1 Department of Computer Science 2 Department of Systems and Information Engineering University

More information

4 Bias-Variance for Ridge Regression (24 points)

4 Bias-Variance for Ridge Regression (24 points) Implement Ridge Regression with λ = 0.00001. Plot the Squared Euclidean test error for the following values of k (the dimensions you reduce to): k = {0, 50, 100, 150, 200, 250, 300, 350, 400, 450, 500,

More information

An Information-Theoretic Analysis of Thompson Sampling

An Information-Theoretic Analysis of Thompson Sampling Journal of Machine Learning Research (2015) Submitted ; Published An Information-Theoretic Analysis of Thompson Sampling Daniel Russo Department of Management Science and Engineering Stanford University

More information

arxiv: v4 [cs.lg] 22 Jul 2014

arxiv: v4 [cs.lg] 22 Jul 2014 Learning to Optimize Via Information-Directed Sampling Daniel Russo and Benjamin Van Roy July 23, 2014 arxiv:1403.5556v4 cs.lg] 22 Jul 2014 Abstract We propose information-directed sampling a new algorithm

More information

Orthogonal tensor decomposition

Orthogonal tensor decomposition Orthogonal tensor decomposition Daniel Hsu Columbia University Largely based on 2012 arxiv report Tensor decompositions for learning latent variable models, with Anandkumar, Ge, Kakade, and Telgarsky.

More information

Analysis of Thompson Sampling for the multi-armed bandit problem

Analysis of Thompson Sampling for the multi-armed bandit problem Analysis of Thompson Sampling for the multi-armed bandit problem Shipra Agrawal Microsoft Research India shipra@microsoft.com avin Goyal Microsoft Research India navingo@microsoft.com Abstract We show

More information

Reinforcement Learning

Reinforcement Learning Reinforcement Learning Lecture 6: RL algorithms 2.0 Alexandre Proutiere, Sadegh Talebi, Jungseul Ok KTH, The Royal Institute of Technology Objectives of this lecture Present and analyse two online algorithms

More information

Tutorial: PART 2. Online Convex Optimization, A Game- Theoretic Approach to Learning

Tutorial: PART 2. Online Convex Optimization, A Game- Theoretic Approach to Learning Tutorial: PART 2 Online Convex Optimization, A Game- Theoretic Approach to Learning Elad Hazan Princeton University Satyen Kale Yahoo Research Exploiting curvature: logarithmic regret Logarithmic regret

More information

Contextual Combinatorial Bandit and its Application on Diversified Online Recommendation

Contextual Combinatorial Bandit and its Application on Diversified Online Recommendation Contextual Combinatorial Bandit and its Application on Diversified Online Recommendation Lijing Qin Shouyuan Chen Xiaoyan Zhu Abstract Recommender systems are faced with new challenges that are beyond

More information

Supplementary: Battle of Bandits

Supplementary: Battle of Bandits Supplementary: Battle of Bandits A Proof of Lemma Proof. We sta by noting that PX i PX i, i {U, V } Pi {U, V } PX i i {U, V } PX i i {U, V } j,j i j,j i j,j i PX i {U, V } {i, j} Q Q aia a ia j j, j,j

More information

Generalization theory

Generalization theory Generalization theory Daniel Hsu Columbia TRIPODS Bootcamp 1 Motivation 2 Support vector machines X = R d, Y = { 1, +1}. Return solution ŵ R d to following optimization problem: λ min w R d 2 w 2 2 + 1

More information

Support Vector Machine (SVM) and Kernel Methods

Support Vector Machine (SVM) and Kernel Methods Support Vector Machine (SVM) and Kernel Methods CE-717: Machine Learning Sharif University of Technology Fall 2016 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin

More information

Reducing contextual bandits to supervised learning

Reducing contextual bandits to supervised learning Reducing contextual bandits to supervised learning Daniel Hsu Columbia University Based on joint work with A. Agarwal, S. Kale, J. Langford, L. Li, and R. Schapire 1 Learning to interact: example #1 Practicing

More information

Solving linear systems (6 lectures)

Solving linear systems (6 lectures) Chapter 2 Solving linear systems (6 lectures) 2.1 Solving linear systems: LU factorization (1 lectures) Reference: [Trefethen, Bau III] Lecture 20, 21 How do you solve Ax = b? (2.1.1) In numerical linear

More information

1-bit Matrix Completion. PAC-Bayes and Variational Approximation

1-bit Matrix Completion. PAC-Bayes and Variational Approximation : PAC-Bayes and Variational Approximation (with P. Alquier) PhD Supervisor: N. Chopin Bayes In Paris, 5 January 2017 (Happy New Year!) Various Topics covered Matrix Completion PAC-Bayesian Estimation Variational

More information

Introduction to Algorithms / Algorithms I Lecturer: Michael Dinitz Topic: Intro to Learning Theory Date: 12/8/16

Introduction to Algorithms / Algorithms I Lecturer: Michael Dinitz Topic: Intro to Learning Theory Date: 12/8/16 600.463 Introduction to Algorithms / Algorithms I Lecturer: Michael Dinitz Topic: Intro to Learning Theory Date: 12/8/16 25.1 Introduction Today we re going to talk about machine learning, but from an

More information

Foundations of Computer Vision

Foundations of Computer Vision Foundations of Computer Vision Wesley. E. Snyder North Carolina State University Hairong Qi University of Tennessee, Knoxville Last Edited February 8, 2017 1 3.2. A BRIEF REVIEW OF LINEAR ALGEBRA Apply

More information

Data dependent operators for the spatial-spectral fusion problem

Data dependent operators for the spatial-spectral fusion problem Data dependent operators for the spatial-spectral fusion problem Wien, December 3, 2012 Joint work with: University of Maryland: J. J. Benedetto, J. A. Dobrosotskaya, T. Doster, K. W. Duke, M. Ehler, A.

More information

Data Analysis and Manifold Learning Lecture 6: Probabilistic PCA and Factor Analysis

Data Analysis and Manifold Learning Lecture 6: Probabilistic PCA and Factor Analysis Data Analysis and Manifold Learning Lecture 6: Probabilistic PCA and Factor Analysis Radu Horaud INRIA Grenoble Rhone-Alpes, France Radu.Horaud@inrialpes.fr http://perception.inrialpes.fr/ Outline of Lecture

More information

PSDDs for Tractable Learning in Structured and Unstructured Spaces

PSDDs for Tractable Learning in Structured and Unstructured Spaces PSDDs for Tractable Learning in Structured and Unstructured Spaces Guy Van den Broeck DeLBP Aug 18, 2017 References Probabilistic Sentential Decision Diagrams Doga Kisa, Guy Van den Broeck, Arthur Choi

More information

New Algorithms for Contextual Bandits

New Algorithms for Contextual Bandits New Algorithms for Contextual Bandits Lev Reyzin Georgia Institute of Technology Work done at Yahoo! 1 S A. Beygelzimer, J. Langford, L. Li, L. Reyzin, R.E. Schapire Contextual Bandit Algorithms with Supervised

More information

Machine Learning, Fall 2009: Midterm

Machine Learning, Fall 2009: Midterm 10-601 Machine Learning, Fall 009: Midterm Monday, November nd hours 1. Personal info: Name: Andrew account: E-mail address:. You are permitted two pages of notes and a calculator. Please turn off all

More information

Introduction to Machine Learning. Introduction to ML - TAU 2016/7 1

Introduction to Machine Learning. Introduction to ML - TAU 2016/7 1 Introduction to Machine Learning Introduction to ML - TAU 2016/7 1 Course Administration Lecturers: Amir Globerson (gamir@post.tau.ac.il) Yishay Mansour (Mansour@tau.ac.il) Teaching Assistance: Regev Schweiger

More information