Outline. Offline Evaluation of Online Metrics Counterfactual Estimation Advanced Estimators. Case Studies & Demo Summary

Size: px
Start display at page:

Download "Outline. Offline Evaluation of Online Metrics Counterfactual Estimation Advanced Estimators. Case Studies & Demo Summary"

Transcription

1 Outline Offline Evaluation of Online Metrics Counterfactual Estimation Advanced Estimators 1. Self-Normalized Estimator 2. Doubly Robust Estimator 3. Slates Estimator Case Studies & Demo Summary

2 IPS: Issues U ips π = 1 n i π(y i x i ) π 0 (y i x i ) δ i Suppose δ 1 E[Constant] Constant Two variables in IPS δ π(y x) Τπ 0 (y x)

3 Fix: Control Variates Use correlated quantities to control Multiplicative θ E[s] E s i = θ known Τ π(y x) π 0 (y x) variability Additive π(y x) π y x E[ δ] E π 0 (y x) π 0 y x δ s + θ e.g. Self-Normalization e.g. Doubly Robust Estimator

4 Self-Normalized Estimator Use expected sample size as multiplicative control variate sƹ π = 1 n i π(y i x i ) π 0 (y i x i ) E sƹ π = 1 U SNips π = i π(y i x i ) π 0 (y i x i ) δ i i π(y i x i ) π 0 (y i x i ) [Trotter & Tukey, 1952] [Hesterberg, 1983] [Swaminathan & Joachims, 2016]

5 Self-Normalization: Properties E[Constant] = Constant Equivariant Asymptotically consistent Pr lim n U SNips π = U π = 1 Small bias which decays O( 1 n ) while variance decays O( 1 n )

6 News Recommender: Results snips often achieves a better bias-variance trade-off

7 Doubly Robust Estimator Reward Prediction U rp π = 1 E n y~π xi [ መδ x i, y ] i Low variance, High bias U ips π = 1 n i IPS π(y i x i ) π 0 (y i x i ) δ i High variance, No bias U dr π = 1 n i π y i x i π 0 y i x i δ i መδ x i, y i + E y~π xi [ መδ x i, y ] [Langford, Li, Dudik, 2011] [Jiang & Li, 2016]

8 Doubly Robust: Properties Useful when using estimated propensities p i π 0 (y i x i ) Unbiased if, either መδ(x, y) = δ(x, y) Or, p i = π 0 (y i x i ) Default in Vowpal Wabbit

9 News Recommender: Results DR dominates IPS even with a noisy መδ(x, y)

10 Evaluating rankings (slates) y i π(x i ) Exact match of composite actions in logs unlikely Idea: Count per-slot matches

11 Slates Estimator If π 0 samples l documents from a multinomial μ(d x), with replacement U slates π = 1 n i l I y i j = π x i 1 l + μ y i j x i ) j=1 j δ i For general π 0, need to record π 0 (y j = d, y k = d x)

12 Slates Estimator: General π 0 Define: Γ π0 x d, j ; d, k = π 0 y j = d, y k = d x 1 y d, j = I y j = d U slates π = 1 n i E y~π xi 1 T y Γ π0 (x i ) 1 yi δ i Can also develop self-normalized/doubly robust variants [Swaminathan et al, 2016]

13 Slates Estimator: Properties Typically, exponentially better sample complexity than IPS Unbiased if reward decomposes per-slot Φ x s. t. δ x, y = Φ x + Φ x + Φ x + Φ x Can capture higher-order interactions with suitable Γ π0 (x)

14 News Recommender: Results (sn)slates better than IPS but can have asymptotic bias

15 Outline Offline Evaluation of Online Metrics Counterfactual Estimation Advanced Estimators Case Studies & Demo 1. Yahoo Front Page [Li,Chu,Langford,Wang; 2011] 2. Bing Speller [Li,Chen,Kleban,Gupta; 2014] 3. Search Ranking [Swaminathan et al; 2016] Summary

16 Yahoo Front Page y~π(x) Pick STORY F1, F20 to highlight for different users Metric δ: CTR Logging π 0 : Uniform random Setup: Deploy policies, check online IPS correlation

17 Yahoo Front Page: Case Study IPS is quite accurate for several (spatio-temporal) policies

18 Yahoo Front Page: Case Study IPS indeed gives unbiased CTR estimates for different articles

19 Bing Speller Logging π 0 : Pick (possibly many) reformulation candidates Independent Bernoulli per rank Pr Pick d j = Τ α exp β score d j score(d 1 ) Setup: Deploy policies, check online IPS correlation

20 Bing Speller: Case Study IPS reliably estimates CTR, etc. despite non-uniform logging

21 Search Ranking Re-rank 5 out of 8 candidates Metric δ: Logging π 0 : Setup: Time-to-success Utility rate Bootstrap from Uniform/Plackett-Luce Report RMSE vs. bootstrap sample size

22 Plackett-Luce for Slates SoftMax/Multinomial without replacement (renormalize probabilities after each draw)

23 Search Ranking: Case Study Slates RP Slates estimator dominates IPS

24 Outline Offline Evaluation of Online Metrics Counterfactual Estimation Advanced Estimators Case Studies & Demo Summary

25 Evaluation: Key Questions What system π 0 should we deploy? What should we record in our logs? y~π 0 (x) π 0 How can we estimate U(π) using logged data from π 0?

26 What system π 0 to deploy? To enable reliable counterfactual estimation If possible, stochastic with logged randomization If not, log enough to estimate propensities p i Uniform? Around current system? How to explore? typically, bad less risk & better targeted [Also see Online contextual Bandit Algorithms ]

27 What should we record? Log EVERYTHING! x i, y i, δ i, p i To reliably replay π on logged data, Candidate set of actions Features for each candidate Action at the point of randomization Y i f(x i, y) y i [Bottou et al, 2013] [Li,Chen,Kleban,Gupta, 2015]

28 How can we evaluate U(π)? Offline Evaluation of Online Metrics Related: Test collections for offline metrics Model the world Reward Prediction Related: Click models; Collaborative filtering Model the bias in data Off-policy estimator Randomization is essential [Carterette et al, 2010] [Aslam et al, 2009] [Schnabel et al, 2016b] [Chuklin et al, 2015] [Schnabel et al, 2016a]

29 Summary: Off-policy Estimators IPS Estimator Simple, effective fix for non-uniform (biased) data Self-Normalized Estimator Doubly Robust Estimator Slates Estimator

30 Demo: Code Samples Visit Download (Recommend) Code_Data.zip Install Anaconda-Python3; joblib Run experiment: python EvalExperiment.py Plot results: python plot_sigir.py --mode [estimate/error] --path../../logs/ssynth

31 QUESTIONS?

Counterfactual Evaluation and Learning

Counterfactual Evaluation and Learning SIGIR 26 Tutorial Counterfactual Evaluation and Learning Adith Swaminathan, Thorsten Joachims Department of Computer Science & Department of Information Science Cornell University Website: http://www.cs.cornell.edu/~adith/cfactsigir26/

More information

Counterfactual Model for Learning Systems

Counterfactual Model for Learning Systems Counterfactual Model for Learning Systems CS 7792 - Fall 28 Thorsten Joachims Department of Computer Science & Department of Information Science Cornell University Imbens, Rubin, Causal Inference for Statistical

More information

Offline Evaluation of Ranking Policies with Click Models

Offline Evaluation of Ranking Policies with Click Models Offline Evaluation of Ranking Policies with Click Models Shuai Li The Chinese University of Hong Kong Joint work with Yasin Abbasi-Yadkori (Adobe Research) Branislav Kveton (Google Research, was in Adobe

More information

Learning with Exploration

Learning with Exploration Learning with Exploration John Langford (Yahoo!) { With help from many } Austin, March 24, 2011 Yahoo! wants to interactively choose content and use the observed feedback to improve future content choices.

More information

Doubly Robust Policy Evaluation and Learning

Doubly Robust Policy Evaluation and Learning Doubly Robust Policy Evaluation and Learning Miroslav Dudik, John Langford and Lihong Li Yahoo! Research Discussed by Miao Liu October 9, 2011 October 9, 2011 1 / 17 1 Introduction 2 Problem Definition

More information

New Algorithms for Contextual Bandits

New Algorithms for Contextual Bandits New Algorithms for Contextual Bandits Lev Reyzin Georgia Institute of Technology Work done at Yahoo! 1 S A. Beygelzimer, J. Langford, L. Li, L. Reyzin, R.E. Schapire Contextual Bandit Algorithms with Supervised

More information

Recommendations as Treatments: Debiasing Learning and Evaluation

Recommendations as Treatments: Debiasing Learning and Evaluation ICML 2016, NYC Tobias Schnabel, Adith Swaminathan, Ashudeep Singh, Navin Chandak, Thorsten Joachims Cornell University, Google Funded in part through NSF Awards IIS-1247637, IIS-1217686, IIS-1513692. Romance

More information

Bandit Algorithms. Zhifeng Wang ... Department of Statistics Florida State University

Bandit Algorithms. Zhifeng Wang ... Department of Statistics Florida State University Bandit Algorithms Zhifeng Wang Department of Statistics Florida State University Outline Multi-Armed Bandits (MAB) Exploration-First Epsilon-Greedy Softmax UCB Thompson Sampling Adversarial Bandits Exp3

More information

Large-scale Information Processing, Summer Recommender Systems (part 2)

Large-scale Information Processing, Summer Recommender Systems (part 2) Large-scale Information Processing, Summer 2015 5 th Exercise Recommender Systems (part 2) Emmanouil Tzouridis tzouridis@kma.informatik.tu-darmstadt.de Knowledge Mining & Assessment SVM question When a

More information

Slides at:

Slides at: Learning to Interact John Langford @ Microsoft Research (with help from many) Slides at: http://hunch.net/~jl/interact.pdf For demo: Raw RCV1 CCAT-or-not: http://hunch.net/~jl/vw_raw.tar.gz Simple converter:

More information

arxiv: v2 [cs.lg] 26 Jun 2017

arxiv: v2 [cs.lg] 26 Jun 2017 Effective Evaluation Using Logged Bandit Feedback from Multiple Loggers Aman Agarwal, Soumya Basu, Tobias Schnabel, Thorsten Joachims Cornell University, Dept. of Computer Science Ithaca, NY, USA [aa2398,sb2352,tbs49,tj36]@cornell.edu

More information

Information Retrieval

Information Retrieval Information Retrieval Online Evaluation Ilya Markov i.markov@uva.nl University of Amsterdam Ilya Markov i.markov@uva.nl Information Retrieval 1 Course overview Offline Data Acquisition Data Processing

More information

Reducing contextual bandits to supervised learning

Reducing contextual bandits to supervised learning Reducing contextual bandits to supervised learning Daniel Hsu Columbia University Based on joint work with A. Agarwal, S. Kale, J. Langford, L. Li, and R. Schapire 1 Learning to interact: example #1 Practicing

More information

1 [15 points] Frequent Itemsets Generation With Map-Reduce

1 [15 points] Frequent Itemsets Generation With Map-Reduce Data Mining Learning from Large Data Sets Final Exam Date: 15 August 2013 Time limit: 120 minutes Number of pages: 11 Maximum score: 100 points You can use the back of the pages if you run out of space.

More information

Importance Sampling for Fair Policy Selection

Importance Sampling for Fair Policy Selection Importance Sampling for Fair Policy Selection Shayan Doroudi Computer Science Department Carnegie Mellon University Pittsburgh, PA 15213 shayand@cs.cmu.edu Philip S. Thomas Computer Science Department

More information

2.3 Estimating PDFs and PDF Parameters

2.3 Estimating PDFs and PDF Parameters .3 Estimating PDFs and PDF Parameters estimating means - discrete and continuous estimating variance using a known mean estimating variance with an estimated mean estimating a discrete pdf estimating a

More information

arxiv: v2 [cs.lg] 27 May 2016

arxiv: v2 [cs.lg] 27 May 2016 Tobias Schnabel, Adith Swaminathan, Ashudeep Singh, Navin Chandak, Thorsten Joachims Cornell University, Ithaca, NY, USA {TBS49, FA234, AS3354, NC475, TJ36}@CORNELL.EDU arxiv:602.05352v2 [cs.lg] 27 May

More information

Discrete Latent Variable Models

Discrete Latent Variable Models Discrete Latent Variable Models Stefano Ermon, Aditya Grover Stanford University Lecture 14 Stefano Ermon, Aditya Grover (AI Lab) Deep Generative Models Lecture 14 1 / 29 Summary Major themes in the course

More information

Bayesian Contextual Multi-armed Bandits

Bayesian Contextual Multi-armed Bandits Bayesian Contextual Multi-armed Bandits Xiaoting Zhao Joint Work with Peter I. Frazier School of Operations Research and Information Engineering Cornell University October 22, 2012 1 / 33 Outline 1 Motivating

More information

arxiv: v1 [stat.ml] 12 Feb 2018

arxiv: v1 [stat.ml] 12 Feb 2018 Practical Evaluation and Optimization of Contextual Bandit Algorithms arxiv:1802.04064v1 [stat.ml] 12 Feb 2018 Alberto Bietti Inria, MSR-Inria Joint Center alberto.bietti@inria.fr Alekh Agarwal Microsoft

More information

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables THE UNIVERSITY OF MANCHESTER. 21 June :45 11:45

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables THE UNIVERSITY OF MANCHESTER. 21 June :45 11:45 Two hours MATH20802 To be supplied by the Examinations Office: Mathematical Formula Tables THE UNIVERSITY OF MANCHESTER STATISTICAL METHODS 21 June 2010 9:45 11:45 Answer any FOUR of the questions. University-approved

More information

Automated Tuning of Ad Auctions

Automated Tuning of Ad Auctions Automated Tuning of Ad Auctions Dilan Gorur, Debabrata Sengupta, Levi Boyles, Patrick Jordan, Eren Manavoglu, Elon Portugaly, Meng Wei, Yaojia Zhu Microsoft {dgorur, desengup, leboyles, pajordan, ermana,

More information

Counterfactual Evaluation and Learning from Logged User Feedback

Counterfactual Evaluation and Learning from Logged User Feedback Counterfactual Evaluation and Learning from Logged User Feedback Adith Swaminathan adith@cs.cornell.edu Department of Computer Science Cornell University Committee: Thorsten Joachims (chair), Johannes

More information

Learning from Rational * Behavior

Learning from Rational * Behavior Learning from Rational * Behavior Josef Broder, Olivier Chapelle, Geri Gay, Arpita Ghosh, Laura Granka, Thorsten Joachims, Bobby Kleinberg, Madhu Kurup, Filip Radlinski, Karthik Raman, Tobias Schnabel,

More information

Learning from Logged Implicit Exploration Data

Learning from Logged Implicit Exploration Data Learning from Logged Implicit Exploration Data Alexander L. Strehl Facebook Inc. 1601 S California Ave Palo Alto, CA 94304 astrehl@facebook.com Lihong Li Yahoo! Research 4401 Great America Parkway Santa

More information

Practical Agnostic Active Learning

Practical Agnostic Active Learning Practical Agnostic Active Learning Alina Beygelzimer Yahoo Research based on joint work with Sanjoy Dasgupta, Daniel Hsu, John Langford, Francesco Orabona, Chicheng Zhang, and Tong Zhang * * introductory

More information

Administration. CSCI567 Machine Learning (Fall 2018) Outline. Outline. HW5 is available, due on 11/18. Practice final will also be available soon.

Administration. CSCI567 Machine Learning (Fall 2018) Outline. Outline. HW5 is available, due on 11/18. Practice final will also be available soon. Administration CSCI567 Machine Learning Fall 2018 Prof. Haipeng Luo U of Southern California Nov 7, 2018 HW5 is available, due on 11/18. Practice final will also be available soon. Remaining weeks: 11/14,

More information

Lecture 3: Policy Evaluation Without Knowing How the World Works / Model Free Policy Evaluation

Lecture 3: Policy Evaluation Without Knowing How the World Works / Model Free Policy Evaluation Lecture 3: Policy Evaluation Without Knowing How the World Works / Model Free Policy Evaluation CS234: RL Emma Brunskill Winter 2018 Material builds on structure from David SIlver s Lecture 4: Model-Free

More information

Counterfactual Learning-to-Rank for Additive Metrics and Deep Models

Counterfactual Learning-to-Rank for Additive Metrics and Deep Models Counterfactual Learning-to-Rank for Additive Metrics and Deep Models Aman Agarwal Cornell University Ithaca, NY aman@cs.cornell.edu Ivan Zaitsev Cornell University Ithaca, NY iz44@cornell.edu Thorsten

More information

COUNTERFACTUAL REASONING AND LEARNING SYSTEMS LÉON BOTTOU

COUNTERFACTUAL REASONING AND LEARNING SYSTEMS LÉON BOTTOU COUNTERFACTUAL REASONING AND LEARNING SYSTEMS LÉON BOTTOU MSR Joint work with Jonas Peters Joaquin Quiñonero Candela Denis Xavier Charles D. Max Chickering Elon Portugaly Dipankar Ray Patrice Simard Ed

More information

Doubly Robust Policy Evaluation and Learning

Doubly Robust Policy Evaluation and Learning Miroslav Dudík John Langford Yahoo! Research, New York, NY, USA 10018 Lihong Li Yahoo! Research, Santa Clara, CA, USA 95054 MDUDIK@YAHOO-INC.COM JL@YAHOO-INC.COM LIHONG@YAHOO-INC.COM Abstract We study

More information

Collaborative topic models: motivations cont

Collaborative topic models: motivations cont Collaborative topic models: motivations cont Two topics: machine learning social network analysis Two people: " boy Two articles: article A! girl article B Preferences: The boy likes A and B --- no problem.

More information

Weighting Methods. Harvard University STAT186/GOV2002 CAUSAL INFERENCE. Fall Kosuke Imai

Weighting Methods. Harvard University STAT186/GOV2002 CAUSAL INFERENCE. Fall Kosuke Imai Weighting Methods Kosuke Imai Harvard University STAT186/GOV2002 CAUSAL INFERENCE Fall 2018 Kosuke Imai (Harvard) Weighting Methods Stat186/Gov2002 Fall 2018 1 / 13 Motivation Matching methods for improving

More information

Modern Methods of Data Analysis - WS 07/08

Modern Methods of Data Analysis - WS 07/08 Modern Methods of Data Analysis Lecture VIc (19.11.07) Contents: Maximum Likelihood Fit Maximum Likelihood (I) Assume N measurements of a random variable Assume them to be independent and distributed according

More information

The Self-Normalized Estimator for Counterfactual Learning

The Self-Normalized Estimator for Counterfactual Learning The Self-Normalized Estimator for Counterfactual Learning Adith Swaminathan Department of Computer Science Cornell University adith@cs.cornell.edu Thorsten Joachims Department of Computer Science Cornell

More information

Exploration Scavenging

Exploration Scavenging John Langford jl@yahoo-inc.com Alexander Strehl strehl@yahoo-inc.com Yahoo! Research, 111 W. 40th Street, New York, New York 10018 Jennifer Wortman wortmanj@seas.upenn.edu Department of Computer and Information

More information

Econometric Analysis of Cross Section and Panel Data

Econometric Analysis of Cross Section and Panel Data Econometric Analysis of Cross Section and Panel Data Jeffrey M. Wooldridge / The MIT Press Cambridge, Massachusetts London, England Contents Preface Acknowledgments xvii xxiii I INTRODUCTION AND BACKGROUND

More information

Off-policy Evaluation and Learning

Off-policy Evaluation and Learning COMS E6998.001: Bandits and Reinforcement Learning Fall 2017 Off-policy Evaluation and Learning Lecturer: Alekh Agarwal Lecture 7, October 11 Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer:

More information

Estimation Considerations in Contextual Bandits

Estimation Considerations in Contextual Bandits Estimation Considerations in Contextual Bandits Maria Dimakopoulou Zhengyuan Zhou Susan Athey Guido Imbens arxiv:1711.07077v4 [stat.ml] 16 Dec 2018 Abstract Contextual bandit algorithms are sensitive to

More information

Learning for Contextual Bandits

Learning for Contextual Bandits Learning for Contextual Bandits Alina Beygelzimer 1 John Langford 2 IBM Research 1 Yahoo! Research 2 NYC ML Meetup, Sept 21, 2010 Example of Learning through Exploration Repeatedly: 1. A user comes to

More information

EXAMINATIONS OF THE HONG KONG STATISTICAL SOCIETY

EXAMINATIONS OF THE HONG KONG STATISTICAL SOCIETY EXAMINATIONS OF THE HONG KONG STATISTICAL SOCIETY HIGHER CERTIFICATE IN STATISTICS, 2013 MODULE 5 : Further probability and inference Time allowed: One and a half hours Candidates should answer THREE questions.

More information

Bounded Off-Policy Evaluation with Missing Data for Course Recommendation and Curriculum Design

Bounded Off-Policy Evaluation with Missing Data for Course Recommendation and Curriculum Design Bounded Off-Policy Evaluation with Missing Data for Course Recommendation and Curriculum Design William Hoiles University of California, Los Angeles Mihaela van der Schaar University of California, Los

More information

arxiv: v2 [cs.lg] 13 Jun 2018

arxiv: v2 [cs.lg] 13 Jun 2018 Offline Evaluation of Ranking Policies with Click odels Shuai Li The Chinese University of Hong Kong shuaili@cse.cuhk.edu.hk S. uthukrishnan Rutgers University muthu@cs.rutgers.edu Yasin Abbasi-Yadkori

More information

Reinforcement Learning

Reinforcement Learning Reinforcement Learning Policy gradients Daniel Hennes 26.06.2017 University Stuttgart - IPVS - Machine Learning & Robotics 1 Policy based reinforcement learning So far we approximated the action value

More information

Lecture 7 Introduction to Statistical Decision Theory

Lecture 7 Introduction to Statistical Decision Theory Lecture 7 Introduction to Statistical Decision Theory I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 20, 2016 1 / 55 I-Hsiang Wang IT Lecture 7

More information

Learning Tetris. 1 Tetris. February 3, 2009

Learning Tetris. 1 Tetris. February 3, 2009 Learning Tetris Matt Zucker Andrew Maas February 3, 2009 1 Tetris The Tetris game has been used as a benchmark for Machine Learning tasks because its large state space (over 2 200 cell configurations are

More information

An Efficient Estimation Method for Longitudinal Surveys with Monotone Missing Data

An Efficient Estimation Method for Longitudinal Surveys with Monotone Missing Data An Efficient Estimation Method for Longitudinal Surveys with Monotone Missing Data Jae-Kwang Kim 1 Iowa State University June 28, 2012 1 Joint work with Dr. Ming Zhou (when he was a PhD student at ISU)

More information

Reinforcement Learning. Spring 2018 Defining MDPs, Planning

Reinforcement Learning. Spring 2018 Defining MDPs, Planning Reinforcement Learning Spring 2018 Defining MDPs, Planning understandability 0 Slide 10 time You are here Markov Process Where you will go depends only on where you are Markov Process: Information state

More information

Deep-Treat: Learning Optimal Personalized Treatments from Observational Data using Neural Networks

Deep-Treat: Learning Optimal Personalized Treatments from Observational Data using Neural Networks Deep-Treat: Learning Optimal Personalized Treatments from Observational Data using Neural Networks Onur Atan 1, James Jordon 2, Mihaela van der Schaar 2,3,1 1 Electrical Engineering Department, University

More information

Data Integration for Big Data Analysis for finite population inference

Data Integration for Big Data Analysis for finite population inference for Big Data Analysis for finite population inference Jae-kwang Kim ISU January 23, 2018 1 / 36 What is big data? 2 / 36 Data do not speak for themselves Knowledge Reproducibility Information Intepretation

More information

Doubly Robust Policy Evaluation and Optimization 1

Doubly Robust Policy Evaluation and Optimization 1 Statistical Science 2014, Vol. 29, No. 4, 485 511 DOI: 10.1214/14-STS500 c Institute of Mathematical Statistics, 2014 Doubly Robust Policy Evaluation and Optimization 1 Miroslav Dudík, Dumitru Erhan, John

More information

Doubly Robust Policy Evaluation and Optimization 1

Doubly Robust Policy Evaluation and Optimization 1 Statistical Science 2014, Vol. 29, No. 4, 485 511 DOI: 10.1214/14-STS500 Institute of Mathematical Statistics, 2014 Doubly Robust Policy valuation and Optimization 1 Miroslav Dudík, Dumitru rhan, John

More information

Learning to Bid Without Knowing your Value. Chara Podimata Harvard University June 4, 2018

Learning to Bid Without Knowing your Value. Chara Podimata Harvard University June 4, 2018 Learning to Bid Without Knowing your Value Zhe Feng Harvard University zhe feng@g.harvard.edu Chara Podimata Harvard University podimata@g.harvard.edu June 4, 208 Vasilis Syrgkanis Microsoft Research vasy@microsoft.com

More information

Crowd-Learning: Improving the Quality of Crowdsourcing Using Sequential Learning

Crowd-Learning: Improving the Quality of Crowdsourcing Using Sequential Learning Crowd-Learning: Improving the Quality of Crowdsourcing Using Sequential Learning Mingyan Liu (Joint work with Yang Liu) Department of Electrical Engineering and Computer Science University of Michigan,

More information

Recommendation Systems

Recommendation Systems Recommendation Systems Pawan Goyal CSE, IITKGP October 21, 2014 Pawan Goyal (IIT Kharagpur) Recommendation Systems October 21, 2014 1 / 52 Recommendation System? Pawan Goyal (IIT Kharagpur) Recommendation

More information

Advanced Machine Learning

Advanced Machine Learning Advanced Machine Learning Bandit Problems MEHRYAR MOHRI MOHRI@ COURANT INSTITUTE & GOOGLE RESEARCH. Multi-Armed Bandit Problem Problem: which arm of a K-slot machine should a gambler pull to maximize his

More information

Hybrid Machine Learning Algorithms

Hybrid Machine Learning Algorithms Hybrid Machine Learning Algorithms Umar Syed Princeton University Includes joint work with: Rob Schapire (Princeton) Nina Mishra, Alex Slivkins (Microsoft) Common Approaches to Machine Learning!! Supervised

More information

DEEP LEARNING WITH LOGGED BANDIT FEEDBACK

DEEP LEARNING WITH LOGGED BANDIT FEEDBACK DP LARNING WITH LOGGD BANDIT FDBACK Thorsten Joachims Cornell University tj@cs.cornell.edu Adith Saminathan Microsoft Research adsamin@microsoft.com Maarten de Rijke University of Amsterdam derijke@uva.nl

More information

Lecture 4 January 23

Lecture 4 January 23 STAT 263/363: Experimental Design Winter 2016/17 Lecture 4 January 23 Lecturer: Art B. Owen Scribe: Zachary del Rosario 4.1 Bandits Bandits are a form of online (adaptive) experiments; i.e. samples are

More information

Learning to play K-armed bandit problems

Learning to play K-armed bandit problems Learning to play K-armed bandit problems Francis Maes 1, Louis Wehenkel 1 and Damien Ernst 1 1 University of Liège Dept. of Electrical Engineering and Computer Science Institut Montefiore, B28, B-4000,

More information

Optimal and Adaptive Online Learning

Optimal and Adaptive Online Learning Optimal and Adaptive Online Learning Haipeng Luo Advisor: Robert Schapire Computer Science Department Princeton University Examples of Online Learning (a) Spam detection 2 / 34 Examples of Online Learning

More information

Double Robustness. Bang and Robins (2005) Kang and Schafer (2007)

Double Robustness. Bang and Robins (2005) Kang and Schafer (2007) Double Robustness Bang and Robins (2005) Kang and Schafer (2007) Set-Up Assume throughout that treatment assignment is ignorable given covariates (similar to assumption that data are missing at random

More information

STAT 830 Non-parametric Inference Basics

STAT 830 Non-parametric Inference Basics STAT 830 Non-parametric Inference Basics Richard Lockhart Simon Fraser University STAT 801=830 Fall 2012 Richard Lockhart (Simon Fraser University)STAT 830 Non-parametric Inference Basics STAT 801=830

More information

Restricted Boltzmann Machines for Collaborative Filtering

Restricted Boltzmann Machines for Collaborative Filtering Restricted Boltzmann Machines for Collaborative Filtering Authors: Ruslan Salakhutdinov Andriy Mnih Geoffrey Hinton Benjamin Schwehn Presentation by: Ioan Stanculescu 1 Overview The Netflix prize problem

More information

Terminology Suppose we have N observations {x(n)} N 1. Estimators as Random Variables. {x(n)} N 1

Terminology Suppose we have N observations {x(n)} N 1. Estimators as Random Variables. {x(n)} N 1 Estimation Theory Overview Properties Bias, Variance, and Mean Square Error Cramér-Rao lower bound Maximum likelihood Consistency Confidence intervals Properties of the mean estimator Properties of the

More information

Temporal Difference. Learning KENNETH TRAN. Principal Research Engineer, MSR AI

Temporal Difference. Learning KENNETH TRAN. Principal Research Engineer, MSR AI Temporal Difference Learning KENNETH TRAN Principal Research Engineer, MSR AI Temporal Difference Learning Policy Evaluation Intro to model-free learning Monte Carlo Learning Temporal Difference Learning

More information

Estimation of Conditional Kendall s Tau for Bivariate Interval Censored Data

Estimation of Conditional Kendall s Tau for Bivariate Interval Censored Data Communications for Statistical Applications and Methods 2015, Vol. 22, No. 6, 599 604 DOI: http://dx.doi.org/10.5351/csam.2015.22.6.599 Print ISSN 2287-7843 / Online ISSN 2383-4757 Estimation of Conditional

More information

Collaborative Filtering. Radek Pelánek

Collaborative Filtering. Radek Pelánek Collaborative Filtering Radek Pelánek 2017 Notes on Lecture the most technical lecture of the course includes some scary looking math, but typically with intuitive interpretation use of standard machine

More information

Robust Bayesian Variable Selection for Modeling Mean Medical Costs

Robust Bayesian Variable Selection for Modeling Mean Medical Costs Robust Bayesian Variable Selection for Modeling Mean Medical Costs Grace Yoon 1,, Wenxin Jiang 2, Lei Liu 3 and Ya-Chen T. Shih 4 1 Department of Statistics, Texas A&M University 2 Department of Statistics,

More information

Mixed Membership Matrix Factorization

Mixed Membership Matrix Factorization Mixed Membership Matrix Factorization Lester Mackey 1 David Weiss 2 Michael I. Jordan 1 1 University of California, Berkeley 2 University of Pennsylvania International Conference on Machine Learning, 2010

More information

Online Learning and Sequential Decision Making

Online Learning and Sequential Decision Making Online Learning and Sequential Decision Making Emilie Kaufmann CNRS & CRIStAL, Inria SequeL, emilie.kaufmann@univ-lille.fr Research School, ENS Lyon, Novembre 12-13th 2018 Emilie Kaufmann Sequential Decision

More information

Section 8.1: Interval Estimation

Section 8.1: Interval Estimation Section 8.1: Interval Estimation Discrete-Event Simulation: A First Course c 2006 Pearson Ed., Inc. 0-13-142917-5 Discrete-Event Simulation: A First Course Section 8.1: Interval Estimation 1/ 35 Section

More information

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review STATS 200: Introduction to Statistical Inference Lecture 29: Course review Course review We started in Lecture 1 with a fundamental assumption: Data is a realization of a random process. The goal throughout

More information

Introduction to Bayesian Statistics with WinBUGS Part 4 Priors and Hierarchical Models

Introduction to Bayesian Statistics with WinBUGS Part 4 Priors and Hierarchical Models Introduction to Bayesian Statistics with WinBUGS Part 4 Priors and Hierarchical Models Matthew S. Johnson New York ASA Chapter Workshop CUNY Graduate Center New York, NY hspace1in December 17, 2009 December

More information

CS 188: Artificial Intelligence. Outline

CS 188: Artificial Intelligence. Outline CS 188: Artificial Intelligence Lecture 21: Perceptrons Pieter Abbeel UC Berkeley Many slides adapted from Dan Klein. Outline Generative vs. Discriminative Binary Linear Classifiers Perceptron Multi-class

More information

MA 575 Linear Models: Cedric E. Ginestet, Boston University Non-parametric Inference, Polynomial Regression Week 9, Lecture 2

MA 575 Linear Models: Cedric E. Ginestet, Boston University Non-parametric Inference, Polynomial Regression Week 9, Lecture 2 MA 575 Linear Models: Cedric E. Ginestet, Boston University Non-parametric Inference, Polynomial Regression Week 9, Lecture 2 1 Bootstrapped Bias and CIs Given a multiple regression model with mean and

More information

Chapter 8. Some Approximations to Probability Distributions: Limit Theorems

Chapter 8. Some Approximations to Probability Distributions: Limit Theorems Chapter 8. Some Approximations to Probability Distributions: Limit Theorems Sections 8.2 -- 8.3: Convergence in Probability and in Distribution Jiaping Wang Department of Mathematical Science 04/22/2013,

More information

Algebra of Random Variables: Optimal Average and Optimal Scaling Minimising

Algebra of Random Variables: Optimal Average and Optimal Scaling Minimising Review: Optimal Average/Scaling is equivalent to Minimise χ Two 1-parameter models: Estimating < > : Scaling a pattern: Two equivalent methods: Algebra of Random Variables: Optimal Average and Optimal

More information

Semi-Parametric Importance Sampling for Rare-event probability Estimation

Semi-Parametric Importance Sampling for Rare-event probability Estimation Semi-Parametric Importance Sampling for Rare-event probability Estimation Z. I. Botev and P. L Ecuyer IMACS Seminar 2011 Borovets, Bulgaria Semi-Parametric Importance Sampling for Rare-event probability

More information

Trust Region Policy Optimization

Trust Region Policy Optimization Trust Region Policy Optimization Yixin Lin Duke University yixin.lin@duke.edu March 28, 2017 Yixin Lin (Duke) TRPO March 28, 2017 1 / 21 Overview 1 Preliminaries Markov Decision Processes Policy iteration

More information

Basics of reinforcement learning

Basics of reinforcement learning Basics of reinforcement learning Lucian Buşoniu TMLSS, 20 July 2018 Main idea of reinforcement learning (RL) Learn a sequential decision policy to optimize the cumulative performance of an unknown system

More information

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER.

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER. Two hours MATH38181 To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER EXTREME VALUES AND FINANCIAL RISK Examiner: Answer any FOUR

More information

arxiv: v1 [cs.ai] 1 Jul 2015

arxiv: v1 [cs.ai] 1 Jul 2015 arxiv:507.00353v [cs.ai] Jul 205 Harm van Seijen harm.vanseijen@ualberta.ca A. Rupam Mahmood ashique@ualberta.ca Patrick M. Pilarski patrick.pilarski@ualberta.ca Richard S. Sutton sutton@cs.ualberta.ca

More information

Mixed Membership Matrix Factorization

Mixed Membership Matrix Factorization Mixed Membership Matrix Factorization Lester Mackey University of California, Berkeley Collaborators: David Weiss, University of Pennsylvania Michael I. Jordan, University of California, Berkeley 2011

More information

arxiv: v2 [math.oc] 29 Aug 2015

arxiv: v2 [math.oc] 29 Aug 2015 Parameter estimation in softmax decision-making models with linear objective functions Paul Reverdy and Naomi E. Leonard 2 arxiv:52.4635v2 [math.oc] 29 Aug 25 Abstract With an eye towards human-centered

More information

Stat 260/CS Learning in Sequential Decision Problems. Peter Bartlett

Stat 260/CS Learning in Sequential Decision Problems. Peter Bartlett Stat 260/CS 294-102. Learning in Sequential Decision Problems. Peter Bartlett 1. Thompson sampling Bernoulli strategy Regret bounds Extensions the flexibility of Bayesian strategies 1 Bayesian bandit strategies

More information

Bandit models: a tutorial

Bandit models: a tutorial Gdt COS, December 3rd, 2015 Multi-Armed Bandit model: general setting K arms: for a {1,..., K}, (X a,t ) t N is a stochastic process. (unknown distributions) Bandit game: a each round t, an agent chooses

More information

Probabilistic Graphical Models for Image Analysis - Lecture 4

Probabilistic Graphical Models for Image Analysis - Lecture 4 Probabilistic Graphical Models for Image Analysis - Lecture 4 Stefan Bauer 12 October 2018 Max Planck ETH Center for Learning Systems Overview 1. Repetition 2. α-divergence 3. Variational Inference 4.

More information

LECTURE 5 NOTES. n t. t Γ(a)Γ(b) pt+a 1 (1 p) n t+b 1. The marginal density of t is. Γ(t + a)γ(n t + b) Γ(n + a + b)

LECTURE 5 NOTES. n t. t Γ(a)Γ(b) pt+a 1 (1 p) n t+b 1. The marginal density of t is. Γ(t + a)γ(n t + b) Γ(n + a + b) LECTURE 5 NOTES 1. Bayesian point estimators. In the conventional (frequentist) approach to statistical inference, the parameter θ Θ is considered a fixed quantity. In the Bayesian approach, it is considered

More information

Recall the Basics of Hypothesis Testing

Recall the Basics of Hypothesis Testing Recall the Basics of Hypothesis Testing The level of significance α, (size of test) is defined as the probability of X falling in w (rejecting H 0 ) when H 0 is true: P(X w H 0 ) = α. H 0 TRUE H 1 TRUE

More information

Stochastic Gradient Descent. CS 584: Big Data Analytics

Stochastic Gradient Descent. CS 584: Big Data Analytics Stochastic Gradient Descent CS 584: Big Data Analytics Gradient Descent Recap Simplest and extremely popular Main Idea: take a step proportional to the negative of the gradient Easy to implement Each iteration

More information

Introduction to Restricted Boltzmann Machines

Introduction to Restricted Boltzmann Machines Introduction to Restricted Boltzmann Machines Ilija Bogunovic and Edo Collins EPFL {ilija.bogunovic,edo.collins}@epfl.ch October 13, 2014 Introduction Ingredients: 1. Probabilistic graphical models (undirected,

More information

Scalable Hierarchical Recommendations Using Spatial Autocorrelation

Scalable Hierarchical Recommendations Using Spatial Autocorrelation Scalable Hierarchical Recommendations Using Spatial Autocorrelation Ayushi Dalmia, Joydeep Das, Prosenjit Gupta, Subhashis Majumder, Debarshi Dutta Ayushi Dalmia, JoydeepScalable Das, Prosenjit Hierarchical

More information

Max. Likelihood Estimation. Outline. Econometrics II. Ricardo Mora. Notes. Notes

Max. Likelihood Estimation. Outline. Econometrics II. Ricardo Mora. Notes. Notes Maximum Likelihood Estimation Econometrics II Department of Economics Universidad Carlos III de Madrid Máster Universitario en Desarrollo y Crecimiento Económico Outline 1 3 4 General Approaches to Parameter

More information

The information complexity of sequential resource allocation

The information complexity of sequential resource allocation The information complexity of sequential resource allocation Emilie Kaufmann, joint work with Olivier Cappé, Aurélien Garivier and Shivaram Kalyanakrishan SMILE Seminar, ENS, June 8th, 205 Sequential allocation

More information

Computer Science, Informatik 4 Communication and Distributed Systems. Simulation. Discrete-Event System Simulation. Dr.

Computer Science, Informatik 4 Communication and Distributed Systems. Simulation. Discrete-Event System Simulation. Dr. Simulation Discrete-Event System Simulation Chapter 0 Output Analysis for a Single Model Purpose Objective: Estimate system performance via simulation If θ is the system performance, the precision of the

More information

Some Mathematical Models to Turn Social Media into Knowledge

Some Mathematical Models to Turn Social Media into Knowledge Some Mathematical Models to Turn Social Media into Knowledge Xiaojin Zhu University of Wisconsin Madison, USA NLP&CC 2013 Collaborators Amy Bellmore Aniruddha Bhargava Kwang-Sung Jun Robert Nowak Jun-Ming

More information

Reinforcement Learning and NLP

Reinforcement Learning and NLP 1 Reinforcement Learning and NLP Kapil Thadani kapil@cs.columbia.edu RESEARCH Outline 2 Model-free RL Markov decision processes (MDPs) Derivative-free optimization Policy gradients Variance reduction Value

More information

Lecture 9: Policy Gradient II (Post lecture) 2

Lecture 9: Policy Gradient II (Post lecture) 2 Lecture 9: Policy Gradient II (Post lecture) 2 Emma Brunskill CS234 Reinforcement Learning. Winter 2018 Additional reading: Sutton and Barto 2018 Chp. 13 2 With many slides from or derived from David Silver

More information

Bias-Variance in Machine Learning

Bias-Variance in Machine Learning Bias-Variance in Machine Learning Bias-Variance: Outline Underfitting/overfitting: Why are complex hypotheses bad? Simple example of bias/variance Error as bias+variance for regression brief comments on

More information