Bayesian Decision Theory Tutorial Visual Recognition Tutorial 1

Size: px
Start display at page:

Download "Bayesian Decision Theory Tutorial Visual Recognition Tutorial 1"

Transcription

1 Bayesian Decision Theory Tutorial C Visual Recognition Tutorial

2 Tutorial the outline Bayesian decision making with discrete probabilities an eample Looking at continuous densities Bayesian decision making with continuous probabilities an eample The Bayesian Doctor Eample Visual Recognition Tutorial

3 Eample checking on a course A student needs to achieve a decision on which courses to take, based only on his first lecture. From his previous eperience, he knows: Quality of the course good fair bad Probability prior These are prior probabilities Visual Recognition Tutorial 3

4 Eample continued The student also knows the class-conditionals: Pr j good fair bad Interesting lecture Boring lecture The loss function is given by the matri λa i j good course fair course bad course Taking the course Not taking the course Visual Recognition Tutorial 4

5 Eample continued The student wants to make an optimal decision> minimal possible Rά The probability to get the interesting lecture interesting: Printeresting Printerestinggood course* Prgood course + Printerestingfair course* Prfair course + Printerestingbad course* Prbad course 0.8*0.+0.5*0.4+0.* Consequently, Prboring Suppose the lecture was interesting. Then we want to compute the posterior probabilities of each one of the 3 possible states of nature Visual Recognition Tutorial 5

6 Eample continued Prgood courseinteresting lecture PrinterestinggoodPrgood 0.8*0. Printeresting Prfairinteresting PrinterestingfairPrfair 0.5* Printeresting 0.4 We can get Prbadinteresting0. either by the same method, or by noting that it complements to the above two. Now, we have all we need for making an intelligent decision about an optimal action Visual Recognition Tutorial 6

7 Eample conclusion The student needs to minimize the conditional risk; c R αi λ αi j P j j he can either take the course: Rtakinginteresting Prgoodinterestingλtakinggood course +Prfairinterestingλtakingfair course +Prbadinterestingλtakingbad course 0.4*0+0.5*5+0.*03.5 or drop it: Rnot takinginteresting Prgoodinterestingλnot takinggood course +Prfairinterestingλnot takingfair course +Prbadinterestingλnot takingbad course 0.4*0+0.5*5+0.* Visual Recognition Tutorial 7

8 Constructing an optimal decision function So, if the first lecture was interesting, the student will minimize the conditional risk by taking the course. In order to construct the full decision function, we need to define the risk minimization action for the case of boring lecture, as well. Do it! Visual Recognition Tutorial 8

9 Eample continuous density Let X be a real value r.v., representing a number randomly picked from the interval [0,]; its distribution is known to be uniform. Then let Y be a real r.v. whose value is chosen at random from [0, X] also with uniform distribution. We are presented with the value of Y, and need to guess the most likely value of X. In a more formal fashion:given the value of Y, find the probability density function p.d.f. of X and determine its maima Visual Recognition Tutorial 9

10 Eample continued Let w denote the state of nature, when X ; What we look for is Pw Yy that is, the p.d.f. The class-conditional given the value of X:, y P Y y w 0, < y For the given evidence: P Y y d ln y y using total probability Visual Recognition Tutorial 0

11 Eample conclusion Applying Bayes rule: p y w p w p w y p y ln y This is monotonically decreasing function, over [y,]. So informally the most likely value of X the one with highest probability density value is Xy Visual Recognition Tutorial

12 Illustration conditional p.d.f. 3.5 The conditional density py Visual Recognition Tutorial

13 Eample 3: hiring a secretary A manager needs to hire a new secretary, and a good one. Unfortunately, good secretary are hard to find: Prw g 0., Prw b 0.8 The manager decides to use a new test. The grade is a real number in the range from 0 to 00. The manager s estimation of the possible losses: λdecision,w i w g w b Hire Reject Visual Recognition Tutorial 3

14 Eample 3: continued The class conditional densities are known to be approimated by a normal p.d.f.: p grade good sec retary ~ N85,5 p grade bad sec retary ~ N40,3 pgradebadprbad pgradegoodprgood Visual Recognition Tutorial 4

15 Eample 3: continued The resulting probability density for the grade looks as follows: pp w b p w b + p w g p w g 0.05 p Visual Recognition Tutorial 5

16 Eample 3: continued We need to know for which grade values hiring the secretary would minimize the risk: Rhire < Rreject p w λhire, w + p w λhire, w b b g g < p w λreject, w + p w λreject, w b b g g [ λhire, w λreject, w ] p w < [ λreject, w λhire, w ] p w b b b b g g The posteriors are given by p w i p w p w i p i Visual Recognition Tutorial 6

17 Eample 3: continued The posteriors scaled by the loss differences, [ λ λ and look like: 30 hire, w reject, w ] p w b b b [ λreject, w λhire, w ] p w b g g 5 0 bad good Visual Recognition Tutorial 7

18 Numerically, we have: Eample 3: continued p e + e 5 π 3 π e e p w 3 5 b π, p wg π p p We need to solve 0 p w > 5 p w b g Solving numerically yields one solution in [0, 00]: Visual Recognition Tutorial 8

19 The Bayesian Doctor Eample A person doesn t feel well and goes to the doctor. Assume two states of nature: : The person has a common flue. : The person is really sick a vicious bacterial infection. The doctors prior is: p 0.9 p This doctor has two possible actions: ``prescribe hot tea or antibiotics. Doctor can use prior and predict optimally: always flue. Therefore doctor will always prescribe hot tea Visual Recognition Tutorial 9

20 The Bayesian Doctor - Cntd. But there is very high risk: Although this doctor can diagnose with very high rate of success using the prior, she can lose a patient once in a while. Denote the two possible actions: a prescribe hot tea a prescribe antibiotics Now assume the following cost loss matri: λ i, j a a Visual Recognition Tutorial 0

21 The Bayesian Doctor - Cntd. Choosing a results in epected risk of R a p λ, + p λ, Choosing a results in epected risk of R a p λ, + p λ, So, considering the costs it s much better and optimal! to always give antibiotics Visual Recognition Tutorial

22 The Bayesian Doctor - Cntd. But doctors can do more. For eample, they can take some observations. A reasonable observation is to perform a blood test. Suppose the possible results of the blood test are: negative no bacterial infection positive infection But blood tests can often fail. Suppose p 0.3 p 0.7 p 0. p class conditional probabilities Visual Recognition Tutorial

23 The Bayesian Doctor - Cntd. Define the conditional risk given the observation R ai p j λ We would like to compute the conditional risk for each action and observation so that the doctor can choose an optimal action that minimizes risk. How can we compute p? We use the class conditional probabilities and Bayes inversion rule. j j i, j Visual Recognition Tutorial 3

24 The Bayesian Doctor - Cntd. Let s calculate first p and p p p p + p p p is complementary to p, so p Visual Recognition Tutorial 4

25 The Bayesian Doctor - Cntd. R a p λ + p λ R a,, 0 + p 0 0 p p p p p p λ p p p, + p 0 λ, Visual Recognition Tutorial 5

26 36607 Visual Recognition Tutorial ,, + + p p p p p p a R λ λ ,, + + p p p p p p p a R λ λ The Bayesian Doctor - Cntd.

27 The Bayesian Doctor - Cntd. To summarize: R R R R a a a a Whenever we encounter an observation, we can minimize the epected loss by minimizing the conditional risk. Makes sense: Doctor chooses hot tea if blood test is negative, and antibiotics otherwise Visual Recognition Tutorial 7

28 Optimal Bayes Decision Strategies A strategy or decision function α is a mapping from observations to actions. The total risk of a decision function is given by E p [ R α ] p R α A decision function is optimal if it minimizes the total risk. This optimal total risk is called Bayes risk. In the Bayesian doctor eample: Total risk if doctor always gives antibiotics: 0.9 Bayes risk: 0.48 How have we got it? Visual Recognition Tutorial 8

Error Rates. Error vs Threshold. ROC Curve. Biometrics: A Pattern Recognition System. Pattern classification. Biometrics CSE 190 Lecture 3

Error Rates. Error vs Threshold. ROC Curve. Biometrics: A Pattern Recognition System. Pattern classification. Biometrics CSE 190 Lecture 3 Biometrics: A Pattern Recognition System Yes/No Pattern classification Biometrics CSE 190 Lecture 3 Authentication False accept rate (FAR): Proportion of imposters accepted False reject rate (FRR): Proportion

More information

Pattern Classification

Pattern Classification Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley & Sons, 2000 with the permission of the authors

More information

Probabilistic classification CE-717: Machine Learning Sharif University of Technology. M. Soleymani Fall 2016

Probabilistic classification CE-717: Machine Learning Sharif University of Technology. M. Soleymani Fall 2016 Probabilistic classification CE-717: Machine Learning Sharif University of Technology M. Soleymani Fall 2016 Topics Probabilistic approach Bayes decision theory Generative models Gaussian Bayes classifier

More information

Machine Learning Lecture 2

Machine Learning Lecture 2 Announcements Machine Learning Lecture 2 Eceptional number of lecture participants this year Current count: 449 participants This is very nice, but it stretches our resources to their limits Probability

More information

Topic -2. Probability. Larson & Farber, Elementary Statistics: Picturing the World, 3e 1

Topic -2. Probability. Larson & Farber, Elementary Statistics: Picturing the World, 3e 1 Topic -2 Probability Larson & Farber, Elementary Statistics: Picturing the World, 3e 1 Probability Experiments Experiment : An experiment is an act that can be repeated under given condition. Rolling a

More information

Machine Learning Lecture 2

Machine Learning Lecture 2 Machine Perceptual Learning and Sensory Summer Augmented 6 Computing Announcements Machine Learning Lecture 2 Course webpage http://www.vision.rwth-aachen.de/teaching/ Slides will be made available on

More information

STAT Chapter 3: Probability

STAT Chapter 3: Probability Basic Definitions STAT 515 --- Chapter 3: Probability Experiment: A process which leads to a single outcome (called a sample point) that cannot be predicted with certainty. Sample Space (of an experiment):

More information

Grundlagen der Künstlichen Intelligenz

Grundlagen der Künstlichen Intelligenz Grundlagen der Künstlichen Intelligenz Uncertainty & Probabilities & Bandits Daniel Hennes 16.11.2017 (WS 2017/18) University Stuttgart - IPVS - Machine Learning & Robotics 1 Today Uncertainty Probability

More information

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 3

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 3 CS434a/541a: attern Recognition rof. Olga Veksler Lecture 3 1 Announcements Link to error data in the book Reading assignment Assignment 1 handed out, due Oct. 4 lease send me an email with your name and

More information

Bayesian Decision Theory

Bayesian Decision Theory Bayesian Decision Theory Dr. Shuang LIANG School of Software Engineering TongJi University Fall, 2012 Today s Topics Bayesian Decision Theory Bayesian classification for normal distributions Error Probabilities

More information

What does Bayes theorem give us? Lets revisit the ball in the box example.

What does Bayes theorem give us? Lets revisit the ball in the box example. ECE 6430 Pattern Recognition and Analysis Fall 2011 Lecture Notes - 2 What does Bayes theorem give us? Lets revisit the ball in the box example. Figure 1: Boxes with colored balls Last class we answered

More information

Bayesian decision theory Introduction to Pattern Recognition. Lectures 4 and 5: Bayesian decision theory

Bayesian decision theory Introduction to Pattern Recognition. Lectures 4 and 5: Bayesian decision theory Bayesian decision theory 8001652 Introduction to Pattern Recognition. Lectures 4 and 5: Bayesian decision theory Jussi Tohka jussi.tohka@tut.fi Institute of Signal Processing Tampere University of Technology

More information

Lecture 3. STAT161/261 Introduction to Pattern Recognition and Machine Learning Spring 2018 Prof. Allie Fletcher

Lecture 3. STAT161/261 Introduction to Pattern Recognition and Machine Learning Spring 2018 Prof. Allie Fletcher Lecture 3 STAT161/261 Introduction to Pattern Recognition and Machine Learning Spring 2018 Prof. Allie Fletcher Previous lectures What is machine learning? Objectives of machine learning Supervised and

More information

Short course A vademecum of statistical pattern recognition techniques with applications to image and video analysis. Agenda

Short course A vademecum of statistical pattern recognition techniques with applications to image and video analysis. Agenda Short course A vademecum of statistical pattern recognition techniques with applications to image and video analysis Lecture Recalls of probability theory Massimo Piccardi University of Technology, Sydney,

More information

Bayesian decision theory. Nuno Vasconcelos ECE Department, UCSD

Bayesian decision theory. Nuno Vasconcelos ECE Department, UCSD Bayesian decision theory Nuno Vasconcelos ECE Department, UCSD Notation the notation in DHS is quite sloppy e.g. show that ( error = ( error z ( z dz really not clear what this means we will use the following

More information

Bayesian Decision and Bayesian Learning

Bayesian Decision and Bayesian Learning Bayesian Decision and Bayesian Learning Ying Wu Electrical Engineering and Computer Science Northwestern University Evanston, IL 60208 http://www.eecs.northwestern.edu/~yingwu 1 / 30 Bayes Rule p(x ω i

More information

14.30 Introduction to Statistical Methods in Economics Spring 2009

14.30 Introduction to Statistical Methods in Economics Spring 2009 MIT OpenCourseWare http://ocw.mit.edu 4.0 Introduction to Statistical Methods in Economics Spring 009 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.

More information

From Bayes Theorem to Pattern Recognition via Bayes Rule

From Bayes Theorem to Pattern Recognition via Bayes Rule From Bayes Theorem to Pattern Recognition via Bayes Rule Slecture by Varun Vasudevan (partially based on Prof. Mireille Boutin s ECE 662 lecture) February 12, 2014 What will you learn from this slecture?

More information

Bayes Decision Theory - I

Bayes Decision Theory - I Bayes Decision Theory - I Nuno Vasconcelos (Ken Kreutz-Delgado) UCSD Statistical Learning from Data Goal: Given a relationship between a feature vector and a vector y, and iid data samples ( i,y i ), find

More information

Bayesian Decision Theory

Bayesian Decision Theory Introduction to Pattern Recognition [ Part 4 ] Mahdi Vasighi Remarks It is quite common to assume that the data in each class are adequately described by a Gaussian distribution. Bayesian classifier is

More information

Intro. ANN & Fuzzy Systems. Lecture 15. Pattern Classification (I): Statistical Formulation

Intro. ANN & Fuzzy Systems. Lecture 15. Pattern Classification (I): Statistical Formulation Lecture 15. Pattern Classification (I): Statistical Formulation Outline Statistical Pattern Recognition Maximum Posterior Probability (MAP) Classifier Maximum Likelihood (ML) Classifier K-Nearest Neighbor

More information

EECS 126 Probability and Random Processes University of California, Berkeley: Spring 2015 Abhay Parekh February 17, 2015.

EECS 126 Probability and Random Processes University of California, Berkeley: Spring 2015 Abhay Parekh February 17, 2015. EECS 126 Probability and Random Processes University of California, Berkeley: Spring 2015 Abhay Parekh February 17, 2015 Midterm Exam Last name First name SID Rules. You have 80 mins (5:10pm - 6:30pm)

More information

Probability Exercises. Problem 1.

Probability Exercises. Problem 1. Probability Exercises. Ma 162 Spring 2010 Ma 162 Spring 2010 April 21, 2010 Problem 1. ˆ Conditional Probability: It is known that a student who does his online homework on a regular basis has a chance

More information

II. The Binomial Distribution

II. The Binomial Distribution 88 CHAPTER 4 PROBABILITY DISTRIBUTIONS 進佳數學團隊 Dr. Herbert Lam 林康榮博士 HKDSE Mathematics M1 II. The Binomial Distribution 1. Bernoulli distribution A Bernoulli eperiment results in any one of two possible

More information

Math 180A. Lecture 16 Friday May 7 th. Expectation. Recall the three main probability density functions so far (1) Uniform (2) Exponential.

Math 180A. Lecture 16 Friday May 7 th. Expectation. Recall the three main probability density functions so far (1) Uniform (2) Exponential. Math 8A Lecture 6 Friday May 7 th Epectation Recall the three main probability density functions so far () Uniform () Eponential (3) Power Law e, ( ), Math 8A Lecture 6 Friday May 7 th Epectation Eample

More information

HST.582J / 6.555J / J Biomedical Signal and Image Processing Spring 2007

HST.582J / 6.555J / J Biomedical Signal and Image Processing Spring 2007 MIT OpenCourseWare http://ocw.mit.edu HST.582J / 6.555J / 16.456J Biomedical Signal and Image Processing Spring 2007 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.

More information

COS 424: Interacting with Data. Lecturer: Dave Blei Lecture #2 Scribe: Ellen Kim February 7, 2008

COS 424: Interacting with Data. Lecturer: Dave Blei Lecture #2 Scribe: Ellen Kim February 7, 2008 COS 424: Interacting with Data Lecturer: Dave Blei Lecture #2 Scribe: Ellen Kim February 7, 2008 1 Probability Review 1.1 Random Variables A random variable can be used to model any probabilistic outcome.

More information

Lecture Notes 2 Random Variables. Discrete Random Variables: Probability mass function (pmf)

Lecture Notes 2 Random Variables. Discrete Random Variables: Probability mass function (pmf) Lecture Notes 2 Random Variables Definition Discrete Random Variables: Probability mass function (pmf) Continuous Random Variables: Probability density function (pdf) Mean and Variance Cumulative Distribution

More information

Final Examination. Adrian Georgi Josh Karen Lee Min Nikos Tina. There are 12 problems totaling 150 points. Total time is 170 minutes.

Final Examination. Adrian Georgi Josh Karen Lee Min Nikos Tina. There are 12 problems totaling 150 points. Total time is 170 minutes. Massachusetts Institute of Technology 6.042J/18.062J, Fall 02: Mathematics for Computer Science Prof. Albert Meyer and Dr. Radhika Nagpal Final Examination Your name: Circle the name of your Tutorial Instructor:

More information

CS626 Data Analysis and Simulation

CS626 Data Analysis and Simulation CS626 Data Analysis and Simulation Instructor: Peter Kemper R 104A, phone 221-3462, email:kemper@cs.wm.edu Today: Probability Primer Quick Reference: Sheldon Ross: Introduction to Probability Models 9th

More information

Bayes Decision Theory

Bayes Decision Theory Bayes Decision Theory Minimum-Error-Rate Classification Classifiers, Discriminant Functions and Decision Surfaces The Normal Density 0 Minimum-Error-Rate Classification Actions are decisions on classes

More information

Deep Learning for Computer Vision

Deep Learning for Computer Vision Deep Learning for Computer Vision Lecture 3: Probability, Bayes Theorem, and Bayes Classification Peter Belhumeur Computer Science Columbia University Probability Should you play this game? Game: A fair

More information

ECO421: Communication

ECO421: Communication ECO421: Communication Marcin P ski February 9, 2018 Plan Introduction Asymmetric information means some players know more than the others. In real life, information can be transmitted A simple model of

More information

Model Averaging (Bayesian Learning)

Model Averaging (Bayesian Learning) Model Averaging (Bayesian Learning) We want to predict the output Y of a new case that has input X = x given the training examples e: p(y x e) = m M P(Y m x e) = m M P(Y m x e)p(m x e) = m M P(Y m x)p(m

More information

Chapter 7 Probability Basics

Chapter 7 Probability Basics Making Hard Decisions Chapter 7 Probability Basics Slide 1 of 62 Introduction A A,, 1 An Let be an event with possible outcomes: 1 A = Flipping a coin A = {Heads} A 2 = Ω {Tails} The total event (or sample

More information

22. RADICALS. x add 5. multiply by 7

22. RADICALS. x add 5. multiply by 7 22. RADICALS doing something, then undoing it The concept of doing something and then undoing it is very important in mathematics. Here are some eamples: Take a number. Add 5 to it. How can you get back

More information

Probability, Statistics, and Bayes Theorem Session 3

Probability, Statistics, and Bayes Theorem Session 3 Probability, Statistics, and Bayes Theorem Session 3 1 Introduction Now that we know what Bayes Theorem is, we want to explore some of the ways that it can be used in real-life situations. Often the results

More information

PATTERN RECOGNITION AND MACHINE LEARNING

PATTERN RECOGNITION AND MACHINE LEARNING PATTERN RECOGNITION AND MACHINE LEARNING Chapter 1. Introduction Shuai Huang April 21, 2014 Outline 1 What is Machine Learning? 2 Curve Fitting 3 Probability Theory 4 Model Selection 5 The curse of dimensionality

More information

ebay/google short course: Problem set 2

ebay/google short course: Problem set 2 18 Jan 013 ebay/google short course: Problem set 1. (the Echange Parado) You are playing the following game against an opponent, with a referee also taking part. The referee has two envelopes (numbered

More information

Intelligent Systems Statistical Machine Learning

Intelligent Systems Statistical Machine Learning Intelligent Systems Statistical Machine Learning Carsten Rother, Dmitrij Schlesinger WS2014/2015, Our tasks (recap) The model: two variables are usually present: - the first one is typically discrete k

More information

Recall from last time: Conditional probabilities. Lecture 2: Belief (Bayesian) networks. Bayes ball. Example (continued) Example: Inference problem

Recall from last time: Conditional probabilities. Lecture 2: Belief (Bayesian) networks. Bayes ball. Example (continued) Example: Inference problem Recall from last time: Conditional probabilities Our probabilistic models will compute and manipulate conditional probabilities. Given two random variables X, Y, we denote by Lecture 2: Belief (Bayesian)

More information

Basics of Probability

Basics of Probability Basics of Probability Lecture 1 Doug Downey, Northwestern EECS 474 Events Event space E.g. for dice, = {1, 2, 3, 4, 5, 6} Set of measurable events S 2 E.g., = event we roll an even number = {2, 4, 6} S

More information

Classifier Evaluation. Learning Curve cleval testc. The Apparent Classification Error. Error Estimation by Test Set. Classifier

Classifier Evaluation. Learning Curve cleval testc. The Apparent Classification Error. Error Estimation by Test Set. Classifier Classifier Learning Curve How to estimate classifier performance. Learning curves Feature curves Rejects and ROC curves True classification error ε Bayes error ε* Sub-optimal classifier Bayes consistent

More information

Dr. Allen Back. Nov. 21, 2016

Dr. Allen Back. Nov. 21, 2016 Dr. Allen Back Nov. 21, 2016 , Type I/II Errors, α, and β Given H 0 : p = p 0, there are two ways an HT can report an inaccurate result: , Type I/II Errors, α, and β Given H 0 : p = p 0, there are two

More information

Machine Learning Lecture 7

Machine Learning Lecture 7 Course Outline Machine Learning Lecture 7 Fundamentals (2 weeks) Bayes Decision Theory Probability Density Estimation Statistical Learning Theory 23.05.2016 Discriminative Approaches (5 weeks) Linear Discriminant

More information

Name Date. Show all work! Exact answers only unless the problem asks for an approximation.

Name Date. Show all work! Exact answers only unless the problem asks for an approximation. Advanced Calculus & AP Calculus AB Summer Assignment Name Date Show all work! Eact answers only unless the problem asks for an approimation. These are important topics from previous courses that you must

More information

Review of elements of Calculus (functions in one variable)

Review of elements of Calculus (functions in one variable) Review of elements of Calculus (functions in one variable) Mainly adapted from the lectures of prof Greg Kelly Hanford High School, Richland Washington http://online.math.uh.edu/houstonact/ https://sites.google.com/site/gkellymath/home/calculuspowerpoints

More information

Inconsistency of Bayesian inference when the model is wrong, and how to repair it

Inconsistency of Bayesian inference when the model is wrong, and how to repair it Inconsistency of Bayesian inference when the model is wrong, and how to repair it Peter Grünwald Thijs van Ommen Centrum Wiskunde & Informatica, Amsterdam Universiteit Leiden June 3, 2015 Outline 1 Introduction

More information

Machine Learning Lecture 2

Machine Learning Lecture 2 Machine Perceptual Learning and Sensory Summer Augmented 15 Computing Many slides adapted from B. Schiele Machine Learning Lecture 2 Probability Density Estimation 16.04.2015 Bastian Leibe RWTH Aachen

More information

Introduction to Machine Learning (Pattern recognition and model fitting) for Master students

Introduction to Machine Learning (Pattern recognition and model fitting) for Master students Introduction to Machine Learning (Pattern recognition and model fitting) for Master students Spring 007, ÖU/RAP Thorsteinn Rögnvaldsson thorsteinn.rognvaldsson@tech.oru.se Contents Machine learning algorithms

More information

IOE 202: lecture 14 outline

IOE 202: lecture 14 outline IOE 202: lecture 14 outline Announcements Last time... Value of information (perfect and imperfect) in decision analysis Course wrap-up: topics covered and where to go from here IOE 202: Operations Modeling,

More information

Physics 509: Bootstrap and Robust Parameter Estimation

Physics 509: Bootstrap and Robust Parameter Estimation Physics 509: Bootstrap and Robust Parameter Estimation Scott Oser Lecture #20 Physics 509 1 Nonparametric parameter estimation Question: what error estimate should you assign to the slope and intercept

More information

2. A Basic Statistical Toolbox

2. A Basic Statistical Toolbox . A Basic Statistical Toolbo Statistics is a mathematical science pertaining to the collection, analysis, interpretation, and presentation of data. Wikipedia definition Mathematical statistics: concerned

More information

Lecture Notes 2 Random Variables. Random Variable

Lecture Notes 2 Random Variables. Random Variable Lecture Notes 2 Random Variables Definition Discrete Random Variables: Probability mass function (pmf) Continuous Random Variables: Probability density function (pdf) Mean and Variance Cumulative Distribution

More information

Random Variables. Definition: A random variable (r.v.) X on the probability space (Ω, F, P) is a mapping

Random Variables. Definition: A random variable (r.v.) X on the probability space (Ω, F, P) is a mapping Random Variables Example: We roll a fair die 6 times. Suppose we are interested in the number of 5 s in the 6 rolls. Let X = number of 5 s. Then X could be 0, 1, 2, 3, 4, 5, 6. X = 0 corresponds to the

More information

2 Statistical Estimation: Basic Concepts

2 Statistical Estimation: Basic Concepts Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof. N. Shimkin 2 Statistical Estimation:

More information

PubH 7470: STATISTICS FOR TRANSLATIONAL & CLINICAL RESEARCH

PubH 7470: STATISTICS FOR TRANSLATIONAL & CLINICAL RESEARCH PubH 7470: STATISTICS FOR TRANSLATIONAL & CLINICAL RESEARCH The First Step: SAMPLE SIZE DETERMINATION THE ULTIMATE GOAL The most important, ultimate step of any of clinical research is to do draw inferences;

More information

Lecture Slides - Part 1

Lecture Slides - Part 1 Lecture Slides - Part 1 Bengt Holmstrom MIT February 2, 2016. Bengt Holmstrom (MIT) Lecture Slides - Part 1 February 2, 2016. 1 / 36 Going to raise the level a little because 14.281 is now taught by Juuso

More information

Business Statistics: A Decision-Making Approach 6 th Edition. Chapter Goals

Business Statistics: A Decision-Making Approach 6 th Edition. Chapter Goals Chapter 6 Student Lecture Notes 6-1 Business Statistics: A Decision-Making Approach 6 th Edition Chapter 6 Introduction to Sampling Distributions Chap 6-1 Chapter Goals To use information from the sample

More information

Intelligent Systems Statistical Machine Learning

Intelligent Systems Statistical Machine Learning Intelligent Systems Statistical Machine Learning Carsten Rother, Dmitrij Schlesinger WS2015/2016, Our model and tasks The model: two variables are usually present: - the first one is typically discrete

More information

Conditional probability

Conditional probability CHAPTER 4 Conditional probability 4.1. Introduction Suppose there are 200 men, of which 100 are smokers, and 100 women, of which 20 are smokers. What is the probability that a person chosen at random will

More information

Machine Learning Lecture 3

Machine Learning Lecture 3 Announcements Machine Learning Lecture 3 Eam dates We re in the process of fiing the first eam date Probability Density Estimation II 9.0.207 Eercises The first eercise sheet is available on L2P now First

More information

Do students sleep the recommended 8 hours a night on average?

Do students sleep the recommended 8 hours a night on average? BIEB100. Professor Rifkin. Notes on Section 2.2, lecture of 27 January 2014. Do students sleep the recommended 8 hours a night on average? We first set up our null and alternative hypotheses: H0: μ= 8

More information

Lecture 10: Introduction to reasoning under uncertainty. Uncertainty

Lecture 10: Introduction to reasoning under uncertainty. Uncertainty Lecture 10: Introduction to reasoning under uncertainty Introduction to reasoning under uncertainty Review of probability Axioms and inference Conditional probability Probability distributions COMP-424,

More information

Math 416 Lecture 2 DEFINITION. Here are the multivariate versions: X, Y, Z iff P(X = x, Y = y, Z =z) = p(x, y, z) of X, Y, Z iff for all sets A, B, C,

Math 416 Lecture 2 DEFINITION. Here are the multivariate versions: X, Y, Z iff P(X = x, Y = y, Z =z) = p(x, y, z) of X, Y, Z iff for all sets A, B, C, Math 416 Lecture 2 DEFINITION. Here are the multivariate versions: PMF case: p(x, y, z) is the joint Probability Mass Function of X, Y, Z iff P(X = x, Y = y, Z =z) = p(x, y, z) PDF case: f(x, y, z) is

More information

COURSE INTRODUCTION. J. Elder CSE 6390/PSYC 6225 Computational Modeling of Visual Perception

COURSE INTRODUCTION. J. Elder CSE 6390/PSYC 6225 Computational Modeling of Visual Perception COURSE INTRODUCTION COMPUTATIONAL MODELING OF VISUAL PERCEPTION 2 The goal of this course is to provide a framework and computational tools for modeling visual inference, motivated by interesting examples

More information

Week 4 Spring Lecture 7. Empirical Bayes, hierarchical Bayes and random effects

Week 4 Spring Lecture 7. Empirical Bayes, hierarchical Bayes and random effects Week 4 Spring 9 Lecture 7 Empirical Bayes, hierarchical Bayes and random effects Empirical Bayes and Hierarchical Bayes eview: Let X N (; ) Consider a normal prior N (; ) Then the posterior distribution

More information

Solution to HW 12. Since B and B 2 form a partition, we have P (A) = P (A B 1 )P (B 1 ) + P (A B 2 )P (B 2 ). Using P (A) = 21.

Solution to HW 12. Since B and B 2 form a partition, we have P (A) = P (A B 1 )P (B 1 ) + P (A B 2 )P (B 2 ). Using P (A) = 21. Solution to HW 12 (1) (10 pts) Sec 12.3 Problem A screening test for a disease shows a positive result in 92% of all cases when the disease is actually present and in 7% of all cases when it is not. Assume

More information

Quantitative Understanding in Biology 1.7 Bayesian Methods

Quantitative Understanding in Biology 1.7 Bayesian Methods Quantitative Understanding in Biology 1.7 Bayesian Methods Jason Banfelder October 25th, 2018 1 Introduction So far, most of the methods we ve looked at fall under the heading of classical, or frequentist

More information

Frequentist Statistics and Hypothesis Testing Spring

Frequentist Statistics and Hypothesis Testing Spring Frequentist Statistics and Hypothesis Testing 18.05 Spring 2018 http://xkcd.com/539/ Agenda Introduction to the frequentist way of life. What is a statistic? NHST ingredients; rejection regions Simple

More information

MAE 493G, CpE 493M, Mobile Robotics. 6. Basic Probability

MAE 493G, CpE 493M, Mobile Robotics. 6. Basic Probability MAE 493G, CpE 493M, Mobile Robotics 6. Basic Probability Instructor: Yu Gu, Fall 2013 Uncertainties in Robotics Robot environments are inherently unpredictable; Sensors and data acquisition systems are

More information

Bayes Rule for Minimizing Risk

Bayes Rule for Minimizing Risk Bayes Rule for Minimizing Risk Dennis Lee April 1, 014 Introduction In class we discussed Bayes rule for minimizing the probability of error. Our goal is to generalize this rule to minimize risk instead

More information

CS 4491/CS 7990 SPECIAL TOPICS IN BIOINFORMATICS

CS 4491/CS 7990 SPECIAL TOPICS IN BIOINFORMATICS CS 4491/CS 7990 SPECIAL TOPICS IN BIOINFORMATICS * The contents are adapted from Dr. Jean Gao at UT Arlington Mingon Kang, Ph.D. Computer Science, Kennesaw State University Primer on Probability Random

More information

This exam contains 13 pages (including this cover page) and 10 questions. A Formulae sheet is provided with the exam.

This exam contains 13 pages (including this cover page) and 10 questions. A Formulae sheet is provided with the exam. Probability and Statistics FS 2017 Session Exam 22.08.2017 Time Limit: 180 Minutes Name: Student ID: This exam contains 13 pages (including this cover page) and 10 questions. A Formulae sheet is provided

More information

Be able to define the following terms and answer basic questions about them:

Be able to define the following terms and answer basic questions about them: CS440/ECE448 Fall 2016 Final Review Be able to define the following terms and answer basic questions about them: Probability o Random variables o Axioms of probability o Joint, marginal, conditional probability

More information

Human-Oriented Robotics. Probability Refresher. Kai Arras Social Robotics Lab, University of Freiburg Winter term 2014/2015

Human-Oriented Robotics. Probability Refresher. Kai Arras Social Robotics Lab, University of Freiburg Winter term 2014/2015 Probability Refresher Kai Arras, University of Freiburg Winter term 2014/2015 Probability Refresher Introduction to Probability Random variables Joint distribution Marginalization Conditional probability

More information

Hypothesis testing I. - In particular, we are talking about statistical hypotheses. [get everyone s finger length!] n =

Hypothesis testing I. - In particular, we are talking about statistical hypotheses. [get everyone s finger length!] n = Hypothesis testing I I. What is hypothesis testing? [Note we re temporarily bouncing around in the book a lot! Things will settle down again in a week or so] - Exactly what it says. We develop a hypothesis,

More information

A Short Introduction to the Theory of Statistical Learning. Kee Siong Ng, PhD

A Short Introduction to the Theory of Statistical Learning. Kee Siong Ng, PhD A Short Introduction to the Theory of Statistical Learning Kee Siong Ng, PhD keesiong.ng@gmail.com Outline Introduction Bayesian Probability Theory Sequence Prediction and Data Compression Bayesian Networks

More information

With Question/Answer Animations. Chapter 7

With Question/Answer Animations. Chapter 7 With Question/Answer Animations Chapter 7 Chapter Summary Introduction to Discrete Probability Probability Theory Bayes Theorem Section 7.1 Section Summary Finite Probability Probabilities of Complements

More information

Parameter estimation Conditional risk

Parameter estimation Conditional risk Parameter estimation Conditional risk Formalizing the problem Specify random variables we care about e.g., Commute Time e.g., Heights of buildings in a city We might then pick a particular distribution

More information

Machine Learning. Classification. Bayes Classifier. Representing data: Choosing hypothesis class. Learning: h:x a Y. Eric Xing

Machine Learning. Classification. Bayes Classifier. Representing data: Choosing hypothesis class. Learning: h:x a Y. Eric Xing Machine Learning 10-701/15 701/15-781, 781, Spring 2008 Naïve Bayes Classifier Eric Xing Lecture 3, January 23, 2006 Reading: Chap. 4 CB and handouts Classification Representing data: Choosing hypothesis

More information

Chapter 2 Bayesian Decision Theory. Pattern Recognition Soochow, Fall Semester 1

Chapter 2 Bayesian Decision Theory. Pattern Recognition Soochow, Fall Semester 1 Chapter 2 Bayesian Decision Theory Pattern Recognition Soochow, Fall Semester 1 Decision Theory Decision Make choice under uncertainty Pattern Recognition Pattern Category Given a test sample, its category

More information

ECE521 week 3: 23/26 January 2017

ECE521 week 3: 23/26 January 2017 ECE521 week 3: 23/26 January 2017 Outline Probabilistic interpretation of linear regression - Maximum likelihood estimation (MLE) - Maximum a posteriori (MAP) estimation Bias-variance trade-off Linear

More information

Bayesian Updating: Discrete Priors: Spring

Bayesian Updating: Discrete Priors: Spring Bayesian Updating: Discrete Priors: 18.05 Spring 2017 http://xkcd.com/1236/ Learning from experience Which treatment would you choose? 1. Treatment 1: cured 100% of patients in a trial. 2. Treatment 2:

More information

Basic Probabilistic Reasoning SEG

Basic Probabilistic Reasoning SEG Basic Probabilistic Reasoning SEG 7450 1 Introduction Reasoning under uncertainty using probability theory Dealing with uncertainty is one of the main advantages of an expert system over a simple decision

More information

Statistical Decision Theory and Bayesian Analysis Chapter1 Basic Concepts. Outline. Introduction. Introduction(cont.) Basic Elements (cont.

Statistical Decision Theory and Bayesian Analysis Chapter1 Basic Concepts. Outline. Introduction. Introduction(cont.) Basic Elements (cont. Statistical Decision Theory and Bayesian Analysis Chapter Basic Concepts 939 Outline Introduction Basic Elements Bayesian Epected Loss Frequentist Risk Randomized Decision Rules Decision Principles Misuse

More information

ECE531 Homework Assignment Number 2

ECE531 Homework Assignment Number 2 ECE53 Homework Assignment Number 2 Due by 8:5pm on Thursday 5-Feb-29 Make sure your reasoning and work are clear to receive full credit for each problem 3 points A city has two taxi companies distinguished

More information

Roberto s Notes on Differential Calculus Chapter 1: Limits and continuity Section 7. Discontinuities. is the tool to use,

Roberto s Notes on Differential Calculus Chapter 1: Limits and continuity Section 7. Discontinuities. is the tool to use, Roberto s Notes on Differential Calculus Chapter 1: Limits and continuity Section 7 Discontinuities What you need to know already: The concept and definition of continuity. What you can learn here: The

More information

Machine Learning. Gaussian Mixture Models. Zhiyao Duan & Bryan Pardo, Machine Learning: EECS 349 Fall

Machine Learning. Gaussian Mixture Models. Zhiyao Duan & Bryan Pardo, Machine Learning: EECS 349 Fall Machine Learning Gaussian Mixture Models Zhiyao Duan & Bryan Pardo, Machine Learning: EECS 349 Fall 2012 1 The Generative Model POV We think of the data as being generated from some process. We assume

More information

P (B)P (A B) = P (AB) = P (A)P (B A), and dividing both sides by P (B) gives Bayes rule: P (A B) = P (A) P (B A) P (B),

P (B)P (A B) = P (AB) = P (A)P (B A), and dividing both sides by P (B) gives Bayes rule: P (A B) = P (A) P (B A) P (B), Conditional probability 18.600 Problem Set 3, due March 2 Welcome to your third 18.600 problem set! Conditional probability is defined by P (A B) = P (AB)/P (B), which implies P (B)P (A B) = P (AB) = P

More information

Class 26: review for final exam 18.05, Spring 2014

Class 26: review for final exam 18.05, Spring 2014 Probability Class 26: review for final eam 8.05, Spring 204 Counting Sets Inclusion-eclusion principle Rule of product (multiplication rule) Permutation and combinations Basics Outcome, sample space, event

More information

Bayesian Networks Structure Learning (cont.)

Bayesian Networks Structure Learning (cont.) Koller & Friedman Chapters (handed out): Chapter 11 (short) Chapter 1: 1.1, 1., 1.3 (covered in the beginning of semester) 1.4 (Learning parameters for BNs) Chapter 13: 13.1, 13.3.1, 13.4.1, 13.4.3 (basic

More information

CS Homework 2: Combinatorics & Discrete Events Due Date: September 25, 2018 at 2:20 PM

CS Homework 2: Combinatorics & Discrete Events Due Date: September 25, 2018 at 2:20 PM CS1450 - Homework 2: Combinatorics & Discrete Events Due Date: September 25, 2018 at 2:20 PM Question 1 A website allows the user to create an 8-character password that consists of lower case letters (a-z)

More information

Lecture 27: More on Rotational Kinematics

Lecture 27: More on Rotational Kinematics Lecture 27: More on Rotational Kinematics Let s work out the kinematics of rotational motion if α is constant: dω α = 1 2 α dω αt = ω ω ω = αt + ω ( t ) dφ α + ω = dφ t 2 α + ωo = φ φo = 1 2 = t o 2 φ

More information

Discrete Mathematics and Probability Theory Fall 2010 Tse/Wagner MT 2 Soln

Discrete Mathematics and Probability Theory Fall 2010 Tse/Wagner MT 2 Soln CS 70 Discrete Mathematics and Probability heory Fall 00 se/wagner M Soln Problem. [Rolling Dice] (5 points) You roll a fair die three times. Consider the following events: A first roll is a 3 B second

More information

Bayes Rule: A Tutorial Introduction ( www45w9klk)

Bayes Rule: A Tutorial Introduction (  www45w9klk) Bayes Rule: A Tutorial Introduction (https://docs.google.com/document/pub?id=1qm4hj4xmmowfvsklgcqpudojya5qcnii_ www45w9klk) by JV Stone 1 Introduction All decisions are based on data, but the best decisions

More information

9 - Markov processes and Burt & Allison 1963 AGEC

9 - Markov processes and Burt & Allison 1963 AGEC This document was generated at 8:37 PM on Saturday, March 17, 2018 Copyright 2018 Richard T. Woodward 9 - Markov processes and Burt & Allison 1963 AGEC 642-2018 I. What is a Markov Chain? A Markov chain

More information

Lecture 5: The Principle of Deferred Decisions. Chernoff Bounds

Lecture 5: The Principle of Deferred Decisions. Chernoff Bounds Randomized Algorithms Lecture 5: The Principle of Deferred Decisions. Chernoff Bounds Sotiris Nikoletseas Associate Professor CEID - ETY Course 2013-2014 Sotiris Nikoletseas, Associate Professor Randomized

More information

SYDE 372 Introduction to Pattern Recognition. Probability Measures for Classification: Part I

SYDE 372 Introduction to Pattern Recognition. Probability Measures for Classification: Part I SYDE 372 Introduction to Pattern Recognition Probability Measures for Classification: Part I Alexander Wong Department of Systems Design Engineering University of Waterloo Outline 1 2 3 4 Why use probability

More information

ENGR 200 ENGR 200. What did we do last week?

ENGR 200 ENGR 200. What did we do last week? ENGR 200 What did we do last week? Definition of probability xioms of probability Sample space robability laws Conditional probability ENGR 200 Lecture 3: genda. Conditional probability 2. Multiplication

More information