Why I think Bayesian Analysis is the greatest thing since NHST (bread)

Size: px
Start display at page:

Download "Why I think Bayesian Analysis is the greatest thing since NHST (bread)"

Transcription

1 Why I think Bayesian Analysis is the greatest thing since NHST (bread) Data Binge Nate Powell May 12, 2017 NHST = Null Hypothesis Significance TesIng

2 Textbook Doing Bayesian Data Analysis by John K. Kruschke Memory and the ComputaIonal Brain by Gallistel and King

3 What is Bayesian Analysis? hrp://xkcd.com/1132/

4 Advantages of Bayesian StaIsIcs vs. FrequenIst StaIsIcs Degree of belief is graded (not binary) Bayesian Update (incorporates new data) Can provide reasonable answers given limited data Can describe the probability of events that have not yet occurred

5 Advantages of Bayesian StaIsIcs vs. FrequenIst StaIsIcs Degree of belief is graded (not binary) what if your {X} test returns p =.06? or p =.055? a bayesian analysis will reflect this as a small change in posterior probability, not a complete shib between true and false. Bayesian Update (incorporates new data) Can provide reasonable answers given limited data Can describe the probability of events that have not yet occurred

6 Advantages of Bayesian StaIsIcs vs. FrequenIst StaIsIcs Degree of belief is graded (not binary) Bayesian Update (incorporates new data) if your p-value is 0.06, are you allowed to add new samples? A posterior distribuion can always be the prior for a bayesian update step if more informaion is collected. Can provide reasonable answers given limited data Can describe the probability of events that have not yet occurred

7 Advantages of Bayesian StaIsIcs vs. FrequenIst StaIsIcs Degree of belief is graded (not binary) Bayesian Update (incorporates new data) Can provide reasonable answers given limited data Limited data (which may not be significant by most tests) can be enough to modify or help confirm an untrusted prior, but won t disturb a trusted one... BUT HERE BE DRAGONS! Can describe the probability of events that have not yet occurred

8 Advantages of Bayesian StaIsIcs vs. FrequenIst StaIsIcs Degree of belief is graded (not binary) Bayesian Update (incorporates new data) Can provide reasonable answers given limited data Can describe the probability of events that have not yet occurred Bayesian methods were used to assess the likelihood of nuclear disasters for US policy (prior to any having occurred) with impressive accuracy.

9 Advantages of Bayesian StaIsIcs vs. Terminology, Conditional Probability FrequenIst StaIsIcs By the way, who s afraid of math? Who s afraid of equations? FrequenIst concern: P(D H) D = Data H = Hypothesis Bayesian Concern: P(ɵ D) or P(M D,ɵ) D = Data ɵ = model parameters M = Model hrp://xkcd.com/892/

10 Bayesian Update Learning from New experience to update previous beliefs P(M D) = P(D M) P(M) / P(D) Posterior Liklihood Prior Evidence This also allows us to incorporate new data without having to change any assumpions hrp://xkcd.com/1236/

11 Bayesian Update : twins example A physicist is having Twins, and wants to know the odds that her twins will be idenical (vs. fraternal). The Dr. told her that 2/3 of twin births are fraternal, and 1/3 are idenical. Also, a sonogram confirmed her twins are both male. (All idenical twins are the same sex, Fraternal can be either the same or mixed.) Efron, B. (2013). A 250-year argument: Belief, behavior, and the bootstrap. Bulle)n of the American Mathema)cal Society, 50(1),

12 Bayesian Update : twins example Prior DistribuIon: P(M = idenical) = 1/3 P(M = fraternal) = 2/3 AddiIonal Evidence: D = Same Sex P(D) = 2/3* Likelihood: P(same sex M = idenical) = 1 P(same sex M = fraternal) = 1/2 Posterior: P(M = idenical same sex) = 1/2 P(M = fraternal same sex) = 1/2 Posterior Prior DistribuIon: Efron, B. (2013). A 250-year argument: Belief, behavior, and the bootstrap. Bulle)n of the American Mathema)cal Society, 50(1), Identical Twins are: Fraternal Sonogram shows: Same sex 1/3 Physicist a Different 1/3 0 c 1/3 1/3 2/3 0 P(fraternal) P(idenIcal) b d Doctor

13 Bayesian Update : twins example But if you prefer a visual confirmaion: Sonogram shows: 1 Prior Same sex Different a b Identical Twins are: Fraternal 1/3 0 c 1/3 1/3 d 1/3 2/3 Doctor 1 0 P(frat) P(id) Posterior Physicist 0 P(frat) P(id) Efron, B. (2013). A 250-year argument: Belief, behavior, and the bootstrap. Bulle)n of the American Mathema)cal Society, 50(1),

14 Bayesian probability vs. FrequenIst probability FrequenIst Probability: The frequency with which events happen in a large number of trials Bayesian Probability: The Degree of belief you have that a certain event will happen (on any given trial) hrp://oikosjournal.wordpress.com/2011/10/11/frequenist-vs-bayesian-staisics-resources-to-helpyou-choose/

15 so, what, if anything, is wrong with NHST? Suppose we flip a coin N = 26 Imes and we come up with z = 8 heads is this coin biased or fair? Note, a bayesian would ask what is the probability that the coin is fair...

16 NHST N = 26 z = 8 is this coin biased? if the coin were fair we would expect N = 26 z = 13 The quesion we must answer is can we reject the null hypothesis, in this case, that the coin is fair and we saw these results by chance alone... If we repeated this experiment many Imes, would we expect to see z = 8 more than 5% of the Ime (with a fair coin?)

17 NHST The probability of seeing z heads in N flips is given by the binomial probability distribuion: where: If we repeated this experiment many Imes, would we expect to see z = 8 more than 5% of the Ime (with a fair coin?)

18 NHST If we repeated this experiment many Imes, would we expect to see z = 8 more than 5% of the Ime (with a fair coin?) The probability of seeing z heads in N flips is given by the binomial probability distribuion: Hypothetical Population Implied Sampling Distribution p(y) = 0.5 N = 26 p(z N,) tail y head z

19 NHST Hypothetical Population Implied Sampling Distribution p(y) tail y = 0.5 head N = 26 p(z N,) We want to know the probability of finding a result as low (or lower) than 8 heads out of 26 flips... This is the p-value of z = 8 for this sampling distribuion! z

20 NHST Hypothetical Population Implied Sampling Distribution p(y) tail y = 0.5 head N = 26 p(z N,) In this case we will consider the two-sided p-value (shaded) we can reject the null if z 7 or z 19 Thus we FAIL to reject the Null Hypothesis... z

21 NHST Thus we FAIL to reject the Null Hypothesis... What does that mean? Failing to reject the null hypothesis is like kissing your sister... we don t have an answer, but there s not a good reason to say the coin is biased.

22 NHST Thus we FAIL to reject the Null Hypothesis... But its more probable the coin is fair right? Do we know how likely the coin is to be fair? What if we had Really REALLY good reasons to believe it was fair?

23 NHST Thus we FAIL to reject the Null Hypothesis... But at least that s it right... I mean, it s an unsaisfying answer, but at least we know what it is... Unless...

24 NHST we used a sampling distribuion for N = 26 Hypothetical Population Implied Sampling Distribution p(y) tail y = 0.5 head N = 26 p(z N,) What if instead they sampled unil they found z = 8? We would have to use a DIFFERENT sampling distribuion... z

25 NHST Sampled coin unil we got z = 8, it took N = 26 flips... which gives us: Hypothetical Population Implied Sampling Distribution p(y) = 0.5 z = 8 p(n z,) tail y head N

26 NHST Sampled coin unil we got z = 8, it took N = 26 flips... Hypothetical Population Implied Sampling Distribution p(y) tail y = 0.5 head z = 8 p(n z,) N In this case, for p <.05, our values are N 9 and N 26 Therefore we REJECT the Null hypothesis and conclude the coin is biased!

27 NHST Therefore we REJECT the Null hypothesis and conclude the coin is biased! Now, do we have an expression of how likely the coin is to be biased? Or fair? And again, what if we were REALLY REALLY sure that the coin was fair?

28 NHST But we have an even bigger problem here! Thus we FAIL to reject the Null Hypothesis... Therefore we REJECT the Null hypothesis and conclude the coin is biased! We both failed to reject and rejected the null hypothesis based on the same data... WHICH IS IT?

29 NHST We both failed to reject and rejected the null hypothesis based on the same data... WHICH IS IT? in order to know which interpretaion is correct, we need to know the intenions of the experimenter... WHAT?!?

30 Bayes The Bayesian interpretaion of this data is a bit simpler. The likelihood is: much simpler, this depends only on the number of flips and number of trials, no explicit assumpions about intenions...

31 Bayes Plus, for Bayesian Analyses we can consider prior beliefs: p() Prior beta( 11,11) p() Prior Strongish fair prior beta( 11,11) beta( 2,20) p() Prior beta( 2,20) p(d ) 0.0e e e e 07 p( D) Likelihood Data: z=8,n=26 Posterior 95% HDI beta( 19,29) p(d)=2.93e 08 p(d ) 0.0e e e e 07 p( D) Likelihood Data: z=8,n=26 Posterior beta( 19,29) biased beta( 10,38) prior 95% HDI 95% HDI Strong p(d)=2.93e 08 p(d)=8.11e 09 p(d ) 0.0e e e e 07 p( D) Likelihood Data: z=8,n=26 95% HDI Posterior beta( 10,38) p(d)=8.11e 09 And in either case, we can say how probable the coin is to be fair (or any other value..)

32 Highest Density Interval (HDI) Of Course an HDI will not always be so simple as on a Normal distribution, and might not be Unimodal... Defined as the interval of possibility space that accounts for x (usually 95)% of the probability mass. 95% HDI x = % HDI x = 0.4

33 Bayes Plus, for Bayesian Analyses we can consider prior beliefs: p() Prior beta( 11,11) p() Prior Strongish fair prior beta( 11,11) beta( 2,20) p() Prior beta( 2,20) p(d ) 0.0e e e e 07 p( D) Likelihood Data: z=8,n=26 Posterior 95% HDI beta( 19,29) p(d)=2.93e 08 p(d ) 0.0e e e e 07 p( D) Likelihood Data: z=8,n=26 Posterior beta( 19,29) biased beta( 10,38) prior 95% HDI 95% HDI Strong p(d)=2.93e 08 p(d)=8.11e 09 p(d ) 0.0e e e e 07 p( D) Likelihood Data: z=8,n=26 95% HDI Posterior beta( 10,38) p(d)=8.11e 09 And in either case, we can say how probable the coin is to be fair (or any other value..)

34 Here be Dragons: Dangers of Bayesian StaIsIcs The Prior DistribuIon...

35 Bayes Plus, for Bayesian Analyses we can consider prior beliefs: p() Prior beta( 11,11) p() Prior Strongish fair prior beta( 11,11) beta( 2,20) p() Prior beta( 2,20) p(d ) 0.0e e e e 07 p( D) Likelihood Data: z=8,n=26 Posterior 95% HDI beta( 19,29) p(d)=2.93e 08 p(d ) 0.0e e e e 07 p( D) Likelihood Data: z=8,n=26 Posterior beta( 19,29) biased beta( 10,38) prior 95% HDI 95% HDI Strong p(d)=2.93e 08 p(d)=8.11e 09 p(d ) 0.0e e e e 07 p( D) Likelihood Data: z=8,n=26 95% HDI Posterior beta( 10,38) p(d)=8.11e 09 And in either case, we can say how probable the coin is to be fair (or any other value..)

36 NHST vs. Bayes Another thorny issue that comes up A LOT: MulIple Comparisons! How many NHSTs can you run on a set of data without changing your p-value?

37 NHST vs. Bayes MulIple Comparisons! Per Comparison false alarm rate: ExperimentWise false alarm rate: (where c is number of comparisons) > so to get <.05, we have to lower SOMETIMES A LOT

38 NHST vs. Bayes MulIple Comparisons! There are a lot of correcions for muliple comparisons, most notably Bonferonni correcion: Some actually depend on whether the comparison was planned in advance or not (?!?!) But in ALL cases, one has to trade off power for exhausive checking of the space...

39 NHST vs. Bayes MulIple Comparisons! There are a lot of correcions for muliple comparisons, most notably Bonferonni correcion: Well, unless you use Bayesian Methods Some actually depend on whether the comparison was planned in advance or not (?!?!) But in ALL cases, one has to trade off power for exhausive checking of the space...

40 NHST vs. Bayes MulIple Comparisons! In a Bayesian analysis, you calculate the enire posterior from the data you have and the prior, then you can examine the posterior however you d like. NO correcions for MulIple Comparisons Necessary! Bayesian methods correct for false posiives by incorporaing a reasonable Prior, and examining the full parameter space.

41 NHST vs. Bayes MulIple Comparisons! In a Bayesian analysis, you calculate the enire posterior from the data you have and the prior, then you can examine the posterior however you d like. NO correcions for MulIple Comparisons Necessary! Bayesian methods correct for false posiives by incorporaing a reasonable Prior, and examining the full parameter space.

42 NHST vs. Bayes So, why have NHST methods been so popular for so long? Maybe because instead of these:

43 NHST vs. Bayes So, why have NHST methods been so popular for so long? We had these:

44 NHST vs. Bayes So, why have NHST methods been so popular for so long? For all their hidden assumpions, reliance on intenions, problems with muliple comparisons, and other issues, Most of the math necessary for a T-test had been done. Would you like to try Gibbs sampling by hand?

45 NHST vs. Bayes Fortunately, The speed of modern computers has made the world a safe and producive place for Bayesians everywhere...

46 PracIcal Bayes Ok, so all this is great, but what do we actually DO with Bayes? For many experiments, frequenist stats are appropriate and all that is necessary And full Bayesian models can be tricky to set up and fully implement. If you ARE interested in learning how to employ those, READ THE BOOK! (and/or talk to me about it)

47 PracIcal Bayes But more importantly, Bayesian principles can help us understand what kind of quesions we ARE asking and what kind of quesions we COULD (and maybe should) ask. Neuroscience example: Decoding!

48 hrps://en.wikipedia.org/wiki/place_cell (Bayesian) Decoding Place Cells in Hippocampus These cells fire when the animal is in a certain spaial locaion. What is the data we have here? Spatial Tuning Curve cell Spatial Tuning Curve cell Tuning curves for HPC cells P(FR(c) Pos)

49 (Bayesian) Decoding Place Cells in Hippocampus Spatial Tuning Curve cell What is more interesing than P(FR(c) Pos)? How about P(Pos FR(c) )? 0 P(Pos FR(c) ) = P(FR(c) Pos) * P(Pos) P(FR(c))

50 (Bayesian) Decoding Place Cells in Hippocampus Decoded vs. Actual Position, HPC PFC Decoded vs. Actual Position, HPC Decoded Position F2 F1 CP 0.35 So, we can decode 0.3 F2 the animal s posiion 0.25 F1 0.2 from HPC what else P(Pos FR(c) ) = P(FR(c) 0.15 Pos) can * we P(Pos) CP look at? 0.1 P(FR(c)) 0.05 Decoded Position CP F1 F2 Actual Position CP F1 F2 Actual Position 0.02

51 (Bayesian) Decoding Place Cells in Hippocampus P(Pos FR(c) ) = P(FR(c) Pos) * P(Pos) P(FR(c)) What about P(Pos)? What should we use for our Prior? As usual, it depends on the quesion you want to ask

52 Anyhoo, I could keep going, but we re probably about out of Ime QuesIons? Or is it just Ime for Pizza?

CS 361: Probability & Statistics

CS 361: Probability & Statistics October 17, 2017 CS 361: Probability & Statistics Inference Maximum likelihood: drawbacks A couple of things might trip up max likelihood estimation: 1) Finding the maximum of some functions can be quite

More information

CS 361: Probability & Statistics

CS 361: Probability & Statistics March 14, 2018 CS 361: Probability & Statistics Inference The prior From Bayes rule, we know that we can express our function of interest as Likelihood Prior Posterior The right hand side contains the

More information

Bayes Formula. MATH 107: Finite Mathematics University of Louisville. March 26, 2014

Bayes Formula. MATH 107: Finite Mathematics University of Louisville. March 26, 2014 Bayes Formula MATH 07: Finite Mathematics University of Louisville March 26, 204 Test Accuracy Conditional reversal 2 / 5 A motivating question A rare disease occurs in out of every 0,000 people. A test

More information

Frequentist Statistics and Hypothesis Testing Spring

Frequentist Statistics and Hypothesis Testing Spring Frequentist Statistics and Hypothesis Testing 18.05 Spring 2018 http://xkcd.com/539/ Agenda Introduction to the frequentist way of life. What is a statistic? NHST ingredients; rejection regions Simple

More information

Comparison of Bayesian and Frequentist Inference

Comparison of Bayesian and Frequentist Inference Comparison of Bayesian and Frequentist Inference 18.05 Spring 2014 First discuss last class 19 board question, January 1, 2017 1 /10 Compare Bayesian inference Uses priors Logically impeccable Probabilities

More information

Quantitative Understanding in Biology 1.7 Bayesian Methods

Quantitative Understanding in Biology 1.7 Bayesian Methods Quantitative Understanding in Biology 1.7 Bayesian Methods Jason Banfelder October 25th, 2018 1 Introduction So far, most of the methods we ve looked at fall under the heading of classical, or frequentist

More information

Discrete Binary Distributions

Discrete Binary Distributions Discrete Binary Distributions Carl Edward Rasmussen November th, 26 Carl Edward Rasmussen Discrete Binary Distributions November th, 26 / 5 Key concepts Bernoulli: probabilities over binary variables Binomial:

More information

Computational Perception. Bayesian Inference

Computational Perception. Bayesian Inference Computational Perception 15-485/785 January 24, 2008 Bayesian Inference The process of probabilistic inference 1. define model of problem 2. derive posterior distributions and estimators 3. estimate parameters

More information

Hypothesis tests

Hypothesis tests 6.1 6.4 Hypothesis tests Prof. Tesler Math 186 February 26, 2014 Prof. Tesler 6.1 6.4 Hypothesis tests Math 186 / February 26, 2014 1 / 41 6.1 6.2 Intro to hypothesis tests and decision rules Hypothesis

More information

P (E) = P (A 1 )P (A 2 )... P (A n ).

P (E) = P (A 1 )P (A 2 )... P (A n ). Lecture 9: Conditional probability II: breaking complex events into smaller events, methods to solve probability problems, Bayes rule, law of total probability, Bayes theorem Discrete Structures II (Summer

More information

Bayesian RL Seminar. Chris Mansley September 9, 2008

Bayesian RL Seminar. Chris Mansley September 9, 2008 Bayesian RL Seminar Chris Mansley September 9, 2008 Bayes Basic Probability One of the basic principles of probability theory, the chain rule, will allow us to derive most of the background material in

More information

CPSC 340: Machine Learning and Data Mining

CPSC 340: Machine Learning and Data Mining CPSC 340: Machine Learning and Data Mining MLE and MAP Original version of these slides by Mark Schmidt, with modifications by Mike Gelbart. 1 Admin Assignment 4: Due tonight. Assignment 5: Will be released

More information

6.041SC Probabilistic Systems Analysis and Applied Probability, Fall 2013 Transcript Tutorial:A Random Number of Coin Flips

6.041SC Probabilistic Systems Analysis and Applied Probability, Fall 2013 Transcript Tutorial:A Random Number of Coin Flips 6.041SC Probabilistic Systems Analysis and Applied Probability, Fall 2013 Transcript Tutorial:A Random Number of Coin Flips Hey, everyone. Welcome back. Today, we're going to do another fun problem that

More information

Chapter Three. Hypothesis Testing

Chapter Three. Hypothesis Testing 3.1 Introduction The final phase of analyzing data is to make a decision concerning a set of choices or options. Should I invest in stocks or bonds? Should a new product be marketed? Are my products being

More information

1 A simple example. A short introduction to Bayesian statistics, part I Math 217 Probability and Statistics Prof. D.

1 A simple example. A short introduction to Bayesian statistics, part I Math 217 Probability and Statistics Prof. D. probabilities, we ll use Bayes formula. We can easily compute the reverse probabilities A short introduction to Bayesian statistics, part I Math 17 Probability and Statistics Prof. D. Joyce, Fall 014 I

More information

Exam 2 Practice Questions, 18.05, Spring 2014

Exam 2 Practice Questions, 18.05, Spring 2014 Exam 2 Practice Questions, 18.05, Spring 2014 Note: This is a set of practice problems for exam 2. The actual exam will be much shorter. Within each section we ve arranged the problems roughly in order

More information

CS 124 Math Review Section January 29, 2018

CS 124 Math Review Section January 29, 2018 CS 124 Math Review Section CS 124 is more math intensive than most of the introductory courses in the department. You re going to need to be able to do two things: 1. Perform some clever calculations to

More information

Introductory Econometrics. Review of statistics (Part II: Inference)

Introductory Econometrics. Review of statistics (Part II: Inference) Introductory Econometrics Review of statistics (Part II: Inference) Jun Ma School of Economics Renmin University of China October 1, 2018 1/16 Null and alternative hypotheses Usually, we have two competing

More information

Bayesian Updating with Discrete Priors Class 11, Jeremy Orloff and Jonathan Bloom

Bayesian Updating with Discrete Priors Class 11, Jeremy Orloff and Jonathan Bloom 1 Learning Goals ian Updating with Discrete Priors Class 11, 18.05 Jeremy Orloff and Jonathan Bloom 1. Be able to apply theorem to compute probabilities. 2. Be able to define the and to identify the roles

More information

Module 03 Lecture 14 Inferential Statistics ANOVA and TOI

Module 03 Lecture 14 Inferential Statistics ANOVA and TOI Introduction of Data Analytics Prof. Nandan Sudarsanam and Prof. B Ravindran Department of Management Studies and Department of Computer Science and Engineering Indian Institute of Technology, Madras Module

More information

Bayesian data analysis using JASP

Bayesian data analysis using JASP Bayesian data analysis using JASP Dani Navarro compcogscisydney.com/jasp-tute.html Part 1: Theory Philosophy of probability Introducing Bayes rule Bayesian reasoning A simple example Bayesian hypothesis

More information

Introduction to Bayesian Statistics

Introduction to Bayesian Statistics Bayesian Parameter Estimation Introduction to Bayesian Statistics Harvey Thornburg Center for Computer Research in Music and Acoustics (CCRMA) Department of Music, Stanford University Stanford, California

More information

Ling 289 Contingency Table Statistics

Ling 289 Contingency Table Statistics Ling 289 Contingency Table Statistics Roger Levy and Christopher Manning This is a summary of the material that we ve covered on contingency tables. Contingency tables: introduction Odds ratios Counting,

More information

DS-GA 1003: Machine Learning and Computational Statistics Homework 7: Bayesian Modeling

DS-GA 1003: Machine Learning and Computational Statistics Homework 7: Bayesian Modeling DS-GA 1003: Machine Learning and Computational Statistics Homework 7: Bayesian Modeling Due: Tuesday, May 10, 2016, at 6pm (Submit via NYU Classes) Instructions: Your answers to the questions below, including

More information

Chapter 1 Review of Equations and Inequalities

Chapter 1 Review of Equations and Inequalities Chapter 1 Review of Equations and Inequalities Part I Review of Basic Equations Recall that an equation is an expression with an equal sign in the middle. Also recall that, if a question asks you to solve

More information

MATH MW Elementary Probability Course Notes Part I: Models and Counting

MATH MW Elementary Probability Course Notes Part I: Models and Counting MATH 2030 3.00MW Elementary Probability Course Notes Part I: Models and Counting Tom Salisbury salt@yorku.ca York University Winter 2010 Introduction [Jan 5] Probability: the mathematics used for Statistics

More information

Reading for Lecture 6 Release v10

Reading for Lecture 6 Release v10 Reading for Lecture 6 Release v10 Christopher Lee October 11, 2011 Contents 1 The Basics ii 1.1 What is a Hypothesis Test?........................................ ii Example..................................................

More information

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14 CS 70 Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14 Introduction One of the key properties of coin flips is independence: if you flip a fair coin ten times and get ten

More information

Discrete Probability and State Estimation

Discrete Probability and State Estimation 6.01, Fall Semester, 2007 Lecture 12 Notes 1 MASSACHVSETTS INSTITVTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science 6.01 Introduction to EECS I Fall Semester, 2007 Lecture 12 Notes

More information

Lecture 4: Parameter Es/ma/on and Confidence Intervals. GENOME 560, Spring 2015 Doug Fowler, GS

Lecture 4: Parameter Es/ma/on and Confidence Intervals. GENOME 560, Spring 2015 Doug Fowler, GS Lecture 4: Parameter Es/ma/on and Confidence Intervals GENOME 560, Spring 2015 Doug Fowler, GS (dfowler@uw.edu) 1 Review: Probability DistribuIons Discrete: Binomial distribuion Hypergeometric distribuion

More information

Introduction: MLE, MAP, Bayesian reasoning (28/8/13)

Introduction: MLE, MAP, Bayesian reasoning (28/8/13) STA561: Probabilistic machine learning Introduction: MLE, MAP, Bayesian reasoning (28/8/13) Lecturer: Barbara Engelhardt Scribes: K. Ulrich, J. Subramanian, N. Raval, J. O Hollaren 1 Classifiers In this

More information

Introduction to Bayesian Learning. Machine Learning Fall 2018

Introduction to Bayesian Learning. Machine Learning Fall 2018 Introduction to Bayesian Learning Machine Learning Fall 2018 1 What we have seen so far What does it mean to learn? Mistake-driven learning Learning by counting (and bounding) number of mistakes PAC learnability

More information

Compute f(x θ)f(θ) dθ

Compute f(x θ)f(θ) dθ Bayesian Updating: Continuous Priors 18.05 Spring 2014 b a Compute f(x θ)f(θ) dθ January 1, 2017 1 /26 Beta distribution Beta(a, b) has density (a + b 1)! f (θ) = θ a 1 (1 θ) b 1 (a 1)!(b 1)! http://mathlets.org/mathlets/beta-distribution/

More information

Machine Learning CMPT 726 Simon Fraser University. Binomial Parameter Estimation

Machine Learning CMPT 726 Simon Fraser University. Binomial Parameter Estimation Machine Learning CMPT 726 Simon Fraser University Binomial Parameter Estimation Outline Maximum Likelihood Estimation Smoothed Frequencies, Laplace Correction. Bayesian Approach. Conjugate Prior. Uniform

More information

ORF 245 Fundamentals of Statistics Chapter 9 Hypothesis Testing

ORF 245 Fundamentals of Statistics Chapter 9 Hypothesis Testing ORF 245 Fundamentals of Statistics Chapter 9 Hypothesis Testing Robert Vanderbei Fall 2014 Slides last edited on November 24, 2014 http://www.princeton.edu/ rvdb Coin Tossing Example Consider two coins.

More information

PHP2510: Principles of Biostatistics & Data Analysis. Lecture X: Hypothesis testing. PHP 2510 Lec 10: Hypothesis testing 1

PHP2510: Principles of Biostatistics & Data Analysis. Lecture X: Hypothesis testing. PHP 2510 Lec 10: Hypothesis testing 1 PHP2510: Principles of Biostatistics & Data Analysis Lecture X: Hypothesis testing PHP 2510 Lec 10: Hypothesis testing 1 In previous lectures we have encountered problems of estimating an unknown population

More information

Bayesian Inference for Normal Mean

Bayesian Inference for Normal Mean Al Nosedal. University of Toronto. November 18, 2015 Likelihood of Single Observation The conditional observation distribution of y µ is Normal with mean µ and variance σ 2, which is known. Its density

More information

Bayesian Models in Machine Learning

Bayesian Models in Machine Learning Bayesian Models in Machine Learning Lukáš Burget Escuela de Ciencias Informáticas 2017 Buenos Aires, July 24-29 2017 Frequentist vs. Bayesian Frequentist point of view: Probability is the frequency of

More information

Introduc)on to Bayesian methods (con)nued) - Lecture 16

Introduc)on to Bayesian methods (con)nued) - Lecture 16 Introduc)on to Bayesian methods (con)nued) - Lecture 16 David Sontag New York University Slides adapted from Luke Zettlemoyer, Carlos Guestrin, Dan Klein, and Vibhav Gogate Outline of lectures Review of

More information

Which coin will I use? Which coin will I use? Logistics. Statistical Learning. Topics. 573 Topics. Coin Flip. Coin Flip

Which coin will I use? Which coin will I use? Logistics. Statistical Learning. Topics. 573 Topics. Coin Flip. Coin Flip Logistics Statistical Learning CSE 57 Team Meetings Midterm Open book, notes Studying See AIMA exercises Daniel S. Weld Daniel S. Weld Supervised Learning 57 Topics Logic-Based Reinforcement Learning Planning

More information

MITOCW watch?v=vjzv6wjttnc

MITOCW watch?v=vjzv6wjttnc MITOCW watch?v=vjzv6wjttnc PROFESSOR: We just saw some random variables come up in the bigger number game. And we're going to be talking now about random variables, just formally what they are and their

More information

( )( b + c) = ab + ac, but it can also be ( )( a) = ba + ca. Let s use the distributive property on a couple of

( )( b + c) = ab + ac, but it can also be ( )( a) = ba + ca. Let s use the distributive property on a couple of Factoring Review for Algebra II The saddest thing about not doing well in Algebra II is that almost any math teacher can tell you going into it what s going to trip you up. One of the first things they

More information

Computational Cognitive Science

Computational Cognitive Science Computational Cognitive Science Lecture 9: A Bayesian model of concept learning Chris Lucas School of Informatics University of Edinburgh October 16, 218 Reading Rules and Similarity in Concept Learning

More information

Hypothesis Testing. ) the hypothesis that suggests no change from previous experience

Hypothesis Testing. ) the hypothesis that suggests no change from previous experience Hypothesis Testing Definitions Hypothesis a claim about something Null hypothesis ( H 0 ) the hypothesis that suggests no change from previous experience Alternative hypothesis ( H 1 ) the hypothesis that

More information

It applies to discrete and continuous random variables, and a mix of the two.

It applies to discrete and continuous random variables, and a mix of the two. 3. Bayes Theorem 3.1 Bayes Theorem A straightforward application of conditioning: using p(x, y) = p(x y) p(y) = p(y x) p(x), we obtain Bayes theorem (also called Bayes rule) p(x y) = p(y x) p(x). p(y)

More information

Chapter 26: Comparing Counts (Chi Square)

Chapter 26: Comparing Counts (Chi Square) Chapter 6: Comparing Counts (Chi Square) We ve seen that you can turn a qualitative variable into a quantitative one (by counting the number of successes and failures), but that s a compromise it forces

More information

CSC321 Lecture 18: Learning Probabilistic Models

CSC321 Lecture 18: Learning Probabilistic Models CSC321 Lecture 18: Learning Probabilistic Models Roger Grosse Roger Grosse CSC321 Lecture 18: Learning Probabilistic Models 1 / 25 Overview So far in this course: mainly supervised learning Language modeling

More information

Solution to Proof Questions from September 1st

Solution to Proof Questions from September 1st Solution to Proof Questions from September 1st Olena Bormashenko September 4, 2011 What is a proof? A proof is an airtight logical argument that proves a certain statement in general. In a sense, it s

More information

Some slides from Carlos Guestrin, Luke Zettlemoyer & K Gajos 2

Some slides from Carlos Guestrin, Luke Zettlemoyer & K Gajos 2 Logistics CSE 446: Point Estimation Winter 2012 PS2 out shortly Dan Weld Some slides from Carlos Guestrin, Luke Zettlemoyer & K Gajos 2 Last Time Random variables, distributions Marginal, joint & conditional

More information

Statistical testing. Samantha Kleinberg. October 20, 2009

Statistical testing. Samantha Kleinberg. October 20, 2009 October 20, 2009 Intro to significance testing Significance testing and bioinformatics Gene expression: Frequently have microarray data for some group of subjects with/without the disease. Want to find

More information

Discrete Mathematics and Probability Theory Spring 2014 Anant Sahai Note 10

Discrete Mathematics and Probability Theory Spring 2014 Anant Sahai Note 10 EECS 70 Discrete Mathematics and Probability Theory Spring 2014 Anant Sahai Note 10 Introduction to Basic Discrete Probability In the last note we considered the probabilistic experiment where we flipped

More information

Using Probability to do Statistics.

Using Probability to do Statistics. Al Nosedal. University of Toronto. November 5, 2015 Milk and honey and hemoglobin Animal experiments suggested that honey in a diet might raise hemoglobin level. A researcher designed a study involving

More information

Why should you care?? Intellectual curiosity. Gambling. Mathematically the same as the ESP decision problem we discussed in Week 4.

Why should you care?? Intellectual curiosity. Gambling. Mathematically the same as the ESP decision problem we discussed in Week 4. I. Probability basics (Sections 4.1 and 4.2) Flip a fair (probability of HEADS is 1/2) coin ten times. What is the probability of getting exactly 5 HEADS? What is the probability of getting exactly 10

More information

Bayesian Analysis. Justin Chin. Spring 2018

Bayesian Analysis. Justin Chin. Spring 2018 Bayesian Analysis Justin Chin Spring 2018 Abstract We often think of the field of Statistics simply as data collection and analysis. While the essence of Statistics lies in numeric analysis of observed

More information

Conditional probabilities and graphical models

Conditional probabilities and graphical models Conditional probabilities and graphical models Thomas Mailund Bioinformatics Research Centre (BiRC), Aarhus University Probability theory allows us to describe uncertainty in the processes we model within

More information

Bayesian Updating with Continuous Priors Class 13, Jeremy Orloff and Jonathan Bloom

Bayesian Updating with Continuous Priors Class 13, Jeremy Orloff and Jonathan Bloom Bayesian Updating with Continuous Priors Class 3, 8.05 Jeremy Orloff and Jonathan Bloom Learning Goals. Understand a parameterized family of distributions as representing a continuous range of hypotheses

More information

CPSC 340: Machine Learning and Data Mining. MLE and MAP Fall 2017

CPSC 340: Machine Learning and Data Mining. MLE and MAP Fall 2017 CPSC 340: Machine Learning and Data Mining MLE and MAP Fall 2017 Assignment 3: Admin 1 late day to hand in tonight, 2 late days for Wednesday. Assignment 4: Due Friday of next week. Last Time: Multi-Class

More information

What are the Findings?

What are the Findings? What are the Findings? James B. Rawlings Department of Chemical and Biological Engineering University of Wisconsin Madison Madison, Wisconsin April 2010 Rawlings (Wisconsin) Stating the findings 1 / 33

More information

Physics 403. Segev BenZvi. Choosing Priors and the Principle of Maximum Entropy. Department of Physics and Astronomy University of Rochester

Physics 403. Segev BenZvi. Choosing Priors and the Principle of Maximum Entropy. Department of Physics and Astronomy University of Rochester Physics 403 Choosing Priors and the Principle of Maximum Entropy Segev BenZvi Department of Physics and Astronomy University of Rochester Table of Contents 1 Review of Last Class Odds Ratio Occam Factors

More information

Today. Statistical Learning. Coin Flip. Coin Flip. Experiment 1: Heads. Experiment 1: Heads. Which coin will I use? Which coin will I use?

Today. Statistical Learning. Coin Flip. Coin Flip. Experiment 1: Heads. Experiment 1: Heads. Which coin will I use? Which coin will I use? Today Statistical Learning Parameter Estimation: Maximum Likelihood (ML) Maximum A Posteriori (MAP) Bayesian Continuous case Learning Parameters for a Bayesian Network Naive Bayes Maximum Likelihood estimates

More information

1 Probabilities. 1.1 Basics 1 PROBABILITIES

1 Probabilities. 1.1 Basics 1 PROBABILITIES 1 PROBABILITIES 1 Probabilities Probability is a tricky word usually meaning the likelyhood of something occuring or how frequent something is. Obviously, if something happens frequently, then its probability

More information

Bayesian Inference. STA 121: Regression Analysis Artin Armagan

Bayesian Inference. STA 121: Regression Analysis Artin Armagan Bayesian Inference STA 121: Regression Analysis Artin Armagan Bayes Rule...s! Reverend Thomas Bayes Posterior Prior p(θ y) = p(y θ)p(θ)/p(y) Likelihood - Sampling Distribution Normalizing Constant: p(y

More information

Ways to make neural networks generalize better

Ways to make neural networks generalize better Ways to make neural networks generalize better Seminar in Deep Learning University of Tartu 04 / 10 / 2014 Pihel Saatmann Topics Overview of ways to improve generalization Limiting the size of the weights

More information

CS 188: Artificial Intelligence Spring Today

CS 188: Artificial Intelligence Spring Today CS 188: Artificial Intelligence Spring 2006 Lecture 9: Naïve Bayes 2/14/2006 Dan Klein UC Berkeley Many slides from either Stuart Russell or Andrew Moore Bayes rule Today Expectations and utilities Naïve

More information

MAT Mathematics in Today's World

MAT Mathematics in Today's World MAT 1000 Mathematics in Today's World Last Time We discussed the four rules that govern probabilities: 1. Probabilities are numbers between 0 and 1 2. The probability an event does not occur is 1 minus

More information

Uni- and Bivariate Power

Uni- and Bivariate Power Uni- and Bivariate Power Copyright 2002, 2014, J. Toby Mordkoff Note that the relationship between risk and power is unidirectional. Power depends on risk, but risk is completely independent of power.

More information

STOR 435 Lecture 5. Conditional Probability and Independence - I

STOR 435 Lecture 5. Conditional Probability and Independence - I STOR 435 Lecture 5 Conditional Probability and Independence - I Jan Hannig UNC Chapel Hill 1 / 16 Motivation Basic point Think of probability as the amount of belief we have in a particular outcome. If

More information

Hypothesis testing I. - In particular, we are talking about statistical hypotheses. [get everyone s finger length!] n =

Hypothesis testing I. - In particular, we are talking about statistical hypotheses. [get everyone s finger length!] n = Hypothesis testing I I. What is hypothesis testing? [Note we re temporarily bouncing around in the book a lot! Things will settle down again in a week or so] - Exactly what it says. We develop a hypothesis,

More information

Announcements. Proposals graded

Announcements. Proposals graded Announcements Proposals graded Kevin Jamieson 2018 1 Hypothesis testing Machine Learning CSE546 Kevin Jamieson University of Washington October 30, 2018 2018 Kevin Jamieson 2 Anomaly detection You are

More information

COMP 551 Applied Machine Learning Lecture 19: Bayesian Inference

COMP 551 Applied Machine Learning Lecture 19: Bayesian Inference COMP 551 Applied Machine Learning Lecture 19: Bayesian Inference Associate Instructor: (herke.vanhoof@mcgill.ca) Class web page: www.cs.mcgill.ca/~jpineau/comp551 Unless otherwise noted, all material posted

More information

BEGINNING BAYES IN R. Bayes with discrete models

BEGINNING BAYES IN R. Bayes with discrete models BEGINNING BAYES IN R Bayes with discrete models Beginning Bayes in R Survey on eating out What is your favorite day for eating out? Construct a prior for p Define p: proportion of all students who answer

More information

With Question/Answer Animations. Chapter 7

With Question/Answer Animations. Chapter 7 With Question/Answer Animations Chapter 7 Chapter Summary Introduction to Discrete Probability Probability Theory Bayes Theorem Section 7.1 Section Summary Finite Probability Probabilities of Complements

More information

STAT 285 Fall Assignment 1 Solutions

STAT 285 Fall Assignment 1 Solutions STAT 285 Fall 2014 Assignment 1 Solutions 1. An environmental agency sets a standard of 200 ppb for the concentration of cadmium in a lake. The concentration of cadmium in one lake is measured 17 times.

More information

CENTRAL LIMIT THEOREM (CLT)

CENTRAL LIMIT THEOREM (CLT) CENTRAL LIMIT THEOREM (CLT) A sampling distribution is the probability distribution of the sample statistic that is formed when samples of size n are repeatedly taken from a population. If the sample statistic

More information

Introduc)on to Bayesian Methods

Introduc)on to Bayesian Methods Introduc)on to Bayesian Methods Bayes Rule py x)px) = px! y) = px y)py) py x) = px y)py) px) px) =! px! y) = px y)py) y py x) = py x) =! y "! y px y)py) px y)py) px y)py) px y)py)dy Bayes Rule py x) =

More information

Data Analysis and Monte Carlo Methods

Data Analysis and Monte Carlo Methods Lecturer: Allen Caldwell, Max Planck Institute for Physics & TUM Recitation Instructor: Oleksander (Alex) Volynets, MPP & TUM General Information: - Lectures will be held in English, Mondays 16-18:00 -

More information

Lecture 9: Naive Bayes, SVM, Kernels. Saravanan Thirumuruganathan

Lecture 9: Naive Bayes, SVM, Kernels. Saravanan Thirumuruganathan Lecture 9: Naive Bayes, SVM, Kernels Instructor: Outline 1 Probability basics 2 Probabilistic Interpretation of Classification 3 Bayesian Classifiers, Naive Bayes 4 Support Vector Machines Probability

More information

You separate binary numbers into columns in a similar fashion. 2 5 = 32

You separate binary numbers into columns in a similar fashion. 2 5 = 32 RSA Encryption 2 At the end of Part I of this article, we stated that RSA encryption works because it s impractical to factor n, which determines P 1 and P 2, which determines our private key, d, which

More information

PIER HLM Course July 30, 2011 Howard Seltman. Discussion Guide for Bayes and BUGS

PIER HLM Course July 30, 2011 Howard Seltman. Discussion Guide for Bayes and BUGS PIER HLM Course July 30, 2011 Howard Seltman Discussion Guide for Bayes and BUGS 1. Classical Statistics is based on parameters as fixed unknown values. a. The standard approach is to try to discover,

More information

Parameter Learning With Binary Variables

Parameter Learning With Binary Variables With Binary Variables University of Nebraska Lincoln CSCE 970 Pattern Recognition Outline Outline 1 Learning a Single Parameter 2 More on the Beta Density Function 3 Computing a Probability Interval Outline

More information

MITOCW ocw f99-lec30_300k

MITOCW ocw f99-lec30_300k MITOCW ocw-18.06-f99-lec30_300k OK, this is the lecture on linear transformations. Actually, linear algebra courses used to begin with this lecture, so you could say I'm beginning this course again by

More information

One sided tests. An example of a two sided alternative is what we ve been using for our two sample tests:

One sided tests. An example of a two sided alternative is what we ve been using for our two sample tests: One sided tests So far all of our tests have been two sided. While this may be a bit easier to understand, this is often not the best way to do a hypothesis test. One simple thing that we can do to get

More information

FERMAT S TEST KEITH CONRAD

FERMAT S TEST KEITH CONRAD FERMAT S TEST KEITH CONRAD 1. Introduction Fermat s little theorem says for prime p that a p 1 1 mod p for all a 0 mod p. A naive extension of this to a composite modulus n 2 would be: for all a 0 mod

More information

HST.582J / 6.555J / J Biomedical Signal and Image Processing Spring 2007

HST.582J / 6.555J / J Biomedical Signal and Image Processing Spring 2007 MIT OpenCourseWare http://ocw.mit.edu HST.582J / 6.555J / 16.456J Biomedical Signal and Image Processing Spring 2007 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.

More information

Hypothesis testing. Data to decisions

Hypothesis testing. Data to decisions Hypothesis testing Data to decisions The idea Null hypothesis: H 0 : the DGP/population has property P Under the null, a sample statistic has a known distribution If, under that that distribution, the

More information

Modeling Environment

Modeling Environment Topic Model Modeling Environment What does it mean to understand/ your environment? Ability to predict Two approaches to ing environment of words and text Latent Semantic Analysis (LSA) Topic Model LSA

More information

Joint, Conditional, & Marginal Probabilities

Joint, Conditional, & Marginal Probabilities Joint, Conditional, & Marginal Probabilities The three axioms for probability don t discuss how to create probabilities for combined events such as P [A B] or for the likelihood of an event A given that

More information

Intermediate Math Circles November 15, 2017 Probability III

Intermediate Math Circles November 15, 2017 Probability III Intermediate Math Circles November 5, 07 Probability III Example : You have bins in which there are coloured balls. The balls are identical except for their colours. The contents of the containers are:

More information

ECE295, Data Assimila0on and Inverse Problems, Spring 2015

ECE295, Data Assimila0on and Inverse Problems, Spring 2015 ECE295, Data Assimila0on and Inverse Problems, Spring 2015 1 April, Intro; Linear discrete Inverse problems (Aster Ch 1 and 2) Slides 8 April, SVD (Aster ch 2 and 3) Slides 15 April, RegularizaFon (ch

More information

Probability Theory and Simulation Methods

Probability Theory and Simulation Methods Feb 28th, 2018 Lecture 10: Random variables Countdown to midterm (March 21st): 28 days Week 1 Chapter 1: Axioms of probability Week 2 Chapter 3: Conditional probability and independence Week 4 Chapters

More information

z and t tests for the mean of a normal distribution Confidence intervals for the mean Binomial tests

z and t tests for the mean of a normal distribution Confidence intervals for the mean Binomial tests z and t tests for the mean of a normal distribution Confidence intervals for the mean Binomial tests Chapters 3.5.1 3.5.2, 3.3.2 Prof. Tesler Math 283 Fall 2018 Prof. Tesler z and t tests for mean Math

More information

Solving Quadratic & Higher Degree Equations

Solving Quadratic & Higher Degree Equations Chapter 9 Solving Quadratic & Higher Degree Equations Sec 1. Zero Product Property Back in the third grade students were taught when they multiplied a number by zero, the product would be zero. In algebra,

More information

STAT 201 Chapter 5. Probability

STAT 201 Chapter 5. Probability STAT 201 Chapter 5 Probability 1 2 Introduction to Probability Probability The way we quantify uncertainty. Subjective Probability A probability derived from an individual's personal judgment about whether

More information

Basics of Proofs. 1 The Basics. 2 Proof Strategies. 2.1 Understand What s Going On

Basics of Proofs. 1 The Basics. 2 Proof Strategies. 2.1 Understand What s Going On Basics of Proofs The Putnam is a proof based exam and will expect you to write proofs in your solutions Similarly, Math 96 will also require you to write proofs in your homework solutions If you ve seen

More information

Discrete Probability and State Estimation

Discrete Probability and State Estimation 6.01, Spring Semester, 2008 Week 12 Course Notes 1 MASSACHVSETTS INSTITVTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science 6.01 Introduction to EECS I Spring Semester, 2008 Week

More information

Introduction to AI Learning Bayesian networks. Vibhav Gogate

Introduction to AI Learning Bayesian networks. Vibhav Gogate Introduction to AI Learning Bayesian networks Vibhav Gogate Inductive Learning in a nutshell Given: Data Examples of a function (X, F(X)) Predict function F(X) for new examples X Discrete F(X): Classification

More information

Standard & Conditional Probability

Standard & Conditional Probability Biostatistics 050 Standard & Conditional Probability 1 ORIGIN 0 Probability as a Concept: Standard & Conditional Probability "The probability of an event is the likelihood of that event expressed either

More information

CIS 2033 Lecture 5, Fall

CIS 2033 Lecture 5, Fall CIS 2033 Lecture 5, Fall 2016 1 Instructor: David Dobor September 13, 2016 1 Supplemental reading from Dekking s textbook: Chapter2, 3. We mentioned at the beginning of this class that calculus was a prerequisite

More information

Bayesian Inference. Introduction

Bayesian Inference. Introduction Bayesian Inference Introduction The frequentist approach to inference holds that probabilities are intrinsicially tied (unsurprisingly) to frequencies. This interpretation is actually quite natural. What,

More information