Why I think Bayesian Analysis is the greatest thing since NHST (bread)
|
|
- Osborn Brooks
- 5 years ago
- Views:
Transcription
1 Why I think Bayesian Analysis is the greatest thing since NHST (bread) Data Binge Nate Powell May 12, 2017 NHST = Null Hypothesis Significance TesIng
2 Textbook Doing Bayesian Data Analysis by John K. Kruschke Memory and the ComputaIonal Brain by Gallistel and King
3 What is Bayesian Analysis? hrp://xkcd.com/1132/
4 Advantages of Bayesian StaIsIcs vs. FrequenIst StaIsIcs Degree of belief is graded (not binary) Bayesian Update (incorporates new data) Can provide reasonable answers given limited data Can describe the probability of events that have not yet occurred
5 Advantages of Bayesian StaIsIcs vs. FrequenIst StaIsIcs Degree of belief is graded (not binary) what if your {X} test returns p =.06? or p =.055? a bayesian analysis will reflect this as a small change in posterior probability, not a complete shib between true and false. Bayesian Update (incorporates new data) Can provide reasonable answers given limited data Can describe the probability of events that have not yet occurred
6 Advantages of Bayesian StaIsIcs vs. FrequenIst StaIsIcs Degree of belief is graded (not binary) Bayesian Update (incorporates new data) if your p-value is 0.06, are you allowed to add new samples? A posterior distribuion can always be the prior for a bayesian update step if more informaion is collected. Can provide reasonable answers given limited data Can describe the probability of events that have not yet occurred
7 Advantages of Bayesian StaIsIcs vs. FrequenIst StaIsIcs Degree of belief is graded (not binary) Bayesian Update (incorporates new data) Can provide reasonable answers given limited data Limited data (which may not be significant by most tests) can be enough to modify or help confirm an untrusted prior, but won t disturb a trusted one... BUT HERE BE DRAGONS! Can describe the probability of events that have not yet occurred
8 Advantages of Bayesian StaIsIcs vs. FrequenIst StaIsIcs Degree of belief is graded (not binary) Bayesian Update (incorporates new data) Can provide reasonable answers given limited data Can describe the probability of events that have not yet occurred Bayesian methods were used to assess the likelihood of nuclear disasters for US policy (prior to any having occurred) with impressive accuracy.
9 Advantages of Bayesian StaIsIcs vs. Terminology, Conditional Probability FrequenIst StaIsIcs By the way, who s afraid of math? Who s afraid of equations? FrequenIst concern: P(D H) D = Data H = Hypothesis Bayesian Concern: P(ɵ D) or P(M D,ɵ) D = Data ɵ = model parameters M = Model hrp://xkcd.com/892/
10 Bayesian Update Learning from New experience to update previous beliefs P(M D) = P(D M) P(M) / P(D) Posterior Liklihood Prior Evidence This also allows us to incorporate new data without having to change any assumpions hrp://xkcd.com/1236/
11 Bayesian Update : twins example A physicist is having Twins, and wants to know the odds that her twins will be idenical (vs. fraternal). The Dr. told her that 2/3 of twin births are fraternal, and 1/3 are idenical. Also, a sonogram confirmed her twins are both male. (All idenical twins are the same sex, Fraternal can be either the same or mixed.) Efron, B. (2013). A 250-year argument: Belief, behavior, and the bootstrap. Bulle)n of the American Mathema)cal Society, 50(1),
12 Bayesian Update : twins example Prior DistribuIon: P(M = idenical) = 1/3 P(M = fraternal) = 2/3 AddiIonal Evidence: D = Same Sex P(D) = 2/3* Likelihood: P(same sex M = idenical) = 1 P(same sex M = fraternal) = 1/2 Posterior: P(M = idenical same sex) = 1/2 P(M = fraternal same sex) = 1/2 Posterior Prior DistribuIon: Efron, B. (2013). A 250-year argument: Belief, behavior, and the bootstrap. Bulle)n of the American Mathema)cal Society, 50(1), Identical Twins are: Fraternal Sonogram shows: Same sex 1/3 Physicist a Different 1/3 0 c 1/3 1/3 2/3 0 P(fraternal) P(idenIcal) b d Doctor
13 Bayesian Update : twins example But if you prefer a visual confirmaion: Sonogram shows: 1 Prior Same sex Different a b Identical Twins are: Fraternal 1/3 0 c 1/3 1/3 d 1/3 2/3 Doctor 1 0 P(frat) P(id) Posterior Physicist 0 P(frat) P(id) Efron, B. (2013). A 250-year argument: Belief, behavior, and the bootstrap. Bulle)n of the American Mathema)cal Society, 50(1),
14 Bayesian probability vs. FrequenIst probability FrequenIst Probability: The frequency with which events happen in a large number of trials Bayesian Probability: The Degree of belief you have that a certain event will happen (on any given trial) hrp://oikosjournal.wordpress.com/2011/10/11/frequenist-vs-bayesian-staisics-resources-to-helpyou-choose/
15 so, what, if anything, is wrong with NHST? Suppose we flip a coin N = 26 Imes and we come up with z = 8 heads is this coin biased or fair? Note, a bayesian would ask what is the probability that the coin is fair...
16 NHST N = 26 z = 8 is this coin biased? if the coin were fair we would expect N = 26 z = 13 The quesion we must answer is can we reject the null hypothesis, in this case, that the coin is fair and we saw these results by chance alone... If we repeated this experiment many Imes, would we expect to see z = 8 more than 5% of the Ime (with a fair coin?)
17 NHST The probability of seeing z heads in N flips is given by the binomial probability distribuion: where: If we repeated this experiment many Imes, would we expect to see z = 8 more than 5% of the Ime (with a fair coin?)
18 NHST If we repeated this experiment many Imes, would we expect to see z = 8 more than 5% of the Ime (with a fair coin?) The probability of seeing z heads in N flips is given by the binomial probability distribuion: Hypothetical Population Implied Sampling Distribution p(y) = 0.5 N = 26 p(z N,) tail y head z
19 NHST Hypothetical Population Implied Sampling Distribution p(y) tail y = 0.5 head N = 26 p(z N,) We want to know the probability of finding a result as low (or lower) than 8 heads out of 26 flips... This is the p-value of z = 8 for this sampling distribuion! z
20 NHST Hypothetical Population Implied Sampling Distribution p(y) tail y = 0.5 head N = 26 p(z N,) In this case we will consider the two-sided p-value (shaded) we can reject the null if z 7 or z 19 Thus we FAIL to reject the Null Hypothesis... z
21 NHST Thus we FAIL to reject the Null Hypothesis... What does that mean? Failing to reject the null hypothesis is like kissing your sister... we don t have an answer, but there s not a good reason to say the coin is biased.
22 NHST Thus we FAIL to reject the Null Hypothesis... But its more probable the coin is fair right? Do we know how likely the coin is to be fair? What if we had Really REALLY good reasons to believe it was fair?
23 NHST Thus we FAIL to reject the Null Hypothesis... But at least that s it right... I mean, it s an unsaisfying answer, but at least we know what it is... Unless...
24 NHST we used a sampling distribuion for N = 26 Hypothetical Population Implied Sampling Distribution p(y) tail y = 0.5 head N = 26 p(z N,) What if instead they sampled unil they found z = 8? We would have to use a DIFFERENT sampling distribuion... z
25 NHST Sampled coin unil we got z = 8, it took N = 26 flips... which gives us: Hypothetical Population Implied Sampling Distribution p(y) = 0.5 z = 8 p(n z,) tail y head N
26 NHST Sampled coin unil we got z = 8, it took N = 26 flips... Hypothetical Population Implied Sampling Distribution p(y) tail y = 0.5 head z = 8 p(n z,) N In this case, for p <.05, our values are N 9 and N 26 Therefore we REJECT the Null hypothesis and conclude the coin is biased!
27 NHST Therefore we REJECT the Null hypothesis and conclude the coin is biased! Now, do we have an expression of how likely the coin is to be biased? Or fair? And again, what if we were REALLY REALLY sure that the coin was fair?
28 NHST But we have an even bigger problem here! Thus we FAIL to reject the Null Hypothesis... Therefore we REJECT the Null hypothesis and conclude the coin is biased! We both failed to reject and rejected the null hypothesis based on the same data... WHICH IS IT?
29 NHST We both failed to reject and rejected the null hypothesis based on the same data... WHICH IS IT? in order to know which interpretaion is correct, we need to know the intenions of the experimenter... WHAT?!?
30 Bayes The Bayesian interpretaion of this data is a bit simpler. The likelihood is: much simpler, this depends only on the number of flips and number of trials, no explicit assumpions about intenions...
31 Bayes Plus, for Bayesian Analyses we can consider prior beliefs: p() Prior beta( 11,11) p() Prior Strongish fair prior beta( 11,11) beta( 2,20) p() Prior beta( 2,20) p(d ) 0.0e e e e 07 p( D) Likelihood Data: z=8,n=26 Posterior 95% HDI beta( 19,29) p(d)=2.93e 08 p(d ) 0.0e e e e 07 p( D) Likelihood Data: z=8,n=26 Posterior beta( 19,29) biased beta( 10,38) prior 95% HDI 95% HDI Strong p(d)=2.93e 08 p(d)=8.11e 09 p(d ) 0.0e e e e 07 p( D) Likelihood Data: z=8,n=26 95% HDI Posterior beta( 10,38) p(d)=8.11e 09 And in either case, we can say how probable the coin is to be fair (or any other value..)
32 Highest Density Interval (HDI) Of Course an HDI will not always be so simple as on a Normal distribution, and might not be Unimodal... Defined as the interval of possibility space that accounts for x (usually 95)% of the probability mass. 95% HDI x = % HDI x = 0.4
33 Bayes Plus, for Bayesian Analyses we can consider prior beliefs: p() Prior beta( 11,11) p() Prior Strongish fair prior beta( 11,11) beta( 2,20) p() Prior beta( 2,20) p(d ) 0.0e e e e 07 p( D) Likelihood Data: z=8,n=26 Posterior 95% HDI beta( 19,29) p(d)=2.93e 08 p(d ) 0.0e e e e 07 p( D) Likelihood Data: z=8,n=26 Posterior beta( 19,29) biased beta( 10,38) prior 95% HDI 95% HDI Strong p(d)=2.93e 08 p(d)=8.11e 09 p(d ) 0.0e e e e 07 p( D) Likelihood Data: z=8,n=26 95% HDI Posterior beta( 10,38) p(d)=8.11e 09 And in either case, we can say how probable the coin is to be fair (or any other value..)
34 Here be Dragons: Dangers of Bayesian StaIsIcs The Prior DistribuIon...
35 Bayes Plus, for Bayesian Analyses we can consider prior beliefs: p() Prior beta( 11,11) p() Prior Strongish fair prior beta( 11,11) beta( 2,20) p() Prior beta( 2,20) p(d ) 0.0e e e e 07 p( D) Likelihood Data: z=8,n=26 Posterior 95% HDI beta( 19,29) p(d)=2.93e 08 p(d ) 0.0e e e e 07 p( D) Likelihood Data: z=8,n=26 Posterior beta( 19,29) biased beta( 10,38) prior 95% HDI 95% HDI Strong p(d)=2.93e 08 p(d)=8.11e 09 p(d ) 0.0e e e e 07 p( D) Likelihood Data: z=8,n=26 95% HDI Posterior beta( 10,38) p(d)=8.11e 09 And in either case, we can say how probable the coin is to be fair (or any other value..)
36 NHST vs. Bayes Another thorny issue that comes up A LOT: MulIple Comparisons! How many NHSTs can you run on a set of data without changing your p-value?
37 NHST vs. Bayes MulIple Comparisons! Per Comparison false alarm rate: ExperimentWise false alarm rate: (where c is number of comparisons) > so to get <.05, we have to lower SOMETIMES A LOT
38 NHST vs. Bayes MulIple Comparisons! There are a lot of correcions for muliple comparisons, most notably Bonferonni correcion: Some actually depend on whether the comparison was planned in advance or not (?!?!) But in ALL cases, one has to trade off power for exhausive checking of the space...
39 NHST vs. Bayes MulIple Comparisons! There are a lot of correcions for muliple comparisons, most notably Bonferonni correcion: Well, unless you use Bayesian Methods Some actually depend on whether the comparison was planned in advance or not (?!?!) But in ALL cases, one has to trade off power for exhausive checking of the space...
40 NHST vs. Bayes MulIple Comparisons! In a Bayesian analysis, you calculate the enire posterior from the data you have and the prior, then you can examine the posterior however you d like. NO correcions for MulIple Comparisons Necessary! Bayesian methods correct for false posiives by incorporaing a reasonable Prior, and examining the full parameter space.
41 NHST vs. Bayes MulIple Comparisons! In a Bayesian analysis, you calculate the enire posterior from the data you have and the prior, then you can examine the posterior however you d like. NO correcions for MulIple Comparisons Necessary! Bayesian methods correct for false posiives by incorporaing a reasonable Prior, and examining the full parameter space.
42 NHST vs. Bayes So, why have NHST methods been so popular for so long? Maybe because instead of these:
43 NHST vs. Bayes So, why have NHST methods been so popular for so long? We had these:
44 NHST vs. Bayes So, why have NHST methods been so popular for so long? For all their hidden assumpions, reliance on intenions, problems with muliple comparisons, and other issues, Most of the math necessary for a T-test had been done. Would you like to try Gibbs sampling by hand?
45 NHST vs. Bayes Fortunately, The speed of modern computers has made the world a safe and producive place for Bayesians everywhere...
46 PracIcal Bayes Ok, so all this is great, but what do we actually DO with Bayes? For many experiments, frequenist stats are appropriate and all that is necessary And full Bayesian models can be tricky to set up and fully implement. If you ARE interested in learning how to employ those, READ THE BOOK! (and/or talk to me about it)
47 PracIcal Bayes But more importantly, Bayesian principles can help us understand what kind of quesions we ARE asking and what kind of quesions we COULD (and maybe should) ask. Neuroscience example: Decoding!
48 hrps://en.wikipedia.org/wiki/place_cell (Bayesian) Decoding Place Cells in Hippocampus These cells fire when the animal is in a certain spaial locaion. What is the data we have here? Spatial Tuning Curve cell Spatial Tuning Curve cell Tuning curves for HPC cells P(FR(c) Pos)
49 (Bayesian) Decoding Place Cells in Hippocampus Spatial Tuning Curve cell What is more interesing than P(FR(c) Pos)? How about P(Pos FR(c) )? 0 P(Pos FR(c) ) = P(FR(c) Pos) * P(Pos) P(FR(c))
50 (Bayesian) Decoding Place Cells in Hippocampus Decoded vs. Actual Position, HPC PFC Decoded vs. Actual Position, HPC Decoded Position F2 F1 CP 0.35 So, we can decode 0.3 F2 the animal s posiion 0.25 F1 0.2 from HPC what else P(Pos FR(c) ) = P(FR(c) 0.15 Pos) can * we P(Pos) CP look at? 0.1 P(FR(c)) 0.05 Decoded Position CP F1 F2 Actual Position CP F1 F2 Actual Position 0.02
51 (Bayesian) Decoding Place Cells in Hippocampus P(Pos FR(c) ) = P(FR(c) Pos) * P(Pos) P(FR(c)) What about P(Pos)? What should we use for our Prior? As usual, it depends on the quesion you want to ask
52 Anyhoo, I could keep going, but we re probably about out of Ime QuesIons? Or is it just Ime for Pizza?
CS 361: Probability & Statistics
October 17, 2017 CS 361: Probability & Statistics Inference Maximum likelihood: drawbacks A couple of things might trip up max likelihood estimation: 1) Finding the maximum of some functions can be quite
More informationCS 361: Probability & Statistics
March 14, 2018 CS 361: Probability & Statistics Inference The prior From Bayes rule, we know that we can express our function of interest as Likelihood Prior Posterior The right hand side contains the
More informationBayes Formula. MATH 107: Finite Mathematics University of Louisville. March 26, 2014
Bayes Formula MATH 07: Finite Mathematics University of Louisville March 26, 204 Test Accuracy Conditional reversal 2 / 5 A motivating question A rare disease occurs in out of every 0,000 people. A test
More informationFrequentist Statistics and Hypothesis Testing Spring
Frequentist Statistics and Hypothesis Testing 18.05 Spring 2018 http://xkcd.com/539/ Agenda Introduction to the frequentist way of life. What is a statistic? NHST ingredients; rejection regions Simple
More informationComparison of Bayesian and Frequentist Inference
Comparison of Bayesian and Frequentist Inference 18.05 Spring 2014 First discuss last class 19 board question, January 1, 2017 1 /10 Compare Bayesian inference Uses priors Logically impeccable Probabilities
More informationQuantitative Understanding in Biology 1.7 Bayesian Methods
Quantitative Understanding in Biology 1.7 Bayesian Methods Jason Banfelder October 25th, 2018 1 Introduction So far, most of the methods we ve looked at fall under the heading of classical, or frequentist
More informationDiscrete Binary Distributions
Discrete Binary Distributions Carl Edward Rasmussen November th, 26 Carl Edward Rasmussen Discrete Binary Distributions November th, 26 / 5 Key concepts Bernoulli: probabilities over binary variables Binomial:
More informationComputational Perception. Bayesian Inference
Computational Perception 15-485/785 January 24, 2008 Bayesian Inference The process of probabilistic inference 1. define model of problem 2. derive posterior distributions and estimators 3. estimate parameters
More informationHypothesis tests
6.1 6.4 Hypothesis tests Prof. Tesler Math 186 February 26, 2014 Prof. Tesler 6.1 6.4 Hypothesis tests Math 186 / February 26, 2014 1 / 41 6.1 6.2 Intro to hypothesis tests and decision rules Hypothesis
More informationP (E) = P (A 1 )P (A 2 )... P (A n ).
Lecture 9: Conditional probability II: breaking complex events into smaller events, methods to solve probability problems, Bayes rule, law of total probability, Bayes theorem Discrete Structures II (Summer
More informationBayesian RL Seminar. Chris Mansley September 9, 2008
Bayesian RL Seminar Chris Mansley September 9, 2008 Bayes Basic Probability One of the basic principles of probability theory, the chain rule, will allow us to derive most of the background material in
More informationCPSC 340: Machine Learning and Data Mining
CPSC 340: Machine Learning and Data Mining MLE and MAP Original version of these slides by Mark Schmidt, with modifications by Mike Gelbart. 1 Admin Assignment 4: Due tonight. Assignment 5: Will be released
More information6.041SC Probabilistic Systems Analysis and Applied Probability, Fall 2013 Transcript Tutorial:A Random Number of Coin Flips
6.041SC Probabilistic Systems Analysis and Applied Probability, Fall 2013 Transcript Tutorial:A Random Number of Coin Flips Hey, everyone. Welcome back. Today, we're going to do another fun problem that
More informationChapter Three. Hypothesis Testing
3.1 Introduction The final phase of analyzing data is to make a decision concerning a set of choices or options. Should I invest in stocks or bonds? Should a new product be marketed? Are my products being
More information1 A simple example. A short introduction to Bayesian statistics, part I Math 217 Probability and Statistics Prof. D.
probabilities, we ll use Bayes formula. We can easily compute the reverse probabilities A short introduction to Bayesian statistics, part I Math 17 Probability and Statistics Prof. D. Joyce, Fall 014 I
More informationExam 2 Practice Questions, 18.05, Spring 2014
Exam 2 Practice Questions, 18.05, Spring 2014 Note: This is a set of practice problems for exam 2. The actual exam will be much shorter. Within each section we ve arranged the problems roughly in order
More informationCS 124 Math Review Section January 29, 2018
CS 124 Math Review Section CS 124 is more math intensive than most of the introductory courses in the department. You re going to need to be able to do two things: 1. Perform some clever calculations to
More informationIntroductory Econometrics. Review of statistics (Part II: Inference)
Introductory Econometrics Review of statistics (Part II: Inference) Jun Ma School of Economics Renmin University of China October 1, 2018 1/16 Null and alternative hypotheses Usually, we have two competing
More informationBayesian Updating with Discrete Priors Class 11, Jeremy Orloff and Jonathan Bloom
1 Learning Goals ian Updating with Discrete Priors Class 11, 18.05 Jeremy Orloff and Jonathan Bloom 1. Be able to apply theorem to compute probabilities. 2. Be able to define the and to identify the roles
More informationModule 03 Lecture 14 Inferential Statistics ANOVA and TOI
Introduction of Data Analytics Prof. Nandan Sudarsanam and Prof. B Ravindran Department of Management Studies and Department of Computer Science and Engineering Indian Institute of Technology, Madras Module
More informationBayesian data analysis using JASP
Bayesian data analysis using JASP Dani Navarro compcogscisydney.com/jasp-tute.html Part 1: Theory Philosophy of probability Introducing Bayes rule Bayesian reasoning A simple example Bayesian hypothesis
More informationIntroduction to Bayesian Statistics
Bayesian Parameter Estimation Introduction to Bayesian Statistics Harvey Thornburg Center for Computer Research in Music and Acoustics (CCRMA) Department of Music, Stanford University Stanford, California
More informationLing 289 Contingency Table Statistics
Ling 289 Contingency Table Statistics Roger Levy and Christopher Manning This is a summary of the material that we ve covered on contingency tables. Contingency tables: introduction Odds ratios Counting,
More informationDS-GA 1003: Machine Learning and Computational Statistics Homework 7: Bayesian Modeling
DS-GA 1003: Machine Learning and Computational Statistics Homework 7: Bayesian Modeling Due: Tuesday, May 10, 2016, at 6pm (Submit via NYU Classes) Instructions: Your answers to the questions below, including
More informationChapter 1 Review of Equations and Inequalities
Chapter 1 Review of Equations and Inequalities Part I Review of Basic Equations Recall that an equation is an expression with an equal sign in the middle. Also recall that, if a question asks you to solve
More informationMATH MW Elementary Probability Course Notes Part I: Models and Counting
MATH 2030 3.00MW Elementary Probability Course Notes Part I: Models and Counting Tom Salisbury salt@yorku.ca York University Winter 2010 Introduction [Jan 5] Probability: the mathematics used for Statistics
More informationReading for Lecture 6 Release v10
Reading for Lecture 6 Release v10 Christopher Lee October 11, 2011 Contents 1 The Basics ii 1.1 What is a Hypothesis Test?........................................ ii Example..................................................
More informationDiscrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14
CS 70 Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14 Introduction One of the key properties of coin flips is independence: if you flip a fair coin ten times and get ten
More informationDiscrete Probability and State Estimation
6.01, Fall Semester, 2007 Lecture 12 Notes 1 MASSACHVSETTS INSTITVTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science 6.01 Introduction to EECS I Fall Semester, 2007 Lecture 12 Notes
More informationLecture 4: Parameter Es/ma/on and Confidence Intervals. GENOME 560, Spring 2015 Doug Fowler, GS
Lecture 4: Parameter Es/ma/on and Confidence Intervals GENOME 560, Spring 2015 Doug Fowler, GS (dfowler@uw.edu) 1 Review: Probability DistribuIons Discrete: Binomial distribuion Hypergeometric distribuion
More informationIntroduction: MLE, MAP, Bayesian reasoning (28/8/13)
STA561: Probabilistic machine learning Introduction: MLE, MAP, Bayesian reasoning (28/8/13) Lecturer: Barbara Engelhardt Scribes: K. Ulrich, J. Subramanian, N. Raval, J. O Hollaren 1 Classifiers In this
More informationIntroduction to Bayesian Learning. Machine Learning Fall 2018
Introduction to Bayesian Learning Machine Learning Fall 2018 1 What we have seen so far What does it mean to learn? Mistake-driven learning Learning by counting (and bounding) number of mistakes PAC learnability
More informationCompute f(x θ)f(θ) dθ
Bayesian Updating: Continuous Priors 18.05 Spring 2014 b a Compute f(x θ)f(θ) dθ January 1, 2017 1 /26 Beta distribution Beta(a, b) has density (a + b 1)! f (θ) = θ a 1 (1 θ) b 1 (a 1)!(b 1)! http://mathlets.org/mathlets/beta-distribution/
More informationMachine Learning CMPT 726 Simon Fraser University. Binomial Parameter Estimation
Machine Learning CMPT 726 Simon Fraser University Binomial Parameter Estimation Outline Maximum Likelihood Estimation Smoothed Frequencies, Laplace Correction. Bayesian Approach. Conjugate Prior. Uniform
More informationORF 245 Fundamentals of Statistics Chapter 9 Hypothesis Testing
ORF 245 Fundamentals of Statistics Chapter 9 Hypothesis Testing Robert Vanderbei Fall 2014 Slides last edited on November 24, 2014 http://www.princeton.edu/ rvdb Coin Tossing Example Consider two coins.
More informationPHP2510: Principles of Biostatistics & Data Analysis. Lecture X: Hypothesis testing. PHP 2510 Lec 10: Hypothesis testing 1
PHP2510: Principles of Biostatistics & Data Analysis Lecture X: Hypothesis testing PHP 2510 Lec 10: Hypothesis testing 1 In previous lectures we have encountered problems of estimating an unknown population
More informationBayesian Inference for Normal Mean
Al Nosedal. University of Toronto. November 18, 2015 Likelihood of Single Observation The conditional observation distribution of y µ is Normal with mean µ and variance σ 2, which is known. Its density
More informationBayesian Models in Machine Learning
Bayesian Models in Machine Learning Lukáš Burget Escuela de Ciencias Informáticas 2017 Buenos Aires, July 24-29 2017 Frequentist vs. Bayesian Frequentist point of view: Probability is the frequency of
More informationIntroduc)on to Bayesian methods (con)nued) - Lecture 16
Introduc)on to Bayesian methods (con)nued) - Lecture 16 David Sontag New York University Slides adapted from Luke Zettlemoyer, Carlos Guestrin, Dan Klein, and Vibhav Gogate Outline of lectures Review of
More informationWhich coin will I use? Which coin will I use? Logistics. Statistical Learning. Topics. 573 Topics. Coin Flip. Coin Flip
Logistics Statistical Learning CSE 57 Team Meetings Midterm Open book, notes Studying See AIMA exercises Daniel S. Weld Daniel S. Weld Supervised Learning 57 Topics Logic-Based Reinforcement Learning Planning
More informationMITOCW watch?v=vjzv6wjttnc
MITOCW watch?v=vjzv6wjttnc PROFESSOR: We just saw some random variables come up in the bigger number game. And we're going to be talking now about random variables, just formally what they are and their
More information( )( b + c) = ab + ac, but it can also be ( )( a) = ba + ca. Let s use the distributive property on a couple of
Factoring Review for Algebra II The saddest thing about not doing well in Algebra II is that almost any math teacher can tell you going into it what s going to trip you up. One of the first things they
More informationComputational Cognitive Science
Computational Cognitive Science Lecture 9: A Bayesian model of concept learning Chris Lucas School of Informatics University of Edinburgh October 16, 218 Reading Rules and Similarity in Concept Learning
More informationHypothesis Testing. ) the hypothesis that suggests no change from previous experience
Hypothesis Testing Definitions Hypothesis a claim about something Null hypothesis ( H 0 ) the hypothesis that suggests no change from previous experience Alternative hypothesis ( H 1 ) the hypothesis that
More informationIt applies to discrete and continuous random variables, and a mix of the two.
3. Bayes Theorem 3.1 Bayes Theorem A straightforward application of conditioning: using p(x, y) = p(x y) p(y) = p(y x) p(x), we obtain Bayes theorem (also called Bayes rule) p(x y) = p(y x) p(x). p(y)
More informationChapter 26: Comparing Counts (Chi Square)
Chapter 6: Comparing Counts (Chi Square) We ve seen that you can turn a qualitative variable into a quantitative one (by counting the number of successes and failures), but that s a compromise it forces
More informationCSC321 Lecture 18: Learning Probabilistic Models
CSC321 Lecture 18: Learning Probabilistic Models Roger Grosse Roger Grosse CSC321 Lecture 18: Learning Probabilistic Models 1 / 25 Overview So far in this course: mainly supervised learning Language modeling
More informationSolution to Proof Questions from September 1st
Solution to Proof Questions from September 1st Olena Bormashenko September 4, 2011 What is a proof? A proof is an airtight logical argument that proves a certain statement in general. In a sense, it s
More informationSome slides from Carlos Guestrin, Luke Zettlemoyer & K Gajos 2
Logistics CSE 446: Point Estimation Winter 2012 PS2 out shortly Dan Weld Some slides from Carlos Guestrin, Luke Zettlemoyer & K Gajos 2 Last Time Random variables, distributions Marginal, joint & conditional
More informationStatistical testing. Samantha Kleinberg. October 20, 2009
October 20, 2009 Intro to significance testing Significance testing and bioinformatics Gene expression: Frequently have microarray data for some group of subjects with/without the disease. Want to find
More informationDiscrete Mathematics and Probability Theory Spring 2014 Anant Sahai Note 10
EECS 70 Discrete Mathematics and Probability Theory Spring 2014 Anant Sahai Note 10 Introduction to Basic Discrete Probability In the last note we considered the probabilistic experiment where we flipped
More informationUsing Probability to do Statistics.
Al Nosedal. University of Toronto. November 5, 2015 Milk and honey and hemoglobin Animal experiments suggested that honey in a diet might raise hemoglobin level. A researcher designed a study involving
More informationWhy should you care?? Intellectual curiosity. Gambling. Mathematically the same as the ESP decision problem we discussed in Week 4.
I. Probability basics (Sections 4.1 and 4.2) Flip a fair (probability of HEADS is 1/2) coin ten times. What is the probability of getting exactly 5 HEADS? What is the probability of getting exactly 10
More informationBayesian Analysis. Justin Chin. Spring 2018
Bayesian Analysis Justin Chin Spring 2018 Abstract We often think of the field of Statistics simply as data collection and analysis. While the essence of Statistics lies in numeric analysis of observed
More informationConditional probabilities and graphical models
Conditional probabilities and graphical models Thomas Mailund Bioinformatics Research Centre (BiRC), Aarhus University Probability theory allows us to describe uncertainty in the processes we model within
More informationBayesian Updating with Continuous Priors Class 13, Jeremy Orloff and Jonathan Bloom
Bayesian Updating with Continuous Priors Class 3, 8.05 Jeremy Orloff and Jonathan Bloom Learning Goals. Understand a parameterized family of distributions as representing a continuous range of hypotheses
More informationCPSC 340: Machine Learning and Data Mining. MLE and MAP Fall 2017
CPSC 340: Machine Learning and Data Mining MLE and MAP Fall 2017 Assignment 3: Admin 1 late day to hand in tonight, 2 late days for Wednesday. Assignment 4: Due Friday of next week. Last Time: Multi-Class
More informationWhat are the Findings?
What are the Findings? James B. Rawlings Department of Chemical and Biological Engineering University of Wisconsin Madison Madison, Wisconsin April 2010 Rawlings (Wisconsin) Stating the findings 1 / 33
More informationPhysics 403. Segev BenZvi. Choosing Priors and the Principle of Maximum Entropy. Department of Physics and Astronomy University of Rochester
Physics 403 Choosing Priors and the Principle of Maximum Entropy Segev BenZvi Department of Physics and Astronomy University of Rochester Table of Contents 1 Review of Last Class Odds Ratio Occam Factors
More informationToday. Statistical Learning. Coin Flip. Coin Flip. Experiment 1: Heads. Experiment 1: Heads. Which coin will I use? Which coin will I use?
Today Statistical Learning Parameter Estimation: Maximum Likelihood (ML) Maximum A Posteriori (MAP) Bayesian Continuous case Learning Parameters for a Bayesian Network Naive Bayes Maximum Likelihood estimates
More information1 Probabilities. 1.1 Basics 1 PROBABILITIES
1 PROBABILITIES 1 Probabilities Probability is a tricky word usually meaning the likelyhood of something occuring or how frequent something is. Obviously, if something happens frequently, then its probability
More informationBayesian Inference. STA 121: Regression Analysis Artin Armagan
Bayesian Inference STA 121: Regression Analysis Artin Armagan Bayes Rule...s! Reverend Thomas Bayes Posterior Prior p(θ y) = p(y θ)p(θ)/p(y) Likelihood - Sampling Distribution Normalizing Constant: p(y
More informationWays to make neural networks generalize better
Ways to make neural networks generalize better Seminar in Deep Learning University of Tartu 04 / 10 / 2014 Pihel Saatmann Topics Overview of ways to improve generalization Limiting the size of the weights
More informationCS 188: Artificial Intelligence Spring Today
CS 188: Artificial Intelligence Spring 2006 Lecture 9: Naïve Bayes 2/14/2006 Dan Klein UC Berkeley Many slides from either Stuart Russell or Andrew Moore Bayes rule Today Expectations and utilities Naïve
More informationMAT Mathematics in Today's World
MAT 1000 Mathematics in Today's World Last Time We discussed the four rules that govern probabilities: 1. Probabilities are numbers between 0 and 1 2. The probability an event does not occur is 1 minus
More informationUni- and Bivariate Power
Uni- and Bivariate Power Copyright 2002, 2014, J. Toby Mordkoff Note that the relationship between risk and power is unidirectional. Power depends on risk, but risk is completely independent of power.
More informationSTOR 435 Lecture 5. Conditional Probability and Independence - I
STOR 435 Lecture 5 Conditional Probability and Independence - I Jan Hannig UNC Chapel Hill 1 / 16 Motivation Basic point Think of probability as the amount of belief we have in a particular outcome. If
More informationHypothesis testing I. - In particular, we are talking about statistical hypotheses. [get everyone s finger length!] n =
Hypothesis testing I I. What is hypothesis testing? [Note we re temporarily bouncing around in the book a lot! Things will settle down again in a week or so] - Exactly what it says. We develop a hypothesis,
More informationAnnouncements. Proposals graded
Announcements Proposals graded Kevin Jamieson 2018 1 Hypothesis testing Machine Learning CSE546 Kevin Jamieson University of Washington October 30, 2018 2018 Kevin Jamieson 2 Anomaly detection You are
More informationCOMP 551 Applied Machine Learning Lecture 19: Bayesian Inference
COMP 551 Applied Machine Learning Lecture 19: Bayesian Inference Associate Instructor: (herke.vanhoof@mcgill.ca) Class web page: www.cs.mcgill.ca/~jpineau/comp551 Unless otherwise noted, all material posted
More informationBEGINNING BAYES IN R. Bayes with discrete models
BEGINNING BAYES IN R Bayes with discrete models Beginning Bayes in R Survey on eating out What is your favorite day for eating out? Construct a prior for p Define p: proportion of all students who answer
More informationWith Question/Answer Animations. Chapter 7
With Question/Answer Animations Chapter 7 Chapter Summary Introduction to Discrete Probability Probability Theory Bayes Theorem Section 7.1 Section Summary Finite Probability Probabilities of Complements
More informationSTAT 285 Fall Assignment 1 Solutions
STAT 285 Fall 2014 Assignment 1 Solutions 1. An environmental agency sets a standard of 200 ppb for the concentration of cadmium in a lake. The concentration of cadmium in one lake is measured 17 times.
More informationCENTRAL LIMIT THEOREM (CLT)
CENTRAL LIMIT THEOREM (CLT) A sampling distribution is the probability distribution of the sample statistic that is formed when samples of size n are repeatedly taken from a population. If the sample statistic
More informationIntroduc)on to Bayesian Methods
Introduc)on to Bayesian Methods Bayes Rule py x)px) = px! y) = px y)py) py x) = px y)py) px) px) =! px! y) = px y)py) y py x) = py x) =! y "! y px y)py) px y)py) px y)py) px y)py)dy Bayes Rule py x) =
More informationData Analysis and Monte Carlo Methods
Lecturer: Allen Caldwell, Max Planck Institute for Physics & TUM Recitation Instructor: Oleksander (Alex) Volynets, MPP & TUM General Information: - Lectures will be held in English, Mondays 16-18:00 -
More informationLecture 9: Naive Bayes, SVM, Kernels. Saravanan Thirumuruganathan
Lecture 9: Naive Bayes, SVM, Kernels Instructor: Outline 1 Probability basics 2 Probabilistic Interpretation of Classification 3 Bayesian Classifiers, Naive Bayes 4 Support Vector Machines Probability
More informationYou separate binary numbers into columns in a similar fashion. 2 5 = 32
RSA Encryption 2 At the end of Part I of this article, we stated that RSA encryption works because it s impractical to factor n, which determines P 1 and P 2, which determines our private key, d, which
More informationPIER HLM Course July 30, 2011 Howard Seltman. Discussion Guide for Bayes and BUGS
PIER HLM Course July 30, 2011 Howard Seltman Discussion Guide for Bayes and BUGS 1. Classical Statistics is based on parameters as fixed unknown values. a. The standard approach is to try to discover,
More informationParameter Learning With Binary Variables
With Binary Variables University of Nebraska Lincoln CSCE 970 Pattern Recognition Outline Outline 1 Learning a Single Parameter 2 More on the Beta Density Function 3 Computing a Probability Interval Outline
More informationMITOCW ocw f99-lec30_300k
MITOCW ocw-18.06-f99-lec30_300k OK, this is the lecture on linear transformations. Actually, linear algebra courses used to begin with this lecture, so you could say I'm beginning this course again by
More informationOne sided tests. An example of a two sided alternative is what we ve been using for our two sample tests:
One sided tests So far all of our tests have been two sided. While this may be a bit easier to understand, this is often not the best way to do a hypothesis test. One simple thing that we can do to get
More informationFERMAT S TEST KEITH CONRAD
FERMAT S TEST KEITH CONRAD 1. Introduction Fermat s little theorem says for prime p that a p 1 1 mod p for all a 0 mod p. A naive extension of this to a composite modulus n 2 would be: for all a 0 mod
More informationHST.582J / 6.555J / J Biomedical Signal and Image Processing Spring 2007
MIT OpenCourseWare http://ocw.mit.edu HST.582J / 6.555J / 16.456J Biomedical Signal and Image Processing Spring 2007 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.
More informationHypothesis testing. Data to decisions
Hypothesis testing Data to decisions The idea Null hypothesis: H 0 : the DGP/population has property P Under the null, a sample statistic has a known distribution If, under that that distribution, the
More informationModeling Environment
Topic Model Modeling Environment What does it mean to understand/ your environment? Ability to predict Two approaches to ing environment of words and text Latent Semantic Analysis (LSA) Topic Model LSA
More informationJoint, Conditional, & Marginal Probabilities
Joint, Conditional, & Marginal Probabilities The three axioms for probability don t discuss how to create probabilities for combined events such as P [A B] or for the likelihood of an event A given that
More informationIntermediate Math Circles November 15, 2017 Probability III
Intermediate Math Circles November 5, 07 Probability III Example : You have bins in which there are coloured balls. The balls are identical except for their colours. The contents of the containers are:
More informationECE295, Data Assimila0on and Inverse Problems, Spring 2015
ECE295, Data Assimila0on and Inverse Problems, Spring 2015 1 April, Intro; Linear discrete Inverse problems (Aster Ch 1 and 2) Slides 8 April, SVD (Aster ch 2 and 3) Slides 15 April, RegularizaFon (ch
More informationProbability Theory and Simulation Methods
Feb 28th, 2018 Lecture 10: Random variables Countdown to midterm (March 21st): 28 days Week 1 Chapter 1: Axioms of probability Week 2 Chapter 3: Conditional probability and independence Week 4 Chapters
More informationz and t tests for the mean of a normal distribution Confidence intervals for the mean Binomial tests
z and t tests for the mean of a normal distribution Confidence intervals for the mean Binomial tests Chapters 3.5.1 3.5.2, 3.3.2 Prof. Tesler Math 283 Fall 2018 Prof. Tesler z and t tests for mean Math
More informationSolving Quadratic & Higher Degree Equations
Chapter 9 Solving Quadratic & Higher Degree Equations Sec 1. Zero Product Property Back in the third grade students were taught when they multiplied a number by zero, the product would be zero. In algebra,
More informationSTAT 201 Chapter 5. Probability
STAT 201 Chapter 5 Probability 1 2 Introduction to Probability Probability The way we quantify uncertainty. Subjective Probability A probability derived from an individual's personal judgment about whether
More informationBasics of Proofs. 1 The Basics. 2 Proof Strategies. 2.1 Understand What s Going On
Basics of Proofs The Putnam is a proof based exam and will expect you to write proofs in your solutions Similarly, Math 96 will also require you to write proofs in your homework solutions If you ve seen
More informationDiscrete Probability and State Estimation
6.01, Spring Semester, 2008 Week 12 Course Notes 1 MASSACHVSETTS INSTITVTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science 6.01 Introduction to EECS I Spring Semester, 2008 Week
More informationIntroduction to AI Learning Bayesian networks. Vibhav Gogate
Introduction to AI Learning Bayesian networks Vibhav Gogate Inductive Learning in a nutshell Given: Data Examples of a function (X, F(X)) Predict function F(X) for new examples X Discrete F(X): Classification
More informationStandard & Conditional Probability
Biostatistics 050 Standard & Conditional Probability 1 ORIGIN 0 Probability as a Concept: Standard & Conditional Probability "The probability of an event is the likelihood of that event expressed either
More informationCIS 2033 Lecture 5, Fall
CIS 2033 Lecture 5, Fall 2016 1 Instructor: David Dobor September 13, 2016 1 Supplemental reading from Dekking s textbook: Chapter2, 3. We mentioned at the beginning of this class that calculus was a prerequisite
More informationBayesian Inference. Introduction
Bayesian Inference Introduction The frequentist approach to inference holds that probabilities are intrinsicially tied (unsurprisingly) to frequencies. This interpretation is actually quite natural. What,
More information