Comparison of Bayesian and Frequentist Inference

Similar documents
Compute f(x θ)f(θ) dθ

Exam 2 Practice Questions, 18.05, Spring 2014

Confidence Intervals for Normal Data Spring 2014

Bayesian Updating: Discrete Priors: Spring 2014

Null Hypothesis Significance Testing p-values, significance level, power, t-tests

Introductory Econometrics. Review of statistics (Part II: Inference)

Frequentist Statistics and Hypothesis Testing Spring

Studio 8: NHST: t-tests and Rejection Regions Spring 2014

Computational Perception. Bayesian Inference

Conditional Probability, Independence, Bayes Theorem Spring January 1, / 23

Bayesian Updating with Continuous Priors Class 13, Jeremy Orloff and Jonathan Bloom

Conjugate Priors: Beta and Normal Spring January 1, /15

18.05 Final Exam. Good luck! Name. No calculators. Number of problems 16 concept questions, 16 problems, 21 pages

Conditional Probability, Independence, Bayes Theorem Spring January 1, / 28

Class 26: review for final exam 18.05, Spring 2014

Continuous Data with Continuous Priors Class 14, Jeremy Orloff and Jonathan Bloom

Statistical Preliminaries. Stony Brook University CSE545, Fall 2016

Chapter 7: Hypothesis Testing

Choosing priors Class 15, Jeremy Orloff and Jonathan Bloom

Choosing Priors Probability Intervals Spring January 1, /25

6.4 Type I and Type II Errors

Introduction to Bayesian Statistics

Bayesian Updating: Odds Class 12, Jeremy Orloff and Jonathan Bloom

(1) Introduction to Bayesian statistics

Hypothesis Testing. Testing Hypotheses MIT Dr. Kempthorne. Spring MIT Testing Hypotheses

Null Hypothesis Significance Testing p-values, significance level, power, t-tests Spring 2017

Choosing Priors Probability Intervals Spring 2014

ECE521 W17 Tutorial 6. Min Bai and Yuhuai (Tony) Wu

Announcements. Proposals graded

Estimation of reliability parameters from Experimental data (Parte 2) Prof. Enrico Zio

Conditional Probability, Independence and Bayes Theorem Class 3, Jeremy Orloff and Jonathan Bloom

18.05 Problem Set 5, Spring 2014 Solutions

Probability: Terminology and Examples Class 2, Jeremy Orloff and Jonathan Bloom

Stat 535 C - Statistical Computing & Monte Carlo Methods. Arnaud Doucet.

Bayesian Updating: Discrete Priors: Spring

18.05 Practice Final Exam

HST.582J / 6.555J / J Biomedical Signal and Image Processing Spring 2007

Hypothesis Tests Solutions COR1-GB.1305 Statistics and Data Analysis

Bayesian Inference. STA 121: Regression Analysis Artin Armagan

Introduction to Statistics

Bayesian Updating: Discrete Priors: Spring

Hypothesis tests

Linear Models: Comparing Variables. Stony Brook University CSE545, Fall 2017

Bayes Formula. MATH 107: Finite Mathematics University of Louisville. March 26, 2014

Lecture 23. Random walks

Statistics of Small Signals

Statistical Data Analysis Stat 3: p-values, parameter estimation

Example. χ 2 = Continued on the next page. All cells

Hypothesis Testing. Part I. James J. Heckman University of Chicago. Econ 312 This draft, April 20, 2006

Mathematical Statistics

Continuous Expectation and Variance, the Law of Large Numbers, and the Central Limit Theorem Spring 2014

Chapter Three. Hypothesis Testing

Inference for a Population Proportion

Accouncements. You should turn in a PDF and a python file(s) Figure for problem 9 should be in the PDF

Bayesian Analysis (Optional)

Chapter 5. Bayesian Statistics

Lecture Testing Hypotheses: The Neyman-Pearson Paradigm

For use only in [the name of your school] 2014 S4 Note. S4 Notes (Edexcel)

Central Limit Theorem and the Law of Large Numbers Class 6, Jeremy Orloff and Jonathan Bloom

Why I think Bayesian Analysis is the greatest thing since NHST (bread)

Bayesian Inference. p(y)

Visual interpretation with normal approximation

Topic 15: Simple Hypotheses

Bayesian Models in Machine Learning

ST 740: Model Selection

Intro to Bayesian Methods

Classical and Bayesian inference

18.05 Exam 1. Table of normal probabilities: The last page of the exam contains a table of standard normal cdf values.

Unobservable Parameter. Observed Random Sample. Calculate Posterior. Choosing Prior. Conjugate prior. population proportion, p prior:

ORF 245 Fundamentals of Statistics Chapter 9 Hypothesis Testing

Bayesian Learning (II)

MTH U481 : SPRING 2009: PRACTICE PROBLEMS FOR FINAL

STAT 515 fa 2016 Lec Statistical inference - hypothesis testing

The Central Limit Theorem

A Bayesian Approach to Phylogenetics

Definition 3.1 A statistical hypothesis is a statement about the unknown values of the parameters of the population distribution.

Discrete Binary Distributions

Quantitative Understanding in Biology 1.7 Bayesian Methods

Inconsistency of Bayesian inference when the model is wrong, and how to repair it

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14

Bayesian Methods: Naïve Bayes

Hierarchical Models & Bayesian Model Selection

8.01x Classical Mechanics, Fall 2016 Massachusetts Institute of Technology. Problem Set 8

Lecture 14. More discrete random variables

COMP 551 Applied Machine Learning Lecture 19: Bayesian Inference

Lecture Slides - Part 1

Bayesian inference. Fredrik Ronquist and Peter Beerli. October 3, 2007

BEGINNING BAYES IN R. Bayes with discrete models

Business Statistics: Lecture 8: Introduction to Estimation & Hypothesis Testing

Bayesian Inference: Posterior Intervals

Bootstrapping Spring 2014

Probability Theory. Probability and Statistics for Data Science CSE594 - Spring 2016

DS-GA 1002 Lecture notes 11 Fall Bayesian statistics

Bootstrap and Linear Regression Spring You should have downloaded studio12.zip and unzipped it into your working directory.

Bayesian Analysis for Natural Language Processing Lecture 2

Statistical Inference

Class 8 Review Problems 18.05, Spring 2014

Some Asymptotic Bayesian Inference (background to Chapter 2 of Tanner s book)

Introduction to Machine Learning

CS 543 Page 1 John E. Boon, Jr.

Transcription:

Comparison of Bayesian and Frequentist Inference 18.05 Spring 2014 First discuss last class 19 board question, January 1, 2017 1 /10

Compare Bayesian inference Uses priors Logically impeccable Probabilities can be interpreted Prior is subjective Frequentist inference No prior Objective everyone gets the same answer Logically complex Conditional probability of error is often misinterpreted as total probability of error Requires complete description of experimental protocol and data analysis protocol before starting the experiment. (This is both good and bad) January 1, 2017 2 /10

Concept question Three different tests are run all with significance level α = 0.05. 1. Experiment 1: finds p = 0.03 and rejects its null hypothesis H 0. 2. Experiment 2: finds p = 0.049 and rejects its null hypothesis. 3. Experiment 3: finds p = 0.15 and fails to rejects its null hypothesis. Which result has the highest probability of being correct? (Click 4 if you don t know.) answer: 4. You can t know probabilities of hypotheses based just on p values. January 1, 2017 3 /10

Board question: Stop! Experiments are run to test a coin that is suspected of being biased towards heads. The significance level is set to α = 0.1 Experiment 1: Toss a coin 5 times. Report the sequence of tosses. Experiment 2: Toss a coin until the first tails. Report the sequence of tosses. 1. Give the test statistic, null distribution and rejection region for each experiment. List all sequences of tosses that produce a test statistic in the rejection region for each experiment. 2. Suppose the data is HHHHT. (a) Do the significance test for both types of experiment. (b) Do a Bayesian update starting from a flat prior: Beta(1,1). Draw some conclusions about the fairness of coin from your posterior. (Use R: pbeta for computation in part (b).) January 1, 2017 4 /10

Solution 1. Experiment 1: The test statistic is the number of heads x out of 5 tosses. The null distribution is binomial(5,0.5). The rejection region is {x = 5}. The sequence of tosses HHHHH. is the only one that leads to rejection. Experiment 2: The test statistic is the number of heads x until the first tails. The null distribution is geom(0.5). The rejection region {x 4}. The sequences of tosses that lead to rejection are {HHHHT, HHHHH T }, where means an arbitrary length string of heads. 2a. For experiment 1 and the given data, as or more extreme means 4 or 5 heads. So for experiment 1 the p-value is P(4 or 5 heads fair coin) = 6/32 0.20. For experiment 2 and the given data as or more extreme means at least 4 heads at the start. So p = 1 - pgeom(3,0.5) = 0.0625. (Solution continued.) January 1, 2017 5 /10

Solution continued 2b. Let θ be the probability of heads, Four heads and a tail updates the prior on θ, Beta(1,1) to the posterior Beta(5,2). Using R we can compute P(Coin is biased to heads) = P(θ > 0.5) = 1 -pbeta(0.5,5,2) = 0.89. If the prior is good then the probability the coin is biased towards heads is 0.89. January 1, 2017 6 /10

Board question: Stop II For each of the following experiments (all done with α = 0.05) (a) Comment on the validity of the claims. (b) Find the true probability of a type I error in each experimental setup. 1 2 By design Ruthi did 50 trials and computed p = 0.04. She reports p = 0.04 with n = 50 and declares it significant. Ani did 50 trials and computed p = 0.06. Since this was not significant, she then did 50 more trials and computed p = 0.04 based on all 100 trials. She reports p = 0.04 with n = 100 and declares it significant. 3 Efrat did 50 trials and computed p = 0.06. Since this was not significant, she started over and computed p = 0.04 based on the next 50 trials. She reports p = 0.04 with n = 50 and declares it statistically significant. January 1, 2017 7 /10

Solution 1. (a) This is a reasonable NHST experiment. (b) The probability of a type I error is 0.05. 2. (a) The actual experiment run: (i) Do 50 trials. (ii) If p < 0.05 then stop. (iii) If not run another 50 trials. (iv) Compute p again, pretending that all 100 trials were run without any possibility of stopping. This is not a reasonable NHST experimental setup because the second p-values are computed using the wrong null distribution. (b) If H 0 is true then the probability of rejecting is already 0.05 by step (ii). It can only increase by allowing steps (iii) and (iv). So the probability of rejecting given H 0 is more than 0.05. We can t say how much more without more details. January 1, 2017 8 /10

Solution continued 3. (a) See answer to (2a). (b) The total probability of a type I error is more than 0.05. We can compute it using a probability tree. Since we are looking at type I errors all probabilities are computed assume H 0 is true. First 50 trials Second 50 trials.05 Reject 0.05.95 Continue Reject Don t reject The total probability of falsely rejecting H 0 is 0.05 + 0.05 0.95 = 0.0975 January 1, 2017 9 /10

MIT OpenCourseWare https://ocw.mit.edu 18.05 Introduction to Probability and Statistics Spring 2014 For information about citing these materials or our Terms of Use, visit: https://ocw.mit.edu/terms.