Chapter 8: Sampling distributions of estimators Sections

Similar documents
Bayesian statistics, simulation and software

Part 4: Multi-parameter and normal models

Statistical Theory MT 2007 Problems 4: Solution sketches

Statistical Theory MT 2006 Problems 4: Solution sketches

An Introduction to Bayesian Linear Regression

Pattern Recognition and Machine Learning. Bishop Chapter 2: Probability Distributions

Introduction to Applied Bayesian Modeling. ICPSR Day 4

Beta statistics. Keywords. Bayes theorem. Bayes rule

Foundations of Statistical Inference

Classical and Bayesian inference

Classical and Bayesian inference

1 Introduction. P (n = 1 red ball drawn) =

HPD Intervals / Regions

Problem Selected Scores

Chapter 5 continued. Chapter 5 sections

Bayesian Inference for Normal Mean

Bayesian Inference and Decision Theory

STAT J535: Chapter 5: Classes of Bayesian Priors

Bayesian Regression Linear and Logistic Regression

Bayesian Inference in a Normal Population

INTRODUCTION TO BAYESIAN METHODS II

Bayesian Inference in a Normal Population

Bayesian inference. Rasmus Waagepetersen Department of Mathematics Aalborg University Denmark. April 10, 2017

Bayesian Inference. STA 121: Regression Analysis Artin Armagan

MATH c UNIVERSITY OF LEEDS Examination for the Module MATH2715 (January 2015) STATISTICAL METHODS. Time allowed: 2 hours

The linear model is the most fundamental of all serious statistical models encompassing:

Bayesian Inference: Concept and Practice

Chapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued

Predictive Distributions

Generalized Bayesian Inference with Sets of Conjugate Priors for Dealing with Prior-Data Conflict

Module 22: Bayesian Methods Lecture 9 A: Default prior selection

Multivariate Normal & Wishart

Bayesian Linear Regression

Bayesian Inference for Regression Parameters

Distributions of Functions of Random Variables. 5.1 Functions of One Random Variable

Introduction. Start with a probability distribution f(y θ) for the data. where η is a vector of hyperparameters

Lecture 4: Probabilistic Learning. Estimation Theory. Classification with Probability Distributions

The binomial model. Assume a uniform prior distribution on p(θ). Write the pdf for this distribution.

Stat 535 C - Statistical Computing & Monte Carlo Methods. Arnaud Doucet.

Bayesian statistics, simulation and software

Statistics 360/601 Modern Bayesian Theory

Remarks on Improper Ignorance Priors

Introduction to Bayesian Methods. Introduction to Bayesian Methods p.1/??

Chapter 5. Bayesian Statistics


Chapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued

USEFUL PROPERTIES OF THE MULTIVARIATE NORMAL*

STATISTICAL METHODS FOR SIGNAL PROCESSING c Alfred Hero

Conjugate Analysis for the Linear Model

STA 2201/442 Assignment 2

The Jeffreys Prior. Yingbo Li MATH Clemson University. Yingbo Li (Clemson) The Jeffreys Prior MATH / 13

Foundations of Statistical Inference

The exponential family: Conjugate priors

Stat 5101 Lecture Notes


MAS3301 Bayesian Statistics

Probability Distributions

Part 6: Multivariate Normal and Linear Models

MAS3301 Bayesian Statistics

Bayesian Linear Models

Chapter 6: Large Random Samples Sections

COS513 LECTURE 8 STATISTICAL CONCEPTS

STAT 730 Chapter 4: Estimation

Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics

Lecture 4: Probabilistic Learning

F & B Approaches to a simple model

All other items including (and especially) CELL PHONES must be left at the front of the room.

Sampling Distributions

Introduction to Bayesian Methods

Bayesian Linear Models

Bayesian Inference. Chapter 9. Linear models and regression

Conjugate Priors for Normal Data

Generalized Linear Models Introduction

Hypothesis Testing. Econ 690. Purdue University. Justin L. Tobias (Purdue) Testing 1 / 33

A Few Special Distributions and Their Properties

STA 4322 Exam I Name: Introduction to Statistics Theory

1 Hypothesis Testing and Model Selection

Bayesian inference. Fredrik Ronquist and Peter Beerli. October 3, 2007

Model Checking and Improvement

General Bayesian Inference I

Topic 12 Overview of Estimation

Bayesian Inference. Chapter 2: Conjugate models

1 Data Arrays and Decompositions

Hierarchical models. Dr. Jarad Niemi. August 31, Iowa State University. Jarad Niemi (Iowa State) Hierarchical models August 31, / 31

Sampling Distributions

Chapter 8.8.1: A factorization theorem

Ph.D. Qualifying Exam: Algebra I

Chapter 9: Hypothesis Testing Sections

Divergence Based priors for the problem of hypothesis testing

Principles of Bayesian Inference

Statistical Inference

Gibbs Sampling in Linear Models #2

Bayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework

MAXIMUM LIKELIHOOD, SET ESTIMATION, MODEL CRITICISM

This exam is closed book and closed notes. (You will have access to a copy of the Table of Common Distributions given in the back of the text.

Statistics STAT:5100 (22S:193), Fall Sample Final Exam B

Multivariate Bayesian Linear Regression MLAI Lecture 11

Introduction to Bayesian Statistics. James Swain University of Alabama in Huntsville ISEEM Department

Bayesian Inference: Posterior Intervals

Homework 1: Solution

Transcription:

Chapter 8: Sampling distributions of estimators Sections 8.1 Sampling distribution of a statistic 8.2 The Chi-square distributions 8.3 Joint Distribution of the sample mean and sample variance Skip: p. 476-478 8.4 The t distributions Skip: derivation of the pdf, p. 483-484 8.5 Confidence intervals 8.7 Unbiased Estimators 8.8 Fisher Information STA 611 (Lecture 16) Sampling Distributions Nov 1, 2012 1 / 13

Bayesian alternative to confidence intervals Bayesian inference is the posterior distribution. Reporting a whole distribution may not be what you (or your client) want For point estimates we use Bayesian estimators Minimize the expected loss If we want an interval to go with the point estimator we simply use quantiles of the posterior distribution For example: We can find constants c 1 and c 2 so that P(c 1 < θ < c 1 X = x) γ The interval (c 1, c 2 ) is called a 100γ% Credible interval for θ Note: The interpretation is very different from interpretation of confidence intervals STA 611 (Lecture 16) Sampling Distributions Nov 1, 2012 2 / 13

Bayesian Analysis for the normal distribution Let X 1,..., X n be a random sample for N(µ, σ 2 ) In Chapter 7.3 we saw: If σ 2 is known, the normal distribution is a conjugate prior for µ Theorem 7.3.3: If the prior is µ N(µ 0, ν0 2 ) the posterior of µ is also normal with mean and variance µ 1 = σ2 µ 0 + nν 2 0 x n σ 2 + nν 2 0 and ν 2 1 = σ2 ν 2 0 σ 2 + nν 2 0 We can obtain credible intervals for µ from this N(µ 1, ν 2 1 ) posterior distribution STA 611 (Lecture 16) Sampling Distributions Nov 1, 2012 3 / 13

Bayesian Analysis for the normal distribution Let X 1,..., X n be a random sample for N(µ, σ 2 ) In Chapter 7.3 we saw: If µ is known, the Inverse-Gamma distribution is a conjugate prior for σ 2 Example 7.3.15: If the prior is σ 2 IG(α 0, β 0 ) the posterior of σ 2 is also Inverse-Gamma with parameters α 1 = α 0 + n 2 and β 1 = β + 1 2 n (x i µ) 2 i=1 We can obtain credible intervals for σ 2 from this IG(α 1, β 1 ) posterior distribution What if both µ and σ 2 are unknown? We need the joint posterior distribution of µ and σ 2 STA 611 (Lecture 16) Sampling Distributions Nov 1, 2012 4 / 13

Bayesian Analysis for the normal distribution When both µ and σ 2 are unknown Def: Precision The precision of a normal distribution is the reciprocal of the variance: τ = 1 σ 2 It is somewhat simpler to work with the precision than the variance The pdf of the normal distribution is then written as f (x µ, τ) = τ 1/2 exp ( 12 ) τ(x µ)2 2π STA 611 (Lecture 16) Sampling Distributions Nov 1, 2012 5 / 13

Normal-Gamma distribution Def: Normal-Gamma distribution Let µ and τ be random variables where µ τ has the normal distribution with mean µ and precision λ 0 τ and τ has the Gamma distribution: µ τ N(µ 0, 1/λ 0 τ) and τ Gamma(α 0, β 0 ) Then the joint distribution of µ and τ is called the Normal-Gamma distribution with parameters µ 0, λ 0, α 0 and β 0. The pdf for this distribution is f (µ, τ µ 0, λ 0, α 0, β 0 ) = (λ 0τ) 1/2 2π exp ( 12 ) λ 0τ(µ µ 0 ) 2 βα 0 0 Γ(α 0 ) τ α0 1 exp ( β 0 τ) STA 611 (Lecture 16) Sampling Distributions Nov 1, 2012 6 / 13

z Chapter 8 continued Pdf of the Normal-Gamma distribution µ 0 = 0, λ 0 = 1, α 0 = 0.5 and β 0 = 0.5 mu tau STA 611 (Lecture 16) Sampling Distributions Nov 1, 2012 7 / 13

Bayesian Analysis for the normal distribution When both µ and σ 2 are unknown Theorem 8.6.1: Conjugate prior for µ and τ Let X 1,..., X n be a random sample from N(µ, 1/τ). The Normal-Gamma distribution with parameters µ 0, λ 0, α 0 and β 0 is a conjugate prior distribution for (µ, τ) and the posterior distribution has (hyper)parameters µ 1 = λ 0µ 0 + nx n, λ 1 = λ 0 + n λ 0 + n α 1 = α 0 + 1 2 and β 1 = β 0 + 1 2 s2 n + nλ 0(x n µ 0 ) 2 2(λ 0 + n) where x n = 1 n n x i and sn 2 = n (x i x n ) 2 i=1 i=1 STA 611 (Lecture 16) Sampling Distributions Nov 1, 2012 8 / 13

Bayesian Analysis for the normal distribution To give credible intervals for µ and σ 2 individually we need the marginal posterior distributions From the structure of the Normal-Gamma distribution we immediately get the marginal for τ: τ X = x Gamma(α 1, β 1 ) This distribution can be used to obtain credible intervals for τ, or any function of τ. STA 611 (Lecture 16) Sampling Distributions Nov 1, 2012 9 / 13

Marginal distribution of the mean Theorem 8.6.2: The marginal distribution of µ Let the joint distribution of µ and τ be the Normal-Gamma distribution with parameters µ 0, λ 0, α 0 and β 0. Then ( ) λ0 α 1/2 0 U = (µ µ 0 ) t 2α0 It is then easy to show that (Theorem 8.6.3): β 0 E(µ) = µ 0 (if α 0 > 1/2) and Var(µ) = β 0 λ 0 (α 0 1) (if α 0 > 1) STA 611 (Lecture 16) Sampling Distributions Nov 1, 2012 10 / 13

Marginal distribution of the mean Say we have done a Bayesian Analysis and end up with the Normal-Gamma posterior with parameters µ 1, λ 1, α 1 and β 1. We can then calculate the marginal posterior means of µ and τ E(µ x) = µ 1 (if α 1 > 1/2) and E(τ x) = α 1 β 1 Can also obtain credible intervals for µ Find quantiles c 1 and c 2 such that P(c 1 < U < c 2 ) = γ Then ( ( ) 1/2 ( ) ) 1/2 β1 β1 P µ 1 + c 1 µ µ 1 + c 2 λ 1 α 1 λ 1 α 1 x = γ E.g: for a 0.95% symmetric credible interval we set c 1 = T 1 2α 1 (0.975) and c 2 = T 1 2α 1 (0.975) STA 611 (Lecture 16) Sampling Distributions Nov 1, 2012 11 / 13

Example Hotdogs (Exercise 8.5.7 in the book) Data on calorie content in 20 different beef hot dogs from Consumer Reports (June 1986 issue): 186, 181, 176, 149, 184, 190, 158, 139, 175, 148, 152, 111, 141, 153, 190, 157, 131, 149, 135, 132 Assume that these numbers are observed values from a random sample of twenty independent N(µ, σ 2 ) random variables, where µ and σ 2 are unknown. n x n = 156.85 and sn 2 = (x i x n ) 2 = 9740.55 Consider the Normal-Gamma prior for µ and τ with parameters µ 0 = 100, λ 0 = 3, α 0 = 2 and β 0 = 2500. Construct the apriori symmetric 95% credible interval for µ i=1 Find the posterior symmetric 95% credible interval for µ STA 611 (Lecture 16) Sampling Distributions Nov 1, 2012 12 / 13

Other priors The usual improper prior for (µ, τ) is p(µ, τ) = 1 τ < µ, τ > 0 The posterior in this case is the Normal-Gamma distribution with parameters µ 1 = x, λ 1 = n, α 1 = (n 1)/2, β 1 = s 2 n/2 Credible intervals turn out to be the same as confidence intervals (common for improper priors) Other common options: µ and τ independent (a priori) with µ N(µ 0, ν 2 0 ) and τ Gamma(α 0, β 0 ). µ and τ will be dependent in the posterior. τ Gamma(α 0, β 0 ) but improper for µ: p(µ) = 1 STA 611 (Lecture 16) Sampling Distributions Nov 1, 2012 13 / 13