Poisson CI s. Robert L. Wolpert Department of Statistical Science Duke University, Durham, NC, USA

Similar documents
Confidence & Credible Interval Estimates

Classical and Bayesian inference

PARAMETER ESTIMATION: BAYESIAN APPROACH. These notes summarize the lectures on Bayesian parameter estimation.

Statistical Inference

Chapter 5. Bayesian Statistics

Statistical Inference

Chapter 4. Bayesian inference. 4.1 Estimation. Point estimates. Interval estimates

Statistical Theory MT 2007 Problems 4: Solution sketches

Interval Estimation. Chapter 9

Statistical Theory MT 2006 Problems 4: Solution sketches

Hypothesis Testing. Robert L. Wolpert Department of Statistical Science Duke University, Durham, NC, USA

Module 22: Bayesian Methods Lecture 9 A: Default prior selection

Chapter 9: Interval Estimation and Confidence Sets Lecture 16: Confidence sets and credible sets

Department of Statistics

Statistical Inference

A Very Brief Summary of Statistical Inference, and Examples

Bayesian Inference: Posterior Intervals

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

STAT 425: Introduction to Bayesian Analysis

Chapter 4 HOMEWORK ASSIGNMENTS. 4.1 Homework #1

POSTERIOR PROPRIETY IN SOME HIERARCHICAL EXPONENTIAL FAMILY MODELS

Part 2: One-parameter models

A Very Brief Summary of Bayesian Inference, and Examples

Introduction to Bayesian Methods

Introduction to Applied Bayesian Modeling. ICPSR Day 4

Beta statistics. Keywords. Bayes theorem. Bayes rule

Fisher Information & Efficiency

Confidence Distribution

Bayesian inference. Rasmus Waagepetersen Department of Mathematics Aalborg University Denmark. April 10, 2017

STAT J535: Chapter 5: Classes of Bayesian Priors

The binomial model. Assume a uniform prior distribution on p(θ). Write the pdf for this distribution.

Testing Simple Hypotheses R.L. Wolpert Institute of Statistics and Decision Sciences Duke University, Box Durham, NC 27708, USA

Stable Limit Laws for Marginal Probabilities from MCMC Streams: Acceleration of Convergence

Remarks on Improper Ignorance Priors

HPD Intervals / Regions

STA 2201/442 Assignment 2

Bayesian statistics, simulation and software

Final Examination a. STA 532: Statistical Inference. Wednesday, 2015 Apr 29, 7:00 10:00pm. Thisisaclosed bookexam books&phonesonthefloor.

Part III. A Decision-Theoretic Approach and Bayesian testing

Stat 5102 Final Exam May 14, 2015

Homework 1: Solution

Eco517 Fall 2004 C. Sims MIDTERM EXAM

Foundations of Statistical Inference

Chapter 8: Sampling distributions of estimators Sections

One Parameter Models

Two examples of the use of fuzzy set theory in statistics. Glen Meeden University of Minnesota.

ABC methods for phase-type distributions with applications in insurance risk problems

STAT 499/962 Topics in Statistics Bayesian Inference and Decision Theory Jan 2018, Handout 01

Estimation of reliability parameters from Experimental data (Parte 2) Prof. Enrico Zio

Statistics Ph.D. Qualifying Exam: Part II November 9, 2002

Classical and Bayesian inference

HT Introduction. P(X i = x i ) = e λ λ x i

1 Introduction. P (n = 1 red ball drawn) =

Lecture 2. (See Exercise 7.22, 7.23, 7.24 in Casella & Berger)

The Bayesian Choice. Christian P. Robert. From Decision-Theoretic Foundations to Computational Implementation. Second Edition.

Theory of Statistical Tests

Integer-Valued Polynomials

Simulating Random Variables

Fundamentals. CS 281A: Statistical Learning Theory. Yangqing Jia. August, Based on tutorial slides by Lester Mackey and Ariel Kleiner

Statistical Inference

Better Bootstrap Confidence Intervals

Statistical Inference: Maximum Likelihood and Bayesian Approaches

The comparative studies on reliability for Rayleigh models

Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing

2 Functions of random variables

An Overview of Objective Bayesian Analysis

Monte Carlo-based statistical methods (MASM11/FMS091)

A few basics of credibility theory

Qualifying Exam CS 661: System Simulation Summer 2013 Prof. Marvin K. Nakayama

Part 4: Multi-parameter and normal models

Statistics: Learning models from data

Other Noninformative Priors

The linear model is the most fundamental of all serious statistical models encompassing:

Introduction to Bayesian Computation

New Bayesian methods for model comparison

Bernoulli and Poisson models

The problem is to infer on the underlying probability distribution that gives rise to the data S.

Data Mining Chapter 4: Data Analysis and Uncertainty Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University

Bayesian Regression Linear and Logistic Regression

Introduction to Probabilistic Machine Learning

Mathematical Statistics

MAS223 Statistical Inference and Modelling Exercises

INTRODUCTION TO BAYESIAN ANALYSIS

Probability and Measure

Post-exam 2 practice questions 18.05, Spring 2014

INTRODUCTION TO BAYESIAN METHODS II

1 Hypothesis Testing and Model Selection

Estimation of Quantiles

Bayesian Linear Models

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3

Hybrid Censoring; An Introduction 2

Noninformative Priors for the Ratio of the Scale Parameters in the Inverted Exponential Distributions

Hyperparameter estimation in Dirichlet process mixture models

Lecture 1: Introduction

Multiplication of Polynomials

Bayesian Linear Models

USEFUL PROPERTIES OF THE MULTIVARIATE NORMAL*

Inference for Stochastic Processes

Approximate Bayesian computation for spatial extremes via open-faced sandwich adjustment

1 A simple example. A short introduction to Bayesian statistics, part I Math 217 Probability and Statistics Prof. D.

Transcription:

Poisson CI s Robert L. Wolpert Department of Statistical Science Duke University, Durham, NC, USA 1 Interval Estimates Point estimates of unknown parameters θ governing the distribution of an observed quantity X are unsatisfying if they come with no measure of accuracy or precision. One approach to giving such a measure is to offer interval estimates for θ, rather than point estimates; upon observing X, we construct an interval [a,b] which is very likely to contain θ, and which is very short. The approach to exactly how these are constructed and interpreted is different for inference in the Sampling Theory tradition, and in the Bayesian tradition. In these notes I ll present both approaches to estimating the mean of a Poisson random variable on the basis of a random sample of fixed size n. 1.1 Sampling Theory: Confidence Intervals Let {X j } be n iid observations from the Poisson distribution with unknown mean θ, and let 0 < γ < 1 be a number between zero and one. Denote by X the vector of all n components, and by X = Z n + the space of its possible values, non-negative n-tuples of integers; also denote by Θ = R + the possible values of θ. A γ-confidence Interval is a pair of functions A : X R and B : X R with the property that ( θ Θ) P θ [A(X) < θ < B(X)] γ. (1) Notice that this probability is for each fixed θ; it is the endpoints of the interval (A,B) that are random in this calculation, not θ. Let s try to find a γ-confidence Interval that is 1

symmetric in the sense that the two possible errors each have the same error bound, P θ [θ A(X)] 1 γ, P θ [B(X) θ] 1 γ () as short as possible, subject to the error bound. Clearly each function A and B will be a monotonically increasing function of the sufficient statistic S = ΣX j Po(nθ); let s write A S and B S for those functions, and consider A S first. Before we do, remember that the arrival time T k for the k th event in a unit-rate Poisson process X t has the Ga(k,1) distribution, and that X t k if and only if T k t (at least k fish by time t if and only if the k th fish arrives before time t) hence, in R, the CDF functions for Gamma and Poisson are related for all k N and t > 0 by 1 ppois(k 1,t) = pgamma(t,k,1). Also recall the Gamma quantile function in R, an inverse for the CDF function, which satisfies p = pgamma(t,k,b) if and only if t = qgamma(p,k,b), and that if Y Ga(α,β) and b > 0 then Y/b Ga(α,β b), so pgamma(b θ,α,1) = pgamma(θ,α,b). Fix any positive integer k. To achieve Equation () for θ A k, we need: 1 γ P θ [θ A(X)] P θ [S k] = 1 ppois(k 1,nθ) = pgamma(nθ,k,1) = pgamma(θ,k,n), i.e. qgamma( 1 γ,k,n) θ for each θ A k. Evidently this happens if and only if: qgamma( 1 γ,k,n) A k. (3)

Similarly, for nonnegative k and B k θ, Equation () requires: 1 γ P θ [B(X) θ] P θ [S k] = ppois(k, nθ) = 1 pgamma(nθ,k +1,1), i.e. 1 + γ pgamma(θ,k +1,n), i.e. qgamma( 1 + γ,k +1,n) θ for each θ B k. This happens if and only if: qgamma( 1 + γ,k +1,n) B k. (4) The shortest interval subject to the two constraints of Equations (3, 4) is: A k = qgamma( 1 γ,k,n) B k = qgamma( 1 + γ,k +1,n). (5) 1. Bayesian Credible Intervals A conjugate Bayesian analysis for iid Poisson data {X j } iid Po(θ) begins with the selection of parmeters α > 0, β > 0 for a Ga(α,β) prior density π(θ) θ α 1 e βθ and calculation of the likelihood function f(x θ) = n [ θ x j ] x j! e θ j=1 θ S e nθ, where again S = n j=1 X j. The posterior distribution is π(θ x) θ α+s 1 e (β+n)θ Ga(α + S, β + n). 3

P(A(X) < θ, θ < B(X)) 0.90 0.9 0.94 0.96 0.98 1.00 0 0 40 60 80 100 θ Figure 1: Exact coverage probability for 95% Poisson Confidence Intervals Thus a symmetric γ posterior ( credible ) interval for θ can be given by where a(x) = a S and b(x) = b S with γ = P[a(X) < θ < b(x) X] (6) a k = qgamma( 1 γ,α+k,β+n) b k = qgamma( 1+γ,α+k,β+n). (7) 1.3 Comparison The two probability statements in Equations (1, 6) are different in Equation (1) the value of θ is fixed while X (and hence the sufficient statistic S) is random. Because S has a discrete distribution it is not possible to achieve exact equality for all θ; the probability P θ [A(X) < θ < B(X)] (as a function of θ) jumps at each of the points {A k,b k } (see Figure(1)). Instead we guarantee a minumum probability of γ (γ = 0.95 in Figure (1)) that θ will be captured by the interval. In Equation (6), however, X (and hence S) are fixed, and we consider θ to be random; it has a continuous distribution, and it is possible to achieve exact equality. The formulas for the interval endpoints given in Equations (5, 7) are similar if we take β = 0 and α = 1 they will be as close as possible to each other. 4

Note that this corresponds to an improper Ga( 1,0) prior distribution π(θ) θ 1/ 1 {θ>0}, but the posterior distribution π(θ X) Ga(S+ 1,n) is proper for any X X. For any α and β, all the intervals have the same asymptotic behavior for large n; by the central limit theorem, A(X),a(X) X z γ X/n, B(X),b(X) X + zγ X/n where Φ(z γ ) = (1 + γ)/, so γ = Φ(z γ ) Φ( z γ ). 5