Minimax estimators of the coverage probability of the impermissible error for a location family

Size: px
Start display at page:

Download "Minimax estimators of the coverage probability of the impermissible error for a location family"

Transcription

1 Minimax estimators of the coverage probability of the impermissible error for a location family by Miguel A. Arcones Binghamton University arcones@math.binghamton.edu Talk based on: Arcones, M. A. (2008). Minimax estimators of the coverage probability of the impermissible error for a location family. Statistics & Decisions. 28, Available at 1

2 1 Confidence regions with a constrained Lebesgue measure for a location family. Let {X j } j=1 be a sequence of Rd valued i.i.d.r.v. s with a p.d.f. belonging to the family {f( θ) : θ R d }. Fix L > 0. We would like to find a translation equivariant confidence region C HL,L (X 1,..., X n ) such that: 1. For each x := (x 1,..., x n ), R m I(y C HL,L ( x)) dy L. (1) 2. For any confidence region C(X 1,..., X n ) satisfying (1), inf P θ{θ C HL,L ( X)} inf P θ{θ C( X)} θ R d θ R d 2

3 If C( X) is a translation equivariant confidence region, then C( X) = X n + C(Z 1,..., Z n 1, 0) = X n + C ( Z). We would like to find C ( Z) maximizing inf θ R d P θ {θ X n + C ( Z)} = P 0 { X n C ( Z)} (2) = I( x n C ( z)) n 1 j=1 f(z j + x n ) f(x n ) dx n d z = I(y C ( z)) n 1 j=1 f(z j y) f( y) dy d z, subject that for each z, I(y x n +C ( z)) dy = I(y C ( z)) dy L. (3) R m R m Conditioning on z, we get that the optimal region is C HL( z) = {y R d : λ HL ( z) where λ HL ( z) is such that It is easy to see that n 1 j=1 λ HL ( z) n 1 j=1 f(z j y) f( y)) f(z j y) f( y))}. dy = L. C HL,L (x 1,..., x n ) = x n + C HL ( z) = {θ R d : λ HL ( z) n j=1 f(x j θ)}. 3

4 Theorem 1. Suppose that: (i) For each 0 < M <, inf{f(x) : x M} > 0. (ii) lim x f(x) = 0. (iii) sup x R d f(x) <. Let C(X 1,..., X n ) be a confidence region for θ such that for each x 1,..., x n, R d I(y C(x 1,..., x n )) dy L. Then, inf θ R d P θ {θ C HL,L ( X)} inf θ R d P θ {θ C( X)}. 4

5 2 Impermissible error Let T n := T n ( X) be an estimator of θ. The error of the estimator T n is T n θ. Suppose that we select a number b, b > 0, such that any error of estimation bigger than b is impermissible. The maximum coverage probability of the impermissible error of the estimator T n is sup P θ { T n θ > b}. (4) θ Θ We would like to obtain an estimator T HL,b minimizing (4) for a location family. Finding T HL,b is equivalent to find a confidence interval of length 2b with coverage probability as large as possible. If d = 1, and C HL,2b ( X) is an interval, we take T n,hl,b ( X) as the middle point of C HL,2b ( X). By Theorem 1, for each estimator T n, sup P θ { T n θ > b} sup P θ { T n,hl,b θ > b}. θ R θ R 5

6 3 Estimation via the m.l.e. An mle ˆθ n of θ, over a parametric family {f(x, θ) : θ Θ} is a r.v. such that n n 1 log f(x j, ˆθ n ) = sup n 1 θ Θ j=1 n log f(x j, θ). (5) Suppose that there exists a r.v. ˆθn = ˆθ n (X 1,..., X n ) satisfying (5) when f(x, θ) = f(x θ). We may assume that ˆθ n is equivariant. This implies that ˆθ n θ is a pivotal quantity. Suppose that θ ˆθ n has a distribution absolutely continuous with respect to the Lebesgue measure, when θ obtains. Let h n ( ) be the pdf of θ ˆθ n when θ obtains. Let L > 0. Given C B(R d ), a confidence region for θ based on θ ˆθ n is determined by {θ R d : θ ˆθ n C} = {θ R d : θ ˆθ n + C}. j=1 6

7 Between all sets C such that I(θ ˆθ n + C) dθ = a set C which maximizes is P θ {θ ˆθ n C} = C mle,l = {x R d : λ mle,l h n (x)} where λ mle,l is such that {x R d :λ mle,l h n (x)} C C 1 dθ L, (6) h n (θ) dθ (7) 1 dx = L. Hence, using the mle, a confidence region satisfying (6) and maximizing (7) is ˆθ n + C mle,l = {θ R d : λ mle,l h n (θ ˆθ n )}. Notice h n (θ x), x R, is the p.d.f. of ˆθ n when θ obtains. 7

8 The obtained confidence regions maximize the coverage probability over all the regions with expected Lebesgue measure less or equal than L: Theorem 2. Let C(ˆθ n ) be a confidence region for θ such that Then, [ ] sup E θ I(y C(ˆθ n )) dy θ R d R d L. (8) inf P θ{θ C(ˆθ n )} inf P θ{θ ˆθ n + C mle,l }. (9) θ R d θ R d The previous theorem implies that if C(ˆθ n ) is a confidence region for θ such that for each x R d, then, (9) holds. R d I(y C(x)) dy L, (10) 8

9 Suppose that d = 1 and we would like to find the estimator T (ˆθ n ) (based on the m.l.e.) minimizing: sup P θ { T (ˆθ n ) θ > b}. (11) θ R If C mle,2b is an interval, then the middle point T n,mle,b of ˆθ n + C mle,2b minimizes (11) between all the estimators based on the m.l.e. 9

10 4 Asymptotic results. Let {X j } j=1 {f( θ) : θ R}. be a sequence of i.i.d.r.v. s from the family Theorem 3. Suppose that: (i) For each x R, f(x) > 0. (ii) log f( ) is a strictly concave function (iii) For each t R, there exists λ t > 0 such that Then, where E 0 [exp(λ t log f(x t) )] <. lim n n 1 log P θ {θ C HL,L (X 1,..., X n )} = S(L), Besides, S(L) = inf λ R log R (f(x L))λ (f(x)) 1 λ dx. (a) S is increasing on [0, ) and S is decreasing in (, 0]. (b) S is a continuous function. (c) lim t S(t) =. 10

11 Theorem 4. Suppose that: (i) For each x R, f(x) > 0. (ii) log f is a strictly concave function. (iii) R f (x) dx = 0. (iv) E[(f(X)) 2 (f (X)) 2 ] <. Then, the m.l.e. ˆθ n is well defined and θ ˆθ n, when θ obtains, has pdf ( n ) h n (t) := E 0 [I j=1 (f(x j + t)) 1 f (X j + t) > 0 ] n j=1 (f(x j)) 1 f (X j ), t R. Besides, h n is nonincreasing in [0, ) and nondecreasing in (, 0]. 11

12 Theorem 5. Suppose that: (i) For each x R, f(x) > 0. (ii) log f is a concave function. (iii) E 0 [(f(x)) 1 f (X)] = 0. (iv) E 0 [(f(x)) 2 (f (X)) 2 ] <. (v) ˆθ n θ satisfies the LDP with speed n and continuous rate function R(t) := inf λ R log E 0[exp(λ(f(X t)) 1 f (X t))], t R. Then, lim n n 1 log P θ {θ ˆθ n C mle,l (ˆθ n )} = S mle (L) where S mle (L) := inf{u 0 : x:r(x) u 1 dx L}. Besides, (i) For each t R, R(t) <. (ii) R(0) = 0 (iii) R is increasing in [0, ) and decreasing in (, 0]. (iv) R is continuous in R. (v) There exists a t 0 (0, L) such that R(t) = R(t 0 L) = S mle (L). 12

13 Example 1. If f(x) = (2π) 1/2 σ 1 exp( 2 1 σ 2 x 2 ), x R, (12) where σ > 0, then, it is easy to see that: (i) T HL,b = T mle,b = ˆθ n = X := n 1 n j=1 X j. (ii) C HL,2b (X 1,..., X n ) = C mle,2b (ˆθ n ) = [ X b, X + b]. (iii) S mle (2b) = S(2b) = R(b) = 2 1 σ 2 b 2. P (iv) For each θ R, T θ HL,b θ. 13

14 Example 2. If f(x) = (Γ(α)) 1 γ α α exp (αγx αe γx ) (13) where α > 0 and γ 0, then, it is possible to see that: (i) ˆθ ( n = γ 1 log n 1 ) n j=1 eγx j. (ii) C HL,2b (X 1,..., X n ) = C mle,2b (ˆθ n ) = [ˆθ n + γ 1 log ( 2 1 γ 1 b 1 (e 2γb 1) ) 2b, ˆθ n + γ 1 log ( 2 1 γ 1 b 1 (e 2γb 1) ) ]. (iii) T HL,b = T mle,b = ˆθ n +γ 1 log ( 2 1 γ 1 b 1 (e γb e γb ) ). (iv) For each b > 0, S(2b) = S mle (2b) = R(b) = α log(2 1 γ 1 b 1 (e 2 γ b 1)) α(e 2 γ b 1) 1 (e 2 γ b 1 2 γ b). (v) For each b > 0 and each θ R, P T θ HL,b θ + γ 1 log(2 1 γ 1 b 1 (e γb e γb )) θ. By Theorem 2 in Ferguson [1] the location families in examples 1 and 2 are the only one dimensional location families, which are exponential families. 14

15 Example 3. If f(x) = c exp( a 1 exp(τ 1 x) a 2 exp( τ 2 x)), (14) where a 1, a 2, τ 1, τ 2 > 0 and c makes f a p.d.f. Then, (i) ( n j=1 eτ 1X j, n j=1 e τ 2X j ) is a minimal sufficient stat. (ii) The family {f( θ) : θ R} is a curved exponential family. (iii) The m.l.e. ˆθ n of θ is (τ 1 + τ 2 ) 1 log (iv) where a 1 2 τ 1 2 a 1τ 1 n j=1 e τ 2X j 1 n j=1 e τ 1X j C HL,2b (X 1,..., X n ) = C mle,2b (ˆθ n ) = ˆθ n + [t 0 2b, t 0 ], (v) t 0 = (τ 1 + τ 2 ) 1 log ( τ 1 1 τ 2(e 2bτ 1 1)(1 e 2bτ 2 ) 1). T HL,b = T mle,b = ˆθ n +(τ 1 + τ 2 ) 1 log ( τ 1 1 τ 2(e bτ 1 e bτ 1)(e bτ 2 e bτ 2) 1). 15.

16 Example 4. If f(x) = 2 1 exp( x ), x R, (15) then, ˆθ n is not uniquely defined. is concave, but not strictly concave. The function log f( ) Let X (1),..., X (n) be the order statistics. X 1,..., X n are all different wit probability one. Assume that X 1,..., X n are all different. If n is odd, ˆθ n = X (2 1 (n+1)). If n is even, then ˆθ n = X (2 1 n) and ˆθ n = X (2 1 n+1) are both m.l.e. s. It is easy to see that whatever choice for the m.l.e. is made theorems 1 5 apply giving that (i) C HL,2b (X 1,..., X n ) C mle,2b (ˆθ n ). (ii) For each b > 0, S(2b) = b log(1 + b). (iii) For each b > 0, S mle (2b) = b 2 1 log ( 2e b 1 ). (iii) For each b > 0, S mle (2b) < S(2b). P (iv) For each θ R, T θ HL,b θ. 16

17 Theorem 6. Suppose that: (i) For each x R, f(x) > 0. (ii) log f is a concave function. (iii) f is even. Then, (i) For each b > 0, R(b) S(2b). (ii) Assume that f has a third derivative and that f is not a p.d.f. from the families of p.d.f. s in examples 1 and 3, then for some b 0, R(b) < S(2b). References [1] Ferguson, T. S. (1962). Location and scale parameters in exponential families of distributions. Ann. Mathemat. Statist

2014/2015 Smester II ST5224 Final Exam Solution

2014/2015 Smester II ST5224 Final Exam Solution 014/015 Smester II ST54 Final Exam Solution 1 Suppose that (X 1,, X n ) is a random sample from a distribution with probability density function f(x; θ) = e (x θ) I [θ, ) (x) (i) Show that the family of

More information

STAT 512 sp 2018 Summary Sheet

STAT 512 sp 2018 Summary Sheet STAT 5 sp 08 Summary Sheet Karl B. Gregory Spring 08. Transformations of a random variable Let X be a rv with support X and let g be a function mapping X to Y with inverse mapping g (A = {x X : g(x A}

More information

Definition 1.1 (Parametric family of distributions) A parametric distribution is a set of distribution functions, each of which is determined by speci

Definition 1.1 (Parametric family of distributions) A parametric distribution is a set of distribution functions, each of which is determined by speci Definition 1.1 (Parametric family of distributions) A parametric distribution is a set of distribution functions, each of which is determined by specifying one or more values called parameters. The number

More information

Chapter 7. Confidence Sets Lecture 30: Pivotal quantities and confidence sets

Chapter 7. Confidence Sets Lecture 30: Pivotal quantities and confidence sets Chapter 7. Confidence Sets Lecture 30: Pivotal quantities and confidence sets Confidence sets X: a sample from a population P P. θ = θ(p): a functional from P to Θ R k for a fixed integer k. C(X): a confidence

More information

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3 Hypothesis Testing CB: chapter 8; section 0.3 Hypothesis: statement about an unknown population parameter Examples: The average age of males in Sweden is 7. (statement about population mean) The lowest

More information

Master s Written Examination

Master s Written Examination Master s Written Examination Option: Statistics and Probability Spring 016 Full points may be obtained for correct answers to eight questions. Each numbered question which may have several parts is worth

More information

Large deviations of empirical processes

Large deviations of empirical processes Large deviations of empirical processes Miguel A. Arcones Abstract. We give necessary and sufficient conditions for the large deviations of empirical processes and of Banach space valued random vectors.

More information

Section 8: Asymptotic Properties of the MLE

Section 8: Asymptotic Properties of the MLE 2 Section 8: Asymptotic Properties of the MLE In this part of the course, we will consider the asymptotic properties of the maximum likelihood estimator. In particular, we will study issues of consistency,

More information

Lecture 17: Likelihood ratio and asymptotic tests

Lecture 17: Likelihood ratio and asymptotic tests Lecture 17: Likelihood ratio and asymptotic tests Likelihood ratio When both H 0 and H 1 are simple (i.e., Θ 0 = {θ 0 } and Θ 1 = {θ 1 }), Theorem 6.1 applies and a UMP test rejects H 0 when f θ1 (X) f

More information

A Very Brief Summary of Bayesian Inference, and Examples

A Very Brief Summary of Bayesian Inference, and Examples A Very Brief Summary of Bayesian Inference, and Examples Trinity Term 009 Prof Gesine Reinert Our starting point are data x = x 1, x,, x n, which we view as realisations of random variables X 1, X,, X

More information

Testing Algebraic Hypotheses

Testing Algebraic Hypotheses Testing Algebraic Hypotheses Mathias Drton Department of Statistics University of Chicago 1 / 18 Example: Factor analysis Multivariate normal model based on conditional independence given hidden variable:

More information

Statistical Inference

Statistical Inference Statistical Inference Classical and Bayesian Methods Revision Class for Midterm Exam AMS-UCSC Th Feb 9, 2012 Winter 2012. Session 1 (Revision Class) AMS-132/206 Th Feb 9, 2012 1 / 23 Topics Topics We will

More information

Lecture 26: Likelihood ratio tests

Lecture 26: Likelihood ratio tests Lecture 26: Likelihood ratio tests Likelihood ratio When both H 0 and H 1 are simple (i.e., Θ 0 = {θ 0 } and Θ 1 = {θ 1 }), Theorem 6.1 applies and a UMP test rejects H 0 when f θ1 (X) f θ0 (X) > c 0 for

More information

An exponential family of distributions is a parametric statistical model having densities with respect to some positive measure λ of the form.

An exponential family of distributions is a parametric statistical model having densities with respect to some positive measure λ of the form. Stat 8112 Lecture Notes Asymptotics of Exponential Families Charles J. Geyer January 23, 2013 1 Exponential Families An exponential family of distributions is a parametric statistical model having densities

More information

Statistical Inference

Statistical Inference Statistical Inference Robert L. Wolpert Institute of Statistics and Decision Sciences Duke University, Durham, NC, USA. Asymptotic Inference in Exponential Families Let X j be a sequence of independent,

More information

Review and continuation from last week Properties of MLEs

Review and continuation from last week Properties of MLEs Review and continuation from last week Properties of MLEs As we have mentioned, MLEs have a nice intuitive property, and as we have seen, they have a certain equivariance property. We will see later that

More information

STAT215: Solutions for Homework 2

STAT215: Solutions for Homework 2 STAT25: Solutions for Homework 2 Due: Wednesday, Feb 4. (0 pt) Suppose we take one observation, X, from the discrete distribution, x 2 0 2 Pr(X x θ) ( θ)/4 θ/2 /2 (3 θ)/2 θ/4, 0 θ Find an unbiased estimator

More information

Econometrics I, Estimation

Econometrics I, Estimation Econometrics I, Estimation Department of Economics Stanford University September, 2008 Part I Parameter, Estimator, Estimate A parametric is a feature of the population. An estimator is a function of the

More information

Lecture 32: Asymptotic confidence sets and likelihoods

Lecture 32: Asymptotic confidence sets and likelihoods Lecture 32: Asymptotic confidence sets and likelihoods Asymptotic criterion In some problems, especially in nonparametric problems, it is difficult to find a reasonable confidence set with a given confidence

More information

) ) = γ. and P ( X. B(a, b) = Γ(a)Γ(b) Γ(a + b) ; (x + y, ) I J}. Then, (rx) a 1 (ry) b 1 e (x+y)r r 2 dxdy Γ(a)Γ(b) D

) ) = γ. and P ( X. B(a, b) = Γ(a)Γ(b) Γ(a + b) ; (x + y, ) I J}. Then, (rx) a 1 (ry) b 1 e (x+y)r r 2 dxdy Γ(a)Γ(b) D 3 Independent Random Variables II: Examples 3.1 Some functions of independent r.v. s. Let X 1, X 2,... be independent r.v. s with the known distributions. Then, one can compute the distribution of a r.v.

More information

ECE598: Information-theoretic methods in high-dimensional statistics Spring 2016

ECE598: Information-theoretic methods in high-dimensional statistics Spring 2016 ECE598: Information-theoretic methods in high-dimensional statistics Spring 06 Lecture : Mutual Information Method Lecturer: Yihong Wu Scribe: Jaeho Lee, Mar, 06 Ed. Mar 9 Quick review: Assouad s lemma

More information

Chapter 2. Binary and M-ary Hypothesis Testing 2.1 Introduction (Levy 2.1)

Chapter 2. Binary and M-ary Hypothesis Testing 2.1 Introduction (Levy 2.1) Chapter 2. Binary and M-ary Hypothesis Testing 2.1 Introduction (Levy 2.1) Detection problems can usually be casted as binary or M-ary hypothesis testing problems. Applications: This chapter: Simple hypothesis

More information

Maximum Likelihood Estimation

Maximum Likelihood Estimation Maximum Likelihood Estimation Assume X P θ, θ Θ, with joint pdf (or pmf) f(x θ). Suppose we observe X = x. The Likelihood function is L(θ x) = f(x θ) as a function of θ (with the data x held fixed). The

More information

Statistics Ph.D. Qualifying Exam

Statistics Ph.D. Qualifying Exam Department of Statistics Carnegie Mellon University May 7 2008 Statistics Ph.D. Qualifying Exam You are not expected to solve all five problems. Complete solutions to few problems will be preferred to

More information

lim n C1/n n := ρ. [f(y) f(x)], y x =1 [f(x) f(y)] [g(x) g(y)]. (x,y) E A E(f, f),

lim n C1/n n := ρ. [f(y) f(x)], y x =1 [f(x) f(y)] [g(x) g(y)]. (x,y) E A E(f, f), 1 Part I Exercise 1.1. Let C n denote the number of self-avoiding random walks starting at the origin in Z of length n. 1. Show that (Hint: Use C n+m C n C m.) lim n C1/n n = inf n C1/n n := ρ.. Show that

More information

Lecture 25: Review. Statistics 104. April 23, Colin Rundel

Lecture 25: Review. Statistics 104. April 23, Colin Rundel Lecture 25: Review Statistics 104 Colin Rundel April 23, 2012 Joint CDF F (x, y) = P [X x, Y y] = P [(X, Y ) lies south-west of the point (x, y)] Y (x,y) X Statistics 104 (Colin Rundel) Lecture 25 April

More information

Optimization. The value x is called a maximizer of f and is written argmax X f. g(λx + (1 λ)y) < λg(x) + (1 λ)g(y) 0 < λ < 1; x, y X.

Optimization. The value x is called a maximizer of f and is written argmax X f. g(λx + (1 λ)y) < λg(x) + (1 λ)g(y) 0 < λ < 1; x, y X. Optimization Background: Problem: given a function f(x) defined on X, find x such that f(x ) f(x) for all x X. The value x is called a maximizer of f and is written argmax X f. In general, argmax X f may

More information

Asymptotics of minimax stochastic programs

Asymptotics of minimax stochastic programs Asymptotics of minimax stochastic programs Alexander Shapiro Abstract. We discuss in this paper asymptotics of the sample average approximation (SAA) of the optimal value of a minimax stochastic programming

More information

Invariant HPD credible sets and MAP estimators

Invariant HPD credible sets and MAP estimators Bayesian Analysis (007), Number 4, pp. 681 69 Invariant HPD credible sets and MAP estimators Pierre Druilhet and Jean-Michel Marin Abstract. MAP estimators and HPD credible sets are often criticized in

More information

Homework 27. Homework 28. Homework 29. Homework 30. Prof. Girardi, Math 703, Fall 2012 Homework: Define f : C C and u, v : R 2 R by

Homework 27. Homework 28. Homework 29. Homework 30. Prof. Girardi, Math 703, Fall 2012 Homework: Define f : C C and u, v : R 2 R by Homework 27 Define f : C C and u, v : R 2 R by f(z) := xy where x := Re z, y := Im z u(x, y) = Re f(x + iy) v(x, y) = Im f(x + iy). Show that 1. u and v satisfies the Cauchy Riemann equations at (x, y)

More information

STAT 135 Lab 2 Confidence Intervals, MLE and the Delta Method

STAT 135 Lab 2 Confidence Intervals, MLE and the Delta Method STAT 135 Lab 2 Confidence Intervals, MLE and the Delta Method Rebecca Barter February 2, 2015 Confidence Intervals Confidence intervals What is a confidence interval? A confidence interval is calculated

More information

Brief Review on Estimation Theory

Brief Review on Estimation Theory Brief Review on Estimation Theory K. Abed-Meraim ENST PARIS, Signal and Image Processing Dept. abed@tsi.enst.fr This presentation is essentially based on the course BASTA by E. Moulines Brief review on

More information

Statistical Inference

Statistical Inference Statistical Inference Robert L. Wolpert Institute of Statistics and Decision Sciences Duke University, Durham, NC, USA Week 12. Testing and Kullback-Leibler Divergence 1. Likelihood Ratios Let 1, 2, 2,...

More information

5.1 Uniformly Most Accurate Families of Confidence

5.1 Uniformly Most Accurate Families of Confidence Chapter 5 Confidence Estimation Let (X, B X, P ), P {P θ θ Θ} be sample space and {Θ, B Θ, µ} be measure space with some σ finite measure µ. 5.1 Uniformly Most Accurate Families of Confidence Sets Def.

More information

Chapter 4. Theory of Tests. 4.1 Introduction

Chapter 4. Theory of Tests. 4.1 Introduction Chapter 4 Theory of Tests 4.1 Introduction Parametric model: (X, B X, P θ ), P θ P = {P θ θ Θ} where Θ = H 0 +H 1 X = K +A : K: critical region = rejection region / A: acceptance region A decision rule

More information

Statistics. Lecture 4 August 9, 2000 Frank Porter Caltech. 1. The Fundamentals; Point Estimation. 2. Maximum Likelihood, Least Squares and All That

Statistics. Lecture 4 August 9, 2000 Frank Porter Caltech. 1. The Fundamentals; Point Estimation. 2. Maximum Likelihood, Least Squares and All That Statistics Lecture 4 August 9, 2000 Frank Porter Caltech The plan for these lectures: 1. The Fundamentals; Point Estimation 2. Maximum Likelihood, Least Squares and All That 3. What is a Confidence Interval?

More information

1 Likelihood. 1.1 Likelihood function. Likelihood & Maximum Likelihood Estimators

1 Likelihood. 1.1 Likelihood function. Likelihood & Maximum Likelihood Estimators Likelihood & Maximum Likelihood Estimators February 26, 2018 Debdeep Pati 1 Likelihood Likelihood is surely one of the most important concepts in statistical theory. We have seen the role it plays in sufficiency,

More information

Stat 5102 Lecture Slides Deck 3. Charles J. Geyer School of Statistics University of Minnesota

Stat 5102 Lecture Slides Deck 3. Charles J. Geyer School of Statistics University of Minnesota Stat 5102 Lecture Slides Deck 3 Charles J. Geyer School of Statistics University of Minnesota 1 Likelihood Inference We have learned one very general method of estimation: method of moments. the Now we

More information

Optimum Joint Detection and Estimation

Optimum Joint Detection and Estimation 20 IEEE International Symposium on Information Theory Proceedings Optimum Joint Detection and Estimation George V. Moustakides Department of Electrical and Computer Engineering University of Patras, 26500

More information

Suggested solutions to written exam Jan 17, 2012

Suggested solutions to written exam Jan 17, 2012 LINKÖPINGS UNIVERSITET Institutionen för datavetenskap Statistik, ANd 73A36 THEORY OF STATISTICS, 6 CDTS Master s program in Statistics and Data Mining Fall semester Written exam Suggested solutions to

More information

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it

More information

Statistics - Lecture One. Outline. Charlotte Wickham 1. Basic ideas about estimation

Statistics - Lecture One. Outline. Charlotte Wickham  1. Basic ideas about estimation Statistics - Lecture One Charlotte Wickham wickham@stat.berkeley.edu http://www.stat.berkeley.edu/~wickham/ Outline 1. Basic ideas about estimation 2. Method of Moments 3. Maximum Likelihood 4. Confidence

More information

Theory of Maximum Likelihood Estimation. Konstantin Kashin

Theory of Maximum Likelihood Estimation. Konstantin Kashin Gov 2001 Section 5: Theory of Maximum Likelihood Estimation Konstantin Kashin February 28, 2013 Outline Introduction Likelihood Examples of MLE Variance of MLE Asymptotic Properties What is Statistical

More information

A BAYESIAN MATHEMATICAL STATISTICS PRIMER. José M. Bernardo Universitat de València, Spain

A BAYESIAN MATHEMATICAL STATISTICS PRIMER. José M. Bernardo Universitat de València, Spain A BAYESIAN MATHEMATICAL STATISTICS PRIMER José M. Bernardo Universitat de València, Spain jose.m.bernardo@uv.es Bayesian Statistics is typically taught, if at all, after a prior exposure to frequentist

More information

Step-Stress Models and Associated Inference

Step-Stress Models and Associated Inference Department of Mathematics & Statistics Indian Institute of Technology Kanpur August 19, 2014 Outline Accelerated Life Test 1 Accelerated Life Test 2 3 4 5 6 7 Outline Accelerated Life Test 1 Accelerated

More information

Likelihood based Statistical Inference. Dottorato in Economia e Finanza Dipartimento di Scienze Economiche Univ. di Verona

Likelihood based Statistical Inference. Dottorato in Economia e Finanza Dipartimento di Scienze Economiche Univ. di Verona Likelihood based Statistical Inference Dottorato in Economia e Finanza Dipartimento di Scienze Economiche Univ. di Verona L. Pace, A. Salvan, N. Sartori Udine, April 2008 Likelihood: observed quantities,

More information

9 Bayesian inference. 9.1 Subjective probability

9 Bayesian inference. 9.1 Subjective probability 9 Bayesian inference 1702-1761 9.1 Subjective probability This is probability regarded as degree of belief. A subjective probability of an event A is assessed as p if you are prepared to stake pm to win

More information

Math 211 Business Calculus TEST 3. Question 1. Section 2.2. Second Derivative Test.

Math 211 Business Calculus TEST 3. Question 1. Section 2.2. Second Derivative Test. Math 211 Business Calculus TEST 3 Question 1. Section 2.2. Second Derivative Test. p. 1/?? Math 211 Business Calculus TEST 3 Question 1. Section 2.2. Second Derivative Test. Question 2. Section 2.3. Graph

More information

Non-parametric Inference and Resampling

Non-parametric Inference and Resampling Non-parametric Inference and Resampling Exercises by David Wozabal (Last update. Juni 010) 1 Basic Facts about Rank and Order Statistics 1.1 10 students were asked about the amount of time they spend surfing

More information

MTH 3102 Complex Variables Solutions: Practice Exam 2 Mar. 26, 2017

MTH 3102 Complex Variables Solutions: Practice Exam 2 Mar. 26, 2017 Name Last name, First name): MTH 31 omplex Variables Solutions: Practice Exam Mar. 6, 17 Exam Instructions: You have 1 hour & 1 minutes to complete the exam. There are a total of 7 problems. You must show

More information

ELEMENTS OF PROBABILITY THEORY

ELEMENTS OF PROBABILITY THEORY ELEMENTS OF PROBABILITY THEORY Elements of Probability Theory A collection of subsets of a set Ω is called a σ algebra if it contains Ω and is closed under the operations of taking complements and countable

More information

HW1 solutions. 1. α Ef(x) β, where Ef(x) is the expected value of f(x), i.e., Ef(x) = n. i=1 p if(a i ). (The function f : R R is given.

HW1 solutions. 1. α Ef(x) β, where Ef(x) is the expected value of f(x), i.e., Ef(x) = n. i=1 p if(a i ). (The function f : R R is given. HW1 solutions Exercise 1 (Some sets of probability distributions.) Let x be a real-valued random variable with Prob(x = a i ) = p i, i = 1,..., n, where a 1 < a 2 < < a n. Of course p R n lies in the standard

More information

STAT 461/561- Assignments, Year 2015

STAT 461/561- Assignments, Year 2015 STAT 461/561- Assignments, Year 2015 This is the second set of assignment problems. When you hand in any problem, include the problem itself and its number. pdf are welcome. If so, use large fonts and

More information

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it

More information

A Very Brief Summary of Statistical Inference, and Examples

A Very Brief Summary of Statistical Inference, and Examples A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2008 Prof. Gesine Reinert 1 Data x = x 1, x 2,..., x n, realisations of random variables X 1, X 2,..., X n with distribution (model)

More information

Chapter 8.8.1: A factorization theorem

Chapter 8.8.1: A factorization theorem LECTURE 14 Chapter 8.8.1: A factorization theorem The characterization of a sufficient statistic in terms of the conditional distribution of the data given the statistic can be difficult to work with.

More information

Statistical Theory MT 2008 Problems 1: Solution sketches

Statistical Theory MT 2008 Problems 1: Solution sketches Statistical Theory MT 008 Problems : Solutio sketches. Which of the followig desities are withi a expoetial family? Explai your reasoig. a) Let 0 < θ < ad put fx, θ) = θ)θ x ; x = 0,,,... b) c) where α

More information

ISyE 6644 Fall 2014 Test 3 Solutions

ISyE 6644 Fall 2014 Test 3 Solutions 1 NAME ISyE 6644 Fall 14 Test 3 Solutions revised 8/4/18 You have 1 minutes for this test. You are allowed three cheat sheets. Circle all final answers. Good luck! 1. [4 points] Suppose that the joint

More information

Statistical Estimation: Data & Non-data Information

Statistical Estimation: Data & Non-data Information Statistical Estimation: Data & Non-data Information Roger J-B Wets University of California, Davis & M.Casey @ Raytheon G.Pflug @ U. Vienna, X. Dong @ EpiRisk, G-M You @ EpiRisk. a little background Decision

More information

Week 2 Spring Lecture 3. The Canonical normal means estimation problem (cont.).! (X) = X+ 1 X X, + (X) = X+ 1

Week 2 Spring Lecture 3. The Canonical normal means estimation problem (cont.).! (X) = X+ 1 X X, + (X) = X+ 1 Week 2 Spring 2009 Lecture 3. The Canonical normal means estimation problem (cont.). Shrink toward a common mean. Theorem. Let X N ; 2 I n. Let 0 < C 2 (n 3) (hence n 4). De ne!! (X) = X+ 1 Then C 2 X

More information

Statistics 3858 : Maximum Likelihood Estimators

Statistics 3858 : Maximum Likelihood Estimators Statistics 3858 : Maximum Likelihood Estimators 1 Method of Maximum Likelihood In this method we construct the so called likelihood function, that is L(θ) = L(θ; X 1, X 2,..., X n ) = f n (X 1, X 2,...,

More information

Part III. A Decision-Theoretic Approach and Bayesian testing

Part III. A Decision-Theoretic Approach and Bayesian testing Part III A Decision-Theoretic Approach and Bayesian testing 1 Chapter 10 Bayesian Inference as a Decision Problem The decision-theoretic framework starts with the following situation. We would like to

More information

Mathematical statistics

Mathematical statistics October 1 st, 2018 Lecture 11: Sufficient statistic Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation

More information

Gibbs Voronoi tessellations: Modeling, Simulation, Estimation.

Gibbs Voronoi tessellations: Modeling, Simulation, Estimation. Gibbs Voronoi tessellations: Modeling, Simulation, Estimation. Frédéric Lavancier, Laboratoire Jean Leray, Nantes (France) Work with D. Dereudre (LAMAV, Valenciennes, France). SPA, Berlin, July 27-31,

More information

UNIVERSITY OF MANITOBA

UNIVERSITY OF MANITOBA Question Points Score INSTRUCTIONS TO STUDENTS: This is a 6 hour examination. No extra time will be given. No texts, notes, or other aids are permitted. There are no calculators, cellphones or electronic

More information

Lecture 7 Introduction to Statistical Decision Theory

Lecture 7 Introduction to Statistical Decision Theory Lecture 7 Introduction to Statistical Decision Theory I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 20, 2016 1 / 55 I-Hsiang Wang IT Lecture 7

More information

Solutions to Exam 1, Math Solution. Because f(x) is one-to-one, we know the inverse function exists. Recall that (f 1 ) (a) =

Solutions to Exam 1, Math Solution. Because f(x) is one-to-one, we know the inverse function exists. Recall that (f 1 ) (a) = Solutions to Exam, Math 56 The function f(x) e x + x 3 + x is one-to-one (there is no need to check this) What is (f ) ( + e )? Solution Because f(x) is one-to-one, we know the inverse function exists

More information

Solutions to the Exercises of Section 2.11.

Solutions to the Exercises of Section 2.11. Solutions to the Exercises of Section 2.11. 2.11.1. Proof. Let ɛ be an arbitrary positive number. Since r(τ n,δ n ) C, we can find an integer n such that r(τ n,δ n ) C ɛ. Then, as in the proof of Theorem

More information

On Consistent Hypotheses Testing

On Consistent Hypotheses Testing Mikhail Ermakov St.Petersburg E-mail: erm2512@gmail.com History is rather distinctive. Stein (1954); Hodges (1954); Hoefding and Wolfowitz (1956), Le Cam and Schwartz (1960). special original setups: Shepp,

More information

Statistical Theory MT 2009 Problems 1: Solution sketches

Statistical Theory MT 2009 Problems 1: Solution sketches Statistical Theory MT 009 Problems : Solutio sketches. Which of the followig desities are withi a expoetial family? Explai your reasoig. (a) Let 0 < θ < ad put f(x, θ) = ( θ)θ x ; x = 0,,,... (b) (c) where

More information

Choosing among models

Choosing among models Eco 515 Fall 2014 Chris Sims Choosing among models September 18, 2014 c 2014 by Christopher A. Sims. This document is licensed under the Creative Commons Attribution-NonCommercial-ShareAlike 3.0 Unported

More information

Statistical Theory MT 2007 Problems 4: Solution sketches

Statistical Theory MT 2007 Problems 4: Solution sketches Statistical Theory MT 007 Problems 4: Solution sketches 1. Consider a 1-parameter exponential family model with density f(x θ) = f(x)g(θ)exp{cφ(θ)h(x)}, x X. Suppose that the prior distribution has the

More information

MIT Spring 2016

MIT Spring 2016 MIT 18.655 Dr. Kempthorne Spring 2016 1 MIT 18.655 Outline 1 2 MIT 18.655 Decision Problem: Basic Components P = {P θ : θ Θ} : parametric model. Θ = {θ}: Parameter space. A{a} : Action space. L(θ, a) :

More information

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed

More information

Hypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes

Hypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes Neyman-Pearson paradigm. Suppose that a researcher is interested in whether the new drug works. The process of determining whether the outcome of the experiment points to yes or no is called hypothesis

More information

Chapter 4: Asymptotic Properties of the MLE

Chapter 4: Asymptotic Properties of the MLE Chapter 4: Asymptotic Properties of the MLE Daniel O. Scharfstein 09/19/13 1 / 1 Maximum Likelihood Maximum likelihood is the most powerful tool for estimation. In this part of the course, we will consider

More information

Likelihood inference in the presence of nuisance parameters

Likelihood inference in the presence of nuisance parameters Likelihood inference in the presence of nuisance parameters Nancy Reid, University of Toronto www.utstat.utoronto.ca/reid/research 1. Notation, Fisher information, orthogonal parameters 2. Likelihood inference

More information

Brownian survival and Lifshitz tail in perturbed lattice disorder

Brownian survival and Lifshitz tail in perturbed lattice disorder Brownian survival and Lifshitz tail in perturbed lattice disorder Ryoki Fukushima Kyoto niversity Random Processes and Systems February 16, 2009 6 B T 1. Model ) ({B t t 0, P x : standard Brownian motion

More information

A Kullback-Leibler Divergence for Bayesian Model Comparison with Applications to Diabetes Studies. Chen-Pin Wang, UTHSCSA Malay Ghosh, U.

A Kullback-Leibler Divergence for Bayesian Model Comparison with Applications to Diabetes Studies. Chen-Pin Wang, UTHSCSA Malay Ghosh, U. A Kullback-Leibler Divergence for Bayesian Model Comparison with Applications to Diabetes Studies Chen-Pin Wang, UTHSCSA Malay Ghosh, U. Florida Lehmann Symposium, May 9, 2011 1 Background KLD: the expected

More information

Probability Theory and Statistics. Peter Jochumzen

Probability Theory and Statistics. Peter Jochumzen Probability Theory and Statistics Peter Jochumzen April 18, 2016 Contents 1 Probability Theory And Statistics 3 1.1 Experiment, Outcome and Event................................ 3 1.2 Probability............................................

More information

HARMONIC ANALYSIS. Date:

HARMONIC ANALYSIS. Date: HARMONIC ANALYSIS Contents. Introduction 2. Hardy-Littlewood maximal function 3. Approximation by convolution 4. Muckenhaupt weights 4.. Calderón-Zygmund decomposition 5. Fourier transform 6. BMO (bounded

More information

Multivariate Analysis and Likelihood Inference

Multivariate Analysis and Likelihood Inference Multivariate Analysis and Likelihood Inference Outline 1 Joint Distribution of Random Variables 2 Principal Component Analysis (PCA) 3 Multivariate Normal Distribution 4 Likelihood Inference Joint density

More information

INDIAN INSTITUTE OF TECHNOLOGY BOMBAY MA205 Complex Analysis Autumn 2012

INDIAN INSTITUTE OF TECHNOLOGY BOMBAY MA205 Complex Analysis Autumn 2012 INDIAN INSTITUTE OF TECHNOLOGY BOMBAY MA205 Complex Analysis Autumn 2012 September 5, 2012 Mapping Properties Lecture 13 We shall once again return to the study of general behaviour of holomorphic functions

More information

Prelim Examination 2010 / 2011 (Assessing Units 1 & 2) MATHEMATICS. Advanced Higher Grade. Time allowed - 2 hours

Prelim Examination 2010 / 2011 (Assessing Units 1 & 2) MATHEMATICS. Advanced Higher Grade. Time allowed - 2 hours Prelim Examination 00 / 0 (Assessing Units & ) MATHEMATICS Advanced Higher Grade Time allowed - hours Read Carefully. Calculators may be used in this paper.. Candidates should answer all questions. Full

More information

Pointwise convergence rates and central limit theorems for kernel density estimators in linear processes

Pointwise convergence rates and central limit theorems for kernel density estimators in linear processes Pointwise convergence rates and central limit theorems for kernel density estimators in linear processes Anton Schick Binghamton University Wolfgang Wefelmeyer Universität zu Köln Abstract Convergence

More information

Information in Data. Sufficiency, Ancillarity, Minimality, and Completeness

Information in Data. Sufficiency, Ancillarity, Minimality, and Completeness Information in Data Sufficiency, Ancillarity, Minimality, and Completeness Important properties of statistics that determine the usefulness of those statistics in statistical inference. These general properties

More information

Berge s Maximum Theorem

Berge s Maximum Theorem Berge s Maximum Theorem References: Acemoglu, Appendix A.6 Stokey-Lucas-Prescott, Section 3.3 Ok, Sections E.1-E.3 Claude Berge, Topological Spaces (1963), Chapter 6 Berge s Maximum Theorem So far, we

More information

We say that the function f obtains a maximum value provided that there. We say that the function f obtains a minimum value provided that there

We say that the function f obtains a maximum value provided that there. We say that the function f obtains a minimum value provided that there Math 311 W08 Day 10 Section 3.2 Extreme Value Theorem (It s EXTREME!) 1. Definition: For a function f: D R we define the image of the function to be the set f(d) = {y y = f(x) for some x in D} We say that

More information

Generation from simple discrete distributions

Generation from simple discrete distributions S-38.3148 Simulation of data networks / Generation of random variables 1(18) Generation from simple discrete distributions Note! This is just a more clear and readable version of the same slide that was

More information

Applied Statistics Lecture Notes

Applied Statistics Lecture Notes Applied Statistics Lecture Notes Kosuke Imai Department of Politics Princeton University February 2, 2008 Making statistical inferences means to learn about what you do not observe, which is called parameters,

More information

Statistical Theory MT 2006 Problems 4: Solution sketches

Statistical Theory MT 2006 Problems 4: Solution sketches Statistical Theory MT 006 Problems 4: Solution sketches 1. Suppose that X has a Poisson distribution with unknown mean θ. Determine the conjugate prior, and associate posterior distribution, for θ. Determine

More information

is a Borel subset of S Θ for each c R (Bertsekas and Shreve, 1978, Proposition 7.36) This always holds in practical applications.

is a Borel subset of S Θ for each c R (Bertsekas and Shreve, 1978, Proposition 7.36) This always holds in practical applications. Stat 811 Lecture Notes The Wald Consistency Theorem Charles J. Geyer April 9, 01 1 Analyticity Assumptions Let { f θ : θ Θ } be a family of subprobability densities 1 with respect to a measure µ on a measurable

More information

Lecture 35: December The fundamental statistical distances

Lecture 35: December The fundamental statistical distances 36-705: Intermediate Statistics Fall 207 Lecturer: Siva Balakrishnan Lecture 35: December 4 Today we will discuss distances and metrics between distributions that are useful in statistics. I will be lose

More information

Statistics: Learning models from data

Statistics: Learning models from data DS-GA 1002 Lecture notes 5 October 19, 2015 Statistics: Learning models from data Learning models from data that are assumed to be generated probabilistically from a certain unknown distribution is a crucial

More information

3 a = 3 b c 2 = a 2 + b 2 = 2 2 = 4 c 2 = 3b 2 + b 2 = 4b 2 = 4 b 2 = 1 b = 1 a = 3b = 3. x 2 3 y2 1 = 1.

3 a = 3 b c 2 = a 2 + b 2 = 2 2 = 4 c 2 = 3b 2 + b 2 = 4b 2 = 4 b 2 = 1 b = 1 a = 3b = 3. x 2 3 y2 1 = 1. MATH 222 LEC SECOND MIDTERM EXAM THU NOV 8 PROBLEM ( 5 points ) Find the standard-form equation for the hyperbola which has its foci at F ± (±2, ) and whose asymptotes are y ± 3 x The calculations b a

More information

Hypothesis testing: theory and methods

Hypothesis testing: theory and methods Statistical Methods Warsaw School of Economics November 3, 2017 Statistical hypothesis is the name of any conjecture about unknown parameters of a population distribution. The hypothesis should be verifiable

More information

4 Invariant Statistical Decision Problems

4 Invariant Statistical Decision Problems 4 Invariant Statistical Decision Problems 4.1 Invariant decision problems Let G be a group of measurable transformations from the sample space X into itself. The group operation is composition. Note that

More information

BEST TESTS. Abstract. We will discuss the Neymann-Pearson theorem and certain best test where the power function is optimized.

BEST TESTS. Abstract. We will discuss the Neymann-Pearson theorem and certain best test where the power function is optimized. BEST TESTS Abstract. We will discuss the Neymann-Pearson theorem and certain best test where the power function is optimized. 1. Most powerful test Let {f θ } θ Θ be a family of pdfs. We will consider

More information

Lecture 4 September 15

Lecture 4 September 15 IFT 6269: Probabilistic Graphical Models Fall 2017 Lecture 4 September 15 Lecturer: Simon Lacoste-Julien Scribe: Philippe Brouillard & Tristan Deleu 4.1 Maximum Likelihood principle Given a parametric

More information

Statistics GIDP Ph.D. Qualifying Exam Theory Jan 11, 2016, 9:00am-1:00pm

Statistics GIDP Ph.D. Qualifying Exam Theory Jan 11, 2016, 9:00am-1:00pm Statistics GIDP Ph.D. Qualifying Exam Theory Jan, 06, 9:00am-:00pm Instructions: Provide answers on the supplied pads of paper; write on only one side of each sheet. Complete exactly 5 of the 6 problems.

More information