Lecture 5. i=1 xi. Ω h(x,y)f X Θ(y θ)µ Θ (dθ) = dµ Θ X

Size: px
Start display at page:

Download "Lecture 5. i=1 xi. Ω h(x,y)f X Θ(y θ)µ Θ (dθ) = dµ Θ X"

Transcription

1 LECURE NOES 25 Lecture 5 9. Minimal sufficient and complete statistics We introduced the notion of sufficient statistics in order to have a function of the data that contains all information about the parameter. However, a sufficient statistic does not have to be any simpler than the data itself. As we have seen, the identity function is a sufficient statistic so this choice does not simplify or summarize anything. A statistic is said to be minimal sufficient if it is as simple as possible in a certain sense. Here is a definition. Definition 11. A sufficient statistic : X is minimal sufficient if for any sufficient statistic U : X U there is a measurable function g : U such that = g(u) µ X Θ ( θ)-a.s. for all θ Ω. How do we check if a statistic is minimal sufficient? It can be inconvenient to check the condition in the definition for all sufficient statistics U. heorem 10. If there exist version of f X Θ (x θ) for each θ and a measurable function : X such that (x) = (y) y D(x), where D(x) = {y X : f X Θ (y θ) = f X Θ (x θ)h(x,y), θ and some function h(x,y) > 0}, then is a minimal sufficient statistic. Example 14. Let {X n } be IID Exp(θ) given Θ = θ and X = (X 1,...,X n ). Put (x) = x 1 + +x n. Let us show is minimal sufficient. he ratio n i=1 xi f X Θ (x θ) f X Θ (y θ) = θn e θ θ n e θ n i=1 yi does not depend on θ if and only if n i=1 x i = n i=1 y i. In this case h(x,y) = 1, D(x) = {y : n i=1 x i = n i=1 y i}, and is minimal sufficient. Proof. Note first that the sets D(x) form a partition of X. Indeed, by putting h(y,x) = 1/h(x,y) we see that y D(x) implies x D(y). Similarly, taking h(x,x) = 1, we see that x D(x) and hence, the different D(x) form a partition. he condition says that the sets D(x) coincide with sets 1 {(x)} and hence D(x) B for each x. By Bayes theorem we have, for y D(x), dµ Θ X f X Θ (x θ) (θ x) = dµ Θ Ω f X Θ(x θ)µ Θ (dθ) = h(x,y)f X Θ(y θ) Ω h(x,y)f X Θ(y θ)µ Θ (dθ) = dµ Θ X (θ y). dµ Θ hat is, the posterior density is constant on D(x). Hence, it is a function of (x) and by Lemma 1 is sufficient. Let us check that is also minimal. ake U : X U to be a sufficient statistic. If we show that U(x) = U(y) implies y D(x), then it follows that U(x) = U(y) implies (x) = (y) and hence that is a function of U(x). hen is minimal. By the factorization theorem (heorem 2, Lecture 6) f X Θ (x θ) = h(x)g(θ,u(x)). We can assume that h(x) > 0 because P θ ({x : h(x) = 0}) = 0. Hence, U(x) = U(y) implies f X Θ (y θ) = h(y) h(x) g(θ,u(x)). hat is, y D(x) with h(x,y) = h(y)/h(x).

2 26 HENRIK HUL he next concept is that of a complete statistic. Definition 12. Let : X be a statistic and {µ Θ ( θ),θ Ω} the family of conditional distributions of (X) given Θ = θ. he family {µ Θ ( θ),θ Ω} is said to be complete if for each measurable function g, E θ [g()] = 0, θ implies P θ (g() = 0) = 1, θ. he family {µ Θ ( θ),θ Ω} is said to be boundedly complete if each bounded measurable function g, E θ [g()] = 0, θ implies P θ (g() = 0) = 1, θ. A statistic is said to be complete if the family {µ Θ ( θ),θ Ω} is complete. A statistic is said to be boundedly complete if the family {µ Θ ( θ),θ Ω} is boundedly complete. Oneshouldnotethatcompletenessisastatementabouttheentirefamily{µ Θ ( θ),θ Ω} and not only about the individual conditional distributions µ Θ ( θ). Example 15. Suppose that has Bin(n,θ) distribution with θ (0,1) and g is a function such that E θ [g()] = 0 θ. hen n ( ) n n ( ) n ( θ ) k. 0 = E θ [g()] = g(k) θ k (1 θ) n k = (1 θ) n g(k) k k 1 θ k=0 If we put r = θ/(1 θ) we see that this equals n ( ) n (1 θ) n g(k) r k k k=0 which is a polynomial in r of degree n. Since this is constant equal to 0 for all r > 0 it must be that g(k) ( n k) = 0 for eachk = 0,...,n, i.e. g(k) = 0 for eachk = 0,...,n. Since, for each θ, is supported on {0,...,n} it follows that P θ (g() = 0) = 1 θ so is complete. k=0 An important result for exponential families is the following. heorem 11. If the natural parameter space Ω of an exponential family contains an open set in R k, then (X) is a complete sufficient statistic. Proof. We will give a proof for k = 1. For larger k one can use induction. We know that the natural statistic has a density c(θ)e θt with respect to ν (see Section 4.2, Lecture 4). Let g be a measurable function such that E θ [g()] = 0 for all θ. hat is, g(t)c(θ)e θt ν (dt) = 0 θ. If we write g + and g for the positive and negative part of g, respectively, then this says g + (t)c(θ)e θt ν (dt) = g (t)c(θ)e θt ν (dt) θ. (9.1) ake a fixed value θ 0 in the interior of Ω. his is possible since Ω contains an open set. Put Z 0 = g + (t)c(θ 0 )e θ0t ν (dt) = g (t)c(θ 0 )e θ0t ν (dt)

3 LECURE NOES 27 and define the probability measures P and Q by P(C) = Z0 1 g + (t)c(θ 0 )e θ0t ν (dt) C Q(C) = Z0 1 g (t)c(θ 0 )e θ0t ν (dt). hen, the equality (9.1) can be written exp{t(θ θ 0 )}P(dt) = C exp{t(θ θ 0 }Q(dt), θ. With u = θ θ 0 we see that this implies that the moment generating function of P, M P (u), equals the mgf of Q, M Q (u) in a neighborhood of u = 0. Hence, by uniqueness ofthe moment generatingfunction P = Q. It follows that g + (t) = g (t) ν -a.e. and hence that µ Θ{t : g(t) = 0 θ} = 1 for all θ. Hence, is complete sufficient statistic. Completeness of a statistic is also related to minimal sufficiency. heorem 12 (Bahadur s theorem). If is a finite-dimensional boundedly complete sufficient statistic, then it is minimal sufficient. Proof. Let U be an arbitrary sufficient statistic. We will show that is a function of U by constructing the appropriate function. Put = ( 1 (X),..., k (X)) and S i () = [1+e i ] 1 so that S i is bounded and bijective. Let X i (u) = E θ [S i () U = u], Y i (t) = E θ [X i (U) = t]. We want to show that S i () = X i (U) P θ -a.s. for all θ. hen, since S i is bijective we have i = S 1 i (X i (U)) and the claim follows. We show S i () = X i (U) P θ -a.s. in two steps. First step: S i () = Y i () P θ -a.s. for all θ. o see this note that E θ [Y i ()] = E θ [E θ [X i (U) ]] = E θ [X i (U)] = E θ [E θ [S i () U]] = E θ [S i ()]. Hence, forallθ, E θ [Y i () S i ()] = 0andsinceS i is bounded, soisy i andbounded completeness implies P θ (S i () = Y i ()) = 1 for all θ. Second step: X i (U) = Y i () P θ -a.s. for all θ. By step one we have E θ [Y i () U] = X i (U) P θ -a.s. So if we show that the conditional variance of Y i () given U is zero we are done. hat is, we need to show Var θ (Y i () U) = 0 P θ -a.s. By the usual rule for conditional variance (heorem B.78 p. 634) Var θ (Y i ()) = E θ [Var θ (Y i () U)]+Var θ (X i (U)) = E θ [Var θ (Y i () U)]+E θ [Var θ (X i (U) )]+Var θ (S i ()). By step one Var θ (Y i ()) = Var θ (S i ()) and E θ [Var θ (X i (U) )] = 0 since X i (U) is known if is known. Combining this we see that Var θ (Y i () U) = 0 P θ -a.s. as we wanted. 10. Ancillary statistics As we have seen a sufficient statistic contains all the information about the parameter. he opposite is when a statistic does not contain any information about the parameter.

4 28 HENRIK HUL Definition 13. A statistic U : X U is called ancillary if the conditional distribution of U given Θ = θ is the same for all θ. Example 16. Let X 1 and X 2 be conditionally independent N(θ,1) distributed given Θ = θ. hen U = X 2 X 1 is ancillary. Indeed, U has N(0,2) distribution, which does not depend on θ. Sometimes a statistic contains a coordinate that is ancillary. Definition 14. If = ( 1, 2 ) is a sufficient statistic and 2 is ancillary, then 1 is called conditionally sufficient given 2. Example 17. Let X = (X 1,...,X n ) be conditionally IID U(θ 1/2,θ+1/2)given Θ = θ. hen f X Θ (x θ) = n I [θ 1/2,θ+1/2] (x i ) = I [θ 1/2, ) (minx i )I (,θ+1/2] (maxx i ). i=1 = ( 1, 2 ) = (maxx i,maxx i minx i ) is minimal sufficient and 2 is ancillary. Note that f X θ (y θ) = f X θ (x θ) maxx i = maxy i and minx i = miny i (x) = (y). Hence, by heorem 10 Lecture 7, is minimal sufficient. he conditional density of ( 1, 2 ) given Θ = θ can be computed as (do this as an exercise) f 1, 2 Θ(t 1,t 2 θ) = n(n 1)t n 2 2 I [0,1] (t 2 )I [θ 1/2+t2,θ+1/2](t 1 ) In particular, the marginal density of 2 is f 2 Θ(t 2 θ) = n(n 1)t n 2 2 (1 t 2 ) and this does not depend on θ. Hence 2 is ancillary. Note that the conditional distribution of 1 given 2 = t 2 and Θ = θ is f 1 2,Θ(t 1 t 2,θ) = 1 (1 t 2 ) I [θ 1/2+t 2,θ+1/2](t 1 ). hat is, it is U(θ 1/2 + t 2,θ + 1/2). Hence, this distribution becomes more concentrated as t 2 becomes large. Although 2 does not tell us something about the parameter, it tells us something about the conditional distribution of 1 given Θ. he usual rule in classical statistics is to (whenever it is possible) perform inference conditional on an ancillary statistic. In our example we can exemplify it. Example 18 (continued). Consider the above example with n = 2 and consider finding a 50% confidence interval for Θ. he naive way to do it is to consider the interval I 1 = [minx i,maxx i ] = [ 1 2, 1 ]. his interval satisfies P θ (Θ I 1 ) = 1/2 since there is probability 1/4 that both observations are above θ and probability 1/4 that both are below θ.

5 LECURE NOES 29 If one performs the inference conditional on the ancillary 2 we get a very different result. We can compute P θ ( 1 2 Θ 1 2 ) = P θ (Θ 1 Θ+ 2 2 = t 2 ) = 1 1 t 2 θ+t2 = t 2 1 t 2 I [0,1/2) (t 2 ). θ I [θ 1/2+t2,θ+1/2](t 1 )dt 1 Hence, the level of confidence depends on t 2. In particular, we can construct an interval I 2 = [ 1 1/4(1+ 2 ), 1 +1/4 3 2 /4] which has the property Indeed, P θ (Θ I 2 2 = t 2 ) = 1/2. P θ (Θ I 2 2 = t 2 ) = P θ (Θ 1/4+3 2 /4 1 Θ+1/4(1+ 2 ) 2 = t 2 ) = θ+(1+t2)/4 θ 1/4+3t 2/4 I [θ 1/2+t2,θ+1/2](t 1 )dt 1 = 1/2. Since this probability does not depend on t 2 it follows that P θ (Θ I 2 ) = 1/2. Let us compare the properties of I 1 and I 2. Suppose we observe 2 small. his does not give us much information about Θ and this is reflected in I 2 being wide. On the contrary, I 1 is very small which is counterintuitive. Similarly, if we observe 2 large, then we know more about Θ and I 2 is short. However, this time I 1 is wide! Suppose is sufficient and U is ancillary and they are conditionally independent given Θ = θ. hen there is no benefit of conditioning on U. Indeed, in this case f U,Θ (t u,θ) = f Θ (t θ) so conditioning on U does not change anything. his situation appear when there is a boundedly complete sufficient statistic. heorem 13 (Basu s theorem). If is boundedly complete sufficient statistic and U is ancillary, then and U are conditionally independent given Θ = θ. Furthermore, for every prior µ Θ they are independent (unconditionally). Proof. For the first claim (to show conditional independence) we want to show that for each measurable set A U µ U Θ (A) = µ U,Θ (A t,θ) µ Θ ( θ) a.e. t, θ. (10.1) Since U is ancillary µ U Θ (A θ) = µ U (A), θ. We also have µ U Θ (A θ) = µ U,Θ (A t,θ)µ Θ (dt θ) = µ U (A t)µ Θ (dt θ), where the second equality follows since is sufficient. Indeed, µ X,Θ (B t,θ) = µ X (B t) and since U = U(X) µ U,Θ (A t,θ) = µ X,Θ (U 1 A t,θ) = µ X (U 1 A t) = µ U (A t).

6 30 HENRIK HUL Combining these two we get [µ U (A) µ U (A t)]µ Θ (dt θ) = 0. By considering the integrand as a function g(t) we see that the above equation is the same as E θ [g()] = 0 for each θ and since is boundedly complete µ Θ ({t : g(t) = 0} θ) = 1 for all θ. hat is (10.1) holds. For the second claim we have by conditional independence that µ U, (A B) = µ U (A t)µ Θ (dt θ)µ Θ (dθ) Ω B = µ U (A)µ Θ (B θ)µ Θ (dθ) so and U are independent. Ω = µ U (A)µ (B) Sometimes a combination of the recent results are useful for computing expected values in an unusual way: Example 19. Let X = (X 1,...,X n ) be conditionally IID Exp(θ) given Θ = θ. Consider computing the expected value of X n g(x) =. X 1 + +X n o do this, note that g(x) is an ancillary statistic. Indeed, if Z = (Z 1,...,Z n ) are IID Exp(1) then X = d θ 1 Z and we see that ( 1 P θ (g(x) x) = P θ x < X X ) n 1 +1 X n X n ( 1 = P θ x < Z Z ) n 1 +1 Z n Z n Since the distribution of Z does not depend on θ we see that g(x) is ancillary. he natural statistic (X) = X X n is complete (by the heorem just proved) and minimal sufficient. By Basu s theorem (heorem 13) (X) and g(x) are independent. Hence, θ = E θ [X n ] = E θ [(X)g(X)] = E θ [(X)]E θ [g(x)] = nθe θ [g(x)] and we see that E θ [g(x)] = n 1.

Lecture 4. f X T, (x t, ) = f X,T (x, t ) f T (t )

Lecture 4. f X T, (x t, ) = f X,T (x, t ) f T (t ) LECURE NOES 21 Lecture 4 7. Sufficient statistics Consider the usual statistical setup: the data is X and the paramter is. o gain information about the parameter we study various functions of the data

More information

Lecture 2. (See Exercise 7.22, 7.23, 7.24 in Casella & Berger)

Lecture 2. (See Exercise 7.22, 7.23, 7.24 in Casella & Berger) 8 HENRIK HULT Lecture 2 3. Some common distributions in classical and Bayesian statistics 3.1. Conjugate prior distributions. In the Bayesian setting it is important to compute posterior distributions.

More information

Fundamentals of Statistics

Fundamentals of Statistics Chapter 2 Fundamentals of Statistics This chapter discusses some fundamental concepts of mathematical statistics. These concepts are essential for the material in later chapters. 2.1 Populations, Samples,

More information

Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics

Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics Data from one or a series of random experiments are collected. Planning experiments and collecting data (not discussed here). Analysis:

More information

Summary. Ancillary Statistics What is an ancillary statistic for θ? .2 Can an ancillary statistic be a sufficient statistic?

Summary. Ancillary Statistics What is an ancillary statistic for θ? .2 Can an ancillary statistic be a sufficient statistic? Biostatistics 62 - Statistical Inference Lecture 5 Hyun Min Kang 1 What is an ancillary statistic for θ? 2 Can an ancillary statistic be a sufficient statistic? 3 What are the location parameter and the

More information

Completeness. On the other hand, the distribution of an ancillary statistic doesn t depend on θ at all.

Completeness. On the other hand, the distribution of an ancillary statistic doesn t depend on θ at all. Completeness A minimal sufficient statistic achieves the maximum amount of data reduction while retaining all the information the sample has concerning θ. On the other hand, the distribution of an ancillary

More information

1 Complete Statistics

1 Complete Statistics Complete Statistics February 4, 2016 Debdeep Pati 1 Complete Statistics Suppose X P θ, θ Θ. Let (X (1),..., X (n) ) denote the order statistics. Definition 1. A statistic T = T (X) is complete if E θ g(t

More information

Lecture 3 September 29

Lecture 3 September 29 STATS 300A: Theory of Statistics Fall 015 Lecture 3 September 9 Lecturer: Lester Mackey Scribe: Konstantin Lopyrev, Karthik Rajkumar Warning: These notes may contain factual and/or typographic errors.

More information

4. CONTINUOUS RANDOM VARIABLES

4. CONTINUOUS RANDOM VARIABLES IA Probability Lent Term 4 CONTINUOUS RANDOM VARIABLES 4 Introduction Up to now we have restricted consideration to sample spaces Ω which are finite, or countable; we will now relax that assumption We

More information

ECE 275B Homework # 1 Solutions Winter 2018

ECE 275B Homework # 1 Solutions Winter 2018 ECE 275B Homework # 1 Solutions Winter 2018 1. (a) Because x i are assumed to be independent realizations of a continuous random variable, it is almost surely (a.s.) 1 the case that x 1 < x 2 < < x n Thus,

More information

ECE 275B Homework # 1 Solutions Version Winter 2015

ECE 275B Homework # 1 Solutions Version Winter 2015 ECE 275B Homework # 1 Solutions Version Winter 2015 1. (a) Because x i are assumed to be independent realizations of a continuous random variable, it is almost surely (a.s.) 1 the case that x 1 < x 2

More information

STA 260: Statistics and Probability II

STA 260: Statistics and Probability II Al Nosedal. University of Toronto. Winter 2017 1 Properties of Point Estimators and Methods of Estimation 2 3 If you can t explain it simply, you don t understand it well enough Albert Einstein. Definition

More information

ST5215: Advanced Statistical Theory

ST5215: Advanced Statistical Theory Department of Statistics & Applied Probability Wednesday, October 19, 2011 Lecture 17: UMVUE and the first method of derivation Estimable parameters Let ϑ be a parameter in the family P. If there exists

More information

Chapter 3: Unbiased Estimation Lecture 22: UMVUE and the method of using a sufficient and complete statistic

Chapter 3: Unbiased Estimation Lecture 22: UMVUE and the method of using a sufficient and complete statistic Chapter 3: Unbiased Estimation Lecture 22: UMVUE and the method of using a sufficient and complete statistic Unbiased estimation Unbiased or asymptotically unbiased estimation plays an important role in

More information

Mathematical Statistics

Mathematical Statistics Mathematical Statistics Chapter Three. Point Estimation 3.4 Uniformly Minimum Variance Unbiased Estimator(UMVUE) Criteria for Best Estimators MSE Criterion Let F = {p(x; θ) : θ Θ} be a parametric distribution

More information

[Chapter 6. Functions of Random Variables]

[Chapter 6. Functions of Random Variables] [Chapter 6. Functions of Random Variables] 6.1 Introduction 6.2 Finding the probability distribution of a function of random variables 6.3 The method of distribution functions 6.5 The method of Moment-generating

More information

MAS3301 Bayesian Statistics

MAS3301 Bayesian Statistics MAS3301 Bayesian Statistics M. Farrow School of Mathematics and Statistics Newcastle University Semester 2, 2008-9 1 11 Conjugate Priors IV: The Dirichlet distribution and multinomial observations 11.1

More information

SOLUTIONS TO MATH68181 EXTREME VALUES AND FINANCIAL RISK EXAM

SOLUTIONS TO MATH68181 EXTREME VALUES AND FINANCIAL RISK EXAM SOLUTIONS TO MATH68181 EXTREME VALUES AND FINANCIAL RISK EXAM Solutions to Question A1 a) The marginal cdfs of F X,Y (x, y) = [1 + exp( x) + exp( y) + (1 α) exp( x y)] 1 are F X (x) = F X,Y (x, ) = [1

More information

Chapter 6. Order Statistics and Quantiles. 6.1 Extreme Order Statistics

Chapter 6. Order Statistics and Quantiles. 6.1 Extreme Order Statistics Chapter 6 Order Statistics and Quantiles 61 Extreme Order Statistics Suppose we have a finite sample X 1,, X n Conditional on this sample, we define the values X 1),, X n) to be a permutation of X 1,,

More information

Lecture 16 November Application of MoUM to our 2-sided testing problem

Lecture 16 November Application of MoUM to our 2-sided testing problem STATS 300A: Theory of Statistics Fall 2015 Lecture 16 November 17 Lecturer: Lester Mackey Scribe: Reginald Long, Colin Wei Warning: These notes may contain factual and/or typographic errors. 16.1 Recap

More information

4 Invariant Statistical Decision Problems

4 Invariant Statistical Decision Problems 4 Invariant Statistical Decision Problems 4.1 Invariant decision problems Let G be a group of measurable transformations from the sample space X into itself. The group operation is composition. Note that

More information

Remarks on Improper Ignorance Priors

Remarks on Improper Ignorance Priors As a limit of proper priors Remarks on Improper Ignorance Priors Two caveats relating to computations with improper priors, based on their relationship with finitely-additive, but not countably-additive

More information

Chapter 7. Confidence Sets Lecture 30: Pivotal quantities and confidence sets

Chapter 7. Confidence Sets Lecture 30: Pivotal quantities and confidence sets Chapter 7. Confidence Sets Lecture 30: Pivotal quantities and confidence sets Confidence sets X: a sample from a population P P. θ = θ(p): a functional from P to Θ R k for a fixed integer k. C(X): a confidence

More information

Normalising constants and maximum likelihood inference

Normalising constants and maximum likelihood inference Normalising constants and maximum likelihood inference Jakob G. Rasmussen Department of Mathematics Aalborg University Denmark March 9, 2011 1/14 Today Normalising constants Approximation of normalising

More information

Chapter 6. Integration. 1. Integrals of Nonnegative Functions. a j µ(e j ) (ca j )µ(e j ) = c X. and ψ =

Chapter 6. Integration. 1. Integrals of Nonnegative Functions. a j µ(e j ) (ca j )µ(e j ) = c X. and ψ = Chapter 6. Integration 1. Integrals of Nonnegative Functions Let (, S, µ) be a measure space. We denote by L + the set of all measurable functions from to [0, ]. Let φ be a simple function in L +. Suppose

More information

Lecture 4: UMVUE and unbiased estimators of 0

Lecture 4: UMVUE and unbiased estimators of 0 Lecture 4: UMVUE and unbiased estimators of Problems of approach 1. Approach 1 has the following two main shortcomings. Even if Theorem 7.3.9 or its extension is applicable, there is no guarantee that

More information

Probability and Distributions

Probability and Distributions Probability and Distributions What is a statistical model? A statistical model is a set of assumptions by which the hypothetical population distribution of data is inferred. It is typically postulated

More information

Part IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Part IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015 Part IA Probability Theorems Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.

More information

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it

More information

MAS223 Statistical Inference and Modelling Exercises

MAS223 Statistical Inference and Modelling Exercises MAS223 Statistical Inference and Modelling Exercises The exercises are grouped into sections, corresponding to chapters of the lecture notes Within each section exercises are divided into warm-up questions,

More information

Chapter 1. Statistical Spaces

Chapter 1. Statistical Spaces Chapter 1 Statistical Spaces Mathematical statistics is a science that studies the statistical regularity of random phenomena, essentially by some observation values of random variable (r.v.) X. Sometimes

More information

Summary. Complete Statistics ... Last Lecture. .1 What is a complete statistic? .2 Why it is called as complete statistic?

Summary. Complete Statistics ... Last Lecture. .1 What is a complete statistic? .2 Why it is called as complete statistic? Last Lecture Biostatistics 602 - Statistical Inference Lecture 06 Hyun Min Kang January 29th, 2013 1 What is a complete statistic? 2 Why it is called as complete statistic? 3 Can the same statistic be

More information

Chapter 3 : Likelihood function and inference

Chapter 3 : Likelihood function and inference Chapter 3 : Likelihood function and inference 4 Likelihood function and inference The likelihood Information and curvature Sufficiency and ancilarity Maximum likelihood estimation Non-regular models EM

More information

Modern Methods of Statistical Learning sf2935 Auxiliary material: Exponential Family of Distributions Timo Koski. Second Quarter 2016

Modern Methods of Statistical Learning sf2935 Auxiliary material: Exponential Family of Distributions Timo Koski. Second Quarter 2016 Auxiliary material: Exponential Family of Distributions Timo Koski Second Quarter 2016 Exponential Families The family of distributions with densities (w.r.t. to a σ-finite measure µ) on X defined by R(θ)

More information

ST5215: Advanced Statistical Theory

ST5215: Advanced Statistical Theory Department of Statistics & Applied Probability Monday, September 26, 2011 Lecture 10: Exponential families and Sufficient statistics Exponential Families Exponential families are important parametric families

More information

Sobolev Spaces. Chapter 10

Sobolev Spaces. Chapter 10 Chapter 1 Sobolev Spaces We now define spaces H 1,p (R n ), known as Sobolev spaces. For u to belong to H 1,p (R n ), we require that u L p (R n ) and that u have weak derivatives of first order in L p

More information

Chapter 6. Convergence. Probability Theory. Four different convergence concepts. Four different convergence concepts. Convergence in probability

Chapter 6. Convergence. Probability Theory. Four different convergence concepts. Four different convergence concepts. Convergence in probability Probability Theory Chapter 6 Convergence Four different convergence concepts Let X 1, X 2, be a sequence of (usually dependent) random variables Definition 1.1. X n converges almost surely (a.s.), or with

More information

conditional cdf, conditional pdf, total probability theorem?

conditional cdf, conditional pdf, total probability theorem? 6 Multiple Random Variables 6.0 INTRODUCTION scalar vs. random variable cdf, pdf transformation of a random variable conditional cdf, conditional pdf, total probability theorem expectation of a random

More information

Rare-Event Simulation

Rare-Event Simulation Rare-Event Simulation Background: Read Chapter 6 of text. 1 Why is Rare-Event Simulation Challenging? Consider the problem of computing α = P(A) when P(A) is small (i.e. rare ). The crude Monte Carlo estimator

More information

Statistical Theory MT 2007 Problems 4: Solution sketches

Statistical Theory MT 2007 Problems 4: Solution sketches Statistical Theory MT 007 Problems 4: Solution sketches 1. Consider a 1-parameter exponential family model with density f(x θ) = f(x)g(θ)exp{cφ(θ)h(x)}, x X. Suppose that the prior distribution has the

More information

is a Borel subset of S Θ for each c R (Bertsekas and Shreve, 1978, Proposition 7.36) This always holds in practical applications.

is a Borel subset of S Θ for each c R (Bertsekas and Shreve, 1978, Proposition 7.36) This always holds in practical applications. Stat 811 Lecture Notes The Wald Consistency Theorem Charles J. Geyer April 9, 01 1 Analyticity Assumptions Let { f θ : θ Θ } be a family of subprobability densities 1 with respect to a measure µ on a measurable

More information

Spring 2012 Math 541A Exam 1. X i, S 2 = 1 n. n 1. X i I(X i < c), T n =

Spring 2012 Math 541A Exam 1. X i, S 2 = 1 n. n 1. X i I(X i < c), T n = Spring 2012 Math 541A Exam 1 1. (a) Let Z i be independent N(0, 1), i = 1, 2,, n. Are Z = 1 n n Z i and S 2 Z = 1 n 1 n (Z i Z) 2 independent? Prove your claim. (b) Let X 1, X 2,, X n be independent identically

More information

Lecture 4: Completion of a Metric Space

Lecture 4: Completion of a Metric Space 15 Lecture 4: Completion of a Metric Space Closure vs. Completeness. Recall the statement of Lemma??(b): A subspace M of a metric space X is closed if and only if every convergent sequence {x n } X satisfying

More information

MAT 257, Handout 13: December 5-7, 2011.

MAT 257, Handout 13: December 5-7, 2011. MAT 257, Handout 13: December 5-7, 2011. The Change of Variables Theorem. In these notes, I try to make more explicit some parts of Spivak s proof of the Change of Variable Theorem, and to supply most

More information

Stat 366 A1 (Fall 2006) Midterm Solutions (October 23) page 1

Stat 366 A1 (Fall 2006) Midterm Solutions (October 23) page 1 Stat 366 A1 Fall 6) Midterm Solutions October 3) page 1 1. The opening prices per share Y 1 and Y measured in dollars) of two similar stocks are independent random variables, each with a density function

More information

Signed Measures. Chapter Basic Properties of Signed Measures. 4.2 Jordan and Hahn Decompositions

Signed Measures. Chapter Basic Properties of Signed Measures. 4.2 Jordan and Hahn Decompositions Chapter 4 Signed Measures Up until now our measures have always assumed values that were greater than or equal to 0. In this chapter we will extend our definition to allow for both positive negative values.

More information

BMIR Lecture Series on Probability and Statistics Fall 2015 Discrete RVs

BMIR Lecture Series on Probability and Statistics Fall 2015 Discrete RVs Lecture #7 BMIR Lecture Series on Probability and Statistics Fall 2015 Department of Biomedical Engineering and Environmental Sciences National Tsing Hua University 7.1 Function of Single Variable Theorem

More information

STAT 512 sp 2018 Summary Sheet

STAT 512 sp 2018 Summary Sheet STAT 5 sp 08 Summary Sheet Karl B. Gregory Spring 08. Transformations of a random variable Let X be a rv with support X and let g be a function mapping X to Y with inverse mapping g (A = {x X : g(x A}

More information

Classical and Bayesian inference

Classical and Bayesian inference Classical and Bayesian inference AMS 132 January 18, 2018 Claudia Wehrhahn (UCSC) Classical and Bayesian inference January 18, 2018 1 / 9 Sampling from a Bernoulli Distribution Theorem (Beta-Bernoulli

More information

1 Solution to Problem 2.1

1 Solution to Problem 2.1 Solution to Problem 2. I incorrectly worked this exercise instead of 2.2, so I decided to include the solution anyway. a) We have X Y /3, which is a - function. It maps the interval, ) where X lives) onto

More information

ADVANCED PROBABILITY: SOLUTIONS TO SHEET 1

ADVANCED PROBABILITY: SOLUTIONS TO SHEET 1 ADVANCED PROBABILITY: SOLUTIONS TO SHEET 1 Last compiled: November 6, 213 1. Conditional expectation Exercise 1.1. To start with, note that P(X Y = P( c R : X > c, Y c or X c, Y > c = P( c Q : X > c, Y

More information

Stat 451 Lecture Notes Numerical Integration

Stat 451 Lecture Notes Numerical Integration Stat 451 Lecture Notes 03 12 Numerical Integration Ryan Martin UIC www.math.uic.edu/~rgmartin 1 Based on Chapter 5 in Givens & Hoeting, and Chapters 4 & 18 of Lange 2 Updated: February 11, 2016 1 / 29

More information

Chapter 4 HOMEWORK ASSIGNMENTS. 4.1 Homework #1

Chapter 4 HOMEWORK ASSIGNMENTS. 4.1 Homework #1 Chapter 4 HOMEWORK ASSIGNMENTS These homeworks may be modified as the semester progresses. It is your responsibility to keep up to date with the correctly assigned homeworks. There may be some errors in

More information

the convolution of f and g) given by

the convolution of f and g) given by 09:53 /5/2000 TOPIC Characteristic functions, cont d This lecture develops an inversion formula for recovering the density of a smooth random variable X from its characteristic function, and uses that

More information

LECTURE NOTES 57. Lecture 9

LECTURE NOTES 57. Lecture 9 LECTURE NOTES 57 Lecture 9 17. Hypothesis testing A special type of decision problem is hypothesis testing. We partition the parameter space into H [ A with H \ A = ;. Wewrite H 2 H A 2 A. A decision problem

More information

λ(x + 1)f g (x) > θ 0

λ(x + 1)f g (x) > θ 0 Stat 8111 Final Exam December 16 Eleven students took the exam, the scores were 92, 78, 4 in the 5 s, 1 in the 4 s, 1 in the 3 s and 3 in the 2 s. 1. i) Let X 1, X 2,..., X n be iid each Bernoulli(θ) where

More information

Recall that in order to prove Theorem 8.8, we argued that under certain regularity conditions, the following facts are true under H 0 : 1 n

Recall that in order to prove Theorem 8.8, we argued that under certain regularity conditions, the following facts are true under H 0 : 1 n Chapter 9 Hypothesis Testing 9.1 Wald, Rao, and Likelihood Ratio Tests Suppose we wish to test H 0 : θ = θ 0 against H 1 : θ θ 0. The likelihood-based results of Chapter 8 give rise to several possible

More information

The Delta Method and Applications

The Delta Method and Applications Chapter 5 The Delta Method and Applications 5.1 Local linear approximations Suppose that a particular random sequence converges in distribution to a particular constant. The idea of using a first-order

More information

A Very Brief Summary of Statistical Inference, and Examples

A Very Brief Summary of Statistical Inference, and Examples A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2008 Prof. Gesine Reinert 1 Data x = x 1, x 2,..., x n, realisations of random variables X 1, X 2,..., X n with distribution (model)

More information

Bayesian statistics, simulation and software

Bayesian statistics, simulation and software Module 10: Bayesian prediction and model checking Department of Mathematical Sciences Aalborg University 1/15 Prior predictions Suppose we want to predict future data x without observing any data x. Assume:

More information

P (A G) dp G P (A G)

P (A G) dp G P (A G) First homework assignment. Due at 12:15 on 22 September 2016. Homework 1. We roll two dices. X is the result of one of them and Z the sum of the results. Find E [X Z. Homework 2. Let X be a r.v.. Assume

More information

INTRODUCTION TO BAYESIAN METHODS II

INTRODUCTION TO BAYESIAN METHODS II INTRODUCTION TO BAYESIAN METHODS II Abstract. We will revisit point estimation and hypothesis testing from the Bayesian perspective.. Bayes estimators Let X = (X,..., X n ) be a random sample from the

More information

Part IB Statistics. Theorems with proof. Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua. Lent 2015

Part IB Statistics. Theorems with proof. Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua. Lent 2015 Part IB Statistics Theorems with proof Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly)

More information

Maximum Likelihood Estimation

Maximum Likelihood Estimation Chapter 7 Maximum Likelihood Estimation 7. Consistency If X is a random variable (or vector) with density or mass function f θ (x) that depends on a parameter θ, then the function f θ (X) viewed as a function

More information

Nuisance parameters and their treatment

Nuisance parameters and their treatment BS2 Statistical Inference, Lecture 2, Hilary Term 2008 April 2, 2008 Ancillarity Inference principles Completeness A statistic A = a(x ) is said to be ancillary if (i) The distribution of A does not depend

More information

Chapter 8: Taylor s theorem and L Hospital s rule

Chapter 8: Taylor s theorem and L Hospital s rule Chapter 8: Taylor s theorem and L Hospital s rule Theorem: [Inverse Mapping Theorem] Suppose that a < b and f : [a, b] R. Given that f (x) > 0 for all x (a, b) then f 1 is differentiable on (f(a), f(b))

More information

Mathematical statistics

Mathematical statistics October 4 th, 2018 Lecture 12: Information Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation Chapter

More information

Bayesian Inference for DSGE Models. Lawrence J. Christiano

Bayesian Inference for DSGE Models. Lawrence J. Christiano Bayesian Inference for DSGE Models Lawrence J. Christiano Outline State space-observer form. convenient for model estimation and many other things. Bayesian inference Bayes rule. Monte Carlo integation.

More information

Optimal Estimation of a Nonsmooth Functional

Optimal Estimation of a Nonsmooth Functional Optimal Estimation of a Nonsmooth Functional T. Tony Cai Department of Statistics The Wharton School University of Pennsylvania http://stat.wharton.upenn.edu/ tcai Joint work with Mark Low 1 Question Suppose

More information

University of Regina. Lecture Notes. Michael Kozdron

University of Regina. Lecture Notes. Michael Kozdron University of Regina Statistics 252 Mathematical Statistics Lecture Notes Winter 2005 Michael Kozdron kozdron@math.uregina.ca www.math.uregina.ca/ kozdron Contents 1 The Basic Idea of Statistics: Estimating

More information

Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed

Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed 18.466 Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed 1. MLEs in exponential families Let f(x,θ) for x X and θ Θ be a likelihood function, that is, for present purposes,

More information

ECE531 Homework Assignment Number 6 Solution

ECE531 Homework Assignment Number 6 Solution ECE53 Homework Assignment Number 6 Solution Due by 8:5pm on Wednesday 3-Mar- Make sure your reasoning and work are clear to receive full credit for each problem.. 6 points. Suppose you have a scalar random

More information

A Very Brief Summary of Statistical Inference, and Examples

A Very Brief Summary of Statistical Inference, and Examples A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2009 Prof. Gesine Reinert Our standard situation is that we have data x = x 1, x 2,..., x n, which we view as realisations of random

More information

The theory of manifolds Lecture 2

The theory of manifolds Lecture 2 The theory of manifolds Lecture 2 Let X be a subset of R N, Y a subset of R n and f : X Y a continuous map. We recall Definition 1. f is a C map if for every p X, there exists a neighborhood, U p, of p

More information

Lecture 1: Introduction

Lecture 1: Introduction Principles of Statistics Part II - Michaelmas 208 Lecturer: Quentin Berthet Lecture : Introduction This course is concerned with presenting some of the mathematical principles of statistical theory. One

More information

Statistical Theory MT 2006 Problems 4: Solution sketches

Statistical Theory MT 2006 Problems 4: Solution sketches Statistical Theory MT 006 Problems 4: Solution sketches 1. Suppose that X has a Poisson distribution with unknown mean θ. Determine the conjugate prior, and associate posterior distribution, for θ. Determine

More information

Sampling Distributions

Sampling Distributions Sampling Distributions In statistics, a random sample is a collection of independent and identically distributed (iid) random variables, and a sampling distribution is the distribution of a function of

More information

Lecture Notes 3 Multiple Random Variables. Joint, Marginal, and Conditional pmfs. Bayes Rule and Independence for pmfs

Lecture Notes 3 Multiple Random Variables. Joint, Marginal, and Conditional pmfs. Bayes Rule and Independence for pmfs Lecture Notes 3 Multiple Random Variables Joint, Marginal, and Conditional pmfs Bayes Rule and Independence for pmfs Joint, Marginal, and Conditional pdfs Bayes Rule and Independence for pdfs Functions

More information

Classical and Bayesian inference

Classical and Bayesian inference Classical and Bayesian inference AMS 132 Claudia Wehrhahn (UCSC) Classical and Bayesian inference January 8 1 / 11 The Prior Distribution Definition Suppose that one has a statistical model with parameter

More information

MIT Spring 2016

MIT Spring 2016 Exponential Families II MIT 18.655 Dr. Kempthorne Spring 2016 1 Outline Exponential Families II 1 Exponential Families II 2 : Expectation and Variance U (k 1) and V (l 1) are random vectors If A (m k),

More information

Will Murray s Probability, XXXII. Moment-Generating Functions 1. We want to study functions of them:

Will Murray s Probability, XXXII. Moment-Generating Functions 1. We want to study functions of them: Will Murray s Probability, XXXII. Moment-Generating Functions XXXII. Moment-Generating Functions Premise We have several random variables, Y, Y, etc. We want to study functions of them: U (Y,..., Y n ).

More information

STAT 3610: Review of Probability Distributions

STAT 3610: Review of Probability Distributions STAT 3610: Review of Probability Distributions Mark Carpenter Professor of Statistics Department of Mathematics and Statistics August 25, 2015 Support of a Random Variable Definition The support of a random

More information

Asymptotics for posterior hazards

Asymptotics for posterior hazards Asymptotics for posterior hazards Pierpaolo De Blasi University of Turin 10th August 2007, BNR Workshop, Isaac Newton Intitute, Cambridge, UK Joint work with Giovanni Peccati (Université Paris VI) and

More information

B553 Lecture 1: Calculus Review

B553 Lecture 1: Calculus Review B553 Lecture 1: Calculus Review Kris Hauser January 10, 2012 This course requires a familiarity with basic calculus, some multivariate calculus, linear algebra, and some basic notions of metric topology.

More information

MIT Spring 2016

MIT Spring 2016 Dr. Kempthorne Spring 2016 1 Outline Building 1 Building 2 Definition Building Let X be a random variable/vector with sample space X R q and probability model P θ. The class of probability models P = {P

More information

Monotone comparative statics Finite Data and GARP

Monotone comparative statics Finite Data and GARP Monotone comparative statics Finite Data and GARP Econ 2100 Fall 2017 Lecture 7, September 19 Problem Set 3 is due in Kelly s mailbox by 5pm today Outline 1 Comparative Statics Without Calculus 2 Supermodularity

More information

Stat 535 C - Statistical Computing & Monte Carlo Methods. Arnaud Doucet.

Stat 535 C - Statistical Computing & Monte Carlo Methods. Arnaud Doucet. Stat 535 C - Statistical Computing & Monte Carlo Methods Arnaud Doucet Email: arnaud@cs.ubc.ca 1 Suggested Projects: www.cs.ubc.ca/~arnaud/projects.html First assignement on the web: capture/recapture.

More information

Differential Equations Practice: 2nd Order Linear: Nonhomogeneous Equations: Variation of Parameters Page 1

Differential Equations Practice: 2nd Order Linear: Nonhomogeneous Equations: Variation of Parameters Page 1 Differential Equations Practice: 2nd Order Linear: Nonhomogeneous Equations: Variation of Parameters Page Questions Example (3.6.) Find a particular solution of the differential equation y 5y + 6y = 2e

More information

STATISTICAL METHODS FOR SIGNAL PROCESSING c Alfred Hero

STATISTICAL METHODS FOR SIGNAL PROCESSING c Alfred Hero STATISTICAL METHODS FOR SIGNAL PROCESSING c Alfred Hero 1999 32 Statistic used Meaning in plain english Reduction ratio T (X) [X 1,..., X n ] T, entire data sample RR 1 T (X) [X (1),..., X (n) ] T, rank

More information

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it

More information

Theory and Applications of Stochastic Systems Lecture Exponential Martingale for Random Walk

Theory and Applications of Stochastic Systems Lecture Exponential Martingale for Random Walk Instructor: Victor F. Araman December 4, 2003 Theory and Applications of Stochastic Systems Lecture 0 B60.432.0 Exponential Martingale for Random Walk Let (S n : n 0) be a random walk with i.i.d. increments

More information

Siegel s formula via Stein s identities

Siegel s formula via Stein s identities Siegel s formula via Stein s identities Jun S. Liu Department of Statistics Harvard University Abstract Inspired by a surprising formula in Siegel (1993), we find it convenient to compute covariances,

More information

Review of Multi-Calculus (Study Guide for Spivak s CHAPTER ONE TO THREE)

Review of Multi-Calculus (Study Guide for Spivak s CHAPTER ONE TO THREE) Review of Multi-Calculus (Study Guide for Spivak s CHPTER ONE TO THREE) This material is for June 9 to 16 (Monday to Monday) Chapter I: Functions on R n Dot product and norm for vectors in R n : Let X

More information

or E ( U(X) ) e zx = e ux e ivx = e ux( cos(vx) + i sin(vx) ), B X := { u R : M X (u) < } (4)

or E ( U(X) ) e zx = e ux e ivx = e ux( cos(vx) + i sin(vx) ), B X := { u R : M X (u) < } (4) :23 /4/2000 TOPIC Characteristic functions This lecture begins our study of the characteristic function φ X (t) := Ee itx = E cos(tx)+ie sin(tx) (t R) of a real random variable X Characteristic functions

More information

Monte-Carlo MMD-MA, Université Paris-Dauphine. Xiaolu Tan

Monte-Carlo MMD-MA, Université Paris-Dauphine. Xiaolu Tan Monte-Carlo MMD-MA, Université Paris-Dauphine Xiaolu Tan tan@ceremade.dauphine.fr Septembre 2015 Contents 1 Introduction 1 1.1 The principle.................................. 1 1.2 The error analysis

More information

A Probability Review

A Probability Review A Probability Review Outline: A probability review Shorthand notation: RV stands for random variable EE 527, Detection and Estimation Theory, # 0b 1 A Probability Review Reading: Go over handouts 2 5 in

More information

Sampling Distributions

Sampling Distributions In statistics, a random sample is a collection of independent and identically distributed (iid) random variables, and a sampling distribution is the distribution of a function of random sample. For example,

More information

University of Pavia. M Estimators. Eduardo Rossi

University of Pavia. M Estimators. Eduardo Rossi University of Pavia M Estimators Eduardo Rossi Criterion Function A basic unifying notion is that most econometric estimators are defined as the minimizers of certain functions constructed from the sample

More information

Course 4 Solutions November 2001 Exams

Course 4 Solutions November 2001 Exams Course 4 Solutions November 001 Exams November, 001 Society of Actuaries Question #1 From the Yule-Walker equations: ρ φ + ρφ 1 1 1. 1 1+ ρ ρφ φ Substituting the given quantities yields: 0.53 φ + 0.53φ

More information