2.2.2 Comparing estimators

Size: px
Start display at page:

Download "2.2.2 Comparing estimators"

Transcription

1 2.2. POINT ESTIMATION Comparing estimators Estimators can be compared through their mean square errors. If they are unbiased, this is equivalent to comparing their variances. In many applications, we try to find an unbiased estimator which has minimum variance, or at least low variance. We can compare the variances of several unbiased estimators, but how do we know if we have found an estimator with the lowest variance among all unbiased estimators? The answer is given by Cramér-Rao lower bound (CRLB). The CRLB for unbiased estimators ofφ = g(ϑ), when there is a single parameter, is CRLB(φ) = E { } 2 dg(ϑ) { }, d2 logf(y ;ϑ) 2 where Y = (Y 1,...,Y n ) T and f(y;ϑ) is the joint density function (pdf) of Y 1,...,Y n when Y is continuous. For discrete r.vs we replace pdf with a pmf. E{ d 2 logf(y;ϑ)/ 2 } is called the Fisher information. When there arepparameters, ϑ = (ϑ 1,...,ϑ p ) T, andφ = g(ϑ), the CRLB is ( g CRLB(φ) =,..., g ) M 1, ϑ 1 ϑ p where the matrix M has (i,j)th element E{ 2 logf(y;ϑ)/ ϑ i ϑ j } and is called the Fisher Information Matrix. g ϑ 1. g ϑ p For the following result, we need to assume that the sample space does not depend on ϑ 1,...,ϑ p, and that the differentiability and continuity constraints for all ϑ Θ hold, so that the order of integration and differentiation can be interchanged, and the Fisher Information is positive and finite. Problems which satisfy these conditions are called regular. Theorem 2.4. Cramér-Rao inequality. For any unbiased estimator,t(y ), ofφif the estimation problem is regular then var{t(y)} CRLB(φ).

2 88 CHAPTER 2. ELEMENTS OF STATISTICAL INFERENCE Proof. For simplicity, we only prove the result in the one-parameter, continuous case. The proof uses the result that, for any two random variables U and V, their correlation coefficient ρ(u,v) lies between -1 and +1, that is, {ρ(u,v)} 2 1, which may be written as var(u) {cov(u,v)}2. (2.2) var(v) LetU = T(Y ) and V = dlogf(y;ϑ)/. Then we may write { } dlogf(y ;ϑ) E(V) = E [ d = E f(y;ϑ) ] property of derivative of log f(y;ϑ) d =... f(y;ϑ) f(y;ϑ) f(y;ϑ)dy 1...dy n df(y;ϑ) =... dy 1...dy n = d... f(y;ϑ)dy 1...dy n = d (1) = 0. Next, sincee(v) = 0, we have { } dlogf(y ;ϑ) cov(u,v) = E(UV) = E T(Y) { } d f(y ;ϑ) = E T(Y) f(y ;ϑ) d =... T(y) f(y;ϑ) f(y;ϑ) f(y;ϑ)dy 1...dy n =... T(y) df(y;ϑ) dy 1...dy n = d... T(y)f(y;ϑ)dy 1...dy n = d dg(ϑ) E{T(Y)} =.

3 2.2. POINT ESTIMATION 89 Finally, again sincee(v) = 0, we have var(v) = E(V 2 ) [ {dlogf(y;ϑ) } ] 2 = E { d dlogf(y;ϑ) = E f(y;ϑ) } f(y;ϑ) { d dlogf(y;ϑ) =... f(y;ϑ) } f(y;ϑ)dy 1...dy n f(y;ϑ) dlogf(y;ϑ) df(y;ϑ) =... dy 1...dy n. Now, since we can write d 2 f = d 2 ( f dlogf ) = df dlogf logf +fd2, 2 we obtain d 2 f(y;ϑ) var(v) =... dy dy n... f(y;ϑ) d2 logf(y;ϑ) dy dy n { } = d2... f(y;ϑ)dy dy n +E d2 logf(y;ϑ) { } { 2 } = d2 2(1)+E d2 logf(y;ϑ) = E d2 logf(y;ϑ). 2 2 The required result then follows immediately from (2.2). Example Suppose thaty 1,...,Y n are independentpoisson(λ) random variables. Let g(λ) = λ. Then we have dg/dλ = 1. Here we have a discrete r.v. whose joint pmf can be written as the following product P(Y 1 = y 1,...,Y n = y n ;λ) = n λ y i e λ y i! = λ n y i e nλ n y i!,

4 90 CHAPTER 2. ELEMENTS OF STATISTICAL INFERENCE and so taking log we obtain logp = ( n ) y i logλ nλ log y i!. Thus, we have dlogp n = y i n dλ λ and d 2 logp n = y i. dλ 2 λ 2 Hence, the Fisher information is ( ) E d2 logp dλ 2 ( ) = 1 λ 2E Y i = 1 E(Y λ 2 i ) = 1 λ 2nλ = n λ and CRLB(λ) = λ/n. The efficiency of an asymptotically unbiased estimator,t(y), ofφis eff(t) = lim n CRLB(φ) var{t(y)}. An asymptotically unbiased estimator, T(Y ), of φ is said to be efficient if its efficiency is equal to one. In this case, for large n, var{t(y)} is approximately equal to the CRLB. Exercise 2.5. Suppose thaty 1,Y 2,...,Y n is an independent sample from abernoulli(p) distribution. Calculate the CRLB(p) and see if Y is an unbiased estimator of p with minimum variance. Exercise 2.6. Suppose thaty 1,Y 2,...,Y n is an independent sample fromn(µ,σ 2 ). Derive the Fisher Information Matrix M for ϑ = (µ,σ 2 ) T and find the Cramer- Rao lower bound for an unbiased estimator ofg(ϑ) = µ+σ 2.

5 2.2. POINT ESTIMATION Minimum Variance Unbiased Estimators If an unbiased estimator has the variance equal to the CRLB, it must have the minimum variance amongst all unbiased estimators. We call it the minimum variance unbiased estimator (MVUE) ofφ. Sufficiency is a powerful property in finding unbiased, minimum variance estimators. If T(Y) is an unbiased estimator of ϑ and S is a statistic sufficient for ϑ, then there is a function of S that is also an unbiased estimator of ϑ and has no larger variance than the variance of T(Y). The following theorem formalizes this statement. Theorem 2.5. Rao-Blackwell theorem. Let Y = (Y 1,Y 2,...,Y n ) T be a random sample, S = (S 1,...,S p ) T be jointly sufficient statistics for ϑ = (ϑ 1,...,ϑ p ) T and T(Y ) (which is not a function of S) be an unbiased estimator of φ = g(ϑ). Then, U = E(T S) is a statistic such that (a) E(U) = φ, so thatu is an unbiased estimator of φ, and (b) var(u) < var(t). Proof. First, we note thatu is a statistic. Indeed, sinces are jointly sufficient for ϑ, the conditional distributiony S does not depend on the parameters and so the conditional distribution of a function T(Y ) given S, T S, does not depend on ϑ either. Thus,U = E(T S) is a function of the random sample only, not a function ofϑ, therefore it is a statistic. Next, we will use the known facts about the conditional expectation and variance given Exercise SinceT is an unbiased estimator ofφ, we have E(U) = E[E(T S)] = E(T) = φ. So U is also an unbiased estimator ofφ, which proves (a). Finally, we get var(t) = var[e(t S)]+E[var(T S)] = var(u)+e[var(t S)]. However, since T is not a function of S we have var(t S) > 0, thus, it follows thate[var(t S)] > 0, and hence (b) is proved.

6 92 CHAPTER 2. ELEMENTS OF STATISTICAL INFERENCE It means that, if we have an unbiased estimator, T, of φ, which is not a function of the sufficient statistics, we can always find an unbiased estimator which has smaller variance, namely U = E(T S 1,...,S p ) which is a function of S. We thus have the following result. Corollary 2.1. MVUEs must be functions of sufficient statistics. Example Suppose thaty 1,...,Y n are independentpoisson(λ) random variables. Then T = Y i, for any i = 1,...,n, is an unbiased estimator of λ. Also, S = n Y i is a sufficient statistic forλand T is not a function ofs. Hence, a better unbiased estimator is given by any ofe(y 1 n Y i),...,e(y n n Y i). Now, sincee(y 1 n Y i) =... = E(Y n n Y i) and E(Y 1 Y i )+E(Y 2 Y i )+...+E(Y n Y i ) = E [ n Y i we have ne(y 1 Y i ) = Y i. ] Y i = Y i, Hence U = E(Y 1 n Y i) = Y is a better unbiased estimator of λ than Y 1 or than any of Y i. In fact, Y is a MVUE of λ as its variance is equal to the Cramer- Rao Lower Bound forλ. (See Example 2.12) Complete sufficient statistics T(Y)is a function of random sampley = (Y 1,Y 2,...,Y n ) and so it is a random variable as well. Hence, we may ask about the distribution oft(y). For example, assume that σ 2 is known equal to σo. 2 Then, to make inference about µ we may reduce the random sample to its mean. We know that T(Y) = Y N ( ) µ, σ2 o n ify i N(µ,σo 2 ) and we may write iid f T (t;µ σ 2 o ) = 1 2πσo e n(t µ)2 /2σ 2 o. Due to this data reduction we can make inference about µ based on the distribution of Y only rather than on the multivariate distribution of the whole random

7 2.2. POINT ESTIMATION 93 sampley. A minimal sufficient statistic reduces data maximally while retaining all the information about the parameter ϑ. We would also like such a statistic to be independent of any so called ancillary functions of the random sample whose distributions do not depend on the parameter of interest. Such an independent statistic is called complete. First, we introduce a notion of a complete family of distributions. Definition 2.8. A family of distributionsp = {P ϑ : ϑ Θ} defined on a common spacey is called complete if for any real measurable functionh(y) E[h(Y)] = 0 implies that P ϑ ( h(y) = 0 ) = 1 for all ϑ Θ. Note: P ϑ ( h(y) = 0 ) = 1 can also be written as Pϑ ( h(y) 0 ) = 0, which means that function h(y) may have non-zero values only on a set B Y such thatp(y B) = 0. Then we say that h(y) = 0 almost surely in Y. Definition 2.9. A statistic T(Y) is called complete for the family P = {P ϑ : ϑ Θ} on Y if the family of probability distributions P T = {P ϑ,t : ϑ Θ} is complete for allϑ, that is, E[h(T)] = 0 implies that P{h(T) = 0} = 1. Example Let Y = (Y 1,...,Y n ) be a random sample from a family of Bernoulli(p) distributions for 0 < p < 1. We will show that T(Y) = n Y i is a complete sufficient statistic forp. Sufficiency The pmf of each Y i isp(y i = y i ) = p y i (1 p) 1 y i and the joint pmf fory can be factorized as follows: n P(Y = y) = p y i (1 p) 1 y i = p n y i (1 p) n n y i 1 Hence, T(Y) = n Y i is a sufficient statistic forp.

8 94 CHAPTER 2. ELEMENTS OF STATISTICAL INFERENCE Completeness Now, we know that a sum of independent Bernoulli rvs has a Binomial distribution, i.e, T Bin(n,p) for 0 < p < 1, t = 0,1,...,n. Let h(t) be such that E[h(T)] = 0. Then 0 = E[h(T)] = h(t) n C t p t (1 p) n t t=0 = (1 p) n t=0 ( ) t p h(t) n C t. 1 p The factor (1 p) n 0 for anyp (0,1). Thus, it must be that 0 = = ( ) t p h(t) n C t 1 p h(t) n C t r t t=0 t=0 for all r = p > 0. The last expression is a polynomial of degree n in r. 1 p For the polynomial to be zero for all r the coefficients h(t) n C t must all be zero. It means that h(t) = 0 for all t {0,1,...,n}. Since P(T = t,t {0,1,...,n}) = 1 it means that P{h(T) = 0} = 1 for all p. Hence, T is a complete statistic forp. The following theorem gives a connection between complete and minimal sufficient statistics: Theorem 2.6. IfT(Y) is a complete sufficient statistic for a family of distributions with parameterϑ, then T(Y)is a minimal sufficient statistic for the family. Exercise 2.7. Suppose that Y 1,Y 2,...,Y n is a random sample from a Poisson(λ) distribution. Show that T(Y) = n Y i is a complete sufficient statistic forλ. The following Theorem establishes the minimum variance property of complete sufficient statistics.

9 2.2. POINT ESTIMATION 95 Theorem 2.7. Lehmann-Scheffé Theorem. Let Y = (Y 1,Y 2,...,Y n ) T be a random sample. If S(Y ) is a jointly complete sufficient statistic and T(Y) is an unbiased estimator for φ = g(ϑ) then U = E[T S] is, with probability 1, a unique MVUE of φ. Proof. First, to prove that U is a MVUE of g(ϑ), we show that whatever unbiased estimatort(y)we take we obtain the samee[t S], i.e., the sameu. Then, by Rao-Blackwell Theorem, condition (b), U must be MVUE ofg(ϑ). Suppose thatt(y)and T (Y)are any two unbiased estimators ofg(ϑ). Let U = E[T S] U = E[T S]. Then we have E{U U } = E{E[T S] E[T S]} = E(T) E(T ) = g(ϑ) g(ϑ) = 0. Hence, by completeness ofs(y ) we get P [ U ( S(Y ) ) = U ( S(Y ) )] = 1 for allϑ. This proves the first part of the theorem. Now, we will show uniqueness. Suppose that U and T are two MVUE of g(ϑ). Then if T is a function of the sufficient statistics S(Y ) then, as shown above, it must be equal to U. If T is not a function of S(Y ) then var(u) < var(t ), hence T cannot be a MVUE. Hence, U is a unique MVUE ofg(ϑ). Note: Lehmann-Scheffé Theorem may be used to construct MVUE of g(ϑ) by two methods. Both are based on complete sufficient statisticss(y ). Method 1: If we can find a function of S = S(Y ), say U(S) such that E[U(S)] = g(ϑ) thenu(s) is a unique MVUE ofg(ϑ).

10 96 CHAPTER 2. ELEMENTS OF STATISTICAL INFERENCE Method 2: If we can find any unbiased estimator T = T(Y) of g(ϑ), then U(S) = E[T S] is a unique MVUE ofg(ϑ). Example Method 1. Let Y i iid Bernoulli(p), i = 1,...,n. Earlier we showed that n Y i is a complete sufficient statistic forp. Denote it bys(y ). Y = 1 n n Y i = 1 S(Y ) is an unbiased estimator ofp, hence, as a function of a n complete sufficient statistic, it is the unique MVUE ofp. Now, let g(p) = var(y) = p(1 p). The sample variance 1 n 1 (Y i Y) 2 is an unbiased estimator of g(p). It is in fact a function of the complete sufficient statistics(y) = n Y i. Hence, it is the unique MVUE ofg(p) = p(1 p). Exercise 2.8. Suppose thaty = (Y 1,...,Y n ) T is a random sample from apoisson(λ) distribution. Find a MVUE ofφ = λ 2.

2.3 Methods of Estimation

2.3 Methods of Estimation 96 CHAPTER 2. ELEMENTS OF STATISTICAL INFERENCE 2.3 Methods of Estimation 2.3. Method of Moments The Method of Moments is a simple technique based on the idea that the sample moments are natural estimators

More information

Mathematical statistics

Mathematical statistics October 4 th, 2018 Lecture 12: Information Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation Chapter

More information

Proof In the CR proof. and

Proof In the CR proof. and Question Under what conditions will we be able to attain the Cramér-Rao bound and find a MVUE? Lecture 4 - Consequences of the Cramér-Rao Lower Bound. Searching for a MVUE. Rao-Blackwell Theorem, Lehmann-Scheffé

More information

2.5 Hypothesis Testing

2.5 Hypothesis Testing 118 CHAPTER 2. ELEMENTS OF STATISTICAL INFERENCE 2.5 Hypothesis Testing We assume that Y 1,...,Y n have a joint distribution which depends on the unknown parametersϑ = ϑ 1,...,ϑ p ) T. The set of all possible

More information

ST5215: Advanced Statistical Theory

ST5215: Advanced Statistical Theory Department of Statistics & Applied Probability Wednesday, October 19, 2011 Lecture 17: UMVUE and the first method of derivation Estimable parameters Let ϑ be a parameter in the family P. If there exists

More information

Chapter 3: Unbiased Estimation Lecture 22: UMVUE and the method of using a sufficient and complete statistic

Chapter 3: Unbiased Estimation Lecture 22: UMVUE and the method of using a sufficient and complete statistic Chapter 3: Unbiased Estimation Lecture 22: UMVUE and the method of using a sufficient and complete statistic Unbiased estimation Unbiased or asymptotically unbiased estimation plays an important role in

More information

Mathematical Statistics

Mathematical Statistics Mathematical Statistics Chapter Three. Point Estimation 3.4 Uniformly Minimum Variance Unbiased Estimator(UMVUE) Criteria for Best Estimators MSE Criterion Let F = {p(x; θ) : θ Θ} be a parametric distribution

More information

Example: An experiment can either result in success or failure with probability θ and (1 θ) respectively. The experiment is performed independently

Example: An experiment can either result in success or failure with probability θ and (1 θ) respectively. The experiment is performed independently Chapter 3 Sufficient statistics and variance reduction Let X 1,X 2,...,X n be a random sample from a certain distribution with p.m/d.f fx θ. A function T X 1,X 2,...,X n = T X of these observations is

More information

2.3.3 Method of Least Squares

2.3.3 Method of Least Squares .3. METHODS OF ESTIMATION 105.3.3 Method of Least Squares IfY 1,...,Y n are independent random variables, which have the same variance and higher-order moments, and, if each E(Y i ) is a linear function

More information

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it

More information

Chapters 9. Properties of Point Estimators

Chapters 9. Properties of Point Estimators Chapters 9. Properties of Point Estimators Recap Target parameter, or population parameter θ. Population distribution f(x; θ). { probability function, discrete case f(x; θ) = density, continuous case The

More information

1. (Regular) Exponential Family

1. (Regular) Exponential Family 1. (Regular) Exponential Family The density function of a regular exponential family is: [ ] Example. Poisson(θ) [ ] Example. Normal. (both unknown). ) [ ] [ ] [ ] [ ] 2. Theorem (Exponential family &

More information

[Chapter 6. Functions of Random Variables]

[Chapter 6. Functions of Random Variables] [Chapter 6. Functions of Random Variables] 6.1 Introduction 6.2 Finding the probability distribution of a function of random variables 6.3 The method of distribution functions 6.5 The method of Moment-generating

More information

Lecture 4: UMVUE and unbiased estimators of 0

Lecture 4: UMVUE and unbiased estimators of 0 Lecture 4: UMVUE and unbiased estimators of Problems of approach 1. Approach 1 has the following two main shortcomings. Even if Theorem 7.3.9 or its extension is applicable, there is no guarantee that

More information

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it

More information

February 26, 2017 COMPLETENESS AND THE LEHMANN-SCHEFFE THEOREM

February 26, 2017 COMPLETENESS AND THE LEHMANN-SCHEFFE THEOREM February 26, 2017 COMPLETENESS AND THE LEHMANN-SCHEFFE THEOREM Abstract. The Rao-Blacwell theorem told us how to improve an estimator. We will discuss conditions on when the Rao-Blacwellization of an estimator

More information

2 Statistical Estimation: Basic Concepts

2 Statistical Estimation: Basic Concepts Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof. N. Shimkin 2 Statistical Estimation:

More information

Methods of evaluating estimators and best unbiased estimators Hamid R. Rabiee

Methods of evaluating estimators and best unbiased estimators Hamid R. Rabiee Stochastic Processes Methods of evaluating estimators and best unbiased estimators Hamid R. Rabiee 1 Outline Methods of Mean Squared Error Bias and Unbiasedness Best Unbiased Estimators CR-Bound for variance

More information

STAT 730 Chapter 4: Estimation

STAT 730 Chapter 4: Estimation STAT 730 Chapter 4: Estimation Timothy Hanson Department of Statistics, University of South Carolina Stat 730: Multivariate Analysis 1 / 23 The likelihood We have iid data, at least initially. Each datum

More information

p y (1 p) 1 y, y = 0, 1 p Y (y p) = 0, otherwise.

p y (1 p) 1 y, y = 0, 1 p Y (y p) = 0, otherwise. 1. Suppose Y 1, Y 2,..., Y n is an iid sample from a Bernoulli(p) population distribution, where 0 < p < 1 is unknown. The population pmf is p y (1 p) 1 y, y = 0, 1 p Y (y p) = (a) Prove that Y is the

More information

Chapter 6 Expectation and Conditional Expectation. Lectures Definition 6.1. Two random variables defined on a probability space are said to be

Chapter 6 Expectation and Conditional Expectation. Lectures Definition 6.1. Two random variables defined on a probability space are said to be Chapter 6 Expectation and Conditional Expectation Lectures 24-30 In this chapter, we introduce expected value or the mean of a random variable. First we define expectation for discrete random variables

More information

ENEE 621 SPRING 2016 DETECTION AND ESTIMATION THEORY THE PARAMETER ESTIMATION PROBLEM

ENEE 621 SPRING 2016 DETECTION AND ESTIMATION THEORY THE PARAMETER ESTIMATION PROBLEM c 2007-2016 by Armand M. Makowski 1 ENEE 621 SPRING 2016 DETECTION AND ESTIMATION THEORY THE PARAMETER ESTIMATION PROBLEM 1 The basic setting Throughout, p, q and k are positive integers. The setup With

More information

Miscellaneous Errors in the Chapter 6 Solutions

Miscellaneous Errors in the Chapter 6 Solutions Miscellaneous Errors in the Chapter 6 Solutions 3.30(b In this problem, early printings of the second edition use the beta(a, b distribution, but later versions use the Poisson(λ distribution. If your

More information

STATISTICAL METHODS FOR SIGNAL PROCESSING c Alfred Hero

STATISTICAL METHODS FOR SIGNAL PROCESSING c Alfred Hero STATISTICAL METHODS FOR SIGNAL PROCESSING c Alfred Hero 1999 32 Statistic used Meaning in plain english Reduction ratio T (X) [X 1,..., X n ] T, entire data sample RR 1 T (X) [X (1),..., X (n) ] T, rank

More information

λ n e λ n f(x;λ) f(y;λ) = λn e λ = e λ ( n = 2 θ θ 1 2 y2 1 3 y3 ] θ = 1 3 θ. E(Y i ) = 3 1 n n1 3 θ = θ. θ 1 3 y3 1 4 y4 ] θ = 1 6 θ2.

λ n e λ n f(x;λ) f(y;λ) = λn e λ = e λ ( n = 2 θ θ 1 2 y2 1 3 y3 ] θ = 1 3 θ. E(Y i ) = 3 1 n n1 3 θ = θ. θ 1 3 y3 1 4 y4 ] θ = 1 6 θ2. CHAPTER Exercise 1 Suppose that Y (Y 1,,Y ) is a radom sample from a Exp(λ) distributio The we may write f Y (y) λe λy i λ e λ y i }{{} }{{} 1 g λ (T(Y )) h(y) It follows thatt(y ) Y i is a sufficiet statistic

More information

f(y θ) = g(t (y) θ)h(y)

f(y θ) = g(t (y) θ)h(y) EXAM3, FINAL REVIEW (and a review for some of the QUAL problems): No notes will be allowed, but you may bring a calculator. Memorize the pmf or pdf f, E(Y ) and V(Y ) for the following RVs: 1) beta(δ,

More information

Estimation theory. Parametric estimation. Properties of estimators. Minimum variance estimator. Cramer-Rao bound. Maximum likelihood estimators

Estimation theory. Parametric estimation. Properties of estimators. Minimum variance estimator. Cramer-Rao bound. Maximum likelihood estimators Estimation theory Parametric estimation Properties of estimators Minimum variance estimator Cramer-Rao bound Maximum likelihood estimators Confidence intervals Bayesian estimation 1 Random Variables Let

More information

A Few Notes on Fisher Information (WIP)

A Few Notes on Fisher Information (WIP) A Few Notes on Fisher Information (WIP) David Meyer dmm@{-4-5.net,uoregon.edu} Last update: April 30, 208 Definitions There are so many interesting things about Fisher Information and its theoretical properties

More information

STATISTICS/ECONOMETRICS PREP COURSE PROF. MASSIMO GUIDOLIN

STATISTICS/ECONOMETRICS PREP COURSE PROF. MASSIMO GUIDOLIN Massimo Guidolin Massimo.Guidolin@unibocconi.it Dept. of Finance STATISTICS/ECONOMETRICS PREP COURSE PROF. MASSIMO GUIDOLIN SECOND PART, LECTURE 2: MODES OF CONVERGENCE AND POINT ESTIMATION Lecture 2:

More information

Review Quiz. 1. Prove that in a one-dimensional canonical exponential family, the complete and sufficient statistic achieves the

Review Quiz. 1. Prove that in a one-dimensional canonical exponential family, the complete and sufficient statistic achieves the Review Quiz 1. Prove that in a one-dimensional canonical exponential family, the complete and sufficient statistic achieves the Cramér Rao lower bound (CRLB). That is, if where { } and are scalars, then

More information

Exercises and Answers to Chapter 1

Exercises and Answers to Chapter 1 Exercises and Answers to Chapter The continuous type of random variable X has the following density function: a x, if < x < a, f (x), otherwise. Answer the following questions. () Find a. () Obtain mean

More information

Parameter Estimation

Parameter Estimation Parameter Estimation Consider a sample of observations on a random variable Y. his generates random variables: (y 1, y 2,, y ). A random sample is a sample (y 1, y 2,, y ) where the random variables y

More information

MATH4427 Notebook 2 Fall Semester 2017/2018

MATH4427 Notebook 2 Fall Semester 2017/2018 MATH4427 Notebook 2 Fall Semester 2017/2018 prepared by Professor Jenny Baglivo c Copyright 2009-2018 by Jenny A. Baglivo. All Rights Reserved. 2 MATH4427 Notebook 2 3 2.1 Definitions and Examples...................................

More information

HT Introduction. P(X i = x i ) = e λ λ x i

HT Introduction. P(X i = x i ) = e λ λ x i MODS STATISTICS Introduction. HT 2012 Simon Myers, Department of Statistics (and The Wellcome Trust Centre for Human Genetics) myers@stats.ox.ac.uk We will be concerned with the mathematical framework

More information

Probability Models. 4. What is the definition of the expectation of a discrete random variable?

Probability Models. 4. What is the definition of the expectation of a discrete random variable? 1 Probability Models The list of questions below is provided in order to help you to prepare for the test and exam. It reflects only the theoretical part of the course. You should expect the questions

More information

i=1 k i=1 g i (Y )] = k

i=1 k i=1 g i (Y )] = k Math 483 EXAM 2 covers 2.4, 2.5, 2.7, 2.8, 3.1, 3.2, 3.3, 3.4, 3.8, 3.9, 4.1, 4.2, 4.3, 4.4, 4.5, 4.6, 4.9, 5.1, 5.2, and 5.3. The exam is on Thursday, Oct. 13. You are allowed THREE SHEETS OF NOTES and

More information

Lecture 1: August 28

Lecture 1: August 28 36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 1: August 28 Our broad goal for the first few lectures is to try to understand the behaviour of sums of independent random

More information

ECE531 Lecture 8: Non-Random Parameter Estimation

ECE531 Lecture 8: Non-Random Parameter Estimation ECE531 Lecture 8: Non-Random Parameter Estimation D. Richard Brown III Worcester Polytechnic Institute 19-March-2009 Worcester Polytechnic Institute D. Richard Brown III 19-March-2009 1 / 25 Introduction

More information

Lecture 13 and 14: Bayesian estimation theory

Lecture 13 and 14: Bayesian estimation theory 1 Lecture 13 and 14: Bayesian estimation theory Spring 2012 - EE 194 Networked estimation and control (Prof. Khan) March 26 2012 I. BAYESIAN ESTIMATORS Mother Nature conducts a random experiment that generates

More information

Mathematical Statistics 1 Math A 6330

Mathematical Statistics 1 Math A 6330 Mathematical Statistics 1 Math A 6330 Chapter 3 Common Families of Distributions Mohamed I. Riffi Department of Mathematics Islamic University of Gaza September 28, 2015 Outline 1 Subjects of Lecture 04

More information

Chapter 5 Class Notes

Chapter 5 Class Notes Chapter 5 Class Notes Sections 5.1 and 5.2 It is quite common to measure several variables (some of which may be correlated) and to examine the corresponding joint probability distribution One example

More information

Final Exam. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given.

Final Exam. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given. (a) If X and Y are independent, Corr(X, Y ) = 0. (b) (c) (d) (e) A consistent estimator must be asymptotically

More information

Completeness. On the other hand, the distribution of an ancillary statistic doesn t depend on θ at all.

Completeness. On the other hand, the distribution of an ancillary statistic doesn t depend on θ at all. Completeness A minimal sufficient statistic achieves the maximum amount of data reduction while retaining all the information the sample has concerning θ. On the other hand, the distribution of an ancillary

More information

Mathematical statistics

Mathematical statistics October 18 th, 2018 Lecture 16: Midterm review Countdown to mid-term exam: 7 days Week 1 Chapter 1: Probability review Week 2 Week 4 Week 7 Chapter 6: Statistics Chapter 7: Point Estimation Chapter 8:

More information

Classical Estimation Topics

Classical Estimation Topics Classical Estimation Topics Namrata Vaswani, Iowa State University February 25, 2014 This note fills in the gaps in the notes already provided (l0.pdf, l1.pdf, l2.pdf, l3.pdf, LeastSquares.pdf). 1 Min

More information

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A. 1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n

More information

ECE 275B Homework # 1 Solutions Version Winter 2015

ECE 275B Homework # 1 Solutions Version Winter 2015 ECE 275B Homework # 1 Solutions Version Winter 2015 1. (a) Because x i are assumed to be independent realizations of a continuous random variable, it is almost surely (a.s.) 1 the case that x 1 < x 2

More information

Variations. ECE 6540, Lecture 10 Maximum Likelihood Estimation

Variations. ECE 6540, Lecture 10 Maximum Likelihood Estimation Variations ECE 6540, Lecture 10 Last Time BLUE (Best Linear Unbiased Estimator) Formulation Advantages Disadvantages 2 The BLUE A simplification Assume the estimator is a linear system For a single parameter

More information

Maximum Likelihood Estimation

Maximum Likelihood Estimation Connexions module: m11446 1 Maximum Likelihood Estimation Clayton Scott Robert Nowak This work is produced by The Connexions Project and licensed under the Creative Commons Attribution License Abstract

More information

STAT 512 sp 2018 Summary Sheet

STAT 512 sp 2018 Summary Sheet STAT 5 sp 08 Summary Sheet Karl B. Gregory Spring 08. Transformations of a random variable Let X be a rv with support X and let g be a function mapping X to Y with inverse mapping g (A = {x X : g(x A}

More information

Mathematical statistics

Mathematical statistics October 1 st, 2018 Lecture 11: Sufficient statistic Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation

More information

Midterm Examination. STA 215: Statistical Inference. Due Wednesday, 2006 Mar 8, 1:15 pm

Midterm Examination. STA 215: Statistical Inference. Due Wednesday, 2006 Mar 8, 1:15 pm Midterm Examination STA 215: Statistical Inference Due Wednesday, 2006 Mar 8, 1:15 pm This is an open-book take-home examination. You may work on it during any consecutive 24-hour period you like; please

More information

March 10, 2017 THE EXPONENTIAL CLASS OF DISTRIBUTIONS

March 10, 2017 THE EXPONENTIAL CLASS OF DISTRIBUTIONS March 10, 2017 THE EXPONENTIAL CLASS OF DISTRIBUTIONS Abstract. We will introduce a class of distributions that will contain many of the discrete and continuous we are familiar with. This class will help

More information

STA 260: Statistics and Probability II

STA 260: Statistics and Probability II Al Nosedal. University of Toronto. Winter 2017 1 Properties of Point Estimators and Methods of Estimation 2 3 If you can t explain it simply, you don t understand it well enough Albert Einstein. Definition

More information

Detection & Estimation Lecture 1

Detection & Estimation Lecture 1 Detection & Estimation Lecture 1 Intro, MVUE, CRLB Xiliang Luo General Course Information Textbooks & References Fundamentals of Statistical Signal Processing: Estimation Theory/Detection Theory, Steven

More information

ECE 275B Homework # 1 Solutions Winter 2018

ECE 275B Homework # 1 Solutions Winter 2018 ECE 275B Homework # 1 Solutions Winter 2018 1. (a) Because x i are assumed to be independent realizations of a continuous random variable, it is almost surely (a.s.) 1 the case that x 1 < x 2 < < x n Thus,

More information

Graduate Econometrics I: Unbiased Estimation

Graduate Econometrics I: Unbiased Estimation Graduate Econometrics I: Unbiased Estimation Yves Dominicy Université libre de Bruxelles Solvay Brussels School of Economics and Management ECARES Yves Dominicy Graduate Econometrics I: Unbiased Estimation

More information

Contents 1. Contents

Contents 1. Contents Contents 1 Contents 6 Distributions of Functions of Random Variables 2 6.1 Transformation of Discrete r.v.s............. 3 6.2 Method of Distribution Functions............. 6 6.3 Method of Transformations................

More information

1 Probability Model. 1.1 Types of models to be discussed in the course

1 Probability Model. 1.1 Types of models to be discussed in the course Sufficiency January 11, 2016 Debdeep Pati 1 Probability Model Model: A family of distributions {P θ : θ Θ}. P θ (B) is the probability of the event B when the parameter takes the value θ. P θ is described

More information

Statistical Inference

Statistical Inference Statistical Inference Classical and Bayesian Methods Revision Class for Midterm Exam AMS-UCSC Th Feb 9, 2012 Winter 2012. Session 1 (Revision Class) AMS-132/206 Th Feb 9, 2012 1 / 23 Topics Topics We will

More information

Lecture 11. Multivariate Normal theory

Lecture 11. Multivariate Normal theory 10. Lecture 11. Multivariate Normal theory Lecture 11. Multivariate Normal theory 1 (1 1) 11. Multivariate Normal theory 11.1. Properties of means and covariances of vectors Properties of means and covariances

More information

Last Lecture. Biostatistics Statistical Inference Lecture 14 Obtaining Best Unbiased Estimator. Related Theorems. Rao-Blackwell Theorem

Last Lecture. Biostatistics Statistical Inference Lecture 14 Obtaining Best Unbiased Estimator. Related Theorems. Rao-Blackwell Theorem Last Lecture Biostatistics 62 - Statistical Inference Lecture 14 Obtaining Best Unbiased Estimator Hyun Min Kang February 28th, 213 For single-parameter exponential family, is Cramer-Rao bound always attainable?

More information

More than one variable

More than one variable Chapter More than one variable.1 Bivariate discrete distributions Suppose that the r.v. s X and Y are discrete and take on the values x j and y j, j 1, respectively. Then the joint p.d.f. of X and Y, to

More information

Define characteristic function. State its properties. State and prove inversion theorem.

Define characteristic function. State its properties. State and prove inversion theorem. ASSIGNMENT - 1, MAY 013. Paper I PROBABILITY AND DISTRIBUTION THEORY (DMSTT 01) 1. (a) Give the Kolmogorov definition of probability. State and prove Borel cantelli lemma. Define : (i) distribution function

More information

Detection & Estimation Lecture 1

Detection & Estimation Lecture 1 Detection & Estimation Lecture 1 Intro, MVUE, CRLB Xiliang Luo General Course Information Textbooks & References Fundamentals of Statistical Signal Processing: Estimation Theory/Detection Theory, Steven

More information

Estimation MLE-Pandemic data MLE-Financial crisis data Evaluating estimators. Estimation. September 24, STAT 151 Class 6 Slide 1

Estimation MLE-Pandemic data MLE-Financial crisis data Evaluating estimators. Estimation. September 24, STAT 151 Class 6 Slide 1 Estimation September 24, 2018 STAT 151 Class 6 Slide 1 Pandemic data Treatment outcome, X, from n = 100 patients in a pandemic: 1 = recovered and 0 = not recovered 1 1 1 0 0 0 1 1 1 0 0 1 0 1 0 0 1 1 1

More information

A General Overview of Parametric Estimation and Inference Techniques.

A General Overview of Parametric Estimation and Inference Techniques. A General Overview of Parametric Estimation and Inference Techniques. Moulinath Banerjee University of Michigan September 11, 2012 The object of statistical inference is to glean information about an underlying

More information

1 Complete Statistics

1 Complete Statistics Complete Statistics February 4, 2016 Debdeep Pati 1 Complete Statistics Suppose X P θ, θ Θ. Let (X (1),..., X (n) ) denote the order statistics. Definition 1. A statistic T = T (X) is complete if E θ g(t

More information

Chapter 8: Least squares (beginning of chapter)

Chapter 8: Least squares (beginning of chapter) Chapter 8: Least squares (beginning of chapter) Least Squares So far, we have been trying to determine an estimator which was unbiased and had minimum variance. Next we ll consider a class of estimators

More information

ECE531 Lecture 10b: Maximum Likelihood Estimation

ECE531 Lecture 10b: Maximum Likelihood Estimation ECE531 Lecture 10b: Maximum Likelihood Estimation D. Richard Brown III Worcester Polytechnic Institute 05-Apr-2011 Worcester Polytechnic Institute D. Richard Brown III 05-Apr-2011 1 / 23 Introduction So

More information

McGill University. Faculty of Science. Department of Mathematics and Statistics. Part A Examination. Statistics: Theory Paper

McGill University. Faculty of Science. Department of Mathematics and Statistics. Part A Examination. Statistics: Theory Paper McGill University Faculty of Science Department of Mathematics and Statistics Part A Examination Statistics: Theory Paper Date: 10th May 2015 Instructions Time: 1pm-5pm Answer only two questions from Section

More information

i=1 k i=1 g i (Y )] = k f(t)dt and f(y) = F (y) except at possibly countably many points, E[g(Y )] = f(y)dy = 1, F(y) = y

i=1 k i=1 g i (Y )] = k f(t)dt and f(y) = F (y) except at possibly countably many points, E[g(Y )] = f(y)dy = 1, F(y) = y Math 480 Exam 2 is Wed. Oct. 31. You are allowed 7 sheets of notes and a calculator. The exam emphasizes HW5-8, and Q5-8. From the 1st exam: The conditional probability of A given B is P(A B) = P(A B)

More information

Joint p.d.f. and Independent Random Variables

Joint p.d.f. and Independent Random Variables 1 Joint p.d.f. and Independent Random Variables Let X and Y be two discrete r.v. s and let R be the corresponding space of X and Y. The joint p.d.f. of X = x and Y = y, denoted by f(x, y) = P(X = x, Y

More information

Review. December 4 th, Review

Review. December 4 th, Review December 4 th, 2017 Att. Final exam: Course evaluation Friday, 12/14/2018, 10:30am 12:30pm Gore Hall 115 Overview Week 2 Week 4 Week 7 Week 10 Week 12 Chapter 6: Statistics and Sampling Distributions Chapter

More information

1 Probability Model. 1.1 Types of models to be discussed in the course

1 Probability Model. 1.1 Types of models to be discussed in the course Sufficiency January 18, 016 Debdeep Pati 1 Probability Model Model: A family of distributions P θ : θ Θ}. P θ (B) is the probability of the event B when the parameter takes the value θ. P θ is described

More information

Solution-Midterm Examination - I STAT 421, Spring 2012

Solution-Midterm Examination - I STAT 421, Spring 2012 Solution-Midterm Examination - I STAT 4, Spring 0 Prof. Prem K. Goel. [0 points] Let X, X,..., X n be a random sample of size n from a Gamma population with β = and α = ν/, where ν is the unknown parameter.

More information

Math 3215 Intro. Probability & Statistics Summer 14. Homework 5: Due 7/3/14

Math 3215 Intro. Probability & Statistics Summer 14. Homework 5: Due 7/3/14 Math 325 Intro. Probability & Statistics Summer Homework 5: Due 7/3/. Let X and Y be continuous random variables with joint/marginal p.d.f. s f(x, y) 2, x y, f (x) 2( x), x, f 2 (y) 2y, y. Find the conditional

More information

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review STATS 200: Introduction to Statistical Inference Lecture 29: Course review Course review We started in Lecture 1 with a fundamental assumption: Data is a realization of a random process. The goal throughout

More information

Evaluating the Performance of Estimators (Section 7.3)

Evaluating the Performance of Estimators (Section 7.3) Evaluating the Performance of Estimators (Section 7.3) Example: Suppose we observe X 1,..., X n iid N(θ, σ 2 0 ), with σ2 0 known, and wish to estimate θ. Two possible estimators are: ˆθ = X sample mean

More information

t x 1 e t dt, and simplify the answer when possible (for example, when r is a positive even number). In particular, confirm that EX 4 = 3.

t x 1 e t dt, and simplify the answer when possible (for example, when r is a positive even number). In particular, confirm that EX 4 = 3. Mathematical Statistics: Homewor problems General guideline. While woring outside the classroom, use any help you want, including people, computer algebra systems, Internet, and solution manuals, but mae

More information

Probability and Distributions

Probability and Distributions Probability and Distributions What is a statistical model? A statistical model is a set of assumptions by which the hypothetical population distribution of data is inferred. It is typically postulated

More information

Theory of Maximum Likelihood Estimation. Konstantin Kashin

Theory of Maximum Likelihood Estimation. Konstantin Kashin Gov 2001 Section 5: Theory of Maximum Likelihood Estimation Konstantin Kashin February 28, 2013 Outline Introduction Likelihood Examples of MLE Variance of MLE Asymptotic Properties What is Statistical

More information

Chapter 8.8.1: A factorization theorem

Chapter 8.8.1: A factorization theorem LECTURE 14 Chapter 8.8.1: A factorization theorem The characterization of a sufficient statistic in terms of the conditional distribution of the data given the statistic can be difficult to work with.

More information

Mathematics Qualifying Examination January 2015 STAT Mathematical Statistics

Mathematics Qualifying Examination January 2015 STAT Mathematical Statistics Mathematics Qualifying Examination January 2015 STAT 52800 - Mathematical Statistics NOTE: Answer all questions completely and justify your derivations and steps. A calculator and statistical tables (normal,

More information

Advanced Signal Processing Introduction to Estimation Theory

Advanced Signal Processing Introduction to Estimation Theory Advanced Signal Processing Introduction to Estimation Theory Danilo Mandic, room 813, ext: 46271 Department of Electrical and Electronic Engineering Imperial College London, UK d.mandic@imperial.ac.uk,

More information

ECE534, Spring 2018: Solutions for Problem Set #3

ECE534, Spring 2018: Solutions for Problem Set #3 ECE534, Spring 08: Solutions for Problem Set #3 Jointly Gaussian Random Variables and MMSE Estimation Suppose that X, Y are jointly Gaussian random variables with µ X = µ Y = 0 and σ X = σ Y = Let their

More information

Limiting Distributions

Limiting Distributions Limiting Distributions We introduce the mode of convergence for a sequence of random variables, and discuss the convergence in probability and in distribution. The concept of convergence leads us to the

More information

Mathematics Ph.D. Qualifying Examination Stat Probability, January 2018

Mathematics Ph.D. Qualifying Examination Stat Probability, January 2018 Mathematics Ph.D. Qualifying Examination Stat 52800 Probability, January 2018 NOTE: Answers all questions completely. Justify every step. Time allowed: 3 hours. 1. Let X 1,..., X n be a random sample from

More information

Chapter 5: Joint Probability Distributions

Chapter 5: Joint Probability Distributions Chapter 5: Joint Probability Distributions Seungchul Baek Department of Statistics, University of South Carolina STAT 509: Statistics for Engineers 1 / 19 Joint pmf Definition: The joint probability mass

More information

1 of 7 7/16/2009 6:12 AM Virtual Laboratories > 7. Point Estimation > 1 2 3 4 5 6 1. Estimators The Basic Statistical Model As usual, our starting point is a random experiment with an underlying sample

More information

Probability Theory and Statistics. Peter Jochumzen

Probability Theory and Statistics. Peter Jochumzen Probability Theory and Statistics Peter Jochumzen April 18, 2016 Contents 1 Probability Theory And Statistics 3 1.1 Experiment, Outcome and Event................................ 3 1.2 Probability............................................

More information

Elements of statistics (MATH0487-1)

Elements of statistics (MATH0487-1) Elements of statistics (MATH0487-1) Prof. Dr. Dr. K. Van Steen University of Liège, Belgium November 12, 2012 Introduction to Statistics Basic Probability Revisited Sampling Exploratory Data Analysis -

More information

The Delta Method and Applications

The Delta Method and Applications Chapter 5 The Delta Method and Applications 5.1 Local linear approximations Suppose that a particular random sequence converges in distribution to a particular constant. The idea of using a first-order

More information

Limiting Distributions

Limiting Distributions We introduce the mode of convergence for a sequence of random variables, and discuss the convergence in probability and in distribution. The concept of convergence leads us to the two fundamental results

More information

EIE6207: Estimation Theory

EIE6207: Estimation Theory EIE6207: Estimation Theory Man-Wai MAK Dept. of Electronic and Information Engineering, The Hong Kong Polytechnic University enmwmak@polyu.edu.hk http://www.eie.polyu.edu.hk/ mwmak References: Steven M.

More information

557: MATHEMATICAL STATISTICS II BIAS AND VARIANCE

557: MATHEMATICAL STATISTICS II BIAS AND VARIANCE 557: MATHEMATICAL STATISTICS II BIAS AND VARIANCE An estimator, T (X), of θ can be evaluated via its statistical properties. Typically, two aspects are considered: Expectation Variance either in terms

More information

Distributions of Functions of Random Variables. 5.1 Functions of One Random Variable

Distributions of Functions of Random Variables. 5.1 Functions of One Random Variable Distributions of Functions of Random Variables 5.1 Functions of One Random Variable 5.2 Transformations of Two Random Variables 5.3 Several Random Variables 5.4 The Moment-Generating Function Technique

More information

ELEG 5633 Detection and Estimation Minimum Variance Unbiased Estimators (MVUE)

ELEG 5633 Detection and Estimation Minimum Variance Unbiased Estimators (MVUE) 1 ELEG 5633 Detection and Estimation Minimum Variance Unbiased Estimators (MVUE) Jingxian Wu Department of Electrical Engineering University of Arkansas Outline Minimum Variance Unbiased Estimators (MVUE)

More information

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems Review of Basic Probability The fundamentals, random variables, probability distributions Probability mass/density functions

More information

STAT 430/510 Probability

STAT 430/510 Probability STAT 430/510 Probability Hui Nie Lecture 16 June 24th, 2009 Review Sum of Independent Normal Random Variables Sum of Independent Poisson Random Variables Sum of Independent Binomial Random Variables Conditional

More information