2.2.2 Comparing estimators
|
|
- Jasper Lawrence
- 5 years ago
- Views:
Transcription
1 2.2. POINT ESTIMATION Comparing estimators Estimators can be compared through their mean square errors. If they are unbiased, this is equivalent to comparing their variances. In many applications, we try to find an unbiased estimator which has minimum variance, or at least low variance. We can compare the variances of several unbiased estimators, but how do we know if we have found an estimator with the lowest variance among all unbiased estimators? The answer is given by Cramér-Rao lower bound (CRLB). The CRLB for unbiased estimators ofφ = g(ϑ), when there is a single parameter, is CRLB(φ) = E { } 2 dg(ϑ) { }, d2 logf(y ;ϑ) 2 where Y = (Y 1,...,Y n ) T and f(y;ϑ) is the joint density function (pdf) of Y 1,...,Y n when Y is continuous. For discrete r.vs we replace pdf with a pmf. E{ d 2 logf(y;ϑ)/ 2 } is called the Fisher information. When there arepparameters, ϑ = (ϑ 1,...,ϑ p ) T, andφ = g(ϑ), the CRLB is ( g CRLB(φ) =,..., g ) M 1, ϑ 1 ϑ p where the matrix M has (i,j)th element E{ 2 logf(y;ϑ)/ ϑ i ϑ j } and is called the Fisher Information Matrix. g ϑ 1. g ϑ p For the following result, we need to assume that the sample space does not depend on ϑ 1,...,ϑ p, and that the differentiability and continuity constraints for all ϑ Θ hold, so that the order of integration and differentiation can be interchanged, and the Fisher Information is positive and finite. Problems which satisfy these conditions are called regular. Theorem 2.4. Cramér-Rao inequality. For any unbiased estimator,t(y ), ofφif the estimation problem is regular then var{t(y)} CRLB(φ).
2 88 CHAPTER 2. ELEMENTS OF STATISTICAL INFERENCE Proof. For simplicity, we only prove the result in the one-parameter, continuous case. The proof uses the result that, for any two random variables U and V, their correlation coefficient ρ(u,v) lies between -1 and +1, that is, {ρ(u,v)} 2 1, which may be written as var(u) {cov(u,v)}2. (2.2) var(v) LetU = T(Y ) and V = dlogf(y;ϑ)/. Then we may write { } dlogf(y ;ϑ) E(V) = E [ d = E f(y;ϑ) ] property of derivative of log f(y;ϑ) d =... f(y;ϑ) f(y;ϑ) f(y;ϑ)dy 1...dy n df(y;ϑ) =... dy 1...dy n = d... f(y;ϑ)dy 1...dy n = d (1) = 0. Next, sincee(v) = 0, we have { } dlogf(y ;ϑ) cov(u,v) = E(UV) = E T(Y) { } d f(y ;ϑ) = E T(Y) f(y ;ϑ) d =... T(y) f(y;ϑ) f(y;ϑ) f(y;ϑ)dy 1...dy n =... T(y) df(y;ϑ) dy 1...dy n = d... T(y)f(y;ϑ)dy 1...dy n = d dg(ϑ) E{T(Y)} =.
3 2.2. POINT ESTIMATION 89 Finally, again sincee(v) = 0, we have var(v) = E(V 2 ) [ {dlogf(y;ϑ) } ] 2 = E { d dlogf(y;ϑ) = E f(y;ϑ) } f(y;ϑ) { d dlogf(y;ϑ) =... f(y;ϑ) } f(y;ϑ)dy 1...dy n f(y;ϑ) dlogf(y;ϑ) df(y;ϑ) =... dy 1...dy n. Now, since we can write d 2 f = d 2 ( f dlogf ) = df dlogf logf +fd2, 2 we obtain d 2 f(y;ϑ) var(v) =... dy dy n... f(y;ϑ) d2 logf(y;ϑ) dy dy n { } = d2... f(y;ϑ)dy dy n +E d2 logf(y;ϑ) { } { 2 } = d2 2(1)+E d2 logf(y;ϑ) = E d2 logf(y;ϑ). 2 2 The required result then follows immediately from (2.2). Example Suppose thaty 1,...,Y n are independentpoisson(λ) random variables. Let g(λ) = λ. Then we have dg/dλ = 1. Here we have a discrete r.v. whose joint pmf can be written as the following product P(Y 1 = y 1,...,Y n = y n ;λ) = n λ y i e λ y i! = λ n y i e nλ n y i!,
4 90 CHAPTER 2. ELEMENTS OF STATISTICAL INFERENCE and so taking log we obtain logp = ( n ) y i logλ nλ log y i!. Thus, we have dlogp n = y i n dλ λ and d 2 logp n = y i. dλ 2 λ 2 Hence, the Fisher information is ( ) E d2 logp dλ 2 ( ) = 1 λ 2E Y i = 1 E(Y λ 2 i ) = 1 λ 2nλ = n λ and CRLB(λ) = λ/n. The efficiency of an asymptotically unbiased estimator,t(y), ofφis eff(t) = lim n CRLB(φ) var{t(y)}. An asymptotically unbiased estimator, T(Y ), of φ is said to be efficient if its efficiency is equal to one. In this case, for large n, var{t(y)} is approximately equal to the CRLB. Exercise 2.5. Suppose thaty 1,Y 2,...,Y n is an independent sample from abernoulli(p) distribution. Calculate the CRLB(p) and see if Y is an unbiased estimator of p with minimum variance. Exercise 2.6. Suppose thaty 1,Y 2,...,Y n is an independent sample fromn(µ,σ 2 ). Derive the Fisher Information Matrix M for ϑ = (µ,σ 2 ) T and find the Cramer- Rao lower bound for an unbiased estimator ofg(ϑ) = µ+σ 2.
5 2.2. POINT ESTIMATION Minimum Variance Unbiased Estimators If an unbiased estimator has the variance equal to the CRLB, it must have the minimum variance amongst all unbiased estimators. We call it the minimum variance unbiased estimator (MVUE) ofφ. Sufficiency is a powerful property in finding unbiased, minimum variance estimators. If T(Y) is an unbiased estimator of ϑ and S is a statistic sufficient for ϑ, then there is a function of S that is also an unbiased estimator of ϑ and has no larger variance than the variance of T(Y). The following theorem formalizes this statement. Theorem 2.5. Rao-Blackwell theorem. Let Y = (Y 1,Y 2,...,Y n ) T be a random sample, S = (S 1,...,S p ) T be jointly sufficient statistics for ϑ = (ϑ 1,...,ϑ p ) T and T(Y ) (which is not a function of S) be an unbiased estimator of φ = g(ϑ). Then, U = E(T S) is a statistic such that (a) E(U) = φ, so thatu is an unbiased estimator of φ, and (b) var(u) < var(t). Proof. First, we note thatu is a statistic. Indeed, sinces are jointly sufficient for ϑ, the conditional distributiony S does not depend on the parameters and so the conditional distribution of a function T(Y ) given S, T S, does not depend on ϑ either. Thus,U = E(T S) is a function of the random sample only, not a function ofϑ, therefore it is a statistic. Next, we will use the known facts about the conditional expectation and variance given Exercise SinceT is an unbiased estimator ofφ, we have E(U) = E[E(T S)] = E(T) = φ. So U is also an unbiased estimator ofφ, which proves (a). Finally, we get var(t) = var[e(t S)]+E[var(T S)] = var(u)+e[var(t S)]. However, since T is not a function of S we have var(t S) > 0, thus, it follows thate[var(t S)] > 0, and hence (b) is proved.
6 92 CHAPTER 2. ELEMENTS OF STATISTICAL INFERENCE It means that, if we have an unbiased estimator, T, of φ, which is not a function of the sufficient statistics, we can always find an unbiased estimator which has smaller variance, namely U = E(T S 1,...,S p ) which is a function of S. We thus have the following result. Corollary 2.1. MVUEs must be functions of sufficient statistics. Example Suppose thaty 1,...,Y n are independentpoisson(λ) random variables. Then T = Y i, for any i = 1,...,n, is an unbiased estimator of λ. Also, S = n Y i is a sufficient statistic forλand T is not a function ofs. Hence, a better unbiased estimator is given by any ofe(y 1 n Y i),...,e(y n n Y i). Now, sincee(y 1 n Y i) =... = E(Y n n Y i) and E(Y 1 Y i )+E(Y 2 Y i )+...+E(Y n Y i ) = E [ n Y i we have ne(y 1 Y i ) = Y i. ] Y i = Y i, Hence U = E(Y 1 n Y i) = Y is a better unbiased estimator of λ than Y 1 or than any of Y i. In fact, Y is a MVUE of λ as its variance is equal to the Cramer- Rao Lower Bound forλ. (See Example 2.12) Complete sufficient statistics T(Y)is a function of random sampley = (Y 1,Y 2,...,Y n ) and so it is a random variable as well. Hence, we may ask about the distribution oft(y). For example, assume that σ 2 is known equal to σo. 2 Then, to make inference about µ we may reduce the random sample to its mean. We know that T(Y) = Y N ( ) µ, σ2 o n ify i N(µ,σo 2 ) and we may write iid f T (t;µ σ 2 o ) = 1 2πσo e n(t µ)2 /2σ 2 o. Due to this data reduction we can make inference about µ based on the distribution of Y only rather than on the multivariate distribution of the whole random
7 2.2. POINT ESTIMATION 93 sampley. A minimal sufficient statistic reduces data maximally while retaining all the information about the parameter ϑ. We would also like such a statistic to be independent of any so called ancillary functions of the random sample whose distributions do not depend on the parameter of interest. Such an independent statistic is called complete. First, we introduce a notion of a complete family of distributions. Definition 2.8. A family of distributionsp = {P ϑ : ϑ Θ} defined on a common spacey is called complete if for any real measurable functionh(y) E[h(Y)] = 0 implies that P ϑ ( h(y) = 0 ) = 1 for all ϑ Θ. Note: P ϑ ( h(y) = 0 ) = 1 can also be written as Pϑ ( h(y) 0 ) = 0, which means that function h(y) may have non-zero values only on a set B Y such thatp(y B) = 0. Then we say that h(y) = 0 almost surely in Y. Definition 2.9. A statistic T(Y) is called complete for the family P = {P ϑ : ϑ Θ} on Y if the family of probability distributions P T = {P ϑ,t : ϑ Θ} is complete for allϑ, that is, E[h(T)] = 0 implies that P{h(T) = 0} = 1. Example Let Y = (Y 1,...,Y n ) be a random sample from a family of Bernoulli(p) distributions for 0 < p < 1. We will show that T(Y) = n Y i is a complete sufficient statistic forp. Sufficiency The pmf of each Y i isp(y i = y i ) = p y i (1 p) 1 y i and the joint pmf fory can be factorized as follows: n P(Y = y) = p y i (1 p) 1 y i = p n y i (1 p) n n y i 1 Hence, T(Y) = n Y i is a sufficient statistic forp.
8 94 CHAPTER 2. ELEMENTS OF STATISTICAL INFERENCE Completeness Now, we know that a sum of independent Bernoulli rvs has a Binomial distribution, i.e, T Bin(n,p) for 0 < p < 1, t = 0,1,...,n. Let h(t) be such that E[h(T)] = 0. Then 0 = E[h(T)] = h(t) n C t p t (1 p) n t t=0 = (1 p) n t=0 ( ) t p h(t) n C t. 1 p The factor (1 p) n 0 for anyp (0,1). Thus, it must be that 0 = = ( ) t p h(t) n C t 1 p h(t) n C t r t t=0 t=0 for all r = p > 0. The last expression is a polynomial of degree n in r. 1 p For the polynomial to be zero for all r the coefficients h(t) n C t must all be zero. It means that h(t) = 0 for all t {0,1,...,n}. Since P(T = t,t {0,1,...,n}) = 1 it means that P{h(T) = 0} = 1 for all p. Hence, T is a complete statistic forp. The following theorem gives a connection between complete and minimal sufficient statistics: Theorem 2.6. IfT(Y) is a complete sufficient statistic for a family of distributions with parameterϑ, then T(Y)is a minimal sufficient statistic for the family. Exercise 2.7. Suppose that Y 1,Y 2,...,Y n is a random sample from a Poisson(λ) distribution. Show that T(Y) = n Y i is a complete sufficient statistic forλ. The following Theorem establishes the minimum variance property of complete sufficient statistics.
9 2.2. POINT ESTIMATION 95 Theorem 2.7. Lehmann-Scheffé Theorem. Let Y = (Y 1,Y 2,...,Y n ) T be a random sample. If S(Y ) is a jointly complete sufficient statistic and T(Y) is an unbiased estimator for φ = g(ϑ) then U = E[T S] is, with probability 1, a unique MVUE of φ. Proof. First, to prove that U is a MVUE of g(ϑ), we show that whatever unbiased estimatort(y)we take we obtain the samee[t S], i.e., the sameu. Then, by Rao-Blackwell Theorem, condition (b), U must be MVUE ofg(ϑ). Suppose thatt(y)and T (Y)are any two unbiased estimators ofg(ϑ). Let U = E[T S] U = E[T S]. Then we have E{U U } = E{E[T S] E[T S]} = E(T) E(T ) = g(ϑ) g(ϑ) = 0. Hence, by completeness ofs(y ) we get P [ U ( S(Y ) ) = U ( S(Y ) )] = 1 for allϑ. This proves the first part of the theorem. Now, we will show uniqueness. Suppose that U and T are two MVUE of g(ϑ). Then if T is a function of the sufficient statistics S(Y ) then, as shown above, it must be equal to U. If T is not a function of S(Y ) then var(u) < var(t ), hence T cannot be a MVUE. Hence, U is a unique MVUE ofg(ϑ). Note: Lehmann-Scheffé Theorem may be used to construct MVUE of g(ϑ) by two methods. Both are based on complete sufficient statisticss(y ). Method 1: If we can find a function of S = S(Y ), say U(S) such that E[U(S)] = g(ϑ) thenu(s) is a unique MVUE ofg(ϑ).
10 96 CHAPTER 2. ELEMENTS OF STATISTICAL INFERENCE Method 2: If we can find any unbiased estimator T = T(Y) of g(ϑ), then U(S) = E[T S] is a unique MVUE ofg(ϑ). Example Method 1. Let Y i iid Bernoulli(p), i = 1,...,n. Earlier we showed that n Y i is a complete sufficient statistic forp. Denote it bys(y ). Y = 1 n n Y i = 1 S(Y ) is an unbiased estimator ofp, hence, as a function of a n complete sufficient statistic, it is the unique MVUE ofp. Now, let g(p) = var(y) = p(1 p). The sample variance 1 n 1 (Y i Y) 2 is an unbiased estimator of g(p). It is in fact a function of the complete sufficient statistics(y) = n Y i. Hence, it is the unique MVUE ofg(p) = p(1 p). Exercise 2.8. Suppose thaty = (Y 1,...,Y n ) T is a random sample from apoisson(λ) distribution. Find a MVUE ofφ = λ 2.
2.3 Methods of Estimation
96 CHAPTER 2. ELEMENTS OF STATISTICAL INFERENCE 2.3 Methods of Estimation 2.3. Method of Moments The Method of Moments is a simple technique based on the idea that the sample moments are natural estimators
More informationMathematical statistics
October 4 th, 2018 Lecture 12: Information Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation Chapter
More informationProof In the CR proof. and
Question Under what conditions will we be able to attain the Cramér-Rao bound and find a MVUE? Lecture 4 - Consequences of the Cramér-Rao Lower Bound. Searching for a MVUE. Rao-Blackwell Theorem, Lehmann-Scheffé
More information2.5 Hypothesis Testing
118 CHAPTER 2. ELEMENTS OF STATISTICAL INFERENCE 2.5 Hypothesis Testing We assume that Y 1,...,Y n have a joint distribution which depends on the unknown parametersϑ = ϑ 1,...,ϑ p ) T. The set of all possible
More informationST5215: Advanced Statistical Theory
Department of Statistics & Applied Probability Wednesday, October 19, 2011 Lecture 17: UMVUE and the first method of derivation Estimable parameters Let ϑ be a parameter in the family P. If there exists
More informationChapter 3: Unbiased Estimation Lecture 22: UMVUE and the method of using a sufficient and complete statistic
Chapter 3: Unbiased Estimation Lecture 22: UMVUE and the method of using a sufficient and complete statistic Unbiased estimation Unbiased or asymptotically unbiased estimation plays an important role in
More informationMathematical Statistics
Mathematical Statistics Chapter Three. Point Estimation 3.4 Uniformly Minimum Variance Unbiased Estimator(UMVUE) Criteria for Best Estimators MSE Criterion Let F = {p(x; θ) : θ Θ} be a parametric distribution
More informationExample: An experiment can either result in success or failure with probability θ and (1 θ) respectively. The experiment is performed independently
Chapter 3 Sufficient statistics and variance reduction Let X 1,X 2,...,X n be a random sample from a certain distribution with p.m/d.f fx θ. A function T X 1,X 2,...,X n = T X of these observations is
More information2.3.3 Method of Least Squares
.3. METHODS OF ESTIMATION 105.3.3 Method of Least Squares IfY 1,...,Y n are independent random variables, which have the same variance and higher-order moments, and, if each E(Y i ) is a linear function
More informationUnbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.
Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it
More informationChapters 9. Properties of Point Estimators
Chapters 9. Properties of Point Estimators Recap Target parameter, or population parameter θ. Population distribution f(x; θ). { probability function, discrete case f(x; θ) = density, continuous case The
More information1. (Regular) Exponential Family
1. (Regular) Exponential Family The density function of a regular exponential family is: [ ] Example. Poisson(θ) [ ] Example. Normal. (both unknown). ) [ ] [ ] [ ] [ ] 2. Theorem (Exponential family &
More information[Chapter 6. Functions of Random Variables]
[Chapter 6. Functions of Random Variables] 6.1 Introduction 6.2 Finding the probability distribution of a function of random variables 6.3 The method of distribution functions 6.5 The method of Moment-generating
More informationLecture 4: UMVUE and unbiased estimators of 0
Lecture 4: UMVUE and unbiased estimators of Problems of approach 1. Approach 1 has the following two main shortcomings. Even if Theorem 7.3.9 or its extension is applicable, there is no guarantee that
More informationUnbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.
Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it
More informationFebruary 26, 2017 COMPLETENESS AND THE LEHMANN-SCHEFFE THEOREM
February 26, 2017 COMPLETENESS AND THE LEHMANN-SCHEFFE THEOREM Abstract. The Rao-Blacwell theorem told us how to improve an estimator. We will discuss conditions on when the Rao-Blacwellization of an estimator
More information2 Statistical Estimation: Basic Concepts
Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof. N. Shimkin 2 Statistical Estimation:
More informationMethods of evaluating estimators and best unbiased estimators Hamid R. Rabiee
Stochastic Processes Methods of evaluating estimators and best unbiased estimators Hamid R. Rabiee 1 Outline Methods of Mean Squared Error Bias and Unbiasedness Best Unbiased Estimators CR-Bound for variance
More informationSTAT 730 Chapter 4: Estimation
STAT 730 Chapter 4: Estimation Timothy Hanson Department of Statistics, University of South Carolina Stat 730: Multivariate Analysis 1 / 23 The likelihood We have iid data, at least initially. Each datum
More informationp y (1 p) 1 y, y = 0, 1 p Y (y p) = 0, otherwise.
1. Suppose Y 1, Y 2,..., Y n is an iid sample from a Bernoulli(p) population distribution, where 0 < p < 1 is unknown. The population pmf is p y (1 p) 1 y, y = 0, 1 p Y (y p) = (a) Prove that Y is the
More informationChapter 6 Expectation and Conditional Expectation. Lectures Definition 6.1. Two random variables defined on a probability space are said to be
Chapter 6 Expectation and Conditional Expectation Lectures 24-30 In this chapter, we introduce expected value or the mean of a random variable. First we define expectation for discrete random variables
More informationENEE 621 SPRING 2016 DETECTION AND ESTIMATION THEORY THE PARAMETER ESTIMATION PROBLEM
c 2007-2016 by Armand M. Makowski 1 ENEE 621 SPRING 2016 DETECTION AND ESTIMATION THEORY THE PARAMETER ESTIMATION PROBLEM 1 The basic setting Throughout, p, q and k are positive integers. The setup With
More informationMiscellaneous Errors in the Chapter 6 Solutions
Miscellaneous Errors in the Chapter 6 Solutions 3.30(b In this problem, early printings of the second edition use the beta(a, b distribution, but later versions use the Poisson(λ distribution. If your
More informationSTATISTICAL METHODS FOR SIGNAL PROCESSING c Alfred Hero
STATISTICAL METHODS FOR SIGNAL PROCESSING c Alfred Hero 1999 32 Statistic used Meaning in plain english Reduction ratio T (X) [X 1,..., X n ] T, entire data sample RR 1 T (X) [X (1),..., X (n) ] T, rank
More informationλ n e λ n f(x;λ) f(y;λ) = λn e λ = e λ ( n = 2 θ θ 1 2 y2 1 3 y3 ] θ = 1 3 θ. E(Y i ) = 3 1 n n1 3 θ = θ. θ 1 3 y3 1 4 y4 ] θ = 1 6 θ2.
CHAPTER Exercise 1 Suppose that Y (Y 1,,Y ) is a radom sample from a Exp(λ) distributio The we may write f Y (y) λe λy i λ e λ y i }{{} }{{} 1 g λ (T(Y )) h(y) It follows thatt(y ) Y i is a sufficiet statistic
More informationf(y θ) = g(t (y) θ)h(y)
EXAM3, FINAL REVIEW (and a review for some of the QUAL problems): No notes will be allowed, but you may bring a calculator. Memorize the pmf or pdf f, E(Y ) and V(Y ) for the following RVs: 1) beta(δ,
More informationEstimation theory. Parametric estimation. Properties of estimators. Minimum variance estimator. Cramer-Rao bound. Maximum likelihood estimators
Estimation theory Parametric estimation Properties of estimators Minimum variance estimator Cramer-Rao bound Maximum likelihood estimators Confidence intervals Bayesian estimation 1 Random Variables Let
More informationA Few Notes on Fisher Information (WIP)
A Few Notes on Fisher Information (WIP) David Meyer dmm@{-4-5.net,uoregon.edu} Last update: April 30, 208 Definitions There are so many interesting things about Fisher Information and its theoretical properties
More informationSTATISTICS/ECONOMETRICS PREP COURSE PROF. MASSIMO GUIDOLIN
Massimo Guidolin Massimo.Guidolin@unibocconi.it Dept. of Finance STATISTICS/ECONOMETRICS PREP COURSE PROF. MASSIMO GUIDOLIN SECOND PART, LECTURE 2: MODES OF CONVERGENCE AND POINT ESTIMATION Lecture 2:
More informationReview Quiz. 1. Prove that in a one-dimensional canonical exponential family, the complete and sufficient statistic achieves the
Review Quiz 1. Prove that in a one-dimensional canonical exponential family, the complete and sufficient statistic achieves the Cramér Rao lower bound (CRLB). That is, if where { } and are scalars, then
More informationExercises and Answers to Chapter 1
Exercises and Answers to Chapter The continuous type of random variable X has the following density function: a x, if < x < a, f (x), otherwise. Answer the following questions. () Find a. () Obtain mean
More informationParameter Estimation
Parameter Estimation Consider a sample of observations on a random variable Y. his generates random variables: (y 1, y 2,, y ). A random sample is a sample (y 1, y 2,, y ) where the random variables y
More informationMATH4427 Notebook 2 Fall Semester 2017/2018
MATH4427 Notebook 2 Fall Semester 2017/2018 prepared by Professor Jenny Baglivo c Copyright 2009-2018 by Jenny A. Baglivo. All Rights Reserved. 2 MATH4427 Notebook 2 3 2.1 Definitions and Examples...................................
More informationHT Introduction. P(X i = x i ) = e λ λ x i
MODS STATISTICS Introduction. HT 2012 Simon Myers, Department of Statistics (and The Wellcome Trust Centre for Human Genetics) myers@stats.ox.ac.uk We will be concerned with the mathematical framework
More informationProbability Models. 4. What is the definition of the expectation of a discrete random variable?
1 Probability Models The list of questions below is provided in order to help you to prepare for the test and exam. It reflects only the theoretical part of the course. You should expect the questions
More informationi=1 k i=1 g i (Y )] = k
Math 483 EXAM 2 covers 2.4, 2.5, 2.7, 2.8, 3.1, 3.2, 3.3, 3.4, 3.8, 3.9, 4.1, 4.2, 4.3, 4.4, 4.5, 4.6, 4.9, 5.1, 5.2, and 5.3. The exam is on Thursday, Oct. 13. You are allowed THREE SHEETS OF NOTES and
More informationLecture 1: August 28
36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 1: August 28 Our broad goal for the first few lectures is to try to understand the behaviour of sums of independent random
More informationECE531 Lecture 8: Non-Random Parameter Estimation
ECE531 Lecture 8: Non-Random Parameter Estimation D. Richard Brown III Worcester Polytechnic Institute 19-March-2009 Worcester Polytechnic Institute D. Richard Brown III 19-March-2009 1 / 25 Introduction
More informationLecture 13 and 14: Bayesian estimation theory
1 Lecture 13 and 14: Bayesian estimation theory Spring 2012 - EE 194 Networked estimation and control (Prof. Khan) March 26 2012 I. BAYESIAN ESTIMATORS Mother Nature conducts a random experiment that generates
More informationMathematical Statistics 1 Math A 6330
Mathematical Statistics 1 Math A 6330 Chapter 3 Common Families of Distributions Mohamed I. Riffi Department of Mathematics Islamic University of Gaza September 28, 2015 Outline 1 Subjects of Lecture 04
More informationChapter 5 Class Notes
Chapter 5 Class Notes Sections 5.1 and 5.2 It is quite common to measure several variables (some of which may be correlated) and to examine the corresponding joint probability distribution One example
More informationFinal Exam. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given.
1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given. (a) If X and Y are independent, Corr(X, Y ) = 0. (b) (c) (d) (e) A consistent estimator must be asymptotically
More informationCompleteness. On the other hand, the distribution of an ancillary statistic doesn t depend on θ at all.
Completeness A minimal sufficient statistic achieves the maximum amount of data reduction while retaining all the information the sample has concerning θ. On the other hand, the distribution of an ancillary
More informationMathematical statistics
October 18 th, 2018 Lecture 16: Midterm review Countdown to mid-term exam: 7 days Week 1 Chapter 1: Probability review Week 2 Week 4 Week 7 Chapter 6: Statistics Chapter 7: Point Estimation Chapter 8:
More informationClassical Estimation Topics
Classical Estimation Topics Namrata Vaswani, Iowa State University February 25, 2014 This note fills in the gaps in the notes already provided (l0.pdf, l1.pdf, l2.pdf, l3.pdf, LeastSquares.pdf). 1 Min
More informationFall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.
1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n
More informationECE 275B Homework # 1 Solutions Version Winter 2015
ECE 275B Homework # 1 Solutions Version Winter 2015 1. (a) Because x i are assumed to be independent realizations of a continuous random variable, it is almost surely (a.s.) 1 the case that x 1 < x 2
More informationVariations. ECE 6540, Lecture 10 Maximum Likelihood Estimation
Variations ECE 6540, Lecture 10 Last Time BLUE (Best Linear Unbiased Estimator) Formulation Advantages Disadvantages 2 The BLUE A simplification Assume the estimator is a linear system For a single parameter
More informationMaximum Likelihood Estimation
Connexions module: m11446 1 Maximum Likelihood Estimation Clayton Scott Robert Nowak This work is produced by The Connexions Project and licensed under the Creative Commons Attribution License Abstract
More informationSTAT 512 sp 2018 Summary Sheet
STAT 5 sp 08 Summary Sheet Karl B. Gregory Spring 08. Transformations of a random variable Let X be a rv with support X and let g be a function mapping X to Y with inverse mapping g (A = {x X : g(x A}
More informationMathematical statistics
October 1 st, 2018 Lecture 11: Sufficient statistic Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation
More informationMidterm Examination. STA 215: Statistical Inference. Due Wednesday, 2006 Mar 8, 1:15 pm
Midterm Examination STA 215: Statistical Inference Due Wednesday, 2006 Mar 8, 1:15 pm This is an open-book take-home examination. You may work on it during any consecutive 24-hour period you like; please
More informationMarch 10, 2017 THE EXPONENTIAL CLASS OF DISTRIBUTIONS
March 10, 2017 THE EXPONENTIAL CLASS OF DISTRIBUTIONS Abstract. We will introduce a class of distributions that will contain many of the discrete and continuous we are familiar with. This class will help
More informationSTA 260: Statistics and Probability II
Al Nosedal. University of Toronto. Winter 2017 1 Properties of Point Estimators and Methods of Estimation 2 3 If you can t explain it simply, you don t understand it well enough Albert Einstein. Definition
More informationDetection & Estimation Lecture 1
Detection & Estimation Lecture 1 Intro, MVUE, CRLB Xiliang Luo General Course Information Textbooks & References Fundamentals of Statistical Signal Processing: Estimation Theory/Detection Theory, Steven
More informationECE 275B Homework # 1 Solutions Winter 2018
ECE 275B Homework # 1 Solutions Winter 2018 1. (a) Because x i are assumed to be independent realizations of a continuous random variable, it is almost surely (a.s.) 1 the case that x 1 < x 2 < < x n Thus,
More informationGraduate Econometrics I: Unbiased Estimation
Graduate Econometrics I: Unbiased Estimation Yves Dominicy Université libre de Bruxelles Solvay Brussels School of Economics and Management ECARES Yves Dominicy Graduate Econometrics I: Unbiased Estimation
More informationContents 1. Contents
Contents 1 Contents 6 Distributions of Functions of Random Variables 2 6.1 Transformation of Discrete r.v.s............. 3 6.2 Method of Distribution Functions............. 6 6.3 Method of Transformations................
More information1 Probability Model. 1.1 Types of models to be discussed in the course
Sufficiency January 11, 2016 Debdeep Pati 1 Probability Model Model: A family of distributions {P θ : θ Θ}. P θ (B) is the probability of the event B when the parameter takes the value θ. P θ is described
More informationStatistical Inference
Statistical Inference Classical and Bayesian Methods Revision Class for Midterm Exam AMS-UCSC Th Feb 9, 2012 Winter 2012. Session 1 (Revision Class) AMS-132/206 Th Feb 9, 2012 1 / 23 Topics Topics We will
More informationLecture 11. Multivariate Normal theory
10. Lecture 11. Multivariate Normal theory Lecture 11. Multivariate Normal theory 1 (1 1) 11. Multivariate Normal theory 11.1. Properties of means and covariances of vectors Properties of means and covariances
More informationLast Lecture. Biostatistics Statistical Inference Lecture 14 Obtaining Best Unbiased Estimator. Related Theorems. Rao-Blackwell Theorem
Last Lecture Biostatistics 62 - Statistical Inference Lecture 14 Obtaining Best Unbiased Estimator Hyun Min Kang February 28th, 213 For single-parameter exponential family, is Cramer-Rao bound always attainable?
More informationMore than one variable
Chapter More than one variable.1 Bivariate discrete distributions Suppose that the r.v. s X and Y are discrete and take on the values x j and y j, j 1, respectively. Then the joint p.d.f. of X and Y, to
More informationDefine characteristic function. State its properties. State and prove inversion theorem.
ASSIGNMENT - 1, MAY 013. Paper I PROBABILITY AND DISTRIBUTION THEORY (DMSTT 01) 1. (a) Give the Kolmogorov definition of probability. State and prove Borel cantelli lemma. Define : (i) distribution function
More informationDetection & Estimation Lecture 1
Detection & Estimation Lecture 1 Intro, MVUE, CRLB Xiliang Luo General Course Information Textbooks & References Fundamentals of Statistical Signal Processing: Estimation Theory/Detection Theory, Steven
More informationEstimation MLE-Pandemic data MLE-Financial crisis data Evaluating estimators. Estimation. September 24, STAT 151 Class 6 Slide 1
Estimation September 24, 2018 STAT 151 Class 6 Slide 1 Pandemic data Treatment outcome, X, from n = 100 patients in a pandemic: 1 = recovered and 0 = not recovered 1 1 1 0 0 0 1 1 1 0 0 1 0 1 0 0 1 1 1
More informationA General Overview of Parametric Estimation and Inference Techniques.
A General Overview of Parametric Estimation and Inference Techniques. Moulinath Banerjee University of Michigan September 11, 2012 The object of statistical inference is to glean information about an underlying
More information1 Complete Statistics
Complete Statistics February 4, 2016 Debdeep Pati 1 Complete Statistics Suppose X P θ, θ Θ. Let (X (1),..., X (n) ) denote the order statistics. Definition 1. A statistic T = T (X) is complete if E θ g(t
More informationChapter 8: Least squares (beginning of chapter)
Chapter 8: Least squares (beginning of chapter) Least Squares So far, we have been trying to determine an estimator which was unbiased and had minimum variance. Next we ll consider a class of estimators
More informationECE531 Lecture 10b: Maximum Likelihood Estimation
ECE531 Lecture 10b: Maximum Likelihood Estimation D. Richard Brown III Worcester Polytechnic Institute 05-Apr-2011 Worcester Polytechnic Institute D. Richard Brown III 05-Apr-2011 1 / 23 Introduction So
More informationMcGill University. Faculty of Science. Department of Mathematics and Statistics. Part A Examination. Statistics: Theory Paper
McGill University Faculty of Science Department of Mathematics and Statistics Part A Examination Statistics: Theory Paper Date: 10th May 2015 Instructions Time: 1pm-5pm Answer only two questions from Section
More informationi=1 k i=1 g i (Y )] = k f(t)dt and f(y) = F (y) except at possibly countably many points, E[g(Y )] = f(y)dy = 1, F(y) = y
Math 480 Exam 2 is Wed. Oct. 31. You are allowed 7 sheets of notes and a calculator. The exam emphasizes HW5-8, and Q5-8. From the 1st exam: The conditional probability of A given B is P(A B) = P(A B)
More informationJoint p.d.f. and Independent Random Variables
1 Joint p.d.f. and Independent Random Variables Let X and Y be two discrete r.v. s and let R be the corresponding space of X and Y. The joint p.d.f. of X = x and Y = y, denoted by f(x, y) = P(X = x, Y
More informationReview. December 4 th, Review
December 4 th, 2017 Att. Final exam: Course evaluation Friday, 12/14/2018, 10:30am 12:30pm Gore Hall 115 Overview Week 2 Week 4 Week 7 Week 10 Week 12 Chapter 6: Statistics and Sampling Distributions Chapter
More information1 Probability Model. 1.1 Types of models to be discussed in the course
Sufficiency January 18, 016 Debdeep Pati 1 Probability Model Model: A family of distributions P θ : θ Θ}. P θ (B) is the probability of the event B when the parameter takes the value θ. P θ is described
More informationSolution-Midterm Examination - I STAT 421, Spring 2012
Solution-Midterm Examination - I STAT 4, Spring 0 Prof. Prem K. Goel. [0 points] Let X, X,..., X n be a random sample of size n from a Gamma population with β = and α = ν/, where ν is the unknown parameter.
More informationMath 3215 Intro. Probability & Statistics Summer 14. Homework 5: Due 7/3/14
Math 325 Intro. Probability & Statistics Summer Homework 5: Due 7/3/. Let X and Y be continuous random variables with joint/marginal p.d.f. s f(x, y) 2, x y, f (x) 2( x), x, f 2 (y) 2y, y. Find the conditional
More informationSTATS 200: Introduction to Statistical Inference. Lecture 29: Course review
STATS 200: Introduction to Statistical Inference Lecture 29: Course review Course review We started in Lecture 1 with a fundamental assumption: Data is a realization of a random process. The goal throughout
More informationEvaluating the Performance of Estimators (Section 7.3)
Evaluating the Performance of Estimators (Section 7.3) Example: Suppose we observe X 1,..., X n iid N(θ, σ 2 0 ), with σ2 0 known, and wish to estimate θ. Two possible estimators are: ˆθ = X sample mean
More informationt x 1 e t dt, and simplify the answer when possible (for example, when r is a positive even number). In particular, confirm that EX 4 = 3.
Mathematical Statistics: Homewor problems General guideline. While woring outside the classroom, use any help you want, including people, computer algebra systems, Internet, and solution manuals, but mae
More informationProbability and Distributions
Probability and Distributions What is a statistical model? A statistical model is a set of assumptions by which the hypothetical population distribution of data is inferred. It is typically postulated
More informationTheory of Maximum Likelihood Estimation. Konstantin Kashin
Gov 2001 Section 5: Theory of Maximum Likelihood Estimation Konstantin Kashin February 28, 2013 Outline Introduction Likelihood Examples of MLE Variance of MLE Asymptotic Properties What is Statistical
More informationChapter 8.8.1: A factorization theorem
LECTURE 14 Chapter 8.8.1: A factorization theorem The characterization of a sufficient statistic in terms of the conditional distribution of the data given the statistic can be difficult to work with.
More informationMathematics Qualifying Examination January 2015 STAT Mathematical Statistics
Mathematics Qualifying Examination January 2015 STAT 52800 - Mathematical Statistics NOTE: Answer all questions completely and justify your derivations and steps. A calculator and statistical tables (normal,
More informationAdvanced Signal Processing Introduction to Estimation Theory
Advanced Signal Processing Introduction to Estimation Theory Danilo Mandic, room 813, ext: 46271 Department of Electrical and Electronic Engineering Imperial College London, UK d.mandic@imperial.ac.uk,
More informationECE534, Spring 2018: Solutions for Problem Set #3
ECE534, Spring 08: Solutions for Problem Set #3 Jointly Gaussian Random Variables and MMSE Estimation Suppose that X, Y are jointly Gaussian random variables with µ X = µ Y = 0 and σ X = σ Y = Let their
More informationLimiting Distributions
Limiting Distributions We introduce the mode of convergence for a sequence of random variables, and discuss the convergence in probability and in distribution. The concept of convergence leads us to the
More informationMathematics Ph.D. Qualifying Examination Stat Probability, January 2018
Mathematics Ph.D. Qualifying Examination Stat 52800 Probability, January 2018 NOTE: Answers all questions completely. Justify every step. Time allowed: 3 hours. 1. Let X 1,..., X n be a random sample from
More informationChapter 5: Joint Probability Distributions
Chapter 5: Joint Probability Distributions Seungchul Baek Department of Statistics, University of South Carolina STAT 509: Statistics for Engineers 1 / 19 Joint pmf Definition: The joint probability mass
More information1 of 7 7/16/2009 6:12 AM Virtual Laboratories > 7. Point Estimation > 1 2 3 4 5 6 1. Estimators The Basic Statistical Model As usual, our starting point is a random experiment with an underlying sample
More informationProbability Theory and Statistics. Peter Jochumzen
Probability Theory and Statistics Peter Jochumzen April 18, 2016 Contents 1 Probability Theory And Statistics 3 1.1 Experiment, Outcome and Event................................ 3 1.2 Probability............................................
More informationElements of statistics (MATH0487-1)
Elements of statistics (MATH0487-1) Prof. Dr. Dr. K. Van Steen University of Liège, Belgium November 12, 2012 Introduction to Statistics Basic Probability Revisited Sampling Exploratory Data Analysis -
More informationThe Delta Method and Applications
Chapter 5 The Delta Method and Applications 5.1 Local linear approximations Suppose that a particular random sequence converges in distribution to a particular constant. The idea of using a first-order
More informationLimiting Distributions
We introduce the mode of convergence for a sequence of random variables, and discuss the convergence in probability and in distribution. The concept of convergence leads us to the two fundamental results
More informationEIE6207: Estimation Theory
EIE6207: Estimation Theory Man-Wai MAK Dept. of Electronic and Information Engineering, The Hong Kong Polytechnic University enmwmak@polyu.edu.hk http://www.eie.polyu.edu.hk/ mwmak References: Steven M.
More information557: MATHEMATICAL STATISTICS II BIAS AND VARIANCE
557: MATHEMATICAL STATISTICS II BIAS AND VARIANCE An estimator, T (X), of θ can be evaluated via its statistical properties. Typically, two aspects are considered: Expectation Variance either in terms
More informationDistributions of Functions of Random Variables. 5.1 Functions of One Random Variable
Distributions of Functions of Random Variables 5.1 Functions of One Random Variable 5.2 Transformations of Two Random Variables 5.3 Several Random Variables 5.4 The Moment-Generating Function Technique
More informationELEG 5633 Detection and Estimation Minimum Variance Unbiased Estimators (MVUE)
1 ELEG 5633 Detection and Estimation Minimum Variance Unbiased Estimators (MVUE) Jingxian Wu Department of Electrical Engineering University of Arkansas Outline Minimum Variance Unbiased Estimators (MVUE)
More informationMA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems
MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems Review of Basic Probability The fundamentals, random variables, probability distributions Probability mass/density functions
More informationSTAT 430/510 Probability
STAT 430/510 Probability Hui Nie Lecture 16 June 24th, 2009 Review Sum of Independent Normal Random Variables Sum of Independent Poisson Random Variables Sum of Independent Binomial Random Variables Conditional
More information