ECE531 Lecture 8: Non-Random Parameter Estimation

Size: px
Start display at page:

Download "ECE531 Lecture 8: Non-Random Parameter Estimation"

Transcription

1 ECE531 Lecture 8: Non-Random Parameter Estimation D. Richard Brown III Worcester Polytechnic Institute 19-March-2009 Worcester Polytechnic Institute D. Richard Brown III 19-March / 25

2 Introduction Recall the two basic classes of estimation problems: Known prior π(θ): Bayesian estimation. Unknown prior: Non-random parameter estimation. In non-random parameter estimation problems, we can still compute the risk of estimator ˆθ(y) when the true parameter is θ: ] R θ (ˆθ) := E θ [C θ (ˆθ(Y )) = C θ (ˆθ(y))p Y (y ; θ)dy Y where E θ means the expectation parameterized by θ and C θ (ˆθ) : Λ Λ R is the cost of the parameter estimate ˆθ Λ given the true parameter θ Λ. We cannot, however, compute any sort of average risk r(ˆθ) = E[R Θ (ˆθ)] since we have no distribution on the random parameter Θ. Worcester Polytechnic Institute D. Richard Brown III 19-March / 25

3 Uniformly Most Powerful Estimators? We would like to find a uniformly most powerful estimator ˆθ(y) that minimizes the conditional risk R θ (ˆθ) for all θ Λ. Example: Consider two distinct points in Λ: θ 0 θ 1. Suppose we have two estimators ˆθa (y) θ 0 for all y Y and ˆθ b (y) θ 1 for all y Y. Note that both of these estimator ignore the observation and always give the same estimate. For any of the cost functions we have considered, we know that C θ0 (ˆθ a (y)) = 0 C θ1 (ˆθ b (y)) = 0. For the MMSE or MMAE estimators, we also know that C θ1 (ˆθ a (y)) > 0 C θ0 (ˆθ b (y)) > 0. It should be clear that a uniformly most powerful estimator is not going to exist in most cases of interest. Worcester Polytechnic Institute D. Richard Brown III 19-March / 25

4 Some Options 1. We could restrict our attention to finding the sort of problems that do admit a uniformly most powerful estimator. 2. We could try find locally most powerful estimators. 3. We could assume a prior π(θ), e.g. perhaps some sort of least favorable prior, and solve the problem in the Bayes framework. 4. We could keep the problem non-random but place restrictions on the class of estimators that we are willing to consider. Worcester Polytechnic Institute D. Richard Brown III 19-March / 25

5 Option 4: Consider Only Unbiased Estimators A reasonable restriction on the class of estimators that we are willing to consider is the class of unbiased estimators. Definition An estimator ˆθ(y) is unbiased if for all θ Λ. Remarks: E θ [ˆθ(Y ) ] = θ This class excludes trivial estimators like ˆθ(y) θ 0. Under the squared-error cost assignment, the parameterized risk of estimators in this class R θ (ˆθ) = E θ h θ ˆθ(Y i ) 2 2 = X i E θi h (ˆθ i(y ) θ i) 2i = X i i var θi hˆθi(y ) The goal: find an unbiased estimator with minimum variance. Worcester Polytechnic Institute D. Richard Brown III 19-March / 25

6 Minimum Variance Unbiased Estimators Definition A minimum-variance unbiased estimator ˆθ mvu (y) is an unbiased estimator satisfying ˆθ mvu (y) = arg min R θ (ˆθ(y)) ˆθ(y) Ω for all θ Λ where Ω is the set of all unbiased estimators. Remarks: Finding an MVU estimator is still a multi-objective optimization problem. MVU estimators do not always exist. The class of problems in which MVU estimators do exist, however, is much larger than that of uniformly most powerful estimators. Worcester Polytechnic Institute D. Richard Brown III 19-March / 25

7 Example: Estimating a Constant in White Gaussian Noise Suppose we have random observations given by Y k = θ + W k k = 0,...,n 1 where W k i.i.d. N(0,σ 2 ). The unknown parameter θ can take on any value on the real line and we have no prior pdf. Suppose we generate estimates with the sample mean: ˆθ(y) = 1 n n 1 y k k=0 Is this estimator unbiased? Yes (easy to check). Is this estimator MVU? The ] variance of the estimator can be calculated as var θ [ˆθ(Y ) = σ2 n. But answering the question as to whether this estimator is MVU or not will require more work. Worcester Polytechnic Institute D. Richard Brown III 19-March / 25

8 A Generalized Model for Estimation Problems parameter space Λ θ p Y (y ; θ) probabilistic model estimator observation space y Y g(θ) RBLS ĝ(y) g[t(y)] estimator T(y) sufficient statistics Worcester Polytechnic Institute D. Richard Brown III 19-March / 25

9 A Procedure for Finding MVU Estimators To find an MVU estimator for g(θ), we can follow a three-step procedure: 1. Find a complete sufficient statistic T for the family of pdfs {p Y (y ; θ); θ Λ} parameterized by θ. 2. Find any unbiased estimator ĝ(y) of g(θ). 3. Compute g[t(y)] = E θ [ĝ(y ) T(Y ) = T(y)]. The Rao-Blackwell-Lehmann-Sheffe Theorem says that g[t(y)] will be a MVU estimator of g(θ). Worcester Polytechnic Institute D. Richard Brown III 19-March / 25

10 Sufficiency and Minimal Sufficiency Definition T : Y is a sufficient statistic for the family of pdfs {p Y (y ; θ); θ Λ} if the distribution of the random observation conditioned on T(Y ), i.e. p Y (y T(Y ) = t ; θ), does not depend on θ for all θ Λ and all t. Intuitively, a sufficient statistic summarizes the information contained in the observation about the unknown parameter. Knowing T(y) is as good as knowing the full observation y when we wish to estimate θ or g(θ). Definition T : Y is said to be minimal sufficient for the family of pdfs {p Y (y ; θ); θ Λ} if it is a function of every other sufficient statistic for this family of pdfs. Intuitively, a minimal sufficient statistic is the most concise summary of the observation. These can be hard to find and don t always exist. Worcester Polytechnic Institute D. Richard Brown III 19-March / 25

11 Example: Sufficiency (part 1) Suppose θ R and we get a vector observation y R n distributed as p Y (y ; θ) = ( n X exp (2πσ 2 ) n/2 2σ 2 k=0 (y k θ) 2 ) = 1 exp (2πσ 2 ) n/2 j y 1θ 2 2 Is T(y) = y = [y 0,...,y n 1 ] sufficient? Intuitively, it should be. To gain some experience with the definition, however, let s check. What happens to p Y (y ; θ) when we condition on an observation Y = [y 0,...,y n 1 ]? 2σ 2 ff. Worcester Polytechnic Institute D. Richard Brown III 19-March / 25

12 Example: Sufficiency (part 2) Is T(y) = ȳ = 1 n n 1 k=0 y k sufficient? Note that Ȳ := 1 n 1 n k=0 Y k is a random variable distributed as N(θ,σ 2 /n). We can apply the definition directly... p Y (y ȳ ; θ) Bayes = pȳ (ȳ y ; θ)p Y (y ; θ) p Ȳ (ȳ ; θ) = δ ( ȳ 1 ) n yk py (y ; θ) p Ȳ (ȳ ; θ) ( = cδ ȳ 1 ) [ ( y 2 yk exp k n(ȳ) 2) ] n 2σ 2 where δ( ) is the Dirac delta function and c does not depend on θ. Hence T(y) = 1 n 1 n k=0 y k is a sufficient statistic. Worcester Polytechnic Institute D. Richard Brown III 19-March / 25

13 Neyman-Fisher Factorization Theorem Guessing and checking sufficient statistics isn t very satisfying. We need a procedure for finding sufficient statistics. Theorem (Fisher 1920, Neyman 1935) A statistic T is sufficient for θ if and only if there exist functions g θ and h such that the pdf of the observation can be factored as for all y Y and all θ Λ. p Y (y ; θ) = g θ (T(y))h(y) The Poor textbook gives the proof for the case when Y is discrete. A general proof can be found in Lehmann Worcester Polytechnic Institute D. Richard Brown III 19-March / 25

14 Example: Neyman-Fisher Factorization Theorem Suppose θ R and p Y (y ; θ) = { } n (2πσ 2 exp ) n/2 2σ 2 (y k θ) 2. k=0 Let T(y) = 1 n 1 n k=0 y k = ȳ. We already know this is a sufficient statistic but let s try the factorization. { ( )} n 1 1 n 1 p Y (y ; θ) = (2πσ 2 exp ) n/2 2σ 2 yk 2 n 2θy k + θ 2 k=0 { 1 n ( = (2πσ 2 exp ) n/2 θ 2 2σ 2 2θȳ ) } { } n 1 1 exp 2σ }{{} 2 yk 2 k=0 }{{} g θ (T(y)) h(y) Worcester Polytechnic Institute D. Richard Brown III 19-March / 25

15 Flashback: Sufficient Statistic For Simple Binary HT Suppose we have a simple binary hypothesis testing problem: { p 0 (y) θ = 0 Y p Y (y ; θ) = p 1 (y) θ = 1 Let T(y) = p 1(y) p 0 (y) g θ (T(y)) = θt(y) + (1 θ) h(y) = p 0 (y) Then it is easy to show that p Y (y ; θ) = g θ (T(y))h(y). Hence the likelihood ratio L(y) = p 1(y) p 0 (y) is a sufficient statistic for simple binary hypothesis testing problems. Worcester Polytechnic Institute D. Richard Brown III 19-March / 25

16 Completeness Definition The family of pdfs {p Y (y ; θ); θ Λ} is said to be complete if the condition E θ [f(y )] = 0 for all θ in Λ implies that Prob θ [f(y ) = 0] = 1 for all θ in Λ. Note that f : Y R can be any function. To get some intuition, consider the case where Y = {y 0,..., y L 1 } is a finite set. Then E θ [f(y )] = L 1 f(y l )Prob θ (Y = y l ) l=0 = f (y)p θ For a fixed θ it is certainly possible to find a non-zero f such that E θ [f(y )] = 0. But we have to satisfy this condition for all θ Λ, i.e. we need a vector f(y) that is orthogonal to the all members of the family of vectors {P θ ; θ Λ}. If the only such vector that satisfies the condition E θ [f(y )] = 0 for all θ Λ is f(y 0 ) = = f(y L 1 ) = 0, then the family {P θ ; θ Λ} is complete. Worcester Polytechnic Institute D. Richard Brown III 19-March / 25

17 Complete Sufficient Statistics Definition Suppose that T is a sufficient statistic for the family of pdfs {p Y (y ; θ); θ Λ}. Let p Z (z ; θ) denote the distribution of Z = T(Y ) when the parameter is θ. If the family of pdfs {p Z (z ; θ); θ Λ} is complete, then T is said to be a complete sufficient statistic for the family {p Y (y ; θ); θ Λ}. parameter space Λ θ p Y (y ; θ) probabilistic model observation space Y Y Z sufficient statistics Worcester Polytechnic Institute D. Richard Brown III 19-March / 25

18 Example: Complete Sufficient Statistic Suppose θ R and p Y (y ; θ) = { } n (2πσ 2 exp ) n/2 2σ 2 (y k θ) 2. k=0 Let T(y) = 1 n complete? n 1 k=0 y k. We know this is a sufficient statistic but is it We know the distribution of T(Y ) = Ȳ is N(θ,σ2 /n). We require E θ [f(ȳ )] = 0 for all θ Λ, i.e. { n n(x θ) 2} s(θ) = f(x) exp 2πσ 2σ 2 dx = 0 for all θ Λ Is there any non-zero f : R R that can do this? Suppose θ = 0. Lots of functions will do this, e.g. f(x) = x, f(x) = sin(x), etc. But we need s(θ) = 0 for all θ Λ. Worcester Polytechnic Institute D. Richard Brown III 19-March / 25

19 Example: Complete Sufficient Statistic (continued) s(θ) = s(θ) = { n n(x θ) 2} f(x) exp 2πσ 2σ 2 dx = 0 for all θ Λ f(x)exp { (θ x) 2} dx = 0 for all θ Λ But this is just the convolution of f(x) with a Gaussian pulse. Recall that convolution in the time domain is multiplication in the frequency domain. Hence, if S(ω) is the Fourier transform of s(θ), then S(ω) = F(ω)G(ω) = 0 for all ω where G(ω) is the Fourier transform of the Gaussian pulse. Note that G(ω) is itself Gaussian and therefore positive for all ω. Hence, the only way to force S(ω) 0 is to have F(ω) 0. Hence the only solution to E θ [f(ȳ )] = 0 for all θ Λ is f(x) 0 for all x and, consequently, T(y) is a complete sufficient statistic. Worcester Polytechnic Institute D. Richard Brown III 19-March / 25

20 Example: Incomplete Sufficient Statistic Suppose θ R and you would like to estimate θ from a scalar observation Y = θ + W where W U [ 1 2, 1 2]. An obvious sufficient statistic then is T(y) = y. But is it complete? Since T(Y ) = Y, we require E θ [f(y )] = 0 for all θ Λ, i.e. s(θ) = θ+ 1 2 θ 1 2 f(x)dx = 0 for all θ Λ Is there any non-zero f : R R that can do this? How about f(x) = sin(2πx)? This definitely forces s(θ) = 0 for all θ Λ. Just need to confirm Prob[f(Y ) = 0] < 1 for at least one θ Λ. Since we found a non-zero f(x) that forced E θ [f(y )] = 0 for all θ Λ, we can say that T(y) = y is not complete. Worcester Polytechnic Institute D. Richard Brown III 19-March / 25

21 Completeness Theorem for Exponential Families Theorem Suppose Y = R n, Λ R m, and { m } p Y (y ; θ) = a(θ)exp θ l T l (y) h(y) where a,t 1,...,T m, and h are all real-valued functions. Then T(y) = [T 1 (y),...,t m (y)] is a complete sufficient statistic for the family {p Y (y ; θ); θ Λ} if Λ contains an m-dimensional rectangle. Remarks: The technical detail about the m-dimensional rectangle ensures that Λ is not missing any dimensions in R m, e.g. Λ is not a two-dimensional plane in R 3. Main idea of proof is similar to how we showed completeness in the Gaussian example. See Poor pp and Lehmann Worcester Polytechnic Institute D. Richard Brown III 19-March / 25 l=1

22 Rao-Blackwell-Lehmann-Sheffe Theorem Theorem If ĝ(y) is any unbiased estimator of g(θ) and T is a sufficient statistic for the family {p Y (y ; θ); θ Λ}, then is g[t(y)] := E θ [ĝ(y ) T(Y ) = T(y)] A valid estimator of g(θ) (not a function of θ) An unbiased estimator of g(θ). Of lesser or equal variance than that of ĝ(y) for all θ Λ Additionally, if T is complete, then g[t(y)] is an MVU estimator of of g(θ). Worcester Polytechnic Institute D. Richard Brown III 19-March / 25

23 Example: Estimating a Constant in White Gaussian Noise Suppose θ R and p Y (y ; θ) = { } n (2πσ 2 exp ) n/2 2σ 2 (y k θ) 2. k=0 Let T(y) = 1 n 1 n k=0 y k. We know this is a complete sufficient statistic. Let s apply the RBLS theorem to find the MVU estimator... Suppose g(θ) = θ is just the identity mapping. We could choose the unbiased estimator ĝ(y) = y 0. Now we need to compute [ g[t(y)] := E θ [ĝ(y ) T(Y ) = T(y)] = E θ Y 0 1 Yk = 1 ] yk n n To solve this, we can use a standard formula for the conditional expectation of a jointly Gaussian random variable... Worcester Polytechnic Institute D. Richard Brown III 19-March / 25

24 Example (continued) Suppose Z = [X,Y ] is jointly Gaussian distributed. It can be shown that E[X Y = y] = E[X] + cov(x,y ) (y E[Y ]). var(y ) In our problem, letting Ȳ = 1 n 1 n k=0 Y k, we can use this result to write [ E θ Y0 Ȳ = t] = E θ [Y 0 ] + cov θ(y 0,Ȳ ) ( 1 ) yk E θ [Ȳ ] ( = θ + σ2 1 σ 2 n = 1 n yk var θ (Y 0 ) n ) yk θ Hence ˆθ mvu (y) = 1 n yk is an MVU estimator of θ. Worcester Polytechnic Institute D. Richard Brown III 19-March / 25

25 Conclusions To approach non-random parameter estimation problems, we had to restrict our attention to the class of unbiased estimators. Under the squared error cost assignment, the performance of these unbiased estimators is measured by the variance of the estimates. Rao-Blackwell-Lehmann-Sheffe theorem establishes a procedure for finding minimum variance unbiased (MVU) estimators. Subtle concepts will require some practice: Sufficiency (Neyman-Fisher factorization theorem) Completeness Following RBLS doesn t guarantee you will find an MVU estimator: It can be difficult/impossible to find a complete sufficient statistic. It is often difficult to check completeness. Computing the conditional expectation can be intractable. Other techniques may be useful for checking if an estimator is MVU. Further restrictions on the class of estimators also facilitate analysis. Worcester Polytechnic Institute D. Richard Brown III 19-March / 25

ECE531 Lecture 10b: Maximum Likelihood Estimation

ECE531 Lecture 10b: Maximum Likelihood Estimation ECE531 Lecture 10b: Maximum Likelihood Estimation D. Richard Brown III Worcester Polytechnic Institute 05-Apr-2011 Worcester Polytechnic Institute D. Richard Brown III 05-Apr-2011 1 / 23 Introduction So

More information

ECE531 Homework Assignment Number 7 Solution

ECE531 Homework Assignment Number 7 Solution ECE53 Homework Assignment Number 7 Solution Due by 8:50pm on Wednesday 30-Mar-20 Make sure your reasoning and work are clear to receive full credit for each problem.. 4 points. Kay I: 2.6. Solution: I

More information

ECE531 Lecture 4b: Composite Hypothesis Testing

ECE531 Lecture 4b: Composite Hypothesis Testing ECE531 Lecture 4b: Composite Hypothesis Testing D. Richard Brown III Worcester Polytechnic Institute 16-February-2011 Worcester Polytechnic Institute D. Richard Brown III 16-February-2011 1 / 44 Introduction

More information

Variations. ECE 6540, Lecture 10 Maximum Likelihood Estimation

Variations. ECE 6540, Lecture 10 Maximum Likelihood Estimation Variations ECE 6540, Lecture 10 Last Time BLUE (Best Linear Unbiased Estimator) Formulation Advantages Disadvantages 2 The BLUE A simplification Assume the estimator is a linear system For a single parameter

More information

ECE531 Lecture 6: Detection of Discrete-Time Signals with Random Parameters

ECE531 Lecture 6: Detection of Discrete-Time Signals with Random Parameters ECE531 Lecture 6: Detection of Discrete-Time Signals with Random Parameters D. Richard Brown III Worcester Polytechnic Institute 26-February-2009 Worcester Polytechnic Institute D. Richard Brown III 26-February-2009

More information

ECE531 Homework Assignment Number 6 Solution

ECE531 Homework Assignment Number 6 Solution ECE53 Homework Assignment Number 6 Solution Due by 8:5pm on Wednesday 3-Mar- Make sure your reasoning and work are clear to receive full credit for each problem.. 6 points. Suppose you have a scalar random

More information

Parameter Estimation

Parameter Estimation 1 / 44 Parameter Estimation Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay October 25, 2012 Motivation System Model used to Derive

More information

ECE531 Screencast 9.2: N-P Detection with an Infinite Number of Possible Observations

ECE531 Screencast 9.2: N-P Detection with an Infinite Number of Possible Observations ECE531 Screencast 9.2: N-P Detection with an Infinite Number of Possible Observations D. Richard Brown III Worcester Polytechnic Institute Worcester Polytechnic Institute D. Richard Brown III 1 / 7 Neyman

More information

Estimation theory. Parametric estimation. Properties of estimators. Minimum variance estimator. Cramer-Rao bound. Maximum likelihood estimators

Estimation theory. Parametric estimation. Properties of estimators. Minimum variance estimator. Cramer-Rao bound. Maximum likelihood estimators Estimation theory Parametric estimation Properties of estimators Minimum variance estimator Cramer-Rao bound Maximum likelihood estimators Confidence intervals Bayesian estimation 1 Random Variables Let

More information

ECE531 Lecture 2b: Bayesian Hypothesis Testing

ECE531 Lecture 2b: Bayesian Hypothesis Testing ECE531 Lecture 2b: Bayesian Hypothesis Testing D. Richard Brown III Worcester Polytechnic Institute 29-January-2009 Worcester Polytechnic Institute D. Richard Brown III 29-January-2009 1 / 39 Minimizing

More information

ECE531 Screencast 5.5: Bayesian Estimation for the Linear Gaussian Model

ECE531 Screencast 5.5: Bayesian Estimation for the Linear Gaussian Model ECE53 Screencast 5.5: Bayesian Estimation for the Linear Gaussian Model D. Richard Brown III Worcester Polytechnic Institute Worcester Polytechnic Institute D. Richard Brown III / 8 Bayesian Estimation

More information

ECE531: Principles of Detection and Estimation Course Introduction

ECE531: Principles of Detection and Estimation Course Introduction ECE531: Principles of Detection and Estimation Course Introduction D. Richard Brown III WPI 22-January-2009 WPI D. Richard Brown III 22-January-2009 1 / 37 Lecture 1 Major Topics 1. Web page. 2. Syllabus

More information

Chapter 8: Least squares (beginning of chapter)

Chapter 8: Least squares (beginning of chapter) Chapter 8: Least squares (beginning of chapter) Least Squares So far, we have been trying to determine an estimator which was unbiased and had minimum variance. Next we ll consider a class of estimators

More information

Lecture 2: Basic Concepts of Statistical Decision Theory

Lecture 2: Basic Concepts of Statistical Decision Theory EE378A Statistical Signal Processing Lecture 2-03/31/2016 Lecture 2: Basic Concepts of Statistical Decision Theory Lecturer: Jiantao Jiao, Tsachy Weissman Scribe: John Miller and Aran Nayebi In this lecture

More information

ECE531: Principles of Detection and Estimation Course Introduction

ECE531: Principles of Detection and Estimation Course Introduction ECE531: Principles of Detection and Estimation Course Introduction D. Richard Brown III WPI 15-January-2013 WPI D. Richard Brown III 15-January-2013 1 / 39 First Lecture: Major Topics 1. Administrative

More information

Economics 520. Lecture Note 19: Hypothesis Testing via the Neyman-Pearson Lemma CB 8.1,

Economics 520. Lecture Note 19: Hypothesis Testing via the Neyman-Pearson Lemma CB 8.1, Economics 520 Lecture Note 9: Hypothesis Testing via the Neyman-Pearson Lemma CB 8., 8.3.-8.3.3 Uniformly Most Powerful Tests and the Neyman-Pearson Lemma Let s return to the hypothesis testing problem

More information

ECE 275A Homework 6 Solutions

ECE 275A Homework 6 Solutions ECE 275A Homework 6 Solutions. The notation used in the solutions for the concentration (hyper) ellipsoid problems is defined in the lecture supplement on concentration ellipsoids. Note that θ T Σ θ =

More information

2 Statistical Estimation: Basic Concepts

2 Statistical Estimation: Basic Concepts Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof. N. Shimkin 2 Statistical Estimation:

More information

ECE531 Lecture 12: Linear Estimation and Causal Wiener-Kolmogorov Filtering

ECE531 Lecture 12: Linear Estimation and Causal Wiener-Kolmogorov Filtering ECE531 Lecture 12: Linear Estimation and Causal Wiener-Kolmogorov Filtering D. Richard Brown III Worcester Polytechnic Institute 16-Apr-2009 Worcester Polytechnic Institute D. Richard Brown III 16-Apr-2009

More information

10. Linear Models and Maximum Likelihood Estimation

10. Linear Models and Maximum Likelihood Estimation 10. Linear Models and Maximum Likelihood Estimation ECE 830, Spring 2017 Rebecca Willett 1 / 34 Primary Goal General problem statement: We observe y i iid pθ, θ Θ and the goal is to determine the θ that

More information

Perhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows.

Perhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows. Chapter 5 Two Random Variables In a practical engineering problem, there is almost always causal relationship between different events. Some relationships are determined by physical laws, e.g., voltage

More information

ECE531 Lecture 13: Sequential Detection of Discrete-Time Signals

ECE531 Lecture 13: Sequential Detection of Discrete-Time Signals ECE531 Lecture 13: Sequential Detection of Discrete-Time Signals D. Richard Brown III Worcester Polytechnic Institute 30-Apr-2009 Worcester Polytechnic Institute D. Richard Brown III 30-Apr-2009 1 / 32

More information

Space Telescope Science Institute statistics mini-course. October Inference I: Estimation, Confidence Intervals, and Tests of Hypotheses

Space Telescope Science Institute statistics mini-course. October Inference I: Estimation, Confidence Intervals, and Tests of Hypotheses Space Telescope Science Institute statistics mini-course October 2011 Inference I: Estimation, Confidence Intervals, and Tests of Hypotheses James L Rosenberger Acknowledgements: Donald Richards, William

More information

ECE531 Screencast 11.5: Uniformly Most Powerful Decision Rules

ECE531 Screencast 11.5: Uniformly Most Powerful Decision Rules ECE531 Screencast 11.5: Uniformly Most Powerful Decision Rules D. Richard Brown III Worcester Polytechnic Institute Worcester Polytechnic Institute D. Richard Brown III 1 / 9 Monotone Likelihood Ratio

More information

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed

More information

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed

More information

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it

More information

ENGG2430A-Homework 2

ENGG2430A-Homework 2 ENGG3A-Homework Due on Feb 9th,. Independence vs correlation a For each of the following cases, compute the marginal pmfs from the joint pmfs. Explain whether the random variables X and Y are independent,

More information

Mathematical Statistics

Mathematical Statistics Mathematical Statistics Chapter Three. Point Estimation 3.4 Uniformly Minimum Variance Unbiased Estimator(UMVUE) Criteria for Best Estimators MSE Criterion Let F = {p(x; θ) : θ Θ} be a parametric distribution

More information

Chapter 2. Binary and M-ary Hypothesis Testing 2.1 Introduction (Levy 2.1)

Chapter 2. Binary and M-ary Hypothesis Testing 2.1 Introduction (Levy 2.1) Chapter 2. Binary and M-ary Hypothesis Testing 2.1 Introduction (Levy 2.1) Detection problems can usually be casted as binary or M-ary hypothesis testing problems. Applications: This chapter: Simple hypothesis

More information

Direction: This test is worth 250 points and each problem worth points. DO ANY SIX

Direction: This test is worth 250 points and each problem worth points. DO ANY SIX Term Test 3 December 5, 2003 Name Math 52 Student Number Direction: This test is worth 250 points and each problem worth 4 points DO ANY SIX PROBLEMS You are required to complete this test within 50 minutes

More information

Advanced Signal Processing Introduction to Estimation Theory

Advanced Signal Processing Introduction to Estimation Theory Advanced Signal Processing Introduction to Estimation Theory Danilo Mandic, room 813, ext: 46271 Department of Electrical and Electronic Engineering Imperial College London, UK d.mandic@imperial.ac.uk,

More information

Module 2. Random Processes. Version 2, ECE IIT, Kharagpur

Module 2. Random Processes. Version 2, ECE IIT, Kharagpur Module Random Processes Version, ECE IIT, Kharagpur Lesson 9 Introduction to Statistical Signal Processing Version, ECE IIT, Kharagpur After reading this lesson, you will learn about Hypotheses testing

More information

ECE 275B Homework # 1 Solutions Version Winter 2015

ECE 275B Homework # 1 Solutions Version Winter 2015 ECE 275B Homework # 1 Solutions Version Winter 2015 1. (a) Because x i are assumed to be independent realizations of a continuous random variable, it is almost surely (a.s.) 1 the case that x 1 < x 2

More information

A Very Brief Summary of Statistical Inference, and Examples

A Very Brief Summary of Statistical Inference, and Examples A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2008 Prof. Gesine Reinert 1 Data x = x 1, x 2,..., x n, realisations of random variables X 1, X 2,..., X n with distribution (model)

More information

Mathematical statistics

Mathematical statistics October 1 st, 2018 Lecture 11: Sufficient statistic Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation

More information

Statistics Ph.D. Qualifying Exam: Part I October 18, 2003

Statistics Ph.D. Qualifying Exam: Part I October 18, 2003 Statistics Ph.D. Qualifying Exam: Part I October 18, 2003 Student Name: 1. Answer 8 out of 12 problems. Mark the problems you selected in the following table. 1 2 3 4 5 6 7 8 9 10 11 12 2. Write your answer

More information

An Introduction to Signal Detection and Estimation - Second Edition Chapter III: Selected Solutions

An Introduction to Signal Detection and Estimation - Second Edition Chapter III: Selected Solutions An Introduction to Signal Detection and Estimation - Second Edition Chapter III: Selected Solutions H. V. Poor Princeton University March 17, 5 Exercise 1: Let {h k,l } denote the impulse response of a

More information

Lecture 7 Introduction to Statistical Decision Theory

Lecture 7 Introduction to Statistical Decision Theory Lecture 7 Introduction to Statistical Decision Theory I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 20, 2016 1 / 55 I-Hsiang Wang IT Lecture 7

More information

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it

More information

Hypothesis Testing: The Generalized Likelihood Ratio Test

Hypothesis Testing: The Generalized Likelihood Ratio Test Hypothesis Testing: The Generalized Likelihood Ratio Test Consider testing the hypotheses H 0 : θ Θ 0 H 1 : θ Θ \ Θ 0 Definition: The Generalized Likelihood Ratio (GLR Let L(θ be a likelihood for a random

More information

First Year Examination Department of Statistics, University of Florida

First Year Examination Department of Statistics, University of Florida First Year Examination Department of Statistics, University of Florida August 19, 010, 8:00 am - 1:00 noon Instructions: 1. You have four hours to answer questions in this examination.. You must show your

More information

Chapter 3: Maximum Likelihood Theory

Chapter 3: Maximum Likelihood Theory Chapter 3: Maximum Likelihood Theory Florian Pelgrin HEC September-December, 2010 Florian Pelgrin (HEC) Maximum Likelihood Theory September-December, 2010 1 / 40 1 Introduction Example 2 Maximum likelihood

More information

Brief Review on Estimation Theory

Brief Review on Estimation Theory Brief Review on Estimation Theory K. Abed-Meraim ENST PARIS, Signal and Image Processing Dept. abed@tsi.enst.fr This presentation is essentially based on the course BASTA by E. Moulines Brief review on

More information

Statistics for scientists and engineers

Statistics for scientists and engineers Statistics for scientists and engineers February 0, 006 Contents Introduction. Motivation - why study statistics?................................... Examples..................................................3

More information

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 26. Estimation: Regression and Least Squares

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 26. Estimation: Regression and Least Squares CS 70 Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 26 Estimation: Regression and Least Squares This note explains how to use observations to estimate unobserved random variables.

More information

STAT215: Solutions for Homework 2

STAT215: Solutions for Homework 2 STAT25: Solutions for Homework 2 Due: Wednesday, Feb 4. (0 pt) Suppose we take one observation, X, from the discrete distribution, x 2 0 2 Pr(X x θ) ( θ)/4 θ/2 /2 (3 θ)/2 θ/4, 0 θ Find an unbiased estimator

More information

ECE 275B Homework # 1 Solutions Winter 2018

ECE 275B Homework # 1 Solutions Winter 2018 ECE 275B Homework # 1 Solutions Winter 2018 1. (a) Because x i are assumed to be independent realizations of a continuous random variable, it is almost surely (a.s.) 1 the case that x 1 < x 2 < < x n Thus,

More information

ECE531 Lecture 11: Dynamic Parameter Estimation: Kalman-Bucy Filter

ECE531 Lecture 11: Dynamic Parameter Estimation: Kalman-Bucy Filter ECE531 Lecture 11: Dynamic Parameter Estimation: Kalman-Bucy Filter D. Richard Brown III Worcester Polytechnic Institute 09-Apr-2009 Worcester Polytechnic Institute D. Richard Brown III 09-Apr-2009 1 /

More information

Mathematical statistics

Mathematical statistics October 4 th, 2018 Lecture 12: Information Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation Chapter

More information

Ch. 8 Math Preliminaries for Lossy Coding. 8.4 Info Theory Revisited

Ch. 8 Math Preliminaries for Lossy Coding. 8.4 Info Theory Revisited Ch. 8 Math Preliminaries for Lossy Coding 8.4 Info Theory Revisited 1 Info Theory Goals for Lossy Coding Again just as for the lossless case Info Theory provides: Basis for Algorithms & Bounds on Performance

More information

Estimation Theory. as Θ = (Θ 1,Θ 2,...,Θ m ) T. An estimator

Estimation Theory. as Θ = (Θ 1,Θ 2,...,Θ m ) T. An estimator Estimation Theory Estimation theory deals with finding numerical values of interesting parameters from given set of data. We start with formulating a family of models that could describe how the data were

More information

ECE531 Lecture 10b: Dynamic Parameter Estimation: System Model

ECE531 Lecture 10b: Dynamic Parameter Estimation: System Model ECE531 Lecture 10b: Dynamic Parameter Estimation: System Model D. Richard Brown III Worcester Polytechnic Institute 02-Apr-2009 Worcester Polytechnic Institute D. Richard Brown III 02-Apr-2009 1 / 14 Introduction

More information

ECE 275A Homework 7 Solutions

ECE 275A Homework 7 Solutions ECE 275A Homework 7 Solutions Solutions 1. For the same specification as in Homework Problem 6.11 we want to determine an estimator for θ using the Method of Moments (MOM). In general, the MOM estimator

More information

A Very Brief Summary of Statistical Inference, and Examples

A Very Brief Summary of Statistical Inference, and Examples A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2009 Prof. Gesine Reinert Our standard situation is that we have data x = x 1, x 2,..., x n, which we view as realisations of random

More information

Estimation Theory Fredrik Rusek. Chapters 6-7

Estimation Theory Fredrik Rusek. Chapters 6-7 Estimation Theory Fredrik Rusek Chapters 6-7 All estimation problems Summary All estimation problems Summary Efficient estimator exists All estimation problems Summary MVU estimator exists Efficient estimator

More information

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3 Hypothesis Testing CB: chapter 8; section 0.3 Hypothesis: statement about an unknown population parameter Examples: The average age of males in Sweden is 7. (statement about population mean) The lowest

More information

Module 22: Bayesian Methods Lecture 9 A: Default prior selection

Module 22: Bayesian Methods Lecture 9 A: Default prior selection Module 22: Bayesian Methods Lecture 9 A: Default prior selection Peter Hoff Departments of Statistics and Biostatistics University of Washington Outline Jeffreys prior Unit information priors Empirical

More information

f(y θ) = g(t (y) θ)h(y)

f(y θ) = g(t (y) θ)h(y) EXAM3, FINAL REVIEW (and a review for some of the QUAL problems): No notes will be allowed, but you may bring a calculator. Memorize the pmf or pdf f, E(Y ) and V(Y ) for the following RVs: 1) beta(δ,

More information

Lecture 25: Review. Statistics 104. April 23, Colin Rundel

Lecture 25: Review. Statistics 104. April 23, Colin Rundel Lecture 25: Review Statistics 104 Colin Rundel April 23, 2012 Joint CDF F (x, y) = P [X x, Y y] = P [(X, Y ) lies south-west of the point (x, y)] Y (x,y) X Statistics 104 (Colin Rundel) Lecture 25 April

More information

Estimation techniques

Estimation techniques Estimation techniques March 2, 2006 Contents 1 Problem Statement 2 2 Bayesian Estimation Techniques 2 2.1 Minimum Mean Squared Error (MMSE) estimation........................ 2 2.1.1 General formulation......................................

More information

MA 575 Linear Models: Cedric E. Ginestet, Boston University Midterm Review Week 7

MA 575 Linear Models: Cedric E. Ginestet, Boston University Midterm Review Week 7 MA 575 Linear Models: Cedric E. Ginestet, Boston University Midterm Review Week 7 1 Random Vectors Let a 0 and y be n 1 vectors, and let A be an n n matrix. Here, a 0 and A are non-random, whereas y is

More information

UCSD ECE250 Handout #27 Prof. Young-Han Kim Friday, June 8, Practice Final Examination (Winter 2017)

UCSD ECE250 Handout #27 Prof. Young-Han Kim Friday, June 8, Practice Final Examination (Winter 2017) UCSD ECE250 Handout #27 Prof. Young-Han Kim Friday, June 8, 208 Practice Final Examination (Winter 207) There are 6 problems, each problem with multiple parts. Your answer should be as clear and readable

More information

Lecture 8: Information Theory and Statistics

Lecture 8: Information Theory and Statistics Lecture 8: Information Theory and Statistics Part II: Hypothesis Testing and I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 23, 2015 1 / 50 I-Hsiang

More information

Estimation Tasks. Short Course on Image Quality. Matthew A. Kupinski. Introduction

Estimation Tasks. Short Course on Image Quality. Matthew A. Kupinski. Introduction Estimation Tasks Short Course on Image Quality Matthew A. Kupinski Introduction Section 13.3 in B&M Keep in mind the similarities between estimation and classification Image-quality is a statistical concept

More information

ECE531 Lecture 2a: A Mathematical Model for Hypothesis Testing (Finite Number of Possible Observations)

ECE531 Lecture 2a: A Mathematical Model for Hypothesis Testing (Finite Number of Possible Observations) ECE531 Lecture 2a: A Mathematical Model for Hypothesis Testing (Finite Number of Possible Observations) D. Richard Brown III Worcester Polytechnic Institute 26-January-2011 Worcester Polytechnic Institute

More information

f X, Y (x, y)dx (x), where f(x,y) is the joint pdf of X and Y. (x) dx

f X, Y (x, y)dx (x), where f(x,y) is the joint pdf of X and Y. (x) dx INDEPENDENCE, COVARIANCE AND CORRELATION Independence: Intuitive idea of "Y is independent of X": The distribution of Y doesn't depend on the value of X. In terms of the conditional pdf's: "f(y x doesn't

More information

Parametric Techniques Lecture 3

Parametric Techniques Lecture 3 Parametric Techniques Lecture 3 Jason Corso SUNY at Buffalo 22 January 2009 J. Corso (SUNY at Buffalo) Parametric Techniques Lecture 3 22 January 2009 1 / 39 Introduction In Lecture 2, we learned how to

More information

LECTURE NOTE #3 PROF. ALAN YUILLE

LECTURE NOTE #3 PROF. ALAN YUILLE LECTURE NOTE #3 PROF. ALAN YUILLE 1. Three Topics (1) Precision and Recall Curves. Receiver Operating Characteristic Curves (ROC). What to do if we do not fix the loss function? (2) The Curse of Dimensionality.

More information

A Few Notes on Fisher Information (WIP)

A Few Notes on Fisher Information (WIP) A Few Notes on Fisher Information (WIP) David Meyer dmm@{-4-5.net,uoregon.edu} Last update: April 30, 208 Definitions There are so many interesting things about Fisher Information and its theoretical properties

More information

ESTIMATION THEORY. Chapter Estimation of Random Variables

ESTIMATION THEORY. Chapter Estimation of Random Variables Chapter ESTIMATION THEORY. Estimation of Random Variables Suppose X,Y,Y 2,...,Y n are random variables defined on the same probability space (Ω, S,P). We consider Y,...,Y n to be the observed random variables

More information

10. Composite Hypothesis Testing. ECE 830, Spring 2014

10. Composite Hypothesis Testing. ECE 830, Spring 2014 10. Composite Hypothesis Testing ECE 830, Spring 2014 1 / 25 In many real world problems, it is difficult to precisely specify probability distributions. Our models for data may involve unknown parameters

More information

6.1 Variational representation of f-divergences

6.1 Variational representation of f-divergences ECE598: Information-theoretic methods in high-dimensional statistics Spring 2016 Lecture 6: Variational representation, HCR and CR lower bounds Lecturer: Yihong Wu Scribe: Georgios Rovatsos, Feb 11, 2016

More information

LECTURE 5 NOTES. n t. t Γ(a)Γ(b) pt+a 1 (1 p) n t+b 1. The marginal density of t is. Γ(t + a)γ(n t + b) Γ(n + a + b)

LECTURE 5 NOTES. n t. t Γ(a)Γ(b) pt+a 1 (1 p) n t+b 1. The marginal density of t is. Γ(t + a)γ(n t + b) Γ(n + a + b) LECTURE 5 NOTES 1. Bayesian point estimators. In the conventional (frequentist) approach to statistical inference, the parameter θ Θ is considered a fixed quantity. In the Bayesian approach, it is considered

More information

Lecture 4: UMVUE and unbiased estimators of 0

Lecture 4: UMVUE and unbiased estimators of 0 Lecture 4: UMVUE and unbiased estimators of Problems of approach 1. Approach 1 has the following two main shortcomings. Even if Theorem 7.3.9 or its extension is applicable, there is no guarantee that

More information

Ch. 5 Hypothesis Testing

Ch. 5 Hypothesis Testing Ch. 5 Hypothesis Testing The current framework of hypothesis testing is largely due to the work of Neyman and Pearson in the late 1920s, early 30s, complementing Fisher s work on estimation. As in estimation,

More information

STA 260: Statistics and Probability II

STA 260: Statistics and Probability II Al Nosedal. University of Toronto. Winter 2017 1 Properties of Point Estimators and Methods of Estimation 2 3 If you can t explain it simply, you don t understand it well enough Albert Einstein. Definition

More information

Ch. 8 Math Preliminaries for Lossy Coding. 8.5 Rate-Distortion Theory

Ch. 8 Math Preliminaries for Lossy Coding. 8.5 Rate-Distortion Theory Ch. 8 Math Preliminaries for Lossy Coding 8.5 Rate-Distortion Theory 1 Introduction Theory provide insight into the trade between Rate & Distortion This theory is needed to answer: What do typical R-D

More information

3.0.1 Multivariate version and tensor product of experiments

3.0.1 Multivariate version and tensor product of experiments ECE598: Information-theoretic methods in high-dimensional statistics Spring 2016 Lecture 3: Minimax risk of GLM and four extensions Lecturer: Yihong Wu Scribe: Ashok Vardhan, Jan 28, 2016 [Ed. Mar 24]

More information

Parametric Techniques

Parametric Techniques Parametric Techniques Jason J. Corso SUNY at Buffalo J. Corso (SUNY at Buffalo) Parametric Techniques 1 / 39 Introduction When covering Bayesian Decision Theory, we assumed the full probabilistic structure

More information

Rowan University Department of Electrical and Computer Engineering

Rowan University Department of Electrical and Computer Engineering Rowan University Department of Electrical and Computer Engineering Estimation and Detection Theory Fall 2013 to Practice Exam II This is a closed book exam. There are 8 problems in the exam. The problems

More information

Machine Learning. Lecture 4: Regularization and Bayesian Statistics. Feng Li. https://funglee.github.io

Machine Learning. Lecture 4: Regularization and Bayesian Statistics. Feng Li. https://funglee.github.io Machine Learning Lecture 4: Regularization and Bayesian Statistics Feng Li fli@sdu.edu.cn https://funglee.github.io School of Computer Science and Technology Shandong University Fall 207 Overfitting Problem

More information

18.440: Lecture 28 Lectures Review

18.440: Lecture 28 Lectures Review 18.440: Lecture 28 Lectures 18-27 Review Scott Sheffield MIT Outline Outline It s the coins, stupid Much of what we have done in this course can be motivated by the i.i.d. sequence X i where each X i is

More information

Master s Written Examination

Master s Written Examination Master s Written Examination Option: Statistics and Probability Spring 05 Full points may be obtained for correct answers to eight questions Each numbered question (which may have several parts) is worth

More information

Estimation and Detection

Estimation and Detection stimation and Detection Lecture 2: Cramér-Rao Lower Bound Dr. ir. Richard C. Hendriks & Dr. Sundeep P. Chepuri 7//207 Remember: Introductory xample Given a process (DC in noise): x[n]=a + w[n], n=0,,n,

More information

BEST TESTS. Abstract. We will discuss the Neymann-Pearson theorem and certain best test where the power function is optimized.

BEST TESTS. Abstract. We will discuss the Neymann-Pearson theorem and certain best test where the power function is optimized. BEST TESTS Abstract. We will discuss the Neymann-Pearson theorem and certain best test where the power function is optimized. 1. Most powerful test Let {f θ } θ Θ be a family of pdfs. We will consider

More information

Probability on a Riemannian Manifold

Probability on a Riemannian Manifold Probability on a Riemannian Manifold Jennifer Pajda-De La O December 2, 2015 1 Introduction We discuss how we can construct probability theory on a Riemannian manifold. We make comparisons to this and

More information

Spring 2012 Math 541A Exam 1. X i, S 2 = 1 n. n 1. X i I(X i < c), T n =

Spring 2012 Math 541A Exam 1. X i, S 2 = 1 n. n 1. X i I(X i < c), T n = Spring 2012 Math 541A Exam 1 1. (a) Let Z i be independent N(0, 1), i = 1, 2,, n. Are Z = 1 n n Z i and S 2 Z = 1 n 1 n (Z i Z) 2 independent? Prove your claim. (b) Let X 1, X 2,, X n be independent identically

More information

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A. 1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n

More information

Appendix A : Introduction to Probability and stochastic processes

Appendix A : Introduction to Probability and stochastic processes A-1 Mathematical methods in communication July 5th, 2009 Appendix A : Introduction to Probability and stochastic processes Lecturer: Haim Permuter Scribe: Shai Shapira and Uri Livnat The probability of

More information

Lecture Notes 4 Vector Detection and Estimation. Vector Detection Reconstruction Problem Detection for Vector AGN Channel

Lecture Notes 4 Vector Detection and Estimation. Vector Detection Reconstruction Problem Detection for Vector AGN Channel Lecture Notes 4 Vector Detection and Estimation Vector Detection Reconstruction Problem Detection for Vector AGN Channel Vector Linear Estimation Linear Innovation Sequence Kalman Filter EE 278B: Random

More information

Problem 1 (20) Log-normal. f(x) Cauchy

Problem 1 (20) Log-normal. f(x) Cauchy ORF 245. Rigollet Date: 11/21/2008 Problem 1 (20) f(x) f(x) 0.0 0.1 0.2 0.3 0.4 0.0 0.2 0.4 0.6 0.8 4 2 0 2 4 Normal (with mean -1) 4 2 0 2 4 Negative-exponential x x f(x) f(x) 0.0 0.1 0.2 0.3 0.4 0.5

More information

Problem Set 2. MAS 622J/1.126J: Pattern Recognition and Analysis. Due: 5:00 p.m. on September 30

Problem Set 2. MAS 622J/1.126J: Pattern Recognition and Analysis. Due: 5:00 p.m. on September 30 Problem Set 2 MAS 622J/1.126J: Pattern Recognition and Analysis Due: 5:00 p.m. on September 30 [Note: All instructions to plot data or write a program should be carried out using Matlab. In order to maintain

More information

1. (Regular) Exponential Family

1. (Regular) Exponential Family 1. (Regular) Exponential Family The density function of a regular exponential family is: [ ] Example. Poisson(θ) [ ] Example. Normal. (both unknown). ) [ ] [ ] [ ] [ ] 2. Theorem (Exponential family &

More information

Statistics: Learning models from data

Statistics: Learning models from data DS-GA 1002 Lecture notes 5 October 19, 2015 Statistics: Learning models from data Learning models from data that are assumed to be generated probabilistically from a certain unknown distribution is a crucial

More information

ST5215: Advanced Statistical Theory

ST5215: Advanced Statistical Theory Department of Statistics & Applied Probability Wednesday, October 19, 2011 Lecture 17: UMVUE and the first method of derivation Estimable parameters Let ϑ be a parameter in the family P. If there exists

More information

Overview. Probabilistic Interpretation of Linear Regression Maximum Likelihood Estimation Bayesian Estimation MAP Estimation

Overview. Probabilistic Interpretation of Linear Regression Maximum Likelihood Estimation Bayesian Estimation MAP Estimation Overview Probabilistic Interpretation of Linear Regression Maximum Likelihood Estimation Bayesian Estimation MAP Estimation Probabilistic Interpretation: Linear Regression Assume output y is generated

More information

F2E5216/TS1002 Adaptive Filtering and Change Detection. Course Organization. Lecture plan. The Books. Lecture 1

F2E5216/TS1002 Adaptive Filtering and Change Detection. Course Organization. Lecture plan. The Books. Lecture 1 Adaptive Filtering and Change Detection Bo Wahlberg (KTH and Fredrik Gustafsson (LiTH Course Organization Lectures and compendium: Theory, Algorithms, Applications, Evaluation Toolbox and manual: Algorithms,

More information

ECE 636: Systems identification

ECE 636: Systems identification ECE 636: Systems identification Lectures 3 4 Random variables/signals (continued) Random/stochastic vectors Random signals and linear systems Random signals in the frequency domain υ ε x S z + y Experimental

More information

Chapter 3: Unbiased Estimation Lecture 22: UMVUE and the method of using a sufficient and complete statistic

Chapter 3: Unbiased Estimation Lecture 22: UMVUE and the method of using a sufficient and complete statistic Chapter 3: Unbiased Estimation Lecture 22: UMVUE and the method of using a sufficient and complete statistic Unbiased estimation Unbiased or asymptotically unbiased estimation plays an important role in

More information