EIE6207: Estimation Theory

Size: px
Start display at page:

Download "EIE6207: Estimation Theory"

Transcription

1 EIE6207: Estimation Theory Man-Wai MAK Dept. of Electronic and Information Engineering, The Hong Kong Polytechnic University mwmak References: Steven M. Kay, Fundamentals of Statistical Signal Processing, Prentice Hall, November 8, 2018 Man-Wai MAK (EIE) Estimation Theory November 8, / 28

2 Overview 1 Motivations 2 Introduction 3 Minimum Variance Unbiased Estimation 4 Cramer-Rao Lower Bound Man-Wai MAK (EIE) Estimation Theory November 8, / 28

3 Motivations Estimation in signal processing is fundamental to many applications: Radar Sonar Speech and audio processing Image and video processing Biomedical and biomedicine Communications Control engineering Seismology All of these applications require to estimate a set of parameters, e.g., the position of aircraft through radar, position of a submarine through sonar, and phonemes of speech waveform. All of these applications have a common characteristics: The parameters are estimated from noisy signals. So, their value can only be an estimate. Man-Wai MAK (EIE) Estimation Theory November 8, / 28

4 Introduction In all of these applications, we estimate the values of parameters based on some waveforms. Specifically, Given N discrete-time {x[0], x[1],..., x[n 1]} that depends on an unknown parameter θ, we define an estimator ˆθ to estimate θ: ˆθ = g(x[0], x[1],..., x[n 1]), where g is some function. For example, for the mean estimator, g is given by g(x[0], x[1],..., x[n 1]) = 1 N N 1 i=0 x[i] Man-Wai MAK (EIE) Estimation Theory November 8, / 28

5 Introduction To determine good estimators, we need to mathematically model the data. We describe the data using probability density function (PDF) parameterized by θ: p(x[0],..., x[n 1]; θ) For example, when N = 1 and assume Gaussian PDF, we have [ 1 p(x[0]) = exp 1 ] (x[0] θ)2 2πσ 2 2σ2 In actual problems, we do not have the PDF but only the data. So, we select an appropriate parametric PDF based on our prior knowledge of the problem. The PDF should also be mathematically tractable. Man-Wai MAK (EIE) Estimation Theory November 8, / 28

6 Introduction The stock price in the figure below seems to increase on the average. To verify this hypothesis, we use the model: x[n] = A + Bn + w[n], n = 0, 1,..., N 1, where w[n] is white Gaussian noise with PDF N (0, σ 2 ). Figure: y-axis: stock price x[n]; x-axis: time point n Man-Wai MAK (EIE) Estimation Theory November 8, / 28

7 Introduction So far, we assume that θ is deterministic but unknown. This is called classical estimation. But we could also bring in prior knowledge about θ, e.g., prior range or distribution of A and B. Then, we have a Bayesian estimator. The data and parameters are modeled by a joint PDF: p(x, θ) = p(x θ)p(θ), where p(θ) is the prior PDF, which summarizes our knowledge about θ before any data are observed. p(x θ) is the conditional PDF, summarizing our knowledge about x when we know θ. Man-Wai MAK (EIE) Estimation Theory November 8, / 28

8 Introduction If w[n] is uncorrelated with x[i], then p(x; θ) = = = N 1 n=1 N 1 n=0 p(x[n]; θ) [ 1 exp 2πσ 2 1 (2πσ 2 exp ) N/2 1 ] (x[n] A Bn)2 2σ2 [ ] 1 N 1 2σ 2 (x[n] A Bn) 2 n=0 where x = [x[0]... x[n 1]] T and θ = [A B] T. Maximizing p(x; θ) with respect to θ, we obtain close-form solutions for A and B. For the diagram shown in previous page, B > 0, meaning the stock price is increasing. Man-Wai MAK (EIE) Estimation Theory November 8, / 28

9 Minimum Variance Unbiased Estimation An estimator is unbiased if on average it yield the true value of the unknown parameter θ. Mathematically, an estimator is unbiased if E(ˆθ) = θ a < θ < b. Example: Unbiased estimator for DC level with white Gaussian noise (WGN): Consider the observations x[n] = A + w[n] n = 0, 1,..., N 1 (1) where < A < is the parameter to be estimated and w[n] is WGN. A reasonable estimator for the average value of x[n] is  = 1 N N 1 n=0 x[n] Man-Wai MAK (EIE) Estimation Theory November 8, / 28

10 Minimum Variance Unbiased Estimation Using the linearity property of expectation: E(Â) = 1 N N 1 n=0 E(x[n]) = A. The sample mean is an unbiased estimator of the DC level in WGN. If multiple estimators of the same parameter θ are available, i.e., {ˆθ 1, ˆθ 2,..., ˆθ K }, a better estimator can be obtained by averaging these estimators: ˆθ = 1 K ˆθ k. K k=1 If all estimators are unbiased, we have E(ˆθ) = θ If all estimators have the same variance, then var(ˆθ) = 1 K K 2 var(ˆθ k ) = var(ˆθ 1 ) = variance decrease k=1 K Man-Wai MAK (EIE) Estimation Theory November 8, / 28

11 Minimum Variance Unbiased Estimation However, a biased estimator is characterized by a systematic error, e.g., E(ˆθ k ) = θ + b(θ). Then, E(ˆθ) = 1 K K E(ˆθ k ) = θ + b(θ) k=1 This means that having more unbiased estimators does not mean that the bias can be reduced. Man-Wai MAK (EIE) Estimation Theory November 8, / 28

12 Minimum Variance Criterion Mean square error (MSE): mse(ˆθ) = E {(ˆθ θ) 2} The MSE has problem if the estimator is biased, because { [ ] } 2 mse(ˆθ) = E (ˆθ E(ˆθ)) + (E(ˆθ) θ) [ ] 2 = var(ˆθ) + E(ˆθ) θ (2) = var(ˆθ) + b 2 (θ) This means that MSE comprises errors due to the variance of the estimators and the bias. Therefore, we should find an unbiased estimator (with b = 0) that produces the minimum variance. Man-Wai MAK (EIE) Estimation Theory November 8, / 28

13 Example of Biased Estimators Let the estimator of A in Eq. 1 be à = α N N 1 n=0 x[n], where α is an arbitrary constant. We would like to find the value of α that minimize { mse(ã) = E (à A)2} [ ] 2 = var(ã) + E{Ã} A = α2 σ 2 N + (α 1)2 A 2 Setting mse(ã) α = 0, we obtain α = A 2 A 2 + σ 2 /N Man-Wai MAK (EIE) Estimation Theory November 8, / 28

14 Example of Biased Estimators The optimal value α depends on the unknown parameter A. Therefore the estimator is not realizable. Man-Wai MAK (EIE) Estimation Theory November 8, / 28

15 Unbiased Estimators We could focus on estimators that have zero bias so that the bias contributes noting to the MSE. Without the bias term, the MSE in Eq. 2 does not involve the unknown parameter. By focusing on estimators with zero bias, we may hope to arrive at a design criterion that yields realizable estimators. Definition: An estimator ˆθ is called unbiased if its bias is zero for all values of the unknown parameter. Recap: For an estimator to be unbiased, we require that on average the estimator will yield the true value of the unknown parameter. Let the estimator be ˆθ = g(x), e.g., g can be the mean function. g is unbiased if θ : E{ˆθ} = g(x)p(x θ)dx = θ, then g(x) is an unbiased estimator. Man-Wai MAK (EIE) Estimation Theory November 8, / 28

16 Minimum Variance Criterion An unbiased estimator is not necessarily a good estimator. For example, Ã = 1 N 1 x[n] 2N n=0 E(Ã) = 1 { 0 if A = 0 = unbiased 2 A = 1A if A 0 = biased 2 Some unbiased estimators are more useful than others. For example, consider two estimators for Eq. 1: Â 1 = x[0] and Â2 = 1 N 1 N n=0 x[n]. Although E(Â1) = E(Â2) = A, Â 2 is better because var(â1) = σ 2 > var(â2) = σ2 N. (3) Man-Wai MAK (EIE) Estimation Theory November 8, / 28

17 Minimum Variance Unbiased Estimators If the bias is zero, the MSE is just the variance. This gives rise to the minimum variance unbiased estimator (MVUE) for θ. Definition: An estimator ˆθ is the MVUE if it is unbiased and has the smallest variance among any unbiased estimator for all values of the unknown parameter θ, i.e., ˆθ, we have var(ˆθ MVUE ) var(ˆθ) It is important to note that a uniformly MVUE may not always exist, and even if it does, we may not be able to find it. This definition and the above discussions apply to vector parameters θ. Man-Wai MAK (EIE) Estimation Theory November 8, / 28

18 Cramer-Rao Lower Bound Given an estimation problem, what is the variance of the best possible estimator? This quantity is given by the Cramer-Rao lower bound (CRLB). The CRLB theorem also provides a method for finding the best estimator. We will use the example in Eq. 1 to explain the idea of CRLB. For simplicity, suppose that we are using only a single observation x[0] to estimate A in Eq. 1. The PDF of x[0] is [ 1 p(x[0]; A) = exp 1 ] (x[0] A)2 2πσ 2 2σ2 Once we have observed x[0], say for example x[0] = 3, some values of A are more likely than others. Man-Wai MAK (EIE) Estimation Theory November 8, / 28

19 Cramer-Rao Lower Bound The pdf of A has the same form as the PDF of x[0]: [ 1 pdf of A = exp 1 ] (3 A)2 2πσ 2 2σ2 0.7 PDF plotted versus A in the case x[0]=3 and σ 2 = 1/ Value of A x[0] 0.7 PDF plotted versus A in the case x[0]=3 and σ 2 = Value of A x[0] We probably will get more accurate estimate of A for the distribution of A in the upper pannel. Man-Wai MAK (EIE) Estimation Theory November 8, / 28

20 Cramer-Rao Lower Bound If the pdf is viewed as a function of the unknown parameter (with x fixed), it is called the likelihood function (function of the unknown parameters). The sharpness of the likelihood function determines how accurately we can estimate the parameter. For example the pdf on the top panel of previous figure is the easier case. If a single sample is observed as x[0] = A + w[0], then we can expect a better estimate if σ 2 is small. How to quantify the sharpness? Is there any measure that would be common for all possible estimators for a specific estimation problem? The second derivative of the likelihood function (or log-likelihood) is one alternative for measuring the sharpness of a function. Man-Wai MAK (EIE) Estimation Theory November 8, / 28

21 Cramer-Rao Lower Bound Recall that the PDF of x[0] is [ 1 p(x[0]; A) = exp 1 ] (x[0] A)2 2πσ 2 2σ2 The log-likelihood function: p(x[0]; A) = log 2πσ 2 1 (x[0] A)2 2σ2 The first and second derivatives w.r.t. A are: log p(x[0]; A) = 1 (x[0] A) A σ2 2 log p(x[0]; A) A 2 = 1 σ 2 Man-Wai MAK (EIE) Estimation Theory November 8, / 28

22 Cramer-Rao Lower Bound Since σ 2 is the smallest possible variance, in this specific case we have an alternative way of finding the minimum variance of all estimators: minimum variance = σ 2 = 1 2 log p(x[0];a) A 2 For the general case, If the function depends on the data x, take the expectation over all x. If the function depends on the parameter θ, evaluate the derivative at the true value of θ, We have the general rule: min. var. of any unbiased estimator = 1 [ ], E 2 log p(x,θ) θ 2 Man-Wai MAK (EIE) Estimation Theory November 8, / 28

23 CRLB Theorem If the pdf p(x; θ) satisfies the regularity condition { } log p(x; θ) E = 0, θ, θ where the expectation is taken with respect to p(x; θ). Then, the variance of any unbiased estimator ˆθ must satisfy 1 var(ˆθ) [ ], E 2 log p(x,θ) θ 2 where the derivative is evalauted at the true value of θ and the expectation is taken with respect to p(x; θ). An unbiased estimator attains the bound for all θ if and only if log p(x; θ) θ = I(θ)(g(x) θ) (4) for some functions g and I. The MVU estimator is ˆθ = g(x) and the minimum variance = 1/I(θ). Man-Wai MAK (EIE) Estimation Theory November 8, / 28

24 CRLB Example 1: Estimation of DC Level in WGN Example: DC level in Gaussian noise: x[n] = A + w[n], n = 0, 1,..., N 1 What is the minimum variance of any unbiased estimator using N samples? The likelihood function is the product of N densities: p(x; A) = = N 1 n=0 [ 1 exp 2πσ 2 1 (2πσ 2 exp ) N/2 1 ] (x[n] A)2 2σ2 [ ] 1 N 1 2σ 2 (x[n] A) 2 n=0 Man-Wai MAK (EIE) Estimation Theory November 8, / 28

25 CRLB Example 1: Estimation of DC Level in WGN The log-likelihood function is [ ] log p(x; A) = log (2πσ 2 ) N 2 1 N 1 2σ 2 (x[n] A) 2 The first derivative is log p(x, A) A where x is the sample mean. The 2nd derivative is n=0 = 1 N 1 σ 2 (x[n] A) = N ( x A), σ2 n=0 2 log p(x, A) A 2 = N σ 2 Therefore, the minimum variance of any unbiased estimator is σ2 var(â) N (5) Man-Wai MAK (EIE) Estimation Theory November 8, / 28

26 CRLB Example 1: Estimation of DC Level in WGN In Eq. 3, the sample mean A 2 gives the minimum variance in Eq. 5. Therefore, the sample mean is an MVUE for this problem. Using the CRLB theorem in Eq. 4, we have log p(x; A) A = N ( x A) σ2 = I(A)(g(x) A), where I(A) = N and g(x) = x. σ 2 So, g(x) is the MVUE and 1/I(A) is its variance. Man-Wai MAK (EIE) Estimation Theory November 8, / 28

27 CRLB Example 2: Estimation of the Phase of Sinusoid Problem: Given N data points and a model with known amplitude A and frequency f 0 : x[n] = A cos(2πf 0 n + φ) + w[n], n = 0,..., N 1 estimate the phase φ of a sinusoid embedded in WGN with variance σ 2. How accurate will the estimator be? Solution: Derivation: See Tutorial var( ˆφ) 2σ 2 }{{} NA 2 approx Man-Wai MAK (EIE) Estimation Theory November 8, / 28

28 CRLB Summary Cramer Rao inequality provides lower bound for the estimation error variance. Minimum attainable variance is often larger than CRLB. We need to know the pdf to evaluate CRLB. Often we dont know this information and cannot evaluate this bound. If the data is multivariate Gaussian or i.i.d. with known distribution, we can evaluate it. Its not guaranteed that MVUE exists or is realizable. Man-Wai MAK (EIE) Estimation Theory November 8, / 28

EIE6207: Maximum-Likelihood and Bayesian Estimation

EIE6207: Maximum-Likelihood and Bayesian Estimation EIE6207: Maximum-Likelihood and Bayesian Estimation Man-Wai MAK Dept. of Electronic and Information Engineering, The Hong Kong Polytechnic University enmwmak@polyu.edu.hk http://www.eie.polyu.edu.hk/ mwmak

More information

A Few Notes on Fisher Information (WIP)

A Few Notes on Fisher Information (WIP) A Few Notes on Fisher Information (WIP) David Meyer dmm@{-4-5.net,uoregon.edu} Last update: April 30, 208 Definitions There are so many interesting things about Fisher Information and its theoretical properties

More information

Advanced Signal Processing Introduction to Estimation Theory

Advanced Signal Processing Introduction to Estimation Theory Advanced Signal Processing Introduction to Estimation Theory Danilo Mandic, room 813, ext: 46271 Department of Electrical and Electronic Engineering Imperial College London, UK d.mandic@imperial.ac.uk,

More information

ELEG 5633 Detection and Estimation Minimum Variance Unbiased Estimators (MVUE)

ELEG 5633 Detection and Estimation Minimum Variance Unbiased Estimators (MVUE) 1 ELEG 5633 Detection and Estimation Minimum Variance Unbiased Estimators (MVUE) Jingxian Wu Department of Electrical Engineering University of Arkansas Outline Minimum Variance Unbiased Estimators (MVUE)

More information

Detection & Estimation Lecture 1

Detection & Estimation Lecture 1 Detection & Estimation Lecture 1 Intro, MVUE, CRLB Xiliang Luo General Course Information Textbooks & References Fundamentals of Statistical Signal Processing: Estimation Theory/Detection Theory, Steven

More information

Detection & Estimation Lecture 1

Detection & Estimation Lecture 1 Detection & Estimation Lecture 1 Intro, MVUE, CRLB Xiliang Luo General Course Information Textbooks & References Fundamentals of Statistical Signal Processing: Estimation Theory/Detection Theory, Steven

More information

Rowan University Department of Electrical and Computer Engineering

Rowan University Department of Electrical and Computer Engineering Rowan University Department of Electrical and Computer Engineering Estimation and Detection Theory Fall 2013 to Practice Exam II This is a closed book exam. There are 8 problems in the exam. The problems

More information

Module 2. Random Processes. Version 2, ECE IIT, Kharagpur

Module 2. Random Processes. Version 2, ECE IIT, Kharagpur Module Random Processes Version, ECE IIT, Kharagpur Lesson 9 Introduction to Statistical Signal Processing Version, ECE IIT, Kharagpur After reading this lesson, you will learn about Hypotheses testing

More information

PATTERN RECOGNITION AND MACHINE LEARNING

PATTERN RECOGNITION AND MACHINE LEARNING PATTERN RECOGNITION AND MACHINE LEARNING Slide Set 2: Estimation Theory January 2018 Heikki Huttunen heikki.huttunen@tut.fi Department of Signal Processing Tampere University of Technology Classical Estimation

More information

Estimation and Detection

Estimation and Detection stimation and Detection Lecture 2: Cramér-Rao Lower Bound Dr. ir. Richard C. Hendriks & Dr. Sundeep P. Chepuri 7//207 Remember: Introductory xample Given a process (DC in noise): x[n]=a + w[n], n=0,,n,

More information

Chapter 8: Least squares (beginning of chapter)

Chapter 8: Least squares (beginning of chapter) Chapter 8: Least squares (beginning of chapter) Least Squares So far, we have been trying to determine an estimator which was unbiased and had minimum variance. Next we ll consider a class of estimators

More information

Linear models. x = Hθ + w, where w N(0, σ 2 I) and H R n p. The matrix H is called the observation matrix or design matrix 1.

Linear models. x = Hθ + w, where w N(0, σ 2 I) and H R n p. The matrix H is called the observation matrix or design matrix 1. Linear models As the first approach to estimator design, we consider the class of problems that can be represented by a linear model. In general, finding the MVUE is difficult. But if the linear model

More information

Advanced Signal Processing Minimum Variance Unbiased Estimation (MVU)

Advanced Signal Processing Minimum Variance Unbiased Estimation (MVU) Advanced Signal Processing Minimum Variance Unbiased Estimation (MVU) Danilo Mandic room 813, ext: 46271 Department of Electrical and Electronic Engineering Imperial College London, UK d.mandic@imperial.ac.uk,

More information

10-704: Information Processing and Learning Fall Lecture 24: Dec 7

10-704: Information Processing and Learning Fall Lecture 24: Dec 7 0-704: Information Processing and Learning Fall 206 Lecturer: Aarti Singh Lecture 24: Dec 7 Note: These notes are based on scribed notes from Spring5 offering of this course. LaTeX template courtesy of

More information

Module 1 - Signal estimation

Module 1 - Signal estimation , Arraial do Cabo, 2009 Module 1 - Signal estimation Sérgio M. Jesus (sjesus@ualg.pt) Universidade do Algarve, PT-8005-139 Faro, Portugal www.siplab.fct.ualg.pt February 2009 Outline of Module 1 Parameter

More information

Detection and Estimation Theory

Detection and Estimation Theory Detection and Estimation Theory Instructor: Prof. Namrata Vaswani Dept. of Electrical and Computer Engineering Iowa State University http://www.ece.iastate.edu/ namrata Slide 1 What is Estimation and Detection

More information

ECE531 Lecture 10b: Maximum Likelihood Estimation

ECE531 Lecture 10b: Maximum Likelihood Estimation ECE531 Lecture 10b: Maximum Likelihood Estimation D. Richard Brown III Worcester Polytechnic Institute 05-Apr-2011 Worcester Polytechnic Institute D. Richard Brown III 05-Apr-2011 1 / 23 Introduction So

More information

Terminology Suppose we have N observations {x(n)} N 1. Estimators as Random Variables. {x(n)} N 1

Terminology Suppose we have N observations {x(n)} N 1. Estimators as Random Variables. {x(n)} N 1 Estimation Theory Overview Properties Bias, Variance, and Mean Square Error Cramér-Rao lower bound Maximum likelihood Consistency Confidence intervals Properties of the mean estimator Properties of the

More information

Mathematical statistics

Mathematical statistics October 4 th, 2018 Lecture 12: Information Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation Chapter

More information

Detection theory. H 0 : x[n] = w[n]

Detection theory. H 0 : x[n] = w[n] Detection Theory Detection theory A the last topic of the course, we will briefly consider detection theory. The methods are based on estimation theory and attempt to answer questions such as Is a signal

More information

Classical Estimation Topics

Classical Estimation Topics Classical Estimation Topics Namrata Vaswani, Iowa State University February 25, 2014 This note fills in the gaps in the notes already provided (l0.pdf, l1.pdf, l2.pdf, l3.pdf, LeastSquares.pdf). 1 Min

More information

Detection theory 101 ELEC-E5410 Signal Processing for Communications

Detection theory 101 ELEC-E5410 Signal Processing for Communications Detection theory 101 ELEC-E5410 Signal Processing for Communications Binary hypothesis testing Null hypothesis H 0 : e.g. noise only Alternative hypothesis H 1 : signal + noise p(x;h 0 ) γ p(x;h 1 ) Trade-off

More information

Regression Estimation - Least Squares and Maximum Likelihood. Dr. Frank Wood

Regression Estimation - Least Squares and Maximum Likelihood. Dr. Frank Wood Regression Estimation - Least Squares and Maximum Likelihood Dr. Frank Wood Least Squares Max(min)imization Function to minimize w.r.t. β 0, β 1 Q = n (Y i (β 0 + β 1 X i )) 2 i=1 Minimize this by maximizing

More information

Estimation Theory Fredrik Rusek. Chapters

Estimation Theory Fredrik Rusek. Chapters Estimation Theory Fredrik Rusek Chapters 3.5-3.10 Recap We deal with unbiased estimators of deterministic parameters Performance of an estimator is measured by the variance of the estimate (due to the

More information

ECE534, Spring 2018: Solutions for Problem Set #3

ECE534, Spring 2018: Solutions for Problem Set #3 ECE534, Spring 08: Solutions for Problem Set #3 Jointly Gaussian Random Variables and MMSE Estimation Suppose that X, Y are jointly Gaussian random variables with µ X = µ Y = 0 and σ X = σ Y = Let their

More information

LINEAR MODELS IN STATISTICAL SIGNAL PROCESSING

LINEAR MODELS IN STATISTICAL SIGNAL PROCESSING TERM PAPER REPORT ON LINEAR MODELS IN STATISTICAL SIGNAL PROCESSING ANIMA MISHRA SHARMA, Y8104001 BISHWAJIT SHARMA, Y8104015 Course Instructor DR. RAJESH HEDGE Electrical Engineering Department IIT Kanpur

More information

Chapters 9. Properties of Point Estimators

Chapters 9. Properties of Point Estimators Chapters 9. Properties of Point Estimators Recap Target parameter, or population parameter θ. Population distribution f(x; θ). { probability function, discrete case f(x; θ) = density, continuous case The

More information

10. Linear Models and Maximum Likelihood Estimation

10. Linear Models and Maximum Likelihood Estimation 10. Linear Models and Maximum Likelihood Estimation ECE 830, Spring 2017 Rebecca Willett 1 / 34 Primary Goal General problem statement: We observe y i iid pθ, θ Θ and the goal is to determine the θ that

More information

Chapter 8.8.1: A factorization theorem

Chapter 8.8.1: A factorization theorem LECTURE 14 Chapter 8.8.1: A factorization theorem The characterization of a sufficient statistic in terms of the conditional distribution of the data given the statistic can be difficult to work with.

More information

STAT 730 Chapter 4: Estimation

STAT 730 Chapter 4: Estimation STAT 730 Chapter 4: Estimation Timothy Hanson Department of Statistics, University of South Carolina Stat 730: Multivariate Analysis 1 / 23 The likelihood We have iid data, at least initially. Each datum

More information

Introduction to Simple Linear Regression

Introduction to Simple Linear Regression Introduction to Simple Linear Regression Yang Feng http://www.stat.columbia.edu/~yangfeng Yang Feng (Columbia University) Introduction to Simple Linear Regression 1 / 68 About me Faculty in the Department

More information

Maximum Likelihood Estimation

Maximum Likelihood Estimation Connexions module: m11446 1 Maximum Likelihood Estimation Clayton Scott Robert Nowak This work is produced by The Connexions Project and licensed under the Creative Commons Attribution License Abstract

More information

13. Parameter Estimation. ECE 830, Spring 2014

13. Parameter Estimation. ECE 830, Spring 2014 13. Parameter Estimation ECE 830, Spring 2014 1 / 18 Primary Goal General problem statement: We observe X p(x θ), θ Θ and the goal is to determine the θ that produced X. Given a collection of observations

More information

Estimation MLE-Pandemic data MLE-Financial crisis data Evaluating estimators. Estimation. September 24, STAT 151 Class 6 Slide 1

Estimation MLE-Pandemic data MLE-Financial crisis data Evaluating estimators. Estimation. September 24, STAT 151 Class 6 Slide 1 Estimation September 24, 2018 STAT 151 Class 6 Slide 1 Pandemic data Treatment outcome, X, from n = 100 patients in a pandemic: 1 = recovered and 0 = not recovered 1 1 1 0 0 0 1 1 1 0 0 1 0 1 0 0 1 1 1

More information

ECE 275A Homework 6 Solutions

ECE 275A Homework 6 Solutions ECE 275A Homework 6 Solutions. The notation used in the solutions for the concentration (hyper) ellipsoid problems is defined in the lecture supplement on concentration ellipsoids. Note that θ T Σ θ =

More information

ECE 275A Homework 7 Solutions

ECE 275A Homework 7 Solutions ECE 275A Homework 7 Solutions Solutions 1. For the same specification as in Homework Problem 6.11 we want to determine an estimator for θ using the Method of Moments (MOM). In general, the MOM estimator

More information

Variations. ECE 6540, Lecture 10 Maximum Likelihood Estimation

Variations. ECE 6540, Lecture 10 Maximum Likelihood Estimation Variations ECE 6540, Lecture 10 Last Time BLUE (Best Linear Unbiased Estimator) Formulation Advantages Disadvantages 2 The BLUE A simplification Assume the estimator is a linear system For a single parameter

More information

DETECTION theory deals primarily with techniques for

DETECTION theory deals primarily with techniques for ADVANCED SIGNAL PROCESSING SE Optimum Detection of Deterministic and Random Signals Stefan Tertinek Graz University of Technology turtle@sbox.tugraz.at Abstract This paper introduces various methods for

More information

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it

More information

FAST AND ACCURATE DIRECTION-OF-ARRIVAL ESTIMATION FOR A SINGLE SOURCE

FAST AND ACCURATE DIRECTION-OF-ARRIVAL ESTIMATION FOR A SINGLE SOURCE Progress In Electromagnetics Research C, Vol. 6, 13 20, 2009 FAST AND ACCURATE DIRECTION-OF-ARRIVAL ESTIMATION FOR A SINGLE SOURCE Y. Wu School of Computer Science and Engineering Wuhan Institute of Technology

More information

Mathematical statistics

Mathematical statistics October 18 th, 2018 Lecture 16: Midterm review Countdown to mid-term exam: 7 days Week 1 Chapter 1: Probability review Week 2 Week 4 Week 7 Chapter 6: Statistics Chapter 7: Point Estimation Chapter 8:

More information

Lecture 8: Information Theory and Statistics

Lecture 8: Information Theory and Statistics Lecture 8: Information Theory and Statistics Part II: Hypothesis Testing and I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 23, 2015 1 / 50 I-Hsiang

More information

Estimation theory. Parametric estimation. Properties of estimators. Minimum variance estimator. Cramer-Rao bound. Maximum likelihood estimators

Estimation theory. Parametric estimation. Properties of estimators. Minimum variance estimator. Cramer-Rao bound. Maximum likelihood estimators Estimation theory Parametric estimation Properties of estimators Minimum variance estimator Cramer-Rao bound Maximum likelihood estimators Confidence intervals Bayesian estimation 1 Random Variables Let

More information

Statistics and Econometrics I

Statistics and Econometrics I Statistics and Econometrics I Point Estimation Shiu-Sheng Chen Department of Economics National Taiwan University September 13, 2016 Shiu-Sheng Chen (NTU Econ) Statistics and Econometrics I September 13,

More information

Methods of evaluating estimators and best unbiased estimators Hamid R. Rabiee

Methods of evaluating estimators and best unbiased estimators Hamid R. Rabiee Stochastic Processes Methods of evaluating estimators and best unbiased estimators Hamid R. Rabiee 1 Outline Methods of Mean Squared Error Bias and Unbiasedness Best Unbiased Estimators CR-Bound for variance

More information

Elements of statistics (MATH0487-1)

Elements of statistics (MATH0487-1) Elements of statistics (MATH0487-1) Prof. Dr. Dr. K. Van Steen University of Liège, Belgium November 12, 2012 Introduction to Statistics Basic Probability Revisited Sampling Exploratory Data Analysis -

More information

6.1 Variational representation of f-divergences

6.1 Variational representation of f-divergences ECE598: Information-theoretic methods in high-dimensional statistics Spring 2016 Lecture 6: Variational representation, HCR and CR lower bounds Lecturer: Yihong Wu Scribe: Georgios Rovatsos, Feb 11, 2016

More information

Graduate Econometrics I: Unbiased Estimation

Graduate Econometrics I: Unbiased Estimation Graduate Econometrics I: Unbiased Estimation Yves Dominicy Université libre de Bruxelles Solvay Brussels School of Economics and Management ECARES Yves Dominicy Graduate Econometrics I: Unbiased Estimation

More information

Brief Review on Estimation Theory

Brief Review on Estimation Theory Brief Review on Estimation Theory K. Abed-Meraim ENST PARIS, Signal and Image Processing Dept. abed@tsi.enst.fr This presentation is essentially based on the course BASTA by E. Moulines Brief review on

More information

10. Composite Hypothesis Testing. ECE 830, Spring 2014

10. Composite Hypothesis Testing. ECE 830, Spring 2014 10. Composite Hypothesis Testing ECE 830, Spring 2014 1 / 25 In many real world problems, it is difficult to precisely specify probability distributions. Our models for data may involve unknown parameters

More information

Math 494: Mathematical Statistics

Math 494: Mathematical Statistics Math 494: Mathematical Statistics Instructor: Jimin Ding jmding@wustl.edu Department of Mathematics Washington University in St. Louis Class materials are available on course website (www.math.wustl.edu/

More information

Estimation, Detection, and Identification CMU 18752

Estimation, Detection, and Identification CMU 18752 Estimation, Detection, and Identification CMU 18752 Graduate Course on the CMU/Portugal ECE PhD Program Spring 2008/2009 Instructor: Prof. Paulo Jorge Oliveira pjcro @ isr.ist.utl.pt Phone: +351 21 8418053

More information

PATTERN RECOGNITION AND MACHINE LEARNING

PATTERN RECOGNITION AND MACHINE LEARNING PATTERN RECOGNITION AND MACHINE LEARNING Slide Set 3: Detection Theory January 2018 Heikki Huttunen heikki.huttunen@tut.fi Department of Signal Processing Tampere University of Technology Detection theory

More information

Regression Estimation Least Squares and Maximum Likelihood

Regression Estimation Least Squares and Maximum Likelihood Regression Estimation Least Squares and Maximum Likelihood Dr. Frank Wood Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 3, Slide 1 Least Squares Max(min)imization Function to minimize

More information

F & B Approaches to a simple model

F & B Approaches to a simple model A6523 Signal Modeling, Statistical Inference and Data Mining in Astrophysics Spring 215 http://www.astro.cornell.edu/~cordes/a6523 Lecture 11 Applications: Model comparison Challenges in large-scale surveys

More information

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it

More information

Cramer-Rao Lower Bound Computation Via the Characteristic Function

Cramer-Rao Lower Bound Computation Via the Characteristic Function Cramer-Rao ower Bound Computation Via the Characteristic Function Steven Kay, Fellow, IEEE, and Cuichun Xu Abstract The Cramer-Rao ower Bound is widely used in statistical signal processing as a benchmark

More information

Estimators as Random Variables

Estimators as Random Variables Estimation Theory Overview Properties Bias, Variance, and Mean Square Error Cramér-Rao lower bound Maimum likelihood Consistency Confidence intervals Properties of the mean estimator Introduction Up until

More information

SGN Advanced Signal Processing: Lecture 8 Parameter estimation for AR and MA models. Model order selection

SGN Advanced Signal Processing: Lecture 8 Parameter estimation for AR and MA models. Model order selection SG 21006 Advanced Signal Processing: Lecture 8 Parameter estimation for AR and MA models. Model order selection Ioan Tabus Department of Signal Processing Tampere University of Technology Finland 1 / 28

More information

Parametric Techniques Lecture 3

Parametric Techniques Lecture 3 Parametric Techniques Lecture 3 Jason Corso SUNY at Buffalo 22 January 2009 J. Corso (SUNY at Buffalo) Parametric Techniques Lecture 3 22 January 2009 1 / 39 Introduction In Lecture 2, we learned how to

More information

Estimation, Detection, and Identification

Estimation, Detection, and Identification Estimation, Detection, and Identification Graduate Course on the CMU/Portugal ECE PhD Program Spring 2008/2009 Chapter 5 Best Linear Unbiased Estimators Instructor: Prof. Paulo Jorge Oliveira pjcro @ isr.ist.utl.pt

More information

Fundamentals of Statistical Signal Processing Volume II Detection Theory

Fundamentals of Statistical Signal Processing Volume II Detection Theory Fundamentals of Statistical Signal Processing Volume II Detection Theory Steven M. Kay University of Rhode Island PH PTR Prentice Hall PTR Upper Saddle River, New Jersey 07458 http://www.phptr.com Contents

More information

Review. December 4 th, Review

Review. December 4 th, Review December 4 th, 2017 Att. Final exam: Course evaluation Friday, 12/14/2018, 10:30am 12:30pm Gore Hall 115 Overview Week 2 Week 4 Week 7 Week 10 Week 12 Chapter 6: Statistics and Sampling Distributions Chapter

More information

Density Estimation. Seungjin Choi

Density Estimation. Seungjin Choi Density Estimation Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr http://mlg.postech.ac.kr/

More information

Parametric Techniques

Parametric Techniques Parametric Techniques Jason J. Corso SUNY at Buffalo J. Corso (SUNY at Buffalo) Parametric Techniques 1 / 39 Introduction When covering Bayesian Decision Theory, we assumed the full probabilistic structure

More information

A General Overview of Parametric Estimation and Inference Techniques.

A General Overview of Parametric Estimation and Inference Techniques. A General Overview of Parametric Estimation and Inference Techniques. Moulinath Banerjee University of Michigan September 11, 2012 The object of statistical inference is to glean information about an underlying

More information

Statistics: Learning models from data

Statistics: Learning models from data DS-GA 1002 Lecture notes 5 October 19, 2015 Statistics: Learning models from data Learning models from data that are assumed to be generated probabilistically from a certain unknown distribution is a crucial

More information

BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation

BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation Yujin Chung November 29th, 2016 Fall 2016 Yujin Chung Lec13: MLE Fall 2016 1/24 Previous Parametric tests Mean comparisons (normality assumption)

More information

MS&E 226: Small Data. Lecture 11: Maximum likelihood (v2) Ramesh Johari

MS&E 226: Small Data. Lecture 11: Maximum likelihood (v2) Ramesh Johari MS&E 226: Small Data Lecture 11: Maximum likelihood (v2) Ramesh Johari ramesh.johari@stanford.edu 1 / 18 The likelihood function 2 / 18 Estimating the parameter This lecture develops the methodology behind

More information

Rethinking Biased Estimation: Improving Maximum Likelihood and the Cramér Rao Bound

Rethinking Biased Estimation: Improving Maximum Likelihood and the Cramér Rao Bound Foundations and Trends R in Signal Processing Vol. 1, No. 4 (2007) 305 449 c 2008 Y. C. Eldar DOI: 10.1561/2000000008 Rethinking Biased Estimation: Improving Maximum Likelihood and the Cramér Rao Bound

More information

Econometrics I, Estimation

Econometrics I, Estimation Econometrics I, Estimation Department of Economics Stanford University September, 2008 Part I Parameter, Estimator, Estimate A parametric is a feature of the population. An estimator is a function of the

More information

System Identification, Lecture 4

System Identification, Lecture 4 System Identification, Lecture 4 Kristiaan Pelckmans (IT/UU, 2338) Course code: 1RT880, Report code: 61800 - Spring 2012 F, FRI Uppsala University, Information Technology 30 Januari 2012 SI-2012 K. Pelckmans

More information

System Identification, Lecture 4

System Identification, Lecture 4 System Identification, Lecture 4 Kristiaan Pelckmans (IT/UU, 2338) Course code: 1RT880, Report code: 61800 - Spring 2016 F, FRI Uppsala University, Information Technology 13 April 2016 SI-2016 K. Pelckmans

More information

Technique for Numerical Computation of Cramér-Rao Bound using MATLAB

Technique for Numerical Computation of Cramér-Rao Bound using MATLAB Technique for Numerical Computation of Cramér-Rao Bound using MATLAB Hing Cheung So http://www.ee.cityu.edu.hk/~hcso Department of Electronic Engineering City University of Hong Kong H. C. So Page 1 The

More information

Part IB Statistics. Theorems with proof. Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua. Lent 2015

Part IB Statistics. Theorems with proof. Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua. Lent 2015 Part IB Statistics Theorems with proof Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly)

More information

Estimation Theory. as Θ = (Θ 1,Θ 2,...,Θ m ) T. An estimator

Estimation Theory. as Θ = (Θ 1,Θ 2,...,Θ m ) T. An estimator Estimation Theory Estimation theory deals with finding numerical values of interesting parameters from given set of data. We start with formulating a family of models that could describe how the data were

More information

Machine Learning 4771

Machine Learning 4771 Machine Learning 4771 Instructor: Tony Jebara Topic 11 Maximum Likelihood as Bayesian Inference Maximum A Posteriori Bayesian Gaussian Estimation Why Maximum Likelihood? So far, assumed max (log) likelihood

More information

Gaussian, Markov and stationary processes

Gaussian, Markov and stationary processes Gaussian, Markov and stationary processes Gonzalo Mateos Dept. of ECE and Goergen Institute for Data Science University of Rochester gmateosb@ece.rochester.edu http://www.ece.rochester.edu/~gmateosb/ November

More information

Introduction to Estimation Methods for Time Series models Lecture 2

Introduction to Estimation Methods for Time Series models Lecture 2 Introduction to Estimation Methods for Time Series models Lecture 2 Fulvio Corsi SNS Pisa Fulvio Corsi Introduction to Estimation () Methods for Time Series models Lecture 2 SNS Pisa 1 / 21 Estimators:

More information

Lecture 7 Introduction to Statistical Decision Theory

Lecture 7 Introduction to Statistical Decision Theory Lecture 7 Introduction to Statistical Decision Theory I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 20, 2016 1 / 55 I-Hsiang Wang IT Lecture 7

More information

Economics 520. Lecture Note 19: Hypothesis Testing via the Neyman-Pearson Lemma CB 8.1,

Economics 520. Lecture Note 19: Hypothesis Testing via the Neyman-Pearson Lemma CB 8.1, Economics 520 Lecture Note 9: Hypothesis Testing via the Neyman-Pearson Lemma CB 8., 8.3.-8.3.3 Uniformly Most Powerful Tests and the Neyman-Pearson Lemma Let s return to the hypothesis testing problem

More information

Parametric Models. Dr. Shuang LIANG. School of Software Engineering TongJi University Fall, 2012

Parametric Models. Dr. Shuang LIANG. School of Software Engineering TongJi University Fall, 2012 Parametric Models Dr. Shuang LIANG School of Software Engineering TongJi University Fall, 2012 Today s Topics Maximum Likelihood Estimation Bayesian Density Estimation Today s Topics Maximum Likelihood

More information

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3 Hypothesis Testing CB: chapter 8; section 0.3 Hypothesis: statement about an unknown population parameter Examples: The average age of males in Sweden is 7. (statement about population mean) The lowest

More information

MA 575 Linear Models: Cedric E. Ginestet, Boston University Midterm Review Week 7

MA 575 Linear Models: Cedric E. Ginestet, Boston University Midterm Review Week 7 MA 575 Linear Models: Cedric E. Ginestet, Boston University Midterm Review Week 7 1 Random Vectors Let a 0 and y be n 1 vectors, and let A be an n n matrix. Here, a 0 and A are non-random, whereas y is

More information

Chapter 8: Estimation 1

Chapter 8: Estimation 1 Chapter 8: Estimation 1 Jae-Kwang Kim Iowa State University Fall, 2014 Kim (ISU) Ch. 8: Estimation 1 Fall, 2014 1 / 33 Introduction 1 Introduction 2 Ratio estimation 3 Regression estimator Kim (ISU) Ch.

More information

MAS223 Statistical Inference and Modelling Exercises

MAS223 Statistical Inference and Modelling Exercises MAS223 Statistical Inference and Modelling Exercises The exercises are grouped into sections, corresponding to chapters of the lecture notes Within each section exercises are divided into warm-up questions,

More information

DS-GA 1002 Lecture notes 11 Fall Bayesian statistics

DS-GA 1002 Lecture notes 11 Fall Bayesian statistics DS-GA 100 Lecture notes 11 Fall 016 Bayesian statistics In the frequentist paradigm we model the data as realizations from a distribution that depends on deterministic parameters. In contrast, in Bayesian

More information

COS513 LECTURE 8 STATISTICAL CONCEPTS

COS513 LECTURE 8 STATISTICAL CONCEPTS COS513 LECTURE 8 STATISTICAL CONCEPTS NIKOLAI SLAVOV AND ANKUR PARIKH 1. MAKING MEANINGFUL STATEMENTS FROM JOINT PROBABILITY DISTRIBUTIONS. A graphical model (GM) represents a family of probability distributions

More information

Parameter Estimation

Parameter Estimation Parameter Estimation Consider a sample of observations on a random variable Y. his generates random variables: (y 1, y 2,, y ). A random sample is a sample (y 1, y 2,, y ) where the random variables y

More information

Fractional Hot Deck Imputation for Robust Inference Under Item Nonresponse in Survey Sampling

Fractional Hot Deck Imputation for Robust Inference Under Item Nonresponse in Survey Sampling Fractional Hot Deck Imputation for Robust Inference Under Item Nonresponse in Survey Sampling Jae-Kwang Kim 1 Iowa State University June 26, 2013 1 Joint work with Shu Yang Introduction 1 Introduction

More information

ECE531 Lecture 8: Non-Random Parameter Estimation

ECE531 Lecture 8: Non-Random Parameter Estimation ECE531 Lecture 8: Non-Random Parameter Estimation D. Richard Brown III Worcester Polytechnic Institute 19-March-2009 Worcester Polytechnic Institute D. Richard Brown III 19-March-2009 1 / 25 Introduction

More information

Composite Hypotheses and Generalized Likelihood Ratio Tests

Composite Hypotheses and Generalized Likelihood Ratio Tests Composite Hypotheses and Generalized Likelihood Ratio Tests Rebecca Willett, 06 In many real world problems, it is difficult to precisely specify probability distributions. Our models for data may involve

More information

STAT 512 sp 2018 Summary Sheet

STAT 512 sp 2018 Summary Sheet STAT 5 sp 08 Summary Sheet Karl B. Gregory Spring 08. Transformations of a random variable Let X be a rv with support X and let g be a function mapping X to Y with inverse mapping g (A = {x X : g(x A}

More information

Parameter Estimation

Parameter Estimation 1 / 44 Parameter Estimation Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay October 25, 2012 Motivation System Model used to Derive

More information

Parametric Inference

Parametric Inference Parametric Inference Moulinath Banerjee University of Michigan April 14, 2004 1 General Discussion The object of statistical inference is to glean information about an underlying population based on a

More information

LECTURE NOTE #3 PROF. ALAN YUILLE

LECTURE NOTE #3 PROF. ALAN YUILLE LECTURE NOTE #3 PROF. ALAN YUILLE 1. Three Topics (1) Precision and Recall Curves. Receiver Operating Characteristic Curves (ROC). What to do if we do not fix the loss function? (2) The Curse of Dimensionality.

More information

Proof In the CR proof. and

Proof In the CR proof. and Question Under what conditions will we be able to attain the Cramér-Rao bound and find a MVUE? Lecture 4 - Consequences of the Cramér-Rao Lower Bound. Searching for a MVUE. Rao-Blackwell Theorem, Lehmann-Scheffé

More information

Central Limit Theorem ( 5.3)

Central Limit Theorem ( 5.3) Central Limit Theorem ( 5.3) Let X 1, X 2,... be a sequence of independent random variables, each having n mean µ and variance σ 2. Then the distribution of the partial sum S n = X i i=1 becomes approximately

More information

Estimation Theory for Engineers

Estimation Theory for Engineers Estimation Theory for Engineers Roberto Togneri 30th August 2005 Applications Modern estimation theory can be found at the heart of many electronic signal processing systems designed to extract information.

More information

2 Statistical Estimation: Basic Concepts

2 Statistical Estimation: Basic Concepts Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof. N. Shimkin 2 Statistical Estimation:

More information