EIE6207: Estimation Theory
|
|
- Asher Reed
- 5 years ago
- Views:
Transcription
1 EIE6207: Estimation Theory Man-Wai MAK Dept. of Electronic and Information Engineering, The Hong Kong Polytechnic University mwmak References: Steven M. Kay, Fundamentals of Statistical Signal Processing, Prentice Hall, November 8, 2018 Man-Wai MAK (EIE) Estimation Theory November 8, / 28
2 Overview 1 Motivations 2 Introduction 3 Minimum Variance Unbiased Estimation 4 Cramer-Rao Lower Bound Man-Wai MAK (EIE) Estimation Theory November 8, / 28
3 Motivations Estimation in signal processing is fundamental to many applications: Radar Sonar Speech and audio processing Image and video processing Biomedical and biomedicine Communications Control engineering Seismology All of these applications require to estimate a set of parameters, e.g., the position of aircraft through radar, position of a submarine through sonar, and phonemes of speech waveform. All of these applications have a common characteristics: The parameters are estimated from noisy signals. So, their value can only be an estimate. Man-Wai MAK (EIE) Estimation Theory November 8, / 28
4 Introduction In all of these applications, we estimate the values of parameters based on some waveforms. Specifically, Given N discrete-time {x[0], x[1],..., x[n 1]} that depends on an unknown parameter θ, we define an estimator ˆθ to estimate θ: ˆθ = g(x[0], x[1],..., x[n 1]), where g is some function. For example, for the mean estimator, g is given by g(x[0], x[1],..., x[n 1]) = 1 N N 1 i=0 x[i] Man-Wai MAK (EIE) Estimation Theory November 8, / 28
5 Introduction To determine good estimators, we need to mathematically model the data. We describe the data using probability density function (PDF) parameterized by θ: p(x[0],..., x[n 1]; θ) For example, when N = 1 and assume Gaussian PDF, we have [ 1 p(x[0]) = exp 1 ] (x[0] θ)2 2πσ 2 2σ2 In actual problems, we do not have the PDF but only the data. So, we select an appropriate parametric PDF based on our prior knowledge of the problem. The PDF should also be mathematically tractable. Man-Wai MAK (EIE) Estimation Theory November 8, / 28
6 Introduction The stock price in the figure below seems to increase on the average. To verify this hypothesis, we use the model: x[n] = A + Bn + w[n], n = 0, 1,..., N 1, where w[n] is white Gaussian noise with PDF N (0, σ 2 ). Figure: y-axis: stock price x[n]; x-axis: time point n Man-Wai MAK (EIE) Estimation Theory November 8, / 28
7 Introduction So far, we assume that θ is deterministic but unknown. This is called classical estimation. But we could also bring in prior knowledge about θ, e.g., prior range or distribution of A and B. Then, we have a Bayesian estimator. The data and parameters are modeled by a joint PDF: p(x, θ) = p(x θ)p(θ), where p(θ) is the prior PDF, which summarizes our knowledge about θ before any data are observed. p(x θ) is the conditional PDF, summarizing our knowledge about x when we know θ. Man-Wai MAK (EIE) Estimation Theory November 8, / 28
8 Introduction If w[n] is uncorrelated with x[i], then p(x; θ) = = = N 1 n=1 N 1 n=0 p(x[n]; θ) [ 1 exp 2πσ 2 1 (2πσ 2 exp ) N/2 1 ] (x[n] A Bn)2 2σ2 [ ] 1 N 1 2σ 2 (x[n] A Bn) 2 n=0 where x = [x[0]... x[n 1]] T and θ = [A B] T. Maximizing p(x; θ) with respect to θ, we obtain close-form solutions for A and B. For the diagram shown in previous page, B > 0, meaning the stock price is increasing. Man-Wai MAK (EIE) Estimation Theory November 8, / 28
9 Minimum Variance Unbiased Estimation An estimator is unbiased if on average it yield the true value of the unknown parameter θ. Mathematically, an estimator is unbiased if E(ˆθ) = θ a < θ < b. Example: Unbiased estimator for DC level with white Gaussian noise (WGN): Consider the observations x[n] = A + w[n] n = 0, 1,..., N 1 (1) where < A < is the parameter to be estimated and w[n] is WGN. A reasonable estimator for the average value of x[n] is  = 1 N N 1 n=0 x[n] Man-Wai MAK (EIE) Estimation Theory November 8, / 28
10 Minimum Variance Unbiased Estimation Using the linearity property of expectation: E(Â) = 1 N N 1 n=0 E(x[n]) = A. The sample mean is an unbiased estimator of the DC level in WGN. If multiple estimators of the same parameter θ are available, i.e., {ˆθ 1, ˆθ 2,..., ˆθ K }, a better estimator can be obtained by averaging these estimators: ˆθ = 1 K ˆθ k. K k=1 If all estimators are unbiased, we have E(ˆθ) = θ If all estimators have the same variance, then var(ˆθ) = 1 K K 2 var(ˆθ k ) = var(ˆθ 1 ) = variance decrease k=1 K Man-Wai MAK (EIE) Estimation Theory November 8, / 28
11 Minimum Variance Unbiased Estimation However, a biased estimator is characterized by a systematic error, e.g., E(ˆθ k ) = θ + b(θ). Then, E(ˆθ) = 1 K K E(ˆθ k ) = θ + b(θ) k=1 This means that having more unbiased estimators does not mean that the bias can be reduced. Man-Wai MAK (EIE) Estimation Theory November 8, / 28
12 Minimum Variance Criterion Mean square error (MSE): mse(ˆθ) = E {(ˆθ θ) 2} The MSE has problem if the estimator is biased, because { [ ] } 2 mse(ˆθ) = E (ˆθ E(ˆθ)) + (E(ˆθ) θ) [ ] 2 = var(ˆθ) + E(ˆθ) θ (2) = var(ˆθ) + b 2 (θ) This means that MSE comprises errors due to the variance of the estimators and the bias. Therefore, we should find an unbiased estimator (with b = 0) that produces the minimum variance. Man-Wai MAK (EIE) Estimation Theory November 8, / 28
13 Example of Biased Estimators Let the estimator of A in Eq. 1 be à = α N N 1 n=0 x[n], where α is an arbitrary constant. We would like to find the value of α that minimize { mse(ã) = E (à A)2} [ ] 2 = var(ã) + E{Ã} A = α2 σ 2 N + (α 1)2 A 2 Setting mse(ã) α = 0, we obtain α = A 2 A 2 + σ 2 /N Man-Wai MAK (EIE) Estimation Theory November 8, / 28
14 Example of Biased Estimators The optimal value α depends on the unknown parameter A. Therefore the estimator is not realizable. Man-Wai MAK (EIE) Estimation Theory November 8, / 28
15 Unbiased Estimators We could focus on estimators that have zero bias so that the bias contributes noting to the MSE. Without the bias term, the MSE in Eq. 2 does not involve the unknown parameter. By focusing on estimators with zero bias, we may hope to arrive at a design criterion that yields realizable estimators. Definition: An estimator ˆθ is called unbiased if its bias is zero for all values of the unknown parameter. Recap: For an estimator to be unbiased, we require that on average the estimator will yield the true value of the unknown parameter. Let the estimator be ˆθ = g(x), e.g., g can be the mean function. g is unbiased if θ : E{ˆθ} = g(x)p(x θ)dx = θ, then g(x) is an unbiased estimator. Man-Wai MAK (EIE) Estimation Theory November 8, / 28
16 Minimum Variance Criterion An unbiased estimator is not necessarily a good estimator. For example, Ã = 1 N 1 x[n] 2N n=0 E(Ã) = 1 { 0 if A = 0 = unbiased 2 A = 1A if A 0 = biased 2 Some unbiased estimators are more useful than others. For example, consider two estimators for Eq. 1: Â 1 = x[0] and Â2 = 1 N 1 N n=0 x[n]. Although E(Â1) = E(Â2) = A, Â 2 is better because var(â1) = σ 2 > var(â2) = σ2 N. (3) Man-Wai MAK (EIE) Estimation Theory November 8, / 28
17 Minimum Variance Unbiased Estimators If the bias is zero, the MSE is just the variance. This gives rise to the minimum variance unbiased estimator (MVUE) for θ. Definition: An estimator ˆθ is the MVUE if it is unbiased and has the smallest variance among any unbiased estimator for all values of the unknown parameter θ, i.e., ˆθ, we have var(ˆθ MVUE ) var(ˆθ) It is important to note that a uniformly MVUE may not always exist, and even if it does, we may not be able to find it. This definition and the above discussions apply to vector parameters θ. Man-Wai MAK (EIE) Estimation Theory November 8, / 28
18 Cramer-Rao Lower Bound Given an estimation problem, what is the variance of the best possible estimator? This quantity is given by the Cramer-Rao lower bound (CRLB). The CRLB theorem also provides a method for finding the best estimator. We will use the example in Eq. 1 to explain the idea of CRLB. For simplicity, suppose that we are using only a single observation x[0] to estimate A in Eq. 1. The PDF of x[0] is [ 1 p(x[0]; A) = exp 1 ] (x[0] A)2 2πσ 2 2σ2 Once we have observed x[0], say for example x[0] = 3, some values of A are more likely than others. Man-Wai MAK (EIE) Estimation Theory November 8, / 28
19 Cramer-Rao Lower Bound The pdf of A has the same form as the PDF of x[0]: [ 1 pdf of A = exp 1 ] (3 A)2 2πσ 2 2σ2 0.7 PDF plotted versus A in the case x[0]=3 and σ 2 = 1/ Value of A x[0] 0.7 PDF plotted versus A in the case x[0]=3 and σ 2 = Value of A x[0] We probably will get more accurate estimate of A for the distribution of A in the upper pannel. Man-Wai MAK (EIE) Estimation Theory November 8, / 28
20 Cramer-Rao Lower Bound If the pdf is viewed as a function of the unknown parameter (with x fixed), it is called the likelihood function (function of the unknown parameters). The sharpness of the likelihood function determines how accurately we can estimate the parameter. For example the pdf on the top panel of previous figure is the easier case. If a single sample is observed as x[0] = A + w[0], then we can expect a better estimate if σ 2 is small. How to quantify the sharpness? Is there any measure that would be common for all possible estimators for a specific estimation problem? The second derivative of the likelihood function (or log-likelihood) is one alternative for measuring the sharpness of a function. Man-Wai MAK (EIE) Estimation Theory November 8, / 28
21 Cramer-Rao Lower Bound Recall that the PDF of x[0] is [ 1 p(x[0]; A) = exp 1 ] (x[0] A)2 2πσ 2 2σ2 The log-likelihood function: p(x[0]; A) = log 2πσ 2 1 (x[0] A)2 2σ2 The first and second derivatives w.r.t. A are: log p(x[0]; A) = 1 (x[0] A) A σ2 2 log p(x[0]; A) A 2 = 1 σ 2 Man-Wai MAK (EIE) Estimation Theory November 8, / 28
22 Cramer-Rao Lower Bound Since σ 2 is the smallest possible variance, in this specific case we have an alternative way of finding the minimum variance of all estimators: minimum variance = σ 2 = 1 2 log p(x[0];a) A 2 For the general case, If the function depends on the data x, take the expectation over all x. If the function depends on the parameter θ, evaluate the derivative at the true value of θ, We have the general rule: min. var. of any unbiased estimator = 1 [ ], E 2 log p(x,θ) θ 2 Man-Wai MAK (EIE) Estimation Theory November 8, / 28
23 CRLB Theorem If the pdf p(x; θ) satisfies the regularity condition { } log p(x; θ) E = 0, θ, θ where the expectation is taken with respect to p(x; θ). Then, the variance of any unbiased estimator ˆθ must satisfy 1 var(ˆθ) [ ], E 2 log p(x,θ) θ 2 where the derivative is evalauted at the true value of θ and the expectation is taken with respect to p(x; θ). An unbiased estimator attains the bound for all θ if and only if log p(x; θ) θ = I(θ)(g(x) θ) (4) for some functions g and I. The MVU estimator is ˆθ = g(x) and the minimum variance = 1/I(θ). Man-Wai MAK (EIE) Estimation Theory November 8, / 28
24 CRLB Example 1: Estimation of DC Level in WGN Example: DC level in Gaussian noise: x[n] = A + w[n], n = 0, 1,..., N 1 What is the minimum variance of any unbiased estimator using N samples? The likelihood function is the product of N densities: p(x; A) = = N 1 n=0 [ 1 exp 2πσ 2 1 (2πσ 2 exp ) N/2 1 ] (x[n] A)2 2σ2 [ ] 1 N 1 2σ 2 (x[n] A) 2 n=0 Man-Wai MAK (EIE) Estimation Theory November 8, / 28
25 CRLB Example 1: Estimation of DC Level in WGN The log-likelihood function is [ ] log p(x; A) = log (2πσ 2 ) N 2 1 N 1 2σ 2 (x[n] A) 2 The first derivative is log p(x, A) A where x is the sample mean. The 2nd derivative is n=0 = 1 N 1 σ 2 (x[n] A) = N ( x A), σ2 n=0 2 log p(x, A) A 2 = N σ 2 Therefore, the minimum variance of any unbiased estimator is σ2 var(â) N (5) Man-Wai MAK (EIE) Estimation Theory November 8, / 28
26 CRLB Example 1: Estimation of DC Level in WGN In Eq. 3, the sample mean A 2 gives the minimum variance in Eq. 5. Therefore, the sample mean is an MVUE for this problem. Using the CRLB theorem in Eq. 4, we have log p(x; A) A = N ( x A) σ2 = I(A)(g(x) A), where I(A) = N and g(x) = x. σ 2 So, g(x) is the MVUE and 1/I(A) is its variance. Man-Wai MAK (EIE) Estimation Theory November 8, / 28
27 CRLB Example 2: Estimation of the Phase of Sinusoid Problem: Given N data points and a model with known amplitude A and frequency f 0 : x[n] = A cos(2πf 0 n + φ) + w[n], n = 0,..., N 1 estimate the phase φ of a sinusoid embedded in WGN with variance σ 2. How accurate will the estimator be? Solution: Derivation: See Tutorial var( ˆφ) 2σ 2 }{{} NA 2 approx Man-Wai MAK (EIE) Estimation Theory November 8, / 28
28 CRLB Summary Cramer Rao inequality provides lower bound for the estimation error variance. Minimum attainable variance is often larger than CRLB. We need to know the pdf to evaluate CRLB. Often we dont know this information and cannot evaluate this bound. If the data is multivariate Gaussian or i.i.d. with known distribution, we can evaluate it. Its not guaranteed that MVUE exists or is realizable. Man-Wai MAK (EIE) Estimation Theory November 8, / 28
EIE6207: Maximum-Likelihood and Bayesian Estimation
EIE6207: Maximum-Likelihood and Bayesian Estimation Man-Wai MAK Dept. of Electronic and Information Engineering, The Hong Kong Polytechnic University enmwmak@polyu.edu.hk http://www.eie.polyu.edu.hk/ mwmak
More informationA Few Notes on Fisher Information (WIP)
A Few Notes on Fisher Information (WIP) David Meyer dmm@{-4-5.net,uoregon.edu} Last update: April 30, 208 Definitions There are so many interesting things about Fisher Information and its theoretical properties
More informationAdvanced Signal Processing Introduction to Estimation Theory
Advanced Signal Processing Introduction to Estimation Theory Danilo Mandic, room 813, ext: 46271 Department of Electrical and Electronic Engineering Imperial College London, UK d.mandic@imperial.ac.uk,
More informationELEG 5633 Detection and Estimation Minimum Variance Unbiased Estimators (MVUE)
1 ELEG 5633 Detection and Estimation Minimum Variance Unbiased Estimators (MVUE) Jingxian Wu Department of Electrical Engineering University of Arkansas Outline Minimum Variance Unbiased Estimators (MVUE)
More informationDetection & Estimation Lecture 1
Detection & Estimation Lecture 1 Intro, MVUE, CRLB Xiliang Luo General Course Information Textbooks & References Fundamentals of Statistical Signal Processing: Estimation Theory/Detection Theory, Steven
More informationDetection & Estimation Lecture 1
Detection & Estimation Lecture 1 Intro, MVUE, CRLB Xiliang Luo General Course Information Textbooks & References Fundamentals of Statistical Signal Processing: Estimation Theory/Detection Theory, Steven
More informationRowan University Department of Electrical and Computer Engineering
Rowan University Department of Electrical and Computer Engineering Estimation and Detection Theory Fall 2013 to Practice Exam II This is a closed book exam. There are 8 problems in the exam. The problems
More informationModule 2. Random Processes. Version 2, ECE IIT, Kharagpur
Module Random Processes Version, ECE IIT, Kharagpur Lesson 9 Introduction to Statistical Signal Processing Version, ECE IIT, Kharagpur After reading this lesson, you will learn about Hypotheses testing
More informationPATTERN RECOGNITION AND MACHINE LEARNING
PATTERN RECOGNITION AND MACHINE LEARNING Slide Set 2: Estimation Theory January 2018 Heikki Huttunen heikki.huttunen@tut.fi Department of Signal Processing Tampere University of Technology Classical Estimation
More informationEstimation and Detection
stimation and Detection Lecture 2: Cramér-Rao Lower Bound Dr. ir. Richard C. Hendriks & Dr. Sundeep P. Chepuri 7//207 Remember: Introductory xample Given a process (DC in noise): x[n]=a + w[n], n=0,,n,
More informationChapter 8: Least squares (beginning of chapter)
Chapter 8: Least squares (beginning of chapter) Least Squares So far, we have been trying to determine an estimator which was unbiased and had minimum variance. Next we ll consider a class of estimators
More informationLinear models. x = Hθ + w, where w N(0, σ 2 I) and H R n p. The matrix H is called the observation matrix or design matrix 1.
Linear models As the first approach to estimator design, we consider the class of problems that can be represented by a linear model. In general, finding the MVUE is difficult. But if the linear model
More informationAdvanced Signal Processing Minimum Variance Unbiased Estimation (MVU)
Advanced Signal Processing Minimum Variance Unbiased Estimation (MVU) Danilo Mandic room 813, ext: 46271 Department of Electrical and Electronic Engineering Imperial College London, UK d.mandic@imperial.ac.uk,
More information10-704: Information Processing and Learning Fall Lecture 24: Dec 7
0-704: Information Processing and Learning Fall 206 Lecturer: Aarti Singh Lecture 24: Dec 7 Note: These notes are based on scribed notes from Spring5 offering of this course. LaTeX template courtesy of
More informationModule 1 - Signal estimation
, Arraial do Cabo, 2009 Module 1 - Signal estimation Sérgio M. Jesus (sjesus@ualg.pt) Universidade do Algarve, PT-8005-139 Faro, Portugal www.siplab.fct.ualg.pt February 2009 Outline of Module 1 Parameter
More informationDetection and Estimation Theory
Detection and Estimation Theory Instructor: Prof. Namrata Vaswani Dept. of Electrical and Computer Engineering Iowa State University http://www.ece.iastate.edu/ namrata Slide 1 What is Estimation and Detection
More informationECE531 Lecture 10b: Maximum Likelihood Estimation
ECE531 Lecture 10b: Maximum Likelihood Estimation D. Richard Brown III Worcester Polytechnic Institute 05-Apr-2011 Worcester Polytechnic Institute D. Richard Brown III 05-Apr-2011 1 / 23 Introduction So
More informationTerminology Suppose we have N observations {x(n)} N 1. Estimators as Random Variables. {x(n)} N 1
Estimation Theory Overview Properties Bias, Variance, and Mean Square Error Cramér-Rao lower bound Maximum likelihood Consistency Confidence intervals Properties of the mean estimator Properties of the
More informationMathematical statistics
October 4 th, 2018 Lecture 12: Information Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation Chapter
More informationDetection theory. H 0 : x[n] = w[n]
Detection Theory Detection theory A the last topic of the course, we will briefly consider detection theory. The methods are based on estimation theory and attempt to answer questions such as Is a signal
More informationClassical Estimation Topics
Classical Estimation Topics Namrata Vaswani, Iowa State University February 25, 2014 This note fills in the gaps in the notes already provided (l0.pdf, l1.pdf, l2.pdf, l3.pdf, LeastSquares.pdf). 1 Min
More informationDetection theory 101 ELEC-E5410 Signal Processing for Communications
Detection theory 101 ELEC-E5410 Signal Processing for Communications Binary hypothesis testing Null hypothesis H 0 : e.g. noise only Alternative hypothesis H 1 : signal + noise p(x;h 0 ) γ p(x;h 1 ) Trade-off
More informationRegression Estimation - Least Squares and Maximum Likelihood. Dr. Frank Wood
Regression Estimation - Least Squares and Maximum Likelihood Dr. Frank Wood Least Squares Max(min)imization Function to minimize w.r.t. β 0, β 1 Q = n (Y i (β 0 + β 1 X i )) 2 i=1 Minimize this by maximizing
More informationEstimation Theory Fredrik Rusek. Chapters
Estimation Theory Fredrik Rusek Chapters 3.5-3.10 Recap We deal with unbiased estimators of deterministic parameters Performance of an estimator is measured by the variance of the estimate (due to the
More informationECE534, Spring 2018: Solutions for Problem Set #3
ECE534, Spring 08: Solutions for Problem Set #3 Jointly Gaussian Random Variables and MMSE Estimation Suppose that X, Y are jointly Gaussian random variables with µ X = µ Y = 0 and σ X = σ Y = Let their
More informationLINEAR MODELS IN STATISTICAL SIGNAL PROCESSING
TERM PAPER REPORT ON LINEAR MODELS IN STATISTICAL SIGNAL PROCESSING ANIMA MISHRA SHARMA, Y8104001 BISHWAJIT SHARMA, Y8104015 Course Instructor DR. RAJESH HEDGE Electrical Engineering Department IIT Kanpur
More informationChapters 9. Properties of Point Estimators
Chapters 9. Properties of Point Estimators Recap Target parameter, or population parameter θ. Population distribution f(x; θ). { probability function, discrete case f(x; θ) = density, continuous case The
More information10. Linear Models and Maximum Likelihood Estimation
10. Linear Models and Maximum Likelihood Estimation ECE 830, Spring 2017 Rebecca Willett 1 / 34 Primary Goal General problem statement: We observe y i iid pθ, θ Θ and the goal is to determine the θ that
More informationChapter 8.8.1: A factorization theorem
LECTURE 14 Chapter 8.8.1: A factorization theorem The characterization of a sufficient statistic in terms of the conditional distribution of the data given the statistic can be difficult to work with.
More informationSTAT 730 Chapter 4: Estimation
STAT 730 Chapter 4: Estimation Timothy Hanson Department of Statistics, University of South Carolina Stat 730: Multivariate Analysis 1 / 23 The likelihood We have iid data, at least initially. Each datum
More informationIntroduction to Simple Linear Regression
Introduction to Simple Linear Regression Yang Feng http://www.stat.columbia.edu/~yangfeng Yang Feng (Columbia University) Introduction to Simple Linear Regression 1 / 68 About me Faculty in the Department
More informationMaximum Likelihood Estimation
Connexions module: m11446 1 Maximum Likelihood Estimation Clayton Scott Robert Nowak This work is produced by The Connexions Project and licensed under the Creative Commons Attribution License Abstract
More information13. Parameter Estimation. ECE 830, Spring 2014
13. Parameter Estimation ECE 830, Spring 2014 1 / 18 Primary Goal General problem statement: We observe X p(x θ), θ Θ and the goal is to determine the θ that produced X. Given a collection of observations
More informationEstimation MLE-Pandemic data MLE-Financial crisis data Evaluating estimators. Estimation. September 24, STAT 151 Class 6 Slide 1
Estimation September 24, 2018 STAT 151 Class 6 Slide 1 Pandemic data Treatment outcome, X, from n = 100 patients in a pandemic: 1 = recovered and 0 = not recovered 1 1 1 0 0 0 1 1 1 0 0 1 0 1 0 0 1 1 1
More informationECE 275A Homework 6 Solutions
ECE 275A Homework 6 Solutions. The notation used in the solutions for the concentration (hyper) ellipsoid problems is defined in the lecture supplement on concentration ellipsoids. Note that θ T Σ θ =
More informationECE 275A Homework 7 Solutions
ECE 275A Homework 7 Solutions Solutions 1. For the same specification as in Homework Problem 6.11 we want to determine an estimator for θ using the Method of Moments (MOM). In general, the MOM estimator
More informationVariations. ECE 6540, Lecture 10 Maximum Likelihood Estimation
Variations ECE 6540, Lecture 10 Last Time BLUE (Best Linear Unbiased Estimator) Formulation Advantages Disadvantages 2 The BLUE A simplification Assume the estimator is a linear system For a single parameter
More informationDETECTION theory deals primarily with techniques for
ADVANCED SIGNAL PROCESSING SE Optimum Detection of Deterministic and Random Signals Stefan Tertinek Graz University of Technology turtle@sbox.tugraz.at Abstract This paper introduces various methods for
More informationUnbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.
Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it
More informationFAST AND ACCURATE DIRECTION-OF-ARRIVAL ESTIMATION FOR A SINGLE SOURCE
Progress In Electromagnetics Research C, Vol. 6, 13 20, 2009 FAST AND ACCURATE DIRECTION-OF-ARRIVAL ESTIMATION FOR A SINGLE SOURCE Y. Wu School of Computer Science and Engineering Wuhan Institute of Technology
More informationMathematical statistics
October 18 th, 2018 Lecture 16: Midterm review Countdown to mid-term exam: 7 days Week 1 Chapter 1: Probability review Week 2 Week 4 Week 7 Chapter 6: Statistics Chapter 7: Point Estimation Chapter 8:
More informationLecture 8: Information Theory and Statistics
Lecture 8: Information Theory and Statistics Part II: Hypothesis Testing and I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 23, 2015 1 / 50 I-Hsiang
More informationEstimation theory. Parametric estimation. Properties of estimators. Minimum variance estimator. Cramer-Rao bound. Maximum likelihood estimators
Estimation theory Parametric estimation Properties of estimators Minimum variance estimator Cramer-Rao bound Maximum likelihood estimators Confidence intervals Bayesian estimation 1 Random Variables Let
More informationStatistics and Econometrics I
Statistics and Econometrics I Point Estimation Shiu-Sheng Chen Department of Economics National Taiwan University September 13, 2016 Shiu-Sheng Chen (NTU Econ) Statistics and Econometrics I September 13,
More informationMethods of evaluating estimators and best unbiased estimators Hamid R. Rabiee
Stochastic Processes Methods of evaluating estimators and best unbiased estimators Hamid R. Rabiee 1 Outline Methods of Mean Squared Error Bias and Unbiasedness Best Unbiased Estimators CR-Bound for variance
More informationElements of statistics (MATH0487-1)
Elements of statistics (MATH0487-1) Prof. Dr. Dr. K. Van Steen University of Liège, Belgium November 12, 2012 Introduction to Statistics Basic Probability Revisited Sampling Exploratory Data Analysis -
More information6.1 Variational representation of f-divergences
ECE598: Information-theoretic methods in high-dimensional statistics Spring 2016 Lecture 6: Variational representation, HCR and CR lower bounds Lecturer: Yihong Wu Scribe: Georgios Rovatsos, Feb 11, 2016
More informationGraduate Econometrics I: Unbiased Estimation
Graduate Econometrics I: Unbiased Estimation Yves Dominicy Université libre de Bruxelles Solvay Brussels School of Economics and Management ECARES Yves Dominicy Graduate Econometrics I: Unbiased Estimation
More informationBrief Review on Estimation Theory
Brief Review on Estimation Theory K. Abed-Meraim ENST PARIS, Signal and Image Processing Dept. abed@tsi.enst.fr This presentation is essentially based on the course BASTA by E. Moulines Brief review on
More information10. Composite Hypothesis Testing. ECE 830, Spring 2014
10. Composite Hypothesis Testing ECE 830, Spring 2014 1 / 25 In many real world problems, it is difficult to precisely specify probability distributions. Our models for data may involve unknown parameters
More informationMath 494: Mathematical Statistics
Math 494: Mathematical Statistics Instructor: Jimin Ding jmding@wustl.edu Department of Mathematics Washington University in St. Louis Class materials are available on course website (www.math.wustl.edu/
More informationEstimation, Detection, and Identification CMU 18752
Estimation, Detection, and Identification CMU 18752 Graduate Course on the CMU/Portugal ECE PhD Program Spring 2008/2009 Instructor: Prof. Paulo Jorge Oliveira pjcro @ isr.ist.utl.pt Phone: +351 21 8418053
More informationPATTERN RECOGNITION AND MACHINE LEARNING
PATTERN RECOGNITION AND MACHINE LEARNING Slide Set 3: Detection Theory January 2018 Heikki Huttunen heikki.huttunen@tut.fi Department of Signal Processing Tampere University of Technology Detection theory
More informationRegression Estimation Least Squares and Maximum Likelihood
Regression Estimation Least Squares and Maximum Likelihood Dr. Frank Wood Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 3, Slide 1 Least Squares Max(min)imization Function to minimize
More informationF & B Approaches to a simple model
A6523 Signal Modeling, Statistical Inference and Data Mining in Astrophysics Spring 215 http://www.astro.cornell.edu/~cordes/a6523 Lecture 11 Applications: Model comparison Challenges in large-scale surveys
More informationUnbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.
Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it
More informationCramer-Rao Lower Bound Computation Via the Characteristic Function
Cramer-Rao ower Bound Computation Via the Characteristic Function Steven Kay, Fellow, IEEE, and Cuichun Xu Abstract The Cramer-Rao ower Bound is widely used in statistical signal processing as a benchmark
More informationEstimators as Random Variables
Estimation Theory Overview Properties Bias, Variance, and Mean Square Error Cramér-Rao lower bound Maimum likelihood Consistency Confidence intervals Properties of the mean estimator Introduction Up until
More informationSGN Advanced Signal Processing: Lecture 8 Parameter estimation for AR and MA models. Model order selection
SG 21006 Advanced Signal Processing: Lecture 8 Parameter estimation for AR and MA models. Model order selection Ioan Tabus Department of Signal Processing Tampere University of Technology Finland 1 / 28
More informationParametric Techniques Lecture 3
Parametric Techniques Lecture 3 Jason Corso SUNY at Buffalo 22 January 2009 J. Corso (SUNY at Buffalo) Parametric Techniques Lecture 3 22 January 2009 1 / 39 Introduction In Lecture 2, we learned how to
More informationEstimation, Detection, and Identification
Estimation, Detection, and Identification Graduate Course on the CMU/Portugal ECE PhD Program Spring 2008/2009 Chapter 5 Best Linear Unbiased Estimators Instructor: Prof. Paulo Jorge Oliveira pjcro @ isr.ist.utl.pt
More informationFundamentals of Statistical Signal Processing Volume II Detection Theory
Fundamentals of Statistical Signal Processing Volume II Detection Theory Steven M. Kay University of Rhode Island PH PTR Prentice Hall PTR Upper Saddle River, New Jersey 07458 http://www.phptr.com Contents
More informationReview. December 4 th, Review
December 4 th, 2017 Att. Final exam: Course evaluation Friday, 12/14/2018, 10:30am 12:30pm Gore Hall 115 Overview Week 2 Week 4 Week 7 Week 10 Week 12 Chapter 6: Statistics and Sampling Distributions Chapter
More informationDensity Estimation. Seungjin Choi
Density Estimation Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr http://mlg.postech.ac.kr/
More informationParametric Techniques
Parametric Techniques Jason J. Corso SUNY at Buffalo J. Corso (SUNY at Buffalo) Parametric Techniques 1 / 39 Introduction When covering Bayesian Decision Theory, we assumed the full probabilistic structure
More informationA General Overview of Parametric Estimation and Inference Techniques.
A General Overview of Parametric Estimation and Inference Techniques. Moulinath Banerjee University of Michigan September 11, 2012 The object of statistical inference is to glean information about an underlying
More informationStatistics: Learning models from data
DS-GA 1002 Lecture notes 5 October 19, 2015 Statistics: Learning models from data Learning models from data that are assumed to be generated probabilistically from a certain unknown distribution is a crucial
More informationBIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation
BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation Yujin Chung November 29th, 2016 Fall 2016 Yujin Chung Lec13: MLE Fall 2016 1/24 Previous Parametric tests Mean comparisons (normality assumption)
More informationMS&E 226: Small Data. Lecture 11: Maximum likelihood (v2) Ramesh Johari
MS&E 226: Small Data Lecture 11: Maximum likelihood (v2) Ramesh Johari ramesh.johari@stanford.edu 1 / 18 The likelihood function 2 / 18 Estimating the parameter This lecture develops the methodology behind
More informationRethinking Biased Estimation: Improving Maximum Likelihood and the Cramér Rao Bound
Foundations and Trends R in Signal Processing Vol. 1, No. 4 (2007) 305 449 c 2008 Y. C. Eldar DOI: 10.1561/2000000008 Rethinking Biased Estimation: Improving Maximum Likelihood and the Cramér Rao Bound
More informationEconometrics I, Estimation
Econometrics I, Estimation Department of Economics Stanford University September, 2008 Part I Parameter, Estimator, Estimate A parametric is a feature of the population. An estimator is a function of the
More informationSystem Identification, Lecture 4
System Identification, Lecture 4 Kristiaan Pelckmans (IT/UU, 2338) Course code: 1RT880, Report code: 61800 - Spring 2012 F, FRI Uppsala University, Information Technology 30 Januari 2012 SI-2012 K. Pelckmans
More informationSystem Identification, Lecture 4
System Identification, Lecture 4 Kristiaan Pelckmans (IT/UU, 2338) Course code: 1RT880, Report code: 61800 - Spring 2016 F, FRI Uppsala University, Information Technology 13 April 2016 SI-2016 K. Pelckmans
More informationTechnique for Numerical Computation of Cramér-Rao Bound using MATLAB
Technique for Numerical Computation of Cramér-Rao Bound using MATLAB Hing Cheung So http://www.ee.cityu.edu.hk/~hcso Department of Electronic Engineering City University of Hong Kong H. C. So Page 1 The
More informationPart IB Statistics. Theorems with proof. Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua. Lent 2015
Part IB Statistics Theorems with proof Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly)
More informationEstimation Theory. as Θ = (Θ 1,Θ 2,...,Θ m ) T. An estimator
Estimation Theory Estimation theory deals with finding numerical values of interesting parameters from given set of data. We start with formulating a family of models that could describe how the data were
More informationMachine Learning 4771
Machine Learning 4771 Instructor: Tony Jebara Topic 11 Maximum Likelihood as Bayesian Inference Maximum A Posteriori Bayesian Gaussian Estimation Why Maximum Likelihood? So far, assumed max (log) likelihood
More informationGaussian, Markov and stationary processes
Gaussian, Markov and stationary processes Gonzalo Mateos Dept. of ECE and Goergen Institute for Data Science University of Rochester gmateosb@ece.rochester.edu http://www.ece.rochester.edu/~gmateosb/ November
More informationIntroduction to Estimation Methods for Time Series models Lecture 2
Introduction to Estimation Methods for Time Series models Lecture 2 Fulvio Corsi SNS Pisa Fulvio Corsi Introduction to Estimation () Methods for Time Series models Lecture 2 SNS Pisa 1 / 21 Estimators:
More informationLecture 7 Introduction to Statistical Decision Theory
Lecture 7 Introduction to Statistical Decision Theory I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 20, 2016 1 / 55 I-Hsiang Wang IT Lecture 7
More informationEconomics 520. Lecture Note 19: Hypothesis Testing via the Neyman-Pearson Lemma CB 8.1,
Economics 520 Lecture Note 9: Hypothesis Testing via the Neyman-Pearson Lemma CB 8., 8.3.-8.3.3 Uniformly Most Powerful Tests and the Neyman-Pearson Lemma Let s return to the hypothesis testing problem
More informationParametric Models. Dr. Shuang LIANG. School of Software Engineering TongJi University Fall, 2012
Parametric Models Dr. Shuang LIANG School of Software Engineering TongJi University Fall, 2012 Today s Topics Maximum Likelihood Estimation Bayesian Density Estimation Today s Topics Maximum Likelihood
More informationHypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3
Hypothesis Testing CB: chapter 8; section 0.3 Hypothesis: statement about an unknown population parameter Examples: The average age of males in Sweden is 7. (statement about population mean) The lowest
More informationMA 575 Linear Models: Cedric E. Ginestet, Boston University Midterm Review Week 7
MA 575 Linear Models: Cedric E. Ginestet, Boston University Midterm Review Week 7 1 Random Vectors Let a 0 and y be n 1 vectors, and let A be an n n matrix. Here, a 0 and A are non-random, whereas y is
More informationChapter 8: Estimation 1
Chapter 8: Estimation 1 Jae-Kwang Kim Iowa State University Fall, 2014 Kim (ISU) Ch. 8: Estimation 1 Fall, 2014 1 / 33 Introduction 1 Introduction 2 Ratio estimation 3 Regression estimator Kim (ISU) Ch.
More informationMAS223 Statistical Inference and Modelling Exercises
MAS223 Statistical Inference and Modelling Exercises The exercises are grouped into sections, corresponding to chapters of the lecture notes Within each section exercises are divided into warm-up questions,
More informationDS-GA 1002 Lecture notes 11 Fall Bayesian statistics
DS-GA 100 Lecture notes 11 Fall 016 Bayesian statistics In the frequentist paradigm we model the data as realizations from a distribution that depends on deterministic parameters. In contrast, in Bayesian
More informationCOS513 LECTURE 8 STATISTICAL CONCEPTS
COS513 LECTURE 8 STATISTICAL CONCEPTS NIKOLAI SLAVOV AND ANKUR PARIKH 1. MAKING MEANINGFUL STATEMENTS FROM JOINT PROBABILITY DISTRIBUTIONS. A graphical model (GM) represents a family of probability distributions
More informationParameter Estimation
Parameter Estimation Consider a sample of observations on a random variable Y. his generates random variables: (y 1, y 2,, y ). A random sample is a sample (y 1, y 2,, y ) where the random variables y
More informationFractional Hot Deck Imputation for Robust Inference Under Item Nonresponse in Survey Sampling
Fractional Hot Deck Imputation for Robust Inference Under Item Nonresponse in Survey Sampling Jae-Kwang Kim 1 Iowa State University June 26, 2013 1 Joint work with Shu Yang Introduction 1 Introduction
More informationECE531 Lecture 8: Non-Random Parameter Estimation
ECE531 Lecture 8: Non-Random Parameter Estimation D. Richard Brown III Worcester Polytechnic Institute 19-March-2009 Worcester Polytechnic Institute D. Richard Brown III 19-March-2009 1 / 25 Introduction
More informationComposite Hypotheses and Generalized Likelihood Ratio Tests
Composite Hypotheses and Generalized Likelihood Ratio Tests Rebecca Willett, 06 In many real world problems, it is difficult to precisely specify probability distributions. Our models for data may involve
More informationSTAT 512 sp 2018 Summary Sheet
STAT 5 sp 08 Summary Sheet Karl B. Gregory Spring 08. Transformations of a random variable Let X be a rv with support X and let g be a function mapping X to Y with inverse mapping g (A = {x X : g(x A}
More informationParameter Estimation
1 / 44 Parameter Estimation Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay October 25, 2012 Motivation System Model used to Derive
More informationParametric Inference
Parametric Inference Moulinath Banerjee University of Michigan April 14, 2004 1 General Discussion The object of statistical inference is to glean information about an underlying population based on a
More informationLECTURE NOTE #3 PROF. ALAN YUILLE
LECTURE NOTE #3 PROF. ALAN YUILLE 1. Three Topics (1) Precision and Recall Curves. Receiver Operating Characteristic Curves (ROC). What to do if we do not fix the loss function? (2) The Curse of Dimensionality.
More informationProof In the CR proof. and
Question Under what conditions will we be able to attain the Cramér-Rao bound and find a MVUE? Lecture 4 - Consequences of the Cramér-Rao Lower Bound. Searching for a MVUE. Rao-Blackwell Theorem, Lehmann-Scheffé
More informationCentral Limit Theorem ( 5.3)
Central Limit Theorem ( 5.3) Let X 1, X 2,... be a sequence of independent random variables, each having n mean µ and variance σ 2. Then the distribution of the partial sum S n = X i i=1 becomes approximately
More informationEstimation Theory for Engineers
Estimation Theory for Engineers Roberto Togneri 30th August 2005 Applications Modern estimation theory can be found at the heart of many electronic signal processing systems designed to extract information.
More information2 Statistical Estimation: Basic Concepts
Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof. N. Shimkin 2 Statistical Estimation:
More information