Estimation, Detection, and Identification

Similar documents
Estimation, Detection, and Identification CMU 18752

Variations. ECE 6540, Lecture 10 Maximum Likelihood Estimation

Rowan University Department of Electrical and Computer Engineering

Estimation Theory Fredrik Rusek. Chapters 6-7

Detection and Estimation Theory

EIE6207: Estimation Theory

Estimation and Detection

Link lecture - Lagrange Multipliers

A TERM PAPER REPORT ON KALMAN FILTER

Terminology Suppose we have N observations {x(n)} N 1. Estimators as Random Variables. {x(n)} N 1

Chapter 8: Least squares (beginning of chapter)

Detection & Estimation Lecture 1

UCSD ECE153 Handout #40 Prof. Young-Han Kim Thursday, May 29, Homework Set #8 Due: Thursday, June 5, 2011

Advanced Signal Processing Introduction to Estimation Theory

The Newton-Raphson Algorithm

Detection & Estimation Lecture 1

ELEG 5633 Detection and Estimation Minimum Variance Unbiased Estimators (MVUE)

F & B Approaches to a simple model

System Identification, Lecture 4

System Identification, Lecture 4

Properties of the least squares estimates

ECE 275A Homework 6 Solutions

To Estimate or Not to Estimate?

1. The Multivariate Classical Linear Regression Model

A6523 Modeling, Inference, and Mining Jim Cordes, Cornell University

Regression Models - Introduction

The Kalman filter is arguably one of the most notable algorithms

Introduction to Estimation Methods for Time Series models. Lecture 1

Cramér-Rao Bounds for Estimation of Linear System Noise Covariances

Lessons in Estimation Theory for Signal Processing, Communications, and Control

Chapter 4: Constrained estimators and tests in the multiple linear regression model (Part III)

26. Filtering. ECE 830, Spring 2014

Probing the covariance matrix

EECE Adaptive Control

Advanced Signal Processing Minimum Variance Unbiased Estimation (MVU)

STAT 100C: Linear models

FUNDAMENTAL FILTERING LIMITATIONS IN LINEAR NON-GAUSSIAN SYSTEMS

MS&E 226: Small Data. Lecture 11: Maximum likelihood (v2) Ramesh Johari

Gaussian, Markov and stationary processes

Course content (will be adapted to the background knowledge of the class):

A Few Notes on Fisher Information (WIP)

VAR Model. (k-variate) VAR(p) model (in the Reduced Form): Y t-2. Y t-1 = A + B 1. Y t + B 2. Y t-p. + ε t. + + B p. where:

Particle Filters. Outline

SIMON FRASER UNIVERSITY School of Engineering Science

Optimal input design for nonlinear dynamical systems: a graph-theory approach

Linear models. x = Hθ + w, where w N(0, σ 2 I) and H R n p. The matrix H is called the observation matrix or design matrix 1.

Bayesian Methods in Positioning Applications

Estimation Theory Fredrik Rusek. Chapters

Outline Lecture 2 2(32)

Econometrics A. Simple linear model (2) Keio University, Faculty of Economics. Simon Clinet (Keio University) Econometrics A October 16, / 11

SELECTIVE ANGLE MEASUREMENTS FOR A 3D-AOA INSTRUMENTAL VARIABLE TMA ALGORITHM

Multivariate Regression

INDIAN INSTITUTE OF SCIENCE STOCHASTIC HYDROLOGY. Lecture -33 Course Instructor : Prof. P. P. MUJUMDAR Department of Civil Engg., IISc.

ELEG 5633 Detection and Estimation Signal Detection: Deterministic Signals

A NEW INFORMATION THEORETIC APPROACH TO ORDER ESTIMATION PROBLEM. Massachusetts Institute of Technology, Cambridge, MA 02139, U.S.A.

ECE531 Lecture 10b: Maximum Likelihood Estimation

Linear models. Linear models are computationally convenient and remain widely used in. applied econometric research

Estimation theory. Parametric estimation. Properties of estimators. Minimum variance estimator. Cramer-Rao bound. Maximum likelihood estimators

LINEAR MODELS IN STATISTICAL SIGNAL PROCESSING

Association studies and regression

Regression #4: Properties of OLS Estimator (Part 2)

DETECTION theory deals primarily with techniques for

Regression Models - Introduction

ME 597: AUTONOMOUS MOBILE ROBOTICS SECTION 2 PROBABILITY. Prof. Steven Waslander

Maximum Likelihood Tests and Quasi-Maximum-Likelihood

3. ESTIMATION OF SIGNALS USING A LEAST SQUARES TECHNIQUE

ADAPTIVE FILTER THEORY

Prof. Dr.-Ing. Armin Dekorsy Department of Communications Engineering. Stochastic Processes and Linear Algebra Recap Slides

OPTIMAL ESTIMATION of DYNAMIC SYSTEMS

1 Appendix A: Matrix Algebra

Estimators as Random Variables

MMSE DECODING FOR ANALOG JOINT SOURCE CHANNEL CODING USING MONTE CARLO IMPORTANCE SAMPLING

Introduction to Maximum Likelihood Estimation

Least squares problems

Brief Review on Estimation Theory

Econometrics II - EXAM Answer each question in separate sheets in three hours

ECE 516: System Control Engineering

ECE531 Lecture 8: Non-Random Parameter Estimation

Modelling and System Identification Microexam 1

New Introduction to Multiple Time Series Analysis

1 Cricket chirps: an example

Fundamentals of Statistical Signal Processing Volume II Detection Theory

12. Prediction Error Methods (PEM)

A Gaussian state-space model for wind fields in the North-East Atlantic

LECTURE 2 LINEAR REGRESSION MODEL AND OLS

Dynamic System Identification using HDMR-Bayesian Technique

LTI Systems, Additive Noise, and Order Estimation

Stochastic Processes

Business Economics BUSINESS ECONOMICS. PAPER No. : 8, FUNDAMENTALS OF ECONOMETRICS MODULE No. : 3, GAUSS MARKOV THEOREM

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8

PROBABILITY AND STOCHASTIC PROCESSES A Friendly Introduction for Electrical and Computer Engineers

EIE6207: Maximum-Likelihood and Bayesian Estimation

Time Series Analysis

On Identification of Cascade Systems 1

Estimation theory and information geometry based on denoising

UCSD ECE 153 Handout #46 Prof. Young-Han Kim Thursday, June 5, Solutions to Homework Set #8 (Prepared by TA Fatemeh Arbabjolfaei)

13. Parameter Estimation. ECE 830, Spring 2014

ECE 5615/4615 Computer Project

Expectation. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda

Module 1 - Signal estimation

Transcription:

Estimation, Detection, and Identification Graduate Course on the CMU/Portugal ECE PhD Program Spring 2008/2009 Chapter 5 Best Linear Unbiased Estimators Instructor: Prof. Paulo Jorge Oliveira pjcro @ isr.ist.utl.pt Phone: +351 21 8418053 ou 2053 (inside IST)

Syllabus: Classical Estimation Theory Chap. 4 - Linear Models in the Presence of Stochastic Signals [1 week] Stationary and transient analysis; White Gaussian noise and linear systems; Examples; Chap. 5 - Best Linear Unbiased Estimators [1 week] Definition of BLUE estimators; White Gaussian noise and bandlimited systems; Examples; Generalized MVU estimation; Chap. 6 - Maximum Likelihood Estimation [1 week] The maximum likelihood estimator; Properties of the ML estimators; Solution for ML estimation; Examples; Monte-Carlo methods; continues

An alternative strategy: FACT: It occurs that the MVU estimator, if it exists, can not be found. e.g. the PDF for the data is not known, the user would not like to assume a model for the PDF, or the problem can be mathematically untreatable. An alternative strategy can be pursued is to study the class of Best Linear Unbiased Estimators Only suboptimal performance can be achieved. The performance degradation, relative to the MVU estimator, is unknown but the resulting performance can be enough for the problem at hand.

BLUE structure: The Best Linear Unbiased Estimator consists of restrict the estimator to be a linear function of the data, i.e. where the a n s are constants to be determined. Optimality in general is lost. ˆθ = n=0 a n x n Examples revisited: DC level in WGN Nonlinear All unbiased estimators Nonlinear ˆθ = N +1 2N max x ( n ) MVU Linear Linear ˆθ = x MVU=BLUE Mean of Uniform noise ˆθ = x BLUE

Finding the BLUE: To find the BLUE we constrain the estimator to be linear to be unbiased to minimize its variance ˆθ = E ˆθ = var ˆθ a n x n n=0 ( a E x n n n=0 = θ (6.2) n=0 E a n x n n=0 ( ) = E a n x n Defining a=[a 0 a 1 a N-1 ] T and x=[x 0 x 1 x N-1 ] T this last expression can be simplified: var ˆθ ( ) = E a T x a T E x ( ) 2 = E a T x E x The problem of finding the BLUE can be stated as, for ( ) ( ) 2 = E at x E x ( )( x E x ) T a min a T Ca subject to a T E x = θ = = at Ca. (6.3) a R N )2

Finding the BLUE: Given the scalar parameter θ, the expected value of the samples can be assumed as E[x[n]]=s[n]θ, where s[n] is known. a n E x n n=0 = a n s n n=0 θ = θ, (from 6.2) Thus the previous problem can be stated as The method of Lagrangian multipliers can be used to solve this problem. Define the Lagrangian function as The gradient of J relative to a is min a T Ca a R N s.t. a T s = 1 J ( a,λ) = a T Ca + λ( a T s 1), λ R J ( a,λ) a = 2Ca + λs = 0

Finding the BLUE: Solving for a produces Using the constrain as Finally, the solution is a = λ 2 C 1 s a T s = λ 2 st C 1 s = 1, or a opt = C 1 s s T C 1 s, with a variance var ˆθ λ = 2 s T C 1 s. ( ) = a T Ca = st C 1 CC 1 s = opt opt ( s T C 1 s) 2 1 s T C 1 s. Taking into account that E[x]=sθ, finally the estimator ˆθ = a T opt x = st C 1 x s T C 1 s, Its expected value is E ˆθ Thus it is unbiased, as required. ( ) = st C 1 E x s T C 1 s = st C 1 sθ s T C 1 s = θ!

Example: Example (DC level in white Gaussian noise revisited): Model of signal: x n = A + w n, n = 0,..., N 1 Where w[n] is zero mean white noise with variance σ 2 (and an unspecified PDF), the problem is to estimate A. Because E[x[n]]=A, we have s=1, where 1=[1 1 1] T. The BLUE is  = 1 T 1 σ 2 Ix 1 T 1 σ 2 I1 = 1 N and the variance is x n = x, var  N =0 ( ) = 1 1 T 1 σ 2 I1 = σ 2 N. Hence the sample mean is the BLUE independent of the PDF of data. It is the MVU estimator for the Gaussian case. And in general: is it optimal?...

Example: Example (DC level UNCORRELATED noise): Model of signal: x n = A + w n, n = 0,..., N 1 Where w[n] is zero mean uncorrelated noise with var(w[n])=σ 2. Once again, the problem is to estimate A. We have again s=1, and C=diag(σ 0 2 σ 12 σ N-12 ), and C -1 =diag(1/σ 0 2 1/σ 12 1/σ N-12 ).. The BLUE is  = 1T C 1 x 1 T C 1 1 = N =0 x n 2 σ n 1 N =0 2 σ n and the variance is var( Â) = 1 1 N =0 2 σ n The BLUE weights those samples more heavily with smallest variances, in an attempt to equalize the noise contribution from each sample Is it optimal? In what cases?...

Extending BLUE to a Vector Parameter: To find the BLUE for a p x 1 vector parameter, we constrain the estimator to be linear ˆθ i = to be unbiased to minimize its variance a in x n i = 1,2,..., p or ˆθ = Ax, A R p x N n=0 E ˆθ var ˆθ i = AE ( ˆθ = θ n=0 E ai n x n n=0 ( ) = E a in x n The problem of finding the BLUE can be stated as, for )2 min var( ˆθ i ) = a T i Ca i subject to a T i E x = θ where E x = Hθ.

Extension to non-white Gaussian noise: Theorem 6.1(Gauss-Markov Theorem) If the data observed are of the general linear model form x = Hθ + w where H is a known N x p observation matrix (with N>p) and rank p, x is a N x 1 vector of observations, θ is a p x 1 vector of parameters to be estimated, and w is an N x 1 noise vector with zero mean and covariance C (for an arbitrary PDF), then the BLUE is and the covariance matrix of estimate is (1bis) C = ˆθ ( HT C 1 H) 1. (2bis)

Bibliography: Further reading Thomas Kailath, Linear Systems, Prentice Hall, 1980. Thomas Kailath, Ali Sayed, and Babak Hassibi, Linear Estimation, Prentice Hall, 2000. Harry L. Van Trees, Detection, Estimation, and Modulation Theory, Parts I to IV, John Wiley, 2001. J. Bibby, H. Toutenburg, Prediction and Improved Estimation in Linear Models, John Wiley, 1977. C.Rao, Linear Statistical Inference and Its Applications, John Wiley, 1973.