LINEAR MMSE ESTIMATION

Similar documents
Ch. 12 Linear Bayesian Estimators

4 Derivations of the Discrete-Time Kalman Filter

2 Statistical Estimation: Basic Concepts

,, rectilinear,, spherical,, cylindrical. (6.1)

Machine Learning. A Bayesian and Optimization Perspective. Academic Press, Sergios Theodoridis 1. of Athens, Athens, Greece.

MATH 167: APPLIED LINEAR ALGEBRA Least-Squares

Advanced Signal Processing Introduction to Estimation Theory

An overview of key ideas

Chapter 2 Wiener Filtering

LINEAR ALGEBRA: THEORY. Version: August 12,

The geometry of least squares

The Hilbert Space of Random Variables

Maximum variance formulation

Linear Algebra. Preliminary Lecture Notes

Linear Regression and Its Applications

Inner Product Spaces

Chapter 0 of Calculus ++, Differential calculus with several variables

MA 575 Linear Models: Cedric E. Ginestet, Boston University Revision: Probability and Linear Algebra Week 1, Lecture 2

Definition 1. A set V is a vector space over the scalar field F {R, C} iff. there are two operations defined on V, called vector addition

Module 2. Random Processes. Version 2, ECE IIT, Kharagpur

Linear Algebra. Preliminary Lecture Notes

UNDERSTANDING THE DIAGONALIZATION PROBLEM. Roy Skjelnes. 1.- Linear Maps 1.1. Linear maps. A map T : R n R m is a linear map if

Singular Value Decomposition. 1 Singular Value Decomposition and the Four Fundamental Subspaces

Math Advanced Calculus II

Chapter 6. Orthogonality and Least Squares

22m:033 Notes: 6.1 Inner Product, Length and Orthogonality

Chapter 2. Vectors and Vector Spaces

v = v 1 2 +v 2 2. Two successive applications of this idea give the length of the vector v R 3 :

Fourier and Wavelet Signal Processing

MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors.

6 Basis. 6.1 Introduction

Lecture 1: Systems of linear equations and their solutions

Topic 14 Notes Jeremy Orloff

which are not all zero. The proof in the case where some vector other than combination of the other vectors in S is similar.

12. Special Transformations 1

Section 6.1. Inner Product, Length, and Orthogonality

Linear Algebra, 4th day, Thursday 7/1/04 REU Info:

Section 3.1: Definition and Examples (Vector Spaces), Completed

Statistics 910, #5 1. Regression Methods

16.584: Random Vectors

Inner product spaces. Layers of structure:

Chapter Two Elements of Linear Algebra

[POLS 8500] Review of Linear Algebra, Probability and Information Theory

3. For a given dataset and linear model, what do you think is true about least squares estimates? Is Ŷ always unique? Yes. Is ˆβ always unique? No.

MTH 2310, FALL Introduction

Basic Linear Algebra in MATLAB

JUST THE MATHS UNIT NUMBER 1.5. ALGEBRA 5 (Manipulation of algebraic expressions) A.J.Hobson

Topics in Probability and Statistics

Vectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1. x 2. x =

Abstract Vector Spaces

Dot Products. K. Behrend. April 3, Abstract A short review of some basic facts on the dot product. Projections. The spectral theorem.

Mon Apr dot product, length, orthogonality, projection onto the span of a single vector. Announcements: Warm-up Exercise:

A matrix over a field F is a rectangular array of elements from F. The symbol

The Four Fundamental Subspaces

ECONOMETRIC METHODS II: TIME SERIES LECTURE NOTES ON THE KALMAN FILTER. The Kalman Filter. We will be concerned with state space systems of the form

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det

Eigenvalues and Eigenvectors. Prepared by Vince Zaccone For Campus Learning Assistance Services at UCSB

Estimation theory. Parametric estimation. Properties of estimators. Minimum variance estimator. Cramer-Rao bound. Maximum likelihood estimators

Solutions to Selected Questions from Denis Sevee s Vector Geometry. (Updated )

Upon successful completion of MATH 220, the student will be able to:

Linear Algebra and Eigenproblems

Math Exam 2, October 14, 2008

Lecture 2: Repetition of probability theory and statistics

Vector Calculus. Lecture Notes

Intro Vectors 2D implicit curves 2D parametric curves. Graphics 2012/2013, 4th quarter. Lecture 2: vectors, curves, and surfaces

LECTURE 10: REVIEW OF POWER SERIES. 1. Motivation

The Gram-Schmidt Process 1

NOTES ON VECTORS, PLANES, AND LINES

SOLUTIONS TO EXERCISES FOR MATHEMATICS 133 Part 1. I. Topics from linear algebra

Vector Spaces. distributive law u,v. Associative Law. 1 v v. Let 1 be the unit element in F, then

6. Vector Random Variables

STATISTICAL LEARNING SYSTEMS

Estimation techniques

Principal Component Analysis

Definition: A vector is a directed line segment which represents a displacement from one point P to another point Q.

Econ 2120: Section 2

Linear Algebra, Summer 2011, pt. 2

Lecture Notes 1: Vector spaces

MATH Linear Algebra

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 26. Estimation: Regression and Least Squares

CS 143 Linear Algebra Review

Final Review Sheet. B = (1, 1 + 3x, 1 + x 2 ) then 2 + 3x + 6x 2

Orthogonality. Orthonormal Bases, Orthogonal Matrices. Orthogonality

Linear Algebra- Final Exam Review

This pre-publication material is for review purposes only. Any typographical or technical errors will be corrected prior to publication.

CS123 INTRODUCTION TO COMPUTER GRAPHICS. Linear Algebra 1/33

Span and Linear Independence

Matrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A =

2. Signal Space Concepts

Multivector Calculus

Chapter 3. More about Vector Spaces Linear Independence, Basis and Dimension. Contents. 1 Linear Combinations, Span

Page 52. Lecture 3: Inner Product Spaces Dual Spaces, Dirac Notation, and Adjoints Date Revised: 2008/10/03 Date Given: 2008/10/03

LINEAR ALGEBRA W W L CHEN

Math 3013 Problem Set 4

Solving Systems of Equations Row Reduction

Intro Vectors 2D implicit curves 2D parametric curves. Graphics 2011/2012, 4th quarter. Lecture 2: vectors, curves, and surfaces

Eigenvalues, Eigenvectors, and an Intro to PCA

Spring 2015 Midterm 1 03/04/15 Lecturer: Jesse Gell-Redman

LINEAR ALGEBRA W W L CHEN

Transcription:

LINEAR MMSE ESTIMATION TERM PAPER FOR EE 602 STATISTICAL SIGNAL PROCESSING By, DHEERAJ KUMAR VARUN KHAITAN 1

Introduction Linear MMSE estimators are chosen in practice because they are simpler than the optimal Bayesian estimators and retain the MMSE criterion. Here the constraint on the estimator is assumed to be linear. We will also show how this estimator depends just on the correlation between the variable to be estimated and the observations. This is a very important advantage as it does not require us to know the joint probability function of the observation and the estimate. Linear MMSE Estimation Let us assume that we have to estimate a scalar parameter based on the data set { x[0],x[l],..., x[n 1] } or in vector form X = [ x[0] x[l]... x[n - I] ] T. We do not assume any specific knowledge of the joint PDF p(x, ), because as we shall see, a knowledge of the first two moments suffices. We show that the correlation between and X is sufficient to estimate. We now consider the class of all linear estimators of the form The task is to choose the weighting coefficients Error) ; to minimize the Bayesian MSE (Mean Square Here the expectation is with respect to the PDF p(x, ). The resultant estimator is termed the linear minimum mean square error (LMMSE) estimator. Here the term has been incorporated to allow for a mismatch between the means of and x. If the means are equal then this coefficient is zero and may be omitted. A LMMSE will be suboptimal unless the MMSE estimator happens to be linear. Otherwise, better estimators will exist which will of course be nonlinear. It is important to note that since the LMMSE estimator relies on the correlation between random variables, a parameter uncorrelated with the data cannot be linearly estimated. We now try to find the optimal values of the coefficients by minimizing the BMSE error. We differentiate the error with respect to and set it to zero. 2

Substituting this into the BMSE expression we now need to minimize over the other N-1 coefficients. Letting a =, we have Where is the NxN covariance matrix of X and is the Nx1 cross covariance vector. Taking the derivative and equating to zero we get Finally the value of is and for this the value of the BMSE error is 3

GEOMETRICAL INTERPRETATION For this interpretation we first assume that and x are zero mean. If that is not the case we subtract their means from the variables and thus make them zero mean random variables. Let and we want to minimize the error Let us now think of the random variables space as shown symbolically., x[0], x[1],..., x[n 1] as elements in a vector If can be expressed as a linear combination of the observations then it will lie in the plane spanned by them. If it cannot, as is usually the case, we picture as only partially lying in the subspace spanned by the x[n]'s. To complete the description of our vector space we require the notion of an inner product between two vectors. We define the inner product as <x,y> = E[xy]. This means that minimization of the MSE is equivalent to a minimization of the squared length of the error vector E =. As can be seen in the figure below the length of the is minimized when E is orthogonal to the subspace spanned by the. 4

If it had a component in the plane then that could always be taken care of by changing the values of the coefficients. Hence, we require This is the important orthogonality principle or projection theorem. It says that in estimating the realization of a random variable by a linear combination of data samples, the optimal estimator is obtained when the error is orthogonal to each data sample. Using the orthogonality principle, the weighting coefficients are easily found as In the matrix form these N equations can be written as Recognizing these matrices as familiar forms we see that the matrix equation can be written as or. Thus the LMMSE is given by. Finally 5

The minimum Bayesian MSE is the squared length of the error vector or After a series of manipulations like We see that the last term is zero because of the orthogonality condition. Thus the estimate simplifies to This is in agreement with the analytical expression we obtained earlier and thus we have shown that the process of obtaining the LMMSE has a very simple geometrical interpretation. 6