IEEE copyright notice

Similar documents
IEEE copyright notice

A Strict Stability Limit for Adaptive Gradient Type Algorithms

Estimation Theory. as Θ = (Θ 1,Θ 2,...,Θ m ) T. An estimator

Article (peer-reviewed)

On the Cramér-Rao lower bound under model mismatch

Frequentist-Bayesian Model Comparisons: A Simple Example

Estimation theory. Parametric estimation. Properties of estimators. Minimum variance estimator. Cramer-Rao bound. Maximum likelihood estimators

Title without the persistently exciting c. works must be obtained from the IEE

Gaussian Estimation under Attack Uncertainty

FAST AND ACCURATE DIRECTION-OF-ARRIVAL ESTIMATION FOR A SINGLE SOURCE

Estimation, Detection, and Identification

SIMON FRASER UNIVERSITY School of Engineering Science

USING STATISTICAL ROOM ACOUSTICS FOR ANALYSING THE OUTPUT SNR OF THE MWF IN ACOUSTIC SENSOR NETWORKS. Toby Christian Lawin-Ore, Simon Doclo

On the Cost of Worst-Case Coding Length Constraints

component risk analysis

1.1 Basis of Statistical Decision Theory

Minimax MMSE Estimator for Sparse System

Optimal Time Division Multiplexing Schemes for DOA Estimation of a Moving Target Using a Colocated MIMO Radar

5. Discriminant analysis

3.0.1 Multivariate version and tensor product of experiments

Publication VI. Esa Ollila On the circularity of a complex random variable. IEEE Signal Processing Letters, volume 15, pages

Elliptically Contoured Distributions

Fast Angular Synchronization for Phase Retrieval via Incomplete Information

A Framework for Optimizing Nonlinear Collusion Attacks on Fingerprinting Systems

Physics 403. Segev BenZvi. Parameter Estimation, Correlations, and Error Bars. Department of Physics and Astronomy University of Rochester

Introduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak

MULTICHANNEL SIGNAL PROCESSING USING SPATIAL RANK COVARIANCE MATRICES

Exploring Granger Causality for Time series via Wald Test on Estimated Models with Guaranteed Stability

ADAPTIVE FILTER THEORY

Applied Linear Algebra in Geoscience Using MATLAB

Rethinking Biased Estimation: Improving Maximum Likelihood and the Cramér Rao Bound

TWO METHODS FOR ESTIMATING OVERCOMPLETE INDEPENDENT COMPONENT BASES. Mika Inki and Aapo Hyvärinen

Restricted Maximum Likelihood in Linear Regression and Linear Mixed-Effects Model

Composite Hypotheses and Generalized Likelihood Ratio Tests

Bayesian Methods in Positioning Applications

DETECTION theory deals primarily with techniques for

Application-Oriented Estimator Selection

A study on regularization parameter choice in Near-field Acoustical Holography

Riccati difference equations to non linear extended Kalman filter constraints

Estimation of the Optimum Rotational Parameter for the Fractional Fourier Transform Using Domain Decomposition

Advanced Signal Processing Introduction to Estimation Theory

MMSE DECODING FOR ANALOG JOINT SOURCE CHANNEL CODING USING MONTE CARLO IMPORTANCE SAMPLING

Integer Least Squares: Sphere Decoding and the LLL Algorithm

Chapter 2 Fundamentals of Adaptive Filter Theory

covariance function, 174 probability structure of; Yule-Walker equations, 174 Moving average process, fluctuations, 5-6, 175 probability structure of

Math 322. Spring 2015 Review Problems for Midterm 2

EIE6207: Maximum-Likelihood and Bayesian Estimation

MULTIPLE-CHANNEL DETECTION IN ACTIVE SENSING. Kaitlyn Beaudet and Douglas Cochran

Title algorithm for active control of mul IEEE TRANSACTIONS ON AUDIO SPEECH A LANGUAGE PROCESSING (2006), 14(1):

An Adaptive Sensor Array Using an Affine Combination of Two Filters

Brief Review on Estimation Theory

Estimation of linear non-gaussian acyclic models for latent factors

THE problem of parameter estimation in linear models is

1 Data Arrays and Decompositions

Statistical and Adaptive Signal Processing

Mark your answers ON THE EXAM ITSELF. If you are not sure of your answer you may wish to provide a brief explanation.

Multiple Degree of Freedom Systems. The Millennium bridge required many degrees of freedom to model and design with.

MULTICHANNEL FAST QR-DECOMPOSITION RLS ALGORITHMS WITH EXPLICIT WEIGHT EXTRACTION

Acoustic Vector Sensor based Speech Source Separation with Mixed Gaussian-Laplacian Distributions

Tracking of Multiple Moving Sources Using Recursive EM Algorithm

J. Liang School of Automation & Information Engineering Xi an University of Technology, China

A Polynomial Chaos Approach to Robust Multiobjective Optimization

Acoustic MIMO Signal Processing

UNIFORMLY MOST POWERFUL CYCLIC PERMUTATION INVARIANT DETECTION FOR DISCRETE-TIME SIGNALS

Energy-Efficient Noncoherent Signal Detection for Networked Sensors Using Ordered Transmissions

LOW COMPLEXITY COVARIANCE-BASED DOA ESTIMATION ALGORITHM

Optimal Mean-Square Noise Benefits in Quantizer-Array Linear Estimation Ashok Patel and Bart Kosko

LTI Systems, Additive Noise, and Order Estimation

ECE 275A Homework 6 Solutions

Mobile Robotics 1. A Compact Course on Linear Algebra. Giorgio Grisetti

Large Sample Properties of Estimators in the Classical Linear Regression Model

Problem Set 2. MAS 622J/1.126J: Pattern Recognition and Analysis. Due: 5:00 p.m. on September 30

EECE Adaptive Control

Logarithmic quantisation of wavelet coefficients for improved texture classification performance

Nonlinear Programming Models

Recursive Estimation

A Simpler Approach to Low-Rank Tensor Canonical Polyadic Decomposition

2 Statistical Estimation: Basic Concepts

A Very Brief Summary of Statistical Inference, and Examples

Bayesian estimation of chaotic signals generated by piecewise-linear maps

Analysis of the AIC Statistic for Optimal Detection of Small Changes in Dynamic Systems

ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications

A Subspace Approach to Estimation of. Measurements 1. Carlos E. Davila. Electrical Engineering Department, Southern Methodist University

REFERENCES AND FURTHER STUDIES

EM-algorithm for Training of State-space Models with Application to Time Series Prediction

A GENERAL CLASS OF LOWER BOUNDS ON THE PROBABILITY OF ERROR IN MULTIPLE HYPOTHESIS TESTING. Tirza Routtenberg and Joseph Tabrikian

Noisy Subsequence Recognition Using Constrained String Editing Involving Substitutions, Insertions, Deletions and Generalized Transpositions 1

MTH Linear Algebra. Study Guide. Dr. Tony Yee Department of Mathematics and Information Technology The Hong Kong Institute of Education

Virtual Array Processing for Active Radar and Sonar Sensing

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Contents. 1 Vectors, Lines and Planes 1. 2 Gaussian Elimination Matrices Vector Spaces and Subspaces 124

Over-enhancement Reduction in Local Histogram Equalization using its Degrees of Freedom. Alireza Avanaki

Math 1060 Linear Algebra Homework Exercises 1 1. Find the complete solutions (if any!) to each of the following systems of simultaneous equations:

ELEG 5633 Detection and Estimation Minimum Variance Unbiased Estimators (MVUE)

Multivariate Distributions

5 Linear Algebra and Inverse Problem

Introduction to Estimation Methods for Time Series models. Lecture 1

COMP 551 Applied Machine Learning Lecture 20: Gaussian processes

Shannon meets Wiener II: On MMSE estimation in successive decoding schemes

On the Estimation of the Mixing Matrix for Underdetermined Blind Source Separation in an Arbitrary Number of Dimensions

Transcription:

IEEE copyright notice Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or redistribution to servers or lists, or to reuse any copyrighted component of this work in other works must be obtained from the IEEE. Contact: Manager, Copyrights and Permissions / IEEE Service Center / 445 Hoes Lane / P.O. Box 1331 / Piscataway, NJ 8855-1331, USA. Telephone: + Intl. 98-56-3966.

MMSE ESTIMATION IN A LINEA SIGNAL MODEL WITH ELLIPSOIDAL CONSTAINTS Stefan Uhlich and Bin Yang Chair of System Theory and Signal Processing Universität Stuttgart, Germany ABSTACT The estimation of an unknown parameter vector in a Gaussian linear model is studied in this paper. Two different cases are analyzed: the parameter vector is assumed to lie either in or on a given ellipsoid. The best estimator in terms of the mean squared error is derived. The performance of this estimator is analyzed and compared with the ordinary least squares, the constrained least squares and the linear minimax approach. Index Terms Parameter estimation, Minimum mean squared error estimation, estricted parameter space 1. INTODUCTION In this paper we consider the problem of estimating an unknown parameter vector θ M from noisy measurements where it is a priori known that the parameter vector is restricted to lie in a subset of M. The two subset cases that are investigated are: the set of all points inside an ellipsoid and the set of all points on the surface of an ellipsoid. The first case has attracted a lot of interest in the past, see e.g. [1 5] where the linear minimax estimator is sought which minimizes the worst case mean squared error. It is well known that the ordinary least squares estimator is outperformed by this estimator. The second case is interesting for DOA estimation problems, see e.g. [6] for an example where a parameter vector has to be estimated that lies on a 3-D sphere which is a special case of the problem we consider. The contribution of this paper is twofold: First, we derive in Sec. 3 the minimum mean squared error (MMSE) estimator for an unknown parameter vector in a linear model with ellipsoidal constraints. We show that the two M-dimensional integrations can be replaced by M one-dimensional integrations. Second, we analyze in Sec. 5 its performance with respect to different error measures and compare it with the ordinary least squares approach, the linear minimax estimator from [, 7], the constrained least squares estimator and modifications of them. The simulation results show that the outperforms the other approaches at the expense of an increased computational complexity. Following notations are used throughout this paper: x denotes a vector, X amatrix,andi the identity matrix. Furthermore, the error function erf(x) is defined as erf(x) x π e t dt.. SIGNAL MODEL We consider the linear signal model x Hθ + z (1) where x N are observations of the unknown parameter vector θ M with N M. H N M is the known model matrix with a full column rank M and z N the observation noise which is Gaussian with z N(,C). It is a priori known that the parameter vector θ satisfies one of the following two constraints: Case A: θ lies inside an ellipsoid,i.e.θ. The ellipsoid is given by {θ :(θ θ ) T Θ(θ θ ) } or Case B: θ lies on an ellipsoid,i.e.θ,with {θ :(θ θ ) T Θ(θ θ ) }. (a) (b) The problem of finding a suitable estimator for θ is a well known problem in signal processing. In [1] the linear minimax estimator for H I and C σ I is derived if θ is restricted to lie in an ellipsoid. This result was later extended in [, 7]. In this paper, we will derive the for this problem by modeling θ as uniformly distributed in or on,respectively. This is the fundamental difference of our approach to the previous works which always assume θ to be deterministic. The uniform distribution of the has the interpretation that we only know that θ is restricted to lie in or on, but no other a priori information is available. Thus, our results are directly comparable to those which were derived in a deterministic framework. 3. CASE A: MMSE ESTIMATO FO θ We now derive the estimator for θ in (1) which has the minimum mean squared error among all estimators. It is well known that the is the mean of the a posteriori distribution [8]: ˆθ E[θ x] M θ p x θ (x θ)p θ (θ)dθ M p x θ (x θ)p θ (θ)dθ. (3) It requires in general two M-dimensional integrations which can often not be solved analytically. In our case, however, we can simplify these to M one-dimensional numerical integrations which are feasible. The derivation of the estimator is divided into three steps: A. Transform the problem to one of estimating a vector in colored noise. The ellipsoid in which θ lies is arbitrarily oriented. B. Transform the problem to one of estimating a vector in white noise where the axes of the ellipsoid are parallel to the coordinate axes. C. Give formulas for the remaining integrals. Step A. As we assume θ and θ is a priori uniformly distributed, it has the probability density function (pdf) ( const θ p θ (θ) (4) otherwise 978-1-444-354-5/9/$5. 9 IEEE 349 ICASSP 9

Furthermore, z N(,C) and therefore we can write (3) as θ exp{ 1 (x ˆθ Hθ)T C 1 (x Hθ)}dθ exp{ 1 (x Hθ)T C 1 (x Hθ)}dθ. (5) Using the identity (x Hθ) T C 1 (x Hθ)(θ x) T C 1 (θ x)+x T C 1 (x H x) (6) with C (H T C 1 H) 1 and x (H T C 1 H) 1 H T C 1 x,we can further write ˆθ θ exp{ 1 (θ x)t C 1 (θ x)}dθ exp{ 1 (θ x)t C 1 (θ x)}dθ. (7) The term x T C 1 (x H x) in (6) is independent of θ and is cancelled. Thus, the problem in (1) can be transformed into a simpler one by left-multiplying (1) with (H T C 1 H) 1 H T C 1 which results in the signal model x θ + z. The problem is now to estimate θ which is observed in colored noise z N(, C). Note, that x represents the ordinary least squares estimate for θ. The improves this estimate further by exploiting the restriction θ and therefore shows a better performance. Thus, we can conclude that the projection into the column space of H as done by the least squares estimate x does not neglect information which we need for the. Step B. We define U as one matrix square root of C,i.e. C U T U. One possibility is the Cholesky factor of C. Furthermore, let V and D diag(d 1,...,d M) be the matrix of eigenvectors and eigenvalues of UΘU T. Using the substitution θ x U T Vy,(7) reads ˆθ ( x + U T Vy)exp{ 1 y }dy exp{ 1 y }dy x + U T V y exp{ 1 y }dy exp{ 1 y }dy : x + U T V i 1 (8) i where is the ellipsoid {y :(y y ) T D(y y ) } with y (U T V) 1 (θ x). Step C. The requires in general a multidimensional integration which is very computational demanding. However, in this particular case the numerical integration can be substantially simplified. In [9, 1] it is shown that i can be expressed as an infinite linear combination of χ distributions, i.e. i (π) M/ P k a kg( /β; M +k) where G( /β; M +k) is the central χ cumulated distribution function (cdf) with M +k degrees of freedom evaluated at position /β and β is an arbitrary constant. The coefficients a k can be found by the recursive rule a r MY β e y d m m m /, a k 1 k 1 X b k l a l k P where b k k M P m1 y mγm k 1 + 1 M m1 (1 ky m)γm, k γ m 1 β d m and y m is the mth component of y. The calculation of i is therefore straightforward where the summation is stopped if the relative error is small enough. An estimate of the truncation error is derived in [9] and was used as stopping criterion. It therefore remains to calculate i 1.Itsmthelement reads i 1m y me 1 y dy y m + y m y me 1 y m B @ m l 1 e 1 ỹ m C dỹ m A dy m (9) where ỹ m P my is identical to y except for the removed mth element. The selection matrix P m (M 1) M is the identity matrix with the mth row erased. Note, that the inner integral is of the same type as i and can thus also be expressed as an infinite series of χ cumulated density functions. The integration area m is given by m {ỹ m :(ỹ m P my ) T P mdp T m(ỹ m P my ) d m(y m y m) } andisafunctionofy m. Eq. (9) can be efficiently solved by an one-dimensional numerical integration. For the special case M 1,i.e. θ θ is a scalar, we can simplify (8) further and give a closed-form solution for ˆθ. Using x h T C 1 x/(h T C 1 h), U (h T C 1 h) 1/, V 1, D d Θ/(h T C 1 h) and y (h T C 1 h) 1/ (θ x), we obtain (1) at the bottom of this page. 4. CASE B: MMSE ESTIMATO FO θ For the case B that θ lies on the surface of an ellipsoid, i.e. θ, we can use the estimator (8) found in Sec. 3 where all integrations are now with respect to from (b) and {y : (y y ) T D(y y ) } opposed to and. Only the ˆθ ht C 1 x h T C 1 h + 1 (h T C 1 h) 1/ y + d ye 1 y dy y d y + d e 1 y dy y d ht C 1 x h T C 1 h + 1 e (h T C 1 h) 1/ p π 1 (y + d ) sinh y d «««(1) y + y erf d erf d 35

third step where the integrals are solved has to be changed. It holds i 1m y me 1 y dy y m + y m i (π) M/ β y me 1 y m B @ m X a k g( /β; M +k) k 1 e 1 ỹ m C dỹ m A dy m (11a) (11b) where g( /β; M +k) is the central χ probability density function (pdf) with M +k degrees of freedom evaluated at position /β. Similar to Sec. 3, we can express i as an infinite series of χ densities and i 1 is found again by M one-dimensional numerical integrations. 5. SIMULATION ESULTS In the following we show some simulation results for case A and B. In both cases, M 3unknowns have to be estimated from N 1 observations. The model matrix H is randomly chosen for each trial and in total we average over 1 trials, except for Fig. where we average over 1 trials. The noise z has the distribution N(,σ I) and the signal-to-noise ratio (SN) is defined as SN 1 log 1 Hθ /(Nσ ). For case A, we assume that the parameter vector θ is restricted to lie in the unit sphere {θ : θ 1} and is uniformly distributed for the simulation. For case B, we assume that θ is uniformly distributed on the unit sphere {θ : θ 1}. 5.1. Case A We compare the derived with the following four estimators for a spherical constraint: Ordinary LS estimator [8] ˆθ LS (H T C 1 H) 1 H T C 1 x x (1) Modified ordinary LS estimator, inspired by [5] ˆθ MLS (ˆθLS estimator [, 7] ˆθ MX Modified minimax estimator [5] ˆθ MMX ˆθ LS > (13) +tr{(h T C 1 H) 1 } ˆθ LS (14) (ˆθMX ˆθ MX ˆθ MX ˆθ MX ˆθ MX > (15) Fig. 1 shows the simulation results for the squared error θ ˆθ averaged over all θ. The derived has clearly the minimal averaged squared error and is therefore superior to the other estimators as expected. It has, however, a higher computational complexity because of the M one-dimensional numerical integrations. The second comparison is in terms of the risk of the linear minimax and the. The risk of an estimator that corresponds to a quadratic loss function is [8] (θ, ˆθ)E x θ θ ˆθ(x) θ ˆθ(x) p(x θ)dx. (16) It is the quadratic loss θ ˆθ averaged over the distribution of the measurements with θ fixed. This comparison is interesting as the linear minimax estimator minimizes the worst case mean squared error for each deterministic θ opposed to the which only considers the overall mean squared error. Note, that the estimator risk is rotational invariant in the parameter space of θ as the model matrix H is chosen randomly. Thus, it is sufficient to plot the estimator risk as a function of the norm of θ only and the risk in (16) is calculated by averaging the squared error over all θ with θ const. Fig. depicts the simulation results for a SN of db. It shows that the is, except for a small region around θ.4, better than the linear minimax estimator. Thus, although we derived the estimator which minimizes the mean squared error averaged over all possible parameter vectors in, it is still almost always better in terms of the risk than the linear minimax estimator which was designed to minimize the worst case error for the deterministic case. 5.. Case B We compare the derived with the ordinary least squares estimator (1) and the following three estimators: Scaled ordinary LS estimator, inspired by [5] ˆθ ScLS ˆθ LS (17) Spherical LS estimator, corresponds to the class of constrained LS estimators [11] ˆθ SLS min θ x Hθ s.t. θ (18) which we solved by expressing θ in spherical coordinates and using a nonlinear least squares optimization procedure. estimator, which is for case B equal to the estimator given in (14) as the worst case MSE to be minimized is always located on the boundary of the ellipsoid. Hence, calculating the linear minimax estimator with respect to is equal to calculating it with respect to. Fig. 3 and 4 show simulation results for case B with respect to the averaged squared error θ ˆθ and the averaged angle between θ and ˆθ in degrees cos 1 θ T ˆθ θ ˆθ. The derived is superior to the other estimators for both error measures. Especially the averaged angle error is interesting for DOA applications as in [6] where it is important that θ and ˆθ point to the same direction. 6. CONCLUSIONS In this paper we derived the for a linear Gaussian model under ellipsoidal constraints. It outperforms all other estimators we know at the expense of a higher computational complexity. A Matlab implementation of the and the other considered estimators can be downloaded from [1]. 351

1 1 Ordinary Least Squares Modified Least Squares Modified linear minimax 1 1 1 Ordinary Least Squares Scaled Least Squares Spherical Least Squares Mean squared error 1 1 1 Mean squared error 1 1 1 1 5 15 1 5 5 1 SN in db Fig. 1. Case A: Comparison of averaged squared error 5 15 1 5 5 1 SN in db Fig. 3. Case B: Comparison of averaged squared error Estimator risk.35.3.5..15.1.5.1..3.4.5.6.7.8.9 1 θ Mean angle between θ and ˆθ in degrees 7 6 5 4 3 1 Ordinary LS + Scaled LS Spherical Least Squares 15 1 5 5 1 SN in db Fig.. Case A: Comparison of estimator risk Fig. 4. Case B: Comparison of averaged angle in degrees 7. EFEENCES [1] M. S. Pinsker, Optimal filtration of square-integrable signals in Gaussian noise, Problems of Information Transmission, vol. 16, pp. 1 133, 198. [] J. Pilz, Minimax linear regression estimation with symmetric parameter restrictions, Journal of Statistical Planning and Inference, vol. 13, pp. 97 318, 1986. [3] Y. C. Eldar, Uniformly improving the Cramer-ao Bound and Maximum-Likelihood estimation, IEEE Trans. Signal Processing, vol. 54, no. 8, pp. 943 956, 6. [4] Y. C. Eldar, Universal weighted MSE improvement of the Least-Squares estimator, IEEE Trans. Signal Processing, vol. 56, no. 5, pp. 1788 18, 8. [5] P. Stoica and G. Ganesan, Linear regression constrained to aball, Digital Signal Processing, vol. 11, no. 1, pp. 8 9,. [6]. Mukai, H. Sawada, S. Araki, and S. Makino, eal-time blind source separation and DOA estimation using small 3-D microphone array, Proc. International Workshop for Acoustic Echo and Noise Control (IWAENC), pp. 45 48, 5. [7] Y.C.Eldar,A.Ben-Tal,andA.Nemirovski, obustmeansquared error estimation in the presence of model uncertainties, IEEE Trans. Signal Processing, vol. 53, no. 1, pp. 168 181, 5. [8] L. L. Scharf, Statistical Signal Processing: Detection, Estimation and Time Series Analysis, Addison-Wesley, 199. [9] H. uben, Probability content of regions under spherical normal distributions, IV: The distribution of homogeneous and non-homogeneous quadratic functions of normal variables, Ann. of Math. Statist., vol. 33, no., pp. 54 57, 196. [1] N. L. Johnson and S. Kotz, Distributions in Statistics: Continuous Univariate Distributions, John Wiley & Sons, 197. [11] T. K. Moon and W. C. Stirling, Mathematical Methods and Algorithms for Signal Processing, Prentice-Hall,. [1] S. Uhlich, Matlab implementation of the MMSE estimator for a restricted parameter vector in a linear Gaussian model, Available: http://www.lss.unistuttgart.de/mitarbeiter/uhlich.en.html. 35