Advanced Signal Processing Minimum Variance Unbiased Estimation (MVU)

Size: px
Start display at page:

Download "Advanced Signal Processing Minimum Variance Unbiased Estimation (MVU)"

Transcription

1 Advanced Signal Processing Minimum Variance Unbiased Estimation (MVU) Danilo Mandic room 813, ext: Department of Electrical and Electronic Engineering Imperial College London, UK URL: mandic c D. P. Mandic Advanced Signal Processing 1

2 Objectives Learn the concept of minumum variance unbiased (MVU) estimation Investigate how the accuracy of an estimator depends upon the relationship between the unknown parameter(s) and the PDF of noise Study the requirements for the design of an efficient estimator Analyse the Cramer Rao Lower Bound (CRLB) for the scalar case Extension to the Cramer Rao Lower Bound (CRLB) for the vector case Sinusoid parameter estimation, linear models Dependence on data length (motivation for sufficient statistics ) Examples: DC level in WGN (frequency estimation in power, bioengineering) Parameters of a sinusoid (scalar case, vector case) A new view of Fourier analysis System identification c D. P. Mandic Advanced Signal Processing 2

3 Example 1: Consider a single observation x[0] = A + w[0], where w[0] N(0,σ 2 ) The simplest estimator of the DC level A in white noise w[0] N(0,σ 2 ) is  = x[0]  is unbiased, with the variance σ2 To show that the estimator accuracy improves as σ 2 decreases: Consider [ ] p i (x[0];a) = 1 exp 1 (x[0] A) 2 2πσ 2 2σ 2 i i for x[0] = 3 and i = 1,2 with σ }{{} 1 = 1 3 and σ 2 = 1 f undamental step, we are f ixing the data value p (x[0]=3;a) 1 p (x[0]=3;a) 2 Case I σ 1=1/3 Case II σ 2=1 3 A 3 Clearly, as σ 1 < σ 2, the DC level A is estimated more accurately with p 1 (x[0];a) Likely candidates for values of A 3 ± 3σ therefore [2, 4] for σ 1 and [0, 6] for σ 2. A c D. P. Mandic Advanced Signal Processing 3

4 Likelihood function When the PDF is viewed as a function of an unknown parameter (with the dataset {x} = x[0],x[1],... fixed) it is termed the likelihood function. The sharpness of the likelihood function determines the accuracy at which the unknown parameter may be estimated. Sharpness is measured by the curvature a negative of the second derivative of the logarithm of the likelihood function at its peak. Example 2: Estimation based on one sample of a DC level in WGN lnp(x[0];a) = ln 2πσ 2 1 2σ 2 ( x[0] A ) 2 then ln p(x[0]; A) A = 1 σ 2 ( x[0] A ) and the curvature 2 lnp(x[0];a) A 2 = 1 σ 2 Therefore, as expected, the curvature increases as σ 2 decreases. c D. P. Mandic Advanced Signal Processing 4

5 Likelihood function: Curvature Since we know that the variance of the estimator equals σ 2, then var(â) = 1 2 lnp(x[0];a) A 2 and the variance decreases as the curvature increases. Generally, the second derivative does depend upon x[0], and hence a more appropriate measure of curvature is the statistical measure [ 2 ] lnp(x[0];a) E A 2 which measures the average curvature of the log-likelihood function Recall: The likelihood function is a random variable due to x[0] Recall: The Mean Square Error MSE = Bias 2 + variance Ê it makes perfect sense to look for a minimum variance unbiased (MVU) solution c D. P. Mandic Advanced Signal Processing 5

6 Link with human perception In the 50s, phsychologist Fred Attneave recorded eye dwellings on objects Example 3: The drawing of a bean (top) andthehistogramofeyedwellings(bottom) Example 4: Read the words below... now read letter by letter... are you still sure? Example 3: Is the drawing on the left still a penguin? So, what is the sufficient information to estimate an object? c D. P. Mandic Advanced Signal Processing 6

7 THE KEY: Cramer-Rao Lower Bound (CRLB) for a scalar parameter (performance of the theoretically best estimator) The Cramer Rao Lower Bound (CRLB) Theorem: [CRLB] Assumption: The PDF p(x; θ) satisfies the regularity condition [ ] lnp(x;θ) E = 0, θ θ where the expectation is taken with respect to p(x; θ). Then, the variance of any unbiased estimator, ˆθ, must satisfy var(ˆθ) E 1 { } 2 lnp(x;θ) θ 2 where the derivative is evaluated at the true value of θ. c D. P. Mandic Advanced Signal Processing 7

8 CRLB for a scalar parameter, continued Moreover, an unbiased estimator may be found that attains the bound for all θ, if and only if for some functions g and I lnp(x;θ) = I(θ) ( g(x) θ ) θ This estimator is the minimum variance unbiased (MVU) estimator, for which ˆθ = g(x) and its minimum variance 1 I(θ) end of CRLB theorem 1 Remark: Since the variance var(ˆθ) { E 2 lnp(x;θ) θ 2 the curvature term gives [ 2 ] lnp(x;θ) 2 lnp(x;θ) E θ 2 = θ 2 p(x;θ)dx }, the evaluation of Obviously, in general the bound depends on the parameter θ and the data length c D. P. Mandic Advanced Signal Processing 8

9 Example 5: Estimation of a DC level in WGN Consider the estimation of a DC level in WGN, assume N observations x[n] = A }{{} unknown DC level + w[n] }{{} noise with known pdf n = 0,1,2,...,N 1 where w[n] N(0,σ 2 ). Determine the CRLB for the unknown DC level A, starting from (θ = A) p(x;θ) = p(x;a) = = N 1 n=0 1 (2πσ 2 ) N/2exp [ 1 exp 1 ] ( ) 2 x[n] A 2πσ 2 2σ 2 [ 1 2σ 2 N 1 n=0 ( x[n] A ) 2 ] Estimation of a DC level is very useful, e.g. in the time-frequency plane a sinusoid of frequency f is represented by a straight line Ê c D. P. Mandic Advanced Signal Processing 9

10 Example 5: DC level in WGN continued Taking the first derivative, we have [ lnp(x;a) = ln [ 2πσ 2] N/2 1 A A 2σ 2 = 1 N 1 σ 2 n=0 where x denotes the sample mean. Connection with CRLB: x = 1 N Upon differentiating again N 1 n=0 ( x[n] A ) = N σ 2 ( x A ) N 1 n=0 2 lnp(x;a) A 2 = N σ 2 ( x[n] A ) 2 ] x(n) = g(x), I(A) = N/σ2 Therefore var(â) σ2 N is the CRLB, which implies that the sample mean estimator attains the Cramer-Rao lower bound and must, therefore, be an MVU estimator in WGN. c D. P. Mandic Advanced Signal Processing 10

11 Efficient estimator concept An estimator which is unbiased and attains the CRLB is said to be efficient in that it is an Minimum Variance Unbiased (MVU) etimator and that it efficiently uses the data. var( ^ θ ) var( θ^ ) CRLB θ^ 2 θ^ 3 θ^ 1 CRLB θ^ 2 ^ θ 3 ^ θ 1 ˆθ 1 is efficient and MVU, ˆθ 2,ˆθ 3 are not θ ˆθ1 may be MVU but is not efficient θ c D. P. Mandic Advanced Signal Processing 11

12 Fisher information The denominator in the CRLB I(θ) = E is referred to as the Fisher information. [ 2 ] lnp(x;θ) Intuitively: the more information available the lower the bound less variance θ 2 Essential properties of an information measure: Ê Non negative Additive for independent observations General CRLB for arbitrary signals in WGN (see the next slide) Ê var(ˆθ) N 1 n=0 σ 2 ( ) 2 s[n;θ] θ Accurate estimators: signals change rapidly with the parameter changes. c D. P. Mandic Advanced Signal Processing 12

13 General case: Arbitrary signal in noise Consider a deterministic signal s[n;θ] observed in WGN, w N(0,σ 2 ) x[n] = s[n;θ]+w[n], n = 0,1,...,N 1 Then the PDF for x parametrised by θ has the form 1 1 N 1 p(x;θ) = (2πσ 2 ) N/2e 2σ 2 n=0 so that ln p(x;θ) θ = 1 N 1 σ 2 2 ln p(x;θ) θ 2 = 1 σ 2 and the Fisher information n=0 N 1 n=0 ( x[n] s[n;θ] ( x[n] s[n;θ] ) s[n;θ] θ [ ( ) 2 s[n;θ] x[n] s[n;θ] }{{} θ 2 E{x[n]}=s[n;θ]} I(θ) = E [ 2 ln p(x;θ) θ 2 ] = 1 σ 2 N 1 n=0 ) 2 ( s[n;θ] θ ( s[n;θ] ) ] 2 θ ) 2 c D. P. Mandic Advanced Signal Processing 13

14 Example 6: Sinusoidal frequency estimation (the CRLB depends both on the unknown parameter f 0 and the data length N) Consider a general sinewave in noise: x[n] = Acos ( 2πf 0 n+φ ) +w[n] If only the freqeuncy f 0 is unknown, then (for normalised frequency) and s[n;f 0 ] = var(ˆf 0 ) A }{{} known A 2 N 1 n=0 cos(2πf 0 n+ Φ }{{} known), 0 < f 0 < 1 2 σ 2 [ 2πn sin(2πf0 n+φ) ] 2 Cramer Rao lower bound 5 x Frequency Note the preferred frequencies, e.g. f 0.03, and that for f 0 {0,1/2} the CRLB Paramet.: N = 10, Φ = 0, SNR=A 2 /σ 2 = 1 c D. P. Mandic Advanced Signal Processing 14

15 Extension to a vector parameter we now have the Fisher Information Matrix I, s.t. [I(θ)] ij = E [ 2 ] ln p(x;θ) θ i θ j Formulation: Estimate a vector parameter θ = [θ 1,θ 2,...,θ p ] T Recall that an unbiased estimator ˆθ is efficient (and therefore an MVU estimator) when it satisfies the conditions of the CRLB It is assumed that the PDF p(x; θ) satisfies the regularity conditions [ ] lnp(x;θ) E = 0, θ θ Then the covariance matrix of any unbiased estimator ˆθ satisfies Cˆθ I 1 (θ) 0 (symbol 0 means that Cˆθ is positive semidefinite) [ ] The Fisher Information Matrix is given by [I(θ)] ij = E 2 lnp(x;θ) θ i θ j An unbiased estimator ˆθ = g(x) exists that satisfies the bound Cˆθ = I 1 (θ) if and only if lnp(x;θ) = I(θ) ( g(x) θ ) θ Ê c D. P. Mandic Advanced Signal Processing 15

16 Extension to a vector parameter: Fisher information Some observations: Elements of the Information Matrix I(θ) are given by [I(θ)] ij = E [ 2 ] lnp(x;θ) θ i θ j where the derivatives are evaluated at the true values of the parameter vector. The CRLB theorem provides a powerful tool for finding MVU estimators for a vector parameter. MVU estimators for linear models are found with the Cramer Rao Lower Bound (CRLB) theorem. Ê c D. P. Mandic Advanced Signal Processing 16

17 Example 7: Sinusoid parameter estimation vector case Consider again a general sinewave s[n] = Acos ( 2πf 0 n+φ ) where A, f 0 and Φ are all unknown. Then, the data model becomes x[n] = A cos(2πf 0 n+φ)+w[n] n = 0,1,...,N 1 where A > 0, 0 < f 0 < 1/2, and w[n] N(0,σ 2 ). Task: Determine CRLB for the parameter vector θ = [A f 0 Φ] T. Solution: The elements of the Fisher Information Matrix become N/2 0 0 I(θ) = 1 0 2A σ 2 2 π 2 N 1 n=0 n2 πa N 1 n=0 n 0 πa N 1 n=0 n NA 2 2 c D. P. Mandic Advanced Signal Processing 17

18 Example 7: Sinusoid parameter estimation continued since Cˆθ = I 1 (θ) (slide 15) make an inverse of the FIM After inversion, the diagonal components yield (where η = A2 2σ 2 is SNR): var(â) 2σ2 N var(ˆf 0 ) 12 (2π) 2 ηn(n 2 1) var(ˆφ) CRLB for Sinusoidal Parameter Estimates at SNR = 3dB (Dashed Lines) and 10dB (Solid Lines) 2(2N 1) ηn(n +1) 0 Amplitude Frequency Phase CRLB [db] Number of data samples, N the variance of the estimated parameters of a sinusoid behaves 1/η and 1/N 3, thus exhibiting strong sensitivity to data length Ê c D. P. Mandic Advanced Signal Processing 18

19 Linear models Generally, it is difficult to determine the MVU estimator. In signal processing, however, a linear data model can often be employed straightforward to determine the MVU estimator. Example 8: Linear model of a straight line in noise x[n] = A+Bn+w[n] n = 0,1,...,N 1 x[n] noisy line where w[n] N(0,σ 2 ), B - slope and A - intercept. A 0 ideal noiseless line n c D. P. Mandic Advanced Signal Processing 19

20 Linear models: Compact notation (Example 8 contd.) This data model can be written more compactly in the matrix notation as where x = and x[0] x[1]. x[n 1] x = Hθ +w or x = Hθ +w = [ x[0],x[1],...,x[n 1] ] T θ = [A B] T w = [w[0],w[1],...,w[n 1]] T w N(0,σ 2 I) I = H = = diag(1,1,...,1) N 1 c D. P. Mandic Advanced Signal Processing 20

21 Linear models: Fisher information matrix NB: p(x;θ) = 1 (2πσ 2 ) N/2e (x Hθ) The CRLB theorem can be used to obtain the MVU estimator for θ The MVU estimator, ˆθ = g(x), will satisfy T (x Hθ) 2σ 2 lnp(x;θ) θ = I(θ) ( g(x) θ ) where I(θ) is the Fisher information matrix, whose elements are [ 2 ln p(x;θ) ] [I] ij = E θ i θ j Applying the Linear Model lnp(x;θ) = [ N θ θ 2 ln(2πσ2 ) 1 ] ( ) T ( ) x Hθ x Hθ 2σ 2 = 1 2σ 2 θ [ Nσ 2 ln(2πσ 2 )+x T x 2x T Hθ +θ T H T Hθ Note that only the quadratic term in θ involves the matrix H ] c D. P. Mandic Advanced Signal Processing 21

22 Linear models: Some useful matrix/vector derivatives the derivation is given in the Lecture Supplement Use the identities (remember that both b T θ and θ T Aθ are scalars) b T θ θ θ T Aθ θ = b xt Hθ θ = 2Aθ θt H T Hθ θ = (x T H) T = H T x = 2H T Hθ (which you should prove for yourself), that is, follow the rules of vector/matrix differentiation. As a rule of thumb, watch for the position of the ( ) T operator Then, for A symmetric: lnp ( x;θ ) θ = 1 σ 2 [ H T x H T Hθ ] c D. P. Mandic Advanced Signal Processing 22

23 Linear models: Cramer-Rao lower bound lnp(x;θ) Find the MVU estimator: θ = I(θ) ( g(x) θ ) Similarly (recall that (H T H) T = H T H), I(θ) = T θ [ ] lnp(x;θ) = 1 H θ σ 2HT Therefore lnp(x;θ) θ = 1 σ 2HT H }{{} I(θ) [ (H T H ) ] 1 H T x θ }{{} g(x) By inspection, the linear estimator ˆθ = g(x) = ( H T H ) 1 H T x provided ( H T H ) 1 is invertible (it is, as H is full rank, with orthogonal rows and columns). The covariance matrix of ˆθ now becomes Cˆθ = I 1 (θ) = σ 2( H T H ) 1 c D. P. Mandic Advanced Signal Processing 23

24 CRLB for linear models continued The MVU estimator for the linear model is efficient it attains the CRLB The columns of H must be linearly independent for ( H T H ) to be invertible Theorem: (Minimum Variance Unbiased Estimator for the Linear Model) If the observed data can be modelled as x = Hθ +w where x is an N 1 vector of observed data H is an N p observation (measurement) matrix of rank p θ is a p 1 unknown parameter vector w is an N 1 additive noise vector N ( 0,σ 2 I ) c D. P. Mandic Advanced Signal Processing 24

25 CRLB for linear models: Theorem Then, the MVU estimator is then given by ˆθ = ( H T H ) 1 H T x for which the covariance matrix Cˆθ = σ2( H T H ) 1 Note that the statistical performance of ˆθ is completely described because ˆθ is a linear transformation of a Gaussian vector x, i.e. ˆθ N (θ,σ 2( H T H ) 1 ) c D. P. Mandic Advanced Signal Processing 25

26 Example 9: Fourier analysis The data model is given by (n = 0,1,...,N 1, w[n] N ( 0,σ 2) ) x[n] = M ( ) 2πkn a k cos + N k=1 M ( ) 2πkn b k sin +w[n] N where the Fourier coefficients, a k and b k, that is the amplitudes of the cosine and sine terms, are to be estimated. k=1 Frequencies are harmonically related, i.e. f 1 = 1 N, and f k = k N. Then, the parameter vector is θ = [a 1,a 2,...,a M,b 1,b 2,...,b M ] T and the observation matrix H ( N }{{} 2M dimensional ) takes the form H = cos 2π N... cos 2πM N sin 2π N... sin 2πM N cos 2πM(N 1) N sin 2π(N 1) N... sin 2πM(N 1) N cos 2π(N 1) N p N 2M c D. P. Mandic Advanced Signal Processing 26

27 Example 9: Fourier analysis continued For H not to be under-determined, it has to satisfy N > p M < N 2, and the columns of H have to be orthogonal. This is obvious if we rewrite the measurement matrix in the form H = [h 1,h 2,...,h 2M ] where h i = h i is the i-th column of H. Then (products of sines and cosines of different frequencies) h T i h j = 0 for i j That is, h i h j, and the columns of matrix H are orthogonal c D. P. Mandic Advanced Signal Processing 27

28 Example 9: Fourier analysis contd. contd. The orthogonality of the columns (for large N) follows from the discrete Fourier transform (DFT) relationships N 1 ( ) ( ) 2πin 2πjn cos cos = N N N 2 δ ij n=0 N 1 n=0 N 1 n=0 where ( ) ( ) 2πin 2πjn sin sin N N ( ) ( ) 2πin 2πjn cos sin N N δ ij = = N 2 δ ij = 0 i,j, s.t. i,j = 1,2,...,M < N 2 { 1, i = j 0, i j In other words: (i) cosiα sinjα, i,j, (ii) cosiα cosjα, i j, (iii) siniα sinjα, i j c D. P. Mandic Advanced Signal Processing 28

29 Example 9: Fourier analysis observation matrix Therefore H T H = ht 1. h T 2M [ h 1,...,h 2M ] = N N N = N 2 I The MVU estimator of the amplitudes of the cosine and sine terms (Fourier coefficients) is therefore given by ˆθ = ( H T H ) 1 H T x ˆθ = 2 N HT x = 2 N h T 1. h T 2M x = that is 2 N ht 1x. 2 N ht 2M x and finally, the estimators of Fourier coefficients become â k = 2 N N 1 n=0 x[n]cos( ) 2πkn N ˆbk = 2 N 1 N n=0 x[n]sin( ) 2πkn N c D. P. Mandic Advanced Signal Processing 29

30 Example 9: Finally Fourier coefficients Therefore, the Fourier analysis is a linear estimator, given by â k = 2 N 1 N n=0 x[n]cos( ) 2πkn N ˆbk = 2 N 1 N n=0 x[n]sin( ) 2πkn N where the a k and b k are discrete Fourier transform coefficients. From CRLB for Linear Model, the covariance matrix of this estimator is Cˆθ = σ2( H T H ) 1 = 2σ 2 N I i) Note that, as ˆθ is a Gaussian random variable and the covariance matrix is diagonal, the amplitude estimates are statistically independent; ii) The orthogonality of the columns of H is fundamental in the computation of the MVU estimator. c D. P. Mandic Advanced Signal Processing 30

31 Example 10: System Identification (SYS ID) Aim: To identify the model of a system (filter coefficients {h}) from input/output data. Assume an FIR filter system model given below u[ n ] 1 Z 1 Z Z 1 h (0) x h (1) x h ( p 1) x Σ Σ p 1 Σ h(k) k =0 u[n k] The input u[n] probes the system, then the output of the FIR filter is given by the convolution x[n] = p 1 k=0 h(k)x[n k] We wish to estimate the filter coefficients [h(0),...,h(p 1)] T In practice, the output is corrupted by additive WGN c D. P. Mandic Advanced Signal Processing 31

32 Example 10: SYS ID data model in noise w N(0,σ 2 ) Data model x[n] = p 1 k=0 Equivalently, in the matrix vector form x[0] x[1]. = x[n 1] } {{ } obs. vec. x that is h(k)u[n k]+w[n] n = 0,1,...,N 1 u[0] u[1] u[0] u[n 1] u[n 2]... u[n p] } {{ } measurement matrix H h(0) h(1). h(p 1) } {{ } coeff. vec. θ x = Hθ +w where w N(0,σ 2 I) The MVU estimator ˆθ = ( H T H ) 1 H T x with Cˆθ = σ 2( H T H ) 1 w[0] + w[1]. w[n 1] } {{ } noise vec. w This representation also lends itslef to state-space modelling c D. P. Mandic Advanced Signal Processing 32

33 Example 10: SYS ID more about H Now, H T H becomes a symmetric Toeplitz autocorrelation matrix, given by r uu (0) r uu (1)... r uu (p 1) H T H = N r uu (1) r uu (0)... r uu (p 2) r uu (p 1) r uu (p 2)... r uu (0) where r uu (k) = 1 N N 1 k n=0 u[n]u[n+k] For H T H to be diagonal, we must have r uu (k) = 0 for k 0, which holds for a pseudorandom (PRN) input sequence. Finally, when H T H = N r uu (0)I then var ( ĥ(i) ) = σ 2 N r uu (0), i = 0,1,...,p 1 c D. P. Mandic Advanced Signal Processing 33

34 Example 10: SYS ID MVU estimator For a PRN sequence, the MVU estimator becomes ˆθ = ( H T H ) 1 H T x Then ĥ(k) = and r ux (k) r uu (0) = 1 N r uu (0) 1 N N 1 n=0 u[n k]x[n] N 1 k n=0 u[n]x[n + k] r uu (0) k = 0,1,...,p 1 Thus, the MVU estimator is the ratio of the input-output cross-correlation to the input autocorrelation. ÊCompare with the Wiener filter in Lecture 7. c D. P. Mandic Advanced Signal Processing 34

35 Theorem: The MVU Estimator for a General Linear Model (GLM) i) Data model x = Hθ + }{{} s +w known signal w N ( 0,C ) ii) Then, the MVU estimator ˆθ = ( H T C 1 H) 1 H T C 1( x s ) iii) with covariance matrix Cˆθ = ( H T CH) 1 c D. P. Mandic Advanced Signal Processing 35

36 What to remember about MVU estimators An estimator is a random variable and as such its performance can only be described statistically by its PDF The use of computer simulations for assessing the performance of an estimator is rarely conclusive Unbiased estimators tend to have symmetric PDFs centred about the true value of θ The minimum mean square error (MMSE) criterion is natural to search for optimal estimators, but it most often leads to unrealisable estimators (those that cannot be written solely as a function of data) Since MSE= Bias 2 + variance, any criterion that depends on bias should be abandoned we need to consider alternative approaches Remedy: Constrain the bias to zero and find an estimator which minimises the variance the minimum variance unbiased (MVU) estim. Minimising the variance of an unbiased estimator also has the effect of concentrating the PDF of the estimation error, ˆθ θ, about zero estimation error less likely to be large c D. P. Mandic Advanced Signal Processing 36

37 A few things about CRLB to remember Even if the MVU estimator exists, there is no turn of the crank procedure to find it. The CRLB sets a lower bound on the variance of any unbiased estimator! This can be extremely useful in several ways: If we find an estimator that achieves the CRLB we known we have found an MVU estimator The CRLB can provide a benchmark against which we can compare the performance of any unbiased estimator The CRLB enables us to rule out impossible estimators. It is physically impossible to find an unbiased estimator that beats the CRLB We may require the estimator to be linear, which is not necessarily a severe restriction, as shown in the example on the estimation of Fourier coefficients c D. P. Mandic Advanced Signal Processing 37

38 Notes: c D. P. Mandic Advanced Signal Processing 38

39 Notes: c D. P. Mandic Advanced Signal Processing 39

40 Notes: c D. P. Mandic Advanced Signal Processing 40

Advanced Signal Processing Introduction to Estimation Theory

Advanced Signal Processing Introduction to Estimation Theory Advanced Signal Processing Introduction to Estimation Theory Danilo Mandic, room 813, ext: 46271 Department of Electrical and Electronic Engineering Imperial College London, UK d.mandic@imperial.ac.uk,

More information

Linear models. x = Hθ + w, where w N(0, σ 2 I) and H R n p. The matrix H is called the observation matrix or design matrix 1.

Linear models. x = Hθ + w, where w N(0, σ 2 I) and H R n p. The matrix H is called the observation matrix or design matrix 1. Linear models As the first approach to estimator design, we consider the class of problems that can be represented by a linear model. In general, finding the MVUE is difficult. But if the linear model

More information

EIE6207: Estimation Theory

EIE6207: Estimation Theory EIE6207: Estimation Theory Man-Wai MAK Dept. of Electronic and Information Engineering, The Hong Kong Polytechnic University enmwmak@polyu.edu.hk http://www.eie.polyu.edu.hk/ mwmak References: Steven M.

More information

Rowan University Department of Electrical and Computer Engineering

Rowan University Department of Electrical and Computer Engineering Rowan University Department of Electrical and Computer Engineering Estimation and Detection Theory Fall 2013 to Practice Exam II This is a closed book exam. There are 8 problems in the exam. The problems

More information

ELEG 5633 Detection and Estimation Minimum Variance Unbiased Estimators (MVUE)

ELEG 5633 Detection and Estimation Minimum Variance Unbiased Estimators (MVUE) 1 ELEG 5633 Detection and Estimation Minimum Variance Unbiased Estimators (MVUE) Jingxian Wu Department of Electrical Engineering University of Arkansas Outline Minimum Variance Unbiased Estimators (MVUE)

More information

Estimation and Detection

Estimation and Detection stimation and Detection Lecture 2: Cramér-Rao Lower Bound Dr. ir. Richard C. Hendriks & Dr. Sundeep P. Chepuri 7//207 Remember: Introductory xample Given a process (DC in noise): x[n]=a + w[n], n=0,,n,

More information

A Few Notes on Fisher Information (WIP)

A Few Notes on Fisher Information (WIP) A Few Notes on Fisher Information (WIP) David Meyer dmm@{-4-5.net,uoregon.edu} Last update: April 30, 208 Definitions There are so many interesting things about Fisher Information and its theoretical properties

More information

LINEAR MODELS IN STATISTICAL SIGNAL PROCESSING

LINEAR MODELS IN STATISTICAL SIGNAL PROCESSING TERM PAPER REPORT ON LINEAR MODELS IN STATISTICAL SIGNAL PROCESSING ANIMA MISHRA SHARMA, Y8104001 BISHWAJIT SHARMA, Y8104015 Course Instructor DR. RAJESH HEDGE Electrical Engineering Department IIT Kanpur

More information

PATTERN RECOGNITION AND MACHINE LEARNING

PATTERN RECOGNITION AND MACHINE LEARNING PATTERN RECOGNITION AND MACHINE LEARNING Slide Set 2: Estimation Theory January 2018 Heikki Huttunen heikki.huttunen@tut.fi Department of Signal Processing Tampere University of Technology Classical Estimation

More information

Chapter 8: Least squares (beginning of chapter)

Chapter 8: Least squares (beginning of chapter) Chapter 8: Least squares (beginning of chapter) Least Squares So far, we have been trying to determine an estimator which was unbiased and had minimum variance. Next we ll consider a class of estimators

More information

ECE 275A Homework 6 Solutions

ECE 275A Homework 6 Solutions ECE 275A Homework 6 Solutions. The notation used in the solutions for the concentration (hyper) ellipsoid problems is defined in the lecture supplement on concentration ellipsoids. Note that θ T Σ θ =

More information

Module 2. Random Processes. Version 2, ECE IIT, Kharagpur

Module 2. Random Processes. Version 2, ECE IIT, Kharagpur Module Random Processes Version, ECE IIT, Kharagpur Lesson 9 Introduction to Statistical Signal Processing Version, ECE IIT, Kharagpur After reading this lesson, you will learn about Hypotheses testing

More information

Variations. ECE 6540, Lecture 10 Maximum Likelihood Estimation

Variations. ECE 6540, Lecture 10 Maximum Likelihood Estimation Variations ECE 6540, Lecture 10 Last Time BLUE (Best Linear Unbiased Estimator) Formulation Advantages Disadvantages 2 The BLUE A simplification Assume the estimator is a linear system For a single parameter

More information

Estimation Theory Fredrik Rusek. Chapters

Estimation Theory Fredrik Rusek. Chapters Estimation Theory Fredrik Rusek Chapters 3.5-3.10 Recap We deal with unbiased estimators of deterministic parameters Performance of an estimator is measured by the variance of the estimate (due to the

More information

Detection & Estimation Lecture 1

Detection & Estimation Lecture 1 Detection & Estimation Lecture 1 Intro, MVUE, CRLB Xiliang Luo General Course Information Textbooks & References Fundamentals of Statistical Signal Processing: Estimation Theory/Detection Theory, Steven

More information

DETECTION theory deals primarily with techniques for

DETECTION theory deals primarily with techniques for ADVANCED SIGNAL PROCESSING SE Optimum Detection of Deterministic and Random Signals Stefan Tertinek Graz University of Technology turtle@sbox.tugraz.at Abstract This paper introduces various methods for

More information

Detection theory. H 0 : x[n] = w[n]

Detection theory. H 0 : x[n] = w[n] Detection Theory Detection theory A the last topic of the course, we will briefly consider detection theory. The methods are based on estimation theory and attempt to answer questions such as Is a signal

More information

Detection & Estimation Lecture 1

Detection & Estimation Lecture 1 Detection & Estimation Lecture 1 Intro, MVUE, CRLB Xiliang Luo General Course Information Textbooks & References Fundamentals of Statistical Signal Processing: Estimation Theory/Detection Theory, Steven

More information

ECE531 Lecture 10b: Maximum Likelihood Estimation

ECE531 Lecture 10b: Maximum Likelihood Estimation ECE531 Lecture 10b: Maximum Likelihood Estimation D. Richard Brown III Worcester Polytechnic Institute 05-Apr-2011 Worcester Polytechnic Institute D. Richard Brown III 05-Apr-2011 1 / 23 Introduction So

More information

A TERM PAPER REPORT ON KALMAN FILTER

A TERM PAPER REPORT ON KALMAN FILTER A TERM PAPER REPORT ON KALMAN FILTER By B. Sivaprasad (Y8104059) CH. Venkata Karunya (Y8104066) Department of Electrical Engineering Indian Institute of Technology, Kanpur Kanpur-208016 SCALAR KALMAN FILTER

More information

Estimation, Detection, and Identification

Estimation, Detection, and Identification Estimation, Detection, and Identification Graduate Course on the CMU/Portugal ECE PhD Program Spring 2008/2009 Chapter 5 Best Linear Unbiased Estimators Instructor: Prof. Paulo Jorge Oliveira pjcro @ isr.ist.utl.pt

More information

10. Linear Models and Maximum Likelihood Estimation

10. Linear Models and Maximum Likelihood Estimation 10. Linear Models and Maximum Likelihood Estimation ECE 830, Spring 2017 Rebecca Willett 1 / 34 Primary Goal General problem statement: We observe y i iid pθ, θ Θ and the goal is to determine the θ that

More information

Terminology Suppose we have N observations {x(n)} N 1. Estimators as Random Variables. {x(n)} N 1

Terminology Suppose we have N observations {x(n)} N 1. Estimators as Random Variables. {x(n)} N 1 Estimation Theory Overview Properties Bias, Variance, and Mean Square Error Cramér-Rao lower bound Maximum likelihood Consistency Confidence intervals Properties of the mean estimator Properties of the

More information

Classical Estimation Topics

Classical Estimation Topics Classical Estimation Topics Namrata Vaswani, Iowa State University February 25, 2014 This note fills in the gaps in the notes already provided (l0.pdf, l1.pdf, l2.pdf, l3.pdf, LeastSquares.pdf). 1 Min

More information

10-704: Information Processing and Learning Fall Lecture 24: Dec 7

10-704: Information Processing and Learning Fall Lecture 24: Dec 7 0-704: Information Processing and Learning Fall 206 Lecturer: Aarti Singh Lecture 24: Dec 7 Note: These notes are based on scribed notes from Spring5 offering of this course. LaTeX template courtesy of

More information

Lecture 4 - Spectral Estimation

Lecture 4 - Spectral Estimation Lecture 4 - Spectral Estimation The Discrete Fourier Transform The Discrete Fourier Transform (DFT) is the equivalent of the continuous Fourier Transform for signals known only at N instants separated

More information

2 Statistical Estimation: Basic Concepts

2 Statistical Estimation: Basic Concepts Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof. N. Shimkin 2 Statistical Estimation:

More information

EIE6207: Maximum-Likelihood and Bayesian Estimation

EIE6207: Maximum-Likelihood and Bayesian Estimation EIE6207: Maximum-Likelihood and Bayesian Estimation Man-Wai MAK Dept. of Electronic and Information Engineering, The Hong Kong Polytechnic University enmwmak@polyu.edu.hk http://www.eie.polyu.edu.hk/ mwmak

More information

Estimators as Random Variables

Estimators as Random Variables Estimation Theory Overview Properties Bias, Variance, and Mean Square Error Cramér-Rao lower bound Maimum likelihood Consistency Confidence intervals Properties of the mean estimator Introduction Up until

More information

ECE531 Lecture 8: Non-Random Parameter Estimation

ECE531 Lecture 8: Non-Random Parameter Estimation ECE531 Lecture 8: Non-Random Parameter Estimation D. Richard Brown III Worcester Polytechnic Institute 19-March-2009 Worcester Polytechnic Institute D. Richard Brown III 19-March-2009 1 / 25 Introduction

More information

Lecture 8: Signal Detection and Noise Assumption

Lecture 8: Signal Detection and Noise Assumption ECE 830 Fall 0 Statistical Signal Processing instructor: R. Nowak Lecture 8: Signal Detection and Noise Assumption Signal Detection : X = W H : X = S + W where W N(0, σ I n n and S = [s, s,..., s n ] T

More information

Module 1 - Signal estimation

Module 1 - Signal estimation , Arraial do Cabo, 2009 Module 1 - Signal estimation Sérgio M. Jesus (sjesus@ualg.pt) Universidade do Algarve, PT-8005-139 Faro, Portugal www.siplab.fct.ualg.pt February 2009 Outline of Module 1 Parameter

More information

Estimation Tasks. Short Course on Image Quality. Matthew A. Kupinski. Introduction

Estimation Tasks. Short Course on Image Quality. Matthew A. Kupinski. Introduction Estimation Tasks Short Course on Image Quality Matthew A. Kupinski Introduction Section 13.3 in B&M Keep in mind the similarities between estimation and classification Image-quality is a statistical concept

More information

PATTERN RECOGNITION AND MACHINE LEARNING

PATTERN RECOGNITION AND MACHINE LEARNING PATTERN RECOGNITION AND MACHINE LEARNING Slide Set 3: Detection Theory January 2018 Heikki Huttunen heikki.huttunen@tut.fi Department of Signal Processing Tampere University of Technology Detection theory

More information

26. Filtering. ECE 830, Spring 2014

26. Filtering. ECE 830, Spring 2014 26. Filtering ECE 830, Spring 2014 1 / 26 Wiener Filtering Wiener filtering is the application of LMMSE estimation to recovery of a signal in additive noise under wide sense sationarity assumptions. Problem

More information

Estimation Theory for Engineers

Estimation Theory for Engineers Estimation Theory for Engineers Roberto Togneri 30th August 2005 Applications Modern estimation theory can be found at the heart of many electronic signal processing systems designed to extract information.

More information

F & B Approaches to a simple model

F & B Approaches to a simple model A6523 Signal Modeling, Statistical Inference and Data Mining in Astrophysics Spring 215 http://www.astro.cornell.edu/~cordes/a6523 Lecture 11 Applications: Model comparison Challenges in large-scale surveys

More information

Technique for Numerical Computation of Cramér-Rao Bound using MATLAB

Technique for Numerical Computation of Cramér-Rao Bound using MATLAB Technique for Numerical Computation of Cramér-Rao Bound using MATLAB Hing Cheung So http://www.ee.cityu.edu.hk/~hcso Department of Electronic Engineering City University of Hong Kong H. C. So Page 1 The

More information

EECE Adaptive Control

EECE Adaptive Control EECE 574 - Adaptive Control Basics of System Identification Guy Dumont Department of Electrical and Computer Engineering University of British Columbia January 2010 Guy Dumont (UBC) EECE574 - Basics of

More information

INTRODUCTION Noise is present in many situations of daily life for ex: Microphones will record noise and speech. Goal: Reconstruct original signal Wie

INTRODUCTION Noise is present in many situations of daily life for ex: Microphones will record noise and speech. Goal: Reconstruct original signal Wie WIENER FILTERING Presented by N.Srikanth(Y8104060), M.Manikanta PhaniKumar(Y8104031). INDIAN INSTITUTE OF TECHNOLOGY KANPUR Electrical Engineering dept. INTRODUCTION Noise is present in many situations

More information

Detecting Parametric Signals in Noise Having Exactly Known Pdf/Pmf

Detecting Parametric Signals in Noise Having Exactly Known Pdf/Pmf Detecting Parametric Signals in Noise Having Exactly Known Pdf/Pmf Reading: Ch. 5 in Kay-II. (Part of) Ch. III.B in Poor. EE 527, Detection and Estimation Theory, # 5c Detecting Parametric Signals in Noise

More information

Mathematical statistics

Mathematical statistics October 18 th, 2018 Lecture 16: Midterm review Countdown to mid-term exam: 7 days Week 1 Chapter 1: Probability review Week 2 Week 4 Week 7 Chapter 6: Statistics Chapter 7: Point Estimation Chapter 8:

More information

Estimation Theory Fredrik Rusek. Chapters 6-7

Estimation Theory Fredrik Rusek. Chapters 6-7 Estimation Theory Fredrik Rusek Chapters 6-7 All estimation problems Summary All estimation problems Summary Efficient estimator exists All estimation problems Summary MVU estimator exists Efficient estimator

More information

ECE534, Spring 2018: Solutions for Problem Set #3

ECE534, Spring 2018: Solutions for Problem Set #3 ECE534, Spring 08: Solutions for Problem Set #3 Jointly Gaussian Random Variables and MMSE Estimation Suppose that X, Y are jointly Gaussian random variables with µ X = µ Y = 0 and σ X = σ Y = Let their

More information

Estimation Theory. as Θ = (Θ 1,Θ 2,...,Θ m ) T. An estimator

Estimation Theory. as Θ = (Θ 1,Θ 2,...,Θ m ) T. An estimator Estimation Theory Estimation theory deals with finding numerical values of interesting parameters from given set of data. We start with formulating a family of models that could describe how the data were

More information

3. ESTIMATION OF SIGNALS USING A LEAST SQUARES TECHNIQUE

3. ESTIMATION OF SIGNALS USING A LEAST SQUARES TECHNIQUE 3. ESTIMATION OF SIGNALS USING A LEAST SQUARES TECHNIQUE 3.0 INTRODUCTION The purpose of this chapter is to introduce estimators shortly. More elaborated courses on System Identification, which are given

More information

Brief Review on Estimation Theory

Brief Review on Estimation Theory Brief Review on Estimation Theory K. Abed-Meraim ENST PARIS, Signal and Image Processing Dept. abed@tsi.enst.fr This presentation is essentially based on the course BASTA by E. Moulines Brief review on

More information

Examination with solution suggestions SSY130 Applied Signal Processing

Examination with solution suggestions SSY130 Applied Signal Processing Examination with solution suggestions SSY3 Applied Signal Processing Jan 8, 28 Rules Allowed aids at exam: L. Råde and B. Westergren, Mathematics Handbook (any edition, including the old editions called

More information

Mathematical statistics

Mathematical statistics October 4 th, 2018 Lecture 12: Information Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation Chapter

More information

SGN Advanced Signal Processing: Lecture 8 Parameter estimation for AR and MA models. Model order selection

SGN Advanced Signal Processing: Lecture 8 Parameter estimation for AR and MA models. Model order selection SG 21006 Advanced Signal Processing: Lecture 8 Parameter estimation for AR and MA models. Model order selection Ioan Tabus Department of Signal Processing Tampere University of Technology Finland 1 / 28

More information

557: MATHEMATICAL STATISTICS II BIAS AND VARIANCE

557: MATHEMATICAL STATISTICS II BIAS AND VARIANCE 557: MATHEMATICAL STATISTICS II BIAS AND VARIANCE An estimator, T (X), of θ can be evaluated via its statistical properties. Typically, two aspects are considered: Expectation Variance either in terms

More information

A NEW BAYESIAN LOWER BOUND ON THE MEAN SQUARE ERROR OF ESTIMATORS. Koby Todros and Joseph Tabrikian

A NEW BAYESIAN LOWER BOUND ON THE MEAN SQUARE ERROR OF ESTIMATORS. Koby Todros and Joseph Tabrikian 16th European Signal Processing Conference EUSIPCO 008) Lausanne Switzerland August 5-9 008 copyright by EURASIP A NEW BAYESIAN LOWER BOUND ON THE MEAN SQUARE ERROR OF ESTIMATORS Koby Todros and Joseph

More information

Advanced Digital Signal Processing -Introduction

Advanced Digital Signal Processing -Introduction Advanced Digital Signal Processing -Introduction LECTURE-2 1 AP9211- ADVANCED DIGITAL SIGNAL PROCESSING UNIT I DISCRETE RANDOM SIGNAL PROCESSING Discrete Random Processes- Ensemble Averages, Stationary

More information

3F1 Random Processes Examples Paper (for all 6 lectures)

3F1 Random Processes Examples Paper (for all 6 lectures) 3F Random Processes Examples Paper (for all 6 lectures). Three factories make the same electrical component. Factory A supplies half of the total number of components to the central depot, while factories

More information

MA 575 Linear Models: Cedric E. Ginestet, Boston University Midterm Review Week 7

MA 575 Linear Models: Cedric E. Ginestet, Boston University Midterm Review Week 7 MA 575 Linear Models: Cedric E. Ginestet, Boston University Midterm Review Week 7 1 Random Vectors Let a 0 and y be n 1 vectors, and let A be an n n matrix. Here, a 0 and A are non-random, whereas y is

More information

ECE 275A Homework 7 Solutions

ECE 275A Homework 7 Solutions ECE 275A Homework 7 Solutions Solutions 1. For the same specification as in Homework Problem 6.11 we want to determine an estimator for θ using the Method of Moments (MOM). In general, the MOM estimator

More information

Identification of ARX, OE, FIR models with the least squares method

Identification of ARX, OE, FIR models with the least squares method Identification of ARX, OE, FIR models with the least squares method CHEM-E7145 Advanced Process Control Methods Lecture 2 Contents Identification of ARX model with the least squares minimizing the equation

More information

ECE531 Homework Assignment Number 7 Solution

ECE531 Homework Assignment Number 7 Solution ECE53 Homework Assignment Number 7 Solution Due by 8:50pm on Wednesday 30-Mar-20 Make sure your reasoning and work are clear to receive full credit for each problem.. 4 points. Kay I: 2.6. Solution: I

More information

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Cheng Soon Ong & Christian Walder. Canberra February June 2018 Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 (Many figures from C. M. Bishop, "Pattern Recognition and ") 1of 254 Part V

More information

DS-GA 1002 Lecture notes 10 November 23, Linear models

DS-GA 1002 Lecture notes 10 November 23, Linear models DS-GA 2 Lecture notes November 23, 2 Linear functions Linear models A linear model encodes the assumption that two quantities are linearly related. Mathematically, this is characterized using linear functions.

More information

HT Introduction. P(X i = x i ) = e λ λ x i

HT Introduction. P(X i = x i ) = e λ λ x i MODS STATISTICS Introduction. HT 2012 Simon Myers, Department of Statistics (and The Wellcome Trust Centre for Human Genetics) myers@stats.ox.ac.uk We will be concerned with the mathematical framework

More information

System Identification, Lecture 4

System Identification, Lecture 4 System Identification, Lecture 4 Kristiaan Pelckmans (IT/UU, 2338) Course code: 1RT880, Report code: 61800 - Spring 2012 F, FRI Uppsala University, Information Technology 30 Januari 2012 SI-2012 K. Pelckmans

More information

System Identification, Lecture 4

System Identification, Lecture 4 System Identification, Lecture 4 Kristiaan Pelckmans (IT/UU, 2338) Course code: 1RT880, Report code: 61800 - Spring 2016 F, FRI Uppsala University, Information Technology 13 April 2016 SI-2016 K. Pelckmans

More information

Lecture 8: Information Theory and Statistics

Lecture 8: Information Theory and Statistics Lecture 8: Information Theory and Statistics Part II: Hypothesis Testing and I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 23, 2015 1 / 50 I-Hsiang

More information

Detection and Estimation Theory

Detection and Estimation Theory Detection and Estimation Theory Instructor: Prof. Namrata Vaswani Dept. of Electrical and Computer Engineering Iowa State University http://www.ece.iastate.edu/ namrata Slide 1 What is Estimation and Detection

More information

A6523 Signal Modeling, Statistical Inference and Data Mining in Astrophysics Spring 2013

A6523 Signal Modeling, Statistical Inference and Data Mining in Astrophysics Spring 2013 A6523 Signal Modeling, Statistical Inference and Data Mining in Astrophysics Spring 2013 Lecture 26 Localization/Matched Filtering (continued) Prewhitening Lectures next week: Reading Bases, principal

More information

A6523 Modeling, Inference, and Mining Jim Cordes, Cornell University

A6523 Modeling, Inference, and Mining Jim Cordes, Cornell University A6523 Modeling, Inference, and Mining Jim Cordes, Cornell University Lecture 19 Modeling Topics plan: Modeling (linear/non- linear least squares) Bayesian inference Bayesian approaches to spectral esbmabon;

More information

Introduction to Biomedical Engineering

Introduction to Biomedical Engineering Introduction to Biomedical Engineering Biosignal processing Kung-Bin Sung 6/11/2007 1 Outline Chapter 10: Biosignal processing Characteristics of biosignals Frequency domain representation and analysis

More information

Linear models. Linear models are computationally convenient and remain widely used in. applied econometric research

Linear models. Linear models are computationally convenient and remain widely used in. applied econometric research Linear models Linear models are computationally convenient and remain widely used in applied econometric research Our main focus in these lectures will be on single equation linear models of the form y

More information

Linear Stochastic Models. Special Types of Random Processes: AR, MA, and ARMA. Digital Signal Processing

Linear Stochastic Models. Special Types of Random Processes: AR, MA, and ARMA. Digital Signal Processing Linear Stochastic Models Special Types of Random Processes: AR, MA, and ARMA Digital Signal Processing Department of Electrical and Electronic Engineering, Imperial College d.mandic@imperial.ac.uk c Danilo

More information

Introduction to Simple Linear Regression

Introduction to Simple Linear Regression Introduction to Simple Linear Regression Yang Feng http://www.stat.columbia.edu/~yangfeng Yang Feng (Columbia University) Introduction to Simple Linear Regression 1 / 68 About me Faculty in the Department

More information

Concentration Ellipsoids

Concentration Ellipsoids Concentration Ellipsoids ECE275A Lecture Supplement Fall 2008 Kenneth Kreutz Delgado Electrical and Computer Engineering Jacobs School of Engineering University of California, San Diego VERSION LSECE275CE

More information

Classifying and building DLMs

Classifying and building DLMs Chapter 3 Classifying and building DLMs In this chapter we will consider a special form of DLM for which the transition matrix F k and the observation matrix H k are constant in time, ie, F k = F and H

More information

Engineering Part IIB: Module 4F10 Statistical Pattern Processing Lecture 5: Single Layer Perceptrons & Estimating Linear Classifiers

Engineering Part IIB: Module 4F10 Statistical Pattern Processing Lecture 5: Single Layer Perceptrons & Estimating Linear Classifiers Engineering Part IIB: Module 4F0 Statistical Pattern Processing Lecture 5: Single Layer Perceptrons & Estimating Linear Classifiers Phil Woodland: pcw@eng.cam.ac.uk Michaelmas 202 Engineering Part IIB:

More information

COMPLEX CONSTRAINED CRB AND ITS APPLICATION TO SEMI-BLIND MIMO AND OFDM CHANNEL ESTIMATION. Aditya K. Jagannatham and Bhaskar D.

COMPLEX CONSTRAINED CRB AND ITS APPLICATION TO SEMI-BLIND MIMO AND OFDM CHANNEL ESTIMATION. Aditya K. Jagannatham and Bhaskar D. COMPLEX CONSTRAINED CRB AND ITS APPLICATION TO SEMI-BLIND MIMO AND OFDM CHANNEL ESTIMATION Aditya K Jagannatham and Bhaskar D Rao University of California, SanDiego 9500 Gilman Drive, La Jolla, CA 92093-0407

More information

3.0.1 Multivariate version and tensor product of experiments

3.0.1 Multivariate version and tensor product of experiments ECE598: Information-theoretic methods in high-dimensional statistics Spring 2016 Lecture 3: Minimax risk of GLM and four extensions Lecturer: Yihong Wu Scribe: Ashok Vardhan, Jan 28, 2016 [Ed. Mar 24]

More information

Statistics. Lecture 2 August 7, 2000 Frank Porter Caltech. The Fundamentals; Point Estimation. Maximum Likelihood, Least Squares and All That

Statistics. Lecture 2 August 7, 2000 Frank Porter Caltech. The Fundamentals; Point Estimation. Maximum Likelihood, Least Squares and All That Statistics Lecture 2 August 7, 2000 Frank Porter Caltech The plan for these lectures: The Fundamentals; Point Estimation Maximum Likelihood, Least Squares and All That What is a Confidence Interval? Interval

More information

DOA Estimation using MUSIC and Root MUSIC Methods

DOA Estimation using MUSIC and Root MUSIC Methods DOA Estimation using MUSIC and Root MUSIC Methods EE602 Statistical signal Processing 4/13/2009 Presented By: Chhavipreet Singh(Y515) Siddharth Sahoo(Y5827447) 2 Table of Contents 1 Introduction... 3 2

More information

Adaptive Systems Homework Assignment 1

Adaptive Systems Homework Assignment 1 Signal Processing and Speech Communication Lab. Graz University of Technology Adaptive Systems Homework Assignment 1 Name(s) Matr.No(s). The analytical part of your homework (your calculation sheets) as

More information

Cramer-Rao Lower Bound Computation Via the Characteristic Function

Cramer-Rao Lower Bound Computation Via the Characteristic Function Cramer-Rao ower Bound Computation Via the Characteristic Function Steven Kay, Fellow, IEEE, and Cuichun Xu Abstract The Cramer-Rao ower Bound is widely used in statistical signal processing as a benchmark

More information

Machine Learning. A Bayesian and Optimization Perspective. Academic Press, Sergios Theodoridis 1. of Athens, Athens, Greece.

Machine Learning. A Bayesian and Optimization Perspective. Academic Press, Sergios Theodoridis 1. of Athens, Athens, Greece. Machine Learning A Bayesian and Optimization Perspective Academic Press, 2015 Sergios Theodoridis 1 1 Dept. of Informatics and Telecommunications, National and Kapodistrian University of Athens, Athens,

More information

Estimation theory. Parametric estimation. Properties of estimators. Minimum variance estimator. Cramer-Rao bound. Maximum likelihood estimators

Estimation theory. Parametric estimation. Properties of estimators. Minimum variance estimator. Cramer-Rao bound. Maximum likelihood estimators Estimation theory Parametric estimation Properties of estimators Minimum variance estimator Cramer-Rao bound Maximum likelihood estimators Confidence intervals Bayesian estimation 1 Random Variables Let

More information

ANALYTICAL MATHEMATICS FOR APPLICATIONS 2018 LECTURE NOTES 3

ANALYTICAL MATHEMATICS FOR APPLICATIONS 2018 LECTURE NOTES 3 ANALYTICAL MATHEMATICS FOR APPLICATIONS 2018 LECTURE NOTES 3 ISSUED 24 FEBRUARY 2018 1 Gaussian elimination Let A be an (m n)-matrix Consider the following row operations on A (1) Swap the positions any

More information

Statistics: Learning models from data

Statistics: Learning models from data DS-GA 1002 Lecture notes 5 October 19, 2015 Statistics: Learning models from data Learning models from data that are assumed to be generated probabilistically from a certain unknown distribution is a crucial

More information

Multiple Random Variables

Multiple Random Variables Multiple Random Variables Joint Probability Density Let X and Y be two random variables. Their joint distribution function is F ( XY x, y) P X x Y y. F XY ( ) 1, < x

More information

Bayesian Linear Regression [DRAFT - In Progress]

Bayesian Linear Regression [DRAFT - In Progress] Bayesian Linear Regression [DRAFT - In Progress] David S. Rosenberg Abstract Here we develop some basics of Bayesian linear regression. Most of the calculations for this document come from the basic theory

More information

Fast Angular Synchronization for Phase Retrieval via Incomplete Information

Fast Angular Synchronization for Phase Retrieval via Incomplete Information Fast Angular Synchronization for Phase Retrieval via Incomplete Information Aditya Viswanathan a and Mark Iwen b a Department of Mathematics, Michigan State University; b Department of Mathematics & Department

More information

Motivation CRLB and Regularity Conditions CRLLB A New Bound Examples Conclusions

Motivation CRLB and Regularity Conditions CRLLB A New Bound Examples Conclusions A Survey of Some Recent Results on the CRLB for Parameter Estimation and its Extension Yaakov 1 1 Electrical and Computer Engineering Department University of Connecticut, Storrs, CT A Survey of Some Recent

More information

Parameter estimation! and! forecasting! Cristiano Porciani! AIfA, Uni-Bonn!

Parameter estimation! and! forecasting! Cristiano Porciani! AIfA, Uni-Bonn! Parameter estimation! and! forecasting! Cristiano Porciani! AIfA, Uni-Bonn! Questions?! C. Porciani! Estimation & forecasting! 2! Cosmological parameters! A branch of modern cosmological research focuses

More information

Lecture Notes 7 Stationary Random Processes. Strict-Sense and Wide-Sense Stationarity. Autocorrelation Function of a Stationary Process

Lecture Notes 7 Stationary Random Processes. Strict-Sense and Wide-Sense Stationarity. Autocorrelation Function of a Stationary Process Lecture Notes 7 Stationary Random Processes Strict-Sense and Wide-Sense Stationarity Autocorrelation Function of a Stationary Process Power Spectral Density Continuity and Integration of Random Processes

More information

Part IB Statistics. Theorems with proof. Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua. Lent 2015

Part IB Statistics. Theorems with proof. Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua. Lent 2015 Part IB Statistics Theorems with proof Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly)

More information

ECE 275A Homework #3 Solutions

ECE 275A Homework #3 Solutions ECE 75A Homework #3 Solutions. Proof of (a). Obviously Ax = 0 y, Ax = 0 for all y. To show sufficiency, note that if y, Ax = 0 for all y, then it must certainly be true for the particular value of y =

More information

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it

More information

Numerical Methods of Approximation

Numerical Methods of Approximation Contents 31 Numerical Methods of Approximation 31.1 Polynomial Approximations 2 31.2 Numerical Integration 28 31.3 Numerical Differentiation 58 31.4 Nonlinear Equations 67 Learning outcomes In this Workbook

More information

Timing Recovery at Low SNR Cramer-Rao bound, and outperforming the PLL

Timing Recovery at Low SNR Cramer-Rao bound, and outperforming the PLL T F T I G E O R G A I N S T I T U T E O H E O F E A L P R O G R ESS S A N D 1 8 8 5 S E R V L O G Y I C E E C H N O Timing Recovery at Low SNR Cramer-Rao bound, and outperforming the PLL Aravind R. Nayak

More information

Communications and Signal Processing Spring 2017 MSE Exam

Communications and Signal Processing Spring 2017 MSE Exam Communications and Signal Processing Spring 2017 MSE Exam Please obtain your Test ID from the following table. You must write your Test ID and name on each of the pages of this exam. A page with missing

More information

Rethinking Biased Estimation: Improving Maximum Likelihood and the Cramér Rao Bound

Rethinking Biased Estimation: Improving Maximum Likelihood and the Cramér Rao Bound Foundations and Trends R in Signal Processing Vol. 1, No. 4 (2007) 305 449 c 2008 Y. C. Eldar DOI: 10.1561/2000000008 Rethinking Biased Estimation: Improving Maximum Likelihood and the Cramér Rao Bound

More information

Lecture 3 September 1

Lecture 3 September 1 STAT 383C: Statistical Modeling I Fall 2016 Lecture 3 September 1 Lecturer: Purnamrita Sarkar Scribe: Giorgio Paulon, Carlos Zanini Disclaimer: These scribe notes have been slightly proofread and may have

More information

Example: An experiment can either result in success or failure with probability θ and (1 θ) respectively. The experiment is performed independently

Example: An experiment can either result in success or failure with probability θ and (1 θ) respectively. The experiment is performed independently Chapter 3 Sufficient statistics and variance reduction Let X 1,X 2,...,X n be a random sample from a certain distribution with p.m/d.f fx θ. A function T X 1,X 2,...,X n = T X of these observations is

More information

Efficient Monte Carlo computation of Fisher information matrix using prior information

Efficient Monte Carlo computation of Fisher information matrix using prior information Efficient Monte Carlo computation of Fisher information matrix using prior information Sonjoy Das, UB James C. Spall, APL/JHU Roger Ghanem, USC SIAM Conference on Data Mining Anaheim, California, USA April

More information