III.C - Linear Transformations: Optimal Filtering

Size: px
Start display at page:

Download "III.C - Linear Transformations: Optimal Filtering"

Transcription

1 1 III.C - Linear Transformations: Optimal Filtering FIR Wiener Filter [p. 3] Mean square signal estimation principles [p. 4] Orthogonality principle [p. 7] FIR Wiener filtering concepts [p. 8] Filter coefficients computation [p. 10] MMSE computation [p. 14] Examples [p. 38] FIR Weiner filter and error surfaces [p. 44] Applications to channel equalization [p. 49] Applications to noise cancellation [p. 54] Applications to spatial filtering (antennas) [p. 60] Appendices [p. 64] References

2 Optimal Filtering Problem: How to estimate one signal from another. In many applications desired signal is not observable directly (convolved with another signal or distorted by noise). Examples: Information signal transmitted over channel gets corrupted with noise. Image recorded by system is subject to distortions. Directional antenna array is vulnerable to string jammers in other directions due to sidelobe leakage, etc. In this course: Emphasis on least square techniques to estimate/recover signal (i.e., case 2 ). Other types of norms could be used to solve problems 06/20/14 EC3410.SuFY14MPF - Section IIIC - FIR 2

3 3 Mean Square Signal Estimation Transmitted signal Possible procedure: s Mean square estimation, i.e., Distorted received signal x { 2 ( ) } minimize ξ = E s d x d ( x) = E[ s x] leads to (proof given in Appendix A) Optimal processor best estimate of transmitted signal s (as a function of received signal x) conditional mean!, usually nonlinear in x [exception when x and s are jointly normal Gauss Markov theorem] Complicated to solve, Restriction to Linear Mean Square Estimator (LMS), estimator of s is forced to be a linear function of measurements x: H d = h x Solution via Wiener Hopf equations using orthogonality principle d

4 Orthogonality Principle Use LMS Criterion: estimate s by where weights {h i } minimize MS error: σ = { 2 ( ) } 2 e E s d x Theorem: Let error e = s - d h { *} { * } d = 2 e H h x minimizes the MSE quantity σ if h is chosen such that E ex = E x e = 0, = 1,, N i i i i.e., the error e is orthogonal to the observations x, i = 1..., N used to compute the filter output. i Corollary: minimum MSE obtained: e min the minimum error obtained for the optimum filter vector. (Proof given in Appendix B) σ = { } E se 2 * where e is 06/20/14 EC3410.SuFY14MPF - Section IIIC - FIR 4

5 P = 2 dn ( ) = hxn ( ) + hxn ( 1) 0 1 s x(n) x(n-1) 06/20/14 EC3410.SuFY14MPF - Section IIIC - FIR 5

6 6 signal s(n) + + noise w(n) x(n) dn ( ) filter n 0 x n d(n) d(n) Typical Wiener Filtering Problems Form of Desired Problem Observations Signal Filtering of x(n) = s(n)+w(n) d(n) = s(n) signal in noise n 0 n 0 n 0 d(n) n q n p q n 1 n+p n d(n) n+1 Prediction of x(n) = s(n)+w(n) d(n) = s(n+p); signal in noise p > 0 Smoothing of x(n) = s(n)+w(n) d(n) = s(n q); signal in noise q > 0 Linear x(n) = s(n 1) d(n) = s(n) prediction

7 FIR Wiener Filtering Concepts Filter criterion used: minimization of mean square error between d(n) and dˆ ( n). d(n) What are we doing here? x(n) d a f n filter - s(n) signal noise w(n) We want to design a filter (in the generic sense can be: filter, smoother, predictor) so that: a f a f a f P 1 * k= 0 d n = h k x n k How d(n) is defined specifies the operation done: filtering: d(n)=s(n) predicting: d(n)=s(n+p) smoothing: d(n)=s(n-p) 06/20/14 EC3410.SuFY14MPF - Section IIIC - FIR 7 e(n)

8 8 How to find h k? Minimize the MSE: { 2 ( ) ˆ ( ) } d n E d n P 1 k = 0 ( k) hx n = h T [, ], ( ), x( n P 1) h= h0 hp 1 x= x n + * k Wiener filter is a linear filter orthogonality principle applies * ( ) ( ) { } E x n i e n = 0, i = 0,..., P 1 * P 1 * E x( n i) d( n) hk x( n k) = 0, i = 0,..., P 1 k = 0 P 1 ( ) ( ) r i hr k i = 0, i= 0,..., P 1 xd k x k = 0 H x T

9 i = 0 i = 1 Matrix form: r 0 R 0 R 1 R P 1 h ( ) P 1 * dx k x k = 0 ( ) ( ) r i = hr k i, i= 0,..., P 1 ( ) ( ) ( ) * dx x x x 0 * r ( 1) R ( 1) ( 0) ( 2) dx x Rx Rx P h 1 = R ( 1) ( 0) x P+ Rx hp 1 Note: different notation than in [Therrien, section 7.3]! 06/20/14 EC3410.SuFY14MPF - Section IIIC - FIR 9

10 10 Minimum MSE (MMSE) obtained when h is obtained from solving WH equations. For best h obtained: σ min { 2 } { } min ( ( ) ( )) min = E e = E dn dn e 2 * e = ( ) { * } min ( n) E d n e P 1 = E d n d n hk x n k k = 0 d * ( ) ( ) ( ) P 1 ( 0) ( ) = R h r k k = 0 σ = R 0 h r 2 e min d ( ) T dx k dx *

11 11 Summary: FIR Wiener Filter Equations s(n) signal + noise w(n) + x(n) filter FIR Wiener filter is a FIR filter such that: d a f n d(n) + - e(n) where P 1 dˆ n = hx k n k ( ) * ( ) k = 0 2 σ ˆ e = E d n d n ( ) ( ) 2 is minimum.

12 12 How d(n) is defined specifies the specific type of Wiener filter designed: filtering: smoothing: predicting: W-H eqs: MMSE: min * 1 * Rh x = rdx => hopt = Rx rdx 2 T T σ e = Rd(0) hopt rdx = Rd(0) rdx h opt

13 13 Example 1: Wiener filter (filter case: d(n) = s(n) & white noise) Assume x(n) is defined by s(n) signal w(n) s(n), w(n) uncorrelated w(n) white noise, zero mean s(n) + noise + x(n) filter R w (n) = 2δ(n) R s (n) = 2 (0.8) n d a f n d(n) + - e(n)

14 14

15 15 Filter length Filter coefficients 2 [0.405, 0.238] [0.382, 0.2, 0.118] 0.76 MMSE 4 [0.377, 0.191, 0.01, 0.06] [0.375, 0.188, 0.095, 0.049, 0.029] 6 [0.3751, , , , , ] 7 [0.3750, , , , , , ] 8 [0.3750, , 0.038, 0.049, , , ,

16 16 Example 2: Application to Wiener filter (filter case: d(n) = s(n) & colored noise) s(n), w(n) uncorrelated, and zero-mean w(n) noise with R w (n) = 2 (0.5) n s(n) signal with R s (n) = 2 (0.8) n

17 17

18 Filter length Filter coefficients MMSE 2 [0.4156, ] [0.4122, , ] [0.4106, , ] 5 [0.4099, , , , ] 6 [0.4095, , , , , ] 7 [0.4094, , , , , , ] 8 [0.4093, , , , , , , ] /20/14 EC3410.SuFY14MPF - Section IIIC - FIR 18

19 19 Example 3: Application of Wiener filter to one-step linear prediction tracking of moving series forecasting of system behavior data compression telephone transmission W-H equations where h opt * ( ) = hx( n ) dˆ n ( ) d n = R r 1 x P 1 = 0 =? l * dx

20 20 Geometric interpretation: Assume a filter of length 2 x(n+1) x(n-1) e(n) ˆd ( n) x(n) ( ) ( ) ˆd n e n = = e(n) is the error between true value x(n+1) and predicted value for x(n+1) based on x(n) and x(n-1) represents the new information in x(n+1) which is not contained in x(n) and x(n-1) e(n) is called the innovation process corresponding to x(n)

21 21 Geometric interpretation: Assume x(n+1) only has NO new information (i.e., information in x(n+1) is that already contained in x(n) and x(n-1). Filter of length 2. ˆ ( ) ( ) Plot xn ( + 1), d n, e n x(n-1) x(n) ( ) ( ) ˆd n e n = =

22 22 Geometric interpretation: Assume x(n+1) only has new information (i.e., information in x(n+1) is NOT contained at all in x(n) and x(n-1) ). Filter of length 2. ˆ ( ) ( ) Plot xn ( + 1), d n, e n x(n-1) x(n) ( ) ( ) ˆd n e n = =

23 23 1-step ahead predictor scenario RP x(n) defined as x(n) =x(n-1) +v(n) a < 1 v(n) is white noise. 1-step predictor of length 2. AR (1) process Predictor a = 0.5 x ( n) =+ a x( n 1) + a x( n 2) 1 2 seed = 1024 L a = + NM O QP

24 24

25 25 Link between Predictor behavior & input signal behavior 1) Case 1: s(n) = process with correlation R ( k) = δ( k) δ( k 1) δ( k+ 1) s Investigate performances of N-step predictor as a function of changes N 2) Case 2: s(n) = process with correlation R k a a k s ( ) =, < 1 Investigate performances of predictor as a function of changes in a

26 26 1) Case 1: s(n) = wss process with R ( k) = δ( k) δ( k 1) δ( k+ 1) s

27 27 2) Case 2: s(n) = wss process with k s ( ) =, < 1 R k a a % EC MPF % Compute FIR filter coefficients for % a 1-step ahead predictor of length 2 % for correlation sequence of type R(k)=a^{ k } % Compute and plot resulting MMSE value A=[-0.9:0.1:0.9]; for k0=1:length(a) a=a(k0); for k=1:3 rs(k)=a^(k-1); end Rs=toeplitz(rs(1:2)); rdx=[rs(2);rs(3)]; h(:,k0)=rs\rdx; mmse(k0)=rs(1)-h(:,k0)'*rdx; end stem(a,mmse) xlabel('value of a') ylabel('mmse(a)') title('mmse(a)for1-step predictor of length 2, for R_s(k)=a^{ k }')

28 28 Example 4: s(n) = process with w(n) = white noise, zero mean s(n), w(n) uncorrelated R n s Design the 1-step ahead predictor of length 2. Compute MMSE. a f a f = n wa f a f = 2δ R n n

29 29 1- step ahead Predictor Filter Length Coefficients MMSE Filter MMSE 2 [0.3238, ] [0.3059, 01.6, ] [0.3015, , , ] 5 [0.3004, , , , 0.04, ] 6 [0.3001, , , , , ] 7 [0.3, 0.15, , , , 0.001, ] 8 [0.3, 0.15, 0.075, , , , , 0.003]

30 Length N-step ahead Predictor 1-step ahead MMSE 2-step ahead MMSE Filter Filter MMSE /20/14 EC3410.SuFY14MPF - Section IIIC - FIR 30

31 31 Example 5: s(n) = process with w(n) = wss noise, zero mean R n s s(n), w(n) uncorrelated Design the 1-step ahead predictor of length 2 Design 1-step back smoother of length 2 R w a f a f = n a f a f n = n

32 32

33 33 1-step ahead predictor (Col. Noise) Length Coefficients MMSE 2 [0.3325, ] [0.3297, 0.06, ] [0.3285, , , ] [0.3279, , 0.04, , ] 6 [0.3276, , , , , ] 7 [0.3275, , , , , , ] 8 [0.3275, , , , 0.018, , , ]

34 34 Length MMSE (Col. Noise) N-step ahead Predictor N-step back Smoother Filter 1-step 2-step 3-step 4-step 1-step 2-step 3-step 4-step

35 Comments 06/20/14 EC3410.SuFY14MPF - Section IIIC - FIR 35

36 36 Example 6: s(n) = cos(ω 0 n+φ), φ~u[0,2π] Design the 1-step ahead predictor of length 2, and associated MMSE Design 1-step back smoother of length 2, and associated MMSE

37 37

38 38 Weiner Filters and Error Surfaces 2 e Recall h opt computed from { 2 ( ) } ( ) d Rh x opt = r * dx {( )( ( ) ) * } H H H σ = E d n h x = E d n h x d n h x H ( ) { H 0 } 2 Real( T ) dx = R + h E xx h h r for real signals d(n), x(n) 2 ( ) T e d x σ = R 0 + hrh 2h r T dx using the fact that H ( h x) * H x h H H ( ) = ( ) d n h x = h d n x

39 39 2 ( ) T e d x σ = R 0 + hrh 2h r T dx For filter length 1 h = h 0 x = x(n) σ = R 0 + h R 0 2hr 0 ( ) ( ) ( ) 2 2 e d 0 x 0 dx 2 σ e 2 σ emin h opt h 0

40 40 For filter length P = 2 2 ( ) T e d x σ = R 0 + hrh 2h r T [ ] ( ) ( ) h= h0, h1 ; x= x n, x n 1 T ( ) 2 T e d x σ = R 0 + hrh 2h r ( 0 ) [ h, h ] [ h h ] ( 0) ( 1) ( 0) x( 1) ( 1) ( 0) x 0 = Rd Rx R x h 1 r dx 2 0, 1 rdx T dx R R h σ = A h + Ah + A h + Ah e Ahh + ( ) Rd 0 T dx

41 2 σ e h opt,0 h opt,0 h 0 h 0 h opt,1 h opt,1 h 1 h opt depends on r dx and R x h 1 moves down ( ) 2 T e d x 2 shape of σ e 2 σ up and e σ = R 0 + hrh 2h r R x : specifies shape of ( ) 2 σ e h r dx : specifies where the bowl is in the 3-d plane but doesn t change the shape of the bowl R d (0): moves bowl up and down in 3-d plane but doesn t change shape or location of bowl depends on R x information: λ(r x ) & eigenvectors 06/20/14 EC3410.SuFY14MPF - Section IIIC - FIR 41 T dx σ e h Rh x = 2Rh 2r = x r dx dx

42 Correlation matrix Eigenvalue Spread Impact on Error Surface Shape see plots Eigenvector Directions for 2 2 Toeplitz Correlation Matrix R x ( ) x( ) ( ) ( ) Rx 0 R 1 normalize 1 a = Rx 1 Rx 0 correlation a 1 eigenvalues of R x ( ) λ a 0 λ a = = 1 + a eigenvectors 1 λ a u11 = 0 a 1 λ u 12 ( λ ) u 11 au = 0 ( + a) u 11 + au 12 = 0 λ 1 = 1 a u11 = u u 12 1 = 1 1 λ2 = 1+ a u2 = 1 06/20/14 EC3410.SuFY14MPF - Section IIIC - FIR 42

43 43 Error surface shape and eigenvalue ratios a=0.1 a=-0.99 a=-0.5

44 44 Application to Channel Equalization s(n) W(z) Goal: Implement the equalization filter H(z) as a stable causal FIR filter Goal: Recover s(n) by estimating channel distortion (applications in communications, control, etc.) Information Available: x n = z n + v n ( ) ( ) ( ) channel output Delay D v(n) d(n) = s(n-d) z(n) x(n) ˆd ( n) + H(z) e(n) additive noise due to sensors d( n) = s( n D) s(n): original data samples Assumptions: 1) v(n) is stationary, zero-mean, uncorrelated with s(n). 2) v(n) = 0 & D=0

45 45 Assume: W( z) = z z Questions: 1 2 1) Assume v(n) = 0 & D=0. Identify the type of filter (FIR/IIR) needed to cancel channel distortions. Identify resulting H(z) 2) Identify whether the equalization filter is causal and stable. 3) Assume v(n) = 0 & D 0. Identify resulting H 2 (z) in terms of H(z). s(n) W(z) Delay D z(n) H(z) ( ) ˆd n + d(n) = s(n-d) e(n)

46 46

47 47

48

49 49 Assume D 0 s(n) W(z) Delay D z(n) H 2 (z) ( ) ˆd n + d(n) = s(n-d) e(n)

50

51 51 Application to Noise Cancellation Goal: Recover s(n) by compensating for the noise distortion while having only access to the related noise distortion signal v(n) (applications in communications, control, etc.) Signal source s(n) d(n)=s(n)+w(n) + - en ( ) = dn ( ) dn ( ) w(n) Information Available: xn ( ) = vn ( ) s(n)+w(n) & v(n) White noise source w(n) Transformation Of white noise into correlated noise A(z) Wiener filter ( ) dn Assumption: w(n) is stationary, zero-mean, uncorrelated with s(n).

52 52 Assume: vn ( ) = avn ( 1) + wn ( ), a= 0.6, wn ( ) white noise with variance σ sn ( ) = sin( ω n+ φ), φ ~ U[0, 2 π] 0 2 w Compute the FIR Wiener filter of length 2 and evaluate filter performances

53 Results Interpretation 06/20/14 EC3410.SuFY14MPF - Section IIIC - FIR 53

54 54 Application to Noise Cancellation (with information leakage) Signal source s(n) White noise source w(n) w(n) K Transformation Of white noise into correlated noise A(z) d(n)=s(n)+w(n) v(n) + xn ( ) = vn ( ) + Ksn ( ) Wiener filter - en ( ) = dn ( ) dn ( ) Information Available: ( ) dn s(n)+w(n) & v(n)+ks(n) Assumption: w(n) is stationary, zero-mean, uncorrelated with s(n).

55 Results Interpretation 06/20/14 EC3410.SuFY14MPF - Section IIIC - FIR 55

56 Application to Spatial Filtering Reference signal d(n)=s(n) e(n) s(n) ± 1 BPSK v 1 (n) v 2 (n) x 1 (n ) x 2 (n ) Noisy received signals x 1 (n) x 2 (n) Filtering ( ) dn Information Available: snapshot in time of received signal retrieved at two antennas & reference signal Assumption: v 1 (n), v 2 (n) zero mean wss white noise RPs independent of each other and of s(n). Goal: Denoise received signal 06/20/14 EC3410.SuFY14MPF - Section IIIC - FIR 56

57 Application to Spatial Filtering, cont 06/20/14 EC3410.SuFY14MPF - Section IIIC - FIR 57

58 58

59 59 Application to Spatial Filtering, cont Did we gain anything by using multiple receivers? s(n) ± 1 BPSK x(n) v(n) Filtering Process Assumption: v(n) zero mean RP independent of s(n). Goal: Compute filter coefficients, filter output, and MMSE.

60 60

61 61 Example: Gain Pattern at filter output Example: N-element array, - desired signal at θ 0 =30, - interferences at θ 1 = -20 θ 2 = 40 - Noise power: 0.1 Array steering vector A w ( θ ) = H w a( θ ) H w w 2

62 Appendices 06/20/14 EC3410.SuFY14MPF - Section IIIC - FIR 62

63 Appendix A: Derivation of proof for Mean Square estimate derivation (p. 3) Proof: ξ = Problem: find ξ = ( ) d x { 2 ( ) } d x ( ( )) ( ) E s { T ( ) } = E s d x s d x L( ss, Define ( x) ) M ( ( )) ( ) ( ) ( ( )) ( ) ( ) = M Ls, d x f s x ds f x d x 0 so that K is minimum : loss function L s, d x f s x f x d x ds, Bayes Rule K ξ is minimized if K is minimized for each value of x K d ( ) ( ) 2 ( s d ( x) ) f ( s x) ds ( s d ( x) ) f ( s x) ds K 2 = s d ( x) f s x ds d d = = 0 = 0 Es [ x] = ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) 1 sf s x ds = d x f s x ds s f s x ds = d x f s x ds s f s x ds = d x d ( x) 06/20/14 EC3410.SuFY14MPF - Section IIIC - FIR 63

64 64 2 K For a minimum we need: > 0 2 d Note: 2 K 2 d ( ( s d ( x) ) f ( s x) ds) d = ( 2) / = ( 2) f =(2) 1 > 0 ( ) s x ds

65 Appendix B: Proof of corollary of orthogonality principle Proof: H e= s h x= s ( h+ h h ) H x h where is the weight vector defined so that the orthogonality principle holds. Resulting error is called e = s h H x σ 2 e 2 H H { } ( ) ( ) H = E e = E s h x+ h h x H H = E e ( h h) x + ( ) 2 H H 2 H H E e = + ( h h) xe + ( h h) x + ex ( h h) { } 2 H ( ) { H } { } 2 { H } { * } E se H 2 H ( ) { }( ) = E e + h h E xe + E h h x + E ex h h { H H } { H } H { H } ( ) = E e = E s h x e = E se h E xe = E se = 06/20/14 EC3410.SuFY14MPF - Section IIIC - FIR 65

66 66 References [1]: Discrete Random Signals and Statistical Signal Processing, C. Therrien, Prentice Hall, 1992 [2]: Statistical and Adaptive Signal Processing, D. Manolakis, V. Ingle & S. Kogon, Artech House, 2005 [3]: Adaptive Filters, 2 nd Ed., S. Haykin, Prentice Hall, 1991

V. Adaptive filtering Widrow-Hopf Learning Rule LMS and Adaline

V. Adaptive filtering Widrow-Hopf Learning Rule LMS and Adaline V. Adaptive filtering Widrow-Hopf Learning Rule LMS and Adaline Goals Introduce Wiener-Hopf (WH) equations Introduce application of the steepest descent method to the WH problem Approximation to the Least

More information

INTRODUCTION Noise is present in many situations of daily life for ex: Microphones will record noise and speech. Goal: Reconstruct original signal Wie

INTRODUCTION Noise is present in many situations of daily life for ex: Microphones will record noise and speech. Goal: Reconstruct original signal Wie WIENER FILTERING Presented by N.Srikanth(Y8104060), M.Manikanta PhaniKumar(Y8104031). INDIAN INSTITUTE OF TECHNOLOGY KANPUR Electrical Engineering dept. INTRODUCTION Noise is present in many situations

More information

Linear Optimum Filtering: Statement

Linear Optimum Filtering: Statement Ch2: Wiener Filters Optimal filters for stationary stochastic models are reviewed and derived in this presentation. Contents: Linear optimal filtering Principle of orthogonality Minimum mean squared error

More information

III.B Linear Transformations: Matched filter

III.B Linear Transformations: Matched filter III.B Linear Transformations: Matched filter [p. ] Definition [p. 3] Deterministic signal case [p. 5] White noise [p. 19] Colored noise [p. 4] Random signal case [p. 4] White noise [p. 4] Colored noise

More information

Chapter 2 Wiener Filtering

Chapter 2 Wiener Filtering Chapter 2 Wiener Filtering Abstract Before moving to the actual adaptive filtering problem, we need to solve the optimum linear filtering problem (particularly, in the mean-square-error sense). We start

More information

Adaptive Filter Theory

Adaptive Filter Theory 0 Adaptive Filter heory Sung Ho Cho Hanyang University Seoul, Korea (Office) +8--0-0390 (Mobile) +8-10-541-5178 dragon@hanyang.ac.kr able of Contents 1 Wiener Filters Gradient Search by Steepest Descent

More information

26. Filtering. ECE 830, Spring 2014

26. Filtering. ECE 830, Spring 2014 26. Filtering ECE 830, Spring 2014 1 / 26 Wiener Filtering Wiener filtering is the application of LMMSE estimation to recovery of a signal in additive noise under wide sense sationarity assumptions. Problem

More information

EE482: Digital Signal Processing Applications

EE482: Digital Signal Processing Applications Professor Brendan Morris, SEB 3216, brendan.morris@unlv.edu EE482: Digital Signal Processing Applications Spring 2014 TTh 14:30-15:45 CBC C222 Lecture 11 Adaptive Filtering 14/03/04 http://www.ee.unlv.edu/~b1morris/ee482/

More information

Optimal and Adaptive Filtering

Optimal and Adaptive Filtering Optimal and Adaptive Filtering Murat Üney M.Uney@ed.ac.uk Institute for Digital Communications (IDCOM) 26/06/2017 Murat Üney (IDCOM) Optimal and Adaptive Filtering 26/06/2017 1 / 69 Table of Contents 1

More information

EE482: Digital Signal Processing Applications

EE482: Digital Signal Processing Applications Professor Brendan Morris, SEB 3216, brendan.morris@unlv.edu EE482: Digital Signal Processing Applications Spring 2014 TTh 14:30-15:45 CBC C222 Lecture 11 Adaptive Filtering 14/03/04 http://www.ee.unlv.edu/~b1morris/ee482/

More information

Statistical and Adaptive Signal Processing

Statistical and Adaptive Signal Processing r Statistical and Adaptive Signal Processing Spectral Estimation, Signal Modeling, Adaptive Filtering and Array Processing Dimitris G. Manolakis Massachusetts Institute of Technology Lincoln Laboratory

More information

Wiener Filtering. EE264: Lecture 12

Wiener Filtering. EE264: Lecture 12 EE264: Lecture 2 Wiener Filtering In this lecture we will take a different view of filtering. Previously, we have depended on frequency-domain specifications to make some sort of LP/ BP/ HP/ BS filter,

More information

Adaptive Filtering Part II

Adaptive Filtering Part II Adaptive Filtering Part II In previous Lecture we saw that: Setting the gradient of cost function equal to zero, we obtain the optimum values of filter coefficients: (Wiener-Hopf equation) Adaptive Filtering,

More information

Machine Learning. A Bayesian and Optimization Perspective. Academic Press, Sergios Theodoridis 1. of Athens, Athens, Greece.

Machine Learning. A Bayesian and Optimization Perspective. Academic Press, Sergios Theodoridis 1. of Athens, Athens, Greece. Machine Learning A Bayesian and Optimization Perspective Academic Press, 2015 Sergios Theodoridis 1 1 Dept. of Informatics and Telecommunications, National and Kapodistrian University of Athens, Athens,

More information

The goal of the Wiener filter is to filter out noise that has corrupted a signal. It is based on a statistical approach.

The goal of the Wiener filter is to filter out noise that has corrupted a signal. It is based on a statistical approach. Wiener filter From Wikipedia, the free encyclopedia In signal processing, the Wiener filter is a filter proposed by Norbert Wiener during the 1940s and published in 1949. [1] Its purpose is to reduce the

More information

An Adaptive Sensor Array Using an Affine Combination of Two Filters

An Adaptive Sensor Array Using an Affine Combination of Two Filters An Adaptive Sensor Array Using an Affine Combination of Two Filters Tõnu Trump Tallinn University of Technology Department of Radio and Telecommunication Engineering Ehitajate tee 5, 19086 Tallinn Estonia

More information

Adaptive Systems Homework Assignment 1

Adaptive Systems Homework Assignment 1 Signal Processing and Speech Communication Lab. Graz University of Technology Adaptive Systems Homework Assignment 1 Name(s) Matr.No(s). The analytical part of your homework (your calculation sheets) as

More information

Adaptive MMSE Equalizer with Optimum Tap-length and Decision Delay

Adaptive MMSE Equalizer with Optimum Tap-length and Decision Delay Adaptive MMSE Equalizer with Optimum Tap-length and Decision Delay Yu Gong, Xia Hong and Khalid F. Abu-Salim School of Systems Engineering The University of Reading, Reading RG6 6AY, UK E-mail: {y.gong,x.hong,k.f.abusalem}@reading.ac.uk

More information

Revision of Lecture 4

Revision of Lecture 4 Revision of Lecture 4 We have discussed all basic components of MODEM Pulse shaping Tx/Rx filter pair Modulator/demodulator Bits map symbols Discussions assume ideal channel, and for dispersive channel

More information

ECE534, Spring 2018: Solutions for Problem Set #5

ECE534, Spring 2018: Solutions for Problem Set #5 ECE534, Spring 08: s for Problem Set #5 Mean Value and Autocorrelation Functions Consider a random process X(t) such that (i) X(t) ± (ii) The number of zero crossings, N(t), in the interval (0, t) is described

More information

Adaptive Systems. Winter Term 2017/18. Instructor: Pejman Mowlaee Beikzadehmahaleh. Assistants: Christian Stetco

Adaptive Systems. Winter Term 2017/18. Instructor: Pejman Mowlaee Beikzadehmahaleh. Assistants: Christian Stetco Adaptive Systems Winter Term 2017/18 Instructor: Pejman Mowlaee Beikzadehmahaleh Assistants: Christian Stetco Signal Processing and Speech Communication Laboratory, Inffeldgasse 16c/EG written by Bernhard

More information

Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes (cont d)

Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes (cont d) Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes (cont d) Electrical & Computer Engineering North Carolina State University Acknowledgment: ECE792-41 slides

More information

Massachusetts Institute of Technology

Massachusetts Institute of Technology Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.011: Introduction to Communication, Control and Signal Processing QUIZ, April 1, 010 QUESTION BOOKLET Your

More information

Laboratory Project 2: Spectral Analysis and Optimal Filtering

Laboratory Project 2: Spectral Analysis and Optimal Filtering Laboratory Project 2: Spectral Analysis and Optimal Filtering Random signals analysis (MVE136) Mats Viberg and Lennart Svensson Department of Signals and Systems Chalmers University of Technology 412 96

More information

A Derivation of the Steady-State MSE of RLS: Stationary and Nonstationary Cases

A Derivation of the Steady-State MSE of RLS: Stationary and Nonstationary Cases A Derivation of the Steady-State MSE of RLS: Stationary and Nonstationary Cases Phil Schniter Nov. 0, 001 Abstract In this report we combine the approach of Yousef and Sayed [1] with that of Rupp and Sayed

More information

SIMON FRASER UNIVERSITY School of Engineering Science

SIMON FRASER UNIVERSITY School of Engineering Science SIMON FRASER UNIVERSITY School of Engineering Science Course Outline ENSC 810-3 Digital Signal Processing Calendar Description This course covers advanced digital signal processing techniques. The main

More information

A TERM PAPER REPORT ON KALMAN FILTER

A TERM PAPER REPORT ON KALMAN FILTER A TERM PAPER REPORT ON KALMAN FILTER By B. Sivaprasad (Y8104059) CH. Venkata Karunya (Y8104066) Department of Electrical Engineering Indian Institute of Technology, Kanpur Kanpur-208016 SCALAR KALMAN FILTER

More information

ELEG-636: Statistical Signal Processing

ELEG-636: Statistical Signal Processing ELEG-636: Statistical Signal Processing Gonzalo R. Arce Department of Electrical and Computer Engineering University of Delaware Spring 2010 Gonzalo R. Arce (ECE, Univ. of Delaware) ELEG-636: Statistical

More information

ADAPTIVE FILTER THEORY

ADAPTIVE FILTER THEORY ADAPTIVE FILTER THEORY Fourth Edition Simon Haykin Communications Research Laboratory McMaster University Hamilton, Ontario, Canada Front ice Hall PRENTICE HALL Upper Saddle River, New Jersey 07458 Preface

More information

MMSE DECISION FEEDBACK EQUALIZER FROM CHANNEL ESTIMATE

MMSE DECISION FEEDBACK EQUALIZER FROM CHANNEL ESTIMATE MMSE DECISION FEEDBACK EQUALIZER FROM CHANNEL ESTIMATE M. Magarini, A. Spalvieri, Dipartimento di Elettronica e Informazione, Politecnico di Milano, Piazza Leonardo da Vinci, 32, I-20133 Milano (Italy),

More information

EE 225a Digital Signal Processing Supplementary Material

EE 225a Digital Signal Processing Supplementary Material EE 225A DIGITAL SIGAL PROCESSIG SUPPLEMETARY MATERIAL EE 225a Digital Signal Processing Supplementary Material. Allpass Sequences A sequence h a ( n) ote that this is true iff which in turn is true iff

More information

Generalized Sidelobe Canceller and MVDR Power Spectrum Estimation. Bhaskar D Rao University of California, San Diego

Generalized Sidelobe Canceller and MVDR Power Spectrum Estimation. Bhaskar D Rao University of California, San Diego Generalized Sidelobe Canceller and MVDR Power Spectrum Estimation Bhaskar D Rao University of California, San Diego Email: brao@ucsd.edu Reference Books 1. Optimum Array Processing, H. L. Van Trees 2.

More information

Analog vs. discrete signals

Analog vs. discrete signals Analog vs. discrete signals Continuous-time signals are also known as analog signals because their amplitude is analogous (i.e., proportional) to the physical quantity they represent. Discrete-time signals

More information

Reduced-Rank Multi-Antenna Cyclic Wiener Filtering for Interference Cancellation

Reduced-Rank Multi-Antenna Cyclic Wiener Filtering for Interference Cancellation Reduced-Rank Multi-Antenna Cyclic Wiener Filtering for Interference Cancellation Hong Zhang, Ali Abdi and Alexander Haimovich Center for Wireless Communications and Signal Processing Research Department

More information

Unbiased Power Prediction of Rayleigh Fading Channels

Unbiased Power Prediction of Rayleigh Fading Channels Unbiased Power Prediction of Rayleigh Fading Channels Torbjörn Ekman UniK PO Box 70, N-2027 Kjeller, Norway Email: torbjorn.ekman@signal.uu.se Mikael Sternad and Anders Ahlén Signals and Systems, Uppsala

More information

Multimedia Communications. Differential Coding

Multimedia Communications. Differential Coding Multimedia Communications Differential Coding Differential Coding In many sources, the source output does not change a great deal from one sample to the next. This means that both the dynamic range and

More information

II. Nonparametric Spectrum Estimation for Stationary Random Signals - Non-parametric Methods -

II. Nonparametric Spectrum Estimation for Stationary Random Signals - Non-parametric Methods - II. onparametric Spectrum Estimation for Stationary Random Signals - on-parametric Methods - - [p. 3] Periodogram - [p. 12] Periodogram properties - [p. 23] Modified periodogram - [p. 25] Bartlett s method

More information

Direction of Arrival Estimation: Subspace Methods. Bhaskar D Rao University of California, San Diego

Direction of Arrival Estimation: Subspace Methods. Bhaskar D Rao University of California, San Diego Direction of Arrival Estimation: Subspace Methods Bhaskar D Rao University of California, San Diego Email: brao@ucsdedu Reference Books and Papers 1 Optimum Array Processing, H L Van Trees 2 Stoica, P,

More information

Noise-Shaped Predictive Coding for Multiple Descriptions of a Colored Gaussian Source

Noise-Shaped Predictive Coding for Multiple Descriptions of a Colored Gaussian Source Noise-Shaped Predictive Coding for Multiple Descriptions of a Colored Gaussian Source Yuval Kochman, Jan Østergaard, and Ram Zamir Abstract It was recently shown that the symmetric multiple-description

More information

ECE4270 Fundamentals of DSP Lecture 20. Fixed-Point Arithmetic in FIR and IIR Filters (part I) Overview of Lecture. Overflow. FIR Digital Filter

ECE4270 Fundamentals of DSP Lecture 20. Fixed-Point Arithmetic in FIR and IIR Filters (part I) Overview of Lecture. Overflow. FIR Digital Filter ECE4270 Fundamentals of DSP Lecture 20 Fixed-Point Arithmetic in FIR and IIR Filters (part I) School of ECE Center for Signal and Information Processing Georgia Institute of Technology Overview of Lecture

More information

Basic Principles of Video Coding

Basic Principles of Video Coding Basic Principles of Video Coding Introduction Categories of Video Coding Schemes Information Theory Overview of Video Coding Techniques Predictive coding Transform coding Quantization Entropy coding Motion

More information

1. Determine if each of the following are valid autocorrelation matrices of WSS processes. (Correlation Matrix),R c =

1. Determine if each of the following are valid autocorrelation matrices of WSS processes. (Correlation Matrix),R c = ENEE630 ADSP Part II w/ solution. Determine if each of the following are valid autocorrelation matrices of WSS processes. (Correlation Matrix) R a = 4 4 4,R b = 0 0,R c = j 0 j 0 j 0 j 0 j,r d = 0 0 0

More information

Convergence Evaluation of a Random Step-Size NLMS Adaptive Algorithm in System Identification and Channel Equalization

Convergence Evaluation of a Random Step-Size NLMS Adaptive Algorithm in System Identification and Channel Equalization Convergence Evaluation of a Random Step-Size NLMS Adaptive Algorithm in System Identification and Channel Equalization 1 Shihab Jimaa Khalifa University of Science, Technology and Research (KUSTAR) Faculty

More information

Sparse Least Mean Square Algorithm for Estimation of Truncated Volterra Kernels

Sparse Least Mean Square Algorithm for Estimation of Truncated Volterra Kernels Sparse Least Mean Square Algorithm for Estimation of Truncated Volterra Kernels Bijit Kumar Das 1, Mrityunjoy Chakraborty 2 Department of Electronics and Electrical Communication Engineering Indian Institute

More information

2.6 The optimum filtering solution is defined by the Wiener-Hopf equation

2.6 The optimum filtering solution is defined by the Wiener-Hopf equation .6 The optimum filtering solution is defined by the Wiener-opf equation w o p for which the minimum mean-square error equals J min σ d p w o () Combine Eqs. and () into a single relation: σ d p p 1 w o

More information

Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes

Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes 1 Discrete-time Stochastic Processes Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes Electrical & Computer Engineering University of Maryland, College Park

More information

Lesson 1. Optimal signalbehandling LTH. September Statistical Digital Signal Processing and Modeling, Hayes, M:

Lesson 1. Optimal signalbehandling LTH. September Statistical Digital Signal Processing and Modeling, Hayes, M: Lesson 1 Optimal Signal Processing Optimal signalbehandling LTH September 2013 Statistical Digital Signal Processing and Modeling, Hayes, M: John Wiley & Sons, 1996. ISBN 0471594318 Nedelko Grbic Mtrl

More information

Optimal and Adaptive Filtering

Optimal and Adaptive Filtering Optimal and Adaptive Filtering Murat Üney M.Uney@ed.ac.uk Institute for Digital Communications (IDCOM) 27/06/2016 Murat Üney (IDCOM) Optimal and Adaptive Filtering 27/06/2016 1 / 69 This presentation aims

More information

Levinson Durbin Recursions: I

Levinson Durbin Recursions: I Levinson Durbin Recursions: I note: B&D and S&S say Durbin Levinson but Levinson Durbin is more commonly used (Levinson, 1947, and Durbin, 1960, are source articles sometimes just Levinson is used) recursions

More information

Adaptive Filters. un [ ] yn [ ] w. yn n wun k. - Adaptive filter (FIR): yn n n w nun k. (1) Identification. Unknown System + (2) Inverse modeling

Adaptive Filters. un [ ] yn [ ] w. yn n wun k. - Adaptive filter (FIR): yn n n w nun k. (1) Identification. Unknown System + (2) Inverse modeling Adaptive Filters - Statistical digital signal processing: in many problems of interest, the signals exhibit some inherent variability plus additive noise we use probabilistic laws to model the statistical

More information

ADAPTIVE ANTENNAS. SPATIAL BF

ADAPTIVE ANTENNAS. SPATIAL BF ADAPTIVE ANTENNAS SPATIAL BF 1 1-Spatial reference BF -Spatial reference beamforming may not use of embedded training sequences. Instead, the directions of arrival (DoA) of the impinging waves are used

More information

Levinson Durbin Recursions: I

Levinson Durbin Recursions: I Levinson Durbin Recursions: I note: B&D and S&S say Durbin Levinson but Levinson Durbin is more commonly used (Levinson, 1947, and Durbin, 1960, are source articles sometimes just Levinson is used) recursions

More information

UCSD ECE153 Handout #40 Prof. Young-Han Kim Thursday, May 29, Homework Set #8 Due: Thursday, June 5, 2011

UCSD ECE153 Handout #40 Prof. Young-Han Kim Thursday, May 29, Homework Set #8 Due: Thursday, June 5, 2011 UCSD ECE53 Handout #40 Prof. Young-Han Kim Thursday, May 9, 04 Homework Set #8 Due: Thursday, June 5, 0. Discrete-time Wiener process. Let Z n, n 0 be a discrete time white Gaussian noise (WGN) process,

More information

Beamspace Adaptive Channel Compensation for Sensor Arrays with Faulty Elements

Beamspace Adaptive Channel Compensation for Sensor Arrays with Faulty Elements 1 2005 Asilomar Conference Beamspace Adaptive Channel Compensation for Sensor Arrays with Faulty Elements Oguz R. Kazanci and Jeffrey L. Krolik Duke University Department of Electrical and Computer Engineering

More information

IS NEGATIVE STEP SIZE LMS ALGORITHM STABLE OPERATION POSSIBLE?

IS NEGATIVE STEP SIZE LMS ALGORITHM STABLE OPERATION POSSIBLE? IS NEGATIVE STEP SIZE LMS ALGORITHM STABLE OPERATION POSSIBLE? Dariusz Bismor Institute of Automatic Control, Silesian University of Technology, ul. Akademicka 16, 44-100 Gliwice, Poland, e-mail: Dariusz.Bismor@polsl.pl

More information

ELEG 5633 Detection and Estimation Signal Detection: Deterministic Signals

ELEG 5633 Detection and Estimation Signal Detection: Deterministic Signals ELEG 5633 Detection and Estimation Signal Detection: Deterministic Signals Jingxian Wu Department of Electrical Engineering University of Arkansas Outline Matched Filter Generalized Matched Filter Signal

More information

Lecture 19: Bayesian Linear Estimators

Lecture 19: Bayesian Linear Estimators ECE 830 Fall 2010 Statistical Signal Processing instructor: R Nowa, scribe: I Rosado-Mendez Lecture 19: Bayesian Linear Estimators 1 Linear Minimum Mean-Square Estimator Suppose our data is set X R n,

More information

A Hilbert Space for Random Processes

A Hilbert Space for Random Processes Gaussian Basics Random Processes Filtering of Random Processes Signal Space Concepts A Hilbert Space for Random Processes I A vector space for random processes X t that is analogous to L 2 (a, b) is of

More information

Improved Detected Data Processing for Decision-Directed Tracking of MIMO Channels

Improved Detected Data Processing for Decision-Directed Tracking of MIMO Channels Improved Detected Data Processing for Decision-Directed Tracking of MIMO Channels Emna Eitel and Joachim Speidel Institute of Telecommunications, University of Stuttgart, Germany Abstract This paper addresses

More information

Least Square Es?ma?on, Filtering, and Predic?on: ECE 5/639 Sta?s?cal Signal Processing II: Linear Es?ma?on

Least Square Es?ma?on, Filtering, and Predic?on: ECE 5/639 Sta?s?cal Signal Processing II: Linear Es?ma?on Least Square Es?ma?on, Filtering, and Predic?on: Sta?s?cal Signal Processing II: Linear Es?ma?on Eric Wan, Ph.D. Fall 2015 1 Mo?va?ons If the second-order sta?s?cs are known, the op?mum es?mator is given

More information

Spatial Smoothing and Broadband Beamforming. Bhaskar D Rao University of California, San Diego

Spatial Smoothing and Broadband Beamforming. Bhaskar D Rao University of California, San Diego Spatial Smoothing and Broadband Beamforming Bhaskar D Rao University of California, San Diego Email: brao@ucsd.edu Reference Books and Papers 1. Optimum Array Processing, H. L. Van Trees 2. Stoica, P.,

More information

Beamforming. A brief introduction. Brian D. Jeffs Associate Professor Dept. of Electrical and Computer Engineering Brigham Young University

Beamforming. A brief introduction. Brian D. Jeffs Associate Professor Dept. of Electrical and Computer Engineering Brigham Young University Beamforming A brief introduction Brian D. Jeffs Associate Professor Dept. of Electrical and Computer Engineering Brigham Young University March 2008 References Barry D. Van Veen and Kevin Buckley, Beamforming:

More information

Machine Learning and Adaptive Systems. Lectures 3 & 4

Machine Learning and Adaptive Systems. Lectures 3 & 4 ECE656- Lectures 3 & 4, Professor Department of Electrical and Computer Engineering Colorado State University Fall 2015 What is Learning? General Definition of Learning: Any change in the behavior or performance

More information

UCSD ECE250 Handout #27 Prof. Young-Han Kim Friday, June 8, Practice Final Examination (Winter 2017)

UCSD ECE250 Handout #27 Prof. Young-Han Kim Friday, June 8, Practice Final Examination (Winter 2017) UCSD ECE250 Handout #27 Prof. Young-Han Kim Friday, June 8, 208 Practice Final Examination (Winter 207) There are 6 problems, each problem with multiple parts. Your answer should be as clear and readable

More information

Chapter [4] "Operations on a Single Random Variable"

Chapter [4] Operations on a Single Random Variable Chapter [4] "Operations on a Single Random Variable" 4.1 Introduction In our study of random variables we use the probability density function or the cumulative distribution function to provide a complete

More information

Estimation techniques

Estimation techniques Estimation techniques March 2, 2006 Contents 1 Problem Statement 2 2 Bayesian Estimation Techniques 2 2.1 Minimum Mean Squared Error (MMSE) estimation........................ 2 2.1.1 General formulation......................................

More information

Ch5: Least Mean-Square Adaptive Filtering

Ch5: Least Mean-Square Adaptive Filtering Ch5: Least Mean-Square Adaptive Filtering Introduction - approximating steepest-descent algorithm Least-mean-square algorithm Stability and performance of the LMS algorithm Robustness of the LMS algorithm

More information

Shannon meets Wiener II: On MMSE estimation in successive decoding schemes

Shannon meets Wiener II: On MMSE estimation in successive decoding schemes Shannon meets Wiener II: On MMSE estimation in successive decoding schemes G. David Forney, Jr. MIT Cambridge, MA 0239 USA forneyd@comcast.net Abstract We continue to discuss why MMSE estimation arises

More information

Stochastic Processes. M. Sami Fadali Professor of Electrical Engineering University of Nevada, Reno

Stochastic Processes. M. Sami Fadali Professor of Electrical Engineering University of Nevada, Reno Stochastic Processes M. Sami Fadali Professor of Electrical Engineering University of Nevada, Reno 1 Outline Stochastic (random) processes. Autocorrelation. Crosscorrelation. Spectral density function.

More information

Interactions of Information Theory and Estimation in Single- and Multi-user Communications

Interactions of Information Theory and Estimation in Single- and Multi-user Communications Interactions of Information Theory and Estimation in Single- and Multi-user Communications Dongning Guo Department of Electrical Engineering Princeton University March 8, 2004 p 1 Dongning Guo Communications

More information

Q3. Derive the Wiener filter (non-causal) for a stationary process with given spectral characteristics

Q3. Derive the Wiener filter (non-causal) for a stationary process with given spectral characteristics Q3. Derive the Wiener filter (non-causal) for a stationary process with given spectral characteristics McMaster University 1 Background x(t) z(t) WF h(t) x(t) n(t) Figure 1: Signal Flow Diagram Goal: filter

More information

Wireless Information Transmission System Lab. Channel Estimation. Institute of Communications Engineering. National Sun Yat-sen University

Wireless Information Transmission System Lab. Channel Estimation. Institute of Communications Engineering. National Sun Yat-sen University Wireless Information Transmission System Lab. Channel Estimation Institute of Communications Engineering National Sun Yat-sen University Table of Contents Introduction to Channel Estimation Generic Pilot

More information

Beamforming Arrays with Faulty Sensors in Dynamic Environments

Beamforming Arrays with Faulty Sensors in Dynamic Environments Beamforming Arrays with Faulty Sensors in Dynamic Environments Jeffrey Krolik and Oguz R. Kazanci Duke University phone: 919-660-5274 email: jk@ee.duke.edu Abstract This paper addresses the problem of

More information

Linear Convolution Using FFT

Linear Convolution Using FFT Linear Convolution Using FFT Another useful property is that we can perform circular convolution and see how many points remain the same as those of linear convolution. When P < L and an L-point circular

More information

Complement on Digital Spectral Analysis and Optimal Filtering: Theory and Exercises

Complement on Digital Spectral Analysis and Optimal Filtering: Theory and Exercises Complement on Digital Spectral Analysis and Optimal Filtering: Theory and Exercises Random Processes With Applications (MVE 135) Mats Viberg Department of Signals and Systems Chalmers University of Technology

More information

Probability Space. J. McNames Portland State University ECE 538/638 Stochastic Signals Ver

Probability Space. J. McNames Portland State University ECE 538/638 Stochastic Signals Ver Stochastic Signals Overview Definitions Second order statistics Stationarity and ergodicity Random signal variability Power spectral density Linear systems with stationary inputs Random signal memory Correlation

More information

University of Waterloo Department of Electrical and Computer Engineering ECE 413 Digital Signal Processing. Spring Home Assignment 2 Solutions

University of Waterloo Department of Electrical and Computer Engineering ECE 413 Digital Signal Processing. Spring Home Assignment 2 Solutions University of Waterloo Department of Electrical and Computer Engineering ECE 13 Digital Signal Processing Spring 017 Home Assignment Solutions Due on June 8, 017 Exercise 1 An LTI system is described by

More information

Special Topics: Data Science

Special Topics: Data Science Special Topics: Data Science L6b: Wiener Filter Victor Solo School of Electrical Engineering University of New South Wales Sydney, AUSTRALIA Topics Norbert Wiener 1894-1964 MIT Professor (Frequency Domain)

More information

GENERALIZED DEFLATION ALGORITHMS FOR THE BLIND SOURCE-FACTOR SEPARATION OF MIMO-FIR CHANNELS. Mitsuru Kawamoto 1,2 and Yujiro Inouye 1

GENERALIZED DEFLATION ALGORITHMS FOR THE BLIND SOURCE-FACTOR SEPARATION OF MIMO-FIR CHANNELS. Mitsuru Kawamoto 1,2 and Yujiro Inouye 1 GENERALIZED DEFLATION ALGORITHMS FOR THE BLIND SOURCE-FACTOR SEPARATION OF MIMO-FIR CHANNELS Mitsuru Kawamoto,2 and Yuiro Inouye. Dept. of Electronic and Control Systems Engineering, Shimane University,

More information

Error Entropy Criterion in Echo State Network Training

Error Entropy Criterion in Echo State Network Training Error Entropy Criterion in Echo State Network Training Levy Boccato 1, Daniel G. Silva 1, Denis Fantinato 1, Kenji Nose Filho 1, Rafael Ferrari 1, Romis Attux 1, Aline Neves 2, Jugurta Montalvão 3 and

More information

Advanced Digital Signal Processing -Introduction

Advanced Digital Signal Processing -Introduction Advanced Digital Signal Processing -Introduction LECTURE-2 1 AP9211- ADVANCED DIGITAL SIGNAL PROCESSING UNIT I DISCRETE RANDOM SIGNAL PROCESSING Discrete Random Processes- Ensemble Averages, Stationary

More information

EFFECTS OF ILL-CONDITIONED DATA ON LEAST SQUARES ADAPTIVE FILTERS. Gary A. Ybarra and S.T. Alexander

EFFECTS OF ILL-CONDITIONED DATA ON LEAST SQUARES ADAPTIVE FILTERS. Gary A. Ybarra and S.T. Alexander EFFECTS OF ILL-CONDITIONED DATA ON LEAST SQUARES ADAPTIVE FILTERS Gary A. Ybarra and S.T. Alexander Center for Communications and Signal Processing Electrical and Computer Engineering Department North

More information

CHAPTER 3 ROBUST ADAPTIVE BEAMFORMING

CHAPTER 3 ROBUST ADAPTIVE BEAMFORMING 50 CHAPTER 3 ROBUST ADAPTIVE BEAMFORMING 3.1 INTRODUCTION Adaptive beamforming is used for enhancing a desired signal while suppressing noise and interference at the output of an array of sensors. It is

More information

ELEN E4810: Digital Signal Processing Topic 2: Time domain

ELEN E4810: Digital Signal Processing Topic 2: Time domain ELEN E4810: Digital Signal Processing Topic 2: Time domain 1. Discrete-time systems 2. Convolution 3. Linear Constant-Coefficient Difference Equations (LCCDEs) 4. Correlation 1 1. Discrete-time systems

More information

Examples. 2-input, 1-output discrete-time systems: 1-input, 1-output discrete-time systems:

Examples. 2-input, 1-output discrete-time systems: 1-input, 1-output discrete-time systems: Discrete-Time s - I Time-Domain Representation CHAPTER 4 These lecture slides are based on "Digital Signal Processing: A Computer-Based Approach, 4th ed." textbook by S.K. Mitra and its instructor materials.

More information

! Spectral Analysis with DFT. ! Windowing. ! Effect of zero-padding. ! Time-dependent Fourier transform. " Aka short-time Fourier transform

! Spectral Analysis with DFT. ! Windowing. ! Effect of zero-padding. ! Time-dependent Fourier transform.  Aka short-time Fourier transform Lecture Outline ESE 531: Digital Signal Processing Spectral Analysis with DFT Windowing Lec 24: April 18, 2019 Spectral Analysis Effect of zero-padding Time-dependent Fourier transform " Aka short-time

More information

Lecture Notes 7 Stationary Random Processes. Strict-Sense and Wide-Sense Stationarity. Autocorrelation Function of a Stationary Process

Lecture Notes 7 Stationary Random Processes. Strict-Sense and Wide-Sense Stationarity. Autocorrelation Function of a Stationary Process Lecture Notes 7 Stationary Random Processes Strict-Sense and Wide-Sense Stationarity Autocorrelation Function of a Stationary Process Power Spectral Density Continuity and Integration of Random Processes

More information

Lessons in Estimation Theory for Signal Processing, Communications, and Control

Lessons in Estimation Theory for Signal Processing, Communications, and Control Lessons in Estimation Theory for Signal Processing, Communications, and Control Jerry M. Mendel Department of Electrical Engineering University of Southern California Los Angeles, California PRENTICE HALL

More information

System Identification and Adaptive Filtering in the Short-Time Fourier Transform Domain

System Identification and Adaptive Filtering in the Short-Time Fourier Transform Domain System Identification and Adaptive Filtering in the Short-Time Fourier Transform Domain Electrical Engineering Department Technion - Israel Institute of Technology Supervised by: Prof. Israel Cohen Outline

More information

Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science. Fall Solutions for Problem Set 2

Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science. Fall Solutions for Problem Set 2 Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Issued: Tuesday, September 5. 6.: Discrete-Time Signal Processing Fall 5 Solutions for Problem Set Problem.

More information

Speaker Tracking and Beamforming

Speaker Tracking and Beamforming Speaker Tracking and Beamforming Dr. John McDonough Spoken Language Systems Saarland University January 13, 2010 Introduction Many problems in science and engineering can be formulated in terms of estimating

More information

DETECTION theory deals primarily with techniques for

DETECTION theory deals primarily with techniques for ADVANCED SIGNAL PROCESSING SE Optimum Detection of Deterministic and Random Signals Stefan Tertinek Graz University of Technology turtle@sbox.tugraz.at Abstract This paper introduces various methods for

More information

Complement on Digital Spectral Analysis and Optimal Filtering: Theory and Exercises

Complement on Digital Spectral Analysis and Optimal Filtering: Theory and Exercises Complement on Digital Spectral Analysis and Optimal Filtering: Theory and Exercises Random Signals Analysis (MVE136) Mats Viberg Department of Signals and Systems Chalmers University of Technology 412

More information

Adaptive Filtering. Squares. Alexander D. Poularikas. Fundamentals of. Least Mean. with MATLABR. University of Alabama, Huntsville, AL.

Adaptive Filtering. Squares. Alexander D. Poularikas. Fundamentals of. Least Mean. with MATLABR. University of Alabama, Huntsville, AL. Adaptive Filtering Fundamentals of Least Mean Squares with MATLABR Alexander D. Poularikas University of Alabama, Huntsville, AL CRC Press Taylor & Francis Croup Boca Raton London New York CRC Press is

More information

Adaptive Beamforming Algorithms

Adaptive Beamforming Algorithms S. R. Zinka srinivasa_zinka@daiict.ac.in October 29, 2014 Outline 1 Least Mean Squares 2 Sample Matrix Inversion 3 Recursive Least Squares 4 Accelerated Gradient Approach 5 Conjugate Gradient Method Outline

More information

ECE 541 Stochastic Signals and Systems Problem Set 11 Solution

ECE 541 Stochastic Signals and Systems Problem Set 11 Solution ECE 54 Stochastic Signals and Systems Problem Set Solution Problem Solutions : Yates and Goodman,..4..7.3.3.4.3.8.3 and.8.0 Problem..4 Solution Since E[Y (t] R Y (0, we use Theorem.(a to evaluate R Y (τ

More information

DSP Applications for Wireless Communications: Linear Equalisation of MIMO Channels

DSP Applications for Wireless Communications: Linear Equalisation of MIMO Channels DSP Applications for Wireless Communications: Dr. Waleed Al-Hanafy waleed alhanafy@yahoo.com Faculty of Electronic Engineering, Menoufia Univ., Egypt Digital Signal Processing (ECE407) Lecture no. 5 August

More information

EE 5345 Biomedical Instrumentation Lecture 12: slides

EE 5345 Biomedical Instrumentation Lecture 12: slides EE 5345 Biomedical Instrumentation Lecture 1: slides 4-6 Carlos E. Davila, Electrical Engineering Dept. Southern Methodist University slides can be viewed at: http:// www.seas.smu.edu/~cd/ee5345.html EE

More information

Statistical signal processing

Statistical signal processing Statistical signal processing Short overview of the fundamentals Outline Random variables Random processes Stationarity Ergodicity Spectral analysis Random variable and processes Intuition: A random variable

More information

HST.582J / 6.555J / J Biomedical Signal and Image Processing Spring 2007

HST.582J / 6.555J / J Biomedical Signal and Image Processing Spring 2007 MIT OpenCourseWare http://ocw.mit.edu HST.58J / 6.555J / 16.456J Biomedical Signal and Image Processing Spring 007 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.

More information