Adaptive Beamforming Algorithms

Size: px
Start display at page:

Download "Adaptive Beamforming Algorithms"

Transcription

1 S. R. Zinka October 29, 2014

2 Outline 1 Least Mean Squares 2 Sample Matrix Inversion 3 Recursive Least Squares 4 Accelerated Gradient Approach 5 Conjugate Gradient Method

3 Outline 1 Least Mean Squares 2 Sample Matrix Inversion 3 Recursive Least Squares 4 Accelerated Gradient Approach 5 Conjugate Gradient Method

4 Gradient Descent Method * x 1 x 2 x 3 x 4 x 0 * x n+1 = x n γ F (x n) (1) In general, γ is allowed to change at every iteration. However, for this algorithm we use constant γ.

5 Gradient Descent Method - γ and Convergence

6 Least Mean Squares Algorithm The squared error (between the reference signal and the array output) is given as ɛ (t) = r (t) w H x (t). (2) Taking the expected values of both sides of the equation (2) gives { ξ = E ɛ (t) 2} = E {ɛ (t) ɛ (t)} { [r = E (t) w H x (t) ] [ r (t) w H x (t) ] } = E {[ r (t) w H x (t) ] [ r (t) x H (t) w ]} { } = E r (t) 2 + w H x (t) x H (t) w w H x (t) r (t) r (t) x H (t) w { = E r (t) 2} + w H Rw w H E {x (t) r (t)} E { r (t) x H (t) } w { = E r (t) 2} + w H Rw w H z z H w, (3) where z = E {x (t) r (t)} is known as signal correlation vector.

7 Least Mean Squares Algorithm... Contd The mean square error ( { E ɛ (t) 2}) surface is a quadratic function of w and is minimized by setting its gradient with respect to w equal to zero, which gives { w E ɛ (t) 2} wopt = 0 Rw opt z = 0 w MSE = R 1 z. (4) In general, we do not know the signal statistics and thus must resort to estimating the array correlation matrix (R) and the signal correlation vector (z) over a range of snapshots or for each instant in time.

8 Least Mean Squares Algorithm... Contd The instantaneous estimates of R and z are as given below: ˆR (k) = x (k) x H (k) (5) ẑ (k) = r (k) x (k) (6)

9 Least Mean Squares Algorithm... Contd The method of steepest descent can be approximated in terms of the weights using the LMS method as shown below: w (k + 1) = w (k) γ w (ξ) = w (k) γ [Rw (k) z] w (k) γ [ ˆR (k + 1)w (k) ẑ (k + 1) ] = w (k) γ [ x (k + 1) x H (k + 1) w (k) r (k + 1) x (k + 1) ] = w (k) γ { x (k + 1) [ x H (k + 1) w (k) r (k + 1) ]} = w (k) γ {x (k + 1) [ w H (k) x (k + 1) r (k + 1) ] } = w (k) + γɛ (k + 1) x (k + 1) (7)

10 LMS - Convergence of Weight Vector According to LMS algorithm, w (k + 1) = w (k) γ [ x (k + 1) x H (k + 1) w (k) r (k + 1) x (k + 1) ]. Taking the expected value on both sides of the above equation E [w (k + 1)] = E [w (k)] γe [ x (k + 1) x H (k + 1) w (k) ] + γe [r (k + 1) x (k + 1)] w (k + 1) = w (k) γr w (k) + γz = [I γr] w (k) + γz. (8) Define a mean error vector v as v (k) = w (k) w MSE (9) where w MSE = R 1 z. Then, v (k + 1) = w (k + 1) w MSE = [I γr] w (k) + γz w MSE = [I γr] [ v (k) + w MSE] + γz w MSE = [I γr] v (k) + [I γr] w MSE + γiz Iw MSE }{{} 0 = [I γr] v (k). (10)

11 LMS - Convergence of Weight Vector... Contd From equation (10), v (k + 1) = [I γr] k+1 v (0). (11) Using eigenvalue decomposition, the above equation an be written as v (k + 1) = [ I γqλq H] k+1 v (0), (12) where Λ is a diagonal matrix of the eigenvalues and QQ H = I. The above equation can further be written as v (k + 1) = Q [I γλ] k+1 Q H v (0). (13) For convergence, as the iteration number increases, each diagonal element of the matrix [I γλ] k+1 should diminish, i.e., 1 γλ i < 1, i. (14)

12 LMS - Convergence of Weight Vector... Contd From the previous slide, step size γ should be (for all λ i values) 1 γλ i < 1 1 < (1 γλ i )< 1 2 < ( γλ i )< 0 2 > γλ i > 0 0 < γλ i < 2 0 < γ < 2 λ i 0 < γ < 2 λ max. (15) As the sum of all eigenvalues of R equals its trace, the sum of its diagonal elements, the gradient step size γ can be selected in terms of measurable quantities using γ < 2 trace (R). (16)

13 Drawbacks One of the drawbacks of the LMS adaptive scheme is that the algorithm must go through many iterations before satisfactory convergence is achieved. The rate of convergence of the weights is dictated by the eigenvalue spread of R.

14 Outline 1 Least Mean Squares 2 Sample Matrix Inversion 3 Recursive Least Squares 4 Accelerated Gradient Approach 5 Conjugate Gradient Method

15 Sample / Direct Matrix Inversion The sample matrix is a time average estimate of the array correlation, matrix using K time samples. For MMSE optimal criteria, optimum weight vector is given as where R = E [ xx H] and z = E [r x]. w MSE = R 1 z, (17) We can estimate the correlation matrix by calculating the time average such that N 2 R ˆR = 1 x (i) x H (i), (18) K i=n 1 where N 2 N 1 = K is the observation interval. Similarly, signal correlation vector can be estimated as z ẑ = 1 N 2 K r (i) x (i). (19) i=n 1 Since we use a K length block of data, this method is called a block adaptive approach. We are thus adapting the weights block-by-block.

16 Calculation of Estimations of R and z Let s define a matrix X K as shown below: X K = x 1 (N 1) x 1 (N 1 + 1) x 1 (N 1 + K) x 2 (N 1) x 2 (N 1 + 1) x L (N 1)... x L (N 1 + K) (20) Then the estimate of the array correlation matrix is given by ˆR = 1 K X KX H K. (21) Similarly, the estimate of the correlation vector is given by ẑ = 1 K rh X K, (22) where r = [ r (N 1) r (N 1 + 1) r (N 1 + K) ].

17 SMI Weights w MSE = ˆR 1 ẑ = [ X K X H ] 1 [ K r H ] X K (23)

18 SMI - Drawbacks The correlation matrix may be ill conditioned resulting in errors or singularities when inverted Computationally intensive because for large arrays, there is the challenge of inverting large matrices

19 Outline 1 Least Mean Squares 2 Sample Matrix Inversion 3 Recursive Least Squares 4 Accelerated Gradient Approach 5 Conjugate Gradient Method

20 Recursive Least Squares Following a similar estimation method used in SMI algorithm, one can estimate R and z as shown below: k R ˆR (k) = x (i) x H (i), (24) i=1 where k is the block length. z ẑ (k) = k r (i) x (i). (25) i=1

21 Recursive Least Squares... Weighted Estimation Both (24) and (25) use rectangular windows, thus they equally consider all previous time samples. Since the signal sources can change or slowly move with time, we might want to deemphasize the earliest data samples and emphasize the most recent ones as shown below: ˆR (k) = ẑ (k) = where α is the forgetting factor such that 0 α 1. k α k i x (i) x H (i), (26) i=1 k α k i r (i) x (i). (27) i=1

22 Recursive Least Squares... Recursive Formulation ˆR (k) = k i=1 α k i x (i) x H (i) = x (k) x H k 1 (k) + α k i x (i) x H (i) i=1 = x (k) x H k 1 (k) + α α k 1 i x (i) x H (i) i=1 = α ˆR (k 1) + x (k) x H (k) ẑ (k) = k i=1 α k i r (i) x (i) = αẑ (k 1) + r (k) x (k)

23 Recursive Least Squares... Recursive Formulation... Contd where g (k) = ˆR (k) = α ˆR (k 1) + x (k) x H (k) ˆR 1 (k) = [ α ˆR (k 1) + x (k) x H (k) ] 1 = α 1 ˆR 1 (k 1) α 1 ˆR 1 (k 1) x (k) x H (k) α 1 ˆR 1 (k 1) 1 + x H (k) α 1 ˆR 1 (k 1) x (k) = α 1 ˆR 1 (k 1) α 2 ˆR 1 (k 1) x (k) x H (k) ˆR 1 (k 1) 1 + α 1 x H (k) ˆR 1 (k 1) x (k) = α 1 ˆR 1 (k 1) α 1 ˆR 1 (k 1) x (k) 1 + α 1 x H (k) ˆR 1 (k 1) x (k) α 1 x H (k) ˆR 1 (k 1) = α 1 ˆR 1 (k 1) α 1 g (k) x H (k) ˆR 1 (k 1) α 1 ˆR 1 (k 1)x(k) 1+α 1 x H (k) ˆR 1. We can also prove that (k 1)x(k) g (k) = [ α 1 ˆR 1 (k 1) α 1 g (k) x H (k) ˆR 1 (k 1) ] x (k) = ˆR 1 (k) x (k).

24 Recursive Least Squares... Recursive Formulation... Contd w (k) = ˆR 1 (k) ẑ (k) = ˆR 1 (k) [αẑ (k 1) + r (k) x (k)] = α ˆR 1 (k) ẑ (k 1) + r (k) ˆR 1 (k) x (k) = α [ α 1 ˆR 1 (k 1) α 1 g (k) x H (k) ˆR 1 (k 1) ] ẑ (k 1) + r (k) ˆR 1 (k) x (k) = [ ˆR 1 (k 1) g (k) x H (k) ˆR 1 (k 1) ] ẑ (k 1) + r (k) ˆR 1 (k) x (k) = ˆR 1 (k 1) ẑ (k 1) g (k) x H (k) ˆR 1 (k 1) ẑ (k 1) + r (k) ˆR 1 (k) x (k) = w (k 1) g (k) x H (k) w (k 1) + r (k) g (k) = w (k 1) + g (k) [ r (k) x H (k) w (k 1) ] = w (k 1) + g (k) ɛ (k + 1) (28)

25 Outline 1 Least Mean Squares 2 Sample Matrix Inversion 3 Recursive Least Squares 4 Accelerated Gradient Approach 5 Conjugate Gradient Method

26 Quadratic Form F (x) = x H Ax b H x x H b + c x F (x) = Ax b

27 The Method of Steepest Descent * x 1 x 2 x 3 x 4 x 0 * x n+1 = x n γ F (x n) (29) Unlike LMS method, in AGM, that the value of γ is allowed to change at every iteration.

28 What is the optimum γ value?

29 The Method of Steepest Descent

30 2 The Method of Steepest Descent γ { F [x n γ F (x n)]} = 0 i.e., γ { F [x n γ [Ax n b]]} = 0

31 The Method of Steepest Descent F [x n+1] = [x n γ [Ax n b]] H A [x n γ [Ax n b]] b H [x n γ [Ax n b]] [x n γ [Ax n b]] H b So, df dγ = xh n A [Ax n b] [Ax n b] H Ax n + 2γ [Ax n b] H A [Ax n b] + b H [Ax n b] + [Ax n b] H b = x H n A [Ax n b] + b H [Ax n b] [Ax n b] H Ax n + [Ax n b] H b + 2γ [Ax n b] H A [Ax n b] = [ x H n A + b H] [Ax n b] [Ax n b] H [Ax n b] + 2γ [Ax n b] H A [Ax n b] = [Ax n b] H [Ax n b] [Ax n b] H [Ax n b] + 2γ [Ax n b] H A [Ax n b] = 2 [Ax n b] H [Ax n b] + 2γ [Ax n b] H A [Ax n b]. Since df dγ = 0, γ = [Ax n b] H [Ax n b] [Ax n b] H A [Ax n b].

32 Consecutive Orthogonal Paths (Ax n b) H (Ax n+1 b) = 0 (Ax n b) H {A [x n γ (Ax n b)] b} = 0 (Ax n b) H {[Ax n γa (Ax n b)] b} = 0 (Ax n b) H {[(Ax n b) γa (Ax n b)]} = 0 {[ ]} (Ax n b) H (Ax n b) γ (Ax n b) H A (Ax n b) = 0 γ = (Ax n b) H (Ax n b) (Ax n b) H A (Ax n b).

33 So, Accelerated Gradient Method is... If we approximate array correlation matrix and signal correlation vectors as R ˆR = x (k + 1) x H (k + 1) and z ẑ = r (k + 1) x (k + 1), w (k + 1) = w (k) γ w (ξ) = w (k) γ [Rw (k) z] ( ) [Rw (k) z] H [Rw (k) z] = w (k) [Rw (k) z] H [Rw (k) z] R [Rw (k) z] ( [ ˆRw (k) ẑ ] H [ ˆRw (k) ẑ ] ) [ w (k) [ ˆRw (k) ẑ ]H ˆR [ ˆRw (k) ẑ ] ˆRw (k) ẑ ] ( = w (k) + ɛ (k + 1) [ = w (k) + ɛ (k + 1) x H (k + 1) x (k + 1) because [Rw (k) z] = ɛ (k + 1) x (k + 1). x H (k + 1) x (k + 1) x H (k + 1) x (k + 1) x H (k + 1) x (k + 1) ) x (k + 1) ] x (k + 1), (30)

34 Outline 1 Least Mean Squares 2 Sample Matrix Inversion 3 Recursive Least Squares 4 Accelerated Gradient Approach 5 Conjugate Gradient Method

35 Steepest Descent vs Conjugate Gradient

36 Before proceeding further, let s recap Gram-Schmidt method...

37 Gram-Schmidt Conjugation d (0) = u 0 d (1) = u 1 β 10 d (0) u 0 u 1 d(0) + u u * d(0) d(1) d (2) = u 2 β 20 d (0) β 21 d (1) i 1 d (i) = u i β ij d (j) j=0 β ij = dh (j) Au i d H (j) Ad (j)

38 Now, let s enter into Conjugate Gradient method...

39 Problem Statement Let s consider a convex cost function as shown below: F (x) = x H Ax b H x x H b + c (31) Let s say that the optimal solution which minimizes the above cost function is x. Then, Ax b = 0. (32) So, solving the cost function for it s stationary point is equivalent to solving the above system of linear equations.

40 Conjugate Gradient Method

41 Expressing Error in terms of Search Vectors e 0 = x x 0 = α 1 p 1 + α 2 p 2 + α 3 p α n p n e 1 = x x 1 = α 2 p 2 + α 3 p α n p n.. e n 1 = x x n 1 = α n p n If {p k } is a simple orthogonal set, α k = ph k e k 1 p H k p. (33) k Also, it can be noted that e k is orthogonal to p k, p k 1, p k 2, etc. However, we can t use this approach because we don t know the value of x. If we knew x in the beginning itself, then we wouldn t worry about solving the optimization problem at all. So, we can t use this approach.

42 Expressing Residual in terms of Search Vectors r 0 = b Ax 0 = α 1 Ap 1 + α 2 Ap 2 + α 3 Ap α n Ap n r 1 = b Ax 1 = α 2 Ap 2 + α 3 Ap α n Ap n.. r n 1 = b Ax n 1 = α n Ap n If {p k } A is an orthogonal (conjugate) set with respect to matrix A, α k = ph k r k 1 p H k Ap. (34) k Instead of e k, in this case, r k is orthogonal to p k, p k 1, p k 2, etc.

43 How to Generate the Orthogonal Set Till now, we assumed that we knew the conjugate (i.e., A orthogonal) set {p k } A. There are several ways to generate this set, both iterative and non-iterative. We will choose Gram-Schmidt iterative process to generate this set.

44 Gram-Schmidt Conjugation It can be easily shown that r 0, r 1, r 2,, r n 1 are independent of each other. So, we can use these residuals to construct an A-orthogonal search set {p 1, p 2, p 3,, p n} A. So, k p k+1 = r k β kj p j (35) j=1 and β kj = ph j Ar k p H j Ap j. (36)

45 So, the Basic Algorithm is... Choose a starting position x 0 Calculate the residual at x 0 and assign this residual vector to p 1, i.e., p 1 = r 0. Now, calculate the value of α 1 using p 1 and r 0 according to the equation (34) Move to the next position according to the equation x k = x k 1 + α k p k Once you have reached to the next position, calculate the next search direction using r k and {p 1, p 2,, p k } according to the equation (36) Repeat this process until you reach x

46 Let s further refine conjugate gradient algorithm...

47 The Residual is Orthogonal to all Previous Residuals Search directions are dictated by the following equation: k p k+1 = r k β kj p j j=1 By taking Hermitian on both sides, we get j=1 p H k k+1 = rh k β kj ph j. Multiply the above equation on both sides by r k+1 to get p H k+1 r k+1 = r H k k r k+1 β kj ph j r k+1. j=1 So, 0 = r H k r k+1 0 r H k r k+1 = 0 (37)

48 Writing α k in terms of Residuals From (34), α k = ph k r k 1 p H k Ap. k Let s re-formulate the above equation in terms of residuals as shown below: r H k r k 1 = 0 (r k 1 α k Ap k ) H r k 1 = 0 ( r H k 1 α k ph k AH) r k 1 = 0 r H k 1 r k 1 α k ph k AH r k 1 = 0 αk = rh k 1 r k 1 p H k AH r k 1 α k = rh k 1 r k 1 r H k 1 Ap k r H k 1 α k = r k 1 ( ) H p k + k 1 j=1 β k 1jp j Apk α k = rh k 1 r k 1 p H k Ap k (38)

49 Writing β kj in terms of Residuals We know that r j = r j 1 α j Ap j. So, r H j = r H j 1 α j ph j A H. Multiply the above equation on both sides by r k. Then we get r H j r k = r H j 1 r k α j ph j A H r k. When A is Hermitian r H j r k = r H j 1 r k α j ph j Ar k r H j r k = r H j 1 r k α j β kjp H j Ap j r H j r k = r H j 1 r k β kj r H j 1 r j 1 β kj = rh j 1 r k r H j r k r H j 1 r. j 1

50 Writing β kj in terms of Residuals... Contd Since j k while calculating β kj values the term r H j 1 r k is always zero. So, β kj = 0 when j < k and rh k r k β kk = r H k 1 r. (39) k 1 So, finally p k+1 = r k β kk p k. (40)

51 Basic Version vs Refined Version α k = ph k r k 1 p H k Ap k k p k+1 = r k β kj p j j=1 β kj = ph j Ar k p H j Ap j α k = rh k 1 r k 1 p H k Ap k p k+1 = r k β kk p k rh k r k β kk = r H k 1 r k 1

52 If A is not Hermitian? Actual problem is solving Ax = b. If A is not Hermitian, then we can reformulate the above problem as shown below: A H Ax = A H b A 1 x = b 1 (41) where A 1 = A H A and b 1 = A H b. Now, we can see that conjugate gradient method can be applied to (41). Also, for this new problem, r new k = b 1 A 1 x k = A H (b A 1 x k ) = A H r k. So, algorithm is α k = ( A H r k 1 ) H ( A H r k 1 ) p H k AH Ap k p k+1 = r new k β kk p k = A H r k β kk p k ( ) A H H ( ) r k A H r k β kk = (A H r k 1 ) H (A H r k 1 )

53 So, Adaptive Algorithm using CGM is... w (k + 1) = w (k) + α k+1 p k+1 p k+1 = R H r k β kk p k where ( ) R H H ( ) r k 1 R H r k 1 α k = p H k RH Rp k ( ) R H H ( ) r k R H r k β kk = (R H r k 1 ) H (R H r k 1 ) r k = z Rw (k) ẑ ˆRw (k) = ɛ (k + 1) x (k + 1)

V. Adaptive filtering Widrow-Hopf Learning Rule LMS and Adaline

V. Adaptive filtering Widrow-Hopf Learning Rule LMS and Adaline V. Adaptive filtering Widrow-Hopf Learning Rule LMS and Adaline Goals Introduce Wiener-Hopf (WH) equations Introduce application of the steepest descent method to the WH problem Approximation to the Least

More information

Numerical Methods - Numerical Linear Algebra

Numerical Methods - Numerical Linear Algebra Numerical Methods - Numerical Linear Algebra Y. K. Goh Universiti Tunku Abdul Rahman 2013 Y. K. Goh (UTAR) Numerical Methods - Numerical Linear Algebra I 2013 1 / 62 Outline 1 Motivation 2 Solving Linear

More information

Mathematical optimization

Mathematical optimization Optimization Mathematical optimization Determine the best solutions to certain mathematically defined problems that are under constrained determine optimality criteria determine the convergence of the

More information

Iterative Methods for Solving A x = b

Iterative Methods for Solving A x = b Iterative Methods for Solving A x = b A good (free) online source for iterative methods for solving A x = b is given in the description of a set of iterative solvers called templates found at netlib: http

More information

Chapter 7 Iterative Techniques in Matrix Algebra

Chapter 7 Iterative Techniques in Matrix Algebra Chapter 7 Iterative Techniques in Matrix Algebra Per-Olof Persson persson@berkeley.edu Department of Mathematics University of California, Berkeley Math 128B Numerical Analysis Vector Norms Definition

More information

Computational Methods CMSC/AMSC/MAPL 460. Eigenvalues and Eigenvectors. Ramani Duraiswami, Dept. of Computer Science

Computational Methods CMSC/AMSC/MAPL 460. Eigenvalues and Eigenvectors. Ramani Duraiswami, Dept. of Computer Science Computational Methods CMSC/AMSC/MAPL 460 Eigenvalues and Eigenvectors Ramani Duraiswami, Dept. of Computer Science Eigen Values of a Matrix Recap: A N N matrix A has an eigenvector x (non-zero) with corresponding

More information

Properties of Matrices and Operations on Matrices

Properties of Matrices and Operations on Matrices Properties of Matrices and Operations on Matrices A common data structure for statistical analysis is a rectangular array or matris. Rows represent individual observational units, or just observations,

More information

Conjugate Gradient (CG) Method

Conjugate Gradient (CG) Method Conjugate Gradient (CG) Method by K. Ozawa 1 Introduction In the series of this lecture, I will introduce the conjugate gradient method, which solves efficiently large scale sparse linear simultaneous

More information

Linear Algebra Massoud Malek

Linear Algebra Massoud Malek CSUEB Linear Algebra Massoud Malek Inner Product and Normed Space In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n An inner product

More information

, b = 0. (2) 1 2 The eigenvectors of A corresponding to the eigenvalues λ 1 = 1, λ 2 = 3 are

, b = 0. (2) 1 2 The eigenvectors of A corresponding to the eigenvalues λ 1 = 1, λ 2 = 3 are Quadratic forms We consider the quadratic function f : R 2 R defined by f(x) = 2 xt Ax b T x with x = (x, x 2 ) T, () where A R 2 2 is symmetric and b R 2. We will see that, depending on the eigenvalues

More information

Conjugate Gradient Method

Conjugate Gradient Method Conjugate Gradient Method Hung M Phan UMass Lowell April 13, 2017 Throughout, A R n n is symmetric and positive definite, and b R n 1 Steepest Descent Method We present the steepest descent method for

More information

Adaptive Filter Theory

Adaptive Filter Theory 0 Adaptive Filter heory Sung Ho Cho Hanyang University Seoul, Korea (Office) +8--0-0390 (Mobile) +8-10-541-5178 dragon@hanyang.ac.kr able of Contents 1 Wiener Filters Gradient Search by Steepest Descent

More information

The Conjugate Gradient Method

The Conjugate Gradient Method The Conjugate Gradient Method Jason E. Hicken Aerospace Design Lab Department of Aeronautics & Astronautics Stanford University 14 July 2011 Lecture Objectives describe when CG can be used to solve Ax

More information

Constrained optimization. Unconstrained optimization. One-dimensional. Multi-dimensional. Newton with equality constraints. Active-set method.

Constrained optimization. Unconstrained optimization. One-dimensional. Multi-dimensional. Newton with equality constraints. Active-set method. Optimization Unconstrained optimization One-dimensional Multi-dimensional Newton s method Basic Newton Gauss- Newton Quasi- Newton Descent methods Gradient descent Conjugate gradient Constrained optimization

More information

7.2 Steepest Descent and Preconditioning

7.2 Steepest Descent and Preconditioning 7.2 Steepest Descent and Preconditioning Descent methods are a broad class of iterative methods for finding solutions of the linear system Ax = b for symmetric positive definite matrix A R n n. Consider

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra)

AMS526: Numerical Analysis I (Numerical Linear Algebra) AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 21: Sensitivity of Eigenvalues and Eigenvectors; Conjugate Gradient Method Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical Analysis

More information

Notes on Some Methods for Solving Linear Systems

Notes on Some Methods for Solving Linear Systems Notes on Some Methods for Solving Linear Systems Dianne P. O Leary, 1983 and 1999 and 2007 September 25, 2007 When the matrix A is symmetric and positive definite, we have a whole new class of algorithms

More information

Statistical and Adaptive Signal Processing

Statistical and Adaptive Signal Processing r Statistical and Adaptive Signal Processing Spectral Estimation, Signal Modeling, Adaptive Filtering and Array Processing Dimitris G. Manolakis Massachusetts Institute of Technology Lincoln Laboratory

More information

Math 5630: Conjugate Gradient Method Hung M. Phan, UMass Lowell March 29, 2019

Math 5630: Conjugate Gradient Method Hung M. Phan, UMass Lowell March 29, 2019 Math 563: Conjugate Gradient Method Hung M. Phan, UMass Lowell March 29, 219 hroughout, A R n n is symmetric and positive definite, and b R n. 1 Steepest Descent Method We present the steepest descent

More information

CHAPTER 6. Projection Methods. Let A R n n. Solve Ax = f. Find an approximate solution ˆx K such that r = f Aˆx L.

CHAPTER 6. Projection Methods. Let A R n n. Solve Ax = f. Find an approximate solution ˆx K such that r = f Aˆx L. Projection Methods CHAPTER 6 Let A R n n. Solve Ax = f. Find an approximate solution ˆx K such that r = f Aˆx L. V (n m) = [v, v 2,..., v m ] basis of K W (n m) = [w, w 2,..., w m ] basis of L Let x 0

More information

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.

More information

Part 3: Trust-region methods for unconstrained optimization. Nick Gould (RAL)

Part 3: Trust-region methods for unconstrained optimization. Nick Gould (RAL) Part 3: Trust-region methods for unconstrained optimization Nick Gould (RAL) minimize x IR n f(x) MSc course on nonlinear optimization UNCONSTRAINED MINIMIZATION minimize x IR n f(x) where the objective

More information

R-Linear Convergence of Limited Memory Steepest Descent

R-Linear Convergence of Limited Memory Steepest Descent R-Linear Convergence of Limited Memory Steepest Descent Frank E. Curtis, Lehigh University joint work with Wei Guo, Lehigh University OP17 Vancouver, British Columbia, Canada 24 May 2017 R-Linear Convergence

More information

Adaptive Filtering. Squares. Alexander D. Poularikas. Fundamentals of. Least Mean. with MATLABR. University of Alabama, Huntsville, AL.

Adaptive Filtering. Squares. Alexander D. Poularikas. Fundamentals of. Least Mean. with MATLABR. University of Alabama, Huntsville, AL. Adaptive Filtering Fundamentals of Least Mean Squares with MATLABR Alexander D. Poularikas University of Alabama, Huntsville, AL CRC Press Taylor & Francis Croup Boca Raton London New York CRC Press is

More information

Iterative Methods for Smooth Objective Functions

Iterative Methods for Smooth Objective Functions Optimization Iterative Methods for Smooth Objective Functions Quadratic Objective Functions Stationary Iterative Methods (first/second order) Steepest Descent Method Landweber/Projected Landweber Methods

More information

Notes on PCG for Sparse Linear Systems

Notes on PCG for Sparse Linear Systems Notes on PCG for Sparse Linear Systems Luca Bergamaschi Department of Civil Environmental and Architectural Engineering University of Padova e-mail luca.bergamaschi@unipd.it webpage www.dmsa.unipd.it/

More information

Ch4: Method of Steepest Descent

Ch4: Method of Steepest Descent Ch4: Method of Steepest Descent The method of steepest descent is recursive in the sense that starting from some initial (arbitrary) value for the tap-weight vector, it improves with the increased number

More information

CS137 Introduction to Scientific Computing Winter Quarter 2004 Solutions to Homework #3

CS137 Introduction to Scientific Computing Winter Quarter 2004 Solutions to Homework #3 CS137 Introduction to Scientific Computing Winter Quarter 2004 Solutions to Homework #3 Felix Kwok February 27, 2004 Written Problems 1. (Heath E3.10) Let B be an n n matrix, and assume that B is both

More information

EE482: Digital Signal Processing Applications

EE482: Digital Signal Processing Applications Professor Brendan Morris, SEB 3216, brendan.morris@unlv.edu EE482: Digital Signal Processing Applications Spring 2014 TTh 14:30-15:45 CBC C222 Lecture 11 Adaptive Filtering 14/03/04 http://www.ee.unlv.edu/~b1morris/ee482/

More information

C&O367: Nonlinear Optimization (Winter 2013) Assignment 4 H. Wolkowicz

C&O367: Nonlinear Optimization (Winter 2013) Assignment 4 H. Wolkowicz C&O367: Nonlinear Optimization (Winter 013) Assignment 4 H. Wolkowicz Posted Mon, Feb. 8 Due: Thursday, Feb. 8 10:00AM (before class), 1 Matrices 1.1 Positive Definite Matrices 1. Let A S n, i.e., let

More information

7.3 The Jacobi and Gauss-Siedel Iterative Techniques. Problem: To solve Ax = b for A R n n. Methodology: Iteratively approximate solution x. No GEPP.

7.3 The Jacobi and Gauss-Siedel Iterative Techniques. Problem: To solve Ax = b for A R n n. Methodology: Iteratively approximate solution x. No GEPP. 7.3 The Jacobi and Gauss-Siedel Iterative Techniques Problem: To solve Ax = b for A R n n. Methodology: Iteratively approximate solution x. No GEPP. 7.3 The Jacobi and Gauss-Siedel Iterative Techniques

More information

Iterative methods for Linear System

Iterative methods for Linear System Iterative methods for Linear System JASS 2009 Student: Rishi Patil Advisor: Prof. Thomas Huckle Outline Basics: Matrices and their properties Eigenvalues, Condition Number Iterative Methods Direct and

More information

Programming, numerics and optimization

Programming, numerics and optimization Programming, numerics and optimization Lecture C-3: Unconstrained optimization II Łukasz Jankowski ljank@ippt.pan.pl Institute of Fundamental Technological Research Room 4.32, Phone +22.8261281 ext. 428

More information

The Singular Value Decomposition

The Singular Value Decomposition The Singular Value Decomposition Philippe B. Laval KSU Fall 2015 Philippe B. Laval (KSU) SVD Fall 2015 1 / 13 Review of Key Concepts We review some key definitions and results about matrices that will

More information

Sample ECE275A Midterm Exam Questions

Sample ECE275A Midterm Exam Questions Sample ECE275A Midterm Exam Questions The questions given below are actual problems taken from exams given in in the past few years. Solutions to these problems will NOT be provided. These problems and

More information

Lecture 22. r i+1 = b Ax i+1 = b A(x i + α i r i ) =(b Ax i ) α i Ar i = r i α i Ar i

Lecture 22. r i+1 = b Ax i+1 = b A(x i + α i r i ) =(b Ax i ) α i Ar i = r i α i Ar i 8.409 An Algorithmist s oolkit December, 009 Lecturer: Jonathan Kelner Lecture Last time Last time, we reduced solving sparse systems of linear equations Ax = b where A is symmetric and positive definite

More information

Least-Squares Systems and The QR factorization

Least-Squares Systems and The QR factorization Least-Squares Systems and The QR factorization Orthogonality Least-squares systems. The Gram-Schmidt and Modified Gram-Schmidt processes. The Householder QR and the Givens QR. Orthogonality The Gram-Schmidt

More information

EECS 275 Matrix Computation

EECS 275 Matrix Computation EECS 275 Matrix Computation Ming-Hsuan Yang Electrical Engineering and Computer Science University of California at Merced Merced, CA 95344 http://faculty.ucmerced.edu/mhyang Lecture 20 1 / 20 Overview

More information

Math Fall Final Exam

Math Fall Final Exam Math 104 - Fall 2008 - Final Exam Name: Student ID: Signature: Instructions: Print your name and student ID number, write your signature to indicate that you accept the honor code. During the test, you

More information

A Quick Tour of Linear Algebra and Optimization for Machine Learning

A Quick Tour of Linear Algebra and Optimization for Machine Learning A Quick Tour of Linear Algebra and Optimization for Machine Learning Masoud Farivar January 8, 2015 1 / 28 Outline of Part I: Review of Basic Linear Algebra Matrices and Vectors Matrix Multiplication Operators

More information

Linear Algebra, part 3 QR and SVD

Linear Algebra, part 3 QR and SVD Linear Algebra, part 3 QR and SVD Anna-Karin Tornberg Mathematical Models, Analysis and Simulation Fall semester, 2012 Going back to least squares (Section 1.4 from Strang, now also see section 5.2). We

More information

Algebra C Numerical Linear Algebra Sample Exam Problems

Algebra C Numerical Linear Algebra Sample Exam Problems Algebra C Numerical Linear Algebra Sample Exam Problems Notation. Denote by V a finite-dimensional Hilbert space with inner product (, ) and corresponding norm. The abbreviation SPD is used for symmetric

More information

I. Multiple Choice Questions (Answer any eight)

I. Multiple Choice Questions (Answer any eight) Name of the student : Roll No : CS65: Linear Algebra and Random Processes Exam - Course Instructor : Prashanth L.A. Date : Sep-24, 27 Duration : 5 minutes INSTRUCTIONS: The test will be evaluated ONLY

More information

6.252 NONLINEAR PROGRAMMING LECTURE 10 ALTERNATIVES TO GRADIENT PROJECTION LECTURE OUTLINE. Three Alternatives/Remedies for Gradient Projection

6.252 NONLINEAR PROGRAMMING LECTURE 10 ALTERNATIVES TO GRADIENT PROJECTION LECTURE OUTLINE. Three Alternatives/Remedies for Gradient Projection 6.252 NONLINEAR PROGRAMMING LECTURE 10 ALTERNATIVES TO GRADIENT PROJECTION LECTURE OUTLINE Three Alternatives/Remedies for Gradient Projection Two-Metric Projection Methods Manifold Suboptimization Methods

More information

Numerical Methods I Solving Square Linear Systems: GEM and LU factorization

Numerical Methods I Solving Square Linear Systems: GEM and LU factorization Numerical Methods I Solving Square Linear Systems: GEM and LU factorization Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 September 18th,

More information

Lecture 11. Linear systems: Cholesky method. Eigensystems: Terminology. Jacobi transformations QR transformation

Lecture 11. Linear systems: Cholesky method. Eigensystems: Terminology. Jacobi transformations QR transformation Lecture Cholesky method QR decomposition Terminology Linear systems: Eigensystems: Jacobi transformations QR transformation Cholesky method: For a symmetric positive definite matrix, one can do an LU decomposition

More information

M.A. Botchev. September 5, 2014

M.A. Botchev. September 5, 2014 Rome-Moscow school of Matrix Methods and Applied Linear Algebra 2014 A short introduction to Krylov subspaces for linear systems, matrix functions and inexact Newton methods. Plan and exercises. M.A. Botchev

More information

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES 48 Arnoldi Iteration, Krylov Subspaces and GMRES We start with the problem of using a similarity transformation to convert an n n matrix A to upper Hessenberg form H, ie, A = QHQ, (30) with an appropriate

More information

SIMON FRASER UNIVERSITY School of Engineering Science

SIMON FRASER UNIVERSITY School of Engineering Science SIMON FRASER UNIVERSITY School of Engineering Science Course Outline ENSC 810-3 Digital Signal Processing Calendar Description This course covers advanced digital signal processing techniques. The main

More information

Conjugate Gradients: Idea

Conjugate Gradients: Idea Overview Steepest Descent often takes steps in the same direction as earlier steps Wouldn t it be better every time we take a step to get it exactly right the first time? Again, in general we choose a

More information

ECE580 Partial Solution to Problem Set 3

ECE580 Partial Solution to Problem Set 3 ECE580 Fall 2015 Solution to Problem Set 3 October 23, 2015 1 ECE580 Partial Solution to Problem Set 3 These problems are from the textbook by Chong and Zak, 4th edition, which is the textbook for the

More information

Image Registration Lecture 2: Vectors and Matrices

Image Registration Lecture 2: Vectors and Matrices Image Registration Lecture 2: Vectors and Matrices Prof. Charlene Tsai Lecture Overview Vectors Matrices Basics Orthogonal matrices Singular Value Decomposition (SVD) 2 1 Preliminary Comments Some of this

More information

Cheng Soon Ong & Christian Walder. Canberra February June 2017

Cheng Soon Ong & Christian Walder. Canberra February June 2017 Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2017 (Many figures from C. M. Bishop, "Pattern Recognition and ") 1of 141 Part III

More information

Lecture 6, Sci. Comp. for DPhil Students

Lecture 6, Sci. Comp. for DPhil Students Lecture 6, Sci. Comp. for DPhil Students Nick Trefethen, Thursday 1.11.18 Today II.3 QR factorization II.4 Computation of the QR factorization II.5 Linear least-squares Handouts Quiz 4 Householder s 4-page

More information

Cheat Sheet for MATH461

Cheat Sheet for MATH461 Cheat Sheet for MATH46 Here is the stuff you really need to remember for the exams Linear systems Ax = b Problem: We consider a linear system of m equations for n unknowns x,,x n : For a given matrix A

More information

Maths for Signals and Systems Linear Algebra in Engineering

Maths for Signals and Systems Linear Algebra in Engineering Maths for Signals and Systems Linear Algebra in Engineering Lectures 13 15, Tuesday 8 th and Friday 11 th November 016 DR TANIA STATHAKI READER (ASSOCIATE PROFFESOR) IN SIGNAL PROCESSING IMPERIAL COLLEGE

More information

Improving the Convergence of Back-Propogation Learning with Second Order Methods

Improving the Convergence of Back-Propogation Learning with Second Order Methods the of Back-Propogation Learning with Second Order Methods Sue Becker and Yann le Cun, Sept 1988 Kasey Bray, October 2017 Table of Contents 1 with Back-Propagation 2 the of BP 3 A Computationally Feasible

More information

Computational Linear Algebra

Computational Linear Algebra Computational Linear Algebra PD Dr. rer. nat. habil. Ralf-Peter Mundani Computation in Engineering / BGU Scientific Computing in Computer Science / INF Winter Term 2018/19 Part 4: Iterative Methods PD

More information

Review of Linear Algebra

Review of Linear Algebra Review of Linear Algebra VBS/MRC Review of Linear Algebra 0 Ok, the Questions Why does algebra sound like cobra? Why do they say abstract algebra? Why bother about all this abstract stuff? What is a vector,

More information

An Iterative Descent Method

An Iterative Descent Method Conjugate Gradient: An Iterative Descent Method The Plan Review Iterative Descent Conjugate Gradient Review : Iterative Descent Iterative Descent is an unconstrained optimization process x (k+1) = x (k)

More information

26. Filtering. ECE 830, Spring 2014

26. Filtering. ECE 830, Spring 2014 26. Filtering ECE 830, Spring 2014 1 / 26 Wiener Filtering Wiener filtering is the application of LMMSE estimation to recovery of a signal in additive noise under wide sense sationarity assumptions. Problem

More information

Lecture # 20 The Preconditioned Conjugate Gradient Method

Lecture # 20 The Preconditioned Conjugate Gradient Method Lecture # 20 The Preconditioned Conjugate Gradient Method We wish to solve Ax = b (1) A R n n is symmetric and positive definite (SPD). We then of n are being VERY LARGE, say, n = 10 6 or n = 10 7. Usually,

More information

Gradient Methods Using Momentum and Memory

Gradient Methods Using Momentum and Memory Chapter 3 Gradient Methods Using Momentum and Memory The steepest descent method described in Chapter always steps in the negative gradient direction, which is orthogonal to the boundary of the level set

More information

Conjugate gradient method. Descent method. Conjugate search direction. Conjugate Gradient Algorithm (294)

Conjugate gradient method. Descent method. Conjugate search direction. Conjugate Gradient Algorithm (294) Conjugate gradient method Descent method Hestenes, Stiefel 1952 For A N N SPD In exact arithmetic, solves in N steps In real arithmetic No guaranteed stopping Often converges in many fewer than N steps

More information

MATH 31 - ADDITIONAL PRACTICE PROBLEMS FOR FINAL

MATH 31 - ADDITIONAL PRACTICE PROBLEMS FOR FINAL MATH 3 - ADDITIONAL PRACTICE PROBLEMS FOR FINAL MAIN TOPICS FOR THE FINAL EXAM:. Vectors. Dot product. Cross product. Geometric applications. 2. Row reduction. Null space, column space, row space, left

More information

Further Mathematical Methods (Linear Algebra)

Further Mathematical Methods (Linear Algebra) Further Mathematical Methods (Linear Algebra) Solutions For The 2 Examination Question (a) For a non-empty subset W of V to be a subspace of V we require that for all vectors x y W and all scalars α R:

More information

Maxwell Equations Waves Helmholtz Wave Equation Dirac Delta Sources Power Matrices Complex Analysis Optimization. EM & Math - Basics

Maxwell Equations Waves Helmholtz Wave Equation Dirac Delta Sources Power Matrices Complex Analysis Optimization. EM & Math - Basics EM & Math - S. R. Zinka srinivasa_zinka@daiict.ac.in October 16, 2014 Outline 1 Maxwell Equations 2 Waves 3 Helmholtz Wave Equation 4 Dirac Delta Sources 5 Power 6 Matrices 7 Complex Analysis 8 Optimization

More information

We describe the generalization of Hazan s algorithm for symmetric programming

We describe the generalization of Hazan s algorithm for symmetric programming ON HAZAN S ALGORITHM FOR SYMMETRIC PROGRAMMING PROBLEMS L. FAYBUSOVICH Abstract. problems We describe the generalization of Hazan s algorithm for symmetric programming Key words. Symmetric programming,

More information

Lecture notes on Quantum Computing. Chapter 1 Mathematical Background

Lecture notes on Quantum Computing. Chapter 1 Mathematical Background Lecture notes on Quantum Computing Chapter 1 Mathematical Background Vector states of a quantum system with n physical states are represented by unique vectors in C n, the set of n 1 column vectors 1 For

More information

Lecture 17 Methods for System of Linear Equations: Part 2. Songting Luo. Department of Mathematics Iowa State University

Lecture 17 Methods for System of Linear Equations: Part 2. Songting Luo. Department of Mathematics Iowa State University Lecture 17 Methods for System of Linear Equations: Part 2 Songting Luo Department of Mathematics Iowa State University MATH 481 Numerical Methods for Differential Equations Songting Luo ( Department of

More information

Foundations of Matrix Analysis

Foundations of Matrix Analysis 1 Foundations of Matrix Analysis In this chapter we recall the basic elements of linear algebra which will be employed in the remainder of the text For most of the proofs as well as for the details, the

More information

Advanced Digital Signal Processing -Introduction

Advanced Digital Signal Processing -Introduction Advanced Digital Signal Processing -Introduction LECTURE-2 1 AP9211- ADVANCED DIGITAL SIGNAL PROCESSING UNIT I DISCRETE RANDOM SIGNAL PROCESSING Discrete Random Processes- Ensemble Averages, Stationary

More information

5.3 The Power Method Approximation of the Eigenvalue of Largest Module

5.3 The Power Method Approximation of the Eigenvalue of Largest Module 192 5 Approximation of Eigenvalues and Eigenvectors 5.3 The Power Method The power method is very good at approximating the extremal eigenvalues of the matrix, that is, the eigenvalues having largest and

More information

Applied Linear Algebra in Geoscience Using MATLAB

Applied Linear Algebra in Geoscience Using MATLAB Applied Linear Algebra in Geoscience Using MATLAB Contents Getting Started Creating Arrays Mathematical Operations with Arrays Using Script Files and Managing Data Two-Dimensional Plots Programming in

More information

Krylov Subspace Methods that Are Based on the Minimization of the Residual

Krylov Subspace Methods that Are Based on the Minimization of the Residual Chapter 5 Krylov Subspace Methods that Are Based on the Minimization of the Residual Remark 51 Goal he goal of these methods consists in determining x k x 0 +K k r 0,A such that the corresponding Euclidean

More information

Numerical Optimization

Numerical Optimization Numerical Optimization Unit 2: Multivariable optimization problems Che-Rung Lee Scribe: February 28, 2011 (UNIT 2) Numerical Optimization February 28, 2011 1 / 17 Partial derivative of a two variable function

More information

Mathematical Optimisation, Chpt 2: Linear Equations and inequalities

Mathematical Optimisation, Chpt 2: Linear Equations and inequalities Mathematical Optimisation, Chpt 2: Linear Equations and inequalities Peter J.C. Dickinson p.j.c.dickinson@utwente.nl http://dickinson.website version: 12/02/18 Monday 5th February 2018 Peter J.C. Dickinson

More information

4.6 Iterative Solvers for Linear Systems

4.6 Iterative Solvers for Linear Systems 4.6 Iterative Solvers for Linear Systems Why use iterative methods? Virtually all direct methods for solving Ax = b require O(n 3 ) floating point operations. In practical applications the matrix A often

More information

MATHEMATICS FOR COMPUTER VISION WEEK 8 OPTIMISATION PART 2. Dr Fabio Cuzzolin MSc in Computer Vision Oxford Brookes University Year

MATHEMATICS FOR COMPUTER VISION WEEK 8 OPTIMISATION PART 2. Dr Fabio Cuzzolin MSc in Computer Vision Oxford Brookes University Year MATHEMATICS FOR COMPUTER VISION WEEK 8 OPTIMISATION PART 2 1 Dr Fabio Cuzzolin MSc in Computer Vision Oxford Brookes University Year 2013-14 OUTLINE OF WEEK 8 topics: quadratic optimisation, least squares,

More information

Conjugate Gradients I: Setup

Conjugate Gradients I: Setup Conjugate Gradients I: Setup CS 205A: Mathematical Methods for Robotics, Vision, and Graphics Justin Solomon CS 205A: Mathematical Methods Conjugate Gradients I: Setup 1 / 22 Time for Gaussian Elimination

More information

Lecture Notes 6: Dynamic Equations Part C: Linear Difference Equation Systems

Lecture Notes 6: Dynamic Equations Part C: Linear Difference Equation Systems University of Warwick, EC9A0 Maths for Economists Peter J. Hammond 1 of 45 Lecture Notes 6: Dynamic Equations Part C: Linear Difference Equation Systems Peter J. Hammond latest revision 2017 September

More information

Topics. The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems

Topics. The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems Topics The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems What about non-spd systems? Methods requiring small history Methods requiring large history Summary of solvers 1 / 52 Conjugate

More information

Fundamentals of Matrices

Fundamentals of Matrices Maschinelles Lernen II Fundamentals of Matrices Christoph Sawade/Niels Landwehr/Blaine Nelson Tobias Scheffer Matrix Examples Recap: Data Linear Model: f i x = w i T x Let X = x x n be the data matrix

More information

6.4 Krylov Subspaces and Conjugate Gradients

6.4 Krylov Subspaces and Conjugate Gradients 6.4 Krylov Subspaces and Conjugate Gradients Our original equation is Ax = b. The preconditioned equation is P Ax = P b. When we write P, we never intend that an inverse will be explicitly computed. P

More information

the method of steepest descent

the method of steepest descent MATH 3511 Spring 2018 the method of steepest descent http://www.phys.uconn.edu/ rozman/courses/m3511_18s/ Last modified: February 6, 2018 Abstract The Steepest Descent is an iterative method for solving

More information

The QR Factorization

The QR Factorization The QR Factorization How to Make Matrices Nicer Radu Trîmbiţaş Babeş-Bolyai University March 11, 2009 Radu Trîmbiţaş ( Babeş-Bolyai University) The QR Factorization March 11, 2009 1 / 25 Projectors A projector

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra)

AMS526: Numerical Analysis I (Numerical Linear Algebra) AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 3: Positive-Definite Systems; Cholesky Factorization Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical Analysis I 1 / 11 Symmetric

More information

ECE 595, Section 10 Numerical Simulations Lecture 7: Optimization and Eigenvalues. Prof. Peter Bermel January 23, 2013

ECE 595, Section 10 Numerical Simulations Lecture 7: Optimization and Eigenvalues. Prof. Peter Bermel January 23, 2013 ECE 595, Section 10 Numerical Simulations Lecture 7: Optimization and Eigenvalues Prof. Peter Bermel January 23, 2013 Outline Recap from Friday Optimization Methods Brent s Method Golden Section Search

More information

Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008

Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008 Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008 Exam 2 will be held on Tuesday, April 8, 7-8pm in 117 MacMillan What will be covered The exam will cover material from the lectures

More information

NPTEL NPTEL ONLINE CERTIFICATION COURSE. Introduction to Machine Learning. Lecture 6

NPTEL NPTEL ONLINE CERTIFICATION COURSE. Introduction to Machine Learning. Lecture 6 (Refer Slide Time: 00:15) NPTEL NPTEL ONLINE CERTIFICATION COURSE Introduction to Machine Learning Lecture 6 Prof. Balaraman Ravibdran Computer Science and Engineering Indian institute of technology Statistical

More information

EE482: Digital Signal Processing Applications

EE482: Digital Signal Processing Applications Professor Brendan Morris, SEB 3216, brendan.morris@unlv.edu EE482: Digital Signal Processing Applications Spring 2014 TTh 14:30-15:45 CBC C222 Lecture 11 Adaptive Filtering 14/03/04 http://www.ee.unlv.edu/~b1morris/ee482/

More information

FEM and sparse linear system solving

FEM and sparse linear system solving FEM & sparse linear system solving, Lecture 9, Nov 19, 2017 1/36 Lecture 9, Nov 17, 2017: Krylov space methods http://people.inf.ethz.ch/arbenz/fem17 Peter Arbenz Computer Science Department, ETH Zürich

More information

Scientific Computing

Scientific Computing Scientific Computing Direct solution methods Martin van Gijzen Delft University of Technology October 3, 2018 1 Program October 3 Matrix norms LU decomposition Basic algorithm Cost Stability Pivoting Pivoting

More information

Scientific Computing II

Scientific Computing II Technische Universität München ST 008 Institut für Informatik Dr. Miriam Mehl Scientific Computing II Final Exam, July, 008 Iterative Solvers (3 pts + 4 extra pts, 60 min) a) Steepest Descent and Conjugate

More information

Parametric Signal Modeling and Linear Prediction Theory 4. The Levinson-Durbin Recursion

Parametric Signal Modeling and Linear Prediction Theory 4. The Levinson-Durbin Recursion Parametric Signal Modeling and Linear Prediction Theory 4. The Levinson-Durbin Recursion Electrical & Computer Engineering North Carolina State University Acknowledgment: ECE792-41 slides were adapted

More information

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2. APPENDIX A Background Mathematics A. Linear Algebra A.. Vector algebra Let x denote the n-dimensional column vector with components 0 x x 2 B C @. A x n Definition 6 (scalar product). The scalar product

More information

17 Solution of Nonlinear Systems

17 Solution of Nonlinear Systems 17 Solution of Nonlinear Systems We now discuss the solution of systems of nonlinear equations. An important ingredient will be the multivariate Taylor theorem. Theorem 17.1 Let D = {x 1, x 2,..., x m

More information

Gradient Descent Methods

Gradient Descent Methods Lab 18 Gradient Descent Methods Lab Objective: Many optimization methods fall under the umbrella of descent algorithms. The idea is to choose an initial guess, identify a direction from this point along

More information

LECTURE 18. Lecture outline Gaussian channels: parallel colored noise inter-symbol interference general case: multiple inputs and outputs

LECTURE 18. Lecture outline Gaussian channels: parallel colored noise inter-symbol interference general case: multiple inputs and outputs LECTURE 18 Last time: White Gaussian noise Bandlimited WGN Additive White Gaussian Noise (AWGN) channel Capacity of AWGN channel Application: DS-CDMA systems Spreading Coding theorem Lecture outline Gaussian

More information

Lecture 3: Linesearch methods (continued). Steepest descent methods

Lecture 3: Linesearch methods (continued). Steepest descent methods Lecture 3: Linesearch methods (continued). Steepest descent methods Coralia Cartis, Mathematical Institute, University of Oxford C6.2/B2: Continuous Optimization Lecture 3: Linesearch methods (continued).

More information