Adaptive Beamforming Algorithms
|
|
- Phillip Newton
- 5 years ago
- Views:
Transcription
1 S. R. Zinka October 29, 2014
2 Outline 1 Least Mean Squares 2 Sample Matrix Inversion 3 Recursive Least Squares 4 Accelerated Gradient Approach 5 Conjugate Gradient Method
3 Outline 1 Least Mean Squares 2 Sample Matrix Inversion 3 Recursive Least Squares 4 Accelerated Gradient Approach 5 Conjugate Gradient Method
4 Gradient Descent Method * x 1 x 2 x 3 x 4 x 0 * x n+1 = x n γ F (x n) (1) In general, γ is allowed to change at every iteration. However, for this algorithm we use constant γ.
5 Gradient Descent Method - γ and Convergence
6 Least Mean Squares Algorithm The squared error (between the reference signal and the array output) is given as ɛ (t) = r (t) w H x (t). (2) Taking the expected values of both sides of the equation (2) gives { ξ = E ɛ (t) 2} = E {ɛ (t) ɛ (t)} { [r = E (t) w H x (t) ] [ r (t) w H x (t) ] } = E {[ r (t) w H x (t) ] [ r (t) x H (t) w ]} { } = E r (t) 2 + w H x (t) x H (t) w w H x (t) r (t) r (t) x H (t) w { = E r (t) 2} + w H Rw w H E {x (t) r (t)} E { r (t) x H (t) } w { = E r (t) 2} + w H Rw w H z z H w, (3) where z = E {x (t) r (t)} is known as signal correlation vector.
7 Least Mean Squares Algorithm... Contd The mean square error ( { E ɛ (t) 2}) surface is a quadratic function of w and is minimized by setting its gradient with respect to w equal to zero, which gives { w E ɛ (t) 2} wopt = 0 Rw opt z = 0 w MSE = R 1 z. (4) In general, we do not know the signal statistics and thus must resort to estimating the array correlation matrix (R) and the signal correlation vector (z) over a range of snapshots or for each instant in time.
8 Least Mean Squares Algorithm... Contd The instantaneous estimates of R and z are as given below: ˆR (k) = x (k) x H (k) (5) ẑ (k) = r (k) x (k) (6)
9 Least Mean Squares Algorithm... Contd The method of steepest descent can be approximated in terms of the weights using the LMS method as shown below: w (k + 1) = w (k) γ w (ξ) = w (k) γ [Rw (k) z] w (k) γ [ ˆR (k + 1)w (k) ẑ (k + 1) ] = w (k) γ [ x (k + 1) x H (k + 1) w (k) r (k + 1) x (k + 1) ] = w (k) γ { x (k + 1) [ x H (k + 1) w (k) r (k + 1) ]} = w (k) γ {x (k + 1) [ w H (k) x (k + 1) r (k + 1) ] } = w (k) + γɛ (k + 1) x (k + 1) (7)
10 LMS - Convergence of Weight Vector According to LMS algorithm, w (k + 1) = w (k) γ [ x (k + 1) x H (k + 1) w (k) r (k + 1) x (k + 1) ]. Taking the expected value on both sides of the above equation E [w (k + 1)] = E [w (k)] γe [ x (k + 1) x H (k + 1) w (k) ] + γe [r (k + 1) x (k + 1)] w (k + 1) = w (k) γr w (k) + γz = [I γr] w (k) + γz. (8) Define a mean error vector v as v (k) = w (k) w MSE (9) where w MSE = R 1 z. Then, v (k + 1) = w (k + 1) w MSE = [I γr] w (k) + γz w MSE = [I γr] [ v (k) + w MSE] + γz w MSE = [I γr] v (k) + [I γr] w MSE + γiz Iw MSE }{{} 0 = [I γr] v (k). (10)
11 LMS - Convergence of Weight Vector... Contd From equation (10), v (k + 1) = [I γr] k+1 v (0). (11) Using eigenvalue decomposition, the above equation an be written as v (k + 1) = [ I γqλq H] k+1 v (0), (12) where Λ is a diagonal matrix of the eigenvalues and QQ H = I. The above equation can further be written as v (k + 1) = Q [I γλ] k+1 Q H v (0). (13) For convergence, as the iteration number increases, each diagonal element of the matrix [I γλ] k+1 should diminish, i.e., 1 γλ i < 1, i. (14)
12 LMS - Convergence of Weight Vector... Contd From the previous slide, step size γ should be (for all λ i values) 1 γλ i < 1 1 < (1 γλ i )< 1 2 < ( γλ i )< 0 2 > γλ i > 0 0 < γλ i < 2 0 < γ < 2 λ i 0 < γ < 2 λ max. (15) As the sum of all eigenvalues of R equals its trace, the sum of its diagonal elements, the gradient step size γ can be selected in terms of measurable quantities using γ < 2 trace (R). (16)
13 Drawbacks One of the drawbacks of the LMS adaptive scheme is that the algorithm must go through many iterations before satisfactory convergence is achieved. The rate of convergence of the weights is dictated by the eigenvalue spread of R.
14 Outline 1 Least Mean Squares 2 Sample Matrix Inversion 3 Recursive Least Squares 4 Accelerated Gradient Approach 5 Conjugate Gradient Method
15 Sample / Direct Matrix Inversion The sample matrix is a time average estimate of the array correlation, matrix using K time samples. For MMSE optimal criteria, optimum weight vector is given as where R = E [ xx H] and z = E [r x]. w MSE = R 1 z, (17) We can estimate the correlation matrix by calculating the time average such that N 2 R ˆR = 1 x (i) x H (i), (18) K i=n 1 where N 2 N 1 = K is the observation interval. Similarly, signal correlation vector can be estimated as z ẑ = 1 N 2 K r (i) x (i). (19) i=n 1 Since we use a K length block of data, this method is called a block adaptive approach. We are thus adapting the weights block-by-block.
16 Calculation of Estimations of R and z Let s define a matrix X K as shown below: X K = x 1 (N 1) x 1 (N 1 + 1) x 1 (N 1 + K) x 2 (N 1) x 2 (N 1 + 1) x L (N 1)... x L (N 1 + K) (20) Then the estimate of the array correlation matrix is given by ˆR = 1 K X KX H K. (21) Similarly, the estimate of the correlation vector is given by ẑ = 1 K rh X K, (22) where r = [ r (N 1) r (N 1 + 1) r (N 1 + K) ].
17 SMI Weights w MSE = ˆR 1 ẑ = [ X K X H ] 1 [ K r H ] X K (23)
18 SMI - Drawbacks The correlation matrix may be ill conditioned resulting in errors or singularities when inverted Computationally intensive because for large arrays, there is the challenge of inverting large matrices
19 Outline 1 Least Mean Squares 2 Sample Matrix Inversion 3 Recursive Least Squares 4 Accelerated Gradient Approach 5 Conjugate Gradient Method
20 Recursive Least Squares Following a similar estimation method used in SMI algorithm, one can estimate R and z as shown below: k R ˆR (k) = x (i) x H (i), (24) i=1 where k is the block length. z ẑ (k) = k r (i) x (i). (25) i=1
21 Recursive Least Squares... Weighted Estimation Both (24) and (25) use rectangular windows, thus they equally consider all previous time samples. Since the signal sources can change or slowly move with time, we might want to deemphasize the earliest data samples and emphasize the most recent ones as shown below: ˆR (k) = ẑ (k) = where α is the forgetting factor such that 0 α 1. k α k i x (i) x H (i), (26) i=1 k α k i r (i) x (i). (27) i=1
22 Recursive Least Squares... Recursive Formulation ˆR (k) = k i=1 α k i x (i) x H (i) = x (k) x H k 1 (k) + α k i x (i) x H (i) i=1 = x (k) x H k 1 (k) + α α k 1 i x (i) x H (i) i=1 = α ˆR (k 1) + x (k) x H (k) ẑ (k) = k i=1 α k i r (i) x (i) = αẑ (k 1) + r (k) x (k)
23 Recursive Least Squares... Recursive Formulation... Contd where g (k) = ˆR (k) = α ˆR (k 1) + x (k) x H (k) ˆR 1 (k) = [ α ˆR (k 1) + x (k) x H (k) ] 1 = α 1 ˆR 1 (k 1) α 1 ˆR 1 (k 1) x (k) x H (k) α 1 ˆR 1 (k 1) 1 + x H (k) α 1 ˆR 1 (k 1) x (k) = α 1 ˆR 1 (k 1) α 2 ˆR 1 (k 1) x (k) x H (k) ˆR 1 (k 1) 1 + α 1 x H (k) ˆR 1 (k 1) x (k) = α 1 ˆR 1 (k 1) α 1 ˆR 1 (k 1) x (k) 1 + α 1 x H (k) ˆR 1 (k 1) x (k) α 1 x H (k) ˆR 1 (k 1) = α 1 ˆR 1 (k 1) α 1 g (k) x H (k) ˆR 1 (k 1) α 1 ˆR 1 (k 1)x(k) 1+α 1 x H (k) ˆR 1. We can also prove that (k 1)x(k) g (k) = [ α 1 ˆR 1 (k 1) α 1 g (k) x H (k) ˆR 1 (k 1) ] x (k) = ˆR 1 (k) x (k).
24 Recursive Least Squares... Recursive Formulation... Contd w (k) = ˆR 1 (k) ẑ (k) = ˆR 1 (k) [αẑ (k 1) + r (k) x (k)] = α ˆR 1 (k) ẑ (k 1) + r (k) ˆR 1 (k) x (k) = α [ α 1 ˆR 1 (k 1) α 1 g (k) x H (k) ˆR 1 (k 1) ] ẑ (k 1) + r (k) ˆR 1 (k) x (k) = [ ˆR 1 (k 1) g (k) x H (k) ˆR 1 (k 1) ] ẑ (k 1) + r (k) ˆR 1 (k) x (k) = ˆR 1 (k 1) ẑ (k 1) g (k) x H (k) ˆR 1 (k 1) ẑ (k 1) + r (k) ˆR 1 (k) x (k) = w (k 1) g (k) x H (k) w (k 1) + r (k) g (k) = w (k 1) + g (k) [ r (k) x H (k) w (k 1) ] = w (k 1) + g (k) ɛ (k + 1) (28)
25 Outline 1 Least Mean Squares 2 Sample Matrix Inversion 3 Recursive Least Squares 4 Accelerated Gradient Approach 5 Conjugate Gradient Method
26 Quadratic Form F (x) = x H Ax b H x x H b + c x F (x) = Ax b
27 The Method of Steepest Descent * x 1 x 2 x 3 x 4 x 0 * x n+1 = x n γ F (x n) (29) Unlike LMS method, in AGM, that the value of γ is allowed to change at every iteration.
28 What is the optimum γ value?
29 The Method of Steepest Descent
30 2 The Method of Steepest Descent γ { F [x n γ F (x n)]} = 0 i.e., γ { F [x n γ [Ax n b]]} = 0
31 The Method of Steepest Descent F [x n+1] = [x n γ [Ax n b]] H A [x n γ [Ax n b]] b H [x n γ [Ax n b]] [x n γ [Ax n b]] H b So, df dγ = xh n A [Ax n b] [Ax n b] H Ax n + 2γ [Ax n b] H A [Ax n b] + b H [Ax n b] + [Ax n b] H b = x H n A [Ax n b] + b H [Ax n b] [Ax n b] H Ax n + [Ax n b] H b + 2γ [Ax n b] H A [Ax n b] = [ x H n A + b H] [Ax n b] [Ax n b] H [Ax n b] + 2γ [Ax n b] H A [Ax n b] = [Ax n b] H [Ax n b] [Ax n b] H [Ax n b] + 2γ [Ax n b] H A [Ax n b] = 2 [Ax n b] H [Ax n b] + 2γ [Ax n b] H A [Ax n b]. Since df dγ = 0, γ = [Ax n b] H [Ax n b] [Ax n b] H A [Ax n b].
32 Consecutive Orthogonal Paths (Ax n b) H (Ax n+1 b) = 0 (Ax n b) H {A [x n γ (Ax n b)] b} = 0 (Ax n b) H {[Ax n γa (Ax n b)] b} = 0 (Ax n b) H {[(Ax n b) γa (Ax n b)]} = 0 {[ ]} (Ax n b) H (Ax n b) γ (Ax n b) H A (Ax n b) = 0 γ = (Ax n b) H (Ax n b) (Ax n b) H A (Ax n b).
33 So, Accelerated Gradient Method is... If we approximate array correlation matrix and signal correlation vectors as R ˆR = x (k + 1) x H (k + 1) and z ẑ = r (k + 1) x (k + 1), w (k + 1) = w (k) γ w (ξ) = w (k) γ [Rw (k) z] ( ) [Rw (k) z] H [Rw (k) z] = w (k) [Rw (k) z] H [Rw (k) z] R [Rw (k) z] ( [ ˆRw (k) ẑ ] H [ ˆRw (k) ẑ ] ) [ w (k) [ ˆRw (k) ẑ ]H ˆR [ ˆRw (k) ẑ ] ˆRw (k) ẑ ] ( = w (k) + ɛ (k + 1) [ = w (k) + ɛ (k + 1) x H (k + 1) x (k + 1) because [Rw (k) z] = ɛ (k + 1) x (k + 1). x H (k + 1) x (k + 1) x H (k + 1) x (k + 1) x H (k + 1) x (k + 1) ) x (k + 1) ] x (k + 1), (30)
34 Outline 1 Least Mean Squares 2 Sample Matrix Inversion 3 Recursive Least Squares 4 Accelerated Gradient Approach 5 Conjugate Gradient Method
35 Steepest Descent vs Conjugate Gradient
36 Before proceeding further, let s recap Gram-Schmidt method...
37 Gram-Schmidt Conjugation d (0) = u 0 d (1) = u 1 β 10 d (0) u 0 u 1 d(0) + u u * d(0) d(1) d (2) = u 2 β 20 d (0) β 21 d (1) i 1 d (i) = u i β ij d (j) j=0 β ij = dh (j) Au i d H (j) Ad (j)
38 Now, let s enter into Conjugate Gradient method...
39 Problem Statement Let s consider a convex cost function as shown below: F (x) = x H Ax b H x x H b + c (31) Let s say that the optimal solution which minimizes the above cost function is x. Then, Ax b = 0. (32) So, solving the cost function for it s stationary point is equivalent to solving the above system of linear equations.
40 Conjugate Gradient Method
41 Expressing Error in terms of Search Vectors e 0 = x x 0 = α 1 p 1 + α 2 p 2 + α 3 p α n p n e 1 = x x 1 = α 2 p 2 + α 3 p α n p n.. e n 1 = x x n 1 = α n p n If {p k } is a simple orthogonal set, α k = ph k e k 1 p H k p. (33) k Also, it can be noted that e k is orthogonal to p k, p k 1, p k 2, etc. However, we can t use this approach because we don t know the value of x. If we knew x in the beginning itself, then we wouldn t worry about solving the optimization problem at all. So, we can t use this approach.
42 Expressing Residual in terms of Search Vectors r 0 = b Ax 0 = α 1 Ap 1 + α 2 Ap 2 + α 3 Ap α n Ap n r 1 = b Ax 1 = α 2 Ap 2 + α 3 Ap α n Ap n.. r n 1 = b Ax n 1 = α n Ap n If {p k } A is an orthogonal (conjugate) set with respect to matrix A, α k = ph k r k 1 p H k Ap. (34) k Instead of e k, in this case, r k is orthogonal to p k, p k 1, p k 2, etc.
43 How to Generate the Orthogonal Set Till now, we assumed that we knew the conjugate (i.e., A orthogonal) set {p k } A. There are several ways to generate this set, both iterative and non-iterative. We will choose Gram-Schmidt iterative process to generate this set.
44 Gram-Schmidt Conjugation It can be easily shown that r 0, r 1, r 2,, r n 1 are independent of each other. So, we can use these residuals to construct an A-orthogonal search set {p 1, p 2, p 3,, p n} A. So, k p k+1 = r k β kj p j (35) j=1 and β kj = ph j Ar k p H j Ap j. (36)
45 So, the Basic Algorithm is... Choose a starting position x 0 Calculate the residual at x 0 and assign this residual vector to p 1, i.e., p 1 = r 0. Now, calculate the value of α 1 using p 1 and r 0 according to the equation (34) Move to the next position according to the equation x k = x k 1 + α k p k Once you have reached to the next position, calculate the next search direction using r k and {p 1, p 2,, p k } according to the equation (36) Repeat this process until you reach x
46 Let s further refine conjugate gradient algorithm...
47 The Residual is Orthogonal to all Previous Residuals Search directions are dictated by the following equation: k p k+1 = r k β kj p j j=1 By taking Hermitian on both sides, we get j=1 p H k k+1 = rh k β kj ph j. Multiply the above equation on both sides by r k+1 to get p H k+1 r k+1 = r H k k r k+1 β kj ph j r k+1. j=1 So, 0 = r H k r k+1 0 r H k r k+1 = 0 (37)
48 Writing α k in terms of Residuals From (34), α k = ph k r k 1 p H k Ap. k Let s re-formulate the above equation in terms of residuals as shown below: r H k r k 1 = 0 (r k 1 α k Ap k ) H r k 1 = 0 ( r H k 1 α k ph k AH) r k 1 = 0 r H k 1 r k 1 α k ph k AH r k 1 = 0 αk = rh k 1 r k 1 p H k AH r k 1 α k = rh k 1 r k 1 r H k 1 Ap k r H k 1 α k = r k 1 ( ) H p k + k 1 j=1 β k 1jp j Apk α k = rh k 1 r k 1 p H k Ap k (38)
49 Writing β kj in terms of Residuals We know that r j = r j 1 α j Ap j. So, r H j = r H j 1 α j ph j A H. Multiply the above equation on both sides by r k. Then we get r H j r k = r H j 1 r k α j ph j A H r k. When A is Hermitian r H j r k = r H j 1 r k α j ph j Ar k r H j r k = r H j 1 r k α j β kjp H j Ap j r H j r k = r H j 1 r k β kj r H j 1 r j 1 β kj = rh j 1 r k r H j r k r H j 1 r. j 1
50 Writing β kj in terms of Residuals... Contd Since j k while calculating β kj values the term r H j 1 r k is always zero. So, β kj = 0 when j < k and rh k r k β kk = r H k 1 r. (39) k 1 So, finally p k+1 = r k β kk p k. (40)
51 Basic Version vs Refined Version α k = ph k r k 1 p H k Ap k k p k+1 = r k β kj p j j=1 β kj = ph j Ar k p H j Ap j α k = rh k 1 r k 1 p H k Ap k p k+1 = r k β kk p k rh k r k β kk = r H k 1 r k 1
52 If A is not Hermitian? Actual problem is solving Ax = b. If A is not Hermitian, then we can reformulate the above problem as shown below: A H Ax = A H b A 1 x = b 1 (41) where A 1 = A H A and b 1 = A H b. Now, we can see that conjugate gradient method can be applied to (41). Also, for this new problem, r new k = b 1 A 1 x k = A H (b A 1 x k ) = A H r k. So, algorithm is α k = ( A H r k 1 ) H ( A H r k 1 ) p H k AH Ap k p k+1 = r new k β kk p k = A H r k β kk p k ( ) A H H ( ) r k A H r k β kk = (A H r k 1 ) H (A H r k 1 )
53 So, Adaptive Algorithm using CGM is... w (k + 1) = w (k) + α k+1 p k+1 p k+1 = R H r k β kk p k where ( ) R H H ( ) r k 1 R H r k 1 α k = p H k RH Rp k ( ) R H H ( ) r k R H r k β kk = (R H r k 1 ) H (R H r k 1 ) r k = z Rw (k) ẑ ˆRw (k) = ɛ (k + 1) x (k + 1)
V. Adaptive filtering Widrow-Hopf Learning Rule LMS and Adaline
V. Adaptive filtering Widrow-Hopf Learning Rule LMS and Adaline Goals Introduce Wiener-Hopf (WH) equations Introduce application of the steepest descent method to the WH problem Approximation to the Least
More informationNumerical Methods - Numerical Linear Algebra
Numerical Methods - Numerical Linear Algebra Y. K. Goh Universiti Tunku Abdul Rahman 2013 Y. K. Goh (UTAR) Numerical Methods - Numerical Linear Algebra I 2013 1 / 62 Outline 1 Motivation 2 Solving Linear
More informationMathematical optimization
Optimization Mathematical optimization Determine the best solutions to certain mathematically defined problems that are under constrained determine optimality criteria determine the convergence of the
More informationIterative Methods for Solving A x = b
Iterative Methods for Solving A x = b A good (free) online source for iterative methods for solving A x = b is given in the description of a set of iterative solvers called templates found at netlib: http
More informationChapter 7 Iterative Techniques in Matrix Algebra
Chapter 7 Iterative Techniques in Matrix Algebra Per-Olof Persson persson@berkeley.edu Department of Mathematics University of California, Berkeley Math 128B Numerical Analysis Vector Norms Definition
More informationComputational Methods CMSC/AMSC/MAPL 460. Eigenvalues and Eigenvectors. Ramani Duraiswami, Dept. of Computer Science
Computational Methods CMSC/AMSC/MAPL 460 Eigenvalues and Eigenvectors Ramani Duraiswami, Dept. of Computer Science Eigen Values of a Matrix Recap: A N N matrix A has an eigenvector x (non-zero) with corresponding
More informationProperties of Matrices and Operations on Matrices
Properties of Matrices and Operations on Matrices A common data structure for statistical analysis is a rectangular array or matris. Rows represent individual observational units, or just observations,
More informationConjugate Gradient (CG) Method
Conjugate Gradient (CG) Method by K. Ozawa 1 Introduction In the series of this lecture, I will introduce the conjugate gradient method, which solves efficiently large scale sparse linear simultaneous
More informationLinear Algebra Massoud Malek
CSUEB Linear Algebra Massoud Malek Inner Product and Normed Space In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n An inner product
More information, b = 0. (2) 1 2 The eigenvectors of A corresponding to the eigenvalues λ 1 = 1, λ 2 = 3 are
Quadratic forms We consider the quadratic function f : R 2 R defined by f(x) = 2 xt Ax b T x with x = (x, x 2 ) T, () where A R 2 2 is symmetric and b R 2. We will see that, depending on the eigenvalues
More informationConjugate Gradient Method
Conjugate Gradient Method Hung M Phan UMass Lowell April 13, 2017 Throughout, A R n n is symmetric and positive definite, and b R n 1 Steepest Descent Method We present the steepest descent method for
More informationAdaptive Filter Theory
0 Adaptive Filter heory Sung Ho Cho Hanyang University Seoul, Korea (Office) +8--0-0390 (Mobile) +8-10-541-5178 dragon@hanyang.ac.kr able of Contents 1 Wiener Filters Gradient Search by Steepest Descent
More informationThe Conjugate Gradient Method
The Conjugate Gradient Method Jason E. Hicken Aerospace Design Lab Department of Aeronautics & Astronautics Stanford University 14 July 2011 Lecture Objectives describe when CG can be used to solve Ax
More informationConstrained optimization. Unconstrained optimization. One-dimensional. Multi-dimensional. Newton with equality constraints. Active-set method.
Optimization Unconstrained optimization One-dimensional Multi-dimensional Newton s method Basic Newton Gauss- Newton Quasi- Newton Descent methods Gradient descent Conjugate gradient Constrained optimization
More information7.2 Steepest Descent and Preconditioning
7.2 Steepest Descent and Preconditioning Descent methods are a broad class of iterative methods for finding solutions of the linear system Ax = b for symmetric positive definite matrix A R n n. Consider
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra)
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 21: Sensitivity of Eigenvalues and Eigenvectors; Conjugate Gradient Method Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical Analysis
More informationNotes on Some Methods for Solving Linear Systems
Notes on Some Methods for Solving Linear Systems Dianne P. O Leary, 1983 and 1999 and 2007 September 25, 2007 When the matrix A is symmetric and positive definite, we have a whole new class of algorithms
More informationStatistical and Adaptive Signal Processing
r Statistical and Adaptive Signal Processing Spectral Estimation, Signal Modeling, Adaptive Filtering and Array Processing Dimitris G. Manolakis Massachusetts Institute of Technology Lincoln Laboratory
More informationMath 5630: Conjugate Gradient Method Hung M. Phan, UMass Lowell March 29, 2019
Math 563: Conjugate Gradient Method Hung M. Phan, UMass Lowell March 29, 219 hroughout, A R n n is symmetric and positive definite, and b R n. 1 Steepest Descent Method We present the steepest descent
More informationCHAPTER 6. Projection Methods. Let A R n n. Solve Ax = f. Find an approximate solution ˆx K such that r = f Aˆx L.
Projection Methods CHAPTER 6 Let A R n n. Solve Ax = f. Find an approximate solution ˆx K such that r = f Aˆx L. V (n m) = [v, v 2,..., v m ] basis of K W (n m) = [w, w 2,..., w m ] basis of L Let x 0
More informationPreliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012
Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.
More informationPart 3: Trust-region methods for unconstrained optimization. Nick Gould (RAL)
Part 3: Trust-region methods for unconstrained optimization Nick Gould (RAL) minimize x IR n f(x) MSc course on nonlinear optimization UNCONSTRAINED MINIMIZATION minimize x IR n f(x) where the objective
More informationR-Linear Convergence of Limited Memory Steepest Descent
R-Linear Convergence of Limited Memory Steepest Descent Frank E. Curtis, Lehigh University joint work with Wei Guo, Lehigh University OP17 Vancouver, British Columbia, Canada 24 May 2017 R-Linear Convergence
More informationAdaptive Filtering. Squares. Alexander D. Poularikas. Fundamentals of. Least Mean. with MATLABR. University of Alabama, Huntsville, AL.
Adaptive Filtering Fundamentals of Least Mean Squares with MATLABR Alexander D. Poularikas University of Alabama, Huntsville, AL CRC Press Taylor & Francis Croup Boca Raton London New York CRC Press is
More informationIterative Methods for Smooth Objective Functions
Optimization Iterative Methods for Smooth Objective Functions Quadratic Objective Functions Stationary Iterative Methods (first/second order) Steepest Descent Method Landweber/Projected Landweber Methods
More informationNotes on PCG for Sparse Linear Systems
Notes on PCG for Sparse Linear Systems Luca Bergamaschi Department of Civil Environmental and Architectural Engineering University of Padova e-mail luca.bergamaschi@unipd.it webpage www.dmsa.unipd.it/
More informationCh4: Method of Steepest Descent
Ch4: Method of Steepest Descent The method of steepest descent is recursive in the sense that starting from some initial (arbitrary) value for the tap-weight vector, it improves with the increased number
More informationCS137 Introduction to Scientific Computing Winter Quarter 2004 Solutions to Homework #3
CS137 Introduction to Scientific Computing Winter Quarter 2004 Solutions to Homework #3 Felix Kwok February 27, 2004 Written Problems 1. (Heath E3.10) Let B be an n n matrix, and assume that B is both
More informationEE482: Digital Signal Processing Applications
Professor Brendan Morris, SEB 3216, brendan.morris@unlv.edu EE482: Digital Signal Processing Applications Spring 2014 TTh 14:30-15:45 CBC C222 Lecture 11 Adaptive Filtering 14/03/04 http://www.ee.unlv.edu/~b1morris/ee482/
More informationC&O367: Nonlinear Optimization (Winter 2013) Assignment 4 H. Wolkowicz
C&O367: Nonlinear Optimization (Winter 013) Assignment 4 H. Wolkowicz Posted Mon, Feb. 8 Due: Thursday, Feb. 8 10:00AM (before class), 1 Matrices 1.1 Positive Definite Matrices 1. Let A S n, i.e., let
More information7.3 The Jacobi and Gauss-Siedel Iterative Techniques. Problem: To solve Ax = b for A R n n. Methodology: Iteratively approximate solution x. No GEPP.
7.3 The Jacobi and Gauss-Siedel Iterative Techniques Problem: To solve Ax = b for A R n n. Methodology: Iteratively approximate solution x. No GEPP. 7.3 The Jacobi and Gauss-Siedel Iterative Techniques
More informationIterative methods for Linear System
Iterative methods for Linear System JASS 2009 Student: Rishi Patil Advisor: Prof. Thomas Huckle Outline Basics: Matrices and their properties Eigenvalues, Condition Number Iterative Methods Direct and
More informationProgramming, numerics and optimization
Programming, numerics and optimization Lecture C-3: Unconstrained optimization II Łukasz Jankowski ljank@ippt.pan.pl Institute of Fundamental Technological Research Room 4.32, Phone +22.8261281 ext. 428
More informationThe Singular Value Decomposition
The Singular Value Decomposition Philippe B. Laval KSU Fall 2015 Philippe B. Laval (KSU) SVD Fall 2015 1 / 13 Review of Key Concepts We review some key definitions and results about matrices that will
More informationSample ECE275A Midterm Exam Questions
Sample ECE275A Midterm Exam Questions The questions given below are actual problems taken from exams given in in the past few years. Solutions to these problems will NOT be provided. These problems and
More informationLecture 22. r i+1 = b Ax i+1 = b A(x i + α i r i ) =(b Ax i ) α i Ar i = r i α i Ar i
8.409 An Algorithmist s oolkit December, 009 Lecturer: Jonathan Kelner Lecture Last time Last time, we reduced solving sparse systems of linear equations Ax = b where A is symmetric and positive definite
More informationLeast-Squares Systems and The QR factorization
Least-Squares Systems and The QR factorization Orthogonality Least-squares systems. The Gram-Schmidt and Modified Gram-Schmidt processes. The Householder QR and the Givens QR. Orthogonality The Gram-Schmidt
More informationEECS 275 Matrix Computation
EECS 275 Matrix Computation Ming-Hsuan Yang Electrical Engineering and Computer Science University of California at Merced Merced, CA 95344 http://faculty.ucmerced.edu/mhyang Lecture 20 1 / 20 Overview
More informationMath Fall Final Exam
Math 104 - Fall 2008 - Final Exam Name: Student ID: Signature: Instructions: Print your name and student ID number, write your signature to indicate that you accept the honor code. During the test, you
More informationA Quick Tour of Linear Algebra and Optimization for Machine Learning
A Quick Tour of Linear Algebra and Optimization for Machine Learning Masoud Farivar January 8, 2015 1 / 28 Outline of Part I: Review of Basic Linear Algebra Matrices and Vectors Matrix Multiplication Operators
More informationLinear Algebra, part 3 QR and SVD
Linear Algebra, part 3 QR and SVD Anna-Karin Tornberg Mathematical Models, Analysis and Simulation Fall semester, 2012 Going back to least squares (Section 1.4 from Strang, now also see section 5.2). We
More informationAlgebra C Numerical Linear Algebra Sample Exam Problems
Algebra C Numerical Linear Algebra Sample Exam Problems Notation. Denote by V a finite-dimensional Hilbert space with inner product (, ) and corresponding norm. The abbreviation SPD is used for symmetric
More informationI. Multiple Choice Questions (Answer any eight)
Name of the student : Roll No : CS65: Linear Algebra and Random Processes Exam - Course Instructor : Prashanth L.A. Date : Sep-24, 27 Duration : 5 minutes INSTRUCTIONS: The test will be evaluated ONLY
More information6.252 NONLINEAR PROGRAMMING LECTURE 10 ALTERNATIVES TO GRADIENT PROJECTION LECTURE OUTLINE. Three Alternatives/Remedies for Gradient Projection
6.252 NONLINEAR PROGRAMMING LECTURE 10 ALTERNATIVES TO GRADIENT PROJECTION LECTURE OUTLINE Three Alternatives/Remedies for Gradient Projection Two-Metric Projection Methods Manifold Suboptimization Methods
More informationNumerical Methods I Solving Square Linear Systems: GEM and LU factorization
Numerical Methods I Solving Square Linear Systems: GEM and LU factorization Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 September 18th,
More informationLecture 11. Linear systems: Cholesky method. Eigensystems: Terminology. Jacobi transformations QR transformation
Lecture Cholesky method QR decomposition Terminology Linear systems: Eigensystems: Jacobi transformations QR transformation Cholesky method: For a symmetric positive definite matrix, one can do an LU decomposition
More informationM.A. Botchev. September 5, 2014
Rome-Moscow school of Matrix Methods and Applied Linear Algebra 2014 A short introduction to Krylov subspaces for linear systems, matrix functions and inexact Newton methods. Plan and exercises. M.A. Botchev
More information4.8 Arnoldi Iteration, Krylov Subspaces and GMRES
48 Arnoldi Iteration, Krylov Subspaces and GMRES We start with the problem of using a similarity transformation to convert an n n matrix A to upper Hessenberg form H, ie, A = QHQ, (30) with an appropriate
More informationSIMON FRASER UNIVERSITY School of Engineering Science
SIMON FRASER UNIVERSITY School of Engineering Science Course Outline ENSC 810-3 Digital Signal Processing Calendar Description This course covers advanced digital signal processing techniques. The main
More informationConjugate Gradients: Idea
Overview Steepest Descent often takes steps in the same direction as earlier steps Wouldn t it be better every time we take a step to get it exactly right the first time? Again, in general we choose a
More informationECE580 Partial Solution to Problem Set 3
ECE580 Fall 2015 Solution to Problem Set 3 October 23, 2015 1 ECE580 Partial Solution to Problem Set 3 These problems are from the textbook by Chong and Zak, 4th edition, which is the textbook for the
More informationImage Registration Lecture 2: Vectors and Matrices
Image Registration Lecture 2: Vectors and Matrices Prof. Charlene Tsai Lecture Overview Vectors Matrices Basics Orthogonal matrices Singular Value Decomposition (SVD) 2 1 Preliminary Comments Some of this
More informationCheng Soon Ong & Christian Walder. Canberra February June 2017
Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2017 (Many figures from C. M. Bishop, "Pattern Recognition and ") 1of 141 Part III
More informationLecture 6, Sci. Comp. for DPhil Students
Lecture 6, Sci. Comp. for DPhil Students Nick Trefethen, Thursday 1.11.18 Today II.3 QR factorization II.4 Computation of the QR factorization II.5 Linear least-squares Handouts Quiz 4 Householder s 4-page
More informationCheat Sheet for MATH461
Cheat Sheet for MATH46 Here is the stuff you really need to remember for the exams Linear systems Ax = b Problem: We consider a linear system of m equations for n unknowns x,,x n : For a given matrix A
More informationMaths for Signals and Systems Linear Algebra in Engineering
Maths for Signals and Systems Linear Algebra in Engineering Lectures 13 15, Tuesday 8 th and Friday 11 th November 016 DR TANIA STATHAKI READER (ASSOCIATE PROFFESOR) IN SIGNAL PROCESSING IMPERIAL COLLEGE
More informationImproving the Convergence of Back-Propogation Learning with Second Order Methods
the of Back-Propogation Learning with Second Order Methods Sue Becker and Yann le Cun, Sept 1988 Kasey Bray, October 2017 Table of Contents 1 with Back-Propagation 2 the of BP 3 A Computationally Feasible
More informationComputational Linear Algebra
Computational Linear Algebra PD Dr. rer. nat. habil. Ralf-Peter Mundani Computation in Engineering / BGU Scientific Computing in Computer Science / INF Winter Term 2018/19 Part 4: Iterative Methods PD
More informationReview of Linear Algebra
Review of Linear Algebra VBS/MRC Review of Linear Algebra 0 Ok, the Questions Why does algebra sound like cobra? Why do they say abstract algebra? Why bother about all this abstract stuff? What is a vector,
More informationAn Iterative Descent Method
Conjugate Gradient: An Iterative Descent Method The Plan Review Iterative Descent Conjugate Gradient Review : Iterative Descent Iterative Descent is an unconstrained optimization process x (k+1) = x (k)
More information26. Filtering. ECE 830, Spring 2014
26. Filtering ECE 830, Spring 2014 1 / 26 Wiener Filtering Wiener filtering is the application of LMMSE estimation to recovery of a signal in additive noise under wide sense sationarity assumptions. Problem
More informationLecture # 20 The Preconditioned Conjugate Gradient Method
Lecture # 20 The Preconditioned Conjugate Gradient Method We wish to solve Ax = b (1) A R n n is symmetric and positive definite (SPD). We then of n are being VERY LARGE, say, n = 10 6 or n = 10 7. Usually,
More informationGradient Methods Using Momentum and Memory
Chapter 3 Gradient Methods Using Momentum and Memory The steepest descent method described in Chapter always steps in the negative gradient direction, which is orthogonal to the boundary of the level set
More informationConjugate gradient method. Descent method. Conjugate search direction. Conjugate Gradient Algorithm (294)
Conjugate gradient method Descent method Hestenes, Stiefel 1952 For A N N SPD In exact arithmetic, solves in N steps In real arithmetic No guaranteed stopping Often converges in many fewer than N steps
More informationMATH 31 - ADDITIONAL PRACTICE PROBLEMS FOR FINAL
MATH 3 - ADDITIONAL PRACTICE PROBLEMS FOR FINAL MAIN TOPICS FOR THE FINAL EXAM:. Vectors. Dot product. Cross product. Geometric applications. 2. Row reduction. Null space, column space, row space, left
More informationFurther Mathematical Methods (Linear Algebra)
Further Mathematical Methods (Linear Algebra) Solutions For The 2 Examination Question (a) For a non-empty subset W of V to be a subspace of V we require that for all vectors x y W and all scalars α R:
More informationMaxwell Equations Waves Helmholtz Wave Equation Dirac Delta Sources Power Matrices Complex Analysis Optimization. EM & Math - Basics
EM & Math - S. R. Zinka srinivasa_zinka@daiict.ac.in October 16, 2014 Outline 1 Maxwell Equations 2 Waves 3 Helmholtz Wave Equation 4 Dirac Delta Sources 5 Power 6 Matrices 7 Complex Analysis 8 Optimization
More informationWe describe the generalization of Hazan s algorithm for symmetric programming
ON HAZAN S ALGORITHM FOR SYMMETRIC PROGRAMMING PROBLEMS L. FAYBUSOVICH Abstract. problems We describe the generalization of Hazan s algorithm for symmetric programming Key words. Symmetric programming,
More informationLecture notes on Quantum Computing. Chapter 1 Mathematical Background
Lecture notes on Quantum Computing Chapter 1 Mathematical Background Vector states of a quantum system with n physical states are represented by unique vectors in C n, the set of n 1 column vectors 1 For
More informationLecture 17 Methods for System of Linear Equations: Part 2. Songting Luo. Department of Mathematics Iowa State University
Lecture 17 Methods for System of Linear Equations: Part 2 Songting Luo Department of Mathematics Iowa State University MATH 481 Numerical Methods for Differential Equations Songting Luo ( Department of
More informationFoundations of Matrix Analysis
1 Foundations of Matrix Analysis In this chapter we recall the basic elements of linear algebra which will be employed in the remainder of the text For most of the proofs as well as for the details, the
More informationAdvanced Digital Signal Processing -Introduction
Advanced Digital Signal Processing -Introduction LECTURE-2 1 AP9211- ADVANCED DIGITAL SIGNAL PROCESSING UNIT I DISCRETE RANDOM SIGNAL PROCESSING Discrete Random Processes- Ensemble Averages, Stationary
More information5.3 The Power Method Approximation of the Eigenvalue of Largest Module
192 5 Approximation of Eigenvalues and Eigenvectors 5.3 The Power Method The power method is very good at approximating the extremal eigenvalues of the matrix, that is, the eigenvalues having largest and
More informationApplied Linear Algebra in Geoscience Using MATLAB
Applied Linear Algebra in Geoscience Using MATLAB Contents Getting Started Creating Arrays Mathematical Operations with Arrays Using Script Files and Managing Data Two-Dimensional Plots Programming in
More informationKrylov Subspace Methods that Are Based on the Minimization of the Residual
Chapter 5 Krylov Subspace Methods that Are Based on the Minimization of the Residual Remark 51 Goal he goal of these methods consists in determining x k x 0 +K k r 0,A such that the corresponding Euclidean
More informationNumerical Optimization
Numerical Optimization Unit 2: Multivariable optimization problems Che-Rung Lee Scribe: February 28, 2011 (UNIT 2) Numerical Optimization February 28, 2011 1 / 17 Partial derivative of a two variable function
More informationMathematical Optimisation, Chpt 2: Linear Equations and inequalities
Mathematical Optimisation, Chpt 2: Linear Equations and inequalities Peter J.C. Dickinson p.j.c.dickinson@utwente.nl http://dickinson.website version: 12/02/18 Monday 5th February 2018 Peter J.C. Dickinson
More information4.6 Iterative Solvers for Linear Systems
4.6 Iterative Solvers for Linear Systems Why use iterative methods? Virtually all direct methods for solving Ax = b require O(n 3 ) floating point operations. In practical applications the matrix A often
More informationMATHEMATICS FOR COMPUTER VISION WEEK 8 OPTIMISATION PART 2. Dr Fabio Cuzzolin MSc in Computer Vision Oxford Brookes University Year
MATHEMATICS FOR COMPUTER VISION WEEK 8 OPTIMISATION PART 2 1 Dr Fabio Cuzzolin MSc in Computer Vision Oxford Brookes University Year 2013-14 OUTLINE OF WEEK 8 topics: quadratic optimisation, least squares,
More informationConjugate Gradients I: Setup
Conjugate Gradients I: Setup CS 205A: Mathematical Methods for Robotics, Vision, and Graphics Justin Solomon CS 205A: Mathematical Methods Conjugate Gradients I: Setup 1 / 22 Time for Gaussian Elimination
More informationLecture Notes 6: Dynamic Equations Part C: Linear Difference Equation Systems
University of Warwick, EC9A0 Maths for Economists Peter J. Hammond 1 of 45 Lecture Notes 6: Dynamic Equations Part C: Linear Difference Equation Systems Peter J. Hammond latest revision 2017 September
More informationTopics. The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems
Topics The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems What about non-spd systems? Methods requiring small history Methods requiring large history Summary of solvers 1 / 52 Conjugate
More informationFundamentals of Matrices
Maschinelles Lernen II Fundamentals of Matrices Christoph Sawade/Niels Landwehr/Blaine Nelson Tobias Scheffer Matrix Examples Recap: Data Linear Model: f i x = w i T x Let X = x x n be the data matrix
More information6.4 Krylov Subspaces and Conjugate Gradients
6.4 Krylov Subspaces and Conjugate Gradients Our original equation is Ax = b. The preconditioned equation is P Ax = P b. When we write P, we never intend that an inverse will be explicitly computed. P
More informationthe method of steepest descent
MATH 3511 Spring 2018 the method of steepest descent http://www.phys.uconn.edu/ rozman/courses/m3511_18s/ Last modified: February 6, 2018 Abstract The Steepest Descent is an iterative method for solving
More informationThe QR Factorization
The QR Factorization How to Make Matrices Nicer Radu Trîmbiţaş Babeş-Bolyai University March 11, 2009 Radu Trîmbiţaş ( Babeş-Bolyai University) The QR Factorization March 11, 2009 1 / 25 Projectors A projector
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra)
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 3: Positive-Definite Systems; Cholesky Factorization Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical Analysis I 1 / 11 Symmetric
More informationECE 595, Section 10 Numerical Simulations Lecture 7: Optimization and Eigenvalues. Prof. Peter Bermel January 23, 2013
ECE 595, Section 10 Numerical Simulations Lecture 7: Optimization and Eigenvalues Prof. Peter Bermel January 23, 2013 Outline Recap from Friday Optimization Methods Brent s Method Golden Section Search
More informationMath 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008
Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008 Exam 2 will be held on Tuesday, April 8, 7-8pm in 117 MacMillan What will be covered The exam will cover material from the lectures
More informationNPTEL NPTEL ONLINE CERTIFICATION COURSE. Introduction to Machine Learning. Lecture 6
(Refer Slide Time: 00:15) NPTEL NPTEL ONLINE CERTIFICATION COURSE Introduction to Machine Learning Lecture 6 Prof. Balaraman Ravibdran Computer Science and Engineering Indian institute of technology Statistical
More informationEE482: Digital Signal Processing Applications
Professor Brendan Morris, SEB 3216, brendan.morris@unlv.edu EE482: Digital Signal Processing Applications Spring 2014 TTh 14:30-15:45 CBC C222 Lecture 11 Adaptive Filtering 14/03/04 http://www.ee.unlv.edu/~b1morris/ee482/
More informationFEM and sparse linear system solving
FEM & sparse linear system solving, Lecture 9, Nov 19, 2017 1/36 Lecture 9, Nov 17, 2017: Krylov space methods http://people.inf.ethz.ch/arbenz/fem17 Peter Arbenz Computer Science Department, ETH Zürich
More informationScientific Computing
Scientific Computing Direct solution methods Martin van Gijzen Delft University of Technology October 3, 2018 1 Program October 3 Matrix norms LU decomposition Basic algorithm Cost Stability Pivoting Pivoting
More informationScientific Computing II
Technische Universität München ST 008 Institut für Informatik Dr. Miriam Mehl Scientific Computing II Final Exam, July, 008 Iterative Solvers (3 pts + 4 extra pts, 60 min) a) Steepest Descent and Conjugate
More informationParametric Signal Modeling and Linear Prediction Theory 4. The Levinson-Durbin Recursion
Parametric Signal Modeling and Linear Prediction Theory 4. The Levinson-Durbin Recursion Electrical & Computer Engineering North Carolina State University Acknowledgment: ECE792-41 slides were adapted
More informationAPPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.
APPENDIX A Background Mathematics A. Linear Algebra A.. Vector algebra Let x denote the n-dimensional column vector with components 0 x x 2 B C @. A x n Definition 6 (scalar product). The scalar product
More information17 Solution of Nonlinear Systems
17 Solution of Nonlinear Systems We now discuss the solution of systems of nonlinear equations. An important ingredient will be the multivariate Taylor theorem. Theorem 17.1 Let D = {x 1, x 2,..., x m
More informationGradient Descent Methods
Lab 18 Gradient Descent Methods Lab Objective: Many optimization methods fall under the umbrella of descent algorithms. The idea is to choose an initial guess, identify a direction from this point along
More informationLECTURE 18. Lecture outline Gaussian channels: parallel colored noise inter-symbol interference general case: multiple inputs and outputs
LECTURE 18 Last time: White Gaussian noise Bandlimited WGN Additive White Gaussian Noise (AWGN) channel Capacity of AWGN channel Application: DS-CDMA systems Spreading Coding theorem Lecture outline Gaussian
More informationLecture 3: Linesearch methods (continued). Steepest descent methods
Lecture 3: Linesearch methods (continued). Steepest descent methods Coralia Cartis, Mathematical Institute, University of Oxford C6.2/B2: Continuous Optimization Lecture 3: Linesearch methods (continued).
More information