Matrices and Moments: Least Squares Problems

Size: px
Start display at page:

Download "Matrices and Moments: Least Squares Problems"

Transcription

1 Matrices and Moments: Least Squares Problems Gene Golub, SungEun Jo, Zheng Su Comparative study (with David Gleich) Computer Science Department Stanford University

2 Matrices and Moments: Least Squares Problems Outlines Part I: Review of least squares problems and gauss quadrature rules Outline of Part I Least Squares Problems Ordinary/Data/Total least squares SVD solutions Secular equation approaches Inverse least squares Gauss Quadrature Rules Gauss quadrature theory Tri-diagonalization for orthonormal polynomials Inverse eigenvalue problem for Gauss-Radau rule Matrices and Moments: Least Squares Problems (2/52)

3 Matrices and Moments: Least Squares Problems Outlines Part II: Application to solving secular equations Outline of Part II Interpolating secular equations A common term of secular functions in TLS/DLS Variations of Newton s methods Seeking the smallest root Conjugate Gradient method with Quadrature (CGQ) Solving TLS and DLS by quadrature and CG method Alternative implementations Comparative study (with David Gleich) Large-scale TLS algorithms Comparison of large-scale TLS results Numerical example CG-based algorithms Conclusion Matrices and Moments: Least Squares Problems (3/52)

4 Least Squares Problems Gauss Quadrature Rules Part I Review Matrices and Moments: Least Squares Problems (4/52)

5 Least Squares Problems Gauss Quadrature Rules Ordinary/Data/Total least squares Approximation problem Approximation problems for a linear system: Ax b, A R m n, b R m 1, m > n. Notations: A and b x T Tr[ ] 2 A F = Tr[A T A] 2 or F A and b OLS/DLS/TLS given data solution to determine transpose the sum of diagonal entries of matrix two-norm of vector Frobenius norm of matrix Euclidean norm residual quantities ordinary/data/total least squares Matrices and Moments: Least Squares Problems (5/52)

6 Least Squares Problems Gauss Quadrature Rules Ordinary/Data/Total least squares Statements and geometric equivalences Table: Problem statements and geometric equivalent statements OLS TLS DLS Problem statement min b 2 s.t. Ax = b+ b x, b min x, A, b [ A, b] F s.t. (A+ A)x = b+ b min A F s.t. (A+ A)x = b x, A a Geometric equivalence min x Ax b 2 2 min x Ax b 2 2 x min x Ax b 2 2 x 2 2 a The TLS/DLS equivalent statements are derived by means of the Lagrange method [GV96, DD93, JK05]. Ordinary Least Squares (OLS): correcting with b Data Least Squares (DLS): correcting with A Total Least Squares (TLS): correcting with A and b TLS is also known as Errors-in-Variables modeling. Matrices and Moments: Least Squares Problems (6/52)

7 Least Squares Problems Gauss Quadrature Rules SVD solutions Singular value decomposition (SVD) approach Table: Singular value decomposition approach TLS: min x Ax b 2 x DLS: min x Ax b 2 x 2 σ 1 min ([A,b]) s.t. v TLS (n+1) 0 σ min (P b A) s.t. bt Av DLS 0 v TLS (n+1) is the last component of v TLS. P b = I 1 b T b bbt 2 1 x TLS = v TLS (n+1) v TLS(1 : n) x DLS = bt b b T v Av DLS DLS 3 [ A TLS, b TLS ] = [A,b]v TLS v T TLS A DLS = P b Av DLSv T DLS 4 [ A TLS, b TLS ] F = σ TLS A DLS F = σ DLS 1 Equivalent singular value problem and feasibility condition. v TLS and v DLS are the right singular vectors associated with the smallest singular values, respectively, of [A,b] and P b A. σ min the minimum singular value; 2 SVD solution; 3 Minimal residual in terms of singular vector; 4 Norm of minimal residuals. Matrices and Moments: Least Squares Problems (7/52)

8 Least Squares Problems Gauss Quadrature Rules Secular equation approaches Secular equation approach Table: Secular equation approach in the generic case 1 TLS: σtls 2 = min Ax b 2 x x A T (Ax b) = σ 2 TLS x b T (b Ax) = σ 2 TLS DLS: σdls 2 = min Ax b 2 x x 2 A T (Ax b) = σ 2 DLS x b T (b Ax) = 0 2 σ min (A) > σ TLS σ min (A) > σ DLS 3 b T b b T A(A T A σ 2 TLS I) 1 A T b = σ 2 TLS b T b b T A(A T A σ 2 DLS I) 1 A T b = 0 4 x TLS = (A T A σ 2 TLS I) 1 A T b x DLS = (A T A σ 2 DLS I) 1 A T b 5 [ A TLS, b TLS ] = r TLS [xt TLS, 1] x TLS 2 +1 A DLS = r DLS xt DLS x DLS 2 1 Normal equations (for stationary points); 2 Generic condition; 3 Secular equation; 4 De-regularized solution; 5 Residuals in terms of solution x and r = b Ax. Matrices and Moments: Least Squares Problems (8/52)

9 Least Squares Problems Gauss Quadrature Rules Secular equation approaches When is the secular equation approach more preferable than the SVD approach? The problem is sensitive: σ TLS/DLS σ min (A) Least squares solution or de-regularized form is needed: OLS provides a good initial guess of solution. The de-regularized form can be easily calculated by adjusting the amount of negative shift. The problem is large: The SVD of a large matrix is very expensive. Instead, we can approximate the secular equation in the large-scaled problem by Gauss quadrature rules. Matrices and Moments: Least Squares Problems (9/52)

10 Least Squares Problems Gauss Quadrature Rules Gauss quadrature theory Riemann-Stieltjes integral M = QΛQ T, Λ = diag(λ i ) 0 < λ 1 λ 2... λ n. M is symmetric positive definite. Q is orthonormal. u T f(m)u = α T f(λ)α = n f(λ i )αi 2 = I[ f] i=1 f(m) is an analytic function of M that is defined on (0, ). α = Q T u for an arbitrary vector u. Riemann-Stieltjes integral I[ f]: b 0 if λ < a = λ 1 I[ f] f(λ)dα(λ), α(λ) = i j=1 α 2 j if λ i λ < λ i+1 a n j=1 α2 j if b = λ n λ where the measure α(λ) is piecewise constant. Matrices and Moments: Least Squares Problems (10/52)

11 Least Squares Problems Gauss Quadrature Rules Gauss quadrature theory Bounds for Riemann-Stieltjes integral The Gauss quadrature theory is formulated in terms of finite summations: b a f(λ)dα(λ) = N j=1 w j f(t j )+ M k=1 v k f(z k )+R[ f] Unknown weights: [w j ] N j=1,[v k] M k=1 ; Unknown nodes: [t j] N j=1. Prescribed nodes: [z k ] M k=1 The remainder term R[ f] is given by R[ f] = f (2N+M) (ξ) b (2N + M)! a M k=1 (λ z k ) [ N 2 (λ t j )] dα(λ), a < ξ < b. j=1 Golub and Meurant [GM93] showed that the sign of the remainder term R[ f] can be adjusted by the prescribed nodes. Setting M = 1, we will use the Gauss-Radau formula to get the bounds of (a part of) secular function. Matrices and Moments: Least Squares Problems (11/52)

12 Least Squares Problems Gauss Quadrature Rules Gauss quadrature theory Orthonormal polynomials Define a sequence of polynomials p 0 (λ), p 1 (λ),... that are orthonormal with respect to α(λ): b { 1 if i = j p i (λ)p j (λ)dα(λ) = 0 otherwise, a and p k (λ) is of exact degree k. Moreover, the roots of p k (λ) are distinct, real and lie in the interval [a, b]. Matrices and Moments: Least Squares Problems (12/52)

13 Least Squares Problems Gauss Quadrature Rules Tri-diagonalization for orthonormal polynomials Three-term recurrence relationship If dα = 1, the set of orthonormal polynomials satisfies: where λ p(λ) = T N p(λ)+γ N p N (λ)e N, p(λ) = [p 0 (λ) p 1 (λ) p N 1 (λ)] T, e N = (0 0 1) T R N, ω 1 γ 1 γ 1 ω 2 γ 2 T N = γ N 2 ω N 1 γ N 1 γ N 1 ω N Matrices and Moments: Least Squares Problems (13/52)

14 Least Squares Problems Gauss Quadrature Rules Tri-diagonalization for orthonormal polynomials Lanczos algorithm for quadratures To obtain the tri-diagonal matrix and hence the Gauss-Radau rule, we will use the Lanczos algorithm with p 1 = u/ u 2 as a starting vector: Eigenvalue decomposition of T N : Function of matrix: M PT N P T T N = QΛQ T u T f(m)u u T f(pt N P T )u = u 2 2 et 1 Q f(λ)qt e 1, where e 1 = (1 0 0) T R N. Thus, the eigenvalues of T N give us the nodes and the squares of the first elements of the eigenvectors give the weights: N w j f(t j ) = u 2 N 2 (Q 1i ) 2 f(λ i ) j=1 i=1 Matrices and Moments: Least Squares Problems (14/52)

15 Least Squares Problems Gauss Quadrature Rules Inverse eigenvalue problem for Gauss-Radau rule Inverse eigenvalue problem To obtain the Gauss-Radau rule, we extend the matrix T N in such a way that it has one prescribed eigenvalue z 1. Lemma The extended tri-diagonal matrix ( ) TN γ ˆT N+1 N e N γ N e T N ŵ N+1 has z 1 as an eigenvalue, where ŵ N+1 = z 1 + δ N, and δ N is the last entry of δ such that (T N z 1 I)δ = γ 2 N e N. (1) Matrices and Moments: Least Squares Problems (15/52)

16 Least Squares Problems Gauss Quadrature Rules Inverse eigenvalue problem for Gauss-Radau rule Proof of the extended tri-diagonal matrix lemma Proof. We can verify that z 1 is an eigenvalue of ˆT N+1 by investigating the following relation to get (1): ˆT N+1 d = z 1 d, where d is a corresponding eigenvector. Now, ˆT N+1 gives the weights and nodes of the Gauss-Radau rule such that N w j f(t j )+v 1 f(z 1 ) = u 2 e T 1 f( ˆT N+1 )e 1. j=1 The remainder is [ R[ f] = u 2 f (2N+1) (ξ) b N 2 (λ z 1 ) (λ t j )] dα(λ). (2N + 1)! a j=1 Matrices and Moments: Least Squares Problems (16/52)

17 Part II Application to solving secular equations Matrices and Moments: Least Squares Problems (17/52)

18 Common function of matrix Secular functions Recall the secular equations: T LS : ψ TLS (λ) = b T b b T A(A T A λ I) 1 A T b λ = 0, DLS : ψ DLS (λ) = b T b b T A(A T A λ I) 1 A T b = 0. ψ TLS/DLS(λ) is referred to as secular function. λ: an estimate of the minimum squared TLS/DLS distance. generic condition: λ < σ 2 min (A) for (AT A λ I) 1 Thus, in the domain of 0 λ < σmin 2 (A), we need to evaluate a matrix function of λ which is common in ψ TLS and ψ DLS : φ(λ) = b T A(A T A λ I) 1 A T b Matrices and Moments: Least Squares Problems (19/52)

19 Common function of matrix Bounds of a common function Now we evaluate the bounds of the scalar quantity φ: φ = g T f 1 x (M)g, M = A T A λ I, g = A T b, where f 1 (x) = 1 x x. Then, the quadrature rule ˆφ N+1 (z 1 ) = g 2 e T 1 f 1 ( ˆT N+1 )e 1 is described in terms of the remainder: x [ b N 2 ˆφ N+1 (z 1 ) = I[ f 1 ]+ g 2 (ξ) (2N+2) (λ z 1 ) (λ t j )] dα(λ). x a j=1 We note that f (2N+1) 1 x (ξ) (2N+1)! = (ξ) (2N+2) < 0, λ < a < ξ < b. Thus, we have the bounds: ˆφ N+1 (ζ b ) < I[ f 1 x ] < ˆφ N+1 (ζ a ), ζ a < a < b < ζ b. Matrices and Moments: Least Squares Problems (20/52)

20 Common function of matrix Comments on bounds f 1(x) is well defined on the proper interval (a, b) such that x the sign of the derivative function f (2N+1) 1 (ξ) is not changed x with the interval ξ (a, b). Since A 1 A > b, we may use ζ b = A 1 A. However, the lower bound of a (the smallest eigenvalue of M) is not easily obtainable. ζ a is determined very roughly. This explains why the upper bound of I[ f 1 x ] is usually poorer than the lower bound. Matrices and Moments: Least Squares Problems (21/52)

21 Common function of matrix Lanczos process with a shift for efficiency The tri-diagonalization is independent of the shift. Tri-diag([g, M]) Tri-diag([g, M+λ I]) (M+λ I)Q N = Q N (T N + λ I) MQ N = Q N T N Then we re-define the extended matrix ˆT N+1 as ( ) TN λ I J N+1 N γ N e N γ N e T, N w where T N and γ N are calculated by Tri-Diag of [g, A T A], (not [g, M]), and w is determined so that J N+1 has a prescribed eigenvalue z 1. Thus, w = z 1 + d N, where d N is the last entry of d such that (T N (z 1 + λ)i N )d = γ 2 N e N. Matrices and Moments: Least Squares Problems (22/52)

22 Common function of matrix Finally, we have the quadrature for bounds: ˆφ N+1 (z 1 ) = g 2 e T 1 f 1 x (J N+1 )e 1 = g 2 e T 1 (J N+1 ) 1 e 1. Once we solve a tri-diagonal J N+1 y = e 1 for y, we have ˆφ N+1 (z 1 ) = g 2 e T 1 y = g 2 y 1. For later interpolation, we need to evaluate the derivatives of the matrix function φ(λ) w.r.t. λ by approximating with f 1 (x) = x 2 and f x 2 1 (x) = x 3 : x 3 ˆφ = g 2 e T 1 f 1 (J N+1)e 1 = g 2 e T x 2 1 (J N+1 ) 2 e 1, ˆφ = 2 g 2 e T 1 f 1 (J N+1)e 1 = 2 g 2 e T x 3 1 (J N+1 ) 3 e 1. By solving J N+1 h = y., we have ˆφ = g 2 y 2, ˆφ = 2 g 2 y T (J N+1 ) 1 y = 2 g 2 y T h. A symmetric, tri-diagonal, and positive definite system requires O(N) flops [GV96] to be solved Ṁatrices and Moments: Least Squares Problems (23/52)

23 Common function of matrix Lemma (Monotonicity of bound sequences) Along with Lanczos processes, a sequence of bound estimates of φ = g T f 1 x (M)g with full-rank symmetric M R n n and f 1(x) = 1 x x is generated by Gauss quadrature rules. Then the estimated sequence ˆφ N+1 is necessarily monotonic. In other words, given each prescribed node ζ a or ζ b such that ζ a < σ min (M) < σ max (M) < ζ b, the lower and upper bound sequences for I[ f 1 x ] satisfy < ˆφ N (ζ b ) < ˆφ N+1 (ζ b ) < < I[ f 1 x ] < < ˆφ N+1 (ζ a ) < ˆφ N (ζ a ) <. Note that the complete Lanczos processes yield the exact evaluation: φ = g 2 e T 1 f 1 x (T n λ I n )e 1. Matrices and Moments: Least Squares Problems (24/52)

24 Newton methods Interpolating the root of secular equations We approximate the common function φ(λ) by using Lanczos processes combined with Gauss quadrature rule, where we proceed the processes until the upper- and lower- bounds of φ(λ) match within a tolerance. Suppose λ k is the current estimate of the minimum distance. In order to interpolate the root of the secular equation ψ(λ k ), we need to evaluate the followings: ψ TLS (λ k ) = b 2 λ k φ(λ k ), ψ DLS (λ k ) = b 2 φ(λ k ). ψ TLS (λ k) = 1 φ (λ k ), ψ DLS (λ k) = φ (λ k ). ψ TLS(λ k ) = φ (λ k ), ψ DLS(λ k ) = φ (λ k ). Then, consider one-point interpolating methods to obtain λ k+1 such that ψ(λ k+1 ) = 0. Note that roots of secular equation consist of the stationary points of the geometrically equivalent cost function. We want to find the smallest root λ k+1 [0, σmin 2 (A)) from the definition of TLS/DLS problem. However, we can not achieve it without using additional information on the locations of poles such as σ 2 min (A) min j i a i j 2 and σ 2 max(a) A 1 A. In the following sections, we will discuss how to use bisection and the upper-bound of the smallest pole. Matrices and Moments: Least Squares Problems (25/52)

25 Newton methods Variations of Newton s method take the form: λ k+1 = λ k ψ(λ k) ψ (λ k ) C k, where C k denotes a convergence factor [Gan78, Gan85] according to methods such as the Newton s method, the Halley s variation, and simple rational approximation in the following Table. Table: Variations of Newton s method Newton s SRA a Halley s Interpolating function h(λ) ψ(λ) h(λ) = c 0 + c 1 λ h(λ) = b 2 c 1 c 2 λ h(λ) = c 0 c 1 ( c 2 λ ) 1 ψ(λ k )ψ (λ k ) 2(ψ (λ k )) 2 b Convergence factor C k 1 2 ψ(λ k ) b 2 1/ Rate of (local) convergence Quadratic Quadratic Cubic Convergence region b Narrow Wide Wider Algebraic interpretation c b g(λ) = ψ(λ) g(λ) = 1 2 b 2 g(λ) = ψ(λ) ψ(λ) ψ (λ) a Simple Rational Approximation b Global convergence in root-finding of secular equations c Equivalently, solve g(λ) = 0 by Newton s method. Matrices and Moments: Least Squares Problems (26/52)

26 Seeking the smallest root Mixing with bisection Whenever we detect that the root estimate is larger than σmin 2 (A), we bisect the estimate to assure that it is less than σmin 2 (A). The monotonicity of the sequence of estimate bounds of Gauss quadratures (GQ) is utilized based on the following scenario: 1. With the initial guess of root, we obtain the sequence of bounds of secular function by means of GQ. 2. If the sequences are not monotonic, we conclude that the root estimate is larger than the squared smallest singular value of A. Then the root estimate is cut by half, and go to Step 1 with the modified estimate of root. 3. Otherwise, we interpolate the root of secular equation by using the estimate of the secular function and its derivatives. 4. If the new root estimate is close to the previous one within a tolerance, then we calculate the de-regularized solution, and stop the algorithm. Otherwise, go to step 1. Although the violation of monotonicity is only a necessary condition, our numerical simulation works well. Matrices and Moments: Least Squares Problems (27/52)

27 Seeking the smallest root Stabilizing with the estimated smallest pole Although the bisection scheme almost always achieves the smallest root, it may suffer from a bi-stability problem which means the estimates are alternating between two values. To get around this, we employ the estimation of the smallest pole by modifying the previous scenario. If we detect the current estimate of root is larger than σmin 2 (A), we cut the estimate by half and set the upper-bound of the smallest pole to the current estimate of root as well. ˆσ 2 min (A) = λ k, λ k+1 = 1 2 λ k. Then, when we interpolate the next estimate of root, we take a harmonic sum between the Newton-based step δ k = ψ(λ k) ψ (λ k ) C k and the distance from the upper-bound of the smallest pole estimation. 1 λ k+1 = λ k + = λ k δ k + δk 2 /(δ k + λ k ˆσ min 2 (A)) δ k ˆσ min 2 (A) λ k Matrices and Moments: Least Squares Problems (28/52)

28 Solving TLS and DLS by quadrature and CG method CGQ as a secular equation approach 1. Find the smallest root of secular equation for TLS or DLS. 1.1 Evaluate the bounds of secular function by Gauss-Radau quadrature rule. 1.2 Interpolate the zero of the function by a variation of Newton method. 1.3 Determine a proper interval for the smallest zero by bisection and harmonic-summation with the upper-bound of the smallest pole. 2. Solve a de-regularized system with a shift of the smallest root. Solve the symmetric, positive-definite system by the conjugate gradient (CG) method, Or, solve the tri-diagonal system with shift. Matrices and Moments: Least Squares Problems (29/52)

29 Alternative implementations Alternative implementations Reuse of Lanczos vectors with sufficient memory Regeneration of Lanczos vectors with knowledge of tri-diagonal entries Avoiding of explicit multiplication of A T A Shifting into Lanczos bi-diagonalization Using backward perturbations Matrices and Moments: Least Squares Problems (30/52)

30 Backward perturbations for linear least squares I min x b Ax 2, A : m n, b : m 1. ξ : arbitrary vector, calculated. ˆx = ξ + e; ˆx = A + b. ξ = (A+δA) + b. ρ = b Aξ. µ(ξ) = min δa δa F. µ(ξ) = min{ ρ 2 ξ 2,σ min (A,B)} (Karlson, Waldén & Sun) B = ρ 2 ξ 2 (I ρ ρt ρ 2 ).

31 Backward perturbations for linear least squares II µ 2 (ξ) = ρ T A(αA T A+β I) 1 A T ρ α = ξ 2 2, β = ρ 2 2 µ(ξ) µ(ξ) lim ξ ˆx µ(ξ) µ( ˆx) = 1 (Grcar)

32 Large-scale TLS algorithms Björck s algorithm Solve the system of nonlinear equations [ A T A A T ][ ] [ ] b x x b T A b T = λ, b 1 1 or equivalently, the system [ f(x,λ) g(x,λ) ] = [ A T ] r λx b T = r+λ [ ] 0 0 with r = b Ax using a Rayleigh-quotient iteration (RQI). This algorithm will always converge to a singular value/vector pair, but we might not get λ = σn+1 2. Björck suggested one initial inverse iteration (i.e. λ = 0) to move closer to the desired λ, and then apply the RQI procedure. Matrices and Moments: Least Squares Problems (34/52)

33 Large-scale TLS algorithms Björck s algorithm details After some manipulations, Björck s algorithm greatly simplifies. The following presentation emphasizes the computationally intensive steps. 1. Solve the least squares system in A,b for x LS. 2. Perform one inverse iteration (solve A T Ax (0) = x LS ) to get the initial x (0). 3. While not converged... solve two systems in (A T A λi)x = b to iterate to x (k+1), but if we detect A T A λi is negative definite, decrease λ and repeat the iteration. Björck suggests using the PCGTLS algorithm to solve each linear system with the Cholesky factor R or A T A as a preconditioner. Inside the CG procedure, we detect A T A λi is negative definite and use the CG vectors to compute a new value of λ. Matrices and Moments: Least Squares Problems (35/52)

34 Large-scale TLS algorithms Notes on our implementation of Björck s algorithm Instead of using PCGTLS with the Cholesky factor of A T A, we use a matrix-free approach for large scale m > 800 and apply the unpreconditioned CGTLS algorithm. The tolerance used in the CG method is To compute the initial least squares solution, we use the LSQR algorithm. We detect convergence when λ changes by less than or the normalized residual increases (theory states the normalized residual always decreases). After we detect convergence, we run one more iteration of the algorithm to ensure that we compute an x for the λ. Matrices and Moments: Least Squares Problems (36/52)

35 Large-scale TLS algorithms Details of the matrix moments based algorithm Algorithm 2 uses the Golub-Kahan bidiagonalization of A and applies the moment algorithm to T = B T B instead of computing T directly from the Lanczos process on A T A. Algorithm 1 restarts the Lanczos process at each iteration. Algorithm 2 never restarts the bidiagonalization process and simply continues the process at each iteration. Matrices and Moments: Least Squares Problems (37/52)

36 Comparison of large-scale TLS results Problems Jo s problems, 15 8 and Björck s problem 1: matrix Large scale problems with and matrices. The large scale problems were generated using random Householder matrices to build the SVD of [ A b ] in product form. Each large-scale matrix was available solely as an operator to all of the algorithms. The singular values of [ ] A b are σ i = log(i)+ N(0,1), where N(0, 1) is a standard normal random variable. Matrices and Moments: Least Squares Problems (38/52)

37 Comparison of large-scale TLS results Parameter choices Algorithm 1 Algorithm 2 λ (0) = 0 λ (0) = 1 λ (0) = ρ λ (0) = 0 λ (0) = 1 λ (0) = ρ newton sra halley newton *8 *8 *7 sra *12 *24 *7 halley *14 *23 *6 newton *20 *7 *10 sra *20 25 *64 halley *55 55 *12 newton *15 *11 *11 sra ++ *15 *25 *14 halley *20 *57 *11 newton sra halley * wrong root; ++ correct w/o convergence; no convergence Matrices and Moments: Least Squares Problems (39/52)

38 Comparison of large-scale TLS results Convergence TEST ALG ITERS ERROR TIME LANZ. jo björck (15,8) Alg σ 2 = Alg jo björck (750,400) Alg 1 > σ 2 = Alg large-scale björck (10000,5000) Alg 1 > σ 2 = Alg large-scale björck (100000,60000) Alg 1 > σ 2 = Alg björck björck (30,15) Alg 1 >100 σ σ 2 = Alg Matrices and Moments: Least Squares Problems (40/52)

39 Numerical data generation The numerical data are generated as follows [GvM91]. Let the error-free data matrix A s = U s Σ s V T s, where U s = I m 2 us ut s u T and V s = I n 2 vs vt s s us v T are Householder matrices and s vs Σ s = diag(σ k ) is a m-by-n diagonal matrix with [σ 1,,σ n ] = [ n,,1]. The vectors u s and v s consist of pseudo-random numbers generated by a function randn providing uniformly distributed random numbers in Matlab TM. An error-free observation vector b s is computed by b s = A s x s. The error-prone data A and b are generated with a given deviation σ n as follows: randn( state,108881);% seed for random number u s = randn(m, 1); v s = randn(n, 1); A s = U s Σ s V T s ; x s = (1./(1 : n)) ;% [1, 1/2,, 1/n] T b s = A s x s ; σ n = 0.3; b = b s + σ n randn(m, 1); A = A s + σ n randn(m, n); Matrices and Moments: Least Squares Problems (42/52)

40 CG-based algorithms CG-based algorithms We can categorize algorithms in the following example as below. Table: CG-based algorithms for solving secular equation. CGQ-BS-KP CGQ-BS CG-BS-KP CG-BS Gauss Quadrature rules Used Used Not Not Bisection Used Used Used Used Upper-bound of the smallest pole Used Not Used Not The previous CG-based approaches in [GJK06] or [BHM00] employ matrix-vector multiplications to evaluate the secular function and derivative from the intermediate solution x(λ) such that (A T A λ I)x(λ) = A T b. Then the secular functions and derivatives are obtained as follows. ψ TLS (λ) = b 2 b T Ax(λ) λ ψ TLS (λ) = x(λ) 2 1 ψ DLS (λ) = b 2 b T Ax(λ) ψ DLS (λ) = x(λ) 2 Matrices and Moments: Least Squares Problems (43/52)

41 Small problem Secular function plots with A R TLS secular function, ψ TLS (λ) 20 DLS secular function, ψ DLS (λ) Squared distances, λ (a) TLS secular function plot Squared distances, λ (b) DLS secular function plot Matrices and Moments: Least Squares Problems (44/52)

42 Small problem Convergent curves in TLS with A R 15 8 Squared distance TLSCGQ Newton KP (6), x = TLSCGQ SRA KP (6), x = TLSCGQ Halley KP (6), x = Iterations (a) Variations of CGQ with the bisection and the smallest pole estimation. Squared distance Squared distance TLSCGQ Newton (6), x = TLSCGQ SRA (5), x = TLSCGQ Halley (5), x = Iterations (b) Variations of CGQ with the bisection only. Squared distance TLSCG Newton KP (5), x = TLSCG HS KP (5), x = Iterations (c) CG with two different interpolations, the bisection and the smallest pole estimation. 0.1 TLSCG RQI (4), x = TLSCG HSI (5), x = Iterations (d) CG with two different inverse iterations, and the bisection only. Matrices and Moments: Least Squares Problems (45/52)

43 Small problem Convergent curves in DLS with A R 15 8 Squared distance DLSCGQ Newton KP (6), x = DLSCGQ SRA KP (6), x = DLSCGQ Halley KP (6), x = Iterations (a) Variations of CGQ with the bisection and the smallest pole estimation. 1 Squared distance DLSCGQ Newton (7), x = DLSCGQ SRA (6), x = DLSCGQ Halley (5), x = Iterations (b) Variations of CGQ with the bisection only Squared distance Squared distance DLSCG Newton KP (5), x = DLSCG HS KP (5), x = Iterations (c) CG with two different interpolations, the bisection and the smallest pole estimation. 0.2 DLSCG RQI (4), x = DLSCG HSI (5), x = Iterations (d) CG with two different inverse iterations, and the bisection only. Matrices and Moments: Least Squares Problems (46/52)

44 Large problem Convergent curves in TLS with A R Squared distance Squared distance TLSCGQ Newton KP (15), x = TLSCGQ SRA KP (9), x = TLSCGQ Halley KP (6), x = Iterations (a) Variations of CGQ with the bisection and the smallest pole estimation. 20 TLSCGQ Newton (20), x = TLSCGQ SRA (20), x = TLSCGQ Halley (20), x = Iterations (b) Variations of CGQ with the bisection only Squared distance 10 5 TLSCG Newton KP (12), x = TLSCG HS KP (12), x = Iterations (c) CG with two different interpolations, the bisection and the smallest pole estimation. Squared distance TLSCG RQI (20), x = TLSCG HSI (20), x = Iterations (d) CG with two different inverse iterations, and the bisection only. Matrices and Moments: Least Squares Problems (47/52)

45 Large problem Convergent curves in DLS with A R Squared distance Squared distance DLSCGQ Newton KP (20), x = DLSCGQ SRA KP (20), x = DLSCGQ Halley KP (13), x = Iterations (a) Variations of CGQ with the bisection and the smallest pole estimation DLSCGQ Newton (20), x = DLSCGQ SRA (20), x = DLSCGQ Halley (20), x = Iterations (b) Variations of CGQ with the bisection only Squared distance Squared distance DLSCG Newton KP (9), x = DLSCG HS KP (9), x = Iterations (c) CG with two different interpolations and the smallest pole estimation. 5 DLSCG RQI (20), x = DLSCG HSI (20), x = Iterations (d) CG with two different inverse iterations, and the bisection only. Matrices and Moments: Least Squares Problems (48/52)

46 Conclusion Conclusion The presented Conjugate Gradient (CG) with Quadrature (CGQ) approximates secular functions by means of Gauss Quadrature (GQ) rules. The previous CG based approaches are exhaustively calculating intermediate solutions to evaluate secular functions. Interpolating the smallest root by variations of Newton method with stabilized modification: Bisection to assure of the convergence to the smallest root. Approximating with rational function of the estimated smallest pole to get around bi-stability problem. CGQ does not pursue the intermediate solution until the root estimate converges. The overall computational complexity can be reduced by adjusting accuracy of GQ. Matrices and Moments: Least Squares Problems (50/52)

47 References Åke Björck, P. Heggernes, and P. Matstoms. Methods for large scale total least squares problems. SIAM J. Matrix Anal. Appl., 22(3): , R. D. DeGroat and E. M. Dowling. The data least squares problem and channel equalization. 41(1): , January Walter Gander. On the linear least squares problem with a quadratic constraint. Technical Report STAN-CS , Stanford University, W. Gander. On halley s iteration method. The American mathematical monthly, 92(2): , Gene H. Golub, SungEun Jo, and Sang Woo Kim. Solving secular equations for total/data least squares problems. SIAM J. Matrix Anal. Appl., submitted for publication, G. H. Golub and G. Meurant. Matrices, moments and quadrature. Report SCCM-93-07, Computer Science Department, Stanford University, Matrices and Moments: Least Squares Problems (51/52)

48 References Gene H. Golub and Charles. F. Van Loan. Matrix Computations. Johns Hopkins Univ. Pr., Baltimore, MD, third edition, November G. H. Golub and U. von Matt. Quadratically constrained least squares and quadratic problems. Numerische Mathematik, 59(1): , December SungEun Jo and Sang Woo Kim. Consistent normalized least mean square filtering with noisy data matrix. (6): , June Matrices and Moments: Least Squares Problems (52/52)

Total least squares. Gérard MEURANT. October, 2008

Total least squares. Gérard MEURANT. October, 2008 Total least squares Gérard MEURANT October, 2008 1 Introduction to total least squares 2 Approximation of the TLS secular equation 3 Numerical experiments Introduction to total least squares In least squares

More information

Introduction. Chapter One

Introduction. Chapter One Chapter One Introduction The aim of this book is to describe and explain the beautiful mathematical relationships between matrices, moments, orthogonal polynomials, quadrature rules and the Lanczos and

More information

The Lanczos and conjugate gradient algorithms

The Lanczos and conjugate gradient algorithms The Lanczos and conjugate gradient algorithms Gérard MEURANT October, 2008 1 The Lanczos algorithm 2 The Lanczos algorithm in finite precision 3 The nonsymmetric Lanczos algorithm 4 The Golub Kahan bidiagonalization

More information

Key words. conjugate gradients, normwise backward error, incremental norm estimation.

Key words. conjugate gradients, normwise backward error, incremental norm estimation. Proceedings of ALGORITMY 2016 pp. 323 332 ON ERROR ESTIMATION IN THE CONJUGATE GRADIENT METHOD: NORMWISE BACKWARD ERROR PETR TICHÝ Abstract. Using an idea of Duff and Vömel [BIT, 42 (2002), pp. 300 322

More information

Block Bidiagonal Decomposition and Least Squares Problems

Block Bidiagonal Decomposition and Least Squares Problems Block Bidiagonal Decomposition and Least Squares Problems Åke Björck Department of Mathematics Linköping University Perspectives in Numerical Analysis, Helsinki, May 27 29, 2008 Outline Bidiagonal Decomposition

More information

Numerical Methods in Matrix Computations

Numerical Methods in Matrix Computations Ake Bjorck Numerical Methods in Matrix Computations Springer Contents 1 Direct Methods for Linear Systems 1 1.1 Elements of Matrix Theory 1 1.1.1 Matrix Algebra 2 1.1.2 Vector Spaces 6 1.1.3 Submatrices

More information

Matrices, Moments and Quadrature, cont d

Matrices, Moments and Quadrature, cont d Jim Lambers CME 335 Spring Quarter 2010-11 Lecture 4 Notes Matrices, Moments and Quadrature, cont d Estimation of the Regularization Parameter Consider the least squares problem of finding x such that

More information

ESTIMATES OF THE TRACE OF THE INVERSE OF A SYMMETRIC MATRIX USING THE MODIFIED CHEBYSHEV ALGORITHM

ESTIMATES OF THE TRACE OF THE INVERSE OF A SYMMETRIC MATRIX USING THE MODIFIED CHEBYSHEV ALGORITHM ESTIMATES OF THE TRACE OF THE INVERSE OF A SYMMETRIC MATRIX USING THE MODIFIED CHEBYSHEV ALGORITHM GÉRARD MEURANT In memory of Gene H. Golub Abstract. In this paper we study how to compute an estimate

More information

Non-stationary extremal eigenvalue approximations in iterative solutions of linear systems and estimators for relative error

Non-stationary extremal eigenvalue approximations in iterative solutions of linear systems and estimators for relative error on-stationary extremal eigenvalue approximations in iterative solutions of linear systems and estimators for relative error Divya Anand Subba and Murugesan Venatapathi* Supercomputer Education and Research

More information

Moments, Model Reduction and Nonlinearity in Solving Linear Algebraic Problems

Moments, Model Reduction and Nonlinearity in Solving Linear Algebraic Problems Moments, Model Reduction and Nonlinearity in Solving Linear Algebraic Problems Zdeněk Strakoš Charles University, Prague http://www.karlin.mff.cuni.cz/ strakos 16th ILAS Meeting, Pisa, June 2010. Thanks

More information

Solving discrete ill posed problems with Tikhonov regularization and generalized cross validation

Solving discrete ill posed problems with Tikhonov regularization and generalized cross validation Solving discrete ill posed problems with Tikhonov regularization and generalized cross validation Gérard MEURANT November 2010 1 Introduction to ill posed problems 2 Examples of ill-posed problems 3 Tikhonov

More information

Golub-Kahan iterative bidiagonalization and determining the noise level in the data

Golub-Kahan iterative bidiagonalization and determining the noise level in the data Golub-Kahan iterative bidiagonalization and determining the noise level in the data Iveta Hnětynková,, Martin Plešinger,, Zdeněk Strakoš, * Charles University, Prague ** Academy of Sciences of the Czech

More information

AN ITERATIVE METHOD WITH ERROR ESTIMATORS

AN ITERATIVE METHOD WITH ERROR ESTIMATORS AN ITERATIVE METHOD WITH ERROR ESTIMATORS D. CALVETTI, S. MORIGI, L. REICHEL, AND F. SGALLARI Abstract. Iterative methods for the solution of linear systems of equations produce a sequence of approximate

More information

Sensitivity of Gauss-Christoffel quadrature and sensitivity of Jacobi matrices to small changes of spectral data

Sensitivity of Gauss-Christoffel quadrature and sensitivity of Jacobi matrices to small changes of spectral data Sensitivity of Gauss-Christoffel quadrature and sensitivity of Jacobi matrices to small changes of spectral data Zdeněk Strakoš Academy of Sciences and Charles University, Prague http://www.cs.cas.cz/

More information

Chapter 12 Solving secular equations

Chapter 12 Solving secular equations Chapter 12 Solving secular equations Gérard MEURANT January-February, 2012 1 Examples of Secular Equations 2 Secular equation solvers 3 Numerical experiments Examples of secular equations What are secular

More information

Computational Methods. Eigenvalues and Singular Values

Computational Methods. Eigenvalues and Singular Values Computational Methods Eigenvalues and Singular Values Manfred Huber 2010 1 Eigenvalues and Singular Values Eigenvalues and singular values describe important aspects of transformations and of data relations

More information

Numerical Methods for Solving Large Scale Eigenvalue Problems

Numerical Methods for Solving Large Scale Eigenvalue Problems Peter Arbenz Computer Science Department, ETH Zürich E-mail: arbenz@inf.ethz.ch arge scale eigenvalue problems, Lecture 2, February 28, 2018 1/46 Numerical Methods for Solving Large Scale Eigenvalue Problems

More information

arxiv: v1 [math.na] 5 May 2011

arxiv: v1 [math.na] 5 May 2011 ITERATIVE METHODS FOR COMPUTING EIGENVALUES AND EIGENVECTORS MAYSUM PANJU arxiv:1105.1185v1 [math.na] 5 May 2011 Abstract. We examine some numerical iterative methods for computing the eigenvalues and

More information

Solving Constrained Rayleigh Quotient Optimization Problem by Projected QEP Method

Solving Constrained Rayleigh Quotient Optimization Problem by Projected QEP Method Solving Constrained Rayleigh Quotient Optimization Problem by Projected QEP Method Yunshen Zhou Advisor: Prof. Zhaojun Bai University of California, Davis yszhou@math.ucdavis.edu June 15, 2017 Yunshen

More information

Efficient and Accurate Rectangular Window Subspace Tracking

Efficient and Accurate Rectangular Window Subspace Tracking Efficient and Accurate Rectangular Window Subspace Tracking Timothy M. Toolan and Donald W. Tufts Dept. of Electrical Engineering, University of Rhode Island, Kingston, RI 88 USA toolan@ele.uri.edu, tufts@ele.uri.edu

More information

A Note on Inverse Iteration

A Note on Inverse Iteration A Note on Inverse Iteration Klaus Neymeyr Universität Rostock, Fachbereich Mathematik, Universitätsplatz 1, 18051 Rostock, Germany; SUMMARY Inverse iteration, if applied to a symmetric positive definite

More information

Total Least Squares Approach in Regression Methods

Total Least Squares Approach in Regression Methods WDS'08 Proceedings of Contributed Papers, Part I, 88 93, 2008. ISBN 978-80-7378-065-4 MATFYZPRESS Total Least Squares Approach in Regression Methods M. Pešta Charles University, Faculty of Mathematics

More information

EIGENVALUE PROBLEMS. EIGENVALUE PROBLEMS p. 1/4

EIGENVALUE PROBLEMS. EIGENVALUE PROBLEMS p. 1/4 EIGENVALUE PROBLEMS EIGENVALUE PROBLEMS p. 1/4 EIGENVALUE PROBLEMS p. 2/4 Eigenvalues and eigenvectors Let A C n n. Suppose Ax = λx, x 0, then x is a (right) eigenvector of A, corresponding to the eigenvalue

More information

Baltzer Journals April 24, Bounds for the Trace of the Inverse and the Determinant. of Symmetric Positive Denite Matrices

Baltzer Journals April 24, Bounds for the Trace of the Inverse and the Determinant. of Symmetric Positive Denite Matrices Baltzer Journals April 4, 996 Bounds for the Trace of the Inverse and the Determinant of Symmetric Positive Denite Matrices haojun Bai ; and Gene H. Golub ; y Department of Mathematics, University of Kentucky,

More information

Numerical Analysis Preliminary Exam 10 am to 1 pm, August 20, 2018

Numerical Analysis Preliminary Exam 10 am to 1 pm, August 20, 2018 Numerical Analysis Preliminary Exam 1 am to 1 pm, August 2, 218 Instructions. You have three hours to complete this exam. Submit solutions to four (and no more) of the following six problems. Please start

More information

ETNA Kent State University

ETNA Kent State University Electronic Transactions on Numerical Analysis. Volume 14, pp. 2-35, 22. Copyright 22,. ISSN 168-9613. ETNA L-CURVE CURVATURE BOUNDS VIA LANCZOS BIDIAGONALIZATION D. CALVETTI, P. C. HANSEN, AND L. REICHEL

More information

M.A. Botchev. September 5, 2014

M.A. Botchev. September 5, 2014 Rome-Moscow school of Matrix Methods and Applied Linear Algebra 2014 A short introduction to Krylov subspaces for linear systems, matrix functions and inexact Newton methods. Plan and exercises. M.A. Botchev

More information

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for 1 Algorithms Notes for 2016-10-31 There are several flavors of symmetric eigenvalue solvers for which there is no equivalent (stable) nonsymmetric solver. We discuss four algorithmic ideas: the workhorse

More information

On Lagrange multipliers of trust-region subproblems

On Lagrange multipliers of trust-region subproblems On Lagrange multipliers of trust-region subproblems Ladislav Lukšan, Ctirad Matonoha, Jan Vlček Institute of Computer Science AS CR, Prague Programy a algoritmy numerické matematiky 14 1.- 6. června 2008

More information

Notes on Eigenvalues, Singular Values and QR

Notes on Eigenvalues, Singular Values and QR Notes on Eigenvalues, Singular Values and QR Michael Overton, Numerical Computing, Spring 2017 March 30, 2017 1 Eigenvalues Everyone who has studied linear algebra knows the definition: given a square

More information

c 1999 Society for Industrial and Applied Mathematics

c 1999 Society for Industrial and Applied Mathematics SIAM J. MATRIX ANAL. APPL. Vol. 21, No. 1, pp. 185 194 c 1999 Society for Industrial and Applied Mathematics TIKHONOV REGULARIZATION AND TOTAL LEAST SQUARES GENE H. GOLUB, PER CHRISTIAN HANSEN, AND DIANNE

More information

FEM and sparse linear system solving

FEM and sparse linear system solving FEM & sparse linear system solving, Lecture 9, Nov 19, 2017 1/36 Lecture 9, Nov 17, 2017: Krylov space methods http://people.inf.ethz.ch/arbenz/fem17 Peter Arbenz Computer Science Department, ETH Zürich

More information

Chap 3. Linear Algebra

Chap 3. Linear Algebra Chap 3. Linear Algebra Outlines 1. Introduction 2. Basis, Representation, and Orthonormalization 3. Linear Algebraic Equations 4. Similarity Transformation 5. Diagonal Form and Jordan Form 6. Functions

More information

COMP 558 lecture 18 Nov. 15, 2010

COMP 558 lecture 18 Nov. 15, 2010 Least squares We have seen several least squares problems thus far, and we will see more in the upcoming lectures. For this reason it is good to have a more general picture of these problems and how to

More information

Comparison of A-Posteriori Parameter Choice Rules for Linear Discrete Ill-Posed Problems

Comparison of A-Posteriori Parameter Choice Rules for Linear Discrete Ill-Posed Problems Comparison of A-Posteriori Parameter Choice Rules for Linear Discrete Ill-Posed Problems Alessandro Buccini a, Yonggi Park a, Lothar Reichel a a Department of Mathematical Sciences, Kent State University,

More information

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.

More information

Numerical Methods - Numerical Linear Algebra

Numerical Methods - Numerical Linear Algebra Numerical Methods - Numerical Linear Algebra Y. K. Goh Universiti Tunku Abdul Rahman 2013 Y. K. Goh (UTAR) Numerical Methods - Numerical Linear Algebra I 2013 1 / 62 Outline 1 Motivation 2 Solving Linear

More information

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection Eigenvalue Problems Last Time Social Network Graphs Betweenness Girvan-Newman Algorithm Graph Laplacian Spectral Bisection λ 2, w 2 Today Small deviation into eigenvalue problems Formulation Standard eigenvalue

More information

Lanczos tridigonalization and Golub - Kahan bidiagonalization: Ideas, connections and impact

Lanczos tridigonalization and Golub - Kahan bidiagonalization: Ideas, connections and impact Lanczos tridigonalization and Golub - Kahan bidiagonalization: Ideas, connections and impact Zdeněk Strakoš Academy of Sciences and Charles University, Prague http://www.cs.cas.cz/ strakos Hong Kong, February

More information

arxiv: v1 [math.na] 1 Sep 2018

arxiv: v1 [math.na] 1 Sep 2018 On the perturbation of an L -orthogonal projection Xuefeng Xu arxiv:18090000v1 [mathna] 1 Sep 018 September 5 018 Abstract The L -orthogonal projection is an important mathematical tool in scientific computing

More information

5 Selected Topics in Numerical Linear Algebra

5 Selected Topics in Numerical Linear Algebra 5 Selected Topics in Numerical Linear Algebra In this chapter we will be concerned mostly with orthogonal factorizations of rectangular m n matrices A The section numbers in the notes do not align with

More information

Eigenvalue Problems. Eigenvalue problems occur in many areas of science and engineering, such as structural analysis

Eigenvalue Problems. Eigenvalue problems occur in many areas of science and engineering, such as structural analysis Eigenvalue Problems Eigenvalue problems occur in many areas of science and engineering, such as structural analysis Eigenvalues also important in analyzing numerical methods Theory and algorithms apply

More information

DELFT UNIVERSITY OF TECHNOLOGY

DELFT UNIVERSITY OF TECHNOLOGY DELFT UNIVERSITY OF TECHNOLOGY REPORT -09 Computational and Sensitivity Aspects of Eigenvalue-Based Methods for the Large-Scale Trust-Region Subproblem Marielba Rojas, Bjørn H. Fotland, and Trond Steihaug

More information

Scientific Computing: An Introductory Survey

Scientific Computing: An Introductory Survey Scientific Computing: An Introductory Survey Chapter 4 Eigenvalue Problems Prof. Michael T. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c 2002. Reproduction

More information

Greedy Tikhonov regularization for large linear ill-posed problems

Greedy Tikhonov regularization for large linear ill-posed problems International Journal of Computer Mathematics Vol. 00, No. 00, Month 200x, 1 20 Greedy Tikhonov regularization for large linear ill-posed problems L. Reichel, H. Sadok, and A. Shyshkov (Received 00 Month

More information

that determines x up to a complex scalar of modulus 1, in the real case ±1. Another condition to normalize x is by requesting that

that determines x up to a complex scalar of modulus 1, in the real case ±1. Another condition to normalize x is by requesting that Chapter 3 Newton methods 3. Linear and nonlinear eigenvalue problems When solving linear eigenvalue problems we want to find values λ C such that λi A is singular. Here A F n n is a given real or complex

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra)

AMS526: Numerical Analysis I (Numerical Linear Algebra) AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 1: Course Overview & Matrix-Vector Multiplication Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 20 Outline 1 Course

More information

On the Computations of Eigenvalues of the Fourth-order Sturm Liouville Problems

On the Computations of Eigenvalues of the Fourth-order Sturm Liouville Problems Int. J. Open Problems Compt. Math., Vol. 4, No. 3, September 2011 ISSN 1998-6262; Copyright c ICSRS Publication, 2011 www.i-csrs.org On the Computations of Eigenvalues of the Fourth-order Sturm Liouville

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 18 Outline

More information

UNIFYING LEAST SQUARES, TOTAL LEAST SQUARES AND DATA LEAST SQUARES

UNIFYING LEAST SQUARES, TOTAL LEAST SQUARES AND DATA LEAST SQUARES UNIFYING LEAST SQUARES, TOTAL LEAST SQUARES AND DATA LEAST SQUARES Christopher C. Paige School of Computer Science, McGill University, Montreal, Quebec, Canada, H3A 2A7 paige@cs.mcgill.ca Zdeněk Strakoš

More information

Lecture 4 Eigenvalue problems

Lecture 4 Eigenvalue problems Lecture 4 Eigenvalue problems Weinan E 1,2 and Tiejun Li 2 1 Department of Mathematics, Princeton University, weinan@princeton.edu 2 School of Mathematical Sciences, Peking University, tieli@pku.edu.cn

More information

CG and MINRES: An empirical comparison

CG and MINRES: An empirical comparison School of Engineering CG and MINRES: An empirical comparison Prequel to LSQR and LSMR: Two least-squares solvers David Fong and Michael Saunders Institute for Computational and Mathematical Engineering

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra)

AMS526: Numerical Analysis I (Numerical Linear Algebra) AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 16: Reduction to Hessenberg and Tridiagonal Forms; Rayleigh Quotient Iteration Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical

More information

SVD and its Application to Generalized Eigenvalue Problems. Thomas Melzer

SVD and its Application to Generalized Eigenvalue Problems. Thomas Melzer SVD and its Application to Generalized Eigenvalue Problems Thomas Melzer June 8, 2004 Contents 0.1 Singular Value Decomposition.................. 2 0.1.1 Range and Nullspace................... 3 0.1.2

More information

CS137 Introduction to Scientific Computing Winter Quarter 2004 Solutions to Homework #3

CS137 Introduction to Scientific Computing Winter Quarter 2004 Solutions to Homework #3 CS137 Introduction to Scientific Computing Winter Quarter 2004 Solutions to Homework #3 Felix Kwok February 27, 2004 Written Problems 1. (Heath E3.10) Let B be an n n matrix, and assume that B is both

More information

Solving Regularized Total Least Squares Problems

Solving Regularized Total Least Squares Problems Solving Regularized Total Least Squares Problems Heinrich Voss voss@tu-harburg.de Hamburg University of Technology Institute of Numerical Simulation Joint work with Jörg Lampe TUHH Heinrich Voss Total

More information

Numerical Methods I Eigenvalue Problems

Numerical Methods I Eigenvalue Problems Numerical Methods I Eigenvalue Problems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 October 2nd, 2014 A. Donev (Courant Institute) Lecture

More information

5.3 The Power Method Approximation of the Eigenvalue of Largest Module

5.3 The Power Method Approximation of the Eigenvalue of Largest Module 192 5 Approximation of Eigenvalues and Eigenvectors 5.3 The Power Method The power method is very good at approximating the extremal eigenvalues of the matrix, that is, the eigenvalues having largest and

More information

Second Order Optimization Algorithms I

Second Order Optimization Algorithms I Second Order Optimization Algorithms I Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A. http://www.stanford.edu/ yyye Chapters 7, 8, 9 and 10 1 The

More information

Lecture 9: Krylov Subspace Methods. 2 Derivation of the Conjugate Gradient Algorithm

Lecture 9: Krylov Subspace Methods. 2 Derivation of the Conjugate Gradient Algorithm CS 622 Data-Sparse Matrix Computations September 19, 217 Lecture 9: Krylov Subspace Methods Lecturer: Anil Damle Scribes: David Eriksson, Marc Aurele Gilles, Ariah Klages-Mundt, Sophia Novitzky 1 Introduction

More information

3D Computer Vision - WT 2004

3D Computer Vision - WT 2004 3D Computer Vision - WT 2004 Singular Value Decomposition Darko Zikic CAMP - Chair for Computer Aided Medical Procedures November 4, 2004 1 2 3 4 5 Properties For any given matrix A R m n there exists

More information

Krylov Subspace Methods for the Evaluation of Matrix Functions. Applications and Algorithms

Krylov Subspace Methods for the Evaluation of Matrix Functions. Applications and Algorithms Krylov Subspace Methods for the Evaluation of Matrix Functions. Applications and Algorithms 4. Monotonicity of the Lanczos Method Michael Eiermann Institut für Numerische Mathematik und Optimierung Technische

More information

AN INVERSE EIGENVALUE PROBLEM AND AN ASSOCIATED APPROXIMATION PROBLEM FOR GENERALIZED K-CENTROHERMITIAN MATRICES

AN INVERSE EIGENVALUE PROBLEM AND AN ASSOCIATED APPROXIMATION PROBLEM FOR GENERALIZED K-CENTROHERMITIAN MATRICES AN INVERSE EIGENVALUE PROBLEM AND AN ASSOCIATED APPROXIMATION PROBLEM FOR GENERALIZED K-CENTROHERMITIAN MATRICES ZHONGYUN LIU AND HEIKE FAßBENDER Abstract: A partially described inverse eigenvalue problem

More information

QR-decomposition. The QR-decomposition of an n k matrix A, k n, is an n n unitary matrix Q and an n k upper triangular matrix R for which A = QR

QR-decomposition. The QR-decomposition of an n k matrix A, k n, is an n n unitary matrix Q and an n k upper triangular matrix R for which A = QR QR-decomposition The QR-decomposition of an n k matrix A, k n, is an n n unitary matrix Q and an n k upper triangular matrix R for which In Matlab A = QR [Q,R]=qr(A); Note. The QR-decomposition is unique

More information

Full-State Feedback Design for a Multi-Input System

Full-State Feedback Design for a Multi-Input System Full-State Feedback Design for a Multi-Input System A. Introduction The open-loop system is described by the following state space model. x(t) = Ax(t)+Bu(t), y(t) =Cx(t)+Du(t) () 4 8.5 A =, B =.5.5, C

More information

SOME PROPERTIES OF SYMPLECTIC RUNGE-KUTTA METHODS

SOME PROPERTIES OF SYMPLECTIC RUNGE-KUTTA METHODS SOME PROPERTIES OF SYMPLECTIC RUNGE-KUTTA METHODS ERNST HAIRER AND PIERRE LEONE Abstract. We prove that to every rational function R(z) satisfying R( z)r(z) = 1, there exists a symplectic Runge-Kutta method

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) Lecture 19: Computing the SVD; Sparse Linear Systems Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical

More information

Orthogonal polynomials

Orthogonal polynomials Orthogonal polynomials Gérard MEURANT October, 2008 1 Definition 2 Moments 3 Existence 4 Three-term recurrences 5 Jacobi matrices 6 Christoffel-Darboux relation 7 Examples of orthogonal polynomials 8 Variable-signed

More information

On the Solution of Constrained and Weighted Linear Least Squares Problems

On the Solution of Constrained and Weighted Linear Least Squares Problems International Mathematical Forum, 1, 2006, no. 22, 1067-1076 On the Solution of Constrained and Weighted Linear Least Squares Problems Mohammedi R. Abdel-Aziz 1 Department of Mathematics and Computer Science

More information

Suppose that the approximate solutions of Eq. (1) satisfy the condition (3). Then (1) if η = 0 in the algorithm Trust Region, then lim inf.

Suppose that the approximate solutions of Eq. (1) satisfy the condition (3). Then (1) if η = 0 in the algorithm Trust Region, then lim inf. Maria Cameron 1. Trust Region Methods At every iteration the trust region methods generate a model m k (p), choose a trust region, and solve the constraint optimization problem of finding the minimum of

More information

The German word eigen is cognate with the Old English word āgen, which became owen in Middle English and own in modern English.

The German word eigen is cognate with the Old English word āgen, which became owen in Middle English and own in modern English. Chapter 4 EIGENVALUE PROBLEM The German word eigen is cognate with the Old English word āgen, which became owen in Middle English and own in modern English. 4.1 Mathematics 4.2 Reduction to Upper Hessenberg

More information

6.4 Krylov Subspaces and Conjugate Gradients

6.4 Krylov Subspaces and Conjugate Gradients 6.4 Krylov Subspaces and Conjugate Gradients Our original equation is Ax = b. The preconditioned equation is P Ax = P b. When we write P, we never intend that an inverse will be explicitly computed. P

More information

A DIVIDE-AND-CONQUER METHOD FOR THE TAKAGI FACTORIZATION

A DIVIDE-AND-CONQUER METHOD FOR THE TAKAGI FACTORIZATION SIAM J MATRIX ANAL APPL Vol 0, No 0, pp 000 000 c XXXX Society for Industrial and Applied Mathematics A DIVIDE-AND-CONQUER METHOD FOR THE TAKAGI FACTORIZATION WEI XU AND SANZHENG QIAO Abstract This paper

More information

Discrete ill posed problems

Discrete ill posed problems Discrete ill posed problems Gérard MEURANT October, 2008 1 Introduction to ill posed problems 2 Tikhonov regularization 3 The L curve criterion 4 Generalized cross validation 5 Comparisons of methods Introduction

More information

Linear Algebra: Matrix Eigenvalue Problems

Linear Algebra: Matrix Eigenvalue Problems CHAPTER8 Linear Algebra: Matrix Eigenvalue Problems Chapter 8 p1 A matrix eigenvalue problem considers the vector equation (1) Ax = λx. 8.0 Linear Algebra: Matrix Eigenvalue Problems Here A is a given

More information

Lecture 5 Singular value decomposition

Lecture 5 Singular value decomposition Lecture 5 Singular value decomposition Weinan E 1,2 and Tiejun Li 2 1 Department of Mathematics, Princeton University, weinan@princeton.edu 2 School of Mathematical Sciences, Peking University, tieli@pku.edu.cn

More information

The Singular Value Decomposition

The Singular Value Decomposition The Singular Value Decomposition An Important topic in NLA Radu Tiberiu Trîmbiţaş Babeş-Bolyai University February 23, 2009 Radu Tiberiu Trîmbiţaş ( Babeş-Bolyai University)The Singular Value Decomposition

More information

Positive Definite Matrix

Positive Definite Matrix 1/29 Chia-Ping Chen Professor Department of Computer Science and Engineering National Sun Yat-sen University Linear Algebra Positive Definite, Negative Definite, Indefinite 2/29 Pure Quadratic Function

More information

Iterative Methods for Solving A x = b

Iterative Methods for Solving A x = b Iterative Methods for Solving A x = b A good (free) online source for iterative methods for solving A x = b is given in the description of a set of iterative solvers called templates found at netlib: http

More information

On Lagrange multipliers of trust region subproblems

On Lagrange multipliers of trust region subproblems On Lagrange multipliers of trust region subproblems Ladislav Lukšan, Ctirad Matonoha, Jan Vlček Institute of Computer Science AS CR, Prague Applied Linear Algebra April 28-30, 2008 Novi Sad, Serbia Outline

More information

Backward Error Estimation

Backward Error Estimation Backward Error Estimation S. Chandrasekaran E. Gomez Y. Karant K. E. Schubert Abstract Estimation of unknowns in the presence of noise and uncertainty is an active area of study, because no method handles

More information

14 Singular Value Decomposition

14 Singular Value Decomposition 14 Singular Value Decomposition For any high-dimensional data analysis, one s first thought should often be: can I use an SVD? The singular value decomposition is an invaluable analysis tool for dealing

More information

Eigenvalues and Eigenvectors

Eigenvalues and Eigenvectors Chapter 1 Eigenvalues and Eigenvectors Among problems in numerical linear algebra, the determination of the eigenvalues and eigenvectors of matrices is second in importance only to the solution of linear

More information

Nonlinear Stationary Subdivision

Nonlinear Stationary Subdivision Nonlinear Stationary Subdivision Michael S. Floater SINTEF P. O. Box 4 Blindern, 034 Oslo, Norway E-mail: michael.floater@math.sintef.no Charles A. Micchelli IBM Corporation T.J. Watson Research Center

More information

15 Singular Value Decomposition

15 Singular Value Decomposition 15 Singular Value Decomposition For any high-dimensional data analysis, one s first thought should often be: can I use an SVD? The singular value decomposition is an invaluable analysis tool for dealing

More information

Preliminary Examination, Numerical Analysis, August 2016

Preliminary Examination, Numerical Analysis, August 2016 Preliminary Examination, Numerical Analysis, August 2016 Instructions: This exam is closed books and notes. The time allowed is three hours and you need to work on any three out of questions 1-4 and any

More information

Solutions and Notes to Selected Problems In: Numerical Optimzation by Jorge Nocedal and Stephen J. Wright.

Solutions and Notes to Selected Problems In: Numerical Optimzation by Jorge Nocedal and Stephen J. Wright. Solutions and Notes to Selected Problems In: Numerical Optimzation by Jorge Nocedal and Stephen J. Wright. John L. Weatherwax July 7, 2010 wax@alum.mit.edu 1 Chapter 5 (Conjugate Gradient Methods) Notes

More information

MATH 5640: Functions of Diagonalizable Matrices

MATH 5640: Functions of Diagonalizable Matrices MATH 5640: Functions of Diagonalizable Matrices Hung Phan, UMass Lowell November 27, 208 Spectral theorem for diagonalizable matrices Definition Let V = X Y Every v V is uniquely decomposed as u = x +

More information

S.F. Xu (Department of Mathematics, Peking University, Beijing)

S.F. Xu (Department of Mathematics, Peking University, Beijing) Journal of Computational Mathematics, Vol.14, No.1, 1996, 23 31. A SMALLEST SINGULAR VALUE METHOD FOR SOLVING INVERSE EIGENVALUE PROBLEMS 1) S.F. Xu (Department of Mathematics, Peking University, Beijing)

More information

Tikhonov Regularization of Large Symmetric Problems

Tikhonov Regularization of Large Symmetric Problems NUMERICAL LINEAR ALGEBRA WITH APPLICATIONS Numer. Linear Algebra Appl. 2000; 00:1 11 [Version: 2000/03/22 v1.0] Tihonov Regularization of Large Symmetric Problems D. Calvetti 1, L. Reichel 2 and A. Shuibi

More information

Review of similarity transformation and Singular Value Decomposition

Review of similarity transformation and Singular Value Decomposition Review of similarity transformation and Singular Value Decomposition Nasser M Abbasi Applied Mathematics Department, California State University, Fullerton July 8 7 page compiled on June 9, 5 at 9:5pm

More information

Math Introduction to Numerical Analysis - Class Notes. Fernando Guevara Vasquez. Version Date: January 17, 2012.

Math Introduction to Numerical Analysis - Class Notes. Fernando Guevara Vasquez. Version Date: January 17, 2012. Math 5620 - Introduction to Numerical Analysis - Class Notes Fernando Guevara Vasquez Version 1990. Date: January 17, 2012. 3 Contents 1. Disclaimer 4 Chapter 1. Iterative methods for solving linear systems

More information

The QR Decomposition

The QR Decomposition The QR Decomposition We have seen one major decomposition of a matrix which is A = LU (and its variants) or more generally PA = LU for a permutation matrix P. This was valid for a square matrix and aided

More information

On the influence of eigenvalues on Bi-CG residual norms

On the influence of eigenvalues on Bi-CG residual norms On the influence of eigenvalues on Bi-CG residual norms Jurjen Duintjer Tebbens Institute of Computer Science Academy of Sciences of the Czech Republic duintjertebbens@cs.cas.cz Gérard Meurant 30, rue

More information

Iterative Methods for Linear Systems of Equations

Iterative Methods for Linear Systems of Equations Iterative Methods for Linear Systems of Equations Projection methods (3) ITMAN PhD-course DTU 20-10-08 till 24-10-08 Martin van Gijzen 1 Delft University of Technology Overview day 4 Bi-Lanczos method

More information

Preconditioning Techniques Analysis for CG Method

Preconditioning Techniques Analysis for CG Method Preconditioning Techniques Analysis for CG Method Huaguang Song Department of Computer Science University of California, Davis hso@ucdavis.edu Abstract Matrix computation issue for solve linear system

More information

Basic Elements of Linear Algebra

Basic Elements of Linear Algebra A Basic Review of Linear Algebra Nick West nickwest@stanfordedu September 16, 2010 Part I Basic Elements of Linear Algebra Although the subject of linear algebra is much broader than just vectors and matrices,

More information

Chapter 7 Iterative Techniques in Matrix Algebra

Chapter 7 Iterative Techniques in Matrix Algebra Chapter 7 Iterative Techniques in Matrix Algebra Per-Olof Persson persson@berkeley.edu Department of Mathematics University of California, Berkeley Math 128B Numerical Analysis Vector Norms Definition

More information

Contribution of Wo¹niakowski, Strako²,... The conjugate gradient method in nite precision computa

Contribution of Wo¹niakowski, Strako²,... The conjugate gradient method in nite precision computa Contribution of Wo¹niakowski, Strako²,... The conjugate gradient method in nite precision computations ªaw University of Technology Institute of Mathematics and Computer Science Warsaw, October 7, 2006

More information

you expect to encounter difficulties when trying to solve A x = b? 4. A composite quadrature rule has error associated with it in the following form

you expect to encounter difficulties when trying to solve A x = b? 4. A composite quadrature rule has error associated with it in the following form Qualifying exam for numerical analysis (Spring 2017) Show your work for full credit. If you are unable to solve some part, attempt the subsequent parts. 1. Consider the following finite difference: f (0)

More information