Matrices and Moments: Least Squares Problems
|
|
- Verity Heath
- 5 years ago
- Views:
Transcription
1 Matrices and Moments: Least Squares Problems Gene Golub, SungEun Jo, Zheng Su Comparative study (with David Gleich) Computer Science Department Stanford University
2 Matrices and Moments: Least Squares Problems Outlines Part I: Review of least squares problems and gauss quadrature rules Outline of Part I Least Squares Problems Ordinary/Data/Total least squares SVD solutions Secular equation approaches Inverse least squares Gauss Quadrature Rules Gauss quadrature theory Tri-diagonalization for orthonormal polynomials Inverse eigenvalue problem for Gauss-Radau rule Matrices and Moments: Least Squares Problems (2/52)
3 Matrices and Moments: Least Squares Problems Outlines Part II: Application to solving secular equations Outline of Part II Interpolating secular equations A common term of secular functions in TLS/DLS Variations of Newton s methods Seeking the smallest root Conjugate Gradient method with Quadrature (CGQ) Solving TLS and DLS by quadrature and CG method Alternative implementations Comparative study (with David Gleich) Large-scale TLS algorithms Comparison of large-scale TLS results Numerical example CG-based algorithms Conclusion Matrices and Moments: Least Squares Problems (3/52)
4 Least Squares Problems Gauss Quadrature Rules Part I Review Matrices and Moments: Least Squares Problems (4/52)
5 Least Squares Problems Gauss Quadrature Rules Ordinary/Data/Total least squares Approximation problem Approximation problems for a linear system: Ax b, A R m n, b R m 1, m > n. Notations: A and b x T Tr[ ] 2 A F = Tr[A T A] 2 or F A and b OLS/DLS/TLS given data solution to determine transpose the sum of diagonal entries of matrix two-norm of vector Frobenius norm of matrix Euclidean norm residual quantities ordinary/data/total least squares Matrices and Moments: Least Squares Problems (5/52)
6 Least Squares Problems Gauss Quadrature Rules Ordinary/Data/Total least squares Statements and geometric equivalences Table: Problem statements and geometric equivalent statements OLS TLS DLS Problem statement min b 2 s.t. Ax = b+ b x, b min x, A, b [ A, b] F s.t. (A+ A)x = b+ b min A F s.t. (A+ A)x = b x, A a Geometric equivalence min x Ax b 2 2 min x Ax b 2 2 x min x Ax b 2 2 x 2 2 a The TLS/DLS equivalent statements are derived by means of the Lagrange method [GV96, DD93, JK05]. Ordinary Least Squares (OLS): correcting with b Data Least Squares (DLS): correcting with A Total Least Squares (TLS): correcting with A and b TLS is also known as Errors-in-Variables modeling. Matrices and Moments: Least Squares Problems (6/52)
7 Least Squares Problems Gauss Quadrature Rules SVD solutions Singular value decomposition (SVD) approach Table: Singular value decomposition approach TLS: min x Ax b 2 x DLS: min x Ax b 2 x 2 σ 1 min ([A,b]) s.t. v TLS (n+1) 0 σ min (P b A) s.t. bt Av DLS 0 v TLS (n+1) is the last component of v TLS. P b = I 1 b T b bbt 2 1 x TLS = v TLS (n+1) v TLS(1 : n) x DLS = bt b b T v Av DLS DLS 3 [ A TLS, b TLS ] = [A,b]v TLS v T TLS A DLS = P b Av DLSv T DLS 4 [ A TLS, b TLS ] F = σ TLS A DLS F = σ DLS 1 Equivalent singular value problem and feasibility condition. v TLS and v DLS are the right singular vectors associated with the smallest singular values, respectively, of [A,b] and P b A. σ min the minimum singular value; 2 SVD solution; 3 Minimal residual in terms of singular vector; 4 Norm of minimal residuals. Matrices and Moments: Least Squares Problems (7/52)
8 Least Squares Problems Gauss Quadrature Rules Secular equation approaches Secular equation approach Table: Secular equation approach in the generic case 1 TLS: σtls 2 = min Ax b 2 x x A T (Ax b) = σ 2 TLS x b T (b Ax) = σ 2 TLS DLS: σdls 2 = min Ax b 2 x x 2 A T (Ax b) = σ 2 DLS x b T (b Ax) = 0 2 σ min (A) > σ TLS σ min (A) > σ DLS 3 b T b b T A(A T A σ 2 TLS I) 1 A T b = σ 2 TLS b T b b T A(A T A σ 2 DLS I) 1 A T b = 0 4 x TLS = (A T A σ 2 TLS I) 1 A T b x DLS = (A T A σ 2 DLS I) 1 A T b 5 [ A TLS, b TLS ] = r TLS [xt TLS, 1] x TLS 2 +1 A DLS = r DLS xt DLS x DLS 2 1 Normal equations (for stationary points); 2 Generic condition; 3 Secular equation; 4 De-regularized solution; 5 Residuals in terms of solution x and r = b Ax. Matrices and Moments: Least Squares Problems (8/52)
9 Least Squares Problems Gauss Quadrature Rules Secular equation approaches When is the secular equation approach more preferable than the SVD approach? The problem is sensitive: σ TLS/DLS σ min (A) Least squares solution or de-regularized form is needed: OLS provides a good initial guess of solution. The de-regularized form can be easily calculated by adjusting the amount of negative shift. The problem is large: The SVD of a large matrix is very expensive. Instead, we can approximate the secular equation in the large-scaled problem by Gauss quadrature rules. Matrices and Moments: Least Squares Problems (9/52)
10 Least Squares Problems Gauss Quadrature Rules Gauss quadrature theory Riemann-Stieltjes integral M = QΛQ T, Λ = diag(λ i ) 0 < λ 1 λ 2... λ n. M is symmetric positive definite. Q is orthonormal. u T f(m)u = α T f(λ)α = n f(λ i )αi 2 = I[ f] i=1 f(m) is an analytic function of M that is defined on (0, ). α = Q T u for an arbitrary vector u. Riemann-Stieltjes integral I[ f]: b 0 if λ < a = λ 1 I[ f] f(λ)dα(λ), α(λ) = i j=1 α 2 j if λ i λ < λ i+1 a n j=1 α2 j if b = λ n λ where the measure α(λ) is piecewise constant. Matrices and Moments: Least Squares Problems (10/52)
11 Least Squares Problems Gauss Quadrature Rules Gauss quadrature theory Bounds for Riemann-Stieltjes integral The Gauss quadrature theory is formulated in terms of finite summations: b a f(λ)dα(λ) = N j=1 w j f(t j )+ M k=1 v k f(z k )+R[ f] Unknown weights: [w j ] N j=1,[v k] M k=1 ; Unknown nodes: [t j] N j=1. Prescribed nodes: [z k ] M k=1 The remainder term R[ f] is given by R[ f] = f (2N+M) (ξ) b (2N + M)! a M k=1 (λ z k ) [ N 2 (λ t j )] dα(λ), a < ξ < b. j=1 Golub and Meurant [GM93] showed that the sign of the remainder term R[ f] can be adjusted by the prescribed nodes. Setting M = 1, we will use the Gauss-Radau formula to get the bounds of (a part of) secular function. Matrices and Moments: Least Squares Problems (11/52)
12 Least Squares Problems Gauss Quadrature Rules Gauss quadrature theory Orthonormal polynomials Define a sequence of polynomials p 0 (λ), p 1 (λ),... that are orthonormal with respect to α(λ): b { 1 if i = j p i (λ)p j (λ)dα(λ) = 0 otherwise, a and p k (λ) is of exact degree k. Moreover, the roots of p k (λ) are distinct, real and lie in the interval [a, b]. Matrices and Moments: Least Squares Problems (12/52)
13 Least Squares Problems Gauss Quadrature Rules Tri-diagonalization for orthonormal polynomials Three-term recurrence relationship If dα = 1, the set of orthonormal polynomials satisfies: where λ p(λ) = T N p(λ)+γ N p N (λ)e N, p(λ) = [p 0 (λ) p 1 (λ) p N 1 (λ)] T, e N = (0 0 1) T R N, ω 1 γ 1 γ 1 ω 2 γ 2 T N = γ N 2 ω N 1 γ N 1 γ N 1 ω N Matrices and Moments: Least Squares Problems (13/52)
14 Least Squares Problems Gauss Quadrature Rules Tri-diagonalization for orthonormal polynomials Lanczos algorithm for quadratures To obtain the tri-diagonal matrix and hence the Gauss-Radau rule, we will use the Lanczos algorithm with p 1 = u/ u 2 as a starting vector: Eigenvalue decomposition of T N : Function of matrix: M PT N P T T N = QΛQ T u T f(m)u u T f(pt N P T )u = u 2 2 et 1 Q f(λ)qt e 1, where e 1 = (1 0 0) T R N. Thus, the eigenvalues of T N give us the nodes and the squares of the first elements of the eigenvectors give the weights: N w j f(t j ) = u 2 N 2 (Q 1i ) 2 f(λ i ) j=1 i=1 Matrices and Moments: Least Squares Problems (14/52)
15 Least Squares Problems Gauss Quadrature Rules Inverse eigenvalue problem for Gauss-Radau rule Inverse eigenvalue problem To obtain the Gauss-Radau rule, we extend the matrix T N in such a way that it has one prescribed eigenvalue z 1. Lemma The extended tri-diagonal matrix ( ) TN γ ˆT N+1 N e N γ N e T N ŵ N+1 has z 1 as an eigenvalue, where ŵ N+1 = z 1 + δ N, and δ N is the last entry of δ such that (T N z 1 I)δ = γ 2 N e N. (1) Matrices and Moments: Least Squares Problems (15/52)
16 Least Squares Problems Gauss Quadrature Rules Inverse eigenvalue problem for Gauss-Radau rule Proof of the extended tri-diagonal matrix lemma Proof. We can verify that z 1 is an eigenvalue of ˆT N+1 by investigating the following relation to get (1): ˆT N+1 d = z 1 d, where d is a corresponding eigenvector. Now, ˆT N+1 gives the weights and nodes of the Gauss-Radau rule such that N w j f(t j )+v 1 f(z 1 ) = u 2 e T 1 f( ˆT N+1 )e 1. j=1 The remainder is [ R[ f] = u 2 f (2N+1) (ξ) b N 2 (λ z 1 ) (λ t j )] dα(λ). (2N + 1)! a j=1 Matrices and Moments: Least Squares Problems (16/52)
17 Part II Application to solving secular equations Matrices and Moments: Least Squares Problems (17/52)
18 Common function of matrix Secular functions Recall the secular equations: T LS : ψ TLS (λ) = b T b b T A(A T A λ I) 1 A T b λ = 0, DLS : ψ DLS (λ) = b T b b T A(A T A λ I) 1 A T b = 0. ψ TLS/DLS(λ) is referred to as secular function. λ: an estimate of the minimum squared TLS/DLS distance. generic condition: λ < σ 2 min (A) for (AT A λ I) 1 Thus, in the domain of 0 λ < σmin 2 (A), we need to evaluate a matrix function of λ which is common in ψ TLS and ψ DLS : φ(λ) = b T A(A T A λ I) 1 A T b Matrices and Moments: Least Squares Problems (19/52)
19 Common function of matrix Bounds of a common function Now we evaluate the bounds of the scalar quantity φ: φ = g T f 1 x (M)g, M = A T A λ I, g = A T b, where f 1 (x) = 1 x x. Then, the quadrature rule ˆφ N+1 (z 1 ) = g 2 e T 1 f 1 ( ˆT N+1 )e 1 is described in terms of the remainder: x [ b N 2 ˆφ N+1 (z 1 ) = I[ f 1 ]+ g 2 (ξ) (2N+2) (λ z 1 ) (λ t j )] dα(λ). x a j=1 We note that f (2N+1) 1 x (ξ) (2N+1)! = (ξ) (2N+2) < 0, λ < a < ξ < b. Thus, we have the bounds: ˆφ N+1 (ζ b ) < I[ f 1 x ] < ˆφ N+1 (ζ a ), ζ a < a < b < ζ b. Matrices and Moments: Least Squares Problems (20/52)
20 Common function of matrix Comments on bounds f 1(x) is well defined on the proper interval (a, b) such that x the sign of the derivative function f (2N+1) 1 (ξ) is not changed x with the interval ξ (a, b). Since A 1 A > b, we may use ζ b = A 1 A. However, the lower bound of a (the smallest eigenvalue of M) is not easily obtainable. ζ a is determined very roughly. This explains why the upper bound of I[ f 1 x ] is usually poorer than the lower bound. Matrices and Moments: Least Squares Problems (21/52)
21 Common function of matrix Lanczos process with a shift for efficiency The tri-diagonalization is independent of the shift. Tri-diag([g, M]) Tri-diag([g, M+λ I]) (M+λ I)Q N = Q N (T N + λ I) MQ N = Q N T N Then we re-define the extended matrix ˆT N+1 as ( ) TN λ I J N+1 N γ N e N γ N e T, N w where T N and γ N are calculated by Tri-Diag of [g, A T A], (not [g, M]), and w is determined so that J N+1 has a prescribed eigenvalue z 1. Thus, w = z 1 + d N, where d N is the last entry of d such that (T N (z 1 + λ)i N )d = γ 2 N e N. Matrices and Moments: Least Squares Problems (22/52)
22 Common function of matrix Finally, we have the quadrature for bounds: ˆφ N+1 (z 1 ) = g 2 e T 1 f 1 x (J N+1 )e 1 = g 2 e T 1 (J N+1 ) 1 e 1. Once we solve a tri-diagonal J N+1 y = e 1 for y, we have ˆφ N+1 (z 1 ) = g 2 e T 1 y = g 2 y 1. For later interpolation, we need to evaluate the derivatives of the matrix function φ(λ) w.r.t. λ by approximating with f 1 (x) = x 2 and f x 2 1 (x) = x 3 : x 3 ˆφ = g 2 e T 1 f 1 (J N+1)e 1 = g 2 e T x 2 1 (J N+1 ) 2 e 1, ˆφ = 2 g 2 e T 1 f 1 (J N+1)e 1 = 2 g 2 e T x 3 1 (J N+1 ) 3 e 1. By solving J N+1 h = y., we have ˆφ = g 2 y 2, ˆφ = 2 g 2 y T (J N+1 ) 1 y = 2 g 2 y T h. A symmetric, tri-diagonal, and positive definite system requires O(N) flops [GV96] to be solved Ṁatrices and Moments: Least Squares Problems (23/52)
23 Common function of matrix Lemma (Monotonicity of bound sequences) Along with Lanczos processes, a sequence of bound estimates of φ = g T f 1 x (M)g with full-rank symmetric M R n n and f 1(x) = 1 x x is generated by Gauss quadrature rules. Then the estimated sequence ˆφ N+1 is necessarily monotonic. In other words, given each prescribed node ζ a or ζ b such that ζ a < σ min (M) < σ max (M) < ζ b, the lower and upper bound sequences for I[ f 1 x ] satisfy < ˆφ N (ζ b ) < ˆφ N+1 (ζ b ) < < I[ f 1 x ] < < ˆφ N+1 (ζ a ) < ˆφ N (ζ a ) <. Note that the complete Lanczos processes yield the exact evaluation: φ = g 2 e T 1 f 1 x (T n λ I n )e 1. Matrices and Moments: Least Squares Problems (24/52)
24 Newton methods Interpolating the root of secular equations We approximate the common function φ(λ) by using Lanczos processes combined with Gauss quadrature rule, where we proceed the processes until the upper- and lower- bounds of φ(λ) match within a tolerance. Suppose λ k is the current estimate of the minimum distance. In order to interpolate the root of the secular equation ψ(λ k ), we need to evaluate the followings: ψ TLS (λ k ) = b 2 λ k φ(λ k ), ψ DLS (λ k ) = b 2 φ(λ k ). ψ TLS (λ k) = 1 φ (λ k ), ψ DLS (λ k) = φ (λ k ). ψ TLS(λ k ) = φ (λ k ), ψ DLS(λ k ) = φ (λ k ). Then, consider one-point interpolating methods to obtain λ k+1 such that ψ(λ k+1 ) = 0. Note that roots of secular equation consist of the stationary points of the geometrically equivalent cost function. We want to find the smallest root λ k+1 [0, σmin 2 (A)) from the definition of TLS/DLS problem. However, we can not achieve it without using additional information on the locations of poles such as σ 2 min (A) min j i a i j 2 and σ 2 max(a) A 1 A. In the following sections, we will discuss how to use bisection and the upper-bound of the smallest pole. Matrices and Moments: Least Squares Problems (25/52)
25 Newton methods Variations of Newton s method take the form: λ k+1 = λ k ψ(λ k) ψ (λ k ) C k, where C k denotes a convergence factor [Gan78, Gan85] according to methods such as the Newton s method, the Halley s variation, and simple rational approximation in the following Table. Table: Variations of Newton s method Newton s SRA a Halley s Interpolating function h(λ) ψ(λ) h(λ) = c 0 + c 1 λ h(λ) = b 2 c 1 c 2 λ h(λ) = c 0 c 1 ( c 2 λ ) 1 ψ(λ k )ψ (λ k ) 2(ψ (λ k )) 2 b Convergence factor C k 1 2 ψ(λ k ) b 2 1/ Rate of (local) convergence Quadratic Quadratic Cubic Convergence region b Narrow Wide Wider Algebraic interpretation c b g(λ) = ψ(λ) g(λ) = 1 2 b 2 g(λ) = ψ(λ) ψ(λ) ψ (λ) a Simple Rational Approximation b Global convergence in root-finding of secular equations c Equivalently, solve g(λ) = 0 by Newton s method. Matrices and Moments: Least Squares Problems (26/52)
26 Seeking the smallest root Mixing with bisection Whenever we detect that the root estimate is larger than σmin 2 (A), we bisect the estimate to assure that it is less than σmin 2 (A). The monotonicity of the sequence of estimate bounds of Gauss quadratures (GQ) is utilized based on the following scenario: 1. With the initial guess of root, we obtain the sequence of bounds of secular function by means of GQ. 2. If the sequences are not monotonic, we conclude that the root estimate is larger than the squared smallest singular value of A. Then the root estimate is cut by half, and go to Step 1 with the modified estimate of root. 3. Otherwise, we interpolate the root of secular equation by using the estimate of the secular function and its derivatives. 4. If the new root estimate is close to the previous one within a tolerance, then we calculate the de-regularized solution, and stop the algorithm. Otherwise, go to step 1. Although the violation of monotonicity is only a necessary condition, our numerical simulation works well. Matrices and Moments: Least Squares Problems (27/52)
27 Seeking the smallest root Stabilizing with the estimated smallest pole Although the bisection scheme almost always achieves the smallest root, it may suffer from a bi-stability problem which means the estimates are alternating between two values. To get around this, we employ the estimation of the smallest pole by modifying the previous scenario. If we detect the current estimate of root is larger than σmin 2 (A), we cut the estimate by half and set the upper-bound of the smallest pole to the current estimate of root as well. ˆσ 2 min (A) = λ k, λ k+1 = 1 2 λ k. Then, when we interpolate the next estimate of root, we take a harmonic sum between the Newton-based step δ k = ψ(λ k) ψ (λ k ) C k and the distance from the upper-bound of the smallest pole estimation. 1 λ k+1 = λ k + = λ k δ k + δk 2 /(δ k + λ k ˆσ min 2 (A)) δ k ˆσ min 2 (A) λ k Matrices and Moments: Least Squares Problems (28/52)
28 Solving TLS and DLS by quadrature and CG method CGQ as a secular equation approach 1. Find the smallest root of secular equation for TLS or DLS. 1.1 Evaluate the bounds of secular function by Gauss-Radau quadrature rule. 1.2 Interpolate the zero of the function by a variation of Newton method. 1.3 Determine a proper interval for the smallest zero by bisection and harmonic-summation with the upper-bound of the smallest pole. 2. Solve a de-regularized system with a shift of the smallest root. Solve the symmetric, positive-definite system by the conjugate gradient (CG) method, Or, solve the tri-diagonal system with shift. Matrices and Moments: Least Squares Problems (29/52)
29 Alternative implementations Alternative implementations Reuse of Lanczos vectors with sufficient memory Regeneration of Lanczos vectors with knowledge of tri-diagonal entries Avoiding of explicit multiplication of A T A Shifting into Lanczos bi-diagonalization Using backward perturbations Matrices and Moments: Least Squares Problems (30/52)
30 Backward perturbations for linear least squares I min x b Ax 2, A : m n, b : m 1. ξ : arbitrary vector, calculated. ˆx = ξ + e; ˆx = A + b. ξ = (A+δA) + b. ρ = b Aξ. µ(ξ) = min δa δa F. µ(ξ) = min{ ρ 2 ξ 2,σ min (A,B)} (Karlson, Waldén & Sun) B = ρ 2 ξ 2 (I ρ ρt ρ 2 ).
31 Backward perturbations for linear least squares II µ 2 (ξ) = ρ T A(αA T A+β I) 1 A T ρ α = ξ 2 2, β = ρ 2 2 µ(ξ) µ(ξ) lim ξ ˆx µ(ξ) µ( ˆx) = 1 (Grcar)
32 Large-scale TLS algorithms Björck s algorithm Solve the system of nonlinear equations [ A T A A T ][ ] [ ] b x x b T A b T = λ, b 1 1 or equivalently, the system [ f(x,λ) g(x,λ) ] = [ A T ] r λx b T = r+λ [ ] 0 0 with r = b Ax using a Rayleigh-quotient iteration (RQI). This algorithm will always converge to a singular value/vector pair, but we might not get λ = σn+1 2. Björck suggested one initial inverse iteration (i.e. λ = 0) to move closer to the desired λ, and then apply the RQI procedure. Matrices and Moments: Least Squares Problems (34/52)
33 Large-scale TLS algorithms Björck s algorithm details After some manipulations, Björck s algorithm greatly simplifies. The following presentation emphasizes the computationally intensive steps. 1. Solve the least squares system in A,b for x LS. 2. Perform one inverse iteration (solve A T Ax (0) = x LS ) to get the initial x (0). 3. While not converged... solve two systems in (A T A λi)x = b to iterate to x (k+1), but if we detect A T A λi is negative definite, decrease λ and repeat the iteration. Björck suggests using the PCGTLS algorithm to solve each linear system with the Cholesky factor R or A T A as a preconditioner. Inside the CG procedure, we detect A T A λi is negative definite and use the CG vectors to compute a new value of λ. Matrices and Moments: Least Squares Problems (35/52)
34 Large-scale TLS algorithms Notes on our implementation of Björck s algorithm Instead of using PCGTLS with the Cholesky factor of A T A, we use a matrix-free approach for large scale m > 800 and apply the unpreconditioned CGTLS algorithm. The tolerance used in the CG method is To compute the initial least squares solution, we use the LSQR algorithm. We detect convergence when λ changes by less than or the normalized residual increases (theory states the normalized residual always decreases). After we detect convergence, we run one more iteration of the algorithm to ensure that we compute an x for the λ. Matrices and Moments: Least Squares Problems (36/52)
35 Large-scale TLS algorithms Details of the matrix moments based algorithm Algorithm 2 uses the Golub-Kahan bidiagonalization of A and applies the moment algorithm to T = B T B instead of computing T directly from the Lanczos process on A T A. Algorithm 1 restarts the Lanczos process at each iteration. Algorithm 2 never restarts the bidiagonalization process and simply continues the process at each iteration. Matrices and Moments: Least Squares Problems (37/52)
36 Comparison of large-scale TLS results Problems Jo s problems, 15 8 and Björck s problem 1: matrix Large scale problems with and matrices. The large scale problems were generated using random Householder matrices to build the SVD of [ A b ] in product form. Each large-scale matrix was available solely as an operator to all of the algorithms. The singular values of [ ] A b are σ i = log(i)+ N(0,1), where N(0, 1) is a standard normal random variable. Matrices and Moments: Least Squares Problems (38/52)
37 Comparison of large-scale TLS results Parameter choices Algorithm 1 Algorithm 2 λ (0) = 0 λ (0) = 1 λ (0) = ρ λ (0) = 0 λ (0) = 1 λ (0) = ρ newton sra halley newton *8 *8 *7 sra *12 *24 *7 halley *14 *23 *6 newton *20 *7 *10 sra *20 25 *64 halley *55 55 *12 newton *15 *11 *11 sra ++ *15 *25 *14 halley *20 *57 *11 newton sra halley * wrong root; ++ correct w/o convergence; no convergence Matrices and Moments: Least Squares Problems (39/52)
38 Comparison of large-scale TLS results Convergence TEST ALG ITERS ERROR TIME LANZ. jo björck (15,8) Alg σ 2 = Alg jo björck (750,400) Alg 1 > σ 2 = Alg large-scale björck (10000,5000) Alg 1 > σ 2 = Alg large-scale björck (100000,60000) Alg 1 > σ 2 = Alg björck björck (30,15) Alg 1 >100 σ σ 2 = Alg Matrices and Moments: Least Squares Problems (40/52)
39 Numerical data generation The numerical data are generated as follows [GvM91]. Let the error-free data matrix A s = U s Σ s V T s, where U s = I m 2 us ut s u T and V s = I n 2 vs vt s s us v T are Householder matrices and s vs Σ s = diag(σ k ) is a m-by-n diagonal matrix with [σ 1,,σ n ] = [ n,,1]. The vectors u s and v s consist of pseudo-random numbers generated by a function randn providing uniformly distributed random numbers in Matlab TM. An error-free observation vector b s is computed by b s = A s x s. The error-prone data A and b are generated with a given deviation σ n as follows: randn( state,108881);% seed for random number u s = randn(m, 1); v s = randn(n, 1); A s = U s Σ s V T s ; x s = (1./(1 : n)) ;% [1, 1/2,, 1/n] T b s = A s x s ; σ n = 0.3; b = b s + σ n randn(m, 1); A = A s + σ n randn(m, n); Matrices and Moments: Least Squares Problems (42/52)
40 CG-based algorithms CG-based algorithms We can categorize algorithms in the following example as below. Table: CG-based algorithms for solving secular equation. CGQ-BS-KP CGQ-BS CG-BS-KP CG-BS Gauss Quadrature rules Used Used Not Not Bisection Used Used Used Used Upper-bound of the smallest pole Used Not Used Not The previous CG-based approaches in [GJK06] or [BHM00] employ matrix-vector multiplications to evaluate the secular function and derivative from the intermediate solution x(λ) such that (A T A λ I)x(λ) = A T b. Then the secular functions and derivatives are obtained as follows. ψ TLS (λ) = b 2 b T Ax(λ) λ ψ TLS (λ) = x(λ) 2 1 ψ DLS (λ) = b 2 b T Ax(λ) ψ DLS (λ) = x(λ) 2 Matrices and Moments: Least Squares Problems (43/52)
41 Small problem Secular function plots with A R TLS secular function, ψ TLS (λ) 20 DLS secular function, ψ DLS (λ) Squared distances, λ (a) TLS secular function plot Squared distances, λ (b) DLS secular function plot Matrices and Moments: Least Squares Problems (44/52)
42 Small problem Convergent curves in TLS with A R 15 8 Squared distance TLSCGQ Newton KP (6), x = TLSCGQ SRA KP (6), x = TLSCGQ Halley KP (6), x = Iterations (a) Variations of CGQ with the bisection and the smallest pole estimation. Squared distance Squared distance TLSCGQ Newton (6), x = TLSCGQ SRA (5), x = TLSCGQ Halley (5), x = Iterations (b) Variations of CGQ with the bisection only. Squared distance TLSCG Newton KP (5), x = TLSCG HS KP (5), x = Iterations (c) CG with two different interpolations, the bisection and the smallest pole estimation. 0.1 TLSCG RQI (4), x = TLSCG HSI (5), x = Iterations (d) CG with two different inverse iterations, and the bisection only. Matrices and Moments: Least Squares Problems (45/52)
43 Small problem Convergent curves in DLS with A R 15 8 Squared distance DLSCGQ Newton KP (6), x = DLSCGQ SRA KP (6), x = DLSCGQ Halley KP (6), x = Iterations (a) Variations of CGQ with the bisection and the smallest pole estimation. 1 Squared distance DLSCGQ Newton (7), x = DLSCGQ SRA (6), x = DLSCGQ Halley (5), x = Iterations (b) Variations of CGQ with the bisection only Squared distance Squared distance DLSCG Newton KP (5), x = DLSCG HS KP (5), x = Iterations (c) CG with two different interpolations, the bisection and the smallest pole estimation. 0.2 DLSCG RQI (4), x = DLSCG HSI (5), x = Iterations (d) CG with two different inverse iterations, and the bisection only. Matrices and Moments: Least Squares Problems (46/52)
44 Large problem Convergent curves in TLS with A R Squared distance Squared distance TLSCGQ Newton KP (15), x = TLSCGQ SRA KP (9), x = TLSCGQ Halley KP (6), x = Iterations (a) Variations of CGQ with the bisection and the smallest pole estimation. 20 TLSCGQ Newton (20), x = TLSCGQ SRA (20), x = TLSCGQ Halley (20), x = Iterations (b) Variations of CGQ with the bisection only Squared distance 10 5 TLSCG Newton KP (12), x = TLSCG HS KP (12), x = Iterations (c) CG with two different interpolations, the bisection and the smallest pole estimation. Squared distance TLSCG RQI (20), x = TLSCG HSI (20), x = Iterations (d) CG with two different inverse iterations, and the bisection only. Matrices and Moments: Least Squares Problems (47/52)
45 Large problem Convergent curves in DLS with A R Squared distance Squared distance DLSCGQ Newton KP (20), x = DLSCGQ SRA KP (20), x = DLSCGQ Halley KP (13), x = Iterations (a) Variations of CGQ with the bisection and the smallest pole estimation DLSCGQ Newton (20), x = DLSCGQ SRA (20), x = DLSCGQ Halley (20), x = Iterations (b) Variations of CGQ with the bisection only Squared distance Squared distance DLSCG Newton KP (9), x = DLSCG HS KP (9), x = Iterations (c) CG with two different interpolations and the smallest pole estimation. 5 DLSCG RQI (20), x = DLSCG HSI (20), x = Iterations (d) CG with two different inverse iterations, and the bisection only. Matrices and Moments: Least Squares Problems (48/52)
46 Conclusion Conclusion The presented Conjugate Gradient (CG) with Quadrature (CGQ) approximates secular functions by means of Gauss Quadrature (GQ) rules. The previous CG based approaches are exhaustively calculating intermediate solutions to evaluate secular functions. Interpolating the smallest root by variations of Newton method with stabilized modification: Bisection to assure of the convergence to the smallest root. Approximating with rational function of the estimated smallest pole to get around bi-stability problem. CGQ does not pursue the intermediate solution until the root estimate converges. The overall computational complexity can be reduced by adjusting accuracy of GQ. Matrices and Moments: Least Squares Problems (50/52)
47 References Åke Björck, P. Heggernes, and P. Matstoms. Methods for large scale total least squares problems. SIAM J. Matrix Anal. Appl., 22(3): , R. D. DeGroat and E. M. Dowling. The data least squares problem and channel equalization. 41(1): , January Walter Gander. On the linear least squares problem with a quadratic constraint. Technical Report STAN-CS , Stanford University, W. Gander. On halley s iteration method. The American mathematical monthly, 92(2): , Gene H. Golub, SungEun Jo, and Sang Woo Kim. Solving secular equations for total/data least squares problems. SIAM J. Matrix Anal. Appl., submitted for publication, G. H. Golub and G. Meurant. Matrices, moments and quadrature. Report SCCM-93-07, Computer Science Department, Stanford University, Matrices and Moments: Least Squares Problems (51/52)
48 References Gene H. Golub and Charles. F. Van Loan. Matrix Computations. Johns Hopkins Univ. Pr., Baltimore, MD, third edition, November G. H. Golub and U. von Matt. Quadratically constrained least squares and quadratic problems. Numerische Mathematik, 59(1): , December SungEun Jo and Sang Woo Kim. Consistent normalized least mean square filtering with noisy data matrix. (6): , June Matrices and Moments: Least Squares Problems (52/52)
Total least squares. Gérard MEURANT. October, 2008
Total least squares Gérard MEURANT October, 2008 1 Introduction to total least squares 2 Approximation of the TLS secular equation 3 Numerical experiments Introduction to total least squares In least squares
More informationIntroduction. Chapter One
Chapter One Introduction The aim of this book is to describe and explain the beautiful mathematical relationships between matrices, moments, orthogonal polynomials, quadrature rules and the Lanczos and
More informationThe Lanczos and conjugate gradient algorithms
The Lanczos and conjugate gradient algorithms Gérard MEURANT October, 2008 1 The Lanczos algorithm 2 The Lanczos algorithm in finite precision 3 The nonsymmetric Lanczos algorithm 4 The Golub Kahan bidiagonalization
More informationKey words. conjugate gradients, normwise backward error, incremental norm estimation.
Proceedings of ALGORITMY 2016 pp. 323 332 ON ERROR ESTIMATION IN THE CONJUGATE GRADIENT METHOD: NORMWISE BACKWARD ERROR PETR TICHÝ Abstract. Using an idea of Duff and Vömel [BIT, 42 (2002), pp. 300 322
More informationBlock Bidiagonal Decomposition and Least Squares Problems
Block Bidiagonal Decomposition and Least Squares Problems Åke Björck Department of Mathematics Linköping University Perspectives in Numerical Analysis, Helsinki, May 27 29, 2008 Outline Bidiagonal Decomposition
More informationNumerical Methods in Matrix Computations
Ake Bjorck Numerical Methods in Matrix Computations Springer Contents 1 Direct Methods for Linear Systems 1 1.1 Elements of Matrix Theory 1 1.1.1 Matrix Algebra 2 1.1.2 Vector Spaces 6 1.1.3 Submatrices
More informationMatrices, Moments and Quadrature, cont d
Jim Lambers CME 335 Spring Quarter 2010-11 Lecture 4 Notes Matrices, Moments and Quadrature, cont d Estimation of the Regularization Parameter Consider the least squares problem of finding x such that
More informationESTIMATES OF THE TRACE OF THE INVERSE OF A SYMMETRIC MATRIX USING THE MODIFIED CHEBYSHEV ALGORITHM
ESTIMATES OF THE TRACE OF THE INVERSE OF A SYMMETRIC MATRIX USING THE MODIFIED CHEBYSHEV ALGORITHM GÉRARD MEURANT In memory of Gene H. Golub Abstract. In this paper we study how to compute an estimate
More informationNon-stationary extremal eigenvalue approximations in iterative solutions of linear systems and estimators for relative error
on-stationary extremal eigenvalue approximations in iterative solutions of linear systems and estimators for relative error Divya Anand Subba and Murugesan Venatapathi* Supercomputer Education and Research
More informationMoments, Model Reduction and Nonlinearity in Solving Linear Algebraic Problems
Moments, Model Reduction and Nonlinearity in Solving Linear Algebraic Problems Zdeněk Strakoš Charles University, Prague http://www.karlin.mff.cuni.cz/ strakos 16th ILAS Meeting, Pisa, June 2010. Thanks
More informationSolving discrete ill posed problems with Tikhonov regularization and generalized cross validation
Solving discrete ill posed problems with Tikhonov regularization and generalized cross validation Gérard MEURANT November 2010 1 Introduction to ill posed problems 2 Examples of ill-posed problems 3 Tikhonov
More informationGolub-Kahan iterative bidiagonalization and determining the noise level in the data
Golub-Kahan iterative bidiagonalization and determining the noise level in the data Iveta Hnětynková,, Martin Plešinger,, Zdeněk Strakoš, * Charles University, Prague ** Academy of Sciences of the Czech
More informationAN ITERATIVE METHOD WITH ERROR ESTIMATORS
AN ITERATIVE METHOD WITH ERROR ESTIMATORS D. CALVETTI, S. MORIGI, L. REICHEL, AND F. SGALLARI Abstract. Iterative methods for the solution of linear systems of equations produce a sequence of approximate
More informationSensitivity of Gauss-Christoffel quadrature and sensitivity of Jacobi matrices to small changes of spectral data
Sensitivity of Gauss-Christoffel quadrature and sensitivity of Jacobi matrices to small changes of spectral data Zdeněk Strakoš Academy of Sciences and Charles University, Prague http://www.cs.cas.cz/
More informationChapter 12 Solving secular equations
Chapter 12 Solving secular equations Gérard MEURANT January-February, 2012 1 Examples of Secular Equations 2 Secular equation solvers 3 Numerical experiments Examples of secular equations What are secular
More informationComputational Methods. Eigenvalues and Singular Values
Computational Methods Eigenvalues and Singular Values Manfred Huber 2010 1 Eigenvalues and Singular Values Eigenvalues and singular values describe important aspects of transformations and of data relations
More informationNumerical Methods for Solving Large Scale Eigenvalue Problems
Peter Arbenz Computer Science Department, ETH Zürich E-mail: arbenz@inf.ethz.ch arge scale eigenvalue problems, Lecture 2, February 28, 2018 1/46 Numerical Methods for Solving Large Scale Eigenvalue Problems
More informationarxiv: v1 [math.na] 5 May 2011
ITERATIVE METHODS FOR COMPUTING EIGENVALUES AND EIGENVECTORS MAYSUM PANJU arxiv:1105.1185v1 [math.na] 5 May 2011 Abstract. We examine some numerical iterative methods for computing the eigenvalues and
More informationSolving Constrained Rayleigh Quotient Optimization Problem by Projected QEP Method
Solving Constrained Rayleigh Quotient Optimization Problem by Projected QEP Method Yunshen Zhou Advisor: Prof. Zhaojun Bai University of California, Davis yszhou@math.ucdavis.edu June 15, 2017 Yunshen
More informationEfficient and Accurate Rectangular Window Subspace Tracking
Efficient and Accurate Rectangular Window Subspace Tracking Timothy M. Toolan and Donald W. Tufts Dept. of Electrical Engineering, University of Rhode Island, Kingston, RI 88 USA toolan@ele.uri.edu, tufts@ele.uri.edu
More informationA Note on Inverse Iteration
A Note on Inverse Iteration Klaus Neymeyr Universität Rostock, Fachbereich Mathematik, Universitätsplatz 1, 18051 Rostock, Germany; SUMMARY Inverse iteration, if applied to a symmetric positive definite
More informationTotal Least Squares Approach in Regression Methods
WDS'08 Proceedings of Contributed Papers, Part I, 88 93, 2008. ISBN 978-80-7378-065-4 MATFYZPRESS Total Least Squares Approach in Regression Methods M. Pešta Charles University, Faculty of Mathematics
More informationEIGENVALUE PROBLEMS. EIGENVALUE PROBLEMS p. 1/4
EIGENVALUE PROBLEMS EIGENVALUE PROBLEMS p. 1/4 EIGENVALUE PROBLEMS p. 2/4 Eigenvalues and eigenvectors Let A C n n. Suppose Ax = λx, x 0, then x is a (right) eigenvector of A, corresponding to the eigenvalue
More informationBaltzer Journals April 24, Bounds for the Trace of the Inverse and the Determinant. of Symmetric Positive Denite Matrices
Baltzer Journals April 4, 996 Bounds for the Trace of the Inverse and the Determinant of Symmetric Positive Denite Matrices haojun Bai ; and Gene H. Golub ; y Department of Mathematics, University of Kentucky,
More informationNumerical Analysis Preliminary Exam 10 am to 1 pm, August 20, 2018
Numerical Analysis Preliminary Exam 1 am to 1 pm, August 2, 218 Instructions. You have three hours to complete this exam. Submit solutions to four (and no more) of the following six problems. Please start
More informationETNA Kent State University
Electronic Transactions on Numerical Analysis. Volume 14, pp. 2-35, 22. Copyright 22,. ISSN 168-9613. ETNA L-CURVE CURVATURE BOUNDS VIA LANCZOS BIDIAGONALIZATION D. CALVETTI, P. C. HANSEN, AND L. REICHEL
More informationM.A. Botchev. September 5, 2014
Rome-Moscow school of Matrix Methods and Applied Linear Algebra 2014 A short introduction to Krylov subspaces for linear systems, matrix functions and inexact Newton methods. Plan and exercises. M.A. Botchev
More informationBindel, Fall 2016 Matrix Computations (CS 6210) Notes for
1 Algorithms Notes for 2016-10-31 There are several flavors of symmetric eigenvalue solvers for which there is no equivalent (stable) nonsymmetric solver. We discuss four algorithmic ideas: the workhorse
More informationOn Lagrange multipliers of trust-region subproblems
On Lagrange multipliers of trust-region subproblems Ladislav Lukšan, Ctirad Matonoha, Jan Vlček Institute of Computer Science AS CR, Prague Programy a algoritmy numerické matematiky 14 1.- 6. června 2008
More informationNotes on Eigenvalues, Singular Values and QR
Notes on Eigenvalues, Singular Values and QR Michael Overton, Numerical Computing, Spring 2017 March 30, 2017 1 Eigenvalues Everyone who has studied linear algebra knows the definition: given a square
More informationc 1999 Society for Industrial and Applied Mathematics
SIAM J. MATRIX ANAL. APPL. Vol. 21, No. 1, pp. 185 194 c 1999 Society for Industrial and Applied Mathematics TIKHONOV REGULARIZATION AND TOTAL LEAST SQUARES GENE H. GOLUB, PER CHRISTIAN HANSEN, AND DIANNE
More informationFEM and sparse linear system solving
FEM & sparse linear system solving, Lecture 9, Nov 19, 2017 1/36 Lecture 9, Nov 17, 2017: Krylov space methods http://people.inf.ethz.ch/arbenz/fem17 Peter Arbenz Computer Science Department, ETH Zürich
More informationChap 3. Linear Algebra
Chap 3. Linear Algebra Outlines 1. Introduction 2. Basis, Representation, and Orthonormalization 3. Linear Algebraic Equations 4. Similarity Transformation 5. Diagonal Form and Jordan Form 6. Functions
More informationCOMP 558 lecture 18 Nov. 15, 2010
Least squares We have seen several least squares problems thus far, and we will see more in the upcoming lectures. For this reason it is good to have a more general picture of these problems and how to
More informationComparison of A-Posteriori Parameter Choice Rules for Linear Discrete Ill-Posed Problems
Comparison of A-Posteriori Parameter Choice Rules for Linear Discrete Ill-Posed Problems Alessandro Buccini a, Yonggi Park a, Lothar Reichel a a Department of Mathematical Sciences, Kent State University,
More informationPreliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012
Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.
More informationNumerical Methods - Numerical Linear Algebra
Numerical Methods - Numerical Linear Algebra Y. K. Goh Universiti Tunku Abdul Rahman 2013 Y. K. Goh (UTAR) Numerical Methods - Numerical Linear Algebra I 2013 1 / 62 Outline 1 Motivation 2 Solving Linear
More informationLast Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection
Eigenvalue Problems Last Time Social Network Graphs Betweenness Girvan-Newman Algorithm Graph Laplacian Spectral Bisection λ 2, w 2 Today Small deviation into eigenvalue problems Formulation Standard eigenvalue
More informationLanczos tridigonalization and Golub - Kahan bidiagonalization: Ideas, connections and impact
Lanczos tridigonalization and Golub - Kahan bidiagonalization: Ideas, connections and impact Zdeněk Strakoš Academy of Sciences and Charles University, Prague http://www.cs.cas.cz/ strakos Hong Kong, February
More informationarxiv: v1 [math.na] 1 Sep 2018
On the perturbation of an L -orthogonal projection Xuefeng Xu arxiv:18090000v1 [mathna] 1 Sep 018 September 5 018 Abstract The L -orthogonal projection is an important mathematical tool in scientific computing
More information5 Selected Topics in Numerical Linear Algebra
5 Selected Topics in Numerical Linear Algebra In this chapter we will be concerned mostly with orthogonal factorizations of rectangular m n matrices A The section numbers in the notes do not align with
More informationEigenvalue Problems. Eigenvalue problems occur in many areas of science and engineering, such as structural analysis
Eigenvalue Problems Eigenvalue problems occur in many areas of science and engineering, such as structural analysis Eigenvalues also important in analyzing numerical methods Theory and algorithms apply
More informationDELFT UNIVERSITY OF TECHNOLOGY
DELFT UNIVERSITY OF TECHNOLOGY REPORT -09 Computational and Sensitivity Aspects of Eigenvalue-Based Methods for the Large-Scale Trust-Region Subproblem Marielba Rojas, Bjørn H. Fotland, and Trond Steihaug
More informationScientific Computing: An Introductory Survey
Scientific Computing: An Introductory Survey Chapter 4 Eigenvalue Problems Prof. Michael T. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c 2002. Reproduction
More informationGreedy Tikhonov regularization for large linear ill-posed problems
International Journal of Computer Mathematics Vol. 00, No. 00, Month 200x, 1 20 Greedy Tikhonov regularization for large linear ill-posed problems L. Reichel, H. Sadok, and A. Shyshkov (Received 00 Month
More informationthat determines x up to a complex scalar of modulus 1, in the real case ±1. Another condition to normalize x is by requesting that
Chapter 3 Newton methods 3. Linear and nonlinear eigenvalue problems When solving linear eigenvalue problems we want to find values λ C such that λi A is singular. Here A F n n is a given real or complex
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra)
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 1: Course Overview & Matrix-Vector Multiplication Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 20 Outline 1 Course
More informationOn the Computations of Eigenvalues of the Fourth-order Sturm Liouville Problems
Int. J. Open Problems Compt. Math., Vol. 4, No. 3, September 2011 ISSN 1998-6262; Copyright c ICSRS Publication, 2011 www.i-csrs.org On the Computations of Eigenvalues of the Fourth-order Sturm Liouville
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 18 Outline
More informationUNIFYING LEAST SQUARES, TOTAL LEAST SQUARES AND DATA LEAST SQUARES
UNIFYING LEAST SQUARES, TOTAL LEAST SQUARES AND DATA LEAST SQUARES Christopher C. Paige School of Computer Science, McGill University, Montreal, Quebec, Canada, H3A 2A7 paige@cs.mcgill.ca Zdeněk Strakoš
More informationLecture 4 Eigenvalue problems
Lecture 4 Eigenvalue problems Weinan E 1,2 and Tiejun Li 2 1 Department of Mathematics, Princeton University, weinan@princeton.edu 2 School of Mathematical Sciences, Peking University, tieli@pku.edu.cn
More informationCG and MINRES: An empirical comparison
School of Engineering CG and MINRES: An empirical comparison Prequel to LSQR and LSMR: Two least-squares solvers David Fong and Michael Saunders Institute for Computational and Mathematical Engineering
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra)
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 16: Reduction to Hessenberg and Tridiagonal Forms; Rayleigh Quotient Iteration Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical
More informationSVD and its Application to Generalized Eigenvalue Problems. Thomas Melzer
SVD and its Application to Generalized Eigenvalue Problems Thomas Melzer June 8, 2004 Contents 0.1 Singular Value Decomposition.................. 2 0.1.1 Range and Nullspace................... 3 0.1.2
More informationCS137 Introduction to Scientific Computing Winter Quarter 2004 Solutions to Homework #3
CS137 Introduction to Scientific Computing Winter Quarter 2004 Solutions to Homework #3 Felix Kwok February 27, 2004 Written Problems 1. (Heath E3.10) Let B be an n n matrix, and assume that B is both
More informationSolving Regularized Total Least Squares Problems
Solving Regularized Total Least Squares Problems Heinrich Voss voss@tu-harburg.de Hamburg University of Technology Institute of Numerical Simulation Joint work with Jörg Lampe TUHH Heinrich Voss Total
More informationNumerical Methods I Eigenvalue Problems
Numerical Methods I Eigenvalue Problems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 October 2nd, 2014 A. Donev (Courant Institute) Lecture
More information5.3 The Power Method Approximation of the Eigenvalue of Largest Module
192 5 Approximation of Eigenvalues and Eigenvectors 5.3 The Power Method The power method is very good at approximating the extremal eigenvalues of the matrix, that is, the eigenvalues having largest and
More informationSecond Order Optimization Algorithms I
Second Order Optimization Algorithms I Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A. http://www.stanford.edu/ yyye Chapters 7, 8, 9 and 10 1 The
More informationLecture 9: Krylov Subspace Methods. 2 Derivation of the Conjugate Gradient Algorithm
CS 622 Data-Sparse Matrix Computations September 19, 217 Lecture 9: Krylov Subspace Methods Lecturer: Anil Damle Scribes: David Eriksson, Marc Aurele Gilles, Ariah Klages-Mundt, Sophia Novitzky 1 Introduction
More information3D Computer Vision - WT 2004
3D Computer Vision - WT 2004 Singular Value Decomposition Darko Zikic CAMP - Chair for Computer Aided Medical Procedures November 4, 2004 1 2 3 4 5 Properties For any given matrix A R m n there exists
More informationKrylov Subspace Methods for the Evaluation of Matrix Functions. Applications and Algorithms
Krylov Subspace Methods for the Evaluation of Matrix Functions. Applications and Algorithms 4. Monotonicity of the Lanczos Method Michael Eiermann Institut für Numerische Mathematik und Optimierung Technische
More informationAN INVERSE EIGENVALUE PROBLEM AND AN ASSOCIATED APPROXIMATION PROBLEM FOR GENERALIZED K-CENTROHERMITIAN MATRICES
AN INVERSE EIGENVALUE PROBLEM AND AN ASSOCIATED APPROXIMATION PROBLEM FOR GENERALIZED K-CENTROHERMITIAN MATRICES ZHONGYUN LIU AND HEIKE FAßBENDER Abstract: A partially described inverse eigenvalue problem
More informationQR-decomposition. The QR-decomposition of an n k matrix A, k n, is an n n unitary matrix Q and an n k upper triangular matrix R for which A = QR
QR-decomposition The QR-decomposition of an n k matrix A, k n, is an n n unitary matrix Q and an n k upper triangular matrix R for which In Matlab A = QR [Q,R]=qr(A); Note. The QR-decomposition is unique
More informationFull-State Feedback Design for a Multi-Input System
Full-State Feedback Design for a Multi-Input System A. Introduction The open-loop system is described by the following state space model. x(t) = Ax(t)+Bu(t), y(t) =Cx(t)+Du(t) () 4 8.5 A =, B =.5.5, C
More informationSOME PROPERTIES OF SYMPLECTIC RUNGE-KUTTA METHODS
SOME PROPERTIES OF SYMPLECTIC RUNGE-KUTTA METHODS ERNST HAIRER AND PIERRE LEONE Abstract. We prove that to every rational function R(z) satisfying R( z)r(z) = 1, there exists a symplectic Runge-Kutta method
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)
AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) Lecture 19: Computing the SVD; Sparse Linear Systems Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical
More informationOrthogonal polynomials
Orthogonal polynomials Gérard MEURANT October, 2008 1 Definition 2 Moments 3 Existence 4 Three-term recurrences 5 Jacobi matrices 6 Christoffel-Darboux relation 7 Examples of orthogonal polynomials 8 Variable-signed
More informationOn the Solution of Constrained and Weighted Linear Least Squares Problems
International Mathematical Forum, 1, 2006, no. 22, 1067-1076 On the Solution of Constrained and Weighted Linear Least Squares Problems Mohammedi R. Abdel-Aziz 1 Department of Mathematics and Computer Science
More informationSuppose that the approximate solutions of Eq. (1) satisfy the condition (3). Then (1) if η = 0 in the algorithm Trust Region, then lim inf.
Maria Cameron 1. Trust Region Methods At every iteration the trust region methods generate a model m k (p), choose a trust region, and solve the constraint optimization problem of finding the minimum of
More informationThe German word eigen is cognate with the Old English word āgen, which became owen in Middle English and own in modern English.
Chapter 4 EIGENVALUE PROBLEM The German word eigen is cognate with the Old English word āgen, which became owen in Middle English and own in modern English. 4.1 Mathematics 4.2 Reduction to Upper Hessenberg
More information6.4 Krylov Subspaces and Conjugate Gradients
6.4 Krylov Subspaces and Conjugate Gradients Our original equation is Ax = b. The preconditioned equation is P Ax = P b. When we write P, we never intend that an inverse will be explicitly computed. P
More informationA DIVIDE-AND-CONQUER METHOD FOR THE TAKAGI FACTORIZATION
SIAM J MATRIX ANAL APPL Vol 0, No 0, pp 000 000 c XXXX Society for Industrial and Applied Mathematics A DIVIDE-AND-CONQUER METHOD FOR THE TAKAGI FACTORIZATION WEI XU AND SANZHENG QIAO Abstract This paper
More informationDiscrete ill posed problems
Discrete ill posed problems Gérard MEURANT October, 2008 1 Introduction to ill posed problems 2 Tikhonov regularization 3 The L curve criterion 4 Generalized cross validation 5 Comparisons of methods Introduction
More informationLinear Algebra: Matrix Eigenvalue Problems
CHAPTER8 Linear Algebra: Matrix Eigenvalue Problems Chapter 8 p1 A matrix eigenvalue problem considers the vector equation (1) Ax = λx. 8.0 Linear Algebra: Matrix Eigenvalue Problems Here A is a given
More informationLecture 5 Singular value decomposition
Lecture 5 Singular value decomposition Weinan E 1,2 and Tiejun Li 2 1 Department of Mathematics, Princeton University, weinan@princeton.edu 2 School of Mathematical Sciences, Peking University, tieli@pku.edu.cn
More informationThe Singular Value Decomposition
The Singular Value Decomposition An Important topic in NLA Radu Tiberiu Trîmbiţaş Babeş-Bolyai University February 23, 2009 Radu Tiberiu Trîmbiţaş ( Babeş-Bolyai University)The Singular Value Decomposition
More informationPositive Definite Matrix
1/29 Chia-Ping Chen Professor Department of Computer Science and Engineering National Sun Yat-sen University Linear Algebra Positive Definite, Negative Definite, Indefinite 2/29 Pure Quadratic Function
More informationIterative Methods for Solving A x = b
Iterative Methods for Solving A x = b A good (free) online source for iterative methods for solving A x = b is given in the description of a set of iterative solvers called templates found at netlib: http
More informationOn Lagrange multipliers of trust region subproblems
On Lagrange multipliers of trust region subproblems Ladislav Lukšan, Ctirad Matonoha, Jan Vlček Institute of Computer Science AS CR, Prague Applied Linear Algebra April 28-30, 2008 Novi Sad, Serbia Outline
More informationBackward Error Estimation
Backward Error Estimation S. Chandrasekaran E. Gomez Y. Karant K. E. Schubert Abstract Estimation of unknowns in the presence of noise and uncertainty is an active area of study, because no method handles
More information14 Singular Value Decomposition
14 Singular Value Decomposition For any high-dimensional data analysis, one s first thought should often be: can I use an SVD? The singular value decomposition is an invaluable analysis tool for dealing
More informationEigenvalues and Eigenvectors
Chapter 1 Eigenvalues and Eigenvectors Among problems in numerical linear algebra, the determination of the eigenvalues and eigenvectors of matrices is second in importance only to the solution of linear
More informationNonlinear Stationary Subdivision
Nonlinear Stationary Subdivision Michael S. Floater SINTEF P. O. Box 4 Blindern, 034 Oslo, Norway E-mail: michael.floater@math.sintef.no Charles A. Micchelli IBM Corporation T.J. Watson Research Center
More information15 Singular Value Decomposition
15 Singular Value Decomposition For any high-dimensional data analysis, one s first thought should often be: can I use an SVD? The singular value decomposition is an invaluable analysis tool for dealing
More informationPreliminary Examination, Numerical Analysis, August 2016
Preliminary Examination, Numerical Analysis, August 2016 Instructions: This exam is closed books and notes. The time allowed is three hours and you need to work on any three out of questions 1-4 and any
More informationSolutions and Notes to Selected Problems In: Numerical Optimzation by Jorge Nocedal and Stephen J. Wright.
Solutions and Notes to Selected Problems In: Numerical Optimzation by Jorge Nocedal and Stephen J. Wright. John L. Weatherwax July 7, 2010 wax@alum.mit.edu 1 Chapter 5 (Conjugate Gradient Methods) Notes
More informationMATH 5640: Functions of Diagonalizable Matrices
MATH 5640: Functions of Diagonalizable Matrices Hung Phan, UMass Lowell November 27, 208 Spectral theorem for diagonalizable matrices Definition Let V = X Y Every v V is uniquely decomposed as u = x +
More informationS.F. Xu (Department of Mathematics, Peking University, Beijing)
Journal of Computational Mathematics, Vol.14, No.1, 1996, 23 31. A SMALLEST SINGULAR VALUE METHOD FOR SOLVING INVERSE EIGENVALUE PROBLEMS 1) S.F. Xu (Department of Mathematics, Peking University, Beijing)
More informationTikhonov Regularization of Large Symmetric Problems
NUMERICAL LINEAR ALGEBRA WITH APPLICATIONS Numer. Linear Algebra Appl. 2000; 00:1 11 [Version: 2000/03/22 v1.0] Tihonov Regularization of Large Symmetric Problems D. Calvetti 1, L. Reichel 2 and A. Shuibi
More informationReview of similarity transformation and Singular Value Decomposition
Review of similarity transformation and Singular Value Decomposition Nasser M Abbasi Applied Mathematics Department, California State University, Fullerton July 8 7 page compiled on June 9, 5 at 9:5pm
More informationMath Introduction to Numerical Analysis - Class Notes. Fernando Guevara Vasquez. Version Date: January 17, 2012.
Math 5620 - Introduction to Numerical Analysis - Class Notes Fernando Guevara Vasquez Version 1990. Date: January 17, 2012. 3 Contents 1. Disclaimer 4 Chapter 1. Iterative methods for solving linear systems
More informationThe QR Decomposition
The QR Decomposition We have seen one major decomposition of a matrix which is A = LU (and its variants) or more generally PA = LU for a permutation matrix P. This was valid for a square matrix and aided
More informationOn the influence of eigenvalues on Bi-CG residual norms
On the influence of eigenvalues on Bi-CG residual norms Jurjen Duintjer Tebbens Institute of Computer Science Academy of Sciences of the Czech Republic duintjertebbens@cs.cas.cz Gérard Meurant 30, rue
More informationIterative Methods for Linear Systems of Equations
Iterative Methods for Linear Systems of Equations Projection methods (3) ITMAN PhD-course DTU 20-10-08 till 24-10-08 Martin van Gijzen 1 Delft University of Technology Overview day 4 Bi-Lanczos method
More informationPreconditioning Techniques Analysis for CG Method
Preconditioning Techniques Analysis for CG Method Huaguang Song Department of Computer Science University of California, Davis hso@ucdavis.edu Abstract Matrix computation issue for solve linear system
More informationBasic Elements of Linear Algebra
A Basic Review of Linear Algebra Nick West nickwest@stanfordedu September 16, 2010 Part I Basic Elements of Linear Algebra Although the subject of linear algebra is much broader than just vectors and matrices,
More informationChapter 7 Iterative Techniques in Matrix Algebra
Chapter 7 Iterative Techniques in Matrix Algebra Per-Olof Persson persson@berkeley.edu Department of Mathematics University of California, Berkeley Math 128B Numerical Analysis Vector Norms Definition
More informationContribution of Wo¹niakowski, Strako²,... The conjugate gradient method in nite precision computa
Contribution of Wo¹niakowski, Strako²,... The conjugate gradient method in nite precision computations ªaw University of Technology Institute of Mathematics and Computer Science Warsaw, October 7, 2006
More informationyou expect to encounter difficulties when trying to solve A x = b? 4. A composite quadrature rule has error associated with it in the following form
Qualifying exam for numerical analysis (Spring 2017) Show your work for full credit. If you are unable to solve some part, attempt the subsequent parts. 1. Consider the following finite difference: f (0)
More information