ON MEINARDUS EXAMPLES FOR THE CONJUGATE GRADIENT METHOD
|
|
- Kristian Wheeler
- 6 years ago
- Views:
Transcription
1 MATHEMATICS OF COMPUTATION Volume 77 Number 61 January 008 Pages S (07)019-9 Article electronically published on September ON MEINARDUS EXAMPLES FOR THE CONJUGATE GRADIENT METHOD REN-CANG LI Abstract. The conjugate gradient (CG) method is widely used to solve a positive definite linear system Ax b of order N. It is well known that the relative residual of the kth approximate solution by CG (with the initial approximation x 0 0) is bounded above by [ ] 1 +1 k + k with 1 where (A) A A 1 is A s spectral condition number. In 1963 Meinardus (Numer. Math. 5 (1963) pp. 14 3) gave an example to achieve this bound for k N 1 but without saying anything about all other 1 k< N 1. This very example can be used to show that the bound is sharp for any given k by constructing examples to attain the bound but such examples depend on k and for them the (k + 1)th residual is exactly zero. Therefore it would be interesting to know if there is any example on which the CG relative residuals are comparable to the bound for all 1 k N 1. There are two contributions in this paper: (1) A closed formula for the CG residuals for all 1 k N 1 on Meinardus example is obtained and in particular it implies that the bound is always within a factor of of the actual residuals; () A complete characterization of extreme positive linear systems for which the kth CG residual achieves the bound is also presented. 1. Introduction The conjugate gradient (CG) method is widely used to solve a positive definite linear system Ax b (often with certain preconditioning). The basic idea is to seek approximate solutions from the so-called Krylov subspaces. While different implementation may render different numerical behavior mathematically 1 the kth approximate solution x k by CG is the optimal one in the sense that the kth approximation error A 1 b x k satisfies [11 Theorem 6:1] (1.1) A 1 b x k A min x K k A 1 b x A Received by the editor September Revised January Mathematics Subject Classification. Primary 65F10. Key words and phrases. Conjugate gradient method Krylov subspace rate of convergence Vandermonde matrix condition number. This work was supported in part by the National Science Foundation CAREER award under Grant No. CCR and by the National Science Foundation under Grant No. DMS Without loss of generality we assume that A is already preconditioned and the initial approximation x c 007 American Mathematical Society Reverts to public domain 8 years from publication
2 336 REN-CANG LI Residual bounds over actual residuals Equidist eigenvalues on [δ 1] δ10-1 [minλ j max λ j ][0.11] δ10 - [min λ j maxλ j ][0.011] N k Residual bounds over actual residuals Random eigenvalues on [δ1] δ 10-1 [minλ j maxλ j ][ ] δ 10 - [minλ j maxλ j ][ ] N k Figure 1.1. Conjugate gradient method for Ax b with A diag(λ 1 λ...λ n )andb the vector of all ones Ratios of (1.4) over the actual residuals for equidistant or random distributed λ j s or equivalently the kth residual r k b Ax k satisfies (1.) r k A 1 min b Ax A 1 x K k where K k K k (A b) isthekth Krylov subspace of A on b defined as (1.3) K k K k (A b) def span{bab...a k 1 b} def and M-vector norm z M z Mz. Here the superscript takes conjugate transpose. In practice x k is computed recursively from x k 1 via short term recurrences [ ]. But exactly how it is computed though extremely crucial in practice is not important to our analysis here in this paper. CG always converges for positive definite A. In fact we have the following well known and frequently referenced error bound (see e.g. [ ]): (1.4) r k A 1 r 0 A 1 A 1 b x k A A 1 b A [ k + k ] 1 where (A) A A 1 is the spectral condition number generic notation is for either the spectral norm (the largest singular value) of a matrix or the euclidian length of a vector and (1.5) t def t +1 t 1 for t>0 that will be used frequently later for different t. The widely cited Kaniel [13] (1966) gave a proof of (1.4) while saying This result is known with a pointer to Meinardus [18] (1963). The same bound was proved to hold for Richardson-like processes [5 pp. 8 31] making it likely that (1.4) could be known before 1963 because their proofs do not differ much. This paper is concerned with how sharp this well-known error bound is. Consider A diag(λ 1 λ...λ n )andb the vector of all ones where λ j [δ 1] is either randomly or equidistantly distributed on the interval. Figure 1.1 plots the ratios of the residual bounds by (1.4) over the actual residuals. What it shows is that initially for small k bounds by (1.4) are good indications of actual residuals but
3 MEINARDUS EXAMPLES FOR CONJUGATE GRADIENT METHOD 337 as k becomes larger and larger this bound overestimates the actual ones too much to be of any use. Are the phenomena in Figure 1.1 representative? Often this is what people observed [0] known as superlinear convergence. In [18] Meinardus devised an N N positive definite linear system Ax b for whichheprovedthat r N 1 A 1 r 0 A 1 [ N 1 ] 1 + (N 1) but without saying anything about all other 1 k<n 1. This example of Meinardus can be easily modified to give examples which achieve the error bound in (1.4) for any given 1 k<n 1; e.g. trivially embed an example of Meinardus of dimension k + 1 with zero blocks to make it of dimension N. Greenbaum [10 p. 5] and [9] claimed that the error bound for a given k can be attained for A having eigenvalues at the extreme points of a translated Chebyshev polynomial of degree k and some particular b whose explicit form was not given however. The same conclusion can be reached from [1 p. 561] if A s eigenvalues are chosen as Greenbaum s. This in a sense shows that the error bound in (1.4) is sharp and cannot be improved in general. But examples i.e. A and b constructed as such depend on the given step-index k and CG on any of these examples for k other than the example it was constructed for behaves much differently and in particular r k+1 0 exactly. So this only proves that the error bound is locally sharp: for given k [ ] k γ + k 1 γ (1.6) max 1. (A)γ r k A 1/ r 0 A 1 What about its global sharpness? For example Is there any positive definite system Ax b for which relative (1.7) residuals r k A 1/ r 0 A 1 achieve the error bounds by (1.4) for all 1 k<n 1? This question turns out to be too strong and the answer is no by Kaniel [13 Theorem 4.4] who showed that if r k attains the bound then it must be r k+1 0 (see also Theorem. below). So instead we ask (1.8) Is modestly bounded? sup (A)γ max 1 k n 1 [ ] k γ + k 1 γ r k A 1/ r 0 A 1 This question has been recently answered positively in Li [14] using A with eigenvalues being the translated zeros of Nth Chebyshev polynomial of the first kind. It is proved there that the ratio in (1.8) can be bounded from above by a bound that asymptotically approaches to γ / γ 1ask goes to infinity. It depends on (A) γ and unfortunately can be arbitrarily large as γ 1 +.Amuchstronger bound on the ratio namely is implied later in this paper. In what follows we shall compute the CG residuals on Meinardus examples for all 1 k N 1 and investigate extreme positive linear systems for which the kth CG residual achieves the error bound in (1.4). Before we set out to do so let us look at some numerical examples. Figure 1. plots the ratios of the error bounds
4 338 REN-CANG LI Residual bounds over actual residuals N50 Residual bounds over actual residuals N k k Figure 1.. Ratios of the error bound in (1.4) over the exact CG residuals for Meinardus example. by (1.4) over the actual CG relative residuals i.e. the right-hand side of (1.4) over its left-hand side on a Meinardus example where the exact CG residuals were carefully computed within MAPLE with a sufficiently high precision. While it is not surprising at all to see that the ratios are not smaller than 1 they seem not to be bigger than as well. This in fact will be confirmed by one of our main results which will also furnish another example for the global sharpness question (1.8) in addition to the one in [14]. The rest of this paper is organized as follows. Section explains Meinardus examples and gives our main results the closed formula for CG residuals for a Meinardus example and a complete characterization of extreme positive linear systems for which the kth CG residual achieves the error bound in (1.4). Proofs for our main results are rather long and thus are given separately in Section 3 and Section 4. Concluding remarks are given in Section 5. Notation. Throughout this paper C n m is the set of all n m complex matrices C n C n 1 andc C 1. Similarly define R n m R n andrexcept replacing the word complex by real. I n (or simply I if its dimension is clear from the context) is the n n identity matrix and e j is its jth column. The superscript T takes transpose only. We shall also adopt MATLAB-like convention to access the entries of vectors and matrices. i : j is the set of integers from i to j inclusive. For vector u and matrix X u (j) is u s jth entry X (ij) is X s (i j)th entry diag(u) isthe diagonal matrix with (diag(u)) (jj) u (j) ; X s submatrices X (k:li:j) X (k:l:) and X (:i:j) consist of intersections of row k to row l and column i to column j rowk to row l andcolumni to column j respectively.. Meinardus examples and main results The mth Chebyshev polynomial of the first kind is (.1) T m (t) cos(m arccos t) for t 1 1 ( t + ) m t 1 ( 1 + t m (.) t 1) for t 1. It frequently shows up in numerical analysis and computations because of its numerous nice properties for example T m (t) 1for t 1and T m (t) grows extremely
5 MEINARDUS EXAMPLES FOR CONJUGATE GRADIENT METHOD 339 fast for t > 1. Later we will need ( ) ( ) (.3) 1+t T m t +1 1 t T m 1 [ m t 1 t + m ] t for 1 t>0. The first equality holds because T m ( t) ( 1) m T m (t). We shall prove the second equality for 0 <t<1 only and a proof for t>1 is similar. For 0 <t<1 we have ( ) 1+t 1+t 1 t t + t 1+ t 1 t 1 t 1 t t which proves (.3) for 0 <t<1. T m (t)hasm+1 extreme points in [ 1 1] so-called the mth Chebyshev extreme nodes: (.4) τ jm cosϑ jm ϑ jm j π 0 j m m at which T m (τ jm ) 1. Givenα<βset (.5) ω β α The linear transformation (.6) t(z) z ω + τ β α > 0 τ α + β β α. ( z α + β ) maps z [α β] one-to-one and onto t [ 1 1]. With its inverse transformation x(t) ω(t τ) we define the so-called mth translated Chebyshev extreme nodes on [α β]: (.7) τjm tr ω(τ jm τ) 0 j m. It can be verified that τ 0m β and τ mm α. Now we are ready to state Meinardus examples. Assume 0 <α<β. For the sake of presentation set n N 1. Let Q be any N N unitary matrix. A Meinardus example is a positive definite system Ax b with (.8) A QΛQ b QΛ 1/ g where (.9) Λ def diag(τ tr 0nτ tr 1n...τ tr nn) g (j+1) def 1/τjn tr for j {0n} /τjn tr for 1 j n 1. So an example of Meinardus is any member of the family parameterized by unitary Q. Theorem.1 is one of the two main results of this paper. Theorem.1. Let 0 <α<βand let A and b be given by (.8) and (.9). r k is the kth CG residual with initially r 0 b. Then (.10) r k A 1 r 0 A 1 ρ k [ k + k for 1 k n where (A) β/α and 1 (.11) < 1 ( ) ( 1+ n n +1 ρ k 1 ] 1 1+ k + (n k) n +1 ) 1.
6 340 REN-CANG LI Remark.1. (1) As far as the equality is concerned (.10) is valid for k 0 as well which corresponds to the very beginning of CG. () The factor ρ k is symmetrical in k about n/ i.e. ρ k ρ n k. This phenomenon certainly showed up in Figure 1. which equivalently plotted ρ 1 k. (3) ρ k 1 with equality if and only if k 0orn. (4) ρ k is strictly decreasing for k n/ (the largest integer that is no bigger than n/) and strictly increasing for k n/ (the smallest integer that is no less than n/) and 1 < min ρ k ρ n/ 1 as n. 0 k n The fact that ρ n 1 has already been established by Meinardus [18]. With it one can easily construct a positive definite linear system Ax b for which the kth CG residual achieves the error bound in (1.4). For example A and b are given by (.8) and (.9) where Λ diag(τ0k tr τtr 1k...τtr kk...) i.e. k +1 of A s eigenvalues are τ0k tr τtr 1k...τtr kk andg (j+1) is 1/τjk tr for j {0k} and /τjk tr for 1 j k 1 and zero for all other j then r k A 1/ r 0 A 1 [ ] 1. k + k For this example r k+1 0 i.e. convergence occurs at the (k + 1)th step! This is not a coincidence as it must be due to Kaniel [13 Theorem 4.4]. The following theorem characterizes all extreme linear systems as such. Theorem.. Let Ax b 0be a positive definite linear system of order N and 1 k<n.ifthekth CG residual r k (initially r 0 b) achieves the error bound in (1.4) i.e. (.1) r k A 1 r 0 A 1 [ k + k ] 1 where (A) A A 1 then the following statements hold. (1) A QΛQ and b QΛ 1/ g for some unitary Q C N N Λdiag(λ 1 λ...λ N ) with 0 <λ 1 λ λ N andg R N with all g (j) 0. () λ j λ 1 g(j) > 0 and λ j λ N g(j) > 0. (3) Let α min j λ j andβ max j λ j andletτjk tr be the translated Chebyshev extreme nodes on [α β]. The distinct λ j s in {λ j : g (j) > 0} consist of exactly τjk tr 0 j k i.e. {τjk tr 0 j k} {λ j : g (j) > 0} and λ i {τjk tr 0 j k} if g (i) > 0. (4) [13 Theorem 4.4] r k+1 0. (5) Let J l {j : λ j τlk tr g (j) > 0}. For some constant µ>0 { 1/τ tr (.13) g Jl µ lk for l {0k} /τ tr lk for 1 l k 1. Remark.. Any Ax b described by items (1) () (3) and (5) is essentially equivalent to an example of Meinardus (.8) and (.9) with N k + 1. Therefore this theorem practically says that the kth CG residual r k (initially r 0 b) achieves the error bound in (1.4) if and only if Ax b is an example of Meinardus. Thus
7 MEINARDUS EXAMPLES FOR CONJUGATE GRADIENT METHOD 341 unless N there is no positive linear system whose kth CG residual achieves the error bound in (1.4) for all 1 k<n. 3. Proof of Theorem.1 We will adopt in whole the notation introduced in Section and assume 0 < α<β. Recall in particular n N 1andA is N N. Theorem.1 will be proved through a restatement. For A as in (.8) and (.9) (3.1) min x K k b Ax A 1 min φ k (0)1 φ k(a)b A 1 min φ k (0)1 φ k(λ)λ 1/ Q b min φ k (0)1 φ k(λ)g min u (1) 1 diag(g) V T k+1nu where φ k (t) is a polynomial of degree k u C k+1 and with α j+1 0 j n def α 1 α α N (3.) V k+1n α1 k α k αn k τ tr jn for a(k +1) N rectangular Vandermonde matrix. Note also that r 0 A 1 g. Therefore after substitution k +1 k Theorem.1 can be equivalently stated as follows. Theorem 3.1. Let 0 <α<β g as in (.9) andv kn as in (3.) with α j+1 τ tr jn for 0 j n. Then diag(g) VkN T (3.3) min u [ ρ k 1 k 1 u (1) 1 g for 1 k N n +1where β/α. ] 1 + (k 1). The rest of this section is devoted to the proof of this theorem. T j (t(z)) T j (z/ω + τ) is a polynomial of degree j in z; sowewrite Notice that T j (z/ω + τ) a jj z j + a j 1 j z j a 1j z + a 0j where a ij a ij (ω τ) are functions of ω and τ in (.5). Their explicit dependence on ω and τ is often suppressed for convenience. For integer m 1 define upper triangular R m R m m a matrix-valued function in ω and τ as (3.4) R m R m (ω τ) def a 00 a 01 a 0 a 0 m 1 a 11 a 1 a 1 m 1 a a m 1....ȧm 1 m 1
8 34 REN-CANG LI i.e. the jth column consists of the coefficients of T j 1 (z/ω + τ). Write V N V NN for short and set T 0 (τ 0n ) T 0 (τ 1n ) T 0 (τ nn ) def T 1 (τ 0n ) T 1 (τ 1n ) T 1 (τ nn ) (3.5) S N.... T n (τ 0n ) T n (τ 1n ) T n (τ nn ) Then V T N R N S T N.SinceR N is upper triangular we have (3.6) V T kn S T knr 1 k a key decomposition of VkN T that will play a vital role later in our proofs where S kn (S N ) (1:k:) is S N s first k rows. Set (3.7) Ω diag( ) R N N Υ def S N ΩS T N. Lemma 3. below says Υ is diagonal. So essentially (3.6) gives a QR-like decomposition of VkN T. Lemma 3.1 below is probably well known but a precise reference is hard to find. However it can be proved by using either Euler identity cos θ (e ιkθ + e ιkθ )/ where ι 1 the imaginary unit or the identities in [8 p. 30]. Detail is omitted. Lemma 3.1. Let ϑ kn πk/n as in (.4). Then N if l mn for some integer m (3.8) cos lϑ kn 0 if l is odd k0 1 if l is even but l mn for any integer m. Lemma 3. is known to [ p. 33] and probably long before that. A proof can be given with the help of Lemma 3.1 and again detail is omitted. Lemma 3.. Let S N Ω andυ be defined as in (3.5) and (3.7). ThenΥ n Ω 1. Lemma 3.3. Let Γdiag(µ+ν cos ϑ 0n µ+ν cos ϑ 1n...µ+ν cos ϑ nn ) and define Υ µν def S N ΩΓS T N whereµ ν C. We have (3.9) Υ µν n 4 Ω 1 (µω+νh)ω 1 where H R N N Proof. Notice that Υ µν µυ 10 + νυ 01 and Υ 10 S N ΩS T N Υ n Ω 1 by Lemma 3.. It is enough to calculate Υ 01. Using to mean the first and last
9 MEINARDUS EXAMPLES FOR CONJUGATE GRADIENT METHOD 343 terms halved we have for 0 i j n (3.10) (S N ΩΓS T N) (i+1j+1) (S N ) (i+1k) (µ + ν cos ϑ kn )(S T N ) (kj+1) µ k0 T i (τ kn )(µ + ν cos ϑ kn )T j (τ kn ) k0 cos iϑ kn (µ + ν cos ϑ kn )cosjϑ kn k0 +ν cos iϑ kn cos jϑ kn k0 cos iϑ kn cos ϑ kn cos jϑ kn. k0 So (Υ 01 ) (i+1j+1) n k0 cos iϑ kn cos ϑ kn cos jϑ kn.now 4 cos iϑ kn cos ϑ kn cos jϑ kn k0 cos(i + j +1)ϑ kn + k0 + cos(i + j 1)ϑ kn k0 cos(i j +1)ϑ kn + k0 cos(i j 1)ϑ kn. Apply Lemma 3.1 to conclude Υ 01 n 4 Ω 1 HΩ 1 whose verification is straightforward albeit tedious. Lemma 3.4. Let m n and ξ C such that ( ξω+h) (1:m1:m) is nonsingular. Then the first entry of the solution to ( ξω+h) (1:m1:m) y e 1 is y (1) k0 γ m γ m + ξ 1(γ m + γm + ) where γ ± ξ ± ξ 1. Proof. Expand y to have a 0th entry y (0) and a (m + 1)th entry y (m+1) satisfying (3.11) y (0) ξy (1) 1 y (m+1) 0. Entry-wise we have y (i 1) ξy (i) + y (i+1) 0 for 1 i m. The general solution has form y (i) c + γ i + + c γ i whereγ ± are the two roots of 1 ξγ + γ 0 i.e. γ ± ξ ± ξ 1. We now determine c + and c by the edge conditions (3.11): (1 ξγ + ) c + + (1 ξγ ) c 1 γ m+1 + c + + γ m+1 c 0.
10 344 REN-CANG LI Notice γ + γ 1and (1 ξγ + )γ m+1 (1 ξγ )γ m+1 + (γ ξ)γ m (γ + ξ)γ m + ξ 1(γ m + γ m + ) to get γ m+1 c + ξ 1(γ m + γ+ m ) c +γ+ m+1 ξ 1(γ m + γ+ m ). Finally y (1) c + γ + + c γ. In its present general form the next lemma was proved in [14]. implied by the proof of [1 Theorem.1]. See also [16]. It was also Lemma 3.5. If Z has full column rank then (3.1) min u (1) 1 Zu [ e T 1 (Z Z) 1 e 1 ] 1/. Proof. Set v Zu. SinceZ has full column rank its Moore-Penrose pseudo-inverse is Z (Z Z) 1 Z [1] and thus u Z v. This gives a one-one and onto mapping between u C m and the column space v span(z). Now (3.13) min Zu Zu min min u (1) 1 u u (1) v span(z) where the last min is achieved at v e T 1 Z v min v v opt ( e T 1 Z ) Z(Z Z) 1 e 1 span(z) v e T 1 Z v et 1 Z 1 which implies the in (3.13) is actually an equality and u opt Z v opt /e T 1 Z v opt. Finally e T 1 Z e T 1 Z (Z ) e 1 e T 1 (Z Z) 1 e 1. This completes the proof. Proof of Theorem 3.1. By Lemma 3.5 diag(g)vkn T (3.14) min u u (1) 1 g [ ( ) ] 1 1/ e T 1 V kn [diag(g)] VkN T e1. g Let Γ diag(τ0n tr τtr 1n...τtr nn) diag(µ + ν cos ϑ 00 µ+ ν cos ϑ 01...µ+ ν cos ϑ nn ) where µ ωτ and ν ω as in (.5). Then (3.15) V kn [diag(g)] VkN T V kn Γ 1 ΩVkN T ( ) e T Γ 1 Ω ( e V k 1N Γ ) ΓVk 1N T ( e T Γ 1 Ωe e T ΩVk 1N T V k 1N Ωe V k 1N ΓΩV T k 1N )
11 MEINARDUS EXAMPLES FOR CONJUGATE GRADIENT METHOD 345 where e ( ) T.NoticeVk 1N T k 1NR 1 k 1 by (3.6) to get (3.16) V k 1N Ωe V k 1N ΩVk 1N T e 1 R T k 1 (Υ 10) (1:k 11:k 1) R 1 k 1 e 1 (3.17) R T k 1 (Υ 10) (1:k 11:k 1) e 1 V k 1N ΓΩVk 1N T R T k 1 S k 1N ΓΩS T k 1NR 1 k 1 R T k 1 (Υ µν) (1:k 11:k 1) R 1 k 1 in the notation introduced in Lemma 3.3. Recall 3 ( ) 1 ( B11 B 1 B B 1 C11 1 C11 1 B 1B 1 B 1 B 1C11 1 B 1 + B 1 B 1C11 1 B 1B 1 assuming all inversions exist where C 11 B 11 B 1 B 1 B 1. We have from (3.15) e T ( 1 VkN [diag(g)] VkN) T 1 (3.18) e1 1 [ ζ e T ΩV T ( k 1N Vk 1N ΓΩV T 1 1 k 1N) Vk 1N Ωe] where ζ e T Γ 1 Ωe. But from (3.16) and (3.17) ) (3.19) e T ΩV T k 1N ( Vk 1N ΓΩV T k 1N) 1 Vk 1N Ωe ] 1 e T 1 (Υ 10 ) (1:k 11:k 1) [(Υ µν ) (1:k 11:k 1) (Υ10 ) (1:k 11:k 1) e 1 [ ] 1 n e T 1 (Υ µν ) (1:k 11:k 1) e1 and for k N by Lemma 3.4 with m k 1andξ τ [ ] 1 (3.0) e T 1 (Υ µν ) (1:k 11:k 1) e1 n 1 e T [ ] 1 1 (µω+νh)(1:k 11:k 1) e1 1 [ ] 1 nω et 1 ( τω+h)(1:k 11:k 1) e1 1 nω γ k 1 τ 1(γ k 1 γ k γ k 1 + ) where γ ± τ ± τ 1. The conditions of Lemma 3.4 are fulfilled because τ > 1 and τω+h is diagonally dominant and thus nonsingular. Since ζ g we have by (3.14) and (3.18) (3.0) diag(g)vkn T (3.1) min u u (1) 1 g [ n 1 ωζ τ 1 γ k 1 γ k 1 We now compute ωζ τ 1. Let f(z) def n tr j0 (z τjn ). Then f(z) η(z τ0n)(z tr τnn) tr U n 1 (z/ω + τ) γ k γk 1 + ] 1/. 3 This is well known. See e.g. [6 pp ] [3 p. 3].
12 346 REN-CANG LI for some constant η where U n 1 (t) isthe(n 1)th Chebyshev polynomial of the second kind. This is because the zeros of U n 1 (z/ω + τ) are precisely τjn tr ω(τ jn τ) j 1...n 1. Then upon noticing τ0n tr β and τnn tr α ζ τjn tr 1 1 τ0n tr + τjn tr 1 ( 1 τnn tr α + 1 ) f (0) β f(0) j0 α + β αβ α + β αβ Recall [3 p. 37] (3.) They yield j0 (α + β) U n 1(τ)+α + βu n 1(τ)/ω αβ U n 1 (τ) ω U n 1(τ) U n 1 (τ). U n 1 (t) (t + t 1) n (t t 1) n t 1 U n 1(t) n (t + t 1) n +(t t 1) n t 1 t [ (t + t 1) n (t t 1) n] (t 1). t 1 U n 1 (τ) γn + γ n τ 1 U n 1 (τ) nγn + + γ n τ 1 τ(γn + γ n ) (τ 1) τ 1. Therefore upon noticing ω (α + β)/ andτ (β + α)/(β α) ζ α + β αβ + n γ n + γ+ n ω τ 1 γ n + τ (3.3) γn + ω τ 1 n γ n + γ+ n ω τ 1 γ n γ+ n ωζ τ 1 n γn + γ+ n (3.4) γ n. γn + Equation (3.1) and (3.4) imply diag(g)vkn T (3.5) min u [ 1 γn γ+ n γ k 1 ] 1/ γk 1 + u (1) 1 g γ n + γn + γ k 1 + γ+ k 1. Because τ ( +1)/( 1) γ k 1 and therefore diag(g)vkn T (3.6) min u [ u (1) 1 g ( 1) k 1 k 1 γ k 1 + ( 1) k 1 (k 1) 1 n n n + n k 1 k 1 (k 1) + (k 1) ] 1/. For k N n + 1 the right-hand side of (3.6) is [ n + n ] 1 aswasshown by Meinardus [18]. For any other k wehave diag(g)vkn T (3.7) min u ρ k 1 u (1) 1 g [ k 1 ] 1 + (k 1).
13 MEINARDUS EXAMPLES FOR CONJUGATE GRADIENT METHOD 347 where ρ k 1 def Right-hand side of (3.6) [ ] 1 k ( k 1 ( k 1 + (k 1) ) + (k 1) 4 )( + (k 1) n (k 1) ( ( n n ) (k 1) n + n ( )( (k 1) +1 [n (k 1)] +1 n +1 4 ( n + n + [n (k 1)] ) 1/ ) (k 1) ) ) 1/ 1/ which yields (.11). 4. Proof of Theorem. We first prove two general lemmas for the Vandermonde matrix V N V NN as defined in (3.) with arbitrary possibly complex nodes α j. Lemma 4.1. Assume one or more of 1) there are fewer than n distinct α j ) some α j 0and3) some g (j) 0occur. Then { 0 min diag(g)v N T u if all αj 0; u (1) 1 αj0 g (j) otherwise. Proof. If all α j 0 only cases 1) and 3) are possible. Let l be the number of distinct α j s exclude those corresponding to g (j) 0. Thenl<n.Bypermuting the rows of diag(g)vn T we may assume that α 1α...α l are distinct and for α j (j >l)eitheritisequaltosomeα i (i l) or corresponding g (j) 0. Setv C N whose v (j) is the coefficient of z j 1 in the polynomial φ(z) l j1 (z α j). Then v (1) l j1 ( α j) 0and min N T u diag(g)vn T (v/v (1) ) 0 u (1) 1 as expected. If some α j 0thensince diag(g)vn T u α j 0 g (j) always for any vector u with u (1) 1 it suffices to find a vector u to annihilate all other rows corresponding to α j 0. Such u can be constructed similarly to what we just did. The next lemma is essentially [17 Theorems.1 and 3.1] but stated differently. The proof below has a slightly different flavor. Lemma 4. ([17]). Let V N V NN be as defined in (3.) with all nodes α j (possibly complex) distinct and let f(z) N j1 (z α j).
14 348 REN-CANG LI (1) If all g (j) 0then diag(g)vn T (4.1) min u u (1) 1 g () (4.) max g min u (1) 1 N j1 diag(g)v T N u g ( ) f(0) N α j f g (j) (α j ) j1 N f(0) α j f (α j ) j1 1/ g (j) where the maximum is achieved if and only if for some constant µ>0 [ ] 1/ f(0) (4.3) g (j) µ α j f for 1 j N. (α j ) Proof. In Lemma 3.5 take Z diag(g)vn T. The assumptions make this Z nonsingular. Therefore [ ] min diag(g)v N T u e T 1 ( V N ΦVN T ) 1 e 1 u (1) 1 e T 1 (V N ΦVN) 1 e 1 (V 1 N e 1) Φ 1 (V 1 N e 1) where Φ [diag(g)] diag(g) and V N is the complex conjugate of V N. Let y V 1 N e 1 the first column of V 1 N which consists of the constant terms of the Lagrangian basis functions l j (z) z α i α i α j i j 1 j N since l j (α i )1fori j and 0 otherwise which means the jth row of V 1 N consists of the coefficients of l j (z). Therefore e T 1 ( V N ( ) f(0) N ΦVN T ) 1 e 1 α j1 j f g (j) (α j ) diag(g)vn T min N ( ) f(0) 1/ (4.4) u (1) 1 g α j1 j f g (j) g (j) (α j ) j1 1/ N f(0) α j f (α j ) 1/ j1 where it is an equality if and only if g (j) are given by (4.3). Remark 4.1. This lemma closely relates to a result of Greenbaum [9 (.) and Theorem 1] which in our notation essentially proved that if all nodes α j > 0 there exist k of α j s: α j1...α jk such that max g min u (1) 1 diag(g)v T kn u g max h min u (1) 1 diag(h)v T k u h.
15 MEINARDUS EXAMPLES FOR CONJUGATE GRADIENT METHOD 349 where V k is the k k Vandermonde matrix with nodes α ji (1 i k). Notice the difference in conditions: Lemma 4.3 only covers k N while this result of Greenbaum s is for all 1 k N but requires all α j > 0. Greenbaum [9 Theorem 1] also obtained an expression for the optimal h but a bit of more complicated than we get from applying Lemma 4.3. It is not clear how to find out the most relevant nodes α ji. Lemma 4.3. Let ω τ C (not necessarily associated with any interval [α β] as previously required) and let n N 1 and τjn tr as in (.7) with any given ω and τ. Suppose the Vandermonde matrix V N has nodes α j+1 τjn tr for 0 j n. (1) If all g (j) 0then (4.5) diag(g) VN T min u u (1) 1 g () 1 (τ0n tr g (1) + ) τ tr 0n τ tr nω N nnu n 1 (τ) g (j) N 1 j j1 1 (τ tr jn ) g (j+1) + 1 (τ tr nn) g (N) 1/ where U n 1 (t) is the (n 1)th Chebyshev polynomial of the second kind as in (3.). (4.6) max g min u (1) 1 diag(g) VN T u g nω τ tr 0n τ tr nnu n 1 (τ) 1 τ tr 0n + n 1 j1 1 τ tr jn + 1 τ tr nn 1 where the maximum is achieved if and only if for some µ>0 µ 1/ τjn tr (4.7) g (j+1) for j {0n} µ for 1 j n 1. / τ tr jn Proof. f(z) N j1 (z α j)admits f(z) η (z τ0n)(z tr τnn)u tr n 1 (z/ω + τ) where η 1 is the coefficient of z n 1 in U n 1 (z/ω + τ). We have f(0) ητ0nτ tr nn tr U n 1 (τ) f (τ0n) tr η (τ0n tr τnn) tr U n 1 (1) η (τ0n tr τnn) tr n η nω f (τnn) tr η (τnn tr τ0n) tr U n 1 ( 1) ( 1) n η (τnn tr τ0n) tr n ( 1) n η nω
16 350 REN-CANG LI and for 1 j n 1 f (τjn) tr η (τjn tr τ0n)(τ tr jn tr τnn)u tr n 1(τ jn )/ω η (τjn tr τ0n)(τ tr jn tr τnn)n/[ω(1 tr τjn)] ηnω. Therefore by Lemma 4. we have (4.5) and (4.6). Remark 4.. As a corollary to (4.6) and the error bound in (1.4) we deduce that the right-hand side of (4.6) is equal to T n (τ) [ n + n ] 1. Proof of Theorem.. Item (1) is always true for any given positive definite system Ax b. In fact let A QΛ Q be its eigendecomposition where Q is unitary and Λ as in the theorem since A is positive definite. Set g Λ 1/ Q b. Define g ( g (1) g ()... g (N) ) T R N. Then g Dg for some diagonal D with D (jj) 1. Finally A QΛQ and b QΛ 1/ g with Q QD still unitary. Next we notice that (4.8) r k A 1 min b Ax A 1 x K k min p k(λ)g p k (0)1 min N p k (λ j ) g(j) p k (0)1 where p k (z) denotes a polynomial of degree no more than k. If either inequality in item () is violated the effective condition number <(A) as far as CG is concerned and the error bound in (1.4) gives r k A 1 r 0 A 1 [ k + k j1 ] 1 [ < k + k ] 1 contradicting (.1). This proves item (). For item (3) we first claim that λ j for which g (j) > 0isin{τjk tr 0 j k}. Otherwise if there was a j 0 such that g (j0 ) > 0andλ j0 {τjk tr 0 j k} then T k (λ j0 /ω + τ) < 1 where ω and τ are given by (.5). Now take p k (z) q k (z) in (4.8) where q k (z) T k (z/ω + τ)/t k (τ) to get r k A 1 q k (λ j0 ) g(j 0 ) + q k (λ j ) g(j) j j 0 < T k (τ) 1 g (j 0 ) + j j 0 g (j) T k (τ) 1 r 0 A 1 contradicting (.1). This proves the claim. On the other hand since r k 0there are at least k + 1 distinct values in {λ j : g (j) > 0} and therefore {λ j : g (j) > 0} {τjk tr 0 j k}. Item(3)isproved. Item (3) says effectively A has k+1 distinct eigenvalues as far as CG is concerned and thus r k+1 0. This is item (4). Kaniel [13] gave a different proof of this fact.
17 MEINARDUS EXAMPLES FOR CONJUGATE GRADIENT METHOD 351 Define ĝ R k+1 by ĝ (l+1) g Jl. (4.8) gives r k A 1 min k p k (τjk tr p k (0)1 ) ĝ(j+1) min diag(ĝ)v k+1u T u (1) 1 j0 where V k+1 V k+1k+1 is the (k +1) (k + 1) Vandermonde matrix as defined in (3.) with nodes α j+1 τjk tr for 0 j k. The condition (.1) and the error bound in (1.4) imply that for ĝ min diag(ĝ)v k+1u T max min diag(h)v k+1u T. u (1) 1 h u (1) 1 Lemma 4.3 shows ĝ (l+1) g Jl must take the form of (.13). 5. Concluding remarks We have found a closed formula for the CG residuals for Meinardus examples. These residuals may deviate from the well-known error bounds in (1.4) by a factor no bigger than 1/ indicating the error bounds by (1.4) governing the CG convergence rate is very tight in general. Three key technical components that made our computations possible are as follows: (1) transforming CG residual computations as minimization problems involving rectangular Vandermonde matrices () the QR-like decomposition VN T ST NR 1 N and (3) the solution to min u(1) 1 Zu. It turns out that QR-like decompositions exist for quite a few Vandermonde matrices and the combination of the three technical components have been used in [14 15] for arriving at the asymptotically optimally conditioned real Vandermonde matrices analyzing the sharpness of existing error bounds for CG and the symmetric Lanczos method for eigenvalue problems. We completely characterized the extreme positive linear systems for which the kth CG residual achieves the error bound in (1.4). Roughly speaking as far as CG is concerned these extreme examples are nothing but one of Meinardus examples of order k + 1. As a consequence unless N there is no positive linear system whose kth CG residual achieves the error bound in (1.4) for all 1 k<n. Acknowledgment The author wishes to acknowledge the anonymous referee to draw his attention to Chapter II in [5] by H. Rutishauser. References [1] O. Axelsson Iterative solution methods Cambridge University Press New York MR (95f:65005) [] B. Beckermann On the numerical condition of polynomial bases: Estimates for the condition number of Vandermonde Krylov and Hankel matrices Habilitationsschrift Universität Hannover April 1996; see Beckermann.pdf. [3] P. Borwein and T. Erdélyi Polynomials and polynomial inequalities Graduate Texts in Mathematics vol. 161 Springer New York MR (97e:41001) [4] J. Demmel Applied numerical linear algebra SIAM Philadelphia MR (98m:65001)
18 35 REN-CANG LI [5] M. Engeli Th. Ginsburg H. Rutishauser and E. Stiefel Refined iterative methods for computation of the solution and the eigenvalues of self-adjoint boundary value problems Birkhäuser Verlag Basel/Stuttgart MR (6:318) [6] V. N. Faddeeva Computational methods of linear algebra Dover Publications New York 1959 Translated from the Russia by Curtis D. Benster. MR (53:4500) [7] G. H. Golub and C. F. Van Loan Matrix computations 3rd ed. Johns Hopkins University Press Baltimore Maryland MR (97g:65006) [8] I. S. Gradshteyn and I. M. Ryzhik Table of integrals series and products Academic Press New York 1980 Corrected and Enlarged Edition prepared by A. Jeffrey incorporated the fourth edition prepared by Yu. V. Geronimus and M. Yu. Tseytlin translaed from the Russian by Scripta Technica Inc. MR (001c:0000) [9] A. Greenbaum Comparision of splittings used with the conjugate gradient algorithm Numer. Math. 33 (1979) MR (80k:65035) [10] Iterative methods for solving linear systems SIAM Philadelphia MR (98j:6503) [11] M. R. Hestenes and E. Stiefel Methods of conjugate gradients for solving linear systems J. Res. Nat. Bur. Standards 49 (195) MR (15651a) [1] I. C. F. Ipsen Expressions and bounds for the GMRES residual BIT 40 (000) no MR (001g:6503) [13] S. Kaniel Estimates for some computational techniques in linear algebra Math. Comp. 0 (1966) no MR (38:934) [14] R.-C. Li Sharpness in rates of convergence for CG and symmetric Lanczos methods Technical Report Department of Mathematics University of Kentucky 005 Available at [15] Vandermonde matrices with Chebyshev nodes Technical Report Department of Mathematics University of Kentucky 005 Available at Lin. Alg. Appl. to appear. [16] J. Liesen M. Rozlozník and Z. Strakos Least squares residuals and minimal residual methods SIAMJ.Sci.Comput.3 (00) no MR (003a:65033) [17] J. Liesen and P. Tichý The worst-case GMRES for normal matrices BIT44 (004) no MR (005d:65046) [18] G. Meinardus Über eine Verallgemeinerung einer Ungleichung von L. V. Kantorowitsch Numer. Math. 5 (1963) MR (8:355) [19] Y. Saad Iterative methods for sparse linear systems nd ed. SIAM Philadelphia 003. MR (004h:6500) [0] G. L. G. Sleijpen and A. van der Sluis Further results on the convergence behavior of conjugate-gradients and Ritz values Linear Algebra Appl. 46 (1996) MR (97j:65067) [1] G. W. Stewart and J.-G. Sun Matrix perturbation theory Academic Press Boston MR (9a:65017) [] L. N. Trefethen and D. Bau III Numerical linear algebra SIAM Philadelphia MR (98k:6500) [3] K.ZhouJ.C.DoyleandK.GloverRobust and optimal control Prentice Hall Department of Mathematics University of Texas at Arlington P.O. Box Arlington Texas address: rcli@.uta.edu URL:
Vandermonde matrices with Chebyshev nodes
Available online at www.sciencedirect.com Linear Algebra and its Applications 48 (008) 803 83 www.elsevier.com/locate/laa Vandermonde matrices with Chebyshev nodes Ren-Cang Li Department of Mathematics,
More informationThe Rate of Convergence of GMRES on a Tridiagonal Toeplitz Linear System
Numerische Mathematik manuscript No. (will be inserted by the editor) The Rate of Convergence of GMRES on a Tridiagonal Toeplitz Linear System Ren-Cang Li 1, Wei Zhang 1 Department of Mathematics, University
More informationNumerische Mathematik
Numer. Math. (009) 11:67 93 DOI 10.1007/s0011-008-006- Numerische Mathematik The rate of convergence of GMRES on a tridiagonal Toeplitz linear system Ren-Cang Li Wei Zhang Received: 1 November 007 / Revised:
More informationABSTRACT OF DISSERTATION. Wei Zhang. The Graduate School. University of Kentucky
ABSTRACT OF DISSERTATION Wei Zhang The Graduate School University of Kentucky 2007 GMRES ON A TRIDIAGONAL TOEPLITZ LINEAR SYSTEM ABSTRACT OF DISSERTATION A dissertation submitted in partial fulfillment
More informationA Note on Eigenvalues of Perturbed Hermitian Matrices
A Note on Eigenvalues of Perturbed Hermitian Matrices Chi-Kwong Li Ren-Cang Li July 2004 Let ( H1 E A = E H 2 Abstract and à = ( H1 H 2 be Hermitian matrices with eigenvalues λ 1 λ k and λ 1 λ k, respectively.
More informationThe Lanczos and conjugate gradient algorithms
The Lanczos and conjugate gradient algorithms Gérard MEURANT October, 2008 1 The Lanczos algorithm 2 The Lanczos algorithm in finite precision 3 The nonsymmetric Lanczos algorithm 4 The Golub Kahan bidiagonalization
More informationKey words. conjugate gradients, normwise backward error, incremental norm estimation.
Proceedings of ALGORITMY 2016 pp. 323 332 ON ERROR ESTIMATION IN THE CONJUGATE GRADIENT METHOD: NORMWISE BACKWARD ERROR PETR TICHÝ Abstract. Using an idea of Duff and Vömel [BIT, 42 (2002), pp. 300 322
More informationON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH
ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH V. FABER, J. LIESEN, AND P. TICHÝ Abstract. Numerous algorithms in numerical linear algebra are based on the reduction of a given matrix
More informationOUTLINE 1. Introduction 1.1 Notation 1.2 Special matrices 2. Gaussian Elimination 2.1 Vector and matrix norms 2.2 Finite precision arithmetic 2.3 Fact
Computational Linear Algebra Course: (MATH: 6800, CSCI: 6800) Semester: Fall 1998 Instructors: { Joseph E. Flaherty, aherje@cs.rpi.edu { Franklin T. Luk, luk@cs.rpi.edu { Wesley Turner, turnerw@cs.rpi.edu
More informationMath Introduction to Numerical Analysis - Class Notes. Fernando Guevara Vasquez. Version Date: January 17, 2012.
Math 5620 - Introduction to Numerical Analysis - Class Notes Fernando Guevara Vasquez Version 1990. Date: January 17, 2012. 3 Contents 1. Disclaimer 4 Chapter 1. Iterative methods for solving linear systems
More informationWe first repeat some well known facts about condition numbers for normwise and componentwise perturbations. Consider the matrix
BIT 39(1), pp. 143 151, 1999 ILL-CONDITIONEDNESS NEEDS NOT BE COMPONENTWISE NEAR TO ILL-POSEDNESS FOR LEAST SQUARES PROBLEMS SIEGFRIED M. RUMP Abstract. The condition number of a problem measures the sensitivity
More informationMultiplicative Perturbation Analysis for QR Factorizations
Multiplicative Perturbation Analysis for QR Factorizations Xiao-Wen Chang Ren-Cang Li Technical Report 011-01 http://www.uta.edu/math/preprint/ Multiplicative Perturbation Analysis for QR Factorizations
More informationMULTIPLICATIVE PERTURBATION ANALYSIS FOR QR FACTORIZATIONS. Xiao-Wen Chang. Ren-Cang Li. (Communicated by Wenyu Sun)
NUMERICAL ALGEBRA, doi:10.3934/naco.011.1.301 CONTROL AND OPTIMIZATION Volume 1, Number, June 011 pp. 301 316 MULTIPLICATIVE PERTURBATION ANALYSIS FOR QR FACTORIZATIONS Xiao-Wen Chang School of Computer
More informationIntroduction to Iterative Solvers of Linear Systems
Introduction to Iterative Solvers of Linear Systems SFB Training Event January 2012 Prof. Dr. Andreas Frommer Typeset by Lukas Krämer, Simon-Wolfgang Mages and Rudolf Rödl 1 Classes of Matrices and their
More informationonto the column space = I P Zj. Embed PZ j
EIGENVALUES OF AN ALIGNMENT MATRIX IN NONLINEAR MANIFOLD LEARNING CHI-KWONG LI, REN-CANG LI, AND QIANG YE Abstract Alignment algorithm is an effective method recently proposed for nonlinear manifold learning
More informationM.A. Botchev. September 5, 2014
Rome-Moscow school of Matrix Methods and Applied Linear Algebra 2014 A short introduction to Krylov subspaces for linear systems, matrix functions and inexact Newton methods. Plan and exercises. M.A. Botchev
More information6.4 Krylov Subspaces and Conjugate Gradients
6.4 Krylov Subspaces and Conjugate Gradients Our original equation is Ax = b. The preconditioned equation is P Ax = P b. When we write P, we never intend that an inverse will be explicitly computed. P
More informationPreliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012
Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.
More informationOn the Preconditioning of the Block Tridiagonal Linear System of Equations
On the Preconditioning of the Block Tridiagonal Linear System of Equations Davod Khojasteh Salkuyeh Department of Mathematics, University of Mohaghegh Ardabili, PO Box 179, Ardabil, Iran E-mail: khojaste@umaacir
More informationLarge-scale eigenvalue problems
ELE 538B: Mathematics of High-Dimensional Data Large-scale eigenvalue problems Yuxin Chen Princeton University, Fall 208 Outline Power method Lanczos algorithm Eigenvalue problems 4-2 Eigendecomposition
More informationarxiv: v1 [math.na] 1 Sep 2018
On the perturbation of an L -orthogonal projection Xuefeng Xu arxiv:18090000v1 [mathna] 1 Sep 018 September 5 018 Abstract The L -orthogonal projection is an important mathematical tool in scientific computing
More informationFoundations of Matrix Analysis
1 Foundations of Matrix Analysis In this chapter we recall the basic elements of linear algebra which will be employed in the remainder of the text For most of the proofs as well as for the details, the
More informationANY FINITE CONVERGENCE CURVE IS POSSIBLE IN THE INITIAL ITERATIONS OF RESTARTED FOM
Electronic Transactions on Numerical Analysis. Volume 45, pp. 133 145, 2016. Copyright c 2016,. ISSN 1068 9613. ETNA ANY FINITE CONVERGENCE CURVE IS POSSIBLE IN THE INITIAL ITERATIONS OF RESTARTED FOM
More informationOn the influence of eigenvalues on Bi-CG residual norms
On the influence of eigenvalues on Bi-CG residual norms Jurjen Duintjer Tebbens Institute of Computer Science Academy of Sciences of the Czech Republic duintjertebbens@cs.cas.cz Gérard Meurant 30, rue
More informationarxiv: v2 [math.na] 27 Dec 2016
An algorithm for constructing Equiangular vectors Azim rivaz a,, Danial Sadeghi a a Department of Mathematics, Shahid Bahonar University of Kerman, Kerman 76169-14111, IRAN arxiv:1412.7552v2 [math.na]
More informationDS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.
DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1
More informationA fast randomized algorithm for overdetermined linear least-squares regression
A fast randomized algorithm for overdetermined linear least-squares regression Vladimir Rokhlin and Mark Tygert Technical Report YALEU/DCS/TR-1403 April 28, 2008 Abstract We introduce a randomized algorithm
More informationDepartment of Mathematics Technical Report May 2000 ABSTRACT. for any matrix norm that is reduced by a pinching. In addition to known
University of Kentucky Lexington Department of Mathematics Technical Report 2000-23 Pinchings and Norms of Scaled Triangular Matrices 1 Rajendra Bhatia 2 William Kahan 3 Ren-Cang Li 4 May 2000 ABSTRACT
More informationOn the Superlinear Convergence of MINRES. Valeria Simoncini and Daniel B. Szyld. Report January 2012
On the Superlinear Convergence of MINRES Valeria Simoncini and Daniel B. Szyld Report 12-01-11 January 2012 This report is available in the World Wide Web at http://www.math.temple.edu/~szyld 0 Chapter
More informationComparative Analysis of Mesh Generators and MIC(0) Preconditioning of FEM Elasticity Systems
Comparative Analysis of Mesh Generators and MIC(0) Preconditioning of FEM Elasticity Systems Nikola Kosturski and Svetozar Margenov Institute for Parallel Processing, Bulgarian Academy of Sciences Abstract.
More informationON KRONECKER PRODUCTS OF CHARACTERS OF THE SYMMETRIC GROUPS WITH FEW COMPONENTS
ON KRONECKER PRODUCTS OF CHARACTERS OF THE SYMMETRIC GROUPS WITH FEW COMPONENTS C. BESSENRODT AND S. VAN WILLIGENBURG Abstract. Confirming a conjecture made by Bessenrodt and Kleshchev in 1999, we classify
More informationJordan Journal of Mathematics and Statistics (JJMS) 5(3), 2012, pp A NEW ITERATIVE METHOD FOR SOLVING LINEAR SYSTEMS OF EQUATIONS
Jordan Journal of Mathematics and Statistics JJMS) 53), 2012, pp.169-184 A NEW ITERATIVE METHOD FOR SOLVING LINEAR SYSTEMS OF EQUATIONS ADEL H. AL-RABTAH Abstract. The Jacobi and Gauss-Seidel iterative
More informationSensitivity of Gauss-Christoffel quadrature and sensitivity of Jacobi matrices to small changes of spectral data
Sensitivity of Gauss-Christoffel quadrature and sensitivity of Jacobi matrices to small changes of spectral data Zdeněk Strakoš Academy of Sciences and Charles University, Prague http://www.cs.cas.cz/
More information4.6 Iterative Solvers for Linear Systems
4.6 Iterative Solvers for Linear Systems Why use iterative methods? Virtually all direct methods for solving Ax = b require O(n 3 ) floating point operations. In practical applications the matrix A often
More informationTikhonov Regularization of Large Symmetric Problems
NUMERICAL LINEAR ALGEBRA WITH APPLICATIONS Numer. Linear Algebra Appl. 2000; 00:1 11 [Version: 2000/03/22 v1.0] Tihonov Regularization of Large Symmetric Problems D. Calvetti 1, L. Reichel 2 and A. Shuibi
More informationIntroduction to Numerical Linear Algebra II
Introduction to Numerical Linear Algebra II Petros Drineas These slides were prepared by Ilse Ipsen for the 2015 Gene Golub SIAM Summer School on RandNLA 1 / 49 Overview We will cover this material in
More informationContribution of Wo¹niakowski, Strako²,... The conjugate gradient method in nite precision computa
Contribution of Wo¹niakowski, Strako²,... The conjugate gradient method in nite precision computations ªaw University of Technology Institute of Mathematics and Computer Science Warsaw, October 7, 2006
More informationOn prescribing Ritz values and GMRES residual norms generated by Arnoldi processes
On prescribing Ritz values and GMRES residual norms generated by Arnoldi processes Jurjen Duintjer Tebbens Institute of Computer Science Academy of Sciences of the Czech Republic joint work with Gérard
More informationAM 205: lecture 8. Last time: Cholesky factorization, QR factorization Today: how to compute the QR factorization, the Singular Value Decomposition
AM 205: lecture 8 Last time: Cholesky factorization, QR factorization Today: how to compute the QR factorization, the Singular Value Decomposition QR Factorization A matrix A R m n, m n, can be factorized
More informationELEMENTARY LINEAR ALGEBRA
ELEMENTARY LINEAR ALGEBRA K R MATTHEWS DEPARTMENT OF MATHEMATICS UNIVERSITY OF QUEENSLAND First Printing, 99 Chapter LINEAR EQUATIONS Introduction to linear equations A linear equation in n unknowns x,
More informationJae Heon Yun and Yu Du Han
Bull. Korean Math. Soc. 39 (2002), No. 3, pp. 495 509 MODIFIED INCOMPLETE CHOLESKY FACTORIZATION PRECONDITIONERS FOR A SYMMETRIC POSITIVE DEFINITE MATRIX Jae Heon Yun and Yu Du Han Abstract. We propose
More informationAlgebra C Numerical Linear Algebra Sample Exam Problems
Algebra C Numerical Linear Algebra Sample Exam Problems Notation. Denote by V a finite-dimensional Hilbert space with inner product (, ) and corresponding norm. The abbreviation SPD is used for symmetric
More informationInterval solutions for interval algebraic equations
Mathematics and Computers in Simulation 66 (2004) 207 217 Interval solutions for interval algebraic equations B.T. Polyak, S.A. Nazin Institute of Control Sciences, Russian Academy of Sciences, 65 Profsoyuznaya
More informationInterlacing Inequalities for Totally Nonnegative Matrices
Interlacing Inequalities for Totally Nonnegative Matrices Chi-Kwong Li and Roy Mathias October 26, 2004 Dedicated to Professor T. Ando on the occasion of his 70th birthday. Abstract Suppose λ 1 λ n 0 are
More informationRITZ VALUE BOUNDS THAT EXPLOIT QUASI-SPARSITY
RITZ VALUE BOUNDS THAT EXPLOIT QUASI-SPARSITY ILSE C.F. IPSEN Abstract. Absolute and relative perturbation bounds for Ritz values of complex square matrices are presented. The bounds exploit quasi-sparsity
More informationOPTIMAL SCALING FOR P -NORMS AND COMPONENTWISE DISTANCE TO SINGULARITY
published in IMA Journal of Numerical Analysis (IMAJNA), Vol. 23, 1-9, 23. OPTIMAL SCALING FOR P -NORMS AND COMPONENTWISE DISTANCE TO SINGULARITY SIEGFRIED M. RUMP Abstract. In this note we give lower
More information1. Introduction. Let µ(t) be a distribution function with infinitely many points of increase in the interval [ π, π] and let
SENSITIVITY ANALYSIS OR SZEGŐ POLYNOMIALS SUN-MI KIM AND LOTHAR REICHEL Abstract Szegő polynomials are orthogonal with respect to an inner product on the unit circle Numerical methods for weighted least-squares
More information9.1 Preconditioned Krylov Subspace Methods
Chapter 9 PRECONDITIONING 9.1 Preconditioned Krylov Subspace Methods 9.2 Preconditioned Conjugate Gradient 9.3 Preconditioned Generalized Minimal Residual 9.4 Relaxation Method Preconditioners 9.5 Incomplete
More informationReduced Synchronization Overhead on. December 3, Abstract. The standard formulation of the conjugate gradient algorithm involves
Lapack Working Note 56 Conjugate Gradient Algorithms with Reduced Synchronization Overhead on Distributed Memory Multiprocessors E. F. D'Azevedo y, V.L. Eijkhout z, C. H. Romine y December 3, 1999 Abstract
More informationYimin Wei a,b,,1, Xiezhang Li c,2, Fanbin Bu d, Fuzhen Zhang e. Abstract
Linear Algebra and its Applications 49 (006) 765 77 wwwelseviercom/locate/laa Relative perturbation bounds for the eigenvalues of diagonalizable and singular matrices Application of perturbation theory
More informationSolution of Linear Equations
Solution of Linear Equations (Com S 477/577 Notes) Yan-Bin Jia Sep 7, 07 We have discussed general methods for solving arbitrary equations, and looked at the special class of polynomial equations A subclass
More informationDEN: Linear algebra numerical view (GEM: Gauss elimination method for reducing a full rank matrix to upper-triangular
form) Given: matrix C = (c i,j ) n,m i,j=1 ODE and num math: Linear algebra (N) [lectures] c phabala 2016 DEN: Linear algebra numerical view (GEM: Gauss elimination method for reducing a full rank matrix
More informationNumerical Analysis: Solving Systems of Linear Equations
Numerical Analysis: Solving Systems of Linear Equations Mirko Navara http://cmpfelkcvutcz/ navara/ Center for Machine Perception, Department of Cybernetics, FEE, CTU Karlovo náměstí, building G, office
More informationb jσ(j), Keywords: Decomposable numerical range, principal character AMS Subject Classification: 15A60
On the Hu-Hurley-Tam Conjecture Concerning The Generalized Numerical Range Che-Man Cheng Faculty of Science and Technology, University of Macau, Macau. E-mail: fstcmc@umac.mo and Chi-Kwong Li Department
More informationCS137 Introduction to Scientific Computing Winter Quarter 2004 Solutions to Homework #3
CS137 Introduction to Scientific Computing Winter Quarter 2004 Solutions to Homework #3 Felix Kwok February 27, 2004 Written Problems 1. (Heath E3.10) Let B be an n n matrix, and assume that B is both
More information1 Conjugate gradients
Notes for 2016-11-18 1 Conjugate gradients We now turn to the method of conjugate gradients (CG), perhaps the best known of the Krylov subspace solvers. The CG iteration can be characterized as the iteration
More informationError estimates for the ESPRIT algorithm
Error estimates for the ESPRIT algorithm Daniel Potts Manfred Tasche Let z j := e f j j = 1,..., M) with f j [ ϕ, 0] + i [ π, π) and small ϕ 0 be distinct nodes. With complex coefficients c j 0, we consider
More informationLecture notes: Applied linear algebra Part 2. Version 1
Lecture notes: Applied linear algebra Part 2. Version 1 Michael Karow Berlin University of Technology karow@math.tu-berlin.de October 2, 2008 First, some exercises: xercise 0.1 (2 Points) Another least
More informationBounding the Spectrum of Large Hermitian Matrices
Bounding the Spectrum of Large Hermitian Matrices Ren-Cang Li Yunkai Zhou Technical Report 2009-07 http://www.uta.edu/math/preprint/ Bounding the Spectrum of Large Hermitian Matrices Ren-Cang Li and Yunkai
More informationGradient Method Based on Roots of A
Journal of Scientific Computing, Vol. 15, No. 4, 2000 Solving Ax Using a Modified Conjugate Gradient Method Based on Roots of A Paul F. Fischer 1 and Sigal Gottlieb 2 Received January 23, 2001; accepted
More informationHere is an example of a block diagonal matrix with Jordan Blocks on the diagonal: J
Class Notes 4: THE SPECTRAL RADIUS, NORM CONVERGENCE AND SOR. Math 639d Due Date: Feb. 7 (updated: February 5, 2018) In the first part of this week s reading, we will prove Theorem 2 of the previous class.
More informationMath 577 Assignment 7
Math 577 Assignment 7 Thanks for Yu Cao 1. Solution. The linear system being solved is Ax = 0, where A is a (n 1 (n 1 matrix such that 2 1 1 2 1 A =......... 1 2 1 1 2 and x = (U 1, U 2,, U n 1. By the
More informationPositive entries of stable matrices
Positive entries of stable matrices Shmuel Friedland Department of Mathematics, Statistics and Computer Science, University of Illinois at Chicago Chicago, Illinois 60607-7045, USA Daniel Hershkowitz,
More informationPrinciples and Analysis of Krylov Subspace Methods
Principles and Analysis of Krylov Subspace Methods Zdeněk Strakoš Institute of Computer Science, Academy of Sciences, Prague www.cs.cas.cz/~strakos Ostrava, February 2005 1 With special thanks to C.C.
More informationThroughout these notes we assume V, W are finite dimensional inner product spaces over C.
Math 342 - Linear Algebra II Notes Throughout these notes we assume V, W are finite dimensional inner product spaces over C 1 Upper Triangular Representation Proposition: Let T L(V ) There exists an orthonormal
More informationCourse Notes: Week 1
Course Notes: Week 1 Math 270C: Applied Numerical Linear Algebra 1 Lecture 1: Introduction (3/28/11) We will focus on iterative methods for solving linear systems of equations (and some discussion of eigenvalues
More informationNumerical Linear Algebra
Numerical Linear Algebra The two principal problems in linear algebra are: Linear system Given an n n matrix A and an n-vector b, determine x IR n such that A x = b Eigenvalue problem Given an n n matrix
More informationFirst, we review some important facts on the location of eigenvalues of matrices.
BLOCK NORMAL MATRICES AND GERSHGORIN-TYPE DISCS JAKUB KIERZKOWSKI AND ALICJA SMOKTUNOWICZ Abstract The block analogues of the theorems on inclusion regions for the eigenvalues of normal matrices are given
More informationACI-matrices all of whose completions have the same rank
ACI-matrices all of whose completions have the same rank Zejun Huang, Xingzhi Zhan Department of Mathematics East China Normal University Shanghai 200241, China Abstract We characterize the ACI-matrices
More informationFundamentals of Engineering Analysis (650163)
Philadelphia University Faculty of Engineering Communications and Electronics Engineering Fundamentals of Engineering Analysis (6563) Part Dr. Omar R Daoud Matrices: Introduction DEFINITION A matrix is
More informationComputing the pth Roots of a Matrix. with Repeated Eigenvalues
Applied Mathematical Sciences, Vol. 5, 2011, no. 53, 2645-2661 Computing the pth Roots of a Matrix with Repeated Eigenvalues Amir Sadeghi 1, Ahmad Izani Md. Ismail and Azhana Ahmad School of Mathematical
More informationA Note on Simple Nonzero Finite Generalized Singular Values
A Note on Simple Nonzero Finite Generalized Singular Values Wei Ma Zheng-Jian Bai December 21 212 Abstract In this paper we study the sensitivity and second order perturbation expansions of simple nonzero
More informationLecture 9: Krylov Subspace Methods. 2 Derivation of the Conjugate Gradient Algorithm
CS 622 Data-Sparse Matrix Computations September 19, 217 Lecture 9: Krylov Subspace Methods Lecturer: Anil Damle Scribes: David Eriksson, Marc Aurele Gilles, Ariah Klages-Mundt, Sophia Novitzky 1 Introduction
More informationIntroduction to Scientific Computing
(Lecture 5: Linear system of equations / Matrix Splitting) Bojana Rosić, Thilo Moshagen Institute of Scientific Computing Motivation Let us resolve the problem scheme by using Kirchhoff s laws: the algebraic
More informationChebyshev semi-iteration in Preconditioning
Report no. 08/14 Chebyshev semi-iteration in Preconditioning Andrew J. Wathen Oxford University Computing Laboratory Tyrone Rees Oxford University Computing Laboratory Dedicated to Victor Pereyra on his
More informationA matrix over a field F is a rectangular array of elements from F. The symbol
Chapter MATRICES Matrix arithmetic A matrix over a field F is a rectangular array of elements from F The symbol M m n (F ) denotes the collection of all m n matrices over F Matrices will usually be denoted
More informationMath 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.
Math 350 Fall 2011 Notes about inner product spaces In this notes we state and prove some important properties of inner product spaces. First, recall the dot product on R n : if x, y R n, say x = (x 1,...,
More informationComputation of eigenvalues and singular values Recall that your solutions to these questions will not be collected or evaluated.
Math 504, Homework 5 Computation of eigenvalues and singular values Recall that your solutions to these questions will not be collected or evaluated 1 Find the eigenvalues and the associated eigenspaces
More informationTHE PERTURBATION BOUND FOR THE SPECTRAL RADIUS OF A NON-NEGATIVE TENSOR
THE PERTURBATION BOUND FOR THE SPECTRAL RADIUS OF A NON-NEGATIVE TENSOR WEN LI AND MICHAEL K. NG Abstract. In this paper, we study the perturbation bound for the spectral radius of an m th - order n-dimensional
More informationthe Unitary Polar Factor æ Ren-Cang Li P.O. Box 2008, Bldg 6012
Relative Perturbation Bounds for the Unitary Polar actor Ren-Cang Li Mathematical Science Section Oak Ridge National Laboratory P.O. Box 2008, Bldg 602 Oak Ridge, TN 3783-6367 èli@msr.epm.ornl.govè LAPACK
More information4.8 Arnoldi Iteration, Krylov Subspaces and GMRES
48 Arnoldi Iteration, Krylov Subspaces and GMRES We start with the problem of using a similarity transformation to convert an n n matrix A to upper Hessenberg form H, ie, A = QHQ, (30) with an appropriate
More informationConjugate Gradient (CG) Method
Conjugate Gradient (CG) Method by K. Ozawa 1 Introduction In the series of this lecture, I will introduce the conjugate gradient method, which solves efficiently large scale sparse linear simultaneous
More informationTwo Results About The Matrix Exponential
Two Results About The Matrix Exponential Hongguo Xu Abstract Two results about the matrix exponential are given. One is to characterize the matrices A which satisfy e A e AH = e AH e A, another is about
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 18 Outline
More informationConjugate Gradient Method
Conjugate Gradient Method direct and indirect methods positive definite linear systems Krylov sequence spectral analysis of Krylov sequence preconditioning Prof. S. Boyd, EE364b, Stanford University Three
More informationOn the Computations of Eigenvalues of the Fourth-order Sturm Liouville Problems
Int. J. Open Problems Compt. Math., Vol. 4, No. 3, September 2011 ISSN 1998-6262; Copyright c ICSRS Publication, 2011 www.i-csrs.org On the Computations of Eigenvalues of the Fourth-order Sturm Liouville
More informationExponentials of Symmetric Matrices through Tridiagonal Reductions
Exponentials of Symmetric Matrices through Tridiagonal Reductions Ya Yan Lu Department of Mathematics City University of Hong Kong Kowloon, Hong Kong Abstract A simple and efficient numerical algorithm
More informationLanczos tridigonalization and Golub - Kahan bidiagonalization: Ideas, connections and impact
Lanczos tridigonalization and Golub - Kahan bidiagonalization: Ideas, connections and impact Zdeněk Strakoš Academy of Sciences and Charles University, Prague http://www.cs.cas.cz/ strakos Hong Kong, February
More informationON THE HÖLDER CONTINUITY OF MATRIX FUNCTIONS FOR NORMAL MATRICES
Volume 10 (2009), Issue 4, Article 91, 5 pp. ON THE HÖLDER CONTINUITY O MATRIX UNCTIONS OR NORMAL MATRICES THOMAS P. WIHLER MATHEMATICS INSTITUTE UNIVERSITY O BERN SIDLERSTRASSE 5, CH-3012 BERN SWITZERLAND.
More informationOctober 25, 2013 INNER PRODUCT SPACES
October 25, 2013 INNER PRODUCT SPACES RODICA D. COSTIN Contents 1. Inner product 2 1.1. Inner product 2 1.2. Inner product spaces 4 2. Orthogonal bases 5 2.1. Existence of an orthogonal basis 7 2.2. Orthogonal
More informationIterative methods for Linear System
Iterative methods for Linear System JASS 2009 Student: Rishi Patil Advisor: Prof. Thomas Huckle Outline Basics: Matrices and their properties Eigenvalues, Condition Number Iterative Methods Direct and
More informationWeaker assumptions for convergence of extended block Kaczmarz and Jacobi projection algorithms
DOI: 10.1515/auom-2017-0004 An. Şt. Univ. Ovidius Constanţa Vol. 25(1),2017, 49 60 Weaker assumptions for convergence of extended block Kaczmarz and Jacobi projection algorithms Doina Carp, Ioana Pomparău,
More informationIterative Methods for Solving A x = b
Iterative Methods for Solving A x = b A good (free) online source for iterative methods for solving A x = b is given in the description of a set of iterative solvers called templates found at netlib: http
More informationLecture notes: Applied linear algebra Part 1. Version 2
Lecture notes: Applied linear algebra Part 1. Version 2 Michael Karow Berlin University of Technology karow@math.tu-berlin.de October 2, 2008 1 Notation, basic notions and facts 1.1 Subspaces, range and
More informationMath 108b: Notes on the Spectral Theorem
Math 108b: Notes on the Spectral Theorem From section 6.3, we know that every linear operator T on a finite dimensional inner product space V has an adjoint. (T is defined as the unique linear operator
More informationClassical iterative methods for linear systems
Classical iterative methods for linear systems Ed Bueler MATH 615 Numerical Analysis of Differential Equations 27 February 1 March, 2017 Ed Bueler (MATH 615 NADEs) Classical iterative methods for linear
More informationIDR(s) Master s thesis Goushani Kisoensingh. Supervisor: Gerard L.G. Sleijpen Department of Mathematics Universiteit Utrecht
IDR(s) Master s thesis Goushani Kisoensingh Supervisor: Gerard L.G. Sleijpen Department of Mathematics Universiteit Utrecht Contents 1 Introduction 2 2 The background of Bi-CGSTAB 3 3 IDR(s) 4 3.1 IDR.............................................
More informationApplied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic
Applied Mathematics 205 Unit V: Eigenvalue Problems Lecturer: Dr. David Knezevic Unit V: Eigenvalue Problems Chapter V.4: Krylov Subspace Methods 2 / 51 Krylov Subspace Methods In this chapter we give
More informationFEM and sparse linear system solving
FEM & sparse linear system solving, Lecture 9, Nov 19, 2017 1/36 Lecture 9, Nov 17, 2017: Krylov space methods http://people.inf.ethz.ch/arbenz/fem17 Peter Arbenz Computer Science Department, ETH Zürich
More informationETNA Kent State University
Electronic Transactions on Numerical Analysis. Volume 41, pp. 159-166, 2014. Copyright 2014,. ISSN 1068-9613. ETNA MAX-MIN AND MIN-MAX APPROXIMATION PROBLEMS FOR NORMAL MATRICES REVISITED JÖRG LIESEN AND
More information