1.1. Contributions. The most important feature of problem (1.1) is that A is

Size: px
Start display at page:

Download "1.1. Contributions. The most important feature of problem (1.1) is that A is"

Transcription

1 FAST AND STABLE ALGORITHMS FOR BANDED PLUS SEMISEPARABLE SYSTEMS OF LINEAR EQUATIONS S. HANDRASEKARAN AND M. GU y Abstract. We present fast and numerically stable algorithms for the solution of linear systems of equations here the coecient matrix can be ritten in the form of a banded plus semiseparable matrix. Such matrices include banded matrices, semiseparable matrices, and block-diagonal plus semiseparable matrices as special cases. Our algorithms are based on novel matrix factorizations developed specically for matrices ith such structures. We also present interesting numerical results ith these algorithms. Key ords. banded matrix, semiseparable matrix, fast algorithms, stable algorithms. AMS subject classications. 5A9, 5A3, 65F5, 65L, 65R.. Introduction. In this paper e consider fast and numerically stable solutions of the n-by-n linear system of equations (.) A x = b; here A is the sum of a banded matrix and a semiseparable matrix. This class of matrices also include block-diagonal + semiseparable matrices. Linear systems of such forms appear in the numerical solution of boundray-value problems for ordinary dierential equations and certain integral equations (see Greengard and Rokhlin [6], Starr [9]) and Lee and Greengard [7]. Some related ork on (.) can also be found in Eidelman and Gohberg [3, 4]... ontributions. The most important feature of problem (.) is that A is a dense but highly structured matrix. Although direct methods have been developed for ecient and numerically stable LU and QR factorizations of banded matrices (see Demmel [, h. ]), such methods do not currently exist for semiseparable matrices, let alone banded+semiseparable matrices. The main diculty is that LU and QR factorizations have tremendous diculties in mentaining the banded+semiseparable structure, and consequently requre O(n 3 ) ops to compute such factorizations of A in (.). Although iterative methods, such as those based on the Lanczos or Arnoldi procedures, can take advantage of the structure in A and can be used to solve (.), their convergence rate is highly problem-dependent and can be very slo ithout eective preconditioning (see Saad [8]). In this paper, e present a number of fast and numerically stable direct methods for solving (.). Our methods are based on some ne matrix factorizations e developed specically for banded+semiseparable matrices. We also present interesting results from our numerical experiments ith these methods in matlab... Notation. To describe the problem precisely, e rst introduce some notation. Reminiscent of matlab notation, e use triu(a; k) to denote the matrix hich is identical to the matrix A on and above the k-th diagonal. k = is the main diagonal, k > is above the main diagonal and k < is belo the main diagonal. Similarly, Department of Electrical and omputer Engineering, University of alifornia, Santa Barbara. address: shivece.ucsb.edu ydepartment of Mathematics, University of alifornia, Los Angeles. address: mgumath.ucla.edu A op is a oating point operation such as +;?; and.

2 S. handrasekaran and M. Gu tril(a; k) denotes the matrix hich is identical to the matrix A on and belo the k-th diagonal. For example, triu A = tril?a = A : As a banded+semiseparable matrix, the matrix A in (.) can be ritten as (.) A = D + triu? u v T ; b u + + tril? p q T ;?b l? here D is an n n banded matrix, ith b u non-zero diagonals strictly above the main diagonal and b l non-zero diagonals strictly belo the main diagonal; u and v are n r u matrices; and p and q are n r l matrices. When b u = b l =, D is a diagonal matrix, and A is a diagonal+semiseparable matrix. When r u = r l =, A = D is a banded matrix. We are interested in the numerical solution of the linear system (.). The rest of this paper provides a set of numerically backard stable algorithms that take approximately O? n(b u + b l + r u + r l ) ops to solve (.) as opposed to O(n 3 ) by using traditional methods involving LU and QR factorizations. The exact constant hidden in the O() notation varies among our algorithms. Throughout this paper, e ill take the liberty to use I to denote an identity matrix of any dimension. The rest of this paper is organized as follos. In x e illustrate the basic ideas behind our algorithms through a simple example. In x3 e describe the algorithms in some detail. In x4 e present our numerical results ith these algorithms.. Basic Idea. In this section e give a description of the basic idea in the simple case hen D is a diagonal matrix, and u, v, p, and q have only one column. The idea is to compute a to-sided decomposition of the form (.) A = W L H ; here W and H can be ritten as the product of n elementary matrices, and L is a loer triangular matrix. The three matrices W, L, H themselves are never explicitly formed, but inverted eciently on-line as the algorithm proceeds. In this section, e ill choose the matrices W and H to be the products of elementary Givens rotations. When e discuss our algorithms in full detail in x3, e ill allo ourselves the additional freedom of choosing W and H to be products of elementary Householder reections or Gaussian elimination matrices ith column and/or ro permutations. More specically consider the 5 5 case: D u v u v u v 3 u v 4 p q D u v u v 3 u v 4 A = B p q p q D u v 3 u v 4 A : p 3 q p 3 q p 3 q D 3 u 3 v 4 p 4 q p 4 q p 4 q p 4 q 3 D 4 For future convenience e also assume that the right-hand side is of the form b b b = B b A?? B p b 3 p 3 b 4 p 4

3 Fast and Stable Solver for Semiseparable Systems 3 here? =. (Of course, the second term on the right-hand side has no eect at this stage, but it ill capture the general form of the recursion as e proceed.) No suppose that W is a Givens rotation such that W u p u A = u + u A (.) bu for any vector. Then if e apply W from the left to A e obtain ba A b ba A b bu v bu v 3 bu v 4 ba = W A = B p q p q D u v 3 u v 4 p 3 q p 3 q p 3 q D 3 u 3 v 4 A : p 4 q p 4 q p 4 q p 4 q 3 D 4 We also apply W to b to obtain b b b = W b = W BB b A?? B b 3 b 4 p p 3 p 4 AA = B b b b b b b 3 b 4 A?? B here e have deliberately ritten the formula in such a ay that it ould be correct even if? had not been zero. We next choose a Givens rotation, H, such that (.3) H T ba ba A = q ba + b A A ea p p 3 p 4 for any vector. Further let H T q = q eq eq and H T ba ea = : ba ea Then No let (.4) ea ea = W A H = A b ea A e bu v bu v 3 bu v 4 H = B p eq p eq D u v 3 u v 4 p 3 eq p 3 eq p 3 q D 3 u 3 v 4 A : p 4 eq p 4 eq p 4 q p 4 q 3 D 4 H? x = H? B x x x x 3 x 4 e ex A B x A ex: x 3 x 4

4 4 S. handrasekaran and M. Gu Then it follos from W A H H? x = e A ex = W b = b b, that Also let e = b b ea : =? + e eq ; e b = b b? e e A and e b = b? p : To reach this stage e needed to compute all the \tilde" and \hat" quantities except ex. They can be computed in constant time, independent of the size of the matrix A. No e can proceed to solve the smaller 4 4 system of equations ea bu v bu v 3 bu v 4 ex e b B p eq D u v 3 u v 4 B x e p 3 eq p 3 q D 3 u 3 v 4 A A = B b x 3 b 3 A? B p 3 p 4 eq p 4 q p 4 q 3 D 4 x 4 b 4 p 4 hich is exactly like the original 5 5 system of equations in form. That is, the coecient matrix is a diagonal matrix plus a semiseparable matrix, and the righthand side is also of the requisite form. Hence e can use this recursion 3 times until the problem size becomes, at hich point e solve the sysmtem directly. Let the 5 numbers obtained by this recursion, e, e, e, e 3, and e 4, be the components of the ve dimensional vector. Then it follos from equation (.4) that the actual solution x to the original 5 5 system of equations is given by (.5) x = H H H ; here the H i 's are the successive Givens transforms computed from the recursion (.3), but set up in such a ay that they only aect ros i and i +. Since there are only 3 of these transforms e retain the linear time complexity of the algorithm. The backard stability of the algorithm follos from the fact that e only use orthogonal transforms and a single forard-substitution. Our factorization is similar in form to the ULV factorization proposed by Steart []. Hoever, the ULV factorization of Steart is developed primarily to reveal potential numerical rank-deciency in a general matrix and can take O(n 3 ) ops to compute; hereas our factorization is designed primarily to take advantage of the banded+semiseparable structure for large savings in computational cost. There are to places in the recursion here elimination is necessary. In equation (.) e chose W to be a Givens rotation to eliminate u, and in equation (.3) e chose P to be another Givens rotation to eliminate b A. These transformations can be replaced by Householder transformations or Gaussian elimination matrices ith ro or column pivoting. This results in several algorithms ith dierent eciency and numerical stability properties. In the next section, e describe a general procedure for solving (.) via the computation of the factorization (.) We also discuss eciency and numerical stability issues for dierent choices of W and H in (.). 3. The Algorithms. We no describe fast algorithms for solving (.), here A is a general banded+semiseparable matrix of the form (.).

5 Fast and Stable Solver for Semiseparable Systems Preprocessing and Basic Linear Algebra Procedures. Some preprocessing is needed before the algorithms formally start. We assume that u is loer triangular. This can be achieved by computing a QR factorization u T = Q R and setting (3.) u := R T and v := v Q: This operation takes roughly 6nru ops using the fact that Q is computed in factored form [5, h. 5]. We also revie a fe ell-knon basic linear algebra routines needed in our algorithms. Let L be an m by s loer triangular matrix ith m > s, l l l. L =..... l s l s l ss : B... A l m l m l ms Algorithm 3. belo is a standard procedure for eciently zeroing out entries l ; l ; ; l ss on the main diagonal of L by using s Givens rotations (see [5, h. ]). Algorithm 3.. Elimination ith Givens Rotations for i := s to step? do hoose c i + s i = such that q ci s i li;i = here?s i c i l i = l i+;i i i;i + l i+;i : Set l i;i :=, l i+;i := i, and compute li; l i;i? ci s := i li; l i;i? : l i+; l i+;i??s i c i l i+; l i+;i? endfor Let W be the product of all the Givens rotations used in the above algorithm. Then its output can be ritten via a matrix-matrix product as L := W L. Similarly, e can zero out the main diagonal of L by using a banded Gassian elimination procedure ith ro pivoting. See Golub and Van Loan [5, h. 4] for details. Let G R ms be a general dense matrix. Then e can choose a Householder tranformation H = I? u u T ; ith kuk = to zero out all the entries in the rst ro of G except the (; ) entry: b (3.) G H = bg G b : The cost for computing u is O(s), and the cost for computing G H is about 4ms ops (see [5, h. 5]).

6 6 S. handrasekaran and M. Gu Alternatively, e can choose H in (3.) to be a Gaussian elimination matrix of the form?h T H = : I olumn pivoting can be used to enhance numerical stability. The cost for computing G H is about ms ops (see Golub and Van Loan [5, h. 3]). 3.. Ne Algorithms. Let ` = b u + r u + and m = ` + b l, e begin by riting A in the folloing form G E (3.3) A = ; F here G R m` is a dense matrix (its banded+semiseparable structure ill be ignored); F R (n?ru?)(n?`) is a banded+semiseparable rectangular matrix; and both R (n?m)` and E R (ru+)(n?`) are lo rank matrices. We caution that strictly speaking equation (3.3) is not a block partitioning of A, since the ro dimension of G is larger than that of E in general. In further detail, e rite = p q T, here p R (n?m)rl and q R`rl contain the last n? m ros of p and the rst ` ros of q, respectively. Similarly, E = u v T, here u R (ru+)ru and v R (n?`)ru contain the rst r u + ros of u and the last n? ` ros of v, respectively. As suggested in x3., e assume that u is a loer triangular matrix. As in x, e ill solve (.) by recursively solving the linear sytem (3.4) A x = b b? ; p here? = R rl is an auxiliary vector that ill play the role of scalar? in the example in x. As before, e ill compute a to-sided decomposition (.) of A and invert matrices W, L, and H on-line. To start the recursion, e choose a matrix W so that W u is a loer triangular matrix ith zeros on its main diagonal. ompute (3.5) bu := W u; b G := W I G; and b := W b? : I p? Linear system (3.4) no becomes B bg p q T v bu T A x = b b: F We further choose H to zero out the rst ro of G b except the (; ) entry. ompute e A := G b H and et (3.6) A := H T q: eg G e eq

7 Fast and Stable Solver for Semiseparable Systems 7 Linear system (3.4) no has the folloing form: e B eg G e bu v T (3.7) A ex = b b; p e p eq T F here ex = e b A x = B ex A and b = b B b I A : ex b b? p? H? No e can perform one step of forard substitution in (3.7) to get e = =e b and eg bu v T b b? e eg e b (3.8) A ex = A A : p eq T F b b? p (? + e e) b b? p This is a system smaller in dimension than (3.4). To complete the recursion, in the folloing e rerite it in the form of (3.4). Rerite v = et p = et ef f et A and F = ev ep ef 3 F e here e T and e T are the rst ros of v and p, respectively; f e R m?ru ; and f e, and ef 3 are column vectors of appropriate dimensions. Similar to equation (3.3), the block form of F above is strictly speaking not a block partitioning of F, since the length of ef is larger than in general. F e is itself a banded+semiseparable rectangular matrix. With this notation, e can no rerite (3.8) in the form of (3.4) as G E _ F ~ ex = b _? (3.9) ep here eg bu e _G = A e T eq T f e ; _ = ep? eq T e ; _E = bu A ev T ; _ B b = b T e b b b? e T ith e T and b T being the (` + )-th and (r u + )-th ros of q and u, respectively. Once again the block form of _G is not a block partitioning. As in x, e can perform elimination and forard substitution steps using formulas (3.4) through (3.9) recursively for some k times to obtain solution components e, e ; : : :; e k?. We stop the recursion hen the problem size n? k in (3.9) becomes so small that n? k m, at hich point e solve it directly to get a solution e.

8 8 S. handrasekaran and M. Gu To recover the solution to our original problem (.), let H, H ; : : :; H k? be the elimination matrices used at the second elimination step dened by equations (3.6) and (3.7), e compute the solution to (.) as (3.) x = I H I A I H I A I H k? I A B e. e k? e here the various identity matrices I, are in general of dierent dimensions Eciency and Numerical Stability onsiderations. In this section e consider special choices of matrices W and H in the recursion and ho they aect the eciency and numerical stability of the procedure. To make op counting simpler, in this section e assume that r l ; r u ; b l ; b u n even though our algorithms ork for general banded + semiseparable matrices. For complete backard stability, e can choose the W matrices in (3.5) to be the product of r u Givens rotations as suggested by Algorithm 3.. The costs for computing bu, G, b and b are about 3r u ops, 6r u` ops, and 6r u ops respectively. Hence the total cost for one step of (3.5) is about 3r u (r u + `) ops. We then choose H in (3.6) as a Householder transformation. The costs for computing G b H and H T q are about 4m` ops and 4r l` ops, respectively. Hence the total cost for one step of (3.6) is about 4(m + r l )` ops. In equation (3.8), the costs for computing e b and are about m ops and r l ops, respectively, leading to a total of (m + r l ) ops. In equation (3.9), the main cost is to explicitly form the last ro and column of G. _ The costs for computing bu e and e T eq T are about ru ops and r l` ops, respectively. There is essentially no cost for f e, hich consists of the non-zero components of a column in the banded matrix D. Hence the total cost in (3.9) is about (ru + r l`) ops. Since there are k n steps of recursion, the total cost for the procedure is about? 3ru (r u + `) + 4(m + r l )` + (ru + r l`) n (3.) =? 5r u + (b u + b l + 3r l + 5r u ) (b u + r u ) n ops: Additionally, there is a cost of about 6run ops for the preprocessing step (3.). With such choices of W and H, e obtain a factorization (.) ith orthorgonal matrix W and H. Since only orthogonal transformations and one forard substitution are used for the solution of (.), this algorithm is backard stable. To reduce computational cost, e can also choose W via the banded Gaussian elimination procedure ith ro pivoting in Golub and Van Loan [5, h. 4]. And e can choose H as a Gaussian elimination matrix, ith column pivoting if necessary. This choice of W and H leads to a factorization (.) ith upper triangular matrices W and H. It is quite interesting to note that factorizations of this form do not seem to have been discussed before in the literature. With this choice of W and H, the cost for one step of (3.5) is about r u (r u + `) ops; the cost for one step of (3.6) is about (m+r l )` ops; and the total cost in (3.9) is about (r u + r l`) ops. With k n steps of recursion, the total cost for the procedure is about?ru (r u + `) + (m + r l )` + (ru + r l`) n (3.) =? 3ru + (b u + b l + r l + r u ) (b u + r u ) n ops:

9 Fast and Stable Solver for Semiseparable Systems 9 Additionally, there is a cost of about 6run ops for the preprocessing step (3.). It is ell-knon that Gaussian elimination ith partial pivoting could occasionally become numerically unstable if certain element groth is too large (see Golub and Van Loan [5, h. 3]). It is likely that by using Gaussian elimination procedures in (3.5) and (3.6), the resulting factorization could become numerically unstable in pathological cases for large values of r u and b u. Alternatively, e can choose only one of W and H to be orthogonal, leading to a factorization (.) ith one of W and H to be orthogonal and the other upper triangular. Furthermore, the choices of W and H can change from one recursion step to another, leading to a factorization (.) ith no obvious structures in W and H. While our algorithms ere presented in such a ay that the only one variable in (3.4) is eliminated in forard substitution at every recursion step. It is straightforard to reorganize the computation to develop a block version here a number of of variables are eliminated together. Given the success of the recent linear algebra package Lapack [] in using block algorithms to speed up numerical computation, it seems clear that hen the dimension becomes very large, the problem (.) can be solved more eciently by block versions of our algorithms. Finally, e note that the problem (.) can be reritten in the folloing form (3.3) B y = S b; B = S A S and x = S y; here S is the matrix ith 's on the main anti-diagonal and zero elsehere. It is easy to verify that B = (S D S) + tril (S u) (S v) T ;?b u? + triu (S p) (S q) T ; b l + : It can be veried that (S D S) is a banded matrix ith b u non-zero diagonals strictly belo the main diagonal and b l non-zero diagonals strictly above the main diagonal. Hence B is itself a banded+semiseparable matrix ith the banded+semiseparable structure of A. Applying the to algorithms e just discussed to solve (3.3), e see that the total costs are? 5r (3.4) l + (b l + b u + 3r u + 5r l ) (b l + r l ) n ops and (3.5)? 3r l + (b l + b u + r u + r l ) (b l + r l ) n ops; respectively. This suggests that one should choose among the to forms (.) and (3.3) according to formulas (3.) through (3.5) to reduce computational cost. 4. Numerical Experiments. In this section, e summarize the results from our numerical experiments ith the algorithms that ere presented in x3. These experiments ere performed on an UltraSparc orkstation in matlab ith double precision?6. We tested to algorithms For example, hen n =, e have S = :

10 S. handrasekaran and M. Gu Algorithm-I: Only Gaussian elimination steps ith partial pivoting are used in computing (.). Algorithm-II: Only Givens rotations and Householder reections are used in computing (.). In all of the test matrices, e chose r l = n=, r u = n=5, b u = and b l =. The matrix entries ere generated randomly. In Table 3., e compared Algorithms I and II in terms of the numbers of ops required to solve (.). The column marked GEPP is the number of ops required for Gaussian elimination ith partial pivoting to solve (.) by treating A as a dense matrix. We see that Algorithm I requires less ops than Algorithm II, and both Algorithms I and II require signicantly less numbers of ops to solve (.) than GEPP. In Table 3., e compared Algorithms I and II in terms of execution times and backard errors. The execution times are in seconds, and the backard error is dened as ka bx? bk kak kbxk ; here bx is the computed solution to (.). This backard error is the smallest relative backard error in -norm (see [5, h. 3]). learly Algorithm I is faster than Algorithm II as expected. Both are comparable in terms of backard errors. Hoever, as e mentioned in x3, Algorithm I could be numerically unstable in pathological cases if any one of b u and r u is very large. REFERENES [] E. Anderson, Z. Bai,. Bischof, J. Demmel, J. Dongarra, J. Du roz, A. Greenbaum, S. Hammarling, A. McKenney, S. Ostrouchov, and D. Sorensen, LAPAK Users' Guide, SIAM, Philadelphia, PA, second ed., 994. [] W. J. Demmel, Numerical Linear Algebra, SIAM, Philadelphia, 997. [3] Y. Eidelman and I. Gohberg, Inversion formulas and linear complexity algorithm for diagonal plus semiseparable matrices, omputers and Mathematics ith Applications, Feb. 997, vol.33, no.4, pp. 69{79. [4] Y. Eidelman and I. Gohberg, A look-ahead block Schur algorithm for diagonal plus semiseparable matrices, omputers and Mathematics ith Applications, May 998, vol.35, no., pp. 5{34. [5] G. Golub and. Van Loan, Matrix omputations, Johns Hopkins University Press, Baltimore, MD, 3nd ed., 996. [6] L. Greengard and V. Rokhlin, On the numerical solution of to-point boundary value problems, omm. Pure Appl. Math., XLIV (99), pp. 49{45. [7] June-Yub Lee and L. Greengard, A fast adaptive numerical method for sti to-point boundary value problems, SIAM Journal on Scientic omputing, March 997, vol.8, no., pp. 43{9. [8] Y. Saad, Iterative Methods for Sparse Linear Systems, PWS Publishing ompany, 996. [9] H. P. Starr, Jr., On the Numerical Solution of One-Dimensional Integral and Dierential Equations, PhD thesis, Department of omputer Science, Yale University, May 99. [] G. W. Steart, Updating a rank-revealing ULV decomposition, SIAM J. Mat. Anal. Appl., 4 (993), pp. 494{499.

11 Fast and Stable Solver for Semiseparable Systems Table 3. Numbers of Flops n Algorithm I Algorithm II GEPP Table 3. Execution Times and Backard Errors Time (seconds) Backard Error n Algorithm I Algorithm II Algorithm I Algorithm II 5 7.6?. 6.?9.6? ?9 5.8? ?.? ?.? ? 6.3? ? 3.8? ?.9? ? 5.3? ? 3.4? ?.?9

Exponentials of Symmetric Matrices through Tridiagonal Reductions

Exponentials of Symmetric Matrices through Tridiagonal Reductions Exponentials of Symmetric Matrices through Tridiagonal Reductions Ya Yan Lu Department of Mathematics City University of Hong Kong Kowloon, Hong Kong Abstract A simple and efficient numerical algorithm

More information

LAPACK-Style Codes for Pivoted Cholesky and QR Updating

LAPACK-Style Codes for Pivoted Cholesky and QR Updating LAPACK-Style Codes for Pivoted Cholesky and QR Updating Sven Hammarling 1, Nicholas J. Higham 2, and Craig Lucas 3 1 NAG Ltd.,Wilkinson House, Jordan Hill Road, Oxford, OX2 8DR, England, sven@nag.co.uk,

More information

LAPACK-Style Codes for Pivoted Cholesky and QR Updating. Hammarling, Sven and Higham, Nicholas J. and Lucas, Craig. MIMS EPrint: 2006.

LAPACK-Style Codes for Pivoted Cholesky and QR Updating. Hammarling, Sven and Higham, Nicholas J. and Lucas, Craig. MIMS EPrint: 2006. LAPACK-Style Codes for Pivoted Cholesky and QR Updating Hammarling, Sven and Higham, Nicholas J. and Lucas, Craig 2007 MIMS EPrint: 2006.385 Manchester Institute for Mathematical Sciences School of Mathematics

More information

CHAPTER 3 THE COMMON FACTOR MODEL IN THE POPULATION. From Exploratory Factor Analysis Ledyard R Tucker and Robert C. MacCallum

CHAPTER 3 THE COMMON FACTOR MODEL IN THE POPULATION. From Exploratory Factor Analysis Ledyard R Tucker and Robert C. MacCallum CHAPTER 3 THE COMMON FACTOR MODEL IN THE POPULATION From Exploratory Factor Analysis Ledyard R Tucker and Robert C. MacCallum 1997 19 CHAPTER 3 THE COMMON FACTOR MODEL IN THE POPULATION 3.0. Introduction

More information

Algorithm 853: an Efficient Algorithm for Solving Rank-Deficient Least Squares Problems

Algorithm 853: an Efficient Algorithm for Solving Rank-Deficient Least Squares Problems Algorithm 853: an Efficient Algorithm for Solving Rank-Deficient Least Squares Problems LESLIE FOSTER and RAJESH KOMMU San Jose State University Existing routines, such as xgelsy or xgelsd in LAPACK, for

More information

Positive Denite Matrix. Ya Yan Lu 1. Department of Mathematics. City University of Hong Kong. Kowloon, Hong Kong. Abstract

Positive Denite Matrix. Ya Yan Lu 1. Department of Mathematics. City University of Hong Kong. Kowloon, Hong Kong. Abstract Computing the Logarithm of a Symmetric Positive Denite Matrix Ya Yan Lu Department of Mathematics City University of Hong Kong Kowloon, Hong Kong Abstract A numerical method for computing the logarithm

More information

Generalized interval arithmetic on compact matrix Lie groups

Generalized interval arithmetic on compact matrix Lie groups myjournal manuscript No. (will be inserted by the editor) Generalized interval arithmetic on compact matrix Lie groups Hermann Schichl, Mihály Csaba Markót, Arnold Neumaier Faculty of Mathematics, University

More information

Institute for Advanced Computer Studies. Department of Computer Science. On the Adjoint Matrix. G. W. Stewart y ABSTRACT

Institute for Advanced Computer Studies. Department of Computer Science. On the Adjoint Matrix. G. W. Stewart y ABSTRACT University of Maryland Institute for Advanced Computer Studies Department of Computer Science College Park TR{97{02 TR{3864 On the Adjoint Matrix G. W. Stewart y ABSTRACT The adjoint A A of a matrix A

More information

MMSE Equalizer Design

MMSE Equalizer Design MMSE Equalizer Design Phil Schniter March 6, 2008 [k] a[m] P a [k] g[k] m[k] h[k] + ṽ[k] q[k] y [k] P y[m] For a trivial channel (i.e., h[k] = δ[k]), e kno that the use of square-root raisedcosine (SRRC)

More information

y(x) = x w + ε(x), (1)

y(x) = x w + ε(x), (1) Linear regression We are ready to consider our first machine-learning problem: linear regression. Suppose that e are interested in the values of a function y(x): R d R, here x is a d-dimensional vector-valued

More information

3 - Vector Spaces Definition vector space linear space u, v,

3 - Vector Spaces Definition vector space linear space u, v, 3 - Vector Spaces Vectors in R and R 3 are essentially matrices. They can be vieed either as column vectors (matrices of size and 3, respectively) or ro vectors ( and 3 matrices). The addition and scalar

More information

ETNA Kent State University

ETNA Kent State University C 8 Electronic Transactions on Numerical Analysis. Volume 17, pp. 76-2, 2004. Copyright 2004,. ISSN 1068-613. etnamcs.kent.edu STRONG RANK REVEALING CHOLESKY FACTORIZATION M. GU AND L. MIRANIAN Abstract.

More information

Matrix decompositions

Matrix decompositions Matrix decompositions How can we solve Ax = b? 1 Linear algebra Typical linear system of equations : x 1 x +x = x 1 +x +9x = 0 x 1 +x x = The variables x 1, x, and x only appear as linear terms (no powers

More information

CANONICAL FORM. University of Minnesota University of Illinois. Computer Science Dept. Electrical Eng. Dept. Abstract

CANONICAL FORM. University of Minnesota University of Illinois. Computer Science Dept. Electrical Eng. Dept. Abstract PLAING ZEROES and the KRONEKER ANONIAL FORM D L oley P Van Dooren University of Minnesota University of Illinois omputer Science Dept Electrical Eng Dept Minneapolis, MN USA Urbana, IL 8 USA boleymailcsumnedu

More information

Matrix decompositions

Matrix decompositions Matrix decompositions How can we solve Ax = b? 1 Linear algebra Typical linear system of equations : x 1 x +x = x 1 +x +9x = 0 x 1 +x x = The variables x 1, x, and x only appear as linear terms (no powers

More information

Applied Linear Algebra in Geoscience Using MATLAB

Applied Linear Algebra in Geoscience Using MATLAB Applied Linear Algebra in Geoscience Using MATLAB Contents Getting Started Creating Arrays Mathematical Operations with Arrays Using Script Files and Managing Data Two-Dimensional Plots Programming in

More information

Institute for Advanced Computer Studies. Department of Computer Science. Two Algorithms for the The Ecient Computation of

Institute for Advanced Computer Studies. Department of Computer Science. Two Algorithms for the The Ecient Computation of University of Maryland Institute for Advanced Computer Studies Department of Computer Science College Park TR{98{12 TR{3875 Two Algorithms for the The Ecient Computation of Truncated Pivoted QR Approximations

More information

problem Au = u by constructing an orthonormal basis V k = [v 1 ; : : : ; v k ], at each k th iteration step, and then nding an approximation for the e

problem Au = u by constructing an orthonormal basis V k = [v 1 ; : : : ; v k ], at each k th iteration step, and then nding an approximation for the e A Parallel Solver for Extreme Eigenpairs 1 Leonardo Borges and Suely Oliveira 2 Computer Science Department, Texas A&M University, College Station, TX 77843-3112, USA. Abstract. In this paper a parallel

More information

Computing least squares condition numbers on hybrid multicore/gpu systems

Computing least squares condition numbers on hybrid multicore/gpu systems Computing least squares condition numbers on hybrid multicore/gpu systems M. Baboulin and J. Dongarra and R. Lacroix Abstract This paper presents an efficient computation for least squares conditioning

More information

Dense LU factorization and its error analysis

Dense LU factorization and its error analysis Dense LU factorization and its error analysis Laura Grigori INRIA and LJLL, UPMC February 2016 Plan Basis of floating point arithmetic and stability analysis Notation, results, proofs taken from [N.J.Higham,

More information

2 MULTIPLYING COMPLEX MATRICES It is rare in matrix computations to be able to produce such a clear-cut computational saving over a standard technique

2 MULTIPLYING COMPLEX MATRICES It is rare in matrix computations to be able to produce such a clear-cut computational saving over a standard technique STABILITY OF A METHOD FOR MULTIPLYING COMPLEX MATRICES WITH THREE REAL MATRIX MULTIPLICATIONS NICHOLAS J. HIGHAM y Abstract. By use of a simple identity, the product of two complex matrices can be formed

More information

OUTLINE 1. Introduction 1.1 Notation 1.2 Special matrices 2. Gaussian Elimination 2.1 Vector and matrix norms 2.2 Finite precision arithmetic 2.3 Fact

OUTLINE 1. Introduction 1.1 Notation 1.2 Special matrices 2. Gaussian Elimination 2.1 Vector and matrix norms 2.2 Finite precision arithmetic 2.3 Fact Computational Linear Algebra Course: (MATH: 6800, CSCI: 6800) Semester: Fall 1998 Instructors: { Joseph E. Flaherty, aherje@cs.rpi.edu { Franklin T. Luk, luk@cs.rpi.edu { Wesley Turner, turnerw@cs.rpi.edu

More information

LU factorization with Panel Rank Revealing Pivoting and its Communication Avoiding version

LU factorization with Panel Rank Revealing Pivoting and its Communication Avoiding version 1 LU factorization with Panel Rank Revealing Pivoting and its Communication Avoiding version Amal Khabou Advisor: Laura Grigori Université Paris Sud 11, INRIA Saclay France SIAMPP12 February 17, 2012 2

More information

BlockMatrixComputations and the Singular Value Decomposition. ATaleofTwoIdeas

BlockMatrixComputations and the Singular Value Decomposition. ATaleofTwoIdeas BlockMatrixComputations and the Singular Value Decomposition ATaleofTwoIdeas Charles F. Van Loan Department of Computer Science Cornell University Supported in part by the NSF contract CCR-9901988. Block

More information

On Orthogonal Block Elimination. Christian Bischof and Xiaobai Sun. Argonne, IL Argonne Preprint MCS-P

On Orthogonal Block Elimination. Christian Bischof and Xiaobai Sun. Argonne, IL Argonne Preprint MCS-P On Orthogonal Block Elimination Christian Bischof and iaobai Sun Mathematics and Computer Science Division Argonne National Laboratory Argonne, IL 6439 fbischof,xiaobaig@mcs.anl.gov Argonne Preprint MCS-P45-794

More information

KEYWORDS. Numerical methods, generalized singular values, products of matrices, quotients of matrices. Introduction The two basic unitary decompositio

KEYWORDS. Numerical methods, generalized singular values, products of matrices, quotients of matrices. Introduction The two basic unitary decompositio COMPUTING THE SVD OF A GENERAL MATRIX PRODUCT/QUOTIENT GENE GOLUB Computer Science Department Stanford University Stanford, CA USA golub@sccm.stanford.edu KNUT SLNA SC-CM Stanford University Stanford,

More information

Block-Partitioned Algorithms for. Solving the Linear Least Squares. Problem. Gregorio Quintana-Orti, Enrique S. Quintana-Orti, and Antoine Petitet

Block-Partitioned Algorithms for. Solving the Linear Least Squares. Problem. Gregorio Quintana-Orti, Enrique S. Quintana-Orti, and Antoine Petitet Block-Partitioned Algorithms for Solving the Linear Least Squares Problem Gregorio Quintana-Orti, Enrique S. Quintana-Orti, and Antoine Petitet CRPC-TR9674-S January 1996 Center for Research on Parallel

More information

These notes give a quick summary of the part of the theory of autonomous ordinary differential equations relevant to modeling zombie epidemics.

These notes give a quick summary of the part of the theory of autonomous ordinary differential equations relevant to modeling zombie epidemics. NOTES ON AUTONOMOUS ORDINARY DIFFERENTIAL EQUATIONS MARCH 2017 These notes give a quick summary of the part of the theory of autonomous ordinary differential equations relevant to modeling zombie epidemics.

More information

ANONSINGULAR tridiagonal linear system of the form

ANONSINGULAR tridiagonal linear system of the form Generalized Diagonal Pivoting Methods for Tridiagonal Systems without Interchanges Jennifer B. Erway, Roummel F. Marcia, and Joseph A. Tyson Abstract It has been shown that a nonsingular symmetric tridiagonal

More information

A New Block Algorithm for Full-Rank Solution of the Sylvester-observer Equation.

A New Block Algorithm for Full-Rank Solution of the Sylvester-observer Equation. 1 A New Block Algorithm for Full-Rank Solution of the Sylvester-observer Equation João Carvalho, DMPA, Universidade Federal do RS, Brasil Karabi Datta, Dep MSc, Northern Illinois University, DeKalb, IL

More information

The restarted QR-algorithm for eigenvalue computation of structured matrices

The restarted QR-algorithm for eigenvalue computation of structured matrices Journal of Computational and Applied Mathematics 149 (2002) 415 422 www.elsevier.com/locate/cam The restarted QR-algorithm for eigenvalue computation of structured matrices Daniela Calvetti a; 1, Sun-Mi

More information

Linear Algebra Linear Algebra : Matrix decompositions Monday, February 11th Math 365 Week #4

Linear Algebra Linear Algebra : Matrix decompositions Monday, February 11th Math 365 Week #4 Linear Algebra Linear Algebra : Matrix decompositions Monday, February 11th Math Week # 1 Saturday, February 1, 1 Linear algebra Typical linear system of equations : x 1 x +x = x 1 +x +9x = 0 x 1 +x x

More information

WEIGHTED LEAST-SQUARES STATE ESTIMATION PART II: SIMULATION. with the normal equation method. problems. Algorithm 1 uses quasi-newton's method

WEIGHTED LEAST-SQUARES STATE ESTIMATION PART II: SIMULATION. with the normal equation method. problems. Algorithm 1 uses quasi-newton's method APPLICAION OF OPIMAL MULIPLIER MEHOD IN WEIGHED LEAS-SQUARES SAE ESIMAION PAR II: SIMULAION Jianping Meng Christopher L. DeMarco Department of Electrical and Computer Engineering University of Wisconsin-Madison

More information

Using Godunov s Two-Sided Sturm Sequences to Accurately Compute Singular Vectors of Bidiagonal Matrices.

Using Godunov s Two-Sided Sturm Sequences to Accurately Compute Singular Vectors of Bidiagonal Matrices. Using Godunov s Two-Sided Sturm Sequences to Accurately Compute Singular Vectors of Bidiagonal Matrices. A.M. Matsekh E.P. Shurina 1 Introduction We present a hybrid scheme for computing singular vectors

More information

Arnoldi Methods in SLEPc

Arnoldi Methods in SLEPc Scalable Library for Eigenvalue Problem Computations SLEPc Technical Report STR-4 Available at http://slepc.upv.es Arnoldi Methods in SLEPc V. Hernández J. E. Román A. Tomás V. Vidal Last update: October,

More information

Sparse BLAS-3 Reduction

Sparse BLAS-3 Reduction Sparse BLAS-3 Reduction to Banded Upper Triangular (Spar3Bnd) Gary Howell, HPC/OIT NC State University gary howell@ncsu.edu Sparse BLAS-3 Reduction p.1/27 Acknowledgements James Demmel, Gene Golub, Franc

More information

Review of matrices. Let m, n IN. A rectangle of numbers written like A =

Review of matrices. Let m, n IN. A rectangle of numbers written like A = Review of matrices Let m, n IN. A rectangle of numbers written like a 11 a 12... a 1n a 21 a 22... a 2n A =...... a m1 a m2... a mn where each a ij IR is called a matrix with m rows and n columns or an

More information

COMPUTING THE CONDITION NUMBER OF TRIDIAGONAL AND DIAGONAL-PLUS-SEMISEPARABLE MATRICES IN LINEAR TIME

COMPUTING THE CONDITION NUMBER OF TRIDIAGONAL AND DIAGONAL-PLUS-SEMISEPARABLE MATRICES IN LINEAR TIME COMPUTING THE CONDITION NUMBER OF TRIDIAGONAL AND DIAGONAL-PLUS-SEMISEPARABLE MATRICES IN LINEAR TIME GARETH I HARGREAVES Abstract For an n n tridiagonal matrix we exploit the structure of its QR factorization

More information

Scientific Computing

Scientific Computing Scientific Computing Direct solution methods Martin van Gijzen Delft University of Technology October 3, 2018 1 Program October 3 Matrix norms LU decomposition Basic algorithm Cost Stability Pivoting Pivoting

More information

Linear Equations and Matrix

Linear Equations and Matrix 1/60 Chia-Ping Chen Professor Department of Computer Science and Engineering National Sun Yat-sen University Linear Algebra Gaussian Elimination 2/60 Alpha Go Linear algebra begins with a system of linear

More information

Numerical Methods in Matrix Computations

Numerical Methods in Matrix Computations Ake Bjorck Numerical Methods in Matrix Computations Springer Contents 1 Direct Methods for Linear Systems 1 1.1 Elements of Matrix Theory 1 1.1.1 Matrix Algebra 2 1.1.2 Vector Spaces 6 1.1.3 Submatrices

More information

Module 6.6: nag nsym gen eig Nonsymmetric Generalized Eigenvalue Problems. Contents

Module 6.6: nag nsym gen eig Nonsymmetric Generalized Eigenvalue Problems. Contents Eigenvalue and Least-squares Problems Module Contents Module 6.6: nag nsym gen eig Nonsymmetric Generalized Eigenvalue Problems nag nsym gen eig provides procedures for solving nonsymmetric generalized

More information

7. LU factorization. factor-solve method. LU factorization. solving Ax = b with A nonsingular. the inverse of a nonsingular matrix

7. LU factorization. factor-solve method. LU factorization. solving Ax = b with A nonsingular. the inverse of a nonsingular matrix EE507 - Computational Techniques for EE 7. LU factorization Jitkomut Songsiri factor-solve method LU factorization solving Ax = b with A nonsingular the inverse of a nonsingular matrix LU factorization

More information

1 :: Mathematical notation

1 :: Mathematical notation 1 :: Mathematical notation x A means x is a member of the set A. A B means the set A is contained in the set B. {a 1,..., a n } means the set hose elements are a 1,..., a n. {x A : P } means the set of

More information

LU Factorization. Marco Chiarandini. DM559 Linear and Integer Programming. Department of Mathematics & Computer Science University of Southern Denmark

LU Factorization. Marco Chiarandini. DM559 Linear and Integer Programming. Department of Mathematics & Computer Science University of Southern Denmark DM559 Linear and Integer Programming LU Factorization Marco Chiarandini Department of Mathematics & Computer Science University of Southern Denmark [Based on slides by Lieven Vandenberghe, UCLA] Outline

More information

D. Gimenez, M. T. Camara, P. Montilla. Aptdo Murcia. Spain. ABSTRACT

D. Gimenez, M. T. Camara, P. Montilla. Aptdo Murcia. Spain.   ABSTRACT Accelerating the Convergence of Blocked Jacobi Methods 1 D. Gimenez, M. T. Camara, P. Montilla Departamento de Informatica y Sistemas. Univ de Murcia. Aptdo 401. 0001 Murcia. Spain. e-mail: fdomingo,cpmcm,cppmmg@dif.um.es

More information

UMIACS-TR July CS-TR 2494 Revised January An Updating Algorithm for. Subspace Tracking. G. W. Stewart. abstract

UMIACS-TR July CS-TR 2494 Revised January An Updating Algorithm for. Subspace Tracking. G. W. Stewart. abstract UMIACS-TR-9-86 July 199 CS-TR 2494 Revised January 1991 An Updating Algorithm for Subspace Tracking G. W. Stewart abstract In certain signal processing applications it is required to compute the null space

More information

Preconditioned Parallel Block Jacobi SVD Algorithm

Preconditioned Parallel Block Jacobi SVD Algorithm Parallel Numerics 5, 15-24 M. Vajteršic, R. Trobec, P. Zinterhof, A. Uhl (Eds.) Chapter 2: Matrix Algebra ISBN 961-633-67-8 Preconditioned Parallel Block Jacobi SVD Algorithm Gabriel Okša 1, Marián Vajteršic

More information

A Backward Stable Hyperbolic QR Factorization Method for Solving Indefinite Least Squares Problem

A Backward Stable Hyperbolic QR Factorization Method for Solving Indefinite Least Squares Problem A Backward Stable Hyperbolic QR Factorization Method for Solving Indefinite Least Suares Problem Hongguo Xu Dedicated to Professor Erxiong Jiang on the occasion of his 7th birthday. Abstract We present

More information

Notes on LU Factorization

Notes on LU Factorization Notes on LU Factorization Robert A van de Geijn Department of Computer Science The University of Texas Austin, TX 78712 rvdg@csutexasedu October 11, 2014 The LU factorization is also known as the LU decomposition

More information

Solving linear equations with Gaussian Elimination (I)

Solving linear equations with Gaussian Elimination (I) Term Projects Solving linear equations with Gaussian Elimination The QR Algorithm for Symmetric Eigenvalue Problem The QR Algorithm for The SVD Quasi-Newton Methods Solving linear equations with Gaussian

More information

A note on eigenvalue computation for a tridiagonal matrix with real eigenvalues Akiko Fukuda

A note on eigenvalue computation for a tridiagonal matrix with real eigenvalues Akiko Fukuda Journal of Math-for-Industry Vol 3 (20A-4) pp 47 52 A note on eigenvalue computation for a tridiagonal matrix with real eigenvalues Aio Fuuda Received on October 6 200 / Revised on February 7 20 Abstract

More information

A turbulence closure based on the maximum entropy method

A turbulence closure based on the maximum entropy method Advances in Fluid Mechanics IX 547 A turbulence closure based on the maximum entropy method R. W. Derksen Department of Mechanical and Manufacturing Engineering University of Manitoba Winnipeg Canada Abstract

More information

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for 1 Logistics Notes for 2016-09-14 1. There was a goof in HW 2, problem 1 (now fixed) please re-download if you have already started looking at it. 2. CS colloquium (4:15 in Gates G01) this Thurs is Margaret

More information

Main matrix factorizations

Main matrix factorizations Main matrix factorizations A P L U P permutation matrix, L lower triangular, U upper triangular Key use: Solve square linear system Ax b. A Q R Q unitary, R upper triangular Key use: Solve square or overdetrmined

More information

Next topics: Solving systems of linear equations

Next topics: Solving systems of linear equations Next topics: Solving systems of linear equations 1 Gaussian elimination (today) 2 Gaussian elimination with partial pivoting (Week 9) 3 The method of LU-decomposition (Week 10) 4 Iterative techniques:

More information

Numerical linear algebra

Numerical linear algebra Numerical linear algebra Purdue University CS 51500 Fall 2017 David Gleich David F. Gleich Call me Prof Gleich Dr. Gleich Please not Hey matrix guy! Huda Nassar Call me Huda Ms. Huda Please not Matrix

More information

A fast randomized algorithm for orthogonal projection

A fast randomized algorithm for orthogonal projection A fast randomized algorithm for orthogonal projection Vladimir Rokhlin and Mark Tygert arxiv:0912.1135v2 [cs.na] 10 Dec 2009 December 10, 2009 Abstract We describe an algorithm that, given any full-rank

More information

Introduction to Mathematical Programming

Introduction to Mathematical Programming Introduction to Mathematical Programming Ming Zhong Lecture 6 September 12, 2018 Ming Zhong (JHU) AMS Fall 2018 1 / 20 Table of Contents 1 Ming Zhong (JHU) AMS Fall 2018 2 / 20 Solving Linear Systems A

More information

6 Linear Systems of Equations

6 Linear Systems of Equations 6 Linear Systems of Equations Read sections 2.1 2.3, 2.4.1 2.4.5, 2.4.7, 2.7 Review questions 2.1 2.37, 2.43 2.67 6.1 Introduction When numerically solving two-point boundary value problems, the differential

More information

Matrix Computations and Semiseparable Matrices

Matrix Computations and Semiseparable Matrices Matrix Computations and Semiseparable Matrices Volume I: Linear Systems Raf Vandebril Department of Computer Science Catholic University of Louvain Marc Van Barel Department of Computer Science Catholic

More information

MS&E 318 (CME 338) Large-Scale Numerical Optimization

MS&E 318 (CME 338) Large-Scale Numerical Optimization Stanford University, Management Science & Engineering (and ICME MS&E 38 (CME 338 Large-Scale Numerical Optimization Course description Instructor: Michael Saunders Spring 28 Notes : Review The course teaches

More information

ETNA Kent State University

ETNA Kent State University Electronic Transactions on Numerical Analysis. Volume 17, pp. 76-92, 2004. Copyright 2004,. ISSN 1068-9613. ETNA STRONG RANK REVEALING CHOLESKY FACTORIZATION M. GU AND L. MIRANIAN Ý Abstract. For any symmetric

More information

c 2007 Society for Industrial and Applied Mathematics

c 2007 Society for Industrial and Applied Mathematics SIAM J MATRIX ANAL APPL Vol 29, No, pp 27 266 c 2007 Society for Industrial and Applied Mathematics A SUPERFAST ALGORITM FOR TOEPLITZ SYSTEMS OF LINEAR EQUATIONS S CANDRASEKARAN, M GU, X SUN, J XIA, AND

More information

A Method for Constructing Diagonally Dominant Preconditioners based on Jacobi Rotations

A Method for Constructing Diagonally Dominant Preconditioners based on Jacobi Rotations A Method for Constructing Diagonally Dominant Preconditioners based on Jacobi Rotations Jin Yun Yuan Plamen Y. Yalamov Abstract A method is presented to make a given matrix strictly diagonally dominant

More information

GAUSSIAN ELIMINATION AND LU DECOMPOSITION (SUPPLEMENT FOR MA511)

GAUSSIAN ELIMINATION AND LU DECOMPOSITION (SUPPLEMENT FOR MA511) GAUSSIAN ELIMINATION AND LU DECOMPOSITION (SUPPLEMENT FOR MA511) D. ARAPURA Gaussian elimination is the go to method for all basic linear classes including this one. We go summarize the main ideas. 1.

More information

On aggressive early deflation in parallel variants of the QR algorithm

On aggressive early deflation in parallel variants of the QR algorithm On aggressive early deflation in parallel variants of the QR algorithm Bo Kågström 1, Daniel Kressner 2, and Meiyue Shao 1 1 Department of Computing Science and HPC2N Umeå University, S-901 87 Umeå, Sweden

More information

The antitriangular factorisation of saddle point matrices

The antitriangular factorisation of saddle point matrices The antitriangular factorisation of saddle point matrices J. Pestana and A. J. Wathen August 29, 2013 Abstract Mastronardi and Van Dooren [this journal, 34 (2013) pp. 173 196] recently introduced the block

More information

Eigenvalue Problems and Singular Value Decomposition

Eigenvalue Problems and Singular Value Decomposition Eigenvalue Problems and Singular Value Decomposition Sanzheng Qiao Department of Computing and Software McMaster University August, 2012 Outline 1 Eigenvalue Problems 2 Singular Value Decomposition 3 Software

More information

9. Numerical linear algebra background

9. Numerical linear algebra background Convex Optimization Boyd & Vandenberghe 9. Numerical linear algebra background matrix structure and algorithm complexity solving linear equations with factored matrices LU, Cholesky, LDL T factorization

More information

NAG Library Routine Document F08FAF (DSYEV)

NAG Library Routine Document F08FAF (DSYEV) NAG Library Routine Document (DSYEV) Note: before using this routine, please read the Users Note for your implementation to check the interpretation of bold italicised terms and other implementation-dependent

More information

Algorithms to solve block Toeplitz systems and. least-squares problems by transforming to Cauchy-like. matrices

Algorithms to solve block Toeplitz systems and. least-squares problems by transforming to Cauchy-like. matrices Algorithms to solve block Toeplitz systems and least-squares problems by transforming to Cauchy-like matrices K. Gallivan S. Thirumalai P. Van Dooren 1 Introduction Fast algorithms to factor Toeplitz matrices

More information

NAG Library Routine Document F07HAF (DPBSV)

NAG Library Routine Document F07HAF (DPBSV) NAG Library Routine Document (DPBSV) Note: before using this routine, please read the Users Note for your implementation to check the interpretation of bold italicised terms and other implementation-dependent

More information

Centro de Processamento de Dados, Universidade Federal do Rio Grande do Sul,

Centro de Processamento de Dados, Universidade Federal do Rio Grande do Sul, A COMPARISON OF ACCELERATION TECHNIQUES APPLIED TO THE METHOD RUDNEI DIAS DA CUNHA Computing Laboratory, University of Kent at Canterbury, U.K. Centro de Processamento de Dados, Universidade Federal do

More information

(17) (18)

(17) (18) Module 4 : Solving Linear Algebraic Equations Section 3 : Direct Solution Techniques 3 Direct Solution Techniques Methods for solving linear algebraic equations can be categorized as direct and iterative

More information

Henk van der Vorst. Abstract. We discuss a novel approach for the computation of a number of eigenvalues and eigenvectors

Henk van der Vorst. Abstract. We discuss a novel approach for the computation of a number of eigenvalues and eigenvectors Subspace Iteration for Eigenproblems Henk van der Vorst Abstract We discuss a novel approach for the computation of a number of eigenvalues and eigenvectors of the standard eigenproblem Ax = x. Our method

More information

I-v k e k. (I-e k h kt ) = Stability of Gauss-Huard Elimination for Solving Linear Systems. 1 x 1 x x x x

I-v k e k. (I-e k h kt ) = Stability of Gauss-Huard Elimination for Solving Linear Systems. 1 x 1 x x x x Technical Report CS-93-08 Department of Computer Systems Faculty of Mathematics and Computer Science University of Amsterdam Stability of Gauss-Huard Elimination for Solving Linear Systems T. J. Dekker

More information

Numerical Linear Algebra

Numerical Linear Algebra Numerical Linear Algebra Decompositions, numerical aspects Gerard Sleijpen and Martin van Gijzen September 27, 2017 1 Delft University of Technology Program Lecture 2 LU-decomposition Basic algorithm Cost

More information

Program Lecture 2. Numerical Linear Algebra. Gaussian elimination (2) Gaussian elimination. Decompositions, numerical aspects

Program Lecture 2. Numerical Linear Algebra. Gaussian elimination (2) Gaussian elimination. Decompositions, numerical aspects Numerical Linear Algebra Decompositions, numerical aspects Program Lecture 2 LU-decomposition Basic algorithm Cost Stability Pivoting Cholesky decomposition Sparse matrices and reorderings Gerard Sleijpen

More information

An exploration of matrix equilibration

An exploration of matrix equilibration An exploration of matrix equilibration Paul Liu Abstract We review three algorithms that scale the innity-norm of each row and column in a matrix to. The rst algorithm applies to unsymmetric matrices,

More information

Eigen-solving via reduction to DPR1 matrices

Eigen-solving via reduction to DPR1 matrices Computers and Mathematics with Applications 56 (2008) 166 171 www.elsevier.com/locate/camwa Eigen-solving via reduction to DPR1 matrices V.Y. Pan a,, B. Murphy a, R.E. Rosholt a, Y. Tang b, X. Wang c,

More information

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.

More information

NAG Library Routine Document F08JDF (DSTEVR)

NAG Library Routine Document F08JDF (DSTEVR) F08 Least-squares and Eigenvalue Problems (LAPACK) NAG Library Routine Document (DSTEVR) Note: before using this routine, please read the Users Note for your implementation to check the interpretation

More information

Reduced Synchronization Overhead on. December 3, Abstract. The standard formulation of the conjugate gradient algorithm involves

Reduced Synchronization Overhead on. December 3, Abstract. The standard formulation of the conjugate gradient algorithm involves Lapack Working Note 56 Conjugate Gradient Algorithms with Reduced Synchronization Overhead on Distributed Memory Multiprocessors E. F. D'Azevedo y, V.L. Eijkhout z, C. H. Romine y December 3, 1999 Abstract

More information

RANA03-02 January Jacobi-Davidson methods and preconditioning with applications in pole-zero analysis

RANA03-02 January Jacobi-Davidson methods and preconditioning with applications in pole-zero analysis RANA03-02 January 2003 Jacobi-Davidson methods and preconditioning with applications in pole-zero analysis by J.Rommes, H.A. van der Vorst, EJ.W. ter Maten Reports on Applied and Numerical Analysis Department

More information

Iterative Methods. Splitting Methods

Iterative Methods. Splitting Methods Iterative Methods Splitting Methods 1 Direct Methods Solving Ax = b using direct methods. Gaussian elimination (using LU decomposition) Variants of LU, including Crout and Doolittle Other decomposition

More information

Enhancing Scalability of Sparse Direct Methods

Enhancing Scalability of Sparse Direct Methods Journal of Physics: Conference Series 78 (007) 0 doi:0.088/7-6596/78//0 Enhancing Scalability of Sparse Direct Methods X.S. Li, J. Demmel, L. Grigori, M. Gu, J. Xia 5, S. Jardin 6, C. Sovinec 7, L.-Q.

More information

LU Factorization with Panel Rank Revealing Pivoting and Its Communication Avoiding Version

LU Factorization with Panel Rank Revealing Pivoting and Its Communication Avoiding Version LU Factorization with Panel Rank Revealing Pivoting and Its Communication Avoiding Version Amal Khabou James Demmel Laura Grigori Ming Gu Electrical Engineering and Computer Sciences University of California

More information

On the approximation of real powers of sparse, infinite, bounded and Hermitian matrices

On the approximation of real powers of sparse, infinite, bounded and Hermitian matrices On the approximation of real poers of sparse, infinite, bounded and Hermitian matrices Roman Werpachoski Center for Theoretical Physics, Al. Lotnikó 32/46 02-668 Warszaa, Poland Abstract We describe a

More information

5.6. PSEUDOINVERSES 101. A H w.

5.6. PSEUDOINVERSES 101. A H w. 5.6. PSEUDOINVERSES 0 Corollary 5.6.4. If A is a matrix such that A H A is invertible, then the least-squares solution to Av = w is v = A H A ) A H w. The matrix A H A ) A H is the left inverse of A and

More information

NAG Toolbox for Matlab nag_lapack_dggev (f08wa)

NAG Toolbox for Matlab nag_lapack_dggev (f08wa) NAG Toolbox for Matlab nag_lapack_dggev () 1 Purpose nag_lapack_dggev () computes for a pair of n by n real nonsymmetric matrices ða; BÞ the generalized eigenvalues and, optionally, the left and/or right

More information

Normalized power iterations for the computation of SVD

Normalized power iterations for the computation of SVD Normalized power iterations for the computation of SVD Per-Gunnar Martinsson Department of Applied Mathematics University of Colorado Boulder, Co. Per-gunnar.Martinsson@Colorado.edu Arthur Szlam Courant

More information

Lecture # 11 The Power Method for Eigenvalues Part II. The power method find the largest (in magnitude) eigenvalue of. A R n n.

Lecture # 11 The Power Method for Eigenvalues Part II. The power method find the largest (in magnitude) eigenvalue of. A R n n. Lecture # 11 The Power Method for Eigenvalues Part II The power method find the largest (in magnitude) eigenvalue of It makes two assumptions. 1. A is diagonalizable. That is, A R n n. A = XΛX 1 for some

More information

Review Questions REVIEW QUESTIONS 71

Review Questions REVIEW QUESTIONS 71 REVIEW QUESTIONS 71 MATLAB, is [42]. For a comprehensive treatment of error analysis and perturbation theory for linear systems and many other problems in linear algebra, see [126, 241]. An overview of

More information

LU Factorization. LU factorization is the most common way of solving linear systems! Ax = b LUx = b

LU Factorization. LU factorization is the most common way of solving linear systems! Ax = b LUx = b AM 205: lecture 7 Last time: LU factorization Today s lecture: Cholesky factorization, timing, QR factorization Reminder: assignment 1 due at 5 PM on Friday September 22 LU Factorization LU factorization

More information

c 2013 Society for Industrial and Applied Mathematics

c 2013 Society for Industrial and Applied Mathematics SIAM J. MATRIX ANAL. APPL. Vol. 34, No. 3, pp. 1401 1429 c 2013 Society for Industrial and Applied Mathematics LU FACTORIZATION WITH PANEL RANK REVEALING PIVOTING AND ITS COMMUNICATION AVOIDING VERSION

More information

Computational Methods. Eigenvalues and Singular Values

Computational Methods. Eigenvalues and Singular Values Computational Methods Eigenvalues and Singular Values Manfred Huber 2010 1 Eigenvalues and Singular Values Eigenvalues and singular values describe important aspects of transformations and of data relations

More information

A fast randomized algorithm for overdetermined linear least-squares regression

A fast randomized algorithm for overdetermined linear least-squares regression A fast randomized algorithm for overdetermined linear least-squares regression Vladimir Rokhlin and Mark Tygert Technical Report YALEU/DCS/TR-1403 April 28, 2008 Abstract We introduce a randomized algorithm

More information

Some Classes of Invertible Matrices in GF(2)

Some Classes of Invertible Matrices in GF(2) Some Classes of Invertible Matrices in GF() James S. Plank Adam L. Buchsbaum Technical Report UT-CS-07-599 Department of Electrical Engineering and Computer Science University of Tennessee August 16, 007

More information

BLAS: Basic Linear Algebra Subroutines Analysis of the Matrix-Vector-Product Analysis of Matrix-Matrix Product

BLAS: Basic Linear Algebra Subroutines Analysis of the Matrix-Vector-Product Analysis of Matrix-Matrix Product Level-1 BLAS: SAXPY BLAS-Notation: S single precision (D for double, C for complex) A α scalar X vector P plus operation Y vector SAXPY: y = αx + y Vectorization of SAXPY (αx + y) by pipelining: page 8

More information