MATLAB TOOLS FOR SOLVING PERIODIC EIGENVALUE PROBLEMS 1
|
|
- Morgan Madeleine Nash
- 11 months ago
- Views:
Transcription
1 MATLAB TOOLS FOR SOLVING PERIODIC EIGENVALUE PROBLEMS 1 Robert Granat Bo Kågström Daniel Kressner,2 Department of Computing Science and HPC2N, Umeå University, SE Umeå, Sweden. Abstract: Software for computing eigenvalues and invariant subspaces of general matrix products is proposed. The implemented algorithms are based on orthogonal transformations of the original data and thus attain numerical backward stability, which enables good accuracy even for small eigenvalues. The prospective toolbox combines the efficiency and robustness of library-style Fortran subroutines based on state-of-the-art algorithms with the convenience of Matlab interfaces. It will be demonstrated that this toolbox can be used to address a number of tasks in systems and control theory, such as the solution of periodic discrete-time algebraic Riccati equations. Keywords: linear periodic system, matrix product, periodic eigenvalue problem, periodic Schur form, library software, Matlab tools. 1. INTRODUCTION Let us consider a linear discrete-time system of the form E k x k+1 = A k x k + B k u k y k = C k x k + D k u k (1) with state, input and output vectors x k, u k, and y k, respectively. The coefficient matrices A k, B k, C k, D k, E k are supposed to be of matching dimensions. Moreover, we also assume that (1) is periodic for some period p 1, i.e., A k+p = A k, B k+p = B k, C k+p = C k, D k+p = D k, E k+p = E k for all integers k. Computational tasks for such periodic systems can often be addressed by solving periodic (or product) eigenvalue problems, see (Varga and Van Dooren 2001) for an overview. For example, if all E k are square and invertible 1 Supported by the Swedish Foundation for Strategic Research under the Frame Programme Grant A3 02: This author was additionally supported by a DFG Emmy Noether fellowship. then (1) is asymptotically stable if and only if all eigenvalues of the monodromy matrix Π = Ep 1 A p Ep 1 1 A p 1 E1 1 A 1 (2) lie strictly inside the unit circle. Theoretically, the eigenvalues of (2) can be computed by first forming the product Π explicitly and then applying any general-purpose eigenvalue solver. In finite-precision arithmetic, however, this approach may lead to potentially disastrous numerical inaccuracies. For example, if any of the matrices E k is nearly singular then its inversion must be avoided whenever possible; a rule that has driven the development of the QZ algorithm for the case p = 1 (Moler and Stewart 1973). Even if all matrices E k are well-conditioned, smaller eigenvalues of Π are severely affected by the lack of numerical backward stability in matrix multiplication (Higham 2002) and cannot be computed by such an approach.
2 There are several alternatives that take the product structure of Π into account and avoid some or all of the numerical disadvantages mentioned above. Lifting (Flamm 1991) is a well-known technique to transform a periodic system (1) into a timeinvariant none. Applied to (2), this leads to a block cyclic matrix pencil A λe = 0 A p E p. A λ E (3) A p 1 0 Ep 1 The eigenvalues of A λe are the p th roots of the eigenvalues of Π. While this embedding enables the application of the standard QZ algorithm to attain good eigenvalue accuracy, the numerical identification of the p th roots is problematic and the computational cost increases cubically with p. Collapsing (Benner and Byers 2001, Van Dooren 1999) is a technique that reduces the long product (2) to the case p = 1 without inverting any of the matrices E k. It can equivalently be viewed as an orthogonal equivalence transformation of a block cyclic matrix pencil closely related to (3). The computational cost of this approach increases linearly with p and numerical inaccuracies due to nearly singular E k are avoided. On the other hand, collapsing does not lead to an overall numerically backward stable method for solving periodic eigenvalue problems. This has been demonstrated in (Granat et al. 2006b) for computing eigenvectors belonging to smaller eigenvalues. The periodic QZ algorithm (Bojanczyk et al. 1992, Hench and Laub 1994) orthogonally transforms the matrices E k and A k into upper (quasi-)triangular form. It achieves numerical backward stability and has a computational cost that increases linearly with p. All publicly available implementations of the periodic QZ algorithm currently support only the case when each E k is an identity matrix. In this case, the periodic QZ reduces to the periodic QR algorithm, see also (Kressner 2006). Because of its numerical advantages, the implementations described in this paper are based on the periodic QZ algorithm and variants thereof. We are developing a Fortran library that admits non-trivial factors E k ; in fact, much more general matrix products are supported, see Section 2. The purpose of this paper is to describe prospective Matlab functions that are based on MEX interfaces to these Fortran routines. In Section 3, we describe the functionality of the toolbox on a level that is suitable for developers and expert users. In Section 4, a Matlab class product is introduced that allows the convenient use of these tools without knowledge of the internal algorithms or data structures. Besides periodic systems there are a variety of other applications requiring the solution of product eigenvalue problems. For example, the structure-preserving solution of time-invariant linearquadratic (LQ) optimal control problems gives rise to matrix products with up to five factors (Benner et al. 2002a). 2. NOTATION In the scope of this paper, we consider product eigenvalue problems in their most general form: A sp p A sp 1 p 1 As1 1, (4) with the indices s 1,..., s p {1, 1}. The dimensions of A 1,..., A p should match, i.e., { R n k+1 n k if s A k k = 1, R n k n k+1 if s k = 1, where n p+1 = n 1 (to simplify the notation we will identify p+1 with 1 in the following). Additionally, we require the condition n k + n k+1 = n k + n k+1, (5) s k =1 s k = 1 s k = 1 s k =1 which ensures that the corresponding lifted pencil (3) is square. The level of generality imposed by (4) allows to cover the other applications mentioned above that do not necessarily lead to a matrix product having the somewhat simpler form (2). It is worth mentioning that we admit matrices A k corresponding to an index s k = 1 to be singular or even rectangular, in which case the matrix product (4) should only be understood in a formal sense. 2.1 Periodic Schur decomposition To define the eigenvalues associated with a formal matrix product, one could consider the canonical form of (4), see (Sergeichuk 2004) for an overview, or an appropriate generalization of the lifted representation (3). A more computationally oriented definition can be derived from the periodic Schur decomposition. Theorem 1. There are orthogonal matrices Q 1 R n1 n1,..., Q p R np np such that each { Q T T k := k+1 A k Q k if s k = 1, Q T (6) k A k Q k+1 if s k = 1.
3 takes the form T k = T 11,k T 12,k T 13,k 0 T 22,k T 23,k 0 0 T 33,k, (7) where all T 22,k are square and of the same order n c n. Moreover, T 22,k is upper quasi-triangular for one index k = l and upper triangular for every other k. The formal matrix products T sp 11,p T s1 11,1 and T sp 33,p T s1 33,1 contain the dimension-induced infinite and zero eigenvalues, respectively. A mathematically precise definition of dimensioninduced infinite and zero eigenvalues is beyond the scope of this paper, see, e.g., (Sergeichuk 2004) for more details. The n c eigenvalues of T sp 22,p T s1 22,1 are called core eigenvalues and can be easily computed by multiplying the diagonal elements of T s1 22,1,..., T sp 22,p explicitly. 3 Core infinite eigenvalues may exist if at least one s k = 1 and T 22,k is singular; otherwise all core eigenvalues are finite. Let us illustrate the implication of Theorem 1 for the important special case when the inverted factors are square, i.e., n k = n k+1 whenever s k = 1. Then the blocks T 11,k in (7) are void, and T sp 33,p T s1 33,1 = 0 R(n1 nc) (n1 nc) with n c = min{n 1, n 2,..., n p }. Note that n c might be smaller than the minimum of all dimensions in the general case. 2.2 Deflating subspaces and eigenvectors A sequence of subspaces X 1 C n1, X 2 C n2,..., X p C np having the same dimension (e.g., dim(x k ) = m) is called a periodic deflating subspace if A k X k X k+1, if s k = 1, A k X k+1 X k, if s k = 1. If n k = n k+1 whenever s k = 1, the periodic Schur decomposition (6) (7) is easily seen to imply that the first m columns of Q 1,..., Q p span a periodic deflating subspace for every m n c such that the (m + 1, m) entry of T 22,l is nonzero. It belongs to the core eigenvalues that reside in the leading m diagonal elements of T s1 22,1,..., T sp 22,p. Periodic deflating subspaces belonging to other eigenvalues can be computed by reordering the periodic Schur decomposition (Granat et al. 2006a), see also Section The numerical treatment of 2 2 blocks in the upper quasi-triangular matrix T 22,l requires some attention (Van Loan 1973). A periodic eigenvector can be defined as a sequence of nonzero vectors x 1 C n1, x 2 C n2,..., x p C np that spans a periodic deflating subspace, with dim(span(x k )) = 1 for all k. In the case n k = n k+1 whenever s k = 1, an eigenvector can be computed by reordering the corresponding eigenvalue to the top-left corner of the product T sp 22,p T s1 22,1. Alternatively, it can be extracted from the null space of the correspondingly shifted block cyclic embedding, see also (Moravitz Martin and Van Loan 2002). 3. ALGORITHMS AND CORE FUNCTIONALITY The computational approach in our toolbox is based on the periodic Schur decomposition (6) (7). 3.1 Data structures In Matlab, we store matrices in a cell array. A cell array is a general purpose matrix. Each of the elements can contain data of a different type, size and dimension. For cell arrays, storage is allocated dynamically. In particular, for the matrices in (4), we may use a code snippet as A = cell(1,p); for :p, A{k} = A(:,:,k);, end to create the cell array A and store the matrices A 1, A 2,..., A p in the array. In contrast to ordinary three-dimensional arrays, this storage scheme does not waste a lot of memory for sequences with strong variation in the matrix dimensions or where the order of the matrices should be changed on runtime or the period p may be increased or decreased. In our Fortran software, the data structure for storing the matrix sequences is an array of Fortran 90 pointers to ordinary two-dimensional data arrays which has the same storage saving benefits as the cell arrays. The signatures s 1,..., s p {1, 1} are provided in a one-dimensional array s of length p. Alternatively, the user can provide text strings s = ones and s = alter to denote the two commonly encountered situations s k 1 and s k = ( 1) k 1, respectively. By default, s = ones. 3.2 Functions with arguments and results In the following, we present Matlab interfaces to the core functions in our toolbox. As usual, a function call provides input to the function by passing the data in an argument list. Data that
4 needs to be returned to the caller is passed back in a list of return values. Typically, the input and output data are cell arrays with signatures. Braces {... } in the function definitions below identify optional arguments as well as result values. We remark that the transformed cell array T of an argument A has the same signature as A Reduction to a square product [{Q,}T] = per square(a{,s}) Deflates dimension-induced zero and infinite eigenvalues from a general matrix product (4). An orthogonal decomposition of the form (6) (7) is computed with the square matrices T 22,1,..., T 22,p in unreduced form. The orthogonal transformation matrices and the transformed matrices are contained in the cell arrays Q and T, respectively Reduction to periodic HT form [{Q,}T] = per hess(a{,s,l}) Reduces a general matrix product (4) to periodic Hessenberg-triangular (HT) form. An orthogonal decomposition of the form (6) (7) is computed, where T 22,l is upper Hessenberg and T 22,k is upper triangular for all k l. By default, l = 1. apply sequences of orthogonal transformations to T 22,1,..., T 22,p (Granat and Kågström 2006, Granat et al. 2006a). One important step in these backward stable reordering methods is the computation of a solution to a Blocked Almost Block Diagonal (BABD) linear system to high accuracy using structured variants of Gaussian elimination with partial pivoting (GEPP) or QR factorization. Currently we do not plan to support reordering of non-core eigenvalues since we are not aware of any applications calling for such an operation Computation of eigenvalues only e = per eig(a{,s, factor }) Returns a one-dimensional array e of length n 1 that contains the eigenvalues of a general matrix product (4). This contains all n c n 1 core eigenvalues and, possibly, dimension-induced zero and infinite eigenvalues. If the optional input argument factor is provided, the eigenvalues are returned in factorized form, i.e., an n 1 p array e is returned whose row products give the eigenvalues (using prod(e,2)). This allows, e.g., computing the logarithms of eigenvalue magnitudes that would otherwise over- or underflow: sum(log(abs(e)),2) Reduction to periodic Schur form [{Q,}T] = per schur(a{,s,l}) Reduces a general matrix product (4) to periodic Schur form. An orthogonal decomposition of the form (6) (7) is computed, where T 22,l is upper quasi-triangular and T 22,k is upper triangular for all k l. By default, l = Reordering eigenvalues [{Q,}T] = per ordqz({q,}t{,s,l}, slct) Reorders eigenvalues so that a selected cluster of eigenvalues appears in the leading (top-left) diagonal blocks of T 22,1,..., T 22,p in the periodic Schur form T. This is done by orthogonal transformations (see below) and the associated periodic deflating subspace is spanned by the corresponding leading columns of Q. The logical vector slct specifies the selected cluster as they appear along the diagonals of the input T 22,1,..., T 22,p. slct can also be a string variable that defines a region of the complex plane, e.g., slct = udi corresponds to all eigenvalues inside the unit disk. Periodic eigenvalue reordering is performed by generalizations of the LAPACK methodology (Bai and Demmel 1993, Kågström and Poromaa 1996) to general products of matrices of the form (4). First, the product is recasted on the exponent structure (2) by insertion of identity matrices. Then, we solve periodic matrix equations and 3.3 Complex matrices If any of the matrices in the cell array A happens to be complex, the Matlab function listed above return complex variants of the periodic decompositions: Q becomes unitary, the transpose in (6) becomes the complex conjugate transpose, and the quasi-triangular matrix T l,22 becomes triangular. 3.4 Additional functionality Additional functionality and tools will be provided, including the solution of various periodic Sylvester-type matrix equations (Varga and Van Dooren 2001, Granat et al. 2006c). The demand of such solvers appear in different control applications but are also useful for condition estimation and error bounds of computed quantities (eigenvalues, and periodic deflating subspaces). Relevant perturbation theory for product eigenvalue problems are presented in (Lin and Sun 2001, Benner et al. 2002b, Sun 2005). 4. THE MATRIX PRODUCT CLASS Matlab classes provide a convenient way to hide the technical complexity imposed by the generality of periodic eigenvalue problems. We have
5 designed a class product that contains the cell and signature arrays A, s, and provides several operations for creating and manipulating objects. 4.1 Construction of a product object The simplest way to construct a product object is to type p = product(a1), where A1 is a twodimensional array. This corresponds to a product of length 1: Π = A 1. Products of longer lengths can be created using the overloaded operations * and inv. For example, p = inv(product(a2)) * product(a1) corresponds to a (formal) product of length 2: Π = A 1 2 A 1. It is important to note that the inversion and multiplication are never carried out explicitly for product objects. In particular, this admits formal inverses of rectangular matrices. 4.2 Main computational functions [{Q,} pt] = square({q,} p) [{Q,} pt] = hess({q,} p) [{Q,} pt] = schur({q,} p) [{Q,} pt] = ordqz({q,} p, slct) The functions listed above are product class equivalents of the Matlab functions described in Sections Q is a cell array containing the orthogonal transformation matrices and T is a product object containing the transformed product Tp sp T sp 1 p 1 T s1 1. The function e = eig(p{, factor }) returns the eigenvalues of the product corresponding to p as described in Section Utilities In the following, we assume that p is a product object corresponding to the matrix product A sp p A sp 1 p 1 As1 1. p transposes the product. p1 + p2 adds the coefficients of two product objects having the same dimensions. p * X for a cell array X computes a product with new transformed coefficients A k X k (if s k = 1) and X k A k (if s k = 1). X * p for a cell array X computes a product with new transformed coefficients X k+1 A k (if s k = 1) and A k X k+1 (if s k = 1). lift(p) returns a lifted, block cyclic representation of the form (3). norm(p, nrm) returns a vector containing the norms of the coefficients. Which norm is controlled by nrm, e.g., nrm = fro corresponds to the Frobenius norm. prod(p) explicitly forms the matrix product if possible. dim(p) returns a vector of length p+1 containing the matrix dimensions n 1, n 2,..., n p+1. signature(p) returns a vector of length p containing the matrix signatures s 1, s 2,..., s p with s k { 1, +1}. 5. CASE STUDY: SOLUTION OF PERIODIC RICCATI EQUATIONS The unique symmetric positive semidefinite solution X k = X k+p of the discrete-time periodic Riccati equation (DPRE) (Bittanti et al. 1991, Hench and Laub 1994) 0 = C T k H k C k E T k 1X k E k 1 + A T k X k+1 A k A T k X k+1 B k (N k + B T k X k+1 B k ) 1 B T k X k+1 A k, (8) can be computed robustly in terms of an associated periodic deflating subspace problem. The DPRE (8) appears in the LQ optimal control problem associated with the system (1) where all E k are nonsingular, and the weighting matrices of the cost functional are p-periodic, i.e., H k+p = H k (= Hk T 0) and N k+p = N k (= Nk T > 0). Similarly as for the case E k = I n (Hench and Laub 1994), it can be shown that the 2n 2n matrix product Mp 1 L p Mp 1 1 L p 1 M1 1 L 1 (9) with [ ] A L k = k 0 Ck T H k C k Ek 1 T, M k = [ Ek 1 B k N 1 k 0 A T k BT k has exactly n eigenvalues inside the unit disk under the reasonable assumption that (1) is d- stabilizable and d-detectable. By reordering the periodic Schur form of the matrix product (9) we can compute a periodic deflating subspace defined by the orthogonal matrices U k, V k R 2n 2n with V k+p = V k such that [ ] Uk T S (k) L k V k = 11 S(k) 12 0 S (k), Uk T M k V k+1 = 22 where the n n matrix product T (p) 1 (p) 11 S 11 T (2) 1 (2) 11 S 11 T (1) 1 (1) 11 S 11 ] [ ] T (k) 11 T (k) 12 0 T (k), 22 contains all eigenvalues inside the unit disk. If we partition [ ] U k =
6 with ij R n n, then [ ] 1 = Xk E k 1, from which X k can be computed. Using our Matlab tools and assuming that we have defined cell arrays L and M corresponding to L k and M k above, the solution sequence X k for k = 1,..., p can be computed as: for k = 1:p, A{2*k-1} = L{k}; A{2*k} = M{k}; end pa = product((-1).^[1:2*p], A); [Q, pt] = schur(pa); [Q, pt] = ordqz(q, pt, udi ); X = cell(1,p); for k = 1:p, X{k} = Q{2*k-1}(n+1:2*n,1:n) /... Q{2*k-1}(1:n,1:n) / L{k}(1:n,1:n); end REFERENCES Bai, Z. and J. W. Demmel (1993). On swapping diagonal blocks in real Schur form. Linear Algebra Appl. 186, Benner, P. and R. Byers (2001). Evaluating products of matrix pencils and collapsing matrix products. Numerical Linear Algebra with Applications 8, Benner, P., R. Byers, V. Mehrmann and H. Xu (2002a). Numerical computation of deflating subspaces of skew-hamiltonian/hamiltonian pencils. SIAM J. Matrix Anal. Appl. 24(1), Benner, P., V. Mehrmann and H. Xu (2002b). Perturbation analysis for the eigenvalue problem of a formal product of matrices. BIT 42(1), Bittanti, S., P. Colaneri and G. De Nicolao (1991). The periodic Riccati equation. In: The Riccati Equation (S. Bittanti, A. J. Laub and J. C. Willems, Eds.). pp Springer-Verlag. Berlin, Heidelberg, Germany. Bojanczyk, A., G. H. Golub and P. Van Dooren (1992). The periodic Schur decomposition; algorithm and applications. In: Proc. SPIE Conference. Vol pp Flamm, D. S. (1991). A new shift-invariant representation for periodic linear systems. Systems Control Lett. 17(1), Granat, R. and B. Kågström (2006). Direct Eigenvalue Reordering in a Product of Matrices in Periodic Schur Form. SIAM J. Matrix Anal. Appl. 28(1), Granat, R., B. Kågström and D. Kressner (2006a). Computing Periodic Deflating Subspaces Associated with a Specified Set of Eigenvalues. BIT Numerical Mathematics (submitted). Granat, R., B. Kågström and D. Kressner (2006b). Reordering the Eigenvalues of a Periodic Matrix Pair with Applications in Control. In: Proc. of 2006 IEEE Conference on Computer Aided Control Systems Design (CACSD). pp ISBN: Granat, R., I. Jonsson and B. Kågström (2006c). Recursive Blocked Algorithms for Solving Periodic Triangular Sylvester-type Matrix Equations. In: Proc. of PARA 06: State-ofthe-art in Scientific and Parallel Computing. LNCS, Springer (to appear). Hench, J. J. and A. J. Laub (1994). Numerical solution of the discrete-time periodic Riccati equation. IEEE Trans. Automat. Control 39(6), Higham, N. J. (2002). Accuracy and Stability of Numerical Algorithms. second ed.. SIAM. Philadelphia, PA. Kågström, B. and P. Poromaa (1996). Computing eigenspaces with specified eigenvalues of a regular matrix pair (A, B) and condition estimation: theory, algorithms and software. Numer. Algorithms 12(3-4), Kressner, D. (2006). The periodic QR algorithm is a disguised QR algorithm. Linear Algebra Appl., 417(2-3): Lin, W.-W. and J.-G. Sun (2001). Perturbation analysis for the eigenproblem of periodic matrix pairs. Linear Algebra Appl. 337, Moler, C. B. and G. W. Stewart (1973). An algorithm for generalized matrix eigenvalue problems. SIAM J. Numer. Anal. 10, Moravitz Martin, C. D. and Van Loan, C. F. (2002). Product triangular systems with shift. SIAM J. Matrix Anal. Appl. 24, Sergeichuk, V. V. (2004). Computation of canonical matrices for chains and cycles of linear mappings. Linear Algebra Appl. 376, Sun, J.-G. (2005). Perturbation bounds for subspaces associated with periodic eigenproblems. Taiwanese J. of Mathematics 9(1), Van Dooren, P. (1999). Two point boundary value and periodic eigenvalue problems. In: Proceedings IEEE CACSD conference. Van Loan, C. F. (1973). Generalized Singular Values with Algorithms and Applications. PhD thesis. The University of Michigan. Varga, A. and P. Van Dooren (2001). Computational methods for periodic systems - an overview. In: Proc. of IFAC Workshop on Periodic Control Systems, Como, Italy. pp
1. Introduction. Applying the QR algorithm to a real square matrix A yields a decomposition of the form
BLOCK ALGORITHMS FOR REORDERING STANDARD AND GENERALIZED SCHUR FORMS LAPACK WORKING NOTE 171 DANIEL KRESSNER Abstract. Block algorithms for reordering a selected set of eigenvalues in a standard or generalized
An SVD-Like Matrix Decomposition and Its Applications
An SVD-Like Matrix Decomposition and Its Applications Hongguo Xu Abstract [ A matrix S C m m is symplectic if SJS = J, where J = I m I m. Symplectic matrices play an important role in the analysis and
ANONSINGULAR tridiagonal linear system of the form
Generalized Diagonal Pivoting Methods for Tridiagonal Systems without Interchanges Jennifer B. Erway, Roummel F. Marcia, and Joseph A. Tyson Abstract It has been shown that a nonsingular symmetric tridiagonal
Properties of Matrices and Operations on Matrices
Properties of Matrices and Operations on Matrices A common data structure for statistical analysis is a rectangular array or matris. Rows represent individual observational units, or just observations,
Lecture 3: QR-Factorization
Lecture 3: QR-Factorization This lecture introduces the Gram Schmidt orthonormalization process and the associated QR-factorization of matrices It also outlines some applications of this factorization
Orthogonal iteration to QR
Week 10: Wednesday and Friday, Oct 24 and 26 Orthogonal iteration to QR On Monday, we went through a somewhat roundabout algbraic path from orthogonal subspace iteration to the QR iteration. Let me start
Eigen-solving via reduction to DPR1 matrices
Computers and Mathematics with Applications 56 (2008) 166 171 www.elsevier.com/locate/camwa Eigen-solving via reduction to DPR1 matrices V.Y. Pan a,, B. Murphy a, R.E. Rosholt a, Y. Tang b, X. Wang c,
CHAPTER 6. Direct Methods for Solving Linear Systems
CHAPTER 6 Direct Methods for Solving Linear Systems. Introduction A direct method for approximating the solution of a system of n linear equations in n unknowns is one that gives the exact solution to
Institute for Advanced Computer Studies. Department of Computer Science. Two Algorithms for the The Ecient Computation of
University of Maryland Institute for Advanced Computer Studies Department of Computer Science College Park TR{98{12 TR{3875 Two Algorithms for the The Ecient Computation of Truncated Pivoted QR Approximations
Maths for Signals and Systems Linear Algebra for Engineering Applications
Maths for Signals and Systems Linear Algebra for Engineering Applications Lectures 1-2, Tuesday 11 th October 2016 DR TANIA STATHAKI READER (ASSOCIATE PROFFESOR) IN SIGNAL PROCESSING IMPERIAL COLLEGE LONDON
EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 2
EE/ACM 150 - Applications of Convex Optimization in Signal Processing and Communications Lecture 2 Andre Tkacenko Signal Processing Research Group Jet Propulsion Laboratory April 5, 2012 Andre Tkacenko
Contents 1 Introduction and Preliminaries 1 Embedding of Extended Matrix Pencils 3 Hamiltonian Triangular Forms 1 4 Skew-Hamiltonian/Hamiltonian Matri
Technische Universitat Chemnitz Sonderforschungsbereich 393 Numerische Simulation auf massiv parallelen Rechnern Peter Benner Ralph Byers Volker Mehrmann Hongguo Xu Numerical Computation of Deating Subspaces
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 1 x 2. x n 8 (4) 3 4 2
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS SYSTEMS OF EQUATIONS AND MATRICES Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a
FINITE-DIMENSIONAL LINEAR ALGEBRA
DISCRETE MATHEMATICS AND ITS APPLICATIONS Series Editor KENNETH H ROSEN FINITE-DIMENSIONAL LINEAR ALGEBRA Mark S Gockenbach Michigan Technological University Houghton, USA CRC Press Taylor & Francis Croup
Lecture 7 Spectral methods
CSE 291: Unsupervised learning Spring 2008 Lecture 7 Spectral methods 7.1 Linear algebra review 7.1.1 Eigenvalues and eigenvectors Definition 1. A d d matrix M has eigenvalue λ if there is a d-dimensional
Image Registration Lecture 2: Vectors and Matrices
Image Registration Lecture 2: Vectors and Matrices Prof. Charlene Tsai Lecture Overview Vectors Matrices Basics Orthogonal matrices Singular Value Decomposition (SVD) 2 1 Preliminary Comments Some of this
Using semiseparable matrices to compute the SVD of a general matrix product/quotient
Using semiseparable matrices to compute the SVD of a general matrix product/quotient Marc Van Barel Yvette Vanberghen Paul Van Dooren Report TW 508, November 007 n Katholieke Universiteit Leuven Department
IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET
IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET This is a (not quite comprehensive) list of definitions and theorems given in Math 1553. Pay particular attention to the ones in red. Study Tip For each
Complementary bases in symplectic matrices and a proof that their determinant is one
Linear Algebra and its Applications 419 (2006) 772 778 www.elsevier.com/locate/laa Complementary bases in symplectic matrices and a proof that their determinant is one Froilán M. Dopico a,, Charles R.
Parallel Singular Value Decomposition. Jiaxing Tan
Parallel Singular Value Decomposition Jiaxing Tan Outline What is SVD? How to calculate SVD? How to parallelize SVD? Future Work What is SVD? Matrix Decomposition Eigen Decomposition A (non-zero) vector
L3: Review of linear algebra and MATLAB
L3: Review of linear algebra and MATLAB Vector and matrix notation Vectors Matrices Vector spaces Linear transformations Eigenvalues and eigenvectors MATLAB primer CSCE 666 Pattern Analysis Ricardo Gutierrez-Osuna
Definition (T -invariant subspace) Example. Example
Eigenvalues, Eigenvectors, Similarity, and Diagonalization We now turn our attention to linear transformations of the form T : V V. To better understand the effect of T on the vector space V, we begin
An error estimate for matrix equations
Applied Numerical Mathematics 50 (2004) 395 407 www.elsevier.com/locate/apnum An error estimate for matrix equations Yang Cao, Linda Petzold 1 Department of Computer Science, University of California,
Math 18, Linear Algebra, Lecture C00, Spring 2017 Review and Practice Problems for Final Exam
Math 8, Linear Algebra, Lecture C, Spring 7 Review and Practice Problems for Final Exam. The augmentedmatrix of a linear system has been transformed by row operations into 5 4 8. Determine if the system
NONCOMMUTATIVE POLYNOMIAL EQUATIONS. Edward S. Letzter. Introduction
NONCOMMUTATIVE POLYNOMIAL EQUATIONS Edward S Letzter Introduction My aim in these notes is twofold: First, to briefly review some linear algebra Second, to provide you with some new tools and techniques
Linear Algebra. Carleton DeTar February 27, 2017
Linear Algebra Carleton DeTar detar@physics.utah.edu February 27, 2017 This document provides some background for various course topics in linear algebra: solving linear systems, determinants, and finding
EIGIFP: A MATLAB Program for Solving Large Symmetric Generalized Eigenvalue Problems
EIGIFP: A MATLAB Program for Solving Large Symmetric Generalized Eigenvalue Problems JAMES H. MONEY and QIANG YE UNIVERSITY OF KENTUCKY eigifp is a MATLAB program for computing a few extreme eigenvalues
Singular Value Decomposition
Chapter 6 Singular Value Decomposition In Chapter 5, we derived a number of algorithms for computing the eigenvalues and eigenvectors of matrices A R n n. Having developed this machinery, we complete our
w T 1 w T 2. w T n 0 if i j 1 if i = j
Lyapunov Operator Let A F n n be given, and define a linear operator L A : C n n C n n as L A (X) := A X + XA Suppose A is diagonalizable (what follows can be generalized even if this is not possible -
Orthogonal iteration revisited
Week 10 11: Friday, Oct 30 and Monday, Nov 2 Orthogonal iteration revisited Last time, we described a generalization of the power methods to compute invariant subspaces. That is, starting from some initial
Algorithms to solve block Toeplitz systems and. least-squares problems by transforming to Cauchy-like. matrices
Algorithms to solve block Toeplitz systems and least-squares problems by transforming to Cauchy-like matrices K. Gallivan S. Thirumalai P. Van Dooren 1 Introduction Fast algorithms to factor Toeplitz matrices
Matrices. Chapter What is a Matrix? We review the basic matrix operations. An array of numbers a a 1n A = a m1...
Chapter Matrices We review the basic matrix operations What is a Matrix? An array of numbers a a n A = a m a mn with m rows and n columns is a m n matrix Element a ij in located in position (i, j The elements
ECS231: Spectral Partitioning. Based on Berkeley s CS267 lecture on graph partition
ECS231: Spectral Partitioning Based on Berkeley s CS267 lecture on graph partition 1 Definition of graph partitioning Given a graph G = (N, E, W N, W E ) N = nodes (or vertices), E = edges W N = node weights
Boolean Inner-Product Spaces and Boolean Matrices
Boolean Inner-Product Spaces and Boolean Matrices Stan Gudder Department of Mathematics, University of Denver, Denver CO 80208 Frédéric Latrémolière Department of Mathematics, University of Denver, Denver
18.06 Problem Set 8 - Solutions Due Wednesday, 14 November 2007 at 4 pm in
806 Problem Set 8 - Solutions Due Wednesday, 4 November 2007 at 4 pm in 2-06 08 03 Problem : 205+5+5+5 Consider the matrix A 02 07 a Check that A is a positive Markov matrix, and find its steady state
On the mathematical background of Google PageRank algorithm
Working Paper Series Department of Economics University of Verona On the mathematical background of Google PageRank algorithm Alberto Peretti, Alberto Roveda WP Number: 25 December 2014 ISSN: 2036-2919
Quick Introduction to Nonnegative Matrix Factorization
Quick Introduction to Nonnegative Matrix Factorization Norm Matloff University of California at Davis 1 The Goal Given an u v matrix A with nonnegative elements, we wish to find nonnegative, rank-k matrices
Conjugate Gradient (CG) Method
Conjugate Gradient (CG) Method by K. Ozawa 1 Introduction In the series of this lecture, I will introduce the conjugate gradient method, which solves efficiently large scale sparse linear simultaneous
Reflections in Hilbert Space III: Eigen-decomposition of Szegedy s operator
Reflections in Hilbert Space III: Eigen-decomposition of Szegedy s operator James Daniel Whitfield March 30, 01 By three methods we may learn wisdom: First, by reflection, which is the noblest; second,
5.6. PSEUDOINVERSES 101. A H w.
5.6. PSEUDOINVERSES 0 Corollary 5.6.4. If A is a matrix such that A H A is invertible, then the least-squares solution to Av = w is v = A H A ) A H w. The matrix A H A ) A H is the left inverse of A and
ETNA Kent State University
Electronic Transactions on Numerical Analysis Volume 38, pp 75-30, 011 Copyright 011, ISSN 1068-9613 ETNA PERTURBATION ANALYSIS FOR COMPLEX SYMMETRIC, SKEW SYMMETRIC, EVEN AND ODD MATRIX POLYNOMIALS SK
GATE Engineering Mathematics SAMPLE STUDY MATERIAL. Postal Correspondence Course GATE. Engineering. Mathematics GATE ENGINEERING MATHEMATICS
SAMPLE STUDY MATERIAL Postal Correspondence Course GATE Engineering Mathematics GATE ENGINEERING MATHEMATICS ENGINEERING MATHEMATICS GATE Syllabus CIVIL ENGINEERING CE CHEMICAL ENGINEERING CH MECHANICAL
The Nearest Doubly Stochastic Matrix to a Real Matrix with the same First Moment
he Nearest Doubly Stochastic Matrix to a Real Matrix with the same First Moment William Glunt 1, homas L. Hayden 2 and Robert Reams 2 1 Department of Mathematics and Computer Science, Austin Peay State
Matrices, Vector Spaces, and Information Retrieval
Matrices, Vector Spaces, and Information Authors: M. W. Berry and Z. Drmac and E. R. Jessup SIAM 1999: Society for Industrial and Applied Mathematics Speaker: Mattia Parigiani 1 Introduction Large volumes
Inverse eigenvalue problems involving multiple spectra
Inverse eigenvalue problems involving multiple spectra G.M.L. Gladwell Department of Civil Engineering University of Waterloo Waterloo, Ontario, Canada N2L 3G1 ggladwell@uwaterloo.ca URL: http://www.civil.uwaterloo.ca/ggladwell
I = i 0,
Special Types of Matrices Certain matrices, such as the identity matrix 0 0 0 0 0 0 I = 0 0 0, 0 0 0 have a special shape, which endows the matrix with helpful properties The identity matrix is an example
EIGENVALUES AND EIGENVECTORS
EIGENVALUES AND EIGENVECTORS Diagonalizable linear transformations and matrices Recall, a matrix, D, is diagonal if it is square and the only non-zero entries are on the diagonal This is equivalent to
Linear Algebra Final Exam Study Guide Solutions Fall 2012
. Let A = Given that v = 7 7 67 5 75 78 Linear Algebra Final Exam Study Guide Solutions Fall 5 explain why it is not possible to diagonalize A. is an eigenvector for A and λ = is an eigenvalue for A diagonalize
MTH50 Spring 07 HW Assignment 7 {From [FIS0]}: Sec 44 #4a h 6; Sec 5 #ad ac 4ae 4 7 The due date for this assignment is 04/05/7 Sec 44 #4a h Evaluate the erminant of the following matrices by any legitimate
Math 25a Practice Final #1 Solutions
Math 25a Practice Final #1 Solutions Problem 1. Suppose U and W are subspaces of V such that V = U W. Suppose also that u 1,..., u m is a basis of U and w 1,..., w n is a basis of W. Prove that is a basis
Numerical behavior of inexact linear solvers
Numerical behavior of inexact linear solvers Miro Rozložník joint results with Zhong-zhi Bai and Pavel Jiránek Institute of Computer Science, Czech Academy of Sciences, Prague, Czech Republic The fourth
Accelerating linear algebra computations with hybrid GPU-multicore systems.
Accelerating linear algebra computations with hybrid GPU-multicore systems. Marc Baboulin INRIA/Université Paris-Sud joint work with Jack Dongarra (University of Tennessee and Oak Ridge National Laboratory)
BUMPLESS SWITCHING CONTROLLERS. William A. Wolovich and Alan B. Arehart 1. December 27, Abstract
BUMPLESS SWITCHING CONTROLLERS William A. Wolovich and Alan B. Arehart 1 December 7, 1995 Abstract This paper outlines the design of bumpless switching controllers that can be used to stabilize MIMO plants
Appendix C Vector and matrix algebra
Appendix C Vector and matrix algebra Concepts Scalars Vectors, rows and columns, matrices Adding and subtracting vectors and matrices Multiplying them by scalars Products of vectors and matrices, scalar
Eigenvalues and Eigenvectors
Eigenvalues and Eigenvectors Philippe B. Laval KSU Fall 2015 Philippe B. Laval (KSU) Eigenvalues and Eigenvectors Fall 2015 1 / 14 Introduction We define eigenvalues and eigenvectors. We discuss how to
KEYWORDS. Numerical methods, generalized singular values, products of matrices, quotients of matrices. Introduction The two basic unitary decompositio
COMPUTING THE SVD OF A GENERAL MATRIX PRODUCT/QUOTIENT GENE GOLUB Computer Science Department Stanford University Stanford, CA USA golub@sccm.stanford.edu KNUT SLNA SC-CM Stanford University Stanford,
where J " 0 I n?i n 0 # () and I n is the nn identity matrix (Note that () is in general not equivalent to L T JL N T JN) In most applications system-
The Symplectic Eigenvalue roblem, the Buttery Form, the SR Algorithm, and the Lanczos Method eter Benner Heike Fabender y August 1, 199 Abstract We discuss some aspects of the recently proposed symplectic
Block-tridiagonal matrices
Block-tridiagonal matrices. p.1/31 Block-tridiagonal matrices - where do these arise? - as a result of a particular mesh-point ordering - as a part of a factorization procedure, for example when we compute
and let s calculate the image of some vectors under the transformation T.
Chapter 5 Eigenvalues and Eigenvectors 5. Eigenvalues and Eigenvectors Let T : R n R n be a linear transformation. Then T can be represented by a matrix (the standard matrix), and we can write T ( v) =
Gaussian Elimination without/with Pivoting and Cholesky Decomposition
Gaussian Elimination without/with Pivoting and Cholesky Decomposition Gaussian Elimination WITHOUT pivoting Notation: For a matrix A R n n we define for k {,,n} the leading principal submatrix a a k A
Topic 2 Quiz 2. choice C implies B and B implies C. correct-choice C implies B, but B does not imply C
Topic 1 Quiz 1 text A reduced row-echelon form of a 3 by 4 matrix can have how many leading one s? choice must have 3 choice may have 1, 2, or 3 correct-choice may have 0, 1, 2, or 3 choice may have 0,
A note on companion pencils
Contemporary Mathematics A note on companion pencils Jared L Aurentz, Thomas Mach, Raf Vandebril, and David S Watkins Abstract Various generalizations of companion matrices to companion pencils are presented
On the -Sylvester Equation AX ± X B = C
Department of Applied Mathematics National Chiao Tung University, Taiwan A joint work with E. K.-W. Chu and W.-W. Lin Oct., 20, 2008 Outline 1 Introduction 2 -Sylvester Equation 3 Numerical Examples 4
Spectrally arbitrary star sign patterns
Linear Algebra and its Applications 400 (2005) 99 119 wwwelseviercom/locate/laa Spectrally arbitrary star sign patterns G MacGillivray, RM Tifenbach, P van den Driessche Department of Mathematics and Statistics,
Math 240 Calculus III
The Calculus III Summer 2015, Session II Wednesday, July 8, 2015 Agenda 1. of the determinant 2. determinants 3. of determinants What is the determinant? Yesterday: Ax = b has a unique solution when A
H 2 -optimal model reduction of MIMO systems
H 2 -optimal model reduction of MIMO systems P. Van Dooren K. A. Gallivan P.-A. Absil Abstract We consider the problem of approximating a p m rational transfer function Hs of high degree by another p m
MATH 304 Linear Algebra Lecture 10: Linear independence. Wronskian.
MATH 304 Linear Algebra Lecture 10: Linear independence. Wronskian. Spanning set Let S be a subset of a vector space V. Definition. The span of the set S is the smallest subspace W V that contains S. If
216 S. Chandrasearan and I.C.F. Isen Our results dier from those of Sun [14] in two asects: we assume that comuted eigenvalues or singular values are
Numer. Math. 68: 215{223 (1994) Numerische Mathemati c Sringer-Verlag 1994 Electronic Edition Bacward errors for eigenvalue and singular value decomositions? S. Chandrasearan??, I.C.F. Isen??? Deartment
Google PageRank. Francesco Ricci Faculty of Computer Science Free University of Bozen-Bolzano
Google PageRank Francesco Ricci Faculty of Computer Science Free University of Bozen-Bolzano fricci@unibz.it 1 Content p Linear Algebra p Matrices p Eigenvalues and eigenvectors p Markov chains p Google
RANA03-02 January Jacobi-Davidson methods and preconditioning with applications in pole-zero analysis
RANA03-02 January 2003 Jacobi-Davidson methods and preconditioning with applications in pole-zero analysis by J.Rommes, H.A. van der Vorst, EJ.W. ter Maten Reports on Applied and Numerical Analysis Department
which is an important tool in measuring invariant subspace sensitivity [10, sec. 7..5], [7, 8]. Here, we are using the Frobenius norm, kak F = ( P i;j
PERTURBATION THEORY AND BACKWARD ERROR FOR AX? XB = C NICHOLAS J. HIGHAM y Abstract. Because of the special structure of the equations AX?XB = C the usual relation for linear equations \backward error
c c c c c c c c c c a 3x3 matrix C= has a determinant determined by
Linear Algebra Determinants and Eigenvalues Introduction: Many important geometric and algebraic properties of square matrices are associated with a single real number revealed by what s known as the determinant.
Filter Design for Linear Time Delay Systems
IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 49, NO. 11, NOVEMBER 2001 2839 ANewH Filter Design for Linear Time Delay Systems E. Fridman Uri Shaked, Fellow, IEEE Abstract A new delay-dependent filtering
Math 20F Practice Final Solutions. Jor-el Briones
Math 2F Practice Final Solutions Jor-el Briones Jor-el Briones / Math 2F Practice Problems for Final Page 2 of 6 NOTE: For the solutions to these problems, I skip all the row reduction calculations. Please
LECTURES 4/5: SYSTEMS OF LINEAR EQUATIONS
LECTURES 4/5: SYSTEMS OF LINEAR EQUATIONS MA1111: LINEAR ALGEBRA I, MICHAELMAS 2016 1 Linear equations We now switch gears to discuss the topic of solving linear equations, and more interestingly, systems
MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)
MATH 20F: LINEAR ALGEBRA LECTURE B00 (T KEMP) Definition 01 If T (x) = Ax is a linear transformation from R n to R m then Nul (T ) = {x R n : T (x) = 0} = Nul (A) Ran (T ) = {Ax R m : x R n } = {b R m
An LQ R weight selection approach to the discrete generalized H 2 control problem
INT. J. CONTROL, 1998, VOL. 71, NO. 1, 93± 11 An LQ R weight selection approach to the discrete generalized H 2 control problem D. A. WILSON², M. A. NEKOUI² and G. D. HALIKIAS² It is known that a generalized
VERSAL DEFORMATIONS OF BILINEAR SYSTEMS UNDER OUTPUT-INJECTION EQUIVALENCE
PHYSCON 2013 San Luis Potosí México August 26 August 29 2013 VERSAL DEFORMATIONS OF BILINEAR SYSTEMS UNDER OUTPUT-INJECTION EQUIVALENCE M Isabel García-Planas Departamento de Matemàtica Aplicada I Universitat
An Introduction To Linear Algebra. Kuttler
An Introduction To Linear Algebra Kuttler April, 7 Contents Introduction 7 F n 9 Outcomes 9 Algebra in F n Systems Of Equations Outcomes Systems Of Equations, Geometric Interpretations Systems Of Equations,
8.1 Bifurcations of Equilibria
1 81 Bifurcations of Equilibria Bifurcation theory studies qualitative changes in solutions as a parameter varies In general one could study the bifurcation theory of ODEs PDEs integro-differential equations
1 Determinants. 1.1 Determinant
1 Determinants [SB], Chapter 9, p.188-196. [SB], Chapter 26, p.719-739. Bellow w ll study the central question: which additional conditions must satisfy a quadratic matrix A to be invertible, that is to
Tangent spaces, normals and extrema
Chapter 3 Tangent spaces, normals and extrema If S is a surface in 3-space, with a point a S where S looks smooth, i.e., without any fold or cusp or self-crossing, we can intuitively define the tangent
Stable iterations for the matrix square root
Numerical Algorithms 5 (997) 7 4 7 Stable iterations for the matrix square root Nicholas J. Higham Department of Mathematics, University of Manchester, Manchester M3 9PL, England E-mail: higham@ma.man.ac.u
c 2002 Society for Industrial and Applied Mathematics
SIAM J. MATRIX ANAL. APPL. Vol. 24, No. 1, pp. 150 164 c 2002 Society for Industrial and Applied Mathematics VARIANTS OF THE GREVILLE FORMULA WITH APPLICATIONS TO EXACT RECURSIVE LEAST SQUARES JIE ZHOU,
3 QR factorization revisited
LINEAR ALGEBRA: NUMERICAL METHODS. Version: August 2, 2000 30 3 QR factorization revisited Now we can explain why A = QR factorization is much better when using it to solve Ax = b than the A = LU factorization
Kernel Methods. Machine Learning A W VO
Kernel Methods Machine Learning A 708.063 07W VO Outline 1. Dual representation 2. The kernel concept 3. Properties of kernels 4. Examples of kernel machines Kernel PCA Support vector regression (Relevance
HOMEWORK PROBLEMS FROM STRANG S LINEAR ALGEBRA AND ITS APPLICATIONS (4TH EDITION)
HOMEWORK PROBLEMS FROM STRANG S LINEAR ALGEBRA AND ITS APPLICATIONS (4TH EDITION) PROFESSOR STEVEN MILLER: BROWN UNIVERSITY: SPRING 2007 1. CHAPTER 1: MATRICES AND GAUSSIAN ELIMINATION Page 9, # 3: Describe
Institute for Advanced Computer Studies. Department of Computer Science. On the Perturbation of. LU and Cholesky Factors. G. W.
University of Maryland Institute for Advanced Computer Studies Department of Computer Science College Park TR{95{93 TR{3535 On the Perturbation of LU and Cholesky Factors G. W. Stewart y October, 1995
arxiv: v1 [math.rt] 7 Oct 2014
A direct approach to the rational normal form arxiv:1410.1683v1 [math.rt] 7 Oct 2014 Klaus Bongartz 8. Oktober 2014 In courses on linear algebra the rational normal form of a matrix is usually derived
ABSTRACT ALGEBRA WITH APPLICATIONS
ABSTRACT ALGEBRA WITH APPLICATIONS IN TWO VOLUMES VOLUME I VECTOR SPACES AND GROUPS KARLHEINZ SPINDLER Darmstadt, Germany Marcel Dekker, Inc. New York Basel Hong Kong Contents f Volume I Preface v VECTOR
18.06 Problem Set 2 Solution
18.06 Problem Set 2 Solution Total: 100 points Section 2.5. Problem 24: Use Gauss-Jordan elimination on [U I] to find the upper triangular U 1 : 1 a b 1 0 UU 1 = I 0 1 c x 1 x 2 x 3 = 0 1 0. 0 0 1 0 0
Clifford Algebras and Spin Groups
Clifford Algebras and Spin Groups Math G4344, Spring 2012 We ll now turn from the general theory to examine a specific class class of groups: the orthogonal groups. Recall that O(n, R) is the group of
Poisson Solvers. William McLean. April 21, Return to Math3301/Math5315 Common Material.
Poisson Solvers William McLean April 21, 2004 Return to Math3301/Math5315 Common Material 1 Introduction Many problems in applied mathematics lead to a partial differential equation of the form a 2 u +
SENSITIVITY OF THE STATIONARY DISTRIBUTION OF A MARKOV CHAIN*
SIAM J Matrix Anal Appl c 1994 Society for Industrial and Applied Mathematics Vol 15, No 3, pp 715-728, July, 1994 001 SENSITIVITY OF THE STATIONARY DISTRIBUTION OF A MARKOV CHAIN* CARL D MEYER Abstract
11 Adjoint and Self-adjoint Matrices
11 Adjoint and Self-adjoint Matrices In this chapter, V denotes a finite dimensional inner product space (unless stated otherwise). 11.1 Theorem (Riesz representation) Let f V, i.e. f is a linear functional
Fast Matrix Multiplication*
mathematics of computation, volume 28, number 125, January, 1974 Triangular Factorization and Inversion by Fast Matrix Multiplication* By James R. Bunch and John E. Hopcroft Abstract. The fast matrix multiplication
Scientific Computing: An Introductory Survey
Scientific Computing: An Introductory Survey Chapter 7 Interpolation Prof. Michael T. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c 2002. Reproduction permitted
TOPIC III LINEAR ALGEBRA
[1] Linear Equations TOPIC III LINEAR ALGEBRA (1) Case of Two Endogenous Variables 1) Linear vs. Nonlinear Equations Linear equation: ax + by = c, where a, b and c are constants. 2 Nonlinear equation:
Quasi-Newton Methods
Quasi-Newton Methods Werner C. Rheinboldt These are excerpts of material relating to the boos [OR00 and [Rhe98 and of write-ups prepared for courses held at the University of Pittsburgh. Some further references
Block companion matrices, discrete-time block diagonal stability and polynomial matrices
Block companion matrices, discrete-time block diagonal stability and polynomial matrices Harald K. Wimmer Mathematisches Institut Universität Würzburg D-97074 Würzburg Germany October 25, 2008 Abstract