On the solution of large Sylvester-observer equations

Similar documents
Tikhonov Regularization of Large Symmetric Problems

Augmented GMRES-type methods

ON THE REGULARIZING PROPERTIES OF THE GMRES METHOD

Krylov subspace iterative methods for nonsymmetric discrete ill-posed problems in image restoration

Partial Eigenvalue Assignment in Linear Systems: Existence, Uniqueness and Numerical Solution

On Solving Large Algebraic. Riccati Matrix Equations

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH

The restarted QR-algorithm for eigenvalue computation of structured matrices

The Lanczos and conjugate gradient algorithms

A MODIFIED TSVD METHOD FOR DISCRETE ILL-POSED PROBLEMS

Computational Methods for Feedback Control in Damped Gyroscopic Second-order Systems 1

AN ITERATIVE METHOD WITH ERROR ESTIMATORS

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES

RITZ VALUE BOUNDS THAT EXPLOIT QUASI-SPARSITY

PARAMETERIZATION OF STATE FEEDBACK GAINS FOR POLE PLACEMENT

ECS231 Handout Subspace projection methods for Solving Large-Scale Eigenvalue Problems. Part I: Review of basic theory of eigenvalue problems

Algorithms that use the Arnoldi Basis

Arnoldi Methods in SLEPc

The quadratic eigenvalue problem (QEP) is to find scalars λ and nonzero vectors u satisfying

Krylov subspace projection methods

A New Block Algorithm for Full-Rank Solution of the Sylvester-observer Equation.

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

QUALITATIVE CONTROLLABILITY AND UNCONTROLLABILITY BY A SINGLE ENTRY

Preconditioned inverse iteration and shift-invert Arnoldi method

A HARMONIC RESTARTED ARNOLDI ALGORITHM FOR CALCULATING EIGENVALUES AND DETERMINING MULTIPLICITY

ITERATIVE REGULARIZATION WITH MINIMUM-RESIDUAL METHODS

Exponentials of Symmetric Matrices through Tridiagonal Reductions

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

Matrices, Moments and Quadrature, cont d

Matrix functions and their approximation. Krylov subspaces

Summary of Iterative Methods for Non-symmetric Linear Equations That Are Related to the Conjugate Gradient (CG) Method

On the Ritz values of normal matrices

DELFT UNIVERSITY OF TECHNOLOGY

DELFT UNIVERSITY OF TECHNOLOGY

Krylov Subspace Methods that Are Based on the Minimization of the Residual

Linear Algebra Massoud Malek

Charles University Faculty of Mathematics and Physics DOCTORAL THESIS. Krylov subspace approximations in linear algebraic problems

On prescribing Ritz values and GMRES residual norms generated by Arnoldi processes

Matrix Algorithms. Volume II: Eigensystems. G. W. Stewart H1HJ1L. University of Maryland College Park, Maryland

c 2005 Society for Industrial and Applied Mathematics

8 The SVD Applied to Signal and Image Deblurring

Alternative correction equations in the Jacobi-Davidson method

8 The SVD Applied to Signal and Image Deblurring

WHEN studying distributed simulations of power systems,

PROJECTED GMRES AND ITS VARIANTS

c 2005 Society for Industrial and Applied Mathematics

Arnoldi-Tikhonov regularization methods

On the loss of orthogonality in the Gram-Schmidt orthogonalization process

APPLIED NUMERICAL LINEAR ALGEBRA

Math 504 (Fall 2011) 1. (*) Consider the matrices

Key words. cyclic subspaces, Krylov subspaces, orthogonal bases, orthogonalization, short recurrences, normal matrices.

6 The SVD Applied to Signal and Image Deblurring

The Solvability Conditions for the Inverse Eigenvalue Problem of Hermitian and Generalized Skew-Hamiltonian Matrices and Its Approximation

Two Results About The Matrix Exponential

Eigenvalue and Eigenvector Problems

On the application of different numerical methods to obtain null-spaces of polynomial matrices. Part 1: block Toeplitz algorithms.

Course Notes: Week 1

ANY FINITE CONVERGENCE CURVE IS POSSIBLE IN THE INITIAL ITERATIONS OF RESTARTED FOM

ON THE GLOBAL KRYLOV SUBSPACE METHODS FOR SOLVING GENERAL COUPLED MATRIX EQUATIONS

AMS526: Numerical Analysis I (Numerical Linear Algebra)

6.4 Krylov Subspaces and Conjugate Gradients

Research Article Some Generalizations and Modifications of Iterative Methods for Solving Large Sparse Symmetric Indefinite Linear Systems

Notes on PCG for Sparse Linear Systems

FEM and sparse linear system solving

EIGENVALUE PROBLEMS. EIGENVALUE PROBLEMS p. 1/4

Computational Methods. Eigenvalues and Singular Values

Outline. Math Numerical Analysis. Errors. Lecture Notes Linear Algebra: Part B. Joseph M. Mahaffy,

A fast randomized algorithm for overdetermined linear least-squares regression

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Iterative Methods for Sparse Linear Systems

1 Extrapolation: A Hint of Things to Come

INVERSE SUBSPACE PROBLEMS WITH APPLICATIONS

Scientific Computing: An Introductory Survey

Solution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method

you expect to encounter difficulties when trying to solve A x = b? 4. A composite quadrature rule has error associated with it in the following form

The Nullspace free eigenvalue problem and the inexact Shift and invert Lanczos method. V. Simoncini. Dipartimento di Matematica, Università di Bologna

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Numerical Methods. Elena loli Piccolomini. Civil Engeneering. piccolom. Metodi Numerici M p. 1/??

Computational Linear Algebra

ON THE QR ITERATIONS OF REAL MATRICES

Krylov Subspaces. Lab 1. The Arnoldi Iteration

Key words. conjugate gradients, normwise backward error, incremental norm estimation.

PERTURBED ARNOLDI FOR COMPUTING MULTIPLE EIGENVALUES

Krylov Space Methods. Nonstationary sounds good. Radu Trîmbiţaş ( Babeş-Bolyai University) Krylov Space Methods 1 / 17

Solving large sparse eigenvalue problems

The Eigenvalue Problem: Perturbation Theory

Chasing the Bulge. Sebastian Gant 5/19/ The Reduction to Hessenberg Form 3

FGMRES for linear discrete ill-posed problems

Large scale continuation using a block eigensolver

OUTLINE 1. Introduction 1.1 Notation 1.2 Special matrices 2. Gaussian Elimination 2.1 Vector and matrix norms 2.2 Finite precision arithmetic 2.3 Fact

DELFT UNIVERSITY OF TECHNOLOGY

Computation of eigenvalues and singular values Recall that your solutions to these questions will not be collected or evaluated.

On the Modification of an Eigenvalue Problem that Preserves an Eigenspace

LARGE SPARSE EIGENVALUE PROBLEMS. General Tools for Solving Large Eigen-Problems

Model reduction via tangential interpolation

Introduction to Iterative Solvers of Linear Systems

Math 405: Numerical Methods for Differential Equations 2016 W1 Topics 10: Matrix Eigenvalues and the Symmetric QR Algorithm

Lecture 5 Basic Iterative Methods II

On solving linear systems arising from Shishkin mesh discretizations

Transcription:

NUMERICAL LINEAR ALGEBRA WITH APPLICATIONS Numer. Linear Algebra Appl. 200; 8: 6 [Version: 2000/03/22 v.0] On the solution of large Sylvester-observer equations D. Calvetti, B. Lewis 2, and L. Reichel 2 Department of Mathematics, Case Western Reserve University, Cleveland, OH 4406. 2 Department of Mathematics and Computer Science, Kent State University, Kent, OH 44242. SUMMARY The design of a Luenberger observer for large control systems is an important problem in Control Theory. Recently, several computational methods have been proposed by Datta and collaborators. The present paper discusses numerical aspects of one of these methods, described by Datta and Saad (99). Copyright c 200 John Wiley & Sons, Ltd. key words: Arnoldi process, Sylvester equation, Luenberger observer, partial fraction representation, control system. Consider the control system. INTRODUCTION ˆx(t) = M ˆx(t) + Bû(t), ˆx(0) = ˆx 0, t 0, ŷ(t) = C ˆx(t), () where M n n, B n k and C m n, and the functions ˆx(t) n and û(t) k are defined for t 0. Throughout this paper we will assume that n is large and m n. In many situations of practical interest neither the initial state ˆx 0 nor the states ˆx(t) for t > 0 are explicitly known. A popular approach to gain information about ˆx(t) is to design a control systems related to (), whose states x(t) approximate ˆx(t); see, e.g., Datta [9, Chapter 2] or Kailath [5, Chapter 4] for discussions. Luenberger [6] proposed the construction of an approximation x(t) of ˆx(t) as follows. Introduce the control system ẋ(t) = H T x(t) + G T ŷ(t) + Dû(t), x(0) = x 0, t 0, (2) email: D. Calvetti (dxc57@po.cwru.edu), B. Lewis (blewis@mcs.kent.edu), L. Reichel (reichel@mcs.kent.edu) Contract/grant sponsor: Research supported in part by NSF grant DMS-9806702. 2 Research supported in part by NSF grant DMS-980643. Received 2 January 2000 Copyright c 200 John Wiley & Sons, Ltd. Revised 5 February 200 Accepted 6 April 200

2 CALVETTI, ET AL. where H, G m m, D m k are matrices to be determined. The system (2) is commonly referred to as a Luenberger observer. Assume that the spectra λ(h) of H and λ(m) of M satisfy The property (3) secures that the Sylvester equation λ(h) λ(m) =. (3) X T M H T X T = G T C (4) has a unique solution X T m n ; see, e.g., [3, Chapter 4] for a discussion. Let For future reference, we express equation (4) in the form A = M T. (5) AX XH = C T G. (6) We will now show that the difference between x(t) and X T ˆx(t), where ˆx(t) solves (), converges to zero as t increases, provided that the matrices H and D in (2) are chosen in a suitable manner. Differentiating the difference and using () and (2) yields e(t) = x(t) X T ˆx(t) (7) ė(t) = H T x(t) + G T ŷ(t) + Dû(t) X T (M ˆx(t) + Bû(t)), = H T e(t) + (H T X T + G T C X T M)ˆx(t) + (D X T B)û(t). (8) Letting D = X T B and substituting the relation (4) into (8) shows that e(t) = exp(ht)(x 0 X T ˆx 0 ), t 0. (9) Assume that the matrix H is asymptotically stable, i.e., that every eigenvalue of H has negative real part. Then it follows from (9) that the difference (7) converges to zero as t increases. The above construction of an approximation x(t) of ˆx(t) showed necessary conditions on the matrix H. However, no requirements were imposed on the matrices X and G. In order to reduce the sensitivity of x(t) to perturbations, the following additional conditions are often imposed:. All eigenvalues of H have smaller real part than any eigenvalue of A, 2. X is well conditioned, and 3. G is such that the matrix pair {H, G} is controllable, i.e., rank([g, H zi m ]) = m, z, (0) where I m denotes the m m identity matrix. Moreover, it is desirable that the distance to uncontrollability d uc (H, G) = min σ m ([G, H zi m ]) z is not tiny, where σ m ([G, H zi m ]) denotes the mth singular value of the matrix [G, H zi m ] and the singular values are enumerated in decreasing order.

LARGE SYLVESTER-OBSERVER EQUATIONS 3 We focus on control systems () with large, possibly sparse, matrices A, and therefore ignore solution methods that require the transformation of the matrix A to a condensed form or demand knowledge of all eigenvalues of A. For references to methods that are well suited for small to medium sized systems (6); see Datta and Saad [] as well as the survey papers by Datta [7, 8]. One approach to solve (6) is to first choose the matrices G and H and then solve (6) for X by a solution method for Sylvester equations, such as the iterative methods discussed in [4, 4]. However, these methods do not guarantee that the solution matrix X is well conditioned. Datta and collaborators have developed several elegant methods for the solution of largescale Sylvester-observer equations (6); see [, 8, 0,, 2]. These methods are based on the Arnoldi process and use the matrix A, defined by (5), only to evaluate matrix-vector products. Sparsity or other structure of A, such as Toeplitz structure, may enable rapid evaluation of the matrix-vector products. Moreover, these methods yield well conditioned matrices X. An insightful and thorough discussion on Sylvester-observer equations is provided in the forthcoming book by Datta [9]. The present paper discusses numerical aspects of the Datta-Saad method []. This method allows an arbitrary choice of distinct eigenvalues of the matrix H. We investigate how the location of the eigenvalues of H affects the performance of the algorithm and propose a strategy for choosing these eigenvalues. In order to keep our presentation simple, we only discuss the method of Datta and Saad [], however, our analysis and strategy for choosing eigenvalues of H also applies, mutatis mutandis, to the related methods developed by Datta and collaborators. A preliminary discussion of the topic of the present paper can be found in [5]. 2. THE DATTA-SAAD METHOD The solution method for the Sylvester-observer equation (6) by Datta and Saad [] is based on the Arnoldi process. The matrix A is used only in matrix-vector product evaluations and this makes the method well-suited to the solution of large-scale Sylvester-observer equations. The method is designed for the special case when the right-hand side matrix C T G in (6) is of rank one. Modifications that allow a right-hand side matrix of higher rank are presented by Bischof et al. [] and Datta and Hetti [0]. Let the matrices A and C be given by (). Following Datta and Saad [], we choose G to be the identity matrix. Then, clearly, equation (0) is satisfied and d uc (H, G). We may assume that the rank-one matrix C T is of the form ce T m, where c n and e m denotes the mth axis vector. This particular form of C can be obtained by an initial orthogonal transformation. Thus, we may write equation (6) as AX XH = ce T m. () We will see below that H m m in () can be chosen to be an upper Hessenberg matrix, and X n m can be chosen to have orthogonal columns, all of the same Euclidean norm. With these choices of H and X, equation () is closely related to the Arnoldi decomposition of A obtained by application of m steps of the Arnoldi process. Specifically, m steps of the Arnoldi process applied to the matrix A with initial vector v of Euclidean norm one yields

4 CALVETTI, ET AL. the Arnoldi decomposition AV m V m H m = η m v m+ e T m, (2) where V m = [v, v 2,..., v m ] n m and Vm T V m = I m, H m m m is an unreduced upper Hessenberg matrix, and v m+ n satisfies vm+ T v m+ = and Vm T v m+ = 0. Another step of the Arnoldi process would give the matrix V m+, whose last column is v m+. We note for future reference that the vectors v, v 2,..., v m+ determined by (2) form an orthonormal basis of the Krylov subspace m+ (A, v ) = span{v, Av,..., A m v }. (3) For certain initial vectors v, the Arnoldi process may break down before the decomposition (2) has been determined. These (rare) cases allow certain simplifications and have to be treated separately. For notational simplicity, we assume throughout this paper that m is chosen small enough so that the Arnoldi decomposition (2) exists with η m 0. The similar form of the equations () and (2) suggests that the Arnoldi process may be applied to determine a solution of (). This observation is the basis of the solution method for () proposed by Datta and Saad []. The following results form the basis of their method. Throughout this paper denotes the Euclidean vector norm or the associated induced matrix norm. Theorem 2.. Let the matrix H m = [h i,j ] m i,j= m m be an unreduced upper Hessenberg matrix with spectrum λ(h m ), and let {µ j } m j= be a set of distinct complex numbers, such that {µ j } m j= λ(h m) =. Introduce the quantities s = m (H m µ j I)e, α = j= m j= h j+,j. (4) Then the upper Hessenberg matrix H m αse T m has the spectrum λ(h m αse T m) = {µ j } m j=. If the set {µ j } m j= is invariant under complex conjugation, then the matrix H m αse T m is real. Proof. Slight modifications of this theorem are formulated by Datta [6] and Datta and Saad []. The proof presented in [6] can be modified to show the theorem. Lemma 2.. Let the matrices A, V m and H m, vectors v and v m+ and scalar η m be those in the Arnoldi decomposition (2). Assume that c m+ (A, v ) and vm+ T c 0. Then there are a vector f m and scalar β m, such that AV m V m (H m fe T m) = β m ce T m. (5) Proof. Let β m = η m /vm+ T c and let the vector f m satisfy β m c = V m f + β m (v T m+ c)v m+ = V m f + η m v m+. (6) The Arnoldi decomposition (2) and formula (6) yield This shows (5). AV m V m (H m fe T m ) = η mv m+ e T m + V mfe T m = η m v m+ e T m + (β m c η m v m+ )e T m.

LARGE SYLVESTER-OBSERVER EQUATIONS 5 Note that the matrix H m fe T m in (5) is of upper Hessenberg form. Therefore Lemma 2. shows that if we determine the vector v so that c m+ (A, v ), then the equation (5) is of the form () up to the scaling factor β m. The following lemmata and theorem show how such a Krylov subspace can be determined and that λ(h m fe T m) = {µ j } m j=. Lemma 2.2. Let the matrices A, V m and H m and vector v be those of the Arnoldi decomposition (2). Let p be a polynomial of degree less than m. Then Proof. It is easy to show that and the lemma follows. Lemma 2.3. Let H m+ = [h i,j ] m+ monic polynomial of degree m. Then p(a)v = V m p(h m )e. (7) A j v = V m H j me, 0 j < m, i,j= (m+) (m+) be an upper Hessenberg matrix and p a e T m+p(h m+ )e = Proof. The result can be shown by induction. m h j+,j. (8) Theorem 2.2. Let A n n and c n be defined by (). Let {µ j } m j= be a set of m distinct complex numbers, such that {µ j } m j= λ(a) =. Introduce the monic polynomial m p m (t) = (t µ j ) (9) and let x be the unique solution of the linear system of equations j= j= p m (A)x = c. (20) Let V m, H m, v m+ and η m be determined by the Arnoldi decomposition (2) with initial vector v = x/ x. Define β m = η m /v T m+c and f = β m V T m c. Then and c m+ (A, v ) (2) λ(h m fe T m ) = {µ j} m j=. (22) Proof. Equation (2) follows from (9) and (20). In order to show (22), we note that f = β m V T m c = β mv T m p m(a)x = β m x V T m p m(a)v. (23) Let p m (t) = m j= (t µ j). Substituting p m (t) = (t µ m )p m (t) into the right-hand side of (23) and applying Lemma 2.2 yield β m x Vm T p m (A)v = β m x Vm T (A µ m I)p m (A)v = β m x Vm T (A µ m I)V m p m (H m )e = β m x (H m µ m I)p m (H m )e = β m x p m (H m )e = β m x s, (24)

6 CALVETTI, ET AL. where we have used that H m = Vm T AV m and s is defined by (4). It follows from equation (20) that β m x = η m vm+ T c x = η m vm+ T p m(a)x x = η m vm+ T p. (25) m(a)v Application of Lemmata 2.2 and 2.3 to the right-hand side of (25) yields η m v T m+ p m(a)v = = η m η m vm+ T V = m+p m (H m+ )e e T m+ p (26) m(h m+ )e η m m j= h j+,j = m j= h, j+,j where H m+ (m+) (m+) is the upper Hessenberg matrix determined by applying m + steps of the Arnoldi process with initial vector v to the matrix A. The last equality in (26) follows from the fact that h m+,m, the last subdiagonal entry of H m+, equals η m. Equations (25)-(26) show that α = β m x, where α is defined by (4). It follows from equations (23)-(24) that f = αs. Hence, (22) is a consequence of Theorem 2.. In the remainder of this section, we outline the computations required in order to determine the vectors x and f of Theorem 2.2. More details are provided in Sections 3-4. We first turn to the system of equations (20). Introduce the partial fraction decomposition m p m (t) = j= The solution of (20) can be expressed as where x j solves α j t µ j, α j = x = p m (µ j). (27) m x j, (28) j= (A µ j I)x j = α j c, j m. (29) The GMRES method is one of the most popular iterative methods for the solution of large linear systems of equations with a nonsymmetric matrix. Following Datta and Saad [], we will apply this method to solve the linear systems (29). A recent description of the GMRES method is provided by Saad [9, Chapter 6]. The standard implementation of the GMRES method for the solution of the jth equation of (29) uses an Arnoldi decomposition of the matrix A µ j I of the form where we note that (A µ j I)V l = V l H (µj) l + η (µj ) l v l+ e T l = V l+ H (µj) l+,l, V le = c/ c, (30) H (µj) l = H l + µ j I l (3)

LARGE SYLVESTER-OBSERVER EQUATIONS 7 and H l is the upper Hessenberg matrix in the Arnoldi decomposition AV l = V l H l + η l v l+ e T l, V l e = c/ c. (32) The matrix H (µj) l+,l (l+) l in (30) consists of the first l columns of the upper Hessenberg matrix H (µj) l+ (l+) (l+) determined by application of l + steps of the Arnoldi process to the matrix A µ j I. The matrix V l+ in (30) is independent of the value of µ j. This fact and the simple form (3) of the matrix H (µj) l made it possible for Datta and Saad [] to solve the m linear systems (29) by the GMRES method by only applying the Arnoldi process once to compute the Arnoldi decomposition (32) for some l sufficiently large, and then determining the decompositions (30) by modifying the matrix H l in (30) according to (3). Using the Arnoldi decomposition (30), the GMRES method determines an approximate solution x (l) j of the jth linear systems of equations (29), such that the associated residual vector satisfies r (l) j = α j c (A µ j )x (l) j (33) r (l) j = α j c (A µ j I)x (l) j = min α jc (A µ j I)x x l (A µ ji,c) = min y l α j c AV l y = min α j c V l+ H (µ) y l l+,l y = min α j c e H (µ) y l l+,ly. (34) Denote the solution of the minimization problem (34) by y (l) j. Having evaluated this solution, we compute x (l) j = V l y (l) j. The following theorem sheds light on the behavior of the linear system (29) as Re(µ j ) decreases. In particular, the theorem shows that the condition number κ(a µ j I) = A µ j I (A µ j I) is not large when Re(µ j ) is of large magnitude. Theorem 2.3. Let x (l) j be the approximate solution of the jth linear system (29) determined by the GMRES method by solving the minimization problem (34), and let r (l) j be the associated residual vector (33). Then r (l) j 0, κ(a µ j I) and x (l) j x j as µ j. (35) Proof. It follows from equation (34) that r (l) j r (l ) j r () j. It is straightforward to bound r () j. Equation (34) with l = yields that x () j = γ j c, where the scalar γ j solves Thus, γ j = r () j = min α j c (A µ j I)(γc). γ α j c T (A T µ j I)c c T (A T µ j I)(A µ j I)c α j µ j as µ j. (36)

8 CALVETTI, ET AL. Substituting x () j = γ j c into (33) yields, in view of (36), that r () j α j c (A µ j I)( α j µ j )c as µ j, (37) and the right-hand side expression in (37) converges to zero as µ j increases. To show that the condition number κ(a µ j I) approaches one as µ j, observe that Finally, A µ j I (A µ j I) = I µ j A (I µ j A) as µ j. x () j x j = (A µ j I) r () j 0 as µ j. In applications of the GMRES method, we may, for instance, choose l in the Arnoldi decomposition (32) as the smallest integer, such that the residual errors satisfy the bound r (l) j ɛ, j m, (38) for some specified ɛ > 0. The theorem suggests that for a fixed vector c and parameter l > 0, the norm of the residual error r (l) j and of the error x () j x j decrease as Re(µ j ) decreases. Indeed, we have observed this behavior in numerous numerical examples, already for fairly small values of Re(µ j ). This indicates that in order for the GMRES method to give rapid convergence, we should choose µ j to have a negative real part of large magnitude. On the other hand, choosing all the eigenvalues µ j of the matrix H in () with real parts much smaller than the real part of the leftmost eigenvalue of the matrix H m in the Arnoldi decomposition (2), makes it necessary to move the eigenvalues of H m a long distance to the µ j in the eigenvalue assignment problem for the matrix H discussed in Theorem 2.. The sensitivity of this eigenvalue assignment problem increases the further the eigenvalues have to be moved; see Mehrmann and Xu [7] for bounds for the condition number of the eigenvalue assignment problem. Related results can be also be found in [3, 8]. Let the Hessenberg matrix H = H m αse T m have spectral factorization with a diagonal matrix Λ H. Then H = W H Λ H W H, κ(w H ) = W H W H provides an estimate of the sensitivity of the eigenvalue assignment problem, see [3, 8], and is displayed in the numerical examples of Section 4. Thus, we would like to choose the real part of the µ j small enough to make the GMRES algorithm converge rapidly and large enough so that the eigenvalue assignment problem discussed in Theorem 2. is not too ill-conditioned. The next section considers a further aspect of the choice of the µ j ; there we discuss how the choice of the µ j affects the magnitude of the coefficients α j in the partial fraction representation (27).

LARGE SYLVESTER-OBSERVER EQUATIONS 9 3. PARTIAL FRACTION REPRESENTATION The eigenvalues µ j of the matrix H in () determine the partial fraction representation (27). The location of the eigenvalues determines how accurately the partial fraction can be evaluated in finite precision arithmetic. Close poles can greatly compromise the accuracy of the computed value of /p m (t). In addition coefficients α j of large magnitude in the representation (27) may make it necessary to compute an Arnoldi decomposition (32) with a large value of l in order to satisfy the residual error bound (38) since residual errors r (l) j determined by the GMRES method are proportional to α j. Example 3.. Let p 2 (t) = (t ɛ)(t + ɛ) with ɛ > 0 tiny. Consider the evaluation of the partial fraction representation p 2 (t) = 2ε t ε 2ε t + ε in finite precision arithmetic and assume that the computations are carried out with three significant digits. Let ɛ = /900. Then evaluation of the representation (39) yields the value 0.00. Evaluating the product form representation p 2 (t) = (t ε)(t + ε) with the same arithmetic and value of ε yields the value.00. In exact arithmetic we have p 2 () = ε 2 =.00000. This example illustrates that the partial fraction representation may yield significantly lower accuracy than the product form representation. The difficulty in the present example stems from loss of significant digits due to partial fraction coefficients of large magnitude and opposite sign. This in turn is caused by the poles being very close. Example 3.2. This example shows that the magnitude of the coefficients in the partial fraction representation depends not only on the distance between poles, but also on the distribution of the poles. Let p m (t) = m (t j m ). j= Then the partial fraction coefficients (27) are given by In particular, for m = 0, we obtain α l = m m (39). (40) m (l j) j = j l α 6 = α 5 3 0 5.

0 CALVETTI, ET AL. Thus, evaluation of the partial fraction representation (27) in finite precision arithmetic may cause severe cancellation of significant digits despite the fact that /p m (t) does not have close poles. Moreover, the large magnitude of the partial fraction coefficients may make the use of an Arnoldi decomposition (32) with a large value of l necessary. Note that some of the coefficients (40) are of much larger magnitude than others. This is measured by the quotient q(m) = max j m α j min j m α j (4) We have q(0) = 26. Note that q(m) is invariant under linear transformation of the zeros µ j of p m (t). In particular, the value q(0) is the same whenever the zeros µ j are equidistant on some interval in the complex plane. Example 3.3. Let p m (t) = cos(m arccos(t/2)), i.e., p m (t) is a Chebyshev polynomial of the first kind for the interval [ 2, 2]. Its zeros are Then yields µ j = 2 cos( 2j π). (42) 2m p m(µ j ) = ( )j m 2 sin( 2j 2m π) α j 2, j m. (43) m The distance between some adjacent zeros µ j is fairly small for large values of m. For instance, µ µ 2 = 2 cos( π 3π ) cos( 2m 2m ) = 4 sin π m sin π 2m < 2π2 m 2. Nevertheless, the bound (43) shows that max j m α j converges to zero as m increases. Moreover, all partial fraction coefficients are of roughly the same order of magnitude. For instance, we have q(0) = 6.4, where q is defined by (4). This value is much smaller than the value for equidistant zeros reported in Example 3.2. Numerical experiments indicate that the partial fraction representation (27) with the zeros (42) does not suffer from severe loss of significant digits when evaluated in finite precision arithmetic. Further examples of difficulties of partial fraction representations, as well as a discussion on how to ameliorate these difficulties, can be found in [2]. 4. CHOICE OF EIGENVALUES OF H The examples in Section 3 illustrate that the use of a partial fraction representation can, but must not, lead to severe loss of accuracy when evaluated in finite precision arithmetic. Partial fraction coefficients of large magnitude can give rise to severe loss of accuracy due to cancellation of significant digits during evaluation of the partial fraction representation. Conversely, numerical experiments display little loss of accuracy when evaluating a partial

LARGE SYLVESTER-OBSERVER EQUATIONS fraction representation whose coefficients all are of roughly same, not very large magnitude. Example 3.3 suggests that choosing the eigenvalues µ j of the matrix H in () as zeros of a Chebyshev polynomial gives coefficients α j in the partial fraction representation (27) of roughly the same magnitude, which typically is not very large. We propose the following selection of eigenvalues {µ j } m j= of the matrix H. In order to secure that the matrices A µ j I are nonsingular, we determine approximations of the eigenvalues of A with smallest real part. The eigenvalues of the upper Hessenberg matrix H l in the Arnoldi decomposition (32) used for the solution of the linear systems (29) by the GMRES method furnishes approximations of the desired eigenvalues of A. If these approximations are not sufficiently accurate then they can be improved, e.g., by the Implicitly Restarted Arnoldi method proposed by Sorensen [20]. We choose the µ j to be zeros of the Chebyshev polynomial of the first kind of degree m for the interval [τ +iρ, τ iρ], i =, where τ < min t λ(a) Re(t) and ρ = max t λ(a) Im(t). The accurate determination of max t λ(a) Im(t) is not crucial for the performance of the the algorithm below, which computes the solution {H, X} of the Sylvesterobserver equation AX XH = β m ce T m, (44) where β m is the same as in Lemma 2.. This equation is obtained by scaling the equation () by the factor β m. Therefore, if {H, X} solves (44), then {H, βm X} solves (). Algorithm Input: A, c, m, l. Output: H, X.. Apply l steps of the Arnoldi process with initial vector c to determine an Arnoldi decomposition of the form (32). Compute approximations of the eigenvalues of A with smallest and largest real parts. 2. If the eigenvalue approximations computed in Step are not sufficiently accurate, then compute improved approximations with the Implicitly Restarted Arnoldi method [20]. 3. Choose the set of eigenvalues {µ j } m j= of the matrix H in (). We propose that µ j be chosen as zeros of Chebyshev polynomials for the interval [τ + iρ, τ iρ] introduced above. This set of eigenvalues is invariant under complex conjugation. 4. Solve the systems of equations (29) by the GMRES method as described in Section 3, i.e., use the available Arnoldi decomposition (32) computed in Step to determine the decompositions (30) for j m. The latter Arnoldi decompositions are used to compute approximations x (l) j the minimization problems (34). If the approximations x (l) j evaluate the approximate solution of x j for j m by the GMRES method, i.e., by solving x (l) = m j= x (l) j are sufficiently accurate, then of the linear system of equations (20), cf. the formula (28), otherwise increase l in the Arnoldi decompositions (32) and (30). 5. Apply m steps of the Arnoldi process with initial vector v = x/ x to determine the Arnoldi decomposition (2). This gives the matrix H m. 6. Compute the vector f m, such that λ(h m fe T m) = {µ j } m j= by using the formulas of Theorem 2. with β m defined in Lemma 2.. Let X = V m /β m, H = H m fe T m.

2 CALVETTI, ET AL. We remark that the vector f used for the eigenvalue assignment in Step 6 of the algorithm can be computed either by the formulas of Theorem 2. or according to Lemma 2.. When the matrix A is large, we generally only solve the equation (20) approximately, and then these two approaches to compute the vector f are not equivalent. When the system (20) is not solved exactly and we compute f by the formulas of Lemma 2., the eigenvalue of the Hessenberg matrix H in general are further away from the µ j than when the formulas of Theorem 2. are applied. We therefore use the latter approach in Algorithm. 5. NUMERICAL EXAMPLES The computations reported in this section were carried out on an Intel Pentium workstation using Matlab 5.3 and floating point arithmetic with 6 significant digits. In all examples, we used the same matrix A 500 500, which we determined by generating its spectral decomposition. The eigenvalues λ j were distributed in the disk = {z : z + = } as follows. We let λ j = ρ j exp(iτ j ), i =, where the ρ j and τ j were uniformly distributed in the intervals [0, ] and [0, π], respectively. When Im(λ j ) > 0, λ j = ρ j exp( iτ j ) was also chosen as an eigenvalue. We determined the eigenvector matrix W with real or complex conjugate columns, the real and imaginary parts of the entries uniformly distributed in the interval [0, ]. This gave an eigenvector matrix with condition number κ(w ) = W W = 5. 0 4. The eigenvalues of A are marked by dots in the figures below. The computations were carried out as described by Algorithm and illustrate the numerical consequences of the choice of eigenvalues µ j of the Hessenberg matrix H in Step 3 of Algorithm. Example 5.. Steps -2 of Algorithm with l = 20 gave the approximations min Re(t) =.9, max t λ(a) Im(t) = 0.9. t λ(a) We let m = 8 and choose the µ j to be the zeros of a Chebyshev polynomial of the first kind of degree 8 for the interval [.9 0.90i,.9 + 0.90i], where i =. The eigenvalues of A and the µ j are displayed by Figure. The solution of the equation (20) is determined by solving the m systems of equations (29) using the Arnoldi decomposition (32) with l = 20. Numerical results are shown in Table I under the heading Nearby µ j. For comparison, we repeat the computations of Step 3-6 of Algorithm with the µ j allocated further to the left in the complex plane. Specifically, we let the µ j be the zeros of the Chebyshev polynomial of the first kind of degree 8 for the interval [ 5.7 0.90i, 5.7 + 0.90i]. The µ j are displayed in Figure 2 and the performance of Algorithm is shown in Table I under the heading Distant µ j. Table I. Effect of the choice of zeros on the solution of the Sylvester-observer equation () Quantity Nearby µ j Distant µ j AX XH β mc 8.4 0 2 3.6 0 3 κ(w H).3 0 3 5. 0

LARGE SYLVESTER-OBSERVER EQUATIONS 3 0.8 0.6 0.4 0.2 0 0.2 0.4 0.6 0.8 2.5 0.5 0 Figure. : Eigenvalues of A, : Eigenvalues of H m, : µ j, + : Eigenvalues of H = H m fe T m We see that the set of µ j further away from the imaginary axis produce substantially lower residual error. However, the condition number of the matrix H is larger for these µ j and this makes the pole placement problem more sensitive to perturbations. Example 5.2. This example illustrates the effect of the distribution of the µ j on the solution of the Sylvester equation. We let m = 4 and solve the equation (20) using the Arnoldi decomposition (32) with l = 50. The µ j are chosen to be zeros of the Chebyshev polynomial of the first kind of degree 4 for the interval [ 2.7 0.96i, 2.7 + 0.96i]. Figure 3 shows the spectra of A, H, and H m, as well as the µ j. Table II displays the performance of Algorithm. For comparison, the computations were also carried out with the µ j allocated equidistantly Table II. Effect of the choice of the µ j on the solution of the Sylvester-observer equation () Quantity µ j equidistant µ j zeros of Chebyshev polynomials AX XH β mc.8 0 0 5.3 0 2 max i m α i.6 0 4 8.9 0 2

4 CALVETTI, ET AL. 0.8 0.6 0.4 0.2 0 0.2 0.4 0.6 0.8 6 5 4 3 2 0 Figure 2. : Eigenvalues of A, : Eigenvalues of H m, : µ j, + : Eigenvalues of H = H m fe T m in the interval [ 2.7 0.96i, 2.7 + 0.96i]; see Figure 4. Numerical results are presented in Table II. Clearly, Algorithm performs better when the µ j are zeros of Chebyshev polynomials. 6. CONCLUSION The performance of the method by Datta and Saad [] for the solution of the Sylvesterobserver equations () depends on the choice of the set of eigenvalues {µ j } m j= of the matrix H. Let τ 0 = min t λ(a) Re(t) and ρ = max t λ(a) Im(t), and assume that τ 0 0 and ρ > 0. We propose to choose the µ j to be zeros of the Chebyshev polynomial of the first kind of degree m for the interval [τ + iρ, τ iρ] for some τ < τ 0. The smaller value of τ, the more accurately we can solve the linear system (20) for a fixed value of l in the Arnoldi decomposition (32), but the more ill-conditioned the eigenvalue assignment problem of Theorem 2.. Typically, we seek to choose τ small, but large enough to be able to solve the eigenvalue assignment problem to desired accuracy. ACKNOWLEDGEMENT

LARGE SYLVESTER-OBSERVER EQUATIONS 5 0.8 0.6 0.4 0.2 0 0.2 0.4 0.6 0.8 3 2.5 2.5 0.5 0 Figure 3. : Eigenvalues of A, : Eigenvalues of H m, : µ j, + : Eigenvalues of H = H m fe T m We would like to thank the referee for comments. REFERENCES. C. Bischof, B. N. Datta and A. Purkyastha, A parallel algorithm for the Sylvester-observer matrix equation, SIAM Journal on Scientific Computing 996; 7:686 698. 2. D. Calvetti, E. Gallopoulos and L. Reichel, Incomplete partial fractions for parallel evaluation of rational matrix functions, Journal of Computational and Applied Mathematics 995; 59:349 380. 3. D. Calvetti, B. Lewis and L. Reichel, On the selection of poles in the single input pole placement problem, Linear Algebra and Its Applications 999; 302-303:33 345. 4. D. Calvetti and L. Reichel, Application of ADI iterative methods to the restoration of noisy images, SIAM Journal on Matrix Analysis and Applications 996; 7:65 86. 5. D. Calvetti and L. Reichel, Numerical aspects of some solution methods for large Sylvester-observer equations, in Proceedings of the 36th IEEE Conference on Decision and Control, IEEE, Piscataway, 997:4389 4393 6. B. N. Datta, An algorithm to assign eigenvalues in a Hessenberg matrix, IEEE Transactions on Automatic Control 987; AC-32:44 47. 7. B. N. Datta, Linear and Numerical Linear Algebra in Control Theory: Some Research Problems, Linear Algebra and Its Applications 994; 98:755 790. 8. B. N. Datta, Krylov subspace methods in control: an overview, in Proceedings of the 36th IEEE Conference on Decision and Control, IEEE, Piscataway, 997:3844 3848.

6 CALVETTI, ET AL. 0.8 0.6 0.4 0.2 0 0.2 0.4 0.6 0.8 3 2.5 2.5 0.5 0 Figure 4. : Eigenvalues of A, : Eigenvalues of H m, : µ j, + : Eigenvalues of H = H m fe T m 9. B. N. Datta, Numerical Methods for Linear Control Systems Design and Analysis, Academic Press, to appear. 0. B. N. Datta and C. Hetti, Generalized Arnoldi methods for the Sylvester-observer equation and the multiinput pole placement problem, in Proceedings of the 36th IEEE Conference on Decision and Control, IEEE, Piscataway, 997:4379 4383.. B. N. Datta and Y. Saad, Arnoldi methods for large Sylvester-like observer matrix equations, and an associated algorithm for partial spectrum assignment, Linear Algebra and Its Applications 99; 54 56:225 244. 2. C. Hetti, On numerical solutions of the Sylvester-observer equation, and the multi-input eigenvalue assignment problem, Ph.D. thesis, Department of Mathematical Sciences, Northern Illinois University, DeKalb, IL, 996. 3. R. A. Horn and C. R. Johnson, Topics in Matrix Analysis, Cambridge Univ. Press, Cambridge, 99. 4. D. Y. Hu and L. Reichel, Krylov subspace methods for the Sylvester equation, Linear Algebra and Its Applications 992; 72:283 33. 5. T. Kailath, Linear Systems, Prentice-Hall, Englewood Cliffs, 980. 6. D. G. Luenberger, Observers for multivariable systems, IEEE Transactions on Automatic Control 966; AC-:90 97. 7. V. Mehrmann and H. Xu, An analysis of the pole placement problem I. The single-input case, Electronic Transactions on Numerical Analysis 996; 4:89 05. 8. V. Mehrmann and H. Xu, Choosing the poles so that the single-input pole placement problem is well conditioned, SIAM Journal on Matrix Analysis and Applications 998; 9:664 68. 9. Y. Saad, Iterative Methods for Sparse Linear Systems, PWS, Boston, 996. 20. D. C. Sorensen, Implicit application of polynomial filters in a k-step Arnoldi method, SIAM Journal on Matrix Analysis and Applications, 992; 3:357 385.