Diagonal matrix solutions of a discrete-time Lyapunov inequality

Similar documents
Block companion matrices, discrete-time block diagonal stability and polynomial matrices

Spectral inequalities and equalities involving products of matrices

Some New Results on Lyapunov-Type Diagonal Stability

POLYNOMIALS WITH A SHARP CAUCHY BOUND AND THEIR ZEROS OF MAXIMAL MODULUS

Z-Pencils. November 20, Abstract

An estimation of the spectral radius of a product of block matrices

Spectral Properties of Matrix Polynomials in the Max Algebra

ISOLATED SEMIDEFINITE SOLUTIONS OF THE CONTINUOUS-TIME ALGEBRAIC RICCATI EQUATION

On the simultaneous diagonal stability of a pair of positive linear systems

Positive entries of stable matrices

THE PERTURBATION BOUND FOR THE SPECTRAL RADIUS OF A NON-NEGATIVE TENSOR

An alternative proof of the Barker, Berman, Plemmons (BBP) result on diagonal stability and extensions - Corrected Version

Some bounds for the spectral radius of the Hadamard product of matrices

Interval solutions for interval algebraic equations

AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES

On Linear Copositive Lyapunov Functions and the Stability of Switched Positive Linear Systems

1 Last time: least-squares problems

Section 3.9. Matrix Norm

Recall the convention that, for us, all vectors are column vectors.

A Characterization of the Hurwitz Stability of Metzler Matrices

DEN: Linear algebra numerical view (GEM: Gauss elimination method for reducing a full rank matrix to upper-triangular

Fast Angular Synchronization for Phase Retrieval via Incomplete Information

Invertibility and stability. Irreducibly diagonally dominant. Invertibility and stability, stronger result. Reducible matrices

Jim Lambers MAT 610 Summer Session Lecture 2 Notes

Topics in Applied Linear Algebra - Part II

Notes on Linear Algebra and Matrix Theory

SIGN PATTERNS THAT REQUIRE OR ALLOW POWER-POSITIVITY. February 16, 2010

Math Introduction to Numerical Analysis - Class Notes. Fernando Guevara Vasquez. Version Date: January 17, 2012.

Maximizing the numerical radii of matrices by permuting their entries

Ma/CS 6b Class 20: Spectral Graph Theory

Spectral radius, symmetric and positive matrices

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88

OPTIMAL SCALING FOR P -NORMS AND COMPONENTWISE DISTANCE TO SINGULARITY

Review of Linear Algebra

Some bounds for the spectral radius of the Hadamard product of matrices

2 Two-Point Boundary Value Problems

Applied Mathematics Letters. Comparison theorems for a subclass of proper splittings of matrices

Ma/CS 6b Class 20: Spectral Graph Theory

Review of Some Concepts from Linear Algebra: Part 2

Contour integral solutions of Sylvester-type matrix equations

Types of convergence of matrices. Olga Pryporova. A dissertation submitted to the graduate faculty

FRIEDRICH-ALEXANDER-UNIVERSITÄT ERLANGEN-NÜRNBERG. Lehrstuhl für Informatik 10 (Systemsimulation)

Stability of interval positive continuous-time linear systems

On the spectra of striped sign patterns

j=1 u 1jv 1j. 1/ 2 Lemma 1. An orthogonal set of vectors must be linearly independent.

A priori bounds on the condition numbers in interior-point methods

The spectral decomposition of near-toeplitz tridiagonal matrices

Quadratic and Copositive Lyapunov Functions and the Stability of Positive Switched Linear Systems

Chapter 7. Canonical Forms. 7.1 Eigenvalues and Eigenvectors

Nonnegative and spectral matrix theory Lecture notes

On the stability of switched positive linear systems

Spectral Graph Theory Lecture 3. Fundamental Graphs. Daniel A. Spielman September 5, 2018

A classification of sharp tridiagonal pairs. Tatsuro Ito, Kazumasa Nomura, Paul Terwilliger

LECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel

Quantum Computing Lecture 2. Review of Linear Algebra

Nonlinear Discrete-Time Observer Design with Linearizable Error Dynamics

First, we review some important facts on the location of eigenvalues of matrices.

CLASSIFICATION OF TREES EACH OF WHOSE ASSOCIATED ACYCLIC MATRICES WITH DISTINCT DIAGONAL ENTRIES HAS DISTINCT EIGENVALUES

Iterative Methods for Solving A x = b

On marker set distance Laplacian eigenvalues in graphs.

U.C. Berkeley CS294: Beyond Worst-Case Analysis Handout 12 Luca Trevisan October 3, 2017

Linear Algebra Massoud Malek

On the Hermitian solutions of the

Fiedler s Theorems on Nodal Domains

An Even Order Symmetric B Tensor is Positive Definite

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

(Inv) Computing Invariant Factors Math 683L (Summer 2003)

Unit 3 Vocabulary. An algebraic expression that can contains. variables, numbers and operators (like +, An equation is a math sentence stating

Math 775 Homework 1. Austin Mohr. February 9, 2011

Fast Linear Iterations for Distributed Averaging 1

Stability Theory for Nonnegative and Compartmental Dynamical Systems with Time Delay

A Cross-Associative Neural Network for SVD of Nonsquared Data Matrix in Signal Processing

Math/Phys/Engr 428, Math 529/Phys 528 Numerical Methods - Summer Homework 3 Due: Tuesday, July 3, 2018

Basic Concepts in Linear Algebra

An Observation on the Positive Real Lemma

The following two problems were posed by de Caen [4] (see also [6]):

Positive systems in the behavioral approach: main issues and recent results

Linear Operators Preserving the Numerical Range (Radius) on Triangular Matrices

A LINEAR PROGRAMMING BASED ANALYSIS OF THE CP-RANK OF COMPLETELY POSITIVE MATRICES

On the adjacency matrix of a block graph

16.31 Fall 2005 Lecture Presentation Mon 31-Oct-05 ver 1.1

Review of Basic Concepts in Linear Algebra

Fiedler s Theorems on Nodal Domains

KOREKTURY wim11203.tex

Mathematical Optimisation, Chpt 2: Linear Equations and inequalities

Generalized Numerical Radius Inequalities for Operator Matrices

MAC Module 2 Systems of Linear Equations and Matrices II. Learning Objectives. Upon completing this module, you should be able to :

Control for stability and Positivity of 2-D linear discrete-time systems

Lectures on Linear Algebra for IT

Abed Elhashash and Daniel B. Szyld. Report August 2007

A TOUR OF LINEAR ALGEBRA FOR JDEP 384H

A NEW EFFECTIVE PRECONDITIONED METHOD FOR L-MATRICES

Chapter III. Stability of Linear Systems

Geometric Mapping Properties of Semipositive Matrices

Clarkson Inequalities With Several Operators

Completely positive matrices of order 5 with ĈP -graph

Inverse Eigenvalue Problems

A Tropical Extremal Problem with Nonlinear Objective Function and Linear Inequality Constraints

Key words. Strongly eventually nonnegative matrix, eventually nonnegative matrix, eventually r-cyclic matrix, Perron-Frobenius.

EE363 homework 7 solutions

Transcription:

Diagonal matrix solutions of a discrete-time Lyapunov inequality Harald K. Wimmer Mathematisches Institut Universität Würzburg D-97074 Würzburg, Germany February 3, 1997

Abstract Diagonal solutions of a Lyapunov inequality for companion matrices are studied. Such solutions are required if states of a discrete-time linear system are computed with a finite precision arithmetic.

1 Introduction Let x(i + 1) = Ax(i), x(0) = x 0, (1.1) ( T be a discrete-time linear system with x(i) = x 1 (i),..., x n (i)) C n. It is well known that the system (1.1) is asymptotically stable if only if there exists a matrix P > 0 (positive-definite) such that A P A P = Q Q (1.2) (A, Q) is observable. (1.3) According to [1] diagonal solutions P of (1.2) are required if a finite precision arithmetic is used to calculate the states x(i) of (1.1). Let g( ) be a scalar function which satisfies g(y) y for all y C, (1.4) ( T let g[x(i)] = g[x 1 (i)],..., g[x n (i)]) be the vector obtained from x(i) using g( ) component-wise as a quantizer operator. It was shown in [1] that the quantized system x(i + 1) = A x(i), x(i + 1) = g[x(i + 1)], (1.5) is asymptotically stable if there exists a diagonal matrix P such that (1.2) (1.3) hold. In the case where A is a companion matrix diagonal solutions of the Lyapunov equation (1.2) were studied in [2]. In this note we clarify some issues of [2] prove the following result. Theorem 1 Let A = be a complex companion matrix. Put a 1... a n 1 a n 1... 1 0 s = n a ν. (i) There exists a diagonal matrix P = diag(p 1,..., p n ) with properties if only if s 1. ν=1 (1.6) P 0 (positive-semidefinite), P 0, (1.7) L(P ) = P A P A 0 (1.8) 1

(ii) There exists a diagonal matrix P > 0 (positive-definite) satisfying (1.8) if only if either s < 1 or both s = 1 a n 0 (1.9) hold. In [2] it was shown that s 1 is necessary for the existence of a diagonal positivedefinite solution P of (1.8). Discrete-time Lyapunov equations (1.2) with a companion matrix A have been investigated by several authors, we refer to [3], [4], [5]. Theorem 1 will be proved in Section 2. A matrix of the form (1.6) which is important for the critical exponent of the row-sum norm [6] will be discussed in Section 3. A counterpart of Theorem 1 for the continuous-time Lyapunov inequality A P + P A 0 will be derived in Section 4. 2 Explicit solutions of L(P ) 0 In the following Lemma the matrix inequality (1.8) will be related to a scalar inequality. It will be convenient to allow denominators to be zero. For α R π = 0 we set α 2 π = 0 α = 0 if α 0. (2.1) Lemma 2 Let a 1,..., a n C p 1,..., p n R be given. The matrix P = diag (p 1,..., p n ) satisfies P 0, P 0 (2.2) if only if the conditions L(P ) = P A P A 0 (2.3) p 1 > 0, p 1 p 2 p n 0 (2.4) hold. Proof. Put 1 p 1 a 1 2 p 1 p 2 + + a n 1 2 p n 1 p n + a n 2 p n (2.5) Π = diag (p 1 p 2,..., p n 1 p n, p n p n+1 ), p n+1 = 0. 2

As in [2] we note that that (2.3) is equivalent to L(P ) = Π p 1 (a 1,..., a n ) (a 1,..., a n ) Π p 1 (a 1,..., a n ) (a 1,..., a n ). (2.6) If a j = a j e iϕ j, j = 1,..., n, set D = diag (e iϕ 1,..., e iϕn ) a = ( a 1,..., a n ). Then D ΠD = Π. Hence (2.6) is equivalent to Π p 1 aa. (2.7) Assume that (2.4) (2.5) hold. Then (2.1) implies a j = 0 if p j = p j+1. Hence in order to prove (2.7) we can discard the indices j with p j p j+1 = 0 because of (2.4) assume Π > 0. Then (2.5) is equivalent to 1 p 1 a Π 1 a = p 1 ( Π 1/2 a ) ( Π 1/2 a ). (2.8) Note that for a vector b C n the eigenvalues of the dyadic product bb are b b, 0,..., 0. Hence (2.8) is equivalent to I p 1 Π 1/2 aa Π 1/2, i.e. to (2.7). Starting from (2.7) the converse part of the lemma can be proved along similar lines. We focus on the inequalities (2.4) (2.5) assuming s 0. Set ˆp n+1 = 0 Then ˆp 1 = 1 ˆp ν = 1 s ( a ν + + a n ), ν = 1,..., n. (2.9) n ν=1 a ν 2 ˆp ν ˆp ν+1 = n a ν s = s 2. Hence if 0 < s 1 then (2.4) (2.5) are satisfied by p ν = ˆp ν, ν = 1,..., n. Now consider the case a n = 0 s < 1. Let k be such that ν=1 a n = = a k+1 = 0, a k 0. (2.10) Then ˆp k+1 = = ˆp n = 0, ˆp k = 1 s a k > 0. For δ R, δ ˆp k, define f(δ) = a 1 2 ˆp 1 ˆp 2 + + a k 1 2 ˆp k 1 ˆp k + a k 2 ˆp k δ. 3

Then f(0) = s 2. By continuity of f(δ) there exists an R, R > 0, such that It can be shown that (2.11) holds for f(δ) 1 if 0 δ R. (2.11) R = a k (1 s 2 ) s s 2 (s a k ) (2.12) that (2.12) is the best possible bound for (2.11). Hence any δ [0, R] yields a solution of (2.4) (2.5). (p 1,..., p n ) = (ˆp 1,..., ˆp k, δ,..., δ) (2.13) Proof of Theorem 1: (i) Assume s 1. If s = 0 then A A = diag(1,..., 1, 0). Hence P = I is a solution of (1.8). If s 0 we use the numbers ˆp ν given by (2.9) set... ˆP = diag(ˆp 1,..., ˆp n ). (2.14) Then Lemma 2 implies that P = ˆP has the properties (1.7) (1.8). Note that ˆP > 0 if a n 0. To show that s 1 is necessary for (1.7) (1.8) consider (1.8) in the equivalent form (2.7). Put e = (1,..., 1) T then (2.7) yields p 1 = e Πe p 1 (e a) 2 = p 1 s 2. Because of (2.4) we have p 1 > 0 therefore 1 s 2. (ii) To show that (1.8) has a positive-definite solution if (1.9) holds consider the three cases s = 0 0 < s 1 with a n 0, 0 < s < 1, a n = 0. (2.15) In the first two cases we know that a solution P > 0 of (1.8) is given by P = I P = ˆP respectively. In the third case (2.15) we assume (2.10) choose 0 < δ R with R as in (2.12). Using the n-tuple (2.13) we obtain the solution P = diag(ˆp 1,..., ˆp n, δ,..., δ) > 0. (2.16) To complete the proof we now assume that (1.8) holds for some P > 0. Then s 1 we have to exclude the case s = 1, a n = 0. (2.17) As before the assumption a ν = a ν, ν = 1,..., n, is no loss of generality. Put g = (1,..., 1, 0) T. Then (2.17) implies Ag = e g T L(P )g = p n. From L(P ) 0 follows 4

p n 0 which is incompatible with P > 0. Let ρ(a) denote the spectral radius of A. Clearly P > 0 L(P ) 0 imply ρ(a) 1, as it was mentioned in Section 1 an observability condition (1.3) ensures ρ(a) < 1. In the case of a companion matrix A a diagonal P we note the following result. Corollary 3 There exists a diagonal matrix P = diag(p 1,..., p n ) such that the pair if only if hold. From (2.20) follows ρ(a) < 1. P > 0 L(P ) = P A P A 0 (2.18) (A, L(P )) is observable (2.19) either s < 1 or both s = 1 0 < a n < 1 (2.20) Proof. If A is of the form (1.6) H = (h (1),..., h (n) ) C m n, then the pair (A, H) is observable if only if h (n) 0. Recall L(P ) = diag(p 1 p 2,..., p n 1 p n, p n ) p 1 (a 1,..., a n ) (a 1,..., a n ). If L(P ) 0 then the last column of L(P ) is nonzero if only if p n p 1 a n 2 > 0. (2.21) Hence (2.18) (2.19) implies (1.9) (2.21). From (2.4) (2.21) we deduce a n < 1 which leads to (2.20). Now assume (2.20) to establish the existence of a diagonal P satisfying (2.18) (2.19) three cases will be considered, namely s = 0, 0 < s < 1, a n = 0, (2.22) 0 < s 1, 0 < a n < 1. (2.23) In the case s = 0 again take P = I. Then L(P ) = diag(0,..., 0, 1), (2.18) (2.19) are satisfied. In the case (2.22) take P as in (2.16). Then p n = δ > p 1 a n 2 = 0 (2.21) equivalently (2.19) hold. In the case (2.23) take the matrix P = ˆP of (2.14). Then p n = 1 s a n, p 1 = 1, p n p 1 a n 2 = 1 s a n (1 s a n ) > 0. 5

We call a matrix A C n n diagonally d-stable if there exists a diagonal matrix P such that P > 0 L(P ) = P A P A > 0. For diagonal stability of discrete- of continuous-time Lyapunov equations we refer to [7]. A slight modification of the proof of Theorem 1 yields the following result. Corollary 4 The companion matrix A in (1.6) is diagonally d-stable if only if s < 1. 3 The critical exponent of the row-sum norm For a matrix B = (b ij ) C n n put B = max i { n } b ij. Mařík Pták [6] define the critical exponent of the row-sum norm on C n n as the positive integer κ which has the property that for which there exists a matrix C with j=1 ρ(b) = 1 if B = B 2 = B κ = 1, (3.1) ρ(c) < 1 C = = C κ 1 = 1; (3.2) i.e. κ is the smallest number with property (3.1). To prove that κ = n 2 n + 1 Mařík Pták use the following matrix C which (up to a permutation of rows columns) is of the form (1.6) satisfies (2.20). Let τ be real, 0 < τ < 1, put C = 0 1............... 1 τ τ 1 0 0 Then Corollary 3 with a n = τ a n 1 = τ 1 implies ρ(c) < 1. Note that C n = [(1 τ)a + τi] n n C n2 n = ( 1) n 1 [(1 τ)a + τi] n 1. (3.3). 6

Again set e = (1,..., 1) T. Then Ce = (1,..., 1, 1),..., C n 1 e = (1, 1,..., 1), (3.3) implies ( n 1 ( ) ) n 1 C n2 n e = ( 1) n 1 τ n 1 j (1 τ) j,... = ( 1) n 1 (1,...). j j=0 Hence the first row of C n2 n has norm 1 from 1 = C C 2... follows C = = C n2 n = 1, (3.4) which shows that κ n 2 n + 1. A different proof of (4.4) using a graph theoretical approach is contained in [8]. 4 The continuous-time Lyapunov inequality For a continuous-time n-dimensional linear system ẋ(t) = Ax(t) the Lyapunov inequality corresponding to (1.8) is S(P ) = A P + P A 0. (4.1) In contrast to the discrete-time inequality (1.8) a matrix A in companion form is hardly an advantage in (4.1) if diagonal solutions P are required. Theorem 5 Let A C n n, n 2, be a companion matrix of the form (1.6). (i) Then there exists a diagonal matrix P such that P 0, P 0 S(P ) 0 hold if only if det(λi λ) = λ n 2 (λ 2 a 1 λ a 2 ) (4.2) ā 1 + a 1 0 a 2 R, a 2 0. (4.3) (ii) If (4.2) (4.3) hold then a diagonal matrix P satisfies P 0 S(P ) 0 if only if P = diag(p 1, a 2 p 1, 0,..., 0), p 1 0. Proof. (i) If P = diag(p 1,..., p n ) has real entries then (ā 1 + a 1 )p 1 a 2 p 1 + p 2 a 3 p... a n p 1 ā 2 p 1 + p 2 0 p 3... 0. S(P ) = ā 3 p 1 p 3............. pn ā n p 1 0 p n 0 7 (4.4)

Assume S(P ) 0 P 0. Then p 3 = = p n = 0 p 1 0, we obtain a 3 = = a n = 0 which is (4.2). Clearly ( ) (ā1 + a 1 )p 1 a 2 p 1 + p 2 0, p ā 2 p 1 + p 2 0 1 > 0, p 2 0 implies (4.3). The converse part of (i) also of (ii) is obvious. In accordance with the discrete time concept we call a matrix A C n n diagonally c-stable if S(P ) = A P + P A < 0 for some positive-definite diagonal matrix P. In mathematical biology [9, p. 199] this property is known as Volterra-Lotka stability. From (4.4) it is clear that a companion matrix can never be diagonally c-stable if n 2. References [1] W. L. Mills, C. T. Mullis, R. A. Roberts, Digital filter realizations without overflow oscillations, IEEE Trans. Acoust., Speech, Signal Proc., vol. AC-26, pp. 334 338, 1978. [2] P. A. Regalia, On finite precision Lyapunov functions for companion matrices, IEEE Trans. Autom. Cont., vol. AC-37, pp. 1640 1644, 1992. [3] R. R. Bitmead, Explicit solutions of discrete-time Lyapunov matrix equations Kalman-Yakubovich equations, IEEE Trans. Autom. Cont., vol. AC-26, pp. 1291 1294, 1981. [4] A. Betser, N. Cohen, E. Zeheb, On solving the Lyapunov Stein equations for a companion matrix, Systems Control Lett., vol. 25, pp. 211 218, 1995. [5] H. K. Wimmer, Linear matrix equations: the module theoretic approach, Linear Algebra Appl., vol. 120, pp. 149 164. 1989. [6] J. Marik V. Pták, Norms, spectral combinatorial properties of matrices, Czechoslovak. Math. J., vol. 10, pp. 181 196, 1960. [7] G. P. Barker, A. Berman, R. J. Plemmons, Positive diagonal solutions to the Lyapunov equations, Linear Multilin. Algebra. vol. 5, pp. 249 256, 1978. [8] G. R. Belitskii Yu. I. Lyubich, Matrix Norms their Applications, Operator Theory: Advances Applications, vol. 36, Basel: Birkhäuser, 1988. [9] J. Hofbauer K. Sigmund, The Theory of Evolution Dynamical Systems, Mathematical Aspects of Selection, Cambridge: Cambridge University Press, 1988. 8