Reduction of Singular Pencils of Matrices 1

Similar documents
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 1 x 2. x n 8 (4) 3 4 2

Linear Algebra. The analysis of many models in the social sciences reduces to the study of systems of equations.

ALTERNATIVE DERIVATION OF SOME REGULAR CONTINUED FRACTIONS

Chapter 7. Linear Algebra: Matrices, Vectors,

Introduction to Quantitative Techniques for MSc Programmes SCHOOL OF ECONOMICS, MATHEMATICS AND STATISTICS MALET STREET LONDON WC1E 7HX

Linear Algebra 1 Exam 1 Solutions 6/12/3

MATRICES. knowledge on matrices Knowledge on matrix operations. Matrix as a tool of solving linear equations with two or three unknowns.

SAMPLE OF THE STUDY MATERIAL PART OF CHAPTER 1 Introduction to Linear Algebra

Introduction to Matrices

Introduction Non-uniqueness of factorization in A[x]... 66

SPACES OF MATRICES WITH SEVERAL ZERO EIGENVALUES

2 Eigenvectors and Eigenvalues in abstract spaces.

SAMPLE OF THE STUDY MATERIAL PART OF CHAPTER 1 Introduction to Linear Algebra

Moore Penrose inverses and commuting elements of C -algebras

A matrix over a field F is a rectangular array of elements from F. The symbol

Course 311: Michaelmas Term 2005 Part III: Topics in Commutative Algebra

Doubly Indexed Infinite Series

ACI-matrices all of whose completions have the same rank

ON DIVISION ALGEBRAS*

Foundations of Matrix Analysis

MATRICES AND ITS APPLICATIONS

ELEMENTARY LINEAR ALGEBRA

4. Matrix inverses. left and right inverse. linear independence. nonsingular matrices. matrices with linearly independent columns

Letting be a field, e.g., of the real numbers, the complex numbers, the rational numbers, the rational functions W(s) of a complex variable s, etc.

Factorization of singular integer matrices

MATH 2331 Linear Algebra. Section 2.1 Matrix Operations. Definition: A : m n, B : n p. Example: Compute AB, if possible.

CHAPTER 10 Shape Preserving Properties of B-splines

3 (Maths) Linear Algebra

Chapter 1. Matrix Algebra

Markov Chains and Stochastic Sampling

Instructions Please answer the five problems on your own paper. These are essay questions: you should write in complete sentences.

S. Lie has thrown much new light on this operation. The assumption

An integer arithmetic method to compute generalized matrix inverse and solve linear equations exactly

Arithmetic Progressions with Constant Weight

Chapter 5. Linear Algebra. A linear (algebraic) equation in. unknowns, x 1, x 2,..., x n, is. an equation of the form

Bare-bones outline of eigenvalue theory and the Jordan canonical form

Math 1060 Linear Algebra Homework Exercises 1 1. Find the complete solutions (if any!) to each of the following systems of simultaneous equations:

ARITHMETIC PROGRESSIONS IN CYCLES OF QUADRATIC POLYNOMIALS

Chapter 5. Linear Algebra. Sections A linear (algebraic) equation in. unknowns, x 1, x 2,..., x n, is. an equation of the form

Chapter 4 - MATRIX ALGEBRA. ... a 2j... a 2n. a i1 a i2... a ij... a in

8. Prime Factorization and Primary Decompositions

x y B =. v u Note that the determinant of B is xu + yv = 1. Thus B is invertible, with inverse u y v x On the other hand, d BA = va + ub 2

Linear Algebra II. 2 Matrices. Notes 2 21st October Matrix algebra

Matrix Representation

A proof of the Jordan normal form theorem

Solutions to the August 2008 Qualifying Examination

A FIRST COURSE IN LINEAR ALGEBRA. An Open Text by Ken Kuttler. Matrix Arithmetic

1. Let m 1 and n 1 be two natural numbers such that m > n. Which of the following is/are true?

On a certain type of matrices with an application to experimental design

GOLDEN SEQUENCES OF MATRICES WITH APPLICATIONS TO FIBONACCI ALGEBRA

Hermite normal form: Computation and applications

arxiv: v1 [math.ra] 15 Dec 2015

Chapter 1: Systems of Linear Equations

LINEAR EQUATIONS WITH UNKNOWNS FROM A MULTIPLICATIVE GROUP IN A FUNCTION FIELD. To Professor Wolfgang Schmidt on his 75th birthday

17. C M 2 (C), the set of all 2 2 matrices with complex entries. 19. Is C 3 a real vector space? Explain.

Chapter 5. Linear Algebra. A linear (algebraic) equation in. unknowns, x 1, x 2,..., x n, is. an equation of the form

ON THE "BANG-BANG" CONTROL PROBLEM*

Linear Equations in Linear Algebra

MTH501- Linear Algebra MCQS MIDTERM EXAMINATION ~ LIBRIANSMINE ~

Chapter 5. Linear Algebra. Sections A linear (algebraic) equation in. unknowns, x 1, x 2,..., x n, is. an equation of the form

a 11 x 1 + a 12 x a 1n x n = b 1 a 21 x 1 + a 22 x a 2n x n = b 2.

Stability of a Class of Singular Difference Equations

Matrix Theory and Differential Equations Homework 6 Solutions, 10/5/6

Glossary of Linear Algebra Terms. Prepared by Vince Zaccone For Campus Learning Assistance Services at UCSB

THE GROUP OF UNITS OF SOME FINITE LOCAL RINGS I

JUST THE MATHS UNIT NUMBER 9.3. MATRICES 3 (Matrix inversion & simultaneous equations) A.J.Hobson

Integral Similarity and Commutators of Integral Matrices Thomas J. Laffey and Robert Reams (University College Dublin, Ireland) Abstract

Chapter 1: Systems of linear equations and matrices. Section 1.1: Introduction to systems of linear equations

64 P. ERDOS AND E. G. STRAUS It is possible that (1.1) has only a finite number of solutions for fixed t and variable k > 2, but we cannot prove this

FACULTY OF MATHEMATICAL STUDIES MATHEMATICS FOR PART I ENGINEERING. Lectures

LEGENDRE S THEOREM, LEGRANGE S DESCENT

ABELIAN FUNCTIONS Abel's theorem and the allied theory of theta functions

Sums of diagonalizable matrices

Theorem 5.3. Let E/F, E = F (u), be a simple field extension. Then u is algebraic if and only if E/F is finite. In this case, [E : F ] = deg f u.

NUMBER SYSTEMS. Number theory is the study of the integers. We denote the set of integers by Z:

Mathematics Course 111: Algebra I Part I: Algebraic Structures, Sets and Permutations

arxiv: v1 [math.rt] 7 Oct 2014

GROUPS. Chapter-1 EXAMPLES 1.1. INTRODUCTION 1.2. BINARY OPERATION

INDEPENDENCE OF RESULTANTS TH. MOTZKIN

Eighth Homework Solutions

SOLUTIONS TO EXERCISES FOR. MATHEMATICS 133 Part 2. I. Topics from linear algebra

22.3. Repeated Eigenvalues and Symmetric Matrices. Introduction. Prerequisites. Learning Outcomes

Systems of Linear Equations and Matrices

Elementary linear algebra

Optimization problems on the rank and inertia of the Hermitian matrix expression A BX (BX) with applications

Lattices and Hermite normal form

Simple Lie algebras. Classification and representations. Roots and weights

Systems of Linear Equations and Matrices

3 Matrix Algebra. 3.1 Operations on matrices

Linear Systems and Matrices

EXERCISE SET 5.1. = (kx + kx + k, ky + ky + k ) = (kx + kx + 1, ky + ky + 1) = ((k + )x + 1, (k + )y + 1)

INTEGER ALGORITHMS TO SOLVE LINEAR EQUATIONS AND SYSTEMS

ECON 186 Class Notes: Linear Algebra

IMPROVEMENT OF AN APPROXIMATE SET OF LATENT ROOTS AND MODAL COLUMNS OF A MATRIX BY METHODS AKIN TO THOSE OF CLASSICAL PERTURBATION THEORY

J. D. Botha* Department of Mathematics University of South Africa P.O. Box 392 Pretoria 0003, South Africa

Affine combinations in affine schemes

Math "Matrix Approach to Solving Systems" Bibiana Lopez. November Crafton Hills College. (CHC) 6.3 November / 25

ELEMENTARY LINEAR ALGEBRA

Ma 227 Review for Systems of DEs

7.6 The Inverse of a Square Matrix

Transcription:

Reduction of Singular Pencils of Matrices 1 By W. LEDERMANX. (Received luh October 1934.. Read 2nd November 1934.) 1. Introduction. Let pa + ob = [pa^ + ab^] be a pencil of type m x in', i.e. with m rows and TO' columns, where A and B are matrices with constant elements which are not mere scalar multiples of each other; and p and a are homogeneous parameters. The pencil pa Y + ob 1 of the same type is said to be equivalent to pa + ob if two non-singular constant square matrices P and Q of degree in and m' respectively can be found of such a kind as to yield an equation (1) P(pA+aB)Q = pa 1 + <xb 1 ; ;P 4=0, [Q = =0. Hence the totality of pencils of type m x m' may be divided up into different classes such that all members of a class are equivalent to one another, while no pencils belonging to different classes can be transformed into each other by an equation (1). The problem which now arises, viz. to carry out this classification, was first solved by Weierstrass and Kronecker in classical papers, and has since been treated by many authors. 2 They have distinguished a certain "canonical" pencil in every class such that any pencil is equivalent to one of these canonical pencils. Weierstrass dealt only with the case in which m = ml and the determinant of pa + ob does not vanish identically. The general case which includes rectangular and singular pencils has been treated by Kronecker. According to Kronecker the general canonical form is (2) diag (A,,, A,,,... A Pn, A' qi, A' Qa, A' f;> M) 1 This paper is intended as a continuation of Prof. Turnbull's paper, pages 67 to 76 above. I should like to express my special thanks to Prof. Turnbull for suggesting this investigation to me, and to thank both him and Dr Aitken for their helpful criticism. - Gf. Turnbull and Aitken, Canonical Matrices (1928), p. 125 Jf, where references may be found.

REDUCTION OF SINGULAR PENCILS OF MATRICES where A^, is a pencil of type (p + 1) x p, thus (3) A, [;].*.. P a p. a P a p.. a p a p. a In (2) A'p is the transposed matrix of A p, and M is a non-singular pencil which may be reduced either to Weierstrass's classical canonical shape, the knowledge of which we shall assume, or to a rational form. 1 Kronecker deduced the canonical form (2) under two conditions. In the first place he excluded degenerate pencils: i.e. although the pencil pa + ab is singular it must not be equivalent to a pencil pa 1 + ob x some rows or columns of which are zero. In particular, no non-zero vector u = [u lt u 2,...., u m ] can be found for which ua = ub = 0. For then we could construct a non-singular square matrix U of degree m whose first row is u. The pencil U (pa + ab) = pa 1 + ab, would be degenerate, its first row being zero. It is easy to see that this assumption is not an essential restriction and we shall therefore adopt it following Kronecker. But there is a second hypothesis which was made by Kronecker and most of the other authors 2 which from one point of view seems to be a loss of generality. They postulated that in pa + ab the rank of B should be as great as the rank of pa + ab (identically in p and a). It is always possible to fulfil this condition by introducing new variables p, a' instead of p, a, where p = a u p + «i2 a > a n a i2 a a 2 j p -f- a 22 a, a 2 i a 22 This may be described as changing the basis A, B of the pencil. This process, however, can in general not be effected by an equivalent 1 Gf. Turnbull and Aitken, Canonical Matrices, Chapter IX. 2 Bromwich, however, deals with the general case (Proc. London Math. Soc. (1), 32 (1900)).

W. LEDERMANN transformation (1) so that we lose some classes of pencils if we admit transformations of basis as well as equivalent transformations. This applies also to the non-singular case of a square pencil pa + ob the determinant of which does not vanish identically. It has mostly been assumed that B is non-singular so that the determinant pa + ab has no root p = 0, a =}= 0 or, putting A = o/p, that the determinant A -j- XB has no infinite elementary divisors. In what follows we shall give a new proof for the fact that every pencil can be reduced to the form (4) diag (A,,, A,,, Pn, N ri, n, A' t, A' 9t,..., A' tt, M) Ap being the same as defined in (3) and A' P being its transposed. Here M is a pencil pa 1 + <JB 1 in which \B 1 \ =J=0 so that the Weierstrassian method may be applied. The pencils N r which do not occur in Kronecker's form (2) correspond to the infinite elementary divisors; thus P op the determinant of N r being p r. degree r and op = PI r In (5) I ] r is the unit matrix of... 1. There is no loss of generality in assuming that in pa + ab the number of rows is at least as great as the number of columns, i.e. m Sg m'. If we had originally m < m', we should consider the transposed pencil pa' + ab'. We can transform this pencil into (4) and hence pa + ab into diag (A',,,..., A' Pn, N' ri,... N' rl, A,,,...,,\, M'), involving N' r instead of N T. But as is well known, N and N' are equivalent (they are, in fact, similar), e.g. [ ;][; ']

REDUCTION OF SINGULAR PENCILS OF MATRICES 95 Our proof will partly be based on the LEMMA : The matrix equation for Z, (7) Z = P + QZK, where P and Q are given constant matrices admits of one and only one solution if a power of K vanishes (or if all latent roots of K are zero). Proof: Let K k = 0. Then r=0 is a solution of (7) as is easily verified. In order to prove that there is but one solution we show that the homogeneous equation (7') Y = QYK has only the trivial solution Y = 0. Let Y o be a solution of (7'), thus By iterating this equation we get Y 0 = QY 0 K. 7 0 = QY 0 K = Q 2 Y 0 K 2 =...= Q k ~i Y o K k ~^ = Q* Y 0 K k = 0, since K k = 0. 2. Special Basis for a System of Vectors. Consider a system of k row-vectors of degree m: (1) Z], Z 2,, Z k. If a row-vector z of the same type can be expressed as a linear aggregate of the vectors (1), we write: z C (zi, z 2,, z k ). It will be convenient to introduce a matrix Z the rows of which are the vectors (1). Thus (2) so that Z is of type k x TO. The vectors (1) need not be linearly independent of one another. Let I be their rank (and the rank

D6 W. LEDERMANN of Z). We may then find I basis vectors z ki, z k, z ki out of the system (1) which are linearly independent themselves and which allow every z k of (1) to be represented as a linear aggregate of the basis vectors. The most natural way to construct such a basis is the following: We go through the sequence (1) beginning with Z\ cancelling every vector that is linearly dependent on its predecessors. In particular every zero-vector has to be dropped. The remaining vectors may be called z ki, z kn,...., z t. This basis is uniquely determined by the process and may be named a " special basis." Every z k is a member of the sequence (1) and we have We put &! < k 2 <...< k t. (3) E.g. Consider the set of vectors z lt z 2, z 3 = az^ + /?z 2, z 4, z 5 = yz 1 + Sz 4 z lt z 2, z 4 being independent of one another. Then we have z ki = z lt z k, = Z 2> z k. = z i- 3. Rough Reduction of the Pencil pa + ab. I. DEFINITION. The k linearly independent vectors x u x 2,...., x k form an A-stair if they satisfy the conditions x x B C(0), (i.e. x 1 B = 0) x 2 B C(arii), x 3 B (Z(x 1 A 1 x 2 A), ' ' x 4 B (Z{x x A x, x 2 A lt x s A), x k B C{xiA u x 2 A lt Xt^A). In the notation of 2 (2), we may write this (2) XB = M-XA, where M is a square matrix of degree k in which only the elements below the diagonal can be non-zero. The number k is, of course, less than or equal to m, since there are only m linearly independent

REDUCTION OF SINGULAR PENCILS OF MATRICES 97 vectors x of degree in. Let us suppose that k < m and that the stair cannot be continued. We may add further rows to X to make a non-singular square matrix of degree m, thus Let the rows of Y be y u y 2, y m -k- The vectors (3) x 1 A, x 2 A,, x k A need not be linearly independent. Let their special basis be (4) x ki A, x^a,..., x k{ A which is represented by the matrix X ** A - = XA, K A the rows of (XA) being independent. We shall now prove that the rows of \ \ are independent. Supposing this were not true, we should have a relation (5) (<*i 2/i+ «2 2/ 2 + +a m -ky m -k) B= (Pi Xk> +Pz ^A 2 + +Pi Zk t ) A - The a cannot all vanish for we should tlien get (Pix kt + p 2 x k2 +...+- p t x ki )A = 0 which is impossible because the vectors (4) are independent. Hence V = a l 2/1 + «2 2/2 + + O-m-k Vm-k is non-zero and independent of x lt x 2, - > x k since the rows of the r x ~\ non-singular matrix are independent. From (5) it now follows that yb = (x ki A, x ki A, or since every x k is a certain x^,x ki A); yb CfaA, x 2 A,...., x,.a) which would prolong our stair by another step in contradiction to our hypothesis. Hence (5) is impossible. We may therefore add

98 W. LEDEEMANN \~XA~\ further rows to to form a non-singular square matrix of L YB-i degree n, rxa (6) YB L Z ] whose rows form a basis for all vectors of degree TO'. Let K = 9- According to the properties of our special basis each of the vectors x x A, x 2 A,... x g _ x A can be expressed by x ki A, x k _ A,..., x k x g B (Z(X}A, x 2 A, (by (I)) we may therefore write or in matrix notation, Xg-xA) x kx B C (x kl A, x k.a,, x kk _ x A) (7) XB = K-XA A. Instead of where K (like M in (2)) has non-zero elements only below the main diagonal. As is known, such a matrix has only the latent root zero and a certain power of it must vanish. Consider the matrix YA. As its rows are vectors of degree n they must be expressible by the rows of the matrix (6); thus (8) YA = PXA +QYB + RZ. It is obvious that XA and XB can be expressed by the rows of XA. Let _ (9) XA=F-XA and XB = G XA by (1). [ X ~~\ we add a certain aggregate of x lt x 2, x k or of x ki, x kn,..., x k to every row of Y the matrix will still be nonsingular. We may for example replace 7 by Y x = Y & X where S is an arbitrary matrix of type (m k) X I which we shall choose in a suitable way. If we carry out this substitution in (8), we get and by (7) T 1 A = (P - S3) XA + Q (F, + SI) B + RZ

REDUCTION OF SINGULAR PENCILS OF MATRICES 99 According to the lemma of 1 we can choose S so as to make vanish. Hence (10) 7 1 A If we now multiply the original pencil by, we get by (9) and (10) fpf+ L 0 +og The last matrix is non-singular, because PQ 01 orj XA' Y 1 B ~ XA ~ Y-, B = Z YB _ XA -&XB Z XA YB-S Z _ ' KXA (7) Hence the pencil = ~ I t <<K I XA' YB Z abr 1 = PQ + is equivalent to the original pencil. But pa 1 -f ab 1 splits up into two pencils with fewer rows and columns unless k = m (p. 93). Therefore if k < m, the proof is completed by induction. II. We shall now suppose that k = m, i.e. the longest jl-stair contains m independent vectors x x, x 2,...., x m. We may assume that the original pencil has this property. According to (2) we have (2) XB = MXA, where now X is a non-singular square matrix of degree m and M is a matrix with zero latent roots only. We have to distinguish two cases. (a) In pa + ob the matrix A has no row dependence: i.e. there is no vector y =(= 0 for which ya = 0. Since we had assumed m 2: m' it follows m = TO' and \A =(=0. The reduction of pa -\- ob can easily be performed; multiply by X: X { P A + ob) = PXA + axb = (pi + om) XA,

100 W. LEDERMANN by (2), where X and XA are non-singular. We may therefore continue by reducing pi + om. Since M has only the latent root 0, the Weierstrassian form of M will be PMP- 1 = diag (H ri, H r.,... H ti ), r x + r 2 +.... + r, = m = m' where H r = 1 Hence P( P I +om)p~ 1 = diag ( P I ri = diag (N r, which proves the theorem. oh ri, PI r.. + ah T., N fl,n r% ) (b) We have now to deal with the more difficult case when a vector 2/ =j= 0 exists for which ya = 0. It is then possible to construct a " 5-stair " in the same way as in I only with A and B interchanged. Every other step remains unaltered: We construct a stair whose length 1 may be I. If I be less than m, we should again be able to split up the pencil and the proof would be concluded by induction. We shall therefore suppose that not only the ^4-stair but also the.b-stair exhausts the whole m-dimensional vector-space. Writing these conditions down in full, we have x x B = 0 (11) x 2 Bd{x x A) (a) x 3 B C (xj A, x 2 A) y 3 A C (y-y B, y 2 B) x m B Cl(x 1 A,x 2 A, x m - t A) y m AC(yiB, y. 2 B,, where x lt x 2,...., x m and y x, y 2 y m are two sets of m linearly independent vectors of degree m. Pencils pa + ob with the properties (1) require a more elaborate study which we are going to explain in 4. 4. Reduction by means of Vector Chains. Let pa + ob be a pencil which fulfils the condition (11) of 3, i.e. we assume that at least one U-stair and one A -stair exists, each of length m. But it is easy to see that every non-zero vector z that annihilates B can be extended to a stair of m elements unless the 1 By saying the stair is of length I, we mean that it consists of I vectors and cannot be continued by another vector. PI r

REDUCTION OF SINGULAR PENCILS OP MATRICES 101 pencil splits up into two pieces. For if a stair beginning with z should break down at the k ih step, i.e. if the stair be of length k (k < m), we could split up the pencil as shown in 3. From 3(11), we see that for every vector u we can find a vector u such that (1) ua = ilb, for u must be a linear aggregate of y lt y 2,... y m whence the existence of u is evident from 3(11)/?. It is not self-evident that the unknown components of the vector u can be calculated from the nonhomogeneous equation (1) because the coefficients of the unknowns do not form a non-singular matrix. The vector u, however, is not uniquely determined. Let v 0 B = 0 (v o =j=o). We may then determine other vectors v 1: v 2, v Pi,..., which form the following "vector chain." (cf. Turnbull, page 72 of this volume.) (2) 0 = v 0 B, v 0 A = v x B, v x A = v 2 B,...., v Pi - X A = v Pi B, v Pi A = v Pi+1 B We can continue the chain as long as we want, but the vectors occurring in it will not be linearly independent. Let v Vi A be the first vector in (2) to be linearly dependent on its predecessors v 0 A,ViA,...., v Pl _i A. We then have the relation (3) ( S a, v Pi _ ) ^4=0, where a 0 =(= 0. It is convenient to put (4) «_4=0, ft= 1,2,3,..., making the equation # _! A = v v B valid also for zero and negative integers. The number p lt i.e. the number of consecutive linearly independent vectors in (2) starting with v 0 A is called the length of the chain. The length is always positive, otherwise we should have v 0 B = v 0 A=0 and the pencil pa + ab would be degenerate ( 1). Let p\ be as small as possible. We derive another chain from (2) by putting (5) dp=va,v t - v (k^pl ). v = 0 In fact, the u^ form a chain, for by (2) u B = ( S a,t; t _,) B = [Z drvt-!-,) A = u^a.

102 W. LEDERMANN In particular u^ B = 0 by (4) and <> 4=(L,»,_,) 4=Oby (3). We have therefore constructed the chain (6) 0--=«<*>5, >.4 =<> B, <>.4 =<> J3,..... u^\_ x A =««5, njm =0 The vectors u^)a,u ( pa,...., vt^^a must be independent, otherwise we could build up a chain of length less than p x which would be contradictory. If there is a vector u^=\=0 which annihilates B and which is independent of the first chain, i.e. of the vectors uf^\ vhp,...., u^l we form another chain like (6) the length p 2 of which shall be taken as small as possible. Naturally px 5S p 2. We then proceed to a third chain provided that its first or " leading " vector «Q 3) is independent of all vectors of the first and second chain its length. p 3 being minimal. In this way we get a whole system of chains 0=>, (7) 0=««>.B, 0=uf>B, u^a =u ( { l) B, u^a = u^b,, v! >_ x A =u^b, u% A =0 As we have shown, this system possesses the following properties : (a) The lengths are increasing (8) p x <,p 2 <* ^p n. (b) The first vector of every chain is independent of all vectors of the preceding chains. (c) Each length is as small as possible, i.e. there is no chain independent of the first chain whose length is less than p 2, nor does a chain exist whose first vector is independent of the first and second chains and the length of which is less than p 3, etc. (d) We have exhausted all chains, i.e. we cannot find any vector u(n+v f or wn ich u (^+1) B=0 unless tt^t 1) is a linear aggregate of the previous chains. We shall now prove that the vectors {9) Um A, «4» A,.., «_! A> u<*>a,ufa,.., «w_ 1 A,.., up A, <U,.., A are independent of one another. If this were not so, we should have a relation S 0«0

REDUCTION OF SINGULAR PENCILS OF MATRICES where (11) q T^p T l, and U' T) is the last element of the -rth chain that really enters therelation (10) with a non-zero coefficient jsw=^o. If the rth chain does not occur at all in (10), we put q r =0 and $, T) = 0. Let q t be the maximum of q u q 2,.., q n ; if several q areequally great, we take g as great as possible so that (12) q g ~^1k (ft = 1, 2..., 0) ; q g > q K (A = g + 1,.., h). We now construct the chain (13)» t = S S j8w ^{?T _^_^T. T = 1 y.t = 0 In fact, the vectors v 0, v lt.., v q form a chain. For \ = v k _ 1 A because according to the chain properties (7) we have- «W B ufl-y A for every v g p r. In particular we get v 0 B = v_\a =0 and v q ^4 = 0 by (10).. Also v 0 reduces to *o = )3{,«u%_ 9g + S<?> u _ tg + + /3<r> <> (by (4) and ft> ^ 0). The suffixes of the u are either 0 or negative since q 2; q (T = 1, 2,.. h). All terms behind the gr th term could be dropped because q g > q K for A > g. v 0 is independent of the first, second,.., (g l) th chain. For, otherwise u$ would be dependent upon its predecessors in contradiction to (b). It is therefore permissible to start the g th chain with v 0 instead of w{f>. But the length of the v- chain is q ^ p 1 or less, viz. if the vectors v 0 A, v^ A,.., v q _ X A be linearly dependent. In any case the length of this modified g th chain would be smaller than p g which contradicts (c). Hence the vectors (9) must be independent of each other. We shall now show that also the vectors (14) <>, <>,..., «; uf\ uf,...,<;... ; <\ < «are linearly independent. If there were a relation between them, it could be written : (15) Yl <> + y 2 «go +... + Yn <) + S S 8 «u<;> = o. l 1

104 W. LEDEBMAKN The S< cannot all vanish. For then the "leading" vectors u ( v, uf },.., u {^> would be dependent in contradiction to (b). Multiplying (15) by B we get (2 S8W)B = 0 since utf B = 0 ; applying the chain properties (7) we have (S S3W +1 «W)i=0 which is incompatible with the vectors (9) being independent. Hence the vectors (14) are independent. What are the connections between the vector chains and the reduction of the pencil pa + ob% Consider one of the chains (7): 0=utfB, u$a = u^b, u^a = u$b, u^_ x A =«W B, ««A = 0. Let (16) U T = u\,,(r) *1>T It follows by (7) that U r {pa + ab) = and U T = + (T = 1, 2,.., n). + and PT P.(17) U T (pa + ob) = a p a p = A PT U T A, where Ap T has been defined in 1 (3). Hence... p. a u 9. U n ( P A + ob) = U n (pa A A Un A

REDUCTION OF SINGULAR PENCILS OF MATRICES 105 or _ (18) U( P A + ob)=a UA, where A = diag (Ax, A 2,..., A n ) and _ u x th (19) U = U * and U = ^2. U n U n _ Obviously, the equations (7) can be interpreted as a vector A-stairinthe sense explained in 2. It contains k=(p 1 + l) + [p 2 + 1) + + (Pn + 1) vectors the independency of which we have proved. We shall show that k = m. If k < m, it must be possible tccontinue the stair by another vector z such that z being independent of all w. By (7) we may write instead of (20) or in full / z ( (1) ^d)_ _ 6 (l) -ji<l)- -...- - e (n)^«) ) ) = 0 e^t) being certain coefficients. Here we should have obtained a vector which is independent of the u and yet annihilates B in contradiction to condition d). Hence k must be m and U has m rows and is therefore square and non-singular. Finally, we shall show that also UA is square (of degree n). If it were not so, we could add further rows to make a non-singular square matrix. From (13) we should then get Hence [A, 0] would be equivalent to pa + ab but it contains null rows and columns which we had excluded. The matrix Z must therefore be illusory and (18) may be written as V ( P A + ab) (UA)- 1 = diag (A b A 2,.., AJ. This completes the proof. In his paper Professor Turnbull has shown how the minimal vector chains are connected with Kronecker's minimal relations between the rows of the pencil pa + ab. In particular, it has been pointed out that the lengths of the vector chains (7) are identical with Kronecker's Minimalgradzahlen.