NEW ALGORITHM FOR MINIMAL SOLUTION OF LINEAR POLYNOMIAL EQUATIONS

Similar documents
How might we evaluate this? Suppose that, by some good luck, we knew that. x 2 5. x 2 dx 5

Lecture 7: Polynomial rings

Basic Algebra. Final Version, August, 2006 For Publication by Birkhäuser Boston Along with a Companion Volume Advanced Algebra In the Series

Linear Cyclic Codes. Polynomial Word 1 + x + x x 4 + x 5 + x x + x

POLYNOMIALS. x + 1 x x 4 + x 3. x x 3 x 2. x x 2 + x. x + 1 x 1

A field F is a set of numbers that includes the two numbers 0 and 1 and satisfies the properties:

(x + 1)(x 2) = 4. x

March Algebra 2 Question 1. March Algebra 2 Question 1

3.3 Dividing Polynomials. Copyright Cengage Learning. All rights reserved.

7x 5 x 2 x + 2. = 7x 5. (x + 1)(x 2). 4 x

Linear Cyclic Codes. Polynomial Word 1 + x + x x 4 + x 5 + x x + x f(x) = q(x)h(x) + r(x),

Ch 7 Summary - POLYNOMIAL FUNCTIONS

We say that a polynomial is in the standard form if it is written in the order of decreasing exponents of x. Operations on polynomials:

Review all the activities leading to Midterm 3. Review all the problems in the previous online homework sets (8+9+10).

Dividing Polynomials: Remainder and Factor Theorems

LINEAR SYSTEMS AND MATRICES

Complex Numbers: Definition: A complex number is a number of the form: z = a + bi where a, b are real numbers and i is a symbol with the property: i

Polynomials. Chapter 4

1.3 Algebraic Expressions. Copyright Cengage Learning. All rights reserved.

An Invitation to Mathematics Prof. Sankaran Vishwanath Institute of Mathematical Science, Chennai. Unit - I Polynomials Lecture 1B Long Division

Example: This theorem is the easiest way to test an ideal (or an element) is prime. Z[x] (x)

8.4 Partial Fractions

Section III.6. Factorization in Polynomial Rings

Algebra III Chapter 2 Note Packet. Section 2.1: Polynomial Functions

L1 2.1 Long Division of Polynomials and The Remainder Theorem Lesson MHF4U Jensen

( 3) ( ) ( ) ( ) ( ) ( )

Part IX. Factorization

MTH310 EXAM 2 REVIEW

D i v i s i b i l i t y

Mathematics 136 Calculus 2 Everything You Need Or Want To Know About Partial Fractions (and maybe more!) October 19 and 21, 2016

Finite Fields: An introduction through exercises Jonathan Buss Spring 2014

L1 2.1 Long Division of Polynomials and The Remainder Theorem Lesson MHF4U Jensen

Contents. 4 Arithmetic and Unique Factorization in Integral Domains. 4.1 Euclidean Domains and Principal Ideal Domains

Chapter 14: Divisibility and factorization

Kybernetika. Vladimír Kučera Linear quadratic control. State space vs. polynomial equations

, a 1. , a 2. ,..., a n

Section 4.3. Polynomial Division; The Remainder Theorem and the Factor Theorem

Mathematical Olympiad Training Polynomials

A quadratic expression is a mathematical expression that can be written in the form 2

ESTIMATION OF POLYNOMIAL ROOTS BY CONTINUED FRACTIONS

Fast Polynomial Multiplication

3 Polynomial and Rational Functions

Rational Functions. Elementary Functions. Algebra with mixed fractions. Algebra with mixed fractions

Partial Fractions. June 27, In this section, we will learn to integrate another class of functions: the rational functions.

Chapter 4. Remember: F will always stand for a field.

Polynomials. In many problems, it is useful to write polynomials as products. For example, when solving equations: Example:

Homework 8 Solutions to Selected Problems

Chapter 2 Formulas and Definitions:

(2) Dividing both sides of the equation in (1) by the divisor, 3, gives: =

be any ring homomorphism and let s S be any element of S. Then there is a unique ring homomorphism

Algebra Review 2. 1 Fields. A field is an extension of the concept of a group.

Partial Fractions. Combining fractions over a common denominator is a familiar operation from algebra: 2 x 3 + 3

Linear Algebra 1 Exam 2 Solutions 7/14/3

Kybernetika. Milan Mareš On fuzzy-quantities with real and integer values. Terms of use:

0 Sets and Induction. Sets

8 Appendix: Polynomial Rings

Chapter Five Notes N P U2C5

Contribution of Problems

Section Properties of Rational Expressions

6x 3 12x 2 7x 2 +16x 7x 2 +14x 2x 4

Closed-Loop Stability of Discrete Linear Single-Variable Systems

ELEMENTARY LINEAR ALGEBRA

Solutions to Exercises, Section 2.4

Today. Polynomials. Secret Sharing.

a + bi by sending α = a + bi to a 2 + b 2. To see properties (1) and (2), it helps to think of complex numbers in polar coordinates:

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Discussion 6B Solution

Proofs. Chapter 2 P P Q Q

ELEMENTARY LINEAR ALGEBRA

On the Numerical Solution of Implicit Two-Point Boundary-Value Problems

PARTIAL FRACTIONS AND POLYNOMIAL LONG DIVISION. The basic aim of this note is to describe how to break rational functions into pieces.

Permutations and Polynomials Sarah Kitchen February 7, 2006

Math 547, Exam 2 Information.

a 11 x 1 + a 12 x a 1n x n = b 1 a 21 x 1 + a 22 x a 2n x n = b 2.

(Rgs) Rings Math 683L (Summer 2003)

Lesson 7.1 Polynomial Degree and Finite Differences

Polynomial Harmonic Decompositions

Questionnaire for CSET Mathematics subset 1


Equations and Inequalities

ELEMENTARY LINEAR ALGEBRA

Groups, Rings, and Finite Fields. Andreas Klappenecker. September 12, 2002

Remainders. We learned how to multiply and divide in elementary

MA22S3 Summary Sheet: Ordinary Differential Equations

Factorization in Polynomial Rings

ALGEBRA. 1. Some elementary number theory 1.1. Primes and divisibility. We denote the collection of integers

PARTIAL FRACTIONS. Introduction

MATH 150 Pre-Calculus

Chapter 2 Polynomial and Rational Functions

COMPUTER ARITHMETIC. 13/05/2010 cryptography - math background pp. 1 / 162

Proofs. Chapter 2 P P Q Q

Selected Math 553 Homework Solutions

Commutative Rings and Fields

11 Division Mod n, Linear Integer Equations, Random Numbers, The Fundamental Theorem of Arithmetic

A matrix over a field F is a rectangular array of elements from F. The symbol

NOTES ON SIMPLE NUMBER THEORY

Chapter 3: Polynomial and Rational Functions

Answers. 2. List all theoretically possible rational roots of the polynomial: P(x) = 2x + 3x + 10x + 14x ) = A( x 4 + 3x 2 4)

Solutions to the Worksheet on Polynomials and Rational Functions

Generator Matrix. Theorem 6: If the generator polynomial g(x) of C has degree n-k then C is an [n,k]-cyclic code. If g(x) = a 0. a 1 a n k 1.

CHAPTER 2 POLYNOMIALS KEY POINTS

Transcription:

KYBERNETIKA - VOLUME 18 (1982), NUMBER 6 NEW ALGORITHM FOR MINIMAL SOLUTION OF LINEAR POLYNOMIAL EQUATIONS JAN JEZEK In this paper, a new efficient algorithm is described for computing the minimal-degree solution of the linear polynomial equation. This problem occurs in linear control synthesis. 0. INTRODUCTION The polynomial equation (1) a(x)x(x) + b(x)y(x) = c(x) plays central role in the linear control theory. It occurs in optimal control synthesis, in dead-beat as well as in quadratic criteria [1, 2, 3]. In this equation, a(x), b(x), c(x) are given polynomials, x(x), y(x) are unknown ones. We suppose a(x), b(x) are not both identically zero. Main tool for studying this Euclidean equation is the Euclidean algorithm. It yields general solution and can serve as a computing algorithm for it. In control theory applications, some particular solutions are of interest: those with degree of x or y (or both) minimal. Algorithms are known for selecting the minimal solution from the general one. In this paper, a new algorithm is described which computes the minimal solution directly. This way is more efficient than the general solution computation. The method is also based on the Euclidean algorithm. The indeterminate X can represent either the derivative operator s in the continuous control theory or the delay operator in the discrete case. We shall also write a instead of a(x) for brevity. By da we mean the degree of polynomial a(x), da = GO means zero polynomial a(x) = 0. The leading coefficient of a(x) is denoted by a da. By gcd (a, b) we mean the greatest common divisor of polynomials a, b. 505

1. PRELIMINARIES Let us summarize the theory of polynomial equation briefly. Let us denote d(x) = gcd (a(x), 6(A)), a 0(X) = *j&, d(x) b 0(X) = %& d(x) Theorem 1. The general solution of the homogeneous equation (2) a(x)x(x) + b(x)y(x) = 0 is (3) x(x) = b 0(X)t(X), y(x)~-a 0(X)t(X) where t(x) is an arbitrary polynomial. Proof - see [2, 3]. Theorem 2. The equation (l) is solvable iff d(x) divides c(x). Proof - see [2, 3]. Theorem 3. The general solution of (l) is (4) x(x) = x p(x) + b 0(X) t(x), y(x) = y p(x) - a 0(X) t(x) where x p(x), y p(x) is any particular solution, t(x) is an arbitrary polynomial. Proof follows from linearity of (1) and from Theorem 1. Theorem 4. The equation (l) with a(x) ^ 0, if solvable, has the unique "y-minimal" solution x(x), y(x) defined by (5) 8y < da - 8d For this solution, (6a) dx < db - dd for dc < da + db - dd (6b) dx = dc - da for dc da + 8b - dd Proof. Let x p(x), y p(x) be a particular solution of (l). As a 0(X) ^ 0, let us take the quotient q(x) and the remainder y(x): y p(x) = a 0(X) q(x) + y(x), According to Theorem 3, the couple y(x) = y p(x) - a 0(X) q(x), dy < 8a 0 = da - dd x(x) = x p(x) + b 0(X) q(x) satisfies (1). Existence of y-minimal solution is proved. Let us prove its uniqueness. Let x u y^ and x 2, y 2 be two ^-minimal solutions, i.e. dy-l < da 0, dy 2 < da 0. The couple x = x l x 2, y = y L y 2 satisfies the homo- 506

geneous equation (2), dy < da 0. But according to (3), for y ^ 0 it must be dy ^ da 0. Therefore y = 0, y t = y 2 - From (2), x = 0, x t = x 2. Now we shall prove the property (6). The equation (l) yields by cancellation of d(x) an equivalent equation (7) a 0 (X)x(X) + b 0 (X) y(x) = c 0 (X) where c 0 (X) = c(x)jd(x). Let us consider the degrees of all three terms: a) c(x) = 0, the j-minimal solution is x(x) = 0, y(x) = 0. For b(x) ^ 0 (6a) is satisfied, for b(x) = 0 (6b) is. b) dc < da + db dd o dc 0 < da 0 + db 0 d(b 0 y) < da 0 + db 0 => d(a 0 x) < da 0 + db 0 => ex < db 0 = db dd, hence (6a) is satisfied. c) dc *i da + db dd o dc 0 ^ da 0 + db 0 d(b 0 y) < da 0 + db 0 S dc 0 => d(a 0 x) = dc 0 => dx = dc 0 - da 0 = dc - da, hence (6b) is satisfied. Theorem 5. The equation (l) with b(l) ^ 0, if solvable, has the unique "x-minimal" solution x(x), y(x) defined by For this solution, dy < da dd for dy = dc db for dx < db - dd dc < da + db dd dc ^ da + 8b dd Proof. If we interchange in (l) a with b and x with y the equation remains invariant. Therefore, Theorem 5 follows from Theorem 4. Theorem 6. The equation (l) with a(x) ^ 0, b(k) ^0, dc < da + db dd, if solvable, has the unique "minimal" solution x, y defined by one of the two equivalent conditions: dx < db dd or dy < da dd Proof. According to Theorem 4, the j-minimal solution exists uniquely. It satisfies (6a), i.e. it is at the same time x minimal. 2. THE ALGORITHM We shall construct the algorithm as a recurrent process with finite number of steps. In every step, the equation is replaced by an equivalent one by a substitution. The final equation can be solved in a trivial way. By backward substitution chain, we shall obtain the solution of the original equation. 507

First, we shall prove four lemmas for the recurrent process. Lemma 1. Let (l) be solvable, a(k) s 0, dc 2: da. Then the equation (8) a(x)x'(x) + b())y(x) = c'(x) where (9) c'u) = c(x) - ^ A ec ~ B a()), dc' < 3c a ca is also solvable, has the >'-minimal solution x', y and the _y-minimal solution x, y of (l) is given by (10) x(x) = x'()) + -^ k ac ' da a ea Proof. Let d = gcd (a, b). From (9), d divides c' if it divides c, i.e. (8) is solvable iff (l) is. As a(x) = = 0, the x', y exists. By substituting (10) into (l) we can see that x, y is a solution. As x', y is v-minimal for (8), By < da dd. Therefore x, y is v-minimal for (1). Lemma 2. Let (l) be solvable, b(x) ^ 0, dc ^ db. Then the equation (11) a(x)x(x) + b(x)y'(x) = c'(x) where (12) c'u) = c(x) - ^ X Se - ib b(x), dc' < dc is also solvable, has the x-minimal solution x, y' and the x-minimal solution x, y of (1) is given by (13) y()) = y'(a)+ C f-) Sc - eb Proof. By interchanging a with b and x with y in Lemma 1. Lemma 3. Let (l) be solvable, a(x) SJE 0, dc < da + db dd, db jg da. Then the equation (14) a())x'()) + b'(x)y(x) = c(x) where (15) b'(x) = b(x) - ^ ) d b - e a a(x), db' < db a da is also solvable, has the j-minimal solution x', y and the minimal solution x, y of (l) is given by (16) x(x) = x'())-^a Sb - da y()) b db 508

Proof. Let d = gcd (a, b), d' = gcd (a, b'). From (15), d divides b' => d divides d'. By the same argument, d' divides b => d' divides d. Therefore d' = d, i.e. (14) is solvable iff (l) is. As a(?) ^ 0,the x', y exists. By substituting (16) into (1) we can see that x, j is a solution. As x', y is j-minimal for (16), By < da dd. Therefore x, y is j'-minimal for (l). As dc < da + db dd, it is minimal. Lemma 4. Let (l) be solvable, b(x) 4= 0, dc < da + db - dd, da ^ db. Then the equation (17) a'(x)x(x) + b(x)y'(x) = c(x) where (18) a'(x) = a(x) - ^2 X Sa - eb b(l), da' < da b db is also solvable, has the x-minimal solution x, y' and the minimal solution x, y of (l) is given by (19) y(x) = y'(x)- a fx s ~ eb x(x) Proof. By interchanging a with b and x with y in Lemma 3. Now we are ready to construct the algorithm. Theorem 7. Let a, b, c be given polynomials in (l). Let p be a two-valued variable with values p x and p y; here p x means "the x-minimal solution is to be found", analogously p y. We suppose the equation is solvable and a ^ 0 for p = p x, b =E 0 for P = Pv Let us construct a sequence a (i \ b ( '>, c (i \ p (i \i = 0, 1,2... with initial values a (0) = = a, b (0) = b, c (0) = c, p (0) = p by the following recurrent rule: If (20a) a (i) = 0, b (i) 0, c (i) = 0, p (i> = p x or (20b) a (i) $ 0, b (» ~ 0, c (i) = 0, p (i > = p y then stop, it is the last element of the sequence. Otherwise, if p (i) = p x we distinguish two cases: 1) dc (i) = db (i). it is always b (i> ^ 0. We use Lemma 2 and define c (i + 1 > by (12), p(h-u = p(f), 2) 5c (i) < 3fc (i) => 5c (i) < da (i) + db (i) - ^, we distinguish two subcases: 2a) db (i > > da (i), for c (i) ^ 0 it is always a (i) $ 0 (case c (i) a 0, a (i) s o see (20)). We use Lemma 3 and define!> (, ' +1) by (15), p (i+1 > = Py. 2b) db (i > g 5a (i), it is always b (i > 0. We use Lemma 4 and define a (,+1) by (18), p«"> = p,. 509

If p"> = p y we analogously have two cases 3) and 4) which are literally the same as 1) and 2) up to interchanging a with b and x with y. Then for every i up to some final value i F the rule is properly defined and a U), b U), c (0, p U) represent the problem of finding x (0 -minimal or }> (0 -minimal solution of equation (21) a u >(X) x u) (X) + b u >(X) / (A) = c U) (X) equivalent to (l). For the final value i F the state (20) is reached whose solution is X = 0, y = 0. Proof. We must prove that the rule is properly defined, i.e. the cases 1) p u > = p x, õc U) 1 ôb u), b U) = 0 2a) P in = p x, дc U) <õb U), c (i) ф 0, a U) = 0 2b) p U) = p x, дc U) < õb U), b U) з 0 3), 4a), 4b) similarly can never happen and that a u), b u ', c u >, p u ' really represent the problem (21), equivalent to (l). We shall do it by induction. For i = 0, cases l) and 2b) are excluded by assumptions of the theorem. In case 2a) the equation is b(x) y(x) = c(x). To be solvable, it must be either 8b _ Be or c = 0, so 2a) is also excluded. Let for some i the rule be properly defined and (21) equivalent to (1). Then a (i+1), /j (l+1), c (i + 1), p { + 1) ' as defined by the rule, always lead to an equivalent equation, solvable and never violating the condition a U) ^ 0 for p u > = p y or b u > ^ 0 for p u> = p x. So the next step is also properly defined and the induction is completed. Now we shall prove that for some finite i we must reach (20a) or (20b). Let us suppose that (20) is not satisfied for i = 0, the case a = 0, b = 0 is excluded. We shall prove that for some i one of polynomials «(0, b U), c u) is identically zero. Let us suppose it is not true for i = 0. In every step, just one of three polynomials a U), b U), c u > is changed, either 8a u+i) < 8a U) or 8b u+1) < db U) or 8c u+1 > < 8c u > is valid. Therefore [(8a + 1) (8b + l)(8c + l)] (i+1) < [(da + 1) (8b + 1) (5c + l)] (0 isalso valid. Having started from a finite positive value, this integer function must reach zero for some i, i.e. some of the a u), b u>, c u > is zero. If c (0 is the case then the sequence continues but only steps 2a) or 2b) are possible. Now [(8a + l)(8b + l)] (i+1) < [(8a + l)(8b + l)] (0 and for some i, a (0 s 0, p (0 = p x (the last step was 2b) or b u)» 0, p U) = p y (2a). If a (0 was the case of zero polynomial then p u) = p x, the equation is b U) u (X) y >(x) = = c U) (X), solvable as proved, i.e. dc U) - 8b U) or c (0 = 0. The sequence continues with steps 1) only possible. For some i we must reach 8c u > < 8b u > i.e. c (i) s o. Similarly in the case of zero b u >. Q 510

3. COMPUTATIONAL ASPECTS The construction just described can serve as an efficient algorithm for a computer or for a paper-and-pencil work. It has the property that neither d(x) nor dd must be known in advance. They are obtained as a result: in (20a), there is fc (0 (A) = d(x), similarly in (20b) a < - i) (X) = d(x). The case of unsolvable equation (d(x) does not divide c(xj) is recognized by reaching the state a (0 = 0, i> (0 = 0, c (0 = = 0, dc 0) < db {i \ p<- l) = p x or a (0 $ 0, fo (0 = 0, c (0 ^ 0, dc (0 < cv, p (0 = p y. The algorithm consists of a forward run and backward one. In the latter one, all "planned" substitutions are performed in backward order, from x (,p), _y ( ' F) to x w, >' (0). In every step of the forward run, some of the polynomials a, b, c are "reduced" by some other one. It is useful to have a subroutine for such a reduction, say, of r(x) by s(x): or ^ ds, r'(x) = r(x) - ^ A dr ~ es s(x) i.e. from r(x) subtract s(x) multiplied by a number q and by l k such that the leading term gets cancelled. The real number q = r Srjs ds and the integer number k = dr ds are also results of the operation. The reduction can be done "in place" with no new memory for r' needed. In the backward run where (16) or (19) is applied, it is useful to have a subroutine for "forced reduction" of r(x) by s(x). It means: for given polynomials r(x), s(x), given real q and integer k ^ 0, compute r'(x) = r(x)- qx*s(x) i.e. from r(x) subtract s(x) multiplied by a prescribed number and a prescribed power of A. It can also be done in place. The operation (10) and (13) means: given a polynomial r(l), given number q and integer k, add qx k to r(x). It can also be done in place and it is useful to have a subroutine or a standard way for doing it. During the forward run, we must keep track of three things: a) the information on which of the four reduction types (9), (12), (15), (18) was used, b) the number q by which the polynomial was multiplied, c) the integer k by which the polynomial was shifted. These data will be used in the backward run for selecting the proper substitution and for doing it. The data are stored and fetched by the stack discipline (first in, last out). If your computer and programming language support recursive procedures then they can be utilized for that, otherwise you must program the stack explicitly in arrays. 511

Flowcharts of the forward and backward runs are in Fig. 1, 2. The four-valued variable f stores the information on the substitution type. Its values are mnemonic names for substitutions. Stack manipulation is done by operations "store" and "fetch". reduce c by b reduce b by a reduce a by b reduce c by a ífij \P~PҲ t: = 'y plus" í: = "X by y' t: = 7 by X " t: = "x plus' store í, q, k Fig. 1. Flowchart of the forward run. 512

Fig. 2. Flowchart of the backward run. 4. EXAMPLES The algorithm was implemented on IBM 370 computer in FORTRAN language. Here presented are four extremely simple examples which can be easily computed with paper and pencil. In Tables 1, 2 the case a(x) = 1 + 32 + 2X 2, b(x) =1 + 22, c(x) = 2 + 2X 2 is solved for the y-minimal solution x(x) = 1, y(x) = 1 and for the x-minimal solution x(x) = 0, y(x) = X. The polynomials a, b, c have the common factor d(x) =1+22. After cancelling it, we have a new equation a(x) =1 + 2, b(x) = 1, c(x) = X with the same solution. This is solved in Tables 3, 4. Table 1. Equation with a common factor,.y-minimal solution a b c P t aѓ t X У 1 + ЗД + 2A 2 1 + 2A A + 2A 2 Py 1 + ЗA + 2A 2 1 + 2A -1-2A Py 1 + 2A 1 + 2A -1-2A Px 1 + 2Д 1 + 2A 0 Px 1 + 2A Py cъy a 1 дrplus aъy b A yъy x cъy b 1 y plus bъy a 1 xъyy 1-1 0-1 0-1 513

Table 2. Equation with a common factor, ^-minimal solution a b c p t aŕ t X У i + з;. + 2x 2 1 + 2Л X+ 2X 2 Px 1 + ЗЛ + 2Л 2 1 + 2Л 0 Px 1 + 2X 1 + 2Л 0 Px 1 -ř- 2X Py cъy b X y plus aъy b X yъy x ЬЪy a 1 xъyy 0 k Table 3. Equation without common factor, >>-minimal solution a ь c p t ax / k X У 1 + X 1 ;. Py 1 + X 1 -i Py 1 1 -i Px 1 1 0 Px 1 Py cъy a 1 xplus aъy b X У DУ X cъy b -1 y plus ЬЪy a 1 xъyy 1 -i 0 1 0-1 Table 4. Equation without common factor, x-minimal solution a b c p t aŕ / X У 1 i +;. i л Px 0 X c Ъy b Л y plus 1 + л > Px aby b Л yъyx 1 0 Px bъy a 1 xъyy i Py 5. COMPUTATIONAL COMPLEXITY Let us estimate the number of numerical operations and the amount of storage needed for the above method and compare them with those for other methods. A suitable unit of computational complexity is an operation of "reduction": sub- 514

trading a multiple of a number from (or adding to) another number. For simplicity, let da = db = 5c = N, Sd = 0. The complexity functions f(n), g(n) for number of operations and number of storage places depending on N are polynomials, e.g. f(n) = a. + /3N + yn 2. Because mainly the behavior of/(at) for great N is interesting, we neglect all but the leading term yn 2. We do this systematically in all stages during complexity function calculation, e.g. we replace N + 1 by At; it simplifies the job significantly. Three methods are investigated: 1) The above method. In the forward run, 5a, db, 5c are lowered from A* to 0 N with 3 YJ n = fn 2 reductions needed. In the backward run, 5x and Sy grow from 0 n = 0 JV to At with 2 n = N 2 operations needed. Total fn 2 operations. Auxiliary storage n = 0 needed is At reals, At integers and N four-valued codewords. 2) The general solution method [2, 3]. It computes polynomials p,q,r,s satisfying ^-řïj-й-ш The matrix P represents accumulated row operations on the vector A. The degrees N dp, Sq, dr, 5s grow from 0 to At, the degrees da, db fall from At to 0 with 6 n = 3/V 2 n = 0 reductions needed. Thenafter the particular solution x 0 = pc, y 0 = qc is recomputed to the minimal one by taking the remainder in division by b resp. by a. This multiplication and division needs AN 2 operations. Total 7At 2 operations is worse than that of the above method by factor 2-8. Auxiliary storage 4At for p, q, r, s is also greater. 3) The probable most wide-spread method of indeterminate coefficients. By it, the polynomial equation is replaced by a set of M = 2At numerical equations for coefficients. The number of operations for solving by elimination is known to be M 3 /3 = At 3, the storage M 2 = 4A/ 2. These formulae show that this method compared with any polynomial one is worse by an order of magnitude: At 3 instead of A 72 operations, At 2 instead of At storage. 6. CONCLUSION The method is very simple, effective and easy to learn. Number of operations needed is less than that in any other method known. ACKNOWLEDGEMENT The author wishes to thank Ing. V. Kucera DrSc. and prof. Ing. V. Strejc DrSc. for valuable comments on the algorithm. (Received March 17, 1982.) 515

REFERENCES [1] V. Kučera: Algebraic theory of discrete optimal control for single-variable systems I Preliminaries. Kybernetika 9 (1973), 2, 94 107. [2] V. Kučera: Algebraická teorie diskrétního lineárního řízení. (Algebraic Theory of Discrete Linear Control). Academia, Praha 1978.. [3] V. Kučera: Discrete Linear Control The Polynomial Equation Approach. Wiley, New York 1979. Ing. Jan Ježek CSc, Ústav teorie informace a automatizace ČSA V (Institute of Information Theory and Automation Czechoslovak Academy of Sciences), Pod vodárenskou věží 4, 182 08 Praha 8. Czechoslovakia. 516