7 Planar systems of linear ODE

Save this PDF as:
Size: px
Start display at page:

Transcription

1 7 Planar systems of linear ODE Here I restrict my attention to a very special class of autonomous ODE: linear ODE with constant coefficients This is arguably the only class of ODE for which explicit solution can always be constructed Linear systems considered as mathematical models of biological processes are of limited use; however, such models still can be used to describe the dynamics of the system during the stages when the interactions between the elements of the system can be disregarded Moreover, the analysis of the linear systems is a necessary step in analysis of a local behavior of nonlinear systems (linearization of the system in a neighborhood of an equilibrium) 71 General theory The linear system of first order ODE with constant coefficients on the plane has the form ẋ 1 = a 11 x 1 + a 12 x 2, ẋ 2 = a 21 x 2 + a 22 x 2, (1) or, in vector notations where x = (x 1, x 2 ), and A = (a ij ) 2 2 matrix with real entries: ẋ = Ax, x R 2, (2) [ ] a11 a A = 12 a 21 a 22 Additionally to (2), consider also the initial condition To present the solution to (2) (3), I first prove x(0) = x 0 (3) Proposition 1 Initial value problem (2) (3) is equivalent to the solution of the integral equation x(t) = x 0 + t 0 Ax(ξ) dξ (4) Proof Assume that x(t) solves (4) Then, by direct inspection we have that x(0) = x 0 Moreover, since the right-hand side is given by the integral this implies that x C (1), therefore, we can take the derivative to find (2) Now other way around, by assuming that x(t) solves (2) (3), integrating (2), and evaluating the constant of integration, we recover (4) Now we can use (4) to approximate the solution to (2) (3) by the method of successive iterations The first approximation is of course the initial condition x 0 : x 1 (t) = x 0 + t 0 Ax 0 dξ = x 0 + Ax 0 t Math 484/684: Mathematical modeling of biological processes by Artem Novozhilov Spring

2 Next, Or, continuing the process, x 2 (t) = x 0 + x n (t) = t 0 Ax 1 (ξ) dξ = x 0 + Ax 0 t + A2 x 0 t 2 2 ( I + At + A2 t An t n ) x 0 1! 2! n! Here I is an identity matrix Please note that what is inside the parenthesis in the last formula is actually a matrix However, the expression is so suggestive, given that you remember the Taylor series for the ordinary exponential function, exp(t) = e t = 1 + t 1! + t2 2! + + tn n! +, that it is impossible to resist the temptation to make Definition 2 The matrix exponent of the matrix A is defined as the infinite series exp(a) = e A = I + A 1! + A2 2! + + An n! + (5) This definition suggests that the solution to the integral equation (4), and, therefore, to the IVP (2) (3), in of the form x(t) = e At x 0, note that here we have to write x 0 on the right to make sure all the operations are well defined Before continuing the analysis of the linear system and actually proving that indeed the solution is given by the presented formula, we have to make sure that the given definition makes sense, ie, all the usual series (there are four of them if matrix A is 2 2) converge Proposition 3 Series (5) converges absolutely Proof Let a ij a For the product AA = A 2 we have that the elements of this product are bonded by 2a 2 Similarly, for A k it is 2 k 1 a k = 2 k a k /2 Since we have that k a k = 1 k! 2 e2a, therefore the series in (5) converges absolutely to the matrix denoted e A k=0 Now it is quite straightforward to prove that exp(at) solves (2) Proposition 4 d dt eat = Ae At Proof Since the series converges absolutely, we are allowed to differentiate the series termwise: d dt A k t k = k! k=0 k=0 d A k t k dt k! = k=1 A k t k 1 (k 1)! = AeAt 2

3 The last proposition indicates that the matrix exponent has some properties similar to the usual exponent Here is a good example to be careful when dealing with the matrix exponent Example 5 Consider two matrices, I claim that Let me prove it I have therefore Similarly, Therefore, Now We have Therefore, C 2 = which proves that e A+B e A e B In the last expression I used A = [ ] 0 1, B = 0 0 e A+B e A e B A 2 = 0, e A = I + A = e B = I + B = e A e B = C = A + B = [ ] 1 0 = I, C 3 = 0 1 [ ] [ ] [ ] [ ] [ ] [ ] 0 1 = C, C 4 = 1 0 [ e C 1 1 = 2! + 1 4! ! + 1 5! + ] ! 1 4! ! + 1 4! + = [ ] 1 0 = I 0 1 [ ] cos 1 sin 1, sin 1 cos 1 cos t = ( 1) k t2k (2k)!, k=0 sin t = ( 1) k 1 t 2k 1 (2k 1)! k=1 Proposition 6 If [A, B] = AB BA = 0, ie, if matrices A and B commute, then e (A+B)t = e At e Bt 3

4 Proof For the matrices that commute the binomial theorem holds: Since, by the Cauchy product (A + B) n = ( i=0 n k=0 ) A i i! j=0 ( ) n A n k B k = n! k B j = j! n=0 i+j=n i+j=n A i i! A i i! B j j! B j j!, one has e A e B = n=0 i+j=n A i i! B j j! (A + B) n = n! n=0 = e A+B We can do these formal manipulations since all the series in the question converge absolutely As an important corollary of the last proposition we have Corollary 7 For the matrix exponent e A(t 1+t 2 ) = e At 1 e At 2, and e At e At = I Proof For the first note that At 1 and At 2 commute For the second put t 1 = t and t 2 = t Now we can actually prove that solution to (2) (3) exists, unique, and defined for all < t < Theorem 8 Solution to the IVP problem (2) (3) is unique and given by Proof First, due to Proposition (4), x(t; x 0 ) = e At x 0, < t < (6) d dt eat x 0 = Ae At x 0, and also e A0 x 0 = Ix 0 = x 0, which proves that (6) is a solution To prove that it is unique, consider any solution x(t) of the IVP and put y(t) = e At x(t) We have ẏ(t) = Ae At x(t) + e At ẋ(t) = Ae At x(t) + e At Ax = 0 4

5 In the expression above, I used two (non-obvious) facts: First, that the usual product rule is still true for the derivative of the product of two matrices, and that e At A = Ae At, ie, that matrices A and e At commute You should fill in the details of the proofs of these statements Hence we have that y(t) = C, therefore, by setting t = 0, we find y(0) = x 0, which implies that any solution is given by x(t) = e At x 0 Since the last formulae is defined for any t, therefore, the solution is defined for < t < 72 Three main matrices and their phase portraits Consider the solutions to (2) and their phase portraits for three matrices: [ ] [ ] [ ] λ1 0 λ 1 α β A 1 =, A 0 λ 2 =, A 2 0 λ 3 =, β α where λ 1, λ 2, λ, α, β are real numbers A 1 We have, using the definition of the matrix exponent that [ ] e A1t e λ 1 t 0 = 0 e λ 2t, therefore the general solution to (2) is given (which can be actually obtained directly, by noting that the equations in the system are decoupled) [ ] [ e λ 1 t 0 e λ 1 x(t; x 0 ) = 0 e λ 2t x 0 = t x 0 ] 1 e λ2t x 0 2 If λ 1 0 and λ 2 0 then we have only one isolated equilibrium ˆx = (0, 0), the phase curves can be found as solutions to the first order ODE dx 2 dx 1 = λ 2x 2 λ 1 x 1, which is separable equation, and the directions on the orbits are easily determined by the signs of λ 1 and λ 2 (ie, if λ 1 < 0 then x 1 (t) 0 as t ) Consider a specific example with 0 < λ 1 < λ 2 In this case we have that all the orbits are parabolas, and the direction is from the origin because both lambdas are positive The only tricky part here is to determine which axis the orbits approach as t, this can be done by looking at the explicit equations for the orbits (you should do it) or by noting that when t e λ1t e λ2t and therefore x 1 component dominates in a small enough neighborhood of (0, 0) (see the figure) The obtained phase portrait is called topological node ( topological is often dropped), and since the arrows point from the origin, it is unstable (we ll come back to the discussion of the stability) 5

6 v 2 x2 v 1 x 1 Figure 1: Unstable node in system (2) with matrix A 1 Vectors v 1 and v 2 indicate the directions of x 1 -axis and x 2 -axis respectively As another example consider the case when λ 2 < 0 < λ 1 In this case (prove it) the orbits are actually hyperbolas on (x 1, x 2 ) plane, and the directions on them can be identifies by noting that on x 1 -axis the movement is from the origin, and on x 2 -axis it is to the origin Such phase portrait is called saddle All the orbits leave a neighborhood of the origin for both t ± except for five special orbits: first, this is of course the origin itself, second two orbits on x 1 -axis that actually approach the origin if t, and two orbits on x 2 -axis, which approach the origin if t The orbits on x 1 -axis form the unstable manifold of ˆx = (0, 0), and the orbits on x 2 -axis form the stable manifold of ˆx These orbits are also called the saddle s separatrices (singular, separatrix) v 2 x2 v 1 x 1 Figure 2: Saddle in system (2) with matrix A 1 Vectors v 1 and v 2 indicate the directions of x 1 -axis and x 2 -axis respectively There are several other cases, which need to be analyzed, let me list them all: 6

7 0 < λ 1 < λ 2 : unstable node (shown in figure) 0 < λ 2 < λ 1 : unstable node 0 < λ 1 = λ 2 : unstable node λ 1 < λ 2 < 0: stable node λ 2 < λ 1 < 0: stable node λ 1 = λ 2 < 0: stable node λ 1 < 0 < λ 2 : saddle λ 2 < 0 < λ 1 : saddle (shown in figure) You should sketch the phase portraits for each of these cases Also keep in mind that for now I exclude cases when one or both λ s are zero A 2 To find e A 2t I will use the fact that A 2 = [ ] λ λ [ ] and these two matrices commute The matrix exponent for the first matrix was found in the previous point, and for the second it is readily seen that [ ] [ ] t exp t = Therefore, [ ] e A2t = e λt 1 t 0 1 Assume that λ < 0 (the cases λ = 0 and λ > 0) left as exercises x(t; x 0 ) 0 as t, moreover, dx 2 0 dx 1 Now, first, we see that as t, therefore the orbits should be tangent to x 1 -axis The figure is given below, the phase portrait is sometimes called the improper stable node A 3 Here I will use as two commuting matrices to find A 3 = [ ] α α [ e A3t = e αt cos βt sin βt [ 0 ] β β 0 ] sin βt cos βt Hence the flow of the system (2) is given by [ x(t; x 0 ) = e A3t x 0 = e αt cos βt sin βt ] sin βt x cos βt 0 7

8 x2 v 1 x 1 Figure 3: Improper node in system (2) with matrix A 2 Vector v 1 indicates the direction of x 1 -axis To determine the phase portrait observe that if α < 0 then all the solutions will approach the origin, and if α > 0, they will go away from origin We also have components of e A 3t which are periodic functions of t, which finally gives us the whole picture: if α < 0 and β > 0 then the orbits are the spirals approaching origin clockwise, if α > 0 and β > 0 then the orbits are spiral unwinding from the origin clockwise, and if α = 0 then the orbits are closed curves Here is an example for α < 0 and β < 0, this phase portrait is called the stable focus (or spiral) x2 x 1 Figure 4: Stable focus in system (2) with matrix A 3 if we take α = 0 and β < 0 then the phase portrait is composed of the closed curves and called the center (recall the Volterra Lotka model) See the figure To determine the direction on the orbits, we can use the original vector field For example, in the case α = 0 β < 0 we have that for any point x 1 = 0 and x 2 > 0 the derivative of x 2 is negative, and therefore the direction is counter clockwise 8

9 x2 x 1 Figure 5: Center in system (2) with matrix A 3 73 A little bit of linear algebra So why actually I spend so much time on studying three quite simple particular matrices A 1, A 2, A 3? It is because the following theorem is true: Theorem 9 Let A be a 2 2 real matrix Then there exists a real invertible 2 2 matrix P such that P 1 AP = J, where matrix J is one of the following three matrices in Jordan normal form [ ] [ ] [ ] λ1 0 λ 1 α β (a), (b), (c) 0 λ 2 0 λ β α Before turning to the proof of this theorem, let us discuss how this theorem can be used for the analysis of a general system (2): ẋ = Ax, x R 2 (2) Consider a linear change of variables x = P y for new unknown vector y We have or therefore, P ẏ = AP y, ẏ = P 1 AP y = Jy, y(t; y 0 ) = e Jt y 0, where we already know how to calculate e Jt Returning to the original variable, we find that x(t; x 0 ) = P e Jt P 1 x 0, full solution to the original problem Moreover, we also showed that e At = P e Jt P 1, 9

10 which is often used to calculate the matrix exponent Finally, the phase portraits for x will be similar to those of y, since the linear invertible transformation amounts to scaling, rotation, and reflection, as we learn in the course of linear algebra The only question is how to find this linear change P For this, let us recall Definition 10 A nonzero vector v is called an eigenvector of matrix A if where λ is called the corresponding eigenvalue Av = λv, Generally eigenvalues and eigenvectors can be complex To find the eigenvalues, we need to find the roots of the characteristic polynomial P (λ) = det(a λi), which is of the second degree if A is a 2 2 matrix Once the eigenvalues are found, the corresponding eigenvectors can be found as solutions to the homogeneous system of linear algebraic equations (A λi)v = 0 Remember that eigenvectors are not unique and determined up to a multiplicative constant Now we are in a position to prove Theorem 9 The proof also gives the way to find the transformation P Proof of Theorem 9 Since the characteristic polynomial has degree two, it may have either two real root, two complex conjugate roots, or one real root multiplicity two Assume first that we have either two distinct real roots λ 1 R λ 2 R with the corresponding eigenvectors v 1 R 2 and v 2 R 2 or a real root λ R multiplicity two which has two linearly independent eigenvectors v 1 R 2 and v 2 R 2 Now consider the matrix P, whose columns are exactly v 1 and v 2, I will use the notation P = (v 1 v 2 ) It is known that the eigenvector corresponding to distinct eigenvalues are linearly independent, hence P is invertible (can you prove this fact?) Now just note that AP = (Av 1 Av 2 ) = (λ 1 v 1 λ 2 v 2 ) = P J, which proves the theorem for case (a) For case (b), assume that there is one real root of characteristic polynomial with the eigenvector v 1 Then there is another vector v 2, which satisfies (A λi)v 2 = v 1, which is linearly independent of v 1 (can you prove it?) Now take P = (v 1 v 2 ), and where J as in (b) AP = (λv 1 v 1 + λv 2 ) = P J, 10

11 Finally, in case (c) we have λ 1,2 = α±iβ as eigenvalues and the corresponding eigenvectors v 1 ±iv 2, where v 1, v 2 are real nonzero vectors Let me take P = (v 1 v 2 ) Since we have Now A(v 1 + iv 2 ) = (α + iβ)(v 1 + iv 2 ), Av 1 = αv 1 βv 2, Av 2 = αv 2 + βv 1 AP = (αv 1 βv 2 βv 1 + αv 2 ) = P J, where J as in (c) The only missing point is to prove that v 1 and v 2 are linearly independent, which is left as an exercise Example 11 Consider system (2) with A = [ ] We find that the eigenvalues and eigenvectors are λ 1 = 2, v 1 = ( 1, 1), λ 2 = 2, v 2 = (3, 1) Therefore, the transformation P here is P = [ ] 1 3, 1 1 and The solution to system J = P 1 AP = ẏ = Jy, [ ] where y = P 1 x is straightforward and given by [ ] e 2t 0 y(t; x 0 ) = 0 e 2t and its phase portrait has the structure of a saddle (see the figure) To see how actually the phase portrait looks in x coordinate, consider solution for x, which takes the form x = P y = (v 1 e λ 1t v 2 e λ 2t )y 0 = C 1 v 1 e λ 1t + C 2 v 2 e λ 2t, where I use C 1, C 2 for arbitrary constants Note that x changing along the straight line with the direction v 1 if C 2 = 0, and along the straight line v 2 when C 1 = 0 The directions of the flow on these lines coincide with the directions of the flow on the y-axes for the system with the matrix in the Jordan normal form (see the figure) This is how we can see the phase portrait for the linear system of two autonomous ODE of the first order Not taking into account the cases when one or both eigenvalues are zero, we therefore saw all possible phase portraits a linear system (2) can have Now it is time discuss stability y 0 11

12 v 2 y2 x2 v 1 y 1 x 1 Figure 6: Saddle point after the linear transformation (left), and the original phase portraits (right) We have x = P y 74 Stability of the linear system (2) In what follows I will assume that det A 0, ie, this means that there is only one isolated equilibrium of system ẋ = Ax, x R 2, (2) which is the origin: ˆx = (0, 0) To define stability of this equilibrium, and therefore stability of the linear system itself, I need a notion of a neighborhood and distance in the set R 2 I will use the following generalization of the absolute value to vectors x = (x 1, x 2 ) R 2 : x = ( (x 1 ) 2 + (x 2 ) 2) 1/2 Then distance between two vectors x 1 R 2 and x 2 R 2 is simply x 1 x 2 Using this convenient notation, now I will verbatim repeat my definition of stability of equilibria of the scalar autonomous ODE To wit, Definition 12 An equilibrium ˆx is called Lyapunov stable if for any ϵ > 0 there exists a δ(ϵ) > 0 such that for any initial conditions ˆx x 0 < δ, the flow of (2) satisfies for any t > t 0 If, additionally, ˆx x(t; x 0 ) < ϵ ˆx x(t; x 0 ) 0, when t, then ˆx is called asymptotically stable If for any initial condition x 0 the orbit γ(x 0 ) leaves a neighborhood of ˆx, then this point is called unstable 12

13 The analysis of linear systems is simple since we actually have the flow of the system given by x(t; x 0 ) = e At x 0 Case by case analysis from this lecture allows to formulate the following big theorem: Theorem 13 Let det A 0 Then the isolated equilibrium point ˆx = (0, 0) of planar system (2) is asymptotically stable if and only if for the eigenvalues of A it is true that Re λ 1,2 < 0 If Re λ 1,2 = 0 then the origin is Lyapunov stable, but not asymptotically stable (center) If for at least one eigenvalue it is true that Re λ i > 0 then the origin is unstable Therefore we can have asymptotically stable nodes, improper nodes, and foci, Lyapunov stable center, and unstable nodes, improper nodes, foci, and saddle (note that the latter are always unstable) We can summarize all the information in one parametric portrait of linear system (2) For this it is useful to consider the characteristic polynomial of A as P (λ) = λ 2 (a 11 + a 22 )λ + (a 11 a 22 a 12 a 21 ) = λ 2 + tr A λ + det A, where I use notation tr A = a 11 + a 22 to denote the trace of A We have λ 1,2 = tr A ± (tr A) 2 4 det A 2 therefore the condition for the asymptotic stability becomes simply tr A < 0, det A > 0 Using the trace and determinant as new parameters we can actually present possible linear systems as in the following figure deta, deta = (tra)2 4 stable nodes stable foci unstable foci unstable nodes tra saddles saddles Figure 7: The type of the linear system depending on the values of tr A and deta The centers here are situated where det A > 0 and tr A = 0 13

14 75 Bifurcations in the linear systems I came to the main point of this lecture We have four parameters in our linear systems, four entries of matrix A We can consider variations of these parameters, and the structure of the phase portrait of the linear system will be changing For example, if we cross the boundary det A = 0 in the last figure for negative tr A, then the saddle becomes the stable node Is this a bifurcation in the system? Or, when crossing the curve det A = 1 4 (tr A)2 for positive values of tr A the unstable focus turns into the unstable node Is this change enough to call it a bifurcation? Recall that for the first order equations the definition of bifurcation was based on the notion of topological equivalence, which identified equations with the same orbit structure as being topologically equivalent But now we can have much richer orbits structures because we are not confined any longer to the phase line, now we are dealing with the phase plane This is a quite complicated subject, therefore, I will mostly state results, proofs can be found elsewhere First, there is a notion of topological equivalence, which includes the one that we already discussed, as a particular case Definition 14 Two planar linear systems ẋ = Ax and ẋ = Bx are called topologically equivalent if there exists a homeomorphism h: R 2 R 2 of the plane, that is, h is continuous with continuous inverse, that maps the orbits of the first system onto the orbits of the second system preserving the direction of time It can be shown that the notion of topological equivalence is indeed an equivalence relation, ie, it divides all possible planar linear system into distinct non-intersecting classes The following theorem gives the topological classification of the linear planar system It is also convenient to have Definition 15 An equilibrium ˆx of ẋ = Ax is called hyperbolic if Re λ 1,2 0, where λ 1,2 are the eigenvalues of A Matrix A as well as system ẋ = Ax itself are also called hyperbolic in this case Theorem 16 Two linear systems with hyperbolic equilibria are topologically equivalent if and only if the number of eigenvalues with positive real part (and hence the number of eigenvalues with negative real part) is the same for both systems We usually have this large zoo of the equilibria: nodes, saddles, foci, but from topological point of view there are only three non-equivalent classes of hyperbolic equilibria: with two negative eigenvalues (a hyperbolic sink), one negative and one positive (a saddle), and with two positive eigenvalues (a hyperbolic source) Definition 17 A bifurcation is a change of the topological type of the system under parameter variation Now we state the result that asserts that it is impossible to have bifurcations in the linear system when the equilibrium is hyperbolic Proposition 18 Let A be hyperbolic Then there is a neighborhood U of A in R 4 such that for any B U system ẋ = Bx is topologically equivalent to ẋ = Ax 14

15 Proof The eigenvalues as the roots of the characteristic polynomial depend continuously on the entries of A Therefore, for any hyperbolic equilibrium a small enough perturbation of the matrix will lead to the eigenvalues that have the same sign of Re λ 1,2 Which implies, by Theorem 16, that this new system will be topologically equivalent to the original system Introducing another term, I can say that a property about 2 2 matrices is generic if the set of matrices possessing this property is dense and open in R 4 It can be proved that hyperbolicity is a generic property I will not go into these details, but rephrase the previous as follows: almost all matrices 2 2 are hyperbolic This brings an important question: Do we really need to study non-hyperbolic matrices and non-hyperbolic equilibria of linear systems? The point is that, in real applications the parameters of the matrix are the data that we collect in our experiments and observations These data always have some noise in it, we never know it exactly Therefore, only hyperbolic systems seem to be observed in the real life However, this is not the case Very often, systems under investigation may possess certain symmetries (such as conservation of energy) Another situation, which is more important for our course, that all the system we study contain parameters, whose values we do not know It is therefore unavoidable that under a continuous parameter change the system matrix will be non-hyperbolic, and at some point we will need to cross the boundary between topologically non equivalent behaviors Taking into account the Jordan normal forms for 2 2 matrices, we can consider the following three parameter dependent matrices: (a) [ ] 1 0, (b) 0 µ [ ] µ1 1, (c) µ 2 µ 1 [ ] µ 1 1 µ These three matrices becomes non-hyperbolic when µ = 0 or µ 1 = µ 2 = 0 For example in the case (a) when we perturb µ around zero the matrix with tr A < 0 and det A < 0 becomes a matrix with tr A < 0 and det A > 0, ie, a topological sink turns into a saddle (or vice verse) In the third case we have the change from tr A > 0 and det A > 0 to tr A < 0 and det A > 0, ie, a topological sink becomes a topological source In the case (b) the situation is more involved since we have two parameters, and I will not discuss it here To conclude, we say that for cases (a) and (c) it is enough to have one parameter, and a bifurcation of codimension one occurs, whereas for the case (b) we face a codimension two bifurcation 15

Section 9.3 Phase Plane Portraits (for Planar Systems)

Section 9.3 Phase Plane Portraits (for Planar Systems) Key Terms: Equilibrium point of planer system yꞌ = Ay o Equilibrium solution Exponential solutions o Half-line solutions Unstable solution Stable

Examples include: (a) the Lorenz system for climate and weather modeling (b) the Hodgkin-Huxley system for neuron modeling

1 Introduction Many natural processes can be viewed as dynamical systems, where the system is represented by a set of state variables and its evolution governed by a set of differential equations. Examples

Math Ordinary Differential Equations

Math 411 - Ordinary Differential Equations Review Notes - 1 1 - Basic Theory A first order ordinary differential equation has the form x = f(t, x) (11) Here x = dx/dt Given an initial data x(t 0 ) = x

14 Periodic phenomena in nature and limit cycles

14 Periodic phenomena in nature and limit cycles 14.1 Periodic phenomena in nature As it was discussed while I talked about the Lotka Volterra model, a great deal of natural phenomena show periodic behavior.

Calculus and Differential Equations II

MATH 250 B Second order autonomous linear systems We are mostly interested with 2 2 first order autonomous systems of the form { x = a x + b y y = c x + d y where x and y are functions of t and a, b, c,

In these chapter 2A notes write vectors in boldface to reduce the ambiguity of the notation.

1 2 Linear Systems In these chapter 2A notes write vectors in boldface to reduce the ambiguity of the notation 21 Matrix ODEs Let and is a scalar A linear function satisfies Linear superposition ) Linear

10 Back to planar nonlinear systems

10 Back to planar nonlinear sstems 10.1 Near the equilibria Recall that I started talking about the Lotka Volterra model as a motivation to stud sstems of two first order autonomous equations of the form

Def. (a, b) is a critical point of the autonomous system. 1 Proper node (stable or unstable) 2 Improper node (stable or unstable)

Types of critical points Def. (a, b) is a critical point of the autonomous system Math 216 Differential Equations Kenneth Harris kaharri@umich.edu Department of Mathematics University of Michigan November

2.10 Saddles, Nodes, Foci and Centers

2.10 Saddles, Nodes, Foci and Centers In Section 1.5, a linear system (1 where x R 2 was said to have a saddle, node, focus or center at the origin if its phase portrait was linearly equivalent to one

Now I switch to nonlinear systems. In this chapter the main object of study will be

Chapter 4 Stability 4.1 Autonomous systems Now I switch to nonlinear systems. In this chapter the main object of study will be ẋ = f(x), x(t) X R k, f : X R k, (4.1) where f is supposed to be locally Lipschitz

Department of Mathematics IIT Guwahati

Stability of Linear Systems in R 2 Department of Mathematics IIT Guwahati A system of first order differential equations is called autonomous if the system can be written in the form dx 1 dt = g 1(x 1,

Math 266: Phase Plane Portrait

Math 266: Phase Plane Portrait Long Jin Purdue, Spring 2018 Review: Phase line for an autonomous equation For a single autonomous equation y = f (y) we used a phase line to illustrate the equilibrium solutions

8.1 Bifurcations of Equilibria

1 81 Bifurcations of Equilibria Bifurcation theory studies qualitative changes in solutions as a parameter varies In general one could study the bifurcation theory of ODEs PDEs integro-differential equations

The goal of this chapter is to study linear systems of ordinary differential equations: dt,..., dx ) T

1 1 Linear Systems The goal of this chapter is to study linear systems of ordinary differential equations: ẋ = Ax, x(0) = x 0, (1) where x R n, A is an n n matrix and ẋ = dx ( dt = dx1 dt,..., dx ) T n.

Lecture Notes for Math 524

Lecture Notes for Math 524 Dr Michael Y Li October 19, 2009 These notes are based on the lecture notes of Professor James S Muldowney, the books of Hale, Copple, Coddington and Levinson, and Perko They

Math 3301 Homework Set Points ( ) ( ) I ll leave it to you to verify that the eigenvalues and eigenvectors for this matrix are, ( ) ( ) ( ) ( )

#7. ( pts) I ll leave it to you to verify that the eigenvalues and eigenvectors for this matrix are, λ 5 λ 7 t t ce The general solution is then : 5 7 c c c x( 0) c c 9 9 c+ c c t 5t 7 e + e A sketch of

Linear Planar Systems Math 246, Spring 2009, Professor David Levermore We now consider linear systems of the form

Linear Planar Systems Math 246, Spring 2009, Professor David Levermore We now consider linear systems of the form d x x 1 = A, where A = dt y y a11 a 12 a 21 a 22 Here the entries of the coefficient matrix

Complex Dynamic Systems: Qualitative vs Quantitative analysis

Complex Dynamic Systems: Qualitative vs Quantitative analysis Complex Dynamic Systems Chiara Mocenni Department of Information Engineering and Mathematics University of Siena (mocenni@diism.unisi.it) Dynamic

Linear Systems Notes for CDS-140a

Linear Systems Notes for CDS-140a October 27, 2008 1 Linear Systems Before beginning our study of linear dynamical systems, it only seems fair to ask the question why study linear systems? One might hope

Homogeneous Constant Matrix Systems, Part II

4 Homogeneous Constant Matrix Systems, Part II Let us now expand our discussions begun in the previous chapter, and consider homogeneous constant matrix systems whose matrices either have complex eigenvalues

APPPHYS217 Tuesday 25 May 2010

APPPHYS7 Tuesday 5 May Our aim today is to take a brief tour of some topics in nonlinear dynamics. Some good references include: [Perko] Lawrence Perko Differential Equations and Dynamical Systems (Springer-Verlag

Linear ODEs. Existence of solutions to linear IVPs. Resolvent matrix. Autonomous linear systems

Linear ODEs p. 1 Linear ODEs Existence of solutions to linear IVPs Resolvent matrix Autonomous linear systems Linear ODEs Definition (Linear ODE) A linear ODE is a differential equation taking the form

Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games

Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games Alberto Bressan ) and Khai T. Nguyen ) *) Department of Mathematics, Penn State University **) Department of Mathematics,

Understand the existence and uniqueness theorems and what they tell you about solutions to initial value problems.

Review Outline To review for the final, look over the following outline and look at problems from the book and on the old exam s and exam reviews to find problems about each of the following topics.. Basics

Math 331 Homework Assignment Chapter 7 Page 1 of 9

Math Homework Assignment Chapter 7 Page of 9 Instructions: Please make sure to demonstrate every step in your calculations. Return your answers including this homework sheet back to the instructor as a

21 Linear State-Space Representations

ME 132, Spring 25, UC Berkeley, A Packard 187 21 Linear State-Space Representations First, let s describe the most general type of dynamic system that we will consider/encounter in this class Systems may

Phase portraits in two dimensions

Phase portraits in two dimensions 8.3, Spring, 999 It [ is convenient to represent the solutions to an autonomous system x = f( x) (where x x = ) by means of a phase portrait. The x, y plane is called

MATH 583A REVIEW SESSION #1

MATH 583A REVIEW SESSION #1 BOJAN DURICKOVIC 1. Vector Spaces Very quick review of the basic linear algebra concepts (see any linear algebra textbook): (finite dimensional) vector space (or linear space),

4 Second-Order Systems

4 Second-Order Systems Second-order autonomous systems occupy an important place in the study of nonlinear systems because solution trajectories can be represented in the plane. This allows for easy visualization

Nonlinear Autonomous Dynamical systems of two dimensions. Part A

Nonlinear Autonomous Dynamical systems of two dimensions Part A Nonlinear Autonomous Dynamical systems of two dimensions x f ( x, y), x(0) x vector field y g( xy, ), y(0) y F ( f, g) 0 0 f, g are continuous

Section 8.2 : Homogeneous Linear Systems

Section 8.2 : Homogeneous Linear Systems Review: Eigenvalues and Eigenvectors Let A be an n n matrix with constant real components a ij. An eigenvector of A is a nonzero n 1 column vector v such that Av

Find the general solution of the system y = Ay, where

Math Homework # March, 9..3. Find the general solution of the system y = Ay, where 5 Answer: The matrix A has characteristic polynomial p(λ = λ + 7λ + = λ + 3(λ +. Hence the eigenvalues are λ = 3and λ

Differential Equations and Modeling

Differential Equations and Modeling Preliminary Lecture Notes Adolfo J. Rumbos c Draft date: March 22, 2018 March 22, 2018 2 Contents 1 Preface 5 2 Introduction to Modeling 7 2.1 Constructing Models.........................

Problem set 6 Math 207A, Fall 2011 Solutions. 1. A two-dimensional gradient system has the form

Problem set 6 Math 207A, Fall 2011 s 1 A two-dimensional gradient sstem has the form x t = W (x,, x t = W (x, where W (x, is a given function (a If W is a quadratic function W (x, = 1 2 ax2 + bx + 1 2

NOTES ON LINEAR ODES

NOTES ON LINEAR ODES JONATHAN LUK We can now use all the discussions we had on linear algebra to study linear ODEs Most of this material appears in the textbook in 21, 22, 23, 26 As always, this is a preliminary

First Midterm Exam Name: Practice Problems September 19, x = ax + sin x.

Math 54 Treibergs First Midterm Exam Name: Practice Problems September 9, 24 Consider the family of differential equations for the parameter a: (a Sketch the phase line when a x ax + sin x (b Use the graphs

= A(λ, t)x. In this chapter we will focus on the case that the matrix A does not depend on time (so that the ODE is autonomous):

Chapter 2 Linear autonomous ODEs 2 Linearity Linear ODEs form an important class of ODEs They are characterized by the fact that the vector field f : R m R p R R m is linear at constant value of the parameters

Problem set 7 Math 207A, Fall 2011 Solutions

Problem set 7 Math 207A, Fall 2011 s 1. Classify the equilibrium (x, y) = (0, 0) of the system x t = x, y t = y + x 2. Is the equilibrium hyperbolic? Find an equation for the trajectories in (x, y)- phase

MIT Final Exam Solutions, Spring 2017

MIT 8.6 Final Exam Solutions, Spring 7 Problem : For some real matrix A, the following vectors form a basis for its column space and null space: C(A) = span,, N(A) = span,,. (a) What is the size m n of

Old Math 330 Exams. David M. McClendon. Department of Mathematics Ferris State University

Old Math 330 Exams David M. McClendon Department of Mathematics Ferris State University Last updated to include exams from Fall 07 Contents Contents General information about these exams 3 Exams from Fall

NONCOMMUTATIVE POLYNOMIAL EQUATIONS. Edward S. Letzter. Introduction

NONCOMMUTATIVE POLYNOMIAL EQUATIONS Edward S Letzter Introduction My aim in these notes is twofold: First, to briefly review some linear algebra Second, to provide you with some new tools and techniques

Part II Problems and Solutions

Problem 1: [Complex and repeated eigenvalues] (a) The population of long-tailed weasels and meadow voles on Nantucket Island has been studied by biologists They measure the populations relative to a baseline,

STABILITY. Phase portraits and local stability

MAS271 Methods for differential equations Dr. R. Jain STABILITY Phase portraits and local stability We are interested in system of ordinary differential equations of the form ẋ = f(x, y), ẏ = g(x, y),

MTH 464: Computational Linear Algebra

MTH 464: Computational Linear Algebra Lecture Outlines Exam 4 Material Prof. M. Beauregard Department of Mathematics & Statistics Stephen F. Austin State University April 15, 2018 Linear Algebra (MTH 464)

MATH 423 Linear Algebra II Lecture 20: Geometry of linear transformations. Eigenvalues and eigenvectors. Characteristic polynomial.

MATH 423 Linear Algebra II Lecture 20: Geometry of linear transformations. Eigenvalues and eigenvectors. Characteristic polynomial. Geometric properties of determinants 2 2 determinants and plane geometry

0.1 Rational Canonical Forms

We have already seen that it is useful and simpler to study linear systems using matrices. But matrices are themselves cumbersome, as they are stuffed with many entries, and it turns out that it s best

Solutions to Dynamical Systems 2010 exam. Each question is worth 25 marks.

Solutions to Dynamical Systems exam Each question is worth marks [Unseen] Consider the following st order differential equation: dy dt Xy yy 4 a Find and classify all the fixed points of Hence draw the

Even-Numbered Homework Solutions

-6 Even-Numbered Homework Solutions Suppose that the matric B has λ = + 5i as an eigenvalue with eigenvector Y 0 = solution to dy = BY Using Euler s formula, we can write the complex-valued solution Y

8 Autonomous first order ODE

8 Autonomous first order ODE There are different ways to approach differential equations. Prior to this lecture we mostly dealt with analytical methods, i.e., with methods that require a formula as a final

Systems of Linear ODEs

P a g e 1 Systems of Linear ODEs Systems of ordinary differential equations can be solved in much the same way as discrete dynamical systems if the differential equations are linear. We will focus here

Math 273 (51) - Final

Name: Id #: Math 273 (5) - Final Autumn Quarter 26 Thursday, December 8, 26-6: to 8: Instructions: Prob. Points Score possible 25 2 25 3 25 TOTAL 75 Read each problem carefully. Write legibly. Show all

From Lay, 5.4. If we always treat a matrix as defining a linear transformation, what role does diagonalisation play?

Overview Last week introduced the important Diagonalisation Theorem: An n n matrix A is diagonalisable if and only if there is a basis for R n consisting of eigenvectors of A. This week we ll continue

Two Dimensional Linear Systems of ODEs

34 CHAPTER 3 Two Dimensional Linear Sstems of ODEs A first-der, autonomous, homogeneous linear sstem of two ODEs has the fm x t ax + b, t cx + d where a, b, c, d are real constants The matrix fm is 31

Math 312 Lecture Notes Linear Two-dimensional Systems of Differential Equations

Math 2 Lecture Notes Linear Two-dimensional Systems of Differential Equations Warren Weckesser Department of Mathematics Colgate University February 2005 In these notes, we consider the linear system of

Math 110, Spring 2015: Midterm Solutions

Math 11, Spring 215: Midterm Solutions These are not intended as model answers ; in many cases far more explanation is provided than would be necessary to receive full credit. The goal here is to make

Nonlinear Autonomous Systems of Differential

Chapter 4 Nonlinear Autonomous Systems of Differential Equations 4.0 The Phase Plane: Linear Systems 4.0.1 Introduction Consider a system of the form x = A(x), (4.0.1) where A is independent of t. Such

Chapter 7. Canonical Forms. 7.1 Eigenvalues and Eigenvectors

Chapter 7 Canonical Forms 7.1 Eigenvalues and Eigenvectors Definition 7.1.1. Let V be a vector space over the field F and let T be a linear operator on V. An eigenvalue of T is a scalar λ F such that there

Linear Algebra Exercises

9. 8.03 Linear Algebra Exercises 9A. Matrix Multiplication, Rank, Echelon Form 9A-. Which of the following matrices is in row-echelon form? 2 0 0 5 0 (i) (ii) (iii) (iv) 0 0 0 (v) [ 0 ] 0 0 0 0 0 0 0 9A-2.

1. General Vector Spaces

1.1. Vector space axioms. 1. General Vector Spaces Definition 1.1. Let V be a nonempty set of objects on which the operations of addition and scalar multiplication are defined. By addition we mean a rule

2.2. PLANAR LINEAR SYSTEMS 3 2.2. Planar Linear Systems We consider the linear system of two first order differential equations or equivalently, = ax + by (2.7) dy = cx + dy [ d x x = A x, where x =, and

Entrance Exam, Differential Equations April, (Solve exactly 6 out of the 8 problems) y + 2y + y cos(x 2 y) = 0, y(0) = 2, y (0) = 4.

Entrance Exam, Differential Equations April, 7 (Solve exactly 6 out of the 8 problems). Consider the following initial value problem: { y + y + y cos(x y) =, y() = y. Find all the values y such that the

MCE693/793: Analysis and Control of Nonlinear Systems

MCE693/793: Analysis and Control of Nonlinear Systems Systems of Differential Equations Phase Plane Analysis Hanz Richter Mechanical Engineering Department Cleveland State University Systems of Nonlinear

Linearization of Differential Equation Models

Linearization of Differential Equation Models 1 Motivation We cannot solve most nonlinear models, so we often instead try to get an overall feel for the way the model behaves: we sometimes talk about looking

Stability of Dynamical systems

Stability of Dynamical systems Stability Isolated equilibria Classification of Isolated Equilibria Attractor and Repeller Almost linear systems Jacobian Matrix Stability Consider an autonomous system u

Course 216: Ordinary Differential Equations

Course 16: Ordinary Differential Equations Notes by Chris Blair These notes cover the ODEs course given in 7-8 by Dr. John Stalker. Contents I Solving Linear ODEs 1 Reduction of Order Computing Matrix

We have already seen that the main problem of linear algebra is solving systems of linear equations.

Notes on Eigenvalues and Eigenvectors by Arunas Rudvalis Definition 1: Given a linear transformation T : R n R n a non-zero vector v in R n is called an eigenvector of T if Tv = λv for some real number

One Dimensional Dynamical Systems

16 CHAPTER 2 One Dimensional Dynamical Systems We begin by analyzing some dynamical systems with one-dimensional phase spaces, and in particular their bifurcations. All equations in this Chapter are scalar

1. In this problem, if the statement is always true, circle T; otherwise, circle F.

Math 1553, Extra Practice for Midterm 3 (sections 45-65) Solutions 1 In this problem, if the statement is always true, circle T; otherwise, circle F a) T F If A is a square matrix and the homogeneous equation

Modeling and Simulation with ODE for MSE

Zentrum Mathematik Technische Universität München Prof. Dr. Massimo Fornasier WS 6/7 Dr. Markus Hansen Sheet 7 Modeling and Simulation with ODE for MSE The exercises can be handed in until Wed, 4..6,.

Jordan normal form notes (version date: 11/21/07)

Jordan normal form notes (version date: /2/7) If A has an eigenbasis {u,, u n }, ie a basis made up of eigenvectors, so that Au j = λ j u j, then A is diagonal with respect to that basis To see this, let

MIDTERM REVIEW AND SAMPLE EXAM. Contents

MIDTERM REVIEW AND SAMPLE EXAM Abstract These notes outline the material for the upcoming exam Note that the review is divided into the two main topics we have covered thus far, namely, ordinary differential

MA5510 Ordinary Differential Equation, Fall, 2014

MA551 Ordinary Differential Equation, Fall, 214 9/3/214 Linear systems of ordinary differential equations: Consider the first order linear equation for some constant a. The solution is given by such that

Hello everyone, Best, Josh

Hello everyone, As promised, the chart mentioned in class about what kind of critical points you get with different types of eigenvalues are included on the following pages (The pages are an ecerpt from

B5.6 Nonlinear Systems

B5.6 Nonlinear Systems 4. Bifurcations Alain Goriely 2018 Mathematical Institute, University of Oxford Table of contents 1. Local bifurcations for vector fields 1.1 The problem 1.2 The extended centre

Math Matrix Algebra

Math 44 - Matrix Algebra Review notes - 4 (Alberto Bressan, Spring 27) Review of complex numbers In this chapter we shall need to work with complex numbers z C These can be written in the form z = a+ib,

Vectors, matrices, eigenvalues and eigenvectors

Vectors, matrices, eigenvalues and eigenvectors 1 ( ) ( ) ( ) Scaling a vector: 0.5V 2 0.5 2 1 = 0.5 = = 1 0.5 1 0.5 ( ) ( ) ( ) ( ) Adding two vectors: V + W 2 1 2 + 1 3 = + = = 1 3 1 + 3 4 ( ) ( ) a

Mathematical Modeling I

Mathematical Modeling I Dr. Zachariah Sinkala Department of Mathematical Sciences Middle Tennessee State University Murfreesboro Tennessee 37132, USA November 5, 2011 1d systems To understand more complex

27. Topological classification of complex linear foliations

27. Topological classification of complex linear foliations 545 H. Find the expression of the corresponding element [Γ ε ] H 1 (L ε, Z) through [Γ 1 ε], [Γ 2 ε], [δ ε ]. Problem 26.24. Prove that for any

Math 4B Notes. Written by Victoria Kala SH 6432u Office Hours: T 12:45 1:45pm Last updated 7/24/2016

Math 4B Notes Written by Victoria Kala vtkala@math.ucsb.edu SH 6432u Office Hours: T 2:45 :45pm Last updated 7/24/206 Classification of Differential Equations The order of a differential equation is the

MATH 2921 VECTOR CALCULUS AND DIFFERENTIAL EQUATIONS LINEAR SYSTEMS NOTES

MATH 9 VECTOR CALCULUS AND DIFFERENTIAL EQUATIONS LINEAR SYSTEMS NOTES R MARANGELL Contents. Higher Order ODEs and First Order Systems: One and the Same. Some First Examples 5 3. Matrix ODEs 6 4. Diagonalization

Lecture Notes for Math 251: ODE and PDE. Lecture 27: 7.8 Repeated Eigenvalues

Lecture Notes for Math 25: ODE and PDE. Lecture 27: 7.8 Repeated Eigenvalues Shawn D. Ryan Spring 22 Repeated Eigenvalues Last Time: We studied phase portraits and systems of differential equations with

EIGENVALUES AND EIGENVECTORS 3

EIGENVALUES AND EIGENVECTORS 3 1. Motivation 1.1. Diagonal matrices. Perhaps the simplest type of linear transformations are those whose matrix is diagonal (in some basis). Consider for example the matrices

Systems of Algebraic Equations and Systems of Differential Equations

Systems of Algebraic Equations and Systems of Differential Equations Topics: 2 by 2 systems of linear equations Matrix expression; Ax = b Solving 2 by 2 homogeneous systems Functions defined on matrices

Tangent spaces, normals and extrema

Chapter 3 Tangent spaces, normals and extrema If S is a surface in 3-space, with a point a S where S looks smooth, i.e., without any fold or cusp or self-crossing, we can intuitively define the tangent

Linear Algebra Practice Problems

Linear Algebra Practice Problems Math 24 Calculus III Summer 25, Session II. Determine whether the given set is a vector space. If not, give at least one axiom that is not satisfied. Unless otherwise stated,

Chapter 6 Nonlinear Systems and Phenomena. Friday, November 2, 12

Chapter 6 Nonlinear Systems and Phenomena 6.1 Stability and the Phase Plane We now move to nonlinear systems Begin with the first-order system for x(t) d dt x = f(x,t), x(0) = x 0 In particular, consider

BIFURCATION PHENOMENA Lecture 1: Qualitative theory of planar ODEs

BIFURCATION PHENOMENA Lecture 1: Qualitative theory of planar ODEs Yuri A. Kuznetsov August, 2010 Contents 1. Solutions and orbits. 2. Equilibria. 3. Periodic orbits and limit cycles. 4. Homoclinic orbits.

University of Sydney MATH3063 LECTURE NOTES

University of Sydney MATH3063 LECTURE NOTES Shane Leviton 6-7-206 Contents Introduction:... 6 Order of ODE:... 6 Explicit ODEs... 6 Notation:... 7 Example:... 7 Systems of ODEs... 7 Example 0: fx = 0...

Linear Algebra - Part II

Linear Algebra - Part II Projection, Eigendecomposition, SVD (Adapted from Sargur Srihari s slides) Brief Review from Part 1 Symmetric Matrix: A = A T Orthogonal Matrix: A T A = AA T = I and A 1 = A T

Homogeneous Linear Systems and Their General Solutions

37 Homogeneous Linear Systems and Their General Solutions We are now going to restrict our attention further to the standard first-order systems of differential equations that are linear, with particular

a11 a A = : a 21 a 22

Matrices The study of linear systems is facilitated by introducing matrices. Matrix theory provides a convenient language and notation to express many of the ideas concisely, and complicated formulas are

Homogeneous Constant Matrix Systems, Part II

4 Homogeneous Constant Matri Systems, Part II Let us now epand our discussions begun in the previous chapter, and consider homogeneous constant matri systems whose matrices either have comple eigenvalues

Symmetric and anti symmetric matrices

Symmetric and anti symmetric matrices In linear algebra, a symmetric matrix is a square matrix that is equal to its transpose. Formally, matrix A is symmetric if. A = A Because equal matrices have equal

18.06 Problem Set 8 - Solutions Due Wednesday, 14 November 2007 at 4 pm in

806 Problem Set 8 - Solutions Due Wednesday, 4 November 2007 at 4 pm in 2-06 08 03 Problem : 205+5+5+5 Consider the matrix A 02 07 a Check that A is a positive Markov matrix, and find its steady state

22.2. Applications of Eigenvalues and Eigenvectors. Introduction. Prerequisites. Learning Outcomes

Applications of Eigenvalues and Eigenvectors 22.2 Introduction Many applications of matrices in both engineering and science utilize eigenvalues and, sometimes, eigenvectors. Control theory, vibration

Nonlinear Control Lecture 2:Phase Plane Analysis

Nonlinear Control Lecture 2:Phase Plane Analysis Farzaneh Abdollahi Department of Electrical Engineering Amirkabir University of Technology Fall 2010 r. Farzaneh Abdollahi Nonlinear Control Lecture 2 1/53

Generalized eigenspaces

Generalized eigenspaces November 30, 2012 Contents 1 Introduction 1 2 Polynomials 2 3 Calculating the characteristic polynomial 5 4 Projections 7 5 Generalized eigenvalues 10 6 Eigenpolynomials 15 1 Introduction

Remark 1 By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

Sec 5 Eigenvectors and Eigenvalues In this chapter, vector means column vector Definition An eigenvector of an n n matrix A is a nonzero vector x such that A x λ x for some scalar λ A scalar λ is called