The Pennsylvania State University Schreyer Honors College. Department of Mathematics. An Optimal Control Problem Arising from an Evolutionary Game

Size: px
Start display at page:

Download "The Pennsylvania State University Schreyer Honors College. Department of Mathematics. An Optimal Control Problem Arising from an Evolutionary Game"

Transcription

1 The Pennsylvania State University Schreyer Honors College Department of Mathematics An Optimal Control Problem Arising from an Evolutionary Game Wente Brian Cai Spring 2014 A thesis submitted in partial fulfillment of the requirements for a baccalaureate degree in Mathematics with honors in Mathematics To be reviewed and approved* by the following: Christopher Griffin Research Associate & Asst. Professor Thesis Supervisor Nate Brown Professor of Mathematics Honors Adviser *Signatures will be on file if this thesis is approved.

2 Abstract This paper is an integrative study of evolutionary game theory and optimal control. We first study the basics of evolutionary game theory and introduce the model that we would like to study based on the game of Rock-Paper-Scissors. We then move on to an introduction of optimal control and identify the requirements that need to be fulfilled in order for a solution to become optimal. Finally, we explore different methods of modeling the Rock-Paper-Scissors game as an optimal control problem and try to find an optimal control that minimizes the cost of the game. Various linearization schemes are attempted and the results are discussed. i

3 Contents List of Figures iii 1 Introduction 1 2 Preliminaries and Literature Review Evolutionary Game Theory Optimal Control Theory Linear Quadratic Control Example Sufficienct Conditions for Optimal Control Problem Statement: Relating Optimal Control to Game Theory 10 3 Optimal Control of Generalized Rock-Paper-Scissors Constructing a General Form of the Problem Linearizing the Problem A Linear Problem Future Work and Conclusions Future Work Conclusion Bibliography 22 ii

4 List of Figures 1 a) An illustration of an unstable Nash Equilibrium fixed point when a = 5; b) An illustration of a non-linear center Nash equilibrium fixed point when a = 0; c) An illustration of a state Nash Equilibrium fixed point when a = Computed optimal control of rock-paper-scissors evolutionary game: x0) = 0.5, y0) = 0.2, z0) = 0.3, T = 40, σ = 1/ Computed optimal control of rock-paper-scissors evolutionary game varitation: x0) = 0.5, y0) = 0.2, z0) = 0.3, T = 40, σ = 1/ Computed optimal control of partially linearized dynamics x0) =.3 1 3, y0) =.3 1 3, z0) = Optimal control of fully linearized dynamics: x0) =.3 1 3, y0) =.3 1 3, z0) =.4 1 3, σ = iii

5 1 Introduction Evolutionary game theory studies the decision making of games over time, such that dominant strategies survive and get passed on to the next generation while failing strategies are phased out. We start with a large population of players playing a two player game. Each player is assigned a strategy by genetics), and the resulting payoff from play will determine which strategy was more successful and thus which strategy reproduces. We can then track the growth or decline of strategies in our population and predict the trajectory of each strategy. Our goal is to study the dynamics of the game while we dynamically alter the payoffs of the game. In the case of Rock-Paper- Scissors, it is understood that the optimal strategy is to pick rock, paper, or scissor in equal amounts 1, 1, 1 ). Consequently, this forms a fixed point of the evolutionary game resulting from rock-paper-scissor play. Under certain payoff assumptions in a population with many scissors and many rocks and only some paper, it is clear that the population of scissors will decrease from lack of reproduction) while the number of paper players increases. For certain classes of payoff, this process will lead to the fixed point 1, 1, 1 ), while in other cases we may observe cyclic behavior In the case of convergent behavior, how quickly that fixed point is reached is dependent on the game. The larger the difference between the winning and losing payoff i.e. the bigger the gap between winning and losing), the quicker the population approach the fixed point strategy. In this thesis, we consider the rock-paper-scissors payoff matrix: a 1 A = a a R 1 + a 1 0 Our objective is to dynamically alter the value of a rather than leaving a static value) in order to drive the population toward the fixed point. However, we assume that altering the value of a may be hard and therefore we attempt to minimize the total cost associated with driving the population to the prescribed fixed point. We study this problem theoretically and numerically showing two solutions for the resulting non-linear optimal control problem with interesting properties. We then attempt to linearize the control problem to obtain a closed form solution. Unfortunately, in so doing we lose some of the characteristics of the control problem and obtain an entirely new control problem with new solution structure. We then discuss in future work how to recover the original problem while maintaining a linear dynamical system, but introducing non-linearity and non-convexity into the objective function. 1

6 2 Preliminaries and Literature Review 2.1 Evolutionary Game Theory Game theory is the study of decision making when an agent s payoff is affected by the decisions of many) other agents and is based on the payoff one gets for using certain strategies [2, 3, 5, 9, 12, 13]. For us a game consists of two players using a set, K of n pure strategies denoted by i = 1, 2,..., n. When Player I uses strategy m K and Player II uses strategy n K, there is a payoff a mn. With values a ij, i, j = 1, 2,..., n, we can create a n n payoff matrix A for player I when we assume the game is symmetric, Player II s payoff matrix is A T ) [12]. Game theory is an extensive subject; the interested reader should consult [2, 3, 5, 9, 12, 13] for details. All standard game theoretic definitions given in this thesis can be found in these references. A player s mixed strategy is a column vector x = x 1, x 2,..., x n ) T where each x i is the probability of using strategy i K. Let S n be the set of all mixed strategies in the game. So if Player I and II uses mixed strategies x, y S n, the expected payoff for Player I is x T Ay and the expected payoff for Player II is x T A T y. The strategy x is said to be the best response to y if x T Ay z T Ay z S n 1) Definition 2.1. A strategy x is a Nash) equilibrium for the symmetric two player game with payoff matrix A if: x ) T Ax y T Ax y S n 2) In evolutionary game theory, we observe how a population of players changes their strategies over time. Let s consider a large population playing the same symmetric game we defined earlier. Let p i t) 0 be the number of individuals at time t using strategy i and let pt) = i K p it) > 0 be the total population. The population state is the vector xt) = x 1 t), x 2 t),...x n t)) T such that each x i t) is the proportion of the population using strategy i at time t. The expected payoff for strategy i K would then be e T i Ax and the population average payoff would be x T Ax. Here e i is the i th standard basis vector in R n, when we have n strategies. Our goal is to derive a differential equation that determines the growth or decline of strategy proportions in the population. In our scenario, the growth and decline of strategies depends on the fitness of the strategy. In other words, the more successful a strategy is, the quicker the population will adopt that strategy over time while poorer strategies die off. The following derivation is due to [15]. We can define the change in the number of individuals using strategy i as ṗ i = [β + e T i Ax δ]p i 3) 2

7 where β 0 is the initial fitness of individuals in the population and δ 0 is the death rate for all individuals. By definition, pt)x i t) = p i t). Differentiating Equation 3 by t, we get: pẋ i = ṗ i ṗx i = [β + e T i Ax δ]p i [β + x T Ax δ]px i 4) If we divide both sides by p, we obtain the replicator dynamics: ẋ i = [e T i Ax x T Ax]x i i = 1,..., n 5) The replicator dynamics tell us that strategies that give a greater than average payoff grow in the population and strategies that give less than the average payoff decline in the population. The rest points of the replicator dynamic are the zeros of the right hand side of the replicator equation; i.e., all points x S n such that e T i Ax = x T Ax). A rest point, x, is stable if every neighborhood B of x contains a neighborhood B such that if we start at x 0, then the solution flow ϕt, x 0 ) B for all t 0. In other words, x is stable if whenever a point starts in a neighborhood of x, it will stay contained in that neighborhood over time. It is asymptotically stable if there exists a B such that lim t ϕt, x 0 ) = x for all x 0 B. The Folk Theorem of Evolutionary Game Theory explains the relationship between resting points and Nash equilibrium [6, 7]. Theorem 2.2 Folk Theorem of Evolutionary Games). Consider the replicator dynamics given in Equation 5 1. If x is a Nash equilibrium, then it is a rest point 2. If x is a strict Nash equilibrium, then it is asymptotically stable 3. If the rest point x is the limit of an interior orbit, then x is a Nash equlilibrium 4. If the rest point x is stable, thein it is a Nash equlibrium For the remainder of thesis, we will be exploring a two-player, symmetric game: Rock, Paper, Scissors. In this game, each pure strategy dominates i.e., beats) exactly one other. Every child in the world knows: Rock beats scissor, scissors beats paper and paper beats rock. So Strategy 1 dominates 2, Strategy 2 dominates 3, and Strategy 3 dominates 1. If two players pick the same strategy, the payoff is 0. Our payoff matrix is the following and is adapted from [7]: a 1 A = a a R 6) 1 + a 1 0 3

8 Zeeman [16] proved the following regarding Rock, Paper, Scissors: Theorem 2.3 Zeeman s Theorem [16]). The following conditions are equivalent for the Rock-Scissor-Paper game: 1. x is asymptotically stable 2. x is globally stable 3. det A > 0 4. x )T Ax > 0 What this tells us is that if we choose a > 0, det A > 0 which results in all orbits in the interior of Sn converging to x. If a < 0, det A < 0 and all orbits in the interior of Sn converge to the boundary Sn. Finally, if a is zero, then det A = 0, and all orbits in the interior of Sn are closed orbits around x. This is illustrated in Figure z z 0.5 z y x y a) x 1.0 b) y x c) Figure 1: a) An illustration of an unstable Nash Equilibrium fixed point when a = 5; b) An illustration of a non-linear center Nash equilibrium fixed point when a = 0; c) An illustration of a state Nash Equilibrium fixed point when a = Optimal Control Theory An optimal control problem [1, 4, 8] deals with finding time varying optimal controls or inputs) needed to maximize or minimize some cost as a functional of time varying state and the control. For the purposes of this thesis, we will investigate optimal control problems that depend only implicitly on time. Let x1 t),...xn t) be the list of state variables and u1 t),..., um t) be the control 4

9 inputs. The system dynamics is the set of n first order differential equations that dictates the path of the control variables: ẋ 1 = g 1 x 1 t),..., x n t), u 1 t),..., u m t), t) 7) ẋ n = g n x 1 t),..., x n t), u 1 t),..., u m t), t) 8) The homogeneous state equation can then be defined as: such that. ẋt) = gxt), ut)) 9) x 1 t) xt) =. 10) x n t) is the vector of our state variables and u 1 t) ut) =. 11) u m t) is the vector of controls. We will focus on a specific form for the optimal control problem. Our goal is to find an optimal control u that can minimize or maximize) our cost functional: J = ΨxT )) + T subject to our state equation constraints 0 fxt), ut))dt 12) ẋ = gxt), ut)) 13) where T is an maximum time we will consider. If T =, then we require infinite time-horizon controls, which we will not consider for simplicity. The interested reader may consult [4, 8, 14] for details on alternate forms of optimal control problems. For the remainder of this thesis, we will consider a minimization problem of Bolza type: T min ΨxT )) + fxt), ut))dt 0 14) s.t. ẋ = gxt), ut)) 5

10 We note that when ΨxT )) 0, then the problem is a Lagrange type problem, which we will also consider. Optimal control problems can be difficult to solve. Fortunately, necessary conditions that an optimal control u must satisfy can be derived. These conditions are generally derived using the Hamiltonian of the control problem: The Hamiltonian is defined as Hx t), u t), λ) = fxt), ut)) + λ T gxt), ut)) 15) where f is the cost function in Expression 12, g is the right hand side of the state equations, and λ is a vector of co-state multipliers 1. The following theorem is proven in most standard references on optimal control; see e.g., [1, 4, 8]. Theorem 2.4 Necessary Conditions for Optimal Control). Consider the Bolza problem given in Expression 14. If u is an optimal control, then Hx t), u t), λ t)) Hx t), ut), λ t)) 16) for all t [0, T ] and for all admissable inputs u U, and must satisfy the following conditions: 1. Pontryagin s Minimim Principle: ut) = H u definite, = 0 and 2 H u 2 is positive 2. Co-State Dynamics: λt) = H = x λt t) gx,u) + fx,u), x x 3. State Dynamics: ẋt) = H λ = gx, u), 4. Initial Condition: x0) = x 0, and 5. Transversality Condition: λt ) = Ψ xt )). x where partial derivatives with respect to vectors, e.g., H, denote a gradient u restricted to only those variables with respect to which we differentiate 2. 1 These co-states can be thought of like Lagrange multipliers [11] in ordinary optimization theory. 2 Note that gx,u) x is a Jacobian matrix. 6

11 2.3 Linear Quadratic Control Example Certain classes of optimal control problems can be solved in closed form. In particular, linear quadratic control LQC) problems have such special structure. We give an example of such a problem and its solution. min T 0 x 2 + u 2 dt s.t. ẋ = ax + bu x0) = x 0 The Hamiltonian for this control problem is: By the minimum principle we require 2 H u 2 for u, we have: 17) H = x 2 + u 2 + λax + bu) 18) > 0 and that H u H u = 2u + bλ = 0 = u = b 2 λ Applying the co-state dynamics to solve for λ yields: λ = H x The state dynamics to solve for ẋ yields: = 0. Solving = 2x + aλ) 19) H λ = gx, u) = ax + bu = ax b 2 λ 20) Note, ax + bu is the original right-hand-side of the state dynamics in the problem constraints. Finally, we have the initial condition x0) = x 0 and the transversality condition λt ) = 0 yielding a two-point boundary value problem: ẋ = ax b 2 λ 21) λ = 2x aλ 22) x0) = x 0 23) λt ) = 0 24) Re-writing in matrix form yields: [ẋ ] [ ] [ ] a b = 2 x λ 2 a λ 25) 7

12 To solve the system of differential equations, we plugged the equations into Mathematica DSolve) which gave us the following solutions: ) be t a 2 +b 1 + e 2t a 2 +b C 1 x = 4 a 2 + b e t a 2 +b a + a 2 + b + ae 2t a 2 +b + ) a 2 + be 2t a 2 +b + 2 a 2 + b e t a 2 +b a + ) a 2 + b ae 2t a 2 +b e 2 a 2 +b C 1 λ = 2 a 2 + b ) e t a 2 +b 1 + e 2 a 2 +b C 2 + a2 + b C 2 26) 27) Next, we solve our two unknown constants C 1 and C 2 by setting x0) = r r being some constant) and λt ) = 0. Again, using Mathematica, we get the following: e 2T ) a 2 +b r C 1 = a + a 2 + b ae 2T a 2 +b + a 2 + be 2T a 2 +b 28) C 2 = r 29) Finally, we can define our xt) and λt) functions: ) be t a 2 +b 1 + e 2t a 2 +b C 1 xt) = 4 a 2 + b e t a 2 +b a + a 2 + b + ae 2t a 2 +b + ) a 2 + be 2t a 2 +b + 2 a 2 + b e t a 2 +b a + ) a 2 + b ae 2t a 2 +b e 2 a 2 +b C 1 λt) = 2 a 2 + b ) e t a 2 +b 1 + e 2 a 2 +b C 2 + a2 + b C 2 30) 31) 8

13 2 1 + e 2T ) a 2 +b r C 1 = a + a 2 + b ae 2T a 2 +b + a 2 + be 2T a 2 +b 32) C 2 = r 33) 2.4 Sufficienct Conditions for Optimal Control We note that any solution u to Problem 14 must satisfy the necessary conditions set forth in Theorem 2.4. However it is possible that a function u satisfies these necessary conditions, but is not an optimal control. Mangaserian [10] and Arrow [4] have shown proved conditions under which the necessary conditions in Theorem 2.4 are sufficient for u to be an optimal control. Theorem 2.5 Mangaserian s Theorem). Suppose the admissible pair x, u ) satisfies all of the relevant continuous-time optimal control problem necessary conditions for OCP, the Hamiltonian H is jointly convex in x and u for all admissible solutions, t 0 and t f are fixed, x 0 is fixed, and there are no terminal time conditions Ψ[xT ), T ] = 0. Then any solution of the continuous-time optimal control necessary conditions is a global minimum. Theorem 2.6 Arrow s Theorem). Let x, u ) be an admissible pair for OCP when the Hamiltonian H is jointly convex in x and u for all admissible solutions, t 0 and t f are fixed, x 0 is fixed, and there are no terminal time conditions Ψ[xT ), T ] = 0. If there exists a continuous and piecewise continuously differentiable function λ = λ 1,..., λ n ) T such that the following conditions are satisfied: λ i = H x i, almost everywhere i = 1,..., n 34) Hx, u, λt), t) Hx, u, λ, t) for all u U and all t [0, T ] 35) Ĥx, λ, t) = minhx, u, λ, t) exists and is convex in x for all t [0, T ] u U 36) then x, u ) solves the OCP for the given. If Ĥx, λ, t) is strictly convex in x for all t, then x is unique but u is not necessarily unique). We will use Mangaserian s theorem in the sequel when we discuss linearization of the optimal control problem we consider. 9

14 2.5 Problem Statement: Relating Optimal Control to Game Theory So far, we have covered the basics of evolutionary game theory and observed how the values of parameters can affect the dynamics of the population see, e.g., Figure 1). We also studied optimal control and identified the conditions of an optimal control problem. Our next goal is to study the effects of parameters on the population when these parameters are considered to be control functions in an optimal control problem. Specifically, we will study an optimal control problem whose differential state evolution is governed by the replicator dynamics. However, such a problem can be difficult because the replicator dynamics are non-linear. Therefore, we focus on rock-paperscissors as an example and study methods to overcome the non-linearity. 10

15 3 Optimal Control of Generalized Rock-Paper- Scissors Recall the Rock-Paper-Scissors dynamics with payoff matrix given in Expression 6. We consider the case when the parameter a is allowed to vary in time as a control function. For rock-paper scissors, our goal is to find a function governing at) to drive the flow towards the Nash equilibrium 1, 1, ), over a finite time horizon T, subject to a cost incurred from a large at). We formulate our problem as: [ T min xt) 1 ) 2 + yt) zt) 0 3 3) 1 ) ] 2 + at) 2 dt 3 s.t. ẋ = x e T i Ax xax ) ẏ = y e T i Ay yay ) 37) ż = z e T i Az zaz ) Expanding the dynamics, we have: [ T min xt) 1 ) 2 + yt) zt) 0 3 3) 1 ) ] 2 + σat) 2 dt 3 s.t. ẋ = ax 2 y ax 2 z axyz + axz xy + xz ẏ = axy 2 axyz + axy ay 2 z + xy yz ż = axyz axz 2 ayz 2 + ayz xz + yz 38) We note that the objective functional is convex in the unknown functions, while the dynamics are highly) non-linear. We also note that the value of at) is not constrained but will be controlled in absolute value by the the objective functional itself; i.e., at) will not get too big or too small. With this system, our Hamiltonian will be: H = xt) 3) yt) 3) zt) 3) σa 2 + λ 1 t)ẋ + λ 2 t)ẏ + λ 3 t)ż 39) The optimal control was solved using Mathematica full equation listed in the Appendix). When we set T to be 40, we get the following plot: Figure 2a) plots our xt), yt), and zt) while b) plots the value of a from time 0 to

16 a) b) Figure 2: Computed optimal control of rock-paper-scissors evolutionary game: x0) = 0.5, y0) = 0.2, z0) = 0.3, T = 40, σ = 1/100 We also explored a modified cost functional such that xt) 3) ) yt) zt) 1 2 3) is outside of the integral. Thus our problem becomes: min xt) yt) 3) 3) zt) 3) 1 2 T + σat) 2 dt 0 s.t. ẋ = ax 2 y ax 2 z axyz + axz xy + xz ẏ = axy 2 axyz + axy ay 2 z + xy yz ż = axyz axz 2 ayz 2 + ayz xz + yz with the Hamiltonian: 40) H = σa 2 + λ 1 t)ẋ + λ 2 t)ẏ + λ 3 ż 41) When we solve an instance of the the optimal control with a computer algebra system, we get the following results: 3.1 Constructing a General Form of the Problem We also explored the method of linearizing the dynamics about the fixed point and then using the results from the LQC to find an optimal solution for a. We are still analyzing the control problem from the Rock-Paper-Scissor 12

17 a) b) Figure 3: Computed optimal control of rock-paper-scissors evolutionary game varitation: x0) = 0.5, y0) = 0.2, z0) = 0.3, T = 40, σ = 1/100 game with the payoff matrix: 0 1 a A = a ) 1 a 0 The resulting replicator dynamics are: ẋ i = x i e T i Ax x T Ax ) 43) This creates a complex convolution of the control problem: min ΦxT )) + T fx, a)dt 0 s.t. ẋ i = x i e T i Ax x T Ax ) 44) Suppose we express A such that: 0 1 a A = a 0 1 = a = B + ac 45) 1 a Denote: F i x) = x i e T i Bx x T Bx ) 46) G i x) = x i e T i Cx x T Cx ) 47) 13

18 Then we can write Problem 44 as: min ΦxT )) + T s.t. ẋ i = Fx) + agx) 0 fx, a)dt 48) Note, when a = 0, we recover fair rock-paper-scissors dynamics as the uncontrolled) state dynamics. Let and Fx) = [F 1 x),..., F n x)] T 49) G = [G 1 x),..., G n x)] T 50) n = 3 for Rock, Paper, Scissors) be vector valued functions. The Hamiltonian for this problem is: Suppose that: such that: Hx, λ, a) fx, a) + λ T Fx) + aλ T Gx) 51) fx, a) = 1 2 x x 2 + σ 2 a2 52) x = is the equilibrium point of rock paper scissors. Then: ) H a = σa + λt Gx) = 0 = a = 1 σ λt Gx) 54) If a is an optimal control, then the transversality condition would ensures that: λt ) = ΨT ) = a T ) = 1 ) T ΨT ) Gx) 55) x σ x In particular, if ΦxT )) 0, then at ) = 0. The resulting necessary conditions are the non-linear differential equations: ) 1 ẋ = Fx) σ λt Gx) Gx) 56) ) λ T = x x ) T λ T Fx) 1 x + σ λt Gx) λ T Gx) 57) x x0) = x 0 58) λt ) = ΨT ) x 14 59)

19 Here: Fx) x, Gx) x are the Jacobians of the vector valued functions F and G. Note, this is true of evolutionary game problem with a parameter a where the payoff matrix can be written as A = B + ac. 3.2 Linearizing the Problem Suppose we linearize the problem about the equilibrium point and suppose that ΦxT )) 0. Let: J = Fx) x 60) x=x H = Gx) 61) x x=x 62) Without loss of generality, suppose we translate the problem so that x = 0 i.e., by replacing x by x x ). Then the quasi-linearized control becomes: From this, we compute: min ẋ T x 2 + σ 2 a2 dt = Jx + ahx 63) a = 1 σ λt Hx 64) a quadratic form in the state and co-state. The necessary conditions for a to be an optimal control are then: ) 1 ẋ = Jx σ λt Hx Hx 65) ) λ T 1 = x T λ T J + σ λt Hx λ T H 66) x0) = x 0 67) λt ) = 0 68) which we can re-write as an almost linear two-point boundary value problem: [ẋ ] [ ] [ ] ) [ ] [ ] J 0 x 1 H 0 x = λ I n J T λ σ λt Hx 0 H T 69) λ x0) = x 0 70) λt ) = 0 71) 15

20 This problem, while non-linear preserves, some of the structure of the original problem, as we can see through numerical solution. Note, the optimal control, a) b) Figure 4: Computed optimal control of partially linearized dynamics x0) =.3 1 3, y0) =.3 1 3, z0) = shown in Figure 4 b), has the decaying characteristic of the control function shown in Figure 2 b) but does not have the oscillation. This is the result of substituting the Jacobians in for the non-linear replicator dynamics. 3.3 A Linear Problem The fundamental problem with this linearization is the resulting problem is still non-linear. To solve this, notice that the linearized dynamics can be written as: ẋ = Jx + ahx = Jx + H ax) 72) For the sake of argument, let u = ax be the control vector. Then a nonequivalent) similar LQC problem to be studied is: min ẋ T Our Hamiltonian for this problem is: x 2 + σ 2 u 2 dt = Jx + Hu 73) H = 1 2 x 2 + σ 2 u 2 + λ T Jx + Hu) 74) 16

21 From this, in the linearized problem we have: H x = σut + λ T H = 0 = u = HT λ σ The necessary conditions for u to be an optimal control are then: 75) ẋ = Jx 1 σ HHT λ 76) λ = x J T λ 77) x0) = x 0 78) λt ) = 0 79) Notice that this time, we have a linear two-point boundary value problem a Ricatti equation [8]) that can be written as the following: [ẋ ] [ ] [ ] J 1 = σ λ HHT x I n J T 80) λ Unfortunately, the solution cannot be written in extended form due to length, but will have the following form: [ ] [ ] ) x = C λ J T 1 exp σ HHT I n J T t 81) where C T is a constant that allows the expression to satisfy the two boundary conditions: x0) = x 0 and λt ) = 0. We illustrate an example solution in Figure 5. Here, we assume the control vector u = [u, v, w] T. Notice, the characteristics of u have no resemblance to the computed characteristics of a from e.g.) Figure 4. This suggests that while we have constructed a linear control problem, it does not provide us with useful information about our original non-linear control problem. We do note however, that Arrow s Theorem ensures that this is the true optimal control as linear quadratic control problems of this type satisfy appropriate convexity assumptions. To see why this is, note that by letter u = ax, we are implicitly allowing the control to operate independently in all three dimensions simultaneously; that is, we assume we can influence the payoff of winning the game independently among the three species rock, paper and scissors). Clearly these extra degrees of freedom allow the control system to achieve a value state value near the equilibrium point much more quickly than in the constrained dynamics of Problem 44 or 48. In our future work, we discuss a way of recovering these 17

22 a) b) Figure 5: Optimal control of fully linearized dynamics: x0) =.3 1 3, y0) =.3 1 3, z0) =.4 1 3, σ = 1 constraints within the objective function, thus moving all the non-convexity and non-linearity) to the objective, rather than the differential equations. 18

23 4 Future Work and Conclusions 4.1 Future Work As noted, when we compare the two methods of solving our linearization problem, we noticed that the dynamics of the second problem approaches the Nash equilibrium quicker than the first but the resulting control functions bear no resemblance to the control computed in the general non-linear system. We can recover the original problem by enforcing constraints of the form: u i x i = u j x j i j = u i x j = u j x i i j 82) and replacing the objective functional with min ẋ T x 2 + σ 2n = Jx + Hu u i x j = u j x i i i j ui x i ) 2 dt 83) The non-linear constraints can be priced out to produce the following again) non-linear control problem but now with linear dynamics: T 1 min 0 2 x 2 + σ ) 2 ui + ρ ij u i x j u j x i ) 2 dt 2n x i 84) ẋ = Jx + Hu i As a future research project, it would be interesting to explore the solutions of this problem and compare it to the original non-linear optimization problem. As additional future work, we note that the structure of the almost linearized problem Problem 48) is deceptively simple. We have not given up hope that it is possible to understand the structure of the solutions of Problem 48 analytically and therefore to shed some light on the behavior of the original non-linear problem near the fixed point. 4.2 Conclusion In this thesis, we studied evolutionary game theory and introduced the payoff matrix and replicator dynamics for the rock-paper-scissors game. Next, we surveyed results in optimal control theory and identified the characteristics of an optimal control. We then provided an example for a linear quadratic 19 i j

24 control problem, a form used later in the thesis. With this knowledge, we formulated an optimal control problem based on the rock-paper-scissor game and solved it numerically. We tried solving a variation of the game by modifying the cost functional and identified with the same problem that global optimality could not be proved. We then attempted to linearize the problem by substituting the Jacobian matrices into the non-linear replicator dynamics, but found that the problem was still not linear because of the way the control interacted. We tried a different method of linearization and were able to derive a closed form solution and found some interesting properties of it. However this method altered the characteristics of our original control problem, which led to an entirely new control problem with new solution structure. Finally, we discussed as future work how to recover the original problem after linearization by enforcing certain constraints and pricing them out in the objective functional 20

25 References [1] S. Aniţa, V. Arnăutu, and V. Capasso, An introduction to optimal control problems in life sciences and economics, Springer, [2] S. J. Brams, Game theory and politics, Dover Press, [3] Y. Freund and R. E. Schapire, Game theory, on-line prediction and boosting, Ninth Annual Conference on Computational Learning Theory, 1996, pp [4] T. L. Friesz, Dynamic Optimization and Differential Games, Springer, [5] C. Griffin, Game Theory: Penn State Math 486 Lecture Notes v 1.1), [6] J. Hofbauer and K. Sigmund, Evolutionary Games and Population Dynamics, Cambridge University Press, [7], Evolutionary Game Dynamics, Bulletin of the American Mathematical Society ), no. 4, [8] D. E. Kirk, Optimal control theory: An introduction, Dover Press, [9] R. D. Luce and H. Raiffa, Games and decisions: Introduction and critical survey, Dover Press, [10] O. L. Mangasarian, Sufficient conditions for the optimal control of nonlinear systems, SIAM J. Control ), no [11] J. E. Marsden and A. Tromba, Vector calculus, 5 ed., W. H. Freeman, [12] P. Morris, Introduction to Game Theory, Springer, [13] R. B. Myerson, Game theory: Analysis of conflict, Harvard University Press, [14] A. Takayama, Mathematical economics, 2 ed., Cambridge University Press, [15] J. W. Weibull, Evolutionary game theory, MIT Press, [16] E. C. Zeeman, Population dynamics from game theory, Global Theory of Dynamical Systems, Springer Lecture Notes in Mathematics, no. 819, Springer,

26 Academic Vita Wente Brian Cai Education The Pennsylvania State University, Spring 2014 Bachelor of Science in Mathematics Honors in Mathematics Activities Pfizer, Epidemiology Intern, May Aug 2012 JPMorgan Chase, Summer Analyst, June Aug 2013 Penn State Dance Marathon, PR Photography Captain, Sep Mar 2014

Mathematical Economics. Lecture Notes (in extracts)

Mathematical Economics. Lecture Notes (in extracts) Prof. Dr. Frank Werner Faculty of Mathematics Institute of Mathematical Optimization (IMO) http://math.uni-magdeburg.de/ werner/math-ec-new.html Mathematical Economics Lecture Notes (in extracts) Winter

More information

CHAPTER 3 THE MAXIMUM PRINCIPLE: MIXED INEQUALITY CONSTRAINTS. p. 1/73

CHAPTER 3 THE MAXIMUM PRINCIPLE: MIXED INEQUALITY CONSTRAINTS. p. 1/73 CHAPTER 3 THE MAXIMUM PRINCIPLE: MIXED INEQUALITY CONSTRAINTS p. 1/73 THE MAXIMUM PRINCIPLE: MIXED INEQUALITY CONSTRAINTS Mixed Inequality Constraints: Inequality constraints involving control and possibly

More information

FIRST-ORDER SYSTEMS OF ORDINARY DIFFERENTIAL EQUATIONS III: Autonomous Planar Systems David Levermore Department of Mathematics University of Maryland

FIRST-ORDER SYSTEMS OF ORDINARY DIFFERENTIAL EQUATIONS III: Autonomous Planar Systems David Levermore Department of Mathematics University of Maryland FIRST-ORDER SYSTEMS OF ORDINARY DIFFERENTIAL EQUATIONS III: Autonomous Planar Systems David Levermore Department of Mathematics University of Maryland 4 May 2012 Because the presentation of this material

More information

Chapter Introduction. A. Bensoussan

Chapter Introduction. A. Bensoussan Chapter 2 EXPLICIT SOLUTIONS OFLINEAR QUADRATIC DIFFERENTIAL GAMES A. Bensoussan International Center for Decision and Risk Analysis University of Texas at Dallas School of Management, P.O. Box 830688

More information

A Model of Evolutionary Dynamics with Quasiperiodic Forcing

A Model of Evolutionary Dynamics with Quasiperiodic Forcing paper presented at Society for Experimental Mechanics (SEM) IMAC XXXIII Conference on Structural Dynamics February 2-5, 205, Orlando FL A Model of Evolutionary Dynamics with Quasiperiodic Forcing Elizabeth

More information

Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games

Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games Alberto Bressan ) and Khai T. Nguyen ) *) Department of Mathematics, Penn State University **) Department of Mathematics,

More information

LMI Methods in Optimal and Robust Control

LMI Methods in Optimal and Robust Control LMI Methods in Optimal and Robust Control Matthew M. Peet Arizona State University Lecture 15: Nonlinear Systems and Lyapunov Functions Overview Our next goal is to extend LMI s and optimization to nonlinear

More information

Second Order Sufficient Conditions for Optimal Control Problems with Non-unique Minimizers

Second Order Sufficient Conditions for Optimal Control Problems with Non-unique Minimizers 2 American Control Conference Marriott Waterfront, Baltimore, MD, USA June 3-July 2, 2 WeA22. Second Order Sufficient Conditions for Optimal Control Problems with Non-unique Minimizers Christos Gavriel

More information

Math Ordinary Differential Equations

Math Ordinary Differential Equations Math 411 - Ordinary Differential Equations Review Notes - 1 1 - Basic Theory A first order ordinary differential equation has the form x = f(t, x) (11) Here x = dx/dt Given an initial data x(t 0 ) = x

More information

An Introduction to Noncooperative Games

An Introduction to Noncooperative Games An Introduction to Noncooperative Games Alberto Bressan Department of Mathematics, Penn State University Alberto Bressan (Penn State) Noncooperative Games 1 / 29 introduction to non-cooperative games:

More information

Deterministic Dynamic Programming

Deterministic Dynamic Programming Deterministic Dynamic Programming 1 Value Function Consider the following optimal control problem in Mayer s form: V (t 0, x 0 ) = inf u U J(t 1, x(t 1 )) (1) subject to ẋ(t) = f(t, x(t), u(t)), x(t 0

More information

Applications of Game Theory to Social Norm Establishment. Michael Andrews. A Thesis Presented to The University of Guelph

Applications of Game Theory to Social Norm Establishment. Michael Andrews. A Thesis Presented to The University of Guelph Applications of Game Theory to Social Norm Establishment by Michael Andrews A Thesis Presented to The University of Guelph In partial fulfilment of requirements for the degree of Master of Science in Mathematics

More information

Alberto Bressan. Department of Mathematics, Penn State University

Alberto Bressan. Department of Mathematics, Penn State University Non-cooperative Differential Games A Homotopy Approach Alberto Bressan Department of Mathematics, Penn State University 1 Differential Games d dt x(t) = G(x(t), u 1(t), u 2 (t)), x(0) = y, u i (t) U i

More information

Dynamical Systems. August 13, 2013

Dynamical Systems. August 13, 2013 Dynamical Systems Joshua Wilde, revised by Isabel Tecu, Takeshi Suzuki and María José Boccardi August 13, 2013 Dynamical Systems are systems, described by one or more equations, that evolve over time.

More information

Problem List MATH 5173 Spring, 2014

Problem List MATH 5173 Spring, 2014 Problem List MATH 5173 Spring, 2014 The notation p/n means the problem with number n on page p of Perko. 1. 5/3 [Due Wednesday, January 15] 2. 6/5 and describe the relationship of the phase portraits [Due

More information

Nonlinear Systems Theory

Nonlinear Systems Theory Nonlinear Systems Theory Matthew M. Peet Arizona State University Lecture 2: Nonlinear Systems Theory Overview Our next goal is to extend LMI s and optimization to nonlinear systems analysis. Today we

More information

minimize x subject to (x 2)(x 4) u,

minimize x subject to (x 2)(x 4) u, Math 6366/6367: Optimization and Variational Methods Sample Preliminary Exam Questions 1. Suppose that f : [, L] R is a C 2 -function with f () on (, L) and that you have explicit formulae for

More information

Duality and dynamics in Hamilton-Jacobi theory for fully convex problems of control

Duality and dynamics in Hamilton-Jacobi theory for fully convex problems of control Duality and dynamics in Hamilton-Jacobi theory for fully convex problems of control RTyrrell Rockafellar and Peter R Wolenski Abstract This paper describes some recent results in Hamilton- Jacobi theory

More information

Optimal Control. Macroeconomics II SMU. Ömer Özak (SMU) Economic Growth Macroeconomics II 1 / 112

Optimal Control. Macroeconomics II SMU. Ömer Özak (SMU) Economic Growth Macroeconomics II 1 / 112 Optimal Control Ömer Özak SMU Macroeconomics II Ömer Özak (SMU) Economic Growth Macroeconomics II 1 / 112 Review of the Theory of Optimal Control Section 1 Review of the Theory of Optimal Control Ömer

More information

EE291E/ME 290Q Lecture Notes 8. Optimal Control and Dynamic Games

EE291E/ME 290Q Lecture Notes 8. Optimal Control and Dynamic Games EE291E/ME 290Q Lecture Notes 8. Optimal Control and Dynamic Games S. S. Sastry REVISED March 29th There exist two main approaches to optimal control and dynamic games: 1. via the Calculus of Variations

More information

Robotics. Control Theory. Marc Toussaint U Stuttgart

Robotics. Control Theory. Marc Toussaint U Stuttgart Robotics Control Theory Topics in control theory, optimal control, HJB equation, infinite horizon case, Linear-Quadratic optimal control, Riccati equations (differential, algebraic, discrete-time), controllability,

More information

Nonlinear Systems and Control Lecture # 12 Converse Lyapunov Functions & Time Varying Systems. p. 1/1

Nonlinear Systems and Control Lecture # 12 Converse Lyapunov Functions & Time Varying Systems. p. 1/1 Nonlinear Systems and Control Lecture # 12 Converse Lyapunov Functions & Time Varying Systems p. 1/1 p. 2/1 Converse Lyapunov Theorem Exponential Stability Let x = 0 be an exponentially stable equilibrium

More information

Linearization problem. The simplest example

Linearization problem. The simplest example Linear Systems Lecture 3 1 problem Consider a non-linear time-invariant system of the form ( ẋ(t f x(t u(t y(t g ( x(t u(t (1 such that x R n u R m y R p and Slide 1 A: f(xu f(xu g(xu and g(xu exist and

More information

Game Theory. Greg Plaxton Theory in Programming Practice, Spring 2004 Department of Computer Science University of Texas at Austin

Game Theory. Greg Plaxton Theory in Programming Practice, Spring 2004 Department of Computer Science University of Texas at Austin Game Theory Greg Plaxton Theory in Programming Practice, Spring 2004 Department of Computer Science University of Texas at Austin Bimatrix Games We are given two real m n matrices A = (a ij ), B = (b ij

More information

Problem 1 Cost of an Infinite Horizon LQR

Problem 1 Cost of an Infinite Horizon LQR THE UNIVERSITY OF TEXAS AT SAN ANTONIO EE 5243 INTRODUCTION TO CYBER-PHYSICAL SYSTEMS H O M E W O R K # 5 Ahmad F. Taha October 12, 215 Homework Instructions: 1. Type your solutions in the LATEX homework

More information

Feedback control for a chemostat with two organisms

Feedback control for a chemostat with two organisms Feedback control for a chemostat with two organisms Patrick De Leenheer and Hal Smith Arizona State University Department of Mathematics and Statistics Tempe, AZ 85287 email: leenheer@math.la.asu.edu,

More information

EN Applied Optimal Control Lecture 8: Dynamic Programming October 10, 2018

EN Applied Optimal Control Lecture 8: Dynamic Programming October 10, 2018 EN530.603 Applied Optimal Control Lecture 8: Dynamic Programming October 0, 08 Lecturer: Marin Kobilarov Dynamic Programming (DP) is conerned with the computation of an optimal policy, i.e. an optimal

More information

DYNAMIC LECTURE 5: DISCRETE TIME INTERTEMPORAL OPTIMIZATION

DYNAMIC LECTURE 5: DISCRETE TIME INTERTEMPORAL OPTIMIZATION DYNAMIC LECTURE 5: DISCRETE TIME INTERTEMPORAL OPTIMIZATION UNIVERSITY OF MARYLAND: ECON 600. Alternative Methods of Discrete Time Intertemporal Optimization We will start by solving a discrete time intertemporal

More information

Available online at ScienceDirect. Procedia IUTAM 19 (2016 ) IUTAM Symposium Analytical Methods in Nonlinear Dynamics

Available online at   ScienceDirect. Procedia IUTAM 19 (2016 ) IUTAM Symposium Analytical Methods in Nonlinear Dynamics Available online at www.sciencedirect.com ScienceDirect Procedia IUTAM 19 (2016 ) 11 18 IUTAM Symposium Analytical Methods in Nonlinear Dynamics A model of evolutionary dynamics with quasiperiodic forcing

More information

Evolutionary Games and Periodic Fitness

Evolutionary Games and Periodic Fitness Evolutionary Games and Periodic Fitness Philippe Uyttendaele Frank Thuijsman Pieter Collins Ralf Peeters Gijs Schoenmakers Ronald Westra March 16, 2012 Abstract One thing that nearly all stability concepts

More information

MATHEMATICAL ECONOMICS: OPTIMIZATION. Contents

MATHEMATICAL ECONOMICS: OPTIMIZATION. Contents MATHEMATICAL ECONOMICS: OPTIMIZATION JOÃO LOPES DIAS Contents 1. Introduction 2 1.1. Preliminaries 2 1.2. Optimal points and values 2 1.3. The optimization problems 3 1.4. Existence of optimal points 4

More information

Brown s Original Fictitious Play

Brown s Original Fictitious Play manuscript No. Brown s Original Fictitious Play Ulrich Berger Vienna University of Economics, Department VW5 Augasse 2-6, A-1090 Vienna, Austria e-mail: ulrich.berger@wu-wien.ac.at March 2005 Abstract

More information

Irrational behavior in the Brown von Neumann Nash dynamics

Irrational behavior in the Brown von Neumann Nash dynamics Irrational behavior in the Brown von Neumann Nash dynamics Ulrich Berger a and Josef Hofbauer b a Vienna University of Economics and Business Administration, Department VW 5, Augasse 2-6, A-1090 Wien,

More information

3.7 Constrained Optimization and Lagrange Multipliers

3.7 Constrained Optimization and Lagrange Multipliers 3.7 Constrained Optimization and Lagrange Multipliers 71 3.7 Constrained Optimization and Lagrange Multipliers Overview: Constrained optimization problems can sometimes be solved using the methods of the

More information

Lecture 4. Chapter 4: Lyapunov Stability. Eugenio Schuster. Mechanical Engineering and Mechanics Lehigh University.

Lecture 4. Chapter 4: Lyapunov Stability. Eugenio Schuster. Mechanical Engineering and Mechanics Lehigh University. Lecture 4 Chapter 4: Lyapunov Stability Eugenio Schuster schuster@lehigh.edu Mechanical Engineering and Mechanics Lehigh University Lecture 4 p. 1/86 Autonomous Systems Consider the autonomous system ẋ

More information

Poincaré Map, Floquet Theory, and Stability of Periodic Orbits

Poincaré Map, Floquet Theory, and Stability of Periodic Orbits Poincaré Map, Floquet Theory, and Stability of Periodic Orbits CDS140A Lecturer: W.S. Koon Fall, 2006 1 Poincaré Maps Definition (Poincaré Map): Consider ẋ = f(x) with periodic solution x(t). Construct

More information

z x = f x (x, y, a, b), z y = f y (x, y, a, b). F(x, y, z, z x, z y ) = 0. This is a PDE for the unknown function of two independent variables.

z x = f x (x, y, a, b), z y = f y (x, y, a, b). F(x, y, z, z x, z y ) = 0. This is a PDE for the unknown function of two independent variables. Chapter 2 First order PDE 2.1 How and Why First order PDE appear? 2.1.1 Physical origins Conservation laws form one of the two fundamental parts of any mathematical model of Continuum Mechanics. These

More information

MATH The Chain Rule Fall 2016 A vector function of a vector variable is a function F: R n R m. In practice, if x 1, x n is the input,

MATH The Chain Rule Fall 2016 A vector function of a vector variable is a function F: R n R m. In practice, if x 1, x n is the input, MATH 20550 The Chain Rule Fall 2016 A vector function of a vector variable is a function F: R n R m. In practice, if x 1, x n is the input, F(x 1,, x n ) F 1 (x 1,, x n ),, F m (x 1,, x n ) where each

More information

A note on linear differential equations with periodic coefficients.

A note on linear differential equations with periodic coefficients. A note on linear differential equations with periodic coefficients. Maite Grau (1) and Daniel Peralta-Salas (2) (1) Departament de Matemàtica. Universitat de Lleida. Avda. Jaume II, 69. 251 Lleida, Spain.

More information

NOTES ON CALCULUS OF VARIATIONS. September 13, 2012

NOTES ON CALCULUS OF VARIATIONS. September 13, 2012 NOTES ON CALCULUS OF VARIATIONS JON JOHNSEN September 13, 212 1. The basic problem In Calculus of Variations one is given a fixed C 2 -function F (t, x, u), where F is defined for t [, t 1 ] and x, u R,

More information

MCE693/793: Analysis and Control of Nonlinear Systems

MCE693/793: Analysis and Control of Nonlinear Systems MCE693/793: Analysis and Control of Nonlinear Systems Lyapunov Stability - I Hanz Richter Mechanical Engineering Department Cleveland State University Definition of Stability - Lyapunov Sense Lyapunov

More information

Math 4B Notes. Written by Victoria Kala SH 6432u Office Hours: T 12:45 1:45pm Last updated 7/24/2016

Math 4B Notes. Written by Victoria Kala SH 6432u Office Hours: T 12:45 1:45pm Last updated 7/24/2016 Math 4B Notes Written by Victoria Kala vtkala@math.ucsb.edu SH 6432u Office Hours: T 2:45 :45pm Last updated 7/24/206 Classification of Differential Equations The order of a differential equation is the

More information

Optimality, Duality, Complementarity for Constrained Optimization

Optimality, Duality, Complementarity for Constrained Optimization Optimality, Duality, Complementarity for Constrained Optimization Stephen Wright University of Wisconsin-Madison May 2014 Wright (UW-Madison) Optimality, Duality, Complementarity May 2014 1 / 41 Linear

More information

OPTIMAL CONTROL THEORY: APPLICATIONS TO MANAGEMENT SCIENCE AND ECONOMICS

OPTIMAL CONTROL THEORY: APPLICATIONS TO MANAGEMENT SCIENCE AND ECONOMICS OPTIMAL CONTROL THEORY: APPLICATIONS TO MANAGEMENT SCIENCE AND ECONOMICS (SECOND EDITION, 2000) Suresh P. Sethi Gerald. L. Thompson Springer Chapter 1 p. 1/37 CHAPTER 1 WHAT IS OPTIMAL CONTROL THEORY?

More information

Principles of Optimal Control Spring 2008

Principles of Optimal Control Spring 2008 MIT OpenCourseWare http://ocw.mit.edu 16.323 Principles of Optimal Control Spring 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. 16.323 Lecture

More information

The Growth Model in Continuous Time (Ramsey Model)

The Growth Model in Continuous Time (Ramsey Model) The Growth Model in Continuous Time (Ramsey Model) Prof. Lutz Hendricks Econ720 September 27, 2017 1 / 32 The Growth Model in Continuous Time We add optimizing households to the Solow model. We first study

More information

1 The Observability Canonical Form

1 The Observability Canonical Form NONLINEAR OBSERVERS AND SEPARATION PRINCIPLE 1 The Observability Canonical Form In this Chapter we discuss the design of observers for nonlinear systems modelled by equations of the form ẋ = f(x, u) (1)

More information

Optimal Control. McGill COMP 765 Oct 3 rd, 2017

Optimal Control. McGill COMP 765 Oct 3 rd, 2017 Optimal Control McGill COMP 765 Oct 3 rd, 2017 Classical Control Quiz Question 1: Can a PID controller be used to balance an inverted pendulum: A) That starts upright? B) That must be swung-up (perhaps

More information

Solutions to Final Exam Sample Problems, Math 246, Spring 2011

Solutions to Final Exam Sample Problems, Math 246, Spring 2011 Solutions to Final Exam Sample Problems, Math 246, Spring 2 () Consider the differential equation dy dt = (9 y2 )y 2 (a) Identify its equilibrium (stationary) points and classify their stability (b) Sketch

More information

Theorem 1. ẋ = Ax is globally exponentially stable (GES) iff A is Hurwitz (i.e., max(re(σ(a))) < 0).

Theorem 1. ẋ = Ax is globally exponentially stable (GES) iff A is Hurwitz (i.e., max(re(σ(a))) < 0). Linear Systems Notes Lecture Proposition. A M n (R) is positive definite iff all nested minors are greater than or equal to zero. n Proof. ( ): Positive definite iff λ i >. Let det(a) = λj and H = {x D

More information

Problem 1: (3 points) Recall that the dot product of two vectors in R 3 is

Problem 1: (3 points) Recall that the dot product of two vectors in R 3 is Linear Algebra, Spring 206 Homework 3 Name: Problem : (3 points) Recall that the dot product of two vectors in R 3 is a x b y = ax + by + cz, c z and this is essentially the same as the matrix multiplication

More information

x 1. x n i + x 2 j (x 1, x 2, x 3 ) = x 1 j + x 3

x 1. x n i + x 2 j (x 1, x 2, x 3 ) = x 1 j + x 3 Version: 4/1/06. Note: These notes are mostly from my 5B course, with the addition of the part on components and projections. Look them over to make sure that we are on the same page as regards inner-products,

More information

Lyapunov Stability Theory

Lyapunov Stability Theory Lyapunov Stability Theory Peter Al Hokayem and Eduardo Gallestey March 16, 2015 1 Introduction In this lecture we consider the stability of equilibrium points of autonomous nonlinear systems, both in continuous

More information

Lotka Volterra Predator-Prey Model with a Predating Scavenger

Lotka Volterra Predator-Prey Model with a Predating Scavenger Lotka Volterra Predator-Prey Model with a Predating Scavenger Monica Pescitelli Georgia College December 13, 2013 Abstract The classic Lotka Volterra equations are used to model the population dynamics

More information

Math 2a Prac Lectures on Differential Equations

Math 2a Prac Lectures on Differential Equations Math 2a Prac Lectures on Differential Equations Prof. Dinakar Ramakrishnan 272 Sloan, 253-37 Caltech Office Hours: Fridays 4 5 PM Based on notes taken in class by Stephanie Laga, with a few added comments

More information

arxiv: v1 [cs.sy] 13 Sep 2017

arxiv: v1 [cs.sy] 13 Sep 2017 On imitation dynamics in potential population games Lorenzo Zino, Giacomo Como, and Fabio Fagnani arxiv:1709.04748v1 [cs.sy] 13 Sep 017 Abstract Imitation dynamics for population games are studied and

More information

Principles of Optimal Control Spring 2008

Principles of Optimal Control Spring 2008 MIT OpenCourseWare http://ocw.mit.edu 16.323 Principles of Optimal Control Spring 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. 16.323 Lecture

More information

1. Find the solution of the following uncontrolled linear system. 2 α 1 1

1. Find the solution of the following uncontrolled linear system. 2 α 1 1 Appendix B Revision Problems 1. Find the solution of the following uncontrolled linear system 0 1 1 ẋ = x, x(0) =. 2 3 1 Class test, August 1998 2. Given the linear system described by 2 α 1 1 ẋ = x +

More information

Nonlinear Control. Nonlinear Control Lecture # 2 Stability of Equilibrium Points

Nonlinear Control. Nonlinear Control Lecture # 2 Stability of Equilibrium Points Nonlinear Control Lecture # 2 Stability of Equilibrium Points Basic Concepts ẋ = f(x) f is locally Lipschitz over a domain D R n Suppose x D is an equilibrium point; that is, f( x) = 0 Characterize and

More information

Formula Sheet for Optimal Control

Formula Sheet for Optimal Control Formula Sheet for Optimal Control Division of Optimization and Systems Theory Royal Institute of Technology 144 Stockholm, Sweden 23 December 1, 29 1 Dynamic Programming 11 Discrete Dynamic Programming

More information

Lecture 10: Singular Perturbations and Averaging 1

Lecture 10: Singular Perturbations and Averaging 1 Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.243j (Fall 2003): DYNAMICS OF NONLINEAR SYSTEMS by A. Megretski Lecture 10: Singular Perturbations and

More information

Numerical Optimal Control Overview. Moritz Diehl

Numerical Optimal Control Overview. Moritz Diehl Numerical Optimal Control Overview Moritz Diehl Simplified Optimal Control Problem in ODE path constraints h(x, u) 0 initial value x0 states x(t) terminal constraint r(x(t )) 0 controls u(t) 0 t T minimize

More information

LINEAR-CONVEX CONTROL AND DUALITY

LINEAR-CONVEX CONTROL AND DUALITY 1 LINEAR-CONVEX CONTROL AND DUALITY R.T. Rockafellar Department of Mathematics, University of Washington Seattle, WA 98195-4350, USA Email: rtr@math.washington.edu R. Goebel 3518 NE 42 St., Seattle, WA

More information

The Big Picture. Discuss Examples of unpredictability. Odds, Stanisław Lem, The New Yorker (1974) Chaos, Scientific American (1986)

The Big Picture. Discuss Examples of unpredictability. Odds, Stanisław Lem, The New Yorker (1974) Chaos, Scientific American (1986) The Big Picture Discuss Examples of unpredictability Odds, Stanisław Lem, The New Yorker (1974) Chaos, Scientific American (1986) Lecture 2: Natural Computation & Self-Organization, Physics 256A (Winter

More information

Multi-Robotic Systems

Multi-Robotic Systems CHAPTER 9 Multi-Robotic Systems The topic of multi-robotic systems is quite popular now. It is believed that such systems can have the following benefits: Improved performance ( winning by numbers ) Distributed

More information

The first order quasi-linear PDEs

The first order quasi-linear PDEs Chapter 2 The first order quasi-linear PDEs The first order quasi-linear PDEs have the following general form: F (x, u, Du) = 0, (2.1) where x = (x 1, x 2,, x 3 ) R n, u = u(x), Du is the gradient of u.

More information

Lecture 4: Optimization. Maximizing a function of a single variable

Lecture 4: Optimization. Maximizing a function of a single variable Lecture 4: Optimization Maximizing or Minimizing a Function of a Single Variable Maximizing or Minimizing a Function of Many Variables Constrained Optimization Maximizing a function of a single variable

More information

Extremity of b-bistochastic Quadratic Stochastic Operators on 2D Simplex

Extremity of b-bistochastic Quadratic Stochastic Operators on 2D Simplex Malaysian Journal of Mathematical Sciences 11(2): 119 139 (2017) MALAYSIAN JOURNAL OF MATHEMATICAL SCIENCES Journal homepage: http://einspem.upm.edu.my/journal Extremity of b-bistochastic Quadratic Stochastic

More information

Replicator Dynamics and Correlated Equilibrium

Replicator Dynamics and Correlated Equilibrium Replicator Dynamics and Correlated Equilibrium Yannick Viossat To cite this version: Yannick Viossat. Replicator Dynamics and Correlated Equilibrium. CECO-208. 2004. HAL Id: hal-00242953

More information

Stabilization and Passivity-Based Control

Stabilization and Passivity-Based Control DISC Systems and Control Theory of Nonlinear Systems, 2010 1 Stabilization and Passivity-Based Control Lecture 8 Nonlinear Dynamical Control Systems, Chapter 10, plus handout from R. Sepulchre, Constructive

More information

On the Stability of the Best Reply Map for Noncooperative Differential Games

On the Stability of the Best Reply Map for Noncooperative Differential Games On the Stability of the Best Reply Map for Noncooperative Differential Games Alberto Bressan and Zipeng Wang Department of Mathematics, Penn State University, University Park, PA, 68, USA DPMMS, University

More information

Evolutionary Game Theory: Overview and Recent Results

Evolutionary Game Theory: Overview and Recent Results Overviews: Evolutionary Game Theory: Overview and Recent Results William H. Sandholm University of Wisconsin nontechnical survey: Evolutionary Game Theory (in Encyclopedia of Complexity and System Science,

More information

16.30/31, Fall 2010 Recitation # 13

16.30/31, Fall 2010 Recitation # 13 16.30/31, Fall 2010 Recitation # 13 Brandon Luders December 6, 2010 In this recitation, we tie the ideas of Lyapunov stability analysis (LSA) back to previous ways we have demonstrated stability - but

More information

An homotopy method for exact tracking of nonlinear nonminimum phase systems: the example of the spherical inverted pendulum

An homotopy method for exact tracking of nonlinear nonminimum phase systems: the example of the spherical inverted pendulum 9 American Control Conference Hyatt Regency Riverfront, St. Louis, MO, USA June -, 9 FrA.5 An homotopy method for exact tracking of nonlinear nonminimum phase systems: the example of the spherical inverted

More information

Nonlinear Systems and Elements of Control

Nonlinear Systems and Elements of Control Nonlinear Systems and Elements of Control Rafal Goebel April 3, 3 Brief review of basic differential equations A differential equation is an equation involving a function and one or more of the function

More information

BEEM103 UNIVERSITY OF EXETER. BUSINESS School. January 2009 Mock Exam, Part A. OPTIMIZATION TECHNIQUES FOR ECONOMISTS solutions

BEEM103 UNIVERSITY OF EXETER. BUSINESS School. January 2009 Mock Exam, Part A. OPTIMIZATION TECHNIQUES FOR ECONOMISTS solutions BEEM03 UNIVERSITY OF EXETER BUSINESS School January 009 Mock Exam, Part A OPTIMIZATION TECHNIQUES FOR ECONOMISTS solutions Duration : TWO HOURS The paper has 3 parts. Your marks on the rst part will be

More information

Pontryagin s Minimum Principle 1

Pontryagin s Minimum Principle 1 ECE 680 Fall 2013 Pontryagin s Minimum Principle 1 In this handout, we provide a derivation of the minimum principle of Pontryagin, which is a generalization of the Euler-Lagrange equations that also includes

More information

L 2 -induced Gains of Switched Systems and Classes of Switching Signals

L 2 -induced Gains of Switched Systems and Classes of Switching Signals L 2 -induced Gains of Switched Systems and Classes of Switching Signals Kenji Hirata and João P. Hespanha Abstract This paper addresses the L 2-induced gain analysis for switched linear systems. We exploit

More information

Game Theory and its Applications to Networks - Part I: Strict Competition

Game Theory and its Applications to Networks - Part I: Strict Competition Game Theory and its Applications to Networks - Part I: Strict Competition Corinne Touati Master ENS Lyon, Fall 200 What is Game Theory and what is it for? Definition (Roger Myerson, Game Theory, Analysis

More information

= m(0) + 4e 2 ( 3e 2 ) 2e 2, 1 (2k + k 2 ) dt. m(0) = u + R 1 B T P x 2 R dt. u + R 1 B T P y 2 R dt +

= m(0) + 4e 2 ( 3e 2 ) 2e 2, 1 (2k + k 2 ) dt. m(0) = u + R 1 B T P x 2 R dt. u + R 1 B T P y 2 R dt + ECE 553, Spring 8 Posted: May nd, 8 Problem Set #7 Solution Solutions: 1. The optimal controller is still the one given in the solution to the Problem 6 in Homework #5: u (x, t) = p(t)x k(t), t. The minimum

More information

Distributed Learning based on Entropy-Driven Game Dynamics

Distributed Learning based on Entropy-Driven Game Dynamics Distributed Learning based on Entropy-Driven Game Dynamics Bruno Gaujal joint work with Pierre Coucheney and Panayotis Mertikopoulos Inria Aug., 2014 Model Shared resource systems (network, processors)

More information

Lecture 4 Continuous time linear quadratic regulator

Lecture 4 Continuous time linear quadratic regulator EE363 Winter 2008-09 Lecture 4 Continuous time linear quadratic regulator continuous-time LQR problem dynamic programming solution Hamiltonian system and two point boundary value problem infinite horizon

More information

Linearization of Differential Equation Models

Linearization of Differential Equation Models Linearization of Differential Equation Models 1 Motivation We cannot solve most nonlinear models, so we often instead try to get an overall feel for the way the model behaves: we sometimes talk about looking

More information

MATH 215/255 Solutions to Additional Practice Problems April dy dt

MATH 215/255 Solutions to Additional Practice Problems April dy dt . For the nonlinear system MATH 5/55 Solutions to Additional Practice Problems April 08 dx dt = x( x y, dy dt = y(.5 y x, x 0, y 0, (a Show that if x(0 > 0 and y(0 = 0, then the solution (x(t, y(t of the

More information

Game interactions and dynamics on networked populations

Game interactions and dynamics on networked populations Game interactions and dynamics on networked populations Chiara Mocenni & Dario Madeo Department of Information Engineering and Mathematics University of Siena (Italy) ({mocenni, madeo}@dii.unisi.it) Siena,

More information

Multivariable Calculus

Multivariable Calculus 2 Multivariable Calculus 2.1 Limits and Continuity Problem 2.1.1 (Fa94) Let the function f : R n R n satisfy the following two conditions: (i) f (K ) is compact whenever K is a compact subset of R n. (ii)

More information

T 1. The value function v(x) is the expected net gain when using the optimal stopping time starting at state x:

T 1. The value function v(x) is the expected net gain when using the optimal stopping time starting at state x: 108 OPTIMAL STOPPING TIME 4.4. Cost functions. The cost function g(x) gives the price you must pay to continue from state x. If T is your stopping time then X T is your stopping state and f(x T ) is your

More information

EC /11. Math for Microeconomics September Course, Part II Problem Set 1 with Solutions. a11 a 12. x 2

EC /11. Math for Microeconomics September Course, Part II Problem Set 1 with Solutions. a11 a 12. x 2 LONDON SCHOOL OF ECONOMICS Professor Leonardo Felli Department of Economics S.478; x7525 EC400 2010/11 Math for Microeconomics September Course, Part II Problem Set 1 with Solutions 1. Show that the general

More information

Summary Notes on Maximization

Summary Notes on Maximization Division of the Humanities and Social Sciences Summary Notes on Maximization KC Border Fall 2005 1 Classical Lagrange Multiplier Theorem 1 Definition A point x is a constrained local maximizer of f subject

More information

MATH 1231 MATHEMATICS 1B Calculus Section 2: - ODEs.

MATH 1231 MATHEMATICS 1B Calculus Section 2: - ODEs. MATH 1231 MATHEMATICS 1B 2007. For use in Dr Chris Tisdell s lectures: Tues 11 + Thur 10 in KBT Calculus Section 2: - ODEs. 1. Motivation 2. What you should already know 3. Types and orders of ODEs 4.

More information

EN Nonlinear Control and Planning in Robotics Lecture 3: Stability February 4, 2015

EN Nonlinear Control and Planning in Robotics Lecture 3: Stability February 4, 2015 EN530.678 Nonlinear Control and Planning in Robotics Lecture 3: Stability February 4, 2015 Prof: Marin Kobilarov 0.1 Model prerequisites Consider ẋ = f(t, x). We will make the following basic assumptions

More information

Linear and non-linear programming

Linear and non-linear programming Linear and non-linear programming Benjamin Recht March 11, 2005 The Gameplan Constrained Optimization Convexity Duality Applications/Taxonomy 1 Constrained Optimization minimize f(x) subject to g j (x)

More information

Entrance Exam, Differential Equations April, (Solve exactly 6 out of the 8 problems) y + 2y + y cos(x 2 y) = 0, y(0) = 2, y (0) = 4.

Entrance Exam, Differential Equations April, (Solve exactly 6 out of the 8 problems) y + 2y + y cos(x 2 y) = 0, y(0) = 2, y (0) = 4. Entrance Exam, Differential Equations April, 7 (Solve exactly 6 out of the 8 problems). Consider the following initial value problem: { y + y + y cos(x y) =, y() = y. Find all the values y such that the

More information

Irrational behavior in the Brown von Neumann Nash dynamics

Irrational behavior in the Brown von Neumann Nash dynamics Games and Economic Behavior 56 (2006) 1 6 www.elsevier.com/locate/geb Irrational behavior in the Brown von Neumann Nash dynamics Ulrich Berger a,, Josef Hofbauer b a Vienna University of Economics and

More information

EG4321/EG7040. Nonlinear Control. Dr. Matt Turner

EG4321/EG7040. Nonlinear Control. Dr. Matt Turner EG4321/EG7040 Nonlinear Control Dr. Matt Turner EG4321/EG7040 [An introduction to] Nonlinear Control Dr. Matt Turner EG4321/EG7040 [An introduction to] Nonlinear [System Analysis] and Control Dr. Matt

More information

State-norm estimators for switched nonlinear systems under average dwell-time

State-norm estimators for switched nonlinear systems under average dwell-time 49th IEEE Conference on Decision and Control December 15-17, 2010 Hilton Atlanta Hotel, Atlanta, GA, USA State-norm estimators for switched nonlinear systems under average dwell-time Matthias A. Müller

More information

Theory and Applications of Constrained Optimal Control Proble

Theory and Applications of Constrained Optimal Control Proble Theory and Applications of Constrained Optimal Control Problems with Delays PART 1 : Mixed Control State Constraints Helmut Maurer 1, Laurenz Göllmann 2 1 Institut für Numerische und Angewandte Mathematik,

More information

1 Introduction. Erik J. Balder, Mathematical Institute, University of Utrecht, Netherlands

1 Introduction. Erik J. Balder, Mathematical Institute, University of Utrecht, Netherlands Comments on Chapters 20 and 21 (Calculus of Variations and Optimal Control Theory) of Introduction to Mathematical Economics (third edition, 2008) by E.T. Dowling Erik J. Balder, Mathematical Institute,

More information

Game Theory, Population Dynamics, Social Aggregation. Daniele Vilone (CSDC - Firenze) Namur

Game Theory, Population Dynamics, Social Aggregation. Daniele Vilone (CSDC - Firenze) Namur Game Theory, Population Dynamics, Social Aggregation Daniele Vilone (CSDC - Firenze) Namur - 18.12.2008 Summary Introduction ( GT ) General concepts of Game Theory Game Theory and Social Dynamics Application:

More information

OPTIMAL CONTROL. Sadegh Bolouki. Lecture slides for ECE 515. University of Illinois, Urbana-Champaign. Fall S. Bolouki (UIUC) 1 / 28

OPTIMAL CONTROL. Sadegh Bolouki. Lecture slides for ECE 515. University of Illinois, Urbana-Champaign. Fall S. Bolouki (UIUC) 1 / 28 OPTIMAL CONTROL Sadegh Bolouki Lecture slides for ECE 515 University of Illinois, Urbana-Champaign Fall 2016 S. Bolouki (UIUC) 1 / 28 (Example from Optimal Control Theory, Kirk) Objective: To get from

More information