Game Theoretic Controllers that Confine a System to a Safe Region in the State Space

Size: px
Start display at page:

Download "Game Theoretic Controllers that Confine a System to a Safe Region in the State Space"

Transcription

1 Milano (Italy) August 8 - September, Game Theoretic Controllers that Confine a System to a Safe Region in the State Space R. B. Vinter EEE Department, Imperial College, London, SW7 AZ, UK ( r.vinter@imperial.ac.uk). Abstract: This paper addresses the problem of controlling a system, subject to disturbances, to prevent exit from a safe region of the state space. Problems of this nature are encountered in many different contexts, for example in the control of tanks to avoid overflow of operation of telecommunications systems to avoid the failing of a communications link. The proposed control methodology is based on solving a differential game, in which a feedback control is found that minimizes the disturbance energy to force the state out of the safe region. Such controls also have interpretations in terms of stochastic exit-time problems. We announce a new class ofexitproblems, including problems with high state dimension, forwhich therelated game can be simply computed. The proposed solution technique involves the on-line solution of N optimal control problems where, typically, N is small. Accordingly, the computational effort is several times larger, but still of the same order of magnitude, as that involved in implementing a stard non-linear model predictive controller. Keywords: Optimal control theory, Generalized solutions of Hamilton-Jacobi equations, Differential or dynamic games. INTRODUCTION We propose a methodology for controlling a system, subject to disturbances, to prevent the state exiting from a safe region of the state space. The system model is taken to be { ẋ(t) = f(x(t),u(t)) + σ(x(t))v(t) x() = x () u(t) Ω, the data for which comprise functions f : R n R m R n σ : R n R n m, a set Ω a point x R n. Here, x denotes the state variable. u is interpreted as a control variable, v as a disturbance variable. The safe region is an open subset A R n which has the functional inequality representation A = N j={x h j (x) < }, expressed in terms of a collection {h j (.) j =,...,N} of C functions, such that the gradients of active constraint functions at each boundary point of A are linearly independent. This paper deals with situations in which magnitude bounds on disturbances are not available the occurence of a state constraint violation, at some time, is inevitable. The object then is to postpone exit from the safe region for as long as possible. Notice that the control design problem is very different to that considered in the Robust Model Predictive control literature, which takes The numerical work reported in Section 6 was carried out by P. Falugi, supported the EPSRC grant Robust Model Predictive Control, ref. EP/PO6675. advantage of known bounds on disturbance signals to provide controllers which avoid exit from the safe region altogether steer the system to a neighborhood of the desired fixed point. See Mayne Rawlings (9). Our approach is to find closed loop control described by the mapping D U, compatible with a state feedback control u = χ(x), that solves the one-sided differential game: sup inf v dt + θ τ u U v V (P x ) subject to ẋ = f(x,u = D(v)) + σ(x)v a.e. t [,τ] x() = x. The data for this problem comprises the functions f, σ sets Ω A, as above, also a design parameter θ. V U denote open loop v-controls closed loop u-controls respectively; these concepts will be defined precisely below. τ is the first time exit time of the state trajectory from A. If θ =, the cost is v dt, that is, the minimum disturbance energy required to drive the state out of the safe region. When θ >, the disturbance energy cost is replaced by a weighted sum of energy time to exit. Thus the controller maximizes some measure of the effort that must be expended by the v-player to achieve violation of the state constraint in a short period of time. Copyright by the International Federation of Automatic Control (IFAC) 57

2 Milano (Italy) August 8 - September, We mention that this differential game is linked to optimal exit time problems in stochastic control. Indeed it is known that the game-theoretic controller also minimizes a risk sensitized modification of the expected exit time τ from the safe region, namely, E x [e θ τ/ǫ], when the dynamics are described by the stochastic stochastic differential equation dx = f(x,u)dt + ǫ σ(x)dw, in which w(.) is Brownian motion, in the limit as the noise intensity ǫ vanishes (see Dupuis McEneaney (997), Bardi Capuzzo-Dolcetta (997), Fleming Soner (997)). There are clearly parallels between solving the above differential game the design of non-linear H controllers, controllers whose purpose is to minimize the norm of the mapping from disturbance signals to output deviations (see Van der Shaft (993)). We point out however that non-linear H controllers only indirectly address the containment problem, because theydonot take account of the specific safe region under consideration. The most general approach to solving thegame is to compute asolution to the associated Isaacs partial differential equation, thereby obtain the optimal strategy (see McEneaney (6), Falcone (987)). The computational dems of this approach increase rapidly with state space dimension. Ourapproach is, instead, torestrict thefamily ofdifferential games considered, insuchamannerthatthe solution can be obtained by analysis alone, by carrying out relatively simple online calculations, or by a combination of the two. According to our approach, the safe set A partitions into regions {D j } N i=, each associated with just one of the state constraints, for each j, it is possible to find a constant control value u j that pushes away from the j th state constraint boundary. The state feedback controller is a multi-valued bang-bang controller of the form: u = u j,where x D j. For the problems considered, the regions {D j } are characterized by examining the relative costs of N (one player) optimal control problems; the jth optimal control problem is constructed from the game, by fixing the u-control at u j, imposing only the j th boundary constraint h j (x(τ)) =. These optimal control problems are relatively simple. If the underlying dynamics are linear θ = then each problem is a linear quadratic problem with an endpointconstraint. Aspecialcase was examined earlier in Clark Vinter (3). We describe a very simple structure for solution to the games problem, namely that it reduces to the solution of a collection ofoptimal control problems indifferent regions of the state space. We give conditions for the solution to have such a structure: they are monotonicity nonintersecting extremals conditions. These conditions are satisfied in a number of engineering applications, some examples of which are analysed in the concluding sections of this paper. In summary then, we formulate the problem of controlling a system to prevent escape from a safe region as a differential game, identify a class of games arising in this way such that a state feedback control strategy can be obtained either by analysis, or by computing online the solution to a small number of optimal control problems. Finally, we supply some examples, pointing to potential areas of the application of the methodology. Within the space ofthis paper, the accompanying analysis is necessarily abbreviated; full details will appear in a forthcoming paper, along with elaborated treatment of the examples.. THE DIFFERENTIAL GAME A precise formulation of the differential game considered is as follows: Maximize inf (P x ) v V v dt + θ τ over D U. Here, τ is the first exit time from A of the state trajectory x(.) satisfying the equation ẋ(t) = f(x(t),u(t) = D(v(.))(t)) + σ(x(t))v(t), with initial state x. That is, we seek a closed loop control for the u-player which maximizes the specified cost, when the v-player, knowing the u-player s strategy, chooses an open loop control v V to minimize the cost. The value of (P x ), for arbitrary x A, is W(x ) = sup inf v dt + θ τ. v V D U This problem description requires definitions of the underlying control spaces. The spaces of open loop controls, U V, for the u v players are respectively U := {measurable functions u : [, ) R m s.t. u(t) Ω a.e. } V := L ([, );R m ). The class U of closed loop controls is taken to be the space of (Elliott-Kalton sense) non-anticipative controllers, namely U := {non-anticipative mappings D : V U} in which a non-anticipative mapping D : V U is a mapping such that, for each pair of open loop controls v, v V each T we have: v [,T] = v [,T] a.e. implies D(v) [,T] = D(v ) [,T] a.e. Denote by exit A the exit boundary from A. This is a subset of the boundary of A, comprising all exit points of state trajectories, originating in A corresponding to some u U v V. 3. DECOMPOSABLE PROBLEMS We introduce a class of games for which the optimal closedloop strategyhas aspecialstructure, which greatly simplifies its computation: the safe set can be partitioned into regions, on each of which the state feedback control takes a constant value. Furthermore, the regions can be identified by solving, for arbitrary initial states, a finite 57

3 Milano (Italy) August 8 - September, number of optimal control problems, each involving just one of the functional inequality endpoint constraints a fixed control which pushes away from the boundary of this constraint. The definitions of this section concern the conditions under which this special structure arises. For j =,...,N write A j = {x h j (x) < } () (A j is a super-set of A, obtained by removing all but the j th state constraint). We have A = N i=a j. For j =,...,N, we select a u-control value u j which, in some sense, pushes away state trajectories from {x h j (x) = }, i.e. the boundary of A j ). For every j x A j consider the optimal control problem Minimize τ v dt + θ τ (Px) j over τ v L ([,τ];r m ) satisfying ẋ = f(x,u j ) + σ(x)v a.e. on [, ) x() = x, h j (x(τ)) =. It is important to observe that each (P j x) is a one player optimal control problem, because the u-player control value is frozen at u j. For each j let W j : A j R be the value function for (P j x) W j (x) = inf{(px)} j for x A. define D j = {x A W j (x) W j (x) for all j j}. (The D j s may overlap). It will be assumed that (H): For each j W j (.) is locally Lipschitz continuous on A j, continuously differentiable on {x A j W(x) > }. Also, for each x {x h j (x ) = } exit A seqence {x i } in A j such that x i x we have lim i W j (x i ) =.. Suppose that u j is the constant u-control associated with the D j region. Fix an initial state x in D j a v-player control v(.) such that the state trajectory x j (.), corresponding to v(.), u(.) u j the given initial state, exits from the domain A j of the (P j x ) problem at time τ. The monotonicity condition requires that the exit time τ for the trajectory x(.), corresponding to v(.) as above, any open loop u- control u the given initial state, will satisfy τ τ, i.e. the state trajectory x(.) will exit before the state trajectory x j (.). The property is illustrated in Fig.. This condition is satisfied, for example, in flow applications: if an inflow disturbance causes a tank to overflow when the outflow rate control is at a maximum, then this same inflow disturbance will certainly cause the tank to overflow, no later than before, if any other strategy is used to control the outflow. General conditions for monotonicity can be obtained, in the linear case for example, using methods of Farina Rinaldi ().. The non-intersecting extremals condition requires that, if the initial state is in region D j for some j, then the corresponding least escape energy trajectory x j (.) always remains in the same region D j. The property is illustrated in Fig.. u uj u(.) t x x(.) j h (x)= x (.) j τ τ t Definition 3.. We say that (P x ) is decomposable w.r.t. {(A j,u j )} if, in addition to (H), the following conditions are satisfied: For each j x A j, (P j x) has a minimizer ( v, τ), with corresponding state trajectory x, such that (a): (monotonicity) For any measurable u : [, τ] Ω the solution x(.) to ẋ(t) = f(x(t),u(t)) + σ(x(t)) v(t) x() = x satisfies h j (x(τ )) = for some τ [, τ], (b): (non-intersecting extremals) If for some j j we have W j (x) = W j (x) then (i): W j (x) > (ii): for all u Ω (W j W j )(x)(f(x,u) + σ(x) v()) <, (Under the hypotheses to be imposed, v(t) is continuous at t =.) (iii): for all u Ω v R m (W j W j )(x)(f(x,u) + σ(x)v) = implies f(x,u) + σ(x)v =. We comment on these conditions: Fig.. Monotonicity condition switching surface h (x)= h (x)= x (.) x (.) region D region D Fig.. Non-intersecting extremals condition 4. OPTIMAL STRATEGIES FOR DECOMPOSABLE PROBLEMS The theorem stated in this section supplies a complete description of the optimal closed loop control for decomposable games, where optimizing means optimizing 57

4 Milano (Italy) August 8 - September, over all non-anticipative u-player controls. Because the optimizing strategy will be described in terms of a state feedback, we must first provide a definition concerning the interpretation of state-feedbacks as non-anticipative controls. It will be helpful to consider the general, setvalued, state feedback relation: u χ(x) (3) where χ : R n Ω is a given set-valued function. (Allowing a set-valued feedback is convenient when dealing with discontinuous state feedback, when there is ambiguity over what control values to associate with the state feedback function at points on switching surfaces.) Definition 4.. A closed loop control D U, is said to be compatible with the feedback relation (3) if, for any v V, the solution x(.) to ẋ(t) = f(x(t),u(t)) + σ(x(t))v(t), x() = x, in which u = D(v), satifies u(t) χ(x(t)) for a.e. t [, ). Conditions for compatibility are given, for example, in Clark et al. (4). Theorem 4.. Consider the differential game (P x ). Assume (H): f(.,.) is continuous, σ(.) is Lipschitz continuous f(.,u) Lipschitz continuous with respect to x, with Lipschitz constant independent of u Ω. Suppose there exist control values u,...,u N Ω such that, by defining A,...,A N according to (), (P x ) is decomposable w.r.t. {(A j,u j )}. Let W j (.), j =,...,N be the value functions for the optimal control problems introduced in Section 3.. Then the value function for (P x ) is the lower envelope of the family of optimal control value functions {W j (.)} N j= : W(x) = min j W j (x) for all x A. (4) Furthermore, anyclosed-loopu-control thatis compatible with the set-valued feedback χ(x) = co {u j j arg min j W j (x)} for x A (5) ( one such control exists) maximizes the cost for Problem (P x ). Sketch of Proof of Thm. 4.: Write J(D(v),v;x ) for the pay-off of the differential game (P x ), for given initial state x, D U v V, i.e. τ J(u = D(v),v;x ) = v dt + θ τ in which τ is the first exit time from A of the state trajectory corresponding to u = D(v) v, originating from x. Denote by D any closed loop u-control compatible with (5). Let W(.) be the function defined by (4). It will suffice to show (A): For any D U inf v V J(D,v;x ) W(x ). (6) (B): Take any non-anticipative control D compatible with the set-valued state feedback u = χ(x) any v V. Take any j arg min j W j (x ) let ( x, v, τ) be a minimizer for (P j x ). Then J( D(v),v;x ) W(x ) (7) J( D( v), v;x ) = W(x ). (8) Indeed, these relations will assure that, for arbitrary x A D U, W(x ) = inf J( D(v),v;x ) inf J(D(v),v;x ). v V v V Consider (A). Select any D U take any j arg min j {W j (x )}. By definition of decomposable problem the optimal control problem (Px j ) has a minimizer ( v(.), τ) with state trajectory x(.). Since τ is the first exit time from A j, h j ( x( τ)) =. Write u = D( v). Write also x(.) for the state trajectory obtained by solving ẋ(t) = f(x(t),u(t)) + σ(x) v(t) with initial value x. From the conditions for decomposability we have h j (x(τ)) = for some τ [, τ]. But this implies that inf J(D(v),v;x ) J(u, v;x ) = v V τ τ v dt + θτ v dt + θ τ = W j (x ) = W(x ). Relation (6) in (A) is confirmed. The function W(.) is not continuously differentiable, but it is locally Lipschitz continuous as a lower envelope of locally Lipschitz continuous functions. Relation (7) in (B), for arbitrary v, is proved by using W(.) as a nonsmooth verification function, making use of invariance techniques as in, for example, Chapter of Vinter (). Proof of the relation (8) in (B), for the open loop v-player control v, makes key use of the non-intersecting extremals condition. We give below an example, with state dimension n =, where the state feedback control function can be obtained by analysis. In other situations where pencil paper solution is not possible, typically when the underlying control system has a high dimensional state space, implementation ofthegame theoretic controlrequires numerical optimization, along the following lines: Let T be the sample period. At time t k = kt measure the state x(t k ) compute the N (optimal control problem) minimum costs: { V j (x(t k )) = Min v dt + θτ ẋ = f(x,uj ) + σ(x)v, } x() = x(t k ) h j (x(τ)) =. Implement the control u(t) = u j on [kt,(k + )T] where j = arg min j {W j (x(t k ))}. We can expect that this control will give an approximation to the game theoretic strategy, the accuracy of which will depend on the sampling period T, if the non-intersecting extremals condition is satisfied. Even if this condition 573

5 Milano (Italy) August 8 - September, is not satisfied, the procedure will yield a sub-optimal controller. Simple online tests are available for testing non-intersection of extremals. 5. EXAMPLE: SECOND ORDER SYSTEM We consider the problem of controlling a surge tank, a buffer device used in process systems engineering, to prevent excessive flow rate fluctuations, as fluids flow between reactors. Let y be the height of the fluid in a surge tank of uniform cross-section. y is related to the rate of change of inflow v the rate of change of outflow u according to: d y/dt = u + v. In this equation, v u are interpreted as a rom disturbance control signal, respectively. The constraints that the tank neither empties nor overflows are captured by the inequalities: y +. (9) On the other h, a condition that the maximum rate of change of outflow must be bounded in magnitude can be expressed as u +. () We now seek a state feedback control, with values in Ω, u = χ(y,ẏ), depending on the current height rate of change of height, to avoid violation of the state constraint (9). In this illustrative example, we set all constants to one. The differential game for this setup is sup D inf τ,v v dt + θ τ over D U, τ, v V subject to ẋ = Fx + b( D(v) + v) a.e. t [,τ] u(t) [,+] a.e. x() = x h j (x(τ)) = for some j, in which F = h j (x) = [ ] b = [ ] { c T x if j = c T x if j = in which c T = [ ]. The state constraint sets A A for the two optimal control problems are A = {x h (x) < } = (,+) R A = {x h (x) < } = (, ) R, with boundaries {x h (x) = } = {+} R {x h (x) = } = { } R. The controls pushing away from these boundaries are, respectively, u = + u =. It is a straightforward exercise to show that the game is monotone, in the sense of Def. 3.. We must also show that it has non-intersecting extremals. To this end, we need to examine the two optimal control problems (Px ) (Px ): inf v dt + θ τ(x) (Px v, τ ) over τ v L ([,τ];r) satisfying ẋ = Fx + b( + v) a.e. t [,τ] x() = x, x(τ) {+} R, inf v dt + θ τ(x) (Px v, τ ) over τ v L ([,τ];r) satisfying ẋ = Fx + b(+ + v) a.e. t [,τ] x() = x, x(τ) { } R. It is helpful to consider the two cases θ > θ = separately. Case (i): θ >. Define the mappings η : (, ) (, ) R η : (, ) (, ) R to be: η (q,τ) = ( τ + 6 qτ3 θτq,τ qτ + θq), η (r,σ) = ( + σ 6 rσ3 + θσr, σ + rσ θr). It is possible to show that there exists open sets O j, j =,, such that, for each j, η j is one-to-one on O j η j (O j ) = A j. Furthermore, for any intial state x A, the value function, first exit time optimal control for problem (Px) are W (x) = 6 q τ 3 + θτ first exit time =τ v(t) = q(τ t), for a.e. t [,τ], in which (q,τ) = (η ) (x). For any initial state x A on the other h, the value function, first exit time optimal control for problem (Px) are W (x) = 6 r σ 3 + θσ first exit time =σ v(t) = r(σ t), for a.e. t [,σ], in which (r,σ) = (η ) (x). We observe that optimal controls are continuous (one of the non-intersecting extremals conditions that needs to be checked), the set A decomposes into disjoint regions A = {x A W (x) < W (x)} Σ {x A W (x) < W (x)}. The switching set Σ, a smooth submanifold, is Σ = {x A 6 q τ 3 + θτ = 6 r σ 3 + θσ}, in which (q,τ) = (η ) (x) (r,σ) = (η ) (x)}. It can be shown that at points x A for which W (x) = W (x), we have W (x) >, (W W )(x)(f(x,u) + σ(x) v ()) <, (W W )(x)(f(x,u) + σ(x) v ()) <, 574

6 Milano (Italy) August 8 - September,, furthermore, f(x,u) + σ(x)v = if (W W )(x)(f(x,u) + σ(x)v) =. Here if v (.) v (.) are optimal controls for (P x) (P x) respectively, u v arbitrary points on Ω R m. These are the non-intersecting extremals conditions. Case(ii): θ=. The reason this case has to treated separately is that, now, W W are not continuously differentiable on the sets A = {x = (x,x ) x < +} A = {x = (x,x ) x > }, respectively. But checking theconditions for decomposable problems merely required that these functions are continuously differentiable on their support sets (sets on which they are strictly positive), as indeed they are. These two supports sets are: A support = {x A x < + x } A support = {x A x > + x } (The significance of these sets is that if, for either j = or, the initial condition x lies in A\A j support, then the related state trajectory with zero v control exits from A, whatever the u-player control happens to be, therefore incurs zero cost.) Take the functions η η as before (with θ = ). We can choose O O such that η j is one-to-one on O j for j =, η(o j ) = {(x,x ) x < } A support, η(o j ) = {(x,x ) x > } A support. It can be shown that, for any intial state x A support, the value function, first exit time optimal control for problem (Px ) are W (x) = 6 q τ 3 first exit time =τ v(t) = q(τ t), for a.e. t [,τ], in which (q,τ) = (η ) (x). Also, for any initial state x A support, the value function, first exit time optimal control for problem (Px ) are W (x) = 6 r σ 3 first exit time =σ v(t) = r(σ t), for a.e. t [,σ], in which (r,σ) = (η ) (x). These are the non-intersecting extremals conditions, also in this case. When θ =, the switching surface is generated by the points x (α) = ( + 3α α + α 3 3α 4 α 5 ) ( + 3α α α 3 + 3α 4 + α 5 ) α(α ) x (α) = ( + α) / ( + 3α α α 3 + 3α 4 + α 5 ), as α ranges of over the interval α (The denominators in these formulae are positive for all non-negative α s.) Comments: The Maximum Rate of Change of Outflow (MROC) is a significant performance indicator in surge tank control. Indeed, it is the rate of change of outflow, rather than the flow itself, that has adverse effects on downstream process units (disturbed sediments, turbulence, etc) (see McDonald McAvoy (986) Kantor (989)). Ourformulation of thecontrol problem (proposed in Clark Vinter (3)) is to maximize in some sense the elapsed time before the surge tank fills or empties, while observing a constraint on the maximum rate of change of outflow. Previous treatments have limited attention to deterministic disturbances with a particular profile (step changes in inflow rate). The formulation of this paper, by contrast, allows for rom inputs. Fig. 3 provides plots of the switching sets for a number of values ofconstantθ. Fig. 4illustrates the non-intersecting properties of extremals; we observe that optimal state trajectories forthetwo optimal controlproblems remainin regions where the u-control has constant value, they depart from the switching surface at a positive angle. x θ= θ= θ= Switching curve for θ=, θ = θ= x Fig. 3. Switching curves x u =- θ=, switching curve u = x Fig. 4. Non-intersecting extremals 575

7 Milano (Italy) August 8 - September, 6. A THIRD ORDER SYSTEM CONTROL REQUIRING NUMERICAL OPTIMIZATION Let us now look at a refinement of the surge tank control problem of the previous section, in which the disturbance acts on the system, not directly, but through a first order lag. (We set the time constant to one). The games theoretic controller is, in this case, the stochastic controller which minimizes a risk sensitive cost, in the limit as the noise intensity parameter vanishes, for the colored noise stochastic system The two problems were solved numerically for initial states at grid points in the box set { (x,x,x 3 ) x +, R x +R, R x 3 +R}, for some suitably large number R. The switching surface was then obtained bycomparing the minimum costs for the two problems for initial states at all points on the grid. The computed optimal trajectories confirm validity of the nonintersecting extremals condition. The switching surface is illustrated in Figure 5. dz = Fz + b( + e) dt de = e dt + ǫ / dw(t), where F b are as above. Changing the noise model in this manner gives rise to the following game, having state dimension n = 3: Minimize τ v dt + θτ subject to [ ] [ ] [ ] ẋ = x u + v u(t) [,+], x(t) (,+) for t [,τ) x() = x, [ ] x(τ) = { } {+}. The safe region in the state space is A = {x c T x < c T x < } = [,+] R R where, now, c T = []. The enlarged state constraint sets, resulting from retaining only one of the state constraints, are A = {x c T x < } = (,+) R R A = {x c T x < } = (, ) R R with boundaries {x c T x = } = {+} R R {x c T x = } = { } R R. The pushing away controls are u = + u =. It is straightforward to check the monotonicity condition is satisfied. The two optimal control problems to be considered are Minimize τ v dt + θτ subject to [ ] [ ] [ ] ẋ = x + v x(t) (,+) for t [,τ) x() = x, [ ] x(τ) = +. Minimize τ v dt + θτ subject to [ ] [ ] [ ] ẋ = x + + v x(t) (, ) for t [,τ) x() = x, [ ] x(τ) =. x x Fig. 5. Switching surface.5 x REFERENCES Bardi, M. Capuzzo-Dolcetta, I. (997). Optimal Control Viscosity Solutions of Hamilton-Jacobi Equations. Birkhäuser, Boston. Clark, J., James, M., Vinter, R. (4). The interpretation of discontinuous feedback strategies as nonanticipative feedback strategies in differential games. IEEE Transactions on Automatic Control, 49, Clark, J. Vinter, R. (3). A differential dynamic games approach to flow control. 43rd Conference on Decision Control, Hawaii. Dupuis, P. McEneaney, W. (997). Risk-sensitive robust escape criteria. SIAM Journal of Control Optimization, 35, 49. Falcone, M. (987). A numerical approach to the infinite horizon problem of deterministic control theory. Appl. Math. Optim., 5, 3. Farina, L. Rinaldi, S. (). Positive Linear Systems: Theory Applications. Wiley, New York. Fleming, W. Soner, H. (997). Controlled Markov Processes Viscosity Solutions. Birkhäuser, Boston,. Kantor, J. (989). Non-linear sliding mode controller objective function for surge tanks. Int. J. Control, 5, Mayne, D. Rawlings, J. (9). Model Predictive Control: Theory Design. Nob Hill Publishing, Madison. McDonald, K. McAvoy, T. (986). Optimal avering level control. AIChE Journal, 3, McEneaney, W. (6). Max-Plus Methods for Nonlinear Control Estimation. Birkhäuser, Boston. Van der Shaft, A. (993). Nonlinear state space H control theory. in Perspectives in Control, Birkhauser, Boston. Vinter, R. (). Optimal Control. Birkhäuser, Boston

Differential Games Controllers that Confine a System to a Safe Region in the State Space, with Applications to Surge Tank Control

Differential Games Controllers that Confine a System to a Safe Region in the State Space, with Applications to Surge Tank Control Differential Games Controllers that Confine a System to a Safe Region in the State Space, with Applications to Surge Tank Control P. Falugi, P.A. Kountouriotis, R.B. Vinter, Fellow, IEEE Abstract Surge

More information

Optimal Control and Viscosity Solutions of Hamilton-Jacobi-Bellman Equations

Optimal Control and Viscosity Solutions of Hamilton-Jacobi-Bellman Equations Martino Bardi Italo Capuzzo-Dolcetta Optimal Control and Viscosity Solutions of Hamilton-Jacobi-Bellman Equations Birkhauser Boston Basel Berlin Contents Preface Basic notations xi xv Chapter I. Outline

More information

Risk-Sensitive Control with HARA Utility

Risk-Sensitive Control with HARA Utility IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 46, NO. 4, APRIL 2001 563 Risk-Sensitive Control with HARA Utility Andrew E. B. Lim Xun Yu Zhou, Senior Member, IEEE Abstract In this paper, a control methodology

More information

AC&ST AUTOMATIC CONTROL AND SYSTEM THEORY SYSTEMS AND MODELS. Claudio Melchiorri

AC&ST AUTOMATIC CONTROL AND SYSTEM THEORY SYSTEMS AND MODELS. Claudio Melchiorri C. Melchiorri (DEI) Automatic Control & System Theory 1 AUTOMATIC CONTROL AND SYSTEM THEORY SYSTEMS AND MODELS Claudio Melchiorri Dipartimento di Ingegneria dell Energia Elettrica e dell Informazione (DEI)

More information

Mean-Field Games with non-convex Hamiltonian

Mean-Field Games with non-convex Hamiltonian Mean-Field Games with non-convex Hamiltonian Martino Bardi Dipartimento di Matematica "Tullio Levi-Civita" Università di Padova Workshop "Optimal Control and MFGs" Pavia, September 19 21, 2018 Martino

More information

Controlled Diffusions and Hamilton-Jacobi Bellman Equations

Controlled Diffusions and Hamilton-Jacobi Bellman Equations Controlled Diffusions and Hamilton-Jacobi Bellman Equations Emo Todorov Applied Mathematics and Computer Science & Engineering University of Washington Winter 2014 Emo Todorov (UW) AMATH/CSE 579, Winter

More information

Neighboring feasible trajectories in infinite dimension

Neighboring feasible trajectories in infinite dimension Neighboring feasible trajectories in infinite dimension Marco Mazzola Université Pierre et Marie Curie (Paris 6) H. Frankowska and E. M. Marchini Control of State Constrained Dynamical Systems Padova,

More information

Computation of an Over-Approximation of the Backward Reachable Set using Subsystem Level Set Functions. Stanford University, Stanford, CA 94305

Computation of an Over-Approximation of the Backward Reachable Set using Subsystem Level Set Functions. Stanford University, Stanford, CA 94305 To appear in Dynamics of Continuous, Discrete and Impulsive Systems http:monotone.uwaterloo.ca/ journal Computation of an Over-Approximation of the Backward Reachable Set using Subsystem Level Set Functions

More information

Viscosity Solutions for Dummies (including Economists)

Viscosity Solutions for Dummies (including Economists) Viscosity Solutions for Dummies (including Economists) Online Appendix to Income and Wealth Distribution in Macroeconomics: A Continuous-Time Approach written by Benjamin Moll August 13, 2017 1 Viscosity

More information

Lecture 5: Lyapunov Functions and Storage Functions 1

Lecture 5: Lyapunov Functions and Storage Functions 1 Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.243j (Fall 2003): DYNAMICS OF NONLINEAR SYSTEMS by A. Megretski Lecture 5: Lyapunov Functions and Storage

More information

Numerical approximation for optimal control problems via MPC and HJB. Giulia Fabrini

Numerical approximation for optimal control problems via MPC and HJB. Giulia Fabrini Numerical approximation for optimal control problems via MPC and HJB Giulia Fabrini Konstanz Women In Mathematics 15 May, 2018 G. Fabrini (University of Konstanz) Numerical approximation for OCP 1 / 33

More information

Multi-dimensional Stochastic Singular Control Via Dynkin Game and Dirichlet Form

Multi-dimensional Stochastic Singular Control Via Dynkin Game and Dirichlet Form Multi-dimensional Stochastic Singular Control Via Dynkin Game and Dirichlet Form Yipeng Yang * Under the supervision of Dr. Michael Taksar Department of Mathematics University of Missouri-Columbia Oct

More information

An Overview of Risk-Sensitive Stochastic Optimal Control

An Overview of Risk-Sensitive Stochastic Optimal Control An Overview of Risk-Sensitive Stochastic Optimal Control Matt James Australian National University Workshop: Stochastic control, communications and numerics perspective UniSA, Adelaide, Sept. 2004. 1 Outline

More information

A Generalization of Barbalat s Lemma with Applications to Robust Model Predictive Control

A Generalization of Barbalat s Lemma with Applications to Robust Model Predictive Control A Generalization of Barbalat s Lemma with Applications to Robust Model Predictive Control Fernando A. C. C. Fontes 1 and Lalo Magni 2 1 Officina Mathematica, Departamento de Matemática para a Ciência e

More information

Nonlinear L 2 -gain analysis via a cascade

Nonlinear L 2 -gain analysis via a cascade 9th IEEE Conference on Decision and Control December -7, Hilton Atlanta Hotel, Atlanta, GA, USA Nonlinear L -gain analysis via a cascade Peter M Dower, Huan Zhang and Christopher M Kellett Abstract A nonlinear

More information

Second Order Sufficient Conditions for Optimal Control Problems with Non-unique Minimizers

Second Order Sufficient Conditions for Optimal Control Problems with Non-unique Minimizers 2 American Control Conference Marriott Waterfront, Baltimore, MD, USA June 3-July 2, 2 WeA22. Second Order Sufficient Conditions for Optimal Control Problems with Non-unique Minimizers Christos Gavriel

More information

Lecture Note 7: Switching Stabilization via Control-Lyapunov Function

Lecture Note 7: Switching Stabilization via Control-Lyapunov Function ECE7850: Hybrid Systems:Theory and Applications Lecture Note 7: Switching Stabilization via Control-Lyapunov Function Wei Zhang Assistant Professor Department of Electrical and Computer Engineering Ohio

More information

A Remark on IVP and TVP Non-Smooth Viscosity Solutions to Hamilton-Jacobi Equations

A Remark on IVP and TVP Non-Smooth Viscosity Solutions to Hamilton-Jacobi Equations 2005 American Control Conference June 8-10, 2005. Portland, OR, USA WeB10.3 A Remark on IVP and TVP Non-Smooth Viscosity Solutions to Hamilton-Jacobi Equations Arik Melikyan, Andrei Akhmetzhanov and Naira

More information

Viscosity Solutions of the Bellman Equation for Perturbed Optimal Control Problems with Exit Times 0

Viscosity Solutions of the Bellman Equation for Perturbed Optimal Control Problems with Exit Times 0 Viscosity Solutions of the Bellman Equation for Perturbed Optimal Control Problems with Exit Times Michael Malisoff Department of Mathematics Louisiana State University Baton Rouge, LA 783-4918 USA malisoff@mathlsuedu

More information

Duality and dynamics in Hamilton-Jacobi theory for fully convex problems of control

Duality and dynamics in Hamilton-Jacobi theory for fully convex problems of control Duality and dynamics in Hamilton-Jacobi theory for fully convex problems of control RTyrrell Rockafellar and Peter R Wolenski Abstract This paper describes some recent results in Hamilton- Jacobi theory

More information

On the Bellman equation for control problems with exit times and unbounded cost functionals 1

On the Bellman equation for control problems with exit times and unbounded cost functionals 1 On the Bellman equation for control problems with exit times and unbounded cost functionals 1 Michael Malisoff Department of Mathematics, Hill Center-Busch Campus Rutgers University, 11 Frelinghuysen Road

More information

Linear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard June 15, 2013

Linear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard June 15, 2013 Linear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard June 15, 2013 Abstract As in optimal control theory, linear quadratic (LQ) differential games (DG) can be solved, even in high dimension,

More information

Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games

Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games Alberto Bressan ) and Khai T. Nguyen ) *) Department of Mathematics, Penn State University **) Department of Mathematics,

More information

Maximum Process Problems in Optimal Control Theory

Maximum Process Problems in Optimal Control Theory J. Appl. Math. Stochastic Anal. Vol. 25, No., 25, (77-88) Research Report No. 423, 2, Dept. Theoret. Statist. Aarhus (2 pp) Maximum Process Problems in Optimal Control Theory GORAN PESKIR 3 Given a standard

More information

Solution of Stochastic Optimal Control Problems and Financial Applications

Solution of Stochastic Optimal Control Problems and Financial Applications Journal of Mathematical Extension Vol. 11, No. 4, (2017), 27-44 ISSN: 1735-8299 URL: http://www.ijmex.com Solution of Stochastic Optimal Control Problems and Financial Applications 2 Mat B. Kafash 1 Faculty

More information

On continuous time contract theory

On continuous time contract theory Ecole Polytechnique, France Journée de rentrée du CMAP, 3 octobre, 218 Outline 1 2 Semimartingale measures on the canonical space Random horizon 2nd order backward SDEs (Static) Principal-Agent Problem

More information

An Integral-type Constraint Qualification for Optimal Control Problems with State Constraints

An Integral-type Constraint Qualification for Optimal Control Problems with State Constraints An Integral-type Constraint Qualification for Optimal Control Problems with State Constraints S. Lopes, F. A. C. C. Fontes and M. d. R. de Pinho Officina Mathematica report, April 4, 27 Abstract Standard

More information

Hausdorff Continuous Viscosity Solutions of Hamilton-Jacobi Equations

Hausdorff Continuous Viscosity Solutions of Hamilton-Jacobi Equations Hausdorff Continuous Viscosity Solutions of Hamilton-Jacobi Equations R Anguelov 1,2, S Markov 2,, F Minani 3 1 Department of Mathematics and Applied Mathematics, University of Pretoria 2 Institute of

More information

On the stability of receding horizon control with a general terminal cost

On the stability of receding horizon control with a general terminal cost On the stability of receding horizon control with a general terminal cost Ali Jadbabaie and John Hauser Abstract We study the stability and region of attraction properties of a family of receding horizon

More information

Min-Max Certainty Equivalence Principle and Differential Games

Min-Max Certainty Equivalence Principle and Differential Games Min-Max Certainty Equivalence Principle and Differential Games Pierre Bernhard and Alain Rapaport INRIA Sophia-Antipolis August 1994 Abstract This paper presents a version of the Certainty Equivalence

More information

Recent Advances in State Constrained Optimal Control

Recent Advances in State Constrained Optimal Control Recent Advances in State Constrained Optimal Control Richard B. Vinter Imperial College London Control for Energy and Sustainability Seminar Cambridge University, 6th November 2009 Outline of the talk

More information

Positive Markov Jump Linear Systems (PMJLS) with applications

Positive Markov Jump Linear Systems (PMJLS) with applications Positive Markov Jump Linear Systems (PMJLS) with applications P. Bolzern, P. Colaneri DEIB, Politecnico di Milano - Italy December 12, 2015 Summary Positive Markov Jump Linear Systems Mean stability Input-output

More information

An introduction to Mathematical Theory of Control

An introduction to Mathematical Theory of Control An introduction to Mathematical Theory of Control Vasile Staicu University of Aveiro UNICA, May 2018 Vasile Staicu (University of Aveiro) An introduction to Mathematical Theory of Control UNICA, May 2018

More information

Lecture 8. Chapter 5: Input-Output Stability Chapter 6: Passivity Chapter 14: Passivity-Based Control. Eugenio Schuster.

Lecture 8. Chapter 5: Input-Output Stability Chapter 6: Passivity Chapter 14: Passivity-Based Control. Eugenio Schuster. Lecture 8 Chapter 5: Input-Output Stability Chapter 6: Passivity Chapter 14: Passivity-Based Control Eugenio Schuster schuster@lehigh.edu Mechanical Engineering and Mechanics Lehigh University Lecture

More information

Chain differentials with an application to the mathematical fear operator

Chain differentials with an application to the mathematical fear operator Chain differentials with an application to the mathematical fear operator Pierre Bernhard I3S, University of Nice Sophia Antipolis and CNRS, ESSI, B.P. 145, 06903 Sophia Antipolis cedex, France January

More information

Adaptive Nonlinear Model Predictive Control with Suboptimality and Stability Guarantees

Adaptive Nonlinear Model Predictive Control with Suboptimality and Stability Guarantees Adaptive Nonlinear Model Predictive Control with Suboptimality and Stability Guarantees Pontus Giselsson Department of Automatic Control LTH Lund University Box 118, SE-221 00 Lund, Sweden pontusg@control.lth.se

More information

Hybrid Systems - Lecture n. 3 Lyapunov stability

Hybrid Systems - Lecture n. 3 Lyapunov stability OUTLINE Focus: stability of equilibrium point Hybrid Systems - Lecture n. 3 Lyapunov stability Maria Prandini DEI - Politecnico di Milano E-mail: prandini@elet.polimi.it continuous systems decribed by

More information

Passivity-based Stabilization of Non-Compact Sets

Passivity-based Stabilization of Non-Compact Sets Passivity-based Stabilization of Non-Compact Sets Mohamed I. El-Hawwary and Manfredi Maggiore Abstract We investigate the stabilization of closed sets for passive nonlinear systems which are contained

More information

Further Results on the Bellman Equation for Optimal Control Problems with Exit Times and Nonnegative Instantaneous Costs

Further Results on the Bellman Equation for Optimal Control Problems with Exit Times and Nonnegative Instantaneous Costs Further Results on the Bellman Equation for Optimal Control Problems with Exit Times and Nonnegative Instantaneous Costs Michael Malisoff Systems Science and Mathematics Dept. Washington University in

More information

On the Stability of the Best Reply Map for Noncooperative Differential Games

On the Stability of the Best Reply Map for Noncooperative Differential Games On the Stability of the Best Reply Map for Noncooperative Differential Games Alberto Bressan and Zipeng Wang Department of Mathematics, Penn State University, University Park, PA, 68, USA DPMMS, University

More information

The principle of least action and two-point boundary value problems in orbital mechanics

The principle of least action and two-point boundary value problems in orbital mechanics The principle of least action and two-point boundary value problems in orbital mechanics Seung Hak Han and William M McEneaney Abstract We consider a two-point boundary value problem (TPBVP) in orbital

More information

1. Numerical Methods for Stochastic Control Problems in Continuous Time (with H. J. Kushner), Second Revised Edition, Springer-Verlag, New York, 2001.

1. Numerical Methods for Stochastic Control Problems in Continuous Time (with H. J. Kushner), Second Revised Edition, Springer-Verlag, New York, 2001. Paul Dupuis Professor of Applied Mathematics Division of Applied Mathematics Brown University Publications Books: 1. Numerical Methods for Stochastic Control Problems in Continuous Time (with H. J. Kushner),

More information

A CHARACTERIZATION OF STRICT LOCAL MINIMIZERS OF ORDER ONE FOR STATIC MINMAX PROBLEMS IN THE PARAMETRIC CONSTRAINT CASE

A CHARACTERIZATION OF STRICT LOCAL MINIMIZERS OF ORDER ONE FOR STATIC MINMAX PROBLEMS IN THE PARAMETRIC CONSTRAINT CASE Journal of Applied Analysis Vol. 6, No. 1 (2000), pp. 139 148 A CHARACTERIZATION OF STRICT LOCAL MINIMIZERS OF ORDER ONE FOR STATIC MINMAX PROBLEMS IN THE PARAMETRIC CONSTRAINT CASE A. W. A. TAHA Received

More information

Some Properties of the Augmented Lagrangian in Cone Constrained Optimization

Some Properties of the Augmented Lagrangian in Cone Constrained Optimization MATHEMATICS OF OPERATIONS RESEARCH Vol. 29, No. 3, August 2004, pp. 479 491 issn 0364-765X eissn 1526-5471 04 2903 0479 informs doi 10.1287/moor.1040.0103 2004 INFORMS Some Properties of the Augmented

More information

Reflected Brownian Motion

Reflected Brownian Motion Chapter 6 Reflected Brownian Motion Often we encounter Diffusions in regions with boundary. If the process can reach the boundary from the interior in finite time with positive probability we need to decide

More information

L 2 -induced Gains of Switched Systems and Classes of Switching Signals

L 2 -induced Gains of Switched Systems and Classes of Switching Signals L 2 -induced Gains of Switched Systems and Classes of Switching Signals Kenji Hirata and João P. Hespanha Abstract This paper addresses the L 2-induced gain analysis for switched linear systems. We exploit

More information

Nonlinear Control Systems

Nonlinear Control Systems Nonlinear Control Systems António Pedro Aguiar pedro@isr.ist.utl.pt 5. Input-Output Stability DEEC PhD Course http://users.isr.ist.utl.pt/%7epedro/ncs2012/ 2012 1 Input-Output Stability y = Hu H denotes

More information

Nonlinear Control. Nonlinear Control Lecture # 3 Stability of Equilibrium Points

Nonlinear Control. Nonlinear Control Lecture # 3 Stability of Equilibrium Points Nonlinear Control Lecture # 3 Stability of Equilibrium Points The Invariance Principle Definitions Let x(t) be a solution of ẋ = f(x) A point p is a positive limit point of x(t) if there is a sequence

More information

A Globally Stabilizing Receding Horizon Controller for Neutrally Stable Linear Systems with Input Constraints 1

A Globally Stabilizing Receding Horizon Controller for Neutrally Stable Linear Systems with Input Constraints 1 A Globally Stabilizing Receding Horizon Controller for Neutrally Stable Linear Systems with Input Constraints 1 Ali Jadbabaie, Claudio De Persis, and Tae-Woong Yoon 2 Department of Electrical Engineering

More information

Asymptotic Perron Method for Stochastic Games and Control

Asymptotic Perron Method for Stochastic Games and Control Asymptotic Perron Method for Stochastic Games and Control Mihai Sîrbu, The University of Texas at Austin Methods of Mathematical Finance a conference in honor of Steve Shreve s 65th birthday Carnegie Mellon

More information

Differential games withcontinuous, switching and impulse controls

Differential games withcontinuous, switching and impulse controls Differential games withcontinuous, switching and impulse controls A.J. Shaiju a,1, Sheetal Dharmatti b,,2 a TIFR Centre, IISc Campus, Bangalore 5612, India b Department of Mathematics, Indian Institute

More information

1 Lyapunov theory of stability

1 Lyapunov theory of stability M.Kawski, APM 581 Diff Equns Intro to Lyapunov theory. November 15, 29 1 1 Lyapunov theory of stability Introduction. Lyapunov s second (or direct) method provides tools for studying (asymptotic) stability

More information

An asymptotic ratio characterization of input-to-state stability

An asymptotic ratio characterization of input-to-state stability 1 An asymptotic ratio characterization of input-to-state stability Daniel Liberzon and Hyungbo Shim Abstract For continuous-time nonlinear systems with inputs, we introduce the notion of an asymptotic

More information

Mean field games and related models

Mean field games and related models Mean field games and related models Fabio Camilli SBAI-Dipartimento di Scienze di Base e Applicate per l Ingegneria Facoltà di Ingegneria Civile ed Industriale Email: Camilli@dmmm.uniroma1.it Web page:

More information

Numerical Methods for Optimal Control Problems. Part I: Hamilton-Jacobi-Bellman Equations and Pontryagin Minimum Principle

Numerical Methods for Optimal Control Problems. Part I: Hamilton-Jacobi-Bellman Equations and Pontryagin Minimum Principle Numerical Methods for Optimal Control Problems. Part I: Hamilton-Jacobi-Bellman Equations and Pontryagin Minimum Principle Ph.D. course in OPTIMAL CONTROL Emiliano Cristiani (IAC CNR) e.cristiani@iac.cnr.it

More information

1 The Observability Canonical Form

1 The Observability Canonical Form NONLINEAR OBSERVERS AND SEPARATION PRINCIPLE 1 The Observability Canonical Form In this Chapter we discuss the design of observers for nonlinear systems modelled by equations of the form ẋ = f(x, u) (1)

More information

Metric Spaces and Topology

Metric Spaces and Topology Chapter 2 Metric Spaces and Topology From an engineering perspective, the most important way to construct a topology on a set is to define the topology in terms of a metric on the set. This approach underlies

More information

Nonlinear representation, backward SDEs, and application to the Principal-Agent problem

Nonlinear representation, backward SDEs, and application to the Principal-Agent problem Nonlinear representation, backward SDEs, and application to the Principal-Agent problem Ecole Polytechnique, France April 4, 218 Outline The Principal-Agent problem Formulation 1 The Principal-Agent problem

More information

Chapter 2 Event-Triggered Sampling

Chapter 2 Event-Triggered Sampling Chapter Event-Triggered Sampling In this chapter, some general ideas and basic results on event-triggered sampling are introduced. The process considered is described by a first-order stochastic differential

More information

On Sontag s Formula for the Input-to-State Practical Stabilization of Retarded Control-Affine Systems

On Sontag s Formula for the Input-to-State Practical Stabilization of Retarded Control-Affine Systems On Sontag s Formula for the Input-to-State Practical Stabilization of Retarded Control-Affine Systems arxiv:1206.4240v1 [math.oc] 19 Jun 2012 P. Pepe Abstract In this paper input-to-state practically stabilizing

More information

An homotopy method for exact tracking of nonlinear nonminimum phase systems: the example of the spherical inverted pendulum

An homotopy method for exact tracking of nonlinear nonminimum phase systems: the example of the spherical inverted pendulum 9 American Control Conference Hyatt Regency Riverfront, St. Louis, MO, USA June -, 9 FrA.5 An homotopy method for exact tracking of nonlinear nonminimum phase systems: the example of the spherical inverted

More information

Nonlinear Control Lecture # 14 Input-Output Stability. Nonlinear Control

Nonlinear Control Lecture # 14 Input-Output Stability. Nonlinear Control Nonlinear Control Lecture # 14 Input-Output Stability L Stability Input-Output Models: y = Hu u(t) is a piecewise continuous function of t and belongs to a linear space of signals The space of bounded

More information

Modeling and Analysis of Dynamic Systems

Modeling and Analysis of Dynamic Systems Modeling and Analysis of Dynamic Systems Dr. Guillaume Ducard Fall 2017 Institute for Dynamic Systems and Control ETH Zurich, Switzerland G. Ducard c 1 / 57 Outline 1 Lecture 13: Linear System - Stability

More information

ON THE POLICY IMPROVEMENT ALGORITHM IN CONTINUOUS TIME

ON THE POLICY IMPROVEMENT ALGORITHM IN CONTINUOUS TIME ON THE POLICY IMPROVEMENT ALGORITHM IN CONTINUOUS TIME SAUL D. JACKA AND ALEKSANDAR MIJATOVIĆ Abstract. We develop a general approach to the Policy Improvement Algorithm (PIA) for stochastic control problems

More information

c 2004 Society for Industrial and Applied Mathematics

c 2004 Society for Industrial and Applied Mathematics SIAM J. COTROL OPTIM. Vol. 43, o. 4, pp. 1222 1233 c 2004 Society for Industrial and Applied Mathematics OZERO-SUM STOCHASTIC DIFFERETIAL GAMES WITH DISCOTIUOUS FEEDBACK PAOLA MAUCCI Abstract. The existence

More information

arxiv: v2 [math.ap] 28 Nov 2016

arxiv: v2 [math.ap] 28 Nov 2016 ONE-DIMENSIONAL SAIONARY MEAN-FIELD GAMES WIH LOCAL COUPLING DIOGO A. GOMES, LEVON NURBEKYAN, AND MARIANA PRAZERES arxiv:1611.8161v [math.ap] 8 Nov 16 Abstract. A standard assumption in mean-field game

More information

Alberto Bressan. Department of Mathematics, Penn State University

Alberto Bressan. Department of Mathematics, Penn State University Non-cooperative Differential Games A Homotopy Approach Alberto Bressan Department of Mathematics, Penn State University 1 Differential Games d dt x(t) = G(x(t), u 1(t), u 2 (t)), x(0) = y, u i (t) U i

More information

2. Dual space is essential for the concept of gradient which, in turn, leads to the variational analysis of Lagrange multipliers.

2. Dual space is essential for the concept of gradient which, in turn, leads to the variational analysis of Lagrange multipliers. Chapter 3 Duality in Banach Space Modern optimization theory largely centers around the interplay of a normed vector space and its corresponding dual. The notion of duality is important for the following

More information

Linear Quadratic Zero-Sum Two-Person Differential Games

Linear Quadratic Zero-Sum Two-Person Differential Games Linear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard To cite this version: Pierre Bernhard. Linear Quadratic Zero-Sum Two-Person Differential Games. Encyclopaedia of Systems and Control,

More information

On the Inherent Robustness of Suboptimal Model Predictive Control

On the Inherent Robustness of Suboptimal Model Predictive Control On the Inherent Robustness of Suboptimal Model Predictive Control James B. Rawlings, Gabriele Pannocchia, Stephen J. Wright, and Cuyler N. Bates Department of Chemical and Biological Engineering and Computer

More information

ESTIMATES ON THE PREDICTION HORIZON LENGTH IN MODEL PREDICTIVE CONTROL

ESTIMATES ON THE PREDICTION HORIZON LENGTH IN MODEL PREDICTIVE CONTROL ESTIMATES ON THE PREDICTION HORIZON LENGTH IN MODEL PREDICTIVE CONTROL K. WORTHMANN Abstract. We are concerned with model predictive control without stabilizing terminal constraints or costs. Here, our

More information

Nonsmooth Analysis in Systems and Control Theory

Nonsmooth Analysis in Systems and Control Theory Nonsmooth Analysis in Systems and Control Theory Francis Clarke Institut universitaire de France et Université de Lyon [January 2008. To appear in the Encyclopedia of Complexity and System Science, Springer.]

More information

The Minimum Speed for a Blocking Problem on the Half Plane

The Minimum Speed for a Blocking Problem on the Half Plane The Minimum Speed for a Blocking Problem on the Half Plane Alberto Bressan and Tao Wang Department of Mathematics, Penn State University University Park, Pa 16802, USA e-mails: bressan@mathpsuedu, wang

More information

WEAK LOWER SEMI-CONTINUITY OF THE OPTIMAL VALUE FUNCTION AND APPLICATIONS TO WORST-CASE ROBUST OPTIMAL CONTROL PROBLEMS

WEAK LOWER SEMI-CONTINUITY OF THE OPTIMAL VALUE FUNCTION AND APPLICATIONS TO WORST-CASE ROBUST OPTIMAL CONTROL PROBLEMS WEAK LOWER SEMI-CONTINUITY OF THE OPTIMAL VALUE FUNCTION AND APPLICATIONS TO WORST-CASE ROBUST OPTIMAL CONTROL PROBLEMS ROLAND HERZOG AND FRANK SCHMIDT Abstract. Sufficient conditions ensuring weak lower

More information

Singular Perturbations of Stochastic Control Problems with Unbounded Fast Variables

Singular Perturbations of Stochastic Control Problems with Unbounded Fast Variables Singular Perturbations of Stochastic Control Problems with Unbounded Fast Variables Joao Meireles joint work with Martino Bardi and Guy Barles University of Padua, Italy Workshop "New Perspectives in Optimal

More information

Limit value of dynamic zero-sum games with vanishing stage duration

Limit value of dynamic zero-sum games with vanishing stage duration Limit value of dynamic zero-sum games with vanishing stage duration Sylvain Sorin IMJ-PRG Université P. et M. Curie - Paris 6 sylvain.sorin@imj-prg.fr Stochastic Methods in Game Theory National University

More information

Some Background Material

Some Background Material Chapter 1 Some Background Material In the first chapter, we present a quick review of elementary - but important - material as a way of dipping our toes in the water. This chapter also introduces important

More information

Networked Control Systems, Event-Triggering, Small-Gain Theorem, Nonlinear

Networked Control Systems, Event-Triggering, Small-Gain Theorem, Nonlinear EVENT-TRIGGERING OF LARGE-SCALE SYSTEMS WITHOUT ZENO BEHAVIOR C. DE PERSIS, R. SAILER, AND F. WIRTH Abstract. We present a Lyapunov based approach to event-triggering for large-scale systems using a small

More information

Min-Max Model Predictive Control of Nonlinear Systems using Discontinuous Feedbacks

Min-Max Model Predictive Control of Nonlinear Systems using Discontinuous Feedbacks Min-Ma Model Predictive Control of Nonlinear Systems using Discontinuous Feedbacks Fernando A. C. C. Fontes and Lalo Magni Abstract This paper proposes a Model Predictive Control (MPC) algorithm for the

More information

Topic # /31 Feedback Control Systems. Analysis of Nonlinear Systems Lyapunov Stability Analysis

Topic # /31 Feedback Control Systems. Analysis of Nonlinear Systems Lyapunov Stability Analysis Topic # 16.30/31 Feedback Control Systems Analysis of Nonlinear Systems Lyapunov Stability Analysis Fall 010 16.30/31 Lyapunov Stability Analysis Very general method to prove (or disprove) stability of

More information

Value Function and Optimal Trajectories for some State Constrained Control Problems

Value Function and Optimal Trajectories for some State Constrained Control Problems Value Function and Optimal Trajectories for some State Constrained Control Problems Hasnaa Zidani ENSTA ParisTech, Univ. of Paris-saclay "Control of State Constrained Dynamical Systems" Università di Padova,

More information

Mean Field Games on networks

Mean Field Games on networks Mean Field Games on networks Claudio Marchi Università di Padova joint works with: S. Cacace (Rome) and F. Camilli (Rome) C. Marchi (Univ. of Padova) Mean Field Games on networks Roma, June 14 th, 2017

More information

Maths 212: Homework Solutions

Maths 212: Homework Solutions Maths 212: Homework Solutions 1. The definition of A ensures that x π for all x A, so π is an upper bound of A. To show it is the least upper bound, suppose x < π and consider two cases. If x < 1, then

More information

Minimal time mean field games

Minimal time mean field games based on joint works with Samer Dweik and Filippo Santambrogio PGMO Days 2017 Session Mean Field Games and applications EDF Lab Paris-Saclay November 14th, 2017 LMO, Université Paris-Sud Université Paris-Saclay

More information

Numerical methods for Mean Field Games: additional material

Numerical methods for Mean Field Games: additional material Numerical methods for Mean Field Games: additional material Y. Achdou (LJLL, Université Paris-Diderot) July, 2017 Luminy Convergence results Outline 1 Convergence results 2 Variational MFGs 3 A numerical

More information

Deterministic Dynamic Programming

Deterministic Dynamic Programming Deterministic Dynamic Programming 1 Value Function Consider the following optimal control problem in Mayer s form: V (t 0, x 0 ) = inf u U J(t 1, x(t 1 )) (1) subject to ẋ(t) = f(t, x(t), u(t)), x(t 0

More information

Filtered scheme and error estimate for first order Hamilton-Jacobi equations

Filtered scheme and error estimate for first order Hamilton-Jacobi equations and error estimate for first order Hamilton-Jacobi equations Olivier Bokanowski 1 Maurizio Falcone 2 2 1 Laboratoire Jacques-Louis Lions, Université Paris-Diderot (Paris 7) 2 SAPIENZA - Università di Roma

More information

The Role of Exosystems in Output Regulation

The Role of Exosystems in Output Regulation 1 The Role of Exosystems in Output Regulation Lassi Paunonen In this paper we study the role of the exosystem in the theory of output regulation for linear infinite-dimensional systems. The main result

More information

EXISTENCE AND UNIQUENESS THEOREMS FOR ORDINARY DIFFERENTIAL EQUATIONS WITH LINEAR PROGRAMS EMBEDDED

EXISTENCE AND UNIQUENESS THEOREMS FOR ORDINARY DIFFERENTIAL EQUATIONS WITH LINEAR PROGRAMS EMBEDDED EXISTENCE AND UNIQUENESS THEOREMS FOR ORDINARY DIFFERENTIAL EQUATIONS WITH LINEAR PROGRAMS EMBEDDED STUART M. HARWOOD AND PAUL I. BARTON Key words. linear programs, ordinary differential equations, embedded

More information

Global Stability and Asymptotic Gain Imply Input-to-State Stability for State-Dependent Switched Systems

Global Stability and Asymptotic Gain Imply Input-to-State Stability for State-Dependent Switched Systems 2018 IEEE Conference on Decision and Control (CDC) Miami Beach, FL, USA, Dec. 17-19, 2018 Global Stability and Asymptotic Gain Imply Input-to-State Stability for State-Dependent Switched Systems Shenyu

More information

Differential Games II. Marc Quincampoix Université de Bretagne Occidentale ( Brest-France) SADCO, London, September 2011

Differential Games II. Marc Quincampoix Université de Bretagne Occidentale ( Brest-France) SADCO, London, September 2011 Differential Games II Marc Quincampoix Université de Bretagne Occidentale ( Brest-France) SADCO, London, September 2011 Contents 1. I Introduction: A Pursuit Game and Isaacs Theory 2. II Strategies 3.

More information

Game Theory Extra Lecture 1 (BoB)

Game Theory Extra Lecture 1 (BoB) Game Theory 2014 Extra Lecture 1 (BoB) Differential games Tools from optimal control Dynamic programming Hamilton-Jacobi-Bellman-Isaacs equation Zerosum linear quadratic games and H control Baser/Olsder,

More information

Noncooperative continuous-time Markov games

Noncooperative continuous-time Markov games Morfismos, Vol. 9, No. 1, 2005, pp. 39 54 Noncooperative continuous-time Markov games Héctor Jasso-Fuentes Abstract This work concerns noncooperative continuous-time Markov games with Polish state and

More information

High-Gain Observers in Nonlinear Feedback Control

High-Gain Observers in Nonlinear Feedback Control High-Gain Observers in Nonlinear Feedback Control Lecture # 1 Introduction & Stabilization High-Gain ObserversinNonlinear Feedback ControlLecture # 1Introduction & Stabilization p. 1/4 Brief History Linear

More information

Converse Lyapunov-Krasovskii Theorems for Systems Described by Neutral Functional Differential Equation in Hale s Form

Converse Lyapunov-Krasovskii Theorems for Systems Described by Neutral Functional Differential Equation in Hale s Form Converse Lyapunov-Krasovskii Theorems for Systems Described by Neutral Functional Differential Equation in Hale s Form arxiv:1206.3504v1 [math.ds] 15 Jun 2012 P. Pepe I. Karafyllis Abstract In this paper

More information

I. D. Landau, A. Karimi: A Course on Adaptive Control Adaptive Control. Part 9: Adaptive Control with Multiple Models and Switching

I. D. Landau, A. Karimi: A Course on Adaptive Control Adaptive Control. Part 9: Adaptive Control with Multiple Models and Switching I. D. Landau, A. Karimi: A Course on Adaptive Control - 5 1 Adaptive Control Part 9: Adaptive Control with Multiple Models and Switching I. D. Landau, A. Karimi: A Course on Adaptive Control - 5 2 Outline

More information

Waves in a Shock Tube

Waves in a Shock Tube Waves in a Shock Tube Ivan Christov c February 5, 005 Abstract. This paper discusses linear-wave solutions and simple-wave solutions to the Navier Stokes equations for an inviscid and compressible fluid

More information

2012 NCTS Workshop on Dynamical Systems

2012 NCTS Workshop on Dynamical Systems Barbara Gentz gentz@math.uni-bielefeld.de http://www.math.uni-bielefeld.de/ gentz 2012 NCTS Workshop on Dynamical Systems National Center for Theoretical Sciences, National Tsing-Hua University Hsinchu,

More information

Input to state Stability

Input to state Stability Input to state Stability Mini course, Universität Stuttgart, November 2004 Lars Grüne, Mathematisches Institut, Universität Bayreuth Part IV: Applications ISS Consider with solutions ϕ(t, x, w) ẋ(t) =

More information

A LaSalle version of Matrosov theorem

A LaSalle version of Matrosov theorem 5th IEEE Conference on Decision Control European Control Conference (CDC-ECC) Orlo, FL, USA, December -5, A LaSalle version of Matrosov theorem Alessro Astolfi Laurent Praly Abstract A weak version of

More information