ON THE ESSENTIAL BOUNDEDNESS OF SOLUTIONS TO PROBLEMS IN PIECEWISE LINEAR-QUADRATIC OPTIMAL CONTROL. R.T. Rockafellar*

Similar documents
Duality and dynamics in Hamilton-Jacobi theory for fully convex problems of control

An introduction to Mathematical Theory of Control

Local strong convexity and local Lipschitz continuity of the gradient of convex functions

DUALIZATION OF SUBGRADIENT CONDITIONS FOR OPTIMALITY

NONSMOOTH ANALYSIS AND PARAMETRIC OPTIMIZATION. R. T. Rockafellar*

Extremal Solutions of Differential Inclusions via Baire Category: a Dual Approach

Convex Optimization Theory. Chapter 5 Exercises and Solutions: Extended Version

A Geometric Framework for Nonconvex Optimization Duality using Augmented Lagrangian Functions

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Some Properties of the Augmented Lagrangian in Cone Constrained Optimization

A Unified Analysis of Nonconvex Optimization Duality and Penalty Methods with General Augmenting Functions

Extended Monotropic Programming and Duality 1

FINITE-DIFFERENCE APPROXIMATIONS AND OPTIMAL CONTROL OF THE SWEEPING PROCESS. BORIS MORDUKHOVICH Wayne State University, USA

BASIC ISSUES IN LAGRANGIAN OPTIMIZATION

LINEAR-CONVEX CONTROL AND DUALITY

ON A CLASS OF NONSMOOTH COMPOSITE FUNCTIONS

On duality theory of conic linear problems

Extended Nonlinear Programming

HAMILTON-JACOBI THEORY AND PARAMETRIC ANALYSIS IN FULLY CONVEX PROBLEMS OF OPTIMAL CONTROL. R. T. Rockafellar 1

Convex Optimization Theory. Athena Scientific, Supplementary Chapter 6 on Convex Optimization Algorithms

A DUALITY APPROXIMATION OF SOME NONLINEAR PDE s

MAXIMALITY OF SUMS OF TWO MAXIMAL MONOTONE OPERATORS

Optimality Conditions for Constrained Optimization

c???? Society for Industrial and Applied Mathematics Vol. 1, No. 1, pp ,???? 000

2 Statement of the problem and assumptions

Unbounded Convex Semialgebraic Sets as Spectrahedral Shadows

McGill University Math 354: Honors Analysis 3

Obstacle problems and isotonicity

An introduction to some aspects of functional analysis

Examples include: (a) the Lorenz system for climate and weather modeling (b) the Hodgkin-Huxley system for neuron modeling

Maximal Monotone Inclusions and Fitzpatrick Functions

ZERO DUALITY GAP FOR CONVEX PROGRAMS: A GENERAL RESULT

Subdifferential representation of convex functions: refinements and applications

An Integral-type Constraint Qualification for Optimal Control Problems with State Constraints

Asymptotic Convergence of the Steepest Descent Method for the Exponential Penalty in Linear Programming

Functional Analysis. Franck Sueur Metric spaces Definitions Completeness Compactness Separability...

Linear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard June 15, 2013

Convex analysis and profit/cost/support functions

("-1/' .. f/ L) I LOCAL BOUNDEDNESS OF NONLINEAR, MONOTONE OPERA TORS. R. T. Rockafellar. MICHIGAN MATHEMATICAL vol. 16 (1969) pp.

(convex combination!). Use convexity of f and multiply by the common denominator to get. Interchanging the role of x and y, we obtain that f is ( 2M ε

Convex Optimization M2

Extreme Abridgment of Boyd and Vandenberghe s Convex Optimization

Perturbation Analysis of Optimization Problems

Research Article Complete Solutions to General Box-Constrained Global Optimization Problems

Duality (Continued) min f ( x), X R R. Recall, the general primal problem is. The Lagrangian is a function. defined by

Lagrangian Duality and Convex Optimization

Radius Theorems for Monotone Mappings

On Semicontinuity of Convex-valued Multifunctions and Cesari s Property (Q)

Subdifferentiability and the Duality Gap

Lecture 7 Monotonicity. September 21, 2008

Inexact alternating projections on nonconvex sets

Calculus of Variations. Final Examination

Convex Sets and Minimal Sublinear Functions

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 4. Subgradient

Jørgen Tind, Department of Statistics and Operations Research, University of Copenhagen, Universitetsparken 5, 2100 Copenhagen O, Denmark.

Continuous Sets and Non-Attaining Functionals in Reflexive Banach Spaces

Existence of Global Minima for Constrained Optimization 1

Lecture 5. The Dual Cone and Dual Problem

Lecture 1: Entropy, convexity, and matrix scaling CSE 599S: Entropy optimality, Winter 2016 Instructor: James R. Lee Last updated: January 24, 2016

Lecture 4. Chapter 4: Lyapunov Stability. Eugenio Schuster. Mechanical Engineering and Mechanics Lehigh University.

Lagrange duality. The Lagrangian. We consider an optimization program of the form

LECTURE 15: COMPLETENESS AND CONVEXITY

POLARS AND DUAL CONES

Approximation of Continuous-Time Infinite-Horizon Optimal Control Problems Arising in Model Predictive Control

COSMIC CONVERGENCE. R.T. Rockafellar Departments of Mathematics and Applied Mathematics University of Washington, Seattle

On the Existence and Convergence of the Central Path for Convex Programming and Some Duality Results

The proximal mapping

Convex Optimization Conjugate, Subdifferential, Proximation

arxiv: v1 [math.oc] 17 Oct 2017

Closing the duality gap in linear vector optimization

The goal of this chapter is to study linear systems of ordinary differential equations: dt,..., dx ) T

Deterministic Dynamic Programming

minimize x subject to (x 2)(x 4) u,

Convex Optimization & Lagrange Duality

INF-SUP CONDITION FOR OPERATOR EQUATIONS

Conjugate convex Lyapunov functions for dual linear differential inclusions

Quadratic Two-Stage Stochastic Optimization with Coherent Measures of Risk

CHAPTER 7. Connectedness

Self-equilibrated Functions in Dual Vector Spaces: a Boundedness Criterion

Enhanced Fritz John Optimality Conditions and Sensitivity Analysis

Linear Programming. Larry Blume Cornell University, IHS Vienna and SFI. Summer 2016

1 Cheeger differentiation

CONVEX OPTIMIZATION VIA LINEARIZATION. Miguel A. Goberna. Universidad de Alicante. Iberian Conference on Optimization Coimbra, November, 2006

Convergence of generalized entropy minimizers in sequences of convex problems

Dedicated to Michel Théra in honor of his 70th birthday

Subgradients. subgradients. strong and weak subgradient calculus. optimality conditions via subgradients. directional derivatives

Convexification by Duality for a Leontief Technology Production Design Problem

A convergence result for an Outer Approximation Scheme

ANALYSIS QUALIFYING EXAM FALL 2017: SOLUTIONS. 1 cos(nx) lim. n 2 x 2. g n (x) = 1 cos(nx) n 2 x 2. x 2.

Optimality Conditions for Nonsmooth Convex Optimization

ON GAP FUNCTIONS OF VARIATIONAL INEQUALITY IN A BANACH SPACE. Sangho Kum and Gue Myung Lee. 1. Introduction

BORIS MORDUKHOVICH Wayne State University Detroit, MI 48202, USA. Talk given at the SPCOM Adelaide, Australia, February 2015

Lecture 7: Semidefinite programming

EVANESCENT SOLUTIONS FOR LINEAR ORDINARY DIFFERENTIAL EQUATIONS

FEEDBACK DIFFERENTIAL SYSTEMS: APPROXIMATE AND LIMITING TRAJECTORIES

8. Conjugate functions

Exact SDP Relaxations for Classes of Nonlinear Semidefinite Programming Problems

Convex Analysis and Optimization Chapter 2 Solutions

GENERALIZED CONVEXITY AND OPTIMALITY CONDITIONS IN SCALAR AND VECTOR OPTIMIZATION

Division of the Humanities and Social Sciences. Supergradients. KC Border Fall 2001 v ::15.45

Transcription:

ON THE ESSENTIAL BOUNDEDNESS OF SOLUTIONS TO PROBLEMS IN PIECEWISE LINEAR-QUADRATIC OPTIMAL CONTROL R.T. Rockafellar* Dedicated to J-L. Lions on his 60 th birthday Abstract. Primal and dual problems of optimal control with linear, quadratic or piecewise linear-quadratic convex objective are considered in which a linear dynamical system is subjected to linear inequality constraints that could jointly involve states and controls. It is shown that when such constraints, except for the ones on controls only, are represented by penalty terms, and a mild coercivity condition is satisfied, the optimal controls for both problems will be essentially bounded in time. The optimal trajectories will thus be Lipschitzian. * This research was supported in part by a grant from the National Science Foundation at the University of Washington, Seattle 1

1 1. Introduction In problems of optimal control of ordinary differential equations, the linear-quadratic case usually refers to a formulation where the dynamics are linear and the objective is quadratic convex (or affine). A broader formulation, termed generalized linear-quadratic in Rockafellar [1], allows the integrand in the objective to be piecewise linear-quadratic (convex) in the state and control vectors. By definition, a function is piecewise linearquadratic if its effective domain can be expressed as the union of finitely many polyhedral convex sets, relative to each of which the function is at most quadratic. The introduction of piecewise linear-quadratic expressions can serve various purposes in mathematical modeling, but one of the most important is the handling of various linear inequality constraints by means of penalty terms. This is especially useful in the case of state constraints or restrictions that make the current set of available controls depend on the current state. Such restrictions, which fall outside the usual patterns of treatment in optimal control, are essential to many applications in economic management and operations research. The theory of piecewise linear-quadratic optimal control, as developed in [1], emphasizes duality in the statement of necessary and sufficient conditions and includes results on the existence of primal and dual solutions. In order to achieve such results, one is obliged to work with control spaces consisting of L 1 functions, at least initially, even though a restriction to L would be more natural for most applications. The purpose of the present paper is to demonstrate that the necessary conditions derived in this way imply, under the supporting assumptions, that the optimal controls are L functions after all. Thus the introduction of L 1 controls can be seen merely as a technical excursion leading in the end to a solid justification for an L framework. 2. Primal and Dual Problems The basic control problem in [1] involves certain ρ expressions defined in general by (2.1) ρ V,Q (r) = sup{r v 2v 1 Qv}, v V where V denotes a nonempty polyhedral convex set in lr l, and Q is a symmetric, positive semidefinite matrix in lr l l (possibly the zero matrix!). The nature of such expressions, the alternative formulas they satisfy, and their roles and special cases, are discussed in detail in [1]. Suffice it to say here that ρ V,Q is in general a piecewise linear-quadratic convex function of r which depends in a convenient manner on the choice of V and Q as parameters. If 0 V, then ρ V,Q is nonnegative everywhere, as usually expected of a 1

2 penalty function. Moreover the set of vectors r where ρ V,Q actually vanishes is a certain polyhedral convex cone K V,Q (possibly a subspace or just the zero vector itself). The primal problem we deal with is the following, where the subscript t refers to time in a fixed interval [t 0, t 1 ] and the subscript e designates endpoint elements : (P 1 ) subject to the dynamical system ẋ t = A t x t + B t u t + b t a.e., quadu t U t a.e., x t0 = B e u e + b e, u e U e, with u : t u t an L 1 function, minimize t1 [p t u t + 1 2u t P t u t c t x t ]dt + [p e u e + 1 2u e P e u e c e x t1 ] t 0 t1 + ρ Vt,Q t (q t C t x t D t u t )dt + ρ V e,qe(q e C e x t1 D e u e ). t 0 Here we have written u t, x t and ẋ t in place of U(t), x(t) and ẋ(t), with u t lr k, x t lr n. The endpoint control vector u e lr ke represents additions parameters supplementary to the instantaneous control vectors u t, but with respect to which some optimization can also take place. The sets U t lr k, U e lr ke, V t lr l, V e lr l e are assumed to be polyhedral convex and nonempty. The matrices P t, P e, Q t, Q e, are assumed to be symmetric and positive semidefinite. Furthermore it is assumed that the elements (2.2) A t, B t, C t, D t, b t, c t, P t, Q t, p t, q t, U t, V t all depend continuously on t [t 0, t 1 ]. Under these assumptions the choice of u and u e, where u denotes the L 1 control function t u t as indicated in the statement of (P), determines by way of the dynamical system in (P) a unique corresponding trajectory x : t x t, which is absolutely continuous. The associated value of the objective functional is then well defined as a real number of, and it is convex in its dependence on (u, u e ); see [1, Thm. 4.2]. The ρ-expressions in (P 1 ) can be viewed broadly as monitoring terms. They monitor the vector C t x t +D t u t versus a given q t, and also C e x t1 +D e u e versus q e. Certain directions or amounts of deviation may be accepted without cost (namely those corresponding to difference vectors in the polyhedral cones K Qt,V t and K Q e,ve, in the case where 0 V t and 2

3 0 V e ), but others may be penalized (or conceivably even rewarded, in cases other than ones where 0 V t and 0 V e ). The problem proposed in [1] as dual to (P) is entirely analogous in form, but with a reversal of roles for the various elements (* denotes transpose): (Q 1 ) Subject to the dynamical system ẏ t = A t y t + C t v t + c t a.e., v t V t a.e., y t0 = C e v e + c e, v e V e, with v : t v t an L 1 function of t, maximize t1 [q t v t 2v 1 t Q t v t b t y t ]dt + [q e v e 1 2v e Q e v e b e y e ] t 0 t1 ρ Ut,P t (Bt y t + Dt v t p t )dt ρ U e,pe (B e y t0 + Dev e p e ). t 0 Again, the choice of the control function v : t v t and vector v e uniquely determines the dual state trajectory y : t y t, which is absolutely continuous. The objective functional in (Q 1 ) is concave in (v, v e ) with values in lr { }. The main results in [1] about the relationship between (P 1 ) and (Q 1 ) concern duality. For present purposes we focus only on the strongest case. Definition 2.1. One says that the primal finiteness condition is satisfied if ρ Vt,Q t < everywhere on lr l and ρ V e,qe < everywhere on lr le. Likewise the dual finiteness condition is satisfied if ρ Ut,P t < everywhere on lr k and ρ U e,pe < everywhere on lr ke. These conditions mean roughly that the monitoring terms in the two control problems do not involve any implicit constraints which are enforced through infinite penalization. It is demonstrated in [1, Prop. 2.4] that the primal finiteness condition holds if and only if (2.3) rc V t nl Q t = {0} for all t [t 0, t 1 ] and rc V e nl Q e = {0}, where rc denotes the recession cone of a convex set [2, 8] and nl denotes the null space of a matrix. Similarly, the dual finiteness condition holds if and only if (2.4) rc U t nl P t = {0} for all t [t 0, t 1 ] and rc U e nl P e = {0}. Condition (2.3) can appropriately be called the primal coercivity condition, and (2.5) the dual coercivity condition. Thus instead of the primal and dual finiteness conditions we could speak of imposing the primal finiteness and coercivity conditions, or equivalently of imposing the dual finiteness and coercivity conditions. 3

4 Theorem 2.2. Under the assumption that the primal and dual finiteness conditions both hold, one has < min(p 1 ) = max(q 1 ) <. In other words, optimal controls exist for both problems, and the corresponding objective values are finite and equal. Proof. This is a special case of [1, Thm. 6.3]. When the finiteness conditions in Theorem 2.2 are invoked separately, each still guarantees the equality of the optimal values in (P 1 ) and (Q 1 ) but only yields one of the existence assertions. See the full version of this result in [1, Thm. 6.3]. Optimality of primal and dual controls in the context of Theorem 2.2 can be expressed as a minimaximum principle. To state this we introduce the quadratic convex-concave functions (2.5) J t (u t, v t ) = p t u t + 1 2u t P t u t + q t v t 1 2v t Q t v t v t D t u t, (2.6) J e (u e, v e ) = p e u e + 1 2u e P e u e + q e v e 1 2v e Q e v e v e D e u e. Theorem 2.3. In the case where min(p 1 ) = max(q 1 ), the following conditions are both necessary and sufficient in order that (ū, ū e ) be optimal in (P 1 ) and ( v, v e ) be optimal in (Q 1 ). In terms of the primal trajectory x corresponding to (ū, ū e ) and the dual trajectory ȳ corresponding to ( v, v e ), one has that (2.7) (ū t, v t ) is a saddle point of J t (u t, v t ) u t B t ȳ t v t C t x t relative to u t U t, v t V t (a.e.), (2.8) (ū e, v e ) is a saddle point of J e (u e, v e ) u e B e ȳ t0 v e C e x t1 relative to u e U e, v e V e, (where ū : t ū t and v : t v t are L 1 functions). Proof. This combines a global saddlepoint criterion for optimality in [1, Thm. 6.2] with a pointwise saddlepoint criterion in [1, Thm. 6.5]. 4

5 3. Reduction to Bounded Controls. Let us denote by (P ) and (Q ) the modified forms of the primal problem (P 1 ) and dual problem (Q 1 ) in which the control functions u : t u t and v : t v t are required to be L rather than L 1. The optimal values in these problems satisfy (3.1) inf(p ) inf(p 1 ) sup(q 1 ) sup(q ) in general (the middle inequality holds always by [1, Thm. 6.2]). This relationship indicates that the prospect of duality between (P ) and (Q ) is less promising than between (P 1 ) and (Q 1 ), despite the fact that (P ) and (Q ) are easier to work with in other respects and more natural as problem formulations for many situations. We demonstrate now, though, that duality between (P ) and (Q ) does hold nonetheless in the important case where the primal and dual finiteness conditions are both satisfied. Theorem 3.1. Under the primal and dual boundedness conditions, every optimal solution (ū, ū e ) to (P 1 ) actually has ū L and therefore is an optimal solution to (P ); likewise, every optimal solution ( v, v e ) to (Q 1 ) has v L and therefore is an optimal solution to (Q ). Thus < min(p ) = max(q ) <. Proof. In view of Theorems 2.2 and 2.3, the task before us is to demonstrate that in the presence of the primal and dual finiteness conditions, the instantaneous saddlepoint condition (2.7) implies ū L, v L. We shall do this in the framework of conjugate convex-concave functions [2]. Define the function J t on lr k lr l by J t (u t, v t ) if u t U t, v t V t, (3.2) Jt (u t, v t ) = if u t U t, v t V t, if u t U t. This is a closed saddlefunction in the terminology of [2, 34], inasmuch as J t is a continuous convex-concave function and the sets U t and V t are closed and convex. The saddlepoint condition (2.5) is equivalent by definition to the subdifferential relation (3.3) (Bt ȳ t, C t x t ) J t (ū t, v t ) a.e. (see [2, p. 374]). By introducing the conjugate (3.4) J t (r, s) = inf v t lr l sup {r u t + s v t J t (u t, v t )} u t lr k = inf sup {r u t + s v t J t (u t, v t )}, v t V t u t U t 5

6 which is another closed saddlefunction [2, Cor. 37.1.1], we can write this subdifferential relation as (3.5) (ū t, v t ) J t (Bt ȳ t, C t x t ), a.e. by [2, Thm. 37.5]. To prove that this implies ū L and v L, it will suffice to show that the multifunction Γ : t [t 0, t 1 ] J t (B t ȳ t, C t x t ) is uniformly bounded when the primal and dual finiteness conditions are satisfied, i.e. has all of its image sets Γ(t) contained in a certain bounded subset of lr k lr l. Because B t ȳ t and C t x t depend continuously on t and therefore themselves remain in a bounded set as t passes through [t 0, t 1 ], it will suffice to show that the multifunctions J t : lr k lr l lr k lr l are uniformly bounded on bounded sets. The formula specified in (3.4) tells us by way of (2.5) that (3.6) J t (r, s) = inf sup {[r p t ] u t + [s q t ] v t 2u 1 v t V t P t u t + 1 2v t Q t v t + v t D t u t } t u t U t = inf v t V t {[s q t ] v t + 1 2v t Q t v t } + sup u t U t {[r p t + D t v t ] u t 1 2u t P t u t } = inf v t V t {[s q t ] v t + 1 2v t Q t v t + ρ Ut,P t (r p t + D t v t )}. The set V t, which is nonempty, closed and convex, depends continuously on t according to our assumption (2.2) so there exists by the selection theorem of Michael [3] a continuous function v : t v t with v V t for all t [t 0, t 1 ]. Then (3.7) J t (r, s) [s q t ] v t + 1 2 v t Q t v t + ρ ut,p t (r p t + Dt v t ) for all (r, s) lr k lr l. Using the fundamental inequality inf sup sup inf on the formula in (3.4), we can obtain similarly that (3.8) J t (r, s) [r p t ] ū t 1 2ū t P t ū t ρ Vt,Q t (q t D t ū t s) for all (v, s) lr k lr l, where ū : t ū t is a continuous function having ū t U t for all t [t 0, t 1 ]. We now invoke the primal and dual finiteness conditions: These say that the estimates in (3.7) and (3.8) are finite. In fact they imply by [1, Prop. 4.1] that ρ Ut,P t (r) is continuous jointly in t and r, and ρ Vt,Q t (s) is continuous jointly in t and s. (The continuous dependence of P t, Q t, 6

7 U t and V t on t in (2.2) comes in here.) This continuity and that of q t, Q t, p t, P t, and D t in (2.2) produces continuity of the right sides of (3.7) and (3.8) with respect to (t, r, s). It follows then that the functions J t subset of lr k lr l. for t [t 0, t 1 ] are uniformly bounded on any bounded This boundedness property must be translated next into a uniform boundedness of the subdifferentials J t. The argument is just a refinement of facts already known in the literature on convex-concave functions. From [2, Thm. 35.2] one sees that the boundedness of the values of the functions J t implies that these functions are equi-lipschitzian on any bounded subset of lr k lr l. But if J t is Lipschitzian with modulus λ on some neighborhood of a point ( r, s), the set J t ( r, s) must be contained in the ball of radius λ around ( r, s), as follows simply from the bounds one gets in this case on directional derivatives. This observation leads to our desired conclusion that the sets J t (r, s) for t [t 0, t 1 ] are uniformly bounded relative to any bounded set of points (r, s). References 1. R.T. Rockafellar, Linear-quadratic programming and optimal control, SIAM J. Control Opt. 26, May, 1987. 2. R.T. Rockafellar, Convex Analysis, Princeton Univ. Press, 1970. 3. E. Michael, Continuous selections, I, Ann. of Math. 63(1956), 361 382. 7