Downloaded 12/29/14 to Redistribution subject to SIAM license or copyright; see

Size: px
Start display at page:

Download "Downloaded 12/29/14 to Redistribution subject to SIAM license or copyright; see"

Transcription

1 SIAM J. OPTIM. Vol. 7, No., pp c 26 Society for Industrial and Applied Mathematics PATH-FOLLOWING METHODS FOR A CLASS OF CONSTRAINED MINIMIZATION PROBLEMS IN FUNCTION SPACE MICHAEL HINTERMÜLLER AND KARL KUNISCH Abstract. Path-following methods for primal-dual active set strategies requiring a regularization parameter are introduced. Existence of a primal-dual path and its differentiability properties are analyzed. Monotonicity and convexity of the primal-dual path value function are investigated. Both feasible and infeasible approximations are considered. Numerical path-following strategies are developed and their efficiency is demonstrated by means of examples. Key words. semismooth Newton methods, path-following methods, active set strategy, primaldual methods AMS subject classifications. 49M5, 49M37, 65K5, 9C33 DOI..37/ Introduction. Primal-dual active set strategies or, in some cases equivalently, semismooth Newton methods, were proved to be efficient methods for solving constrained variational problems in function space [, 9,,, 2, 3]. In certain cases regularization is required, resulting in a family of approximating problems with more favorable properties than those of the original one, [2, 3]. In previous work [3] convergence, and in some cases rate of convergence, with respect to the regularization parameter was proved. In the numerical work the adaptation of these parameters was heuristic, however. The focus of the present investigation is on an efficient control of the regularization parameter in the primal-dual active set strategy for a class of constrained variational problems. To explain the involved issues we proceed mostly formally in this section and consider the problem { min J (v) over v X () s.t. Gv ψ, where J is a quadratic functional on a Hilbert space X, and G: X Y. It is assumed that Y L 2 () is a Hilbert lattice with ordering induced by the natural ordering of L 2 (). We note that () subsumes problems of very different nature. For example, for the control constrained optimal control problem min 2 y z 2 L + α 2 2 u 2 L 2 s.t. Δy = u in, y = on, u ψ a.e. in, with a bounded domain in R n, z L 2 (), α>, one can use y =( Δ) u, where Δ denotes the Laplacian with homogenous Dirichlet boundary conditions, and arrive Received by the editors July 4, 24; accepted for publication (in revised form) November 2, 25; published electronically May 3, 26. This research was partially supported by the Fonds zur Förderung der wissenschaftlichen Forschung under SFB 3 Optimierung und Kontrolle. Department of Computational and Applied Mathematics, Rice University, Houston, TX. Institute of Mathematics and Scientific Computing, University of Graz, Graz, Austria (michael. hintermueller@uni-graz.at, karl.kunisch@uni-graz.at). 59

2 6 MICHAEL HINTERMÜLLER AND KARL KUNISCH at { min 2 ( Δ) u z 2 + α 2 u 2 s.t. u ψ a.e. in, which is clearly of the form (). For J (v) = 2 v 2 dx fv, X = H (), and G = I we obtain the classical obstacle problem. For state constrained control problems with y ψ one has { min 2 ( Δ) u z 2 + α 2 u 2 s.t. ( Δ) u ψ a.e. in, which is also of the form (). From the point of view of duality theory these three problems are very different. While it is straightforward to argue the existence of a Lagrange multiplier in L 2 () for the control constrained optimal control problem, it is already more involved and requires additional assumptions to guarantee its existence in L 2 () for obstacle problems, and for state constrained problems the Lagrange multiplier is only a measure. If we resort to a formal discussion, then in either of these cases we arrive at the optimality system of the form { J (v)+g λ =, (2) λ = max(,λ+ c(g(v) ψ)) for any fixed c>. Here, G denotes the adjoint of G. The second equation in (2) is equivalent to λ, G(v) ψ, and λ(g(v) ψ) =. Continuing formally, the primal-dual active set strategy determines the active set at iteration level k by means of A k+ = {x : λ k (x)+c(g(v k )(x) ψ(x))> }, assigns the inactive set I k+ =\A k+, and updates (v, λ) by means of { J (v k+ )+G λ k+ =, (3) λ k+ = on I k+, (G(v k+ ) ψ)(x) = for x A k+. These auxiliary problems require special attention. For obstacle problems the constraint v k+ = ψ on A k+ induces that the associated Lagrange multiplier λ k+ is in general less regular than the Lagrange multiplier associated with v ψ for the original problem; see, e.g., [3]. For problems with combined control and state constraints it may happen that due to the assignment on I k+ and A k+, (3) has no solution while the original problem does. For these reasons in, e.g., [9, 2, 3] the second equation in (2) was regularized, resulting in the family of equations { J (v)+g λ =, (4) λ = max(, λ + γ(g(v) ψ)), where λ is fixed, possibly λ =, and γ R +. In the above-mentioned references it was shown that under appropriate conditions the solutions (v γ,λ γ ) to (4) exist, the quantity λ γ enjoys extra regularity, and (v γ,λ γ ) converge to the solution of (2) as γ +.

3 PATH-FOLLOWING METHODS 6 In previous numerical implementations the increase of γ to infinity was heuristic. As the system (4) becomes increasingly ill-conditioned as γ tends to, in this paper a framework for a properly controlled increase of γ-values will be developed in order to cope with the conditioning problem. In fact, in a typical algorithmic regime for solving () one uses the solution (v γ,λ γ ) to (4) for some γ as the initial guess for the solution to (4) for the updated γ-value γ + >γ. Typically, if γ + γ, then (v γ,λ γ )isonlyapoor approximation of (v γ +,λ γ +), which in addition to numerical linear algebra issues (like ill-conditioned system matrices) causes severe stability problems for iterative solvers for (4) such as semismooth Newton methods. Together with developing a new γ- update strategy, we aim at solving the auxiliary problems (4) only inexactly to keep the overall computational cost low. To this end we define neighborhoods of the path which allow inexact solutions and which contract in a controlled way towards the path as the iteration proceeds. Our work is inspired by concepts from path-following methods in finite dimensional spaces [4, 5, 6, 8, 9]. We first guarantee the existence of a sufficiently smooth path γ (v γ,λ γ ), with γ (, ) in appropriately chosen function spaces. Once the path is available it can be used as the basis for updating strategies of the path parameter. Given a current value γ k, with associated primal and dual states (v γk,λ γk ), the γ-update should be sufficiently large to make good progress towards satisfying the complementarity conditions. On the other hand, since we are not solving the problems along the path exactly, we have to use safeguards against steps which would lead us too far off the path. Of course, these goals are impeded by the fact that the path is not available numerically. To overcome this difficulty we use qualitative properties of the value function, like monotonicity and convexity, which can be verified analytically. These suggest the introduction of model functions which will be shown to approximate the value functional along the path very well. We use these model functions for our updating strategies of γ. In the case of exact path-following we can even prove convergence of the resulting strategy. In the present paper the program just described is carried out for a class of problems corresponding to contact problems. State constrained optimal control problems require a different approach that will be considered independently. As we shall see, the (infinite dimensional) parameter λ can be used to guarantee that the iterates of the primal variable are feasible. Further, it turns out that the numerical behavior of infeasible approximations is superior to the feasible ones from the point of view of iteration numbers. Interior point methods also require an additional parameter, which, however, enters into (2) differently. For the problem under consideration here, the interiorpoint relaxation replaces the second equation in (2) by (5) λ(x)(ψ G(v))(x) = for x. γ Path-following interior-point methods typically start strictly feasible, with iterates which are required to stay strictly feasible during the iterations while satisfying, or satisfying approximately, the first equation in (2) and (5). Path-following interiorpoint methods have not received much attention for infinite dimensional problems yet. In fact, we are aware of only [7], where such methods are analyzed for optimal control problems related to ordinary differential equations. For the problem classes that we outlined at the beginning of this section, the primal-dual active set strategy proved to be an excellent competitor to interior-point methods, as was demonstrated, for example, in [] comparing these two methods. This paper is organized as follows. Section 2 contains the precise problem formulation and the necessary background on the primal-dual active set strategy. The

4 62 MICHAEL HINTERMÜLLER AND KARL KUNISCH existence and regularity of the primal-dual path is discussed in section 3. Properties of the primal-dual path value functional are analyzed in section 4. Section 5 contains the derivation of the proposed model functions for the primal-dual path value functional. Exact as well as inexact path-following algorithms are proposed in section 6, and their numerical behavior is discussed there as well. (P) 2. Problem statement, regularization, and its motivation. We consider { min 2 a(y, y) (f,y) over y H () s.t. y ψ, where f L 2 (), ψ H (), with ψ, where is a bounded domain in R n with Lipschitz continuous boundary. Throughout, (, ) denotes the standard L 2 ()-inner product, and we assume that a(, ) is a bilinear form on H () H () satisfying (6) a(v, v) ν v 2 H and a(w, z) μ w H z H for some ν >, μ>independent of v H () and w, z H (). Here and throughout we use v H = v L 2 for v H (), which defines a norm on H () due to Friedrichs inequality, and w H =( w 2 L + w 2 2 L ) /2 denotes the standard 2 H -norm; see, e.g., [2]. Moreover, let A: H () H () be defined by a(v, w) = Av,w H,H for all v, w H (). It is well known that (P) admits a unique solution y H () with associated Lagrange multiplier λ = Ay + f, satisfying the optimality system { a(y,v)+ λ,v H,H (7) =(f,v), λ,y ψ H,H =, y ψ, λ,v for all v. This also holds with f H (). Under well-known additional requirements on a, ψ, and, as for example { a(v, w) = ( a ij v xi w xj + dvw), with a ij C ( ),d L (), (8) d, ψ H 2 (), is C,, or is a convex polyhedron, we have (y,λ ) H 2 () L 2 (), and the optimality system can be expressed as { Ay + λ = f in L 2 (), (9) λ =(λ + c(y ψ)) + for some c>, where (v) + = max(,v); for details see, e.g., [4]. Our aim is the development of Newton-type methods for solving (7) or (9), which is complicated by the system of inequalities in (7) and the nondifferentiable maxoperator in (9). In the recent past significant progress was made in the investigation of semismooth Newton methods and primal-dual active set methods for coping with nondifferentiable functionals in infinite dimensional spaces; see, for instance, [, 5]. A direct application of these techniques to (9) results in the following algorithm.

5 PATH-FOLLOWING METHODS 63 Algorithm A. (i) Choose c>, (y,λ ); set k =. (ii) Set A k+ = {x : λ k (x)+c(y k (x) ψ(x)) > }. (iii) Compute y k+ = arg min { 2 a(y, y) (f,y): y = ψ on A k+}. (iv) Let λ k+ be the Lagrange multiplier associated with the constraint in (iii), with λ k+ = on \A k+. (v) Set k := k + and go to (ii). The optimality system for the variational problem in (iii) is given by { a(yk+,v)+ λ k+,v H,H =(f,v) for all v H (), () y k+ = ψ on A k+, λ k+ = on I k+ =\A k+. This corresponds to (3) in our introductory discussion. The Lagrange multiplier associated with the constraint y = ψ on A k+ is in general only a distribution in H () and is not in L 2 (). In fact, λ k+ is related to the jumps in the normal derivatives of y across the interface between A k+ and I k+ [3]. This complicates the convergence analysis for Algorithm A since the calculus of Newton (or slant) differentiability [] does not apply. We note that these difficulties are not present if (7) or (9) is discretized. However, they are crucial for the treatment of infinite dimensional problems, and as such they are generic. Analogous difficulties arise for state constrained optimization problems, for inverse problems with BV-regularization, and for elasticity problems with contract and friction, to mention a few. This suggests the introduction of regularized problems, which in our case are chosen as min (P a(y, y) (f,y)+ ( λ + γ(y ψ)) + 2 over y H γ ) (), 2 2γ where γ>and λ L 2 (), λ are fixed. For later use we denote the objective functional of (P γ )byj(y; γ). The choice of λ will be used to influence the feasibility of the solution y γ of (P γ ). Using Lebesgue s bounded convergence theorem to differentiate the max under the integral in J(y; γ), the first order optimality condition associated with (P γ ) is given by { a(yγ,v)+(λ γ,v)=(f,v) for all v H (), (OC γ ) λ γ =( λ + γ(y γ ψ)) +, where (y γ,λ γ ) H () L 2 (). With (8) holding, we have y γ H 2 (). The primaldual active set strategy, or equivalently the semismooth Newton method, for (P γ )is given next. For its statement and for later use we introduce χ A k+, the characteristic function of the set A k+. Algorithm B. (i) Choose λ, (y,λ ); set k =. (ii) Set A k+ = {x : λ(x)+γ(y k (x) ψ(x)) > }, I k+ =\A k+. (iii) Solve for y k+ H (): a(y k+,v)+(( λ + γ(y k+ ψ))χ Ak+,v)=(f,v) for all v H (). (iv) Set λ k+ = { on Ik+, λ + γ(y k+ ψ) on A k+.

6 64 MICHAEL HINTERMÜLLER AND KARL KUNISCH Algorithm B was analyzed in [3], where global as well as locally superlinear convergence for every fixed γ> were established. However, the choice and adaptation (increase) of γ was heuristic in [3] and earlier work. The focus of the present investigation is the automatic adaptive choice of γ. We shall utilize the following two results, which we recall from [3] where the proofs can also be found. Proposition 2.. The solutions (y γ,λ γ ) to (OC γ ) converge to (y,λ ) in the sense that y γ y strongly in H () and λ γ λ weakly in H () as γ. We say that a satisfies the weak maximum principle if for any v H () () a(v, v + ) implies v + =. Proposition 2.2. Assume that () holds and let <γ γ 2 <. (a) In the infeasible case, i.e., for λ =, we have y y γ2 y γ. (b) In the feasible case, i.e., if (2) λ and λ f + Aψ, v H,H for all v H (), with v, then y γ y γ2 y ψ. 3. The primal-dual path. In this section we introduce the primal-dual path and discuss its smoothness properties. Definition 3.. The family of solutions C = {(y γ,λ γ ): γ (, )} to (OC γ ), considered as subset of H () H (), is called the primal-dual path associated with (P). For r we further set C r = {(y γ,λ γ ): γ [r, )}, and with some abuse of terminology we also refer to C r as a path. In the following lemma we denote by ŷ the solution to the unconstrained problem (ˆP) min J(y) = a(y, y) (f,y) 2 over y H (). Subsequently, in connection with convergence of a sequence in function space we use the subscript weak together with the space to indicate convergence in the weak sense. Lemma 3.2. For each r> the path C r is bounded in H () H (), with lim γ (y γ,λ γ )=(y,λ ) in H () H () weak.for λ =the path C is bounded in H () H (), with lim γ +(y γ,λ γ )=(ŷ, ) in H () L 2 (). Proof. From (OC γ ) we have for every γ> (3) a(y γ,y γ y )+(λ γ,y γ y )=(f,y γ y ). Since λ γ = max(, λ + γ(y γ ψ)) and ψ y wehave ( (λ γ,y γ y )= λ γ, λ γ + y γ ψ + ψ y λ ) γ Combined with (3) this implies that γ (λ γ, λ + γ(y γ ψ)) γ (λ γ, λ) = [ λγ 2 L γ (λ γ, λ) ]. 2 (4) a(y γ,y γ )+ γ λ γ 2 L 2 a(y γ,y )+(f,y γ y )+ γ ( λ, λ γ ).

7 PATH-FOLLOWING METHODS 65 This estimate, (6), (OC γ ), and the Poincaré Friedrichs inequality imply that C r is bounded in H () H () for every r>. In fact, for ω> satisfying ω y 2 H y 2,wehave H ω y γ 2 H + γ λ γ 2 L 2 a(y γ,y γ )+ γ λ γ 2 L 2 and hence μ y γ H y H + f H ( y γ H + y H )+ γ λ L 2 λ γ L 2 ω 4 y γ 2 H + μ2 ω y 2 H + ω 2 y γ 2 H + 2ω f 2 H + 2γ λ γ 2 L 2 + 2γ λ 2 L 2 + f H y H, ω 4 y γ 2 H + 2γ λ γ 2 L μ2 2 ω y 2 H + 2ω f H + f H y H + 2γ λ 2 L. 2 This estimate implies that {y γ : γ r} is bounded in H () for every r>. The first equation of (OC γ ) implies that {λ γ : γ r} is bounded in H () as well. From Proposition 2. we have that lim γ (y γ,λ γ )=(y,λ )inh () H () weak. If λ =, then from (4), (6), and (OC γ ) the path C o is bounded in H () H () and λ γ inl 2 () for γ +. From (OC γ ) and the optimality condition for (ˆP) we have a(y γ ŷ, y γ ŷ)+(λ γ,y γ ŷ) =, and hence lim γ + y γ =ŷ in H (). Proposition 3.3. The path C r is globally Lipschitz in H () H () for every r>. If λ =, then C is globally Lipschitz continuous. Proof. Let γ, γ [r, ) be arbitrary. Then A(y γ y γ )+( λ + γ(y γ ψ)) + ( λ + γ(y γ ψ)) + =. Taking the inner-product with y γ y γ and using the monotonicity and Lipschitz continuity (with constant L =)ofx max(,x), we find a(y γ y γ, y γ y γ ) ( ( λ + γ(y γ ψ)) + ( λ + γ(y γ ψ)) +,y γ y γ ) γ γ y γ ψ L 2 y γ y γ L 2. By Lemma 3.2 the set {y γ } γ r is bounded in H (). Hence there exists K > such that ν y γ y γ 2 H K γ γ y γ y γ L 2, and by Poincaré s inequality there exists K 2 > such that y γ y γ H K 2 γ γ for all γ r, γ r.

8 66 MICHAEL HINTERMÜLLER AND KARL KUNISCH Let us recall here that y H = y L 2. Lipschitz continuity of γ λ γ from [r, ) to H () follows from the first equation in (OC γ ). For λ = the set {y γ } γ is bounded in H (). The remainder of the proof remains identical. Lemma 3.4. For every subset I [r, ), r>, the mapping γ λ γ is globally Lipschitz from I to L 2 (). Proof. For <γ γ 2 we have by (OC γ ) λ γ λ γ2 L 2 = ( λ + γ (y γ ψ)) + ( λ + γ 2 (y γ2 ψ)) + L 2 for some constant K 3 >. We shall use the following notation: Further we set (5) (K 3 γ + K + ψ L 2) γ γ 2 S γ = {x : λ(x)+γ(y γ ψ)(x) > }. g(γ) = λ + γ(y γ ψ). Since γ y γ H () is Lipschitz continuous by Proposition 3.3, there exists a weak accumulation point ẏ(= ẏ γ )of γ γ (y γ y γ )as γ γ>, which is also a strong accumulation point in L 2 (). Further γ γ (g( γ) g(γ)) has ġ(γ): = y γ ψ + γ ẏ γ as a strong accumulation point in L 2 () as γ γ. In case γ approaches γ from above (or below), the associated accumulation points ẏγ r (or ẏγ) l satisfy certain properties which are described next. In what follows we use ġ r (γ) orġ l (γ) whenever ẏ γ in ġ(γ) is replaced by ẏγ r and ẏγ, l respectively. Proposition 3.5. Let γ>, and denote by ẏγ r any weak accumulation point of γ γ (y γ y γ ) in H () as γ γ. Set Then ẏ r γ satisfies (6) (7) S + γ = S γ {x: λ(x)+γ(y γ (x) ψ(x))= ġ r (γ)(x) }. a(ẏ r γ,v)+((y γ ψ + γẏ r γ)χ S + γ,v)= for all v H (). Proof. By (OC γ ) we have for every v H () a(y γ y γ,v)+(( λ + γ(y γ ψ)) + ( λ + γ(y γ ψ)) +,v)=. We multiply (7) by ( γ γ) and discuss separately the two terms in (7). Clearly, we have lim( γ γ) a(y γ y γ,v)=a(ẏ r γ γ γ,v). Here and below the limit is taken on the sequence of γ-values, which provides the accumulation point. Lebesgue s bounded convergence theorem allows us to consider the pointwise limits of the integrands. Considering separately the cases g(γ)(x)<, g(γ)(x) >, and g(γ)(x) =,wehave (8) ( γ γ) ((g( γ)) + (g(γ)) +,v) ((y γ ψ + γ ẏ r γ)χ S + γ,v) as γ γ,

9 PATH-FOLLOWING METHODS 67 which ends the proof. As a consequence of the proof we obtain the following result. Corollary 3.6. Let γ>, and denote by ẏγ l any weak accumulation point of γ γ (y γ y γ ) in H () as γ γ. Set Sγ = S γ {x: λ(x) +γ(y γ (x) ψ(x)) = ġ l (γ)(x) }. Then ẏγ l satisfies (9) a(ẏ l γ,v)+((y γ ψ + γ ẏ l γ)χ S γ,v) = for all v H (). Another corollary of Proposition 3.5 treats the case λ =. Corollary 3.7. Let λ =, and assume that () holds. Then the right- and left- derivatives ẏ r γ and ẏ l γ of γ y γ, γ (, ), exist and are given by (2) (2) a(ẏγ,v)+((y r γ ψ + γ ẏγ)χ r {yγ>ψ},v) = a(ẏγ,v)+((y l γ ψ + γ ẏγ)χ l {yγ ψ},v) = for all v H (), for all v H (). Proof. Let γ γ. By Proposition 2.2 any accumulation point ẏ r γ of ( γ γ) (y γ y γ ) satisfies ẏ r γ and hence Observe that S + γ = {x : y γ (x) >ψ(x)} {x : y γ (x) =ψ(x) ẏ r γ(x) =}. (y γ ψ + γẏ r γ)χ S + γ =(y γ ψ + γẏ r γ)χ {yγ>ψ}. This implies that every accumulation point ẏγ r satisfies (2). Since the solution to (2) is unique, the directional derivative from the right exists. Similarly, if γ γ, by Proposition 2.2 we have Sγ = {x : y γ (x) ψ(x)}, and (2) follows. Henceforth we set S γ = {x : λ(x)+γ(y γ ψ)(x) =}. Corollary 3.8. If meas(s γ)=, then γ y γ H () is differentiable at γ, and the derivative ẏ γ satisfies (22) a(ẏ γ,v)+((y γ ψ + γ ẏ γ )χ Sγ,v)= forall v H (). Proof. Let z denote the difference of two accumulation points of ( γ γ) (y γ y γ ) as γ γ. As a consequence of (6) and (9) a(z,v)+γ(zχ Sγ,v) = for all v H (). This implies that z = by (6). Consequently, accumulation points are unique, and by (6), (9) they satisfy (22). The assumption meas(sγ) o = in Corollary 3.8 reflects the lack of differentiability of the max-operation in (OC γ ). 4. The primal-dual path value functional. In this section we investigate the value function associated with (P γ ) and study its monotonicity and smoothness properties. Definition 4.. The functional γ V (γ) =J(y γ ; γ) = 2 a(y γ,y γ ) (f,y γ )+ 2γ ( λ + γ(y γ ψ)) + 2 L 2

10 68 MICHAEL HINTERMÜLLER AND KARL KUNISCH defined on (, ) is called the primal-dual path value functional. Let us start by studying first order differentiability properties of V. Proposition 4.2. The value function V is differentiable with V (γ) = 2γ 2 ( λ + γ(y γ ψ)) γ ( λ + γ(y γ ψ)) + (y γ ψ). Corollary 4.3. For λ =we have V (γ) = 2 (y γ ψ) + 2 and V (γ) > unless y γ is feasible. For λ satisfying (2) and with () holding, we have y γ ψ and hence V (γ) for γ (, ). In either of the two cases V (γ) =implies that y γ solves (ˆP). Proof. We show only that V (γ) = implies that y γ solves (ˆP). The rest of the assertion follows immediately from Proposition 4.2. If λ =, then V (γ) = yields y γ ψ. Thus, λ γ =, and hence y γ solves (ˆP). If () and (2) are satisfied, then y γ ψ and V (γ) = implies γ(y γ ψ) λ + γ(y γ ψ). As a consequence λ γ =, and y γ solves (ˆP). Proof of Proposition 4.2. For γ,γ (, ) wefind (23) and consequently 2 a(y γ + y γ,y γ y γ ) (f,y γ y γ ) + 2 (( λ + γ(y γ ψ)) + +( λ + γ(y γ ψ)) +,y γ y γ )=, V ( γ) V (γ) = 2 a(y γ,y γ ) 2 a(y γ,y γ ) (f,y γ y γ ) + ( λ + γ(y γ ψ)) + 2 ( λ + γ(y γ ψ)) γ 2γ = ( λ + γ(y γ ψ)) ( λ + γ(y γ ψ)) γ 2γ + (( λ + γ(y γ ψ)) + +( λ + γ(y γ ψ)) + )(y γ y γ ) 2 = z + z + z = I + I 2 + I 3, P γ P γ P γ N γ P γ N γ where z stands for the sum of the kernels on the left of the above equalities, P γ = {x: λ + γ(y γ ψ) > }, N γ = {x: λ + γ(y γ ψ) < }, and P γ, N γ are defined analogously. For I 2 we have I 2 2 P γ N γ γ ( λ + γ(y γ ψ)) 2 + λ + γ(y γ ψ) y γ y γ 2 γ ( γ(y γ ψ) γ(y γ ψ)) 2 + y γ y γ ( γy γ γy γ + γ γ ψ ), and hence by Proposition 3.3 (24) lim γ γ γ γ I 2 =.

11 Analogously one verifies that PATH-FOLLOWING METHODS 69 (25) On P γ P γ we have z = 2 γ ( λ + γ(y γ ψ)) 2 lim γ γ γ γ I 3 =. 2γ ( λ + γ(y γ ψ)) 2 2 (2 λ + γ(y γ ψ)+γ(y γ ψ))(y γ y γ ) = γ γ 2 γγ ( λ + γ(y γ ψ)) 2 + 2γ [ 2 λ( γ(y γ ψ) γ(y γ ψ)) + γ 2 (y γ ψ) 2 γ 2 (y γ ψ) 2] 2 (2 λ + γ(y γ ψ)+γ(y γ ψ))(y γ y γ ) = γ γ 2 γγ ( λ + γ(y γ ψ)) 2 + λ γ [ γ(y γ ψ) γ(y γ ψ)] + [ ] γ 2 2 γ (y γ ψ) 2 γ(y γ ψ) 2 +( γ γ)(y γ ψ)(y γ ψ), and thus on P γ P γ ( γ γ) z = 2 γγ ( λ + γ(y γ ψ)) 2 + λ γ (y γ ψ) + [ ] γ 2 γ (y γ ψ) 2 +(y γ ψ)(y γ ψ). By Lebesgue s bounded convergence theorem, lim γ γ γ γ I = lim zχ γ γ γ γ P γ Pγ = 2γ 2 + γ(y γ ψ)) (( λ + ) 2 + ( λ + γ(y γ ψ)) + (y γ ψ). γ Together with (24) and (25), this implies the claim. Remark 4.. Note that V is characterized without recourse to ẏ γ. The boundedness of {γ 2 V (γ)}γ is established next. In what follows we use (v) = min(,v). Proposition 4.4. If λ = and a(v +,v ) = for all v H (), then {γ 2 V (γ)}γ is bounded. If () and (2) hold, then again {γ 2 V (γ)}γ is bounded. Proof. In the case λ = we have a(y γ ψ, v)+γ((y γ ψ) +,v)=(f,v) a(ψ, v) for all v H (). Since (y γ ψ) H () and a((y γ ψ) +, (y γ ψ) ) = we have, using (6) with v =(y γ ψ) +, ν (y γ ψ) + 2 H () + γ (y γ ψ) + 2 L 2 f L 2 (y γ ψ) + H + μ ψ H y γ ψ H.

12 7 MICHAEL HINTERMÜLLER AND KARL KUNISCH This implies the existence of a constant K, depending on ψ H and f L 2 but independent of γ, such that γ (y γ ψ) + L 2 K. Since V (γ) = 2 (y γ ψ) + 2 the claim follows. Turning to the feasible case with () and (2) holding, we have that y γ ψ for every γ>, and hence ( λ + γ(y γ ψ))(x) > if and only if λ(x) >γ(ψ y γ )(x). Consequently, V (γ) 2γ 2 ( λ + γ(y γ ψ)) ( λ + γ(y γ ψ)) + (ψ y γ ) γ 3 2γ 2 λ 2 L 2, which again implies the claim. Before we investigate V, we state a result which connects γ V (γ), y y γ H, and V V (γ), where V = lim γ V (γ). It will be used in section 6. for designing a γ-update strategy. Proposition 4.5. In the feasible and infeasible cases the following estimate holds true: y y γ 2 H 2 ( V V (γ) γ ν V ) (γ). Proof. We have V V (γ) =J(y ) J(y γ ; γ) and J(y ) J(y γ ; γ) ν 2 y y γ 2 H + a(y γ,y y γ ) (f,y y γ ) 2γ ( λ + γ(y γ ψ)) + 2 L 2, where we have used (6). From (OC γ )wehave and hence a(y γ,y y γ ) (f,y y γ )= (( λ + γ(y γ ψ)) +,y y γ ), J(y ) J(y γ ; γ) ν 2 y y γ 2 H (( λ + γ(y γ ψ)) +,y y γ ) 2γ ( λ + γ(y γ ψ)) + 2 L 2 ν 2 y y γ 2 H 2γ ( λ + γ(y γ ψ)) + 2 L 2 +(( λ + γ(y γ ψ)) +,y γ ψ) = ν 2 y y γ 2 H + γ V (γ). This completes the proof. Below we shall assume that y γ ψ C( ). Recall that for dimension n 3 and with (6) and (8) holding, we have y γ H 2 () C( ). Proposition 4.6. Let ẏ γ denote any accumulation point of ( γ γ) (y γ y γ ) as γ γ. (a) If λ =, y γ ψ C( ), and (8) is satisfied, then γ V (γ) is twice differentiable at γ with (26) V (γ) = (y γ ψ) + ẏ γ.

13 PATH-FOLLOWING METHODS 7 (b) For arbitrary λ, ifmeas(sγ)=, then γ V (γ) is twice differentiable at γ with V (γ) = γ 3 ( λ + γ(y γ ψ)) (27) γ 2 ( λ + γ(y γ ψ)) + (y γ ψ) + (y γ ψ)(y γ ψ + γ ẏ γ )χ Sγ. γ Proof. (a) On the subsequence γ n realizing the accumulation point, we have that lim n (γ n γ) ( V (γ n ) V (γ)) equals the right-hand side of (26). The claim will be established by verifying that the accumulation points ẏ γ restricted to S γ = {x: y γ (x) ψ(x) > } are unique. Let z denote the difference of two accumulation points. By (6) and (9) we have a(z,v)+γ(z,v) = for all v H () with v = on \ S γ. Using (8) and the fact that S γ is an open set relative to due to the continuity of y γ ψ, we find that z =ins γ, as desired. (b) Let ẏ γ denote any accumulation point of ( γ γ) (y γ y γ )as γ γ, and recall the notation g(γ) = λ + γ(y γ ψ) and S γ + from section 3. On the subsequence realizing the accumulation point we find lim γ γ γ γ ( V ( γ) V (γ)) = γ 3 ( λ + γ(y γ ψ)) (28) γ 2 ( λ + γ(y γ ψ)) + (y γ ψ) + (y γ ψ)(y γ ψ + γ ẏ γ )χ γ S + γ. By assumption, meas(s γ) = and, hence the right-hand sides of (27) and (28) coincide. Since ẏ γ is unique by Corollary 3.8 the claim is established. 5. Model functions. In this section we derive low-parameter families of functions which approximate the value functional V and share some of its qualitative properties. We will make use of these models in the numerics section when devising path-following algorithms. 5.. Infeasible case. Throughout this subsection we assume (8) and (29) λ =,y γ ψ C( ) for all γ (, ). Observe that (8), together with the general assumption (6), implies (). In fact, for any v H () we have a(v, v + ) γ v + 2, and hence a(v, v + ) implies v + =. Proposition 5.. The value function V satisfies V (γ) and V (γ) for γ (, ). Proof. Proposition 4.2 implies that V (γ). Moreover, y γ2 y γ for γ 2 γ > and hence ẏ γ a.e. on S γ. Consequently V (γ) by Proposition 4.6. A model function m for the value function V should reflect the sign properties of V and its derivatives. Moreover, V () gives the value of (ˆP), and hence we shall require that m() = V (). Finally from Lemma 3.2 we conclude that V is bounded on [, ). All these properties are satisfied by functions of the form (3) m(γ) =C C 2 E + γ

14 72 MICHAEL HINTERMÜLLER AND KARL KUNISCH with C R. Here C 2, E> satisfy (3) m() = V () = C C 2 E. Other choices for model functions are also conceivable, for example, γ C C (E+γ) r with r>. Note, however, that the asymptotic behavior of the model in (3) is such that γ 2 ṁ(γ) is bounded for γ. This is consistent with the boundedness of γ 2 V (γ) for γ asserted in Proposition 4.4. Another reason for choosing (3) is illustrated next. Choosing v =(y γ ψ) + in (OC γ ), we find (32) a(ẏ γ, (y γ ψ) + )+ (y γ ψ) + 2 L + γ (y 2 γ ψ) + ẏ γ =. For the following discussion we (33) replace a(, ) bye(, ) with E> a constant, and V by m. By Proposition 4.2 and (26) the following ordinary differential equation is obtained for m: (34) (E + γ) m(γ)+ 2 ṁ(γ) =. The solutions to (34) are given by (3). To get an account for the quality of our model in (3) we refer to the left-hand plot of Figure 4 in section Feasible case. Throughout this subsection we assume (35) (), λ satisfies (2), and meas (S γ ) = for all γ (, ). Proposition 5.2. The value function V satisfies V (γ) and V (γ) for γ (, ). Proof. By Proposition 2.2 we have y γ ψ and hence V (γ) by Proposition 4.2. A short computation based on (27) shows that (36) V (γ) = γ 3 χ λ 2 + χ(y γ ψ)ẏ γ γ χ(y γ ψ) 2 + χ (y γ ψ)ẏ γ, where χ is the characteristic function of the set S γ = { λ + γ(y γ ψ) > }. From (22) we have γ ẏ γ L 2 (S γ) ψ y γ L 2 (S γ), and hence V (γ). An immediate consequence is stated next. Lemma 5.3. If the solution to the unconstrained problem is not feasible, then lim γ V (γ) =. Proof. Assume that lim γ V (γ) is finite. Then, using (P γ ), there exists a sequence γ n and ỹ H () such that y γn ỹ weakly in H (), with y γn the solution to (P γn ), and λ γn = max(, λ + γ n (y n ψ)) inl 2 (). Consequently ỹ ψ. Taking the limit with respect to n in (OC γn ), it follows that ỹ ψ is the solution to (ˆP), which contradicts our assumption.

15 PATH-FOLLOWING METHODS 73 From Lemmas 3.2 and 5.3 and Proposition 5.2 it follows that γ V (γ), γ (, ), is a monotonically strictly decreasing convex function with lim γ + V (γ) =. All these properties are also satisfied by functions of the form (37) m(γ) =C C 2 E + γ + B γ, provided that C R, C 2, E>, B>, and C 2 B. We now give the motivation for choosing the model function m for V as in (37). From (22) with v =(y γ ψ)χ we get a(ẏ γ, (y ψ)χ)+γ(ẏ γ χ, y γ ψ)+((y γ ψ)χ, y γ ψ) =, where χ = χ Sγ. As in the infeasible case we replace a(, )bye(, ), with E a constant, and using (22), we arrive at The choice v = y γ ψ implies (38) (E + γ)(ẏ γ χ, v)+((y γ ψ)χ, v) =. (E + γ)(ẏ γ χ, y γ ψ)+((y γ ψ)χ, y γ ψ) =. Note that V (γ) can be expressed as V (γ) = λ 2 2γ 2 χ + (39) (y γ ψ) 2 χ. 2 Using (36) and (39) in (38), and replacing V by m, due to the substitution for a(, ), we find (E + γ) m +2ṁ Eγ 3 χ λ 2 =. We further replace χ λ 2, which is a bounded quantity depending on γ, by2b, and obtain, as the ordinary differential equation that we propose for the model function m in the feasible case, (4) (E + γ) m +2ṁ 2γ 3 EB=. The family of solutions is given by (37). In the right-hand plot of Figure 4 in section 6 we depict the approximation quality of m(γ). 6. Path-following algorithms. In this section we study the basic Algorithm B together with a variety of adjustment schemes for the path parameter γ. For this purpose recall that, depending on the shift parameter λ, the elements y γ along the primal-dual path are feasible or infeasible. As we have seen in the previous section, this implies different models for approximating the value function V. We will see, however, that for γ>in both cases similar strategies for updating γ may be used. When referring to the infeasible or feasible case, (29), respectively (35), is assumed to hold. The subsequent discussion is based on the following two-dimensional test problems. We point out that the bound ψ in problem P below does not satisfy ψ H (). However, as we shall see, the feasible and infeasible primal-dual path as well as the algorithms introduced subsequently still perform satisfactorily. We include this example since discontinuous obstacles are of practical relevance.

16 74 MICHAEL HINTERMÜLLER AND KARL KUNISCH Test problem P. We consider (8) with a ij = δ ij, with δ ij the Kronecker symbol, d =,and=(, ) 2. We choose f(x,x 2 ) = 5x sin(5x ) cos(x 2 ) and ψ on \ K, and ψ onk with K = {x : 5 x ( 2, 2 ) }. The solution y, the obstacle ψ, and the active set A at the solution are shown in Figure. Fig.. Optimal solution y (upper left plot), obstacle ψ (upper right plot), and the active set A (lower plot) for test problem P. Test problem P2. Again we consider (8), with a ij, d, and as before, and define x on T := {x :x 2 x x 2 x }, y x := 2 on T 2 := {x :x 2 x x 2 x }, (4) x on T 3 := {x :x 2 x x 2 x }, x 2 on T 4 := {x :x 2 x x 2 x }. The obstacle ψ is defined by ψ y on S := {x : x ( 2, 2 ) 4 }, ψ 4 on S 2 \ S, and 2x on T ( \ S 2 ), 4 ψ := 2(x ) on T 2 ( \ S 2 ), 4 2(x 7 8 ) on T 3 ( \ S 2 ), 2x 2 on T 4 ( \ S 2 ), with S 2 := {x : x ( 2, 2 ) 3 8 }. The forcing term is given by (f,φ) L 2 = φ(s)ds +(χ S,φ) L 2 + φ(s)ds for all φ H (), + S +

17 PATH-FOLLOWING METHODS 75 where + := {x :x 2 = x } {x :x 2 = x }. We recall that for φ H (), R 2, the traces along smooth curves are well defined. The solution y is given by y = y. The active or coincidence set at the solution is A = S. The Lagrange multiplier λ = f +Δy is in H () and enjoys no extra regularity. In Figure 2 we display the optimal solution y, the obstacle ψ, and the active set A. Fig. 2. Optimal solution y (upper left plot), obstacle ψ (upper right plot), and the active set A (lower plot) for test problem P2. Test problem P3. For this test problem (8) is satisfied. We therefore obtain y H 2 () and λ L 2 (). The coefficients a ij and d as well as are as before. The volume force f is given by f = Δv with v(x,x 2 ) = sin(3πx ) sin(3πx 2 ). Further, we have ψ = 4 sin(πx ) sin(πx 2 ). The optimal solution y, the Lagrange multiplier λ, and the active set at y are displayed in Figure 3. Unless specified otherwise, the subsequent algorithms are initialized by y = ( Δ) f, where Δ denotes the Laplacian with homogeneous Dirichlet boundary conditions. The initial Lagrange multiplier is chosen as λ = γ χ {y>ψ}(y ψ). The discretization of Δ is based on the classical five-point finite difference stencil. We denote the mesh size by h, which we occasionally drop for convenience. The forcing term f in P2 is discretized by f = Δy + χ S e + χ S ( Δy ), where e is the vector of all ones and χ S represents a diagonal matrix with entry (χ S ) ii = for grid points x i S and (χ S ) ii = otherwise. Above y denotes the grid function corresponding to (4). 6.. A strategy based on model functions exact path-following. As outlined in section 5, there are good reasons to trust our model functions (3) and (37) in the infeasible and feasible cases, respectively. Let us start by focusing on the infeasible case. The model is given by m(γ) =C C 2 (E + γ). For determining

18 76 MICHAEL HINTERMÜLLER AND KARL KUNISCH Active set P3 Fig. 3. Optimal solution y (upper left plot), obstacle ψ (upper right plot), and the active set A (lower plot) for test problem P3. the three parameters C,C 2, and E, we use the information V (),V(γ), V (γ), which, by Proposition 4.2, is available from one solve of the unconstrained problem (ˆP) and one solve for (Pγ). The conditions (42) yield (43) m() = V (), m(γ) =V (γ), ṁ(γ) = V (γ) E = γ 2 V (γ) (V (γ) V () γ V (γ)), C 2 = γ E(E + γ)(v (γ) V ()), C = V () + C 2 E. We could have used an alternative reference value γ r (,γ) and computed m(γ r )= V (γ r ) instead of m() = V (). In Figure 4 we compare V (γ) tom(γ) for different values of the coefficients (C,C 2,E). These coefficients depend on different values y f for γ (in (42)) produced by Algorithm EP (see below) for problem P. The solid line corresponds to V (γ). The corresponding γ-values γ f for (42) are depicted in the legend of the left plot in Figure 4. The dotted and dashed line belong to rather small γ-values, and the dashed-dotted and the circled lines to large γ f in (42). As we can see, the dotted line is accurate in the range of relatively small γ f, while the other lines are more accurate for large γ f. From now on we consider only the choices γ r = and γ = γ k in (42) when updating γ k. Next we discuss properties of the model parameters E, C, C 2 according to (43). For this purpose assume that the solution ŷ to (ˆP) is not feasible for (P). Then

19 PATH-FOLLOWING METHODS 77 m(γ), V(γ) Model m(γ) vs V(γ) (solid); infeasible case V(γ) γ f =.35E3 γ f =7.37E5 γ f =7.55E8 γ f =.93E γ m(γ) V(γ) m(γ) V(γ) ; feasible case γ γ f = γ f =.56E2 γ f =4.E3 γ f =.9E5 γ f =.52E7 Fig. 4. Left: Model m(γ) vs. V (γ) (solid) in the infeasible case for P. Right: Model m(γ) vs. V (γ) in the feasible case. by Corollary 4.3 we have V (γ) > for all γ>. Consequently V (γ) >V() and V (γ) V () γ V (γ) = γ γ V (σ)dσds >, and hence E>and C s 2 > for all γ (, + ). This implies m(γ) C and m(γ) C for γ +. We propose the following update strategy for γ: Let {τ k } satisfy τ k (, ) for all k N and τ k ask, and assume that V (γ k ) is available. Then, given γ k, the updated value γ k+ should ideally satisfy (44) V V (γ k+ ) τ k V V (γ k ). Since V and V (γ k+ ) are unknown, we use C,k and our model m k (γ) =C,k C 2,k /(E k + γ) atγ = γ k+ instead. Thus, (44) is replaced by (45) C,k m k (γ k+ ) τ k C,k V (γ k ) =: β k. Solving the equation C,k m k (γ k+ )=β k, we obtain γ k+ = C 2,k (46) E k. β k In Theorem 6. we shall show that γ k+ κγ k, with κ>, independently of k N. Before we turn to the feasible case, we interpret (44) in view of Proposition 4.5 in the infeasible case. Recall that V V (γ), and observe that V V (γ) = O( y y γ H ). Proposition 4.5 yields y y γ 2 H 2 ν (V V (γ)) since V (γ) >. Setting τ k = ω 2 k V V (γ k ), with ω k, in (44) yields Consequently, we obtain y y γk+ 2 H C τ ωk y 2 y γk 2 H. y y γk+ H y y γk H C τ ω k,

20 78 MICHAEL HINTERMÜLLER AND KARL KUNISCH which implies q-superlinear convergence of {y γk } in H (). In the feasible case, i.e., when λ satisfies (2), we use the model m(γ) =C C 2 (E + γ) + Bγ with C 2 and E,B > ; see (37). Let γ r >, γ r γ, denote a reference γ-value; then we use the conditions m(γ r )=V (γ r ), ṁ(γ r )= V (γ r ), m(γ) =V (γ), ṁ(γ) = V (γ) for fixing B, C, C2, E. Solving the corresponding system of nonlinear equations, we get (( (γ r γ)( V (γ r )γr 2 + V )) (γ)γ 2 )+2γ r γ(v (γ) V (γ r ) E = ( ( V (γ)γ + V ) (γ r )γ r )(γ γ r )+(γ r + γ)(v (γ r ) V (γ)) and B = ( γr 2 γ 2 (V (γ) V (γ r )) 2 V (γ) V ) (γ r )(γ γ r ) 2 ( (γ γ r ) 2 ( V (γ r )γr 2 + V ) (γ)γ 2 )+2(γ γ r )γ r γ(v (γ r ) V (γ)) Then the parameters C and C 2 are given by ( ) B C 2 =(E + γ) 2 γ 2 + V (γ), C = V (γ)+ C 2 E + γ B γ. In the right plot of Figure 4 we show m(γ) V (γ) with m(γ) produced by the iterates of Algorithm EP for P similar to the infeasible case. Again we can see that our model yields a close approximation of the value function V. If we require that (45) be satisfied in the feasible case, then we obtain the following update strategy for γ: (47) γ k+ = D k 2 + D 2 k 4 + B ke k β k, where D k = E k +(C 2,k B k )/β k. In Theorem 6. we shall establish γ k+ κγ k for all k N with κ> independent of k. Next we describe an exact path-following version of Algorithm B, which utilizes the update strategy (45) for updating γ. Algorithm EP. (i) Select γ r. Compute V (γ r ), and choose γ > max(,γ r ); set k =. (ii) Apply Algorithm B to obtain y γk. (iii) Compute V (γ k ), V (γk ), and γ k+ according to (46) in the infeasible case or (47) in the feasible case. (iv) Set k = k +, and go to (ii). Concerning the choice of γ r note that in the infeasible case we have γ r, and in the feasible case γ r >. Convergence of Algorithm EP is addressed next. Theorem 6.. Assume that the solution to (ˆP) is not feasible for (P). Then the iterates γ k of Algorithm EP tend to as k, and consequently lim k (y γk,λ γk )= (y,λ ) in H () H () weak.

21 PATH-FOLLOWING METHODS 79 Proof. Let us consider the infeasible case. Then (45) is equivalent to (48) <C,k m k (γ k+ ) <τ k (C,k m k (γ k )). Since γ m k (γ) is strictly increasing and τ k (, ), it follows that γ k+ >γ k for every k =,,... If lim k γ k =, then lim k (y γk,λ γk )=(y,λ ). Otherwise there exists γ such that lim k γ k = γ. Since γ V (γ) and γ V (γ) are continuous on (, ), it follows from (42) and (43) that lim k E k = E( γ), lim k C,k = C ( γ), and lim k C 2,k = C 2 ( γ), where E( γ), C ( γ), C 2 ( γ) are given by (43) with γ replaced by γ. Taking the limit with respect to k in (48), we arrive at C 2 ( γ) E( γ)+ γ =, which is impossible, since C 2 ( γ) > and E( γ) > if the solution to (ˆP) is not feasible for (P). Thus lim k γ k =. The feasible case is treated analogously. Numerically we stop the algorithm as soon as (r,h k,r2,h,r3,h) 2 ɛ M, where r,h k r 2,h k = yγ h k +( Δ h ) (λ h γ k f h ) H,h/ f h H,h, = λ h γ k max(,λ h γ k + yγ h k ψ h ) H,h, r 3,h k = max(,yγ h k ψ h ) L h 2, and ɛ M denotes the machine accuracy. Here H,h denotes the discrete version of H. For some vector v it is realized as v H = h ( Δ h ) v L h 2 with L h 2 the discrete L 2 -norm and h a forward difference approximation of the gradient operator; see [8]. The inner iteration, i.e., Algorithm B for γ = γ k, is terminated if successive active sets coincide or Δ h yγ h,l k + λ h,l γ k f h H,h f h ɛ M. H,h Here the superscript l = l(k) denotes the iteration index of Algorithm B for fixed k. For a discussion and numerical results in the case where the approximation errors due to the discretization of the underlying function space problems are incorporated into the algorithmic framework, e.g., when stopping the algorithm, we refer to the next section 6.2. The initialization of γ is as follows: In the infeasible case we propose a choice of γ based on the deviation of the linearization of V (γ) atγ = γ r from the objective value of the unconstrained problem (ˆP) at the projection of y γr onto the feasible set. In our realization of this heuristic we choose γ r = and compute ŷ, V (), and V (). Then we set { γ = max,ζ J(y } b) V () (49), V () where ζ (, ] is some fixed constant, y b (x) = min(ŷ, ψ(x)), and J denotes the objective function of (ˆP). Note that ŷ is the minimizer of the unconstrained problem (ˆP). For the examples below we use ζ =. In the feasible case we choose a reference value γ r, e.g., γ r =, and solve the path problem (P γ ). Then we choose γ = γ r + J(ŷ) V (γ r) (5), V (γ r ) k k

22 8 MICHAEL HINTERMÜLLER AND KARL KUNISCH where ŷ denotes the minimizer of the discretized unconstrained problem (ˆP). If ŷ is not feasible for (P), then one has J(ŷ) <V(γ r ) and hence γ >γ r. When applied to P, P2, and P3 for h =/28 and with τ k =. k+, we obtain the results shown in Figure 5 and Table 6.. # inner iterations inner iterations P outer iteration inner iterations inner iterations P outer iteration # inner iterations inner iterations P outer iteration Fig. 5. Number of inner iterations (vertical axis) per outer iteration for P (left plot), P2 (middle plot), and P3 (right plot); solid line infeasible case, dashed line feasible case. Table 6. Comparison of iteration counts. P P2 P3 Version # outer # inner # outer # inner # outer # inner Feasible Infeasible From our test runs, also for other test problems, we observe the following characteristics: For the feasible version the number of inner iterations exhibits an increasing tendency until a saturation value is reached, and then, unless the algorithm stops at an approximate solution, it starts to decrease. For the infeasible version we typically observe that the first couple of iterations require several inner iterations. As the outer iterations proceed the number of inner iterations drops eventually to one. We also tested less aggressive γ-updates compared to the ones used here, e.g., updates based on γ k+ = ξγ k with ξ> fixed. The numerically observable convergence speed of y γk towards y in H () is typically superlinear. This can be seen from Figure 6, where the plots for the discrete versions qk h of the quotients q k = y γ k+ y H y γk y H are shown. Note that the vertical axis uses a logarithmic scale. In the first row, for P we depict the behavior of qk h for h =2 i, i =5, 6, 7, 8, for the infeasible case (left plot) and the feasible case (right plot). We observe that the convergence rate is stable with respect to decreasing mesh size h. In the second row we see the behavior of qk h for P2 and P3, with h =2 7. Again, we observe a superlinear rate of convergence. With respect to decreasing h the same conclusion as for P holds true. These stability results provide a link between our function space theory and the numerical realization of the algorithms.

23 PATH-FOLLOWING METHODS 8 Convergence behavior of q k h ; P outer iteration feasible infeasible Convergence behavior of q k h ; P3 infeasible feasible 2 3 outer iteration Fig. 6. Discrete quotients qk h for P and various mesh sizes h (upper row) and for P2 (lower left) and P3 (lower right) for h =/28. number of grid points Difference of active sets outer iteration Fig. 7. Difference in active sets for P and P2; solid line infeasible case, dashed line feasible case. In connection with the convergence speed it is of interest how the detection process of the correct active set works. For the rather aggressive γ-updates used in Algorithm EP the difference between two successive active sets is zero typically only in the last iteration. However, if a less aggressive strategy for updating γ is used, then it is to be expected, that the difference of active sets might become zero earlier along the iteration. In Figure 7, for the strategy γ k+ =2γ k, we show the difference of successive active sets; i.e., the vertical axis relates to the number of grid points that are in A k+ but not in A k and vice versa. We detect that for the infeasible case there exists an iteration index k after which the difference is constantly zero. This behavior is a strong indication that the correct active set was detected. It suggests that we fix this set A k and set ȳ A k = ψ A k, I k =\A k, and λ I k =. Then one computes ȳ I k and λ A k such that a(ȳ, v)+ λ, v H,H =(f,v) for all v H (), and checks whether (ȳ, λ) satisfies (7). If this is the case, then the solution is

24 82 MICHAEL HINTERMÜLLER AND KARL KUNISCH found; otherwise γ k is updated and the iteration continued. If we apply this technique for P in the infeasible case, then the algorithm stops at iteration 5 (35 inner iterations) with the exact discrete solution, as compared to 28 outer and 47 inner iterations without the additional stopping rule. There were four iterations where the additional system solve was necessary but without obtaining the numerical solution. Hence, with respect to system solves, the amount of work drops from 47 solves to 39 (= ). A similar observation is true for P2 and P3. In the feasible case, however, this strategy yields no reduction of iterations. Here, typically the correct active set is determined in the last iteration (for large enough γ). The dependence of the iteration number on the mesh size of the discretization for P is depicted in Table 6.2 (those for P2 and P3 are similar). In parenthesis we show the number of inner iterations. The results clearly indicate that the outer iterations are mesh independent, while the number of inner iterations increases as the mesh size decreases. In the third row we display the results obtained by applying Algorithm A for the solution of the unregularized problem (P) with data according to P. If we compare these results with those of the infeasible exact path-following algorithm, we find that for sufficiently small mesh sizes h the infeasible version of Algorithm EP requires significantly fewer iterations than does Algorithm A, which is also an infeasible algorithm. Also, the number of iterations required by Algorithm A exhibits a relatively strong dependence on h when compared to Algorithm EP in the infeasible case. Similar observations apply also to P2 and P3. This shows that taking into account the function space theoretic properties when regularizing problem (P) results in an algorithmic framework which performs stably with respect to decreasing mesh size of the discretization. 6 γ values; P outer iteration Table 6.2 Comparison of iteration counts for different mesh sizes. Mesh size h Version /6 /32 /64 /28 /256 EP feasible 5(9) 5(23) 5(3) 5(44) 5(72) EP infeasible 4(8) 4() 4(3) 4(5) 4(9) Algorithm A inner iteration γ values; P outer iteration 2 γ values; P outer iteration Fig. 8. γ-updates; solid line infeasible case, dashed line feasible case. From the plots in Figure 8, where the y-axis again has a logarithmic scale, it can be seen that our strategy (45) produces a rapidly increasing sequence

25 PATH-FOLLOWING METHODS 83 {γ k }. The plots in Figure 8 depict the increase of γ k as a function of the iteration number. The question arises of whether one could increase γ more rapidly. Numerical examples employing an ad hoc strategy show that if γ is increased too quickly, then the numerical error may prevent the residuals of the first order system from dropping below ɛ M. This effect is due to the ill-conditioning of the linear systems for large γ. On the other hand, small increases in γ result in a slow convergence speed of Algorithm EP. Further, in our test runs and as can be seen from Figure 8, the feasible version of Algorithm EP is less aggressive in enlarging γ k Inexact path-following. While exact path-following is primarily of theoretical interest, the development of inexact path-following techniques that keep the number of iterations as small as possible is of more practical importance. The strategy in the previous section relies on the fact that for every γ k the corresponding point on the primal-dual path is computed. This, however, is not the case for inexact techniques and, as a consequence, a different update strategy for the path parameter γ is necessary. A common concept in inexact path-following methods is based on the definition of an appropriate neighborhood of the path; see, e.g., [3] and the references therein for a noninterior neighborhood-based path-following method, or [5, 6, 8, 9] for path-following techniques related to interior point methods. It is typically required that the primal-dual iterates stay within the neighborhood of the path, with the goal to reduce the computational burden while still maintaining convergence of the method. We define (5a) (5b) rγ(y, λ) = Δy + λ f H, rγ(y, 2 λ) = λ max(,λ+ γ(y ψ)) H, and the neighborhood { N (γ) := (y, λ) H () L 2 () : (rγ(y, λ),rγ(y, 2 λ)) 2 τ } (52) γ in the infeasible case and { N (γ) := (y, λ) H () L 2 () : (rγ(y, λ),rγ(y, 2 λ)) 2 (53) } J(y; γ) γ in the feasible case. Above, τ>denotes some fixed parameter. Note that adding the condition γj(y; γ) in (52) yields no further restriction, since this condition is automatically satisfied by the structure of J(y; γ). We also point out that the conditions on the derivative of J(y; γ) are included in (52) and (53), respectively, in order to qualitatively capture (up to first order) the analytical properties of the primal-dual path. Next we specify our framework for an inexact path-following algorithm. Algorithm IP. (i) Initialize γ according to (49) in the infeasible case or (5) in the feasible case; set k :=. τ γ

c 2004 Society for Industrial and Applied Mathematics

c 2004 Society for Industrial and Applied Mathematics SIAM J. OPTIM. Vol. 15, No. 1, pp. 39 62 c 2004 Society for Industrial and Applied Mathematics SEMISMOOTH NEWTON AND AUGMENTED LAGRANGIAN METHODS FOR A SIMPLIFIED FRICTION PROBLEM GEORG STADLER Abstract.

More information

SEVERAL PATH-FOLLOWING METHODS FOR A CLASS OF GRADIENT CONSTRAINED VARIATIONAL INEQUALITIES

SEVERAL PATH-FOLLOWING METHODS FOR A CLASS OF GRADIENT CONSTRAINED VARIATIONAL INEQUALITIES SEVERAL PATH-FOLLOWING METHODS FOR A CLASS OF GRADIENT CONSTRAINED VARIATIONAL INEQUALITIES M. HINTERMÜLLER AND J. RASCH Abstract. Path-following splitting and semismooth Newton methods for solving a class

More information

RESEARCH ARTICLE. A strategy of finding an initial active set for inequality constrained quadratic programming problems

RESEARCH ARTICLE. A strategy of finding an initial active set for inequality constrained quadratic programming problems Optimization Methods and Software Vol. 00, No. 00, July 200, 8 RESEARCH ARTICLE A strategy of finding an initial active set for inequality constrained quadratic programming problems Jungho Lee Computer

More information

Affine covariant Semi-smooth Newton in function space

Affine covariant Semi-smooth Newton in function space Affine covariant Semi-smooth Newton in function space Anton Schiela March 14, 2018 These are lecture notes of my talks given for the Winter School Modern Methods in Nonsmooth Optimization that was held

More information

A full-newton step infeasible interior-point algorithm for linear programming based on a kernel function

A full-newton step infeasible interior-point algorithm for linear programming based on a kernel function A full-newton step infeasible interior-point algorithm for linear programming based on a kernel function Zhongyi Liu, Wenyu Sun Abstract This paper proposes an infeasible interior-point algorithm with

More information

SUPERCONVERGENCE PROPERTIES FOR OPTIMAL CONTROL PROBLEMS DISCRETIZED BY PIECEWISE LINEAR AND DISCONTINUOUS FUNCTIONS

SUPERCONVERGENCE PROPERTIES FOR OPTIMAL CONTROL PROBLEMS DISCRETIZED BY PIECEWISE LINEAR AND DISCONTINUOUS FUNCTIONS SUPERCONVERGENCE PROPERTIES FOR OPTIMAL CONTROL PROBLEMS DISCRETIZED BY PIECEWISE LINEAR AND DISCONTINUOUS FUNCTIONS A. RÖSCH AND R. SIMON Abstract. An optimal control problem for an elliptic equation

More information

Primal/Dual Decomposition Methods

Primal/Dual Decomposition Methods Primal/Dual Decomposition Methods Daniel P. Palomar Hong Kong University of Science and Technology (HKUST) ELEC5470 - Convex Optimization Fall 2018-19, HKUST, Hong Kong Outline of Lecture Subgradients

More information

Robust error estimates for regularization and discretization of bang-bang control problems

Robust error estimates for regularization and discretization of bang-bang control problems Robust error estimates for regularization and discretization of bang-bang control problems Daniel Wachsmuth September 2, 205 Abstract We investigate the simultaneous regularization and discretization of

More information

1.The anisotropic plate model

1.The anisotropic plate model Pré-Publicações do Departamento de Matemática Universidade de Coimbra Preprint Number 6 OPTIMAL CONTROL OF PIEZOELECTRIC ANISOTROPIC PLATES ISABEL NARRA FIGUEIREDO AND GEORG STADLER ABSTRACT: This paper

More information

Path-following and augmented Lagrangian methods for contact problems in linear elasticity

Path-following and augmented Lagrangian methods for contact problems in linear elasticity Journal of Computational and Applied Mathematics 203 (2007) 533 547 www.elsevier.com/locate/cam Path-following and augmented Lagrangian methods for contact problems in linear elasticity Georg Stadler Center

More information

A Second Full-Newton Step O(n) Infeasible Interior-Point Algorithm for Linear Optimization

A Second Full-Newton Step O(n) Infeasible Interior-Point Algorithm for Linear Optimization A Second Full-Newton Step On Infeasible Interior-Point Algorithm for Linear Optimization H. Mansouri C. Roos August 1, 005 July 1, 005 Department of Electrical Engineering, Mathematics and Computer Science,

More information

On the Local Quadratic Convergence of the Primal-Dual Augmented Lagrangian Method

On the Local Quadratic Convergence of the Primal-Dual Augmented Lagrangian Method Optimization Methods and Software Vol. 00, No. 00, Month 200x, 1 11 On the Local Quadratic Convergence of the Primal-Dual Augmented Lagrangian Method ROMAN A. POLYAK Department of SEOR and Mathematical

More information

A Unified Analysis of Nonconvex Optimization Duality and Penalty Methods with General Augmenting Functions

A Unified Analysis of Nonconvex Optimization Duality and Penalty Methods with General Augmenting Functions A Unified Analysis of Nonconvex Optimization Duality and Penalty Methods with General Augmenting Functions Angelia Nedić and Asuman Ozdaglar April 16, 2006 Abstract In this paper, we study a unifying framework

More information

An interior-point gradient method for large-scale totally nonnegative least squares problems

An interior-point gradient method for large-scale totally nonnegative least squares problems An interior-point gradient method for large-scale totally nonnegative least squares problems Michael Merritt and Yin Zhang Technical Report TR04-08 Department of Computational and Applied Mathematics Rice

More information

Lecture 1: Entropy, convexity, and matrix scaling CSE 599S: Entropy optimality, Winter 2016 Instructor: James R. Lee Last updated: January 24, 2016

Lecture 1: Entropy, convexity, and matrix scaling CSE 599S: Entropy optimality, Winter 2016 Instructor: James R. Lee Last updated: January 24, 2016 Lecture 1: Entropy, convexity, and matrix scaling CSE 599S: Entropy optimality, Winter 2016 Instructor: James R. Lee Last updated: January 24, 2016 1 Entropy Since this course is about entropy maximization,

More information

An introduction to Mathematical Theory of Control

An introduction to Mathematical Theory of Control An introduction to Mathematical Theory of Control Vasile Staicu University of Aveiro UNICA, May 2018 Vasile Staicu (University of Aveiro) An introduction to Mathematical Theory of Control UNICA, May 2018

More information

LECTURE 1: SOURCES OF ERRORS MATHEMATICAL TOOLS A PRIORI ERROR ESTIMATES. Sergey Korotov,

LECTURE 1: SOURCES OF ERRORS MATHEMATICAL TOOLS A PRIORI ERROR ESTIMATES. Sergey Korotov, LECTURE 1: SOURCES OF ERRORS MATHEMATICAL TOOLS A PRIORI ERROR ESTIMATES Sergey Korotov, Institute of Mathematics Helsinki University of Technology, Finland Academy of Finland 1 Main Problem in Mathematical

More information

Research Note. A New Infeasible Interior-Point Algorithm with Full Nesterov-Todd Step for Semi-Definite Optimization

Research Note. A New Infeasible Interior-Point Algorithm with Full Nesterov-Todd Step for Semi-Definite Optimization Iranian Journal of Operations Research Vol. 4, No. 1, 2013, pp. 88-107 Research Note A New Infeasible Interior-Point Algorithm with Full Nesterov-Todd Step for Semi-Definite Optimization B. Kheirfam We

More information

An Infeasible Interior-Point Algorithm with full-newton Step for Linear Optimization

An Infeasible Interior-Point Algorithm with full-newton Step for Linear Optimization An Infeasible Interior-Point Algorithm with full-newton Step for Linear Optimization H. Mansouri M. Zangiabadi Y. Bai C. Roos Department of Mathematical Science, Shahrekord University, P.O. Box 115, Shahrekord,

More information

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44 Convex Optimization Newton s method ENSAE: Optimisation 1/44 Unconstrained minimization minimize f(x) f convex, twice continuously differentiable (hence dom f open) we assume optimal value p = inf x f(x)

More information

Hamburger Beiträge zur Angewandten Mathematik

Hamburger Beiträge zur Angewandten Mathematik Hamburger Beiträge zur Angewandten Mathematik Numerical analysis of a control and state constrained elliptic control problem with piecewise constant control approximations Klaus Deckelnick and Michael

More information

Semismooth Newton Methods for an Optimal Boundary Control Problem of Wave Equations

Semismooth Newton Methods for an Optimal Boundary Control Problem of Wave Equations Semismooth Newton Methods for an Optimal Boundary Control Problem of Wave Equations Axel Kröner 1 Karl Kunisch 2 and Boris Vexler 3 1 Lehrstuhl für Mathematische Optimierung Technische Universität München

More information

CO 250 Final Exam Guide

CO 250 Final Exam Guide Spring 2017 CO 250 Final Exam Guide TABLE OF CONTENTS richardwu.ca CO 250 Final Exam Guide Introduction to Optimization Kanstantsin Pashkovich Spring 2017 University of Waterloo Last Revision: March 4,

More information

An Alternative Proof of Primitivity of Indecomposable Nonnegative Matrices with a Positive Trace

An Alternative Proof of Primitivity of Indecomposable Nonnegative Matrices with a Positive Trace An Alternative Proof of Primitivity of Indecomposable Nonnegative Matrices with a Positive Trace Takao Fujimoto Abstract. This research memorandum is aimed at presenting an alternative proof to a well

More information

REGULAR LAGRANGE MULTIPLIERS FOR CONTROL PROBLEMS WITH MIXED POINTWISE CONTROL-STATE CONSTRAINTS

REGULAR LAGRANGE MULTIPLIERS FOR CONTROL PROBLEMS WITH MIXED POINTWISE CONTROL-STATE CONSTRAINTS REGULAR LAGRANGE MULTIPLIERS FOR CONTROL PROBLEMS WITH MIXED POINTWISE CONTROL-STATE CONSTRAINTS fredi tröltzsch 1 Abstract. A class of quadratic optimization problems in Hilbert spaces is considered,

More information

Functional Analysis. Franck Sueur Metric spaces Definitions Completeness Compactness Separability...

Functional Analysis. Franck Sueur Metric spaces Definitions Completeness Compactness Separability... Functional Analysis Franck Sueur 2018-2019 Contents 1 Metric spaces 1 1.1 Definitions........................................ 1 1.2 Completeness...................................... 3 1.3 Compactness......................................

More information

K. Krumbiegel I. Neitzel A. Rösch

K. Krumbiegel I. Neitzel A. Rösch SUFFICIENT OPTIMALITY CONDITIONS FOR THE MOREAU-YOSIDA-TYPE REGULARIZATION CONCEPT APPLIED TO SEMILINEAR ELLIPTIC OPTIMAL CONTROL PROBLEMS WITH POINTWISE STATE CONSTRAINTS K. Krumbiegel I. Neitzel A. Rösch

More information

min f(x). (2.1) Objectives consisting of a smooth convex term plus a nonconvex regularization term;

min f(x). (2.1) Objectives consisting of a smooth convex term plus a nonconvex regularization term; Chapter 2 Gradient Methods The gradient method forms the foundation of all of the schemes studied in this book. We will provide several complementary perspectives on this algorithm that highlight the many

More information

Kaisa Joki Adil M. Bagirov Napsu Karmitsa Marko M. Mäkelä. New Proximal Bundle Method for Nonsmooth DC Optimization

Kaisa Joki Adil M. Bagirov Napsu Karmitsa Marko M. Mäkelä. New Proximal Bundle Method for Nonsmooth DC Optimization Kaisa Joki Adil M. Bagirov Napsu Karmitsa Marko M. Mäkelä New Proximal Bundle Method for Nonsmooth DC Optimization TUCS Technical Report No 1130, February 2015 New Proximal Bundle Method for Nonsmooth

More information

A Distributed Newton Method for Network Utility Maximization, II: Convergence

A Distributed Newton Method for Network Utility Maximization, II: Convergence A Distributed Newton Method for Network Utility Maximization, II: Convergence Ermin Wei, Asuman Ozdaglar, and Ali Jadbabaie October 31, 2012 Abstract The existing distributed algorithms for Network Utility

More information

A Posteriori Estimates for Cost Functionals of Optimal Control Problems

A Posteriori Estimates for Cost Functionals of Optimal Control Problems A Posteriori Estimates for Cost Functionals of Optimal Control Problems Alexandra Gaevskaya, Ronald H.W. Hoppe,2 and Sergey Repin 3 Institute of Mathematics, Universität Augsburg, D-8659 Augsburg, Germany

More information

Primal-Dual Interior-Point Methods for Linear Programming based on Newton s Method

Primal-Dual Interior-Point Methods for Linear Programming based on Newton s Method Primal-Dual Interior-Point Methods for Linear Programming based on Newton s Method Robert M. Freund March, 2004 2004 Massachusetts Institute of Technology. The Problem The logarithmic barrier approach

More information

Barrier Method. Javier Peña Convex Optimization /36-725

Barrier Method. Javier Peña Convex Optimization /36-725 Barrier Method Javier Peña Convex Optimization 10-725/36-725 1 Last time: Newton s method For root-finding F (x) = 0 x + = x F (x) 1 F (x) For optimization x f(x) x + = x 2 f(x) 1 f(x) Assume f strongly

More information

A duality-based approach to elliptic control problems in non-reflexive Banach spaces

A duality-based approach to elliptic control problems in non-reflexive Banach spaces A duality-based approach to elliptic control problems in non-reflexive Banach spaces Christian Clason Karl Kunisch June 3, 29 Convex duality is a powerful framework for solving non-smooth optimal control

More information

12. Interior-point methods

12. Interior-point methods 12. Interior-point methods Convex Optimization Boyd & Vandenberghe inequality constrained minimization logarithmic barrier function and central path barrier method feasibility and phase I methods complexity

More information

5 Handling Constraints

5 Handling Constraints 5 Handling Constraints Engineering design optimization problems are very rarely unconstrained. Moreover, the constraints that appear in these problems are typically nonlinear. This motivates our interest

More information

Linear Programming Redux

Linear Programming Redux Linear Programming Redux Jim Bremer May 12, 2008 The purpose of these notes is to review the basics of linear programming and the simplex method in a clear, concise, and comprehensive way. The book contains

More information

Simple Examples on Rectangular Domains

Simple Examples on Rectangular Domains 84 Chapter 5 Simple Examples on Rectangular Domains In this chapter we consider simple elliptic boundary value problems in rectangular domains in R 2 or R 3 ; our prototype example is the Poisson equation

More information

Nonlinear Optimization for Optimal Control

Nonlinear Optimization for Optimal Control Nonlinear Optimization for Optimal Control Pieter Abbeel UC Berkeley EECS Many slides and figures adapted from Stephen Boyd [optional] Boyd and Vandenberghe, Convex Optimization, Chapters 9 11 [optional]

More information

Lecture notes for Analysis of Algorithms : Markov decision processes

Lecture notes for Analysis of Algorithms : Markov decision processes Lecture notes for Analysis of Algorithms : Markov decision processes Lecturer: Thomas Dueholm Hansen June 6, 013 Abstract We give an introduction to infinite-horizon Markov decision processes (MDPs) with

More information

Convex Optimization Notes

Convex Optimization Notes Convex Optimization Notes Jonathan Siegel January 2017 1 Convex Analysis This section is devoted to the study of convex functions f : B R {+ } and convex sets U B, for B a Banach space. The case of B =

More information

September Math Course: First Order Derivative

September Math Course: First Order Derivative September Math Course: First Order Derivative Arina Nikandrova Functions Function y = f (x), where x is either be a scalar or a vector of several variables (x,..., x n ), can be thought of as a rule which

More information

Spring 2017 CO 250 Course Notes TABLE OF CONTENTS. richardwu.ca. CO 250 Course Notes. Introduction to Optimization

Spring 2017 CO 250 Course Notes TABLE OF CONTENTS. richardwu.ca. CO 250 Course Notes. Introduction to Optimization Spring 2017 CO 250 Course Notes TABLE OF CONTENTS richardwu.ca CO 250 Course Notes Introduction to Optimization Kanstantsin Pashkovich Spring 2017 University of Waterloo Last Revision: March 4, 2018 Table

More information

Algorithms for Constrained Optimization

Algorithms for Constrained Optimization 1 / 42 Algorithms for Constrained Optimization ME598/494 Lecture Max Yi Ren Department of Mechanical Engineering, Arizona State University April 19, 2015 2 / 42 Outline 1. Convergence 2. Sequential quadratic

More information

Constrained Optimization and Lagrangian Duality

Constrained Optimization and Lagrangian Duality CIS 520: Machine Learning Oct 02, 2017 Constrained Optimization and Lagrangian Duality Lecturer: Shivani Agarwal Disclaimer: These notes are designed to be a supplement to the lecture. They may or may

More information

AN AUGMENTED LAGRANGIAN AFFINE SCALING METHOD FOR NONLINEAR PROGRAMMING

AN AUGMENTED LAGRANGIAN AFFINE SCALING METHOD FOR NONLINEAR PROGRAMMING AN AUGMENTED LAGRANGIAN AFFINE SCALING METHOD FOR NONLINEAR PROGRAMMING XIAO WANG AND HONGCHAO ZHANG Abstract. In this paper, we propose an Augmented Lagrangian Affine Scaling (ALAS) algorithm for general

More information

Structural and Multidisciplinary Optimization. P. Duysinx and P. Tossings

Structural and Multidisciplinary Optimization. P. Duysinx and P. Tossings Structural and Multidisciplinary Optimization P. Duysinx and P. Tossings 2018-2019 CONTACTS Pierre Duysinx Institut de Mécanique et du Génie Civil (B52/3) Phone number: 04/366.91.94 Email: P.Duysinx@uliege.be

More information

Optimality, Duality, Complementarity for Constrained Optimization

Optimality, Duality, Complementarity for Constrained Optimization Optimality, Duality, Complementarity for Constrained Optimization Stephen Wright University of Wisconsin-Madison May 2014 Wright (UW-Madison) Optimality, Duality, Complementarity May 2014 1 / 41 Linear

More information

Key words. preconditioned conjugate gradient method, saddle point problems, optimal control of PDEs, control and state constraints, multigrid method

Key words. preconditioned conjugate gradient method, saddle point problems, optimal control of PDEs, control and state constraints, multigrid method PRECONDITIONED CONJUGATE GRADIENT METHOD FOR OPTIMAL CONTROL PROBLEMS WITH CONTROL AND STATE CONSTRAINTS ROLAND HERZOG AND EKKEHARD SACHS Abstract. Optimality systems and their linearizations arising in

More information

Unconstrained minimization of smooth functions

Unconstrained minimization of smooth functions Unconstrained minimization of smooth functions We want to solve min x R N f(x), where f is convex. In this section, we will assume that f is differentiable (so its gradient exists at every point), and

More information

1 Lyapunov theory of stability

1 Lyapunov theory of stability M.Kawski, APM 581 Diff Equns Intro to Lyapunov theory. November 15, 29 1 1 Lyapunov theory of stability Introduction. Lyapunov s second (or direct) method provides tools for studying (asymptotic) stability

More information

SPECTRAL PROPERTIES OF THE LAPLACIAN ON BOUNDED DOMAINS

SPECTRAL PROPERTIES OF THE LAPLACIAN ON BOUNDED DOMAINS SPECTRAL PROPERTIES OF THE LAPLACIAN ON BOUNDED DOMAINS TSOGTGEREL GANTUMUR Abstract. After establishing discrete spectra for a large class of elliptic operators, we present some fundamental spectral properties

More information

A PREDICTOR-CORRECTOR PATH-FOLLOWING ALGORITHM FOR SYMMETRIC OPTIMIZATION BASED ON DARVAY'S TECHNIQUE

A PREDICTOR-CORRECTOR PATH-FOLLOWING ALGORITHM FOR SYMMETRIC OPTIMIZATION BASED ON DARVAY'S TECHNIQUE Yugoslav Journal of Operations Research 24 (2014) Number 1, 35-51 DOI: 10.2298/YJOR120904016K A PREDICTOR-CORRECTOR PATH-FOLLOWING ALGORITHM FOR SYMMETRIC OPTIMIZATION BASED ON DARVAY'S TECHNIQUE BEHROUZ

More information

Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games

Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games Alberto Bressan ) and Khai T. Nguyen ) *) Department of Mathematics, Penn State University **) Department of Mathematics,

More information

x 2 x n r n J(x + t(x x ))(x x )dt. For warming-up we start with methods for solving a single equation of one variable.

x 2 x n r n J(x + t(x x ))(x x )dt. For warming-up we start with methods for solving a single equation of one variable. Maria Cameron 1. Fixed point methods for solving nonlinear equations We address the problem of solving an equation of the form (1) r(x) = 0, where F (x) : R n R n is a vector-function. Eq. (1) can be written

More information

The optimal partial transport problem

The optimal partial transport problem The optimal partial transport problem Alessio Figalli Abstract Given two densities f and g, we consider the problem of transporting a fraction m [0, min{ f L 1, g L 1}] of the mass of f onto g minimizing

More information

An Inexact Newton Method for Nonlinear Constrained Optimization

An Inexact Newton Method for Nonlinear Constrained Optimization An Inexact Newton Method for Nonlinear Constrained Optimization Frank E. Curtis Numerical Analysis Seminar, January 23, 2009 Outline Motivation and background Algorithm development and theoretical results

More information

An optimal control problem for a parabolic PDE with control constraints

An optimal control problem for a parabolic PDE with control constraints An optimal control problem for a parabolic PDE with control constraints PhD Summer School on Reduced Basis Methods, Ulm Martin Gubisch University of Konstanz October 7 Martin Gubisch (University of Konstanz)

More information

Constrained Minimization and Multigrid

Constrained Minimization and Multigrid Constrained Minimization and Multigrid C. Gräser (FU Berlin), R. Kornhuber (FU Berlin), and O. Sander (FU Berlin) Workshop on PDE Constrained Optimization Hamburg, March 27-29, 2008 Matheon Outline Successive

More information

Newton s Method. Javier Peña Convex Optimization /36-725

Newton s Method. Javier Peña Convex Optimization /36-725 Newton s Method Javier Peña Convex Optimization 10-725/36-725 1 Last time: dual correspondences Given a function f : R n R, we define its conjugate f : R n R, f ( (y) = max y T x f(x) ) x Properties and

More information

Adaptive methods for control problems with finite-dimensional control space

Adaptive methods for control problems with finite-dimensional control space Adaptive methods for control problems with finite-dimensional control space Saheed Akindeinde and Daniel Wachsmuth Johann Radon Institute for Computational and Applied Mathematics (RICAM) Austrian Academy

More information

An Introduction to Variational Inequalities

An Introduction to Variational Inequalities An Introduction to Variational Inequalities Stefan Rosenberger Supervisor: Prof. DI. Dr. techn. Karl Kunisch Institut für Mathematik und wissenschaftliches Rechnen Universität Graz January 26, 2012 Stefan

More information

3.10 Lagrangian relaxation

3.10 Lagrangian relaxation 3.10 Lagrangian relaxation Consider a generic ILP problem min {c t x : Ax b, Dx d, x Z n } with integer coefficients. Suppose Dx d are the complicating constraints. Often the linear relaxation and the

More information

A FULL-NEWTON STEP INFEASIBLE-INTERIOR-POINT ALGORITHM COMPLEMENTARITY PROBLEMS

A FULL-NEWTON STEP INFEASIBLE-INTERIOR-POINT ALGORITHM COMPLEMENTARITY PROBLEMS Yugoslav Journal of Operations Research 25 (205), Number, 57 72 DOI: 0.2298/YJOR3055034A A FULL-NEWTON STEP INFEASIBLE-INTERIOR-POINT ALGORITHM FOR P (κ)-horizontal LINEAR COMPLEMENTARITY PROBLEMS Soodabeh

More information

Conditional Gradient (Frank-Wolfe) Method

Conditional Gradient (Frank-Wolfe) Method Conditional Gradient (Frank-Wolfe) Method Lecturer: Aarti Singh Co-instructor: Pradeep Ravikumar Convex Optimization 10-725/36-725 1 Outline Today: Conditional gradient method Convergence analysis Properties

More information

Optimization methods

Optimization methods Lecture notes 3 February 8, 016 1 Introduction Optimization methods In these notes we provide an overview of a selection of optimization methods. We focus on methods which rely on first-order information,

More information

Stochastic Optimization with Inequality Constraints Using Simultaneous Perturbations and Penalty Functions

Stochastic Optimization with Inequality Constraints Using Simultaneous Perturbations and Penalty Functions International Journal of Control Vol. 00, No. 00, January 2007, 1 10 Stochastic Optimization with Inequality Constraints Using Simultaneous Perturbations and Penalty Functions I-JENG WANG and JAMES C.

More information

Self-Concordant Barrier Functions for Convex Optimization

Self-Concordant Barrier Functions for Convex Optimization Appendix F Self-Concordant Barrier Functions for Convex Optimization F.1 Introduction In this Appendix we present a framework for developing polynomial-time algorithms for the solution of convex optimization

More information

Linear algebra issues in Interior Point methods for bound-constrained least-squares problems

Linear algebra issues in Interior Point methods for bound-constrained least-squares problems Linear algebra issues in Interior Point methods for bound-constrained least-squares problems Stefania Bellavia Dipartimento di Energetica S. Stecco Università degli Studi di Firenze Joint work with Jacek

More information

A Geometric Framework for Nonconvex Optimization Duality using Augmented Lagrangian Functions

A Geometric Framework for Nonconvex Optimization Duality using Augmented Lagrangian Functions A Geometric Framework for Nonconvex Optimization Duality using Augmented Lagrangian Functions Angelia Nedić and Asuman Ozdaglar April 15, 2006 Abstract We provide a unifying geometric framework for the

More information

Finite Element Methods for Fourth Order Variational Inequalities

Finite Element Methods for Fourth Order Variational Inequalities Louisiana State University LSU Digital Commons LSU Doctoral Dissertations Graduate School 2013 Finite Element Methods for Fourth Order Variational Inequalities Yi Zhang Louisiana State University and Agricultural

More information

Inference For High Dimensional M-estimates. Fixed Design Results

Inference For High Dimensional M-estimates. Fixed Design Results : Fixed Design Results Lihua Lei Advisors: Peter J. Bickel, Michael I. Jordan joint work with Peter J. Bickel and Noureddine El Karoui Dec. 8, 2016 1/57 Table of Contents 1 Background 2 Main Results and

More information

Some Properties of the Augmented Lagrangian in Cone Constrained Optimization

Some Properties of the Augmented Lagrangian in Cone Constrained Optimization MATHEMATICS OF OPERATIONS RESEARCH Vol. 29, No. 3, August 2004, pp. 479 491 issn 0364-765X eissn 1526-5471 04 2903 0479 informs doi 10.1287/moor.1040.0103 2004 INFORMS Some Properties of the Augmented

More information

Analysis of an Adaptive Finite Element Method for Recovering the Robin Coefficient

Analysis of an Adaptive Finite Element Method for Recovering the Robin Coefficient Analysis of an Adaptive Finite Element Method for Recovering the Robin Coefficient Yifeng Xu 1 Jun Zou 2 Abstract Based on a new a posteriori error estimator, an adaptive finite element method is proposed

More information

Penalty and Barrier Methods. So we again build on our unconstrained algorithms, but in a different way.

Penalty and Barrier Methods. So we again build on our unconstrained algorithms, but in a different way. AMSC 607 / CMSC 878o Advanced Numerical Optimization Fall 2008 UNIT 3: Constrained Optimization PART 3: Penalty and Barrier Methods Dianne P. O Leary c 2008 Reference: N&S Chapter 16 Penalty and Barrier

More information

Differentiable exact penalty functions for nonlinear optimization with easy constraints. Takuma NISHIMURA

Differentiable exact penalty functions for nonlinear optimization with easy constraints. Takuma NISHIMURA Master s Thesis Differentiable exact penalty functions for nonlinear optimization with easy constraints Guidance Assistant Professor Ellen Hidemi FUKUDA Takuma NISHIMURA Department of Applied Mathematics

More information

CONSTRAINED PERCOLATION ON Z 2

CONSTRAINED PERCOLATION ON Z 2 CONSTRAINED PERCOLATION ON Z 2 ZHONGYANG LI Abstract. We study a constrained percolation process on Z 2, and prove the almost sure nonexistence of infinite clusters and contours for a large class of probability

More information

Lecture 8 Plus properties, merit functions and gap functions. September 28, 2008

Lecture 8 Plus properties, merit functions and gap functions. September 28, 2008 Lecture 8 Plus properties, merit functions and gap functions September 28, 2008 Outline Plus-properties and F-uniqueness Equation reformulations of VI/CPs Merit functions Gap merit functions FP-I book:

More information

Laplace s Equation. Chapter Mean Value Formulas

Laplace s Equation. Chapter Mean Value Formulas Chapter 1 Laplace s Equation Let be an open set in R n. A function u C 2 () is called harmonic in if it satisfies Laplace s equation n (1.1) u := D ii u = 0 in. i=1 A function u C 2 () is called subharmonic

More information

AM 205: lecture 19. Last time: Conditions for optimality, Newton s method for optimization Today: survey of optimization methods

AM 205: lecture 19. Last time: Conditions for optimality, Newton s method for optimization Today: survey of optimization methods AM 205: lecture 19 Last time: Conditions for optimality, Newton s method for optimization Today: survey of optimization methods Quasi-Newton Methods General form of quasi-newton methods: x k+1 = x k α

More information

Lecture 7: Semidefinite programming

Lecture 7: Semidefinite programming CS 766/QIC 820 Theory of Quantum Information (Fall 2011) Lecture 7: Semidefinite programming This lecture is on semidefinite programming, which is a powerful technique from both an analytic and computational

More information

UNIQUENESS OF POSITIVE SOLUTION TO SOME COUPLED COOPERATIVE VARIATIONAL ELLIPTIC SYSTEMS

UNIQUENESS OF POSITIVE SOLUTION TO SOME COUPLED COOPERATIVE VARIATIONAL ELLIPTIC SYSTEMS TRANSACTIONS OF THE AMERICAN MATHEMATICAL SOCIETY Volume 00, Number 0, Pages 000 000 S 0002-9947(XX)0000-0 UNIQUENESS OF POSITIVE SOLUTION TO SOME COUPLED COOPERATIVE VARIATIONAL ELLIPTIC SYSTEMS YULIAN

More information

Frequency functions, monotonicity formulas, and the thin obstacle problem

Frequency functions, monotonicity formulas, and the thin obstacle problem Frequency functions, monotonicity formulas, and the thin obstacle problem IMA - University of Minnesota March 4, 2013 Thank you for the invitation! In this talk we will present an overview of the parabolic

More information

6.252 NONLINEAR PROGRAMMING LECTURE 10 ALTERNATIVES TO GRADIENT PROJECTION LECTURE OUTLINE. Three Alternatives/Remedies for Gradient Projection

6.252 NONLINEAR PROGRAMMING LECTURE 10 ALTERNATIVES TO GRADIENT PROJECTION LECTURE OUTLINE. Three Alternatives/Remedies for Gradient Projection 6.252 NONLINEAR PROGRAMMING LECTURE 10 ALTERNATIVES TO GRADIENT PROJECTION LECTURE OUTLINE Three Alternatives/Remedies for Gradient Projection Two-Metric Projection Methods Manifold Suboptimization Methods

More information

Appendix A Taylor Approximations and Definite Matrices

Appendix A Taylor Approximations and Definite Matrices Appendix A Taylor Approximations and Definite Matrices Taylor approximations provide an easy way to approximate a function as a polynomial, using the derivatives of the function. We know, from elementary

More information

On some weighted fractional porous media equations

On some weighted fractional porous media equations On some weighted fractional porous media equations Gabriele Grillo Politecnico di Milano September 16 th, 2015 Anacapri Joint works with M. Muratori and F. Punzo Gabriele Grillo Weighted Fractional PME

More information

Lecture 5: Gradient Descent. 5.1 Unconstrained minimization problems and Gradient descent

Lecture 5: Gradient Descent. 5.1 Unconstrained minimization problems and Gradient descent 10-725/36-725: Convex Optimization Spring 2015 Lecturer: Ryan Tibshirani Lecture 5: Gradient Descent Scribes: Loc Do,2,3 Disclaimer: These notes have not been subjected to the usual scrutiny reserved for

More information

CONVERGENCE PROPERTIES OF COMBINED RELAXATION METHODS

CONVERGENCE PROPERTIES OF COMBINED RELAXATION METHODS CONVERGENCE PROPERTIES OF COMBINED RELAXATION METHODS Igor V. Konnov Department of Applied Mathematics, Kazan University Kazan 420008, Russia Preprint, March 2002 ISBN 951-42-6687-0 AMS classification:

More information

Newton s Method. Ryan Tibshirani Convex Optimization /36-725

Newton s Method. Ryan Tibshirani Convex Optimization /36-725 Newton s Method Ryan Tibshirani Convex Optimization 10-725/36-725 1 Last time: dual correspondences Given a function f : R n R, we define its conjugate f : R n R, Properties and examples: f (y) = max x

More information

Primal-dual relationship between Levenberg-Marquardt and central trajectories for linearly constrained convex optimization

Primal-dual relationship between Levenberg-Marquardt and central trajectories for linearly constrained convex optimization Primal-dual relationship between Levenberg-Marquardt and central trajectories for linearly constrained convex optimization Roger Behling a, Clovis Gonzaga b and Gabriel Haeser c March 21, 2013 a Department

More information

Numerical algorithms for one and two target optimal controls

Numerical algorithms for one and two target optimal controls Numerical algorithms for one and two target optimal controls Sung-Sik Kwon Department of Mathematics and Computer Science, North Carolina Central University 80 Fayetteville St. Durham, NC 27707 Email:

More information

SYMMETRY RESULTS FOR PERTURBED PROBLEMS AND RELATED QUESTIONS. Massimo Grosi Filomena Pacella S. L. Yadava. 1. Introduction

SYMMETRY RESULTS FOR PERTURBED PROBLEMS AND RELATED QUESTIONS. Massimo Grosi Filomena Pacella S. L. Yadava. 1. Introduction Topological Methods in Nonlinear Analysis Journal of the Juliusz Schauder Center Volume 21, 2003, 211 226 SYMMETRY RESULTS FOR PERTURBED PROBLEMS AND RELATED QUESTIONS Massimo Grosi Filomena Pacella S.

More information

PDE Constrained Optimization selected Proofs

PDE Constrained Optimization selected Proofs PDE Constrained Optimization selected Proofs Jeff Snider jeff@snider.com George Mason University Department of Mathematical Sciences April, 214 Outline 1 Prelim 2 Thms 3.9 3.11 3 Thm 3.12 4 Thm 3.13 5

More information

INVERSE FUNCTION THEOREM and SURFACES IN R n

INVERSE FUNCTION THEOREM and SURFACES IN R n INVERSE FUNCTION THEOREM and SURFACES IN R n Let f C k (U; R n ), with U R n open. Assume df(a) GL(R n ), where a U. The Inverse Function Theorem says there is an open neighborhood V U of a in R n so that

More information

Interior-Point Methods for Linear Optimization

Interior-Point Methods for Linear Optimization Interior-Point Methods for Linear Optimization Robert M. Freund and Jorge Vera March, 204 c 204 Robert M. Freund and Jorge Vera. All rights reserved. Linear Optimization with a Logarithmic Barrier Function

More information

OPTIMALITY CONDITIONS AND ERROR ANALYSIS OF SEMILINEAR ELLIPTIC CONTROL PROBLEMS WITH L 1 COST FUNCTIONAL

OPTIMALITY CONDITIONS AND ERROR ANALYSIS OF SEMILINEAR ELLIPTIC CONTROL PROBLEMS WITH L 1 COST FUNCTIONAL OPTIMALITY CONDITIONS AND ERROR ANALYSIS OF SEMILINEAR ELLIPTIC CONTROL PROBLEMS WITH L 1 COST FUNCTIONAL EDUARDO CASAS, ROLAND HERZOG, AND GERD WACHSMUTH Abstract. Semilinear elliptic optimal control

More information

Part 3: Trust-region methods for unconstrained optimization. Nick Gould (RAL)

Part 3: Trust-region methods for unconstrained optimization. Nick Gould (RAL) Part 3: Trust-region methods for unconstrained optimization Nick Gould (RAL) minimize x IR n f(x) MSc course on nonlinear optimization UNCONSTRAINED MINIMIZATION minimize x IR n f(x) where the objective

More information

Penalty and Barrier Methods General classical constrained minimization problem minimize f(x) subject to g(x) 0 h(x) =0 Penalty methods are motivated by the desire to use unconstrained optimization techniques

More information

We describe the generalization of Hazan s algorithm for symmetric programming

We describe the generalization of Hazan s algorithm for symmetric programming ON HAZAN S ALGORITHM FOR SYMMETRIC PROGRAMMING PROBLEMS L. FAYBUSOVICH Abstract. problems We describe the generalization of Hazan s algorithm for symmetric programming Key words. Symmetric programming,

More information

Optimization Tutorial 1. Basic Gradient Descent

Optimization Tutorial 1. Basic Gradient Descent E0 270 Machine Learning Jan 16, 2015 Optimization Tutorial 1 Basic Gradient Descent Lecture by Harikrishna Narasimhan Note: This tutorial shall assume background in elementary calculus and linear algebra.

More information