Downloaded 12/29/14 to Redistribution subject to SIAM license or copyright; see

Similar documents
c 2004 Society for Industrial and Applied Mathematics

SEVERAL PATH-FOLLOWING METHODS FOR A CLASS OF GRADIENT CONSTRAINED VARIATIONAL INEQUALITIES

RESEARCH ARTICLE. A strategy of finding an initial active set for inequality constrained quadratic programming problems

Affine covariant Semi-smooth Newton in function space

A full-newton step infeasible interior-point algorithm for linear programming based on a kernel function

SUPERCONVERGENCE PROPERTIES FOR OPTIMAL CONTROL PROBLEMS DISCRETIZED BY PIECEWISE LINEAR AND DISCONTINUOUS FUNCTIONS

Primal/Dual Decomposition Methods

Robust error estimates for regularization and discretization of bang-bang control problems

1.The anisotropic plate model

Path-following and augmented Lagrangian methods for contact problems in linear elasticity

A Second Full-Newton Step O(n) Infeasible Interior-Point Algorithm for Linear Optimization

On the Local Quadratic Convergence of the Primal-Dual Augmented Lagrangian Method

A Unified Analysis of Nonconvex Optimization Duality and Penalty Methods with General Augmenting Functions

An interior-point gradient method for large-scale totally nonnegative least squares problems

Lecture 1: Entropy, convexity, and matrix scaling CSE 599S: Entropy optimality, Winter 2016 Instructor: James R. Lee Last updated: January 24, 2016

An introduction to Mathematical Theory of Control

LECTURE 1: SOURCES OF ERRORS MATHEMATICAL TOOLS A PRIORI ERROR ESTIMATES. Sergey Korotov,

Research Note. A New Infeasible Interior-Point Algorithm with Full Nesterov-Todd Step for Semi-Definite Optimization

An Infeasible Interior-Point Algorithm with full-newton Step for Linear Optimization

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44

Hamburger Beiträge zur Angewandten Mathematik

Semismooth Newton Methods for an Optimal Boundary Control Problem of Wave Equations

CO 250 Final Exam Guide

An Alternative Proof of Primitivity of Indecomposable Nonnegative Matrices with a Positive Trace

REGULAR LAGRANGE MULTIPLIERS FOR CONTROL PROBLEMS WITH MIXED POINTWISE CONTROL-STATE CONSTRAINTS

Functional Analysis. Franck Sueur Metric spaces Definitions Completeness Compactness Separability...

K. Krumbiegel I. Neitzel A. Rösch

min f(x). (2.1) Objectives consisting of a smooth convex term plus a nonconvex regularization term;

Kaisa Joki Adil M. Bagirov Napsu Karmitsa Marko M. Mäkelä. New Proximal Bundle Method for Nonsmooth DC Optimization

A Distributed Newton Method for Network Utility Maximization, II: Convergence

A Posteriori Estimates for Cost Functionals of Optimal Control Problems

Primal-Dual Interior-Point Methods for Linear Programming based on Newton s Method

Barrier Method. Javier Peña Convex Optimization /36-725

A duality-based approach to elliptic control problems in non-reflexive Banach spaces

12. Interior-point methods

5 Handling Constraints

Linear Programming Redux

Simple Examples on Rectangular Domains

Nonlinear Optimization for Optimal Control

Lecture notes for Analysis of Algorithms : Markov decision processes

Convex Optimization Notes

September Math Course: First Order Derivative

Spring 2017 CO 250 Course Notes TABLE OF CONTENTS. richardwu.ca. CO 250 Course Notes. Introduction to Optimization

Algorithms for Constrained Optimization

Constrained Optimization and Lagrangian Duality

AN AUGMENTED LAGRANGIAN AFFINE SCALING METHOD FOR NONLINEAR PROGRAMMING

Structural and Multidisciplinary Optimization. P. Duysinx and P. Tossings

Optimality, Duality, Complementarity for Constrained Optimization

Key words. preconditioned conjugate gradient method, saddle point problems, optimal control of PDEs, control and state constraints, multigrid method

Unconstrained minimization of smooth functions

1 Lyapunov theory of stability

SPECTRAL PROPERTIES OF THE LAPLACIAN ON BOUNDED DOMAINS

A PREDICTOR-CORRECTOR PATH-FOLLOWING ALGORITHM FOR SYMMETRIC OPTIMIZATION BASED ON DARVAY'S TECHNIQUE

Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games

x 2 x n r n J(x + t(x x ))(x x )dt. For warming-up we start with methods for solving a single equation of one variable.

The optimal partial transport problem

An Inexact Newton Method for Nonlinear Constrained Optimization

An optimal control problem for a parabolic PDE with control constraints

Constrained Minimization and Multigrid

Newton s Method. Javier Peña Convex Optimization /36-725

Adaptive methods for control problems with finite-dimensional control space

An Introduction to Variational Inequalities

3.10 Lagrangian relaxation

A FULL-NEWTON STEP INFEASIBLE-INTERIOR-POINT ALGORITHM COMPLEMENTARITY PROBLEMS

Conditional Gradient (Frank-Wolfe) Method

Optimization methods

Stochastic Optimization with Inequality Constraints Using Simultaneous Perturbations and Penalty Functions

Self-Concordant Barrier Functions for Convex Optimization

Linear algebra issues in Interior Point methods for bound-constrained least-squares problems

A Geometric Framework for Nonconvex Optimization Duality using Augmented Lagrangian Functions

Finite Element Methods for Fourth Order Variational Inequalities

Inference For High Dimensional M-estimates. Fixed Design Results

Some Properties of the Augmented Lagrangian in Cone Constrained Optimization

Analysis of an Adaptive Finite Element Method for Recovering the Robin Coefficient

Penalty and Barrier Methods. So we again build on our unconstrained algorithms, but in a different way.

Differentiable exact penalty functions for nonlinear optimization with easy constraints. Takuma NISHIMURA

CONSTRAINED PERCOLATION ON Z 2

Lecture 8 Plus properties, merit functions and gap functions. September 28, 2008

Laplace s Equation. Chapter Mean Value Formulas

AM 205: lecture 19. Last time: Conditions for optimality, Newton s method for optimization Today: survey of optimization methods

Lecture 7: Semidefinite programming

UNIQUENESS OF POSITIVE SOLUTION TO SOME COUPLED COOPERATIVE VARIATIONAL ELLIPTIC SYSTEMS

Frequency functions, monotonicity formulas, and the thin obstacle problem

6.252 NONLINEAR PROGRAMMING LECTURE 10 ALTERNATIVES TO GRADIENT PROJECTION LECTURE OUTLINE. Three Alternatives/Remedies for Gradient Projection

Appendix A Taylor Approximations and Definite Matrices

On some weighted fractional porous media equations

Lecture 5: Gradient Descent. 5.1 Unconstrained minimization problems and Gradient descent

CONVERGENCE PROPERTIES OF COMBINED RELAXATION METHODS

Newton s Method. Ryan Tibshirani Convex Optimization /36-725

Primal-dual relationship between Levenberg-Marquardt and central trajectories for linearly constrained convex optimization

Numerical algorithms for one and two target optimal controls

SYMMETRY RESULTS FOR PERTURBED PROBLEMS AND RELATED QUESTIONS. Massimo Grosi Filomena Pacella S. L. Yadava. 1. Introduction

PDE Constrained Optimization selected Proofs

INVERSE FUNCTION THEOREM and SURFACES IN R n

Interior-Point Methods for Linear Optimization

OPTIMALITY CONDITIONS AND ERROR ANALYSIS OF SEMILINEAR ELLIPTIC CONTROL PROBLEMS WITH L 1 COST FUNCTIONAL

Part 3: Trust-region methods for unconstrained optimization. Nick Gould (RAL)


We describe the generalization of Hazan s algorithm for symmetric programming

Optimization Tutorial 1. Basic Gradient Descent

Transcription:

SIAM J. OPTIM. Vol. 7, No., pp. 59 87 c 26 Society for Industrial and Applied Mathematics PATH-FOLLOWING METHODS FOR A CLASS OF CONSTRAINED MINIMIZATION PROBLEMS IN FUNCTION SPACE MICHAEL HINTERMÜLLER AND KARL KUNISCH Abstract. Path-following methods for primal-dual active set strategies requiring a regularization parameter are introduced. Existence of a primal-dual path and its differentiability properties are analyzed. Monotonicity and convexity of the primal-dual path value function are investigated. Both feasible and infeasible approximations are considered. Numerical path-following strategies are developed and their efficiency is demonstrated by means of examples. Key words. semismooth Newton methods, path-following methods, active set strategy, primaldual methods AMS subject classifications. 49M5, 49M37, 65K5, 9C33 DOI..37/46598. Introduction. Primal-dual active set strategies or, in some cases equivalently, semismooth Newton methods, were proved to be efficient methods for solving constrained variational problems in function space [, 9,,, 2, 3]. In certain cases regularization is required, resulting in a family of approximating problems with more favorable properties than those of the original one, [2, 3]. In previous work [3] convergence, and in some cases rate of convergence, with respect to the regularization parameter was proved. In the numerical work the adaptation of these parameters was heuristic, however. The focus of the present investigation is on an efficient control of the regularization parameter in the primal-dual active set strategy for a class of constrained variational problems. To explain the involved issues we proceed mostly formally in this section and consider the problem { min J (v) over v X () s.t. Gv ψ, where J is a quadratic functional on a Hilbert space X, and G: X Y. It is assumed that Y L 2 () is a Hilbert lattice with ordering induced by the natural ordering of L 2 (). We note that () subsumes problems of very different nature. For example, for the control constrained optimal control problem min 2 y z 2 L + α 2 2 u 2 L 2 s.t. Δy = u in, y = on, u ψ a.e. in, with a bounded domain in R n, z L 2 (), α>, one can use y =( Δ) u, where Δ denotes the Laplacian with homogenous Dirichlet boundary conditions, and arrive Received by the editors July 4, 24; accepted for publication (in revised form) November 2, 25; published electronically May 3, 26. This research was partially supported by the Fonds zur Förderung der wissenschaftlichen Forschung under SFB 3 Optimierung und Kontrolle. http://www.siam.org/journals/siopt/7-/659.html Department of Computational and Applied Mathematics, Rice University, Houston, TX. Institute of Mathematics and Scientific Computing, University of Graz, Graz, Austria (michael. hintermueller@uni-graz.at, karl.kunisch@uni-graz.at). 59

6 MICHAEL HINTERMÜLLER AND KARL KUNISCH at { min 2 ( Δ) u z 2 + α 2 u 2 s.t. u ψ a.e. in, which is clearly of the form (). For J (v) = 2 v 2 dx fv, X = H (), and G = I we obtain the classical obstacle problem. For state constrained control problems with y ψ one has { min 2 ( Δ) u z 2 + α 2 u 2 s.t. ( Δ) u ψ a.e. in, which is also of the form (). From the point of view of duality theory these three problems are very different. While it is straightforward to argue the existence of a Lagrange multiplier in L 2 () for the control constrained optimal control problem, it is already more involved and requires additional assumptions to guarantee its existence in L 2 () for obstacle problems, and for state constrained problems the Lagrange multiplier is only a measure. If we resort to a formal discussion, then in either of these cases we arrive at the optimality system of the form { J (v)+g λ =, (2) λ = max(,λ+ c(g(v) ψ)) for any fixed c>. Here, G denotes the adjoint of G. The second equation in (2) is equivalent to λ, G(v) ψ, and λ(g(v) ψ) =. Continuing formally, the primal-dual active set strategy determines the active set at iteration level k by means of A k+ = {x : λ k (x)+c(g(v k )(x) ψ(x))> }, assigns the inactive set I k+ =\A k+, and updates (v, λ) by means of { J (v k+ )+G λ k+ =, (3) λ k+ = on I k+, (G(v k+ ) ψ)(x) = for x A k+. These auxiliary problems require special attention. For obstacle problems the constraint v k+ = ψ on A k+ induces that the associated Lagrange multiplier λ k+ is in general less regular than the Lagrange multiplier associated with v ψ for the original problem; see, e.g., [3]. For problems with combined control and state constraints it may happen that due to the assignment on I k+ and A k+, (3) has no solution while the original problem does. For these reasons in, e.g., [9, 2, 3] the second equation in (2) was regularized, resulting in the family of equations { J (v)+g λ =, (4) λ = max(, λ + γ(g(v) ψ)), where λ is fixed, possibly λ =, and γ R +. In the above-mentioned references it was shown that under appropriate conditions the solutions (v γ,λ γ ) to (4) exist, the quantity λ γ enjoys extra regularity, and (v γ,λ γ ) converge to the solution of (2) as γ +.

PATH-FOLLOWING METHODS 6 In previous numerical implementations the increase of γ to infinity was heuristic. As the system (4) becomes increasingly ill-conditioned as γ tends to, in this paper a framework for a properly controlled increase of γ-values will be developed in order to cope with the conditioning problem. In fact, in a typical algorithmic regime for solving () one uses the solution (v γ,λ γ ) to (4) for some γ as the initial guess for the solution to (4) for the updated γ-value γ + >γ. Typically, if γ + γ, then (v γ,λ γ )isonlyapoor approximation of (v γ +,λ γ +), which in addition to numerical linear algebra issues (like ill-conditioned system matrices) causes severe stability problems for iterative solvers for (4) such as semismooth Newton methods. Together with developing a new γ- update strategy, we aim at solving the auxiliary problems (4) only inexactly to keep the overall computational cost low. To this end we define neighborhoods of the path which allow inexact solutions and which contract in a controlled way towards the path as the iteration proceeds. Our work is inspired by concepts from path-following methods in finite dimensional spaces [4, 5, 6, 8, 9]. We first guarantee the existence of a sufficiently smooth path γ (v γ,λ γ ), with γ (, ) in appropriately chosen function spaces. Once the path is available it can be used as the basis for updating strategies of the path parameter. Given a current value γ k, with associated primal and dual states (v γk,λ γk ), the γ-update should be sufficiently large to make good progress towards satisfying the complementarity conditions. On the other hand, since we are not solving the problems along the path exactly, we have to use safeguards against steps which would lead us too far off the path. Of course, these goals are impeded by the fact that the path is not available numerically. To overcome this difficulty we use qualitative properties of the value function, like monotonicity and convexity, which can be verified analytically. These suggest the introduction of model functions which will be shown to approximate the value functional along the path very well. We use these model functions for our updating strategies of γ. In the case of exact path-following we can even prove convergence of the resulting strategy. In the present paper the program just described is carried out for a class of problems corresponding to contact problems. State constrained optimal control problems require a different approach that will be considered independently. As we shall see, the (infinite dimensional) parameter λ can be used to guarantee that the iterates of the primal variable are feasible. Further, it turns out that the numerical behavior of infeasible approximations is superior to the feasible ones from the point of view of iteration numbers. Interior point methods also require an additional parameter, which, however, enters into (2) differently. For the problem under consideration here, the interiorpoint relaxation replaces the second equation in (2) by (5) λ(x)(ψ G(v))(x) = for x. γ Path-following interior-point methods typically start strictly feasible, with iterates which are required to stay strictly feasible during the iterations while satisfying, or satisfying approximately, the first equation in (2) and (5). Path-following interiorpoint methods have not received much attention for infinite dimensional problems yet. In fact, we are aware of only [7], where such methods are analyzed for optimal control problems related to ordinary differential equations. For the problem classes that we outlined at the beginning of this section, the primal-dual active set strategy proved to be an excellent competitor to interior-point methods, as was demonstrated, for example, in [] comparing these two methods. This paper is organized as follows. Section 2 contains the precise problem formulation and the necessary background on the primal-dual active set strategy. The

62 MICHAEL HINTERMÜLLER AND KARL KUNISCH existence and regularity of the primal-dual path is discussed in section 3. Properties of the primal-dual path value functional are analyzed in section 4. Section 5 contains the derivation of the proposed model functions for the primal-dual path value functional. Exact as well as inexact path-following algorithms are proposed in section 6, and their numerical behavior is discussed there as well. (P) 2. Problem statement, regularization, and its motivation. We consider { min 2 a(y, y) (f,y) over y H () s.t. y ψ, where f L 2 (), ψ H (), with ψ, where is a bounded domain in R n with Lipschitz continuous boundary. Throughout, (, ) denotes the standard L 2 ()-inner product, and we assume that a(, ) is a bilinear form on H () H () satisfying (6) a(v, v) ν v 2 H and a(w, z) μ w H z H for some ν >, μ>independent of v H () and w, z H (). Here and throughout we use v H = v L 2 for v H (), which defines a norm on H () due to Friedrichs inequality, and w H =( w 2 L + w 2 2 L ) /2 denotes the standard 2 H -norm; see, e.g., [2]. Moreover, let A: H () H () be defined by a(v, w) = Av,w H,H for all v, w H (). It is well known that (P) admits a unique solution y H () with associated Lagrange multiplier λ = Ay + f, satisfying the optimality system { a(y,v)+ λ,v H,H (7) =(f,v), λ,y ψ H,H =, y ψ, λ,v for all v. This also holds with f H (). Under well-known additional requirements on a, ψ, and, as for example { a(v, w) = ( a ij v xi w xj + dvw), with a ij C ( ),d L (), (8) d, ψ H 2 (), is C,, or is a convex polyhedron, we have (y,λ ) H 2 () L 2 (), and the optimality system can be expressed as { Ay + λ = f in L 2 (), (9) λ =(λ + c(y ψ)) + for some c>, where (v) + = max(,v); for details see, e.g., [4]. Our aim is the development of Newton-type methods for solving (7) or (9), which is complicated by the system of inequalities in (7) and the nondifferentiable maxoperator in (9). In the recent past significant progress was made in the investigation of semismooth Newton methods and primal-dual active set methods for coping with nondifferentiable functionals in infinite dimensional spaces; see, for instance, [, 5]. A direct application of these techniques to (9) results in the following algorithm.

PATH-FOLLOWING METHODS 63 Algorithm A. (i) Choose c>, (y,λ ); set k =. (ii) Set A k+ = {x : λ k (x)+c(y k (x) ψ(x)) > }. (iii) Compute y k+ = arg min { 2 a(y, y) (f,y): y = ψ on A k+}. (iv) Let λ k+ be the Lagrange multiplier associated with the constraint in (iii), with λ k+ = on \A k+. (v) Set k := k + and go to (ii). The optimality system for the variational problem in (iii) is given by { a(yk+,v)+ λ k+,v H,H =(f,v) for all v H (), () y k+ = ψ on A k+, λ k+ = on I k+ =\A k+. This corresponds to (3) in our introductory discussion. The Lagrange multiplier associated with the constraint y = ψ on A k+ is in general only a distribution in H () and is not in L 2 (). In fact, λ k+ is related to the jumps in the normal derivatives of y across the interface between A k+ and I k+ [3]. This complicates the convergence analysis for Algorithm A since the calculus of Newton (or slant) differentiability [] does not apply. We note that these difficulties are not present if (7) or (9) is discretized. However, they are crucial for the treatment of infinite dimensional problems, and as such they are generic. Analogous difficulties arise for state constrained optimization problems, for inverse problems with BV-regularization, and for elasticity problems with contract and friction, to mention a few. This suggests the introduction of regularized problems, which in our case are chosen as min (P a(y, y) (f,y)+ ( λ + γ(y ψ)) + 2 over y H γ ) (), 2 2γ where γ>and λ L 2 (), λ are fixed. For later use we denote the objective functional of (P γ )byj(y; γ). The choice of λ will be used to influence the feasibility of the solution y γ of (P γ ). Using Lebesgue s bounded convergence theorem to differentiate the max under the integral in J(y; γ), the first order optimality condition associated with (P γ ) is given by { a(yγ,v)+(λ γ,v)=(f,v) for all v H (), (OC γ ) λ γ =( λ + γ(y γ ψ)) +, where (y γ,λ γ ) H () L 2 (). With (8) holding, we have y γ H 2 (). The primaldual active set strategy, or equivalently the semismooth Newton method, for (P γ )is given next. For its statement and for later use we introduce χ A k+, the characteristic function of the set A k+. Algorithm B. (i) Choose λ, (y,λ ); set k =. (ii) Set A k+ = {x : λ(x)+γ(y k (x) ψ(x)) > }, I k+ =\A k+. (iii) Solve for y k+ H (): a(y k+,v)+(( λ + γ(y k+ ψ))χ Ak+,v)=(f,v) for all v H (). (iv) Set λ k+ = { on Ik+, λ + γ(y k+ ψ) on A k+.

64 MICHAEL HINTERMÜLLER AND KARL KUNISCH Algorithm B was analyzed in [3], where global as well as locally superlinear convergence for every fixed γ> were established. However, the choice and adaptation (increase) of γ was heuristic in [3] and earlier work. The focus of the present investigation is the automatic adaptive choice of γ. We shall utilize the following two results, which we recall from [3] where the proofs can also be found. Proposition 2.. The solutions (y γ,λ γ ) to (OC γ ) converge to (y,λ ) in the sense that y γ y strongly in H () and λ γ λ weakly in H () as γ. We say that a satisfies the weak maximum principle if for any v H () () a(v, v + ) implies v + =. Proposition 2.2. Assume that () holds and let <γ γ 2 <. (a) In the infeasible case, i.e., for λ =, we have y y γ2 y γ. (b) In the feasible case, i.e., if (2) λ and λ f + Aψ, v H,H for all v H (), with v, then y γ y γ2 y ψ. 3. The primal-dual path. In this section we introduce the primal-dual path and discuss its smoothness properties. Definition 3.. The family of solutions C = {(y γ,λ γ ): γ (, )} to (OC γ ), considered as subset of H () H (), is called the primal-dual path associated with (P). For r we further set C r = {(y γ,λ γ ): γ [r, )}, and with some abuse of terminology we also refer to C r as a path. In the following lemma we denote by ŷ the solution to the unconstrained problem (ˆP) min J(y) = a(y, y) (f,y) 2 over y H (). Subsequently, in connection with convergence of a sequence in function space we use the subscript weak together with the space to indicate convergence in the weak sense. Lemma 3.2. For each r> the path C r is bounded in H () H (), with lim γ (y γ,λ γ )=(y,λ ) in H () H () weak.for λ =the path C is bounded in H () H (), with lim γ +(y γ,λ γ )=(ŷ, ) in H () L 2 (). Proof. From (OC γ ) we have for every γ> (3) a(y γ,y γ y )+(λ γ,y γ y )=(f,y γ y ). Since λ γ = max(, λ + γ(y γ ψ)) and ψ y wehave ( (λ γ,y γ y )= λ γ, λ γ + y γ ψ + ψ y λ ) γ Combined with (3) this implies that γ (λ γ, λ + γ(y γ ψ)) γ (λ γ, λ) = [ λγ 2 L γ (λ γ, λ) ]. 2 (4) a(y γ,y γ )+ γ λ γ 2 L 2 a(y γ,y )+(f,y γ y )+ γ ( λ, λ γ ).

PATH-FOLLOWING METHODS 65 This estimate, (6), (OC γ ), and the Poincaré Friedrichs inequality imply that C r is bounded in H () H () for every r>. In fact, for ω> satisfying ω y 2 H y 2,wehave H ω y γ 2 H + γ λ γ 2 L 2 a(y γ,y γ )+ γ λ γ 2 L 2 and hence μ y γ H y H + f H ( y γ H + y H )+ γ λ L 2 λ γ L 2 ω 4 y γ 2 H + μ2 ω y 2 H + ω 2 y γ 2 H + 2ω f 2 H + 2γ λ γ 2 L 2 + 2γ λ 2 L 2 + f H y H, ω 4 y γ 2 H + 2γ λ γ 2 L μ2 2 ω y 2 H + 2ω f H + f H y H + 2γ λ 2 L. 2 This estimate implies that {y γ : γ r} is bounded in H () for every r>. The first equation of (OC γ ) implies that {λ γ : γ r} is bounded in H () as well. From Proposition 2. we have that lim γ (y γ,λ γ )=(y,λ )inh () H () weak. If λ =, then from (4), (6), and (OC γ ) the path C o is bounded in H () H () and λ γ inl 2 () for γ +. From (OC γ ) and the optimality condition for (ˆP) we have a(y γ ŷ, y γ ŷ)+(λ γ,y γ ŷ) =, and hence lim γ + y γ =ŷ in H (). Proposition 3.3. The path C r is globally Lipschitz in H () H () for every r>. If λ =, then C is globally Lipschitz continuous. Proof. Let γ, γ [r, ) be arbitrary. Then A(y γ y γ )+( λ + γ(y γ ψ)) + ( λ + γ(y γ ψ)) + =. Taking the inner-product with y γ y γ and using the monotonicity and Lipschitz continuity (with constant L =)ofx max(,x), we find a(y γ y γ, y γ y γ ) ( ( λ + γ(y γ ψ)) + ( λ + γ(y γ ψ)) +,y γ y γ ) γ γ y γ ψ L 2 y γ y γ L 2. By Lemma 3.2 the set {y γ } γ r is bounded in H (). Hence there exists K > such that ν y γ y γ 2 H K γ γ y γ y γ L 2, and by Poincaré s inequality there exists K 2 > such that y γ y γ H K 2 γ γ for all γ r, γ r.

66 MICHAEL HINTERMÜLLER AND KARL KUNISCH Let us recall here that y H = y L 2. Lipschitz continuity of γ λ γ from [r, ) to H () follows from the first equation in (OC γ ). For λ = the set {y γ } γ is bounded in H (). The remainder of the proof remains identical. Lemma 3.4. For every subset I [r, ), r>, the mapping γ λ γ is globally Lipschitz from I to L 2 (). Proof. For <γ γ 2 we have by (OC γ ) λ γ λ γ2 L 2 = ( λ + γ (y γ ψ)) + ( λ + γ 2 (y γ2 ψ)) + L 2 for some constant K 3 >. We shall use the following notation: Further we set (5) (K 3 γ + K + ψ L 2) γ γ 2 S γ = {x : λ(x)+γ(y γ ψ)(x) > }. g(γ) = λ + γ(y γ ψ). Since γ y γ H () is Lipschitz continuous by Proposition 3.3, there exists a weak accumulation point ẏ(= ẏ γ )of γ γ (y γ y γ )as γ γ>, which is also a strong accumulation point in L 2 (). Further γ γ (g( γ) g(γ)) has ġ(γ): = y γ ψ + γ ẏ γ as a strong accumulation point in L 2 () as γ γ. In case γ approaches γ from above (or below), the associated accumulation points ẏγ r (or ẏγ) l satisfy certain properties which are described next. In what follows we use ġ r (γ) orġ l (γ) whenever ẏ γ in ġ(γ) is replaced by ẏγ r and ẏγ, l respectively. Proposition 3.5. Let γ>, and denote by ẏγ r any weak accumulation point of γ γ (y γ y γ ) in H () as γ γ. Set Then ẏ r γ satisfies (6) (7) S + γ = S γ {x: λ(x)+γ(y γ (x) ψ(x))= ġ r (γ)(x) }. a(ẏ r γ,v)+((y γ ψ + γẏ r γ)χ S + γ,v)= for all v H (). Proof. By (OC γ ) we have for every v H () a(y γ y γ,v)+(( λ + γ(y γ ψ)) + ( λ + γ(y γ ψ)) +,v)=. We multiply (7) by ( γ γ) and discuss separately the two terms in (7). Clearly, we have lim( γ γ) a(y γ y γ,v)=a(ẏ r γ γ γ,v). Here and below the limit is taken on the sequence of γ-values, which provides the accumulation point. Lebesgue s bounded convergence theorem allows us to consider the pointwise limits of the integrands. Considering separately the cases g(γ)(x)<, g(γ)(x) >, and g(γ)(x) =,wehave (8) ( γ γ) ((g( γ)) + (g(γ)) +,v) ((y γ ψ + γ ẏ r γ)χ S + γ,v) as γ γ,

PATH-FOLLOWING METHODS 67 which ends the proof. As a consequence of the proof we obtain the following result. Corollary 3.6. Let γ>, and denote by ẏγ l any weak accumulation point of γ γ (y γ y γ ) in H () as γ γ. Set Sγ = S γ {x: λ(x) +γ(y γ (x) ψ(x)) = ġ l (γ)(x) }. Then ẏγ l satisfies (9) a(ẏ l γ,v)+((y γ ψ + γ ẏ l γ)χ S γ,v) = for all v H (). Another corollary of Proposition 3.5 treats the case λ =. Corollary 3.7. Let λ =, and assume that () holds. Then the right- and left- derivatives ẏ r γ and ẏ l γ of γ y γ, γ (, ), exist and are given by (2) (2) a(ẏγ,v)+((y r γ ψ + γ ẏγ)χ r {yγ>ψ},v) = a(ẏγ,v)+((y l γ ψ + γ ẏγ)χ l {yγ ψ},v) = for all v H (), for all v H (). Proof. Let γ γ. By Proposition 2.2 any accumulation point ẏ r γ of ( γ γ) (y γ y γ ) satisfies ẏ r γ and hence Observe that S + γ = {x : y γ (x) >ψ(x)} {x : y γ (x) =ψ(x) ẏ r γ(x) =}. (y γ ψ + γẏ r γ)χ S + γ =(y γ ψ + γẏ r γ)χ {yγ>ψ}. This implies that every accumulation point ẏγ r satisfies (2). Since the solution to (2) is unique, the directional derivative from the right exists. Similarly, if γ γ, by Proposition 2.2 we have Sγ = {x : y γ (x) ψ(x)}, and (2) follows. Henceforth we set S γ = {x : λ(x)+γ(y γ ψ)(x) =}. Corollary 3.8. If meas(s γ)=, then γ y γ H () is differentiable at γ, and the derivative ẏ γ satisfies (22) a(ẏ γ,v)+((y γ ψ + γ ẏ γ )χ Sγ,v)= forall v H (). Proof. Let z denote the difference of two accumulation points of ( γ γ) (y γ y γ ) as γ γ. As a consequence of (6) and (9) a(z,v)+γ(zχ Sγ,v) = for all v H (). This implies that z = by (6). Consequently, accumulation points are unique, and by (6), (9) they satisfy (22). The assumption meas(sγ) o = in Corollary 3.8 reflects the lack of differentiability of the max-operation in (OC γ ). 4. The primal-dual path value functional. In this section we investigate the value function associated with (P γ ) and study its monotonicity and smoothness properties. Definition 4.. The functional γ V (γ) =J(y γ ; γ) = 2 a(y γ,y γ ) (f,y γ )+ 2γ ( λ + γ(y γ ψ)) + 2 L 2

68 MICHAEL HINTERMÜLLER AND KARL KUNISCH defined on (, ) is called the primal-dual path value functional. Let us start by studying first order differentiability properties of V. Proposition 4.2. The value function V is differentiable with V (γ) = 2γ 2 ( λ + γ(y γ ψ)) + 2 + γ ( λ + γ(y γ ψ)) + (y γ ψ). Corollary 4.3. For λ =we have V (γ) = 2 (y γ ψ) + 2 and V (γ) > unless y γ is feasible. For λ satisfying (2) and with () holding, we have y γ ψ and hence V (γ) for γ (, ). In either of the two cases V (γ) =implies that y γ solves (ˆP). Proof. We show only that V (γ) = implies that y γ solves (ˆP). The rest of the assertion follows immediately from Proposition 4.2. If λ =, then V (γ) = yields y γ ψ. Thus, λ γ =, and hence y γ solves (ˆP). If () and (2) are satisfied, then y γ ψ and V (γ) = implies γ(y γ ψ) λ + γ(y γ ψ). As a consequence λ γ =, and y γ solves (ˆP). Proof of Proposition 4.2. For γ,γ (, ) wefind (23) and consequently 2 a(y γ + y γ,y γ y γ ) (f,y γ y γ ) + 2 (( λ + γ(y γ ψ)) + +( λ + γ(y γ ψ)) +,y γ y γ )=, V ( γ) V (γ) = 2 a(y γ,y γ ) 2 a(y γ,y γ ) (f,y γ y γ ) + ( λ + γ(y γ ψ)) + 2 ( λ + γ(y γ ψ)) + 2 2 γ 2γ = ( λ + γ(y γ ψ)) + 2 + ( λ + γ(y γ ψ)) + 2 2 γ 2γ + (( λ + γ(y γ ψ)) + +( λ + γ(y γ ψ)) + )(y γ y γ ) 2 = z + z + z = I + I 2 + I 3, P γ P γ P γ N γ P γ N γ where z stands for the sum of the kernels on the left of the above equalities, P γ = {x: λ + γ(y γ ψ) > }, N γ = {x: λ + γ(y γ ψ) < }, and P γ, N γ are defined analogously. For I 2 we have I 2 2 P γ N γ γ ( λ + γ(y γ ψ)) 2 + λ + γ(y γ ψ) y γ y γ 2 γ ( γ(y γ ψ) γ(y γ ψ)) 2 + y γ y γ ( γy γ γy γ + γ γ ψ ), and hence by Proposition 3.3 (24) lim γ γ γ γ I 2 =.

Analogously one verifies that PATH-FOLLOWING METHODS 69 (25) On P γ P γ we have z = 2 γ ( λ + γ(y γ ψ)) 2 lim γ γ γ γ I 3 =. 2γ ( λ + γ(y γ ψ)) 2 2 (2 λ + γ(y γ ψ)+γ(y γ ψ))(y γ y γ ) = γ γ 2 γγ ( λ + γ(y γ ψ)) 2 + 2γ [ 2 λ( γ(y γ ψ) γ(y γ ψ)) + γ 2 (y γ ψ) 2 γ 2 (y γ ψ) 2] 2 (2 λ + γ(y γ ψ)+γ(y γ ψ))(y γ y γ ) = γ γ 2 γγ ( λ + γ(y γ ψ)) 2 + λ γ [ γ(y γ ψ) γ(y γ ψ)] + [ ] γ 2 2 γ (y γ ψ) 2 γ(y γ ψ) 2 +( γ γ)(y γ ψ)(y γ ψ), and thus on P γ P γ ( γ γ) z = 2 γγ ( λ + γ(y γ ψ)) 2 + λ γ (y γ ψ) + [ ] γ 2 γ (y γ ψ) 2 +(y γ ψ)(y γ ψ). By Lebesgue s bounded convergence theorem, lim γ γ γ γ I = lim zχ γ γ γ γ P γ Pγ = 2γ 2 + γ(y γ ψ)) (( λ + ) 2 + ( λ + γ(y γ ψ)) + (y γ ψ). γ Together with (24) and (25), this implies the claim. Remark 4.. Note that V is characterized without recourse to ẏ γ. The boundedness of {γ 2 V (γ)}γ is established next. In what follows we use (v) = min(,v). Proposition 4.4. If λ = and a(v +,v ) = for all v H (), then {γ 2 V (γ)}γ is bounded. If () and (2) hold, then again {γ 2 V (γ)}γ is bounded. Proof. In the case λ = we have a(y γ ψ, v)+γ((y γ ψ) +,v)=(f,v) a(ψ, v) for all v H (). Since (y γ ψ) H () and a((y γ ψ) +, (y γ ψ) ) = we have, using (6) with v =(y γ ψ) +, ν (y γ ψ) + 2 H () + γ (y γ ψ) + 2 L 2 f L 2 (y γ ψ) + H + μ ψ H y γ ψ H.

7 MICHAEL HINTERMÜLLER AND KARL KUNISCH This implies the existence of a constant K, depending on ψ H and f L 2 but independent of γ, such that γ (y γ ψ) + L 2 K. Since V (γ) = 2 (y γ ψ) + 2 the claim follows. Turning to the feasible case with () and (2) holding, we have that y γ ψ for every γ>, and hence ( λ + γ(y γ ψ))(x) > if and only if λ(x) >γ(ψ y γ )(x). Consequently, V (γ) 2γ 2 ( λ + γ(y γ ψ)) + 2 + ( λ + γ(y γ ψ)) + (ψ y γ ) γ 3 2γ 2 λ 2 L 2, which again implies the claim. Before we investigate V, we state a result which connects γ V (γ), y y γ H, and V V (γ), where V = lim γ V (γ). It will be used in section 6. for designing a γ-update strategy. Proposition 4.5. In the feasible and infeasible cases the following estimate holds true: y y γ 2 H 2 ( V V (γ) γ ν V ) (γ). Proof. We have V V (γ) =J(y ) J(y γ ; γ) and J(y ) J(y γ ; γ) ν 2 y y γ 2 H + a(y γ,y y γ ) (f,y y γ ) 2γ ( λ + γ(y γ ψ)) + 2 L 2, where we have used (6). From (OC γ )wehave and hence a(y γ,y y γ ) (f,y y γ )= (( λ + γ(y γ ψ)) +,y y γ ), J(y ) J(y γ ; γ) ν 2 y y γ 2 H (( λ + γ(y γ ψ)) +,y y γ ) 2γ ( λ + γ(y γ ψ)) + 2 L 2 ν 2 y y γ 2 H 2γ ( λ + γ(y γ ψ)) + 2 L 2 +(( λ + γ(y γ ψ)) +,y γ ψ) = ν 2 y y γ 2 H + γ V (γ). This completes the proof. Below we shall assume that y γ ψ C( ). Recall that for dimension n 3 and with (6) and (8) holding, we have y γ H 2 () C( ). Proposition 4.6. Let ẏ γ denote any accumulation point of ( γ γ) (y γ y γ ) as γ γ. (a) If λ =, y γ ψ C( ), and (8) is satisfied, then γ V (γ) is twice differentiable at γ with (26) V (γ) = (y γ ψ) + ẏ γ.

PATH-FOLLOWING METHODS 7 (b) For arbitrary λ, ifmeas(sγ)=, then γ V (γ) is twice differentiable at γ with V (γ) = γ 3 ( λ + γ(y γ ψ)) + 2 2 (27) γ 2 ( λ + γ(y γ ψ)) + (y γ ψ) + (y γ ψ)(y γ ψ + γ ẏ γ )χ Sγ. γ Proof. (a) On the subsequence γ n realizing the accumulation point, we have that lim n (γ n γ) ( V (γ n ) V (γ)) equals the right-hand side of (26). The claim will be established by verifying that the accumulation points ẏ γ restricted to S γ = {x: y γ (x) ψ(x) > } are unique. Let z denote the difference of two accumulation points. By (6) and (9) we have a(z,v)+γ(z,v) = for all v H () with v = on \ S γ. Using (8) and the fact that S γ is an open set relative to due to the continuity of y γ ψ, we find that z =ins γ, as desired. (b) Let ẏ γ denote any accumulation point of ( γ γ) (y γ y γ )as γ γ, and recall the notation g(γ) = λ + γ(y γ ψ) and S γ + from section 3. On the subsequence realizing the accumulation point we find lim γ γ γ γ ( V ( γ) V (γ)) = γ 3 ( λ + γ(y γ ψ)) + 2 2 (28) γ 2 ( λ + γ(y γ ψ)) + (y γ ψ) + (y γ ψ)(y γ ψ + γ ẏ γ )χ γ S + γ. By assumption, meas(s γ) = and, hence the right-hand sides of (27) and (28) coincide. Since ẏ γ is unique by Corollary 3.8 the claim is established. 5. Model functions. In this section we derive low-parameter families of functions which approximate the value functional V and share some of its qualitative properties. We will make use of these models in the numerics section when devising path-following algorithms. 5.. Infeasible case. Throughout this subsection we assume (8) and (29) λ =,y γ ψ C( ) for all γ (, ). Observe that (8), together with the general assumption (6), implies (). In fact, for any v H () we have a(v, v + ) γ v + 2, and hence a(v, v + ) implies v + =. Proposition 5.. The value function V satisfies V (γ) and V (γ) for γ (, ). Proof. Proposition 4.2 implies that V (γ). Moreover, y γ2 y γ for γ 2 γ > and hence ẏ γ a.e. on S γ. Consequently V (γ) by Proposition 4.6. A model function m for the value function V should reflect the sign properties of V and its derivatives. Moreover, V () gives the value of (ˆP), and hence we shall require that m() = V (). Finally from Lemma 3.2 we conclude that V is bounded on [, ). All these properties are satisfied by functions of the form (3) m(γ) =C C 2 E + γ

72 MICHAEL HINTERMÜLLER AND KARL KUNISCH with C R. Here C 2, E> satisfy (3) m() = V () = C C 2 E. Other choices for model functions are also conceivable, for example, γ C C (E+γ) r with r>. Note, however, that the asymptotic behavior of the model in (3) is such that γ 2 ṁ(γ) is bounded for γ. This is consistent with the boundedness of γ 2 V (γ) for γ asserted in Proposition 4.4. Another reason for choosing (3) is illustrated next. Choosing v =(y γ ψ) + in (OC γ ), we find (32) a(ẏ γ, (y γ ψ) + )+ (y γ ψ) + 2 L + γ (y 2 γ ψ) + ẏ γ =. For the following discussion we (33) replace a(, ) bye(, ) with E> a constant, and V by m. By Proposition 4.2 and (26) the following ordinary differential equation is obtained for m: (34) (E + γ) m(γ)+ 2 ṁ(γ) =. The solutions to (34) are given by (3). To get an account for the quality of our model in (3) we refer to the left-hand plot of Figure 4 in section 6. 5.2. Feasible case. Throughout this subsection we assume (35) (), λ satisfies (2), and meas (S γ ) = for all γ (, ). Proposition 5.2. The value function V satisfies V (γ) and V (γ) for γ (, ). Proof. By Proposition 2.2 we have y γ ψ and hence V (γ) by Proposition 4.2. A short computation based on (27) shows that (36) V (γ) = γ 3 χ λ 2 + χ(y γ ψ)ẏ γ γ χ(y γ ψ) 2 + χ (y γ ψ)ẏ γ, where χ is the characteristic function of the set S γ = { λ + γ(y γ ψ) > }. From (22) we have γ ẏ γ L 2 (S γ) ψ y γ L 2 (S γ), and hence V (γ). An immediate consequence is stated next. Lemma 5.3. If the solution to the unconstrained problem is not feasible, then lim γ V (γ) =. Proof. Assume that lim γ V (γ) is finite. Then, using (P γ ), there exists a sequence γ n and ỹ H () such that y γn ỹ weakly in H (), with y γn the solution to (P γn ), and λ γn = max(, λ + γ n (y n ψ)) inl 2 (). Consequently ỹ ψ. Taking the limit with respect to n in (OC γn ), it follows that ỹ ψ is the solution to (ˆP), which contradicts our assumption.

PATH-FOLLOWING METHODS 73 From Lemmas 3.2 and 5.3 and Proposition 5.2 it follows that γ V (γ), γ (, ), is a monotonically strictly decreasing convex function with lim γ + V (γ) =. All these properties are also satisfied by functions of the form (37) m(γ) =C C 2 E + γ + B γ, provided that C R, C 2, E>, B>, and C 2 B. We now give the motivation for choosing the model function m for V as in (37). From (22) with v =(y γ ψ)χ we get a(ẏ γ, (y ψ)χ)+γ(ẏ γ χ, y γ ψ)+((y γ ψ)χ, y γ ψ) =, where χ = χ Sγ. As in the infeasible case we replace a(, )bye(, ), with E a constant, and using (22), we arrive at The choice v = y γ ψ implies (38) (E + γ)(ẏ γ χ, v)+((y γ ψ)χ, v) =. (E + γ)(ẏ γ χ, y γ ψ)+((y γ ψ)χ, y γ ψ) =. Note that V (γ) can be expressed as V (γ) = λ 2 2γ 2 χ + (39) (y γ ψ) 2 χ. 2 Using (36) and (39) in (38), and replacing V by m, due to the substitution for a(, ), we find (E + γ) m +2ṁ Eγ 3 χ λ 2 =. We further replace χ λ 2, which is a bounded quantity depending on γ, by2b, and obtain, as the ordinary differential equation that we propose for the model function m in the feasible case, (4) (E + γ) m +2ṁ 2γ 3 EB=. The family of solutions is given by (37). In the right-hand plot of Figure 4 in section 6 we depict the approximation quality of m(γ). 6. Path-following algorithms. In this section we study the basic Algorithm B together with a variety of adjustment schemes for the path parameter γ. For this purpose recall that, depending on the shift parameter λ, the elements y γ along the primal-dual path are feasible or infeasible. As we have seen in the previous section, this implies different models for approximating the value function V. We will see, however, that for γ>in both cases similar strategies for updating γ may be used. When referring to the infeasible or feasible case, (29), respectively (35), is assumed to hold. The subsequent discussion is based on the following two-dimensional test problems. We point out that the bound ψ in problem P below does not satisfy ψ H (). However, as we shall see, the feasible and infeasible primal-dual path as well as the algorithms introduced subsequently still perform satisfactorily. We include this example since discontinuous obstacles are of practical relevance.

74 MICHAEL HINTERMÜLLER AND KARL KUNISCH Test problem P. We consider (8) with a ij = δ ij, with δ ij the Kronecker symbol, d =,and=(, ) 2. We choose f(x,x 2 ) = 5x sin(5x ) cos(x 2 ) and ψ on \ K, and ψ onk with K = {x : 5 x ( 2, 2 ) 2 2 5 }. The solution y, the obstacle ψ, and the active set A at the solution are shown in Figure. Fig.. Optimal solution y (upper left plot), obstacle ψ (upper right plot), and the active set A (lower plot) for test problem P. Test problem P2. Again we consider (8), with a ij, d, and as before, and define x on T := {x :x 2 x x 2 x }, y x := 2 on T 2 := {x :x 2 x x 2 x }, (4) x on T 3 := {x :x 2 x x 2 x }, x 2 on T 4 := {x :x 2 x x 2 x }. The obstacle ψ is defined by ψ y on S := {x : x ( 2, 2 ) 4 }, ψ 4 on S 2 \ S, and 2x on T ( \ S 2 ), 4 ψ := 2(x 2 7 8 ) on T 2 ( \ S 2 ), 4 2(x 7 8 ) on T 3 ( \ S 2 ), 2x 2 on T 4 ( \ S 2 ), with S 2 := {x : x ( 2, 2 ) 3 8 }. The forcing term is given by (f,φ) L 2 = φ(s)ds +(χ S,φ) L 2 + φ(s)ds for all φ H (), + S +

PATH-FOLLOWING METHODS 75 where + := {x :x 2 = x } {x :x 2 = x }. We recall that for φ H (), R 2, the traces along smooth curves are well defined. The solution y is given by y = y. The active or coincidence set at the solution is A = S. The Lagrange multiplier λ = f +Δy is in H () and enjoys no extra regularity. In Figure 2 we display the optimal solution y, the obstacle ψ, and the active set A. Fig. 2. Optimal solution y (upper left plot), obstacle ψ (upper right plot), and the active set A (lower plot) for test problem P2. Test problem P3. For this test problem (8) is satisfied. We therefore obtain y H 2 () and λ L 2 (). The coefficients a ij and d as well as are as before. The volume force f is given by f = Δv with v(x,x 2 ) = sin(3πx ) sin(3πx 2 ). Further, we have ψ = 4 sin(πx ) sin(πx 2 ). The optimal solution y, the Lagrange multiplier λ, and the active set at y are displayed in Figure 3. Unless specified otherwise, the subsequent algorithms are initialized by y = ( Δ) f, where Δ denotes the Laplacian with homogeneous Dirichlet boundary conditions. The initial Lagrange multiplier is chosen as λ = γ χ {y>ψ}(y ψ). The discretization of Δ is based on the classical five-point finite difference stencil. We denote the mesh size by h, which we occasionally drop for convenience. The forcing term f in P2 is discretized by f = Δy + χ S e + χ S ( Δy ), where e is the vector of all ones and χ S represents a diagonal matrix with entry (χ S ) ii = for grid points x i S and (χ S ) ii = otherwise. Above y denotes the grid function corresponding to (4). 6.. A strategy based on model functions exact path-following. As outlined in section 5, there are good reasons to trust our model functions (3) and (37) in the infeasible and feasible cases, respectively. Let us start by focusing on the infeasible case. The model is given by m(γ) =C C 2 (E + γ). For determining

76 MICHAEL HINTERMÜLLER AND KARL KUNISCH Active set P3 Fig. 3. Optimal solution y (upper left plot), obstacle ψ (upper right plot), and the active set A (lower plot) for test problem P3. the three parameters C,C 2, and E, we use the information V (),V(γ), V (γ), which, by Proposition 4.2, is available from one solve of the unconstrained problem (ˆP) and one solve for (Pγ). The conditions (42) yield (43) m() = V (), m(γ) =V (γ), ṁ(γ) = V (γ) E = γ 2 V (γ) (V (γ) V () γ V (γ)), C 2 = γ E(E + γ)(v (γ) V ()), C = V () + C 2 E. We could have used an alternative reference value γ r (,γ) and computed m(γ r )= V (γ r ) instead of m() = V (). In Figure 4 we compare V (γ) tom(γ) for different values of the coefficients (C,C 2,E). These coefficients depend on different values y f for γ (in (42)) produced by Algorithm EP (see below) for problem P. The solid line corresponds to V (γ). The corresponding γ-values γ f for (42) are depicted in the legend of the left plot in Figure 4. The dotted and dashed line belong to rather small γ-values, and the dashed-dotted and the circled lines to large γ f in (42). As we can see, the dotted line is accurate in the range of relatively small γ f, while the other lines are more accurate for large γ f. From now on we consider only the choices γ r = and γ = γ k in (42) when updating γ k. Next we discuss properties of the model parameters E, C, C 2 according to (43). For this purpose assume that the solution ŷ to (ˆP) is not feasible for (P). Then

PATH-FOLLOWING METHODS 77 m(γ), V(γ) 82 83 84 85 86 87 Model m(γ) vs V(γ) (solid); infeasible case V(γ) γ f =.35E3 γ f =7.37E5 γ f =7.55E8 γ f =.93E 88 2 4 6 8 γ m(γ) V(γ).7.6.5.4.3.2. m(γ) V(γ) ; feasible case 2 4 6 8 γ γ f = γ f =.56E2 γ f =4.E3 γ f =.9E5 γ f =.52E7 Fig. 4. Left: Model m(γ) vs. V (γ) (solid) in the infeasible case for P. Right: Model m(γ) vs. V (γ) in the feasible case. by Corollary 4.3 we have V (γ) > for all γ>. Consequently V (γ) >V() and V (γ) V () γ V (γ) = γ γ V (σ)dσds >, and hence E>and C s 2 > for all γ (, + ). This implies m(γ) C and m(γ) C for γ +. We propose the following update strategy for γ: Let {τ k } satisfy τ k (, ) for all k N and τ k ask, and assume that V (γ k ) is available. Then, given γ k, the updated value γ k+ should ideally satisfy (44) V V (γ k+ ) τ k V V (γ k ). Since V and V (γ k+ ) are unknown, we use C,k and our model m k (γ) =C,k C 2,k /(E k + γ) atγ = γ k+ instead. Thus, (44) is replaced by (45) C,k m k (γ k+ ) τ k C,k V (γ k ) =: β k. Solving the equation C,k m k (γ k+ )=β k, we obtain γ k+ = C 2,k (46) E k. β k In Theorem 6. we shall show that γ k+ κγ k, with κ>, independently of k N. Before we turn to the feasible case, we interpret (44) in view of Proposition 4.5 in the infeasible case. Recall that V V (γ), and observe that V V (γ) = O( y y γ H ). Proposition 4.5 yields y y γ 2 H 2 ν (V V (γ)) since V (γ) >. Setting τ k = ω 2 k V V (γ k ), with ω k, in (44) yields Consequently, we obtain y y γk+ 2 H C τ ωk y 2 y γk 2 H. y y γk+ H y y γk H C τ ω k,

78 MICHAEL HINTERMÜLLER AND KARL KUNISCH which implies q-superlinear convergence of {y γk } in H (). In the feasible case, i.e., when λ satisfies (2), we use the model m(γ) =C C 2 (E + γ) + Bγ with C 2 and E,B > ; see (37). Let γ r >, γ r γ, denote a reference γ-value; then we use the conditions m(γ r )=V (γ r ), ṁ(γ r )= V (γ r ), m(γ) =V (γ), ṁ(γ) = V (γ) for fixing B, C, C2, E. Solving the corresponding system of nonlinear equations, we get (( (γ r γ)( V (γ r )γr 2 + V )) (γ)γ 2 )+2γ r γ(v (γ) V (γ r ) E = ( ( V (γ)γ + V ) (γ r )γ r )(γ γ r )+(γ r + γ)(v (γ r ) V (γ)) and B = ( γr 2 γ 2 (V (γ) V (γ r )) 2 V (γ) V ) (γ r )(γ γ r ) 2 ( (γ γ r ) 2 ( V (γ r )γr 2 + V ) (γ)γ 2 )+2(γ γ r )γ r γ(v (γ r ) V (γ)) Then the parameters C and C 2 are given by ( ) B C 2 =(E + γ) 2 γ 2 + V (γ), C = V (γ)+ C 2 E + γ B γ. In the right plot of Figure 4 we show m(γ) V (γ) with m(γ) produced by the iterates of Algorithm EP for P similar to the infeasible case. Again we can see that our model yields a close approximation of the value function V. If we require that (45) be satisfied in the feasible case, then we obtain the following update strategy for γ: (47) γ k+ = D k 2 + D 2 k 4 + B ke k β k, where D k = E k +(C 2,k B k )/β k. In Theorem 6. we shall establish γ k+ κγ k for all k N with κ> independent of k. Next we describe an exact path-following version of Algorithm B, which utilizes the update strategy (45) for updating γ. Algorithm EP. (i) Select γ r. Compute V (γ r ), and choose γ > max(,γ r ); set k =. (ii) Apply Algorithm B to obtain y γk. (iii) Compute V (γ k ), V (γk ), and γ k+ according to (46) in the infeasible case or (47) in the feasible case. (iv) Set k = k +, and go to (ii). Concerning the choice of γ r note that in the infeasible case we have γ r, and in the feasible case γ r >. Convergence of Algorithm EP is addressed next. Theorem 6.. Assume that the solution to (ˆP) is not feasible for (P). Then the iterates γ k of Algorithm EP tend to as k, and consequently lim k (y γk,λ γk )= (y,λ ) in H () H () weak.

PATH-FOLLOWING METHODS 79 Proof. Let us consider the infeasible case. Then (45) is equivalent to (48) <C,k m k (γ k+ ) <τ k (C,k m k (γ k )). Since γ m k (γ) is strictly increasing and τ k (, ), it follows that γ k+ >γ k for every k =,,... If lim k γ k =, then lim k (y γk,λ γk )=(y,λ ). Otherwise there exists γ such that lim k γ k = γ. Since γ V (γ) and γ V (γ) are continuous on (, ), it follows from (42) and (43) that lim k E k = E( γ), lim k C,k = C ( γ), and lim k C 2,k = C 2 ( γ), where E( γ), C ( γ), C 2 ( γ) are given by (43) with γ replaced by γ. Taking the limit with respect to k in (48), we arrive at C 2 ( γ) E( γ)+ γ =, which is impossible, since C 2 ( γ) > and E( γ) > if the solution to (ˆP) is not feasible for (P). Thus lim k γ k =. The feasible case is treated analogously. Numerically we stop the algorithm as soon as (r,h k,r2,h,r3,h) 2 ɛ M, where r,h k r 2,h k = yγ h k +( Δ h ) (λ h γ k f h ) H,h/ f h H,h, = λ h γ k max(,λ h γ k + yγ h k ψ h ) H,h, r 3,h k = max(,yγ h k ψ h ) L h 2, and ɛ M denotes the machine accuracy. Here H,h denotes the discrete version of H. For some vector v it is realized as v H = h ( Δ h ) v L h 2 with L h 2 the discrete L 2 -norm and h a forward difference approximation of the gradient operator; see [8]. The inner iteration, i.e., Algorithm B for γ = γ k, is terminated if successive active sets coincide or Δ h yγ h,l k + λ h,l γ k f h H,h f h ɛ M. H,h Here the superscript l = l(k) denotes the iteration index of Algorithm B for fixed k. For a discussion and numerical results in the case where the approximation errors due to the discretization of the underlying function space problems are incorporated into the algorithmic framework, e.g., when stopping the algorithm, we refer to the next section 6.2. The initialization of γ is as follows: In the infeasible case we propose a choice of γ based on the deviation of the linearization of V (γ) atγ = γ r from the objective value of the unconstrained problem (ˆP) at the projection of y γr onto the feasible set. In our realization of this heuristic we choose γ r = and compute ŷ, V (), and V (). Then we set { γ = max,ζ J(y } b) V () (49), V () where ζ (, ] is some fixed constant, y b (x) = min(ŷ, ψ(x)), and J denotes the objective function of (ˆP). Note that ŷ is the minimizer of the unconstrained problem (ˆP). For the examples below we use ζ =. In the feasible case we choose a reference value γ r, e.g., γ r =, and solve the path problem (P γ ). Then we choose γ = γ r + J(ŷ) V (γ r) (5), V (γ r ) k k

8 MICHAEL HINTERMÜLLER AND KARL KUNISCH where ŷ denotes the minimizer of the discretized unconstrained problem (ˆP). If ŷ is not feasible for (P), then one has J(ŷ) <V(γ r ) and hence γ >γ r. When applied to P, P2, and P3 for h =/28 and with τ k =. k+, we obtain the results shown in Figure 5 and Table 6.. # inner iterations 25 2 5 5 inner iterations P 2 3 4 5 outer iteration inner iterations 6 5.5 5 4.5 4 3.5 3 2.5 2.5 inner iterations P2 2 3 4 outer iteration # inner iterations 6 4 2 8 6 4 2 inner iterations P3 2 3 4 outer iteration Fig. 5. Number of inner iterations (vertical axis) per outer iteration for P (left plot), P2 (middle plot), and P3 (right plot); solid line infeasible case, dashed line feasible case. Table 6. Comparison of iteration counts. P P2 P3 Version # outer # inner # outer # inner # outer # inner Feasible 5 44 4 4 3 Infeasible 4 5 4 4 6 From our test runs, also for other test problems, we observe the following characteristics: For the feasible version the number of inner iterations exhibits an increasing tendency until a saturation value is reached, and then, unless the algorithm stops at an approximate solution, it starts to decrease. For the infeasible version we typically observe that the first couple of iterations require several inner iterations. As the outer iterations proceed the number of inner iterations drops eventually to one. We also tested less aggressive γ-updates compared to the ones used here, e.g., updates based on γ k+ = ξγ k with ξ> fixed. The numerically observable convergence speed of y γk towards y in H () is typically superlinear. This can be seen from Figure 6, where the plots for the discrete versions qk h of the quotients q k = y γ k+ y H y γk y H are shown. Note that the vertical axis uses a logarithmic scale. In the first row, for P we depict the behavior of qk h for h =2 i, i =5, 6, 7, 8, for the infeasible case (left plot) and the feasible case (right plot). We observe that the convergence rate is stable with respect to decreasing mesh size h. In the second row we see the behavior of qk h for P2 and P3, with h =2 7. Again, we observe a superlinear rate of convergence. With respect to decreasing h the same conclusion as for P holds true. These stability results provide a link between our function space theory and the numerical realization of the algorithms.

PATH-FOLLOWING METHODS 8 Convergence behavior of q k h ; P2 2 3 4 5 2 3 outer iteration feasible infeasible 2 3 4 Convergence behavior of q k h ; P3 infeasible feasible 2 3 outer iteration Fig. 6. Discrete quotients qk h for P and various mesh sizes h (upper row) and for P2 (lower left) and P3 (lower right) for h =/28. number of grid points 2 8 6 4 2 8 6 4 2 Difference of active sets 5 5 2 25 3 35 outer iteration Fig. 7. Difference in active sets for P and P2; solid line infeasible case, dashed line feasible case. In connection with the convergence speed it is of interest how the detection process of the correct active set works. For the rather aggressive γ-updates used in Algorithm EP the difference between two successive active sets is zero typically only in the last iteration. However, if a less aggressive strategy for updating γ is used, then it is to be expected, that the difference of active sets might become zero earlier along the iteration. In Figure 7, for the strategy γ k+ =2γ k, we show the difference of successive active sets; i.e., the vertical axis relates to the number of grid points that are in A k+ but not in A k and vice versa. We detect that for the infeasible case there exists an iteration index k after which the difference is constantly zero. This behavior is a strong indication that the correct active set was detected. It suggests that we fix this set A k and set ȳ A k = ψ A k, I k =\A k, and λ I k =. Then one computes ȳ I k and λ A k such that a(ȳ, v)+ λ, v H,H =(f,v) for all v H (), and checks whether (ȳ, λ) satisfies (7). If this is the case, then the solution is

82 MICHAEL HINTERMÜLLER AND KARL KUNISCH found; otherwise γ k is updated and the iteration continued. If we apply this technique for P in the infeasible case, then the algorithm stops at iteration 5 (35 inner iterations) with the exact discrete solution, as compared to 28 outer and 47 inner iterations without the additional stopping rule. There were four iterations where the additional system solve was necessary but without obtaining the numerical solution. Hence, with respect to system solves, the amount of work drops from 47 solves to 39 (= 35 + 4). A similar observation is true for P2 and P3. In the feasible case, however, this strategy yields no reduction of iterations. Here, typically the correct active set is determined in the last iteration (for large enough γ). The dependence of the iteration number on the mesh size of the discretization for P is depicted in Table 6.2 (those for P2 and P3 are similar). In parenthesis we show the number of inner iterations. The results clearly indicate that the outer iterations are mesh independent, while the number of inner iterations increases as the mesh size decreases. In the third row we display the results obtained by applying Algorithm A for the solution of the unregularized problem (P) with data according to P. If we compare these results with those of the infeasible exact path-following algorithm, we find that for sufficiently small mesh sizes h the infeasible version of Algorithm EP requires significantly fewer iterations than does Algorithm A, which is also an infeasible algorithm. Also, the number of iterations required by Algorithm A exhibits a relatively strong dependence on h when compared to Algorithm EP in the infeasible case. Similar observations apply also to P2 and P3. This shows that taking into account the function space theoretic properties when regularizing problem (P) results in an algorithmic framework which performs stably with respect to decreasing mesh size of the discretization. 6 γ values; P 4 2 8 6 4 2.5 2 2.5 3 3.5 4 4.5 5 outer iteration Table 6.2 Comparison of iteration counts for different mesh sizes. Mesh size h Version /6 /32 /64 /28 /256 EP feasible 5(9) 5(23) 5(3) 5(44) 5(72) EP infeasible 4(8) 4() 4(3) 4(5) 4(9) Algorithm A 4 8 4 26 48 inner iteration 8 6 4 2 γ values; P2.5 2 2.5 3 3.5 4 outer iteration 2 γ values; P3 8 6 4 2 2 3 4 outer iteration Fig. 8. γ-updates; solid line infeasible case, dashed line feasible case. From the plots in Figure 8, where the y-axis again has a logarithmic scale, it can be seen that our strategy (45) produces a rapidly increasing sequence

PATH-FOLLOWING METHODS 83 {γ k }. The plots in Figure 8 depict the increase of γ k as a function of the iteration number. The question arises of whether one could increase γ more rapidly. Numerical examples employing an ad hoc strategy show that if γ is increased too quickly, then the numerical error may prevent the residuals of the first order system from dropping below ɛ M. This effect is due to the ill-conditioning of the linear systems for large γ. On the other hand, small increases in γ result in a slow convergence speed of Algorithm EP. Further, in our test runs and as can be seen from Figure 8, the feasible version of Algorithm EP is less aggressive in enlarging γ k. 6.2. Inexact path-following. While exact path-following is primarily of theoretical interest, the development of inexact path-following techniques that keep the number of iterations as small as possible is of more practical importance. The strategy in the previous section relies on the fact that for every γ k the corresponding point on the primal-dual path is computed. This, however, is not the case for inexact techniques and, as a consequence, a different update strategy for the path parameter γ is necessary. A common concept in inexact path-following methods is based on the definition of an appropriate neighborhood of the path; see, e.g., [3] and the references therein for a noninterior neighborhood-based path-following method, or [5, 6, 8, 9] for path-following techniques related to interior point methods. It is typically required that the primal-dual iterates stay within the neighborhood of the path, with the goal to reduce the computational burden while still maintaining convergence of the method. We define (5a) (5b) rγ(y, λ) = Δy + λ f H, rγ(y, 2 λ) = λ max(,λ+ γ(y ψ)) H, and the neighborhood { N (γ) := (y, λ) H () L 2 () : (rγ(y, λ),rγ(y, 2 λ)) 2 τ } (52) γ in the infeasible case and { N (γ) := (y, λ) H () L 2 () : (rγ(y, λ),rγ(y, 2 λ)) 2 (53) } J(y; γ) γ in the feasible case. Above, τ>denotes some fixed parameter. Note that adding the condition γj(y; γ) in (52) yields no further restriction, since this condition is automatically satisfied by the structure of J(y; γ). We also point out that the conditions on the derivative of J(y; γ) are included in (52) and (53), respectively, in order to qualitatively capture (up to first order) the analytical properties of the primal-dual path. Next we specify our framework for an inexact path-following algorithm. Algorithm IP. (i) Initialize γ according to (49) in the infeasible case or (5) in the feasible case; set k :=. τ γ