RESEARCH ARTICLE. A strategy of finding an initial active set for inequality constrained quadratic programming problems

Similar documents
1.The anisotropic plate model

Scalable BETI for Variational Inequalities

Short title: Total FETI. Corresponding author: Zdenek Dostal, VŠB-Technical University of Ostrava, 17 listopadu 15, CZ Ostrava, Czech Republic


Constrained Minimization and Multigrid

5 Handling Constraints

Lecture 3. Optimization Problems and Iterative Algorithms

Affine covariant Semi-smooth Newton in function space

A hybrid domain decomposition method based on one-level FETI and BDDC algorithms

Lectures Notes Algorithms and Preconditioning in PDE-Constrained Optimization. Prof. Dr. R. Herzog

Domain Decomposition Preconditioners for Spectral Nédélec Elements in Two and Three Dimensions

20. A Dual-Primal FETI Method for solving Stokes/Navier-Stokes Equations

On the Local Quadratic Convergence of the Primal-Dual Augmented Lagrangian Method

An Infeasible Interior-Point Algorithm with full-newton Step for Linear Optimization

1 Non-negative Matrix Factorization (NMF)

Primal-Dual Interior-Point Methods for Linear Programming based on Newton s Method

Linear algebra issues in Interior Point methods for bound-constrained least-squares problems

MS&E 318 (CME 338) Large-Scale Numerical Optimization

FETI domain decomposition method to solution of contact problems with large displacements

1 Computing with constraints

Numerical Methods for PDE-Constrained Optimization

Lecture 18: Optimization Programming

CSC Linear Programming and Combinatorial Optimization Lecture 10: Semidefinite Programming

Interior Point Methods in Mathematical Programming

CSC 576: Linear System

Basic Concepts in Linear Algebra

March 8, 2010 MATH 408 FINAL EXAM SAMPLE

An Iterative Descent Method

Structural and Multidisciplinary Optimization. P. Duysinx and P. Tossings

Key words. preconditioned conjugate gradient method, saddle point problems, optimal control of PDEs, control and state constraints, multigrid method

Coordinate Descent and Ascent Methods

6.252 NONLINEAR PROGRAMMING LECTURE 10 ALTERNATIVES TO GRADIENT PROJECTION LECTURE OUTLINE. Three Alternatives/Remedies for Gradient Projection

1. Introduction. We consider the model problem that seeks an unknown function u = u(x) satisfying

A PREDICTOR-CORRECTOR PATH-FOLLOWING ALGORITHM FOR SYMMETRIC OPTIMIZATION BASED ON DARVAY'S TECHNIQUE

Numerical Methods I Non-Square and Sparse Linear Systems

Review of Basic Concepts in Linear Algebra

Determination of Feasible Directions by Successive Quadratic Programming and Zoutendijk Algorithms: A Comparative Study

Chapter 2. Optimization. Gradients, convexity, and ALS

Iterative Methods for Solving A x = b

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for

Part 4: Active-set methods for linearly constrained optimization. Nick Gould (RAL)

Matrix Derivatives and Descent Optimization Methods

SOR- and Jacobi-type Iterative Methods for Solving l 1 -l 2 Problems by Way of Fenchel Duality 1

Solving Obstacle Problems by Using a New Interior Point Algorithm. Abstract

Linear Solvers. Andrew Hazel

THE solution of the absolute value equation (AVE) of

Lecture 11: CMSC 878R/AMSC698R. Iterative Methods An introduction. Outline. Inverse, LU decomposition, Cholesky, SVD, etc.

An interior-point gradient method for large-scale totally nonnegative least squares problems

Numerical Optimization of Partial Differential Equations

Numerical optimization

Optimization. Escuela de Ingeniería Informática de Oviedo. (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30

Numerical Optimization Professor Horst Cerjak, Horst Bischof, Thomas Pock Mat Vis-Gra SS09

Review of matrices. Let m, n IN. A rectangle of numbers written like A =

Constrained Optimization Theory

Copositive Plus Matrices

Newton s Method. Ryan Tibshirani Convex Optimization /36-725

Dual methods and ADMM. Barnabas Poczos & Ryan Tibshirani Convex Optimization /36-725

A DECOMPOSITION PROCEDURE BASED ON APPROXIMATE NEWTON DIRECTIONS

Functions of Several Variables

ON A CLASS OF NONSMOOTH COMPOSITE FUNCTIONS

Adaptive Coarse Space Selection in BDDC and FETI-DP Iterative Substructuring Methods: Towards Fast and Robust Solvers

Numerical optimization. Numerical optimization. Longest Shortest where Maximal Minimal. Fastest. Largest. Optimization problems

Uniform Boundedness of a Preconditioned Normal Matrix Used in Interior Point Methods

Research Article Solving the Matrix Nearness Problem in the Maximum Norm by Applying a Projection and Contraction Method

An interior-point stochastic approximation method and an L1-regularized delta rule

9.1 Preconditioned Krylov Subspace Methods

Numerical algorithms for one and two target optimal controls

Research Article Residual Iterative Method for Solving Absolute Value Equations

Semidefinite Programming Basics and Applications

A semidefinite relaxation scheme for quadratically constrained quadratic problems with an additional linear constraint

Examination paper for TMA4180 Optimization I

Primal-Dual Interior-Point Methods. Ryan Tibshirani Convex Optimization

On Generalized Primal-Dual Interior-Point Methods with Non-uniform Complementarity Perturbations for Quadratic Programming

Quadratic Programming

Multispace and Multilevel BDDC. Jan Mandel University of Colorado at Denver and Health Sciences Center

Numerical Methods for Large-Scale Nonlinear Systems

Multilevel and Adaptive Iterative Substructuring Methods. Jan Mandel University of Colorado Denver

On the interior of the simplex, we have the Hessian of d(x), Hd(x) is diagonal with ith. µd(w) + w T c. minimize. subject to w T 1 = 1,

Conjugate Gradient (CG) Method

Preconditioning via Diagonal Scaling

SOME PRACTICAL ASPECTS OF PARALLEL ADAPTIVE BDDC METHOD

Algorithms for PDE-Constrained Optimization

Uses of duality. Geoff Gordon & Ryan Tibshirani Optimization /

Optimisation in Higher Dimensions

Primal/Dual Decomposition Methods

NONLINEAR CLASSIFICATION AND REGRESSION. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition

You should be able to...

Next topics: Solving systems of linear equations

Solving Linear Systems

A FULL-NEWTON STEP INFEASIBLE-INTERIOR-POINT ALGORITHM COMPLEMENTARITY PROBLEMS

Numerisches Rechnen. (für Informatiker) M. Grepl P. Esser & G. Welper & L. Zhang. Institut für Geometrie und Praktische Mathematik RWTH Aachen

Solution Methods. Richard Lusby. Department of Management Engineering Technical University of Denmark

Newton-Krylov-Schwarz Method for a Spherical Shallow Water Model

Optimal Preconditioners for Interval Gauss Seidel Methods

Numerical Optimization Prof. Shirish K. Shevade Department of Computer Science and Automation Indian Institute of Science, Bangalore

Nonlinear Programming

Proximal Newton Method. Ryan Tibshirani Convex Optimization /36-725

, b = 0. (2) 1 2 The eigenvectors of A corresponding to the eigenvalues λ 1 = 1, λ 2 = 3 are

Absolute value equations

A globally and R-linearly convergent hybrid HS and PRP method and its inexact version with applications

Transcription:

Optimization Methods and Software Vol. 00, No. 00, July 200, 8 RESEARCH ARTICLE A strategy of finding an initial active set for inequality constrained quadratic programming problems Jungho Lee Computer Science and Mathematics Division, Oak Ridge National Laboratory, Oak Ridge, TN, USA (Received July 200) It is well known that an active set method can often be slow due to a poor guess of the optimal active set. In this paper, we provide an efficient strategy of finding an initial active set and an initial guess for a quadratic programming problem with inequality constraints. We use a combination of this strategy with a primal-dual active set method as a nonsmooth Newton method [2] to solve a model problem. Keywords: quadratic programming; inequality constraints; initial active set; primal-dual active set method; nonsmooth Newton method; domain decomposition methods AMS Subject Classification: 49M29; 65N55; 90C33; 90C59. Introduction The idea of an active set method, or a working set method, is to reduce an inequality constrained problem to a sequence of equality constrained problems; see [, Chapter 5], [5]. It starts with an initial guess as to which face of the boundary of the feasible region the optimal solution lies on, or equivalently, the set of constraints that are satisfied as equalities at the solution. Such a guess is called an active set, or a working set. In the rest of this paper we are going to use the terminology an active set and an active set method. An active set method has a finite termination property but also has a very pessimistic upper bound on the number of iterations needed to reach the correct solution; this is due to the fact that the number of possible active sets is 2 I, where I is the number of inequality constraints. This is a phenomenon known as the combinatorial difficulty [5, Chapter 5]. On the other hand, it is possible to cast an inequality constrained quadratic programming problem as a nonlinear equation so that a semismooth Newton method can be used; see [2] and the references therein. A semismooth Newton method has a superlinear convergence property, but such a convergence is guaranteed only when the initial guess is sufficiently close to the right solution. One can easily see that finding a good initial guess, again, becomes an issue of critical importance. In this paper, we propose a strategy for finding an initial active set and an initial Computer Science and Mathematics Division, Oak Ridge National Laboratory, PO BOX 2008 MS664, Oak Ridge, TN 3783-664, USA. Telephone: -865-574-3653, Fax: -865-24-48, Email: jungho@ornl.gov ISSN: 055-6788 print/issn 029-4937 online c 200 Taylor & Francis DOI: 0.080/055678xxxxxxxxxxxxx http://www.informaworld.com

2 J. Lee guess for a quadratic programming problem with inequality constraints. This work originated in the Ph.D. thesis of the author [3] concerning domain decomposition methods for contact problems. The rest of the paper is organised as follows. In Section 2, we provide the primal and the dual formulations of an inequality constrained quadratic programming problem that we consider in this paper. We also outline our strategy of finding an initial active set and an initial guess. In Section 3, we provide the details of the strategy. In Section 4, we describe a primal-dual active set method as a nonsmooth Newton method [2]. In Section 5, we provide the numerical results of a model problem using a combination of our strategy of finding an initial active set and an initial guess and the primal-dual active set method [4]. We provide a few concluding remarks in Section 6... Some notation We adopt the notation similar to the one established in [, Chapter ]. In the rest of the paper, the ith component of a vector v R n is denoted by [v] i. The relations between two vectors u, v R n are defined component-wise; that is, u = v and u v are equivalent to [u] i = [v] i, i and [u] i [v] i, i, respectively. For given u R n, the vector u + R n is defined by [u + ] i = max([u] i, 0), i. The zero vector in R n is denoted by o n ; when the dimension of the zero vector is clear from the context, we drop the subscript n. Analogously, the (i, j)th component of a matrix A R m n is denoted by [A] ij. The zero matrix in R m n is denoted by O mn ; when the dimension of the matrix is clear from the context, we drop the subscript mn. 2. Primal and Dual formulations of a quadratic programming problem with inequality constraints We consider a quadratic programming problem of the following form, where A R n n, b R n,b R m n, m n, and A is symmetric and positive-definite: min x R n 2 xt Ax b T x, with Bx o. () The dual form of () is as follows (see, for instance, []): min λ R m 2 λt BA B T λ d T λ, with λ o, (2) where d = BA b. The dual form (2) has so-called box constraints, i.e., constraints of the form a [λ] i b, whereas the primal form () does not. We propose the following strategy to determine an initial active set for the problem (): Algorithm. Our strategy of finding an initial active set. () Solve the unconstrained version of (2) with a preconditioned conjugate gradient (PCG) method, using a zero initial guess. (2) Apply a projection-like operator, ˆP, to the resulting unconstrained solution, λ. (3) Define the initial active set as the set of constraints satisfied by ˆPλ.

Optimization Methods and Software 3 3. Details of the strategy of finding an initial active set/guess We first solve the unconstrained version of (2): min λ R m 2 λt BA B T λ d T λ, (3) and denote the solution of (3) by λ. In Figure, we illustrate the projection of λ in the original coordinate system with the standard basis {e i } m i= and the transformed coordinate system defined by {ē i : M /2 ē i = e i } m i=, where M := BA B T, in two dimensions. The concentric ellipses on the left in Figure indicate the level sets of f(λ) := 2 λt BA B T λ d T λ, whereas the concentric circles on the right in Figure indicate the level sets of the transformed function f( λ) := 2 λ T M /2 BA B T M /2 λ d T M /2 λ = 2 λ T λ d T M /2 λ. The feasible region Ω B := {λ : λ o} has been transformed into { λ : M /2 λ o}. We make the following key observation: whereas the projection of λ onto Ω B in the original coordinate system does not necessarily coincide with λ, the minimiser of the inequality constrained problem (2), the projection of λ onto { λ : M /2 λ o} in the transformed coordinate system coincides with λ. In practice the preconditioner will not be equal to the system matrix BA B T and thus we cannot expect this to happen, but we can still expect the projection of λ in the transformed coordinate system to be a better approximation of λ than the projection in the original coordinate system. We now explain what we mean by a projection-like operator ˆP. Note that Ω B = {λ : λ, e i 0, i} and the orthogonal projection of a vector v = i v ie i onto Ω B can be written as i max( v, ei, 0)e i. Motivated by this, letting M /2 e i new := e i, we define Pw where w = i w ie i new as Pw := i ( e i ) new e i max w, e i new, 0 new e i new. Let u old = λ and let u new be defined by the relation M /2 u new = u old. (4) Recalling the assumption that λ was obtained by a PCG method with a zero initial guess (Algorithm ), there exists a vector u such that u old = λ = M u. (5) This u can be obtained by adding just a few lines to the original PCG algorithm and without any additional computational cost. Our operator ˆP is defined as follows: given u old, we apply P to the corresponding u new. We then express the resulting vector in terms of the standard basis by leftmultiplying it with M /2 : M /2 i= max( u new, e i new, 0) e i new, e i e i new new

4 J. Lee max( M /2 u old,m /2 e i old, 0) M /2 e i old,m/2 e i old M /2 e i old = M /2 i= = max( Mu old, e i old, 0) [M] i= ii = i= e i old max( u, e i old, 0) e i old [M]. (6) ii Noting that the entries of M are easily available while those of M are not, we replace /[M] ii of (6) by [M ] ii : i= max( u, e i old, 0) e i old [M] [M ] diag, u + =: ˆλ =: ˆPλ, (7) ii where M diag Rm is a vector consisting of the diagonal elements of M. We recall the KKT conditions for (), which are satisfied by an optimal pair (x, λ): Bx o, λ o, λ T (Bx) = 0, Ax b + B T λ = o. (8) The second and the third equations of (8) indicate that [λ] i > 0 implies [Bx] i = 0. This motivates us to set I 0 = {i : [ˆλ] i > 0}, λ 0 = ˆλ. (9) λ proj (λ ) ΩB ~ ΩB λ λ proj (λ )=λ ~ Figure. The projection of λ onto the feasible region in original and transformed coordinates, respectively. When the preconditioner is equal to the inverse of the system matrix (as shown in right), the projection of the solution of the unconstrained problem, λ, onto the feasible region coincides with the solution of the constrained problem, λ. Therefore we can expect proj(λ ) λ with a good preconditioner. 4. A primal-dual active set method as a semismooth Newton method In this section, we briefly describe a primal-dual active set method; for details, see [2] and the references therein.

Optimization Methods and Software 5 Again, we consider the quadratic programming problem with inequality constraints (), which is equivalent to the following problem { Ax + B T λ = b, Bx o, λ o, λ T Bx = 0. (0) The complementarity condition given in the second line is equivalent to C(x, λ, c) := λ max(o, λ + cbx) = o, () for each c > 0. Here, the max function is to be understood component-wise. In the following, we will suppress the dependence on c and use the notation C(x, λ), for the sake of brevity; we will also use c = and choose not to investigate different choices of c. The system (0) can thus be expressed as the following nonlinear system of equations: { Ax + B T λ = b, C(x, λ) = o. It follows that a (semismooth) Newton step for the nonlinear system (2) is [ ] [ ] A B T δx k B Ak I Ik δλ k = [ f (Ax k + B T λ k ] ) C(x k, λ k ) (2) (3) and x k+ = x k + δx k, λ k+ = λ k + δλ k, (4) where I k = {i : [λ k + Bx k ] i 0}, A k = {i : [λ k + Bx k ] i > 0}, (5) and B Ak results from replacing the rows of B for which the index does not belong to A k with zero row vectors. The matrix I Ik is defined similarly. We can rewrite the second equation of (3) as follows: [Bx k ] i = [Bx k ] i, i A k, and [δλ k ] i = [λ k ] i, i I k. (6) We also rewrite the first equation: Ax k + B T A k (δλ k ) Ak + B T I k (δλ k ) Ik = b (Ax k + B T A k (λ k ) Ak + B T I k (λ k ) Ik ), (7) where (δλ k ) Ak results from replacing the components of δλ k for which the index does not belong to A k with zeros; (δλ k ) Ik, etc., are defined similarly. This equation is equivalent to Ax k + B T A k (δλ k ) Ak = b (Ax k + B T A k (λ k ) Ak ), (8) due to (6). Consequently, we can rewrite the Newton step defined by (3) and (4) as [ A B T Ak O B Ak ] [ ] δx k δλ k = [ b (Ax k + B T A (λ k ] ) Ak ) B Ak x k (9)

6 J. Lee and x k+ = x k + δx k, λ k+ = λ k + δλ k, where [δλ k ] i = [λ k ] i, i I k. (20) The following algorithm is simply the primal-dual active set strategy [2] combined with our strategy of finding an initial active set. Algorithm 2. Primal-Dual active set method combined with our strategy of finding an initial active set. () Choose λ 0 as described in Section 3. Set x 0 = 0. Set k = 0. (2) Set I k = {i : [λ k + Bx k ] i 0}, A k = {i : [λ k + Bx k ] i > 0}. (3) Solve [ A B T Ak O B Ak ] [ ] [ x k+ b = o] λ k+ and set λ k+ = o on I k. (4) Stop if A k+ = A k and I k+ = I k. Otherwise return to 2. (2) 5. Numerical examples In this section, we provide the numerical results of Algorithm 2 applied to the following model problem, taken from [, Chapter 8]: 2 ( ) min u i 2 dx fu i dx 2 i= Ω i Ω i where u i H (Ω i ), i =, 2, Ω = (0, ) (0, ), Ω 2 = (, 2) (0, ), u = 0 on Γ u = {0} (0, ), u 2 u 0 on Γ c = {} (0, ). (22) The results we provide here are taken from [4]. As mentioned in [, Chapter 8], we can view the solution of this problem as the displacement of two membranes, Ω and Ω 2, under a body force. The left edge of the right membrane Ω 2 is not allowed to go below the right edge of the left membrane Ω, and the left edge of the left membrane is fixed. This problem is coercive and thus has a unique solution; see [, Chapter 8] and the references therein. We use a domain decomposition approach to solve this problem, in particular, a hybrid algorithm described in [4]. The membranes Ω and Ω 2 are decomposed into N N subdomains, which in turn are divided into n n bilinear elements. The side lengths of a subdomain and of an element are H := /N and h := /(Nn), respectively. The finite element discretisation of the problem (22) with a hybrid domain decomposition method [4] is a quadratic programming problem with inequality constraints of the form (), and we solve it by combining our strategy of finding an initial active set and an initial guess and a primal-dual active set strategy. The results are summarised in Table 5. It is shown that for the combinations of H and h we tried, the number of outer iterations of the primal-dual active set method is at most 2 when combined with our strategy of finding an initial active set and an initial guess.

Optimization Methods and Software 7 Table. Results: primal-dual active set method + hybrid method. outer it. denotes the number of outer iterations of the primal-dual active set method; inner it. denotes the number of iterations needed to solve the inner minimisation problems by the PCR (preconditioned conjugate residual) method, until the norm of the residual has been reduced by 0 5, on the active faces identified in the outer iterations. total it. denotes the total number of inner iterations. N sub (/H) H/h N dof (λ) N dof (total) outer it. inner it. total it. 6(4) 4 7 56 2 6 6 32 6(4) 8 33 245 2 20 9 39 6(4) 2 49 4753 2 22 20 42 6(4) 6 65 8385 2 26 24 50 64(8) 4 33 245 2 8 7 35 64(8) 8 65 8385 23 23 64(8) 2 97 872 27 27 64(8) 6 29 3353 29 29 44(2) 4 49 4753 9 9 44(2) 8 97 872 2 24 22 46 44(2) 2 45 4905 2 28 24 52 44(2) 6 93 74305 2 30 27 57 256(6) 4 65 8385 9 9 256(6) 8 29 3353 26 26 256(6) 2 93 74305 28 28 256(6) 6 257 384 32 32 6. Concluding remarks In this paper, we have considered an efficient strategy of finding an initial active set and an initial guess for a quadratic programming problem with inequality constraints. Numerical results from the application of this strategy to a simple model problem were presented, which show that our strategy finds the optimal active set quite accurately for the cases that were considered. The effectiveness of this strategy in more complicated problems, such as contact problems in linear elasticity, remains to be shown. Acknowledgements The author would like to thank Olof Widlund and Zdeněk Dostál for carefully reading the manuscript and for their encouragement. This author s work was supported in part by the U.S. Department of Energy under contracts DE-FG02-06ER2578 and DE-FC02-0ER25482 and in part by National Science Foundation grant DMS- 05325. This submission was sponsored by a contractor of the United States Government under contract DE-AC05-00OR22725 with the United States Department of Energy. The United States Government retains, and the publisher, by accepting this submission for publication, acknowledges that the United States Government retains, a nonexclusive, paid-up, irrevocable, worldwide license to publish or reproduce the published form of this submission, or allow others to do so, for United States Government purposes. References [] Z. Dostál, Optimal quadratic programming algorithms. With applications to variational inequalities, Springer Optimization and Its Applications Vol. 23, Springer, New York, 2009. [2] M. Hintermüller, K. Ito, and K. Kunisch, The Primal-Dual Active Set Strategy as a Semismooth Newton Method, SIAM J. on Optimization 3 (2002), pp. 865 888.

8 Taylor & Francis and I.T. Consultant [3] J. Lee, A Hybrid Domain Decomposition Method and its Applications to Contact Problems, Ph.D. diss., Courant Institute of Mathematical Sciences, New York University, New York, NY, 2009. [4] J. Lee, A hybrid domain decomposition method based on one-level FETI and BDDC algorithms., preprint (200), submitted to SIAM J. Sci. Comput. [5] J. Nocedal and S.J. Wright, Numerical Optimization, Springer Series in Operations Research, Springer, 2000.