Relaxation schemes for mathematical programs with switching constraints

Similar documents
Fakultät für Mathematik und Informatik

FIRST- AND SECOND-ORDER OPTIMALITY CONDITIONS FOR MATHEMATICAL PROGRAMS WITH VANISHING CONSTRAINTS 1. Tim Hoheisel and Christian Kanzow

A FRITZ JOHN APPROACH TO FIRST ORDER OPTIMALITY CONDITIONS FOR MATHEMATICAL PROGRAMS WITH EQUILIBRIUM CONSTRAINTS

AN ABADIE-TYPE CONSTRAINT QUALIFICATION FOR MATHEMATICAL PROGRAMS WITH EQUILIBRIUM CONSTRAINTS. Michael L. Flegel and Christian Kanzow

First-order optimality conditions for mathematical programs with second-order cone complementarity constraints

Constraint qualifications for nonlinear programming

Priority Programme 1962

ON LICQ AND THE UNIQUENESS OF LAGRANGE MULTIPLIERS

Implications of the Constant Rank Constraint Qualification

Preprint ISSN

First order optimality conditions for mathematical programs with second-order cone complementarity constraints

1. Introduction The nonlinear complementarity problem (NCP) is to nd a point x 2 IR n such that hx; F (x)i = ; x 2 IR n + ; F (x) 2 IRn + ; where F is

Numerical Optimization

Mingbin Feng, John E. Mitchell, Jong-Shi Pang, Xin Shen, Andreas Wächter

Mathematical programs with complementarity constraints in Banach spaces

Complementarity Formulations of l 0 -norm Optimization Problems

The limiting normal cone to pointwise defined sets in Lebesgue spaces

INTERIOR-POINT METHODS FOR NONCONVEX NONLINEAR PROGRAMMING: CONVERGENCE ANALYSIS AND COMPUTATIONAL PERFORMANCE

Constrained Optimization Theory

ON GENERALIZED-CONVEX CONSTRAINED MULTI-OBJECTIVE OPTIMIZATION

Complementarity Formulations of l 0 -norm Optimization Problems

Optimality Conditions for Constrained Optimization

ON AUGMENTED LAGRANGIAN METHODS WITH GENERAL LOWER-LEVEL CONSTRAINTS. 1. Introduction. Many practical optimization problems have the form (1.

Nonlinear Programming, Elastic Mode, SQP, MPEC, MPCC, complementarity

1. Introduction. We consider mathematical programs with equilibrium constraints in the form of complementarity constraints:

Characterizations of Solution Sets of Fréchet Differentiable Problems with Quasiconvex Objective Function

Optimization over Sparse Symmetric Sets via a Nonmonotone Projected Gradient Method

The Relation Between Pseudonormality and Quasiregularity in Constrained Optimization 1

5 Handling Constraints

SOME STABILITY RESULTS FOR THE SEMI-AFFINE VARIATIONAL INEQUALITY PROBLEM. 1. Introduction

An Enhanced Spatial Branch-and-Bound Method in Global Optimization with Nonconvex Constraints

On smoothness properties of optimal value functions at the boundary of their domain under complete convexity

Interior Methods for Mathematical Programs with Complementarity Constraints

Combinatorial Structures in Nonlinear Programming

Priority Programme 1962

TMA 4180 Optimeringsteori KARUSH-KUHN-TUCKER THEOREM

AN AUGMENTED LAGRANGIAN AFFINE SCALING METHOD FOR NONLINEAR PROGRAMMING

Topological properties

A Sequential NCP Algorithm for Solving Equilibrium Problems with Equilibrium Constraints

1 Directional Derivatives and Differentiability

Semi-infinite programming, duality, discretization and optimality conditions

Gerd Wachsmuth. January 22, 2016

arxiv: v2 [math.oc] 23 Nov 2016

On sequential optimality conditions for constrained optimization. José Mario Martínez martinez

Structural and Multidisciplinary Optimization. P. Duysinx and P. Tossings

Finite Convergence for Feasible Solution Sequence of Variational Inequality Problems

Necessary optimality conditions for optimal control problems with nonsmooth mixed state and control constraints

Second Order Optimality Conditions for Constrained Nonlinear Programming

IE 5531: Engineering Optimization I

Robust error estimates for regularization and discretization of bang-bang control problems

CONVERGENCE ANALYSIS OF AN INTERIOR-POINT METHOD FOR NONCONVEX NONLINEAR PROGRAMMING

Constraint Qualifications and Stationarity Concepts for Mathematical Programs with Equilibrium Constraints

ARE202A, Fall Contents

Linear Algebra (part 1) : Vector Spaces (by Evan Dummit, 2017, v. 1.07) 1.1 The Formal Denition of a Vector Space

system of equations. In particular, we give a complete characterization of the Q-superlinear

Lecture 13: Constrained optimization

UNDERGROUND LECTURE NOTES 1: Optimality Conditions for Constrained Optimization Problems

New Class of duality models in discrete minmax fractional programming based on second-order univexities

Mathematical Programs with Complementarity Constraints in the Context of Inverse Optimal Control for Locomotion

Technische Universität Dresden Herausgeber: Der Rektor

Preprint Patrick Mehlitz, Gerd Wachsmuth Weak and strong stationarity in generalized bilevel programming and bilevel optimal control

Congurations of periodic orbits for equations with delayed positive feedback

INTRODUCTION TO NETS. limits to coincide, since it can be deduced: i.e. x

IE 5531: Engineering Optimization I

On Augmented Lagrangian Methods with General Lower-Level Constraints

Some Properties of the Augmented Lagrangian in Cone Constrained Optimization

Date: July 5, Contents

CHAPTER 7. Connectedness

Thus, X is connected by Problem 4. Case 3: X = (a, b]. This case is analogous to Case 2. Case 4: X = (a, b). Choose ε < b a

c 2012 Society for Industrial and Applied Mathematics

Normally admissible stratifications and calculation of normal cones to a finite union of polyhedral sets

ON A CLASS OF NONSMOOTH COMPOSITE FUNCTIONS

Constraint Identification and Algorithm Stabilization for Degenerate Nonlinear Programs

Examination paper for TMA4180 Optimization I

1. Introduction. We consider the mathematical programming problem

Exhaustive Classication of Finite Classical Probability Spaces with Regard to the Notion of Causal Up-to-n-closedness

Optimality, identifiability, and sensitivity

arxiv: v1 [math.oc] 30 Mar 2017

MODIFYING SQP FOR DEGENERATE PROBLEMS

Identifying Active Constraints via Partial Smoothness and Prox-Regularity

GENERALIZED CONVEXITY AND OPTIMALITY CONDITIONS IN SCALAR AND VECTOR OPTIMIZATION

Some new facts about sequential quadratic programming methods employing second derivatives

CONSTRAINT QUALIFICATIONS, LAGRANGIAN DUALITY & SADDLE POINT OPTIMALITY CONDITIONS

AN INTERIOR-POINT METHOD FOR NONLINEAR OPTIMIZATION PROBLEMS WITH LOCATABLE AND SEPARABLE NONSMOOTHNESS

Part III. 10 Topological Space Basics. Topological Spaces

Lecture 9 Monotone VIs/CPs Properties of cones and some existence results. October 6, 2008

On Augmented Lagrangian methods with general lower-level constraints

WHEN ARE THE (UN)CONSTRAINED STATIONARY POINTS OF THE IMPLICIT LAGRANGIAN GLOBAL SOLUTIONS?

University of California. Berkeley, CA fzhangjun johans lygeros Abstract

Optimization. A first course on mathematics for economists

A DC (DIFFERENCE OF CONVEX FUNCTIONS) APPROACH OF THE MPECS. Matthieu Marechal. Rafael Correa. (Communicated by the associate editor name)

An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization

Optimality, identifiability, and sensitivity

Least Sparsity of p-norm based Optimization Problems with p > 1

p. 5 First line of Section 1.4. Change nonzero vector v R n to nonzero vector v C n int([x, y]) = (x, y)

From Fractional Brownian Motion to Multifractional Brownian Motion

Nonlinear Chance Constrained Problems: Optimality Conditions, Regularization and Solvers

c 2000 Society for Industrial and Applied Mathematics

A Geometric Framework for Nonconvex Optimization Duality using Augmented Lagrangian Functions

IE 5531: Engineering Optimization I

Transcription:

Relaxation schemes for mathematical programs with switching constraints Christian Kanzow Patrick Mehlitz Daniel Steck September 6, 2018 Switching-constrained optimization problems form a dicult class of mathematical programs since their feasible set is almost disconnected while standard constraint qualications are likely to fail at several feasible points. That is why the application of standard methods from nonlinear programming does not seem to be promising in order to solve such problems. In this paper, we adapt the relaxation method from Kanzow and Schwartz (SIAM J. Optim., 23(2):770798, 2013) for the numerical treatment of mathematical programs with complementarity constraints to the setting of switching-constrained optimization. It is shown that the proposed method computes M-stationary points under mild assumptions. Furthermore, we comment on other possible relaxation approaches which can be used to tackle mathematical programs with switching constraints. As it turns out, adapted versions of Scholtes' global relaxation scheme as well as the relaxation scheme of Steensen and Ulbrich only nd W-stationary points of switching-constrained optimization problems in general. Some computational experiments visualize the performance of the proposed relaxation method. Keywords: Constraint qualications, Mathematical program with switching constraints, Relaxation methods, Global convergence. MSC: 65K05, 90C30, 90C33. University of Würzburg, Institute of Mathematics, 97074 Würzburg, Germany, kanzow@mathematik.uni-wuerzburg.de, https://www.mathematik.uni-wuerzburg.de/~kanzow/, ORCID: 0000-0003-2897-2509 Brandenburgische Technische Universität CottbusSenftenberg, Institute of Mathematics, 03046 Cottbus, Germany, mehlitz@b-tu.de, https://www.b-tu.de/fg-optimale-steuerung/team/ dr-patrick-mehlitz, ORCID: 0000-0002-9355-850X University of Würzburg, Institute of Mathematics, 97074 Würzburg, Germany, daniel.steck@mathematik.uni-wuerzburg.de, ~steck/, ORCID: 0000-0001-5335-5428 https://www.mathematik.uni-wuerzburg.de/ 1

1 Introduction This paper is dedicated to so-called mathematical programs with switching constraints, MPSCs for short. These are optimization problems of the form f(x) min g i (x) 0, i M, h j (x) = 0, j P, G l (x)h l (x) = 0, l Q, (MPSC) where M := {1,..., m}, P := {1,..., p}, Q := {1,..., q} are index sets and the functions f, g i, h j, G l, H l : R n R are continuously dierentiable for all i M, j P, and l Q. For brevity, g : R n R m, h: R n R p, G: R n R q, and H : R n R q are the mappings which possess the component functions g i (i M), h j (j P), G l (l Q), and H l (l Q), respectively. The last q constraints in (MPSC) force G l (x) or H l (x) to be zero for all l Q, which gives rise to the terminology switching constraints. Switching structures appear frequently in the context of optimal control, see Clason et al. [2017, Gugat [2008, Hante and Sager [2013, Liberzon [2003, Seidman [2013, Wang and Yan [2015, Zuazua [2011, and the references therein, or as a reformulation of so-called either-or constraints, see [Mehlitz, 2018, Section 7. Naturally, (MPSC) is related to other problem classes from disjunctive programming such as mathematical programs with complementarity constraints, MPCCs for short, see Luo et al. [1996, Outrata et al. [1998, or mathematical programs with vanishing constraints, MPVCs for short, see Achtziger and Kanzow [2008, Hoheisel and Kanzow [2008. Indeed, similarly to MPCCs and MPVCs, standard constraint qualications are likely to be violated at the feasible points of (MPSC). Recently, stationarity conditions and constraint qualications for (MPSC) were introduced in Mehlitz [2018. Here, we focus on the computational treatment of (MPSC). Clearly, standard methods from nonlinear programming may run into diculties when applied to (MPSC) due to two reasons: rst, the feasible set of (MPSC) is likely to be disconnected or at least almost disconnected. Secondly, standard regularity conditions like the Mangasarian Fromovitz constraint qualication, MFCQ for short, are likely to fail at the feasible points of (MPSC) under mild assumptions, see [Mehlitz, 2018, Lemma 4.1. Similar issues appear in the context of MPCCs and MPVCs where several dierent relaxation schemes were introduced to overcome these shortcomings, see Hoheisel et al. [2012, 2013 and the references therein. Basically, the idea is to relax the irregular constraints using a relaxation parameter such that the resulting surrogate problems are (regular) standard nonlinear problems which can be tackled by common methods. The relaxation parameter is then iteratively reduced to zero and, in each iteration, a KarushKuhnTucker (KKT) point of the surrogate problem is computed. Ideally, the resulting sequence possesses a limit point and, under some problem-tailored constraint qualication, this point satises a suitable stationarity condition. Furthermore, it is desirable that the relaxed problems satisfy standard constraint qualications in a neighborhood of the limit point under reasonable assumptions. 2

In this paper, we show that the relaxation scheme from Kanzow and Schwartz [2013, which was designed for the numerical investigation of MPCCs, can be adapted for the computational treatment of (MPSC). Particularly, it will be shown that the modied method can be used to nd M-stationary points of (MPSC). By means of examples it is demonstrated that other relaxation methods which are well known from the theory of MPCCs only yield W-stationary points of (MPSC). We present the results of some numerical experiments which show the performance of the proposed method. The remaining part of the paper is structured as follows. In Section 2, we describe the general notation used throughout the paper and recall some fundamental theory on nonlinear programming and switching constraints. Section 3 is dedicated to the main relaxation approach and contains various properties of the resulting algorithm as well as convergence results. In Section 4, we describe how some common regularization methods from MPCCs can be carried over to the switching-constrained setting, and discuss the convergence properties of the resulting algorithms. Section 5 contains some numerical applications, including either-or constraints, switching-constrained optimal control, and semi-continuous optimization problems arising in portfolio optimization. We conclude the paper with some nal remarks in Section 6. 2 Notation and preliminaries 2.1 Basic notation The subsequently introduced tools of variational analysis can be found in Rockafellar and Wets [1998. For a nonempty set A R n, we call A := {y R n x A: x y 0} the polar cone of A. Here, x y denotes the Euclidean inner product of the two vectors x, y R n. It is well known that A is a nonempty, closed, convex cone. For any two sets B 1, B 2 R n, the polarization rule (B 1 B 2 ) = B1 B 2 holds by denition. The polar of a polyhedral cone can be characterized by means of, e.g., Motzkin's theorem of alternatives. Note that we interpret the relations and for vectors componentwise. Lemma 2.1. For matrices C R m n and D R p n, let K R n be the polyhedral cone K := {d R n Cd 0, Dd = 0}. Then K = {C λ + D ρ λ R m, λ 0, ρ R p }. Let A R n be a nonempty set and x A. Then the closed cone T A ( x) := {d R n {x k } k N A {τ k } k N R + : x k x, τ k 0, (x k x)/τ k d} is called tangent or Bouligand cone to A at x. Here, R + := {r R r > 0} denotes the set of all positive reals. 3

The union {v 1,..., v r } {w 1,..., w s } of sets {v 1,..., v r }, {w 1,..., w s } R n is called positive-linearly dependent if there exist vectors α R r, α 0, and β R s which do not vanish at the same time such that 0 = r α i v i + i=1 s β j w j. Otherwise, {v 1,..., v r } {w 1,..., w s } is called positive-linearly independent. Clearly, if the set {v 1,..., v r } is empty, then the above denitions reduce to linear dependence and independence, respectively. The following lemma will be useful in this paper; its proof is similar to that of [Qi and Wei, 2000, Proposition 2.2 and therefore omitted. Lemma 2.2. Let {v 1,..., v r }, {w 1,..., w s } R n be given sets whose union {v 1,..., v r } {w 1,..., w s } is positive-linearly independent. Then there exists ε > 0 such that, for all vectors ṽ 1,..., ṽ r, w 1,..., w s {z R n z 2 ε}, the union {v 1 + ṽ 1,..., v r + ṽ r } {w 1 + w 1,..., w s + w s } is positive-linearly independent. For some vector z R n and an index set I {1,..., n}, z I R I denotes the vector which results from z by deleting all z i with i {1,..., n} \ I. Finally, let us mention that supp z := {i {1,..., n} z i 0} is called the support of the vector z R n. 2.2 Standard nonlinear programs Here, we recall some fundamental constraint qualications from standard nonlinear programming, see, e.g., Bazaraa et al. [1993. Therefore, we consider the nonlinear program j=1 f(x) min g i (x) 0, i M, h j (x) = 0, j P, (NLP) i.e., we forget about the switching constraints in (MPSC) for a moment. Let X R n denote the feasible set of (NLP) and x some point x X. Then I g ( x) := {i M g i ( x) = 0} is called the index set of active inequality constraints at x. Furthermore, the set { } L X( x) := d R n g i ( x) d 0 i I g ( x) h j ( x) d = 0 j P is called the linearization cone to X at x. Obviously, L X( x) is a polyhedral cone, and thus closed and convex. It is well known that T X( x) L X( x) is always satised. The converse inclusion generally only holds under some constraint qualication. In the denition below, we recall several standard constraint qualications which are applicable to (NLP). 4

Denition 2.3. Let x R n be a feasible point of (NLP). Then x is said to satisfy the (a) linear independence constraint qualication (LICQ) if the following vectors are linearly independent: { g i ( x) i I g ( x)} { h j ( x) j P}. (1) (b) MangasarianFromovitz constraint qualication (MFCQ) if the vectors in (1) are positive-linearly independent. (c) constant positive linear dependence condition (CPLD) if, for any sets I I g ( x) and J P such that the gradients { g i ( x) i I} { h j ( x) j J} are positive-linearly dependent, there exists a neighborhood U R n of x such that the gradients { g i (x) i I} { h j (x) j J} are linearly dependent for all x U. (d) Abadie constraint qualication (ACQ) if T X( x) = L X( x). (e) Guignard constraint qualication (GCQ) if T X( x) = L X( x). Note that the following relations hold between the constraint qualications from Definition 2.3: LICQ = MFCQ = CPLD = ACQ = GCQ, see [Hoheisel et al., 2013, Section 2.1 for some additional information. It is well known that the validity of GCQ at some local minimizer x R n of (NLP) implies that the KKT conditions 0 = f( x) + ρ j h j ( x), i I g ( x): λ i 0 i I g ( x) λ i g i ( x) + j P provide a necessary optimality condition. Thus, the same holds for the stronger constraint qualications ACQ, CPLD, MFCQ, and LICQ. 2.3 Mathematical programs with switching constraints The statements of this section are taken from Mehlitz [2018. Let X R n denote the feasible set of (MPSC) and x a point x X. Then the index sets I G ( x) := {l Q G l ( x) = 0, H l ( x) 0}, I H ( x) := {l Q G l ( x) 0, H l ( x) = 0}, I GH ( x) := {l Q G l ( x) = 0, H l ( x) = 0} 5

form a disjoint partition of Q. It is easily seen that MFCQ (and thus LICQ) cannot hold for (MPSC) at x if I GH ( x). Taking a look at the associated linearization cone g i ( x) d 0 i I g ( x) L X ( x) = d R n h j ( x) d = 0 j P G k ( x) d = 0 k I G, ( x) H k ( x) d = 0 k I H ( x) which is always convex, one can imagine that ACQ is likely to fail as well if I GH ( x) since, in the latter situation, T X ( x) might be nonconvex. Note that GCQ may hold for (MPSC) even in the aforementioned context. Due to the inherent lack of regularity, stationarity conditions for (MPSC) which are weaker than the associated KKT conditions were introduced. Denition 2.4. A feasible point x X of (MPSC) is called (a) weakly stationary (W-stationary) if there exist multipliers λ i (i I g ( x)), ρ j (j P), µ l (l Q), and ν l (l Q) which solve the following system: 0 = f( x) + ρ j h j ( x) i I g ( x): λ i 0, i I g ( x) λ i g i ( x) + j P + l Q[ µl G l ( x) + ν l H l ( x), l I H ( x): µ l = 0, l I G ( x): ν l = 0. (b) Mordukhovich-stationary (M-stationary) if it is W-stationary and the associated multipliers additionally satisfy l I GH ( x): µ l ν l = 0. (c) strongly stationary (S-stationary) if it is W-stationary while the associated multipliers additionally satisfy Clearly, the following implications hold: l I GH ( x): µ l = 0 ν l = 0. S-stationarity = M-stationarity = W-stationarity. Moreover, the KKT conditions of (MPSC) are equivalent to the S-stationarity conditions from Denition 2.4. One may check Figure 1 for a geometric interpretation of the Lagrange multipliers associated with the switching conditions from I GH ( x). 6

ν l ν l ν l µ l µ l µ l Figure 1: Geometric illustrations of weak, M-, and S-stationarity for an index l I GH ( x). In order to ensure that one of these stationarity notions plays the role of a necessary optimality condition for (MPSC), suitable problem-tailored constraint qualications need to be valid. For the denition of such conditions, the following so-called tightened nonlinear problem is of interest: f(x) min g i (x) 0, i M, h j (x) = 0, j P, G l (x) = 0, l I G ( x) I GH ( x), H l (x) = 0, l I H ( x) I GH ( x). (TNLP) Note that (TNLP) is a standard nonlinear program. Denition 2.5. Let x X be a feasible point of (MPSC). Then MPSC-LICQ (MPSC- MFCQ) is said to hold for (MPSC) at x if LICQ (MFCQ) holds for (TNLP) at x, i.e., if the vectors [ { g i ( x) i I g ( x)} { h j ( x) j P} are linearly independent (positive-linearly independent). { G l ( x) l I G ( x) I GH ( x)} { H l ( x) l I H ( x) I GH ( x)} It is obvious that MPSC-LICQ is stronger than MPSC-MFCQ. Furthermore, MPSC- LICQ implies standard GCQ for (MPSC) at the reference point. In this paper, we will use another MPSC-tailored constraint qualication called MPSC- NNAMCQ, where NNAMCQ stands for the No Nonzero Abnormal Multiplier Constraint Qualication which has been introduced to investigate optimization problems whose feasible sets are preimages of closed but not necessarily convex sets under continuously dierentiable mappings, see [Rockafellar and Wets, 1998, Section 6.D and Ye and Ye [1997. Clearly, (MPSC) belongs to this problem class as well if one reformulates the switching constraints as (G l (x), H l (x)) C, l Q, 7

with C := {(a, b) R 2 ab = 0}. Denition 2.6. Let x X be a feasible point of (MPSC). Then MPSC-NNAMCQ is said to hold for (MPSC) at x if the following condition is valid: 0 = ρ j h j ( x) λ i g i ( x) + i I g ( x) j P + [ µ l G l ( x) + ν l H l ( x) l Q i I g ( x): λ i 0, = λ = 0, ρ = 0, µ = 0, ν = 0. l I H ( x): µ l = 0, l I G ( x): ν l = 0, l I GH ( x): µ l ν l = 0 It is easy to check that MPSC-NNAMCQ is implied by MPSC-MFCQ. Note that a local minimizer of (MPSC) where MPSC-LICQ holds is an S-stationary point. Furthermore, one can easily check that the associated multipliers which solve the system of S-stationarity are uniquely determined. Under MPSC-MFCQ (and, thus, MPSC- NNAMCQ), a local minimizer of (MPSC) is, in general, only M-stationary. Finally, there exist several problem-tailored constraint qualications for (MPSC) which are weaker than MPSC-MFCQ but also imply that local solutions are M-stationary, see Mehlitz [2018. 3 The relaxation scheme and its convergence properties 3.1 On the relaxation scheme For our relaxation approach, we will make use of the function ϕ: R 2 R dened below: { (a, b) R 2 ab if a + b 0, : ϕ(a, b) := 1 2 (a2 + b 2 ) if a + b < 0. The function ϕ was introduced in Kanzow and Schwartz [2013 to study a relaxation method for the numerical solution of MPCCs. In the following lemma, which parallels [Kanzow and Schwartz, 2013, Lemma 3.1, some properties of ϕ are summarized. Lemma 3.1. (a) The function ϕ is an NCP-function, i.e. it satises (a, b) R 2 : ϕ(a, b) = 0 a 0 b 0 ab = 0. (b) The function ϕ is continuously dierentiable and satises ( ) b if a + b 0, (a, b) R 2 a : ϕ(a, b) = ( ) a if a + b < 0. b 8

For some parameter t 0 as well as indices s S := {1, 2, 3, 4} and l Q, we dene functions Φ s l ( ; t): Rn R via Φ 1 l (x; t) := ϕ(g l(x) t, H l (x) t), Φ 2 l (x; t) := ϕ( G l(x) t, H l (x) t), Φ 3 l (x; t) := ϕ( G l(x) t, H l (x) t), Φ 4 l (x; t) := ϕ(g l(x) t, H l (x) t) for any x R n. interest: Now, we are in position to introduce the surrogate problem of our f(x) min g i (x) 0 h j (x) = 0 i M j P (P(t)) Φ s l (x; t) 0 s S, l Q. The feasible set of (P(t)) will be denoted by X(t). Figure 2 provides an illustration of X(t). H l(x) ( t, t) (t, t) G l(x) ( t, t) (t, t) Figure 2: Geometric interpretation of the relaxed feasible set X(t). Let x X(t) be a feasible point of (P(t)) for some xed parameter t > 0. Later on, it will be benecial to work with the index sets dened below: It,1(x) 00 := {l Q G l (x) = t, H l (x) = t}, It,2(x) 00 := {l Q G l (x) = t, H l (x) = t}, It,1 0+ l(x) = t, H l (x) > t}, It,2 0+ l(x) = t, H l (x) > t}, I t,1 +0 l(x) > t, H l (x) = t}, It,2 0 l(x) < t, H l (x) = t}, It,3(x) 00 := {l Q G l (x) = t, H l (x) = t}, It,4(x) 00 := {l Q G l (x) = t, H l (x) = t}, It,3 0 l(x) = t, H l (x) < t}, It,4 0 l(x) = t, H l (x) < t}, I 0 t,3 (x) := {l Q G l(x) < t, H l (x) = t}, I +0 t,4 (x) := {l Q G l(x) > t, H l (x) = t}. Note that all these sets are pairwise disjoint. The index sets It,1 00 (x), I0+ t,1 (x), and I+0 t,1 (x) subsume the three possible cases where the constraints Φ 1 l (x; t) 0 (l Q) are active. Similarly, the other index sets cover those indices where the constraints Φ s l (x; t) 0 (l Q, s {2, 3, 4}) are active. It follows that an index l Q which does not belong to any of the above sets is inactive for (P(t)) and can therefore be disregarded (locally). In order to address any of the four quadrants separately, we will exploit It,1(x) 0 := It,1(x) 00 It,1 0+ (x) I+0 t,1 (x), I0 t,2(x) := I 00 t,2(x) It,2 0+ (x) I 0 t,2 (x), 9

It,3(x) 0 := It,3(x) 00 It,3 0 (x) I 0 t,3 (x), I0 t,4(x) := I 00 t,4(x) It,4 0 (x) I+0 t,4 (x), i.e., for xed s S, It,s(x) 0 collects all indices l Q where the constraint Φ s l (x; t) 0 is active. For brevity, we set It 00 (x) := It,s(x), 00 s S I 0± t (x) := I 0+ I ±0 t,1 t (x) := I +0 t,1 (x) I0+ t,2 (x) I 0 t,2 (x) I0 t,3 (x) I0 t,4 (x), (x) I 0 t,3 (x) I+0 t,4 (x). Thus, we collect all indices in It 0± (x) where G l (x) { t, t} holds while H l (x) > t is valid. Similarly, I t ±0 (x) comprises all indices where H l (x) { t, t} and G l (x) > t hold true. The set It 00 (x) contains all those indices where G l (x), H l (x) { t, t} is valid. If x X is feasible to (MPSC), x lies in a suciently small neighborhood of x, and t 0 is suciently small, then the following inclusions follow from the continuity of all appearing functions: I 00 t (x) It 0± (x) I G ( x) I GH ( x), t (x) I ±0 (x) I H ( x) I GH ( x). I 00 t In the lemma below, we present explicit formulas for the gradients of Φ s l ( ; t) with l Q and s S. They can be derived exploiting Lemma 3.1 as well as the chain rule. Lemma 3.2. For x R n, t > 0, and l Q, the following formulas are valid: { x Φ 1 l (x; t) = (H l (x) t) G l (x) + (G l (x) t) H l (x) if G l (x) + H l (x) 2t, (G l (x) t) G l (x) (H l (x) t) H l (x) if G l (x) + H l (x) < 2t, x Φ 2 l (x; t) = { (t H l (x)) G l (x) (G l (x) + t) H l (x) if G l (x) + H l (x) 2t, (G l (x) + t) G l (x) (H l (x) t) H l (x) if G l (x) + H l (x) < 2t, x Φ 3 l (x; t) = { (H l (x) + t) G l (x) + (G l (x) + t) H l (x) if G l (x) H l (x) 2t, (G l (x) + t) G l (x) (H l (x) + t) H l (x) if G l (x) H l (x) < 2t, x Φ 4 l (x; t) = { (H l (x) + t) G l (x) + (t G l (x)) H l (x) if G l (x) H l (x) 2t, (G l (x) t) G l (x) (H l (x) + t) H l (x) if G l (x) H l (x) < 2t. Particularly, we have l It,1(x): 0 x Φ 1 l (x; t) = l It,2(x): 0 x Φ 2 l (x; t) = (G l (x) t) H l (x) if l I t,1 +0 (x), (H l (x) t) G l (x) if l It,1 0+ (x), 0 if l It,1 00 (x), (G l (x) + t) H l (x) if l It,2 0 (x), (t H l (x)) G l (x) if l It,2 0+ (x), 0 if l It,2 00(x), (2) 10

l It,3(x): 0 x Φ 3 l (x; t) = l It,4(x): 0 x Φ 4 l (x; t) = (G l (x) + t) H l (x) if l It,3 0 (x), (H l (x) + t) G l (x) if l It,3 0 (x), 0 if l It,3 00 (x), (t G l (x)) H l (x) if l I t,4 +0 (x), (H l (x) + t) G l (x) if l It,4 0 (x), 0 if l It,4 00(x). As a corollary of the above lemma, we obtain an explicit formula for the linearization cone associated with (P(t)). Corollary 3.3. Fix t > 0 and a feasible point x X(t) of (P(t)). Then the following formula is valid: g i (x) d 0 i I g (x) h j (x) d 0 j P L X(t) (x) = d R n G l (x) d 0 l It,1 0+ (x) I0 t,4 (x) G l (x) d 0 l It,2 0+ (x) I0 t,3 (x). H l (x) d 0 l I t,1 +0 (x) I 0 t,2 (x) H l (x) d 0 l It,3 0 (x) I+0 t,4 (x) The upcoming lemma justies that (P(t)) is indeed a relaxation of (MPSC). Its proof parallels the one of [Kanzow and Schwartz, 2013, Lemma 3.2. Lemma 3.4. The following statements hold: (P1) X(0) = X, (P2) 0 t 1 t 2 = X(t 1 ) X(t 2 ), (P3) t>0 X(t) = X. Now, we are in position to characterize a conceptual method for the numerical solution of (MPSC): First, a sequence {t k } k N R + of positive relaxation parameters is chosen which converges to zero. Next, one solves the surrogate problem (P(t k )) via standard methods. If one computes a (local) minimizer of one of these surrogate problems which is feasible to (MPSC), then it is already a (local) minimizer of (MPSC). In general, it will be only possible to compute KKT points of the surrogate problem. However, if such a sequence of KKT points converges to some point x R n, then this point must be feasible to (MPSC) by construction. Furthermore, it will be shown in Section 3.2 that whenever MPSC-NNAMCQ is valid at x, then it is an M-stationary point of (MPSC). 3.2 Convergence properties In this section, we analyze the theoretical properties of our relaxation scheme. In order to do so, we x some standing assumptions below. 11

Assumption 3.5. Let {t k } k N R + be a sequence of positive relaxation parameters converging to zero. For each k N, let x k X(t k ) be a KKT point of (P(t k )). We assume that {x k } k N converges to x R n. Note that x X holds by Lemma 3.4. First, we will show that whenever MPSC-NNAMCQ is valid at x, then it is an M- stationary point of (MPSC). Second, it will be demonstrated that MPSC-LICQ at x implies that GCQ holds for the surrogate problem (P(t)) at all feasible points from a suciently small neighborhood of x and suciently small t > 0. This property ensures that local minima of the surrogate problem (P(t)) which are located near x are in fact KKT points. This way, it is shown that Assumption 3.5 is reasonable. Theorem 3.6. Let Assumption 3.5 be valid. Suppose that MPSC-NNAMCQ holds at x. Then x is an M-stationary point of (MPSC). Proof. Since x k is a KKT point of (P(t k )), there exist multipliers λ k R m, ρ k R p, and α k, β k, γ k, δ k R q which solve the following system: 0 = f(x k ) + ρ k j h j (x k ) λ k i g i (x k ) + i I g (x k ) j P + [ αl k xφ 1 l (x k; t k ) + βl k xφ 2 l (x k; t k ) + γl k xφ 3 l (x k; t k ) + δl k xφ 4 l (x k; t k ), l Q i I g (x k ): λ k i 0; i M \ I g (x k ): λ k l = 0, l I 0 t k,1(x k ): α k l 0; l Q \ I 0 t k,1(x k ): α k l = 0, l I 0 t k,2(x k ): β k l 0; l Q \ I 0 t k,2(x k ): β k l = 0, l I 0 t k,3(x k ): γ k l 0; l Q \ I 0 t k,3(x k ): γ k l = 0, l I 0 t k,4(x k ): δ k l 0; l Q \ I 0 t k,4(x k ): δ k l = 0. Next, let us dene new multipliers α k G, αk H, βk G, βk H, γk G, γk H, δk G, δk H Rq as stated below for all l Q: α k G,l := { α k l (H l(x k ) t k ) l I 0+ t k,1 (x k), 0 otherwise, β k G,l := { β k l (t k H l (x k )) l I 0+ t k,2 (x k), 0 otherwise, γ k G,l := { γ k l (H l(x k ) + t k ) l I 0 t k,3 (x k), 0 otherwise, δ k G,l := { δ k l ( H l(x k ) t k ) l I 0 t k,4 (x k), 0 otherwise, α k H,l := { α k l (G l(x k ) t k ) l I +0 t k,1 (x k), 0 otherwise, β k H,l := { β k l ( G l(x k ) t k ) l I 0 t k,2 (x k), 0 otherwise, γ k H,l := { γ k l (G l(x k ) + t k ) l I 0 t k,3 (x k), 0 otherwise, { δh,l k := δl k(t k G l (x k )) l I t +0 k,4 (x k), 0 otherwise. Furthermore, we set µ k := αg k +βk G +γk G +δk G and νk := αh k +βk H +γk H +δk H. By denition, we have supp µ k It 0± k (x k ) as well as supp ν k I ±0 (x k ). Thus, for suciently large t k 12

k N, (2) yields Using Lemma 3.2, 0 = f(x k ) + supp µ k I G ( x) I GH ( x), i I g (x k ) λ k i g i (x k ) + j P ρ k j h j (x k ) + l Q supp ν k I H ( x) I GH ( x). [ µ k l G l(x k ) + νl k H l(x k ) is obtained. Next, it will be shown that the sequence {(λ k, ρ k, µ k, ν k )} k N is bounded. Assuming the contrary, we dene k N: ( λ k, ρ k, µ k, ν k ) := (λk, ρ k, µ k, ν k ) (λ k, ρ k, µ k, ν k ) 2. Clearly, {( λ k, ρ k, µ k, ν k )} k N is bounded and, thus, possesses a converging subsequence (without relabeling) with nonvanishing limit ( λ, ρ, µ, ν). The continuity of g yields supp λ I g ( x). The above considerations yield supp µ I G ( x) I GH ( x), supp ν I H ( x) I GH ( x). Choose l I GH ( x) arbitrarily. If µ l 0 holds true, then µ k l 0 must be valid for suciently large k N. By denition of µ k l, l I0± t k (x k ) follows. Since It 0± k (x k ) and I t ±0 k (x k ) are disjoint, νl k = 0 holds for suciently large k N. This yields ν l k = 0 for suciently large k N, i.e. ν l = 0 is obtained. This shows l I GH ( x): µ l ν l = 0. Dividing (3) by (λ k, ρ k, µ k, ν k ) 2, taking the limit k, respecting the continuous dierentiability of f, g, h, G, as well as H, and invoking the above arguments, we obtain 0 = λ i g i ( x) + ρ j h j ( x) + [ µ l G l ( x) + ν l H l ( x), i I g ( x) j P l Q i I g ( x): λ i 0; i M \ I g ( x): λ i = 0, l I H ( x): µ l = 0, l I G ( x): ν l = 0, l I GH ( x): µ l ν l = 0. Due to the fact that ( λ, ρ, µ, ν) does not vanish, this is a contradiction to the validity of MPSC-NNAMCQ. Thus, {(λ k, ρ k, µ k, ν k )} k N is bounded. We assume w.l.o.g. that {(λ k, ρ k, µ k, ν k )} k N converges to (λ, ρ, µ, ν) (otherwise, we choose an appropriate subsequence). Reprising the above arguments, we have i I g ( x): λ i 0; i M \ I g ( x): λ i = 0, l I H ( x): µ l = 0, l I G ( x): ν l = 0, l I GH ( x): µ l ν l = 0. (3) 13

Taking the limit in (3) and respecting the continuous dierentiability of all appearing mappings, we obtain 0 = f( x) + [ µ l G l ( x) + ν l H l ( x). i I g ( x) λ i g i ( x) + j P ρ j h j ( x) + l Q This shows that x is an M-stationary point of (MPSC). In order to show that the validity of MPSC-LICQ at the limit point x ensures that GCQ holds at all feasible points of (P(t)) suciently close to x where t is suciently small, we need to study the variational geometry of the sets X(t) in some more detail. Fix t > 0 and some point x X(t). For an arbitrary index set I It 00 ( x), we consider the subsequent program: f(x) min g i (x) 0, i M, h j (x) = 0, j P, t G l (x) t, l It 0± ( x) I, t H l (x) t, l I t ±0 ( x) (It 00 ( x) \ I), Φ s l (x; t) 0, l Q \ I0 t,s( x), s S. (P(t, x, I)) The feasible set of (P(t, x, I)) will be denoted by X(t, x, I). Clearly, x is a feasible point of (P(t, x, I)) for arbitrary I It 00 ( x). Furthermore, X(t, x, I) X(t) is valid for any choice of I It 00 ( x). Lemma 3.7. For xed t > 0 and x X(t), we have T X(t) ( x) = T X(t, x,i) ( x). I It 00( x) Proof. We show both inclusions separately. Fix an arbitrary direction d T X(t) ( x). Then we nd sequences {y k } k N X(t) and {τ k } k N R + such that y k x, τ k 0, and (y k x)/τ k d as k. It is sucient to verify the existence of an index set Ī I00 t ( x) such that {y k } k N X(t, x, Ī) possesses innite cardinality since this already gives us d T X(t, x, Ī) ( x). Fix k N suciently large and l It 0± ( x). Then, due to continuity of G l, H l, as well as ϕ and feasibility of y k to (P(t)), we either have l It 0± (y k ) and, thus, G l (y k ) { t, t}, or t < G l (y k ) < t. Similarly, we obtain t H l (y k ) t for all l I t ±0 ( x). Due to feasibility of y k to (P(t)), we have t G l (y k ) t or t H l (y k ) t. Thus, setting I k := It 00 ( x) {l Q t G l (y k ) t}, y k X(t, x, I k ) is valid. Since there are only nitely many subsets of It 00 ( x) while {y k } k N is innite, there must exist Ī I00 t ( x) such that {y k } k N X(t, x, Ī) is of innite cardinality. By denition of the tangent cone, we easily obtain T X(t, x,i) ( x) T X(t) ( x) for any I It 00 ( x). Taking the union over all subsets of It 00 ( x) yields the desired inclusion. 14

Theorem 3.8. Let x X be a feasible point of (MPSC) where MPSC-LICQ is satised. Then there exist t > 0 and a neighborhood U R n of x such that GCQ holds for (P(t)) at all points from X(t) U for all t (0, t. Proof. Due to the validity of MPSC-LICQ at x and the continuous dierentiability of g, h, G, and H, the gradients { g i (x) i I g ( x)} { h j (x) j P} { G l (x) l I G ( x) I GH ( x)} { H l (x) l I H ( x) I GH ( x)} are linearly independent for all x which are chosen from a suciently small neighborhood V of x, see Lemma 2.2. Invoking (2), we can choose a neighborhood U V of x and t > 0 such that for any x X(t) U, where t (0, t holds, we have I 00 I g ( x) I g ( x), t ( x) It 0± ( x) I G ( x) I GH ( x), t ( x) I ±0 ( x) I H ( x) I GH ( x). I 00 t (4) Particularly, for any such x R n and I It 00 ( x), the gradients { g i ( x) i I g ( x)} { h j ( x) j P} { G l ( x) l It 0± ( x) I} { H l ( x) l I t ±0 ( x) (It 00 ( x) \ I)} are linearly independent, i.e., standard LICQ is valid for (P(t, x, I)) at x for any set I It 00 ( x). This implies T X(t, x,i) ( x) = L X(t, x,i) ( x) for any I It 00 ( x). Exploiting Lemma 3.7, we obtain T X(t) ( x) = L X(t, x,i) ( x). I It 00( x) Computing the polar cone on both sides yields T X(t) ( x) = L X(t, x,i) ( x). (5) I It 00( x) Dene I + G I G I + H I H (t, x, I) := I0+ t,1 ( x) I0 t,4 ( x) [ I ( It,1( x) 00 It,4( x) 00 ), (t, x, I) := I0+ t,2 ( x) I0 t,3 ( x) [ I ( It,2( x) 00 It,3( x) 00 ), (t, x, I) := I+0 t,1 ( x) I 0 t,2 ( x) [( It 00 ( x) \ I ) ( It,1( x) 00 It,2( x) 00 ), (t, x, I) := I 0 t,3 ( x) I+0 t,4 ( x) [( It 00 ( x) \ I ) ( It,3( x) 00 It,4( x) 00 ) 15

and observe that these sets characterize the indices l Q where the constraints G l ( x) t, G l ( x) t, H l ( x) t, and H l ( x) t, respectively, are active in (P(t, x, I)). We therefore obtain g i ( x) d 0 i I g ( x) h j ( x) d = 0 j P L X(t, x,i) ( x) = d R n G l ( x) d 0 l I + G (t, x, I) G l ( x) d 0 l I. G (t, x, I) H l ( x) d 0 l I + H (t, x, I) H l ( x) d 0 l I H (t, x, I) Exploiting Lemma 2.1, the polar of this cone is easily computed: η = λ i g i ( x) + ρ j h j ( x) i I g ( x) j P + [ µ l G l ( x) + ν l H l ( x) l Q i I g ( x): λ i 0 L X(t, x,i) ( x) = η R n 0 if l I + G (t, x, I),. (6) l Q: µ l 0 if l I G (t, x, I), = 0 otherwise, 0 if l I + H (t, x, I), l Q: ν l 0 if l I H (t, x, I), = 0 otherwise Observing that (P(t)) is a standard nonlinear problem, T X(t) ( x) L X(t) ( x) and, thus, L X(t) ( x) T X(t) ( x) are inherent. It remains to show T X(t) ( x) L X(t) ( x). Thus, choose η T X(t) ( x) arbitrarily. Then, in particular, (5) yields η L X(t, x, ) ( x) L X(t, x,i 00 t ( x))( x). Exploiting the representation (6), we nd λ i, λ i 0 (i Ig ( x)), ρ, ρ R p, µ, µ R q, as well as ν, ν R q which satisfy η = [ µ l G l ( x) + ν l H l ( x) and λ i g i ( x) + ρ j h j ( x) + i I g ( x) j P l Q = [ µ l G l( x) + ν l H l( x) i I g ( x) λ i g i ( x) + j P ρ j h j ( x) + l Q µ l 0 if l It,1 0+ ( x) I0 t,4 ( x), 0 if l It,2 0+ ( x) I0 t,3 ( x), = 0 otherwise, 16

µ l ν l ν l for all l Q. Thus, we obtain 0 = 0 if l It,1 0+ ( x) I0 t,4 ( x) I00 t,1 ( x) I00 t,4 ( x), 0 if l It,2 0+ ( x) I0 t,3 ( x) I00 t,2 ( x) I00 t,3 ( x), = 0 otherwise, 0 if l I t,1 +0 ( x) I 0 t,2 ( x) I00 t,1 ( x) I00 t,2 ( x), 0 if l It,3 0 ( x) I+0 t,4 ( x) I00 t,3 ( x) I00 t,4 ( x), = 0 otherwise, 0 if l I t,1 +0 ( x) I 0 t,2 ( x), 0 if l It,3 0 ( x) I+0 t,4 ( x), = 0 otherwise (ρ j ρ j) h j ( x) i λ i I ( x)(λ i) g i ( x) + g j P + [ (µ l µ l ) G l( x) + (ν l ν l ) H l( x). l Q Observing supp(µ µ ) It 00 ( x) It 0± ( x) as well as supp(ν ν ) It 00 ( x) I t ±0 ( x) and using (4), we obtain λ i = λ i (i Ig ( x)), ρ = ρ, µ = µ, as well as ν = ν. Particularly, η = [ µ l G l ( x) + ν l H l( x). i I g ( x) λ i g i ( x) + j P ρ j h j ( x) + l Q is obtained where supp µ It 0± ( x) and supp ν I t ±0 ( x) hold true. Finally, we exploit Corollary 3.3 in order to see η L X(t) ( x). This shows the validity of the inclusion T X(t) ( x) L X(t) ( x) and, thereby, GCQ holds true for (P(t)) at x. Since t (0, t and x X(t) U were arbitrarily chosen, the proof is completed. 4 Remarks on other possible relaxation schemes In this section, we discuss three more relaxation approaches for the numerical treatment of (MPSC) which are inspired by the rich theory on MPCCs. Particularly, the relaxation schemes of Scholtes [2001, Steensen and Ulbrich [2010, as well as Kadrani et al. [2009, are adapted to the setting of switching-constrained optimization. 4.1 The relaxation scheme of Scholtes For some parameter t 0, let us consider the surrogate problem f(x) min g i (x) 0, i M, h j (x) = 0, j P, t G l (x)h l (x) t, l Q. (P S (t)) 17

This idea is inspired by Scholtes' global relaxation method which was designed for the computational treatment of MPCCs, see Scholtes [2001 and [Hoheisel et al., 2013, Section 3.1. The feasible set of (P S (t)) is denoted by X S (t) and visualized in Figure 3. Note that the family {X S (t)} t 0 possesses the same properties as the family {X(t)} t 0 described in Lemma 3.4. Thus, Scholtes' relaxation is reasonable for switching-constrained problems as well. In contrast to (P(t)), where we need four inequality constraints in order to replace one original switching constraint, one only needs two inequality constraints in (P S (t)) for the same purpose. This is a signicant advantage of (P S (t)) over the surrogate (P(t)). H l(x) ( t, t) ( t, t) G l(x) ( t, t) ( t, t) Figure 3: Geometric interpretation of the relaxed feasible set X S (t). It is well known from [Hoheisel et al., 2013, Theorem 3.1 that Scholtes' relaxation approach nds Clarke-stationary points of MPCCs under an MPCC-tailored version of MFCQ. Note that, in the context of MPCCs, Clarke-stationarity is stronger than weak stationarity but weaker than Mordukhovich-stationarity. Below, we want to generalize the result from Hoheisel et al. [2013 to the problem (MPSC). For the xed parameter t > 0 and a feasible point x X S (t) of (P S (t)), we introduce the index sets I + t (x) := {l Q G l(x)h l (x) = t}, I t (x) := {l Q G l(x)h l (x) = t}. In the upcoming theorem, we provide a convergence result of Scholtes' relaxation scheme for the problem (MPSC). Theorem 4.1. Let {t k } k N R + be a sequence of positive relaxation parameters converging to zero. For each k N, let x k X S (t k ) be a KKT point of (P S (t k )). Assume that the sequence {x k } k N converges to a point x X where MPSC-MFCQ holds. Then x is a W-stationary point of (MPSC). Proof. Noting that x k is a KKT point of (P S (t k )), we nd multipliers λ k R m, ρ k R p, and ξ k R q which satisfy the following conditions: 0 = f(x k ) + ρ k j h j (x k ) λ k i g i (x k ) + i I g (x k ) j P + [ ξl k H l (x k ) G l (x k ) + G l (x k ) H l (x k ), l Q 18

i I g (x k ): λ k i 0, i M \ I g (x k ): λ k i = 0, l I + t (x k): ξ k l 0, l I t (x k): ξ k l 0, l Q \ (I + t (x k) I t (x k)): ξ k l = 0. For any k N and l Q, let us dene articial multipliers µ k l, νk l R as stated below: { { µ k ξl k l := H l(x k ) l I G ( x) I GH ( x), 0 l I H νl k ξl k := G l(x k ) l I H ( x) I GH ( x), ( x), 0 l I G ( x). Thus, we obtain 0 = f(x k ) + λ k i g i (x k ) + ρ k j h j (x k ) + i I g (x k ) j P l Q + ξl k H l(x k ) G l (x k ) + ξl k G l(x k ) H l (x k ). l I H ( x) l I G ( x) [ µ k l G l(x k ) + νl k H l(x k ) Next, we are going to show that the sequence {(λ k, ρ k, µ k, ν k, ξ k I )} k N is bounded where we used I := I G ( x) I H ( x) for brevity. We assume on the contrary that this is not the case and dene k N: ( λ k, ρ k, µ k, ν k, ξ I k ) := (λk, ρ k, µ k, ν k, ξ I k) (λ k, ρ k, µ k, ν k, ξi k). 2 Clearly, {( λ k, ρ k, µ k, ν k, ξ I k)} k N is bounded and, thus, converges w.l.o.g. to a nonvanishing vector ( λ, ρ, µ, ν, ξ I ) (otherwise, a suitable subsequence is chosen). The continuity of g ensures that I g (x k ) I g ( x) is valid for suciently large k N. Dividing (7) by (λ k, ρ k, µ k, ν k, ξi k) 2 and taking the limit k while respecting the continuous dierentiability of all involved functions, we come up with 0 = λ i g i ( x) + ρ j h j ( x) + [ µ l G l ( x) + ν l H l ( x), i I g ( x) j P l Q i I g ( x): λ i 0, i M \ I g ( x): λ i = 0, l I H ( x): µ l = 0, l I G ( x): ν l = 0. Now, the validity of MPSC-MFCQ yields λ = 0, ρ = 0, µ = 0, and ν = 0. Hence, ξ l0 0 holds for at least one index l 0 I. Let us assume l 0 I H ( x). Then we have νl k 0 = ξl k 0 G l0 (x k ), which leads to ν l0 = lim k ν k l 0 (λ k, ρ k, µ k, ν k, ξi k) = lim 2 k ξl k 0 G l0 (x k ) (λ k, ρ k, µ k, ν k, ξi k) = ξ l0 G l0 ( x) 0. 2 (7) 19

This, however, is a contradiction since ν vanishes due to the above arguments. Similarly, the case l 0 I G ( x) leads to a contradiction. As a consequence, the sequence {(λ k, ρ k, µ k, ν k, ξi k)} k N is bounded. Thus, we may assume w.l.o.g. that this sequence converges to (λ, ρ, µ, ν, ξ I ). Again, we take the limit in (7) and obtain 0 = f( x) + [ µ l G l ( x) + ν l H l ( x), i I g ( x) λ i g i ( x) + j P ρ j h j ( x) + l Q i I g ( x): λ i 0, i M \ I g ( x): λ i = 0, l I H ( x): µ l = 0, l I G ( x): ν l = 0 which shows that x is a W-stationary point of (MPSC). Noting that no suitable denition of Clarke-stationarity (in particular, a reasonable stationarity concept which is stronger than W- but weaker than M-stationarity) seems to be available for (MPSC), Theorem 4.1 does not seem to be too surprising at all. The following example shows that we cannot expect any stronger results in general. Thus, the qualitative properties of Scholtes' relaxation method are substantially weaker than those of the relaxation scheme proposed in Section 3.1. Example 4.2. Let us consider the switching-constrained optimization problem 1 2 (x 1 1) 2 + 1 2 (x 2 1) 2 min x 1 x 2 = 0. The globally optimal solutions of this program are given by (1, 0) as well as (0, 1), and these points are S-stationary. Furthermore, there exists a W-stationary point at x = (0, 0) which is no local minimizer. One can easily check that the associated problem (P S (t)) possesses a KKT point at ( t, t) for any t (0, 1. Taking the limit t 0, this point tends to x which is, as we already mentioned above, only W-stationary for the switching-constrained problem of interest. Clearly, MPSC-LICQ is valid at x. Although the theoretical properties of Scholtes' relaxation approach do not seem to be promising in light of (MPSC), we check the applicability of the approach. More precisely, we analyze the restrictiveness of the assumption that a sequence of KKT points associated with (P S (t)) can be chosen. In order to guarantee that locally optimal solutions of the nonlinear relaxed surrogate problems (P S (t)) which are located closely to the limit point from Theorem 4.1 are KKT points, a constraint qualication needs to be satised. Adapting [Hoheisel et al., 2013, Theorem 3.2 to the switching-constrained situation, it is possible to show that whenever MPSC-MFCQ is valid at a feasible point x X of (MPSC), then standard MFCQ is valid for (P S (t)) in a neighborhood of x. 20

Theorem 4.3. Let x X be a feasible point of (MPSC) where MPSC-MFCQ is satised. Then there exists a neighborhood U R n of x such that MFCQ holds for (P S (t)) at all points from X S (t) U for all t > 0. Proof. Due to the validity of MPSC-MFCQ at x, the union [ { g i ( x) i I g ( x)} { h j ( x) j P} { G l ( x) l I G ( x) I GH ( x)} { H l ( x) l I H ( x) I GH ( x)} is positive-linearly independent. Invoking Lemma 2.2, there is a neighborhood U of x such that the vectors [ { g i (x) i I g ( x)} { h j (x) j P} { G l (x) l I G ( x) I GH ( x)} { H l (x) l I H ( x) I GH ( x)} are positive-linearly independent for any choice of x U. Now, x t > 0 as well as x X S (t) U and set It a (x) := I t + (x) I t t > 0 guarantees I t + (x) I t (x) =. Clearly, we have (x). Note that l I H ( x): G l (x) 0 H l (x) 0, l I G ( x): G l (x) 0 H l (x) 0 if U is suciently small. Exploiting Lemma 2.2 once more while recalling that G and H are continuously dierentiable, we obtain that the vectors [ { g i (x) i I g ( x)} { h j (x) j P} {H l (x) G l (x) + G l (x) H l (x) l I G ( x) It a (x)} {H l (x) G l (x) + G l (x) H l (x) l I H ( x) It a (x)} { G l (x) l I GH ( x) It a (x)} { H l (x) l I GH ( x) It a (x)} (8) are positive-linearly independent if the neighborhood U is chosen small enough. Suppose that there are vectors λ R m, ρ R p, and ξ R q which satisfy 0 = ξ l [H l (x) G l (x) + G l (x) H l (x), i I g (x) λ i g i (x) + j P ρ j h j (x) + l Q i I g (x): λ i 0, i M \ I g (x): λ i = 0, l I + t (x): ξ l 0, l I t (x): ξ l 0, l Q \ I a t (x): ξ l = 0. 21

In order to show the validity of MFCQ for (P S (t)) at x, λ = 0, ρ = 0, and ξ = 0 have to be deduced. We get 0 = ρ j h j (x) λ i g i (x) + i I g (x) j P + + l (I G ( x) I H ( x)) I a t (x) ξ l l I GH ( x) I a t (x) ξ l G l (x) H l (x) + [ H l (x) G l (x) + G l (x) H l (x), l I GH ( x) I a t (x) ξ l H l (x) G l (x). Noting that λ i 0 holds for all i I g (x) while I g (x) I g ( x) holds whenever U is chosen suciently small, we obtain λ = 0, ρ = 0, ξ l = 0 (l (I G ( x) I H ( x)) I a t (x)), ξ l G l (x) = 0 (l I GH ( x) I a t (x)), and ξ l H l (x) = 0 (l I GH ( x) I a t (x)) from the positivelinear independence of the vectors in (8). Since we have G l (x) 0 and H l (x) 0 for all l I a t (x) from t > 0, ξ l = 0 follows for all l I a t (x) since I G ( x) I H ( x) I GH ( x) = Q is valid. This yields ξ = 0. Thus, MFCQ holds for (P S (t)) at x. 4.2 The relaxation scheme of Steensen and Ulbrich Here, we adapt the relaxation scheme from Steensen and Ulbrich [2010 for the numerical treatment of (MPSC). For any t > 0, let us introduce φ( ; t): R R by means of { z if z t, z R: φ(z; t) := t θ(z/t) if z < t, where θ : [ 1, 1 R is a twice continuously dierentiable function with the following properties: (a) θ(1) = θ( 1) = 1, (b) θ ( 1) = 1 and θ (1) = 1, (c) θ ( 1) = θ (1) = 0, (d) θ (z) > 0 for all z ( 1, 1). A typical example for a function θ with the above properties is given by z [ 1, 1: θ(z) := 2 π sin ( π 2 z + 3π ) 2 + 1, (9) see [Steensen and Ulbrich, 2010, Section 3. Noting that the function φ( ; t) is smooth for any choice of t > 0, it can be used to regularize the feasible set of (MPSC). A suitable surrogate problem is given by f(x) min g i (x) 0, i M, h j (x) = 0, j P, G l (x) + H l (x) φ(g l (x) H l (x); t) 0, l Q, G l (x) H l (x) φ(g l (x) + H l (x); t) 0, l Q, G l (x) + H l (x) φ( G l (x) H l (x); t) 0, l Q, G l (x) H l (x) φ( G l (x) + H l (x); t) 0, l Q. (P SU (t)) 22

Its feasible set will be denoted by X SU (t) and is visualized in Figure 4. Adapting the proof of [Steensen and Ulbrich, 2010, Lemma 3.3, the family {X SU (t)} t>0 possesses the properties (P2) and (P3) from Lemma 3.4. This justies that (P SU (t)) is a relaxation of (MPSC). Note that we need to introduce four inequality constraints to replace one of the original switching constraints. H l (x) (0, t) ( t, 0) (t, 0) G l (x) (0, t) Figure 4: Geometric interpretation of the relaxed feasible set X SU (t). It has been mentioned in Hoheisel et al. [2013 that the relaxation scheme of Steensen and Ulbrich computes Clarke-stationary points of MPCCs under an MPCC-tailored version of CPLD, see [Hoheisel et al., 2013, Section 3.4 as well. Recalling some arguments from Section 4.1, the adapted method may only nd W-stationary points of (MPSC) in general. The upcoming example conrms this conjecture. Example 4.4. Let us consider the switching-constrained optimization problem x 1 x 2 x 1 x 2 min x 2 1 + x 2 2 1 0, x 1 x 2 = 0. (10) Obviously, the globally optimal solutions of this problem are given by (1, 0) as well as (0, 1), and these points are S-stationary. Furthermore, there is a W-stationary point at x = (0, 0) which is no local minimizer. The global maximizers ( 1, 0) and (0, 1) do not satisfy any of the introduced stationarity concepts. Let us consider the associated family of nonlinear problems (P SU (t)) for t (0, 1 where the function θ is chosen as in (9). It can easily be checked that x(t) := ( t 2 (1 2 π ), t 2 (1 2 π )) is a KKT point of (P SU (t)). Note that x(t) x as t 0, and that MPSC-LICQ holds in x. However, x is only a W-stationary point of the switching-constrained problem (10). 4.3 The relaxation scheme of Kadrani, Dussault, and Benchakroun Finally, we want to take a closer look at the relaxation approach which was suggested by Kadrani et al. [2009 for the treatment of MPCCs. For any t 0, let us consider the 23

optimization problem f(x) min g i (x) 0, i M, h j (x) = 0, j P, (G l (x) t)(h l (x) t) 0, l Q, ( G l (x) t)(h l (x) t) 0, l Q, (G l (x) + t)(h l (x) + t) 0, l Q, (G l (x) t)( H l (x) t) 0, l Q, (P KDB (t)) whose feasible set will be denoted by X KDB (t). The family {X KDB (t)} t 0 only satises property (P1) from Lemma 3.4 while (P2) and (P3) are violated in general. Thus, the surrogate problem (P KDB (t)) does not induce a relaxation technique for (MPSC) in the narrower sense. Figure 5 depicts that X KBD (t) is almost disconnected, i.e., it is close to crumbling into four disjoint sets for any t > 0. This may cause serious problems when standard techniques are used to solve the associated surrogate problem (P KDB (t)). Moreover, four inequality constraints are necessary to replace one switching constraint from (MPSC) in (P KDB (t)). H l(x) ( t, t) (t, t) G l(x) ( t, t) (t, t) Figure 5: Geometric interpretation of the relaxed feasible set X KDB (t). On the other hand, it is clear from [Hoheisel et al., 2013, Section 3.3 that the regularization approach of Kadrani, Dussault, and Benchakroun computes M-stationary points of MPCCs under an MPCC-tailored version of CPLD at the limit point. Furthermore, if an MPCC-tailored LICQ holds at the limit point, then standard GCQ holds for the surrogate problems in a neighborhood of the point for suciently small relaxation parameters. These results are closely related to those for the relaxation approach from Kanzow and Schwartz [2013 which we generalized to (MPSC) in Sections 3.1 and 3.2. Although we abstain from a detailed analysis of the regularization method which is induced by the surrogate problem (P KDB (t)) due to the aforementioned shortcomings, the above arguments motivate the formulation of the following two conjectures. Conjecture 4.5. Let {t k } k N R + be a sequence of positive regularization parameters converging to zero. For each k N, let x k X KDB (t k ) be a KKT point of (P KDB (t k )). Assume that the sequence {x k } k N converges to a point x X where MPSC-MFCQ holds. Then x is an M-stationary point of (MPSC). 24

Conjecture 4.6. Let x X be a feasible point of (MPSC) where MPSC-LICQ is satis- ed. Then there exist t > 0 and a neighborhood U R n of x such that GCQ holds for (P KDB (t)) at all points from X KDB (t) U for all t (0, t. 5 Numerical results This section is dedicated to a detailed analysis and comparison of various numerical methods for (MPSC). To obtain a meaningful comparison, we apply the relaxation scheme from Section 3 as well as a collection of other algorithms (see below) to multiple classes of MPSCs which possess signicant practical relevance. The particular examples we analyze are: an either-or constrained problem with known local and global solutions a switching-constrained optimal control problem involving the non-stationary heat equation in two dimensions optimization problems involving semi-continuous variables, in particular special instances of portfolio optimization For each of the examples, we rst provide an overview of the corresponding problem structure, and then give some numerical results. To facilitate a quantitative comparison of the used algorithms, we use performance proles (see Dolan and Moré [2002) based on the computed function values. 5.1 Implementation The numerical experiments in this section were all done in MATLAB R2018a. particular algorithms we use for our computations are the following: The KS: the adapted KanzowSchwartz relaxation scheme from this paper FMC: the fmincon function from the MATLAB Optimization Toolbox SNOPT: the SNOPT nonlinear programming solver from Gill et al. [2002, called through the TOMLAB programming environment IPOPT: the IPOPT interior-point algorithm from Wächter and Biegler [2006 The overall implementation is done in MATLAB, and each algorithm is called with usersupplied gradients of the objective functions and constraints. The stopping tolerance for all algorithms is set to 10 4 (although it should be noted that the methods use dierent stopping criteria, i.e., they impose the accuracy in dierent ways). For the KS algorithm, the relaxation parameters are chosen as t k := 0.01 k, and the method is also terminated as soon as t k drops below 10 8. Finally, to solve the relaxed subproblems in the relaxation method, we employ the SNOPT algorithm with an accuracy of 10 6. 25

Judging by past experience with MPCCs, the SNOPT algorithm can be expected to rival the relaxation scheme in terms of robustness. To accurately measure the performance of the solvers, it is important to note that MPSCs can, in general, admit a substantial amount of local minimizers. Therefore, the robustness is best measured by comparing the obtained function values (using dierent methods and starting points) with the globally optimal function valueif the latter is known; otherwise, a suitable approximate is used. To avoid placing too much emphasis on the accuracy of the nal output (which does not make sense since the algorithms use completely dierent stopping criteria), we use the quantity { f(x) f min + δ, if x is feasible within tolerance, Q δ (x) := (11) +, otherwise as the base metric for the performance proles, where x is the nal iterate of the given algorithmic run, f min the (approximate) global minimal value of the underlying problem, and δ 0 is an additional parameter which reduces the sensitivity of the values to numerical accuracy. We have found that an appropriate choice of δ can signicantly improve the meaningfulness of the results. 5.2 Numerical examples The following pages contain three examples of MPSCs. In Section 5.2.1, we deal with an either-or constrained problem, which can be reformulated as an MPSC, see Mehlitz [2018. Section 5.2.2 is dedicated to a switching-constrained optimal control problem based on the framework from Clason et al. [2017. Finally, in Section 5.2.3, we deal with a class of optimization problems with semi-continuous variables, which can again be reformulated as MPSCs. This section contains a particular example from portfolio optimization which originates from Frangioni and Gentile [2007. 5.2.1 An either-or constrained example Let us consider the optimization problem (x 1 8) 2 + (x 2 + 3) 2 min x 1 2x 2 + 4 0 x 1 2 0, x 2 1 4x 2 0 (x 1 3) 2 + (x 2 1) 2 10 0. (E2) Here, denotes the logical or. The feasible set of this program is visualized in Figure 6. It is easily seen that (E2) possesses the unique global minimizer x = (2, 2) and another local minimizer x = (4, 4). Arguing as in [Mehlitz, 2018, Section 7, we can transform (E2) into a switching-constrained optimization problem by introducing additional variables: (x 1 8) 2 + (x 2 + 3) 2 min x,z z 1, z 2, z 3, z 4 0, (x 1 2x 2 + 4 z 1 )(x 1 2 z 2 ) = 0, (x 2 1 4x 2 z 3 )((x 1 3) 2 + (x 2 1) 2 10 z 4 ) = 0. (12) 26