Technische Universität Dresden Herausgeber: Der Rektor

Size: px
Start display at page:

Download "Technische Universität Dresden Herausgeber: Der Rektor"

Transcription

1 Als Manuskript gedruckt Technische Universität Dresden Herausgeber: Der Rektor The Gradient of the Squared Residual as Error Bound an Application to Karush-Kuhn-Tucker Systems Andreas Fischer MATH-NM August 2002

2 The Gradient of the Squared Residual as Error Bound an Application to Karush-Kuhn-Tucker Systems Andreas Fischer Department of Mathematics University of Dresden Dresden Germany August 2002 Abstract. A general relationship between the natural residual of a system of equations the necessary optimality conditions associated to the corresponding least squares problem is presented. Based on this an error bound result for the gradient of a least squares reformulation of Karush-Kuhn-Tucker systems will be derived. 1 Introduction Error bounds have been turned out an essential tool for analyzing both the global the local convergence behavior of algorithms for solving equations or more general problems. Besides their use for theoretical purposes appropriate error bounds play a central role in several techniques for globalizing locally convergent algorithms. More recently, computable error bounds have been successfully employed for achieving or improving certain local properties of algorithmn cases where classical regularity conditions are violated. Let us first mention techniques for the locally accurate identification of constraints which are active at a solution of a Karush-Kuhn-Tucker system. Secondly, several modifications of Newton-type methods have been developed that guarantee superlinear convergence properties even in cases where no isolated solution exists. For such modifications the existence of an appropriate error bound is one of the key assumptions for superlinear convergence. Let us consider the problem of solving the equation H(z) = 0, (1) where H : R l 1 R l 2 is a given map. Its solution set is assumed to be nonempty 2

3 denoted by Σ, i.e., Σ := {z R l 1 H(z) = 0}. (2) In [7] [3] Levenberg-Marquardt type algorithms have been suggested for solving (1) provided that H is continuously differentiable. The local Q-quadratic rate of convergence in [7] relies, besides further assumptions, on an error bound condition like µd[z, Σ] H(z) z Σ 0 + ɛb (3) for some ɛ, µ > 0, where Σ 0 denotes a nonempty closed subset of Σ d[z, Σ] the distance of z to Σ. In [3] a general approach for solving generalized equations with nonisolated solutions has been developed. It exploits the upper Lipschitz-continuity of the solution set map belonging to a perturbed generalized equation. In the case of problem (1) this assumption is equivalent to the error bound condition above. As an application of the approach just mentioned a prox-regularized Newton-type method for solving the necessary optimality conditions Q(z) := H(z)H(z) = 0 (4) associated to the least squares problem Ψ(z) := 1 2 H(z)T H(z) min is suggested in [3, Section 5.4]. Therefore, instead of (3), an error bound condition for problem (4) is of importance. This condition reads for some ɛ, µ > 0. Note that where µd[z, Σ] Q(z) z Σ 0 + ɛb (5) d[z, S] d[z, Σ] z R l 1, S := {z R l 1 Q(z) = 0} Σ denotes the solution set of problem (4). Obviously, if H(z) is bounded in a neighborhood of Σ, the error bound condition (5) implies condition (3) with µ > 0 suitably chosen. However, since (3) is a natural error bound condition for problem (1), the question arises under which assumptions (3) implies (5). An answer to this basic question will be given by Theorem 1 in Section 2. This theorem is not only applicable to continuously differentiable maps but also to a certain class of nondifferentiable maps H. In particular, maps Φ can be dealt with that are frequently used to reformulate Karush-Kuhn-Tucker (KKT) systems as systems of nondifferentiable equations. To this end, we show in Section 3 that if the KKT system satisfies certain assumptions the map Φ belongs to the class of nondifferentiable maps covered by Theorem 1. In Section 4, it is shown that there is a reasonably large class of KKT systems that satisfy this assumption. In particular, affine KKT systems do so. If (3) implies (5) then this can be exploited for the analysis of the local behavior of algorithms. Suppose that a line search algorithm makes use of search directions d := M(z) H(z)H(z) = M(z)Q(z) 3

4 with a certain matrix M(z) R l 1 l 1. Then, (3) (5) can be helpful for estimating the local progress, in particular, if the step length is computed by decreasing a merit function like Ψ. A corresponding example is provided in the forthcoming paper [4]. There, a class of algorithms for solving KKT systems with nonisolated solutionnvestigated in respect to the limiting behavior of the iterates of active set estimates. Another advantage of having (5) is that any solution of Q(z) = 0 in a neighborhood of Σ 0 also solves H(z) = 0 without any further condition. Notation: Throughout the paper denotes the Euclidean vector norm or the induced matrix norm. The unit ball (of appropriate dimension) is always denoted by B. For a set S R l a point z R l the distance of z to S is defined as d[z, S] := inf{ z s s S} if S d[z, S] := otherwise. Moreover, by Π(z, S) := {s S z s = d[z, S]} we denote the set of all pointn S that have minimal distance to S. If S is nonempty closed then Π(z, S) is nonempty for any z R l. Let G : R d 1 R d 2 be a locally Lipschitz-continuous function. Then, Clarke s generalized Jacobian of G at z R d 1 exists is denoted by G(z). If G is continuously differentiable at z then G(z) = G(z) T. Definition further properties of G(z) can be found in [1]. 2 The Gradient of the Squared Residual as Error Bound Let us first discuss the question whether assumptions (3) implies (5) in the classical situation where H is continuously differentibale H(z ) is nonsingular at a solution z of (1). Then, with the continuity of H, Taylors formula shows that z is an isolated solution. Moreover, in a neighborhood of z, H(z) 1 exists its norm is bounded above. Therefore, setting Σ 0 := {z }, condition (3) for ɛ, µ > 0 sufficiently small implies (5) with µ > 0 suitably chosen. If, however, z is a nonisolated solution of (1), then H(z ) must be singular as long as H is continuous. Moreover, even if H(z) 1 exists for z close to z, its norm cannot be bounded. Nevertheless, (3) implies (5). This follows from Theorem 9 in [3]. This theorem is applicable not only to continuously differentiable maps but also to a certain class of nondifferentiable maps H. We will now present a corresponding theorem for a larger class of nondifferentiable maps H. To this end, let H : R l 1 R l 2 denote a locally Lipschitz-continuous map consider problem (1), i.e., H(z) = 0. According to (2), this problem is assumed to have a nonmepty solution set denoted by Σ. The necessary optimality condition for minimizing the squared residual Ψ(z) = 1 2 H(z)T H(z). reads 0 Ψ(z), 4

5 where Ψ(z) = H(z) T H(z) holds according to an appropriate chain rule [1]. We will now derive conditions under which the norm of elementn Ψ(z) can serve as an error bound for d[z, Σ], i.e., for the distance of z to the solution set of (1). Theorem 1 Let Σ 0 Σ be nonempty closed. Assume that there are ɛ, µ > 0, σ 1 so that, for any z Σ 0 + ɛb, there is ẑ z + σd[z, Σ]B V H(z) so that H(z) + V (ẑ z) 1 µd[z, Σ] (6) 2 Then, with µ := µ(2σ) 1, is valid for all z Σ 0 + ɛb. µd[z, Σ] H(z). (7) µd[z, Σ] V T H(z) (8) Proof. Choose any z Σ 0 + ɛb any V H(z). If z Σ, then inequality (8) is obviously valid. Otherwise, if z (Σ 0 + ɛb) \ Σ, multiply the vector within the norm in (6) by H(z) T. With (6), this yields H(z) 2 H(z) T V (ẑ z) H(z) T H(z) + H(z) T V (ẑ z) µ H(z) d[z, Σ]. From ẑ z + σd[z, Σ]B it follows that H(z) µ H(z) d[z, Σ] σd[z, Σ] V T H(z). Dividing this by d[z, Σ] taking into account (7), we obtain 1 2 µ H(z) σ V T H(z). The previous theorem refines [3, Theorem 9]. In particular, (6) is a weaker assumption than the corresponding condition in [3]. Due to this refinement, it will be possible to apply Theorem 1 for the map H := Φ with Φ defined below. 3 Application to Karush-Kuhn-Tucker Systems In this section the case is dealt with that H(z) = 0 reformulates the KKT system in a particular but frequently used manner. Assumptions will be provided which ensure that (6) is satisfied, thus, that Theorem 1 is applicable to this case

6 We first need to describe the KKT system its reformulation as system of equations in more detail. Let F : R n R n be a continuously differentiable function, g : R n R m h : R n R p twice continuously differentiable functions consider the system L(x, u, v) = 0 h(x) = 0 g(x) 0 u 0 u T g(x) = 0 (9) with the Lagrangian L : R n+m+p R n given by L(x, u, v) := F (x) + h(x)v g(x)u. System (9) is well known as the Karush-Kuhn-Tucker (KKT) system belonging to the variational inequality problem Find x G so that F (x) T (ξ x) 0 for all ξ G, (10) where G := {x R n h(x) = 0, g(x) 0}. If x is a solution of (10) if a certain constraint qualification is satisfied at x then (u, v) exists so that (x, u, v) solves (9). Moreover, under a certain constraint qualification, the system (9), with F := f for f : R n R sufficiently smooth, states necessary optimality conditions associated to the programming problem f(x) min s.t. x G. Therefore, a basic approach for solving such programs or the variational inequality problem (10) is to determine a solution of the KKT system (9). To this end, (9) is often reformualated as a system of equations. A frequently used approach is based on the function ϕ : R 2 R given by ϕ(a, b) := a 2 + b 2 a b. Since ϕ equals to zero if only if a 0, b 0, ab = 0, it is easily verified that (9) is equivalent to Φ(z) = 0, where z := (x, u, v) with Now, if we set Φ(z) := L(z) h(x) φ(z) φ(z) := (ϕ(g 1 (x), u 1 ),..., ϕ(g m (x), u m )) T. H := Φ, (11) the same question an Sections 1 2 becomes of interest. Namely, which assumptions ensure that the error bound condition (3) implies (5). However, the function Q as defined in (4) employed in (5) is not well defined now. This due to the fact that ϕ is nondifferentiable at (0, 0) so that H is not necessarily everywhere differentiable. Nevertheless, the merit function Ψ(z) = 1 2 Φ(z)T Φ(z) 6

7 is continously differentiable [2, 5] the function Q can be defined by It holds that Any matrix V Φ(z) can be written as V = Q(z) := Ψ(z). Ψ(z) = V T Φ(z) for all V Φ(z). (12) x L(x, u, v) h(x) g(x)d a (g(x), u) h(x) T 0 0 g(x) T 0 D b (g(x), u) with diagonal matrices D a (g(x), u) D b (g(x), u). Their i-th diagonal entries a(g i (x), u i ) b(g i (x), u i ), respectively, are given by T (13) a(g i (x), u i ) = a ϕ(g i (x), u i ), b(g i (x), u i ) = b ϕ(g i (x), u i ) (14) if (g i (x), u i ) (0, 0), where a ϕ(a, b) = a a2 + b 2 1, bϕ(a, b) = b a2 + b 2 1 (15) for (a, b) (0, 0). Otherwise, if (g i (x), u i ) = (0, 0), there are α i, β i R so that a(g i (x), u i ) = α i 1, b(g i (x), u i ) = β i 1 with α 2 i + β 2 i 1. (16) To answer the main question under which assumptions (3) implies (5) we would like to apply Theorem 1. Therefore, besides the error bound condition (7) which corresponds to (3), condition (6) need to be satisfied. To achieve thin the case that H = Φ the subsequent assumption plays a key role. Its formulation the analysis thereafter require some index sets. With I := {1,..., m} let I C (z) := {i I g i (x) = u i = 0} denote the set of all indices that are complementary at z R n+m+p. z R n+m+p t 0 define I(z, t) := {i I max{u i, g i (x)} t}. Note that I C (z) I(z, t) for any z Σ any t 0. Moreover, for Assumption 1 Let Σ 0 Σ be nonempty closed ɛ 1 > 0 be given. For any N 1 there are σ > 0 τ > 0 so that, for any y (Σ 0 + ɛ 1 B) Σ any t [0, τ], there is y t Σ with y y t σt I(y t, N max{ y y t, t}) = I C (y t ). 7

8 Roughly speaking, this assumption requires that for any point in the set (Σ 0 + ɛ 1 B) Σ there is a point in Σ so that both points are not too far away from each other that the complementary indices of the latter point are stable in a certain sense. Before exploiting Assumption 1 let us refer to Section 4. There, an error bound condition is presented under which Assumption 1 can be fulfilled. Besides the smoothness conditions stated at the beginning of this section we will make use of the following additional Lipschitz-continuity conditions. Assumption 2 There is L 1 so that, for all z, z Σ 0 + B, a) g i (x) g i (x ) L x x for all i I, b) g i (x) g i (x ) L x x for all i I, h(x) h(x ) L x x, L(z) L(z ) L z z. If Σ 0 is bounded then Assumption 2 except the third inequality in part b) is satisfied. If, in addition, F, 2 g, 2 h are locally Lipschitz-continuous then the latter inequality holds as well. Lemma 1 Suppose that Assumption 1 is satisfied. Then, for any N 1, there is ˆρ > 0 so that, for any z Σ 0 + ˆρB, a vector ẑ Σ (17) exists with z ẑ (σ + 1)d[z, Σ] (18) g i (ˆx) + û i > 1 2 N max{ z ẑ, d[z, Σ]} i I \ I C(ẑ). (19) If, in addition, Assumption 2 a) holds N 4L, then I C (z) I C (ẑ). (20) Proof. Let N 1 be arbitrary but fixed. Then, with ɛ 1, σ > 0, τ > 0 existing due to Assumption 1, define ˆρ := min{1, τ, 1 2 ɛ 1 1, }. (21) σ + 2 Now, choose any z Σ 0 + ˆρB. Then, there is y Π(z, Σ). From (21) it follows that d[y, Σ 0 ] z y + d[z, Σ 0 ] 2d[z, Σ 0 ] 2ˆρ ɛ 1. Thus, To apply Assumption 1, define Due to (21), thimplies y (Σ 0 + ɛ 1 B) Σ. (22) t := d[z, Σ]. (23) t d[z, Σ 0 ] ˆρ τ. (24) 8

9 Assumption 1 together with (22) (24) ensures that ẑ := y t Σ exists with y ẑ σt = σd[z, Σ] We therefore obtain (18) by g i (ˆx) + û i > N max{ y ẑ, t} i I \ I C (ẑ). z ẑ z y + y ẑ (1 + σ)d[z, Σ] (19) since g i (ˆx) + û i > N max{ y ẑ, t} N max{ z ẑ y z, t} = N max{ z ẑ t, t} 1 N max{ z ẑ, t} 2 for all i I \ I C (ẑ). To verify (20) we first note that, by (21) (18), z Σ 0 + ˆρB Σ 0 + B, ẑ z + (σ + 1)d[z, Σ]B Σ 0 + ˆρB + (σ + 1)ˆρB Σ 0 + B. (25) Hence, Assumption 2 a), N 4L, L 1 provide g i (ˆx) + û i g i (ˆx) g i (x) + û i u i L ˆx x + û u 1 N ẑ z 2 for any i I C (z). Therefore, (19) yield I C (ẑ). Lemma 2 Suppose that Assumptions 1 2 are satisfied that µ > 0 is given. Then, there is ρ > 0 so that, for any z Σ 0 + ρb, a vector ẑ Σ exists with z ẑ (σ + 1)d[z, Σ] (26) Φ(z) + V (ẑ z) 1 µd[z, Σ] for all V Φ(z). (27) 2 Proof. To apply Lemma 1 choose N so that N 4L. (28) Then, according to Lemma 1, ˆρ > 0 exists so that, for any z Σ 0 + ˆρB, there is ẑ so that (17) (20) are satisfied. Based on this we will show that (z, ẑ) also satisfies (26) (27) if z Σ 0 + ρb, where ρ is given by ρ := min{ˆρ, µ 8 }. (29) m(σ + 1) 2 L Obviously, (26) directly follows from (18). To prove (27), the term R(z, ẑ) := Φ(z) + V (ẑ z) (30) 9

10 will be investigated componentwise for V Φ(z) arbitrary but fixed. The first n + p components of R(z, ẑ) read as follows R 1 n (z, ẑ) = L(z) + L(z) T (ẑ z) R n+1 n+p (z, ẑ) = h(x) + h(x) T (ˆx x). Taylor s formula, L(ẑ) = 0 due to (17), Assumption 2 b) with z, ẑ Σ 0 + B (an (25)) yield R 1 n (z, ẑ) = 1 Taking into account (26) (29) we further get In the same way one can show that 0 ( L(z + s(ẑ z)) L(z)) T (ẑ z)ds L ẑ z 2. R 1 n (z, ẑ) L(σ + 1) 2 d[z, Σ] 2 µ d[z, Σ]. (31) 8 We now consider the last m components of (30), i.e., R n+1 n+p (z, ẑ) µ d[z, Σ]. (32) 8 R n+p+i (z, ẑ) = φ i (z) + v i (ẑ z) i = 1,..., m, where v i is the (n + p + i)-th row of V, thus v i φ i (z). Taylor s formula yields with r i (x, ˆx) := g i (x) T (ˆx x) = g i (ˆx) g(x) r i (x, ˆx) (33) 1 Similar to showing (31), we get 0 ( g i (x + s(ˆx x)) g i (x)) T (ˆx x)ds. r i (x, ˆx) µ 8 d[z, Σ]. (34) m Now, for any i I two cases are distinguished: a) i I C (z). Due to (20) in Lemma 1, i I C (ẑ) follows. Thus, i I C (z) I C (ẑ) g i (x) = g i (ˆx) = u i = û i = φ i (z) = 0. (35) Using the representation (13) of matrices V contained in Φ(z) together with (16), we get R n+p+i (z, ẑ) = φ i (z) + v i (ẑ z) = (α i 1) g i (x) T (ˆx x) + (β i 1)(û i u i ) = (α i 1) g i (x) T (ˆx x). Therefore, (33), (35), (34), α i 1 2 (by 16) imply R n+p+i (z, ẑ) µ 4 d[z, Σ]. m 10

11 b) i I \ I C (z). Then, (g i (x), u i ) (0, 0) so that φ i is continuously differentiable at z = (x, u, v). With (13) (14), we obtain φ i (z) T (ẑ z) = a ϕ(g i (x), u i ) g i (x) T (ˆx x) + b ϕ(g i (x), u i )(û i u i ). Having (15) in mind setting := g i (x) 2 + u 2 i, we further get φ i (z) T (ẑ z) = ( g i(x) Together with (33), we have 1) g i (x) T (ˆx x) + ( u i 1)(û i u i ). R n+p+i (z, ẑ) = φ i (z) + φ i (z) T (ẑ z) = g i (x) u i + ( g i(x) 1) g i (x) T (ˆx x) + ( u i 1)(û i u i ) = g i (x) + ( g i(x) 1)(g i (ˆx) g i (x) r i (x, ˆx)) u2 i + ( u i 1)û i = g i(x) 2 +u 2 i + ( g i(x) 1)(g i (ˆx) r i (x, ˆx)) + ( u i 1)û i, by the definition of, R n+p+i (z, ẑ) = ( g i(x) Now, three subcases of case b) are considered. b1) i I C (ẑ). Then, g i (ˆx) = û i = 0. From (36) (34) 1)(g i (ˆx) r i (x, ˆx)). + ( u i 1)û i (36) R n+p+i (z, ẑ) = g i(x) 1 r i (x, ˆx) µ 4 d[z, Σ]. m follows. b2) i I \ I C (ẑ) g i (ˆx) > 0. Then, by (17), û i = 0. Moreover, (19) can be exploited. Therefore, having Assumption 2 a) (28) in mind, we get g i (x) g i (ˆx) L x ˆx > 1g 2 i(ˆx) + 1 N max{ z ẑ, d[z, Σ]} L x ˆx 4 1g 2 i(ˆx) > 0 (37) g i (x) 1 2 g i(ˆx) > 1 N max{ z ẑ, d[z, Σ]}. (38) 4 This, Assumption 2 a), (28) yield g i (x ) g i (x) L x x g i (x) Ld[z, Σ] > 0 11

12 for z = (x, u ) Π(z, Σ). Since z Σ, thimplies u i = 0 By an appropriate Taylor expansion we have u i = u i u i z z = d[z, Σ]. (39) a a 2 + b 2 b2 2a (a, b) (0, ) R. Setting a := g i (x) b := u i, it follows with (37) (39) that g i (x) = g i(x) gi (x) 2 + u 2 i u2 i 2 g i (x) d[z, Σ]2 2g i (x) 2. Therefore, with (37) (34), we further get R n+p+i (z, ẑ) = g i(x) g i (ˆx) + r i (x, ˆx) This (38) lead to Obviously, for N sufficiently large, R n+p+i (z, ẑ) 4 N d[z, Σ] + µ 4N m d[z, Σ]2. R n+p+i (z, ẑ) µ 4 d[z, Σ] m d[z, Σ]2 2g i (x) (2g i(x) + µ 2 8 d[z, Σ]). m follows. b3) i I \ I C (ẑ) û i > 0. In a very similar way the same estimate an case b2) can be obtained by carefully interchanging certain terms, (g i (ˆx) with û i or g i (x ) with u i, for instance). The results for the cases a) b1) b3) together with (31) (32) show that Φ(z) + V (ẑ z) = R(z, ẑ) 1 µd[z, Σ]. 2 holds for all V Φ(z). Theorem 2 Suppose that Assumptions 1 2 are satisfied. Moreover, assume that there are ɛ, µ > 0 so that µd[z, Σ] Φ(z) z Σ 0 + ɛb. (40) Then, there is µ > 0 so that µd[z, Σ] Ψ(z) z Σ 0 + ɛb. Proof. Apply Theorem 1 for H := Φ with σ + 1 instead of σ, where (12) Lemma 2 with µ > 0 from (40) have to be taken into account. 12

13 4 Pleasant Karush-Kuhn-Tucker Systems In this section conditions are provided under which Assumption 1 is satisfied. To proceed let us first define investigate the activity pattern belonging to any z Σ. Definition 1 For any z Σ the activity pattern p(z) := (g(z), u(z)) is defined by g(z) := {i I g i (x) = 0}, u(z) := {i I u i = 0}. All activity patterns of pointn z Σ are collected in the set In addition, let P(Σ) := {p(z) z Σ}. P := {p = (g, u) g, u {1,..., m}, g u = {1,..., m}}. Elements p 1 = (g 1, u 1 ) p 2 = (g 2, u 2 ) of P are said to be related, p 1 p 2 for short, if g 1 g 2 u 1 u 2. In addition, p 1 p 2 is used to denote that p 1 p 2 p 1 p 2. Any element p P(Σ) is called maximal if no q P(Σ) exists with p q. The set of all maximal elementn P(Σ) is denoted by P max (Σ). Finally, let P a (Σ) := {p P p q P(Σ)}. Obviously, P collects all those activity patterns that are potentially possible but need not occur in the solution set of a particular KKT system. The only requirement an element (g, u) of P has to satisfy is the complementarity condition, i.e. that each index i I is contained in at least one of the sets g u. Moreover, the inclusions P(Σ) P a (Σ) P. can easily be verified. For any p = (g, u) P, let the map Φ p the cone K p be defined by L(z) h(x) F p (z) := g g (x) u u, K p := {0} n+p+ g + u R m+m +, g(x) u where Then, for any p = (g, u) P, the set g g := ( g i ) T i g u u := ( u i ) T i u. Σ p := {z R n+m+p F p (z) K p } is possibly empty contained in Σ. For affine KKT systems any set Σ p with p P is a closed polyhedron. 13

14 Lemma 3 a) If p 1, p 2 P, then p 1 p 2 implies Σ p2 Σ p1. b) For any p P a (Σ), there is p max P max (Σ) so that p p max. c) The set Σ p is nonempty if only if p P a (Σ). Proof. Obvious. Lemma 4 Let Σ 2 Σ be nonempty compact. Then, there is κ 0 (0, 1] so that g i (x) + u i κ 0 i I \ I C (z) for any z Σ 2 with p(z) = (g(z), u(z)) P max (Σ). Proof. Assume the contrary. Then, a sequence {z ν } Σ 2, z Σ 2, p P max (Σ), i I must exist so that (i, i) / p(z ν ) = p P max (Σ) ν N lim (g i(x ν ) + u ν i ) = 0, ν lim z ν = z. ν With the continuity of g it follows that p p(z ). Thus, p cannot be maximal. Lemma 5 Let Σ 2 Σ be nonempty compact. Then, for any η > 0, there is κ (0, κ 0 ] so that, for all z Σ 2 all i I, g i (x) + u i κ (41) implies p(z) (i, i) P a (Σ) (42) inf{ z s s Σ p(z) (i,i) } η. (43) Proof. Fix η > 0. Assume first that there is no κ (0, κ 0 ] so that (41) implies (42) for all z Σ 2 all i I. Then, sequences {κ ν } (0, κ 0 ] {z ν } Σ 2, j I, ẑ = (ˆx, û, ˆv) Σ 2 must exist with lim κ ν = 0, ν lim ν z ν = ẑ, (44) g j (x ν ) + u ν j κ ν ν N, (45) p(z ν ) (j, j) P \ P a (Σ) ν N. (46) Taking suitable subsequencef necessary we have that, without loss of generality, p(z ν ) = ˆp ν N (47) for some fixed ˆp = (ĝ, û) P(Σ). Now consider any i I. If i ĝ, we have from (44) by the continuity of g i that lim g i(x ν ) = g i (ˆx) = 0, ν lim u ν i = û i 0 ν 14

15 If i û, we get Hence, with ẑ Σ 2 Σ, follows. Since (45) implies lim g i(x ν ) = g i (ˆx) 0, ν lim u ν i = û i = 0. ν ˆp p(ẑ) P(Σ) (48) lim g j(x ν ) + u ν j = g j (ˆx) + û j = 0, ν we further get (j, j) p(ẑ). This together with (48) yields ˆp ˆp (j, j) p(ẑ) P(Σ). (49) Thus, ˆp (j, j) P a (Σ) which contradicts (46) (47). Therefore, (42) is valid for all z Σ 2. To show that (43) implied by (41) first note that, due to (42) Lemma 3 c), Σ p(z) (i,i) for all z Σ 2 all i I satisfying (41). Thus, the left term in (43) is well defined. Let us assume that (43) does not hold. Then, we can repeat all steps of the previous part of the proof until formula (49) with the only modification that (46) is replaced by From (49), Lemma 3 a), (47) we have Therefore, since ẑ Σ p(ẑ), inf{ z ν y y Σ p(z ν ) (j,j)} > η ν N. (50) Σ p(ẑ) Σˆp (j,j) = Σ p(z ν ) (j,j) ν N. inf{ z ν y y Σ p(z ν ) (j,j)} z ν ẑ follows. By (44), this contradicts (50) for ν N sufficiently large. Assumption 3 (Pleasant KKT System) There are ω (0, 1] δ > 0 so that, for any p P a (Σ), ωd[z, Σ p ] inf F p (z) f z Σ p + δb. f K p Theorem 3 Let Assumption 3 be satisfied suppose that Σ 0 Σ is nonempty compact. Moreover, let ɛ 1 > 0 be given. Then, Assumption 1 is satisfied. Proof. Let N 1 be arbitrary but fixed define σ the sets Σ 1 Σ 2 Σ by σ := (1 + ω 1 N) m Σ 1 := (Σ 0 + ɛ 1 B) Σ, Σ 2 := (Σ 1 + σb) Σ. 15

16 Since Σ 0 is compact by assumption the same holds for Σ 1 Σ 2. Therefore, with η := δ (δ > 0 from Assumption 3), Lemma 5 provides some κ (0, κ 0 ] (0, 1] we can define τ := κσ 1 < 1. (51) Now, choose any (y, t) Σ 1 [0, τ] define vectors z 0,..., z m numbers σ 0,..., σ m recursively as follows. First, let z 0 := y. To define z k+1 from z k for k {0,..., m 1} choose If then set choose Otherwise, set z k+1 := z k. Finally, let i k argmin{g i (x k ) + u k i i I \ I C (z k )}. g ik (x k ) + u k i k N max{ y z k, t}, (52) p k := p(z k ) (i k, i k ) (53) z k+1 argmin{ z z k z Σ p k}. (54) σ k := (1 + ω 1 N) k (55) for k = 0,..., m. We now show by induction that z 0,..., z m are well defined that y z k σ k t z k Σ 2. (56) holds for k = 0, 1,..., m. For k := 0 we get that z 0 = y σ 0 = 1 so that (56) is obviously satisfied. Now, let (56) be valid for some k {0,..., m 1}. If (52) is violated, then z k+1 = z k (56) must hold. Therefore, we only need to consider the case if (52) is satisfied. In view of σ k 1, t [0, τ], (56), (55), (51), thimplies g ik (x k ) + u k i k Nσ k t N(1 + ω 1 N) k τ σ k+1 τ στ = κ. (57) Therefore since z k Σ 2, we can apply Lemma 5. Together with (53), p(z k ) p k = p(z k ) (i k, i k ) P a (Σ) (58) follows. Thus, by Lemma 3 c), the closed set Σ p k is nonempty so that z k+1 is well defined by (54). Moreover, (43) in Lemma 5 (with η := δ) gives inf{ z k s s Σ p k} = z k z k+1 η = δ. Hence, since z k+1 Σ p k, Assumption 3 can be exploited for z := z k p := p k leads to z k z k+1 = d[z k, Σ p k] ω 1 inf F p k(z k ) f. 16 f K p k

17 Due to z k Σ p(z k ) (53), the definition of F p k yields so that, with (57), inf F p k(z k ) f = g ik (x k ) + u k i f K k p k z k z k+1 ω 1 Nσ k t. follows. Therefore, by (56), t [0, τ], τ < 1 from (51), (55), we have y z k+1 y z k + z k z k+1 (1 + ω 1 N)σ k t = σ k+1 t σ. Thus, (56) is true for k + 1 instead of k so for all k {0,..., m}. From (58) for k = 0,..., m 1 it follows that Together with p(z k ) p k p(z k+1 ) k {0,..., m 1}. I g(z 0 ) u(z 0 ), for p(z 0 ) = (g(z 0 ), u(z 0 )), we have that there is k 0 {0,..., m} so that p(z k ) P max (Σ) k {k 0,..., m}. According to Lemma 4 this means that g i (x k ) + u k i κ 0 i I \ I C (z k ) k {k 0,..., m}. Therefore, (52) is violated for all k {k 0,..., m}. In particular, together with (56) for k = m, it follows that y z m σt, g i (x m ) + u m i > N max{ y z m, t} i I \ I C (z m ). Hence, y t := z m has exactly the properties required in Assumption 1. Corollary 1 Suppose that Σ 0 Σ is nonempty compact. Moreover, let Assumptions 2 3 be satisfied. If there are ɛ, µ > 0 so that µd[z, Σ] Φ(z) z Σ 0 + ɛb, then there are ɛ, µ > 0 so that µd[z, Σ] Ψ(z) z Σ 0 + ɛb. Proof. The assertion directly follows from Theorem 2 Theorem 3. For affine KKT systems, i.e., if F, g, h are affine functions, Corollary 1 can be simplified as follows. Corollary 2 Suppose that Σ 0 Σ is nonempty compact. If the KKT system is affine then there are ɛ, µ > 0 so that µd[z, Σ] Ψ(z) z Σ 0 + ɛb. 17

18 Proof. Since F, g, h are affine, the function F p is affine for any p P. Thus, Assumption 2 is satisfied. Moreover, by Lemma 3 c), Σ p is nonempty for all p P a (Σ) Therefore, Hoffman s error bound [6] for affine systems of inequalities ensures that, for any p P a (Σ), there are are δ p, ω p > 0 so that ω p d[z, Σ p ] inf f K p F p (z) f z Σ p + δ p B. Since P a (Σ) is a finite set, Assumption 3 is satisfied with δ := min{δ p p P a (Σ)} ω := min{ω p p P a (Σ)}. Hence, by Theorem 3, Assumption 1 holds for any ɛ 1 > 0. Altogether, Theorem 2 provides the desired result. Affine KKT systems are of particular interest for future research. We think that the compactness of Σ 0 Σ as often assumed in this paper can be removed for affine KKT systems. Another improvement of the resultn this section might be obtained by using a more local version of Assumption 3 so that not all Σ p with p P a (Σ) occur. References [1] Clarke, F.H. (1983): Optimization Nonsmooth Analysis. John Wiley Sons, NY [2] Facchinei, F., Soares, J. (1997): A new merit function for nonlinear complementarity problems a related algorithms. SIAM Journal on Optimization, 7, [3] Fischer, A. (2001): Local behavior of an iterative framework for generalized equations with nonisolated solutions. Applied Mathematics Report 203, Department of Mathematics, University of Dortmund, Dortmund (revised 2002) [4] Fischer, A. (2002): Limiting behavior of an algorithmic framework for Karush- Kuhn-Tucker systems. Forthcoming. [5] Geiger, C., Kanzow, C. (1996): On the resolution of monotone complementarity problems. Computational Optimization Applications, 5, [6] Hoffman, A. J. (1952): On approximate solutions of systems of linear inequalities, Journal of Research of the National Bureau of Stards, 49, [7] Yamashita, N., Fukushima, M. (2001): On the rate of convergence of the Levenberg- Marquardt method. Computing, 15,

A SIMPLY CONSTRAINED OPTIMIZATION REFORMULATION OF KKT SYSTEMS ARISING FROM VARIATIONAL INEQUALITIES

A SIMPLY CONSTRAINED OPTIMIZATION REFORMULATION OF KKT SYSTEMS ARISING FROM VARIATIONAL INEQUALITIES A SIMPLY CONSTRAINED OPTIMIZATION REFORMULATION OF KKT SYSTEMS ARISING FROM VARIATIONAL INEQUALITIES Francisco Facchinei 1, Andreas Fischer 2, Christian Kanzow 3, and Ji-Ming Peng 4 1 Università di Roma

More information

Technische Universität Dresden Institut für Numerische Mathematik. An LP-Newton Method: Nonsmooth Equations, KKT Systems, and Nonisolated Solutions

Technische Universität Dresden Institut für Numerische Mathematik. An LP-Newton Method: Nonsmooth Equations, KKT Systems, and Nonisolated Solutions Als Manuskript gedruckt Technische Universität Dresden Institut für Numerische Mathematik An LP-Newton Method: Nonsmooth Equations, KKT Systems, and Nonisolated Solutions F. Facchinei, A. Fischer, and

More information

WHEN ARE THE (UN)CONSTRAINED STATIONARY POINTS OF THE IMPLICIT LAGRANGIAN GLOBAL SOLUTIONS?

WHEN ARE THE (UN)CONSTRAINED STATIONARY POINTS OF THE IMPLICIT LAGRANGIAN GLOBAL SOLUTIONS? WHEN ARE THE (UN)CONSTRAINED STATIONARY POINTS OF THE IMPLICIT LAGRANGIAN GLOBAL SOLUTIONS? Francisco Facchinei a,1 and Christian Kanzow b a Università di Roma La Sapienza Dipartimento di Informatica e

More information

The effect of calmness on the solution set of systems of nonlinear equations

The effect of calmness on the solution set of systems of nonlinear equations Mathematical Programming manuscript No. (will be inserted by the editor) The effect of calmness on the solution set of systems of nonlinear equations Roger Behling Alfredo Iusem Received: date / Accepted:

More information

ON REGULARITY CONDITIONS FOR COMPLEMENTARITY PROBLEMS

ON REGULARITY CONDITIONS FOR COMPLEMENTARITY PROBLEMS ON REGULARITY CONDITIONS FOR COMPLEMENTARITY PROBLEMS A. F. Izmailov and A. S. Kurennoy December 011 ABSTRACT In the context of mixed complementarity problems various concepts of solution regularity are

More information

A PENALIZED FISCHER-BURMEISTER NCP-FUNCTION. September 1997 (revised May 1998 and March 1999)

A PENALIZED FISCHER-BURMEISTER NCP-FUNCTION. September 1997 (revised May 1998 and March 1999) A PENALIZED FISCHER-BURMEISTER NCP-FUNCTION Bintong Chen 1 Xiaojun Chen 2 Christian Kanzow 3 September 1997 revised May 1998 and March 1999 Abstract: We introduce a new NCP-function in order to reformulate

More information

UNDERGROUND LECTURE NOTES 1: Optimality Conditions for Constrained Optimization Problems

UNDERGROUND LECTURE NOTES 1: Optimality Conditions for Constrained Optimization Problems UNDERGROUND LECTURE NOTES 1: Optimality Conditions for Constrained Optimization Problems Robert M. Freund February 2016 c 2016 Massachusetts Institute of Technology. All rights reserved. 1 1 Introduction

More information

A QP-FREE CONSTRAINED NEWTON-TYPE METHOD FOR VARIATIONAL INEQUALITY PROBLEMS. Christian Kanzow 1 and Hou-Duo Qi 2

A QP-FREE CONSTRAINED NEWTON-TYPE METHOD FOR VARIATIONAL INEQUALITY PROBLEMS. Christian Kanzow 1 and Hou-Duo Qi 2 A QP-FREE CONSTRAINED NEWTON-TYPE METHOD FOR VARIATIONAL INEQUALITY PROBLEMS Christian Kanzow 1 and Hou-Duo Qi 2 1 University of Hamburg Institute of Applied Mathematics Bundesstrasse 55, D-20146 Hamburg,

More information

Technische Universität Dresden Institut für Numerische Mathematik

Technische Universität Dresden Institut für Numerische Mathematik Als Manuskript gedruckt Technische Universität Dresden Institut für Numerische Mathematik Convergence Conditions for Newton-type Methods Applied to Complementarity Systems with Nonisolated Solutions A.

More information

Optimality Conditions for Constrained Optimization

Optimality Conditions for Constrained Optimization 72 CHAPTER 7 Optimality Conditions for Constrained Optimization 1. First Order Conditions In this section we consider first order optimality conditions for the constrained problem P : minimize f 0 (x)

More information

Solving generalized semi-infinite programs by reduction to simpler problems.

Solving generalized semi-infinite programs by reduction to simpler problems. Solving generalized semi-infinite programs by reduction to simpler problems. G. Still, University of Twente January 20, 2004 Abstract. The paper intends to give a unifying treatment of different approaches

More information

First-order optimality conditions for mathematical programs with second-order cone complementarity constraints

First-order optimality conditions for mathematical programs with second-order cone complementarity constraints First-order optimality conditions for mathematical programs with second-order cone complementarity constraints Jane J. Ye Jinchuan Zhou Abstract In this paper we consider a mathematical program with second-order

More information

Numerical Optimization

Numerical Optimization Constrained Optimization Computer Science and Automation Indian Institute of Science Bangalore 560 012, India. NPTEL Course on Constrained Optimization Constrained Optimization Problem: min h j (x) 0,

More information

Implications of the Constant Rank Constraint Qualification

Implications of the Constant Rank Constraint Qualification Mathematical Programming manuscript No. (will be inserted by the editor) Implications of the Constant Rank Constraint Qualification Shu Lu Received: date / Accepted: date Abstract This paper investigates

More information

A Novel Inexact Smoothing Method for Second-Order Cone Complementarity Problems

A Novel Inexact Smoothing Method for Second-Order Cone Complementarity Problems A Novel Inexact Smoothing Method for Second-Order Cone Complementarity Problems Xiaoni Chi Guilin University of Electronic Technology School of Math & Comput Science Guilin Guangxi 541004 CHINA chixiaoni@126.com

More information

Newton-type Methods for Solving the Nonsmooth Equations with Finitely Many Maximum Functions

Newton-type Methods for Solving the Nonsmooth Equations with Finitely Many Maximum Functions 260 Journal of Advances in Applied Mathematics, Vol. 1, No. 4, October 2016 https://dx.doi.org/10.22606/jaam.2016.14006 Newton-type Methods for Solving the Nonsmooth Equations with Finitely Many Maximum

More information

system of equations. In particular, we give a complete characterization of the Q-superlinear

system of equations. In particular, we give a complete characterization of the Q-superlinear INEXACT NEWTON METHODS FOR SEMISMOOTH EQUATIONS WITH APPLICATIONS TO VARIATIONAL INEQUALITY PROBLEMS Francisco Facchinei 1, Andreas Fischer 2 and Christian Kanzow 3 1 Dipartimento di Informatica e Sistemistica

More information

A CHARACTERIZATION OF STRICT LOCAL MINIMIZERS OF ORDER ONE FOR STATIC MINMAX PROBLEMS IN THE PARAMETRIC CONSTRAINT CASE

A CHARACTERIZATION OF STRICT LOCAL MINIMIZERS OF ORDER ONE FOR STATIC MINMAX PROBLEMS IN THE PARAMETRIC CONSTRAINT CASE Journal of Applied Analysis Vol. 6, No. 1 (2000), pp. 139 148 A CHARACTERIZATION OF STRICT LOCAL MINIMIZERS OF ORDER ONE FOR STATIC MINMAX PROBLEMS IN THE PARAMETRIC CONSTRAINT CASE A. W. A. TAHA Received

More information

On the Coerciveness of Merit Functions for the Second-Order Cone Complementarity Problem

On the Coerciveness of Merit Functions for the Second-Order Cone Complementarity Problem On the Coerciveness of Merit Functions for the Second-Order Cone Complementarity Problem Guidance Professor Assistant Professor Masao Fukushima Nobuo Yamashita Shunsuke Hayashi 000 Graduate Course in Department

More information

Introduction to Optimization Techniques. Nonlinear Optimization in Function Spaces

Introduction to Optimization Techniques. Nonlinear Optimization in Function Spaces Introduction to Optimization Techniques Nonlinear Optimization in Function Spaces X : T : Gateaux and Fréchet Differentials Gateaux and Fréchet Differentials a vector space, Y : a normed space transformation

More information

FIXED POINT ITERATIONS

FIXED POINT ITERATIONS FIXED POINT ITERATIONS MARKUS GRASMAIR 1. Fixed Point Iteration for Non-linear Equations Our goal is the solution of an equation (1) F (x) = 0, where F : R n R n is a continuous vector valued mapping in

More information

FIRST- AND SECOND-ORDER OPTIMALITY CONDITIONS FOR MATHEMATICAL PROGRAMS WITH VANISHING CONSTRAINTS 1. Tim Hoheisel and Christian Kanzow

FIRST- AND SECOND-ORDER OPTIMALITY CONDITIONS FOR MATHEMATICAL PROGRAMS WITH VANISHING CONSTRAINTS 1. Tim Hoheisel and Christian Kanzow FIRST- AND SECOND-ORDER OPTIMALITY CONDITIONS FOR MATHEMATICAL PROGRAMS WITH VANISHING CONSTRAINTS 1 Tim Hoheisel and Christian Kanzow Dedicated to Jiří Outrata on the occasion of his 60th birthday Preprint

More information

5 Handling Constraints

5 Handling Constraints 5 Handling Constraints Engineering design optimization problems are very rarely unconstrained. Moreover, the constraints that appear in these problems are typically nonlinear. This motivates our interest

More information

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 4. Subgradient

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 4. Subgradient Shiqian Ma, MAT-258A: Numerical Optimization 1 Chapter 4 Subgradient Shiqian Ma, MAT-258A: Numerical Optimization 2 4.1. Subgradients definition subgradient calculus duality and optimality conditions Shiqian

More information

The general programming problem is the nonlinear programming problem where a given function is maximized subject to a set of inequality constraints.

The general programming problem is the nonlinear programming problem where a given function is maximized subject to a set of inequality constraints. 1 Optimization Mathematical programming refers to the basic mathematical problem of finding a maximum to a function, f, subject to some constraints. 1 In other words, the objective is to find a point,

More information

Constraint qualifications for nonlinear programming

Constraint qualifications for nonlinear programming Constraint qualifications for nonlinear programming Consider the standard nonlinear program min f (x) s.t. g i (x) 0 i = 1,..., m, h j (x) = 0 1 = 1,..., p, (NLP) with continuously differentiable functions

More information

Lecture 13 Newton-type Methods A Newton Method for VIs. October 20, 2008

Lecture 13 Newton-type Methods A Newton Method for VIs. October 20, 2008 Lecture 13 Newton-type Methods A Newton Method for VIs October 20, 2008 Outline Quick recap of Newton methods for composite functions Josephy-Newton methods for VIs A special case: mixed complementarity

More information

20 J.-S. CHEN, C.-H. KO AND X.-R. WU. : R 2 R is given by. Recently, the generalized Fischer-Burmeister function ϕ p : R2 R, which includes

20 J.-S. CHEN, C.-H. KO AND X.-R. WU. : R 2 R is given by. Recently, the generalized Fischer-Burmeister function ϕ p : R2 R, which includes 016 0 J.-S. CHEN, C.-H. KO AND X.-R. WU whereas the natural residual function ϕ : R R is given by ϕ (a, b) = a (a b) + = min{a, b}. Recently, the generalized Fischer-Burmeister function ϕ p : R R, which

More information

1. Introduction. We consider the general smooth constrained optimization problem:

1. Introduction. We consider the general smooth constrained optimization problem: OPTIMIZATION TECHNICAL REPORT 02-05, AUGUST 2002, COMPUTER SCIENCES DEPT, UNIV. OF WISCONSIN TEXAS-WISCONSIN MODELING AND CONTROL CONSORTIUM REPORT TWMCC-2002-01 REVISED SEPTEMBER 2003. A FEASIBLE TRUST-REGION

More information

1. Introduction The nonlinear complementarity problem (NCP) is to nd a point x 2 IR n such that hx; F (x)i = ; x 2 IR n + ; F (x) 2 IRn + ; where F is

1. Introduction The nonlinear complementarity problem (NCP) is to nd a point x 2 IR n such that hx; F (x)i = ; x 2 IR n + ; F (x) 2 IRn + ; where F is New NCP-Functions and Their Properties 3 by Christian Kanzow y, Nobuo Yamashita z and Masao Fukushima z y University of Hamburg, Institute of Applied Mathematics, Bundesstrasse 55, D-2146 Hamburg, Germany,

More information

Constrained Optimization and Lagrangian Duality

Constrained Optimization and Lagrangian Duality CIS 520: Machine Learning Oct 02, 2017 Constrained Optimization and Lagrangian Duality Lecturer: Shivani Agarwal Disclaimer: These notes are designed to be a supplement to the lecture. They may or may

More information

Laplace s Equation. Chapter Mean Value Formulas

Laplace s Equation. Chapter Mean Value Formulas Chapter 1 Laplace s Equation Let be an open set in R n. A function u C 2 () is called harmonic in if it satisfies Laplace s equation n (1.1) u := D ii u = 0 in. i=1 A function u C 2 () is called subharmonic

More information

Differentiable exact penalty functions for nonlinear optimization with easy constraints. Takuma NISHIMURA

Differentiable exact penalty functions for nonlinear optimization with easy constraints. Takuma NISHIMURA Master s Thesis Differentiable exact penalty functions for nonlinear optimization with easy constraints Guidance Assistant Professor Ellen Hidemi FUKUDA Takuma NISHIMURA Department of Applied Mathematics

More information

Affine scaling interior Levenberg-Marquardt method for KKT systems. C S:Levenberg-Marquardt{)KKTXÚ

Affine scaling interior Levenberg-Marquardt method for KKT systems. C S:Levenberg-Marquardt{)KKTXÚ 2013c6 $ Ê Æ Æ 117ò 12Ï June, 2013 Operations Research Transactions Vol.17 No.2 Affine scaling interior Levenberg-Marquardt method for KKT systems WANG Yunjuan 1, ZHU Detong 2 Abstract We develop and analyze

More information

Convex Optimization Theory. Chapter 5 Exercises and Solutions: Extended Version

Convex Optimization Theory. Chapter 5 Exercises and Solutions: Extended Version Convex Optimization Theory Chapter 5 Exercises and Solutions: Extended Version Dimitri P. Bertsekas Massachusetts Institute of Technology Athena Scientific, Belmont, Massachusetts http://www.athenasc.com

More information

Lecture 8 Plus properties, merit functions and gap functions. September 28, 2008

Lecture 8 Plus properties, merit functions and gap functions. September 28, 2008 Lecture 8 Plus properties, merit functions and gap functions September 28, 2008 Outline Plus-properties and F-uniqueness Equation reformulations of VI/CPs Merit functions Gap merit functions FP-I book:

More information

Lecture 19 Algorithms for VIs KKT Conditions-based Ideas. November 16, 2008

Lecture 19 Algorithms for VIs KKT Conditions-based Ideas. November 16, 2008 Lecture 19 Algorithms for VIs KKT Conditions-based Ideas November 16, 2008 Outline for solution of VIs Algorithms for general VIs Two basic approaches: First approach reformulates (and solves) the KKT

More information

Lecture 3. Optimization Problems and Iterative Algorithms

Lecture 3. Optimization Problems and Iterative Algorithms Lecture 3 Optimization Problems and Iterative Algorithms January 13, 2016 This material was jointly developed with Angelia Nedić at UIUC for IE 598ns Outline Special Functions: Linear, Quadratic, Convex

More information

1 The Observability Canonical Form

1 The Observability Canonical Form NONLINEAR OBSERVERS AND SEPARATION PRINCIPLE 1 The Observability Canonical Form In this Chapter we discuss the design of observers for nonlinear systems modelled by equations of the form ẋ = f(x, u) (1)

More information

Constraint Identification and Algorithm Stabilization for Degenerate Nonlinear Programs

Constraint Identification and Algorithm Stabilization for Degenerate Nonlinear Programs Preprint ANL/MCS-P865-1200, Dec. 2000 (Revised Nov. 2001) Mathematics and Computer Science Division Argonne National Laboratory Stephen J. Wright Constraint Identification and Algorithm Stabilization for

More information

ON GAP FUNCTIONS OF VARIATIONAL INEQUALITY IN A BANACH SPACE. Sangho Kum and Gue Myung Lee. 1. Introduction

ON GAP FUNCTIONS OF VARIATIONAL INEQUALITY IN A BANACH SPACE. Sangho Kum and Gue Myung Lee. 1. Introduction J. Korean Math. Soc. 38 (2001), No. 3, pp. 683 695 ON GAP FUNCTIONS OF VARIATIONAL INEQUALITY IN A BANACH SPACE Sangho Kum and Gue Myung Lee Abstract. In this paper we are concerned with theoretical properties

More information

First order optimality conditions for mathematical programs with second-order cone complementarity constraints

First order optimality conditions for mathematical programs with second-order cone complementarity constraints First order optimality conditions for mathematical programs with second-order cone complementarity constraints Jane J. Ye and Jinchuan Zhou April 9, 05 Abstract In this paper we consider a mathematical

More information

A Regularized Directional Derivative-Based Newton Method for Inverse Singular Value Problems

A Regularized Directional Derivative-Based Newton Method for Inverse Singular Value Problems A Regularized Directional Derivative-Based Newton Method for Inverse Singular Value Problems Wei Ma Zheng-Jian Bai September 18, 2012 Abstract In this paper, we give a regularized directional derivative-based

More information

TMA 4180 Optimeringsteori KARUSH-KUHN-TUCKER THEOREM

TMA 4180 Optimeringsteori KARUSH-KUHN-TUCKER THEOREM TMA 4180 Optimeringsteori KARUSH-KUHN-TUCKER THEOREM H. E. Krogstad, IMF, Spring 2012 Karush-Kuhn-Tucker (KKT) Theorem is the most central theorem in constrained optimization, and since the proof is scattered

More information

INTERIOR-POINT METHODS FOR NONCONVEX NONLINEAR PROGRAMMING: CONVERGENCE ANALYSIS AND COMPUTATIONAL PERFORMANCE

INTERIOR-POINT METHODS FOR NONCONVEX NONLINEAR PROGRAMMING: CONVERGENCE ANALYSIS AND COMPUTATIONAL PERFORMANCE INTERIOR-POINT METHODS FOR NONCONVEX NONLINEAR PROGRAMMING: CONVERGENCE ANALYSIS AND COMPUTATIONAL PERFORMANCE HANDE Y. BENSON, ARUN SEN, AND DAVID F. SHANNO Abstract. In this paper, we present global

More information

CONVERGENCE ANALYSIS OF AN INTERIOR-POINT METHOD FOR NONCONVEX NONLINEAR PROGRAMMING

CONVERGENCE ANALYSIS OF AN INTERIOR-POINT METHOD FOR NONCONVEX NONLINEAR PROGRAMMING CONVERGENCE ANALYSIS OF AN INTERIOR-POINT METHOD FOR NONCONVEX NONLINEAR PROGRAMMING HANDE Y. BENSON, ARUN SEN, AND DAVID F. SHANNO Abstract. In this paper, we present global and local convergence results

More information

SEMI-SMOOTH SECOND-ORDER TYPE METHODS FOR COMPOSITE CONVEX PROGRAMS

SEMI-SMOOTH SECOND-ORDER TYPE METHODS FOR COMPOSITE CONVEX PROGRAMS SEMI-SMOOTH SECOND-ORDER TYPE METHODS FOR COMPOSITE CONVEX PROGRAMS XIANTAO XIAO, YONGFENG LI, ZAIWEN WEN, AND LIWEI ZHANG Abstract. The goal of this paper is to study approaches to bridge the gap between

More information

Inexact alternating projections on nonconvex sets

Inexact alternating projections on nonconvex sets Inexact alternating projections on nonconvex sets D. Drusvyatskiy A.S. Lewis November 3, 2018 Dedicated to our friend, colleague, and inspiration, Alex Ioffe, on the occasion of his 80th birthday. Abstract

More information

Subgradient. Acknowledgement: this slides is based on Prof. Lieven Vandenberghes lecture notes. definition. subgradient calculus

Subgradient. Acknowledgement: this slides is based on Prof. Lieven Vandenberghes lecture notes. definition. subgradient calculus 1/41 Subgradient Acknowledgement: this slides is based on Prof. Lieven Vandenberghes lecture notes definition subgradient calculus duality and optimality conditions directional derivative Basic inequality

More information

The Relation Between Pseudonormality and Quasiregularity in Constrained Optimization 1

The Relation Between Pseudonormality and Quasiregularity in Constrained Optimization 1 October 2003 The Relation Between Pseudonormality and Quasiregularity in Constrained Optimization 1 by Asuman E. Ozdaglar and Dimitri P. Bertsekas 2 Abstract We consider optimization problems with equality,

More information

MODIFYING SQP FOR DEGENERATE PROBLEMS

MODIFYING SQP FOR DEGENERATE PROBLEMS PREPRINT ANL/MCS-P699-1097, OCTOBER, 1997, (REVISED JUNE, 2000; MARCH, 2002), MATHEMATICS AND COMPUTER SCIENCE DIVISION, ARGONNE NATIONAL LABORATORY MODIFYING SQP FOR DEGENERATE PROBLEMS STEPHEN J. WRIGHT

More information

On the Quadratic Convergence of the Cubic Regularization Method under a Local Error Bound Condition

On the Quadratic Convergence of the Cubic Regularization Method under a Local Error Bound Condition On the Quadratic Convergence of the Cubic Regularization Method under a Local Error Bound Condition Man-Chung Yue Zirui Zhou Anthony Man-Cho So Abstract In this paper we consider the cubic regularization

More information

Extreme Abridgment of Boyd and Vandenberghe s Convex Optimization

Extreme Abridgment of Boyd and Vandenberghe s Convex Optimization Extreme Abridgment of Boyd and Vandenberghe s Convex Optimization Compiled by David Rosenberg Abstract Boyd and Vandenberghe s Convex Optimization book is very well-written and a pleasure to read. The

More information

A strongly polynomial algorithm for linear systems having a binary solution

A strongly polynomial algorithm for linear systems having a binary solution A strongly polynomial algorithm for linear systems having a binary solution Sergei Chubanov Institute of Information Systems at the University of Siegen, Germany e-mail: sergei.chubanov@uni-siegen.de 7th

More information

A Continuation Method for the Solution of Monotone Variational Inequality Problems

A Continuation Method for the Solution of Monotone Variational Inequality Problems A Continuation Method for the Solution of Monotone Variational Inequality Problems Christian Kanzow Institute of Applied Mathematics University of Hamburg Bundesstrasse 55 D 20146 Hamburg Germany e-mail:

More information

A FRITZ JOHN APPROACH TO FIRST ORDER OPTIMALITY CONDITIONS FOR MATHEMATICAL PROGRAMS WITH EQUILIBRIUM CONSTRAINTS

A FRITZ JOHN APPROACH TO FIRST ORDER OPTIMALITY CONDITIONS FOR MATHEMATICAL PROGRAMS WITH EQUILIBRIUM CONSTRAINTS A FRITZ JOHN APPROACH TO FIRST ORDER OPTIMALITY CONDITIONS FOR MATHEMATICAL PROGRAMS WITH EQUILIBRIUM CONSTRAINTS Michael L. Flegel and Christian Kanzow University of Würzburg Institute of Applied Mathematics

More information

On the Local Quadratic Convergence of the Primal-Dual Augmented Lagrangian Method

On the Local Quadratic Convergence of the Primal-Dual Augmented Lagrangian Method Optimization Methods and Software Vol. 00, No. 00, Month 200x, 1 11 On the Local Quadratic Convergence of the Primal-Dual Augmented Lagrangian Method ROMAN A. POLYAK Department of SEOR and Mathematical

More information

Manual of ReSNA. matlab software for mixed nonlinear second-order cone complementarity problems based on Regularized Smoothing Newton Algorithm

Manual of ReSNA. matlab software for mixed nonlinear second-order cone complementarity problems based on Regularized Smoothing Newton Algorithm Manual of ReSNA matlab software for mixed nonlinear second-order cone complementarity problems based on Regularized Smoothing Newton Algorithm Shunsuke Hayashi September 4, 2013 1 Introduction ReSNA (Regularized

More information

AN ABADIE-TYPE CONSTRAINT QUALIFICATION FOR MATHEMATICAL PROGRAMS WITH EQUILIBRIUM CONSTRAINTS. Michael L. Flegel and Christian Kanzow

AN ABADIE-TYPE CONSTRAINT QUALIFICATION FOR MATHEMATICAL PROGRAMS WITH EQUILIBRIUM CONSTRAINTS. Michael L. Flegel and Christian Kanzow AN ABADIE-TYPE CONSTRAINT QUALIFICATION FOR MATHEMATICAL PROGRAMS WITH EQUILIBRIUM CONSTRAINTS Michael L. Flegel and Christian Kanzow University of Würzburg Institute of Applied Mathematics and Statistics

More information

1. Introduction. Consider the following parameterized optimization problem:

1. Introduction. Consider the following parameterized optimization problem: SIAM J. OPTIM. c 1998 Society for Industrial and Applied Mathematics Vol. 8, No. 4, pp. 940 946, November 1998 004 NONDEGENERACY AND QUANTITATIVE STABILITY OF PARAMETERIZED OPTIMIZATION PROBLEMS WITH MULTIPLE

More information

SOME STABILITY RESULTS FOR THE SEMI-AFFINE VARIATIONAL INEQUALITY PROBLEM. 1. Introduction

SOME STABILITY RESULTS FOR THE SEMI-AFFINE VARIATIONAL INEQUALITY PROBLEM. 1. Introduction ACTA MATHEMATICA VIETNAMICA 271 Volume 29, Number 3, 2004, pp. 271-280 SOME STABILITY RESULTS FOR THE SEMI-AFFINE VARIATIONAL INEQUALITY PROBLEM NGUYEN NANG TAM Abstract. This paper establishes two theorems

More information

A null-space primal-dual interior-point algorithm for nonlinear optimization with nice convergence properties

A null-space primal-dual interior-point algorithm for nonlinear optimization with nice convergence properties A null-space primal-dual interior-point algorithm for nonlinear optimization with nice convergence properties Xinwei Liu and Yaxiang Yuan Abstract. We present a null-space primal-dual interior-point algorithm

More information

Stationary Points of Bound Constrained Minimization Reformulations of Complementarity Problems1,2

Stationary Points of Bound Constrained Minimization Reformulations of Complementarity Problems1,2 JOURNAL OF OPTIMIZATION THEORY AND APPLICATIONS: Vol. 94, No. 2, pp. 449-467, AUGUST 1997 Stationary Points of Bound Constrained Minimization Reformulations of Complementarity Problems1,2 M. V. SOLODOV3

More information

Constrained Optimization

Constrained Optimization 1 / 22 Constrained Optimization ME598/494 Lecture Max Yi Ren Department of Mechanical Engineering, Arizona State University March 30, 2015 2 / 22 1. Equality constraints only 1.1 Reduced gradient 1.2 Lagrange

More information

Nonlinear Programming, Elastic Mode, SQP, MPEC, MPCC, complementarity

Nonlinear Programming, Elastic Mode, SQP, MPEC, MPCC, complementarity Preprint ANL/MCS-P864-1200 ON USING THE ELASTIC MODE IN NONLINEAR PROGRAMMING APPROACHES TO MATHEMATICAL PROGRAMS WITH COMPLEMENTARITY CONSTRAINTS MIHAI ANITESCU Abstract. We investigate the possibility

More information

AN AUGMENTED LAGRANGIAN AFFINE SCALING METHOD FOR NONLINEAR PROGRAMMING

AN AUGMENTED LAGRANGIAN AFFINE SCALING METHOD FOR NONLINEAR PROGRAMMING AN AUGMENTED LAGRANGIAN AFFINE SCALING METHOD FOR NONLINEAR PROGRAMMING XIAO WANG AND HONGCHAO ZHANG Abstract. In this paper, we propose an Augmented Lagrangian Affine Scaling (ALAS) algorithm for general

More information

On the Quadratic Convergence of the Cubic Regularization Method under a Local Error Bound Condition

On the Quadratic Convergence of the Cubic Regularization Method under a Local Error Bound Condition On the Quadratic Convergence of the Cubic Regularization Method under a Local Error Bound Condition Man-Chung Yue Zirui Zhou Anthony Man-Cho So Abstract In this paper we consider the cubic regularization

More information

1. Introduction. We consider the mathematical programming problem

1. Introduction. We consider the mathematical programming problem SIAM J. OPTIM. Vol. 15, No. 1, pp. 210 228 c 2004 Society for Industrial and Applied Mathematics NEWTON-TYPE METHODS FOR OPTIMIZATION PROBLEMS WITHOUT CONSTRAINT QUALIFICATIONS A. F. IZMAILOV AND M. V.

More information

WEIERSTRASS THEOREMS AND RINGS OF HOLOMORPHIC FUNCTIONS

WEIERSTRASS THEOREMS AND RINGS OF HOLOMORPHIC FUNCTIONS WEIERSTRASS THEOREMS AND RINGS OF HOLOMORPHIC FUNCTIONS YIFEI ZHAO Contents. The Weierstrass factorization theorem 2. The Weierstrass preparation theorem 6 3. The Weierstrass division theorem 8 References

More information

Primal-dual relationship between Levenberg-Marquardt and central trajectories for linearly constrained convex optimization

Primal-dual relationship between Levenberg-Marquardt and central trajectories for linearly constrained convex optimization Primal-dual relationship between Levenberg-Marquardt and central trajectories for linearly constrained convex optimization Roger Behling a, Clovis Gonzaga b and Gabriel Haeser c March 21, 2013 a Department

More information

A SHIFTED PRIMAL-DUAL PENALTY-BARRIER METHOD FOR NONLINEAR OPTIMIZATION

A SHIFTED PRIMAL-DUAL PENALTY-BARRIER METHOD FOR NONLINEAR OPTIMIZATION A SHIFTED PRIMAL-DUAL PENALTY-BARRIER METHOD FOR NONLINEAR OPTIMIZATION Philip E. Gill Vyacheslav Kungurtsev Daniel P. Robinson UCSD Center for Computational Mathematics Technical Report CCoM-19-3 March

More information

Radius Theorems for Monotone Mappings

Radius Theorems for Monotone Mappings Radius Theorems for Monotone Mappings A. L. Dontchev, A. Eberhard and R. T. Rockafellar Abstract. For a Hilbert space X and a mapping F : X X (potentially set-valued) that is maximal monotone locally around

More information

Department of Social Systems and Management. Discussion Paper Series

Department of Social Systems and Management. Discussion Paper Series Department of Social Systems and Management Discussion Paper Series No. 1262 Complementarity Problems over Symmetric Cones: A Survey of Recent Developments in Several Aspects by Akiko YOSHISE July 2010

More information

CONVERGENCE PROPERTIES OF COMBINED RELAXATION METHODS

CONVERGENCE PROPERTIES OF COMBINED RELAXATION METHODS CONVERGENCE PROPERTIES OF COMBINED RELAXATION METHODS Igor V. Konnov Department of Applied Mathematics, Kazan University Kazan 420008, Russia Preprint, March 2002 ISBN 951-42-6687-0 AMS classification:

More information

A GLOBALLY CONVERGENT STABILIZED SQP METHOD: SUPERLINEAR CONVERGENCE

A GLOBALLY CONVERGENT STABILIZED SQP METHOD: SUPERLINEAR CONVERGENCE A GLOBALLY CONVERGENT STABILIZED SQP METHOD: SUPERLINEAR CONVERGENCE Philip E. Gill Vyacheslav Kungurtsev Daniel P. Robinson UCSD Center for Computational Mathematics Technical Report CCoM-14-1 June 30,

More information

LAGRANGIAN TRANSFORMATION IN CONVEX OPTIMIZATION

LAGRANGIAN TRANSFORMATION IN CONVEX OPTIMIZATION LAGRANGIAN TRANSFORMATION IN CONVEX OPTIMIZATION ROMAN A. POLYAK Abstract. We introduce the Lagrangian Transformation(LT) and develop a general LT method for convex optimization problems. A class Ψ of

More information

Chapter 1. Optimality Conditions: Unconstrained Optimization. 1.1 Differentiable Problems

Chapter 1. Optimality Conditions: Unconstrained Optimization. 1.1 Differentiable Problems Chapter 1 Optimality Conditions: Unconstrained Optimization 1.1 Differentiable Problems Consider the problem of minimizing the function f : R n R where f is twice continuously differentiable on R n : P

More information

SOLUTION OF NONLINEAR COMPLEMENTARITY PROBLEMS

SOLUTION OF NONLINEAR COMPLEMENTARITY PROBLEMS A SEMISMOOTH EQUATION APPROACH TO THE SOLUTION OF NONLINEAR COMPLEMENTARITY PROBLEMS Tecla De Luca 1, Francisco Facchinei 1 and Christian Kanzow 2 1 Universita di Roma \La Sapienza" Dipartimento di Informatica

More information

A globally and R-linearly convergent hybrid HS and PRP method and its inexact version with applications

A globally and R-linearly convergent hybrid HS and PRP method and its inexact version with applications A globally and R-linearly convergent hybrid HS and PRP method and its inexact version with applications Weijun Zhou 28 October 20 Abstract A hybrid HS and PRP type conjugate gradient method for smooth

More information

UC Berkeley Department of Electrical Engineering and Computer Science. EECS 227A Nonlinear and Convex Optimization. Solutions 5 Fall 2009

UC Berkeley Department of Electrical Engineering and Computer Science. EECS 227A Nonlinear and Convex Optimization. Solutions 5 Fall 2009 UC Berkeley Department of Electrical Engineering and Computer Science EECS 227A Nonlinear and Convex Optimization Solutions 5 Fall 2009 Reading: Boyd and Vandenberghe, Chapter 5 Solution 5.1 Note that

More information

SEMISMOOTH LEAST SQUARES METHODS FOR COMPLEMENTARITY PROBLEMS

SEMISMOOTH LEAST SQUARES METHODS FOR COMPLEMENTARITY PROBLEMS SEMISMOOTH LEAST SQUARES METHODS FOR COMPLEMENTARITY PROBLEMS Dissertation zur Erlangung des naturwissenschaftlichen Doktorgrades der Bayerischen Julius Maximilians Universität Würzburg vorgelegt von STEFANIA

More information

YURI LEVIN, MIKHAIL NEDIAK, AND ADI BEN-ISRAEL

YURI LEVIN, MIKHAIL NEDIAK, AND ADI BEN-ISRAEL Journal of Comput. & Applied Mathematics 139(2001), 197 213 DIRECT APPROACH TO CALCULUS OF VARIATIONS VIA NEWTON-RAPHSON METHOD YURI LEVIN, MIKHAIL NEDIAK, AND ADI BEN-ISRAEL Abstract. Consider m functions

More information

1. Nonlinear Equations. This lecture note excerpted parts from Michael Heath and Max Gunzburger. f(x) = 0

1. Nonlinear Equations. This lecture note excerpted parts from Michael Heath and Max Gunzburger. f(x) = 0 Numerical Analysis 1 1. Nonlinear Equations This lecture note excerpted parts from Michael Heath and Max Gunzburger. Given function f, we seek value x for which where f : D R n R n is nonlinear. f(x) =

More information

An Alternative Proof of Primitivity of Indecomposable Nonnegative Matrices with a Positive Trace

An Alternative Proof of Primitivity of Indecomposable Nonnegative Matrices with a Positive Trace An Alternative Proof of Primitivity of Indecomposable Nonnegative Matrices with a Positive Trace Takao Fujimoto Abstract. This research memorandum is aimed at presenting an alternative proof to a well

More information

Using exact penalties to derive a new equation reformulation of KKT systems associated to variational inequalities

Using exact penalties to derive a new equation reformulation of KKT systems associated to variational inequalities Using exact penalties to derive a new equation reformulation of KKT systems associated to variational inequalities Thiago A. de André Paulo J. S. Silva March 24, 2007 Abstract In this paper, we present

More information

The Karush-Kuhn-Tucker (KKT) conditions

The Karush-Kuhn-Tucker (KKT) conditions The Karush-Kuhn-Tucker (KKT) conditions In this section, we will give a set of sufficient (and at most times necessary) conditions for a x to be the solution of a given convex optimization problem. These

More information

Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem

Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem Michael Patriksson 0-0 The Relaxation Theorem 1 Problem: find f := infimum f(x), x subject to x S, (1a) (1b) where f : R n R

More information

ON A CLASS OF NONSMOOTH COMPOSITE FUNCTIONS

ON A CLASS OF NONSMOOTH COMPOSITE FUNCTIONS MATHEMATICS OF OPERATIONS RESEARCH Vol. 28, No. 4, November 2003, pp. 677 692 Printed in U.S.A. ON A CLASS OF NONSMOOTH COMPOSITE FUNCTIONS ALEXANDER SHAPIRO We discuss in this paper a class of nonsmooth

More information

4TE3/6TE3. Algorithms for. Continuous Optimization

4TE3/6TE3. Algorithms for. Continuous Optimization 4TE3/6TE3 Algorithms for Continuous Optimization (Duality in Nonlinear Optimization ) Tamás TERLAKY Computing and Software McMaster University Hamilton, January 2004 terlaky@mcmaster.ca Tel: 27780 Optimality

More information

2.3 Linear Programming

2.3 Linear Programming 2.3 Linear Programming Linear Programming (LP) is the term used to define a wide range of optimization problems in which the objective function is linear in the unknown variables and the constraints are

More information

Convex Analysis and Optimization Chapter 2 Solutions

Convex Analysis and Optimization Chapter 2 Solutions Convex Analysis and Optimization Chapter 2 Solutions Dimitri P. Bertsekas with Angelia Nedić and Asuman E. Ozdaglar Massachusetts Institute of Technology Athena Scientific, Belmont, Massachusetts http://www.athenasc.com

More information

Some new facts about sequential quadratic programming methods employing second derivatives

Some new facts about sequential quadratic programming methods employing second derivatives To appear in Optimization Methods and Software Vol. 00, No. 00, Month 20XX, 1 24 Some new facts about sequential quadratic programming methods employing second derivatives A.F. Izmailov a and M.V. Solodov

More information

Key words. linear complementarity problem, non-interior-point algorithm, Tikhonov regularization, P 0 matrix, regularized central path

Key words. linear complementarity problem, non-interior-point algorithm, Tikhonov regularization, P 0 matrix, regularized central path A GLOBALLY AND LOCALLY SUPERLINEARLY CONVERGENT NON-INTERIOR-POINT ALGORITHM FOR P 0 LCPS YUN-BIN ZHAO AND DUAN LI Abstract Based on the concept of the regularized central path, a new non-interior-point

More information

Iteration-complexity of first-order penalty methods for convex programming

Iteration-complexity of first-order penalty methods for convex programming Iteration-complexity of first-order penalty methods for convex programming Guanghui Lan Renato D.C. Monteiro July 24, 2008 Abstract This paper considers a special but broad class of convex programing CP)

More information

Existence of minimizers

Existence of minimizers Existence of imizers We have just talked a lot about how to find the imizer of an unconstrained convex optimization problem. We have not talked too much, at least not in concrete mathematical terms, about

More information

Sharpening the Karush-John optimality conditions

Sharpening the Karush-John optimality conditions Sharpening the Karush-John optimality conditions Arnold Neumaier and Hermann Schichl Institut für Mathematik, Universität Wien Strudlhofgasse 4, A-1090 Wien, Austria email: Arnold.Neumaier@univie.ac.at,

More information

A note on upper Lipschitz stability, error bounds, and critical multipliers for Lipschitz-continuous KKT systems

A note on upper Lipschitz stability, error bounds, and critical multipliers for Lipschitz-continuous KKT systems Math. Program., Ser. A (2013) 142:591 604 DOI 10.1007/s10107-012-0586-z SHORT COMMUNICATION A note on upper Lipschitz stability, error bounds, and critical multipliers for Lipschitz-continuous KKT systems

More information

Pacific Journal of Optimization (Vol. 2, No. 3, September 2006) ABSTRACT

Pacific Journal of Optimization (Vol. 2, No. 3, September 2006) ABSTRACT Pacific Journal of Optimization Vol., No. 3, September 006) PRIMAL ERROR BOUNDS BASED ON THE AUGMENTED LAGRANGIAN AND LAGRANGIAN RELAXATION ALGORITHMS A. F. Izmailov and M. V. Solodov ABSTRACT For a given

More information

ON LICQ AND THE UNIQUENESS OF LAGRANGE MULTIPLIERS

ON LICQ AND THE UNIQUENESS OF LAGRANGE MULTIPLIERS ON LICQ AND THE UNIQUENESS OF LAGRANGE MULTIPLIERS GERD WACHSMUTH Abstract. Kyparisis proved in 1985 that a strict version of the Mangasarian- Fromovitz constraint qualification (MFCQ) is equivalent to

More information

On the Local Convergence of Regula-falsi-type Method for Generalized Equations

On the Local Convergence of Regula-falsi-type Method for Generalized Equations Journal of Advances in Applied Mathematics, Vol., No. 3, July 017 https://dx.doi.org/10.606/jaam.017.300 115 On the Local Convergence of Regula-falsi-type Method for Generalized Equations Farhana Alam

More information