A Global Regularization Method for Solving. the Finite Min-Max Problem. O. Barrientos LAO June 1996

Size: px
Start display at page:

Download "A Global Regularization Method for Solving. the Finite Min-Max Problem. O. Barrientos LAO June 1996"

Transcription

1 A Global Regularization Method for Solving the Finite Min-Max Problem O. Barrientos LAO June 1996 LABORATOIRE APPROXIMATION ET OPTIMISATION Universite Paul Sabatier, 118 route de Narbonne Toulouse cedex - France Tel lao@cict.fr

2 A Global Regularization Method for Solving the Finite Min-Max Problem 1 O. BARRIENTOS 2 Abstract. In this paper, we present a method for solving the nite nonlinear min-max problem. By using quasi-newton methods, we approximately solve a sequence of dierentiable subproblems where, for each subproblem, the cost function to minimize is a global regularization underestimating the nite maximum function. We show that every cluster point of the sequence generated is a stationary point of the min-max problem, and therefore, in the convex case, is a solution of the problem. Moreover, we give numerical results for a large set of test problems which show that the method is ecient in practice. Key words: min-max, regularization technique, nondierentiable optimization, quasi-newton methods. where 1. Introduction The so-called nite min-max problem can be stated as follows: '(x) = min '(x); (1) x2ir n max j2f1;:::;mg f j(x); (2) and for each j 2 f1; :::; mg the function f j is dened on IR n. In this work, we suppose that all the functions f j are continuously dierentiable on IR n. It is a well known fact that the function ' is generally nondierentiable at points where the maximum in (2) is attained for more than one function f j. In this paper, we present a method to solve the problem (1) by generating dierentiable subproblems whose solutions cluster to stationary points of the original problem. These subproblems are obtained by a global regularization of the function ', while for obtaining the approximated solutions of the subproblems we use a quasi-newton method. This regularization technique was introduced by Ggola & Gomez in [1]; we have added one parameter what will allow us to show, for example, that the sequence generated by the method is bounded. Moreover, in the convex case, we establish that this global regularization of the function ' is essentially the Moreau-Yosida's regularization of the support function of the 1 This work was accomplished while the author was visiting the Laboratoire d'approximation et Optimisation at Universite Paul Sabatier, Toulouse, France. The author is indebted to Professor J.-B. Hiriart-Urruty for bringing to his attention the work of Ggola & Gomez referenced in this paper and for supervising this work. In addition his lucid comments substantially improved this article. 2 Depto. Ingeniera Matematica, Universidad de Chile, Casilla 170/3 Correo 3, Santiago, CHILE. 1

3 unit-simplex in IR m. We also extend to the nonconvex case the results shown in [1] in a clear, simple and direct fashion. On the other hand, some other methods have been proposed in the literature to solve the min-max problem. Most of them use the function ' as a `merit function'. In order to nd a descent direction, Charalambous & Conn [2] and Demyanov & Malozemov [3] use the "?feasibility concept, Han [4] and Murray & Overton [5] solve quadratic programming subproblems. Zang [6] uses a local regularization-type method to approximate ' by smooth functions. The paper is organised as follows: in the second section, we dene the regularized function ' " of the function ', this function will be dened on IR m IR n and we say that IR m is the space of the dual variables. We also show some properties of the regularized function. Next, in section 3, we analyse two dierent theoretical methods for solving the nite min-max problem. First, a primal method, in the sense that we consider the dual variables as parameters. The method consists in the approximated minimization of dierentiable subproblems dened via the regularized function ' ". We show that every cluster point of the sequence of approximated solutions of these subproblems is a stationary point of the problem (1) (see Theorem 3.1 in section 3). Next, a primal-dual method based on the proximal method for saddle functions. In this case, the method is an application of the proximal point theory developed by Martinet in [7] and Rockafellar in [8]. We show that the sequence generated by the method converges to a solution point of the min-max problem (see Proposition 3.4 in section 3). Unfortunately, we could not show that the conditions for obtaining this result are attainable in practice. In section 4, we rstly explain the algorithm proposed to solve the min-max problem and we next describe in detail the method as it is implemented. Finally, in the last section, we show the performance of the proposed method by testing rst, the well known so-called MAXQUAD problem and next, a large set of test problems where the cost function is a nite maximum of random (convex and nonconvex) quadratic functions. 2. Regularization technique In this part, we present the regularization technique that we use to obtain the regularized function underestimating the function '. For simplicity, we consider m convex functions f j : IR n! IR j 2 f1; :::; mg, but it is easy to see that this property is not necessary to obtain the results we are going to derive. Moreover, if we dene the set (unit-simplex of IR n ) U = fu 2 IR m : mx u j = 1; u j 0 for all j 2 f1; :::; mgg; (3) 2

4 it is straightforward to see that the function ' dened in (2) veries the relation '(x) = max u2u mx We now dene the function # : IR m! IR by #(y) = u j f j (x): (4) max j2f1;:::;mg y j for all y = (y 1 ; :::; y m ) T 2 IR m ; (5) and we therefore have the equality '(x) = #(f 1 (x); :::; f m (x)) for all x 2 IR n. On the other hand, for each " > 0 and for each v 2 IR m, we consider the function # " (v; ) : IR m! IR dened by # " (v; y) = max u2u f<u; y>? " 2 ku? vk2 g for all y 2 IR m : (6) We will subsequent see that the function # " (v; ) is a global regularization of the function #. This regularization technique was introduced, in the case " = 1, by Ggola and Gomez in [1]. We will also see that the introduction of the parameter " has interesting properties from the algorithmic viewpoint (see for example Proposition 3.1, Corollary 3.1 or Theorem 3.1 in section 3). We then can write the equality # " (v; y) = sup u2ir m f<u; y>? [ " 2 ku? vk2 + 0 (u)] g; where 0 denotes the indicator function of the set U, that is, 0 (u) = 0 if u 2 U, 0(u) = +1 otherwise. We now consider the perturbated function " of the function 0 dened by (v; u) 2 IR m IR m?! " (v; u) = " 2 ku? vk2 + 0 (u) " > 0; we then have the following equalities # " (v; y) =[ " (v; )] (y) = h? 1 + U _ <v; > + 2" k k2 i (y) =? " 2 kvk2 + inf u2ir mf U(u) + 1 2" k("v + y)? uk2 g; (7) where [ " (v; )] denotes the Legendre-Fenchel conjugate of the function " (v; ), U is the support function of the set U, that is, (y) = max <u; y> for all y 2 u2u IRm, and + _ denotes the inmal convolution operation. 3

5 Thus, if T " denotes the Moreau-Yosida's regularization of the function U, we then can write the following equality # " (v; y) =? " 2 kvk2 + T " ("v + y) for all y 2 IR m ; (8) it tells us that the function # " (v; ) is essentially the Moreau-Yosida's regularization of the function U. We now show some properties of the functions # and # ". In particular, we prove that the function # " (v; ) is a global regularization of the function #. Proposition 2.1 i) For all " > 0 and for all v 2 IR m, the function # " (v; ) is convex dierentiable and its gradient mapping is Lipschitz on IR m. ii) For all " > 0, we have the equality #(y) = max # "(v; y) for all y 2 IR m. v2u iii) For all " > 0 and for all v 2 U, we have the inequality #(y)? " 2 h 1+kvk 2?2 min j2f1;:::;mg v j i # " (v; y) #(y) for all y 2 IR m : (9) Proof. i) The convexity of the function # " (v; ) follows from the rst equality in (7), whereas its regularity follows from the equality (8). ii) It follows from the denition of the function # " (v; ) in (6). iii) The inequality in the right of (9) is straightforward from the denition (6), whereas the inequality in the left follows from the following inequality y j? " 2 hx i6=j v 2 i + (1? v j ) 2i # " (v; y) for all j 2 f1; :::; mg: From now on, for each " > 0, we consider the regularized function ' " of the function ' dened on IR m IR n as ' " (v; x) = # " (v; f(x)) = max u2u f<u; f(x)>? " 2 ku? vk2 g; (10) where f(x) = (f 1 (x); :::; f m (x)) T 2 IR m for all x 2 IR n. For simplicity, for each " > 0, for each x 2 IR n and for each v 2 IR m, we denote by u(v; x) the element in U verifying the following equality ' " (v; x) = <u(v; x); f(x)>? " 2 ku(v; x)? vk2 (11) (for calculation details, see section 4.2). 4

6 Moreover, from the inequality (9), it is easy to deduce the following inequality '(x)? " ' " (v; x) '(x) for all (v; x) 2 U IR n : (12) We next show some properties of the function ' ". In particular, we prove that the function ' " is indeed a global regularization of the function '. Proposition 2.2 Let us suppose that all the functions f j are dierentiable on IR n. Then, i) For all " > 0 and for all v 2 IR m, the function ' " (v; ) is dierentiable on IR P n and r x ' " (v; x) = m u j (v; x)rf j (x) where u(v; x) is dened in (11). ii) For all " > 0 and for all x 2 IR n, the function ' " (; x) is concave dierentiable on IR m and r v ' " (v; x) = "[u(v; x)? v] where u(v; x) is dened in (11). iii) For all " > 0, for all x 2 IR n, for all v 1 ; v 2 2 IR m and for all 2 [0; 1], we have the following inequality ' " (v 1 + (1? )v 2 ; x) ' " (v 1 ; x) + (1? )' " (v 2 ; x) + 1 2" (1? )kr v' " (v 1 ; x)? r v ' " (v 2 ; x)k 2 : (13) Proof. i) The dierentiability of the function ' " (v; ) follows from the dierentiability of the functions # " (v; ) and f j for all j 2 f1; :::; mg. ii) The dierentiability of the function ' " (; x) follows from the denition (10), whereas its concavity will be a consequence of the inequality (13). iii) Given " > 0, x 2 IR n, v 1 ; v 2 2 IR m and 2 [0; 1], from the denition (10), we have the following inequality ' " (v 1 + (1? )v 2 ; x) <u; f(x)>? " 2 kv 1 + (1? )v 2? uk 2 for all u 2 U: We now set u = u 1 + (1? )u 2 with u 1 ; u 2 2 U for obtaining the equality <u; f(x)>? " 2 kv 1 + (1? )v 2? uk 2 = h<u 1 ; f(x)>? " 2 kv 1? u 1 k 2i h + (1? ) <u 2 ; f(x)>? " 2 kv 2? u 2 k 2i + " 2 (1? )k(v 1? u 1 )? (v 2? u 2 )k 2 ; and we end the proof by considering u k dened in (11). = u(v k ; x) k = 1; 2 where u(v k ; x) is 5

7 3. Convergence Results In this part, we rstly are interested in knowing if there exists or not a relation between the set of minima points of the functions ' " (v; ) and the set of minima points of the function '. Next, we will also study if there exists or not a relation between the set of the saddle points of the problem max min ' "(v; x) and the set v2u x2ir n of saddle points of the problem max min <v; f(x)>. v2u x2ir n For simplicity, we suppose that the m functions f j are continuous on IR n and there exists a function f j0 coercive on IR n some j 0 2 f1; :::; mg, that is, lim kxk!+1 f j 0 (x) = +1. Then, for each " > 0 and for each v 2 U, the set of minima points of the functions ' " (v; ) and ' are nonempty. Proposition 3.1 Let us consider the three following sequences: f" k g a sequence of positive real numbers converging to zero, fv k g a sequence in U and ff k g the sequence of functions dened on IR n by If we dene the sequence fx k g IR n as verifying f k (x) = ' "k (v k ; x) for all x 2 IR n : (14) f k (x k ) = min x2ir n f k(x): Then, every cluster point of the sequence fx k g is a solution point of the problem (1). Proof. To prove the proposition we use the following result due to Attouch & Wets: Proposition. (See Ref. [9] or Ref. [10]) If x k is a minimum point of the function f k and if the sequence ff k g epiconverges to f, then every cluster point of the sequence fx k g is contained in the set of minima points of the function f. } Therefore, we only must show that the sequence ff k = ' "k (v k ; )g epiconverges to the function ', but this follows from the inequality (12) and the continuity of the functions ' and ' ". Consequently, we give a sucient condition ensuring that the sequence fx k g is bounded. Note that the second condition is veried if the point x k+1 is simply obtained by using any descent method on the function ' "k (v k ; ) starting from the point x k. 6

8 Proposition 3.2 Let us consider the three following sequences: f" k g a sequence of positive real numbers, fv k g a sequence in U and fx k g a sequence in IR n verifying the following properties: i) P k2in " k < +1, ii) " k + ' "k (v k ; x k ) ' "k (v k ; x k+1 ) for all k 2 IN. Then, the sequence fx k g is contained in the set fx2ir n : '(x)'(x 0 )+2 P k2in " k g. Proof. From the inequality (12) and condition ii), we obtain the following inequality '(x k ) + " k ' "k (v k ; x k+1 ) for all k 2 IN : By using again the inequality (12), we can write '(x k ) + 2" k '(x k+1 ) for all k 2 IN, and we therefore conclude '(x i ) '(x 0 ) + 2 X k2in " k for all i 2 IN : Corollary 3.1 If, in addition to the hypotheses in Proposition 3.2, we suppose that there exists a function f j0 coercive on IR n some j 0 2 f1; :::; mg. Then, the sequence fx k g is bounded. We now present an interesting property of the sequence fv k g U when v k is `optimally' chosen. It tells us that the cluster points of this sequence are always optimal points for the limits of the primal sequence fx k g. Moreover, we characterize the stationarity condition of a point x 2 IR n. More precisely, we have: Lemma 3.1 i) Given a sequence f" k g converging to zero and a sequence fx k g IR n. Let us consider the sequence fv k g U recursively dened by v 0 2 U and v k+1 = u(v k ; x k+1 ) where u(v k ; x k+1 ) is dened in (11). Then, every cluster point (v ; x ) 2 UIR n of the sequence f(v k ; x k )g satises the following relationship '(x ) = <v ; f(x )>: ii) If, in addition to the above equality, the point (v ; x ) 2 U IR n veries mp v j rf j(x ) = 0. Then, 0 ), ) denotes the Clarke's subdierential of the function ' at point x (see Clarke [11]). 7

9 Proof. i) From the denitions (10) and (11), for every u 2 U, we have <v k+1 ; f(x k+1 )>? " k 2 kvk+1? v k k 2 <u; f(x k+1 )>? " k 2 ku? vk k 2 : Now, let us suppose that the sequence f(v k+1 ; x k+1 )g converges to the point (v ; x ); then, by taking limits, we obtain <v ; f(x )> <u; f(x )> for all u 2 U, that is, '(x ) = <v ; f(x )>. ii) It is straightforward by remembering that the rst-order directional derivative ' 0 (x ; d) of the function ' at the point x 2 IR n in the direction d 2 IR n veries the following formula ' 0 (x ; d) = max mx v j <rf j (x ); d>; v2i(x ;d) where the set I(x ; d) = fv 2 U : '(x ) = <v; f(x )>g. In Proposition 3.1, we have shown that the minima points of the functions f k (dened in (14)) cluster to minima points of the function '. From a practical viewpoint, this result is not interesting because the cost for obtaining one minimum point of the function f k may be as expensive as to solve the problem (1). Consequently, we next show one more (computationally) implementable version of this proposition which will be the base to design our algorithm for solving the problem (1) in section 4. Theorem 3.1 Let us suppose that all the functions f j are continuously dierentiable on IR n. Given two sequences f k g, f" k g converging to zero and a point (v 0 ; x 0 ) 2 U IR n, let us consider the sequence f(v k ; x k )g U IR n recursively dened as follows: a) Starting from the point x k, the function ' "k (v k ; ) is minimized until nding a point x k+1 2 IR n verifying kr x ' "k (v k ; x k+1 )k k ; (15) b) v k+1 = u(v k ; x k+1 ) where u(v k ; x k+1 ) is dened in (11). Then, every cluster point x 2 IR n of the sequence fx k g is a stationary point of the problem (1), that is, the point x veries 0 ), ) denotes the Clarke's subdierential of the function ' at point x (see Clarke [11]). 8

10 Proof. Let fx k0 g be the subsequence of fx k g converging to x. From the compacity of the set U, we can suppose that the subsequence fv k0 g of fv k g converges to v 2 U. From Proposition 2.2, i) and (15), we have the following inequality kr x ' "k 0?1 (v k0?1 ; x k 0 )k = k and by taking limits, we obtain mp mx v j rf j(x ) = 0. v k0 j rf j (x k0 )k k 0?1; On the other hand, from Lemma 3.1, we know that the point (v ; x ) veries the relationship '(x ) = <v ; f(x )>. We therefore can conclude that 0 = mx v j rf j(x ) ): P Corollary 3.2 If, in addition to the hypotheses in the above theorem, we suppose that the sequence f" k g veries " k < +1, all the functions f j are convex on IR n, and there exists f j0 k2in coercive on IR n some j 0 2 f1; :::; mg. Then, every cluster point of the sequence fx k g is a solution point of the problem (1). Moreover, we have the following equalities lim k!+1 ' " k (v k ; x k+1 ) = lim k!+1 '(xk+1 ) = min '(x): x2ir n Proof. Firstly, from Corollary 3.1, we know that the sequence fx k g is bounded. Now, given x 1 ; x 2 2 IR n, x 1 6= x 2, two cluster points of the sequence fx k g, from the above theorem we have that 0 1)\@'(x 2). Since ' is a convex function, that means that both points are solutions of the problem (1), that is, = '(x 1) = '(x 2) = min x2ir n '(x): From the above reasoning, we also can deduce that the sequence f'(x k+1 )g is bounded and that it has as unique cluster point, it therefore converges to. Finally, by using (12), the sequence f' "k (v k ; x k+1 )g converges to. P Remark. In Theorem 3.1, if we suppose that the sequence fx k g is bounded (from Corollary 3.1, this holds if " k < +1) and the function ' has only one k2in stationary point x 2 IR n, then the sequence fx k g converges to x. 9

11 So far, we have used v 2 U as a parameter, we now are interested to know what are the consequences when we consider v as an independent variable. In this way, a well known relation exists between the solution points of the problem (1) and the saddle points of the problem max v2u min <v; f(x)>. Then, it is interesting n min ' "(v; x) and if there x2ir n min <v; f(x)>. n x2ir to know what are the saddle points of the problem max v2u exists any relation with those of the problem max v2u x2ir We next completely characterise the saddle points of these problems. Proposition 3.3 Let us suppose that all the functions f j are convex dierentiable on IR n. Given " > 0, let us consider the two following problems: (i) max v2u min <v; f(x)> (ii) max min ' n "(v; x): v2u x2ir n x2ir Then, the following statements are equivalent: a) (v ; x ) 2 U IR n is a saddle point of the problem (i). b) (v ; x ) 2 U IR n is a saddle point of the problem (ii). c) (v ; x ) 2 U IR n is a point verifying the equalities m P <v ; f(x )> = '(x ). v j rf j(x ) = 0 and Proof. a) ) b). Let (v ; x ) 2 U IR n be a point verifying a), that is, <u; f(x )> <v ; f(x )> <v ; f(x)> for all u 2 U; x 2 IR n : From the inequality in the left, we have <u; f(x )>? " 2 ku? vk2 <u; f(x )> <v ; f(x )> for all u; v 2 U; and therefore by taking the supremum on u 2 U, we obtain ' " (v; x ) <v ; f(x )> for all v 2 U: On the other hand, from the denition (10), we have the following inequality <v ; f(x)> = <v ; f(x)>? " 2 kv? v k 2 ' " (v ; x) for all x 2 IR n : Now, it is straightforward to deduce that <v ; f(x )> = ' " (v ; x ), and therefore that the point (v ; x ) satises b). 10

12 b) ) c). Let (v ; x ) 2 U IR n be a point verifying b), that is, ' " (v; x ) ' " (v ; x ) for all v 2 U; (16) ' " (v ; x ) ' " (v ; x) for all x 2 IR n : (17) First, we write the rst order optimality condition to v in (16), that is, <r v ' " (v ; x ); v? v > = "<u(v ; x )? v ; v? v > 0 for all v 2 U: We then can conclude that u(v ; x ) = v, ' " (v ; x ) = <v ; f(x )> and, P from the inequality (17), 0 = r x ' " (v ; x ) = m v j rf j(x ). Now, from the inequality (16) and Proposition 2.1, ii), we also see that '(x ) = max v2u ' "(v; x ) = ' " (v ; x ) = <v ; f(x )>: Therefore, we have shown that the point (v ; x ) 2 U IR n satises c). To nish the proof, we show that c) ) a). Let (v ; x ) 2 U IR n be a point verifying the statement c). Let us consider the function g dened by g(x) = <v ; f(x)> for all x 2 IR n. Since the function g is convex dierentiable on IR n, we have the equivalence: rg(x ) = mx v j rf j(x ) = 0, <v ; f(x )> <v ; f(x)> for all x 2 IR n : On the other hand, from (4), it is easy to see that we have the inequality <v ; f(x )> = '(x ) = max u2u <u; f(x )> <v; f(x )> for all v 2 U: Therefore, the point (v ; x ) 2 U IR n satises a). Remark. Let (v ; x ) 2 U IR n be a saddle point of the problem (ii), we have shown while proving b) ) c) that r v ' " (v ; x ) = 0, that is, the point v is a maximum of the concave function ' " (; x ) on IR m. Now, since r v ' " (v ; x ) = 0 implies v 2 U, we can conclude that the set of saddle points of the problem (ii) is not modied if in the denition of this problem we change the set U by IR m. On the other hand, from Lemma 3.1, we see that if the point (v ; x ) 2 UIR n veries the statement c) in above proposition then x is a solution point of the problem (1). 11

13 To end this section, we present a result showing that the sequence generated by the proximal method, developed by Rockafellar in [8], converges to a saddle point of the problem (i) or (ii) in Proposition 3.3 above. Unfortunately, the method has only a theoretical interest because the conditions (18) and (19) below can not be simultaneously veried in practice. Proposition 3.4 Let us suppose that all the functions f j are convex dierentiable on IR n, there exists a function f j0 coercive on IR n some j 0 2 f1; :::; mg, fc k g is a sequence of positive real numbers bounded away from zero, that is, there exists c > 0 such that P c k c for all k 2 IN, f k g is a sequence of positive real numbers such that k < +1. Let us consider the sequence f(v k ; x k )g 2 U IR n recursively k2in generated as (v 0 ; x 0 ) 2 U IR n and (v k+1 ; x k+1 ) 2 U IR n such that? k? r v ' 1 (v k+1 ; x k+1 ) + c?1 k v k+1? k v k k (18) c k? k? r x ' 1 (v k+1 ; x k+1 ) + c?1 k x k+1? k x k k : (19) c k Then, the sequence f(v k ; x k )g converges to a point (v ; x ) 2 UIR n verifying the statement b) in Proposition 3.3 (with " = 1). Proof. To prove the proposition we use the following result due to Rockafellar: Theorem. (See Ref. [8]) Let fz k g be any sequence generated by the proximal point algorithm under the criterion dist? 0; S k (z k+1 ) k c k with X k2in k < +1; (20) where S k (z) = T (z) + c?1 k (z? zk ), T is a maximal monotone operator and fc k g is a sequence bounded away from zero. Suppose fz k g is bounded; this holds under the preceding assumption if and only if there exists at least one solution to 0 2 T (z). Then, fz k g converges weakly to a point z 1 satisfying 0 2 T (z 1 ), and lim k2in kzk+1? z k k = 0. } For applying the above result, we dene the function L on IR m IR n by L(v; x) = ' 1 (v; x)? 0 (v) (v; x) 2 IR m IR n ; where 0 is the indicator function of the set U. 12

14 It is easy to see that the function L is a saddle function in the sense of [12], Section 33, and from the compacity of the set U, we have the following equalities sup v2ir m x2ir inf L(v; x) = sup inf ' n 1(v; x) = inf sup ' 1 (v; x) x2ir n x2ir n v2u = inf x2ir n v2ir sup L(v; x) = inf m We now consider the operator T dened on IR m IR n by v2u (21) '(x): x2ir n T (v; x) = f (u; y) 2 IR m IR n : u x) (v); y )(x) g; denotes the convex subdierential (see Rockafellar [12], Section 23). It is not dicult to show that T is a maximal monotone operator and that we moreover have the following properties dom T = f (v; x) 2 IR m IR n : T (v; x) 6= g dom L = U IR n ; (0; 0) 2 T (v ; x ), (v ; x ) is a saddle point of the problems (21): Finally, we nish the proof by noting that the condition (18) and (19) are sucient to obtain the criterion (20). 4. The Method in Practice In line with theoretical results of the previous section (Theorem 3.1 principally), we present the method designed for solving the problem (1). 4.1 The Algorithm At each iteration we minimize the regularized function ' "k (v k ; ) (" k > 0; v k 2 U xed), where the starting point for this minimization is the point x k obtained for the last regularized function ' "k?1 (v k?1 ; ), that is, we nd x k+1 2 IR n such that ' "k (v k ; x k+1 ) ' "k (v k ; x k ) and kr x ' "k (v k ; x k+1 )k k ; (22) where f k g is a sequence converging to zero. We now update (if necessary) the value of the paramenter v k 2 U, that is, if kv k? u(v k ; x k+1 )k is large where u(v k ; x k+1 ) is dened in (11), we then set v k+1 u(v k ; x k+1 ). Consequently, we can describe precisely the algorithm designed for solving the nite min-max problem (1): Algorithm 0. Let (v 0 ; x 0 ) 2 UIR n be a initial point, let > 0 be a tolerance P threshold, let f" k g be a sequence of positive real numbers such that " k < +1 and let 13 k2in

15 0 < < 1 be a parameter. We dene the rst element of the sequence fg k g by G 0 = kr x ' "0 (v 0 ; x 0 )k, " = " 0 and j = 0. At the k-th iteration we have a point (v k ; x k ) 2 U IR n and two real numbers G k and ". 1. If kr x ' " (v k ; x k )k and ku(v k ; x k )? v k k STOP, because we have found the point (v k ; x k ) satisfying the EXIT. Otherwise go to If kr x ' " (v k ; x k )k >, starting from point x k, we minimize the function ' " (v k ; ) until nding a point x 0 2 IR n verifying kr x ' " (v k ; x 0 )k G k : (23) We update G k+1 kr x ' " (v k ; x 0 )k and x k x 0. Go to If ku(v k ; x k )? v k k >, we update the parameters v k u(v k ; x k ), j j + 1 and " " j. Go to We update x k+1 x k, v k+1 v k and k k + 1. Go to 1. Remark. We have introduced the counting number j because from the proof of Proposition 3.2, the updating of the parameter " must only be made while we update the paramenter v. 4.2 The Implementation First, we have chosen the values " k = [1+(k?1)10?4 ]?2 k 2 IN and = 0:99 for the parameters. Moreover, from the computational experiments, we have seen that these values are not of a vital importance for the algorithm work well. Next, for a given point (v; x) 2 UIR n and " > 0 we must solve the minimization problem in step 2 above. That is, for each regularized function ' " (v; ), we must nd a point satisfying the condition (23). For it, we apply a quasi-newton method, more precisely, we generated a sequence of points fx i g IR n with x 1 = x k using descent directions as follows: x i+1 = x i + i d i, where d i is a descent direction for the function ' " (v; ) at point x i. We take d i =?H i r' " (v; x i ), where H i is an approximation to the inverse of the Hessian of the function r' " (v; ) at point x i (a positive denite matrix). To update the Hessian we use the Broyden-Fletcher- Goldfarb-Shanno (BFGS) updating formula with Powell's strategy to preserve positive deniteness (see Minoux [13], Vol. I, chapter 4). The step-length i is found using Wolfe's method for the one-dimensional search. In this case, the parameters set are 0.1 and 0.7 (see Hiriart-Urruty & Lemarechal [14], chapter II). On the other hand, at every point x 2 IR n, we must evaluate the objetive function ' " (v; x), so we need to compute u(v; x) 2 U verifying the equality (11), that is, u(v; x) 2 U for which the following maximum is attained (' " (v; x) =) max u2u f<u; f(x)>? " 2 ku? vk2 g: 14

16 For solving this problem, we use the same technique as the one described in Ggola and Gomez [1], that is, we apply a Lagrange multiplier technique to the above problem. Thus, we obtain the following formula for the solution: u j (v; x) = 8 < f j (x) + "v j? (v; x) " : 0 otherwise; P where (v; x) 2 IR saties m P u j (v; x) = m if f j (x) + "v j? (v; x) > 0 max fj (x)+"v j?(v;x) ; 0 " = 1. This last equation can be solved very easily because it is piecewise linear in the variable. In practice we calcule (v; x) using the following procedure: (24) i) Generate the vector 2 IR m, ordering in a decreasing order the elements of j = f j (x) + "v j j 2 f1; :::; mg, P ii) nd j 0 as the rst index in the set f1; :::; m?1g such that j0 j j 0 j0+1+", otherwise set j 0 = m. Then, we dene (v; x) = h j0 P i. j? " j 0. Therefore, we can suppose that we have obtained x 0 2 IR n verifying the condition (23), we then proceed (if necessary) to change the regularized function, that is, we take new values of the parameters v and ", we set v = u(v; x 0 ) and " = " j+1 where u(v; x 0 ) 2 U is obtained by formula (24). Finally, in order to justify our updating of the parameter v, we consider of vector v() 2 IR m dened by v() = v + [u(v; x)? v] 0: Since r v ' " (v; x) = "[u(v; x)? v], we see that v() is the direction of maximum increase for the function ' " (; x). Moreover, it is easy to show that if r v ' " (v; x) 6= 0 then the largest step-length max ( 1) such that v() 2 U for all 2 [0; max ] is given by max = min vj : 0 u j < v j : v j? u j We next estimate the increase of the function 2 [0; max ]! ' " (v(); x). Moreover, it is straightforward to see that to set = 1 is the best a priori estimation of the solution of the problem max ' "(v(); x). 2[0; max] 15

17 Proposition 4.1 For all 2 [0; max ], we have the following inequality ' " (v; x)+ (2?) "ku(v; x)?vk 2 ' " (v(); x) ' " (v; x)+"ku(v; x)?vk 2 : (25) 2 Proof. From the denition (10), for all 2 [0; max ], we have the equality ' " (v(); x) =?" 2 2 ku(v; x)? vk2? "<v; u(v; x)? v> + supf<; f(x) + "[u(v; x)? v]>? " 2U 2 k? vk2 g; by setting = u(v; x), we then obtain the inequality in the left of (25). On the other hand, from the concavity of the function ' " (; x), we have the following inequality ' " (v(); x) ' " (v; x) + <r v ' " (v; x); v()? v>; and as r v ' " (v; x) = "[u(v; x)? v] we obtain the inequality in the right of (25). 5. Numerical Results The computational results presented below were all obtained using a CD 9460 computer (Silicon Graphics). The CD 9460 is a multiprocessor computer, in this case, with 4 MIPS R4400 MC 64 bits processors. The R4400 is a RISC-processor of 3-th generation with a SIMM memory of 1 Mb. The CD 9460 has a SIMM memory of 512 Mb and a CPU speed of 150 MHz. Test Problems. dened on IR n as For every test problems, we consider m quadratic functions f j f j (x) = <Q j x; x>? <c j ; x> + d j ; (26) where Q j = (q j ik ) is a symmetric nn matrix, cj = (c j i ) is a vector in IRn and d j is a real number. Test Problem 5.1 It is the well known so-called MAXQUAD problem, that is, the data in (26) are given by the fancy formulae: q j ik = exp( i k ) cos(i k) sin(j) i < k; cj i = exp( i j ) sin(i j); dj = 0 P and the diagonals are q j ii = i n j sin(j)j + jq j ikj. See Ref. [15], pp , for i6=k more details on this test problem. 16

18 We solved the above problem for dimensions n=10 and m=5 starting from two dierents points x 0 = (0; :::; 0) T 2 IR 10 and x 1 = (1; :::; 1) T 2 IR 10 ; for both the starting dual point was v 0 = (1; 0; :::; 0) T 2 IR 5. Table 5.1 below summarizes the computational results with tolerance threshold = 5 10?9. Table 5.1 func jac mult iter time norm x x func=number of function evaluations mult=number of quadratic pbs. solved time=cpu time (in seconds) jac=number of Jacobian evaluations iter=number of algorithm iterations norm=starting gradient norm Test Problem 5.2 In this case, the data in (26), q j ik ; cj i ; dj 2 [?10; 10] P i < k are randomly generated real numbers and the diagonals are q j ii = 1 + jq j ik j. i6=k Thus, Q j is a symmetric diagonal dominating matrix. For every problems we have considered as starting points x 0 = (1; :::; 1) 2 IR n and v 0 = (1; 0; :::; 0) T 2 U. Test Problem We solved 7 problems (10 for each dimension), the smallest of dimension 50 (10 functions to maximize) and the largest of dimension 500 (10 functions to maximize). The average CPU time was 5.7 secs. for n=50 and m=10, to secs. for n=500 and m=10. The minimum starting gradient norm was norm=3638. Table below summarizes the computational results with tolerance threshold = 5 10?4. Test Problem We solved 7 problems (10 for each dimension), the smallest of dimension 50 (10 functions to maximize) and the largest of dimension 100 (70 functions to maximize). The average CPU time was 8.1 secs. for n=50 and m=10, to secs. for n=100 and m=70. The minimum starting gradient norm was norm=3554. Table below summarizes the computational results with tolerance threshold = 5 10?6. Test Problem 5.3 In this case, in formula (26), Q 1 is a symmetric diagonal dominating matrix generated as in Test Problem 2 and for each j 2 f2; :::; mg the data q j ik ; cj i ; dj 2 [?10; 10] i k are randomly generated real numbers. Note that for all j 2 f2; :::; mg the symmetric matrix Q j is not necessarily positive semidenite. For every problems we have considered as starting points x 0 = (1; :::; 1) 2 IR n and v 0 = (1; 0; :::; 0) T 2 U. We solved 7 problems (10 for each dimension), the smallest of dimension 50 (10 functions to maximize) and the largest of dimension 300 (10 functions to maximize). The average CPU time was 9.6 secs. for n=50 and m=10, to secs. for n=300 and m=10. The minimum starting gradient norm was norm=

19 Table 5.3 below summarizes the computational results with tolerance threshold = 5 10?6. Table n m func jac mult iter time min max norm Table n m func jac mult iter time min max norm Table 5.3 n m func jac mult iter time min max norm n=number of variables func=average number of function evaluations mult=average number of quadratic pbs. solved time=average CPU time (in seconds) max=maximun CPU time (in seconds) m=number of functions jac=average number of Jacobian evaluations iter=average number of algorithm iterations min=minimum CPU time (in seconds) norm=average starting gradient norm 6. Conclusions We have presented a method for solving the nite min-max problem based on the idea of regularizing to obtain a sequence of dierentiable subproblems. This method principally diers from the others in the literature because rstly, we globally regularize the nondierentiable nite maximum function (to dierence of 18

20 [6]), and secondly, we do not use the "-feasibility concept (see [2] or [3]). We have shown in a clear and direct fashion that every cluster point of the sequence of approximated solutions of these problems is a stationary point of the min-max problem, and therefore, in the convex case, is a solution point of this problem. The implementation becomes very simple because for solving (approximately) the dierentiable subproblems we have used a quasi-newton method, thus we do not need to solve quadratic programming problems for obtaining the descent directions of these subproblems. Finally, from the numerical results we conclude that the proposed method is probably one of the most robust and ecient methods for nite min-max optimization, even in the nonconvex case. Therefore, we think that the present method is of practical interest. References [1] GIGOLA, C. & G OMEZ, S., A regularization method for solving the nite convex min-max problem, SIAM J. Numer. Anal., Vol. 27, pp , [2] CHARALAMBOUS, C. & CONN, A.R., An ecient method to solve the minmax problem directly, SIAM J. Numer. Anal., Vol. 15, pp , [3] DEMYANOV, V.F. & MALOZEMOV, V.N., On the theory of non-linear min-max problems, Russian Math. Surveys, Vol. 26, pp , [4] HAN, S.P., Variable metric methods for minimizing a class of non-dierentiable functions, Math. Programming, Vol. 20, pp. 1-13, [5] MURRAY, W. & OVERTON, M.L., A projective Lagrange algorithm for nonlinear min-max optimization, SIAM J. Sci. Statist. Comput., Vol. 1, pp , [6] ZANG, I., A smoothing-out technique for min-max optimization, Math. Programming, Vol. 19, pp , [7] MARTINET, B., Regularisation d'inequations variationnelles par approximations succesives, Rev. Francaise Inf. Rech. Oper., Vol. R-3, pp , [8] ROCKAFELLAR, R.T., Monotone operators and proximal point algorithm, SIAM J. Control and Optimization, Vol. 14, pp , [9] ATTOUCH, H. & WETS, R.J.B., Approximation and convergence in Non-linear optimization, Nonlinear Programming, Edited by O. Mangasarian, R.R. Meyer and S.M. Robinson, Academic Press New York, New York, Vol. 4, pp , [10] HIGLE, J.L. & SEN, S., Epigraphical Nesting: A unifying theory for the convergence of algorithms, J. of Optimization Theory and Applications, Vol. 84, pp , [11] CLARKE, F.H., Generalized gradients and applications, Trans. Amer. Math. Soc. Vol. 205, pp , [12] ROCKAFELLAR, R.T., Convex Analysis, Princeton University Press, Princeton, New Jersey, [13] MINOUX, M., Programmation mathematique, theorie et algorithmes, Dunod, Paris, France, [14] HIRIART-URRUTY, J.B. & LEMAR ECHAL, C., Convex Analysis and Minimization Algorithms, Grundlehren der mathematischen Wissenschaften 305, Springer- Verlag, New York, [15] LEMAR ECHAL, C. & MIFFLIN, R., Editors, Nonsmooth optimization: Proceedings of a IIASA Workshop 1977 Laxenburg, IIASA proceedings series, Vol. 3, Pergamon Press Ltd., Oxford, England,

Methods for a Class of Convex. Functions. Stephen M. Robinson WP April 1996

Methods for a Class of Convex. Functions. Stephen M. Robinson WP April 1996 Working Paper Linear Convergence of Epsilon-Subgradient Descent Methods for a Class of Convex Functions Stephen M. Robinson WP-96-041 April 1996 IIASA International Institute for Applied Systems Analysis

More information

A Proximal Method for Identifying Active Manifolds

A Proximal Method for Identifying Active Manifolds A Proximal Method for Identifying Active Manifolds W.L. Hare April 18, 2006 Abstract The minimization of an objective function over a constraint set can often be simplified if the active manifold of the

More information

Some Inexact Hybrid Proximal Augmented Lagrangian Algorithms

Some Inexact Hybrid Proximal Augmented Lagrangian Algorithms Some Inexact Hybrid Proximal Augmented Lagrangian Algorithms Carlos Humes Jr. a, Benar F. Svaiter b, Paulo J. S. Silva a, a Dept. of Computer Science, University of São Paulo, Brazil Email: {humes,rsilva}@ime.usp.br

More information

1 Introduction We consider the problem nd x 2 H such that 0 2 T (x); (1.1) where H is a real Hilbert space, and T () is a maximal monotone operator (o

1 Introduction We consider the problem nd x 2 H such that 0 2 T (x); (1.1) where H is a real Hilbert space, and T () is a maximal monotone operator (o Journal of Convex Analysis Volume 6 (1999), No. 1, pp. xx-xx. cheldermann Verlag A HYBRID PROJECTION{PROXIMAL POINT ALGORITHM M. V. Solodov y and B. F. Svaiter y January 27, 1997 (Revised August 24, 1998)

More information

McMaster University. Advanced Optimization Laboratory. Title: A Proximal Method for Identifying Active Manifolds. Authors: Warren L.

McMaster University. Advanced Optimization Laboratory. Title: A Proximal Method for Identifying Active Manifolds. Authors: Warren L. McMaster University Advanced Optimization Laboratory Title: A Proximal Method for Identifying Active Manifolds Authors: Warren L. Hare AdvOl-Report No. 2006/07 April 2006, Hamilton, Ontario, Canada A Proximal

More information

Some Properties of the Augmented Lagrangian in Cone Constrained Optimization

Some Properties of the Augmented Lagrangian in Cone Constrained Optimization MATHEMATICS OF OPERATIONS RESEARCH Vol. 29, No. 3, August 2004, pp. 479 491 issn 0364-765X eissn 1526-5471 04 2903 0479 informs doi 10.1287/moor.1040.0103 2004 INFORMS Some Properties of the Augmented

More information

Brøndsted-Rockafellar property of subdifferentials of prox-bounded functions. Marc Lassonde Université des Antilles et de la Guyane

Brøndsted-Rockafellar property of subdifferentials of prox-bounded functions. Marc Lassonde Université des Antilles et de la Guyane Conference ADGO 2013 October 16, 2013 Brøndsted-Rockafellar property of subdifferentials of prox-bounded functions Marc Lassonde Université des Antilles et de la Guyane Playa Blanca, Tongoy, Chile SUBDIFFERENTIAL

More information

A NOTE ON A GLOBALLY CONVERGENT NEWTON METHOD FOR SOLVING. Patrice MARCOTTE. Jean-Pierre DUSSAULT

A NOTE ON A GLOBALLY CONVERGENT NEWTON METHOD FOR SOLVING. Patrice MARCOTTE. Jean-Pierre DUSSAULT A NOTE ON A GLOBALLY CONVERGENT NEWTON METHOD FOR SOLVING MONOTONE VARIATIONAL INEQUALITIES Patrice MARCOTTE Jean-Pierre DUSSAULT Resume. Il est bien connu que la methode de Newton, lorsqu'appliquee a

More information

1. Introduction The nonlinear complementarity problem (NCP) is to nd a point x 2 IR n such that hx; F (x)i = ; x 2 IR n + ; F (x) 2 IRn + ; where F is

1. Introduction The nonlinear complementarity problem (NCP) is to nd a point x 2 IR n such that hx; F (x)i = ; x 2 IR n + ; F (x) 2 IRn + ; where F is New NCP-Functions and Their Properties 3 by Christian Kanzow y, Nobuo Yamashita z and Masao Fukushima z y University of Hamburg, Institute of Applied Mathematics, Bundesstrasse 55, D-2146 Hamburg, Germany,

More information

ON GAP FUNCTIONS OF VARIATIONAL INEQUALITY IN A BANACH SPACE. Sangho Kum and Gue Myung Lee. 1. Introduction

ON GAP FUNCTIONS OF VARIATIONAL INEQUALITY IN A BANACH SPACE. Sangho Kum and Gue Myung Lee. 1. Introduction J. Korean Math. Soc. 38 (2001), No. 3, pp. 683 695 ON GAP FUNCTIONS OF VARIATIONAL INEQUALITY IN A BANACH SPACE Sangho Kum and Gue Myung Lee Abstract. In this paper we are concerned with theoretical properties

More information

2 JOSE HERSKOVITS The rst stage to get an optimal design is to dene the Optimization Model. That is, to select appropriate design variables, an object

2 JOSE HERSKOVITS The rst stage to get an optimal design is to dene the Optimization Model. That is, to select appropriate design variables, an object A VIEW ON NONLINEAR OPTIMIZATION JOSE HERSKOVITS Mechanical Engineering Program COPPE / Federal University of Rio de Janeiro 1 Caixa Postal 68503, 21945-970 Rio de Janeiro, BRAZIL 1. Introduction Once

More information

Subdifferential representation of convex functions: refinements and applications

Subdifferential representation of convex functions: refinements and applications Subdifferential representation of convex functions: refinements and applications Joël Benoist & Aris Daniilidis Abstract Every lower semicontinuous convex function can be represented through its subdifferential

More information

On the Convergence and O(1/N) Complexity of a Class of Nonlinear Proximal Point Algorithms for Monotonic Variational Inequalities

On the Convergence and O(1/N) Complexity of a Class of Nonlinear Proximal Point Algorithms for Monotonic Variational Inequalities STATISTICS,OPTIMIZATION AND INFORMATION COMPUTING Stat., Optim. Inf. Comput., Vol. 2, June 204, pp 05 3. Published online in International Academic Press (www.iapress.org) On the Convergence and O(/N)

More information

1 Introduction CONVEXIFYING THE SET OF MATRICES OF BOUNDED RANK. APPLICATIONS TO THE QUASICONVEXIFICATION AND CONVEXIFICATION OF THE RANK FUNCTION

1 Introduction CONVEXIFYING THE SET OF MATRICES OF BOUNDED RANK. APPLICATIONS TO THE QUASICONVEXIFICATION AND CONVEXIFICATION OF THE RANK FUNCTION CONVEXIFYING THE SET OF MATRICES OF BOUNDED RANK. APPLICATIONS TO THE QUASICONVEXIFICATION AND CONVEXIFICATION OF THE RANK FUNCTION Jean-Baptiste Hiriart-Urruty and Hai Yen Le Institut de mathématiques

More information

Algorithms for Nonsmooth Optimization

Algorithms for Nonsmooth Optimization Algorithms for Nonsmooth Optimization Frank E. Curtis, Lehigh University presented at Center for Optimization and Statistical Learning, Northwestern University 2 March 2018 Algorithms for Nonsmooth Optimization

More information

system of equations. In particular, we give a complete characterization of the Q-superlinear

system of equations. In particular, we give a complete characterization of the Q-superlinear INEXACT NEWTON METHODS FOR SEMISMOOTH EQUATIONS WITH APPLICATIONS TO VARIATIONAL INEQUALITY PROBLEMS Francisco Facchinei 1, Andreas Fischer 2 and Christian Kanzow 3 1 Dipartimento di Informatica e Sistemistica

More information

NONSMOOTH VARIANTS OF POWELL S BFGS CONVERGENCE THEOREM

NONSMOOTH VARIANTS OF POWELL S BFGS CONVERGENCE THEOREM NONSMOOTH VARIANTS OF POWELL S BFGS CONVERGENCE THEOREM JIAYI GUO AND A.S. LEWIS Abstract. The popular BFGS quasi-newton minimization algorithm under reasonable conditions converges globally on smooth

More information

ON A CLASS OF NONSMOOTH COMPOSITE FUNCTIONS

ON A CLASS OF NONSMOOTH COMPOSITE FUNCTIONS MATHEMATICS OF OPERATIONS RESEARCH Vol. 28, No. 4, November 2003, pp. 677 692 Printed in U.S.A. ON A CLASS OF NONSMOOTH COMPOSITE FUNCTIONS ALEXANDER SHAPIRO We discuss in this paper a class of nonsmooth

More information

A quasisecant method for minimizing nonsmooth functions

A quasisecant method for minimizing nonsmooth functions A quasisecant method for minimizing nonsmooth functions Adil M. Bagirov and Asef Nazari Ganjehlou Centre for Informatics and Applied Optimization, School of Information Technology and Mathematical Sciences,

More information

Epiconvergence and ε-subgradients of Convex Functions

Epiconvergence and ε-subgradients of Convex Functions Journal of Convex Analysis Volume 1 (1994), No.1, 87 100 Epiconvergence and ε-subgradients of Convex Functions Andrei Verona Department of Mathematics, California State University Los Angeles, CA 90032,

More information

IE 5531: Engineering Optimization I

IE 5531: Engineering Optimization I IE 5531: Engineering Optimization I Lecture 15: Nonlinear optimization Prof. John Gunnar Carlsson November 1, 2010 Prof. John Gunnar Carlsson IE 5531: Engineering Optimization I November 1, 2010 1 / 24

More information

A new ane scaling interior point algorithm for nonlinear optimization subject to linear equality and inequality constraints

A new ane scaling interior point algorithm for nonlinear optimization subject to linear equality and inequality constraints Journal of Computational and Applied Mathematics 161 (003) 1 5 www.elsevier.com/locate/cam A new ane scaling interior point algorithm for nonlinear optimization subject to linear equality and inequality

More information

The Pacic Institute for the Mathematical Sciences http://www.pims.math.ca pims@pims.math.ca Surprise Maximization D. Borwein Department of Mathematics University of Western Ontario London, Ontario, Canada

More information

On the Convergence of Newton Iterations to Non-Stationary Points Richard H. Byrd Marcelo Marazzi y Jorge Nocedal z April 23, 2001 Report OTC 2001/01 Optimization Technology Center Northwestern University,

More information

A characterization of essentially strictly convex functions on reflexive Banach spaces

A characterization of essentially strictly convex functions on reflexive Banach spaces A characterization of essentially strictly convex functions on reflexive Banach spaces Michel Volle Département de Mathématiques Université d Avignon et des Pays de Vaucluse 74, rue Louis Pasteur 84029

More information

Step lengths in BFGS method for monotone gradients

Step lengths in BFGS method for monotone gradients Noname manuscript No. (will be inserted by the editor) Step lengths in BFGS method for monotone gradients Yunda Dong Received: date / Accepted: date Abstract In this paper, we consider how to directly

More information

Nonlinear Programming

Nonlinear Programming Nonlinear Programming Kees Roos e-mail: C.Roos@ewi.tudelft.nl URL: http://www.isa.ewi.tudelft.nl/ roos LNMB Course De Uithof, Utrecht February 6 - May 8, A.D. 2006 Optimization Group 1 Outline for week

More information

Numerical Comparisons of. Path-Following Strategies for a. Basic Interior-Point Method for. Revised August Rice University

Numerical Comparisons of. Path-Following Strategies for a. Basic Interior-Point Method for. Revised August Rice University Numerical Comparisons of Path-Following Strategies for a Basic Interior-Point Method for Nonlinear Programming M. A rg a e z, R.A. T a p ia, a n d L. V e l a z q u e z CRPC-TR97777-S Revised August 1998

More information

Global Optimality Conditions in Maximizing a Convex Quadratic Function under Convex Quadratic Constraints

Global Optimality Conditions in Maximizing a Convex Quadratic Function under Convex Quadratic Constraints Journal of Global Optimization 21: 445 455, 2001. 2001 Kluwer Academic Publishers. Printed in the Netherlands. 445 Global Optimality Conditions in Maximizing a Convex Quadratic Function under Convex Quadratic

More information

Local strong convexity and local Lipschitz continuity of the gradient of convex functions

Local strong convexity and local Lipschitz continuity of the gradient of convex functions Local strong convexity and local Lipschitz continuity of the gradient of convex functions R. Goebel and R.T. Rockafellar May 23, 2007 Abstract. Given a pair of convex conjugate functions f and f, we investigate

More information

QUADRATICALLY AND SUPERLINEARLY CONVERGENT ALGORITHMS FOR THE SOLUTION OF INEQUALITY CONSTRAINED MINIMIZATION PROBLEMS 1

QUADRATICALLY AND SUPERLINEARLY CONVERGENT ALGORITHMS FOR THE SOLUTION OF INEQUALITY CONSTRAINED MINIMIZATION PROBLEMS 1 QUADRATICALLY AND SUPERLINEARLY CONVERGENT ALGORITHMS FOR THE SOLUTION OF INEQUALITY CONSTRAINED MINIMIZATION PROBLEMS 1 F. FACCHINEI 2 AND S. LUCIDI 3 Communicated by L.C.W. Dixon 1 This research was

More information

Projected Gradient Methods for NCP 57. Complementarity Problems via Normal Maps

Projected Gradient Methods for NCP 57. Complementarity Problems via Normal Maps Projected Gradient Methods for NCP 57 Recent Advances in Nonsmooth Optimization, pp. 57-86 Eds..-Z. u, L. Qi and R.S. Womersley c1995 World Scientic Publishers Projected Gradient Methods for Nonlinear

More information

quantitative information on the error caused by using the solution of the linear problem to describe the response of the elastic material on a corner

quantitative information on the error caused by using the solution of the linear problem to describe the response of the elastic material on a corner Quantitative Justication of Linearization in Nonlinear Hencky Material Problems 1 Weimin Han and Hong-ci Huang 3 Abstract. The classical linear elasticity theory is based on the assumption that the size

More information

A convergence result for an Outer Approximation Scheme

A convergence result for an Outer Approximation Scheme A convergence result for an Outer Approximation Scheme R. S. Burachik Engenharia de Sistemas e Computação, COPPE-UFRJ, CP 68511, Rio de Janeiro, RJ, CEP 21941-972, Brazil regi@cos.ufrj.br J. O. Lopes Departamento

More information

A proximal approach to the inversion of ill-conditioned matrices

A proximal approach to the inversion of ill-conditioned matrices A proximal approach to the inversion of ill-conditioned matrices Pierre Maréchal, Aude Rondepierre To cite this version: Pierre Maréchal, Aude Rondepierre. A proximal approach to the inversion of ill-conditioned

More information

ALGORITHMS FOR MINIMIZING DIFFERENCES OF CONVEX FUNCTIONS AND APPLICATIONS

ALGORITHMS FOR MINIMIZING DIFFERENCES OF CONVEX FUNCTIONS AND APPLICATIONS ALGORITHMS FOR MINIMIZING DIFFERENCES OF CONVEX FUNCTIONS AND APPLICATIONS Mau Nam Nguyen (joint work with D. Giles and R. B. Rector) Fariborz Maseeh Department of Mathematics and Statistics Portland State

More information

On the Midpoint Method for Solving Generalized Equations

On the Midpoint Method for Solving Generalized Equations Punjab University Journal of Mathematics (ISSN 1016-56) Vol. 40 (008) pp. 63-70 On the Midpoint Method for Solving Generalized Equations Ioannis K. Argyros Cameron University Department of Mathematics

More information

Downloaded 09/27/13 to Redistribution subject to SIAM license or copyright; see

Downloaded 09/27/13 to Redistribution subject to SIAM license or copyright; see SIAM J. OPTIM. Vol. 23, No., pp. 256 267 c 203 Society for Industrial and Applied Mathematics TILT STABILITY, UNIFORM QUADRATIC GROWTH, AND STRONG METRIC REGULARITY OF THE SUBDIFFERENTIAL D. DRUSVYATSKIY

More information

SOLUTION OF NONLINEAR COMPLEMENTARITY PROBLEMS

SOLUTION OF NONLINEAR COMPLEMENTARITY PROBLEMS A SEMISMOOTH EQUATION APPROACH TO THE SOLUTION OF NONLINEAR COMPLEMENTARITY PROBLEMS Tecla De Luca 1, Francisco Facchinei 1 and Christian Kanzow 2 1 Universita di Roma \La Sapienza" Dipartimento di Informatica

More information

and P RP k = gt k (g k? g k? ) kg k? k ; (.5) where kk is the Euclidean norm. This paper deals with another conjugate gradient method, the method of s

and P RP k = gt k (g k? g k? ) kg k? k ; (.5) where kk is the Euclidean norm. This paper deals with another conjugate gradient method, the method of s Global Convergence of the Method of Shortest Residuals Yu-hong Dai and Ya-xiang Yuan State Key Laboratory of Scientic and Engineering Computing, Institute of Computational Mathematics and Scientic/Engineering

More information

On Second-order Properties of the Moreau-Yosida Regularization for Constrained Nonsmooth Convex Programs

On Second-order Properties of the Moreau-Yosida Regularization for Constrained Nonsmooth Convex Programs On Second-order Properties of the Moreau-Yosida Regularization for Constrained Nonsmooth Convex Programs Fanwen Meng,2 ; Gongyun Zhao Department of Mathematics National University of Singapore 2 Science

More information

Unconstrained optimization

Unconstrained optimization Chapter 4 Unconstrained optimization An unconstrained optimization problem takes the form min x Rnf(x) (4.1) for a target functional (also called objective function) f : R n R. In this chapter and throughout

More information

The common-line problem in congested transit networks

The common-line problem in congested transit networks The common-line problem in congested transit networks R. Cominetti, J. Correa Abstract We analyze a general (Wardrop) equilibrium model for the common-line problem in transit networks under congestion

More information

INDEFINITE TRUST REGION SUBPROBLEMS AND NONSYMMETRIC EIGENVALUE PERTURBATIONS. Ronald J. Stern. Concordia University

INDEFINITE TRUST REGION SUBPROBLEMS AND NONSYMMETRIC EIGENVALUE PERTURBATIONS. Ronald J. Stern. Concordia University INDEFINITE TRUST REGION SUBPROBLEMS AND NONSYMMETRIC EIGENVALUE PERTURBATIONS Ronald J. Stern Concordia University Department of Mathematics and Statistics Montreal, Quebec H4B 1R6, Canada and Henry Wolkowicz

More information

1. Introduction Let the least value of an objective function F (x), x2r n, be required, where F (x) can be calculated for any vector of variables x2r

1. Introduction Let the least value of an objective function F (x), x2r n, be required, where F (x) can be calculated for any vector of variables x2r DAMTP 2002/NA08 Least Frobenius norm updating of quadratic models that satisfy interpolation conditions 1 M.J.D. Powell Abstract: Quadratic models of objective functions are highly useful in many optimization

More information

Optimization Methods. Lecture 18: Optimality Conditions and. Gradient Methods. for Unconstrained Optimization

Optimization Methods. Lecture 18: Optimality Conditions and. Gradient Methods. for Unconstrained Optimization 5.93 Optimization Methods Lecture 8: Optimality Conditions and Gradient Methods for Unconstrained Optimization Outline. Necessary and sucient optimality conditions Slide. Gradient m e t h o d s 3. The

More information

On convergence rate of the Douglas-Rachford operator splitting method

On convergence rate of the Douglas-Rachford operator splitting method On convergence rate of the Douglas-Rachford operator splitting method Bingsheng He and Xiaoming Yuan 2 Abstract. This note provides a simple proof on a O(/k) convergence rate for the Douglas- Rachford

More information

Improving the Convergence of Back-Propogation Learning with Second Order Methods

Improving the Convergence of Back-Propogation Learning with Second Order Methods the of Back-Propogation Learning with Second Order Methods Sue Becker and Yann le Cun, Sept 1988 Kasey Bray, October 2017 Table of Contents 1 with Back-Propagation 2 the of BP 3 A Computationally Feasible

More information

A Generalized Homogeneous and Self-Dual Algorithm. for Linear Programming. February 1994 (revised December 1994)

A Generalized Homogeneous and Self-Dual Algorithm. for Linear Programming. February 1994 (revised December 1994) A Generalized Homogeneous and Self-Dual Algorithm for Linear Programming Xiaojie Xu Yinyu Ye y February 994 (revised December 994) Abstract: A generalized homogeneous and self-dual (HSD) infeasible-interior-point

More information

A Unified Approach to Proximal Algorithms using Bregman Distance

A Unified Approach to Proximal Algorithms using Bregman Distance A Unified Approach to Proximal Algorithms using Bregman Distance Yi Zhou a,, Yingbin Liang a, Lixin Shen b a Department of Electrical Engineering and Computer Science, Syracuse University b Department

More information

Nonlinear Optimization: What s important?

Nonlinear Optimization: What s important? Nonlinear Optimization: What s important? Julian Hall 10th May 2012 Convexity: convex problems A local minimizer is a global minimizer A solution of f (x) = 0 (stationary point) is a minimizer A global

More information

Fast proximal gradient methods

Fast proximal gradient methods L. Vandenberghe EE236C (Spring 2013-14) Fast proximal gradient methods fast proximal gradient method (FISTA) FISTA with line search FISTA as descent method Nesterov s second method 1 Fast (proximal) gradient

More information

AN INEXACT HYBRIDGENERALIZEDPROXIMAL POINT ALGORITHM ANDSOME NEW RESULTS ON THE THEORY OF BREGMAN FUNCTIONS

AN INEXACT HYBRIDGENERALIZEDPROXIMAL POINT ALGORITHM ANDSOME NEW RESULTS ON THE THEORY OF BREGMAN FUNCTIONS MATHEMATICS OF OPERATIONS RESEARCH Vol. 25, No. 2, May 2000, pp. 214 230 Printed in U.S.A. AN INEXACT HYBRIDGENERALIZEDPROXIMAL POINT ALGORITHM ANDSOME NEW RESULTS ON THE THEORY OF BREGMAN FUNCTIONS M.

More information

On the iterate convergence of descent methods for convex optimization

On the iterate convergence of descent methods for convex optimization On the iterate convergence of descent methods for convex optimization Clovis C. Gonzaga March 1, 2014 Abstract We study the iterate convergence of strong descent algorithms applied to convex functions.

More information

On the Local Quadratic Convergence of the Primal-Dual Augmented Lagrangian Method

On the Local Quadratic Convergence of the Primal-Dual Augmented Lagrangian Method Optimization Methods and Software Vol. 00, No. 00, Month 200x, 1 11 On the Local Quadratic Convergence of the Primal-Dual Augmented Lagrangian Method ROMAN A. POLYAK Department of SEOR and Mathematical

More information

MS&E 318 (CME 338) Large-Scale Numerical Optimization

MS&E 318 (CME 338) Large-Scale Numerical Optimization Stanford University, Management Science & Engineering (and ICME) MS&E 318 (CME 338) Large-Scale Numerical Optimization 1 Origins Instructor: Michael Saunders Spring 2015 Notes 9: Augmented Lagrangian Methods

More information

Relationships between upper exhausters and the basic subdifferential in variational analysis

Relationships between upper exhausters and the basic subdifferential in variational analysis J. Math. Anal. Appl. 334 (2007) 261 272 www.elsevier.com/locate/jmaa Relationships between upper exhausters and the basic subdifferential in variational analysis Vera Roshchina City University of Hong

More information

MAXIMALITY OF SUMS OF TWO MAXIMAL MONOTONE OPERATORS

MAXIMALITY OF SUMS OF TWO MAXIMAL MONOTONE OPERATORS MAXIMALITY OF SUMS OF TWO MAXIMAL MONOTONE OPERATORS JONATHAN M. BORWEIN, FRSC Abstract. We use methods from convex analysis convex, relying on an ingenious function of Simon Fitzpatrick, to prove maximality

More information

Iterative Convex Optimization Algorithms; Part One: Using the Baillon Haddad Theorem

Iterative Convex Optimization Algorithms; Part One: Using the Baillon Haddad Theorem Iterative Convex Optimization Algorithms; Part One: Using the Baillon Haddad Theorem Charles Byrne (Charles Byrne@uml.edu) http://faculty.uml.edu/cbyrne/cbyrne.html Department of Mathematical Sciences

More information

Spurious Chaotic Solutions of Dierential. Equations. Sigitas Keras. September Department of Applied Mathematics and Theoretical Physics

Spurious Chaotic Solutions of Dierential. Equations. Sigitas Keras. September Department of Applied Mathematics and Theoretical Physics UNIVERSITY OF CAMBRIDGE Numerical Analysis Reports Spurious Chaotic Solutions of Dierential Equations Sigitas Keras DAMTP 994/NA6 September 994 Department of Applied Mathematics and Theoretical Physics

More information

INEXACT CUTS IN BENDERS' DECOMPOSITION GOLBON ZAKERI, ANDREW B. PHILPOTT AND DAVID M. RYAN y Abstract. Benders' decomposition is a well-known techniqu

INEXACT CUTS IN BENDERS' DECOMPOSITION GOLBON ZAKERI, ANDREW B. PHILPOTT AND DAVID M. RYAN y Abstract. Benders' decomposition is a well-known techniqu INEXACT CUTS IN BENDERS' DECOMPOSITION GOLBON ZAKERI, ANDREW B. PHILPOTT AND DAVID M. RYAN y Abstract. Benders' decomposition is a well-known technique for solving large linear programs with a special

More information

c 1998 Society for Industrial and Applied Mathematics

c 1998 Society for Industrial and Applied Mathematics SIAM J. OPTIM. Vol. 9, No. 1, pp. 179 189 c 1998 Society for Industrial and Applied Mathematics WEAK SHARP SOLUTIONS OF VARIATIONAL INEQUALITIES PATRICE MARCOTTE AND DAOLI ZHU Abstract. In this work we

More information

National Institute of Standards and Technology USA. Jon W. Tolle. Departments of Mathematics and Operations Research. University of North Carolina USA

National Institute of Standards and Technology USA. Jon W. Tolle. Departments of Mathematics and Operations Research. University of North Carolina USA Acta Numerica (1996), pp. 1{000 Sequential Quadratic Programming Paul T. Boggs Applied and Computational Mathematics Division National Institute of Standards and Technology Gaithersburg, Maryland 20899

More information

Extra-Updates Criterion for the Limited Memory BFGS Algorithm for Large Scale Nonlinear Optimization M. Al-Baali y December 7, 2000 Abstract This pape

Extra-Updates Criterion for the Limited Memory BFGS Algorithm for Large Scale Nonlinear Optimization M. Al-Baali y December 7, 2000 Abstract This pape SULTAN QABOOS UNIVERSITY Department of Mathematics and Statistics Extra-Updates Criterion for the Limited Memory BFGS Algorithm for Large Scale Nonlinear Optimization by M. Al-Baali December 2000 Extra-Updates

More information

A derivative-free nonmonotone line search and its application to the spectral residual method

A derivative-free nonmonotone line search and its application to the spectral residual method IMA Journal of Numerical Analysis (2009) 29, 814 825 doi:10.1093/imanum/drn019 Advance Access publication on November 14, 2008 A derivative-free nonmonotone line search and its application to the spectral

More information

Journal of Inequalities in Pure and Applied Mathematics

Journal of Inequalities in Pure and Applied Mathematics Journal of Inequalities in Pure and Applied Mathematics EPI-DIFFERENTIABILITY AND OPTIMALITY CONDITIONS FOR AN EXTREMAL PROBLEM UNDER INCLUSION CONSTRAINTS T. AMAHROQ, N. GADHI AND PR H. RIAHI Université

More information

Necessary and Sufficient Conditions for the Existence of a Global Maximum for Convex Functions in Reflexive Banach Spaces

Necessary and Sufficient Conditions for the Existence of a Global Maximum for Convex Functions in Reflexive Banach Spaces Laboratoire d Arithmétique, Calcul formel et d Optimisation UMR CNRS 6090 Necessary and Sufficient Conditions for the Existence of a Global Maximum for Convex Functions in Reflexive Banach Spaces Emil

More information

Optimization: Interior-Point Methods and. January,1995 USA. and Cooperative Research Centre for Robust and Adaptive Systems.

Optimization: Interior-Point Methods and. January,1995 USA. and Cooperative Research Centre for Robust and Adaptive Systems. Innite Dimensional Quadratic Optimization: Interior-Point Methods and Control Applications January,995 Leonid Faybusovich John B. Moore y Department of Mathematics University of Notre Dame Mail Distribution

More information

On the interior of the simplex, we have the Hessian of d(x), Hd(x) is diagonal with ith. µd(w) + w T c. minimize. subject to w T 1 = 1,

On the interior of the simplex, we have the Hessian of d(x), Hd(x) is diagonal with ith. µd(w) + w T c. minimize. subject to w T 1 = 1, Math 30 Winter 05 Solution to Homework 3. Recognizing the convexity of g(x) := x log x, from Jensen s inequality we get d(x) n x + + x n n log x + + x n n where the equality is attained only at x = (/n,...,

More information

Optimization for Machine Learning

Optimization for Machine Learning Optimization for Machine Learning (Problems; Algorithms - A) SUVRIT SRA Massachusetts Institute of Technology PKU Summer School on Data Science (July 2017) Course materials http://suvrit.de/teaching.html

More information

Finite Convergence for Feasible Solution Sequence of Variational Inequality Problems

Finite Convergence for Feasible Solution Sequence of Variational Inequality Problems Mathematical and Computational Applications Article Finite Convergence for Feasible Solution Sequence of Variational Inequality Problems Wenling Zhao *, Ruyu Wang and Hongxiang Zhang School of Science,

More information

GLOBAL CONVERGENCE OF CONJUGATE GRADIENT METHODS WITHOUT LINE SEARCH

GLOBAL CONVERGENCE OF CONJUGATE GRADIENT METHODS WITHOUT LINE SEARCH GLOBAL CONVERGENCE OF CONJUGATE GRADIENT METHODS WITHOUT LINE SEARCH Jie Sun 1 Department of Decision Sciences National University of Singapore, Republic of Singapore Jiapu Zhang 2 Department of Mathematics

More information

Spectral gradient projection method for solving nonlinear monotone equations

Spectral gradient projection method for solving nonlinear monotone equations Journal of Computational and Applied Mathematics 196 (2006) 478 484 www.elsevier.com/locate/cam Spectral gradient projection method for solving nonlinear monotone equations Li Zhang, Weijun Zhou Department

More information

Global Maximum of a Convex Function: Necessary and Sufficient Conditions

Global Maximum of a Convex Function: Necessary and Sufficient Conditions Journal of Convex Analysis Volume 13 2006), No. 3+4, 687 694 Global Maximum of a Convex Function: Necessary and Sufficient Conditions Emil Ernst Laboratoire de Modélisation en Mécaniue et Thermodynamiue,

More information

Continuous Sets and Non-Attaining Functionals in Reflexive Banach Spaces

Continuous Sets and Non-Attaining Functionals in Reflexive Banach Spaces Laboratoire d Arithmétique, Calcul formel et d Optimisation UMR CNRS 6090 Continuous Sets and Non-Attaining Functionals in Reflexive Banach Spaces Emil Ernst Michel Théra Rapport de recherche n 2004-04

More information

Pointwise convergence rate for nonlinear conservation. Eitan Tadmor and Tao Tang

Pointwise convergence rate for nonlinear conservation. Eitan Tadmor and Tao Tang Pointwise convergence rate for nonlinear conservation laws Eitan Tadmor and Tao Tang Abstract. We introduce a new method to obtain pointwise error estimates for vanishing viscosity and nite dierence approximations

More information

2. Quasi-Newton methods

2. Quasi-Newton methods L. Vandenberghe EE236C (Spring 2016) 2. Quasi-Newton methods variable metric methods quasi-newton methods BFGS update limited-memory quasi-newton methods 2-1 Newton method for unconstrained minimization

More information

Nondegenerate Solutions and Related Concepts in. Ane Variational Inequalities. M.C. Ferris. Madison, Wisconsin

Nondegenerate Solutions and Related Concepts in. Ane Variational Inequalities. M.C. Ferris. Madison, Wisconsin Nondegenerate Solutions and Related Concepts in Ane Variational Inequalities M.C. Ferris Computer Sciences Department University of Wisconsin Madison, Wisconsin 53706 Email: ferris@cs.wisc.edu J.S. Pang

More information

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 3. Gradient Method

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 3. Gradient Method Shiqian Ma, MAT-258A: Numerical Optimization 1 Chapter 3 Gradient Method Shiqian Ma, MAT-258A: Numerical Optimization 2 3.1. Gradient method Classical gradient method: to minimize a differentiable convex

More information

Convex Quadratic Approximation

Convex Quadratic Approximation Convex Quadratic Approximation J. Ben Rosen 1 and Roummel F. Marcia 2 Abstract. For some applications it is desired to approximate a set of m data points in IR n with a convex quadratic function. Furthermore,

More information

STABLE AND TOTAL FENCHEL DUALITY FOR CONVEX OPTIMIZATION PROBLEMS IN LOCALLY CONVEX SPACES

STABLE AND TOTAL FENCHEL DUALITY FOR CONVEX OPTIMIZATION PROBLEMS IN LOCALLY CONVEX SPACES STABLE AND TOTAL FENCHEL DUALITY FOR CONVEX OPTIMIZATION PROBLEMS IN LOCALLY CONVEX SPACES CHONG LI, DONGHUI FANG, GENARO LÓPEZ, AND MARCO A. LÓPEZ Abstract. We consider the optimization problem (P A )

More information

ON THE DIAMETER OF THE ATTRACTOR OF AN IFS Serge Dubuc Raouf Hamzaoui Abstract We investigate methods for the evaluation of the diameter of the attrac

ON THE DIAMETER OF THE ATTRACTOR OF AN IFS Serge Dubuc Raouf Hamzaoui Abstract We investigate methods for the evaluation of the diameter of the attrac ON THE DIAMETER OF THE ATTRACTOR OF AN IFS Serge Dubuc Raouf Hamzaoui Abstract We investigate methods for the evaluation of the diameter of the attractor of an IFS. We propose an upper bound for the diameter

More information

Additional Homework Problems

Additional Homework Problems Additional Homework Problems Robert M. Freund April, 2004 2004 Massachusetts Institute of Technology. 1 2 1 Exercises 1. Let IR n + denote the nonnegative orthant, namely IR + n = {x IR n x j ( ) 0,j =1,...,n}.

More information

Extended Monotropic Programming and Duality 1

Extended Monotropic Programming and Duality 1 March 2006 (Revised February 2010) Report LIDS - 2692 Extended Monotropic Programming and Duality 1 by Dimitri P. Bertsekas 2 Abstract We consider the problem minimize f i (x i ) subject to x S, where

More information

t x 0.25

t x 0.25 Journal of ELECTRICAL ENGINEERING, VOL. 52, NO. /s, 2, 48{52 COMPARISON OF BROYDEN AND NEWTON METHODS FOR SOLVING NONLINEAR PARABOLIC EQUATIONS Ivan Cimrak If the time discretization of a nonlinear parabolic

More information

Self-dual Smooth Approximations of Convex Functions via the Proximal Average

Self-dual Smooth Approximations of Convex Functions via the Proximal Average Chapter Self-dual Smooth Approximations of Convex Functions via the Proximal Average Heinz H. Bauschke, Sarah M. Moffat, and Xianfu Wang Abstract The proximal average of two convex functions has proven

More information

Identifying Active Constraints via Partial Smoothness and Prox-Regularity

Identifying Active Constraints via Partial Smoothness and Prox-Regularity Journal of Convex Analysis Volume 11 (2004), No. 2, 251 266 Identifying Active Constraints via Partial Smoothness and Prox-Regularity W. L. Hare Department of Mathematics, Simon Fraser University, Burnaby,

More information

Sequential Quadratic Programming Method for Nonlinear Second-Order Cone Programming Problems. Hirokazu KATO

Sequential Quadratic Programming Method for Nonlinear Second-Order Cone Programming Problems. Hirokazu KATO Sequential Quadratic Programming Method for Nonlinear Second-Order Cone Programming Problems Guidance Professor Masao FUKUSHIMA Hirokazu KATO 2004 Graduate Course in Department of Applied Mathematics and

More information

Expanding the reach of optimal methods

Expanding the reach of optimal methods Expanding the reach of optimal methods Dmitriy Drusvyatskiy Mathematics, University of Washington Joint work with C. Kempton (UW), M. Fazel (UW), A.S. Lewis (Cornell), and S. Roy (UW) BURKAPALOOZA! WCOM

More information

230 L. HEI if ρ k is satisfactory enough, and to reduce it by a constant fraction (say, ahalf): k+1 = fi 2 k (0 <fi 2 < 1); (1.7) in the case ρ k is n

230 L. HEI if ρ k is satisfactory enough, and to reduce it by a constant fraction (say, ahalf): k+1 = fi 2 k (0 <fi 2 < 1); (1.7) in the case ρ k is n Journal of Computational Mathematics, Vol.21, No.2, 2003, 229 236. A SELF-ADAPTIVE TRUST REGION ALGORITHM Λ1) Long Hei y (Institute of Computational Mathematics and Scientific/Engineering Computing, Academy

More information

A CHARACTERIZATION OF STRICT LOCAL MINIMIZERS OF ORDER ONE FOR STATIC MINMAX PROBLEMS IN THE PARAMETRIC CONSTRAINT CASE

A CHARACTERIZATION OF STRICT LOCAL MINIMIZERS OF ORDER ONE FOR STATIC MINMAX PROBLEMS IN THE PARAMETRIC CONSTRAINT CASE Journal of Applied Analysis Vol. 6, No. 1 (2000), pp. 139 148 A CHARACTERIZATION OF STRICT LOCAL MINIMIZERS OF ORDER ONE FOR STATIC MINMAX PROBLEMS IN THE PARAMETRIC CONSTRAINT CASE A. W. A. TAHA Received

More information

On the acceleration of augmented Lagrangian method for linearly constrained optimization

On the acceleration of augmented Lagrangian method for linearly constrained optimization On the acceleration of augmented Lagrangian method for linearly constrained optimization Bingsheng He and Xiaoming Yuan October, 2 Abstract. The classical augmented Lagrangian method (ALM plays a fundamental

More information

A double projection method for solving variational inequalities without monotonicity

A double projection method for solving variational inequalities without monotonicity A double projection method for solving variational inequalities without monotonicity Minglu Ye Yiran He Accepted by Computational Optimization and Applications, DOI: 10.1007/s10589-014-9659-7,Apr 05, 2014

More information

Higher-Order Methods

Higher-Order Methods Higher-Order Methods Stephen J. Wright 1 2 Computer Sciences Department, University of Wisconsin-Madison. PCMI, July 2016 Stephen Wright (UW-Madison) Higher-Order Methods PCMI, July 2016 1 / 25 Smooth

More information

DUALIZATION OF SUBGRADIENT CONDITIONS FOR OPTIMALITY

DUALIZATION OF SUBGRADIENT CONDITIONS FOR OPTIMALITY DUALIZATION OF SUBGRADIENT CONDITIONS FOR OPTIMALITY R. T. Rockafellar* Abstract. A basic relationship is derived between generalized subgradients of a given function, possibly nonsmooth and nonconvex,

More information

Unconstrained minimization: assumptions

Unconstrained minimization: assumptions Unconstrained minimization I terminology and assumptions I gradient descent method I steepest descent method I Newton s method I self-concordant functions I implementation IOE 611: Nonlinear Programming,

More information

Converse Lyapunov Functions for Inclusions 2 Basic denitions Given a set A, A stands for the closure of A, A stands for the interior set of A, coa sta

Converse Lyapunov Functions for Inclusions 2 Basic denitions Given a set A, A stands for the closure of A, A stands for the interior set of A, coa sta A smooth Lyapunov function from a class-kl estimate involving two positive semidenite functions Andrew R. Teel y ECE Dept. University of California Santa Barbara, CA 93106 teel@ece.ucsb.edu Laurent Praly

More information

Nonmonotonic back-tracking trust region interior point algorithm for linear constrained optimization

Nonmonotonic back-tracking trust region interior point algorithm for linear constrained optimization Journal of Computational and Applied Mathematics 155 (2003) 285 305 www.elsevier.com/locate/cam Nonmonotonic bac-tracing trust region interior point algorithm for linear constrained optimization Detong

More information

Enhanced Fritz John Optimality Conditions and Sensitivity Analysis

Enhanced Fritz John Optimality Conditions and Sensitivity Analysis Enhanced Fritz John Optimality Conditions and Sensitivity Analysis Dimitri P. Bertsekas Laboratory for Information and Decision Systems Massachusetts Institute of Technology March 2016 1 / 27 Constrained

More information

FLUX IDENTIFICATION IN CONSERVATION LAWS 2 It is quite natural to formulate this problem more or less like an optimal control problem: for any functio

FLUX IDENTIFICATION IN CONSERVATION LAWS 2 It is quite natural to formulate this problem more or less like an optimal control problem: for any functio CONVERGENCE RESULTS FOR THE FLUX IDENTIFICATION IN A SCALAR CONSERVATION LAW FRANCOIS JAMES y AND MAURICIO SEP ULVEDA z Abstract. Here we study an inverse problem for a quasilinear hyperbolic equation.

More information