Remarks On Second Order Generalized Derivatives For Differentiable Functions With Lipschitzian Jacobian

Similar documents
A survey on C 1,1 functions: theory, numerical methods and applications

ON GENERALIZED DERIVATIVES FOR C 1,1 VECTOR OPTIMIZATION PROBLEMS

On second order conditions in vector optimization

Applications of Mathematics

Second-order mollified derivatives and optimization

SECOND-ORDER CHARACTERIZATIONS OF CONVEX AND PSEUDOCONVEX FUNCTIONS

generalized Jacobians, nonsmooth analysis, mean value conditions, optimality

ON GAP FUNCTIONS OF VARIATIONAL INEQUALITY IN A BANACH SPACE. Sangho Kum and Gue Myung Lee. 1. Introduction

Chapter 1. Optimality Conditions: Unconstrained Optimization. 1.1 Differentiable Problems

BASICS OF CONVEX ANALYSIS

Journal of Inequalities in Pure and Applied Mathematics

On the iterate convergence of descent methods for convex optimization

Nonlinear equations. Norms for R n. Convergence orders for iterative methods

ON GENERALIZED-CONVEX CONSTRAINED MULTI-OBJECTIVE OPTIMIZATION

FIRST ORDER CHARACTERIZATIONS OF PSEUDOCONVEX FUNCTIONS. Vsevolod Ivanov Ivanov

A Finite Newton Method for Classification Problems

Methods for a Class of Convex. Functions. Stephen M. Robinson WP April 1996

ON A CLASS OF NONSMOOTH COMPOSITE FUNCTIONS

Introduction. New Nonsmooth Trust Region Method for Unconstraint Locally Lipschitz Optimization Problems

A quasisecant method for minimizing nonsmooth functions

A Unified Analysis of Nonconvex Optimization Duality and Penalty Methods with General Augmenting Functions

Some Properties of the Augmented Lagrangian in Cone Constrained Optimization

Optimization and Optimal Control in Banach Spaces

Stability of optimization problems with stochastic dominance constraints

Generalized Hessians of C 1,1 - functions and second-order viscosity subjets

Structural and Multidisciplinary Optimization. P. Duysinx and P. Tossings

Examination paper for TMA4180 Optimization I

Review of Multi-Calculus (Study Guide for Spivak s CHAPTER ONE TO THREE)

On constraint qualifications with generalized convexity and optimality conditions

An Infeasible Interior Proximal Method for Convex Programming Problems with Linear Constraints 1

Newton-type Methods for Solving the Nonsmooth Equations with Finitely Many Maximum Functions

A Geometric Framework for Nonconvex Optimization Duality using Augmented Lagrangian Functions

Bootcamp. Christoph Thiele. Summer As in the case of separability we have the following two observations: Lemma 1 Finite sets are compact.

Keywords. 1. Introduction.

Journal of Inequalities in Pure and Applied Mathematics

Functional Analysis. Franck Sueur Metric spaces Definitions Completeness Compactness Separability...

QUADRATIC MAJORIZATION 1. INTRODUCTION

1. Bounded linear maps. A linear map T : E F of real Banach

20 J.-S. CHEN, C.-H. KO AND X.-R. WU. : R 2 R is given by. Recently, the generalized Fischer-Burmeister function ϕ p : R2 R, which includes

Preprint Stephan Dempe and Patrick Mehlitz Lipschitz continuity of the optimal value function in parametric optimization ISSN

1 Directional Derivatives and Differentiability

Optimality Conditions for Nonsmooth Convex Optimization

Key words. nonlinear programming, pattern search algorithm, derivative-free optimization, convergence analysis, second order optimality conditions

A Note on Nonconvex Minimax Theorem with Separable Homogeneous Polynomials

A Common Fixed Points in Cone and rectangular cone Metric Spaces

Journal of Inequalities in Pure and Applied Mathematics

arxiv: v1 [math.fa] 30 Oct 2018

CONVEX FUNCTIONS AND PEANO DERIVATIVES

Convex Optimization Notes

Sequential Pareto Subdifferential Sum Rule And Sequential Effi ciency

Constraint qualifications for convex inequality systems with applications in constrained optimization

Relationships between upper exhausters and the basic subdifferential in variational analysis

A globally and R-linearly convergent hybrid HS and PRP method and its inexact version with applications

Constrained Optimization and Lagrangian Duality

On the Local Quadratic Convergence of the Primal-Dual Augmented Lagrangian Method

Convergence rate estimates for the gradient differential inclusion

A CHARACTERIZATION OF STRICT LOCAL MINIMIZERS OF ORDER ONE FOR STATIC MINMAX PROBLEMS IN THE PARAMETRIC CONSTRAINT CASE

Problem Set 6: Solutions Math 201A: Fall a n x n,

Convex Optimization and Modeling

McMaster University. Advanced Optimization Laboratory. Title: A Proximal Method for Identifying Active Manifolds. Authors: Warren L.

Alternative Theorems and Necessary Optimality Conditions for Directionally Differentiable Multiobjective Programs

Robust Duality in Parametric Convex Optimization

Dual Proximal Gradient Method

From now on, we will represent a metric space with (X, d). Here are some examples: i=1 (x i y i ) p ) 1 p, p 1.

Weak sharp minima on Riemannian manifolds 1

OPTIMALITY CONDITIONS FOR D.C. VECTOR OPTIMIZATION PROBLEMS UNDER D.C. CONSTRAINTS. N. Gadhi, A. Metrane

TEST CODE: PMB SYLLABUS

A Proximal Method for Identifying Active Manifolds

Complexity Analysis of Interior Point Algorithms for Non-Lipschitz and Nonconvex Minimization

The exact absolute value penalty function method for identifying strict global minima of order m in nonconvex nonsmooth programming

Subdifferential representation of convex functions: refinements and applications

Lecture 15 Newton Method and Self-Concordance. October 23, 2008

Convergence of a Two-parameter Family of Conjugate Gradient Methods with a Fixed Formula of Stepsize

Stochastic Optimization with Inequality Constraints Using Simultaneous Perturbations and Penalty Functions

ON THE CONSTRUCTION OF DUALLY FLAT FINSLER METRICS

A REVIEW OF OPTIMIZATION

Analysis Finite and Infinite Sets The Real Numbers The Cantor Set

Thai Journal of Mathematics Volume 14 (2016) Number 1 : ISSN

On Second-order Properties of the Moreau-Yosida Regularization for Constrained Nonsmooth Convex Programs

1. Gradient method. gradient method, first-order methods. quadratic bounds on convex functions. analysis of gradient method

Course 212: Academic Year Section 1: Metric Spaces

OPTIMALITY CONDITIONS FOR GLOBAL MINIMA OF NONCONVEX FUNCTIONS ON RIEMANNIAN MANIFOLDS

8 Numerical methods for unconstrained problems

(7: ) minimize f (x) subject to g(x) E C, (P) (also see Pietrzykowski [5]). We shall establish an equivalence between the notion of

Some Inexact Hybrid Proximal Augmented Lagrangian Algorithms

Chapter 2 Convex Analysis

Unconstrained minimization of smooth functions

Some Background Math Notes on Limsups, Sets, and Convexity

Notions such as convergent sequence and Cauchy sequence make sense for any metric space. Convergent Sequences are Cauchy

Sobolev Spaces. Chapter 10

Duality (Continued) min f ( x), X R R. Recall, the general primal problem is. The Lagrangian is a function. defined by

SOME REMARKS ON SUBDIFFERENTIABILITY OF CONVEX FUNCTIONS

A Dual Condition for the Convex Subdifferential Sum Formula with Applications

Tools from Lebesgue integration

Optimality Conditions for Constrained Optimization

MINIMAL GRAPHS PART I: EXISTENCE OF LIPSCHITZ WEAK SOLUTIONS TO THE DIRICHLET PROBLEM WITH C 2 BOUNDARY DATA

g 2 (x) (1/3)M 1 = (1/3)(2/3)M.

Lecture 3: Basics of set-constrained and unconstrained optimization

Characterizations of the solution set for non-essentially quasiconvex programming

Cubic regularization of Newton s method for convex problems with constraints

Transcription:

Applied Mathematics E-Notes, 3(2003), 130-137 c ISSN 1607-2510 Available free at mirror sites of http://www.math.nthu.edu.tw/ amen/ Remarks On Second Order Generalized Derivatives For Differentiable Functions With Lipschitzian Jacobian Davide La Torre, Matteo Rocca Received 2 October 2002 Abstract Many definitions of second-order generalized derivatives have been introduced to obtain optimality conditions for optimization problems with C 1,1 data. The aim of this note is to show some relations among these definitions. The class of C 1,1 functions, that is the class of differentiable functions with Lipschitzian Jacobian, was first brought to attention by Hiriart-Urruty, Strodiot and Hien Nguyen [8]. The need for investigating such functions, as pointed out in [8] and [9], comes from the fact that several problems of applied mathematics including variational inequalities, semi-infinite programming, penalty functions, augmented Lagrangian, proximal point methods, iterated local minimization by decomposition etc. involve differentiable functions with no hope of being twice differentiable. Many secondorder generalized derivatives have been introduced to obtain optimality conditions for optimization problems with C 1,1 data and for this class of functions numerical methods and minimization algorithms have been proposed too [20, 21]. In this paper we will give some relations among several definitions that one can find in literature and we will compare the necessary second-order optimality conditions expressed by means of these derivatives. We will focus our attention on the definitions due to Hiriart-Urruty, Strodiot and Hien Nguyen [8], Liu [14], Yang and Jeyakumar [23], Peano [18] and Riemann [22]. Some of these definitions do not require the hypothesis of C 1,1 regularity; however, under this assumption, each derivative in the previous list is finite. The definitions introduced by Hiriart-Urruty and Yang-Jeyakumar extend to the secondorder, respectively, the notions due to Clarke and Michel-Penot for the first-order case. Peano and Riemann definitions are classical ones. Peano introduced the homonymous definition while he was studying Taylor s expansion formula for real functions. Peano derivatives were studied and generalized in recent years by Ben-Tal and Zowe [1] and Liu [14], who also obtained optimality conditions. Riemann higher-order derivatives were introduced in the theory of trigonometric series. Furthermore they were developed by several authors. Mathematics Subject Classifications: 26A24, 26A16. Department of Economics, University of Milan, via Conservatorio,7, 20122, Milano, Italy. Department of Economics, University of Insubria, via Ravasi, 2, 21100, Varese, Italy. 130

D. L. Torre and M. Rocca 131 In the following Ω will denote an open subset of R n. DEFINITION 1. A function f : Ω R is locally Lipschitz at x 0 when there exist a constant K and a neighbourhood U of x 0 such that f(x) f(y) K x y, x, y U. DEFINITION 2. A function f : Ω R is of class C 1,1 at x 0 when its first order partial derivatives exist in a neighbourhood of x 0 and are locally Lipschitz at x 0. Some possible applications of C 1,1 functions are shown in the following examples. EXAMPLE 1. Let g : Ω R n R be twice continuously differentiable on Ω and consider f(x) =[g + (x)] 2 where g + (x) =max{g(x), 0}. Thenf is C 1,1 on Ω. EXAMPLE 2. Let f i : R n R, i =0,...,m, and consider the following minimization problem: min f 0 (x) over all x R n such that f 1 (x) 0,..., f m (x) 0. Letting r denote a positive parameter, the augmented Lagrangian L r is defined on R n R m as L r (x, y) =f 0 (x)+ 1 4r m {[y i +2rf i (x)] + } 2 yi 2. i=1 From the general theory of duality which yields L r as a particular Lagrangian, we know that L r (x, ) is concave and also that L r (,y) is convex whenever the minimization problem is a convex minimization problem. By stating y = 0 in the previous expression, we observe that m L r (x, 0) = f 0 (x)+r [f + i (x)]2 is the ordinary penalized version of the minimization problem. L r is differentiable everywhere on R n R m with x L r (x, y) = f 0 (x)+ L r (x, y) =max y i i=1 m [y j +2rf j (x)] + f j (x), j=1 f i (x), y i 2r,i=1,...,m. When the f i are C 2 on R n, L r is C 1,1 on R n+m. The dual problem corresponding to L r is by definition max g r (y) over y R m,whereg r (y) =inf x R n L r (x, y). In the convex case, with r>0, g r is again a C 1,1 concave function with the following uniform Lipschitz property on g r : g r (y) g r (x) 1 2r y y, y, y R m.

132 Generalized Derivatives In [11] the authors have proved the following result which gives a characterization of C 1,1 functions by divided differences. THEOREM 1. [11] Assume that the function f : Ω R is bounded on a neighbourhood of the point x 0 Ω. Then f is of class C 1,1 at x 0 if and only if there exist neighbourhoods U of x 0 and V of 0 R such that δd 2 f(x;h) h is bounded on U V \{0}, 2 d S 1 = {d R n : d =1} where δ d 2f(x; h) =f(x +2hd) 2f(x + hd)+f(x). It is known [23] that if a function f is of class C 1,1 at x 0 then it can be expressed (in a neighbourhood of x 0 ) as the difference of two convex functions. The following corollary strenghtens the results in [23]. COROLLARY 1. If f is of class C 1,1 at x 0,thenf = f + p where f is convex and p is a polynomial of degree at most two. PROOF. From the previous theorem we know that a function f is of class C 1,1 at x 0 if and only if the following difference: δ d 2f(x; d) h 2 = f(x +2hd) 2f(x + hd)+f(x) h 2 is bounded by a constant M for each x in a neighbourhood U of x 0,foreachh in a neighbourhood V of 0 and d S 1. If p(x) =p(x 1,x 2,...,x n )= n i,j=1 c ijx i x j is a polynomial of degree 2 in the variables x 1, x 2,..., x n,itiswellknownthat δ d 2p(x; d) h 2 = n c ij d i d j i,j=1 where d =(d 1,d 2,...d n ). So we can choose the polynomial p such that δ2 d p(x; d) sup d S 1 h 2 M and then for the function f(x) =f(x) p(x) the following inequality holds δ d 2 f(x; d) h 2 0, x U, h V,d S 1 that is f is locally convex. Now we remember the notions of second-order generalized derivative on which we will focus our attention. The following definitions can be introduced without the hypothesis that f is of class C 1,1 but were investigated mostly under this hypothesis. DEFINITION 3. Let us consider a function f : Ω R of class C 1,1 at x 0. i) Peano s second order derivative of f at x 0 in the direction d R n is defined as f f(x 0 + td) f(x 0 ) t f(x 0 )d P (x 0 ; d) =2limsup t 0 t 2. +

D. L. Torre and M. Rocca 133 ii) Riemann s second upper derivative of f at x 0 in the direction d R n is defined as f f(x 0 +2td) 2f(x 0 + td)+f(x 0 ) R(x 0 ; d) = lim sup t 0 t 2. + iii) Yang-Jeyakumar s second upper derivative of f at x 0 in the directions u, v R n is defined as f Y (x 0 ; u, v) = sup z R n lim sup t 0 + f(x 0 + tu + tz)v f(x 0 + tz)v. t iv) Hiriart-Urruty s second upper derivative of f at x 0 in the direction u, v R n is defined as f f(x + tu)v f(x )v H(x 0 ; u, v) = lim sup. x x, t 0 t + REMARK 1. Analogously one can define lower derivatives. In particular Peano s second lower derivative is defined by f (x f(x 0 + td) f(x 0 ) t f(x 0 )d P 0; d) = 2 lim inf t 0 + t 2. f(x 0 +td)d f(x 0 )d REMARK 2. We will denote by f D(x 0 ; d) = lim sup t 0 + t the classical second upper derivative of f at x 0 in the direction d R n. The following result is trivial. THEOREM 2. If f : Ω R is of class C 1,1 at x 0,theneachderivativeinthe previous list is finite (the same holds for lower derivatives). It is known that the following necessary optimality conditions for an unconstrained minimization problem hold. THEOREM 3. [6, 8, 14, 23] Let f : Ω R be a C 1,1 function and assume that x 0 Ω is a local minimum point of f. Then d R n the derivatives f H(x 0 ; d, d), f Y (x 0 ; d, d), f P (x 0 ; d), f R(x 0 ; d), f (x P 0; d), are greater or equal than zero. In the following we will give some relations among the derivatives in Definition 3 that we will allow us to state that the best necessary optimality conditions for an unconstrained minimization problem are those expressed by means of Peano s second upper derivative. REMARK 3. In [23] is given the following chain of inequalities f D(x 0 ; d) f Y (x 0 ; d, d) f H(x 0 ; d, d). Furthermore f Y (x 0 ; d, d) = f H(x 0 ; d, d) if and only if the map f Y ( ; d, d) is upper semicontinuous [23]. REMARK 4. In [5] is given the following characterization of Hiriart-Urruty s generalized derivative f u,v 2 f(y; s, t) H(x 0 ; u, v) = lim sup st y x 0,s,t 0 +

134 Generalized Derivatives where u,v 2 f(y; s, t) =f(y + su + tv) f(y + su) f(y + tv)+f(y). From this characterization one can trivially deduce that f R(x 0 ; d) f H(x 0 ; d, d). LEMMA 1. Let f : Ω R n R be a given C 1,1 function. Then f R(x; d) 2f P (x; d) f (x; d). P PROOF. Choose t n 0 + as n +, such that L = lim n + Obviously L f R(x; d). Let now f(x +2t n d) 2f(x + t n d)+f(x) t 2. n s 1,n =2 f(x + t nd) f(x) t n f(x)d t 2 n and s 2,n =2 f(x +2t nd) f(x) 2t n f(x)d t 2, n eventually by extracting subsequences, s 1,n s 1 and s 2,n s 2 with s 1 f (x; d) P and s 2 4f P (x; d). By a simple calculation, we have s 2,n 2s 1,n =2 f(x +2t nd) 2f(x + t n d)+f(x) t 2 n 2L and then 2L = lim s 2,n 2s 1,n 4f n + P (x; d) 2f (x; d). P THEOREM 4. Let f be a function of class C 1,1 at x 0.Then i) f P (x 0 ; d) f D(x 0 ; d) f Y (x 0 ; d, d) f H(x 0 ; d, d). ii) f P (x 0 ; d) f R(x 0 ; d) f Y (x 0 ; d, d) f H(x 0 ; d, d). PROOF. i) From the previous remarks, it is only necessary to prove the inequality f P (x 0 ; d) f D(x 0 ; d). If we take the function φ 1 (t) =f(x 0 + td) t f(x 0 )d and φ 2 (t) =t 2, applying Cauchy s theorem, we obtain 2 f(x 0 + td) f(x 0 ) t f(x 0 )d t 2 =2 φ 1(t) φ 1 (0) φ 2 (t) φ 2 (0) = 2 φ 1(ξ) φ 2 (ξ) = f(x 0 + ξd)d f(x 0 )d, ξ where ξ = ξ(t) (0,t), and then f P (x 0 ; d) f D(x 0 ; d).

D. L. Torre and M. Rocca 135 ii) From the previous remarks, it is only necessary to prove the inequalities f P (x 0 ; d) f R(x 0 ; d) f Y (x 0 ; d, d). Concerning the first inequality, from the definition of f P (x 0 ; d) wehave where lim sup t 0 + where lim sup t 0 + f(x 0 + td) =f(x 0 )+t f(x 0 )d + t2 2 f P (x 0 ; d)+g(t) g(t) t 2 =0and f(x 0 +2td) =f(x 0 )+2t f(x 0 )d +2t 2 f P (x 0; d)+g(2t) g(2t) g(2t) t = 4 lim sup 2 t 0 + 4t =0. Then 2 f(x 0 +2td) 2f(x 0 + td)+f(x 0 ) t 2 = t2 f P (x 0 ; d)+g(2t) g(t) t 2 f g(2t) g(t) P (x 0 ; d) + lim sup t 0 t + 2 lim sup t 0 t + 2. Then f R(x 0 ; d) f P (x 0 ; d). For the second inequality, we define φ 1 (t) =f(x 0 + 2td) 2f(x 0 + td) andφ 2 (t) =t 2. Then, by Cauchy s theorem, we obtain f(x 0 +2td) 2f(x 0 + td)+f(x 0 ) t 2 = φ 1(t) φ 1 (0) φ 2 (t) φ 2 (0) = φ 1(ξ) φ 2 (ξ) = f(x 0 +2ξd)d f(x 0 + tξ)d, ξ where ξ = ξ(t) (0,t), and then f R(x 0 ; d) f Y (x 0 ; d, d). THEOREM 5. Let f be a function of class C 1,1 at x 0.Ifx 0 is a local minimum for f, then f(x 0 ) = 0 and the following chain of inequalities holds 0 f P (x 0; d) f P (x 0 ; d) f R(x 0 ; d) 2f D(x 0 ; d) 2f Y (x 0 ; d, d) 2f H(x 0 ; d, d) Indeed, the proof follows from Theorem 4 and Lemma 1. REMARK 5. From the previous theorem one can conclude that the best optimality conditions for an unconstrained minimization problem with C 1,1 objective function are those expressed by means of Peano s derivatives. REMARK 6. A similar chain of inequalities holds x Ω when f is convex and of class C 1,1. In fact in this case it is not difficult to prove that f R(x; d) 4f D(x; d) and then, reasoning as in the proof of Theorem 5, we obtain the following chain of inequalities 0 f P (x; d) f P (x; d) f R(x; d) 4f D(x; d) 4f Y (x; d, d) 4f H(x; d, d).

136 Generalized Derivatives References [1] A. Ben-Tal and J. Zowe, Directional derivatives in nonsmooth optimization, Journal of Optimization Theory and Applications, 47(4)(1985), 483 490. [2]J.M.Borwein J.M.,S.P.Fitzpatrick and J.R.Giles,The differentiability of real functions on normed linear space using generalized subgradients, Journal of Mathematical Analysis and Applications, 128(1987), 512 534. [3] W. L. Chan, L. R. Huang and K. F. Ng, On generalized second-order derivatives and Taylor expansion formula in nonsmooth optimization, SIAM Journal on Control and Optimization, 32(3)(1994), 591 611. [4] F. H. Clarke, Optimization and Nonsmooth Analysis, Wiley, New York, 1983. [5] R. Cominetti and R. Correa, A generalized second-order derivative in nonsmooth optimization, SIAM Journal Control and Optimization, 28(1990), 789 809. [6] I. Ginchev and A. Guerraggio, Second order optimality conditions in nonsmooth unconstrained optimization, Pliska Studia Mathematica Bulgarica, 12(1998), 39 50. [7] A. Guerraggio and D.T. Luc, On optimality conditions for C 1,1 vector optimization problems, Journal of Optimization Theory and Applications, 109(3)(2001), 615 629. [8] J. B. Hiriart-Urruty, J. J. Strodiot and H. V. Nguyen, Generalized Hessian matrix and second order optimality conditions for problems with C 1,1 data, Applied Mathematics and Optimization, 11(1984), 43 56. [9] D. Klatte and K. Tammer, On second-order sufficient optimality conditions for C 1,1 optimization problems, Optimization, 19(1988), 169 179. [10] D. Klatte, Upper Lipschitz behavior of solutions to perturbed C 1,1 programs. Mathematical Programming (Ser. B), 88(2000), 285 311. [11] D. La Torre and M. Rocca, C 1,1 functions and optimality conditions, Journal of Computational Analysis and Applications, to appear. [12] D. La Torre and M. Rocca, On C 1,1 constrained optimization problems, Journal of Computational Analysis and Applications, to appear. [13] D. La Torre and M. Rocca, A characterization of C k,1 functions, Real Analysis Exchange, 27(2)(2002), 515 534. [14] L. Liu and M. Krisek, Second order optimality conditions for nondominated solutions of multiobjective programming with C 1,1 data, Applications of Mathematics, 45(5)(2000), 381 397. [15] D.T. Luc, Taylor s formula for C k,1 functions, SIAM Journal on Optimization, 5(1995), 659 669.

D. L. Torre and M. Rocca 137 [16] P. Michel and J. P. Penot, Calcul sous-différential pour des fonctions lischitziennes an nonlipschitziennes, Comptes Rendus de l Academie des Sciences Paris, 298(1984), 269 272. [17] H. W. Oliver, The exact Peano derivative, Transaction Amererican Mathematical Sociecty, 76(1954), 444 456. [18] G. Peano, Sulla formula di Taylor, Atti dell Accademia delle Science di Torino, 27(1891-92), 40 46. [19] L. Qi and W. Sun, A nonsmooth version of Newton s method, Mathematical Programming, 58(1993), 353 367. [20] L. Qi, Superlinearly convergent approximate Newton methods for LC 1 optimization problems, Mathematical Programming, 64(1994), 277-294. [21] L. Qi, LC 1 functions and LC 1 optimization, Operations Research and its applications (D.Z. Du, X.S. Zhang and K. Cheng eds.), World Publishing, Beijing, 1996, 4 13. [22] B. Riemann, Uber die darstellbarkeit einen function durch eine trigonometrische reihe, Ges. Werke, 2 Aufl., Leipzig (1982), 227 271. [23] X. Q. Yang and V. Jeyakumar, Generalized second-order directional derivatives and optimization with C 1,1 functions, Optimization, 26(1992), 165 185.