GENERALIZED second-order cone complementarity

Similar documents
Sequential Quadratic Programming Method for Nonlinear Second-Order Cone Programming Problems. Hirokazu KATO

A Novel Inexact Smoothing Method for Second-Order Cone Complementarity Problems

On the Coerciveness of Merit Functions for the Second-Order Cone Complementarity Problem

Error bounds for symmetric cone complementarity problems

Constrained Optimization Theory

Convex Expected Residual Models for Stochastic Affine Variational Inequality Problems and Its Application to the Traffic Equilibrium Problem

1. Introduction The nonlinear complementarity problem (NCP) is to nd a point x 2 IR n such that hx; F (x)i = ; x 2 IR n + ; F (x) 2 IRn + ; where F is

Asymptotics of minimax stochastic programs

Lecture 8 Plus properties, merit functions and gap functions. September 28, 2008

Expected Residual Minimization Method for Stochastic Linear Complementarity Problems 1

20 J.-S. CHEN, C.-H. KO AND X.-R. WU. : R 2 R is given by. Recently, the generalized Fischer-Burmeister function ϕ p : R2 R, which includes

Numerical Optimization

Applying a type of SOC-functions to solve a system of equalities and inequalities under the order induced by second-order cone

A smoothing Newton-type method for second-order cone programming problems based on a new smoothing Fischer-Burmeister function

Differentiable exact penalty functions for nonlinear optimization with easy constraints. Takuma NISHIMURA

Stochastic Mathematical Programs with Equilibrium Constraints, Modeling and Sample Average Approximation

Exact Augmented Lagrangian Functions for Nonlinear Semidefinite Programming

A Power Penalty Method for a Bounded Nonlinear Complementarity Problem

A derivative-free nonmonotone line search and its application to the spectral residual method

Finite Convergence for Feasible Solution Sequence of Variational Inequality Problems

Research Article The Solution Set Characterization and Error Bound for the Extended Mixed Linear Complementarity Problem

Polynomial complementarity problems

The Relation Between Pseudonormality and Quasiregularity in Constrained Optimization 1

8. Constrained Optimization

A PENALIZED FISCHER-BURMEISTER NCP-FUNCTION. September 1997 (revised May 1998 and March 1999)

Affine scaling interior Levenberg-Marquardt method for KKT systems. C S:Levenberg-Marquardt{)KKTXÚ

A SIMPLY CONSTRAINED OPTIMIZATION REFORMULATION OF KKT SYSTEMS ARISING FROM VARIATIONAL INEQUALITIES

Fischer-Burmeister Complementarity Function on Euclidean Jordan Algebras

Algorithms for nonlinear programming problems II

Assignment 1: From the Definition of Convexity to Helley Theorem

5. Duality. Lagrangian

Stochastic Approximation Approaches to the Stochastic Variational Inequality Problem

ON LICQ AND THE UNIQUENESS OF LAGRANGE MULTIPLIERS

A class of Smoothing Method for Linear Second-Order Cone Programming

Enhanced Fritz John Optimality Conditions and Sensitivity Analysis

Optimality Conditions for Constrained Optimization

Convex Optimization & Lagrange Duality

WHEN ARE THE (UN)CONSTRAINED STATIONARY POINTS OF THE IMPLICIT LAGRANGIAN GLOBAL SOLUTIONS?

Convex Optimization M2

A CHARACTERIZATION OF STRICT LOCAL MINIMIZERS OF ORDER ONE FOR STATIC MINMAX PROBLEMS IN THE PARAMETRIC CONSTRAINT CASE

A proximal-like algorithm for a class of nonconvex programming

On deterministic reformulations of distributionally robust joint chance constrained optimization problems

Properties of Solution Set of Tensor Complementarity Problem

Reformulation of chance constrained problems using penalty functions

Constrained Optimization and Lagrangian Duality

Decision Science Letters

A QP-FREE CONSTRAINED NEWTON-TYPE METHOD FOR VARIATIONAL INEQUALITY PROBLEMS. Christian Kanzow 1 and Hou-Duo Qi 2

A double projection method for solving variational inequalities without monotonicity

Some Properties of the Augmented Lagrangian in Cone Constrained Optimization

minimize x subject to (x 2)(x 4) u,

ON GAP FUNCTIONS OF VARIATIONAL INEQUALITY IN A BANACH SPACE. Sangho Kum and Gue Myung Lee. 1. Introduction

A SHIFTED PRIMAL-DUAL INTERIOR METHOD FOR NONLINEAR OPTIMIZATION

Lecture 3. Optimization Problems and Iterative Algorithms

Introduction to Nonlinear Control Lecture # 3 Time-Varying and Perturbed Systems

Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.16

Confidence regions for stochastic variational inequalities

A Smoothing Newton Method for Solving Absolute Value Equations

Lecture 7 Monotonicity. September 21, 2008

Algorithms for nonlinear programming problems II

Convergence of Stationary Points of Sample Average Two-Stage Stochastic Programs: A Generalized Equation Approach

First-order optimality conditions for mathematical programs with second-order cone complementarity constraints

AN AUGMENTED LAGRANGIAN AFFINE SCALING METHOD FOR NONLINEAR PROGRAMMING

A smoothing augmented Lagrangian method for solving simple bilevel programs

The Karush-Kuhn-Tucker (KKT) conditions

DUALITY, OPTIMALITY CONDITIONS AND PERTURBATION ANALYSIS

UNDERGROUND LECTURE NOTES 1: Optimality Conditions for Constrained Optimization Problems

Local Convergence of Sequential Convex Programming for Nonconvex Optimization

Research Article Solving the Matrix Nearness Problem in the Maximum Norm by Applying a Projection and Contraction Method

A sensitivity result for quadratic semidefinite programs with an application to a sequential quadratic semidefinite programming algorithm

Efficient Methods for Stochastic Composite Optimization

Author(s) Huang, Feimin; Matsumura, Akitaka; Citation Osaka Journal of Mathematics. 41(1)

Introduction to Optimization Techniques. Nonlinear Optimization in Function Spaces

UC Berkeley Department of Electrical Engineering and Computer Science. EECS 227A Nonlinear and Convex Optimization. Solutions 5 Fall 2009

On Generalized Primal-Dual Interior-Point Methods with Non-uniform Complementarity Perturbations for Quadratic Programming

Technische Universität Dresden Herausgeber: Der Rektor

Research Article Algorithms for a System of General Variational Inequalities in Banach Spaces

Convex Optimization Boyd & Vandenberghe. 5. Duality

GENERAL NONCONVEX SPLIT VARIATIONAL INEQUALITY PROBLEMS. Jong Kyu Kim, Salahuddin, and Won Hee Lim

Complexity Analysis of Interior Point Algorithms for Non-Lipschitz and Nonconvex Minimization

A generalized forward-backward method for solving split equality quasi inclusion problems in Banach spaces

N. L. P. NONLINEAR PROGRAMMING (NLP) deals with optimization models with at least one nonlinear function. NLP. Optimization. Models of following form:

1. Introduction. We consider mathematical programs with equilibrium constraints in the form of complementarity constraints:

SOME STABILITY RESULTS FOR THE SEMI-AFFINE VARIATIONAL INEQUALITY PROBLEM. 1. Introduction

Spectral gradient projection method for solving nonlinear monotone equations

Lectures on Parametric Optimization: An Introduction

E5295/5B5749 Convex optimization with engineering applications. Lecture 5. Convex programming and semidefinite programming

Implications of the Constant Rank Constraint Qualification

Additional Homework Problems

APPLICATIONS OF DIFFERENTIABILITY IN R n.

Nonlinear Optimization

Lecture 5. Theorems of Alternatives and Self-Dual Embedding

A Smoothing SQP Method for Mathematical Programs with Linear Second-Order Cone Complementarity Constraints

Stationary Points of Bound Constrained Minimization Reformulations of Complementarity Problems1,2

Lectures 9 and 10: Constrained optimization problems and their optimality conditions

On sequential optimality conditions for constrained optimization. José Mario Martínez martinez

A SHIFTED PRIMAL-DUAL PENALTY-BARRIER METHOD FOR NONLINEAR OPTIMIZATION

Variational Inequalities. Anna Nagurney Isenberg School of Management University of Massachusetts Amherst, MA 01003

Stability of optimization problems with stochastic dominance constraints

Sample Average Approximation Methods for A Class of Stochastic Variational Inequality Problems

system of equations. In particular, we give a complete characterization of the Q-superlinear

Transcription:

Stochastic Generalized Complementarity Problems in Second-Order Cone: Box-Constrained Minimization Reformulation and Solving Methods Mei-Ju Luo and Yan Zhang Abstract In this paper, we reformulate the stochastic generalized second-order cone complementarity problems as boxconstrained optimization problems. If satisfy the condition that the reformulation s objective value is zero, the solutions of box-constrained optimization problems are also solutions of stochastic generalized second-order cone complementarity problems. Since the box-constrained minimization problems contain an expectation function, we then employ sample average approximation method to give approximation problems of the box-constrained minimization reformulation. Meanwhile the convergence results of global solutions and stationary points for the corresponding approximation problems are considered. Index Terms box-constrained minimization problems, sample average approximation, convergence. I. INTRODUCTION GENERALIZED second-order cone complementarity problems, denoted by GSOCCP(F,G,K) are to find a vector x R n satisfying F (x) K, G(x) K, F (x) T G(x) = 0, where F (x), G(x) : R n R n are continuously differentiable respect to x R n. K is the convex cone n K = {x R n x 2 1 a 2 jx 2 j, x 1 0} j=2 and its polar cone K denoted by K = {x R n y K, x, y 0}. Here, a j, j = 2,..., n is the given parameter. The generalized second-order cone complementarity problems play a basic role and occupy an important position in the minimization theory, which have vital applications in many fields. In [2] the authors analyze the case where K is a polyhedral cone. In [3], [4], based on the well know Fischer-Burmeister NCPfunction, smooth merit function is presented for the secondorder cone product of Lorentz cones. In [9], the authors study the special case of GSOCCP(F,G,K), where G(x) = x and K is Cartesian product of Lorentz cones. Moreover, in [1] R. Andreani etc. consider a special cone, which is denoted as K = {x R n x 2 1 (x 2,..., x p ) T 2 }. Since the origin is not a regular point, the authors make up for Manuscript received May 10, 2016; revised August 29, December 5, 2016. This work was supported in part by National Natural Science Foundation of China under Grant NO. 11501275 and Scientific Research Fund of Liaoning Provincial Education Department under Grant NO. L2015199. M.J. Luo is with the School of Mathematics, Liaoning University, Shenyang, Liaoning, 110036, China e-mail: meijuluolnu@126.com. Corresponding author. Y. Zhang is with the School of Mathematics, Liaoning University, Shenyang, Liaoning, 110036, China e-mail: yanzhangdalnu@163.com. this lack by combining two functions which are mentioned in [2]. Then the authors construct two reformulations of GSOCCP(F,G,K) which are nonlinear minimization problems with box constraints. More basic theories, effective algorithms and important applications of GSOCCP(F,G,K) can be found in [7], [13]. However, in practice, stochastic elements are usually involved in several problems. Paying no attention to these stochastic elements will make mistake decision. Therefore, it is meaningful and interesting to study the stochastic generalized second-order cone complementarity problems. Stochastic generalized second-order cone complementarity problems, denoted by SGSOCCP(F,G,K) are to find a vector x R n such that E[F (x, ω)] K, E[G(x, ω)] K, E[F (x, ω)] T E[G(x, ω)] = 0, where E denotes mathematical expectation, ω Ω R m denotes stochastic variable, F (x, ω), G(x, ω) : R n Ω R n are continuously differentiable respect to x R n, the convex cones K and K are defined as above. Without loss of generality, set A = diag(1, a 2 2, a 2 3,..., a 2 p, 0,..., 0), where a i 0, for 2 i p, 1 A = diag(1,, 0,..., 0) and M = a 2 2 diag(m 1, m 2,..., m n ), where m i = 0, for 1 i p, and m i = 1 for i > p, then the convex cone K can be expressed by matrix form, that is,..., 1 a 2 p K = {x R n 1 2 xt Ax 0, x 1 0} and its corresponding polar cone K is K = {x R n 1 2 xt Ax 0, x 1 0, Mx = 0}. In this paper, we suppose a i = 1, for i = 2,..., p. Then, we can rewrite SGSOCCP(F,G,K) as follows: Find x R n such that where E[F (x, ω)] K, E[G(x, ω)] K, E[F (x, ω)] T E[G(x, ω)]=0, (1) K = {x R n x 2 1 (x 2,..., x p ) T 2, x 1 0} and its polar cone K = {x R n x 2 1 (x 2,..., x p ) T 2, x 1 0, Mx = 0}. Since p may be strictly less than n, the cone considered in this paper is more general. This implies, in particular, that K may be different from K.

In fact, problem (1) is also a stochastic complementarity problem. About the classical stochastic complementarity problems(denoted by SCP), there are many good results have been published. For example, the expected value (EV) model suggested by Gürkan et al. [8]. The Expected residual minimization (ERM) model presented by Chen and Fukushima [5]. The Condition value-at-risk (CVaR) model presented by Chen and Lin [6]. The CVaR-constrained stochastic programming model presented by Xu and Yu [18]. More recently research of SCP can be seen in [11], [10], [19]. Based on [1], in this paper, we investigate SGSOCCP(F,G,K), where a special cone denoted as K = {x R n x 2 1 (x 2,..., x p ) T 2 } is considered. The main difference between two papers is that SGSOCCP(F,G,K) contain expectation which is usually difficult to evaluate. Therefore, we employ sample average approximation (SAA) method to solve the corresponding box-constrained minimization reformulation. However, whether the solutions of SAA approximation problems are regarded as the solutions of SGSOCCP(F,G,K) is reasonable? To answer this question, we consider the convergence results of global solutions and stationary points for the corresponding approximation problems in theory. The remainder of the paper is organized as follows: In Section 2, we introduce the box-constrained reformulation and the corresponding approximation problems of SGSOCCP(F, G, K). In Section 3, the convergence results of global solutions and stationary points for the approximation problems are considered. In Section 4, we give the conclusions. II. BOX-CONSTRAINED REFORMULATION OF SGSOCCP(F,G,K) In this section, we will study the equivalence problems of SGSOCCP(F,G,K) and the approximation problems of box-constrained minimization problems. Inspired by [1], in order to reformulate SGSOCCP(F,G,K) via nonlinear programming, we employ the KKT conditions of minimization problems which are related to SGSOCCP(F,G,K) to construct two merit functions denoted by f(x, t) = E[F (x, ω)] t 1 AE[G(x, ω)] t 2 e 1 2 +( 1 2 E[G(x, ω)]t AE[G(x, ω)] t 3 ) 2 +(E[G 1 (x, ω)] t 4 ) 2 + (t 1 t 3 ) 2 + (t 2 t 4 ) 2, g(x, s, ζ) = E[G(x, ω)] s 1 AE[F (x, ω)] s 2 e 1 Mζ 2 +( 1 2 E[F (x, ω)]t AE[F (x, ω)] s 3 ) 2 +(s 1 s 3 ) 2 +(E[F 1 (x, ω)] s 4 ) 2 + ME[F (x, ω)] 2 +(s 2 s 4 ) 2, where x, ζ R n, setting t = (t 1, t 2, t 3, t 4 ) T R 4, s = (s 1, s 2, s 3, s 4 ) T R 4, and e 1 = (1, 0,..., 0) T R n. Then, SGSOCCP(F,G,K) is equivalent to the following box-constrained minimization problem: min (x,t,s,ζ,r) θ(x, t, s, ζ, r)=rf(x, t)+(1 r)g(x, s, ζ) s.t. 0 r 1, t 0, s 0. (2) If (x, t, s, ζ, r) is a solution of (2) with objective value zero, then x is a solution to (1). The detail proof see [1]. Noting that, (2) is a box-constrained minimization problem which is easily to solve and to deal. Since the objective function of problem (2) contains an expectation which is generally difficult to evaluate, we use the sample average approximation (SAA) method to solve (2). The detail can be seen in [15]. We suppose that ψ : Ω R is integrable and Ω k := {ω 1,, ω } represents sample space of random ω, we then use 1 ψ(ω j ) to approximate E[ψ(ω)]. If { } is not a decreasing function respect to k, the strong law of large numbers guarantees that the following result holds with probability one (abbreviated by w.p.1). That is lim k 1 For simplicity, we assume ψ(ω j ) = E[ψ(ω)]. l 1 (x, t 1, t 2 ) = E[F (x, ω)] t 1 AE[G(x, ω)] t 2 e 1, l 2 (x, t 3 ) = 1 2 E[G(x, ω)]t AE[G(x, ω)] t 3, l 3 (x, t 4 ) = E[G 1 (x, ω)] t 4, l 4 (x, s 1, s 2, ζ) = E[G(x, ω)] s 1 AE[F (x, ω)] s 2 e 1 Mζ, l 5 (x, s 3 ) = 1 2 E[F (x, ω)]t AE[F (x, ω)] s 3, l 6 (x, s 4 ) = E[F 1 (x, ω)] s 4, l 7 (x) = ME[F (x, ω)]. Corresponding, we let l1(x, k t 1, t 2 ) = 1 F (x, ω j ( 1 ) t 1 A G(x, ω j ) ) t2 e 1, l2(x, k t 3 ) = ( 1 G(x, ω j ) TA ( 1 ) G(x, ω j ) ) t3, 2 l3(x, k t 4 ) = ( 1 G 1 (x, ω j ) ) t4, l4(x, k s 1, s 2, ζ) = ( 1 G(x, ω j ) ) s1 A ( 1 F (x, ω j ) ) s 2 e 1 Mζ, l5(x, k s 3 ) = ( 1 F (x, ω j ) TA ( 1 ) F (x, ω j ) ) s3, 2 l6(x, k s 4 ) = ( 1 F 1 (x, ω j ) ) s4, l7(x) k = M ( 1 F (x, ω j ) ). Thus, we have f k (x, t) = l k 1(x, t 1, t 2 ) 2 +(l k 2(x, t 3 )) 2 +(l k 3(x, t 4 )) 2 +(t 1 t 3 ) 2 +(t 2 t 4 ) 2, g k (x, s, ζ) = l k 4(x, s 1, s 2, ζ) 2 +(l k 5(x, s 3 )) 2 +(l k 6(x, s 4 )) 2 + l k 7(x) 2 +(s 1 s 3 ) 2 +(s 2 s 4 ) 2. Then, by SAA method one may construct, for each k, an approximation of problem (2) can be constructed as follows: min (x,t,s,ζ,r) θ k (x, t, s, ζ, r)=rf k (x, t)+(1 r)g k (x, s, ζ) s.t. 0 r 1, t 0, s 0. (3)

In this paper, we assume that, for each x R n, F (x, ω) and G(x, ω) are integrable over ω Ω, and for any ω Ω, F (x, ω) and G(x, ω) are continuously differentiable respect to x. Furthermore, we suppose that Ω is a nonempty set and for all x R n and max{ x F (x, ω) 2, x G(x, ω) 2, F (x, ω) 2, G(x, ω) 2 } σ(ω), (4) E[σ(ω)] < +, (5) from (4),(5) and Cauchy-Schwarz inequality, for x R n, we have and max{ x F ( x, ω), x G(x, ω), F (x, ω), G(x, ω) } σ(ω), (6) E[ σ(ω)] < +. (7) In addition, for an n n matrix A, A denotes spectrum norm and A F denotes Frobenius norm, that is n n A F := ( a ij 2 ) 1 2, j=1 where a ij be the factor of matrix A. By the definition, we have A A F. For a function η(x): R n R, x η(x) denotes gradient of η(x) respect to x. III. CONVERGENCE ANALYSIS Before considering the convergence results of (3), we give the following Definitions and Lemmas, which have been given in [12] firstly. Definition 1: Let h(x) : R n R n, for each N, h N (x) : R n R n is a mapping. If for any ϵ > 0, there exists N(ϵ), when N > N(ϵ), we have h N (x) h(x) < ϵ, x R n. Then we call {h N (x)} is uniformly convergent to h(x) on R n. Definition 2: Let h(x) : R n R n, for each N, h N (x) : R n R n is a mapping. If for any sequence {x N } x, when N, we have lim h N(x N ) = h(x), x R n. N Then we call {h N (x N )} is continuously convergent to h(x) on R n. Lemma 1: Let h : R n R n, for each N, h N : R n R n be a mapping, then h N is continuously convergent to h on compact set D if and only if h N is uniformly convergent to h on compact set D. Theorem 1: Suppose that for each k, {(x k, t k, s k, ζ k, r k )} is a global optimal solution of problem (3) and (x, t, s, ζ, r ) is an accumulation point of {(x k, t k, s k, ζ k, r k )}. Then, we have (x, t, s, ζ, r ) is a global optimal solution of problem (2). Proof: Without loss of generality, we may assume lim k (x k, t k, s k, ζ k, r k ) = (x, t, s, ζ, r ). Let B S be a compact convex set containing the sequence {(x k, t k, s k, ζ k, r k )}, where S := {(x, t, s, ζ, r) x R n, t 0, s 0, ζ R n, 0 r 1} denotes the feasible region of (2). Let ϕ(x) = E[F (x, ω)], ϕ k (x) = 1 F (x, ω j ), φ(x) = E[G(x, ω)], φ k (x) = 1 G(x, ω j ). From (6), (7) and Propositon 7 in Chapter 6 of [16], ϕ(x), φ(x) are continuous respect to x and ϕ k (x), φ k (x) are uniformly convergent to ϕ(x), φ(x) on B, respectively. Moreover, from Lemma 1, we have ϕ k (x k ), φ k (x k ) are continuously convergent to ϕ(x ), φ(x ) on B, respectively. Taking a limit for l k 1(x k, t k 1, t k 2),..., l k 7(x k ), we obtain l k 1(x k, t k 1, t k 2),..., l k 7(x k ) are continuously convergent to l 1 (x, t 1, t 2),..., l 7 (x ) on B, respectively. From the definitions of θ(x, t, s, ζ, r) and θ k (x, t, s, ζ, r), we have that θ k (x k, t k, s k, ζ k, r k ) is continuously convergent to θ(x, t, s, ζ, r ) on B. On the other hand, for each k, since (x k, t k, s k, ζ k, r k ) is a global solution of (3), for (x, t, s, ζ, r) S there holds θ k (x k, t k, s k, ζ k, r k ) θ k (x, t, s, ζ, r). (8) Letting k in (8) and by the fact that θ k (x k, t k, s k, ζ k, r k ) is continuously convergent to θ(x, t, s, ζ, r ) on B, we have that θ(x, t, s, ζ, r ) θ(x, t, s, ζ, r), (x, t, s, ζ, r) S, which indicates (x, t, s, ζ, r ) is a global optimal solution of (2). For simplicity, we set θ ω (x, t, s, ζ, r) = rf ω (x, t) + (1 r)g ω (x, s, ζ), where f ω (x, t) = F (x, ω) t 1 AG(x, ω) t 2 e 1 2 +( 1 2 G(x, ω)t AG(x, ω) t 3 ) 2 +(G 1 (x, ω) t 4 ) 2 + (t 1 t 3 ) 2 + (t 2 t 4 ) 2, g ω (x, s, ζ) = G(x, ω) s 1 AF (x, ω) s 2 e 1 Mζ 2 +( 1 2 F (x, ω)t AF (x, ω) s 3 ) 2 +(s 1 s 3 ) 2 +F 1 (x, ω s 4 ) 2 + MF (x, ω) 2 +(s 2 s 4 ) 2. We then show that with the increase of sample size, the optimal solutions of the approximation problem (3) converge exponentially to a solution of problem (2) with probability approaching one. Theorem 2: Suppose that for each k, {(x k, t k, s k, ζ k, r k )} is a global optimal solution of problem (3) and {(x k, t k, s k, ζ k, r k )} itself converges to (x, t, s, ζ, r ). Let B be a compact set that contains the whole sequence {(x k, t k, s k, ζ k, r k )}. Then, for any ε > 0 there exist positive constants C(ε) and β(ε), independent of, such that P rob { sup (x,t,s,ζ,r) B θ k (x, t, s, ζ, r) θ(x, t, s, ζ, r) ε } C(ε)e β(ε). Proof: For all ω Ω and all (x, t, s, ζ, r) B, noting that θ ω and θ are all continuously differentiable with respect to (x, t, s, ζ, r), by the norm inequality or the absolute

value inequality, following from (4)(5)(6)(7), after a simple calculation, we have that there exist κ 1 (ω), κ 2 (ω) satisfying θ ω (x, t, s, ζ, r) θ(x, t, s, ζ, r) κ 1 (ω), θ ω (x, t, s, ζ, r ) θ ω (x, t, s, ζ, r) κ 2 (ω) ( x x + t t + s s + ζ ζ + r r ) and E[κ 1 (ω)] < +, E[κ 2 (ω)] < +. From above results, it is easy to obtain that the conditions of Theorem 5.1 in reference [17] are all hold, and hence completes the proof. However, in general, approximation problems are nonconvex programming. Then we probably get the stationary points rather than the global optimal solutions. Therefore, considering the stationary points are necessary. We next give some definitions associated with stationary points. Definition 3: If there exist Lagrange multipliers β k = (β k 1, β k 2, β k 3, β k 4 ) T R 4, π k = (π k 1, π k 2, π k 3, π k 4 ) T R 4, γ k 1 R, γ k 2 R satisfying x θ k (x k, t k, s k, ζ k, r k ) = 0, (9) 4 t θ k (x k, t k, s k, ζ k, r k ) βi k e i = 0, (10) t k i 0, βi k 0, t k i βi k = 0, (11) 4 s θ k (x k, t k, s k, ζ k, r k ) πj k e j = 0, (12) j=1 s k j 0, π k j 0, s k j π k j = 0, (13) ζ θ k (x k, t k, s k, ζ k, r k ) = 0, (14) r θ k (x k, t k, s k, ζ k, r k ) γ k 1 + γ k 2 = 0, (15) γ k 1 0, γ k 2 0, r k γ k 1 =0, (1 r k )γ k 2 =0, 0 r k 1,(16) where e i denotes a unit vector with the ith component is 1, then we call (x k, t k, s k, ζ k, r k ) is a stationary point of (3). Definition 4: If there exist lagrange multipliers β = (β 1, β 2, β 3, β 4) T R 4, π = (π 1, π 2, π 3, π 4) T R 4, γ 1 R, γ 2 R satisfying x θ(x, t, s, ζ, r ) = 0, (17) 4 t θ(x, t, s, ζ, r ) βi e i = 0, (18) t 0, βi 0, t i βi = 0, (19) 4 s θ(x, t, s, ζ, r ) πj e j = 0, (20) j=1 s j 0, π j 0, s j π = 0, (21) ζ θ(x, t, s, ζ, r ) = 0, (22) r θ(x, t, s, ζ, r ) γ 1 + γ 2 = 0, (23) γ 1 0, γ 2 0, r γ 1 =0, (1 r )γ 2 =0, 0 r 1, (24) then we call (x, t, s, ζ, r ) is a stationary point of (2). We now turn our attentions to Slater constraint qualification, which can ensure the boundedness of Lagrange multipliers in Definition 3. Definition 5: If there exists a vector y R 4, for each i I(t ) := { i t i = 0, 1 i 4} such that y i > 0. Then we call Slater constraint qualification holds. Note that by the condtion (6) and Theory 16.8 of [14], we obtain that f(x, t) is continuously differentiable with respect to x and x f(x, t) = 2l 1 (x, t 1, t 2 ) T E[ x F (x, ω) t 1 A x G(x, ω)] +2l 2 (x, t 3 )E[ x G(x, ω)] T AE[G(x, ω)] +2l 3 (x, t 4 )E[ x G 1 (x, ω)]. Similarly, g(x, s, ζ) is continuously differentiable over x and x g(x, s, ζ) = 2l 4 (x, s 1, s 2 ) T E[ x G(x, ω) s 1 A x F (x, ω)] +2l 5 (x, s 3 )E[ x F (x, ω)] T AE[F (x, ω)] +2l 6 (x, s 4 )E[ x F 1 (x, ω)] + 2l T 7 (x)me[ x F (x, ω)]. Therefore, by the definition of θ(x, t, s, ζ, r), it is easy to obtain that θ(x, t, s, ζ, r) is continuously differentiable with respect to t, s, ζ, r, respectively. What means that θ(x, t, s, ζ, r) is continuously differentiable on S. Lemma 2: θ k (x k, t k, s k, ζ k, r k ) is continuously convergent to θ(x, t, s, ζ, r ) on N (x, t, s, ζ, r ). Proof: We assume lim k (x k, t k, s k, ζ k, r k ) = (x, t, s, ζ, r ), and N (x, t, s, ζ, r ) S is a neighborhood of (x, t, s, ζ, r ) contains {(x k, t k, s k, ζ k, r k )}. For each ω Ω, combining (6), (7) and Proposition 7 in Chapter 6 of [16], we have θ k (x, t, s, ζ, r ) is uniformly convergent to θ(x, t, s, ζ, r ). Combining with Lemma 1, we obtain the conclusion immediately. Theorem 3: Suppose that for each k, (x k, t k, s k, ζ k, r k ) is a stationary point of (3), (x, t, s, ζ, r ) is an accumulation point of {(x k, t k, s k, ζ k, r k )}. Then (x, t, s, ζ, r ) is a stationary point of (2). Proof: For simplicity, we assume lim k (x k, t k, s k, ζ k, r k ) = (x, t, s, ζ, r ). Let β k, π k, γ1 k, γ2 k are the Lagrange multipliers of (9) (16). (i) We first show that the sequence of {β k } is bounded for all k. To this end, we set 4 τ k := βi k. (25) Suppose by contradiction that {β k } is not bounded, which means that there exists a sequence {τ k } such that lim k τ k = +. We may further assume that the limit βi k β i := lim, (i = 1,..., 4) k τ k exists. By (11), we have for each i / I(t ) = { i t i = 0, 1 i 4}, β i = 0. Then following from (25), we obtain i I(t ) β i = 4 β i = 1. (26) For (10), dividing by τ k and taking a limit, following from Lemma 2 and formulation (26), we have 4 β i e i = β i e i = 0. (27) i I(t ) Since Slater constraint qualification holds, then there exists a vector y R 4, for all i I(t ) satisfying y i > 0 and there holds: (y t ) T e i = y i t i = y i > 0, i I(t ). (28)

Let (27) multiplying (y t ) T, by the fact that β i 0 and (28) holds, we have β i = 0 for each i I(t ), which contradicts (26). Therefore {β k } is bounded. Similarly, we can obtain the boundedness of Lagrange multipliers {π k }, {γ k 1 }, {γ k 2 }. (ii) Since {β k }, {π k }, {γ k 1 }, {γ k 2 } are bounded, we may assume that β = lim k β k, π = lim k π k, γ 1 = lim k γ k 1, γ 2 = lim k γ k 2 exist, which together with Lemma 2 and Definition 2, taking a limit in (9) (16), we get (17) (24), which mean (x, t, s, ζ, r ) is a stationary point of problem (2). IV. CONCLUSION In this paper, we present an equivalent problem of SGSOCCP(F, G, K), which is constructed by the convex combination of two given merit functions. Since the boxconstrained minimization problems contain an expectation function, we then use SAA method to give approximation problems. Furthermore the convergence results of global solutions and stationary points for the corresponding approximation problems are considered, the conclusions ensure that it is feasible to regard the solutions of approximation problems as the solutions of SGSOCCP(F, G, K) in theory. [13] M. Lobo, L. Vandenberghe, S. Boyd and H. Lebret. Applications of second-order cone programming, Linear and Algebra and its Applications, vol. 284, no. 1-3, pp. 193-228, Nov. 1998. [14] B. Patrick, Probability and Measure. New York, USA: Wiley- Interscience, 1995. [15] S.M. Robinson. Analysis of sample-path optimization, Mathematics of Operations Research, vol. 21, no. 3, pp. 513-528, Aug. 1996. [16] A. Ruszczynski and A. Shapiro, Stochastic Programming, Handbook in Operations Research and Management Science. Amsterdam, Holland: Elsevier, 2003. [17] A. Shapiro and H. Xu. Stochastic mathematical programs with equilibrium constraints, modelling and sample average approximation, Optimization, vol. 57, no. 3, pp. 395-418, Jun. 2008. [18] L. Xu and B. Yu. CVaR-constrained stochastic programming reformulation for stochastic nonlinear complementarity problems, Computational Optimization and Applications, vol. 58, no. 2, pp. 483-501, Jun. 2014. [19] H. Yu. Minimum mean-squared deviation method for stochastic complementarity problems, Journal of Computational and Applied Mathematics, vol. 93, no. 7, pp. 1173-1187, 2016. ACKNOWLEDGMENT The authors are grateful to the two anonymous referees and editor whose helpful suggestions have led to much improvement of the paper. REFERENCES [1] R. Andreani, A. Friedlander, M.P. Mello and S.A. Santos. Boxconstrained minimization reformulations of complementarity problems in second-order cones, Journal of Global Optimization, vol. 40, no. 4, pp. 505-527, Apr. 2008. [2] R. Andreani, A. Friedlander and S.A. Santos. On the resolution of generalized nonlinear complementarity problems, SIAM journal on Optimization, vol. 12, no. 2, pp. 303-321, Nov. 2001. [3] J.S. Chen. A new merit function and its related properties for the second-order cone complementarity problem, Pacfic Optimization, vol. 2, no. 1, pp. 167-179, Jan. 2006. [4] J.S. Chen and P. Tseng. An unconstrained smooth minimization reformulation of the second cone complementarity problems, Mathematic Programming, vol. 104, no. 2, pp. 293-327, Nov. 2005. [5] X.J. Chen and M. Fukushima. Expected residual minimization method for stochastic linear complementarity problems, Mathematics of Operations Research, vol. 30, no. 4, pp. 1022-1038, Nov. 2005. [6] X.J. Chen and G.H. Lin. CVaR-based formulation and approximation method for stochastic variational inequalities, Numerical Algebra Control and Optimization, vol. 1, no. 1, pp. 35-48, Mar. 2011. [7] M. Fukushima, Z.Q. Luo and P. Tseng. Smoothing functions for second-order cone complementarity problems, SIAM journal on Optimization, vol. 12, no. 2, pp. 436-460, Dec. 2001. [8] G. Gürkan, A.Y. Ozge and S.M. Robinson. Sample-path solution of stochastic variational inequalities, Mathematical Programming, vol. 84, no. 2, pp. 313-333, Feb. 1999. [9] S. Hayashi, N. Yamashita and M. Fukushima. A Combined Smoothing and Regularization Method for Monotone Second-Order Cone Complementarity Problems, SIAM Journal on Optimization, vol. 15, no. 2, pp. 593-615, Feb. 2005. [10] S. He, M. Wei and H. Tong. A smooth penalty-based sample average approximation method for stochastic complementarity problems, Journal of Computational and Applied Mathematics, vol. 287, no. 15, pp. 20-31, Oct. 2015. [11] Y. Huang, H. Liu and S. Zhou. A Barzilai-Borwein type method for stochastic linear complementarity problems, Numerical Algorithms, vol. 67, no. 3, pp. 477-489, Nov. 2014. [12] P. Kall. Approximation to optimization problems: An elementary review, Mathematics of Operations Research, vol. 11, no. 1, pp. 8-17, Feb. 1986.