Multi-step hybrid steepest-descent methods for split feasibility problems with hierarchical variational inequality problem constraints

Similar documents
A Relaxed Explicit Extragradient-Like Method for Solving Generalized Mixed Equilibria, Variational Inequalities and Constrained Convex Minimization

Iterative algorithms based on the hybrid steepest descent method for the split feasibility problem

Algorithms for finding minimum norm solution of equilibrium and fixed point problems for nonexpansive semigroups in Hilbert spaces

Strong Convergence Theorem by a Hybrid Extragradient-like Approximation Method for Variational Inequalities and Fixed Point Problems

A general iterative algorithm for equilibrium problems and strict pseudo-contractions in Hilbert spaces

An Algorithm for Solving Triple Hierarchical Pseudomonotone Variational Inequalities

Iterative common solutions of fixed point and variational inequality problems

Some unified algorithms for finding minimum norm fixed point of nonexpansive semigroups in Hilbert spaces

An iterative method for fixed point problems and variational inequality problems

New Iterative Algorithm for Variational Inequality Problem and Fixed Point Problem in Hilbert Spaces

A General Iterative Method for Constrained Convex Minimization Problems in Hilbert Spaces

Research Article Strong Convergence of a Projected Gradient Method

On an iterative algorithm for variational inequalities in. Banach space

Viscosity approximation methods for the implicit midpoint rule of asymptotically nonexpansive mappings in Hilbert spaces

Iterative Method for a Finite Family of Nonexpansive Mappings in Hilbert Spaces with Applications

On nonexpansive and accretive operators in Banach spaces

The Journal of Nonlinear Science and Applications

A New Modified Gradient-Projection Algorithm for Solution of Constrained Convex Minimization Problem in Hilbert Spaces

ON A HYBRID PROXIMAL POINT ALGORITHM IN BANACH SPACES

Research Article Algorithms for a System of General Variational Inequalities in Banach Spaces

STRONG CONVERGENCE OF AN ITERATIVE METHOD FOR VARIATIONAL INEQUALITY PROBLEMS AND FIXED POINT PROBLEMS

WEAK CONVERGENCE THEOREMS FOR EQUILIBRIUM PROBLEMS WITH NONLINEAR OPERATORS IN HILBERT SPACES

A VISCOSITY APPROXIMATIVE METHOD TO CESÀRO MEANS FOR SOLVING A COMMON ELEMENT OF MIXED EQUILIBRIUM, VARIATIONAL INEQUALITIES AND FIXED POINT PROBLEMS

Monotone variational inequalities, generalized equilibrium problems and fixed point methods

A generalized forward-backward method for solving split equality quasi inclusion problems in Banach spaces

Hybrid extragradient viscosity method for general system of variational inequalities

A NEW ITERATIVE METHOD FOR THE SPLIT COMMON FIXED POINT PROBLEM IN HILBERT SPACES. Fenghui Wang

Split equality monotone variational inclusions and fixed point problem of set-valued operator

Convergence theorems for a finite family. of nonspreading and nonexpansive. multivalued mappings and equilibrium. problems with application

The Split Hierarchical Monotone Variational Inclusions Problems and Fixed Point Problems for Nonexpansive Semigroup

Thai Journal of Mathematics Volume 14 (2016) Number 1 : ISSN

STRONG CONVERGENCE THEOREMS BY A HYBRID STEEPEST DESCENT METHOD FOR COUNTABLE NONEXPANSIVE MAPPINGS IN HILBERT SPACES

On the split equality common fixed point problem for quasi-nonexpansive multi-valued mappings in Banach spaces

On the Weak Convergence of the Extragradient Method for Solving Pseudo-Monotone Variational Inequalities

WEAK AND STRONG CONVERGENCE OF AN ITERATIVE METHOD FOR NONEXPANSIVE MAPPINGS IN HILBERT SPACES

Split hierarchical variational inequality problems and fixed point problems for nonexpansive mappings

The Generalized Viscosity Implicit Rules of Asymptotically Nonexpansive Mappings in Hilbert Spaces

Strong Convergence of Two Iterative Algorithms for a Countable Family of Nonexpansive Mappings in Hilbert Spaces

Split equality problem with equilibrium problem, variational inequality problem, and fixed point problem of nonexpansive semigroups

Convergence Theorems of Approximate Proximal Point Algorithm for Zeroes of Maximal Monotone Operators in Hilbert Spaces 1

Viscosity Approximation Methods for Equilibrium Problems and a Finite Family of Nonspreading Mappings in a Hilbert Space

Research Article Cyclic Iterative Method for Strictly Pseudononspreading in Hilbert Space

Viscosity approximation methods for equilibrium problems and a finite family of nonspreading mappings in a Hilbert space

STRONG CONVERGENCE THEOREMS FOR COMMUTATIVE FAMILIES OF LINEAR CONTRACTIVE OPERATORS IN BANACH SPACES

WEAK CONVERGENCE OF RESOLVENTS OF MAXIMAL MONOTONE OPERATORS AND MOSCO CONVERGENCE

A GENERALIZATION OF THE REGULARIZATION PROXIMAL POINT METHOD

Common fixed points of two generalized asymptotically quasi-nonexpansive mappings

Shih-sen Chang, Yeol Je Cho, and Haiyun Zhou

Iterative Convex Optimization Algorithms; Part One: Using the Baillon Haddad Theorem

Existence and convergence theorems for the split quasi variational inequality problems on proximally smooth sets

Research Article Generalized Mann Iterations for Approximating Fixed Points of a Family of Hemicontractions

SHRINKING PROJECTION METHOD FOR A SEQUENCE OF RELATIVELY QUASI-NONEXPANSIVE MULTIVALUED MAPPINGS AND EQUILIBRIUM PROBLEM IN BANACH SPACES

Strong convergence theorems for fixed point problems, variational inequality problems, and equilibrium problems

GENERAL NONCONVEX SPLIT VARIATIONAL INEQUALITY PROBLEMS. Jong Kyu Kim, Salahuddin, and Won Hee Lim

THROUGHOUT this paper, we let C be a nonempty

The Split Common Fixed Point Problem for Asymptotically Quasi-Nonexpansive Mappings in the Intermediate Sense

Extensions of the CQ Algorithm for the Split Feasibility and Split Equality Problems

Available online at J. Nonlinear Sci. Appl., 10 (2017), Research Article

PROXIMAL POINT ALGORITHMS INVOLVING FIXED POINT OF NONSPREADING-TYPE MULTIVALUED MAPPINGS IN HILBERT SPACES

Research Article Some Krasnonsel skiĭ-mann Algorithms and the Multiple-Set Split Feasibility Problem

Graph Convergence for H(, )-co-accretive Mapping with over-relaxed Proximal Point Method for Solving a Generalized Variational Inclusion Problem

Convergence theorems for mixed type asymptotically nonexpansive mappings in the intermediate sense

CONVERGENCE OF HYBRID FIXED POINT FOR A PAIR OF NONLINEAR MAPPINGS IN BANACH SPACES

A regularization projection algorithm for various problems with nonlinear mappings in Hilbert spaces

Hybrid Steepest Descent Method for Variational Inequality Problem over Fixed Point Sets of Certain Quasi-Nonexpansive Mappings

A Viscosity Method for Solving a General System of Finite Variational Inequalities for Finite Accretive Operators

Viscosity approximation method for m-accretive mapping and variational inequality in Banach space

Viscosity Iterative Approximating the Common Fixed Points of Non-expansive Semigroups in Banach Spaces

Generalized vector equilibrium problems on Hadamard manifolds

Synchronal Algorithm For a Countable Family of Strict Pseudocontractions in q-uniformly Smooth Banach Spaces

CONVERGENCE THEOREMS FOR STRICTLY ASYMPTOTICALLY PSEUDOCONTRACTIVE MAPPINGS IN HILBERT SPACES. Gurucharan Singh Saluja

Strong convergence theorem for amenable semigroups of nonexpansive mappings and variational inequalities

Weak and strong convergence theorems of modified SP-iterations for generalized asymptotically quasi-nonexpansive mappings

Convergence to Common Fixed Point for Two Asymptotically Quasi-nonexpansive Mappings in the Intermediate Sense in Banach Spaces

On Generalized Set-Valued Variational Inclusions

The Strong Convergence Theorems for the Split Equality Fixed Point Problems in Hilbert Spaces

Research Article Some Results on Strictly Pseudocontractive Nonself-Mappings and Equilibrium Problems in Hilbert Spaces

HAIYUN ZHOU, RAVI P. AGARWAL, YEOL JE CHO, AND YONG SOO KIM

INERTIAL ACCELERATED ALGORITHMS FOR SOLVING SPLIT FEASIBILITY PROBLEMS. Yazheng Dang. Jie Sun. Honglei Xu

On the simplest expression of the perturbed Moore Penrose metric generalized inverse

On The Convergence Of Modified Noor Iteration For Nearly Lipschitzian Maps In Real Banach Spaces

Correspondence should be addressed to S. Imnang,

STRONG CONVERGENCE RESULTS FOR NEARLY WEAK UNIFORMLY L-LIPSCHITZIAN MAPPINGS

Developments on Variational Inclusions

Well-posedness for generalized mixed vector variational-like inequality problems in Banach space

ON THE CONVERGENCE OF MODIFIED NOOR ITERATION METHOD FOR NEARLY LIPSCHITZIAN MAPPINGS IN ARBITRARY REAL BANACH SPACES

A double projection method for solving variational inequalities without monotonicity

Strong Convergence Theorems for Nonself I-Asymptotically Quasi-Nonexpansive Mappings 1

ITERATIVE SCHEMES FOR APPROXIMATING SOLUTIONS OF ACCRETIVE OPERATORS IN BANACH SPACES SHOJI KAMIMURA AND WATARU TAKAHASHI. Received December 14, 1999

Viscosity approximation methods for the implicit midpoint rule of nonexpansive mappings in CAT(0) Spaces

arxiv: v1 [math.oc] 20 Sep 2016

638 W. TAKAHASHI, N.-C. WONG, AND J.-C. YAO

A projection-type method for generalized variational inequalities with dual solutions

Research Article An Iterative Algorithm for the Split Equality and Multiple-Sets Split Equality Problem

Convergence Theorems for Bregman Strongly Nonexpansive Mappings in Reflexive Banach Spaces

CONVERGENCE THEOREMS FOR TOTAL ASYMPTOTICALLY NONEXPANSIVE MAPPINGS

Algorithms for Bilevel Pseudomonotone Variational Inequality Problems

Bulletin of the Iranian Mathematical Society Vol. 39 No.6 (2013), pp

1 Introduction and preliminaries

Transcription:

Available online at www.tjnsa.com J. Nonlinear Sci. Appl. 9 (2016), 4148 4166 Research Article Multi-step hybrid steepest-descent methods for split feasibility problems with hierarchical variational inequality problem constraints L. C. Ceng a, Y. C. Liou b,, D. R. Sahu c a Department of Mathematics, Shanghai Normal University, Shanghai 200234, P. R. China. b Department of Healthcare Administration and Medical Informatics, Kaohsiung Medical University, Kaohsiung 80708, Taiwan. c Department of Mathematics, Institute of Science, Banaras Hindu University, Varanasi-221005, India. Communicated by Y. Yao Abstract In this paper, we introduce and analyze a multi-step hybrid steepest-descent algorithm by combining Korpelevich s extragradient method, viscosity approximation method, hybrid steepest-descent method, Mann s iteration method and gradient-projection method (GPM) with regularization in the setting of infinitedimensional Hilbert spaces. Strong convergence was established. c 2016 All rights reserved. Keywords: Hybrid steepest-descent method, split feasibility problem, generalized mixed equilibrium problem, variational inclusion, maximal monotone mapping, nonexpansive mapping. 2010 MSC: 49J30, 47H09, 47J20, 49M05. 1. Introduction Let C be a nonempty closed convex subset of a real Hilbert space H and P C be the metric projection of H onto C. Let S : C H be a nonlinear mapping on C. We denote by Fix(S) the set of fixed points of S and by R the set of all real numbers. A mapping S : C H is called L-Lipschitz continuous (or L-Lipschitzian) if there exists a constant L 0 such that Sx Sy L x y, x, y C. In particular, if L = 1 then S is called a nonexpansive mapping; if L [0, 1) then S is called a contraction. Let A : C H Corresponding author Email addresses: zenglc@hotmail.com (L. C. Ceng), simplex_liou@hotmail.com (Y. C. Liou), drsahudr@gmail.com (D. R. Sahu) Received 2016-06-23

L. C. Ceng, Y. C. Liou, D. R. Sahu, J. Nonlinear Sci. Appl. 9 (2016), 4148 4166 4149 be a nonlinear mapping on C. We consider the following variational inequality problem (VIP): find a point x C such that Ax, y x 0, y C. (1.1) The solution set of VIP (1.1) is denoted by VI(C, A). In 1976, Korpelevich [26] proposed an iterative algorithm for solving the VIP (1.1) in Euclidean space R n, which is known as the extragradient method. Subsequently, many authors improved it in various ways; see e.g., [9 11, 15, 19, 27, 28] and references therein. On the other hand, let C and Q be nonempty closed convex subsets of infinite-dimensional real Hilbert spaces H and H, respectively. The split feasibility problem (SFP) is to find a point x with the property: x C and Ax Q, (1.2) where A B(H, H) and B(H, H) denotes the family of all bounded linear operators from H to H. We denote by Γ the solution set of the SFP (1.2). In 1994, the SFP (1.2) was first introduced by Censor and Elfving [22], in finite-dimensional Hilbert spaces, for modeling inverse problems which arise from phase retrievals and in medical image reconstruction. Assume that the SFP (1.2) is consistent, that is, the solution set Γ of the SFP (1.2) is nonempty. Let f : H R be a continuous differentiable function. The minimization problem min x C f(x) := Ax P Q Ax 2 /2 is ill-posed. In 2010, Xu [35] considered the following Tikhonov regularization problem: min x C f α(x) := 1 2 Ax P QAx 2 + 1 2 α x 2, where α > 0 is the regularization parameter. Very recently, by combining the gradient-projection method with regularization and extragradient method due to Nadezhkina and Takahashi [27], Ceng, Ansari and Yao [12] proposed a Mann type extragradient-like algorithm, and proved that the sequences generated by the proposed algorithm converge weakly to a common solution of the SFP (1.2) and the fixed point problem of a nonexpansive mapping. On the other hand, let S and T be two nonexpansive self-mappings on a nonempty closed convex subset C of a real Hilbert space H. In 2009, Yao, Liou and Marino [40, Theorem 3.2] considered the following hierarchical variational inequality problem (HVIP): find hierarchically a fixed point of T, which is a solution to the VIP for monotone mapping I S; namely, find x Fix(T ) such that (I S) x, p x 0, p Fix(T ). (1.3) The solution set of the HVIP (1.3) is denoted by. It is not hard to check that solving the HVIP (1.3) is equivalent to the fixed point problem of the composite mapping P Fix(T ) S, that is, find x C such that x = P Fix(T ) S x. They introduced and analyzed an iterative algorithm for solving the HVIP (1.3). Furthermore, let ϕ : C R be a real-valued function, A : H H be a nonlinear mapping and Θ : C C R be a bifunction. In 2008, Peng and Yao [28] introduced the following generalized mixed equilibrium problem (GMEP) of finding x C such that Θ(x, y) + ϕ(y) ϕ(x) + Ax, y x 0, y C. (1.4) We denote the set of solutions of GMEP (1.4) by GMEP(Θ, ϕ, A). The GMEP (1.4) is very general in the sense that it includes, as special cases, optimization problems, variational inequalities, minimax problems, Nash equilibrium problems in noncooperative games and others. The GMEP is further considered and studied; see e.g., [7, 9, 13 15, 19]. It was assumed in [28] that Θ : C C R is a bifunction satisfying conditions (A1)-(A4) and ϕ : C R is a lower semicontinuous and convex function with restriction (B1) or (B2): (A1) Θ(x, x) = 0 for all x C;

L. C. Ceng, Y. C. Liou, D. R. Sahu, J. Nonlinear Sci. Appl. 9 (2016), 4148 4166 4150 (A2) Θ is monotone, that is, Θ(x, y) + Θ(y, x) 0 for any x, y C; (A3) Θ is upper-hemicontinuous, that is, for each x, y, z C, lim sup Θ(tz + (1 t)x, y) Θ(x, y); t 0 + (A4) Θ(x, ) is convex and lower semicontinuous for each x C; (B1) for each x H and r > 0, there exists a bounded subset D x C and y x C such that for any z C \ D x, Θ(z, y x ) + ϕ(y x ) ϕ(z) + 1 r y x z, z x < 0; (B2) C is a bounded set. In addition, let B be a single-valued mapping of C into H and R be a multivalued mapping with D(R) = C. Consider the following variational inclusion: find a point x C such that 0 Bx + Rx. (1.5) We denote by I(B, R) the solution set of the variational inclusion (1.5). Let a set-valued mapping R : D(R) H 2 H be maximal monotone. We define the resolvent operator J R,λ : H D(R) associated with R and λ by J R,λ x = (I + λr) 1 x, x H, where λ is a positive number. In 1998, Huang [25] studied problem (1.5) in the case where R is maximal monotone and B is strongly monotone and Lipschitz continuous with D(R) = C = H. Subsequently, Zeng, Guu and Yao [45] further studied problem (1.5) in the case which is more general than Huang s one [25]. In this paper, we introduce and analyze a multi-step hybrid steepest-descent algorithm by combining Korpelevich s extragradient method, viscosity approximation method, hybrid steepest-descent method, Mann s iteration method and gradient-projection method (GPM) with regularization in the setting of infinite-dimensional Hilbert spaces. It is proven that under appropriate assumptions the proposed algorithm converges strongly to a solution of the SFP (1.2) with constraints of several problems: finitely many GMEPs, finitely many variational inclusions and the fixed point problem of an infinite family of nonexpansive mappings. Our results improve, extend and develop the corresponding results in the literature; see e.g., [40, Theorems 3.1 and 3.2] and [12, Theorem 3.2]. Recent results in this directions can be also found, e.g., in [1, 2, 4, 5, 8, 16, 17, 20, 21, 32, 33, 37 39, 42 44]. 2. Preliminaries Throughout this paper, we assume that H is a real Hilbert space whose inner product and norm are denoted by, and, respectively. Let C be a nonempty closed convex subset of H. We write x n x (resp. x n x) to indicate that the sequence {x n } converges weakly (resp. strongly) to x. Moreover, we use ω w (x n ) to denote the weak ω-limit set of {x n }, that is, ω w (x n ) := {x H : x ni x for some subsequence {x ni } of {x n }}. The metric projection from H onto C is the mapping P C : H C which assigns to each point x H, the unique point P C x C such that x P C x = inf y C x y. Definition 2.1. Let T be a mapping with domain D(T ) H and range R(T ) H. Then T is said to be (i) monotone if T x T y, x y 0, x, y D(T ); (ii) β-strongly monotone if T x T y, x y η x y 2, x, y D(T ), for some β > 0; (iii) ν-inverse-strongly monotone if T x T y, x y ν T x T y 2, x, y D(T ), for some ν > 0. It is clear that if T is ν-inverse-strongly monotone, then T is monotone and 1 ν -Lipschitz continuous. Moreover, one also has that, for all u, v D(T ) and λ > 0, (I λt )u (I λt )v 2 u v 2 + λ(λ 2ν) T u T v 2. (2.1) So, if λ 2ν, then I λt is a nonexpansive mapping. Next, some important properties of projections are

L. C. Ceng, Y. C. Liou, D. R. Sahu, J. Nonlinear Sci. Appl. 9 (2016), 4148 4166 4151 gathered in the following proposition. Proposition 2.2 ([24]). For given x H and z C: (i) z = P C x x z, y z 0, y C; (ii) z = P C x x z 2 x y 2 y z 2, y C; (iii) P C x P C y, x y P C x P C y 2, y H. Definition 2.3. A mapping T : H H is said to be firmly nonexpansive if 2T I is nonexpansive, or equivalently, if T is 1-inverse strongly monotone (1-ism); alternatively, T is firmly nonexpansive if and only if T is expressed as T = (I + S)/2, where S is nonexpansive on H. Proposition 2.4 ([18]). Assume that Θ : C C R satisfies (A1)-(A4) and let ϕ : C R be a proper lower semicontinuous and convex function. Assume that either (B1) or (B2) holds. For r > 0 and x H, define a mapping T r (Θ,ϕ) : H C as follows: T r (Θ,ϕ) (x) = {z C : Θ(z, y) + ϕ(y) ϕ(z) + 1 y z, z x 0, y C} r for all x H. Then the following hold: (i) for each x H, T r (Θ,ϕ) (x) is nonempty and single-valued; (ii) T (Θ,ϕ) r is firmly nonexpansive, that is, T (Θ,ϕ) r x T (Θ,ϕ) r y 2 T r (Θ,ϕ) x T r (Θ,ϕ) y, x y for any x, y H; (iii) Fix(T r (Θ,ϕ) ) = MEP(Θ, ϕ); (iv) MEP(Θ, ϕ) is closed and convex; (v) T (Θ,ϕ) s x T (Θ,ϕ) t x 2 s t s (Θ,ϕ) T s x T (Θ,ϕ) t x, T (Θ,ϕ) s x x for all s, t > 0 and x H. Definition 2.5. A mapping T : H H is said to be an averaged mapping if it can be written as the average of the identity I and a nonexpansive mapping, that is, T (1 α)i + αs where α (0, 1) and S : H H is nonexpansive. More precisely, when the last equality holds, we say that T is α-averaged. Thus firmly nonexpansive mappings (in particular, projections) are 1 2-averaged mappings. Proposition 2.6 ([6]). Let T : H H be a given mapping. (i) T is nonexpansive if and only if the complement I T is 1 2 -ism. (ii) If T is ν-ism, then for γ > 0, γt is ν γ -ism. (iii) T is averaged if and only if the complement I T is ν-ism for some ν > 1/2. Indeed, for α (0, 1), T is α-averaged if and only if I T is 1 2α -ism. Proposition 2.7 ([6, 23]). Let S, T, V : H H be given operators. (i) If T = (1 α)s + αv for some α (0, 1) and if S is averaged and V is nonexpansive, then T is averaged. (ii) T is firmly nonexpansive if and only if the complement I T is firmly nonexpansive.

L. C. Ceng, Y. C. Liou, D. R. Sahu, J. Nonlinear Sci. Appl. 9 (2016), 4148 4166 4152 (iii) If T = (1 α)s + αv for some α (0, 1) and if S is firmly nonexpansive and V is nonexpansive, then T is averaged. (iv) The composite of finitely many averaged mappings is averaged. That is, if each of the mappings {T i } N is averaged, then so is the composite T 1 T N. In particular, if T 1 is α 1 -averaged and T 2 is α 2 -averaged, where α 1, α 2 (0, 1), then the composite T 1 T 2 is α-averaged, where α = α 1 + α 2 α 1 α 2. We need some facts and tools in a real Hilbert space H which are listed as lemmas below. Lemma 2.8 ([31]). Let X be a real inner product space. Then the following inequality holds: x + y 2 x 2 + 2 y, x + y, x, y X. Lemma 2.9 ([31]). Let H be a real Hilbert space. Then the following hold: (a) x y 2 = x 2 y 2 2 x y, y for all x, y H; (b) λx + µy 2 = λ x 2 + µ y 2 λµ x y 2 for all x, y H and λ, µ [0, 1] with λ + µ = 1. Let {T n } n=1 be an infinite family of nonexpansive self-mappings on C and {ρ n} n=1 nonnegative numbers in [0, 1]. For any n 1, define a self-mapping W n on C as follows: U n,n+1 = I, U n,n = ρ n T n U n,n+1 + (1 ρ n )I, U n,k = ρ k T k U n,k+1 + (1 ρ k )I, U n,2 = ρ 2 T 2 U n,3 + (1 ρ 2 )I, W n = U n,1 = ρ 1 T 1 U n,2 + (1 ρ 1 )I. be a sequence of (2.2) Such a mapping W n is called the W -mapping generated by T n, T n 1,..., T 1 and ρ n, ρ n 1,..., ρ 1. Lemma 2.10 ([30]). Let C be a nonempty closed convex subset of a real Hilbert space H. Let {T n } n=1 be a sequence of nonexpansive self-mappings on C such that n=1 Fix(T n) and let {ρ n } be a sequence in (0, θ] for some θ (0, 1). Then, for every x C and k 1 the limit lim n U n,k x exists. Remark 2.11 ([41], Remark 3.1). It can be known from Lemma 2.10 that if D is a nonempty bounded subset of C, then for ɛ > 0 there exists n 0 k such that for all n > n 0, sup x D U n,k x U k x ɛ. Remark 2.12 ([41]). Utilizing Lemma 2.10, we define a mapping W : C C by W x = lim n W n x = lim n U n,1 x, x C. Such a W is called the W -mapping generated by T 1, T 2,... and ρ 1, ρ 2,... Since W n is nonexpansive, W : C C is also nonexpansive. If {x n } is a bounded sequence in C, then from Remark 2.12, one can show that lim n W n x n W x n = 0. Lemma 2.13 ([30]). Let C be a nonempty closed convex subset of a real Hilbert space H. Let {T n } n=1 be a sequence of nonexpansive self-mappings on C such that n=1 Fix(T n), and let {ρ n } be a sequence in (0, θ] for some θ (0, 1). Then, Fix(W ) = n=1 Fix(T n). Lemma 2.14 ([24]). (Demiclosedness principle). Let C be a nonempty closed convex subset of a real Hilbert space H. Let S be a nonexpansive self-mapping on C. Then I S is demiclosed. That is, whenever {x n } is a sequence in C weakly converging to some x C and the sequence {(I S)x n } strongly converges to some y, it follows that (I S)x = y. Here I is the identity operator of H. Let C be a nonempty closed convex subset of a real Hilbert space H. We introduce some notations. Let λ be a number in (0, 1] and let µ > 0. Associating with a nonexpansive mapping T : C H, we define the

L. C. Ceng, Y. C. Liou, D. R. Sahu, J. Nonlinear Sci. Appl. 9 (2016), 4148 4166 4153 mapping T λ : C H by T λ x := T x λµf (T x), x C, where F : H H is an operator such that, for some positive constants κ, η > 0, F is κ-lipschitzian and η-strongly monotone on H; that is, F satisfies the condition that for all x, y H, F x F y κ x y and F x F y, x y η x y 2. Lemma 2.15 ([36], Lemma 3.1). T λ is a contraction provided 0 < µ < 2η κ 2 ; that is, where τ = 1 1 µ(2η µκ 2 ) (0, 1]. T λ x T λ y (1 λτ) x y, x, y C, Lemma 2.16 ([34]). Let {a n } be a sequence of nonnegative real numbers satisfying the property: where {s n } (0, 1] and {b n } are such that (i) n=1 s n = ; (ii) either lim sup n b n 0 or n=0 s nb n < ; (iii) n=1 t n < where t n 0, for all n 1. Then lim n a n = 0. a n+1 (1 s n )a n + s n b n + t n, n 1, Recall that a set-valued mapping T : D(T ) H 2 H is called monotone if for all x, y D(T ), f T x and g T y imply that f g, x y 0. A set-valued mapping T is called maximal monotone if T is monotone and (I + λt )D(T ) = H for each λ > 0, where I is the identity mapping of H. We denote by G(T ) the graph of T. It is known that a monotone mapping T is maximal if and only if, for (x, f) H H, f g, x y 0 for every (y, g) G(T ) implies f T x. Next we provide an example to illustrate the concept of maximal monotone mapping. Let A : C H be a monotone, k-lipschitz-continuous mapping and let N C v be the normal cone to C at v C, that is, N C v = {u H : v p, u 0, p C}. Define { Av + NC v, if v C, T v =, if v C. Then, T is maximal monotone (see [29]) such that 0 T v v VI(C, A). Let R : D(R) H 2 H be a maximal monotone mapping. Let λ, µ > 0 be two positive numbers. Lemma 2.17 ([3]). There holds the resolvent identity ( µ J R,λ x = J R,µ λ x + (1 µ ) λ )J R,λx, x H. Remark 2.18. For λ, µ > 0, there holds the following relation ( 1 J R,λ x J R,µ y x y + λ µ λ J R,λx y + 1 ) µ x J R,µy, x, y H. (2.3) In terms of Huang [25] (see also [45]), there holds the following property for the resolvent operator J R,λ : H D(R). Lemma 2.19. J R,λ is single-valued and firmly nonexpansive, that is, J R,λ x J R,λ y, x y J R,λ x J R,λ y 2, x, y H. Consequently, J R,λ is nonexpansive and monotone.

L. C. Ceng, Y. C. Liou, D. R. Sahu, J. Nonlinear Sci. Appl. 9 (2016), 4148 4166 4154 Lemma 2.20 ([10]). Let R be a maximal monotone mapping with D(R) = C. Then for any given λ > 0, u C is a solution of problem (1.5) if and only if u C satisfies u = J R,λ (u λbu). Lemma 2.21 ([45]). Let R be a maximal monotone mapping with D(R) = C and let B : C H be a strongly monotone, continuous and single-valued mapping. Then for each z H, the equation z (B + λr)x has a unique solution x λ for λ > 0. Lemma 2.22 ([10]). Let R be a maximal monotone mapping with D(R) = C and B : C H be a monotone, continuous and single-valued mapping. Then (I + λ(r + B))C = H for each λ > 0. In this case, R + B is maximal monotone. 3. Main Results We now state and prove the main result of this paper. Let H be a real Hilbert space and f : H R be a function. Then the minimization problem min x C f(x) := 1 2 Ax P QAx 2 is ill-posed. Xu [35] considered the following Tikhonov s regularization problem: min f α(x) := 1 x C 2 Ax P QAx 2 + 1 2 α x 2, where α > 0 is the regularization parameter. It is clear that the gradient f α = f +αi = A (I P Q )A+αI is (α + A 2 )-Lipschitz continuous. We are now in a position to state and prove the main result in this paper. Theorem 3.1. Let C be a nonempty closed convex subset of a real Hilbert space H. Let M, N be two positive integers. Let Θ k be a bifunction from C C to R satisfying (A1)-(A4) and ϕ k : C R {+ } be a proper lower semicontinuous and convex function with restriction (B1) or (B2), where k {1, 2,..., M}. Let R i : C 2 H be a maximal monotone mapping and let A k : H H and B i : C H be µ k -inverse strongly monotone and η i -inverse strongly monotone, respectively, where k {1, 2,..., M}, i {1, 2,..., N}. Let S : H H be a nonexpansive mapping and V : H H be a ρ-contraction with coefficient ρ [0, 1). Let {T n } n=1 be a sequence of nonexpansive self-mappings on C and {ρ n} be a sequence in (0, θ] for some θ (0, 1). Let F : H H be κ-lipschitzian and η-strongly monotone with positive constants κ, η > 0 such that 0 γ < τ and 0 < µ < 2η where τ = 1 1 µ(2η µκ κ 2 ). Assume that Ω := 2 n=1 Fix(T n) M GMEP(Θ k, ϕ k, A k ) N I(B 2 i, R i ) Γ. Let {λ n } [a, b] (0, ), {α A 2 n } (0, ) with n=1 α n <, {ɛ n }, { }, {β n }, {γ n }, {σ n } (0, 1) with β n + γ n + σ n = 1, and {λ i,n } [a i, b i ] (0, 2η i ), {r k,n } [c k, d k ] (0, 2µ k ) where i {1, 2,..., N} and k {1, 2,..., M}. For arbitrarily given x 1 H, let {x n } be a sequence generated by u n = T (Θ M,ϕ M ) r M,n (I r M,n A M )T (Θ M 1,ϕ M 1 ) r M 1,n (I r M 1,n A M 1 ) T (Θ 1,ϕ 1 ) r 1,n (I r 1,n A 1 )x n, v n = J RN,λ N,n (I λ N,n B N )J RN 1,λ N 1,n (I λ N 1,n B N 1 ) J R1,λ 1,n (I λ 1,n B 1 )u n, y n = β n x n + γ n P C (I λ n f αn )v n + σ n W n P C (I λ n f αn )v n, x n+1 = ɛ n γ( V x n + (1 )Sx n ) + (I ɛ n µf )y n, n 1, where f αn = α n I + f and W n is the W -mapping generated by (2.2). Suppose that (C1) lim n ɛ n = 0, n=1 ɛ 1 n = and lim n ɛ n 1 1 = 0; (C2) lim sup n ɛn <, lim n 1 ɛ n 1 1 1 = 0 and lim n 1 1 ɛ n 1 ɛ n = 0; (3.1)

L. C. Ceng, Y. C. Liou, D. R. Sahu, J. Nonlinear Sci. Appl. 9 (2016), 4148 4166 4155 (C3) lim n θ n β ɛ n = 0, lim n β n 1 γ n ɛ n = 0 and lim n γ n 1 n ɛ n = 0; (C4) lim n λ i,n λ i,n 1 ɛ n = 0 and lim n r k,n r k,n 1 ɛ n = 0 for i = 1, 2,..., N and k = 1, 2,..., M; λ (C5) lim n λ n 1 λ n ɛ n = 0, lim nα n λ n 1 α n 1 α n ɛ n = 0 and lim n n δn = 0; (C6) 0 < lim inf n β n lim sup n (β n + σ n ) < 1 and lim inf n σ n > 0. Then we have: (i) lim n x n+1 x n = 0; (ii) ω w (x n ) Ω; (iii) {x n } converges strongly to a solution x of the SFP (1.2), which is a unique solution in Ω to the HVIP Proof. First of all, observe that µη τ κ η, and (µf γs)x, p x 0, p Ω. (µf γs)x (µf γs)y, x y (µη γ) x y 2, x, y H. Since 0 γ < τ and κ η, we know that µη τ > γ and hence the mapping µf γs is (µη γ)-strongly monotone. Moreover, it is clear that the mapping µf γs is (µκ + γ)-lipschitzian. Thus, there exists a unique solution x in Ω to the VIP That is, {x } = VI(Ω, µf γs). Now, we put (µf γs)x, p x 0, p Ω. k n = T (Θ k,ϕ k ) r k,n (I r k,n A k )T (Θ k 1,ϕ k 1 ) r k 1,n (I r k 1,n A k 1 ) T (Θ 1,ϕ 1 ) (I r 1,n A 1 )x n for all k {1, 2,..., M} and n 1, Λ i n = J Ri,λ i,n (I λ i,n B i )J Ri 1,λ i 1,n (I λ i 1,n B i 1 ) J R1,λ 1,n (I λ 1,n B 1 ) for all i {1, 2,..., N}, 0 n = I and Λ 0 n = I, where I is the identity mapping on H. Then we have that u n = M n x n and v n = Λ N n u n. In addition, in terms of condition (C6), we may assume, without loss of generality, that {β n } r 1,n ( 0, 2 α+ A 2 ), where ζ = [c, d] (0, 1). Now, we show that P C (I λ f α ) is ζ-averaged for each λ ( 2 + λ(α + A 2 ) ) /4 (0, 1). Indeed, since f = A 1 (I P Q )A is -ism, it is easy to see that A 2 (α + A 2 ) f α (x) f α (y), x y f α (x) f α (y) 2. Hence, it follows that f α = αi + A 1 (I P Q )A is -ism. Thus, by Proposition 2.6 (ii), λ f α+ A 2 α is 1 λ(α+ A 2 ) -ism. From Proposition 2.6 (iii), the complement I λ f α is λ(α+ A 2 ) 2 -averaged. Therefore, noting that P C is 1 2 -averaged and utilizing Proposition 2.7 (iv), we obtain that for each λ (0, 2 ), P α+ A 2 C (I λ f α ) is ζ-averaged with ζ = ( 2 + λ(α + A 2 ) ) /4 (0, 1). This shows that P C (I λ f α ) is nonexpansive. 2 Taking into account that {λ n } [a, b] (0, ) and α A 2 n 0, we get lim sup n 2 + λ n (α n + A 2 ) 4 2 + b A 2 4 < 1.

L. C. Ceng, Y. C. Liou, D. R. Sahu, J. Nonlinear Sci. Appl. 9 (2016), 4148 4166 4156 Without loss of generality, we may assume that ζ n := 2+λn(αn+ A 2 ) 4 < 1 for each n 1. So, P C (I λ n f αn ) λ is nonexpansive for each n 1. Since lim sup n(α n+ A 2 ) n 2 b A 2 2 < 1, it is known similarly that I λ n f αn is nonexpansive for each n 1. Next, we divide the rest of the proof into several steps. Step 1. We prove that {x n } is bounded. Indeed, take a fixed p Ω arbitrarily. Utilizing (2.1) and Proposition 2.4 (ii) we have u n p M 1 n x n M 1 n p 0 nx n 0 np = x n p. (3.2) Utilizing (2.1) and Lemma 2.19 we have v n p = Λ N n u n Λ N n p Λn N 1 u n Λ N 1 n p Λ 0 nu n Λ 0 np = u n p. (3.3) Combining (3.2) and (3.3), we have v n p x n p. (3.4) For simplicity, put t n = P C (I λ n f αn )v n for each n 0. Note that P C (I λ f)p = p for λ (0, 2 A 2 ). Hence, from (3.4), it follows that t n p P C (I λ n f αn )v n P C (I λ n f αn )p + P C (I λ n f αn )p P C (I λ n f)p v n p + λ n α n p x n p + λ n α n p. (3.5) Since W n p = p for all n 1 and W n is a nonexpansive mapping, we obtain from (3.1) and (3.5) that y n p β n x n p + γ n t n p + σ n W n t n p β n x n p + (γ n + σ n ) ( x n p + λ n α n p ) x n p + λ n α n p. (3.6) Utilizing Lemma 2.16, we deduce from (3.1), (3.6), {λ n } [a, b] (0, 2 A 2 ) and 0 γ < τ that for all n 1 x n+1 p ɛ n (γv x n µf p) + (1 )(γsx n µf p) + (I ɛ n µf )y n (I ɛ n µf )p ɛ n [ (γρ x n p + γv p µf p + (1 )(γ x n p + γsp µf p )] +(1 ɛ n τ) y n p ɛ n γ x n p + ɛ n max { γv p µf p, γsp µf p } +(1 ɛ n τ) x n p + α n b p max { x n p, γv p µf p τ γ, γsp µf p τ γ } + α n b p. By induction, we get { x n+1 p max x 1 p, γv p µf p, τ γ } γsp µf p + τ γ n α j b p j=1 for all n 1. and {y n }. Thus, {x n } is bounded (due to n=1 α n < ) and so are the sequences {t n }, {u n }, {v n }, Step 2. We prove that lim n x n+1 x n = 0. Indeed, utilizing (2.1) and (2.3), we obtain that v n+1 v n = Λ N n+1 u n+1 Λ N n u n J RN,λ N,n+1 (I λ N,n+1 B N )Λ N 1 n+1 u n+1 J RN,λ N,n+1 (I λ N,n B N )Λ N 1 n+1 u n+1

L. C. Ceng, Y. C. Liou, D. R. Sahu, J. Nonlinear Sci. Appl. 9 (2016), 4148 4166 4157 where sup n 1 + J RN,λ N,n+1 (I λ N,n B N )Λ N 1 n+1 u n+1 J RN,λ N,n (I λ N,n B N )Λ N 1 n u n ( λ N,n+1 λ N,n B N Λ N 1 n+1 u n+1 + M ) + Λ N 1 n+1 u n+1 Λ N 1 n u n ( λ N,n+1 λ N,n B N Λ N 1 n+1 u n+1 + M ) ( + λ N 1,n+1 λ N 1,n B N 1 Λ N 2 n+1 u n+1 + M ) ( + + λ 1,n+1 λ 1,n B 1 Λ 0 n+1u n+1 + M ) + Λ 0 n+1u n+1 Λ 0 nu n M 0 N λ i,n+1 λ i,n + u n+1 u n, 1 JRN {,λ λ N,n+1 (I λ N,n B N )Λ N 1 n+1 u n+1 (I λ N,n B N )Λ N 1 n u n N,n+1 + 1 (I λn,n B N )Λn+1 N 1 λ u n+1 J RN,λ N,n (I λ N,n B N )Λ N 1 } n u n M N,n for some M { N > 0 and sup n 0 B iλ i 1 n+1 u n+1 + M } M 0 for some M 0 > 0. Utilizing Proposition 2.4 (ii), (v), we deduce that u n+1 u n = M n+1x n+1 M n x n T (Θ M,ϕ M ) r M,n+1 (I r M,n+1 A M ) M 1 + T (Θ M,ϕ M ) r M,n (I r M,n A M ) M 1 r M,n+1 r M,n [ A M M 1 n+1 x n+1 + M 1 1 + T (Θ M,ϕ M ) r r M,n+1 M,n+1 r M,n n+1 x n+1 T (Θ M,ϕ M ) (I r M,n A M ) M 1 n+1 x n+1 n+1 x n+1 T (Θ M,ϕ M ) (I r M,n A M ) M 1 n x n r M,n n+1 x n+1 M 1 n x n (I r M,n+1 A M ) M 1 n+1 x n+1 (I r M,n+1 A M ) M 1 n+1 x n+1 ] r M,n+1 r M,n [ A M M 1 n+1 x n+1 1 + T (Θ M,ϕ M ) r r M,n+1 (I r M,n+1 A M ) M 1 n+1 x n+1 M,n+1 (I r M,n+1 A M ) M 1 n+1 x n+1 ] + + r 1,n+1 r 1,n [ A 1 0 n+1x n+1 + 1 T (Θ 1,ϕ 1 ) r r 1,n+1 (I r 1,n+1 A 1 ) 0 n+1x n+1 (I r 1,n+1 A 1 ) 0 n+1x n+1 ] 1,n+1 + 0 n+1x n+1 0 nx n M M 1 r k,n+1 r k,n + x n+1 x n, where M 1 > 0 is a constant such that for each n 1 M [ A k k 1 n+1 x n+1 + 1 T (Θ k,ϕ k ) r r k,n+1 (I r k,n+1 A k ) k 1 n+1 x n+1 (I r k,n+1 A k ) k 1 n+1 x n+1 ] M 1. k,n+1 Furthermore, we define y n = β n x n + ( )w n for all n 1. It follows that w n+1 w n = γ n+1(t n+1 t n ) + σ n+1 (W n+1 t n+1 W n t n ) +1 γ n+1 + ( γ n σ n+1 )t n + ( σ n )W n t n. +1 +1 (3.7) (3.8) (3.9)

L. C. Ceng, Y. C. Liou, D. R. Sahu, J. Nonlinear Sci. Appl. 9 (2016), 4148 4166 4158 Taking into account the nonexpansivity of W n, T k and U n,k, from (2.2) we get n+1 W n+1 t n+1 W n t n t n+1 t n + ρ j T n+1 t n t n. (3.10) By the nonexpansivity of P C (I λ n f αn ) we have t n+1 t n v n+1 v n + λ n+1 α n+1 λ n α n v n + λ n+1 λ n f(v n ). (3.11) Hence it follows from (3.7)-(3.11) and {ρ n } (0, θ] (0, 1) that j=1 Note that w n+1 w n γ n+1 t n+1 t n + σ n+1 W n+1 t n+1 W n t n γ n+1 + γ n t n +1 +1 σ n+1 + σ n W n t n +1 γ n+1 t n+1 t n + γ n+1 n ( t n + W n t n ) + ρ j T n+1 t n t n +1 v n+1 v n + λ n+1 α n+1 λ n α n v n + λ n+1 λ n f(v n ) γ n+1 j=1 + γ n ( t n + W n t n ) + θ n+1 T n+1 t n t n +1 M 0 N λ i,n+1 λ i,n + M M 1 r k,n+1 r k,n + x n+1 x n + λ n+1 α n+1 λ n α n v n + λ n+1 λ n f(v n ) γ n+1 + γ n ( t n + W n t n ) + θ n+1 T n+1 t n t n. +1 (3.12) It follows from (3.12) that y n+1 y n = β n (x n+1 x n ) + ( )(w n+1 w n ) + (β n+ )(x n+1 w n+1 ). y n+1 y n β n x n+1 x n + ( ) w n+1 w n + β n+ x n+1 w n+1 x n+1 x n + M N 0 λ i,n+1 λ i,n + M M 1 r k,n+1 r k,n + γ n+1 γ n ( ) + γ n β n+ +1 ( t n + W n t n ) + λ n+1 α n+1 λ n α n v n + λ n+1 λ n f(v n ) + θ n+1 T n+1 t n t n + β n+ x n+1 w n+1 x n+1 x n + M ( N M 2 λ i,n+1 λ i,n + r k,n+1 r k,n + γ n+1 γ n + β n+ + λ n+1 α n+1 λ n α n + λ n+1 λ n + θ n+1), where sup n 1 { x n+1 w n+1 + tn + Wntn 1 d + v n + f(v n ) + T n+1 t n t n + M 0 + M 1 } M 2 (3.13) some M 2 > 0. On the other hand, we define z n := V x n + (1 )Sx n for all n 1. Then it is known that x n+1 = ɛ n γz n + (I ɛ n µf )y n for all n 1. Simple calculations show that z n+1 z n = (+1 )(V x n Sx n ) + +1 (V x n+1 V x n ) + (1 +1 )(Sx n+1 Sx n ), x n+2 x n+1 = (ɛ n+1 ɛ n )(γz n µf y n ) + ɛ n+1 γ(z n+1 z n ) + (I λ n+1 µf )y n+1 (I λ n+1 µf )y n. for

L. C. Ceng, Y. C. Liou, D. R. Sahu, J. Nonlinear Sci. Appl. 9 (2016), 4148 4166 4159 Since V is a ρ-contraction with coefficient ρ [0, 1) and S is a nonexpansive mapping, we conclude that z n+1 z n +1 V x n Sx n + +1 V x n+1 V x n + (1 +1 ) Sx n+1 Sx n +1 V x n Sx n + +1 ρ x n+1 x n + (1 +1 ) x n+1 x n which together with (3.13) and 0 γ < τ, implies that = (1 +1 (1 ρ)) x n+1 x n + +1 V x n Sx n, x n+2 x n+1 ɛ n+1 ɛ n γz n µf y n + ɛ n+1 γ z n+1 z n + (I ɛ n+1 µf )y n+1 (I ɛ n+1 µf )y n ɛ n+1 ɛ n γz n µf y n + ɛ n+1 γ [ (1 +1 (1 ρ)) x n+1 x n + +1 V x n Sx n ] + (1 ɛ n+1 τ) [ x n+1 x n + M ( N M 2 λ i,n+1 λ i,n + r k,n+1 r k,n + γ n+1 γ n + β n+ + λ n+1 α n+1 λ n α n + λ n+1 λ n + θ n+1)] (1 ɛ n+1 (τ γ)) x n+1 x n + M { N 3 λ i,n+1 λ i,n + M r k,n+1 r k,n + ɛ n+1 ɛ n + +1 + β n+ + γ n+1 γ n + λ n+1 α n+1 λ n α n + λ n+1 λ n + θ n+1}, { where sup n 0 γz n µf y n + V x n Sx n + M } 2 M 3 for some M 3 > 0. Consequently, x n+1 x n (1 ɛ n (τ γ)) x n x n 1 + M 3 { N λ i,n λ i,n 1 + M r k,n r k,n 1 + ɛ n ɛ n 1 + 1 + β n β n 1 + γ n γ n 1 + λ n+1α n+1 λ n α n + λ n+1 λ n + θn+1 } (1 ɛ n (τ γ)) x n x n 1 M 4 { 1 + ɛ n (τ γ) 1 1 + 1 τ γ ɛ n 1 + M N λ i,n λ i,n 1 ɛ n r k,n r k,n 1 ɛ n + 1 1 ɛ n 1 ɛ n + 1 ɛ n 1 1 + β n β n 1 ɛ n + γ n γ n 1 ɛ n + λ nα n λ n 1 α n 1 + λ n λ n 1 + θn }, ɛ n ɛ n ɛ n { where sup n 1 x n x n 1 + M } 3 M 4 for some M 4 > 0. n=0 ɛ n(τ γ) = and lim n M 4 { 1 1 1 + τ γ ɛ n 1 N λ i,n λ i,n 1 ɛ n + M (3.14) From conditions (C1)-(C5) it follows that r k,n r k,n 1 ɛ n + 1 1 ɛ n 1 ɛ n

L. C. Ceng, Y. C. Liou, D. R. Sahu, J. Nonlinear Sci. Appl. 9 (2016), 4148 4166 4160 + 1 1 1 + β n β n 1 + γ n γ n 1 + λ nα n λ n 1 α n 1 ɛ n ɛ n ɛ n ɛ n + λ n λ n 1 + θn } = 0. ɛ n ɛ n Thus, utilizing Lemma 2.17, we immediately conclude that lim n x n+1 x n / = 0. So, from 0 it follows that lim n x n+1 x n = 0. Step 3. We prove that lim n x n u n = 0, lim n x n v n = 0, lim n v n t n = 0 and lim n t n W t n = 0. Indeed, by Lemmas 2.8 and 2.9 (b), from (3.1), (3.4)-(3.5) and 0 γ < τ one has and hence y n p 2 = β n (x n p) + ( )( γ nt n + σ n W n t n p) 2 = β n x n p 2 + γ n t n p 2 + σ n W n t n p 2 γ nσ n t n W n t n 2 β n y n x n 2 β n x n p 2 + ( )( x n p + λ n α n p ) 2 γ nσ n t n W n t n 2 β n y n x n 2 ( x n p + λ n α n p ) 2 γ nσ n t n W n t n 2 β n y n x n 2, x n+1 p 2 = ɛ n [ (γv x n γv p) + (1 )(γsx n γsp)] + (I ɛ n µf )y n (I ɛ n µf )p + ɛ n [ (γv p µf p) + (1 )(γsp µf p)] 2 [ ɛ n (1 (1 ρ))γ x n p + (1 ɛ n τ) y n p ] 2 + 2ɛ n (γv p µf p), x n+1 p + 2ɛ n (1 ) (γsp µf p), x n+1 p ɛ nγ 2 x n p 2 + (1 ɛ n τ) [ ( x n p + α n b p ) 2 γ nσ n t n W n t n 2 τ β n y n x n 2] + 2ɛ n (γv p µf p), x n+1 p + 2ɛ n (1 ) (γsp µf p), x n+1 p (1 ɛ n τ 2 γ 2 τ )( x n p + α n b p ) 2 γ nσ n (1 ɛ n τ) t n W n t n 2 β n(1 ɛ n τ) y n x n 2 + 2ɛ n (γv p µf p), x n+1 p + 2ɛ n (1 ) (γsp µf p), x n+1 p, which together with {β n } [c, d] (0, 1), immediately yields γ n σ n (1 ɛ n τ) t n W n t n 2 + c(1 ɛ nτ) y n x n 2 1 c 1 c ( x n x n+1 + α n b p )( x n p + x n+1 p + α n b p ) + 2ɛ n γv p µf p x n+1 p + 2ɛ n γsp µf p x n+1 p. (3.15) (3.16) In terms of (C6), we find that lim inf n γ n σ n 0. Since ɛ n 0, 0, α n 0, x n+1 x n 0 and {x n } is bounded, we have lim n t n W n t n = 0 and lim n y n x n = 0. (3.17)

L. C. Ceng, Y. C. Liou, D. R. Sahu, J. Nonlinear Sci. Appl. 9 (2016), 4148 4166 4161 Noting that y n x n =γ n (t n x n ) + σ n (W n t n x n ) =( )(t n x n ) + σ n (W n t n t n ), we conclude from (3.17) and {β n } [c, d] (0, 1) that as n, Observe that (1 d) t n x n ( )(t n x n ) y n x n + W n t n t n 0. (3.18) k nx n p 2 x n p 2 + r k,n (r k,n 2µ k ) A k k 1 n x n A k p 2, (3.19) Λ i nu n p 2 x n p 2 + λ i,n (λ i,n 2η i ) B i Λ i 1 n u n B i p 2 (3.20) for i {1, 2,..., N} and k {1, 2,..., M}. Combining (3.5), (3.15), (3.19) and (3.20), we get y n p 2 β n x n p 2 + ( ) t n p 2 β n x n p 2 + ( )( v n p + λ n α n p ) 2 β n x n p 2 + ( ) Λ i nu n p 2 + α n b p (2 v n p + α n b p ) β n x n p 2 + ( ) [ u n p 2 + λ i,n (λ i,n 2η i ) B i Λn i 1 u n B i p 2] + α n b p (2 v n p + α n b p ) x n p 2 + ( ) [ r k,n (r k,n 2µ k ) A k k 1 n x n A k p 2 + λ i,n (λ i,n 2η i ) B i Λ i 1 n u n B i p 2] + α n b p (2 v n p + α n b p ), (3.21) which immediately leads to ( ) [ r k,n (2µ k r k,n ) A k k 1 n x n A k p 2 + λ i,n (2η i λ i,n ) B i Λ i 1 n u n B i p 2] x n y n ( x n p + y n p ) + α n b p (2 v n p + α n b p ). Since {β n } [c, d] (0, 1), {λ i,n } [a i, b i ] (0, 2η i ), {r k,n } [c k, d k ] (0, 2µ k ), i {1, 2,..., N}, k {1, 2,..., M} and {v n }, {x n }, {y n } are bounded sequences, from (3.17) and α n 0 we have lim A k k 1 n n x n A k p = 0 and lim B iλ i 1 n n u n B i p = 0 (3.22) for all k {1, 2,..., M} and i {1, 2,..., N}. Furthermore, by Proposition 2.4 (ii) and Lemma 2.9 (a) we have k nx n p 2 (I r k,n A k ) k 1 n x n (I r k,n A k )p, k nx n p which implies that 1 2 ( k 1 n k nx n p 2 x n p 2 k 1 n x n p 2 + k nx n p 2 k 1 By Lemma 2.9 (a) and Lemma 2.20, we obtain n x n k nx n r k,n (A k k 1 n x n A k p) 2 ), x n k nx n 2 + 2r k,n k 1 n x n k nx n A k k 1 n x n A k p. (3.23) Λ i nu n p 2 (I λ i,n B i )Λ i 1 n u n (I λ i,n B i )p, Λ i nu n p which immediately leads to 1 2 ( x n p 2 + Λ i nu n p 2 Λ i 1 n u n Λ i nu n λ i,n (B i Λn i 1 u n B i p) 2 ), Λ i nu n p 2 x n p 2 Λ i 1 n u n Λ i nu n 2 + 2λ i,n Λn i 1 u n Λ i nu n B i Λn i 1 u n B i p. (3.24)

L. C. Ceng, Y. C. Liou, D. R. Sahu, J. Nonlinear Sci. Appl. 9 (2016), 4148 4166 4162 Combining (3.21) and (3.24) we conclude that y n p 2 x n p 2 ( ) Λ i 1 n u n Λ i nu n 2 + 2λ i,n Λn i 1 u n Λ i nu n B i Λn i 1 u n B i p + α n b p (2 v n p + α n b p ), which yields ( ) Λ i 1 n u n Λ i nu n 2 x n y n ( x n p + y n p ) + 2λ i,n Λ i 1 n + α n b p (2 v n p + α n b p ). u n Λ i nu n B i Λ i 1 n u n B i p Since {β n } [c, d] (0, 1), {λ i,n } [a i, b i ] (0, 2η i ), i = 1, 2,..., N, and {u n }, {x n } and {y n } are bounded sequences, we deduce from (3.17), (3.22) and α n 0 that Also, combining (3.3), (3.21) and (3.23) we deduce that lim n Λi 1 n u n Λ i nu n = 0, i {1, 2,..., N}. (3.25) y n p 2 x n p 2 ( ) k 1 n x n k nx n 2 + 2r k,n k 1 n x n k nx n A k k 1 n x n A k p + α n b p (2 v n p + α n b p ), which yields ( ) k 1 n x n k nx n 2 x n y n ( x n p + y n p ) + 2r k,n n k 1 x n k nx n A k k 1 n x n A k p + α n b p (2 v n p + α n b p ). Since {β n } [c, d] (0, 1), {r k,n } [c k, d k ] (0, 2µ k ) for k = 1, 2,..., M, and {v n }, {x n }, {y n } are bounded sequences, we deduce from (3.17), (3.22) and α n 0 that Hence from (3.25) and (3.26) we obtain that as n, lim n k 1 n x n k nx n = 0, k {1, 2,..., M}. (3.26) x n u n 0 nx n 1 nx n + 1 nx n 2 nx n + + M 1 n x n M n x n 0, (3.27) u n v n Λ 0 nu n Λ 1 nu n + Λ 1 nu n Λ 2 nu n + + Λ N 1 n u n Λ N n u n 0. (3.28) Thus, from (3.27) and (3.28) we obtain which together with (3.18), attains x n v n x n u n + u n v n 0 as n, (3.29) t n v n t n x n + x n v n 0 as n. (3.30) In addition, it is clear that t n W t n t n W n t n + W n t n W t n. Thus, we conclude from Remark 2.12, (3.17) and the boundedness of {t n } that lim t n W t n = 0. (3.31) n Step 4. We prove that ω w (x n ) Ω. Indeed, since H is reflexive and {x n } is bounded, there exists at least a weak convergence subsequence of {x n }. Hence it is known that ω w (x n ). Now, take an arbitrary w ω w (x n ). Then there exists a subsequence {x ni } of {x n } such that x ni w. From (3.18), (3.25)-(3.27)

L. C. Ceng, Y. C. Liou, D. R. Sahu, J. Nonlinear Sci. Appl. 9 (2016), 4148 4166 4163 and (3.29) we have that u ni w, v ni w, t ni w, Λ m n i u ni w and k n i x ni w, where m {1, 2,..., N} and k {1, 2,..., M}. Utilizing Lemma 2.14, we deduce from t ni w and (3.31) that w Fix(W ) = n=1 Fix(T n) (due to Lemma 2.13). Next, we claim that w N m=1 I(B m, R m ) M GMEP(Θ k, ϕ k, A k ). As a matter of fact, repeating the same arguments as those of w N m=1 I(B m, R m ) M GMEP(Θ k, ϕ k, A k ) in Step 4 of the proof of [15, Theorem 3.1], we obtain the desired assertion. Thus, w n=1 Fix(T n) M GMEP(Θ k, ϕ k, A k ) N m=1 I(B m, R m ). Furthermore, let us define { f(v) + NC v, if v C, T v =, if v C, where N C v = {u H : v x, u 0, x C}. Then, T is maximal monotone and 0 T v if and only if v VI(C, f); see [29]. By standard argument we can show that w T 1 0 and hence, w VI(C, f) = Γ. Consequently, w Ω. This shows that ω w (x n ) Ω. Step 5. We prove that x n x where {x } = VI(Ω, γs µf ). Indeed, take an arbitrary w ω w (x n ). Then there exists a subsequence {x ni } of {x n } such that x ni w. Utilizing (3.16), we obtain that for all p Ω which implies that x n+1 p 2 ( x n p + α n b p ) 2 + 2ɛ n (γv p µf p), x n+1 p + 2ɛ n (1 ) (γsp µf p), x n+1 p, (µf γs)p, x n p (µf γs)p, x n x n+1 + (µf γs)p, x n+1 p (µf γs)p x n x n+1 + ( x n p + α n b p ) 2 x n+1 p 2 2ɛ n (1 ) + 1 (γv µf )p, x n+1 p. (3.32) δ Since lim sup n α n ɛn <, n δn 0 and xn x n+1 0 (due to Step 2), from (3.32) we get (µf γs)p, w p lim sup (µf γs)p, x n p 0, p Ω. Since µf γs is (µη γ)-strongly monotone n and (µκ + γ)-lipschitz continuous, by Minty s Lemma [24] we know that w VI(Ω, µf γs). Noticing {x } = VI(Ω, µf γs), we have w = x. Thus, ω w (x n ) = {x }; that is, x n x. Finally, we prove that lim n x n x = 0. Indeed, utilizing (3.16) with p = x, we get x n+1 x 2 τ 2 γ 2 (1 ɛ n )( x n x + α n b x ) 2 + 2ɛ n (γv µf )x x n+1 x τ + 2ɛ n (1 ) (γs µf )x, x n+1 x τ 2 γ 2 (1 ɛ n ) x n x 2 τ 2 γ 2 2τ + ɛ n τ τ τ 2 γ 2 [ (γv µf )x x n+1 x + (1 ) (γs µf )x ), x n+1 x ] + α n b x (2 x n x + α n b x ). Note that n=1 α nb x (2 x n x + α n b x ) <, lim n n=1 ɛ n τ 2 γ 2 τ =, and 2τ τ 2 γ 2 [ δn (γv µf )x x n+1 x + (1 ) (γs µf )x, x n+1 x ] = 0. So, applying Lemma 2.17 we derive lim n x n x = 0. Remark 3.2. The scheme (3.9) in [12, Theorem 3.2] is extended to develop our scheme (3.1) for the SFP (1.2) with constraints of finite many GMEPs, finite many variational inclusions and the fixed point problem

L. C. Ceng, Y. C. Liou, D. R. Sahu, J. Nonlinear Sci. Appl. 9 (2016), 4148 4166 4164 of nonexpansive mappings {T n } n=1. Under the lack of the assumptions similar to those in [40, Theorem 3.2], e.g., {x n } is bounded, Fix(T ) intc and x T x kdist(x, Fix(T )), x C for some k > 0, the sequence {x n } generated by (3.1) converges strongly to a point x Ω, which is a unique solution of the HVIP (over the fixed point set of nonexpansive mappings {T n } n=1 ), that is, (µf γs)x, p x 0, p Ω. Recalling the argument process of Theorem 3.1, we can also derive the following Theorem 3.3. In Theorem 3.1, if the conditions (C1)-(C6) are replaced by the following ones (C1)-(C5): (C1) lim n = 0, lim n ɛ n = 0 and n=1 ɛ n = ; (C2) n=2 ɛ n ɛ n 1 < or lim n ɛ n 1 ɛ n = 1; (C3) n=2 ( N λ i,n λ i,n 1 + M r k,n r k,n 1 ) < or lim ( N M λ i,n λ i,n 1 + r k,n r k,n 1 )/ɛ n = 0; n (C4) n=2 ( 1 + β n β n 1 + γ n γ n 1 + λ n λ n 1 ) < or lim ( 1 + β n β n 1 + γ n γ n 1 + λ n λ n 1 )/ɛ n = 0; n (C5) lim inf n σ n > 0 and 0 < lim inf n β n lim sup n (β n + σ n ) < 1. Then {x n } converges strongly to a solution x of the SFP provided x n+1 x n + α n = o(ɛ n ), which is a unique solution in Ω to the HVIP: (µf γs)x, p x 0, p Ω. Acknowledgment This research was partially supported by the Program for Outstanding Academic Leaders in Shanghai City (15XD1503100), Innovation Program of Shanghai Municipal Education Commission (15ZZ068) and Ph.D. Program Foundation of Ministry of Education of China (20123127110002). This research was partially supported by a grant from NSC. References [1] A. E. Al-Mazrooei, B. A. Bin Dehaish, A. Latif, J. C. Yao, On general system of variational inequalities in Banach spaces, J. Nonlinear Convex Anal., 16 (2015), 639 658. 1 [2] A. S. M. Alofi, A. Latif, A. E. Al-Mazrooei, J. C. Yao, Composite viscosity iterative methods for general systems of variational inequalities and fixed point problem in Hilbert spaces, J. Nonlinear Convex Anal., 17 (2016), 669 682. 1 [3] V. Barbu, Nonlinear semigroups and differential equations in Banach spaces, Noordhoff International Publishing, Leiden, (1976). 2.17 [4] A. Bnouhachem, Q. H. Ansari, J. C. Yao, An iterative algorithm for hierarchical fixed point problems for a finite family of nonexpansive mappings, Fixed Point Theory Appl., 2015 (2015), 13 pages. 1 [5] A. Bnouhachem, Q. H. Ansari, J. C. Yao, Strong convergence algorithm for hierarchical fixed point problems of a finite family of nonexpansive mappings, Fixed Point Theory, 17 (2016), 47 62. 1 [6] C. Byrne, A unified treatment of some iterative algorithms in signal processing and image reconstruction, Inverse Problems, 20 (2004), 103 120. 2.6, 2.7 [7] L. C. Ceng, S. Al-Homidan, Algorithms of common solutions for generalized mixed equilibria, variational inclusions, and constrained convex minimization, Abstr. Appl. Anal., 2014 (2014), 25 pages. 1 [8] L. C. Ceng, Q. H. Ansari, A. Petruşel, J. C. Yao, Approximation methods for triple hierarchical variational inequalities (II), Fixed Point Theory, 16 (2015), 237 260. 1 [9] L. C. Ceng, Q. H. Ansari, S. Schaible, Hybrid extragradient-like methods for generalized mixed equilibrium problems, systems of generalized equilibrium problems and optimization problems, J. Global Optim., 53 (2012), 69 96. 1, 1

L. C. Ceng, Y. C. Liou, D. R. Sahu, J. Nonlinear Sci. Appl. 9 (2016), 4148 4166 4165 [10] L. C. Ceng, Q. H. Ansari, M. M. Wong, J. C. Yao, Mann type hybrid extragradient method for variational inequalities, variational inclusions and fixed point problems, Fixed Point Theory, 13 (2012), 403 422. 2.20, 2.22 [11] L. C. Ceng, Q. H. Ansari, J. C. Yao, An extragradient method for solving split feasibility and fixed point problems, Comput. Math. Appl., 64 (2012), 633 642. 1 [12] L. C. Ceng, Q. H. Ansari, J. C. Yao, Relaxed extragradient methods for finding minimum-norm solutions of the split feasibility problem, Nonlinear Anal., 75 (2012), 2116 2125. 1, 1, 3.2 [13] L. C. Ceng, S. M. Guu, J. C. Yao, Hybrid iterative method for finding common solutions of generalized mixed equilibrium and fixed point problems, Fixed Point Theory Appl., 2012 (2012), 19 pages. 1 [14] L. C. Ceng, H. Y. Hu, M. M. Wong, Strong and weak convergence theorems for generalized mixed equilibrium problem with perturbation and fixed pointed problem of infinitely many nonexpansive mappings, Taiwanese J. Math., 15 (2011), 1341 1367. [15] L. C. Ceng, A. Latif, Q. H. Ansari, J. C. Yao, Hybrid extragradient method for hierarchical variational inequalities, Fixed Point Theory Appl., 2014 (2014), 35 pages. 1, 1, 3 [16] L. C. Ceng, A. Latif, J. C. Yao, On solutions of system of variational inequalities and fixed point problems in Banach spaces, Fixed Point Theory Appl., 2013 (2013), 34 pages. 1 [17] L. C. Ceng, M. M. Wong, A. Petruşel, J. C. Yao, Relaxed implicit extragradient-like methods for finding minimumnorm solutions of the split feasibility problem, Fixed Point Theory, 14 (2013), 327 344. 1 [18] L. C. Ceng, J. C. Yao, A hybrid iterative scheme for mixed equilibrium problems and fixed point problems, J. Comput. Appl. Math., 214 (2008), 186 201. 2.4 [19] L. C. Ceng, J. C. Yao, A relaxed extragradient-like method for a generalized mixed equilibrium problem, a general system of generalized equilibria and a fixed point problem, Nonlinear Anal., 72 (2010), 1922 1937. 1, 1 [20] L. C. Ceng, J. C. Yao, Relaxed and hybrid viscosity methods for general system of variational inequalities with split feasibility problem constraint, Fixed Point Theory App., 2013 (2013), 50 pages. 1 [21] L. C. Ceng, J. C. Yao, On the triple hierarchical variational inequalities with constraints of mixed equilibria, variational inclusions and systems of generalized equilibria, Tamkang J. Math., 45 (2014), 297 334. 1 [22] Y. Censor, T. Elfving, A multiprojection algorithm using Bregman projections in a product space, Numer. Algorithms, 8 (1994) 221 239. 1 [23] P. L. Combettes, Solving monotone inclusions via compositions of nonexpansive averaged operators, Optimization, 53 (2004), 475 504. 2.7 [24] K. Goebel, W. A. Kirk, Topics in metric fixed point theory, Cambridge University Press, Cambridge, England, (1990). 2.2, 2.14, 3 [25] N. J. Huang, A new completely general class of variational inclusions with noncompact valued mappings, Comput. Math. Appl., 35 (1998), 9 14. 1, 2 [26] G. M. Korpelevich, The extragradient method for finding saddle points and other problems, Matecon, 12 (1976), 747 756. 1 [27] N. Nadezhkina, W. Takahashi, Weak convergence theorem by an extragradient method for nonexpansive mappings and monotone mappings, J. Optim. Theory Appl., 128 (2006), 191 201. 1, 1 [28] J. W. Peng, J. C. Yao, A new hybrid-extragradient method for generalized mixed equilibrium problems, fixed point problems and variational inequality problems, Taiwanese J. Math., 12 (2008), 1401 1432. 1, 1, 1 [29] R. T. Rockafellar, Monotone operators and the proximal point algorithms, SIAM J. Control Optimization, 14 (1976), 877 898. 2, 3 [30] K. Shimoji, W. Takahashi, Strong convergence to common fixed points of infinite nonexpansive mappings and applications, Taiwanese J. Math., 5 (2001), 387 404. 2.10, 2.13 [31] W. Takahashi, Introduction to nonlinear and convex analysis, Yokohama Publishers, Yokohama, (2009). 2.8, 2.9 [32] W. Takahashi, H. K. Xu, J. C. Yao, Iterative methods for generalized split feasibility problems in Hilbert spaces, Set-Valued Var. Anal., 23 (2015), 205 221. 1 [33] X. M. Wang, C. Li, J. C. Yao, Subgradient projection algorithms for convex feasibility on Riemannian manifolds with lower bounded curvatures, J. Optim. Theory Appl., 164 (2015), 202 217. 1 [34] H. K. Xu, Iterative algorithms for nonlinear operators, J. London Math. Soc., 66 (2002), 240 256. 2.16 [35] H. K. Xu, Iterative methods for the split feasibility problem in infinite-dimensional Hilbert spaces, Inverse Problems, 26 (2010), 17 pages. 1, 3 [36] H. K. Xu, T. H. Kim, Convergence of hybrid steepest-descent methods for variational inequalities, J. Optim. Theory. Appl., 119 (2003), 185 201. 2.15 [37] Y. Yao, R. Chen, H. K. Xu, Schemes for finding minimum-norm solutions of variational inequalities, Nonlinear Anal., 72 (2010), 3447 3456. 1 [38] Y. Yao, W. Jigang, Y. C. Liou, Regularized methods for the split feasibility problem, Abstr. Appl. Anal., 2012 (2012), 15 pages. [39] Y. Yao, Y. C. Liou, S. M. Kang, Two-step projection methods for a system of variational inequality problems in Banach spaces, J. Global Optim., 55 (2013), 801 811. 1 [40] Y. Yao, Y. C. Liou, G. Marino, Two-step iterative algorithms for hierarchical fixed point problems and variational inequality problems, J. Appl. Math. Comput., 31 (2009), 433 445. 1, 1, 3.2 [41] Y. Yao, Y. C. Liou, J. C. Yao, Convergence theorem for equilibrium problems and fixed point problems of infinite

L. C. Ceng, Y. C. Liou, D. R. Sahu, J. Nonlinear Sci. Appl. 9 (2016), 4148 4166 4166 family of nonexpansive mappings, Fixed Point Theory Appl., 2007 (2007), 12 pages. 2.11, 2.12 [42] Y. Yao, Y. C. Liou, J. C. Yao, Split common fixed point problem for two quasi-pseudo-contractive operators and its algorithm construction, Fixed Point Theory Appl., 2015 (2015), 19 pages. 1 [43] Y. Yao, M. A. Noor, Y. C. Liou, Strong convergence of a modified extragradient method to the minimum-norm solution of variational inequalities, Abstr. Appl. Anal., 2012 (2012), 9 pages. [44] Y. Yao, M. Postolache, Y. C. Liou, Strong convergence of a self-adaptive method for the split feasibility problem, Fixed Point Theory Appl., 2013 (2013), 12 pages. 1 [45] L. C. Zeng, S. M. Guu, J. C. Yao, Characterization of H-monotone operators with applications to variational inclusions, Comput. Math. Appl., 50 (2005), 329 337. 1, 2, 2.21