AN INEXACT HYBRID GENERALIZED PROXIMAL POINT ALGORITHM AND SOME NEW RESULTS ON THE THEORY OF BREGMAN FUNCTIONS. M. V. Solodov and B. F.

Similar documents
AN INEXACT HYBRID GENERALIZED PROXIMAL POINT ALGORITHM AND SOME NEW RESULTS ON THE THEORY OF BREGMAN FUNCTIONS. May 14, 1998 (Revised March 12, 1999)

AN INEXACT HYBRIDGENERALIZEDPROXIMAL POINT ALGORITHM ANDSOME NEW RESULTS ON THE THEORY OF BREGMAN FUNCTIONS

Journal of Convex Analysis (accepted for publication) A HYBRID PROJECTION PROXIMAL POINT ALGORITHM. M. V. Solodov and B. F.

Forcing strong convergence of proximal point iterations in a Hilbert space

1 Introduction We consider the problem nd x 2 H such that 0 2 T (x); (1.1) where H is a real Hilbert space, and T () is a maximal monotone operator (o

Merit functions and error bounds for generalized variational inequalities

A convergence result for an Outer Approximation Scheme

Error bounds for proximal point subproblems and associated inexact proximal point algorithms

ENLARGEMENT OF MONOTONE OPERATORS WITH APPLICATIONS TO VARIATIONAL INEQUALITIES. Abstract

Some Inexact Hybrid Proximal Augmented Lagrangian Algorithms

On the complexity of the hybrid proximal extragradient method for the iterates and the ergodic mean

An Accelerated Hybrid Proximal Extragradient Method for Convex Optimization and its Implications to Second-Order Methods

Spectral gradient projection method for solving nonlinear monotone equations

FIXED POINTS IN THE FAMILY OF CONVEX REPRESENTATIONS OF A MAXIMAL MONOTONE OPERATOR

On the convergence properties of the projected gradient method for convex optimization

Journal of Convex Analysis Vol. 14, No. 2, March 2007 AN EXPLICIT DESCENT METHOD FOR BILEVEL CONVEX OPTIMIZATION. Mikhail Solodov. September 12, 2005

A Unified Approach to Proximal Algorithms using Bregman Distance

Maximal monotone operators, convex functions and a special family of enlargements.

Convergence rate of inexact proximal point methods with relative error criteria for convex optimization

1. Introduction. We consider the classical variational inequality problem [1, 3, 7] VI(F, C), which is to find a point x such that

Pacific Journal of Optimization (Vol. 2, No. 3, September 2006) ABSTRACT

Maximal Monotone Operators with a Unique Extension to the Bidual

A proximal-like algorithm for a class of nonconvex programming

An inexact subgradient algorithm for Equilibrium Problems

On Total Convexity, Bregman Projections and Stability in Banach Spaces

On the Convergence and O(1/N) Complexity of a Class of Nonlinear Proximal Point Algorithms for Monotonic Variational Inequalities

An Infeasible Interior Proximal Method for Convex Programming Problems with Linear Constraints 1

Finite Convergence for Feasible Solution Sequence of Variational Inequality Problems

A projection-type method for generalized variational inequalities with dual solutions

Stationary Points of Bound Constrained Minimization Reformulations of Complementarity Problems1,2

Convergence Analysis of Perturbed Feasible Descent Methods 1

On well definedness of the Central Path

A strongly convergent hybrid proximal method in Banach spaces

WEAK CONVERGENCE OF RESOLVENTS OF MAXIMAL MONOTONE OPERATORS AND MOSCO CONVERGENCE

Optimization and Optimal Control in Banach Spaces

On convergence rate of the Douglas-Rachford operator splitting method

A Strongly Convergent Method for Nonsmooth Convex Minimization in Hilbert Spaces

arxiv: v1 [math.na] 25 Sep 2012

Near Equality, Near Convexity, Sums of Maximally Monotone Operators, and Averages of Firmly Nonexpansive Mappings

A doubly stabilized bundle method for nonsmooth convex optimization

Fixed points in the family of convex representations of a maximal monotone operator

arxiv: v1 [math.fa] 16 Jun 2011

Monotone operators and bigger conjugate functions

A Dykstra-like algorithm for two monotone operators

Iterative Convex Optimization Algorithms; Part One: Using the Baillon Haddad Theorem

A double projection method for solving variational inequalities without monotonicity

Subdifferential representation of convex functions: refinements and applications

arxiv: v3 [math.oc] 18 Apr 2012

Existence and Approximation of Fixed Points of. Bregman Nonexpansive Operators. Banach Spaces

ε-enlargements of maximal monotone operators: theory and applications

Convergence rate estimates for the gradient differential inclusion

Complexity of the relaxed Peaceman-Rachford splitting method for the sum of two maximal strongly monotone operators

On the iterate convergence of descent methods for convex optimization

Convex Optimization Notes

INERTIAL ACCELERATED ALGORITHMS FOR SOLVING SPLIT FEASIBILITY PROBLEMS. Yazheng Dang. Jie Sun. Honglei Xu

An Accelerated Hybrid Proximal Extragradient Method for Convex Optimization and its Implications to Second-Order Methods

Kantorovich s Majorants Principle for Newton s Method

1. Introduction The nonlinear complementarity problem (NCP) is to nd a point x 2 IR n such that hx; F (x)i = ; x 2 IR n + ; F (x) 2 IRn + ; where F is

On Slater s condition and finite convergence of the Douglas Rachford algorithm for solving convex feasibility problems in Euclidean spaces

On the order of the operators in the Douglas Rachford algorithm

PARALLEL SUBGRADIENT METHOD FOR NONSMOOTH CONVEX OPTIMIZATION WITH A SIMPLE CONSTRAINT

M. Marques Alves Marina Geremia. November 30, 2017

References 1. Aleyner, A., & Reich, S. (2008). Block-iterative algorithms for solving convex feasibility problems in Hilbert and Banach spaces. Journa

Extensions of Korpelevich s Extragradient Method for the Variational Inequality Problem in Euclidean Space

Iteration-complexity of a Rockafellar s proximal method of multipliers for convex programming based on second-order approximations

Convex Feasibility Problems

Proximal Point Methods and Augmented Lagrangian Methods for Equilibrium Problems

Variational Inequalities. Anna Nagurney Isenberg School of Management University of Massachusetts Amherst, MA 01003

Heinz H. Bauschke and Walaa M. Moursi. December 1, Abstract

Brøndsted-Rockafellar property of subdifferentials of prox-bounded functions. Marc Lassonde Université des Antilles et de la Guyane

A Geometric Framework for Nonconvex Optimization Duality using Augmented Lagrangian Functions

ON A CLASS OF NONSMOOTH COMPOSITE FUNCTIONS

Sequential Unconstrained Minimization: A Survey

On a result of Pazy concerning the asymptotic behaviour of nonexpansive mappings

On proximal-like methods for equilibrium programming

The effect of calmness on the solution set of systems of nonlinear equations

A Unified Analysis of Nonconvex Optimization Duality and Penalty Methods with General Augmenting Functions

Iteration-Complexity of a Newton Proximal Extragradient Method for Monotone Variational Inequalities and Inclusion Problems

c 2013 Society for Industrial and Applied Mathematics

A derivative-free nonmonotone line search and its application to the spectral residual method

MOSCO STABILITY OF PROXIMAL MAPPINGS IN REFLEXIVE BANACH SPACES

3.10 Lagrangian relaxation

Downloaded 12/13/16 to Redistribution subject to SIAM license or copyright; see

Strong Convergence Theorem by a Hybrid Extragradient-like Approximation Method for Variational Inequalities and Fixed Point Problems

BREGMAN DISTANCES, TOTALLY

c 1998 Society for Industrial and Applied Mathematics

A characterization of essentially strictly convex functions on reflexive Banach spaces

The Relation Between Pseudonormality and Quasiregularity in Constrained Optimization 1

AN OUTER APPROXIMATION METHOD FOR THE VARIATIONAL INEQUALITY PROBLEM

ON A HYBRID PROXIMAL POINT ALGORITHM IN BANACH SPACES

A globally convergent Levenberg Marquardt method for equality-constrained optimization

Maximal monotone operators are selfdual vector fields and vice-versa

Maximal Monotone Inclusions and Fitzpatrick Functions

On Penalty and Gap Function Methods for Bilevel Equilibrium Problems

A NEW ITERATIVE METHOD FOR THE SPLIT COMMON FIXED POINT PROBLEM IN HILBERT SPACES. Fenghui Wang

A WEAK-TO-STRONGCONVERGENCE PRINCIPLE FOR FEJÉR-MONOTONE METHODS IN HILBERT SPACES

A GENERALIZATION OF THE REGULARIZATION PROXIMAL POINT METHOD

On the Weak Convergence of the Extragradient Method for Solving Pseudo-Monotone Variational Inequalities

Splitting methods for decomposing separable convex programs

Transcription:

AN INEXACT HYBRID GENERALIZED PROXIMAL POINT ALGORITHM AND SOME NEW RESULTS ON THE THEORY OF BREGMAN FUNCTIONS M. V. Solodov and B. F. Svaiter May 14, 1998 (Revised July 8, 1999) ABSTRACT We present a new Bregman-function-based algorithm which is a modification of the generalized proximal point method for solving the variational inequality problem with a maximal monotone operator. The principal advantage of the presented algorithm is that it allows a more constructive error tolerance criterion in solving the proximal point subproblems. Furthermore, we eliminate the assumption of pseudomonotonicity which was, until now, standard in proving convergence for paramonotone operators. Thus we obtain a convergence result which is new even for exact generalized proximal point methods. Finally, we present some new results on the theory of Bregman functions. For example, we show that the standard assumption of convergence consistency is a consequence of the other properties of Bregman functions, and is therefore superfluous. Key words. Proximal point method, Bregman function, maximal monotone operator, variational inequality. AMS subject classifications. 90C25, 90C30. Research of the first author is supported by CNPq Grant 300734/95-6 and by PRONEX Optimization, research of the second author is supported by CNPq Grant 301200/93-9(RN) and by PRONEX Optimization. Instituto de Matemática Pura e Aplicada, Estrada Dona Castorina 110, Jardim Botânico, Rio de Janeiro, RJ 22460-320, Brazil. Email : solodov@impa.br and benar@impa.br.

1 Introduction In this paper, we are concerned with proximal point algorithms for solving the variational inequality problem. Specifically, we consider the methods which are based on Bregman distance regularization. Our objective is twofold. First of all, we develop a hybrid algorithm based on inexact solution of proximal subproblems. The important new feature of the proposed method is that the error tolerance criterion imposed on inexact subproblem solution is constructive and easily implementable for a wide range of applications. Second, we obtain a number of new results on the theory of Bregman functions and on the convergence of related proximal point methods. In particular, we show that one of the standard assumptions on the Bregman function (convergence consistency), as well as one of the standard assumptions on the operator defining the problem (pseudomonotonicity, in the paramonotone operator case), are extraneous. Given an operator T on R n (point-to-set, in general) and a closed convex subset C of R n, the associated variational inequality problem [12], from now on VIP(T, C), is to find a pair x and v such that x C, v T (x ), x x, v 0 x C, (1) where, stands for the usual inner product in R n. The operator T : R n P(R n ), where P(R n ) stands for the family of subsets of R n, is monotone if u v, x y 0 for any x, y R n and any u T (x), v T (y). T is maximal monotone if it is monotone and its graph G(T ) = {(x, u) R n R n u T (x)} is not contained in the graph of any other monotone operator. Throughout this paper we assume that T is maximal monotone. It is well known that VIP(T, C) is closely related to the problem of finding a zero of a maximal monotone operator ˆT : 0 ˆT (z), z R n. (2) Recall that we assume that T is maximal monotone. Therefore, (2) is a particular case of VIP(T, C) for C = R n. On the other hand, define N C as the normal cone operator, that is N C : R n P(R n ) is given by { {v R N C (x) = n v, y x 0 y C} if x C, otherwise. 1

The operator T + N C is monotone and x solves VIP(T, C) (with some v T (x )) if and only if 0 (T + N C )(x ). Additionally, if the relative interiors of C and of the domain of T intersect, then T +N C is maximal monotone [31], and the above inclusion is a particular case of (2), i.e., the problem of finding a zero of a maximal monotone operator. Hence, in this case, VIP(T, C) can be solved using the classical proximal point method for finding a zero of the operator ˆT = T + N C. The proximal point method was introduced by Martinet [26] and further developed by Rockafellar [34]. Some other relevant papers on this method, its applications and modifications, are [27, 33, 3, 29, 25, 17, 18, 15]; see [24] for a survey. The classical proximal point algorithm generates a sequence {x k } by solving a sequence of proximal subproblems. The iterate x k+1 is the solution of 0 c k ˆT (x) + x x k, where c k > 0 is a regularization parameter. For the method to be implementable, it is important to handle approximate solutions of subproblems. This consideration gives rise to the inexact version of the method [34], which can be written as e k+1 + x k c k ˆT (x k+1 ) + x k+1, where e k+1 is the associated error term. To guarantee convergence, it is typically assumed that (see, for example, [34, 8]) e k <. k=0 Note that even though the proximal subproblems are better conditioned than the original problem, structurally they are as difficult to solve. This observation motivates the development of the nonlinear or generalized proximal point method [16, 13, 11, 19, 23, 22, 20, 6]. In the generalized proximal point method, x k+1 is obtained solving the generalized proximal point subproblem 0 c k T (x) + f(x) f(x k ). The function f is the Bregman function [2], namely it is strictly convex, differentiable in the interior of C and its gradient is divergent on the boundary 2

of C (f also has to satisfy some additional technical conditions, which we shall discuss in Section 2). All information about the feasible set C is embedded in the function f, which is both a regularization and a penalization term. Properties of f (discussed in Section 2) ensure that solutions of subproblems belong to the interior of C without any explicit consideration of constraints. The advantage of the generalized proximal point method is that the subproblems are essentially unconstrained. For example, if VIP(T, C) is the classical nonlinear complementarity problem [28], then a reasonable choice of f gives proximal subproblems which are (unconstrained!) systems of nonlinear equations. By contrast, subproblems given by the classical proximal algorithm are themselves nonlinear complementarity problems, which are structurally considerably more difficult to solve than systems of equations. We refer the reader to [6] for a detailed example. As in the case of the classical method, implementable versions of the generalized proximal point algorithm must take into consideration inexact solution of subproblems: In [14], it was established that if e k+1 + f(x k ) c k T (x k+1 ) + f(x k+1 ). k=0 e k < and k=0 e k, x k exists and is finite, (3) then the generated sequence converges to a solution (provided it exists) under basically the same assumptions that are needed for the convergence of the exact method. Other inexact generalized proximal algorithms are [7, 23, 41]. However, the approach of [14] is the simplest and the easiest to use in practical computation (see the discussion in [14]). Still, the error criterion given by (3) is not totally satisfactory. Obviously, there exist many error sequences {e k } that satisfy the first relation in (3), and it is not very clear which e k should be considered acceptable for each specific iteration k. In this sense, criterion (3) is not quite constructive. The second relation in (3) is even somewhat more problematic. In this paper, we present a hybrid generalized proximal-based algorithm which employs a more constructive error criterion than (3). Our method is completely implementable when the gradient of f is easily invertible, which is a common case for many important applications. The inexact solution is used to obtain the new iterate in a way very similar to Bregman generalized 3

projections. When the error is zero, our algorithm coincides with the generalized proximal point method. However, for nonzero error, it is different from the inexact method of [14] described above. Our new method is motivated by [40], where a constructive error tolerance was introduced for the classical proximal point method. This approach has already proved to be very useful in a number of applications [38, 37, 35, 39, 36]. Besides the algorithm, we also present a theoretical result which is new even for exact methods. In particular, we prove convergence of the method for paramonotone operators, without the previously used assumption of pseudomonotonicity (paramonotone operators were introduced in [4, 5], see also [9, 21]; we shall state this definition in Section 3, together with the definition of pseudomonotonicity). It is important to note that the subgradient of a proper closed convex function is paramonotone, but need not be pseudomonotone. Hence, among other things, our result unifies the proof of convergence for paramonotone operators and for minimization. We also remove the condition of convergence consistency which has been used to characterize Bregman functions, proving it to be a consequence of the other properties. This work is organized as follows. In Section 2, we discuss Bregman functions and derive some new results on their properties. In Section 3, the error tolerance to be used is formally defined, the new algorithm is described and the convergence result is stated. Section 4 contains convergence analysis. A few words about our notation are in order. Given a (convex) set A, ri(a) will denote the relative interior, Ā will denote the closure, int(a) will denote the interior, and bdry(a) will denote the boundary of A. For an operator T, Dom(T ) stands for its domain, i.e., all points x R n such that T (x). 2 Bregman Function and Bregman Distance Given a convex function f on R n, finite at x, y R n and differentiable at y, the Bregman distance [2] between x and y, determined by f, is D f (x, y) := f(x) f(y) f(y), x y. Note that, by the convexity of f, the Bregman distance is always nonnegative. We mention here the recent article [1] as one good reference on Bregman functions and their properties. 4

Definition 2.1 Given S, a convex open subset of R n, we say that f : S R is a Bregman function with zone S if 1. f is strictly convex and continuous in S, 2. f is continuously differentiable in S, 3. for any x S and α R, the right partial level set is bounded, L(x, α) = {y S D f (x, y) α} 4. If {y k } is a sequence in S converging to y, then lim D f(y, y k ) = 0. k Some remarks are in order regarding this definition. In addition to the above four items, there is one more standard requirement for Bregman function, namely Convergence Consistency : If {x k } S is bounded, {y k } S converges to y, and lim k D f (x k, y k ) = 0, then {x k } also converges to y. This requirement has been imposed in all previous studies of Bregman functions and related algorithms [10, 11, 13, 19, 9, 14, 1, 22, 20, 6]. In what follows, we shall establish that convergence consistency holds automatically as a consequence of Definition 2.1 (we shall actually prove a stronger result). The original definition of a Bregman function also requires the left partial level sets L (α, y) = {x S D f (x, y) α} to be bounded for any y S. However, it has been already observed that this condition is not needed to prove convergence of proximal methods (e.g., [14]). And it is known that this boundedness condition is extraneous regardless, since it is also a consequence of Definition 2.1 (e.g., see [1]). Indeed, observe that for any y, the level set L (0, y) = {y}, so it is nonempty and bounded. Also Definition 2.1 implies that D f (, y) is a proper closed convex function. Because this function has one level set which in nonempty and bounded, it follows that all of its level sets are bounded (i.e., L (α, y) is bounded for every α) [32, Corollary 8.7.1]. To prove convergence consistency using the properties given in Definition 2.1, we start with the following results. 5

Lemma 2.2 (The Restricted Triangular Inequality) Let f be a convex function satisfying items 1 and 2 of Definition 2.1. If x S, y S and w is a proper convex combination of x and y, i.e., w = (1 θ)x+θy with θ (0, 1), then D f (x, w) + D f (w, y) D f (x, y). Proof. We have that w = y + (1 θ)(x y) with 0 < θ < 1. Clearly, w S. Since f is monotone, f(w), w y f(y), w y. Taking into account that w y = (1 θ)(x y), the latter relation yields Therefore f(w), x y f(y), x y. D f (x, w) + D f (w, y) = [f(x) f(w) f(w), x w ] + [f(w) f(y) f(y), w y ] = f(x) f(y) θ f(w), x y (1 θ) f(y), x y f(x) f(y) θ f(y), x y (1 θ) f(y), x y = f(x) f(y) f(y), x y = D f (x, y). Lemma 2.3 Let f be a convex function satisfying items 1 and 2 of Definition 2.1. If {x k } is a sequence in S converging to x, {y k } is a sequence in S converging to y and y x, then Proof. Define lim inf k D f(x k, y k ) > 0. z k = (1/2)(x k + y k ). Clearly {z k } is a sequence in S converging to z = (1/2)(x + y) S. By the convexity of f, it follows that for all k: f(z k ) f(y k ) + f(y k ), z k y k = f(y k ) + (1/2) f(y k ), x k y k. 6

Therefore f(x k ) + f(y k ) 2 Letting k we obtain f(z k ) f(xk ) + f(y k ) 2 f(x) + f(y) 2 = D f (x k, y k )/2. f(y k ) 1 2 f(yk ), x k y k f( x + y ) 1 lim inf 2 2 D f(x k, y k ). k Using the strict convexity of f and the hypothesis x y, the desired result follows. We are now ready to prove a result which is actually stronger than the property of convergence consistency discussed above. This result will be crucial for strengthening convergence properties of proximal point methods, carried out in this paper. Theorem 2.4 Let f be a convex function satisfying items 1 and 2 of Definition 2.1. If {x k } is a sequence in S, {y k } is a sequence in S, lim D f(x k, y k ) = 0, k and one of the sequences ({x k } or {y k }) converges, then the other also converges to the same limit. Proof. Suppose, by contradiction, that one of the sequences converges and the other does not converge or does not converge to the same limit. Then there exist some ε > 0 and a subsequence of indices {k j } satisfying Suppose first that {y k } converges and x k j y k j > ε. lim k yk = y. Define x j = y k j + ε x k j y k j (x k j y k j ), 7

i.e., x j is a proper convex combination of x k j and y k j. Using Lemma 2.2 we conclude that D f ( x j, y k j ) D f (x k j, y k j ), which implies that lim D f( x j, y k j ) = 0. j Since x j y k j = ε and {y k j } converges, it follows that { x j } is bounded and there exists a subsequence { x j i } converging to some x. Therefore we have the following set of relations lim i D f ( x j i, y k j i ) = 0, lim i y k j i = y, lim i x j i = x, x y = ε > 0, which is in contradiction with Lemma 2.3. If we assume that the sequence {x k } converges, then reversing the roles of {x k } and {y k } in the argument above, we reach a contradiction with Lemma 2.3 in exactly the same manner. It is easy to see that Convergence Consistency is an immediate consequence of Theorem 2.4. We next state a well-known result which is widely used in the analysis of generalized proximal point methods. Lemma 2.5 (Three-Point Lemma)[11] Let f be a Bregman function with zone S as in Definition 2.1. x, z S and y S, it holds that For any D f (y, x) = D f (z, x) + D f (y, z) + f(x) f(z), z y. In the sequel, we shall use the following consequence of Lemma 2.5, which can be obtained by subtracting the three-point inequalities written with y, x, z and s, x, z. Corollary 2.6 (Four-Point Lemma) Let f be a Bregman function with zone S as in Definition 2.1. x, z S and y, s S, it holds that For any D f (s, z) = D f (s, x) + f(x) f(z), s y + D f (y, z) D f (y, x). 8

3 The Inexact Generalized Proximal Point Method We start with some assumptions which are standard in the study and development of Bregman-function-based algorithms. Suppose C, the feasible set of VIP(T, C), has nonempty interior, and we have chosen f, an associated Bregman function with zone int(c). We also assume that int(c) Dom(T ), so that T + N C is maximal monotone [31]. The solution set of VIP(T, C) is X := {s C v s T (s), y C v s, y s 0}. We assume this set to be nonempty, since this is the more interesting case. In principle, following standard analysis, results regarding unboundedness of the iterates can be obtained for the case when no solution exists. Additionally, we need the assumptions which guarantee that proximal subproblem solutions exist and belong to the interior of C. H1 For any x int(c) and c > 0, the generalized proximal subproblem has a solution. 0 ct ( ) + f( ) f(x). H2 For any x int(c), if {y k } is a sequence in int(c) and lim k yk = y bdry(c) then lim f(y k ), y k x = +. k A simple sufficient condition for H1 is that the image of f is the whole space R n (see [6, Proposition 3]). Assumption H2 is called boundary coerciveness and it is the key concept in the context of proximal point methods for constrained problems for the following reason. It is clear from Definition 2.1 that if f is a Bregman function with zone int(c) and P is any open subset of int(c), then f is also a Bregman function with zone P, which means that one cannot recover C from f. Therefore in order to use the Bregman 9

distance D f for penalization purposes, f has to possess an additional property. In particular, f should contain information about C. This is precisely the role of H2 because it implies divergence of f on bdry(c), which makes C defined by f: Dom( f) = int(c). (4) Divergence of f also implies that the proximal subproblems cannot have solutions on the boundary of C. We refer the readers to [9, 6] for further details on boundary coercive Bregman functions. Note also that boundary coerciveness is equivalent to f being essentially smooth on int(c) [1, Theorem 4.5 (i)]. It is further worth to note that if the domain of f is the interior of C, and the image of f is R n, then H1 and H2 hold automatically (see [6, Proposition 3] and [9, Proposition 7]). We are now ready to describe our error tolerance criterion. Take any x int(c) and c > 0, and consider the proximal subproblem 0 ct ( ) + f( ) f(x), (5) which is to find a pair (y, v) satisfying the proximal system The latter is in turn equivalent to v T (y), cv + f(y) f(x) = 0. (6) v T (y), y = ( f) 1 ( f(x) cv). (7) Therefore, an approximate solution of (5) (or (6) or (7)) should satisfy v T (y), y ( f) 1 ( f(x) cv). (8) We next formally define the concept of inexact solutions of (6), taking the approach of (8). Definition 3.1 Let x int(c), c > 0 and σ [0, 1). We say that a pair (y, v) is an inexact solution with tolerance σ of the proximal subproblem (6) if v T (y) and z, the solution of equation satisfies cv + f(z) f(x) = 0, D f (y, z) σ 2 D f (y, x). 10

Note that from (4) (which is a consequence of H2), it follows that Note that equivalently z is given by z int(c). z = f 1 ( f(x) cv). Therefore z, and hence D f (y, z), are easily computable from x, y and v whenever f is explicitly invertible. In that case it is trivial to check whether a given pair (y, v) is an admissible approximate solution in the sense of Definition 3.1: it is enough to obtain z = f 1 ( f(x) cv) and verify if D f (y, z) σ 2 D f (y, x). Since our algorithm is based on this test, it is most easy to implement when f is explicitly invertible. We point out that this case covers a wide range of important applications. For example, Bregman functions with this property are readily available when the feasible set C is an orthant, a polyhedron, a box, or a ball (see [9]). Another important observation is that for σ = 0, we have that y = z. Hence, the only point which satisfies Definition 3.1 for σ = 0, is precisely the exact solution of the proximal subproblem. Therefore our view of inexact solution of generalized proximal subproblems is quite natural. We note, in the passing, that it is motivated by the approach developed in [40] for the classical ( linear ) proximal point method. In that case, Definition 3.1 (albeit slightly modified) is equivalent to saying that the subproblems are solved within fixed relative error tolerance (see also [37]). Such an approach seems to be computationally more realistic/constructive than the common summable-error-type requirements. Regarding the existence of inexact solutions, the situation is clearly even easier than for exact methods. Since we are supposing that the generalized proximal problem (5) has always an exact solution in int(c), this problem will certainly always have (possibly many) inexact solutions (y, v) satisfying also y C. Now we can formally state our inexact generalized proximal method. Algorithm 1 Inexact Generalized Proximal Method. Initialization: Choose some c > 0, and the error tolerance parameter σ [0, 1). Choose some x 0 int(c). Set k := 0. Iteration k: Choose the regularization parameter c k c, and find (y k, v k ), an inexact solution with tolerance σ of 0 c k T ( ) + f( ) f(x k ), (9) 11

satisfying Define Set k := k + 1; and repeat. y k C. (10) x k+1 = f 1 ( f(x k ) c k v k ). (11) We have already discussed the possibility of solving inexactly (9) with condition (10). Another important observation is that since for σ = 0 inexact subproblem solution coincides with the exact one, in that case Algorithm 1 produces the same iterates as the standard exact generalized proximal method. Hence, all our convergence results (some of them are new!) apply also to the exact method. For σ 0 however, there is no direct relation between the iterates of Algorithm 1 and e k+1 + f(x k ) c k T (x k+1 ) + f(x k+1 ), considered in [14]. The advantage of our approach is that it allows an attractive constructive stopping criterion (given by Definition 3.1) for approximate solution of subproblems (at least, when f is invertible). Under our hypothesis, Algorithm 1 is well-defined. From now on, {x k } and {(y k, v k )} are sequences generated by Algorithm 1. Therefore, by the construction of Algorithm 1 and by Definition 3.1, for all k it holds that y k C, x k int(c), (12) v k T (y k ), (13) c k v k + f(x k+1 ) f(x k ) = 0, (14) D f (y k, x k+1 ) σ 2 D f (y k, x k ). (15) We now state our main convergence result. First, recall that a maximal monotone operator T is paramonotone ([4, 5], see also [9, 21]) if u T (x), v T (y) and u v, x y = 0 u T (y) and v T (x). Some examples of paramonotone operators are subdifferentials of proper closed convex functions, and strictly monotone maximal monotone operators. Theorem 3.2 Suppose that VIP(T, C) has solutions and one of the following two conditions holds : 12

1. X int(c). 2. T is paramonotone. Then the sequence {x k } converges to a solution of VIP(T, C). Thus we establish convergence of our inexact algorithm under assumptions which are even weaker than the ones that have been used, until now, for exact algorithms. Specifically, in the paramonotone case, we get rid of the pseudomonotonicity assumption on T [6] which can be stated as follows: Take any sequence {y k } Dom(T ) converging to y and any sequence {v k }, v k T (y k ). Then for each x Dom(T ) there exists and element v T (y) such that v, y x lim inf k vk, y k x. Until now, this (or some other, related) technical assumption was employed in the analysis of all generalized proximal methods (e.g., [14, 7, 6]). Among other things, this resulted in splitting the proof of convergence for the case of minimization and for paramonotone operators (the subdifferential of a convex function is paramonotone, but it need not satisfy the above condition). And of course, the additional requirement of pseudomonotonicity makes the convergence result for paramonotone operators weaker. Since for the tolerance parameter σ = 0 our Algorithm 1 reduces to the exact generalized proximal method, Theorem 3.2 also constitutes a new convergence result for the standard setting of exact proximal algorithms. We note that the stronger than convergence consistency property of Bregman functions established in this paper is crucial for obtaining this new result. To obtain this stronger result, the proof will be somewhat more involved than the usual, and some auxiliary analysis will be needed. However, we think that this is worthwhile since it allows us to remove some (rather awkward) additional assumptions. 4 Convergence Analysis Given sequences {x k }, {y k } and {v k } generated by Algorithm 1, define X as all points x C for which the index set {k N v k, y k x < 0} 13

is finite. For x X, define k(x) as the smallest integer such that k k(x) v k, y k x 0. Of course, the set X and the application k( ) depend on the particular sequences generated by the algorithm. These definitions will facilitate the subsequent analysis. Note that, by monotonicity of T, X X; and in fact, k(x) = 0 for any x X. Lemma 4.1 For any s X and k k(s), it holds that D f (s, x k+1 ) D f (s, x k ) c k v k, y k s (1 σ 2 )D f (y k, x k ) D f (s, x k ). (16) Proof. Take s X and k k(s). Using Lemma 2.6, we get D f (s, x k+1 ) = D f (s, x k ) + f(x k ) f(x k+1 ), s y k + D f (y k, x k+1 ) D f (y k, x k ) By (14) and (15), we further obtain D f (s, x k+1 ) D f (s, x k ) c k v k, y k s (1 σ 2 )D f (y k, x k ), which proves the first inequality in (16). always nonnegative and σ [0, 1), we have Since the Bregman distance is D f (s, x k+1 ) D f (s, x k ) c k v k, y k s. The last inequality in (16) follows directly from the hypothesis s X, k k(s) and the respective definitions. As an immediate consequence, we obtain that the sequence {D f ( x, x k )} is decreasing for any x X. 14

Corollary 4.2 If the sequence {x k } has an accumulation point x X then the whole sequence converges to x. Proof. Suppose that some subsequence {x k j } converges to x X. Using Defintion 2.1 (item 4), we conclude that lim D f( x, x k j ) = 0. j Since the whole sequence {D f ( x, x k )} is decreasing and it has a convergent subsequence, it follows that it converges : lim D f( x, x k ) = 0. k Now the desired result follows from Theorem 2.4. Corollary 4.3 Suppose that X. Then the following statements hold: 1. The sequence {x k } is bounded ; 2. k=0 D f (y k, x k ) < ; 3. For any s X, k=0 v k, y k s < ; 4. The sequence {y k } is bounded. Proof. Take some s X. From Lemma 4.1 it follows that for all k greater than k(s), D f (s, x k ) D f (s, x k(s) ). Therefore, D f (s, x k ) is bounded and from Definition 2.1 (item 3), it follows that {x k } is bounded. By Lemma 4.1, it follows that for any r N D f (s, x k(s)+r+1 ) D f (s, x k(s) ) Therefore k(s)+r k=k(s) k(s)+r k=k(s) k(s)+r c k v k, y k s + (1 σ 2 ) ( ck v k, y k s + (1 σ 2 )D f (y k, x k ) ). k=k(s) D f (y k, x k ) D f (s, x k(s) ). Since r is arbitrary and the terms of both summations are nonnegative, (recall the definition of k(s)), it follows that we can take the limit as r in both 15

sides of the latter relation. Taking further into account that {c k } is bounded away from zero, the second and third assertions of the Corollary easily follow. As consequences, we also obtain that and lim D f(y k, x k ) = 0 (17) k lim k vk, y k s = 0, s X. (18) Suppose now that {y k } is unbounded. Then there exists a pair of subsequences {x k j } and {y k j } such that {x k j } converges but {y k j } diverges. However, by (17) and Theorem 2.4, {y k j } must converge (to the same limit as {x k j }), which contradicts the assumption. Hence, {y k } is bounded. The next proposition establishes the first part of Theorem 3.2, namely the convergence of the inexact generalized proximal algorithm in the case when X int(c). Proposition 4.4 If X int(c) then {x k } converges to some x int(c) which is a solution of VIP(T, C). Proof. By Corollary 4.3, it follows that {x k } is bounded, so it has some accumulation point x C, and for some subsequence {x k j }, lim j xk j = x. Take any ˆx X int(c). Suppose that x bdry(c). Since and, by H2, it follows that D f (ˆx, x k j ) = f(ˆx) f(x k j ) f(x k j ), ˆx x k j lim j f(xk j ), x k j ˆx = +, lim D f(ˆx, x k j ) =. j But the latter is impossible because D f (ˆx, x k ) is a decreasing sequence, at least for k k(ˆx) (by Lemma 4.1). Hence, x int(c). 16

Next, we prove that x is a solution of VIP(T, C). By (17), we have that lim D f(y k j, x k j ) = 0. (19) j Because, by (15), D f (y k j, x k j+1 ) σ 2 D f (y k j, x k j ), it follows that lim D f(y k j, x kj+1 ) = 0. (20) j Since {x k j } converges to x, Theorem 2.4 and (19) imply that {y k j } also converges to x. Applying Theorem 2.4 once again, this time with (20), we conclude that {x kj+1 } also converges to x. Since x int(c) and f is continuous in int(c), we therefore conclude that Since c k c > 0, using (14) we get lim j f(xk j+1 ) f(x k j ) = 0. lim j vk j = 0, where v k j T (y k j ). Now the fact that {y k j } x, together with the maximality of T, implies that 0 T ( x). Thus we have a subsequence {x k j } converging to x X. By Corollary 4.2, the whole sequence {x k } converges to x. We proceed to analyze the case when T is paramonotone. By (18), we already know that if s X then v k, y k s 0. If we could pass onto the limit with respect to v k (for example, using the technical assumption of pseudomonotonicity stated above), then we could conclude that 0 v, x s v s, x s 0, where x is an accumulation point of {x k } (hence also of {y k }), and v T ( x), v s T (s). By paramonotonicity, it follows that v s T ( x). Now by monotonicity, we further obtain that for any x C v s, x x = v s, x s + v s, s x = v s, x s 0, which means that x X. However, in the absence of the assumption of pseudomonotonicity one cannot use this well-established line of argument. To overcome the difficulty resulting from the impossibility of directly passing onto the limit as was done above, we shall need some auxiliary constructions. Let A be the affine hull of the domain of T. Then there exists some V, a subspace of R n, such that A = V + x 17

for any x Dom(T ). Denote by P the orthogonal projection onto V, and for each k define u k = P V (v k ). The idea is to show the following key facts: u k T (y k ), u k, y k x = v k, y k x x Dom(T ), {u k } has an accumulation point. With these facts in hand, we could pass onto the limit in a manner similar to the above, and complete the proof. First, note that x R n, P T (x) = T (x) V. (21) This can be verified rather easily: if x Dom(T ) then both sets in (21) are empty, so it is enough to consider x Dom(T ). Clearly, if some u T (x) V then u P T (x) because u T (x) and also u V. Now take some u P T (x), so that u = P V (v) and v T (x). By monotonicity of T, for any z Dom(T ) and any w T (z), it holds that v w, x z 0. Since x z V and u = P V (v), it holds that Therefore v, x z = u, x z. u w, x z 0, which implies that u T (x) by the maximality of T. Since also u V, it follows that u T (x) V, which establishes (21). Lemma 4.5 If X int(c) = and X, then some subsequence of {u k } is bounded. Proof. We assumed that Dom(T ) int(c). Therefore ri(dom(t )) int(c). Take some ˆx ri(dom(t )) int(c). Then A = V + ˆx, and let P be the projection operator onto V discussed above. In particular, 18

u k = P V (v k ) P T (y k ) = T (y k ) V, so that u k T (y k ). Furthermore, the operator ˆT : V P(V ) defined by ˆT (ξ) = P (T (ξ + ˆx)) is maximal monotone as an operator on the space V (this can be easily verified using the maximal monotonicity of T on R n ). We also have that 0 int(dom( ˆT )). Therefore ˆT is bounded around zero [30]. So, P T is bounded around ˆx, i.e., there exist some r > 0 and M 0 such that x ˆx V, x ˆx r w P T (x), w M. But P T (x) = T (x) V. Hence, x ˆx V, x ˆx r v T (x), v M. Since X int(c) =, it follows that ˆx / X. Therefore, by the definition of X, there exists an infinite subsequence of indices {k j } such that v k j, y k j ˆx < 0. Note that u k j T (y k j ) and, since y k j ˆx V, it holds that Define, for each j, u k j, y k j ˆx = v k j, y k j ˆx < 0. (22) ˆx j = ˆx + (r/ u k j )u k j. Then for each j there exists ˆv j T (ˆx j ) such that Furthermore, ˆv j M. ˆv j, ˆx j y k j u k j, ˆx j y k j = u k j, ˆx j ˆx + u k j, ˆx y k j > u k j, ˆx j ˆx = r u k j, where the first inequality is by the monotonicity of T, and the second is by (22). Using further the Cauchy-Schwarz and triangular inequalities, we obtain r u k j ˆv j ( ˆx j ˆx + ˆx y k j ) M(r + ˆx y k j ). 19

Since the sequence {y k } is bounded (Corollary 4.3, item 4), it follows that { u k j } is bounded. We conclude the analysis by establishing the second part of Theorem 3.2. Proposition 4.6 Suppose X and T is paramonotone. Then {x k } converges to some x X. Proof. If X int(c), then the conclusion follows from Proposition 4.4. Suppose now that X int(c) =. By Lemma 4.5, it follows that some subsequence of {u k } is bounded. Since X X, from Corollary 4.3 it follows that the whole sequence {x k } is bounded. Hence, there exist two subsequences {x k j }, {u k j } which both converge : lim j xk j = x, lim j uk j = ū. Recall from the proof of Lemma 4.5 that u k j T (y k j ). By Corollary 4.3 (item 2), we have that lim j D f(y k j, x k j ) = 0. Therefore, by Theorem 2.4, and lim j yk j = x, ū T ( x), by the maximality of T. Take now some s X. There exists some v s T (s) such that v s, x s 0 for all x C. Therefore, using also the monotonicity of T, Note that for any x Dom(T ) 0 v s, x s ū, x s. (23) v k j, y k j x = u k j, y k j x. 20

Taking x = s, and passing onto the limit as j, (18) implies that Together with (23), this implies that ū, x s = 0. v s, x s = ū, x s = 0. Using now the paramonotonicity of T, we conclude that Finally, for any x C, we obtain v s T ( x). v s, x x = v s, x s + v s, s x = v s, x s 0. Therefore x X. Since we have a subsequence {x k j } converging to x X, from Corollary 4.2 it follows that the whole sequence {x k } converges to x. References [1] H.H. Bauschke and J.M. Borwein. Legendre functions and the method of random Bregman projections. Journal of Convex Analysis, 4:27 67, 1997. [2] L.M. Bregman. The relaxation method of finding the common points of convex sets and its application to the solution of problems in convex programming. USSR Computational Mathematics and Mathematical Physics, 7(3):200 217, 1967. [3] H. Brézis and P.L. Lions. Produits infinis de résolvantes. Israel Journal of Mathematics, 29:329 345, 1978. [4] R.E. Bruck. An iterative solution of a variational inequality for certain monotone operators in a Hilbert space. Bulletin of the American Mathematical Society, 81:890 892, 1975. [5] R.E. Bruck. Corrigendum to [4]. Bulletin of the American Mathematical Society, 82:353, 1976. 21

[6] R.S. Burachik and A.N. Iusem. A generalized proximal point algorithm for the variational inequality problem in a Hilbert space. SIAM Journal on Optimization, 8:197 216, 1998. [7] R.S. Burachik, A.N. Iusem, and B.F. Svaiter. Enlargement of monotone operators with applications to variational inequalities. Set-Valued Analysis, 5:159 180, 1997. [8] J.V. Burke and M. Qian. A variable metric proximal point algorithm for monotone operators. SIAM Journal on Control and Optimization, 37:353 375, 1998. [9] Y. Censor, A.N. Iusem, and S.A. Zenios. An interior point method with Bregman functions for the variational inequality problem with paramonotone operators. Mathematical Programming, 81:373 400, 1998. [10] Y. Censor and S. Zenios. The proximal minimization algorithm with D-functions. Journal of Optimization Theory and Applications, 73:451 464, 1992. [11] G. Chen and M. Teboulle. Convergence analysis of proximal-like optimization algorithm using Bregman functions. SIAM Journal on Optimization, 3:538 543, 1993. [12] R.W. Cottle, F. Giannessi, and J.-L. Lions. Variational Inequalities and Complementarity Problems : Theory and Applications. Wiley, New York, 1980. [13] J. Eckstein. Nonlinear proximal point algorithms using Bregman functions, with applications to convex programming. Mathematics of Operations Research, 18:202 226, 1993. [14] J. Eckstein. Approximate iterations in Bregman-function-based proximal algorithms. Mathematical Programming, 83:113 123, 1998. [15] J. Eckstein and D.P. Bertsekas. On the Douglas-Rachford splitting method and the proximal point algorithm for maximal monotone operators. Mathematical Programming, 55:293 318, 1992. [16] P.P.B. Eggermont. Multiplicative iterative algorithms for convex programming. Linear Algebra and Its Applications, 130:25 32, 1990. 22

[17] M.C. Ferris. Finite termination of the proximal point algorithm. Mathematical Programming, 50:359 366, 1991. [18] O. Güler. New proximal point algorithms for convex minimization. SIAM Journal on Optimization, 2:649 664, 1992. [19] A. N. Iusem. On some properties of generalized proximal point methods for quadratic and linear programming. Journal of Optimization Theory and Applications, 85:596 612, 1995. [20] A. N. Iusem. On some properties of generalized proximal point methods for the variational inequality problem. Journal of Optimization Theory and Applications, 96:337 362, 1998. [21] A. N. Iusem. On some properties of paramonotone operators. Journal of Convex Analysis, 5:269 278, 1998. [22] A.N. Iusem and M.V. Solodov. Newton-type methods with generalized distances for constrained optimization. Optimization, 41:257 278, 1997. [23] K.C. Kiwiel. Proximal minimization methods with generalized Bregman functions. SIAM Journal on Control and Optimization, 35:1142 1168, 1997. [24] B. Lemaire. The proximal algorithm. In J.P. Penot, editor, New Methods of Optimization and Their Industrial Use. International Series of Numerical Mathematics 87, pages 73 87. Birkhauser, Basel, 1989. [25] F.J. Luque. Asymptotic convergence analysis of the proximal point algorithm. SIAM Journal on Control and Optimization, 22:277 293, 1984. [26] B. Martinet. Regularisation d inequations variationelles par approximations successives. Revue Française d Informatique et de Recherche Opérationelle, 4:154 159, 1970. [27] J.-J. Moreau. Proximité et dualité dans un espace Hilbertien. Bulletin de la Société Mathématique de France, 93:273 299, 1965. [28] J.-S. Pang. Complementarity problems. In R. Horst and P. Pardalos, editors, Handbook of Global Optimization, pages 271 338. Kluwer Academic Publishers, Boston, Massachusetts, 1995. 23

[29] G.B. Passty. Weak convergence theorems for nonexpansive mappings in Banach spaces. Journal of Mathematical Analysis and Applications, 67:274 276, 1979. [30] R. T. Rockafellar. Local boundedness of nonlinear monotone operators. Michigan Mathematical Journal, 16:397 407, 1969. [31] R. T. Rockafellar. On the maximality of sums of nonlinear monotone operators. Transactions of the American Mathematical Society, 149:75 88, 1970. [32] R.T. Rockafellar. Convex Analysis. Princeton University Press, Princeton, NJ, 1970. [33] R.T. Rockafellar. Augmented Lagrangians and applications of the proximal point algorithm in convex programming. Mathematics of Operations Research, 1:97 116, 1976. [34] R.T. Rockafellar. Monotone operators and the proximal point algorithm. SIAM Journal on Control and Optimization, 14:877 898, 1976. [35] M. V. Solodov and B. F. Svaiter. A truly globally convergent Newtontype method for the monotone nonlinear complementarity problem, 1998. SIAM Journal on Optimization, to appear. [36] M.V. Solodov and B.F. Svaiter. A comparison of rates of convergence of two inexact proximal point algorithms. In G. Di Pillo and F. Giannessi, editors, Nonlinear Optimization and Applications 2. Kluwer Academic Publishers. To appear. [37] M.V. Solodov and B.F. Svaiter. Forcing strong convergence of proximal point iterations in a Hilbert space, 1997. Mathematical Programming, to appear. [38] M.V. Solodov and B.F. Svaiter. A globally convergent inexact Newton method for systems of monotone equations. In M. Fukushima and L. Qi, editors, Reformulation - Nonsmooth, Piecewise Smooth, Semismooth and Smoothing Methods, pages 355 369. Kluwer Academic Publishers, 1998. [39] M.V. Solodov and B.F. Svaiter. A hybrid approximate extragradient proximal point algorithm using the enlargement of a maximal monotone operator, 1998. Set-Valued Analysis, submitted. 24

[40] M.V. Solodov and B.F. Svaiter. A hybrid projection proximal point algorithm. Journal of Convex Analysis, 6(1), 1999. [41] M. Teboulle. Convergence of proximal-like algorithms. SIAM Journal on Optimization, 7:1069 1083, 1997. 25