A Strongly Convergent Method for Nonsmooth Convex Minimization in Hilbert Spaces

Similar documents
On the convergence properties of the projected gradient method for convex optimization

Nacional de La Pampa, Santa Rosa, La Pampa, Argentina b Instituto de Matemática Aplicada San Luis, Consejo Nacional de Investigaciones Científicas

Forcing strong convergence of proximal point iterations in a Hilbert space

Guangzhou, P.R. China

Precise Large Deviations for Sums of Negatively Dependent Random Variables with Common Long-Tailed Distributions

University, Tempe, Arizona, USA b Department of Mathematics and Statistics, University of New. Mexico, Albuquerque, New Mexico, USA

Dissipation Function in Hyperbolic Thermoelasticity

Communications in Algebra Publication details, including instructions for authors and subscription information:

Open problems. Christian Berg a a Department of Mathematical Sciences, University of. Copenhagen, Copenhagen, Denmark Published online: 07 Nov 2014.

A convergence result for an Outer Approximation Scheme

Testing Goodness-of-Fit for Exponential Distribution Based on Cumulative Residual Entropy

Version of record first published: 01 Sep 2006.

Gilles Bourgeois a, Richard A. Cunjak a, Daniel Caissie a & Nassir El-Jabi b a Science Brunch, Department of Fisheries and Oceans, Box

Journal of Convex Analysis (accepted for publication) A HYBRID PROJECTION PROXIMAL POINT ALGORITHM. M. V. Solodov and B. F.

The Homogeneous Markov System (HMS) as an Elastic Medium. The Three-Dimensional Case

PARALLEL SUBGRADIENT METHOD FOR NONSMOOTH CONVEX OPTIMIZATION WITH A SIMPLE CONSTRAINT

arxiv: v1 [math.fa] 16 Jun 2011

PLEASE SCROLL DOWN FOR ARTICLE

Use and Abuse of Regression

On Total Convexity, Bregman Projections and Stability in Banach Spaces

FIXED POINTS IN THE FAMILY OF CONVEX REPRESENTATIONS OF A MAXIMAL MONOTONE OPERATOR

Characterizations of Student's t-distribution via regressions of order statistics George P. Yanev a ; M. Ahsanullah b a

The Fourier transform of the unit step function B. L. Burrows a ; D. J. Colwell a a

Ankara, Turkey Published online: 20 Sep 2013.

Online publication date: 01 March 2010 PLEASE SCROLL DOWN FOR ARTICLE

Full terms and conditions of use:

Derivation of SPDEs for Correlated Random Walk Transport Models in One and Two Dimensions

Diatom Research Publication details, including instructions for authors and subscription information:

Erciyes University, Kayseri, Turkey

Park, Pennsylvania, USA. Full terms and conditions of use:

Global Existence of Large BV Solutions in a Model of Granular Flow

A strongly convergent hybrid proximal method in Banach spaces

Full terms and conditions of use:

Merit functions and error bounds for generalized variational inequalities

Discussion on Change-Points: From Sequential Detection to Biology and Back by David Siegmund

Published online: 05 Oct 2006.

Strong convergence to a common fixed point. of nonexpansive mappings semigroups

Kantorovich s Majorants Principle for Newton s Method

OF SCIENCE AND TECHNOLOGY, TAEJON, KOREA

A double projection method for solving variational inequalities without monotonicity

George L. Fischer a, Thomas R. Moore b c & Robert W. Boyd b a Department of Physics and The Institute of Optics,

FB 4, University of Osnabrück, Osnabrück

Convergence rate estimates for the gradient differential inclusion

Fixed points in the family of convex representations of a maximal monotone operator

Online publication date: 30 March 2011

arxiv: v3 [math.oc] 18 Apr 2012

G. S. Denisov a, G. V. Gusakova b & A. L. Smolyansky b a Institute of Physics, Leningrad State University, Leningrad, B-

Journal of Convex Analysis Vol. 14, No. 2, March 2007 AN EXPLICIT DESCENT METHOD FOR BILEVEL CONVEX OPTIMIZATION. Mikhail Solodov. September 12, 2005

Móstoles, Madrid. Spain. Universidad de Málaga. Málaga. Spain

AN INEXACT HYBRID GENERALIZED PROXIMAL POINT ALGORITHM AND SOME NEW RESULTS ON THE THEORY OF BREGMAN FUNCTIONS. May 14, 1998 (Revised March 12, 1999)

Geometrical optics and blackbody radiation Pablo BenÍTez ab ; Roland Winston a ;Juan C. Miñano b a

An Accelerated Hybrid Proximal Extragradient Method for Convex Optimization and its Implications to Second-Order Methods

An inexact subgradient algorithm for Equilibrium Problems

PLEASE SCROLL DOWN FOR ARTICLE

University, Wuhan, China c College of Physical Science and Technology, Central China Normal. University, Wuhan, China Published online: 25 Apr 2014.

On the order of the operators in the Douglas Rachford algorithm

Online publication date: 22 March 2010

PLEASE SCROLL DOWN FOR ARTICLE. Full terms and conditions of use:

The American Statistician Publication details, including instructions for authors and subscription information:

Acyclic, Cyclic and Polycyclic P n

AN INEXACT HYBRID GENERALIZED PROXIMAL POINT ALGORITHM AND SOME NEW RESULTS ON THE THEORY OF BREGMAN FUNCTIONS. M. V. Solodov and B. F.

On the iterate convergence of descent methods for convex optimization

Maximal monotone operators, convex functions and a special family of enlargements.

Some Inexact Hybrid Proximal Augmented Lagrangian Algorithms

Upper sign-continuity for equilibrium problems

WEAK CONVERGENCE OF RESOLVENTS OF MAXIMAL MONOTONE OPERATORS AND MOSCO CONVERGENCE

ON A HYBRID PROXIMAL POINT ALGORITHM IN BANACH SPACES

arxiv: v1 [math.na] 25 Sep 2012

MSE Performance and Minimax Regret Significance Points for a HPT Estimator when each Individual Regression Coefficient is Estimated

Tore Henriksen a & Geir Ulfstein b a Faculty of Law, University of Tromsø, Tromsø, Norway. Available online: 18 Feb 2011

35-959, Rzeszów, Poland b Institute of Computer Science, Jagiellonian University,

A NEW ITERATIVE METHOD FOR THE SPLIT COMMON FIXED POINT PROBLEM IN HILBERT SPACES. Fenghui Wang

Monotone operators and bigger conjugate functions

Pacific Journal of Optimization (Vol. 2, No. 3, September 2006) ABSTRACT

On the Weak Convergence of the Extragradient Method for Solving Pseudo-Monotone Variational Inequalities

PLEASE SCROLL DOWN FOR ARTICLE

A WEAK-TO-STRONGCONVERGENCE PRINCIPLE FOR FEJÉR-MONOTONE METHODS IN HILBERT SPACES

A GENERALIZATION OF THE REGULARIZATION PROXIMAL POINT METHOD

AN INEXACT HYBRIDGENERALIZEDPROXIMAL POINT ALGORITHM ANDSOME NEW RESULTS ON THE THEORY OF BREGMAN FUNCTIONS

Spectral gradient projection method for solving nonlinear monotone equations

Brøndsted-Rockafellar property of subdifferentials of prox-bounded functions. Marc Lassonde Université des Antilles et de la Guyane

ENLARGEMENT OF MONOTONE OPERATORS WITH APPLICATIONS TO VARIATIONAL INEQUALITIES. Abstract

A projection-type method for generalized variational inequalities with dual solutions

Convergence Theorems of Approximate Proximal Point Algorithm for Zeroes of Maximal Monotone Operators in Hilbert Spaces 1

On the complexity of the hybrid proximal extragradient method for the iterates and the ergodic mean

University of Thessaloniki, Thessaloniki, Greece

Tong University, Shanghai , China Published online: 27 May 2014.

Strong Convergence Theorem by a Hybrid Extragradient-like Approximation Method for Variational Inequalities and Fixed Point Problems

Maximal Monotone Operators with a Unique Extension to the Bidual

CCSM: Cross correlogram spectral matching F. Van Der Meer & W. Bakker Published online: 25 Nov 2010.

Communications in Algebra Publication details, including instructions for authors and subscription information:

ON GAP FUNCTIONS OF VARIATIONAL INEQUALITY IN A BANACH SPACE. Sangho Kum and Gue Myung Lee. 1. Introduction

Convergence Theorems for Bregman Strongly Nonexpansive Mappings in Reflexive Banach Spaces

THE CYCLIC DOUGLAS RACHFORD METHOD FOR INCONSISTENT FEASIBILITY PROBLEMS

A New Modified Gradient-Projection Algorithm for Solution of Constrained Convex Minimization Problem in Hilbert Spaces

1. Introduction. We consider the classical variational inequality problem [1, 3, 7] VI(F, C), which is to find a point x such that

Convergence rate of inexact proximal point methods with relative error criteria for convex optimization

Published online: 04 Jun 2015.

Existence results for quasi-equilibrium problems

Transcription:

This article was downloaded by: [IMPA Inst de Matematica Pura & Aplicada] On: 11 November 2011, At: 05:10 Publisher: Taylor & Francis Informa Ltd Registered in England and Wales Registered Number: 1072954 Registered office: Mortimer House, 37-41 Mortimer Street, London W1T 3JH, UK Numerical Functional Analysis and Optimization Publication details, including instructions for authors and subscription information: http://www.tandfonline.com/loi/lnfa20 A Strongly Convergent Method for Nonsmooth Convex Minimization in Hilbert Spaces J. Y. Bello Cruz a & A. N. Iusem b a Instituto de Matemática e Estatística, Universidade Federal de Goiás, Campus Samambaia, Goiânia, Brazil b Instituto de Matemática Pura e Aplicada, Jardim Botânico, Rio de Janeiro, Brazil Available online: 10 Aug 2011 To cite this article: J. Y. Bello Cruz & A. N. Iusem (2011): A Strongly Convergent Method for Nonsmooth Convex Minimization in Hilbert Spaces, Numerical Functional Analysis and Optimization, 32:10, 1009-1018 To link to this article: http://dx.doi.org/10.1080/01630563.2011.590914 PLEASE SCROLL DOWN FOR ARTICLE Full terms and conditions of use: http://www.tandfonline.com/page/terms-and-conditions This article may be used for research, teaching, and private study purposes. Any substantial or systematic reproduction, redistribution, reselling, loan, sub-licensing, systematic supply, or distribution in any form to anyone is expressly forbidden. The publisher does not give any warranty express or implied or make any representation that the contents will be complete or accurate or up to date. The accuracy of any instructions, formulae, and drug doses should be independently verified with primary sources. The publisher shall not be liable for any loss, actions, claims, proceedings, demand, or costs or damages whatsoever or howsoever caused arising directly or indirectly in connection with or arising out of the use of this material.

Numerical Functional Analysis and Optimization, 32(10):1009 1018, 2011 Copyright Taylor & Francis Group, LLC ISSN: 0163-0563 print/1532-2467 online DOI: 10.1080/01630563.2011.590914 A STRONGLY CONVERGENT METHOD FOR NONSMOOTH CONVEX MINIMIZATION IN HILBERT SPACES J. Y. Bello Cruz 1 and A. N. Iusem 2 1 Instituto de Matemática e Estatística, Universidade Federal de Goiás, Campus Samambaia, Goiânia, Brazil 2 Instituto de Matemática Pura e Aplicada, Jardim Botânico, Rio de Janeiro, Brazil In this article, we propose a strongly convergent variant on the projected subgradient method for constrained convex minimization problems in Hilbert spaces. The advantage of the proposed method is that it converges strongly when the problem has solutions, without additional assumptions. The method also has the following desirable property: the sequence converges to the solution of the problem which lies closest to the initial iterate. Keywords Convex minimization; Nonsmooth optimization; Projected subgradient algorithm; Projection method; Strong convergence. AMS Subject Classification 90C25; 65C25; 90C30. 1. INTRODUCTION First, we briefly describe our notation. Let C be a nonempty, closed and convex subset of a real Hilbert space H and f : H be a continuous, proper and convex function. The inner product in H is denoted by,. The norm determined by the inner product is denoted by. For an element x H, the orthogonal projection of x onto C is denoted by P C (x). The problem under consideration is the following: min f (x) s.t. x C (1) We denote f := inf x C f (x) and the solution set of this problem by S := x C : f (x) = f. Received 22 November 2010; Revised and Accepted 17 May 2011. Address correspondence to J. Y. Bello Cruz, Instituto de Matemática e Estatística, Universidade Federal de Goiás, Campus Samambaia, Goiânia, CEP 74001-970 GO Brazil; E-mail: yunier@impa.br 1009

1010 J. Y. Bello Cruz and A. N. Iusem We remind that the subdifferential of f at x is the subset f (x) of H given by f (x) = u H : f (y) f (x) + u, y x y H (2) The elements of the subdifferential of f at x are called subgradients of f at x. 1.1. The Projected Subgradient Method Many methods have been proposed to solve Problem (1), with the simplest being the projected subgradient method [1, 2, 17, 18]. The projected subgradient method has several advantages. Primarily, it is easy to implement (specifically, for optimization problems with relatively simple constraints), it uses little storage and readily exploits any sparsity or separable structure of f or C. Furthermore, it is able to drop or add active constraints during the iterations. Thus, the projected subgradient method has been widely used for solving various cases of Problem (1). Each iteration of the projected subgradient method, consists basically of two stages: Starting from the iterate x k, we move a certain length in the direction opposite to a subgradient of f at x k, and next the resulting vector is projected onto C. This is not a trivial idea, since the function will not, in general, decrease along a direction selected in this way. When C = H and f is differentiable this is just the steepest descent method [8]. For differentiable functions the projected subgradient method coincides with the projected gradient method and different variants rules are used to choose the stepsizes in order to ensure functional decrease at each iteration; see [13]. Its extension to the generalized convex case is studied in [7, 14, 21]. 1.2. On Strong Convergence 1.2.1. The New Method In this article, a modification of the projected subgradient method forcing strong convergence in Hilbert spaces is considered. The new method is related to the classical projected subgradient method and it uses a similar idea exposed in [6, 19], with the same goal, upgrading weak convergence to strong one. Strong convergence is forced by combining a subgradient iteration with a simple projection step onto the intersection of C and two halfspaces, containing S. Additionally, our algorithm has the distinctive feature that the limit of the generated sequence is the closest solution of the problem to the initial iterate x 0. This property is useful in many specific applications, e.g. in image reconstruction [9, 12, 16]. We emphasize that this feature is of

Nonsmooth Convex Minimization in Hilbert Spaces 1011 interest also in finite dimension, differently from the strong versus weak convergence issue. We mention that the proposed method requires in each iteration the computation of the exact projection onto the intersection of C with two halfspaces. The presence of the halfspaces does not entail any significant additional cost over the computation of the projection onto C itself. The computational cost of this projection is similar to the cost of a projected subgradient step. In order to prove the convergence of the new method, we assume that f is finite valued, so that the effective domain is H. Also we suppose that f is bounded on bounded sets, which holds automatically in finite dimension, as a consequence of the maximal monotonicity of the subdifferential of lower semicontinuous convex functions. We also mention that these assumptions are required in the analysis of [2] for proving weak convergence of the projected subgradient method in infinite-dimensional Hilbert spaces. 1.2.2. Why Strong Convergence? It is important to mention that many real-world problems in economics and engineering are modeled in infinite-dimensional spaces. These include optimal control and structural design problems, among others. It is clear that weak and strong convergence are only distinguishable in the infinite-dimensional setting. On the other hand, even when we have to solve infinite-dimensional problems, numerical implementations of algorithms are certainly applied to finite-dimensional approximations of the problems. Nevertheless, it is important to have convergence theory for the infinite-dimensional case, because it guarantees robustness and stability with respect to discretization schemes employed for obtaining finitedimensional approximations of infinite-dimensional problems. This issue is closely related to the so-called Mesh Independence Principle [3, 4, 15]. This principle relies on infinite-dimensional convergence to predict the convergence properties of a discretized finite-dimensional method. Furthermore, the mesh independence provides theoretical justification for the design of refinement strategies. Note that fine discretization is crucial for the obtained discrete solution to be an appropriate approximation to the true solution of the infinitedimensional problem being solved. Many important real-world problems in economics and engineering are modeled in infinite-dimensional spaces. These include optimal control and structural design problems, and the problem of minimal area surface with obstacles, among others. We refer the reader to [11], where a variety of applications are described. A strong convergence principle for Fejér-monotone methods in Hilbert spaces was presented in [5].

1012 J. Y. Bello Cruz and A. N. Iusem The importance of strong convergence is also underlined in [10], where a convex function f is minimized via the proximal-point algorithm: It is shown that the rate of convergence of the value sequence f (x k ) is better when x k converges strongly than when it converges weakly. Such properties have a direct impact when the algorithm is executed directly in the underlying infinite-dimensional space, as is the case, for instance, in optical signal processing, see [20]. 1.3. Preliminary Material In this subsection, we present some results that are used in the convergence analysis. First, we state a well known fact on orthogonal projections. Proposition 1. Let K be any nonempty closed and convex set in H. For all x, y H and all z K, x P K (x), z P K (x) 0. Proof. Elementary. We continue with an elementary property of the solution set of Problem (1). Proposition 2. Assume that f : H is proper, finite-valued, convex and continuous. Then S, if nonempty, is a closed and convex set. Proof. Closedness of S = x C : f (x) f follows easily from the continuity of f. Convexity of S is elementary. 2. THE ALGORITHM In this section, we state the method, which generates a sequence strongly convergent to a point belonging to S, differently from the projected subgradient method, for which only weak convergence has been established. We assume that the optimal value f of Problem (1) is available. Algorithm A Initialization step. Take x 0 C, 0 := f (x 0 ) f and u 0 f (x 0 ). Iterative step. Given x k and u k f (x k ), define k := f (x k ) f, H k := x H : x x k, u k + k 0, (3)

Nonsmooth Convex Minimization in Hilbert Spaces 1013 and W k := x H : x x k, x 0 x k 0 Compute x k+1 := P Hk W k C(x 0 ) (4) If x k+1 = x k then stop. Before the formal analysis of the convergence properties of the algorithm, we make the following remarks: 1. Regarding the projection step, we shall prove that the set H k W k C is nonempty, ensuring that the algorithm is well defined. 2. Concerning the complexity of the projection step in Algorithm A, the presence of the halfspaces does not entail any significant additional cost over the computation of the projection onto C itself. The computational cost of this projection is similar to the cost of a projected subgradient step. Even though we are working in an infinitedimensional space, projection onto an intersection of two halfspaces amounts to solving, at most, a linear system of two equations with two unknowns. Thus, the cost of each iteration of Algorithm A is about the same as that of an iteration of the projected subgradient method. Next we proceed to establish strong convergence of the sequence generated by Algorithm A. 3. CONVERGENCE ANALYSIS First, we establish some properties of the iterates. Proposition 3. and For all k 0 it holds that x k+1 x 0 2 x k x 0 2 + x k+1 x k 2, (5) x k+1 x k f (x k ) f (6) u k Proof. Since x k+1 W k, 0 x k+1 x k, x 0 x k = 1 ( x k+1 x k 2 x k+1 x 0 2 + x k x 0 2), 2 which implies (5).

1014 J. Y. Bello Cruz and A. N. Iusem Now, since x k P Hk (x k ) z x k for all z H k have that and x k+1 H k,we x k+1 x k x k P Hk (x k ) (7) Using (3) and the fact that P Hk (x k ) = x k k, we obtain from (7) u k 2 u k x k+1 x k x k P Hk (x k ) = k u k = f (x k ) f u k Now we show feasibility of the sequence generated by Algorithm A and the validity of the stopping criterion. Proposition 4. If the Algorithm A generates an infinite sequence x k, then x k C. Furthermore, if Algorithm A stops, x k S. Proof. First suppose that Algorithm A generates an infinite sequence. By (4), we have x k C. If Algorithm A stops, then x k+1 = x k. It follows from (6) that f (x k ) f 0, so x k S. Next we prove optimality of the weak cluster points of x k. Theorem 1. Suppose that Algorithm A generates an infinite sequence x k. Then either x k is bounded and each of its weak cluster points belongs to S =,or S = and lim x k =. Proof. If x k is bounded, we obtain from (5) that the sequence x k x 0 is nondecreasing and bounded, hence convergent. By (5) again, 0 x k+1 x k 2 x k+1 x 0 2 x k x 0 2, and we conclude that By Proposition 4, f (x k ) f 0 for all k. So, lim x k+1 x k 2 = 0 (8) 0 = lim x k+1 x k lim f (x k ) f u k 0, using (6) and (8). The boundedness of u k follows from the boundedness of x k and the assumption that f is bounded on bounded sets. Thus, lim f (x k ) f = 0, (9)

Nonsmooth Convex Minimization in Hilbert Spaces 1015 Since x k is bounded, there exists a weakly convergent subsequence. Let x i k be any weakly convergent subsequence of x k, and let x C be its weak limit. Since f is weakly lower semicontinuous we get, using (9), f f ( x) lim inf f (x i k ) = lim f (x k ) = f Thus, f ( x) = f, implying that x S. We conclude that all weak cluster points of x k belongs to S. Suppose now that S =. Using the previous assertion in this theorem, we obtain that x k has no bounded subsequence, and consequently lim x k =. Next, we shall prove that when S = Algorithm A is well defined and the generated sequence x k converges strongly to a solution. We assume from now on that S is nonempty. Having chosen the initial iterate x 0, define V 0 := x H : z x, x 0 x 0 z S (10) Next, we show that the generated sequence x k is contained in V 0 and that the set H k W k C always contains the solution set S. Proposition 5. If x k V 0 then i) S H k W k C, ii) x k+1 is well defined and x k+1 V 0. Proof. (i) Since k = f (x k ) f by (3), we get, using (2), k + u k, x x k =f (x k ) f (x ) + u k, x x k 0, (11) for any x S. It follows from (11) that S H k. Since x k V 0, we have that x x k, x 0 x k 0 for all x S.By the definition of W k, we obtain that S W k. Therefore, we conclude that S H k W k C. (ii) Since S H k W k C and S is nonempty, it follows that H k W k C is nonempty. Thus, the next iterate x k+1 is well defined, in view of (4). Using (4) and Proposition 1, we have that z x k+1, x 0 x k+1 0 z H k W k C (12) By (i), S H k W k C for all k. Since (12) holds for all z S,sox k+1 V 0 by (10).

1016 J. Y. Bello Cruz and A. N. Iusem Corollary 1. all k. Algorithm A is well defined, x k V 0 and S H k W k C for Proof. It is enough to observe that x 0 V 0 and apply inductively Proposition 5. Corollary 2. The sequence x k generated by Algorithm A is bounded and each of its weak cluster points belong to S. Proof. If the solution set is nonempty, in view of (4) we have that x k+1 x 0 z x 0 for all z H k W k C. Since S H k W k C by Corollary 1, it follows that x k+1 x 0 x x 0 for all x S. Thus, x k is bounded, and by Theorem 1, all its weak cluster points belong to S. Finally, we are now ready to prove strong convergence of the sequence x k generated by Algorithm A to the solution which lies closest to x 0. Theorem 2. Assume that S = and let x k be a sequence generated by Algorithm A. Define x = P S (x 0 ). Then x k converges strongly to x. Proof. By Proposition 2, S is closed and convex. Therefore x, the orthogonal projection of x 0 onto S, exists. By the definition of x k+1,we have that x k+1 x 0 z x 0 z H k W k C (13) Since x S H k W k C for all k, it follows from (13) that x k x 0 x x 0, (14) for all k. By Corollary 2, x k is bounded and each of its weak cluster points belongs to S. Let x i k be any weakly convergent subsequence of x k, and let ˆx S be its weak limit. Observe that x i k x 2 = x i k x 0 (x x 0 ) 2 = x i k x 0 2 + x x 0 2 2 x i k x 0, x x 0 2 x x 0 2 2 x i k x 0, x x 0, where the inequality follows from (14). By the weak convergence of x i k to ˆx, we obtain lim sup x i k x 2 2( x x 0 2 ˆx x 0, x x 0 ) (15)

Nonsmooth Convex Minimization in Hilbert Spaces 1017 Applying Proposition 1 with K = S, x = x 0 and z =ˆx S, and taking into account that x is the projection of x 0 onto S, we have that Now, using (16) we have x 0 x, ˆx x 0 (16) 0 ˆx x, x x 0 = x 0 x, x x 0 ˆx x 0, x x 0 x x 0 2 ˆx x 0, x x 0 (17) Combining (17) with (15), we conclude that x i k converges strongly to x. Thus, we have shown that every weakly convergent subsequence of x k converges strongly to x. Hence, the whole sequence x k converges strongly to x S. ACKNOWLEDGMENTS Research for this article was partially supported by PROCADnf-UFG/UnB/IMPA research and PRONEX-CNPq-FAPERJ-Optimization Research. J. Y. B. C thanks the Institute for Pure and Applied Mathematics (IMPA), at Rio de Janeiro, Brazil, where he was a visitor while working on this article. REFERENCES 1. Ya.I. Alber and A.N. Iusem (2001). Extension of subgradient techniques for nonsmooth optimization in Banach spaces. Set-Valued Analysis 9:315 335. 2. Ya.I. Alber, A.N. Iusem, and M.V. Solodov (1998). On the projected subgradient method for nonsmooth convex optimization in a Hilbert space. Math. Prog. 81:23 37. 3. E.L. Allgower and K. Böhmer (1987). Application of the mesh-independence principle to mesh re_nement strategies. SIAM J. Numer. Anal. 24:1335 1351. 4. E.L. Allgower, K. Böhmer, F.A. Potra, and W.C. Rheinboldt (1986). A mesh-independence principle for operator equations and their discretizations. SIAM J. Numer. Anal. 23:160 169. 5. H.H. Bauschke and P.L. Combettes (2001). A weak-to-strong convergence principle for Fejérmonotone methods in Hilbert spaces. Math. Oper. Res. 26:248 264. 6. J.Y. Bello Cruz and A.N. Iusem (2009). A strongly convergent direct method for monotone variational inequalities in Hilbert spaces. Numer. Funct. Anal. Optim. 30:23 36. 7. J.Y. Bello Cruz and L.R. Lucambio Pérez (2010). Convergence of a projected gradient method variant for quasiconvex objectives. Nonlinear Anal. 9:2917 2922. 8. R. Burachik, L.M. Graña Drummond, A.N. Iusem, and B.F. Svaiter (1995). Full convergence of the steepest descent method with inexact line searches. Optimization 32:137 146. 9. R. Gordon and G.T. Herman (1971). Reconstruction of pictures from their projections. Communications of the Association for Computing Machinery 14:759 768. 10. O. Güler (1992). New proximal point algorithms for convex minimization. SIAM J. Optim. 2:649 664. 11. J. Henry and J.-P. Yvon (1994). System Modelling and Optimization. Lecture Notes in Control and Information Sciences 197. Springer, London. 12. H.M. Hudson and R.S. Larkin (1994). Accelerated image reconstruction using ordered subsets of projection data. IEEE Trans. Med. Imaging 13:601 609.

1018 J. Y. Bello Cruz and A. N. Iusem 13. A.N. Iusem (2003). On the convergence properties of the projected gradient method for convex optimization. Comput. Appl. Maths. 22:37 52. 14. K.C. Kiwiel and K. Murty (1996). Convergence of the steepest descent method for minimizing quasiconvex functions. J. Optim. Theory Appl. 89:221 226. 15. M. Laumen (1999) Newton s mesh independence principle for a class of optimal shape design problems. SIAM J. Control Optim. 37:1070 1088. 16. A.J. Rockmore and A. Macovski (1977). A maximum likelihood approach to transmission image reconstruction from projections. IEEE Trans. Nuclear Sci. 24:1929 1935. 17. B.T. Polyak (1969). Minimization of unsmooth functionals. U.S.S.R. Computational Mathematics and Mathematical Physics 9:14 29. 18. N.Z. Shor (1985). Minimization Methods for Nondifferentiable Functions. Springer, Berlin. 19. M.V. Solodov and B.F. Svaiter (2000). Forcing strong convergence of proximal point iterations in a Hilbert space. Math. Prog. 87:189 202. 20. A. Vanderlugt (1992). Optical Signal Processing. Wiley, New York. 21. C. Wang and N. Xiu (2000). Convergence of the gradient projection method for generalized convex minimization. Comput. Optim. Appl. 16:111 120.