A New Method for Nonsmooth

Similar documents
Journal of Convex Analysis (accepted for publication) A HYBRID PROJECTION PROXIMAL POINT ALGORITHM. M. V. Solodov and B. F.

1 Introduction We consider the problem nd x 2 H such that 0 2 T (x); (1.1) where H is a real Hilbert space, and T () is a maximal monotone operator (o

Finite Convergence for Feasible Solution Sequence of Variational Inequality Problems

A Proximal Method for Identifying Active Manifolds

A CHARACTERIZATION OF STRICT LOCAL MINIMIZERS OF ORDER ONE FOR STATIC MINMAX PROBLEMS IN THE PARAMETRIC CONSTRAINT CASE

Some Properties of the Augmented Lagrangian in Cone Constrained Optimization

McMaster University. Advanced Optimization Laboratory. Title: A Proximal Method for Identifying Active Manifolds. Authors: Warren L.

A proximal-like algorithm for a class of nonconvex programming

A convergence result for an Outer Approximation Scheme

ON A CLASS OF NONSMOOTH COMPOSITE FUNCTIONS

A Note on Nonconvex Minimax Theorem with Separable Homogeneous Polynomials

A Dykstra-like algorithm for two monotone operators

LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE

Some Inexact Hybrid Proximal Augmented Lagrangian Algorithms

Structural and Multidisciplinary Optimization. P. Duysinx and P. Tossings

Self-dual Smooth Approximations of Convex Functions via the Proximal Average

An Infeasible Interior Proximal Method for Convex Programming Problems with Linear Constraints 1

Lagrange Multipliers with Optimal Sensitivity Properties in Constrained Optimization

Convergence Theorems of Approximate Proximal Point Algorithm for Zeroes of Maximal Monotone Operators in Hilbert Spaces 1

On the Local Quadratic Convergence of the Primal-Dual Augmented Lagrangian Method

On convergence rate of the Douglas-Rachford operator splitting method

Convergence rate estimates for the gradient differential inclusion

(7: ) minimize f (x) subject to g(x) E C, (P) (also see Pietrzykowski [5]). We shall establish an equivalence between the notion of

An inexact subgradient algorithm for Equilibrium Problems

JOURNAL OF OPTIMIZATION THEORY AND APPLICATIONS: Vol. 89, No. 1, pp , APRIL 1996 TECHNICAL NOTE

A Geometric Framework for Nonconvex Optimization Duality using Augmented Lagrangian Functions

On Second-order Properties of the Moreau-Yosida Regularization for Constrained Nonsmooth Convex Programs

On Total Convexity, Bregman Projections and Stability in Banach Spaces

("-1/' .. f/ L) I LOCAL BOUNDEDNESS OF NONLINEAR, MONOTONE OPERA TORS. R. T. Rockafellar. MICHIGAN MATHEMATICAL vol. 16 (1969) pp.

A Unified Approach to Proximal Algorithms using Bregman Distance

Coordinate Update Algorithm Short Course Proximal Operators and Algorithms

Math 273a: Optimization Subgradient Methods

The Newton Bracketing method for the minimization of convex functions subject to affine constraints

A quasisecant method for minimizing nonsmooth functions

Spectral gradient projection method for solving nonlinear monotone equations

Existence of Global Minima for Constrained Optimization 1

On the simplest expression of the perturbed Moore Penrose metric generalized inverse

Pacific Journal of Optimization (Vol. 2, No. 3, September 2006) ABSTRACT

A Novel Inexact Smoothing Method for Second-Order Cone Complementarity Problems

A derivative-free nonmonotone line search and its application to the spectral residual method

SOLVING A MINIMIZATION PROBLEM FOR A CLASS OF CONSTRAINED MAXIMUM EIGENVALUE FUNCTION

Relationships between upper exhausters and the basic subdifferential in variational analysis

AN INEXACT HYBRID GENERALIZED PROXIMAL POINT ALGORITHM AND SOME NEW RESULTS ON THE THEORY OF BREGMAN FUNCTIONS. M. V. Solodov and B. F.

PROXIMAL POINT ALGORITHMS INVOLVING FIXED POINT OF NONSPREADING-TYPE MULTIVALUED MAPPINGS IN HILBERT SPACES

Splitting methods for decomposing separable convex programs

The Relation Between Pseudonormality and Quasiregularity in Constrained Optimization 1

MAT 570 REAL ANALYSIS LECTURE NOTES. Contents. 1. Sets Functions Countability Axiom of choice Equivalence relations 9

Splitting Techniques in the Face of Huge Problem Sizes: Block-Coordinate and Block-Iterative Approaches

Convex Optimization Theory. Chapter 5 Exercises and Solutions: Extended Version

4y Springer NONLINEAR INTEGER PROGRAMMING

A Unified Analysis of Nonconvex Optimization Duality and Penalty Methods with General Augmenting Functions

Strong Convergence Theorem by a Hybrid Extragradient-like Approximation Method for Variational Inequalities and Fixed Point Problems

SOME COMMENTS ON WOLFE'S 'AWAY STEP'

ON GAP FUNCTIONS OF VARIATIONAL INEQUALITY IN A BANACH SPACE. Sangho Kum and Gue Myung Lee. 1. Introduction

Iterative Convex Optimization Algorithms; Part One: Using the Baillon Haddad Theorem

On the Convergence and O(1/N) Complexity of a Class of Nonlinear Proximal Point Algorithms for Monotonic Variational Inequalities

1. Gradient method. gradient method, first-order methods. quadratic bounds on convex functions. analysis of gradient method

Downloaded 12/13/16 to Redistribution subject to SIAM license or copyright; see

Coordinate Update Algorithm Short Course Subgradients and Subgradient Methods

AN INEXACT HYBRID GENERALIZED PROXIMAL POINT ALGORITHM AND SOME NEW RESULTS ON THE THEORY OF BREGMAN FUNCTIONS. May 14, 1998 (Revised March 12, 1999)

c 1998 Society for Industrial and Applied Mathematics

Merit functions and error bounds for generalized variational inequalities

1. Introduction The nonlinear complementarity problem (NCP) is to nd a point x 2 IR n such that hx; F (x)i = ; x 2 IR n + ; F (x) 2 IRn + ; where F is

Lecture 3. Optimization Problems and Iterative Algorithms

On the convexity of piecewise-defined functions

Sequential Unconstrained Minimization: A Survey

Math 273a: Optimization Subgradients of convex functions

Recent Developments of Alternating Direction Method of Multipliers with Multi-Block Variables

1. Nonlinear Equations. This lecture note excerpted parts from Michael Heath and Max Gunzburger. f(x) = 0

A globally and R-linearly convergent hybrid HS and PRP method and its inexact version with applications

Lecture 11 and 12: Penalty methods and augmented Lagrangian methods for nonlinear programming

Convex Optimization & Lagrange Duality

Numerisches Rechnen. (für Informatiker) M. Grepl P. Esser & G. Welper & L. Zhang. Institut für Geometrie und Praktische Mathematik RWTH Aachen

INERTIAL ACCELERATED ALGORITHMS FOR SOLVING SPLIT FEASIBILITY PROBLEMS. Yazheng Dang. Jie Sun. Honglei Xu

Optimality Conditions for Nonsmooth Convex Optimization

Functional Analysis. Franck Sueur Metric spaces Definitions Completeness Compactness Separability...

You should be able to...

CONSTRAINT QUALIFICATIONS, LAGRANGIAN DUALITY & SADDLE POINT OPTIMALITY CONDITIONS

Course 212: Academic Year Section 1: Metric Spaces

Quiz Discussion. IE417: Nonlinear Programming: Lecture 12. Motivation. Why do we care? Jeff Linderoth. 16th March 2006

Dual and primal-dual methods

Convergence Analysis of Perturbed Feasible Descent Methods 1

Solving generalized semi-infinite programs by reduction to simpler problems.

Optimization and Optimal Control in Banach Spaces

Algorithms for constrained local optimization

Algorithms for Nonsmooth Optimization

FIXED POINTS IN THE FAMILY OF CONVEX REPRESENTATIONS OF A MAXIMAL MONOTONE OPERATOR

TECHNICAL NOTE. A Finite Algorithm for Finding the Projection of a Point onto the Canonical Simplex of R"

Enhanced Fritz John Optimality Conditions and Sensitivity Analysis

The speed of Shor s R-algorithm

Continuous Sets and Non-Attaining Functionals in Reflexive Banach Spaces

A NOTE ON Q-ORDER OF CONVERGENCE

Optimality Conditions for Constrained Optimization

Sparse Optimization Lecture: Dual Methods, Part I

Convex Analysis and Optimization Chapter 2 Solutions

ITERATIVE SCHEMES FOR APPROXIMATING SOLUTIONS OF ACCRETIVE OPERATORS IN BANACH SPACES SHOJI KAMIMURA AND WATARU TAKAHASHI. Received December 14, 1999

On the relations between different duals assigned to composed optimization problems

Two-Step Iteration Scheme for Nonexpansive Mappings in Banach Space

On nonexpansive and accretive operators in Banach spaces

Near Equality, Near Convexity, Sums of Maximally Monotone Operators, and Averages of Firmly Nonexpansive Mappings

Transcription:

J. oflnequal. & Appl., 1998, Vol. 2, pp. 157-179 Reprints available directly from the publisher Photocopying permitted by license only (C) 1998 OPA (Overseas Publishers Association) Amsterdam B.V. Published under license under the Gordon and Breach Science Publishers imprint. Printed in Malaysia. A New Method for Nonsmooth Convex Optimization* Z. WEI a, L. QI a,l and J.R. BIRGE b a School of Mathematics, University of New South Wales, Sydney NSW 2052, Australia, b Department of Industrial and Operations Engineering, University of Michigan, Ann Arbor, M148109, USA (Received 3 March 1997) A new method for minimizing a proper closed convex function f is proposed and its convergence properties are studied. The convergence rate depends on both the growth speed off at minimizers and the choice of proximal parameters. An application of the method extends the corresponding results given by Kort and Bertsekas for proximal minimization algorithms to the case in which the iteration points are calculated approximately. In particular, it relaxes the convergence conditions of Rockafellar s results for the proximal point algorithm. Keywords." Proper closed convex function; Proximal minimization; Global and superlinear convergence AMS Subject Classification." Primary 90C47, Secondary 90C05 1 INTRODUCTION Consider the proximal minimization algorithm for solving the following convex minimization problem min{f(x)- x E R n} (1) * This material is based on work supported by the Australian Research Council and the National Science Foundation under Award Number DDM-9215921. Corresponding author. 157

158 Z. WEI et al. defined by - where f is a proper closed convex function defined in R n, R, the proximal parameters Ag and are positive constants, [1"11 denotes the standard Euclidean norm on (see Martinet [15] which - is the case -= based on [16]). The algorithm is globally convergent, and its local speed of convergence depends on: (i) The growth speed off at the minimizers; (ii) The choice of ); and (iii) The choice of (see [1,2,11]). For a survey of results of the proximal type minimization we refer the reader to [3-6,8-10,14,20-24], especially to [3,5,9,20]. Xk+l argmin{f(x) + [Ix-- xkll +1" x E Rn}, Rockafellar [20] made a nice extension for the proximal point algorithm. In his algorithm, the iteration points are calculated approximately. A question is: do corresponding results of Kort and Bertsekas [11] hold for the algorithm of Rockafellar [20] when it is applied to a proper closed convex function? In the next section, a new method for solving nonsmooth convex optimization problems is presented. This algorithm is globally convergent, and its local speed of convergence depends on both the growth speed at the minimizers and the choice of proximal parameters. These results are similar to those in [1,2,11]. Furthermore, a general proximal point algorithm introduced in [20] can be regarded as a special case of this method. In Section 3, an application of this algorithm relaxes a key condition of Rockafellar [20] for convergence of the proximal point algorithm. The condition, becomes k=l 6k < +cxz, lim sup{6k} < 1, where 6 is a parameter in the algorithm. In the same section, some special cases of the algorithm are also discussed. Some concluding remarks are in Section 4.

A NEW METHOD FOR NONSMOOTH CONVEX OPTIMIZATION 159 2 METHOD AND ITS CONVERGENCE PROPERTIES 2.1 The Algorithm and its Global Convergence Algorithm 1 Let q > be a given positive number. Let X E R be arbitrary. In general, given Xk R n, generate a proximal parameter tk > 0 and Xk + R such that there exists gk + Of(Xk + l) satisfying the following inequality: T g+ (x X+l) tl]g+ q (2) THEOREM 2.1 Let {xk}kl be any sequence generated by Algorithm 1. (i) If k=l t: +cx, then either f(xk)--- -x or liminfg Ilgkll 0. In the latter case,/f(xg}l is a bounded set, then every accumulation point of {xk}g l is an optimal point of(l). (ii) If inf{ t} > 0, then either f(x)- or gk O. In the latter case, every accumulation point of {Xg}k=l (if there is any) is an optimal point of (1). x Furthermore, Let X* be the set of minimizers off, f*= inf{f(x): Rn}. Then f(xk+l) f* < inf{llxk/l X* X* e X*)[t-;*llxk/ Xkl[] (q-l)- Proof From the subgradient inequality, f(xzc) > + By (2), we have f(x) >_ f(xc+l) + t[ig+ q. (3) This implies that f(xk+l)<_f(xlc). If the decreasing sequence {f( Xk)}=l tends to -oo, we are done. Otherwise, this sequence is bounded from below. By (3), k=l tl[g+ll q < +c.

160 Z. WEI et al. If Ilgll is bounded below by a positive number, then the above inequality implies that k tk < +o, which contradicts our assumptions in (i). Thus, lim inf IIgll 0. If{x} is a bounded set, then there exists an infinite index set K such that lim /gk --0. Without loss of generality, we may assume that lim/ x x*. Applying T f (x) >_ f(x) + gk (x xk) and letting k oe, k E K, we have that for all x Rn, f(x) >_f(x*), which implies that x* is a minimum point off. By f(xk + l) <f(xk), we deduce that every accumulation point of {x}l has the same optimal value, f(x*), and is, hence, also a minimum point of (1). We now prove the second conclusion. From (3), we have that either f(x) --+ -oe or gg 0 by using inf{tg} > 0. Let {x: k K} be any convergence subsequence of {xg}g=l and limke/xg x Then, lim Of(xg) C_ Of(x*) kgk yields that 0 Of(x*). _ So x* is an optimal point of (1), since f is a convex function. For each x* X* and any k, since f* f(x*) >_f(x+l) + gk+ (x* x+ ), by (2), we have f(x+) f* <_ IIg+llllx+ x*ll [[Xk+l X*[[[/ -1 [[Xk+l XkII](q-1) -1 Hence f(xk+l) f* <_ inf{llx+ x*ll" x* X*}[t-l[Ixk+l xll] (q-l)- The second conclusion follows. COROLLARY 2.1 Suppose that X* is nonempty, compact and {tk}ke=l is bounded away from zero. Let {xk} l be any sequence generated by

A NEW METHOD FOR NONSMOOTH CONVEX OPTIMIZATION 161 Algorithm 1. Then {Xk}kZ=l is bounded, f(xk) --f*and lim p(xk; X*) 0, k- x where p(x; X*) is the distance from x to X given by p(x;x*) min{llx- x*l[: x* X*}. (4) Proof By Theorems 8.4 and 8.7 of[18], the level sets offare bounded. On the other hand, for all k, f(x+l)>_f(x). Hence { X)= c is contained in some level sets off. Therefore, {xk}k= is a bounded set. By (ii) of Theorem 2.1 and the compactness of X*, we havef(xk)f* and (4). 2.2 Local Convergence We from now on assume that inf{tg} > 0. Like [3,11], we need the following key assumption which was used to analyze the convergence rate of the proximal minimization algorithm for quadratic as well as certain types of nonquadratic proximal terms (see [3,11]). Assumption I X* is nonempty and compact. Furthermore, there exist scalars a > 0, p > and 6 > 0 such that for any x with p(x; X*) < 6, LEMMA 2.1 and g Of(x), f(x)-f* > a(p(x;x*)) p. Suppose that (5) holds. Then for each x with p(x; X*) <_ Ilgll o/p[f(x) -f*](p-)/p. (6) Proof Since for any g Of(x) and any x* X, So we have f* f(x*) >_f(x) + gt(x* x). f*-f(x) > max{gt(x *- x)" x* X*} > max{-ilgllllx* xll" x* x*) _llgllp(x;x*).

162 Z. WEI et al. Hence, f (x) f* <_ I[gl[p(x; X*) [Igll f(x f* /p by using (5). This implies (6). LEMMA 2.2 Suppose that (5) holds, {xk}k= is generated by Algorithm 1. Then for all large k, f(x+l) --f* < f(x) f + t:cq/p[ f(xc+l f,]((q-1)(p-1)-l)/p (7) Proof Since p(xk; X*) 0 and f(x:) >f(x + 1), we have kl > such that for all k > kl, p(x; X*) <_ (5. Using g + E Of(x + ), we have T f(xlc) >_ f(x+l) + g+l (x/c by Lemma 2.1. Therefore, X+l) >_ f(x/c+l) + tllg/ q >_ f(xk+l) + tkoz q/p[ f(xk+l) _f,]q(p-l)/p f(xk) f* >_f(xk+l) f* -+-tkctq/p[f(xk+l)_f,]q(p-1)/p. This implies the conclusion by using f(xk + 1) >f*. THEORZM 2.2 Suppose that (5) holds, {Xk}k l is generated by Algorithm 1. Then f(x) tends to f* superlinearly under any one of the following conditions ((a), (b) or (c)): (a) (q- 1)(p- 1) < 1. (b) (q-1)(p-1) and t +oe. (If t: t, < +oc, then f(x) tends to f* linearly.) (c) (q- 1)(p- 1) > and lim t[ f (x:) f,]((q-1)p-q)/p nt_cx3" k-- Ifp- 1, then there exhsts k2 > kl such that x f can be determined within a finite number ofsteps. (8) X*, i.e., the minimizer of

A NEW METHOD FOR NONSMOOTH CONVEX OPTIMIZATION 163 Furthermore, if there ex&ts a scalar < / xz such that f(x) f* <_ (p(x;x*)) p for all x with p(x; X*) <_ 6, (9) then xk tends to X*superlinearly under any of the above conditions ((a), (b) or (c)). Proof The fact, f(xk + l) --+f*, implies that lim t:q/p[ f(x:+) f,]((q-1)(p-1)-l)/p -+-00, (10) k--+ if one of the conditions (a)-(b) holds. Then by Lemma 2.2. In case (b) with t lim f(xk+l) --f* 0 (11) f(x) f t, < +, from (7), we have lim f(x+i)-f* < f(x)-f* + t, ceq/p Hence f(x) tends to f* linearly. For (c), using f(x:+l f* f(x) -f* 1/{1 -I- oqlp[f(xk+l) --f*/f(xl,:) _f,]((q-1)(p-1)-l)lp t[ f (x:) f*l ((q-)(p-1)-)lp }, (11) follows. Recalling (5) and p(x; X*) <_ 5 for all k > kl, which has been used in the proof of Lemma 2.2, we have But from the proof of the Lemma 2.1, f(xk) f* >_ (p(x;x*)) p. (12) f (Xk) f* <_ IIgllp(x; (13)

164 Z. WEI et al. Thus, Ifp 1, since limk gk 0 and X* is a compact set, we have k2 > kl such that Xk2 E X* (otherwise a _< 0). We now prove the second conclusion. Let Ilxk--X*ll denote p(xk; X*). By (5) and (9) we have k3 > kl such that for all k > k3, al]xk X*ll p <_ f(xk) f* </3l[Xk x*ll p. (14) This yields thatf(xk) tends tof* superlinearly if and only if Xk tends to X* superlinearly. The second conclusion follows from (11). COROLLARY 2.2 Suppose that X* is singleton, i.e., X*= {x*}, {Xk}kZ=l is generated by Algorithm 1. (i) Assume that (5) and (9) hold. If there exist o > max{0, (q-1)p-q} and l > O, such that for all k, then if, for all k, tgllxg Xg_l 7-1, lim inf [IXk/l x*ll 0; k- Ilxg-- x*ll then lim I[Xk+l x*l[ O. k- Ilxk-x*ll (ii) Assume that (5) holds. If there exists, > 0 such that for all k, tk[ f(xk) f(xk-1)] q-1 _ 7"1,

A NEW METHOD FOR NONSMOOTH CONVEX OPTIMIZATION 165 then lim inf f(xk+) --f* O; k-cx f (Xk) f if, for all k, tk[ f(xk+l f (xk)] q-1 > then Furthermore, if (9) is true, then -lim f (xk+ f* O. f(x)-.f* lim inf IlXk+l X*ll 0 or holds respectively. -lim IlXk+ x*ll 0 IIx-x*ll Proof (i) It is very clear that tk--+-t-cx. Hence, we only need to consider case (c) of Theorem 2.2. If the conclusion is false, then we have a positive number -2 < + oc, such that for all k, Thus, we have

166 Z. WEI et al. This inequality and (5) yield that This last result implies that (8) holds. Hence xk tends to x* superlinearly by Theorem 2.2 and This is a contradiction. So the first conclusion of (i) follows. We now prove the second conclusion of (i). If the conclusion is false, then we have a positive number 7" 2 < -[- OQ and K such that for all k E K, I1>_. IIx x 11-2 Ilxz+,- x* Using this inequality, similarly to the proof of the first conclusion, we have for all k E K, t[ f(xk+l f,] ((q-1)p-q)/p [.1_[_7"21 > TlO((q_l)p_q)/p This and Lemma 2.2 imply that 1 [(q-1)p-q] Ilx+ Xkl[ --[(q-1)p-q] lim f (xl+ f* O. ki,: f(xk) f*

A NEW METHOD FOR NONSMOOTH CONVEX OPTIMIZATION 167 By (5) and (9), we have lim ]]x:+l- x*[] [OZ] lip So 7"2 + OO. Again, this is a contradiction. This proves the second conclusion of (i). (ii) Assume that (5) holds. We prove the first conclusion. Using f(xk) --f* f(xl) --f(x-l) >- + (f(x-l) --f*)l(f(x) --f*) and tl[ f (xl) f,l((q-1)p-q)/p ---tk[f(xk)--f(xk-1)]q-1 [f(xk X [f(xk)--f(xk_l)] -qlp, )f(xk)-f*_f(xk_l) 1 ((q-1)p-q)/p similarly to the proof of (i), if the conclusion is false, we have lim tk[f(xk) f,]((q-1)p-q)/p -P-O0. k---+ Using conclusion (c) of Theorem 2.2, we have f(x) tends to f* superlinearly which contradicts our assumption that the conclusion is false. Hence, the conclusion follows. Similarly to the proof of the first conclusion of (ii) and the proof of the second conclusion of (i), we can prove the second conclusion of (ii). If (9) holds, by IIx + x*ll f(xk+l) --f* --X -; <- f x f ]1 /P

168 Z. WEI et al. and the conclusions which we have just proved, the last conclusion of the corollary follows. Using the results of this subsection, we may give some special cases of Algorithm according to the choice of tk. One such special case follows. Algorithm 1.1 In Algorithm 1, generate xk+ E Of(x+ 1) satisfying the following inequality: or T (xz Ilxz IIg+ q T gk+l (Xk Xk+l) >_ [f(xk) f(xk-1)] 1-q[Igk+l q In general, we can choose q 2. It is worth to note that, if we choose tk as following: tz r ilgllq where r > is a constant, then from (2), we can deduce that for all k, tk > r k. Hence tk 4-. Hence if p--q= 2, then, by Theorem 2.2, f(xk) tends to f* superlinearly under the condition (5). Kort and Bertsekas in their novel paper [11] presented a combined primal-dual and penalty method for solving constrained convex programming. It essentially is a proximal point algorithm (see [2,19]). Hence, from the results of [1,2,11], under (5), the convergence rate of the ordinary proximal minimization algorithm is shown to depend on the growth speed of f at the minimizers as well as the proximal parameters Ak and 7-. More precisely, if 7-1, the ordinary minimization algorithm has finite, superlinear and linear convergence rate depending on whether p= 1, <p < 2 and p 2, respectively. The convergence rate is also superlinear if p--2 and Ak oo. In the case where the proximal term has a growth rate 7- > other than quadratic (7-1), the convergence rate is superlinear if < p < 7-4- even in the case where p >_ 2 (see [1,2,11]). From Theorem 2.2 (in the case of q + 1/7-), we obtain corresponding results for our algorithm.

A NEW METHOD FOR NONSMOOTH CONVEX OPTIMIZATION 169 3 APPLICATION TO THE PROXIMAL POINT METHOD In [20], Rockafellar presented two general proximal point algorithms for solving (1). One of them is as follows. Algorithm A For Xk, generate xk+l E R satisfying where ck > 0, and k=l Ck 6k <, (16) &(x) of(),) + (),- x) (17) In the following, we present a generalization of Algorithm A. In the new algorithm, we relax (16) to the following condition: lim sup{6k} < 1. (18) The new algorithm possesses a convergence rate similar to the rate for Algorithm A. The relaxation may be important in practice. Algorithm B For any given 7- > 0, let xl E R n. In general, for xk R n, generate xk + R" satisfying p(0; S(x+)) IIx+ xll, (19) -l Sk(X) Of(x) +- IIx- x:ll (x- x:). Ck (20) The following theorem proves that Algorithm B is a special case of Algorithm 1, hence the conclusions oftheorem 2.2 hold for Algorithm B. It is very clear that the ordinary proximal minimization described in the introduction of this paper is a special case of Algorithm B. Therefore,

170 Z. WEI et al. we extended the corresponding results [1,2,11], for ordinary proximal minimization algorithms to the case when the iteration points are calculated approximately. THEOREM 3.1 Suppose that xk + is generated by Algorithm B, andgk + is a vector in Of(x+ 1) such that the norm ofg+ + 1/cllx+ xl[ *-1 (x+l-x) is not bigger than the right-hand side of (19). Then for all large k, (xk + 1, g + 1) satisfies (2) with q-- + 1/- and 6k / (1 + 6k)1+1/7- Ck Proof Without loss of generality, we may assume that By (19), (21) and (20), we have sup{6/c} < 1. (21) 7"--1 g+ +-IIx+ xll (x+ x) Ck (22) The inequality, x/c) (x/c+l x/c) 7"-1 + +-IIx+ xll (x+ Ck 7"-1 -< g/c+l +--]]x/c+l xgll (Xk+l Xk) IlXk+l Xkl], Ck and (22) imply that T gk+l (Xk+l Xk) <_ Ck Therefore, gt /C+I (X/C Xk+l) (23)

A NEW METHOD FOR NONSMOOTH CONVEX OPTIMIZATION 171 On the other hand, using (22), we have Ilgk+l _ IlXk+l l+6k Ck xkll. (24) Now, the conclusion of the theorem follows (23) and (24). By the above results, we may state global convergence and convergence rate results for Algorithm B. Noticing that tk + oo if and only if c + oo, we only need to change q to + 1/- in Theorem 2.2 and Corollary 2.2. We do not go into this direction in details. We are more interested in the special case in which -= 1. In the following analysis, we always assume -= 1. THEOREM 3.2 Let {Xk}k=l be any sequence generated by Algorithm B oo with {ck}k- bounded awayfrom zero. Suppose that { Xk}k= oo is bounded. Then every accumulation point of {xk}k l is an optimal solution of (1). Furthermore, if x* is an optimal solution of (1), then f(xk+l)-f(x*) <_ IIx +a- x*ll X [[Igk+l -t-" -1 (Xk+ Xk)] nl -111Xk+l Proof Ck By Theorem 2.1, we have the first conclusion. The proof of the first inequality in the second conclusion is similar to the proof of Theorem 4 in [18]. The second inequality follows (22). For rates of convergence in optimization methods, a key condition is called the quadratic growth at x* off, i.e., in the neighbourhood of x*, O(x*;,1)- (x: IIx-x*ll < f(x) >f(x*) + cllx- x*ll 2 for all x 60(x*;r), (25) where c >0. Under (25) and some other additional assumptions (ek --+ + oc and k--1 k < +oe), Rockafellar proved the superlinear convergence of Algorithm A. The conditions in (16) and ck ---, + oe are, however, quite strict. Very large ck and very small k, in fact, imply that the calculation of xk+l can be almost as difficult as the original minimization problem (0 E Of(xk + 1)). Moreover, the condition (25) is

172 Z. WEI et al. false for a large class convex functions (for example,f(x) I[x[[ p, p > 2). So it is necessary to discuss the possibility of relaxation of the requirements for ok, 6k and (25). For given xe R and S c_ R + =(0, +oc), a collection of proper closed convex functions (PCCF) is defined by CC(x; S) {f" f PCCF and there exists p G S, 0 < ap(f; x) <_ p f x) < o( }, where and inf{ ap(f;x)-limy._.,x f(y)[[y_-f(x) x[[ } p f(y)> f(x) (26) /3p(f;x)-lyim x sup{ f(y)[ly--f(x) x[i }. p f(y)> f(x) (27) We now have the following theorem by Theorems 2.2 and 3.2. THEOREM 3.3 Suppose that the conditions of Theorem 3.2 hold, {x:): is generated by Algorithm B and x x*. Then f(x) tends to f(x*) superlinearly if one of the following conditions ((a), (b) or (Cl)) holds: (al) p (1,2), Cp(f; x*) > 0; (bl) p 2, Ctz(f; x*) > 0 and c: (Cl) p (2, + oc), ap(f x*) > O, and limk_ ck[ f(x) f(x* )](P-2)/P If c(f, x*) > O, finite number of steps. then the minimizer off can be determined after a Furthermore, x tends to x*superlinearly if one of the following conditions ((a2) or (b2) or (c2)) holds: (a2) fg CC (x*; (1,2)); (b2) fe CC (x*; {2}) and ck ---, + o; (c2) f CC(x*; (2, + cx)) andlim_, c:[f(xg) f(x*)] (p-z)/p

A NEW METHOD FOR NONSMOOTH CONVEX OPTIMIZATION 173 COROLLARY 3.1 Assume that f has a unique minimum, say x*. Let x { k}k=l be a sequence generated by Algorithm B. (i) IffE CC (x*; {p}) and there exist 7-3 > max{0, p-2}, and 7"4 > 0, such that for all k - IIx- then xk tends to the unique minima off, i.e., Xk-- x*, and iffor all k, then liminf Ilx+l x l[ x*ll - [Ix+l x*[i lim Ilx- x*ll (ii) Assume oo(f; x*) > O. If there exists 7"1, > 0 such that for all k, then if, for all k, then O; ck [f(x) --f(xk-1)] _> 7"4, f(xk+l) --f(x*) ko f(xk) f(x*) =0; lim inf c [f(x+l) -f(x)] > 7"4,, lim f(x+l) f(x*) O. -o f(x) f(x*)

174 Z. WEI et al. Furthermore, iffe CC (x*;p), then lim inf [IXk+l x* Ilia- x*[i or lim Ilxk+ x*[i 0 0 holds respectively. Proof Using the assumption of this theorem, we can deduce that { X }g-1 is bounded (see the proof of Corollary 2.1). By the boundedness of {Xk}kC) l and Theorem 3.2, we have that xk tends to x*, the unique minimum off, i.e., xk x*. The conclusions follow Corollary 2.2 now. It should be noted that the condition oz2(f; X*)> 0 is weaker than (25). In fact, oz2(f;x*)>0 does not imply the uniqueness of the minimizers off, but the following proposition holds. If (25) holds, then X*--{x*}. Let Y* be the set of all accumulation points of {xg}gl generated by PROPOSITION 3.1 Algorithm B. Then we have the following result. LEMMA 3.1 Then Suppose that Y* 7 ( has an isolated point and oe {xk}= is a convergent sequence. lim IlXk+l Xkl] O. (28) k-- cx Proof By the assumption on Y*, there is a point y* E Y* such that it is an isolation point in Y*. Then there exist r2 >0 and O(y*;r2)= {x: Ilx-Y*ll<r} such that Y*fqO(y*;rz)\{y*}=. Let K(y*)= {k: x O(y*; r2)}. Then K(y*) is an infinite set and lim x y*. kk(y*)

A NEW METHOD FOR NONSMOOTH CONVEX OPTIMIZATION 175 Furthermore, let K(y*) {kj: jl < j2 implies kj, <_ kj2 } and N be the set of natural numbers. IfN\K(y*) is an infinite set, then there exists an infinite set KI(y*) c_ K(y*) such that for all kie KI(y*), ki+ lk(y*). Let ll be the smallest number in N\K(y*). In general, suppose that we have {/1, 12,...,/j}. Then let/j+l be the smallest number in N\K(y*)\{ll, ll + 1,...,/j,/j+ 1}. By induction, we have an infinite set {/i: ien}. It is clear that Kl(y*)={ki=l-1: N} is the desired set. Since limki o xki y*, we have limk; Xk+ y* by (28). This is impossible from the definition of K(y*) and k+l _K(y*). So N\K(y*) is a finite set. This implies that limk_ Xk y* by the definition of K(y*). Indeed, if both L(f;xl)= {x: f(x)<_ f(xl)} and {Ck} are bounded, then Theorem 3.2 implies that limk_ [IXk+l Xk[I 0. In addition, if the condition (a2) in Theorem 3.3 holds and Y* is not a dense subset of R n, then Xk tends to one of the minimizers off superlinearly. In this sense, (25) relaxed by a2(f; x*) > 0 is a useful consideration for those problems with nonunique minima. Example 3.1 Let f(x) LEC, and f(x*) 0. For any given number < -(p 2)/p, choose c>0 ifl <_p<2, c- [/(x)]t ifp_>2 (29) in Algorithm B, thenf(xk) tends tof(x*) superlinearly if Cp(f; x*) > O. Furthermore, xk tends to x* superlinearly iff CC(x*; {p}). In Examples 3.2-3.4, we assume that c satisfies (29) in Algorithm B and that xk tends to x*. Example 3.2 Suppose that we wish to find a zero point of a function h(x)lec. Then we can choose f(x)=max{o,h(x)} and start Algorithm B from the point xl, h(xl)> 0. Since f(x)>f(x+ 1)-- f(x*), also h(xk) > h(x + 1)-- h(x*). Thus, h(x*)= 0. Furthermore, we have the following proposition. PROPOSITION 3.2 Ifcp(h; x*) > O, then h(xk) tends to 0 superlinearly. If h CC(x*; {p}), then x tends to x*superlinearly. Proof Since Cp(f; x*) Cp(h; x*) and 3p(J x*) p(h; x*), the conclusions follow from Theorem 3.3.

176 Z. WEI et al. Example 3.3 An important special case of Example 3.1 is called the inequality feasibility problem which is the most fundamental in convex mathematical programming. Let f. (i E I a finite index set) be convex functions. Find x E R such that j(x) < 0 for all I. Let f(x) max{o, f,.(x): I}. Then f is a convex function, f(x) > O. Furthermore f(x)-o if and only ifj(x)<_o for all iei. Hence, for the inequality feasibility problem, we have a method to solve it quickly if there exists p(0,+oc) such that for all EI(x*)= {i:f.(x*)=o}, f. CC(x*; {p}). More precisely, we have the following proposition. PROPOSITION 3.3 Iffor all e I(x*), ap(j}; x*) > 0, thenf(xk) tends to 0 superlinearly; iffor all ii(x*), f. CC(x*; {p}), then xk tends to x* superlinearly. Proof First of all, we note that I(x*)O from f(x)>0 and f(x)f(x*)=o. On the other hand, for q I(x*), since f-(x*)< 0, so there exists r2 > 0, such that for all x E O(x*;r2), f-(x) < 0. This implies that for i I(x*), Hence, if for all f.(x) f(x) for all x O(x*;r2). I(x*), Cp(J; x*) > O, then ap(f; x*) lm, { inf f(x) f(x*) x x [Ix-x lip > lim inf { fi(x) j}(x*) f(x)>f(x*) I(x*), f.(x) --f(x) > 0 x x* [[x x*[[p > min{ap f.; x*)" I(x*) } > O. Therefore, the first conclusion follows Theorem 3.3.

A NEW METHOD FOR NONSMOOTH CONVEX OPTIMIZATION 177 By the same approach, if for all iei(x*),f.e CC(x*; {p}), we have /3p(f;x*) < The second conclusion follows Theorem 3.3. Example 3.4 Another case of Example 3.1 is finding the zero points of a given function F(x) R Rm. More precisely, find x* Rn, such that F(x*) 0 if f(x)-1/2f(x)rf(x) is a convex function (example: F(x)=Ax + b, A R n m). For this problem, we have the following result. PROPOSITION 3.4 Suppose that f is a convex function from R --+ R, lim inf( [IF(x)! [2 x-,x, p" I[x- x[i F(x) O} > O. Then IIF(xz)ll tends to 0 superlinearly. Furthermore, if the following condition x- x, IIx x-ii F(x) # 0 < +cx is added, then xk tends to x* superl&early. 4 CONCLUDING REMARKS The original motivation for the Moreau-Yosida regularization was to solve ill-conditioned systems of linear equations. It is noticed that Algorithm B always converges superlinearly when p < 2, ck constant, < and, for a large class of functions, a smaller p ( < 2) implies greater ill-conditioning. Comparing with other methods (for example, Newton or quasi-newton methods), the proximal point method takes advantage of solving ill-conditioned problems (also see [13,17,7] for more details on this topic). The benefit of Algorithm B can be explained by using the univariate functionf(x)-1/2]xl3. For this function, the algorithm given in [12] may

178 Z. WEI et al. not converge superlinearly when starting from X > 0 (see Proposition 15 of [12]). If we choose Ck--[f(xk)] (/<--1/2 is a constant) and apply Algorithm B, then { Xk}k=l converges to the solution 0 superlinearly by the results of Example 3.1. Also, for this function, since (25) does not hold, the superlinear convergence results of [20] cannot be applied. The results of Corollaries 2.2 and 3.1 are important for constructing a good algorithm in the spirit of the variable metric methods (see [12]). In [12], the authors gave a way to construct Ck such that Ck +. They proved the algorithm possesses superlinear convergence under (25) when f is an LC function (in this case, fe CC(x*; {2})). Since Algorithm 13 of[12] is a special case of Algorithm B (if for all k, 6k 0), Xk tends to x* superlinearly from Theorem 3.3. On the other hand, if we let Ck IlXk- Xk-l][ - r3 in Algorithm B, then we can expect that Xk tends to x* superlinearly even if f is not a smooth function using Corollary 3.1. Hence, the results of Corollary 3.1 provide a rule to construct the variable matrices based on the ideas of [12] (for penaltytype bundle algorithms) and quasi-newton formulae. It is not very hard to give a partial proximal minimization algorithm (see [3]) by a way similar to Algorithm 1. In fact, in Algorithm 1, if we let gk + be the vector with components zero for E I (where I is a subset of the index set { 1,..., n}, see [3]) then one of the special cases of the PPM algorithm given by [3] is contained in the Algorithm 1. The reader can go into this direction in details by following [3]. The essential content of this paper is a theoretical investigation of algorithms for nonsmooth convex optimization. We extended convergence results from those previously given but one question remains. How can one solve (2) effectively iffis a nonsmooth function? Acknowledgement The authors are thankful to the referee for his or her helpful comments. References [1] D.P. Bertsekas, Necessary and sufficient conditions for a penalty method to be exact, Mathematical Programming, 9 (1975), 87-99. [2] D.P. Bertsekas, Constrained Optimization andlagrange Multiplier Methods, Academic Press, New York, 1982.

A NEW METHOD FOR NONSMOOTH CONVEX OPTIMIZATION 179 [3] D.P. Bertsekas and P. Tseng, Partial proximal minimization algorithms for convex programming, SlAM Journal on Optimization, 4 (1994), 551-572. [4] J.R. Birge, L. Qi and Z. Wei, Convergence analysis of some methods for minimizing a nonsmooth convex function, Journal of Optimization Theory and Applications, to appear. [5] G. Chen and M. Teboulle, A proximal-based decomposition method for convex minimization problems, Mathematical Programming, 64 (1994), 81 101. [6] J. Eckstein and D.P. Bertsekas, On the Douglas-Rachford splitting method and the proximal point algorithm for maximal monotone operators. Mathematical Programming, 55 (1992), 293-318. [7] M. Fukushima and L. Qi, A globally and superlinearly convergent algorithm for nonsmooth convex minimization, SlAM Journal on Optimization, 6 (1996), 1106-1120. [8] O. Gtiler, On the convergence of the proximal point algorithm for convex minimization, SIAM Journal on Control and Optimization, 29 (1991), 403-419. [9] O. Gtiler, New Proximal point algorithms for convex minimization, SlAM Journal on Optimization, 4 (1992), 649-664. [10] C.D. Ha, A generalization of the proximal point algorithm, SlAM Journal on Control and Optimization, 28 (1990), 503-512. [11] B.W. Kort and D.P. Bertsekas, Combined primal-dual and penalty methods for convex programming, SIAM Journal on Control and Optimization, 2 (1976), 268-294. [12] C. Lemar6chal and C. Sagastizfibal, An approach to variable metric methods, in: Systems Modelling and Optimization, J. Henry and J.-P. Yvon, Eds., no. 197 in Lecture Notes in Control and Information Sciences, Springer-Verlag, 1994. [13] C. Lemarchal and C. Sagastizfibal, Practical aspects of the Moreau-Yosida regularization I: Theoretical properties, SlAM Journal on Optimization, to appear. [14] F.J. Luque, Asymptotic convergence analysis of the proximal point algorithm, SlAM Journal on Control and Optimization, 22 (1984), 277-293. [15] B. Martinet, Perturbation des methods d optimisation: Application, RAIRO Anal. France, 93(12) (1978), 153-171. [16] J.J. Moreau, Proximite et dualite dans un espace hilbrtien, Bulletin de la Societe Mathematique de France, 93 (1965), 273-299. [17] L. Qi, Second-order analysis of the Moreau-Yosida regularization of a convex function, AMR 94/20, Applied Mathematics Report, University of New South Wales, Revised, 1995. [18] R.T. Rockafellar, Convex Analysis, Princeton University Press, Princeton, NJ, 1970. [19] R.T. Rockafellar, A dual approach to solving nonlinear programming problems by unconstrained minimization, Mathematical Programming, 5; (1973), 354-373. [20] R.T. Rockafellar, Monotone operators and proximal point algorithm, SlAM Journal on Control and Optimization, 14 (1976), 877-898. [21] R.T. Rockafellar, Augmented Lagrangians and applications of the proximal point algorithm in convex programming, Mathematics of Operations Research, 1 (1976), 97-116. [22] C. Zhu, Modified proximal point algorithm for extended linear-quadratic programming, Computational Optimization and Applications, 1 (1992), 185-206. [23] C. Zhu, Asymptotic convergence analysis of the forward-backward splitting algorithm, Mathematics of Operations Research, 20 (1995), 449-464. [24] C. Zhu, Asymptotic convergence analysis of some inexact proximal point algorithms for minimization, SlAM Journal on Optimization, to appear.