A UNIFIED CHARACTERIZATION OF PROXIMAL ALGORITHMS VIA THE CONJUGATE OF REGULARIZATION TERM

Size: px
Start display at page:

Download "A UNIFIED CHARACTERIZATION OF PROXIMAL ALGORITHMS VIA THE CONJUGATE OF REGULARIZATION TERM"

Transcription

1 A UNIFIED CHARACTERIZATION OF PROXIMAL ALGORITHMS VIA THE CONJUGATE OF REGULARIZATION TERM OUHEI HARADA September 19, 2018 Abstract. We consider proximal algorithms with generalized regularization term. General requirements for the conjugate of regularization term are proposed to make the corresponding proximal algorithm globally converge. Applying the proposed technique to the case of ϕ-divergence, we have proved that ϕ-divergence with nonsmooth kernel function is acceptable as a regularization term. ey words. proximal point method, ϕ-divergence AMS subject classifications. 90C22, 90C25 1. Introduction In this paper, we are interested in the problem (P): minimize f(x), x X, where f : R n R is a convex, inf-compact and possibly nonsmooth function and X is a closed convex subset of R n. In order to solve the problem, a typical approach is a proximal algorithm, where the following subproblems, minimize f(x k + d) + 1 2t k d 2, x k + d X are iteratively solved with a positive sequence {t k }. Let d k be an optimal solution of the subproblem, the point x k is updated to x k + d k. The procedure continues until d k become sufficiently small. The term d 2 /2t is called a regularization term. In this paper, we try to generalize the regularization term as a sequence of positive convex functions {P k }, and iteratively solve the following subproblems. minimize f(x k + d) + P k (d), x k + d X. The extension of the regularization term in proximal algorithms has been widely researched, although some of them are given in the context of Augmented Lagrangian methods or bundle methods. Power regularization, P k (d) = d α /t k α, α > 1, has been considered by [13]. Indicator function, P k (d) = I B(0,tk )(d), is covered by [10]. Variable metric function, P k (d) = d T M k d/2, M k 0, has been considered NTTDATA Mathematical Systems Inc, 1F, Shinanomachi Renga-an, 35, Shinanomachi, Shinjuku-ku, Tokyo, JAPAN, harada@msi.co.jp. 1

2 2 OUHEI HARADA by [4]. The application of Bregman distance as a regularization term has been proposed by [5, 6]. The extension of nonsmooth Bregman function has been considered by [14]. ϕ-divergence is another candidate of regularization term. The application of ϕ-divergence as a regularization term has been proposed by [7, 12, 3, 17, 19]. A review of proximal point methods and Augmented Lagrangian methods with Bregman distance and ϕ-divergence are given by [11]. A unified approach which covers both Bregman distance and ϕ-divergence has been proposed by [2]. Quasi distance, where P k is not assumed to be convex, has been considered by [16]. To prove the global convergence of the algorithm, we focus on the structure of Pk, the conjugate function of P k. This approach has originally given by Auslender [1]. However, they assumed that P k is invariable with respect to k and P k is of Legendre type [18, Section 26]. This approach has been essentially extended and developed by Frangioni [8] in the context of bundle methods. They remove the both assumptions mentioned above. However, they assume that P k = Q t k, where Q t is a parametric function with respect to t which implies that the effect of the index k is limited to t k. We further extend their approach and show that more generalized P k is acceptable to proximal algorithms. Although most of the convergence properties for generalized P k proposed in this paper has been already researched independently, our result is meaningful because we can see the convergence properties from a unified viewpoint of Pk. Furthermore, in terms of ϕ-divergence based regularization term, we extend the previous result of [3]. We have proved that non-differentiable ϕ- divergence kernel is acceptable as a regularization term in proximal algorithms. This paper is organized as follows. In section 2, we introduce the notation used in the paper and briefly give some preliminary results which are necessary for the later sections. In section 3, we give the fundamental requirements for the regularization term to make the corresponding generalized proximal algorithm converges globally. In particular, a new requirement is given from a viewpoint of Pk, and it plays an important role. We also introduce some examples of regularization terms and confirm their convergence. In section 4, we apply the result of section 3 to Bregman distance. Unfortunately, our approach is not useful in this case, and no new results are proposed in this section. In section 5, we apply the result of section 3 to ϕ- divergence. In this case, our approach works well and we can extend the previous results. It is proved that the kernel function ϕ is not necessary to be smooth except for its minimum point. 2. Preliminaries We use the following standard notation which are mainly derived from [18], [9], and [10]. For a closed convex set S, I S denotes the indicator function of S. For a proper convex and lower semicontinuous(lsc) function F : R n R {+ }, its effective domain is defined by dom F = { x F (x) < + }. For a given set S, its interior, relative interior, and bound are denoted by int(s), ri(s), and bd(s) respectively. A proper convex function F is said to be essentially smooth if F is smooth on int(dom F ) and F (x k ) + whenever x k x such that x k int(dom (F )) and x bd(dom (F )). A proper convex function F is said to be of Legendre type if both F and F are essentially smooth. For all ɛ 0, the ɛ-subdifferential of a function F at x is defined by ɛ F (x) = { s R n F (y) F (x) + s, y x ɛ, y R n }. For all α R, the level set of F with respect to α is defined by L α (F ) = { x F (x) α }. We say that a function F

3 A UNIFIED CHARACTERIZATION OF PROXIMAL ALGORITHMS VIA THE CONJUGATE OF REGULARIZATION TERM3 is inf-compact if L α (F ) is inf-compact for every α R. The domain of F = 0 F is defined by dom F = { x F (x) }. The set of n-vectors with nonnegative(positive) components is denoted by R n +(R n ++). B(c, r) = { x x c r } denotes the Euqulidian ball of radius r with center c. A recession function of f with respect to direction d 0 is defined by f (d) = f(x + td) f(d) lim = sup t + t t 0 f(x + td) f(d). t The distance between a point x and a convex set S is denoted by dist(x, S) = inf s S x s. The projection point of x onto a closed cone is denoted by p (x) = arg min s x s. Following lemmas are necessary for the later sections. Lemma 2.1. [10, Chapter X, Corollary 1.4.4] Let f be a proper, lsc, and convex function, the following holds. z f(x) x f (z) f(x) + f (z) = x, z. Lemma 2.2. [18, Theorem 23.8] [10, Chapter XI, Theorem 3.1.1] Let f 1 and f 2 be a proper convex function on R n and ri(dom f 1 ) ri(dom f 2 ), then we have (f 1 + f 2 )(x) = f 1 (x) + f 2 (x), x. Lemma 2.3. Let f be a proper, lsc, and convex function such that f(0) = 0, then we have L ɛ (f ) = ɛ f(0) for all ɛ 0. Proof. By definition of ɛ-subdifferential and conjugate function, we have that z ɛ f(0) f(x) z, x ɛ, x ɛ f (z) z L ɛ (f ). We have completed the proof. Lemma 2.4. [9, Chapter III, Theorem 3.2.5] Let be a closed convex cone and = { z x, z 0, x } be its polar cone, then for all x R n, we have that x = p (x) + p (x), p (x), p (x) = Fundamentals In this section, we introduce fundamental requirements for regularization term P k to make the corresponding proximal algorithm globally converge. We firstly propose the two basic requirements in the following: (P1) P k is convex, lower semicontinuous, and positive. (P2) P k (0) = 0. From (P1) and (P2), following symmetric conditions for Pk clearly hold. (P*1) Pk is convex, lower semicontinuous, and positive. (P*2) Pk (0) = 0. In each iteration, we minimize f(x k + d) + P k (d) on x k + d X and update x k by (3.1) x k+1 = x k + d k, where d k is a minimizer of f(x k + d) + P k (d). To make the algorithm well-defined, we assume that

4 4 OUHEI HARADA (PF1) ri(dom [f + I X ]) ri(dom P k ). We use the symbol (PF) to stress that the assumption depends not only on P k but also f. Therefore, by means of Lemma 2.2, there exists a sequence {z k } which satisfies the following: (3.2) (3.3) z k [f + I X ](x k + d k ), z k P k (d k ). Proposition 3.1. Let {x k }, {z k } and {d k } be sequences which enjoy (3.1), (3.2), and (3.3), and {P k } be sequences of positive functions which enjoy (P1), (P2), and (PF), then we have the following: (i) There exists a convergent subsequece {x k } k such that x k x. (ii) k=0 ( Pk (d k ) + P k ( zk ) ) < +, in particular, P k ( zk ) 0. Proof. In view of (3.2), we have [f + I X ](x k ) [f + I X ](x k + d k ) + z k, x k (x k + d k ). On the other hand, in view of (3.3) and Lemma 2.1, we have Therefore, we have that (3.4) P k (d k ) + P k ( z k ) = d k, z k. [f + I X ](x k ) [f + I X ](x k + d k ) + P k (d k ) + P k ( z k ). In view of (P2) and (P*2), the sequence {[f +I X ](x k )} is a nonincreasing sequence. Therefore, we have that x k L [f+ix ](x 0 )(f + I X ) for all k, and hence there exists a convergent subsequence {x k } k such that x k x. We have proved item (i). Again, in view of (P2), (P*2), and (3.4), it follows that there exists a limit f R such that [f + I X ](x k ) f. Thus, by summing up (3.4) from k = 0 to, we have [f + I X ](x 0 ) f ( + Pk (d k ) + Pk ( z k ) ). k=0 Furthermore, by means of (P2) and (P*2), we have Pk ( zk ) 0. Therefore, we have proved item (ii). As pointed out by Frangioni [8], the role of P k ( zk ) has been overlooked. Many researches so far has been proved the convergence properties not via P k ( zk ) 0, but via P k (d k ) 0. We consider the difference in the latter of the section. Lemma 3.2. Let {x k } and {z k } be sequences which satisfies z k [f + I X ](x k+1 ), x k x and z k 0, then x minimizes f on X. Proof. x k 0, z k 0, and z k [f + I X ](x k+1 ) implies that 0 [f + I X ](x ); Indeed, z k f(x k+1 ) is equivalent to [f + I X ](x) [f + I X ](x k+1 ) + z k, x x k+1, x R n. Take k, we have that [f + I X ](x) [f + I X ](x ) + 0, x x, x R n. Therefore, we have completed the proof.

5 A UNIFIED CHARACTERIZATION OF PROXIMAL ALGORITHMS VIA THE CONJUGATE OF REGULARIZATION TERM5 Proposition 3.1 (i) and Lemma 3.2 suggest that any limit point x generalized by the proximal algorithm is an optimal solution of the original problem (P) if z k 0 holds. Therefore, by means of Proposition 3.1 (i), what remains is how to reach z k 0 from Pk ( zk ) 0. In this sense, the structure of Pk plays the key role. In addition to (P1) and (P2), we assume the following new requirement to prove z k 0. (P3) lim inf k + Pk (z) > 0 for all z 0. Following lemma clarifies the role of the assumption (P3). Lemma 3.3. Let F k : R n R + be a sequence of lsc convex functions such that F k (0) = 0 for all k, L 0 (F ) = {0} where F (y) = lim inf k + F k (y) and there exists a sequence {y k } such that F k (y k ) 0, then y k 0. Proof. Assume to achieve at a contradiction that there exists a ɛ > 0 and infinite subsequence {y k } k 0 such that y k / B(0, ɛ) for all k 0. For sufficiently large k, we have that ( ) y F k (y k k ) F k y k ɛ inf F k(y) inf F (y). y =ɛ y =ɛ First inequality follows from the convexity of F k and F k (0) = 0. Since L 0 (F ) = {0}, inf y =ɛ F (y) > 0. Therefore, F k (y k ) does not converges to 0 on k 0, and this is a contradiction. We have completed the proof. The structure of the assumption (P3) is a bit confusing. If P k is invariable with respect to k, (P3) is equivalent to L 0 (P ) = {0} where P = P k for all k. However, the following assumption is not equivalent to (P3). (P3A) L 0 (Pk ) = {0} for all k. For example, Pk (z) = tk z 2 /2 with t k 0 satisfies (P3A), but not satisfies (P3). Intuitively, we need to avoid Pk to become flat around 0 for {P k} to enjoy (P3). One way to meet (P3) is to add the following strong assumption. (P3B) There exists a k such that P k Pk for all k. With this assumption, we can confirm (P3) holds since P k (z) lim inf k + Pk (z) and L 0 (P k ) = {0}. If P k depends only on t k, there exists a positive, lsc, convex, and parametric function Q t with respect to t such that P k = Q t k. In this case, (P3B) can be reduced to the following: (P3C) Q t Q τ if t τ and there exists a t L > 0 such that t L t k for all k. By an easy algebra, Q t Q τ is equivalent to Q t Q τ. Except for the requirements of {t k }, (P3C) is equivalent to [8, (P4)(P4*) Section 3]. Theorem 3.4. Let {x k } and {z k } be sequences which enjoy (3.1), (3.2), and (3.3), then x x and 0 [f + I X ](x ), if (P1), (P2), (P3) and (PF1) hold. Proof. In view of Proposition 3.1 (i) and Lemma 3.2, we have that 0 [f +I X ](x ) if z k 0 holds. On the other hand, in view of Proposition 3.1 (ii) and Lemma 3.3, we have z k 0. Therefore, we have completed the proof. Let us review some regularization terms of which convergence properties can be confirmed in our approach. In other words, in order to apply Theorem 3.4, let us confirm that whether (P1), (P2), (PF1), and (P3) hold or not in every case.

6 6 OUHEI HARADA 1. Power regularization. This type of regularization term is defined by P k (d) = 1 t k α d α, α > 1, where {t k } is a positive sequence such that 0 < t L t k for all k. In this case, (P1),(P2), and (PF1) are obviously hold. By means of [18], we have P k (z) = t k z β /β where 1/α + 1/β = 1. Therefore, (P3A) and (P3C) hold, and hence (P3) holds. However if α = 1, we have P k (z) = I B(0,t k )(z). It implies that (P3) does not hold. 2. Trust region. This type of regularization term is defined by P k (d) = I B(0,tk )(d). The confirmation of (P1), (P2), and (PF1) are obvious if there exists a t L > 0 such that t k t L. In this casem, P k (z) = z /tk. Thus, (P3A) and (P3C) hold, and hence (P3) hold. 3. Variable metric. This type of regularization term is defined by P k (d) = 1 2 dt W k d, where {W k } is a sequence of n-dimentional positive definite functions which often depends on x k. In this case, Pk (z) = zt (W k ) 1 z/2, and (P3A) obviously holds. However, to hold (P3), we need another assumption: there exists a λ L > 0 such that λ min (W k ) λ L for all k. Without the assumption, we can easily construct a counter example W k = I k /t k with t k Indiscriminate variable. Surprisingly, the assumption (P3) does not require that every P k is generated by a parametric function with respect to t or x. All we need is just (P3), lim inf k Pk (z) > 0 for all z 0. Therefore, for example, following absurd sequence of functions 1 t k α d α, k = 3m P k (d) = I B(0,tk )(d), k = 3m + 1, 1 2 dt W k d, k = 3m + 2 where α > 1 and there exists a t L > 0 such that t k t L, is acceptable. We have proved z k 0 by means of P k ( zk ) 0 with (P3). However, Proposition 3.1 also implies that P k (d k ) 0, and one may try to prove z k 0 from P k (d k ) 0 via d k 0. To reach d k 0 from P k (d k ) 0, an assumption similar to (P3) is necessary for P k. In addition, we need to prove z k 0 from d k 0, and hence some appropriate but not obvious assumptions are necessary. Unfortunately, even if the assumptions become clear, this approach is less attractive, because it does not cover all the regularization terms proved from P k ( zk ) 0. For example, consider the case of trust region, P k (d) = I B(0,tk )(d). In this case, we have confirmed that (P3) holds and z k 0. On the other hand, P k (d k ) = I B(0,tk )(d) 0 does not imply d k 0. Therefore, in this case, the convergence property cannot be proved by this approach. In short, z k 0 and d k 0 are not equivalent, in general. If P k is invariable with respect to k and the uniform function P = P k is of Legendre type, then the corespondence between z and d is one-to-one, thus z k 0 and d k 0 become equivalent. Further, the one-to-one relation is not necessary to hold on the entire

7 A UNIFIED CHARACTERIZATION OF PROXIMAL ALGORITHMS VIA THE CONJUGATE OF REGULARIZATION TERM7 domain, its necessary to hold just on B(0, ɛ) for sufficiently small ɛ > 0. z P ɛ O ɛ d Figure 1. Locally of Legendre type We do not consider the entire convergence of {x k } so far. In the case of P k (d) = d 2 /2t k, the entire convergence can be shown through the relation in the following. (3.5) x x k+1 2 = x x k 2 x k+1 x k 2 2 x k+1 x k, x x k+1, (3.6) 0 x k+1 x k, x x k+1. Equality (3.5) is just an algebra. Inequality (3.6) is deduced from [f + I X ](x ) [f + I X ](x k ) and [f + I X ](x ) [f + I X ](x k+1 ) + z k, x x k+1. What is deficient to prove the entire convergence of {x k }? For a general regularization term, we can not prove (3.6). What we can prove in general is that (3.7) 0 z k, x x k+1. Thus, we wish to prove something which mimics (3.5). This desire is partially achievable as follows. However, it is still not sufficient to prove the entire convergence. Lemma 3.5. Let {x k }, {z k }, and {d k } be sequences such that (3.1), (3.2), and (3.3) hold and P k be a sequence of positive convex functions such that (P1) and (P2) hold, then there exists a positive semidefinite symmetirc matrix M k such that M k (x k+1 x k ) = z k for all k such that z k, d k 0. Proof. Let W k = ( z) k ( z k ) T / z k, d k, then we can easily confirm that W k d k = z k. W k is positive semidefinite, since z k, d k = P k (d k ) + P k ( zk ) 0 by means of (P2) and (P*2), and we assume that z k, d k 0. To make W k positive definite, we need to prepare n 1 vertical vectors q i, i = 1,, n 1 such that q i, q j = δ ij and q i, d k = 0. These vectors, indeed, exist since there exists a (n 1)-dimentional hyperplane of which d k is a vertical vector. Note that z k is independent of {q i } n 1 i=1 since zk, d k 0 and q i, d k = 0. Therefore, adding q i q T i to W k, we can prepare a symmetric positive definite matrix M k = W k + n 1 i=1 q iqi T which satisfies M k (x k+1 x k ) = z k. We have completed the proof. By means of Lemma 3.5, we can obtain x x k+1 2 M k = x x k 2 M k x k+1 x k 2 M k 2 x k+1 x k, x x k+1 Mk, 0 x k+1 x k, x x k+1 Mk.

8 8 OUHEI HARADA At first glance, this relation is similar to (3.5) and (3.6). The essential difference is that the positive definite matrices M k are not invariable. Therefore, it is not sufficient to deduce x k x. However, by preparing a sufficient condition for f+i X, we can deduce dist(x k, X ) 0, where X = { x R n [f + I X ](x) [f + I X ](y), y } is the optimal set of (P). (F1) τ > 0, dist(x, X ) τ z, z f(x). Condition (F1) is known as error bounds which has been proposed by Luo [15]. By means of (F1), we can deduce dist(x k, X ) 0 directly from z k 0. Theorem 3.6. Let {x k } and {z k } be sequences which enjoy (3.1), (3.2), and (3.3), then dist(x k, X ) 0, if (P1), (P2), (P3), (PF1) and (F1) hold. To attack the entire convergence, other ideas are necessary. An attempt for this subject can be seen at [2], where the distance like function H(x, y) is introduced for {P k } such that H(x, x k+1 ) = H(x, x k ) H(x k+1, x k ) 2 z k, x x k+1, 0 z k, x x k+1, to hold. Of course, much stronger assumptions are necessary for the family of P k. For more detailed discussion, see [2]. 4. Bregman functions In this section, we consider the case that the regularization term P k is defined as D-function of Bregman functions. This approach has been developed by [5]. For any differentiable convex function h, D-function of h is defined by D h (x, y) = h(x) h(y) h(y), x y. For an open convex set S, a convex function h is said to be a Bregman function with zone S, if the following conditions hold [7, Definition 1]. (h1) h is continuously differentiable on S. (h2) h is strictly convex and continuous on S. (h3) D h (x, y) is inf-compact with respect to both x and y. (h4) If x k x, then D h (x, x k ) 0. (h5) If x k and y k are sequences such that y k y S, {x k } is bounded, and D h (x k, y k ) 0, then x k y. We show that P k (d) = (t k ) 1 D h (x k + d, x k ) is suitable for a regularization term. (P1) is obvious. (P2) follows from D h (x, x) = 0. To make the algorithm welldefined, we need ri(dom [f + I X ]) S. Then, condition (PF1) holds. In order to hold P t,x (0) = 0, strict convexity of h is necessary. We can confirm that (P3) also holds if P t,x (0) = 0 and there exists a t L > 0 such that t L t k for all k. Therefore, we can apply Theorem 3.4 and conclude that every limit point x is an optimal solution of Problem (P). In the case of Bregman functions, (P3) is not tractable to prove z k 0. By definition, we can see that z k = (t k ) 1 ( h(x k+1 ) h(x k )) 0 directly if t k t L for all k. Furthermore, it is well known that much stronger convergence result, the entire convergence of x k x can be proved. Theorem 4.1. [7, Theorem 1] Let {x k } be a sequence such that x k x and enjoys (3.2) and (3.3) with P k (d) = (t k ) 1 D h (x k + d, x k ), then we have x k x.

9 A UNIFIED CHARACTERIZATION OF PROXIMAL ALGORITHMS VIA THE CONJUGATE OF REGULARIZATION TERM9 Proof. By definition, we have that (4.1) D h (x, x k+1 ) = D h (x, x k ) D h (x k+1, x k ) h(x k+1 ) h(x k ), x x k+1. On the other hand, since x is an optimal solution of (P) and h(x k ) h(x k+1 ) is the derivative of f + I X at x k+1, we have f(x ) f(x k+1 ) f(x ) f(x k+1 ) h(x k+1 ) h(x k ), x x k+1 Therefore, we have h(x k+1 ) h(x k ), x x k+1 0. Together with D h (x k+1, x k ) 0, and (4.1), we have that the sequence {D h (x, x k )} is a non increasing sequence and hence entirely converges to some positive value. On the other hand, we have D h (x, x k ) 0 by means of x k x and (h4). This yields D h (x, x k ) 0. As a more general case, we can consider the case that h is not assumed to be smooth. In this case, P k is defined by P k (d) = D k h(x k + d, x k ) = h(x k + d) h(x k ) s, d, s h(x k ). Convergence property in this case has been proved by iwiel [14]. However, in our approach, it is hard to show z k 0. Let us consider that n = 1, X = R, f(x) = x and h(x) = x + x 2 /2. In this case, if we reach optimal x k+1 = 0 at some k, 0 f(x k ) + P k (0) implies not z k = 0 but z k [ 1, 1]. Although we have reached the point such that 0 f(x k+1 ), we cannot confirm it from just Pk ( zk ) ϕ-divergence In this section, we consider the case that P k itself plays the role of I X if necessary. In this case, the subproblem is replaced with minimize f(x k + d) + P k (d), x k + d R n, and hence the condition (3.2) becomes (5.1) z k f(x k + d k ). It implies that z k 0 does not always secure 0 (f + I X )(x ). Similarly, the assumption (PF1) should be replaced with (PF1 ) ri(dom f) ri(dom P k ). On the other hand, we can easily confirm that Proposition 3.1 holds by just replacing f + I X with f in its proof. We first explore a sufficient condition for {z k } to achieve 0 (f + I X )(x ). The situation differs whether x int(x) or not. The normal cone N X (x ) and its polar cone T X (x ) play important roles. To simplify the description, we denote the projection of a vector z onto N X (x ) and T X (x ) as p N (z) and p T (z) respectively. Proposition 5.1. Let {x k } and {z k } be sequences such that x k x and p T (z k ) 0, then 0 (f + I X )(x ). Proof. In view of Lemma 2.4, we have z k = p T ( z k ) + p N ( z k ). Note that p T ( z k ) p T (z k ), in general. Since ( z k ) = z k f(x k+1 ), we have p T ( z k ) p N ( z k ) f(x k+1 ). By definition, p N ( z k ) N X (x ) = I X (x ), and hence

10 10 OUHEI HARADA p T ( z k ) f(x k+1 ) + I X (x ). Therefore, condition p T ( z k ) 0 implies that 0 (f + I X )(x ). In particular, if {z k } converges to some z, p T ( z k ) 0 is equivalent to z N X (x ). Moreover, N X (x ) = I X (x ) = {0} holds if x int(x). In this case, p T ( z k ) 0 is equivalent to z k 0. The problem is how to prepare P k to satisfy p T ( z k ) 0. Let us consider the special case that X = R n +. In this case, ϕ-divergence defined in the following are well known. n ( ) P k = Q tk,x k, Q xi + d i (5.2) t,x(d) = t i x i ϕ, where Q t,x is a parametric positive convex function with respect to t R n + and x, and ϕ : R n R + is a positive, lsc, and convex function such that the following conditions hold. i=1 (ϕ0) ϕ is smooth on (0, + ). (ϕ1) ϕ(1) = 0. (ϕ2) ϕ (1) = 0. (ϕ3) (0, + ) dom ϕ (0, + ]. (ϕ4) lim d 0 ϕ (d) = + and b > 0, ϕ (1) = b. ϕ is often said to be a kernel function of ϕ-divergence Q t,x. We can easily confirm that P k defined as (5.2) satisfies the requirements of the regularization term (P1), (P2), and (PF). However, in our setting, we can see that shifting ϕ by 1 to minus direction make the function more tractable and symmetric with respect to the conjugate operator. Furthermore, the smooth condition (ϕ0) is not necessary except for its minimum point. For that reason, we introduce another kernel function ˆϕ(d) which mimics ϕ(1+d). Then, the regularization term is defined as follows. n ( ) P k = Q t k,x k, Q di (5.3) t,x(d) = t i x i ˆϕ, where ˆϕ : R R + is a positive, lsc, and convex function which satisfies the following requirements: ( ˆϕ1) ˆϕ(0) = 0. ( ˆϕ2) ˆϕ(0) = {0} ( ˆϕ is differentiable at 0 and ˆϕ (0) = 0). ( ˆϕ3) ( 1, + ) dom ˆϕ [ 1, + ). ( ˆϕ4) ˆϕ ( 1) = + and b > 0, ˆϕ (1) = b. We can easily confirm that if a positive function ϕ satisfies (ϕ1) to (ϕ4), then the shifted function ˆϕ(d) = ϕ(1 + d) satisfies ( ˆϕ1) to ( ˆϕ4). Let ˆψ = ˆϕ be a conjugate function of ˆϕ, following properties hold. ( ˆψ1) ˆψ(0) = 0. ( ˆψ2) L 0 ( ˆψ) = {0}. ( ˆψ3) (, b) dom ˆψ (, b]. ( ˆψ4) ˆψ ( 1) = 1 and ˆψ (1) = +. Lemma 5.2. If a positive function ˆϕ satisfies ( ˆϕ1) to ( ˆϕ4), then its conjugate function ˆψ = ˆϕ satisfies ( ˆψ1) to ( ˆψ4). i=1 x i x i

11 A UNIFIED CHARACTERIZATION OF PROXIMAL ALGORITHMS VIA THE CONJUGATE OF REGULARIZATION TERM11 Proof. By definition, any conjugate function is convex and lsc, and hence ( ˆψ0) holds. ( ˆψ1) is obvious; ˆψ(0) = supd ( ˆϕ(d)) = 0. By means of Lemma, we have L 0 ( ˆψ) = ˆϕ(0) = {0}. Therefore ( ˆψ2) is proved. In view of ( ˆϕ4), there exists a d dom ˆϕ such that s ˆϕ(d) d ˆψ(s) for all s (, b). Therefore, we have (, b) dom ˆψ. On the other hand, by means of ˆϕ ( 1) = +, we have ˆψ(s) = sup d d(s ˆϕ(d)/d) sup d 0 d(s b) = + for all s > b. It follows that dom ˆψ (b, + ) =, and hence we have dom ˆψ (, 1]. We have proved ( ˆψ3). In view of ( ˆϕ3), there exists a s ˆϕ(d) d ˆψ(s) for all d ( 1, + ). It implies that ˆψ ( 1) 1 and ˆψ (1) = +. On the other hand dom ˆϕ [ 1, + ) implies that dom ˆϕ (, 1) =, and hence ˆψ ( 1) 1. We have proved ( ˆψ4). If we add L 0 ( ˆϕ) = {0} in ( ˆϕ2) and set b = 1 at ( ˆϕ4), then ˆψ(0) = {0} is added to ( ˆψ2) and ( ˆψ3) becomes (, 1) dom ˆψ (, 1]. Therefore, conditions between ˆϕ and ˆψ become symmetric with respect to the vertical line passing through the origin. A typical example is ˆϕ(d) = log(1 + d) + d and ˆψ(z) = log(1 z) z. ˆϕ ˆψ 1 O b Figure 2. ˆϕ and ˆψ In view of Proposition 3.1 and 5.1, our goal is to prove p T ( z k ) 0. We begin with the following lemmas. Lemma 5.3. The conjugate function of Q t,x is given as follows: n ( ) Q zi t,x(z) = Ψ ti,x i (z i ), Ψ ti,x i (z i ) = t i x i ˆψ. i=1 Proof. By definition, we have that { n [ Q t,x(z) = sup z i d i t i x i ˆϕ d i=1 ( di Therefore, we have completed the proof. x i )] } = n i=1 t i [ { zi t i x i sup di ˆϕ d t i x i ( di x i )}]. Lemma 5.4. Let {a k }, {v k }, and {w k } be sequences such that Ψ ak,v k(wk ) 0, a k ā > 0 for all k, and 0 is not a limit point of {v k }, then w k 0. Proof. Since 0 is not a limit point of {v k }, there exists a sufficiently large k such that v k 0 for all k k. It follows that the sequence of the function {Ψ a k,v k} k k and the sequence {w k } k k satisfies the requirements of Lemma 3.3. Therefore, we have w k 0 and completed the proof. Lemma 5.5. Let {a k }, {v k }, and {w k } be sequences such that Ψ a k,v k(wk ) 0 and a k 0, then (w k ) + 0.

12 12 OUHEI HARADA Proof. By definition, Ψ is a positive and convex function. In view of ( ˆψ3), dom Ψ (, ba k ]. Therefore w k (, ba k ]. By means of a k 0, it follows that (w k ) + 0. To make use of Lemma 5.4 and 5.5, we need to elaborate how to update t k which satisfies the following requirement. (T1) t k i 0 if x i = 0, otherwise there exists a t > 0 such that t k i t for all k. Under the requirement, we can establish the following proposition. Proposition 5.6. Let the regularization term P k is defined by (5.3), {t k }, {x k }, and {z k } be sequences such that P k ( z k ) 0, x k x, and {t k } satisfies (T1), then p T ( z k ) 0. Proof. In view of Lemma 5.3, P k ( z k ) 0 implies that Ψ t k i,x k( zk i i ) 0 for all i. Let I be the set of indices such that I = {i x i = 0}. For every i / I, we have zi k 0 from Lemma 5.4. Similarly, for every i I, we have ( zi k) + 0 from Lemma 5.5. On the other hand, N R n + (x ) = {z z R n, z i 0, i I}, and hence T R n + (x ) = {d d R n, d i 0, i I}. Therefore, we conclude that p T ( z k ) = ( z k ) + 0 and we have completed the proof. Corollary 5.7. Let the regularization term P k is defined by (5.3), {t k }, {x k }, and {z k } be sequences such that P k ( z k ) 0, x k x int(x), and {t k } satisfies (T1), then z k 0. The left is how to construct an update rule of t k to satisfy (T1). A simple rule is to set t k = x k. Ben-tal [3] proposed nondecreasing sublinear functions π k : R + R + such that there exists a scalar c > 0 which satisfies π k (x) cx for all positive schalar x > 0, and they update t k by means of π k such that t k i = πk (x k i ). This rule enjoys (T1), to be sure. However, more generalized π k is acceptable. (π1) π k is nonincreasing. (π2) lim t 0 π k (t) = 0. (π3) L 0 (π k ) = {0}. If π k satisfies (π1-3), then (T1) holds. For example, π k (x) = x is not sublinear, but satisfies (π1-3), and hence (T1) holds. In view of Proposition 5.1 and 5.6, we conclude the following. Theorem 5.8. Let {x k }, {z k } and {d k } be sequences which enjoy (3.2), (5.1), and (PF1 ), and {t k } is updated by t k = π k (x k ) such that (π1-3) holds, then there exists a convergent subsequence x k x and every limit point is a minimizer of f on R n +. The result obtained in this section can be seen an extention of [3]. We do not assume that ˆϕ is essentially smooth, and the somewhat artificial evaluation of t i x i ˆϕ(d i /x i ), [3, Lemma 2], is not necessary in our approach. In addition, our requirements for π k is more general. References [1] Alfred Auslender. Numerical methods for nondifferentiable convex optimization. In Nonlinear Analysis and Optimization, pages Springer, 1987.

13 A UNIFIED CHARACTERIZATION OF PROXIMAL ALGORITHMS VIA THE CONJUGATE OF REGULARIZATION TERM13 [2] Alfred Auslender and Marc Teboulle. Interior gradient and proximal methods for convex and conic optimization. SIAM Journal on Optimization, 16(3): , [3] Aharon Ben-Tal and Michael Zibulevsky. Penalty/barrier multiplier methods for convex programming problems. SIAM Journal on Optimization, 7(2): , [4] Joseph Frédéric Bonnans, Jean Charles Gilbert, Claude Lemaréchal, and Claudia Sagastizábal. A family of variable metric proximal methods. Mathematical Programming, 68(1-3):15 47, [5] Yair Censor and Stavros Andrea Zenios. Proximal minimization algorithm withd-functions. Journal of Optimization Theory and Applications, 73(3): , [6] Gong Chen and Marc Teboulle. Convergence analysis of a proximal-like minimization algorithm using Bregman functions. SIAM Journal on Optimization, 3(3): , [7] Jonathan Eckstein. Nonlinear proximal point algorithms using bregman functions, with applications to convex programming. Mathematics of Operations Research, 18(1): , [8] Antonio Frangioni. Generalized bundle methods. SIAM Journal on Optimization, 13(1): , [9] JB Hiriart-Urruty and Claude Lemaréchal. Convex Analysis and Minimization Algorithms I. Fundamentals. Springer-Verlag, New York, [10] JB Hiriart-Urruty and Claude Lemaréchal. Convex Analysis and Minimization Algorithms II: Advanced Theory and Bundle Methods. Springer-Verlag, New York, [11] Alfredo N Iusem. Augmented Lagrangian methods and proximal point methods for convex optimization. Investigación Operativa, 8(11-49):7, [12] Alfredo N Iusem, BF Svaiter, and Marc Teboulle. Entropy-like proximal methods in convex programming. Mathematics of Operations Research, 19(4): , [13] Sehun im, un-nyeong Chang, and Jun-Yeon Lee. A descent method with linear programming subproblems for nondifferentiable convex optimization. Mathematical programming, 71(1):17 28, [14] rzysztof C iwiel. Proximal minimization methods with generalized Bregman functions. SIAM Journal on Control and Optimization, 35(4): , [15] Zhi-Quan Luo and Paul Tseng. Error bounds and convergence analysis of feasible descent methods: a general approach. Annals of Operations Research, 46(1): , [16] Felipe Garcia Moreno, Paulo Roberto Oliveira, and Antoine Soubeyran. A proximal algorithm with quasi distance. application to habit s formation. Optimization, 61(12): , [17] Roman Polyak and Marc Teboulle. Nonlinear rescaling and proximal-like methods in convex optimization. Mathematical Programming, 76(2): , [18] R Tyrrell Rockafellar. Convex Analysis (Princeton Mathematical Series). Princeton University Press, [19] Marc Teboulle. Convergence of proximal-like algorithms. SIAM Journal on Optimization, 7(4): , 1997.

Optimality Conditions for Nonsmooth Convex Optimization

Optimality Conditions for Nonsmooth Convex Optimization Optimality Conditions for Nonsmooth Convex Optimization Sangkyun Lee Oct 22, 2014 Let us consider a convex function f : R n R, where R is the extended real field, R := R {, + }, which is proper (f never

More information

Some Properties of the Augmented Lagrangian in Cone Constrained Optimization

Some Properties of the Augmented Lagrangian in Cone Constrained Optimization MATHEMATICS OF OPERATIONS RESEARCH Vol. 29, No. 3, August 2004, pp. 479 491 issn 0364-765X eissn 1526-5471 04 2903 0479 informs doi 10.1287/moor.1040.0103 2004 INFORMS Some Properties of the Augmented

More information

A convergence result for an Outer Approximation Scheme

A convergence result for an Outer Approximation Scheme A convergence result for an Outer Approximation Scheme R. S. Burachik Engenharia de Sistemas e Computação, COPPE-UFRJ, CP 68511, Rio de Janeiro, RJ, CEP 21941-972, Brazil regi@cos.ufrj.br J. O. Lopes Departamento

More information

An Infeasible Interior Proximal Method for Convex Programming Problems with Linear Constraints 1

An Infeasible Interior Proximal Method for Convex Programming Problems with Linear Constraints 1 An Infeasible Interior Proximal Method for Convex Programming Problems with Linear Constraints 1 Nobuo Yamashita 2, Christian Kanzow 3, Tomoyui Morimoto 2, and Masao Fuushima 2 2 Department of Applied

More information

A Geometric Framework for Nonconvex Optimization Duality using Augmented Lagrangian Functions

A Geometric Framework for Nonconvex Optimization Duality using Augmented Lagrangian Functions A Geometric Framework for Nonconvex Optimization Duality using Augmented Lagrangian Functions Angelia Nedić and Asuman Ozdaglar April 15, 2006 Abstract We provide a unifying geometric framework for the

More information

WEAK CONVERGENCE OF RESOLVENTS OF MAXIMAL MONOTONE OPERATORS AND MOSCO CONVERGENCE

WEAK CONVERGENCE OF RESOLVENTS OF MAXIMAL MONOTONE OPERATORS AND MOSCO CONVERGENCE Fixed Point Theory, Volume 6, No. 1, 2005, 59-69 http://www.math.ubbcluj.ro/ nodeacj/sfptcj.htm WEAK CONVERGENCE OF RESOLVENTS OF MAXIMAL MONOTONE OPERATORS AND MOSCO CONVERGENCE YASUNORI KIMURA Department

More information

Hessian Riemannian Gradient Flows in Convex Programming

Hessian Riemannian Gradient Flows in Convex Programming Hessian Riemannian Gradient Flows in Convex Programming Felipe Alvarez, Jérôme Bolte, Olivier Brahic INTERNATIONAL CONFERENCE ON MODELING AND OPTIMIZATION MODOPT 2004 Universidad de La Frontera, Temuco,

More information

A Unified Analysis of Nonconvex Optimization Duality and Penalty Methods with General Augmenting Functions

A Unified Analysis of Nonconvex Optimization Duality and Penalty Methods with General Augmenting Functions A Unified Analysis of Nonconvex Optimization Duality and Penalty Methods with General Augmenting Functions Angelia Nedić and Asuman Ozdaglar April 16, 2006 Abstract In this paper, we study a unifying framework

More information

A Unified Approach to Proximal Algorithms using Bregman Distance

A Unified Approach to Proximal Algorithms using Bregman Distance A Unified Approach to Proximal Algorithms using Bregman Distance Yi Zhou a,, Yingbin Liang a, Lixin Shen b a Department of Electrical Engineering and Computer Science, Syracuse University b Department

More information

FIXED POINTS IN THE FAMILY OF CONVEX REPRESENTATIONS OF A MAXIMAL MONOTONE OPERATOR

FIXED POINTS IN THE FAMILY OF CONVEX REPRESENTATIONS OF A MAXIMAL MONOTONE OPERATOR PROCEEDINGS OF THE AMERICAN MATHEMATICAL SOCIETY Volume 00, Number 0, Pages 000 000 S 0002-9939(XX)0000-0 FIXED POINTS IN THE FAMILY OF CONVEX REPRESENTATIONS OF A MAXIMAL MONOTONE OPERATOR B. F. SVAITER

More information

Journal of Convex Analysis (accepted for publication) A HYBRID PROJECTION PROXIMAL POINT ALGORITHM. M. V. Solodov and B. F.

Journal of Convex Analysis (accepted for publication) A HYBRID PROJECTION PROXIMAL POINT ALGORITHM. M. V. Solodov and B. F. Journal of Convex Analysis (accepted for publication) A HYBRID PROJECTION PROXIMAL POINT ALGORITHM M. V. Solodov and B. F. Svaiter January 27, 1997 (Revised August 24, 1998) ABSTRACT We propose a modification

More information

A New Look at First Order Methods Lifting the Lipschitz Gradient Continuity Restriction

A New Look at First Order Methods Lifting the Lipschitz Gradient Continuity Restriction A New Look at First Order Methods Lifting the Lipschitz Gradient Continuity Restriction Marc Teboulle School of Mathematical Sciences Tel Aviv University Joint work with H. Bauschke and J. Bolte Optimization

More information

On Total Convexity, Bregman Projections and Stability in Banach Spaces

On Total Convexity, Bregman Projections and Stability in Banach Spaces Journal of Convex Analysis Volume 11 (2004), No. 1, 1 16 On Total Convexity, Bregman Projections and Stability in Banach Spaces Elena Resmerita Department of Mathematics, University of Haifa, 31905 Haifa,

More information

Positive semidefinite matrix approximation with a trace constraint

Positive semidefinite matrix approximation with a trace constraint Positive semidefinite matrix approximation with a trace constraint Kouhei Harada August 8, 208 We propose an efficient algorithm to solve positive a semidefinite matrix approximation problem with a trace

More information

On the Convergence and O(1/N) Complexity of a Class of Nonlinear Proximal Point Algorithms for Monotonic Variational Inequalities

On the Convergence and O(1/N) Complexity of a Class of Nonlinear Proximal Point Algorithms for Monotonic Variational Inequalities STATISTICS,OPTIMIZATION AND INFORMATION COMPUTING Stat., Optim. Inf. Comput., Vol. 2, June 204, pp 05 3. Published online in International Academic Press (www.iapress.org) On the Convergence and O(/N)

More information

ON GAP FUNCTIONS OF VARIATIONAL INEQUALITY IN A BANACH SPACE. Sangho Kum and Gue Myung Lee. 1. Introduction

ON GAP FUNCTIONS OF VARIATIONAL INEQUALITY IN A BANACH SPACE. Sangho Kum and Gue Myung Lee. 1. Introduction J. Korean Math. Soc. 38 (2001), No. 3, pp. 683 695 ON GAP FUNCTIONS OF VARIATIONAL INEQUALITY IN A BANACH SPACE Sangho Kum and Gue Myung Lee Abstract. In this paper we are concerned with theoretical properties

More information

A proximal-like algorithm for a class of nonconvex programming

A proximal-like algorithm for a class of nonconvex programming Pacific Journal of Optimization, vol. 4, pp. 319-333, 2008 A proximal-like algorithm for a class of nonconvex programming Jein-Shan Chen 1 Department of Mathematics National Taiwan Normal University Taipei,

More information

LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE

LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE CONVEX ANALYSIS AND DUALITY Basic concepts of convex analysis Basic concepts of convex optimization Geometric duality framework - MC/MC Constrained optimization

More information

Local strong convexity and local Lipschitz continuity of the gradient of convex functions

Local strong convexity and local Lipschitz continuity of the gradient of convex functions Local strong convexity and local Lipschitz continuity of the gradient of convex functions R. Goebel and R.T. Rockafellar May 23, 2007 Abstract. Given a pair of convex conjugate functions f and f, we investigate

More information

Existence and Approximation of Fixed Points of. Bregman Nonexpansive Operators. Banach Spaces

Existence and Approximation of Fixed Points of. Bregman Nonexpansive Operators. Banach Spaces Existence and Approximation of Fixed Points of in Reflexive Banach Spaces Department of Mathematics The Technion Israel Institute of Technology Haifa 22.07.2010 Joint work with Prof. Simeon Reich General

More information

Convergence Theorems for Bregman Strongly Nonexpansive Mappings in Reflexive Banach Spaces

Convergence Theorems for Bregman Strongly Nonexpansive Mappings in Reflexive Banach Spaces Filomat 28:7 (2014), 1525 1536 DOI 10.2298/FIL1407525Z Published by Faculty of Sciences and Mathematics, University of Niš, Serbia Available at: http://www.pmf.ni.ac.rs/filomat Convergence Theorems for

More information

Optimality Conditions for Constrained Optimization

Optimality Conditions for Constrained Optimization 72 CHAPTER 7 Optimality Conditions for Constrained Optimization 1. First Order Conditions In this section we consider first order optimality conditions for the constrained problem P : minimize f 0 (x)

More information

Subdifferential representation of convex functions: refinements and applications

Subdifferential representation of convex functions: refinements and applications Subdifferential representation of convex functions: refinements and applications Joël Benoist & Aris Daniilidis Abstract Every lower semicontinuous convex function can be represented through its subdifferential

More information

McMaster University. Advanced Optimization Laboratory. Title: A Proximal Method for Identifying Active Manifolds. Authors: Warren L.

McMaster University. Advanced Optimization Laboratory. Title: A Proximal Method for Identifying Active Manifolds. Authors: Warren L. McMaster University Advanced Optimization Laboratory Title: A Proximal Method for Identifying Active Manifolds Authors: Warren L. Hare AdvOl-Report No. 2006/07 April 2006, Hamilton, Ontario, Canada A Proximal

More information

Extended Monotropic Programming and Duality 1

Extended Monotropic Programming and Duality 1 March 2006 (Revised February 2010) Report LIDS - 2692 Extended Monotropic Programming and Duality 1 by Dimitri P. Bertsekas 2 Abstract We consider the problem minimize f i (x i ) subject to x S, where

More information

A Proximal Method for Identifying Active Manifolds

A Proximal Method for Identifying Active Manifolds A Proximal Method for Identifying Active Manifolds W.L. Hare April 18, 2006 Abstract The minimization of an objective function over a constraint set can often be simplified if the active manifold of the

More information

Iterative Convex Optimization Algorithms; Part One: Using the Baillon Haddad Theorem

Iterative Convex Optimization Algorithms; Part One: Using the Baillon Haddad Theorem Iterative Convex Optimization Algorithms; Part One: Using the Baillon Haddad Theorem Charles Byrne (Charles Byrne@uml.edu) http://faculty.uml.edu/cbyrne/cbyrne.html Department of Mathematical Sciences

More information

On the iterate convergence of descent methods for convex optimization

On the iterate convergence of descent methods for convex optimization On the iterate convergence of descent methods for convex optimization Clovis C. Gonzaga March 1, 2014 Abstract We study the iterate convergence of strong descent algorithms applied to convex functions.

More information

ON A CLASS OF NONSMOOTH COMPOSITE FUNCTIONS

ON A CLASS OF NONSMOOTH COMPOSITE FUNCTIONS MATHEMATICS OF OPERATIONS RESEARCH Vol. 28, No. 4, November 2003, pp. 677 692 Printed in U.S.A. ON A CLASS OF NONSMOOTH COMPOSITE FUNCTIONS ALEXANDER SHAPIRO We discuss in this paper a class of nonsmooth

More information

Finite Convergence for Feasible Solution Sequence of Variational Inequality Problems

Finite Convergence for Feasible Solution Sequence of Variational Inequality Problems Mathematical and Computational Applications Article Finite Convergence for Feasible Solution Sequence of Variational Inequality Problems Wenling Zhao *, Ruyu Wang and Hongxiang Zhang School of Science,

More information

ON GENERALIZED-CONVEX CONSTRAINED MULTI-OBJECTIVE OPTIMIZATION

ON GENERALIZED-CONVEX CONSTRAINED MULTI-OBJECTIVE OPTIMIZATION ON GENERALIZED-CONVEX CONSTRAINED MULTI-OBJECTIVE OPTIMIZATION CHRISTIAN GÜNTHER AND CHRISTIANE TAMMER Abstract. In this paper, we consider multi-objective optimization problems involving not necessarily

More information

I P IANO : I NERTIAL P ROXIMAL A LGORITHM FOR N ON -C ONVEX O PTIMIZATION

I P IANO : I NERTIAL P ROXIMAL A LGORITHM FOR N ON -C ONVEX O PTIMIZATION I P IANO : I NERTIAL P ROXIMAL A LGORITHM FOR N ON -C ONVEX O PTIMIZATION Peter Ochs University of Freiburg Germany 17.01.2017 joint work with: Thomas Brox and Thomas Pock c 2017 Peter Ochs ipiano c 1

More information

Journal of Convex Analysis Vol. 14, No. 2, March 2007 AN EXPLICIT DESCENT METHOD FOR BILEVEL CONVEX OPTIMIZATION. Mikhail Solodov. September 12, 2005

Journal of Convex Analysis Vol. 14, No. 2, March 2007 AN EXPLICIT DESCENT METHOD FOR BILEVEL CONVEX OPTIMIZATION. Mikhail Solodov. September 12, 2005 Journal of Convex Analysis Vol. 14, No. 2, March 2007 AN EXPLICIT DESCENT METHOD FOR BILEVEL CONVEX OPTIMIZATION Mikhail Solodov September 12, 2005 ABSTRACT We consider the problem of minimizing a smooth

More information

Convergence rate estimates for the gradient differential inclusion

Convergence rate estimates for the gradient differential inclusion Convergence rate estimates for the gradient differential inclusion Osman Güler November 23 Abstract Let f : H R { } be a proper, lower semi continuous, convex function in a Hilbert space H. The gradient

More information

Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem

Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem Michael Patriksson 0-0 The Relaxation Theorem 1 Problem: find f := infimum f(x), x subject to x S, (1a) (1b) where f : R n R

More information

CONVEX OPTIMIZATION VIA LINEARIZATION. Miguel A. Goberna. Universidad de Alicante. Iberian Conference on Optimization Coimbra, November, 2006

CONVEX OPTIMIZATION VIA LINEARIZATION. Miguel A. Goberna. Universidad de Alicante. Iberian Conference on Optimization Coimbra, November, 2006 CONVEX OPTIMIZATION VIA LINEARIZATION Miguel A. Goberna Universidad de Alicante Iberian Conference on Optimization Coimbra, 16-18 November, 2006 Notation X denotes a l.c. Hausdorff t.v.s and X its topological

More information

Inequality Constraints

Inequality Constraints Chapter 2 Inequality Constraints 2.1 Optimality Conditions Early in multivariate calculus we learn the significance of differentiability in finding minimizers. In this section we begin our study of the

More information

ALGORITHMS FOR MINIMIZING DIFFERENCES OF CONVEX FUNCTIONS AND APPLICATIONS

ALGORITHMS FOR MINIMIZING DIFFERENCES OF CONVEX FUNCTIONS AND APPLICATIONS ALGORITHMS FOR MINIMIZING DIFFERENCES OF CONVEX FUNCTIONS AND APPLICATIONS Mau Nam Nguyen (joint work with D. Giles and R. B. Rector) Fariborz Maseeh Department of Mathematics and Statistics Portland State

More information

1 Introduction We consider the problem nd x 2 H such that 0 2 T (x); (1.1) where H is a real Hilbert space, and T () is a maximal monotone operator (o

1 Introduction We consider the problem nd x 2 H such that 0 2 T (x); (1.1) where H is a real Hilbert space, and T () is a maximal monotone operator (o Journal of Convex Analysis Volume 6 (1999), No. 1, pp. xx-xx. cheldermann Verlag A HYBRID PROJECTION{PROXIMAL POINT ALGORITHM M. V. Solodov y and B. F. Svaiter y January 27, 1997 (Revised August 24, 1998)

More information

NONSMOOTH VARIANTS OF POWELL S BFGS CONVERGENCE THEOREM

NONSMOOTH VARIANTS OF POWELL S BFGS CONVERGENCE THEOREM NONSMOOTH VARIANTS OF POWELL S BFGS CONVERGENCE THEOREM JIAYI GUO AND A.S. LEWIS Abstract. The popular BFGS quasi-newton minimization algorithm under reasonable conditions converges globally on smooth

More information

arxiv: v1 [math.fa] 16 Jun 2011

arxiv: v1 [math.fa] 16 Jun 2011 arxiv:1106.3342v1 [math.fa] 16 Jun 2011 Gauge functions for convex cones B. F. Svaiter August 20, 2018 Abstract We analyze a class of sublinear functionals which characterize the interior and the exterior

More information

A quasisecant method for minimizing nonsmooth functions

A quasisecant method for minimizing nonsmooth functions A quasisecant method for minimizing nonsmooth functions Adil M. Bagirov and Asef Nazari Ganjehlou Centre for Informatics and Applied Optimization, School of Information Technology and Mathematical Sciences,

More information

ENLARGEMENT OF MONOTONE OPERATORS WITH APPLICATIONS TO VARIATIONAL INEQUALITIES. Abstract

ENLARGEMENT OF MONOTONE OPERATORS WITH APPLICATIONS TO VARIATIONAL INEQUALITIES. Abstract ENLARGEMENT OF MONOTONE OPERATORS WITH APPLICATIONS TO VARIATIONAL INEQUALITIES Regina S. Burachik* Departamento de Matemática Pontíficia Universidade Católica de Rio de Janeiro Rua Marques de São Vicente,

More information

CONTINUOUS CONVEX SETS AND ZERO DUALITY GAP FOR CONVEX PROGRAMS

CONTINUOUS CONVEX SETS AND ZERO DUALITY GAP FOR CONVEX PROGRAMS CONTINUOUS CONVEX SETS AND ZERO DUALITY GAP FOR CONVEX PROGRAMS EMIL ERNST AND MICHEL VOLLE ABSTRACT. This article uses classical notions of convex analysis over euclidean spaces, like Gale & Klee s boundary

More information

A CHARACTERIZATION OF STRICT LOCAL MINIMIZERS OF ORDER ONE FOR STATIC MINMAX PROBLEMS IN THE PARAMETRIC CONSTRAINT CASE

A CHARACTERIZATION OF STRICT LOCAL MINIMIZERS OF ORDER ONE FOR STATIC MINMAX PROBLEMS IN THE PARAMETRIC CONSTRAINT CASE Journal of Applied Analysis Vol. 6, No. 1 (2000), pp. 139 148 A CHARACTERIZATION OF STRICT LOCAL MINIMIZERS OF ORDER ONE FOR STATIC MINMAX PROBLEMS IN THE PARAMETRIC CONSTRAINT CASE A. W. A. TAHA Received

More information

On Semicontinuity of Convex-valued Multifunctions and Cesari s Property (Q)

On Semicontinuity of Convex-valued Multifunctions and Cesari s Property (Q) On Semicontinuity of Convex-valued Multifunctions and Cesari s Property (Q) Andreas Löhne May 2, 2005 (last update: November 22, 2005) Abstract We investigate two types of semicontinuity for set-valued

More information

Chapter 2 Convex Analysis

Chapter 2 Convex Analysis Chapter 2 Convex Analysis The theory of nonsmooth analysis is based on convex analysis. Thus, we start this chapter by giving basic concepts and results of convexity (for further readings see also [202,

More information

Auxiliary-Function Methods in Iterative Optimization

Auxiliary-Function Methods in Iterative Optimization Auxiliary-Function Methods in Iterative Optimization Charles L. Byrne April 6, 2015 Abstract Let C X be a nonempty subset of an arbitrary set X and f : X R. The problem is to minimize f over C. In auxiliary-function

More information

Downloaded 09/27/13 to Redistribution subject to SIAM license or copyright; see

Downloaded 09/27/13 to Redistribution subject to SIAM license or copyright; see SIAM J. OPTIM. Vol. 23, No., pp. 256 267 c 203 Society for Industrial and Applied Mathematics TILT STABILITY, UNIFORM QUADRATIC GROWTH, AND STRONG METRIC REGULARITY OF THE SUBDIFFERENTIAL D. DRUSVYATSKIY

More information

BREGMAN DISTANCES, TOTALLY

BREGMAN DISTANCES, TOTALLY BREGMAN DISTANCES, TOTALLY CONVEX FUNCTIONS AND A METHOD FOR SOLVING OPERATOR EQUATIONS IN BANACH SPACES DAN BUTNARIU AND ELENA RESMERITA January 18, 2005 Abstract The aim of this paper is twofold. First,

More information

Legendre-Fenchel transforms in a nutshell

Legendre-Fenchel transforms in a nutshell 1 2 3 Legendre-Fenchel transforms in a nutshell Hugo Touchette School of Mathematical Sciences, Queen Mary, University of London, London E1 4NS, UK Started: July 11, 2005; last compiled: August 14, 2007

More information

Refined optimality conditions for differences of convex functions

Refined optimality conditions for differences of convex functions Noname manuscript No. (will be inserted by the editor) Refined optimality conditions for differences of convex functions Tuomo Valkonen the date of receipt and acceptance should be inserted later Abstract

More information

Variational inequalities for set-valued vector fields on Riemannian manifolds

Variational inequalities for set-valued vector fields on Riemannian manifolds Variational inequalities for set-valued vector fields on Riemannian manifolds Chong LI Department of Mathematics Zhejiang University Joint with Jen-Chih YAO Chong LI (Zhejiang University) VI on RM 1 /

More information

arxiv: v3 [math.oc] 17 Dec 2017 Received: date / Accepted: date

arxiv: v3 [math.oc] 17 Dec 2017 Received: date / Accepted: date Noname manuscript No. (will be inserted by the editor) A Simple Convergence Analysis of Bregman Proximal Gradient Algorithm Yi Zhou Yingbin Liang Lixin Shen arxiv:1503.05601v3 [math.oc] 17 Dec 2017 Received:

More information

Fixed points in the family of convex representations of a maximal monotone operator

Fixed points in the family of convex representations of a maximal monotone operator arxiv:0802.1347v2 [math.fa] 12 Feb 2008 Fixed points in the family of convex representations of a maximal monotone operator published on: Proc. Amer. Math. Soc. 131 (2003) 3851 3859. B. F. Svaiter IMPA

More information

Convex Optimization Theory. Chapter 5 Exercises and Solutions: Extended Version

Convex Optimization Theory. Chapter 5 Exercises and Solutions: Extended Version Convex Optimization Theory Chapter 5 Exercises and Solutions: Extended Version Dimitri P. Bertsekas Massachusetts Institute of Technology Athena Scientific, Belmont, Massachusetts http://www.athenasc.com

More information

Extensions of Korpelevich s Extragradient Method for the Variational Inequality Problem in Euclidean Space

Extensions of Korpelevich s Extragradient Method for the Variational Inequality Problem in Euclidean Space Extensions of Korpelevich s Extragradient Method for the Variational Inequality Problem in Euclidean Space Yair Censor 1,AvivGibali 2 andsimeonreich 2 1 Department of Mathematics, University of Haifa,

More information

Radial Subgradient Descent

Radial Subgradient Descent Radial Subgradient Descent Benja Grimmer Abstract We present a subgradient method for imizing non-smooth, non-lipschitz convex optimization problems. The only structure assumed is that a strictly feasible

More information

Accelerated Proximal Gradient Methods for Convex Optimization

Accelerated Proximal Gradient Methods for Convex Optimization Accelerated Proximal Gradient Methods for Convex Optimization Paul Tseng Mathematics, University of Washington Seattle MOPTA, University of Guelph August 18, 2008 ACCELERATED PROXIMAL GRADIENT METHODS

More information

Existence of Global Minima for Constrained Optimization 1

Existence of Global Minima for Constrained Optimization 1 Existence of Global Minima for Constrained Optimization 1 A. E. Ozdaglar 2 and P. Tseng 3 Communicated by A. Miele 1 We thank Professor Dimitri Bertsekas for his comments and support in the writing of

More information

A characterization of essentially strictly convex functions on reflexive Banach spaces

A characterization of essentially strictly convex functions on reflexive Banach spaces A characterization of essentially strictly convex functions on reflexive Banach spaces Michel Volle Département de Mathématiques Université d Avignon et des Pays de Vaucluse 74, rue Louis Pasteur 84029

More information

Some Contributions to Convex Infinite-Dimensional Optimization Duality

Some Contributions to Convex Infinite-Dimensional Optimization Duality Some Contributions to Convex Infinite-Dimensional Optimization Duality Marco A. López Alicante University King s College London Strand Campus June 2014 Introduction Consider the convex infinite programming

More information

POLARS AND DUAL CONES

POLARS AND DUAL CONES POLARS AND DUAL CONES VERA ROSHCHINA Abstract. The goal of this note is to remind the basic definitions of convex sets and their polars. For more details see the classic references [1, 2] and [3] for polytopes.

More information

Convex Optimization Notes

Convex Optimization Notes Convex Optimization Notes Jonathan Siegel January 2017 1 Convex Analysis This section is devoted to the study of convex functions f : B R {+ } and convex sets U B, for B a Banach space. The case of B =

More information

A double projection method for solving variational inequalities without monotonicity

A double projection method for solving variational inequalities without monotonicity A double projection method for solving variational inequalities without monotonicity Minglu Ye Yiran He Accepted by Computational Optimization and Applications, DOI: 10.1007/s10589-014-9659-7,Apr 05, 2014

More information

Set, functions and Euclidean space. Seungjin Han

Set, functions and Euclidean space. Seungjin Han Set, functions and Euclidean space Seungjin Han September, 2018 1 Some Basics LOGIC A is necessary for B : If B holds, then A holds. B A A B is the contraposition of B A. A is sufficient for B: If A holds,

More information

Subgradient. Acknowledgement: this slides is based on Prof. Lieven Vandenberghes lecture notes. definition. subgradient calculus

Subgradient. Acknowledgement: this slides is based on Prof. Lieven Vandenberghes lecture notes. definition. subgradient calculus 1/41 Subgradient Acknowledgement: this slides is based on Prof. Lieven Vandenberghes lecture notes definition subgradient calculus duality and optimality conditions directional derivative Basic inequality

More information

PARALLEL SUBGRADIENT METHOD FOR NONSMOOTH CONVEX OPTIMIZATION WITH A SIMPLE CONSTRAINT

PARALLEL SUBGRADIENT METHOD FOR NONSMOOTH CONVEX OPTIMIZATION WITH A SIMPLE CONSTRAINT Linear and Nonlinear Analysis Volume 1, Number 1, 2015, 1 PARALLEL SUBGRADIENT METHOD FOR NONSMOOTH CONVEX OPTIMIZATION WITH A SIMPLE CONSTRAINT KAZUHIRO HISHINUMA AND HIDEAKI IIDUKA Abstract. In this

More information

Constraint qualifications for convex inequality systems with applications in constrained optimization

Constraint qualifications for convex inequality systems with applications in constrained optimization Constraint qualifications for convex inequality systems with applications in constrained optimization Chong Li, K. F. Ng and T. K. Pong Abstract. For an inequality system defined by an infinite family

More information

The local equicontinuity of a maximal monotone operator

The local equicontinuity of a maximal monotone operator arxiv:1410.3328v2 [math.fa] 3 Nov 2014 The local equicontinuity of a maximal monotone operator M.D. Voisei Abstract The local equicontinuity of an operator T : X X with proper Fitzpatrick function ϕ T

More information

On Optimality Conditions for Pseudoconvex Programming in Terms of Dini Subdifferentials

On Optimality Conditions for Pseudoconvex Programming in Terms of Dini Subdifferentials Int. Journal of Math. Analysis, Vol. 7, 2013, no. 18, 891-898 HIKARI Ltd, www.m-hikari.com On Optimality Conditions for Pseudoconvex Programming in Terms of Dini Subdifferentials Jaddar Abdessamad Mohamed

More information

Continuous Sets and Non-Attaining Functionals in Reflexive Banach Spaces

Continuous Sets and Non-Attaining Functionals in Reflexive Banach Spaces Laboratoire d Arithmétique, Calcul formel et d Optimisation UMR CNRS 6090 Continuous Sets and Non-Attaining Functionals in Reflexive Banach Spaces Emil Ernst Michel Théra Rapport de recherche n 2004-04

More information

OPTIMALITY CONDITIONS FOR D.C. VECTOR OPTIMIZATION PROBLEMS UNDER D.C. CONSTRAINTS. N. Gadhi, A. Metrane

OPTIMALITY CONDITIONS FOR D.C. VECTOR OPTIMIZATION PROBLEMS UNDER D.C. CONSTRAINTS. N. Gadhi, A. Metrane Serdica Math. J. 30 (2004), 17 32 OPTIMALITY CONDITIONS FOR D.C. VECTOR OPTIMIZATION PROBLEMS UNDER D.C. CONSTRAINTS N. Gadhi, A. Metrane Communicated by A. L. Dontchev Abstract. In this paper, we establish

More information

A user s guide to Lojasiewicz/KL inequalities

A user s guide to Lojasiewicz/KL inequalities Other A user s guide to Lojasiewicz/KL inequalities Toulouse School of Economics, Université Toulouse I SLRA, Grenoble, 2015 Motivations behind KL f : R n R smooth ẋ(t) = f (x(t)) or x k+1 = x k λ k f

More information

Optimization and Optimal Control in Banach Spaces

Optimization and Optimal Control in Banach Spaces Optimization and Optimal Control in Banach Spaces Bernhard Schmitzer October 19, 2017 1 Convex non-smooth optimization with proximal operators Remark 1.1 (Motivation). Convex optimization: easier to solve,

More information

4. Algebra and Duality

4. Algebra and Duality 4-1 Algebra and Duality P. Parrilo and S. Lall, CDC 2003 2003.12.07.01 4. Algebra and Duality Example: non-convex polynomial optimization Weak duality and duality gap The dual is not intrinsic The cone

More information

Convex series of convex functions with applications to Statistical Mechanics

Convex series of convex functions with applications to Statistical Mechanics Convex series of convex functions with applications to Statistical Mechanics Constantin Zălinescu University Alexandru Ioan Cuza Iaşi Faculty of Mathematics zalinesc@uaic.ro Melbourne, MODU2016 Motivation

More information

Maximal monotone operators, convex functions and a special family of enlargements.

Maximal monotone operators, convex functions and a special family of enlargements. Maximal monotone operators, convex functions and a special family of enlargements. Regina Sandra Burachik Engenharia de Sistemas e Computação, COPPE UFRJ, CP 68511, Rio de Janeiro RJ, 21945 970, Brazil.

More information

An exponential family of distributions is a parametric statistical model having densities with respect to some positive measure λ of the form.

An exponential family of distributions is a parametric statistical model having densities with respect to some positive measure λ of the form. Stat 8112 Lecture Notes Asymptotics of Exponential Families Charles J. Geyer January 23, 2013 1 Exponential Families An exponential family of distributions is a parametric statistical model having densities

More information

Convex Analysis Notes. Lecturer: Adrian Lewis, Cornell ORIE Scribe: Kevin Kircher, Cornell MAE

Convex Analysis Notes. Lecturer: Adrian Lewis, Cornell ORIE Scribe: Kevin Kircher, Cornell MAE Convex Analysis Notes Lecturer: Adrian Lewis, Cornell ORIE Scribe: Kevin Kircher, Cornell MAE These are notes from ORIE 6328, Convex Analysis, as taught by Prof. Adrian Lewis at Cornell University in the

More information

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 17

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 17 EE/ACM 150 - Applications of Convex Optimization in Signal Processing and Communications Lecture 17 Andre Tkacenko Signal Processing Research Group Jet Propulsion Laboratory May 29, 2012 Andre Tkacenko

More information

AN INEXACT HYBRID GENERALIZED PROXIMAL POINT ALGORITHM AND SOME NEW RESULTS ON THE THEORY OF BREGMAN FUNCTIONS. May 14, 1998 (Revised March 12, 1999)

AN INEXACT HYBRID GENERALIZED PROXIMAL POINT ALGORITHM AND SOME NEW RESULTS ON THE THEORY OF BREGMAN FUNCTIONS. May 14, 1998 (Revised March 12, 1999) AN INEXACT HYBRID GENERALIZED PROXIMAL POINT ALGORITHM AND SOME NEW RESULTS ON THE THEORY OF BREGMAN FUNCTIONS M. V. Solodov and B. F. Svaiter May 14, 1998 (Revised March 12, 1999) ABSTRACT We present

More information

Weak sharp minima on Riemannian manifolds 1

Weak sharp minima on Riemannian manifolds 1 1 Chong Li Department of Mathematics Zhejiang University Hangzhou, 310027, P R China cli@zju.edu.cn April. 2010 Outline 1 2 Extensions of some results for optimization problems on Banach spaces 3 4 Some

More information

Primal-dual Subgradient Method for Convex Problems with Functional Constraints

Primal-dual Subgradient Method for Convex Problems with Functional Constraints Primal-dual Subgradient Method for Convex Problems with Functional Constraints Yurii Nesterov, CORE/INMA (UCL) Workshop on embedded optimization EMBOPT2014 September 9, 2014 (Lucca) Yu. Nesterov Primal-dual

More information

An inexact subgradient algorithm for Equilibrium Problems

An inexact subgradient algorithm for Equilibrium Problems Volume 30, N. 1, pp. 91 107, 2011 Copyright 2011 SBMAC ISSN 0101-8205 www.scielo.br/cam An inexact subgradient algorithm for Equilibrium Problems PAULO SANTOS 1 and SUSANA SCHEIMBERG 2 1 DM, UFPI, Teresina,

More information

On the convergence properties of the projected gradient method for convex optimization

On the convergence properties of the projected gradient method for convex optimization Computational and Applied Mathematics Vol. 22, N. 1, pp. 37 52, 2003 Copyright 2003 SBMAC On the convergence properties of the projected gradient method for convex optimization A. N. IUSEM* Instituto de

More information

Week 3: Faces of convex sets

Week 3: Faces of convex sets Week 3: Faces of convex sets Conic Optimisation MATH515 Semester 018 Vera Roshchina School of Mathematics and Statistics, UNSW August 9, 018 Contents 1. Faces of convex sets 1. Minkowski theorem 3 3. Minimal

More information

Relationships between upper exhausters and the basic subdifferential in variational analysis

Relationships between upper exhausters and the basic subdifferential in variational analysis J. Math. Anal. Appl. 334 (2007) 261 272 www.elsevier.com/locate/jmaa Relationships between upper exhausters and the basic subdifferential in variational analysis Vera Roshchina City University of Hong

More information

Active sets, steepest descent, and smooth approximation of functions

Active sets, steepest descent, and smooth approximation of functions Active sets, steepest descent, and smooth approximation of functions Dmitriy Drusvyatskiy School of ORIE, Cornell University Joint work with Alex D. Ioffe (Technion), Martin Larsson (EPFL), and Adrian

More information

The sum of two maximal monotone operator is of type FPV

The sum of two maximal monotone operator is of type FPV CJMS. 5(1)(2016), 17-21 Caspian Journal of Mathematical Sciences (CJMS) University of Mazandaran, Iran http://cjms.journals.umz.ac.ir ISSN: 1735-0611 The sum of two maximal monotone operator is of type

More information

Stability of efficient solutions for semi-infinite vector optimization problems

Stability of efficient solutions for semi-infinite vector optimization problems Stability of efficient solutions for semi-infinite vector optimization problems Z. Y. Peng, J. T. Zhou February 6, 2016 Abstract This paper is devoted to the study of the stability of efficient solutions

More information

Math 273a: Optimization Subgradients of convex functions

Math 273a: Optimization Subgradients of convex functions Math 273a: Optimization Subgradients of convex functions Made by: Damek Davis Edited by Wotao Yin Department of Mathematics, UCLA Fall 2015 online discussions on piazza.com 1 / 42 Subgradients Assumptions

More information

One Mirror Descent Algorithm for Convex Constrained Optimization Problems with Non-Standard Growth Properties

One Mirror Descent Algorithm for Convex Constrained Optimization Problems with Non-Standard Growth Properties One Mirror Descent Algorithm for Convex Constrained Optimization Problems with Non-Standard Growth Properties Fedor S. Stonyakin 1 and Alexander A. Titov 1 V. I. Vernadsky Crimean Federal University, Simferopol,

More information

Convex Functions. Pontus Giselsson

Convex Functions. Pontus Giselsson Convex Functions Pontus Giselsson 1 Today s lecture lower semicontinuity, closure, convex hull convexity preserving operations precomposition with affine mapping infimal convolution image function supremum

More information

Methods for a Class of Convex. Functions. Stephen M. Robinson WP April 1996

Methods for a Class of Convex. Functions. Stephen M. Robinson WP April 1996 Working Paper Linear Convergence of Epsilon-Subgradient Descent Methods for a Class of Convex Functions Stephen M. Robinson WP-96-041 April 1996 IIASA International Institute for Applied Systems Analysis

More information

CONVEX ANALYSIS APPROACH TO D. C. PROGRAMMING: THEORY, ALGORITHMS AND APPLICATIONS

CONVEX ANALYSIS APPROACH TO D. C. PROGRAMMING: THEORY, ALGORITHMS AND APPLICATIONS ACTA MATHEMATICA VIETNAMICA Volume 22, Number 1, 1997, pp. 289 355 289 CONVEX ANALYSIS APPROACH TO D. C. PROGRAMMING: THEORY, ALGORITHMS AND APPLICATIONS PHAM DINH TAO AND LE THI HOAI AN Dedicated to Hoang

More information

The Trust Region Subproblem with Non-Intersecting Linear Constraints

The Trust Region Subproblem with Non-Intersecting Linear Constraints The Trust Region Subproblem with Non-Intersecting Linear Constraints Samuel Burer Boshi Yang February 21, 2013 Abstract This paper studies an extended trust region subproblem (etrs in which the trust region

More information

On the simplest expression of the perturbed Moore Penrose metric generalized inverse

On the simplest expression of the perturbed Moore Penrose metric generalized inverse Annals of the University of Bucharest (mathematical series) 4 (LXII) (2013), 433 446 On the simplest expression of the perturbed Moore Penrose metric generalized inverse Jianbing Cao and Yifeng Xue Communicated

More information

Maximal monotone operators are selfdual vector fields and vice-versa

Maximal monotone operators are selfdual vector fields and vice-versa Maximal monotone operators are selfdual vector fields and vice-versa Nassif Ghoussoub Department of Mathematics, University of British Columbia, Vancouver BC Canada V6T 1Z2 nassif@math.ubc.ca February

More information

Convexity, Duality, and Lagrange Multipliers

Convexity, Duality, and Lagrange Multipliers LECTURE NOTES Convexity, Duality, and Lagrange Multipliers Dimitri P. Bertsekas with assistance from Angelia Geary-Nedic and Asuman Koksal Massachusetts Institute of Technology Spring 2001 These notes

More information