A user s guide to Lojasiewicz/KL inequalities
|
|
- Morgan Norman Foster
- 6 years ago
- Views:
Transcription
1 Other A user s guide to Lojasiewicz/KL inequalities Toulouse School of Economics, Université Toulouse I SLRA, Grenoble, 2015
2 Motivations behind KL f : R n R smooth ẋ(t) = f (x(t)) or x k+1 = x k λ k f (x k ) Other KL answers a counterintuitive phenomenon of gradient systems. Bounded curves/sequences of with f C may oscillate indefinitely (for any choices of steps...) and generate paths with INFINITE LENGTH
3 Smooth counter-examples: Palis-De Melo, Absil et al. ẋ(t) = f (x(t)) Gradient: x k+1 = x k λ k f (x k ) Idea: Spiraling bump yields spiraling curves Other ( f (r, θ) = e (1 r 2 ) r 4 ) ( 1 r 4 + (1 r 2 ) 4 sin θ 1 ) 1 r 2, r 1
4 Counter-examples (continued) Other Oscillations in Optimization: Very bad predictions arbitrarily bad complexity Even worst behaviors for nonsmooth Same awful behaviors for more complex : e.g. Forward-Backward Solutions to these bad behaviors Topological approach: to avoid oscillations use (or definable) Ad hoc approach: study deformations of a class of nonsmooth sharp function
5 Definition : Other A subset of R n is a finite union of sets of the form {x R n : f i (x) = 0, g j (x) < 0, i I }, where I, J are finite and f i, g j : R n R are real polynomial. Stability by finite,, and complementation. A function or a mapping is if its graph is a set (same definition for real-extended function or multivalued mappings)
6 How to recognize sets /? Other Theorem (Tarski-Seidenberg) The image of a set by a linear projection is. Example: Let A be a subset of R n and f : R n R p. Then f (A) is. Proof. Then f (A) = {y R p : x A, y = f (x)}= Π x (graph f (A R p )), where Π x (x, y) = y for all (x, y) R n R p.
7 How to recognize sets /? Other Theorem (Tarski-Seidenberg) The image of a set by a linear projection is. Example: Let A be a subset of R n and f : R n R p. Then f (A) is. Proof. Then f (A) = {y R p : x A, y = f (x)}= Π x (graph f (A R p )), where Π x (x, y) = y for all (x, y) R n R p.
8 Checking ity is easy Other Typical illustration: The closure of a set A is semialgebraic. Proof. Observe that { A = Set then y R n : ɛ ]0, + [, x A, B = {(x, ɛ, y) R n ]0, + [ R n : where } n (x i y i ) 2 < ɛ 2. i=1 n (x i y i ) 2 < ɛ 2 } i=1 A = R n \ Π y ((R n R n \ Π ɛ (B))) Π ɛ (x, ɛ, y) = (x, y) Π y (x, y) = x.
9 Checking ity is easy Other Typical illustration: The closure of a set A is semialgebraic. Proof. Observe that { A = Set then y R n : ɛ ]0, + [, x A, B = {(x, ɛ, y) R n ]0, + [ R n : where } n (x i y i ) 2 < ɛ 2. i=1 n (x i y i ) 2 < ɛ 2 } i=1 A = R n \ Π y ((R n R n \ Π ɛ (B))) Π ɛ (x, ɛ, y) = (x, y) Π y (x, y) = x.
10 Other examples Other Other examples: Is the derivative? F : U R n R m, L = df (x) f (x + h) = f (x) + Lh + o( h ) ɛ > 0, δ > 0, graph df = {(x, L) U R n m, L = df (x)} ( h < δ, x + h U) f (x + h) f (x) Lh < ɛ h ) Exercise: show that g(x) = max{f (x, y) : y K} is whenever f, K.
11 Examples Other examples: Other min 1 2 Ax b 2 + λ x p (p rational) { } 1 min 2 A X B 2 + λrank (X ) : X matrix { } 1 min 2 AB M 2 : A 0, B 0 { } 1 min 2 AB M 2 : A 0, B 0, A 0 r, B s...
12 Fréchet subdifferential Other f : R n R {+ }. First-order formulation: Let x in dom f (Fréchet) subdifferential : p ˆ f (x) iff f (u) f (x) + p, u x + o( u x ), u R n. u f (u) u f (x) + p.u x
13 Subdifferential Other Definition (Subdifferential) It is denoted by f and defined through: x f (x) iff (x k, x k ) (x, x ) such that f (x k ) f (x) and f (u) f (x k ) + x k, u x k + o( u x k ). Example: f (x) = x 1, f (0) = B = [ 1, 1] n Set f (x) = min{ x : x f (x)} Properties (Critical point) Fermat s rule if f has a minimizer at x then f (x) 0. Conversely when 0 f (x), the point x is called critical.
14 An elementary remedy to gradient oscillation : Sharpness A function f : R n R {+ } is called sharp on the slice [r 0 < f < r 1 ] := {x R n : r 0 < f (x) < r 1 }, if there exists c > 0 f (x) c > 0, x [r 0 < f < r 1 ] Basic example f (x) = x Other Many works since 78, Rockafellar, Polyak, Ferris, Burke and many others..
15 Sharpness: example Nonconvex illustration with a continuum of minimizers Other
16 Finite convergence Other Why is sharpness interesting? Gradient curves reach the valley within a finite time. Easy to see the phenomenon on proximal descent Prox descent = formal setting for implicit gradient x + = x step. f (x + ) Prox operator: prox s f x = argmin{f (u) + 1 2s u x 2 : u R n } (Moreau) x + = prox step f (x) When f is sharp, convergence occurs in finite time!!
17 Proof Assume f is sharp, then if x k+1 f (x k+1 ) f (x k ) δ. is non critical Other Write (assume step is constant) f (x k+1 ) f (x k ) 1 2.step x k+1 x k 2 x k+1 x k step. f (x k+1 ) thus x k+1 x k c.step f (x k+1 ) f (x k ) c2.step 2 Set δ = 0.5c 2.step.
18 Measuring the default of sharpness Other 0 is a critical value of f (true up to a translation). Set [0 < f < r 0 ] := {x R n : 0 < f (x) < r 0 } and assume that there are no critical points in [0 < f < r 0 ]. f has the KL property on [0 < f < r 0 ] if there exists a function g : [0 < f < r 0 ] R {+ } whose sublevel sets are those of f, ie [f r] r (0,r0 ), and such that g(x) 1 for all x in [0 < f < r 0 ]. EX (Concentric circles) f (x) = x 2 and g(x) = x
19 Formal KL property Other Formally : Desingularizing on (0, r 0 ) : ϕ C([0, r 0 ), R + ), concave, ϕ C 1 (0, r 0 ), ϕ > 0 and ϕ(0) = 0. Definition f has the KL property on [0 < f < r 0 ] if there exists a desingularizing function ϕ such that (ϕ f )(x) 1, x [0 < f < r 0 ]. Local version : replace [0 < f < r 0 ] by the intersection of [0 < f < r 0 ] with a closed ball.
20 Key results Other Theorem [ Lojasiewicz (Hörmander) 1968; Kurdyka 98] If f is analytic or smooth and semialgebraic, it satisfies the Lojasiewicz inequality around each point of R n. Nonsmooth semialgebraic/subanalytic case 2006: [B-Daniilidis-Lewis] Take f : R n R {+ } lower semicontinuous and semialgebraic then f has KL property around each point. Many many satisfy KL inequality (B-Daniilidis- Lewis-Shiota 2007, Attouch-B-Redont ) More to come
21 Descent at large Other Let f : R n R {+ } be a proper lower semicontinuous function; a, b > 0. Let x k be a sequence in dom f such that Sufficient decrease condition f (x k+1 ) + a x k+1 x k 2 f (x k ); k 0 Relative error condition For each k N, there exists w k+1 f (x k+1 ) such that w k+1 b x k+1 x k ;
22 An abstract convergence theorem Other Theorem (Attouch-B-Svaiter 2012) Let f be a KL function and x k a descent sequence for f. If x k is bounded then it converges to a critical point of f. Corollary Let f be a coercive function and x k a descent sequence for f. Then x k converges to a critical point of f.
23 Rate of convergence Other Let x the (unique) limit point of x k. Assume that ϕ(s) = cs 1 θ with c > 0, θ [0, 1) (cover case). Theorem (Attouch-B 2009) (i) If θ = 0, the sequence (x k ) k N converges in a finite number of steps, (ii) If θ (0, 1 2 ] then there exist c > 0 and Q [0, 1) such that x k x c Q k, (iii) If θ ( 1 2, 1) then there exists c > 0 such that x k x c k 1 θ 2θ 1.
24 (Nonconvex) forward-backward splitting algorithm Minimizing nonsmooth+smooth structure: f = g + h Other with { h C 1 and h L Lipschitz continuous g lsc bounded from below + prox is easily computable Forward-backward splitting (Lions-Mercier 79): Let γ k be such that 0 < γ < γ k < γ < 1 L x k+1 prox γk g (x k γ k h(x k )) Theorem If the problem is coercive x k is a converging sequence
25 Sparse solutions of under-determined systems Other min{λ x Ax b 2 }, (Blumensath Davis 2009) Forward-backward splitting ) x k+1 prox γk λ 0 (x k γ k (A T Ax k A T b) where 0 < γ < γ k < γ < 1 A T A F, Theorem (Attouch-B-Svaiter) The sequence x k converges to a critical point of λ x Ax b 2
26 Low rank solutions of under-determined systems Other rank M= rank of M or a (or definable) surrogate of the rank whose prox is easily computable. See Hiriart-Urruty talk. min {λrank (M) + 12 } A M B 2, where A = linear operator Forward-backward splitting M k+1 prox γk λrank where 0 < γ < γ k < γ < 1 A T A F, ( ) x k γ k (A T A M k A T B) Theorem The sequence M k converges to a critical point of λrank M A M B 2
27 Gradient projection Other C a set, f a function. ) x k+1 P C (x k γ k f (x k ) Theorem The sequence x k converges to a critical point of the problem min C f
28 Gradient projection (Example) Other 1 2 A M B 2 + i Cs (x) with C s the set of matrix of at most rank s. P Cs M is known in closed form threhsold the inner diagonal matrix in the SVD Gradient-Projection method: ) M k+1 P Cs (M k γ k (A T A M k A T B) where γ k (ɛ, 1 A T A F ). No oscillations, M k converges to a critical point!!
29 Another illustration: Averaged projections Other x k+1 (1 θ) x k + θ ( 1 p ) p P Fi (x k ) + ɛ k, i=1 where θ (0, 1) and ɛ k M x k+1 x k Theorem (Inexact averaged projection method) F 1,..., F p be with C 2 boundaries or convex which satisfy p i=1 F i. If x 0 is sufficiently close to p i=1 F i, then x k converges to a feasible point x, i.e. such that x p F i. i=1
30 Alternating versions of prox algorithm Other F : R m 1... R mp R {+ } lsc Structure of F : p F (x 1,..., x p ) = f i (x i ) + Q(x 1,..., x p ) i=1 f i are proper lsc and Q is C 1. Gauss-Seidel method/ Prox (Auslender 1993, Grippo-Sciandrone 2000) x k+1 1 argmin u R m 1 x k+1 p F (u, x k 2,..., x k p ) µ 1 k u x 1 k 2 argmin F (x k+1 u R mp 1,..., xp 1 k+1, u) + 1 2µ p u x p k 2 k
31 Alternating versions of prox algorithm Other F : R m 1... R mp R {+ } lsc Structure of F : p F (x 1,..., x p ) = f i (x i ) + Q(x 1,..., x p ) i=1 f i are proper lsc and Q is C 1. Gauss-Seidel method/ Prox (Auslender 1993, Grippo-Sciandrone 2000) x k+1 1 argmin u R m 1 x k+1 p F (u, x k 2,..., x k p ) µ 1 k u x 1 k 2 argmin F (x k+1 u R mp 1,..., xp 1 k+1, u) + 1 2µ p u x p k 2 k
32 Convergence Other Theorem (Attouch-B.-Redont-Soubeyran 2010, B.-Combettes-Pesquet, 2010) Assume that F is a KL function (e.g. f i, Q are ), then the sequence (x k 1,..., x k p ) satisfies either (x k 1,..., x k p ) + (x k 1,..., x k p ) converges to a critical point of F Note that convergence automatically happens if either one of the f i or Q is coercive.
33 Proximal alternating linearized method Other F (x) = Q(x 1, x 2 ) + f 1 (x 1 ) + f 2 (x 2 ), x 1, Q(x 1, ) is C 1 with L 2 (x 1 ) > 0 Lipschitz continuous gradient (same for x 2 with L 1 (x 2 )) ( x1 k+1 prox 1 L 1 (x 2 k f x k ) x k+1 2 prox 1 L 2 (x k 1 ) f 1 Set x k+1 = (x k+1 1, x k+1 2 ). ) L 1 (x2 k) x 1 Q(x1 k, x2 k ) ( x k 1 1 L 2 (x k 1 ) x 2 Q(x k+1 1, x k 2 ). ). Theorem (Convergence of PALM (B., Teboulle, Sabach)) Any bounded sequence generated by PALM converges to a critical point of F.
34 Sparse Nonnegative Matrix Factorization Other Consider in NMF some sparsity measure as X 0 = number of nonzero entries { 1 min 2 A XY 2 F : X 0, X 0 r, } Y 0, Y 0 s
35 PALM for Sparse NMF Other 1. Take X 0 R m r + and Y 0 R+ r n. 2. Generate a sequence {( X k, Y k)} k N : 2.1. Take γ 1 > 1, set c k = γ 1 Y k ( Y k) T F and compute U k = X k 1 c k ( X k Y k A ) ( Y k) T ; X k+1 prox 1 c k ( U k ) = T α ( P+ ( U k )) Take γ 2 > 1, set d k = γ 2 X k+1 ( X k+1) T F and compute V k = Y k 1 d k ( X k+1 ) T ( X k+1 Y k A ) Y k+1 prox R2 d k ( V k ) = T β ( P+ ( V k )). We thus get the desired convergence result by our general theorem
36 Some complications: A SQP method Other We wish to solve min{f (x) : x R n, f i (x) 0} which can be written min{f + i C } with C = [f i 0, i]. We assume: f is L f Lipschitz i, f i is L fi Lipschitz continuous. Prox operator here are out of reach: too complex!!
37 Moving balls Other Classical Sequential Quadratic Programming (SQP) idea, replace by some simple approximations. Moving balls method (Auslender-Shefi-Teboulle, Math. Prog., 2010) min x R n f (x k) + f (x k )(x x k )+ L f 2 x x k 2 f 1 (x k ) + f 1(x k )(x x k )+ L f 1 2 x x k f m (x k ) + f m(x k )(x x k )+ L f m 2 x x k 2 0 Bad surprise: x k is not a descent sequence for f + i C.
38 Moving balls Introduce Other Theorem val(x) = min y R n f (x) + f (x)(y x)+ L f y x 2 2 f 1 (x) + f 1(x)(y x)+ L f 1 2 y x f m (x) + f m(x)(y x)+ L f m 2 y x 2 0 Good surprise: x k is a descent sequence for val. Assume Mangasarian-Fromovitz qualification condition. If x k is bounded it converges to a KKT point of the original problem.
A semi-algebraic look at first-order methods
splitting A semi-algebraic look at first-order Université de Toulouse / TSE Nesterov s 60th birthday, Les Houches, 2016 in large-scale first-order optimization splitting Start with a reasonable FOM (some
More informationSequential convex programming,: value function and convergence
Sequential convex programming,: value function and convergence Edouard Pauwels joint work with Jérôme Bolte Journées MODE Toulouse March 23 2016 1 / 16 Introduction Local search methods for finite dimensional
More informationI P IANO : I NERTIAL P ROXIMAL A LGORITHM FOR N ON -C ONVEX O PTIMIZATION
I P IANO : I NERTIAL P ROXIMAL A LGORITHM FOR N ON -C ONVEX O PTIMIZATION Peter Ochs University of Freiburg Germany 17.01.2017 joint work with: Thomas Brox and Thomas Pock c 2017 Peter Ochs ipiano c 1
More informationHedy Attouch, Jérôme Bolte, Benar Svaiter. To cite this version: HAL Id: hal
Convergence of descent methods for semi-algebraic and tame problems: proximal algorithms, forward-backward splitting, and regularized Gauss-Seidel methods Hedy Attouch, Jérôme Bolte, Benar Svaiter To cite
More informationFrom error bounds to the complexity of first-order descent methods for convex functions
From error bounds to the complexity of first-order descent methods for convex functions Nguyen Trong Phong-TSE Joint work with Jérôme Bolte, Juan Peypouquet, Bruce Suter. Toulouse, 23-25, March, 2016 Journées
More informationReceived: 15 December 2010 / Accepted: 30 July 2011 / Published online: 20 August 2011 Springer and Mathematical Optimization Society 2011
Math. Program., Ser. A (2013) 137:91 129 DOI 10.1007/s10107-011-0484-9 FULL LENGTH PAPER Convergence of descent methods for semi-algebraic and tame problems: proximal algorithms, forward backward splitting,
More informationVariable Metric Forward-Backward Algorithm
Variable Metric Forward-Backward Algorithm 1/37 Variable Metric Forward-Backward Algorithm for minimizing the sum of a differentiable function and a convex function E. Chouzenoux in collaboration with
More informationProximal Alternating Linearized Minimization for Nonconvex and Nonsmooth Problems
Proximal Alternating Linearized Minimization for Nonconvex and Nonsmooth Problems Jérôme Bolte Shoham Sabach Marc Teboulle Abstract We introduce a proximal alternating linearized minimization PALM) algorithm
More informationConvergence of descent methods for semi-algebraic and tame problems.
OPTIMIZATION, GAMES, AND DYNAMICS Institut Henri Poincaré November 28-29, 2011 Convergence of descent methods for semi-algebraic and tame problems. Hedy ATTOUCH Institut de Mathématiques et Modélisation
More informationTame variational analysis
Tame variational analysis Dmitriy Drusvyatskiy Mathematics, University of Washington Joint work with Daniilidis (Chile), Ioffe (Technion), and Lewis (Cornell) May 19, 2015 Theme: Semi-algebraic geometry
More informationMajorization-minimization procedures and convergence of SQP methods for semi-algebraic and tame programs
Majorization-minimization procedures and convergence of SQP methods for semi-algebraic and tame programs Jérôme Bolte and Edouard Pauwels September 29, 2014 Abstract In view of solving nonsmooth and nonconvex
More informationAn inertial forward-backward algorithm for the minimization of the sum of two nonconvex functions
An inertial forward-backward algorithm for the minimization of the sum of two nonconvex functions Radu Ioan Boţ Ernö Robert Csetnek Szilárd Csaba László October, 1 Abstract. We propose a forward-backward
More informationActive sets, steepest descent, and smooth approximation of functions
Active sets, steepest descent, and smooth approximation of functions Dmitriy Drusvyatskiy School of ORIE, Cornell University Joint work with Alex D. Ioffe (Technion), Martin Larsson (EPFL), and Adrian
More informationarxiv: v2 [math.oc] 6 Oct 2016
The value function approach to convergence analysis in composite optimization Edouard Pauwels IRIT-UPS, 118 route de Narbonne, 31062 Toulouse, France. arxiv:1604.01654v2 [math.oc] 6 Oct 2016 Abstract This
More informationConvergence rates for an inertial algorithm of gradient type associated to a smooth nonconvex minimization
Convergence rates for an inertial algorithm of gradient type associated to a smooth nonconvex minimization Szilárd Csaba László November, 08 Abstract. We investigate an inertial algorithm of gradient type
More informationVariational Analysis and Tame Optimization
Variational Analysis and Tame Optimization Aris Daniilidis http://mat.uab.cat/~arisd Universitat Autònoma de Barcelona April 15 17, 2010 PLAN OF THE TALK Nonsmooth analysis Genericity of pathological situations
More informationA memory gradient algorithm for l 2 -l 0 regularization with applications to image restoration
A memory gradient algorithm for l 2 -l 0 regularization with applications to image restoration E. Chouzenoux, A. Jezierska, J.-C. Pesquet and H. Talbot Université Paris-Est Lab. d Informatique Gaspard
More informationDouglas-Rachford splitting for nonconvex feasibility problems
Douglas-Rachford splitting for nonconvex feasibility problems Guoyin Li Ting Kei Pong Jan 3, 015 Abstract We adapt the Douglas-Rachford DR) splitting method to solve nonconvex feasibility problems by studying
More informationEE 546, Univ of Washington, Spring Proximal mapping. introduction. review of conjugate functions. proximal mapping. Proximal mapping 6 1
EE 546, Univ of Washington, Spring 2012 6. Proximal mapping introduction review of conjugate functions proximal mapping Proximal mapping 6 1 Proximal mapping the proximal mapping (prox-operator) of a convex
More informationFirst Order Methods beyond Convexity and Lipschitz Gradient Continuity with Applications to Quadratic Inverse Problems
First Order Methods beyond Convexity and Lipschitz Gradient Continuity with Applications to Quadratic Inverse Problems Jérôme Bolte Shoham Sabach Marc Teboulle Yakov Vaisbourd June 20, 2017 Abstract We
More informationA gradient type algorithm with backward inertial steps for a nonconvex minimization
A gradient type algorithm with backward inertial steps for a nonconvex minimization Cristian Daniel Alecsa Szilárd Csaba László Adrian Viorel November 22, 208 Abstract. We investigate an algorithm of gradient
More informationOn the iterate convergence of descent methods for convex optimization
On the iterate convergence of descent methods for convex optimization Clovis C. Gonzaga March 1, 2014 Abstract We study the iterate convergence of strong descent algorithms applied to convex functions.
More informationarxiv: v2 [math.oc] 21 Nov 2017
Unifying abstract inexact convergence theorems and block coordinate variable metric ipiano arxiv:1602.07283v2 [math.oc] 21 Nov 2017 Peter Ochs Mathematical Optimization Group Saarland University Germany
More informationConvex Functions. Pontus Giselsson
Convex Functions Pontus Giselsson 1 Today s lecture lower semicontinuity, closure, convex hull convexity preserving operations precomposition with affine mapping infimal convolution image function supremum
More informationBlock Coordinate Descent for Regularized Multi-convex Optimization
Block Coordinate Descent for Regularized Multi-convex Optimization Yangyang Xu and Wotao Yin CAAM Department, Rice University February 15, 2013 Multi-convex optimization Model definition Applications Outline
More informationSpring School on Variational Analysis Paseky nad Jizerou, Czech Republic (April 20 24, 2009)
Spring School on Variational Analysis Paseky nad Jizerou, Czech Republic (April 20 24, 2009) GRADIENT DYNAMICAL SYSTEMS, TAME OPTIMIZATION AND APPLICATIONS Aris DANIILIDIS Departament de Matemàtiques Universitat
More information6. Proximal gradient method
L. Vandenberghe EE236C (Spring 2016) 6. Proximal gradient method motivation proximal mapping proximal gradient method with fixed step size proximal gradient method with line search 6-1 Proximal mapping
More informationAlgorithmes de minimisation alternée avec frottement: H. Attouch
Algorithmes de minimisation alternée avec frottement: Applications aux E.D.P., problèmes inverses et jeux dynamiques. H. Attouch Institut de Mathématiques et de Modélisation de Montpellier UMR CNRS 5149
More informationComposite nonlinear models at scale
Composite nonlinear models at scale Dmitriy Drusvyatskiy Mathematics, University of Washington Joint work with D. Davis (Cornell), M. Fazel (UW), A.S. Lewis (Cornell) C. Paquette (Lehigh), and S. Roy (UW)
More informationOn the convergence of a regularized Jacobi algorithm for convex optimization
On the convergence of a regularized Jacobi algorithm for convex optimization Goran Banjac, Kostas Margellos, and Paul J. Goulart Abstract In this paper we consider the regularized version of the Jacobi
More informationProximal methods. S. Villa. October 7, 2014
Proximal methods S. Villa October 7, 2014 1 Review of the basics Often machine learning problems require the solution of minimization problems. For instance, the ERM algorithm requires to solve a problem
More informationApproaching monotone inclusion problems via second order dynamical systems with linear and anisotropic damping
March 0, 206 3:4 WSPC Proceedings - 9in x 6in secondorderanisotropicdamping206030 page Approaching monotone inclusion problems via second order dynamical systems with linear and anisotropic damping Radu
More informationStrongly convex functions, Moreau envelopes and the generic nature of convex functions with strong minimizers
University of Wollongong Research Online Faculty of Engineering and Information Sciences - Papers: Part B Faculty of Engineering and Information Sciences 206 Strongly convex functions, Moreau envelopes
More informationarxiv: v1 [math.oc] 11 Jan 2008
Alternating minimization and projection methods for nonconvex problems 1 Hedy ATTOUCH 2, Jérôme BOLTE 3, Patrick REDONT 2, Antoine SOUBEYRAN 4. arxiv:0801.1780v1 [math.oc] 11 Jan 2008 Abstract We study
More informationConvex Optimization Conjugate, Subdifferential, Proximation
1 Lecture Notes, HCI, 3.11.211 Chapter 6 Convex Optimization Conjugate, Subdifferential, Proximation Bastian Goldlücke Computer Vision Group Technical University of Munich 2 Bastian Goldlücke Overview
More informationA New Look at First Order Methods Lifting the Lipschitz Gradient Continuity Restriction
A New Look at First Order Methods Lifting the Lipschitz Gradient Continuity Restriction Marc Teboulle School of Mathematical Sciences Tel Aviv University Joint work with H. Bauschke and J. Bolte Optimization
More informationLocal strong convexity and local Lipschitz continuity of the gradient of convex functions
Local strong convexity and local Lipschitz continuity of the gradient of convex functions R. Goebel and R.T. Rockafellar May 23, 2007 Abstract. Given a pair of convex conjugate functions f and f, we investigate
More informationThe proximal mapping
The proximal mapping http://bicmr.pku.edu.cn/~wenzw/opt-2016-fall.html Acknowledgement: this slides is based on Prof. Lieven Vandenberghes lecture notes Outline 2/37 1 closed function 2 Conjugate function
More informationminimize x subject to (x 2)(x 4) u,
Math 6366/6367: Optimization and Variational Methods Sample Preliminary Exam Questions 1. Suppose that f : [, L] R is a C 2 -function with f () on (, L) and that you have explicit formulae for
More informationA splitting minimization method on geodesic spaces
A splitting minimization method on geodesic spaces J.X. Cruz Neto DM, Universidade Federal do Piauí, Teresina, PI 64049-500, BR B.P. Lima DM, Universidade Federal do Piauí, Teresina, PI 64049-500, BR P.A.
More informationarxiv: v2 [math.oc] 22 Jun 2017
A Proximal Difference-of-convex Algorithm with Extrapolation Bo Wen Xiaojun Chen Ting Kei Pong arxiv:1612.06265v2 [math.oc] 22 Jun 2017 June 17, 2017 Abstract We consider a class of difference-of-convex
More informationSubgradient. Acknowledgement: this slides is based on Prof. Lieven Vandenberghes lecture notes. definition. subgradient calculus
1/41 Subgradient Acknowledgement: this slides is based on Prof. Lieven Vandenberghes lecture notes definition subgradient calculus duality and optimality conditions directional derivative Basic inequality
More informationA Majorize-Minimize subspace approach for l 2 -l 0 regularization with applications to image processing
A Majorize-Minimize subspace approach for l 2 -l 0 regularization with applications to image processing Emilie Chouzenoux emilie.chouzenoux@univ-mlv.fr Université Paris-Est Lab. d Informatique Gaspard
More informationDownloaded 09/27/13 to Redistribution subject to SIAM license or copyright; see
SIAM J. OPTIM. Vol. 23, No., pp. 256 267 c 203 Society for Industrial and Applied Mathematics TILT STABILITY, UNIFORM QUADRATIC GROWTH, AND STRONG METRIC REGULARITY OF THE SUBDIFFERENTIAL D. DRUSVYATSKIY
More informationIterative Convex Optimization Algorithms; Part One: Using the Baillon Haddad Theorem
Iterative Convex Optimization Algorithms; Part One: Using the Baillon Haddad Theorem Charles Byrne (Charles Byrne@uml.edu) http://faculty.uml.edu/cbyrne/cbyrne.html Department of Mathematical Sciences
More information(x k ) sequence in F, lim x k = x x F. If F : R n R is a function, level sets and sublevel sets of F are any sets of the form (respectively);
STABILITY OF EQUILIBRIA AND LIAPUNOV FUNCTIONS. By topological properties in general we mean qualitative geometric properties (of subsets of R n or of functions in R n ), that is, those that don t depend
More informationOptimality, identifiability, and sensitivity
Noname manuscript No. (will be inserted by the editor) Optimality, identifiability, and sensitivity D. Drusvyatskiy A. S. Lewis Received: date / Accepted: date Abstract Around a solution of an optimization
More informationChapter 2: Unconstrained Extrema
Chapter 2: Unconstrained Extrema Math 368 c Copyright 2012, 2013 R Clark Robinson May 22, 2013 Chapter 2: Unconstrained Extrema 1 Types of Sets Definition For p R n and r > 0, the open ball about p of
More informationAuxiliary-Function Methods in Optimization
Auxiliary-Function Methods in Optimization Charles Byrne (Charles Byrne@uml.edu) http://faculty.uml.edu/cbyrne/cbyrne.html Department of Mathematical Sciences University of Massachusetts Lowell Lowell,
More informationIdentifying Active Constraints via Partial Smoothness and Prox-Regularity
Journal of Convex Analysis Volume 11 (2004), No. 2, 251 266 Identifying Active Constraints via Partial Smoothness and Prox-Regularity W. L. Hare Department of Mathematics, Simon Fraser University, Burnaby,
More informationDuality in Linear Programs. Lecturer: Ryan Tibshirani Convex Optimization /36-725
Duality in Linear Programs Lecturer: Ryan Tibshirani Convex Optimization 10-725/36-725 1 Last time: proximal gradient descent Consider the problem x g(x) + h(x) with g, h convex, g differentiable, and
More informationON GAP FUNCTIONS OF VARIATIONAL INEQUALITY IN A BANACH SPACE. Sangho Kum and Gue Myung Lee. 1. Introduction
J. Korean Math. Soc. 38 (2001), No. 3, pp. 683 695 ON GAP FUNCTIONS OF VARIATIONAL INEQUALITY IN A BANACH SPACE Sangho Kum and Gue Myung Lee Abstract. In this paper we are concerned with theoretical properties
More informationALGORITHMS FOR MINIMIZING DIFFERENCES OF CONVEX FUNCTIONS AND APPLICATIONS
ALGORITHMS FOR MINIMIZING DIFFERENCES OF CONVEX FUNCTIONS AND APPLICATIONS Mau Nam Nguyen (joint work with D. Giles and R. B. Rector) Fariborz Maseeh Department of Mathematics and Statistics Portland State
More informationA Unified Approach to Proximal Algorithms using Bregman Distance
A Unified Approach to Proximal Algorithms using Bregman Distance Yi Zhou a,, Yingbin Liang a, Lixin Shen b a Department of Electrical Engineering and Computer Science, Syracuse University b Department
More informationOptimization and Optimal Control in Banach Spaces
Optimization and Optimal Control in Banach Spaces Bernhard Schmitzer October 19, 2017 1 Convex non-smooth optimization with proximal operators Remark 1.1 (Motivation). Convex optimization: easier to solve,
More informationarxiv: v1 [math.oc] 7 Jan 2019
LOCAL MINIMIZERS OF SEMI-ALGEBRAIC FUNCTIONS TIÊ N-SO. N PHẠM arxiv:1901.01698v1 [math.oc] 7 Jan 2019 Abstract. Consider a semi-algebraic function f: R n R, which is continuous around a point x R n. Using
More informationA Proximal Alternating Direction Method for Semi-Definite Rank Minimization (Supplementary Material)
A Proximal Alternating Direction Method for Semi-Definite Rank Minimization (Supplementary Material) Ganzhao Yuan and Bernard Ghanem King Abdullah University of Science and Technology (KAUST), Saudi Arabia
More informationarxiv: v3 [math.oc] 22 Jan 2013
Proximal alternating minimization and projection methods for nonconvex problems. An approach based on the Kurdyka- Lojasiewicz inequality 1 Hedy ATTOUCH 2, Jérôme BOLTE 3, Patrick REDONT 2, Antoine SOUBEYRAN
More informationMathematical methods for Image Processing
Mathematical methods for Image Processing François Malgouyres Institut de Mathématiques de Toulouse, France invitation by Jidesh P., NITK Surathkal funding Global Initiative on Academic Network Oct. 23
More informationSolution Methods. Richard Lusby. Department of Management Engineering Technical University of Denmark
Solution Methods Richard Lusby Department of Management Engineering Technical University of Denmark Lecture Overview (jg Unconstrained Several Variables Quadratic Programming Separable Programming SUMT
More informationA characterization of essentially strictly convex functions on reflexive Banach spaces
A characterization of essentially strictly convex functions on reflexive Banach spaces Michel Volle Département de Mathématiques Université d Avignon et des Pays de Vaucluse 74, rue Louis Pasteur 84029
More informationSplitting methods for decomposing separable convex programs
Splitting methods for decomposing separable convex programs Philippe Mahey LIMOS - ISIMA - Université Blaise Pascal PGMO, ENSTA 2013 October 4, 2013 1 / 30 Plan 1 Max Monotone Operators Proximal techniques
More informationCharacterizations of Lojasiewicz inequalities: subgradient flows, talweg, convexity.
Characterizations of Lojasiewicz inequalities: subgradient flows, talweg, convexity. Jérôme BOLTE, Aris DANIILIDIS, Olivier LEY, Laurent MAZET. Abstract The classical Lojasiewicz inequality and its extensions
More informationStructural and Multidisciplinary Optimization. P. Duysinx and P. Tossings
Structural and Multidisciplinary Optimization P. Duysinx and P. Tossings 2018-2019 CONTACTS Pierre Duysinx Institut de Mécanique et du Génie Civil (B52/3) Phone number: 04/366.91.94 Email: P.Duysinx@uliege.be
More informationMaster 2 MathBigData. 3 novembre CMAP - Ecole Polytechnique
Master 2 MathBigData S. Gaïffas 1 3 novembre 2014 1 CMAP - Ecole Polytechnique 1 Supervised learning recap Introduction Loss functions, linearity 2 Penalization Introduction Ridge Sparsity Lasso 3 Some
More informationProximal Gradient Descent and Acceleration. Ryan Tibshirani Convex Optimization /36-725
Proximal Gradient Descent and Acceleration Ryan Tibshirani Convex Optimization 10-725/36-725 Last time: subgradient method Consider the problem min f(x) with f convex, and dom(f) = R n. Subgradient method:
More informationSequential Unconstrained Minimization: A Survey
Sequential Unconstrained Minimization: A Survey Charles L. Byrne February 21, 2013 Abstract The problem is to minimize a function f : X (, ], over a non-empty subset C of X, where X is an arbitrary set.
More informationConvex Optimization. (EE227A: UC Berkeley) Lecture 15. Suvrit Sra. (Gradient methods III) 12 March, 2013
Convex Optimization (EE227A: UC Berkeley) Lecture 15 (Gradient methods III) 12 March, 2013 Suvrit Sra Optimal gradient methods 2 / 27 Optimal gradient methods We saw following efficiency estimates for
More informationHence, (f(x) f(x 0 )) 2 + (g(x) g(x 0 )) 2 < ɛ
Matthew Straughn Math 402 Homework 5 Homework 5 (p. 429) 13.3.5, 13.3.6 (p. 432) 13.4.1, 13.4.2, 13.4.7*, 13.4.9 (p. 448-449) 14.2.1, 14.2.2 Exercise 13.3.5. Let (X, d X ) be a metric space, and let f
More informationOptimization for Learning and Big Data
Optimization for Learning and Big Data Donald Goldfarb Department of IEOR Columbia University Department of Mathematics Distinguished Lecture Series May 17-19, 2016. Lecture 1. First-Order Methods for
More informationOptimality, identifiability, and sensitivity
Noname manuscript No. (will be inserted by the editor) Optimality, identifiability, and sensitivity D. Drusvyatskiy A. S. Lewis Received: date / Accepted: date Abstract Around a solution of an optimization
More informationAN AUGMENTED LAGRANGIAN METHOD FOR l 1 -REGULARIZED OPTIMIZATION PROBLEMS WITH ORTHOGONALITY CONSTRAINTS
AN AUGMENTED LAGRANGIAN METHOD FOR l 1 -REGULARIZED OPTIMIZATION PROBLEMS WITH ORTHOGONALITY CONSTRAINTS WEIQIANG CHEN, HUI JI, AND YANFEI YOU Abstract. A class of l 1 -regularized optimization problems
More informationc???? Society for Industrial and Applied Mathematics Vol. 1, No. 1, pp ,???? 000
SIAM J. OPTIMIZATION c???? Society for Industrial and Applied Mathematics Vol. 1, No. 1, pp. 000 000,???? 000 TILT STABILITY OF A LOCAL MINIMUM * R. A. POLIQUIN AND R. T. ROCKAFELLAR Abstract. The behavior
More informationA Geometric Framework for Nonconvex Optimization Duality using Augmented Lagrangian Functions
A Geometric Framework for Nonconvex Optimization Duality using Augmented Lagrangian Functions Angelia Nedić and Asuman Ozdaglar April 15, 2006 Abstract We provide a unifying geometric framework for the
More informationA Convergent Incoherent Dictionary Learning Algorithm for Sparse Coding
A Convergent Incoherent Dictionary Learning Algorithm for Sparse Coding Chenglong Bao, Yuhui Quan, Hui Ji Department of Mathematics National University of Singapore Abstract. Recently, sparse coding has
More informationA proximal-newton method for monotone inclusions in Hilbert spaces with complexity O(1/k 2 ).
H. ATTOUCH (Univ. Montpellier 2) Fast proximal-newton method Sept. 8-12, 2014 1 / 40 A proximal-newton method for monotone inclusions in Hilbert spaces with complexity O(1/k 2 ). Hedy ATTOUCH Université
More informationNon-smooth Non-convex Bregman Minimization: Unification and New Algorithms
JOTA manuscript No. (will be inserted by the editor) Non-smooth Non-convex Bregman Minimization: Unification and New Algorithms Peter Ochs Jalal Fadili Thomas Brox Received: date / Accepted: date Abstract
More informationOuter limits of subdifferentials for min-max type functions
Outer limits of subdifferentials for min-max type functions Andrew Eberhard, Vera Roshchina, Tian Sang August 31, 2017 arxiv:1701.02852v2 [math.oc] 30 Aug 2017 Abstract We generalise the outer subdifferential
More informationarxiv: v3 [math.oc] 27 Oct 2016
A Multi-step Inertial Forward Backward Splitting Method for Non-convex Optimization arxiv:1606.02118v3 [math.oc] 27 Oct 2016 Jingwei Liang Jalal M. Fadili Gabriel Peyré Abstract In this paper, we propose
More informationLARGE SCALE COMPOSITE OPTIMIZATION PROBLEMS WITH COUPLED OBJECTIVE FUNCTIONS: THEORY, ALGORITHMS AND APPLICATIONS CUI YING. (B.Sc.
LARGE SCALE COMPOSITE OPTIMIZATION PROBLEMS WITH COUPLED OBJECTIVE FUNCTIONS: THEORY, ALGORITHMS AND APPLICATIONS CUI YING (B.Sc., ZJU, China) A THESIS SUBMITTED FOR THE DEGREE OF DOCTOR OF PHILOSOPHY
More information6. Proximal gradient method
L. Vandenberghe EE236C (Spring 2013-14) 6. Proximal gradient method motivation proximal mapping proximal gradient method with fixed step size proximal gradient method with line search 6-1 Proximal mapping
More informationRecent Developments of Alternating Direction Method of Multipliers with Multi-Block Variables
Recent Developments of Alternating Direction Method of Multipliers with Multi-Block Variables Department of Systems Engineering and Engineering Management The Chinese University of Hong Kong 2014 Workshop
More informationNumerical Linear Algebra Primer. Ryan Tibshirani Convex Optimization /36-725
Numerical Linear Algebra Primer Ryan Tibshirani Convex Optimization 10-725/36-725 Last time: proximal gradient descent Consider the problem min g(x) + h(x) with g, h convex, g differentiable, and h simple
More information6-1 The Positivstellensatz P. Parrilo and S. Lall, ECC
6-1 The Positivstellensatz P. Parrilo and S. Lall, ECC 2003 2003.09.02.10 6. The Positivstellensatz Basic semialgebraic sets Semialgebraic sets Tarski-Seidenberg and quantifier elimination Feasibility
More informationSubdifferential representation of convex functions: refinements and applications
Subdifferential representation of convex functions: refinements and applications Joël Benoist & Aris Daniilidis Abstract Every lower semicontinuous convex function can be represented through its subdifferential
More informationPenalty PALM Method for Sparse Portfolio Selection Problems
Penalty PALM Method for Sparse Portfolio Selection Problems Yue Teng a, Li Yang b, Bo Yu a and Xiaoliang Song a a School of Mathematical Sciences, Dalian University of Technology, Dalian 116024, P.R. China;
More informationExistence and Approximation of Fixed Points of. Bregman Nonexpansive Operators. Banach Spaces
Existence and Approximation of Fixed Points of in Reflexive Banach Spaces Department of Mathematics The Technion Israel Institute of Technology Haifa 22.07.2010 Joint work with Prof. Simeon Reich General
More informationProximal Newton Method. Ryan Tibshirani Convex Optimization /36-725
Proximal Newton Method Ryan Tibshirani Convex Optimization 10-725/36-725 1 Last time: primal-dual interior-point method Given the problem min x subject to f(x) h i (x) 0, i = 1,... m Ax = b where f, h
More informationAmir Beck August 7, Abstract. 1 Introduction and Problem/Model Formulation. In this paper we consider the following minimization problem:
On the Convergence of Alternating Minimization for Convex Programming with Applications to Iteratively Reweighted Least Squares and Decomposition Schemes Amir Beck August 7, 04 Abstract This paper is concerned
More informationOPTIMALITY CONDITIONS FOR GLOBAL MINIMA OF NONCONVEX FUNCTIONS ON RIEMANNIAN MANIFOLDS
OPTIMALITY CONDITIONS FOR GLOBAL MINIMA OF NONCONVEX FUNCTIONS ON RIEMANNIAN MANIFOLDS S. HOSSEINI Abstract. A version of Lagrange multipliers rule for locally Lipschitz functions is presented. Using Lagrange
More informationA Unified Analysis of Nonconvex Optimization Duality and Penalty Methods with General Augmenting Functions
A Unified Analysis of Nonconvex Optimization Duality and Penalty Methods with General Augmenting Functions Angelia Nedić and Asuman Ozdaglar April 16, 2006 Abstract In this paper, we study a unifying framework
More informationCourse Summary Math 211
Course Summary Math 211 table of contents I. Functions of several variables. II. R n. III. Derivatives. IV. Taylor s Theorem. V. Differential Geometry. VI. Applications. 1. Best affine approximations.
More informationCompressed Sensing: a Subgradient Descent Method for Missing Data Problems
Compressed Sensing: a Subgradient Descent Method for Missing Data Problems ANZIAM, Jan 30 Feb 3, 2011 Jonathan M. Borwein Jointly with D. Russell Luke, University of Goettingen FRSC FAAAS FBAS FAA Director,
More informationTaylor-like models in nonsmooth optimization
Taylor-like models in nonsmooth optimization Dmitriy Drusvyatskiy Mathematics, University of Washington Joint work with Ioffe (Technion), Lewis (Cornell), and Paquette (UW) SIAM Optimization 2017 AFOSR,
More informationA proximal minimization algorithm for structured nonconvex and nonsmooth problems
A proximal minimization algorithm for structured nonconvex and nonsmooth problems Radu Ioan Boţ Ernö Robert Csetnek Dang-Khoa Nguyen May 8, 08 Abstract. We propose a proximal algorithm for minimizing objective
More informationProximal Newton Method. Zico Kolter (notes by Ryan Tibshirani) Convex Optimization
Proximal Newton Method Zico Kolter (notes by Ryan Tibshirani) Convex Optimization 10-725 Consider the problem Last time: quasi-newton methods min x f(x) with f convex, twice differentiable, dom(f) = R
More information1 Strict local optimality in unconstrained optimization
ORF 53 Lecture 14 Spring 016, Princeton University Instructor: A.A. Ahmadi Scribe: G. Hall Thursday, April 14, 016 When in doubt on the accuracy of these notes, please cross check with the instructor s
More informationInequality Constraints
Chapter 2 Inequality Constraints 2.1 Optimality Conditions Early in multivariate calculus we learn the significance of differentiability in finding minimizers. In this section we begin our study of the
More informationOptimality Conditions for Constrained Optimization
72 CHAPTER 7 Optimality Conditions for Constrained Optimization 1. First Order Conditions In this section we consider first order optimality conditions for the constrained problem P : minimize f 0 (x)
More information1 Introduction and preliminaries
Proximal Methods for a Class of Relaxed Nonlinear Variational Inclusions Abdellatif Moudafi Université des Antilles et de la Guyane, Grimaag B.P. 7209, 97275 Schoelcher, Martinique abdellatif.moudafi@martinique.univ-ag.fr
More information