Hessian Riemannian Gradient Flows in Convex Programming

Size: px
Start display at page:

Download "Hessian Riemannian Gradient Flows in Convex Programming"

Transcription

1 Hessian Riemannian Gradient Flows in Convex Programming Felipe Alvarez, Jérôme Bolte, Olivier Brahic INTERNATIONAL CONFERENCE ON MODELING AND OPTIMIZATION MODOPT 2004 Universidad de La Frontera, Temuco, Chile January 19-22, Hessian Riemannian Flows. p. 1/25

2 Outline 1. Motivation: scaling the Euclidean gradient. 2. Riemannian gradient flows on convex sets. 3. Hessian metrics, existence, convergence and examples. Hessian Riemannian Flows. p. 2/25

3 1. Motivation: gradient method Let f : R n R be a smooth function, x 0 R n. x k+1 = x k + λ k d k, Hessian Riemannian Flows. p. 3/25

4 1. Motivation: gradient method Let f : R n R be a smooth function, x 0 R n. x k+1 = x k + λ k d k, where λ k > 0 is a stepsize Hessian Riemannian Flows. p. 3/25

5 1. Motivation: gradient method Let f : R n R be a smooth function, x 0 R n. x k+1 = x k + λ k d k, where λ k > 0 is a stepsize and d k = f(x k ) is the steepest descent direction. Hessian Riemannian Flows. p. 3/25

6 1. Motivation: gradient method Let f : R n R be a smooth function, x 0 R n. x k+1 = x k + λ k d k, where λ k > 0 is a stepsize and d k = f(x k ) is the steepest descent direction. Continuous gradient method: dx dt = f(x), t > 0. Continuous flow discrete method. Hessian Riemannian Flows. p. 3/25

7 1. Scaling and Newton s method Newton s correction: x k+1 = x k λ k 2 f(x k ) 1 f(x k ). Continuous Newton s method: dx dt = 2 f(x) 1 f(x). Hessian Riemannian Flows. p. 4/25

8 1. Scaling and Newton s method Newton s correction: x k+1 = x k λ k 2 f(x k ) 1 f(x k ). Continuous Newton s method: dx dt = 2 f(x) 1 f(x). d f(x) = f(x). dt Hessian Riemannian Flows. p. 4/25

9 1. Scaling and Newton s method Newton s correction: x k+1 = x k λ k 2 f(x k ) 1 f(x k ). Continuous Newton s method: dx dt = 2 f(x) 1 f(x). d f(x) = f(x). dt f(x(t)) = e t f(x 0 ) Scale invariant rate of convergence on a straight line! back. Hessian Riemannian Flows. p. 4/25

10 1. Scaling and constraints Problem: min{f(x) x 0}. Hessian Riemannian Flows. p. 5/25

11 1. Scaling and constraints Problem: min{f(x) x 0}. ODE approach: dx i dt = x f i (x), i = 1,...,n. x i Hessian Riemannian Flows. p. 5/25

12 1. Scaling and constraints Problem: min{f(x) x 0}. ODE approach: Properties: d dt f(x) = n i=1 dx i dt = x f i (x), i = 1,...,n. x i f x i x i (x) 2 0 descent method on R n +. x i (0) > 0 t > 0, x i (t) > 0 interior point trajectory. Hessian Riemannian Flows. p. 5/25

13 1. Scaling and constraints Problem: min{f(x) x 0}. ODE approach: Properties: d dt f(x) = n i=1 dx i dt = x f i (x), i = 1,...,n. x i f x i x i (x) 2 0 descent method on R n +. x i (0) > 0 t > 0, x i (t) > 0 interior point trajectory. The equation may be written as d dt log(x i) = f (x) x i Scaling logarithmic barrier to force x(t) > 0! Hessian Riemannian Flows. p. 5/25

14 1. Hessian type scaling d dt log(x i) = f (x) x i Hessian Riemannian Flows. p. 6/25

15 1. Hessian type scaling where d dt log(x i) = f (x) x i d h(x) = f(x), dt h(x) = n x i log(x i ) x i i=1 Hessian Riemannian Flows. p. 6/25

16 1. Hessian type scaling where Thus d dt log(x i) = f (x) x i d h(x) = f(x), dt h(x) = n x i log(x i ) x i i=1 dx i dt = x f i (x) dx x i dt = 2 h(x) 1 f(x) Remark the analogy with Newton s method Hessian Riemannian Flows. p. 6/25

17 2. Riemannian gradient flows Problem: min f(x), with C = {x x C Rn x Q, Ax = b}. Q R n nonempty, open and convex. A R m n and b R m with m n. Hessian Riemannian Flows. p. 7/25

18 2. Riemannian gradient flows Problem: min f(x), with C = {x x C Rn x Q, Ax = b}. Q R n nonempty, open and convex. A R m n and b R m with m n. Strategy: Introduce a Riemannian metric H(x) S n ++ on Q, (u,v) x = H(x)u,v = Consider the gradient flow n i=1 H ij (x)u i v j, x Q. dx dt (t) = Hf(x(t)),t > 0, Hessian Riemannian Flows. p. 7/25

19 2. Riemannian gradient Let M, (, ) be a Riemannian manifold. T x M: tangent space to M at x M. Hessian Riemannian Flows. p. 8/25

20 2. Riemannian gradient Let M, (, ) be a Riemannian manifold. T x M: tangent space to M at x M. The gradient gradf of f C 1 (M; R) is uniquely determined by Tangency: for all x M, gradf(x) T x M. Duality: for all x M, v T x M, df(x)v = (gradf(x),v) x, where df(x) : T x M R is the differential of f. Hessian Riemannian Flows. p. 8/25

21 2. Riemannian gradient in our case M = Q {x R n Ax = b} with Q open set, then T x M Ker A. (, ) x = H(x), with a barrier/penalty effect near Q. Hessian Riemannian Flows. p. 9/25

22 2. Riemannian gradient in our case M = Q {x R n Ax = b} with Q open set, then T x M Ker A. (, ) x = H(x), with a barrier/penalty effect near Q. H f = H 1 [I A T (AH 1 A T ) 1 AH 1 ] f, where H stands for grad to stress the dependence on H. Hessian Riemannian Flows. p. 9/25

23 2. Riemannian gradient in our case M = Q {x R n Ax = b} with Q open set, then T x M Ker A. (, ) x = H(x), with a barrier/penalty effect near Q. H f = H 1 [I A T (AH 1 A T ) 1 AH 1 ] f, where H stands for grad to stress the dependence on H. Projection vector field in the tangent space. Scaling interior point method x(t) Q. Hessian Riemannian Flows. p. 9/25

24 2. Example C = n 1 := {x R n x 0, n i=1 x i = 1} Hessian Riemannian Flows. p. 10/25

25 2. Example C = n 1 := {x R n x 0, n i=1 x i = 1} Q = R n ++, A = (1,...,1) R 1 n and b = 1. M = {x R n x > 0, n i=1 x i = 1}. T x M = {v R n n i=1 v i = 0}. Hessian Riemannian Flows. p. 10/25

26 2. Example C = n 1 := {x R n x 0, n i=1 x i = 1} Q = R n ++, A = (1,...,1) R 1 n and b = 1. M = {x R n x > 0, n i=1 x i = 1}. T x M = {v R n n i=1 v i = 0}. Take H(x) = diag(1/x 1,...,1/x n ), then (u,v) x = n i=1 u i v j x i Shahshahani metric. dx i dt = x f i + x i n j=1 x i x j f x j Lotka-Volterra type eq. Karmarkar 90, Faybusovich 91,... Hessian Riemannian Flows. p. 10/25

27 2. Barrier effect: Legendre functions We focus on the case H(x) = 2 h(x), x Q with h : R n R {+ } is closed, convex and proper. int dom h = Q. (H 0 ) h is of Legendre type. h Q C 2 (Q; R) and 2 h(x) > 0. 2 h is locally Lipschitz. Hessian Riemannian Flows. p. 11/25

28 2. Barrier effect: Legendre functions We focus on the case H(x) = 2 h(x), x Q with h : R n R {+ } is closed, convex and proper. int dom h = Q. (H 0 ) h is of Legendre type. h Q C 2 (Q; R) and 2 h(x) > 0. 2 h is locally Lipschitz. h is strictly convex and C 1 on int dom h. int dom h x j x int dom h, h(x j ) +. back Hessian Riemannian Flows. p. 11/25

29 2. Example of Legendre function h(x) = i I θ(g i (x)). where I = {1,...,p}, g i C 3 (R n ) concave. Q = {x R n g i (x) > 0,i I}. x Q, span { g i (x) i I} = R n, and (i) (0, ) domθ [0, ), θ C 3 (0, ). (ii) lim s 0 + θ (s) = and s > 0, θ (s) > 0. (iii) Either θ is nonincreasing or i I,g i is affine. Hessian Riemannian Flows. p. 12/25

30 3. Questions Given x 0 M = Q {x R n Ax = b}: dx dt (t) = Hf(x(t)),t > 0, Hessian Riemannian Flows. p. 13/25

31 3. Questions Given x 0 M = Q {x R n Ax = b}: dx dt (t) = Hf(x(t)),t > 0, Well posedness: global existence for all t > 0. Asymptotic behavior: convergence to an equilibrium as t, rate of convergence,... Hessian Riemannian Flows. p. 13/25

32 3. Questions Given x 0 M = Q {x R n Ax = b}: dx dt (t) = Hf(x(t)),t > 0, Well posedness: global existence for all t > 0. Asymptotic behavior: convergence to an equilibrium as t, rate of convergence,... Main difficulty: singular behavior near Q. Classical results do not apply. Hessian Riemannian Flows. p. 13/25

33 3. Well posedness: global existence Thm. 2 The trajectory x(t) is defined for all t 0 under any of the following conditions: (C 1 ) {x C f(x) f(x 0 )} is bounded. (i) dom h = Q (C 2 ) (ii) y Q, γ R, {x C D h (y,x) γ} is bdd. (iii) Argmin C f and f quasiconvex. (C 3 ) K 0, L R such that x Q, H(x) 1 K x + L. Hessian Riemannian Flows. p. 14/25

34 3. Why Hessian metrics? Suppose f is convex and A = 0 and b = 0. Euclidean case: y Argmin C f x C, f(x),x y 0. Hessian Riemannian Flows. p. 15/25

35 3. Why Hessian metrics? Suppose f is convex and A = 0 and b = 0. Euclidean case: y Argmin C f x C, f(x),x y 0. ϕ y (x) = 1 x y 2 2 ẋ = f(x) d dt ϕ y(x) = ϕ y (x),ẋ = x y, f(x) 0. Hessian Riemannian Flows. p. 15/25

36 3. Why Hessian metrics? Suppose f is convex and A = 0 and b = 0. Euclidean case: y Argmin C f x C, f(x),x y 0. ϕ y (x) = 1 x y 2 2 ẋ = f(x) d dt ϕ y(x) = ϕ y (x),ẋ = x y, f(x) 0. ϕ y (x) is a Lyapunov function for the gradient flow Hessian Riemannian Flows. p. 15/25

37 3. Characterization of Hessian metrics Riemannian case: y Argmin Q f x Q, ( H f(x),x y) x 0. Hessian Riemannian Flows. p. 16/25

38 3. Characterization of Hessian metrics Riemannian case: y Argmin Q f x Q, ( H f(x),x y) x 0. Thm. 1 H C 1 (Q; S n ++) satisfies y Q, ϕ y C 1 (Q; R), H ϕ y (x) = x y? Hessian Riemannian Flows. p. 16/25

39 3. Characterization of Hessian metrics Riemannian case: y Argmin Q f x Q, ( H f(x),x y) x 0. Thm. 1 H C 1 (Q; S n ++) satisfies y Q, ϕ y C 1 (Q; R), H ϕ y (x) = x y h C 3 (Q) such that H = 2 h on Q and ϕ y (x) = D h (y,x) = h(y) h(x) h(x),y x = Bregman pseudo-distance induced by h. Hessian Riemannian Flows. p. 16/25

40 3. Implicit proximal iteration x k+1 Argmin { } f(x) + 1 λ k D h (x,x k ) Ax = b, 1 λ k [ h(x k+1 ) h(x k )] f(x k+1 ) + Im A T, Ax k+1 = b Bregman 67, Censor-Zenios 92, Teboulle 92, Eckstein 93, Kiwiel 97,... Hessian Riemannian Flows. p. 17/25

41 3. Implicit proximal iteration x k+1 Argmin { } f(x) + 1 λ k D h (x,x k ) Ax = b, 1 λ k [ h(x k+1 ) h(x k )] f(x k+1 ) + Im A T, Ax k+1 = b Bregman 67, Censor-Zenios 92, Teboulle 92, Eckstein 93, Kiwiel 97,... But dx dt = Hf(x) d h(x) f(x) + Im dt AT, Ax(t) = b, t 0. This link was already noticed by Iusem-Svaiter-Da Cruz Neto 99, together with convergence results for a linear objective function. Hessian Riemannian Flows. p. 17/25

42 3. Convergence: Bregman functions A Legendre function h with domh = Q is of Bregman type if (i) {x Q D h (y,x) γ} is bdd. y Q, γ R. (ii) y Q, y j y with y j Q, D h (y,y j ) 0. Hessian Riemannian Flows. p. 18/25

43 3. Convergence: Bregman functions A Legendre function h with domh = Q is of Bregman type if (i) {x Q D h (y,x) γ} is bdd. y Q, γ R. (ii) y Q, y j y with y j Q, D h (y,y j ) 0. Thm. 3 Suppose (H 0 ) with h of Bregman type. f is quasiconvex and Argmin C f. Then x C such that x(t) x as t + with f(x ) N Q (x ) + Ker A, where N Q (x ) is the normal cone to Q at x. Hessian Riemannian Flows. p. 18/25

44 3. Examples on n 1 Boltzmann-Shanon entropy: h(x) = n i=1 x i log(x i ) x i. Shahshahani metric: H(x) = 2 h(x) = diag(1/x 1,...,1/x n ). Kullback-Liebler divergence: D h (y,x) = n y i log(y i /x i ) + x i y i. i= x x(0)=(1/4,1/4,1/2) 2.5 O=(0,0,0) h(x)=x*log(x) x x x Lotka-Volterra type flow 1.30 Hessian Riemannian Flows. p. 19/25

45 3. Other examples h(x) = 2 n i=1 xi H(x) = 2 h(x) = 1 2 diag(1/x3/2 1,...,1/x 3/2 n ). D h (y,x) = n i=1 ( y i x i )/ x i. ( Flow: dx i = 2x 3/2 f dt i x i n j=1 x 3/2 j n k=1 x3/2 k f x j ). Hessian Riemannian Flows. p. 20/25

46 3. Other examples h(x) = 2 n i=1 xi H(x) = 2 h(x) = 1 2 diag(1/x3/2 1,...,1/x 3/2 n ). D h (y,x) = n i=1 ( y i x i )/ x i. ( Flow: dx i = 2x 3/2 f dt i x i n j=1 x 3/2 j n k=1 x3/2 k f x j ). h(x) = n i=1 log(x i) (h(0) = + so that h is not Bregman). H(x) = 2 h(x) = diag(1/x 2 1,...,1/x 2 n). D h (y,x) = n i=1 log(x i/y i ) + (y i x i )/x i. ( Flow: dx i = x 2 f dt i x i n f j=1 x j ). x 2 j n k=1 x2 k Hessian Riemannian Flows. p. 20/25

47 3. Further developments Rate of convergence. Dual trajectory and its convergence. Geodesic type characterization of trajectories. Connections with completely integrable Hamiltonian systems. Reference: A.-Bolte-Brahic, to appear in SIAM J. on Control Optim. Hessian Riemannian Flows. p. 21/25

48 3. Further developments Rate of convergence. Dual trajectory and its convergence. Geodesic type characterization of trajectories. Connections with completely integrable Hamiltonian systems. Reference: A.-Bolte-Brahic, to appear in SIAM J. on Control Optim. Continuous version of similar results for proximal iterations: Iusem-Monteiro 00. Hessian Riemannian Flows. p. 21/25

49 3. Further developments Rate of convergence. Dual trajectory and its convergence. Geodesic type characterization of trajectories. Connections with completely integrable Hamiltonian systems. Reference: A.-Bolte-Brahic, to appear in SIAM J. on Control Optim. Continuous version of similar results for proximal iterations: Iusem-Monteiro 00. Generalizations of results on the log-metric in linear programming: Bayer-Lagarias 89. Hessian Riemannian Flows. p. 21/25

50 4. Duality when Q = R n ++ (P) min{f(x) x 0, Ax = b} Assume: f is convex and S(P). x 0 R n, x 0 > 0, Ax 0 = b. Dual problem: (D) min{p(λ) λ 0} where p(λ) = sup{ λ,x f(x) Ax = b}. Then S(D) = {λ R n λ 0, λ f(x ) + ImA T, λ,x = 0}, where x is any solution of (P). Hessian Riemannian Flows. p. 22/25

51 4. Dual trajectory Integrating the differential inclusion we obtain where c(t) = 1 t d dt h(x(t)) f(x(t)) + Im AT, t 0 λ(t) c(t) + ImA T, f(x(τ))dτ and λ(t) = 1 t [ h(x0 ) h(x(t))]. If h(x) = n i=1 θ(x i ), then λ i (t) = 1 t [θ (x 0 i) θ (x i (t))]. Hessian Riemannian Flows. p. 23/25

52 4. Dual penalty scheme We have: λ(t) = 1 t [ h(x0 ) h(x(t))]. But h is Legendre h 1 = h, with h (λ) = n i=1 θ (λ i ) being the Fenchel conjugate of h. Hence x(t) = h ( h(x 0 ) tλ(t)), where Take A x = b. Since Ax(t) = b, we have x h ( h(x 0 ) tλ(t)) KerA. Then, λ(t) solves { min x,λ + 1 λ t n i=1 } θ (θ (x 0 i) tλ i ) λ c(t) + ImAT Hessian Riemannian Flows. p. 24/25

53 4. Dual trajectory: convergence Example: θ(s) = s log(s) s θ (s ) = exp(s ), s R. Then min λ { x,λ + 1 t n i=1 } x 0 i exp( tλ i ) λ c(t) + ImAT Convergence: f(x) = c,x c(t) = 1 t f(x(τ))dτ c t 0 by Cominetti-San Martin 96, Auslender et al. 97, Cominetti 00,... convergence to the θ -center of S(D). Otherwise, x(t) bounded f(x(t)) f(x ) for x S(P) c(t) f(x ) convergence by Iusem-Monteiro 00. Hessian Riemannian Flows. p. 25/25

Hessian Riemannian gradient flows in convex programming

Hessian Riemannian gradient flows in convex programming Hessian Riemannian gradient flows in convex programming Felipe Alvarez, Jérôme Bolte, Olivier Brahic. Abstract. Motivated by a class of constrained minimization problems, it is studied the gradient flows

More information

c 2004 Society for Industrial and Applied Mathematics

c 2004 Society for Industrial and Applied Mathematics SIAM J. CONTROL OPTIM. Vol. 43, No. 2, pp. 477 501 c 2004 Society for Industrial and Applied Mathematics HESSIAN RIEMANNIAN GRADIENT FLOWS IN CONVEX PROGRAMMING FELIPE ALVAREZ, JÉRÔME BOLTE, AND OLIVIER

More information

A Unified Approach to Proximal Algorithms using Bregman Distance

A Unified Approach to Proximal Algorithms using Bregman Distance A Unified Approach to Proximal Algorithms using Bregman Distance Yi Zhou a,, Yingbin Liang a, Lixin Shen b a Department of Electrical Engineering and Computer Science, Syracuse University b Department

More information

Steepest descent method on a Riemannian manifold: the convex case

Steepest descent method on a Riemannian manifold: the convex case Steepest descent method on a Riemannian manifold: the convex case Julien Munier Abstract. In this paper we are interested in the asymptotic behavior of the trajectories of the famous steepest descent evolution

More information

Central Paths in Semidefinite Programming, Generalized Proximal Point Method and Cauchy Trajectories in Riemannian Manifolds

Central Paths in Semidefinite Programming, Generalized Proximal Point Method and Cauchy Trajectories in Riemannian Manifolds Central Paths in Semidefinite Programming, Generalized Proximal Point Method and Cauchy Trajectories in Riemannian Manifolds da Cruz Neto, J. X., Ferreira, O. P., Oliveira, P. R. and Silva, R. C. M. February

More information

On well definedness of the Central Path

On well definedness of the Central Path On well definedness of the Central Path L.M.Graña Drummond B. F. Svaiter IMPA-Instituto de Matemática Pura e Aplicada Estrada Dona Castorina 110, Jardim Botânico, Rio de Janeiro-RJ CEP 22460-320 Brasil

More information

A New Look at First Order Methods Lifting the Lipschitz Gradient Continuity Restriction

A New Look at First Order Methods Lifting the Lipschitz Gradient Continuity Restriction A New Look at First Order Methods Lifting the Lipschitz Gradient Continuity Restriction Marc Teboulle School of Mathematical Sciences Tel Aviv University Joint work with H. Bauschke and J. Bolte Optimization

More information

Unconstrained minimization

Unconstrained minimization CSCI5254: Convex Optimization & Its Applications Unconstrained minimization terminology and assumptions gradient descent method steepest descent method Newton s method self-concordant functions 1 Unconstrained

More information

Steepest descent method on a Riemannian manifold: the convex case

Steepest descent method on a Riemannian manifold: the convex case Steepest descent method on a Riemannian manifold: the convex case Julien Munier To cite this version: Julien Munier. Steepest descent method on a Riemannian manifold: the convex case. 2006.

More information

Lecture 3. Optimization Problems and Iterative Algorithms

Lecture 3. Optimization Problems and Iterative Algorithms Lecture 3 Optimization Problems and Iterative Algorithms January 13, 2016 This material was jointly developed with Angelia Nedić at UIUC for IE 598ns Outline Special Functions: Linear, Quadratic, Convex

More information

Accelerated Proximal Gradient Methods for Convex Optimization

Accelerated Proximal Gradient Methods for Convex Optimization Accelerated Proximal Gradient Methods for Convex Optimization Paul Tseng Mathematics, University of Washington Seattle MOPTA, University of Guelph August 18, 2008 ACCELERATED PROXIMAL GRADIENT METHODS

More information

arxiv: v3 [math.oc] 17 Dec 2017 Received: date / Accepted: date

arxiv: v3 [math.oc] 17 Dec 2017 Received: date / Accepted: date Noname manuscript No. (will be inserted by the editor) A Simple Convergence Analysis of Bregman Proximal Gradient Algorithm Yi Zhou Yingbin Liang Lixin Shen arxiv:1503.05601v3 [math.oc] 17 Dec 2017 Received:

More information

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44 Convex Optimization Newton s method ENSAE: Optimisation 1/44 Unconstrained minimization minimize f(x) f convex, twice continuously differentiable (hence dom f open) we assume optimal value p = inf x f(x)

More information

10. Unconstrained minimization

10. Unconstrained minimization Convex Optimization Boyd & Vandenberghe 10. Unconstrained minimization terminology and assumptions gradient descent method steepest descent method Newton s method self-concordant functions implementation

More information

Proximal Point Method for a Class of Bregman Distances on Riemannian Manifolds

Proximal Point Method for a Class of Bregman Distances on Riemannian Manifolds Proximal Point Method for a Class of Bregman Distances on Riemannian Manifolds E. A. Papa Quiroz and P. Roberto Oliveira PESC-COPPE Federal University of Rio de Janeiro Rio de Janeiro, Brazil erik@cos.ufrj.br,

More information

2 Preliminaries. is an application P γ,t0,t 1. : T γt0 M T γt1 M given by P γ,t0,t 1

2 Preliminaries. is an application P γ,t0,t 1. : T γt0 M T γt1 M given by P γ,t0,t 1 804 JX DA C NETO, P SANTOS AND S SOUZA where 2 denotes the Euclidean norm We extend the definition of sufficient descent direction for optimization on Riemannian manifolds and we relax that definition

More information

Convex Functions. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)

Convex Functions. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST) Convex Functions Daniel P. Palomar Hong Kong University of Science and Technology (HKUST) ELEC5470 - Convex Optimization Fall 2017-18, HKUST, Hong Kong Outline of Lecture Definition convex function Examples

More information

Supplement: Universal Self-Concordant Barrier Functions

Supplement: Universal Self-Concordant Barrier Functions IE 8534 1 Supplement: Universal Self-Concordant Barrier Functions IE 8534 2 Recall that a self-concordant barrier function for K is a barrier function satisfying 3 F (x)[h, h, h] 2( 2 F (x)[h, h]) 3/2,

More information

Information geometry of mirror descent

Information geometry of mirror descent Information geometry of mirror descent Geometric Science of Information Anthea Monod Department of Statistical Science Duke University Information Initiative at Duke G. Raskutti (UW Madison) and S. Mukherjee

More information

Half of Final Exam Name: Practice Problems October 28, 2014

Half of Final Exam Name: Practice Problems October 28, 2014 Math 54. Treibergs Half of Final Exam Name: Practice Problems October 28, 24 Half of the final will be over material since the last midterm exam, such as the practice problems given here. The other half

More information

BISTA: a Bregmanian proximal gradient method without the global Lipschitz continuity assumption

BISTA: a Bregmanian proximal gradient method without the global Lipschitz continuity assumption BISTA: a Bregmanian proximal gradient method without the global Lipschitz continuity assumption Daniel Reem (joint work with Simeon Reich and Alvaro De Pierro) Department of Mathematics, The Technion,

More information

Optimization and Optimal Control in Banach Spaces

Optimization and Optimal Control in Banach Spaces Optimization and Optimal Control in Banach Spaces Bernhard Schmitzer October 19, 2017 1 Convex non-smooth optimization with proximal operators Remark 1.1 (Motivation). Convex optimization: easier to solve,

More information

arxiv: v1 [math.fa] 16 Jun 2011

arxiv: v1 [math.fa] 16 Jun 2011 arxiv:1106.3342v1 [math.fa] 16 Jun 2011 Gauge functions for convex cones B. F. Svaiter August 20, 2018 Abstract We analyze a class of sublinear functionals which characterize the interior and the exterior

More information

An Infeasible Interior Proximal Method for Convex Programming Problems with Linear Constraints 1

An Infeasible Interior Proximal Method for Convex Programming Problems with Linear Constraints 1 An Infeasible Interior Proximal Method for Convex Programming Problems with Linear Constraints 1 Nobuo Yamashita 2, Christian Kanzow 3, Tomoyui Morimoto 2, and Masao Fuushima 2 2 Department of Applied

More information

The proximal mapping

The proximal mapping The proximal mapping http://bicmr.pku.edu.cn/~wenzw/opt-2016-fall.html Acknowledgement: this slides is based on Prof. Lieven Vandenberghes lecture notes Outline 2/37 1 closed function 2 Conjugate function

More information

arxiv: v1 [math.oc] 29 Dec 2018

arxiv: v1 [math.oc] 29 Dec 2018 DeGroot-Friedkin Map in Opinion Dynamics is Mirror Descent Abhishek Halder ariv:1812.11293v1 [math.oc] 29 Dec 2018 Abstract We provide a variational interpretation of the DeGroot-Friedkin map in opinion

More information

Variational inequalities for set-valued vector fields on Riemannian manifolds

Variational inequalities for set-valued vector fields on Riemannian manifolds Variational inequalities for set-valued vector fields on Riemannian manifolds Chong LI Department of Mathematics Zhejiang University Joint with Jen-Chih YAO Chong LI (Zhejiang University) VI on RM 1 /

More information

Conditional Gradient (Frank-Wolfe) Method

Conditional Gradient (Frank-Wolfe) Method Conditional Gradient (Frank-Wolfe) Method Lecturer: Aarti Singh Co-instructor: Pradeep Ravikumar Convex Optimization 10-725/36-725 1 Outline Today: Conditional gradient method Convergence analysis Properties

More information

WEAK CONVERGENCE OF RESOLVENTS OF MAXIMAL MONOTONE OPERATORS AND MOSCO CONVERGENCE

WEAK CONVERGENCE OF RESOLVENTS OF MAXIMAL MONOTONE OPERATORS AND MOSCO CONVERGENCE Fixed Point Theory, Volume 6, No. 1, 2005, 59-69 http://www.math.ubbcluj.ro/ nodeacj/sfptcj.htm WEAK CONVERGENCE OF RESOLVENTS OF MAXIMAL MONOTONE OPERATORS AND MOSCO CONVERGENCE YASUNORI KIMURA Department

More information

Lecture 15 Newton Method and Self-Concordance. October 23, 2008

Lecture 15 Newton Method and Self-Concordance. October 23, 2008 Newton Method and Self-Concordance October 23, 2008 Outline Lecture 15 Self-concordance Notion Self-concordant Functions Operations Preserving Self-concordance Properties of Self-concordant Functions Implications

More information

Inertial Game Dynamics

Inertial Game Dynamics ... Inertial Game Dynamics R. Laraki P. Mertikopoulos CNRS LAMSADE laboratory CNRS LIG laboratory ADGO'13 Playa Blanca, October 15, 2013 ... Motivation Main Idea: use second order tools to derive efficient

More information

CSCI : Optimization and Control of Networks. Review on Convex Optimization

CSCI : Optimization and Control of Networks. Review on Convex Optimization CSCI7000-016: Optimization and Control of Networks Review on Convex Optimization 1 Convex set S R n is convex if x,y S, λ,µ 0, λ+µ = 1 λx+µy S geometrically: x,y S line segment through x,y S examples (one

More information

Newton s Method. Javier Peña Convex Optimization /36-725

Newton s Method. Javier Peña Convex Optimization /36-725 Newton s Method Javier Peña Convex Optimization 10-725/36-725 1 Last time: dual correspondences Given a function f : R n R, we define its conjugate f : R n R, f ( (y) = max y T x f(x) ) x Properties and

More information

Numerical Algorithms as Dynamical Systems

Numerical Algorithms as Dynamical Systems A Study on Numerical Algorithms as Dynamical Systems Moody Chu North Carolina State University What This Study Is About? To recast many numerical algorithms as special dynamical systems, whence to derive

More information

A UNIFIED CHARACTERIZATION OF PROXIMAL ALGORITHMS VIA THE CONJUGATE OF REGULARIZATION TERM

A UNIFIED CHARACTERIZATION OF PROXIMAL ALGORITHMS VIA THE CONJUGATE OF REGULARIZATION TERM A UNIFIED CHARACTERIZATION OF PROXIMAL ALGORITHMS VIA THE CONJUGATE OF REGULARIZATION TERM OUHEI HARADA September 19, 2018 Abstract. We consider proximal algorithms with generalized regularization term.

More information

Dual convergence of the proximal point method with Bregman distances for linear programming

Dual convergence of the proximal point method with Bregman distances for linear programming Optimization Methods and Software Vol. 22, No. 2, April 2007, 339 360 Dual convergence of the proximal point method with Bregman distances for linear programming J. X. CRUZ NETO, O. P. FERREIRA, A. N.

More information

Applications of Linear Programming

Applications of Linear Programming Applications of Linear Programming lecturer: András London University of Szeged Institute of Informatics Department of Computational Optimization Lecture 9 Non-linear programming In case of LP, the goal

More information

Convex Optimization Boyd & Vandenberghe. 5. Duality

Convex Optimization Boyd & Vandenberghe. 5. Duality 5. Duality Convex Optimization Boyd & Vandenberghe Lagrange dual problem weak and strong duality geometric interpretation optimality conditions perturbation and sensitivity analysis examples generalized

More information

Lecture: Smoothing.

Lecture: Smoothing. Lecture: Smoothing http://bicmr.pku.edu.cn/~wenzw/opt-2018-fall.html Acknowledgement: this slides is based on Prof. Lieven Vandenberghe s lecture notes Smoothing 2/26 introduction smoothing via conjugate

More information

Interior Point Methods in Mathematical Programming

Interior Point Methods in Mathematical Programming Interior Point Methods in Mathematical Programming Clóvis C. Gonzaga Federal University of Santa Catarina, Brazil Journées en l honneur de Pierre Huard Paris, novembre 2008 01 00 11 00 000 000 000 000

More information

Convex Optimization. Dani Yogatama. School of Computer Science, Carnegie Mellon University, Pittsburgh, PA, USA. February 12, 2014

Convex Optimization. Dani Yogatama. School of Computer Science, Carnegie Mellon University, Pittsburgh, PA, USA. February 12, 2014 Convex Optimization Dani Yogatama School of Computer Science, Carnegie Mellon University, Pittsburgh, PA, USA February 12, 2014 Dani Yogatama (Carnegie Mellon University) Convex Optimization February 12,

More information

Logistic Map, Euler & Runge-Kutta Method and Lotka-Volterra Equations

Logistic Map, Euler & Runge-Kutta Method and Lotka-Volterra Equations Logistic Map, Euler & Runge-Kutta Method and Lotka-Volterra Equations S. Y. Ha and J. Park Department of Mathematical Sciences Seoul National University Sep 23, 2013 Contents 1 Logistic Map 2 Euler and

More information

Unconstrained minimization: assumptions

Unconstrained minimization: assumptions Unconstrained minimization I terminology and assumptions I gradient descent method I steepest descent method I Newton s method I self-concordant functions I implementation IOE 611: Nonlinear Programming,

More information

A Brief Review on Convex Optimization

A Brief Review on Convex Optimization A Brief Review on Convex Optimization 1 Convex set S R n is convex if x,y S, λ,µ 0, λ+µ = 1 λx+µy S geometrically: x,y S line segment through x,y S examples (one convex, two nonconvex sets): A Brief Review

More information

An Accelerated Hybrid Proximal Extragradient Method for Convex Optimization and its Implications to Second-Order Methods

An Accelerated Hybrid Proximal Extragradient Method for Convex Optimization and its Implications to Second-Order Methods An Accelerated Hybrid Proximal Extragradient Method for Convex Optimization and its Implications to Second-Order Methods Renato D.C. Monteiro B. F. Svaiter May 10, 011 Revised: May 4, 01) Abstract This

More information

Higher order derivative

Higher order derivative 2 Î 3 á Higher order derivative Î 1 Å Iterated partial derivative Iterated partial derivative Suppose f has f/ x, f/ y and ( f ) = 2 f x x x 2 ( f ) = 2 f x y x y ( f ) = 2 f y x y x ( f ) = 2 f y y y

More information

Active sets, steepest descent, and smooth approximation of functions

Active sets, steepest descent, and smooth approximation of functions Active sets, steepest descent, and smooth approximation of functions Dmitriy Drusvyatskiy School of ORIE, Cornell University Joint work with Alex D. Ioffe (Technion), Martin Larsson (EPFL), and Adrian

More information

Proximal methods. S. Villa. October 7, 2014

Proximal methods. S. Villa. October 7, 2014 Proximal methods S. Villa October 7, 2014 1 Review of the basics Often machine learning problems require the solution of minimization problems. For instance, the ERM algorithm requires to solve a problem

More information

On the interior of the simplex, we have the Hessian of d(x), Hd(x) is diagonal with ith. µd(w) + w T c. minimize. subject to w T 1 = 1,

On the interior of the simplex, we have the Hessian of d(x), Hd(x) is diagonal with ith. µd(w) + w T c. minimize. subject to w T 1 = 1, Math 30 Winter 05 Solution to Homework 3. Recognizing the convexity of g(x) := x log x, from Jensen s inequality we get d(x) n x + + x n n log x + + x n n where the equality is attained only at x = (/n,...,

More information

ORIE 6326: Convex Optimization. Quasi-Newton Methods

ORIE 6326: Convex Optimization. Quasi-Newton Methods ORIE 6326: Convex Optimization Quasi-Newton Methods Professor Udell Operations Research and Information Engineering Cornell April 10, 2017 Slides on steepest descent and analysis of Newton s method adapted

More information

Steepest descent method with a generalized Armijo search for quasiconvex functions on Riemannian manifolds

Steepest descent method with a generalized Armijo search for quasiconvex functions on Riemannian manifolds J. Math. Anal. Appl. 341 (2008) 467 477 www.elsevier.com/locate/jmaa Steepest descent method with a generalized Armijo search for quasiconvex functions on Riemannian manifolds E.A. Papa Quiroz, E.M. Quispe,

More information

MA102: Multivariable Calculus

MA102: Multivariable Calculus MA102: Multivariable Calculus Rupam Barman and Shreemayee Bora Department of Mathematics IIT Guwahati Differentiability of f : U R n R m Definition: Let U R n be open. Then f : U R n R m is differentiable

More information

More First-Order Optimization Algorithms

More First-Order Optimization Algorithms More First-Order Optimization Algorithms Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A. http://www.stanford.edu/ yyye Chapters 3, 8, 3 The SDM

More information

Interior Point Methods for Mathematical Programming

Interior Point Methods for Mathematical Programming Interior Point Methods for Mathematical Programming Clóvis C. Gonzaga Federal University of Santa Catarina, Florianópolis, Brazil EURO - 2013 Roma Our heroes Cauchy Newton Lagrange Early results Unconstrained

More information

Algorithms for constrained local optimization

Algorithms for constrained local optimization Algorithms for constrained local optimization Fabio Schoen 2008 http://gol.dsi.unifi.it/users/schoen Algorithms for constrained local optimization p. Feasible direction methods Algorithms for constrained

More information

Sequential Unconstrained Minimization: A Survey

Sequential Unconstrained Minimization: A Survey Sequential Unconstrained Minimization: A Survey Charles L. Byrne February 21, 2013 Abstract The problem is to minimize a function f : X (, ], over a non-empty subset C of X, where X is an arbitrary set.

More information

Sequential convex programming,: value function and convergence

Sequential convex programming,: value function and convergence Sequential convex programming,: value function and convergence Edouard Pauwels joint work with Jérôme Bolte Journées MODE Toulouse March 23 2016 1 / 16 Introduction Local search methods for finite dimensional

More information

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 17

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 17 EE/ACM 150 - Applications of Convex Optimization in Signal Processing and Communications Lecture 17 Andre Tkacenko Signal Processing Research Group Jet Propulsion Laboratory May 29, 2012 Andre Tkacenko

More information

Convex Optimization. Problem set 2. Due Monday April 26th

Convex Optimization. Problem set 2. Due Monday April 26th Convex Optimization Problem set 2 Due Monday April 26th 1 Gradient Decent without Line-search In this problem we will consider gradient descent with predetermined step sizes. That is, instead of determining

More information

Auxiliary-Function Methods in Iterative Optimization

Auxiliary-Function Methods in Iterative Optimization Auxiliary-Function Methods in Iterative Optimization Charles L. Byrne April 6, 2015 Abstract Let C X be a nonempty subset of an arbitrary set X and f : X R. The problem is to minimize f over C. In auxiliary-function

More information

Convex Optimization. 9. Unconstrained minimization. Prof. Ying Cui. Department of Electrical Engineering Shanghai Jiao Tong University

Convex Optimization. 9. Unconstrained minimization. Prof. Ying Cui. Department of Electrical Engineering Shanghai Jiao Tong University Convex Optimization 9. Unconstrained minimization Prof. Ying Cui Department of Electrical Engineering Shanghai Jiao Tong University 2017 Autumn Semester SJTU Ying Cui 1 / 40 Outline Unconstrained minimization

More information

The general programming problem is the nonlinear programming problem where a given function is maximized subject to a set of inequality constraints.

The general programming problem is the nonlinear programming problem where a given function is maximized subject to a set of inequality constraints. 1 Optimization Mathematical programming refers to the basic mathematical problem of finding a maximum to a function, f, subject to some constraints. 1 In other words, the objective is to find a point,

More information

A proximal-like algorithm for a class of nonconvex programming

A proximal-like algorithm for a class of nonconvex programming Pacific Journal of Optimization, vol. 4, pp. 319-333, 2008 A proximal-like algorithm for a class of nonconvex programming Jein-Shan Chen 1 Department of Mathematics National Taiwan Normal University Taipei,

More information

Optimization Methods. Lecture 19: Line Searches and Newton s Method

Optimization Methods. Lecture 19: Line Searches and Newton s Method 15.93 Optimization Methods Lecture 19: Line Searches and Newton s Method 1 Last Lecture Necessary Conditions for Optimality (identifies candidates) x local min f(x ) =, f(x ) PSD Slide 1 Sufficient Conditions

More information

3. Convex functions. basic properties and examples. operations that preserve convexity. the conjugate function. quasiconvex functions

3. Convex functions. basic properties and examples. operations that preserve convexity. the conjugate function. quasiconvex functions 3. Convex functions Convex Optimization Boyd & Vandenberghe basic properties and examples operations that preserve convexity the conjugate function quasiconvex functions log-concave and log-convex functions

More information

Constrained Optimization and Lagrangian Duality

Constrained Optimization and Lagrangian Duality CIS 520: Machine Learning Oct 02, 2017 Constrained Optimization and Lagrangian Duality Lecturer: Shivani Agarwal Disclaimer: These notes are designed to be a supplement to the lecture. They may or may

More information

3. Convex functions. basic properties and examples. operations that preserve convexity. the conjugate function. quasiconvex functions

3. Convex functions. basic properties and examples. operations that preserve convexity. the conjugate function. quasiconvex functions 3. Convex functions Convex Optimization Boyd & Vandenberghe basic properties and examples operations that preserve convexity the conjugate function quasiconvex functions log-concave and log-convex functions

More information

Agenda. Fast proximal gradient methods. 1 Accelerated first-order methods. 2 Auxiliary sequences. 3 Convergence analysis. 4 Numerical examples

Agenda. Fast proximal gradient methods. 1 Accelerated first-order methods. 2 Auxiliary sequences. 3 Convergence analysis. 4 Numerical examples Agenda Fast proximal gradient methods 1 Accelerated first-order methods 2 Auxiliary sequences 3 Convergence analysis 4 Numerical examples 5 Optimality of Nesterov s scheme Last time Proximal gradient method

More information

Optimization. Escuela de Ingeniería Informática de Oviedo. (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30

Optimization. Escuela de Ingeniería Informática de Oviedo. (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30 Optimization Escuela de Ingeniería Informática de Oviedo (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30 Unconstrained optimization Outline 1 Unconstrained optimization 2 Constrained

More information

10 Numerical methods for constrained problems

10 Numerical methods for constrained problems 10 Numerical methods for constrained problems min s.t. f(x) h(x) = 0 (l), g(x) 0 (m), x X The algorithms can be roughly divided the following way: ˆ primal methods: find descent direction keeping inside

More information

12. Interior-point methods

12. Interior-point methods 12. Interior-point methods Convex Optimization Boyd & Vandenberghe inequality constrained minimization logarithmic barrier function and central path barrier method feasibility and phase I methods complexity

More information

Asymptotic Convergence of the Steepest Descent Method for the Exponential Penalty in Linear Programming

Asymptotic Convergence of the Steepest Descent Method for the Exponential Penalty in Linear Programming Journal of Convex Analysis Volume 2 (1995), No.1/2, 145 152 Asymptotic Convergence of the Steepest Descent Method for the Exponential Penalty in Linear Programming R. Cominetti 1 Universidad de Chile,

More information

Lecture Notes on Iterative Optimization Algorithms

Lecture Notes on Iterative Optimization Algorithms Charles L. Byrne Department of Mathematical Sciences University of Massachusetts Lowell December 8, 2014 Lecture Notes on Iterative Optimization Algorithms Contents Preface vii 1 Overview and Examples

More information

On the Convergence of the Entropy-Exponential Penalty Trajectories and Generalized Proximal Point Methods in Semidefinite Optimization

On the Convergence of the Entropy-Exponential Penalty Trajectories and Generalized Proximal Point Methods in Semidefinite Optimization On the Convergence of the Entropy-Exponential Penalty Trajectories and Generalized Proximal Point Methods in Semidefinite Optimization Ferreira, O. P. Oliveira, P. R. Silva, R. C. M. June 02, 2007 Abstract

More information

5. Duality. Lagrangian

5. Duality. Lagrangian 5. Duality Convex Optimization Boyd & Vandenberghe Lagrange dual problem weak and strong duality geometric interpretation optimality conditions perturbation and sensitivity analysis examples generalized

More information

HW1 solutions. 1. α Ef(x) β, where Ef(x) is the expected value of f(x), i.e., Ef(x) = n. i=1 p if(a i ). (The function f : R R is given.

HW1 solutions. 1. α Ef(x) β, where Ef(x) is the expected value of f(x), i.e., Ef(x) = n. i=1 p if(a i ). (The function f : R R is given. HW1 solutions Exercise 1 (Some sets of probability distributions.) Let x be a real-valued random variable with Prob(x = a i ) = p i, i = 1,..., n, where a 1 < a 2 < < a n. Of course p R n lies in the standard

More information

SOR- and Jacobi-type Iterative Methods for Solving l 1 -l 2 Problems by Way of Fenchel Duality 1

SOR- and Jacobi-type Iterative Methods for Solving l 1 -l 2 Problems by Way of Fenchel Duality 1 SOR- and Jacobi-type Iterative Methods for Solving l 1 -l 2 Problems by Way of Fenchel Duality 1 Masao Fukushima 2 July 17 2010; revised February 4 2011 Abstract We present an SOR-type algorithm and a

More information

EE 546, Univ of Washington, Spring Proximal mapping. introduction. review of conjugate functions. proximal mapping. Proximal mapping 6 1

EE 546, Univ of Washington, Spring Proximal mapping. introduction. review of conjugate functions. proximal mapping. Proximal mapping 6 1 EE 546, Univ of Washington, Spring 2012 6. Proximal mapping introduction review of conjugate functions proximal mapping Proximal mapping 6 1 Proximal mapping the proximal mapping (prox-operator) of a convex

More information

Lecture 5. The Dual Cone and Dual Problem

Lecture 5. The Dual Cone and Dual Problem IE 8534 1 Lecture 5. The Dual Cone and Dual Problem IE 8534 2 For a convex cone K, its dual cone is defined as K = {y x, y 0, x K}. The inner-product can be replaced by x T y if the coordinates of the

More information

Math 273a: Optimization Convex Conjugacy

Math 273a: Optimization Convex Conjugacy Math 273a: Optimization Convex Conjugacy Instructor: Wotao Yin Department of Mathematics, UCLA Fall 2015 online discussions on piazza.com Convex conjugate (the Legendre transform) Let f be a closed proper

More information

Newton s Method. Ryan Tibshirani Convex Optimization /36-725

Newton s Method. Ryan Tibshirani Convex Optimization /36-725 Newton s Method Ryan Tibshirani Convex Optimization 10-725/36-725 1 Last time: dual correspondences Given a function f : R n R, we define its conjugate f : R n R, Properties and examples: f (y) = max x

More information

6. Proximal gradient method

6. Proximal gradient method L. Vandenberghe EE236C (Spring 2016) 6. Proximal gradient method motivation proximal mapping proximal gradient method with fixed step size proximal gradient method with line search 6-1 Proximal mapping

More information

On the Convergence and O(1/N) Complexity of a Class of Nonlinear Proximal Point Algorithms for Monotonic Variational Inequalities

On the Convergence and O(1/N) Complexity of a Class of Nonlinear Proximal Point Algorithms for Monotonic Variational Inequalities STATISTICS,OPTIMIZATION AND INFORMATION COMPUTING Stat., Optim. Inf. Comput., Vol. 2, June 204, pp 05 3. Published online in International Academic Press (www.iapress.org) On the Convergence and O(/N)

More information

Regularity of solutions to Hamilton-Jacobi equations for Tonelli Hamiltonians

Regularity of solutions to Hamilton-Jacobi equations for Tonelli Hamiltonians Regularity of solutions to Hamilton-Jacobi equations for Tonelli Hamiltonians Université Nice Sophia Antipolis & Institut Universitaire de France Nonlinear Analysis and Optimization Royal Society, London,

More information

LEARNING IN CONCAVE GAMES

LEARNING IN CONCAVE GAMES LEARNING IN CONCAVE GAMES P. Mertikopoulos French National Center for Scientific Research (CNRS) Laboratoire d Informatique de Grenoble GSBE ETBC seminar Maastricht, October 22, 2015 Motivation and Preliminaries

More information

Lecture 1: Introduction. Outline. B9824 Foundations of Optimization. Fall Administrative matters. 2. Introduction. 3. Existence of optima

Lecture 1: Introduction. Outline. B9824 Foundations of Optimization. Fall Administrative matters. 2. Introduction. 3. Existence of optima B9824 Foundations of Optimization Lecture 1: Introduction Fall 2009 Copyright 2009 Ciamac Moallemi Outline 1. Administrative matters 2. Introduction 3. Existence of optima 4. Local theory of unconstrained

More information

Dually Flat Geometries in the State Space of Statistical Models

Dually Flat Geometries in the State Space of Statistical Models 1/ 12 Dually Flat Geometries in the State Space of Statistical Models Jan Naudts Universiteit Antwerpen ECEA, November 2016 J. Naudts, Dually Flat Geometries in the State Space of Statistical Models. In

More information

DUALIZATION OF SUBGRADIENT CONDITIONS FOR OPTIMALITY

DUALIZATION OF SUBGRADIENT CONDITIONS FOR OPTIMALITY DUALIZATION OF SUBGRADIENT CONDITIONS FOR OPTIMALITY R. T. Rockafellar* Abstract. A basic relationship is derived between generalized subgradients of a given function, possibly nonsmooth and nonconvex,

More information

LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE

LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE CONVEX ANALYSIS AND DUALITY Basic concepts of convex analysis Basic concepts of convex optimization Geometric duality framework - MC/MC Constrained optimization

More information

ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications

ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications Professor M. Chiang Electrical Engineering Department, Princeton University March

More information

Local strong convexity and local Lipschitz continuity of the gradient of convex functions

Local strong convexity and local Lipschitz continuity of the gradient of convex functions Local strong convexity and local Lipschitz continuity of the gradient of convex functions R. Goebel and R.T. Rockafellar May 23, 2007 Abstract. Given a pair of convex conjugate functions f and f, we investigate

More information

Fast proximal gradient methods

Fast proximal gradient methods L. Vandenberghe EE236C (Spring 2013-14) Fast proximal gradient methods fast proximal gradient method (FISTA) FISTA with line search FISTA as descent method Nesterov s second method 1 Fast (proximal) gradient

More information

Quasi-invariant Measures on Path Space. Denis Bell University of North Florida

Quasi-invariant Measures on Path Space. Denis Bell University of North Florida Quasi-invariant Measures on Path Space Denis Bell University of North Florida Transformation of measure under the flow of a vector field Let E be a vector space (or a manifold), equipped with a finite

More information

Key words. saddle-point dynamics, asymptotic convergence, convex-concave functions, proximal calculus, center manifold theory, nonsmooth dynamics

Key words. saddle-point dynamics, asymptotic convergence, convex-concave functions, proximal calculus, center manifold theory, nonsmooth dynamics SADDLE-POINT DYNAMICS: CONDITIONS FOR ASYMPTOTIC STABILITY OF SADDLE POINTS ASHISH CHERUKURI, BAHMAN GHARESIFARD, AND JORGE CORTÉS Abstract. This paper considers continuously differentiable functions of

More information

Nonsmooth optimization: conditioning, convergence, and semi-algebraic models

Nonsmooth optimization: conditioning, convergence, and semi-algebraic models Nonsmooth optimization: conditioning, convergence, and semi-algebraic models Adrian Lewis ORIE Cornell International Congress of Mathematicians Seoul, August 2014 1/16 Outline I Optimization and inverse

More information

minimize x subject to (x 2)(x 4) u,

minimize x subject to (x 2)(x 4) u, Math 6366/6367: Optimization and Variational Methods Sample Preliminary Exam Questions 1. Suppose that f : [, L] R is a C 2 -function with f () on (, L) and that you have explicit formulae for

More information

Introduction to Information Geometry

Introduction to Information Geometry Introduction to Information Geometry based on the book Methods of Information Geometry written by Shun-Ichi Amari and Hiroshi Nagaoka Yunshu Liu 2012-02-17 Outline 1 Introduction to differential geometry

More information

Non-smooth Non-convex Bregman Minimization: Unification and new Algorithms

Non-smooth Non-convex Bregman Minimization: Unification and new Algorithms Non-smooth Non-convex Bregman Minimization: Unification and new Algorithms Peter Ochs, Jalal Fadili, and Thomas Brox Saarland University, Saarbrücken, Germany Normandie Univ, ENSICAEN, CNRS, GREYC, France

More information

Frank-Wolfe Method. Ryan Tibshirani Convex Optimization

Frank-Wolfe Method. Ryan Tibshirani Convex Optimization Frank-Wolfe Method Ryan Tibshirani Convex Optimization 10-725 Last time: ADMM For the problem min x,z f(x) + g(z) subject to Ax + Bz = c we form augmented Lagrangian (scaled form): L ρ (x, z, w) = f(x)

More information

Primal-Dual Symmetric Interior-Point Methods from SDP to Hyperbolic Cone Programming and Beyond

Primal-Dual Symmetric Interior-Point Methods from SDP to Hyperbolic Cone Programming and Beyond Primal-Dual Symmetric Interior-Point Methods from SDP to Hyperbolic Cone Programming and Beyond Tor Myklebust Levent Tunçel September 26, 2014 Convex Optimization in Conic Form (P) inf c, x A(x) = b, x

More information