Hessian Riemannian Gradient Flows in Convex Programming
|
|
- Ambrose Gray
- 5 years ago
- Views:
Transcription
1 Hessian Riemannian Gradient Flows in Convex Programming Felipe Alvarez, Jérôme Bolte, Olivier Brahic INTERNATIONAL CONFERENCE ON MODELING AND OPTIMIZATION MODOPT 2004 Universidad de La Frontera, Temuco, Chile January 19-22, Hessian Riemannian Flows. p. 1/25
2 Outline 1. Motivation: scaling the Euclidean gradient. 2. Riemannian gradient flows on convex sets. 3. Hessian metrics, existence, convergence and examples. Hessian Riemannian Flows. p. 2/25
3 1. Motivation: gradient method Let f : R n R be a smooth function, x 0 R n. x k+1 = x k + λ k d k, Hessian Riemannian Flows. p. 3/25
4 1. Motivation: gradient method Let f : R n R be a smooth function, x 0 R n. x k+1 = x k + λ k d k, where λ k > 0 is a stepsize Hessian Riemannian Flows. p. 3/25
5 1. Motivation: gradient method Let f : R n R be a smooth function, x 0 R n. x k+1 = x k + λ k d k, where λ k > 0 is a stepsize and d k = f(x k ) is the steepest descent direction. Hessian Riemannian Flows. p. 3/25
6 1. Motivation: gradient method Let f : R n R be a smooth function, x 0 R n. x k+1 = x k + λ k d k, where λ k > 0 is a stepsize and d k = f(x k ) is the steepest descent direction. Continuous gradient method: dx dt = f(x), t > 0. Continuous flow discrete method. Hessian Riemannian Flows. p. 3/25
7 1. Scaling and Newton s method Newton s correction: x k+1 = x k λ k 2 f(x k ) 1 f(x k ). Continuous Newton s method: dx dt = 2 f(x) 1 f(x). Hessian Riemannian Flows. p. 4/25
8 1. Scaling and Newton s method Newton s correction: x k+1 = x k λ k 2 f(x k ) 1 f(x k ). Continuous Newton s method: dx dt = 2 f(x) 1 f(x). d f(x) = f(x). dt Hessian Riemannian Flows. p. 4/25
9 1. Scaling and Newton s method Newton s correction: x k+1 = x k λ k 2 f(x k ) 1 f(x k ). Continuous Newton s method: dx dt = 2 f(x) 1 f(x). d f(x) = f(x). dt f(x(t)) = e t f(x 0 ) Scale invariant rate of convergence on a straight line! back. Hessian Riemannian Flows. p. 4/25
10 1. Scaling and constraints Problem: min{f(x) x 0}. Hessian Riemannian Flows. p. 5/25
11 1. Scaling and constraints Problem: min{f(x) x 0}. ODE approach: dx i dt = x f i (x), i = 1,...,n. x i Hessian Riemannian Flows. p. 5/25
12 1. Scaling and constraints Problem: min{f(x) x 0}. ODE approach: Properties: d dt f(x) = n i=1 dx i dt = x f i (x), i = 1,...,n. x i f x i x i (x) 2 0 descent method on R n +. x i (0) > 0 t > 0, x i (t) > 0 interior point trajectory. Hessian Riemannian Flows. p. 5/25
13 1. Scaling and constraints Problem: min{f(x) x 0}. ODE approach: Properties: d dt f(x) = n i=1 dx i dt = x f i (x), i = 1,...,n. x i f x i x i (x) 2 0 descent method on R n +. x i (0) > 0 t > 0, x i (t) > 0 interior point trajectory. The equation may be written as d dt log(x i) = f (x) x i Scaling logarithmic barrier to force x(t) > 0! Hessian Riemannian Flows. p. 5/25
14 1. Hessian type scaling d dt log(x i) = f (x) x i Hessian Riemannian Flows. p. 6/25
15 1. Hessian type scaling where d dt log(x i) = f (x) x i d h(x) = f(x), dt h(x) = n x i log(x i ) x i i=1 Hessian Riemannian Flows. p. 6/25
16 1. Hessian type scaling where Thus d dt log(x i) = f (x) x i d h(x) = f(x), dt h(x) = n x i log(x i ) x i i=1 dx i dt = x f i (x) dx x i dt = 2 h(x) 1 f(x) Remark the analogy with Newton s method Hessian Riemannian Flows. p. 6/25
17 2. Riemannian gradient flows Problem: min f(x), with C = {x x C Rn x Q, Ax = b}. Q R n nonempty, open and convex. A R m n and b R m with m n. Hessian Riemannian Flows. p. 7/25
18 2. Riemannian gradient flows Problem: min f(x), with C = {x x C Rn x Q, Ax = b}. Q R n nonempty, open and convex. A R m n and b R m with m n. Strategy: Introduce a Riemannian metric H(x) S n ++ on Q, (u,v) x = H(x)u,v = Consider the gradient flow n i=1 H ij (x)u i v j, x Q. dx dt (t) = Hf(x(t)),t > 0, Hessian Riemannian Flows. p. 7/25
19 2. Riemannian gradient Let M, (, ) be a Riemannian manifold. T x M: tangent space to M at x M. Hessian Riemannian Flows. p. 8/25
20 2. Riemannian gradient Let M, (, ) be a Riemannian manifold. T x M: tangent space to M at x M. The gradient gradf of f C 1 (M; R) is uniquely determined by Tangency: for all x M, gradf(x) T x M. Duality: for all x M, v T x M, df(x)v = (gradf(x),v) x, where df(x) : T x M R is the differential of f. Hessian Riemannian Flows. p. 8/25
21 2. Riemannian gradient in our case M = Q {x R n Ax = b} with Q open set, then T x M Ker A. (, ) x = H(x), with a barrier/penalty effect near Q. Hessian Riemannian Flows. p. 9/25
22 2. Riemannian gradient in our case M = Q {x R n Ax = b} with Q open set, then T x M Ker A. (, ) x = H(x), with a barrier/penalty effect near Q. H f = H 1 [I A T (AH 1 A T ) 1 AH 1 ] f, where H stands for grad to stress the dependence on H. Hessian Riemannian Flows. p. 9/25
23 2. Riemannian gradient in our case M = Q {x R n Ax = b} with Q open set, then T x M Ker A. (, ) x = H(x), with a barrier/penalty effect near Q. H f = H 1 [I A T (AH 1 A T ) 1 AH 1 ] f, where H stands for grad to stress the dependence on H. Projection vector field in the tangent space. Scaling interior point method x(t) Q. Hessian Riemannian Flows. p. 9/25
24 2. Example C = n 1 := {x R n x 0, n i=1 x i = 1} Hessian Riemannian Flows. p. 10/25
25 2. Example C = n 1 := {x R n x 0, n i=1 x i = 1} Q = R n ++, A = (1,...,1) R 1 n and b = 1. M = {x R n x > 0, n i=1 x i = 1}. T x M = {v R n n i=1 v i = 0}. Hessian Riemannian Flows. p. 10/25
26 2. Example C = n 1 := {x R n x 0, n i=1 x i = 1} Q = R n ++, A = (1,...,1) R 1 n and b = 1. M = {x R n x > 0, n i=1 x i = 1}. T x M = {v R n n i=1 v i = 0}. Take H(x) = diag(1/x 1,...,1/x n ), then (u,v) x = n i=1 u i v j x i Shahshahani metric. dx i dt = x f i + x i n j=1 x i x j f x j Lotka-Volterra type eq. Karmarkar 90, Faybusovich 91,... Hessian Riemannian Flows. p. 10/25
27 2. Barrier effect: Legendre functions We focus on the case H(x) = 2 h(x), x Q with h : R n R {+ } is closed, convex and proper. int dom h = Q. (H 0 ) h is of Legendre type. h Q C 2 (Q; R) and 2 h(x) > 0. 2 h is locally Lipschitz. Hessian Riemannian Flows. p. 11/25
28 2. Barrier effect: Legendre functions We focus on the case H(x) = 2 h(x), x Q with h : R n R {+ } is closed, convex and proper. int dom h = Q. (H 0 ) h is of Legendre type. h Q C 2 (Q; R) and 2 h(x) > 0. 2 h is locally Lipschitz. h is strictly convex and C 1 on int dom h. int dom h x j x int dom h, h(x j ) +. back Hessian Riemannian Flows. p. 11/25
29 2. Example of Legendre function h(x) = i I θ(g i (x)). where I = {1,...,p}, g i C 3 (R n ) concave. Q = {x R n g i (x) > 0,i I}. x Q, span { g i (x) i I} = R n, and (i) (0, ) domθ [0, ), θ C 3 (0, ). (ii) lim s 0 + θ (s) = and s > 0, θ (s) > 0. (iii) Either θ is nonincreasing or i I,g i is affine. Hessian Riemannian Flows. p. 12/25
30 3. Questions Given x 0 M = Q {x R n Ax = b}: dx dt (t) = Hf(x(t)),t > 0, Hessian Riemannian Flows. p. 13/25
31 3. Questions Given x 0 M = Q {x R n Ax = b}: dx dt (t) = Hf(x(t)),t > 0, Well posedness: global existence for all t > 0. Asymptotic behavior: convergence to an equilibrium as t, rate of convergence,... Hessian Riemannian Flows. p. 13/25
32 3. Questions Given x 0 M = Q {x R n Ax = b}: dx dt (t) = Hf(x(t)),t > 0, Well posedness: global existence for all t > 0. Asymptotic behavior: convergence to an equilibrium as t, rate of convergence,... Main difficulty: singular behavior near Q. Classical results do not apply. Hessian Riemannian Flows. p. 13/25
33 3. Well posedness: global existence Thm. 2 The trajectory x(t) is defined for all t 0 under any of the following conditions: (C 1 ) {x C f(x) f(x 0 )} is bounded. (i) dom h = Q (C 2 ) (ii) y Q, γ R, {x C D h (y,x) γ} is bdd. (iii) Argmin C f and f quasiconvex. (C 3 ) K 0, L R such that x Q, H(x) 1 K x + L. Hessian Riemannian Flows. p. 14/25
34 3. Why Hessian metrics? Suppose f is convex and A = 0 and b = 0. Euclidean case: y Argmin C f x C, f(x),x y 0. Hessian Riemannian Flows. p. 15/25
35 3. Why Hessian metrics? Suppose f is convex and A = 0 and b = 0. Euclidean case: y Argmin C f x C, f(x),x y 0. ϕ y (x) = 1 x y 2 2 ẋ = f(x) d dt ϕ y(x) = ϕ y (x),ẋ = x y, f(x) 0. Hessian Riemannian Flows. p. 15/25
36 3. Why Hessian metrics? Suppose f is convex and A = 0 and b = 0. Euclidean case: y Argmin C f x C, f(x),x y 0. ϕ y (x) = 1 x y 2 2 ẋ = f(x) d dt ϕ y(x) = ϕ y (x),ẋ = x y, f(x) 0. ϕ y (x) is a Lyapunov function for the gradient flow Hessian Riemannian Flows. p. 15/25
37 3. Characterization of Hessian metrics Riemannian case: y Argmin Q f x Q, ( H f(x),x y) x 0. Hessian Riemannian Flows. p. 16/25
38 3. Characterization of Hessian metrics Riemannian case: y Argmin Q f x Q, ( H f(x),x y) x 0. Thm. 1 H C 1 (Q; S n ++) satisfies y Q, ϕ y C 1 (Q; R), H ϕ y (x) = x y? Hessian Riemannian Flows. p. 16/25
39 3. Characterization of Hessian metrics Riemannian case: y Argmin Q f x Q, ( H f(x),x y) x 0. Thm. 1 H C 1 (Q; S n ++) satisfies y Q, ϕ y C 1 (Q; R), H ϕ y (x) = x y h C 3 (Q) such that H = 2 h on Q and ϕ y (x) = D h (y,x) = h(y) h(x) h(x),y x = Bregman pseudo-distance induced by h. Hessian Riemannian Flows. p. 16/25
40 3. Implicit proximal iteration x k+1 Argmin { } f(x) + 1 λ k D h (x,x k ) Ax = b, 1 λ k [ h(x k+1 ) h(x k )] f(x k+1 ) + Im A T, Ax k+1 = b Bregman 67, Censor-Zenios 92, Teboulle 92, Eckstein 93, Kiwiel 97,... Hessian Riemannian Flows. p. 17/25
41 3. Implicit proximal iteration x k+1 Argmin { } f(x) + 1 λ k D h (x,x k ) Ax = b, 1 λ k [ h(x k+1 ) h(x k )] f(x k+1 ) + Im A T, Ax k+1 = b Bregman 67, Censor-Zenios 92, Teboulle 92, Eckstein 93, Kiwiel 97,... But dx dt = Hf(x) d h(x) f(x) + Im dt AT, Ax(t) = b, t 0. This link was already noticed by Iusem-Svaiter-Da Cruz Neto 99, together with convergence results for a linear objective function. Hessian Riemannian Flows. p. 17/25
42 3. Convergence: Bregman functions A Legendre function h with domh = Q is of Bregman type if (i) {x Q D h (y,x) γ} is bdd. y Q, γ R. (ii) y Q, y j y with y j Q, D h (y,y j ) 0. Hessian Riemannian Flows. p. 18/25
43 3. Convergence: Bregman functions A Legendre function h with domh = Q is of Bregman type if (i) {x Q D h (y,x) γ} is bdd. y Q, γ R. (ii) y Q, y j y with y j Q, D h (y,y j ) 0. Thm. 3 Suppose (H 0 ) with h of Bregman type. f is quasiconvex and Argmin C f. Then x C such that x(t) x as t + with f(x ) N Q (x ) + Ker A, where N Q (x ) is the normal cone to Q at x. Hessian Riemannian Flows. p. 18/25
44 3. Examples on n 1 Boltzmann-Shanon entropy: h(x) = n i=1 x i log(x i ) x i. Shahshahani metric: H(x) = 2 h(x) = diag(1/x 1,...,1/x n ). Kullback-Liebler divergence: D h (y,x) = n y i log(y i /x i ) + x i y i. i= x x(0)=(1/4,1/4,1/2) 2.5 O=(0,0,0) h(x)=x*log(x) x x x Lotka-Volterra type flow 1.30 Hessian Riemannian Flows. p. 19/25
45 3. Other examples h(x) = 2 n i=1 xi H(x) = 2 h(x) = 1 2 diag(1/x3/2 1,...,1/x 3/2 n ). D h (y,x) = n i=1 ( y i x i )/ x i. ( Flow: dx i = 2x 3/2 f dt i x i n j=1 x 3/2 j n k=1 x3/2 k f x j ). Hessian Riemannian Flows. p. 20/25
46 3. Other examples h(x) = 2 n i=1 xi H(x) = 2 h(x) = 1 2 diag(1/x3/2 1,...,1/x 3/2 n ). D h (y,x) = n i=1 ( y i x i )/ x i. ( Flow: dx i = 2x 3/2 f dt i x i n j=1 x 3/2 j n k=1 x3/2 k f x j ). h(x) = n i=1 log(x i) (h(0) = + so that h is not Bregman). H(x) = 2 h(x) = diag(1/x 2 1,...,1/x 2 n). D h (y,x) = n i=1 log(x i/y i ) + (y i x i )/x i. ( Flow: dx i = x 2 f dt i x i n f j=1 x j ). x 2 j n k=1 x2 k Hessian Riemannian Flows. p. 20/25
47 3. Further developments Rate of convergence. Dual trajectory and its convergence. Geodesic type characterization of trajectories. Connections with completely integrable Hamiltonian systems. Reference: A.-Bolte-Brahic, to appear in SIAM J. on Control Optim. Hessian Riemannian Flows. p. 21/25
48 3. Further developments Rate of convergence. Dual trajectory and its convergence. Geodesic type characterization of trajectories. Connections with completely integrable Hamiltonian systems. Reference: A.-Bolte-Brahic, to appear in SIAM J. on Control Optim. Continuous version of similar results for proximal iterations: Iusem-Monteiro 00. Hessian Riemannian Flows. p. 21/25
49 3. Further developments Rate of convergence. Dual trajectory and its convergence. Geodesic type characterization of trajectories. Connections with completely integrable Hamiltonian systems. Reference: A.-Bolte-Brahic, to appear in SIAM J. on Control Optim. Continuous version of similar results for proximal iterations: Iusem-Monteiro 00. Generalizations of results on the log-metric in linear programming: Bayer-Lagarias 89. Hessian Riemannian Flows. p. 21/25
50 4. Duality when Q = R n ++ (P) min{f(x) x 0, Ax = b} Assume: f is convex and S(P). x 0 R n, x 0 > 0, Ax 0 = b. Dual problem: (D) min{p(λ) λ 0} where p(λ) = sup{ λ,x f(x) Ax = b}. Then S(D) = {λ R n λ 0, λ f(x ) + ImA T, λ,x = 0}, where x is any solution of (P). Hessian Riemannian Flows. p. 22/25
51 4. Dual trajectory Integrating the differential inclusion we obtain where c(t) = 1 t d dt h(x(t)) f(x(t)) + Im AT, t 0 λ(t) c(t) + ImA T, f(x(τ))dτ and λ(t) = 1 t [ h(x0 ) h(x(t))]. If h(x) = n i=1 θ(x i ), then λ i (t) = 1 t [θ (x 0 i) θ (x i (t))]. Hessian Riemannian Flows. p. 23/25
52 4. Dual penalty scheme We have: λ(t) = 1 t [ h(x0 ) h(x(t))]. But h is Legendre h 1 = h, with h (λ) = n i=1 θ (λ i ) being the Fenchel conjugate of h. Hence x(t) = h ( h(x 0 ) tλ(t)), where Take A x = b. Since Ax(t) = b, we have x h ( h(x 0 ) tλ(t)) KerA. Then, λ(t) solves { min x,λ + 1 λ t n i=1 } θ (θ (x 0 i) tλ i ) λ c(t) + ImAT Hessian Riemannian Flows. p. 24/25
53 4. Dual trajectory: convergence Example: θ(s) = s log(s) s θ (s ) = exp(s ), s R. Then min λ { x,λ + 1 t n i=1 } x 0 i exp( tλ i ) λ c(t) + ImAT Convergence: f(x) = c,x c(t) = 1 t f(x(τ))dτ c t 0 by Cominetti-San Martin 96, Auslender et al. 97, Cominetti 00,... convergence to the θ -center of S(D). Otherwise, x(t) bounded f(x(t)) f(x ) for x S(P) c(t) f(x ) convergence by Iusem-Monteiro 00. Hessian Riemannian Flows. p. 25/25
Hessian Riemannian gradient flows in convex programming
Hessian Riemannian gradient flows in convex programming Felipe Alvarez, Jérôme Bolte, Olivier Brahic. Abstract. Motivated by a class of constrained minimization problems, it is studied the gradient flows
More informationc 2004 Society for Industrial and Applied Mathematics
SIAM J. CONTROL OPTIM. Vol. 43, No. 2, pp. 477 501 c 2004 Society for Industrial and Applied Mathematics HESSIAN RIEMANNIAN GRADIENT FLOWS IN CONVEX PROGRAMMING FELIPE ALVAREZ, JÉRÔME BOLTE, AND OLIVIER
More informationA Unified Approach to Proximal Algorithms using Bregman Distance
A Unified Approach to Proximal Algorithms using Bregman Distance Yi Zhou a,, Yingbin Liang a, Lixin Shen b a Department of Electrical Engineering and Computer Science, Syracuse University b Department
More informationSteepest descent method on a Riemannian manifold: the convex case
Steepest descent method on a Riemannian manifold: the convex case Julien Munier Abstract. In this paper we are interested in the asymptotic behavior of the trajectories of the famous steepest descent evolution
More informationCentral Paths in Semidefinite Programming, Generalized Proximal Point Method and Cauchy Trajectories in Riemannian Manifolds
Central Paths in Semidefinite Programming, Generalized Proximal Point Method and Cauchy Trajectories in Riemannian Manifolds da Cruz Neto, J. X., Ferreira, O. P., Oliveira, P. R. and Silva, R. C. M. February
More informationOn well definedness of the Central Path
On well definedness of the Central Path L.M.Graña Drummond B. F. Svaiter IMPA-Instituto de Matemática Pura e Aplicada Estrada Dona Castorina 110, Jardim Botânico, Rio de Janeiro-RJ CEP 22460-320 Brasil
More informationA New Look at First Order Methods Lifting the Lipschitz Gradient Continuity Restriction
A New Look at First Order Methods Lifting the Lipschitz Gradient Continuity Restriction Marc Teboulle School of Mathematical Sciences Tel Aviv University Joint work with H. Bauschke and J. Bolte Optimization
More informationUnconstrained minimization
CSCI5254: Convex Optimization & Its Applications Unconstrained minimization terminology and assumptions gradient descent method steepest descent method Newton s method self-concordant functions 1 Unconstrained
More informationSteepest descent method on a Riemannian manifold: the convex case
Steepest descent method on a Riemannian manifold: the convex case Julien Munier To cite this version: Julien Munier. Steepest descent method on a Riemannian manifold: the convex case. 2006.
More informationLecture 3. Optimization Problems and Iterative Algorithms
Lecture 3 Optimization Problems and Iterative Algorithms January 13, 2016 This material was jointly developed with Angelia Nedić at UIUC for IE 598ns Outline Special Functions: Linear, Quadratic, Convex
More informationAccelerated Proximal Gradient Methods for Convex Optimization
Accelerated Proximal Gradient Methods for Convex Optimization Paul Tseng Mathematics, University of Washington Seattle MOPTA, University of Guelph August 18, 2008 ACCELERATED PROXIMAL GRADIENT METHODS
More informationarxiv: v3 [math.oc] 17 Dec 2017 Received: date / Accepted: date
Noname manuscript No. (will be inserted by the editor) A Simple Convergence Analysis of Bregman Proximal Gradient Algorithm Yi Zhou Yingbin Liang Lixin Shen arxiv:1503.05601v3 [math.oc] 17 Dec 2017 Received:
More informationConvex Optimization. Newton s method. ENSAE: Optimisation 1/44
Convex Optimization Newton s method ENSAE: Optimisation 1/44 Unconstrained minimization minimize f(x) f convex, twice continuously differentiable (hence dom f open) we assume optimal value p = inf x f(x)
More information10. Unconstrained minimization
Convex Optimization Boyd & Vandenberghe 10. Unconstrained minimization terminology and assumptions gradient descent method steepest descent method Newton s method self-concordant functions implementation
More informationProximal Point Method for a Class of Bregman Distances on Riemannian Manifolds
Proximal Point Method for a Class of Bregman Distances on Riemannian Manifolds E. A. Papa Quiroz and P. Roberto Oliveira PESC-COPPE Federal University of Rio de Janeiro Rio de Janeiro, Brazil erik@cos.ufrj.br,
More information2 Preliminaries. is an application P γ,t0,t 1. : T γt0 M T γt1 M given by P γ,t0,t 1
804 JX DA C NETO, P SANTOS AND S SOUZA where 2 denotes the Euclidean norm We extend the definition of sufficient descent direction for optimization on Riemannian manifolds and we relax that definition
More informationConvex Functions. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)
Convex Functions Daniel P. Palomar Hong Kong University of Science and Technology (HKUST) ELEC5470 - Convex Optimization Fall 2017-18, HKUST, Hong Kong Outline of Lecture Definition convex function Examples
More informationSupplement: Universal Self-Concordant Barrier Functions
IE 8534 1 Supplement: Universal Self-Concordant Barrier Functions IE 8534 2 Recall that a self-concordant barrier function for K is a barrier function satisfying 3 F (x)[h, h, h] 2( 2 F (x)[h, h]) 3/2,
More informationInformation geometry of mirror descent
Information geometry of mirror descent Geometric Science of Information Anthea Monod Department of Statistical Science Duke University Information Initiative at Duke G. Raskutti (UW Madison) and S. Mukherjee
More informationHalf of Final Exam Name: Practice Problems October 28, 2014
Math 54. Treibergs Half of Final Exam Name: Practice Problems October 28, 24 Half of the final will be over material since the last midterm exam, such as the practice problems given here. The other half
More informationBISTA: a Bregmanian proximal gradient method without the global Lipschitz continuity assumption
BISTA: a Bregmanian proximal gradient method without the global Lipschitz continuity assumption Daniel Reem (joint work with Simeon Reich and Alvaro De Pierro) Department of Mathematics, The Technion,
More informationOptimization and Optimal Control in Banach Spaces
Optimization and Optimal Control in Banach Spaces Bernhard Schmitzer October 19, 2017 1 Convex non-smooth optimization with proximal operators Remark 1.1 (Motivation). Convex optimization: easier to solve,
More informationarxiv: v1 [math.fa] 16 Jun 2011
arxiv:1106.3342v1 [math.fa] 16 Jun 2011 Gauge functions for convex cones B. F. Svaiter August 20, 2018 Abstract We analyze a class of sublinear functionals which characterize the interior and the exterior
More informationAn Infeasible Interior Proximal Method for Convex Programming Problems with Linear Constraints 1
An Infeasible Interior Proximal Method for Convex Programming Problems with Linear Constraints 1 Nobuo Yamashita 2, Christian Kanzow 3, Tomoyui Morimoto 2, and Masao Fuushima 2 2 Department of Applied
More informationThe proximal mapping
The proximal mapping http://bicmr.pku.edu.cn/~wenzw/opt-2016-fall.html Acknowledgement: this slides is based on Prof. Lieven Vandenberghes lecture notes Outline 2/37 1 closed function 2 Conjugate function
More informationarxiv: v1 [math.oc] 29 Dec 2018
DeGroot-Friedkin Map in Opinion Dynamics is Mirror Descent Abhishek Halder ariv:1812.11293v1 [math.oc] 29 Dec 2018 Abstract We provide a variational interpretation of the DeGroot-Friedkin map in opinion
More informationVariational inequalities for set-valued vector fields on Riemannian manifolds
Variational inequalities for set-valued vector fields on Riemannian manifolds Chong LI Department of Mathematics Zhejiang University Joint with Jen-Chih YAO Chong LI (Zhejiang University) VI on RM 1 /
More informationConditional Gradient (Frank-Wolfe) Method
Conditional Gradient (Frank-Wolfe) Method Lecturer: Aarti Singh Co-instructor: Pradeep Ravikumar Convex Optimization 10-725/36-725 1 Outline Today: Conditional gradient method Convergence analysis Properties
More informationWEAK CONVERGENCE OF RESOLVENTS OF MAXIMAL MONOTONE OPERATORS AND MOSCO CONVERGENCE
Fixed Point Theory, Volume 6, No. 1, 2005, 59-69 http://www.math.ubbcluj.ro/ nodeacj/sfptcj.htm WEAK CONVERGENCE OF RESOLVENTS OF MAXIMAL MONOTONE OPERATORS AND MOSCO CONVERGENCE YASUNORI KIMURA Department
More informationLecture 15 Newton Method and Self-Concordance. October 23, 2008
Newton Method and Self-Concordance October 23, 2008 Outline Lecture 15 Self-concordance Notion Self-concordant Functions Operations Preserving Self-concordance Properties of Self-concordant Functions Implications
More informationInertial Game Dynamics
... Inertial Game Dynamics R. Laraki P. Mertikopoulos CNRS LAMSADE laboratory CNRS LIG laboratory ADGO'13 Playa Blanca, October 15, 2013 ... Motivation Main Idea: use second order tools to derive efficient
More informationCSCI : Optimization and Control of Networks. Review on Convex Optimization
CSCI7000-016: Optimization and Control of Networks Review on Convex Optimization 1 Convex set S R n is convex if x,y S, λ,µ 0, λ+µ = 1 λx+µy S geometrically: x,y S line segment through x,y S examples (one
More informationNewton s Method. Javier Peña Convex Optimization /36-725
Newton s Method Javier Peña Convex Optimization 10-725/36-725 1 Last time: dual correspondences Given a function f : R n R, we define its conjugate f : R n R, f ( (y) = max y T x f(x) ) x Properties and
More informationNumerical Algorithms as Dynamical Systems
A Study on Numerical Algorithms as Dynamical Systems Moody Chu North Carolina State University What This Study Is About? To recast many numerical algorithms as special dynamical systems, whence to derive
More informationA UNIFIED CHARACTERIZATION OF PROXIMAL ALGORITHMS VIA THE CONJUGATE OF REGULARIZATION TERM
A UNIFIED CHARACTERIZATION OF PROXIMAL ALGORITHMS VIA THE CONJUGATE OF REGULARIZATION TERM OUHEI HARADA September 19, 2018 Abstract. We consider proximal algorithms with generalized regularization term.
More informationDual convergence of the proximal point method with Bregman distances for linear programming
Optimization Methods and Software Vol. 22, No. 2, April 2007, 339 360 Dual convergence of the proximal point method with Bregman distances for linear programming J. X. CRUZ NETO, O. P. FERREIRA, A. N.
More informationApplications of Linear Programming
Applications of Linear Programming lecturer: András London University of Szeged Institute of Informatics Department of Computational Optimization Lecture 9 Non-linear programming In case of LP, the goal
More informationConvex Optimization Boyd & Vandenberghe. 5. Duality
5. Duality Convex Optimization Boyd & Vandenberghe Lagrange dual problem weak and strong duality geometric interpretation optimality conditions perturbation and sensitivity analysis examples generalized
More informationLecture: Smoothing.
Lecture: Smoothing http://bicmr.pku.edu.cn/~wenzw/opt-2018-fall.html Acknowledgement: this slides is based on Prof. Lieven Vandenberghe s lecture notes Smoothing 2/26 introduction smoothing via conjugate
More informationInterior Point Methods in Mathematical Programming
Interior Point Methods in Mathematical Programming Clóvis C. Gonzaga Federal University of Santa Catarina, Brazil Journées en l honneur de Pierre Huard Paris, novembre 2008 01 00 11 00 000 000 000 000
More informationConvex Optimization. Dani Yogatama. School of Computer Science, Carnegie Mellon University, Pittsburgh, PA, USA. February 12, 2014
Convex Optimization Dani Yogatama School of Computer Science, Carnegie Mellon University, Pittsburgh, PA, USA February 12, 2014 Dani Yogatama (Carnegie Mellon University) Convex Optimization February 12,
More informationLogistic Map, Euler & Runge-Kutta Method and Lotka-Volterra Equations
Logistic Map, Euler & Runge-Kutta Method and Lotka-Volterra Equations S. Y. Ha and J. Park Department of Mathematical Sciences Seoul National University Sep 23, 2013 Contents 1 Logistic Map 2 Euler and
More informationUnconstrained minimization: assumptions
Unconstrained minimization I terminology and assumptions I gradient descent method I steepest descent method I Newton s method I self-concordant functions I implementation IOE 611: Nonlinear Programming,
More informationA Brief Review on Convex Optimization
A Brief Review on Convex Optimization 1 Convex set S R n is convex if x,y S, λ,µ 0, λ+µ = 1 λx+µy S geometrically: x,y S line segment through x,y S examples (one convex, two nonconvex sets): A Brief Review
More informationAn Accelerated Hybrid Proximal Extragradient Method for Convex Optimization and its Implications to Second-Order Methods
An Accelerated Hybrid Proximal Extragradient Method for Convex Optimization and its Implications to Second-Order Methods Renato D.C. Monteiro B. F. Svaiter May 10, 011 Revised: May 4, 01) Abstract This
More informationHigher order derivative
2 Î 3 á Higher order derivative Î 1 Å Iterated partial derivative Iterated partial derivative Suppose f has f/ x, f/ y and ( f ) = 2 f x x x 2 ( f ) = 2 f x y x y ( f ) = 2 f y x y x ( f ) = 2 f y y y
More informationActive sets, steepest descent, and smooth approximation of functions
Active sets, steepest descent, and smooth approximation of functions Dmitriy Drusvyatskiy School of ORIE, Cornell University Joint work with Alex D. Ioffe (Technion), Martin Larsson (EPFL), and Adrian
More informationProximal methods. S. Villa. October 7, 2014
Proximal methods S. Villa October 7, 2014 1 Review of the basics Often machine learning problems require the solution of minimization problems. For instance, the ERM algorithm requires to solve a problem
More informationOn the interior of the simplex, we have the Hessian of d(x), Hd(x) is diagonal with ith. µd(w) + w T c. minimize. subject to w T 1 = 1,
Math 30 Winter 05 Solution to Homework 3. Recognizing the convexity of g(x) := x log x, from Jensen s inequality we get d(x) n x + + x n n log x + + x n n where the equality is attained only at x = (/n,...,
More informationORIE 6326: Convex Optimization. Quasi-Newton Methods
ORIE 6326: Convex Optimization Quasi-Newton Methods Professor Udell Operations Research and Information Engineering Cornell April 10, 2017 Slides on steepest descent and analysis of Newton s method adapted
More informationSteepest descent method with a generalized Armijo search for quasiconvex functions on Riemannian manifolds
J. Math. Anal. Appl. 341 (2008) 467 477 www.elsevier.com/locate/jmaa Steepest descent method with a generalized Armijo search for quasiconvex functions on Riemannian manifolds E.A. Papa Quiroz, E.M. Quispe,
More informationMA102: Multivariable Calculus
MA102: Multivariable Calculus Rupam Barman and Shreemayee Bora Department of Mathematics IIT Guwahati Differentiability of f : U R n R m Definition: Let U R n be open. Then f : U R n R m is differentiable
More informationMore First-Order Optimization Algorithms
More First-Order Optimization Algorithms Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A. http://www.stanford.edu/ yyye Chapters 3, 8, 3 The SDM
More informationInterior Point Methods for Mathematical Programming
Interior Point Methods for Mathematical Programming Clóvis C. Gonzaga Federal University of Santa Catarina, Florianópolis, Brazil EURO - 2013 Roma Our heroes Cauchy Newton Lagrange Early results Unconstrained
More informationAlgorithms for constrained local optimization
Algorithms for constrained local optimization Fabio Schoen 2008 http://gol.dsi.unifi.it/users/schoen Algorithms for constrained local optimization p. Feasible direction methods Algorithms for constrained
More informationSequential Unconstrained Minimization: A Survey
Sequential Unconstrained Minimization: A Survey Charles L. Byrne February 21, 2013 Abstract The problem is to minimize a function f : X (, ], over a non-empty subset C of X, where X is an arbitrary set.
More informationSequential convex programming,: value function and convergence
Sequential convex programming,: value function and convergence Edouard Pauwels joint work with Jérôme Bolte Journées MODE Toulouse March 23 2016 1 / 16 Introduction Local search methods for finite dimensional
More informationEE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 17
EE/ACM 150 - Applications of Convex Optimization in Signal Processing and Communications Lecture 17 Andre Tkacenko Signal Processing Research Group Jet Propulsion Laboratory May 29, 2012 Andre Tkacenko
More informationConvex Optimization. Problem set 2. Due Monday April 26th
Convex Optimization Problem set 2 Due Monday April 26th 1 Gradient Decent without Line-search In this problem we will consider gradient descent with predetermined step sizes. That is, instead of determining
More informationAuxiliary-Function Methods in Iterative Optimization
Auxiliary-Function Methods in Iterative Optimization Charles L. Byrne April 6, 2015 Abstract Let C X be a nonempty subset of an arbitrary set X and f : X R. The problem is to minimize f over C. In auxiliary-function
More informationConvex Optimization. 9. Unconstrained minimization. Prof. Ying Cui. Department of Electrical Engineering Shanghai Jiao Tong University
Convex Optimization 9. Unconstrained minimization Prof. Ying Cui Department of Electrical Engineering Shanghai Jiao Tong University 2017 Autumn Semester SJTU Ying Cui 1 / 40 Outline Unconstrained minimization
More informationThe general programming problem is the nonlinear programming problem where a given function is maximized subject to a set of inequality constraints.
1 Optimization Mathematical programming refers to the basic mathematical problem of finding a maximum to a function, f, subject to some constraints. 1 In other words, the objective is to find a point,
More informationA proximal-like algorithm for a class of nonconvex programming
Pacific Journal of Optimization, vol. 4, pp. 319-333, 2008 A proximal-like algorithm for a class of nonconvex programming Jein-Shan Chen 1 Department of Mathematics National Taiwan Normal University Taipei,
More informationOptimization Methods. Lecture 19: Line Searches and Newton s Method
15.93 Optimization Methods Lecture 19: Line Searches and Newton s Method 1 Last Lecture Necessary Conditions for Optimality (identifies candidates) x local min f(x ) =, f(x ) PSD Slide 1 Sufficient Conditions
More information3. Convex functions. basic properties and examples. operations that preserve convexity. the conjugate function. quasiconvex functions
3. Convex functions Convex Optimization Boyd & Vandenberghe basic properties and examples operations that preserve convexity the conjugate function quasiconvex functions log-concave and log-convex functions
More informationConstrained Optimization and Lagrangian Duality
CIS 520: Machine Learning Oct 02, 2017 Constrained Optimization and Lagrangian Duality Lecturer: Shivani Agarwal Disclaimer: These notes are designed to be a supplement to the lecture. They may or may
More information3. Convex functions. basic properties and examples. operations that preserve convexity. the conjugate function. quasiconvex functions
3. Convex functions Convex Optimization Boyd & Vandenberghe basic properties and examples operations that preserve convexity the conjugate function quasiconvex functions log-concave and log-convex functions
More informationAgenda. Fast proximal gradient methods. 1 Accelerated first-order methods. 2 Auxiliary sequences. 3 Convergence analysis. 4 Numerical examples
Agenda Fast proximal gradient methods 1 Accelerated first-order methods 2 Auxiliary sequences 3 Convergence analysis 4 Numerical examples 5 Optimality of Nesterov s scheme Last time Proximal gradient method
More informationOptimization. Escuela de Ingeniería Informática de Oviedo. (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30
Optimization Escuela de Ingeniería Informática de Oviedo (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30 Unconstrained optimization Outline 1 Unconstrained optimization 2 Constrained
More information10 Numerical methods for constrained problems
10 Numerical methods for constrained problems min s.t. f(x) h(x) = 0 (l), g(x) 0 (m), x X The algorithms can be roughly divided the following way: ˆ primal methods: find descent direction keeping inside
More information12. Interior-point methods
12. Interior-point methods Convex Optimization Boyd & Vandenberghe inequality constrained minimization logarithmic barrier function and central path barrier method feasibility and phase I methods complexity
More informationAsymptotic Convergence of the Steepest Descent Method for the Exponential Penalty in Linear Programming
Journal of Convex Analysis Volume 2 (1995), No.1/2, 145 152 Asymptotic Convergence of the Steepest Descent Method for the Exponential Penalty in Linear Programming R. Cominetti 1 Universidad de Chile,
More informationLecture Notes on Iterative Optimization Algorithms
Charles L. Byrne Department of Mathematical Sciences University of Massachusetts Lowell December 8, 2014 Lecture Notes on Iterative Optimization Algorithms Contents Preface vii 1 Overview and Examples
More informationOn the Convergence of the Entropy-Exponential Penalty Trajectories and Generalized Proximal Point Methods in Semidefinite Optimization
On the Convergence of the Entropy-Exponential Penalty Trajectories and Generalized Proximal Point Methods in Semidefinite Optimization Ferreira, O. P. Oliveira, P. R. Silva, R. C. M. June 02, 2007 Abstract
More information5. Duality. Lagrangian
5. Duality Convex Optimization Boyd & Vandenberghe Lagrange dual problem weak and strong duality geometric interpretation optimality conditions perturbation and sensitivity analysis examples generalized
More informationHW1 solutions. 1. α Ef(x) β, where Ef(x) is the expected value of f(x), i.e., Ef(x) = n. i=1 p if(a i ). (The function f : R R is given.
HW1 solutions Exercise 1 (Some sets of probability distributions.) Let x be a real-valued random variable with Prob(x = a i ) = p i, i = 1,..., n, where a 1 < a 2 < < a n. Of course p R n lies in the standard
More informationSOR- and Jacobi-type Iterative Methods for Solving l 1 -l 2 Problems by Way of Fenchel Duality 1
SOR- and Jacobi-type Iterative Methods for Solving l 1 -l 2 Problems by Way of Fenchel Duality 1 Masao Fukushima 2 July 17 2010; revised February 4 2011 Abstract We present an SOR-type algorithm and a
More informationEE 546, Univ of Washington, Spring Proximal mapping. introduction. review of conjugate functions. proximal mapping. Proximal mapping 6 1
EE 546, Univ of Washington, Spring 2012 6. Proximal mapping introduction review of conjugate functions proximal mapping Proximal mapping 6 1 Proximal mapping the proximal mapping (prox-operator) of a convex
More informationLecture 5. The Dual Cone and Dual Problem
IE 8534 1 Lecture 5. The Dual Cone and Dual Problem IE 8534 2 For a convex cone K, its dual cone is defined as K = {y x, y 0, x K}. The inner-product can be replaced by x T y if the coordinates of the
More informationMath 273a: Optimization Convex Conjugacy
Math 273a: Optimization Convex Conjugacy Instructor: Wotao Yin Department of Mathematics, UCLA Fall 2015 online discussions on piazza.com Convex conjugate (the Legendre transform) Let f be a closed proper
More informationNewton s Method. Ryan Tibshirani Convex Optimization /36-725
Newton s Method Ryan Tibshirani Convex Optimization 10-725/36-725 1 Last time: dual correspondences Given a function f : R n R, we define its conjugate f : R n R, Properties and examples: f (y) = max x
More information6. Proximal gradient method
L. Vandenberghe EE236C (Spring 2016) 6. Proximal gradient method motivation proximal mapping proximal gradient method with fixed step size proximal gradient method with line search 6-1 Proximal mapping
More informationOn the Convergence and O(1/N) Complexity of a Class of Nonlinear Proximal Point Algorithms for Monotonic Variational Inequalities
STATISTICS,OPTIMIZATION AND INFORMATION COMPUTING Stat., Optim. Inf. Comput., Vol. 2, June 204, pp 05 3. Published online in International Academic Press (www.iapress.org) On the Convergence and O(/N)
More informationRegularity of solutions to Hamilton-Jacobi equations for Tonelli Hamiltonians
Regularity of solutions to Hamilton-Jacobi equations for Tonelli Hamiltonians Université Nice Sophia Antipolis & Institut Universitaire de France Nonlinear Analysis and Optimization Royal Society, London,
More informationLEARNING IN CONCAVE GAMES
LEARNING IN CONCAVE GAMES P. Mertikopoulos French National Center for Scientific Research (CNRS) Laboratoire d Informatique de Grenoble GSBE ETBC seminar Maastricht, October 22, 2015 Motivation and Preliminaries
More informationLecture 1: Introduction. Outline. B9824 Foundations of Optimization. Fall Administrative matters. 2. Introduction. 3. Existence of optima
B9824 Foundations of Optimization Lecture 1: Introduction Fall 2009 Copyright 2009 Ciamac Moallemi Outline 1. Administrative matters 2. Introduction 3. Existence of optima 4. Local theory of unconstrained
More informationDually Flat Geometries in the State Space of Statistical Models
1/ 12 Dually Flat Geometries in the State Space of Statistical Models Jan Naudts Universiteit Antwerpen ECEA, November 2016 J. Naudts, Dually Flat Geometries in the State Space of Statistical Models. In
More informationDUALIZATION OF SUBGRADIENT CONDITIONS FOR OPTIMALITY
DUALIZATION OF SUBGRADIENT CONDITIONS FOR OPTIMALITY R. T. Rockafellar* Abstract. A basic relationship is derived between generalized subgradients of a given function, possibly nonsmooth and nonconvex,
More informationLECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE
LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE CONVEX ANALYSIS AND DUALITY Basic concepts of convex analysis Basic concepts of convex optimization Geometric duality framework - MC/MC Constrained optimization
More informationELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications
ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications Professor M. Chiang Electrical Engineering Department, Princeton University March
More informationLocal strong convexity and local Lipschitz continuity of the gradient of convex functions
Local strong convexity and local Lipschitz continuity of the gradient of convex functions R. Goebel and R.T. Rockafellar May 23, 2007 Abstract. Given a pair of convex conjugate functions f and f, we investigate
More informationFast proximal gradient methods
L. Vandenberghe EE236C (Spring 2013-14) Fast proximal gradient methods fast proximal gradient method (FISTA) FISTA with line search FISTA as descent method Nesterov s second method 1 Fast (proximal) gradient
More informationQuasi-invariant Measures on Path Space. Denis Bell University of North Florida
Quasi-invariant Measures on Path Space Denis Bell University of North Florida Transformation of measure under the flow of a vector field Let E be a vector space (or a manifold), equipped with a finite
More informationKey words. saddle-point dynamics, asymptotic convergence, convex-concave functions, proximal calculus, center manifold theory, nonsmooth dynamics
SADDLE-POINT DYNAMICS: CONDITIONS FOR ASYMPTOTIC STABILITY OF SADDLE POINTS ASHISH CHERUKURI, BAHMAN GHARESIFARD, AND JORGE CORTÉS Abstract. This paper considers continuously differentiable functions of
More informationNonsmooth optimization: conditioning, convergence, and semi-algebraic models
Nonsmooth optimization: conditioning, convergence, and semi-algebraic models Adrian Lewis ORIE Cornell International Congress of Mathematicians Seoul, August 2014 1/16 Outline I Optimization and inverse
More informationminimize x subject to (x 2)(x 4) u,
Math 6366/6367: Optimization and Variational Methods Sample Preliminary Exam Questions 1. Suppose that f : [, L] R is a C 2 -function with f () on (, L) and that you have explicit formulae for
More informationIntroduction to Information Geometry
Introduction to Information Geometry based on the book Methods of Information Geometry written by Shun-Ichi Amari and Hiroshi Nagaoka Yunshu Liu 2012-02-17 Outline 1 Introduction to differential geometry
More informationNon-smooth Non-convex Bregman Minimization: Unification and new Algorithms
Non-smooth Non-convex Bregman Minimization: Unification and new Algorithms Peter Ochs, Jalal Fadili, and Thomas Brox Saarland University, Saarbrücken, Germany Normandie Univ, ENSICAEN, CNRS, GREYC, France
More informationFrank-Wolfe Method. Ryan Tibshirani Convex Optimization
Frank-Wolfe Method Ryan Tibshirani Convex Optimization 10-725 Last time: ADMM For the problem min x,z f(x) + g(z) subject to Ax + Bz = c we form augmented Lagrangian (scaled form): L ρ (x, z, w) = f(x)
More informationPrimal-Dual Symmetric Interior-Point Methods from SDP to Hyperbolic Cone Programming and Beyond
Primal-Dual Symmetric Interior-Point Methods from SDP to Hyperbolic Cone Programming and Beyond Tor Myklebust Levent Tunçel September 26, 2014 Convex Optimization in Conic Form (P) inf c, x A(x) = b, x
More information