A Feasible Method for Optimization with Orthogonality Constraints

Size: px
Start display at page:

Download "A Feasible Method for Optimization with Orthogonality Constraints"

Transcription

1 A Feasible Method for Optimization with Orthogonality Constraints Zaiwen Wen 1 Wotao Yin 2 1 Shanghai Jiaotong University, Math Dept 2 Rice University, CAAM Dept May 2011

2 Orthogonality Constraints Unknowns: q orthogonal matrices X (1),..., X (q) R n p Optimization problem: min F (X), subject to X (i) X (i) = I, i = 1,..., q. X (i) is constrained to the Stiefel manifold.

3 The Vector Case If p = 1 X (i) reduces to x (i) R n X(i) X (i) = I becomes x (i) 2 = 1 or x (i) S n 1. Example: n = 2, q = 2, the feasible set is homeomorphic to a torus By Dave Burke

4 Why are orthogonality/normalization constraints interesting? wide applications; non-convex in general, causing possibly local minima; expensive to stay feasible in the matrix case. Applications arise with 1. Direction fields: x ij 2 = 1 2. Homogeneous objectives and normalization: x 2 = 1, X F = 1 3. Eigenvalue problems: X X = I 4. Certain SDPs (Burer et al): 0 Z = X X, Z ii = x i 2 = 1 5. Certain combinatorial problems: Square X is a permutation matrix X X = I and X

5 The Method

6 Stay Feasible vs Infeasible Infeasible algorithms Penalty, augmented Lagrangian methods SDP relaxations / lifting Feasible algorithms iterative projection descent along geodesics (or retractions) Advantages: Work when cost functions are meaningless outside of feas. set In case of early termination, a feasible sol is returned Fewer parameters (e.g., for penalty) Disadvantages: maintaining feasibility can be expensive

7 Basic Idea Consider At iteration k min F(X), subject to X X = I. X k+1 Orthogonalize X k + τproj( F k ) X k+1 Orthogonalize X k + τh k X k X k+1 solutiony (τ) of Y = X k + τ 2 H k(x k + Y ) Need H k such that 1. Y (τ) Y (τ) = Xk X k; 2. Y (τ) is a descent curve; 3. solving for Y (τ) is fast.

8 Example: S 2 = {x R 3 : x 2 = 1} S

9 Example: S 2 = {x R 3 : x 2 = 1} x x S

10 Example: S 2 = {x R 3 : x 2 = 1} x f(x) x S 2 geodesic

11 Example: S 2 = {x R 3 : x 2 = 1} x f(x) S 2 x f(x) x geodesic Hx Proj( f ) = x ( f x)

12 Example: S 2 = {x R 3 : x 2 = 1} x f(x) S 2 x f(x) x geodesic Hx Proj( f ) = x ( f x) = ( x f f x ) x = Hx.

13 Example: S 2 = {x R 3 : x 2 = 1} x f(x) S 2 x f(x) x 1 2 Hx geodesic Hx Proj( f ) = x ( f x) = ( x f f x ) x = Hx.

14 Example: S 2 = {x R 3 : x 2 = 1} x f(x) S 2 x f(x) x 1 2 Hx geodesic 1 2 Hy Hx 0.7 y Proj( f ) = ( f x) x = ( x f f x ) x = Hx. Solution y of y = x + 1 2H(x + y) is feasible. Solution {y(τ)} of y = x + τ 2 H(x + y) is a geodesic.

15 Relation to Crank-Nicolson Discretize 1D PDE u t = F (u, x, t, u x, u x ): (forward Euler) (backward Euler) (Crank-Nicolson) u n+1 j u n+1 j u n+1 j uj n t uj n t uj n t = F n j ( ) = F n+1 j ( ) = 1 2 ( F n+1 j ) ( ) + Fj n ( ) Crank-Nicolson is the average forward Euler and backward Euler:

16 The Matrix Update Given X k and G k = F(X k ), we let H = X k G k G kx k and solve Y = X + τ 2 H(X + Y ) for Y (τ). Update X k+1 Y (τ) = where τ is a step size. Properties of Y (τ): 1. Y (τ) Y (τ) = X k X k; 2. Y (τ) is a descent curve; 3. Y (τ) has fast implementations. (I τ 2 H ) 1 (I + τ 2 H ) X k,

17 Skew-Symmetric H Preserves X X Theorem (Cayley s Transform) If H = H, then 1. (I H) is nonsingular; 2. Y = (I H) 1 (I + H)X satisfies Y Y = X X. Proof. 1. x (I H)x = x x for any vector x. 2. Y Y = X (I + H) (I H) T (I H) 1 (I + H)X = X (I H) (I + H) 1 (I H) 1 (I + H)X = X (I H) (I H) 1 (I + H) 1 (I + H)X = X X.

18 Generalization For M 0 and complex-valued constraints X MX = C. The update Y = (I HM) 1 (I + HM)X preserves Y MY = X MX. provided that H is a skew-hermitian matrix.

19 We now know that X X can be preserved by X k+1 Y (τ) = (I τ 2 H ) 1 (I + τ 2 H ) X k. This is known as Cayley s transform, which has been used in matrix computations such as inverse eigenvalue problems [Friedland-Nocedal- Overton 87]. It is also called Crank-Nicolson schemes for heat PDEs, etc. It appears that it has not been systematically studied for minimizing a general differentiable F(X) subject to X X = I.

20 Properties of Y (τ): 1. Y (τ) Y (τ) = Xk X k; 2. Y (τ) is a descent curve; 3. Y (τ) has fast implementations.

21 Y (τ) is a descent curve since Y (0) = negative projected gradient. Y (0) = HX k = Proj c ( F k ; X k ). Proj c is projection to the tangent of {X : X X = I} under the canonical metric: Y, Z c = tr(y (I 1 2 XX )Z ), Y, Z T X. For projection under the Euclidean metric: Y, Z e = tr(y Z ), use H = XG k (I 1 2 X kx k ) (I 1 2 X kx k )G kx. Generally, given any tangent direction D T X (CG, Newton, etc.), leads to Y (0) = D. H = (I 1 2 XX )DX XD (I 1 2 XX )

22 Properties of Y (τ): 1. Y (τ) Y (τ) = Xk X k; 2. Y (τ) is a descent curve; 3. Y (τ) has fast implementations.

23 Comparing with Descent along Geodesics Moving along geodesics is a sensible choice (c.f. Smith 93, Mahony 94; Edelman-Arias-Smith 98) for optimization on manifold, but geodesics of {X : X X = I} are difficult to compute. Y (τ) = ( I τ 2 H) 1 ( I + τ 2 H) X is not a geodesic. It is a retraction ( approximate geodesic ) of the Stiefel manifold; see book by Absil, Mahony, and Sepulchre.

24 Computing Cost Y (τ) = (I τ 2 H ) 1 (I + τ 2 H ) X. p = 1: Y (τ) is closed-form, a linear combination of X and F, small p: apply the SMW formula and solve a smaller (2p 2p) system, large p: solve an n n system, OR perform low-rank approximates. Cheaper than: geodesic steps: approximating matrix exponentials or solving 2nd-order PDEs; gradient projection: matrix orthogonalization, n p SVD (e.g., Manton 02); gradient + orthogonalization: QR decomposition.

25 In short, we proposed a curvilinear descent path, which maintains feasible and easy to compute. The path Y (τ) is neither straight nor a geodesic, and τ is not linear in the length of Y (τ).

26 Convergence Properties Theorem Consider min F (X), s.t. X X = I. X is a stationary point if and only if X X = I and HX = 0. If HX 0, then Y (τ) is a descent curve. We also proved X k globally converges to a stationary point, near a stationary point, τ is lower bounded, HX = 0 if and only if H = 0. As X k X station, (I τ 2 H) I, so (I τ 2 H) is well-behaved.

27 Full Algorithm and Numerical Results

28 Algorithm: 1. Input: feasible X 0 2. k 0 3. While stopping conditions not met do 4. H X k ( F k ) ( F k )Xk 5. Compute an initial τ by Barzilai-Borwein 6. τ non-monotonic line search 7. X k+1 Y (τ) = ( I τ 2 H) 1 ( I + τ 2 H) X k 8. k k + 1 Code was written in MATLAB.

29 Three types of problems: 1. Guaranteed global minimum; 2. Numerical global minimum, but no proof yet; or just lucky? 3. Local (non-global) minima; we restart algorithm from random points, and pick the best solution.

30 Max-Cut SDP SDP: max X tr(cx), s.t. X ii = 1, i = 1,, n, X 0. NLP: write X = V V where V = [v 1,..., v n ] R p n max V R p n c ij v i v j, s.t. v i = 1, i = 1,..., n. i,j If p rank(x ), NLP is equivalent to SDP. Theorem (a slight extension of Burer-Monteiro 05) There exists p(c) n such that, if p p(c), any local minimizer of NLP is globally optimal. A simple upper bound: p(c) n + 1 inf{rank(c + D) : D is diagonal}.

31 Max-Cut SDP SDPLR for max-cut, C language Our code Name n obj CPU p obj CPU nfe feasi t..g e e e-14 t..pm e e e-14 G e e e-15 G e e e-15 G e e e-14 G e e e-14 G e e e-14 G e e e-14 G e e e-14 G e e e-14 G e e e-14 G e e e-14 G e e e-14 G e e e-14 G e e e-14 G e e e-14 G e e e-14 G e e e-14 G e e e-14 G e e e-14 SDRLP for max-cut by S. Burer and R. Monteiro

32 Low-Rank Nearest Correlation Matrix Estimation Given a weight matrix W and a symmetric matrix C: min X W (X C) 2 F, X ii = 1, i = 1,..., n, rank(x) r. Low-rank approximation: X = V V, V R r n 1 min V R r n 2 W (V V C) 2 F, s.t. V i 2 = 1, i = 1,..., n. Test examples: Ex1: n = 500, G ij = e 0.05 i j, W = I Ex3: n = 500, G ij = e 0.05 i j, random W Ex5: n = 943, G is based on the Movielens data sets. W is given by T. Fushiki

33 Low-Rank nearest correlation problem Major PenCorr Our code r obj CPU feasi obj CPU feasi obj CPU feasi Ex e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e-15 Ex e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e-15 Ex e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e-14 Major: Pietersz-Groenen, PenCorr: Y.Gao-D.Sun

34 Linear eigenvalue problem Let λ 1... λ n be the eigenvalues of A: p i=1 λ i := max X R n p tr(x AX) s.t. X X = I Compared with MATLAB s eigs (which calls Fortran library ARPACK and incurs overhead) on: Dense random matrices UF Sparse Matrix Collection: 39 large sparse matrices with n 4000

35 Dense matrix: sum of 6 largest eigenvalues n eigs obj 1.153e e e e e e+05 feasi 5.972e e e e e e-15 nax cpu Our code obj 1.153e e e e e e+05 feasi 6.468e e e e e e-15 nfe cpu err 1.255e e e e e e-06

36 Dense matrix: fix n = 5000 p eigs obj 1.986e e e e e e e+05 nax feasi 1.776e e e e e e e-14 cpu Our code obj 1.986e e e e e e e+05 feasi 4.441e e e e e e e-15 nfe cpu err 1.325e e e e e e e-07

37 Sparse matrix: sum of 2 largest eigenvalues eigs Our code Name n obj CPU nax obj err CPU nfe crystm e e e ct20stif e e e msc e e e msc e e e pwtk e e e s3rmt3m e e e bcsstk e e e bcsstm e e e Kuu e e e Muu e e e finan e e e nd3k e e e nasa e e e nasasrb e e e fv e fv e fv e t2dal e e aft e e e cfd e e e

38 Quadratic Assignment Problem (QAP) Minimizing a quadratic over permutation matrices: min tr(a XBX ), s.t. X X = I, X 0 X R n n Tested 134 problems in QAPLIB: rewrite the objective as tr(a (X X)B(X X) ) apply augmented Lagrangian to X 0 with parameter µ run our code 40 times from random X 0 with µ = 0.1, 1, 10 each pick the best out of 120 rounded solutions

39 QAPLIB: exact recovery gap = best upper bound v best upper bound 100%, CPU: average cpu Our code Name n obj µ min gap % med gap % max gap % CPU feasi chr12b e e+00 esc e e+00 esc16a e e+00 esc16b e e+00 esc16c e e+00 esc16d e e+00 esc16e e e+00 esc16f e e+00 esc16g e e+00 esc16h e e+00 esc16i e e+00 esc16j e e+00 esc32b e e+00 esc32c e e+00 esc32d e e+00 esc32e e e+00 esc32g e e+00 esc64a e e+00

40 QAPLIB: exact recovery Our code Name n obj µ min gap % med gap % max gap % CPU feasi had e e+00 had e e+00 had e e+00 had e e+00 had e e+00 lipa20a e e+00 lipa20b e e+00 lipa30a e e+00 lipa30b e e+00 lipa40b e e+00 lipa50b e e+00 lipa60b e e+00 lipa70b e e+00 lipa80b e e+00 lipa90b e e+00 nug e e+00 nug e e+00 nug16a e e+00 nug16b e e+00 tai12a e e+00

41 QAPLIB: n 80 Our code Name n obj µ min gap% med gap% max gap% CPU feasi esc e e+00 lipa80a e e+00 lipa80b e e+00 lipa90a e e+00 lipa90b e e+00 sko100a e e+00 sko100b e e+00 sko100c e e+00 sko100d e e+00 sko100e e e+00 sko100f e e+00 sko e e+00 sko e e+00 tai100a e e+00 tai100b e e+00 tai150b e e+00 tai256c e e+00 tai80a e e+00 tai80b e e+00 tho e e+00 wil e e+00 exact recovery

42 Polynomial Optimization Consider minimizing the following polynomials subject to x 2 = 1: 1. ( i j + k + l)x i x j x k x l i<j<k<l 50 1 i<j<k 49 1 i 20 1 i<j<k 20 x i x j x k + x 2 i x j x 2 i x k + x j x 2 k x 6 i + 1 i 19 x 3 i x 3 i+1 x 2 i x 2 j x 2 k + x 3 i x 2 j x k + x 2 i x 3 j x k + x i x 3 j x 2 k SDP (Nie) Our code F(x) f hmg sos cpu rep F(x) (min, mean, max) CPU feasi :30:00 10 ( , , ) e :00:00 10 ( , , ) e e-5 10:00: (6e-5, , ) e :00: ( , , ) e-15 Problems from J.Nie, Regularization Methods for Sum of Squares Relaxations... Nie s code solves an SDP relaxation, which has approximation guarantees.

43 Stability Number of Graph Given a graph G = (V, E), Motzkin and Straus showed α(g) 1 min n xi (i,j) E x i 2 xj 2, s.t. x 2 = 1 i=1 Test random graphs G whose edges e i,j for {i, j} M are generated with probability 1 2, where M is a subset chosen randomly from V with M = n/2. SDP (Nie) Our code n CPU α(g) rep CPU (min, mean, max) feasi 20 0:03: ( 0.004, 0.006, 0.007) 1.7e :39: ( 0.007, 0.009, 0.011) 7.8e :21: ( 0.005, 0.007, 0.010) 1.4e :27: ( 0.008, 0.010, 0.012) 5.5e :46: ( 0.009, 0.012, 0.015) 3.6e ( 0.013, 0.015, 0.020) 2.0e ( 0.018, 0.023, 0.036) 2.4e-15

44 1-Bit CS Recover x from min x 1, s.t. Ax 0, x 2 = 1. Our approach (joint J. Laska, Z. Wen, and R. Baraniuk) 1. Outer: form augmented Lagrangian subproblem and update µ, B and b. min µ x Bx b 2 2, s.t. x 2 = 1, 2. Inner: solve the above subproblem by iterating x k+1 min x 2 =1 µτ x x (x k + τg k ) 2 2

45 Shrinkage on Sphere Shrinkage (soft-thresholding): shrink(y, µ) min x µ x x y 2 2. Shrinkage on sphere: add constraint x 2 = 1 x opt { shrink(y,µ) shrink(y,µ) 2, shrink(y, µ) 0, sign(y i )e i, o.w., where y i is the largest (in magnitude) entry of y. Matrix case: min X µ X X Y 2 F, s.t. X F = 1. reduces to shrinkage on sphere over singular values.

A Feasible Method for Optimization with Orthogonality Constraints

A Feasible Method for Optimization with Orthogonality Constraints Noname manuscript No. (will be inserted by the editor) A Feasible Method for Optimization with Orthogonality Constraints Zaiwen Wen Wotao Yin Received: date / Accepted: date Abstract Minimization with

More information

A FEASIBLE METHOD FOR OPTIMIZATION WITH ORTHOGONALITY CONSTRAINTS. November 7, 2010

A FEASIBLE METHOD FOR OPTIMIZATION WITH ORTHOGONALITY CONSTRAINTS. November 7, 2010 A FEASIBLE METHOD FOR OPTIMIZATION WITH ORTHOGONALITY CONSTRAINTS ZAIWEN WEN AND WOTAO YIN November 7, 2010 Abstract. Minimization with orthogonality constraints (e.g., X X = I) and/or spherical constraints

More information

Lift me up but not too high Fast algorithms to solve SDP s with block-diagonal constraints

Lift me up but not too high Fast algorithms to solve SDP s with block-diagonal constraints Lift me up but not too high Fast algorithms to solve SDP s with block-diagonal constraints Nicolas Boumal Université catholique de Louvain (Belgium) IDeAS seminar, May 13 th, 2014, Princeton The Riemannian

More information

Low-Rank Factorization Models for Matrix Completion and Matrix Separation

Low-Rank Factorization Models for Matrix Completion and Matrix Separation for Matrix Completion and Matrix Separation Joint work with Wotao Yin, Yin Zhang and Shen Yuan IPAM, UCLA Oct. 5, 2010 Low rank minimization problems Matrix completion: find a low-rank matrix W R m n so

More information

Robust Principal Component Pursuit via Alternating Minimization Scheme on Matrix Manifolds

Robust Principal Component Pursuit via Alternating Minimization Scheme on Matrix Manifolds Robust Principal Component Pursuit via Alternating Minimization Scheme on Matrix Manifolds Tao Wu Institute for Mathematics and Scientific Computing Karl-Franzens-University of Graz joint work with Prof.

More information

Coordinate Update Algorithm Short Course Proximal Operators and Algorithms

Coordinate Update Algorithm Short Course Proximal Operators and Algorithms Coordinate Update Algorithm Short Course Proximal Operators and Algorithms Instructor: Wotao Yin (UCLA Math) Summer 2016 1 / 36 Why proximal? Newton s method: for C 2 -smooth, unconstrained problems allow

More information

A New Trust Region Algorithm Using Radial Basis Function Models

A New Trust Region Algorithm Using Radial Basis Function Models A New Trust Region Algorithm Using Radial Basis Function Models Seppo Pulkkinen University of Turku Department of Mathematics July 14, 2010 Outline 1 Introduction 2 Background Taylor series approximations

More information

Coordinate Update Algorithm Short Course Operator Splitting

Coordinate Update Algorithm Short Course Operator Splitting Coordinate Update Algorithm Short Course Operator Splitting Instructor: Wotao Yin (UCLA Math) Summer 2016 1 / 25 Operator splitting pipeline 1. Formulate a problem as 0 A(x) + B(x) with monotone operators

More information

Estimating Bounds for Quadratic Assignment Problems Associated with Hamming and Manhattan Distance Matrices based on Semidefinite Programming

Estimating Bounds for Quadratic Assignment Problems Associated with Hamming and Manhattan Distance Matrices based on Semidefinite Programming Estimating Bounds for Quadratic Assignment Problems Associated with Hamming and Manhattan Distance Matrices based on Semidefinite Programming Hans Mittelmann Jiming Peng September 17, 010 Abstract Quadratic

More information

Infeasible Primal-Dual (Path-Following) Interior-Point Methods for Semidefinite Programming*

Infeasible Primal-Dual (Path-Following) Interior-Point Methods for Semidefinite Programming* Infeasible Primal-Dual (Path-Following) Interior-Point Methods for Semidefinite Programming* Notes for CAAM 564, Spring 2012 Dept of CAAM, Rice University Outline (1) Introduction (2) Formulation & a complexity

More information

Sparse Optimization Lecture: Basic Sparse Optimization Models

Sparse Optimization Lecture: Basic Sparse Optimization Models Sparse Optimization Lecture: Basic Sparse Optimization Models Instructor: Wotao Yin July 2013 online discussions on piazza.com Those who complete this lecture will know basic l 1, l 2,1, and nuclear-norm

More information

A Continuation Approach Using NCP Function for Solving Max-Cut Problem

A Continuation Approach Using NCP Function for Solving Max-Cut Problem A Continuation Approach Using NCP Function for Solving Max-Cut Problem Xu Fengmin Xu Chengxian Ren Jiuquan Abstract A continuous approach using NCP function for approximating the solution of the max-cut

More information

N. L. P. NONLINEAR PROGRAMMING (NLP) deals with optimization models with at least one nonlinear function. NLP. Optimization. Models of following form:

N. L. P. NONLINEAR PROGRAMMING (NLP) deals with optimization models with at least one nonlinear function. NLP. Optimization. Models of following form: 0.1 N. L. P. Katta G. Murty, IOE 611 Lecture slides Introductory Lecture NONLINEAR PROGRAMMING (NLP) deals with optimization models with at least one nonlinear function. NLP does not include everything

More information

Solving large Semidefinite Programs - Part 1 and 2

Solving large Semidefinite Programs - Part 1 and 2 Solving large Semidefinite Programs - Part 1 and 2 Franz Rendl http://www.math.uni-klu.ac.at Alpen-Adria-Universität Klagenfurt Austria F. Rendl, Singapore workshop 2006 p.1/34 Overview Limits of Interior

More information

Tightening a Discrete Formulation of the Quadratic Assignment Problem

Tightening a Discrete Formulation of the Quadratic Assignment Problem A publication of 1309 CHEMICAL ENGINEERING TRANSACTIONS VOL. 32, 2013 Chief Editors: Sauro Pierucci, Jiří J. Klemeš Copyright 2013, AIDIC Servizi S.r.l., ISBN 978-88-95608-23-5; ISSN 1974-9791 The Italian

More information

Inverse Singular Value Problems

Inverse Singular Value Problems Chapter 8 Inverse Singular Value Problems IEP versus ISVP Existence question A continuous approach An iterative method for the IEP An iterative method for the ISVP 139 140 Lecture 8 IEP versus ISVP Inverse

More information

Copositive Programming and Combinatorial Optimization

Copositive Programming and Combinatorial Optimization Copositive Programming and Combinatorial Optimization Franz Rendl http://www.math.uni-klu.ac.at Alpen-Adria-Universität Klagenfurt Austria joint work with I.M. Bomze (Wien) and F. Jarre (Düsseldorf) IMA

More information

A non-monotone Inexact Restoration approach for minimization with orthogonality constraints

A non-monotone Inexact Restoration approach for minimization with orthogonality constraints A non-monotone Inexact Restoration approach for minimization with orthogonality constraints Juliano B. Francisco * Douglas S. Gonçalves Fermín S. V. Bazán Lila L. T. Paredes Federal University of Santa

More information

c 2010 Society for Industrial and Applied Mathematics

c 2010 Society for Industrial and Applied Mathematics SIAM J. OPTIM. Vol. 20, No. 5, pp. 2327 2351 c 2010 Society for Industrial and Applied Mathematics LOW-RANK OPTIMIZATION ON THE CONE OF POSITIVE SEMIDEFINITE MATRICES M. JOURNÉE, F. BACH, P.-A. ABSIL,

More information

Nonlinear Optimization: What s important?

Nonlinear Optimization: What s important? Nonlinear Optimization: What s important? Julian Hall 10th May 2012 Convexity: convex problems A local minimizer is a global minimizer A solution of f (x) = 0 (stationary point) is a minimizer A global

More information

Quadratic reformulation techniques for 0-1 quadratic programs

Quadratic reformulation techniques for 0-1 quadratic programs OSE SEMINAR 2014 Quadratic reformulation techniques for 0-1 quadratic programs Ray Pörn CENTER OF EXCELLENCE IN OPTIMIZATION AND SYSTEMS ENGINEERING ÅBO AKADEMI UNIVERSITY ÅBO NOVEMBER 14th 2014 2 Structure

More information

Acyclic Semidefinite Approximations of Quadratically Constrained Quadratic Programs

Acyclic Semidefinite Approximations of Quadratically Constrained Quadratic Programs Acyclic Semidefinite Approximations of Quadratically Constrained Quadratic Programs Raphael Louca & Eilyan Bitar School of Electrical and Computer Engineering American Control Conference (ACC) Chicago,

More information

LINEAR SYSTEMS (11) Intensive Computation

LINEAR SYSTEMS (11) Intensive Computation LINEAR SYSTEMS () Intensive Computation 27-8 prof. Annalisa Massini Viviana Arrigoni EXACT METHODS:. GAUSSIAN ELIMINATION. 2. CHOLESKY DECOMPOSITION. ITERATIVE METHODS:. JACOBI. 2. GAUSS-SEIDEL 2 CHOLESKY

More information

Math 273a: Optimization Overview of First-Order Optimization Algorithms

Math 273a: Optimization Overview of First-Order Optimization Algorithms Math 273a: Optimization Overview of First-Order Optimization Algorithms Wotao Yin Department of Mathematics, UCLA online discussions on piazza.com 1 / 9 Typical flow of numerical optimization Optimization

More information

Line Search Methods for Unconstrained Optimisation

Line Search Methods for Unconstrained Optimisation Line Search Methods for Unconstrained Optimisation Lecture 8, Numerical Linear Algebra and Optimisation Oxford University Computing Laboratory, MT 2007 Dr Raphael Hauser (hauser@comlab.ox.ac.uk) The Generic

More information

Approximation Algorithms

Approximation Algorithms Approximation Algorithms Chapter 26 Semidefinite Programming Zacharias Pitouras 1 Introduction LP place a good lower bound on OPT for NP-hard problems Are there other ways of doing this? Vector programs

More information

, b = 0. (2) 1 2 The eigenvectors of A corresponding to the eigenvalues λ 1 = 1, λ 2 = 3 are

, b = 0. (2) 1 2 The eigenvectors of A corresponding to the eigenvalues λ 1 = 1, λ 2 = 3 are Quadratic forms We consider the quadratic function f : R 2 R defined by f(x) = 2 xt Ax b T x with x = (x, x 2 ) T, () where A R 2 2 is symmetric and b R 2. We will see that, depending on the eigenvalues

More information

1 Non-negative Matrix Factorization (NMF)

1 Non-negative Matrix Factorization (NMF) 2018-06-21 1 Non-negative Matrix Factorization NMF) In the last lecture, we considered low rank approximations to data matrices. We started with the optimal rank k approximation to A R m n via the SVD,

More information

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.

More information

Bindel, Fall 2009 Matrix Computations (CS 6210) Week 8: Friday, Oct 17

Bindel, Fall 2009 Matrix Computations (CS 6210) Week 8: Friday, Oct 17 Logistics Week 8: Friday, Oct 17 1. HW 3 errata: in Problem 1, I meant to say p i < i, not that p i is strictly ascending my apologies. You would want p i > i if you were simply forming the matrices and

More information

1 Newton s Method. Suppose we want to solve: x R. At x = x, f (x) can be approximated by:

1 Newton s Method. Suppose we want to solve: x R. At x = x, f (x) can be approximated by: Newton s Method Suppose we want to solve: (P:) min f (x) At x = x, f (x) can be approximated by: n x R. f (x) h(x) := f ( x)+ f ( x) T (x x)+ (x x) t H ( x)(x x), 2 which is the quadratic Taylor expansion

More information

Infeasible Primal-Dual (Path-Following) Interior-Point Methods for Semidefinite Programming*

Infeasible Primal-Dual (Path-Following) Interior-Point Methods for Semidefinite Programming* Infeasible Primal-Dual (Path-Following) Interior-Point Methods for Semidefinite Programming* Yin Zhang Dept of CAAM, Rice University Outline (1) Introduction (2) Formulation & a complexity theorem (3)

More information

Riemannian Optimization and its Application to Phase Retrieval Problem

Riemannian Optimization and its Application to Phase Retrieval Problem Riemannian Optimization and its Application to Phase Retrieval Problem 1/46 Introduction Motivations Optimization History Phase Retrieval Problem Summary Joint work with: Pierre-Antoine Absil, Professor

More information

PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.1/39

PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.1/39 PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP Michal Kočvara Institute of Information Theory and Automation Academy of Sciences of the Czech Republic and Czech Technical University

More information

Fast and Robust Phase Retrieval

Fast and Robust Phase Retrieval Fast and Robust Phase Retrieval Aditya Viswanathan aditya@math.msu.edu CCAM Lunch Seminar Purdue University April 18 2014 0 / 27 Joint work with Yang Wang Mark Iwen Research supported in part by National

More information

Sparse Solutions of an Undetermined Linear System

Sparse Solutions of an Undetermined Linear System 1 Sparse Solutions of an Undetermined Linear System Maddullah Almerdasy New York University Tandon School of Engineering arxiv:1702.07096v1 [math.oc] 23 Feb 2017 Abstract This work proposes a research

More information

XLVI Pesquisa Operacional na Gestão da Segurança Pública

XLVI Pesquisa Operacional na Gestão da Segurança Pública A linear formulation with O(n 2 ) variables for the quadratic assignment problem Serigne Gueye and Philippe Michelon Université d Avignon et des Pays du Vaucluse, Laboratoire d Informatique d Avignon (LIA),

More information

FINITE-DIMENSIONAL LINEAR ALGEBRA

FINITE-DIMENSIONAL LINEAR ALGEBRA DISCRETE MATHEMATICS AND ITS APPLICATIONS Series Editor KENNETH H ROSEN FINITE-DIMENSIONAL LINEAR ALGEBRA Mark S Gockenbach Michigan Technological University Houghton, USA CRC Press Taylor & Francis Croup

More information

Outline Introduction: Problem Description Diculties Algebraic Structure: Algebraic Varieties Rank Decient Toeplitz Matrices Constructing Lower Rank St

Outline Introduction: Problem Description Diculties Algebraic Structure: Algebraic Varieties Rank Decient Toeplitz Matrices Constructing Lower Rank St Structured Lower Rank Approximation by Moody T. Chu (NCSU) joint with Robert E. Funderlic (NCSU) and Robert J. Plemmons (Wake Forest) March 5, 1998 Outline Introduction: Problem Description Diculties Algebraic

More information

Linear Algebra Massoud Malek

Linear Algebra Massoud Malek CSUEB Linear Algebra Massoud Malek Inner Product and Normed Space In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n An inner product

More information

Solving linear equations with Gaussian Elimination (I)

Solving linear equations with Gaussian Elimination (I) Term Projects Solving linear equations with Gaussian Elimination The QR Algorithm for Symmetric Eigenvalue Problem The QR Algorithm for The SVD Quasi-Newton Methods Solving linear equations with Gaussian

More information

A solution approach for linear optimization with completely positive matrices

A solution approach for linear optimization with completely positive matrices A solution approach for linear optimization with completely positive matrices Franz Rendl http://www.math.uni-klu.ac.at Alpen-Adria-Universität Klagenfurt Austria joint work with M. Bomze (Wien) and F.

More information

Sparse Covariance Selection using Semidefinite Programming

Sparse Covariance Selection using Semidefinite Programming Sparse Covariance Selection using Semidefinite Programming A. d Aspremont ORFE, Princeton University Joint work with O. Banerjee, L. El Ghaoui & G. Natsoulis, U.C. Berkeley & Iconix Pharmaceuticals Support

More information

ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications

ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications Professor M. Chiang Electrical Engineering Department, Princeton University March

More information

Projection methods to solve SDP

Projection methods to solve SDP Projection methods to solve SDP Franz Rendl http://www.math.uni-klu.ac.at Alpen-Adria-Universität Klagenfurt Austria F. Rendl, Oberwolfach Seminar, May 2010 p.1/32 Overview Augmented Primal-Dual Method

More information

1 Computing with constraints

1 Computing with constraints Notes for 2017-04-26 1 Computing with constraints Recall that our basic problem is minimize φ(x) s.t. x Ω where the feasible set Ω is defined by equality and inequality conditions Ω = {x R n : c i (x)

More information

Lecture 6 Positive Definite Matrices

Lecture 6 Positive Definite Matrices Linear Algebra Lecture 6 Positive Definite Matrices Prof. Chun-Hung Liu Dept. of Electrical and Computer Engineering National Chiao Tung University Spring 2017 2017/6/8 Lecture 6: Positive Definite Matrices

More information

Sparse Optimization Lecture: Dual Methods, Part I

Sparse Optimization Lecture: Dual Methods, Part I Sparse Optimization Lecture: Dual Methods, Part I Instructor: Wotao Yin July 2013 online discussions on piazza.com Those who complete this lecture will know dual (sub)gradient iteration augmented l 1 iteration

More information

Linear algebra issues in Interior Point methods for bound-constrained least-squares problems

Linear algebra issues in Interior Point methods for bound-constrained least-squares problems Linear algebra issues in Interior Point methods for bound-constrained least-squares problems Stefania Bellavia Dipartimento di Energetica S. Stecco Università degli Studi di Firenze Joint work with Jacek

More information

An Iterative Descent Method

An Iterative Descent Method Conjugate Gradient: An Iterative Descent Method The Plan Review Iterative Descent Conjugate Gradient Review : Iterative Descent Iterative Descent is an unconstrained optimization process x (k+1) = x (k)

More information

A NEW FIRST-ORDER ALGORITHMIC FRAMEWORK FOR OPTIMIZATION PROBLEMS WITH ORTHOGONALITY CONSTRAINTS. min

A NEW FIRST-ORDER ALGORITHMIC FRAMEWORK FOR OPTIMIZATION PROBLEMS WITH ORTHOGONALITY CONSTRAINTS. min A NEW FIRST-ORDER ALGORITHMIC FRAMEWORK FOR OPTIMIZATION PROBLEMS WITH ORTHOGONALITY CONSTRAINTS BIN GAO, XIN LIU, XIAOJUN CHEN, AND YA-XIANG YUAN Abstract. In this paper, we consider a class of optimization

More information

Approximation algorithms for nonnegative polynomial optimization problems over unit spheres

Approximation algorithms for nonnegative polynomial optimization problems over unit spheres Front. Math. China 2017, 12(6): 1409 1426 https://doi.org/10.1007/s11464-017-0644-1 Approximation algorithms for nonnegative polynomial optimization problems over unit spheres Xinzhen ZHANG 1, Guanglu

More information

Large-Scale L1-Related Minimization in Compressive Sensing and Beyond

Large-Scale L1-Related Minimization in Compressive Sensing and Beyond Large-Scale L1-Related Minimization in Compressive Sensing and Beyond Yin Zhang Department of Computational and Applied Mathematics Rice University, Houston, Texas, U.S.A. Arizona State University March

More information

Determination of Feasible Directions by Successive Quadratic Programming and Zoutendijk Algorithms: A Comparative Study

Determination of Feasible Directions by Successive Quadratic Programming and Zoutendijk Algorithms: A Comparative Study International Journal of Mathematics And Its Applications Vol.2 No.4 (2014), pp.47-56. ISSN: 2347-1557(online) Determination of Feasible Directions by Successive Quadratic Programming and Zoutendijk Algorithms:

More information

minimize x subject to (x 2)(x 4) u,

minimize x subject to (x 2)(x 4) u, Math 6366/6367: Optimization and Variational Methods Sample Preliminary Exam Questions 1. Suppose that f : [, L] R is a C 2 -function with f () on (, L) and that you have explicit formulae for

More information

Block Coordinate Descent for Regularized Multi-convex Optimization

Block Coordinate Descent for Regularized Multi-convex Optimization Block Coordinate Descent for Regularized Multi-convex Optimization Yangyang Xu and Wotao Yin CAAM Department, Rice University February 15, 2013 Multi-convex optimization Model definition Applications Outline

More information

Riemannian Optimization Method on the Flag Manifold for Independent Subspace Analysis

Riemannian Optimization Method on the Flag Manifold for Independent Subspace Analysis Riemannian Optimization Method on the Flag Manifold for Independent Subspace Analysis Yasunori Nishimori 1, Shotaro Akaho 1, and Mark D. Plumbley 2 1 Neuroscience Research Institute, National Institute

More information

5 Handling Constraints

5 Handling Constraints 5 Handling Constraints Engineering design optimization problems are very rarely unconstrained. Moreover, the constraints that appear in these problems are typically nonlinear. This motivates our interest

More information

Lecture: Algorithms for LP, SOCP and SDP

Lecture: Algorithms for LP, SOCP and SDP 1/53 Lecture: Algorithms for LP, SOCP and SDP Zaiwen Wen Beijing International Center For Mathematical Research Peking University http://bicmr.pku.edu.cn/~wenzw/bigdata2018.html wenzw@pku.edu.cn Acknowledgement:

More information

Unconstrained Optimization Models for Computing Several Extreme Eigenpairs of Real Symmetric Matrices

Unconstrained Optimization Models for Computing Several Extreme Eigenpairs of Real Symmetric Matrices Unconstrained Optimization Models for Computing Several Extreme Eigenpairs of Real Symmetric Matrices Yu-Hong Dai LSEC, ICMSEC Academy of Mathematics and Systems Science Chinese Academy of Sciences Joint

More information

Numerical Methods I Non-Square and Sparse Linear Systems

Numerical Methods I Non-Square and Sparse Linear Systems Numerical Methods I Non-Square and Sparse Linear Systems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 September 25th, 2014 A. Donev (Courant

More information

The nonsmooth Newton method on Riemannian manifolds

The nonsmooth Newton method on Riemannian manifolds The nonsmooth Newton method on Riemannian manifolds C. Lageman, U. Helmke, J.H. Manton 1 Introduction Solving nonlinear equations in Euclidean space is a frequently occurring problem in optimization and

More information

A Quick Tour of Linear Algebra and Optimization for Machine Learning

A Quick Tour of Linear Algebra and Optimization for Machine Learning A Quick Tour of Linear Algebra and Optimization for Machine Learning Masoud Farivar January 8, 2015 1 / 28 Outline of Part I: Review of Basic Linear Algebra Matrices and Vectors Matrix Multiplication Operators

More information

A direct formulation for sparse PCA using semidefinite programming

A direct formulation for sparse PCA using semidefinite programming A direct formulation for sparse PCA using semidefinite programming A. d Aspremont, L. El Ghaoui, M. Jordan, G. Lanckriet ORFE, Princeton University & EECS, U.C. Berkeley Available online at www.princeton.edu/~aspremon

More information

PDE-Constrained and Nonsmooth Optimization

PDE-Constrained and Nonsmooth Optimization Frank E. Curtis October 1, 2009 Outline PDE-Constrained Optimization Introduction Newton s method Inexactness Results Summary and future work Nonsmooth Optimization Sequential quadratic programming (SQP)

More information

Preliminary Examination, Numerical Analysis, August 2016

Preliminary Examination, Numerical Analysis, August 2016 Preliminary Examination, Numerical Analysis, August 2016 Instructions: This exam is closed books and notes. The time allowed is three hours and you need to work on any three out of questions 1-4 and any

More information

An Inexact Newton Method for Nonlinear Constrained Optimization

An Inexact Newton Method for Nonlinear Constrained Optimization An Inexact Newton Method for Nonlinear Constrained Optimization Frank E. Curtis Numerical Analysis Seminar, January 23, 2009 Outline Motivation and background Algorithm development and theoretical results

More information

An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization

An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization Frank E. Curtis, Lehigh University involving joint work with Travis Johnson, Northwestern University Daniel P. Robinson, Johns

More information

Part 4: Active-set methods for linearly constrained optimization. Nick Gould (RAL)

Part 4: Active-set methods for linearly constrained optimization. Nick Gould (RAL) Part 4: Active-set methods for linearly constrained optimization Nick Gould RAL fx subject to Ax b Part C course on continuoue optimization LINEARLY CONSTRAINED MINIMIZATION fx subject to Ax { } b where

More information

A Customized ADMM for Rank-Constrained Optimization Problems with Approximate Formulations

A Customized ADMM for Rank-Constrained Optimization Problems with Approximate Formulations A Customized ADMM for Rank-Constrained Optimization Problems with Approximate Formulations Chuangchuang Sun and Ran Dai Abstract This paper proposes a customized Alternating Direction Method of Multipliers

More information

CONSTRAINED NONLINEAR PROGRAMMING

CONSTRAINED NONLINEAR PROGRAMMING 149 CONSTRAINED NONLINEAR PROGRAMMING We now turn to methods for general constrained nonlinear programming. These may be broadly classified into two categories: 1. TRANSFORMATION METHODS: In this approach

More information

Conditional Gradient Algorithms for Rank-One Matrix Approximations with a Sparsity Constraint

Conditional Gradient Algorithms for Rank-One Matrix Approximations with a Sparsity Constraint Conditional Gradient Algorithms for Rank-One Matrix Approximations with a Sparsity Constraint Marc Teboulle School of Mathematical Sciences Tel Aviv University Joint work with Ronny Luss Optimization and

More information

Alternating Direction Augmented Lagrangian Algorithms for Convex Optimization

Alternating Direction Augmented Lagrangian Algorithms for Convex Optimization Alternating Direction Augmented Lagrangian Algorithms for Convex Optimization Joint work with Bo Huang, Shiqian Ma, Tony Qin, Katya Scheinberg, Zaiwen Wen and Wotao Yin Department of IEOR, Columbia University

More information

Basic Math for

Basic Math for Basic Math for 16-720 August 23, 2002 1 Linear Algebra 1.1 Vectors and Matrices First, a reminder of a few basic notations, definitions, and terminology: Unless indicated otherwise, vectors are always

More information

A Note on Semidefinite Programming Relaxations For Polynomial Optimization Over a Single Sphere

A Note on Semidefinite Programming Relaxations For Polynomial Optimization Over a Single Sphere . ARTICLES. doi: A Note on Semidefinite Programming Relaxations For Polynomial Optimization Over a Single Sphere Jiang Hu,, Bo jiang, Xin Liu 3 & Zaiwen Wen Beijing International Center for Mathematical

More information

The Q Method for Symmetric Cone Programmin

The Q Method for Symmetric Cone Programmin The Q Method for Symmetric Cone Programming The Q Method for Symmetric Cone Programmin Farid Alizadeh and Yu Xia alizadeh@rutcor.rutgers.edu, xiay@optlab.mcma Large Scale Nonlinear and Semidefinite Progra

More information

Optimisation on Manifolds

Optimisation on Manifolds Optimisation on Manifolds K. Hüper MPI Tübingen & Univ. Würzburg K. Hüper (MPI Tübingen & Univ. Würzburg) Applications in Computer Vision Grenoble 18/9/08 1 / 29 Contents 2 Examples Essential matrix estimation

More information

ECE580 Exam 1 October 4, Please do not write on the back of the exam pages. Extra paper is available from the instructor.

ECE580 Exam 1 October 4, Please do not write on the back of the exam pages. Extra paper is available from the instructor. ECE580 Exam 1 October 4, 2012 1 Name: Solution Score: /100 You must show ALL of your work for full credit. This exam is closed-book. Calculators may NOT be used. Please leave fractions as fractions, etc.

More information

SDPNAL+: A Majorized Semismooth Newton-CG Augmented Lagrangian Method for Semidefinite Programming with Nonnegative Constraints

SDPNAL+: A Majorized Semismooth Newton-CG Augmented Lagrangian Method for Semidefinite Programming with Nonnegative Constraints SDPNAL+: A Majorized Semismooth Newton-CG Augmented Lagrangian Method for Semidefinite Programming with Nonnegative Constraints Liuqin Yang, Defeng Sun, Kim-Chuan Toh May 28, 2014; Revised on Mar 10, 2015

More information

SDPNAL+: A Majorized Semismooth Newton-CG Augmented Lagrangian Method for Semidefinite Programming with Nonnegative Constraints

SDPNAL+: A Majorized Semismooth Newton-CG Augmented Lagrangian Method for Semidefinite Programming with Nonnegative Constraints arxiv:1406.0942v1 [math.oc] 4 Jun 2014 SDPNAL+: A Majorized Semismooth Newton-CG Augmented Lagrangian Method for Semidefinite Programming with Nonnegative Constraints Liuqin Yang, Defeng Sun, Kim-Chuan

More information

An Active Set Strategy for Solving Optimization Problems with up to 200,000,000 Nonlinear Constraints

An Active Set Strategy for Solving Optimization Problems with up to 200,000,000 Nonlinear Constraints An Active Set Strategy for Solving Optimization Problems with up to 200,000,000 Nonlinear Constraints Klaus Schittkowski Department of Computer Science, University of Bayreuth 95440 Bayreuth, Germany e-mail:

More information

CHAPTER 11. A Revision. 1. The Computers and Numbers therein

CHAPTER 11. A Revision. 1. The Computers and Numbers therein CHAPTER A Revision. The Computers and Numbers therein Traditional computer science begins with a finite alphabet. By stringing elements of the alphabet one after another, one obtains strings. A set of

More information

Proximal Gradient Descent and Acceleration. Ryan Tibshirani Convex Optimization /36-725

Proximal Gradient Descent and Acceleration. Ryan Tibshirani Convex Optimization /36-725 Proximal Gradient Descent and Acceleration Ryan Tibshirani Convex Optimization 10-725/36-725 Last time: subgradient method Consider the problem min f(x) with f convex, and dom(f) = R n. Subgradient method:

More information

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection Eigenvalue Problems Last Time Social Network Graphs Betweenness Girvan-Newman Algorithm Graph Laplacian Spectral Bisection λ 2, w 2 Today Small deviation into eigenvalue problems Formulation Standard eigenvalue

More information

Computation. For QDA we need to calculate: Lets first consider the case that

Computation. For QDA we need to calculate: Lets first consider the case that Computation For QDA we need to calculate: δ (x) = 1 2 log( Σ ) 1 2 (x µ ) Σ 1 (x µ ) + log(π ) Lets first consider the case that Σ = I,. This is the case where each distribution is spherical, around the

More information

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2013 PROBLEM SET 2

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2013 PROBLEM SET 2 STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2013 PROBLEM SET 2 1. You are not allowed to use the svd for this problem, i.e. no arguments should depend on the svd of A or A. Let W be a subspace of C n. The

More information

Robust PCA. CS5240 Theoretical Foundations in Multimedia. Leow Wee Kheng

Robust PCA. CS5240 Theoretical Foundations in Multimedia. Leow Wee Kheng Robust PCA CS5240 Theoretical Foundations in Multimedia Leow Wee Kheng Department of Computer Science School of Computing National University of Singapore Leow Wee Kheng (NUS) Robust PCA 1 / 52 Previously...

More information

Copositive Programming and Combinatorial Optimization

Copositive Programming and Combinatorial Optimization Copositive Programming and Combinatorial Optimization Franz Rendl http://www.math.uni-klu.ac.at Alpen-Adria-Universität Klagenfurt Austria joint work with M. Bomze (Wien) and F. Jarre (Düsseldorf) and

More information

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44 Convex Optimization Newton s method ENSAE: Optimisation 1/44 Unconstrained minimization minimize f(x) f convex, twice continuously differentiable (hence dom f open) we assume optimal value p = inf x f(x)

More information

Lecture 8: February 9

Lecture 8: February 9 0-725/36-725: Convex Optimiation Spring 205 Lecturer: Ryan Tibshirani Lecture 8: February 9 Scribes: Kartikeya Bhardwaj, Sangwon Hyun, Irina Caan 8 Proximal Gradient Descent In the previous lecture, we

More information

Newton s Method and Linear Approximations

Newton s Method and Linear Approximations Newton s Method and Linear Approximations Curves are tricky. Lines aren t. Newton s Method and Linear Approximations Newton s Method for finding roots Goal: Where is f (x) = 0? f (x) = x 7 + 3x 3 + 7x

More information

Part 5: Penalty and augmented Lagrangian methods for equality constrained optimization. Nick Gould (RAL)

Part 5: Penalty and augmented Lagrangian methods for equality constrained optimization. Nick Gould (RAL) Part 5: Penalty and augmented Lagrangian methods for equality constrained optimization Nick Gould (RAL) x IR n f(x) subject to c(x) = Part C course on continuoue optimization CONSTRAINED MINIMIZATION x

More information

Preface to Second Edition... vii. Preface to First Edition...

Preface to Second Edition... vii. Preface to First Edition... Contents Preface to Second Edition..................................... vii Preface to First Edition....................................... ix Part I Linear Algebra 1 Basic Vector/Matrix Structure and

More information

Minimizing the Difference of L 1 and L 2 Norms with Applications

Minimizing the Difference of L 1 and L 2 Norms with Applications 1/36 Minimizing the Difference of L 1 and L 2 Norms with Department of Mathematical Sciences University of Texas Dallas May 31, 2017 Partially supported by NSF DMS 1522786 2/36 Outline 1 A nonconvex approach:

More information

ONP-MF: An Orthogonal Nonnegative Matrix Factorization Algorithm with Application to Clustering

ONP-MF: An Orthogonal Nonnegative Matrix Factorization Algorithm with Application to Clustering ONP-MF: An Orthogonal Nonnegative Matrix Factorization Algorithm with Application to Clustering Filippo Pompili 1, Nicolas Gillis 2, P.-A. Absil 2,andFrançois Glineur 2,3 1- University of Perugia, Department

More information

Optimization. Yuh-Jye Lee. March 21, Data Science and Machine Intelligence Lab National Chiao Tung University 1 / 29

Optimization. Yuh-Jye Lee. March 21, Data Science and Machine Intelligence Lab National Chiao Tung University 1 / 29 Optimization Yuh-Jye Lee Data Science and Machine Intelligence Lab National Chiao Tung University March 21, 2017 1 / 29 You Have Learned (Unconstrained) Optimization in Your High School Let f (x) = ax

More information

A direct formulation for sparse PCA using semidefinite programming

A direct formulation for sparse PCA using semidefinite programming A direct formulation for sparse PCA using semidefinite programming A. d Aspremont, L. El Ghaoui, M. Jordan, G. Lanckriet ORFE, Princeton University & EECS, U.C. Berkeley A. d Aspremont, INFORMS, Denver,

More information

Non-negative Matrix Factorization via accelerated Projected Gradient Descent

Non-negative Matrix Factorization via accelerated Projected Gradient Descent Non-negative Matrix Factorization via accelerated Projected Gradient Descent Andersen Ang Mathématique et recherche opérationnelle UMONS, Belgium Email: manshun.ang@umons.ac.be Homepage: angms.science

More information

Nonlinear Optimization for Optimal Control

Nonlinear Optimization for Optimal Control Nonlinear Optimization for Optimal Control Pieter Abbeel UC Berkeley EECS Many slides and figures adapted from Stephen Boyd [optional] Boyd and Vandenberghe, Convex Optimization, Chapters 9 11 [optional]

More information

Numerical Methods in Matrix Computations

Numerical Methods in Matrix Computations Ake Bjorck Numerical Methods in Matrix Computations Springer Contents 1 Direct Methods for Linear Systems 1 1.1 Elements of Matrix Theory 1 1.1.1 Matrix Algebra 2 1.1.2 Vector Spaces 6 1.1.3 Submatrices

More information