A Feasible Method for Optimization with Orthogonality Constraints
|
|
- Lilian Jacobs
- 5 years ago
- Views:
Transcription
1 A Feasible Method for Optimization with Orthogonality Constraints Zaiwen Wen 1 Wotao Yin 2 1 Shanghai Jiaotong University, Math Dept 2 Rice University, CAAM Dept May 2011
2 Orthogonality Constraints Unknowns: q orthogonal matrices X (1),..., X (q) R n p Optimization problem: min F (X), subject to X (i) X (i) = I, i = 1,..., q. X (i) is constrained to the Stiefel manifold.
3 The Vector Case If p = 1 X (i) reduces to x (i) R n X(i) X (i) = I becomes x (i) 2 = 1 or x (i) S n 1. Example: n = 2, q = 2, the feasible set is homeomorphic to a torus By Dave Burke
4 Why are orthogonality/normalization constraints interesting? wide applications; non-convex in general, causing possibly local minima; expensive to stay feasible in the matrix case. Applications arise with 1. Direction fields: x ij 2 = 1 2. Homogeneous objectives and normalization: x 2 = 1, X F = 1 3. Eigenvalue problems: X X = I 4. Certain SDPs (Burer et al): 0 Z = X X, Z ii = x i 2 = 1 5. Certain combinatorial problems: Square X is a permutation matrix X X = I and X
5 The Method
6 Stay Feasible vs Infeasible Infeasible algorithms Penalty, augmented Lagrangian methods SDP relaxations / lifting Feasible algorithms iterative projection descent along geodesics (or retractions) Advantages: Work when cost functions are meaningless outside of feas. set In case of early termination, a feasible sol is returned Fewer parameters (e.g., for penalty) Disadvantages: maintaining feasibility can be expensive
7 Basic Idea Consider At iteration k min F(X), subject to X X = I. X k+1 Orthogonalize X k + τproj( F k ) X k+1 Orthogonalize X k + τh k X k X k+1 solutiony (τ) of Y = X k + τ 2 H k(x k + Y ) Need H k such that 1. Y (τ) Y (τ) = Xk X k; 2. Y (τ) is a descent curve; 3. solving for Y (τ) is fast.
8 Example: S 2 = {x R 3 : x 2 = 1} S
9 Example: S 2 = {x R 3 : x 2 = 1} x x S
10 Example: S 2 = {x R 3 : x 2 = 1} x f(x) x S 2 geodesic
11 Example: S 2 = {x R 3 : x 2 = 1} x f(x) S 2 x f(x) x geodesic Hx Proj( f ) = x ( f x)
12 Example: S 2 = {x R 3 : x 2 = 1} x f(x) S 2 x f(x) x geodesic Hx Proj( f ) = x ( f x) = ( x f f x ) x = Hx.
13 Example: S 2 = {x R 3 : x 2 = 1} x f(x) S 2 x f(x) x 1 2 Hx geodesic Hx Proj( f ) = x ( f x) = ( x f f x ) x = Hx.
14 Example: S 2 = {x R 3 : x 2 = 1} x f(x) S 2 x f(x) x 1 2 Hx geodesic 1 2 Hy Hx 0.7 y Proj( f ) = ( f x) x = ( x f f x ) x = Hx. Solution y of y = x + 1 2H(x + y) is feasible. Solution {y(τ)} of y = x + τ 2 H(x + y) is a geodesic.
15 Relation to Crank-Nicolson Discretize 1D PDE u t = F (u, x, t, u x, u x ): (forward Euler) (backward Euler) (Crank-Nicolson) u n+1 j u n+1 j u n+1 j uj n t uj n t uj n t = F n j ( ) = F n+1 j ( ) = 1 2 ( F n+1 j ) ( ) + Fj n ( ) Crank-Nicolson is the average forward Euler and backward Euler:
16 The Matrix Update Given X k and G k = F(X k ), we let H = X k G k G kx k and solve Y = X + τ 2 H(X + Y ) for Y (τ). Update X k+1 Y (τ) = where τ is a step size. Properties of Y (τ): 1. Y (τ) Y (τ) = X k X k; 2. Y (τ) is a descent curve; 3. Y (τ) has fast implementations. (I τ 2 H ) 1 (I + τ 2 H ) X k,
17 Skew-Symmetric H Preserves X X Theorem (Cayley s Transform) If H = H, then 1. (I H) is nonsingular; 2. Y = (I H) 1 (I + H)X satisfies Y Y = X X. Proof. 1. x (I H)x = x x for any vector x. 2. Y Y = X (I + H) (I H) T (I H) 1 (I + H)X = X (I H) (I + H) 1 (I H) 1 (I + H)X = X (I H) (I H) 1 (I + H) 1 (I + H)X = X X.
18 Generalization For M 0 and complex-valued constraints X MX = C. The update Y = (I HM) 1 (I + HM)X preserves Y MY = X MX. provided that H is a skew-hermitian matrix.
19 We now know that X X can be preserved by X k+1 Y (τ) = (I τ 2 H ) 1 (I + τ 2 H ) X k. This is known as Cayley s transform, which has been used in matrix computations such as inverse eigenvalue problems [Friedland-Nocedal- Overton 87]. It is also called Crank-Nicolson schemes for heat PDEs, etc. It appears that it has not been systematically studied for minimizing a general differentiable F(X) subject to X X = I.
20 Properties of Y (τ): 1. Y (τ) Y (τ) = Xk X k; 2. Y (τ) is a descent curve; 3. Y (τ) has fast implementations.
21 Y (τ) is a descent curve since Y (0) = negative projected gradient. Y (0) = HX k = Proj c ( F k ; X k ). Proj c is projection to the tangent of {X : X X = I} under the canonical metric: Y, Z c = tr(y (I 1 2 XX )Z ), Y, Z T X. For projection under the Euclidean metric: Y, Z e = tr(y Z ), use H = XG k (I 1 2 X kx k ) (I 1 2 X kx k )G kx. Generally, given any tangent direction D T X (CG, Newton, etc.), leads to Y (0) = D. H = (I 1 2 XX )DX XD (I 1 2 XX )
22 Properties of Y (τ): 1. Y (τ) Y (τ) = Xk X k; 2. Y (τ) is a descent curve; 3. Y (τ) has fast implementations.
23 Comparing with Descent along Geodesics Moving along geodesics is a sensible choice (c.f. Smith 93, Mahony 94; Edelman-Arias-Smith 98) for optimization on manifold, but geodesics of {X : X X = I} are difficult to compute. Y (τ) = ( I τ 2 H) 1 ( I + τ 2 H) X is not a geodesic. It is a retraction ( approximate geodesic ) of the Stiefel manifold; see book by Absil, Mahony, and Sepulchre.
24 Computing Cost Y (τ) = (I τ 2 H ) 1 (I + τ 2 H ) X. p = 1: Y (τ) is closed-form, a linear combination of X and F, small p: apply the SMW formula and solve a smaller (2p 2p) system, large p: solve an n n system, OR perform low-rank approximates. Cheaper than: geodesic steps: approximating matrix exponentials or solving 2nd-order PDEs; gradient projection: matrix orthogonalization, n p SVD (e.g., Manton 02); gradient + orthogonalization: QR decomposition.
25 In short, we proposed a curvilinear descent path, which maintains feasible and easy to compute. The path Y (τ) is neither straight nor a geodesic, and τ is not linear in the length of Y (τ).
26 Convergence Properties Theorem Consider min F (X), s.t. X X = I. X is a stationary point if and only if X X = I and HX = 0. If HX 0, then Y (τ) is a descent curve. We also proved X k globally converges to a stationary point, near a stationary point, τ is lower bounded, HX = 0 if and only if H = 0. As X k X station, (I τ 2 H) I, so (I τ 2 H) is well-behaved.
27 Full Algorithm and Numerical Results
28 Algorithm: 1. Input: feasible X 0 2. k 0 3. While stopping conditions not met do 4. H X k ( F k ) ( F k )Xk 5. Compute an initial τ by Barzilai-Borwein 6. τ non-monotonic line search 7. X k+1 Y (τ) = ( I τ 2 H) 1 ( I + τ 2 H) X k 8. k k + 1 Code was written in MATLAB.
29 Three types of problems: 1. Guaranteed global minimum; 2. Numerical global minimum, but no proof yet; or just lucky? 3. Local (non-global) minima; we restart algorithm from random points, and pick the best solution.
30 Max-Cut SDP SDP: max X tr(cx), s.t. X ii = 1, i = 1,, n, X 0. NLP: write X = V V where V = [v 1,..., v n ] R p n max V R p n c ij v i v j, s.t. v i = 1, i = 1,..., n. i,j If p rank(x ), NLP is equivalent to SDP. Theorem (a slight extension of Burer-Monteiro 05) There exists p(c) n such that, if p p(c), any local minimizer of NLP is globally optimal. A simple upper bound: p(c) n + 1 inf{rank(c + D) : D is diagonal}.
31 Max-Cut SDP SDPLR for max-cut, C language Our code Name n obj CPU p obj CPU nfe feasi t..g e e e-14 t..pm e e e-14 G e e e-15 G e e e-15 G e e e-14 G e e e-14 G e e e-14 G e e e-14 G e e e-14 G e e e-14 G e e e-14 G e e e-14 G e e e-14 G e e e-14 G e e e-14 G e e e-14 G e e e-14 G e e e-14 G e e e-14 G e e e-14 SDRLP for max-cut by S. Burer and R. Monteiro
32 Low-Rank Nearest Correlation Matrix Estimation Given a weight matrix W and a symmetric matrix C: min X W (X C) 2 F, X ii = 1, i = 1,..., n, rank(x) r. Low-rank approximation: X = V V, V R r n 1 min V R r n 2 W (V V C) 2 F, s.t. V i 2 = 1, i = 1,..., n. Test examples: Ex1: n = 500, G ij = e 0.05 i j, W = I Ex3: n = 500, G ij = e 0.05 i j, random W Ex5: n = 943, G is based on the Movielens data sets. W is given by T. Fushiki
33 Low-Rank nearest correlation problem Major PenCorr Our code r obj CPU feasi obj CPU feasi obj CPU feasi Ex e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e-15 Ex e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e-15 Ex e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e-14 Major: Pietersz-Groenen, PenCorr: Y.Gao-D.Sun
34 Linear eigenvalue problem Let λ 1... λ n be the eigenvalues of A: p i=1 λ i := max X R n p tr(x AX) s.t. X X = I Compared with MATLAB s eigs (which calls Fortran library ARPACK and incurs overhead) on: Dense random matrices UF Sparse Matrix Collection: 39 large sparse matrices with n 4000
35 Dense matrix: sum of 6 largest eigenvalues n eigs obj 1.153e e e e e e+05 feasi 5.972e e e e e e-15 nax cpu Our code obj 1.153e e e e e e+05 feasi 6.468e e e e e e-15 nfe cpu err 1.255e e e e e e-06
36 Dense matrix: fix n = 5000 p eigs obj 1.986e e e e e e e+05 nax feasi 1.776e e e e e e e-14 cpu Our code obj 1.986e e e e e e e+05 feasi 4.441e e e e e e e-15 nfe cpu err 1.325e e e e e e e-07
37 Sparse matrix: sum of 2 largest eigenvalues eigs Our code Name n obj CPU nax obj err CPU nfe crystm e e e ct20stif e e e msc e e e msc e e e pwtk e e e s3rmt3m e e e bcsstk e e e bcsstm e e e Kuu e e e Muu e e e finan e e e nd3k e e e nasa e e e nasasrb e e e fv e fv e fv e t2dal e e aft e e e cfd e e e
38 Quadratic Assignment Problem (QAP) Minimizing a quadratic over permutation matrices: min tr(a XBX ), s.t. X X = I, X 0 X R n n Tested 134 problems in QAPLIB: rewrite the objective as tr(a (X X)B(X X) ) apply augmented Lagrangian to X 0 with parameter µ run our code 40 times from random X 0 with µ = 0.1, 1, 10 each pick the best out of 120 rounded solutions
39 QAPLIB: exact recovery gap = best upper bound v best upper bound 100%, CPU: average cpu Our code Name n obj µ min gap % med gap % max gap % CPU feasi chr12b e e+00 esc e e+00 esc16a e e+00 esc16b e e+00 esc16c e e+00 esc16d e e+00 esc16e e e+00 esc16f e e+00 esc16g e e+00 esc16h e e+00 esc16i e e+00 esc16j e e+00 esc32b e e+00 esc32c e e+00 esc32d e e+00 esc32e e e+00 esc32g e e+00 esc64a e e+00
40 QAPLIB: exact recovery Our code Name n obj µ min gap % med gap % max gap % CPU feasi had e e+00 had e e+00 had e e+00 had e e+00 had e e+00 lipa20a e e+00 lipa20b e e+00 lipa30a e e+00 lipa30b e e+00 lipa40b e e+00 lipa50b e e+00 lipa60b e e+00 lipa70b e e+00 lipa80b e e+00 lipa90b e e+00 nug e e+00 nug e e+00 nug16a e e+00 nug16b e e+00 tai12a e e+00
41 QAPLIB: n 80 Our code Name n obj µ min gap% med gap% max gap% CPU feasi esc e e+00 lipa80a e e+00 lipa80b e e+00 lipa90a e e+00 lipa90b e e+00 sko100a e e+00 sko100b e e+00 sko100c e e+00 sko100d e e+00 sko100e e e+00 sko100f e e+00 sko e e+00 sko e e+00 tai100a e e+00 tai100b e e+00 tai150b e e+00 tai256c e e+00 tai80a e e+00 tai80b e e+00 tho e e+00 wil e e+00 exact recovery
42 Polynomial Optimization Consider minimizing the following polynomials subject to x 2 = 1: 1. ( i j + k + l)x i x j x k x l i<j<k<l 50 1 i<j<k 49 1 i 20 1 i<j<k 20 x i x j x k + x 2 i x j x 2 i x k + x j x 2 k x 6 i + 1 i 19 x 3 i x 3 i+1 x 2 i x 2 j x 2 k + x 3 i x 2 j x k + x 2 i x 3 j x k + x i x 3 j x 2 k SDP (Nie) Our code F(x) f hmg sos cpu rep F(x) (min, mean, max) CPU feasi :30:00 10 ( , , ) e :00:00 10 ( , , ) e e-5 10:00: (6e-5, , ) e :00: ( , , ) e-15 Problems from J.Nie, Regularization Methods for Sum of Squares Relaxations... Nie s code solves an SDP relaxation, which has approximation guarantees.
43 Stability Number of Graph Given a graph G = (V, E), Motzkin and Straus showed α(g) 1 min n xi (i,j) E x i 2 xj 2, s.t. x 2 = 1 i=1 Test random graphs G whose edges e i,j for {i, j} M are generated with probability 1 2, where M is a subset chosen randomly from V with M = n/2. SDP (Nie) Our code n CPU α(g) rep CPU (min, mean, max) feasi 20 0:03: ( 0.004, 0.006, 0.007) 1.7e :39: ( 0.007, 0.009, 0.011) 7.8e :21: ( 0.005, 0.007, 0.010) 1.4e :27: ( 0.008, 0.010, 0.012) 5.5e :46: ( 0.009, 0.012, 0.015) 3.6e ( 0.013, 0.015, 0.020) 2.0e ( 0.018, 0.023, 0.036) 2.4e-15
44 1-Bit CS Recover x from min x 1, s.t. Ax 0, x 2 = 1. Our approach (joint J. Laska, Z. Wen, and R. Baraniuk) 1. Outer: form augmented Lagrangian subproblem and update µ, B and b. min µ x Bx b 2 2, s.t. x 2 = 1, 2. Inner: solve the above subproblem by iterating x k+1 min x 2 =1 µτ x x (x k + τg k ) 2 2
45 Shrinkage on Sphere Shrinkage (soft-thresholding): shrink(y, µ) min x µ x x y 2 2. Shrinkage on sphere: add constraint x 2 = 1 x opt { shrink(y,µ) shrink(y,µ) 2, shrink(y, µ) 0, sign(y i )e i, o.w., where y i is the largest (in magnitude) entry of y. Matrix case: min X µ X X Y 2 F, s.t. X F = 1. reduces to shrinkage on sphere over singular values.
A Feasible Method for Optimization with Orthogonality Constraints
Noname manuscript No. (will be inserted by the editor) A Feasible Method for Optimization with Orthogonality Constraints Zaiwen Wen Wotao Yin Received: date / Accepted: date Abstract Minimization with
More informationA FEASIBLE METHOD FOR OPTIMIZATION WITH ORTHOGONALITY CONSTRAINTS. November 7, 2010
A FEASIBLE METHOD FOR OPTIMIZATION WITH ORTHOGONALITY CONSTRAINTS ZAIWEN WEN AND WOTAO YIN November 7, 2010 Abstract. Minimization with orthogonality constraints (e.g., X X = I) and/or spherical constraints
More informationLift me up but not too high Fast algorithms to solve SDP s with block-diagonal constraints
Lift me up but not too high Fast algorithms to solve SDP s with block-diagonal constraints Nicolas Boumal Université catholique de Louvain (Belgium) IDeAS seminar, May 13 th, 2014, Princeton The Riemannian
More informationLow-Rank Factorization Models for Matrix Completion and Matrix Separation
for Matrix Completion and Matrix Separation Joint work with Wotao Yin, Yin Zhang and Shen Yuan IPAM, UCLA Oct. 5, 2010 Low rank minimization problems Matrix completion: find a low-rank matrix W R m n so
More informationRobust Principal Component Pursuit via Alternating Minimization Scheme on Matrix Manifolds
Robust Principal Component Pursuit via Alternating Minimization Scheme on Matrix Manifolds Tao Wu Institute for Mathematics and Scientific Computing Karl-Franzens-University of Graz joint work with Prof.
More informationCoordinate Update Algorithm Short Course Proximal Operators and Algorithms
Coordinate Update Algorithm Short Course Proximal Operators and Algorithms Instructor: Wotao Yin (UCLA Math) Summer 2016 1 / 36 Why proximal? Newton s method: for C 2 -smooth, unconstrained problems allow
More informationA New Trust Region Algorithm Using Radial Basis Function Models
A New Trust Region Algorithm Using Radial Basis Function Models Seppo Pulkkinen University of Turku Department of Mathematics July 14, 2010 Outline 1 Introduction 2 Background Taylor series approximations
More informationCoordinate Update Algorithm Short Course Operator Splitting
Coordinate Update Algorithm Short Course Operator Splitting Instructor: Wotao Yin (UCLA Math) Summer 2016 1 / 25 Operator splitting pipeline 1. Formulate a problem as 0 A(x) + B(x) with monotone operators
More informationEstimating Bounds for Quadratic Assignment Problems Associated with Hamming and Manhattan Distance Matrices based on Semidefinite Programming
Estimating Bounds for Quadratic Assignment Problems Associated with Hamming and Manhattan Distance Matrices based on Semidefinite Programming Hans Mittelmann Jiming Peng September 17, 010 Abstract Quadratic
More informationInfeasible Primal-Dual (Path-Following) Interior-Point Methods for Semidefinite Programming*
Infeasible Primal-Dual (Path-Following) Interior-Point Methods for Semidefinite Programming* Notes for CAAM 564, Spring 2012 Dept of CAAM, Rice University Outline (1) Introduction (2) Formulation & a complexity
More informationSparse Optimization Lecture: Basic Sparse Optimization Models
Sparse Optimization Lecture: Basic Sparse Optimization Models Instructor: Wotao Yin July 2013 online discussions on piazza.com Those who complete this lecture will know basic l 1, l 2,1, and nuclear-norm
More informationA Continuation Approach Using NCP Function for Solving Max-Cut Problem
A Continuation Approach Using NCP Function for Solving Max-Cut Problem Xu Fengmin Xu Chengxian Ren Jiuquan Abstract A continuous approach using NCP function for approximating the solution of the max-cut
More informationN. L. P. NONLINEAR PROGRAMMING (NLP) deals with optimization models with at least one nonlinear function. NLP. Optimization. Models of following form:
0.1 N. L. P. Katta G. Murty, IOE 611 Lecture slides Introductory Lecture NONLINEAR PROGRAMMING (NLP) deals with optimization models with at least one nonlinear function. NLP does not include everything
More informationSolving large Semidefinite Programs - Part 1 and 2
Solving large Semidefinite Programs - Part 1 and 2 Franz Rendl http://www.math.uni-klu.ac.at Alpen-Adria-Universität Klagenfurt Austria F. Rendl, Singapore workshop 2006 p.1/34 Overview Limits of Interior
More informationTightening a Discrete Formulation of the Quadratic Assignment Problem
A publication of 1309 CHEMICAL ENGINEERING TRANSACTIONS VOL. 32, 2013 Chief Editors: Sauro Pierucci, Jiří J. Klemeš Copyright 2013, AIDIC Servizi S.r.l., ISBN 978-88-95608-23-5; ISSN 1974-9791 The Italian
More informationInverse Singular Value Problems
Chapter 8 Inverse Singular Value Problems IEP versus ISVP Existence question A continuous approach An iterative method for the IEP An iterative method for the ISVP 139 140 Lecture 8 IEP versus ISVP Inverse
More informationCopositive Programming and Combinatorial Optimization
Copositive Programming and Combinatorial Optimization Franz Rendl http://www.math.uni-klu.ac.at Alpen-Adria-Universität Klagenfurt Austria joint work with I.M. Bomze (Wien) and F. Jarre (Düsseldorf) IMA
More informationA non-monotone Inexact Restoration approach for minimization with orthogonality constraints
A non-monotone Inexact Restoration approach for minimization with orthogonality constraints Juliano B. Francisco * Douglas S. Gonçalves Fermín S. V. Bazán Lila L. T. Paredes Federal University of Santa
More informationc 2010 Society for Industrial and Applied Mathematics
SIAM J. OPTIM. Vol. 20, No. 5, pp. 2327 2351 c 2010 Society for Industrial and Applied Mathematics LOW-RANK OPTIMIZATION ON THE CONE OF POSITIVE SEMIDEFINITE MATRICES M. JOURNÉE, F. BACH, P.-A. ABSIL,
More informationNonlinear Optimization: What s important?
Nonlinear Optimization: What s important? Julian Hall 10th May 2012 Convexity: convex problems A local minimizer is a global minimizer A solution of f (x) = 0 (stationary point) is a minimizer A global
More informationQuadratic reformulation techniques for 0-1 quadratic programs
OSE SEMINAR 2014 Quadratic reformulation techniques for 0-1 quadratic programs Ray Pörn CENTER OF EXCELLENCE IN OPTIMIZATION AND SYSTEMS ENGINEERING ÅBO AKADEMI UNIVERSITY ÅBO NOVEMBER 14th 2014 2 Structure
More informationAcyclic Semidefinite Approximations of Quadratically Constrained Quadratic Programs
Acyclic Semidefinite Approximations of Quadratically Constrained Quadratic Programs Raphael Louca & Eilyan Bitar School of Electrical and Computer Engineering American Control Conference (ACC) Chicago,
More informationLINEAR SYSTEMS (11) Intensive Computation
LINEAR SYSTEMS () Intensive Computation 27-8 prof. Annalisa Massini Viviana Arrigoni EXACT METHODS:. GAUSSIAN ELIMINATION. 2. CHOLESKY DECOMPOSITION. ITERATIVE METHODS:. JACOBI. 2. GAUSS-SEIDEL 2 CHOLESKY
More informationMath 273a: Optimization Overview of First-Order Optimization Algorithms
Math 273a: Optimization Overview of First-Order Optimization Algorithms Wotao Yin Department of Mathematics, UCLA online discussions on piazza.com 1 / 9 Typical flow of numerical optimization Optimization
More informationLine Search Methods for Unconstrained Optimisation
Line Search Methods for Unconstrained Optimisation Lecture 8, Numerical Linear Algebra and Optimisation Oxford University Computing Laboratory, MT 2007 Dr Raphael Hauser (hauser@comlab.ox.ac.uk) The Generic
More informationApproximation Algorithms
Approximation Algorithms Chapter 26 Semidefinite Programming Zacharias Pitouras 1 Introduction LP place a good lower bound on OPT for NP-hard problems Are there other ways of doing this? Vector programs
More information, b = 0. (2) 1 2 The eigenvectors of A corresponding to the eigenvalues λ 1 = 1, λ 2 = 3 are
Quadratic forms We consider the quadratic function f : R 2 R defined by f(x) = 2 xt Ax b T x with x = (x, x 2 ) T, () where A R 2 2 is symmetric and b R 2. We will see that, depending on the eigenvalues
More information1 Non-negative Matrix Factorization (NMF)
2018-06-21 1 Non-negative Matrix Factorization NMF) In the last lecture, we considered low rank approximations to data matrices. We started with the optimal rank k approximation to A R m n via the SVD,
More informationPreliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012
Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.
More informationBindel, Fall 2009 Matrix Computations (CS 6210) Week 8: Friday, Oct 17
Logistics Week 8: Friday, Oct 17 1. HW 3 errata: in Problem 1, I meant to say p i < i, not that p i is strictly ascending my apologies. You would want p i > i if you were simply forming the matrices and
More information1 Newton s Method. Suppose we want to solve: x R. At x = x, f (x) can be approximated by:
Newton s Method Suppose we want to solve: (P:) min f (x) At x = x, f (x) can be approximated by: n x R. f (x) h(x) := f ( x)+ f ( x) T (x x)+ (x x) t H ( x)(x x), 2 which is the quadratic Taylor expansion
More informationInfeasible Primal-Dual (Path-Following) Interior-Point Methods for Semidefinite Programming*
Infeasible Primal-Dual (Path-Following) Interior-Point Methods for Semidefinite Programming* Yin Zhang Dept of CAAM, Rice University Outline (1) Introduction (2) Formulation & a complexity theorem (3)
More informationRiemannian Optimization and its Application to Phase Retrieval Problem
Riemannian Optimization and its Application to Phase Retrieval Problem 1/46 Introduction Motivations Optimization History Phase Retrieval Problem Summary Joint work with: Pierre-Antoine Absil, Professor
More informationPENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP p.1/39
PENNON A Generalized Augmented Lagrangian Method for Convex NLP and SDP Michal Kočvara Institute of Information Theory and Automation Academy of Sciences of the Czech Republic and Czech Technical University
More informationFast and Robust Phase Retrieval
Fast and Robust Phase Retrieval Aditya Viswanathan aditya@math.msu.edu CCAM Lunch Seminar Purdue University April 18 2014 0 / 27 Joint work with Yang Wang Mark Iwen Research supported in part by National
More informationSparse Solutions of an Undetermined Linear System
1 Sparse Solutions of an Undetermined Linear System Maddullah Almerdasy New York University Tandon School of Engineering arxiv:1702.07096v1 [math.oc] 23 Feb 2017 Abstract This work proposes a research
More informationXLVI Pesquisa Operacional na Gestão da Segurança Pública
A linear formulation with O(n 2 ) variables for the quadratic assignment problem Serigne Gueye and Philippe Michelon Université d Avignon et des Pays du Vaucluse, Laboratoire d Informatique d Avignon (LIA),
More informationFINITE-DIMENSIONAL LINEAR ALGEBRA
DISCRETE MATHEMATICS AND ITS APPLICATIONS Series Editor KENNETH H ROSEN FINITE-DIMENSIONAL LINEAR ALGEBRA Mark S Gockenbach Michigan Technological University Houghton, USA CRC Press Taylor & Francis Croup
More informationOutline Introduction: Problem Description Diculties Algebraic Structure: Algebraic Varieties Rank Decient Toeplitz Matrices Constructing Lower Rank St
Structured Lower Rank Approximation by Moody T. Chu (NCSU) joint with Robert E. Funderlic (NCSU) and Robert J. Plemmons (Wake Forest) March 5, 1998 Outline Introduction: Problem Description Diculties Algebraic
More informationLinear Algebra Massoud Malek
CSUEB Linear Algebra Massoud Malek Inner Product and Normed Space In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n An inner product
More informationSolving linear equations with Gaussian Elimination (I)
Term Projects Solving linear equations with Gaussian Elimination The QR Algorithm for Symmetric Eigenvalue Problem The QR Algorithm for The SVD Quasi-Newton Methods Solving linear equations with Gaussian
More informationA solution approach for linear optimization with completely positive matrices
A solution approach for linear optimization with completely positive matrices Franz Rendl http://www.math.uni-klu.ac.at Alpen-Adria-Universität Klagenfurt Austria joint work with M. Bomze (Wien) and F.
More informationSparse Covariance Selection using Semidefinite Programming
Sparse Covariance Selection using Semidefinite Programming A. d Aspremont ORFE, Princeton University Joint work with O. Banerjee, L. El Ghaoui & G. Natsoulis, U.C. Berkeley & Iconix Pharmaceuticals Support
More informationELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications
ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications Professor M. Chiang Electrical Engineering Department, Princeton University March
More informationProjection methods to solve SDP
Projection methods to solve SDP Franz Rendl http://www.math.uni-klu.ac.at Alpen-Adria-Universität Klagenfurt Austria F. Rendl, Oberwolfach Seminar, May 2010 p.1/32 Overview Augmented Primal-Dual Method
More information1 Computing with constraints
Notes for 2017-04-26 1 Computing with constraints Recall that our basic problem is minimize φ(x) s.t. x Ω where the feasible set Ω is defined by equality and inequality conditions Ω = {x R n : c i (x)
More informationLecture 6 Positive Definite Matrices
Linear Algebra Lecture 6 Positive Definite Matrices Prof. Chun-Hung Liu Dept. of Electrical and Computer Engineering National Chiao Tung University Spring 2017 2017/6/8 Lecture 6: Positive Definite Matrices
More informationSparse Optimization Lecture: Dual Methods, Part I
Sparse Optimization Lecture: Dual Methods, Part I Instructor: Wotao Yin July 2013 online discussions on piazza.com Those who complete this lecture will know dual (sub)gradient iteration augmented l 1 iteration
More informationLinear algebra issues in Interior Point methods for bound-constrained least-squares problems
Linear algebra issues in Interior Point methods for bound-constrained least-squares problems Stefania Bellavia Dipartimento di Energetica S. Stecco Università degli Studi di Firenze Joint work with Jacek
More informationAn Iterative Descent Method
Conjugate Gradient: An Iterative Descent Method The Plan Review Iterative Descent Conjugate Gradient Review : Iterative Descent Iterative Descent is an unconstrained optimization process x (k+1) = x (k)
More informationA NEW FIRST-ORDER ALGORITHMIC FRAMEWORK FOR OPTIMIZATION PROBLEMS WITH ORTHOGONALITY CONSTRAINTS. min
A NEW FIRST-ORDER ALGORITHMIC FRAMEWORK FOR OPTIMIZATION PROBLEMS WITH ORTHOGONALITY CONSTRAINTS BIN GAO, XIN LIU, XIAOJUN CHEN, AND YA-XIANG YUAN Abstract. In this paper, we consider a class of optimization
More informationApproximation algorithms for nonnegative polynomial optimization problems over unit spheres
Front. Math. China 2017, 12(6): 1409 1426 https://doi.org/10.1007/s11464-017-0644-1 Approximation algorithms for nonnegative polynomial optimization problems over unit spheres Xinzhen ZHANG 1, Guanglu
More informationLarge-Scale L1-Related Minimization in Compressive Sensing and Beyond
Large-Scale L1-Related Minimization in Compressive Sensing and Beyond Yin Zhang Department of Computational and Applied Mathematics Rice University, Houston, Texas, U.S.A. Arizona State University March
More informationDetermination of Feasible Directions by Successive Quadratic Programming and Zoutendijk Algorithms: A Comparative Study
International Journal of Mathematics And Its Applications Vol.2 No.4 (2014), pp.47-56. ISSN: 2347-1557(online) Determination of Feasible Directions by Successive Quadratic Programming and Zoutendijk Algorithms:
More informationminimize x subject to (x 2)(x 4) u,
Math 6366/6367: Optimization and Variational Methods Sample Preliminary Exam Questions 1. Suppose that f : [, L] R is a C 2 -function with f () on (, L) and that you have explicit formulae for
More informationBlock Coordinate Descent for Regularized Multi-convex Optimization
Block Coordinate Descent for Regularized Multi-convex Optimization Yangyang Xu and Wotao Yin CAAM Department, Rice University February 15, 2013 Multi-convex optimization Model definition Applications Outline
More informationRiemannian Optimization Method on the Flag Manifold for Independent Subspace Analysis
Riemannian Optimization Method on the Flag Manifold for Independent Subspace Analysis Yasunori Nishimori 1, Shotaro Akaho 1, and Mark D. Plumbley 2 1 Neuroscience Research Institute, National Institute
More information5 Handling Constraints
5 Handling Constraints Engineering design optimization problems are very rarely unconstrained. Moreover, the constraints that appear in these problems are typically nonlinear. This motivates our interest
More informationLecture: Algorithms for LP, SOCP and SDP
1/53 Lecture: Algorithms for LP, SOCP and SDP Zaiwen Wen Beijing International Center For Mathematical Research Peking University http://bicmr.pku.edu.cn/~wenzw/bigdata2018.html wenzw@pku.edu.cn Acknowledgement:
More informationUnconstrained Optimization Models for Computing Several Extreme Eigenpairs of Real Symmetric Matrices
Unconstrained Optimization Models for Computing Several Extreme Eigenpairs of Real Symmetric Matrices Yu-Hong Dai LSEC, ICMSEC Academy of Mathematics and Systems Science Chinese Academy of Sciences Joint
More informationNumerical Methods I Non-Square and Sparse Linear Systems
Numerical Methods I Non-Square and Sparse Linear Systems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 September 25th, 2014 A. Donev (Courant
More informationThe nonsmooth Newton method on Riemannian manifolds
The nonsmooth Newton method on Riemannian manifolds C. Lageman, U. Helmke, J.H. Manton 1 Introduction Solving nonlinear equations in Euclidean space is a frequently occurring problem in optimization and
More informationA Quick Tour of Linear Algebra and Optimization for Machine Learning
A Quick Tour of Linear Algebra and Optimization for Machine Learning Masoud Farivar January 8, 2015 1 / 28 Outline of Part I: Review of Basic Linear Algebra Matrices and Vectors Matrix Multiplication Operators
More informationA direct formulation for sparse PCA using semidefinite programming
A direct formulation for sparse PCA using semidefinite programming A. d Aspremont, L. El Ghaoui, M. Jordan, G. Lanckriet ORFE, Princeton University & EECS, U.C. Berkeley Available online at www.princeton.edu/~aspremon
More informationPDE-Constrained and Nonsmooth Optimization
Frank E. Curtis October 1, 2009 Outline PDE-Constrained Optimization Introduction Newton s method Inexactness Results Summary and future work Nonsmooth Optimization Sequential quadratic programming (SQP)
More informationPreliminary Examination, Numerical Analysis, August 2016
Preliminary Examination, Numerical Analysis, August 2016 Instructions: This exam is closed books and notes. The time allowed is three hours and you need to work on any three out of questions 1-4 and any
More informationAn Inexact Newton Method for Nonlinear Constrained Optimization
An Inexact Newton Method for Nonlinear Constrained Optimization Frank E. Curtis Numerical Analysis Seminar, January 23, 2009 Outline Motivation and background Algorithm development and theoretical results
More informationAn Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization
An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization Frank E. Curtis, Lehigh University involving joint work with Travis Johnson, Northwestern University Daniel P. Robinson, Johns
More informationPart 4: Active-set methods for linearly constrained optimization. Nick Gould (RAL)
Part 4: Active-set methods for linearly constrained optimization Nick Gould RAL fx subject to Ax b Part C course on continuoue optimization LINEARLY CONSTRAINED MINIMIZATION fx subject to Ax { } b where
More informationA Customized ADMM for Rank-Constrained Optimization Problems with Approximate Formulations
A Customized ADMM for Rank-Constrained Optimization Problems with Approximate Formulations Chuangchuang Sun and Ran Dai Abstract This paper proposes a customized Alternating Direction Method of Multipliers
More informationCONSTRAINED NONLINEAR PROGRAMMING
149 CONSTRAINED NONLINEAR PROGRAMMING We now turn to methods for general constrained nonlinear programming. These may be broadly classified into two categories: 1. TRANSFORMATION METHODS: In this approach
More informationConditional Gradient Algorithms for Rank-One Matrix Approximations with a Sparsity Constraint
Conditional Gradient Algorithms for Rank-One Matrix Approximations with a Sparsity Constraint Marc Teboulle School of Mathematical Sciences Tel Aviv University Joint work with Ronny Luss Optimization and
More informationAlternating Direction Augmented Lagrangian Algorithms for Convex Optimization
Alternating Direction Augmented Lagrangian Algorithms for Convex Optimization Joint work with Bo Huang, Shiqian Ma, Tony Qin, Katya Scheinberg, Zaiwen Wen and Wotao Yin Department of IEOR, Columbia University
More informationBasic Math for
Basic Math for 16-720 August 23, 2002 1 Linear Algebra 1.1 Vectors and Matrices First, a reminder of a few basic notations, definitions, and terminology: Unless indicated otherwise, vectors are always
More informationA Note on Semidefinite Programming Relaxations For Polynomial Optimization Over a Single Sphere
. ARTICLES. doi: A Note on Semidefinite Programming Relaxations For Polynomial Optimization Over a Single Sphere Jiang Hu,, Bo jiang, Xin Liu 3 & Zaiwen Wen Beijing International Center for Mathematical
More informationThe Q Method for Symmetric Cone Programmin
The Q Method for Symmetric Cone Programming The Q Method for Symmetric Cone Programmin Farid Alizadeh and Yu Xia alizadeh@rutcor.rutgers.edu, xiay@optlab.mcma Large Scale Nonlinear and Semidefinite Progra
More informationOptimisation on Manifolds
Optimisation on Manifolds K. Hüper MPI Tübingen & Univ. Würzburg K. Hüper (MPI Tübingen & Univ. Würzburg) Applications in Computer Vision Grenoble 18/9/08 1 / 29 Contents 2 Examples Essential matrix estimation
More informationECE580 Exam 1 October 4, Please do not write on the back of the exam pages. Extra paper is available from the instructor.
ECE580 Exam 1 October 4, 2012 1 Name: Solution Score: /100 You must show ALL of your work for full credit. This exam is closed-book. Calculators may NOT be used. Please leave fractions as fractions, etc.
More informationSDPNAL+: A Majorized Semismooth Newton-CG Augmented Lagrangian Method for Semidefinite Programming with Nonnegative Constraints
SDPNAL+: A Majorized Semismooth Newton-CG Augmented Lagrangian Method for Semidefinite Programming with Nonnegative Constraints Liuqin Yang, Defeng Sun, Kim-Chuan Toh May 28, 2014; Revised on Mar 10, 2015
More informationSDPNAL+: A Majorized Semismooth Newton-CG Augmented Lagrangian Method for Semidefinite Programming with Nonnegative Constraints
arxiv:1406.0942v1 [math.oc] 4 Jun 2014 SDPNAL+: A Majorized Semismooth Newton-CG Augmented Lagrangian Method for Semidefinite Programming with Nonnegative Constraints Liuqin Yang, Defeng Sun, Kim-Chuan
More informationAn Active Set Strategy for Solving Optimization Problems with up to 200,000,000 Nonlinear Constraints
An Active Set Strategy for Solving Optimization Problems with up to 200,000,000 Nonlinear Constraints Klaus Schittkowski Department of Computer Science, University of Bayreuth 95440 Bayreuth, Germany e-mail:
More informationCHAPTER 11. A Revision. 1. The Computers and Numbers therein
CHAPTER A Revision. The Computers and Numbers therein Traditional computer science begins with a finite alphabet. By stringing elements of the alphabet one after another, one obtains strings. A set of
More informationProximal Gradient Descent and Acceleration. Ryan Tibshirani Convex Optimization /36-725
Proximal Gradient Descent and Acceleration Ryan Tibshirani Convex Optimization 10-725/36-725 Last time: subgradient method Consider the problem min f(x) with f convex, and dom(f) = R n. Subgradient method:
More informationLast Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection
Eigenvalue Problems Last Time Social Network Graphs Betweenness Girvan-Newman Algorithm Graph Laplacian Spectral Bisection λ 2, w 2 Today Small deviation into eigenvalue problems Formulation Standard eigenvalue
More informationComputation. For QDA we need to calculate: Lets first consider the case that
Computation For QDA we need to calculate: δ (x) = 1 2 log( Σ ) 1 2 (x µ ) Σ 1 (x µ ) + log(π ) Lets first consider the case that Σ = I,. This is the case where each distribution is spherical, around the
More informationSTAT 309: MATHEMATICAL COMPUTATIONS I FALL 2013 PROBLEM SET 2
STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2013 PROBLEM SET 2 1. You are not allowed to use the svd for this problem, i.e. no arguments should depend on the svd of A or A. Let W be a subspace of C n. The
More informationRobust PCA. CS5240 Theoretical Foundations in Multimedia. Leow Wee Kheng
Robust PCA CS5240 Theoretical Foundations in Multimedia Leow Wee Kheng Department of Computer Science School of Computing National University of Singapore Leow Wee Kheng (NUS) Robust PCA 1 / 52 Previously...
More informationCopositive Programming and Combinatorial Optimization
Copositive Programming and Combinatorial Optimization Franz Rendl http://www.math.uni-klu.ac.at Alpen-Adria-Universität Klagenfurt Austria joint work with M. Bomze (Wien) and F. Jarre (Düsseldorf) and
More informationConvex Optimization. Newton s method. ENSAE: Optimisation 1/44
Convex Optimization Newton s method ENSAE: Optimisation 1/44 Unconstrained minimization minimize f(x) f convex, twice continuously differentiable (hence dom f open) we assume optimal value p = inf x f(x)
More informationLecture 8: February 9
0-725/36-725: Convex Optimiation Spring 205 Lecturer: Ryan Tibshirani Lecture 8: February 9 Scribes: Kartikeya Bhardwaj, Sangwon Hyun, Irina Caan 8 Proximal Gradient Descent In the previous lecture, we
More informationNewton s Method and Linear Approximations
Newton s Method and Linear Approximations Curves are tricky. Lines aren t. Newton s Method and Linear Approximations Newton s Method for finding roots Goal: Where is f (x) = 0? f (x) = x 7 + 3x 3 + 7x
More informationPart 5: Penalty and augmented Lagrangian methods for equality constrained optimization. Nick Gould (RAL)
Part 5: Penalty and augmented Lagrangian methods for equality constrained optimization Nick Gould (RAL) x IR n f(x) subject to c(x) = Part C course on continuoue optimization CONSTRAINED MINIMIZATION x
More informationPreface to Second Edition... vii. Preface to First Edition...
Contents Preface to Second Edition..................................... vii Preface to First Edition....................................... ix Part I Linear Algebra 1 Basic Vector/Matrix Structure and
More informationMinimizing the Difference of L 1 and L 2 Norms with Applications
1/36 Minimizing the Difference of L 1 and L 2 Norms with Department of Mathematical Sciences University of Texas Dallas May 31, 2017 Partially supported by NSF DMS 1522786 2/36 Outline 1 A nonconvex approach:
More informationONP-MF: An Orthogonal Nonnegative Matrix Factorization Algorithm with Application to Clustering
ONP-MF: An Orthogonal Nonnegative Matrix Factorization Algorithm with Application to Clustering Filippo Pompili 1, Nicolas Gillis 2, P.-A. Absil 2,andFrançois Glineur 2,3 1- University of Perugia, Department
More informationOptimization. Yuh-Jye Lee. March 21, Data Science and Machine Intelligence Lab National Chiao Tung University 1 / 29
Optimization Yuh-Jye Lee Data Science and Machine Intelligence Lab National Chiao Tung University March 21, 2017 1 / 29 You Have Learned (Unconstrained) Optimization in Your High School Let f (x) = ax
More informationA direct formulation for sparse PCA using semidefinite programming
A direct formulation for sparse PCA using semidefinite programming A. d Aspremont, L. El Ghaoui, M. Jordan, G. Lanckriet ORFE, Princeton University & EECS, U.C. Berkeley A. d Aspremont, INFORMS, Denver,
More informationNon-negative Matrix Factorization via accelerated Projected Gradient Descent
Non-negative Matrix Factorization via accelerated Projected Gradient Descent Andersen Ang Mathématique et recherche opérationnelle UMONS, Belgium Email: manshun.ang@umons.ac.be Homepage: angms.science
More informationNonlinear Optimization for Optimal Control
Nonlinear Optimization for Optimal Control Pieter Abbeel UC Berkeley EECS Many slides and figures adapted from Stephen Boyd [optional] Boyd and Vandenberghe, Convex Optimization, Chapters 9 11 [optional]
More informationNumerical Methods in Matrix Computations
Ake Bjorck Numerical Methods in Matrix Computations Springer Contents 1 Direct Methods for Linear Systems 1 1.1 Elements of Matrix Theory 1 1.1.1 Matrix Algebra 2 1.1.2 Vector Spaces 6 1.1.3 Submatrices
More information