Dimension reduction for semidefinite programming

Similar documents
MIT Algebraic techniques and semidefinite optimization May 9, Lecture 21. Lecturer: Pablo A. Parrilo Scribe:???

Reduction methods in semidefinite and conic optimization

Applications of semidefinite programming in Algebraic Combinatorics

Convex sets, conic matrix factorizations and conic rank lower bounds

DSOS/SDOS Programming: New Tools for Optimization over Nonnegative Polynomials

Fast ADMM for Sum of Squares Programs Using Partial Orthogonality

Equivariant semidefinite lifts and sum of squares hierarchies

Semidefinite Programming Basics and Applications

Semidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 5

CSCI 1951-G Optimization Methods in Finance Part 10: Conic Optimization

Iterative LP and SOCP-based. approximations to. sum of squares programs. Georgina Hall Princeton University. Joint work with:

Robust and Optimal Control, Spring 2015

Semidefinite programming and convex algebraic geometry

Convex Optimization of Graph Laplacian Eigenvalues

A Numerical Algorithm for Block-Diagonal Decomposition of Matrix -Algebras, Part I: Proposed Approach and Application to Semidefinite Programming

The convex algebraic geometry of rank minimization

Facial Reduction and Geometry on Conic Programming

1.4 Solvable Lie algebras

Real Symmetric Matrices and Semidefinite Programming

Convex Optimization of Graph Laplacian Eigenvalues

Facial reduction for Euclidean distance matrix problems

A NUMERICAL ALGORITHM FOR BLOCK-DIAGONAL DECOMPOSITION OF MATRIX -ALGEBRAS

Positive semidefinite rank

V&V MURI Overview Caltech, October 2008

REPRESENTATION THEORY WEEK 5. B : V V k

Semidefinite programming lifts and sparse sums-of-squares

Problems in Linear Algebra and Representation Theory

Introduction to Association Schemes

Sta$s$cal models and graphs

Semidefinite Programming and Harmonic Analysis

CSC Linear Programming and Combinatorial Optimization Lecture 10: Semidefinite Programming

An LMI description for the cone of Lorentz-positive maps II

SEMIDEFINITE PROGRAM BASICS. Contents

Symmetric Sums of Squares

Lift-and-Project Techniques and SDP Hierarchies

A notion of Total Dual Integrality for Convex, Semidefinite and Extended Formulations

A STRUCTURAL GEOMETRICAL ANALYSIS OF WEAKLY INFEASIBLE SDPS

REPRESENTATIONS OF U(N) CLASSIFICATION BY HIGHEST WEIGHTS NOTES FOR MATH 261, FALL 2001

Lecture: Examples of LP, SOCP and SDP

Lecture 5 : Projections

Notes on nilpotent orbits Computational Theory of Real Reductive Groups Workshop. Eric Sommers

INTERIOR-POINT METHODS ROBERT J. VANDERBEI JOINT WORK WITH H. YURTTAN BENSON REAL-WORLD EXAMPLES BY J.O. COLEMAN, NAVAL RESEARCH LAB

Relaxations of combinatorial problems via association schemes

Linear Algebra problems

SDP Relaxations for MAXCUT

arxiv: v1 [math.oc] 26 Sep 2015

Non-commutative polynomial optimization

Math Linear Algebra II. 1. Inner Products and Norms

Acyclic Semidefinite Approximations of Quadratically Constrained Quadratic Programs

Lower bounds on the size of semidefinite relaxations. David Steurer Cornell

Graph Partitioning Using Random Walks

The Graph Realization Problem

JOHNSON SCHEMES AND CERTAIN MATRICES WITH INTEGRAL EIGENVALUES

Complexity of Deciding Convexity in Polynomial Optimization

Analysis of Control Systems on Symmetric Cones

L(C G (x) 0 ) c g (x). Proof. Recall C G (x) = {g G xgx 1 = g} and c g (x) = {X g Ad xx = X}. In general, it is obvious that

arxiv: v4 [math.oc] 12 Apr 2017

SDP and eigenvalue bounds for the graph partition problem

Commuting birth-and-death processes

The maximal stable set problem : Copositive programming and Semidefinite Relaxations

LINEAR ALGEBRA BOOT CAMP WEEK 1: THE BASICS

A CONIC DANTZIG-WOLFE DECOMPOSITION APPROACH FOR LARGE SCALE SEMIDEFINITE PROGRAMMING

Uniqueness of the Solutions of Some Completion Problems

Optimization over Polynomials with Sums of Squares and Moment Matrices

Chordal structure in computer algebra: Permanents

COURSE ON LMI PART I.2 GEOMETRY OF LMI SETS. Didier HENRION henrion

Introduction to Semidefinite Programming I: Basic properties a

There are six more problems on the next two pages

approximation algorithms I

Mathematical Optimisation, Chpt 2: Linear Equations and inequalities

Semidefinite programming and eigenvalue bounds for the graph partition problem

Degeneracy in Maximal Clique Decomposition for Semidefinite Programs

Semidefinite Programming

Elementary linear algebra

Analysis and synthesis: a complexity perspective

Block-Diagonal Decomposition of Matrices based on -Algebra for Structural Optimization with Symmetry Property

Decentralized Control of Stochastic Systems

Network Localization via Schatten Quasi-Norm Minimization

LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM

Control System Analysis on Symmetric Cones

Copositive Programming and Combinatorial Optimization

LMI MODELLING 4. CONVEX LMI MODELLING. Didier HENRION. LAAS-CNRS Toulouse, FR Czech Tech Univ Prague, CZ. Universidad de Valladolid, SP March 2009

Bare-bones outline of eigenvalue theory and the Jordan canonical form

Sparse Matrix Theory and Semidefinite Optimization

A proof of the Jordan normal form theorem

The Trust Region Subproblem with Non-Intersecting Linear Constraints

Convex sets, matrix factorizations and positive semidefinite rank

Representations of Positive Polynomials: Theory, Practice, and

Signatures of GL n Multiplicity Spaces

Relaxations and Randomized Methods for Nonconvex QCQPs

On bounding the bandwidth of graphs with symmetry

Matrix Mathematics. Theory, Facts, and Formulas with Application to Linear Systems Theory. Dennis S. Bernstein

Numerical block diagonalization of matrix -algebras with application to semidefinite programming

Approximation Algorithms

Lecture: Introduction to LP, SDP and SOCP

Compression, Matrix Range and Completely Positive Map

The value of a problem is not so much coming up with the answer as in the ideas and attempted ideas it forces on the would be solver I.N.

Equations in simple algebras: variations on a theme of A. Borel. Boris Kunyavskĭı Bar-Ilan University RAGE conference Emory University, May 18, 2011

Problem Set (T) If A is an m n matrix, B is an n p matrix and D is a p s matrix, then show

Background on Chevalley Groups Constructed from a Root System

Transcription:

1 / 22 Dimension reduction for semidefinite programming Pablo A. Parrilo Laboratory for Information and Decision Systems Electrical Engineering and Computer Science Massachusetts Institute of Technology Joint work with Frank Permenter (MIT) arxiv:1608.02090 CDC 2016 - Las Vegas

2 / 22 Semidefinite programs (SDPs) minimize Tr CX subject to X A S n + Formulated over vector space S n of n n symmetric matrices. variable X S n A S n an affine subspace, C S n cost matrix S n + cone of psd matrices Efficiently solvable in theory; in practice, solving some instances impossible unless special structure is exploited.

Dimension reduction Reformulate problem over subspace S S n intersecting set of optimal solutions minimize Tr CX subject to X A S n + minimize subject to Tr CX X A S n + S (Reformulation) S A S n + opt. solns where S n + S equals product K i K m of simple cones. Reduction methods: symmetry reduction and facial reduction 3 / 22

4 / 22 Symmetry reduction (MAXCUT relaxation example ) minimize Tr CX subject to X A S n + A := {X S n : X ii = 1} C := adjacency matrix

Symmetry reduction (MAXCUT relaxation example ) minimize Tr CX subject to X A S n + A := {X S n : X ii = 1} C := adjacency matrix range P A S n + opt. solns (e.g., Schrijver 79; Gatermann-P. 03) Idea: find special projection map P P(X) optimal when X optimal. P explicitly constructed from automorphism group of graph. Range block-diagonal a direct-sum of matrix algebras. 4 / 22

5 / 22 Facial reduction H S n + A minimize Tr CX subject to X A S n + First, find face of S n + containing feasible set. There exists a hyperplane H containing A. S n + H a face isomorphic to S d + for d < n. Face S n + H contains feasible set A S n +.

5 / 22 Facial reduction H S n + A minimize Tr CX subject to X A S n + First, find face of S n + containing feasible set. There exists a hyperplane H containing A. S n + H a face isomorphic to S d + for d < n. Face S n + H contains feasible set A S n +. Next, reformulate SDP over face: minimize subject to Tr CX X A S n + H

Facial reduction H S n + A minimize Tr CX subject to X A S n + First, find face of S n + containing feasible set. There exists a hyperplane H containing A. S n + H a face isomorphic to S d + for d < n. Face S n + H contains feasible set A S n +. Next, reformulate SDP over face: minimize subject to Tr CX X A S n + H Borwein-Wolkowicz 81; Pataki 00; Permenter-P. 14 5 / 22

6 / 22 Application specific approaches Facial reduction: MAXCUT (Anjos, Wolkowicz) QAP (Zhao,Wolkowicz) Sums-of-squares optimization (Permenter-P., Waki-Muramatsu) Matrix completion (Krislock,Wolkowicz)... Symmetry reduction: MAXCUT (earlier example), QAP (de Klerk, Sotirov); Markov chains (Boyd et al.); codes (Schrijver; Laurent)...

7 / 22 Our approach This talk: new reduction method subsuming symmetry reduction Notion of optimal reductions. A general purpose algorithm with optimality guarantees Jordan algebra interpretation; hence, easy extension to symmetric cone optimization (e.g., LP, SOCP). Combinatorial refinements for computational efficiency

8 / 22 How does symmetry reduction work? Given SDP min X A S n + Tr CX, method finds special orthogonal projection P : S n S n range of P A S n + opt. solns If X feas./optimal, P(X) feas./optimal.

8 / 22 How does symmetry reduction work? Given SDP min X A S n + Tr CX, method finds special orthogonal projection P : S n S n range of P A S n + opt. solns If X feas./optimal, P(X) feas./optimal. P satisfies following conditions: P(A) A, P(S n +) S n +, P(C) = C

How does symmetry reduction work? Given SDP min X A S n + Tr CX, method finds special orthogonal projection P : S n S n range of P A S n + opt. solns If X feas./optimal, P(X) feas./optimal. P satisfies following conditions: P(A) A, P(S n +) S n +, P(C) = C Hence, if X feasible then P(X) feasible with equal cost: 8 / 22

9 / 22 Example: a MAXCUT SDP relaxation minimize Tr CX subject to X A S n + A := {X S n : X ii = 1} C := adjacency matrix

9 / 22 Example: a MAXCUT SDP relaxation minimize Tr CX subject to X A S n + A := {X S n : X ii = 1} C := adjacency matrix

9 / 22 Example: a MAXCUT SDP relaxation minimize Tr CX subject to X A S n + A := {X S n : X ii = 1} C := adjacency matrix Let G denote group of permutation matrices (automorphisms) G := {U a permutation matrix : U T CU = C}

9 / 22 Example: a MAXCUT SDP relaxation minimize Tr CX subject to X A S n + A := {X S n : X ii = 1} C := adjacency matrix Let G denote group of permutation matrices (automorphisms) G := {U a permutation matrix : U T CU = C} Taking P(X) := 1 G U G UT XU, desired conditions hold: P(S n +) S n + P(A) A, P(C) = C

9 / 22 Example: a MAXCUT SDP relaxation minimize Tr CX subject to X A S n + A := {X S n : X ii = 1} C := adjacency matrix Let G denote group of permutation matrices (automorphisms) G := {U a permutation matrix : U T CU = C} Taking P(X) := 1 G U G UT XU, desired conditions hold: P(S n +) S n + P(A) A, P(C) = C Hence, range of P contains solutions: when X feasible, P(X) feasible with equal cost.

10 / 22 Our approach: optimize over projections Given SDP min X A S n + C, X, find map P that solves minimize subject to rank P P(C) = C, P(I) = I P(A) A P(S n +) S n + P : S n S n an orthogonal projection.

10 / 22 Our approach: optimize over projections Given SDP min X A S n + C, X, find map P that solves minimize subject to rank P P(C) = C, P(I) = I P(A) A P(S n +) S n + P : S n S n an orthogonal projection. Main properties: Can be solved in polynomial time (!) Range of P structured: a Jordan subalgebra of S n. S n + range P equals a product of symmetric cones.

11 / 22 Invariance characterization of optimal subspace Theorem (Permenter-P.) Orthogonal projection P : S n S n solves minimize subject to rank P P(C) = C, P(I) = I P(A) A P(S n +) S n + iff the range of P solves minimize dim S subject to S {I, X L, C} S P L (S) S {X 2 : X S}, where A = X L + L, and X L is the min-norm point of A.

12 / 22 Subspace optimization and solution algorithm minimize subject to dim S S C, X L, I S P L (S) S {X 2 : X S} S span{c, X L, I} repeat S S + P L (S) S S + span{x 2 : X S} until converged.

12 / 22 Subspace optimization and solution algorithm minimize subject to dim S S C, X L, I S P L (S) S {X 2 : X S} S span{c, X L, I} repeat S S + P L (S) S S + span{x 2 : X S} until converged. Properties of algorithm: Optimal subspace contains each iterate (induction)

12 / 22 Subspace optimization and solution algorithm minimize subject to dim S S C, X L, I S P L (S) S {X 2 : X S} S span{c, X L, I} repeat S S + P L (S) S S + span{x 2 : X S} until converged. Properties of algorithm: Optimal subspace contains each iterate (induction) Computes ascending chain of subspaces terminates.

12 / 22 Subspace optimization and solution algorithm minimize subject to dim S S C, X L, I S P L (S) S {X 2 : X S} S span{c, X L, I} repeat S S + P L (S) S S + span{x 2 : X S} until converged. Properties of algorithm: Optimal subspace contains each iterate (induction) Computes ascending chain of subspaces terminates. At termination, subspace feasible; hence, optimal.

12 / 22 Subspace optimization and solution algorithm minimize subject to dim S S C, X L, I S P L (S) S {X 2 : X S} S span{c, X L, I} repeat S S + P L (S) S S + span{x 2 : X S} until converged. Properties of algorithm: Optimal subspace contains each iterate (induction) Computes ascending chain of subspaces terminates. At termination, subspace feasible; hence, optimal. Properties of minimization problem: Feasible set closed under intersection (lattice)

Subspace optimization and solution algorithm minimize subject to dim S S C, X L, I S P L (S) S {X 2 : X S} S span{c, X L, I} repeat S S + P L (S) S S + span{x 2 : X S} until converged. Properties of algorithm: Optimal subspace contains each iterate (induction) Computes ascending chain of subspaces terminates. At termination, subspace feasible; hence, optimal. Properties of minimization problem: Feasible set closed under intersection (lattice) A unique solution. 12 / 22

13 / 22 Combinatorial descriptions Great! Now, we can easily compute the optimal subspace S.

13 / 22 Combinatorial descriptions Great! Now, we can easily compute the optimal subspace S. But, often want/need additional properties (e.g., dense subspaces may not be very efficient). Can tradeoff dimension with sparsity of a basis?

13 / 22 Combinatorial descriptions Great! Now, we can easily compute the optimal subspace S. But, often want/need additional properties (e.g., dense subspaces may not be very efficient). Can tradeoff dimension with sparsity of a basis? Yes! Three kinds of sparse bases for S: E.g., Partition subspaces: defined by a partition of [n] [n]. Coordinate subspaces: defined by a sparsity pattern Combinatorial subspaces: orthogonal basis of 0/1 matrices a a b a a b b b c vs. a b 0 b c 0 0 0 d vs. a 0 b 0 a c b c b

14 / 22 Finding optimal structured subspaces The main algorithm can be adapted to compute the optimal subspace for each of these three cases.

14 / 22 Finding optimal structured subspaces The main algorithm can be adapted to compute the optimal subspace for each of these three cases. Key property (again): lattice structure (closedness under intersection)

14 / 22 Finding optimal structured subspaces The main algorithm can be adapted to compute the optimal subspace for each of these three cases. Key property (again): lattice structure (closedness under intersection) E.g., for partition subspaces, instead of optimizing over lattice of subspaces, use the lattice of partitions: minimize subject to dim S S C, X L, I S P L (S) S {X 2 : X S} S is a partition subspace P Part{C, X L, I} repeat P refine(p, P L ) P refine(p, X X 2 ) until converged.

14 / 22 Finding optimal structured subspaces The main algorithm can be adapted to compute the optimal subspace for each of these three cases. Key property (again): lattice structure (closedness under intersection) E.g., for partition subspaces, instead of optimizing over lattice of subspaces, use the lattice of partitions: minimize subject to dim S S C, X L, I S P L (S) S {X 2 : X S} S is a partition subspace P Part{C, X L, I} repeat P refine(p, P L ) P refine(p, X X 2 ) until converged. Great! But there s more...

15 / 22 Decomposition via Jordan algebras Given SDP min X A S n + C, X, we ve found a subspace invariant under X X 2 containing optimal solutions: S A S n + opt. solns S {X 2 : X S}

15 / 22 Decomposition via Jordan algebras Given SDP min X A S n + C, X, we ve found a subspace invariant under X X 2 containing optimal solutions: S A S n + opt. solns S {X 2 : X S} Subspaces invariant under X X 2 have decomposition S 1 0... 0 0 S 2... 0 S = Q...... QT, 0 0 0 S m S i are simple Jordan algebras

Decomposition via Jordan algebras Given SDP min X A S n + C, X, we ve found a subspace invariant under X X 2 containing optimal solutions: S A S n + opt. solns S {X 2 : X S} Subspaces invariant under X X 2 have decomposition S 1 0... 0 0 S 2... 0 S = Q...... QT, 0 0 0 S m S i are simple Jordan algebras Number of distinct eigenvalues of generic element equals rank of S i a complexity measure. 15 / 22

16 / 22 Minimizing dimension optimizes decomposition minimize subject to dim S S X L, C, I S P L (S) S {X 2 : X S}, All feasible subspaces have decomp. S = d S i=1 S i. In what sense does solution S optimize the ranks of each S i?

16 / 22 Minimizing dimension optimizes decomposition minimize subject to dim S S X L, C, I S P L (S) S {X 2 : X S}, All feasible subspaces have decomp. S = d S i=1 S i. In what sense does solution S optimize the ranks of each S i? Thm. (Permenter-P.): S minimizes i rank S i and max i rank S i

Minimizing dimension optimizes decomposition minimize subject to dim S S X L, C, I S P L (S) S {X 2 : X S}, All feasible subspaces have decomp. S = d S i=1 S i. In what sense does solution S optimize the ranks of each S i? Thm. (Permenter-P.): S minimizes i rank S i and max i rank S i Majorization inequalities hold, i.e., for each m 1 m m rank Si rank S i i=1 (ranks sorted in decreasing order) i=1 16 / 22

17 / 22 Majorization example Subspaces (parametrized by u i and v i ) and their rank vectors u 1 u 2 0 0 0 u 2 u 3 0 0 0 0 0 u 4 0 0 0 0 0 u 5 u 6 0 0 0 u 6 u 7 v 1 v 2 0 0 0 v 2 v 3 0 0 0 0 0 v 4 v 5 v 6 0 0 v 5 v 7 v 8 0 0 v 6 v 8 v 9 r u = (2, 1, 2) r v = (2, 3)

17 / 22 Majorization example Subspaces (parametrized by u i and v i ) and their rank vectors u 1 u 2 0 0 0 u 2 u 3 0 0 0 0 0 u 4 0 0 0 0 0 u 5 u 6 0 0 0 u 6 u 7 v 1 v 2 0 0 0 v 2 v 3 0 0 0 0 0 v 4 v 5 v 6 0 0 v 5 v 7 v 8 0 0 v 6 v 8 v 9 r u = (2, 1, 2) r v = (2, 3) Vector r u = (2, 2, 1) majorized by r v = (3, 2, 0): 2 3, 2 + 2 3 + 2, 2 + 2 + 1 3 + 2 + 0

18 / 22 Jordan algebras Jordan algebras are commutative algebras satisfying Jordan identity (X Y ) X 2 = X (Y X 2 ) The vector space S n a Jordan algebra if equipped with product X Y := 1 (XY + YX) 2 The subalgebras of S n precisely the sets closed under squaring map X X 2 since XY + YX = (X + Y ) 2 X 2 Y 2. Structure theorem of Jordan-von Neumann-Wigner describes subalgebras of S n...

Decomposition of S S n + If S S n a Jordan subalgebra, it equals direct-sum m i=1 S i, where each S i is isomorphic to one of the following: Algebra of Hermitian matrices with real, complex or quaternion entries A spin-factor algebra Implies cone-of-squares S S n + isomorphic to product of PSD cones with real/complex/quaternion entries Lorentz cones Yields reformulation of original SDP over this product minimize Tr CX subject to X A S n + minimize Tr CX subject to X A T (K 1 K m ) }{{} S S n + 19 / 22

20 / 22 Computational results Comparison with reduction method of de Klerk 10 survey (generating *-algebras from data): instance S S data hamming_7_5_6 5 8256 hamming_8_3_4 5 32896 hamming_9_5_6 6 131328 hamming_9_8 6 131328 hamming_10_2 7 524800 Table list dimension of our subspace S S n and subspace S data S n found by generating *-algebra. Decomposing S yields a linear program.

Results: SOSOPT (Seiler 13) Demo scripts Script Name n (before) n (after) sosoptdemo2 13, 3 3, 2 3, 1 7 sosoptdemo4 35 5 5, 1 10 gsosoptdemo1 9, 5 6, 3 2, 2 IOGainDemo_3 15, 8 10, 5 2, 3 Chesi(1 2)_IterationWithVlin 9, 5 6, 3 2, 2 Chesi3_GlobalStability 14, 5 8, 6, 3, 2 Chesi(3 4)_IterationWithVlin 9, 5 6, 3 2, 2 Chesi(5 6)_Bootstrap 19, 9 13, 6 2, 3 Chesi(5 6)_IterationWithVlin 19, 9 13, 6 2, 3 Coutinho3_IterationWithVlin 9, 5 6, 3 2, 2 HachichoTibken_Bootstrap 19, 9 12, 7, 6, 3 HachichoTibken_IterationWithVlin 19, 9 12, 7, 6, 3 Hahn_IterationWithVlin 9, 5 6, 3, 3, 2 KuChen_IterationWithVlin 19, 9 13, 6 2, 3 Parrilo1_GlobalStabilityWithVec 3, 2 2, 1 3 Parrilo2_GlobalStabilityWithMat 3, 2 2, 1 3 VDP_IterationWithVball 5, 4 3 2, 2, 1 VDP_IterationWithVlin 9, 5 6, 3 2, 2 VDP_LinearizedLyap 9, 5 6, 3 2, 2 VannelliVidyasagar2_Bootstrap 19, 9 13, 6 2, 3 VannelliVidyasagar2_IterationWithVlin 19, 9 13, 6 2, 3 VincentGrantham_IterationWithVlin 9, 5 6, 3 2, 2 WTBenchmark_IterationWithVlin 19, 9 13, 6 2, 3 21 / 22

22 / 22 Conclusions New reduction method for SDP. Generalizes symmetry reduction and *-algebra-methods Fully algorithmic, don t need to compute automorphisms! Yields optimal block-diagonalization (majorization) Can exploit combinatorial description of subspace Through Jordan algebra theory, extends to LP/SOCP/...

22 / 22 Conclusions New reduction method for SDP. Generalizes symmetry reduction and *-algebra-methods Fully algorithmic, don t need to compute automorphisms! Yields optimal block-diagonalization (majorization) Can exploit combinatorial description of subspace Through Jordan algebra theory, extends to LP/SOCP/... Preprint at arxiv:1608.02090.

22 / 22 Conclusions New reduction method for SDP. Generalizes symmetry reduction and *-algebra-methods Fully algorithmic, don t need to compute automorphisms! Yields optimal block-diagonalization (majorization) Can exploit combinatorial description of subspace Through Jordan algebra theory, extends to LP/SOCP/... Preprint at arxiv:1608.02090. Thanks for your attention!