Sparse Matrix Theory and Semidefinite Optimization
|
|
- Dulcie Poole
- 5 years ago
- Views:
Transcription
1 Sparse Matrix Theory and Semidefinite Optimization Lieven Vandenberghe Department of Electrical Engineering University of California, Los Angeles Joint work with Martin S. Andersen and Yifan Sun Third VORACE Workshop Toulouse, September, 2016
2 Semidefinite program (SDP) minimize tr(c X) subject to tr(a i X) = b i, i = 1,..., m X 0 variable X is n n symmetric matrix; X 0 means X is positive semidefinite matrix inequalities arise naturally in many areas (for example, control, statistics) used in convex modeling systems (CVX, YALMIP, CVXPY,... ) relaxations of nonconvex quadratic and polynomial optimization in many applications the coefficients A i, C are sparse optimal X is typically dense, even for sparse A i, C 1
3 SDP with band structure cost of solving SDP with banded matrices (bandwidth 11, 100 constraints) Time per iteration (seconds) O(n 2 ) SDPT3 SeDuMi for bandwidth 1 (linear program), cost/iteration is linear in n for bandwidth > 1, cost grows as n 2 or faster n (from Andersen, Dahl, Vandenberghe 2010) 2
4 Power flow optimization an optimization problem with non-convex quadratic constraints Variables complex voltage v i at each node (bus) of the network complex power flow s i j entering the link (line) from node i to node j Non-convex constraints (lower) bounds on voltage magnitudes v min v i v max flow balance equations: bus i s i j g i j s ji bus j s i j + s ji = ḡ i j v i v j 2 g i j is admittance of line from node i to j 3
5 Semidefinite relaxation of optimal power flow problem introduce matrix variable X = Re (vv H ), i.e., with elements X i j = Re (v i v j ) voltage bounds and flow balance equations are convex in X: v min v i v max vmin 2 X ii vmax 2 s i j + s ji = ḡ i j v i v j 2 s i j + s ji = ḡ i j (X ii + X j j 2X i j ) replace constraint X = Re (vv H ) with weaker constraint X 0 relaxation is exact if optimal X has rank two Sparsity in SDP relaxation: off-diagonal X i j appears in constraints only if there is a line between buses i and j (Jabr 2006, Bai et al. 2008, Lavaei and Low 2012,... ) 4
6 Outline minimize tr(c X) subject to tr(a i X) = b i, i = 1,..., m X 0 structure in solution X that results from sparsity in coefficients A i, C applications of sparse matrix theory to semidefinite optimization algorithms Outline 1. Chordal graphs 2. Decomposition of sparse matrix cones 3. Multifrontal algorithms for barrier functions 4. Minimum rank positive semidefinite completion
7 Symmetric sparsity pattern A = 3 4 A 11 A 21 A 31 0 A 51 A 21 A 22 0 A 42 0 A 31 0 A 33 0 A 53 0 A 42 0 A 44 A 54 A 51 0 A 53 A 54 A 55 sparsity pattern of symmetric n n matrix is set of nonzero positions E {{i, j} i, j {1, 2,..., n}} A has sparsity pattern E if A i j = 0 if i j and {i, j} E notation: A S n E represented by undirected graph (V, E) with edges E, vertices V = {1,..., n} clique (maximal complete subgraph) forms maximal dense principal submatrix 5
8 Symmetric sparsity pattern A = 3 4 A 11 A 21 A 31 0 A 51 A 21 A 22 0 A 42 0 A 31 0 A 33 0 A 53 0 A 42 0 A 44 A 54 A 51 0 A 53 A 54 A 55 sparsity pattern of symmetric n n matrix is set of nonzero positions E {{i, j} i, j {1, 2,..., n}} A has sparsity pattern E if A i j = 0 if i j and {i, j} E notation: A S n E represented by undirected graph (V, E) with edges E, vertices V = {1,..., n} clique (maximal complete subgraph) forms maximal dense principal submatrix 5
9 Chordal graph undirected graph with vertex set V, edge set E {{v, w} v, w V } G = (V, E) a chord of a cycle is an edge between non-consecutive vertices G is chordal if every cycle of length greater than three has a chord a a f b f b e c e c d not chordal d chordal also known as triangulated, decomposable, rigid circuit graph,... 6
10 History chordal graphs have been studied in many disciplines since the 1960s combinatorial optimization (a class of perfect graphs) linear algebra (sparse factorization, completion problems) computer science (database theory) machine learning (graphical models, probabilistic networks) nonlinear optimization (partial separability) first used in semidefinite optimization by Fujisawa, Kojima, Nakata (1997) 7
11 Chordal sparsity and Cholesky factorization Cholesky factorization of positive definite A S n E : PAP T = LDL T P a permutation, L unit lower triangular, D positive diagonal if E is chordal, then there exists a permutation for which P T (L + L T )P S n E A has a zero fill Cholesky factorization if E is not chordal, then for every P there exist positive definite A S n E for which P T (L + L T )P S n E (Rose 1970) 8
12 Examples Simple patterns Sparsity pattern of a Cholesky factor : edges of non-chordal sparsity pattern : fill entries in Cholesky factorization a chordal extension of non-chordal pattern 9
13 Perfect elimination ordering ordered graph (V, E, σ): undirected graph, plus numbering σ of vertices a 4 e 2 f b 3 f b a 1 e c 6 c d 5 d higher neighborhood of vertex v: adj + (v) = {w {w, v} E, σ 1 (w) > σ 1 (v)} σ is perfect elimination ordering if all higher neighborhoods are complete perfect elimination ordering exists if and only if (V, E) is chordal (Fulkerson and Gross 1965) 10
14 Perfect elimination ordering ordered graph (V, E, σ): undirected graph, plus numbering σ of vertices a 4 e 2 f b 3 f b a 1 e c 6 c d 5 d higher neighborhood of vertex v: adj + (v) = {w {w, v} E, σ 1 (w) > σ 1 (v)} σ is perfect elimination ordering if all higher neighborhoods are complete perfect elimination ordering exists if and only if (V, E) is chordal (Fulkerson and Gross 1965) 10
15 Elimination tree chordal graph with perfect elimination ordering σ = (1, 2,..., n) parent p(v) of vertex v in elimination tree is first vertex in adj + (v) 11
16 Expanded elimination tree , 9 7 7, , , 7, 8 2 5, 8 3 2, 7 1 each block represents a column: bottom row is v, top row is adj + (v) vertices of tree are complete subgraphs (and include all cliques) for each v, tree vertices that contain v form a subtree of elimination tree 12
17 Expanded elimination tree , 9 7 7, , , 7, 8 2 5, 8 3 2, 7 1 each block represents a column: bottom row is v, top row is adj + (v) vertices of tree are complete subgraphs (and include all cliques) for each v, tree vertices that contain v form a subtree of elimination tree 12
18 Supernodal elimination tree , 10 5, 6, 7 6, , 14, 15, 16, 17 16, 17 14, 15, , 12 10, 16 8, 9 9, 16 9, , 2 vertices of tree are cliques of sparsity graph top row of each block is intersection of clique with parent clique bottom rows are supernodes; form a partition of {1, 2,..., n} for each v, cliques that contain v form a subtree of elimination tree 13
19 Supernodal elimination tree , 10 5, 6, 7 6, , 14, 15, 16, 17 16, 17 14, 15, , 12 10, 16 8, 9 9, 16 9, , 2 vertices of tree are cliques of sparsity graph top row of each block is intersection of clique with parent clique bottom rows are supernodes; form a partition of {1, 2,..., n} for each v, cliques that contain v form a subtree of elimination tree 13
20 Outline 1. Chordal graphs 2. Decomposition of sparse matrix cones 3. Multifrontal algorithms for barrier functions 4. Minimum rank positive semidefinite completion
21 Sparse matrix cones we define two matrix cones in S n E (symmetric n n matrices with pattern E) positive semidefinite matrices S n + S n E = {X Sn E X 0} matrices with a positive semidefinite completion Π E (S n +) = {Π E (X) X 0} Π E is projection on S n E Properties two cones are convex closed, pointed, with nonempty interior (relative to S n E ) form a pair of dual cones (for the trace inner product) 14
22 Positive semidefinite matrices with chordal sparsity pattern A S n E is positive semidefinite if and only if it can be expressed as A = cliques γ i P T γ i H i P γi with H i 0 (for an index set β, P β is 0-1 matrix of size β n with P β x = x β for all x) = + + A 0 P T γ 1 H 1 P γ1 0 P T γ 2 H 2 P γ2 0 P T γ 3 H 3 P γ3 0 (Griewank and Toint 1984, Agler, Helton, McCullough, Rodman 1988) 15
23 Decomposition from Cholesky factorization example with two cliques: = H 1 + H 2 H 1 and H 2 follow by combining columns in Cholesky factorization = + readily computed from update matrices in multifrontal Cholesky factorization 16
24 Non-chordal example (Griewank and Toint 1984) graph with chordless cycle (1, 2,..., k, 1) of length k > 3 the following matrix A S n E is positive semidefinite for λ 2 cos(π/k): A = [ à ], à = λ ( 1) k 1 1 λ λ λ 1 ( 1) k λ for λ < 2 there exists no decomposition à = λ u 1 0 ( 1) k ( 1) k 1 0 u k + u λ u with positive semidefinite terms 17
25 Partial separability a function f : R n R is partially separable in variables x i and x j (with i j) if f (x + se i + te j ) = f (x + se i ) + f (x + te j ) f (x) x R n, s, t R f is separable in the variables x i and x j if the other variables are held constant can be represented by an interaction graph with vertices V = {1, 2,..., n}, {i, j} E = f is partially separable in x i and x j Example: f (x) = f 1 (x 1, x 4, x 5 ) + f 2 (x 1, x 3 ) + f 3 (x 2, x 3 ) + f 4 (x 2, x 4 ) (Griewank and Toint 1982, 1984) 18
26 Convex quadratic function interaction graph of f (x) = x T Ax is the sparsity graph of A if G is chordal, with l cliques γ i then there exist PSD matrices H i such that f (x) = x T ( l i=1 P T γ i H i P γi )x = l i=1 x T γ i H i x γi Example 4 2 5, 6 5 3, 4 3, 4 1 f (x) = + T x 1 x 1 x 3 H 1 x 3 x 4 + x 4 [ ] T [ ] x2 x2 H x 3 4 x 4 + T x 3 x 3 x 4 H 2 x 4 x 5 x 5 [ ] T [ ] x5 x5 H x 4 6 x 6 clique tree convex decomposition of f 19
27 Sum of squares a polynomial f : R n R is a sum of squares of degree 2d or less if f (x) = A α β x α x β = v(x) T Av(x) with A 0 α N n d β N n d x α with α = (α 1,..., α n ) denotes the monomial x α 1 1 xα 2 2 xα n n N n d is the set of nonnegative integer vectors (α 1,..., α n ) with i α i d ( ) n + d v(x) is vector of monomials in x of degree d or less d Sparse sums of squares: exploit partial separability to reduce size of A (Lasserre 2006, Kojima and Muramatsu 2009) 20
28 Sums of squares with chordal interaction graph f (x) = v(x) T Av(x) define a graph G = (V, E) with V = {1,..., n} and {i, j} E = f contains no monomials that depend on x i and x j (A α β = 0 if α i + β i > 0 and α j + β j > 0) sparsity pattern of A is chordal if G is chordal Example (n = 6, d = 2): clique tree of G and clique tree of sparsity graph of A 5, 6 5 3, 4 1, x 5, x 6, x 2 5, x 5x 6, x 2 6 1, x 5, x 2 5 x 3, x 4, x 2 3, x 3x 4, x 3 x 5, x 2 4, x 4x 5, x , 4 1 1, x 4, x 2 4 x 2, x 2 2, x 2x 4 1, x 3, x 4, x2 3, x 3x 4, x 2 4 x 1, x 2 1, x 1x 3, x 1 x 4 21
29 Clique decomposition clique decomposition gives more compact sum-of-squares representation v(x γi ) is the l f (x) = v(x γi ) T H i v(x γi ), H i 0 for i = 1,..., l i=1 ( γi + d d ) -vector with monomials in x γi of degree d or less Example (n = 6, d = 2): clique tree of G and clique tree of sparsity graph of A 4 2 5, 6 5 3, 4 3, 4 1 f (x) = v(x 1, x 3, x 4 ) T H 1 v(x 1, x 3, x 4 ) + v(x 3, x 4, x 5 ) T H 2 v(x 3, x 4, x 5 ) + v(x 2, x 4 ) T H 3 v(x 2, x 4 ) + v(x 5, x 6 ) T H 4 + v(x 5, x 6 ) clique tree sparse sum of squares decomposition of f 22
30 PSD-completable cone with chordal sparsity A S n E has a positive semidefinite completion if and only if A γi γ i 0 for all cliques γ i follows from duality and clique decomposition of positive semidefinite cone Example (three cliques γ 1, γ 2, γ 3 ) A γ1 γ 1 0 PSD completable A A γ2 γ 2 0 A γ3 γ 3 0 (Grone, Johnson, Sá, Wolkowicz, 1984) 23
31 Sparse semidefinite optimization minimize tr(c X) subject to tr(a i X) = b i, i = 1,..., m X 0 aggregate sparsity pattern E is union of patterns of C, A 1,..., A m without loss of generality, can assume E is chordal constraint X 0 can be replaced by X γi γ i 0 for all cliques γ i Decomposition algorithms replace X with smaller matrix variables X i = X γi γ i add equality constraints to enforce consistency of the variables X i first proposed for interior-point methods (Fukuda et al. 2000, Nakata et al. 2003) first-order (dual decomposition) methods (Lu et al. 2007, Lam et al. 2011,... ) 24
32 Example: sparse nearest matrix problems find nearest sparse PSD-completable matrix with given sparsity pattern minimize X A F 2 subject to X Π E (S+ n ) find nearest sparse PSD matrix with given sparsity pattern minimize subject to S + A 2 F S S n + Sn E these two problems are duals: K = Π E (S n + ) K = (S n + Sn E ) A 25
33 Decomposition methods from the decomposition theorems, the two problems can be written primal: minimize X A 2 F subject to X γi γ i 0, i = 1,..., l dual: minimize A + l i=1 P T γ i H i P γi 2 F subject to H i 0, i = 1,..., l Algorithms Dykstra s algorithm (dual block coordinate ascent) (fast) dual projected gradient algorithm (FISTA) Douglas-Rachford splitting, ADMM sequence of projections on PSD cones of order γ i (eigenvalue decomposition) 26
34 Results sparsity patterns from University of Florida Sparse Matrix Collection n density #cliques avg. clique size max. clique e e e e e total runtime (sec) time/iteration (sec) n FISTA Dykstra DR FISTA Dykstra DR e2 3.9e1 3.8e e2 4.7e1 6.2e > 4hr 1.4e2 1.1e e3 3.0e2 2.4e e2 5.5e1 1.0e (Sun and Vandenberghe 2015) 27
35 Outline 1. Chordal graphs 2. Decomposition of sparse matrix cones 3. Multifrontal algorithms for barrier functions 4. Minimum rank positive semidefinite completion
36 Sparse SDP as nonsymmetric conic linear program minimize tr(c X) subject to tr(a i X) = b i, i = 1,..., m X 0 aggregate sparsity pattern E is union of patterns of C, A 1,..., A m Equivalent conic LP minimize tr(c X) subject to tr(a i X) = b i, i = 1,..., m X K K Π E (S+ n ) is cone of PSD completable matrices with pattern E optimization problem in a lower-dimensional space K is not self-dual; no symmetric interior-point methods 28
37 Barrier function for positive semidefinite cone φ(s) = log det S, dom φ = {S S n E S 0} gradient (negative projected inverse) φ(s) = Π E (S 1 ) requires entries of dense inverse S 1 on diagonal and for {i, j} E Hessian applied to sparse Y S n E : 2 φ(s)[y] = d dt φ(s + ty ) t=0 = Π E ( S 1 Y S 1) requires projection of dense product S 1 Y S 1 29
38 Multifrontal algorithms assume E is a chordal sparsity pattern (or chordal extension) Cholesky factorization (Duff and Reid 1983) factorization S = LDL T gives function value of barrier: φ(s) = i log D ii recursion on elimination tree in topological order Gradient (Campbell and Davis 1995, Andersen et al. 2013) compute φ(s) = Π E (S 1 ) from equation S 1 L = L T D 1 recursion on elimination tree in inverse topological order Hessian compute 2 φ(s)[y] = Π E (S 1 Y S 1 ) by linearizing recursion for gradient two recursions on elimination tree (topological and inverse topological order) 30
39 Projected inverse versus Cholesky factorization 10 6 Problem statistics 10 1 Time comparison Number of nonzeros in A Projected inverse Order n Cholesky factorization 667 patterns from University of Florida Sparse Matrix Collection time in seconds for projected inverse and Cholesky factorization code at github.com/cvxopt/chompack 31
40 Barrier for positive semidefinite completable cone φ (X) = sup S ( tr(x S) φ(s)), dom φ = {X = Π E (Y ) Y 0} this is the conjugate of the barrier φ(s) = log det S for the sparse PSD cone inverse Z = Ŝ 1 of optimal Ŝ is maximum determinant PD completion of X: maximize subject to log det Z Π E (Z) = X gradient and Hessian of φ at X are φ (X) = Ŝ, 2 φ (X) = 2 φ(ŝ) 1 for chordal E, efficient multifrontal algorithms for Cholesky factors of Ŝ, given X 32
41 Inverse completion versus Cholesky factorization 10 1 Inverse completion factorization Cholesky factorization time for Cholesky factorization of inverse of maximum determinant PD completion 33
42 Nonsymmetric interior-point methods minimize tr(c X) subject to tr(a i X) = b i, i = 1,..., m X Π E (S n + ) can be solved by nonsymmetric primal or dual barrier methods logarithmic barriers for cone Π E (S n + ) and its dual cone Sn + Sn E : φ (X) = sup S ( tr(x S) + log det S ), φ(s) = log det S fast evaluation of barrier values and derivatives if pattern is chordal examples: linear complexity per iteration for band or arrow pattern code and numerical results at github.com/cvxopt/smcp (Fukuda et al. 2000, Burer 2003, Srijungtongsiri and Vavasis 2004, Andersen et al. 2010) 34
43 Sparsity patterns sparsity patterns from University of Florida Sparse Matrix Collection m = 200 constraints random data with 0.05% nonzeros in Ai relative to E rs228 rs35 rs200 rs365 n = 1,919 n = 2,003 n = 3,025 n = 4, rs1555 rs828 rs1184 rs1288 n = 7,479 n = 10,800 n = 14,822 n = 30,401 35
44 Results n DSDP SDPA SDPA-C SDPT3 SeDuMi SMCP n DSDP SDPA-C SMCP mem average time per iteration for different solvers SMCP uses nonsymmetric matrix cone approach (Andersen et al. 2010) 36
45 Outline 1. Chordal graphs 2. Decomposition of sparse matrix cones 3. Multifrontal algorithms for logarithmic barriers 4. Minimum rank positive semidefinite completion
46 Minimum rank PSD completion with chordal sparsity recall that A S n E has a positive semidefinite completion if and only if A γi γ i 0 for all cliques γ i A γ1 γ 1 0 PSD completable A A γ2 γ 2 0 A γ3 γ 3 0 the minimum rank PSD completion has rank equal to max rank(a γi γ i ) cliques γ i (Dancis 1992) 37
47 Two-block completion problem we first consider the simple two-block completion problem?? A = A 11 A 12 0 A 21 A 22 A 23 0 A 32 A 33 a completion exists if and only if C 1 = [ ] A11 A 12 A 21 A 22 0, C 2 = [ ] A22 A 23 A 32 A 33 0 we construct a PSD completion of rank r = max{rank(c 1 ), rank(c 2 )} 38
48 Two-block completion algorithm compute matrices U, V, Ṽ, W of column dimension r such that [ ] A11 A 12 A 21 A 22 = [ U V ] [ U V ] T, [ ] A22 A 23 A 32 A 33 = [ Ṽ W ] [ Ṽ W ] T since VV T = ṼṼ T, the matrices V and Ṽ have SVDs V = PΣQ T 1, Ṽ = PΣQT 2 hence V = ṼQ with Q = Q 2 Q T 1 an orthogonal r r matrix a completion of rank r is given by UQ T Ṽ W UQ T Ṽ W T = A 11 A 12 UQ T W T A 21 A 22 A 23 WQU T A 32 A 33 39
49 Minimum rank completion for general chordal pattern recursion over supernodal elimination tree in inverse topological order , 10 5, 6, 7 6, , 14, 15, 16, 17 16, 17 14, 15, , 12 10, 16 8, 9 9, 16 9, , 2 vertices of tree are cliques γ i top row of vertex i is intersection α i of clique γ i with parent clique bottom rows are supernodes ν i = γ i \ α i ; these form a partition of {1, 2,..., n} 40
50 Minimum rank completion for general chordal pattern we compute an n r matrix Y with Π E (YY T ) = A and r = max cliques γ i rank(a γi γ i ) the block rows Y νj are computed in inverse topological order hence Y α j is known when we compute Y νj Algorithm: enumerate cliques in inverse topological order at vertex γ j, compute matrices U j, V j with column dimension r such that [ ] Aνj ν j A νj α j A α j ν j A α j α j = [ Uj V j ] [ Uj V j ] T if j is the root of the supernodal elimination tree, set Y νj = U j otherwise, compute orthogonal Q such that V j = Y α j Q and set Y νj = U j Q T 41
51 Sparse semidefinite optimization minimize tr(c X) subject to tr(a i X) = b i, i = 1,..., m X 0 aggregate sparsity pattern E is union of patterns of C, A 1,..., A m any feasible X can be replaced by PSD completion of Π E (X) if E is chordal, there is a completion with rank bounded by size of largest clique proves exactness of some simple SDP relaxations useful for rounding solution of SDP relaxations to minimum rank solution 42
52 Summary Sparse matrix theory: PSD and PSD-completable matrices with chordal pattern decomposition of matrix cones as sum or intersection of simple cones fast algorithms for evaluating barrier functions and derivatives simple algorithms for maximum determinant and minimum rank completion simple algorithms for Euclidean distance matrix completion Sparse semidefinite optimization minimize tr(c X) subject to tr(a i X) = b i, i = 1,..., m X 0 decomposition and splitting methods nonsymmetric interior-point methods 43
III. Applications in convex optimization
III. Applications in convex optimization nonsymmetric interior-point methods partial separability and decomposition partial separability first order methods interior-point methods Conic linear optimization
More informationarxiv: v1 [math.oc] 26 Sep 2015
arxiv:1509.08021v1 [math.oc] 26 Sep 2015 Degeneracy in Maximal Clique Decomposition for Semidefinite Programs Arvind U. Raghunathan and Andrew V. Knyazev Mitsubishi Electric Research Laboratories 201 Broadway,
More informationResearch Reports on Mathematical and Computing Sciences
ISSN 1342-284 Research Reports on Mathematical and Computing Sciences Exploiting Sparsity in Linear and Nonlinear Matrix Inequalities via Positive Semidefinite Matrix Completion Sunyoung Kim, Masakazu
More informationDegeneracy in Maximal Clique Decomposition for Semidefinite Programs
MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com Degeneracy in Maximal Clique Decomposition for Semidefinite Programs Raghunathan, A.U.; Knyazev, A. TR2016-040 July 2016 Abstract Exploiting
More informationDistributed Control of Connected Vehicles and Fast ADMM for Sparse SDPs
Distributed Control of Connected Vehicles and Fast ADMM for Sparse SDPs Yang Zheng Department of Engineering Science, University of Oxford Homepage: http://users.ox.ac.uk/~ball4503/index.html Talk at Cranfield
More informationChordal Sparsity in Interior-Point Methods for Conic Optimization
University of California Los Angeles Chordal Sparsity in Interior-Point Methods for Conic Optimization A dissertation submitted in partial satisfaction of the requirements for the degree Doctor of Philosophy
More informationSemidefinite Programming
Semidefinite Programming Basics and SOS Fernando Mário de Oliveira Filho Campos do Jordão, 2 November 23 Available at: www.ime.usp.br/~fmario under talks Conic programming V is a real vector space h, i
More informationParallel implementation of primal-dual interior-point methods for semidefinite programs
Parallel implementation of primal-dual interior-point methods for semidefinite programs Masakazu Kojima, Kazuhide Nakata Katsuki Fujisawa and Makoto Yamashita 3rd Annual McMaster Optimization Conference:
More informationLMI MODELLING 4. CONVEX LMI MODELLING. Didier HENRION. LAAS-CNRS Toulouse, FR Czech Tech Univ Prague, CZ. Universidad de Valladolid, SP March 2009
LMI MODELLING 4. CONVEX LMI MODELLING Didier HENRION LAAS-CNRS Toulouse, FR Czech Tech Univ Prague, CZ Universidad de Valladolid, SP March 2009 Minors A minor of a matrix F is the determinant of a submatrix
More informationResearch Reports on Mathematical and Computing Sciences
ISSN 1342-2804 Research Reports on Mathematical and Computing Sciences Sums of Squares and Semidefinite Programming Relaxations for Polynomial Optimization Problems with Structured Sparsity Hayato Waki,
More informationFast Algorithms for SDPs derived from the Kalman-Yakubovich-Popov Lemma
Fast Algorithms for SDPs derived from the Kalman-Yakubovich-Popov Lemma Venkataramanan (Ragu) Balakrishnan School of ECE, Purdue University 8 September 2003 European Union RTN Summer School on Multi-Agent
More informationLecture Note 5: Semidefinite Programming for Stability Analysis
ECE7850: Hybrid Systems:Theory and Applications Lecture Note 5: Semidefinite Programming for Stability Analysis Wei Zhang Assistant Professor Department of Electrical and Computer Engineering Ohio State
More informationPreliminaries Overview OPF and Extensions. Convex Optimization. Lecture 8 - Applications in Smart Grids. Instructor: Yuanzhang Xiao
Convex Optimization Lecture 8 - Applications in Smart Grids Instructor: Yuanzhang Xiao University of Hawaii at Manoa Fall 2017 1 / 32 Today s Lecture 1 Generalized Inequalities and Semidefinite Programming
More informationEquivalent relaxations of optimal power flow
Equivalent relaxations of optimal power flow 1 Subhonmesh Bose 1, Steven H. Low 2,1, Thanchanok Teeraratkul 1, Babak Hassibi 1 1 Electrical Engineering, 2 Computational and Mathematical Sciences California
More informationResearch Reports on Mathematical and Computing Sciences
ISSN 1342-2804 Research Reports on Mathematical and Computing Sciences Correlative sparsity in primal-dual interior-point methods for LP, SDP and SOCP Kazuhiro Kobayashi, Sunyoung Kim and Masakazu Kojima
More informationMATHEMATICAL ENGINEERING TECHNICAL REPORTS. Exact SDP Relaxations with Truncated Moment Matrix for Binary Polynomial Optimization Problems
MATHEMATICAL ENGINEERING TECHNICAL REPORTS Exact SDP Relaxations with Truncated Moment Matrix for Binary Polynomial Optimization Problems Shinsaku SAKAUE, Akiko TAKEDA, Sunyoung KIM, and Naoki ITO METR
More informationResearch Reports on Mathematical and Computing Sciences
ISSN 1342-2804 Research Reports on Mathematical and Computing Sciences Doubly Nonnegative Relaxations for Quadratic and Polynomial Optimization Problems with Binary and Box Constraints Sunyoung Kim, Masakazu
More informationPositive Semidefinite Matrix Completions on Chordal Graphs and Constraint Nondegeneracy in Semidefinite Programming
Positive Semidefinite Matrix Completions on Chordal Graphs and Constraint Nondegeneracy in Semidefinite Programming Houduo Qi February 1, 008 and Revised October 8, 008 Abstract Let G = (V, E) be a graph
More informationFast ADMM for Sum of Squares Programs Using Partial Orthogonality
Fast ADMM for Sum of Squares Programs Using Partial Orthogonality Antonis Papachristodoulou Department of Engineering Science University of Oxford www.eng.ox.ac.uk/control/sysos antonis@eng.ox.ac.uk with
More informationPOLYNOMIAL OPTIMIZATION WITH SUMS-OF-SQUARES INTERPOLANTS
POLYNOMIAL OPTIMIZATION WITH SUMS-OF-SQUARES INTERPOLANTS Sercan Yıldız syildiz@samsi.info in collaboration with Dávid Papp (NCSU) OPT Transition Workshop May 02, 2017 OUTLINE Polynomial optimization and
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Brown University CSCI 295-P, Spring 213 Prof. Erik Sudderth Lecture 11: Inference & Learning Overview, Gaussian Graphical Models Some figures courtesy Michael Jordan s draft
More informationSMCP Documentation. Release Martin S. Andersen and Lieven Vandenberghe
SMCP Documentation Release 0.4.5 Martin S. Andersen and Lieven Vandenberghe May 18, 2018 Contents 1 Current release 3 2 Future releases 5 3 Availability 7 4 Authors 9 5 Feedback and bug reports 11 5.1
More informationEE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 17
EE/ACM 150 - Applications of Convex Optimization in Signal Processing and Communications Lecture 17 Andre Tkacenko Signal Processing Research Group Jet Propulsion Laboratory May 29, 2012 Andre Tkacenko
More informationApproximation Algorithms
Approximation Algorithms Chapter 26 Semidefinite Programming Zacharias Pitouras 1 Introduction LP place a good lower bound on OPT for NP-hard problems Are there other ways of doing this? Vector programs
More informationCOURSE ON LMI PART I.2 GEOMETRY OF LMI SETS. Didier HENRION henrion
COURSE ON LMI PART I.2 GEOMETRY OF LMI SETS Didier HENRION www.laas.fr/ henrion October 2006 Geometry of LMI sets Given symmetric matrices F i we want to characterize the shape in R n of the LMI set F
More informationIntroduction to Semidefinite Programming I: Basic properties a
Introduction to Semidefinite Programming I: Basic properties and variations on the Goemans-Williamson approximation algorithm for max-cut MFO seminar on Semidefinite Programming May 30, 2010 Semidefinite
More information18. Primal-dual interior-point methods
L. Vandenberghe EE236C (Spring 213-14) 18. Primal-dual interior-point methods primal-dual central path equations infeasible primal-dual method primal-dual method for self-dual embedding 18-1 Symmetric
More informationSemidefinite Programming, Combinatorial Optimization and Real Algebraic Geometry
Semidefinite Programming, Combinatorial Optimization and Real Algebraic Geometry assoc. prof., Ph.D. 1 1 UNM - Faculty of information studies Edinburgh, 16. September 2014 Outline Introduction Definition
More informationSemidefinite Programming Basics and Applications
Semidefinite Programming Basics and Applications Ray Pörn, principal lecturer Åbo Akademi University Novia University of Applied Sciences Content What is semidefinite programming (SDP)? How to represent
More informationLecture: Examples of LP, SOCP and SDP
1/34 Lecture: Examples of LP, SOCP and SDP Zaiwen Wen Beijing International Center For Mathematical Research Peking University http://bicmr.pku.edu.cn/~wenzw/bigdata2018.html wenzw@pku.edu.cn Acknowledgement:
More informationA Julia JuMP-based module for polynomial optimization with complex variables applied to ACOPF
JuMP-dev workshop 018 A Julia JuMP-based module for polynomial optimization with complex variables applied to ACOPF Gilles Bareilles, Manuel Ruiz, Julie Sliwak 1. MathProgComplex.jl: A toolbox for Polynomial
More informationELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications
ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications Professor M. Chiang Electrical Engineering Department, Princeton University March
More informationExplicit Sensor Network Localization using Semidefinite Programming and Clique Reductions
Explicit Sensor Network Localization using Semidefinite Programming and Clique Reductions Nathan Krislock and Henry Wolkowicz Dept. of Combinatorics and Optimization University of Waterloo Haifa Matrix
More information15. Conic optimization
L. Vandenberghe EE236C (Spring 216) 15. Conic optimization conic linear program examples modeling duality 15-1 Generalized (conic) inequalities Conic inequality: a constraint x K where K is a convex cone
More informationA NEW SECOND-ORDER CONE PROGRAMMING RELAXATION FOR MAX-CUT PROBLEMS
Journal of the Operations Research Society of Japan 2003, Vol. 46, No. 2, 164-177 2003 The Operations Research Society of Japan A NEW SECOND-ORDER CONE PROGRAMMING RELAXATION FOR MAX-CUT PROBLEMS Masakazu
More information1. Introduction. We consider conic linear optimization problems (conic LPs) minimize c T x (1.1)
SIAM J. OPTIM. Vol. 24, No. 2, pp. 873 897 c 2014 Society for Industrial and Applied Mathematics DECOMPOSITION IN CONIC OPTIMIZATION WITH PARTIALLY SEPARABLE STRUCTURE YIFAN SUN, MARTIN S. ANDERSEN, AND
More informationReal Symmetric Matrices and Semidefinite Programming
Real Symmetric Matrices and Semidefinite Programming Tatsiana Maskalevich Abstract Symmetric real matrices attain an important property stating that all their eigenvalues are real. This gives rise to many
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra)
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 3: Positive-Definite Systems; Cholesky Factorization Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical Analysis I 1 / 11 Symmetric
More informationL. Vandenberghe EE236C (Spring 2016) 18. Symmetric cones. definition. spectral decomposition. quadratic representation. log-det barrier 18-1
L. Vandenberghe EE236C (Spring 2016) 18. Symmetric cones definition spectral decomposition quadratic representation log-det barrier 18-1 Introduction This lecture: theoretical properties of the following
More informationSemidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 5
Semidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 5 Instructor: Farid Alizadeh Scribe: Anton Riabov 10/08/2001 1 Overview We continue studying the maximum eigenvalue SDP, and generalize
More informationCSCI 1951-G Optimization Methods in Finance Part 10: Conic Optimization
CSCI 1951-G Optimization Methods in Finance Part 10: Conic Optimization April 6, 2018 1 / 34 This material is covered in the textbook, Chapters 9 and 10. Some of the materials are taken from it. Some of
More informationNonlinear Optimization for Optimal Control
Nonlinear Optimization for Optimal Control Pieter Abbeel UC Berkeley EECS Many slides and figures adapted from Stephen Boyd [optional] Boyd and Vandenberghe, Convex Optimization, Chapters 9 11 [optional]
More informationSemidefinite Programming
Semidefinite Programming Notes by Bernd Sturmfels for the lecture on June 26, 208, in the IMPRS Ringvorlesung Introduction to Nonlinear Algebra The transition from linear algebra to nonlinear algebra has
More informationThe Ongoing Development of CSDP
The Ongoing Development of CSDP Brian Borchers Department of Mathematics New Mexico Tech Socorro, NM 87801 borchers@nmt.edu Joseph Young Department of Mathematics New Mexico Tech (Now at Rice University)
More informationIntroduction to Semidefinite Programs
Introduction to Semidefinite Programs Masakazu Kojima, Tokyo Institute of Technology Semidefinite Programming and Its Application January, 2006 Institute for Mathematical Sciences National University of
More informationE5295/5B5749 Convex optimization with engineering applications. Lecture 5. Convex programming and semidefinite programming
E5295/5B5749 Convex optimization with engineering applications Lecture 5 Convex programming and semidefinite programming A. Forsgren, KTH 1 Lecture 5 Convex optimization 2006/2007 Convex quadratic program
More informationHow to generate weakly infeasible semidefinite programs via Lasserre s relaxations for polynomial optimization
CS-11-01 How to generate weakly infeasible semidefinite programs via Lasserre s relaxations for polynomial optimization Hayato Waki Department of Computer Science, The University of Electro-Communications
More informationSparse Optimization Lecture: Basic Sparse Optimization Models
Sparse Optimization Lecture: Basic Sparse Optimization Models Instructor: Wotao Yin July 2013 online discussions on piazza.com Those who complete this lecture will know basic l 1, l 2,1, and nuclear-norm
More informationarxiv: v2 [math.oc] 15 Aug 2018
Noname manuscript No. (will be inserted by the editor) Sparse Semidefinite Programs with Guaranteed Near-Linear Time Complexity via Dualized Clique Tree Conversion Richard Y. Zhang Javad Lavaei arxiv:1710.03475v2
More informationProblem structure in semidefinite programs arising in control and signal processing
Problem structure in semidefinite programs arising in control and signal processing Lieven Vandenberghe Electrical Engineering Department, UCLA Joint work with: Mehrdad Nouralishahi, Tae Roh Semidefinite
More informationCSC Linear Programming and Combinatorial Optimization Lecture 10: Semidefinite Programming
CSC2411 - Linear Programming and Combinatorial Optimization Lecture 10: Semidefinite Programming Notes taken by Mike Jamieson March 28, 2005 Summary: In this lecture, we introduce semidefinite programming
More informationFurther Relaxations of the SDP Approach to Sensor Network Localization
Further Relaxations of the SDP Approach to Sensor Network Localization Zizhuo Wang, Song Zheng, Stephen Boyd, and inyu e September 27, 2006 Abstract Recently, a semidefinite programming (SDP) relaxation
More informationV C V L T I 0 C V B 1 V T 0 I. l nk
Multifrontal Method Kailai Xu September 16, 2017 Main observation. Consider the LDL T decomposition of a SPD matrix [ ] [ ] [ ] [ ] B V T L 0 I 0 L T L A = = 1 V T V C V L T I 0 C V B 1 V T, 0 I where
More information9. Numerical linear algebra background
Convex Optimization Boyd & Vandenberghe 9. Numerical linear algebra background matrix structure and algorithm complexity solving linear equations with factored matrices LU, Cholesky, LDL T factorization
More informationPreliminaries and Complexity Theory
Preliminaries and Complexity Theory Oleksandr Romanko CAS 746 - Advanced Topics in Combinatorial Optimization McMaster University, January 16, 2006 Introduction Book structure: 2 Part I Linear Algebra
More informationChordally Sparse Semidefinite Programs
Chordally Sparse Semidefinite Programs Raphael Louca Abstract We consider semidefinite programs where the collective sparsity pattern of the data matrices is characterized by a given chordal graph. Acyclic,
More informationA solution approach for linear optimization with completely positive matrices
A solution approach for linear optimization with completely positive matrices Franz Rendl http://www.math.uni-klu.ac.at Alpen-Adria-Universität Klagenfurt Austria joint work with M. Bomze (Wien) and F.
More information6-1 The Positivstellensatz P. Parrilo and S. Lall, ECC
6-1 The Positivstellensatz P. Parrilo and S. Lall, ECC 2003 2003.09.02.10 6. The Positivstellensatz Basic semialgebraic sets Semialgebraic sets Tarski-Seidenberg and quantifier elimination Feasibility
More informationCS675: Convex and Combinatorial Optimization Fall 2016 Combinatorial Problems as Linear and Convex Programs. Instructor: Shaddin Dughmi
CS675: Convex and Combinatorial Optimization Fall 2016 Combinatorial Problems as Linear and Convex Programs Instructor: Shaddin Dughmi Outline 1 Introduction 2 Shortest Path 3 Algorithms for Single-Source
More informationMIT Algebraic techniques and semidefinite optimization February 14, Lecture 3
MI 6.97 Algebraic techniques and semidefinite optimization February 4, 6 Lecture 3 Lecturer: Pablo A. Parrilo Scribe: Pablo A. Parrilo In this lecture, we will discuss one of the most important applications
More informationRobust and Optimal Control, Spring 2015
Robust and Optimal Control, Spring 2015 Instructor: Prof. Masayuki Fujita (S5-303B) G. Sum of Squares (SOS) G.1 SOS Program: SOS/PSD and SDP G.2 Duality, valid ineqalities and Cone G.3 Feasibility/Optimization
More information9. Numerical linear algebra background
Convex Optimization Boyd & Vandenberghe 9. Numerical linear algebra background matrix structure and algorithm complexity solving linear equations with factored matrices LU, Cholesky, LDL T factorization
More informationThe maximal stable set problem : Copositive programming and Semidefinite Relaxations
The maximal stable set problem : Copositive programming and Semidefinite Relaxations Kartik Krishnan Department of Mathematical Sciences Rensselaer Polytechnic Institute Troy, NY 12180 USA kartis@rpi.edu
More informationFisher Information in Gaussian Graphical Models
Fisher Information in Gaussian Graphical Models Jason K. Johnson September 21, 2006 Abstract This note summarizes various derivations, formulas and computational algorithms relevant to the Fisher information
More informationLagrange Duality. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)
Lagrange Duality Daniel P. Palomar Hong Kong University of Science and Technology (HKUST) ELEC5470 - Convex Optimization Fall 2017-18, HKUST, Hong Kong Outline of Lecture Lagrangian Dual function Dual
More informationChapter 3. Some Applications. 3.1 The Cone of Positive Semidefinite Matrices
Chapter 3 Some Applications Having developed the basic theory of cone programming, it is time to apply it to our actual subject, namely that of semidefinite programming. Indeed, any semidefinite program
More informationHomework 2 Foundations of Computational Math 2 Spring 2019
Homework 2 Foundations of Computational Math 2 Spring 2019 Problem 2.1 (2.1.a) Suppose (v 1,λ 1 )and(v 2,λ 2 ) are eigenpairs for a matrix A C n n. Show that if λ 1 λ 2 then v 1 and v 2 are linearly independent.
More informationSEMIDEFINITE PROGRAM BASICS. Contents
SEMIDEFINITE PROGRAM BASICS BRIAN AXELROD Abstract. A introduction to the basics of Semidefinite programs. Contents 1. Definitions and Preliminaries 1 1.1. Linear Algebra 1 1.2. Convex Analysis (on R n
More informationConvex Optimization of Graph Laplacian Eigenvalues
Convex Optimization of Graph Laplacian Eigenvalues Stephen Boyd Abstract. We consider the problem of choosing the edge weights of an undirected graph so as to maximize or minimize some function of the
More information12. Cholesky factorization
L. Vandenberghe ECE133A (Winter 2018) 12. Cholesky factorization positive definite matrices examples Cholesky factorization complex positive definite matrices kernel methods 12-1 Definitions a symmetric
More informationA Modular AC Optimal Power Flow Implementation for Distribution Grid Planning
A Modular AC Optimal Power Flow Implementation for Distribution Grid Planning Adrian Hauswirth Tyler Summers Joseph Warrington John Lygeros Automatic Control Laboratory Swiss Federal Institute of Technology
More informationModeling with semidefinite and copositive matrices
Modeling with semidefinite and copositive matrices Franz Rendl http://www.math.uni-klu.ac.at Alpen-Adria-Universität Klagenfurt Austria F. Rendl, Singapore workshop 2006 p.1/24 Overview Node and Edge relaxations
More informationDimension reduction for semidefinite programming
1 / 22 Dimension reduction for semidefinite programming Pablo A. Parrilo Laboratory for Information and Decision Systems Electrical Engineering and Computer Science Massachusetts Institute of Technology
More informationThe doubly negative matrix completion problem
The doubly negative matrix completion problem C Mendes Araújo, Juan R Torregrosa and Ana M Urbano CMAT - Centro de Matemática / Dpto de Matemática Aplicada Universidade do Minho / Universidad Politécnica
More informationLecture 2: Linear Algebra Review
EE 227A: Convex Optimization and Applications January 19 Lecture 2: Linear Algebra Review Lecturer: Mert Pilanci Reading assignment: Appendix C of BV. Sections 2-6 of the web textbook 1 2.1 Vectors 2.1.1
More informationSecond Order Cone Programming Relaxation of Nonconvex Quadratic Optimization Problems
Second Order Cone Programming Relaxation of Nonconvex Quadratic Optimization Problems Sunyoung Kim Department of Mathematics, Ewha Women s University 11-1 Dahyun-dong, Sudaemoon-gu, Seoul 120-750 Korea
More informationCopositive Plus Matrices
Copositive Plus Matrices Willemieke van Vliet Master Thesis in Applied Mathematics October 2011 Copositive Plus Matrices Summary In this report we discuss the set of copositive plus matrices and their
More informationStrong duality in Lasserre s hierarchy for polynomial optimization
Strong duality in Lasserre s hierarchy for polynomial optimization arxiv:1405.7334v1 [math.oc] 28 May 2014 Cédric Josz 1,2, Didier Henrion 3,4,5 Draft of January 24, 2018 Abstract A polynomial optimization
More informationConic Approximation with Provable Guarantee for Traffic Signal Offset Optimization
Conic Approximation with Provable Guarantee for Traffic Signal Offset Optimization Yi Ouyang, Richard Y. Zhang, Javad Lavaei, and Pravin Varaiya Abstract We consider the offset optimization problem that
More informationINTERIOR-POINT METHODS ROBERT J. VANDERBEI JOINT WORK WITH H. YURTTAN BENSON REAL-WORLD EXAMPLES BY J.O. COLEMAN, NAVAL RESEARCH LAB
1 INTERIOR-POINT METHODS FOR SECOND-ORDER-CONE AND SEMIDEFINITE PROGRAMMING ROBERT J. VANDERBEI JOINT WORK WITH H. YURTTAN BENSON REAL-WORLD EXAMPLES BY J.O. COLEMAN, NAVAL RESEARCH LAB Outline 2 Introduction
More informationInterior Point Algorithms for Constrained Convex Optimization
Interior Point Algorithms for Constrained Convex Optimization Chee Wei Tan CS 8292 : Advanced Topics in Convex Optimization and its Applications Fall 2010 Outline Inequality constrained minimization problems
More information6.854J / J Advanced Algorithms Fall 2008
MIT OpenCourseWare http://ocw.mit.edu 6.85J / 8.5J Advanced Algorithms Fall 008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. 8.5/6.85 Advanced Algorithms
More informationFurther Relaxations of the SDP Approach to Sensor Network Localization
Further Relaxations of the SDP Approach to Sensor Network Localization Zizhuo Wang, Song Zheng, Stephen Boyd, and Yinyu Ye September 27, 2006; Revised May 2, 2007 Abstract Recently, a semi-definite programming
More informationDistance Geometry-Matrix Completion
Christos Konaxis Algs in Struct BioInfo 2010 Outline Outline Tertiary structure Incomplete data Tertiary structure Measure diffrence of matched sets Def. Root Mean Square Deviation RMSD = 1 n x i y i 2,
More informationConvex relaxation. In example below, we have N = 6, and the cut we are considering
Convex relaxation The art and science of convex relaxation revolves around taking a non-convex problem that you want to solve, and replacing it with a convex problem which you can actually solve the solution
More informationA PARALLEL INTERIOR POINT DECOMPOSITION ALGORITHM FOR BLOCK-ANGULAR SEMIDEFINITE PROGRAMS IN POLYNOMIAL OPTIMIZATION
A PARALLEL INTERIOR POINT DECOMPOSITION ALGORITHM FOR BLOCK-ANGULAR SEMIDEFINITE PROGRAMS IN POLYNOMIAL OPTIMIZATION Kartik K. Sivaramakrishnan Department of Mathematics North Carolina State University
More informationA direct formulation for sparse PCA using semidefinite programming
A direct formulation for sparse PCA using semidefinite programming A. d Aspremont, L. El Ghaoui, M. Jordan, G. Lanckriet ORFE, Princeton University & EECS, U.C. Berkeley Available online at www.princeton.edu/~aspremon
More informationIntroduction and Math Preliminaries
Introduction and Math Preliminaries Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A. http://www.stanford.edu/ yyye Appendices A, B, and C, Chapter
More informationConvex Optimization of Graph Laplacian Eigenvalues
Convex Optimization of Graph Laplacian Eigenvalues Stephen Boyd Stanford University (Joint work with Persi Diaconis, Arpita Ghosh, Seung-Jean Kim, Sanjay Lall, Pablo Parrilo, Amin Saberi, Jun Sun, Lin
More informationEfficient Algorithm for Large-and-Sparse LMI Feasibility Problems
Efficient Algorithm for Large-and-Sparse LMI Feasibility Problems Richard Y. Zhang and Javad Lavaei Abstract Linear matrix inequalities (LMIs) play a fundamental role in robust and optimal control theory.
More informationTamás Terlaky George N. and Soteria Kledaras 87 Endowed Chair Professor. Chair, Department of Industrial and Systems Engineering Lehigh University
5th SJOM Bejing, 2011 Cone Linear Optimization (CLO) From LO, SOCO and SDO Towards Mixed-Integer CLO Tamás Terlaky George N. and Soteria Kledaras 87 Endowed Chair Professor. Chair, Department of Industrial
More information11. Equality constrained minimization
Convex Optimization Boyd & Vandenberghe 11. Equality constrained minimization equality constrained minimization eliminating equality constraints Newton s method with equality constraints infeasible start
More informationThe Solution of Euclidean Norm Trust Region SQP Subproblems via Second Order Cone Programs, an Overview and Elementary Introduction
The Solution of Euclidean Norm Trust Region SQP Subproblems via Second Order Cone Programs, an Overview and Elementary Introduction Florian Jarre, Felix Lieder, Mathematisches Institut, Heinrich-Heine
More informationCopositive Programming and Combinatorial Optimization
Copositive Programming and Combinatorial Optimization Franz Rendl http://www.math.uni-klu.ac.at Alpen-Adria-Universität Klagenfurt Austria joint work with I.M. Bomze (Wien) and F. Jarre (Düsseldorf) IMA
More informationSta$s$cal models and graphs
Sta$s$cal models and graphs AFOSR FA9550-11-1-0305 Fundamental methodology and applica$ons Convex op$miza$on and graphs as underlying engines Examples: Sparse graph es$ma$on, latent-variable modeling,
More informationLU Factorization. Marco Chiarandini. DM559 Linear and Integer Programming. Department of Mathematics & Computer Science University of Southern Denmark
DM559 Linear and Integer Programming LU Factorization Marco Chiarandini Department of Mathematics & Computer Science University of Southern Denmark [Based on slides by Lieven Vandenberghe, UCLA] Outline
More informationExploiting Chordal Structure in Polynomial Ideals: A Gröbner Bases Approach
Exploiting Chordal Structure in Polynomial Ideals: A Gröbner Bases Approach The MIT Faculty has made this article openly available. Please share how this access benefits you. Your story matters. Citation
More informationConvex relaxation. In example below, we have N = 6, and the cut we are considering
Convex relaxation The art and science of convex relaxation revolves around taking a non-convex problem that you want to solve, and replacing it with a convex problem which you can actually solve the solution
More informationSemidefinite Programming Duality and Linear Time-invariant Systems
Semidefinite Programming Duality and Linear Time-invariant Systems Venkataramanan (Ragu) Balakrishnan School of ECE, Purdue University 2 July 2004 Workshop on Linear Matrix Inequalities in Control LAAS-CNRS,
More informationGraph structure in polynomial systems: chordal networks
Graph structure in polynomial systems: chordal networks Pablo A. Parrilo Laboratory for Information and Decision Systems Electrical Engineering and Computer Science Massachusetts Institute of Technology
More information