Disentangling Orthogonal Matrices
|
|
- Lindsey Anis Atkins
- 5 years ago
- Views:
Transcription
1 Disentangling Orthogonal Matrices Teng Zhang Department of Mathematics, University of Central Florida Orlando, Florida 3286, USA arxiv: v2 [math.oc] 8 May 206 Amit Singer 2 Department of Mathematics and The Program in Applied and Computational Mathematics (PACM), Princeton University Princeton, New Jersey 08544, USA Abstract Motivated by a certain molecular reconstruction methodology in cryo-electron microscopy, we consider the problem of solving a linear system with two unknown orthogonal matrices, which is a generalization of the well-known orthogonal Procrustes problem. We propose an algorithm based on a semi-definite programming (SDP) relaxation, and give a theoretical guarantee for its performance. Both theoretically and empirically, the proposed algorithm performs better than the naïve approach of solving the linear system directly without the orthogonal constraints. We also consider the generalization to linear systems with more than two unknown orthogonal matrices. Keywords: SDP relaxation, orthogonal Procrustes problem, cryo-em MSC[200] 5A24, 90C22. Introduction In this paper, we consider the following problem: given known matrices X, X 2 R N D and unknown orthogonal matrices V, V 2 O(D), recover V and V 2 from X 3 R N D defined by X 3 = X V + X 2 V 2. () A naïve approach would be solving () while dropping the constraints of orthogonality on V and V 2. This linear system has ND linear constraints and 2D 2 unknown variables, therefore, this approach can recover V and V 2 when N 2D. The question is, can we develop an algorithm that takes the constraints of orthogonality into consideration, so that it is able to recover V and teng.zhang@ucf.edu 2 amits@math.princeton.edu Preprint submitted to Elsevier May 0, 206
2 V 2 when N < 2D, and more stably when the observation X 3 is contaminated by noise? The associated least squares problem min V,V 2 O(D) X V + X 2 V 2 X 3 2 F (2) can be considered as a generalization of the well-known orthogonal Procrustes problem []: min X V X 2 2 F, (3) V O(D) with the main difference being that the minimization in (2) is over two orthogonal matrices instead of just one in (3). Although the orthogonal Procrustes problem has a closed form solution using the singular value decomposition, the problem (2) does not enjoy this property. Still, (2) can be reformulated so that it belongs to a wider class of problems called the little Grothendieck problem [2], which again belongs to QO-OC (Quadratic Optimization under Orthogonality Constraints) considered by Nemirovski [3]. QO-OCs have been well studied and include many important problems as special cases, such as Max-Cut [4] and generalized orthogonal Procrustes [5, 6, 7] min X i V i X j V j 2 F, V,...,V n O(D) i,j n which has applications to areas such as psychometrics, image and shape analysis and biometric identification. The non-commutative little Grothendieck problem is defined by: min n V,...,V n O(D) i,j= tr(c ij V i V T j ). (4) Problem (2) can be considered as a special case of (4) when n = 3. The argument is as follows. For convenience, we homogenize () by introducing a slack unitary variable V 3 O(D) and consider the augmented linear system X V + X 2 V 2 + X 3 V 3 = 0 (5) Clearly, if (V, V 2, V 3 ) is a solution to (5), then the pair (V V T 3, V 2 V T 3 ) is a solution to the original linear system (). The least squares formulation corresponding to (5) is min X V + X 2 V 2 + X 3 V 3 2 F. (6) V,V 2,V 3 O(D) Let C R 3D 3D be a Hermitian matrix with the (i, j) th D D block given by C ij = X T i X j. The least squares problem (6) is equivalent to min 3 V,V 2,V 3 O(D) i,j= which is the little Grothendieck problem (4) with n = 3. tr(c ij V j V T i ), (7) 2
3 .. Motivation Our problem arises naturally in single particle reconstruction (SPR) from cryo-electron microscopy (EM), where the goal is to determine the 3D structure of a macromolecule complex from 2D projection images of identical, but randomly oriented, copies of the macromolecule. Zvi Kam [8] showed that the spherical harmonic expansion coefficients of the 3D molecule, when arranged as matrices, can be estimated from 2D projection images up to an orthogonal matrix (for each degree of spherical harmonics). Based on this observation, Bhamre et al. [9] recently proposed Orthogonal Replacement (OR), an orthogonal matrix retrieval procedure in which cryo-em projection images are available for two unknown structures ϕ () and ϕ (2) whose difference ϕ (2) ϕ () is known. It follows from Kam s theory that we are given the spherical harmonic expansion coefficients of ϕ () and ϕ (2) up to an orthogonal matrix, and their difference. Then the problem of recovering the spherical harmonic expansion coefficients of ϕ () and ϕ (2) is reduced to the mathematical problem (). If () can be solved for smaller N, then we can reconstruct ϕ () and ϕ (2) with higher resolution. The cryo-em application serves as the main motivation of this paper. We refer the reader to [9] for further details regarding the specific application to cryo-em. 2. Algorithm and Main result The little Grothendieck problem and QO-OCs are generally intractable, for example, it is well-known that the Max-Cut Problem is NP-hard. Many approximation algorithms have been proposed and analyzed [4, 3, 0,, 2, 2], and the principle of these algorithms is to apply a semi-definite programming (SDP) relaxation followed by a rounding procedure. The SDP can be solved in polynomial time (for any finite precision). Based on the same principle, we relax the problem (7) to an SDP as follows. Let H R 3D 3D be a Hermitian matrix with the (i, j) th D D block given by H ij = V i Vj T. Then (7) is equivalent to min tr(ch), H 0,H ii=i,rank(h)=d where H 0 denotes that H is a positive semidefinite matrix. The only constraint which is non-convex is the rank constraint. Dropping it leads to the following SDP: min tr(ch), subject to H 0 and H ii = I. (8) If the solution satisfies rank(h) = D, then V, V 2 and V 3 are extracted by applying Cholesky decomposition to H. Notice that the solution to (5) is not unique: if (V, V 2, V 3 ) satisfies (5), then for any V O(D), the triplet (V V, V 2 V, V 3 V ) satisfies (5) as well. Although the solution to (5) is not unique, the solution to the original problem (2) is uniquely given by (V V3 T, V 2 V3 T ). 3
4 When the solution matrix H has rank greater than D, we employ the rounding procedure described in [2]: assuming that Cholesky decomposition of H gives H = U T U, where U R 3D 3D and U = (U, U 2, U 3 ), where U i R 3D D. Then we let the approximate solutions to (2) be the nearest orthogonal matrices to U T U 3 and U T 2 U 3 by the procedure in [3]. The procedure is as follows: for a matrix Z with SVD decomposition Z = U Z Σ Z V Z, the nearest orthogonal matrix is U Z V Z = Z(Z T Z) 2. Specifically, our solutions to U T U 3 are given by V = U T U 3 (U T 3 U U T U 3 ) 2 and V2 = U T 2 U 3 (U T 3 U 2 U T 2 U 3 ) 2. (9) 2.. Notations Without loss of generality, we assume that V = V 2 = I in the theoretical analysis. We define the subspace L and the set A as follows: L = {x C 3D : x = (v, v, v) for some v C D }, A = {x C 3D : x = (c v, c 2 v, c 3 v) for some c, c 2, c 3 C and v C D }. For any matrix A C m n, Sp(A) represents the subspace spanned by the row vectors of A. L + L represents the subspace {x + y : x L, y L }, P L is the projector to the subspace L and P L (L ) denotes the orthogonal projection of the subspace L onto L, i.e., {P L x : x L }. For any matrix A with 3D columns, we use the the same notation with subscripts to denote its decomposition into three parts: for A R m 3D, A = (A, A 2, A 3 ), where A i R m D Main results The main contribution of this paper is a particular theoretical guarantee for the SDP approach to return a solution of rank D and recover V and V 2 exactly. We start with a theorem that controls the lower bound of the objective function in (8). Theorem 2.. For any U = (U, U 2, U 3 ) C k 3D that satisfies U i C k D, Ui T U i = I, ( UPL XU T F c(x) ) 2 2D F. 200D Here c(x) is a constant defined as follows: c(x) = min v A L, v = Xv. Based on Theorem 2. and XU T 2 F = tr(ch), this paper proves that when N D +, the SDP method recovers the orthogonal matrices for generic cases, i.e., the property holds for (X, X 2 ) that lies in a dense open subset of R N D R N D. The statement is as follows, and part (b) shows that the SDP method is stable to noise. 4
5 Theorem 2.2. (a) For generic X, X 2 R N D with N D +, the SDP method recovers V and V 2 exactly. (b) For generic X, X 2 R N D with N D+, suppose that the input matrices satisfy ˆX i X i F ɛ for i 3, then the SDP method recovers {V i } 2 i= approximately in the sense that the error between the recovered orthogonal matrix ˆV i and the true orthogonal matrix V i, ˆV i V i F, is bounded above by O(ɛ 2 2D ). This result shows that the SDP method successfully recovers the orthogonal matrices as long as N D+, compared with the stringent requirement N 2D for the naïve least squares approach. The condition N D+ is nearly optimal. In (), there are ND constraints and D(D ) variables. Hence, it is impossible to recover V and V 2 when N < D. While existing works on the sensitivity of SDP problems give similar bounds as the one in Theorem 2.2(b) [4], this bound is a worst-case guarantee and in practice, ˆV i V i F is usually much smaller, as shown in Table 2 of Section 3. We also remark that Theorem 2.2 can be generalized to the complex case the proof applies to the case of unitary matrices as well. For the complex case, there are 2ND constraints and 2D 2 degrees of freedom. Therefore, it is impossible to recover V and V 2 when N < D. Moreover, we suspect that the recovery is impossible even for N = D, which would suggest that the sufficient condition N D + in Theorem 2.2(a) is also necessary: in fact, it is easy to verify the impossibility of recovering V and V 2 when N = D = Generalization A natural generalization of () is the following problem: given known matrices X, X 2,..., X K R N D and unknown orthogonal matrices V, V 2,..., V K O(D), recover {V i } K i= from X K = K i= X i V i. (0) For this generalized problem, the SDP method is formulated as follow: let H R KD KD be a Hermitian matrix with the (i, j) th D D block given by H ij = V i Vj T. Then the SDP method solves min tr(ch), subject to H 0 and H ii = I for all i K, () where H ii represents the (i, i) th D D block. Then we extract the orthogonal matrices by the procedure (9). For this generalized problem and its associated SDP approach, we have the following theoretical guarantee. Theorem 2.3. For generic {X i } K i= RN D, where N (K 2)D +, and X K = ( K i= X iv i )VK T, then the SDP method recovers {V i} K i= exactly. Theorem 2.2(a) can be considered as a special case of Theorem 2.3 when K = 3. However, for K > 3, the condition N (K 2)D + is not close to 5
6 optimal. Since (0) has ND constraints and D(D )(K )/2 variables, the information-theoretic limit is N = (D )(K )/2. Simulations in Section 3 also show that the SDP approach empirically recovers the orthogonal matrices even when N is smaller than (K 2)D +. However, the theoretical guarantee in Theorem 2.3 is still more powerful than the least squares approach of solving min {Vi} K K i= RD D i= X iv i X K 2 F, which requires N (K )D + to recover {V i } K i=. 3. Numerical Experiments In this section, we compare several methods for solving () and (0) on artificial data sets. The data sets are generated as follows: {X i } K i= are random matrices with i.i.d standard Gaussian entries N (0, ), {V i } K i= are random orthogonal matrices, and X K is generated by (0). We compare the following five methods:. The SDP relaxation approach (SDP) described in Section The naïve least squares approach (LS): min {V i} K i= K X K i= X i V i 2 F 3. Since the convex hull of the set of orthogonal matrices is the set of matrices with operator norm non greater than one, we can strengthen the LS approach by constraining its domain (C-LS): min {V i} K i= K X K i= X i V i 2 F, subject to V i, i K 4. This is an approach suggested to us by Afonso Bandeira. Let us start with the case K = 3. If V 3 = V V2 T, then from (), X 3 V2 T = X V 3 + X 2 and X 3 V T = X + X 2 V3 T. Then we solve the expanded least squares problem based on these three equations (LS+): min X 3 X V X 2 V 2 2 F + X 3 V2 T X V 3 X 2 2 F + X 3 V T X X 2 V3 T 2 F V,V 2,V 3 In general, for K 3, this method can be written as min tr(ch 2 ), subject to H = H T and H ii = I for all i K, H R KD KD where H ij represents the ij-th D D block of H. 5. The LS+ approach with constraints on the operator norm of H ij (C-LS+): min tr(ch 2 ), subject to H = H T, H ii = I and H ij for all i, j K. H R KD KD 6
7 To compare the SDP/LS+/C-LS+ approaches, we summarize their objective functions and their constraints in Table. There are two main differences. First, the objective functions are different. However, since tr(ch) = 0 if and only if tr(ch 2 ) = 0 (considering C 0 and H 0), this difference does not affect the property of exact recovery. Second, the constraints of the SDP approach is more restrictive than that of the C-LS+ approach (H ii = H jj = I and H 0 imply H ij ), which is more restrictive than the C-LS approach. This observation partially justifies the fact that SDP performs better than C-LS+, and C-LS+ performs better than C-LS. However, these interpretations do not justify the empirical finding in Figures 2 and 3 that C-LS+ and SDP behave very similarly in the absence of noise. We leave the explanation of this observation as an open question. Table : Comparison between SDP, LS+ and C-LS+ approaches. objective function common constraint other constraints SDP tr(ch) H 0 LS+ tr(ch 2 ) H ii = I, H = H T C-LS+ tr(ch 2 ) H ij Among these optimization approaches, the LS method has an explicit solution by decomposing it into D sub-problems, where each is a regression problem that estimates KD regression parameters. All other methods are convex and can be solved by CVX [5], where the default solver SeDuMi is used [6]. While the LS+ approach can also be written as a least squares problem with an explicit solution, this problem is not decomposable (unlike the LS method). When the solution matrices of LS/C-LS are not orthogonal, they are rounded to the nearest orthogonal matrices using the approach in [3]. The rounding procedure of LS+/C-LS+ is the same as that of the SDP method in (9). In the first simulation, we aim to find the size of N such that the orthogonal matrices be exactly recovered by the suggested algorithms for K = 3. We let D = 0 or 20 and choose various values for N, and record the mean recovery error of V (in Frobenius norm) over 50 repeated simulations in Figure 2. The performance of LS verifies our theoretical analysis: it recovers the orthogonal matrices for N 2D. LS fails when N < 2D because the null space of [X, X 2 ] is nontrivial and there are infinite solutions. Besides, LS+ succeeds when N 3D/2. SDP and C-LS+ are the best approaches and they succeed when N D +, which verifies Theorem 2.2. In the second simulation, we test the stability of the suggested algorithms when K = 3 and the measurement matrix X 3 is contaminated elementwisely by Gaussian noise N (0, σ 2 ). We use the setting N = 2, 6, 22, D = 0 and σ = 0.0 or 0. and record the mean recovery error over 50 runs in Table 2, which shows that the SDP relaxation approach is more stable to noise than competing approaches. This motivates our interest in studying the SDP approach. In the third simulation, we compare these methods for K = 5 and D = 5, 0. The results are shown in Figure 3. This simulation verifies Theorem 2.3 by 7
8 Figure : The dependence of the mean recovery error with respect to N, when D = 0 (left panel) and D = 20 (right panel). The y-axis represents the mean recovery error of V in Frobenius norm. Table 2: The mean recovery error in the noisy setting for K = 3 and D = 0. N σ SDP C-LS+ LS+ C-LS LS
9 Figure 2: The dependence of the mean recovery error with respect to N, when K = 5, D = 5 (left panel) and K = 5, D = 0 (right panel). The y-axis represents the mean recovery error of V in Frobenius norm. showing that the SDP approach successfully recovers the orthogonal matrices for N (K 2)D +. Indeed, the empirical performance of the SDP approach is even better: it recovers {V i } 5 i= at N = 2 and 25 respectively, which are smaller than (K 2)D +. Compared with LS/LS+/C-LS, the SDP and C- LS+ approaches recover the orthogonal matrices with smaller N. At last, we record the running time for all approaches in Table 3. Although the running time is not the main focus of this paper, and CVX is not optimized for the approaches, this table get a sense of the running times. Table 3 clearly shows that the LS approach is much faster than the other approaches, and the SDP approach is consistently faster than C-LS+. We suspect that it is due to the fact that SDP has fewer constraint, even though the constraint of SDP is more restrictive than that of C-LS+. Table 3: The average running time (in seconds) when σ = 0., K = 3, D = 0, N = 5 (first row) and σ = 0., K = 5, D = 0, N = 35 (second row). SDP C-LS+ LS+ C-LS LS Proof of the main result 4.. Auxiliary lemmas This section presents two auxiliary lemmas and their proofs. Lemma 4.. Given linear maps f, f 2 : C n V, where Im(f ) Im(f 2 ) = {0} and Im(f ) + Im(f 2 ) = V, for any given subspace L V with dim(l) n, 9
10 there exists a nonzero vector x L, c, c 2 C and v C n such that x = c f (v) + c 2 f 2 (v). Proof. WLOG assume that dim(l) = n. Let {e,..., e n } be the coordinate system in C n and {f (e ),..., f (e n ), f 2 (e ),..., f 2 (e n )} be the coordinate system in V, and assume that L = Im(C), where C C 2n n, ( ) C C =, and C, C 2 C n n. C 2 Then it is sufficient to show that there exist y C n and λ C such that C y = λc 2 y, or C 2 y = 0. When C 2 is invertible, let y to be any eigenvector of C2 C, then C y = λc 2 y for some λ C. When C 2 is not invertible, C 2 is singular and we can find y such that C 2 y = 0. Therefore, Lemma 4. is proved. Lemma 4.2. AB F A B F. (2) Proof. To prove it, assume that B = (b,..., b ), then AB = (Ab,..., Ab n ). Adding up the inequalities Ab i 2 A 2 b i 2 for i n, (2) is proved Proof of Theorem 2. The main idea of the proof is to find Ũ () such that its row space intersects A, and approximately contained in the row space of U. Then XU T can be bounded below by c(x) and the approximation error. By SVD decomposition of UP L, there exists a unitary matrix U 0 R k k such that the columns of U 0 UP L are orthogonal and have norms σ σ 2... σ k. We write U 0 U into two parts: ( ) U () U 0 U =, U (2) where U () C d 3D and U (2) C (k d) 3D. Then for any unit vector x in C d, Let ker(u () ) = {x CD : U () x = 0} and xu () P L σ d, and (3) U (2) P L σ d+. (4) Ũ () = (U () P ker(u () ), U () 2 P ker(u () ), U () 3 P ker(u () ) ). The argument for Theorem 2. is based on the following three properties. Their proofs are is deferred to Section
11 . Ũ () U () is small: Ũ () U () 4 σ d+. (5) 2. The row space of Ũ () contains vectors from A if σ d 4 σ d+ > Assume σ i = 0 for k + i 2D +, then max σ i 8 σ i+ ( σ i 2D 8 )4D. (6) By property 2 and property 3, there exists d such that there is a vector in both A and Sp(Ũ () ), denoted by v Ũ (), where v C d and v =. Then by (3) and (5), v Ũ () P L v U () P L Ũ () U () σ d 4 σ d+. As a result, Then v U () P L v Ũ () P L U () Ũ () σ d 8 σ d+. XU () F = XP L U () XP L Ũ () v (σ d 8 σ d+ ) P L Ũ () v (σ d 8 σ d+ )c(x). (7) Combining it with (6) and XU F XU () F, Theorem 2. is proved Proof of the three properties The argument for the first property is as follows. For any x ker(u () ) with x =, U () (2) x = 0 and U x = U x =. Applying (4), we have U (2) (2) U 2 = U (2) [I, I, 0] T = U 2 P L [I, I, 0] T U 2 P L [I, I, 0] 2σ d+ and U (2) (2) 2 x (U U (2) 2 )x 2σ d+. As a result, U () 2 x 2 σ d+. Since it holds for any unit vector x, U () 2 P ker(u () ) 2 σ d+. Similarly, U () 3 P ker(u () ) 2 σ d+. Property is concluded as follows: Ũ () U () = (0, U () 2 P ker(u () ), U () 3 P ker(u () )) 4 σ d+. (8) The prove the second property, combining (5) with (3), this implies that when σ d 4 σ d+ > 0, dim(sp(ũ () )) = dim(sp(u () )) dim(sp(u () )) = dim(ker(u () ) ). Let f, f 2 : ker(u () ) C 3D defined by f (x) = [x, x, 0] and f 2 (x) = [x, 0, x], then Lemma 4. implies that Sp(Ũ () ) contains a vector in A. Finally, the argument for the last property (6) is as follows. Assume max i 2D σ i 8 σ i+ = a, then a < 3 since a < max i 2D σ i U 3 i= U i = 3. Applying the definition of a with σ 2D+ = 0, σ 2D a. Then σ 2D a + 8 σ 2D a + 8 a a 2 < 00a 2. Now use the argument of induction. Assume that for d 2D we have σ d 00a 2d 2D, then σ d a + 8 σ d a + 80a 2d 2D < 00a 2d 2D. Therefore, σ < 00a 2 2D and a ( σ 00 )22D. 2D Since σ i= σ2 i /2D > UP L F /2D, (6) is proved.
12 4.3. Proof of Theorem 2.2 Part (a) follows from the result in part (b) with ɛ = 0: then we have UP L F = 0 and U = U 2 = U 3. Therefore, H = (I, I, I) T (I, I, I) and has a rank of D, which means that the SDP approach recovers V and V 2. In the proof of part (b), we represent the noisy setting in (8) by ˆX, Ĉ and Ĥ, the clean setting by X, C and H, and write the Cholesky decompositions of H and Ĥ by H = U T U and Ĥ = Û T Û. Since Û i T Û i = I, Û 3 i= Ûi = 3, (2) implies XÛ T F ˆXÛ T F (X ˆX)Û T F 3 X ˆX F 9ɛ (9) and following the same argument, ˆXU T F 9ɛ + XU T F. (20) Since ˆXÛ T F ˆXU T F and XU T F = 0, (20) implies Combining (9), (2) with Theorem 2., and we have ˆXÛ T F 9ɛ. (2) 9ɛ ˆXÛ F XÛ F 9ɛ c(x)( ÛP L F 2D 200D )2 9ɛ (22) ( 8ɛ ) 2 2D ÛP L F 200D. (23) c(x) Then ( 8ɛ ) 2 2D Û Û2 F = ÛP L (I, I, 0)T F ÛP L F (I, I, 0) T 400D, c(x) and Û T ( 8ɛ ) 2 2D Û2 I F = Û T (Û Û) F Û Û Û F 400D. c(x) Since the post-processing step of V 2 in (9) is a continuous and differentiable function with respect to Û T Û2 at Û T Û2 = I, the recovered orthogonal matrix V 2 has an error bounded above by a constant that is in the order of O(ɛ 2 2D ) if c(x) 0. To prove part (b), it suffices to show that when N > D +, c(x) > 0 for generic X. Let XP L = [C, C 2 ], then this problem is equivalent to: given generic matrices C, C 2 R N D with entries i.i.d. sampled from N(0, ), there do not exist y C D and λ C such that y 0, C y = λc 2 y, or C 2 y = 0. This is a rectangular generalized eigenvalue problem. It is a well-known fact that, for generic C, C 2 with more rows than columns, there is no eigenvalue for such eigenvalue problems [7]. 2
13 4.4. Proof of Theorem 2.3 Proof. The proof can be divided into two steps. First, we prove that if the SDP fails to recover {V i } K i=, then X satisfies a condition (27). Second, we show that this condition does not hold for generic X. In the proof, we define linear operators f,..., f K : R D R KD f i (x) = (0,..., 0, x, 0,..., 0). ith position We also defined L slightly differently from previous sections (but following the same principle) by L = {z R KD : z = (x, x,..., x) for some x R D }. Since X X K = 0, ( ) ( ) tr H(I,..., I) T (I,..., I) = tr (I,..., I)H(I,..., I) T = X +...+X K 2 F = 0. Considering that tr(hc) 0 for any H 0, if the solution to the SDP approach is not uniquely given by (I,..., I) T (I,..., I), then there exists H (I,..., I) T (I,..., I) such that tr(ch) = 0. Let H = U T U for a matrix U R k KD, then we have Sp(X) Sp(U), (24) for a U = (U, U 2,..., U K ) such that U i R k KD and Ui T U i = I for i K, and {U i } K i= are not all the same. (25) Let d = k dim(sp(u) L ), then we can choose an unitary matrix U 0 C k k such that Ũ = U 0U has the following property: the first k d rows of Ũ lie in L and the span of the other d rows only intersects L at origin. Since we can choose U from H = U T U up to a multiplication of any orthogonal matrix in O(k), WLOG we may assume that U 0 = I and U has this property. Define U i and U i2 by ( ) Ui U i =, where U i C (k d) D and U i2 C d D, U i2 and let Ũ = (U, U 2,..., U K ) R (k d) KD and Û = (U 2, U 22,..., U K2 ) R d KD. Since the first (k d) rows of U lie in L, U = U 2 =... = U K. Combining it with Ui2 T U i2 = Ui T U i Ui T U i = I Ui T U i, we have Sp(U 2 ) = Sp(U 22 ) =... = Sp(U K2 ). Let L 0 = Sp(U 2 ) R D, then Sp(U) = Sp(Û) + Sp(Ũ), where Sp(Ũ) L, Sp(Û) L = {0}, Sp(Û) f (L 0 ) + f 2 (L 0 ) f K (L 0 ), and dim(sp(û)) dim(l 0 ). Let ˆL = P L Sp(Û), then ˆL (f (L 0 ) + f 2 (L 0 ) f K (L 0 )) L and dim(ˆl) = dim(sp(û)) dim(l 0). (26) 3
14 Applying (24), if the SDP method fails to recover {V i } K i=, then there exists ˆL P L (Sp(X) ) and L 0 R D such that (26) is satissp(ũ) L fied. For a generic X, P L (Sp(X) ) is a KD D N-dimension subspace P L (Sp(X) ) in L, and we denote the manifold of all such KD D N- dimension subspaces by M. Applying the previous analysis, all P L (Sp(X) ) with property (26) is in the set M M 2... M D, where M d is the set of all L M such that: There exists a d-dimensional subspace L 0 R D and a d-dimensional subspace ˆL such that L ˆL, and ˆL f (L 0 ) + f 2 (L 0 ) f K (L 0 ). (27) Indeed, every element in M d can be obtained by the following procedure: Pick L 0 by a random d-dimensional subspace in R D. Pick ˆL by a random d-dimensional subspace in [f (L 0 ) + f 2 (L 0 ) f K (L 0 )] L, a (K )d-dimensional space. Choose a random N d-dimensional subspace L in the KD D d- dimensional space (ˆL + L ). Then L = L + ˆL is an element in M d. This constructive procedure gives a mapping from G(D, d) G(Kd d, d) G(KD D d, N d) to M d M, where G(D, d) represents the manifold of all d-dimensional subspaces in R D. The dimension of G(D, d) G(Kd d, d) G(KD D d, N d) is (D d)d + (K 2)d 2 + N(KD D N d). Since N (K 2)D +, the dimensionality of M is N(KD D N) and it is larger than the dimension of G(D, d) G(Kd d, d) G(KD D d, N d). Applying this difference in dimensionality, we can prove that (M M 2... M D ) c is a open dense set in M by constructing a ɛ-net for M d as follows. First, find a ɛ-net for the L 0 in step, and denote the ɛ-net by {L (i) 0 }N i= with N = O(ɛ (D d)d ). Second, for each L (i) 0, find a ɛ-net for the corresponding ˆL in step 2, denoted by {ˆL (i)(j) } N2 j= with N 2 = O(ɛ (K 2)d2 ). Finally, for each ˆL (i)(j), find a ɛ-net for the corresponding L in step 3, denoted by {L (i)(j)(k) } N3 k= with N 3 = O(ɛ N(KD D N d) ). By analyzing the continuity of the constructive procedure, {L (i)(j)(k) } N,N2,N3 i,j,k= is a cɛ-net of M d for some c > 0. Denote the haar measure of M by µ, then the measure on a ɛ-ball is O(ɛ N(KD N N) ). By counting the measure of the ɛ-net and letting ɛ 0, we have µ(m d ) = 0. In addition, by the continuity in the constructive procedure, M d is closed and M c d is open. Therefore, M M 2... M D is a closed set with measure zero, which means that its complement in M is an open and dense set. Therefore, the SDP approach recovers {V i } K i= for generic P L (Sp(X) ) M and also for general X. 4
15 Acknowledgment The authors were partially supported by Award Number R0GM from the NIGMS, FA and FA from AFOSR, LTR DTD from the Simons Foundation, and the Moore Foundation Data- Driven Discovery Investigator Award. References References [] J. Gower, G. Dijksterhuis, Procrustes Problems, Oxford Statistical Science Series. [2] A. S. Bandeira, C. Kennedy, A. Singer, Approximating the little Grothendieck problem over the orthogonal and unitary groups, Mathematical Programming (206) 43. [3] A. Nemirovski, Sums of random symmetric matrices and quadratic optimization under orthogonality constraints, Mathematical Programming 09 (2-3) (2007) [4] M. X. Goemans, D. P. Williamson, Improved approximation algorithms for maximum cut and satisfiability problems using semidefinite programming, J. ACM 42 (6) (995) [5] J. Gower, Generalized procrustes analysis, Psychometrika 40 () (975) [6] J. Ten Berge, Orthogonal procrustes rotation for two or more matrices, Psychometrika 42 (2) (977) [7] A. Shapiro, J. D. Botha, Dual algorithm for orthogonal procrustes rotations, SIAM Journal on Matrix Analysis and Applications 9 (3) (988) [8] Z. Kam, The reconstruction of structure from electron micrographs of randomly oriented particles, Journal of Theoretical Biology 82 () (980) [9] T. Bhamre, T. Zhang, A. Singer, Orthogonal matrix retrieval in cryoelectron microscopy, 2th IEEE International Symposium on Biomedical Imaging (ISBI 205). [0] A. M.-C. So, J. Zhang, Y. Ye, On approximating complex quadratic optimization problems via semidefinite programming relaxations, Math. Program. 0 () (2007)
16 [] A.-C. So, Moment inequalities for sums of random matrices and their applications in optimization, Mathematical Programming 30 () (20) [2] A. Naor, O. Regev, T. Vidick, Efficient rounding for the noncommutative grothendieck inequality, in: Proceedings of the Forty-fifth Annual ACM Symposium on Theory of Computing, STOC 3, 203, pp [3] J. B. Keller, Closest unitary, orthogonal and Hermitian operators to a given operator, Mathematics Magazine 48 (4) (975) [4] Y.-L. Cheung, H. Wolkowicz, Sensitivity analysis of semidefinite programs without strong duality, Technical report, University of Waterloo, Waterloo, Ontario, 204. Available online at HTML/204/06/446.html. [5] M. Grant, S. Boyd, CVX: Matlab software for disciplined convex programming, version 2., (Mar. 204). [6] H. R. Tütüncü, C. K. Toh, J. M. Todd, Solving semidefinite-quadraticlinear programs using sdpt3, Mathematical Programming 95 (2) (2003) [7] J. W. Demmel, A. Edelman, The dimension of matrices (matrix pencils) with given Jordan (Kronecker) canonical forms, Linear Algebra and its Applications 230 (995)
Some Optimization and Statistical Learning Problems in Structural Biology
Some Optimization and Statistical Learning Problems in Structural Biology Amit Singer Princeton University, Department of Mathematics and PACM January 8, 2013 Amit Singer (Princeton University) January
More informationThree-dimensional structure determination of molecules without crystallization
Three-dimensional structure determination of molecules without crystallization Amit Singer Princeton University, Department of Mathematics and PACM June 4, 214 Amit Singer (Princeton University) June 214
More informationOrientation Assignment in Cryo-EM
Orientation Assignment in Cryo-EM Amit Singer Princeton University Department of Mathematics and Program in Applied and Computational Mathematics July 24, 2014 Amit Singer (Princeton University) July 2014
More informationMath 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination
Math 0, Winter 07 Final Exam Review Chapter. Matrices and Gaussian Elimination { x + x =,. Different forms of a system of linear equations. Example: The x + 4x = 4. [ ] [ ] [ ] vector form (or the column
More informationChapter 3 Transformations
Chapter 3 Transformations An Introduction to Optimization Spring, 2014 Wei-Ta Chu 1 Linear Transformations A function is called a linear transformation if 1. for every and 2. for every If we fix the bases
More informationLecture 9: Low Rank Approximation
CSE 521: Design and Analysis of Algorithms I Fall 2018 Lecture 9: Low Rank Approximation Lecturer: Shayan Oveis Gharan February 8th Scribe: Jun Qi Disclaimer: These notes have not been subjected to the
More informationPHASE RETRIEVAL OF SPARSE SIGNALS FROM MAGNITUDE INFORMATION. A Thesis MELTEM APAYDIN
PHASE RETRIEVAL OF SPARSE SIGNALS FROM MAGNITUDE INFORMATION A Thesis by MELTEM APAYDIN Submitted to the Office of Graduate and Professional Studies of Texas A&M University in partial fulfillment of the
More informationHandout 6: Some Applications of Conic Linear Programming
ENGG 550: Foundations of Optimization 08 9 First Term Handout 6: Some Applications of Conic Linear Programming Instructor: Anthony Man Cho So November, 08 Introduction Conic linear programming CLP, and
More informationMaths for Signals and Systems Linear Algebra in Engineering
Maths for Signals and Systems Linear Algebra in Engineering Lectures 13 15, Tuesday 8 th and Friday 11 th November 016 DR TANIA STATHAKI READER (ASSOCIATE PROFFESOR) IN SIGNAL PROCESSING IMPERIAL COLLEGE
More informationMatrix Mathematics. Theory, Facts, and Formulas with Application to Linear Systems Theory. Dennis S. Bernstein
Matrix Mathematics Theory, Facts, and Formulas with Application to Linear Systems Theory Dennis S. Bernstein PRINCETON UNIVERSITY PRESS PRINCETON AND OXFORD Contents Special Symbols xv Conventions, Notation,
More informationSparse and Low-Rank Matrix Decompositions
Forty-Seventh Annual Allerton Conference Allerton House, UIUC, Illinois, USA September 30 - October 2, 2009 Sparse and Low-Rank Matrix Decompositions Venkat Chandrasekaran, Sujay Sanghavi, Pablo A. Parrilo,
More informationPreliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012
Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.
More informationA direct formulation for sparse PCA using semidefinite programming
A direct formulation for sparse PCA using semidefinite programming A. d Aspremont, L. El Ghaoui, M. Jordan, G. Lanckriet ORFE, Princeton University & EECS, U.C. Berkeley Available online at www.princeton.edu/~aspremon
More informationOctober 25, 2013 INNER PRODUCT SPACES
October 25, 2013 INNER PRODUCT SPACES RODICA D. COSTIN Contents 1. Inner product 2 1.1. Inner product 2 1.2. Inner product spaces 4 2. Orthogonal bases 5 2.1. Existence of an orthogonal basis 7 2.2. Orthogonal
More informationSparse PCA with applications in finance
Sparse PCA with applications in finance A. d Aspremont, L. El Ghaoui, M. Jordan, G. Lanckriet ORFE, Princeton University & EECS, U.C. Berkeley Available online at www.princeton.edu/~aspremon 1 Introduction
More informationMATH 240 Spring, Chapter 1: Linear Equations and Matrices
MATH 240 Spring, 2006 Chapter Summaries for Kolman / Hill, Elementary Linear Algebra, 8th Ed. Sections 1.1 1.6, 2.1 2.2, 3.2 3.8, 4.3 4.5, 5.1 5.3, 5.5, 6.1 6.5, 7.1 7.2, 7.4 DEFINITIONS Chapter 1: Linear
More informationLinear Algebra. Session 12
Linear Algebra. Session 12 Dr. Marco A Roque Sol 08/01/2017 Example 12.1 Find the constant function that is the least squares fit to the following data x 0 1 2 3 f(x) 1 0 1 2 Solution c = 1 c = 0 f (x)
More informationReview of Some Concepts from Linear Algebra: Part 2
Review of Some Concepts from Linear Algebra: Part 2 Department of Mathematics Boise State University January 16, 2019 Math 566 Linear Algebra Review: Part 2 January 16, 2019 1 / 22 Vector spaces A set
More informationLecture 1. 1 Conic programming. MA 796S: Convex Optimization and Interior Point Methods October 8, Consider the conic program. min.
MA 796S: Convex Optimization and Interior Point Methods October 8, 2007 Lecture 1 Lecturer: Kartik Sivaramakrishnan Scribe: Kartik Sivaramakrishnan 1 Conic programming Consider the conic program min s.t.
More informationCompressed Sensing and Robust Recovery of Low Rank Matrices
Compressed Sensing and Robust Recovery of Low Rank Matrices M. Fazel, E. Candès, B. Recht, P. Parrilo Electrical Engineering, University of Washington Applied and Computational Mathematics Dept., Caltech
More informationApproximating the little Grothendieck problem over the orthogonal and unitary groups
Math. Program., Ser. A 06 60:433 475 DOI 0.007/s007-06-0993-7 FULL LENGTH PAPER Approximating the little Grothendieck problem over the orthogonal and unitary groups Afonso S. Bandeira Christopher Kennedy
More informationI. Multiple Choice Questions (Answer any eight)
Name of the student : Roll No : CS65: Linear Algebra and Random Processes Exam - Course Instructor : Prashanth L.A. Date : Sep-24, 27 Duration : 5 minutes INSTRUCTIONS: The test will be evaluated ONLY
More informationarxiv: v1 [math.oc] 26 Sep 2015
arxiv:1509.08021v1 [math.oc] 26 Sep 2015 Degeneracy in Maximal Clique Decomposition for Semidefinite Programs Arvind U. Raghunathan and Andrew V. Knyazev Mitsubishi Electric Research Laboratories 201 Broadway,
More informationConditions for Robust Principal Component Analysis
Rose-Hulman Undergraduate Mathematics Journal Volume 12 Issue 2 Article 9 Conditions for Robust Principal Component Analysis Michael Hornstein Stanford University, mdhornstein@gmail.com Follow this and
More informationThroughout these notes we assume V, W are finite dimensional inner product spaces over C.
Math 342 - Linear Algebra II Notes Throughout these notes we assume V, W are finite dimensional inner product spaces over C 1 Upper Triangular Representation Proposition: Let T L(V ) There exists an orthonormal
More informationFall 2016 MATH*1160 Final Exam
Fall 2016 MATH*1160 Final Exam Last name: (PRINT) First name: Student #: Instructor: M. R. Garvie Dec 16, 2016 INSTRUCTIONS: 1. The exam is 2 hours long. Do NOT start until instructed. You may use blank
More informationMAT 585: Johnson-Lindenstrauss, Group testing, and Compressed Sensing
MAT 585: Johnson-Lindenstrauss, Group testing, and Compressed Sensing Afonso S. Bandeira April 9, 2015 1 The Johnson-Lindenstrauss Lemma Suppose one has n points, X = {x 1,..., x n }, in R d with d very
More informationCovariance Matrix Estimation for the Cryo-EM Heterogeneity Problem
Covariance Matrix Estimation for the Cryo-EM Heterogeneity Problem Amit Singer Princeton University Department of Mathematics and Program in Applied and Computational Mathematics July 25, 2014 Joint work
More informationPrincipal Component Analysis
Machine Learning Michaelmas 2017 James Worrell Principal Component Analysis 1 Introduction 1.1 Goals of PCA Principal components analysis (PCA) is a dimensionality reduction technique that can be used
More informationOn Optimal Frame Conditioners
On Optimal Frame Conditioners Chae A. Clark Department of Mathematics University of Maryland, College Park Email: cclark18@math.umd.edu Kasso A. Okoudjou Department of Mathematics University of Maryland,
More informationVector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis.
Vector spaces DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_fall17/index.html Carlos Fernandez-Granda Vector space Consists of: A set V A scalar
More informationThe value of a problem is not so much coming up with the answer as in the ideas and attempted ideas it forces on the would be solver I.N.
Math 410 Homework Problems In the following pages you will find all of the homework problems for the semester. Homework should be written out neatly and stapled and turned in at the beginning of class
More informationAcyclic Semidefinite Approximations of Quadratically Constrained Quadratic Programs
2015 American Control Conference Palmer House Hilton July 1-3, 2015. Chicago, IL, USA Acyclic Semidefinite Approximations of Quadratically Constrained Quadratic Programs Raphael Louca and Eilyan Bitar
More informationReview problems for MA 54, Fall 2004.
Review problems for MA 54, Fall 2004. Below are the review problems for the final. They are mostly homework problems, or very similar. If you are comfortable doing these problems, you should be fine on
More informationLecture Notes 9: Constrained Optimization
Optimization-based data analysis Fall 017 Lecture Notes 9: Constrained Optimization 1 Compressed sensing 1.1 Underdetermined linear inverse problems Linear inverse problems model measurements of the form
More informationSupremum of simple stochastic processes
Subspace embeddings Daniel Hsu COMS 4772 1 Supremum of simple stochastic processes 2 Recap: JL lemma JL lemma. For any ε (0, 1/2), point set S R d of cardinality 16 ln n S = n, and k N such that k, there
More informationExact Low-rank Matrix Recovery via Nonconvex M p -Minimization
Exact Low-rank Matrix Recovery via Nonconvex M p -Minimization Lingchen Kong and Naihua Xiu Department of Applied Mathematics, Beijing Jiaotong University, Beijing, 100044, People s Republic of China E-mail:
More informationECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis
ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis Lecture 7: Matrix completion Yuejie Chi The Ohio State University Page 1 Reference Guaranteed Minimum-Rank Solutions of Linear
More informationThe convex algebraic geometry of linear inverse problems
The convex algebraic geometry of linear inverse problems The MIT Faculty has made this article openly available. Please share how this access benefits you. Your story matters. Citation As Published Publisher
More information1. Let m 1 and n 1 be two natural numbers such that m > n. Which of the following is/are true?
. Let m and n be two natural numbers such that m > n. Which of the following is/are true? (i) A linear system of m equations in n variables is always consistent. (ii) A linear system of n equations in
More informationAssignment #10: Diagonalization of Symmetric Matrices, Quadratic Forms, Optimization, Singular Value Decomposition. Name:
Assignment #10: Diagonalization of Symmetric Matrices, Quadratic Forms, Optimization, Singular Value Decomposition Due date: Friday, May 4, 2018 (1:35pm) Name: Section Number Assignment #10: Diagonalization
More informationMATH 1120 (LINEAR ALGEBRA 1), FINAL EXAM FALL 2011 SOLUTIONS TO PRACTICE VERSION
MATH (LINEAR ALGEBRA ) FINAL EXAM FALL SOLUTIONS TO PRACTICE VERSION Problem (a) For each matrix below (i) find a basis for its column space (ii) find a basis for its row space (iii) determine whether
More informationLift me up but not too high Fast algorithms to solve SDP s with block-diagonal constraints
Lift me up but not too high Fast algorithms to solve SDP s with block-diagonal constraints Nicolas Boumal Université catholique de Louvain (Belgium) IDeAS seminar, May 13 th, 2014, Princeton The Riemannian
More informationSparse Solutions of an Undetermined Linear System
1 Sparse Solutions of an Undetermined Linear System Maddullah Almerdasy New York University Tandon School of Engineering arxiv:1702.07096v1 [math.oc] 23 Feb 2017 Abstract This work proposes a research
More informationTHE SINGULAR VALUE DECOMPOSITION MARKUS GRASMAIR
THE SINGULAR VALUE DECOMPOSITION MARKUS GRASMAIR 1. Definition Existence Theorem 1. Assume that A R m n. Then there exist orthogonal matrices U R m m V R n n, values σ 1 σ 2... σ p 0 with p = min{m, n},
More informationLinear Algebra. Min Yan
Linear Algebra Min Yan January 2, 2018 2 Contents 1 Vector Space 7 1.1 Definition................................. 7 1.1.1 Axioms of Vector Space..................... 7 1.1.2 Consequence of Axiom......................
More informationFast Angular Synchronization for Phase Retrieval via Incomplete Information
Fast Angular Synchronization for Phase Retrieval via Incomplete Information Aditya Viswanathan a and Mark Iwen b a Department of Mathematics, Michigan State University; b Department of Mathematics & Department
More informationLinear Algebra- Final Exam Review
Linear Algebra- Final Exam Review. Let A be invertible. Show that, if v, v, v 3 are linearly independent vectors, so are Av, Av, Av 3. NOTE: It should be clear from your answer that you know the definition.
More informationLecture 1: Review of linear algebra
Lecture 1: Review of linear algebra Linear functions and linearization Inverse matrix, least-squares and least-norm solutions Subspaces, basis, and dimension Change of basis and similarity transformations
More informationUniversity of Colorado at Denver Mathematics Department Applied Linear Algebra Preliminary Exam With Solutions 16 January 2009, 10:00 am 2:00 pm
University of Colorado at Denver Mathematics Department Applied Linear Algebra Preliminary Exam With Solutions 16 January 2009, 10:00 am 2:00 pm Name: The proctor will let you read the following conditions
More information1 9/5 Matrices, vectors, and their applications
1 9/5 Matrices, vectors, and their applications Algebra: study of objects and operations on them. Linear algebra: object: matrices and vectors. operations: addition, multiplication etc. Algorithms/Geometric
More informationarxiv: v3 [cs.cv] 3 Sep 2013
GLOBAL REGISTRATION OF MULTIPLE POINT CLOUDS USING SEMIDEFINITE PROGRAMMING KUNAL N. CHAUDHURY, YUEHAW KHOO, AND AMIT SINGER arxiv:1306.5226v3 [cs.cv] 3 Sep 2013 ABSTRACT. Consider N points in R d and
More informationLimitations in Approximating RIP
Alok Puranik Mentor: Adrian Vladu Fifth Annual PRIMES Conference, 2015 Outline 1 Background The Problem Motivation Construction Certification 2 Planted model Planting eigenvalues Analysis Distinguishing
More informationOn the local stability of semidefinite relaxations
On the local stability of semidefinite relaxations Diego Cifuentes Department of Mathematics Massachusetts Institute of Technology Joint work with Sameer Agarwal (Google), Pablo Parrilo (MIT), Rekha Thomas
More information2. LINEAR ALGEBRA. 1. Definitions. 2. Linear least squares problem. 3. QR factorization. 4. Singular value decomposition (SVD) 5.
2. LINEAR ALGEBRA Outline 1. Definitions 2. Linear least squares problem 3. QR factorization 4. Singular value decomposition (SVD) 5. Pseudo-inverse 6. Eigenvalue decomposition (EVD) 1 Definitions Vector
More informationDS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.
DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1
More informationCS 229r: Algorithms for Big Data Fall Lecture 17 10/28
CS 229r: Algorithms for Big Data Fall 2015 Prof. Jelani Nelson Lecture 17 10/28 Scribe: Morris Yau 1 Overview In the last lecture we defined subspace embeddings a subspace embedding is a linear transformation
More informationLecture 7: Positive Semidefinite Matrices
Lecture 7: Positive Semidefinite Matrices Rajat Mittal IIT Kanpur The main aim of this lecture note is to prepare your background for semidefinite programming. We have already seen some linear algebra.
More informationCSCI 1951-G Optimization Methods in Finance Part 10: Conic Optimization
CSCI 1951-G Optimization Methods in Finance Part 10: Conic Optimization April 6, 2018 1 / 34 This material is covered in the textbook, Chapters 9 and 10. Some of the materials are taken from it. Some of
More informationEE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 2
EE/ACM 150 - Applications of Convex Optimization in Signal Processing and Communications Lecture 2 Andre Tkacenko Signal Processing Research Group Jet Propulsion Laboratory April 5, 2012 Andre Tkacenko
More informationMatrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A =
30 MATHEMATICS REVIEW G A.1.1 Matrices and Vectors Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A = a 11 a 12... a 1N a 21 a 22... a 2N...... a M1 a M2... a MN A matrix can
More informationB553 Lecture 5: Matrix Algebra Review
B553 Lecture 5: Matrix Algebra Review Kris Hauser January 19, 2012 We have seen in prior lectures how vectors represent points in R n and gradients of functions. Matrices represent linear transformations
More informationConvex sets, conic matrix factorizations and conic rank lower bounds
Convex sets, conic matrix factorizations and conic rank lower bounds Pablo A. Parrilo Laboratory for Information and Decision Systems Electrical Engineering and Computer Science Massachusetts Institute
More informationA Continuation Approach Using NCP Function for Solving Max-Cut Problem
A Continuation Approach Using NCP Function for Solving Max-Cut Problem Xu Fengmin Xu Chengxian Ren Jiuquan Abstract A continuous approach using NCP function for approximating the solution of the max-cut
More informationDesign of Spectrally Shaped Binary Sequences via Randomized Convex Relaxations
Design of Spectrally Shaped Binary Sequences via Randomized Convex Relaxations Dian Mo Department of Electrical and Computer Engineering University of Massachusetts Amherst, MA 3 mo@umass.edu Marco F.
More informationUniversity of Colorado Denver Department of Mathematical and Statistical Sciences Applied Linear Algebra Ph.D. Preliminary Exam January 23, 2015
University of Colorado Denver Department of Mathematical and Statistical Sciences Applied Linear Algebra PhD Preliminary Exam January 23, 2015 Name: Exam Rules: This exam lasts 4 hours and consists of
More informationMATH 315 Linear Algebra Homework #1 Assigned: August 20, 2018
Homework #1 Assigned: August 20, 2018 Review the following subjects involving systems of equations and matrices from Calculus II. Linear systems of equations Converting systems to matrix form Pivot entry
More informationThe Singular Value Decomposition
The Singular Value Decomposition We are interested in more than just sym+def matrices. But the eigenvalue decompositions discussed in the last section of notes will play a major role in solving general
More information8.1 Concentration inequality for Gaussian random matrix (cont d)
MGMT 69: Topics in High-dimensional Data Analysis Falll 26 Lecture 8: Spectral clustering and Laplacian matrices Lecturer: Jiaming Xu Scribe: Hyun-Ju Oh and Taotao He, October 4, 26 Outline Concentration
More informationLECTURE NOTE #11 PROF. ALAN YUILLE
LECTURE NOTE #11 PROF. ALAN YUILLE 1. NonLinear Dimension Reduction Spectral Methods. The basic idea is to assume that the data lies on a manifold/surface in D-dimensional space, see figure (1) Perform
More informationRank minimization via the γ 2 norm
Rank minimization via the γ 2 norm Troy Lee Columbia University Adi Shraibman Weizmann Institute Rank Minimization Problem Consider the following problem min X rank(x) A i, X b i for i = 1,..., k Arises
More informationDS-GA 1002 Lecture notes 10 November 23, Linear models
DS-GA 2 Lecture notes November 23, 2 Linear functions Linear models A linear model encodes the assumption that two quantities are linearly related. Mathematically, this is characterized using linear functions.
More informationarxiv: v1 [math.oc] 14 Oct 2014
arxiv:110.3571v1 [math.oc] 1 Oct 01 An Improved Analysis of Semidefinite Approximation Bound for Nonconvex Nonhomogeneous Quadratic Optimization with Ellipsoid Constraints Yong Hsia a, Shu Wang a, Zi Xu
More informationApplied Mathematics 205. Unit II: Numerical Linear Algebra. Lecturer: Dr. David Knezevic
Applied Mathematics 205 Unit II: Numerical Linear Algebra Lecturer: Dr. David Knezevic Unit II: Numerical Linear Algebra Chapter II.3: QR Factorization, SVD 2 / 66 QR Factorization 3 / 66 QR Factorization
More informationRelaxations and Randomized Methods for Nonconvex QCQPs
Relaxations and Randomized Methods for Nonconvex QCQPs Alexandre d Aspremont, Stephen Boyd EE392o, Stanford University Autumn, 2003 Introduction While some special classes of nonconvex problems can be
More informationBounds on the radius of nonsingularity
Towards bounds on the radius of nonsingularity, Milan Hladík Department of Applied Mathematics, Charles University, Prague and Institute of Computer Science of the Academy of Czech Republic SCAN 2014,
More informationLecture 2: Linear operators
Lecture 2: Linear operators Rajat Mittal IIT Kanpur The mathematical formulation of Quantum computing requires vector spaces and linear operators So, we need to be comfortable with linear algebra to study
More informationConvex Optimization M2
Convex Optimization M2 Lecture 8 A. d Aspremont. Convex Optimization M2. 1/57 Applications A. d Aspremont. Convex Optimization M2. 2/57 Outline Geometrical problems Approximation problems Combinatorial
More informationDimension reduction for semidefinite programming
1 / 22 Dimension reduction for semidefinite programming Pablo A. Parrilo Laboratory for Information and Decision Systems Electrical Engineering and Computer Science Massachusetts Institute of Technology
More informationConvexity of the Joint Numerical Range
Convexity of the Joint Numerical Range Chi-Kwong Li and Yiu-Tung Poon October 26, 2004 Dedicated to Professor Yik-Hoi Au-Yeung on the occasion of his retirement. Abstract Let A = (A 1,..., A m ) be an
More informationLecture notes: Applied linear algebra Part 1. Version 2
Lecture notes: Applied linear algebra Part 1. Version 2 Michael Karow Berlin University of Technology karow@math.tu-berlin.de October 2, 2008 1 Notation, basic notions and facts 1.1 Subspaces, range and
More informationBindel, Fall 2016 Matrix Computations (CS 6210) Notes for
1 Logistics Notes for 2016-08-29 General announcement: we are switching from weekly to bi-weekly homeworks (mostly because the course is much bigger than planned). If you want to do HW but are not formally
More informationLecture Notes 1: Vector spaces
Optimization-based data analysis Fall 2017 Lecture Notes 1: Vector spaces In this chapter we review certain basic concepts of linear algebra, highlighting their application to signal processing. 1 Vector
More informationElementary linear algebra
Chapter 1 Elementary linear algebra 1.1 Vector spaces Vector spaces owe their importance to the fact that so many models arising in the solutions of specific problems turn out to be vector spaces. The
More informationSPARSE signal representations have gained popularity in recent
6958 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 57, NO. 10, OCTOBER 2011 Blind Compressed Sensing Sivan Gleichman and Yonina C. Eldar, Senior Member, IEEE Abstract The fundamental principle underlying
More informationMAT Linear Algebra Collection of sample exams
MAT 342 - Linear Algebra Collection of sample exams A-x. (0 pts Give the precise definition of the row echelon form. 2. ( 0 pts After performing row reductions on the augmented matrix for a certain system
More information2. Every linear system with the same number of equations as unknowns has a unique solution.
1. For matrices A, B, C, A + B = A + C if and only if A = B. 2. Every linear system with the same number of equations as unknowns has a unique solution. 3. Every linear system with the same number of equations
More informationMatrix Factorizations
1 Stat 540, Matrix Factorizations Matrix Factorizations LU Factorization Definition... Given a square k k matrix S, the LU factorization (or decomposition) represents S as the product of two triangular
More informationLinear Algebra M1 - FIB. Contents: 5. Matrices, systems of linear equations and determinants 6. Vector space 7. Linear maps 8.
Linear Algebra M1 - FIB Contents: 5 Matrices, systems of linear equations and determinants 6 Vector space 7 Linear maps 8 Diagonalization Anna de Mier Montserrat Maureso Dept Matemàtica Aplicada II Translation:
More informationA Unified Theorem on SDP Rank Reduction. yyye
SDP Rank Reduction Yinyu Ye, EURO XXII 1 A Unified Theorem on SDP Rank Reduction Yinyu Ye Department of Management Science and Engineering and Institute of Computational and Mathematical Engineering Stanford
More informationNotes on Eigenvalues, Singular Values and QR
Notes on Eigenvalues, Singular Values and QR Michael Overton, Numerical Computing, Spring 2017 March 30, 2017 1 Eigenvalues Everyone who has studied linear algebra knows the definition: given a square
More informationapproximation algorithms I
SUM-OF-SQUARES method and approximation algorithms I David Steurer Cornell Cargese Workshop, 201 meta-task encoded as low-degree polynomial in R x example: f(x) = i,j n w ij x i x j 2 given: functions
More information1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det
What is the determinant of the following matrix? 3 4 3 4 3 4 4 3 A 0 B 8 C 55 D 0 E 60 If det a a a 3 b b b 3 c c c 3 = 4, then det a a 4a 3 a b b 4b 3 b c c c 3 c = A 8 B 6 C 4 D E 3 Let A be an n n matrix
More informationSingular Value Decompsition
Singular Value Decompsition Massoud Malek One of the most useful results from linear algebra, is a matrix decomposition known as the singular value decomposition It has many useful applications in almost
More informationBeyond Scalar Affinities for Network Analysis or Vector Diffusion Maps and the Connection Laplacian
Beyond Scalar Affinities for Network Analysis or Vector Diffusion Maps and the Connection Laplacian Amit Singer Princeton University Department of Mathematics and Program in Applied and Computational Mathematics
More informationGlossary of Linear Algebra Terms. Prepared by Vince Zaccone For Campus Learning Assistance Services at UCSB
Glossary of Linear Algebra Terms Basis (for a subspace) A linearly independent set of vectors that spans the space Basic Variable A variable in a linear system that corresponds to a pivot column in the
More informationMATH36001 Generalized Inverses and the SVD 2015
MATH36001 Generalized Inverses and the SVD 201 1 Generalized Inverses of Matrices A matrix has an inverse only if it is square and nonsingular. However there are theoretical and practical applications
More informationThe Singular Value Decomposition (SVD) and Principal Component Analysis (PCA)
Chapter 5 The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA) 5.1 Basics of SVD 5.1.1 Review of Key Concepts We review some key definitions and results about matrices that will
More informationConstrained optimization
Constrained optimization DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_fall17/index.html Carlos Fernandez-Granda Compressed sensing Convex constrained
More informationQALGO workshop, Riga. 1 / 26. Quantum algorithms for linear algebra.
QALGO workshop, Riga. 1 / 26 Quantum algorithms for linear algebra., Center for Quantum Technologies and Nanyang Technological University, Singapore. September 22, 2015 QALGO workshop, Riga. 2 / 26 Overview
More information