The total least squares problem in AX B. A new classification with the relationship to the classical works

Size: px
Start display at page:

Download "The total least squares problem in AX B. A new classification with the relationship to the classical works"

Transcription

1 Eidgenössische Technische Hochschule Zürich Ecole polytechnique fédérale de Zurich Politecnico federale di Zurigo Swiss Federal Institute of Technology Zurich The total least squares problem in AX B. A new classification with the relationship to the classical works I. Hnětynková, M. Plešinger, D.M. Sima, Z. Strakoš and S. Van Huffel Research Report No November 2010 Seminar für Angewandte Mathematik Eidgenössische Technische Hochschule CH-8092 Zürich Switzerland Faculty of Mathematics and Physics, Charles University, Prague and Institute of Computer Science, Academy of Sciences of the Czech Republic Seminar for Applied Mathematics, Department of Mathematics, ETH, Zurich, Switzerland, Institute of Computer Science, Academy of Sciences of the Czech Republic, and Faculty of Mechatronics, Technical University of Liberec, Czech Republic ESAT-SCD, Department of Electrical Engineering, Katholieke Universiteit Leuven (KUL), Belgium

2 THE TOTAL LEAST SQUARES PROBLEM IN AX B. A NEW CLASSIFICATION WITH THE RELATIONSHIP TO THE CLASSICAL WORKS IVETA HNĚTYNKOVÁ, MARTIN PLEŠINGER, DIANA MARIA SIMA, ZDENĚK STRAKOŠ, AND SABINE VAN HUFFEL Abstract. The presented paper revisits the analysis of the total least squares (TLS) problem AX B with multiple right-hand sides given by Sabine Van Huffel and Joos Vandewalle, in the monograph: The Total Least Squares Problem: Computational Aspects and Analysis, SIAM Publications, Philadelphia The newly proposed classification is based on properties of the singular value decomposition of the extended matrix B A. It aims at identifying the cases when a TLS solution does or does not exist, and when the output computed by the classical TLS algorithm, given by Van Huffel and Vandewalle, is actually a TLS solution. The presented results on existence and uniqueness of the TLS solution reveal subtleties that were not captured in the known literature. Key words. total least squares (TLS), multiple right-hand sides, linear approximation problems, orthogonally invariant problems, orthogonal regression, errors-in-variables modeling. AMS subject classifications. 15A24, 15A60, 65F20, 65F Introduction. This paper focuses on the total least squares (TLS) formulation of the linear approximation problem with multiple right-hand sides AX B, A R m n, X R n d, B R m d, A T B 0, (1.1) or, equivalently, B A I d X 0. (1.2) We concentrate on the incompatible problem (1.1), i.e. R(B) R(A). The compatible case reduces to finding a solution of a system of linear algebraic equations. In TLS, contrary to the ordinary least squares, the correction is allowed to compensate for errors in the system (data) matrix A as well as in the right-hand side (observation) matrix B, and the matrices E and G are sought to minimize the Frobenius norm in min G E F subject to (A + E)X B + G. (1.3) X,E,G The research of I. Hnětynková and Z. Strakoš was supported by the research project MSM financed by MŠMT ČR, and by the GACR grant 201/09/0917. The research of M. Plešinger was supported by the GAAV grant IAA and by Institutional Research Plan AV0Z D. M. Sima, a postdoctoral fellow of the Fund for Scientific Research Flanders, and S. Van Huffel acknowledge the Research Council KUL: GOA MaNet, CoE EF/05/006 Optimization in Engineering (OPTEC), and the Belgian Federal Science Policy Office IUAP P6/04 (DYSCO, Dynamical systems, control and optimization, ). Faculty of Mathematics and Physics, Charles University, Prague and Institute of Computer Science, Academy of Sciences of the Czech Republic ({hnetynkova, strakos}@cs.cas.cz). Seminar for Applied Mathematics, Department of Mathematics, ETH, Zurich, Switzerland, Institute of Computer Science, Academy of Sciences of the Czech Republic, and Faculty of Mechatronics, Technical University of Liberec, Czech Republic (martin.plesinger@cs.cas.cz). Department of Electrical Engineering, ESAT-SCD, Katholieke Universiteit Leuven (KUL), and IBBT-KUL Future Health Department, Leuven, Belgium ({diana.sima, sabine.vanhuffel}@esat.kuleuven.be). 1

3 2 I. HNĚTYNKOVÁ, M. PLEŠINGER, D. M. SIMA, Z. STRAKOŠ, S. VAN HUFFEL Throughout the whole paper, any matrix X which solves the corrected system in (1.3) is called a TLS solution. Similarly to the ordinary least squares, we are often interested in TLS solutions minimal in the 2-norm and/or in the Frobenius norm. Mathematically equivalent problems have been independently investigated in several areas as orthogonal regression and errors-in-variables modeling, see 18, 19. It is worth noting that other norms than the Frobenius norm in (1.3) can also be relevant in practice, see, e.g., 20. The TLS problem (1.1) (1.3) has been investigated in its algebraic setting for decades, see the early works 6, 4, Section 6, 14. In 7 it is shown that even with d 1 (which gives Ax b, where b is an m-vector) the TLS problem may not have a solution and, when the solution exists, it may not be unique; see also 5, pp The classical book 17 introduces the generic nongeneric terminology representing the basic classification of TLS problems. If d 1, then the generic problems simply represent problems that have a (possibly nonunique) solution, whereas nongeneric problems do not have a solution in the sense of (1.3). This is no longer true for multiple right-hand sides, where d>1. The monograph 17 analyzes only two particular cases characterized by the special distribution of singular values of the extended matrix B A. The so called classical TLS algorithm presented in 17, however, for any A, B computes some output X. The relationship of this output to the original problem is not always clear. For d 1, the TLS problem does not have a solution when the collinearities among columns of A are stronger than the collinearities between R(A) and b; see 9, 10, 11 for a recent description. An analogous situation may occur for d>1, but here the difficulty can be caused for different columns of B by different subsets of columns of A. Therefore it is no longer possible to stay with the generic nongeneric classification of TLS problems. This is also the reason why the question remained open in 17. In this paper we try to fill this gap and investigate existence and uniqueness of the TLS solution with d>1 in full generality. The organization of this paper is as follows. Section 2 recalls some basic results. Section 3 introduces problems of what we call the 1st class. After recalling known results for two special distributions of singular values in Sections 3.1 and 3.2, we turn to the general case in Section 3.3. The new classification is introduced in Section 4. Section 5 introduces problems of the 2nd class. Section 6 links the new classification with the classical TLS algorithm from 17 and Section 7 concludes the paper. 2. Preliminaries. As usual, σ j (M) denotes the jth largest singular value, R(M) and N (M) the range and the null space, M F and M the Frobenius norm and the 2-norm of the given matrix M, respectively, and M denotes the Moore-Penrose pseudoinverse of M. Further, v denotes the 2-norm of the given vector v, I k R k k denotes the k-by-k identity matrix. In order to simplify the notation we assume, with no loss of generality, m n + d (otherwise, we can simply add zero rows). Consider the SVD of A, r rank(a), A U Σ (V ) T, (2.1) where (U ) 1 (U ) T,(V ) 1 (V ) T, Σ diag(σ 1,..., σ r, 0) Rm n, and σ 1... σ r > σ r+1... σ n 0. (2.2) Similarly, consider the SVD of B A, s rank(b A), B A UΣV T, (2.3)

4 THE TOTAL LEAST SQUARES PROBLEM IN AX B 3 where U 1 U T, V 1 V T, Σ diag(σ 1,...,σ s, 0) R m (n+d), and σ 1... σ s > σ s+1... σ n+d 0. (2.4) If s n+d (which implies r n), then Σ and Σ have no zero singular values. Among the singular values, a key role is played by σ n+1, where n represents the number of columns of A. In order to handle possible higher multiplicity of σ n+1, we introduce the following notation σ p σ n q > σ n q+1... σ n }{{} q σ n+1... σ n+e > σ n+e+1, (2.5) }{{} e where q singular values to the left and e 1 singular values to the right are equal to σ n+1, and hence q 0, e 1. For convenience we denote n q p. (Clearly σ p σ n q is not defined iff q n, similarly σ n+e+1 is not defined iff e d.) For an integer (not necessarily nonnegative) it will be useful to consider the following partitioning Σ n {}}{ d + {}}{ Σ ( ) 1 Σ ( ) 2 m, V n {}}{ d + {}}{ V ( ) 11 V ( ) V ( ) 21 V ( ) d n, (2.6) where Σ ( ) 1 R m (n ), Σ ( ) 2 R m (d+ ), and V ( ) 11 R d (n ), V ( ) R d (d+ ), V ( ) 21 R n (n ), V ( ) R n (d+ ). When 0, the partitioning conforms to the fact that B A is created by A appended by the matrix B with d columns and in this case the upper index is omitted, Σ 1 Σ (0) 1, etc. The classical analysis of the TLS problem with a single right-hand side (d 1) in 7, and the theory developed in 17 were based on relationships between the singular values of A and B A. For d 1, in particular, σ n > σ n+1 represents a sufficient (but not necessary) condition for the existence and uniqueness of the solution. In order to extend this condition to the case d>1, the following generalization of 7, Theorem 4.1 is useful. Theorem 2.1. Let (2.1) be the SVD of A and (2.3) the SVD of B A with the partitioning given by (2.6), m n + d, 0. If σ n > σ n +1, (2.7) then σ n > σ n +1. Moreover, V ( ) is of full row rank equal to d, and V ( ) 21 is of full column rank equal to (n ). The first part follows immediately from the interlacing theorem for singular values 17, Theorem 2.4, p. 32 (see also 13). For the proof of the second part see 21, Lemma 2.1 or 17, Lemma 3.1, pp (Please note the different ordering of the partitioning of V in 21, 17.) We start our analysis with the following definition. Definition 2.2 (Problems of the 1st class and of the 2nd class). Consider a TLS problem (1.1) (1.3), m n + d. Let (2.3) be the SVD of B A with the partitioning given by (2.6). Take q, where q is the left multiplicity of σ n+1 given by (2.5).

5 4 I. HNĚTYNKOVÁ, M. PLEŠINGER, D. M. SIMA, Z. STRAKOŠ, S. VAN HUFFEL If is of full row rank d, then we call (1.1) (1.3) a TLS problem of the 1st class. If is rank deficient (i.e. has linearly dependent rows), then we call (1.1) (1.3) a TLS problem of the 2st class. The set of all problems of the 1st class will be denoted by F. The set of all problems of the 2nd class will be denoted by S. 3. Problems of the 1st class. For d 1, the right singular vector subspace corresponding to the smallest singular value σ n+1 of b A contains for a TLS problem of the 1st class a singular vector with a nonzero first component. Consequently, the TLS problem has a (possibly nonunique) solution. As we will see, for d > 1 an analogous property does not hold. The TLS problem of the 1st class with d>1 may not have a solution. First we recall known results for two special cases of problems of the 1st class Problems of the 1st class with unique TLS solution. Consider a TLS problem of the 1st class. Assume that σ n > σ n+1, i.e. q 0 (p n). Setting q 0 in (2.6), V is a square (and nonsingular) matrix. Define the correction matrix G E U 0 Σ2 V T U Σ 2 V T V T. (3.1) Clearly, G E F ( n+d jn+1 σ2 j )1/2, and the corrected matrix B + G A + E represents, by the Eckart-Young-Mirsky theorem 1, 8, the unique rank n approximation of B A with minimal G E in the Frobenius norm. The columns of the matrix V T V T T represent a basis for the null space of the corrected matrix B + G A + E UΣ 1 V11 V T 21. T Since V is square and nonsingular, B + G A + E I d V V 1 which gives the uniquely determined TLS solution 0, X TLS X (0) V V 1. (3.2) We summarize these observations in the following theorem, see 17, Theorem 3.1, pp Theorem 3.1. Consider a TLS problem of the 1st class. If σ n > σ n+1, (3.3) then with the partitioning of the SVD of B A given by (2.6), q 0, V R d d is square and nonsingular, and (3.2) represents the unique TLS solution of the problem (1.1) (1.3) with the corresponding correction G E given by (3.1). Theorem 2.1 gives the following corollary. Corollary 3.2. Let (2.1) be the SVD of A and (2.3) the SVD of B A with the partitioning given by (2.6), m n + d, 0. If σ n > σ n+1, (3.4)

6 THE TOTAL LEAST SQUARES PROBLEM IN AX B 5 then (1.1) (1.3) is a problem of the 1st class, σ n > σ n+1, and (3.2) represents the unique TLS solution of the problem (1.1) (1.3) with the corresponding correction matrix G E given by (3.1). We see that (3.4) represents a sufficient condition for the existence and uniqueness of the TLS solution of the problem (1.1) (1.3). This condition is, however, intricate. It may look as the key to the analysis of the TLS problem, in particular when one considers the following corollary of the interlacing theorem for singular values and Theorem 2.1; see 17, Corollary 3.4, p. 65. Corollary 3.3. Let (2.1) be the SVD of A and (2.3) the SVD of B A with the partitioning given by (2.6), m n + d, q 0. Then the following conditions are equivalent: (i) σ n q > σ n q+1... σ n+d, (ii) σ n q > σ n q+1... σ n+d and is of (full row) rank d. In the following discussion we restrict ourselves to the single right-hand side case. The condition (i) implies that the TLS problem is of the 1st class. If d 1 and q 0, then (i) reduces to (3.4) and the statement of Corollary 3.3 says that σ n > σ n+1 if and only if σ n > σ n+1 and 1, 0,..., 0 T v n+1 0. In order to show the difficulty and motivate the classification in the sequel, we now consider all remaining possibilities for the case d 1. It should be, however, understood that they go beyond the problems of the 1st class and the unique TLS solution. If σ n σ n+1, then it may happen either σ n > σ n+1 and i T 1 v n+1 0, which means that the TLS problem is not of the 1st class and it does not have a solution, or σ n σ n+1. In the latter case, depending on the relationship between σ n q and σ n q+1... σ n+1 for some q>0, see Corollary 3.3, the TLS problem may have a nonunique solution, if the TLS problem is of the 1st class (see the next section), or the solution may not exist. We see that an attempt to base the analysis on the relationship between σ n and σ n+1 becomes very involved. The situation becomes more transparent with the use of the core problem concept from 11. For any linear approximation problem Ax b (we still consider d 1) there are orthogonal matrices P, R such that where: P T b A R b1 A A, (3.5) (i) A 11 is of minimal dimensions and A is of maximal dimensions (A may also have zero number of rows and/or columns) over all orthogonal transformations of b A yielding the structure (3.5) of zero and nonzero blocks. Suppose b R(A) has nonzero projections on exactly l left singular vector subspaces of A corresponding to distinct (nonzero) singular values. Then among all decompositions of the form (3.5) the minimally dimensioned A 11 is l l if Ax b is compatible, and (l + 1) l if Ax b is incompatible (see 11, Theorem 2.2). (ii) All singular values of A 11 are simple and nonzero, all singular values of b 1 A 11 are simple and, since b R(A), nonzero (recall that we consider only the incompatible problems), (iii) first components of all right singular vectors of b 1 A 11 are nonzero, (iv) σ min (A 11 ) > σ min (b 1 A 11 ). Moreover, singular values of A 11 strictly interlace singular values of b 1 A 11,

7 6 I. HNĚTYNKOVÁ, M. PLEŠINGER, D. M. SIMA, Z. STRAKOŠ, S. VAN HUFFEL see 11, Section 3. The minimally dimensioned subproblem A 11 x 1 b 1 is then called the core problem within Ax b. The SVD of the block structured matrix on the right-hand side in (3.5) can be obtained as a direct sum of the SVD decompositions of the blocks b 1 A 11 and A, just by extending the singular vectors corresponding to the first block by zeros on the bottom and the singular vectors corresponding to the second block by zeros on the top. Consequently, considering the special structure of the orthogonal transformation diag(1,r) in (3.5), which does not change the first components of the right singular vectors, all right singular vectors of b A with nonzero first components correspond to the block b 1 A 11, and all right singular vectors of b A with zero first component correspond to A. Moreover, σ n σ min (A) min{σ min (A 11 ), σ min (A )}, σ n+1 σ min (b A) min{σ min (b 1 A 11 ), σ min (A )}. We will review all possible situations: Case 1: σ n > σ n+1. This happens if and only if σ min (A ) > σ min (b 1 A 11 ) σ n+1 which is equivalent to the existence of the unique TLS solution. Case 2: σ min (A ) σ n σ n+1. Here we have to distinguish two cases: Case 2a: σ min (A) σ min (b A) σ min (b 1 A 11 ). This guarantees the existence of the (minimum norm) TLS solution. All singular values of A equal to σ min (A) are the singular values of the block A. Consequently, the multiplicity of σ min (b A) is larger by one than the multiplicity of σ min (A). Case 2b: σ min (A) σ min (b A) < σ min (b 1 A 11 ). Then the multiplicities of σ min (A) and σ min (b A) are equal, all right singular vectors of b A corresponding to σ min (b A) have zero first components, and the TLS solution does not exist. Summarizing, the TLS solution exists if and only if either σ min (A) > σ min (b A), or σ min (A) σ min (b A) with different multiplicities for σ min (A) and σ min (b A). In terms of the singular values of subblocks in the core reduction (3.5), σ min (A ) > σ min (b 1 A 11 ) TLS solution exists and is unique, σ min (A )σ min (b 1 A 11 ) TLS solution exists and is not unique, σ min (A ) < σ min (b 1 A 11 ) TLS solution does not exist. If the TLS solution exists, then the minimum norm TLS solution can always be computed, and it is automatically given by the core problem formulation. If the TLS solution does not exist, then the core problem formulation gives the solution equivalent to the minimum norm nongeneric solution constructed in 17. We will see that in the multiple right-hand sides case the situation is much more complicated Problems of the 1st class with nonunique TLS solutions a special case. Consider a TLS problem of the 1st class. Assume that e d in (2.5), i.e. let all the singular values starting from σ n q+1 σ p+1 be equal, σ 1... σ p > σ p+1... σ n+1... σ n+d 0. (3.6)

8 THE TOTAL LEAST SQUARES PROBLEM IN AX B 7 The case q 0 (p n) reduces to the problem with unique TLS solution discussed in Section 3.1. If q n (p 0), i.e. σ 1... σ n+d, then the columns of B A are mutually orthogonal and B A T B A σ1 2I n+d. Then it seems meaningless to approximate B by the columns of A, and we will get consistently with 17 the trivial solution X TLS 0 (this case does not satisfy the nontriviality assumption A T B 0 in (1.1)). Therefore in this section the interesting case is represented by n > q > 0 (0 < p < n). We first construct the solution minimal in norm. Since Rd (q+d) is of full row rank, there exists an orthogonal matrix Q R (q+d) (q+d) such that Q 0 Γ v p+1,..., v n+d Q Y Z, (3.7) where Γ R d d is square and nonsingular. Such an orthogonal matrix Q can be obtained, e.g., using the LQ decomposition of. Consider the partitioning Q Q 1 Q 2, where Q 2 R (q+d) d has d columns. Then the columns of Q 2 form an orthonormal basis of the subspace spanned by the columns of T, Q 1 R (q+d) q is an orthonormal basis of its orthogonal complement, and Γ Z Q 2, ΓQT 2. (3.8) Define the correction matrix T Γ Γ G E B A (3.9) Z Z T UΣV T Q 2 Q T 2 σ n+1 up+1,...,u n+d Q2 Q T T 2 vp+1,...,v n+d, where u j and v j represent left and right singular vectors of the matrix B A, respectively. If σ p+1... σ n+d 0, then the correction matrix is a zero matrix (σ n+1 0) and the problem is compatible, thus we consider σ p+1... σ n+d > 0. Note that with the choice of any other matrix Q Q 1 Q 2 giving a decomposition of the form (3.7), Q 2 represents an orthonormal basis of the subspace spanned by the columns of T, and therefore Q 2 Q 2Ψ for some orthogonal matrix Ψ R d d. Consequently, (3.9) is uniquely determined independently on the choice of Q in (3.7). Clearly, G E F σ n+1 Q 2 Q T 2 F σ n+1 d and the corrected matrix T ) B + G A + E B A ( I n+d Γ Z Γ Z represents the rank n approximation of B A such that the Frobenius norm of the correction matrix G E is minimal, by the Eckart-Young-Mirsky theorem. The columns of the matrix Γ T Z T T represent a basis for the null space of the corrected matrix B + G A + E. Since Γ is square and nonsingular, I B + G A + E d ZΓ 1 0,

9 8 I. HNĚTYNKOVÁ, M. PLEŠINGER, D. M. SIMA, Z. STRAKOŠ, S. VAN HUFFEL which gives the TLS solution X TLS ZΓ 1 Y Z Q T Q 0 Γ 1 X (q). (3.10) This can be expressed as X TLS ( A T A σ 2 n+1i n ) A T B, see 17, Theorem 3.10, pp The solution (3.10) and the correction (3.9) do not depend on the choice of the matrix Q in (3.7). We summarize these observations in the following theorem (see 17, Theorem 3.9, pp ). Theorem 3.4. Consider a TLS problem of the 1st class. Let (2.3) be the SVD of B A with the partitioning given by (2.6), q < n, p n q. If σ p > σ p+1... σ n+d, (3.11) then (3.10) represents a TLS solution X TLS of the problem (1.1) (1.3). This is the unique solution of the minimal Frobenius norm and 2-norm, with the corresponding unique correction matrix G E given by (3.9). Using Corollary 3.3 we get σ p > σ p+1... σ n+d, (3.) which represents a sufficient condition for the existence of the TLS solution of the TLS problem (1.1) (1.3) minimal in the Frobenius norm and the 2-norm. The correction matrix minimal in the Frobenius norm can be in this special case constructed from any d vectors selected among q + d columns v p+1,..., v n+d (or their orthogonal linear transformation) of the matrix V such that their top d-subvectors create a d-by-d square nonsingular matrix. The equality of the last q+d singular values ensures that the Frobenius norm of the corresponding correction matrix is still equal to σ n+1 d. It can be shown, that for any such choice a norm of the corresponding solution X is larger than or equal to the norm of X (q) given by (3.10), and any such X represents a TLS solution. Consequently, the special TLS problem satisfying (3.6) has infinitely many solutions Problems of the 1st class the general case. Here we consider a TLS problem of the 1st class with a general distribution of singular values. We will discuss only the remaining cases not covered in the previous two sections, i.e., n q>0 (0 p<n, recall that p n q) and e<d, giving σ 1... σ p > σ p+1... σ n+1... σ n+e > σ n+e+1... σ n+d 0 (note that σ p does not exist for q n (p 0)). We will see that in this general case the problem (1.1) (1.3) may not have a solution. We try to construct a TLS solution with the same approach as in Section 3.2, and we will show that it may fail. Since, with the partitioning (2.6), q, the matrix Rd (q+d) is of full row rank, there exists an orthogonal matrix Q R (q+d) (q+d) such that Q 0 Γ v p+1,..., v n+d Q Y Z, (3.13)

10 THE TOTAL LEAST SQUARES PROBLEM IN AX B 9 where Γ R d d is square and nonsingular. With the partitioning Q Q 1 Q 2, where Q 1 R (q+d) q, Q 2 R (q+d) d, the columns of Q 2 form an orthonormal basis of the subspace spanned by the columns of T, and Γ Z Q 2, Following 17, it is tempting to define the correction matrix ΓQT 2. (3.14) T Γ Γ G E B A (3.15) Z Z T UΣV T Q 2 Q T 2 u p+1,..., u n+d diag(σp+1,...,σ n+d )Q 2 Q T T 2 vp+1,...,v n+d, which differs from (3.9) because the diagonal factor is no longer a scalar multiple of the identity matrix. Analogously to the previous section, the matrix (3.15) is uniquely determined independently on the choice of Q in (3.13). The columns of the matrix Γ T Z T T are in the null space of the corrected matrix T ) B + G A + E B A ( I n+d Γ Z Γ Z. (3.16) In general the columns of Γ T Z T T do not represent a basis for the null space of the corrected matrix. If A is not of full column rank, the extended matrix B A has a zero singular value with the corresponding right singular vector having the first d entries equal to zero. Such a right singular vector is in the null space of the corrected matrix but it can not be obtained as a linear combination of the columns of Γ T Z T T. Since Γ is square and nonsingular, and we can construct B + G A + E I d ZΓ 1 0, X (q) ZΓ 1. (3.17) The matrices (3.17) and (3.15) do not depend on the choice of Q in (3.13). The matrix X (q) given by (3.17) is a natural generalization of X (q) given by (3.10). The classical TLS algorithm 15, 16 (see also 17) applied to a TLS problem of the 1st class returns as output the matrix X (q) given by (3.17) with the matrices G, E given by (3.15). We will show, however, that X (q) is not necessarily a TLS solution. We first focus on the question whether there exists another correction Ẽ, G corresponding to the last q + d columns of V that makes the corrected system compatible. Such a correction can be constructed analogously to (3.13) by considering an orthogonal matrix Q Q 1 Q 2 such that Q Ω v p+1,...,v Γ n+d Q, (3.18) Ỹ Z

11 10 I. HNĚTYNKOVÁ, M. PLEŠINGER, D. M. SIMA, Z. STRAKOŠ, S. VAN HUFFEL where Γ R d d is nonsingular and Ω is a matrix not necessarily equal to zero. Then define the correction matrix G Ẽ B A T Γ Γ. (3.19) Z Z The corrected system (A + Ẽ)X B + G is compatible and the matrix X Z Γ 1 ( Q 2 QT 2 ) (3.20) solves this corrected system. The columns of Γ T Z T T have to be in the null space of the corrected matrix B + G A + Ẽ. As above, they do not necessarily represent a basis of this null space. Now we show that X (q) does not necessarily represent a TLS solution, i.e., the Frobenius norm of the correction matrix (3.15) need not be minimal. This can be illustrated by a simple example. Let q n and e < d. Then in (3.13) we set Q T T. (Notice that V ( ) 11 and V ( ) 21 in the partitioning (2.6) vanish for q n.) Therefore T T 0 Id I n 0, i.e., Γ I d, Z 0, which gives from (3.13) G E B 0, and, analogously, X (q) 0, see (3.17). If we solve the same problem in the ordinary least squares sense, then the corresponding correction matrix is Ḡ Ē (AA I)B 0 having in general smaller Frobenius norm than G E B 0, given by (3.15). Therefore the constructed matrix X (q) given by (3.17) does not, in general, represent a TLS solution. Summarizing, the classical TLS algorithm of Van Huffel computes for TLS problems of the 1st class the output (3.2), (3.10), or (3.17), which are formally analogous, but with different relationship to the TLS solution. While (3.2) and (in the particular case of a very special distribution of the singular values) (3.10) represent TLS solutions (having minimal Frobenius and 2-norm), the interpretation of (3.17) remains unclear. The partitioning of the set F of TLS problems of the 1st class according to the conditions valid in (3.2), (3.10), and (3.17) is unsatisfactory. In particular, apart from the simple case (3.2) and the very special case (3.10) we do not know whether a TLS solution exists. 1 We will therefore develop a different partitioning of the set F in Section 4. First we briefly discuss some properties of matrices X (q) and X Note on the norms of matrices X (q) and X. It is obvious that X (q) given by (3.17) is a special case of X given by (3.20). The following Lemma 3.5 gives simple formulas for the Frobenius norm and 2-norm of X. Lemma 3.6 shows that X (q) has the minimal norms among all X of the form (3.20). The proofs are fully analogous to the proofs of 17, Theorems 3.6 and 3.9. Lemma 3.5. Let Γ T Z T T R (n+d) d have orthonormal columns and assume Γ R d d is nonsingular. Then the matrix X Z Γ 1 has the norms X 2 F Γ 1 2 F d, and X 2 1 σ2 min ( Γ) σmin 2 ( Γ), (3.21) 1 The problems in the set F are in 17 called generic. Since a problem in this set may not have a TLS solution, we will not further use the generic-nongeneric terminology.

12 THE TOTAL LEAST SQUARES PROBLEM IN AX B 11 where σ min ( Γ) is the minimal singular value of Γ. Lemma 3.6. Consider X (q) ZΓ 1 given by (3.13) (3.17) and X Z Γ 1 given by (3.18) (3.20). Then X F X (q) F, and X X (q). (3.) Moreover, equality holds for the Frobenius norms if and only if X X (q). These lemmas can be easily seen as follows. A matrix X of the form (3.20) is going to be minimal in the Frobenius or the 2-norm when Γ 1 F is minimized or σ min ( Γ) σ d ( Γ) is maximized, respectively. The minimization/maximization are with respect to the orthogonal matrix Q which is considered a free variable, with the constraint that Γ has to be nonsingular. The interlacing theorem for singular values applied to the matrices Ω Γ Q and Γ gives ( σ j (Γ) σ j ( )σ j Ω Γ ) σ j ( Γ), j 1,..., d, with all the inequalities becoming equalities if and only if Ω 0. The minimum for the 2-norm is reached when the smallest singular values are equal, i.e., σ d (Γ) σ d ( Γ). Note that there can be more than one matrix of the form (3.20) reaching the minimum of the 2-norm. If the corrected matrix (A+Ẽ) has linearly dependent columns, then the corrected system with the correction G Ẽ of the form (3.19) can have more than one solution. The following lemma shows that under some additional assumptions on the structure of Q, the matrix (A + Ẽ) is of full column rank, and therefore the matrix X of the form (3.20) is the unique solution of the corrected system. (Note that the correction (3.15) is a special case of the correction (3.19).) Lemma 3.7. Consider a TLS problem of the 1st class. Let G Ẽ be the correction matrix given by (3.19) and let X be the matrix given by (3.20). If Q in (3.18) has the block diagonal form Q diag(q,i d e ), where Q R (q+e) (q+e) is an orthogonal matrix, then (A + Ẽ) is of full column rank and X represents the unique solution of the corrected system (A + Ẽ) X B + G. Proof. Since Q diag(q,i d e ) has the block diagonal structure, B A UΣV T U I p Q I m n d Σ ( Ip 0 V 0 Q ) T ŪΣ V T, i.e. ŪΣ V T represents the SVD of B A with Ū ū 1,...,ū m, V v1,..., v n+d Ω Γ Ỹ Z. Using this SVD, the corrected matrix can be written as B + G A + Ẽ ū 1,..., ū n diag(σ1,...,σ n ) 11 Ω 21 Ỹ T.

13 I. HNĚTYNKOVÁ, M. PLEŠINGER, D. M. SIMA, Z. STRAKOŠ, S. VAN HUFFEL If σ n 0, then G Ẽ 0 and the original system is compatible, i.e. R(B) R(A); therefore assume σ n > 0. From the CS decomposition of V it follows that since Γ is square nonsingular, the matrix 21 Ỹ is square nonsingular. Since ū 1,...,ū n is of full column rank, the matrix (A + Ẽ) ū 1,..., ū n diag(σ1,...,σ n ) is of full column rank. The matrix X is then the unique solution of the corrected system (A + Ẽ) X B + G. We will see in the next section that the form Q diag(q,i d e ) appears in a natural way. 4. Partitioning of the set of problems of the 1st class. We will base our partitioning and the subsequent classification of TLS problems with multiple righthand sides on the following theorem. Theorem 4.1. Consider a TLS problem of the 1st class. Let (2.3) be the SVD of B A with the partitioning given by (2.6), q n, where q is the left multiplicity of σ n+1 given by (2.5), p n q. Consider an orthogonal matrix Q such that Q Ω Ỹ where Q 1 R (q+d) q, Q 2 R (q+d) d, and define 21 Γ, Q Q1 Q2, (4.1) Z Ỹ T G Ẽ B A Γ Z Γ Z T (4.2) u p+1,...,u n+d diag(σp+1,..., σ n+d ) Q 2 QT 2 vp+1,..., v n+d T. Then the following two assertions are equivalent: (i) There exists an orthonormal matrix Ψ R d d, such that Q Q diag(i q, Ψ) has the block diagonal structure Q 0 Q R (q+d) (q+d), Q R (q+e) (q+e), (4.3) 0 I d e and using Q in (4.1) (4.2) instead of Q yields the same G Ẽ. (ii) The matrix G Ẽ satisfies G Ẽ F ( n+d jn+1 σ2 j ) 1/2. (4.4) Proof. First we prove the implication (i) (ii). We partition Q Q 1 Q 2, where Q 1 R (q+d) q, Q2 R (q+d) d, and Q Q 1 Q 2, where Q 1 R(q+e) q, Q 2 R(q+e) e. Then Q Q 2 QT Q T I d e 0 I d e Q 2 0 Q 2 0 T + 0 I d e 0 I d e T,

14 which gives, using (4.2) and (2.5) G Ẽ 2 THE TOTAL LEAST SQUARES PROBLEM IN AX B 13 F diag(σ p+1,..., σ n+d ) Q 2 QT 2 2 F σ 2 n+1 Q 2(Q 2) T 2 F + n+d jn+e+1 σ2 j σ 2 n+1e + n+d jn+e+1 σ2 j, i.e. (4.4). The implication (i) (ii) is proved. Now we prove the implication (ii) (i). Let G Ẽ be given by (4.1), (4.2) and assume that (4.4) holds. We prove that there exists Q of the form (4.3) giving the same G Ẽ. Define the splitting Q11 Q Q Q1 Q2 such that Q 11 R (q+e) q, Q21 R (d e) q, Q R (q+e) d, Q R (d e) d. The matrix G Ẽ given by (4.2) satisfies G Ẽ 2 Q 21 Q F diag(σ p+1,..., σ n+d ) Q 2 2 F σ 2 n+1 Q 2 F + D Q 2 F, where D diag(σ n+e+1,..., σ n+d ). Note that Q 2 F d Q 2 F, since the matrix Q 2 consists of d orthonormal columns. Thus G Ẽ Using (4.4) this gives 2 σ 2 n+1(d e) n+d F σ2 n+1 (d Q 2 F )+ D Q 2 F σ 2 n+1 d (σ2 n+1 I d e D 2 ) 1/2 Q 2 F. jn+e+1 σ2 j (σ 2 n+1i d e D 2 ) 1/2 Q 2 F. Since σ n+1 > σ n+e+l for all l 1,..., d e, this implies that all rows of Q have norm equal to one. Consequently, since Q is an orthogonal matrix, Q 21 0, i.e. Q11 Q Q Q1 Q2, 0 Q and the matrix Q has orthonormal rows. Consider the SVD Q SI d e 0P T S 0P T, where S R (d e) (d e), P R d d are square orthogonal matrices. Define orthogonal matrices 0 S T Ψ P I e 0 R d d Iq 0 and Q Q 0 Ψ Q11 Q Ψ 0 0 I d e Because Q is orthogonal, the last d e columns of Q Ψ (i.e. corresponding to the block I d e ) are zero and Q diag(q,i d e ).

15 14 I. HNĚTYNKOVÁ, M. PLEŠINGER, D. M. SIMA, Z. STRAKOŠ, S. VAN HUFFEL is in the form (4.3) with Q Q 11 Q ΨI (e) q+d R(q+e) (q+e), where I (e) q+d represents the first e columns of I q+d. Because Q 2 QT 2 ( Q 2 Ψ)( Q 2 Ψ) T Q 2 QT 2, the matrix Q yields the same correction (4.2) as Q. The statement of this theorem says that any correction G Ẽ (reducing rank of B A to at most n) having the norm given by (4.4) can be obtained as in (4.1) (4.2) with Q in the block diagonal form (4.3). Now we describe three disjoint subsets of problems of the 1st class representing the core of the proposed classification. Define the partitioning of the matrix with respect to e, the right multiplicity of σ n+1, given by (2.5) q + e {}}{ d e {}}{ W (q,e) V ( e) d, (4.5) } {{ } d where W (q,e) R d (q+e), V ( e) R d (d e). Note that since rank( )d, i.e. a problem is of the 1st class, rank(v ( e) ) d e implies that rank(w (q,e) ) e. On the other hand rank(w (q,e) )e implies that rank(v ( e) )d e. Definition 4.2 (Partitioning of the set of problems of the 1st class). Consider a TLS problem (1.1) (1.3), m n + d. Let (2.3) be the SVD of B A with the partitioning given by (2.6), q, and the partitioning of given by (4.5), where q and e are the integers related to the multiplicity of σ n+1, given by (2.5). Let the problem (1.1) (1.3) be of the 1st class (i.e., rank( )d). The set of all problems for which rank(w (q,e) )e and rank(v ( e) )d e (V ( e) has full column rank), rank(w (q,e) ) >e and rank(v ( e) )d e (V ( e) has full column rank), rank(w (q,e) ) >e and rank(v ( e) ) <d e (V ( e) is rank deficient), will be denoted by F 1, F 2, and F 3, respectively. Clearly, F 1, F 2, and F 3 are mutually disjoint and F 1 F 2 F 3 F The set F 1 problems of the 1st class having a TLS solution in the form X (q). Consider a TLS problem of the 1st class from the set F 1, i.e. rank(w (q,e) )ein (4.5) which implies V ( e) is of full column rank, i.e. rank(v ( e) ) d e. First we give a lemma which allows to relate the partitioning (4.5) to the construction of a solution in (3.13) (3.17). Lemma 4.3. Let (2.3) be the SVD of B A with the partitioning (2.6), m n+d, q n. Consider the partitioning (4.5) of. The following two assertions are equivalent: (i) The matrix W (q,e) has rank equal to e. (ii) There exists Q in the block diagonal form (4.3) satisfying (3.13). Proof. Let W (q,e) R d (q+e) have rank equal to e. Then rank(v ( e) )d e. There exists an orthogonal matrix H R (q+e) (q+e) (e.g., a product of Householder

16 THE TOTAL LEAST SQUARES PROBLEM IN AX B 15 transformation matrices) such that W (q,e) H 0 M where M R d e is of full column rank. Putting Q diag(h, I d e ) yields Q 0 Γ, where the square matrix Γ M V ( e) R d d is nonsingular. Conversely, let Q diag(q,i d e ) and satisfy (3.13). Denote Γ Γ 1 Γ 2, where Γ 1 R d e, Γ 2 R d (d e). Obviously 0 Γ 1 W (q,e) Q, Γ 2 V ( e) I d e V ( e). Since Γ is nonsingular, rank(γ 1 )e. Q is an orthogonal matrix and thus rank(w (q,e) )e. The following theorem formulates results for the set F 1. Theorem 4.4. Let (2.3) be the SVD of B A with the partitioning (2.6), m n + d, q n (p n q). Let the TLS problem (1.1) (1.3) be of the 1st class, i.e. is of full row rank equal to d. Let σ p > σ p+1... σ n+1... σ n+e, 1 e d (if q n, then σ p is not defined). Consider the partitioning of given by (4.5). If rank(w (q,e) )e, (4.6) (the problem is from the set F 1 ), then X TLS X (q) given by (3.17) represents the TLS solution having the minimality property (3.). The corresponding correction G E given by (3.15) has the norm (4.4). The proof follows immediately from Lemma 4.3, Lemma 3.6, and Lemma 3.7. The problems of the 1st class discussed earlier in Sections 3.1 and 3.2 belong to the set F 1. In the first case q 0 and V is square nonsingular. Thus independently on the value of e (4.5) yields W (0,e) with the (full column) rank equal to e and the matrix Q from Q diag(q,i d e ) in the assertion (ii) of Lemma 4.3 can be always chosen equal to the identity matrix I e, i.e. Q I d. In the second case e d. Thus W (q,d) is of (full row) rank equal to d. Here the identity block I d e in the assertion (ii) of Lemma 4.3 disappears, i.e. Q Q The set F 2 problems of the 1st class having a TLS solution but not in the form X (q). Consider a TLS problem of the 1st class from the set F 2, i.e. rank(v ( e) )d e and rank(w (q,e) ) >ein (4.5). Because V ( e) is of full column rank, there exists Q diag(q,i d e ) having the block diagonal form (4.3) such that (4.1) holds, Q W (q,e) Q V ( e) Ω Γ 1 V ( e), (4.7) with Γ Γ 1 V ( e) nonsingular. Consequently, the correction G Ẽ defined by (4.2) is minimal in the Frobenius norm, see Theorem 4.1, and the corresponding matrix X Z Γ 1 given by (3.20) represents a TLS solution (which is, by Lemma 3.7, the unique solution of the corrected system with the given fixed correction G Ẽ). Because rank(w (q,e) ) >eand Q is orthogonal, the product W (q,e) Q Ω Γ 1 where rank( Γ 1 )e ( Γ is nonsingular) leads always to a nonzero Ω. On the other hand, the construction (3.15) (3.17) always leads to Ω 0. Hence, the matrix X (q) given by (3.17) does not represent a TLS solution. The following theorem completes the argument by showing that any problem from the set F 2 has always a minimum norm TLS solution.

17 16 I. HNĚTYNKOVÁ, M. PLEŠINGER, D. M. SIMA, Z. STRAKOŠ, S. VAN HUFFEL Theorem 4.5. Let (1.1) (1.3) be the TLS problem of the 1st class belonging to the set F 2. Then there exist TLS solutions given by (3.18) (3.20) minimal in the 2-norm, and in the Frobenius norm, respectively. Proof. A TLS solution X Z Γ 1 is obtained from the formula V Q (q) Q 1 Q 2 0 Ω Γ, 0 0 I d e Ỹ Z where the block diagonal matrix Q is the orthogonal matrix (4.3) from Theorem 4.1. The TLS solution is uniquely determined by the orthogonal matrix Q Q 1 Q 2 R (q+e) (q+e). In our construction, Q R (q+e) (q+e) is required to lead to a nonsingular Γ. Since the matrix inversion is a continuous function of entries of a nonsingular matrix, and matrix multiplication is a continuous function of entries of both factors, the matrix X Z Γ 1 is a continuous matrix-valued function of Q. Define two nonnegative functionals N 2 (Q ):R (q+e) (q+e) 0, + and N F (Q ):R (q+e) (q+e) 0, + on a set of all (q + e)-by-(q + e) orthogonal matrices such that { N 2 (Q X(Q ) ) 2, +, if Q gives Γ(Q ) nonsingular, if Q gives Γ(Q ) singular. The functional N F (Q ) is defined analogously. Note that both functionals are nonnegative and lower semi-continuous on the compact set of all (q + e)-by-(q + e) orthogonal matrices, and thus both functionals have a minimum on this set. Theorem 4.5 does not address the uniqueness of the minimum norm solutions, and it also does not give any practical algorithm for computing them. Further note that the sets of solutions minimal in 2-norm and minimal in the Frobenius norm can be different or even disjoint. This fact can be illustrated with the following example. Consider the problem given by its SVD decomposition B A U T, (4.8) where A R 4 2, B R 4 2 (it is easy to verify that A T B 0). Here q 1, e 1, W (q,e) , V ( e) have rank two and one, respectively. This problem is of the 1st class and belongs to the set F 2. The TLS solution is determined by the orthogonal matrix Q Q 1 Q cos(φ) sin(φ) sin(φ) cos(φ) I d e which depends only on one real variable φ. Figure 4.1 shows how the 2-norm and the Frobenius norm of the TLS solution depend on the value of φ. From the behavior of,

18 THE TOTAL LEAST SQUARES PROBLEM IN AX B X(φ) 2 X(φ) F min φ X(φ) 2 min X(φ) φ F X(φ) 2 X(φ) F min φ X(φ) 2 min X(φ) φ F norm of X(φ) norm of X(φ) π, π, π 0.5π 1π 1.5π 2π φ 1 0.9π 1.0π 1.1π 1.2π 1.3π φ Fig (Left plot) The 2-norm and the Frobenius norm of TLS solutions of the problem (4.8) belonging to the set F 2. Solutions minimal in different norms are distinct. (Right plot) Detail of the solutions minimal in the 2-norm and in the Frobenius norm. the norms it is clear that the set of solutions minimal in 2-norm has no intersection with the set of solutions minimal in the Frobenius norm. If we use in the previous example (4.8) the matrix of the right singular vectors V then there exists a solution which is minimal in both 2-norm and the Frobenius norm The set F 3 problems of the 1st class which do not have a TLS solution. Consider a TLS problem of the 1st class from the set F 3, i.e. the case with rank(v ( e) ) <d e. Since V ( e) in (4.5) is rank deficient, Q R (q+d) (q+d) in the block diagonal form (4.3) leads to (4.7) with Γ Γ 1 V ( e) containing linearly dependent column(s). Thus Γ in (4.1) is always singular. Consequently, in this case there does not exist Q in the block diagonal form yielding Γ nonsingular. Therefore, there is no correction G Ẽ having the norm (4.4) which makes the system (1.1) compatible, see Theorem 4.1. Now we show that a TLS solution does not exist for the problems from the set F 3. Using a general matrix Q, see (3.18) we construct a correction (3.19) which makes the system compatible, and the norm of this correction is arbitrarily close to the lower bound (4.4). Denote ρ (d e) rank(v ( e) ) the rank defect of V ( e). Analogously to Section 4.2, there exists an orthogonal matrix Q R (q+e) (q+e) such that diag(q,i d e ) W (q,e) Q V ( e) Ω Γ 1 V ( e) with rank( Γ 1 V ( e) ) d ρ, compare with (4.7). Let J {j 1,...,j ρ } denote indices of any ρ columns of V ( e) such that the remaining columns of V ( e) (with indices {1,..., d e} \J) are linearly independent. Because rank( )d, the matrix Ω has ρ linearly independent columns which are not in R( Γ 1 V ( e) ), let,

19 18 I. HNĚTYNKOVÁ, M. PLEŠINGER, D. M. SIMA, Z. STRAKOŠ, S. VAN HUFFEL K {k 1,..., k ρ } denote their indices. Consider an angle θ, 0< θ < π. A Givens rotation corresponding to θ applied subsequently on pairs of columns with indices j l and k l, for l 1,..., ρ, can be written as an orthogonal transformation Ω Γ 1 V ( e) C 11 0 S 0 I e 0 S T 0 C Ω Γ1 V ( e) where C 11 R q q and C R (d e) (d e) are diagonal matrices having ρ diagonal entries (on the positions (k l,k l ) and (j l,j l ), l 1,..., ρ, respectively) equal to cos(θ) (the other diagonal entries are equal to 1), and S R q (d e) has entries on positions (k l,j l ), l 1,..., ρ, equal to sin(θ) (the other entries are zero). Since 0 < θ < π, the matrix Γ Γ 1 V ( e) is nonsingular, and thus the corresponding correction makes the system compatible. The transformation matrix Q Q diag(c 11,I e ) Q S 0 S T 0 C can be, with θ 0, arbitrarily close to the block diagonal form (4.3), and moreover the Frobenius norm of the corresponding correction G Ẽ F ( n+d, jn+1 σ2 j + sin 2 (θ) j J (σ2 n+1 σ 2 n+e+j), ) 1/2 can be arbitrarily close to the lower bound given by (4.4). Consequently, there is no minimal correction that makes the system (1.1) compatible. The TLS problem (1.1) (1.3) with rank deficient V ( e) does not have a solution Correction corresponding to the matrix X (q). In the previous three sections we have shown that a TLS solution (if exists) always has the correction matrix with the Frobenius norm (4.4). We can formulate the following Corollary. Corollary 4.6. Consider a TLS problem (1.1) (1.3) of the 1st class. The construction (3.13) (3.17) yields the TLS solution X TLS X (q) if and only if there exists an orthogonal matrix Q in the block diagonal form (4.3) such that substituting Q for Q in (3.13) (3.15) gives the same correction G E. Now we focus on the properties of the correction G E given by (3.15) in general. First we prove an auxiliary lemma. Lemma 4.7. Let G E be the correction matrix given by (3.15). Denote s rank(b A). Then the ranks of the correction and corrected matrix satisfy min{s, d} rank ( G E ) max{0,s n}, (4.9) max{0,s d} rank ( B + G A + E ) min{s, n}. (4.10) Proof. The upper bound in (4.9) follows immediately from (3.15). The lower bound in (4.9) follows from the fact that the correction matrix makes the system compatible, i.e. the resulting rank of B + G A + E is at most n, which also proves the upper bound in (4.10). Since the rank of G E is at most d, the lower bounds in (4.10) follows trivially.

20 THE TOTAL LEAST SQUARES PROBLEM IN AX B 19 The result of the following theorem can also be found in, Equation (5.4). Theorem 4.8. Let G E be the correction matrix given by (3.15). Then its Frobenius norm satisfies ( p+d jp+1 σ2 j ) 1/2 G E F ( n+d jn+1 σ2 j ) 1/2. (4.11) Proof. The lower bound in (4.11) is trivial. The matrix G E has from (4.9) the rank not greater than min{s, d} which immediately gives the upper bound. From the construction (3.15) a rank d matrix of the given form can not have Frobenius norm larger than (4.11). Since the Frobenius norm of the correction G E given by (3.15) can be larger than ( n+d jn+1 σ2 j )1/2, the correction need not be minimal and (3.17) need not represent (as described above) a TLS solution. Further note the inequalities in (4.11) become equalities if and only if σ p+j σ n+j, j 1,..., d, (recall that n p + q). This happens either if q 0 (the case with the unique solution discussed in Section 3.1), or if σ p+1... σ n+d (the special case discussed in Section 3.2). 5. Problems of the 2nd class. In this section we briefly describe problems (1.1) (1.3) of the 2nd class, i.e. the problems for which does not have full row rank, see Definition 2.2. Here the right singular vector subspace given by the last (q + d) singular vectors v p+1,..., v n+d does not contain sufficient information for constructing a solution (3.20) and the problems of the 2nd class do not have a TLS solution (the argumentation is analogous as in Section 4.3). The classical TLS algorithm, which gives an output also for problems of the 2nd class, is derived in 17 by a straightforward generalization of the single righthand side concept. The right singular vector subspace R(T T T ) used for the construction (3.13) (3.17) in previous cases is extended with additional right singular vectors until, for some t, a full row rank block V (t) Rd (t+d) is found in the upper right corner of V (and V (t 1) is, at the same time, rank deficient), V q d {}}{{}}{ } d }{{}}{{}}{{} n t t d n V (t) 11 V (t) V (t) 21 V (t) Then the matrix X (t) V (t) V (t) with the corresponding correction can be constructed analogously to (3.13) (3.17), with q replaced by t. Obviously, this matrix might not be uniquely defined when σ n t+1 is not simple, in particular, when σ n t σ n t+1. In order to handle a possible multiplicity of σ n t+1, it is convenient to consider the following notation σ n eq > σ n eq+1... σ n t σ n t+1 σ n t+2,.

21 20 I. HNĚTYNKOVÁ, M. PLEŠINGER, D. M. SIMA, Z. STRAKOŠ, S. VAN HUFFEL where q t; put for simplicity n q p. (If such σ n eq σ ep does not exist, then put q n.) The condition that V (eq) is of full row rank equal to d is readily satisfied, since V (eq) extends V (t). Then X(eq) and G E can be constructed as in (3.13) (3.17) with q replaced by q. Thus the matrix X (eq) V (eq) V (eq) represents a solution of the compatible corrected system (A + E)X B + G. The Frobenius and the 2-norm of the matrix X (eq) are given by Lemma 3.5. Similarly to the problems of the 1st class, the minimality property (3.) of X (eq) can be shown. Thus X (eq) has minimal Frobenius and 2-norm over all matrices X that can be obtained from the construction analogous to (3.18) (3.20) with q replaced by q. The substitution of q for t ensures the uniqueness of the construction, and leads to the matrix with the smallest norm. On the other hand it inevitably increases the norm of the correction, with G E ( n+d ) 1/2 F > jn+1 σ2 j. The Frobenius norm of G E is strictly larger than the smallest possible correction reducing the rank of B A to n, and the matrix X (eq) does not represent a TLS solution Summary of the relationship to the classical TLS algorithm. The classical TLS algorithm gives for any data the output X (κ) which is equal (in exact Algorithm 1 (The classical TLS algorithm) A fully documented Fortran 77 implementation is given in 15, 16. The code can be obtained through Netlib.org, cf. Require: A R m n, B R m d {here the SVD of B A in the form (2.3) (2.6)} 1: 0 2: if rank(v ( ) )d and n, then goto 6 3: if rank(v ( ) )d and σ n > σ n +1, then goto 6 4: +1 5: goto 2 6: κ 7: X (κ) V (κ) V (κ) 8: return κ, X (κ) arithmetic) either to X (q) given by (3.2), or (3.10), or by (3.17), or to X (eq) described in Section 5. The output X (κ) is called generic (or TLS) solution in 17 for any problem of the 1st class, and it is called nongeneric solution in 17 for any problem of the 2nd class. As our new partitioning and the included classification reveals, (i) if the problem is of the 1st class and rank(w (q,e) )e, i.e., the problem belongs to the set F 1, then X (κ) X TLS represents a TLS solution (it solves the TLS problem (1.1) (1.3)), κ q, (ii) if the problem is of the 1st class and rank(w (q,e) ) >e, i.e., the problem belongs to the set F 2 F 3, then X (κ) does not represent a TLS solution, which exists for the problems in the set F 2 but does not exist for the problems in the set F 3, κ q, 2 The matrix X (eq) V (eq) V (eq) is called nongeneric solution in 17, Definition 3.3, p. 78.

UNIFYING LEAST SQUARES, TOTAL LEAST SQUARES AND DATA LEAST SQUARES

UNIFYING LEAST SQUARES, TOTAL LEAST SQUARES AND DATA LEAST SQUARES UNIFYING LEAST SQUARES, TOTAL LEAST SQUARES AND DATA LEAST SQUARES Christopher C. Paige School of Computer Science, McGill University, Montreal, Quebec, Canada, H3A 2A7 paige@cs.mcgill.ca Zdeněk Strakoš

More information

Characterization of half-radial matrices

Characterization of half-radial matrices Characterization of half-radial matrices Iveta Hnětynková, Petr Tichý Faculty of Mathematics and Physics, Charles University, Sokolovská 83, Prague 8, Czech Republic Abstract Numerical radius r(a) is the

More information

THE CORE PROBLEM WITHIN A LINEAR APPROXIMATION PROBLEM AX B WITH MULTIPLE RIGHT-HAND SIDES. 1. Introduction. Consider a linear approximation problem

THE CORE PROBLEM WITHIN A LINEAR APPROXIMATION PROBLEM AX B WITH MULTIPLE RIGHT-HAND SIDES. 1. Introduction. Consider a linear approximation problem HE CORE PROBLEM WIHIN A LINEAR APPROXIMAION PROBLEM AX B WIH MULIPLE RIGH-HAND SIDES IVEA HNĚYNKOVÁ, MARIN PLEŠINGER, AND ZDENĚK SRAKOŠ Abstract his paper focuses on total least squares (LS) problems AX

More information

Lecture notes: Applied linear algebra Part 1. Version 2

Lecture notes: Applied linear algebra Part 1. Version 2 Lecture notes: Applied linear algebra Part 1. Version 2 Michael Karow Berlin University of Technology karow@math.tu-berlin.de October 2, 2008 1 Notation, basic notions and facts 1.1 Subspaces, range and

More information

Linear Systems. Carlo Tomasi

Linear Systems. Carlo Tomasi Linear Systems Carlo Tomasi Section 1 characterizes the existence and multiplicity of the solutions of a linear system in terms of the four fundamental spaces associated with the system s matrix and of

More information

Notes on Eigenvalues, Singular Values and QR

Notes on Eigenvalues, Singular Values and QR Notes on Eigenvalues, Singular Values and QR Michael Overton, Numerical Computing, Spring 2017 March 30, 2017 1 Eigenvalues Everyone who has studied linear algebra knows the definition: given a square

More information

Linear Systems. Carlo Tomasi. June 12, r = rank(a) b range(a) n r solutions

Linear Systems. Carlo Tomasi. June 12, r = rank(a) b range(a) n r solutions Linear Systems Carlo Tomasi June, 08 Section characterizes the existence and multiplicity of the solutions of a linear system in terms of the four fundamental spaces associated with the system s matrix

More information

Total Least Squares Approach in Regression Methods

Total Least Squares Approach in Regression Methods WDS'08 Proceedings of Contributed Papers, Part I, 88 93, 2008. ISBN 978-80-7378-065-4 MATFYZPRESS Total Least Squares Approach in Regression Methods M. Pešta Charles University, Faculty of Mathematics

More information

ECE 275A Homework #3 Solutions

ECE 275A Homework #3 Solutions ECE 75A Homework #3 Solutions. Proof of (a). Obviously Ax = 0 y, Ax = 0 for all y. To show sufficiency, note that if y, Ax = 0 for all y, then it must certainly be true for the particular value of y =

More information

IV. Matrix Approximation using Least-Squares

IV. Matrix Approximation using Least-Squares IV. Matrix Approximation using Least-Squares The SVD and Matrix Approximation We begin with the following fundamental question. Let A be an M N matrix with rank R. What is the closest matrix to A that

More information

Properties of Matrices and Operations on Matrices

Properties of Matrices and Operations on Matrices Properties of Matrices and Operations on Matrices A common data structure for statistical analysis is a rectangular array or matris. Rows represent individual observational units, or just observations,

More information

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH V. FABER, J. LIESEN, AND P. TICHÝ Abstract. Numerous algorithms in numerical linear algebra are based on the reduction of a given matrix

More information

COMP 558 lecture 18 Nov. 15, 2010

COMP 558 lecture 18 Nov. 15, 2010 Least squares We have seen several least squares problems thus far, and we will see more in the upcoming lectures. For this reason it is good to have a more general picture of these problems and how to

More information

Stat 159/259: Linear Algebra Notes

Stat 159/259: Linear Algebra Notes Stat 159/259: Linear Algebra Notes Jarrod Millman November 16, 2015 Abstract These notes assume you ve taken a semester of undergraduate linear algebra. In particular, I assume you are familiar with the

More information

Chapter 3 Transformations

Chapter 3 Transformations Chapter 3 Transformations An Introduction to Optimization Spring, 2014 Wei-Ta Chu 1 Linear Transformations A function is called a linear transformation if 1. for every and 2. for every If we fix the bases

More information

Maths for Signals and Systems Linear Algebra in Engineering

Maths for Signals and Systems Linear Algebra in Engineering Maths for Signals and Systems Linear Algebra in Engineering Lectures 13 15, Tuesday 8 th and Friday 11 th November 016 DR TANIA STATHAKI READER (ASSOCIATE PROFFESOR) IN SIGNAL PROCESSING IMPERIAL COLLEGE

More information

7. Symmetric Matrices and Quadratic Forms

7. Symmetric Matrices and Quadratic Forms Linear Algebra 7. Symmetric Matrices and Quadratic Forms CSIE NCU 1 7. Symmetric Matrices and Quadratic Forms 7.1 Diagonalization of symmetric matrices 2 7.2 Quadratic forms.. 9 7.4 The singular value

More information

EE731 Lecture Notes: Matrix Computations for Signal Processing

EE731 Lecture Notes: Matrix Computations for Signal Processing EE731 Lecture Notes: Matrix Computations for Signal Processing James P. Reilly c Department of Electrical and Computer Engineering McMaster University September 22, 2005 0 Preface This collection of ten

More information

Recall the convention that, for us, all vectors are column vectors.

Recall the convention that, for us, all vectors are column vectors. Some linear algebra Recall the convention that, for us, all vectors are column vectors. 1. Symmetric matrices Let A be a real matrix. Recall that a complex number λ is an eigenvalue of A if there exists

More information

EE731 Lecture Notes: Matrix Computations for Signal Processing

EE731 Lecture Notes: Matrix Computations for Signal Processing EE731 Lecture Notes: Matrix Computations for Signal Processing James P. Reilly c Department of Electrical and Computer Engineering McMaster University October 17, 005 Lecture 3 3 he Singular Value Decomposition

More information

Matrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A =

Matrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A = 30 MATHEMATICS REVIEW G A.1.1 Matrices and Vectors Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A = a 11 a 12... a 1N a 21 a 22... a 2N...... a M1 a M2... a MN A matrix can

More information

Lecture 1: Review of linear algebra

Lecture 1: Review of linear algebra Lecture 1: Review of linear algebra Linear functions and linearization Inverse matrix, least-squares and least-norm solutions Subspaces, basis, and dimension Change of basis and similarity transformations

More information

Review of Some Concepts from Linear Algebra: Part 2

Review of Some Concepts from Linear Algebra: Part 2 Review of Some Concepts from Linear Algebra: Part 2 Department of Mathematics Boise State University January 16, 2019 Math 566 Linear Algebra Review: Part 2 January 16, 2019 1 / 22 Vector spaces A set

More information

DS-GA 1002 Lecture notes 10 November 23, Linear models

DS-GA 1002 Lecture notes 10 November 23, Linear models DS-GA 2 Lecture notes November 23, 2 Linear functions Linear models A linear model encodes the assumption that two quantities are linearly related. Mathematically, this is characterized using linear functions.

More information

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88 Math Camp 2010 Lecture 4: Linear Algebra Xiao Yu Wang MIT Aug 2010 Xiao Yu Wang (MIT) Math Camp 2010 08/10 1 / 88 Linear Algebra Game Plan Vector Spaces Linear Transformations and Matrices Determinant

More information

7. Dimension and Structure.

7. Dimension and Structure. 7. Dimension and Structure 7.1. Basis and Dimension Bases for Subspaces Example 2 The standard unit vectors e 1, e 2,, e n are linearly independent, for if we write (2) in component form, then we obtain

More information

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.

More information

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 2

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 2 EE/ACM 150 - Applications of Convex Optimization in Signal Processing and Communications Lecture 2 Andre Tkacenko Signal Processing Research Group Jet Propulsion Laboratory April 5, 2012 Andre Tkacenko

More information

Linear Algebra. Session 12

Linear Algebra. Session 12 Linear Algebra. Session 12 Dr. Marco A Roque Sol 08/01/2017 Example 12.1 Find the constant function that is the least squares fit to the following data x 0 1 2 3 f(x) 1 0 1 2 Solution c = 1 c = 0 f (x)

More information

Optimization problems on the rank and inertia of the Hermitian matrix expression A BX (BX) with applications

Optimization problems on the rank and inertia of the Hermitian matrix expression A BX (BX) with applications Optimization problems on the rank and inertia of the Hermitian matrix expression A BX (BX) with applications Yongge Tian China Economics and Management Academy, Central University of Finance and Economics,

More information

1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det

1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det What is the determinant of the following matrix? 3 4 3 4 3 4 4 3 A 0 B 8 C 55 D 0 E 60 If det a a a 3 b b b 3 c c c 3 = 4, then det a a 4a 3 a b b 4b 3 b c c c 3 c = A 8 B 6 C 4 D E 3 Let A be an n n matrix

More information

Normed & Inner Product Vector Spaces

Normed & Inner Product Vector Spaces Normed & Inner Product Vector Spaces ECE 174 Introduction to Linear & Nonlinear Optimization Ken Kreutz-Delgado ECE Department, UC San Diego Ken Kreutz-Delgado (UC San Diego) ECE 174 Fall 2016 1 / 27 Normed

More information

Elementary linear algebra

Elementary linear algebra Chapter 1 Elementary linear algebra 1.1 Vector spaces Vector spaces owe their importance to the fact that so many models arising in the solutions of specific problems turn out to be vector spaces. The

More information

j=1 u 1jv 1j. 1/ 2 Lemma 1. An orthogonal set of vectors must be linearly independent.

j=1 u 1jv 1j. 1/ 2 Lemma 1. An orthogonal set of vectors must be linearly independent. Lecture Notes: Orthogonal and Symmetric Matrices Yufei Tao Department of Computer Science and Engineering Chinese University of Hong Kong taoyf@cse.cuhk.edu.hk Orthogonal Matrix Definition. Let u = [u

More information

Numerical Methods. Elena loli Piccolomini. Civil Engeneering. piccolom. Metodi Numerici M p. 1/??

Numerical Methods. Elena loli Piccolomini. Civil Engeneering.  piccolom. Metodi Numerici M p. 1/?? Metodi Numerici M p. 1/?? Numerical Methods Elena loli Piccolomini Civil Engeneering http://www.dm.unibo.it/ piccolom elena.loli@unibo.it Metodi Numerici M p. 2/?? Least Squares Data Fitting Measurement

More information

Singular Value Decomposition

Singular Value Decomposition Chapter 6 Singular Value Decomposition In Chapter 5, we derived a number of algorithms for computing the eigenvalues and eigenvectors of matrices A R n n. Having developed this machinery, we complete our

More information

Singular Value Decomposition

Singular Value Decomposition Chapter 5 Singular Value Decomposition We now reach an important Chapter in this course concerned with the Singular Value Decomposition of a matrix A. SVD, as it is commonly referred to, is one of the

More information

Block Bidiagonal Decomposition and Least Squares Problems

Block Bidiagonal Decomposition and Least Squares Problems Block Bidiagonal Decomposition and Least Squares Problems Åke Björck Department of Mathematics Linköping University Perspectives in Numerical Analysis, Helsinki, May 27 29, 2008 Outline Bidiagonal Decomposition

More information

NORMS ON SPACE OF MATRICES

NORMS ON SPACE OF MATRICES NORMS ON SPACE OF MATRICES. Operator Norms on Space of linear maps Let A be an n n real matrix and x 0 be a vector in R n. We would like to use the Picard iteration method to solve for the following system

More information

MATH 2331 Linear Algebra. Section 2.1 Matrix Operations. Definition: A : m n, B : n p. Example: Compute AB, if possible.

MATH 2331 Linear Algebra. Section 2.1 Matrix Operations. Definition: A : m n, B : n p. Example: Compute AB, if possible. MATH 2331 Linear Algebra Section 2.1 Matrix Operations Definition: A : m n, B : n p ( 1 2 p ) ( 1 2 p ) AB = A b b b = Ab Ab Ab Example: Compute AB, if possible. 1 Row-column rule: i-j-th entry of AB:

More information

Singular Value Decomposition (SVD)

Singular Value Decomposition (SVD) School of Computing National University of Singapore CS CS524 Theoretical Foundations of Multimedia More Linear Algebra Singular Value Decomposition (SVD) The highpoint of linear algebra Gilbert Strang

More information

NONCOMMUTATIVE POLYNOMIAL EQUATIONS. Edward S. Letzter. Introduction

NONCOMMUTATIVE POLYNOMIAL EQUATIONS. Edward S. Letzter. Introduction NONCOMMUTATIVE POLYNOMIAL EQUATIONS Edward S Letzter Introduction My aim in these notes is twofold: First, to briefly review some linear algebra Second, to provide you with some new tools and techniques

More information

Linear Algebra March 16, 2019

Linear Algebra March 16, 2019 Linear Algebra March 16, 2019 2 Contents 0.1 Notation................................ 4 1 Systems of linear equations, and matrices 5 1.1 Systems of linear equations..................... 5 1.2 Augmented

More information

σ 11 σ 22 σ pp 0 with p = min(n, m) The σ ii s are the singular values. Notation change σ ii A 1 σ 2

σ 11 σ 22 σ pp 0 with p = min(n, m) The σ ii s are the singular values. Notation change σ ii A 1 σ 2 HE SINGULAR VALUE DECOMPOSIION he SVD existence - properties. Pseudo-inverses and the SVD Use of SVD for least-squares problems Applications of the SVD he Singular Value Decomposition (SVD) heorem For

More information

Chap 3. Linear Algebra

Chap 3. Linear Algebra Chap 3. Linear Algebra Outlines 1. Introduction 2. Basis, Representation, and Orthonormalization 3. Linear Algebraic Equations 4. Similarity Transformation 5. Diagonal Form and Jordan Form 6. Functions

More information

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2017 LECTURE 5

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2017 LECTURE 5 STAT 39: MATHEMATICAL COMPUTATIONS I FALL 17 LECTURE 5 1 existence of svd Theorem 1 (Existence of SVD) Every matrix has a singular value decomposition (condensed version) Proof Let A C m n and for simplicity

More information

Econ Slides from Lecture 7

Econ Slides from Lecture 7 Econ 205 Sobel Econ 205 - Slides from Lecture 7 Joel Sobel August 31, 2010 Linear Algebra: Main Theory A linear combination of a collection of vectors {x 1,..., x k } is a vector of the form k λ ix i for

More information

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra. DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1

More information

4.3 - Linear Combinations and Independence of Vectors

4.3 - Linear Combinations and Independence of Vectors - Linear Combinations and Independence of Vectors De nitions, Theorems, and Examples De nition 1 A vector v in a vector space V is called a linear combination of the vectors u 1, u,,u k in V if v can be

More information

Multiplicative Perturbation Bounds of the Group Inverse and Oblique Projection

Multiplicative Perturbation Bounds of the Group Inverse and Oblique Projection Filomat 30: 06, 37 375 DOI 0.98/FIL67M Published by Faculty of Sciences Mathematics, University of Niš, Serbia Available at: http://www.pmf.ni.ac.rs/filomat Multiplicative Perturbation Bounds of the Group

More information

Vector Space Concepts

Vector Space Concepts Vector Space Concepts ECE 174 Introduction to Linear & Nonlinear Optimization Ken Kreutz-Delgado ECE Department, UC San Diego Ken Kreutz-Delgado (UC San Diego) ECE 174 Fall 2016 1 / 25 Vector Space Theory

More information

Applied Mathematics 205. Unit II: Numerical Linear Algebra. Lecturer: Dr. David Knezevic

Applied Mathematics 205. Unit II: Numerical Linear Algebra. Lecturer: Dr. David Knezevic Applied Mathematics 205 Unit II: Numerical Linear Algebra Lecturer: Dr. David Knezevic Unit II: Numerical Linear Algebra Chapter II.3: QR Factorization, SVD 2 / 66 QR Factorization 3 / 66 QR Factorization

More information

Preliminary Linear Algebra 1. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 100

Preliminary Linear Algebra 1. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 100 Preliminary Linear Algebra 1 Copyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 1 / 100 Notation for all there exists such that therefore because end of proof (QED) Copyright c 2012

More information

LINEAR ALGEBRA: THEORY. Version: August 12,

LINEAR ALGEBRA: THEORY. Version: August 12, LINEAR ALGEBRA: THEORY. Version: August 12, 2000 13 2 Basic concepts We will assume that the following concepts are known: Vector, column vector, row vector, transpose. Recall that x is a column vector,

More information

POSITIVE SEMIDEFINITE INTERVALS FOR MATRIX PENCILS

POSITIVE SEMIDEFINITE INTERVALS FOR MATRIX PENCILS POSITIVE SEMIDEFINITE INTERVALS FOR MATRIX PENCILS RICHARD J. CARON, HUIMING SONG, AND TIM TRAYNOR Abstract. Let A and E be real symmetric matrices. In this paper we are concerned with the determination

More information

ECE 275A Homework # 3 Due Thursday 10/27/2016

ECE 275A Homework # 3 Due Thursday 10/27/2016 ECE 275A Homework # 3 Due Thursday 10/27/2016 Reading: In addition to the lecture material presented in class, students are to read and study the following: A. The material in Section 4.11 of Moon & Stirling

More information

Golub-Kahan iterative bidiagonalization and determining the noise level in the data

Golub-Kahan iterative bidiagonalization and determining the noise level in the data Golub-Kahan iterative bidiagonalization and determining the noise level in the data Iveta Hnětynková,, Martin Plešinger,, Zdeněk Strakoš, * Charles University, Prague ** Academy of Sciences of the Czech

More information

Notes on singular value decomposition for Math 54. Recall that if A is a symmetric n n matrix, then A has real eigenvalues A = P DP 1 A = P DP T.

Notes on singular value decomposition for Math 54. Recall that if A is a symmetric n n matrix, then A has real eigenvalues A = P DP 1 A = P DP T. Notes on singular value decomposition for Math 54 Recall that if A is a symmetric n n matrix, then A has real eigenvalues λ 1,, λ n (possibly repeated), and R n has an orthonormal basis v 1,, v n, where

More information

MATH36001 Generalized Inverses and the SVD 2015

MATH36001 Generalized Inverses and the SVD 2015 MATH36001 Generalized Inverses and the SVD 201 1 Generalized Inverses of Matrices A matrix has an inverse only if it is square and nonsingular. However there are theoretical and practical applications

More information

. = V c = V [x]v (5.1) c 1. c k

. = V c = V [x]v (5.1) c 1. c k Chapter 5 Linear Algebra It can be argued that all of linear algebra can be understood using the four fundamental subspaces associated with a matrix Because they form the foundation on which we later work,

More information

4.2. ORTHOGONALITY 161

4.2. ORTHOGONALITY 161 4.2. ORTHOGONALITY 161 Definition 4.2.9 An affine space (E, E ) is a Euclidean affine space iff its underlying vector space E is a Euclidean vector space. Given any two points a, b E, we define the distance

More information

A MODIFIED TSVD METHOD FOR DISCRETE ILL-POSED PROBLEMS

A MODIFIED TSVD METHOD FOR DISCRETE ILL-POSED PROBLEMS A MODIFIED TSVD METHOD FOR DISCRETE ILL-POSED PROBLEMS SILVIA NOSCHESE AND LOTHAR REICHEL Abstract. Truncated singular value decomposition (TSVD) is a popular method for solving linear discrete ill-posed

More information

Consistent Histories. Chapter Chain Operators and Weights

Consistent Histories. Chapter Chain Operators and Weights Chapter 10 Consistent Histories 10.1 Chain Operators and Weights The previous chapter showed how the Born rule can be used to assign probabilities to a sample space of histories based upon an initial state

More information

The Singular Value Decomposition

The Singular Value Decomposition The Singular Value Decomposition An Important topic in NLA Radu Tiberiu Trîmbiţaş Babeş-Bolyai University February 23, 2009 Radu Tiberiu Trîmbiţaş ( Babeş-Bolyai University)The Singular Value Decomposition

More information

ON THE SINGULAR DECOMPOSITION OF MATRICES

ON THE SINGULAR DECOMPOSITION OF MATRICES An. Şt. Univ. Ovidius Constanţa Vol. 8, 00, 55 6 ON THE SINGULAR DECOMPOSITION OF MATRICES Alina PETRESCU-NIŢǍ Abstract This paper is an original presentation of the algorithm of the singular decomposition

More information

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces. Math 350 Fall 2011 Notes about inner product spaces In this notes we state and prove some important properties of inner product spaces. First, recall the dot product on R n : if x, y R n, say x = (x 1,...,

More information

Orthogonality. 6.1 Orthogonal Vectors and Subspaces. Chapter 6

Orthogonality. 6.1 Orthogonal Vectors and Subspaces. Chapter 6 Chapter 6 Orthogonality 6.1 Orthogonal Vectors and Subspaces Recall that if nonzero vectors x, y R n are linearly independent then the subspace of all vectors αx + βy, α, β R (the space spanned by x and

More information

Linear Algebra Massoud Malek

Linear Algebra Massoud Malek CSUEB Linear Algebra Massoud Malek Inner Product and Normed Space In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n An inner product

More information

The Singular Value Decomposition

The Singular Value Decomposition The Singular Value Decomposition Philippe B. Laval KSU Fall 2015 Philippe B. Laval (KSU) SVD Fall 2015 1 / 13 Review of Key Concepts We review some key definitions and results about matrices that will

More information

Introduction to Numerical Linear Algebra II

Introduction to Numerical Linear Algebra II Introduction to Numerical Linear Algebra II Petros Drineas These slides were prepared by Ilse Ipsen for the 2015 Gene Golub SIAM Summer School on RandNLA 1 / 49 Overview We will cover this material in

More information

Topological Data Analysis - Spring 2018

Topological Data Analysis - Spring 2018 Topological Data Analysis - Spring 2018 Simplicial Homology Slightly rearranged, but mostly copy-pasted from Harer s and Edelsbrunner s Computational Topology, Verovsek s class notes. Gunnar Carlsson s

More information

THE SINGULAR VALUE DECOMPOSITION MARKUS GRASMAIR

THE SINGULAR VALUE DECOMPOSITION MARKUS GRASMAIR THE SINGULAR VALUE DECOMPOSITION MARKUS GRASMAIR 1. Definition Existence Theorem 1. Assume that A R m n. Then there exist orthogonal matrices U R m m V R n n, values σ 1 σ 2... σ p 0 with p = min{m, n},

More information

(v, w) = arccos( < v, w >

(v, w) = arccos( < v, w > MA322 F all206 Notes on Inner Products Notes on Chapter 6 Inner product. Given a real vector space V, an inner product is defined to be a bilinear map F : V V R such that the following holds: Commutativity:

More information

Finite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product

Finite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product Chapter 4 Hilbert Spaces 4.1 Inner Product Spaces Inner Product Space. A complex vector space E is called an inner product space (or a pre-hilbert space, or a unitary space) if there is a mapping (, )

More information

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2013 PROBLEM SET 2

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2013 PROBLEM SET 2 STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2013 PROBLEM SET 2 1. You are not allowed to use the svd for this problem, i.e. no arguments should depend on the svd of A or A. Let W be a subspace of C n. The

More information

OHSx XM511 Linear Algebra: Solutions to Online True/False Exercises

OHSx XM511 Linear Algebra: Solutions to Online True/False Exercises This document gives the solutions to all of the online exercises for OHSx XM511. The section ( ) numbers refer to the textbook. TYPE I are True/False. Answers are in square brackets [. Lecture 02 ( 1.1)

More information

Algebraic Methods in Combinatorics

Algebraic Methods in Combinatorics Algebraic Methods in Combinatorics Po-Shen Loh 27 June 2008 1 Warm-up 1. (A result of Bourbaki on finite geometries, from Răzvan) Let X be a finite set, and let F be a family of distinct proper subsets

More information

(v, w) = arccos( < v, w >

(v, w) = arccos( < v, w > MA322 F all203 Notes on Inner Products Notes on Chapter 6 Inner product. Given a real vector space V, an inner product is defined to be a bilinear map F : V V R such that the following holds: For all v,

More information

STRUCTURE OF GEODESICS IN A 13-DIMENSIONAL GROUP OF HEISENBERG TYPE

STRUCTURE OF GEODESICS IN A 13-DIMENSIONAL GROUP OF HEISENBERG TYPE Steps in Differential Geometry, Proceedings of the Colloquium on Differential Geometry, 25 30 July, 2000, Debrecen, Hungary STRUCTURE OF GEODESICS IN A 13-DIMENSIONAL GROUP OF HEISENBERG TYPE Abstract.

More information

Lebesgue Measure on R n

Lebesgue Measure on R n 8 CHAPTER 2 Lebesgue Measure on R n Our goal is to construct a notion of the volume, or Lebesgue measure, of rather general subsets of R n that reduces to the usual volume of elementary geometrical sets

More information

MINIMAL GENERATING SETS OF GROUPS, RINGS, AND FIELDS

MINIMAL GENERATING SETS OF GROUPS, RINGS, AND FIELDS MINIMAL GENERATING SETS OF GROUPS, RINGS, AND FIELDS LORENZ HALBEISEN, MARTIN HAMILTON, AND PAVEL RŮŽIČKA Abstract. A subset X of a group (or a ring, or a field) is called generating, if the smallest subgroup

More information

UNIT 6: The singular value decomposition.

UNIT 6: The singular value decomposition. UNIT 6: The singular value decomposition. María Barbero Liñán Universidad Carlos III de Madrid Bachelor in Statistics and Business Mathematical methods II 2011-2012 A square matrix is symmetric if A T

More information

Lecture 2: Linear Algebra Review

Lecture 2: Linear Algebra Review EE 227A: Convex Optimization and Applications January 19 Lecture 2: Linear Algebra Review Lecturer: Mert Pilanci Reading assignment: Appendix C of BV. Sections 2-6 of the web textbook 1 2.1 Vectors 2.1.1

More information

ELA THE OPTIMAL PERTURBATION BOUNDS FOR THE WEIGHTED MOORE-PENROSE INVERSE. 1. Introduction. Let C m n be the set of complex m n matrices and C m n

ELA THE OPTIMAL PERTURBATION BOUNDS FOR THE WEIGHTED MOORE-PENROSE INVERSE. 1. Introduction. Let C m n be the set of complex m n matrices and C m n Electronic Journal of Linear Algebra ISSN 08-380 Volume 22, pp. 52-538, May 20 THE OPTIMAL PERTURBATION BOUNDS FOR THE WEIGHTED MOORE-PENROSE INVERSE WEI-WEI XU, LI-XIA CAI, AND WEN LI Abstract. In this

More information

SOLVING FUZZY LINEAR SYSTEMS BY USING THE SCHUR COMPLEMENT WHEN COEFFICIENT MATRIX IS AN M-MATRIX

SOLVING FUZZY LINEAR SYSTEMS BY USING THE SCHUR COMPLEMENT WHEN COEFFICIENT MATRIX IS AN M-MATRIX Iranian Journal of Fuzzy Systems Vol 5, No 3, 2008 pp 15-29 15 SOLVING FUZZY LINEAR SYSTEMS BY USING THE SCHUR COMPLEMENT WHEN COEFFICIENT MATRIX IS AN M-MATRIX M S HASHEMI, M K MIRNIA AND S SHAHMORAD

More information

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators. MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators. Adjoint operator and adjoint matrix Given a linear operator L on an inner product space V, the adjoint of L is a transformation

More information

The Cartan Dieudonné Theorem

The Cartan Dieudonné Theorem Chapter 7 The Cartan Dieudonné Theorem 7.1 Orthogonal Reflections Orthogonal symmetries are a very important example of isometries. First let us review the definition of a (linear) projection. Given a

More information

I. Multiple Choice Questions (Answer any eight)

I. Multiple Choice Questions (Answer any eight) Name of the student : Roll No : CS65: Linear Algebra and Random Processes Exam - Course Instructor : Prashanth L.A. Date : Sep-24, 27 Duration : 5 minutes INSTRUCTIONS: The test will be evaluated ONLY

More information

October 25, 2013 INNER PRODUCT SPACES

October 25, 2013 INNER PRODUCT SPACES October 25, 2013 INNER PRODUCT SPACES RODICA D. COSTIN Contents 1. Inner product 2 1.1. Inner product 2 1.2. Inner product spaces 4 2. Orthogonal bases 5 2.1. Existence of an orthogonal basis 7 2.2. Orthogonal

More information

Lanczos tridigonalization and Golub - Kahan bidiagonalization: Ideas, connections and impact

Lanczos tridigonalization and Golub - Kahan bidiagonalization: Ideas, connections and impact Lanczos tridigonalization and Golub - Kahan bidiagonalization: Ideas, connections and impact Zdeněk Strakoš Academy of Sciences and Charles University, Prague http://www.cs.cas.cz/ strakos Hong Kong, February

More information

Intrinsic products and factorizations of matrices

Intrinsic products and factorizations of matrices Available online at www.sciencedirect.com Linear Algebra and its Applications 428 (2008) 5 3 www.elsevier.com/locate/laa Intrinsic products and factorizations of matrices Miroslav Fiedler Academy of Sciences

More information

Linear Algebra (part 1) : Vector Spaces (by Evan Dummit, 2017, v. 1.07) 1.1 The Formal Denition of a Vector Space

Linear Algebra (part 1) : Vector Spaces (by Evan Dummit, 2017, v. 1.07) 1.1 The Formal Denition of a Vector Space Linear Algebra (part 1) : Vector Spaces (by Evan Dummit, 2017, v. 1.07) Contents 1 Vector Spaces 1 1.1 The Formal Denition of a Vector Space.................................. 1 1.2 Subspaces...................................................

More information

Pseudoinverse & Orthogonal Projection Operators

Pseudoinverse & Orthogonal Projection Operators Pseudoinverse & Orthogonal Projection Operators ECE 174 Linear & Nonlinear Optimization Ken Kreutz-Delgado ECE Department, UC San Diego Ken Kreutz-Delgado (UC San Diego) ECE 174 Fall 2016 1 / 48 The Four

More information

Linear Algebra Fundamentals

Linear Algebra Fundamentals Linear Algebra Fundamentals It can be argued that all of linear algebra can be understood using the four fundamental subspaces associated with a matrix. Because they form the foundation on which we later

More information

Theorem A.1. If A is any nonzero m x n matrix, then A is equivalent to a partitioned matrix of the form. k k n-k. m-k k m-k n-k

Theorem A.1. If A is any nonzero m x n matrix, then A is equivalent to a partitioned matrix of the form. k k n-k. m-k k m-k n-k I. REVIEW OF LINEAR ALGEBRA A. Equivalence Definition A1. If A and B are two m x n matrices, then A is equivalent to B if we can obtain B from A by a finite sequence of elementary row or elementary column

More information

Chapter 6 - Orthogonality

Chapter 6 - Orthogonality Chapter 6 - Orthogonality Maggie Myers Robert A. van de Geijn The University of Texas at Austin Orthogonality Fall 2009 http://z.cs.utexas.edu/wiki/pla.wiki/ 1 Orthogonal Vectors and Subspaces http://z.cs.utexas.edu/wiki/pla.wiki/

More information

(a) If A is a 3 by 4 matrix, what does this tell us about its nullspace? Solution: dim N(A) 1, since rank(a) 3. Ax =

(a) If A is a 3 by 4 matrix, what does this tell us about its nullspace? Solution: dim N(A) 1, since rank(a) 3. Ax = . (5 points) (a) If A is a 3 by 4 matrix, what does this tell us about its nullspace? dim N(A), since rank(a) 3. (b) If we also know that Ax = has no solution, what do we know about the rank of A? C(A)

More information

be a Householder matrix. Then prove the followings H = I 2 uut Hu = (I 2 uu u T u )u = u 2 uut u

be a Householder matrix. Then prove the followings H = I 2 uut Hu = (I 2 uu u T u )u = u 2 uut u MATH 434/534 Theoretical Assignment 7 Solution Chapter 7 (71) Let H = I 2uuT Hu = u (ii) Hv = v if = 0 be a Householder matrix Then prove the followings H = I 2 uut Hu = (I 2 uu )u = u 2 uut u = u 2u =

More information

Least Squares. Tom Lyche. October 26, Centre of Mathematics for Applications, Department of Informatics, University of Oslo

Least Squares. Tom Lyche. October 26, Centre of Mathematics for Applications, Department of Informatics, University of Oslo Least Squares Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University of Oslo October 26, 2010 Linear system Linear system Ax = b, A C m,n, b C m, x C n. under-determined

More information

12. Perturbed Matrices

12. Perturbed Matrices MAT334 : Applied Linear Algebra Mike Newman, winter 208 2. Perturbed Matrices motivation We want to solve a system Ax = b in a context where A and b are not known exactly. There might be experimental errors,

More information