Computing the nearest reversible Markov chain

Size: px
Start display at page:

Download "Computing the nearest reversible Markov chain"

Transcription

1 Konrad-Zuse-Zentrum für Informationstechnik Berlin Takustraße 7 D Berlin-Dahlem Germany ADAM NIELSEN AND MARCUS WEBER Computing the nearest reversible Markov chain ZIB Report (December 2014)

2 Herausgegeben vom Konrad-Zuse-Zentrum für Informationstechnik Berlin Takustraße 7 D Berlin-Dahlem Telefon: Telefax: bibliothek@zib.de URL: ZIB-Report (Print) ISSN ZIB-Report (Internet) ISSN

3 Computing the nearest reversible Markov chain Adam Nielsen Marcus Weber Abstract Reversible Markov chains are the basis of many applications. However, computing transition probabilities by a finite sampling of a Markov chain can lead to truncation errors. Even if the original Markov chain is reversible, the approximated Markov chain might be non-reversible will lose important properties, like the real valued spectrum. In this paper, we show how to find the closest reversible Markov chain to a given transition matrix. It turns out that this matrix can be computed by solving a convex minimization problem. 1 Introduction A Markov chain with n N states is described through a transition matrix P R n n, i.e. P ij = 1 P ij 0 for i, j = 1,..., n. A vector π R n is called a probability distribution if π i = 1 π j 0 i=1 for j = 1,..., n holds. A Markov chain or its corresponding transition matrix P is called reversible according to a probability distribution π if the following equation DP = P T D is valid for the diagonal matrix D = diag(π 1,..., π n ). stationary distribution of the Markov chain, i.e. π T P = π T. In this case, π is a The main result of this article is that for any transition Matrix P, any probability distribution π, any norm on R n n which is induced by a scalar product, there exists a unique transition matrix P which is reversible according to π has minimal distance to P with respect to the norm. This article is structured as follows. First, we will prove the above conjecture. Then, we show how to numerically obtain the reversible matrix P, discuss the computational cost include a perturbation analysis. Finally, we will give an application a numerical example. nielsen@zib.de weber@zib.de 1

4 2 Existence Proof Given a stochastic matrix P a stationary distribution π, it will be proven in the following that there exists a unique transition matrix P which minimizes P P is reversible according to π. First, let us consider U = {A R n n DA = A T D k R with A ij = k for all i = 1,..., n} where D = diag(π 1,..., π n ) denotes the diagonal matrix with values π i on the diagonal. Notice that U is a subspace of R n n because for A, B U with n a 1j = k 1 n i=1 b 1j = k 2 we obtain for any α, β R that D(αA + βb) = αda + βdb = αa T D + βb T D = (αa + βb) T D αa ij + βb ij = αk 1 + βk 2 for all i = 1,..., n. We can show the following property: Lemma 2.1. For A U with n A ij = k it holds π T A = kπ T. Proof. Since A U, we have π i A ij = π j A ji. Therefore, we obtain (π T A) i = π l A li = π i l=1 n l=1 A il = π i k. For an understing of the space U, the following matrices are of interest: π s if i = r j = s, π r if i = s j = r, A [r,s] 1 π s if i = j = r, ij = 1 π r if i = j = s, 1 if i = j s i r, 0 else, δ [r,s] ij = 1 if i = j i r, 1 if i = r j = s, 0 else. Proposition 2.1. The set {A [r,s] (r, s) I A } {δ [r,s], δ [s,r] (r, s) I B } {Id} 2

5 is a basis of U where Id is the identity matrix I = {(r, s) 1 r < s n}, A = {i: π i 0}, B = {i: π i = 0}, The dimension of U is given by I A = {(r, s) I r A or s A}, I B = {(r, s) I r, s B}. dim(u) = ( ) n ( ) B. 2 Proof. One may notice that the row sum of Id, A [r,s] δ [r,s] is always one that ( DA [r,s] = A [r,s]) T D holds. The latter can be seen by ( DA [r,s]) ij = k=1 = π i A [r,s] ij ( ) D ik A [r,s] kj = π j A [r,s] ji = A [r,s] ki D kj = k=1 ( ) (A [r,s] ) T D where ( ) holds because for i = r j = s or vice versa we have π i A [r,s] ij ij = π r π s = π s π r = π j A [r,s] ji, for i = j the equation in ( ) is trivial in all other cases we have A [r,s] A [r,s] ji = 0, therefore A [r,s] U. Also, we have Dδ [r,s] = δ [r,s]t D for all (r, s) I B. This is because π r = 0, therefore, { Dδ [r,s] π i if i = j, i,j = 0 else., ij = Since Dδ [r,s] is just a diagonal matrix, it is in particular symmetric we obtain Dδ [r,s] = (Dδ [r,s] ) T = δ [r,s]t D, the argument is analogous for δ [s,r]. To prove that these matrices are indeed a basis of U, it remains to show that they are linearly independent that they span the subspace U. 3

6 To see the linearly independence, consider now an arbitrary linear combination of zero: α r,s A [r,s] + α r,s δ [r,s] + β r,s δ [s,r] + αi = 0. (r,s) I A (r,s) I B For (r, s) I A the matrix A [r,s] is the only matrix in the above linear combination that could have a non-zero entry in row r column s in row s column r. Therefore, we obtain 0 = α r,s π s 0 = α r,s π r. Thus, π s 0 or π r 0 provides α r,s = 0. For (r, s) I B, we obtain analogously α r,s = 0 β r,s = 0. The linear combination reduces to α Id = 0 which, finally, leads us to α = 0. It remains to show that the given matrices span the subspace U. Thus, let us consider a matrix C U with n C ij = k for some k R. For (r, s) I A define α r,s := Csr π r if π r 0 otherwise α r,s := Crs π s. From π r C rs = π s C sr we get that For (r, s) I B choose α r,s A [r,s] rs = C rs α r,s A [r,s] sr = C sr. α r,s = C r,s β r,s = C s,r. Since each off-diagonal element appears in exactly one matrix, C differs from C := α r,s A [r,s] + α r,s δ [r,s] + β r,s δ [s,r] (r,s) I A (r,s) I B only in the diagonal. We also know that there exists l R with n C ij = l for i = 1,..., n since C U. Therefore, the matrix Ĉ := C + (k l) Id has row-sum k Ĉ ii = k Ĉ ij = k C ij = C ii, holds. Therefore,,j i C = Ĉ U,,j i which shows that the given matrices span U. Furthermore, dim U = I A + I B + I B + 1 = I + I B + 1. The statement follows from I = I B = ( ) n 2 ( ) B. 2 4

7 Our goal is to find a matrix P X = {A U A ij 0 for i, j = 1,..., n a 1j = 1} such that: P P A P for all A X. (1) To do so, let us characterize the set X. To simplify notation, let us denote the basis of Proposition 2.1 by (v i ) i=1,...,m with m = dim U v m = Id. For any A X we have a unique coefficent vector x R m with A = m i=1 x iv i. From n a ij = 1 we get that 1 = a ij = = = ( m ) x l v l (i, j) m l=1 x l m x l. l=1 l=1 v l (i, j) This can be rewritten in 1 T x = 1 where 1 R m is the constant vector 1 i = 1 for i = 1,..., m. Further, we have a ij 0 for all i j if only if x l 0 for all l = 1,..., m 1. This can be rewritten as xe i 0 for i = 1,..., m 1. The diagonal entries of A can be positive even if x m is negative. Let us quickly go back to the old notation to see how we have to define the restriction here. So let A be given by A = α r,s A [r,s] + α r,s δ [r,s] + β r,s δ [s,r] + αid, (r,s) I A (r,s) I B thus, a ii = α r,s A [r,s] ii + α r,s δ [r,s] ii (r,s) I A (r,s) I B Which leads to a ii = α r,s (1 π s )+ α r,s (1 π r )+ + β r,s δ [s,r] ii + αi ii. α r,s + α r,s + β r,s +α. (r,s) I A r=i (r,s) I A s=i (r,s) I A r i s (r,s) I B r i (r,s) I B s i The condition a ii 0 is equivalent to xg i 0 where 1 π s if v j = A [i,s] for some s > i, 1 π r if v j = A [r,i] for some r < i, g i (j) = 0 if v j = δ [i,s] for some s, 1 else, 5

8 for i = 1,..., n j = 1,..., m. Given the matrix e T 1. C = 1 e T m 1 g T R (n+m 1) m, 1. the condition that A = m i=1 x iv i is in set X is equivalent to g T n Cx 0 1 T x = 1. In order to find a matrix P X which satisfies inequality (1), let us recall that the norm is induced by a scalar product,, i.e. A = A, A for any matrix A R n n. We want to minimize the term with where m x i v i P i=1 2 = m m x i x j v i, v j 2 x i v i, P + P, P i, i=1 = 1 2 xt Qx + x T f + c (2) Cx 0 1 T x = 1, Q(i, j) := 2 v i, v j, f(i) = 2 v i, P c = P, P. Since Q is a Gram matrix of linear independent vectors, it is positive definite. This follows because for any x R n it holds x T Qx = n x i Q ij x j = x i v i, x i v i. i,j i=1 i=1 Since v 1,..., v n is a basis, we have that n i=1 x iv i 0 for x 0, in consequence, x T Qx > 0 for x 0. Since Q is positive definite, the quadratic function (2) is strongly convex. Therefore, we have formulated the problem into a strongly convex quadratic programming problem that attains its global minimum, since the quadratic function is coercive, continuous the set X is non-empty because of Id X, see [9] Theorem Also, the global minimum is unique because the quadratic function is strongly convex. Thus, we have proven the existence of a unique matrix P X which fulfills inequality (1). We will now discuss how to compute this matrix. 6

9 3 Complexity Perturbation To avoid technical difficulties, we will assume in this chapter that π i > 0 for i = 1,..., n. For A R n n the trace tr(a) is given by the sum of the diagonal elements tr(a) = A ii. It is known that i=1 A, A F := tr(a T B) is a scalar product on R n n for A, B R n n. This scalar product induces the Frobenius norm A F = A, A F = a ij 2. i=1 The following complexity analysis will be given according to the Frobenius norm. 3.1 Complexity Unfortunately, the matrix Q is quite large, in particular we have Q R m m where m = dim(u) = 1 + n2 2. Each entry of Q is given by a trace of two sparse matrices which can be computed by the following formula n π r π r π s π s if r, r, s, s are distinct, n 1 2π s + (1 π r )(1 π r ) if r r, s = s, A [r,s], A [r,s ] = n 1 2π r + (1 π s )(1 π s ) if r = r, s s, F n 1 + (1 π s )(1 π s ) + (1 π r )(1 π r ) if r = s or r = s, n 2 + (1 π r ) 2 + πr 2 + (1 π s ) 2 + πs 2 if r = r s = s. Note that because of r < s r < s other cases are not possible. We only give a proof for the case r r s = s, all other cases can be obtained analogously. First, if A [r,s] ij 0 for i j it follows that either i = r, j = s or i = s, j = r holds. Since r r, s = s we have A [r,s ] ij = 0, therefore, A [r,s], A [r,s ] = F i=1 A [r,s] ii A [r,s ] ii = n 3 + A [r,s] rr A [r,s ] rr + A [r,s],s ] r r A[r r r + A[r,s] ss A [r,s ] ss = n π s + 1 π s + (1 π r )(1 π r ) = n 1 2π s + (1 π r )(1 π r ). Therefore, the computational cost to evaluate a single entry of Q does not increase with n the effort to compute store Q is O(n 4 ). The convex minimization problem can be solved with a barrier method, e.g. the interior point method. The method consist of N Newton iterations.. 7

10 The number N of Newton iterations to find a strictly feasible point is bounded by N ( n + n 2 ) n + n 2 /2 1 /2 1 log γ ε where ε > 0 is the demed accuracy γ is a constant depending on the choice of backtracking parameter, see [1], section Therefore, the amount of Newton iterations is bounded by O(n log n). Unfortunately, each Newton iteration has to solve a linear equation system cost O((n 2 ) 3 ) since Q R ( n2 n2 2 +1) ( 2 +1). Therefore, the total cost for the optimization problem is bounded by O(n 7 log n). In the field of convex optimization, the upper bound for the amount of Newton iterations is known to be a large overestimation [1]. If we assume this amount to be independent from n, then we can expect that the time to find a solution with the convex optimization problem should be given by g(n) = αn 6 where n is the size of the matrix P R n n. If we include the bad estimation for the Newton iteration, then the time should be represented by f(n) = βn 7 log(n). In order to explore the computation time of the convex optimization problem, we generated for each n = 10,..., 100 a stochastic matrix P R n n a stochastic vector π R n where each entry was drawn from the stard uniform distribution on the open interval (0,1) then we normalized P π. We used Matlab R2012b on a 3 gigahertz computer with 8 gigabyte memory. We solved the convex optimization problem with the interior-point-convex algorithm from the provided Matlab method quadprog with default options, i.e. relative dual feasibility =2.31e 15 with TolFun=1e 0.8, complementarity measure =1.68e 10 with TolFun=1e 0.8 relative max constraint violoation = 0 with TolCon=1e 0.8. For n = 45 the execution time was about seconds. The scalars α, β where chosen such that f(45) = g(45) = holds. In Figure 1 one can see that g seems to be a reasonable approximation of the execution time. Also one can see that the execution time takes only seconds for matrices in R 50 50, but already 32 minutes for matrices in R Perturbation analysis For we obtain k := min r,r,s,s A [r,s], A [r,s ] F k Id Q n Id where the inequality has to be read componentwise. The upper bound follows from v i, v j F v i, v i F v j, v j F max { v l, v l F } l=1,...,m the fact that A [r,s], A [r,s] = n 2 + (1 π r ) 2 + πr 2 + (1 π s ) 2 + πs 2 n F 8

11 70 Minutes Time of minmization algorithm g f n Figure 1: Duration of convex minimization problem to find closest reversible Markov chain. Id, Id F = n holds. Since Q is symmetric, its condition number according to the spectral norm is given by κ(q) = λ max λ min n k. Since k = n c for some 0 < c < 4 we have that κ(q) 1 1 c/n 1 for n. Therefore, Q is well conditioned. We assume now that we are interested in finding a reversible matrix ˆP X, but that we only have a perturbed version P = ˆP + E of ˆP which is not necessarily reversible. With the above scheme, we can find a reversible matrix P X which is closest to P according to the Frobenius norm. The question arises, how eigenvalues change between P ˆP. In order to answer this question, the weighted Frobenius norm is introduced with A F := D 1 2 AD 1 2 F D = diag(π 1,..., π n ), D 1 2 D 1 2 = D D 1 2 := (D 1 2 ) 1. Note that the weighted Frobenius norm is given by A 2 F = i, a 2 π i i,j We can then give the following estimation about the eigenvalues: π j. 9

12 Theorem 3.1. Let A, B X, let λ 1,..., λ n be the eigenvalues of A ˆλ 1,..., ˆλ n be the eigenvalues of B. Then there exists a permutation σ of the integers 1, 2,..., n such that ˆλ σ(i) λ i 2 A B 2 F. i=1 Proof. The matrices A, B X are self-adjoint according to the scalar product x, y π := x T Dy. Lets denote with {w 1,..., w n } a, π -orthonormal basis, denote with W the matrix whose columns contain the vectors w i, i.e. W = w 1 w 2... w n. Then A := W 1 AW B = W 1 BW are symmetric, see [3] Chapter By the Hoffman Wielt Theorem (see Theorem in [5]) we obtain ˆλ σ(i) λ i 2 A B 2 F i=1 for a permutation σ, because similar matrices have the same eigenvalues. remains to show It or equivalently W 1 CW 2 F = C 2 F C 2 F = W CW 1 2 F for any matrix C R n. By construction of W we have Therefore, W T DW = I W 1 D 1 (W T ) 1 = I. (3) W CW 1 2 F = D 1 2 W CW 1 D F = tr(d 1 2 (W 1 ) T C T W T D 1 2 D 1 2 W CW 1 D 1 2 ) ( ) = tr(w 1 D 1 (W 1 ) T C T W T DW C) (3) = tr(c T C) = C 2 F, where in ( ) it is used that the trace is invariant under cyclic permutations. This shows that in order to guarantee good approximations for the eigenvalues, one has to assure a good approximation of a ij for those i, j where π j << π i. These transitions are also known as rare events often difficult to compute. We will end this chapter with a short experiment motivated by Theorem 3.1. We create 100 reversible Markov chains ( ˆP i ) i=1,...,100 R 5 5 as follows. First, 10

13 we generate a symmetric matrix A where each entry A ij with i j is chosen equally distributed from [0, 1] A ij := A ji for i > j. After normalizing A the resulting matrix is reversible [7] will be denoted as ˆP i. We perturb normalize ˆP i to obtain a perturbed version P i of ˆP i. For each P i we compute the unique reversible matrix P i which is closest to P i according to weighted Frobenius norm according to the stationary distribution of ˆP i. The matrix P i exists can be computed by the introduced convex optimization problem, because the weighted Frobenius norm is induced by the scalar product A, B F = tr(dad 1 B T ). Also we compute the unique reversible matrix ˇP i which is closest to P i according to stard Frobenius norm according to the stationary distribution of ˆPi. For each tuple ( ˆP, P, P, ˇP ) we compute the corresponding eigenvalues (ˆλ j, λ j, λ j, ˇλ j ),...,5. Then, we compute the numbers c 1 = min 5 ˆλ j λ σ(j) 2, c 2 = min 5 ˆλ j λ σ(j) 2 σ Π 5 σ Π 5 c 3 = min 5 ˆλ j ˇλ σ(j) 2, σ Π 5 where Π 5 denotes the set of all permutations of the numbers {1,..., 5}. The experiment is visualized in Figure 2. The result is that the closest reversible matrix ˇP according the Frobenius norm does not maintain the spectrum of ˆP. However, the closest reversible matrix according the weighted Frobenius norm P gives a good approximation for the spectrum of ˆP improves always the spectrum of the perturbed version P Weighted correction c1 Perturbation c Unweighted correction c3 Pertubation c ˆP P F ˆP ˇP F Figure 2: Comparison of c 1 c 2. Figure 3: Comparison of c 2 c 3. Three remarks may be in order: The resulting matrix P = α r,s A [r,s] + α Id (r,s) I 11

14 is reversible up to machine precision. In fact, it is reversible if π r α r,s π s = π s α r,s π r holds 1. A numerical error in the solution (x) i=1,...,m 1 = (α r,s ) (r,s) I of the convex optimization problem results in an error of the entry P (r, s) = α r,s π s P (s, r) = α r,s π r for r s. In order to guarantee that P has row sum one, it might be advantages to set p ii = 1 p ij. j i 4 Application numerical example In the past decades, analysis of a certain class of stochastic processes has been formulated in terms of approximating a transfer operator T : L 2 (µ) L 2 (µ) [12, 8, 11] which is self-adjoint according to the scalar product f, g µ = f(x) g(x) µ(dx), i.e. E T f, g µ = f, T g µ for all f, g L 2 (µ), where (E, Σ, µ) is a measure space for some set E R n, L 2 (µ) = {f : E R f, f µ <, fµ-measurable}/n with N = {f : E R A Σ with µ(a) = 0 f(x) = 0 for all x E\A}. Furthermore, the operator has the property that for f L 2 (µ) with f 0 it follows T f 0 almost surely. Let us assume that we have a set of non-negative functions {φ 1,..., φ n } given, such that n i=1 φ i(x) = 1 for all x E φ i, 1 µ > 0 for i = 1,..., n where 1(x) := 1 for all x E. If we define π i := φ i, 1 µ T ij := T φ i, φ j µ φ i, 1 µ, then it is straightforward to verify the properties (i) πt = π, (ii) DT = T T D, (iii) n T ij = 1 for all i = 1,..., n, (iv) T ij 0, 1 One may note that I A = I because of the assumption at the beginning of this chapter. 12

15 where D = diag(π 1,..., π n ) denotes the diagonal matrix of π. In other words, T is a reversible Markov chain. This matrix plays an essential role in the Galerkin discretization of T. Theorem 4.1. Let {φ 1,..., φ n } L 2 (µ) be a basis with φ i, 1 µ > 0 of a subspace D, Q: L 2 (µ) D the orthogonal projection onto D. For any self-adjoint continuous operator T : L 2 (µ) L 2 (µ) we have M = S 1 T, T ij = T φ i, φ j µ φ i, 1 µ, S ij = φ i, φ j µ φ i, 1 µ is a right matrix representation of QT Q according to the basis A = {φ 1,..., φ n }, i.e. for any f = α i φ i, QT Qf = β i φ i it holds i=1 i=1 M(α 1,..., α n ) T = (β 1,..., β n ) T. Proof. Consider the Gram matrix of {φ 1,..., φ n } Ŝ ij = φ i, φ j µ. This matrix is invertible since {φ 1,..., φ n } is a basis the orthogonal projection Q can be represented as Qv = Ŝ 1 ij v, φ i µ φ j. i, This can be verified by checking Qv v, g µ = 0 for all g D, v L 2 (µ). From ) S = D 1 Ŝ with D = diag ( φ 1, 1 µ,..., φ n, 1 µ we obtain S 1 = Ŝ 1 D, therefore, Ŝ 1 ij = S 1 ij 1 = S 1 1 ji, φ j, 1 µ φ i, 1 µ in the last step it was used that Ŝ 1 is symmetric since Ŝ is symmetric. This implies Qv = S 1 φ j ji v, φ i µ. φ i, 1 µ Therefore, QT Qφ k = QT φ k = = = i, i, i, S 1 φ j ij T φ k, φ i µ φ i, 1 µ S 1 T φ i, φ k µ ij φ j = φ i, 1 µ ( S 1 T ) kj φ j. ( n i=1 S 1 ji T ik ) φ j 13

16 For the more general case that the operator is not self-adjoint, one can find a similar Galerkin approximation, see [10], Theorem 3. Note that S is a special case of the matrix T with T = I, where I denotes the identity operator. Therefore, we can use our machinery to correct T S if π is given. One possible choice for the functions φ i are set based functions, i.e. we have disjoint sets A 1,..., A n with n i=1 A i = E { 1 if x A i φ i (x) = 1 Ai (x) = 0 else. In this case, the matrix S from Theorem 4.1 turns out to be the identity matrix we are only left with the task of computing T. Other possible choices are radial basis functions or commitor functions which can be found in [13, 10] So far, we know how to maintain reversibility for S T separately. The question arises, if S 1 T also inherits some properties that might get lost due to numerical errors. First, we can rewrite S = D 1 Ŝ, T = D 1 ˆT with Ŝij = φ i, φ j µ ˆT ij = T φ i, φ j µ, hence S 1 T = Ŝ 1 DD 1 ˆT = Ŝ 1 ˆT. Analogously as shown for Q also Ŝ Ŝ 1 are symmetric positive definite matrices. Now, since Ŝ 1 is positive definite, we know the existence of a symmetric square matrix A such that A 2 = Ŝ 1. Thus, A 1 Ŝ 1 ˆT A = A ˆT A. Consequently, Ŝ 1 ˆT is similar to a symmetric matrix hence diagonalizable. This shows that the spectrum of Ŝ 1 ˆT is real that we know the existence of a basis of eigenvectors of S 1 T. For numerical estimation, the following procedure may be advantageous: First, approximate T S by the convex minimization problem stated above. Then, calculate DS which is symmetric up to machine precision. For the symmetric matrix (DS ) 1 one can solve a convex minimization problem that finds the closest symmetric positive semi definite matrix ((DS ) 1 ) according to the Frobenius norm, see [4]. Finally, ((DS ) 1 ) T is the corrected matrix. Also it would be possible to correct S 1 with our explained optimization scheme, because this matrix has also row sum one is reversible, this follows from S1 = 1 1 = S 1 1 DS = S T D (S T ) 1 D = DS 1 (S 1 ) T D = DS 1. Therefore, by dropping the constraints given by the matrix C, one can also approximate S 1 as a convex quadratic programing problem with only linear constrains. 14

17 A Galerkin approximation of the transfer operator can be used to apply spectral clustering algorithms such as PCCA+ [2] which relies on the real spectrum of T S 1. Especially for a set-based reduction, we can guarantee with our machinery that T will have a real spectrum,, therefore, spectral clustering such as PCCA+ is applicable. 4.1 Numerical example Consider the one dimensional, 2π-periodic function V B : R R V B (x) = a + b cos(x) + c cos 2 (x) + d cos 3 (x) with a=2.0567, b= , c= d= This can be seen as an approximation of a butane potential energy function, where x is the central dihedral angle. We are going to realize a trajectory of the dihedral angles X t = X t mod 2π of butane from the stochastic differential equation d X t = V B ( X t )dt + σdb t. If we divide [0, 2π] in the sets A i = [ i 1 30, i 30 ] for i = 1,..., 30, then the Galerkin discretization of T reduces to T ij = P(X τ A j X 0 A i ), see [8]. To gain the associated Markov State Model T R 30 30, we compute a longterm trajectory (X i ) i=0,...,n 1 by performing n timesteps of size dt using the Euler-Maruyama discretization X i+1 = X i V B ( X i )dt + σ dt η i, where η i are i.i.d rom variables distributed according to the stard normal distribution. This trajectory is chopped into pieces of length 400 yielding M subtrajectories (X k i ) i=1,...,l := (X lk,..., X l(k+1) 1 ) for k = 0,..., M 1. We can estimate T by counting transitions C ij = M 1 k=0 1 Ai (X k 1 ) 1 Aj (X k l ) C ij T ij = 30 i=1 C. ij The Markov State Model created in this way becomes only reasonable when considered for a trajectory longer then 10 6 timesteps, due to rare transition events. We will now compare the eigenvalues of the Markov State Model T to the eigenvalues of the above introduced corrected estimation T with respect to the weighted Frobenius norm F. We will also compare them for different length of a given trajectory, starting from length n = 10 6 until n = We used the timestep dt = Since the eigenvalues of T turned out to be complex sometimes, we simply set the imaginary part to zero, in order to compare the eigenvalues with T. 15

18 Corrected estimation Stard estimation nd eigenvalue Amount of trajectories (x10 7 ) Discussion In the figures the stard estimation the corrected estimation do not essentially differ for the eigenvalues close to 1. Our algorithm of finding the nearest reversible Markov chain especially preserves the dominant eigenvalues of the underlying operator, thus, is suitable for cluster analysis 2 of Molecular simulation data. In contrast to T, the resulting transition matrix T from our algorithm is guaranteed to be applicable to spectral clustering methods, such as PCCA+, because of its real spectrum. Acknowledgment We gratefully thank the Berlin Mathematical School for financial support of Adam Nielsen. 2 For the connection between cluster analysis eigenvalues we refer to [6]. 16

19 0.85 Corrected estimation Stard estimation 0.8 3rd eigenvalue Amount of trajectories (x10 7 ) Corrected estimation Stard estimation 0.3 4th eigenvalue Amount of trajectories (x10 7 ) 17

20 Corrected estimation Stard estimation 0 5th eigenvalue Amount of trajectories (x10 7 ) Corrected estimation Stard estimation 0 6th eigenvalue Amount of trajectories (x10 7 ) 18

21 References [1] S. Boyd L. Venberghe, Convex optimization, Cambridge University Press, [2] Peter Deuflhard M. Weber, Robust Perron cluster analysis in conformation dynamics, In M. Dellnitz, S. Kirkl, M. Neumann, C. Schütte, editors, Linear Algebra Its Applications - Special Issue on Matrices Mathematical Biology,volume 398C (2005), [3] G. Fischer, Lineare algebra, 17 ed., Vieweg + Teubner, [4] Nicholas J. Higham, Computing a nearest symmetric positive semidefinite matrix, Linear Algebra its Applications 103 (1988), ii iii. [5] R. Horn C. Johnson, Matrix analysis, Cambridge University Press, [6] Wilhelm Huisinga Bernd Schmidt, Metastability dominan eigenvaues of transfer operators, Lecture Notes in Computational Science Engineering 49 (2006), [7] Ph. Metzner, F. Noé, Ch. Schütte, Estimating the sampling error: Distribution of transition matrices functions of transition matrices for given trajectory data, Physical Review E, 80 (2) [8] A. Nielsen, K.Fackeldey, M.Weber, Transfer operator for arbitary measures, in submission. [9] N. V. Thoai R. Horst, P. M. Pardalos, Introduction to global optimization, Kluwer Academic Publishers, [10] M. Sarich, Projected transfer operators - discretization of Markov processes in high-dimensional state space., Ph.D. thesis, Freie Universität Berlin, [11] C. Schütte, Conformational dynamics: Modelling, theory, algorithm, application to biomolecules., habilitation, Freie Universität Berlin, [12] Christof Schütte, Wilhelm Huisinga, Peter Deuflhard, Transfer operator approach to conformational dynamics in biomolecular systems, Applied Mathematics Entering the 21th Century (2003). [13] Marcus Weber, A subspace approach to molecular Markov state models via a new infinitesimal generator., habilitation, Freie Universität Berlin,

The Monte Carlo Computation Error of Transition Probabilities

The Monte Carlo Computation Error of Transition Probabilities Zuse Institute Berlin Takustr. 7 495 Berlin Germany ADAM NILSN The Monte Carlo Computation rror of Transition Probabilities ZIB Report 6-37 (July 206) Zuse Institute Berlin Takustr. 7 D-495 Berlin Telefon:

More information

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra. DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1

More information

Review problems for MA 54, Fall 2004.

Review problems for MA 54, Fall 2004. Review problems for MA 54, Fall 2004. Below are the review problems for the final. They are mostly homework problems, or very similar. If you are comfortable doing these problems, you should be fine on

More information

Boolean Inner-Product Spaces and Boolean Matrices

Boolean Inner-Product Spaces and Boolean Matrices Boolean Inner-Product Spaces and Boolean Matrices Stan Gudder Department of Mathematics, University of Denver, Denver CO 80208 Frédéric Latrémolière Department of Mathematics, University of Denver, Denver

More information

Throughout these notes we assume V, W are finite dimensional inner product spaces over C.

Throughout these notes we assume V, W are finite dimensional inner product spaces over C. Math 342 - Linear Algebra II Notes Throughout these notes we assume V, W are finite dimensional inner product spaces over C 1 Upper Triangular Representation Proposition: Let T L(V ) There exists an orthonormal

More information

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP) MATH 20F: LINEAR ALGEBRA LECTURE B00 (T KEMP) Definition 01 If T (x) = Ax is a linear transformation from R n to R m then Nul (T ) = {x R n : T (x) = 0} = Nul (A) Ran (T ) = {Ax R m : x R n } = {b R m

More information

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces. Math 350 Fall 2011 Notes about inner product spaces In this notes we state and prove some important properties of inner product spaces. First, recall the dot product on R n : if x, y R n, say x = (x 1,...,

More information

Linear Algebra Massoud Malek

Linear Algebra Massoud Malek CSUEB Linear Algebra Massoud Malek Inner Product and Normed Space In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n An inner product

More information

Mathematical Optimisation, Chpt 2: Linear Equations and inequalities

Mathematical Optimisation, Chpt 2: Linear Equations and inequalities Mathematical Optimisation, Chpt 2: Linear Equations and inequalities Peter J.C. Dickinson p.j.c.dickinson@utwente.nl http://dickinson.website version: 12/02/18 Monday 5th February 2018 Peter J.C. Dickinson

More information

Spectral Theorem for Self-adjoint Linear Operators

Spectral Theorem for Self-adjoint Linear Operators Notes for the undergraduate lecture by David Adams. (These are the notes I would write if I was teaching a course on this topic. I have included more material than I will cover in the 45 minute lecture;

More information

REVIEW FOR EXAM III SIMILARITY AND DIAGONALIZATION

REVIEW FOR EXAM III SIMILARITY AND DIAGONALIZATION REVIEW FOR EXAM III The exam covers sections 4.4, the portions of 4. on systems of differential equations and on Markov chains, and..4. SIMILARITY AND DIAGONALIZATION. Two matrices A and B are similar

More information

The University of Texas at Austin Department of Electrical and Computer Engineering. EE381V: Large Scale Learning Spring 2013.

The University of Texas at Austin Department of Electrical and Computer Engineering. EE381V: Large Scale Learning Spring 2013. The University of Texas at Austin Department of Electrical and Computer Engineering EE381V: Large Scale Learning Spring 2013 Assignment Two Caramanis/Sanghavi Due: Tuesday, Feb. 19, 2013. Computational

More information

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2. APPENDIX A Background Mathematics A. Linear Algebra A.. Vector algebra Let x denote the n-dimensional column vector with components 0 x x 2 B C @. A x n Definition 6 (scalar product). The scalar product

More information

Math 443 Differential Geometry Spring Handout 3: Bilinear and Quadratic Forms This handout should be read just before Chapter 4 of the textbook.

Math 443 Differential Geometry Spring Handout 3: Bilinear and Quadratic Forms This handout should be read just before Chapter 4 of the textbook. Math 443 Differential Geometry Spring 2013 Handout 3: Bilinear and Quadratic Forms This handout should be read just before Chapter 4 of the textbook. Endomorphisms of a Vector Space This handout discusses

More information

Clustering by using a simplex structure

Clustering by using a simplex structure Konrad-Zuse-Zentrum für Informationstechnik Berlin Takustraße 7 D-14195 Berlin-Dahlem Germany MARCUS WEBER Clustering by using a simplex structure supported by DFG Research Center Mathematics for Key Technologies

More information

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.

More information

Lecture 1. 1 Conic programming. MA 796S: Convex Optimization and Interior Point Methods October 8, Consider the conic program. min.

Lecture 1. 1 Conic programming. MA 796S: Convex Optimization and Interior Point Methods October 8, Consider the conic program. min. MA 796S: Convex Optimization and Interior Point Methods October 8, 2007 Lecture 1 Lecturer: Kartik Sivaramakrishnan Scribe: Kartik Sivaramakrishnan 1 Conic programming Consider the conic program min s.t.

More information

14 Singular Value Decomposition

14 Singular Value Decomposition 14 Singular Value Decomposition For any high-dimensional data analysis, one s first thought should often be: can I use an SVD? The singular value decomposition is an invaluable analysis tool for dealing

More information

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v ) Section 3.2 Theorem 3.6. Let A be an m n matrix of rank r. Then r m, r n, and, by means of a finite number of elementary row and column operations, A can be transformed into the matrix ( ) Ir O D = 1 O

More information

CS 246 Review of Linear Algebra 01/17/19

CS 246 Review of Linear Algebra 01/17/19 1 Linear algebra In this section we will discuss vectors and matrices. We denote the (i, j)th entry of a matrix A as A ij, and the ith entry of a vector as v i. 1.1 Vectors and vector operations A vector

More information

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS LINEAR ALGEBRA, -I PARTIAL EXAM SOLUTIONS TO PRACTICE PROBLEMS Problem (a) For each of the two matrices below, (i) determine whether it is diagonalizable, (ii) determine whether it is orthogonally diagonalizable,

More information

Jim Lambers MAT 610 Summer Session Lecture 2 Notes

Jim Lambers MAT 610 Summer Session Lecture 2 Notes Jim Lambers MAT 610 Summer Session 2009-10 Lecture 2 Notes These notes correspond to Sections 2.2-2.4 in the text. Vector Norms Given vectors x and y of length one, which are simply scalars x and y, the

More information

Linear Algebra 2 Spectral Notes

Linear Algebra 2 Spectral Notes Linear Algebra 2 Spectral Notes In what follows, V is an inner product vector space over F, where F = R or C. We will use results seen so far; in particular that every linear operator T L(V ) has a complex

More information

Chapter 7: Symmetric Matrices and Quadratic Forms

Chapter 7: Symmetric Matrices and Quadratic Forms Chapter 7: Symmetric Matrices and Quadratic Forms (Last Updated: December, 06) These notes are derived primarily from Linear Algebra and its applications by David Lay (4ed). A few theorems have been moved

More information

Quadratic forms. Here. Thus symmetric matrices are diagonalizable, and the diagonalization can be performed by means of an orthogonal matrix.

Quadratic forms. Here. Thus symmetric matrices are diagonalizable, and the diagonalization can be performed by means of an orthogonal matrix. Quadratic forms 1. Symmetric matrices An n n matrix (a ij ) n ij=1 with entries on R is called symmetric if A T, that is, if a ij = a ji for all 1 i, j n. We denote by S n (R) the set of all n n symmetric

More information

7. Symmetric Matrices and Quadratic Forms

7. Symmetric Matrices and Quadratic Forms Linear Algebra 7. Symmetric Matrices and Quadratic Forms CSIE NCU 1 7. Symmetric Matrices and Quadratic Forms 7.1 Diagonalization of symmetric matrices 2 7.2 Quadratic forms.. 9 7.4 The singular value

More information

MATH 1120 (LINEAR ALGEBRA 1), FINAL EXAM FALL 2011 SOLUTIONS TO PRACTICE VERSION

MATH 1120 (LINEAR ALGEBRA 1), FINAL EXAM FALL 2011 SOLUTIONS TO PRACTICE VERSION MATH (LINEAR ALGEBRA ) FINAL EXAM FALL SOLUTIONS TO PRACTICE VERSION Problem (a) For each matrix below (i) find a basis for its column space (ii) find a basis for its row space (iii) determine whether

More information

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET This is a (not quite comprehensive) list of definitions and theorems given in Math 1553. Pay particular attention to the ones in red. Study Tip For each

More information

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 2

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 2 EE/ACM 150 - Applications of Convex Optimization in Signal Processing and Communications Lecture 2 Andre Tkacenko Signal Processing Research Group Jet Propulsion Laboratory April 5, 2012 Andre Tkacenko

More information

Lecture 7: Positive Semidefinite Matrices

Lecture 7: Positive Semidefinite Matrices Lecture 7: Positive Semidefinite Matrices Rajat Mittal IIT Kanpur The main aim of this lecture note is to prepare your background for semidefinite programming. We have already seen some linear algebra.

More information

An indicator for the number of clusters using a linear map to simplex structure

An indicator for the number of clusters using a linear map to simplex structure An indicator for the number of clusters using a linear map to simplex structure Marcus Weber, Wasinee Rungsarityotin, and Alexander Schliep Zuse Institute Berlin ZIB Takustraße 7, D-495 Berlin, Germany

More information

Recall the convention that, for us, all vectors are column vectors.

Recall the convention that, for us, all vectors are column vectors. Some linear algebra Recall the convention that, for us, all vectors are column vectors. 1. Symmetric matrices Let A be a real matrix. Recall that a complex number λ is an eigenvalue of A if there exists

More information

1 Last time: least-squares problems

1 Last time: least-squares problems MATH Linear algebra (Fall 07) Lecture Last time: least-squares problems Definition. If A is an m n matrix and b R m, then a least-squares solution to the linear system Ax = b is a vector x R n such that

More information

Diagonalizing Matrices

Diagonalizing Matrices Diagonalizing Matrices Massoud Malek A A Let A = A k be an n n non-singular matrix and let B = A = [B, B,, B k,, B n ] Then A n A B = A A 0 0 A k [B, B,, B k,, B n ] = 0 0 = I n 0 A n Notice that A i B

More information

22m:033 Notes: 7.1 Diagonalization of Symmetric Matrices

22m:033 Notes: 7.1 Diagonalization of Symmetric Matrices m:33 Notes: 7. Diagonalization of Symmetric Matrices Dennis Roseman University of Iowa Iowa City, IA http://www.math.uiowa.edu/ roseman May 3, Symmetric matrices Definition. A symmetric matrix is a matrix

More information

The following definition is fundamental.

The following definition is fundamental. 1. Some Basics from Linear Algebra With these notes, I will try and clarify certain topics that I only quickly mention in class. First and foremost, I will assume that you are familiar with many basic

More information

Chapter 3 Transformations

Chapter 3 Transformations Chapter 3 Transformations An Introduction to Optimization Spring, 2014 Wei-Ta Chu 1 Linear Transformations A function is called a linear transformation if 1. for every and 2. for every If we fix the bases

More information

Linear Algebra- Final Exam Review

Linear Algebra- Final Exam Review Linear Algebra- Final Exam Review. Let A be invertible. Show that, if v, v, v 3 are linearly independent vectors, so are Av, Av, Av 3. NOTE: It should be clear from your answer that you know the definition.

More information

18.S34 linear algebra problems (2007)

18.S34 linear algebra problems (2007) 18.S34 linear algebra problems (2007) Useful ideas for evaluating determinants 1. Row reduction, expanding by minors, or combinations thereof; sometimes these are useful in combination with an induction

More information

0.1 Rational Canonical Forms

0.1 Rational Canonical Forms We have already seen that it is useful and simpler to study linear systems using matrices. But matrices are themselves cumbersome, as they are stuffed with many entries, and it turns out that it s best

More information

CSC Linear Programming and Combinatorial Optimization Lecture 10: Semidefinite Programming

CSC Linear Programming and Combinatorial Optimization Lecture 10: Semidefinite Programming CSC2411 - Linear Programming and Combinatorial Optimization Lecture 10: Semidefinite Programming Notes taken by Mike Jamieson March 28, 2005 Summary: In this lecture, we introduce semidefinite programming

More information

Math 321: Linear Algebra

Math 321: Linear Algebra Math 32: Linear Algebra T. Kapitula Department of Mathematics and Statistics University of New Mexico September 8, 24 Textbook: Linear Algebra,by J. Hefferon E-mail: kapitula@math.unm.edu Prof. Kapitula,

More information

Methods for sparse analysis of high-dimensional data, II

Methods for sparse analysis of high-dimensional data, II Methods for sparse analysis of high-dimensional data, II Rachel Ward May 26, 2011 High dimensional data with low-dimensional structure 300 by 300 pixel images = 90, 000 dimensions 2 / 55 High dimensional

More information

Mathematical Methods wk 2: Linear Operators

Mathematical Methods wk 2: Linear Operators John Magorrian, magog@thphysoxacuk These are work-in-progress notes for the second-year course on mathematical methods The most up-to-date version is available from http://www-thphysphysicsoxacuk/people/johnmagorrian/mm

More information

CONCENTRATION OF THE MIXED DISCRIMINANT OF WELL-CONDITIONED MATRICES. Alexander Barvinok

CONCENTRATION OF THE MIXED DISCRIMINANT OF WELL-CONDITIONED MATRICES. Alexander Barvinok CONCENTRATION OF THE MIXED DISCRIMINANT OF WELL-CONDITIONED MATRICES Alexander Barvinok Abstract. We call an n-tuple Q 1,...,Q n of positive definite n n real matrices α-conditioned for some α 1 if for

More information

MATH SOLUTIONS TO PRACTICE MIDTERM LECTURE 1, SUMMER Given vector spaces V and W, V W is the vector space given by

MATH SOLUTIONS TO PRACTICE MIDTERM LECTURE 1, SUMMER Given vector spaces V and W, V W is the vector space given by MATH 110 - SOLUTIONS TO PRACTICE MIDTERM LECTURE 1, SUMMER 2009 GSI: SANTIAGO CAÑEZ 1. Given vector spaces V and W, V W is the vector space given by V W = {(v, w) v V and w W }, with addition and scalar

More information

Kernels of Directed Graph Laplacians. J. S. Caughman and J.J.P. Veerman

Kernels of Directed Graph Laplacians. J. S. Caughman and J.J.P. Veerman Kernels of Directed Graph Laplacians J. S. Caughman and J.J.P. Veerman Department of Mathematics and Statistics Portland State University PO Box 751, Portland, OR 97207. caughman@pdx.edu, veerman@pdx.edu

More information

Basic Elements of Linear Algebra

Basic Elements of Linear Algebra A Basic Review of Linear Algebra Nick West nickwest@stanfordedu September 16, 2010 Part I Basic Elements of Linear Algebra Although the subject of linear algebra is much broader than just vectors and matrices,

More information

1 9/5 Matrices, vectors, and their applications

1 9/5 Matrices, vectors, and their applications 1 9/5 Matrices, vectors, and their applications Algebra: study of objects and operations on them. Linear algebra: object: matrices and vectors. operations: addition, multiplication etc. Algorithms/Geometric

More information

Math Linear Algebra II. 1. Inner Products and Norms

Math Linear Algebra II. 1. Inner Products and Norms Math 342 - Linear Algebra II Notes 1. Inner Products and Norms One knows from a basic introduction to vectors in R n Math 254 at OSU) that the length of a vector x = x 1 x 2... x n ) T R n, denoted x,

More information

MATH 304 Linear Algebra Lecture 23: Diagonalization. Review for Test 2.

MATH 304 Linear Algebra Lecture 23: Diagonalization. Review for Test 2. MATH 304 Linear Algebra Lecture 23: Diagonalization. Review for Test 2. Diagonalization Let L be a linear operator on a finite-dimensional vector space V. Then the following conditions are equivalent:

More information

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88 Math Camp 2010 Lecture 4: Linear Algebra Xiao Yu Wang MIT Aug 2010 Xiao Yu Wang (MIT) Math Camp 2010 08/10 1 / 88 Linear Algebra Game Plan Vector Spaces Linear Transformations and Matrices Determinant

More information

Spanning and Independence Properties of Finite Frames

Spanning and Independence Properties of Finite Frames Chapter 1 Spanning and Independence Properties of Finite Frames Peter G. Casazza and Darrin Speegle Abstract The fundamental notion of frame theory is redundancy. It is this property which makes frames

More information

On the projection onto a finitely generated cone

On the projection onto a finitely generated cone Acta Cybernetica 00 (0000) 1 15. On the projection onto a finitely generated cone Miklós Ujvári Abstract In the paper we study the properties of the projection onto a finitely generated cone. We show for

More information

1. General Vector Spaces

1. General Vector Spaces 1.1. Vector space axioms. 1. General Vector Spaces Definition 1.1. Let V be a nonempty set of objects on which the operations of addition and scalar multiplication are defined. By addition we mean a rule

More information

Then x 1,..., x n is a basis as desired. Indeed, it suffices to verify that it spans V, since n = dim(v ). We may write any v V as r

Then x 1,..., x n is a basis as desired. Indeed, it suffices to verify that it spans V, since n = dim(v ). We may write any v V as r Practice final solutions. I did not include definitions which you can find in Axler or in the course notes. These solutions are on the terse side, but would be acceptable in the final. However, if you

More information

Uniqueness of the Solutions of Some Completion Problems

Uniqueness of the Solutions of Some Completion Problems Uniqueness of the Solutions of Some Completion Problems Chi-Kwong Li and Tom Milligan Abstract We determine the conditions for uniqueness of the solutions of several completion problems including the positive

More information

Linear Algebra Primer

Linear Algebra Primer Linear Algebra Primer David Doria daviddoria@gmail.com Wednesday 3 rd December, 2008 Contents Why is it called Linear Algebra? 4 2 What is a Matrix? 4 2. Input and Output.....................................

More information

MAT Linear Algebra Collection of sample exams

MAT Linear Algebra Collection of sample exams MAT 342 - Linear Algebra Collection of sample exams A-x. (0 pts Give the precise definition of the row echelon form. 2. ( 0 pts After performing row reductions on the augmented matrix for a certain system

More information

MATH 235: Inner Product Spaces, Assignment 7

MATH 235: Inner Product Spaces, Assignment 7 MATH 235: Inner Product Spaces, Assignment 7 Hand in questions 3,4,5,6,9, by 9:3 am on Wednesday March 26, 28. Contents Orthogonal Basis for Inner Product Space 2 2 Inner-Product Function Space 2 3 Weighted

More information

Signal Recovery from Permuted Observations

Signal Recovery from Permuted Observations EE381V Course Project Signal Recovery from Permuted Observations 1 Problem Shanshan Wu (sw33323) May 8th, 2015 We start with the following problem: let s R n be an unknown n-dimensional real-valued signal,

More information

Real symmetric matrices/1. 1 Eigenvalues and eigenvectors

Real symmetric matrices/1. 1 Eigenvalues and eigenvectors Real symmetric matrices 1 Eigenvalues and eigenvectors We use the convention that vectors are row vectors and matrices act on the right. Let A be a square matrix with entries in a field F; suppose that

More information

forms Christopher Engström November 14, 2014 MAA704: Matrix factorization and canonical forms Matrix properties Matrix factorization Canonical forms

forms Christopher Engström November 14, 2014 MAA704: Matrix factorization and canonical forms Matrix properties Matrix factorization Canonical forms Christopher Engström November 14, 2014 Hermitian LU QR echelon Contents of todays lecture Some interesting / useful / important of matrices Hermitian LU QR echelon Rewriting a as a product of several matrices.

More information

Semidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 5

Semidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 5 Semidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 5 Instructor: Farid Alizadeh Scribe: Anton Riabov 10/08/2001 1 Overview We continue studying the maximum eigenvalue SDP, and generalize

More information

Linear Algebra and Eigenproblems

Linear Algebra and Eigenproblems Appendix A A Linear Algebra and Eigenproblems A working knowledge of linear algebra is key to understanding many of the issues raised in this work. In particular, many of the discussions of the details

More information

Exercises * on Linear Algebra

Exercises * on Linear Algebra Exercises * on Linear Algebra Laurenz Wiskott Institut für Neuroinformatik Ruhr-Universität Bochum, Germany, EU 4 February 7 Contents Vector spaces 4. Definition...............................................

More information

Linear Algebra Solutions 1

Linear Algebra Solutions 1 Math Camp 1 Do the following: Linear Algebra Solutions 1 1. Let A = and B = 3 8 5 A B = 3 5 9 A + B = 9 11 14 4 AB = 69 3 16 BA = 1 4 ( 1 3. Let v = and u = 5 uv = 13 u v = 13 v u = 13 Math Camp 1 ( 7

More information

An Introduction to Correlation Stress Testing

An Introduction to Correlation Stress Testing An Introduction to Correlation Stress Testing Defeng Sun Department of Mathematics and Risk Management Institute National University of Singapore This is based on a joint work with GAO Yan at NUS March

More information

The Hilbert Space of Random Variables

The Hilbert Space of Random Variables The Hilbert Space of Random Variables Electrical Engineering 126 (UC Berkeley) Spring 2018 1 Outline Fix a probability space and consider the set H := {X : X is a real-valued random variable with E[X 2

More information

Copositive matrices and periodic dynamical systems

Copositive matrices and periodic dynamical systems Extreme copositive matrices and periodic dynamical systems Weierstrass Institute (WIAS), Berlin Optimization without borders Dedicated to Yuri Nesterovs 60th birthday February 11, 2016 and periodic dynamical

More information

Research Note. A New Infeasible Interior-Point Algorithm with Full Nesterov-Todd Step for Semi-Definite Optimization

Research Note. A New Infeasible Interior-Point Algorithm with Full Nesterov-Todd Step for Semi-Definite Optimization Iranian Journal of Operations Research Vol. 4, No. 1, 2013, pp. 88-107 Research Note A New Infeasible Interior-Point Algorithm with Full Nesterov-Todd Step for Semi-Definite Optimization B. Kheirfam We

More information

Lecture 2: Linear Algebra Review

Lecture 2: Linear Algebra Review EE 227A: Convex Optimization and Applications January 19 Lecture 2: Linear Algebra Review Lecturer: Mert Pilanci Reading assignment: Appendix C of BV. Sections 2-6 of the web textbook 1 2.1 Vectors 2.1.1

More information

8. Diagonalization.

8. Diagonalization. 8. Diagonalization 8.1. Matrix Representations of Linear Transformations Matrix of A Linear Operator with Respect to A Basis We know that every linear transformation T: R n R m has an associated standard

More information

MATHEMATICS 217 NOTES

MATHEMATICS 217 NOTES MATHEMATICS 27 NOTES PART I THE JORDAN CANONICAL FORM The characteristic polynomial of an n n matrix A is the polynomial χ A (λ) = det(λi A), a monic polynomial of degree n; a monic polynomial in the variable

More information

Methods for sparse analysis of high-dimensional data, II

Methods for sparse analysis of high-dimensional data, II Methods for sparse analysis of high-dimensional data, II Rachel Ward May 23, 2011 High dimensional data with low-dimensional structure 300 by 300 pixel images = 90, 000 dimensions 2 / 47 High dimensional

More information

For δa E, this motivates the definition of the Bauer-Skeel condition number ([2], [3], [14], [15])

For δa E, this motivates the definition of the Bauer-Skeel condition number ([2], [3], [14], [15]) LAA 278, pp.2-32, 998 STRUCTURED PERTURBATIONS AND SYMMETRIC MATRICES SIEGFRIED M. RUMP Abstract. For a given n by n matrix the ratio between the componentwise distance to the nearest singular matrix and

More information

CHAPTER 3 Further properties of splines and B-splines

CHAPTER 3 Further properties of splines and B-splines CHAPTER 3 Further properties of splines and B-splines In Chapter 2 we established some of the most elementary properties of B-splines. In this chapter our focus is on the question What kind of functions

More information

Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008

Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008 Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008 Exam 2 will be held on Tuesday, April 8, 7-8pm in 117 MacMillan What will be covered The exam will cover material from the lectures

More information

1 Vectors. Notes for Bindel, Spring 2017 Numerical Analysis (CS 4220)

1 Vectors. Notes for Bindel, Spring 2017 Numerical Analysis (CS 4220) Notes for 2017-01-30 Most of mathematics is best learned by doing. Linear algebra is no exception. You have had a previous class in which you learned the basics of linear algebra, and you will have plenty

More information

Homework 1. Yuan Yao. September 18, 2011

Homework 1. Yuan Yao. September 18, 2011 Homework 1 Yuan Yao September 18, 2011 1. Singular Value Decomposition: The goal of this exercise is to refresh your memory about the singular value decomposition and matrix norms. A good reference to

More information

GQE ALGEBRA PROBLEMS

GQE ALGEBRA PROBLEMS GQE ALGEBRA PROBLEMS JAKOB STREIPEL Contents. Eigenthings 2. Norms, Inner Products, Orthogonality, and Such 6 3. Determinants, Inverses, and Linear (In)dependence 4. (Invariant) Subspaces 3 Throughout

More information

Math 3191 Applied Linear Algebra

Math 3191 Applied Linear Algebra Math 9 Applied Linear Algebra Lecture 9: Diagonalization Stephen Billups University of Colorado at Denver Math 9Applied Linear Algebra p./9 Section. Diagonalization The goal here is to develop a useful

More information

Detailed Proof of The PerronFrobenius Theorem

Detailed Proof of The PerronFrobenius Theorem Detailed Proof of The PerronFrobenius Theorem Arseny M Shur Ural Federal University October 30, 2016 1 Introduction This famous theorem has numerous applications, but to apply it you should understand

More information

Mathematical Methods wk 1: Vectors

Mathematical Methods wk 1: Vectors Mathematical Methods wk : Vectors John Magorrian, magog@thphysoxacuk These are work-in-progress notes for the second-year course on mathematical methods The most up-to-date version is available from http://www-thphysphysicsoxacuk/people/johnmagorrian/mm

More information

Mathematical Methods wk 1: Vectors

Mathematical Methods wk 1: Vectors Mathematical Methods wk : Vectors John Magorrian, magog@thphysoxacuk These are work-in-progress notes for the second-year course on mathematical methods The most up-to-date version is available from http://www-thphysphysicsoxacuk/people/johnmagorrian/mm

More information

Lecture 2: Linear operators

Lecture 2: Linear operators Lecture 2: Linear operators Rajat Mittal IIT Kanpur The mathematical formulation of Quantum computing requires vector spaces and linear operators So, we need to be comfortable with linear algebra to study

More information

2nd Symposium on System, Structure and Control, Oaxaca, 2004

2nd Symposium on System, Structure and Control, Oaxaca, 2004 263 2nd Symposium on System, Structure and Control, Oaxaca, 2004 A PROJECTIVE ALGORITHM FOR STATIC OUTPUT FEEDBACK STABILIZATION Kaiyang Yang, Robert Orsi and John B. Moore Department of Systems Engineering,

More information

642:550, Summer 2004, Supplement 6 The Perron-Frobenius Theorem. Summer 2004

642:550, Summer 2004, Supplement 6 The Perron-Frobenius Theorem. Summer 2004 642:550, Summer 2004, Supplement 6 The Perron-Frobenius Theorem. Summer 2004 Introduction Square matrices whose entries are all nonnegative have special properties. This was mentioned briefly in Section

More information

Lecture 7 Spectral methods

Lecture 7 Spectral methods CSE 291: Unsupervised learning Spring 2008 Lecture 7 Spectral methods 7.1 Linear algebra review 7.1.1 Eigenvalues and eigenvectors Definition 1. A d d matrix M has eigenvalue λ if there is a d-dimensional

More information

Balanced Truncation 1

Balanced Truncation 1 Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.242, Fall 2004: MODEL REDUCTION Balanced Truncation This lecture introduces balanced truncation for LTI

More information

Chapter 6: Orthogonality

Chapter 6: Orthogonality Chapter 6: Orthogonality (Last Updated: November 7, 7) These notes are derived primarily from Linear Algebra and its applications by David Lay (4ed). A few theorems have been moved around.. Inner products

More information

Honors Algebra II MATH251 Course Notes by Dr. Eyal Goren McGill University Winter 2007

Honors Algebra II MATH251 Course Notes by Dr. Eyal Goren McGill University Winter 2007 Honors Algebra II MATH251 Course Notes by Dr Eyal Goren McGill University Winter 2007 Last updated: April 4, 2014 c All rights reserved to the author, Eyal Goren, Department of Mathematics and Statistics,

More information

Lecture 8: Linear Algebra Background

Lecture 8: Linear Algebra Background CSE 521: Design and Analysis of Algorithms I Winter 2017 Lecture 8: Linear Algebra Background Lecturer: Shayan Oveis Gharan 2/1/2017 Scribe: Swati Padmanabhan Disclaimer: These notes have not been subjected

More information

2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian

2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian FE661 - Statistical Methods for Financial Engineering 2. Linear algebra Jitkomut Songsiri matrices and vectors linear equations range and nullspace of matrices function of vectors, gradient and Hessian

More information

5.) For each of the given sets of vectors, determine whether or not the set spans R 3. Give reasons for your answers.

5.) For each of the given sets of vectors, determine whether or not the set spans R 3. Give reasons for your answers. Linear Algebra - Test File - Spring Test # For problems - consider the following system of equations. x + y - z = x + y + 4z = x + y + 6z =.) Solve the system without using your calculator..) Find the

More information

Knowledge Discovery and Data Mining 1 (VO) ( )

Knowledge Discovery and Data Mining 1 (VO) ( ) Knowledge Discovery and Data Mining 1 (VO) (707.003) Review of Linear Algebra Denis Helic KTI, TU Graz Oct 9, 2014 Denis Helic (KTI, TU Graz) KDDM1 Oct 9, 2014 1 / 74 Big picture: KDDM Probability Theory

More information

HOMEWORK PROBLEMS FROM STRANG S LINEAR ALGEBRA AND ITS APPLICATIONS (4TH EDITION)

HOMEWORK PROBLEMS FROM STRANG S LINEAR ALGEBRA AND ITS APPLICATIONS (4TH EDITION) HOMEWORK PROBLEMS FROM STRANG S LINEAR ALGEBRA AND ITS APPLICATIONS (4TH EDITION) PROFESSOR STEVEN MILLER: BROWN UNIVERSITY: SPRING 2007 1. CHAPTER 1: MATRICES AND GAUSSIAN ELIMINATION Page 9, # 3: Describe

More information

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination Math 0, Winter 07 Final Exam Review Chapter. Matrices and Gaussian Elimination { x + x =,. Different forms of a system of linear equations. Example: The x + 4x = 4. [ ] [ ] [ ] vector form (or the column

More information

Problem Set (T) If A is an m n matrix, B is an n p matrix and D is a p s matrix, then show

Problem Set (T) If A is an m n matrix, B is an n p matrix and D is a p s matrix, then show MTH 0: Linear Algebra Department of Mathematics and Statistics Indian Institute of Technology - Kanpur Problem Set Problems marked (T) are for discussions in Tutorial sessions (T) If A is an m n matrix,

More information

Perron-Frobenius theorem

Perron-Frobenius theorem Perron-Frobenius theorem V.S. Sunder Institute of Mathematical Sciences sunder@imsc.res.in Unity in Mathematics Lecture Chennai Math Institute December 18, 2009 Overview The aim of the talk is to describe

More information