Computing the nearest reversible Markov chain
|
|
- Ethel Dennis
- 5 years ago
- Views:
Transcription
1 Konrad-Zuse-Zentrum für Informationstechnik Berlin Takustraße 7 D Berlin-Dahlem Germany ADAM NIELSEN AND MARCUS WEBER Computing the nearest reversible Markov chain ZIB Report (December 2014)
2 Herausgegeben vom Konrad-Zuse-Zentrum für Informationstechnik Berlin Takustraße 7 D Berlin-Dahlem Telefon: Telefax: bibliothek@zib.de URL: ZIB-Report (Print) ISSN ZIB-Report (Internet) ISSN
3 Computing the nearest reversible Markov chain Adam Nielsen Marcus Weber Abstract Reversible Markov chains are the basis of many applications. However, computing transition probabilities by a finite sampling of a Markov chain can lead to truncation errors. Even if the original Markov chain is reversible, the approximated Markov chain might be non-reversible will lose important properties, like the real valued spectrum. In this paper, we show how to find the closest reversible Markov chain to a given transition matrix. It turns out that this matrix can be computed by solving a convex minimization problem. 1 Introduction A Markov chain with n N states is described through a transition matrix P R n n, i.e. P ij = 1 P ij 0 for i, j = 1,..., n. A vector π R n is called a probability distribution if π i = 1 π j 0 i=1 for j = 1,..., n holds. A Markov chain or its corresponding transition matrix P is called reversible according to a probability distribution π if the following equation DP = P T D is valid for the diagonal matrix D = diag(π 1,..., π n ). stationary distribution of the Markov chain, i.e. π T P = π T. In this case, π is a The main result of this article is that for any transition Matrix P, any probability distribution π, any norm on R n n which is induced by a scalar product, there exists a unique transition matrix P which is reversible according to π has minimal distance to P with respect to the norm. This article is structured as follows. First, we will prove the above conjecture. Then, we show how to numerically obtain the reversible matrix P, discuss the computational cost include a perturbation analysis. Finally, we will give an application a numerical example. nielsen@zib.de weber@zib.de 1
4 2 Existence Proof Given a stochastic matrix P a stationary distribution π, it will be proven in the following that there exists a unique transition matrix P which minimizes P P is reversible according to π. First, let us consider U = {A R n n DA = A T D k R with A ij = k for all i = 1,..., n} where D = diag(π 1,..., π n ) denotes the diagonal matrix with values π i on the diagonal. Notice that U is a subspace of R n n because for A, B U with n a 1j = k 1 n i=1 b 1j = k 2 we obtain for any α, β R that D(αA + βb) = αda + βdb = αa T D + βb T D = (αa + βb) T D αa ij + βb ij = αk 1 + βk 2 for all i = 1,..., n. We can show the following property: Lemma 2.1. For A U with n A ij = k it holds π T A = kπ T. Proof. Since A U, we have π i A ij = π j A ji. Therefore, we obtain (π T A) i = π l A li = π i l=1 n l=1 A il = π i k. For an understing of the space U, the following matrices are of interest: π s if i = r j = s, π r if i = s j = r, A [r,s] 1 π s if i = j = r, ij = 1 π r if i = j = s, 1 if i = j s i r, 0 else, δ [r,s] ij = 1 if i = j i r, 1 if i = r j = s, 0 else. Proposition 2.1. The set {A [r,s] (r, s) I A } {δ [r,s], δ [s,r] (r, s) I B } {Id} 2
5 is a basis of U where Id is the identity matrix I = {(r, s) 1 r < s n}, A = {i: π i 0}, B = {i: π i = 0}, The dimension of U is given by I A = {(r, s) I r A or s A}, I B = {(r, s) I r, s B}. dim(u) = ( ) n ( ) B. 2 Proof. One may notice that the row sum of Id, A [r,s] δ [r,s] is always one that ( DA [r,s] = A [r,s]) T D holds. The latter can be seen by ( DA [r,s]) ij = k=1 = π i A [r,s] ij ( ) D ik A [r,s] kj = π j A [r,s] ji = A [r,s] ki D kj = k=1 ( ) (A [r,s] ) T D where ( ) holds because for i = r j = s or vice versa we have π i A [r,s] ij ij = π r π s = π s π r = π j A [r,s] ji, for i = j the equation in ( ) is trivial in all other cases we have A [r,s] A [r,s] ji = 0, therefore A [r,s] U. Also, we have Dδ [r,s] = δ [r,s]t D for all (r, s) I B. This is because π r = 0, therefore, { Dδ [r,s] π i if i = j, i,j = 0 else., ij = Since Dδ [r,s] is just a diagonal matrix, it is in particular symmetric we obtain Dδ [r,s] = (Dδ [r,s] ) T = δ [r,s]t D, the argument is analogous for δ [s,r]. To prove that these matrices are indeed a basis of U, it remains to show that they are linearly independent that they span the subspace U. 3
6 To see the linearly independence, consider now an arbitrary linear combination of zero: α r,s A [r,s] + α r,s δ [r,s] + β r,s δ [s,r] + αi = 0. (r,s) I A (r,s) I B For (r, s) I A the matrix A [r,s] is the only matrix in the above linear combination that could have a non-zero entry in row r column s in row s column r. Therefore, we obtain 0 = α r,s π s 0 = α r,s π r. Thus, π s 0 or π r 0 provides α r,s = 0. For (r, s) I B, we obtain analogously α r,s = 0 β r,s = 0. The linear combination reduces to α Id = 0 which, finally, leads us to α = 0. It remains to show that the given matrices span the subspace U. Thus, let us consider a matrix C U with n C ij = k for some k R. For (r, s) I A define α r,s := Csr π r if π r 0 otherwise α r,s := Crs π s. From π r C rs = π s C sr we get that For (r, s) I B choose α r,s A [r,s] rs = C rs α r,s A [r,s] sr = C sr. α r,s = C r,s β r,s = C s,r. Since each off-diagonal element appears in exactly one matrix, C differs from C := α r,s A [r,s] + α r,s δ [r,s] + β r,s δ [s,r] (r,s) I A (r,s) I B only in the diagonal. We also know that there exists l R with n C ij = l for i = 1,..., n since C U. Therefore, the matrix Ĉ := C + (k l) Id has row-sum k Ĉ ii = k Ĉ ij = k C ij = C ii, holds. Therefore,,j i C = Ĉ U,,j i which shows that the given matrices span U. Furthermore, dim U = I A + I B + I B + 1 = I + I B + 1. The statement follows from I = I B = ( ) n 2 ( ) B. 2 4
7 Our goal is to find a matrix P X = {A U A ij 0 for i, j = 1,..., n a 1j = 1} such that: P P A P for all A X. (1) To do so, let us characterize the set X. To simplify notation, let us denote the basis of Proposition 2.1 by (v i ) i=1,...,m with m = dim U v m = Id. For any A X we have a unique coefficent vector x R m with A = m i=1 x iv i. From n a ij = 1 we get that 1 = a ij = = = ( m ) x l v l (i, j) m l=1 x l m x l. l=1 l=1 v l (i, j) This can be rewritten in 1 T x = 1 where 1 R m is the constant vector 1 i = 1 for i = 1,..., m. Further, we have a ij 0 for all i j if only if x l 0 for all l = 1,..., m 1. This can be rewritten as xe i 0 for i = 1,..., m 1. The diagonal entries of A can be positive even if x m is negative. Let us quickly go back to the old notation to see how we have to define the restriction here. So let A be given by A = α r,s A [r,s] + α r,s δ [r,s] + β r,s δ [s,r] + αid, (r,s) I A (r,s) I B thus, a ii = α r,s A [r,s] ii + α r,s δ [r,s] ii (r,s) I A (r,s) I B Which leads to a ii = α r,s (1 π s )+ α r,s (1 π r )+ + β r,s δ [s,r] ii + αi ii. α r,s + α r,s + β r,s +α. (r,s) I A r=i (r,s) I A s=i (r,s) I A r i s (r,s) I B r i (r,s) I B s i The condition a ii 0 is equivalent to xg i 0 where 1 π s if v j = A [i,s] for some s > i, 1 π r if v j = A [r,i] for some r < i, g i (j) = 0 if v j = δ [i,s] for some s, 1 else, 5
8 for i = 1,..., n j = 1,..., m. Given the matrix e T 1. C = 1 e T m 1 g T R (n+m 1) m, 1. the condition that A = m i=1 x iv i is in set X is equivalent to g T n Cx 0 1 T x = 1. In order to find a matrix P X which satisfies inequality (1), let us recall that the norm is induced by a scalar product,, i.e. A = A, A for any matrix A R n n. We want to minimize the term with where m x i v i P i=1 2 = m m x i x j v i, v j 2 x i v i, P + P, P i, i=1 = 1 2 xt Qx + x T f + c (2) Cx 0 1 T x = 1, Q(i, j) := 2 v i, v j, f(i) = 2 v i, P c = P, P. Since Q is a Gram matrix of linear independent vectors, it is positive definite. This follows because for any x R n it holds x T Qx = n x i Q ij x j = x i v i, x i v i. i,j i=1 i=1 Since v 1,..., v n is a basis, we have that n i=1 x iv i 0 for x 0, in consequence, x T Qx > 0 for x 0. Since Q is positive definite, the quadratic function (2) is strongly convex. Therefore, we have formulated the problem into a strongly convex quadratic programming problem that attains its global minimum, since the quadratic function is coercive, continuous the set X is non-empty because of Id X, see [9] Theorem Also, the global minimum is unique because the quadratic function is strongly convex. Thus, we have proven the existence of a unique matrix P X which fulfills inequality (1). We will now discuss how to compute this matrix. 6
9 3 Complexity Perturbation To avoid technical difficulties, we will assume in this chapter that π i > 0 for i = 1,..., n. For A R n n the trace tr(a) is given by the sum of the diagonal elements tr(a) = A ii. It is known that i=1 A, A F := tr(a T B) is a scalar product on R n n for A, B R n n. This scalar product induces the Frobenius norm A F = A, A F = a ij 2. i=1 The following complexity analysis will be given according to the Frobenius norm. 3.1 Complexity Unfortunately, the matrix Q is quite large, in particular we have Q R m m where m = dim(u) = 1 + n2 2. Each entry of Q is given by a trace of two sparse matrices which can be computed by the following formula n π r π r π s π s if r, r, s, s are distinct, n 1 2π s + (1 π r )(1 π r ) if r r, s = s, A [r,s], A [r,s ] = n 1 2π r + (1 π s )(1 π s ) if r = r, s s, F n 1 + (1 π s )(1 π s ) + (1 π r )(1 π r ) if r = s or r = s, n 2 + (1 π r ) 2 + πr 2 + (1 π s ) 2 + πs 2 if r = r s = s. Note that because of r < s r < s other cases are not possible. We only give a proof for the case r r s = s, all other cases can be obtained analogously. First, if A [r,s] ij 0 for i j it follows that either i = r, j = s or i = s, j = r holds. Since r r, s = s we have A [r,s ] ij = 0, therefore, A [r,s], A [r,s ] = F i=1 A [r,s] ii A [r,s ] ii = n 3 + A [r,s] rr A [r,s ] rr + A [r,s],s ] r r A[r r r + A[r,s] ss A [r,s ] ss = n π s + 1 π s + (1 π r )(1 π r ) = n 1 2π s + (1 π r )(1 π r ). Therefore, the computational cost to evaluate a single entry of Q does not increase with n the effort to compute store Q is O(n 4 ). The convex minimization problem can be solved with a barrier method, e.g. the interior point method. The method consist of N Newton iterations.. 7
10 The number N of Newton iterations to find a strictly feasible point is bounded by N ( n + n 2 ) n + n 2 /2 1 /2 1 log γ ε where ε > 0 is the demed accuracy γ is a constant depending on the choice of backtracking parameter, see [1], section Therefore, the amount of Newton iterations is bounded by O(n log n). Unfortunately, each Newton iteration has to solve a linear equation system cost O((n 2 ) 3 ) since Q R ( n2 n2 2 +1) ( 2 +1). Therefore, the total cost for the optimization problem is bounded by O(n 7 log n). In the field of convex optimization, the upper bound for the amount of Newton iterations is known to be a large overestimation [1]. If we assume this amount to be independent from n, then we can expect that the time to find a solution with the convex optimization problem should be given by g(n) = αn 6 where n is the size of the matrix P R n n. If we include the bad estimation for the Newton iteration, then the time should be represented by f(n) = βn 7 log(n). In order to explore the computation time of the convex optimization problem, we generated for each n = 10,..., 100 a stochastic matrix P R n n a stochastic vector π R n where each entry was drawn from the stard uniform distribution on the open interval (0,1) then we normalized P π. We used Matlab R2012b on a 3 gigahertz computer with 8 gigabyte memory. We solved the convex optimization problem with the interior-point-convex algorithm from the provided Matlab method quadprog with default options, i.e. relative dual feasibility =2.31e 15 with TolFun=1e 0.8, complementarity measure =1.68e 10 with TolFun=1e 0.8 relative max constraint violoation = 0 with TolCon=1e 0.8. For n = 45 the execution time was about seconds. The scalars α, β where chosen such that f(45) = g(45) = holds. In Figure 1 one can see that g seems to be a reasonable approximation of the execution time. Also one can see that the execution time takes only seconds for matrices in R 50 50, but already 32 minutes for matrices in R Perturbation analysis For we obtain k := min r,r,s,s A [r,s], A [r,s ] F k Id Q n Id where the inequality has to be read componentwise. The upper bound follows from v i, v j F v i, v i F v j, v j F max { v l, v l F } l=1,...,m the fact that A [r,s], A [r,s] = n 2 + (1 π r ) 2 + πr 2 + (1 π s ) 2 + πs 2 n F 8
11 70 Minutes Time of minmization algorithm g f n Figure 1: Duration of convex minimization problem to find closest reversible Markov chain. Id, Id F = n holds. Since Q is symmetric, its condition number according to the spectral norm is given by κ(q) = λ max λ min n k. Since k = n c for some 0 < c < 4 we have that κ(q) 1 1 c/n 1 for n. Therefore, Q is well conditioned. We assume now that we are interested in finding a reversible matrix ˆP X, but that we only have a perturbed version P = ˆP + E of ˆP which is not necessarily reversible. With the above scheme, we can find a reversible matrix P X which is closest to P according to the Frobenius norm. The question arises, how eigenvalues change between P ˆP. In order to answer this question, the weighted Frobenius norm is introduced with A F := D 1 2 AD 1 2 F D = diag(π 1,..., π n ), D 1 2 D 1 2 = D D 1 2 := (D 1 2 ) 1. Note that the weighted Frobenius norm is given by A 2 F = i, a 2 π i i,j We can then give the following estimation about the eigenvalues: π j. 9
12 Theorem 3.1. Let A, B X, let λ 1,..., λ n be the eigenvalues of A ˆλ 1,..., ˆλ n be the eigenvalues of B. Then there exists a permutation σ of the integers 1, 2,..., n such that ˆλ σ(i) λ i 2 A B 2 F. i=1 Proof. The matrices A, B X are self-adjoint according to the scalar product x, y π := x T Dy. Lets denote with {w 1,..., w n } a, π -orthonormal basis, denote with W the matrix whose columns contain the vectors w i, i.e. W = w 1 w 2... w n. Then A := W 1 AW B = W 1 BW are symmetric, see [3] Chapter By the Hoffman Wielt Theorem (see Theorem in [5]) we obtain ˆλ σ(i) λ i 2 A B 2 F i=1 for a permutation σ, because similar matrices have the same eigenvalues. remains to show It or equivalently W 1 CW 2 F = C 2 F C 2 F = W CW 1 2 F for any matrix C R n. By construction of W we have Therefore, W T DW = I W 1 D 1 (W T ) 1 = I. (3) W CW 1 2 F = D 1 2 W CW 1 D F = tr(d 1 2 (W 1 ) T C T W T D 1 2 D 1 2 W CW 1 D 1 2 ) ( ) = tr(w 1 D 1 (W 1 ) T C T W T DW C) (3) = tr(c T C) = C 2 F, where in ( ) it is used that the trace is invariant under cyclic permutations. This shows that in order to guarantee good approximations for the eigenvalues, one has to assure a good approximation of a ij for those i, j where π j << π i. These transitions are also known as rare events often difficult to compute. We will end this chapter with a short experiment motivated by Theorem 3.1. We create 100 reversible Markov chains ( ˆP i ) i=1,...,100 R 5 5 as follows. First, 10
13 we generate a symmetric matrix A where each entry A ij with i j is chosen equally distributed from [0, 1] A ij := A ji for i > j. After normalizing A the resulting matrix is reversible [7] will be denoted as ˆP i. We perturb normalize ˆP i to obtain a perturbed version P i of ˆP i. For each P i we compute the unique reversible matrix P i which is closest to P i according to weighted Frobenius norm according to the stationary distribution of ˆP i. The matrix P i exists can be computed by the introduced convex optimization problem, because the weighted Frobenius norm is induced by the scalar product A, B F = tr(dad 1 B T ). Also we compute the unique reversible matrix ˇP i which is closest to P i according to stard Frobenius norm according to the stationary distribution of ˆPi. For each tuple ( ˆP, P, P, ˇP ) we compute the corresponding eigenvalues (ˆλ j, λ j, λ j, ˇλ j ),...,5. Then, we compute the numbers c 1 = min 5 ˆλ j λ σ(j) 2, c 2 = min 5 ˆλ j λ σ(j) 2 σ Π 5 σ Π 5 c 3 = min 5 ˆλ j ˇλ σ(j) 2, σ Π 5 where Π 5 denotes the set of all permutations of the numbers {1,..., 5}. The experiment is visualized in Figure 2. The result is that the closest reversible matrix ˇP according the Frobenius norm does not maintain the spectrum of ˆP. However, the closest reversible matrix according the weighted Frobenius norm P gives a good approximation for the spectrum of ˆP improves always the spectrum of the perturbed version P Weighted correction c1 Perturbation c Unweighted correction c3 Pertubation c ˆP P F ˆP ˇP F Figure 2: Comparison of c 1 c 2. Figure 3: Comparison of c 2 c 3. Three remarks may be in order: The resulting matrix P = α r,s A [r,s] + α Id (r,s) I 11
14 is reversible up to machine precision. In fact, it is reversible if π r α r,s π s = π s α r,s π r holds 1. A numerical error in the solution (x) i=1,...,m 1 = (α r,s ) (r,s) I of the convex optimization problem results in an error of the entry P (r, s) = α r,s π s P (s, r) = α r,s π r for r s. In order to guarantee that P has row sum one, it might be advantages to set p ii = 1 p ij. j i 4 Application numerical example In the past decades, analysis of a certain class of stochastic processes has been formulated in terms of approximating a transfer operator T : L 2 (µ) L 2 (µ) [12, 8, 11] which is self-adjoint according to the scalar product f, g µ = f(x) g(x) µ(dx), i.e. E T f, g µ = f, T g µ for all f, g L 2 (µ), where (E, Σ, µ) is a measure space for some set E R n, L 2 (µ) = {f : E R f, f µ <, fµ-measurable}/n with N = {f : E R A Σ with µ(a) = 0 f(x) = 0 for all x E\A}. Furthermore, the operator has the property that for f L 2 (µ) with f 0 it follows T f 0 almost surely. Let us assume that we have a set of non-negative functions {φ 1,..., φ n } given, such that n i=1 φ i(x) = 1 for all x E φ i, 1 µ > 0 for i = 1,..., n where 1(x) := 1 for all x E. If we define π i := φ i, 1 µ T ij := T φ i, φ j µ φ i, 1 µ, then it is straightforward to verify the properties (i) πt = π, (ii) DT = T T D, (iii) n T ij = 1 for all i = 1,..., n, (iv) T ij 0, 1 One may note that I A = I because of the assumption at the beginning of this chapter. 12
15 where D = diag(π 1,..., π n ) denotes the diagonal matrix of π. In other words, T is a reversible Markov chain. This matrix plays an essential role in the Galerkin discretization of T. Theorem 4.1. Let {φ 1,..., φ n } L 2 (µ) be a basis with φ i, 1 µ > 0 of a subspace D, Q: L 2 (µ) D the orthogonal projection onto D. For any self-adjoint continuous operator T : L 2 (µ) L 2 (µ) we have M = S 1 T, T ij = T φ i, φ j µ φ i, 1 µ, S ij = φ i, φ j µ φ i, 1 µ is a right matrix representation of QT Q according to the basis A = {φ 1,..., φ n }, i.e. for any f = α i φ i, QT Qf = β i φ i it holds i=1 i=1 M(α 1,..., α n ) T = (β 1,..., β n ) T. Proof. Consider the Gram matrix of {φ 1,..., φ n } Ŝ ij = φ i, φ j µ. This matrix is invertible since {φ 1,..., φ n } is a basis the orthogonal projection Q can be represented as Qv = Ŝ 1 ij v, φ i µ φ j. i, This can be verified by checking Qv v, g µ = 0 for all g D, v L 2 (µ). From ) S = D 1 Ŝ with D = diag ( φ 1, 1 µ,..., φ n, 1 µ we obtain S 1 = Ŝ 1 D, therefore, Ŝ 1 ij = S 1 ij 1 = S 1 1 ji, φ j, 1 µ φ i, 1 µ in the last step it was used that Ŝ 1 is symmetric since Ŝ is symmetric. This implies Qv = S 1 φ j ji v, φ i µ. φ i, 1 µ Therefore, QT Qφ k = QT φ k = = = i, i, i, S 1 φ j ij T φ k, φ i µ φ i, 1 µ S 1 T φ i, φ k µ ij φ j = φ i, 1 µ ( S 1 T ) kj φ j. ( n i=1 S 1 ji T ik ) φ j 13
16 For the more general case that the operator is not self-adjoint, one can find a similar Galerkin approximation, see [10], Theorem 3. Note that S is a special case of the matrix T with T = I, where I denotes the identity operator. Therefore, we can use our machinery to correct T S if π is given. One possible choice for the functions φ i are set based functions, i.e. we have disjoint sets A 1,..., A n with n i=1 A i = E { 1 if x A i φ i (x) = 1 Ai (x) = 0 else. In this case, the matrix S from Theorem 4.1 turns out to be the identity matrix we are only left with the task of computing T. Other possible choices are radial basis functions or commitor functions which can be found in [13, 10] So far, we know how to maintain reversibility for S T separately. The question arises, if S 1 T also inherits some properties that might get lost due to numerical errors. First, we can rewrite S = D 1 Ŝ, T = D 1 ˆT with Ŝij = φ i, φ j µ ˆT ij = T φ i, φ j µ, hence S 1 T = Ŝ 1 DD 1 ˆT = Ŝ 1 ˆT. Analogously as shown for Q also Ŝ Ŝ 1 are symmetric positive definite matrices. Now, since Ŝ 1 is positive definite, we know the existence of a symmetric square matrix A such that A 2 = Ŝ 1. Thus, A 1 Ŝ 1 ˆT A = A ˆT A. Consequently, Ŝ 1 ˆT is similar to a symmetric matrix hence diagonalizable. This shows that the spectrum of Ŝ 1 ˆT is real that we know the existence of a basis of eigenvectors of S 1 T. For numerical estimation, the following procedure may be advantageous: First, approximate T S by the convex minimization problem stated above. Then, calculate DS which is symmetric up to machine precision. For the symmetric matrix (DS ) 1 one can solve a convex minimization problem that finds the closest symmetric positive semi definite matrix ((DS ) 1 ) according to the Frobenius norm, see [4]. Finally, ((DS ) 1 ) T is the corrected matrix. Also it would be possible to correct S 1 with our explained optimization scheme, because this matrix has also row sum one is reversible, this follows from S1 = 1 1 = S 1 1 DS = S T D (S T ) 1 D = DS 1 (S 1 ) T D = DS 1. Therefore, by dropping the constraints given by the matrix C, one can also approximate S 1 as a convex quadratic programing problem with only linear constrains. 14
17 A Galerkin approximation of the transfer operator can be used to apply spectral clustering algorithms such as PCCA+ [2] which relies on the real spectrum of T S 1. Especially for a set-based reduction, we can guarantee with our machinery that T will have a real spectrum,, therefore, spectral clustering such as PCCA+ is applicable. 4.1 Numerical example Consider the one dimensional, 2π-periodic function V B : R R V B (x) = a + b cos(x) + c cos 2 (x) + d cos 3 (x) with a=2.0567, b= , c= d= This can be seen as an approximation of a butane potential energy function, where x is the central dihedral angle. We are going to realize a trajectory of the dihedral angles X t = X t mod 2π of butane from the stochastic differential equation d X t = V B ( X t )dt + σdb t. If we divide [0, 2π] in the sets A i = [ i 1 30, i 30 ] for i = 1,..., 30, then the Galerkin discretization of T reduces to T ij = P(X τ A j X 0 A i ), see [8]. To gain the associated Markov State Model T R 30 30, we compute a longterm trajectory (X i ) i=0,...,n 1 by performing n timesteps of size dt using the Euler-Maruyama discretization X i+1 = X i V B ( X i )dt + σ dt η i, where η i are i.i.d rom variables distributed according to the stard normal distribution. This trajectory is chopped into pieces of length 400 yielding M subtrajectories (X k i ) i=1,...,l := (X lk,..., X l(k+1) 1 ) for k = 0,..., M 1. We can estimate T by counting transitions C ij = M 1 k=0 1 Ai (X k 1 ) 1 Aj (X k l ) C ij T ij = 30 i=1 C. ij The Markov State Model created in this way becomes only reasonable when considered for a trajectory longer then 10 6 timesteps, due to rare transition events. We will now compare the eigenvalues of the Markov State Model T to the eigenvalues of the above introduced corrected estimation T with respect to the weighted Frobenius norm F. We will also compare them for different length of a given trajectory, starting from length n = 10 6 until n = We used the timestep dt = Since the eigenvalues of T turned out to be complex sometimes, we simply set the imaginary part to zero, in order to compare the eigenvalues with T. 15
18 Corrected estimation Stard estimation nd eigenvalue Amount of trajectories (x10 7 ) Discussion In the figures the stard estimation the corrected estimation do not essentially differ for the eigenvalues close to 1. Our algorithm of finding the nearest reversible Markov chain especially preserves the dominant eigenvalues of the underlying operator, thus, is suitable for cluster analysis 2 of Molecular simulation data. In contrast to T, the resulting transition matrix T from our algorithm is guaranteed to be applicable to spectral clustering methods, such as PCCA+, because of its real spectrum. Acknowledgment We gratefully thank the Berlin Mathematical School for financial support of Adam Nielsen. 2 For the connection between cluster analysis eigenvalues we refer to [6]. 16
19 0.85 Corrected estimation Stard estimation 0.8 3rd eigenvalue Amount of trajectories (x10 7 ) Corrected estimation Stard estimation 0.3 4th eigenvalue Amount of trajectories (x10 7 ) 17
20 Corrected estimation Stard estimation 0 5th eigenvalue Amount of trajectories (x10 7 ) Corrected estimation Stard estimation 0 6th eigenvalue Amount of trajectories (x10 7 ) 18
21 References [1] S. Boyd L. Venberghe, Convex optimization, Cambridge University Press, [2] Peter Deuflhard M. Weber, Robust Perron cluster analysis in conformation dynamics, In M. Dellnitz, S. Kirkl, M. Neumann, C. Schütte, editors, Linear Algebra Its Applications - Special Issue on Matrices Mathematical Biology,volume 398C (2005), [3] G. Fischer, Lineare algebra, 17 ed., Vieweg + Teubner, [4] Nicholas J. Higham, Computing a nearest symmetric positive semidefinite matrix, Linear Algebra its Applications 103 (1988), ii iii. [5] R. Horn C. Johnson, Matrix analysis, Cambridge University Press, [6] Wilhelm Huisinga Bernd Schmidt, Metastability dominan eigenvaues of transfer operators, Lecture Notes in Computational Science Engineering 49 (2006), [7] Ph. Metzner, F. Noé, Ch. Schütte, Estimating the sampling error: Distribution of transition matrices functions of transition matrices for given trajectory data, Physical Review E, 80 (2) [8] A. Nielsen, K.Fackeldey, M.Weber, Transfer operator for arbitary measures, in submission. [9] N. V. Thoai R. Horst, P. M. Pardalos, Introduction to global optimization, Kluwer Academic Publishers, [10] M. Sarich, Projected transfer operators - discretization of Markov processes in high-dimensional state space., Ph.D. thesis, Freie Universität Berlin, [11] C. Schütte, Conformational dynamics: Modelling, theory, algorithm, application to biomolecules., habilitation, Freie Universität Berlin, [12] Christof Schütte, Wilhelm Huisinga, Peter Deuflhard, Transfer operator approach to conformational dynamics in biomolecular systems, Applied Mathematics Entering the 21th Century (2003). [13] Marcus Weber, A subspace approach to molecular Markov state models via a new infinitesimal generator., habilitation, Freie Universität Berlin,
The Monte Carlo Computation Error of Transition Probabilities
Zuse Institute Berlin Takustr. 7 495 Berlin Germany ADAM NILSN The Monte Carlo Computation rror of Transition Probabilities ZIB Report 6-37 (July 206) Zuse Institute Berlin Takustr. 7 D-495 Berlin Telefon:
More informationDS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.
DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1
More informationReview problems for MA 54, Fall 2004.
Review problems for MA 54, Fall 2004. Below are the review problems for the final. They are mostly homework problems, or very similar. If you are comfortable doing these problems, you should be fine on
More informationBoolean Inner-Product Spaces and Boolean Matrices
Boolean Inner-Product Spaces and Boolean Matrices Stan Gudder Department of Mathematics, University of Denver, Denver CO 80208 Frédéric Latrémolière Department of Mathematics, University of Denver, Denver
More informationThroughout these notes we assume V, W are finite dimensional inner product spaces over C.
Math 342 - Linear Algebra II Notes Throughout these notes we assume V, W are finite dimensional inner product spaces over C 1 Upper Triangular Representation Proposition: Let T L(V ) There exists an orthonormal
More informationMATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)
MATH 20F: LINEAR ALGEBRA LECTURE B00 (T KEMP) Definition 01 If T (x) = Ax is a linear transformation from R n to R m then Nul (T ) = {x R n : T (x) = 0} = Nul (A) Ran (T ) = {Ax R m : x R n } = {b R m
More informationMath 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.
Math 350 Fall 2011 Notes about inner product spaces In this notes we state and prove some important properties of inner product spaces. First, recall the dot product on R n : if x, y R n, say x = (x 1,...,
More informationLinear Algebra Massoud Malek
CSUEB Linear Algebra Massoud Malek Inner Product and Normed Space In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n An inner product
More informationMathematical Optimisation, Chpt 2: Linear Equations and inequalities
Mathematical Optimisation, Chpt 2: Linear Equations and inequalities Peter J.C. Dickinson p.j.c.dickinson@utwente.nl http://dickinson.website version: 12/02/18 Monday 5th February 2018 Peter J.C. Dickinson
More informationSpectral Theorem for Self-adjoint Linear Operators
Notes for the undergraduate lecture by David Adams. (These are the notes I would write if I was teaching a course on this topic. I have included more material than I will cover in the 45 minute lecture;
More informationREVIEW FOR EXAM III SIMILARITY AND DIAGONALIZATION
REVIEW FOR EXAM III The exam covers sections 4.4, the portions of 4. on systems of differential equations and on Markov chains, and..4. SIMILARITY AND DIAGONALIZATION. Two matrices A and B are similar
More informationThe University of Texas at Austin Department of Electrical and Computer Engineering. EE381V: Large Scale Learning Spring 2013.
The University of Texas at Austin Department of Electrical and Computer Engineering EE381V: Large Scale Learning Spring 2013 Assignment Two Caramanis/Sanghavi Due: Tuesday, Feb. 19, 2013. Computational
More informationAPPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.
APPENDIX A Background Mathematics A. Linear Algebra A.. Vector algebra Let x denote the n-dimensional column vector with components 0 x x 2 B C @. A x n Definition 6 (scalar product). The scalar product
More informationMath 443 Differential Geometry Spring Handout 3: Bilinear and Quadratic Forms This handout should be read just before Chapter 4 of the textbook.
Math 443 Differential Geometry Spring 2013 Handout 3: Bilinear and Quadratic Forms This handout should be read just before Chapter 4 of the textbook. Endomorphisms of a Vector Space This handout discusses
More informationClustering by using a simplex structure
Konrad-Zuse-Zentrum für Informationstechnik Berlin Takustraße 7 D-14195 Berlin-Dahlem Germany MARCUS WEBER Clustering by using a simplex structure supported by DFG Research Center Mathematics for Key Technologies
More informationPreliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012
Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.
More informationLecture 1. 1 Conic programming. MA 796S: Convex Optimization and Interior Point Methods October 8, Consider the conic program. min.
MA 796S: Convex Optimization and Interior Point Methods October 8, 2007 Lecture 1 Lecturer: Kartik Sivaramakrishnan Scribe: Kartik Sivaramakrishnan 1 Conic programming Consider the conic program min s.t.
More information14 Singular Value Decomposition
14 Singular Value Decomposition For any high-dimensional data analysis, one s first thought should often be: can I use an SVD? The singular value decomposition is an invaluable analysis tool for dealing
More informationIr O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )
Section 3.2 Theorem 3.6. Let A be an m n matrix of rank r. Then r m, r n, and, by means of a finite number of elementary row and column operations, A can be transformed into the matrix ( ) Ir O D = 1 O
More informationCS 246 Review of Linear Algebra 01/17/19
1 Linear algebra In this section we will discuss vectors and matrices. We denote the (i, j)th entry of a matrix A as A ij, and the ith entry of a vector as v i. 1.1 Vectors and vector operations A vector
More informationLINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS
LINEAR ALGEBRA, -I PARTIAL EXAM SOLUTIONS TO PRACTICE PROBLEMS Problem (a) For each of the two matrices below, (i) determine whether it is diagonalizable, (ii) determine whether it is orthogonally diagonalizable,
More informationJim Lambers MAT 610 Summer Session Lecture 2 Notes
Jim Lambers MAT 610 Summer Session 2009-10 Lecture 2 Notes These notes correspond to Sections 2.2-2.4 in the text. Vector Norms Given vectors x and y of length one, which are simply scalars x and y, the
More informationLinear Algebra 2 Spectral Notes
Linear Algebra 2 Spectral Notes In what follows, V is an inner product vector space over F, where F = R or C. We will use results seen so far; in particular that every linear operator T L(V ) has a complex
More informationChapter 7: Symmetric Matrices and Quadratic Forms
Chapter 7: Symmetric Matrices and Quadratic Forms (Last Updated: December, 06) These notes are derived primarily from Linear Algebra and its applications by David Lay (4ed). A few theorems have been moved
More informationQuadratic forms. Here. Thus symmetric matrices are diagonalizable, and the diagonalization can be performed by means of an orthogonal matrix.
Quadratic forms 1. Symmetric matrices An n n matrix (a ij ) n ij=1 with entries on R is called symmetric if A T, that is, if a ij = a ji for all 1 i, j n. We denote by S n (R) the set of all n n symmetric
More information7. Symmetric Matrices and Quadratic Forms
Linear Algebra 7. Symmetric Matrices and Quadratic Forms CSIE NCU 1 7. Symmetric Matrices and Quadratic Forms 7.1 Diagonalization of symmetric matrices 2 7.2 Quadratic forms.. 9 7.4 The singular value
More informationMATH 1120 (LINEAR ALGEBRA 1), FINAL EXAM FALL 2011 SOLUTIONS TO PRACTICE VERSION
MATH (LINEAR ALGEBRA ) FINAL EXAM FALL SOLUTIONS TO PRACTICE VERSION Problem (a) For each matrix below (i) find a basis for its column space (ii) find a basis for its row space (iii) determine whether
More informationIMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET
IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET This is a (not quite comprehensive) list of definitions and theorems given in Math 1553. Pay particular attention to the ones in red. Study Tip For each
More informationEE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 2
EE/ACM 150 - Applications of Convex Optimization in Signal Processing and Communications Lecture 2 Andre Tkacenko Signal Processing Research Group Jet Propulsion Laboratory April 5, 2012 Andre Tkacenko
More informationLecture 7: Positive Semidefinite Matrices
Lecture 7: Positive Semidefinite Matrices Rajat Mittal IIT Kanpur The main aim of this lecture note is to prepare your background for semidefinite programming. We have already seen some linear algebra.
More informationAn indicator for the number of clusters using a linear map to simplex structure
An indicator for the number of clusters using a linear map to simplex structure Marcus Weber, Wasinee Rungsarityotin, and Alexander Schliep Zuse Institute Berlin ZIB Takustraße 7, D-495 Berlin, Germany
More informationRecall the convention that, for us, all vectors are column vectors.
Some linear algebra Recall the convention that, for us, all vectors are column vectors. 1. Symmetric matrices Let A be a real matrix. Recall that a complex number λ is an eigenvalue of A if there exists
More information1 Last time: least-squares problems
MATH Linear algebra (Fall 07) Lecture Last time: least-squares problems Definition. If A is an m n matrix and b R m, then a least-squares solution to the linear system Ax = b is a vector x R n such that
More informationDiagonalizing Matrices
Diagonalizing Matrices Massoud Malek A A Let A = A k be an n n non-singular matrix and let B = A = [B, B,, B k,, B n ] Then A n A B = A A 0 0 A k [B, B,, B k,, B n ] = 0 0 = I n 0 A n Notice that A i B
More information22m:033 Notes: 7.1 Diagonalization of Symmetric Matrices
m:33 Notes: 7. Diagonalization of Symmetric Matrices Dennis Roseman University of Iowa Iowa City, IA http://www.math.uiowa.edu/ roseman May 3, Symmetric matrices Definition. A symmetric matrix is a matrix
More informationThe following definition is fundamental.
1. Some Basics from Linear Algebra With these notes, I will try and clarify certain topics that I only quickly mention in class. First and foremost, I will assume that you are familiar with many basic
More informationChapter 3 Transformations
Chapter 3 Transformations An Introduction to Optimization Spring, 2014 Wei-Ta Chu 1 Linear Transformations A function is called a linear transformation if 1. for every and 2. for every If we fix the bases
More informationLinear Algebra- Final Exam Review
Linear Algebra- Final Exam Review. Let A be invertible. Show that, if v, v, v 3 are linearly independent vectors, so are Av, Av, Av 3. NOTE: It should be clear from your answer that you know the definition.
More information18.S34 linear algebra problems (2007)
18.S34 linear algebra problems (2007) Useful ideas for evaluating determinants 1. Row reduction, expanding by minors, or combinations thereof; sometimes these are useful in combination with an induction
More information0.1 Rational Canonical Forms
We have already seen that it is useful and simpler to study linear systems using matrices. But matrices are themselves cumbersome, as they are stuffed with many entries, and it turns out that it s best
More informationCSC Linear Programming and Combinatorial Optimization Lecture 10: Semidefinite Programming
CSC2411 - Linear Programming and Combinatorial Optimization Lecture 10: Semidefinite Programming Notes taken by Mike Jamieson March 28, 2005 Summary: In this lecture, we introduce semidefinite programming
More informationMath 321: Linear Algebra
Math 32: Linear Algebra T. Kapitula Department of Mathematics and Statistics University of New Mexico September 8, 24 Textbook: Linear Algebra,by J. Hefferon E-mail: kapitula@math.unm.edu Prof. Kapitula,
More informationMethods for sparse analysis of high-dimensional data, II
Methods for sparse analysis of high-dimensional data, II Rachel Ward May 26, 2011 High dimensional data with low-dimensional structure 300 by 300 pixel images = 90, 000 dimensions 2 / 55 High dimensional
More informationMathematical Methods wk 2: Linear Operators
John Magorrian, magog@thphysoxacuk These are work-in-progress notes for the second-year course on mathematical methods The most up-to-date version is available from http://www-thphysphysicsoxacuk/people/johnmagorrian/mm
More informationCONCENTRATION OF THE MIXED DISCRIMINANT OF WELL-CONDITIONED MATRICES. Alexander Barvinok
CONCENTRATION OF THE MIXED DISCRIMINANT OF WELL-CONDITIONED MATRICES Alexander Barvinok Abstract. We call an n-tuple Q 1,...,Q n of positive definite n n real matrices α-conditioned for some α 1 if for
More informationMATH SOLUTIONS TO PRACTICE MIDTERM LECTURE 1, SUMMER Given vector spaces V and W, V W is the vector space given by
MATH 110 - SOLUTIONS TO PRACTICE MIDTERM LECTURE 1, SUMMER 2009 GSI: SANTIAGO CAÑEZ 1. Given vector spaces V and W, V W is the vector space given by V W = {(v, w) v V and w W }, with addition and scalar
More informationKernels of Directed Graph Laplacians. J. S. Caughman and J.J.P. Veerman
Kernels of Directed Graph Laplacians J. S. Caughman and J.J.P. Veerman Department of Mathematics and Statistics Portland State University PO Box 751, Portland, OR 97207. caughman@pdx.edu, veerman@pdx.edu
More informationBasic Elements of Linear Algebra
A Basic Review of Linear Algebra Nick West nickwest@stanfordedu September 16, 2010 Part I Basic Elements of Linear Algebra Although the subject of linear algebra is much broader than just vectors and matrices,
More information1 9/5 Matrices, vectors, and their applications
1 9/5 Matrices, vectors, and their applications Algebra: study of objects and operations on them. Linear algebra: object: matrices and vectors. operations: addition, multiplication etc. Algorithms/Geometric
More informationMath Linear Algebra II. 1. Inner Products and Norms
Math 342 - Linear Algebra II Notes 1. Inner Products and Norms One knows from a basic introduction to vectors in R n Math 254 at OSU) that the length of a vector x = x 1 x 2... x n ) T R n, denoted x,
More informationMATH 304 Linear Algebra Lecture 23: Diagonalization. Review for Test 2.
MATH 304 Linear Algebra Lecture 23: Diagonalization. Review for Test 2. Diagonalization Let L be a linear operator on a finite-dimensional vector space V. Then the following conditions are equivalent:
More informationMath Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88
Math Camp 2010 Lecture 4: Linear Algebra Xiao Yu Wang MIT Aug 2010 Xiao Yu Wang (MIT) Math Camp 2010 08/10 1 / 88 Linear Algebra Game Plan Vector Spaces Linear Transformations and Matrices Determinant
More informationSpanning and Independence Properties of Finite Frames
Chapter 1 Spanning and Independence Properties of Finite Frames Peter G. Casazza and Darrin Speegle Abstract The fundamental notion of frame theory is redundancy. It is this property which makes frames
More informationOn the projection onto a finitely generated cone
Acta Cybernetica 00 (0000) 1 15. On the projection onto a finitely generated cone Miklós Ujvári Abstract In the paper we study the properties of the projection onto a finitely generated cone. We show for
More information1. General Vector Spaces
1.1. Vector space axioms. 1. General Vector Spaces Definition 1.1. Let V be a nonempty set of objects on which the operations of addition and scalar multiplication are defined. By addition we mean a rule
More informationThen x 1,..., x n is a basis as desired. Indeed, it suffices to verify that it spans V, since n = dim(v ). We may write any v V as r
Practice final solutions. I did not include definitions which you can find in Axler or in the course notes. These solutions are on the terse side, but would be acceptable in the final. However, if you
More informationUniqueness of the Solutions of Some Completion Problems
Uniqueness of the Solutions of Some Completion Problems Chi-Kwong Li and Tom Milligan Abstract We determine the conditions for uniqueness of the solutions of several completion problems including the positive
More informationLinear Algebra Primer
Linear Algebra Primer David Doria daviddoria@gmail.com Wednesday 3 rd December, 2008 Contents Why is it called Linear Algebra? 4 2 What is a Matrix? 4 2. Input and Output.....................................
More informationMAT Linear Algebra Collection of sample exams
MAT 342 - Linear Algebra Collection of sample exams A-x. (0 pts Give the precise definition of the row echelon form. 2. ( 0 pts After performing row reductions on the augmented matrix for a certain system
More informationMATH 235: Inner Product Spaces, Assignment 7
MATH 235: Inner Product Spaces, Assignment 7 Hand in questions 3,4,5,6,9, by 9:3 am on Wednesday March 26, 28. Contents Orthogonal Basis for Inner Product Space 2 2 Inner-Product Function Space 2 3 Weighted
More informationSignal Recovery from Permuted Observations
EE381V Course Project Signal Recovery from Permuted Observations 1 Problem Shanshan Wu (sw33323) May 8th, 2015 We start with the following problem: let s R n be an unknown n-dimensional real-valued signal,
More informationReal symmetric matrices/1. 1 Eigenvalues and eigenvectors
Real symmetric matrices 1 Eigenvalues and eigenvectors We use the convention that vectors are row vectors and matrices act on the right. Let A be a square matrix with entries in a field F; suppose that
More informationforms Christopher Engström November 14, 2014 MAA704: Matrix factorization and canonical forms Matrix properties Matrix factorization Canonical forms
Christopher Engström November 14, 2014 Hermitian LU QR echelon Contents of todays lecture Some interesting / useful / important of matrices Hermitian LU QR echelon Rewriting a as a product of several matrices.
More informationSemidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 5
Semidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 5 Instructor: Farid Alizadeh Scribe: Anton Riabov 10/08/2001 1 Overview We continue studying the maximum eigenvalue SDP, and generalize
More informationLinear Algebra and Eigenproblems
Appendix A A Linear Algebra and Eigenproblems A working knowledge of linear algebra is key to understanding many of the issues raised in this work. In particular, many of the discussions of the details
More informationExercises * on Linear Algebra
Exercises * on Linear Algebra Laurenz Wiskott Institut für Neuroinformatik Ruhr-Universität Bochum, Germany, EU 4 February 7 Contents Vector spaces 4. Definition...............................................
More informationLinear Algebra Solutions 1
Math Camp 1 Do the following: Linear Algebra Solutions 1 1. Let A = and B = 3 8 5 A B = 3 5 9 A + B = 9 11 14 4 AB = 69 3 16 BA = 1 4 ( 1 3. Let v = and u = 5 uv = 13 u v = 13 v u = 13 Math Camp 1 ( 7
More informationAn Introduction to Correlation Stress Testing
An Introduction to Correlation Stress Testing Defeng Sun Department of Mathematics and Risk Management Institute National University of Singapore This is based on a joint work with GAO Yan at NUS March
More informationThe Hilbert Space of Random Variables
The Hilbert Space of Random Variables Electrical Engineering 126 (UC Berkeley) Spring 2018 1 Outline Fix a probability space and consider the set H := {X : X is a real-valued random variable with E[X 2
More informationCopositive matrices and periodic dynamical systems
Extreme copositive matrices and periodic dynamical systems Weierstrass Institute (WIAS), Berlin Optimization without borders Dedicated to Yuri Nesterovs 60th birthday February 11, 2016 and periodic dynamical
More informationResearch Note. A New Infeasible Interior-Point Algorithm with Full Nesterov-Todd Step for Semi-Definite Optimization
Iranian Journal of Operations Research Vol. 4, No. 1, 2013, pp. 88-107 Research Note A New Infeasible Interior-Point Algorithm with Full Nesterov-Todd Step for Semi-Definite Optimization B. Kheirfam We
More informationLecture 2: Linear Algebra Review
EE 227A: Convex Optimization and Applications January 19 Lecture 2: Linear Algebra Review Lecturer: Mert Pilanci Reading assignment: Appendix C of BV. Sections 2-6 of the web textbook 1 2.1 Vectors 2.1.1
More information8. Diagonalization.
8. Diagonalization 8.1. Matrix Representations of Linear Transformations Matrix of A Linear Operator with Respect to A Basis We know that every linear transformation T: R n R m has an associated standard
More informationMATHEMATICS 217 NOTES
MATHEMATICS 27 NOTES PART I THE JORDAN CANONICAL FORM The characteristic polynomial of an n n matrix A is the polynomial χ A (λ) = det(λi A), a monic polynomial of degree n; a monic polynomial in the variable
More informationMethods for sparse analysis of high-dimensional data, II
Methods for sparse analysis of high-dimensional data, II Rachel Ward May 23, 2011 High dimensional data with low-dimensional structure 300 by 300 pixel images = 90, 000 dimensions 2 / 47 High dimensional
More informationFor δa E, this motivates the definition of the Bauer-Skeel condition number ([2], [3], [14], [15])
LAA 278, pp.2-32, 998 STRUCTURED PERTURBATIONS AND SYMMETRIC MATRICES SIEGFRIED M. RUMP Abstract. For a given n by n matrix the ratio between the componentwise distance to the nearest singular matrix and
More informationCHAPTER 3 Further properties of splines and B-splines
CHAPTER 3 Further properties of splines and B-splines In Chapter 2 we established some of the most elementary properties of B-splines. In this chapter our focus is on the question What kind of functions
More informationMath 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008
Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008 Exam 2 will be held on Tuesday, April 8, 7-8pm in 117 MacMillan What will be covered The exam will cover material from the lectures
More information1 Vectors. Notes for Bindel, Spring 2017 Numerical Analysis (CS 4220)
Notes for 2017-01-30 Most of mathematics is best learned by doing. Linear algebra is no exception. You have had a previous class in which you learned the basics of linear algebra, and you will have plenty
More informationHomework 1. Yuan Yao. September 18, 2011
Homework 1 Yuan Yao September 18, 2011 1. Singular Value Decomposition: The goal of this exercise is to refresh your memory about the singular value decomposition and matrix norms. A good reference to
More informationGQE ALGEBRA PROBLEMS
GQE ALGEBRA PROBLEMS JAKOB STREIPEL Contents. Eigenthings 2. Norms, Inner Products, Orthogonality, and Such 6 3. Determinants, Inverses, and Linear (In)dependence 4. (Invariant) Subspaces 3 Throughout
More informationMath 3191 Applied Linear Algebra
Math 9 Applied Linear Algebra Lecture 9: Diagonalization Stephen Billups University of Colorado at Denver Math 9Applied Linear Algebra p./9 Section. Diagonalization The goal here is to develop a useful
More informationDetailed Proof of The PerronFrobenius Theorem
Detailed Proof of The PerronFrobenius Theorem Arseny M Shur Ural Federal University October 30, 2016 1 Introduction This famous theorem has numerous applications, but to apply it you should understand
More informationMathematical Methods wk 1: Vectors
Mathematical Methods wk : Vectors John Magorrian, magog@thphysoxacuk These are work-in-progress notes for the second-year course on mathematical methods The most up-to-date version is available from http://www-thphysphysicsoxacuk/people/johnmagorrian/mm
More informationMathematical Methods wk 1: Vectors
Mathematical Methods wk : Vectors John Magorrian, magog@thphysoxacuk These are work-in-progress notes for the second-year course on mathematical methods The most up-to-date version is available from http://www-thphysphysicsoxacuk/people/johnmagorrian/mm
More informationLecture 2: Linear operators
Lecture 2: Linear operators Rajat Mittal IIT Kanpur The mathematical formulation of Quantum computing requires vector spaces and linear operators So, we need to be comfortable with linear algebra to study
More information2nd Symposium on System, Structure and Control, Oaxaca, 2004
263 2nd Symposium on System, Structure and Control, Oaxaca, 2004 A PROJECTIVE ALGORITHM FOR STATIC OUTPUT FEEDBACK STABILIZATION Kaiyang Yang, Robert Orsi and John B. Moore Department of Systems Engineering,
More information642:550, Summer 2004, Supplement 6 The Perron-Frobenius Theorem. Summer 2004
642:550, Summer 2004, Supplement 6 The Perron-Frobenius Theorem. Summer 2004 Introduction Square matrices whose entries are all nonnegative have special properties. This was mentioned briefly in Section
More informationLecture 7 Spectral methods
CSE 291: Unsupervised learning Spring 2008 Lecture 7 Spectral methods 7.1 Linear algebra review 7.1.1 Eigenvalues and eigenvectors Definition 1. A d d matrix M has eigenvalue λ if there is a d-dimensional
More informationBalanced Truncation 1
Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.242, Fall 2004: MODEL REDUCTION Balanced Truncation This lecture introduces balanced truncation for LTI
More informationChapter 6: Orthogonality
Chapter 6: Orthogonality (Last Updated: November 7, 7) These notes are derived primarily from Linear Algebra and its applications by David Lay (4ed). A few theorems have been moved around.. Inner products
More informationHonors Algebra II MATH251 Course Notes by Dr. Eyal Goren McGill University Winter 2007
Honors Algebra II MATH251 Course Notes by Dr Eyal Goren McGill University Winter 2007 Last updated: April 4, 2014 c All rights reserved to the author, Eyal Goren, Department of Mathematics and Statistics,
More informationLecture 8: Linear Algebra Background
CSE 521: Design and Analysis of Algorithms I Winter 2017 Lecture 8: Linear Algebra Background Lecturer: Shayan Oveis Gharan 2/1/2017 Scribe: Swati Padmanabhan Disclaimer: These notes have not been subjected
More information2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian
FE661 - Statistical Methods for Financial Engineering 2. Linear algebra Jitkomut Songsiri matrices and vectors linear equations range and nullspace of matrices function of vectors, gradient and Hessian
More information5.) For each of the given sets of vectors, determine whether or not the set spans R 3. Give reasons for your answers.
Linear Algebra - Test File - Spring Test # For problems - consider the following system of equations. x + y - z = x + y + 4z = x + y + 6z =.) Solve the system without using your calculator..) Find the
More informationKnowledge Discovery and Data Mining 1 (VO) ( )
Knowledge Discovery and Data Mining 1 (VO) (707.003) Review of Linear Algebra Denis Helic KTI, TU Graz Oct 9, 2014 Denis Helic (KTI, TU Graz) KDDM1 Oct 9, 2014 1 / 74 Big picture: KDDM Probability Theory
More informationHOMEWORK PROBLEMS FROM STRANG S LINEAR ALGEBRA AND ITS APPLICATIONS (4TH EDITION)
HOMEWORK PROBLEMS FROM STRANG S LINEAR ALGEBRA AND ITS APPLICATIONS (4TH EDITION) PROFESSOR STEVEN MILLER: BROWN UNIVERSITY: SPRING 2007 1. CHAPTER 1: MATRICES AND GAUSSIAN ELIMINATION Page 9, # 3: Describe
More informationMath 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination
Math 0, Winter 07 Final Exam Review Chapter. Matrices and Gaussian Elimination { x + x =,. Different forms of a system of linear equations. Example: The x + 4x = 4. [ ] [ ] [ ] vector form (or the column
More informationProblem Set (T) If A is an m n matrix, B is an n p matrix and D is a p s matrix, then show
MTH 0: Linear Algebra Department of Mathematics and Statistics Indian Institute of Technology - Kanpur Problem Set Problems marked (T) are for discussions in Tutorial sessions (T) If A is an m n matrix,
More informationPerron-Frobenius theorem
Perron-Frobenius theorem V.S. Sunder Institute of Mathematical Sciences sunder@imsc.res.in Unity in Mathematics Lecture Chennai Math Institute December 18, 2009 Overview The aim of the talk is to describe
More information