Analysis of Transient Behavior in Linear Differential Algebraic Equations

Size: px
Start display at page:

Download "Analysis of Transient Behavior in Linear Differential Algebraic Equations"

Transcription

1 Analysis of Transient Behavior in Linear Differential Algebraic Equations Dr. Mark Embree and Blake Keeler May 6, 5 Abstract The linearized Navier-Stokes equations form a type of system called a differential algebraic equation (DAE). These equations impose both algebraic and differential constraints on the solution, and present significant challenges to solving. This paper discusses a method for solving DAEs by the Schur factorization to avoid some of the obstacles that are typically present. We also investigate the transient behavior of the solution using results from pseudospectral theory, which will help us understand whether the solution to the linearized Navier-Stokes system will become turbulent if its steady state is disturbed. If a flow does indeed grow substantially in energy, then we must be cautious about using the linearized model. Thus, having information about transient behavior is quite instructive. Introduction The Navier-Stokes equations are system of highly nonlinear PDEs that describe how the velocity and pressure fields of a fluid behave. Given a particular fluid flow problem, one often seeks to understand whether that flow will be stable (laminar) or unstable (turbulent) if it is perturbed from its steady state. One of the most common ways to answer this question is to perform a finite element discretization in space, resulting in a system of linear ODEs in time. That system of ODEs turns out to be a linear differential algebraic equation (DAE), a system in which the coefficient matrix of the derivative vector is singular. This makes solving these equations much more complex, as the generalized eigenvalue problem becomes more difficult to interpret and connect to the question of stability. In particular, there will be infinite eigenvalues that must be dealt with. We will see that these infinite eigenvalues impose algebraic constraints on the solution, forcing it to evolve on a lower dimensional subspace of R n. Hence the name differential algebraic equation. This paper uses a method by which these equations may be solved without any direct need for the generalized eigenvalue problem by surgically removing the infinite eigenvalues from the system without breaking the dynamics, resulting in a physically meaningful solution in exponential form. We investigate how to analyze the transient behavior of the solution. The asymptotic properties of the solutions to differential equations are well-understood. It is commonly known that if a solution to a system of differential equations x (t) = Ax(t) has the form

2 e ta x(), then the eigenvalues of A are the determining factors in the behavior of the solution as t. However, this tells us nothing of what happens between the initial state and the inevitable decay at infinity. It is often possible that the solution experiences growth by several orders of magnitude before settling down and going to zero. This behavior is referred to as transient growth and is a consequence of the departure of the matrix A from normality. One manifestation of this phenomenon is eigenvectors of A associated with distinct eigenvalues being nearly parallel. Unfortunately, the amount of information gained from inspecting the eigenvectors alone is quite limited. They will not tell us how much growth will occur, only that such growth is likely. We can, however, obtain such information by looking not at the spectrum of the matrix, but at the pseudospectra instead. The second portion of this paper applies results from pseudospectral theory to analyze transient behavior in the Navier-Stokes problem. Differential Algebraic Equations Systems of linear differential equations of the form Bx (t) = Ax(t) (.) are commonplace in many applications, and are quite easy to solve when B is invertible. We can simply invert B and use the matrix exponential to find x(t) = e BAt x(). For more practical formulations, one can seek the generalized eigenvalues and eigenvectors of the matrix pencil (A, B) and see that any function of the form x(t) = e µt v is a solution to the differential equation for any eigenpair (µ, v). However, this paper is primarily concerned with the more interesting case where B is singular, making the system a DAE. Problems of this type are much more difficult to solve, since the pencil (A, B) now possesses infinite eigenvalues. This can be seen by writing down the generalized eigenvalue problem as Av = Bv (.) µ and noting that since B is singular, there exist nonzero vectors v for which Bv =. For these v, taking µ = satisfies the eigenvalue problem, making (µ, v) an infinite eigenpair of the matrix pencil. So how shall we handle such an equation? We clearly cannot write x(t) = e t v as a solution. The following section details a method by which we may construct a sensible solution to (.) using the Schur decomposition, thereby avoiding the generalized eigenvalue problem entirely.. Solving a DAE by the Schur Decomposition Details about the solution of DAEs can be found in [. In this section we discuss one particular method that employs the Schur decomposition. Consider (.) where B is singular, but assume that (A, B) is a nonsingular pencil, so there exists some real number λ such that A λb is invertible. We define A λ = (A λb) B and B λ = (A λb) B. Then if we multiply (.) by (A λb) we obtain B λ x (t) = A λ x(t). (.3)

3 Since (A λb) is nonsingular, (.3) is equivalent to (.). Notice also that A λ λb λ = (A λb) A λ(a λb) B = (A λb) (A λb) = I. Therefore, A λ = I + λb λ, so (.3) can be written as Now we write the Schur decomposition of B λ as B λ x (t) = (I + λb λ )x(t). (.4) B λ = QTQ, where Q is a unitary matrix, and T is an upper triangular matrix in which the eigenvalues of B λ may appear in any order. Suppose the pencil has an infinite eigenvalue of algebraic multiplicity m, so that B λ has a zero eigenvalue of algebraic multiplicity m. Then it is convenient to order the eigenvalues so that [ G S T =, N where N is an m m upper triangular matrix corresponding to the zero eigenvalues of B λ (i.e. its diagonal elements are all zero), G is the (n m) (n m) upper triangular matrix corresponding to the nonzero eigenvalues (so G is invertible), and U is some (n m) m matrix. Hence, if we premultiply (.3) by Q, we obtain which can be rewritten as [ G S N Q B λ QQ x (t) = Q (I + λb λ )QQ x(t), (.5) [ I + λg λs Q x (t) = I + λn Q x(t). (.6) Define y(t) = Q x(t), where y(t) = [y, y T with y of length n m and y of length m, so we have [ [ [ [ G S y I + λg λs y =. (.7) N I + λn y From the second block-row, we have Ny = (I + λn)y. Since the diagonal of N is zero, it must be a nilpotent matrix. Let us assume that it has index k, that is N k =, but N k. Therefore, premultiplying by N k we find (N k + λn k )y = N k y =. So N k y =. Then if we premultiply by N k we have (N k + λn k )y = N k y = (N k y ) =. So N k y =. This process can be continued inductively to obtain N y =. 3 y

4 Therefore, y =. Let us consider for a moment what this implies about the set of valid initial conditions x(). Let us first partition Q = [Q, Q so that Q R n (n m) and Q R n m. Then since x(t) = Qy(t) with y (t) =, we have that x() = Q y (). (.8) In other words, the only initial conditions that satisfy the algebraic constraints of the DAE are those in Ran(Q ). Similarly, the relationship x(t) = Q y (t) implies that the solution evolves on a subspace of R n of dimension n m. Hence, the infinite eigenvalues of the pencil (A, B) serve to constrain the solution to a subspace of lower dimension. We will see a very clear example of this in section 3.. Continuing from (.7) with the knowledge that y (t) =, we have that Gy = (I + λg)y, and so the solution is y (t) = e (G +λi)t y (). The full solution to (.7) can now be written as [ [ y (t) e (G +λi)t = y (t) [ y (). (.9) Notice that the entries in the (,) and (,) blocks of the above matrix are irrelevant, but makes a convenient choice. We can then write [ y (t) x(t) = Q y (t) ( [ ) [ G = Q exp t + λi Q y () Q. For simplicity, we will write the solution to (.) as ( [ M x(t) = exp tq Q ) x(), (.) where M = G + λi. Thus, we have formulated a solution to (.) that maintains standard exponential form, despite the fact that the pencil (A, B) had infinite eigenvalues. The next few sections will outline some of the interesting properties of this solution.. The Eigenvalues of M As mentioned previously, if we have a standard system of equations like (.) with B nonsingular, we can simply write the solution in terms of the eigenvalues of the pencil (A, B). For a DAE this is complicated by the existence of infinite eigenvalues, but it is reasonable to expect that the finite eigenvalues of the pencil will still dictate the behavior of the solution. In the following, we prove that this is indeed the case, as the eigenvalues of M in (.) are the same as the finite eigenvalues of the pencil. 4

5 Theorem.. If A, B, M, and λ are as in section., then the spectrum of M is the set of all finite eigenvalues of the pencil (A, B). Proof. Suppose η is a finite eigenvalue of the pencil (A, B). nonzero x. So we can subtract λbx to obtain Then Ax = ηbx for some (A λb)x = (η λ)bx. Since A λb is invertible, η λ must be nonzero. We can rearrange the above to obtain (A λb) Bx = η λ x. Hence is a nonzero eigenvalue of B η λ λ = (A λb) B. If the Schur decomposition of B λ is as in section., then is an eigenvalue of G. Hence, η λ is an eigenvalue of η λ G. Finally, this implies that η is an eigenvalue of M = G + λi. The reverse inclusion follows the same argument, and so the theorem follows. This result, while it is what we had hoped for, is interesting in that no matter what shift λ is applied to make A λb invertible, the key matrix in the solution always has the same eigenvalues. In fact, as we shall see, the solution is entirely independent of λ..3 The Solution s Independence of λ In this section, we will prove that our solution (.) is independent of the parameter λ, whose only purpose was to make the pencil (A, B) nonsingular. We begin with the following lemmas. Lemma.. Let A, B C n n and suppose λ and µ are any real numbers such that A λb and A µb are invertible. Define B λ = (A λb) B and B µ = (A µb) B. Then η σ(b λ ) = η + (λ µ)η σ(b µ). Proof. Suppose η σ(b λ ). Then (A λb) Bx = ηx for some nonzero vector x. Premultiplying by (A λb) and adding and subtracting ηµbx from the right gives A rearrangement yields Bx = η(a µb)x + η(µ λ)bx. ( + η(λ µ))bx = η(a µb)x. If η, then since A µb is invertible, + η(λ µ) must be nonzero. If η =, then clearly + η(λ µ). Therefore, we can write Bx = η (A µb)x. + η(λ µ) Premultiplying by (A µb) demonstrates the result. 5

6 Lemma.3. Let A, B, B λ and B µ be as in Lemma. and suppose the pencil (A, B) has an infinite eigenvalue of algebraic multiplicity m, so that B λ and B µ each have a zero eigenvalue of algebraic multiplicity m. Let the Schur decompositions of B λ and B µ be given by [ Gλ S B λ = Q λ N λ Q [ Gµ S B µ = U µ N µ U, respectively, where Q, U are unitary, N λ, N µ R m m have zero diagonal entries, and G λ, G µ R (n m) (n m) are invertible. Partition Q = [Q, Q and U = [U, U so that Q, U R n (n m). Then Ran(Q ) = Ran(U ). Proof. Let us write B λ in Jordan form as B λ X = XJ where the nonzero eigenvalues are written in the upper left (n m) (n m) block of J. If we partition X = [X, X in the same way that we partitioned Q, then Ran(X ) is the invariant subspace of B λ associated with the nonzero eigenvalues. Observe that Ran(Q ) is also an invariant subspace of B λ associated with the nonzero eigenvalues. Hence Ran(X ) = Ran(Q ). Next we can premultiply B λ X = XJ by A λb and add and subtract µbxj on the right to obtain Simple rearrangement yields BX = (A µb)xj + (µ λ)bxj. BX(I + (λ µ)j) = (A µb)xj. For µ sufficiently close to λ, I + (λ µ)j is invertible, so we can write (A µb) BX = XJ(I + (λ µ)j). (.) It is important to note that J(I + (λ µ)j) will retain the block upper triangular form of J. This is shown by partitioning J by its Jordan blocks as J = diag{j,..., J p } so that J (I + (λ µ)j ) J(I + (λ µ)j) =..., J p (I p + (λ µ)j p ) where I,..., I p are the appropriate diagonal partitions of the identity matrix. Therefore, the diagonal of J(I + (λ µ)j) will retain the structure of the diagonal of J. That is, the nonzero eigenvalues (of B µ this time) will be listed as the first elements along the diagonal. Hence by (.) X is the invariant subspace associated with the nonzero eigenvalues of B µ. Again, since such invariant subspaces are unique, we have that Ran(U ) = Ran(X ). When we recall that Ran(X ) = Ran(Q ), we have the desired result. Thus far, we have only proved the theorem for µ sufficiently close to λ, since that condition guaranteed that I + (λ µ)j would be invertible. However, we claim that this matrix is invertible for all valid choices of λ and µ. To see this, note that I+(λ µ)j will be invertible provided + (λ µ)η is nonzero for all eigenvalues η of B λ. If η σ(b λ ), Lemma. states that η/( + (λ µ)η) is an eigenvalue of B µ. Since the eigenvalues of B µ must be finite, + (λ µ)η must be nonzero. The result therefore generalizes to all λ and µ for which A λb and A µb are invertible. 6

7 Lemma.4. If all matrices are as in Lemma.3, [ Q Ĝµ Ŝ B µ Q = µ Nµ, where Ĝµ = Q U G µ U Q. Proof. Direct computation shows that [ Q Q B µ Q = Q B µ [ Q Q = Q B µ Q Q B µ Q Q B µ Q Q B µ Q [ Ĝ =: µ Ŝ µ. Q B µ Q Nµ By Lemma.3 we know that Ran(Q ) = Ran(U ) is an invariant subspace of B µ, so we have Q B µ Q = Q Q Z =, for some matrix Z R (n m) (n m), since the columns of Q are orthogonal to the columns of Q. Therefore, [ Q Ĝµ Ŝ B µ Q = µ. Nµ Also, if we expand the Schur form of B µ, we obtain and hence, B µ = U G µ U + (U S µ + U N µ )U, Ĝ µ = Q B µ Q = Q U G µ U Q, since the columns of Q are orthogonal to the columns of U. We have therefore proven that Q B µ Q remains upper triangular with an explicit formula for the upper left block. Lemma.5. B λ B µ = (λ µ)b λ B µ. Proof. Notice that (A µb) (A λb) = (λ µ)b. If we premultiply by (A λb) and postmultiply by (A µb), we obtain (A λb) (A µb) = (λ µ)(a λb) B(A µb), which is a generalized form of the First Resolvent Identity. If we then postmultiply this equation by B, we have the desired result. 7

8 This is the final lemma that we need in order to prove the following theorem. Theorem.6. The solution formulated in section. is independent of the choice of the parameter λ, provided A λb is nonsingular. That is, in the context of the preceding lemmas, [ [ G µ + µi G U U λ = Q + λi Q. Proof. By Lemma.5, we have B λ B µ = (λ µ)b λ B µ. Multiplying by Q on the left and Q on the right, we obtain Q B λ Q + Q B µ Q = (λ µ)q B λ QQ B µ Q. If we replace B λ and B µ by their Schur decompositions and invoke Lemma.4, we can write [ [ [ [ Gλ S λ Ĝµ Ŝ µ Gλ S = (λ µ) λ Ĝ µ Ŝ µ. N λ Nµ N λ Nµ Restricting our attention to the (,) block of this matrix equation, we see that G λ Ĝµ = (λ µ)g λ Ĝ µ. Since Ĝµ = Q U G µ U Q, we can multiply on the left by Q and on the right by Q to obtain Q G λ Q (Q Q U )G µ (U Q Q ) = (λ µ)(q G λ Q )U G µ (U Q Q ). (.) Since Q Q and U U are both orthogonal projectors onto Ran(Q ) = Ran(U ), we have that Q Q = U U. So we have and similarly Q Q U = U U U = U U U Q = Q Q Q = Q. We will use these identities freely in the following calculations. Continuing from (.) we have Q G λ Q U G µ U = (λ µ)(q G λ Q )(U G µ U ). Then if we multiply on the left by Q G λ Q, we obtain Q Q (Q G λ Q )(U G µ U ) = (λ µ)u G µ U. Next we can multiply on the right by U G µ U, which gives U G µ U Q G λ Q = (λ µ)u U. 8

9 Since U U = Q Q, we can rearrange the above to see that U (G µ + µi)u = Q (G λ + λi)q. It follows that [ G µ + µi U [ G U λ = Q + λi Q. Now that we know the solution will not be affected by our choice of the parameter λ, we can be confident in using this approach to solve DAEs..4 Linearized Navier-Stokes As mentioned in the introduction, the primary application that we wish to study is the numerical solution of the incompressible Navier-Stokes equations, ν u + u u + p = u t (.3) u =, (.4) where u is the velocity field of the fluid and p is the pressure field. Equations (.3) and (.4) form a system of nonlinear time-dependent partial differential equations, for which finding a solution is a difficult task. One common approach is to find a steady state solution and linearize the system around that steady state using a finite element formulation, the details of which we will not discuss here. For a full description of this process, see [3. The key idea is to approximate u u h and p p h where u h = n a i (t)φ i (x) and m b j (t)ψ j (x) for some finite collections of basis functions {φ,..., φ n } and {ψ,..., ψ m }. With this discretization, the linearized problem is reduced to a first order system of linear differential equations in time, given by [ A i= [ K C x (t) = C j= x(t), (.5) [ a(t) where x(t) = with a(t) = [a b(t) (t),..., a n (t) and b(t) = [b (t),..., b m (t), the vectors of time-dependent coefficients in u h and p h. Also, A C n n is Hermitian positive definite, C C n m has full column rank (m n), and K C n n. From this we can see the connection to DAEs, since the matrix on the left side of (.5) is singular, as it has m zero columns. It should be noted that (.5) describes how a fluid flow system will behave in time, but only under the assumption that it stays sufficiently close to the steady state. Therefore, the transient properties of the system are very important. If we can conclude that the solution will likely grow substantially from its initial state, then the necessary assumptions for the linearized analysis will not be satisfied. For more details about linear stability analysis, see [ and [8. The next sections describe the necessary tools to analyze how the pseudospectral properties of the matrix M in the solution to (.5) influence transient behavior, allowing us to predict whether or not the linearization will remain descriptive. 9

10 3 Pseudospectra As we turn our attention to transient growth in the solutions to DAEs, we will find the concept of pseudospectra to be quite useful. Conceptually, the pseudospectra of a matrix A are sets of numbers in the complex plane that are almost eigenvalues. Before we introduce the more mathematically precise definition, notice that if z is not an eigenvalue of A (z / σ(a)), the resolvent norm (A zi) is defined and will grow in magnitude as A zi gets closer to being singular, which is generally when z approaches an eigenvalue of A. Therefore, it is reasonable to take the convention (A zi) = if z σ(a). This motivates the definition of pseudospectra that we will use here, taken from [6, which states that if ε > is arbitrary and A C n n, then the ε-pseudospectrum σ ε (A) of A is the set of z C such that (A zi) > ε. We will not look too deeply into the theory of pseudospectra in this paper; we will simply use a few results that will prove to be useful in studying the transient behavior of our DAE solutions, as discussed in the next section. 3. Pseudospectra and Transient Growth Recall that our solution to a general DAE in section had the form ( [ ) M x(t) = exp tq Q x(). Therefore, since x() is constant, the transient behavior of the solution is determined primarily by the behavior of the matrix exponential as time increases. Specifically, we only need to know about e tm since Q and Q simply make an orthogonal similarity transformation, under which the -norm is invariant. This is where some results from pseudospectral theory prove quite useful. In particular, we can use pseudospectra to put a lower bound on the maximum value of the matrix exponential. From Theorem 5.4 in [6, we have that if the pseudospectral abscissa, α ε (A) is defined by α ε (A) = sup Re(z), (3.) z σ ε (A) which in simple terms is the real part of the rightmost point in the ε-pseudospectrum, then Taking the supremum over ε gives sup e ta α ε (A)/ε ε >. (3.) t sup t e ta K(A) = sup α ε (A)/ε, (3.3) ε>

11 where K(A) is known as the Kreiss constant. Therefore, if the ε-pseudospectrum of A extends farther than ε into the right-half plane, we can guarantee that e ta must grow by at least a factor of α ε (A)/ε from the initial state. It is important to note that this only guarantees growth in the same norm that defines the pseudospectra. Determining an effective method for computing pseudospectra in the appropriate norm will be discussed in section 3.3. Also, notice that we can only guarantee such growth in the matrix exponential e ta. If the solution to a DAE has the form x(t) = e tm x(), then we have only that sup t x(t) sup e tm x(). (3.4) t Hence, we need not see transient growth for all initial conditions x(). However, by definition of an operator norm, e tm = max e tm x. x = Therefore, there does exist an initial state x() such that maximal transient growth at time t is realized. The next section details a simple three-dimensional example where we see a clear connection between the Kreiss constant and the growth of the solution. 3. A Three-Dimensional Example Consider the dynamical system where B = Bx (t) = Ax(t), (3.5) and A = The matrix A was chosen arbitrarily to simply be nonsingular. B, on the other hand, was specifically constructed so that it has a zero eigenvalue of algebraic multiplicity one. When we follow the procedure outlined in section. with any appropriate λ, we obtain where Q = x(t) = Q exp t Q x(), (3.6) and x() must be in the range of the first two columns of Q, by (.8). A suitable choice for the initial condition was x() = Notice that in the context of (.), M = [

12 The eigenvectors of this matrix are [, and [, /, which are close to parallel, so from the standard qualitative perspective we should expect transient growth. The pseudospectra will reflect this and give us a more quantitative idea of exactly how much growth to expect. Before we look at the pseudospectra, however, recall that since B (and therefore B λ ) had a zero eigenvalue of multiplicity one, we should expect the solution to evolve on a twodimensional subspace of R 3. And indeed, as shown in figure 3. we see that the solution is constrained to evolve on a plane. z Initial Point Final Point y 5 5 x Figure 3.: The path x(t) and the D subspace on which it lives Figure 3.: Pseudospectra of M We see also that there is some transient growth in x(t), so this should be reflected in the pseudospectral properties of M. Based on our discussion in section 3., we should expect

13 the ε-pseudospectra to extend farther than ε into the right half-plane for some ε >. As shown in figure 3., this is exactly what we see. The black dots show the eigenvalues of M, which are both in the left half-plane, which is sufficient to guarantee that the solution is asymptotically stable. However, the fact that the pseudospectra, denoted by the colored contours, extend so far to the right of the imaginary axis gives intermediate growth before the decay in the solution at infinity. To understand just how much growth we should expect in x(t), we compute the ratios α ε (M)/ε for various ε between. and to find a close approximation to the Kreiss constant. Figure 3.3 demonstrates the relationship between ε and the ratios α ε (M)/ε. The peak that is highlighted in the figure occurs at ε =.398 and takes on a value of α ε (M)/ε = 3.488, which gives us that K(M) Hence, we should expect the norm of x(t) to grow by at least a factor of from its initial value before its final decay. This behavior is demonstrated in figure 3.3 where the red line indicates our predicted level of growth. We see that the norm grows to a point just past our prediction before quickly shrinking to its steady state at zero αε(m) ε ε Figure 3.3: α ε (M)/ε as a function of ε. x(t) αε ε x t Figure 3.4: Transient Behavior of x(t). 3

14 This example provides a clear demonstration of the relationship between the pseudospectra of M and the transient behavior of the solution norm. However, it is worth noting that we have simply defaulted to using the -norm to compute both the pseudospectra and the growth of the solution. In this case, it made sense because we were only interested in the path of the particle, but in the case of Navier-Stokes it is not so obvious how growth should be described. We discuss this issue in the following section. 3.3 Finding the Appropriate Norm We are interested [ in transient behavior in the solution to the Navier-Stokes problem, which u takes the form, where u and p are the velocity and pressure fields, respectively. The p task at hand is to determine a meaningful way to describe what we mean by growth. For example, if the -norm of the velocity-pressure vectors exhibits transient growth, does that imply that there will be any noticeable behavior in the physical system? In general, it is difficult to assign any physical meaning to such a norm. However, one norm that proves quite useful in this case is the energy-norm, denoted E. If we let Ω denote the physical domain of the problem, then the energy norm for Navier-Stokes is given by [ u p E = ( u H + ) / p L = Ω u u + p dω where H is the H semi-norm and L is the L norm. Suppose now that /, [ uh a finite element approximation to the solution, with u h = n a i φ i (x) and v h = m b i ψ i (x), where {φ,..., φ n } and {ψ,..., ψ m } are the finite element basis functions. Note that any boundary conditions enforced will vary with the particular flow being studied, but this does not affect our calculations. Computing the energy norm of our finite element approximation gives [ u p E [ uh p h E = = Ω n i= u h u h + p hdω n a i a j j= Ω φ i φ j dω + i= m i= m b i b j j= Ω i= p h ψ i (x)ψ j (x)dω. is 4

15 If we let a = [a,..., a n, b = [b,..., b m, (K) ij = φ Ω i φ j dω, and (A) ij = ψ Ω i(x)ψ j (x)dω, we can write [ u [ uh = a Ka + b Ab p E p h E = [ a b = x Hx. [ K A [ a b [ [ K a where H = and x =. Notice that under the typical restriction that φ A b i and ψ i obey Dirichlet boundary conditions, both K and A are Gram matrices and therefore are symmetric positive definite. Hence, H is symmetric positive definite, and so the function u, v H = v Hu defines an inner product that induces the corresponding norm u H = (u Hu) /. Therefore, we have that [ [ u uh p = x H. E Coincidentally, K and A are the same as in (.4), so it requires no additional effort to construct the appropriate H. Additionally, if RR = H is a Cholesky factorization of H (which must exist since H is symmetric positive definite), then x H = x Hx = x RR x = (R x) R x = R x. Therefore, we can compute the energy norm of our finite element solution by simply computing the -norm of the vector of finite element coefficients, x, transformed by R. Similarly, if we wish to compute the energy norm of a matrix B, we can first note that p h E By E y E = y B HBy y Hy = (R By) R By (R y) Ry = R By R y = R BR z z where z = R y. Since R is invertible, maximizing over y is equivalent to maximizing over z, and therefore B E = R BR. (3.7) Thus, we can compute the energy norm of a linear operator by performing the similarity transformation with respect to R and then computing the -norm. 5

16 3.4 Results for Navier-Stokes We have finally developed all of the necessary tools to analyze the transient behavior of the linearized Navier-Stokes problem given by [ [ A K C x (t) = C x(t) (3.8) Before we compute any solutions to this system, we perform the change of coordinates as described in section 3.3, resulting in the modified system [ [ A K C R R y (t) = R C R y(t), (3.9) where y(t) = R x(t). Now when we compute the -norm of the solution to (3.9), we obtain the energy norm of the solution to (3.8). We will compute solutions to this system for two classical fluid flow problems in two physical dimensions, namely flow over a step and flow around an obstacle. All of the linearized systems that we analyze here were generated using the IFISS software package from the University of Manchester with grid size 4 ( 4 vertical elements) [4,[5. This is the smallest grid size on which we can obtain reasonable resolution of the eddie currents. With this grid, the coefficient matrices in the step flow problem are of size The A and K blocks have dimension , while the bottom right zero blocks have size Therefore, when we follow the procedure outlined in section., we obtain a solution of the form x(t) = exp ( [ M tq Q ) x(), (3.) where M has dimension (797 99). For the obstacle flow problem, the coefficient matrices are smaller, having dimension A and K have dimension 9 9, and M in the solution has size 896. In the following sections we will show how the pseudospectral properties of M change with the viscosity ν by computing approximations to the Kreiss constant, which will in turn tell us how the transient behavior of the system changes with ν Flow Over a Step The step flow system involves a viscous fluid flowing off the edge of a small step, creating an eddy current at the base of the step as well as a noticeable dip at the surface of the fluid. A typical steady state solution to this problem is shown in figure 3.5. Figures show the pseudospectra plots of M for viscosity values of ν =,,, and

17 Streamlines: uniform Pressure field Figure 3.5: Flow over a Step with ν = Figure 3.6: Pseudospectra for step flow, ν =. 7

18 Figure 3.7: Pseudospectra for step flow, ν = Figure 3.8: Pseudospectra for step flow, ν = 4. 8

19 Figure 3.9: Pseudospectra for step flow, ν = 5. It is easy to see that as the viscosity decreases the spectrum of M drifts closer to the imaginary axis, but all of the eigenvalues remain in the left half-plane for these viscosities, keeping the system asymptotically stable. Perhaps less obvious is the fact that some of the pseudospectral contours also begin to shift farther into the right half-plane, indicating that the Kreiss constant is increasing as the viscosity decreases. As aesthetically pleasing as these plots are, it is difficult to obtain any quantitative information from them, so table 3. lists α approximations to the Kreiss constant, K(M) = sup ε (M) ε>, for each ν. From this we can ε clearly see how the viscosity increases the level of transient growth in the solution. For a nice value of the viscosity like /, maximum energy growth is only guaranteed to reach about. times the energy of its original state, but for ν = /5 the solution s maximal energy growth is guaranteed to reach a level of about 7.8 times the initial value. Therefore, we can feel fairly comfortable using the linearized system (3.8) to model the step flow for viscosities of / or lower since we remain within a factor of of the initial energy, but for higher viscosities we see that it is possible for the energy to grow substantially from the initial value, making linear analysis an unwise choice. Table 3.: Approximate Kreiss constants for various step flow viscosities. ν / / /4 /5 Approximate K(M) Next we will look at similar results for flow around an obstacle. 9

20 3.4. Flow Around an Obstacle The obstacle flow problem involves a viscous fluid flowing around a rectangular object in the center of the flow path. A typical steady state solution is shown in figure 3.. In a similar fashion to the previous section, figures show the pseudospectra plots of the M matrix for viscosities ν =,,, and. We did not use ν = as in the step flow problem because IFISS did not converge to a steady state solution for that value. Streamlines: uniform Pressure field Figure 3.: Flow around an obstacle, ν = Figure 3.: Pseudospectra for obstacle flow, ν =.

21 Figure 3.: Pseudospectra for obstacle flow, ν = Figure 3.3: Pseudospectra for obstacle flow, ν = 4.

22 Figure 3.4: Pseudospectra for obstacle flow, ν = 45. Despite the fact that the pseudospectral contours corresponding to larger ε values extend farther into the right-half plane than for the step, table 3.4. shows that the approximate Kreiss constants follow similar patterns to those of the step flow. We see once again that for low viscosities, the linearized system will serve us just fine, but for the higher values we see enough potential growth that using a linearized system is not a viable choice. Table 3.: Approximate Kriess constants for various obstacle flow viscosities. ν / / /4 /45 Approximate K(M) Future Work Our study of these two Navier-Stokes problems demonstrates how we can combine our method for solving DAEs from section. with results from pseudospectral theory to get a better understanding of the transient properties of fluid systems that are disturbed from their steady state. From this we can predict how valid it is to use the linearized system to model the fluid system. However, one key point that we have not considered in this paper is the convergence of the Kreiss constants for the Navier-Stokes matrices as we refine the grid. That is, are the computed values on grid 4 a reasonable approximation to the true values, or would they change substantially if we went to grid 5 or 6? These numerical experiments would involve a great deal more time and computational power to perform. Thus, a point of further research would be implementing these finer grid experiments on more powerful machines, using techniques from sparse linear algebra.

23 It would also be interesting to investigate what conditions x() must satisfy in order to guarantee transient growth. We know that such initial conditions must exist, so having a way to determine whether a given initial state will or will not lead to transient growth would be an incredibly useful tool. 4 Conclusions In this paper we have covered a great deal of ground. We have developed a method for solving an arbitrary DAE via the Schur decomposition and proved that the solution that we obtain is not affected by our parameter λ. This proof required us to dig deep into the structure of the orthogonal transformations of the Schur form and how they behave when we change λ. We then discussed how the pseudospectra of M affect the potential for transient growth in the solution, giving a three-dimensional example to make the connection clear. Equipped with these tools we turned our attention to the linearized Navier-Stokes equations, deriving the appropriate change of coordinates so that the -norm of the solution that we compute is equivalent to the energy norm of the solution to the original system. We then investigated the effect of the viscosity on the pseudospectra of M and in turn the effect on the possible energy growth of the solution. This information allows us to be better informed about how valid it is to use the linearized system to analyze fluid stability. 3

24 References [ J. S. Baggett, T. A. Driscoll, L. N. Trefethen, A Mostly Linear Model of Transition to Turbulence, Physics of Fluids. 7, (995). [ S. L. Campbell, C. D. Meyer Jr., Generalized Inverses of Linear Transformations, Pitman, London, 979. [3 H. C. Elman, D. Silvester, A. Wathen, Finite Elements and Fast Iterative Solvers, Oxford University Press, 5. [4 H. C. Elman, A. Ramage, D. J. Silvester, IFISS: A Computational Laboratory for Investigating Incompressible Flow Problems, SIAM Review, 56, 6-73 (4). [5 H. C. Elman, A. Ramage, D. J. Silvester, Algorithm 866: IFISS, A MATLAB Toolbox for Modelling Incompressible Flow, ACM Transactions on Mathematical Software, 33 (7). [6 M. Embree, L. Trefethen, Spectra and Pseudospectra, Princeton University Press, 5. [7 M. Gunzburger, Finite Element Methods for Viscous Incompressible Fluids, Academic Press, 989. [8 L. N. Trefethen, A. E. Trefethen, S. C. Reddy, T. A. Driscoll, Hydrodynamic Stability without Eigenvalues, Science, 6, (993). [9 T. G. Wright, (), EigTool, Software available at 4

arxiv: v3 [math.na] 27 Jun 2017

arxiv: v3 [math.na] 27 Jun 2017 PSEUDOSPECTRA OF MATRIX PENCILS FOR TRANSIENT ANALYSIS OF DIFFERENTIAL-ALGEBRAIC EQUATIONS MARK EMBREE AND BLAKE KEELER arxiv:6.44v3 [math.na] 7 Jun 7 Abstract. To understand the solution of a linear,

More information

MATH 5524 MATRIX THEORY Problem Set 5

MATH 5524 MATRIX THEORY Problem Set 5 MATH 554 MATRIX THEORY Problem Set 5 Posted Tuesday April 07. Due Tuesday 8 April 07. [Late work is due on Wednesday 9 April 07.] Complete any four problems, 5 points each. Recall the definitions of the

More information

LINEAR ALGEBRA KNOWLEDGE SURVEY

LINEAR ALGEBRA KNOWLEDGE SURVEY LINEAR ALGEBRA KNOWLEDGE SURVEY Instructions: This is a Knowledge Survey. For this assignment, I am only interested in your level of confidence about your ability to do the tasks on the following pages.

More information

Finding Rightmost Eigenvalues of Large, Sparse, Nonsymmetric Parameterized Eigenvalue Problems

Finding Rightmost Eigenvalues of Large, Sparse, Nonsymmetric Parameterized Eigenvalue Problems Finding Rightmost Eigenvalues of Large, Sparse, Nonsymmetric Parameterized Eigenvalue Problems AMSC 663-664 Final Report Minghao Wu AMSC Program mwu@math.umd.edu Dr. Howard Elman Department of Computer

More information

MATH 583A REVIEW SESSION #1

MATH 583A REVIEW SESSION #1 MATH 583A REVIEW SESSION #1 BOJAN DURICKOVIC 1. Vector Spaces Very quick review of the basic linear algebra concepts (see any linear algebra textbook): (finite dimensional) vector space (or linear space),

More information

Linear Algebra Massoud Malek

Linear Algebra Massoud Malek CSUEB Linear Algebra Massoud Malek Inner Product and Normed Space In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n An inner product

More information

Eigenvectors and Hermitian Operators

Eigenvectors and Hermitian Operators 7 71 Eigenvalues and Eigenvectors Basic Definitions Let L be a linear operator on some given vector space V A scalar λ and a nonzero vector v are referred to, respectively, as an eigenvalue and corresponding

More information

Numerical Linear Algebra Homework Assignment - Week 2

Numerical Linear Algebra Homework Assignment - Week 2 Numerical Linear Algebra Homework Assignment - Week 2 Đoàn Trần Nguyên Tùng Student ID: 1411352 8th October 2016 Exercise 2.1: Show that if a matrix A is both triangular and unitary, then it is diagonal.

More information

Pseudospectra and Nonnormal Dynamical Systems

Pseudospectra and Nonnormal Dynamical Systems Pseudospectra and Nonnormal Dynamical Systems Mark Embree and Russell Carden Computational and Applied Mathematics Rice University Houston, Texas ELGERSBURG MARCH 1 Overview of the Course These lectures

More information

Chapter 7. Canonical Forms. 7.1 Eigenvalues and Eigenvectors

Chapter 7. Canonical Forms. 7.1 Eigenvalues and Eigenvectors Chapter 7 Canonical Forms 7.1 Eigenvalues and Eigenvectors Definition 7.1.1. Let V be a vector space over the field F and let T be a linear operator on V. An eigenvalue of T is a scalar λ F such that there

More information

w T 1 w T 2. w T n 0 if i j 1 if i = j

w T 1 w T 2. w T n 0 if i j 1 if i = j Lyapunov Operator Let A F n n be given, and define a linear operator L A : C n n C n n as L A (X) := A X + XA Suppose A is diagonalizable (what follows can be generalized even if this is not possible -

More information

Eigenvalue and Eigenvector Problems

Eigenvalue and Eigenvector Problems Eigenvalue and Eigenvector Problems An attempt to introduce eigenproblems Radu Trîmbiţaş Babeş-Bolyai University April 8, 2009 Radu Trîmbiţaş ( Babeş-Bolyai University) Eigenvalue and Eigenvector Problems

More information

Iterative Methods for Solving A x = b

Iterative Methods for Solving A x = b Iterative Methods for Solving A x = b A good (free) online source for iterative methods for solving A x = b is given in the description of a set of iterative solvers called templates found at netlib: http

More information

Conceptual Questions for Review

Conceptual Questions for Review Conceptual Questions for Review Chapter 1 1.1 Which vectors are linear combinations of v = (3, 1) and w = (4, 3)? 1.2 Compare the dot product of v = (3, 1) and w = (4, 3) to the product of their lengths.

More information

Review problems for MA 54, Fall 2004.

Review problems for MA 54, Fall 2004. Review problems for MA 54, Fall 2004. Below are the review problems for the final. They are mostly homework problems, or very similar. If you are comfortable doing these problems, you should be fine on

More information

5.3 The Power Method Approximation of the Eigenvalue of Largest Module

5.3 The Power Method Approximation of the Eigenvalue of Largest Module 192 5 Approximation of Eigenvalues and Eigenvectors 5.3 The Power Method The power method is very good at approximating the extremal eigenvalues of the matrix, that is, the eigenvalues having largest and

More information

A Robust Preconditioned Iterative Method for the Navier-Stokes Equations with High Reynolds Numbers

A Robust Preconditioned Iterative Method for the Navier-Stokes Equations with High Reynolds Numbers Applied and Computational Mathematics 2017; 6(4): 202-207 http://www.sciencepublishinggroup.com/j/acm doi: 10.11648/j.acm.20170604.18 ISSN: 2328-5605 (Print); ISSN: 2328-5613 (Online) A Robust Preconditioned

More information

Lecture 3: QR-Factorization

Lecture 3: QR-Factorization Lecture 3: QR-Factorization This lecture introduces the Gram Schmidt orthonormalization process and the associated QR-factorization of matrices It also outlines some applications of this factorization

More information

arxiv: v1 [math.na] 5 May 2011

arxiv: v1 [math.na] 5 May 2011 ITERATIVE METHODS FOR COMPUTING EIGENVALUES AND EIGENVECTORS MAYSUM PANJU arxiv:1105.1185v1 [math.na] 5 May 2011 Abstract. We examine some numerical iterative methods for computing the eigenvalues and

More information

Notes on Eigenvalues, Singular Values and QR

Notes on Eigenvalues, Singular Values and QR Notes on Eigenvalues, Singular Values and QR Michael Overton, Numerical Computing, Spring 2017 March 30, 2017 1 Eigenvalues Everyone who has studied linear algebra knows the definition: given a square

More information

Balanced Truncation 1

Balanced Truncation 1 Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.242, Fall 2004: MODEL REDUCTION Balanced Truncation This lecture introduces balanced truncation for LTI

More information

Lecture 4: Analysis of MIMO Systems

Lecture 4: Analysis of MIMO Systems Lecture 4: Analysis of MIMO Systems Norms The concept of norm will be extremely useful for evaluating signals and systems quantitatively during this course In the following, we will present vector norms

More information

Perturbation theory for eigenvalues of Hermitian pencils. Christian Mehl Institut für Mathematik TU Berlin, Germany. 9th Elgersburg Workshop

Perturbation theory for eigenvalues of Hermitian pencils. Christian Mehl Institut für Mathematik TU Berlin, Germany. 9th Elgersburg Workshop Perturbation theory for eigenvalues of Hermitian pencils Christian Mehl Institut für Mathematik TU Berlin, Germany 9th Elgersburg Workshop Elgersburg, March 3, 2014 joint work with Shreemayee Bora, Michael

More information

B553 Lecture 5: Matrix Algebra Review

B553 Lecture 5: Matrix Algebra Review B553 Lecture 5: Matrix Algebra Review Kris Hauser January 19, 2012 We have seen in prior lectures how vectors represent points in R n and gradients of functions. Matrices represent linear transformations

More information

Definite versus Indefinite Linear Algebra. Christian Mehl Institut für Mathematik TU Berlin Germany. 10th SIAM Conference on Applied Linear Algebra

Definite versus Indefinite Linear Algebra. Christian Mehl Institut für Mathematik TU Berlin Germany. 10th SIAM Conference on Applied Linear Algebra Definite versus Indefinite Linear Algebra Christian Mehl Institut für Mathematik TU Berlin Germany 10th SIAM Conference on Applied Linear Algebra Monterey Bay Seaside, October 26-29, 2009 Indefinite Linear

More information

The Eigenvalue Problem: Perturbation Theory

The Eigenvalue Problem: Perturbation Theory Jim Lambers MAT 610 Summer Session 2009-10 Lecture 13 Notes These notes correspond to Sections 7.2 and 8.1 in the text. The Eigenvalue Problem: Perturbation Theory The Unsymmetric Eigenvalue Problem Just

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 18 Outline

More information

EIGENVALUE PROBLEMS. Background on eigenvalues/ eigenvectors / decompositions. Perturbation analysis, condition numbers..

EIGENVALUE PROBLEMS. Background on eigenvalues/ eigenvectors / decompositions. Perturbation analysis, condition numbers.. EIGENVALUE PROBLEMS Background on eigenvalues/ eigenvectors / decompositions Perturbation analysis, condition numbers.. Power method The QR algorithm Practical QR algorithms: use of Hessenberg form and

More information

Notes for CS542G (Iterative Solvers for Linear Systems)

Notes for CS542G (Iterative Solvers for Linear Systems) Notes for CS542G (Iterative Solvers for Linear Systems) Robert Bridson November 20, 2007 1 The Basics We re now looking at efficient ways to solve the linear system of equations Ax = b where in this course,

More information

Extreme Values and Positive/ Negative Definite Matrix Conditions

Extreme Values and Positive/ Negative Definite Matrix Conditions Extreme Values and Positive/ Negative Definite Matrix Conditions James K. Peterson Department of Biological Sciences and Department of Mathematical Sciences Clemson University November 8, 016 Outline 1

More information

Linear Algebra Methods for Data Mining

Linear Algebra Methods for Data Mining Linear Algebra Methods for Data Mining Saara Hyvönen, Saara.Hyvonen@cs.helsinki.fi Spring 2007 1. Basic Linear Algebra Linear Algebra Methods for Data Mining, Spring 2007, University of Helsinki Example

More information

Robust control of resistive wall modes using pseudospectra

Robust control of resistive wall modes using pseudospectra Robust control of resistive wall modes using pseudospectra M. Sempf, P. Merkel, E. Strumberger, C. Tichmann, and S. Günter Max-Planck-Institut für Plasmaphysik, EURATOM Association, Garching, Germany GOTiT

More information

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.

More information

MATH 5524 MATRIX THEORY Problem Set 4

MATH 5524 MATRIX THEORY Problem Set 4 MATH 5524 MATRIX THEORY Problem Set 4 Posted Tuesday 28 March 217. Due Tuesday 4 April 217. [Corrected 3 April 217.] [Late work is due on Wednesday 5 April.] Complete any four problems, 25 points each.

More information

Numerical Linear Algebra

Numerical Linear Algebra Numerical Linear Algebra The two principal problems in linear algebra are: Linear system Given an n n matrix A and an n-vector b, determine x IR n such that A x = b Eigenvalue problem Given an n n matrix

More information

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators. MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators. Adjoint operator and adjoint matrix Given a linear operator L on an inner product space V, the adjoint of L is a transformation

More information

MIT Final Exam Solutions, Spring 2017

MIT Final Exam Solutions, Spring 2017 MIT 8.6 Final Exam Solutions, Spring 7 Problem : For some real matrix A, the following vectors form a basis for its column space and null space: C(A) = span,, N(A) = span,,. (a) What is the size m n of

More information

COMP 558 lecture 18 Nov. 15, 2010

COMP 558 lecture 18 Nov. 15, 2010 Least squares We have seen several least squares problems thus far, and we will see more in the upcoming lectures. For this reason it is good to have a more general picture of these problems and how to

More information

Lecture 7. Econ August 18

Lecture 7. Econ August 18 Lecture 7 Econ 2001 2015 August 18 Lecture 7 Outline First, the theorem of the maximum, an amazing result about continuity in optimization problems. Then, we start linear algebra, mostly looking at familiar

More information

Numerical Methods for Solving Large Scale Eigenvalue Problems

Numerical Methods for Solving Large Scale Eigenvalue Problems Peter Arbenz Computer Science Department, ETH Zürich E-mail: arbenz@inf.ethz.ch arge scale eigenvalue problems, Lecture 2, February 28, 2018 1/46 Numerical Methods for Solving Large Scale Eigenvalue Problems

More information

A Review of Preconditioning Techniques for Steady Incompressible Flow

A Review of Preconditioning Techniques for Steady Incompressible Flow Zeist 2009 p. 1/43 A Review of Preconditioning Techniques for Steady Incompressible Flow David Silvester School of Mathematics University of Manchester Zeist 2009 p. 2/43 PDEs Review : 1984 2005 Update

More information

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra. DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1

More information

21 Linear State-Space Representations

21 Linear State-Space Representations ME 132, Spring 25, UC Berkeley, A Packard 187 21 Linear State-Space Representations First, let s describe the most general type of dynamic system that we will consider/encounter in this class Systems may

More information

Termination criteria for inexact fixed point methods

Termination criteria for inexact fixed point methods Termination criteria for inexact fixed point methods Philipp Birken 1 October 1, 2013 1 Institute of Mathematics, University of Kassel, Heinrich-Plett-Str. 40, D-34132 Kassel, Germany Department of Mathematics/Computer

More information

Lecture 8 : Eigenvalues and Eigenvectors

Lecture 8 : Eigenvalues and Eigenvectors CPS290: Algorithmic Foundations of Data Science February 24, 2017 Lecture 8 : Eigenvalues and Eigenvectors Lecturer: Kamesh Munagala Scribe: Kamesh Munagala Hermitian Matrices It is simpler to begin with

More information

Numerical Methods I Eigenvalue Problems

Numerical Methods I Eigenvalue Problems Numerical Methods I Eigenvalue Problems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 October 2nd, 2014 A. Donev (Courant Institute) Lecture

More information

Throughout these notes we assume V, W are finite dimensional inner product spaces over C.

Throughout these notes we assume V, W are finite dimensional inner product spaces over C. Math 342 - Linear Algebra II Notes Throughout these notes we assume V, W are finite dimensional inner product spaces over C 1 Upper Triangular Representation Proposition: Let T L(V ) There exists an orthonormal

More information

Efficient Augmented Lagrangian-type Preconditioning for the Oseen Problem using Grad-Div Stabilization

Efficient Augmented Lagrangian-type Preconditioning for the Oseen Problem using Grad-Div Stabilization Efficient Augmented Lagrangian-type Preconditioning for the Oseen Problem using Grad-Div Stabilization Timo Heister, Texas A&M University 2013-02-28 SIAM CSE 2 Setting Stationary, incompressible flow problems

More information

Matrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A =

Matrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A = 30 MATHEMATICS REVIEW G A.1.1 Matrices and Vectors Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A = a 11 a 12... a 1N a 21 a 22... a 2N...... a M1 a M2... a MN A matrix can

More information

The Singular Value Decomposition

The Singular Value Decomposition The Singular Value Decomposition Philippe B. Laval KSU Fall 2015 Philippe B. Laval (KSU) SVD Fall 2015 1 / 13 Review of Key Concepts We review some key definitions and results about matrices that will

More information

Eigenvalues and eigenvectors

Eigenvalues and eigenvectors Chapter 6 Eigenvalues and eigenvectors An eigenvalue of a square matrix represents the linear operator as a scaling of the associated eigenvector, and the action of certain matrices on general vectors

More information

1 Last time: least-squares problems

1 Last time: least-squares problems MATH Linear algebra (Fall 07) Lecture Last time: least-squares problems Definition. If A is an m n matrix and b R m, then a least-squares solution to the linear system Ax = b is a vector x R n such that

More information

VISUALIZING PSEUDOSPECTRA FOR POLYNOMIAL EIGENVALUE PROBLEMS. Adéla Klimentová *, Michael Šebek ** Czech Technical University in Prague

VISUALIZING PSEUDOSPECTRA FOR POLYNOMIAL EIGENVALUE PROBLEMS. Adéla Klimentová *, Michael Šebek ** Czech Technical University in Prague VSUALZNG PSEUDOSPECTRA FOR POLYNOMAL EGENVALUE PROBLEMS Adéla Klimentová *, Michael Šebek ** * Department of Control Engineering Czech Technical University in Prague ** nstitute of nformation Theory and

More information

Linear Algebra. The analysis of many models in the social sciences reduces to the study of systems of equations.

Linear Algebra. The analysis of many models in the social sciences reduces to the study of systems of equations. POLI 7 - Mathematical and Statistical Foundations Prof S Saiegh Fall Lecture Notes - Class 4 October 4, Linear Algebra The analysis of many models in the social sciences reduces to the study of systems

More information

NORMS ON SPACE OF MATRICES

NORMS ON SPACE OF MATRICES NORMS ON SPACE OF MATRICES. Operator Norms on Space of linear maps Let A be an n n real matrix and x 0 be a vector in R n. We would like to use the Picard iteration method to solve for the following system

More information

Operations On Networks Of Discrete And Generalized Conductors

Operations On Networks Of Discrete And Generalized Conductors Operations On Networks Of Discrete And Generalized Conductors Kevin Rosema e-mail: bigyouth@hardy.u.washington.edu 8/18/92 1 Introduction The most basic unit of transaction will be the discrete conductor.

More information

Linear Algebra March 16, 2019

Linear Algebra March 16, 2019 Linear Algebra March 16, 2019 2 Contents 0.1 Notation................................ 4 1 Systems of linear equations, and matrices 5 1.1 Systems of linear equations..................... 5 1.2 Augmented

More information

Bare-bones outline of eigenvalue theory and the Jordan canonical form

Bare-bones outline of eigenvalue theory and the Jordan canonical form Bare-bones outline of eigenvalue theory and the Jordan canonical form April 3, 2007 N.B.: You should also consult the text/class notes for worked examples. Let F be a field, let V be a finite-dimensional

More information

Math Ordinary Differential Equations

Math Ordinary Differential Equations Math 411 - Ordinary Differential Equations Review Notes - 1 1 - Basic Theory A first order ordinary differential equation has the form x = f(t, x) (11) Here x = dx/dt Given an initial data x(t 0 ) = x

More information

Draft. Lecture 01 Introduction & Matrix-Vector Multiplication. MATH 562 Numerical Analysis II. Songting Luo

Draft. Lecture 01 Introduction & Matrix-Vector Multiplication. MATH 562 Numerical Analysis II. Songting Luo Lecture 01 Introduction & Matrix-Vector Multiplication Songting Luo Department of Mathematics Iowa State University MATH 562 Numerical Analysis II Songting Luo ( Department of Mathematics Iowa State University[0.5in]

More information

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic Applied Mathematics 205 Unit V: Eigenvalue Problems Lecturer: Dr. David Knezevic Unit V: Eigenvalue Problems Chapter V.4: Krylov Subspace Methods 2 / 51 Krylov Subspace Methods In this chapter we give

More information

Problem set 2. Math 212b February 8, 2001 due Feb. 27

Problem set 2. Math 212b February 8, 2001 due Feb. 27 Problem set 2 Math 212b February 8, 2001 due Feb. 27 Contents 1 The L 2 Euler operator 1 2 Symplectic vector spaces. 2 2.1 Special kinds of subspaces....................... 3 2.2 Normal forms..............................

More information

ACM 104. Homework Set 5 Solutions. February 21, 2001

ACM 104. Homework Set 5 Solutions. February 21, 2001 ACM 04 Homework Set 5 Solutions February, 00 Franklin Chapter 4, Problem 4, page 0 Let A be an n n non-hermitian matrix Suppose that A has distinct eigenvalues λ,, λ n Show that A has the eigenvalues λ,,

More information

EIGENVALUE PROBLEMS. EIGENVALUE PROBLEMS p. 1/4

EIGENVALUE PROBLEMS. EIGENVALUE PROBLEMS p. 1/4 EIGENVALUE PROBLEMS EIGENVALUE PROBLEMS p. 1/4 EIGENVALUE PROBLEMS p. 2/4 Eigenvalues and eigenvectors Let A C n n. Suppose Ax = λx, x 0, then x is a (right) eigenvector of A, corresponding to the eigenvalue

More information

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH V. FABER, J. LIESEN, AND P. TICHÝ Abstract. Numerous algorithms in numerical linear algebra are based on the reduction of a given matrix

More information

Matrix Mathematics. Theory, Facts, and Formulas with Application to Linear Systems Theory. Dennis S. Bernstein

Matrix Mathematics. Theory, Facts, and Formulas with Application to Linear Systems Theory. Dennis S. Bernstein Matrix Mathematics Theory, Facts, and Formulas with Application to Linear Systems Theory Dennis S. Bernstein PRINCETON UNIVERSITY PRESS PRINCETON AND OXFORD Contents Special Symbols xv Conventions, Notation,

More information

Linear Algebra: Matrix Eigenvalue Problems

Linear Algebra: Matrix Eigenvalue Problems CHAPTER8 Linear Algebra: Matrix Eigenvalue Problems Chapter 8 p1 A matrix eigenvalue problem considers the vector equation (1) Ax = λx. 8.0 Linear Algebra: Matrix Eigenvalue Problems Here A is a given

More information

Fundamentals of Linear Algebra. Marcel B. Finan Arkansas Tech University c All Rights Reserved

Fundamentals of Linear Algebra. Marcel B. Finan Arkansas Tech University c All Rights Reserved Fundamentals of Linear Algebra Marcel B. Finan Arkansas Tech University c All Rights Reserved 2 PREFACE Linear algebra has evolved as a branch of mathematics with wide range of applications to the natural

More information

Preconditioners for the incompressible Navier Stokes equations

Preconditioners for the incompressible Navier Stokes equations Preconditioners for the incompressible Navier Stokes equations C. Vuik M. ur Rehman A. Segal Delft Institute of Applied Mathematics, TU Delft, The Netherlands SIAM Conference on Computational Science and

More information

Stat 159/259: Linear Algebra Notes

Stat 159/259: Linear Algebra Notes Stat 159/259: Linear Algebra Notes Jarrod Millman November 16, 2015 Abstract These notes assume you ve taken a semester of undergraduate linear algebra. In particular, I assume you are familiar with the

More information

NONCOMMUTATIVE POLYNOMIAL EQUATIONS. Edward S. Letzter. Introduction

NONCOMMUTATIVE POLYNOMIAL EQUATIONS. Edward S. Letzter. Introduction NONCOMMUTATIVE POLYNOMIAL EQUATIONS Edward S Letzter Introduction My aim in these notes is twofold: First, to briefly review some linear algebra Second, to provide you with some new tools and techniques

More information

EECS 275 Matrix Computation

EECS 275 Matrix Computation EECS 275 Matrix Computation Ming-Hsuan Yang Electrical Engineering and Computer Science University of California at Merced Merced, CA 95344 http://faculty.ucmerced.edu/mhyang Lecture 16 1 / 21 Overview

More information

Final A. Problem Points Score Total 100. Math115A Nadja Hempel 03/23/2017

Final A. Problem Points Score Total 100. Math115A Nadja Hempel 03/23/2017 Final A Math115A Nadja Hempel 03/23/2017 nadja@math.ucla.edu Name: UID: Problem Points Score 1 10 2 20 3 5 4 5 5 9 6 5 7 7 8 13 9 16 10 10 Total 100 1 2 Exercise 1. (10pt) Let T : V V be a linear transformation.

More information

Here is an example of a block diagonal matrix with Jordan Blocks on the diagonal: J

Here is an example of a block diagonal matrix with Jordan Blocks on the diagonal: J Class Notes 4: THE SPECTRAL RADIUS, NORM CONVERGENCE AND SOR. Math 639d Due Date: Feb. 7 (updated: February 5, 2018) In the first part of this week s reading, we will prove Theorem 2 of the previous class.

More information

Applied Mathematics 205. Unit II: Numerical Linear Algebra. Lecturer: Dr. David Knezevic

Applied Mathematics 205. Unit II: Numerical Linear Algebra. Lecturer: Dr. David Knezevic Applied Mathematics 205 Unit II: Numerical Linear Algebra Lecturer: Dr. David Knezevic Unit II: Numerical Linear Algebra Chapter II.3: QR Factorization, SVD 2 / 66 QR Factorization 3 / 66 QR Factorization

More information

Solving the steady state diffusion equation with uncertainty Final Presentation

Solving the steady state diffusion equation with uncertainty Final Presentation Solving the steady state diffusion equation with uncertainty Final Presentation Virginia Forstall vhfors@gmail.com Advisor: Howard Elman elman@cs.umd.edu Department of Computer Science May 6, 2012 Problem

More information

Lecture 11. Linear systems: Cholesky method. Eigensystems: Terminology. Jacobi transformations QR transformation

Lecture 11. Linear systems: Cholesky method. Eigensystems: Terminology. Jacobi transformations QR transformation Lecture Cholesky method QR decomposition Terminology Linear systems: Eigensystems: Jacobi transformations QR transformation Cholesky method: For a symmetric positive definite matrix, one can do an LU decomposition

More information

EECS 275 Matrix Computation

EECS 275 Matrix Computation EECS 275 Matrix Computation Ming-Hsuan Yang Electrical Engineering and Computer Science University of California at Merced Merced, CA 95344 http://faculty.ucmerced.edu/mhyang Lecture 17 1 / 26 Overview

More information

The Jordan Canonical Form

The Jordan Canonical Form The Jordan Canonical Form The Jordan canonical form describes the structure of an arbitrary linear transformation on a finite-dimensional vector space over an algebraically closed field. Here we develop

More information

Lecture notes: Applied linear algebra Part 1. Version 2

Lecture notes: Applied linear algebra Part 1. Version 2 Lecture notes: Applied linear algebra Part 1. Version 2 Michael Karow Berlin University of Technology karow@math.tu-berlin.de October 2, 2008 1 Notation, basic notions and facts 1.1 Subspaces, range and

More information

12. Perturbed Matrices

12. Perturbed Matrices MAT334 : Applied Linear Algebra Mike Newman, winter 208 2. Perturbed Matrices motivation We want to solve a system Ax = b in a context where A and b are not known exactly. There might be experimental errors,

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra)

AMS526: Numerical Analysis I (Numerical Linear Algebra) AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 1: Course Overview & Matrix-Vector Multiplication Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 20 Outline 1 Course

More information

University of Maryland Department of Computer Science TR-5009 University of Maryland Institute for Advanced Computer Studies TR April 2012

University of Maryland Department of Computer Science TR-5009 University of Maryland Institute for Advanced Computer Studies TR April 2012 University of Maryland Department of Computer Science TR-5009 University of Maryland Institute for Advanced Computer Studies TR-202-07 April 202 LYAPUNOV INVERSE ITERATION FOR COMPUTING A FEW RIGHTMOST

More information

Fast Iterative Solution of Saddle Point Problems

Fast Iterative Solution of Saddle Point Problems Michele Benzi Department of Mathematics and Computer Science Emory University Atlanta, GA Acknowledgments NSF (Computational Mathematics) Maxim Olshanskii (Mech-Math, Moscow State U.) Zhen Wang (PhD student,

More information

MATH 5640: Functions of Diagonalizable Matrices

MATH 5640: Functions of Diagonalizable Matrices MATH 5640: Functions of Diagonalizable Matrices Hung Phan, UMass Lowell November 27, 208 Spectral theorem for diagonalizable matrices Definition Let V = X Y Every v V is uniquely decomposed as u = x +

More information

6 EIGENVALUES AND EIGENVECTORS

6 EIGENVALUES AND EIGENVECTORS 6 EIGENVALUES AND EIGENVECTORS INTRODUCTION TO EIGENVALUES 61 Linear equations Ax = b come from steady state problems Eigenvalues have their greatest importance in dynamic problems The solution of du/dt

More information

Spectral Processing. Misha Kazhdan

Spectral Processing. Misha Kazhdan Spectral Processing Misha Kazhdan [Taubin, 1995] A Signal Processing Approach to Fair Surface Design [Desbrun, et al., 1999] Implicit Fairing of Arbitrary Meshes [Vallet and Levy, 2008] Spectral Geometry

More information

ON A GENERAL CLASS OF PRECONDITIONERS FOR NONSYMMETRIC GENERALIZED SADDLE POINT PROBLEMS

ON A GENERAL CLASS OF PRECONDITIONERS FOR NONSYMMETRIC GENERALIZED SADDLE POINT PROBLEMS U..B. Sci. Bull., Series A, Vol. 78, Iss. 4, 06 ISSN 3-707 ON A GENERAL CLASS OF RECONDIIONERS FOR NONSYMMERIC GENERALIZED SADDLE OIN ROBLE Fatemeh anjeh Ali BEIK his paper deals with applying a class

More information

Notes on SU(3) and the Quark Model

Notes on SU(3) and the Quark Model Notes on SU() and the Quark Model Contents. SU() and the Quark Model. Raising and Lowering Operators: The Weight Diagram 4.. Triangular Weight Diagrams (I) 6.. Triangular Weight Diagrams (II) 8.. Hexagonal

More information

14 Singular Value Decomposition

14 Singular Value Decomposition 14 Singular Value Decomposition For any high-dimensional data analysis, one s first thought should often be: can I use an SVD? The singular value decomposition is an invaluable analysis tool for dealing

More information

Physics 202 Laboratory 5. Linear Algebra 1. Laboratory 5. Physics 202 Laboratory

Physics 202 Laboratory 5. Linear Algebra 1. Laboratory 5. Physics 202 Laboratory Physics 202 Laboratory 5 Linear Algebra Laboratory 5 Physics 202 Laboratory We close our whirlwind tour of numerical methods by advertising some elements of (numerical) linear algebra. There are three

More information

Linear Algebra (part 1) : Vector Spaces (by Evan Dummit, 2017, v. 1.07) 1.1 The Formal Denition of a Vector Space

Linear Algebra (part 1) : Vector Spaces (by Evan Dummit, 2017, v. 1.07) 1.1 The Formal Denition of a Vector Space Linear Algebra (part 1) : Vector Spaces (by Evan Dummit, 2017, v. 1.07) Contents 1 Vector Spaces 1 1.1 The Formal Denition of a Vector Space.................................. 1 1.2 Subspaces...................................................

More information

Lecture 4 Eigenvalue problems

Lecture 4 Eigenvalue problems Lecture 4 Eigenvalue problems Weinan E 1,2 and Tiejun Li 2 1 Department of Mathematics, Princeton University, weinan@princeton.edu 2 School of Mathematical Sciences, Peking University, tieli@pku.edu.cn

More information

Putzer s Algorithm. Norman Lebovitz. September 8, 2016

Putzer s Algorithm. Norman Lebovitz. September 8, 2016 Putzer s Algorithm Norman Lebovitz September 8, 2016 1 Putzer s algorithm The differential equation dx = Ax, (1) dt where A is an n n matrix of constants, possesses the fundamental matrix solution exp(at),

More information

Elementary linear algebra

Elementary linear algebra Chapter 1 Elementary linear algebra 1.1 Vector spaces Vector spaces owe their importance to the fact that so many models arising in the solutions of specific problems turn out to be vector spaces. The

More information

Lecture notes: Applied linear algebra Part 2. Version 1

Lecture notes: Applied linear algebra Part 2. Version 1 Lecture notes: Applied linear algebra Part 2. Version 1 Michael Karow Berlin University of Technology karow@math.tu-berlin.de October 2, 2008 First, some exercises: xercise 0.1 (2 Points) Another least

More information

Spectral Theorem for Self-adjoint Linear Operators

Spectral Theorem for Self-adjoint Linear Operators Notes for the undergraduate lecture by David Adams. (These are the notes I would write if I was teaching a course on this topic. I have included more material than I will cover in the 45 minute lecture;

More information

MATH 315 Linear Algebra Homework #1 Assigned: August 20, 2018

MATH 315 Linear Algebra Homework #1 Assigned: August 20, 2018 Homework #1 Assigned: August 20, 2018 Review the following subjects involving systems of equations and matrices from Calculus II. Linear systems of equations Converting systems to matrix form Pivot entry

More information

Lecture notes on Quantum Computing. Chapter 1 Mathematical Background

Lecture notes on Quantum Computing. Chapter 1 Mathematical Background Lecture notes on Quantum Computing Chapter 1 Mathematical Background Vector states of a quantum system with n physical states are represented by unique vectors in C n, the set of n 1 column vectors 1 For

More information

Review of some mathematical tools

Review of some mathematical tools MATHEMATICAL FOUNDATIONS OF SIGNAL PROCESSING Fall 2016 Benjamín Béjar Haro, Mihailo Kolundžija, Reza Parhizkar, Adam Scholefield Teaching assistants: Golnoosh Elhami, Hanjie Pan Review of some mathematical

More information