Analysis of Transient Behavior in Linear Differential Algebraic Equations
|
|
- Edmund Fowler
- 5 years ago
- Views:
Transcription
1 Analysis of Transient Behavior in Linear Differential Algebraic Equations Dr. Mark Embree and Blake Keeler May 6, 5 Abstract The linearized Navier-Stokes equations form a type of system called a differential algebraic equation (DAE). These equations impose both algebraic and differential constraints on the solution, and present significant challenges to solving. This paper discusses a method for solving DAEs by the Schur factorization to avoid some of the obstacles that are typically present. We also investigate the transient behavior of the solution using results from pseudospectral theory, which will help us understand whether the solution to the linearized Navier-Stokes system will become turbulent if its steady state is disturbed. If a flow does indeed grow substantially in energy, then we must be cautious about using the linearized model. Thus, having information about transient behavior is quite instructive. Introduction The Navier-Stokes equations are system of highly nonlinear PDEs that describe how the velocity and pressure fields of a fluid behave. Given a particular fluid flow problem, one often seeks to understand whether that flow will be stable (laminar) or unstable (turbulent) if it is perturbed from its steady state. One of the most common ways to answer this question is to perform a finite element discretization in space, resulting in a system of linear ODEs in time. That system of ODEs turns out to be a linear differential algebraic equation (DAE), a system in which the coefficient matrix of the derivative vector is singular. This makes solving these equations much more complex, as the generalized eigenvalue problem becomes more difficult to interpret and connect to the question of stability. In particular, there will be infinite eigenvalues that must be dealt with. We will see that these infinite eigenvalues impose algebraic constraints on the solution, forcing it to evolve on a lower dimensional subspace of R n. Hence the name differential algebraic equation. This paper uses a method by which these equations may be solved without any direct need for the generalized eigenvalue problem by surgically removing the infinite eigenvalues from the system without breaking the dynamics, resulting in a physically meaningful solution in exponential form. We investigate how to analyze the transient behavior of the solution. The asymptotic properties of the solutions to differential equations are well-understood. It is commonly known that if a solution to a system of differential equations x (t) = Ax(t) has the form
2 e ta x(), then the eigenvalues of A are the determining factors in the behavior of the solution as t. However, this tells us nothing of what happens between the initial state and the inevitable decay at infinity. It is often possible that the solution experiences growth by several orders of magnitude before settling down and going to zero. This behavior is referred to as transient growth and is a consequence of the departure of the matrix A from normality. One manifestation of this phenomenon is eigenvectors of A associated with distinct eigenvalues being nearly parallel. Unfortunately, the amount of information gained from inspecting the eigenvectors alone is quite limited. They will not tell us how much growth will occur, only that such growth is likely. We can, however, obtain such information by looking not at the spectrum of the matrix, but at the pseudospectra instead. The second portion of this paper applies results from pseudospectral theory to analyze transient behavior in the Navier-Stokes problem. Differential Algebraic Equations Systems of linear differential equations of the form Bx (t) = Ax(t) (.) are commonplace in many applications, and are quite easy to solve when B is invertible. We can simply invert B and use the matrix exponential to find x(t) = e BAt x(). For more practical formulations, one can seek the generalized eigenvalues and eigenvectors of the matrix pencil (A, B) and see that any function of the form x(t) = e µt v is a solution to the differential equation for any eigenpair (µ, v). However, this paper is primarily concerned with the more interesting case where B is singular, making the system a DAE. Problems of this type are much more difficult to solve, since the pencil (A, B) now possesses infinite eigenvalues. This can be seen by writing down the generalized eigenvalue problem as Av = Bv (.) µ and noting that since B is singular, there exist nonzero vectors v for which Bv =. For these v, taking µ = satisfies the eigenvalue problem, making (µ, v) an infinite eigenpair of the matrix pencil. So how shall we handle such an equation? We clearly cannot write x(t) = e t v as a solution. The following section details a method by which we may construct a sensible solution to (.) using the Schur decomposition, thereby avoiding the generalized eigenvalue problem entirely.. Solving a DAE by the Schur Decomposition Details about the solution of DAEs can be found in [. In this section we discuss one particular method that employs the Schur decomposition. Consider (.) where B is singular, but assume that (A, B) is a nonsingular pencil, so there exists some real number λ such that A λb is invertible. We define A λ = (A λb) B and B λ = (A λb) B. Then if we multiply (.) by (A λb) we obtain B λ x (t) = A λ x(t). (.3)
3 Since (A λb) is nonsingular, (.3) is equivalent to (.). Notice also that A λ λb λ = (A λb) A λ(a λb) B = (A λb) (A λb) = I. Therefore, A λ = I + λb λ, so (.3) can be written as Now we write the Schur decomposition of B λ as B λ x (t) = (I + λb λ )x(t). (.4) B λ = QTQ, where Q is a unitary matrix, and T is an upper triangular matrix in which the eigenvalues of B λ may appear in any order. Suppose the pencil has an infinite eigenvalue of algebraic multiplicity m, so that B λ has a zero eigenvalue of algebraic multiplicity m. Then it is convenient to order the eigenvalues so that [ G S T =, N where N is an m m upper triangular matrix corresponding to the zero eigenvalues of B λ (i.e. its diagonal elements are all zero), G is the (n m) (n m) upper triangular matrix corresponding to the nonzero eigenvalues (so G is invertible), and U is some (n m) m matrix. Hence, if we premultiply (.3) by Q, we obtain which can be rewritten as [ G S N Q B λ QQ x (t) = Q (I + λb λ )QQ x(t), (.5) [ I + λg λs Q x (t) = I + λn Q x(t). (.6) Define y(t) = Q x(t), where y(t) = [y, y T with y of length n m and y of length m, so we have [ [ [ [ G S y I + λg λs y =. (.7) N I + λn y From the second block-row, we have Ny = (I + λn)y. Since the diagonal of N is zero, it must be a nilpotent matrix. Let us assume that it has index k, that is N k =, but N k. Therefore, premultiplying by N k we find (N k + λn k )y = N k y =. So N k y =. Then if we premultiply by N k we have (N k + λn k )y = N k y = (N k y ) =. So N k y =. This process can be continued inductively to obtain N y =. 3 y
4 Therefore, y =. Let us consider for a moment what this implies about the set of valid initial conditions x(). Let us first partition Q = [Q, Q so that Q R n (n m) and Q R n m. Then since x(t) = Qy(t) with y (t) =, we have that x() = Q y (). (.8) In other words, the only initial conditions that satisfy the algebraic constraints of the DAE are those in Ran(Q ). Similarly, the relationship x(t) = Q y (t) implies that the solution evolves on a subspace of R n of dimension n m. Hence, the infinite eigenvalues of the pencil (A, B) serve to constrain the solution to a subspace of lower dimension. We will see a very clear example of this in section 3.. Continuing from (.7) with the knowledge that y (t) =, we have that Gy = (I + λg)y, and so the solution is y (t) = e (G +λi)t y (). The full solution to (.7) can now be written as [ [ y (t) e (G +λi)t = y (t) [ y (). (.9) Notice that the entries in the (,) and (,) blocks of the above matrix are irrelevant, but makes a convenient choice. We can then write [ y (t) x(t) = Q y (t) ( [ ) [ G = Q exp t + λi Q y () Q. For simplicity, we will write the solution to (.) as ( [ M x(t) = exp tq Q ) x(), (.) where M = G + λi. Thus, we have formulated a solution to (.) that maintains standard exponential form, despite the fact that the pencil (A, B) had infinite eigenvalues. The next few sections will outline some of the interesting properties of this solution.. The Eigenvalues of M As mentioned previously, if we have a standard system of equations like (.) with B nonsingular, we can simply write the solution in terms of the eigenvalues of the pencil (A, B). For a DAE this is complicated by the existence of infinite eigenvalues, but it is reasonable to expect that the finite eigenvalues of the pencil will still dictate the behavior of the solution. In the following, we prove that this is indeed the case, as the eigenvalues of M in (.) are the same as the finite eigenvalues of the pencil. 4
5 Theorem.. If A, B, M, and λ are as in section., then the spectrum of M is the set of all finite eigenvalues of the pencil (A, B). Proof. Suppose η is a finite eigenvalue of the pencil (A, B). nonzero x. So we can subtract λbx to obtain Then Ax = ηbx for some (A λb)x = (η λ)bx. Since A λb is invertible, η λ must be nonzero. We can rearrange the above to obtain (A λb) Bx = η λ x. Hence is a nonzero eigenvalue of B η λ λ = (A λb) B. If the Schur decomposition of B λ is as in section., then is an eigenvalue of G. Hence, η λ is an eigenvalue of η λ G. Finally, this implies that η is an eigenvalue of M = G + λi. The reverse inclusion follows the same argument, and so the theorem follows. This result, while it is what we had hoped for, is interesting in that no matter what shift λ is applied to make A λb invertible, the key matrix in the solution always has the same eigenvalues. In fact, as we shall see, the solution is entirely independent of λ..3 The Solution s Independence of λ In this section, we will prove that our solution (.) is independent of the parameter λ, whose only purpose was to make the pencil (A, B) nonsingular. We begin with the following lemmas. Lemma.. Let A, B C n n and suppose λ and µ are any real numbers such that A λb and A µb are invertible. Define B λ = (A λb) B and B µ = (A µb) B. Then η σ(b λ ) = η + (λ µ)η σ(b µ). Proof. Suppose η σ(b λ ). Then (A λb) Bx = ηx for some nonzero vector x. Premultiplying by (A λb) and adding and subtracting ηµbx from the right gives A rearrangement yields Bx = η(a µb)x + η(µ λ)bx. ( + η(λ µ))bx = η(a µb)x. If η, then since A µb is invertible, + η(λ µ) must be nonzero. If η =, then clearly + η(λ µ). Therefore, we can write Bx = η (A µb)x. + η(λ µ) Premultiplying by (A µb) demonstrates the result. 5
6 Lemma.3. Let A, B, B λ and B µ be as in Lemma. and suppose the pencil (A, B) has an infinite eigenvalue of algebraic multiplicity m, so that B λ and B µ each have a zero eigenvalue of algebraic multiplicity m. Let the Schur decompositions of B λ and B µ be given by [ Gλ S B λ = Q λ N λ Q [ Gµ S B µ = U µ N µ U, respectively, where Q, U are unitary, N λ, N µ R m m have zero diagonal entries, and G λ, G µ R (n m) (n m) are invertible. Partition Q = [Q, Q and U = [U, U so that Q, U R n (n m). Then Ran(Q ) = Ran(U ). Proof. Let us write B λ in Jordan form as B λ X = XJ where the nonzero eigenvalues are written in the upper left (n m) (n m) block of J. If we partition X = [X, X in the same way that we partitioned Q, then Ran(X ) is the invariant subspace of B λ associated with the nonzero eigenvalues. Observe that Ran(Q ) is also an invariant subspace of B λ associated with the nonzero eigenvalues. Hence Ran(X ) = Ran(Q ). Next we can premultiply B λ X = XJ by A λb and add and subtract µbxj on the right to obtain Simple rearrangement yields BX = (A µb)xj + (µ λ)bxj. BX(I + (λ µ)j) = (A µb)xj. For µ sufficiently close to λ, I + (λ µ)j is invertible, so we can write (A µb) BX = XJ(I + (λ µ)j). (.) It is important to note that J(I + (λ µ)j) will retain the block upper triangular form of J. This is shown by partitioning J by its Jordan blocks as J = diag{j,..., J p } so that J (I + (λ µ)j ) J(I + (λ µ)j) =..., J p (I p + (λ µ)j p ) where I,..., I p are the appropriate diagonal partitions of the identity matrix. Therefore, the diagonal of J(I + (λ µ)j) will retain the structure of the diagonal of J. That is, the nonzero eigenvalues (of B µ this time) will be listed as the first elements along the diagonal. Hence by (.) X is the invariant subspace associated with the nonzero eigenvalues of B µ. Again, since such invariant subspaces are unique, we have that Ran(U ) = Ran(X ). When we recall that Ran(X ) = Ran(Q ), we have the desired result. Thus far, we have only proved the theorem for µ sufficiently close to λ, since that condition guaranteed that I + (λ µ)j would be invertible. However, we claim that this matrix is invertible for all valid choices of λ and µ. To see this, note that I+(λ µ)j will be invertible provided + (λ µ)η is nonzero for all eigenvalues η of B λ. If η σ(b λ ), Lemma. states that η/( + (λ µ)η) is an eigenvalue of B µ. Since the eigenvalues of B µ must be finite, + (λ µ)η must be nonzero. The result therefore generalizes to all λ and µ for which A λb and A µb are invertible. 6
7 Lemma.4. If all matrices are as in Lemma.3, [ Q Ĝµ Ŝ B µ Q = µ Nµ, where Ĝµ = Q U G µ U Q. Proof. Direct computation shows that [ Q Q B µ Q = Q B µ [ Q Q = Q B µ Q Q B µ Q Q B µ Q Q B µ Q [ Ĝ =: µ Ŝ µ. Q B µ Q Nµ By Lemma.3 we know that Ran(Q ) = Ran(U ) is an invariant subspace of B µ, so we have Q B µ Q = Q Q Z =, for some matrix Z R (n m) (n m), since the columns of Q are orthogonal to the columns of Q. Therefore, [ Q Ĝµ Ŝ B µ Q = µ. Nµ Also, if we expand the Schur form of B µ, we obtain and hence, B µ = U G µ U + (U S µ + U N µ )U, Ĝ µ = Q B µ Q = Q U G µ U Q, since the columns of Q are orthogonal to the columns of U. We have therefore proven that Q B µ Q remains upper triangular with an explicit formula for the upper left block. Lemma.5. B λ B µ = (λ µ)b λ B µ. Proof. Notice that (A µb) (A λb) = (λ µ)b. If we premultiply by (A λb) and postmultiply by (A µb), we obtain (A λb) (A µb) = (λ µ)(a λb) B(A µb), which is a generalized form of the First Resolvent Identity. If we then postmultiply this equation by B, we have the desired result. 7
8 This is the final lemma that we need in order to prove the following theorem. Theorem.6. The solution formulated in section. is independent of the choice of the parameter λ, provided A λb is nonsingular. That is, in the context of the preceding lemmas, [ [ G µ + µi G U U λ = Q + λi Q. Proof. By Lemma.5, we have B λ B µ = (λ µ)b λ B µ. Multiplying by Q on the left and Q on the right, we obtain Q B λ Q + Q B µ Q = (λ µ)q B λ QQ B µ Q. If we replace B λ and B µ by their Schur decompositions and invoke Lemma.4, we can write [ [ [ [ Gλ S λ Ĝµ Ŝ µ Gλ S = (λ µ) λ Ĝ µ Ŝ µ. N λ Nµ N λ Nµ Restricting our attention to the (,) block of this matrix equation, we see that G λ Ĝµ = (λ µ)g λ Ĝ µ. Since Ĝµ = Q U G µ U Q, we can multiply on the left by Q and on the right by Q to obtain Q G λ Q (Q Q U )G µ (U Q Q ) = (λ µ)(q G λ Q )U G µ (U Q Q ). (.) Since Q Q and U U are both orthogonal projectors onto Ran(Q ) = Ran(U ), we have that Q Q = U U. So we have and similarly Q Q U = U U U = U U U Q = Q Q Q = Q. We will use these identities freely in the following calculations. Continuing from (.) we have Q G λ Q U G µ U = (λ µ)(q G λ Q )(U G µ U ). Then if we multiply on the left by Q G λ Q, we obtain Q Q (Q G λ Q )(U G µ U ) = (λ µ)u G µ U. Next we can multiply on the right by U G µ U, which gives U G µ U Q G λ Q = (λ µ)u U. 8
9 Since U U = Q Q, we can rearrange the above to see that U (G µ + µi)u = Q (G λ + λi)q. It follows that [ G µ + µi U [ G U λ = Q + λi Q. Now that we know the solution will not be affected by our choice of the parameter λ, we can be confident in using this approach to solve DAEs..4 Linearized Navier-Stokes As mentioned in the introduction, the primary application that we wish to study is the numerical solution of the incompressible Navier-Stokes equations, ν u + u u + p = u t (.3) u =, (.4) where u is the velocity field of the fluid and p is the pressure field. Equations (.3) and (.4) form a system of nonlinear time-dependent partial differential equations, for which finding a solution is a difficult task. One common approach is to find a steady state solution and linearize the system around that steady state using a finite element formulation, the details of which we will not discuss here. For a full description of this process, see [3. The key idea is to approximate u u h and p p h where u h = n a i (t)φ i (x) and m b j (t)ψ j (x) for some finite collections of basis functions {φ,..., φ n } and {ψ,..., ψ m }. With this discretization, the linearized problem is reduced to a first order system of linear differential equations in time, given by [ A i= [ K C x (t) = C j= x(t), (.5) [ a(t) where x(t) = with a(t) = [a b(t) (t),..., a n (t) and b(t) = [b (t),..., b m (t), the vectors of time-dependent coefficients in u h and p h. Also, A C n n is Hermitian positive definite, C C n m has full column rank (m n), and K C n n. From this we can see the connection to DAEs, since the matrix on the left side of (.5) is singular, as it has m zero columns. It should be noted that (.5) describes how a fluid flow system will behave in time, but only under the assumption that it stays sufficiently close to the steady state. Therefore, the transient properties of the system are very important. If we can conclude that the solution will likely grow substantially from its initial state, then the necessary assumptions for the linearized analysis will not be satisfied. For more details about linear stability analysis, see [ and [8. The next sections describe the necessary tools to analyze how the pseudospectral properties of the matrix M in the solution to (.5) influence transient behavior, allowing us to predict whether or not the linearization will remain descriptive. 9
10 3 Pseudospectra As we turn our attention to transient growth in the solutions to DAEs, we will find the concept of pseudospectra to be quite useful. Conceptually, the pseudospectra of a matrix A are sets of numbers in the complex plane that are almost eigenvalues. Before we introduce the more mathematically precise definition, notice that if z is not an eigenvalue of A (z / σ(a)), the resolvent norm (A zi) is defined and will grow in magnitude as A zi gets closer to being singular, which is generally when z approaches an eigenvalue of A. Therefore, it is reasonable to take the convention (A zi) = if z σ(a). This motivates the definition of pseudospectra that we will use here, taken from [6, which states that if ε > is arbitrary and A C n n, then the ε-pseudospectrum σ ε (A) of A is the set of z C such that (A zi) > ε. We will not look too deeply into the theory of pseudospectra in this paper; we will simply use a few results that will prove to be useful in studying the transient behavior of our DAE solutions, as discussed in the next section. 3. Pseudospectra and Transient Growth Recall that our solution to a general DAE in section had the form ( [ ) M x(t) = exp tq Q x(). Therefore, since x() is constant, the transient behavior of the solution is determined primarily by the behavior of the matrix exponential as time increases. Specifically, we only need to know about e tm since Q and Q simply make an orthogonal similarity transformation, under which the -norm is invariant. This is where some results from pseudospectral theory prove quite useful. In particular, we can use pseudospectra to put a lower bound on the maximum value of the matrix exponential. From Theorem 5.4 in [6, we have that if the pseudospectral abscissa, α ε (A) is defined by α ε (A) = sup Re(z), (3.) z σ ε (A) which in simple terms is the real part of the rightmost point in the ε-pseudospectrum, then Taking the supremum over ε gives sup e ta α ε (A)/ε ε >. (3.) t sup t e ta K(A) = sup α ε (A)/ε, (3.3) ε>
11 where K(A) is known as the Kreiss constant. Therefore, if the ε-pseudospectrum of A extends farther than ε into the right-half plane, we can guarantee that e ta must grow by at least a factor of α ε (A)/ε from the initial state. It is important to note that this only guarantees growth in the same norm that defines the pseudospectra. Determining an effective method for computing pseudospectra in the appropriate norm will be discussed in section 3.3. Also, notice that we can only guarantee such growth in the matrix exponential e ta. If the solution to a DAE has the form x(t) = e tm x(), then we have only that sup t x(t) sup e tm x(). (3.4) t Hence, we need not see transient growth for all initial conditions x(). However, by definition of an operator norm, e tm = max e tm x. x = Therefore, there does exist an initial state x() such that maximal transient growth at time t is realized. The next section details a simple three-dimensional example where we see a clear connection between the Kreiss constant and the growth of the solution. 3. A Three-Dimensional Example Consider the dynamical system where B = Bx (t) = Ax(t), (3.5) and A = The matrix A was chosen arbitrarily to simply be nonsingular. B, on the other hand, was specifically constructed so that it has a zero eigenvalue of algebraic multiplicity one. When we follow the procedure outlined in section. with any appropriate λ, we obtain where Q = x(t) = Q exp t Q x(), (3.6) and x() must be in the range of the first two columns of Q, by (.8). A suitable choice for the initial condition was x() = Notice that in the context of (.), M = [
12 The eigenvectors of this matrix are [, and [, /, which are close to parallel, so from the standard qualitative perspective we should expect transient growth. The pseudospectra will reflect this and give us a more quantitative idea of exactly how much growth to expect. Before we look at the pseudospectra, however, recall that since B (and therefore B λ ) had a zero eigenvalue of multiplicity one, we should expect the solution to evolve on a twodimensional subspace of R 3. And indeed, as shown in figure 3. we see that the solution is constrained to evolve on a plane. z Initial Point Final Point y 5 5 x Figure 3.: The path x(t) and the D subspace on which it lives Figure 3.: Pseudospectra of M We see also that there is some transient growth in x(t), so this should be reflected in the pseudospectral properties of M. Based on our discussion in section 3., we should expect
13 the ε-pseudospectra to extend farther than ε into the right half-plane for some ε >. As shown in figure 3., this is exactly what we see. The black dots show the eigenvalues of M, which are both in the left half-plane, which is sufficient to guarantee that the solution is asymptotically stable. However, the fact that the pseudospectra, denoted by the colored contours, extend so far to the right of the imaginary axis gives intermediate growth before the decay in the solution at infinity. To understand just how much growth we should expect in x(t), we compute the ratios α ε (M)/ε for various ε between. and to find a close approximation to the Kreiss constant. Figure 3.3 demonstrates the relationship between ε and the ratios α ε (M)/ε. The peak that is highlighted in the figure occurs at ε =.398 and takes on a value of α ε (M)/ε = 3.488, which gives us that K(M) Hence, we should expect the norm of x(t) to grow by at least a factor of from its initial value before its final decay. This behavior is demonstrated in figure 3.3 where the red line indicates our predicted level of growth. We see that the norm grows to a point just past our prediction before quickly shrinking to its steady state at zero αε(m) ε ε Figure 3.3: α ε (M)/ε as a function of ε. x(t) αε ε x t Figure 3.4: Transient Behavior of x(t). 3
14 This example provides a clear demonstration of the relationship between the pseudospectra of M and the transient behavior of the solution norm. However, it is worth noting that we have simply defaulted to using the -norm to compute both the pseudospectra and the growth of the solution. In this case, it made sense because we were only interested in the path of the particle, but in the case of Navier-Stokes it is not so obvious how growth should be described. We discuss this issue in the following section. 3.3 Finding the Appropriate Norm We are interested [ in transient behavior in the solution to the Navier-Stokes problem, which u takes the form, where u and p are the velocity and pressure fields, respectively. The p task at hand is to determine a meaningful way to describe what we mean by growth. For example, if the -norm of the velocity-pressure vectors exhibits transient growth, does that imply that there will be any noticeable behavior in the physical system? In general, it is difficult to assign any physical meaning to such a norm. However, one norm that proves quite useful in this case is the energy-norm, denoted E. If we let Ω denote the physical domain of the problem, then the energy norm for Navier-Stokes is given by [ u p E = ( u H + ) / p L = Ω u u + p dω where H is the H semi-norm and L is the L norm. Suppose now that /, [ uh a finite element approximation to the solution, with u h = n a i φ i (x) and v h = m b i ψ i (x), where {φ,..., φ n } and {ψ,..., ψ m } are the finite element basis functions. Note that any boundary conditions enforced will vary with the particular flow being studied, but this does not affect our calculations. Computing the energy norm of our finite element approximation gives [ u p E [ uh p h E = = Ω n i= u h u h + p hdω n a i a j j= Ω φ i φ j dω + i= m i= m b i b j j= Ω i= p h ψ i (x)ψ j (x)dω. is 4
15 If we let a = [a,..., a n, b = [b,..., b m, (K) ij = φ Ω i φ j dω, and (A) ij = ψ Ω i(x)ψ j (x)dω, we can write [ u [ uh = a Ka + b Ab p E p h E = [ a b = x Hx. [ K A [ a b [ [ K a where H = and x =. Notice that under the typical restriction that φ A b i and ψ i obey Dirichlet boundary conditions, both K and A are Gram matrices and therefore are symmetric positive definite. Hence, H is symmetric positive definite, and so the function u, v H = v Hu defines an inner product that induces the corresponding norm u H = (u Hu) /. Therefore, we have that [ [ u uh p = x H. E Coincidentally, K and A are the same as in (.4), so it requires no additional effort to construct the appropriate H. Additionally, if RR = H is a Cholesky factorization of H (which must exist since H is symmetric positive definite), then x H = x Hx = x RR x = (R x) R x = R x. Therefore, we can compute the energy norm of our finite element solution by simply computing the -norm of the vector of finite element coefficients, x, transformed by R. Similarly, if we wish to compute the energy norm of a matrix B, we can first note that p h E By E y E = y B HBy y Hy = (R By) R By (R y) Ry = R By R y = R BR z z where z = R y. Since R is invertible, maximizing over y is equivalent to maximizing over z, and therefore B E = R BR. (3.7) Thus, we can compute the energy norm of a linear operator by performing the similarity transformation with respect to R and then computing the -norm. 5
16 3.4 Results for Navier-Stokes We have finally developed all of the necessary tools to analyze the transient behavior of the linearized Navier-Stokes problem given by [ [ A K C x (t) = C x(t) (3.8) Before we compute any solutions to this system, we perform the change of coordinates as described in section 3.3, resulting in the modified system [ [ A K C R R y (t) = R C R y(t), (3.9) where y(t) = R x(t). Now when we compute the -norm of the solution to (3.9), we obtain the energy norm of the solution to (3.8). We will compute solutions to this system for two classical fluid flow problems in two physical dimensions, namely flow over a step and flow around an obstacle. All of the linearized systems that we analyze here were generated using the IFISS software package from the University of Manchester with grid size 4 ( 4 vertical elements) [4,[5. This is the smallest grid size on which we can obtain reasonable resolution of the eddie currents. With this grid, the coefficient matrices in the step flow problem are of size The A and K blocks have dimension , while the bottom right zero blocks have size Therefore, when we follow the procedure outlined in section., we obtain a solution of the form x(t) = exp ( [ M tq Q ) x(), (3.) where M has dimension (797 99). For the obstacle flow problem, the coefficient matrices are smaller, having dimension A and K have dimension 9 9, and M in the solution has size 896. In the following sections we will show how the pseudospectral properties of M change with the viscosity ν by computing approximations to the Kreiss constant, which will in turn tell us how the transient behavior of the system changes with ν Flow Over a Step The step flow system involves a viscous fluid flowing off the edge of a small step, creating an eddy current at the base of the step as well as a noticeable dip at the surface of the fluid. A typical steady state solution to this problem is shown in figure 3.5. Figures show the pseudospectra plots of M for viscosity values of ν =,,, and
17 Streamlines: uniform Pressure field Figure 3.5: Flow over a Step with ν = Figure 3.6: Pseudospectra for step flow, ν =. 7
18 Figure 3.7: Pseudospectra for step flow, ν = Figure 3.8: Pseudospectra for step flow, ν = 4. 8
19 Figure 3.9: Pseudospectra for step flow, ν = 5. It is easy to see that as the viscosity decreases the spectrum of M drifts closer to the imaginary axis, but all of the eigenvalues remain in the left half-plane for these viscosities, keeping the system asymptotically stable. Perhaps less obvious is the fact that some of the pseudospectral contours also begin to shift farther into the right half-plane, indicating that the Kreiss constant is increasing as the viscosity decreases. As aesthetically pleasing as these plots are, it is difficult to obtain any quantitative information from them, so table 3. lists α approximations to the Kreiss constant, K(M) = sup ε (M) ε>, for each ν. From this we can ε clearly see how the viscosity increases the level of transient growth in the solution. For a nice value of the viscosity like /, maximum energy growth is only guaranteed to reach about. times the energy of its original state, but for ν = /5 the solution s maximal energy growth is guaranteed to reach a level of about 7.8 times the initial value. Therefore, we can feel fairly comfortable using the linearized system (3.8) to model the step flow for viscosities of / or lower since we remain within a factor of of the initial energy, but for higher viscosities we see that it is possible for the energy to grow substantially from the initial value, making linear analysis an unwise choice. Table 3.: Approximate Kreiss constants for various step flow viscosities. ν / / /4 /5 Approximate K(M) Next we will look at similar results for flow around an obstacle. 9
20 3.4. Flow Around an Obstacle The obstacle flow problem involves a viscous fluid flowing around a rectangular object in the center of the flow path. A typical steady state solution is shown in figure 3.. In a similar fashion to the previous section, figures show the pseudospectra plots of the M matrix for viscosities ν =,,, and. We did not use ν = as in the step flow problem because IFISS did not converge to a steady state solution for that value. Streamlines: uniform Pressure field Figure 3.: Flow around an obstacle, ν = Figure 3.: Pseudospectra for obstacle flow, ν =.
21 Figure 3.: Pseudospectra for obstacle flow, ν = Figure 3.3: Pseudospectra for obstacle flow, ν = 4.
22 Figure 3.4: Pseudospectra for obstacle flow, ν = 45. Despite the fact that the pseudospectral contours corresponding to larger ε values extend farther into the right-half plane than for the step, table 3.4. shows that the approximate Kreiss constants follow similar patterns to those of the step flow. We see once again that for low viscosities, the linearized system will serve us just fine, but for the higher values we see enough potential growth that using a linearized system is not a viable choice. Table 3.: Approximate Kriess constants for various obstacle flow viscosities. ν / / /4 /45 Approximate K(M) Future Work Our study of these two Navier-Stokes problems demonstrates how we can combine our method for solving DAEs from section. with results from pseudospectral theory to get a better understanding of the transient properties of fluid systems that are disturbed from their steady state. From this we can predict how valid it is to use the linearized system to model the fluid system. However, one key point that we have not considered in this paper is the convergence of the Kreiss constants for the Navier-Stokes matrices as we refine the grid. That is, are the computed values on grid 4 a reasonable approximation to the true values, or would they change substantially if we went to grid 5 or 6? These numerical experiments would involve a great deal more time and computational power to perform. Thus, a point of further research would be implementing these finer grid experiments on more powerful machines, using techniques from sparse linear algebra.
23 It would also be interesting to investigate what conditions x() must satisfy in order to guarantee transient growth. We know that such initial conditions must exist, so having a way to determine whether a given initial state will or will not lead to transient growth would be an incredibly useful tool. 4 Conclusions In this paper we have covered a great deal of ground. We have developed a method for solving an arbitrary DAE via the Schur decomposition and proved that the solution that we obtain is not affected by our parameter λ. This proof required us to dig deep into the structure of the orthogonal transformations of the Schur form and how they behave when we change λ. We then discussed how the pseudospectra of M affect the potential for transient growth in the solution, giving a three-dimensional example to make the connection clear. Equipped with these tools we turned our attention to the linearized Navier-Stokes equations, deriving the appropriate change of coordinates so that the -norm of the solution that we compute is equivalent to the energy norm of the solution to the original system. We then investigated the effect of the viscosity on the pseudospectra of M and in turn the effect on the possible energy growth of the solution. This information allows us to be better informed about how valid it is to use the linearized system to analyze fluid stability. 3
24 References [ J. S. Baggett, T. A. Driscoll, L. N. Trefethen, A Mostly Linear Model of Transition to Turbulence, Physics of Fluids. 7, (995). [ S. L. Campbell, C. D. Meyer Jr., Generalized Inverses of Linear Transformations, Pitman, London, 979. [3 H. C. Elman, D. Silvester, A. Wathen, Finite Elements and Fast Iterative Solvers, Oxford University Press, 5. [4 H. C. Elman, A. Ramage, D. J. Silvester, IFISS: A Computational Laboratory for Investigating Incompressible Flow Problems, SIAM Review, 56, 6-73 (4). [5 H. C. Elman, A. Ramage, D. J. Silvester, Algorithm 866: IFISS, A MATLAB Toolbox for Modelling Incompressible Flow, ACM Transactions on Mathematical Software, 33 (7). [6 M. Embree, L. Trefethen, Spectra and Pseudospectra, Princeton University Press, 5. [7 M. Gunzburger, Finite Element Methods for Viscous Incompressible Fluids, Academic Press, 989. [8 L. N. Trefethen, A. E. Trefethen, S. C. Reddy, T. A. Driscoll, Hydrodynamic Stability without Eigenvalues, Science, 6, (993). [9 T. G. Wright, (), EigTool, Software available at 4
arxiv: v3 [math.na] 27 Jun 2017
PSEUDOSPECTRA OF MATRIX PENCILS FOR TRANSIENT ANALYSIS OF DIFFERENTIAL-ALGEBRAIC EQUATIONS MARK EMBREE AND BLAKE KEELER arxiv:6.44v3 [math.na] 7 Jun 7 Abstract. To understand the solution of a linear,
More informationMATH 5524 MATRIX THEORY Problem Set 5
MATH 554 MATRIX THEORY Problem Set 5 Posted Tuesday April 07. Due Tuesday 8 April 07. [Late work is due on Wednesday 9 April 07.] Complete any four problems, 5 points each. Recall the definitions of the
More informationLINEAR ALGEBRA KNOWLEDGE SURVEY
LINEAR ALGEBRA KNOWLEDGE SURVEY Instructions: This is a Knowledge Survey. For this assignment, I am only interested in your level of confidence about your ability to do the tasks on the following pages.
More informationFinding Rightmost Eigenvalues of Large, Sparse, Nonsymmetric Parameterized Eigenvalue Problems
Finding Rightmost Eigenvalues of Large, Sparse, Nonsymmetric Parameterized Eigenvalue Problems AMSC 663-664 Final Report Minghao Wu AMSC Program mwu@math.umd.edu Dr. Howard Elman Department of Computer
More informationMATH 583A REVIEW SESSION #1
MATH 583A REVIEW SESSION #1 BOJAN DURICKOVIC 1. Vector Spaces Very quick review of the basic linear algebra concepts (see any linear algebra textbook): (finite dimensional) vector space (or linear space),
More informationLinear Algebra Massoud Malek
CSUEB Linear Algebra Massoud Malek Inner Product and Normed Space In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n An inner product
More informationEigenvectors and Hermitian Operators
7 71 Eigenvalues and Eigenvectors Basic Definitions Let L be a linear operator on some given vector space V A scalar λ and a nonzero vector v are referred to, respectively, as an eigenvalue and corresponding
More informationNumerical Linear Algebra Homework Assignment - Week 2
Numerical Linear Algebra Homework Assignment - Week 2 Đoàn Trần Nguyên Tùng Student ID: 1411352 8th October 2016 Exercise 2.1: Show that if a matrix A is both triangular and unitary, then it is diagonal.
More informationPseudospectra and Nonnormal Dynamical Systems
Pseudospectra and Nonnormal Dynamical Systems Mark Embree and Russell Carden Computational and Applied Mathematics Rice University Houston, Texas ELGERSBURG MARCH 1 Overview of the Course These lectures
More informationChapter 7. Canonical Forms. 7.1 Eigenvalues and Eigenvectors
Chapter 7 Canonical Forms 7.1 Eigenvalues and Eigenvectors Definition 7.1.1. Let V be a vector space over the field F and let T be a linear operator on V. An eigenvalue of T is a scalar λ F such that there
More informationw T 1 w T 2. w T n 0 if i j 1 if i = j
Lyapunov Operator Let A F n n be given, and define a linear operator L A : C n n C n n as L A (X) := A X + XA Suppose A is diagonalizable (what follows can be generalized even if this is not possible -
More informationEigenvalue and Eigenvector Problems
Eigenvalue and Eigenvector Problems An attempt to introduce eigenproblems Radu Trîmbiţaş Babeş-Bolyai University April 8, 2009 Radu Trîmbiţaş ( Babeş-Bolyai University) Eigenvalue and Eigenvector Problems
More informationIterative Methods for Solving A x = b
Iterative Methods for Solving A x = b A good (free) online source for iterative methods for solving A x = b is given in the description of a set of iterative solvers called templates found at netlib: http
More informationConceptual Questions for Review
Conceptual Questions for Review Chapter 1 1.1 Which vectors are linear combinations of v = (3, 1) and w = (4, 3)? 1.2 Compare the dot product of v = (3, 1) and w = (4, 3) to the product of their lengths.
More informationReview problems for MA 54, Fall 2004.
Review problems for MA 54, Fall 2004. Below are the review problems for the final. They are mostly homework problems, or very similar. If you are comfortable doing these problems, you should be fine on
More information5.3 The Power Method Approximation of the Eigenvalue of Largest Module
192 5 Approximation of Eigenvalues and Eigenvectors 5.3 The Power Method The power method is very good at approximating the extremal eigenvalues of the matrix, that is, the eigenvalues having largest and
More informationA Robust Preconditioned Iterative Method for the Navier-Stokes Equations with High Reynolds Numbers
Applied and Computational Mathematics 2017; 6(4): 202-207 http://www.sciencepublishinggroup.com/j/acm doi: 10.11648/j.acm.20170604.18 ISSN: 2328-5605 (Print); ISSN: 2328-5613 (Online) A Robust Preconditioned
More informationLecture 3: QR-Factorization
Lecture 3: QR-Factorization This lecture introduces the Gram Schmidt orthonormalization process and the associated QR-factorization of matrices It also outlines some applications of this factorization
More informationarxiv: v1 [math.na] 5 May 2011
ITERATIVE METHODS FOR COMPUTING EIGENVALUES AND EIGENVECTORS MAYSUM PANJU arxiv:1105.1185v1 [math.na] 5 May 2011 Abstract. We examine some numerical iterative methods for computing the eigenvalues and
More informationNotes on Eigenvalues, Singular Values and QR
Notes on Eigenvalues, Singular Values and QR Michael Overton, Numerical Computing, Spring 2017 March 30, 2017 1 Eigenvalues Everyone who has studied linear algebra knows the definition: given a square
More informationBalanced Truncation 1
Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.242, Fall 2004: MODEL REDUCTION Balanced Truncation This lecture introduces balanced truncation for LTI
More informationLecture 4: Analysis of MIMO Systems
Lecture 4: Analysis of MIMO Systems Norms The concept of norm will be extremely useful for evaluating signals and systems quantitatively during this course In the following, we will present vector norms
More informationPerturbation theory for eigenvalues of Hermitian pencils. Christian Mehl Institut für Mathematik TU Berlin, Germany. 9th Elgersburg Workshop
Perturbation theory for eigenvalues of Hermitian pencils Christian Mehl Institut für Mathematik TU Berlin, Germany 9th Elgersburg Workshop Elgersburg, March 3, 2014 joint work with Shreemayee Bora, Michael
More informationB553 Lecture 5: Matrix Algebra Review
B553 Lecture 5: Matrix Algebra Review Kris Hauser January 19, 2012 We have seen in prior lectures how vectors represent points in R n and gradients of functions. Matrices represent linear transformations
More informationDefinite versus Indefinite Linear Algebra. Christian Mehl Institut für Mathematik TU Berlin Germany. 10th SIAM Conference on Applied Linear Algebra
Definite versus Indefinite Linear Algebra Christian Mehl Institut für Mathematik TU Berlin Germany 10th SIAM Conference on Applied Linear Algebra Monterey Bay Seaside, October 26-29, 2009 Indefinite Linear
More informationThe Eigenvalue Problem: Perturbation Theory
Jim Lambers MAT 610 Summer Session 2009-10 Lecture 13 Notes These notes correspond to Sections 7.2 and 8.1 in the text. The Eigenvalue Problem: Perturbation Theory The Unsymmetric Eigenvalue Problem Just
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 18 Outline
More informationEIGENVALUE PROBLEMS. Background on eigenvalues/ eigenvectors / decompositions. Perturbation analysis, condition numbers..
EIGENVALUE PROBLEMS Background on eigenvalues/ eigenvectors / decompositions Perturbation analysis, condition numbers.. Power method The QR algorithm Practical QR algorithms: use of Hessenberg form and
More informationNotes for CS542G (Iterative Solvers for Linear Systems)
Notes for CS542G (Iterative Solvers for Linear Systems) Robert Bridson November 20, 2007 1 The Basics We re now looking at efficient ways to solve the linear system of equations Ax = b where in this course,
More informationExtreme Values and Positive/ Negative Definite Matrix Conditions
Extreme Values and Positive/ Negative Definite Matrix Conditions James K. Peterson Department of Biological Sciences and Department of Mathematical Sciences Clemson University November 8, 016 Outline 1
More informationLinear Algebra Methods for Data Mining
Linear Algebra Methods for Data Mining Saara Hyvönen, Saara.Hyvonen@cs.helsinki.fi Spring 2007 1. Basic Linear Algebra Linear Algebra Methods for Data Mining, Spring 2007, University of Helsinki Example
More informationRobust control of resistive wall modes using pseudospectra
Robust control of resistive wall modes using pseudospectra M. Sempf, P. Merkel, E. Strumberger, C. Tichmann, and S. Günter Max-Planck-Institut für Plasmaphysik, EURATOM Association, Garching, Germany GOTiT
More informationPreliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012
Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.
More informationMATH 5524 MATRIX THEORY Problem Set 4
MATH 5524 MATRIX THEORY Problem Set 4 Posted Tuesday 28 March 217. Due Tuesday 4 April 217. [Corrected 3 April 217.] [Late work is due on Wednesday 5 April.] Complete any four problems, 25 points each.
More informationNumerical Linear Algebra
Numerical Linear Algebra The two principal problems in linear algebra are: Linear system Given an n n matrix A and an n-vector b, determine x IR n such that A x = b Eigenvalue problem Given an n n matrix
More informationMATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.
MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators. Adjoint operator and adjoint matrix Given a linear operator L on an inner product space V, the adjoint of L is a transformation
More informationMIT Final Exam Solutions, Spring 2017
MIT 8.6 Final Exam Solutions, Spring 7 Problem : For some real matrix A, the following vectors form a basis for its column space and null space: C(A) = span,, N(A) = span,,. (a) What is the size m n of
More informationCOMP 558 lecture 18 Nov. 15, 2010
Least squares We have seen several least squares problems thus far, and we will see more in the upcoming lectures. For this reason it is good to have a more general picture of these problems and how to
More informationLecture 7. Econ August 18
Lecture 7 Econ 2001 2015 August 18 Lecture 7 Outline First, the theorem of the maximum, an amazing result about continuity in optimization problems. Then, we start linear algebra, mostly looking at familiar
More informationNumerical Methods for Solving Large Scale Eigenvalue Problems
Peter Arbenz Computer Science Department, ETH Zürich E-mail: arbenz@inf.ethz.ch arge scale eigenvalue problems, Lecture 2, February 28, 2018 1/46 Numerical Methods for Solving Large Scale Eigenvalue Problems
More informationA Review of Preconditioning Techniques for Steady Incompressible Flow
Zeist 2009 p. 1/43 A Review of Preconditioning Techniques for Steady Incompressible Flow David Silvester School of Mathematics University of Manchester Zeist 2009 p. 2/43 PDEs Review : 1984 2005 Update
More informationDS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.
DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1
More information21 Linear State-Space Representations
ME 132, Spring 25, UC Berkeley, A Packard 187 21 Linear State-Space Representations First, let s describe the most general type of dynamic system that we will consider/encounter in this class Systems may
More informationTermination criteria for inexact fixed point methods
Termination criteria for inexact fixed point methods Philipp Birken 1 October 1, 2013 1 Institute of Mathematics, University of Kassel, Heinrich-Plett-Str. 40, D-34132 Kassel, Germany Department of Mathematics/Computer
More informationLecture 8 : Eigenvalues and Eigenvectors
CPS290: Algorithmic Foundations of Data Science February 24, 2017 Lecture 8 : Eigenvalues and Eigenvectors Lecturer: Kamesh Munagala Scribe: Kamesh Munagala Hermitian Matrices It is simpler to begin with
More informationNumerical Methods I Eigenvalue Problems
Numerical Methods I Eigenvalue Problems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 October 2nd, 2014 A. Donev (Courant Institute) Lecture
More informationThroughout these notes we assume V, W are finite dimensional inner product spaces over C.
Math 342 - Linear Algebra II Notes Throughout these notes we assume V, W are finite dimensional inner product spaces over C 1 Upper Triangular Representation Proposition: Let T L(V ) There exists an orthonormal
More informationEfficient Augmented Lagrangian-type Preconditioning for the Oseen Problem using Grad-Div Stabilization
Efficient Augmented Lagrangian-type Preconditioning for the Oseen Problem using Grad-Div Stabilization Timo Heister, Texas A&M University 2013-02-28 SIAM CSE 2 Setting Stationary, incompressible flow problems
More informationMatrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A =
30 MATHEMATICS REVIEW G A.1.1 Matrices and Vectors Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A = a 11 a 12... a 1N a 21 a 22... a 2N...... a M1 a M2... a MN A matrix can
More informationThe Singular Value Decomposition
The Singular Value Decomposition Philippe B. Laval KSU Fall 2015 Philippe B. Laval (KSU) SVD Fall 2015 1 / 13 Review of Key Concepts We review some key definitions and results about matrices that will
More informationEigenvalues and eigenvectors
Chapter 6 Eigenvalues and eigenvectors An eigenvalue of a square matrix represents the linear operator as a scaling of the associated eigenvector, and the action of certain matrices on general vectors
More information1 Last time: least-squares problems
MATH Linear algebra (Fall 07) Lecture Last time: least-squares problems Definition. If A is an m n matrix and b R m, then a least-squares solution to the linear system Ax = b is a vector x R n such that
More informationVISUALIZING PSEUDOSPECTRA FOR POLYNOMIAL EIGENVALUE PROBLEMS. Adéla Klimentová *, Michael Šebek ** Czech Technical University in Prague
VSUALZNG PSEUDOSPECTRA FOR POLYNOMAL EGENVALUE PROBLEMS Adéla Klimentová *, Michael Šebek ** * Department of Control Engineering Czech Technical University in Prague ** nstitute of nformation Theory and
More informationLinear Algebra. The analysis of many models in the social sciences reduces to the study of systems of equations.
POLI 7 - Mathematical and Statistical Foundations Prof S Saiegh Fall Lecture Notes - Class 4 October 4, Linear Algebra The analysis of many models in the social sciences reduces to the study of systems
More informationNORMS ON SPACE OF MATRICES
NORMS ON SPACE OF MATRICES. Operator Norms on Space of linear maps Let A be an n n real matrix and x 0 be a vector in R n. We would like to use the Picard iteration method to solve for the following system
More informationOperations On Networks Of Discrete And Generalized Conductors
Operations On Networks Of Discrete And Generalized Conductors Kevin Rosema e-mail: bigyouth@hardy.u.washington.edu 8/18/92 1 Introduction The most basic unit of transaction will be the discrete conductor.
More informationLinear Algebra March 16, 2019
Linear Algebra March 16, 2019 2 Contents 0.1 Notation................................ 4 1 Systems of linear equations, and matrices 5 1.1 Systems of linear equations..................... 5 1.2 Augmented
More informationBare-bones outline of eigenvalue theory and the Jordan canonical form
Bare-bones outline of eigenvalue theory and the Jordan canonical form April 3, 2007 N.B.: You should also consult the text/class notes for worked examples. Let F be a field, let V be a finite-dimensional
More informationMath Ordinary Differential Equations
Math 411 - Ordinary Differential Equations Review Notes - 1 1 - Basic Theory A first order ordinary differential equation has the form x = f(t, x) (11) Here x = dx/dt Given an initial data x(t 0 ) = x
More informationDraft. Lecture 01 Introduction & Matrix-Vector Multiplication. MATH 562 Numerical Analysis II. Songting Luo
Lecture 01 Introduction & Matrix-Vector Multiplication Songting Luo Department of Mathematics Iowa State University MATH 562 Numerical Analysis II Songting Luo ( Department of Mathematics Iowa State University[0.5in]
More informationApplied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic
Applied Mathematics 205 Unit V: Eigenvalue Problems Lecturer: Dr. David Knezevic Unit V: Eigenvalue Problems Chapter V.4: Krylov Subspace Methods 2 / 51 Krylov Subspace Methods In this chapter we give
More informationProblem set 2. Math 212b February 8, 2001 due Feb. 27
Problem set 2 Math 212b February 8, 2001 due Feb. 27 Contents 1 The L 2 Euler operator 1 2 Symplectic vector spaces. 2 2.1 Special kinds of subspaces....................... 3 2.2 Normal forms..............................
More informationACM 104. Homework Set 5 Solutions. February 21, 2001
ACM 04 Homework Set 5 Solutions February, 00 Franklin Chapter 4, Problem 4, page 0 Let A be an n n non-hermitian matrix Suppose that A has distinct eigenvalues λ,, λ n Show that A has the eigenvalues λ,,
More informationEIGENVALUE PROBLEMS. EIGENVALUE PROBLEMS p. 1/4
EIGENVALUE PROBLEMS EIGENVALUE PROBLEMS p. 1/4 EIGENVALUE PROBLEMS p. 2/4 Eigenvalues and eigenvectors Let A C n n. Suppose Ax = λx, x 0, then x is a (right) eigenvector of A, corresponding to the eigenvalue
More informationON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH
ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH V. FABER, J. LIESEN, AND P. TICHÝ Abstract. Numerous algorithms in numerical linear algebra are based on the reduction of a given matrix
More informationMatrix Mathematics. Theory, Facts, and Formulas with Application to Linear Systems Theory. Dennis S. Bernstein
Matrix Mathematics Theory, Facts, and Formulas with Application to Linear Systems Theory Dennis S. Bernstein PRINCETON UNIVERSITY PRESS PRINCETON AND OXFORD Contents Special Symbols xv Conventions, Notation,
More informationLinear Algebra: Matrix Eigenvalue Problems
CHAPTER8 Linear Algebra: Matrix Eigenvalue Problems Chapter 8 p1 A matrix eigenvalue problem considers the vector equation (1) Ax = λx. 8.0 Linear Algebra: Matrix Eigenvalue Problems Here A is a given
More informationFundamentals of Linear Algebra. Marcel B. Finan Arkansas Tech University c All Rights Reserved
Fundamentals of Linear Algebra Marcel B. Finan Arkansas Tech University c All Rights Reserved 2 PREFACE Linear algebra has evolved as a branch of mathematics with wide range of applications to the natural
More informationPreconditioners for the incompressible Navier Stokes equations
Preconditioners for the incompressible Navier Stokes equations C. Vuik M. ur Rehman A. Segal Delft Institute of Applied Mathematics, TU Delft, The Netherlands SIAM Conference on Computational Science and
More informationStat 159/259: Linear Algebra Notes
Stat 159/259: Linear Algebra Notes Jarrod Millman November 16, 2015 Abstract These notes assume you ve taken a semester of undergraduate linear algebra. In particular, I assume you are familiar with the
More informationNONCOMMUTATIVE POLYNOMIAL EQUATIONS. Edward S. Letzter. Introduction
NONCOMMUTATIVE POLYNOMIAL EQUATIONS Edward S Letzter Introduction My aim in these notes is twofold: First, to briefly review some linear algebra Second, to provide you with some new tools and techniques
More informationEECS 275 Matrix Computation
EECS 275 Matrix Computation Ming-Hsuan Yang Electrical Engineering and Computer Science University of California at Merced Merced, CA 95344 http://faculty.ucmerced.edu/mhyang Lecture 16 1 / 21 Overview
More informationFinal A. Problem Points Score Total 100. Math115A Nadja Hempel 03/23/2017
Final A Math115A Nadja Hempel 03/23/2017 nadja@math.ucla.edu Name: UID: Problem Points Score 1 10 2 20 3 5 4 5 5 9 6 5 7 7 8 13 9 16 10 10 Total 100 1 2 Exercise 1. (10pt) Let T : V V be a linear transformation.
More informationHere is an example of a block diagonal matrix with Jordan Blocks on the diagonal: J
Class Notes 4: THE SPECTRAL RADIUS, NORM CONVERGENCE AND SOR. Math 639d Due Date: Feb. 7 (updated: February 5, 2018) In the first part of this week s reading, we will prove Theorem 2 of the previous class.
More informationApplied Mathematics 205. Unit II: Numerical Linear Algebra. Lecturer: Dr. David Knezevic
Applied Mathematics 205 Unit II: Numerical Linear Algebra Lecturer: Dr. David Knezevic Unit II: Numerical Linear Algebra Chapter II.3: QR Factorization, SVD 2 / 66 QR Factorization 3 / 66 QR Factorization
More informationSolving the steady state diffusion equation with uncertainty Final Presentation
Solving the steady state diffusion equation with uncertainty Final Presentation Virginia Forstall vhfors@gmail.com Advisor: Howard Elman elman@cs.umd.edu Department of Computer Science May 6, 2012 Problem
More informationLecture 11. Linear systems: Cholesky method. Eigensystems: Terminology. Jacobi transformations QR transformation
Lecture Cholesky method QR decomposition Terminology Linear systems: Eigensystems: Jacobi transformations QR transformation Cholesky method: For a symmetric positive definite matrix, one can do an LU decomposition
More informationEECS 275 Matrix Computation
EECS 275 Matrix Computation Ming-Hsuan Yang Electrical Engineering and Computer Science University of California at Merced Merced, CA 95344 http://faculty.ucmerced.edu/mhyang Lecture 17 1 / 26 Overview
More informationThe Jordan Canonical Form
The Jordan Canonical Form The Jordan canonical form describes the structure of an arbitrary linear transformation on a finite-dimensional vector space over an algebraically closed field. Here we develop
More informationLecture notes: Applied linear algebra Part 1. Version 2
Lecture notes: Applied linear algebra Part 1. Version 2 Michael Karow Berlin University of Technology karow@math.tu-berlin.de October 2, 2008 1 Notation, basic notions and facts 1.1 Subspaces, range and
More information12. Perturbed Matrices
MAT334 : Applied Linear Algebra Mike Newman, winter 208 2. Perturbed Matrices motivation We want to solve a system Ax = b in a context where A and b are not known exactly. There might be experimental errors,
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra)
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 1: Course Overview & Matrix-Vector Multiplication Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 20 Outline 1 Course
More informationUniversity of Maryland Department of Computer Science TR-5009 University of Maryland Institute for Advanced Computer Studies TR April 2012
University of Maryland Department of Computer Science TR-5009 University of Maryland Institute for Advanced Computer Studies TR-202-07 April 202 LYAPUNOV INVERSE ITERATION FOR COMPUTING A FEW RIGHTMOST
More informationFast Iterative Solution of Saddle Point Problems
Michele Benzi Department of Mathematics and Computer Science Emory University Atlanta, GA Acknowledgments NSF (Computational Mathematics) Maxim Olshanskii (Mech-Math, Moscow State U.) Zhen Wang (PhD student,
More informationMATH 5640: Functions of Diagonalizable Matrices
MATH 5640: Functions of Diagonalizable Matrices Hung Phan, UMass Lowell November 27, 208 Spectral theorem for diagonalizable matrices Definition Let V = X Y Every v V is uniquely decomposed as u = x +
More information6 EIGENVALUES AND EIGENVECTORS
6 EIGENVALUES AND EIGENVECTORS INTRODUCTION TO EIGENVALUES 61 Linear equations Ax = b come from steady state problems Eigenvalues have their greatest importance in dynamic problems The solution of du/dt
More informationSpectral Processing. Misha Kazhdan
Spectral Processing Misha Kazhdan [Taubin, 1995] A Signal Processing Approach to Fair Surface Design [Desbrun, et al., 1999] Implicit Fairing of Arbitrary Meshes [Vallet and Levy, 2008] Spectral Geometry
More informationON A GENERAL CLASS OF PRECONDITIONERS FOR NONSYMMETRIC GENERALIZED SADDLE POINT PROBLEMS
U..B. Sci. Bull., Series A, Vol. 78, Iss. 4, 06 ISSN 3-707 ON A GENERAL CLASS OF RECONDIIONERS FOR NONSYMMERIC GENERALIZED SADDLE OIN ROBLE Fatemeh anjeh Ali BEIK his paper deals with applying a class
More informationNotes on SU(3) and the Quark Model
Notes on SU() and the Quark Model Contents. SU() and the Quark Model. Raising and Lowering Operators: The Weight Diagram 4.. Triangular Weight Diagrams (I) 6.. Triangular Weight Diagrams (II) 8.. Hexagonal
More information14 Singular Value Decomposition
14 Singular Value Decomposition For any high-dimensional data analysis, one s first thought should often be: can I use an SVD? The singular value decomposition is an invaluable analysis tool for dealing
More informationPhysics 202 Laboratory 5. Linear Algebra 1. Laboratory 5. Physics 202 Laboratory
Physics 202 Laboratory 5 Linear Algebra Laboratory 5 Physics 202 Laboratory We close our whirlwind tour of numerical methods by advertising some elements of (numerical) linear algebra. There are three
More informationLinear Algebra (part 1) : Vector Spaces (by Evan Dummit, 2017, v. 1.07) 1.1 The Formal Denition of a Vector Space
Linear Algebra (part 1) : Vector Spaces (by Evan Dummit, 2017, v. 1.07) Contents 1 Vector Spaces 1 1.1 The Formal Denition of a Vector Space.................................. 1 1.2 Subspaces...................................................
More informationLecture 4 Eigenvalue problems
Lecture 4 Eigenvalue problems Weinan E 1,2 and Tiejun Li 2 1 Department of Mathematics, Princeton University, weinan@princeton.edu 2 School of Mathematical Sciences, Peking University, tieli@pku.edu.cn
More informationPutzer s Algorithm. Norman Lebovitz. September 8, 2016
Putzer s Algorithm Norman Lebovitz September 8, 2016 1 Putzer s algorithm The differential equation dx = Ax, (1) dt where A is an n n matrix of constants, possesses the fundamental matrix solution exp(at),
More informationElementary linear algebra
Chapter 1 Elementary linear algebra 1.1 Vector spaces Vector spaces owe their importance to the fact that so many models arising in the solutions of specific problems turn out to be vector spaces. The
More informationLecture notes: Applied linear algebra Part 2. Version 1
Lecture notes: Applied linear algebra Part 2. Version 1 Michael Karow Berlin University of Technology karow@math.tu-berlin.de October 2, 2008 First, some exercises: xercise 0.1 (2 Points) Another least
More informationSpectral Theorem for Self-adjoint Linear Operators
Notes for the undergraduate lecture by David Adams. (These are the notes I would write if I was teaching a course on this topic. I have included more material than I will cover in the 45 minute lecture;
More informationMATH 315 Linear Algebra Homework #1 Assigned: August 20, 2018
Homework #1 Assigned: August 20, 2018 Review the following subjects involving systems of equations and matrices from Calculus II. Linear systems of equations Converting systems to matrix form Pivot entry
More informationLecture notes on Quantum Computing. Chapter 1 Mathematical Background
Lecture notes on Quantum Computing Chapter 1 Mathematical Background Vector states of a quantum system with n physical states are represented by unique vectors in C n, the set of n 1 column vectors 1 For
More informationReview of some mathematical tools
MATHEMATICAL FOUNDATIONS OF SIGNAL PROCESSING Fall 2016 Benjamín Béjar Haro, Mihailo Kolundžija, Reza Parhizkar, Adam Scholefield Teaching assistants: Golnoosh Elhami, Hanjie Pan Review of some mathematical
More information