Iterative methods for linear systems

Size: px
Start display at page:

Download "Iterative methods for linear systems"

Transcription

1 Iterative methods for linear sstems Chris H. Rcroft November 9th, 26 Introduction For man elliptic PDE problems, finite-difference and finite-element methods are the techniques of choice. In a finite-difference approach, a solution u k on a set of discrete gridpoints,..., k is searched for. The discretized partial differential equation and boundar conditions create linear relationships between the different values of u k. In the finite-element method, the solution is epressed as a linear combination u k of basis functions λ k on the domain, and the corresponding finite-element variational problem again gives linear relationships between the different values of u k. Regardless of the precise details, all of these approaches ultimatel end up with having to find the u k that satisf all the linear relationships prescribed b the PDE. This can be written as a matri equation of the form Au = b () where one aims to find a solution u, given that A is a matri capturing the differentiation operator, and b corresponds to an source or boundar terms. Theoreticall, this problem could be solved on a computer b an of the standard methods for dealing with matrices. However, the real challenge for PDEs is that frequentl the dimensionalit of the problem can be enormous. For eample, for a two-dimensional PDE problem, a grid would be a perfectl reasonable size to consider. Thus u would be a vector with 4 elements, and A would be a matri with 8 elements. Even allocating memor for such a large matri ma be problematic. Direct approaches, such as the eplicit construction of A, are impractical. The ke to making progress is to note that in general the matri A is etremel sparse, since the linear relationships usuall onl relate nearb gridpoints together. It is therefore natural to seek methods that do not require ever eplicitl specifing all the elements of A, but eploit its special structure directl. Man of these methods are iterative starting with Electronic address: chr@seas.harvard.edu.

2 a guess u k, a process is applied that ields a closer solution u k+. Tpicall, these iterative methods are based on a splitting of A. This is a decomposition A = M K where M is non-singular. An splitting creates a possible iterative process. Equation can be rewritten as and hence a possible iteration is (M K)u = b, (2) Mu = Ku + b, (3) u = M Ku + M b, (4) u k+ = M Ku k + M b. (5) Of course, there is no guarantee that an arbitrar splitting will result in an iterative method that converges. To stud convergence, one must look at the properties of the matri R = M K. For convergence analsis, it is helpful to introduce the spectral radius ρ(r) = ma { λ j, (6) j where the λ j are the eigenvalues of R. It can be shown [2] that an iterative scheme converges if and onl if ρ(r) <. The size of the spectral radius determines the convergence rate, and ideall one would like to find splittings that result in as small a ρ(r) as possible. An eample: a two-dimensional Poisson problem In the convergence analsis later, we will consider a two-dimensional Poisson problem on the square,, given b the equation 2 u = f, (7) subject to the Dirichlet conditions that u(, ) vanishes on the boundar. We use a source function of the form { if <.5 and <.5 f (, ) = otherwise. (8) This is plotted on a grid in figure. For convergence properties, the eigenfunctions and eigenvalues of this function are ver important, and to determine these, it is helpful to consider an associated one-dimensional Poisson problem on the interval, d2 u = f (), (9) d2 2

3 subject to Dirichlet boundar conditions u( ) = u() =. We consider a discretization into N + 2 gridpoints such that j = + 2j/(N + ) for j =,..., N +. When constructing the corresponding matri problem, u and u N+ need not be considered, since their values are alwas fied to zero. B discretizing the second derivative according to d 2 u d 2 = u j + u j+ 2u j =j h 2 () where h = 2/(N + ), the corresponding linear sstem is 2 u. u T N. =... u... N 2 2 u u. u N = h2 f f. f N. () We epect that the eigenvectors of T N ma be based on sine functions. A reasonable guess for the jth eigenfunction is 2 z j (k) = N + sin πkj N +. (2) To verif this is an eigenfunction, and find its eigenvalue, we appl T N to obtain ( ( ) 2 TN z j (k) = 2 sin πkj ) π(k + )j π(k )j sin sin. (3) N + N + N + N + Note that this epression will alwas be valid for the range k =, 2,..., N, and the boundar values will automaticall work out. The last two sine functions can be rewritten using a trigonometric identit to give ( ( ) 2 TN z j (k) = 2 sin πkj N + N + 2 sin ( 2 = N + 2 πj cos N + = 2 ( cos πj N + ) sin πkj N + cos πkj N + ) πj N + ) z j (k) (4) and hence z j is an eigenvector with eigenvalue λ j = 2( cos πj/(n + )). The smallest eigenvalue is λ = 2( cos π/(n + )) and the largest is λ N = 2( cos Nπ/(N + )). Returning to the two-dimensional problem, the corresponding derivative matri T N N can be written as the tensor product of two one-dimensional problems T N. Its eigenvectors are the tensor products of the one-dimensional eigenvectors, namel z i,j (k, l) = z i (k)z j (l), (5) 3

4 f Figure : A sample source function f (, ) on a grid. u Figure 2: The eact solution to the 2D Poisson problem 2 u = f, with zero boundar conditions and a source term given in figure. 4

5 . u Figure 3: The solution to the eample 2D Poisson problem after ten iterations of the Jacobi method. and the corresponding eigenvalues are λ i,j = λ i + λ j. (6) The Jacobi Method The Jacobi method is one of the simplest iterations to implement. While its convergence properties make it too slow for use in man problems, it is worthwhile to consider since it forms the basis of other methods. Starting with an initial guess u, one successivel applies an iteration of the form for j = to N do ) u m+,j = b j k =j a jk u m,k end for a jj ( In other words, the jth component of u is set so that it eactl satisfies equation j of the linear sstem. For the two-dimensional Poisson problem considered above, this corresponds to an iteration of the form for i = to N do for j = to N do 5

6 u m+,i,j = ( h 2 f j + u m,i,j+ + u m,i,j + u m,i+,j + u m,i,j ) /4 end for end for To find the corresponding matri form, write A = D L U where D is diagonal, L is lower-triangular, and U is upper-triangular. Then the above iteration is u m+ = D (L + U)u m + D b. (7) The convergence properties, discussed later, are then set b the matri R J = D (L + U). The Jacobi method has the advantage that for each m, the order in which the components of u m+ are computed has no effect this ma be a favorable propert to have in some parallel implementations. However, it can also be seen that u m must be retained until after u m+ is constructed, meaning we must store u m+ in a different part of the memor. The listing given in Appendi A. carries out the Jacobi iteration on the Poisson test function. It makes use of two arras for the storage of u, computing the odd u k in one and the even u k in the other. Figure 3 shows a the progress of the Jacobi method after ten iterations. The Gauss Seidel Method The Gauss Seidel method improves on the Jacobi algorithm, b noting that when updating a particular point u m+,j, one might as well reference the alread updated values u m+,,...,u m+,j in the calculation, rather than using the original values u m,,..., u m,j. The iteration can be written as for j = to N ( do ) u m+,j = a b jj j j k= a jku m+,k k=j+ N u m,k end for The Gauss Seidel algorithm has the advantage that in a computer implementation, it is not necessar to allocate two arras for u m+ and u m. Instead, a single arra for u m can be used and all updates can be carried out in situ. However, the Gauss Seidel implementation introduces an additional complication that the order in which the updates are applied affects the values of u m. For a two-dimensional problem, two particular orderings are worth special attention: Natural ordering this is the tpical ordering that would result in a for loop. We first loop successivel through all elements of the first row (, ),..., (, n) before moving onto the second row, and so on. Since this frequentl corresponds to the order that the grid is stored within the computer memor, this order can lead to performance improvements due to good cache efficienc. 6

7 .2 u Figure 4: The Gauss Seidel solution to the eample 2D Poisson problem after ten iterations. The crinkles in the solution are due to the Red Black update procedure. Red Black ordering this is the ordering that results b coloring the gridpoints red and black in a checkerboard pattern. Specificall, a gridpoint (i, j) is colored red if i + j is even and colored black if i + j is odd. During the Gauss Seidel update, all red points are updated before the black points. For the two-dimensional Poisson problem, updating a red grid point onl requires information from the black gridpoints, and vice versa. Hence the order in which points in each set are updated does not matter. The whole Gauss Seidel update is divided into a red gridpoint update and black gridpoint update, and this can be helpful in the convergence analsis. From the algorithm above, the corresponding matri splitting for the Gauss Seidel method is (D L)u m+ = Uu m + b, (8) u m+ = (D L) Uu m + (D L) b. (9) Appendi A.2 contains a C++ code to carr out a Gauss Seidel method on the eample problem, and the result after ten iterations is shown in figure 4. Successive Over-Relaation Successive Over-Relaation (SOR) is a refinement to the Gauss Seidel algorithm. At each stage in the Gauss Seidel algorithm, a value u m,j is updated to a new one u m+,j, which is 7

8 u Figure 5: The SOR solution (using the theoreticall optimal ω) to the eample 2D Poisson problem after ten iterations. The solution is closer to the answer than the Jacobi or Gauss Seidel methods. equivalent to displacing u m,j b an amount u = u m+,j u m,j. The SOR algorithm works b displacing the values b an amount ω u, where tpicall ω >, in the hope that if u is a good direction to move in, one might as well move further in that direction. The iteration is for j = to N do ( ) u m+,j = ( ω)u m,j + a ω b jj j j k= a jku m+,k k=j+ N u m,k end for The corresponding matri form is (D + ωl)u m+ = [( ω)d Uω] u m + ωb, (2) u m+ = (D + ωl) [( ω)d Uω] u m + (D + ωl) ωb. (2) Appendi A.3 contains a C++ code to carr out the SOR iteration on the eample problem, and the result is shown in figure 5. In the SOR algorithm, the value of ω can be chosen freel, and the best choices can be determined b considering the eigenfunctions of the associated problem. This is discussed in more detail below. 8

9 Convergence analsis and compleit To eamine the convergence properties of the different methods, we need to look at the associated spectral radii. For the Jacobi method, the matri of interest is R J = D (L + U). For the 2D Poisson problem, D = 4I and hence R J = (4I) (4I T N N ) = I T N N. (22) 4 The largest eigenvalue of R J corresponds to the smallest of T N N, namel ( ) ( ) π π λ, = 4 2 cos 2 cos N + N + ( ) π = 4 4 cos. (23) N + Hence ( ) π ρ(r J ) = cos. (24) N + Epanding the cosine as a Talor series shows that ρ(r J ) π2. The time to gain 2(N+) an etra digit of accurac is approimatel 2 log ρ(r J ) N2, (25) so the algorithm must be run for O(N 2 ) iterations to attain a specific level of accurac. Since there are O(N 2 ) gridpoints for the 2D problem, the total running time is O(N 4 ). For detailed proofs of the convergence properties of the other methods, the reader should consult the book b Demmel [2]. It can be shown that ( ) π ρ(r GS ) = cos 2, (26) N + so that one iteration of the Gauss Seidel method is equivalent to two Jacobi iterations. Note however the compleit is the same: O(N 2 ) iterations are still required to reach a desired level of accurac. For the SOR algorithm, it can be shown that the optimal value of ω is 2. (27) + ρ(r J ) 2 and that for this value ρ(r SOR ) 2 2π N +. (28) Since there is a factor of N in the denominator as opposed to N 2, the order of computation decreases to O(N) per gridpoint. 9

10 5 Jacobi Gauss Seidel Optimal SOR L 2 -norm error Iterations Figure 6: Errors versus the number of effective iterations for the Jacobi, Gauss Seidel, and SOR methods, applied to the eample 2D Poisson problem on a grid. The plots are in line with the theoretical results of the tet. The Gauss Seidel method is faster than the Jacobi method, but has still not reached double numerical precision after iterations. The SOR method is significantl faster, but still requires 2 iterations to reach double numerical precision. Figure 6 shows a plot of mean square error against the number of iterations for the model problem with the Jacobi, Gauss Seidel, and optimal SOR method. The lines agree with the above results. The SOR method reaches numerical precision within 2 iterations, while the other two methods have not full converged even after 4 iterations. Multigrid One of the major problems with the three methods considered so far is that the onl appl locall. Information about different cell values onl propagates b one or two gridpoints per iteration. However, for man elliptic problems, a point source ma cause an effect over the entire domain. The above methods have a fundamental limitation that the will need to be applied for at least as man iterations as it takes for information to propagate across the grid. As such, we should not epect to ever do better than O(N) operations per point. This can also be seen b considering the eigenvalues. The maimal eigenvalue of R J was set b the λ,, corresponding to the lowest order mode. While the methods ma effectivel damp out high frequenc oscillations, it takes a ver long time to correctl capture the lowest modes with the largest wavelengths. The multigrid method circumvents these limitations b introducing a hierarch of

11 f () f () f (2) f (3) Figure 7: The restriction of the eample source term f using the multigrid method with v down =, v up = 2. Top left: the initial grid, with gridpoints. Top right: grid, with 7 7 gridpoints. Bottom left: grid 2, with 9 9 gridpoints. Bottom right: grid 3, with 5 5 gridpoints. coarser and coarser grids. Tpicall, at each level, the number of gridpoints is reduced b a factor of two in each direction, with the coarsest grid having roughl ten to twent points. To find a solution, we restrict the source term to the coarse grids, refine the solution on each, and interpolate up to the original grid. On the coarser grids, the lower frequenc modes in the final solution can be dealt with much more effectivel. Since the coarser grids have progressivel fewer gridpoints, the time spent computing them is minimal. Because of this, the multigrid algorithm requires onl O() computation per point, which is the best order of compleit that can be achieved. To be more specific, let the original problem be on grid, and we then introduce a sequence of successivel coarser grids,..., g. Write u (i) and b (i) to represent the solution and source terms on the ith grid. A multigrid algorithm requires the following: A solution operator S(u (i), b (i) ) that returns a better approimation u (i) to the solution on the ith level, An interpolation operator T(u (i) ) that returns an interpolation u (i ) on the (i )th level, A restriction operator R(b (i) ) that returns a restriction b (i+) on the (i + )th level. The interpolation and restriction operators can be thought of as rectangular matrices. As an eample, consider a problem with nine equall-spaced gridpoints on the unit interval,

12 u (3) u (2) u () u () Figure 8: The solution of the v in the first multigrid V-ccle for the eample 2D Poisson problem. Top left: the eact solution on grid 3 to the f (3) source term in figure 7. Top right: the interpolation and refinement on grid 2. Bottom left: the interpolation and refinement on grid. Bottom right: the interpolation and refinement on grid. Even after a single V-ccle, the solution is closer to the eact solution than the plots of 3, 4, and 5. at (, /8, 2/8,..., ). Let grid have five points at (, /4, 2/4, 3/4, ), and let grid 2 have three gridpoints at (, /2, ). Interpolation operators between the grids can be written as T () = /2 /2 /2 /2 /2 /2 /2 /2, T (2) /2 /2 /2 /2 where the function values at gridpoints common between the two levels are kept, and the function values at midpoints are calculated using linear interpolation. Similarl, the (29) 2

13 restriction operators can be written as R () = /4 /2 /4 /4 /2 /4 /4 /2 /4 R () = /4 /2 /4 (3) Without the boundar points, R (i) is the proportional to the transpose of T (i+), and this propert is ver useful in proving the convergence of the multigrid method. However, this propert is not strictl necessar to create an efficient multigrid algorithm. Other methods of interpolating and restricting are also possible, and it should also be noted that while a grid of size 2 n + has a particularl convenient multigrid formulation, the multigrid method can be applied to grids of arbitrar size. Given a differential operator matri A = A () on the top level, corresponding matrices on the lower levels can be calculated according to A (i) = R (i ) A (i ) T (i). (3) With this definition, a solution operator S(u (i), b (i) ) can be constructed as a single red black Gauss Seidel sweep. The multigrid algorithm is then given b function Multi(u (i), b (i) ) if i = g then compute eact solution to A (i) u (i) = b (i) return u (i) else for j = to v down do u (i) = S(u (i), b (i) ) end for r (i) = b (i) A (i) u (i) d (i) = T(Multi( (i+), R(r (i) ))) u (i) = u (i) + d (i) for j = to v up do u (i) = S(u (i), b (i) ) end for end if Here, (i+) represents a zero vector on grid level i +. The multigrid function is applied recursivel, and at each stage, the remainder of the problem on the level above is sent to the lower level. The algorithm starts on level, descends to level g, and then ascends to level again, following the shape of a V. It is therefore referred to as the multigrid V-ccle. Other more elaborate methods of moving between grids are possible, although the V-ccle is ver efficient in man situations. In the algorithm, we are free to choose the number of. 3

14 5 L 2 -norm error Jacobi Gauss Seidel Optimal SOR Multigrid, v down =, v up = 2 Multigrid, v down = v up = Effective iterations Figure 9: Errors versus the number of effective iterations for the several different iteration techniques. Here, to allow a direct comparison, effective iterations for the multigrid methods is defined b the number of Gauss Seidel iterations that are applied on the top grid level, since the Gauss Seidel iterations on the coarser grids are small in comparison. For the v down =, v up = 2 method, the effective number of iterations is twice the number of V-ccles. For the v down = v up = 2 method, the algorithm was slightl modified, so that onl two Gauss Seidel iterations were applied at the top level each time, instead of the epected four. Thus the effective number of iterations is also twice the number of V-ccles. The speed of the multigrid methods is startling when compared to an of the other three iterations. times the solution operator is applied on the wa down and on the wa up, and tpical good values to tr ma be v down = v up = 2, or even v down =, v up = 2. It ma also be worthwhile to carr out more iterations on the coarser grids, since the computation is much cheaper there. Figure 7 shows the restriction of the source term in the test problem on the coarser grids. Figure 8 shows the solution being successivel refined on the grids. Even after a single V-ccle, the solution is close to the eact answer. Figure 9 shows the computation times for two different multigrid algorithms, compared with the previous three methods considered. The multigrid algorithms reach numerical precision etremel quickl, much faster even than SOR. Onl twent Gauss Seidel iterations are applied at the top level before double numerical precision has been reached. A Code listings The following codes were used to generate the Jacobi, Gauss Seidel, and SOR diagrams in these notes. The are written in C++ and were compiled using the GNU C++ compiler. 4

15 Each of the first three routines calls a common code listed in appendi A.4 for setting up useful constants and defining common routines. This common code also contains a function for outputting the 2D matrices in a matri binar format that is readable b the plotting program Gnuplot []. This output routine could be replaced in order to save to different plotting programs. A. Jacobi method jacobi.cc // Load common routines and constants #include "common.cc" int main() { int i,j,ij,k; double error,u[m n],v[m n],z; double a, b; // Set initial guess to be identicall zero for(ij=;ij<m n;ij++) u[ij]=v[ij]=; output and error("jacobi out",u,); // Carr out Jacobi iterations for(k=;k<=total iters;k++) { // Alternatel flip input and output matrices if (k%2==) {a=u;b=v; else {a=v;b=u; // Compute Jacobi iteration for(j=;j<n ;j++) { for(i=;i<m ;i++) { ij=i+m j; a[ij]=(f(i,j)+dinv (b[ij ]+b[ij+]) +dinv (b[ij m]+b[ij+m])) dcent; // Save and compute error if necessar output and error("jacobi out",a,k); A.2 Gauss Seidel gsrb.cc // Load common routines and constants #include "common.cc" int main() { 5

16 int i,j,ij,k; double error,u[m n],z; // Set initial guess to be identicall zero for(ij=;ij<m n;ij++) u[ij]=; output and error("gsrb out",u,); // Compute Red Black Gauss Seidel iteration for(k=;k<=total iters;k++) { for(j=;j<n ;j++) { for(i=+(j&);i<m ;i+=2) { ij=i+m j; u[ij]=(f(i,j)+dinv (u[ij ]+u[ij+]) +dinv (u[ij m]+u[ij+m])) dcent; for(j=;j<n ;j++) { for(i=2 (j&);i<m ;i+=2) { ij=i+m j; u[ij]=(f(i,j)+dinv (u[ij ]+u[ij+]) +dinv (u[ij m]+u[ij+m])) dcent; // Save the result and compute error if necessar output and error("gsrb out",u,k); A.3 Successive Over-Relaation sor.cc // Load common routines and constants #include "common.cc" int main() { int i,j,ij,k; double error,u[m n],z; // Set initial guess to be identicall zero for(ij=;ij<m n;ij++) u[ij]=; output and error("sor out",u,); // Compute SOR Red Black iterations for(k=;k<=total iters;k++) { for(j=;j<n ;j++) { for(i=+(j&);i<m ;i+=2) { 6

17 ij=i+m j; u[ij]=u[ij] ( omega)+omega (f(i,j) +dinv (u[ij ]+u[ij+]) +dinv (u[ij m]+u[ij+m])) dcent; for(j=;j<n ;j++) { for(i=2 (j&);i<m ;i+=2) { ij=i+m j; u[ij]=u[ij] ( omega)+omega (f(i,j) +dinv (u[ij ]+u[ij+]) +dinv (u[ij m]+u[ij+m])) dcent; // Save the result and compute error if necessar output and error("sor out",u,k); A.4 Common routine for setup and output common.cc // Load standard libraries #include <cstdio> #include <cstdlib> #include <iostream> #include <fstream> #include <cmath> using namespace std; // Set grid size and number of iterations const int save iters=2; const int total iters=2; const int error ever=2; const int m=33,n=33; const double min=,ma=; const double min=,ma=; // Compute useful constants const double pi= ; const double omega=2/(+sin(2 pi/n)); const double d=(ma min)/(m ); const double d=(ma min)/(n ); const double dinv=/(d d); const double dinv=/(d d); const double dcent=/(2 (dinv+dinv)); 7

18 // Input function inline double f(int i,int j) { double =min+i d,=min+j d; return abs() >.5 abs()>.5?:; // Common output and error routine void output and error(char filename,double a,const int sn) { // Computes the error if sn%error ever== if(sn%error ever==) { double z,error=;int ij; for(int j=;j<n ;j++) { for(int i=;i<m ;i++) { ij=i+m j; z=f(i,j) a[ij] (2 dinv+2 dinv) +dinv (a[ij ]+a[ij+]) +dinv (a[ij m]+a[ij+m]); error+=z z; cout << sn << " " << error d d << endl; // Saves the matri if sn<=save iters if(sn<=save iters) { int i,j,ij=,ds=sizeof(float); float,,data float;const char pfloat; pfloat=(const char )&data float; ofstream outfile; static char fname[256]; sprintf(fname,"%s.%d",filename,sn); outfile.open(fname,fstream::out fstream::trunc fstream::binar); data float=m;outfile.write(pfloat,ds); for(i=;i<m;i++) { =min+i d; data float=;outfile.write(pfloat,ds); for(j=;j<n;j++) { =min+j d; data float=; 8

19 outfile.write(pfloat,ds); for(i=;i<m;i++) { data float=a[ij++]; outfile.write(pfloat,ds); outfile.close(); References [] [2] J. W. Demmel, Applied numerical linear algebra, SIAM, 997. [3] G. H. Golub and C. H. Van Loan, Matri computations, Johns Hopkins Universit Publishers,

Krylov Integration Factor Method on Sparse Grids for High Spatial Dimension Convection Diffusion Equations

Krylov Integration Factor Method on Sparse Grids for High Spatial Dimension Convection Diffusion Equations DOI.7/s95-6-6-7 Krlov Integration Factor Method on Sparse Grids for High Spatial Dimension Convection Diffusion Equations Dong Lu Yong-Tao Zhang Received: September 5 / Revised: 9 March 6 / Accepted: 8

More information

Lecture 18 Classical Iterative Methods

Lecture 18 Classical Iterative Methods Lecture 18 Classical Iterative Methods MIT 18.335J / 6.337J Introduction to Numerical Methods Per-Olof Persson November 14, 2006 1 Iterative Methods for Linear Systems Direct methods for solving Ax = b,

More information

x y plane is the plane in which the stresses act, yy xy xy Figure 3.5.1: non-zero stress components acting in the x y plane

x y plane is the plane in which the stresses act, yy xy xy Figure 3.5.1: non-zero stress components acting in the x y plane 3.5 Plane Stress This section is concerned with a special two-dimensional state of stress called plane stress. It is important for two reasons: () it arises in real components (particularl in thin components

More information

Kasetsart University Workshop. Multigrid methods: An introduction

Kasetsart University Workshop. Multigrid methods: An introduction Kasetsart University Workshop Multigrid methods: An introduction Dr. Anand Pardhanani Mathematics Department Earlham College Richmond, Indiana USA pardhan@earlham.edu A copy of these slides is available

More information

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for 1 Iteration basics Notes for 2016-11-07 An iterative solver for Ax = b is produces a sequence of approximations x (k) x. We always stop after finitely many steps, based on some convergence criterion, e.g.

More information

FRIEDRICH-ALEXANDER-UNIVERSITÄT ERLANGEN-NÜRNBERG. Lehrstuhl für Informatik 10 (Systemsimulation)

FRIEDRICH-ALEXANDER-UNIVERSITÄT ERLANGEN-NÜRNBERG. Lehrstuhl für Informatik 10 (Systemsimulation) FRIEDRICH-ALEXANDER-UNIVERSITÄT ERLANGEN-NÜRNBERG INSTITUT FÜR INFORMATIK (MATHEMATISCHE MASCHINEN UND DATENVERARBEITUNG) Lehrstuhl für Informatik 1 (Sstemsimulation) Efficient hierarchical grid coarsening

More information

Notes for CS542G (Iterative Solvers for Linear Systems)

Notes for CS542G (Iterative Solvers for Linear Systems) Notes for CS542G (Iterative Solvers for Linear Systems) Robert Bridson November 20, 2007 1 The Basics We re now looking at efficient ways to solve the linear system of equations Ax = b where in this course,

More information

Iterative Methods and Multigrid

Iterative Methods and Multigrid Iterative Methods and Multigrid Part 1: Introduction to Multigrid 1 12/02/09 MG02.prz Error Smoothing 1 0.9 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0 Initial Solution=-Error 0 10 20 30 40 50 60 70 80 90 100 DCT:

More information

Periodic Structures in FDTD

Periodic Structures in FDTD EE 5303 Electromagnetic Analsis Using Finite Difference Time Domain Lecture #19 Periodic Structures in FDTD Lecture 19 These notes ma contain coprighted material obtained under fair use rules. Distribution

More information

9. Iterative Methods for Large Linear Systems

9. Iterative Methods for Large Linear Systems EE507 - Computational Techniques for EE Jitkomut Songsiri 9. Iterative Methods for Large Linear Systems introduction splitting method Jacobi method Gauss-Seidel method successive overrelaxation (SOR) 9-1

More information

Mathematics 309 Conic sections and their applicationsn. Chapter 2. Quadric figures. ai,j x i x j + b i x i + c =0. 1. Coordinate changes

Mathematics 309 Conic sections and their applicationsn. Chapter 2. Quadric figures. ai,j x i x j + b i x i + c =0. 1. Coordinate changes Mathematics 309 Conic sections and their applicationsn Chapter 2. Quadric figures In this chapter want to outline quickl how to decide what figure associated in 2D and 3D to quadratic equations look like.

More information

We have examined power functions like f (x) = x 2. Interchanging x

We have examined power functions like f (x) = x 2. Interchanging x CHAPTER 5 Eponential and Logarithmic Functions We have eamined power functions like f =. Interchanging and ields a different function f =. This new function is radicall different from a power function

More information

Background. Background. C. T. Kelley NC State University tim C. T. Kelley Background NCSU, Spring / 58

Background. Background. C. T. Kelley NC State University tim C. T. Kelley Background NCSU, Spring / 58 Background C. T. Kelley NC State University tim kelley@ncsu.edu C. T. Kelley Background NCSU, Spring 2012 1 / 58 Notation vectors, matrices, norms l 1 : max col sum... spectral radius scaled integral norms

More information

Iterative Methods. Splitting Methods

Iterative Methods. Splitting Methods Iterative Methods Splitting Methods 1 Direct Methods Solving Ax = b using direct methods. Gaussian elimination (using LU decomposition) Variants of LU, including Crout and Doolittle Other decomposition

More information

Von Neumann Analysis of Jacobi and Gauss-Seidel Iterations

Von Neumann Analysis of Jacobi and Gauss-Seidel Iterations Von Neumann Analysis of Jacobi and Gauss-Seidel Iterations We consider the FDA to the 1D Poisson equation on a grid x covering [0,1] with uniform spacing h, h (u +1 u + u 1 ) = f whose exact solution (to

More information

6 = 1 2. The right endpoints of the subintervals are then 2 5, 3, 7 2, 4, 2 9, 5, while the left endpoints are 2, 5 2, 3, 7 2, 4, 9 2.

6 = 1 2. The right endpoints of the subintervals are then 2 5, 3, 7 2, 4, 2 9, 5, while the left endpoints are 2, 5 2, 3, 7 2, 4, 9 2. 5 THE ITEGRAL 5. Approimating and Computing Area Preliminar Questions. What are the right and left endpoints if [, 5] is divided into si subintervals? If the interval [, 5] is divided into si subintervals,

More information

Computational Linear Algebra

Computational Linear Algebra Computational Linear Algebra PD Dr. rer. nat. habil. Ralf-Peter Mundani Computation in Engineering / BGU Scientific Computing in Computer Science / INF Winter Term 2018/19 Part 4: Iterative Methods PD

More information

Introduction to 3D Game Programming with DirectX 9.0c: A Shader Approach

Introduction to 3D Game Programming with DirectX 9.0c: A Shader Approach Introduction to 3D Game Programming with DirectX 90c: A Shader Approach Part I Solutions Note : Please email to frank@moon-labscom if ou find an errors Note : Use onl after ou have tried, and struggled

More information

Multigrid Methods and their application in CFD

Multigrid Methods and their application in CFD Multigrid Methods and their application in CFD Michael Wurst TU München 16.06.2009 1 Multigrid Methods Definition Multigrid (MG) methods in numerical analysis are a group of algorithms for solving differential

More information

Computation of three-dimensional standing water waves

Computation of three-dimensional standing water waves Computation of three-dimensional standing water waves Chris H. Rcroft a,b,c,d, Jon Wilkening b,c a Department of Phsics, Universit of California, Berkele, CA 9472, United States b Department of Mathematics,

More information

FACULTY OF MATHEMATICAL STUDIES MATHEMATICS FOR PART I ENGINEERING. Lectures AB = BA = I,

FACULTY OF MATHEMATICAL STUDIES MATHEMATICS FOR PART I ENGINEERING. Lectures AB = BA = I, FACULTY OF MATHEMATICAL STUDIES MATHEMATICS FOR PART I ENGINEERING Lectures MODULE 7 MATRICES II Inverse of a matri Sstems of linear equations Solution of sets of linear equations elimination methods 4

More information

11.4 Polar Coordinates

11.4 Polar Coordinates 11. Polar Coordinates 917 11. Polar Coordinates In Section 1.1, we introduced the Cartesian coordinates of a point in the plane as a means of assigning ordered pairs of numbers to points in the plane.

More information

Analytic Trigonometry

Analytic Trigonometry CHAPTER 5 Analtic Trigonometr 5. Fundamental Identities 5. Proving Trigonometric Identities 5.3 Sum and Difference Identities 5.4 Multiple-Angle Identities 5.5 The Law of Sines 5.6 The Law of Cosines It

More information

Determinants. We said in Section 3.3 that a 2 2 matrix a b. Determinant of an n n Matrix

Determinants. We said in Section 3.3 that a 2 2 matrix a b. Determinant of an n n Matrix 3.6 Determinants We said in Section 3.3 that a 2 2 matri a b is invertible if and onl if its c d erminant, ad bc, is nonzero, and we saw the erminant used in the formula for the inverse of a 2 2 matri.

More information

5.3.3 The general solution for plane waves incident on a layered halfspace. The general solution to the Helmholz equation in rectangular coordinates

5.3.3 The general solution for plane waves incident on a layered halfspace. The general solution to the Helmholz equation in rectangular coordinates 5.3.3 The general solution for plane waves incident on a laered halfspace The general solution to the elmhol equation in rectangular coordinates The vector propagation constant Vector relationships between

More information

Scientific Computing WS 2018/2019. Lecture 9. Jürgen Fuhrmann Lecture 9 Slide 1

Scientific Computing WS 2018/2019. Lecture 9. Jürgen Fuhrmann Lecture 9 Slide 1 Scientific Computing WS 2018/2019 Lecture 9 Jürgen Fuhrmann juergen.fuhrmann@wias-berlin.de Lecture 9 Slide 1 Lecture 9 Slide 2 Simple iteration with preconditioning Idea: Aû = b iterative scheme û = û

More information

4 Strain true strain engineering strain plane strain strain transformation formulae

4 Strain true strain engineering strain plane strain strain transformation formulae 4 Strain The concept of strain is introduced in this Chapter. The approimation to the true strain of the engineering strain is discussed. The practical case of two dimensional plane strain is discussed,

More information

15. Eigenvalues, Eigenvectors

15. Eigenvalues, Eigenvectors 5 Eigenvalues, Eigenvectors Matri of a Linear Transformation Consider a linear ( transformation ) L : a b R 2 R 2 Suppose we know that L and L Then c d because of linearit, we can determine what L does

More information

MA3232 Numerical Analysis Week 9. James Cooley (1926-)

MA3232 Numerical Analysis Week 9. James Cooley (1926-) MA umerical Analysis Week 9 James Cooley (96-) James Cooley is an American mathematician. His most significant contribution to the world of mathematics and digital signal processing is the Fast Fourier

More information

4.7. Newton s Method. Procedure for Newton s Method HISTORICAL BIOGRAPHY

4.7. Newton s Method. Procedure for Newton s Method HISTORICAL BIOGRAPHY 4. Newton s Method 99 4. Newton s Method HISTORICAL BIOGRAPHY Niels Henrik Abel (18 189) One of the basic problems of mathematics is solving equations. Using the quadratic root formula, we know how to

More information

Numerical Methods Process Systems Engineering ITERATIVE METHODS. Numerical methods in chemical engineering Edwin Zondervan

Numerical Methods Process Systems Engineering ITERATIVE METHODS. Numerical methods in chemical engineering Edwin Zondervan IERAIVE MEHODS Numerical methods in chemical engineering Edwin Zondervan 1 OVERVIEW Iterative methods for large systems of equations We will solve Laplace s equation for steady state heat conduction LAPLACE

More information

16.5. Maclaurin and Taylor Series. Introduction. Prerequisites. Learning Outcomes

16.5. Maclaurin and Taylor Series. Introduction. Prerequisites. Learning Outcomes Maclaurin and Talor Series 6.5 Introduction In this Section we eamine how functions ma be epressed in terms of power series. This is an etremel useful wa of epressing a function since (as we shall see)

More information

8. BOOLEAN ALGEBRAS x x

8. BOOLEAN ALGEBRAS x x 8. BOOLEAN ALGEBRAS 8.1. Definition of a Boolean Algebra There are man sstems of interest to computing scientists that have a common underling structure. It makes sense to describe such a mathematical

More information

Sparse Linear Systems. Iterative Methods for Sparse Linear Systems. Motivation for Studying Sparse Linear Systems. Partial Differential Equations

Sparse Linear Systems. Iterative Methods for Sparse Linear Systems. Motivation for Studying Sparse Linear Systems. Partial Differential Equations Sparse Linear Systems Iterative Methods for Sparse Linear Systems Matrix Computations and Applications, Lecture C11 Fredrik Bengzon, Robert Söderlund We consider the problem of solving the linear system

More information

SEPARABLE EQUATIONS 2.2

SEPARABLE EQUATIONS 2.2 46 CHAPTER FIRST-ORDER DIFFERENTIAL EQUATIONS 4. Chemical Reactions When certain kinds of chemicals are combined, the rate at which the new compound is formed is modeled b the autonomous differential equation

More information

KEY IDEAS. Chapter 1 Function Transformations. 1.1 Horizontal and Vertical Translations Pre-Calculus 12 Student Workbook MHR 1

KEY IDEAS. Chapter 1 Function Transformations. 1.1 Horizontal and Vertical Translations Pre-Calculus 12 Student Workbook MHR 1 Chapter Function Transformations. Horizontal and Vertical Translations A translation can move the graph of a function up or down (vertical translation) and right or left (horizontal translation). A translation

More information

2.2 SEPARABLE VARIABLES

2.2 SEPARABLE VARIABLES 44 CHAPTER FIRST-ORDER DIFFERENTIAL EQUATIONS 6 Consider the autonomous DE 6 Use our ideas from Problem 5 to find intervals on the -ais for which solution curves are concave up and intervals for which

More information

Math 471 (Numerical methods) Chapter 3 (second half). System of equations

Math 471 (Numerical methods) Chapter 3 (second half). System of equations Math 47 (Numerical methods) Chapter 3 (second half). System of equations Overlap 3.5 3.8 of Bradie 3.5 LU factorization w/o pivoting. Motivation: ( ) A I Gaussian Elimination (U L ) where U is upper triangular

More information

RELATIONS AND FUNCTIONS through

RELATIONS AND FUNCTIONS through RELATIONS AND FUNCTIONS 11.1.2 through 11.1. Relations and Functions establish a correspondence between the input values (usuall ) and the output values (usuall ) according to the particular relation or

More information

10.2 The Unit Circle: Cosine and Sine

10.2 The Unit Circle: Cosine and Sine 0. The Unit Circle: Cosine and Sine 77 0. The Unit Circle: Cosine and Sine In Section 0.., we introduced circular motion and derived a formula which describes the linear velocit of an object moving on

More information

Improved Rotated Finite Difference Method for Solving Fractional Elliptic Partial Differential Equations

Improved Rotated Finite Difference Method for Solving Fractional Elliptic Partial Differential Equations American Scientific Researc Journal for Engineering, Tecnolog, and Sciences (ASRJETS) ISSN (Print) 33-44, ISSN (Online) 33-44 Global Societ of Scientific Researc and Researcers ttp://asrjetsjournal.org/

More information

Course Requirements. Course Mechanics. Projects & Exams. Homework. Week 1. Introduction. Fast Multipole Methods: Fundamentals & Applications

Course Requirements. Course Mechanics. Projects & Exams. Homework. Week 1. Introduction. Fast Multipole Methods: Fundamentals & Applications Week 1. Introduction. Fast Multipole Methods: Fundamentals & Applications Ramani Duraiswami Nail A. Gumerov What are multipole methods and what is this course about. Problems from phsics, mathematics,

More information

MAT 1275: Introduction to Mathematical Analysis. Graphs and Simplest Equations for Basic Trigonometric Functions. y=sin( x) Function

MAT 1275: Introduction to Mathematical Analysis. Graphs and Simplest Equations for Basic Trigonometric Functions. y=sin( x) Function MAT 275: Introduction to Mathematical Analsis Dr. A. Rozenblum Graphs and Simplest Equations for Basic Trigonometric Functions We consider here three basic functions: sine, cosine and tangent. For them,

More information

On the Extension of Goal-Oriented Error Estimation and Hierarchical Modeling to Discrete Lattice Models

On the Extension of Goal-Oriented Error Estimation and Hierarchical Modeling to Discrete Lattice Models On the Etension of Goal-Oriented Error Estimation and Hierarchical Modeling to Discrete Lattice Models J. T. Oden, S. Prudhomme, and P. Bauman Institute for Computational Engineering and Sciences The Universit

More information

Numerical Programming I (for CSE)

Numerical Programming I (for CSE) Technische Universität München WT 1/13 Fakultät für Mathematik Prof. Dr. M. Mehl B. Gatzhammer January 1, 13 Numerical Programming I (for CSE) Tutorial 1: Iterative Methods 1) Relaxation Methods a) Let

More information

Iterative Methods and Multigrid

Iterative Methods and Multigrid Iterative Methods and Multigrid Part 1: Introduction to Multigrid 2000 Eric de Sturler 1 12/02/09 MG01.prz Basic Iterative Methods (1) Nonlinear equation: f(x) = 0 Rewrite as x = F(x), and iterate x i+1

More information

CLASSICAL ITERATIVE METHODS

CLASSICAL ITERATIVE METHODS CLASSICAL ITERATIVE METHODS LONG CHEN In this notes we discuss classic iterative methods on solving the linear operator equation (1) Au = f, posed on a finite dimensional Hilbert space V = R N equipped

More information

Applications of Gauss-Radau and Gauss-Lobatto Numerical Integrations Over a Four Node Quadrilateral Finite Element

Applications of Gauss-Radau and Gauss-Lobatto Numerical Integrations Over a Four Node Quadrilateral Finite Element Avaiable online at www.banglaol.info angladesh J. Sci. Ind. Res. (), 77-86, 008 ANGLADESH JOURNAL OF SCIENTIFIC AND INDUSTRIAL RESEARCH CSIR E-mail: bsir07gmail.com Abstract Applications of Gauss-Radau

More information

Math 214 Spring problem set (a) Consider these two first order equations. (I) dy dx = x + 1 dy

Math 214 Spring problem set (a) Consider these two first order equations. (I) dy dx = x + 1 dy Math 4 Spring 08 problem set. (a) Consider these two first order equations. (I) d d = + d (II) d = Below are four direction fields. Match the differential equations above to their direction fields. Provide

More information

9.1 Preconditioned Krylov Subspace Methods

9.1 Preconditioned Krylov Subspace Methods Chapter 9 PRECONDITIONING 9.1 Preconditioned Krylov Subspace Methods 9.2 Preconditioned Conjugate Gradient 9.3 Preconditioned Generalized Minimal Residual 9.4 Relaxation Method Preconditioners 9.5 Incomplete

More information

8.1 Exponents and Roots

8.1 Exponents and Roots Section 8. Eponents and Roots 75 8. Eponents and Roots Before defining the net famil of functions, the eponential functions, we will need to discuss eponent notation in detail. As we shall see, eponents

More information

39. (a) Use trigonometric substitution to verify that. 40. The parabola y 2x divides the disk into two

39. (a) Use trigonometric substitution to verify that. 40. The parabola y 2x divides the disk into two 35. Prove the formula A r for the area of a sector of a circle with radius r and central angle. [Hint: Assume 0 and place the center of the circle at the origin so it has the equation. Then is the sum

More information

1 HOMOGENEOUS TRANSFORMATIONS

1 HOMOGENEOUS TRANSFORMATIONS HOMOGENEOUS TRANSFORMATIONS Purpose: The purpose of this chapter is to introduce ou to the Homogeneous Transformation. This simple 4 4 transformation is used in the geometr engines of CAD sstems and in

More information

10.5 Graphs of the Trigonometric Functions

10.5 Graphs of the Trigonometric Functions 790 Foundations of Trigonometr 0.5 Graphs of the Trigonometric Functions In this section, we return to our discussion of the circular (trigonometric functions as functions of real numbers and pick up where

More information

Iterative Methods for Solving A x = b

Iterative Methods for Solving A x = b Iterative Methods for Solving A x = b A good (free) online source for iterative methods for solving A x = b is given in the description of a set of iterative solvers called templates found at netlib: http

More information

CONTINUOUS SPATIAL DATA ANALYSIS

CONTINUOUS SPATIAL DATA ANALYSIS CONTINUOUS SPATIAL DATA ANALSIS 1. Overview of Spatial Stochastic Processes The ke difference between continuous spatial data and point patterns is that there is now assumed to be a meaningful value, s

More information

QUALITATIVE ANALYSIS OF DIFFERENTIAL EQUATIONS

QUALITATIVE ANALYSIS OF DIFFERENTIAL EQUATIONS arxiv:1803.0591v1 [math.gm] QUALITATIVE ANALYSIS OF DIFFERENTIAL EQUATIONS Aleander Panfilov stable spiral det A 6 3 5 4 non stable spiral D=0 stable node center non stable node saddle 1 tr A QUALITATIVE

More information

c 2016 Society for Industrial and Applied Mathematics

c 2016 Society for Industrial and Applied Mathematics SIAM J. SCI. COMPUT. Vol. 38, No. 2, pp. A765 A788 c 206 Societ for Industrial and Applied Mathematics ROOTS OF BIVARIATE POLYNOMIAL SYSTEMS VIA DETERMINANTAL REPRESENTATIONS BOR PLESTENJAK AND MICHIEL

More information

COMPACT IMPLICIT INTEGRATION FACTOR METHODS FOR A FAMILY OF SEMILINEAR FOURTH-ORDER PARABOLIC EQUATIONS. Lili Ju. Xinfeng Liu.

COMPACT IMPLICIT INTEGRATION FACTOR METHODS FOR A FAMILY OF SEMILINEAR FOURTH-ORDER PARABOLIC EQUATIONS. Lili Ju. Xinfeng Liu. DISCRETE AND CONTINUOUS doi:13934/dcdsb214191667 DYNAMICAL SYSTEMS SERIES B Volume 19, Number 6, August 214 pp 1667 1687 COMPACT IMPLICIT INTEGRATION FACTOR METHODS FOR A FAMILY OF SEMILINEAR FOURTH-ORDER

More information

6. Iterative Methods for Linear Systems. The stepwise approach to the solution...

6. Iterative Methods for Linear Systems. The stepwise approach to the solution... 6 Iterative Methods for Linear Systems The stepwise approach to the solution Miriam Mehl: 6 Iterative Methods for Linear Systems The stepwise approach to the solution, January 18, 2013 1 61 Large Sparse

More information

Chapter 7 Iterative Techniques in Matrix Algebra

Chapter 7 Iterative Techniques in Matrix Algebra Chapter 7 Iterative Techniques in Matrix Algebra Per-Olof Persson persson@berkeley.edu Department of Mathematics University of California, Berkeley Math 128B Numerical Analysis Vector Norms Definition

More information

Here is an example of a block diagonal matrix with Jordan Blocks on the diagonal: J

Here is an example of a block diagonal matrix with Jordan Blocks on the diagonal: J Class Notes 4: THE SPECTRAL RADIUS, NORM CONVERGENCE AND SOR. Math 639d Due Date: Feb. 7 (updated: February 5, 2018) In the first part of this week s reading, we will prove Theorem 2 of the previous class.

More information

Exact and Approximate Numbers:

Exact and Approximate Numbers: Eact and Approimate Numbers: The numbers that arise in technical applications are better described as eact numbers because there is not the sort of uncertainty in their values that was described above.

More information

The first change comes in how we associate operators with classical observables. In one dimension, we had. p p ˆ

The first change comes in how we associate operators with classical observables. In one dimension, we had. p p ˆ VI. Angular momentum Up to this point, we have been dealing primaril with one dimensional sstems. In practice, of course, most of the sstems we deal with live in three dimensions and 1D quantum mechanics

More information

Methods for Advanced Mathematics (C3) Coursework Numerical Methods

Methods for Advanced Mathematics (C3) Coursework Numerical Methods Woodhouse College 0 Page Introduction... 3 Terminolog... 3 Activit... 4 Wh use numerical methods?... Change of sign... Activit... 6 Interval Bisection... 7 Decimal Search... 8 Coursework Requirements on

More information

EP225 Note No. 4 Wave Motion

EP225 Note No. 4 Wave Motion EP5 Note No. 4 Wave Motion 4. Sinusoidal Waves, Wave Number Waves propagate in space in contrast to oscillations which are con ned in limited regions. In describing wave motion, spatial coordinates enter

More information

Chapter 6. Nonlinear Equations. 6.1 The Problem of Nonlinear Root-finding. 6.2 Rate of Convergence

Chapter 6. Nonlinear Equations. 6.1 The Problem of Nonlinear Root-finding. 6.2 Rate of Convergence Chapter 6 Nonlinear Equations 6. The Problem of Nonlinear Root-finding In this module we consider the problem of using numerical techniques to find the roots of nonlinear equations, f () =. Initially we

More information

Scientific Computing: An Introductory Survey

Scientific Computing: An Introductory Survey Scientific Computing: An Introductory Survey Chapter 11 Partial Differential Equations Prof. Michael T. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c 2002.

More information

DEN: Linear algebra numerical view (GEM: Gauss elimination method for reducing a full rank matrix to upper-triangular

DEN: Linear algebra numerical view (GEM: Gauss elimination method for reducing a full rank matrix to upper-triangular form) Given: matrix C = (c i,j ) n,m i,j=1 ODE and num math: Linear algebra (N) [lectures] c phabala 2016 DEN: Linear algebra numerical view (GEM: Gauss elimination method for reducing a full rank matrix

More information

Solving PDEs with Multigrid Methods p.1

Solving PDEs with Multigrid Methods p.1 Solving PDEs with Multigrid Methods Scott MacLachlan maclachl@colorado.edu Department of Applied Mathematics, University of Colorado at Boulder Solving PDEs with Multigrid Methods p.1 Support and Collaboration

More information

Introduction to Differential Equations. National Chiao Tung University Chun-Jen Tsai 9/14/2011

Introduction to Differential Equations. National Chiao Tung University Chun-Jen Tsai 9/14/2011 Introduction to Differential Equations National Chiao Tung Universit Chun-Jen Tsai 9/14/011 Differential Equations Definition: An equation containing the derivatives of one or more dependent variables,

More information

MMJ1153 COMPUTATIONAL METHOD IN SOLID MECHANICS PRELIMINARIES TO FEM

MMJ1153 COMPUTATIONAL METHOD IN SOLID MECHANICS PRELIMINARIES TO FEM B Course Content: A INTRODUCTION AND OVERVIEW Numerical method and Computer-Aided Engineering; Phsical problems; Mathematical models; Finite element method;. B Elements and nodes, natural coordinates,

More information

Elliptic Equations. Chapter Definitions. Contents. 4.2 Properties of Laplace s and Poisson s Equations

Elliptic Equations. Chapter Definitions. Contents. 4.2 Properties of Laplace s and Poisson s Equations 5 4. Properties of Laplace s and Poisson s Equations Chapter 4 Elliptic Equations Contents. Neumann conditions the normal derivative, / = n u is prescribed on the boundar second BP. In this case we have

More information

Lab 5 Forces Part 1. Physics 211 Lab. You will be using Newton s 2 nd Law to help you examine the nature of these forces.

Lab 5 Forces Part 1. Physics 211 Lab. You will be using Newton s 2 nd Law to help you examine the nature of these forces. b Lab 5 Forces Part 1 Phsics 211 Lab Introduction This is the first week of a two part lab that deals with forces and related concepts. A force is a push or a pull on an object that can be caused b a variet

More information

Trigonometric. equations. Topic: Periodic functions and applications. Simple trigonometric. equations. Equations using radians Further trigonometric

Trigonometric. equations. Topic: Periodic functions and applications. Simple trigonometric. equations. Equations using radians Further trigonometric Trigonometric equations 6 sllabusref eferenceence Topic: Periodic functions and applications In this cha 6A 6B 6C 6D 6E chapter Simple trigonometric equations Equations using radians Further trigonometric

More information

INTRODUCTION TO DIFFERENTIAL EQUATIONS

INTRODUCTION TO DIFFERENTIAL EQUATIONS INTRODUCTION TO DIFFERENTIAL EQUATIONS. Definitions and Terminolog. Initial-Value Problems.3 Differential Equations as Mathematical Models CHAPTER IN REVIEW The words differential and equations certainl

More information

MATHEMATICAL FUNDAMENTALS I. Michele Fitzpatrick

MATHEMATICAL FUNDAMENTALS I. Michele Fitzpatrick MTHEMTICL FUNDMENTLS I Michele Fitpatrick OVERVIEW Vectors and arras Matrices Linear algebra Del( operator Tensors DEFINITIONS vector is a single row or column of numbers. n arra is a collection of vectors

More information

The Conjugate Gradient Method

The Conjugate Gradient Method The Conjugate Gradient Method Classical Iterations We have a problem, We assume that the matrix comes from a discretization of a PDE. The best and most popular model problem is, The matrix will be as large

More information

COURSE Iterative methods for solving linear systems

COURSE Iterative methods for solving linear systems COURSE 0 4.3. Iterative methods for solving linear systems Because of round-off errors, direct methods become less efficient than iterative methods for large systems (>00 000 variables). An iterative scheme

More information

Additional Topics in Differential Equations

Additional Topics in Differential Equations 6 Additional Topics in Differential Equations 6. Eact First-Order Equations 6. Second-Order Homogeneous Linear Equations 6.3 Second-Order Nonhomogeneous Linear Equations 6.4 Series Solutions of Differential

More information

Stabilization and Acceleration of Algebraic Multigrid Method

Stabilization and Acceleration of Algebraic Multigrid Method Stabilization and Acceleration of Algebraic Multigrid Method Recursive Projection Algorithm A. Jemcov J.P. Maruszewski Fluent Inc. October 24, 2006 Outline 1 Need for Algorithm Stabilization and Acceleration

More information

Time-Frequency Analysis: Fourier Transforms and Wavelets

Time-Frequency Analysis: Fourier Transforms and Wavelets Chapter 4 Time-Frequenc Analsis: Fourier Transforms and Wavelets 4. Basics of Fourier Series 4.. Introduction Joseph Fourier (768-83) who gave his name to Fourier series, was not the first to use Fourier

More information

Hamiltonicity and Fault Tolerance

Hamiltonicity and Fault Tolerance Hamiltonicit and Fault Tolerance in the k-ar n-cube B Clifford R. Haithcock Portland State Universit Department of Mathematics and Statistics 006 In partial fulfillment of the requirements of the degree

More information

REVISION SHEET FP2 (MEI) CALCULUS. x x 0.5. x x 1.5. π π. Standard Calculus of Inverse Trig and Hyperbolic Trig Functions = + = + arcsin x = +

REVISION SHEET FP2 (MEI) CALCULUS. x x 0.5. x x 1.5. π π. Standard Calculus of Inverse Trig and Hyperbolic Trig Functions = + = + arcsin x = + the Further Mathematics network www.fmnetwork.org.uk V 07 REVISION SHEET FP (MEI) CALCULUS The main ideas are: Calculus using inverse trig functions & hperbolic trig functions and their inverses. Maclaurin

More information

Introduction to Multigrid Method

Introduction to Multigrid Method Introduction to Multigrid Metod Presented by: Bogojeska Jasmina /08/005 JASS, 005, St. Petersburg 1 Te ultimate upsot of MLAT Te amount of computational work sould be proportional to te amount of real

More information

520 Chapter 9. Nonlinear Differential Equations and Stability. dt =

520 Chapter 9. Nonlinear Differential Equations and Stability. dt = 5 Chapter 9. Nonlinear Differential Equations and Stabilit dt L dθ. g cos θ cos α Wh was the negative square root chosen in the last equation? (b) If T is the natural period of oscillation, derive the

More information

MAE 323: Chapter 4. Plane Stress and Plane Strain. The Stress Equilibrium Equation

MAE 323: Chapter 4. Plane Stress and Plane Strain. The Stress Equilibrium Equation The Stress Equilibrium Equation As we mentioned in Chapter 2, using the Galerkin formulation and a choice of shape functions, we can derive a discretized form of most differential equations. In Structural

More information

RANGE CONTROL MPC APPROACH FOR TWO-DIMENSIONAL SYSTEM 1

RANGE CONTROL MPC APPROACH FOR TWO-DIMENSIONAL SYSTEM 1 RANGE CONTROL MPC APPROACH FOR TWO-DIMENSIONAL SYSTEM Jirka Roubal Vladimír Havlena Department of Control Engineering, Facult of Electrical Engineering, Czech Technical Universit in Prague Karlovo náměstí

More information

Chapter 6. Self-Adjusting Data Structures

Chapter 6. Self-Adjusting Data Structures Chapter 6 Self-Adjusting Data Structures Chapter 5 describes a data structure that is able to achieve an epected quer time that is proportional to the entrop of the quer distribution. The most interesting

More information

INTRODUCTION TO MULTIGRID METHODS

INTRODUCTION TO MULTIGRID METHODS INTRODUCTION TO MULTIGRID METHODS LONG CHEN 1. ALGEBRAIC EQUATION OF TWO POINT BOUNDARY VALUE PROBLEM We consider the discretization of Poisson equation in one dimension: (1) u = f, x (0, 1) u(0) = u(1)

More information

Simplified Fast Multipole Methods for Micromagnetic Modeling

Simplified Fast Multipole Methods for Micromagnetic Modeling implified Fast Multipole Methods for Micromagnetic Modeling D. M. Apalkov and P. B.Visscher Department of Phsics and Astronom and MIT Center The Universit of Alabama This project was supported b F grants

More information

BASE VECTORS FOR SOLVING OF PARTIAL DIFFERENTIAL EQUATIONS

BASE VECTORS FOR SOLVING OF PARTIAL DIFFERENTIAL EQUATIONS BASE VECTORS FOR SOLVING OF PARTIAL DIFFERENTIAL EQUATIONS J. Roubal, V. Havlena Department of Control Engineering, Facult of Electrical Engineering, Czech Technical Universit in Prague Abstract The distributed

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) Lecture 19: Computing the SVD; Sparse Linear Systems Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical

More information

Solving Variable-Coefficient Fourth-Order ODEs with Polynomial Nonlinearity by Symmetric Homotopy Method

Solving Variable-Coefficient Fourth-Order ODEs with Polynomial Nonlinearity by Symmetric Homotopy Method Applied and Computational Mathematics 218; 7(2): 58-7 http://www.sciencepublishinggroup.com/j/acm doi: 1.11648/j.acm.21872.14 ISSN: 2328-565 (Print); ISSN: 2328-5613 (Online) Solving Variable-Coefficient

More information

Sample-based Optimal Transport and Barycenter Problems

Sample-based Optimal Transport and Barycenter Problems Sample-based Optimal Transport and Barcenter Problems MAX UANG New York Universit, Courant Institute of Mathematical Sciences AND ESTEBAN G. TABA New York Universit, Courant Institute of Mathematical Sciences

More information

Eigenvectors and Eigenvalues 1

Eigenvectors and Eigenvalues 1 Ma 2015 page 1 Eigenvectors and Eigenvalues 1 In this handout, we will eplore eigenvectors and eigenvalues. We will begin with an eploration, then provide some direct eplanation and worked eamples, and

More information

Lecture Note 7: Iterative methods for solving linear systems. Xiaoqun Zhang Shanghai Jiao Tong University

Lecture Note 7: Iterative methods for solving linear systems. Xiaoqun Zhang Shanghai Jiao Tong University Lecture Note 7: Iterative methods for solving linear systems Xiaoqun Zhang Shanghai Jiao Tong University Last updated: December 24, 2014 1.1 Review on linear algebra Norms of vectors and matrices vector

More information

Process Model Formulation and Solution, 3E4

Process Model Formulation and Solution, 3E4 Process Model Formulation and Solution, 3E4 Section B: Linear Algebraic Equations Instructor: Kevin Dunn dunnkg@mcmasterca Department of Chemical Engineering Course notes: Dr Benoît Chachuat 06 October

More information

CHAPTER 7 EARTHQUAKE RESPONSE OF INELASTIC SYSTEMS. Base shear force in a linearly elastic system due to ground excitation is Vb

CHAPTER 7 EARTHQUAKE RESPONSE OF INELASTIC SYSTEMS. Base shear force in a linearly elastic system due to ground excitation is Vb CHAPTER 7 EARTHQUAKE RESPONSE OF INELASTIC SYSTEMS Base shear force in a linearl elastic sstem due to ground excitation is Vb = ( A/ g) w where A is the pseudo-acceleration corresponding to natural period

More information

Linear Algebraic Equations

Linear Algebraic Equations Linear Algebraic Equations Linear Equations: a + a + a + a +... + a = c 11 1 12 2 13 3 14 4 1n n 1 a + a + a + a +... + a = c 21 2 2 23 3 24 4 2n n 2 a + a + a + a +... + a = c 31 1 32 2 33 3 34 4 3n n

More information