Computational methods of elementary particle physics
|
|
- Kerry Norman
- 5 years ago
- Views:
Transcription
1 Computational methods of elementary particle physics Markus Huber Lecture in SS017 Version: March 7, 017
2 Contents 0.0 Preface Introduction Overview of computational methods in elementary particle physics Overview of this course Basic techniques 7.1 Numerical integration Newton-Cotes formula Gauss quadrature Interpolation Polynomial interpolation Chebyshev interpolation Spline interpolation
3 0.0. PREFACE Contents 0.0 Preface These lecture notes are based on the lecture Computational methods of elementary particle physics held at the University of Graz in the summer term 017. I welcome everybody to send remarks or report any errors to markus.huber@unigraz.at. 3 Version: March 7, 017
4 1 Introduction 1.1 Overview of computational methods in elementary particle physics Computers play a crucial role in elementary particle physics, or in physics in general. As a matter of fact, physics was always pushing the development of faster and better computers. Computers were used from early ballistic calculations for military use to data analysis of events in particle collisions at the LHC. Nowadays they are omnipresent in modeling, data analysis, simulations and analytic calculations. Applications range from one-liners in Mathematica to complex simulations on super computers. Some examples for computer applications in elementary particle physics are the following: Lattice field theory: In field theory a central quantity is the path integral. E.g., for quantum chromodynamics (QCD) it reads: S[A, ψ,ψ] Z = D[A ψψ]e (1.1) The integral is over the elementary fields and S[A, ψ, ψ] is the action of QCD. It can be approximated by putting space-time on a lattice with lattice spacing a and length L, see Fig However, the remaining sum can still not be performed exactly and is calculated using Monte Carlo methods, viz., using a stochastic sampling of configuration space. Since continuum physics corresponds to a = 0 and L =, the corresponding limits have to be performed. In practice, this is done by repeating simulations with different values for a and L and extrapolating to the so-called continuum limit. Lattice simulations account for the most resource intensive calculations in theoretical physics, but their predictive power makes them an extremely useful tool. Monte Carlo event generators: It is not possible to describe collisions of particles fully from first principles because of the complexity of the perturbative part and difficulties in describing the nonperturbative part. However, this is necessary to confront experimental results with theoretical predictions and to separate signals from the background. Event generators produce large numbers of possible outcomes which are then compared to what is actually measured. Data analysis: Experiments as those carried out at the LHC produce immense amounts of data (estimated PB per year for the LHC). Processing and storing them requires large computational resources. In the case of the LHC this is handled via a worldwide network of computing centers, the so-called Worldwide LHC Computing Grid. It handles the distribution of the data and provides distributed computing resources. 4
5 1.1. OVERVIEW OF COMPUTATIONAL METHODS IN ELEMENTARY PARTICLE PHYSICS CHAPTER 1. INTRODUCTION Figure 1.1: Discretizing space-time on a rectangular lattice with lattice spacing a and lattice size L. Functional methods: Functional methods consist of integro-differential, differential or integral equations which, for most quantitative cases, need to be solved numerically. Compared to other methods they are far less resourceintensive. However, the size of the systems of equations solved increases steadily and with that the required computational resources. Depending on the problem, various architectures are used for computations. For some problems desktop PCs are sufficient. However, even then parallelization is advantageous due to the increasing number of cores in modern CPUs. Computer clusters naturally take advantage of parallelized programs. Clusters can have a few nodes, with several CPUs each, up to millions. In some cases it is even useful to run several single-core instances of the same program, e.g., for Monte Carlo methods when the same program is run over and over again to accumulate statistics. Possibilities for parallelization are MPI or OpenMP. In the last decade graphical processing units (GPUs) have become also very important. Programs are written with CUDA or OpenCL. Another possibility is Intel s Xeon Phi coprocessor, which offers a highly-parallel multicore environment similar to a GPU. Various programming languages are used in elementary particles physics. Historically, FORTRAN was always strong and, despite constant rumors to the contrary, it is still widely used. One reason is that many libraries, e.g., LAPACK, used in physics are written in FORTRAN. The fact that FORTRAN was heavily used in physics, or in science in general, is reflected in many physics friendly features, e.g., a complex number type was included early on. Another often used language is C++. For GPU and coprocessor architectures languages like CUDA and OpenCL are required. On the symbolic side Mathematica must be mentioned, which was founded by Stephen Wolfram, a particle physicist, in Special purpose packages can add additional functionalities useful for particle physicists. Special mention deserves the program FORM. It has a narrow focus on use in particle physics, where it is unrivaled in speed when it comes to perform large symbolic calculations. Finally, for (larger) projects additional auxiliary tools exist that can be very useful: make (for the compilation of the code), IDEs (integrated development environments, or at least editors that offer more support than syntax highlighting), version control 5 Version: March 7, 017
6 1.. OVERVIEW OF THIS COURSE CHAPTER 1. INTRODUCTION S(k ) Γ(p, P ) = K(k, p, P ) S(k + ) Γ(k, P ) Figure 1.: Diagrammatic representation of a meson BSE. systems (git can be considered state-of-the-art, but also subversion or others offer the basic features), debuggers, memory error detection programs and profilers. 1. Overview of this course In this course we will develop the methods necessary to calculate a meson mass in QCD. This will be done by solving the appropriate bound state equation, a Bethe- Salpeter equation (BSE, which, however, could also mean bound state equation), see Fig. 1.. This equation depends on several quantities. Some of them will be taken from models, another one, the quark propagator, will be calculated in this course. On the technical side this will require numeric integration, interpolation and solving an eigenvalue system. The complete code for this project will be developed step by step and should result in a final program. The following tasks need to be implemented which will be done in individual projects: 1. Functions for numeric integration. Functions for interpolation 3. Implementation of the quark propagator 4. Implementation of the BSE 5. Solution of the BSE through an eigenvalue system Programmes suitable for numeric computation should be used, e.g., FORTRAN, C, C++, Python, Julia. 6 Version: March 7, 017
7 Basic techniques.1 Numerical integration A numeric evaluation of an integral is necessary if its solution is not known in closed form or if the integrand itself is only known numerically. Depending on the problem, one can choose from many methods for numerical integration. Alternatively one speaks of (numerical) quadrature, which can refer to one- or multi-dimensional integration, whereas cubature is used for multi-dimensional integration only. In this section we will first look at some simple integration formulas to explain the basic idea and get some feeling for errors. Then we will turn towards integration techniques which are actually used for functional integrals..1.1 Newton-Cotes formula A simple integration method is the Newton-Cotes formula. To calculate an integral I one replaces the integrand f(x) by an approximating function p m (x), typically an m-th order polynomial: I = b a dx f(x) b a dx p m (x). (.1) The approximating function p m (x) is chosen such that it agrees with f(x) at a fixed number of points. For a polynomial this determines the values for the coefficients a i : p m (x) = a 0 + a 1 x a m x m + a m x m. (.) For practical calculations one splits the integration interval [a, b] into smaller intervals with width and defines the points h = x = b a n (.3) x i = a + i h, i = 0,..., n. (.4) The simplest approximation is to use a constant p m (x). This is called the rectangular rule. It is rarely used, because its truncation error is quite large for general functions. We can approximate the function f(x) in the interval x i x x i+1 either by f i = f(x i ) or f i+1 = f(x i+q ). The value of the integral in the interval [x i, x i+1 ] is then h f i and h f i+1, respectively. The total integral can be calculated correspondingly as n I left h f i (.5) i=0 7
8 .1. NUMERICAL INTEGRATION CHAPTER. BASIC TECHNIQUES or n I right h f i+1 = h i=0 n f i. (.6) These formulas over- or underestimate integrals of functions that increase or decrease monotonically. This behavior can be improved using the average of the two forms instead: n I r h i=0 i=1 f i + f i+1. (.7) As we will see, this is equivalent to the trapezoidal rule. The trapezoidal rule approximates the integrand piecewise by a polynomial of order one: p 1 (x) = a 0 + a 1 x. The area under the curve in the interval [x i, x i+1 ] is then given by the area of a trapezoid: h (f i + f i+1 )/. Summing up all intervals yields n I t h i=0 f i + f i+1 = h (f 0 + f f n + f n ). (.8) The error for a single segment is with this method given by E 1t = b a dx f(x) f(a) + f(b) (b a). (.9) Using a Taylor expansion around x = (b + a)/ with y = x x, we can perform the integral: b a dx f(x) = f(x) = f( x) + yf ( x) + y! f ( x) +..., (.10) h h ) dy (f( x) + yf ( x) + y! f ( x) +... = = h f( x) h3 f ( x) (.11) We use the Taylor expansion also at the endpoints of the interval: f(a) = f( x) h f ( x) + 1 ( ) h f ( x)..., f(b) = f( x) + h f ( x) + 1 ( ) h f ( x) (.1) Plugging all this into Eq. (.9) we obtain ( E 1t = h f( x) + 1 ) ( 4 h3 f ( x) +... h f( x) + 1 ) 8 h3 f ( x) h3 f ( x). (.13) 8 Version: March 7, 017
9 .1. NUMERICAL INTEGRATION CHAPTER. BASIC TECHNIQUES Thus the error is proportional to f ( x) and h 3. When a multisegment trapezoidal rule is used, the individual errors need to be summed: E t 1 ( ) 3 b a n f ( x i ), (.14) 1 n where x i is the midpoint between x i and x i+1. Using an average value for the second derivative, this can be rewritten as E t 1 1 (b a) ( b a n i=0 f = 1 n f ( x i ), (.15) n n=0 ) f = 1 1 (b a)h f. (.16) Thus, the error is O(h ), since b a is fixed. Repeating a similar analysis for the rectangular rules shows that the error is O(h): E r 1 (b a)hf. (.17) Exercise: Derive eq. Eq. (.17). To improve the precision of the results, the straightforward approach is to lower the step size h. However, the usefulness of this approach is limited by round-off errors becoming too large at some points. Another method is to increase the degree of the employed polynomial n. For n = this leads to Simpson s one-third rule, for n = 3 to Simpson s three-eight rule and for n = 4 to Boole s rule. An alternative possibility is to reduce the error by combining the results from calculations with different step sizes (Richardson s extrapolation). For example, consider the errors obtained for two step sizes h 1 and h with the trapezoidal method: E(h 1 ) 1 1 (b a)h 1f, (.18) E(h ) 1 1 (b a)h f. (.19) We assume that f is independent of the step size. The true value for the integral I can then be obtained from each calculation as I I(h 1 ) + c h 1, (.0) I I(h ) + c h, (.1) where I(h) is the result obtained by using the step size h. equations we can calculate the constant c: Plugging this into Eq. (.1) yields Equating these two c = 1 1 (b a)f I(h ) I(h 1 ). (.) h 1 h I I(h ) + I(h ) I(h 1 ) ( ). (.3) h 1 h 1 9 Version: March 7, 017
10 .1. NUMERICAL INTEGRATION CHAPTER. BASIC TECHNIQUES One can show that the error is now O(h 4 ). A particular application is to use h = h 1 /. This process can be used to combine results from calculations with different step sizes. For example, if we have the results I(h 1 ), I(h ) and I(h 3 ), we can combine I(h 1 ) and I(h ) to obtain a result I(h 1, h ) with an error O(h 4 ), but we can also combine I(h ) and I(h 3 ) to obtain a result I(h, h 3 ). Combining the two new results, we can get a result with an error O(h 6 ). Systematically applying this procedure is known as Romberg integration..1. Gauss quadrature The Newton-Cotes formula calculates an integral from values at equidistant points. Gauss quadrature, on the other hand, requires knowledge of the integrand at not equidistant given points. The general formula is c c 1 dx g(x)f(x) n w i f(x i ). (.4) i=1 Gauss quadrature does not only use the freedom to fix the weights w i, but also the positions x i where the function needs to be known. Due to this enlarged freedom, fewer function evaluations are required compared to the Newton-Cotes method. For a given value of n, the formula is exact for polynomials of degree n 1. Note that the function g(x) appears only on the left-hand side. It allows to make the quadrature exact for polynomials up to order n times the function g(x). The values of the positions x i and of the weights w i depend on the choice of g(x). For now we will choose g(x) = 1. This is called Gauss-Legendre quadrature, which is based on Legendre polynomials. The choice of g(x) also determines the values for the boundaries, c 1 and c. For Gauss-Legendre they are set to c 1 = and c = 1. Hence, an integral with different boundaries must be transformed to this interval. In principle any transformation from an interval z [a, b] is possible that transforms this interval to x [, 1]. For example, the linear transformation reads x = z a b, (.5) b a where z is the original variable. The inverse transformation is z = (b a)x + a + b. (.6) When such a transformation of variables is performed, the Jacobian of the transformation must also be included: Thus, a general integral is calculated as dz = b a dx. (.7) b a dzf(z) b a n w i f(z(x i )). (.8) i 10 Version: March 7, 017
11 .1. NUMERICAL INTEGRATION CHAPTER. BASIC TECHNIQUES The Jacobian can be taken into account directly in the weights. For integrands that stretch over several orders of magnitude a logarithmic transformation is advantageous: x = A + B ln z, z = e x A B. (.9) The coefficients A and B are determined by the conditions x(a) = and x(b) = 1 as A = ln(a b) ln(b/a), B = ln(b/a). (.30) The Jacobian is dz = e x A B B dx. (.31) Let us first consider the two-point formula, viz., the result is exact for polynomials of order 3 and below: dxf(x) = w 1 f(x 1 ) + w f(x ). (.3) From the requirement that it should be exact for f(x) equal to 1, x, x and x 3 we obtain These four conditions leads to dx = = w 1 + w, dx x = 0 = w 1 x 1 + w x, dx x = 3 = w 1 x 1 + w x dx x 3 = 0 = w 1 x w x 3. (.33) w 1 = w = 1, x 1 = x = 1 3. (.34) Note the symmetry x 1 = x. As Simpson s three-eight rule, Eq. (.3) is exact for polynomials up to order three. However, here less evaluations of the function are required ( vs. 4). For higher n the determination of the x i and w i is done using Legendre polynomials, for useful relation see, e.g., [3]. They are given by P 0 (x) = 1, P 1 (x) = x, P n (x) = n 1 x P n (x) n 1 n n P n (x), n =, 3,... (.35) 11 Version: March 7, 017
12 .1. NUMERICAL INTEGRATION CHAPTER. BASIC TECHNIQUES and fulfill the orthogonality relation The weights w i are w i = dxp n (x)p m (x) = δ nm n + 1. (.36) (1 x i )P n(x i ) = (1 x i ) (n + 1) P n+1 (x i ). (.37) The derivative can be calculated from the recurrence relation (1 x )P n(x) = (n + 1)xP n (x) (n + 1)P n+1 (x). (.38) The values for x i correspond to the zeros of the Legendre polynomial of order n: P n (x i ) = 0. There are exactly n zeros. For finding the zeros, the standard Newton method is sufficient, viz., by iterating x i = x i f(x i) f (x i ) (.39) until x i x i < ɛ. ɛ must be chosen sufficiently small, e.g., By using a good initial guess x 0, this method converges in a few steps. For the Legendre polynomials a good initial guess is x j,0 = cos(π (j 0.5)/(n + 0.5)), where j labels the zero. In principle tabulated values for the weights and zeros can be used. However, more flexibility is achieved when they are calculated in the program. The Gauss-Legendre quadrature has an error estimate of E GL n+1 (n!) 4 (n + 1)((n)!) 3 f (n) (x), < ξ < 1. (.40) From this is can be seen that polynomials of order up to n 1 is evaluated exactly, since the derivative f (n) vanishes in that case. Also, convergence is much faster than for Newton-Cotes. In principle it is possible to construct the weights and zeros for a generic function g(x), see []. Some choices lead to well-known orthogonal polynomials. The most useful cases are: g(x) Polynomials Boundaries 1 Legendre [, 1] 1/ 1 x Chebyshev first kind [, 1] 1 x Chebyshev second kind [, 1] e x Laguerre [0, ] e x Hermite [, ] A typical application for Gauss-Chebyshev quadrature is the integration over angles. For example, with u = cos(θ) the following integral, where a factor sin(θ) appears from the integral measure, can be rewritten as π 0 sin(θ) dθ = 1 u du. (.41) 1 Version: March 7, 017
13 .. INTERPOLATION CHAPTER. BASIC TECHNIQUES The generalization to multi-dimensional integrals is straightforward. becomes then a multiple sum. For example: dxdy f(x, y) = n i=1 The sum m wi x w y j f(x i, y i ). (.4) For many applications a fixed order of the quadrature is sufficient. One can easily find out how many points are needed by increasing their number until the result does not change anymore and stay with that number. Alternatively, adaptive integration algorithms exist that rely on subdividing the integral into smaller integrals until a certain convergence criterion is reached. Ideally one can reuse the points already calculated. Gauss-Kronrod quadrature is an example of such an adaptive integration.. Interpolation The dressing function we calculate with functional equations are only known at discrete values. However, in the equations they are typically required for arbitrary arguments. One can try to choose the arguments, where the dressings are calculated, such that they coincide with the arguments used in the equations, but this only works in some cases. Typically we need a way to describe the function also away from the calculated points. This is done via interpolation. In general one can distinguish between two ways to describe a function known at discrete points. One is to try to approximate the general trend of the available data. This is useful when it is known that the data itself has significant error or noise, for example, when it comes from an experimental measurement. Approximating the data in such a generic way is called least-squares regression. Another approach is to make the approximating function pass through every known point. This is done when the data is very precise and known as interpolation. In the following we will only treat this case. Once a method for interpolation is defined, one can determine all the required parameters and use it also outside of the data. This is then known as extrapolation. If no further information about the behavior of the function outside the domain of the available data is known, this is a very delicate procedure and should always be done with caution...1 Polynomial interpolation As we did for integration, one can use a polynomial to describe data. If n + 1 points are known, a polynomial of order n is required to make the polynomial pass through the data: j=1 f(x) = y a 0 + a 1 x a n x n. (.43) A naive implementation requires the following system of equations to be solved for the coefficients a i : f(x i ) = y i = a 0 + a 1 x i a n x n i, i = 0, 1,,..., n. (.44) This system has a unique solution, but typically solving it is not very efficient and problems like an ill-conditioned matrices can appear. 13 Version: March 7, 017
14 .. INTERPOLATION CHAPTER. BASIC TECHNIQUES e -x x Figure.1: Lagrange fit of e x using 101 points between 0 and 100. Clearly some artifacts are seen for high x. If another representation for the polynomial is chosen, the determination of the coefficients is easier. Lagrange interpolation uses this representation: n n f(x) = y (x x j ). (.45) For example, for n =, this yields i=0 b i j=0,j i f(x) = y b 0 (x x 1 )(x x ) + b 1 (x x 0 )(x x ) + b (x x 0 )(x x 1 ). (.46) The coefficients can easily be determined from the f(x i ): f(x 0 ) = y 0 = b 0 (x 0 x 1 )(x 0 x ), f(x 1 ) = y 1 = b 1 (x 1 x 0 )(x 1 x ), f(x ) = y = b (x x 0 )(x x 1 ). (.47) Solving for the b i and plugging these expressions back into Eq. (.46) yields (x x 1 )(x x ) f(x) = y y 0 (x 0 x 1 )(x 0 x ) + y (x x 0 )(x x ) 1 (x 1 x 0 )(x 1 x ) + y (x x 0 )(x x 1 ) (x x 0 )(x x 1 ). For general n the formula is f(x) = y n i=0 y i n j=0,j i (.48) x x j x i x j. (.49) If the function to be described is a polynomial of order n or less, using polynomials is exact. However, for the Lagrange it is difficult to estimate the error in general. Another disadvantage of the Lagrange method is that results of a computation cannot be reused for interpolation at another point. Also, if another point is added to the known data, the complete calculation has to be redone. A method that overcomes the last two problems is using a Newton polynomial. The approximating polynomial is in this case written as f(x) = y c 0 + c 1 (x x 0 ) + c (x x 0 )(x x 1 ) c n (x x 0 )(x x 1 ) (x x n ). (.50) 14 Version: March 7, 017
15 .. INTERPOLATION CHAPTER. BASIC TECHNIQUES The coefficients c i can be determined recursively starting with a constant interpolation c 0 (x 0 ) = f(x 0 ). Adding the point x 1 one determine the coefficient for linear interpolation: c 1 (x 1, x 0 ) = f(x 1) f(x 0 ) x 1 x 0. (.51) The next point x is used to fix the quadratic coefficient: ( 1 f(x ) f(x 1 ) c (x, x 1, x 0 ) = f(x ) 1) f(x 0 ). (.5) x x 0 x x 1 x 1 x 0 The i-th coefficient can be expressed as: c i (x i, x i,..., x 0 ) = c i(x i, x i,..., x 1 ) c i (x i,..., x 0 ) x i x 0. (.53) The recursive definitions of the c i are known as divided differences and hence this method is also known as Newton s divided-difference interpolation polynomial. The Newton polynomial, Eq. (.50), is in structure similar to a Taylor series. The error can thus be estimated from the next order n + 1: E n Newton c n+1 (x n+1,..., x 0 )(x x 0 ) (x x n ). (.54) Since we already used up all data points, we create a data point (x n+1, y n+1 ) from the polynomial itself from an arbitrarily chosen point x n+1... Chebyshev interpolation One problem of the interpolation methods mentioned up to now is that the error of interpolation is larger at the ends of the domain. To have a uniformly distributed error, one can use orthogonal polynomials as approximating functions. This also belongs to the class of polynomial interpolations. Chebyshev polynomials The recursive definition of Chebyshev polynomials is ( x 1): P 0 (x) = 1, (.55) P 1 (x) = x, (.56) P k (x) = xp k (x) P k (x), k =, 3,... (.57) Chebyshev polynomials can be expressed in terms of cosines: P k (x) = cos(k θ) (.58) with cos θ = x. The equivalence can be seen from P 0 (x) = cos 0 = 1, P 1 (x) = cos(cos x) = x and the trigonometric identity cos((k + 1)θ) + cos((k 1)θ) = cos θ cos(k θ), (.59) 15 Version: March 7, 017
16 .. INTERPOLATION CHAPTER. BASIC TECHNIQUES which yields the recursive relation Eq. (.57). The orthogonality relation of the polynomials is dx P n(x)p m (x) 0 n m = π n = m = 0 1 x π n = m 0 (.60) The k zeros x j can be determined from the zeros of the cosine at ±π/, ±3π/,... in Eq. (.58) as a k j + 1 k θ = π, j = 1,,... k, (.61) ( ) k j + 1 x j = cos π. (.6) k a The indices are put such that the zeros are ordered from (k 1)π/ to π/ which in turn yields the zeros in x from low to high. In principle, one can create a standard polynomial by expressing single polynomials via Chebyshev polynomials. For example: 1 = P 0 (x), (.63) x = P 1 (x), (.64) x = 1 (P 0(x) + P (x)). (.65) However, it is more convenient to make use of the orthogonality properties of the polynomials and write it as f(x) = y d 0 + d i P i (x), (.66) where the coefficients d i can be determined from Eq. (.60): d i = π i=1 dx f(x)p i(x), i = 0, 1,.... (.67) 1 x 16 Version: March 7, 017
17 Bibliography [1] Singiresu S. Rao, Applied numerical methods for engineers and scientists, Prentice Hall, New Jersey, 00. [] William H. Press et al., Numerical recipes - The art of scientific computing, Cambridge University Press, Cambridge, 007. [3] Weisstein, Eric W., Legendre-Gauss Quadrature. From MathWorld A Wolfram Web Resource, Legendre-GaussQuadrature.html 17
CS 450 Numerical Analysis. Chapter 8: Numerical Integration and Differentiation
Lecture slides based on the textbook Scientific Computing: An Introductory Survey by Michael T. Heath, copyright c 2018 by the Society for Industrial and Applied Mathematics. http://www.siam.org/books/cl80
More informationIntegration, differentiation, and root finding. Phys 420/580 Lecture 7
Integration, differentiation, and root finding Phys 420/580 Lecture 7 Numerical integration Compute an approximation to the definite integral I = b Find area under the curve in the interval Trapezoid Rule:
More informationJim Lambers MAT 460/560 Fall Semester Practice Final Exam
Jim Lambers MAT 460/560 Fall Semester 2009-10 Practice Final Exam 1. Let f(x) = sin 2x + cos 2x. (a) Write down the 2nd Taylor polynomial P 2 (x) of f(x) centered around x 0 = 0. (b) Write down the corresponding
More informationDifferentiation and Integration
Differentiation and Integration (Lectures on Numerical Analysis for Economists II) Jesús Fernández-Villaverde 1 and Pablo Guerrón 2 February 12, 2018 1 University of Pennsylvania 2 Boston College Motivation
More informationNumerical integration and differentiation. Unit IV. Numerical Integration and Differentiation. Plan of attack. Numerical integration.
Unit IV Numerical Integration and Differentiation Numerical integration and differentiation quadrature classical formulas for equally spaced nodes improper integrals Gaussian quadrature and orthogonal
More informationNumerical Integration (Quadrature) Another application for our interpolation tools!
Numerical Integration (Quadrature) Another application for our interpolation tools! Integration: Area under a curve Curve = data or function Integrating data Finite number of data points spacing specified
More informationChapter 5: Numerical Integration and Differentiation
Chapter 5: Numerical Integration and Differentiation PART I: Numerical Integration Newton-Cotes Integration Formulas The idea of Newton-Cotes formulas is to replace a complicated function or tabulated
More informationCOURSE Numerical integration of functions (continuation) 3.3. The Romberg s iterative generation method
COURSE 7 3. Numerical integration of functions (continuation) 3.3. The Romberg s iterative generation method The presence of derivatives in the remainder difficulties in applicability to practical problems
More informationNumerical Integra/on
Numerical Integra/on Applica/ons The Trapezoidal Rule is a technique to approximate the definite integral where For 1 st order: f(a) f(b) a b Error Es/mate of Trapezoidal Rule Truncation error: From Newton-Gregory
More informationPHYS-4007/5007: Computational Physics Course Lecture Notes Appendix G
PHYS-4007/5007: Computational Physics Course Lecture Notes Appendix G Dr. Donald G. Luttermoser East Tennessee State University Version 7.0 Abstract These class notes are designed for use of the instructor
More informationFixed point iteration and root finding
Fixed point iteration and root finding The sign function is defined as x > 0 sign(x) = 0 x = 0 x < 0. It can be evaluated via an iteration which is useful for some problems. One such iteration is given
More informationScientific Computing: An Introductory Survey
Scientific Computing: An Introductory Survey Chapter 7 Interpolation Prof. Michael T. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c 2002. Reproduction permitted
More informationRomberg Integration and Gaussian Quadrature
Romberg Integration and Gaussian Quadrature P. Sam Johnson October 17, 014 P. Sam Johnson (NITK) Romberg Integration and Gaussian Quadrature October 17, 014 1 / 19 Overview We discuss two methods for integration.
More informationarxiv: v1 [physics.comp-ph] 22 Jul 2010
Gaussian integration with rescaling of abscissas and weights arxiv:007.38v [physics.comp-ph] 22 Jul 200 A. Odrzywolek M. Smoluchowski Institute of Physics, Jagiellonian University, Cracov, Poland Abstract
More information(f(x) P 3 (x)) dx. (a) The Lagrange formula for the error is given by
1. QUESTION (a) Given a nth degree Taylor polynomial P n (x) of a function f(x), expanded about x = x 0, write down the Lagrange formula for the truncation error, carefully defining all its elements. How
More informationScientific Computing: Numerical Integration
Scientific Computing: Numerical Integration Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 Course MATH-GA.2043 or CSCI-GA.2112, Fall 2015 Nov 5th, 2015 A. Donev (Courant Institute) Lecture
More informationyou expect to encounter difficulties when trying to solve A x = b? 4. A composite quadrature rule has error associated with it in the following form
Qualifying exam for numerical analysis (Spring 2017) Show your work for full credit. If you are unable to solve some part, attempt the subsequent parts. 1. Consider the following finite difference: f (0)
More informationIn numerical analysis quadrature refers to the computation of definite integrals.
Numerical Quadrature In numerical analysis quadrature refers to the computation of definite integrals. f(x) a x i x i+1 x i+2 b x A traditional way to perform numerical integration is to take a piece of
More informationEAD 115. Numerical Solution of Engineering and Scientific Problems. David M. Rocke Department of Applied Science
EAD 115 Numerical Solution of Engineering and Scientific Problems David M. Rocke Department of Applied Science Multidimensional Unconstrained Optimization Suppose we have a function f() of more than one
More information(0, 0), (1, ), (2, ), (3, ), (4, ), (5, ), (6, ).
1 Interpolation: The method of constructing new data points within the range of a finite set of known data points That is if (x i, y i ), i = 1, N are known, with y i the dependent variable and x i [x
More informationScientific Computing
2301678 Scientific Computing Chapter 2 Interpolation and Approximation Paisan Nakmahachalasint Paisan.N@chula.ac.th Chapter 2 Interpolation and Approximation p. 1/66 Contents 1. Polynomial interpolation
More informationExamination paper for TMA4215 Numerical Mathematics
Department of Mathematical Sciences Examination paper for TMA425 Numerical Mathematics Academic contact during examination: Trond Kvamsdal Phone: 93058702 Examination date: 6th of December 207 Examination
More informationCh. 03 Numerical Quadrature. Andrea Mignone Physics Department, University of Torino AA
Ch. 03 Numerical Quadrature Andrea Mignone Physics Department, University of Torino AA 2017-2018 Numerical Quadrature In numerical analysis quadrature refers to the computation of definite integrals. y
More informationNumerical Methods. King Saud University
Numerical Methods King Saud University Aims In this lecture, we will... find the approximate solutions of derivative (first- and second-order) and antiderivative (definite integral only). Numerical Differentiation
More informationNumerical Analysis. A Comprehensive Introduction. H. R. Schwarz University of Zürich Switzerland. with a contribution by
Numerical Analysis A Comprehensive Introduction H. R. Schwarz University of Zürich Switzerland with a contribution by J. Waldvogel Swiss Federal Institute of Technology, Zürich JOHN WILEY & SONS Chichester
More informationComputing Gauss-Jacobi quadrature rules
.... Computing Gauss-Jacobi quadrature rules Alex Townsend Joint work with Nick Hale NA Internal Seminar Trinity Term 2012 . Gauss Jacobi Quadrature An n-point Gauss Jacobi quadrature rule: 1 1 w(x)f (x)dx
More informationx x2 2 + x3 3 x4 3. Use the divided-difference method to find a polynomial of least degree that fits the values shown: (b)
Numerical Methods - PROBLEMS. The Taylor series, about the origin, for log( + x) is x x2 2 + x3 3 x4 4 + Find an upper bound on the magnitude of the truncation error on the interval x.5 when log( + x)
More informationOutline. 1 Interpolation. 2 Polynomial Interpolation. 3 Piecewise Polynomial Interpolation
Outline Interpolation 1 Interpolation 2 3 Michael T. Heath Scientific Computing 2 / 56 Interpolation Motivation Choosing Interpolant Existence and Uniqueness Basic interpolation problem: for given data
More informationInterpolation. Chapter Interpolation. 7.2 Existence, Uniqueness and conditioning
76 Chapter 7 Interpolation 7.1 Interpolation Definition 7.1.1. Interpolation of a given function f defined on an interval [a,b] by a polynomial p: Given a set of specified points {(t i,y i } n with {t
More informationOutline. 1 Numerical Integration. 2 Numerical Differentiation. 3 Richardson Extrapolation
Outline Numerical Integration Numerical Differentiation Numerical Integration Numerical Differentiation 3 Michael T. Heath Scientific Computing / 6 Main Ideas Quadrature based on polynomial interpolation:
More informationNumerical Methods I: Numerical Integration/Quadrature
1/20 Numerical Methods I: Numerical Integration/Quadrature Georg Stadler Courant Institute, NYU stadler@cims.nyu.edu November 30, 2017 umerical integration 2/20 We want to approximate the definite integral
More informationDepartment of Applied Mathematics and Theoretical Physics. AMA 204 Numerical analysis. Exam Winter 2004
Department of Applied Mathematics and Theoretical Physics AMA 204 Numerical analysis Exam Winter 2004 The best six answers will be credited All questions carry equal marks Answer all parts of each question
More informationMathematics for Engineers. Numerical mathematics
Mathematics for Engineers Numerical mathematics Integers Determine the largest representable integer with the intmax command. intmax ans = int32 2147483647 2147483647+1 ans = 2.1475e+09 Remark The set
More informationM.SC. PHYSICS - II YEAR
MANONMANIAM SUNDARANAR UNIVERSITY DIRECTORATE OF DISTANCE & CONTINUING EDUCATION TIRUNELVELI 627012, TAMIL NADU M.SC. PHYSICS - II YEAR DKP26 - NUMERICAL METHODS (From the academic year 2016-17) Most Student
More information3. Numerical Quadrature. Where analytical abilities end...
3. Numerical Quadrature Where analytical abilities end... Numerisches Programmieren, Hans-Joachim Bungartz page 1 of 32 3.1. Preliminary Remarks The Integration Problem Numerical quadrature denotes numerical
More informationNUMERICAL METHODS. x n+1 = 2x n x 2 n. In particular: which of them gives faster convergence, and why? [Work to four decimal places.
NUMERICAL METHODS 1. Rearranging the equation x 3 =.5 gives the iterative formula x n+1 = g(x n ), where g(x) = (2x 2 ) 1. (a) Starting with x = 1, compute the x n up to n = 6, and describe what is happening.
More informationBindel, Spring 2012 Intro to Scientific Computing (CS 3220) Week 12: Monday, Apr 16. f(x) dx,
Panel integration Week 12: Monday, Apr 16 Suppose we want to compute the integral b a f(x) dx In estimating a derivative, it makes sense to use a locally accurate approximation to the function around the
More informationPreface. 2 Linear Equations and Eigenvalue Problem 22
Contents Preface xv 1 Errors in Computation 1 1.1 Introduction 1 1.2 Floating Point Representation of Number 1 1.3 Binary Numbers 2 1.3.1 Binary number representation in computer 3 1.4 Significant Digits
More informationLECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel
LECTURE NOTES on ELEMENTARY NUMERICAL METHODS Eusebius Doedel TABLE OF CONTENTS Vector and Matrix Norms 1 Banach Lemma 20 The Numerical Solution of Linear Systems 25 Gauss Elimination 25 Operation Count
More informationApplied Numerical Analysis (AE2220-I) R. Klees and R.P. Dwight
Applied Numerical Analysis (AE0-I) R. Klees and R.P. Dwight February 018 Contents 1 Preliminaries: Motivation, Computer arithmetic, Taylor series 1 1.1 Numerical Analysis Motivation..........................
More informationPHYS 404 Lecture 1: Legendre Functions
PHYS 404 Lecture 1: Legendre Functions Dr. Vasileios Lempesis PHYS 404 - LECTURE 1 DR. V. LEMPESIS 1 Legendre Functions physical justification Legendre functions or Legendre polynomials are the solutions
More informationTABLE OF CONTENTS INTRODUCTION, APPROXIMATION & ERRORS 1. Chapter Introduction to numerical methods 1 Multiple-choice test 7 Problem set 9
TABLE OF CONTENTS INTRODUCTION, APPROXIMATION & ERRORS 1 Chapter 01.01 Introduction to numerical methods 1 Multiple-choice test 7 Problem set 9 Chapter 01.02 Measuring errors 11 True error 11 Relative
More informationChapter 4: Interpolation and Approximation. October 28, 2005
Chapter 4: Interpolation and Approximation October 28, 2005 Outline 1 2.4 Linear Interpolation 2 4.1 Lagrange Interpolation 3 4.2 Newton Interpolation and Divided Differences 4 4.3 Interpolation Error
More informationMA2501 Numerical Methods Spring 2015
Norwegian University of Science and Technology Department of Mathematics MA5 Numerical Methods Spring 5 Solutions to exercise set 9 Find approximate values of the following integrals using the adaptive
More informationPhysics 115/242 Romberg Integration
Physics 5/242 Romberg Integration Peter Young In this handout we will see how, starting from the trapezium rule, we can obtain much more accurate values for the integral by repeatedly eliminating the leading
More informationNUMERICAL COMPUTATION IN SCIENCE AND ENGINEERING
NUMERICAL COMPUTATION IN SCIENCE AND ENGINEERING C. Pozrikidis University of California, San Diego New York Oxford OXFORD UNIVERSITY PRESS 1998 CONTENTS Preface ix Pseudocode Language Commands xi 1 Numerical
More informationNumerical Analysis Exam with Solutions
Numerical Analysis Exam with Solutions Richard T. Bumby Fall 000 June 13, 001 You are expected to have books, notes and calculators available, but computers of telephones are not to be used during the
More informationINTERPOLATION. and y i = cos x i, i = 0, 1, 2 This gives us the three points. Now find a quadratic polynomial. p(x) = a 0 + a 1 x + a 2 x 2.
INTERPOLATION Interpolation is a process of finding a formula (often a polynomial) whose graph will pass through a given set of points (x, y). As an example, consider defining and x 0 = 0, x 1 = π/4, x
More informationSection 6.6 Gaussian Quadrature
Section 6.6 Gaussian Quadrature Key Terms: Method of undetermined coefficients Nonlinear systems Gaussian quadrature Error Legendre polynomials Inner product Adapted from http://pathfinder.scar.utoronto.ca/~dyer/csca57/book_p/node44.html
More informationIntroduction to numerical computations on the GPU
Introduction to numerical computations on the GPU Lucian Covaci http://lucian.covaci.org/cuda.pdf Tuesday 1 November 11 1 2 Outline: NVIDIA Tesla and Geforce video cards: architecture CUDA - C: programming
More informationCS 257: Numerical Methods
CS 57: Numerical Methods Final Exam Study Guide Version 1.00 Created by Charles Feng http://www.fenguin.net CS 57: Numerical Methods Final Exam Study Guide 1 Contents 1 Introductory Matter 3 1.1 Calculus
More informationFunction Approximation
1 Function Approximation This is page i Printer: Opaque this 1.1 Introduction In this chapter we discuss approximating functional forms. Both in econometric and in numerical problems, the need for an approximating
More informationA THEORETICAL INTRODUCTION TO NUMERICAL ANALYSIS
A THEORETICAL INTRODUCTION TO NUMERICAL ANALYSIS Victor S. Ryaben'kii Semyon V. Tsynkov Chapman &. Hall/CRC Taylor & Francis Group Boca Raton London New York Chapman & Hall/CRC is an imprint of the Taylor
More information4. Numerical Quadrature. Where analytical abilities end... continued
4. Numerical Quadrature Where analytical abilities end... continued Where analytical abilities end... continued, November 30, 22 1 4.3. Extrapolation Increasing the Order Using Linear Combinations Once
More informationNumerical Integra/on
Numerical Integra/on The Trapezoidal Rule is a technique to approximate the definite integral where For 1 st order: f(a) f(b) a b Error Es/mate of Trapezoidal Rule Truncation error: From Newton-Gregory
More informationInfinite Series. 1 Introduction. 2 General discussion on convergence
Infinite Series 1 Introduction I will only cover a few topics in this lecture, choosing to discuss those which I have used over the years. The text covers substantially more material and is available for
More informationNumerical Methods. Scientists. Engineers
Third Edition Numerical Methods for Scientists and Engineers K. Sankara Rao Numerical Methods for Scientists and Engineers Numerical Methods for Scientists and Engineers Third Edition K. SANKARA RAO Formerly,
More informationIntegration. Topic: Trapezoidal Rule. Major: General Engineering. Author: Autar Kaw, Charlie Barker.
Integration Topic: Trapezoidal Rule Major: General Engineering Author: Autar Kaw, Charlie Barker 1 What is Integration Integration: The process of measuring the area under a function plotted on a graph.
More informationWe consider the problem of finding a polynomial that interpolates a given set of values:
Chapter 5 Interpolation 5. Polynomial Interpolation We consider the problem of finding a polynomial that interpolates a given set of values: x x 0 x... x n y y 0 y... y n where the x i are all distinct.
More informationReview. Numerical Methods Lecture 22. Prof. Jinbo Bi CSE, UConn
Review Taylor Series and Error Analysis Roots of Equations Linear Algebraic Equations Optimization Numerical Differentiation and Integration Ordinary Differential Equations Partial Differential Equations
More informationMA2501 Numerical Methods Spring 2015
Norwegian University of Science and Technology Department of Mathematics MA2501 Numerical Methods Spring 2015 Solutions to exercise set 7 1 Cf. Cheney and Kincaid, Exercise 4.1.9 Consider the data points
More informationSolving PDEs with PGI CUDA Fortran Part 4: Initial value problems for ordinary differential equations
Solving PDEs with PGI CUDA Fortran Part 4: Initial value problems for ordinary differential equations Outline ODEs and initial conditions. Explicit and implicit Euler methods. Runge-Kutta methods. Multistep
More informationAy190 Computational Astrophysics
CALIFORNIA INSTITUTE OF TECHNOLOGY Division of Physics, Mathematics, and Astronomy Ay190 Computational Astrophysics Christian D. Ott and Andrew Benson cott@tapir.caltech.edu, abenson@tapir.caltech.edu
More informationLectures 9-10: Polynomial and piecewise polynomial interpolation
Lectures 9-1: Polynomial and piecewise polynomial interpolation Let f be a function, which is only known at the nodes x 1, x,, x n, ie, all we know about the function f are its values y j = f(x j ), j
More informationIntroduction to Numerical Analysis
Introduction to Numerical Analysis S. Baskar and S. Sivaji Ganesh Department of Mathematics Indian Institute of Technology Bombay Powai, Mumbai 400 076. Introduction to Numerical Analysis Lecture Notes
More informationNumerical Methods for Engineers
Numerical Methods for Engineers SEVENTH EDITION Steven C Chopra Berger Chair in Computing and Engineering Tufts University Raymond P. Canal Professor Emeritus of Civil Engineering of Michiaan University
More informationNUMERICAL METHODS FOR ENGINEERING APPLICATION
NUMERICAL METHODS FOR ENGINEERING APPLICATION Second Edition JOEL H. FERZIGER A Wiley-Interscience Publication JOHN WILEY & SONS, INC. New York / Chichester / Weinheim / Brisbane / Singapore / Toronto
More informationLECTURE 16 GAUSS QUADRATURE In general for Newton-Cotes (equispaced interpolation points/ data points/ integration points/ nodes).
CE 025 - Lecture 6 LECTURE 6 GAUSS QUADRATURE In general for ewton-cotes (equispaced interpolation points/ data points/ integration points/ nodes). x E x S fx dx hw' o f o + w' f + + w' f + E 84 f 0 f
More informationGENG2140, S2, 2012 Week 7: Curve fitting
GENG2140, S2, 2012 Week 7: Curve fitting Curve fitting is the process of constructing a curve, or mathematical function, f(x) that has the best fit to a series of data points Involves fitting lines and
More informationCOURSE Numerical integration of functions
COURSE 6 3. Numerical integration of functions The need: for evaluating definite integrals of functions that has no explicit antiderivatives or whose antiderivatives are not easy to obtain. Let f : [a,
More informationNumerical techniques to solve equations
Programming for Applications in Geomatics, Physical Geography and Ecosystem Science (NGEN13) Numerical techniques to solve equations vaughan.phillips@nateko.lu.se Vaughan Phillips Associate Professor,
More informationNumerical Analysis: Approximation of Functions
Numerical Analysis: Approximation of Functions Mirko Navara http://cmp.felk.cvut.cz/ navara/ Center for Machine Perception, Department of Cybernetics, FEE, CTU Karlovo náměstí, building G, office 104a
More informationIntroduction to Uncertainty Quantification in Computational Science Handout #3
Introduction to Uncertainty Quantification in Computational Science Handout #3 Gianluca Iaccarino Department of Mechanical Engineering Stanford University June 29 - July 1, 2009 Scuola di Dottorato di
More informationNumerical Methods in Physics and Astrophysics
Kostas Kokkotas 2 October 20, 2014 2 http://www.tat.physik.uni-tuebingen.de/ kokkotas Kostas Kokkotas 3 TOPICS 1. Solving nonlinear equations 2. Solving linear systems of equations 3. Interpolation, approximation
More informationAM205: Assignment 3 (due 5 PM, October 20)
AM25: Assignment 3 (due 5 PM, October 2) For this assignment, first complete problems 1, 2, 3, and 4, and then complete either problem 5 (on theory) or problem 6 (on an application). If you submit answers
More informationStat 451 Lecture Notes Numerical Integration
Stat 451 Lecture Notes 03 12 Numerical Integration Ryan Martin UIC www.math.uic.edu/~rgmartin 1 Based on Chapter 5 in Givens & Hoeting, and Chapters 4 & 18 of Lange 2 Updated: February 11, 2016 1 / 29
More informationNumerical Methods for Engineers. and Scientists. Applications using MATLAB. An Introduction with. Vish- Subramaniam. Third Edition. Amos Gilat.
Numerical Methods for Engineers An Introduction with and Scientists Applications using MATLAB Third Edition Amos Gilat Vish- Subramaniam Department of Mechanical Engineering The Ohio State University Wiley
More informationNumerical Integration exact integration is not needed to achieve the optimal convergence rate of nite element solutions ([, 9, 11], and Chapter 7). In
Chapter 6 Numerical Integration 6.1 Introduction After transformation to a canonical element,typical integrals in the element stiness or mass matrices (cf. (5.5.8)) have the forms Q = T ( )N s Nt det(j
More informationEngg. Math. II (Unit-IV) Numerical Analysis
Dr. Satish Shukla of 33 Engg. Math. II (Unit-IV) Numerical Analysis Syllabus. Interpolation and Curve Fitting: Introduction to Interpolation; Calculus of Finite Differences; Finite Difference and Divided
More informationIntroductory Numerical Analysis
Introductory Numerical Analysis Lecture Notes December 16, 017 Contents 1 Introduction to 1 11 Floating Point Numbers 1 1 Computational Errors 13 Algorithm 3 14 Calculus Review 3 Root Finding 5 1 Bisection
More informationExam 2. Average: 85.6 Median: 87.0 Maximum: Minimum: 55.0 Standard Deviation: Numerical Methods Fall 2011 Lecture 20
Exam 2 Average: 85.6 Median: 87.0 Maximum: 100.0 Minimum: 55.0 Standard Deviation: 10.42 Fall 2011 1 Today s class Multiple Variable Linear Regression Polynomial Interpolation Lagrange Interpolation Newton
More informationApproximation theory
Approximation theory Xiaojing Ye, Math & Stat, Georgia State University Spring 2019 Numerical Analysis II Xiaojing Ye, Math & Stat, Georgia State University 1 1 1.3 6 8.8 2 3.5 7 10.1 Least 3squares 4.2
More informationComputational Methods
Numerical Computational Methods Revised Edition P. B. Patil U. P. Verma Alpha Science International Ltd. Oxford, U.K. CONTENTS Preface List ofprograms v vii 1. NUMER1CAL METHOD, ERROR AND ALGORITHM 1 1.1
More informationSimpson s 1/3 Rule Simpson s 1/3 rule assumes 3 equispaced data/interpolation/integration points
CE 05 - Lecture 5 LECTURE 5 UMERICAL ITEGRATIO COTIUED Simpson s / Rule Simpson s / rule assumes equispaced data/interpolation/integration points Te integration rule is based on approximating fx using
More informationNumerical Analysis: Interpolation Part 1
Numerical Analysis: Interpolation Part 1 Computer Science, Ben-Gurion University (slides based mostly on Prof. Ben-Shahar s notes) 2018/2019, Fall Semester BGU CS Interpolation (ver. 1.00) AY 2018/2019,
More informationNumerical Methods in Physics and Astrophysics
Kostas Kokkotas 2 October 17, 2017 2 http://www.tat.physik.uni-tuebingen.de/ kokkotas Kostas Kokkotas 3 TOPICS 1. Solving nonlinear equations 2. Solving linear systems of equations 3. Interpolation, approximation
More informationSlow Growth for Gauss Legendre Sparse Grids
Slow Growth for Gauss Legendre Sparse Grids John Burkardt, Clayton Webster April 4, 2014 Abstract A sparse grid for multidimensional quadrature can be constructed from products of 1D rules. For multidimensional
More information3.1 Interpolation and the Lagrange Polynomial
MATH 4073 Chapter 3 Interpolation and Polynomial Approximation Fall 2003 1 Consider a sample x x 0 x 1 x n y y 0 y 1 y n. Can we get a function out of discrete data above that gives a reasonable estimate
More informationContents. I Basic Methods 13
Preface xiii 1 Introduction 1 I Basic Methods 13 2 Convergent and Divergent Series 15 2.1 Introduction... 15 2.1.1 Power series: First steps... 15 2.1.2 Further practical aspects... 17 2.2 Differential
More informationSTOCHASTIC SAMPLING METHODS
STOCHASTIC SAMPLING METHODS APPROXIMATING QUANTITIES OF INTEREST USING SAMPLING METHODS Recall that quantities of interest often require the evaluation of stochastic integrals of functions of the solutions
More informationResearch Computing with Python, Lecture 7, Numerical Integration and Solving Ordinary Differential Equations
Research Computing with Python, Lecture 7, Numerical Integration and Solving Ordinary Differential Equations Ramses van Zon SciNet HPC Consortium November 25, 2014 Ramses van Zon (SciNet HPC Consortium)Research
More information(x x 0 )(x x 1 )... (x x n ) (x x 0 ) + y 0.
> 5. Numerical Integration Review of Interpolation Find p n (x) with p n (x j ) = y j, j = 0, 1,,..., n. Solution: p n (x) = y 0 l 0 (x) + y 1 l 1 (x) +... + y n l n (x), l k (x) = n j=1,j k Theorem Let
More informationNumerical Analysis. Introduction to. Rostam K. Saeed Karwan H.F. Jwamer Faraidun K. Hamasalh
Iraq Kurdistan Region Ministry of Higher Education and Scientific Research University of Sulaimani Faculty of Science and Science Education School of Science Education-Mathematics Department Introduction
More informationApplied Numerical Analysis
Applied Numerical Analysis Using MATLAB Second Edition Laurene V. Fausett Texas A&M University-Commerce PEARSON Prentice Hall Upper Saddle River, NJ 07458 Contents Preface xi 1 Foundations 1 1.1 Introductory
More informationNumerical Methods with MATLAB
Numerical Methods with MATLAB A Resource for Scientists and Engineers G. J. BÖRSE Lehigh University PWS Publishing Company I(T)P AN!NTERNATIONAL THOMSON PUBLISHING COMPANY Boston Albany Bonn Cincinnati
More informationChapter 5 - Quadrature
Chapter 5 - Quadrature The Goal: Adaptive Quadrature Historically in mathematics, quadrature refers to the act of trying to find a square with the same area of a given circle. In mathematical computing,
More informationNUMERICAL METHODS USING MATLAB
NUMERICAL METHODS USING MATLAB Dr John Penny George Lindfield Department of Mechanical Engineering, Aston University ELLIS HORWOOD NEW YORK LONDON TORONTO SYDNEY TOKYO SINGAPORE Preface 1 An introduction
More informationInterpolation. P. Sam Johnson. January 30, P. Sam Johnson (NITK) Interpolation January 30, / 75
Interpolation P. Sam Johnson January 30, 2015 P. Sam Johnson (NITK) Interpolation January 30, 2015 1 / 75 Overview One of the basic ideas in Mathematics is that of a function and most useful tool of numerical
More informationNUMERICAL ANALYSIS I. MARTIN LOTZ School of Mathematics The University of Manchester. May 2016
NUMERICAL ANALYSIS I by MARTIN LOTZ School of Mathematics The University of Manchester May 06 Contents Contents ii Week. Computational Complexity....................... Accuracy...............................
More informationNumerical interpolation, extrapolation and fi tting of data
Chapter 6 Numerical interpolation, extrapolation and fi tting of data 6.1 Introduction Numerical interpolation and extrapolation is perhaps one of the most used tools in numerical applications to physics.
More information