The answer in each case is the error in evaluating the taylor series for ln(1 x) for x = which is 6.9.

Size: px
Start display at page:

Download "The answer in each case is the error in evaluating the taylor series for ln(1 x) for x = which is 6.9."

Transcription

1 Brad Nelson Math 26 Homework #2 /23/2. a MATLAB outputs: >> a=(+3.4e-6)-.e-6;a- ans = 4.449e-6 >> a=+(3.4e-6-.e-6);a- ans = 2.224e-6 And the exact answer for both operations is 2.3e-6. The reason why way I is so inaccurate is that the number +3.4e-6 is stored on the computer as ( )( + ɛ mach ) which corresponds to an error of magnitude 6 in double precision. Then the floating point operation ( ). 6 is accurate to (( ). 6 )( + ɛ mach ), which also introduces an error of magnitude 6. The final step, a, doesn t introduce significant error into this calculation, since the resulting number is on magnitude of 6 so the error is of the magnitude 32. The two operations which introduce error of magnitude 6 result in an inaccurate answer since the real answer is of the same magnitude. Way II is more accurate since ( = ( + ɛ mach ) only has an error 32, which is not significant for this calculation. Then = ( )( + ɛ mach ) introduces an error of magnitude 6 which is significant since the final answer is of this magnitude. Once again, the final operation, a does not introduce significant error. While the returned answer is inaccurate because the introduced error was of the same magnitude as the answer, there was only one such step that introduced such a large error, which compared to two such steps in way I, is more accurate. b MATLAB outputs: >> x =.999; a = ; for j=:6, a = a-x^j/j; end; a-log(-x) ans = 3.553e-3 >> x =.999; a = ; for j=6:-:, a = a-x^j/j; end; a-log(-x) ans = e-6 The answer in each case is the error in evaluating the taylor series for ln( x) for x =.999 which is For way I, the series is evaluated as a = j For way II, a = j 6 j. 6 j j= j= Looking at way I, the first term evaluated is.999 and the second term is.5. Adding these two terms together introduces an error of.5ɛ mach which is of the magnitude 6. Since the partial sum monotonically approaches , it is always of magnitude. This means that adding each term to the series introduces an error of 6, which compounded over 6 terms turns into a total error of 2. The actual observed error

2 is close to this, at 3, and may be smaller than the prediction due to errors in opposite directions from the true answer canceling each other out. As for way II, the first term evaluated is of order 3. Performing operations on numbers of this magnitude will introduce a relative error of 6, which is an absolute error of 47. Since these errors are so small, the introduced error doesn t reach a magnitude of 6 until towards the end of the calculation, so the compounded error is still 6. From these observations, we can conclude that when summing together a list of numbers using floating-point arithmetic, it is best to start with the smallest numbers first. This way the total error of the summation stays small relative to the size of the answer until the last terms are added, which prevents a large relative error from compounding over many iterations. 2. a f (x) = 2x computed via x x. On a machine, this algorithm is f (x) = f l(x) f l(x). To prove backwards stability, we want to show that x X, f x x (x) = f ( x) for some x : = O(ɛ x mach ). To see if this is possible, we set f (x) = (x( + ɛ ) + x( + ɛ ))( + ɛ 2 ) = x( + ɛ 3 ) + x( + ɛ 3 ) where ɛ,2 ɛ mach. It follows that ɛ 3 = ɛ + ɛ 2 + ɛ ɛ 2. Thus, ɛ 3 2ɛ mach + ɛ 2 mach, and ɛ 2 = O(ɛ mach ). Thus, the algorithm is backwards stable. b f (x) = + x computed via x. Set: f (x) = ( + x( + ɛ ))( + ɛ 2 ) = + x( + ɛ 3 ) = f ( x) where ɛ,2 ɛ mach. Ignoring high-order terms (because of their small size), we have ɛ 3 = ɛ + ɛ 2 + ɛ 2 x Since for choices of x, ɛ 3 will blow up, the algorithm is not backwards stable. However to test stability, consider f (x) and f ( x). If we force ɛ 3 < ɛ mach, then f (x) f ( x) f (x) = ɛ 2 + x(ɛ + ɛ 2 ɛ 3 ɛ ɛ 2 ) + x( + ɛ 3 ) We see that for x, the algorithm is backwards stable, so focusing on x <, if x is still close to, the numerator is O(ɛ mach ), while the denominator 2, thus f (x) f ( x) f = (x) O(ɛ mach ). For x, the numerator is ɛ 2 and the denominator is, so f (x) f ( x) f = (x) O(ɛ mach ). Thus, while the algorithm is not backwards stable, it is stable. 3. Computing the eigenvalues of a symmetric matrix via the characteristic polynomial: i Restricting to 2x2 diagonal matrices with diagonal entries a and b, the characteristic polynomial is λ 2 (a + b)λ + ab. If we choose A = I, then we have λ 2 2λ +. Since the coefficients of this polynomial are stored with error ɛ mach, the polynomial can be written ( + ɛ a )λ 2 2( + ɛ b )λ + ( + ɛ c ) where ɛ a,b,c ɛ mach. To find the roots of this polynomial using the quadratic formula, we have, λ = 2( + ɛ b) ± 4( + ɛ b ) 2 4( + ɛ a )( + ɛ c ) 2( + ɛ a ) By expanding the inside of the square root, we see that the largest order error terms will be O( ɛ mach ). 2

3 ii While the above was done for A = I, it is easy to see that for any quadratic polynomial ax 2 + bx + c, perturbation of O(ɛ mach ) in the coefficients will lead to a perturbation of O( ɛ mach ) in the roots of a quadratic polynomial by placing the coefficients in the quadratic formula. Thus if ɛ mach = 6, then there will be an error of O( 8 ) in the roots of the polynomial. In general, for a degree n polynomial, we would expect to see O( n ɛ mach ) if a similar method is used. Since n ɛ goes to slower than ɛ, f (x) f ( x) O(ɛ mach ) for any x x f (x) x : = O(ɛ x mach ), and the algorithm is not stable. Since stability is a weaker condition than backwards stable, it is also not backwards stable. Thus, the algorithm is unstable. 4. Code: lagrange.m, lagrange eq.m. The following plot was produced: 5 lagrange basis functions for 2 equally spaced points on [,] Note how for each node, all polynomials except for one go to, and the one that does not goes to. 5. Code: interp error.m, hw2fns.m The two functions in the following questions look like: 3

4 f (x) = sin(2e x ) f (x) = ( + 25x 2 ) The following two plots used 25 equally spaced points on [, ] to create an interpolated polynomial for each function. The interpolation error E n (x) := f (x) (L n f )(x) is plotted. 4

5 .5 x 9 interpolation error for sin(2e x ) (a) Note that the error is at most 9. f (x) = sin(2e x ) 3 interpolation error for (+25*x 2 ) (b) f (x) = ( + 25x 2 ) Note that the error blows up to 25 at edges. 5

6 interpolation error as function of n points Sup norm of the interpolation error for f (x) = sin(2e x ) as a function of n equally spaced points on [, ]. Note how line appears to be roughly linear between n= and 27. Since this is on a semilog plot, this indicates exponential convergence in this region. To explain the reverse trend when n > 27, we note that sup x [,],k {...n} l k (x) grows exponentially: 9 sup of n lagrange basis functions on [,] Sup norm over all n Lagrange basis functions on [, ] as a function of n. Since the supremum is growing exponentially, the supremum in the error of the sum over the 6

7 these functions is going to grow exponentially as ɛ mach sup x [,],k {...n} l k (x), following the rules of floating point arithmetic. Note that at n 27, the supremum is 6, so the supremum of the error will be if ɛ mach = 6. This is where the graph of the sup norm of the interpolation error above suddenly reverses course, and begins to be less accurate, as the increase in error due to the sup norm of the basis functions overtakes the decrease in error due to adding more terms to the interpolant. Code: lbasis sup.m 6. Code: interp error cheby.m, lagrange cheby.m After updating the codes to use Chebychev nodes instead of equally spaced nodes, the following plots were produced:.2 lagrange basis functions for 2 Chebychev points on [,] Lagrange basis functions for 2 Chebychev nodes on [, ]. By inspecting the graph, it is clear that for each l k (x), sup x [,] l k (x). 7

8 5 x 2 interpolation error for sin(2e x ) interpolation error for f (x) = sin(2e x ) plotted for 25 Chebychev points on [, ]. Maximum error is 2, which is better than for equally-spaced points. x 3 interpolation error for (+25*x 2 ) interpolation error for f (x) = ( + 25x 2 ) plotted for 25 Chebychev points on [, ]. The maximum error for this function is now 9 3, which is a drastic improvement over the scenario with equally-spaced nodes. 8

9 interpolation error as function of n points Sup norm of the interpolation error for f (x) = sin(2e x ) as a function of n Chebychev points on [, ]. interpolation error as function of n points Sup norm of the interpolation error for f (x) = ( + 25x 2 ) as a function of n Chebychev points on [, ]. Note that the interpolation error is converging exponentially to. Thus, the Runge phenomenon does not persist. Comparing the graphs for the sup norm of the interpolation error of f (x) = sin(2e x ) for Chebychev points and equally spaced points, it appears that the Chebychev points allow a better achievable interpolation error. For equally spaced points, the best possible error is 9

10 about while Chebychev points have a best possible error of 5. We can t reasonably expect the error to get much better, as this is close to ɛ mach, since the maximum for this function is.

Numerical Algorithms. IE 496 Lecture 20

Numerical Algorithms. IE 496 Lecture 20 Numerical Algorithms IE 496 Lecture 20 Reading for This Lecture Primary Miller and Boxer, Pages 124-128 Forsythe and Mohler, Sections 1 and 2 Numerical Algorithms Numerical Analysis So far, we have looked

More information

Lecture Notes 7, Math/Comp 128, Math 250

Lecture Notes 7, Math/Comp 128, Math 250 Lecture Notes 7, Math/Comp 128, Math 250 Misha Kilmer Tufts University October 23, 2005 Floating Point Arithmetic We talked last time about how the computer represents floating point numbers. In a floating

More information

Two hours. To be provided by Examinations Office: Mathematical Formula Tables. THE UNIVERSITY OF MANCHESTER. 29 May :45 11:45

Two hours. To be provided by Examinations Office: Mathematical Formula Tables. THE UNIVERSITY OF MANCHESTER. 29 May :45 11:45 Two hours MATH20602 To be provided by Examinations Office: Mathematical Formula Tables. THE UNIVERSITY OF MANCHESTER NUMERICAL ANALYSIS 1 29 May 2015 9:45 11:45 Answer THREE of the FOUR questions. If more

More information

Interpolation and the Lagrange Polynomial

Interpolation and the Lagrange Polynomial Interpolation and the Lagrange Polynomial MATH 375 J. Robert Buchanan Department of Mathematics Fall 2013 Introduction We often choose polynomials to approximate other classes of functions. Theorem (Weierstrass

More information

Lecture 7. Floating point arithmetic and stability

Lecture 7. Floating point arithmetic and stability Lecture 7 Floating point arithmetic and stability 2.5 Machine representation of numbers Scientific notation: 23 }{{} }{{} } 3.14159265 {{} }{{} 10 sign mantissa base exponent (significand) s m β e A floating

More information

Math 2142 Homework 5 Part 1 Solutions

Math 2142 Homework 5 Part 1 Solutions Math 2142 Homework 5 Part 1 Solutions Problem 1. For the following homogeneous second order differential equations, give the general solution and the particular solution satisfying the given initial conditions.

More information

you expect to encounter difficulties when trying to solve A x = b? 4. A composite quadrature rule has error associated with it in the following form

you expect to encounter difficulties when trying to solve A x = b? 4. A composite quadrature rule has error associated with it in the following form Qualifying exam for numerical analysis (Spring 2017) Show your work for full credit. If you are unable to solve some part, attempt the subsequent parts. 1. Consider the following finite difference: f (0)

More information

Math/Phys/Engr 428, Math 529/Phys 528 Numerical Methods - Summer Homework 3 Due: Tuesday, July 3, 2018

Math/Phys/Engr 428, Math 529/Phys 528 Numerical Methods - Summer Homework 3 Due: Tuesday, July 3, 2018 Math/Phys/Engr 428, Math 529/Phys 528 Numerical Methods - Summer 28. (Vector and Matrix Norms) Homework 3 Due: Tuesday, July 3, 28 Show that the l vector norm satisfies the three properties (a) x for x

More information

Jim Lambers MAT 610 Summer Session Lecture 2 Notes

Jim Lambers MAT 610 Summer Session Lecture 2 Notes Jim Lambers MAT 610 Summer Session 2009-10 Lecture 2 Notes These notes correspond to Sections 2.2-2.4 in the text. Vector Norms Given vectors x and y of length one, which are simply scalars x and y, the

More information

1 Finding and fixing floating point problems

1 Finding and fixing floating point problems Notes for 2016-09-09 1 Finding and fixing floating point problems Floating point arithmetic is not the same as real arithmetic. Even simple properties like associativity or distributivity of addition and

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra)

AMS526: Numerical Analysis I (Numerical Linear Algebra) AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 09: Accuracy and Stability Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 12 Outline 1 Condition Number of Matrices

More information

Numerical Analysis Preliminary Exam 10 am to 1 pm, August 20, 2018

Numerical Analysis Preliminary Exam 10 am to 1 pm, August 20, 2018 Numerical Analysis Preliminary Exam 1 am to 1 pm, August 2, 218 Instructions. You have three hours to complete this exam. Submit solutions to four (and no more) of the following six problems. Please start

More information

Math Numerical Analysis Mid-Term Test Solutions

Math Numerical Analysis Mid-Term Test Solutions Math 400 - Numerical Analysis Mid-Term Test Solutions. Short Answers (a) A sufficient and necessary condition for the bisection method to find a root of f(x) on the interval [a,b] is f(a)f(b) < 0 or f(a)

More information

Homework and Computer Problems for Math*2130 (W17).

Homework and Computer Problems for Math*2130 (W17). Homework and Computer Problems for Math*2130 (W17). MARCUS R. GARVIE 1 December 21, 2016 1 Department of Mathematics & Statistics, University of Guelph NOTES: These questions are a bare minimum. You should

More information

Outline. Math Numerical Analysis. Errors. Lecture Notes Linear Algebra: Part B. Joseph M. Mahaffy,

Outline. Math Numerical Analysis. Errors. Lecture Notes Linear Algebra: Part B. Joseph M. Mahaffy, Math 54 - Numerical Analysis Lecture Notes Linear Algebra: Part B Outline Joseph M. Mahaffy, jmahaffy@mail.sdsu.edu Department of Mathematics and Statistics Dynamical Systems Group Computational Sciences

More information

Homework 2. Matthew Jin. April 10, 2014

Homework 2. Matthew Jin. April 10, 2014 Homework Matthew Jin April 10, 014 1a) The relative error is given by ŷ y y, where ŷ represents the observed output value, and y represents the theoretical output value. In this case, the observed output

More information

Review for Exam 2 Ben Wang and Mark Styczynski

Review for Exam 2 Ben Wang and Mark Styczynski Review for Exam Ben Wang and Mark Styczynski This is a rough approximation of what we went over in the review session. This is actually more detailed in portions than what we went over. Also, please note

More information

Newton s Method and Linear Approximations

Newton s Method and Linear Approximations Newton s Method and Linear Approximations Curves are tricky. Lines aren t. Newton s Method and Linear Approximations Newton s Method for finding roots Goal: Where is f (x) = 0? f (x) = x 7 + 3x 3 + 7x

More information

Chapter 6. Nonlinear Equations. 6.1 The Problem of Nonlinear Root-finding. 6.2 Rate of Convergence

Chapter 6. Nonlinear Equations. 6.1 The Problem of Nonlinear Root-finding. 6.2 Rate of Convergence Chapter 6 Nonlinear Equations 6. The Problem of Nonlinear Root-finding In this module we consider the problem of using numerical techniques to find the roots of nonlinear equations, f () =. Initially we

More information

9.5. Polynomial and Rational Inequalities. Objectives. Solve quadratic inequalities. Solve polynomial inequalities of degree 3 or greater.

9.5. Polynomial and Rational Inequalities. Objectives. Solve quadratic inequalities. Solve polynomial inequalities of degree 3 or greater. Chapter 9 Section 5 9.5 Polynomial and Rational Inequalities Objectives 1 3 Solve quadratic inequalities. Solve polynomial inequalities of degree 3 or greater. Solve rational inequalities. Objective 1

More information

Math 56 Homework 1 Michael Downs. ne n 10 + ne n (1)

Math 56 Homework 1 Michael Downs. ne n 10 + ne n (1) . Problem (a) Yes. The following equation: ne n + ne n () holds for all n R but, since we re only concerned with the asymptotic behavior as n, let us only consider n >. Dividing both sides by n( + ne n

More information

Iterative solvers for linear equations

Iterative solvers for linear equations Spectral Graph Theory Lecture 23 Iterative solvers for linear equations Daniel A. Spielman November 26, 2018 23.1 Overview In this and the next lecture, I will discuss iterative algorithms for solving

More information

CM2104: Computational Mathematics General Maths: 2. Algebra - Factorisation

CM2104: Computational Mathematics General Maths: 2. Algebra - Factorisation CM204: Computational Mathematics General Maths: 2. Algebra - Factorisation Prof. David Marshall School of Computer Science & Informatics Factorisation Factorisation is a way of simplifying algebraic expressions.

More information

Math 365 Homework 5 Due April 6, 2018 Grady Wright

Math 365 Homework 5 Due April 6, 2018 Grady Wright Math 365 Homework 5 Due April 6, 018 Grady Wright 1. (Polynomial interpolation, 10pts) Consider the following three samples of some function f(x): x f(x) -1/ -1 1 1 1/ (a) Determine the unique second degree

More information

x x2 2 + x3 3 x4 3. Use the divided-difference method to find a polynomial of least degree that fits the values shown: (b)

x x2 2 + x3 3 x4 3. Use the divided-difference method to find a polynomial of least degree that fits the values shown: (b) Numerical Methods - PROBLEMS. The Taylor series, about the origin, for log( + x) is x x2 2 + x3 3 x4 4 + Find an upper bound on the magnitude of the truncation error on the interval x.5 when log( + x)

More information

The numerical stability of barycentric Lagrange interpolation

The numerical stability of barycentric Lagrange interpolation IMA Journal of Numerical Analysis (2004) 24, 547 556 The numerical stability of barycentric Lagrange interpolation NICHOLAS J. HIGHAM Department of Mathematics, University of Manchester, Manchester M13

More information

Iterative solvers for linear equations

Iterative solvers for linear equations Spectral Graph Theory Lecture 17 Iterative solvers for linear equations Daniel A. Spielman October 31, 2012 17.1 About these notes These notes are not necessarily an accurate representation of what happened

More information

Essential Mathematics

Essential Mathematics Appendix B 1211 Appendix B Essential Mathematics Exponential Arithmetic Exponential notation is used to express very large and very small numbers as a product of two numbers. The first number of the product,

More information

Integration of Rational Functions by Partial Fractions

Integration of Rational Functions by Partial Fractions Title Integration of Rational Functions by MATH 1700 MATH 1700 1 / 11 Readings Readings Readings: Section 7.4 MATH 1700 2 / 11 Rational functions A rational function is one of the form where P and Q are

More information

Lecture 8 : Eigenvalues and Eigenvectors

Lecture 8 : Eigenvalues and Eigenvectors CPS290: Algorithmic Foundations of Data Science February 24, 2017 Lecture 8 : Eigenvalues and Eigenvectors Lecturer: Kamesh Munagala Scribe: Kamesh Munagala Hermitian Matrices It is simpler to begin with

More information

Midterm Review. Igor Yanovsky (Math 151A TA)

Midterm Review. Igor Yanovsky (Math 151A TA) Midterm Review Igor Yanovsky (Math 5A TA) Root-Finding Methods Rootfinding methods are designed to find a zero of a function f, that is, to find a value of x such that f(x) =0 Bisection Method To apply

More information

Math 473: Practice Problems for Test 1, Fall 2011, SOLUTIONS

Math 473: Practice Problems for Test 1, Fall 2011, SOLUTIONS Math 473: Practice Problems for Test 1, Fall 011, SOLUTIONS Show your work: 1. (a) Compute the Taylor polynomials P n (x) for f(x) = sin x and x 0 = 0. Solution: Compute f(x) = sin x, f (x) = cos x, f

More information

3.1 Interpolation and the Lagrange Polynomial

3.1 Interpolation and the Lagrange Polynomial MATH 4073 Chapter 3 Interpolation and Polynomial Approximation Fall 2003 1 Consider a sample x x 0 x 1 x n y y 0 y 1 y n. Can we get a function out of discrete data above that gives a reasonable estimate

More information

Newton s Method and Linear Approximations

Newton s Method and Linear Approximations Newton s Method and Linear Approximations Newton s Method for finding roots Goal: Where is f (x) =0? f (x) =x 7 +3x 3 +7x 2 1 2-1 -0.5 0.5-2 Newton s Method for finding roots Goal: Where is f (x) =0? f

More information

Integration of Rational Functions by Partial Fractions

Integration of Rational Functions by Partial Fractions Title Integration of Rational Functions by Partial Fractions MATH 1700 December 6, 2016 MATH 1700 Partial Fractions December 6, 2016 1 / 11 Readings Readings Readings: Section 7.4 MATH 1700 Partial Fractions

More information

1 Backward and Forward Error

1 Backward and Forward Error Math 515 Fall, 2008 Brief Notes on Conditioning, Stability and Finite Precision Arithmetic Most books on numerical analysis, numerical linear algebra, and matrix computations have a lot of material covering

More information

Justify all your answers and write down all important steps. Unsupported answers will be disregarded.

Justify all your answers and write down all important steps. Unsupported answers will be disregarded. Numerical Analysis FMN011 10058 The exam lasts 4 hours and has 13 questions. A minimum of 35 points out of the total 70 are required to get a passing grade. These points will be added to those you obtained

More information

CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 9

CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 9 CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 9 GENE H GOLUB 1 Error Analysis of Gaussian Elimination In this section, we will consider the case of Gaussian elimination and perform a detailed

More information

Scientific Computing

Scientific Computing Scientific Computing Direct solution methods Martin van Gijzen Delft University of Technology October 3, 2018 1 Program October 3 Matrix norms LU decomposition Basic algorithm Cost Stability Pivoting Pivoting

More information

Newton s Method and Linear Approximations 10/19/2011

Newton s Method and Linear Approximations 10/19/2011 Newton s Method and Linear Approximations 10/19/2011 Curves are tricky. Lines aren t. Newton s Method and Linear Approximations 10/19/2011 Newton s Method Goal: Where is f (x) =0? f (x) =x 7 +3x 3 +7x

More information

Mathematics for Engineers. Numerical mathematics

Mathematics for Engineers. Numerical mathematics Mathematics for Engineers Numerical mathematics Integers Determine the largest representable integer with the intmax command. intmax ans = int32 2147483647 2147483647+1 ans = 2.1475e+09 Remark The set

More information

8.7 Taylor s Inequality Math 2300 Section 005 Calculus II. f(x) = ln(1 + x) f(0) = 0

8.7 Taylor s Inequality Math 2300 Section 005 Calculus II. f(x) = ln(1 + x) f(0) = 0 8.7 Taylor s Inequality Math 00 Section 005 Calculus II Name: ANSWER KEY Taylor s Inequality: If f (n+) is continuous and f (n+) < M between the center a and some point x, then f(x) T n (x) M x a n+ (n

More information

Applied Mathematics 205. Unit 0: Overview of Scientific Computing. Lecturer: Dr. David Knezevic

Applied Mathematics 205. Unit 0: Overview of Scientific Computing. Lecturer: Dr. David Knezevic Applied Mathematics 205 Unit 0: Overview of Scientific Computing Lecturer: Dr. David Knezevic Scientific Computing Computation is now recognized as the third pillar of science (along with theory and experiment)

More information

Equations in Quadratic Form

Equations in Quadratic Form Equations in Quadratic Form MATH 101 College Algebra J. Robert Buchanan Department of Mathematics Summer 2012 Objectives In this lesson we will learn to: make substitutions that allow equations to be written

More information

Roots and Coefficients Polynomials Preliminary Maths Extension 1

Roots and Coefficients Polynomials Preliminary Maths Extension 1 Preliminary Maths Extension Question If, and are the roots of x 5x x 0, find the following. (d) (e) Question If p, q and r are the roots of x x x 4 0, evaluate the following. pq r pq qr rp p q q r r p

More information

CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 5. Ax = b.

CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 5. Ax = b. CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 5 GENE H GOLUB Suppose we want to solve We actually have an approximation ξ such that 1 Perturbation Theory Ax = b x = ξ + e The question is, how

More information

Non-polynomial Least-squares fitting

Non-polynomial Least-squares fitting Applied Math 205 Last time: piecewise polynomial interpolation, least-squares fitting Today: underdetermined least squares, nonlinear least squares Homework 1 (and subsequent homeworks) have several parts

More information

Computational Methods. Least Squares Approximation/Optimization

Computational Methods. Least Squares Approximation/Optimization Computational Methods Least Squares Approximation/Optimization Manfred Huber 2011 1 Least Squares Least squares methods are aimed at finding approximate solutions when no precise solution exists Find the

More information

Review I: Interpolation

Review I: Interpolation Review I: Interpolation Varun Shankar January, 206 Introduction In this document, we review interpolation by polynomials. Unlike many reviews, we will not stop there: we will discuss how to differentiate

More information

Binary floating point

Binary floating point Binary floating point Notes for 2017-02-03 Why do we study conditioning of problems? One reason is that we may have input data contaminated by noise, resulting in a bad solution even if the intermediate

More information

Math 502 Fall 2005 Solutions to Homework 3

Math 502 Fall 2005 Solutions to Homework 3 Math 502 Fall 2005 Solutions to Homework 3 (1) As shown in class, the relative distance between adjacent binary floating points numbers is 2 1 t, where t is the number of digits in the mantissa. Since

More information

Numerical Methods. Elena loli Piccolomini. Civil Engeneering. piccolom. Metodi Numerici M p. 1/??

Numerical Methods. Elena loli Piccolomini. Civil Engeneering.  piccolom. Metodi Numerici M p. 1/?? Metodi Numerici M p. 1/?? Numerical Methods Elena loli Piccolomini Civil Engeneering http://www.dm.unibo.it/ piccolom elena.loli@unibo.it Metodi Numerici M p. 2/?? Least Squares Data Fitting Measurement

More information

CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 6

CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 6 CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 6 GENE H GOLUB Issues with Floating-point Arithmetic We conclude our discussion of floating-point arithmetic by highlighting two issues that frequently

More information

Department of Applied Mathematics and Theoretical Physics. AMA 204 Numerical analysis. Exam Winter 2004

Department of Applied Mathematics and Theoretical Physics. AMA 204 Numerical analysis. Exam Winter 2004 Department of Applied Mathematics and Theoretical Physics AMA 204 Numerical analysis Exam Winter 2004 The best six answers will be credited All questions carry equal marks Answer all parts of each question

More information

Math 411 Preliminaries

Math 411 Preliminaries Math 411 Preliminaries Provide a list of preliminary vocabulary and concepts Preliminary Basic Netwon s method, Taylor series expansion (for single and multiple variables), Eigenvalue, Eigenvector, Vector

More information

Notes on floating point number, numerical computations and pitfalls

Notes on floating point number, numerical computations and pitfalls Notes on floating point number, numerical computations and pitfalls November 6, 212 1 Floating point numbers An n-digit floating point number in base β has the form x = ±(.d 1 d 2 d n ) β β e where.d 1

More information

Lecture 4: Numerical solution of ordinary differential equations

Lecture 4: Numerical solution of ordinary differential equations Lecture 4: Numerical solution of ordinary differential equations Department of Mathematics, ETH Zürich General explicit one-step method: Consistency; Stability; Convergence. High-order methods: Taylor

More information

Computer Arithmetic. MATH 375 Numerical Analysis. J. Robert Buchanan. Fall Department of Mathematics. J. Robert Buchanan Computer Arithmetic

Computer Arithmetic. MATH 375 Numerical Analysis. J. Robert Buchanan. Fall Department of Mathematics. J. Robert Buchanan Computer Arithmetic Computer Arithmetic MATH 375 Numerical Analysis J. Robert Buchanan Department of Mathematics Fall 2013 Machine Numbers When performing arithmetic on a computer (laptop, desktop, mainframe, cell phone,

More information

Queens College, CUNY, Department of Computer Science Numerical Methods CSCI 361 / 761 Spring 2018 Instructor: Dr. Sateesh Mane.

Queens College, CUNY, Department of Computer Science Numerical Methods CSCI 361 / 761 Spring 2018 Instructor: Dr. Sateesh Mane. Queens College, CUNY, Department of Computer Science Numerical Methods CSCI 361 / 761 Spring 2018 Instructor: Dr. Sateesh Mane c Sateesh R. Mane 2018 3 Lecture 3 3.1 General remarks March 4, 2018 This

More information

Linear Algebra Review

Linear Algebra Review Chapter 1 Linear Algebra Review It is assumed that you have had a course in linear algebra, and are familiar with matrix multiplication, eigenvectors, etc. I will review some of these terms here, but quite

More information

(0, 0), (1, ), (2, ), (3, ), (4, ), (5, ), (6, ).

(0, 0), (1, ), (2, ), (3, ), (4, ), (5, ), (6, ). 1 Interpolation: The method of constructing new data points within the range of a finite set of known data points That is if (x i, y i ), i = 1, N are known, with y i the dependent variable and x i [x

More information

Elements of Floating-point Arithmetic

Elements of Floating-point Arithmetic Elements of Floating-point Arithmetic Sanzheng Qiao Department of Computing and Software McMaster University July, 2012 Outline 1 Floating-point Numbers Representations IEEE Floating-point Standards Underflow

More information

LECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel

LECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel LECTURE NOTES on ELEMENTARY NUMERICAL METHODS Eusebius Doedel TABLE OF CONTENTS Vector and Matrix Norms 1 Banach Lemma 20 The Numerical Solution of Linear Systems 25 Gauss Elimination 25 Operation Count

More information

COMP 558 lecture 18 Nov. 15, 2010

COMP 558 lecture 18 Nov. 15, 2010 Least squares We have seen several least squares problems thus far, and we will see more in the upcoming lectures. For this reason it is good to have a more general picture of these problems and how to

More information

Here is an example of a block diagonal matrix with Jordan Blocks on the diagonal: J

Here is an example of a block diagonal matrix with Jordan Blocks on the diagonal: J Class Notes 4: THE SPECTRAL RADIUS, NORM CONVERGENCE AND SOR. Math 639d Due Date: Feb. 7 (updated: February 5, 2018) In the first part of this week s reading, we will prove Theorem 2 of the previous class.

More information

Lecture2 The implicit function theorem. Bifurcations.

Lecture2 The implicit function theorem. Bifurcations. Lecture2 The implicit function theorem. Bifurcations. 1 Review:Newton s method. The existence theorem - the assumptions. Basins of attraction. 2 The implicit function theorem. 3 Bifurcations of iterations.

More information

Centerville High School Curriculum Mapping Algebra II 1 st Nine Weeks

Centerville High School Curriculum Mapping Algebra II 1 st Nine Weeks Centerville High School Curriculum Mapping Algebra II 1 st Nine Weeks Chapter/ Lesson Common Core Standard(s) 1-1 SMP1 1. How do you use a number line to graph and order real numbers? 2. How do you identify

More information

MATH10001 Mathematical Workshop Graph Fitting Project part 2

MATH10001 Mathematical Workshop Graph Fitting Project part 2 MATH10001 Mathematical Workshop Graph Fitting Project part 2 Polynomial models Modelling a set of data with a polynomial curve can be convenient because polynomial functions are particularly easy to differentiate

More information

Lecture Note 3: Interpolation and Polynomial Approximation. Xiaoqun Zhang Shanghai Jiao Tong University

Lecture Note 3: Interpolation and Polynomial Approximation. Xiaoqun Zhang Shanghai Jiao Tong University Lecture Note 3: Interpolation and Polynomial Approximation Xiaoqun Zhang Shanghai Jiao Tong University Last updated: October 10, 2015 2 Contents 1.1 Introduction................................ 3 1.1.1

More information

Mathematical preliminaries and error analysis

Mathematical preliminaries and error analysis Mathematical preliminaries and error analysis Tsung-Ming Huang Department of Mathematics National Taiwan Normal University, Taiwan September 12, 2015 Outline 1 Round-off errors and computer arithmetic

More information

One Dimensional Dynamical Systems

One Dimensional Dynamical Systems 16 CHAPTER 2 One Dimensional Dynamical Systems We begin by analyzing some dynamical systems with one-dimensional phase spaces, and in particular their bifurcations. All equations in this Chapter are scalar

More information

1. The positive zero of y = x 2 + 2x 3/5 is, to the nearest tenth, equal to

1. The positive zero of y = x 2 + 2x 3/5 is, to the nearest tenth, equal to SAT II - Math Level Test #0 Solution SAT II - Math Level Test No. 1. The positive zero of y = x + x 3/5 is, to the nearest tenth, equal to (A) 0.8 (B) 0.7 + 1.1i (C) 0.7 (D) 0.3 (E). 3 b b 4ac Using Quadratic

More information

The final is cumulative, but with more emphasis on chapters 3 and 4. There will be two parts.

The final is cumulative, but with more emphasis on chapters 3 and 4. There will be two parts. Math 141 Review for Final The final is cumulative, but with more emphasis on chapters 3 and 4. There will be two parts. Part 1 (no calculator) graphing (polynomial, rational, linear, exponential, and logarithmic

More information

One-Sided Difference Formula for the First Derivative

One-Sided Difference Formula for the First Derivative POLYTECHNIC UNIVERSITY Department of Computer and Information Science One-Sided Difference Formula for the First Derivative K. Ming Leung Abstract: Derive a one-sided formula for the first derive of a

More information

Numerical Methods for Differential Equations Mathematical and Computational Tools

Numerical Methods for Differential Equations Mathematical and Computational Tools Numerical Methods for Differential Equations Mathematical and Computational Tools Gustaf Söderlind Numerical Analysis, Lund University Contents V4.16 Part 1. Vector norms, matrix norms and logarithmic

More information

Math Solutions to homework 5

Math Solutions to homework 5 Math 75 - Solutions to homework 5 Cédric De Groote November 9, 207 Problem (7. in the book): Let {e n } be a complete orthonormal sequence in a Hilbert space H and let λ n C for n N. Show that there is

More information

Solutions to Dynamical Systems 2010 exam. Each question is worth 25 marks.

Solutions to Dynamical Systems 2010 exam. Each question is worth 25 marks. Solutions to Dynamical Systems exam Each question is worth marks [Unseen] Consider the following st order differential equation: dy dt Xy yy 4 a Find and classify all the fixed points of Hence draw the

More information

Homework 1 Solutions Math 587

Homework 1 Solutions Math 587 Homework 1 Solutions Math 587 1) Find positive functions f(x),g(x), continuous on (0, ) so that f = O(g(x)) as x 0, but lim x 0 f(x)/g(x) does not exist. One can take f(x) = sin(1/x) and g(x) = 1, for

More information

Introduction. Chapter One

Introduction. Chapter One Chapter One Introduction The aim of this book is to describe and explain the beautiful mathematical relationships between matrices, moments, orthogonal polynomials, quadrature rules and the Lanczos and

More information

ERROR IN LINEAR INTERPOLATION

ERROR IN LINEAR INTERPOLATION ERROR IN LINEAR INTERPOLATION Let P 1 (x) be the linear polynomial interpolating f (x) at x 0 and x 1. Assume f (x) is twice continuously differentiable on an interval [a, b] which contains the points

More information

2-7 Solving Quadratic Inequalities. ax 2 + bx + c > 0 (a 0)

2-7 Solving Quadratic Inequalities. ax 2 + bx + c > 0 (a 0) Quadratic Inequalities In One Variable LOOKS LIKE a quadratic equation but Doesn t have an equal sign (=) Has an inequality sign (>,

More information

PARTIAL DIFFERENTIAL EQUATIONS

PARTIAL DIFFERENTIAL EQUATIONS MATHEMATICAL METHODS PARTIAL DIFFERENTIAL EQUATIONS I YEAR B.Tech By Mr. Y. Prabhaker Reddy Asst. Professor of Mathematics Guru Nanak Engineering College Ibrahimpatnam, Hyderabad. SYLLABUS OF MATHEMATICAL

More information

CS100: DISCRETE STRUCTURES. Lecture 3 Matrices Ch 3 Pages:

CS100: DISCRETE STRUCTURES. Lecture 3 Matrices Ch 3 Pages: CS100: DISCRETE STRUCTURES Lecture 3 Matrices Ch 3 Pages: 246-262 Matrices 2 Introduction DEFINITION 1: A matrix is a rectangular array of numbers. A matrix with m rows and n columns is called an m x n

More information

Preliminary Examination in Numerical Analysis

Preliminary Examination in Numerical Analysis Department of Applied Mathematics Preliminary Examination in Numerical Analysis August 7, 06, 0 am pm. Submit solutions to four (and no more) of the following six problems. Show all your work, and justify

More information

PowerPoints organized by Dr. Michael R. Gustafson II, Duke University

PowerPoints organized by Dr. Michael R. Gustafson II, Duke University Part 1 Chapter 4 Roundoff and Truncation Errors PowerPoints organized by Dr. Michael R. Gustafson II, Duke University All images copyright The McGraw-Hill Companies, Inc. Permission required for reproduction

More information

ERROR AND SENSITIVTY ANALYSIS FOR SYSTEMS OF LINEAR EQUATIONS. Perturbation analysis for linear systems (Ax = b)

ERROR AND SENSITIVTY ANALYSIS FOR SYSTEMS OF LINEAR EQUATIONS. Perturbation analysis for linear systems (Ax = b) ERROR AND SENSITIVTY ANALYSIS FOR SYSTEMS OF LINEAR EQUATIONS Conditioning of linear systems. Estimating errors for solutions of linear systems Backward error analysis Perturbation analysis for linear

More information

MA3457/CS4033: Numerical Methods for Calculus and Differential Equations

MA3457/CS4033: Numerical Methods for Calculus and Differential Equations MA3457/CS4033: Numerical Methods for Calculus and Differential Equations Course Materials P A R T II B 14 2014-2015 1 2. APPROXIMATION CLASS 9 Approximation Key Idea Function approximation is closely related

More information

the method of steepest descent

the method of steepest descent MATH 3511 Spring 2018 the method of steepest descent http://www.phys.uconn.edu/ rozman/courses/m3511_18s/ Last modified: February 6, 2018 Abstract The Steepest Descent is an iterative method for solving

More information

Learning Module 1 - Basic Algebra Review (Appendix A)

Learning Module 1 - Basic Algebra Review (Appendix A) Learning Module 1 - Basic Algebra Review (Appendix A) Element 1 Real Numbers and Operations on Polynomials (A.1, A.2) Use the properties of real numbers and work with subsets of the real numbers Determine

More information

Outline. Math Numerical Analysis. Intermediate Value Theorem. Lecture Notes Zeros and Roots. Joseph M. Mahaffy,

Outline. Math Numerical Analysis. Intermediate Value Theorem. Lecture Notes Zeros and Roots. Joseph M. Mahaffy, Outline Math 541 - Numerical Analysis Lecture Notes Zeros and Roots Joseph M. Mahaffy, jmahaffy@mail.sdsu.edu Department of Mathematics and Statistics Dynamical Systems Group Computational Sciences Research

More information

Taylor Series and Numerical Approximations

Taylor Series and Numerical Approximations Taylor Series and Numerical Approximations Hilary Weller h.weller@reading.ac.uk August 7, 05 An introduction to the concept of a Taylor series and how these are used in numerical analysis to find numerical

More information

b = 2, c = 3, we get x = 0.3 for the positive root. Ans. (D) x 2-2x - 8 < 0, or (x - 4)(x + 2) < 0, Therefore -2 < x < 4 Ans. (C)

b = 2, c = 3, we get x = 0.3 for the positive root. Ans. (D) x 2-2x - 8 < 0, or (x - 4)(x + 2) < 0, Therefore -2 < x < 4 Ans. (C) SAT II - Math Level 2 Test #02 Solution 1. The positive zero of y = x 2 + 2x is, to the nearest tenth, equal to (A) 0.8 (B) 0.7 + 1.1i (C) 0.7 (D) 0.3 (E) 2.2 ± Using Quadratic formula, x =, with a = 1,

More information

Math 221 Notes on Rolle s Theorem, The Mean Value Theorem, l Hôpital s rule, and the Taylor-Maclaurin formula. 1. Two theorems

Math 221 Notes on Rolle s Theorem, The Mean Value Theorem, l Hôpital s rule, and the Taylor-Maclaurin formula. 1. Two theorems Math 221 Notes on Rolle s Theorem, The Mean Value Theorem, l Hôpital s rule, and the Taylor-Maclaurin formula 1. Two theorems Rolle s Theorem. If a function y = f(x) is differentiable for a x b and if

More information

Math Numerical Analysis

Math Numerical Analysis Math 541 - Numerical Analysis Lecture Notes Zeros and Roots Joseph M. Mahaffy, jmahaffy@mail.sdsu.edu Department of Mathematics and Statistics Dynamical Systems Group Computational Sciences Research Center

More information

MATH ASSIGNMENT 07 SOLUTIONS. 8.1 Following is census data showing the population of the US between 1900 and 2000:

MATH ASSIGNMENT 07 SOLUTIONS. 8.1 Following is census data showing the population of the US between 1900 and 2000: MATH4414.01 ASSIGNMENT 07 SOLUTIONS 8.1 Following is census data showing the population of the US between 1900 and 2000: Years after 1900 Population in millions 0 76.0 20 105.7 40 131.7 60 179.3 80 226.5

More information

Parallel programming practices for the solution of Sparse Linear Systems (motivated by computational physics and graphics)

Parallel programming practices for the solution of Sparse Linear Systems (motivated by computational physics and graphics) Parallel programming practices for the solution of Sparse Linear Systems (motivated by computational physics and graphics) Eftychios Sifakis CS758 Guest Lecture - 19 Sept 2012 Introduction Linear systems

More information

The Behavior of Algorithms in Practice 2/21/2002. Lecture 4. ɛ 1 x 1 y ɛ 1 x 1 1 = x y 1 1 = y 1 = 1 y 2 = 1 1 = 0 1 1

The Behavior of Algorithms in Practice 2/21/2002. Lecture 4. ɛ 1 x 1 y ɛ 1 x 1 1 = x y 1 1 = y 1 = 1 y 2 = 1 1 = 0 1 1 8.409 The Behavior of Algorithms in Practice //00 Lecture 4 Lecturer: Dan Spielman Scribe: Matthew Lepinski A Gaussian Elimination Example To solve: [ ] [ ] [ ] x x First factor the matrix to get: [ ]

More information

INTERPOLATION Background Polynomial Approximation Problem:

INTERPOLATION Background Polynomial Approximation Problem: INTERPOLATION Background Polynomial Approximation Problem: given f(x) C[a, b], find P n (x) = a 0 + a 1 x + a 2 x 2 + + a n x n with P n (x) close to f(x) for x [a, b]. Motivations: f(x) might be difficult

More information

Chapter 1: Precalculus Review

Chapter 1: Precalculus Review : Precalculus Review Math 115 17 January 2018 Overview 1 Important Notation 2 Exponents 3 Polynomials 4 Rational Functions 5 Cartesian Coordinates 6 Lines Notation Intervals: Interval Notation (a, b) (a,

More information

Chapter III. Stability of Linear Systems

Chapter III. Stability of Linear Systems 1 Chapter III Stability of Linear Systems 1. Stability and state transition matrix 2. Time-varying (non-autonomous) systems 3. Time-invariant systems 1 STABILITY AND STATE TRANSITION MATRIX 2 In this chapter,

More information