The answer in each case is the error in evaluating the taylor series for ln(1 x) for x = which is 6.9.
|
|
- Myles Barnett
- 5 years ago
- Views:
Transcription
1 Brad Nelson Math 26 Homework #2 /23/2. a MATLAB outputs: >> a=(+3.4e-6)-.e-6;a- ans = 4.449e-6 >> a=+(3.4e-6-.e-6);a- ans = 2.224e-6 And the exact answer for both operations is 2.3e-6. The reason why way I is so inaccurate is that the number +3.4e-6 is stored on the computer as ( )( + ɛ mach ) which corresponds to an error of magnitude 6 in double precision. Then the floating point operation ( ). 6 is accurate to (( ). 6 )( + ɛ mach ), which also introduces an error of magnitude 6. The final step, a, doesn t introduce significant error into this calculation, since the resulting number is on magnitude of 6 so the error is of the magnitude 32. The two operations which introduce error of magnitude 6 result in an inaccurate answer since the real answer is of the same magnitude. Way II is more accurate since ( = ( + ɛ mach ) only has an error 32, which is not significant for this calculation. Then = ( )( + ɛ mach ) introduces an error of magnitude 6 which is significant since the final answer is of this magnitude. Once again, the final operation, a does not introduce significant error. While the returned answer is inaccurate because the introduced error was of the same magnitude as the answer, there was only one such step that introduced such a large error, which compared to two such steps in way I, is more accurate. b MATLAB outputs: >> x =.999; a = ; for j=:6, a = a-x^j/j; end; a-log(-x) ans = 3.553e-3 >> x =.999; a = ; for j=6:-:, a = a-x^j/j; end; a-log(-x) ans = e-6 The answer in each case is the error in evaluating the taylor series for ln( x) for x =.999 which is For way I, the series is evaluated as a = j For way II, a = j 6 j. 6 j j= j= Looking at way I, the first term evaluated is.999 and the second term is.5. Adding these two terms together introduces an error of.5ɛ mach which is of the magnitude 6. Since the partial sum monotonically approaches , it is always of magnitude. This means that adding each term to the series introduces an error of 6, which compounded over 6 terms turns into a total error of 2. The actual observed error
2 is close to this, at 3, and may be smaller than the prediction due to errors in opposite directions from the true answer canceling each other out. As for way II, the first term evaluated is of order 3. Performing operations on numbers of this magnitude will introduce a relative error of 6, which is an absolute error of 47. Since these errors are so small, the introduced error doesn t reach a magnitude of 6 until towards the end of the calculation, so the compounded error is still 6. From these observations, we can conclude that when summing together a list of numbers using floating-point arithmetic, it is best to start with the smallest numbers first. This way the total error of the summation stays small relative to the size of the answer until the last terms are added, which prevents a large relative error from compounding over many iterations. 2. a f (x) = 2x computed via x x. On a machine, this algorithm is f (x) = f l(x) f l(x). To prove backwards stability, we want to show that x X, f x x (x) = f ( x) for some x : = O(ɛ x mach ). To see if this is possible, we set f (x) = (x( + ɛ ) + x( + ɛ ))( + ɛ 2 ) = x( + ɛ 3 ) + x( + ɛ 3 ) where ɛ,2 ɛ mach. It follows that ɛ 3 = ɛ + ɛ 2 + ɛ ɛ 2. Thus, ɛ 3 2ɛ mach + ɛ 2 mach, and ɛ 2 = O(ɛ mach ). Thus, the algorithm is backwards stable. b f (x) = + x computed via x. Set: f (x) = ( + x( + ɛ ))( + ɛ 2 ) = + x( + ɛ 3 ) = f ( x) where ɛ,2 ɛ mach. Ignoring high-order terms (because of their small size), we have ɛ 3 = ɛ + ɛ 2 + ɛ 2 x Since for choices of x, ɛ 3 will blow up, the algorithm is not backwards stable. However to test stability, consider f (x) and f ( x). If we force ɛ 3 < ɛ mach, then f (x) f ( x) f (x) = ɛ 2 + x(ɛ + ɛ 2 ɛ 3 ɛ ɛ 2 ) + x( + ɛ 3 ) We see that for x, the algorithm is backwards stable, so focusing on x <, if x is still close to, the numerator is O(ɛ mach ), while the denominator 2, thus f (x) f ( x) f = (x) O(ɛ mach ). For x, the numerator is ɛ 2 and the denominator is, so f (x) f ( x) f = (x) O(ɛ mach ). Thus, while the algorithm is not backwards stable, it is stable. 3. Computing the eigenvalues of a symmetric matrix via the characteristic polynomial: i Restricting to 2x2 diagonal matrices with diagonal entries a and b, the characteristic polynomial is λ 2 (a + b)λ + ab. If we choose A = I, then we have λ 2 2λ +. Since the coefficients of this polynomial are stored with error ɛ mach, the polynomial can be written ( + ɛ a )λ 2 2( + ɛ b )λ + ( + ɛ c ) where ɛ a,b,c ɛ mach. To find the roots of this polynomial using the quadratic formula, we have, λ = 2( + ɛ b) ± 4( + ɛ b ) 2 4( + ɛ a )( + ɛ c ) 2( + ɛ a ) By expanding the inside of the square root, we see that the largest order error terms will be O( ɛ mach ). 2
3 ii While the above was done for A = I, it is easy to see that for any quadratic polynomial ax 2 + bx + c, perturbation of O(ɛ mach ) in the coefficients will lead to a perturbation of O( ɛ mach ) in the roots of a quadratic polynomial by placing the coefficients in the quadratic formula. Thus if ɛ mach = 6, then there will be an error of O( 8 ) in the roots of the polynomial. In general, for a degree n polynomial, we would expect to see O( n ɛ mach ) if a similar method is used. Since n ɛ goes to slower than ɛ, f (x) f ( x) O(ɛ mach ) for any x x f (x) x : = O(ɛ x mach ), and the algorithm is not stable. Since stability is a weaker condition than backwards stable, it is also not backwards stable. Thus, the algorithm is unstable. 4. Code: lagrange.m, lagrange eq.m. The following plot was produced: 5 lagrange basis functions for 2 equally spaced points on [,] Note how for each node, all polynomials except for one go to, and the one that does not goes to. 5. Code: interp error.m, hw2fns.m The two functions in the following questions look like: 3
4 f (x) = sin(2e x ) f (x) = ( + 25x 2 ) The following two plots used 25 equally spaced points on [, ] to create an interpolated polynomial for each function. The interpolation error E n (x) := f (x) (L n f )(x) is plotted. 4
5 .5 x 9 interpolation error for sin(2e x ) (a) Note that the error is at most 9. f (x) = sin(2e x ) 3 interpolation error for (+25*x 2 ) (b) f (x) = ( + 25x 2 ) Note that the error blows up to 25 at edges. 5
6 interpolation error as function of n points Sup norm of the interpolation error for f (x) = sin(2e x ) as a function of n equally spaced points on [, ]. Note how line appears to be roughly linear between n= and 27. Since this is on a semilog plot, this indicates exponential convergence in this region. To explain the reverse trend when n > 27, we note that sup x [,],k {...n} l k (x) grows exponentially: 9 sup of n lagrange basis functions on [,] Sup norm over all n Lagrange basis functions on [, ] as a function of n. Since the supremum is growing exponentially, the supremum in the error of the sum over the 6
7 these functions is going to grow exponentially as ɛ mach sup x [,],k {...n} l k (x), following the rules of floating point arithmetic. Note that at n 27, the supremum is 6, so the supremum of the error will be if ɛ mach = 6. This is where the graph of the sup norm of the interpolation error above suddenly reverses course, and begins to be less accurate, as the increase in error due to the sup norm of the basis functions overtakes the decrease in error due to adding more terms to the interpolant. Code: lbasis sup.m 6. Code: interp error cheby.m, lagrange cheby.m After updating the codes to use Chebychev nodes instead of equally spaced nodes, the following plots were produced:.2 lagrange basis functions for 2 Chebychev points on [,] Lagrange basis functions for 2 Chebychev nodes on [, ]. By inspecting the graph, it is clear that for each l k (x), sup x [,] l k (x). 7
8 5 x 2 interpolation error for sin(2e x ) interpolation error for f (x) = sin(2e x ) plotted for 25 Chebychev points on [, ]. Maximum error is 2, which is better than for equally-spaced points. x 3 interpolation error for (+25*x 2 ) interpolation error for f (x) = ( + 25x 2 ) plotted for 25 Chebychev points on [, ]. The maximum error for this function is now 9 3, which is a drastic improvement over the scenario with equally-spaced nodes. 8
9 interpolation error as function of n points Sup norm of the interpolation error for f (x) = sin(2e x ) as a function of n Chebychev points on [, ]. interpolation error as function of n points Sup norm of the interpolation error for f (x) = ( + 25x 2 ) as a function of n Chebychev points on [, ]. Note that the interpolation error is converging exponentially to. Thus, the Runge phenomenon does not persist. Comparing the graphs for the sup norm of the interpolation error of f (x) = sin(2e x ) for Chebychev points and equally spaced points, it appears that the Chebychev points allow a better achievable interpolation error. For equally spaced points, the best possible error is 9
10 about while Chebychev points have a best possible error of 5. We can t reasonably expect the error to get much better, as this is close to ɛ mach, since the maximum for this function is.
Numerical Algorithms. IE 496 Lecture 20
Numerical Algorithms IE 496 Lecture 20 Reading for This Lecture Primary Miller and Boxer, Pages 124-128 Forsythe and Mohler, Sections 1 and 2 Numerical Algorithms Numerical Analysis So far, we have looked
More informationLecture Notes 7, Math/Comp 128, Math 250
Lecture Notes 7, Math/Comp 128, Math 250 Misha Kilmer Tufts University October 23, 2005 Floating Point Arithmetic We talked last time about how the computer represents floating point numbers. In a floating
More informationTwo hours. To be provided by Examinations Office: Mathematical Formula Tables. THE UNIVERSITY OF MANCHESTER. 29 May :45 11:45
Two hours MATH20602 To be provided by Examinations Office: Mathematical Formula Tables. THE UNIVERSITY OF MANCHESTER NUMERICAL ANALYSIS 1 29 May 2015 9:45 11:45 Answer THREE of the FOUR questions. If more
More informationInterpolation and the Lagrange Polynomial
Interpolation and the Lagrange Polynomial MATH 375 J. Robert Buchanan Department of Mathematics Fall 2013 Introduction We often choose polynomials to approximate other classes of functions. Theorem (Weierstrass
More informationLecture 7. Floating point arithmetic and stability
Lecture 7 Floating point arithmetic and stability 2.5 Machine representation of numbers Scientific notation: 23 }{{} }{{} } 3.14159265 {{} }{{} 10 sign mantissa base exponent (significand) s m β e A floating
More informationMath 2142 Homework 5 Part 1 Solutions
Math 2142 Homework 5 Part 1 Solutions Problem 1. For the following homogeneous second order differential equations, give the general solution and the particular solution satisfying the given initial conditions.
More informationyou expect to encounter difficulties when trying to solve A x = b? 4. A composite quadrature rule has error associated with it in the following form
Qualifying exam for numerical analysis (Spring 2017) Show your work for full credit. If you are unable to solve some part, attempt the subsequent parts. 1. Consider the following finite difference: f (0)
More informationMath/Phys/Engr 428, Math 529/Phys 528 Numerical Methods - Summer Homework 3 Due: Tuesday, July 3, 2018
Math/Phys/Engr 428, Math 529/Phys 528 Numerical Methods - Summer 28. (Vector and Matrix Norms) Homework 3 Due: Tuesday, July 3, 28 Show that the l vector norm satisfies the three properties (a) x for x
More informationJim Lambers MAT 610 Summer Session Lecture 2 Notes
Jim Lambers MAT 610 Summer Session 2009-10 Lecture 2 Notes These notes correspond to Sections 2.2-2.4 in the text. Vector Norms Given vectors x and y of length one, which are simply scalars x and y, the
More information1 Finding and fixing floating point problems
Notes for 2016-09-09 1 Finding and fixing floating point problems Floating point arithmetic is not the same as real arithmetic. Even simple properties like associativity or distributivity of addition and
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra)
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 09: Accuracy and Stability Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 12 Outline 1 Condition Number of Matrices
More informationNumerical Analysis Preliminary Exam 10 am to 1 pm, August 20, 2018
Numerical Analysis Preliminary Exam 1 am to 1 pm, August 2, 218 Instructions. You have three hours to complete this exam. Submit solutions to four (and no more) of the following six problems. Please start
More informationMath Numerical Analysis Mid-Term Test Solutions
Math 400 - Numerical Analysis Mid-Term Test Solutions. Short Answers (a) A sufficient and necessary condition for the bisection method to find a root of f(x) on the interval [a,b] is f(a)f(b) < 0 or f(a)
More informationHomework and Computer Problems for Math*2130 (W17).
Homework and Computer Problems for Math*2130 (W17). MARCUS R. GARVIE 1 December 21, 2016 1 Department of Mathematics & Statistics, University of Guelph NOTES: These questions are a bare minimum. You should
More informationOutline. Math Numerical Analysis. Errors. Lecture Notes Linear Algebra: Part B. Joseph M. Mahaffy,
Math 54 - Numerical Analysis Lecture Notes Linear Algebra: Part B Outline Joseph M. Mahaffy, jmahaffy@mail.sdsu.edu Department of Mathematics and Statistics Dynamical Systems Group Computational Sciences
More informationHomework 2. Matthew Jin. April 10, 2014
Homework Matthew Jin April 10, 014 1a) The relative error is given by ŷ y y, where ŷ represents the observed output value, and y represents the theoretical output value. In this case, the observed output
More informationReview for Exam 2 Ben Wang and Mark Styczynski
Review for Exam Ben Wang and Mark Styczynski This is a rough approximation of what we went over in the review session. This is actually more detailed in portions than what we went over. Also, please note
More informationNewton s Method and Linear Approximations
Newton s Method and Linear Approximations Curves are tricky. Lines aren t. Newton s Method and Linear Approximations Newton s Method for finding roots Goal: Where is f (x) = 0? f (x) = x 7 + 3x 3 + 7x
More informationChapter 6. Nonlinear Equations. 6.1 The Problem of Nonlinear Root-finding. 6.2 Rate of Convergence
Chapter 6 Nonlinear Equations 6. The Problem of Nonlinear Root-finding In this module we consider the problem of using numerical techniques to find the roots of nonlinear equations, f () =. Initially we
More information9.5. Polynomial and Rational Inequalities. Objectives. Solve quadratic inequalities. Solve polynomial inequalities of degree 3 or greater.
Chapter 9 Section 5 9.5 Polynomial and Rational Inequalities Objectives 1 3 Solve quadratic inequalities. Solve polynomial inequalities of degree 3 or greater. Solve rational inequalities. Objective 1
More informationMath 56 Homework 1 Michael Downs. ne n 10 + ne n (1)
. Problem (a) Yes. The following equation: ne n + ne n () holds for all n R but, since we re only concerned with the asymptotic behavior as n, let us only consider n >. Dividing both sides by n( + ne n
More informationIterative solvers for linear equations
Spectral Graph Theory Lecture 23 Iterative solvers for linear equations Daniel A. Spielman November 26, 2018 23.1 Overview In this and the next lecture, I will discuss iterative algorithms for solving
More informationCM2104: Computational Mathematics General Maths: 2. Algebra - Factorisation
CM204: Computational Mathematics General Maths: 2. Algebra - Factorisation Prof. David Marshall School of Computer Science & Informatics Factorisation Factorisation is a way of simplifying algebraic expressions.
More informationMath 365 Homework 5 Due April 6, 2018 Grady Wright
Math 365 Homework 5 Due April 6, 018 Grady Wright 1. (Polynomial interpolation, 10pts) Consider the following three samples of some function f(x): x f(x) -1/ -1 1 1 1/ (a) Determine the unique second degree
More informationx x2 2 + x3 3 x4 3. Use the divided-difference method to find a polynomial of least degree that fits the values shown: (b)
Numerical Methods - PROBLEMS. The Taylor series, about the origin, for log( + x) is x x2 2 + x3 3 x4 4 + Find an upper bound on the magnitude of the truncation error on the interval x.5 when log( + x)
More informationThe numerical stability of barycentric Lagrange interpolation
IMA Journal of Numerical Analysis (2004) 24, 547 556 The numerical stability of barycentric Lagrange interpolation NICHOLAS J. HIGHAM Department of Mathematics, University of Manchester, Manchester M13
More informationIterative solvers for linear equations
Spectral Graph Theory Lecture 17 Iterative solvers for linear equations Daniel A. Spielman October 31, 2012 17.1 About these notes These notes are not necessarily an accurate representation of what happened
More informationEssential Mathematics
Appendix B 1211 Appendix B Essential Mathematics Exponential Arithmetic Exponential notation is used to express very large and very small numbers as a product of two numbers. The first number of the product,
More informationIntegration of Rational Functions by Partial Fractions
Title Integration of Rational Functions by MATH 1700 MATH 1700 1 / 11 Readings Readings Readings: Section 7.4 MATH 1700 2 / 11 Rational functions A rational function is one of the form where P and Q are
More informationLecture 8 : Eigenvalues and Eigenvectors
CPS290: Algorithmic Foundations of Data Science February 24, 2017 Lecture 8 : Eigenvalues and Eigenvectors Lecturer: Kamesh Munagala Scribe: Kamesh Munagala Hermitian Matrices It is simpler to begin with
More informationMidterm Review. Igor Yanovsky (Math 151A TA)
Midterm Review Igor Yanovsky (Math 5A TA) Root-Finding Methods Rootfinding methods are designed to find a zero of a function f, that is, to find a value of x such that f(x) =0 Bisection Method To apply
More informationMath 473: Practice Problems for Test 1, Fall 2011, SOLUTIONS
Math 473: Practice Problems for Test 1, Fall 011, SOLUTIONS Show your work: 1. (a) Compute the Taylor polynomials P n (x) for f(x) = sin x and x 0 = 0. Solution: Compute f(x) = sin x, f (x) = cos x, f
More information3.1 Interpolation and the Lagrange Polynomial
MATH 4073 Chapter 3 Interpolation and Polynomial Approximation Fall 2003 1 Consider a sample x x 0 x 1 x n y y 0 y 1 y n. Can we get a function out of discrete data above that gives a reasonable estimate
More informationNewton s Method and Linear Approximations
Newton s Method and Linear Approximations Newton s Method for finding roots Goal: Where is f (x) =0? f (x) =x 7 +3x 3 +7x 2 1 2-1 -0.5 0.5-2 Newton s Method for finding roots Goal: Where is f (x) =0? f
More informationIntegration of Rational Functions by Partial Fractions
Title Integration of Rational Functions by Partial Fractions MATH 1700 December 6, 2016 MATH 1700 Partial Fractions December 6, 2016 1 / 11 Readings Readings Readings: Section 7.4 MATH 1700 Partial Fractions
More information1 Backward and Forward Error
Math 515 Fall, 2008 Brief Notes on Conditioning, Stability and Finite Precision Arithmetic Most books on numerical analysis, numerical linear algebra, and matrix computations have a lot of material covering
More informationJustify all your answers and write down all important steps. Unsupported answers will be disregarded.
Numerical Analysis FMN011 10058 The exam lasts 4 hours and has 13 questions. A minimum of 35 points out of the total 70 are required to get a passing grade. These points will be added to those you obtained
More informationCME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 9
CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 9 GENE H GOLUB 1 Error Analysis of Gaussian Elimination In this section, we will consider the case of Gaussian elimination and perform a detailed
More informationScientific Computing
Scientific Computing Direct solution methods Martin van Gijzen Delft University of Technology October 3, 2018 1 Program October 3 Matrix norms LU decomposition Basic algorithm Cost Stability Pivoting Pivoting
More informationNewton s Method and Linear Approximations 10/19/2011
Newton s Method and Linear Approximations 10/19/2011 Curves are tricky. Lines aren t. Newton s Method and Linear Approximations 10/19/2011 Newton s Method Goal: Where is f (x) =0? f (x) =x 7 +3x 3 +7x
More informationMathematics for Engineers. Numerical mathematics
Mathematics for Engineers Numerical mathematics Integers Determine the largest representable integer with the intmax command. intmax ans = int32 2147483647 2147483647+1 ans = 2.1475e+09 Remark The set
More information8.7 Taylor s Inequality Math 2300 Section 005 Calculus II. f(x) = ln(1 + x) f(0) = 0
8.7 Taylor s Inequality Math 00 Section 005 Calculus II Name: ANSWER KEY Taylor s Inequality: If f (n+) is continuous and f (n+) < M between the center a and some point x, then f(x) T n (x) M x a n+ (n
More informationApplied Mathematics 205. Unit 0: Overview of Scientific Computing. Lecturer: Dr. David Knezevic
Applied Mathematics 205 Unit 0: Overview of Scientific Computing Lecturer: Dr. David Knezevic Scientific Computing Computation is now recognized as the third pillar of science (along with theory and experiment)
More informationEquations in Quadratic Form
Equations in Quadratic Form MATH 101 College Algebra J. Robert Buchanan Department of Mathematics Summer 2012 Objectives In this lesson we will learn to: make substitutions that allow equations to be written
More informationRoots and Coefficients Polynomials Preliminary Maths Extension 1
Preliminary Maths Extension Question If, and are the roots of x 5x x 0, find the following. (d) (e) Question If p, q and r are the roots of x x x 4 0, evaluate the following. pq r pq qr rp p q q r r p
More informationCME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 5. Ax = b.
CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 5 GENE H GOLUB Suppose we want to solve We actually have an approximation ξ such that 1 Perturbation Theory Ax = b x = ξ + e The question is, how
More informationNon-polynomial Least-squares fitting
Applied Math 205 Last time: piecewise polynomial interpolation, least-squares fitting Today: underdetermined least squares, nonlinear least squares Homework 1 (and subsequent homeworks) have several parts
More informationComputational Methods. Least Squares Approximation/Optimization
Computational Methods Least Squares Approximation/Optimization Manfred Huber 2011 1 Least Squares Least squares methods are aimed at finding approximate solutions when no precise solution exists Find the
More informationReview I: Interpolation
Review I: Interpolation Varun Shankar January, 206 Introduction In this document, we review interpolation by polynomials. Unlike many reviews, we will not stop there: we will discuss how to differentiate
More informationBinary floating point
Binary floating point Notes for 2017-02-03 Why do we study conditioning of problems? One reason is that we may have input data contaminated by noise, resulting in a bad solution even if the intermediate
More informationMath 502 Fall 2005 Solutions to Homework 3
Math 502 Fall 2005 Solutions to Homework 3 (1) As shown in class, the relative distance between adjacent binary floating points numbers is 2 1 t, where t is the number of digits in the mantissa. Since
More informationNumerical Methods. Elena loli Piccolomini. Civil Engeneering. piccolom. Metodi Numerici M p. 1/??
Metodi Numerici M p. 1/?? Numerical Methods Elena loli Piccolomini Civil Engeneering http://www.dm.unibo.it/ piccolom elena.loli@unibo.it Metodi Numerici M p. 2/?? Least Squares Data Fitting Measurement
More informationCME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 6
CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 6 GENE H GOLUB Issues with Floating-point Arithmetic We conclude our discussion of floating-point arithmetic by highlighting two issues that frequently
More informationDepartment of Applied Mathematics and Theoretical Physics. AMA 204 Numerical analysis. Exam Winter 2004
Department of Applied Mathematics and Theoretical Physics AMA 204 Numerical analysis Exam Winter 2004 The best six answers will be credited All questions carry equal marks Answer all parts of each question
More informationMath 411 Preliminaries
Math 411 Preliminaries Provide a list of preliminary vocabulary and concepts Preliminary Basic Netwon s method, Taylor series expansion (for single and multiple variables), Eigenvalue, Eigenvector, Vector
More informationNotes on floating point number, numerical computations and pitfalls
Notes on floating point number, numerical computations and pitfalls November 6, 212 1 Floating point numbers An n-digit floating point number in base β has the form x = ±(.d 1 d 2 d n ) β β e where.d 1
More informationLecture 4: Numerical solution of ordinary differential equations
Lecture 4: Numerical solution of ordinary differential equations Department of Mathematics, ETH Zürich General explicit one-step method: Consistency; Stability; Convergence. High-order methods: Taylor
More informationComputer Arithmetic. MATH 375 Numerical Analysis. J. Robert Buchanan. Fall Department of Mathematics. J. Robert Buchanan Computer Arithmetic
Computer Arithmetic MATH 375 Numerical Analysis J. Robert Buchanan Department of Mathematics Fall 2013 Machine Numbers When performing arithmetic on a computer (laptop, desktop, mainframe, cell phone,
More informationQueens College, CUNY, Department of Computer Science Numerical Methods CSCI 361 / 761 Spring 2018 Instructor: Dr. Sateesh Mane.
Queens College, CUNY, Department of Computer Science Numerical Methods CSCI 361 / 761 Spring 2018 Instructor: Dr. Sateesh Mane c Sateesh R. Mane 2018 3 Lecture 3 3.1 General remarks March 4, 2018 This
More informationLinear Algebra Review
Chapter 1 Linear Algebra Review It is assumed that you have had a course in linear algebra, and are familiar with matrix multiplication, eigenvectors, etc. I will review some of these terms here, but quite
More information(0, 0), (1, ), (2, ), (3, ), (4, ), (5, ), (6, ).
1 Interpolation: The method of constructing new data points within the range of a finite set of known data points That is if (x i, y i ), i = 1, N are known, with y i the dependent variable and x i [x
More informationElements of Floating-point Arithmetic
Elements of Floating-point Arithmetic Sanzheng Qiao Department of Computing and Software McMaster University July, 2012 Outline 1 Floating-point Numbers Representations IEEE Floating-point Standards Underflow
More informationLECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel
LECTURE NOTES on ELEMENTARY NUMERICAL METHODS Eusebius Doedel TABLE OF CONTENTS Vector and Matrix Norms 1 Banach Lemma 20 The Numerical Solution of Linear Systems 25 Gauss Elimination 25 Operation Count
More informationCOMP 558 lecture 18 Nov. 15, 2010
Least squares We have seen several least squares problems thus far, and we will see more in the upcoming lectures. For this reason it is good to have a more general picture of these problems and how to
More informationHere is an example of a block diagonal matrix with Jordan Blocks on the diagonal: J
Class Notes 4: THE SPECTRAL RADIUS, NORM CONVERGENCE AND SOR. Math 639d Due Date: Feb. 7 (updated: February 5, 2018) In the first part of this week s reading, we will prove Theorem 2 of the previous class.
More informationLecture2 The implicit function theorem. Bifurcations.
Lecture2 The implicit function theorem. Bifurcations. 1 Review:Newton s method. The existence theorem - the assumptions. Basins of attraction. 2 The implicit function theorem. 3 Bifurcations of iterations.
More informationCenterville High School Curriculum Mapping Algebra II 1 st Nine Weeks
Centerville High School Curriculum Mapping Algebra II 1 st Nine Weeks Chapter/ Lesson Common Core Standard(s) 1-1 SMP1 1. How do you use a number line to graph and order real numbers? 2. How do you identify
More informationMATH10001 Mathematical Workshop Graph Fitting Project part 2
MATH10001 Mathematical Workshop Graph Fitting Project part 2 Polynomial models Modelling a set of data with a polynomial curve can be convenient because polynomial functions are particularly easy to differentiate
More informationLecture Note 3: Interpolation and Polynomial Approximation. Xiaoqun Zhang Shanghai Jiao Tong University
Lecture Note 3: Interpolation and Polynomial Approximation Xiaoqun Zhang Shanghai Jiao Tong University Last updated: October 10, 2015 2 Contents 1.1 Introduction................................ 3 1.1.1
More informationMathematical preliminaries and error analysis
Mathematical preliminaries and error analysis Tsung-Ming Huang Department of Mathematics National Taiwan Normal University, Taiwan September 12, 2015 Outline 1 Round-off errors and computer arithmetic
More informationOne Dimensional Dynamical Systems
16 CHAPTER 2 One Dimensional Dynamical Systems We begin by analyzing some dynamical systems with one-dimensional phase spaces, and in particular their bifurcations. All equations in this Chapter are scalar
More information1. The positive zero of y = x 2 + 2x 3/5 is, to the nearest tenth, equal to
SAT II - Math Level Test #0 Solution SAT II - Math Level Test No. 1. The positive zero of y = x + x 3/5 is, to the nearest tenth, equal to (A) 0.8 (B) 0.7 + 1.1i (C) 0.7 (D) 0.3 (E). 3 b b 4ac Using Quadratic
More informationThe final is cumulative, but with more emphasis on chapters 3 and 4. There will be two parts.
Math 141 Review for Final The final is cumulative, but with more emphasis on chapters 3 and 4. There will be two parts. Part 1 (no calculator) graphing (polynomial, rational, linear, exponential, and logarithmic
More informationOne-Sided Difference Formula for the First Derivative
POLYTECHNIC UNIVERSITY Department of Computer and Information Science One-Sided Difference Formula for the First Derivative K. Ming Leung Abstract: Derive a one-sided formula for the first derive of a
More informationNumerical Methods for Differential Equations Mathematical and Computational Tools
Numerical Methods for Differential Equations Mathematical and Computational Tools Gustaf Söderlind Numerical Analysis, Lund University Contents V4.16 Part 1. Vector norms, matrix norms and logarithmic
More informationMath Solutions to homework 5
Math 75 - Solutions to homework 5 Cédric De Groote November 9, 207 Problem (7. in the book): Let {e n } be a complete orthonormal sequence in a Hilbert space H and let λ n C for n N. Show that there is
More informationSolutions to Dynamical Systems 2010 exam. Each question is worth 25 marks.
Solutions to Dynamical Systems exam Each question is worth marks [Unseen] Consider the following st order differential equation: dy dt Xy yy 4 a Find and classify all the fixed points of Hence draw the
More informationHomework 1 Solutions Math 587
Homework 1 Solutions Math 587 1) Find positive functions f(x),g(x), continuous on (0, ) so that f = O(g(x)) as x 0, but lim x 0 f(x)/g(x) does not exist. One can take f(x) = sin(1/x) and g(x) = 1, for
More informationIntroduction. Chapter One
Chapter One Introduction The aim of this book is to describe and explain the beautiful mathematical relationships between matrices, moments, orthogonal polynomials, quadrature rules and the Lanczos and
More informationERROR IN LINEAR INTERPOLATION
ERROR IN LINEAR INTERPOLATION Let P 1 (x) be the linear polynomial interpolating f (x) at x 0 and x 1. Assume f (x) is twice continuously differentiable on an interval [a, b] which contains the points
More information2-7 Solving Quadratic Inequalities. ax 2 + bx + c > 0 (a 0)
Quadratic Inequalities In One Variable LOOKS LIKE a quadratic equation but Doesn t have an equal sign (=) Has an inequality sign (>,
More informationPARTIAL DIFFERENTIAL EQUATIONS
MATHEMATICAL METHODS PARTIAL DIFFERENTIAL EQUATIONS I YEAR B.Tech By Mr. Y. Prabhaker Reddy Asst. Professor of Mathematics Guru Nanak Engineering College Ibrahimpatnam, Hyderabad. SYLLABUS OF MATHEMATICAL
More informationCS100: DISCRETE STRUCTURES. Lecture 3 Matrices Ch 3 Pages:
CS100: DISCRETE STRUCTURES Lecture 3 Matrices Ch 3 Pages: 246-262 Matrices 2 Introduction DEFINITION 1: A matrix is a rectangular array of numbers. A matrix with m rows and n columns is called an m x n
More informationPreliminary Examination in Numerical Analysis
Department of Applied Mathematics Preliminary Examination in Numerical Analysis August 7, 06, 0 am pm. Submit solutions to four (and no more) of the following six problems. Show all your work, and justify
More informationPowerPoints organized by Dr. Michael R. Gustafson II, Duke University
Part 1 Chapter 4 Roundoff and Truncation Errors PowerPoints organized by Dr. Michael R. Gustafson II, Duke University All images copyright The McGraw-Hill Companies, Inc. Permission required for reproduction
More informationERROR AND SENSITIVTY ANALYSIS FOR SYSTEMS OF LINEAR EQUATIONS. Perturbation analysis for linear systems (Ax = b)
ERROR AND SENSITIVTY ANALYSIS FOR SYSTEMS OF LINEAR EQUATIONS Conditioning of linear systems. Estimating errors for solutions of linear systems Backward error analysis Perturbation analysis for linear
More informationMA3457/CS4033: Numerical Methods for Calculus and Differential Equations
MA3457/CS4033: Numerical Methods for Calculus and Differential Equations Course Materials P A R T II B 14 2014-2015 1 2. APPROXIMATION CLASS 9 Approximation Key Idea Function approximation is closely related
More informationthe method of steepest descent
MATH 3511 Spring 2018 the method of steepest descent http://www.phys.uconn.edu/ rozman/courses/m3511_18s/ Last modified: February 6, 2018 Abstract The Steepest Descent is an iterative method for solving
More informationLearning Module 1 - Basic Algebra Review (Appendix A)
Learning Module 1 - Basic Algebra Review (Appendix A) Element 1 Real Numbers and Operations on Polynomials (A.1, A.2) Use the properties of real numbers and work with subsets of the real numbers Determine
More informationOutline. Math Numerical Analysis. Intermediate Value Theorem. Lecture Notes Zeros and Roots. Joseph M. Mahaffy,
Outline Math 541 - Numerical Analysis Lecture Notes Zeros and Roots Joseph M. Mahaffy, jmahaffy@mail.sdsu.edu Department of Mathematics and Statistics Dynamical Systems Group Computational Sciences Research
More informationTaylor Series and Numerical Approximations
Taylor Series and Numerical Approximations Hilary Weller h.weller@reading.ac.uk August 7, 05 An introduction to the concept of a Taylor series and how these are used in numerical analysis to find numerical
More informationb = 2, c = 3, we get x = 0.3 for the positive root. Ans. (D) x 2-2x - 8 < 0, or (x - 4)(x + 2) < 0, Therefore -2 < x < 4 Ans. (C)
SAT II - Math Level 2 Test #02 Solution 1. The positive zero of y = x 2 + 2x is, to the nearest tenth, equal to (A) 0.8 (B) 0.7 + 1.1i (C) 0.7 (D) 0.3 (E) 2.2 ± Using Quadratic formula, x =, with a = 1,
More informationMath 221 Notes on Rolle s Theorem, The Mean Value Theorem, l Hôpital s rule, and the Taylor-Maclaurin formula. 1. Two theorems
Math 221 Notes on Rolle s Theorem, The Mean Value Theorem, l Hôpital s rule, and the Taylor-Maclaurin formula 1. Two theorems Rolle s Theorem. If a function y = f(x) is differentiable for a x b and if
More informationMath Numerical Analysis
Math 541 - Numerical Analysis Lecture Notes Zeros and Roots Joseph M. Mahaffy, jmahaffy@mail.sdsu.edu Department of Mathematics and Statistics Dynamical Systems Group Computational Sciences Research Center
More informationMATH ASSIGNMENT 07 SOLUTIONS. 8.1 Following is census data showing the population of the US between 1900 and 2000:
MATH4414.01 ASSIGNMENT 07 SOLUTIONS 8.1 Following is census data showing the population of the US between 1900 and 2000: Years after 1900 Population in millions 0 76.0 20 105.7 40 131.7 60 179.3 80 226.5
More informationParallel programming practices for the solution of Sparse Linear Systems (motivated by computational physics and graphics)
Parallel programming practices for the solution of Sparse Linear Systems (motivated by computational physics and graphics) Eftychios Sifakis CS758 Guest Lecture - 19 Sept 2012 Introduction Linear systems
More informationThe Behavior of Algorithms in Practice 2/21/2002. Lecture 4. ɛ 1 x 1 y ɛ 1 x 1 1 = x y 1 1 = y 1 = 1 y 2 = 1 1 = 0 1 1
8.409 The Behavior of Algorithms in Practice //00 Lecture 4 Lecturer: Dan Spielman Scribe: Matthew Lepinski A Gaussian Elimination Example To solve: [ ] [ ] [ ] x x First factor the matrix to get: [ ]
More informationINTERPOLATION Background Polynomial Approximation Problem:
INTERPOLATION Background Polynomial Approximation Problem: given f(x) C[a, b], find P n (x) = a 0 + a 1 x + a 2 x 2 + + a n x n with P n (x) close to f(x) for x [a, b]. Motivations: f(x) might be difficult
More informationChapter 1: Precalculus Review
: Precalculus Review Math 115 17 January 2018 Overview 1 Important Notation 2 Exponents 3 Polynomials 4 Rational Functions 5 Cartesian Coordinates 6 Lines Notation Intervals: Interval Notation (a, b) (a,
More informationChapter III. Stability of Linear Systems
1 Chapter III Stability of Linear Systems 1. Stability and state transition matrix 2. Time-varying (non-autonomous) systems 3. Time-invariant systems 1 STABILITY AND STATE TRANSITION MATRIX 2 In this chapter,
More information