Least-Squares Fitting of Data with Polynomials
|
|
- Bartholomew Washington
- 6 years ago
- Views:
Transcription
1 Least-Squares Fitting of Data with Polynomials David Eberly, Geometric Tools, Redmond WA This work is licensed under the Creative Commons Attribution 4.0 International License. To view a copy of this license, visit or send a letter to Creative Commons, PO Box 1866, Mountain View, CA 94042, USA. Created: October 17, 2005 Last Modified: February 9, 2008 Contents 1 Fitting with Standard Polynomials 2 2 Fitting with Orthogonal Polynomials 3 1
2 1 Fitting with Standard Polynomials Given a set of n points x k, y k with x k+1 > x k for all k, the goal is to fit the data with a polynomial of degree d, namely, d y = c i x i The least-squares method chooses the coefficients c i to minimize the squared error function n d 2 Ec 0,..., c d = c i x i k y k This is a nonnegative quadratic function with respect to the coefficients and obtains its global minimum when the gradient of E is zero; that is, the partial derivatives of E must be zero: 0 = E d = 2 c i x i k y k x j k c j for 0 j d. The equation simplifies to d x j+i k c i = x j k y k This is a linear system of d + 1 equations in d + 1 unknowns and is of the form A T Ac = A T y 1 where A = [a ki ] is an n d + 1 matrix whose general entry is a ki = x i k, c = [c j] is a d column vector, and y = [y k ] is an n 1 column vector. The coefficient matrix of the system is A T A = xd k xd+1 k 1 x k x k x2 k xd k xd+1 k x2d k The assumption that the x k are increasing guarantees that A T A is invertible, so the coefficients of the polynomial are c = A T A 1 A T y Although this is a concise mathematical formulation, a computer implementation using a fixed-size floatingpoint number can suffer from numerical problems when inverting the coefficient matrix. The inversion is ill-conditioned when x 0 = min k x k or x = max k x k is large and the degree d is large. In some cases, a preconditioning of the data points will suffice to give a reasonably accurate set of coefficients. Map the input points by translating the x-values and then uniformly scale the x- and y-values so that the translated and scaled x-values lie in the interval [0, 1]. Such a transformation is xk x 0 y k u k, v k =, x x 0 x x 0 2
3 The transformed data points are fit with a polynomial v = d c i u i using the least-squares method. The conditioning of the matrix A is sometimes better by using the transformation approach, but not all the time. The evaluation of the polynomial at an x value is illustrated by the pseudocode, where the degree, coefficients, and data points are assumed to be globally accessible: f l o a t E v a l u a t e P o l y n o m i a l f l o a t x f l o a t u = x x [ 0 ] / x [ ] x [ 0 ] ; f l o a t v = c [ d e g r e e ] ; f o r i n t i = degree 1; i >= 0 ; i v = u ; v += c [ i ] ; y = x [ ] x [ 0 ] v ; Another attempt to solve the matrix system uses the Cholesky decomposition. The matrix A T A is factored into BB T, where B is lower triangular and B T is upper triangular. The matrix B is then inverted using LU methods to produce B T c = B 1 A T y. The matrix B T is similarly inverted to solve for c = B T B 1 A T y. Even this method can suffer from numerical problems with fixed-size floating-point numbers. Alternatively, a computationally expensive method is to use exact rational arithmetic, where the data points have floating-point components that are exactly represented as rational numbers. 2 Fitting with Orthogonal Polynomials Another way to view the fitting problem is to use different polynomials than 1, x, x 2,..., x d, call them p i x for 0 i d with Degreep i = i. The idea is to obtain a set of equations like Equation 1, but where A T A is a diagonal matrix. In this case, the problems with numerical round-off errors are eliminated other than the potential ones associated with divisions. The polynomial we want to fit is of the form The error function to minimize is y = Ec 0,..., c d = d c i p i x 2 n d 2 c i p i x k y k Computing the partial derivatives and setting to zero, 0 = E d = 2 c i p i x k y k p j x k c j 3
4 for 0 j d. The equation simplifies to d p i x k p j x k c i = p j x k y k If we can choose the polynomials so that p i x k p j x k = 0, i j 3 then the set of equations have the easily computed solution c i = p ix k y k p2 i x k 4 The condition in Equation 3 is an orthogonality condition. A set of polynomials satisfying the condition are said to be orthogonal polynomials. Notice that the orthogonality condition depends on the data points themselves. This is different from the standard polynomial fitting where 1, x,..., x d are chosen independently of the input data. Thus, the fitting with orthogonal polynomials may be viewed as a data-driven method. There are a variety of ways to generate orthogonal polynomials. One method is illustrated next. Choose which is a constant polynomial degree 0 for all x. Also choose It is easy to verify the orthogonality of the two polynomials: p 0 x k p 1 x k = x k a 1 = p 0 x = 1 5 p 1 x = x a 1, a 1 = 1 x k 6 n x k na 1 = na 1 na 1 = 0 The remaining polynomials are generated by a second-order recurrence relationship, where p m+1 x = x a m+1 p m x b m p m 1 x, m 1 7 a m+1 = x kp 2 mx k p2 mx k, b m = p2 mx k p2 m 1 x k The expressions for a m+1 and b m may be computed from three equations obtained by multiplying Equation 7 by p m 1 x, by p m x, and by p m+1 x, and then substituting in x k and summing over k. In the process of doing so, you must use the orthogonality condition of Equation 3. The problems of computing the coefficients in Equation 4 and of evaluating the polynomial in Equation 2 both reduce to having to evaluate the orthogonal polynomials of Equations 5, 6, and 7. The pseudocode for this is listed next. It is assumed that a m+1 and b m are globally accessible. 4
5 f l o a t E v a l u a t e O r t h o g o n a l i n t power, f l o a t x f l o a t p0 = 1 ; i f power == 0 r e t u r n p0 ; f l o a t p1 = x a [ 1 ] ; i f power == 1 r e t u r n p1 ; f l o a t p2 ; f o r i n t m = 2 ; m <= power ; m++ p2 = x a [m] p1 b [m 1] p0 ; p0 = p1 ; p1 = p2 ; r e t u r n p2 ; v o i d ComputeAB i n t n, f l o a t x [ ] a [ 1 ] = 0 ; f o r k = 0 ; k < n ; k++ a [ 1 ] += x [ k ] ; a [ 1 ] /= n ; f o r m = 1 ; m < d e g r e e ; m++ f l o a t sum0 = 0, sum1 = 0, sum2 = 0 ; f o r k = 0 ; k < n ; k++ f l o a t tmp0 = E v a l u a t e O r t h o g o n a l m 1,x [ k ] ; f l o a t tmp1 = E v a l u a t e O r t h o g o n a l m, x [ k ] ; sum0 += tmp0 tmp0 ; sum1 += tmp1 tmp1 ; sum2 += x [ k ] tmp1 tmp1 ; a [m+1] = sum2/sum1 ; b [m] = sum1/sum0 ; The function ComputeAB should be called first to initialize the a m and b m values that are used in general polynomial evaluations. 5
Distance Between Ellipses in 2D
Distance Between Ellipses in 2D David Eberly, Geometric Tools, Redmond WA 98052 https://www.geometrictools.com/ This work is licensed under the Creative Commons Attribution 4.0 International License. To
More informationInformation About Ellipses
Information About Ellipses David Eberly, Geometric Tools, Redmond WA 9805 https://www.geometrictools.com/ This work is licensed under the Creative Commons Attribution 4.0 International License. To view
More informationFitting a Natural Spline to Samples of the Form (t, f(t))
Fitting a Natural Spline to Samples of the Form (t, f(t)) David Eberly, Geometric Tools, Redmond WA 9852 https://wwwgeometrictoolscom/ This work is licensed under the Creative Commons Attribution 4 International
More informationA Relationship Between Minimum Bending Energy and Degree Elevation for Bézier Curves
A Relationship Between Minimum Bending Energy and Degree Elevation for Bézier Curves David Eberly, Geometric Tools, Redmond WA 9852 https://www.geometrictools.com/ This work is licensed under the Creative
More informationLow-Degree Polynomial Roots
Low-Degree Polynomial Roots David Eberly, Geometric Tools, Redmond WA 98052 https://www.geometrictools.com/ This work is licensed under the Creative Commons Attribution 4.0 International License. To view
More informationSolving Systems of Polynomial Equations
Solving Systems of Polynomial Equations David Eberly, Geometric Tools, Redmond WA 98052 https://www.geometrictools.com/ This work is licensed under the Creative Commons Attribution 4.0 International License.
More informationIntersection of Ellipses
Intersection of Ellipses David Eberly, Geometric Tools, Redmond WA 98052 https://www.geometrictools.com/ This work is licensed under the Creative Commons Attribution 4.0 International License. To view
More informationPerspective Projection of an Ellipse
Perspective Projection of an Ellipse David Eberly, Geometric ools, Redmond WA 98052 https://www.geometrictools.com/ his work is licensed under the Creative Commons Attribution 4.0 International License.
More informationIntersection of Rectangle and Ellipse
Intersection of Rectangle and Ellipse David Eberly, Geometric Tools, Redmond WA 98052 https://www.geometrictools.com/ This work is licensed under the Creative Commons Attribution 4.0 International License.
More informationThe Laplace Expansion Theorem: Computing the Determinants and Inverses of Matrices
The Laplace Expansion Theorem: Computing the Determinants and Inverses of Matrices David Eberly, Geometric Tools, Redmond WA 98052 https://www.geometrictools.com/ This work is licensed under the Creative
More informationDistance from a Point to an Ellipse, an Ellipsoid, or a Hyperellipsoid
Distance from a Point to an Ellipse, an Ellipsoid, or a Hyperellipsoid David Eberly, Geometric Tools, Redmond WA 9805 https://www.geometrictools.com/ This work is licensed under the Creative Commons Attribution
More informationLeast Squares Fitting of Data by Linear or Quadratic Structures
Least Squares Fitting of Data by Linear or Quadratic Structures David Eberly, Geometric Tools, Redmond WA 98052 https://www.geometrictools.com/ This work is licensed under the Creative Commons Attribution
More informationIntersection of Ellipsoids
Intersection of Ellipsoids David Eberly, Geometric Tools, Redmond WA 9805 https://www.geometrictools.com/ This work is licensed under the Creative Commons Attribution 4.0 International License. To view
More informationReconstructing an Ellipsoid from its Perspective Projection onto a Plane
Reconstructing an Ellipsoid from its Perspective Projection onto a Plane David Eberly, Geometric Tools, Redmond WA 98052 https://www.geometrictools.com/ This work is licensed under the Creative Commons
More informationDistance from Line to Rectangle in 3D
Distance from Line to Rectangle in 3D David Eberly, Geometric Tools, Redmond WA 98052 https://www.geometrictools.com/ This work is licensed under the Creative Commons Attribution 4.0 International License.
More informationIntersection of Infinite Cylinders
Intersection of Infinite Cylinders David Eberly, Geometric Tools, Redmond WA 980 https://www.geometrictools.com/ This work is licensed under the Creative Commons Attribution 4.0 International License.
More informationComputing Orthonormal Sets in 2D, 3D, and 4D
Computing Orthonormal Sets in 2D, 3D, and 4D David Eberly, Geometric Tools, Redmond WA 98052 https://www.geometrictools.com/ This work is licensed under the Creative Commons Attribution 4.0 International
More informationIntersection of Objects with Linear and Angular Velocities using Oriented Bounding Boxes
Intersection of Objects with Linear and Angular Velocities using Oriented Bounding Boxes David Eberly, Geometric Tools, Redmond WA 98052 https://www.geometrictools.com/ This work is licensed under the
More informationA matrix over a field F is a rectangular array of elements from F. The symbol
Chapter MATRICES Matrix arithmetic A matrix over a field F is a rectangular array of elements from F The symbol M m n (F ) denotes the collection of all m n matrices over F Matrices will usually be denoted
More informationInterpolation of Rigid Motions in 3D
Interpolation of Rigid Motions in 3D David Eberly, Geometric Tools, Redmond WA 98052 https://www.geometrictools.com/ This work is licensed under the Creative Commons Attribution 4.0 International License.
More informationPolyhedral Mass Properties (Revisited)
Polyhedral Mass Properties (Revisited) David Eberly, Geometric Tools, Redmond WA 98052 https://www.geometrictools.com/ This work is licensed under the Creative Commons Attribution 4.0 International License.
More informationLinear Algebraic Equations
Linear Algebraic Equations 1 Fundamentals Consider the set of linear algebraic equations n a ij x i b i represented by Ax b j with [A b ] [A b] and (1a) r(a) rank of A (1b) Then Axb has a solution iff
More informationDistance to Circles in 3D
Distance to Circles in 3D David Eberly, Geometric Tools, Redmond WA 98052 https://www.geometrictools.com/ This work is licensed under the Creative Commons Attribution 4.0 International License. To view
More informationThe Area of Intersecting Ellipses
The Area of Intersecting Ellipses David Eberly, Geometric Tools, Redmond WA 98052 https://www.geometrictools.com/ This work is licensed under the Creative Commons Attribution 4.0 International License.
More informationLinear Algebra Section 2.6 : LU Decomposition Section 2.7 : Permutations and transposes Wednesday, February 13th Math 301 Week #4
Linear Algebra Section. : LU Decomposition Section. : Permutations and transposes Wednesday, February 1th Math 01 Week # 1 The LU Decomposition We learned last time that we can factor a invertible matrix
More informationELEMENTARY LINEAR ALGEBRA
ELEMENTARY LINEAR ALGEBRA K R MATTHEWS DEPARTMENT OF MATHEMATICS UNIVERSITY OF QUEENSLAND First Printing, 99 Chapter LINEAR EQUATIONS Introduction to linear equations A linear equation in n unknowns x,
More informationDynamic Collision Detection using Oriented Bounding Boxes
Dynamic Collision Detection using Oriented Bounding Boxes David Eberly, Geometric Tools, Redmond WA 98052 https://www.geometrictools.com/ This work is licensed under the Creative Commons Attribution 4.0
More informationMATH 3511 Lecture 1. Solving Linear Systems 1
MATH 3511 Lecture 1 Solving Linear Systems 1 Dmitriy Leykekhman Spring 2012 Goals Review of basic linear algebra Solution of simple linear systems Gaussian elimination D Leykekhman - MATH 3511 Introduction
More informationInverses. Stephen Boyd. EE103 Stanford University. October 28, 2017
Inverses Stephen Boyd EE103 Stanford University October 28, 2017 Outline Left and right inverses Inverse Solving linear equations Examples Pseudo-inverse Left and right inverses 2 Left inverses a number
More informationNumerical Optimization
Numerical Optimization Unit 2: Multivariable optimization problems Che-Rung Lee Scribe: February 28, 2011 (UNIT 2) Numerical Optimization February 28, 2011 1 / 17 Partial derivative of a two variable function
More informationPrincipal Curvatures of Surfaces
Principal Curvatures of Surfaces David Eberly, Geometric Tools, Redmond WA 98052 https://www.geometrictools.com/ This work is licensed under the Creative Commons Attribution 4.0 International License.
More informationMidterm for Introduction to Numerical Analysis I, AMSC/CMSC 466, on 10/29/2015
Midterm for Introduction to Numerical Analysis I, AMSC/CMSC 466, on 10/29/2015 The test lasts 1 hour and 15 minutes. No documents are allowed. The use of a calculator, cell phone or other equivalent electronic
More informationMatrix decompositions
Matrix decompositions How can we solve Ax = b? 1 Linear algebra Typical linear system of equations : x 1 x +x = x 1 +x +9x = 0 x 1 +x x = The variables x 1, x, and x only appear as linear terms (no powers
More informationAn Approximation for the Inverse Square Root Function
An Approximation for the Inverse Square Root Function David Eberly, Geometric Tools, Redmond WA 98052 https://www.geometrictools.com/ This work is licensed under the Creative Commons Attribution 4.0 International
More informationMatrix decompositions
Matrix decompositions How can we solve Ax = b? 1 Linear algebra Typical linear system of equations : x 1 x +x = x 1 +x +9x = 0 x 1 +x x = The variables x 1, x, and x only appear as linear terms (no powers
More informationa 11 x 1 + a 12 x a 1n x n = b 1 a 21 x 1 + a 22 x a 2n x n = b 2.
Chapter 1 LINEAR EQUATIONS 11 Introduction to linear equations A linear equation in n unknowns x 1, x,, x n is an equation of the form a 1 x 1 + a x + + a n x n = b, where a 1, a,, a n, b are given real
More informationLinear Algebra Linear Algebra : Matrix decompositions Monday, February 11th Math 365 Week #4
Linear Algebra Linear Algebra : Matrix decompositions Monday, February 11th Math Week # 1 Saturday, February 1, 1 Linear algebra Typical linear system of equations : x 1 x +x = x 1 +x +9x = 0 x 1 +x x
More informationMoving Along a Curve with Specified Speed
Moving Along a Curve with Specified Speed David Eberly, Geometric Tools, Redmond WA 98052 https://www.geometrictools.com/ This work is licensed under the Creative Commons Attribution 4.0 International
More informationA A x i x j i j (i, j) (j, i) Let. Compute the value of for and
7.2 - Quadratic Forms quadratic form on is a function defined on whose value at a vector in can be computed by an expression of the form, where is an symmetric matrix. The matrix R n Q R n x R n Q(x) =
More informationCS137 Introduction to Scientific Computing Winter Quarter 2004 Solutions to Homework #3
CS137 Introduction to Scientific Computing Winter Quarter 2004 Solutions to Homework #3 Felix Kwok February 27, 2004 Written Problems 1. (Heath E3.10) Let B be an n n matrix, and assume that B is both
More informationIntegration of Rational Functions by Partial Fractions
Integration of Rational Functions by Partial Fractions Part 2: Integrating Rational Functions Rational Functions Recall that a rational function is the quotient of two polynomials. x + 3 x + 2 x + 2 x
More informationEE731 Lecture Notes: Matrix Computations for Signal Processing
EE731 Lecture Notes: Matrix Computations for Signal Processing James P. Reilly c Department of Electrical and Computer Engineering McMaster University September 22, 2005 0 Preface This collection of ten
More informationMatrices and Matrix Algebra.
Matrices and Matrix Algebra 3.1. Operations on Matrices Matrix Notation and Terminology Matrix: a rectangular array of numbers, called entries. A matrix with m rows and n columns m n A n n matrix : a square
More informationReview Questions REVIEW QUESTIONS 71
REVIEW QUESTIONS 71 MATLAB, is [42]. For a comprehensive treatment of error analysis and perturbation theory for linear systems and many other problems in linear algebra, see [126, 241]. An overview of
More informationReview of matrices. Let m, n IN. A rectangle of numbers written like A =
Review of matrices Let m, n IN. A rectangle of numbers written like a 11 a 12... a 1n a 21 a 22... a 2n A =...... a m1 a m2... a mn where each a ij IR is called a matrix with m rows and n columns or an
More informationNumerical Methods - Numerical Linear Algebra
Numerical Methods - Numerical Linear Algebra Y. K. Goh Universiti Tunku Abdul Rahman 2013 Y. K. Goh (UTAR) Numerical Methods - Numerical Linear Algebra I 2013 1 / 62 Outline 1 Motivation 2 Solving Linear
More informationAlgebra I. Course Outline
Algebra I Course Outline I. The Language of Algebra A. Variables and Expressions B. Order of Operations C. Open Sentences D. Identity and Equality Properties E. The Distributive Property F. Commutative
More informationRobust Computation of Distance Between Line Segments
Robust Computation of Distance Between Line Segments David Eberly, Geometric Tools, Redmond WA 98052 https://www.geometrictools.com/ This work is licensed under the Creative Commons Attribution 4.0 International
More informationProperties of Matrices and Operations on Matrices
Properties of Matrices and Operations on Matrices A common data structure for statistical analysis is a rectangular array or matris. Rows represent individual observational units, or just observations,
More informationMA2501 Numerical Methods Spring 2015
Norwegian University of Science and Technology Department of Mathematics MA2501 Numerical Methods Spring 2015 Solutions to exercise set 3 1 Attempt to verify experimentally the calculation from class that
More informationSCORE BOOSTER JAMB PREPARATION SERIES II
BOOST YOUR JAMB SCORE WITH PAST Polynomials QUESTIONS Part II ALGEBRA by H. O. Aliu J. K. Adewole, PhD (Editor) 1) If 9x 2 + 6xy + 4y 2 is a factor of 27x 3 8y 3, find the other factor. (UTME 2014) 3x
More informationChapter 2 Formulas and Definitions:
Chapter 2 Formulas and Definitions: (from 2.1) Definition of Polynomial Function: Let n be a nonnegative integer and let a n,a n 1,...,a 2,a 1,a 0 be real numbers with a n 0. The function given by f (x)
More informationNumerical Methods in Matrix Computations
Ake Bjorck Numerical Methods in Matrix Computations Springer Contents 1 Direct Methods for Linear Systems 1 1.1 Elements of Matrix Theory 1 1.1.1 Matrix Algebra 2 1.1.2 Vector Spaces 6 1.1.3 Submatrices
More informationB553 Lecture 5: Matrix Algebra Review
B553 Lecture 5: Matrix Algebra Review Kris Hauser January 19, 2012 We have seen in prior lectures how vectors represent points in R n and gradients of functions. Matrices represent linear transformations
More informationIntroduction to Applied Linear Algebra with MATLAB
Sigam Series in Applied Mathematics Volume 7 Rizwan Butt Introduction to Applied Linear Algebra with MATLAB Heldermann Verlag Contents Number Systems and Errors 1 1.1 Introduction 1 1.2 Number Representation
More information. =. a i1 x 1 + a i2 x 2 + a in x n = b i. a 11 a 12 a 1n a 21 a 22 a 1n. i1 a i2 a in
Vectors and Matrices Continued Remember that our goal is to write a system of algebraic equations as a matrix equation. Suppose we have the n linear algebraic equations a x + a 2 x 2 + a n x n = b a 2
More information1 Number Systems and Errors 1
Contents 1 Number Systems and Errors 1 1.1 Introduction................................ 1 1.2 Number Representation and Base of Numbers............. 1 1.2.1 Normalized Floating-point Representation...........
More informationMEMORIAL UNIVERSITY OF NEWFOUNDLAND
MEMORIAL UNIVERSITY OF NEWFOUNDLAND DEPARTMENT OF MATHEMATICS AND STATISTICS Section 5. Math 090 Fall 009 SOLUTIONS. a) Using long division of polynomials, we have x + x x x + ) x 4 4x + x + 0x x 4 6x
More informationSection 8.3 Partial Fraction Decomposition
Section 8.6 Lecture Notes Page 1 of 10 Section 8.3 Partial Fraction Decomposition Partial fraction decomposition involves decomposing a rational function, or reversing the process of combining two or more
More informationThis ensures that we walk downhill. For fixed λ not even this may be the case.
Gradient Descent Objective Function Some differentiable function f : R n R. Gradient Descent Start with some x 0, i = 0 and learning rate λ repeat x i+1 = x i λ f(x i ) until f(x i+1 ) ɛ Line Search Variant
More informationANALYTICAL MATHEMATICS FOR APPLICATIONS 2018 LECTURE NOTES 3
ANALYTICAL MATHEMATICS FOR APPLICATIONS 2018 LECTURE NOTES 3 ISSUED 24 FEBRUARY 2018 1 Gaussian elimination Let A be an (m n)-matrix Consider the following row operations on A (1) Swap the positions any
More informationIntroduction to Arithmetic Geometry Fall 2013 Lecture #2 09/10/2013
18.78 Introduction to Arithmetic Geometry Fall 013 Lecture # 09/10/013.1 Plane conics A conic is a plane projective curve of degree. Such a curve has the form C/k : ax + by + cz + dxy + exz + fyz with
More informationNumerical Methods. Elena loli Piccolomini. Civil Engeneering. piccolom. Metodi Numerici M p. 1/??
Metodi Numerici M p. 1/?? Numerical Methods Elena loli Piccolomini Civil Engeneering http://www.dm.unibo.it/ piccolom elena.loli@unibo.it Metodi Numerici M p. 2/?? Least Squares Data Fitting Measurement
More informationNext topics: Solving systems of linear equations
Next topics: Solving systems of linear equations 1 Gaussian elimination (today) 2 Gaussian elimination with partial pivoting (Week 9) 3 The method of LU-decomposition (Week 10) 4 Iterative techniques:
More informationConverting Between Coordinate Systems
Converting Between Coordinate Systems David Eberly, Geometric Tools, Redmond WA 98052 https://www.geometrictools.com/ This work is licensed under the Creative Commons Attribution 4.0 International License.
More informationImportant Matrix Factorizations
LU Factorization Choleski Factorization The QR Factorization LU Factorization: Gaussian Elimination Matrices Gaussian elimination transforms vectors of the form a α, b where a R k, 0 α R, and b R n k 1,
More informationIntroduction to Matrices
POLS 704 Introduction to Matrices Introduction to Matrices. The Cast of Characters A matrix is a rectangular array (i.e., a table) of numbers. For example, 2 3 X 4 5 6 (4 3) 7 8 9 0 0 0 Thismatrix,with4rowsand3columns,isoforder
More informationMatrix notation. A nm : n m : size of the matrix. m : no of columns, n: no of rows. Row matrix n=1 [b 1, b 2, b 3,. b m ] Column matrix m=1
Matrix notation A nm : n m : size of the matrix m : no of columns, n: no of rows Row matrix n=1 [b 1, b 2, b 3,. b m ] Column matrix m=1 n = m square matrix Symmetric matrix Upper triangular matrix: matrix
More informationToday s class. Linear Algebraic Equations LU Decomposition. Numerical Methods, Fall 2011 Lecture 8. Prof. Jinbo Bi CSE, UConn
Today s class Linear Algebraic Equations LU Decomposition 1 Linear Algebraic Equations Gaussian Elimination works well for solving linear systems of the form: AX = B What if you have to solve the linear
More informationLECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel
LECTURE NOTES on ELEMENTARY NUMERICAL METHODS Eusebius Doedel TABLE OF CONTENTS Vector and Matrix Norms 1 Banach Lemma 20 The Numerical Solution of Linear Systems 25 Gauss Elimination 25 Operation Count
More informationLeast Squares Fitting of Data
Least Squares Fitting of Data David Eberly, Geoetric Tools, Redond WA 98052 https://www.geoetrictools.co/ This work is licensed under the Creative Coons Attribution 4.0 International License. To view a
More informationMAA507, Power method, QR-method and sparse matrix representation.
,, and representation. February 11, 2014 Lecture 7: Overview, Today we will look at:.. If time: A look at representation and fill in. Why do we need numerical s? I think everyone have seen how time consuming
More informationLU Factorization. LU factorization is the most common way of solving linear systems! Ax = b LUx = b
AM 205: lecture 7 Last time: LU factorization Today s lecture: Cholesky factorization, timing, QR factorization Reminder: assignment 1 due at 5 PM on Friday September 22 LU Factorization LU factorization
More informationApril 30, Name: Amy s Solutions. Discussion Section: N/A. Discussion TA: N/A
Math 1151, April 30, 010 Exam 3 (in-class) Name: Amy s Solutions Discussion Section: N/A Discussion TA: N/A This exam has 8 multiple-choice problems, each worth 5 points. When you have decided on a correct
More informationThis can be accomplished by left matrix multiplication as follows: I
1 Numerical Linear Algebra 11 The LU Factorization Recall from linear algebra that Gaussian elimination is a method for solving linear systems of the form Ax = b, where A R m n and bran(a) In this method
More informationMatrix Factorization and Analysis
Chapter 7 Matrix Factorization and Analysis Matrix factorizations are an important part of the practice and analysis of signal processing. They are at the heart of many signal-processing algorithms. Their
More informationEigenvalues and Eigenvectors: An Introduction
Eigenvalues and Eigenvectors: An Introduction The eigenvalue problem is a problem of considerable theoretical interest and wide-ranging application. For example, this problem is crucial in solving systems
More informationNumerical Analysis: Solving Systems of Linear Equations
Numerical Analysis: Solving Systems of Linear Equations Mirko Navara http://cmpfelkcvutcz/ navara/ Center for Machine Perception, Department of Cybernetics, FEE, CTU Karlovo náměstí, building G, office
More information2.1 Gaussian Elimination
2. Gaussian Elimination A common problem encountered in numerical models is the one in which there are n equations and n unknowns. The following is a description of the Gaussian elimination method for
More information(17) (18)
Module 4 : Solving Linear Algebraic Equations Section 3 : Direct Solution Techniques 3 Direct Solution Techniques Methods for solving linear algebraic equations can be categorized as direct and iterative
More informationIndex. book 2009/5/27 page 121. (Page numbers set in bold type indicate the definition of an entry.)
page 121 Index (Page numbers set in bold type indicate the definition of an entry.) A absolute error...26 componentwise...31 in subtraction...27 normwise...31 angle in least squares problem...98,99 approximation
More informationMATH 221, Spring Homework 10 Solutions
MATH 22, Spring 28 - Homework Solutions Due Tuesday, May Section 52 Page 279, Problem 2: 4 λ A λi = and the characteristic polynomial is det(a λi) = ( 4 λ)( λ) ( )(6) = λ 6 λ 2 +λ+2 The solutions to the
More informationCSL361 Problem set 4: Basic linear algebra
CSL361 Problem set 4: Basic linear algebra February 21, 2017 [Note:] If the numerical matrix computations turn out to be tedious, you may use the function rref in Matlab. 1 Row-reduced echelon matrices
More informationIterative methods for Linear System of Equations. Joint Advanced Student School (JASS-2009)
Iterative methods for Linear System of Equations Joint Advanced Student School (JASS-2009) Course #2: Numerical Simulation - from Models to Software Introduction In numerical simulation, Partial Differential
More information30.3. LU Decomposition. Introduction. Prerequisites. Learning Outcomes
LU Decomposition 30.3 Introduction In this Section we consider another direct method for obtaining the solution of systems of equations in the form AX B. Prerequisites Before starting this Section you
More informationMatrix Factorization Reading: Lay 2.5
Matrix Factorization Reading: Lay 2.5 October, 20 You have seen that if we know the inverse A of a matrix A, we can easily solve the equation Ax = b. Solving a large number of equations Ax = b, Ax 2 =
More informationMatrix Basic Concepts
Matrix Basic Concepts Topics: What is a matrix? Matrix terminology Elements or entries Diagonal entries Address/location of entries Rows and columns Size of a matrix A column matrix; vectors Special types
More informationMaths for Signals and Systems Linear Algebra for Engineering Applications
Maths for Signals and Systems Linear Algebra for Engineering Applications Lectures 1-2, Tuesday 11 th October 2016 DR TANIA STATHAKI READER (ASSOCIATE PROFFESOR) IN SIGNAL PROCESSING IMPERIAL COLLEGE LONDON
More informationFitting Linear Statistical Models to Data by Least Squares I: Introduction
Fitting Linear Statistical Models to Data by Least Squares I: Introduction Brian R. Hunt and C. David Levermore University of Maryland, College Park Math 420: Mathematical Modeling February 5, 2014 version
More informationL2-7 Some very stylish matrix decompositions for solving Ax = b 10 Oct 2015
L-7 Some very stylish matrix decompositions for solving Ax = b 10 Oct 015 Marty McFly: Wait a minute, Doc. Ah... Are you telling me you built a time machine... out of a DeLorean? Doc Brown: The way I see
More informationSimultaneous Linear Equations
Simultaneous Linear Equations A farmer had 196 cows. When he rounded them up, he had 200 cows. (Reader s Digest) 1 The name of the person in the picture is A. A$AP Rocky B. Kid Cudi C. MC Hammer D. T.I.
More informationDirect Methods for Solving Linear Systems. Simon Fraser University Surrey Campus MACM 316 Spring 2005 Instructor: Ha Le
Direct Methods for Solving Linear Systems Simon Fraser University Surrey Campus MACM 316 Spring 2005 Instructor: Ha Le 1 Overview General Linear Systems Gaussian Elimination Triangular Systems The LU Factorization
More informationApplied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic
Applied Mathematics 205 Unit V: Eigenvalue Problems Lecturer: Dr. David Knezevic Unit V: Eigenvalue Problems Chapter V.4: Krylov Subspace Methods 2 / 51 Krylov Subspace Methods In this chapter we give
More informationApplied Linear Algebra in Geoscience Using MATLAB
Applied Linear Algebra in Geoscience Using MATLAB Contents Getting Started Creating Arrays Mathematical Operations with Arrays Using Script Files and Managing Data Two-Dimensional Plots Programming in
More information(a) II and III (b) I (c) I and III (d) I and II and III (e) None are true.
1 Which of the following statements is always true? I The null space of an m n matrix is a subspace of R m II If the set B = {v 1,, v n } spans a vector space V and dimv = n, then B is a basis for V III
More informationProcess Model Formulation and Solution, 3E4
Process Model Formulation and Solution, 3E4 Section B: Linear Algebraic Equations Instructor: Kevin Dunn dunnkg@mcmasterca Department of Chemical Engineering Course notes: Dr Benoît Chachuat 06 October
More informationApproximation Algorithms
Approximation Algorithms Chapter 26 Semidefinite Programming Zacharias Pitouras 1 Introduction LP place a good lower bound on OPT for NP-hard problems Are there other ways of doing this? Vector programs
More information1 Multiply Eq. E i by λ 0: (λe i ) (E i ) 2 Multiply Eq. E j by λ and add to Eq. E i : (E i + λe j ) (E i )
Direct Methods for Linear Systems Chapter Direct Methods for Solving Linear Systems Per-Olof Persson persson@berkeleyedu Department of Mathematics University of California, Berkeley Math 18A Numerical
More informationChapter 2. Solving Systems of Equations. 2.1 Gaussian elimination
Chapter 2 Solving Systems of Equations A large number of real life applications which are resolved through mathematical modeling will end up taking the form of the following very simple looking matrix
More information