AMS526: Numerical Analysis I (Numerical Linear Algebra)

Similar documents
AMS526: Numerical Analysis I (Numerical Linear Algebra)

EECS 275 Matrix Computation

AMS526: Numerical Analysis I (Numerical Linear Algebra)

AMS526: Numerical Analysis I (Numerical Linear Algebra)

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Lecture 9. Errors in solving Linear Systems. J. Chaudhry (Zeb) Department of Mathematics and Statistics University of New Mexico

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for At a high level, there are two pieces to solving a least squares problem:

AMS526: Numerical Analysis I (Numerical Linear Algebra)

AMS526: Numerical Analysis I (Numerical Linear Algebra)

AMS526: Numerical Analysis I (Numerical Linear Algebra)

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

Applied Numerical Linear Algebra. Lecture 8

Outline. Math Numerical Analysis. Errors. Lecture Notes Linear Algebra: Part B. Joseph M. Mahaffy,

Least-Squares Systems and The QR factorization

Numerical Methods. Elena loli Piccolomini. Civil Engeneering. piccolom. Metodi Numerici M p. 1/??

NUMERICAL LINEAR ALGEBRA. Lecture notes for MA 660A/B. Rudi Weikard

The QR Factorization

MATH 612 Computational methods for equation solving and function minimization Week # 6

Since the determinant of a diagonal matrix is the product of its diagonal elements it is trivial to see that det(a) = α 2. = max. A 1 x.

Scientific Computing: Dense Linear Systems

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

Applied Mathematics 205. Unit II: Numerical Linear Algebra. Lecturer: Dr. David Knezevic

Lecture notes: Applied linear algebra Part 1. Version 2

AMS526: Numerical Analysis I (Numerical Linear Algebra)

ECS130 Scientific Computing. Lecture 1: Introduction. Monday, January 7, 10:00 10:50 am

MIDTERM. b) [2 points] Compute the LU Decomposition A = LU or explain why one does not exist.

Lecture 3: Linear Algebra Review, Part II

Numerical Methods I Solving Square Linear Systems: GEM and LU factorization

Linear Algebra Massoud Malek

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

Least squares: the big idea

Computational Methods. Least Squares Approximation/Optimization

Problem 1. CS205 Homework #2 Solutions. Solution

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Math Fall Final Exam

5. Orthogonal matrices

Lecture 7. Gaussian Elimination with Pivoting. David Semeraro. University of Illinois at Urbana-Champaign. February 11, 2014

Lecture 3: QR-Factorization

Math 407: Linear Optimization

Dot product and linear least squares problems

Numerical Linear Algebra Notes

Lecture 13: Orthogonal projections and least squares (Section ) Thang Huynh, UC San Diego 2/9/2018

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Q T A = R ))) ) = A n 1 R

Lecture: Linear algebra. 4. Solutions of linear equation systems The fundamental theorem of linear algebra

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

AM 205: lecture 8. Last time: Cholesky factorization, QR factorization Today: how to compute the QR factorization, the Singular Value Decomposition

ANALYTICAL MATHEMATICS FOR APPLICATIONS 2018 LECTURE NOTES 3

Lecture 9: Numerical Linear Algebra Primer (February 11st)

WHEN MODIFIED GRAM-SCHMIDT GENERATES A WELL-CONDITIONED SET OF VECTORS

Orthogonalization and least squares methods

Math Camp II. Basic Linear Algebra. Yiqing Xu. Aug 26, 2014 MIT

AMS526: Numerical Analysis I (Numerical Linear Algebra)

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2018 LECTURE 9

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)

Math 3191 Applied Linear Algebra

DS-GA 1002 Lecture notes 10 November 23, Linear models

Scientific Computing: Solving Linear Systems

Lecture 7. Econ August 18

MATH 3511 Lecture 1. Solving Linear Systems 1

5 Solving Systems of Linear Equations

EECS 275 Matrix Computation

7. Dimension and Structure.

YORK UNIVERSITY. Faculty of Science Department of Mathematics and Statistics MATH M Test #2 Solutions

Lecture 6, Sci. Comp. for DPhil Students

Computational Linear Algebra

AM205: Assignment 2. i=1

Eigenvalue and Eigenvector Problems

Matrix decompositions

Review Questions REVIEW QUESTIONS 71

Cheat Sheet for MATH461

arxiv: v2 [math.na] 27 Dec 2016

Numerical Linear Algebra Primer. Ryan Tibshirani Convex Optimization /36-725

Orthogonality. 6.1 Orthogonal Vectors and Subspaces. Chapter 6

Linear Analysis Lecture 16

Solving Linear Systems of Equations

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Jim Lambers MAT 610 Summer Session Lecture 2 Notes

Orthogonal Transformations

Numerical Linear Algebra Primer. Ryan Tibshirani Convex Optimization

Scientific Computing

Numerical Methods I Non-Square and Sparse Linear Systems

Worksheet for Lecture 23 (due December 4) Section 6.1 Inner product, length, and orthogonality

Math Linear Algebra Final Exam Review Sheet

A fast randomized algorithm for overdetermined linear least-squares regression

Gaussian Elimination for Linear Systems

Numerical Methods for Solving Large Scale Eigenvalue Problems

11 Adjoint and Self-adjoint Matrices

14.2 QR Factorization with Column Pivoting

STA141C: Big Data & High Performance Statistical Computing

AMS526: Numerical Analysis I (Numerical Linear Algebra)

MATH 304 Linear Algebra Lecture 18: Orthogonal projection (continued). Least squares problems. Normed vector spaces.

lecture 2 and 3: algorithms for linear algebra

CSL361 Problem set 4: Basic linear algebra

AM 205: lecture 6. Last time: finished the data fitting topic Today s lecture: numerical linear algebra, LU factorization

Chapter 3 Transformations

1. Let m 1 and n 1 be two natural numbers such that m > n. Which of the following is/are true?

Lecture 6: Geometry of OLS Estimation of Linear Regession

Maths for Signals and Systems Linear Algebra in Engineering

Conceptual Questions for Review

Transcription:

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 13: Conditioning of Least Squares Problems; Stability of Householder Triangularization Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical Analysis I 1 / 16

Outline 1 Conditioning of Least Squares Problems (NLA 18) 2 Stability of Householder Triangularization (NLA 16,19) Xiangmin Jiao Numerical Analysis I 2 / 16

Four Conditioning Problems Least squares problem: Given A R m n with full rank and b R m, min x R n b Ax Its solution is x = A + b. Another quantity is y = Ax = Pb, where P = AA + Consider A and b as input data, and x and y as output. We then have four conditioning problems: Input \ Output y x b κ b y κ b x A κ A y κ A x These conditioning problems are important and subtle. Xiangmin Jiao Numerical Analysis I 3 / 16

Some Prerequisites We focus on the second column, namely κ b x and κ A x However, understanding κ b y and κ A y is prerequisite Three quantities: (All in 2-norms) Condition number of A: κ(a) = A A + = σ 1 /σ n Angle between b and y: θ = arccos y b. (0 θ π/2) Orientation of y with range(a): η = A x y. (1 η κ(a)) Xiangmin Jiao Numerical Analysis I 4 / 16

Sensitivity of y to Perturbations in b Intuition: The larger θ is, the more sensitive y is in terms of relative error Analysis: y = Pb, so where P = 1 κ b y = P y / b = b y = 1 cosθ, Input \ Output y x 1 b cosθ A Question: When is the maximum attained for perturbation δb? Xiangmin Jiao Numerical Analysis I 5 / 16

Sensitivity of y to Perturbations in b Intuition: The larger θ is, the more sensitive y is in terms of relative error Analysis: y = Pb, so where P = 1 κ b y = P y / b = b y = 1 cosθ, Input \ Output y x 1 b cosθ A Question: When is the maximum attained for perturbation δb? Answer: When is δb in range(a) Xiangmin Jiao Numerical Analysis I 5 / 16

Sensitivity of x to Perturbations in b Intuition: It depends on how sensitive y is to b, and how y lies within range(a) Analysis: x = A + b, so κ b x = A+ x / b = A+ b y y x = A+ 1 A cosθ η = κ(a) η cosθ, where η = A x / y Input \ Output y x b 1 cosθ A κ(a) η cosθ Xiangmin Jiao Numerical Analysis I 6 / 16

Sensitivity of x to Perturbations in b Assume cosθ = O(1), κ b x = κ(a) O(κ(A))! η cosθ can lie anywhere between 1 and Question: When is the maximum attained for perturbation δb? Xiangmin Jiao Numerical Analysis I 7 / 16

Sensitivity of x to Perturbations in b Assume cosθ = O(1), κ b x = κ(a) O(κ(A))! η cosθ can lie anywhere between 1 and Question: When is the maximum attained for perturbation δb? Answer: When is δb in subspace spanned by left singular vectors corresponding to smallest singular values Question: What if A is a nonsingular matrix? Xiangmin Jiao Numerical Analysis I 7 / 16

Sensitivity of x to Perturbations in b Assume cosθ = O(1), κ b x = κ(a) O(κ(A))! η cosθ can lie anywhere between 1 and Question: When is the maximum attained for perturbation δb? Answer: When is δb in subspace spanned by left singular vectors corresponding to smallest singular values Question: What if A is a nonsingular matrix? Answer: κ b x can lie anywhere between 1 and κ(a)! Xiangmin Jiao Numerical Analysis I 7 / 16

Sensitivity of y and x to Perturbations in A The relationship are nonlinear, because range(a) changes due to δa Intuitions: The larger θ is, the more sensitive y is in terms of relative error. Tilting of range(a) depends on κ(a). For x, it depends where y lies within range(a) Input \ Output y x 1 b cosθ A κ(a) cosθ κ(a) η cosθ κ(a)+ κ(a)2 tanθ η For second row, bounds are not necessarily tight Assume cosθ = O(1), κ A x can lie anywhere between κ(a) and O(κ(A) 2 ) Xiangmin Jiao Numerical Analysis I 8 / 16

Condition Numbers of Linear Systems Linear system Ax = b for nonsingular A R m m is a special case of least squares problems, where y = b If m = n, then θ = 0, so cosθ = 1 and tanθ = 0. Input \ Output y x b 1 κ(a)/η A κ(a) κ(a) Xiangmin Jiao Numerical Analysis I 9 / 16

Outline 1 Conditioning of Least Squares Problems (NLA 18) 2 Stability of Householder Triangularization (NLA 16,19) Xiangmin Jiao Numerical Analysis I 10 / 16

Solution of Least Squares Problems An efficient and robust approach is to use QR factorization A = ˆQˆR b can be projected onto range(a) by P = ˆQˆQT, and therefore ˆQˆRx = ˆQˆQ T b Left-multiply by ˆQ T and we get ˆRx = ˆQ T b (note A + = ˆR 1ˆQ T ) Least squares via QR Factorization Compute reduced QR factorization A = ˆQˆR Compute vector c = ˆQ T b Solve upper-triangular system ˆRx = c for x Computation is dominated by QR factorization (2mn 2 2 3 n3 ) What about stability? Xiangmin Jiao Numerical Analysis I 11 / 16

Backward Stability of Householder Triangularization For a QR factorization A = QR computed by Householder triangularization, the factors Q and R satisfy Q R = A+δA, δa / A = O(ǫ machine ), i.e., exact QR factorization of a slightly perturbed A R is R computed by algorithm using floating points However, Q is product of exactly orthogonal reflectors Q = Q 1 Q2... Q n where Q k is given by computed ṽ k, since Q is not formed explicitly Xiangmin Jiao Numerical Analysis I 12 / 16

Backward Stability of Solving Ax = b with QR Algorithm: Solving Ax = b by QR Factorization Compute A = QR using Householder, represent Q by reflectors Compute vector y = Q T b implicitly using reflectors Solve upper-triangular system Rx = y for x All three steps are backward stable Overall, we can show that (A+ A) x = b, A / A = O(ǫ machine ) as we prove next Xiangmin Jiao Numerical Analysis I 13 / 16

Backward Stability of Solving Ax = b with Householder Triangularization Proof: Step 2 gives Step 3 gives Therefore, ( Q +δq)ỹ = b, δq = O(ǫ machine ) ( R +δr) x = ỹ, δr / R = O(ǫ machine ) b = ( Q +δq)( R +δr) x = [ Q R +(δq) R + Q(δR)+(δQ)(δR) ] x Step 1 gives where Q R = A+δA b = A+δA+(δQ) R + Q(δR)+(δQ)(δR) x }{{} A Xiangmin Jiao Numerical Analysis I 14 / 16

Proof of Backward Stability Cont d Q R = A+δA where δa / A = O(ǫ machine ), and therefore R A Q T A+δA = O(1) A Now show that each term in A is small Overall, (δq) R A Q(δR) A (δq)(δr) A (δq) R A = O(ǫ machine ) Q δr R R A = O(ǫ machine ) δq δr A = O(ǫ2 machine ) A A δa A + (δq) R + Q(δR) + (δq)(δr) A A A Since the algorithm is backward stable, it is also accurate. = O(ǫ machine ) Xiangmin Jiao Numerical Analysis I 15 / 16

Backward Stability of Householder Triangularization Theorem Let the full-rank least squares problem be solved using Householder triangularization on a computer satisfying the two axioms of floating point numbers. The algorithm is backward stable in the sense that the computed solution x has the property (A+δA) x b) = min, δa A = O(ǫ machine ) for some δa R m n. Backward stability of the algorithm is true whether ˆQ T b is computed via explicit formation of ˆQ or computed implicitly Backward stability also holds for Householder triangularization with arbitrary column pivoting AP = ˆQˆR Xiangmin Jiao Numerical Analysis I 16 / 16