Numerical Methods I: Orthogonalization and Newton s method

Similar documents
Numerical Methods I: Eigenvalues and eigenvectors

MATH 235: Inner Product Spaces, Assignment 7

Orthonormal Bases; Gram-Schmidt Process; QR-Decomposition

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES

Lecture 4 Orthonormal vectors and QR factorization

Inverses. Stephen Boyd. EE103 Stanford University. October 28, 2017

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2018 LECTURE 9

Math 407: Linear Optimization

Worksheet for Lecture 25 Section 6.4 Gram-Schmidt Process

Linear Algebra Massoud Malek

Computational Methods. Least Squares Approximation/Optimization

Lecture 6, Sci. Comp. for DPhil Students

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Chapter 7. Iterative methods for large sparse linear systems. 7.1 Sparse matrix algebra. Large sparse matrices

YORK UNIVERSITY. Faculty of Science Department of Mathematics and Statistics MATH M Test #2 Solutions

Cheat Sheet for MATH461

ELE/MCE 503 Linear Algebra Facts Fall 2018

Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008

Inner products. Theorem (basic properties): Given vectors u, v, w in an inner product space V, and a scalar k, the following properties hold:

Linear Independence. Stephen Boyd. EE103 Stanford University. October 9, 2017

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

5.) For each of the given sets of vectors, determine whether or not the set spans R 3. Give reasons for your answers.

x 3y 2z = 6 1.2) 2x 4y 3z = 8 3x + 6y + 8z = 5 x + 3y 2z + 5t = 4 1.5) 2x + 8y z + 9t = 9 3x + 5y 12z + 17t = 7

Linear Analysis Lecture 16

Applied Mathematics 205. Unit II: Numerical Linear Algebra. Lecturer: Dr. David Knezevic

Problem 1. CS205 Homework #2 Solutions. Solution

235 Final exam review questions

COURSE Iterative methods for solving linear systems

MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors.

Math 250B Midterm II Information Spring 2019 SOLUTIONS TO PRACTICE PROBLEMS

linearly indepedent eigenvectors as the multiplicity of the root, but in general there may be no more than one. For further discussion, assume matrice

ENGG5781 Matrix Analysis and Computations Lecture 8: QR Decomposition

QR Decomposition. When solving an overdetermined system by projection (or a least squares solution) often the following method is used:

(a) If A is a 3 by 4 matrix, what does this tell us about its nullspace? Solution: dim N(A) 1, since rank(a) 3. Ax =

1. Select the unique answer (choice) for each problem. Write only the answer.

AM 205: lecture 8. Last time: Cholesky factorization, QR factorization Today: how to compute the QR factorization, the Singular Value Decomposition

Section 6.4. The Gram Schmidt Process

Introduction to Linear Algebra. Tyrone L. Vincent

This can be accomplished by left matrix multiplication as follows: I

Assignment #9: Orthogonal Projections, Gram-Schmidt, and Least Squares. Name:

Nonlinear equations. Norms for R n. Convergence orders for iterative methods

I. Multiple Choice Questions (Answer any eight)

Krylov Space Methods. Nonstationary sounds good. Radu Trîmbiţaş ( Babeş-Bolyai University) Krylov Space Methods 1 / 17

Solution of Linear Equations

1. Nonlinear Equations. This lecture note excerpted parts from Michael Heath and Max Gunzburger. f(x) = 0

The QR Factorization

Chapter 5 Orthogonality

Eigenvalues and eigenvectors

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

MATH 22A: LINEAR ALGEBRA Chapter 4

Linear Algebra, part 2 Eigenvalues, eigenvectors and least squares solutions

Conceptual Questions for Review

18.06 Quiz 2 April 7, 2010 Professor Strang

1. Diagonalize the matrix A if possible, that is, find an invertible matrix P and a diagonal

Math 416, Spring 2010 Gram-Schmidt, the QR-factorization, Orthogonal Matrices March 4, 2010 GRAM-SCHMIDT, THE QR-FACTORIZATION, ORTHOGONAL MATRICES

Least-Squares Systems and The QR factorization

Stabilization and Acceleration of Algebraic Multigrid Method

MATH 167: APPLIED LINEAR ALGEBRA Least-Squares

Math 224, Fall 2007 Exam 3 Thursday, December 6, 2007

7. Symmetric Matrices and Quadratic Forms

5 Compact linear operators

Linear algebra I Homework #1 due Thursday, Oct Show that the diagonals of a square are orthogonal to one another.

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS

TBP MATH33A Review Sheet. November 24, 2018

M.A. Botchev. September 5, 2014

LINEAR ALGEBRA: NUMERICAL METHODS. Version: August 12,

Iterative Methods. Splitting Methods

MATH 235. Final ANSWERS May 5, 2015

CLASS NOTES Computational Methods for Engineering Applications I Spring 2015

Conjugate Gradient Method

Krylov Subspace Methods that Are Based on the Minimization of the Residual

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

Numerical Methods I Non-Square and Sparse Linear Systems

Numerical Methods I Eigenvalue Problems

MODULE 8 Topics: Null space, range, column space, row space and rank of a matrix

Elementary Linear Algebra Review for Exam 2 Exam is Monday, November 16th.

Problem # Max points possible Actual score Total 120

Chapter 4 Euclid Space

Linear Least squares

7. Dimension and Structure.

6. Orthogonality and Least-Squares

A nonlinear equation is any equation of the form. f(x) = 0. A nonlinear equation can have any number of solutions (finite, countable, uncountable)

Typical Problem: Compute.

Numerical Methods I: Polynomial Interpolation

Orthonormal Transformations

NUMERICALLY COMPUTING THE LYAPUNOV EXPONENTS OF MATRIX-VALUED COCYCLES

We will discuss matrix diagonalization algorithms in Numerical Recipes in the context of the eigenvalue problem in quantum mechanics, m A n = λ m

There are two things that are particularly nice about the first basis

orthogonal relations between vectors and subspaces Then we study some applications in vector spaces and linear systems, including Orthonormal Basis,

COMP 558 lecture 18 Nov. 15, 2010

Orthogonality. 6.1 Orthogonal Vectors and Subspaces. Chapter 6

SOLUTION OF ALGEBRAIC AND TRANSCENDENTAL EQUATIONS BISECTION METHOD

Matrix Algebra: Summary

Linear Algebra, part 3. Going back to least squares. Mathematical Models, Analysis and Simulation = 0. a T 1 e. a T n e. Anna-Karin Tornberg

Matrix Theory, Math6304 Lecture Notes from September 27, 2012 taken by Tasadduk Chowdhury

Problem Set (T) If A is an m n matrix, B is an n p matrix and D is a p s matrix, then show

18.06 Professor Johnson Quiz 1 October 3, 2007

Linear Algebra, part 3 QR and SVD

Designing Information Devices and Systems I Spring 2018 Lecture Notes Note 25

LECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel

Transcription:

1/42 Numerical Methods I: Orthogonalization and Newton s method Georg Stadler Courant Institute, NYU stadler@cims.nyu.edu October 5, 2017

Overview 2/42 Linear least squares and orthogonalization methods

eview: Least-squares problems 3/42 Given data points/measurements (t i,b i ), i =1,...,m and a model function that relates t and b: b = (t; x 1,...,x n ), where x 1,...,x n are model function parameters. If the model is supposed to describe the data, the deviations/errors i = b i (t i,x 1,...,x n ) should be small. Thus, to fit the model to the measurements, one must choose x 1,...,x n appropriately.

eview: Linear least-squares 4/42 We assume (for now) that the model depends linearly on x 1,...,x n, e.g.: (t; x 1,...x n )=a 1 (t)x 1 +...+ a n (t)x n Choosing the least square error, this results in min x kax bk, where x =(x 1,...,x n ) T, b =(b 1,...,b m ) T,anda ij = a j (t i ). In the following, we study the over-determined case, i.e., m n.

inear least-squares problems QR factorization 5/42 Consider non-square matrices A 2 R m n with m rank(a) = n. Thenthesystem Ax = b n and does, in general, not have a solution (more equations than unknowns). We thus instead solve a minimization problem min x kax bk 2. The minimum x of this optimization problem is characterized by the normal equations: A T A x = A T b.

inear least-squares problems QR factorization 6/42 To avoid the multiplication A T A and to use a suitable factorization of A that aids in solving the normal equation, we use the QR-factorization: A = QR = Q 1,Q 2 apple R 1 0 = Q 1 R 1, where Q 2 R m m is an orthonormal matrix (QQ T = I), and R 2 R m n consists of an upper triangular matrix and a block of zeros.

inear least-squares problems QR factorization 7/42 How can the QR factorization be used to solve the normal equation? apple b1 k min x kax bk 2 =min where Q T b = apple b1 b 2. kq T (Ax b)k 2 min x x R 1 x b 2 k 2, Thus, the least squares solution is x = R 1 b 1 and the residual is kb 2 k.

inear least-squares problems QR factorization 8/42 How can we compute the QR factorization? Givens rotations Use sequence of rotations in 2D subspaces: For m n: n 2 /2 square roots, and 4/3n 3 multiplications For m n: nm square roots, and 2mn 2 multiplications Householder reflections Use sequence of reflections in 2D subspaces For m n: 2/3n 3 multiplications For m n: 2mn 2 multiplications These methods compute an orthonormal basis of the columns of A. An alternative is the Gram Schmidt method however, Gram Schmidt is unstable and thus sensitive to rounding errors (there are modified versions that are stable but require more computation).

QR factorization: Orthogonal transformations 9/42

QR factorization: Givens rotations 10 / 42

QR factorization: Givens rotations 11 / 42

QR factorization: Householder reflections 12 / 42

QR factorization: Householder reflections 13 / 42

Overview 14 / 42 Nonlinear systems

ixed point ideas 15 / 42 We intend to solve the nonlinear equation f(x) =0, x 2 R. Reformulation as fixed point method: x = (x) Corresponding iteration: Choose x 0 (initialization) and compute x 1,x 2,... from x k+1 = (x k ) When does this iteration converge?

ixed point ideas 16 / 42 Example: Solve the nonlinear equation 2x tan(x) =0. Iteration #1: x k+1 = 1 (x k )=0.5 tan(x k ) Iteration #2: x k+1 = 2 (x k ) = arctan(2x k ) Iteration #3: x k+1 = 3 (x k )=x k 2x k tan(x k ) 1 tan 2 (x k )

onvergence of fixed point methods 17 / 42

onvergence of fixed point methods 18 / 42 Amapping :[a, b]! R is called contractive on [a, b] if there is a 0 apple < 1 such that (x) (y) apple x y for all x, y 2 [a, b]. If is continuously di erentiable on [a, b], then (x) (y) sup = sup x,y2[a,b] x y z2[a,b] 0 (z)

onvergence of fixed point methods 19 / 42 Let :[a, b]! [a, b] be contractive with constant < 1. Then: I There exists a unique fixed point x with x = ( x) I For any starting guess x 0 in [a, b], the fixed point iteration converges to x and x k+1 x k apple x k x k 1 (linear convergence) x x k apple k 1 x 1 x 0. The second expression allows to estimate the required number of iterations.

Convergence of fixed point methods 20 / 42

ewton s method 21 / 42 In one dimension, solve f(x) =0: Start with x 0, and compute x 1,x 2,... from f(x k ) x k+1 = x k f 0, k =0, 1,... (x k ) Requires f(x k ) 6= 0to be well-defined (i.e., tangent has nonzero slope).

22 / 42 ewton s method Let F : R n! R n, n 1 and solve F (x) =0. Taylor expansion about starting point x 0 : F (x) =F (x 0 )+F 0 (x 0 )(x x 0 )+o( x x 0 ) for x! x 0. Hence: x 1 = x 0 F 0 (x 0 ) 1 F (x 0 ) Newton iteration: Start with x 0 2 R n,andfork =0, 1,... compute F 0 (x k ) x k = F (x k ), x k+1 = x k + x k Requires that F 0 (x k ) 2 R n n is invertible.

ewton s method 23 / 42 Newton iteration: Start with x 0 2 R n,andfork =0, 1,... compute Equivalently: F 0 (x k ) x k = F (x k ), x k+1 = x k + x k x k+1 = x k F 0 (x k ) 1 F (x k ) Newton s method is a ne invariant, thatis,thesequenceis invariant to a ne transformations:

24 / 42 onvergence of Newton s method Assumptions on F : D R n open and convex, F : D! R n continuously di erentiable with F 0 (x) is invertible for all x, and there exists! 0 such that kf 0 (x) 1 (F 0 (x + sv) F 0 (x))vk apples!kvk 2 for all s 2 [0, 1], x 2 D, v 2 R n with x + v 2 D. Assumptions on x and x 0 : There exists a solution x 2 D and a starting point x 0 2 D such that := kx x 0 kapple 2! and B (x ) D Theorem: Then, the Newton sequence x k stays in B (x ) and lim k!1 x k = x,and kx k+1 x kapple! 2 kxk x k 2