STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2018 LECTURE 17. a ij x (k) b i. a ij x (k+1) (D + L)x (k+1) = b Ux (k)

Similar documents
CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 13

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2018 LECTURE 16

Vector Norms. Chapter 7 Iterative Techniques in Matrix Algebra. Cauchy-Bunyakovsky-Schwarz Inequality for Sums. Distances. Convergence.

Chapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems

Relaxation Methods for Iterative Solution to Linear Systems of Equations

Yong Joon Ryang. 1. Introduction Consider the multicommodity transportation problem with convex quadratic cost function. 1 2 (x x0 ) T Q(x x 0 )

MATH Homework #2

A New Refinement of Jacobi Method for Solution of Linear System Equations AX=b

Lecture 21: Numerical methods for pricing American type derivatives

Hongyi Miao, College of Science, Nanjing Forestry University, Nanjing ,China. (Received 20 June 2013, accepted 11 March 2014) I)ϕ (k)

Errors for Linear Systems

1 GSW Iterative Techniques for y = Ax

MMA and GCMMA two methods for nonlinear optimization

Salmon: Lectures on partial differential equations. Consider the general linear, second-order PDE in the form. ,x 2

Eigenvalues of Random Graphs

Feature Selection: Part 1

Some Comments on Accelerating Convergence of Iterative Sequences Using Direct Inversion of the Iterative Subspace (DIIS)

Some modelling aspects for the Matlab implementation of MMA

Solutions to exam in SF1811 Optimization, Jan 14, 2015

Lecture Notes on Linear Regression

APPENDIX A Some Linear Algebra

Deriving the X-Z Identity from Auxiliary Space Method

Dynamic Systems on Graphs

Inexact Newton Methods for Inverse Eigenvalue Problems

for Linear Systems With Strictly Diagonally Dominant Matrix

Ph 219a/CS 219a. Exercises Due: Wednesday 23 October 2013

4DVAR, according to the name, is a four-dimensional variational method.

Norms, Condition Numbers, Eigenvalues and Eigenvectors

First day August 1, Problems and Solutions

6.854J / J Advanced Algorithms Fall 2008

Lecture 12: Discrete Laplacian

VQ widely used in coding speech, image, and video

Perron Vectors of an Irreducible Nonnegative Interval Matrix

Solution of Linear System of Equations and Matrix Inversion Gauss Seidel Iteration Method

Appendix B. The Finite Difference Scheme

Lectures - Week 4 Matrix norms, Conditioning, Vector Spaces, Linear Independence, Spanning sets and Basis, Null space and Range of a Matrix

U.C. Berkeley CS294: Beyond Worst-Case Analysis Luca Trevisan September 5, 2017

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur

By Samuel Schechter. (1) H = i-î -1 with 1 ^ i,j g n. [a, b,j

Singular Value Decomposition: Theory and Applications

Ph 219a/CS 219a. Exercises Due: Wednesday 12 November 2008

NUMERICAL DIFFERENTIATION

On a Parallel Implementation of the One-Sided Block Jacobi SVD Algorithm

SIO 224. m(r) =(ρ(r),k s (r),µ(r))

10-801: Advanced Optimization and Randomized Methods Lecture 2: Convex functions (Jan 15, 2014)

Edge Isoperimetric Inequalities

Estimation: Part 2. Chapter GREG estimation

Appendix B. Criterion of Riemann-Stieltjes Integrability

Week 5: Neural Networks

Chapter 7 Generalized and Weighted Least Squares Estimation. In this method, the deviation between the observed and expected values of

Topic 5: Non-Linear Regression

IV. Performance Optimization

Lecture 3. Ax x i a i. i i

Solutions to Problem Set 6

Graph Reconstruction by Permutations

Statistical pattern recognition

BOUNDEDNESS OF THE RIESZ TRANSFORM WITH MATRIX A 2 WEIGHTS

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X

Lecture 6/7 (February 10/12, 2014) DIRAC EQUATION. The non-relativistic Schrödinger equation was obtained by noting that the Hamiltonian 2

Physics 5153 Classical Mechanics. Principle of Virtual Work-1

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 12 10/21/2013. Martingale Concentration Inequalities and Applications

Poisson brackets and canonical transformations

( ) 2 ( ) ( ) Problem Set 4 Suggested Solutions. Problem 1

Generalized Linear Methods

C4B Machine Learning Answers II. = σ(z) (1 σ(z)) 1 1 e z. e z = σ(1 σ) (1 + e z )

1 Convex Optimization

form, and they present results of tests comparng the new algorthms wth other methods. Recently, Olschowka & Neumaer [7] ntroduced another dea for choo

8.4 COMPLEX VECTOR SPACES AND INNER PRODUCTS

Which Separator? Spring 1

Non-negative Matrices and Distributed Control

n α j x j = 0 j=1 has a nontrivial solution. Here A is the n k matrix whose jth column is the vector for all t j=0

Common loop optimizations. Example to improve locality. Why Dependence Analysis. Data Dependence in Loops. Goal is to find best schedule:

Chapter 12. Ordinary Differential Equation Boundary Value (BV) Problems

2.3 Nilpotent endomorphisms

APPROXIMATE PRICES OF BASKET AND ASIAN OPTIONS DUPONT OLIVIER. Premia 14

Discretization. Consistency. Exact Solution Convergence x, t --> 0. Figure 5.1: Relation between consistency, stability, and convergence.

A Parallel Multisplitting Solution of the Least Squares Problem

DISCRIMINANTS AND RAMIFIED PRIMES. 1. Introduction A prime number p is said to be ramified in a number field K if the prime ideal factorization

,, MRTS is the marginal rate of technical substitution

Limited Dependent Variables

Iterative Solution methods

However, since P is a symmetric idempotent matrix, of P are either 0 or 1 [Eigen-values

Overlapping additive and multiplicative Schwarz iterations for H -matrices

Complete subgraphs in multipartite graphs

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity

ON THE GENERAL ALGEBRAIC INVERSE EIGENVALUE PROBLEMS. 1. Introduction

Stanford University CS359G: Graph Partitioning and Expanders Handout 4 Luca Trevisan January 13, 2011

EEE 241: Linear Systems

Lecture 20: Lift and Project, SDP Duality. Today we will study the Lift and Project method. Then we will prove the SDP duality theorem.

ρ some λ THE INVERSE POWER METHOD (or INVERSE ITERATION) , for , or (more usually) to

Explicit constructions of all separable two-qubits density matrices and related problems for three-qubits systems

Problem Do any of the following determine homomorphisms from GL n (C) to GL n (C)?

The Order Relation and Trace Inequalities for. Hermitian Operators

Global Optimization of Truss. Structure Design INFORMS J. N. Hooker. Tallys Yunes. Slide 1

CSci 6974 and ECSE 6966 Math. Tech. for Vision, Graphics and Robotics Lecture 21, April 17, 2006 Estimating A Plane Homography

1 Matrix representations of canonical matrices

Module 9. Lecture 6. Duality in Assignment Problems

MATH Sensitivity of Eigenvalue Problems

e - c o m p a n i o n

PHYS 705: Classical Mechanics. Calculus of Variations II

Transcription:

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 08 LECTURE 7. sor method remnder: n coordnatewse form, Jacob method s = [ b a x (k) a and Gauss Sedel method s = [ b a = = remnder: n matrx form, Jacob method s and Gauss Sedel s a =+ =+ D = b (L + U)x (k) (D + L) = b Ux (k) a x (k) a x (k) another general adage n numercal computatons s: don t dscard prevous nformaton, try to use t too applyng ths to Gauss Sedel, we could try to use both and x (k) for =,..., to obtan ths yelds the method of successve over relaxaton (sor) ths s gven by the teraton = ω [b a a x (k) + ( ω)x (k) (.) a = =+ the quantty ω s called the relaxaton parameter f ω =, then the sor method reduces to the Gauss Sedel method,.e., = [ b a a x (k) a = the name over relaxaton comes from choosng ω > n matrx form, the teraton can be wrtten as =+ D = ω(b L Ux (k) ) + ( ω)dx (k) whch can be rearranged to obtan or = (D + ωl) = ωb + [( ω)d ωux (k) ( ) [( ) ( ) ω D + L ω D U x (k) + ω D + L b (.) Date: December, 08, verson.. Comments, bug reports: lekheng@galton.uchcago.edu.

the teraton matrx s B ω = ( ) [( ) ω D + L ω D U snce B = B GS, f we pck some ω such that ρ(b ω ) < ρ(b GS ), we would mprove the convergence of Gauss Sedel so sor s at least as fast as Gauss Sedel and often faster n fact, for certan types of matrces, one can pck ω so that ρ(b ω ) s mnmzed for example, f A s () a nonsngular matrx, () ts Jacob teraton matrx B J has only real egenvalues, and () A may be permuted nto the form [ D B A = Π Π B D where Π, Π are permutaton matrces and D, D are dagonal matrces, then the optmal elaxaton parameter s gven by ω opt = + + ρ(b J ) and note that f Ax = b, then ρ(b ωopt ) = + ρ(b J ) + + ρ(b J ) Dx = ω(b Lx Ux) + ( ω)dx and so ( ) [( ) ( ) x = ω D + L ω D U x + ω D + L b (.3) subtractng (.3) from (.), we get note that e (k+) = B ω e (k) ( ) [( det B ω = det ω D + L det = det ( det ω D + L) ω n = n = a = ( ω) n [( ω ( ω) n n = a ω n ) ω ) D U D U therefore n = λ = ( ω) n where λ,..., λ n are the egenvalues of B ω, wth λ λ n hence we get λ n ( ω) n snce we must also have λ = ρ(b ω ) < for convergence t follows that a necessary condton for convergence of sor s 0 < ω <

f A s symmetrc postve defnte, then the condton 0 < ω < s also suffcent a result of Ostrowsk mples that for such an A, ρ(b ω ) < ff 0 < ω < suppose A R n n s a symmetrc matrx, then U = L T and f we set M ω = ω ( ) ( ) ω ω ω D + L D D + LT and defne our teraton as = x (k) M ω (Ax (k) b) ths s called the method of symmetrc successve over relaxaton (ssor) there are yet other varants of sor such as: block sor or block ssor when the matrx has a block structure; lkewse, we may also ntroduce block Jacob or block Gauss Sedel applyng the sor extrapolaton to Jacob method to get = ω [b a x (k) a = =+ a x (k) + ( ω)x (k) (.4) to preserve parallelsm; ths s often called or note the dfference between (.4) and (.) and note that when ω = n (.4), then the or method reduces to Jacob method one may even defne a nonlnear verson of sor for teratons of the form = f(x (k) ) where f s some nonlnear functon f : R n R n : sor = ( ω)x (k) sor + ωf(x (k) sor). Rchardson method unlke the splttng methods n the prevous lecture, the teratve methods here do not requre splttng A nto a sum of two matrces but they are a bt lke sor n that there s a scalar parameter nvolved at each step ths scalar parameter can ether be fxed throughout (e.g., Rchardson) or can vary from one nteraton to the next (e.g., steepest descent, Chebyshev) or there can even be two scalar parameters at each step (e.g., conugate gradent) the smplest one s known as the Rchardson method, where the teraton s smply = (I αa)x (k) + αb (.) = x (k) + α(b Ax (k) ) = x (k) + αr (k) where r (k) := b Ax (k) s the resdual at the kth step note that f x = A b, then we trvally have x = (I αa)x + αb (.) as usual we defne the error e (k) = x x (k), then subtractng (.) from (.) yelds e (k+) = B α e (k) where the teraton matrx s B α = I αa we want to choose the parameter α > 0 a pror so as to mnmze ρ(b α ) suppose A s symmetrc postve defnte, wth egenvalues µ µ µ n > 0 snce B α = I αa, we have λ = αµ for =,..., n 3

f we want α so that ρ(b α ) s mnmzed,.e., mn α max λ (α) = mn max αµ n α = mn max( αµ, αµ n ), n α the optmal parameter α s attaned when and snce these must dffer by a sgn, α µ = α µ n α µ n = ( α µ ), whch yelds α = µ + µ n note that when αµ =, the teraton dverges for some choce of x (0) hence the method converges for 0 < α < µ however ths teraton s senstve to perturbaton and therefore bad numercally for example, f µ = 0 and µ n = 0 4, then the optmal α s /(0 + 0 4 ), but ths s close to a value of α for whch the teraton dverges, α = /0 also, note that and smlarly, therefore λ (α ) = µ + µ n µ = µ n µ µ + µ n = κ(a) + κ(a) 0, λ n (α ) = µ µ n µ + µ n = µ /µ n µ /µ n + = κ(a) κ(a) + 0 ρ(b α ) = κ(a) κ(a) + and we see that the convergence rate depends on κ(a) 3. rate of convergence and spectral radus we kept referrng to the spectral radus of the teraton matrx ρ(b) as the rate of convergence but f you thnk about t, snce e (k+) B e (k), t should be B that controls the convergence rate of course f the teraton matrx B s symmetrc, then t makes no dfference snce B = ρ(b) but n general they are not equal, n fact s possble for ρ(b) = 0 and B 0 the reason s because we look at the average rate of convergence for any teraton matrx B C n n (not necessarly symmetrc) and any consstent norm (not necessarly submultplcatve), after k steps, we get the average reducton n error per step after k steps s then ( ) /k e (k) e (0) e (k) B k e (0) (3.) 4

and by (3.) s bounded by B k /k the spectral radus then drops out when we take lmts lm k Bk /k = ρ(b), whch holds for any consstent norm (.e., satsfes Ax A x for any A C n n and any x C n ) 5