Optimal Preconditioners for Interval Gauss Seidel Methods

Similar documents
EXISTENCE VERIFICATION FOR SINGULAR ZEROS OF REAL NONLINEAR SYSTEMS

Optimal Preconditioning for the Interval Parametric Gauss Seidel Method

Find, with certainty, approximations to all solutions of the nonlinear system

1. Introduction. We consider the general global optimization problem

Key words. numerical linear algebra, global optimization, validated computing, interval analysis

Verified Global Optimization with Taylor Model based Range Bounders

INTERVAL MATHEMATICS TECHNIQUES FOR CONTROL THEORY COMPUTATIONS

arxiv: v1 [math.na] 28 Jun 2013

Advice for Mathematically Rigorous Bounds on Optima of Linear Programs

Solving Linear Systems

Linear Algebraic Equations

Exclusion regions for systems of equations

Homework 6 Solutions

Lecture 8: Fast Linear Solvers (Part 7)

A Comparison of Some Methods for Bounding Connected and Disconnected Solution Sets of Interval Linear Systems

Solving Linear Systems

Iterative Methods. Splitting Methods

Interval Arithmetic An Elementary Introduction and Successful Applications

Solving Linear Systems of Equations

Convergence of Rump s Method for Inverting Arbitrarily Ill-Conditioned Matrices

Review of matrices. Let m, n IN. A rectangle of numbers written like A =

9.1 Preconditioned Krylov Subspace Methods

From Stationary Methods to Krylov Subspaces

COURSE Iterative methods for solving linear systems

Today s class. Linear Algebraic Equations LU Decomposition. Numerical Methods, Fall 2011 Lecture 8. Prof. Jinbo Bi CSE, UConn

Iterative Methods for Solving A x = b

Next topics: Solving systems of linear equations

Global Optimization by Interval Analysis

OUTLINE ffl CFD: elliptic pde's! Ax = b ffl Basic iterative methods ffl Krylov subspace methods ffl Preconditioning techniques: Iterative methods ILU

Improved Newton s method with exact line searches to solve quadratic matrix equation

Math 5630: Iterative Methods for Systems of Equations Hung Phan, UMass Lowell March 22, 2018

A Refinement of Gauss-Seidel Method for Solving. of Linear System of Equations

A NEW EFFECTIVE PRECONDITIONED METHOD FOR L-MATRICES

Verified Branch and Bound for Singular Linear and Nonlinear Programs An Epsilon-Inflation Process

A Method for Constructing Diagonally Dominant Preconditioners based on Jacobi Rotations

Math/Phys/Engr 428, Math 529/Phys 528 Numerical Methods - Summer Homework 3 Due: Tuesday, July 3, 2018

Lecture 11: CMSC 878R/AMSC698R. Iterative Methods An introduction. Outline. Inverse, LU decomposition, Cholesky, SVD, etc.

A NEW METHOD FOR SOLVING ILL-CONDITIONED LINEAR SYSTEMS. Fazlollah Soleymani

Solving Global Optimization Problems by Interval Computation

1 Number Systems and Errors 1

Chapter 7 Iterative Techniques in Matrix Algebra

M.SC. PHYSICS - II YEAR

RESEARCH ARTICLE. A strategy of finding an initial active set for inequality constrained quadratic programming problems

Numerical Analysis: Solutions of System of. Linear Equation. Natasha S. Sharma, PhD

Math 471 (Numerical methods) Chapter 3 (second half). System of equations

NUMERICAL COMPUTATION IN SCIENCE AND ENGINEERING

TMA4125 Matematikk 4N Spring 2017

Multigrid absolute value preconditioning

The flexible incomplete LU preconditioner for large nonsymmetric linear systems. Takatoshi Nakamura Takashi Nodera

The Conjugate Gradient Method

On the Approximation of Linear AE-Solution Sets

if : Xk, then the system of equations unique solution in

G-SPLINE INTERPOLATION FOR APPROXIMATING THE SOLUTION OF FRACTIONAL DIFFERENTIAL EQUATIONS USING LINEAR MULTI- STEP METHODS

S.F. Xu (Department of Mathematics, Peking University, Beijing)

Reduced-Hessian Methods for Constrained Optimization

APPLIED NUMERICAL LINEAR ALGEBRA

Gaussian interval quadrature rule for exponential weights

Linear Algebra Linear Algebra : Matrix decompositions Monday, February 11th Math 365 Week #4

CS760, S. Qiao Part 3 Page 1. Kernighan and Pike [5] present the following general guidelines for testing software.

Reliable Prediction of Phase Stability Using an Interval Newton Method

An Accelerated Block-Parallel Newton Method via Overlapped Partitioning

6. Iterative Methods for Linear Systems. The stepwise approach to the solution...

JACOBI S ITERATION METHOD

10.2 ITERATIVE METHODS FOR SOLVING LINEAR SYSTEMS. The Jacobi Method

On Rigorous Upper Bounds to a Global Optimum

Positive Semidefiniteness and Positive Definiteness of a Linear Parametric Interval Matrix

Numerical Methods for Chemical Engineers

Linear Equation Systems Direct Methods

Fast Verified Solutions of Sparse Linear Systems with H-matrices

Institute of Computer Science. Compact Form of the Hansen-Bliek-Rohn Enclosure

Identify the graph of a function, and obtain information from or about the graph of a function.

Formula for the inverse matrix. Cramer s rule. Review: 3 3 determinants can be computed expanding by any row or column

A new nonmonotone Newton s modification for unconstrained Optimization

Stabilization and Acceleration of Algebraic Multigrid Method

DEN: Linear algebra numerical view (GEM: Gauss elimination method for reducing a full rank matrix to upper-triangular

Block Bidiagonal Decomposition and Least Squares Problems

Introduction to Applied Linear Algebra with MATLAB

Interval Gauss-Seidel Method for Generalized Solution Sets to Interval Linear Systems

ON A HOMOTOPY BASED METHOD FOR SOLVING SYSTEMS OF LINEAR EQUATIONS

Computers and Mathematics with Applications. Convergence analysis of the preconditioned Gauss Seidel method for H-matrices

Numerical Methods in Matrix Computations

A SUFFICIENTLY EXACT INEXACT NEWTON STEP BASED ON REUSING MATRIX INFORMATION

Matrix decompositions

c 1995 Society for Industrial and Applied Mathematics Vol. 37, No. 1, pp , March

College Algebra To learn more about all our offerings Visit Knewton.com

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

LINEAR ALGEBRA: NUMERICAL METHODS. Version: August 12,

Lecture 18 Classical Iterative Methods

Miller Objectives Alignment Math

Poisson Equation in 2D

A MULTIGRID ALGORITHM FOR. Richard E. Ewing and Jian Shen. Institute for Scientic Computation. Texas A&M University. College Station, Texas SUMMARY

5.7 Cramer's Rule 1. Using Determinants to Solve Systems Assumes the system of two equations in two unknowns

SQUARE-ROOT FAMILIES FOR THE SIMULTANEOUS APPROXIMATION OF POLYNOMIAL MULTIPLE ZEROS

Introduction, basic but important concepts

PowerPoints organized by Dr. Michael R. Gustafson II, Duke University

Verified Solutions of Sparse Linear Systems by LU factorization

FINAL (CHAPTERS 7-9) MATH 141 SPRING 2018 KUNIYUKI 250 POINTS TOTAL

Hence a root lies between 1 and 2. Since f a is negative and f(x 0 ) is positive The root lies between a and x 0 i.e. 1 and 1.

Interval Methods for Solving Underdetermined Nonlinear Equations Systems

Experiments with iterative computation of search directions within interior methods for constrained optimization

Transcription:

Optimal Preconditioners for Interval Gauss Seidel Methods R B Kearfott and Xiaofa Shi Introduction Consider the following nonlinear system f 1 x 1, x 2,, x n ) F X) f n x 1, x 2,, x n ) where bounds x i and x i are known such that x i x i x i, for 1 i n, 1) We write X x 1, x 2,, x n ) T and denote by B the box given by the above inequalities on the variables x i A general approach to such problems is to transform the nonlinear system F X) to the interval linear system: F X k ) X k ˇX k ) F ˇX k ), 2) where F X k ) is a suitable interval expansion of the Jacobi matrix over the box X k X B) and ˇX k X k represents a predictor or initial guess point The general goal is to verify existence/uniqueness/non-existence of solutions in the box, and find all existing solutions 1 What is a preconditioner? Why preconditioning? Consider an interval linear system: AX ˇX) b We may multiply a matrix or a row vector Y to both sides of the above system The matrix or the row vector Y is called a preconditioner The main reason for preconditioning is to get improvement when solving the interval linear system

2 R B Kearfott and Xiaofa Shi Example 1 Consider: with initial box [ 1, 1 [ 1, 1 [ 1 [ 1, 1 1 [ x1 x 2 [ When we apply Gauss Seidel steps to the above system, we obtain x 1 [ 1, x 2 [ 1, 1, x 2 x 1 [ 1, 1 Since neither coordinate interval has changed, the Gauss Seidel method does not bring any improvement for this example On the other hand, let Y [ 2/3 1/3 2/3 2/3 If we multiply both sides of the system by Y, we have [ [ [ 1 [ 1/3, 1/3 x1 [2/3, 4/3 Now, if we apply Gauss Seidel steps, we obtain x 2 x 1 [ 1/3, 1/3 x 2 [ 1/3, 1/3, x 2 For x 1, we obtain a smaller interval For x 2, we get an even better result, a point This example tells us that preconditioners are necessary for interval Gauss Seidel steps In this paper, we will concentrate on optimal preconditioners, computed row-by-row, only A preconditioned interval Gauss Seidel method may be used to compute a new interval x k for the k-th variable Suppose Y k y k1, y k2,, y kn ) is the preconditioner for x k Algorithm 1 Preconditioned Gauss Seidel method) 1 Compute Y k F X ˇX) and Y k F Then compute n ) y ki f i + y ki f ij x j ˇx j ) x k ˇx k j1 j k y ki f ik 3)

Preconditioners 3 2 If x k x k, then return, indicating that there is no root of F in X 3 Replace x k by x k x k 2 Four different preconditioners 1 C W preconditioner A preconditioner Y k for x k is called a C W preconditioner, if w x k) min Y k w x k ), where wx) is the width of interval x, and n ) y ki f i + y ki f ij x j ˇx j ) x k ˇx k j1 j k y ki f ik d i Y i ) Then the CW preconditioner Yk the following optimization problem: min wn i Y i )) d i Y i )1 can be computed by solving 2 E W preconditioner d i Y i ) If d iy i ) and n i Y i ) >, then Kahan arithmetic gives n i Y i ) d i Y i ), n iy i ) [ ) ni Y i ) d i Y i ) d i Y i ), A preconditioner Yk for x k is called an E W preconditioner if it is a solution of the following optimization problem: max n w i Y i )1 [ ) 1 d i Y i ), 1 d i Y i ) 3 C M preconditioner

4 R B Kearfott and Xiaofa Shi A preconditioner Y k for x k is called a C M preconditioner, if x k ˇx k min Y k x k ˇx k d i Y i ) Then the CM preconditioner Yk the following optimization problem: 4 E M preconditioner min n i Y i ) d i Y i)1 d i Y i ) If d iy i ) and n i Y i ) >, then n i Y i ) d i Y i ), n i Y i) [ ) ni Y i ) d i Y i ) d i Y i ), can be computed by solving A preconditioner Y k for x k is called an E M preconditioner if it is a solution of the following optimization problem: However, d i Y i ) implies { max n min 1 } i Y i)1 d i Y i ), 1 d i Y i ) 4) { min 1 } d i Y i ), 1 d i Y i ) 1 max{ d i Y i ), d i Y i ) } 1 d i Y i ), and 1/ d i Y i ) is maximum when d i Y i ) is minimum Thus, Problem 4) can be replaced by min n d iy i ) 5) i Y i)1 For C W preconditioners, we have the following existence proposition For the other preconditioners, similar results can be obtained Proposition 1 Hu [1) There exists a C W preconditioner Y k if and only if at least one element of the k-th column of A does not contain For each type of preconditioner discussed here, a nonlinear optimization problem is invoked in its definition Fortunately, these nonlinear optimization problems can be simplified into linear programming problems Thus, to obtain a preconditioner, we only need to solve a linear programming problem Details can be found in [7

Preconditioners 5 3 Applications C W preconditioners minimize the width of x k, are appropriate for finding solutions of nonlinear systems C M preconditioners minimize the absolute value of x k ˇx k, and are appropriate for verifying existence or uniqueness of solutions E W and E M preconditioners maximize the width and the absolute value of the gap that splits x k ˇx k, respectively They are appropriate for verifying that there is no solution in the initial box Example 2 Consider: f i X) x i + x j n 1, 1 i n 1, f n X) 1 j n 1 j n x j 1, with n 5 and initial box X [, 5 [, 5 [, 5 [, 5 [, 17 The mean value extension, with a slope matrix, over the initial box is: F X) [ 6, 135 [ 6, 135 [ 6, 135 [ 6, 135 [ 1996, 625 However, there is no root in the initial box When the E M preconditioner is used, this fact is discovered right after the first preconditioner is applied, but the C W and C M preconditioners will not show that there is no root In experiments to date, we have not found cases where existence could be verified with C M preconditioners, but not C W preconditioners The reason is that the C W preconditioners and the C M preconditioners are the same if we choose the midpoint as the predictor point ˇx k This is stated formally in the following theorem Theorem 1 In 3), if we choose ˇX ˇx 1, ˇx 2,, ˇx n ) T to be the midpoint of X, then the C W preconditioners and the C M preconditioners are the same Proof: If ˇX ˇx1, ˇx 2,, ˇx n ) T is chosen to be the midpoint of X, then X will be symmetric about see [12) Thus, w x k ) x k x k Therefore, the C W and the C M preconditioners are the same

6 R B Kearfott and Xiaofa Shi References [1 C Hu Optimal Preconditioners for the Interval Newton Method PhD thesis, University of Southwestern Louisiana, 199 [2 R B Kearfott Preconditioners for the interval Gauss Seidel method SIAM J Numer Anal, 273):84 822, June 199 [3 R B Kearfott Decomposition of arithmetic expressions to improve the behavior of interval iteration for nonlinear systems Computing, 472):169 191, 1991 [4 R B Kearfott A Fortran 9 environment for research and prototyping of enclosure algorithms for nonlinear equations and global optimization ACM Trans Math Software, 211):63 78, March 1995 [5 R B Kearfott Empirical evaluation of innovations in interval branch and bound algorithms for nonlinear algebraic systems, accepted for publication in the SIAM J Sci Statist Comput [6 R B Kearfott A review of techniques in the verified solution of constrained global optimization problems, to appear in Applications of Interval Computations, Kluwer Academic Publishers [7 R B Kearfott Verified Global Optimization Interval Methods and Software Kluwer Academic Publishers, to appear [8 R B Kearfott, C Hu, and M Novoa A review of preconditioners for the interval Gauss Seidel method Interval Computations, 11):59 85, 1991 [9 R B Kearfott and M Novoa Algorithm 681: INTBIS, a portable interval Newton/bisection package ACM Trans Math Software, 162):152 157, June 199 [1 R B Kearfott and X Shi A preconditioner selection heuristic for efficient iteration with decomposition of arithmetic expressions for nonlinear systems Interval Computations, 19931):15 33, 1993 [11 R B Kearfott and Z Xing An interval step control for continuation methods SIAM J Numer Anal, 313):892 914, June 1994 [12 X Shi Intermediate Expression Preconditioning and Verification for Rigorous Solution of Nonlinear Systems PhD thesis, University of Southwestern Louisiana, Department of Mathematics, August 1995 Addresses:

Preconditioners 7 R B Kearfott, The University of Southwestern Louisiana, Department of Mathematics, Lafayette, LA 754, USA Xiaofa Shi, The University of Southwestern Louisiana, The Center for Advanced Computer Studies, Lafayette, LA 754, USA