Numerical Methods for Large-Scale Nonlinear Equations

Similar documents
Handout on Newton s Method for Systems

DELFT UNIVERSITY OF TECHNOLOGY

Newton s Method and Efficient, Robust Variants

An Accelerated Block-Parallel Newton Method via Overlapped Partitioning

THE INEXACT, INEXACT PERTURBED AND QUASI-NEWTON METHODS ARE EQUIVALENT MODELS

Maria Cameron. f(x) = 1 n

ADAPTIVE ACCURACY CONTROL OF NONLINEAR NEWTON-KRYLOV METHODS FOR MULTISCALE INTEGRATED HYDROLOGIC MODELS

Trust-Region SQP Methods with Inexact Linear System Solves for Large-Scale Optimization

Interior-Point Methods as Inexact Newton Methods. Silvia Bonettini Università di Modena e Reggio Emilia Italy

A SUFFICIENTLY EXACT INEXACT NEWTON STEP BASED ON REUSING MATRIX INFORMATION

2 CAI, KEYES AND MARCINKOWSKI proportional to the relative nonlinearity of the function; i.e., as the relative nonlinearity increases the domain of co

Research Article A Two-Step Matrix-Free Secant Method for Solving Large-Scale Systems of Nonlinear Equations

c 2007 Society for Industrial and Applied Mathematics

7th AIAA/USAF/NASA/ISSMO Symposium on Multidisciplinary Analysis and Optimization September 2 4, 1998/ St. Louis, MO

NOWADAYS, biotechnological processes are widely used. Inexact Newton matrix-free methods for solving complex biotechnological systems

Inexact Newton Methods Applied to Under Determined Systems. Joseph P. Simonis. A Dissertation. Submitted to the Faculty

REPORTS IN INFORMATICS

17 Solution of Nonlinear Systems

Steady-State Optimization Lecture 2: Solution Methods for Systems of Nonlinear Algebraic Equations

A Numerical Study of Globalizations of Newton-GMRES Methods. Joseph P. Simonis. AThesis. Submitted to the Faculty WORCESTER POLYTECHNIC INSTITUTE

An Implicit Multi-Step Diagonal Secant-Type Method for Solving Large-Scale Systems of Nonlinear Equations

A Domain Decomposition Based Jacobi-Davidson Algorithm for Quantum Dot Simulation

Multipoint secant and interpolation methods with nonmonotone line search for solving systems of nonlinear equations

Lecture 8: Fast Linear Solvers (Part 7)

Sample File for SIAM LATEX Book Macro Package, Proceedings Version

Applied Linear Algebra in Geoscience Using MATLAB

High Performance Nonlinear Solvers

Quasi-Newton Methods

Inexactness and flexibility in linear Krylov solvers

Residual iterative schemes for largescale linear systems

Key words. conjugate gradients, normwise backward error, incremental norm estimation.

Stabilization and Acceleration of Algebraic Multigrid Method

8 Numerical methods for unconstrained problems

Linear Solvers. Andrew Hazel

A NOTE ON Q-ORDER OF CONVERGENCE

The Conjugate Gradient Method

DELFT UNIVERSITY OF TECHNOLOGY

An Efficient Solver for Systems of Nonlinear. Equations with Singular Jacobian. via Diagonal Updating

XIAO-CHUAN CAI AND MAKSYMILIAN DRYJA. strongly elliptic equations discretized by the nite element methods.

MATH 3795 Lecture 13. Numerical Solution of Nonlinear Equations in R N.

Chapter 4. Unconstrained optimization

CONVERGENCE BEHAVIOUR OF INEXACT NEWTON METHODS

A high order adaptive finite element method for solving nonlinear hyperbolic conservation laws

A derivative-free nonmonotone line search and its application to the spectral residual method

EXISTENCE VERIFICATION FOR SINGULAR ZEROS OF REAL NONLINEAR SYSTEMS

Topics. The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems

GLOBALIZATION TECHNIQUES FOR NEWTON KRYLOV METHODS AND APPLICATIONS TO THE FULLY-COUPLED SOLUTION OF THE NAVIER STOKES EQUATIONS

Termination criteria for inexact fixed point methods

A new Newton like method for solving nonlinear equations

Keywords: Nonlinear least-squares problems, regularized models, error bound condition, local convergence.

Coupling SUNDIALS with OpenFOAM 6 th OpenFOAM Workshop. E. David Huckaby June 14, 2011

Iterative Methods for Solving A x = b

Bindel, Spring 2016 Numerical Analysis (CS 4220) Notes for

THE solution of the absolute value equation (AVE) of

Search Directions for Unconstrained Optimization

Conjugate gradient method. Descent method. Conjugate search direction. Conjugate Gradient Algorithm (294)

Domain-decomposed Fully Coupled Implicit Methods for a Magnetohydrodynamics Problem

Numerical Methods for Large-Scale Nonlinear Systems

Scientific Computing: An Introductory Survey

A Robust Preconditioned Iterative Method for the Navier-Stokes Equations with High Reynolds Numbers

Convergence Analysis of Inexact Infeasible Interior Point Method. for Linear Optimization

1. Introduction. In this paper we derive an algorithm for solving the nonlinear system

1. Introduction. The problem of interest is a system of nonlinear equations

MULTIGRID METHODS FOR NONLINEAR PROBLEMS: AN OVERVIEW

WHEN studying distributed simulations of power systems,

Our setting is a hybrid of two ideas, splitting methods for nonsmooth nonlinear equations [, ] and pseudo-transient continuation (Ψtc) [ 5], a method

Lecture Notes to Accompany. Scientific Computing An Introductory Survey. by Michael T. Heath. Chapter 5. Nonlinear Equations

Globalization techniques for Newton Krylov methods and applications to the fully-coupled solution of the Navier Stokes equations

Iterative methods for Linear System of Equations. Joint Advanced Student School (JASS-2009)

A Novel Approach for Solving the Power Flow Equations

Preliminary Results of GRAPES Helmholtz solver using GCR and PETSc tools

Improved Damped Quasi-Newton Methods for Unconstrained Optimization

Implicit solvers for large-scale nonlinear problems

Numerical Methods I Solving Square Linear Systems: GEM and LU factorization

IMA Preprint Series # 2286

Optimization. Charles J. Geyer School of Statistics University of Minnesota. Stat 8054 Lecture Notes

M.A. Botchev. September 5, 2014

GMRES ON (NEARLY) SINGULAR SYSTEMS

Newton s method for approximating x, formulated as follows" CHOOSING THE FORCING TERMS IN AN INEXACT NEWTON METHOD*

Linear Algebra Massoud Malek

Linear algebra issues in Interior Point methods for bound-constrained least-squares problems

On fast trust region methods for quadratic models with linear constraints. M.J.D. Powell

CS 450 Numerical Analysis. Chapter 5: Nonlinear Equations

Part 3: Trust-region methods for unconstrained optimization. Nick Gould (RAL)

EECS 275 Matrix Computation

Nonlinear Optimization for Optimal Control

Lecture 11: CMSC 878R/AMSC698R. Iterative Methods An introduction. Outline. Inverse, LU decomposition, Cholesky, SVD, etc.

1. Introduction. In this paper we discuss an algorithm for equality constrained optimization problems of the form. f(x) s.t.

Programming, numerics and optimization

Newton additive and multiplicative Schwarz iterative methods

ITERATIVE METHODS FOR NONLINEAR ELLIPTIC EQUATIONS

Algorithms for Nonlinear and Transient Problems

Properties of Preconditioners for Robust Linear Regression

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

Unconstrained optimization

An Alternative Three-Term Conjugate Gradient Algorithm for Systems of Nonlinear Equations

Recent advances in approximation using Krylov subspaces. V. Simoncini. Dipartimento di Matematica, Università di Bologna.

Optimisation non convexe avec garanties de complexité via Newton+gradient conjugué

Statistics 580 Optimization Methods

Chapter 7 Iterative Techniques in Matrix Algebra

Transcription:

Slide 1 Numerical Methods for Large-Scale Nonlinear Equations Homer Walker MA 512 April 28, 2005 Inexact Newton and Newton Krylov Methods a. Newton-iterative and inexact Newton methods. Slide 2 i. Formulation and local convergence. ii. Globally convergent methods. iii. Choosing the forcing terms. b. Newton Krylov methods. i. General considerations. ii. Matrix-free implementations.

a. Newton-iterative and inexact Newton methods. The model method will be... Slide 3 Newton s Method: Given an initial x. Iterate: Decide whether to stop or continue. Solve J(x)s = F (x). Update x x + s. Here, J(x) = F (x) = ( Fi(x) x j ) IR n n. About Newton s method, recall... Slide 4 Major strength: quadratic local convergence, which is often mesh-independent on discretized PDE problems [6], [1].. For general discussions of stopping, scaling, globalization procedures, local and global convergence, etc., see [6]. Assume throughout: F is continuously differentiable.

Suppose that iterative linear algebra methods are preferred for solving J(x)s = F (x). The resulting method is a Newton iterative (truncated Newton) method. Slide 5 Key aspect: J(x)s = F (x) is solved only approximately. Key issues: When should we stop the linear iterations? How should we globalize the method? Which linear solver should we use? The first two can be well treated in the strictly more general context of inexact Newton methods. An inexact Newton method [4] is any method each step of which reduces the norm of the local linear model of F. Slide 6 Inexact Newton Method [4]: Given an initial x. Iterate: Decide whether to stop or continue. Find some η [0, 1) and s that satisfy F (x) + J(x) s η F (x). Update x x + s.

A Newton iterative method fits naturally into this framework: Choose η [0, 1). Slide 7 Apply the iterative linear solver until F (x) + J(x) s η F (x). Used in this way, η is called a forcing term. The issue of stopping the linear iterations becomes the issue of choosing the forcing terms. Local convergence is controlled by choices of η [4]. Theorem [4]: Suppose F (x ) = 0 and J(x ) is invertible. If {x k } is an inexact Newton sequence with x 0 sufficiently near x, then η k η max < 1 = x k x q-linearly, η k 0 = x k x q-superlinearly, Slide 8 If also J is Lipschitz continuous at x, then η k = O( F (x k ) ) = x k x q-quadratically. For some β < 1, x k+1 x J(x ) β x k x J(x ) for sufficiently large k, where w J(x ) J(x ) w. x k+1 x β k x k x, where β k 0. For some λ, J(x) J(x ) λ x x for x near x. For some C, x k+1 x C x k x 2 for all k.

Proof idea: Suppose F (x) + J(x) s η F (x). Set x + = x + s. We have F (x +) F (x) + J(x) s = F (x +) < η F (x). Slide 9 Near x... F (x) = F (x) F (x ) J(x ) (x x ), F (x +) = F (x +) F (x ) J(x ) (x + x ). So J(x ) (x + x ) < η J(x ) (x x ), i.e., x + x J(x ) < η x x J(x ) A globally convergent algorithm is... Inexact Newton Backtracking (INB) Method [7]: Given an initial x and t (0, 1), η max [0, 1), t (0, 1), and 0 < θ min < θ max < 1. Slide 10 Iterate: Decide whether to stop or continue. Choose initial η [0, η max] and s such that F (x) + J(x) s η F (x). Evaluate F (x + s). While F (x + s) > [1 t(1 η)] F (x), do: Choose θ [θ min, θ max]. Update s θs and η 1 θ(1 η). Revaluate F (x + s). Update x x + s and F (x) F (x + s).

The global convergence result is... Slide 11 Theorem [7, Th.6.1]: Suppose {x k } is produced by the INB method. If {x k } has a limit point x such that J(x ) is nonsingular, then F (x ) = 0 and x k x. Furthermore, the initial s k and η k are accepted for all sufficiently large k. Possibilities: x k. {x k } has limit points, and J is singular at each one. {x k } converges to x such that F (x ) = 0, J(x ) is nonsingular, and asymptotic convergence is determined by the initial η k s. Practical implementation of the INB method. Minor details (most as before): Choose η max near 1, e.g., η max =.9. Slide 12 Choose t small, e.g., t = 10 4. Choose θ min =.1, θ max =.5. Take to be an inner-product norm, e.g., = 2. Choose θ [θ min, θ max ] to minimize a quadratic or cubic that interpolates F (x k + θs k ).

Choosing the forcing terms. From [4], we know... η k constant < 1 = local linear convergence. η k 0 = local superlinear convergence. Slide 13 η k = O( F (x k ) ) = local quadratic convergence. These allow practically implementable choices of the η k s that lead to desirable asymptotic convergence rates. But there remains the danger of oversolving, i.e., imposing an accuracy on an approximate solution s of the Newton equation that leads to significant disagreement between F (x + s) and F (x) + J(x) s. Example: The driven cavity problem. (1/Re) 2 ψ + ψ x 1 ψ ψ x 2 x 2 x 1 ψ = 0 in D = [0, 1] [0, 1], On D, ψ = 0 and ψ { 1 on top. n = 0 on the sides and bottom. Slide 14 Streamlines for Re = 10, 000.

{ For η k = min F (x k ) 2, 1 k + 2 } (from [5]),... 0 2 Slide 15 Log Residual Norm 4 6 8 10 12 Linear Residual Norm Nonlinear Function Norm New Nonlinear Iterations 14 0 50 100 150 200 250 300 Total GMRES(20) Iterations Performance on the driven cavity problem, Re = 500. Gaps indicate oversolving. Forcing term choices have been proposed in [8] that are aimed at reducing oversolving. Here s one... Choice 1: Set η k = min {η max, η k }, where Slide 16 F (x k) F (x k 1 ) + J(x k 1 ) s k 1 η k = F (x k 1 ) This directly reflects the (dis)agreement between F and its local linear model at the previous step. This is invariant under multiplication of F by a scalar.

If we use η k given by Choice 1 in the backtracking method, we can combine a local convergence result in [8] with the previous global result to obtain... Theorem: Suppose {x k } is produced by the INB method with each η k given by Choice 1. If {x k } has a limit point x such that J(x ) is nonsingular and J is Lipschitz continuous at x, then F (x ) = 0 and x k x with Slide 17 x k+1 x β x k x x k 1 x. for some β independent of k. It follows that the convergence is... r-order (1 + 5)/2, q-superlinear, two-step q-quadratic. This and other choices in [8] may become too small too quickly away from a solution. We recommend safeguards that work against this. Slide 18 Rationale: If large forcing terms are appropriate at some point, then dramatically smaller forcing terms should be justified over several iterations before usage. Choice 1 safeguard [8]: Modify η k by whenever η (1+ 5)/2 k 1 >.1. η k max{η k, η (1+ 5)/2 k 1 }

For safeguarded Choice 1 η k s,... 0 2 Slide 19 Log Residual Norm 4 6 8 10 12 Linear Residual Norm Nonlinear Function Norm New Nonlinear Iterations 14 0 50 100 150 200 Total GMRES(20) Iterations Performance on the driven cavity problem, Re = 500. The inverted triangle indicates the safeguard value was used. b. Newton Krylov methods. Slide 20 Idea: Implement a Newton iterative method using a Krylov subspace method as the linear solver. The term appears to have originated with [3]. Naming conventions: Newton-GMRES, Newton Krylov Schwarz (NKS), Newton-Krylov-Multigrid (NKMG),... Truncated Newton originated with [5], which outlined an implementation of Newton with CG.

General considerations. The linear system is J(x) s = F (x). The usual initial approximate solution is s 0 = 0. The linear residual norm F (x) + J(x) s is just the local linear model norm. Slide 21 About preconditioning... Preconditioning on the right retains compatibility between the norms used in the linear and nonlinear inexact Newton strategies. Preconditioning on the left may introduce incompatibilities. It is safe to precondition the problem on the left, i.e., to solve M 1 F (x) = 0 for an M that is used without change throughout the solution process. Matrix-free implementations. Krylov subspace methods require only products of J(x) and sometimes J(x) T with vectors. Slide 22 There are possibilities for producing these without creating and storing J(x). One possibility for products involving either J(x) or J(x) T is automatic differentiation. This is actively being explored in the Mathematics and Computer Science Division, Argonne National Lab. See the ANL Computational Differentiation Project web page, www-unix.mcs.anl.gov/autodiff/index.html.

A very widely used technique, applicable when only products involving J(x) are needed, is finite-difference approximation. For a local convergence analysis, see [2]. Given v IR n, formulas for approximating J(x)v to 1st, 2nd, 4th, and 6th order are... Slide 23 1 [F (x + δv) F (x)], δ 1 [F (x + δv) F (x δv)], 2δ ] 1 [8F (x + δ2 6δ v) 8F (x δ2 v) F (x + δv) + F (x δv), 1 90δ [256F (x + δ4 v) 256F (x δ4 v) 40F (x + δ2 v) + 40F (x δ2 v) + F (x + δv) F (x δv) ]. The 1st-order formula is very commonly used; the others very rarely used (although sometimes they re needed, see [13], [12]). Choosing δ. As before... We try to choose δ to roughly balance truncation and floating point error. Fairly well-justified choices can be made for scalar functions. The justifications weaken with vector functions. Nothing is foolproof. Slide 24 A choice used in [12] that approximately minimizes a bound on the relative error in the difference approximation is based on... [ ] 1/(p+1) (1 + x )ɛf δ =, v where p is the difference order and ɛ F is the relative error in F -evaluations ( function precision ). The main underlying assumption is that F and its derivatives up to order p + 1 have about the same scale. A crude heuristic is δ = ɛ 1/(p+1), where ɛ is machine epsilon.

Additional references and software. In addition to the classic book of Dennis and Schnabel [6], I recommend the books by Kelley, especially [9] and also [10] and [11]. These come with MATLAB software. You can download PDF files of [9] and [10] as well as the associated software from the SIAM sites http://www.siam.org/books/kelley/kelley.html and http://www.siam.org/books/fr18/, resp. See also Kelley s website http://www4.ncsu.edu/eos/users/c/ctkelley/www/tim.html. Slide 25 For challenging large-scale applications, I recommend the following publicly available, downloadable software: NITSOL This is a flexible Newton Krylov code written in Fortran. It is described in [12] and downloadable at http://users.wpi.edu/ ~ walker/nitsol/. NOX and KINSOL These are powerful codes for parallel solution of large-scale nonlinear systems, downloadable at http://software.sandia.gov/trilinos/packages/nox/ and http://www.llnl.gov/casc/sundials/, resp. PETSc This is a broad suite of codes for parallel solution of large-scale linear and nonlinear systems plus ancillary tasks such as preconditioning. It is downloadable at http://www-unix.mcs.anl.gov/petsc/petsc-as/.

References [1] E. L. Allgower, K. Böhmer, F. A. Potra, and W. C. Rheinboldt, A mesh-independence principle for operator equations and their discretizations, SIAM J. Numer. Anal., 23 (1986), pp. 160 169. [2] P. N. Brown, A local convergence theory for combined inexact-newton/finite difference projection methods, SIAM J. Numer. Anal., 24 (1987), pp. 407 434. [3] P. N. Brown and Y. Saad, Hybrid Krylov methods for nonlinear systems of equations, SIAM J. Sci. Stat. Comput., 11 (1990), pp. 450 481. [4] R. S. Dembo, S. C. Eisenstat, and T. Steihaug, Inexact Newton methods, SIAM J. Numer. Anal., 19 (1982), pp. 400 408. [5] R. S. Dembo and T. Steihaug, Truncated Newton algorithms for large-scale optimization, Math. Prog., 26 (1983), pp. 190 212. [6] J. E. Dennis, Jr. and R. B. Schnabel, Numerical Methods for Unconstrained Optimization and Nonlinear Equations, vol. 16 of Classics in Applied Mathematics, SIAM, Philadelphia, PA, 1996. [7] S. C. Eisenstat and H. F. Walker, Globally convergent inexact Newton methods, SIAM J. Optimization, 4 (1994), pp. 393 422. [8], Choosing the forcing terms in an inexact Newton method, SIAM J. Sci. Comput., 17 (1996), pp. 16 32. [9] C. T. Kelley, Iterative Methods for Linear and Nonlinear Equations, vol. 16 of Frontiers in Applied Mathematics, SIAM, Philadelphia, PA, 1995. [10], Iterative Methods for Optimization, vol. 18 of Frontiers in Applied Mathematics, SIAM, Philadelphia, PA, 1999. [11], Solving Nonlinear Equations with Newton s Method, vol. 1 of Fundamentals of Algorithms, SIAM, Philadelphia, PA, 2003. [12] M. Pernice and H. F. Walker, NITSOL: a Newton iterative solver for nonlinear systems, SIAM J. Sci. Comput., 19 (1998), pp. 302 318. [13] K. Turner and H. F. Walker, Efficient high accuracy solutions with GMRES(m), SIAM J. Sci. Stat. Comput., 13 (1992), pp. 815 825.