The convergence of stationary iterations with indefinite splitting

Similar documents
Up to this point, our main theoretical tools for finding eigenvalues without using det{a λi} = 0 have been the trace and determinant formulas

CHAPTER 5. Basic Iterative Methods

CAAM 454/554: Stationary Iterative Methods

Lecture 18 Classical Iterative Methods

JACOBI S ITERATION METHOD

Some definitions. Math 1080: Numerical Linear Algebra Chapter 5, Solving Ax = b by Optimization. A-inner product. Important facts

Modeling, equilibria, power and risk

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for

c 1995 Society for Industrial and Applied Mathematics Vol. 37, No. 1, pp , March

Solving Linear Systems

Optimality conditions and complementarity, Nash equilibria and games, engineering and economic application

Math Introduction to Numerical Analysis - Class Notes. Fernando Guevara Vasquez. Version Date: January 17, 2012.

Solving equilibrium problems using extended mathematical programming and SELKIE

Numerical Methods - Numerical Linear Algebra

Algebra C Numerical Linear Algebra Sample Exam Problems

9.1 Preconditioned Krylov Subspace Methods

Scientific Computing WS 2018/2019. Lecture 9. Jürgen Fuhrmann Lecture 9 Slide 1

Continuous methods for numerical linear algebra problems

Computational Methods. Systems of Linear Equations

Linear Programming: Simplex

DEN: Linear algebra numerical view (GEM: Gauss elimination method for reducing a full rank matrix to upper-triangular

Lecture 1. 1 Conic programming. MA 796S: Convex Optimization and Interior Point Methods October 8, Consider the conic program. min.

Iterative Methods for Solving A x = b

Iterative Methods for Ax=b

Lecture 7: Positive Semidefinite Matrices

Math/Phys/Engr 428, Math 529/Phys 528 Numerical Methods - Summer Homework 3 Due: Tuesday, July 3, 2018

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

MOPEC: Multiple Optimization Problems with Equilibrium Constraints

Symmetric Matrices and Eigendecomposition

EXAMPLES OF CLASSICAL ITERATIVE METHODS

Review of matrices. Let m, n IN. A rectangle of numbers written like A =

Iterative Methods. Splitting Methods

Iterative Methods and Multigrid

Part 4: Active-set methods for linearly constrained optimization. Nick Gould (RAL)

MATRIX SPLITTING PRINCIPLES

Iterative Methods and Multigrid

Copositive Plus Matrices

Algorithms for Constrained Optimization

5.5 Quadratic programming

Solving Linear Systems

UNDERGROUND LECTURE NOTES 1: Optimality Conditions for Constrained Optimization Problems

ANALYTICAL MATHEMATICS FOR APPLICATIONS 2018 LECTURE NOTES 3

Iterative methods for Linear System

Econ Slides from Lecture 7

Math Matrix Algebra

ORIE 6300 Mathematical Programming I August 25, Recitation 1

BOOK REVIEWS 169. minimize (c, x) subject to Ax > b, x > 0.

Next topics: Solving systems of linear equations

Optimality, Duality, Complementarity for Constrained Optimization

3. THE SIMPLEX ALGORITHM

Stabilization and Acceleration of Algebraic Multigrid Method

10.2 ITERATIVE METHODS FOR SOLVING LINEAR SYSTEMS. The Jacobi Method

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88

Linear algebra issues in Interior Point methods for bound-constrained least-squares problems

A Constraint-Reduced MPC Algorithm for Convex Quadratic Programming, with a Modified Active-Set Identification Scheme

Linear Algebra Massoud Malek

Structural and Multidisciplinary Optimization. P. Duysinx and P. Tossings

Notes on Some Methods for Solving Linear Systems

Lecture 11. Fast Linear Solvers: Iterative Methods. J. Chaudhry. Department of Mathematics and Statistics University of New Mexico

9. Iterative Methods for Large Linear Systems

Numerical Methods I Non-Square and Sparse Linear Systems

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Chapter 7 Iterative Techniques in Matrix Algebra

Numerical Analysis: Solutions of System of. Linear Equation. Natasha S. Sharma, PhD

Matrix Operations: Determinant

AM 205: lecture 18. Last time: optimization methods Today: conditions for optimality

A Distributed Newton Method for Network Utility Maximization, II: Convergence

Jordan Journal of Mathematics and Statistics (JJMS) 5(3), 2012, pp A NEW ITERATIVE METHOD FOR SOLVING LINEAR SYSTEMS OF EQUATIONS

Semidefinite Programming Basics and Applications

A DECOMPOSITION PROCEDURE BASED ON APPROXIMATE NEWTON DIRECTIONS

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

Introduction to Scientific Computing

4.6 Iterative Solvers for Linear Systems

Here is an example of a block diagonal matrix with Jordan Blocks on the diagonal: J

Course Notes: Week 1

1 Computing with constraints

6.4 Krylov Subspaces and Conjugate Gradients

In English, this means that if we travel on a straight line between any two points in C, then we never leave C.

Recall : Eigenvalues and Eigenvectors

Computing Solution Concepts of Normal-Form Games. Song Chong EE, KAIST

3 (Maths) Linear Algebra

FEM and Sparse Linear System Solving

Iterative methods for Linear System of Equations. Joint Advanced Student School (JASS-2009)

Lecture 6 Positive Definite Matrices

Lecture 15 Newton Method and Self-Concordance. October 23, 2008

Positive Definite Matrix

A NEW EFFECTIVE PRECONDITIONED METHOD FOR L-MATRICES

Lecture Note 7: Iterative methods for solving linear systems. Xiaoqun Zhang Shanghai Jiao Tong University

Chapter 3. Determinants and Eigenvalues

Written Examination

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Exchange Market Equilibria with Leontief s Utility: Freedom of Pricing Leads to Rationality

In particular, if A is a square matrix and λ is one of its eigenvalues, then we can find a non-zero column vector X with

Unconstrained optimization

EIGENVALUE PROBLEMS. Background on eigenvalues/ eigenvectors / decompositions. Perturbation analysis, condition numbers..

Computational Linear Algebra

Linear programming. Saad Mneimneh. maximize x 1 + x 2 subject to 4x 1 x 2 8 2x 1 + x x 1 2x 2 2

6.254 : Game Theory with Engineering Applications Lecture 7: Supermodular Games

Lecture 16 Methods for System of Linear Equations (Linear Systems) Songting Luo. Department of Mathematics Iowa State University

Constrained Nonlinear Optimization Algorithms

Transcription:

The convergence of stationary iterations with indefinite splitting Michael C. Ferris Joint work with: Tom Rutherford and Andy Wathen University of Wisconsin, Madison 6th International Conference on Complementarity Problems Berlin, Germany August 7, 2014 Ferris (Univ. Wisconsin) MOPEC splitting ICCP 2014 1 / 22

The problems VI(F, C): x C, F (x ), x x 0, x C MOPEC: x, y: x i solves min x i K i (x i,y) θ(x i, x i, y), i y solves VI (F (x, ), C) A 1 A 1,2 A 1,p E 1 A 2,1 A 2............ A p-1,p E p-1 A p,1 A p,p-1 A p E p F 1 F p-1 F p D Ferris (Univ. Wisconsin) MOPEC splitting ICCP 2014 2 / 22

Strongly Convex Nash Equilibria min x 1 0 min x 2 0 No solution for θ 1: 1 2 x 1 2 θx 1 x 2 4x 1 s.t. x 1 + x 2 1 1 2 x 2 2 x 1 x 2 3x 2 x 1 (x 2 ) = (θx 2 + 4) +, x 2 (x 1 ) = (x 1 + 3) + Solution 4 3 θ < 1: x 1 = 4+3θ 1 θ, x 2 = x 1 + 3 Solution θ 4 3 : x 1 = 0, x 1 = 3 Jacobi works provided θ < 1, but theory fails Ferris (Univ. Wisconsin) MOPEC splitting ICCP 2014 3 / 22

The Issues This is not the optimality conditions of a single optimization problem: 1 1 θ x 1 4 x 1 0 1 0 1 p 1 1 p 1 0 1 1 x 2 3 x 2 The matrix A in general is never diagonally dominant except in trivial cases Iterations based on succesive inversion of local blocks (or successive optimization of local strategies) can converge. We establish sufficient conditions which guarantee convergence of block Jacobi and block Gauss-Seidel iterations for such matrices. Ferris (Univ. Wisconsin) MOPEC splitting ICCP 2014 4 / 22

Iteration with Indefinite Splitting Ax = b Splitting A = P N naturally leads to a stationary iteration of the form x 0 arbitrary, Px k+1 = Nx k + b, k = 0, 1,... This iteration may or may not converge; simply applicable sufficient conditions for convergence are particularly valuable. Most well-known such conditions are diagonal dominance: if the preconditioner is P = diag(a) (leading to Jacobi iteration) or P is the lower triangular part of A (leading to Gauss-Seidel iteration), then convergence is guaranteed if the strict diagonal dominance condition a i,i > a i,j, i = 1,..., n (1) j=1,...,n,j i is satisfied by A = {a i,j, i, j = 1,..., n}. Ferris (Univ. Wisconsin) MOPEC splitting ICCP 2014 5 / 22

Weaker diagonal dominance conditions For irreducible matrices, it is well documented that the weaker condition a i,i a i,j, i = 1,..., n (2) j=1,...,n,j i is also sufficient provided strict inequality holds for at least one row index, i The condition (1) or (2) also guarantees that A R n n is invertible, so a unique solution exists. Ferris (Univ. Wisconsin) MOPEC splitting ICCP 2014 6 / 22

The Setting We focus on matrices of the form A 1 A 1,2 A 1,p E 1. A 2,1 A.. 2.. A =....... Ap 1,p E p 1 A p,1 A p,p 1 A p E p F 1 F p 1 F p D (3) where A i = [ Qi B T i B i 0 ], i = 1,..., p (4) with Q i = Q T i R n i n i positive definite and B i R m i n i of full rank m i < n i for each i (m i > 0). These conditions guarantee that each A i is invertible. The submatrix D R s s, s 0 must be symmetric and invertible (unless s = 0). Ferris (Univ. Wisconsin) MOPEC splitting ICCP 2014 7 / 22

Existing Block Theory For the blocked matrix (3) a result of Feingold and Varga (1962) applies: If A is block irreducible and ( A 1 i 2 ) 1 E i 2 + A i,j 2, i = 1,..., p (5) and ( D 1 2 ) 1 j=1,...,p,j i j=1,...,p,j i F i 2 (6) with strict inequality in (6) or for at least one index, i, in (5), then A is invertible (existence and uniqueness) Ferris (Univ. Wisconsin) MOPEC splitting ICCP 2014 8 / 22

Relation to Iteration Before considering these conditions in more detail, consider a block Jacobi or block Gauss-Seidel iteration based on the splitting with P = A 1 A 2... A p D or P = A 1 A 2,1 A 2....... A p,1 A p,p-1 A p F 1 F p-1 F p D Asymptotic convergence of the corresponding stationary (or simple) iteration will be guaranteed for any starting vector if all of the eigenvalues, λ, of I P 1 A lie strictly inside the unit disc. Ferris (Univ. Wisconsin) MOPEC splitting ICCP 2014 9 / 22

The link Such eigenvalues satisfy (I P 1 A)x = λx, x 0 or equivalently (A + (λ 1)P)x = 0, x 0. In the case of block Jacobi, asymptotic convergence will be guaranteed if there does not exist any λ with λ 1 such that the matrix λa 1 A 1,2 A 1,p E 1. A 2,1 λa.. 2.. A(λ) = A + (λ 1)P =....... Ap 1,p E p 1 A p,1 A p,p 1 λa p E p F 1 F p 1 F p λd is singular. But ( (λa i ) 1 2 ) 1 = λ ( A 1 i 2 ) 1 ( A 1 i 2 ) 1 whenever λ 1 with a identical argument holding for D. Hence satisfaction of the conditions (5),(6) not only guarantees invertibility of A, but also guarantees covergence of the block Jacobi iteration. Ferris (Univ. Wisconsin) MOPEC splitting ICCP 2014 10 / 22

Another piece Let µ i denote the smallest eigenvalue of the positive definite matrix Q i and γ i denote the smallest eigenvalue of the positive definite (Schur complement) matrix B i Q 1 i Bi T, then there are no eigenvalues of [ Qi B A i = i T ] B i 0 in the interval ( ( ) ) 1 µ i µ 2i + 4γ i µ i, µ i 2 which contains the origin. Ferris (Univ. Wisconsin) MOPEC splitting ICCP 2014 11 / 22

Finally... If the matrix A given by (3),(4) is block irreducible, then it is invertible and the block Jacobi and block Gauss-Seidel iterations for a linear system Ax = b converge to x for any starting vector if { ) } 1 min ( µ 2i + 4γ i µ i µ i, µ i 2 E i 2 + A i,j 2, i = 1,..., p (7) and d j=1,...,p,j i j=1,...,p,j i F i 2 (8) with strict inequality in (8) 1 or for at least one index, i, in (7). 1 d is the absolute value of eighenvalue of D closest to origin Ferris (Univ. Wisconsin) MOPEC splitting ICCP 2014 12 / 22

A Simplification If for each i = 1,..., p, γ i 2µ i then A is invertible and the block Jacobi and block Gauss-Seidel iterations for a linear system Ax = b converge to x for any starting vector if µ i E i 2 + A i,j 2, i = 1,..., p (9) and d j=1,...,p,j i j=1,...,p,j i F i 2 (10) with strict inequality in (10) or for at least one index, i, in (9). Ferris (Univ. Wisconsin) MOPEC splitting ICCP 2014 13 / 22

Simple example min x 1 0.5x 2 1 θx 1 x 2 4x 1 s.t. 2x 1 + 0.5x 2 = 1 min x 2 0.5x 2 2 x 1 x 2 3x 2 1 2 θ x 1 4 2 0.5 p 1 = 1 1 1 x 2 3 Solution: x 1 = 0.2, x 2 = 2.8, p 1 = 1.4θ 2.1 Jacobi works, but convergence guaranteed if θ < 3/2. Ferris (Univ. Wisconsin) MOPEC splitting ICCP 2014 14 / 22

Extensions Can also prove same result for SOR schemes Can apply regularization (proximal iterations) on the constraints: for ɛ i, α i > 0 [ Qi + α A i = i I Bi T ], ɛ i I B i can be used for some subset (or indeed all) of the indices i = 1,..., p. This increases the value of µ i and γ i in the above and strengthens the theory No rates given here Ferris (Univ. Wisconsin) MOPEC splitting ICCP 2014 15 / 22

Strongly convex optimization 1 min x 1 2 x 1 2 x 1 x 2 4x 1 s.t. x 1 + x 2 = 1 1 min x 2 2 x 2 2 x 1 x 2 3x 2 1 + α 1 1 1 x 1 4 1 ɛ 1 1 p 1 = 1 1 1 x 2 3 Solution: x 1 = 1, x 2 = 2, p 1 = 7 Jacobi fails: after 4 steps back at (1, 1) T Modified Jacobi α 1, ɛ 1 = 0.1 solves in 50 steps Ferris (Univ. Wisconsin) MOPEC splitting ICCP 2014 16 / 22

Extension to Inequality Case (QP iterates) Each Jacobi iterate replaces system of equations with solution of a small(er) scale Quadratic Programs: min x i 0 1 2 x i T Q i x i + c i (x i ) T x i s.t. B i x i = b(x i ) Solution is typically found in (many) fewer iterations than unconstrained case Can use any QP solver for subproblems (and/or VI solver) Ferris (Univ. Wisconsin) MOPEC splitting ICCP 2014 17 / 22

Extension to Inequality Case (Normal Map) Replace systems of equations by (normal map formulation of) complementarity problems i: Q i ((x i ) + ) B T i p + c i + x i (x i ) + = 0 B i ((x i ) + ) = b i Note this is a natural extension of the case considered above Choose active set at each iteration based on prediction from previous iteration Need to employ a regularization on the subproblem constraints Apply theory to all selections of the resulting linear systems Ferris (Univ. Wisconsin) MOPEC splitting ICCP 2014 18 / 22

Extension to Inequality Case (Interior Point) Apply interior point code to solve each QP subproblem min x i 0 1 2 x i T Q i x i + c i (x i ) T x i s.t. B i x i = b(x i ) Resulting systems to solve have form [ Qi + X 1 i W i Bi T B i 0 ] i = 1,..., p where X and W are diagonals of iterates and slacks at previous iteration Update barrier parameter after each Jacobi step Ferris (Univ. Wisconsin) MOPEC splitting ICCP 2014 19 / 22

Economic Application Model is a partial equilibrium, geographic exchange model. Goods are distinguished by region of origin. There is one unit of region r goods. These goods may be consumed in region r or they may be exported. Each region solves: min f r (X, T ) s.t. F (X, T ) = 0, T j = T j, j r X,T r where f r (X, T ) is a quadratic form and F (X, T ) is linear and defines X uniquely as a function of T. F (X, T ) defines an equilibrium; here it is simply a set of equations, not a complementarity problem Ferris (Univ. Wisconsin) MOPEC splitting ICCP 2014 20 / 22

Results Gauss-Seidel residuals Iteration deviation 1 3.14930 2 0.90970 3 0.14224 4 0.02285 5 0.00373 6 0.00061 7 0.00010 8 0.00002 9 0.00000 Tariff revenue region SysOpt MOPEC 1 0.117 0.012 2 0.517 0.407 3 0.496 0.214 4 0.517 0.407 5 0.117 0.012 Note that competitive solution produces much less revenue than system optimal solution Model has non-convex objective, but each subproblem is solved globally (lindoglobal) Ferris (Univ. Wisconsin) MOPEC splitting ICCP 2014 21 / 22

Conclusions MOPEC problems capture complex interactions between optimizing agents Policy implications addressable using MOPEC MOPEC available to use within the GAMS modeling system New sufficient conditions for existence, uniqueness and convergence shown in special cases Many new settings available for deployment; need for more theoretic and algorithmic enhancements Ferris (Univ. Wisconsin) MOPEC splitting ICCP 2014 22 / 22