from the primal-dual interior-point algorithm (Megiddo [16], Kojima, Mizuno, and Yoshise

Similar documents
A Generalized Homogeneous and Self-Dual Algorithm. for Linear Programming. February 1994 (revised December 1994)

1. Introduction A number of recent papers have attempted to analyze the probabilistic behavior of interior point algorithms for linear programming. Ye

Numerical Comparisons of. Path-Following Strategies for a. Basic Interior-Point Method for. Revised August Rice University

Abstract. A new class of continuation methods is presented which, in particular,

AN INTERIOR POINT METHOD, BASED ON RANK-ONE UPDATES, Jos F. Sturm 1 and Shuzhong Zhang 2. Erasmus University Rotterdam ABSTRACT

1 Outline Part I: Linear Programming (LP) Interior-Point Approach 1. Simplex Approach Comparison Part II: Semidenite Programming (SDP) Concludin

Enlarging neighborhoods of interior-point algorithms for linear programming via least values of proximity measure functions

An Infeasible Interior-Point Algorithm with full-newton Step for Linear Optimization

Primal-Dual Interior-Point Methods for Linear Programming based on Newton s Method

Following The Central Trajectory Using The Monomial Method Rather Than Newton's Method

Room 225/CRL, Department of Electrical and Computer Engineering, McMaster University,

An O(nL) Infeasible-Interior-Point Algorithm for Linear Programming arxiv: v2 [math.oc] 29 Jun 2015

IBM Almaden Research Center,650 Harry Road Sun Jose, Calijornia and School of Mathematical Sciences Tel Aviv University Tel Aviv, Israel

New Infeasible Interior Point Algorithm Based on Monomial Method

APPROXIMATING THE COMPLEXITY MEASURE OF. Levent Tuncel. November 10, C&O Research Report: 98{51. Abstract

On well definedness of the Central Path

Lecture 10. Primal-Dual Interior Point Method for LP

A full-newton step feasible interior-point algorithm for P (κ)-lcp based on a new search direction

Primal-dual relationship between Levenberg-Marquardt and central trajectories for linearly constrained convex optimization

A Sublinear Parallel Algorithm for Stable Matching. network. This result is then applied to the stable. matching problem in Section 6.

On Superlinear Convergence of Infeasible Interior-Point Algorithms for Linearly Constrained Convex Programs *

I.3. LMI DUALITY. Didier HENRION EECI Graduate School on Control Supélec - Spring 2010

Interior-Point Methods for Linear Optimization

Research Note. A New Infeasible Interior-Point Algorithm with Full Nesterov-Todd Step for Semi-Definite Optimization

that nds a basis which is optimal for both the primal and the dual problems, given

A Strongly Polynomial Simplex Method for Totally Unimodular LP

A PREDICTOR-CORRECTOR PATH-FOLLOWING ALGORITHM FOR SYMMETRIC OPTIMIZATION BASED ON DARVAY'S TECHNIQUE

Introduction to optimization

Interior Point Methods. We ll discuss linear programming first, followed by three nonlinear problems. Algorithms for Linear Programming Problems


Interior Point Methods for Mathematical Programming

A Second-Order Path-Following Algorithm for Unconstrained Convex Optimization

Lecture 5. Theorems of Alternatives and Self-Dual Embedding

Infeasible Primal-Dual (Path-Following) Interior-Point Methods for Semidefinite Programming*

A Full Newton Step Infeasible Interior Point Algorithm for Linear Optimization

Infeasible Primal-Dual (Path-Following) Interior-Point Methods for Semidefinite Programming*

Exploiting Sparsity in Primal-Dual Interior-Point Methods for Semidefinite Programming.

y Ray of Half-line or ray through in the direction of y

On the behavior of Lagrange multipliers in convex and non-convex infeasible interior point methods

Interior Point Methods in Mathematical Programming

Some Perturbation Theory. James Renegar. School of Operations Research. Cornell University. Ithaca, New York October 1992

Abstract. A complementarity problem with a continuous mapping f from

Semidefinite Programming

SOME GENERALIZATIONS OF THE CRISS-CROSS METHOD. Emil Klafszky Tamas Terlaky 1. Mathematical Institut, Dept. of Op. Res.

A FULL-NEWTON STEP INFEASIBLE-INTERIOR-POINT ALGORITHM COMPLEMENTARITY PROBLEMS

Linear Programming: Simplex

Example Bases and Basic Feasible Solutions 63 Let q = >: ; > and M = >: ;2 > and consider the LCP (q M). The class of ; ;2 complementary cones

COMPARATIVE STUDY BETWEEN LEMKE S METHOD AND THE INTERIOR POINT METHOD FOR THE MONOTONE LINEAR COMPLEMENTARY PROBLEM

A primal-simplex based Tardos algorithm

2 JOSE HERSKOVITS The rst stage to get an optimal design is to dene the Optimization Model. That is, to select appropriate design variables, an object

Springer-Verlag Berlin Heidelberg

A semidefinite relaxation scheme for quadratically constrained quadratic problems with an additional linear constraint

On the Local Quadratic Convergence of the Primal-Dual Augmented Lagrangian Method

Detecting Infeasibility in Infeasible-Interior-Point Methods for Optimization

c 2005 Society for Industrial and Applied Mathematics

ON THE ARITHMETIC-GEOMETRIC MEAN INEQUALITY AND ITS RELATIONSHIP TO LINEAR PROGRAMMING, BAHMAN KALANTARI

An E cient A ne-scaling Algorithm for Hyperbolic Programming

Interior Point Algorithms for Constrained Convex Optimization

Rank-one LMIs and Lyapunov's Inequality. Gjerrit Meinsma 4. Abstract. We describe a new proof of the well-known Lyapunov's matrix inequality about

A Second Full-Newton Step O(n) Infeasible Interior-Point Algorithm for Linear Optimization

Smoothed Fischer-Burmeister Equation Methods for the. Houyuan Jiang. CSIRO Mathematical and Information Sciences

ARE202A, Fall Contents

Limiting behavior of the central path in semidefinite optimization

Numerical Methods for Model Predictive Control. Jing Yang

1 Introduction After over a decade of extraordinarily active research triggered by the seminal work of Karmarkar [5], the eld of interior-point method

Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem

Local Self-concordance of Barrier Functions Based on Kernel-functions

A PRIMAL-DUAL INTERIOR POINT ALGORITHM FOR CONVEX QUADRATIC PROGRAMS. 1. Introduction Consider the quadratic program (PQ) in standard format:

c 2000 Society for Industrial and Applied Mathematics

Advances in Convex Optimization: Theory, Algorithms, and Applications

460 HOLGER DETTE AND WILLIAM J STUDDEN order to examine how a given design behaves in the model g` with respect to the D-optimality criterion one uses

A new ane scaling interior point algorithm for nonlinear optimization subject to linear equality and inequality constraints

Nonlinear Optimization: What s important?

Optimization: Then and Now

2.098/6.255/ Optimization Methods Practice True/False Questions


A NEW PROXIMITY FUNCTION GENERATING THE BEST KNOWN ITERATION BOUNDS FOR BOTH LARGE-UPDATE AND SMALL-UPDATE INTERIOR-POINT METHODS

On Mehrotra-Type Predictor-Corrector Algorithms

Predictor-corrector methods for sufficient linear complementarity problems in a wide neighborhood of the central path

Uniform Boundedness of a Preconditioned Normal Matrix Used in Interior Point Methods

MATHEMATICS AND COMPUTER SCIENCE DIVISION, ARGONNE NATIONAL LABORATORY

CONSTRAINED NONLINEAR PROGRAMMING

א K ٢٠٠٢ א א א א א א E٤

On a wide region of centers and primal-dual interior. Abstract

INDEFINITE TRUST REGION SUBPROBLEMS AND NONSYMMETRIC EIGENVALUE PERTURBATIONS. Ronald J. Stern. Concordia University

The iterative convex minorant algorithm for nonparametric estimation

An upper bound for the number of different solutions generated by the primal simplex method with any selection rule of entering variables

IE 5531: Engineering Optimization I

290 J.M. Carnicer, J.M. Pe~na basis (u 1 ; : : : ; u n ) consisting of minimally supported elements, yet also has a basis (v 1 ; : : : ; v n ) which f

Structural and Multidisciplinary Optimization. P. Duysinx and P. Tossings

An interior-point gradient method for large-scale totally nonnegative least squares problems

LOWER BOUNDS FOR THE MAXIMUM NUMBER OF SOLUTIONS GENERATED BY THE SIMPLEX METHOD

4TE3/6TE3. Algorithms for. Continuous Optimization

A MULTIGRID ALGORITHM FOR. Richard E. Ewing and Jian Shen. Institute for Scientic Computation. Texas A&M University. College Station, Texas SUMMARY

Machine Learning. Support Vector Machines. Manfred Huber

An EP theorem for dual linear complementarity problems

A Note on Exchange Market Equilibria with Leontief s Utility: Freedom of Pricing Leads to Rationality

A SUFFICIENTLY EXACT INEXACT NEWTON STEP BASED ON REUSING MATRIX INFORMATION

A numerical implementation of a predictor-corrector algorithm for sufcient linear complementarity problem

Research Reports on Mathematical and Computing Sciences

Transcription:

1. Introduction The primal-dual infeasible-interior-point algorithm which we will discuss has stemmed from the primal-dual interior-point algorithm (Megiddo [16], Kojima, Mizuno, and Yoshise [7], Monteiro and Adler [19], Tanabe [22]) for linear programs. It has already been studied by many researchers (Lustig [12], Lustig, Marsten, and Shanno [13], Marsten, Subramanian, Saltzman, Lustig, and Shanno [14], Tanabe [22, 23], Vanderbei and Carpenter [24], etc.), and is known as practically ecient algorithms among numerous variations and extensions of the primal-dual interior-point algorithm. 1 Many numerical studies show that the algorithm solves large scale practical problems very eciently ([12, 13, 14], etc.). Theoretically, however, neither polynomial-time nor global convergence of the algorithm has been shown. The aim of the present paper is to propose a rule of controlling step lengths to ensure its global convergence. Let A be an m n matrix, b 2 R m,andc2r n. Consider the standard form linear program P: Minimize c T x subject to Ax = b x 0 and its dual D: Maximize b T y subject to A T y + z = c z 0: Throughout the paper, we assume that the matrix A has full row rank, i.e., rank A = m. If feasible solutions x of P and (y z) ofd satisfy x > 0 and z > 0, we call them interior feasible solutions of P and D, respectively. Wealsosay that (x y z) is a feasible solution (resp. an interior feasible solution, an optimal solution) of the primal-dual pair of P and D if x and (y z) are feasible solutions (resp. interior feasible solutions, optimal solutions) of the problems P and D, respectively. Let (x 1 y 1 z 1 ) be an initial point such that x 1 > 0 and z 1 > 0. At each iterate (x k y k z k ) of a primal-dual infeasible-interior-point algorithm, we compute a Newton direction (x y z) towards a point on the central trajectory (the path of centers, Megiddo [16], see also Fiacco and McCormick [3]), and then generate a new point (x k+1 y k+1 z k+1 ) such that x k+1 = x k + k px > 0 y k+1 = y k + k dy z k+1 = z k + k dz > 0: Here k p 0 and k d 0 denote primal and dual step lengths. Thus the infeasible-interiorpoint algorithm shares this basic structure with many of the primal-dual interior-point 1 Mehrotra [17] proposed more ecient methods using a predictor-corrector strategy. 1

algorithms developed so far (Choi, Monma, and Shanno [2], Kojima, Megiddo, Noma, and Yoshise [6], Kojima, Mizuno, and Yoshise [7, 8], Lustig [11, 12], McShane, Monma, and Shanno [15], Mizuno, Todd, and Ye[18],Tanabe [22, 23], Monteiro and Adler [19, 20], Ye [25], Ye, Guler, Tapia, and Zhang [26], etc.). The generated sequence f(x k y k z k )g is, however, not restricted to the interior of the feasible region an iterate (x k y k z k ) is required to satisfy neither the equality constraints Ax = b of P nor A T y +z = c of D, but only the positivity x > 0 and z > 0. Therefore, we can start from arbitrary (x 1 y 1 z 1 ) with strictly positive x 1 and z 1 to approach optimal solutions moving through not only the interior but also the outside of the feasible region of the primal-dual pair of P and D. The distinctive feature of the infeasible-interior-point algorithm mentioned above isa main advantage over primal-dual interior-point algorithms. When we apply a primal-dual interior-point algorithm to the problems P and D, we usually need to prepare an articial primal-dual pair of linear programs having a known interior feasible solution from which the algorithm starts. Lustig [12] pointed out a drawback ofthisapproachthatthe articial primal-dual pair of linear programs involves large constants called the big M and articial dense columns which often cause numerical instability and computational ineciency. He derived the limiting feasible direction (Newton direction) as the constant \big M" tends to innity, and showed that the primal-dual interior-point algorithm using the limiting feasible direction leads to the infeasible-interior-point algorithm. Lustig, Marsten and Shanno [13] showed the equivalence of the the limiting feasible direction and the Newton direction to the Karush-Kuhn-Tucker condition for the linear program P and D. To mitigate the drawback ofinterior-point algorithms, Kojima, Mizuno, and Yoshise [9] recently proposed an articial self-dual linear program with a single big M as well as a numerical method for updating the big M. But the computation of the Newton direction in each iteration of a primal-dual interior-point algorithm applied to the articial self-dual linear program is still a little more expensive thanthatofthe infeasible-interior-point algorithm. We alsomention that the infeasible-interior-point algorithm can be interpreted as an application of an interior-point algorithm to the articial self-dual linear program (see Section 4 of [9]). This paper proposes a rule for controlling the step length with which the primal-dual infeasible-interior-point algorithm takes large distinct step lengths k p in the primal space and k d in the dual space, and generates a sequence f(x k y k z k )g satisfying the following properties: (a) For any accuracy >0 required for the total complementarity,any tolerance p > 0 for the primal feasibility, any tolerance d for the dual feasibility andany large!, there exists a number k such that after k iterations the algorithm using the rule generates a point (x k y k z k ) which is either an approximate optimal solution 2

(x k y k z k ) satisfying or satises (x k ) T z k kax k ; bk p and ka T y k + z k ; ck d (1) k(x k z k )k 1 >! : (2) Here kuk 1 denotes the 1-norm of a vector u 2 R`, i.e., kuk 1 = P` i=1 j u i j. (b) If (2) holds, we can derive information on the infeasibilitysuch that the primal-dual pair of P and D has no feasible solution in a certain wide region of the primaldual space. (See [9] for a method of getting such information in interior-point algorithms.) In Section 2 we give the details of the primal-dual infeasible-interior-point algorithm using the step length control rule, where the inequalities (1) and (2) serve as stopping criteria. In Section 3 we establish that the algorithm enjoys property (a). In Sections 4 and 5, we discuss property (b). 2. An Infeasible-Interior-Point Algorithm It is convenient to denote the feasible region of the primal-dual pair of P and D as Q + \S, where Q + = f(x y z) 2 R n+m+n : x 0 z 0g S = f(x y z) 2 R n+m+n : Ax = b A T y + z = cg: We also denote by Q ++ the interior of the set Q + Q ++ = f(x y z) 2 R n+m+n : x > 0 z > 0g: The interior of the feasible region Q + \ S can be denoted by Q ++ \ S. Let 0 <<1, p > 0, and d > 0. The algorithm generates a sequence f(x k y k z k )g in the neighborhood N = f(x y z) 2 Q ++ : x i z i x T z=n (i =1 2... n) x T z p kax ; bk or kax ; bk p x T z d ka T y + z ; ck or ka T y + z ; ck d g of the central trajectory (the path of centers) consisting of the solutions (x y z) 2 Q ++ to the system of equations 0 1 Ax ; b B @ A T y + z ; c C A = 0 (3) Xz ; e 3

for all >0. Here X denotes the nn diagonal matrix with the coordinates of a vector x 2 R n and e =(1... 1) T 2 R n. Let 0 < 1 < 2 < 3 < 1. At each iteration, we assign the value 1 (x k ) T z k =n to the parameter, and then compute the Newton direction (x y z) at(x k y k z k )for the system (3) of equations. More precisely, (x y z) is the unique solution of the system of linear equations 0 B @ A 0 0 0 A T I Z k 0 X k 1 0 C B A @ x y z 1 C A = ; 0 B @ Ax k ; b A T y k + z k ; c X k z k ; e The parameters 2 and 3 control the primal and dual step lengths. 1 C A : (4) We can take an arbitrary initial point f(x 1 y 1 z 1 )g with x 1 > 0 and z 1 > 0, but we must choose the parameters, p, d and! such that (x 1 y 1 z 1 ) 2N and k(x 1 z 1 )k 1! : All the parameters p d p d! 1 2 and 3 may ormaynotdependon the input data for the problems P and D. Now we are ready to state our algorithm. Algorithm 2.1. Step 1: Let k =1. Step 2: If (1) or (2) holds then stop. Step 3: Let = 1 (x k ) T z k =n. Compute the unique solution (x y z) at (x k y k z k ) of the system (4) of equations. Step 4: Let k be the maximum of ~'s 1 such that the relations ) (x k y k z k )+(x y z) 2N (x k + x) T (z k + z) (1 ; (1 ; 2 ))(x k ) T z k (5) hold for every 2 [0 ~]. See Remark 3.1 for the computation of k. Step 5: Choose a primal step length k p 2 (0 1], a dual step length k d 2 (0 1] and a new iterate (x k+1 y k+1 z k+1 ) such that (x k+1 y k+1 z k+1 )=(x k + k px y k + k dy z k + k dz) 2N (x k+1 ) T z k+1 (1 ; k (1 ; 3 ))(x k ) T z k : Step 6: Increase k by 1. Go to Step 2. Since 0 < 2 < 3 < 1, the common value k is always available for both the primal step length k p and the dual step length k d although we can assign distinct values to them. In the next section, we will show existence of a positive number such that k is 4 ) (6)

not less than for every k as long as the iteration continues. This will lead to a nite termination of the algorithm at Step 2. We consider how large step lengths k p and k d we can choose subject to the condition (6) at Step 5 of the algorithm. The second inequality of (6) requires a reduction (1; k (1; 3 )) in the total complementarity x T z. By the denition of k and 0 < 2 < 3 < 1, a bigger reduction (1 ; k (1 ; 2 )) than (1 ; k (1 ; 3 )) is always possible. So the inequality seems reasonable. If we take a positive 3 less than but suciently close to 1, the inequality does little harm to take large step lengths. Now we focus our attention on the rst constraint of (6) which the step lengths k p and k d must satisfy. The denition of N consists of three kinds of relations x i z i x T z=n (i =1 2... n) (7) x T z p kax ; bk or kax ; bk p (8) x T z d ka T y + z ; ck or ka T y + z ; ck d : (9) As in the primal-dual interior-point algorithms (Kojima, Mizuno, and Yoshise [7], Mizuno, Todd, and Ye [18], etc.), the inequalities (7) prevent the generated sequence f(x k y k z k )g from reaching the boundary of Q ++ before the total complementarity(x k ) T z k attains 0. The other relations (8) and (9) play the role of excluding the possibility that the generated sequence f(x k y k z k )g might converge to an infeasible complementary solution (x y z ) 2 Q + (x ) T z = 0 such that kax ; bk p and/or ka T y + z ; ck d : Besides the primal feasibility tolerance p > 0 and the dual feasibility tolerance d > 0, which we assume xed in what follows, the set N involves the positive parameters, p, and d. To clarify the dependency on these parameters, we will write the set as N ( p d ). Then N ( p d ) N( 0 0 p 0 d) if 0 < ( p d ) ( 0 0 p 0 d) [fn ( p d ) : ( p d ) > 0g = Q ++ : Therefore, as we take smaller positive p and d, the set N ( p d )covers a larger subregion of Q ++ hence, we can take larger step lengths p and d satisfying (6). McShane, Monma, and Shanno [15], proposed taking large step lengths k p and k d such that k p =0:9995^k p and k d =0:9995^k d (10) where ^ k p = maxf : x k + x 0g ^ k d = maxf : z k + z 0g: 5

This choice of the step lengths is known to work very eciently in practice ([12, 13, 14, 15], etc.), but has not been shown to ensure the global convergence. The above observation on the set N ( p d ) suggests a combination of their step lengths with ours to ensure the global convergence: Take the large step lengths k and p k d given in (10) when they satisfy (6), and the common step length k p = k d = k otherwise. If wechoose suciently small positive p and d,we can expect that the large step lengths k p and k d given in (10) usually satisfy (6). Remark 2.2. Kojima, Megiddo, and Noma [5] proposed a continuation method that traces a trajectory leading to a solution of the complementarity problem. If we take a positive less than but close to 1 and small positive p, d, then the set N constitutes a narrow neighborhood of the central trajectory. In this case our infeasible-interior-point algorithm may be regarded as path-following or continuation method, which generates a sequence f(x k y k z k )g tracing the central trajectory in its narrow neighborhood. It makes a main dierence between the Kojima-Megiddo-Noma continuation algorithm and our algorithm that the trajectory traced by their algorithm runs through the outside of the feasible region while the central trajectory traced by our algorithm runs through the interior of the feasible region. See also Kojima, Megiddo, and Mizuno [4] for a more general framework of continuation methods for complementarity problems. Remark 2.3. If we restricted N to the set S, N would turn out to be f(x y z) 2 Q ++ \ S : x 0 z 0 x i z i x T z=n (i =1 2... n)g: This type of neighborhood of the central trajectory relative to S has been utilized in many primal-dual interior-point algorithms ([6, 7, 8, 18, 19, 20, 22, 23], etc.). Specically, it coincides with the one used by Mizuno, Todd, and Ye [18]. The generated sequence f(x k y k z k )g satises the following relations which will be used in the succeeding sections: A(x k + x) ; b =(1; )(Ax k ; b) A T (y k + y)+(z k + z) ; c =(1; )(A T y k + z k ; c) 9 >= > for every 0 (11) (x k+1 ) T z k+1 (1 ; k (1 ; 3 ))(x k ) T z k (x 1 ) T z 1 (12) 9 x k i zi k (x k ) T z k =n (i =1 2... n) >= (x k ) T z k p kax k ; bk or kax k ; bk p (13) (x k ) T z k d ka T y k + z k ; ck or ka T y k + z k ; > ck d ) (x k ) T z +x T z k = ;(1 ; 1 )(x k ) T z k (14) x k i z i +x i zi k = 1 (x k ) T z k =n ; x k i zi k : 6

Here the equalities in (11) follow from the Newton equation (4), the inequality (12) follows from Step 5 of the algorithm (see the last inequality of (6)), the inequalities in (13) follow from (x k y k z k ) 2N, and the equalities in (14) follow from the Newton equation (4) with = 1 (x k ) T z k. 3. Global Convergence In this section we show that the algorithm presented in the previous section terminates at Step 2 in a nite number of iterations for any positive p d and! associated with its stopping criteria (1) and (2). We assume, on the contrary, that the algorithm never stops and derive a contradiction. We rst observe that in addition to (11) (14) the inequalities hold for every k (k =1 2...), where (x k ) T z k and k(x k z k )k 1! (15) = minf p p d d g (16) because otherwise (x k y k z k )would satisfy either of the stopping criteria (1) and (2) for some k. Hence, the entire sequence f(x k y k z k )g lies in the compact set N = f(x y z) 2N : x T z and k(x z)k 1! g: On the other hand, the Newton direction (x y z) determined by the system (4) of equations is a continuous function of the location of (x k y k z k ) 2N. This is easily seen because the coecient matrix on the left hand side of (4) is nonsingular for any (x k y k z k ) 2N, and the coecient matrix as well as the right hand side of the system (4) is continuous in (x k y k z k ) 2N. Therefore the Newton direction (x y z) is uniformly bounded for all (x k y k z k )over the compact set N. Therefore we can nd a positive constant such that the Newton direction (x y z) computed at Step 3 of every iteration satises the inequalities which will be utilized below. j x i z i ; x T z=n j and j x T z j (17) Let k be xed arbitrarily. Dene the real-valued quadratic functions f i (i =1 2... n), g p, g d, and h as follows. f i () =(x k i + x i )(z k i + z i ) ; (x k + x) T (z k + z)=n g p () =(x k + x) T (z k + z) ; p (1 ; )kax k ; bk g d () =(x k + x) T (z k + z) ; d (1 ; )ka T y k ; z k ; ck h() =(1; (1 ; 2 ))(x k ) T z k ; (x k + x) T (z k + z): 7

By (11), we see that the terms coincide with (1 ; )kax k ; bk and (1 ; )ka T y k ; z k ; ck ka(x k + x) ; bk and ka T (y k + y)+(z k + z) ; ck respectively. Hence, we can rewrite the relation (5) to determine the k as f i () 0 (i =1 2... n) g p () 0 or (1 ; )kax k ; bk p g d () 0 or (1 ; )ka T y k ; z k ; ck d h() 0: Remark 3.1. Since all the functions in the inequalities above are linear or quadratic, we can easily compute the value of k by solving them for. We can verify that for every i (i =1 2... n)and 2 [0 1], f i () = x k i z k i + (z k i x i + x k i z i )+ 2 x i z i ; (x k ) T z k + ((z k ) T x +(x k ) T z)+ 2 (x) T z =n = x k i z k i ; (x k i z k i ; 1 (x k ) T z k =n)+ 2 x i z i ; (x k ) T z k ; (1 ; 1 )(x k ) T z k )+ 2 (x) T z =n (by (14)) = x k i z k i (1 ; )+ 1 (x k ) T z k =n + 2 x i z i ; (x k ) T z k (1 ; )+ 1 (x k ) T z k )+ 2 (x) T z =n = (x k i z k i ; (x k ) T z k =n)(1 ; )+ 1 (1 ; )((x k ) T z k =n)) +(x i z i ; x T z=n) 2 1 (1 ; )( =n) ; 2 (by (13), (15) and (17)) Similarly, for every 2 [0 1], g p () 1 ; 2 (1 ; )kax k ; bk p if g p (0) < 0 g d () 1 ; 2 (1 ; )ka T y k + z k ; ck d if g d (0) < 0 h() ( 2 ; 1 ) ; 2 : if g p (0) = (x k ) T z k ; p kax k ; bk0 if g d (0)=(x k ) T z k ; d ka T y k + z k ; ck0 Hence, letting =min ( 1 1 (1 ; ) n 1 ) ( 2 ; 1 ) 8

we obtain that the inequalities f i () 0 (i =1 2... n) g p () 0 if g p (0) 0 (1 ; )kax k ; bk p if g p (0) < 0 g d () 0 if g d (0) 0 (1 ; )ka T y k + z k ; ck d if g d (0) < 0 h() 0 hold for every 2 [0 ]. By the construction of the real-valued functions f i (i = 1 2... n), g p, g d,andh, this can be restated as: the relation (5) holds for every 2 [0 ]. Thus we have shown that the inequality k holds for every k (k =1 2 ). Finally, by the inequality (12) and k, (x k ) T z k (1 ; ( )(1 ; 3 )) k;1 (x 1 ) T z 1 (k =2 3... ): Obviously, the right-hand side of the inequalityconverges to zero as k tends to 1 hence, so does the left-hand side. This contradicts the rst inequality of (15). 4. Detecting Infeasibility {I We showed in the previous section that the algorithm stops at Step 2 in a nite number of iterations for any small >0, p > 0, d > 0, and any large! > 0. If the algorithm stops with the stopping criterion (1), we obtain an approximate optimal solution (x k y k z k ) of the primal-dual pair of P and D. In this section and the next one we will derive information on the infeasibility of the primal-dual pair of P and D when the algorithm stops with the criterion (2). For every pair of nonnegative real numbers and!, dene S(!)=f(x y z) 2 Q + : e x e z and k(x z)k 1!g: We will be concerned with the question whether the region S(!)contains a feasible solution (x y z) of the primal-dual pair of P and D. Theorem 4.1. Take positive numbers,!, and! satisfying (x 1 y 1 z 1 ) 2 S(!) and (!) 2 +(x 1 ) T z 1! : Assume that the algorithm has stopped at Step 2 with the stopping criterion (2), i.e. k(x k z k )k 1 >!. Then, the region S(!) contains no feasible solution (x y z) of the primal-dual pair of P and D. 9

In the remainder of this section we prove Theorem 4.1. It is convenient tointroduce the following primal-dual pair of parametric linear programs with parameters p d 2 [0 1]: P( p d ) Minimize c( d ) T x subject to Ax = b( p ) x 0: D( p d ) Maximize b( p ) T y subject to A T y + z = c( d ) z 0: Here, We obtain from (11) that b() =Ax 1 +(1; )b c() =(A T y 1 + z 1 )+(1; )c: and that Ax k ; b = A T y k + z k ; c = k;1 Y (1 ; j p)(ax 1 ; b) j=1 k;1 Y (1 ; j d)(a T y 1 + z 1 ; c): j=1 These properties were shown in the paper [1]. Geometrically, this implies that the point Ax k ;b lies in the line segment connecting the point Ax 1 ;b and the origin 0 2 R m, and that the point A T y k +z k ;c lies in the line segment connecting the point A T y 1 +z 1 ;c and the origin 0 2 R n. Specically, weknowthat Hence, by dening we have Ax k ; b =0 for every k 1if Ax 1 ; b = 0 A T y k + z k ; c =0 for every k 1if A T y 1 + z k ; c = 0: ( k = kaxk ; bk=kax 1 ; bk if kax 1 ; bk > 0 p 0 if kax 1 ; bk =0 ( k kat y d = k + z k ; ck=ka T y 1 + z 1 ; ck if ka T y 1 + z 1 ; ck > 0 0 if ka T y 1 + z 1 ; ck =0 Ax k ; b = k p (Ax1 ; b) A T y k + z k ; c = k d(a T y 1 + z 1 ; c) for each k. Thus, we can measure the primal and dual infeasibilities of the k'th iterate (x k y k z k ) in terms of k p and k d, respectively. It should be noted that both of the 10 ) (18)

sequences f k pg and f k dg are monotone nonincreasing. The relation (18) together with x k > 0 and z k > 0 implies that (x k y k z k )isaninterior feasible solution of the primaldual pair of P( k p k d) and D( k p k d). We also observe thatif k p =1(or k d = 1 ) holds at a k'th iteration then, for every j (j = k +1 k+2...), x j is a feasible solution of P and j p =0((y j z j ) is a feasible solution of D and j d =0). The following lemma plays an essential role in proving not only Theorem 4.1 but also the theorems in the next section. Lemma 4.2. Suppose that k(x k z k )k 1 >!. Assume that (x y z) is a feasible solution of the primal-dual pair of P( k p k d ) and D(k p k d ) satisfying kxk 1! p kzk 1! d e x e z: (19) Then! p! d +(x 1 ) T z 1 >! : Proof: By (18) and the assumption of the lemma, we see that both (x k y k z k ) and (x y z) are feasible solutions of the primal-dual pair of P(p k k d )andd(k p k d ). It follows that Hence, or Thus we obtain A(x ; x k )=0 and z ; z k = ;A T (y ; y k ):! p! d +(x 1 ) T z 1 x T z +(x k ) T z k (x ; x k ) T (z ; z k )=0 x T z +(x k ) T z k = x T z k +(x k ) T z: (since! p kxk 1,! d kzk 1 and (x 1 ) T z 1 (x k ) T z k ) = x T z k +(x k ) T z e T z k + e T x k (since x e and z e) = k(x k z k )k 1 >! Proof of Theorem 4.1: Assume, on the contrary, that the region S(!)contains a feasible solution (~x ~y ~z) of the primal-dual pair of P and D. Let x =(1; k p)~x + k px 1 (y z) =(1; k d)(~y ~z)+ k d(y 1 z 1 ) and! p =! d =!: 11

Then, we can easily verify that (x y z) is a feasible solution of the primal-dual pair of P( k p k d)andd( k p k d), satisfying all the assumptions in (19) of Lemma 4.2. Hence, by Lemma 4.2, we have (!) 2 +(x 1 ) T z 1 >! which contradicts the assumption of the theorem. This completes the proof. The conclusion of Theorem 4.1 does not necessarily imply the infeasibility of the primal problem P nor the dual problem D. That is, there may exist a feasible solution (x y z) of the primal-dual pair of P and D outside of S( u). Specically, Theorem 4.1 can not be applied to degenerate cases where both problems P and D are feasible but P or D has no interior feasible solution. In such cases, there exists no feasible solution (x y z) 2 S(!) of the primal-dual pair of P and D for any small positive. 5. Detecting Infeasibility {II In order to overcome the shortcomings of Theorem 4.1, we need to somewhat modify the algorithm. The modication is done by replacing Step 3 by Step 3' below. It is designed so that once the primal feasibility errorkax k ; bk becomes less than or equal to the tolerance p (or the dual feasibility error ka T y k + z k ; ck becomes less than or equal to the tolerance p ) at some iteration k, the error will never be improved but maintained from then on kax j ; bk = kax k ; bk if kax k ; bk p and j k ka T y j + z j ; ck = ka T y k + z k ; ck if ka T y k + z k ; ck p and j k: Step 3': Let = 1 (x k ) T z k =n. In the system (4) of equation, replace Ax k ; b by 0 if kax k ;bk p and replace A T y k +z k ;c by 0 if ka T y k +z k ;ck d : Compute the unique solution (x y z) at(x k y k z k ) of the system (4) or its modication mentioned just above whenkax k ; bk p and/or ka T y k + z k ; ck d. It is easily seen that all the relations in (12) through (14) and (18) remain valid for the sequence f(x k y k z k )g generated by the modied algorithm. Hence, each (x k y k z k )is an interior feasible solution of the primal-dual pair of P(p k d)andd( k p k d), k so that Lemma 4.2 also remains valid. The equalities in (11), however, need some modication. For every 0, ( (1 ; A(x k )(Axk ; b) if kax + x) ; b = k ; bk > p Ax k ; b if kax k ; bk p A T (y k + y)+(z k + z) ; c ( (1 ; )(AT y k + z k ; c) if ka T y k + z k ; ck > = d A T y k + z k ; c if ka T y k + z k ; ck d : 12

Furthermore, we can show that the modied algorithm stops at Step 2 in a nite number of iterations. The proof is omitted here it is similar but needs some additional arguments to the proof given in Section 3 for the original algorithm. Let 0 p = 0 d = ( p =kax 1 ; bk if kax 1 ; bk > 0 +1 if kax 1 ; bk =0 ( d =ka T y 1 + z 1 ; ck if ka T y 1 + z 1 ; ck > 0 +1 if ka T y 1 + z 1 ; ck =0: Then, the stopping criterion (1) can be rewritten as (x k ) T z k k p 0 p and k d 0 d: (20) Along the sequence generated by the modied algorithm, we dene k = maxf : e x k e z k g! k = k(x k z k )k 1 = e T x k + e T z k : We now assume that the modied algorithm has stopped at an s'th iteration with satisfying the criterion (2) but not (1). Then k(x s z s )k 1! and (x s ) T z s where denotes the positive constant given in (16), and one of the following four cases occurs: (a) kax s ; bk p and ka T y s + z s ; ck d (i.e., p s 0 and p s d 0 ). d (b) kax s ; bk p and ka T y s + z s ; ck d (i.e., p s 0 p and d s 0 d). (c) kax s ; bk p and ka T y s + z s ; ck d (i.e., p s 0 p and d s 0 d). (d) kax s ; bk p and ka T y s + z s ; ck d (i.e., p s 0 and p s d 0 ). d Let q be the number of the rst iteration such that q p 0 p in the cases (a) and (b), and r the rst iteration number such that r d 0 d in the cases (a) and (c). We rst deal with the case (a). In this case, we have kb( q p) ; bk p Ax k = b( q p) for every k q kc( r d) ; ck d A T y k + z k = c( r d) for every k r: Letting ` =maxfq rg, we see that for every k `, the modied algorithm works as Mizuno-Todd-Ye's interior-point algorithm [18] applied to the primal-dual pair of the 13

linear programs P(p q d)andd( r p q d) r starting from (x` y` z`). Their algorithm is known to reduce the duality gap (x k ) T z k by a constant factor which is independent of the location of the iterate (x k y k z k ). Hence, if we neglect the stopping criterion (2) and continue running the modied algorithm beyond the s'th iteration, we eventually obtain an (x k y k z k ) satisfying the stopping criterion (1) (x k y k z k )gives a desired approximate optimal solution of the primal-dual pair of P and D. In the case (b), we obtain that kb( q p) ; bk p Ax k = b( q p) for every k q ka T y s + z s ; ck = s dka T y 1 + z 1 ; ck > d (i:e: s d >0 d ) The second relation above together with x k > 0 (k q) implies that x k solution of a common primal problem P( q p 0) for every k q. is a feasible Theorem 5.1. Suppose that the case (b) occurs. Let!! 1 and = minf 0 d 1 q g. Assume that!!q! +(x 1 ) T z 1 : Then there isnofeasible solution (~y ~z) of the problem D such that k~zk 1!. Proof: Assume, on the contrary, that there exists a feasible solution (~y ~z) ofthe problem D such that k~zk 1!. Let x = x q (y z) =(1; s d)(~y ~z)+ s d(y 1 z 1 )! p =! q and! d =!: Then, we can easily verify that (x y z) is a feasible solution of the primal-dual pair of P( s p s d)andd( s p s d), satisfying all the assumptions in (19) of Lemma 4.2. Hence,! q! +(x 1 ) T z 1 >! which contradicts the assumption of the theorem. Similarly, we obtain the following result. Theorem 5.2. Suppose that the case (c) occurs. Let!! 1 and = minf 0 p 1 r g. Assume that!!r! +(x 1 ) T z 1 : Then, there isnofeasible solution ~x of the problem P such that k~xk 1!. 14

It should be noted that! q and q in Theorem 5.1 (! r and r in Theorem 5.2) are not known prior to the starting the algorithm. So, to apply Theorems 5.1 or 5.2, we need to adjust the values of! and/or! during the execution of the algorithm. Finally, we consider the case (d). Theorem 5.3. Suppose that the case (d) occurs. Let!! 1 and =minf 0 p 1 0 d 1 g. Assume that Then, the set +(x! (!)2 1 ) T z 1 : S(0!)=f(x y z) 2 Q + : k(x z)k 1!g contains no feasible solution of the primal-dual pair of P and D. Proof: In the case (d), we have kax s ; bk = s pkax 1 ; bk > p (i:e: s p >0 p ) ka T y s + z s ; ck = s dka T y 1 + z 1 ; ck > d (i:e: s d > 0 d). Now assume, on the contrary, that there exists a feasible solution (~x ~y ~z) ofthe primal-dual pair of P and D such thatk(~x ~z)k 1!. Let x =(1; s p )~x + s px 1 (y z) =(1; s d )(~y ~z)+s d (y1 z 1 )! p =! d =!: Then, (x y z) is a feasible solution of the primal-dual pair of P( s p s d) and D( s p s d) satisfying all the assumptions in (19) of Lemma 4.2. Hence, (!) 2 +(x 1 ) T z 1 >! which contradicts the assumption of the theorem. 6. Concluding Remarks. The primal-dual interior-point algorithm for linear programs has been extended to various mathematical programming problems such as convex quadratic programs, convex programs, linear and nonlinear complementarity problems ([4, 6, 8, 10, 20, 21], etc.). It is possible to apply the basic idea of controlling the step length in the way proposed in this paper to infeasible-interior-point algorithms for such problems. Among others, we can easily modify the infeasible-interior-point algorithm using the step length control rule, which has been described in Section 2, so as to adapt it to the linear complementarity problem with an n n positive semi-denite matrix M: nd (x z) 2 R 2n such 15

that (x z) 0 z = Mx + q and x T z = 0. In fact, we dene the central trajectory (Megiddo [16]) as the set of solutions (x z) > 0 to the system of equations z = Mx+ q and Xz = e for every >0, the Newton direction at a k'th iterate (x k z k ) > 0 as the unique solution (x z) of the system of linear equations z ; Mx = ;z k + Mx k + q Z k x + X k z = ;X k z k +( 1 (x k ) T z k =n)e and the neighborhood N of the central trajectory as N = f(x z) > 0 : x i z i x T z=n (i =1 2... n) x T z 0 kz ; Mx; qk or kz ; Mx; qk 0 g where 0 < 1 < 1 0 <<1 0 < 0 and 0 < 0. Starting from an arbitrary point (x 1 z 1 )inn, the algorithm iteratively generates a new point (x k+1 z k+1 ) such that (x k+1 z k+1 )=(x k z k )+ k (x z) 2N (x k+1 ) T z k+1 (1 ; k (1 ; 3 ))(x k ) T z k where k be the maximum of ~'s 1such that the relations (x k z k )+(x z) 2N (x k + x) T (z k + z) (1 ; (1 ; 2 ))(x k ) T z k and 0 < 1 < 2 < 3 < 1. We could show similar results to the ones stated in Sections 3, 4 and 5, but the details are omitted. We called an infeasible-interior-point algorithm \an exterior point algorithm" in the original manuscript. But many people havepointed out to us that the terminology \exterior" is inappropriate because the algorithm still generates a sequence of points within the interior of the region determined by the nonnegativity constraints and an exterior point algorithm usually means an algorithm that relaxes the nonnegativity constrains (see, for example, [3]). Y. Zhang proposed to use the new name \an infeasible interior-point algorithm." But it may give an impression \an infeasible algorithm" to the readers. To avoid such an impression, we have modied it to \an infeasible-interior-point algorithm." References 16

[1] T. J. Carpenter, I. J. Lustig, J. Mulvey, and D. F. Shanno, \Higher order predictorcorrector interior point methods with applications to quadratic objectives," em SIAM Journal on Optimization to appear. [2] I. C. Choi, C. L. Monma, and D. F. Shanno, \Further development of a primal-dual interior point method," ORSA Journal on Computing 2 (1990) 304-311. [3] A. V. Fiacco and G. P. McCormick, Nonlinear Programming: Sequential Unconstrained Minimization Technique John Wiley and Sons, New York, 1968. [4] M. Kojima, N. Megiddo, and S. Mizuno, \A general framework of continuation methods for complementarity problems," Research Report RJ 7720, IBM Almaden Research Center, San Jose, California 95120, 1990). [5] M. Kojima, N. Megiddo, and T. Noma, \Homotopy continuation methods for nonlinear complementarity problems," Mathematic of Operations Research, to appear. [6] M. Kojima, N. Megiddo, T. Noma, and A. Yoshise, A Unied Approach to Interior Point Algorithms for Linear Complementarity Problems, Lecture Notes in Computer Science 538, Springer-Verlag, New York, 1991. [7] M. Kojima, S. Mizuno, and A. Yoshise, \A primal-dual interior point algorithm for linear programming," In N. Megiddo, ed., Progress in Mathematical Programming, Interior-Point and Related Methods, Springer-Verlag, New York, 1989, pp. 29{47. [8] M. Kojima, S. Mizuno, and A. Yoshise, \A polynomial-time algorithm for a class of linear complementary problems," Mathematical Programming 44 (1989) 1{26. [9] M. Kojima, S. Mizuno, and A. Yoshise, \A little theorem of the big M in interior point algorithms," Research Report on Information Sciences B-239, Department of Information Sciences, Tokyo Institute of Technology, Oh-Okayama, Meguro-ku, Tokyo 152, Japan, Feb. 1991, Revised Sep. 1991. [10] K. O. Kortanek, F. Potra, and Y. Ye, \On some ecient interior point methods for nonlinear convex programming," Linear Algebra and Its Applications 152 (1991) 169{189. [11] I. J. Lustig, \A generic primal-dual interior point algorithm," Technical Report SOR 88-3, Program in Statistics and Operations Research, Department of Civil Engineering and Operations Research, School of Engineering and Applied Science, Princeton University, Princeton, New Jersey, 08544, 1988. [12] I. J. Lustig, \Feasibility issues in a primal-dual interior-point method for linear programming," Mathematical Programming 49 (1990/91) 145{162. [13] I. J. Lustig, R. E. Marsten, and D. F. Shanno, \Computational experience with a primal-dual interior point method for linear programming," Linear Algebra and Its Applications 152 (1991) 191{222. [14] R. Marsten, R. Subramanian, M. Saltzman, I Lustig, and D. Shanno, \Interior point methods for linear programming: Just call Newton, Lagrange, and Fiacco and McCormick!," Interfaces 20 (1990) 105{116. [15] K. A. McShane, C. L. Monma, and D. F. Shanno, \An implementation of a primaldual interior point method for linear programming," ORSA Journal on computing 17

1 (1989) 70{83. [16] N. Megiddo, \Pathways to the optimal set in linear programming," In N. Megiddo, ed., Progress in Mathematical Programming, Interior-Point and Related Methods (Springer-Verlag, New York, 1989) pp. 131{158. [17] S. Mehrotra, \On the implementation of a (primal-dual) interior point method," Technical Report 90-03, Department of Industrial Engineering and Management Sciences Northwestern University (Evanston, Illinois, 1990). [18] S. Mizuno, M. J. Todd, and Y. Ye, \On adaptive-step primal-dual interior-point algorithms for linear programming," Technical Report No. 944, School of Operations Research and Industrial Engineering, College of Engineering, Cornell University (Ithaca, New York, 14853-3801, 1990). [19] R. D. C. Monteiro and I. Adler, \Interior path following primal-dual algorithms. Part I: linear programming," Mathematical Programming 44 (1989) 27{41. [20] R. D. C. Monteiro and I. Adler, \Interior path following primal-dual algorithms. Part II: convex quadratic programming," Mathematical Programming 44 (1989) 43{66. [21] T. Noma, A globally convergent iterative algorithm for complementarity problems : A modication of interior point algorithms for linear complementarity problems. PhD thesis, Dept. of Systems Science, Tokyo Institute of Technology, 2{12{1 Oh{ Okayama, Meguro{ku, Tokyo 152, Japan, 1991. [22] K. Tanabe, \Centered Newton method for mathematical programming," In M. Iri and K. Yajima, eds., System Modeling and Optimization, Springer-Verlag, New York, 1988, pp. 197{206. [23] K. Tanabe, \Centered newton method for linear programming: Interior and `exterior' point method' (in Japanese)," In K. Tone, ed., New Methods for Linear Programming 3 The Institute of Statistical Mathematics, 4-6-7 Minami-Azabu, Minatoku, Tokyo 106, Japan, 1990, pp. 98{100. [24] R. J. Vanderbei and T. J. Carpenter, \Symmetric indenite systems for interiorpoint methods," SOR-91-7, School of Engineering and Applied Science Department of Civil Engineering and Operations Research, Princeton University, Princeton, New Jersey 08544, 1991. [25] Y. Ye, \On the Q-order of convergence of interior-point algorithms for linear programming," Working Paper, Department of Management Sciences, The University of Iowa, Iowa City, Iowa 53342, 1991. [26] Y. Ye, O. Guler, R. A. Tapia, and Y. Zhang, \A quadratically convergent O( p nl)- iteration algorithm for linear programming," TR91-26, Department of Mathematical Sciences, Rice University, Houston, Texas 77251, 1991. 18