GYULA FARKAS WOULD ALSO FEEL PROUD

Similar documents
On Hoffman s celebrated cycling LP example

Updating and downdating an upper trapezoidal sparse orthogonal factorization

Linear Programming: Simplex

Introduction to optimization

Answers to problems. Chapter 1. Chapter (0, 0) (3.5,0) (0,4.5) (2, 3) 2.1(a) Last tableau. (b) Last tableau /2 -3/ /4 3/4 1/4 2.

1 date: February 23, 1998 le: papar1. coecient pivoting rule. a particular form of the simplex algorithm.

SOME GENERALIZATIONS OF THE CRISS-CROSS METHOD. Emil Klafszky Tamas Terlaky 1. Mathematical Institut, Dept. of Op. Res.

Note 3: LP Duality. If the primal problem (P) in the canonical form is min Z = n (1) then the dual problem (D) in the canonical form is max W = m (2)

New Artificial-Free Phase 1 Simplex Method

Chapter 3, Operations Research (OR)

2.098/6.255/ Optimization Methods Practice True/False Questions

4.5 Simplex method. LP in standard form: min z = c T x s.t. Ax = b

3. THE SIMPLEX ALGORITHM

LP Duality: outline. Duality theory for Linear Programming. alternatives. optimization I Idea: polyhedra

Lecture: Algorithms for LP, SOCP and SDP

Week 2. The Simplex method was developed by Dantzig in the late 40-ties.

Part 4: Active-set methods for linearly constrained optimization. Nick Gould (RAL)

4. Duality Duality 4.1 Duality of LPs and the duality theorem. min c T x x R n, c R n. s.t. ai Tx = b i i M a i R n

Part 1. The Review of Linear Programming

The Simplex Algorithm

A primal-simplex based Tardos algorithm

An Infeasible Interior-Point Algorithm with full-newton Step for Linear Optimization

CS261: A Second Course in Algorithms Lecture #9: Linear Programming Duality (Part 2)

Introduction to Mathematical Programming IE406. Lecture 10. Dr. Ted Ralphs

Input: System of inequalities or equalities over the reals R. Output: Value for variables that minimizes cost function

Today: Linear Programming (con t.)

The Primal-Dual Algorithm P&S Chapter 5 Last Revised October 30, 2006

Midterm Review. Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A.

A NEW FACE METHOD FOR LINEAR PROGRAMMING. 1. Introduction We are concerned with the standard linear programming (LP) problem

Chapter 1. Preliminaries

Discrete Optimization

min 4x 1 5x 2 + 3x 3 s.t. x 1 + 2x 2 + x 3 = 10 x 1 x 2 6 x 1 + 3x 2 + x 3 14

arxiv:cs/ v1 [cs.ms] 13 Aug 2004

BBM402-Lecture 20: LP Duality

Yinyu Ye, MS&E, Stanford MS&E310 Lecture Note #06. The Simplex Method

Summary of the simplex method

Appendix A Taylor Approximations and Definite Matrices

A SUFFICIENTLY EXACT INEXACT NEWTON STEP BASED ON REUSING MATRIX INFORMATION

Applications. Stephen J. Stoyan, Maged M. Dessouky*, and Xiaoqing Wang

On the projection onto a finitely generated cone

Multicommodity Flows and Column Generation

A Strongly Polynomial Simplex Method for Totally Unimodular LP

FRACTIONAL PACKING OF T-JOINS. 1. Introduction

CO350 Linear Programming Chapter 8: Degeneracy and Finite Termination

1 Review Session. 1.1 Lecture 2

Block Bidiagonal Decomposition and Least Squares Problems

LINEAR PROGRAMMING ALGORITHMS USING LEAST-SQUARES METHOD

Primal-dual relationship between Levenberg-Marquardt and central trajectories for linearly constrained convex optimization

CO 250 Final Exam Guide

Math 273a: Optimization The Simplex method

A Polynomial Column-wise Rescaling von Neumann Algorithm

Criss-cross Method for Solving the Fuzzy Linear Complementarity Problem

1. Algebraic and geometric treatments Consider an LP problem in the standard form. x 0. Solutions to the system of linear equations

Robust Farkas Lemma for Uncertain Linear Systems with Applications

A full-newton step infeasible interior-point algorithm for linear programming based on a kernel function

ON THE PARAMETRIC LCP: A HISTORICAL PERSPECTIVE

Algorithmic Game Theory and Applications. Lecture 7: The LP Duality Theorem

Linear Programming: Chapter 5 Duality

A Simpler and Tighter Redundant Klee-Minty Construction

Math 5593 Linear Programming Week 1

Summary of the simplex method

ORIE 6300 Mathematical Programming I August 25, Lecture 2

Lecture 10: Linear programming duality and sensitivity 0-0

4.5 Simplex method. min z = c T x s.v. Ax = b. LP in standard form

COMPUTATIONAL COMPLEXITY OF PARAMETRIC LINEAR PROGRAMMING +

Linear Programming Redux

Introduction to Linear and Combinatorial Optimization (ADM I)

MS&E 318 (CME 338) Large-Scale Numerical Optimization

Basis Pursuit Denoising and the Dantzig Selector

A PRIMAL-DUAL INTERIOR POINT ALGORITHM FOR CONVEX QUADRATIC PROGRAMS. 1. Introduction Consider the quadratic program (PQ) in standard format:

Lecture slides by Kevin Wayne

9.1 Linear Programs in canonical form

4TE3/6TE3. Algorithms for. Continuous Optimization

LP. Lecture 3. Chapter 3: degeneracy. degeneracy example cycling the lexicographic method other pivot rules the fundamental theorem of LP

LINEAR AND NONLINEAR PROGRAMMING

MS&E 318 (CME 338) Large-Scale Numerical Optimization

A Second Full-Newton Step O(n) Infeasible Interior-Point Algorithm for Linear Optimization

Linear Programming. Linear Programming I. Lecture 1. Linear Programming. Linear Programming

Linear Programming. Chapter Introduction

Nonlinear Optimization: What s important?

1 Overview. 2 Extreme Points. AM 221: Advanced Optimization Spring 2016

::::: OFTECHY. .0D 0 ::: ::_ I;. :.!:: t;0i f::t l. :- - :.. :?:: : ;. :--- :-.-i. .. r : : a o er -,:I :,--:-':: : :.:

An interior-point gradient method for large-scale totally nonnegative least squares problems

Scholars Research Library

The simplex algorithm

Absolute Value Programming

AM 121: Intro to Optimization

Ω R n is called the constraint set or feasible set. x 1

LINEAR PROGRAMMING I. a refreshing example standard form fundamental questions geometry linear algebra simplex algorithm

Example Problem. Linear Program (standard form) CSCI5654 (Linear Programming, Fall 2013) Lecture-7. Duality

Numerical Optimization

Farkas Lemma, Dual Simplex and Sensitivity Analysis

3 The Simplex Method. 3.1 Basic Solutions

Contents. Introduction

Linear Programming using NNLS

Formulation of L 1 Norm Minimization in Gauss-Markov Models

Developing an Algorithm for LP Preamble to Section 3 (Simplex Method)

Polynomiality of Linear Programming

Review Solutions, Exam 2, Operations Research

CSCI 1951-G Optimization Methods in Finance Part 01: Linear Programming

Transcription:

GYULA FARKAS WOULD ALSO FEEL PROUD PABLO GUERRERO-GARCÍA a, ÁNGEL SANTOS-PALOMO a a Department of Applied Mathematics, University of Málaga, 29071 Málaga, Spain (Received 26 September 2003; In final form 26 September 2003) Starting from the fact that the 1974 version of the NNLS algorithm of Lawson and Hanson does indeed constitute by itself a new constructive proof of Farkas lemma (in primal and dual form), we apply it to solve the restricted feasibility subproblems that occur in two primal-dual active-set linear programming algorithms, thus establishing the relationship with a recently proposed (dated in October, 2002) primal-dual active-set scheme. As a result, the proposed algorithms are impervious to degeneracy, both in Phase I and in Phase II, and finite in exact arithmetic without having to resort to any anti-cycling pivoting rule, and hence they are by themselves constructive proofs of the strong duality theorem. Finally, we describe the sparse implementation of both algorithms in terms of certain sparse Cholesky factorization. Keywords: Linear programming; Non-negative least squares; Primal-dual active-set method; Farkas lemma; Degeneracy Technical Report, Dept. Applied Mathematics, Univ. Málaga, November 2002 Corresponding author. Email addresses: pablito@ctima.uma.es (PABLO GUERRERO-GARCÍA), santos@ctima.uma.es (ÁNGEL SANTOS-PALOMO). Preprint submitted to Optimization Methods and Software 13 November 2002

1 CONSTRUCTIVE PROOF OF FARKAS LEMMA IMPERVIOUS TO DEGENERACY Let us consider the usual unsymmetric primal-dual pair of linear programs using a non-standard notation (we have deliberately exchanged the usual roles of b and c, x and y, n and m, and (P ) and (D), as in e.g. [16, 2]): (P ) min l(x). = c T x, x R n (D) max L(y). = b T y, y R m s.t A T x b s.t Ay = c, y O where A R n m with m n and rank(a) = n. We denote with F and G the feasible region of (P ) and (D), respectively. In his elementary proof of Farkas lemma, or y R m : Ay = c, y O, or else d R n : c T d < 0, A T d O, Dax (cf. [8] and also [7]) links the dual feasibility problem find y R m : Ay = c, y O (1) with the NNLS convex problem min{ Ay c 2 : y O}. (2) He gave a particular method to deal with this problem, in which he had to manage the possibility of degeneracy. Indeed, as was pointed out in [3], this possibility seems to be an intrinsic part of the problem. Our proposal is to apply the NNLS algorithm of Lawson and Hanson [13, 23.3] directly to this problem, because in exact arithmetic the algorithm is finite and so we do not have to resort to anti- 2

cycling rules to prove Farkas lemma. Note that problem (2) usually appears in the literature in stricly convex form, but a careful view of the proof of the finiteness in [13, 23.3] does not rule out the convex case. We obtain a dual feasible point (in particular, a point y =. [y B ; O] such that A B y B = c, y B > O and A B has full column rank, so y is a basic feasible solution of (1)), or else we obtain a descent direction d of F. Note that the number of columns in A B is less than or equal to n. Gale s theorem is equivalent to the dual form of Farkas lemma (e.g., [20]), or x R n : A T x b, or else δ R m : b T δ > 0, Aδ = O, δ O. Our favourite constructive proof consists on linking the primal feasibility problem find x R n : A T x b (3) with the least distance problem min{ x 2 : A T x b}. (4) We then proceed with Cline s method as in [13, 23.4], solving b T 1 min δ A O 2 : δ O (5) via NNLS instead of (4), so we also escape from degeneracy. We obtain an x F (i.e., a solution of (3)), or else we obtain an ascent direction δ of G. 3

2 THE PROPOSED PRIMAL-DUAL ACTIVE-SET ALGORITHM The well-known primal-dual active-set method of [5] can be applied to our primaldual pair if we have a primal feasible point x available. The solution procedure proceeds by solving a sequence of restricted dual feasibility subproblems, namely find y R m : A A y A = c, y A O, (6) where A A is constructed with the columns of A corresponding to active primal constraints in x. Note that A A can have more than n columns when x is a degenerate primal vertex. We do not reproduce the details here: the interested reader can consult [14, pp. 263 265]. The original proposal was to solve the restricted feasibility subproblem (6) by constructing a Phase I problem min{e T t : A A y A + t = c, y A O, t O} and solving it by the simplex method with t R n as the initial artificial set of basic variables. The main point is that we have to resort to anti-cycling rules as the inefficient Bland s one to ensure the convergence of the overall procedure. Several authors (e.g., [11], [12], [10]) have proposed to solve the Phase I subproblem using a criss-cross variant, whose efficiency is an open question [11, 7]. Once again, Dax [7] was the first one to link the restricted feasibility subproblem (6) with the NNLS convex subproblem min{ A A y A c 2 : y A O}. (7) 4

He gave a particular method to deal with this subproblem, in which he had to manage the possibility of degeneracy; this method is known as Dax s steepest-descent primal-dual active-set algorithm since [9]. Our proposal is to resort to our favourite proof of Farkas lemma and then to apply the NNLS algorithm directly to subproblem (7), obtaining an optimal y A O or else a descent direction d such that c T d < 0 and A T Ad O; such a direction maintains primal feasibility and strictly decreases the primal objective function l. This is the algorithm proposed in [1], but Dantzig et al. [6] apply NNLS to (2) nine years ago and claim that several NNLS variants can be applied to solve the Phase II linear program efficiently by means of a sequence of improving approximations to the optimal value of the objective function. Note that in [6] no explicit Farkas connection was given. But what about the computation of the initial primal feasible x? Dax [7] proposes to apply a similar algorithm with a modified piecewise linear objective function with the sum of primal infeasibilities, and [1] refer to [17, 5] where the classical Beale s device (also in [5]) modifying the dual feasible region is described. Note that both proposed Phase I s modify the original problem in a somewhat anti-natural sense. Our proposal is to apply our constructive proof of the dual form of Farkas lemma to find the initial x F. As a result, we can detect if (D) is unbounded above during Phase I with the dual form of Farkas lemma. If not, we solve a sequence of restricted dual feasibility subproblems with the primal form of Farkas lemma, using the 1974 version of NNLS all the time to do not have to worry about degeneracy in exact arithmetic. This proposal is by itself a constructive proof of the strong duality theorem (e.g., [20]), also known as the existence-duality theorem. Indeed, it is related to the im- 5

plicit one given by Klafszky and Terlaky (cf. [11], but also the older ones [7] and [14, pp. 261 262]). We think that the first proposed criss-cross-based implicit proof of Farkas lemma was done in [11] (using Bland s least index rule), although they made it explicit later in [12]. A new criss-cross-based proof of Farkas lemma (using two different anti-cycling rules) was given recently in [10], but their proof of the strong duality theorem does not rely upon the hungarian primal-dual connection. 3 THE PROPOSED DUAL-PRIMAL ACTIVE-SET ALGORITHM But we can also see the problem from the dual perspective (and hence obtain an alternative proof of the strong duality theorem), namely to apply our constructive proof of the primal form of Farkas lemma to find an initial y G. As a result, we can detect if (P ) is unbounded below during Phase I with the primal form of Farkas lemma. If not, we solve a sequence of restricted primal feasibility subproblems using the dual form of Farkas lemma. To the best of our knowledge, the first author that proposed to solve a sequence of restricted primal feasibility subproblems was Murty [14, pp. 261 262], and so we gave this method the name Murty s steepest-ascent dual-primal active-set algorithm. Although he pointed out in his book that it was described in [5], they indeed only included in 1956 the description of the above section for a linear program in standard form. The solution procedure has to compute first a y G and then proceeds by solving a sequence of restricted primal feasibility subproblems, namely find x R n : A T Ax = b A, A T Nx b N, (8) where A A is constructed with the columns of A corresponding to positive compo- 6

nents of y (and A N with those to zero components). Note that A N can have more than m n columns when y is a degenerate dual vertex. The original proposal was to solve this restricted primal feasibility subproblem by a Phase I for the primal simplex method. But since the dual of (8) is find δ R m : b T δ > 0, Aδ = O, δ N O, (9) our proposal is to link it with b T 1 min δ A O 2 : δ N O, (10) obtaining an optimal x or else an ascent direction δ such that it maintains dual feasibility and strictly increases the dual objective function L. Note that we cannot solve this problem with the 1974 version of the NNLS algorithm, because not all the components of δ R m are restricted to be non-negative. Nevertheless, Dantzig et al. [6, 3.1] describe the minor changes necessary to allow NNLS to deal with this possibility, and they do not compromise the imperviousity to degeneracy. 4 SPARSE IMPLEMENTATION: THE CHOLESKY CONNECTION The necessity to obtain least-squares solutions occurring in the intermediate steps of the NNLS algorithm lead us to work with the QR factorization of A k, where A k is a full column rank matrix with a subset of columns of the fixed matrix A. This is the implementation chosen in [7], in [6] and in [1] when they are faced with a dense problem. 7

When the problem is sparse, iterative relaxation techniques are recommended in [7, p. 321] to solve the restricted dual feasibility subproblem, but in this way we do not take advantage of the existing relationship between A k and A k+1, where A k+1 is the resulting full column rank matrix after adding (or deleting) columns of A to (or from) A k. Both in [6] and [1], a sparse NNLS implementation using QR update techniques is asked for. A suitable sparse approach is to adapt the methodology of Björck [2] and Oreborn [15] to be able to apply the sparse NNLS algorithm with a short-and-fat matrix A. They proposed an active set algorithm for the sparse least squares problem minimize 1/2 y T Cy + d T y, y R m subject to l y u with C > 0. In our problem C = A T A d = A T c i 1: m, l i = 0 u i = + but C 0, so to maintain a sparse QR factorization of A k we have had to adapt the proposed static technique as in [4], but without forming C. Details can be found in [18]. We do not maintain the orthonormal factor in the sparse QR factorization of A k ; instead, we take advantage of the fact that its triangular factor is the Cholesky factor of A T k A k, and we use corrected seminormal equations to solve the least squares subproblems with the desired accuracy. As was claimed by Saunders [19], major Cholesky would feel proud, but we think that Gyula Farkas would also feel proud! Acknowledgements The authors thank Michael Saunders at Stanford who, with respect to the techni- 8

cal report prior to [6], commented us: Their method is equivalent to the known (and very simple) NNLS algorithm for non-negative least squares. Their real contribution was to apply it to linear programs to find a feasible solution. We also thank Åke Björck at Linköping and Achiya Dax at Hydrological Service for providing us the references [2] and [7] that we were not aware of, and John Gilbert at Xerox for confirming us that the sparse orthogonal factorization in MATLAB 5 is row-oriented, Givens-based and providing us technical details about the interface. References [1] E. Barnes, V. Chen, B. Gopalakrishnan, and E. L. Johnson. A least-squares primal-dual algorithm for solving linear programming problems. Oper. Res. Lett., 30:289 294, October 2002. [2] Å. Björck. A direct method for sparse least squares problems with lower and upper bounds. Numer. Math., 54:19 32, 1988. [3] C. G. Broyden. A simple algebraic proof of Farkas s lemma and related theorems. Optim. Meth. Soft., 8(3 4):185 199, 1998. [4] T. F. Coleman and L. A. Hulbert. A direct active set algorithm for large sparse quadratic programs with simple lower bounds. Math. Progr., 45:373 406, 1989. [5] G. B. Dantzig, L. R. Ford, and D. R. Fulkerson. A primal-dual algorithm for linear programs. In H. W. Kuhn and A. W. Tucker, editors, Linear Inequalities and Related Systems, number 38 in Annals of Mathematics Studies, pages 171 181. Princeton University Press, Princeton, NJ, USA, 1956. [6] G. B. Dantzig, S. A. Leichner, and J. W. Davis. A strictly improving linear programming Phase I algorithm. Ann. Oper. Res., 46/47:409 430, 1993. [7] A. Dax. Linear programming via least squares. Linear Algebra Appl., 111:313 324, 9

1988. [8] A. Dax. An elementary proof of Farkas lemma. SIAM Rev., 39(3):503 507, September 1997. [9] K. George and M. R. Osborne. On degeneracy in linear programming and related problems. Ann. Oper. Res., 46/47:343 359, 1993. [10] T. Illés and K. Mészáros. A new and constructive proof of two basic results of linear programming. Yugosl. J. Oper. Res., 11:15 31, 2001. [11] E. Klafszky and T. Terlaky. Variants of the Hungarian method for solving linear programming problems. Math. Oper. und Stat. ser. Optimization, 20(1):79 91, 1989. [12] E. Klafszky and T. Terlaky. The role of pivoting in proving some fundamental theorems of linear algebra. Linear Algebra Appl., 151:97 118, 1991. [13] C. L. Lawson and R. J. Hanson. Solving Least Squares Problems. Original edition, Prentice-Hall, Englewood Cliffs, NJ, USA, 1974. [14] K. G. Murty. Linear Programming. John Wiley and Sons, New York, NY, USA, 1983. [15] U. Oreborn. A Direct Method for Sparse Nonnegative Least Squares Problems. Licentiat thesis, Department of Mathematics, Linköping University, Sweden, 1986. [16] M. R. Osborne. Finite Algorithms in Optimization and Data Analysis. John Wiley and Sons, Chichester, 1985. [17] C. H. Papadimitriou and K. Steiglitz. Combinatorial Optimization: Algorithms and Complexity. Original edition, Prentice-Hall, Englewood Cliffs, NJ, USA, 1982. [18] Á. Santos-Palomo and P. Guerrero-García. Solving a sequence of sparse least squares problems. Technical report, Department of Applied Mathematics, University of Málaga, September 2001. Submitted for publication to BIT. [19] M. A. Saunders. Major Cholesky would feel proud. ORSA J. Computing, 6(1):23 27, 1994. 10

[20] T. Terlaky and C. Roos. Linear programming. In P. M. Pardalos and M. G. C. Resende, editors, Handbook of Applied Optimization, pages 1 39. Oxford University Press, New York, NY, USA, 2002. 11