PRIMAL-DUAL ENTROPY-BASED INTERIOR-POINT ALGORITHMS FOR LINEAR OPTIMIZATION

Similar documents
A full-newton step infeasible interior-point algorithm for linear programming based on a kernel function

McMaster University. Advanced Optimization Laboratory. Title: Computational Experience with Self-Regular Based Interior Point Methods

A Second Full-Newton Step O(n) Infeasible Interior-Point Algorithm for Linear Optimization

A full-newton step feasible interior-point algorithm for P (κ)-lcp based on a new search direction

Research Note. A New Infeasible Interior-Point Algorithm with Full Nesterov-Todd Step for Semi-Definite Optimization

An Infeasible Interior-Point Algorithm with full-newton Step for Linear Optimization

A FULL-NEWTON STEP INFEASIBLE-INTERIOR-POINT ALGORITHM COMPLEMENTARITY PROBLEMS

A PREDICTOR-CORRECTOR PATH-FOLLOWING ALGORITHM FOR SYMMETRIC OPTIMIZATION BASED ON DARVAY'S TECHNIQUE

A new primal-dual path-following method for convex quadratic programming

A path following interior-point algorithm for semidefinite optimization problem based on new kernel function. djeffal

Primal-Dual Interior-Point Methods for Linear Programming based on Newton s Method

Enlarging neighborhoods of interior-point algorithms for linear programming via least values of proximity measure functions

On Mehrotra-Type Predictor-Corrector Algorithms

On Mehrotra-Type Predictor-Corrector Algorithms

A Full Newton Step Infeasible Interior Point Algorithm for Linear Optimization

A PRIMAL-DUAL INTERIOR POINT ALGORITHM FOR CONVEX QUADRATIC PROGRAMS. 1. Introduction Consider the quadratic program (PQ) in standard format:

Lecture 5. Theorems of Alternatives and Self-Dual Embedding

Interior Point Methods in Mathematical Programming

Improved Full-Newton Step O(nL) Infeasible Interior-Point Method for Linear Optimization

Interior-Point Methods for Linear Optimization

CCO Commun. Comb. Optim.

A Generalized Homogeneous and Self-Dual Algorithm. for Linear Programming. February 1994 (revised December 1994)

A full-newton step infeasible interior-point algorithm for linear complementarity problems based on a kernel function

Local Self-concordance of Barrier Functions Based on Kernel-functions

Interior Point Methods for Mathematical Programming

A NEW PROXIMITY FUNCTION GENERATING THE BEST KNOWN ITERATION BOUNDS FOR BOTH LARGE-UPDATE AND SMALL-UPDATE INTERIOR-POINT METHODS

APPROXIMATING THE COMPLEXITY MEASURE OF. Levent Tuncel. November 10, C&O Research Report: 98{51. Abstract

Lecture 15 Newton Method and Self-Concordance. October 23, 2008

Full Newton step polynomial time methods for LO based on locally self concordant barrier functions

Interior-point algorithm for linear optimization based on a new trigonometric kernel function

PREDICTOR-CORRECTOR SMOOTHING METHODS FOR LINEAR PROGRAMS WITH A MORE FLEXIBLE UPDATE OF THE SMOOTHING PARAMETER 1

On Generalized Primal-Dual Interior-Point Methods with Non-uniform Complementarity Perturbations for Quadratic Programming

On self-concordant barriers for generalized power cones

Improved Full-Newton-Step Infeasible Interior- Point Method for Linear Complementarity Problems

12. Interior-point methods

IMPLEMENTATION OF INTERIOR POINT METHODS

On Two Measures of Problem Instance Complexity and their Correlation with the Performance of SeDuMi on Second-Order Cone Problems

Interior-Point Methods

Research overview. Seminar September 4, Lehigh University Department of Industrial & Systems Engineering. Research overview.

New stopping criteria for detecting infeasibility in conic optimization

A New Class of Polynomial Primal-Dual Methods for Linear and Semidefinite Optimization

A Full-Newton Step O(n) Infeasible Interior-Point Algorithm for Linear Optimization

Primal-dual IPM with Asymmetric Barrier

An interior-point gradient method for large-scale totally nonnegative least squares problems

A Simpler and Tighter Redundant Klee-Minty Construction

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44

A semidefinite relaxation scheme for quadratically constrained quadratic problems with an additional linear constraint

arxiv: v1 [math.oc] 21 Jan 2019

A WIDE NEIGHBORHOOD PRIMAL-DUAL INTERIOR-POINT ALGORITHM WITH ARC-SEARCH FOR LINEAR COMPLEMENTARITY PROBLEMS 1. INTRODUCTION

Nonlinear Optimization for Optimal Control

arxiv:math/ v1 [math.co] 23 May 2000

Linear & nonlinear classifiers

Nonsymmetric potential-reduction methods for general cones

w Kluwer Academic Publishers Boston/Dordrecht/London HANDBOOK OF SEMIDEFINITE PROGRAMMING Theory, Algorithms, and Applications

Using Schur Complement Theorem to prove convexity of some SOC-functions

A strongly polynomial algorithm for linear systems having a binary solution

Computational Experience with Rigorous Error Bounds for the Netlib Linear Programming Library

15. Conic optimization

A priori bounds on the condition numbers in interior-point methods

Lecture 9 Sequential unconstrained minimization

Lecture 1. 1 Conic programming. MA 796S: Convex Optimization and Interior Point Methods October 8, Consider the conic program. min.

Largest dual ellipsoids inscribed in dual cones

Interior Point Methods. We ll discuss linear programming first, followed by three nonlinear problems. Algorithms for Linear Programming Problems

Barrier Method. Javier Peña Convex Optimization /36-725

On implementing a primal-dual interior-point method for conic quadratic optimization

Semidefinite Programming

A Redundant Klee-Minty Construction with All the Redundant Constraints Touching the Feasible Region

Spring 2017 CO 250 Course Notes TABLE OF CONTENTS. richardwu.ca. CO 250 Course Notes. Introduction to Optimization

Lagrangian-Conic Relaxations, Part I: A Unified Framework and Its Applications to Quadratic Optimization Problems

Some Properties of the Augmented Lagrangian in Cone Constrained Optimization

On Superlinear Convergence of Infeasible Interior-Point Algorithms for Linearly Constrained Convex Programs *

Duality Theory of Constrained Optimization

Lecture 8 Plus properties, merit functions and gap functions. September 28, 2008


Following The Central Trajectory Using The Monomial Method Rather Than Newton's Method

minimize x subject to (x 2)(x 4) u,

c 2000 Society for Industrial and Applied Mathematics

Limiting behavior of the central path in semidefinite optimization

An O(nL) Infeasible-Interior-Point Algorithm for Linear Programming arxiv: v2 [math.oc] 29 Jun 2015

IMPLEMENTING THE NEW SELF-REGULAR PROXIMITY BASED IPMS

10-725/ Optimization Midterm Exam

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

A tight iteration-complexity upper bound for the MTY predictor-corrector algorithm via redundant Klee-Minty cubes

Selected Examples of CONIC DUALITY AT WORK Robust Linear Optimization Synthesis of Linear Controllers Matrix Cube Theorem A.

Solving Obstacle Problems by Using a New Interior Point Algorithm. Abstract

CO 250 Final Exam Guide

AN EFFICIENT APPROACH TO UPDATING SIMPLEX MULTIPLIERS IN THE SIMPLEX ALGORITHM

A Second-Order Path-Following Algorithm for Unconstrained Convex Optimization

Optimization methods

CSC Linear Programming and Combinatorial Optimization Lecture 10: Semidefinite Programming

Chapter 1. Preliminaries

2. Dual space is essential for the concept of gradient which, in turn, leads to the variational analysis of Lagrange multipliers.

AN INTERIOR POINT METHOD, BASED ON RANK-ONE UPDATES, Jos F. Sturm 1 and Shuzhong Zhang 2. Erasmus University Rotterdam ABSTRACT

Lecture 24: August 28

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

4TE3/6TE3. Algorithms for. Continuous Optimization

We describe the generalization of Hazan s algorithm for symmetric programming

Lecture: Algorithms for LP, SOCP and SDP

Assignment 1: From the Definition of Convexity to Helley Theorem

10 Numerical methods for constrained problems

Transcription:

PRIMAL-DUAL ENTROPY-BASED INTERIOR-POINT ALGORITHMS FOR LINEAR OPTIMIZATION MEHDI KARIMI, SHEN LUO, AND LEVENT TUNÇEL Abstract. We propose a family of search directions based on primal-dual entropy in the context of interior-point methods for linear optimization. We show that by using entropy-based search directions in the predictor step of a predictor-corrector algorithm together with a homogeneous self-dual embedding, we can achieve the current best iteration complexity bound for linear optimization. Then, we focus on some wide neighborhood algorithms and show that in our family of entropy-based search directions, we can find the best search direction and step size combination by performing a plane search at each iteration. For this purpose, we propose a heuristic plane search algorithm as well as an exact one. Finally, we perform computational experiments to study the performance of entropy-based search directions in wide neighborhoods of the central path, with and without utilizing the plane search algorithms. Keywords: interior-point methods, primal-dual entropy, central path, homogeneous and selfdual embedding, search direction.. Introduction Primal-dual interior-point methods have been proven to be one of the most useful algorithms in the area of modern interior-point methods for solving linear programming LP) problems. In this paper, we are interested in a class of path-following algorithms that generate a sequence of primal-dual iterates within certain neighbourhoods of the central path. Several algorithms in this class have been studied, which can be distinguished by the choice of search direction. We introduce a family of search directions inspired by nonlinear reparametrizations of the central Date: October 04, revised: August 05. Mehdi Karimi: m7karimi@uwaterloo.ca) Department of Combinatorics and Optimization, Faculty of Mathematics, University of Waterloo, Waterloo, Ontario NL 3G, Canada. Research of this author was supported in part by OGS scholarships from the government of Ontario, a Discovery Grant from NSERC, and by ONR Research Grant N0004---0049. Shen Luo: shenluo@alumni.uwaterloo.ca). Toronto, Ontario, Canada. Research of this author was supported in part by an NSERC Discovery Grant. Levent Tunçel: ltuncel@uwaterloo.ca) Department of Combinatorics and Optimization, Faculty of Mathematics, University of Waterloo, Waterloo, Ontario NL 3G, Canada. Research of this author was supported in part by Discovery Grants from NSERC and by ONR Research Grant N0004---0049.

KARIMI, LUO, and TUNÇEL path equations, as well as the concept of entropy. Entropy and the underlying functions have been playing important roles in many different areas in mathematics, mathematical sciences, and engineering; such as partial differential equations [3], information theory [5, 3], signal and image processing [4, 9, 3], smoothing techniques [9], dynamical systems [8], and various topics in optimization [7,, 3, 4, 8]. In the context of primal-dual algorithms, we use the entropy function in determining the search directions as well as measuring centrality of primaldual iterates. Consider the following form of LP and its dual problem: P) minimize c x subject to Ax = b, x 0, D) maximize b y subject to A y + s = c, s 0, where c R n, A R m n, and b R m are given data. Without loss of generality, we always assume A has full row rank, i.e., ranka) = m. We assume throughout this paper that positive integers m and n satisfy n m +. Let us define F and F + as F := {x, s) : Ax = b, A y + s = c, x 0, s 0, y R m }, F + := {x, s) : Ax = b, A y + s = c, x > 0, s > 0, y R m }. Next, we define the standard primal-dual central path with parameter > 0, i.e. C := {x, s ) : > 0}, as the solutions of the following system: ) A y + s = c, s > 0 Ax = b, x > 0 Xs = e, where e is the all ones vector whose dimension will be clear from the context in this case n). The above system has a unique solution for each > 0. For every pair x, s) F, we define the average duality gap as := xt s n. In standard primal-dual algorithms, search direction is found by applying a Newton-like method to the equations in system ) with an appropriate value of + and the current point as the starting point. Explicitly, the search direction at a point x, s) F + is the solution of the following linear system of equations:

Interior-Point Algorithms Based on Primal-Dual Entropy 3 ) 0 A I A 0 0 S 0 X d x d y d s = 0 0 XSe + + e. The first two blocks of equations in ) are linear and as a result, they are perfectly handled by Newton s method. The nonlinear equation Xs = + e plays a very critical role in Newton s method. Now, if we apply a continuously differentiable strictly monotone function f : R + R to both sides of Xs = + e element-wise), clearly the set of solutions of ) does not change, but the solutions of Newton system might change dramatically. This reparametrization of the KKT system can potentially give us an infinite number of search directions, but not all of them would have desirable properties. For a diagonal matrix V, let fv ) and f V ) denote diagonal matrices with the jth diagonal entry equal to fv j ) and f v j ), respectively. Replacing Xs = + e with fxs) = f + e) and applying Newton s method gives us the same system as ) with the last equation replaced by see [35]): 3) Sd x + Xd s = f XS)) f + e) fxs)). This kind of reparametrization has connections to Kernel functions in interior-point methods see our discussion in Appendix A). Every choice of a continuously differentiable strictly monotone function f in 3) gives us a search direction. These search directions include some of the previously studied ones. For example, the choice of fx) = /x gives the search direction proposed in [4] also see [5] for another connection to the entropy function), and the choice of fx) = x leads to the work in [6]. A natural choice for f is ln ), which has been studied in [35], [38], and [8]. Substituting fx) = lnx) in 3) results in 4) ) Xs Sd x + Xd s = XS) ln. This is the place where entropy function comes into play. In this paper, we study the behaviour of the search direction derived by using 4), with an appropriate choice of +. This search direction corresponds to the gradient of the primal-dual entropy-based potential function ψx, s) := n x js j lnx j s j ) see [35]). As in [35], we define a proximity measure δx, s) as: + 5) δx, s) := ) x j s j n ln xj s j.

4 KARIMI, LUO, and TUNÇEL We sometimes drop x, s) in δx, s) when the argument of δ is clear from the context. If we choose + such that ln 6) ) + = δx, s), then 4) is reduced to [ )] Xs Sd x + Xd s = Xs + δxs XS) ln. This is exactly the search direction studied in [35] for the following neighborhood of the central path) { N E β) := x, s) F + : ) β ln xj s j } + β, for all j, where β. It is proved in [35] that we can obtain the iteration complexity bound of O n ln )) ɛ for N E 3/). We will generalize 6) to define our family of entropy-based search directions. In the vast literature on primal-dual interior-point methods, two of the closest treatments to ours are [35] and [38]. Our search directions unify and generalize the search directions introduced in [35] and [38]. Besides that, for infeasible start algorithms, we use homogeneous self-dual embedding proposed in [37]. In this approach, we combine the primal and dual problems into an equivalent homogeneous self-dual LP with an available starting point of our choice. It is proved in [37] that we can achieve the current best iteration complexity bound of O n ln )) ɛ by using this approach. See Appendix B for a definition of homogeneous self-dual embedding and the properties of it that we need. In Section, we introduce our family of search directions that generalizes and unifies those proposed in [35] and [38], and prove some basic properties. In Section 3, we use the entropybased search direction in the predictor step of a predictor-corrector algorithm for the narrow neighborhood of the central path N β) := { x, s) F + : Xs e β and prove that we can obtain the current best iteration complexity bound of O n ln )) ɛ. After that, we focus on the wide neighborhood N β) := { x, s) F + : x js j }, } β, for all j, and work with our new family of search directions, parameterized by η which indicates the weight of a component of the search direction that is based on primal-dual entropy). For various primal-dual interior-point algorithms utilizing the wide neighborhood, see [, 6, 30, 34] and the references therein. In Section 4, we derive some theoretical results for the wide neighborhood. However, our main goal in the context of wide neighborhood algorithms is to investigate the best practical performance for this class of search directions, in terms of total number of iterations. At each iteration, to find the best search direction in the family i.e. the best value of η) that gives us the longest step and hence the largest decrease in the duality gap), we perform a plane

Interior-Point Algorithms Based on Primal-Dual Entropy 5 search. For this purpose, we propose a heuristic plane search algorithm as well as an exact one in Section 5. Then, in Section 6, we perform computational experiments to study the performance of entropy-based search directions with and without utilizing the plane search. Our computational experiments are on a class of classical small dimensional problems from NETLIB library [7]. Section 7 is the conclusion of this paper.. Entropy-based search directions and Basic properties In this section, we derive some useful properties for analyzing our algorithms. convenient to work in the scaled v-space. Let us define It is more v := X / S / e, 7) u := Xs = V v. We define the scaled right-hand-side vector with parameter η R + as [ )] V v 8) wη) := v + η δv V ln. This definition generalizes and unifies the search )} directions proposed in [35] η = ) and [38] η = σ {, with σ 0.5, ) and σ < min ln β ). For simplicity, we write w := w), which is the scaled right-hand-side vector of 6). By using 7), we can also write δ = n n u j lnu j ). If we define [wη)] p as the projection of wη) on the null space of the scaled matrix Ā := AD, where D := X / S /, and define [wη)] q := wη) [wη)] p, then in the original space, the primal and dual search directions are d x = D d x and d s = D ds, respectively, where d x := [wη)] p and d s := [wη)] q. In other words, the scaled search directions, i.e., d x and d s, can be obtained from the unique solution of the following system: 9) 0 Ā I Ā 0 0 I 0 I d x d y d s = 0 0 wη). Note that d y in the above system 9) is the same d y as in ) since A d y + d s = 0 if and only if 0 = DA d y + Dd s = Ā d y + d s. Most of the upcoming results in this section are for the neighborhood N defined as: N β) := { x, s) F + : Xs e β We also use some of these results for N since N β) N β) for all β > 0, this is valid). Let us start with the following lemma see [35]): }.

6 KARIMI, LUO, and TUNÇEL Lemma.. For every x > 0, s > 0, we have. δ 0;. equality holds above if and only if Xs = e. The following lemma is well-known and is commonly used in the interior-point literature and elsewhere. See, for instance, Lemma 4. in [6] and Lemma in [36]. Lemma.. For every α R such that α, we have : α α ln + α) α. α ) Remark.. The right-hand-side inequality above holds for every α, + ). Next, we relate the primal-dual proximity measure δx, s) to a more commonly used -norm proximity measure for the central path. Lemma.3. Let β [0, ) such that x, s) N β). Then, 3β β)n Xs e δx, s) Xs n e Proof. The right-hand-side inequality was proved in [35]. We prove the left-hand-side inequality here. Let β [0, ), such that x, s) N β). Then, we have estimations are done in the u-space): δu) = n u j ln u j ) n [ u j u j u j ) ] u j ) = n u e n u j u j ) u j ) n u + β) e n β) u e = 3β Xs β)n e. In the above, the first inequality uses Lemma., the first equality uses n u j = n, and the second inequality follows from the fact that x, s) N β). Corollary.. For every x, s) N 4), δ Xs. Moreover, for every x, s) N 0 ), δ 7 8n Xs e. 6n e.

Interior-Point Algorithms Based on Primal-Dual Entropy 7 Next, we want to study the behaviour of the search direction w = v + δv V ln V v ). We already have upper and lower bounds on δ, so we can easily estimate v + δv. Next, we estimate V ln V v ) within the neighborhood N β). Lemma.4. Let β [0, ). Then, for every x, s) N β), we have: δu) β 4β 6β + ) v + V e w δu) )v + V e. Proof. Let x, s) N β) for some β [0, V v ). Then, β)e + β)e. On the one hand, using Lemma., we have ) V v V ln = V lnv e) = V lne + V e e) V V e e) = V e v. On the other hand, using Lemma. again and the fact that x, s) N β), β [0, ), for every j {,,..., n}, we have ) v j ln v j = ) u j ln u j u j u j + To justify some of the remarks following this proof, we focus on the cases ) uj ). u j u j [ β, ], u j [, + β]. Case u j [ β, ]): Using the derivation above, we further compute ) ) v j ln vj uj u j + ) u j u j = v j + u j) ) v j uj = v j v j [ v j v j [ + u j ] + u j) u j u j ) β β) β) ].

8 KARIMI, LUO, and TUNÇEL Case u j [, + β]): Again, using the derivation before this case analysis, we further compute ) v j ln v j u j u j + = v j v j [ + u j) = v j v j u j + u j v j v j [ + v j v j [ + u j ) β + β) ) uj ) u j ] ] β β) β) where the last inequality uses the fact that β 3β + + β for β [0, ). Therefore, within the neighborhood N β), for β [0, ), we conclude that the claimed relation holds. Remark.. Focusing on the case analysis in the last proof, we see that for those j with x j s j Case ), the corresponding component w j of w is very close to the corresponding component computed for a generic primal-dual search direction. For example, for β [0, /4], δu) 40 ], ) v j + v j w j δu) ) v j + v j. Corollary.. For every x, s) N 4), δu) ) v + V e w δu) )v + V e. Remark.3. Recall that in a generic primal-dual search direction, w is replaced by [ v + γv e ], γ [0, ] being the centering parameter. The above corollary shows that inside the neighborhood N 4 ), [ ] v + δu)) δu) V e w δu) v + δu) V e. Since by Lemma.3, inside the neighborhood N /4) we have δu) /6, working with w is close to setting the centrality parameter γ :. Let us define the following quantities which play an important role in analysis of our algorithms: u) := u j lnu j ), u) := u j ln u j ), u) := u j ln u j ).

Interior-Point Algorithms Based on Primal-Dual Entropy 9 We drop the argument u, e.g. we write ij instead of ij u)) when u is clear from the context. The next few results provide bounds on the above quantities. Lemma.5. Let β [0, 4 ] and assume that x, s) N β). Then, 0) ξ ij nδu) ij ζ ij nδu), ij {, }, where ξ := 3 β) + β) ln β), ζ := 3 + β) + + β) ln + β), ξ := β) + 6 β) ln β) + 6 β) ln β), ζ := + β) + 6 + β) ln + β) + 6 + β) ln + β). Proof. See Appendix C. Corollary.3. For every x, s) N 4), we have.8nδu) 9 nδu), and < 5nδu). Lemma.6. Let β [0, ] and assume that x, s) N β). Then, 0 ζ nδu), where ζ := lnn)+). Furthermore, the upper bound is tight within a constant factor for large n. Proof. The left-hand-side inequality obviously holds due to the nonnegativity of the vectors x, s, u and ln Uu). For ζ = lnn) + ), let us define F := ζ nδu), then F u) = lnn) + )e + lnn) + ) lnu) Diaglnu)) lnu) lnu), F u) = lnn)u Diaglnu))U. We consider the constrained optimization problem minimize u R n F u) subject to e u n = 0, u e 0. The Lagrangian has the form L u, λ) = F u) λ e u n) λ u e). Then, F is positive definite if u < ne. Since we know that u n+ e within N ), we conclude that F is strictly convex here. Moreover, for u = e, the Lagrange multipliers λ = ζ and λ = 0 satisfy the KKT conditions. Therefore, u is the global minimizer of the optimization problem. We notice that F u ) = 0 which implies the desired conclusion.

0 KARIMI, LUO, and TUNÇEL Let u R n ++ be a vector with n ) entries equal to / and one entry equal to n + )/. Then, for large n, we have: n nδu) = ln /) + n+ ln ) n+ ) n + n ln/) + n+ ln ) n+ ln = lnn + ) ln). Thus, the upper bound is tight within a constant factor, for large n. Lemma.7. Let x > 0, s > 0. Then nδ. Moreover, equality holds if and only if Xs = e. Proof. Let x > 0, s > 0. Since u j > 0, u j > 0 and u j lnu j ) 0. Using Cauchy-Schwarz inequality, we have u j ln u j ) u j lnu j ) u j lnu j ). u j Then, the claimed inequality follows. Moreover, by utilizing n u j = n, we have equality if and only if u = e we used Cauchy-Schwarz inequality), or equivalently Xs = e. Now, we have all the tools to state and analyze our algorithms. 3. Iteration Complexity Analysis for Predictor-Corrector Algorithm As stated in[ previous section, )] our search directions are the solutions of system 9), where wη) := v + η δv V ln V v. Here, η R + parameterizes the family of search directions. [35] and [38] studied these search directions for special η from iteration complexity point of view. It is proved in [35] that, using w) as the search direction i.e., η = ), we can obtain the iteration complexity bound of O n ln )) ɛ for NE 3/), for feasible start algorithms. These search directions have also been studied { in [38], in )} the wide neighborhood, for the special case that η = σ with σ 0.5, ) and σ < min, ln β. It was shown in [38] that the underlying infeasiblestart algorithm, utilizing a wide neighborhood, has iteration complexity of O n ln )) ɛ. In this section, we show that the current best iteration complexity bound O n ln ɛ )) can be achieved if we use the entropy-based search direction in the predictor step of the standard predictor-corrector algorithm proposed by Mizuno, Todd and Ye [], together with homogeneous self-dual embedding. Here is the algorithm: Algorithm 3.. Input: A, x 0), s 0), b, c, ɛ), where x 0), s 0) ) N 4 ), and ɛ > 0 is the desired tolerance. x, s) x 0), s 0) ),

Interior-Point Algorithms Based on Primal-Dual Entropy while x s > ɛ, predictor step: solve 9) with η = for d x and d s. xα) := x + αd d x, sα) := s + αd ds, where D = X / S /. α := max{α : xα), sα)) N )}. Let x xα ), and s sα ) corrector step: solve 9) for d x and d s, where wη) is replaced by v + V e, Let x x + D d x, s s + D ds. end {while}. The O n ln ɛ )) iteration complexity bound is the conclusion of a series of lemmas. Lemma 3.. For every point x, s) N 4 ), the following condition on α guarantees that xα), sα)) N ): d 4 α 4 + d 3 α 3 + d α + d α + d 0 0, where d 0 := 3 0, d := 3 δ x js j nδ + nδ + 6 = 3 δ x js j + 6, d := 6 δ x j s j ) + δ + C d + 3, d 3 := 3δ )C 3B, d 4 := 6 w p ) jw q ) j, B := C := x j s j ln u j ) w p ) j w q ) j, x j s j w p ) j w q ) j. Proof. See Appendix C. Lemma 3.. For every point x, s) N 4 ), we have the following bounds on d, d, d 3 and d 4 defined in Lemma 3.. d 0, d 34n, d 3 64n 3, d 4 5n.

KARIMI, LUO, and TUNÇEL Proof. See Appendix C. We state the following well-known lemma without proof. Lemma 3.3. [] For every point x, s) N ), the corrector step of Algorithm 3. returns a point in the neighborhood N 4 ). Now, we can prove the iteration complexity bound for Algorithm 3.. Theorem 3.. Algorithm 3. gives an ɛ-solution in O n ln ɛ )) iterations. Proof. By Lemma 3. and Lemma 3., in the predictor step, it is sufficient for α to satisfy ) 5n α 4 + 64n 3 α 3 + 34nα + 0α 3. It is easy to check that α = 50 satisfies this inequality. Lemma 3.3 shows that we have a point n x, s) N 4 ) at the beginning of each predictor step and the algorithm is consistent. Since xα) sα) = α)x s by part b) of Lemma 3. of [35], we deduce that the algorithm will reach an ɛ-solution in O n ln )) ɛ iterations. 4. Algorithm for the Wide Neighborhoods In the rest of the paper, we study the behaviour of the entropy-based search directions in a wide neighborhood. As mentioned before, for [ each η, our search )] direction is derived from the solution of system 9), where wη) := v + η δv V ln V v. These search directions have been studied { in [38] in the )} wide neighborhood for the special case that η = σ with σ 0.5, ) and σ < min, ln β. In this paper, we study these search directions for a wider range of η. We prove some results on iteration complexity bounds in this section. However, in the rest of the paper, we mainly focus on the practical performance of our search directions in the wide neighborhood. The algorithm in a wide neighborhood with a value of η 0 fixed by the user) is: Algorithm 4.. Input A, x 0), s 0), b, c, ɛ, η), ɛ > 0 is the desired tolerance. x, s) x 0), s 0) ), while x s > ɛ solve 9) for d x and d s, xα) := x + αd d x, sα) := s + αd ds, where D = X / S /.

α := max{α : xα), sα)) N β)}. Let x xα ); s sα ) end {while}. Interior-Point Algorithms Based on Primal-Dual Entropy 3 Lemma 4.. In Algorithm 4., for every choice of η R +, we have xα) sα) = α) n. Proof. We proceed as in the proof of Lemma 3. of [35], part b): xα) sα) = x s + αv d x + d s ) = x s + αv wη) = α)x s. ) For the last equation, we used the facts that v v = x s, and v and δv V ln V v are orthogonal. This lemma shows that the reduction in the duality gap is independent of η and is exactly the same as in the primal-dual affine scaling algorithm. So, Lemma 4. includes part b) of Lemma 3. of [35] and part c) of Theorem 3. of [] as special cases. We show later that by performing a plane search, we can find an η that gives the largest value of α in the algorithm and hence the largest possible reduction in duality gap, per iteration). Lemma 4.. Let x > 0, s > 0. Then, for η 0, we have wη) = n + η nη δ. Proof. Let x > 0, s > 0, and η 0. Then, )) wη) = vj xj s j δη η ln = ) x j s j δ η + + η ln xj s j + η ln = nδ η + n + η + nηδ δηn nδη δ = n + η nη δ ) xj s j δη δη ln )) xj s j Theorem 4.. If we apply Algorithm 4. with N ), then the algorithm converges to an ɛ- solution in at most O n lnn) ln )) ɛ iterations for every nonnegative η = O). Proof. See Appendix C. In the above algorithm, the value of η is constant for all values of j. In the following, we show that if η is allowed to take one of two constant values for each j one of the values being zero),

4 KARIMI, LUO, and TUNÇEL we get a better iteration complexity bound. For each j {,..., n}, let us define v j, if u j > [ )] 3 4, ) [wη)] j := vj v j + η δv j v j ln, if u j 3 4, where η := δ+ln)). Now we have the following theorem: Theorem 4.. If we apply the Algorithm 4. with wη) defined in ) to N ), the algorithm converges to an ɛ-solution in at most On ln ) ɛ ) iterations. Proof. See Appendix C. 5. Plane Search Algorithms In the previous section, we showed how to fix two parameters α and η to achieve iteration complexity bounds. However, in practice we may consider performing a plane search to choose the best α and η in each iteration. Here, our goal is to choose a direction in our family of directions [ that gives the most reduction in the duality gap. As before, we have wη) = v + η δv V ln V v ]. ) For simplicity, in this section we drop parameter η and write w = wη), so w p = P AD w, and w q = w w p, where P AD is the projection operator onto the null space of AD. Our goal is to solve the following optimization problem. maximize α subject to 0 < α <, 3) η 0, w p ) j w q ) j α + α u j δη u j lnu j )η u j + ) + u j ) 0, j {,..., n}. In the above optimization problem, the objective function is linear and the main constraints are quadratic. Let us define )) )) V v V v t p := P AD δv V ln, t q := δv V ln t p, v p := P AD v), v q := v v p. By these definitions, the quadratic inequalities in formulation 3) become a j η α + b j ηα + c j ηα + d j α) + e j α 0, where a j := t p) j t q ) j, b j := u j δ u j lnu j ), c j := v p) j t p ) j + v q ) j t p ) j, e j := v p) j v q ) j, d j := u j.

Interior-Point Algorithms Based on Primal-Dual Entropy 5 In this section, we propose two algorithms, an exact one and a heuristic one, to solve the twovariable optimization problem 3). 5.. Exact plane search algorithm. We define a new variable z := αη. Then, the quadratic form can be written as: g j z, α) := a j z + b j z + c j zα + d j α) + e j α, j {,..., n}. We are optimizing in the plane of α and z, actually working in the one-sided strip in R, defined by 0 α and z 0. The following proposition establishes that it suffices to check On ) points to find an optimal solution: Proposition 5.. Let α, η ) be an optimal solution of 3). Then, one of the following is true: ) α = ; ) there exists z 0 such that z, α ) is a solution of system g jz, α) = 0 for g i z, α) 0 some pair i, j {,..., n}; 3) α is a solution of j α) := b j + αc j ) 4a j d j α) + e j α ) = 0, j {,..., n}, where j α) is the discriminant of g j z, α) with respect to z. Proof. Assume that α, η ) is a solution to 3), and z := α η. Therefore, we have g j z, α ) 0, j {,..., n}. By continuity, we must have g j z, α ) = 0 for at least one j, and because z is real, we have j α ) 0. If j α ) = 0, then condition 3) is satisfied; otherwise, by continuity, we can increase α so that j remains positive. In this case, if there does not exist another i {,..., n} such that g i z, α ) = 0, continuity gives us another point ᾱ, η) that is feasible to 3) and ᾱ > α, which is a contradiction. Hence, condition ) must hold. The above proposition tells us that to find a solution for 3), it suffices to check On ) values for α. For calculating each of these values, we find the roots of a quartic equation. 5.. Heuristic plane search algorithm. The idea of the heuristic algorithm is that we start with α = and see if there exists η such that η, α) is feasible for 3). If not, we keep reducing α and repeat this process. We can reduce α by a small amount for example 0.0) if α is close to for example α 0.95), and by a larger amount for example 0.05) otherwise. This approach tries to favor the larger α values over the smaller ones. The difficult part is checking if there exists η for the current α in the algorithm. To do that, we need to verify if there exists a positive η which satisfies the n inequalities in the constraints 3) of 3). Each constraint is a quadratic form in η and can induce a feasible interval for η. If

6 KARIMI, LUO, and TUNÇEL the intersection of all the intervals corresponding to these n inequality constraints is not empty, we then find the η corresponding to a step length α. We use the following procedure to determine the feasible interval of η for a given step length α. Assume that we fix α. For each quadratic constraint of 3), we can solve for η and find the feasible interval. One form is the union of two open intervals, i.e.,, r j)] and [r j), ); denote the indexes in this class as K. Another is the convex interval [r 3 j), r 4 j)]; denote the indexes in this class by K. It is easy to find the intersection of the convex intervals: [t, t ] := [max j K r 3 j), min j K r 4 j)]. Now we have to intersect [t, t ] with the intervals in class K. First, we handle the intervals [t, t ] that intersect only one of, r j)] and [r j), ); in that case we can update [t, t ] [t, r j)] or [t, t ] [r j), t ] for each of these intervals. At the end of this step, we can assume that for the rest of the intervals in K we denote them by K ), [t, t ] intersects both, r j)] and [r j), ). Then, we can define two intervals: [t, t 3 := min j K r j))], [t 4 := max j K r j)), t ] If one of these intervals is non-empty, then there exists η such that η, α) is feasible for 3), and we return α. For a more detailed introduction to this heuristic see [9]. To evaluate the performance of our heuristic algorithm, note that the set of feasible points α, η) of 3) in R is not necessarily a connected region. We can think of it as the union of many connected components. In our heuristic algorithm, we check a few discrete values of α = ᾱ. However, for each value we check, we can precisely decide if there exists a feasible η for that value of α. If one of the lines α = ᾱ intersects a component of feasible region that contains a point with maximum α, then our heuristic algorithm returns an α that is close the optimal value. However, if none of the lines α = ᾱ that we check for large values of ᾱ intersects the right component, the heuristic algorithm may return a very bad estimate of the optimal value. In the next section, we observe that see Figures 5 8) our heuristic algorithm in the worst-case may return values for α very close to zero while the optimal value is close to. 6. Computational Experiments with the Entropic Search Direction Family We performed some computational experiments using the software MATLAB R04a, on a 48- core AMD Opteron 676 machine with 56GB of memory. The test LP problems are well-known among those in the problem set of NETLIB [7]. We implemented Algorithm 4. for a fixed value of η and then ran it for each fixed η {,, 3, 4}. We also implemented Algorithm 4. with η being calculated using the exact and

Interior-Point Algorithms Based on Primal-Dual Entropy 7 heuristic plane search algorithms. β = / was set for the algorithm, therefore our results are for the wide neighborhood N /). We used homogeneous self-dual embedding for the LP problems as shown in Appendix B. The initial feasible solution is y 0) := 0, x 0) := e, s 0) := e, θ :=, t := and κ :=. In the statements of Algorithms 3. and 4., we used the stopping criterion x T s ɛ, which is an abstract criterion assuming exact arithmetic computation. In practice, we may encounter numerical inaccuracies and we need to take that into account for our stopping criterion. We used the stopping criterion proposed and studied in [5], which is very closely related to the stopping criterion in SeDuMi [33]. Let us define x, ȳ, s) := x τ, y τ, s τ ), and their residuals: r p := b A x, r d := A ȳ + z c, r g := c x b ȳ. The following stopping criterion for general convex optimization problems using homogeneous self-dual embedding was proposed in [5]: r p + r d max{0, r g } + + b + c max { c x, b ȳ, } r max. In our algorithm, we used the above stopping criterion for r max := 0 9. Table shows the number of iterations for each problem. The first four columns show the number of iterations of Algorithm 4. with a fixed value of η {,, 3, 4}. Let use define η and η as the η found at each iteration of the plane search algorithm using the heuristic and exact plane search algorithms, respectively. The fifth and sixth columns of the table are the number of iterations when we perform a plane search, using the heuristic plane search and exact plane search algorithms, respectively. The problems in the table are sorted based on the value of η {,..., 4} that gives the smallest number of iterations. For each η, the problems are sorted alphabetically.

8 KARIMI, LUO, and TUNÇEL Table : The number of iterations of Algorithm 4.. NETLIB Name Dimensions Nonzeros η = η = η = 3 η = 4 η η af iro 8 3 88 3 37 45 55 9 8 beaconf d 74 6 370 3 37 46 56 blend 75 83 5 35 37 4 48 3 9 grow7 4 30 633 47 47 55 65 36 3 grow5 30 645 5665 4 47 54 66 37 3 sc05 06 03 8 8 36 43 54 3 9 sc05 06 03 55 8 37 4 53 5 sc50a 5 48 3 8 35 43 5 8 sc50b 5 48 9 6 33 43 50 9 6 scagr7 30 40 533 34 39 47 55 8 5 scsd 78 760 348 3 35 43 5 6 8 scsd8 398 750 334 3 34 4 48 5 9 shareb 97 79 730 35 40 45 54 5 3 adlittle 57 97 465 40 40 49 56 7 4 kb 44 4 9 43 43 49 58 3 30 agg 489 63 54 58 50 58 68 46 39 agg 57 30 455 6 5 56 66 40 35 agg3 57 30 453 70 55 6 68 4 37 boeing 67 43 339 7 53 54 6 39 37 brandy 49 506 7 56 56 57 43 36 capri 7 353 786 60 49 53 57 39 37 degen 445 534 4449 4 38 4 48 37 4

Interior-Point Algorithms Based on Primal-Dual Entropy 9 NETLIB Name Dimensions Nonzeros η = η = η = 3 η = 4 η η degen3 504 88 630 40 38 40 45 30 6 f itd 5 06 4430 68 53 64 64 38 37 f orplan 6 4 496 08 77 79 9 60 50 ganges 30 68 706 5 48 54 63 4 38 gf rd pnc 67 09 3467 50 47 53 63 35 30 grow 44 946 838 5 50 55 67 38 34 lotf i 54 308 086 54 46 50 58 35 3 scagr5 47 500 09 43 4 5 58 33 3 scsd6 48 350 5666 35 36 44 5 5 sctap 09 880 84 56 48 49 5 34 3 ship04s 403 458 590 53 45 5 54 34 30 ship04l 403 8 8450 53 49 56 58 35 9 stocf or 8 474 5 48 55 6 30 6 woodp 45 594 706 0 6 04 65 53 5 f itp 68 677 0894 63 54 54 59 37 35 bandm 305 47 659 67 54 5 58 40 35 boeing 35 384 3865 9 68 65 68 5 46 e6 4 8 767 66 53 5 55 40 36 israel 75 4 358 96 69 67 73 46 40 d6cube 404 684 37704 76 58 55 56 39 3 modszk 686 6 370 0 87 77 8 75 53 scf xm 33 457 6 3 8 73 74 53 4

0 KARIMI, LUO, and TUNÇEL NETLIB Name Dimensions Nonzeros η = η = η = 3 η = 4 η η scrs8 49 69 409 43 05 95 00 50 44 sctap3 48 480 0734 64 53 53 57 37 5 ship08s 779 387 950 8 6 6 63 54 3 vtp base 99 03 94 7 84 76 77 50 36 scf xm3 99 37 7846 43 98 84 84 70 47 5f v45 8 57 7 60 9 79 77 57 bnl 644 75 69 83 03 95 87 64 bnl 35 3489 64 97 37 0 00 78 7 czprob 930 353 473 36 56 9 9 70 6 etamacro 40 688 489 8 34 09 98 66 59 pilot4 4 000 545 50 07 9 88 87 77 pilot we 73 789 98 34 64 39 5 9 8 perold 66 376 606 90 3 0 95 94 93 scf xm 66 94 59 46 97 83 8 69 47 sctap 30 480 05 9 8 75 67 37 35 seba 56 08 4874 70 0 03 94 78 50 shareb 8 5 8 8 84 74 73 58 5 shipl 5 5437 597 7 64 33 0 75 50 ships 5 763 094 8 34 06 97 74 4 stocf or 58 03 949 9 95 83 8 63 56 standata 360 075 3038 08 7 67 66 33 7 standmps 468 075 3686 5 85 73 70 63 35

Interior-Point Algorithms Based on Primal-Dual Entropy As we mentioned above, our family of search directions is a common generalization of the search direction in [35] that uses { η = and )} the search directions in [38] and [8] that use η = σ with σ 0.5, ) and σ < min, ln β, so η. As we observe from Table, our generalization to consider using larger values of η is justified. Among the problems solved and among the fixed values for η {,, 3, 4}, η = had the smallest iteration count for 5 problems, η = won for 4 problems, η = 3 won for 3 problems, and η = 4 had the smallest iteration count for 8 problems ties counted as wins for both winning η s). Table also shows that using plane search algorithms can be crucial in reducing the number of iterations in addition to making the behaviour of the underlying algorithms more robust; as ) for most of the problems, there is a large gap between the number of iterations of the plane search and the best constant η algorithms, and ) we do not know which η is the best one before solving the problem. The exact plane search algorithm gives a lower bound for our heuristic plane search algorithm. As we observe from Table, for most of the problems, exact and heuristic plane search algorithms have similar performances in terms of the number of iterations. In Figures 4, we plot the value of η at each iteration for four of the problems of NETLIB, for both exact and heuristic plane search algorithms. For beaconfd and capri the performances are close and for degen and ship08s there is a large gap. An interesting point is that the plane search algorithms sometimes lead to values of η as large as 0 or 0 as can be seen in Figures 3 and 4. Figures 5 8 provide a more reasonable comparison between the exact and heuristic plane search algorithms for problems degen and ship08s. In Figure 5 for degen ) and Figure 7 for ship08s), we plot the values of η and α for the heuristic algorithm, as well as the corresponding values that would have been computed by the exact algorithm at each iteration for the same current iterates x k), s k) )). In Figure 6 for degen ) and Figure 8 for ship08s), we plot the values of η and α for the exact algorithm, as well as the corresponding values that would have been given by the heuristic algorithm at each iteration. Note that in Figures 5 8, the comparison is iteration-wise. The plot in solid line is the main algorithm and the plot in dotted line is the value that would have been returned by the other algorithm using the iterates generated by the main algorithm. We observe from the figures that when the optimal value of α is close to or 0, the heuristic algorithm cannot keep up with the exact algorithm. A conclusion of the above discussion is that utilization of plane search algorithms improves the number of iterations significantly. If the plane search algorithm is fast enough, then we can also improve the running time. Our heuristic plane search algorithm is much faster than the exact one. For the exact plane search algorithm, we solve On ) quartic equations, and in each iteration of the primal-dual algorithm, we perform On 3 ) operations. Therefore, if we can speed up our exact plane search algorithm, this would have a potential impact on practical performance of algorithms in this paper as well as some other related algorithms. Note that our main focus

KARIMI, LUO, and TUNÇEL in these preliminary computational experiments is on the number of iterations. To speed up the plane search algorithms, one may even use tools from computational geometry, analogous to those used for solving two-dimensional or O)-dimensional) LP problems with n constraints in On) time see [0], [0], and the book []). 7. Conclusion In this paper, we introduced a family of search directions parameterized by η. We proved that if we use our search direction with η = in the predictor step of standard predictor-corrector algorithm, we can achieve the current best iteration complexity bound. Then, we focused on the wide neighborhoods, and after the derivation of some theoretical results, we studied the practical performance of our family of search directions. To find the best search direction in our family, which gives the largest decrease in the duality gap, we proposed a heuristic plane search algorithm as well as an exact one. Our experimental results showed that using plane search algorithms improves the performance of the primal-dual algorithm significantly in terms of the number of iterations. Although our heuristic algorithm works efficiently, there is more room here to work on other heuristic plane search algorithms or improving the practical performance of the exact one, so that we also obtain a significant improvement in the overall running time of the primal-dual algorithm. The idea of using a plane search in each iteration of a primal-dual algorithm has been used by many other researchers. For example, relatively recently, Ai and Zhang [] defined a new wide neighborhood which contains the conventional wide neighborhood for suitable choices of parameter values) and introduced a new search direction by decomposing the right-hand-side vector of ) into positive and negative parts and performing a plane search to find the step size for each vector. By this approach, they obtained the current best iteration complexity bound for their wide neighborhood. Their approach together with ours inspires the following question: are there other efficient decompositions which in combination with a plane search, give good theoretical as well as computational performances in the wide neighborhoods of the central path? This is an interesting question left for future work. References [] W. Ai and S. Zhang. An O nl) iteration primal-dual path-following method, based on wide neighborhoods and large updates, for monotone LCP. SIAM Journal on Optimization, 6):400 47, 005. [] Y.Q. Bai, M. El Ghami, and C. Roos. A comparative study of kernel functions for primal-dual interior-point algorithms in linear optimization. SIAM Journal on Optimization, 5):0 8, 004. [3] X.Z. Cai, G. Q. Wang, M. El Ghami, and Y.J. Yue. Complexity analysis of primal-dual interior-point methods for linear optimization based on a new parametric kernel function with a trigonometric barrier term. Abstract and Applied Analysis, 04Article ID 7058).

Interior-Point Algorithms Based on Primal-Dual Entropy 3 [4] C.I. Chang, Y. Du, J. Wang, S.M. Guo, and P.D. Thouin. Survey and comparative analysis of entropy and relative entropy thresholding techniques. In Vision, Image and Signal Processing, IEE Proceedings-, volume 53, pages 837 850. IET, 006. [5] T.M. Cover and J.A. Thomas. Elements of information theory. John Wiley & Sons, 0. [6] Zs. Darvay. A new algorithm for solving self-dual linear optimization problems. Studia University Babȩs-Bolyai, 47):5 6, 003. [7] A. Decarreau, D. Hilhorst, C. Lemaréchal, and J. Navaza. Dual methods in entropy maximization. application to some problems in crystallography. SIAM Journal on Optimization, ):73 97, 99. [8] T. Downarowicz. Entropy in dynamical systems. Cambridge University Press, 0. [9] N. J. Dusaussoy and I. E. Abdou. The extended ment algorithm: a maximum entropy type algorithm using prior knowledge for computerized tomography. Signal Processing, IEEE Transactions on, 395):64 80, 99. [0] M. Dyer. A class of convex programs with applications to computational geometry. In Proceedings of the eighth annual symposium on Computational geometry, pages 9 5. ACM, 99. [] H. Edelsbrunner. Algorithms in Computational Geometry. Springer New York, 987. [] S. Erlander. Accessibility, entropy and the distribution and assignment of traffic. Transportation Research, 3):49 53, 977. [3] S. Erlander. Entropy in linear programs. Mathematical Programming, ):37 5, 98. [4] S. Fang, J. R. Rajasekera, and H. J. Tsao. Entropy optimization and mathematical programming, volume 8. Springer, 997. [5] R.M. Freund. On the behavior of the homogeneous self-dual model for conic convex optimization. Mathematical programming, 063):57 545, 006. [6] N. Karmarkar. A new polynomial time algorithm for linear programming. Combinatorica, 44):373 395, 984. [7] Y.H. Lee, J.H. Jin, and G.M. Cho. Kernel function based interior-point algorithms for semidefinite optimization. Mathematical Inequalities & Applications, 64):79 94, 03. [8] X.S. Li. Entropy and optimization. PhD thesis, University of Liverpool, U.K., 987. [9] S. Luo. Interior-Point Algorithms Based on Primal-Dual Entropy. Master s thesis, University of Waterloo, Canada, 006. [0] N. Megiddo. Linear-time algorithms for linear programming in R 3 and related problems. In Foundations of Computer Science, 98. SFCS 08. 3rd Annual Symposium on, pages 39 338. IEEE, 98. [] S. Mizuno, M.J. Todd, and Y. Ye. On adaptive-step primal-dual interior-point algorithms for linear programming. Math. Oper. Res., 84):964 98, 993. [] R. D.C. Monteiro, I. Adler, and M. G.C. Resende. A polynomial-time primal-dual affine scaling algorithm for linear and convex quadratic programming and its power series extension. Mathematics of Operations Research, 5):9 4, 990. [3] C. Nadeu and M. Bertran. A flatness-based generalized optimization approach to spectral estimation. Signal Processing, 94):3 30, 990. [4] J.L. Nazareth. A reformulation of the central path equations and its algorithmic implications. Department of Pure and Applied Math., Washington State Univ., 994. [5] J.L. Nazareth. Deriving potential functions via a symmetry principle for nonlinear equations. Operations Research Letters, 3):47 5, 997. [6] Y. Nesterov. Parabolic target space and primal dual interior-point methods. Discrete Applied Mathematics, 56):079 00, 008. [7] NETLIB. http://www.netlib.org/lp/data/. [8] S. Pan, X. Li, and S. He. An infeasible primal dual interior point algorithm for linear programs based on logarithmic equivalent transformation. Journal of mathematical analysis and applications, 34):644 660, 006. [9] M.Ç. Pınar and S.A. Zenios. An entropic approximation of l penalty function. Transactions on Operational Research, pages 0 0, 995. [30] F. A. Potra. Primal-dual affine scaling interior point methods for linear complementarity problems. SIAM Journal on Optimization, 9):4 43, 008. [3] T. Pynchon. A survey of entropy methods for partial differential equations. American Mathematical Society, 44):409 438, 004.

4 KARIMI, LUO, and TUNÇEL [3] C. E. Shannon. A mathematical theory of communication. ACM SIGMOBILE Mobile Computing and Communications Review, 5):3 55, 00. [33] J.F. Sturm. Implementation of interior point methods for mixed semidefinite and second order cone optimization problems. Optimization Methods and Software, 76):05 54, 00. [34] L. Tunçel. On the convergence of primal-dual interior-point methods with wide neighborhoods. Computational Optimization and Applications, 4):39 58, 995. [35] L. Tunçel and M.J. Todd. On the interplay among entropy, variable metrics and potential functions in interiorpoint algorithms. Computational Optimization and Applications, 8):5 9, 997. [36] Y. Ye. An On 3 L) potential reduction algorithm for linear programming. Mathematical Programming, 50-3):39 58, 99. [37] Y. Ye, M.J. Todd, and S. Mizuno. An O nl)-iteration homogeneous and self-dual linear programming algorithm. Mathematics of Operations Research, 9):53 67, 994. [38] P. Zhang and X. Li. An infeasible-start path-following method for monotone LCPs. Mathematical and Computer Modelling, 38-):3 3, 003.

Interior-Point Algorithms Based on Primal-Dual Entropy 5 5 4.5 Heuristic algorithm Exact algorithm 4 3.5 3.5.5 0.5 0 0 5 0 5 0 5 Figure. Values of η for the heuristic algorithm) and η for the exact algorithm) in each iteration for problem beaconfd. 9 8 Heuristic algorithm Exact algorithm 7 6 5 4 3 0 0 5 0 5 0 5 30 35 40 Figure. Values of η for the heuristic algorithm) and η for the exact algorithm) in each iteration for problem capri.

6 KARIMI, LUO, and TUNÇEL 0 9 Heuristic algorithm Exact algorithm 8 7 6 5 4 3 0 0 5 0 5 0 5 30 35 40 Figure 3. Values of η for the heuristic algorithm) and η for the exact algorithm) in each iteration for problem degen. 5 Heuristic algorithm Exact algorithm 0 5 0 5 0 0 0 0 30 40 50 60 Figure 4. Values of η for the heuristic algorithm) and η for the exact algorithm) in each iteration for problem ship08s.

Interior-Point Algorithms Based on Primal-Dual Entropy 7 0 Heuristic algorithm Corresponding by the exact algorithm Heuristic algorithm Corresponding by the exact algorithm 8 0.8 6 0.6 4 0.4 0. 0 0 5 0 5 0 5 30 35 a) 0 0 5 0 5 0 5 30 35 b) Figure 5. Values of a) η b) α for the heuristic algorithm, and the corresponding values calculated by the exact algorithm at each iteration of it, for problem degen. 0 Exact algorithm Corresponding by the heuristic algorithm Exact algorithm Corresponding by the heuristic algorithm 8 0.8 6 0.6 4 0.4 0. 0 0 5 0 5 0 a) 0 0 5 0 5 0 b) Figure 6. Values of a) η b) α for the exact algorithm, and the corresponding values calculated by the heuristic algorithm at each iteration of it, for problem degen.

8 KARIMI, LUO, and TUNÇEL 0 8 6 Heuristic algorithm Corresponding by the exact algorithm Heuristic algorithm Corresponding by the exact algorithm 4 0.8 0 0.6 8 6 0.4 4 0. 0 0 5 0 5 0 5 30 35 40 45 50 a) 0 0 0 0 30 40 50 b) Figure 7. Values of a) η b) α for the heuristic algorithm, and the corresponding values calculated by the exact algorithm at each iteration of it, for problem ship08s. 5 Exact algorithm Corresponding by the heuristic algorithm Exact algorithm Corresponding by the heuristic algorithm 0 0.8 5 0.6 0 0.4 5 0. 0 0 5 0 5 0 5 30 a) 0 0 5 0 5 0 5 30 b) Figure 8. Values of a) η b) α for the exact algorithm, and the corresponding values calculated by the heuristic algorithm at each iteration of it, for problem ship08s.

Interior-Point Algorithms Based on Primal-Dual Entropy 9 Appendix A. Connection with Kernel functions In this section, we introduce the Kernel function approach for interior-point methods [, 7, 3] and discuss its connection with our approach. Let Ψv) : R n ++ R be a strictly convex function such that Ψv) is minimal at v = e and Ψe) = 0. In the Kernel function approach we replace the last equation of ) with 4) Sd x + Xd s = ) v V Ψ, ) where v := X / S / e []. Note that by the definition of Ψ, Ψ v = 0 if and only if x, s) is on the central path. To simplify the matters, we assume that Ψv) = ψv j ), where ψt) : R ++ R is an strictly convex function with unique minimizer at t = and ψ) = 0. We call the univariate function ψt) the Kernel function of Ψv). It has been shown that the short update primal-dual path following algorithms using special Kernel functions obtain the current best iteration complexity bound []. Comparing 3) and 4), we observe that the two approaches are similar in the sense that the left-hand-side of the last equation in ) is replaced by a nonlinear function of Xs. The question here is whether there exists a continuously differentiable strictly monotone function f for each Kernel function ψ or vice versa so that ) and 4) give the same search direction. In other words, can we solve ) t tψ = K f) ft ) 5) f t, ) for f or ψ, for a constant scalar K? For t =, both sides of 5) are equal to zero, so the equation ) is consistent in that sense. ψt) is a strictly convex function with minimum at t =, so ψ t < 0 for t < ) and ψ t > 0 for t >. This makes both sides of 5) consistent for a strictly monotone function f. Hence, 5) may be solved for f or ψ, however the result depends on in general. Table shows five pairs of functions. Some of the Kernel functions in the table are from the set of functions studied in [], and we solved 5) for the corresponding fx). In the last two ones, we picked fx) = lnx) and fx) = x and derived the corresponding ψt).

30 KARIMI, LUO, and TUNÇEL Table : Some ψt) and their corresponding fx) in view of 5). t ψt) fx) lnt) x ) t t x t ) + t q+ q+, q > xq t + ) t x t + t q q, q > x q+ t ) x t lnt) t + lnx) As an example, we see the derivation of fx) for the third ψt): we have ψ t) = t t q, then ) ) t tψ = t t q/ t q = t + q+ t q = q+ t q+ t q = f) ft ) f t ), fx) = x q+. For the forth pair, the function ψt) = t lnt) t + obtains its minimum at t = with ψ) = 0, and is decreasing before t = and increasing after that. The function is also convex around t =, but it is not convex on the whole range of t > 0. As mentioned above, for each Kernel function ψt), solving 5) for fx) may result in a function depending on. We can cover that by generalizing our method as follows. At each iteration, instead of applying f ) to both sides of Xs = e, we apply a function of, i.e. f; ). The rationale behind it is that we expect different behaviours from the algorithm when > and ; e.g., we expect quadratic or at least super-linear convergence when. Hence, it is reasonable to apply a function f ) that depends on. We saw above that 5) gives a non-convex function ψt) for fx) = lnx) and the Kernel function approach does not cover our approach. However, our generalized method contains the Kernel function approach and is strictly more general in that sense. Consider 5) for K = / and assume, without loss of generality, that f ) = 0. Then, from 5), for t we have: 6) tf t )) ) t ft ) f) = ψ d )) t dt [ln ft ) f) )] = ψ [ )) ] t ft ) = exp ψ dt,