A NEW PROXIMITY FUNCTION GENERATING THE BEST KNOWN ITERATION BOUNDS FOR BOTH LARGE-UPDATE AND SMALL-UPDATE INTERIOR-POINT METHODS

Similar documents
A full-newton step infeasible interior-point algorithm for linear programming based on a kernel function

Research Note. A New Infeasible Interior-Point Algorithm with Full Nesterov-Todd Step for Semi-Definite Optimization

A Second Full-Newton Step O(n) Infeasible Interior-Point Algorithm for Linear Optimization

An Infeasible Interior-Point Algorithm with full-newton Step for Linear Optimization

Interior-point algorithm for linear optimization based on a new trigonometric kernel function

Local Self-concordance of Barrier Functions Based on Kernel-functions

A PRIMAL-DUAL INTERIOR POINT ALGORITHM FOR CONVEX QUADRATIC PROGRAMS. 1. Introduction Consider the quadratic program (PQ) in standard format:

A new primal-dual path-following method for convex quadratic programming

Full Newton step polynomial time methods for LO based on locally self concordant barrier functions

A path following interior-point algorithm for semidefinite optimization problem based on new kernel function. djeffal

A FULL-NEWTON STEP INFEASIBLE-INTERIOR-POINT ALGORITHM COMPLEMENTARITY PROBLEMS

A full-newton step infeasible interior-point algorithm for linear complementarity problems based on a kernel function

A Full Newton Step Infeasible Interior Point Algorithm for Linear Optimization

A PREDICTOR-CORRECTOR PATH-FOLLOWING ALGORITHM FOR SYMMETRIC OPTIMIZATION BASED ON DARVAY'S TECHNIQUE

A full-newton step feasible interior-point algorithm for P (κ)-lcp based on a new search direction

Improved Full-Newton Step O(nL) Infeasible Interior-Point Method for Linear Optimization

A New Class of Polynomial Primal-Dual Methods for Linear and Semidefinite Optimization

Interior Point Methods for Linear Programming: Motivation & Theory

A Redundant Klee-Minty Construction with All the Redundant Constraints Touching the Feasible Region

CCO Commun. Comb. Optim.

A WIDE NEIGHBORHOOD PRIMAL-DUAL INTERIOR-POINT ALGORITHM WITH ARC-SEARCH FOR LINEAR COMPLEMENTARITY PROBLEMS 1. INTRODUCTION

Enlarging neighborhoods of interior-point algorithms for linear programming via least values of proximity measure functions

Introduction to Nonlinear Stochastic Programming

4TE3/6TE3. Algorithms for. Continuous Optimization

A Simpler and Tighter Redundant Klee-Minty Construction

On Mehrotra-Type Predictor-Corrector Algorithms

Primal-Dual Interior-Point Methods for Linear Programming based on Newton s Method

A Full-Newton Step O(n) Infeasible Interior-Point Algorithm for Linear Optimization

An EP theorem for dual linear complementarity problems

New Interior Point Algorithms in Linear Programming

Using Schur Complement Theorem to prove convexity of some SOC-functions

2.1. Jordan algebras. In this subsection, we introduce Jordan algebras as well as some of their basic properties.

A Generalized Homogeneous and Self-Dual Algorithm. for Linear Programming. February 1994 (revised December 1994)

PRIMAL-DUAL ALGORITHMS FOR SEMIDEFINIT OPTIMIZATION PROBLEMS BASED ON GENERALIZED TRIGONOMETRIC BARRIER FUNCTION

An E cient A ne-scaling Algorithm for Hyperbolic Programming

On well definedness of the Central Path

An Infeasible Interior Point Method for the Monotone Linear Complementarity Problem

A new Primal-Dual Interior-Point Algorithm for Second-Order Cone Optimization

1. Introduction and background. Consider the primal-dual linear programs (LPs)

Primal-dual relationship between Levenberg-Marquardt and central trajectories for linearly constrained convex optimization

On intermediate value theorem in ordered Banach spaces for noncompact and discontinuous mappings

Interior-Point Methods for Linear Optimization

Lecture 14 Barrier method

AN INTERIOR POINT METHOD, BASED ON RANK-ONE UPDATES, Jos F. Sturm 1 and Shuzhong Zhang 2. Erasmus University Rotterdam ABSTRACT

On the acceleration of augmented Lagrangian method for linearly constrained optimization

Robust Farkas Lemma for Uncertain Linear Systems with Applications

Introduction to optimization

Interior Point Algorithms for Constrained Convex Optimization

Primal-Dual Interior-Point Methods. Javier Peña Convex Optimization /36-725

Interior Point Methods for Nonlinear Optimization

McMaster University. Advanced Optimization Laboratory. Title: A Proximal Method for Identifying Active Manifolds. Authors: Warren L.

EXPLICIT SOLUTION TO MODULAR OPERATOR EQUATION T XS SX T = A

Convergence Analysis of Inexact Infeasible Interior Point Method. for Linear Optimization

A Proximal Method for Identifying Active Manifolds

Interior Point Methods for Convex Quadratic and Convex Nonlinear Programming

Numerical Optimization


2.098/6.255/ Optimization Methods Practice True/False Questions

IMPLEMENTATION OF INTERIOR POINT METHODS

A priori bounds on the condition numbers in interior-point methods

Lecture 17: Primal-dual interior-point methods part II

Largest dual ellipsoids inscribed in dual cones

A Second-Order Path-Following Algorithm for Unconstrained Convex Optimization

Primal-Dual Symmetric Interior-Point Methods from SDP to Hyperbolic Cone Programming and Beyond

A Weighted-Path-Following Interior-Point Algorithm for Second-Order Cone Optimization

ON THE DYNAMICAL SYSTEMS METHOD FOR SOLVING NONLINEAR EQUATIONS WITH MONOTONE OPERATORS

A Solution Method for Semidefinite Variational Inequality with Coupled Constraints

4TE3/6TE3. Algorithms for. Continuous Optimization

Interior-point methods Optimization Geoff Gordon Ryan Tibshirani

Interior Point Methods in Mathematical Programming

Primal-dual path-following algorithms for circular programming

Chapter 6 Interior-Point Approach to Linear Programming

12. Interior-point methods

New Primal-dual Interior-point Methods Based on Kernel Functions

A Polynomial Column-wise Rescaling von Neumann Algorithm

A projection algorithm for strictly monotone linear complementarity problems.

Second-order cone programming

GYULA FARKAS WOULD ALSO FEEL PROUD

EXPLICIT SOLUTION TO MODULAR OPERATOR EQUATION T XS SX T = A

Lecture 5. Theorems of Alternatives and Self-Dual Embedding

Interior Point Methods: Second-Order Cone Programming and Semidefinite Programming

McMaster University. Advanced Optimization Laboratory. Title: Computational Experience with Self-Regular Based Interior Point Methods

Absolute value equations

Limiting behavior of the central path in semidefinite optimization

More First-Order Optimization Algorithms

IMPLEMENTING THE NEW SELF-REGULAR PROXIMITY BASED IPMS

More Powerful Tests for Homogeneity of Multivariate Normal Mean Vectors under an Order Restriction

Lecture: Cone programming. Approximating the Lorentz cone.

Lecture 5. The Dual Cone and Dual Problem

The Nearest Doubly Stochastic Matrix to a Real Matrix with the same First Moment

The continuous d-step conjecture for polytopes

A semidefinite relaxation scheme for quadratically constrained quadratic problems with an additional linear constraint

1 Introduction Semidenite programming (SDP) has been an active research area following the seminal work of Nesterov and Nemirovski [9] see also Alizad

WEAK CONVERGENCE OF RESOLVENTS OF MAXIMAL MONOTONE OPERATORS AND MOSCO CONVERGENCE

On matrix equations X ± A X 2 A = I

Improved Full-Newton-Step Infeasible Interior- Point Method for Linear Complementarity Problems

An interior-point trust-region polynomial algorithm for convex programming

CONVERGENCE OF A SHORT-STEP PRIMAL-DUAL ALGORITHM BASED ON THE GAUSS-NEWTON DIRECTION

Interior-point methods Optimization Geoff Gordon Ryan Tibshirani

POLYNOMIAL OPTIMIZATION WITH SUMS-OF-SQUARES INTERPOLANTS

Transcription:

ANZIAM J. 49(007), 59 70 A NEW PROXIMITY FUNCTION GENERATING THE BEST KNOWN ITERATION BOUNDS FOR BOTH LARGE-UPDATE AND SMALL-UPDATE INTERIOR-POINT METHODS KEYVAN AMINI and ARASH HASELI (Received 6 December, 006; revised 6 August, 007) Abstract Interior-Point Methods (IPMs) are not only very effective in practice for solving linear optimization problems but also have polynomial-time complexity. Despite the practical efficiency of large-update algorithms, from a theoretical point of view, these algorithms have a weaker iteration bound with respect to small-update algorithms. In fact, there is a significant gap between theory and practice for large-update algorithms. By introducing self-regular barrier functions, Peng, Roos and Terlaky improved this gap up to a factor of log n. However, checking these self-regular functions is not simple and proofs of theorems involving these functions are very complicated. Roos et al. by presenting a new class of barrier functions which are not necessarily self-regular, achieved very good results through some much simpler theorems. In this paper we introduce a new kernel function in this class which yields the best known complexity bound, both for large-update and small-update methods. 000 Mathematics subject classification: primary 90C05; secondary 90C5. Keywords and phrases: linear optimization, interior-point method, primal-dual method, kernel function, large-update, small-update, polynomial complexity.. Introduction We deal with the standard linear optimization problem min { c x : Ax = b; x 0 } (.) where A R m n is a real m n matrix with rank.a/ = m, b R m and c R n.the dual problem of (.) isgivenby max { b y : A y + s = c; s 0 } : (.) Department of Sciences, Razi University, Kermanshah, Iran; email: keyvanamini353@yahoo.com or kamini@razi.ac.ir. Islamic Azad University, Kermanshah branch, Kermanshah, Iran; email: arhaseli@yahoo.com. c Australian Mathematical Society 007, Serial-fee code 446-8735/07 59

60 Keyvan Amini and Arash Haseli [] Kernel functions play an important role in defining new search directions in primaldual interior point algorithms for solving linear optimization problems. A kernel function is a univariate strictly convex function which is defined for all positive reals t and is minimal at t = where the minimal value equals zero. In other words.t/ : D R + with R ++ D, is a kernel function when it satisfies./ =./ = 0;.t/ >0: Moreover,.t/ has the barrier property, that is,.t/ goes to infinity if either t 0 or t. Note that the above properties imply that.t/ is completely determined by its second derivative.t/ = t ¾. /d d¾: This kernel function may be extended to a positive n-dimensional vector ¼ by n 9.¹/ =.¹ i /; (.3) i= yielding the scaled barrier function 9.¹/. Note that the barrier function 9.¹/ is nonnegative, and zero if and only if ¹ is a vector of ones. Therefore, the value of the barrier function can be considered as a measure for the closeness of x and.y; s/ to the ¼-centres of (.) and (.). Hence some authors also call 9.¹/ a proximity function. In the next section we briefly describe how any such barrier function defines a primal-dual interior-point method. The iteration bound for so-called large-update methods is obtained by showing that each iteration decreases 9.¹/ by a sufficient amount. Table gives some examples of kernel functions that have been analyzed so far, and the complexity results for the corresponding algorithms. Note that all kernel functions in this table depend on a parameter and the mentioned iteration bounds in the table occur by choosing this parameter as indicated in the third column of the table. The first kernel function, the so-called self-regular function, was introduced and analyzed by Peng et al. Also the second kernel function is self regular and is the special case of the self regular function p;q.t/, forp =. The third function is not self regular and has been proposed and analyzed by Bai et al. A surprising feature of this kernel function is that it is finite for t = 0, a property which separates it from self-regular functions, because self-regular functions become unbounded when t approaches zero. Also Bai et al. introduced the following kernel function in []: t + e=t e : e

[3] A new proximity function 6 TABLE. Kernel functions with the best known iteration bound for large-update methods. The iteration bound for all kernel functions is O. n log n/ log.n=ž/. Kernel function Parameter References t t t t + t q q, q > q = log n [5, 6] + t q q.q / q.t /, q > q + ( e. t/ ) ; >0 = O.log n/ [] t + e q.=¾ / d¾, q q = O.log. + n// [3] They showed that the iteration bound for the corresponding algorithm is O ( n.log n/ log.n=ž/ ). This bound is a factor log n worse than the bound in Table. In this paper we introduce a new kernel function as follows:.t/ = t + eq.=t / ; q : (.4) q We show that this kernel function yields the best-known iteration bound for largeupdate methods, that is, O. n log n/ log.n=ž/. Figure depicts the graph of.t/. 0 8 6 4 0 0 0.5.5.5 3 3.5 4 FIGURE. The graph of for q = 4. The paper is organized as follows. In Section, we briefly recall how a given kernel function defines a primal-dual interior-point algorithm. In Section 3, we describe the

6 Keyvan Amini and Arash Haseli [4] simple conditions on the kernel function that define a general class of kernel functions introduced by Bai et al. in []. Then we show that the new kernel function introduced in (.4) satisfies these conditions. In what follows, we use the general scheme for analyzing the generic algorithm, as presented in []. We obtain the iteration bounds for both large-update and small-update methods based on a new kernel function in Section 4 and finally, Section 5 contains some concluding remarks.. The generic primal-dual interior-point algorithm It is well known that finding an optimal solution of (.) and (.) is equivalent to solving the following system: Ax = b; x 0; A y + s = c; s 0; xs = 0: The basic idea of the primal-dual IPMs is to replace the third equation, the socalled complementarity condition, by the parameterized equation xs = ¼e, with ¼>0. Thus we consider the following system: Ax = b; x 0; A y + s = c; s 0; xs = ¼e: We assume that the primal and dual problem in (.) and (.) satisfies the interiorpoint condition (IPC), that is, there exists.x 0 ; s 0 ; y 0 / such that Ax 0 = b; x 0 > 0; A y 0 + s 0 = c; s 0 > 0: It is well known that the IPC can be assumed without loss of generality. By using the self-dual embedding model, we will assume that x 0 = s 0 = e [4, 7]. If rank.a/ = m and the IPC holds, then for each ¼>0 the above parameterized system has a unique solution. We denote this solution as ( x.¼/; y.¼/; s.¼/ ) and call x.¼/ the ¼-centre of (.)and ( y.¼/; s.¼/ ) the ¼-centre of (.). The set of ¼-centres (with ¼ running through all positive real numbers) gives a homotopy path, which is called the central path of (.) and (.). If ¼ 0 then the limit of the central path exists and since the limit point satisfies the complementarity condition, the limit yields to optimal solutions for (.) and (.). If ( x.¼/; y.¼/; s.¼/ ) is known for some positive ¼, then we decrease ¼ to ¼ :=. /¼,forsomefixed.0; / and solve the following

[5] A new proximity function 63 system: Ax = 0; A y + s = 0; sx + xs = ¼e xs: (.) This system uniquely defines a search direction.x;s;y/. This direction approximates the next ¼-centre. Hence all IPMs follow the central path approximately. Now we define xs ¹ = ¼ : One can easily check that System (.) which defines the search direction can be rewritten as follows: Ād x = 0; Ā y + d s = 0; (.) d x + d s = ¹ ¹; where d x = ¹x=x, d s = ¹s=s and Ā = AV X, with V = diag.¹/ and X = diag.x/. The third equation in (.) is called the scaled centring equation. The right-hand side in the scaled centring equation equals minus the gradient of the basic logarithmic barrier function, that is, 9 c.¹/ = n i= ( ) v i log v i : The basic idea in IPMs is to replace the scaled barrier function 9 c.¹/ by an arbitrary strictly convex function 9.¹/, such that 9.¹/ is minimal at ¹ = e with 9.e/ = 0. Thus System (.) converts to Ād x = 0; Ā y + d s = 0; d x + d s = 9.¹/: Since the vectors d x and d s belong to the null and row spaces of the matrix Ā, these vectors are orthogonal. By taking a suitable step size Þ.0; /, these search directions construct a new triple.x + ; y + ; s + / with x + = x + Þx; y + = y + Þy; s + = s + Þs:

64 Keyvan Amini and Arash Haseli [6] We repeat the procedure until we find an iterate in a certain neighbourhood of ( x.¼/; y.¼/; s.¼/ ). Then ¼ is again reduced by the factor and we apply Newton s method targeting the new ¼-centres, and so on. This process is repeated until ¼ is small enough and at this stage we have found an ž-solution of the problems (.) and (.). The generic form of the algorithm is shown in Figure. Input: A threshold parameter >0; An accuracy parameter ž>0; A fixed barrier update parameter ;0 <<; begin x := e; s := e; ¼ := ; while n¼ >ždo begin ¼ :=. /¼; while 9.¹/ > do begin x + := x + Þx; y + := y + Þy; ¹ := xs ; ¼ end end end FIGURE. A generic primal-dual algorithm for Lo 3. A general class of the kernel functions In [] Bai, El Ghami and Roos introduced a general class of kernel functions by using the following conditions: t.t/ +.t/ >0; t < ; (3.) t.t/.t/ >0; t > ; (3.).t/ <0; t > 0; (3.3).t/.t/.t/ >0; t < ; (3.4).t/.þt/ þ.t/.þt/ >0; t > ;þ >: (3.5) The kernel function.t/ is called eligible if it satisfies (3.) and (3.3) to(3.5). It was shown in [] that(3.) and (3.3) imply (3.5). So a kernel function.t/ is

[7] A new proximity function 65 eligible if it satisfies conditions (3.) to(3.4). First we show that the kernel function introduced in (.4) satisfies these conditions. The first three derivatives of.t/ are shown in the following table: TABLE. and its derivatives..t/ t + eq=t e q qe q.t/ t eq.=t / t.t/ + q + t e q.=t / t 4.t/ q + 6qt + 6t t 6 e q.=t / Since for 0 < t < andq [ q + t t.t/ +.t/ = t + ] ( q + t e q.=t / = t + t 3 t t 3 Condition (3.) is satisfied. Furthermore, for t > ) e q.=t / > 0; t.t/.t/ = q + 3t e q.=t / > 0: t 3 So (3.) is satisfied as well. Also from Table it is obvious that (3.3) is satisfied. Finally for satisfying (3.4), it is seen that.t/.t/.t/ = [ t 4 +.q + t/e q.=t /] + e q.=t /.q + 6qt + 6t / ( t 3 e q.=t /) t 8 : If we set A = e q.=t /, after some elementary reduction we have [.t/.t/.t/ = A t + qt + q ] + A [ 4t + 0qt + q t 3] + t 8 : t 8 On the other hand, because 0 < t <, it follows that =t > 0 and for q we have A = e q.=t / > :

66 Keyvan Amini and Arash Haseli [8] So we have.t/.t/.t/ >0 and hence condition (3.4) holds. Therefore.t/ lies in the general class introduced in []. Having such a kernel function, we can now construct the barrier function 9.¹/ by (.3). The function 9.¹/ not only serves to define a search direction, but also as a measure of closeness of the current iterates to the ¼-centre. We use the norm-based proximity measure Ž.¹/ defined by Ž.¹/ = 9.¹/ = n..¹ i //: Since 9.¹/ is strictly convex and minimal at ¹ = e wehave 9.¹/ = 0 Ž.¹/ = 0 ¹ = e: In other words, the proximity measure Ž.¹/ is zero if and only if the current iterates are the ¼-centres. For any kernel function in the mentioned class, iteration bounds for both small-update and large-update methods can be obtained by using the following scheme: Step. Solve the equation.t/= = s to get ².s/, the inverse function of.t/=, t.0; ]. If the equation is hard to solve, derive a lower bound for ².s/. Step. Calculate the decrease of 9.¹/ in terms of Ž for the default step size Þ = = ".².Ž// from f. Þ/ = Ž = ".².Ž//. Step 3. Solve the equation.t/ = s to get Q.s/, the inverse function of.t/, t. If the equation is hard to solve, derive lower and upper bound for Q.s/. Step 4. Derive a lower bound for Ž in term of 9.¹/ by using Ž.¹/ ( Q.9.¹// ) =. Step 5. Using the result of Steps 3 and 4 find positive constants and, with.0; ], such that f. Þ/ 9.¹/. Step 6. Calculate the upper bound for 9 0 from ( ) Q. =n/ 9 0 L.n;; /= n : i= Step 7. Derive an upper bound for the total number of iterations from 9 0 log n ž : Step 8. To calculate a complexity bound for large-update algorithms set = O.n/ and =./ and for small-update method algorithms set = O./ and =.= n/. In the next section, we apply this scheme to obtain an iteration bound for the algorithm generated by the kernel function introduced in (.4).

[9] A new proximity function 67 4. Iteration bounds of the algorithm Since we are unable to get explicit expressions for the inverse functions ² and Q in the above steps, we recall two lemmas from []. LEMMA 4.. Let Q :[0; ] [; ] be the inverse function of.t/ for t. Then we have + s Q.s/ + s; s 0: LEMMA 4.. Let b.t/ be the barrier term of.t/ (.t/ =.t /= + b.t/) and let ² :[0; /.0; ] be the inverse function of the restriction of b.t/ to the interval.0; ]. Then one has ².s/ ². + s/: Now by using these two lemmas, we derive some bounds for ² and Q. Step. From the equation b.t/ = s we have eq.=t / b.t/ = = s; ( t ) e q.=t / = st q t = log s + logt: From 0 < t we find that log t and as a result q ( ) t log s t = ².s/ + q log s : Now by using Lemma 4.,since².s/ ². + s/, we derive a lower bound for ².s/ as follows: ².s/ ; s 0: (4.) + q log. + s/ Step. The function.t/ is monotonically decreasing, hence Ž f. Þ/.².Ž// Ž.². + 4Ž// : Putting t = ². + 4Ž/,wehavet and can write f. Þ/ Ž.t/ Ž : (4.) +.q + t/t 4 eq.=t /

68 Keyvan Amini and Arash Haseli [0] Note that t = ². + 4Ž/. + 4Ž/ = eq.=t / b.t/ = : t Substituting this equation in (4.), we have Ž f. Þ/ +.q + t/t. + 4Ž/ Ž + 3qt. + 4Ž/ : On the other hand, t = ². + 4Ž/ ².Ž/ ( + q log. + 4Ž/ ) : So Step 3. Ž f. Þ/ + 3q. + 4Ž/. + q log. + 4Ž// : (4.3) By Lemma 4. the inverse function of.t/ for t [; / satisfies +.t/ Q..t// +.t/: (4.4) Step 4. Using Ž.¹/ ( Q.9.¹// ) =, we have ( Q.9.¹// ) ( ) Ž e q.= +9 / + 9 : (4.5). + 9/ Note that 9, so. + 9/ 3= e q.= +9 / and + 9 39. So we can write Ž ( ) 9 + 9 = 9 9 + 9 + 9 39 3 : (4.6) Step 5. Let 9 0 9 3. We deduced that Ž and 9 3Ž Ž. Now by using (4.3)wehave Ž f. Þ/ 6qŽ. + q log. + 4Ž// Ž 6q. + q log. + 4Ž// ; that is, 9 f. Þ/ 48q ( + q log. + 9 0 / ) : (4.7) Thus it follows that 9 k+ 9 k.9 k / ; k = 0; ;:::;K ; (4.8) where = 48q ( + q log. + 9 0 / ) ; = and K denotes the number of inner iterations.

[] A new proximity function 69 Step 6. From Lemma 4. we have Q. =n/ + =n. As a consequence ( ) ( ) Q. =n/ + =n 9 0 L.n;; /= n n : Since.t/.t /=fort, 9 0 n =n + =n = + n : (4.9) Step 7. By inequality (4.8), the number of inner iterations is bounded by ( ( K 9 0 = 96q + q log + )) 9 0 9 = 0 : Substituting (4.9) in this inequality gives + ( n K 96q + q log + + ) = n : Thus the total number of iterations is bounded above by K log n ž 96q + n + q log + ( + ) = n log n ž : (4.0) Step 8. For large-update methods set = O.n/ and =./. As a consequence, 9 0 + n = O.n/: By choosing + n q = log + = O.log n/ the total iteration bound in (4.0) becomes O ( n.log n/ ) log n ž : Setting = O./ and =.= n/ for small-update methods, we can obtain the best known bound as follows. By Lemma.4 in [], we have.t/ <./.t / =, t >. Since./ = q + 3, we achieve an upper bound for 9 0 of 9 0 = n (./ + =n ( ) + n = q + 3 ; ) n./ ( ) + =n

70 Keyvan Amini and Arash Haseli [] where we also use that. Now 9 0 = O.q/ and the iteration bound becomes O ( q qn ) log n ž : By choosing the parameter q as a constant which is independent of n, this is the best bound for small-update methods. 5. Concluding remarks In this paper, we introduced a new kernel function and showed that it generates the best possible iteration bounds, both for small-update and for large-update methods. This paper was inspired as a result of the work on primal-dual interior-point methods (IPMs) for linear optimization based on kernel functions and the scheme for analyzing such methods. The new kernel function is a parameterized version of a kernel function introduced in []. References [] Y. Q. Bai, M. El Ghami and C. Roos, A new efficient large-update primal-dual interior-point method based on a finite barrier, SIAM J. Optim. 3 (003) 766 78 (electronic). [] Y. Q. Bai, M. El Ghami and C. Roos, A comparative study of new barrier functions for primal-dual interior-point algorithms in linear optimization, SIAM J. Optim. 5 (004) 0 8. [3] Y. Q. Bai, J. Guo and C. Roos, A new kernel function yielding the best known iteration bounds for primal-dual interior-point algorithms, 006, working paper, available at http://www.isa.ewi.tudelft.nl/~roos/wpapers.html. [4] N. Megiddo, Pathways to the optimal set in linear programming, in Progress in Mathematical Programming: Interior Point and Related Methods (ed. N. Megiddo), Identical version in: Proceedings of the 6th Mathematical Programming Symposium of Japan, Nagoya, Japan, pages -35, 986, (Springer Verlag, New York, 989), 3 58. [5] J. Peng, C. Roos and T. Terlaky, Self-regular functions and new search directions for linear and semidefinite optimization, Math. Program. 93 (00) 9 7. [6] J. Peng, C. Roos and T. Terlaky, Self-Regularity. A New Paradigm for Primal-Dual Interior-Point Algorithms (Princeton University Press, Princeton, New Jersey, 00). [7] C. Roos, T. Terlaky and J.-Ph. Vial, Theory and Algorithms for Linear Optimization. An Interior Point Approach (John Wiley & Sons, Chichester, 997).