Central Paths in Semidefinite Programming, Generalized Proximal Point Method and Cauchy Trajectories in Riemannian Manifolds

Similar documents
On the Convergence of the Entropy-Exponential Penalty Trajectories and Generalized Proximal Point Methods in Semidefinite Optimization

On the convergence of the entropy-exponential penalty trajectories and generalized proximal point methods in semidefinite optimization

On well definedness of the Central Path

Steepest descent method on a Riemannian manifold: the convex case

DEPARTMENT OF MATHEMATICS

Monotone Point-to-Set Vector Fields

Hessian Riemannian Gradient Flows in Convex Programming

Primal-dual relationship between Levenberg-Marquardt and central trajectories for linearly constrained convex optimization

Using Schur Complement Theorem to prove convexity of some SOC-functions

Largest dual ellipsoids inscribed in dual cones

Steepest descent method on a Riemannian manifold: the convex case

Steepest descent method with a generalized Armijo search for quasiconvex functions on Riemannian manifolds

Proximal Point Method for a Class of Bregman Distances on Riemannian Manifolds

Research Note. A New Infeasible Interior-Point Algorithm with Full Nesterov-Todd Step for Semi-Definite Optimization

Dual convergence of the proximal point method with Bregman distances for linear programming

Limiting behavior of the central path in semidefinite optimization

A PREDICTOR-CORRECTOR PATH-FOLLOWING ALGORITHM FOR SYMMETRIC OPTIMIZATION BASED ON DARVAY'S TECHNIQUE

Unconstrained Steepest Descent Method for Multicriteria Optimization on Riemannian Manifolds

A convergence result for an Outer Approximation Scheme

A path following interior-point algorithm for semidefinite optimization problem based on new kernel function. djeffal

A Second-Order Path-Following Algorithm for Unconstrained Convex Optimization

arxiv: v1 [math.oc] 2 Apr 2015

Supplement: Universal Self-Concordant Barrier Functions

A proximal-like algorithm for a class of nonconvex programming

Lecture 7: Semidefinite programming

2 Preliminaries. is an application P γ,t0,t 1. : T γt0 M T γt1 M given by P γ,t0,t 1

Lecture 5. The Dual Cone and Dual Problem

OPTIMALITY CONDITIONS FOR GLOBAL MINIMA OF NONCONVEX FUNCTIONS ON RIEMANNIAN MANIFOLDS

Primal-dual IPM with Asymmetric Barrier

Semidefinite Programming

Lecture 1. 1 Conic programming. MA 796S: Convex Optimization and Interior Point Methods October 8, Consider the conic program. min.

On Total Convexity, Bregman Projections and Stability in Banach Spaces

A sensitivity result for quadratic semidefinite programs with an application to a sequential quadratic semidefinite programming algorithm

A Unified Approach to Proximal Algorithms using Bregman Distance

L. Vandenberghe EE236C (Spring 2016) 18. Symmetric cones. definition. spectral decomposition. quadratic representation. log-det barrier 18-1

Kantorovich s Majorants Principle for Newton s Method

PRIMAL-DUAL INTERIOR-POINT METHODS FOR SELF-SCALED CONES

2.1. Jordan algebras. In this subsection, we introduce Jordan algebras as well as some of their basic properties.

Real Symmetric Matrices and Semidefinite Programming

A full-newton step infeasible interior-point algorithm for linear programming based on a kernel function

A semidefinite relaxation scheme for quadratically constrained quadratic problems with an additional linear constraint

A hybrid proximal extragradient self-concordant primal barrier method for monotone variational inequalities

Lecture 5. Theorems of Alternatives and Self-Dual Embedding

Variational inequalities for set-valued vector fields on Riemannian manifolds

A splitting minimization method on geodesic spaces

Some Properties of the Augmented Lagrangian in Cone Constrained Optimization

Nonsymmetric potential-reduction methods for general cones

Summer School: Semidefinite Optimization

Primal-Dual Geometry of Level Sets and their Explanatory Value of the Practical Performance of Interior-Point Methods for Conic Optimization

Weak sharp minima on Riemannian manifolds 1

The Trust Region Subproblem with Non-Intersecting Linear Constraints

A Primal-Dual Second-Order Cone Approximations Algorithm For Symmetric Cone Programming

1 Directional Derivatives and Differentiability

Agenda. Interior Point Methods. 1 Barrier functions. 2 Analytic center. 3 Central path. 4 Barrier method. 5 Primal-dual path following algorithms

Positive semidefinite matrix approximation with a trace constraint

A new primal-dual path-following method for convex quadratic programming

Fidelity and trace norm

Key words. Complementarity set, Lyapunov rank, Bishop-Phelps cone, Irreducible cone

Advances in Convex Optimization: Theory, Algorithms, and Applications

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Interior Point Methods: Second-Order Cone Programming and Semidefinite Programming

A Simple Derivation of a Facial Reduction Algorithm and Extended Dual Systems

12. Interior-point methods

POLYNOMIAL OPTIMIZATION WITH SUMS-OF-SQUARES INTERPOLANTS

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

Lecture: Introduction to LP, SDP and SOCP

arxiv: v1 [math.na] 29 Oct 2010

Exact SDP Relaxations for Classes of Nonlinear Semidefinite Programming Problems

A QUADRATIC CONE RELAXATION-BASED ALGORITHM FOR LINEAR PROGRAMMING

A PRIMAL-DUAL INTERIOR POINT ALGORITHM FOR CONVEX QUADRATIC PROGRAMS. 1. Introduction Consider the quadratic program (PQ) in standard format:

08a. Operators on Hilbert spaces. 1. Boundedness, continuity, operator norms

More First-Order Optimization Algorithms

CONVEX OPTIMIZATION OVER POSITIVE POLYNOMIALS AND FILTER DESIGN. Y. Genin, Y. Hachez, Yu. Nesterov, P. Van Dooren

An Infeasible Interior Proximal Method for Convex Programming Problems with Linear Constraints 1

Primal-dual path-following algorithms for circular programming

New stopping criteria for detecting infeasibility in conic optimization

How to Reach his Desires: Variational Rationality and the Equilibrium Problem on Hadamard Manifolds

Self-concordant functions for optimization on smooth manifolds

An E cient A ne-scaling Algorithm for Hyperbolic Programming

A double projection method for solving variational inequalities without monotonicity

H-convex Riemannian submanifolds

ON THE ARITHMETIC-GEOMETRIC MEAN INEQUALITY AND ITS RELATIONSHIP TO LINEAR PROGRAMMING, BAHMAN KALANTARI

On the complexity of the hybrid proximal extragradient method for the iterates and the ergodic mean

A New Class of Polynomial Primal-Dual Methods for Linear and Semidefinite Optimization

Geometric problems. Chapter Projection on a set. The distance of a point x 0 R n to a closed set C R n, in the norm, is defined as

A UNIFIED CHARACTERIZATION OF PROXIMAL ALGORITHMS VIA THE CONJUGATE OF REGULARIZATION TERM

E5295/5B5749 Convex optimization with engineering applications. Lecture 5. Convex programming and semidefinite programming

POLARS AND DUAL CONES

Affine iterations on nonnegative vectors

Lecture: Smoothing.

A priori bounds on the condition numbers in interior-point methods

4TE3/6TE3. Algorithms for. Continuous Optimization

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 2

An adaptive accelerated first-order method for convex optimization

arxiv: v1 [math.oc] 26 Sep 2015

A new Primal-Dual Interior-Point Algorithm for Second-Order Cone Optimization

Constrained Optimization Theory

Primal-Dual Symmetric Interior-Point Methods from SDP to Hyperbolic Cone Programming and Beyond

The Simplest Semidefinite Programs are Trivial

An accelerated non-euclidean hybrid proximal extragradient-type algorithm for convex concave saddle-point problems

Transcription:

Central Paths in Semidefinite Programming, Generalized Proximal Point Method and Cauchy Trajectories in Riemannian Manifolds da Cruz Neto, J. X., Ferreira, O. P., Oliveira, P. R. and Silva, R. C. M. February 01, 2006 Abstract The relationships among central path in the context of semidefinite programming, generalized proximal point method and Cauchy trajectory in Riemannian manifolds is studied in this paper. First it is proved that the central path associated to the general function is well defined. The convergence and characterization of its limit point is established for functions satisfying a certain continuous property. Also, the generalized proximal point method is considered, and it is proved that the corresponding generated sequence is contained in the central path. As a consequence, both converge to the same point. Finally, it is proved that the central path coincides with the Cauchy trajectory in the Riemannian manifold. Keywords: central path, generalized proximal point methods, Cauchy trajectory, semidefinite programming, Riemannian manifold. 1 Introduction The extension of concepts and techniques from linear programming to semidefinite programming became attractive after the seminal works due to Alizadeh (1995) and Nesterov and Nemirovskii (1994). It is well known that the concept of central path, with respect to log barrier function, is very important in several subjects including linear programming and semidefinite programming, see for example Güler (1994) and Todd (2001). The central path for semidefinite programming problems DM, Universidade Federal do Piauí, Teresina, PI 64049-500, BR (Email: jxavier@ufpi.br). This author was supported in part by PRONEX(CNPq). IME, Universidade Federal de Goiás, Goiânia, GO 74001-970, BR (Email: orizon@mat.ufg.br). This author was supported in part by FUNAPE/UFG and PRONEX(CNPq). COPPE-Sistemas, Universidade Federal do Rio de Janeiro, Rio de Janeiro, RJ 21945-970, BR (Email: poliveir@cos.ufrj.br). This author was supported in part by CNPq. ICE, Universidade Federal de Amazonas, Manaus, AM 69077-000, BR (Email: rmesquita@ufam.edu.br). This author was supported in part by CAPES-PICDT. 1

converges, see Halická et al. (2002). More general, Graña Drummond and Peterzil (2002) established its convergence for analytic convex nonlinear semidefinite programming problems. However, the central path does not converges to the analytic center of the solution set, see Halická et Al. (2002). Partial characterizations of the limit point has been given by Luo et al. (1998), Sporre and Forsgren (2002) and Halická et al. (2005). Several generalizations of the classical proximal point method studied in Rockafellar (1976) have been considered, due to the important role they play in the development of Augmented Lagrangian algorithms. Early works include Censor and Zenios (1992), Chen and Teboulle (1993), Eckstein (1993), Iusem (1998) and Doljansky and Teboule (1998) who introduced a generalized proximal method for unconstrained convex semidefinite programming problems. Extensions of concepts and techniques from Euclidean space to Riemannian manifold are natural. It has been done frequently in the last few years, with theoretical objectives and also in order to obtain effective algorithms of optimization on Riemannian manifold setting. Severals works dealing with this issue include Karmarkar (1990), Smith (1994), Udriste (1994), Rapcsák, T. and Thang (1996), Rapcsák (1997), da Cruz Neto et al. (1998), Ferreira and Oliveira (1998), Ferreira and Oliveira (2002), Nesterov and Todd (2002) and Nesterov and Nemirovski (2003). A couple of paper have dealt with the behavior of the Cauchy trajectories in Riemannian manifolds including Karmarkar (1990), Helmke and Moore (1994), Balogh et al. (2004) and Alvarez et al. (2004). The central path with respect to general barrier function, for monotone variational inequality problem, has been considered by Iusem et al. (1999) and its well definition and convergence properties was obtained. Characterizations of the limit point for some specific problems including linear programming was given, i.e., it was proved that the central path converges to the analytic center of the solution set. Also, Iusem et al. (1999) provided a connection among central path, generalized proximal point sequence and Cauchy trajectory (or gradient trajectories) in Riemannian manifold. It was showed that in some cases, including linear programming, these three concepts are in a certain way equivalent. In particular, this relationship allowed to show that the generalized proximal point sequence converges to the analytic center of the solution set. In this paper we will prove the equivalence among three concepts, namely, central path, generalized proximal point sequence and Cauchy trajectory in Riemannian manifold, in the context of semidefinite programming. The results obtained are natural extensions of the results of Iusem et al. (1999). We begin by studying the central path for semidefinite programming problems associated to the general function. By assuming that this function satisfies some specific properties, we prove that the central path is well defined. The convergence and characterization is established for functions satisfying a certain continuous property, i.e, we prove that the central path converges to the analytic center of the solution set. After the study of the central path we obtain its equivalence with the generalized proximal point sequence and Cauchy trajectory in Riemannian manifold. The organization of our paper is as follows. In Subsection 1.1, we list some basic notation and terminology used in our presentation. In Section 2, we describe the semidefinite programming problem and the basic assumptions that will be used throughout the paper. In Section 3, we introduce some assumptions in order to guarantee well definedness of the central path and establish 2

some results about it. In Section 4, we describe the proximal point method and establish its connection with the central path. In Section 5, we present the relationship among Cauchy trajectory in Riemannian manifold and central path. 1.1 Notation and terminology The following notations and results of matrix analysis are used throughout our presentation, they can be found in Horn and Johnson (1985). IR n denotes the n-dimensional Euclidean space. IR + n = {(x 1,..., x n ) IR n ; x i 0 i = 1,..., n} and IR ++ n = {(x 1,..., x n ) IR n ; x i > 0 i = 1,..., n} denote nonnegative and positive orthant, respectively. The set of all matrices n m is denoted by IR n m. The (i, j)-th entry of a matrix X IR n m is denoted by X ij and the j-th column is denoted by X j. The transpose of X IR n m is denoted by X T. The set of all symmetric n n matrices is denoted by S n. The cone of positive semidefinite (resp., definite) n n symmetric matrices is denoted by S+ n (resp., S++) n and S+ n denotes the boundary of S+. n The trace of a matrix X IR n n is denoted by tr X p i=1 X ii. Given X and Y in IR n m, the inner product between them is defined as X, Y tr X T Y = n,m i=1,j=1 X ijy ij. The Frobenius norm of the matrix X is defined as X ( X, X ) 1/2. The image (or range) space and the null space of a linear operator A will be denoted by Im(A) and Null(A) respectively; the dimension of the subspace Im(A), referred to as the rank of A, will be denoted by rank(a). Given a linear operator A : E F between two finite dimensional inner product spaces (E,, E ) and (F,, F ), its adjoint is the unique operator A : F E satisfying A(u), v F = u, A (v) E for all u E and v F. Given A 1,..., A m S n, define the linear application A : S n IR m by AX = ( A 1, X,..., A m, X ) T. (1) Note that the adjoint application A : IR m S n of A is given by A v = m i=1 v ia i. Let λ(x) = (λ 1 (X),..., λ n (X)) T denote the vector of eigenvalues of an n n matrix A. We assume that the eigenvalues are ordered, e.g., λ 1 (X) λ 2 (X)... λ n (X). Lemma 1.1. For any X, Y S n, tr(xy ) λ(x) T λ(y ). 2 Preliminaries In this section, we describe our problem and two assumptions on it that will be used throughout the paper. We consider the semidefinite programming problem (SDP) (P ) min { C, X : AX = b, X 0}, and its associated dual (D) max { b T y : A y + S = C, S 0 }, 3

where the data consist of C S n, b IR m and a linear operator A : S n IR m, the primal variable is X S n, and the dual variable consists of (S, y) S n IR m. We write F(P ) and F(D) for the sets of feasible solutions to (P ) and (D) respectively, and by F 0 (P ) and F 0 (D) its relative interior. We also write F (P ) and F (D) for the sets of optimal solutions of (P ) and (D) respectively. Throughout this paper, we assume that the following two conditions hold without explicitly mentioning them in the statements of our results. A1) A : S n IR m is a surjective linear operator; A2) F 0 (P ) and F 0 (D). Assumption A1 is not really crucial for our analysis but it is convenient to ensure that the variables S and y are in one-to-one correspondence. Assumption A2 ensures that both (P ) and (D) have optimal solutions, the optimal values of (P ) and (D) are equals and the solutions sets F (P ) and F (P ) are limited (see for example Todd (2001)). It is also important to ensure the existence of the central path. 3 Central paths in semidefinite programming In this section we describe the central path associated to a function ϕ : S n ++ IR. By assuming that ϕ satisfies some assumptions, we prove that the central path is well defined, bounded and converges. Moreover, if ϕ can be continuously extended to S n + we prove also that the central path converges to the analytic center of the solution set of the problem (P). Let ϕ : S n ++ IR be a strictly convex function and C 2. We assume this assumption in all our results without mention it. The central path to the Problem (P) with respect to ϕ is the set of points {X(µ) : µ > 0} defined by X(µ) = argmin X 0 { C, X + µϕ(x) : AX = b}, µ IR ++. (2) Some of our results require one of the following assumptions on ϕ: A3) i) The function ϕ can be continuously extended to S n + and for all α IR the sub-level set L α = {X S n + : ϕ(x) α}, is bounded; ii) For all sequence {X k } S n ++ such that lim k X k = X S n +, there holds for all X S n ++. lim ϕ(x k), X Xk =, k 4

A4) i) The function ϕ goes to + as X goes to the boundary S n + of S n +, i.e., lim ϕ(x) = +. X S+ n ii) For each V S n ++ and µ > 0 the function φ V,µ : S n ++ IR defined by φ V,µ (X) = V, X + µϕ(x) satisfies lim X + φ V,µ(X) = +. The assumptions A3 and A4 are important to assure the well definition of the central path. Now, we are going to show that the assumption A4 implies that φ V,µ has a compact sub-level. Lemma 3.1. Under assumption A4 the sub-level set K α,µ (V ) = {X S n ++ : φ V,µ (X) α}, is compact for each α IR. As a consequence, φ V,µ has a minimizer in S n ++. Proof. Let α IR. We claim that K α,µ (V ) is bounded. Indeed, assume by contradiction that K α,µ (V ) is unbounded. Then, there exists a sequence {X k } K α,µ (V ) such that lim k + X k = +. But, the assumption A4.ii, implies lim Xk + φ V,µ (X k ) = + which is an absurd, since φ V,µ (X k ) α, for all k. So, we have established the claim. Now we are going to show that K α,µ (V ) is closed. Let {X k } K α,µ (V ) be such that lim k + X k = X. Since {X k } S n ++, we have two possibilities: X S n ++ or X S n ++, where S n ++ denotes the boundary of S n ++. Since µ > 0, the assumption A4.i implies that X / S n ++. So, X S n ++. Now, the continuity of φ V,µ in S n ++ implies that α lim k + φ V,µ (X k ) = φ V,µ ( X), i.e., X Kα,µ. Thus, K α,µ (V ) is closed. Therefore K α,µ (V ) is compact, and it easy to conclude that φ V,µ has a minimizer in S n ++. The above assumptions will be applied to the function ϕ : S n ++ IR in the following examples: Example 3.1. Let ϕ : S n ++ IR be given by ϕ(x) = tr ( X ln(x) ). Clearly ϕ extends continuously to S n + with the convention that 0 ln 0 = 0. The gradient of ϕ is given by ϕ(x) = ln(x) + I. It is easy to see that the function ϕ is strictly convex and has a unique minimizer X = e I. Therefore, L α = {X S n + : ϕ(x) α} is bounded. Let us consider {X k } S n ++ such that lim k X k = X S n + and X S n ++. So, ϕ(x K ), X Xk = ln(x K ) + I, X Xk = ln(x K ), X ln(xk ), X k + I, X I, Xk n λ i ( X)λ n i (ln(x k )) ϕ(x k ) + λ i ( X) n λ i (X k ). i=1 Since X S n ++ and lim k X k = X S n + the first term of the right hand side of last inequality goes to as k goes to, and due the fact that the other ones have a finite limit we obtain that lim k ϕ(x k ), X Xk =. Hence, ϕ satisfies A3. 5 i=1 i=1

For details on properties of ln(x) see Horn and Johnson (1991). Example 3.2. Let ϕ : S n ++ IR be given by ϕ(x) = ln det ( X ). It easy to see that ϕ is strictly convex. So, for each V S n ++ and µ > 0 the function φ V,µ (X) = V, X µ ln det (X) is also strictly convex. Since φ V,µ is convex we obtain φ V,µ (X) φ V,µ (2µV 1 ) + φ V,µ (2µV 1 ), X 2µV 1, or equivalently φ V,µ (X) 1 2 V, X + µn µ ln det ( 2µV 1). As V S++ n we have lim X + V, X = + thus latter inequality implies lim X + φ V,µ (X) = +. Now, lim ln det (X) = +. X S++ n Therefore, ϕ satisfies A4. Example 3.3. Let ϕ : S n ++ IR be given by ϕ(x) = det X α, where α > 0. It is straightforward to show that the gradient and the Hessian of ϕ are given, respectively, by ϕ(x) = α det X α X 1 and 2 ϕ(x)h = α det X α ( α X 1, H X 1 + X 1 HX 1), where H S n. Hence we obtain 2 ϕ(x)h, H = α det X α ( α X 1, H 2 + X 1/2 HX 1/2 2) > 0, for all H 0. So, ϕ is strictly convex, as a consequence φ V,µ (X) = V, X + µ det X α is also strictly convex, for all V S++ n and µ > 0. Now, since µ > 0, det X α > 0 and V S++ n we have lim φ V,µ(X) lim V, X = +. X + X + Finally, as lim X S n ++ det X α = +, we have that ϕ satisfies A4. Example 3.4. Let ϕ : S n ++ IR be given by ϕ(x) = tr X 1. So, ϕ(x) = X 2 and 2 ϕ(x)h = X 2 HX 1 + X 1 HX 2, where H S n. Then, we obtain 2 ϕ(x)h, H = X 1 HX 1/2 2 + X 1/2 HX 1 2 > 0, for all H 0, which implies that ϕ is strictly convex. Thus, φ V,µ (X) = V, X + µ tr X 1 strictly convex, for all V S++ n and µ > 0. As tr X 1 > 0, µ > 0 and V S++ n we have is also lim φ V,µ (X) lim V, X = +. X S++ n X + Now, since lim X S n ++ tr X 1 = +, we obtain that ϕ satisfies A4. 6

Theorem 3.1. If ϕ satisfies A3 or A4, then the central path {X(µ) : µ > 0} is well defined and is in F 0 (P ). Proof. Take X 0 F 0 (P ) and (y 0, S 0 ) F 0 (D). For µ > 0, define φ (S0,µ) : S n ++ IR by φ (S0,µ)(X) = S 0, X + µϕ(x). First, assume that ϕ satisfies A3. It easy to see that C, X = S 0, X + b T y 0, for all X F(P ), hence (2) is equivalent to X(µ) = argmin X 0 { φ(s0,µ)(x) : AX = b, φ (S0,µ)(X) φ (S0,µ)(X 0 ) }. (3) Let us consider the sub-level set L β = {X S+ n : φ (S0,µ)(X) β}, where β = φ (S0,µ)(X 0 ). Note that L β L α = {X S+ n : ϕ(x) α}, where α = β/µ. From A3.i we have that L α is bounded. So, L β is also bounded. As φ (S0,µ) is continuous in S+ n we have that L β is compact, which implies that L β {X S+ n AX = b} is also compact. Since φ (S0,µ) is strictly convex we have that there exists a unique minimizer X(µ) F(P ) and therefore (3) is well defined. Thus the central path {X(µ) : µ > 0} is also well defined. Now, we are going to show that X(µ) F 0 (P ). Assume by contradiction that X(µ) F(P ) = {X S+ n : AX = b} and define the sequence Z k = (1 ε k )X(µ) + ε k X 0, where {ε k } is a sequence satisfying ε k (0, 1) and lim k ε k = 0. Then, as X 0 F 0 (P ), X(µ) F(P ), ε k (0, 1) and F 0 (P ) is convex, we conclude that Z k F 0 (P ) for all ε k (0, 1). Now, combining definitions of X(µ) and sequence {Z k } with convexity of ϕ we obtain The latter inequality implies that 0 C, Z k + µϕ(z k ) C, X(µ) µϕ(x(µ)) C, (Z k X(µ)) + µ ϕ(z k ), Z k X(µ) = ε k C, (X 0 X(µ)) + µ ε k 1 ε k ϕ(z k ), X 0 Z k. (1 ε k ) C, (X(µ) X 0 ) ϕ(z k ), X 0 Z k. µ Since lim k Z k = X(µ) F(P ), A3.ii implies that the right hand side of the above inequality goes to, as k goes to, however the left hand side of this inequality has a finite limit. Therefore, this contradiction implies that X(µ) F 0 (P ). Finally, assume that ϕ satisfies A4. Let us consider the sub-level set K α,µ (S 0 ) = {X S n ++ : φ (S0,µ)(X) α}, where α = φ (S0,µ)(X 0 ). Now, from Lemma 3.1 it easy to see that K α,µ (S 0 ) {X S n + AX = b} is compact. So, a similar argument used in the first part permit to conclude that the central path {X(µ) : µ > 0} is well defined and definition of K α,µ (S 0 ) implies that it is in F 0 (P ). 7

For ϕ satisfying A3 or A4, the above Theorem implies that the central path {X(µ) : µ > 0}, with respect to ϕ, is well defined and is in F 0 (P ). So, for all µ > 0, we have from (2) that for some y(µ) IR m. µ ϕ(x(µ)) = C + A y(µ), (4) Proposition 3.1. Suppose that ϕ satisfies A3 or A4. Then, the following statements hold: i) the function 0 < µ ϕ(x(µ)) is non-increasing; ii) the set {X(µ) : 0 < µ < µ} is bounded, for each µ > 0; iii) all cluster points of {X(µ) : µ > 0} are solutions of the Problem (P). Proof. The proof of the items i), ii) and iii) are similar to the proof of the Proposition 3.i, Proposition 4 and Proposition 5 of Iusem et al. (1999), respectively. Theorem 3.2. Assume that ϕ satisfies A3. Let X S+ n be the analytic center of F (P ) with respect to ϕ, i.e., the unique solution of the problem { min X ϕ(x) (S) s.t. X F (P ). Then lim µ + X(µ) = X, where {X(µ) : µ > 0} is the central path with respect to ϕ. Proof. Take X a cluster point of {X(µ) : µ > 0} and a sequence of positive numbers {µ k } such that lim k + µ k = 0 and lim k + X(µ k ) = X. Now, from (4) we have C + µ k ϕ (X(µ k )) = A y(µ k ), for some y(µ k ) IR m. So, µ k ϕ (X(µ k )), X X(µ k ) = A y(µ k ) C, X X(µ k ), for all X F (P ). Using that X X(µ k ) Null(A), this equality becomes µ k ϕ (X(µ k )), X X(µ k ) = C, X X(µ k ). Since ϕ is convex, the above equality implies that µ k (ϕ (X(µ k )) ϕ(x)) C, X C, X(µ k ). Because X F (P ) and µ k > 0, it follows from the latter inequality that ϕ (X(µ k )) ϕ(x). Now, as ϕ is continuous we can take limits, as k goes to +, in this inequality to conclude that ϕ( X) ϕ(x), for all X F (P ). Thus, any cluster point X of {X(µ) : µ > 0} is a solution of the problem (S). Therefore, since X is the unique solution of the problem (S), the central path converges to it and the theorem is proved. Theorem 3.3. The central path {X(µ) : µ > 0} with respect to the function ϕ(x) = ln det(x) converges, as µ goes to 0. Proof. See Halická et Al. (2002) or Graña Drummond and Peterzil (2002). 8

4 Central paths and generalized proximal point methods In this section we describe a generalized proximal point method to solve semidefinite programming problem and present some convergence results for it. It is worthwhile to mention that our goal in this section is to bring to semidefinite programming context the ideas of Iusem et al. (1999). We begin by defining a generalized distance. Let ϕ : S n ++ IR be a C 2 strictly convex function. Define the generalized distance D ϕ : S n ++ S n ++ IR, with respect to ϕ, as D ϕ (X, Y ) = ϕ(x) ϕ(y ) ϕ(y ), X Y. Example 4.1. For the functions of the Examples 3.1-3.4 their correspondent generalized distances are given, respectively, by: i) D ϕ (X, Y ) = tr ( X ln X X ln Y + Y X ) ; ii) D ϕ (X, Y ) = tr ( XY 1) ln det ( XY 1) n; iii) D ϕ (X, Y ) = (det X) α + (det Y ) α (α tr(xy 1 ) αn 1), where α > 0; iv) D ϕ (X, Y ) = tr(x 1 ) 2 tr(y 1 ) + tr(xy 2 ). Remark 4.1. Note that if ϕ is C 2 and strictly convex then for each fixed Y S n ++ the function D ϕ (., Y ) is also C 2 and strictly convex. Moreover, the following statements hold: 1) if ϕ satisfies A3 then D ϕ (., Y ) satisfies A3; 2) if ϕ satisfies A4 then D ϕ (., Y ) satisfies A4. Indeed, by assuming that ϕ satisfies A3, it is immediate to conclude that D ϕ (., Y ) also can be extended continuously to S n +. Now, since D ϕ (., Y ) has a minimizer and is strictly convex in S n ++, we have that the sub-level set L α = {X S n + : D ϕ (X, Y ) α} is compact, for all α IR. Hence D ϕ (., Y ) satisfies A3.i. Let us consider {X k } S n ++ such that lim k X k = X S n +. So, D ϕ (X k, Y ), X Xk = ϕ(x k ), X Xk ϕ(y ), X Xk, for all X S n ++. By A3.ii we have that the first term of the right side of last equality goes to and the second one converges, as k goes to. Therefore, D ϕ (., Y ) satisfies A3. Now, assume that ϕ satisfies A4. Let V S++ n and µ > 0. Since D ϕ (X, Y ) 0, for X, Y S++ n, and V S++ n it is easy to see that lim ( V, X + µd ϕ(x, Y )) lim V, X = +. X + X + On the other hand, if ϕ satisfies A4.i then we have lim D ϕ (X, Y ) = lim (ϕ(x) ϕ(y ) ϕ(y ), X Y ) = +. X S++ n X S++ n Therefore D ϕ (., Y ) satisfies A4, and the statements are proved. 9

Let X 0 F 0 (P ). The proximal point method with generalized distance D ϕ, for solving the problem (P ), generates a sequence {X k } S n ++ with starting point X 0 F 0 (P ) and X k+1 = arg min X S n ++ { C, X + λ k D ϕ (X, X k ) : AX = b}, where the sequence {λ k } IR ++ satisfies λ 1 k = +. (5) k=0 From now on we refer to the above sequence {X k } as proximal point sequence with respect to D ϕ, associated to {λ k } and starting point X 0. If ϕ satisfies A3 or A4, then Remark 4.1 permits to use a similar argument to prove the well-definedness of the proximal point sequence {X k }, Moreover, {X k } satisfies C + λ k ( ϕ(x k+1 ) ϕ(x k )) = A z k, (6) for some sequence {z k } in IR m and k = 0, 1, 2,. Let {X(µ) : µ > 0} be the central path for the Problem (P), with respect to the function D ϕ (., X 0 ), i.e., X(µ) = argmin X S n ++ { C, X + µd ϕ (X, X 0 ) : AX = b}, for all µ > 0. If ϕ satisfies A3 or A4, then from Remark 4.1 and Theorem 3.1 we obtain that the central path {X(µ) : µ > 0} is well defined and is the unique solution of the system for some path {y(µ) : µ > 0} in IR m. C + µ ( ϕ(x(µ)) ϕ(x 0 )) = A y(µ), (7) Proposition 4.1. Suppose that ϕ satisfies A3 or A4. Then the central path {X(µ) : µ > 0}, with respect to the function D ϕ (., X 0 ), is bounded and all its cluster points are solutions to the Problem (P). Moreover, if ϕ satisfies A3 then X(µ) converges, as µ goes to 0, to the analytic center of F (P ), with respect to D ϕ (., X 0 ). Proof. First, note that by Remark 4.1 the function D ϕ (., X 0 ) satisfies the corresponding properties to ϕ. So, the limitation of {X(µ) : µ > µ > 0} follows from Proposition 3.1. Moreover, Proposition 3.1 guarantees that all cluster points of {X(µ) : µ > 0} are solutions to the Problem (P ). Finally, if ϕ satisfies A3, then from Theorem 3.2 we obtain that {X(µ) : µ > 0} converges to the analytic center of F (P ) with respect to D ϕ (., X 0 ). Theorem 4.1. Assume that ϕ satisfies A3 or A4. Let {X(µ) : µ > 0} be the central path with respect to D ϕ (., X 0 ) and let {X k } be the proximal point sequence. If the sequence {µ k } is defined as 1 k 1 µ k = 1 λ j, for k = 1, 2. (8) j=0 10

Then X k = X(µ k ), for k = 1, 2. Moreover, for each positive decreasing sequence {µ k }, there exists a positive sequence {λ k } satisfying (5) such that the proximal sequence {X k } associated to it satisfies X k = X(µ k ). Proof. From (6) we have that X j satisfies C + λ j ( ϕ(x j+1 ) ϕ(x j )) = A z j or equivalently ( ) C + ϕ(x j+1 ) ϕ(x j ) = A zj, for j = 0, 1,... λ j λ j Summing this equality from j = 0 to k 1, letting µ k as in (8) and v k = µ k k 1 j=0 λ j 1 z j we obtain C + µ k ( ϕ(x k ) ϕ(x 0 )) = A v k, for all k 1. So, the above equality and (7) implies that X k = X(µ k ) and the first part is proved. For the second part, let {X(µ) : µ > 0} be the central path and let {y(µ) : µ > 0} be given by (7). Take a positive decreasing sequence {µ k } and define the sequences X k = X(µ k ) and y k = y(µ k ). It follows from (7) that ( ) C + ϕ(x k ) ϕ(x 0 ) = A yk. µ k µ k Now, letting λ k = ((µ k+1 ) 1 (µ k ) 1 ) 1 and z k = λ k ((µ k+1 ) 1 y k+1 (µ k ) 1 y k it easy to see from the last equality that C + λ k ( ϕ(x k+1 ) ϕ(x k )) = A z k. Therefore, since λ k = ((µ k+1 ) 1 (µ k ) 1 ) 1 satisfies (5) the result follows from the latter equality and (6). It is worth pointing out that the Theorem 4.1 is a natural extension to semidefinite programming of the Theorem 3 of Iusem et al. (1999). Theorem 4.2. Assume that ϕ satisfies A3 or A4. Let {X k } be the proximal point sequence. Then {X k } is bounded and all its cluster points are solutions of the Problem (P ). Moreover, if ϕ satisfies A3, then {X k } converges to the analytic center of solution set of the Problem (P ) with respect to D ϕ (., X 0 ). Proof. If {X k } is the proximal point sequence, then setting µ k, as in (8), we obtain from Theorem 4.1 that X k = X(µ k ). Since lim k + µ k = 0, the result follows from Proposition 4.1. Lemma 4.1. Suppose that ϕ satisfies A3 or A4. Let {X(µ) : µ > 0} be the central path to the Problem (P ) with respect to D ϕ (., X 0 ). If ϕ(x 0 ) Im A then {X(µ) : µ > 0} is also the central path to the Problem (P ) with respect to ϕ. 11

Proof. Take ϕ satisfying A3 or A4. From Remark 4.1 we have that D ϕ (., X 0 ), also satisfies A3 or A4. So, from Theorem 3.1 the central path {X(µ) : µ > 0} is well defined and satisfies C + µ ( ϕ(x(µ)) ϕ(x 0 )) = A (y(µ)), for some path {y(µ) : µ > 0} in IR m. As ϕ(x 0 ) Im A there exists y 0 IR m such that A y 0 = ϕ(x 0 ), thus combining last two equalities we obtain C + µ ϕ(x(µ)) = A (y(µ) + µy 0 ). So, {X(µ) : µ > 0} is a solution to (4), however (4) has unique solution, namely, the central path to the Problem (P ) with respect to ϕ. Lemma 4.2. Let {X(µ) : µ > 0} be the central path to the Problem (P ) with respect to the D ϕ (., X 0 ), where ϕ(x) = ln det(x). If X 1 0 Im A, then X(µ) converges as µ goes to 0. Proof. From Example 3.2 we have that ϕ(x) = ln det(x) satisfies A4. Since ϕ(x 0 ) = X 1 0 Im A we have from Lemma 4.1 that the central paths with respect to D ϕ (., X 0 ) and to ϕ are equals. Therefore the statement follows from Theorem 3.3. Theorem 4.3. Let {X(µ) : µ > 0} be the central path with respect to the D ϕ (., X 0 ), where ϕ(x) = ln det(x). Suppose that X0 1 Im A. Then the proximal point sequence {X k } with starting point X 0, converges. Proof. If {X k } is the proximal point sequence with starting point X 0, then setting µ k, as in (8), we obtain from Theorem 4.1 that X k = X(µ k ). Since lim k + µ k = 0, the result follows from Lemma 4.2. 5 Central paths and Cauchy trajectories in Riemannian manifolds In this section we are going to prove that the central path, with respect to the function ϕ for the Problem (P ), becomes a Cauchy trajectory on the Riemannian manifold endowed with the metric given by the Hessian of ϕ. This result extends to semidefinite programming context the corresponding result of linear programming, see Section 4 of Iusem et al. (1999). We begin with some basics results of Riemannian geometry. Consider the set of positive definite n n symmetric matrices S n ++ with its usual differentiable structure and endowed with the Euclidean metric.,.. The tangent space to S n ++ at X is given by T X S n ++ = {Y X; Y S n } = S n. Let ϕ : S n ++ IR be strictly convex and C 2. Define a new metric in S n ++ as U, V 2 ϕ(x) = 2 ϕ(x)u, V, 12

so that M := (S n ++, 2 ϕ) is now a Riemannian Manifold. The metric of M induces a map φ grad φ which associates to each φ C 1 ( S n ++) its gradient grad φ S n by the rule dφ X (V ) = grad φ(x), V 2 ϕ(x), where X S n ++ and V S n. It is easy to see that the gradient vector field in M is given by grad φ(x) = ( 2 ϕ(x)) 1 φ(x), where φ(x) is the Euclidean gradient vector field at X, i.e., φ(x) is the gradient with respect to the Euclidean metric. Let A : S n IR m be the linear operator as defined in (1). The assumptions A1 and A2 imply that the set F 0 (P ) = {X S n : AX = b, X S n ++} is a Riemannian submanifold of M with the induced metric and tangent space at X given by T X F 0 (P ) = {V S n ; AV = 0}. The adjoint operator of A with respect to the metric of M is ( 2 ϕ(x)) 1 A, where A : IR m S n is the usual adjoint operator of A. In this case, the orthogonal projection Π X : S n T X F 0 (P ) with respect to the metric of M is Π X = I ( 2 ϕ(x)) 1 A ( A( 2 ϕ(x)) 1 A ) 1 A. The gradient vector field of the function φ F 0 (P ) : F 0 (P ) IR, with respect to the metric of M, is given by grad φ F 0 (P ) = Π grad φ, i.e., grad φ F 0 (P ) = (I ( 2 ϕ(x)) 1 A ( A( 2 ϕ(x)) 1 A ) 1 A ) ( 2 ϕ(x)) 1 φ. (9) Finally, the Cauchy trajectory for the function φ F 0 (P ), with respect to ϕ, is the differentiable curve Z : [0, β) F 0 (P ) given by { Z (t) = grad φ F0(P )( Z(t) ) Z(0) = Z 0, for the starting point Z 0 and some β > 0. Remark 5.1. It is well known that for each Z 0 F 0 (P ), there exists β > 0 such that (10) has a unique solution Z(t) defined in [0, β). Consider the following parametrization of the central path {X(t) : t 0}, where (10) X(t) = argmin X 0 {t C, X + ϕ(x) : AX = b}. (11) Next result extends to semidefinite programming, namely, the corresponding one in Iusem et al. (1999) obtained for linear programming. 13

Theorem 5.1. Let {X(t) : t 0} be the central path with respect to ϕ, as defined in (11), and let φ(x) = C, X. If Z 0 F 0 (P ) satisfies ϕ(z 0 ) = A z 0, for some z 0 IR m, then the central path {X(t) : t 0} is a solution for (10) with X(0) = Z 0, i.e., the central path is the Cauchy trajectory for φ F 0 (P ) in F 0 (P ) with respect to ϕ and starting point Z 0. Proof. First, note that from optimality condition for (11) we have tc + ϕ(x(t)) = A y(t), for all t 0 and some y(t) IR m. So, ϕ(x(0)) = A y(0). Since ϕ(z 0 ) = A z 0 and ϕ is strictly convex we have X(0) = Z 0 and y(0) = z 0. Now, taking derivative in the above equality we obtain C + 2 ϕ(x(t))x (t) = A y (t), or equivalently ( 2 ϕ(x(t)) ) 1 C + X (t) = ( 2 ϕ(x(t)) ) 1 A y (t). (12) Applying A in this equality we have A( 2 ϕ(x(t))) 1 C + AX (t) = A( 2 ϕ(x(t))) 1 A y (t). Because X (t) T X F 0 (P ), it follows from last equality that A( 2 ϕ(x(t))) 1 C = A( 2 ϕ(x(t))) 1 A y (t). Now, due the fact that A( 2 ϕ(x(t))) 1 A is nonsingular, is easy to see from latter equality that ( y (t) = A( 2 ϕ(x(t))) 1 A ) 1 A( 2 ϕ(x(t))) 1 C. Now, substituting y (t) in (12) we obtain X (t) = (I ( ( 2 ϕ(x(t))) 1 A A( 2 ϕ(x(t))) 1 A ) ) 1 A ( 2 ϕ(x(t))) 1 C. Finally, as φ(x) = C, X it follows from last equation and (9) that X(t) satisfies (10) and the statement of the theorem is proved. The next result is a consequence of the latter theorem. Corollary 5.1. The central path {X(t) : t 0} for the problem (P) with starting point X 0 F 0 (P ), where X 0 satisfies ϕ(x 0 ) = A y 0 for some y 0 IR m, is bounded. Proof. Let φ(x) = C, X and define ψ(t) = φ ( X(t) ). It follows from Theorem 5.1 that ψ (t) = grad φ F 0 (P ) ( X(t) ), X 2 ϕ = grad φ F 0 (P ) ( X(t) ), grad φ F 0 (P ) ( X(t) ) 2 ϕ = 2 ϕ(x(t)) grad φ F 0 (P ) ( X(t) ), grad φ F 0 (P ) ( X(t) ) = ( 2 ϕ(x(t))) 1/2 grad φ F 0 (P ) ( X(t) ) 2 < 0, t (0, + ). 14

Then φ is decreasing along to the central path, which implies that {X(t) : t > 0} {X S+; n φ(x) φ(x 0 )}. Since that optimal solutions set of (P) is compact, it follows from convexity of φ that {X S+; n φ(x) φ(x 0 ), AX = b} is also compact. Therefore, {X(t) : t > 0} is bounded. 6 Final Remarks In this paper we have studied the convergence properties of central path, for semidefinite programming problems, associated to a function satisfying some specific assumptions. We have showed that the central path is well defined and bounded. Moreover, whether that function can be continuously extended to the boundary of its domain, we have proved that the central path converges to the analytic center of the solution set of the problem. Partial characterizations of the limit point of the central path with respect to the log-barrier function for semidefinite programming problems has been obtained by Sporre and Forsgren (2002), Halická et al. (2005) and and da Cruz Neto et al. (2005). For a more general class of functions, including the functions presented in the Examples 3.3 and 3.4, the convergence and characterization of the limit point of central path associated to them is an open problem. As application of the study of the central path, we have presented some convergence properties of the generalized proximal point method for semidefinite programming problems. Also, convergence results for the generalized proximal point methods associated to a class of functions including the functions presented in the Examples 3.3 and 3.4 is lacking. References [1] Alizadeh, F., Interior Point Methods in Semidefinite Programming with Applications to Combinatorial Optimization, SIAM Journal on Optimization Vol.5, No.1, pp.13-51, 1995. [2] Alvarez, F., Bolte, J., Brahic, O., Hessian Riemannian Gradient Flows in Convex Programming, SIAM Journal on Control and Optimization, Vol.43, No.2, pp.477-501, 2004. [3] Balogh, J., Csendes, T., Rapcsák, T., Some Global Optimization Problems on Stiefel Manifolds, Journal of Global Optimization Vol.30, No.1, pp.91-101, 2004. [4] Censor, Y. and Zenios, S., The Proximal Minimization Algorithm with D-Functions, Journal of Optimization Theory and Applications, Vol.73, No.3, pp.451-464, 1992. [5] Chen, G. and Teboulle, M., Convergence Analysis of a Proximal-Like Minimization Algorithm using Bregman Functions, SIAM Journal on Optimization Vol.3, No.3, pp.538-543, 1993. 15

[6] da Cruz Neto, J.X., Ferreira, O.P., Monteiro, R.D.C., Assymptotic Behavior of the Central Path for a Special Class of Degenerate SDP Problems, Mathematical Programming, Vol.103, No.3, pp.487-514, 2005. [7] da Cruz Neto, J. X.; de Lima, L. L.; Oliveira, P. R., Geodesic Algorithms in Riemannian Geometry, Balkan Journal of Geometry and Its Applications, Vol.3, No.2, pp.89-100, 1998. [8] Doljansky, M. and Teboulle, M., An Interior Proximal Algorithm and the Exponential Multiplier Method for Semidefinite Programming, SIAM Journal on Optimization, Vol.9, No.1, pp.1-13, 1998. [9] Eckstein, J., Nonlinear Proximal Point Algorithm using Bregman Functions, with Applications to Convex Programming, Mathematics of Operations Research, Vol.18, No.1, pp.202-226, 1993. [10] Ferreira, O. P. and Oliveira, P. R., Proximal Point Algorithm on Riemannian Manifolds, Optimization, Vol.51, No.2, pp.257-270, 2002. [11] Ferreira, O. P. and Oliveira, P. R., Subgradient Algorithm on Riemannian Manifolds, Journal of Optimization Theory and Applications, Vol.97, No.1, pp.93-104, 1998. [12] Graña Drummond, L.M., Peterzil, H.Y., The Central Path in Smooth Convex Semidefinite Programs, Optmization Vol.51, No. 2, pp.207-233, 2002. [13] Güler, O., Limiting Behavior of Weighted Central Paths in Linear Programming, Mathematical Programming, Vol. 65, pp.347-363, 1994. [14] Halická, M., de Klerk, E., Roos, C., Limiting Behavior of the Central Path in Semidefinite Optimization, Optimization Methods Software, Vol.20, No.1, pp.99-113, 2005. [15] Halická, M., de Klerk, E., Roos, C., On The Convergence of the Central Path in Semidefinite Optimization, SIAM Journal on Optimization Vol.12, No. 4, pp.1090-1099, 2002. [16] Helmke, U. and Moore, J.B., Optmization and Dynamical Systems, Springer Verlag, New York, New York, 1994. [17] Horn, R.A., and Johnson, C.R., Topics in Matrix Analysis, Cambridge University Press, New York, 1991. [18] Horn, R.A., and Johnson, C.R., Matrix Analysis, Cambridge University Press, New York, 1985. [19] Iusem, A.N., Svaiter, B.F., da Cruz Neto, J.X., Central Paths, Generalized Proximal Point Methods and Cauchy Trajectories in Riemannian Manifolds, SIAM Journal on Control and Optimization, Vol.37, No.2, pp.566-588, 1999. 16

[20] Iusem, A.N., On Some Properties of Generalized Proximal Point Methods for Variational Inequalities, Journal of Optimization Theory and Applications, Vol.96, No.2, pp.337-362, 1998. [21] Karmarkar, N., Riemannian Geometry Underlying Interior Point Methods for Linear Programming, in Mathematical Developments Arising from Linear Programming (J.C. Lagarias and M.J. Todd, eds.), Contemporary Mathematics, Vol. 114, American Mathematical Society, Providence, RI, pp.51-75, 1990. [22] Luo, Z-Q., Sturm, J,F., Zhang, S., Superlinear Convergence of a Symmetric Primal-Dual Path Following Algorithm for Semidefinite Programming, SIAM Journal on Optimization Vol.8, No.1, pp.59-81, 1998. [23] Nesterov, Yu. and Nemirovski, A., Central Paths and Riemannian Distances, Technical report, CORE Discussion Paper, CORE, Université Catholique de Louvain, Belgium, 2003. [24] Nesterov, Yu. and Nemirovski, A., Interior Point Polynomial Methods in Convex Programming, SIAM, Philadelphia, PA, 1994. [25] Nesterov, Y. and Todd, M.J., On the Riemannian Geometry Defined by Self-Concordant Barriers and Interior-Point Methods, Foundations of Computational Mathematics, Vol.2, No.4, pp.333-361, 2002. [26] Rapcsák, T., Smooth Nonlinear Optimization in R n, Kluwer Academic Publishers, Dordrecht, Holland, 1997. [27] Rapcsák, T. and Thang, T. T., A Class of Polynomial Variable-Metric Algorithms for Linear Optimization, Mathematical Programming, Vol.74, pp.319-331, 1996. [28] Rockafellar, R.T., Monotone Operators and the Proximal Point Algorithm, SIAM Journal on Control and Optimization, Vol.14, No.5, pp.877-898, 1976. [29] Smith, S.T., Optimization Techniques on Riemannian Manifolds, Fields Institute Communications, American Mathematical Society, Providence, Rhode Island, Vol. 3, pp.113-136, 1994. [30] Sporre, G., Forsgren, A., Characterization of the Limit Point of the Central Path in Semidefinite Programming, Technical Report TRITA-MAT-2002-OS12, Departament of Mathematics, Royal Institute of Technology, SE-100 44 Stockholm, Sweden, June 2002. [31] Todd, M.J., Semidefinite Optimization, Acta Numerica, Vol. 10, pp.515-560, 2001. [32] Udriste, C., Convex Functions and Optimization Methods on Riemannian Manifolds, Kluwer Academic Publishers, Dordrecht, Holland, 1994. 17