The Q Method for Second-Order Cone Programming

Size: px
Start display at page:

Download "The Q Method for Second-Order Cone Programming"

Transcription

1 The Q Method for Second-Order Cone Programming Yu Xia Farid Alizadeh July 5, 005 Key words. Second-order cone programming, infeasible interior point method, the Q method Abstract We develop the Q method for the Second Order Cone Programming problem. The algorithm is the adaptation of the Q method for semidefinite programming originally developed by Alizadeh, Haeberly and Overton, [3] and []. We take advantage of the special algebraic structure associated with second order cone programs to formulate the Q method. Furthermore we prove that our algorithm is globally convergent. Finally Some numerical results are presented. Introduction The Second Order Cone Programming SOCP problem is a family of convex optimization problems more general than linear programming; however it is a special case of semidefinite programming SDP problem. In [] a review of this problem and its analytic, algorithmic and numerical properties, as well a number of its applications in various areas are presented. We refer the reader to this paper and references in it for further information. In this paper we present a new interior point algorithm which is an adaptation of the Q method for semidefinite programming introduced in [3] and []. However we will not rely on the knowledge of SDP algorithms and will develop the Q method for the SOCP problem from ground up. We only mention that in [8] we analyze the Q method for optimization problems over any symmetric cone which include both the cone of positive semidefinite matrices and the second order cone. In the following sections we first develop the algebraic foundations in. In 3 we develop the Q method and derive the associated Newton direction. The global convergence analysis of this method is given in Appendix A. In 4 numerical results for both randomly generated problems, and also problems arising from Steiner tree problems are presented. Finally in 5 a modification of the Q method along with additional computational results are presented. We also present global convergence of the modified algorithm in Appendix A. Notation For the most part we use the notation of []. In particular, boldface letters such as a, b, x, etc. represent column vectors and capital letters such as A, X, etc. represent matrices. To refer to the i th entry of the vector x we write x i. For transpose of matrix or vector A we write A. The The Institute of Statistical Mathematics, Minami-Azabu, Minato-ku Tokyo , JAPAN xiay@optlab.mcmaster.ca Department of Management Science and Information Systems and Rutgers Center for Operations Research, Rutgers, the State University of New Jersey, U.S.A. alizadeh@rutcor.rutgers.edu. Research supported in part by the U.S. National Science Foundation Grant #NSF-CCR and Office of Naval Research through contract #N

2 vector of all zeros is written as 0, and the vector of all ones as. If we need to join vectors and matrices, we adapt the syntax of MATLAB and similar programming languages. Thus, is used to join matrices and vectors in a row and ; is used to join them in a column. For example, x, y x a b = = x; y and = a, b; c, d y c d If a vector is partitioned into several blocks we write x = x ;... ; x r ; therefore, x i refers to the i th block and should not be confused with x i. Also x ij means the j th entry of the i th block. For norms we use. In particular, x, x and x are the Euclidean, the and the infinity norms of x, respectively. For a vector x Diagx is a diagonal matrix whose diagonal is x. In general, if X,..., X n are a sequence of matrices, then DiagX,..., X n is a block diagonal matrix made up of X i. The simple second order cone Q n+ is defined by Q n+ = {x 0 ; x x 0 x } Vectors x in Q n+ are indexed from zero, and the subvector consisting of x through x n is represented by x. We also right x for the subvector formed from entries x through x n. Also a matrix multiplied from left by x has its columns indexed from zero. The identity matrix is I and the matrix R n+ =, 0 ; 0, I; R is indexed from zero in both rows and columns. The vector e n+ = ; 0, a zero indexed n + -dimensional vector will play a fundamental role. For reasons to become clear shortly e is called the identity element. More generally, we call any direct sum of simple second order cones the second order cone and write Q. Thus, Q = {x ;... ; x r x i Q ni+ for i =... r for simple second order cones Q ni+} In this context, e = e n+;... ; e nr+, R = Diag R n+,..., R nr+, and when we say x Q it is understood that x is decomposed into blocks: x = x ;... ; x r consistent with Q and each x i Q ni. Foundations The standard second order cone programming problem is defined as a pair of mutually dual optimization problems Primal: min s.t. c x + + c n x r A x + + A n x r = b x i Q ni+ i =,..., r Dual: max b y s.t. A i y + z i = c i i =,..., r z i Q ni+ i =,..., r where Q ni+ are simple second order cones. This problem can also be written more succinctly as Primal: min s.t. c x Ax = b x Q Dual: max s.t. b y Ay + z = c z Q where Q is direct sum of Q ni+ and x, c and z can be decomposed into blocks x i, c i and z i as in In this paper we switch back and forth between and.

3 . Duality and complementarity It can be shown that under certain constraint qualifications the optimal values of Primal and Dual problems are equal. That is at the optimal point x; y; z we have c x b y = x z = 0. Furthermore, at the optimal point a series of complementary slackness conditions analogous to those in linear programming hold. To express this complementarity conditions we first define an algebraic binary operation as follows: For x, z R ni+, both of which are indexed from zero define x z x 0 z + x z 0 x z =. x ni z 0 + x 0 z ni Note that the operation is bilinear in x and z and it is commutative, that is x z = z x. Bilinear property of implies that there is a matrix X such that x z = Xz for every z. This matrix is nothing other than the arrow-shaped matrix representation of Lorentz transformation and is defined by x0 x Arwx = x x 0 I Thus, x z = Arwxz = Arwx Arwze, where e = ; 0. Notice that e acts as the identity element with respect to, that is e x = x e = x. In the case of multiple blocks where x = x ;... ; x r and z = z i ;... ; z r x z = x z ;... ; x r z r Also, for a multiple block x, we set Arwx = Diag Arwx,..., Arwx r. Now the complementary slackness theorem for SOCP can be formulated as follows: Proposition. Complementary slackness If x = x ;..., x r is optimal for Primal problem in and y; z with z = z ;... ; z r is optimal for Dual, then for i =,..., r. x i z i = 0 The significance of the complementary slackness theorem is that the equalities for primal and dual feasibility and complementarity 3 Ax = b A y + z = c x z = 0 constitutes a square system of equations which in the absence of various degeneracies, determine the optimal solution uniquely, see [].. The traditional interior point methods in a nutshell Let Q ni+ be a simple second order cone and let x Int Q ni+. Then the function bx = ln x 0 x is a barrier function defined on the interior of Q ni+. This means that bx is defined on Int Q ni+ and tends to infinity as x tends to any point on the boundary of Q ni+. Also observe that bx is strictly and strongly convex on Int Q ni+. When Q is the direct sum of simple second order cones Q ni+, define bx = i bx i. 3

4 In traditional interior point methods, Primal in is replaced by 4 min s.t. c x + µbx Ax = b with the understanding that x is restricted to the interior of Q. Applying Karush-Kuhn-Tucker conditions to 4 and simplifying we arrive at the following system of equations: 5 Ax = b A y + z = c x z = µe where e = e ;... ; e k and each e i = ; 0. Definition The set of points x µ ; y µ ; z µ that satisfy 5 as µ varies traverse a path called the central path. Note that x i µ z i µ = µ and thus x µ z µ = µr; therefore as µ tends to zero, the duality gap x µ z µ also tends to zero, and thus x µ, y µ and z µ tend to the optimal solution. It can be shown that for every x µ and z µ on the central path the matrices Arwx µ and Arwz µ commute, but outside of the central path this need not hold. The next step in solving SOCP problems in traditional interior point methods is to apply Newton s method to the nonlinear system of equations 5. This amounts to replacing x; y; z by x+ x; y + y; z+ z in 5, with x; y; z as our unknowns. In this approach, all nonlinear terms in the s are removed and the Newton system, that is the following linear system is solved: A x = b Ax A y + z = c A y z z x + x z = µe x z Most interior point algorithms apply some linear transformation to x and z before this step in order to simplify the analysis; refer to [] and the relevant references there for details. Note that the last equation can be written as Arwz x + Arwx z = µe x z Thus in block matrix form the system of equations 6 can be written as A A I x r p r p = b Ax y = r d where r d = c A y z Arwz 0 Arwx z r c = µe x z The algorithms proceed as follows: At each iteration the system 9 is solved for a given value of µ and a new solution x new = x + α x; y new = y + β y; z new = z + β y is calculated. Then µ is reduced by some factor, and x new ; y new ; z new is used in place of x; y; z in 9. The new system is solved again, and the process is continued until we get sufficiently close to the optimal solution. With judicious choice of step lengths, reduction factor of µ and scaling of x, y, z it can be shown that the resulting algorithm or class of algorithms converge to the optimal solution in polynomial time; see references in []. There are two issues with this class of algorithms just described. When solving the system 6 we will end up solving equations with the Schur complement r c A Arwz ArwxA 4

5 Without any scaling of x and z, Arwx and Arwz do not commute and the Schur complement is not a symmetric matrix for points not on the central path. With certain scaling applied to x and z at each iteration, it is possible to transform the problem to an equivalent one where Arwx and Arwz in the transformed space commute making the Schur complement symmetric and positive semidefinite; see [] for details and references. However, since the Newton method is applied to a different system at each iteration, technically it is not Newton method any more, and in particular the Newton system may converge to a singular system. The Q method is devised to ensure that we deal with a symmetric Schur complement at each iteration while keeping the Newton system nonsingular even at the optimum..3 Further algebraic properties of second order cones Before describing the Q method we need to develop some machinery that simplifies presentation. For the moment let us assume that we are dealing with a simple second order cone Q n+ and vectors are assumed to have just one block. As we saw the binary operation x z is commutative, however the associative law does not hold in general. Instead a weaker condition holds. Writing x = x x it is straightforward to verify that x x y = x x y This identity is equivalent to stating that Arwx and Arwx commute. An important fact is that the second order cone Q is the cone of squares under operation: 0 Q n+ = {x x R n+ } There are two elementary but very important identities that are satisfied by all vectors. The first one is the quadratic identity x x 0 x + x 0 x e = 0. The polynomial λ x 0 λ + x 0 x is called the characteristic polynomial of x and plays a similar role as the characteristic polynomials of symmetric matrices. The roots of this polynomials are x 0 ± x and are called the eigenvalues of x. Only when x = 0, that is when x is a scalar multiple of the identity element e, the two eigenvalues are equal. The second important identity is x = x 0 + x x x + x 0 x x. x Defining def c = def x, c = x x, x λ = x 0 + x, λ = x 0 x, it follows that can be written as 3 x = λ c + λ c. The relationships and 3 are spectral decomposition of x. They play much the same role as the eigenvalue decomposition of symmetric matrices. Observe that c c = 0, and c c = 0 c = c and c = c, c + c = e, c = Rc and c = Rc, c, c bd Q. 5

6 Any pair of vectors {c, c } that satisfies properties 4,5, and 6 is called a Jordan frame. It can be shown that every Jordan frame {c, c } is of the form, that is Jordan frames are pairs of vectors ; ±c with c =. When a vector has distinct eigenvalues that is when x 0, the Jordan frame is determined uniquely. Therefore, the only vectors whose associated Jordan frames are ambiguous are multiples of the identity vector e. For these vectors both eigenvalues are equal, and it turns out that we can take any Jordan frame as their associated Jordan frame. This issue will cause a slight difficulty in our algorithm which we will address when we discuss the Q method. c x 0 e x c - -x - x x x Figure : Construction of Jordan frames from a vector x. We can also define the trace and determinant of x: trx = λ + λ = x 0 and detx = λ λ = x 0 x. Thus the barrier function defined earlier can be written as bx = ln det x. From the spectral decomposition we can prove that the binary operation is power associative, that is as long as we are multiplying many copies of x by itself, the order of multiplication does not matter, Thus for any nonnegative integer p we can write x p unambiguously. Another consequence is that the second order cone consists of those vectors x whose both eigenvalues are nonnegative and the interior of Q consists of vectors whose both eigenvalues are positive. Using the spectral decomposition one can extend any real valued continuous functions f defined on R to vectors: x = λ c + λ c : fx def = fλ c + fλ c whenever fλ, fλ are defined In particular, when x = λ c + λ c, we may define 9 x def = λ c + λ c when λ 0 and λ 0 It is easily verified that x x = e. Notice that the condition x z = e for a given x may be satisfied by an infinite number of vectors. In other words for the binary operation there is no unique inverse in general. Therefore when we say inverse we mean the one defined by 9. Also note that the definition of the central path in 5 is not ambiguous. It turns out that the only 6

7 vectors x for which there are more than one z with x z = µe are those with x 0 = 0. Since in 5 we assume that x IntQ it follows that x 0 > 0 and z is unique. As we saw earlier, for two vectors x and z, it is not necessarily the case that the matrices Arwx and Arwz commute. It turns out that the case when Arwx and Arwz commute is an important property that can be exploited in design of interior point algorithms, including the Q method. Therefore we present the following Definition Two vectors x and z operator commute if Arwx and Arwz commute. However two matrices commute if and only if they share a system of common eigenvectors. Moreover, eigenvalues and eigenvectors of Arwx are described by the Jordan frame associated with x. Lemma. Let the spectral decomposition of x = λ c + λ c, where λ, and c, are as in. Then eigenvalues and eigenvectors of Arwx are given as follows: i. x 0 + x is an eigenvalue with corresponding eigenvector c. ii. x 0 x is an eigenvalue with corresponding eigenvector c. iii. All other eigenvalues are equal to x 0 and their corresponding eigenvectors are any set of orthonormal vectors, orthogonal to c and c. It is fairly simple to verify this lemma. As a direct consequence we get Theorem. Two vectors x and z operator commute if and only if they share a common Jordan frame. In other words if x and z operator commute, then there is a Jordan frame {c, c } such that x = λ c + λ c and z = ω c + ω c. On the central path defined by 5 the vectors x and z operator commute since x = µz for µ > 0. Similarly it is not hard to show that at the optimum x and z operator commute. Therefore, the complementary slackness theorem may be restated as: Theorem. x = x ;... ; x r is primal optimal and y, z with z = z,..., z r is dual optimal solutions of if and only if i. there is exist Jordan frames {c i, c i } for i =,..., r and pairs of real numbers λ i, λ i and ω i, ω i such that x i = λ i c i + λ i c i and z i = ω i c i + ω i c i, for i =,..., r, and ii. λ ij ω ij = 0 for i =,..., r and j =,. It is possible to apply an orthogonal transformation to any Jordan frame {c, c } to change it to another Jordan frame {f, f }. Such transformations are essential for the Q method. More specifically if {c, c } and {f, f } are two pairs of Jordan frames, then we wish to find a linear transformation that maps c to f, c to f and the second order cone Q n+ back to itself. However, in general there may be many such transformations. To make this choice unique we choose to make this transformation the unique rotation in the plane spanned by c and f fixing the identity element e. This rotation will also map c, to f and thus the frame {c, c } to {f, f }. In particular when we use the standard frame {d, d }, defined by setting d = ; ; 0 and d = ; ; 0, then the rotation that maps the frame {c, c } to {d, d } is 0 Q x def = 0 0 x 0 x x x x xx 0 x I x x +x = c c 0 c I 4cc +c.

8 c x 0 d d e c c d c θ d x x Figure : Regardless of the dimension, a two dimensional rotation θ in the plane spanned by c and d can map a frame {c, c } to frame {d, d }. where x = x ;... ; x n, and similarly for c. Also c is the entry of c indexed by noting that c = c 0 ; c ;... is indexed from 0. Observe that this transformation really depends on the Jordan frame of x rather than x itself. Thus all vectors sharing the same Jordan frame will have the same Q x. We also note two special cases. First, if c = d, c = d, then Q x = I. Second if c = d, c = d then Q x = Q flip, where Q flip def = I Thus, Q flip corresponds to rotation of π which effectively maps d to d and d to d. It is well-known that if S is a skew-symmetric matrix then the matrix exps = k=0 Sk k! is orthogonal. However we are not interested in all orthogonal matrices, but only those that are of the form Q x for some x. It turns out that a particular subset of skew-symmetric matrices generates all Q x through the matrix exponential map. More precisely, let 0 0 L def = 0 c c c = c 0 ; c ; c and {c, Rc} is a Jordan frame 0 c I 4cc +c Now define the set l l def = S s = s s R n, 0 s 0 and l π = {S s l s < π}. Then we have Proposition. The exponential map exp maps l to L and l π to L\{Q flip }; the latter is one-toone. 8

9 Proof: For any S s l, Ss = 0 s s 0, 0 0 ss and Hence for s 0, 3 exps s = I + S s s [ S k+ s = s s k S s, S k+ s = s s k S s. i= i+ s i i! ] + Ss s [ i=0 i s i+ i+! ] = I + cos s S s s + sin s s S s. When s = 0, then S = 0 and exp0 = I. Now let c R n+ such that {c, Rc} is a Jordan frame; thus in particular, c = /. Assume c < /. Then there is a unique 0 < α < π such that cos α = c and sin α = c. Since c 0, let s = αc, then expss = Q c. On the other hand, S s l, S s 0, let 4 c = ; cos s ; sin s s Then exps s = Q c L. For c = / Q c = I and for c = / note that as s π, the matrix Q c tends to Q flip. Corollary. For any Q in L, QQ = Q. Proof: It is well-known that the set of orthogonal transformations that fix the coordinate x 0 and act on x of vectors x Q are part of the automorphism group of Q, see [4]. And L is a subset of the orthogonal transformations that fix e. Instead of the exponential map we may use the Cayley transform to map skew symmetric matrices to orthogonal matrices: cays def = I + I S S For the special case of S s we write cays for cays s. Proposition.3 Under the Cayley transform l is mapped onto L. Furthermore, this map is oneto-one. Proof: When s <, all eigenvalues of S s are smaller than two in modulus. Therefore, the power series expansion of the cays is defined and can be expanded by the following power series: 5 + I + S si S s = I + k= s k S k s = I s S s + 4+ s S s. Since the right hand side of 5 is well defined even for s, we use the right hand side of 5 as the definition of Cayley transformation for any S l It is not hard to see that given S l, the Cayley transformation of S is Q c where 6 c = ; 4 s 4+ s ; 4s 4+ s Next let Q c L, where {c, Rc} is a Jordan frame, c = / and c. Set s = c c. Then 0+ cays = Q c. Note that as s and s s, cays converges to Q d where d = ; ; 0. Propositions. and.3 show that the tangent space to L at the identity I is l.. 9

10 3 Description of the Q Method As stated earlier, the traditional interior point methods for SOCP are based on applying Newton s method to the system of equations 5. Assume that x; y; z is our current estimate of the optimal solution. In general there is no reason for x and z to operator commute, even though the optimal x and z do. In many traditional interior point algorithms a transformation is applied to x and z so that in the transformed space they operator commute. The price paid for this operation is that the algorithm deviates from Newton s method. In particular the Schur complement A Arwz ArwxA may be undefined or singular at the optimal point. 3. Change of variables by spectral decomposition The Q method imposes the requirement that x and z should operator commute all the time. To achieve this, we make a change of variables by replacing blocks x i and z i with their spectral decomposition. Furthermore, we require that in their spectral decomposition x i and z i share a common Jordan frame {c i, c i }. Thus, if x = x,... ; x r and z = z ;... ; z r where each x i and z i are in the simple second order cone Q ni+, then λ i c i + λ i c i x i ω i c i + ω i c i z i With this change of variables the system 5 becomes 7 r A i λ i c i + λ i c i = b i= A i y + ω i c i + ω i c i = c i for i =,..., r λ i ω i = λ i ω i = µ for i =,..., r The difficulty with this change of variables is that we have to somehow ensure that {c i, c i } is a Jordan frame. To this end we write c i = Q ci d i, c i = Q ci d i, where {d i, d i } is the standard frame, and Q cij L. Next we express each Q ci by either exps si or cays si where S si l. With these modifications the system 7 can be written as A i exps si λ i d i + λ i d i = b for i =,..., r 8 3. The Newton direction i A i y + exps si ω i d i + ω i d i = ci for i =,..., r λ i ω i = λ i ω i = µ for i =,..., r The next step is to apply Newton s method to 8. First, for multiple block system, as usual, we set λ = λ ; λ ;... ; λ r ; λ r, ω = ω ; ω ;... ; ω r ; ω r, s = s ;... ; s r, S s = DiagS s,..., S sr, and Q = Q,..., Q r. The same applies to λ, ω, etc. To derive the the next approximation to the optimal solution using Newton s method we make the following substitution in the primal and dual feasibility and complementarity relations 8: λ λ + λ y y + y ω ω + ω Q Q exps s or Q Q cays s 0

11 To update Q i we may use cays i instead of the exponential map. Once we make this substitution in 8 we drop all nonlinear terms involving s i, λ, y, and ω and solve the resulting linear system in these variables. For exps s we simply write I + S s and drop higher order terms of its power series expansion. Since cay and exp have the same first two terms in their power series expansion we make the same replacement if we are using the Cayley transform. After simplification the primal feasibility, dual feasibility and complementarity conditions result in the following linear Newton system: λ i + λ i i B i λ i λ i λi λ = r p i s 9 ω i + ω i B i y + ω i ω i ωi ω = Q i r d i for i =,..., r i s λ ij ω ij + ω ij λ ij = µ λ ij ω ij for i =,..., r, j =,, where B i = A i Q i r p = b Ax r d = c A y z r c = µ λ ω Here λ ω is a vector which is the component-wise product of entries of λ and ω. Also r d = rd ;... ; r d r. Note that when x is primal-feasible then rp = 0. Similarly when y; z is dualfeasible then r d = 0. In general even if x; y; z is feasible the solution to the Newton system 9 is not necessarily primal or dual feasible direction. This is due to the fact that the change of variables has turned both primal and dual feasibility equations nonlinear. In contrast, in traditional interior point methods, if x; y; z is primal respectively dual feasible, then the Newton direction is primal respectively dual feasible. Let P = DiagP,..., P r where P i =, ;,. Note that P = P. Finally let Λ = Diagλ and Ω = Diagω. After collecting the first two columns of each B i into B, the remaining columns of each B i into ˆB, and splitting Q T r d accordingly to r d and ˆr d, we can rewrite the Newton system as P ω + B T y = r d, 30 ω i ω i s i + ˆB i T y = ˆr d i i =,..., n, n λi λ i BP λ + ˆB i s i = r p, i= Λ ω + Ω λ = r c. def Let E i = ωi ωi def I, E = DiagE,..., E r, D i = λi λi I, and D = DiagD,..., D r. Then the solution to the Newton linear system is given by y = BP Ω ΛP T BT ˆBDE ˆBT r p BP Ω r c ˆBDE ˆr d + BP Ω ΛP T r d ω = P r d 3 B T y λ = Ω ˆr r c Λ ω s = E d ˆB T y.

12 Comparison of Newton systems in the Q and traditional methods. In traditional methods our variables are x, y, z. When we have r blocks, i n i = n and A has m rows then we will have a total n + m variables. In the Q method for each block of size n i we have λ i, λ i, ω i, ω i and s i, for a total of n + r + m variables.. The Schur complement BP Ω ΛP T BT ˆBDE ˆBT is a symmetric positive definite matrix. This can be seen by writing the Schur complement as AQDiag P Ω i Λ i P T, D i E i Q T A T, The traditional interior point methods require a scaling in order to get a symmetric Schur complement. 3.3 Centrality requirement To ensure that the sequence of iterates in the Q method converges to an optimal solution when one exists, we define a region that includes the optimal solution and require that each new approximation remain in this region. The region or neighborhood N γ c, γ p, γ d depends on several parameters. Let ɛ p, ɛ d, and ɛ c be our tolerance for, respectively, primal infeasibility, dual infeasibility, and duality gap. In Appendix A we will explain the details of dependence of γ p, γ d, γ c on ɛ p, ɛ d, ɛ c and other parameters derived from the data. For now let us simply assume that the γ s are given. Then the neighborhood N is defined as { N γ c, γ p, γ d def = λ, ω, y, Q λ R r, ω R r, y R m, Q L, λ > 0, ω > 0, λ λ ij ω ij γ ω c r i =,..., r, j =, and λ AQ AQ ω γ p P λ b or P λ b ɛp and λ A ω γ d y + Q P ω c or A y + Q P } ω c ɛ d The first inequality is the centrality condition; it ensures that none of the products λ ij ω ij approach zero much faster than others. The second and third inequalities guarantee that the complementarity is not attained before primal or the dual feasibility. Obviously, when γ c, γ p, γ d γ c, γ p, γ d, N γ c, γ p, γ d N γ c, γ p, γ d, and γ c,γ p,γ d >0 N γ c, γ p, γ d = {λ, ω, y, Q: λ > 0, ω > 0}. Therefore, if λ k, ω k, y k, Q k N γ c, γ p, γ d is a sequence of points where λ k ω k approaches zero, then λ k, ω k, y k, Q k necessarily approaches a point in the optimal solution set of. The Q method is summarized in Figure 3.3. Here we have used the Cayley transformation, however, updating the Q i through the exponential map can be achieved in a similar way; and the analysis presented in Appendix A carries over with slight modifications of constants. One issue that we need to address here is that if λ k i = λk i or ωk i = ωk i then xk i is a multiple of the identity vector. In this case the Jordan frame is not unique. Indeed, as was stated earlier, every Jordan frame can be taken as the associated frame of multiples of the identity vector. The result is that the vector s may not be uniquely determined, and in fact the Newton system may be singular. To avoid this situation, we take care to ensure that eigenvalues of primal and dual blocks are distinct

13 Algorithm Choose 0 < σ < σ < σ 3 < and Υ > 0. To start from an arbitrary point λ 0, ω 0, y 0, Q 0, one may select 0 < γ c <, γ p > 0, γ d > 0, so that λ 0, ω 0, y 0, Q 0 N γ c, γ p, γ d. Do. Set µ = σ λ k ω k r.. Compute the search direction λ, ω, y, s from Choose step sizes α, β, γ and set λ k+ = λ k + α λ, y k+ = y k + β y, ω k+ = ω k + β ω, Q k+ = Q c k+ where c k+ is computed from 4 for expγ c S sk and 6 for cayγ c s k 4. k k +. Until either: r k p < ɛ p, r k d < ɛ d, and λ k ω k < ɛ c ; or: λ k, ω k > Υ. End eigenvalues of different blocks may be equal and this will not cause any difficulty. By careful choice of step sizes we can always ensure that λ k+ i > λ k+ i > 0 and ω k+ i > ω k+ i > 0 for each block i. First, if ω k+ i > ω k+ i, or λ k i > λk i, then we simply swap the the corresponding Jordan frame. This means that c i, c i is replaced by c i, c i. This transformation geometrically is equivalent to a rotation of π and algebraically equivalent to multiplication by Q flip. Now let ωi k > ωk i. Only when ω i ω i and β = ωk i ωk i ω i ω i, is it possible that ωi k + β ω i = ωi k + β ω i. In this case, instead of taking a uniform step size of length β we simply take a smaller step size β i for each block i Thus β i are not necessarily equal. For instance β i can be set to at least β. In this way we guarantee that in each block eigenvalues are distinct for both the x i and z i. Note that Algorithm does not require a large number of numerical operations to calculate updates of Q i, since we will use 5 for the Cayley transform and 3 for the exponential map. Let ˆα k be the maximum of α [0, ], so that for any α [0, α], and λ k + α λ, ω k + α ω, y k + α y, Q k I + α SI α S N, λ k + α λ ω k + α ω [ α σ ] λ k ω k. The step sizes α 0, ], β 0, ], γ 0, ] are chosen so that λ k+, ω k+, y k+, Q k+ N γ c, γ p, γ d, λ k+ ω k+ [ ˆα k σ 3 ] λ k ω k. Since σ < σ < σ 3, the primal and dual step sizes are not necessarily equal. One important property of the Q method is that in the absence of primal and dual nondegeneracy and presence of strict complementarity see [] for definitions the Newton system 9 is nonsingular 3

14 at the optimum. A similar result holds for the so-called AHO method for SOCP see [] but not for many other traditional methods. We will not prove nonsingularity of the Newton system here since it is proved in a more general setting in [8]. In Appendix A we prove that with appropriate choice of step lengths the algorithm converges to an ɛ p, ɛ d, ɛ c -optimal solution in finite number of iterations. 4 Numerical Results To test the Q method, we have implemented a version of Algorithm in MATLAB. Below we present the results of our test on,000 randomly generated problems with known solutions. We test the algorithm using the Cayley transform for the most part. For the step sizes we set α = min, τα, β = min, τβ, γ = αβ, where α is the maximum possible step size such that x + α x Q. Similarly β is the maximum step size such that z + β z Q. These choices are in line with implementation of traditional methods. 4. Test results on randomly generated problems We use x 0 i = ; ; 0, s0 i = ; ; 0, y0 = 0 as starting point. We pick σ = 0.5, τ = Our code reduced the l norm of primal infeasibility, l norm of dual infeasibility, and the duality gap to less than 5.0e for all problems. The data for A, b and c were randomly generated between -0.5 and 0.5. Thus, in most cases the magnitude of the solutions were comparable and there is no need to use relative accuracy. bk dimension of each block type of each block m r p0 r d 0 it 0 [,,,,,,,,,] [b,i,o,b,i,b,o,i,i,b] [0,0,0,0,0,0,0,0, [b,o,i,b,b,i,o,b,b, ,0] o] 0 [3,0,8,9,,4,6,3,4,8] [b,i,o,b,i,o,i,i,b,o] [0,0,8,9,,5,6,3,4,8] [b,i,b,i,i,o,b,i,b,o] [0,5,5,5,5,5,5,5, [b,i,b,i,i,o,b,i,b,o] ,5] [0,0,0,0,0,0,0,0, [b,o,i,b,b,i,o,b,b,o, ,0,0,0] b,i] 5 [0,0,0,0,0,0,0,0 [b,o,i,b,b,i,o,b,b,o, , 0,0,0,0,0,0] b,o,i,i,o] 5 [5,5,5,5,5,5,5,5, [i,o,b,i,i,b,o,i,b,b, ,5,5,5,5,5,5,] i,o,b,b,o] 0 [0,0,3,0,4,0,3,8,6, [b,o,i,b,b,i,o,b,b,o, ,9,,,3,,3,5,,0, b,b,i,o,i,b,b,b,i,b] 8] 0 [0,0,0,0,0,0,0,0 [b,o,i,b,b,i,o,b,b,o, ,0,0,0,0,0,0,0, b,b,i,o,i,b,b,b] 0,0,0,0] Figure 3: Computational test results of performance of the Q method for 000 randomly generated SOCP s. For each of 0 different classes the size and type of each block along with the average number of iterations for the 00 generated instances of that type is given. For our tests, we generated 00 random instances of similar classes of SOCP problems. In each class, the number of blocks r, dimensions n i for i =,..., r, number of rows of matrix A, and the type of each block at the optimum is fixed. Each block can be one of three types: Type i: x i Int Q ni+, z i = 0, Type o: x i = 0, z i Int Q ni+, 4

15 Type b: x i, z i bd Q ni+. In the table in Figure 4. bk stands for the number of blocks, type of each block is a string made of i, o and b, and m is the number of constraints. r p 0 and r d 0 give the average l norm of initial primal and dual infeasibility for the 00 instances in each row, and it is the average number of iterations for the 00 instances. Every instance was solved to optimality in fewer than 50 iterations, indicating the robustness of the algorithm. Notice that the problem type and size have little effect on the average number of iterations, which is typical of most interior point methods for SOCP. In Figure 4. a plot indicating progress of one run of the algorithm is presented. The logarithm of primal and dual feasibility as represented by euclidean norm of r p and r d and duality gap are plotted against iteration count. The plot shows that, at least in this case, the order of convergence Convergence in a typical run of Algorithm 4 primal infeasibility dual infeasibility duality gap log 0 gap iteration is well above. We have observed that this is a typical behavior for almost all runs. It is possible to adapt Mehrotra s predictor-corrector method developed for linear programming in [7] to the Q method for SOCP. However, once this predictor-corrector method was added to our code, we observed an inconsistent behavior. In most cases the number of iterations was reduced by as much as one third. In other cases though the number of iterations was increased significantly and some times, no convergence was attained at all. We believe that adaption of the predictor-corrector method for the Q method requires further research and experimentation. Finally we have tested the algorithm using the exponential map instead of the Cayley transform. The results were entirely compatible with Cayley transform version of the algorithm. 4. Testing the Q method on the Steiner minimal tree problem We also tested the Q method on problems arising from euclidean minimal Steiner tree problems. Details can be found in the paper of Xue and Ye, [9]. Here we briefly mention that in the Steiner tree problem we are searching for a tree consisted of a set of N given points called regular points and possibly additional vertices called Steiner points such that the sum of euclidean length of edges in the tree is minimum. The problem may be decomposed into two subproblems. In the first subproblem we need to find a particular topology, that is a set of Steiner points. In the second subproblem, given the topology we need to find the optimal position of the Steiner points. The first 5

16 subproblem finding a particular topology is combinatorial in nature, the second can be formulated as an SOCP problem. It can be shown that in the optimal tree the number of Steiner points is at most N and the degree of each steiner point is three. We tested our code on Example in [9]. We use the same starting points as we did in the randomly generated SOCP s of the previous section. The progress of the Q method in each iteration is summarized in Figure 4. it network-cost r p r d gap e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e-0 Figure 4: Progress of the Q method on one instance of a Steiner tree problem with a given topology. In contrast to the algorithm used in Xue and Ye [9] we start ours from an infeasible point. We achieve the same accuracy as Xue and Ye did at iteration number Taking advantage of sparsity Our implementation of Algorithm in MATLAB does not take advantage of sparsity in the input matrix A. Here we briefly state how sparse input matrices A can be exploited and compare the situation with traditional interior point methods. Let both A and AA be sparse matrices. It is well-known that in interior point methods for linear programming the Schur complement in all versions are of the form AF A for some diagonal matrix F. Also the sparsity pattern of AA and AF A are identical. The most difficult part for a sparse Cholesky factorization is to find a good order for pivoting. Once such an order is found for AA, the same ordering may be used for all subsequent iterations for FACTORING AF A. 6

17 The situation for SOCP is more involved, see [] for a detailed analysis for the case of traditional interior point methods. However, in all methods we have the fact that the Schur complement at each iteration is of the form AF A + L where L is of rank r, r or 3r, and is symmetric for most methods except for the so-called AHO method. We now show that the Schur complement for the Q method is also of the same format. For convenience let us temporarily assume that r =. In this case, as stated in 3, the Schur complement is given by AQDiag P Ω i Λ i P T, D i E i Q T A T, The Cholesky factorization of this matrix is the main computational effort at each iteration. However it is easily verified that the matrices Q c L are all of the form of I + rank- matrix. Furthermore, with some algebraic manipulation it can be verified that the QDiag P Ω i Λ i P T, D i E i is also sum of a diagonal matrix with positive entries and a positive semidefinite rank- matrix. Thus, the Schur complement is of the form AF A +a rank- matrix. With r blocks, the Schur complement is AF A + a rank-r matrix. In both cases F is a diagonal. As in linear programming, sparse factorization of AA will yield sparse factorization of ADA. Sophisticated numerical techniques are available to extend this sparse factorization to ones with low rank updates, see [] for details and references. 5 A Modified Q Method In this section, we give a variant of the Q method for SOCP, which has similar properties and convergence results as Algorithm. 5. The Newton system We can write dual and primal feasibility and complementarity relations alternatively in the following form Q T 3 Q P ω + A y = c, AQ P λ = b, λ ω = µ where P = Diag P,..., P r where each P i = d i, d i, made up of the standard frame in R ni. Thus each block P i is an n i matrix with only the top part nonzero. Now in 3 The matrix Q appears only as Q P. Thus for each block i only columns indexed zero and of each Q i is involved. Since each Q i is of the form Q c L, the 0 th column is all zeros except the first entry. The column indexed is of the form 0; c where {c, Rc} is a Jordan frame. In particular, defining q = c it follows that the only relevant part of Q i in 3 is the vector q i, of the column indexed. Furthermore q i q i =. Partitioning A i into A i = A i 0, Āi we write Ā def = Ā,, Ān. Observe that xi and z i may be written as 33 x i = λ i+λ i ; λi λi q i, zi = ω i+ω i ; ωi ωi q i. We set q i = ; 0 when x i = 0. Then the decomposition 33 is unique when λ i λ i and ω i ω i. We now substitute 33 into 3, and add constraints q i q i = for i =,..., r. The 7

18 residuals r p, r d and r c are defined as before. Let r d k i be the first element of r d k i, and r d k i the remaining subvector. Then, similar to Algorithm we make the substitution λ λ + λ, ω ω + ω, and y y + y. But now, instead of exponential map or the Cayley transform we simply substitute q q + q. After making these substitutions in 3 and dropping nonlinear terms in s we get the linear Newton system ω i+ ω i + A i 0 y = r d k i 34 ω i ω i q k i + ωk i ωk i q i + Āi y = r d k i n λi+ λ i λi λi A i 0 + Āi q k i + Āi λk i λk i q k i = r k p i= q k i qi = 0 i =,..., n Λ k ω + Ω k λ = r k c. where, as before Λ = Diagλ and Ω = Diagω. The modified algorithm is similar to Algorithm with the exception that the new q i should be normalized: q k+ i = qk i +γ qi q k+γ qi. i 5. The solution to modified Q method Let def u i = λi ω i + λi ω i, def v i = λi ω i λi ω i ; E i and D i are defined as that in section 3, but with dimensions adjusted properly. Omitting the superscript k, the Schur complement M can be written as M = n i= Ai v i 0 q i Ā i + Āiq i A i 0 + ĀDiag u i, D i E i + ui I q i q i D i E i Ā. The solution to 34 is y = M r p + n i= A i 0 r c i ω i Āiq i r c i + u i A i 0 r d i + v i Ā i q i r d i + v i A i 0 q i r d i + u i + D i E i Āiq i q i r d i D i E i Ā i r d i ω i = q i r d i q i r d i + q i Ā i A i 0 y q i = E i rd i q i r d i + q i ω i Ā i q i A i 0 y ω i = r d i A i 0 y ω i λ = Ω r c Λ ω. Properties of the solution ω i. The number of unknowns in the modified Q method is the same as the Q method: Instead of the vector s i we now have q i in each block i.. a The dimension of each block M i of M is one smaller than the Schur complement of the traditional methods. 8

19 b To keep each iterate in Q, we only need to compute α max{ λ i /λ i : λ i < 0}. In comparison, in traditional methods we need to solve a quadratic equation for each block. c When λ i > λ i > 0 and ω i > ω i > 0, we have u i > v i > 0, and D i E i is a positive scalar matrix. Therefore it is easily verified that each M i is symmetric positive definite and we may use Cholesky factorization to solve systems of equations coefficient matrix M. 3. Accuracy and stability. It can be shown that at the optimum the Newton system 34 is nonsingular if primal and dual nondegeneracy and strict complementarity holds. This in turn implies that the algorithm converges quadratically near the optimal solution. 4. We can show that with careful choice of parameters the modified Q method converges to the optimal solution in finite number of iterations. This is argued in Appendix A. 5.3 Numerical examples We have implemented the basic algorithm for the modified Q method in MATLAB and have tested on,000 randomly generated problems. Step sizes α, β, γ are chosen similar to our implementation of Algorithm. We applied our modified Q code to the same set 0 classes of SOCP problems each with 00 instances. The results are summarized in Figure 5.3. Although the algorithm finds an problem r p0 r d 0 it ɛ p, ɛ d, ɛ c -optimal solution for all of the, 000 problems, a small portion of them need more than 00 iterations to reach the required accuracy. Finally we applied our implementation of the modified Q method to the same instance of the Steiner tree problem we used for our implementation of Algorithm. The progress of the algorithm is reported in Figure 5.3 The performance is similar to Algorithm. 6 Conclusions We have developed and analyzed the Q method and a modified variant for SOCP. Preliminary numerical results show that the algorithm is robust. We conclude by mentioning some future research directions. First, it remains open whether the Q method has polynomial time iteration complexity. The finite convergence result presented in this paper depends on the condition number of the input matrix A. For practical and stable implementation of the method, Mehotra s predictor-corrector method should be adjusted to perform more robustly. Finally, a serious implementation should take advantage of sparse factorization of the Schur complement. 9

20 it network-cost r p r d gap e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e e-06.49e e e e e e e e e e e-0 Figure 5: Progress of the modified Q method for one instance of a Steiner tree problem with given topology. A Appendix: Convergence Analysis of the Q and Modified Q Methods A. Convergence analysis Global convergence of Algorithm can be established by adopting ideas of Kojima, Megiddo and Mizuno, [6] for infeasible interior point linear programming. Theorem A. If Algorithm does not stop after a finite number of steps, the smallest singular value of the Jacobian of 30 must converge to zero. Proof: Let J be the Jacobian of 30. The key to the proof is to show that the step sizes are bounded below. Assume the algorithm doesn t stop after a finite number of steps. Define ɛ def = minɛ c, γ p ɛ p, γ d ɛ d. Then for each iteration k, λ k ω k ɛ, and λ k ; ω k Υ, otherwise, the termination criterion will force the algorithm to stop. Boundedness of y k is explained by the dual feasibility constraint. Also Q k is orthogonal and the set of orthogonal matrices is compact. Assume the smallest singular value of J does not converge to zero. Then there must exist a positive scalar d, and a subsequence {λ mi, ω mi, y mi, Q mi } i= such that for all m i, the largest singular value of J is at most d. Both the right hand side and the left hand side of 30 0

21 depend continuously on λ, ω, y, Q. Since the set λ, ω, y, Q derived from the Newton system forms a compact set the Newton direction in 30 is a continuous function of λ, ω, y, Q. Therefore, solution of 30 is uniformly bounded for the subsequence {m i }. Hence, there exists a positive constant η, such that the search direction computed by 30 satisfies λ ij ω ij γ c n λ ω η, λ ω η, λ η, ω η, s i η, for i =,..., n; j =,. Note that S i = s i for i =,..., n, S = max i s i. For k {m i } i=, adopting the notation used in [6], we define f ij α def = [ λ k ij + α λ ij ] [ ω k ij + α ω ij ] γ c n λk + α λ ω k + α ω, g p α def = λ k + α λ ω k + α ω AQ γ k p I + α SI α S P λ k + α λ b, g d α def = λ k + α λ ω k + α ω A γ d y k + y + Q k I + α SI α S P ω k + α ω c, hα def = [ α σ ] λ k ω k λ k + α λ ω k + α ω. Therefore, ˆα k is the largest α satisfying the following inequalities: f ij α 0 i =,..., n; j =,, g p α 0 or AQ k P λ k b ɛ p, g d α 0 or A y k + Q k P ω k c ɛ d, hα 0. Next, we give a lower bound for the set {ˆα k }. Recall that each block of the Cayley transformation is given by 35 I + αs i α3 s i 4 + α s i S i + α 4 + α s i S i. The inequalities for f ij and h can be obtained by similar arguments as those in [6]: ɛ f ij α σ n γ cα ηα, hα σ σ ɛ α ηα. Next, we estimate g p α and g d α. Note that the first column of Si is zero; and the only nonzero entry of its second column is s i s i. Let Q k be the matrix consisting of only the nd column of each block of Q k, λ be the vector made of all of the larger of the two eigenvalues of x i and λ be the vector made up of all the smaller of the two eigenvalues of x i, for i =,..., r. When Recall that S, the spectral norm of S, is the given by the modulus of the largest-modulus eigenvalue of S.

22 λ k ω k AQ k γ p P λ k b, 36 g p α αλ k ω k + ασ λ k ω k + α λ ω γ p α AQ k P λ k b [ AQ γ p α k S P λ + max i s i AQ k 4 λ k λ k + α λ α λ ] + α max i s i AQ k S 4 P λ k + α λ ασ ɛ α η γ p α A η 3/ + 4 ηυ + η + 4 η3/ Υ + 4 η5/. The first inequality follows from the Newton linear system for the search directions 30, the fact that λ k ω γp AQ k P λ b, and the Cayley transform 35. The second inequality follows from the bound on the variables and search directions, the fact that α, and If AQ k P λ k b ɛ p, then P λ = λ λ. 37 AQ k I + α SI α S P λ k + α λ b α AQ k P λ k b [ AQ + α k S P λ + max i s i AQ k 4 λ k λ k + α λ α λ ] + α max i s i AQ k S 4 P λ k + α λ k αɛ p + α A η 3/ + 4 ηυ + η + 4 η3/ Υ + 4 η5/. Therefore when ɛ p α, A η 3/ + 4 ηυ + η + 4 η3/ Υ + 4 η5/ we must have AQ k+ P λ k+ b ɛ p. Next, we will consider the dual constraints. When λ k ω k A γ d y k + Q k P ω k c, 38 g d α αλ k ω k + ασ λ k ω k + α λ ω [ Q γ d α A y k + Q k P ω k c γ d α k S P ω ] + Q k S P ω k + α ω + α max i s i Q k S 4 4 P ω k + α ω ασ ɛ α η γ d α η 3/ + 4 ηυ + η + 4 η3/ Υ + 4 η5/.

The Q Method for Symmetric Cone Programmin

The Q Method for Symmetric Cone Programmin The Q Method for Symmetric Cone Programming The Q Method for Symmetric Cone Programmin Farid Alizadeh and Yu Xia alizadeh@rutcor.rutgers.edu, xiay@optlab.mcma Large Scale Nonlinear and Semidefinite Progra

More information

McMaster University. Advanced Optimization Laboratory. Authors: Farid Alizadeh and Yu Xia. AdvOl-Report No. 2004/15

McMaster University. Advanced Optimization Laboratory. Authors: Farid Alizadeh and Yu Xia. AdvOl-Report No. 2004/15 McMaster University Advanced Optimization Laboratory Title: The Q Method for Second-order Cone Programming Authors: Farid Alizadeh and Yu Xia AdvOl-Report No 004/5 October 004, Hamilton, Ontario, Canada

More information

Lecture: Algorithms for LP, SOCP and SDP

Lecture: Algorithms for LP, SOCP and SDP 1/53 Lecture: Algorithms for LP, SOCP and SDP Zaiwen Wen Beijing International Center For Mathematical Research Peking University http://bicmr.pku.edu.cn/~wenzw/bigdata2018.html wenzw@pku.edu.cn Acknowledgement:

More information

Lecture 1. 1 Conic programming. MA 796S: Convex Optimization and Interior Point Methods October 8, Consider the conic program. min.

Lecture 1. 1 Conic programming. MA 796S: Convex Optimization and Interior Point Methods October 8, Consider the conic program. min. MA 796S: Convex Optimization and Interior Point Methods October 8, 2007 Lecture 1 Lecturer: Kartik Sivaramakrishnan Scribe: Kartik Sivaramakrishnan 1 Conic programming Consider the conic program min s.t.

More information

The Q Method for Symmetric Cone Programming

The Q Method for Symmetric Cone Programming The Q Method for Symmetric Cone Programming Farid Alizadeh Yu Xia October 5, 010 Communicated by F. Potra Abstract The Q method of semidefinite programming, developed by Alizadeh, Haeberly and Overton,

More information

Lecture: Introduction to LP, SDP and SOCP

Lecture: Introduction to LP, SDP and SOCP Lecture: Introduction to LP, SDP and SOCP Zaiwen Wen Beijing International Center For Mathematical Research Peking University http://bicmr.pku.edu.cn/~wenzw/bigdata2015.html wenzw@pku.edu.cn Acknowledgement:

More information

Interior Point Methods. We ll discuss linear programming first, followed by three nonlinear problems. Algorithms for Linear Programming Problems

Interior Point Methods. We ll discuss linear programming first, followed by three nonlinear problems. Algorithms for Linear Programming Problems AMSC 607 / CMSC 764 Advanced Numerical Optimization Fall 2008 UNIT 3: Constrained Optimization PART 4: Introduction to Interior Point Methods Dianne P. O Leary c 2008 Interior Point Methods We ll discuss

More information

Semidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 4

Semidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 4 Semidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 4 Instructor: Farid Alizadeh Scribe: Haengju Lee 10/1/2001 1 Overview We examine the dual of the Fermat-Weber Problem. Next we will

More information

A PREDICTOR-CORRECTOR PATH-FOLLOWING ALGORITHM FOR SYMMETRIC OPTIMIZATION BASED ON DARVAY'S TECHNIQUE

A PREDICTOR-CORRECTOR PATH-FOLLOWING ALGORITHM FOR SYMMETRIC OPTIMIZATION BASED ON DARVAY'S TECHNIQUE Yugoslav Journal of Operations Research 24 (2014) Number 1, 35-51 DOI: 10.2298/YJOR120904016K A PREDICTOR-CORRECTOR PATH-FOLLOWING ALGORITHM FOR SYMMETRIC OPTIMIZATION BASED ON DARVAY'S TECHNIQUE BEHROUZ

More information

Linear Programming: Simplex

Linear Programming: Simplex Linear Programming: Simplex Stephen J. Wright 1 2 Computer Sciences Department, University of Wisconsin-Madison. IMA, August 2016 Stephen Wright (UW-Madison) Linear Programming: Simplex IMA, August 2016

More information

Semidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 5

Semidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 5 Semidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 5 Instructor: Farid Alizadeh Scribe: Anton Riabov 10/08/2001 1 Overview We continue studying the maximum eigenvalue SDP, and generalize

More information

Chapter 3 Transformations

Chapter 3 Transformations Chapter 3 Transformations An Introduction to Optimization Spring, 2014 Wei-Ta Chu 1 Linear Transformations A function is called a linear transformation if 1. for every and 2. for every If we fix the bases

More information

A class of Smoothing Method for Linear Second-Order Cone Programming

A class of Smoothing Method for Linear Second-Order Cone Programming Columbia International Publishing Journal of Advanced Computing (13) 1: 9-4 doi:1776/jac1313 Research Article A class of Smoothing Method for Linear Second-Order Cone Programming Zhuqing Gui *, Zhibin

More information

We describe the generalization of Hazan s algorithm for symmetric programming

We describe the generalization of Hazan s algorithm for symmetric programming ON HAZAN S ALGORITHM FOR SYMMETRIC PROGRAMMING PROBLEMS L. FAYBUSOVICH Abstract. problems We describe the generalization of Hazan s algorithm for symmetric programming Key words. Symmetric programming,

More information

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.

More information

Primal-Dual Interior-Point Methods for Linear Programming based on Newton s Method

Primal-Dual Interior-Point Methods for Linear Programming based on Newton s Method Primal-Dual Interior-Point Methods for Linear Programming based on Newton s Method Robert M. Freund March, 2004 2004 Massachusetts Institute of Technology. The Problem The logarithmic barrier approach

More information

A Generalized Homogeneous and Self-Dual Algorithm. for Linear Programming. February 1994 (revised December 1994)

A Generalized Homogeneous and Self-Dual Algorithm. for Linear Programming. February 1994 (revised December 1994) A Generalized Homogeneous and Self-Dual Algorithm for Linear Programming Xiaojie Xu Yinyu Ye y February 994 (revised December 994) Abstract: A generalized homogeneous and self-dual (HSD) infeasible-interior-point

More information

c 2005 Society for Industrial and Applied Mathematics

c 2005 Society for Industrial and Applied Mathematics SIAM J. OPTIM. Vol. 15, No. 4, pp. 1147 1154 c 2005 Society for Industrial and Applied Mathematics A NOTE ON THE LOCAL CONVERGENCE OF A PREDICTOR-CORRECTOR INTERIOR-POINT ALGORITHM FOR THE SEMIDEFINITE

More information

Lecture 5. Theorems of Alternatives and Self-Dual Embedding

Lecture 5. Theorems of Alternatives and Self-Dual Embedding IE 8534 1 Lecture 5. Theorems of Alternatives and Self-Dual Embedding IE 8534 2 A system of linear equations may not have a solution. It is well known that either Ax = c has a solution, or A T y = 0, c

More information

Largest dual ellipsoids inscribed in dual cones

Largest dual ellipsoids inscribed in dual cones Largest dual ellipsoids inscribed in dual cones M. J. Todd June 23, 2005 Abstract Suppose x and s lie in the interiors of a cone K and its dual K respectively. We seek dual ellipsoidal norms such that

More information

Optimality Conditions for Constrained Optimization

Optimality Conditions for Constrained Optimization 72 CHAPTER 7 Optimality Conditions for Constrained Optimization 1. First Order Conditions In this section we consider first order optimality conditions for the constrained problem P : minimize f 0 (x)

More information

12. Interior-point methods

12. Interior-point methods 12. Interior-point methods Convex Optimization Boyd & Vandenberghe inequality constrained minimization logarithmic barrier function and central path barrier method feasibility and phase I methods complexity

More information

NORMS ON SPACE OF MATRICES

NORMS ON SPACE OF MATRICES NORMS ON SPACE OF MATRICES. Operator Norms on Space of linear maps Let A be an n n real matrix and x 0 be a vector in R n. We would like to use the Picard iteration method to solve for the following system

More information

A Review of Linear Programming

A Review of Linear Programming A Review of Linear Programming Instructor: Farid Alizadeh IEOR 4600y Spring 2001 February 14, 2001 1 Overview In this note we review the basic properties of linear programming including the primal simplex

More information

Linear Programming Redux

Linear Programming Redux Linear Programming Redux Jim Bremer May 12, 2008 The purpose of these notes is to review the basics of linear programming and the simplex method in a clear, concise, and comprehensive way. The book contains

More information

UNDERGROUND LECTURE NOTES 1: Optimality Conditions for Constrained Optimization Problems

UNDERGROUND LECTURE NOTES 1: Optimality Conditions for Constrained Optimization Problems UNDERGROUND LECTURE NOTES 1: Optimality Conditions for Constrained Optimization Problems Robert M. Freund February 2016 c 2016 Massachusetts Institute of Technology. All rights reserved. 1 1 Introduction

More information

Optimality, Duality, Complementarity for Constrained Optimization

Optimality, Duality, Complementarity for Constrained Optimization Optimality, Duality, Complementarity for Constrained Optimization Stephen Wright University of Wisconsin-Madison May 2014 Wright (UW-Madison) Optimality, Duality, Complementarity May 2014 1 / 41 Linear

More information

Chapter 1. Matrix Algebra

Chapter 1. Matrix Algebra ST4233, Linear Models, Semester 1 2008-2009 Chapter 1. Matrix Algebra 1 Matrix and vector notation Definition 1.1 A matrix is a rectangular or square array of numbers of variables. We use uppercase boldface

More information

More First-Order Optimization Algorithms

More First-Order Optimization Algorithms More First-Order Optimization Algorithms Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A. http://www.stanford.edu/ yyye Chapters 3, 8, 3 The SDM

More information

Optimization Theory. A Concise Introduction. Jiongmin Yong

Optimization Theory. A Concise Introduction. Jiongmin Yong October 11, 017 16:5 ws-book9x6 Book Title Optimization Theory 017-08-Lecture Notes page 1 1 Optimization Theory A Concise Introduction Jiongmin Yong Optimization Theory 017-08-Lecture Notes page Optimization

More information

Uniform Boundedness of a Preconditioned Normal Matrix Used in Interior Point Methods

Uniform Boundedness of a Preconditioned Normal Matrix Used in Interior Point Methods Uniform Boundedness of a Preconditioned Normal Matrix Used in Interior Point Methods Renato D. C. Monteiro Jerome W. O Neal Takashi Tsuchiya March 31, 2003 (Revised: December 3, 2003) Abstract Solving

More information

Lecture 15 Newton Method and Self-Concordance. October 23, 2008

Lecture 15 Newton Method and Self-Concordance. October 23, 2008 Newton Method and Self-Concordance October 23, 2008 Outline Lecture 15 Self-concordance Notion Self-concordant Functions Operations Preserving Self-concordance Properties of Self-concordant Functions Implications

More information

Research Note. A New Infeasible Interior-Point Algorithm with Full Nesterov-Todd Step for Semi-Definite Optimization

Research Note. A New Infeasible Interior-Point Algorithm with Full Nesterov-Todd Step for Semi-Definite Optimization Iranian Journal of Operations Research Vol. 4, No. 1, 2013, pp. 88-107 Research Note A New Infeasible Interior-Point Algorithm with Full Nesterov-Todd Step for Semi-Definite Optimization B. Kheirfam We

More information

A priori bounds on the condition numbers in interior-point methods

A priori bounds on the condition numbers in interior-point methods A priori bounds on the condition numbers in interior-point methods Florian Jarre, Mathematisches Institut, Heinrich-Heine Universität Düsseldorf, Germany. Abstract Interior-point methods are known to be

More information

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra. DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1

More information

I.3. LMI DUALITY. Didier HENRION EECI Graduate School on Control Supélec - Spring 2010

I.3. LMI DUALITY. Didier HENRION EECI Graduate School on Control Supélec - Spring 2010 I.3. LMI DUALITY Didier HENRION henrion@laas.fr EECI Graduate School on Control Supélec - Spring 2010 Primal and dual For primal problem p = inf x g 0 (x) s.t. g i (x) 0 define Lagrangian L(x, z) = g 0

More information

Relation of Pure Minimum Cost Flow Model to Linear Programming

Relation of Pure Minimum Cost Flow Model to Linear Programming Appendix A Page 1 Relation of Pure Minimum Cost Flow Model to Linear Programming The Network Model The network pure minimum cost flow model has m nodes. The external flows given by the vector b with m

More information

MATH 5720: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 2018

MATH 5720: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 2018 MATH 57: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 18 1 Global and Local Optima Let a function f : S R be defined on a set S R n Definition 1 (minimizers and maximizers) (i) x S

More information

Second-order cone programming

Second-order cone programming Outline Second-order cone programming, PhD Lehigh University Department of Industrial and Systems Engineering February 10, 2009 Outline 1 Basic properties Spectral decomposition The cone of squares The

More information

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination Math 0, Winter 07 Final Exam Review Chapter. Matrices and Gaussian Elimination { x + x =,. Different forms of a system of linear equations. Example: The x + 4x = 4. [ ] [ ] [ ] vector form (or the column

More information

Using Schur Complement Theorem to prove convexity of some SOC-functions

Using Schur Complement Theorem to prove convexity of some SOC-functions Journal of Nonlinear and Convex Analysis, vol. 13, no. 3, pp. 41-431, 01 Using Schur Complement Theorem to prove convexity of some SOC-functions Jein-Shan Chen 1 Department of Mathematics National Taiwan

More information

5 Handling Constraints

5 Handling Constraints 5 Handling Constraints Engineering design optimization problems are very rarely unconstrained. Moreover, the constraints that appear in these problems are typically nonlinear. This motivates our interest

More information

A Distributed Newton Method for Network Utility Maximization, II: Convergence

A Distributed Newton Method for Network Utility Maximization, II: Convergence A Distributed Newton Method for Network Utility Maximization, II: Convergence Ermin Wei, Asuman Ozdaglar, and Ali Jadbabaie October 31, 2012 Abstract The existing distributed algorithms for Network Utility

More information

10 Numerical methods for constrained problems

10 Numerical methods for constrained problems 10 Numerical methods for constrained problems min s.t. f(x) h(x) = 0 (l), g(x) 0 (m), x X The algorithms can be roughly divided the following way: ˆ primal methods: find descent direction keeping inside

More information

Matrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A =

Matrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A = 30 MATHEMATICS REVIEW G A.1.1 Matrices and Vectors Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A = a 11 a 12... a 1N a 21 a 22... a 2N...... a M1 a M2... a MN A matrix can

More information

NOTES ON LINEAR ALGEBRA CLASS HANDOUT

NOTES ON LINEAR ALGEBRA CLASS HANDOUT NOTES ON LINEAR ALGEBRA CLASS HANDOUT ANTHONY S. MAIDA CONTENTS 1. Introduction 2 2. Basis Vectors 2 3. Linear Transformations 2 3.1. Example: Rotation Transformation 3 4. Matrix Multiplication and Function

More information

Projection methods to solve SDP

Projection methods to solve SDP Projection methods to solve SDP Franz Rendl http://www.math.uni-klu.ac.at Alpen-Adria-Universität Klagenfurt Austria F. Rendl, Oberwolfach Seminar, May 2010 p.1/32 Overview Augmented Primal-Dual Method

More information

Primal-dual path-following algorithms for circular programming

Primal-dual path-following algorithms for circular programming Primal-dual path-following algorithms for circular programming Baha Alzalg Department of Mathematics, The University of Jordan, Amman 1194, Jordan July, 015 Abstract Circular programming problems are a

More information

Primal-dual relationship between Levenberg-Marquardt and central trajectories for linearly constrained convex optimization

Primal-dual relationship between Levenberg-Marquardt and central trajectories for linearly constrained convex optimization Primal-dual relationship between Levenberg-Marquardt and central trajectories for linearly constrained convex optimization Roger Behling a, Clovis Gonzaga b and Gabriel Haeser c March 21, 2013 a Department

More information

Semidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 2

Semidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 2 Semidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 2 Instructor: Farid Alizadeh Scribe: Xuan Li 9/17/2001 1 Overview We survey the basic notions of cones and cone-lp and give several

More information

Mathematical Methods wk 2: Linear Operators

Mathematical Methods wk 2: Linear Operators John Magorrian, magog@thphysoxacuk These are work-in-progress notes for the second-year course on mathematical methods The most up-to-date version is available from http://www-thphysphysicsoxacuk/people/johnmagorrian/mm

More information

Penalty and Barrier Methods General classical constrained minimization problem minimize f(x) subject to g(x) 0 h(x) =0 Penalty methods are motivated by the desire to use unconstrained optimization techniques

More information

Lecture 2: Linear Algebra Review

Lecture 2: Linear Algebra Review EE 227A: Convex Optimization and Applications January 19 Lecture 2: Linear Algebra Review Lecturer: Mert Pilanci Reading assignment: Appendix C of BV. Sections 2-6 of the web textbook 1 2.1 Vectors 2.1.1

More information

This appendix provides a very basic introduction to linear algebra concepts.

This appendix provides a very basic introduction to linear algebra concepts. APPENDIX Basic Linear Algebra Concepts This appendix provides a very basic introduction to linear algebra concepts. Some of these concepts are intentionally presented here in a somewhat simplified (not

More information

An Alternative Proof of Primitivity of Indecomposable Nonnegative Matrices with a Positive Trace

An Alternative Proof of Primitivity of Indecomposable Nonnegative Matrices with a Positive Trace An Alternative Proof of Primitivity of Indecomposable Nonnegative Matrices with a Positive Trace Takao Fujimoto Abstract. This research memorandum is aimed at presenting an alternative proof to a well

More information

Foundations of Matrix Analysis

Foundations of Matrix Analysis 1 Foundations of Matrix Analysis In this chapter we recall the basic elements of linear algebra which will be employed in the remainder of the text For most of the proofs as well as for the details, the

More information

Recall the convention that, for us, all vectors are column vectors.

Recall the convention that, for us, all vectors are column vectors. Some linear algebra Recall the convention that, for us, all vectors are column vectors. 1. Symmetric matrices Let A be a real matrix. Recall that a complex number λ is an eigenvalue of A if there exists

More information

Interior-Point Methods for Linear Optimization

Interior-Point Methods for Linear Optimization Interior-Point Methods for Linear Optimization Robert M. Freund and Jorge Vera March, 204 c 204 Robert M. Freund and Jorge Vera. All rights reserved. Linear Optimization with a Logarithmic Barrier Function

More information

The Jordan Algebraic Structure of the Circular Cone

The Jordan Algebraic Structure of the Circular Cone The Jordan Algebraic Structure of the Circular Cone Baha Alzalg Department of Mathematics, The University of Jordan, Amman 1194, Jordan Abstract In this paper, we study and analyze the algebraic structure

More information

Linear Algebra March 16, 2019

Linear Algebra March 16, 2019 Linear Algebra March 16, 2019 2 Contents 0.1 Notation................................ 4 1 Systems of linear equations, and matrices 5 1.1 Systems of linear equations..................... 5 1.2 Augmented

More information

An Infeasible Interior-Point Algorithm with full-newton Step for Linear Optimization

An Infeasible Interior-Point Algorithm with full-newton Step for Linear Optimization An Infeasible Interior-Point Algorithm with full-newton Step for Linear Optimization H. Mansouri M. Zangiabadi Y. Bai C. Roos Department of Mathematical Science, Shahrekord University, P.O. Box 115, Shahrekord,

More information

AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES

AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES JOEL A. TROPP Abstract. We present an elementary proof that the spectral radius of a matrix A may be obtained using the formula ρ(a) lim

More information

Rigid Geometric Transformations

Rigid Geometric Transformations Rigid Geometric Transformations Carlo Tomasi This note is a quick refresher of the geometry of rigid transformations in three-dimensional space, expressed in Cartesian coordinates. 1 Cartesian Coordinates

More information

Lecture 7: Positive Semidefinite Matrices

Lecture 7: Positive Semidefinite Matrices Lecture 7: Positive Semidefinite Matrices Rajat Mittal IIT Kanpur The main aim of this lecture note is to prepare your background for semidefinite programming. We have already seen some linear algebra.

More information

Chap 3. Linear Algebra

Chap 3. Linear Algebra Chap 3. Linear Algebra Outlines 1. Introduction 2. Basis, Representation, and Orthonormalization 3. Linear Algebraic Equations 4. Similarity Transformation 5. Diagonal Form and Jordan Form 6. Functions

More information

Research Reports on Mathematical and Computing Sciences

Research Reports on Mathematical and Computing Sciences ISSN 1342-284 Research Reports on Mathematical and Computing Sciences Exploiting Sparsity in Linear and Nonlinear Matrix Inequalities via Positive Semidefinite Matrix Completion Sunyoung Kim, Masakazu

More information

Conceptual Questions for Review

Conceptual Questions for Review Conceptual Questions for Review Chapter 1 1.1 Which vectors are linear combinations of v = (3, 1) and w = (4, 3)? 1.2 Compare the dot product of v = (3, 1) and w = (4, 3) to the product of their lengths.

More information

Interior Point Methods in Mathematical Programming

Interior Point Methods in Mathematical Programming Interior Point Methods in Mathematical Programming Clóvis C. Gonzaga Federal University of Santa Catarina, Brazil Journées en l honneur de Pierre Huard Paris, novembre 2008 01 00 11 00 000 000 000 000

More information

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88 Math Camp 2010 Lecture 4: Linear Algebra Xiao Yu Wang MIT Aug 2010 Xiao Yu Wang (MIT) Math Camp 2010 08/10 1 / 88 Linear Algebra Game Plan Vector Spaces Linear Transformations and Matrices Determinant

More information

L. Vandenberghe EE236C (Spring 2016) 18. Symmetric cones. definition. spectral decomposition. quadratic representation. log-det barrier 18-1

L. Vandenberghe EE236C (Spring 2016) 18. Symmetric cones. definition. spectral decomposition. quadratic representation. log-det barrier 18-1 L. Vandenberghe EE236C (Spring 2016) 18. Symmetric cones definition spectral decomposition quadratic representation log-det barrier 18-1 Introduction This lecture: theoretical properties of the following

More information

ECE 275A Homework #3 Solutions

ECE 275A Homework #3 Solutions ECE 75A Homework #3 Solutions. Proof of (a). Obviously Ax = 0 y, Ax = 0 for all y. To show sufficiency, note that if y, Ax = 0 for all y, then it must certainly be true for the particular value of y =

More information

Symmetric matrices and dot products

Symmetric matrices and dot products Symmetric matrices and dot products Proposition An n n matrix A is symmetric iff, for all x, y in R n, (Ax) y = x (Ay). Proof. If A is symmetric, then (Ax) y = x T A T y = x T Ay = x (Ay). If equality

More information

Lecture: Examples of LP, SOCP and SDP

Lecture: Examples of LP, SOCP and SDP 1/34 Lecture: Examples of LP, SOCP and SDP Zaiwen Wen Beijing International Center For Mathematical Research Peking University http://bicmr.pku.edu.cn/~wenzw/bigdata2018.html wenzw@pku.edu.cn Acknowledgement:

More information

Linear Algebra Review. Vectors

Linear Algebra Review. Vectors Linear Algebra Review 9/4/7 Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka http://cs.gmu.edu/~kosecka/cs682.html Virginia de Sa (UCSD) Cogsci 8F Linear Algebra review Vectors

More information

Semidefinite Programming

Semidefinite Programming Semidefinite Programming Notes by Bernd Sturmfels for the lecture on June 26, 208, in the IMPRS Ringvorlesung Introduction to Nonlinear Algebra The transition from linear algebra to nonlinear algebra has

More information

Properties of Matrices and Operations on Matrices

Properties of Matrices and Operations on Matrices Properties of Matrices and Operations on Matrices A common data structure for statistical analysis is a rectangular array or matris. Rows represent individual observational units, or just observations,

More information

A Second Full-Newton Step O(n) Infeasible Interior-Point Algorithm for Linear Optimization

A Second Full-Newton Step O(n) Infeasible Interior-Point Algorithm for Linear Optimization A Second Full-Newton Step On Infeasible Interior-Point Algorithm for Linear Optimization H. Mansouri C. Roos August 1, 005 July 1, 005 Department of Electrical Engineering, Mathematics and Computer Science,

More information

Linear Algebra. Min Yan

Linear Algebra. Min Yan Linear Algebra Min Yan January 2, 2018 2 Contents 1 Vector Space 7 1.1 Definition................................. 7 1.1.1 Axioms of Vector Space..................... 7 1.1.2 Consequence of Axiom......................

More information

j=1 u 1jv 1j. 1/ 2 Lemma 1. An orthogonal set of vectors must be linearly independent.

j=1 u 1jv 1j. 1/ 2 Lemma 1. An orthogonal set of vectors must be linearly independent. Lecture Notes: Orthogonal and Symmetric Matrices Yufei Tao Department of Computer Science and Engineering Chinese University of Hong Kong taoyf@cse.cuhk.edu.hk Orthogonal Matrix Definition. Let u = [u

More information

CONSTRAINED NONLINEAR PROGRAMMING

CONSTRAINED NONLINEAR PROGRAMMING 149 CONSTRAINED NONLINEAR PROGRAMMING We now turn to methods for general constrained nonlinear programming. These may be broadly classified into two categories: 1. TRANSFORMATION METHODS: In this approach

More information

STAT200C: Review of Linear Algebra

STAT200C: Review of Linear Algebra Stat200C Instructor: Zhaoxia Yu STAT200C: Review of Linear Algebra 1 Review of Linear Algebra 1.1 Vector Spaces, Rank, Trace, and Linear Equations 1.1.1 Rank and Vector Spaces Definition A vector whose

More information

LMI MODELLING 4. CONVEX LMI MODELLING. Didier HENRION. LAAS-CNRS Toulouse, FR Czech Tech Univ Prague, CZ. Universidad de Valladolid, SP March 2009

LMI MODELLING 4. CONVEX LMI MODELLING. Didier HENRION. LAAS-CNRS Toulouse, FR Czech Tech Univ Prague, CZ. Universidad de Valladolid, SP March 2009 LMI MODELLING 4. CONVEX LMI MODELLING Didier HENRION LAAS-CNRS Toulouse, FR Czech Tech Univ Prague, CZ Universidad de Valladolid, SP March 2009 Minors A minor of a matrix F is the determinant of a submatrix

More information

Solving large Semidefinite Programs - Part 1 and 2

Solving large Semidefinite Programs - Part 1 and 2 Solving large Semidefinite Programs - Part 1 and 2 Franz Rendl http://www.math.uni-klu.ac.at Alpen-Adria-Universität Klagenfurt Austria F. Rendl, Singapore workshop 2006 p.1/34 Overview Limits of Interior

More information

Second Order Cone Programming Relaxation of Positive Semidefinite Constraint

Second Order Cone Programming Relaxation of Positive Semidefinite Constraint Research Reports on Mathematical and Computing Sciences Series B : Operations Research Department of Mathematical and Computing Sciences Tokyo Institute of Technology 2-12-1 Oh-Okayama, Meguro-ku, Tokyo

More information

A Primal-Dual Second-Order Cone Approximations Algorithm For Symmetric Cone Programming

A Primal-Dual Second-Order Cone Approximations Algorithm For Symmetric Cone Programming A Primal-Dual Second-Order Cone Approximations Algorithm For Symmetric Cone Programming Chek Beng Chua Abstract This paper presents the new concept of second-order cone approximations for convex conic

More information

Knowledge Discovery and Data Mining 1 (VO) ( )

Knowledge Discovery and Data Mining 1 (VO) ( ) Knowledge Discovery and Data Mining 1 (VO) (707.003) Review of Linear Algebra Denis Helic KTI, TU Graz Oct 9, 2014 Denis Helic (KTI, TU Graz) KDDM1 Oct 9, 2014 1 / 74 Big picture: KDDM Probability Theory

More information

An Introduction to Correlation Stress Testing

An Introduction to Correlation Stress Testing An Introduction to Correlation Stress Testing Defeng Sun Department of Mathematics and Risk Management Institute National University of Singapore This is based on a joint work with GAO Yan at NUS March

More information

Appendix A Taylor Approximations and Definite Matrices

Appendix A Taylor Approximations and Definite Matrices Appendix A Taylor Approximations and Definite Matrices Taylor approximations provide an easy way to approximate a function as a polynomial, using the derivatives of the function. We know, from elementary

More information

A full-newton step infeasible interior-point algorithm for linear programming based on a kernel function

A full-newton step infeasible interior-point algorithm for linear programming based on a kernel function A full-newton step infeasible interior-point algorithm for linear programming based on a kernel function Zhongyi Liu, Wenyu Sun Abstract This paper proposes an infeasible interior-point algorithm with

More information

Solving Dual Problems

Solving Dual Problems Lecture 20 Solving Dual Problems We consider a constrained problem where, in addition to the constraint set X, there are also inequality and linear equality constraints. Specifically the minimization problem

More information

On implementing a primal-dual interior-point method for conic quadratic optimization

On implementing a primal-dual interior-point method for conic quadratic optimization On implementing a primal-dual interior-point method for conic quadratic optimization E. D. Andersen, C. Roos, and T. Terlaky December 18, 2000 Abstract Conic quadratic optimization is the problem of minimizing

More information

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 2

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 2 EE/ACM 150 - Applications of Convex Optimization in Signal Processing and Communications Lecture 2 Andre Tkacenko Signal Processing Research Group Jet Propulsion Laboratory April 5, 2012 Andre Tkacenko

More information

1 Outline Part I: Linear Programming (LP) Interior-Point Approach 1. Simplex Approach Comparison Part II: Semidenite Programming (SDP) Concludin

1 Outline Part I: Linear Programming (LP) Interior-Point Approach 1. Simplex Approach Comparison Part II: Semidenite Programming (SDP) Concludin Sensitivity Analysis in LP and SDP Using Interior-Point Methods E. Alper Yldrm School of Operations Research and Industrial Engineering Cornell University Ithaca, NY joint with Michael J. Todd INFORMS

More information

1. General Vector Spaces

1. General Vector Spaces 1.1. Vector space axioms. 1. General Vector Spaces Definition 1.1. Let V be a nonempty set of objects on which the operations of addition and scalar multiplication are defined. By addition we mean a rule

More information

Lecture notes: Applied linear algebra Part 1. Version 2

Lecture notes: Applied linear algebra Part 1. Version 2 Lecture notes: Applied linear algebra Part 1. Version 2 Michael Karow Berlin University of Technology karow@math.tu-berlin.de October 2, 2008 1 Notation, basic notions and facts 1.1 Subspaces, range and

More information

Chapter 7. Linear Algebra: Matrices, Vectors,

Chapter 7. Linear Algebra: Matrices, Vectors, Chapter 7. Linear Algebra: Matrices, Vectors, Determinants. Linear Systems Linear algebra includes the theory and application of linear systems of equations, linear transformations, and eigenvalue problems.

More information

Ω R n is called the constraint set or feasible set. x 1

Ω R n is called the constraint set or feasible set. x 1 1 Chapter 5 Linear Programming (LP) General constrained optimization problem: minimize subject to f(x) x Ω Ω R n is called the constraint set or feasible set. any point x Ω is called a feasible point We

More information

ALGEBRA QUALIFYING EXAM PROBLEMS LINEAR ALGEBRA

ALGEBRA QUALIFYING EXAM PROBLEMS LINEAR ALGEBRA ALGEBRA QUALIFYING EXAM PROBLEMS LINEAR ALGEBRA Kent State University Department of Mathematical Sciences Compiled and Maintained by Donald L. White Version: August 29, 2017 CONTENTS LINEAR ALGEBRA AND

More information

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v ) Section 3.2 Theorem 3.6. Let A be an m n matrix of rank r. Then r m, r n, and, by means of a finite number of elementary row and column operations, A can be transformed into the matrix ( ) Ir O D = 1 O

More information

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2. APPENDIX A Background Mathematics A. Linear Algebra A.. Vector algebra Let x denote the n-dimensional column vector with components 0 x x 2 B C @. A x n Definition 6 (scalar product). The scalar product

More information