CONJUGATE GRADIENT WITH SUBSPACE OPTIMIZATION

Size: px
Start display at page:

Download "CONJUGATE GRADIENT WITH SUBSPACE OPTIMIZATION"

Transcription

1 CONJUGATE GRADIENT WITH SUBSPACE OPTIMIZATION SAHAR KARIMI AND STEPHEN VAVASIS Abstract. In this paper we present a variant of the conjugate gradient (CG) agorithm in which we invoke a subspace minimization subprobem on each iteration. We ca this agorithm CGSO for conjugate gradient with subspace optimization. It is reated to earier work by Nemirovsky and Yudin. We appy the agorithm to sove unconstrained stricty convex probems. As with other CG agorithms, the update step on each iteration is a inear combination of the ast gradient and ast update. Unike some other conjugate gradient methods, our agorithm attains a theoretica compexity bound of O( / og(1/ɛ)), where the ratio / characterizes the strong convexity of the objective function. In practice, CGSO competes with other CG-type agorithms by incorporating some second order information in each iteration. Key words. Conjugate Gradient, First Order Agorithms, Convex Optimization AMS subject cassifications. 1. Introduction. The method of conjugate gradients (CG) was introduced by Hestenes and Stiefe in 1952 [8] to minimize quadratic objective functions of the form f(x) = x t Ax/2 b t x (or equivaenty, to sove Ax = b), where A is a symmetric positive definite matrix. We refer to this agorithm as inear conjugate gradient. ater, the agorithm was generaized to noninear conjugate gradient which addresses unconstrained minimization of arbitrary differentiabe objective functions, by Fetcher and Reeves [3] and Poak and Ribiére [13], and others. Noninear CG agorithms a have the property that when appied to a quadratic objective function and couped with an exact ine search, they reduce to inear CG. None of these agorithms, however, has a known compexity bound when appied to nonquadratic functions. Indeed, Nemirovsky and Yudin [9] argue that their worst-case compexity for stricty convex functions is quite poor. Nemirovsky and Yudin, on the other hand, found a different generaization of CG. Actuay, their agorithm, which we refer to as NYCG, is not a generaization of the CG agorithm itsef but rather of its derivation from some equations and inequaities that underie it. Their agorithm achieves a worst-case compexity bound of O(n(1/ɛ) /). Here ɛ is the desired reative accuracy, that is, ɛ = (f(x n ) f(x ))/(f(x 0 ) f(x )), where x 0 is the starting point, x is the optimizer, and x n is the fina iterate, and and characterize the convexity of f. In particuar, we assume that for a x, y ying in the eve set of x 0, f(x) f(y) x y, (1.1) f(y) f(x) f(x), y x + 2 y x 2. (1.2) For exampe, in the case of a convex quadratic function f(x) = x t Ax/2 b t x, / is the condition number of A. From inequaity (1.1), it foows that f(y) f(x) f(x), y x + 2 y x 2, (1.3) which wi be usefu in our anaysis. One drawback of the NYCG agorithm is that in the case of quadratic objective functions, it does not reduce to inear CG (and in fact, is much sower in practice). A second drawback is that it requires prior knowedge of the ratio /. This is because 1

2 2 Sahar Karimi and Stephen Vavasis the agorithm needs to be restarted every O( /) iterations in order to achieve the above-mentioned compexity bound. A further disadvantage of NYCG is that it has a reativey expensive computation on every iteration, namey, one must sove a two-dimensiona convex optimization probem using the eipsoid method. This drawback was ater remedied by a different agorithm due to Nesterov [10]. Nesterov s agorithm generates two sequences of iterates; the first sequence of iterates is generated such that a sufficient reduction in objective function is achieved, and the new iterate in the second sequence is an affine combination of the ast two iterates of the first sequence. Nesterov s agorithm, however, sti has the two disadvantages mentioned in the previous paragraph. The purpose of this paper is to deveop an agorithm that we ca conjugate gradient with subspace optimization (CGSO) that mainy avoids the disadvantages mentioned above. CGSO is mosty cosey reated to NYCG. In particuar, we seek a conjugate-gradient-ike agorithm with the foowing properties. 1. The agorithm shoud reduce to inear CG when the objective function is quadratic. 2. The agorithm shoud have the compexity bound of O(n(1/ɛ) /) iterations for convex functions satisfying (1.1) and (1.2). 3. The agorithm shoud not require prior knowedge of any parameters describing f.. The cost per iteration shoud not be excessive. The agorithm we propose achieves goas 1 3, and mosty achieves goa. ike NYCG, our agorithm must sove a convex optimization subprobem on every iteration. The dimension of the subprobem is aways at east 2 and is determined adaptivey by the agorithm. The worst-case upper bound we are abe to prove for the dimension of this subprobem is O(og j), where j is the number of iterations so far. However, in our testing, the dimension of the subprobem was 2 in amost every case; in one test case it reached the vaue 3 for a few iterations, but it never exceeded 3. Furthermore, we find that soving the subprobem is usuay fairy efficient because we usuay appy Newton s method, using automatic differentiation to obtain the necessary first and second derivatives. Goa 3, the condition of no prior knowedge of parameters, has the obvious benefit of making the agorithm easier to appy in practice. It aso has a second subter benefit. For some convex probems, e.g., minimization of a og-barrier function, there is no prior bound on / over the domain of the function since the derivatives tend to infinity at the boundaries. However, as the minimizer is approached, the bad behavior at the boundaries becomes irreevant, and there is a new smaer ratio / reevant for eve sets in the neighborhood of the minimizer. In this case, CGSO automaticay adapts to the improved vaue of /. Such adaptation is possibe aso with NYCG, and Nesterov s agorithm too, but, as far as we know, the adaptation must be done by the user and cannot be easiy automated. Methods reviewed in this paper are among techniques that are generay referred to as first-order agorithms, because they use ony the first derivative information of the function in each iteration. Due to the successfu theory of NYCG and Nesterov techniques, first-order agorithms have attracted many researchers during the ast decade and have been extended to soving different casses of probems. Nesterov in [11] propsed a variation of his earier agorithms for minimizing a nonsmooth function. In addition to nonsmooth optimization, Nesterov agorithm has been adapted for constrained probems with simpe enough feasibe regions so that a projection on these

3 Conjugate Gradient with Subspace Optimization 3 sets can be easiy computed. One may refer to [15] and references therein for a more eaborated discussion on different adaptations of Nesterov s agorithm. The focus of this paper, however, is more on CG agorithm and not on first-order techniques in genera. We now provide further background on conjugate gradient. The origina inear CG has the foowing form: ( ) r x j+1 = x j j t d j (d j ) t Ad, (1.a) j d j+1 = r j+1 + (r j+1) t Ad j (d j ) t Ad j d j. (1.b) In the above equations r j is f(x) = Ax j b and d 0 = r 0. It is possibe to show that the number of iterations in inear CG is bounded by the dimension of the probem, n. For more detais on inear CG, one may refer to [5] or [12]. Noninear CG was proposed by Fetcher and Reeves [3] as an adaptation of the above agorithm for minimizing a genera noninear function. The genera form of this agorithm is as foows: x j+1 = x j + α j d j, (1.5a) d j+1 = g j+1 + β j d j. (1.5b) Here, d j is the search direction at each iteration, g j+1 is the gradient of the function at (j +1)th iterate, i.e., f(x j+1 ); and α j is the step size, usuay determined by a ine search. Different updating rues for β j give us different variants of noninear CG. The most common formuas for computing β j are: Fetcher-Reeves (196): β F R = gj+1 g j, Poak-Ribière (1969): β P R = (gj+1 ) t (g j+1 g j ) g j. Hager and Zhang [6] present a compete ist of a updating rues in their survey on noninear CG. The convergence of noninear CG is highy dependent on the ine search; for some the exact ine search is crucia. There are numerous papers devoted to the study of goba convergence of noninear CG agorithms, most of which discuss variants of noninear CG that do not rey on exact ine search to be gobay convergent. A- Baai [1] discusses the convergence of Fetcher-Reeves agorithm with exact ine search. Gibert and Noceda [] study the convergence of noninear CG agorithms with no restart and no exact ine search. Dai and Yuan [2] present a noninear CG for which the standard Wofe condition suffices. A recent variant of CG has been proposed by Hager and Zhang [7] that reies on a ine search satisfying the Wofe Conditions. Furthermore this agorithm has the advantage that every search direction is a descent direction, which is not necessariy the case in noninear CG. From Yuan and Stoer s perspective [16], CG is a technique in which the search direction d j+1 ies in the subspace spanned by Sp{g j, d j }. In the agorithm they propose they compute the new search direction by minimizing a quadratic approximation of the objective function over the mentioned subspace. A more generaized

4 Sahar Karimi and Stephen Vavasis form of CG caed Heavy Ba Method, was introduced by Poyak [1], in which x j+1 is x j +α( g j )+β(x j x j 1 ). He proved a geometric progression rate for this agorithm when α and β beong to a specific range. In the same work, Poyak reviews the CG method; our simpest form of CGSO, which is given in section 2.1, coincides with his presentation of CG. The remainder of this paper is devoted to CGSO and its properties. In section 2 we describe the agorithm. Our main resut on the convergence of the agorithm is presented in section 2.3. The impementation of the agorithm is discussed in section 3; furthermore the resuts and comparison of CGSO with CG are presented in this section. 2. CGSO. In this section we present CGSO for soving the probem min f (x), (2.1) x Rn where f(x) is a stricty convex function characterized by parameters and. We foow the standard notation throughout this paper..,. represents the inner product of two vectors in proper dimension, and. stands for the 2-norm of a vector uness otherwise is stated. Bod ower case characters and upper case characters are used for vectors and matrices respectivey; and their superscript states the iteration count The Agorithm. et g(x) denote the derivative of f(x); CGSO, in its preiminary form, is as foows: x 0 = arbitrary; for j = 1, 2,... x j+1 = x j + α j g(x j ) + β j d j where d j = x j x j 1 and α j, β j = argmin α,β f(x j + αg(x j ) + βd j ) As we sha see, some modifications are necessary to achieve the desired compexity bound. The above agorithm is certainy a generaization of noninear CG, since each search direction is a combination of previous gradients. Notice that the above agorithm is a descent method (i.e. f(x j+1 ) f(x j )). Furthermore, the above agorithm performs at east as we as steepest descent in each iteration; hence it converges to a stationary point, which, for the cass of convex probems, coincides with the minimizer of the function. For convenience, et us represent g(x j ) by g j, and et v f (x) denote the residua of the function, i.e. f(x) f(x ). By strong convexity of the function, we get the foowing properties for the above agorithm: (a) f(x j+1 ) f(x j ) 1 2 gj 2 (b) g j, x x j f(x ) f(x j ) (c) v f (x 0 ) = f(x 0 ) f 2 (x x 0 ) 2 Property (a) foows from (1.3) and the fact that f(x j+1 ) f(x j 1 gj ). Property (b) is true by convexity of the function, and property (c) is a direct derivation from inequaity (1.2). We are now ready to present the main emma on the compexity bound of the agorithm. emma 2.1. Suppose m 8ρ and ( f(x ) f(x 0 ) ) λ j + λ j g j, x j x 0 < 0, (2.2)

5 Conjugate Gradient with Subspace Optimization 5 and λ j g j ρ (λ j ) 2 g j 2, (2.3) are satisfied, where ρ is a constant 1, and λ j f(x = j ) f(x j+1 ) g j 2. Then the residua of the function is divided in haf after m iterations; i.e. v f (x m ) 1 2 v f (x 0 ). Proof. Our proof is an extension of the proof in section 7.3 in [9]. Suppose by contradiction that m 8ρ, (2.2) and (2.3) are satisfied; but v f (x m ) > v f (x 0 ) 2. By definition of λ j, hence f(x j+1 ) = f(x j ) ( λ j) 2 g j 2, v f (x j+1 ) = v f (x j ) ( λ j) 2 g j 2. Summing these inequaities over j = 0,, m 1, we get: or equivaenty, 0 v f (x m ) = v f (x 0 ) ( λ j ) 2 g j 2, ( λ j ) 2 g j 2 v f (x 0 ). (2.) By convexity of the function we have, g j, x x j f(x ) + f(x j ) = v f (x j ), and so g j, x x 0 g j, x j x 0 v f (x j ) v f (x m ) < v f (x 0 ). 2 et s consider the weighted sum of a the above inequaities for j = 0,... m 1 with weights λ j s to get: λ j g j, x x 0 λ j g j, x j x 0 < v f (x 0 ) λ j, 2 which can be rearranged to the foowing form, λ j g j, x x 0 < v f (x 0 ) 2 λ j + λ j g j, x j x 0.

6 6 Sahar Karimi and Stephen Vavasis Equivaenty we can rewrite the above inequaity as: λ j g j, x x 0 < v f (x 0 ) λ j + f(x ) f(x 0 ) λ j + λ j g j, x j x 0. Using inequaity (2.2) aong with the facts that f(x ) f(x j ) and λ j 0 for a j, we get: λ j g j, x x 0 < v f (x 0 ) λ j. (2.5) By the Cauchy-Schwarz inequaity we have λ j g j x x 0 λ j g j, x x 0 < v f (x 0 ) λ j, hence λ j g j x x 0 > v f (x 0 ) λ j. (2.6) By property (c) we have x x 0 2vf (x 0 ). (2.7) Furthermore, by inequaities (2.3) and (2.) we get: λ j g j ρ (λ j ) 2 g j 2 ρ v f (x 0 ). (2.8) Repacing inequaities (2.7) and (2.8) in inequaity (2.6), we get 2vf ρ v f (x 0 (x ) 0 ) > v f (x 0 ) λ j. (2.9) Notice that by definition of λ and property (a), λ j λ j 1 2 m. Using this fact in inequaity (2.9), we get 2vf ρ v f (x 0 (x ) 0 ) > v f (x 0 ) 1 2 ( ) 1 2 m, for a j, so

7 Conjugate Gradient with Subspace Optimization 7 therefore m < 8ρ, (2.10) which contradicts our assumption on the vaue of m. emma 2.1 shows that under conditions (2.2) and (2.3), the residua of the function is divided in haf every m = O( ) iterations. For the next sequence of m iterations, a further reduction of 1 2 is achieved provided (2.2) and (2.3) hod, with x m substituted in pace of x 0. Hence by etting x m be the new x 0 and repeating the same agorithm, we can find the ɛ-optima soution in 1 og 2 ɛ 8ρ iterations. Of course, m is not known to our agorithm, a point to which we return in the next section Restarting. As mentioned before, the agorithm shoud use first 0, then m, then 2m and so on when checking (2.2) and (2.3) in order to achieve the desired compexity bound. We use the term restarting to refer to the process of repacing x 0 in inequaities (2.2) and (2.3) with x m for some m > 0. Notice that this process does not change any of the iterates, and it ony changes the interva of indices in which we check inequaities (2.2) and (2.3). If we know parameters of the function, i.e. and, we can compute m directy; hence we woud be abe to determine exacty when we need to restart the agorithm. In many cases, however, finding the parameters of the function is a nontrivia probem itsef. In order to have an agorithm which does not rey on any prior knowedge about the function, we propose the foowing technique. Since 2 p m 2 p+1 for some p, by restarting the agorithm every 2 p+1 iterations, we are guaranteed that the residua of the function is divided in haf between every two consecutive restarts which contains at most 2m = 16ρ iterations, where the ast statement foows from the fact that 2 p+1 2m. Since the exact vaue of p is usuay unknown, we repeat the above procedure for every vaue of p {P,..., P u }. In the section on impementation of the agorithm we wi discuss the range of p we used, however note that P u does not need to exceed og 2 j, where j is the current iteration count. This is because inequaities (2.2) and (2.3) are the same for a p og 2 j. We can now state the CGSO agorithm in a more compete form: x 0 = arbitrary; ρ : given; For j = 1, 2,... x j+1 = argmin f(x) where E j = x j + Sp { g j, d j} x E j et λ j = f(x j ) f(x j+1 ) g j 2 For p = P,..., og 2 j If j + 1 = k p 2 p for some integer k p et r p = (k p 1) 2 p ; check (f(x j+1 ) f(x rp )) and ( j λ i ) + j λ i g i, x i x rp < 0 j λ i g i j ρ (λ i ) 2 g i 2 If any of the above inequaities fais, take the correction step. (which wi be defined in the subsequent section)

8 8 Sahar Karimi and Stephen Vavasis 2.3. Correction Step. et us refer to the set of iterates between two consecutive restart as a bock of iterates; in other words for any p, the set of iterates x 0, x 1,..., x 2p 1 is the first bock of size 2 p, x 2p, x 2p +1,..., x 2(2p ) 1 is the second bock of size 2 p, and so on. At the end of each bock we check inequaities (2.2) and (2.3). If they are satisfied and 2 p 8ρ, then by emma 2.1 we know that the residua of the function is divided in haf; however if any of these inequaities fais, then as mentioned in the previous section we need to take correction step for the next bock of iterates. The correction step is basicay computing the next bock of iterates in a way that satisfaction of inequaities (2.2) and (2.3) is guaranteed at the end of this bock. Then the correction step is omitted in the subsequent bocks unti the inequaities are vioated again. Reca that in an ordinary step, the new iterate, x j+1 is cacuated by a 2- dimensiona search on the pane passing through x j, and parae to the 2-dimensiona subspace spanned by g j, and d j. Suppose at east one of the inequaities (2.2) and (2.3) is vioated for kth bock of p; i.e. for the bock of iterates x rp,..., x rp+2p 1 where r p = (k 1)2 p. Then for the next bock we search for the new iterate x j+1 on the space of x j + Sp { } g j, d j, q j p, x j x rp where q j p = j λ i g i. Finding the new iterate x j+1 through a search on the space that in addition to g j and d j incudes q j p and x j x rp is what we referred to as correction step. Notice that for each p with the vioated constraints we increase the dimension of the search space by 2. However, the dimension of the search space never exceeds O(P u ) = 2+2 og 2 j, which happens to be the case when the inequaities are vioated for a possibe vaues of p. It is quite easy to see that inequaities (2.2) and (2.3) are satisfied for the (k+1)th bock of p when we take the correction step throughout it. By KKT condition, we have g j, x j x rp = 0 for a j in this bock. Using this, aong with the fact that f(x j ) < f(x rp ) and non-negativity of λ j for a j, we derive (2.2). Simiary one can argue that by KKT g j, q j 1 p r p+2 p 1 = 0 for a j, hence λ i g i = r p+2 p 1 (λ i ) 2 g i 2, which means inequaity (2.3) is satisfied. After finding the iterates of one bock through a correction step, the agorithm switches back to taking a reguar step unti the next faiure of the inequaities. We can now present the agorithm in its entirety. Agorithm 1. x 0 = arbitrary; ρ : given; S =. for j = 1, 2,... x j+1 = argmin f(xj ) f(x j+1 ) g j 2 f(x) where E j = x j + Sp { } g j, d j, p S q j p, p S x j x rp x E j et λ j = for p = P,..., og 2 j if j + 1 = k p 2 p for some integer k p et r p = (k p 1) 2 p if p S S = S \ {p} ese (i.e. if p S), check

9 end Conjugate Gradient with Subspace Optimization 9 (f(x j+1 ) f(x rp )) and ( j λ i ) + j λ i g i, x i x rp < 0 j λ i g i j ρ (λ i ) 2 g i 2 if any of the above inequaities fais, S = S {p} end end end Theorem 2.2. Suppose m 8ρ, and x j is a sequence generated by agorithm 1 for soving probem (2.1); then v f (x (n+)m ) 1 2 v f (x nm ). Proof. et p P be the integer for which 2 p 1 m 2 p ; and et s p stand for 2 p. Using agorithm 1, we are guaranteed that for at east one of any two consecutive bocks of size s p inequaities (2.2) and (2.3) are satisfied. The size of this bock is s p m 8ρ and hence by emma 2.1 we have v f (x nm+2s p ) 1 2 v f (x nm ). (2.11) Since 2s p m, so f(x nm+m ) f(x nm+2s p ); hence v f (x nm+m ) v f (x nm+2s p ). (2.12) (2.11) and (2.12) gives us the resut we wanted to show. In our experiment with the agorithm, however, we ony incude the x j x rp term in the correction step. We wi discuss this and few other remarks on the impementation of the agorithm in the foowing section. 3. Impementation Soving Each Iteration. Obviousy the most important part in CGSO is soving the subspace optimization probem in each iteration. In our impementation of the agorithm, we used Newton s method for this task uness it fais to rapidy converge to the optimum. In the case of faiure of Newton s agorithm, the eipsoid method carries out the task of soving the optimization probem. In other words we assign an upper bound to the number of iterations that Newton s method may take, and if it fais to converge within the given number of iterations the agorithm switches to the eipsoid method for finding the next iterate. Reca that at (j + 1)th iterate, we search for x j+1 in the space of vectors x = x j + αg j + βd j + aq + br, where Q R n S is the matrix formed by coumns q j p for a p S; R is the matrix of the same dimension with coumns x j x rp for a p S; α, β R, and a, b R S are coefficients that we want to find. et y denote the variabe of the subspace optimization probem, i.e., y = [α, β, a t, b t ] t ; in addition et B = [ g j, d j, Q, R ] and K = S. We can now state the forma presentation of the subspace optimization probem, min f ( x j + By ) (3.1) y R K As mentioned above, we sove probem (3.1) with Newton s method. etting f(y) = f(x j +By) and using chain rue we get the foowing formuas for the gradient

10 10 Sahar Karimi and Stephen Vavasis and Hessian of each Newton s iteration, f(y) = B t f(x) (3.2) 2 f(y) = B t 2 f(x)b (3.3) Notice that some second order information of the function comes into pay in equation (3.3); and we beieve that this is one of the reasons that CGSO stands out in practice. We compute f(x) and 2 f(x) directy when f(x) is simpe enough. For more compicated functions we use automatic differentiation (AD) in backward mode to compute f(x) and 2 f(x)b. et B (k) denote kth coumn of matrix B. Backward AD enabes us to keep the computationa cost of f(x) within a constant factor of the objective function evauation cost; and the cost of computing 2 f(x)b (k) within a constant factor of the computationa cost of gradient evauation. The storage space required in backward AD, however, is more than the required storage in forward AD; and in worst case it can be proportiona to the number of operations required for computing f(x). In spite of that, we are abe to keep the storage required by backward AD for the cass of probems we studied in O(s), where s is the space required to compute f(x). Detais on our test probems are presented in section 3.2. For more information on AD, one may refer to [12]. In addition to the storage required by AD, we need to store x j, and matrix B; we aso need to update and store x rp, j λ i, j λ i g i, x i x rp j, λ i g i, j (λ i ) 2 g i 2 for a p {P,..., og 2 j}. In our experiment with CGSO we take P to be. We find that K is equa to 2 in every case except one, in which it reaches the vaue of 3. Hence the required storage space for the above eements is in O(n max{k, og 2 j}). A difficuty we need to overcome in soving a probem with CGSO, especiay when we are trying to achieve high accuracy, is round-off error. In particuar f(x j ) f(x j+1 ), which is required for computing λ j, gets more and more inaccurate as the iterates approach the optimum. To overcome this, we took advantage of the second order Tayor series expansion. We have a subroutine that anayze the absoute error of f(x j ) f(x j+1 ) computed directy and through Tayor series, i.e. f(x j ) f(x j+1 ) ( f(x j ) t (x j+1 x j ) (xj+1 x j ) t 2 f(x j )(x j+1 x j ); the one with smaer error is accepted. In some cases the error anaysis is not easy, hence a heuristic is used to choose the preferred formua. Computing the difference of two objective vaues is appeared in inequaity (2.2) as we, in f(x ) f(x 0 ); the same subroutine is used to compute this term Numerica Resuts. We have tested our agorithm on the foowing casses of probem: f 1 (x) = m i=1 og (at i x b i) f 2 (x) = c t x og (det (C Diag(x))) f 3 (x) = m i=1 (at i x b i) d where d is a given even integer. Functions f 1 and f 2 are og-barrier functions, and f 3 is an approximation to the infinity norm of the vector Ax b. Note that we need to restrict the degree, d, to even numbers to save the convexity of f 3. A the codes for CGSO, Eipsoid method, and automatic differentiation are written in MATAB. The Hessian of f 1 and f 3 can be derived easiy in cosed form. The Hessian of f 2 was obtained by appying backward mode AD by hand to a program that computes f 2.

11 Conjugate Gradient with Subspace Optimization 11 The stopping criteria we used for both CGSO and its subprobems are f(x) ɛ and f(x) ɛ N, respectivey. In our impementation, ɛ N at iteration j is B t f(x j ) 100 ; however, if the subprobem is soved by Newton s method, the obtained soution is usuay more accurate due to the fast oca convergence of the Newton s method. Notice that for f 1 and f 2 we have some hidden constraints, namey a t i x b i > 0 in f 1 and C Diag(x) 0 in f 2. Therefore CGSO switches to eipsoid method if Newton s method does not converge in 15 iterations or if the iterates get infeasibe. In [7], Hager and Zhang compared their variant of CG with -BFGS and PRP +, and estabished the superior performance of their variant of CG. Hence, we compare our agorithm with their variant of CG, represented by CG HZ in Tabe 3.1 that summarizes this comparison. The top portion of the tabe expains the instances that we generated randomy. Parameters m, n, d, and ɛ are defined as above; m f for f 2 represents c = m f e where e is the a ones vector. The point of this parameter is that arger vaues of m f force C Diag(x ) for the optimizer x coser to the boundary of the feasibe region, hence the probem become more i-conditioned. et A be the m by n matrix formed by a t i as its rows; ds. indicates the density, the percentage of nonzero entries of A for f 1 and f 3, and density of C for f 2. It. count for both CG HZ and CGSO indicates the tota number of major iteration the agorithm takes before convergence. S. count is the tota iteration performed by the ine search routine in CG HZ ; for each iterate of S we need one gradient evauation. In CGSO, one gradient and one Hessian evauation is required for each Newton s iterate; and one gradient is cacuated in each iteration of the eipsoid agorithm. As tabe 3.1 shows, the number of iterations CGSO takes is ess than CG HZ in a instances, and the number of iterations for soving subprobems is significanty ess than the number of ine search iterations; especiay for instances 3 and, this difference is considerabe. The ast row of the tabe impies that no correction step was taken in any of these instances, except one. Actuay the purpose of Ins. 7 is to show that correction step might be required for some probems. In this instance A is a square matrix of size with O(10 3 ) condition number. The correction step is taken throughout 7 bocks corresponding to P, for which the size of the subprobems reaches 3. This, however, is consistent with the theory because the arger the condition number is, the arger is; therefore 2 P may not be a good approximation of 8ρ Parameter ρ. Reca that ρ is a parameter of CGSO required for checking inequaity (2.3). As mentioned before, we do not check inequaity (2.3) in our impementation of the agorithm; instead we gather the vaues of ρ and study their trend. Tabe 3.2 summarizes the maximum vaue ρ reached in each instance over a vaues of p. Figure 3.1 and 3.2 depict ρ for Ins. 1 and Ins. 3, respectivey. We chose these instances because they have higher ρ max and iteration count in each category. Our observation suggests that a reasonaby sma bound for ρ shoud suffice. Notice that the potted vaues impy that ρ reaches its maximum for some p and then it decreases for higher vaues of p, so it does not grow with p. Furthermore, as we get coser to the optimum parameter ρ gets coser to 1. This is a common pattern for a test probems we had. Instead of fixing ρ one may adaptivey update this parameter. In other words, we can assign a vaue to ρ and if inequaity (2.3) fais for more than a certain number of bocks, then we increase ρ, and as the agorithm progresses towards the optimum we can decrease ρ.

12 12 Sahar Karimi and Stephen Vavasis Tabe 3.1 Comparison of CGSO and CG HZ f 1 f 2 f 3 Ins. 1 Ins. 2 Ins. 3 Ins. Ins. 5 Ins. 6 Ins. 7 m n m f d ds ɛ 1e-8 1e-8 1e-3 1e-8 1e-18 1e-18 1e-8 It. count CG HZ S. count It. count CGSO Newton eipsoid Corr Tabe 3.2 Parameter ρ Ins. 1 Ins. 2 Ins. 3 Ins. Ins. 5 Ins. 6 Ins. 7 ρ max Concusion. We presented CGSO for soving unconstrained strongy convex functions. CGSO is a variant of CG, since the update step in each iteration is a combination of previous gradients, and it reduces to inear CG when appied to a quadratic function. The coefficients of previous gradients, however, do not foow an updating rue and are found by a subspace optimization subprobem. We have discussed that these subprobems are mosty two dimensiona; in some cases the dimension of the subprobems may reach sighty higher vaues, but it certainy never exceeds O (og j), where j is the iteration count. We have aso shown that CGSO benefits from the optima compexity bound of O ( og ( 1 ɛ ) ) in theory. CGSO does not depend on any prior information about the function and can easiy be impemented. In practice, it outperforms the variant of CG proposed by Hager and Zhang [7] which is known to be stronger than other techniques such as -BFGS or PRP +. Practica efficiency of CGSO can be improved by incorporating some second order information of the function in soving the subprobems with Newton s method as discussed in the previous section.

13 Conjugate Gradient with Subspace Optimization p= p=5 p=6 p=7 p= Fig Parameter ρ for Ins p= p=5 p=6 p=7 p=8 p=9 p= Fig Parameter ρ for Ins. 3 REFERENCES [1] M. A-Baai. Descent property and goba convergence of the fetcher-reeves method with inexact ine search. IMA Journa of Numerica Anaysis, 5:121 12, [2] Y. H. Dai and Y. Yuan. A noninear conjugate gradient method with a strong goba convergence property. Siam Journa on Optimization, 10(1): , [3] R. Fetcher and C. M. Reeves. Function minimization by conjugate gradients. Computer Journa, 7:19 15, 196. [] J. C. Gibert and J. Noceda. Goba convergence properties of conjugate gradient methods for optimization. Siam Journa on Optimization, 2(1):21 2, [5] G. H. Goub and C. F. Van oan. Matrix Computations. Hopkins Fufiment Service, [6] W. W. Hager and H. Zhang. A survey of noninear conjugate gradient methods. [7] W. W. Hager and H. Zhang. A new conjugate gradient method with guaranteed descent and an efficient ine search. SIAM Journa on Optimization, 16(1): , [8] M. R. Hestenes and E. Stiefe. Methods of conjugate gradient for soving inear systems. Journa

14 1 Sahar Karimi and Stephen Vavasis of Research of the Nationa Bureau of Standards, 9:09 36, [9] A. S. Nemirovsky and D. B. Yudin. Probem Compexity and Method Efficiency in Optimization. John Wiey & sons, [10] Y. E. Nesterov. A method of soving a convex programming probem with convergence rate o( 1 k 2 ). Soviet mathematics, Dokady, 27(2): , [11] Y. E. Nesterov. Smooth minimization of nonsmooth functions. Math. Programming, 103: , [12] J. Noceda and S. J. Wright. Numerica Optimization. Springer Science, [13] E. Poak and G. Ribiére. Note sur a convergence de méthodes de directions conjuguées. Revue Francaise d informatique et de Recherche Opeŕationnee, 16:35 3, [1] Boris T. Poyak. Introduction to Optimization. Optimization Software Inc.,, [15] P. Tseng. On acceerated proxima gradient methods for convex-concave optimization [16] Y. Yuan and J. Stoer. A subspace study on conjugate gradient agorithms. ZAMM (Zeitschrift fr angewandte Mathematik und Mechanik), 11:69 77, 1995.

A Brief Introduction to Markov Chains and Hidden Markov Models

A Brief Introduction to Markov Chains and Hidden Markov Models A Brief Introduction to Markov Chains and Hidden Markov Modes Aen B MacKenzie Notes for December 1, 3, &8, 2015 Discrete-Time Markov Chains You may reca that when we first introduced random processes,

More information

CS229 Lecture notes. Andrew Ng

CS229 Lecture notes. Andrew Ng CS229 Lecture notes Andrew Ng Part IX The EM agorithm In the previous set of notes, we taked about the EM agorithm as appied to fitting a mixture of Gaussians. In this set of notes, we give a broader view

More information

Lecture Note 3: Stationary Iterative Methods

Lecture Note 3: Stationary Iterative Methods MATH 5330: Computationa Methods of Linear Agebra Lecture Note 3: Stationary Iterative Methods Xianyi Zeng Department of Mathematica Sciences, UTEP Stationary Iterative Methods The Gaussian eimination (or

More information

Convergence Property of the Iri-Imai Algorithm for Some Smooth Convex Programming Problems

Convergence Property of the Iri-Imai Algorithm for Some Smooth Convex Programming Problems Convergence Property of the Iri-Imai Agorithm for Some Smooth Convex Programming Probems S. Zhang Communicated by Z.Q. Luo Assistant Professor, Department of Econometrics, University of Groningen, Groningen,

More information

Uniprocessor Feasibility of Sporadic Tasks with Constrained Deadlines is Strongly conp-complete

Uniprocessor Feasibility of Sporadic Tasks with Constrained Deadlines is Strongly conp-complete Uniprocessor Feasibiity of Sporadic Tasks with Constrained Deadines is Strongy conp-compete Pontus Ekberg and Wang Yi Uppsaa University, Sweden Emai: {pontus.ekberg yi}@it.uu.se Abstract Deciding the feasibiity

More information

Problem set 6 The Perron Frobenius theorem.

Problem set 6 The Perron Frobenius theorem. Probem set 6 The Perron Frobenius theorem. Math 22a4 Oct 2 204, Due Oct.28 In a future probem set I want to discuss some criteria which aow us to concude that that the ground state of a sef-adjoint operator

More information

First-Order Corrections to Gutzwiller s Trace Formula for Systems with Discrete Symmetries

First-Order Corrections to Gutzwiller s Trace Formula for Systems with Discrete Symmetries c 26 Noninear Phenomena in Compex Systems First-Order Corrections to Gutzwier s Trace Formua for Systems with Discrete Symmetries Hoger Cartarius, Jörg Main, and Günter Wunner Institut für Theoretische

More information

XSAT of linear CNF formulas

XSAT of linear CNF formulas XSAT of inear CN formuas Bernd R. Schuh Dr. Bernd Schuh, D-50968 Kön, Germany; bernd.schuh@netcoogne.de eywords: compexity, XSAT, exact inear formua, -reguarity, -uniformity, NPcompeteness Abstract. Open

More information

Robust Sensitivity Analysis for Linear Programming with Ellipsoidal Perturbation

Robust Sensitivity Analysis for Linear Programming with Ellipsoidal Perturbation Robust Sensitivity Anaysis for Linear Programming with Eipsoida Perturbation Ruotian Gao and Wenxun Xing Department of Mathematica Sciences Tsinghua University, Beijing, China, 100084 September 27, 2017

More information

Componentwise Determination of the Interval Hull Solution for Linear Interval Parameter Systems

Componentwise Determination of the Interval Hull Solution for Linear Interval Parameter Systems Componentwise Determination of the Interva Hu Soution for Linear Interva Parameter Systems L. V. Koev Dept. of Theoretica Eectrotechnics, Facuty of Automatics, Technica University of Sofia, 1000 Sofia,

More information

BALANCING REGULAR MATRIX PENCILS

BALANCING REGULAR MATRIX PENCILS BALANCING REGULAR MATRIX PENCILS DAMIEN LEMONNIER AND PAUL VAN DOOREN Abstract. In this paper we present a new diagona baancing technique for reguar matrix pencis λb A, which aims at reducing the sensitivity

More information

Supplement of Limited-memory Common-directions Method for Distributed Optimization and its Application on Empirical Risk Minimization

Supplement of Limited-memory Common-directions Method for Distributed Optimization and its Application on Empirical Risk Minimization Suppement of Limited-memory Common-directions Method for Distributed Optimization and its Appication on Empirica Risk Minimization Ching-pei Lee Po-Wei Wang Weizhu Chen Chih-Jen Lin I Introduction In this

More information

Separation of Variables and a Spherical Shell with Surface Charge

Separation of Variables and a Spherical Shell with Surface Charge Separation of Variabes and a Spherica She with Surface Charge In cass we worked out the eectrostatic potentia due to a spherica she of radius R with a surface charge density σθ = σ cos θ. This cacuation

More information

Schedulability Analysis of Deferrable Scheduling Algorithms for Maintaining Real-Time Data Freshness

Schedulability Analysis of Deferrable Scheduling Algorithms for Maintaining Real-Time Data Freshness 1 Scheduabiity Anaysis of Deferrabe Scheduing Agorithms for Maintaining Rea-Time Data Freshness Song Han, Deji Chen, Ming Xiong, Kam-yiu Lam, Aoysius K. Mok, Krithi Ramamritham UT Austin, Emerson Process

More information

18-660: Numerical Methods for Engineering Design and Optimization

18-660: Numerical Methods for Engineering Design and Optimization 8-660: Numerica Methods for Engineering esign and Optimization in i epartment of ECE Carnegie Meon University Pittsburgh, PA 523 Side Overview Conjugate Gradient Method (Part 4) Pre-conditioning Noninear

More information

Algorithms to solve massively under-defined systems of multivariate quadratic equations

Algorithms to solve massively under-defined systems of multivariate quadratic equations Agorithms to sove massivey under-defined systems of mutivariate quadratic equations Yasufumi Hashimoto Abstract It is we known that the probem to sove a set of randomy chosen mutivariate quadratic equations

More information

MARKOV CHAINS AND MARKOV DECISION THEORY. Contents

MARKOV CHAINS AND MARKOV DECISION THEORY. Contents MARKOV CHAINS AND MARKOV DECISION THEORY ARINDRIMA DATTA Abstract. In this paper, we begin with a forma introduction to probabiity and expain the concept of random variabes and stochastic processes. After

More information

Primal and dual active-set methods for convex quadratic programming

Primal and dual active-set methods for convex quadratic programming Math. Program., Ser. A 216) 159:469 58 DOI 1.17/s117-15-966-2 FULL LENGTH PAPER Prima and dua active-set methods for convex quadratic programming Anders Forsgren 1 Phiip E. Gi 2 Eizabeth Wong 2 Received:

More information

Explicit overall risk minimization transductive bound

Explicit overall risk minimization transductive bound 1 Expicit overa risk minimization transductive bound Sergio Decherchi, Paoo Gastado, Sandro Ridea, Rodofo Zunino Dept. of Biophysica and Eectronic Engineering (DIBE), Genoa University Via Opera Pia 11a,

More information

Schedulability Analysis of Deferrable Scheduling Algorithms for Maintaining Real-Time Data Freshness

Schedulability Analysis of Deferrable Scheduling Algorithms for Maintaining Real-Time Data Freshness 1 Scheduabiity Anaysis of Deferrabe Scheduing Agorithms for Maintaining Rea- Data Freshness Song Han, Deji Chen, Ming Xiong, Kam-yiu Lam, Aoysius K. Mok, Krithi Ramamritham UT Austin, Emerson Process Management,

More information

Mat 1501 lecture notes, penultimate installment

Mat 1501 lecture notes, penultimate installment Mat 1501 ecture notes, penutimate instament 1. bounded variation: functions of a singe variabe optiona) I beieve that we wi not actuay use the materia in this section the point is mainy to motivate the

More information

4 1-D Boundary Value Problems Heat Equation

4 1-D Boundary Value Problems Heat Equation 4 -D Boundary Vaue Probems Heat Equation The main purpose of this chapter is to study boundary vaue probems for the heat equation on a finite rod a x b. u t (x, t = ku xx (x, t, a < x < b, t > u(x, = ϕ(x

More information

A proposed nonparametric mixture density estimation using B-spline functions

A proposed nonparametric mixture density estimation using B-spline functions A proposed nonparametric mixture density estimation using B-spine functions Atizez Hadrich a,b, Mourad Zribi a, Afif Masmoudi b a Laboratoire d Informatique Signa et Image de a Côte d Opae (LISIC-EA 4491),

More information

Statistical Learning Theory: A Primer

Statistical Learning Theory: A Primer Internationa Journa of Computer Vision 38(), 9 3, 2000 c 2000 uwer Academic Pubishers. Manufactured in The Netherands. Statistica Learning Theory: A Primer THEODOROS EVGENIOU, MASSIMILIANO PONTIL AND TOMASO

More information

The EM Algorithm applied to determining new limit points of Mahler measures

The EM Algorithm applied to determining new limit points of Mahler measures Contro and Cybernetics vo. 39 (2010) No. 4 The EM Agorithm appied to determining new imit points of Maher measures by Souad E Otmani, Georges Rhin and Jean-Marc Sac-Épée Université Pau Veraine-Metz, LMAM,

More information

Cryptanalysis of PKP: A New Approach

Cryptanalysis of PKP: A New Approach Cryptanaysis of PKP: A New Approach Éiane Jaumes and Antoine Joux DCSSI 18, rue du Dr. Zamenhoff F-92131 Issy-es-Mx Cedex France eiane.jaumes@wanadoo.fr Antoine.Joux@ens.fr Abstract. Quite recenty, in

More information

4 Separation of Variables

4 Separation of Variables 4 Separation of Variabes In this chapter we describe a cassica technique for constructing forma soutions to inear boundary vaue probems. The soution of three cassica (paraboic, hyperboic and eiptic) PDE

More information

A BUNDLE METHOD FOR A CLASS OF BILEVEL NONSMOOTH CONVEX MINIMIZATION PROBLEMS

A BUNDLE METHOD FOR A CLASS OF BILEVEL NONSMOOTH CONVEX MINIMIZATION PROBLEMS SIAM J. OPTIM. Vo. 18, No. 1, pp. 242 259 c 2007 Society for Industria and Appied Mathematics A BUNDLE METHOD FOR A CLASS OF BILEVEL NONSMOOTH CONVEX MINIMIZATION PROBLEMS MIKHAIL V. SOLODOV Abstract.

More information

C. Fourier Sine Series Overview

C. Fourier Sine Series Overview 12 PHILIP D. LOEWEN C. Fourier Sine Series Overview Let some constant > be given. The symboic form of the FSS Eigenvaue probem combines an ordinary differentia equation (ODE) on the interva (, ) with a

More information

VALIDATED CONTINUATION FOR EQUILIBRIA OF PDES

VALIDATED CONTINUATION FOR EQUILIBRIA OF PDES VALIDATED CONTINUATION FOR EQUILIBRIA OF PDES SARAH DAY, JEAN-PHILIPPE LESSARD, AND KONSTANTIN MISCHAIKOW Abstract. One of the most efficient methods for determining the equiibria of a continuous parameterized

More information

An explicit Jordan Decomposition of Companion matrices

An explicit Jordan Decomposition of Companion matrices An expicit Jordan Decomposition of Companion matrices Fermín S V Bazán Departamento de Matemática CFM UFSC 88040-900 Forianópois SC E-mai: fermin@mtmufscbr S Gratton CERFACS 42 Av Gaspard Coriois 31057

More information

Data Mining Technology for Failure Prognostic of Avionics

Data Mining Technology for Failure Prognostic of Avionics IEEE Transactions on Aerospace and Eectronic Systems. Voume 38, #, pp.388-403, 00. Data Mining Technoogy for Faiure Prognostic of Avionics V.A. Skormin, Binghamton University, Binghamton, NY, 1390, USA

More information

A Novel Learning Method for Elman Neural Network Using Local Search

A Novel Learning Method for Elman Neural Network Using Local Search Neura Information Processing Letters and Reviews Vo. 11, No. 8, August 2007 LETTER A Nove Learning Method for Eman Neura Networ Using Loca Search Facuty of Engineering, Toyama University, Gofuu 3190 Toyama

More information

Approximated MLC shape matrix decomposition with interleaf collision constraint

Approximated MLC shape matrix decomposition with interleaf collision constraint Approximated MLC shape matrix decomposition with intereaf coision constraint Thomas Kainowski Antje Kiese Abstract Shape matrix decomposition is a subprobem in radiation therapy panning. A given fuence

More information

Lecture 6: Moderately Large Deflection Theory of Beams

Lecture 6: Moderately Large Deflection Theory of Beams Structura Mechanics 2.8 Lecture 6 Semester Yr Lecture 6: Moderatey Large Defection Theory of Beams 6.1 Genera Formuation Compare to the cassica theory of beams with infinitesima deformation, the moderatey

More information

Fitting affine and orthogonal transformations between two sets of points

Fitting affine and orthogonal transformations between two sets of points Mathematica Communications 9(2004), 27-34 27 Fitting affine and orthogona transformations between two sets of points Hemuth Späth Abstract. Let two point sets P and Q be given in R n. We determine a transation

More information

Global Optimality Principles for Polynomial Optimization Problems over Box or Bivalent Constraints by Separable Polynomial Approximations

Global Optimality Principles for Polynomial Optimization Problems over Box or Bivalent Constraints by Separable Polynomial Approximations Goba Optimaity Principes for Poynomia Optimization Probems over Box or Bivaent Constraints by Separabe Poynomia Approximations V. Jeyakumar, G. Li and S. Srisatkunarajah Revised Version II: December 23,

More information

VALIDATED CONTINUATION FOR EQUILIBRIA OF PDES

VALIDATED CONTINUATION FOR EQUILIBRIA OF PDES SIAM J. NUMER. ANAL. Vo. 0, No. 0, pp. 000 000 c 200X Society for Industria and Appied Mathematics VALIDATED CONTINUATION FOR EQUILIBRIA OF PDES SARAH DAY, JEAN-PHILIPPE LESSARD, AND KONSTANTIN MISCHAIKOW

More information

Efficiently Generating Random Bits from Finite State Markov Chains

Efficiently Generating Random Bits from Finite State Markov Chains 1 Efficienty Generating Random Bits from Finite State Markov Chains Hongchao Zhou and Jehoshua Bruck, Feow, IEEE Abstract The probem of random number generation from an uncorreated random source (of unknown

More information

Partial permutation decoding for MacDonald codes

Partial permutation decoding for MacDonald codes Partia permutation decoding for MacDonad codes J.D. Key Department of Mathematics and Appied Mathematics University of the Western Cape 7535 Bevie, South Africa P. Seneviratne Department of Mathematics

More information

BASIC NOTIONS AND RESULTS IN TOPOLOGY. 1. Metric spaces. Sets with finite diameter are called bounded sets. For x X and r > 0 the set

BASIC NOTIONS AND RESULTS IN TOPOLOGY. 1. Metric spaces. Sets with finite diameter are called bounded sets. For x X and r > 0 the set BASIC NOTIONS AND RESULTS IN TOPOLOGY 1. Metric spaces A metric on a set X is a map d : X X R + with the properties: d(x, y) 0 and d(x, y) = 0 x = y, d(x, y) = d(y, x), d(x, y) d(x, z) + d(z, y), for a

More information

Bayesian Learning. You hear a which which could equally be Thanks or Tanks, which would you go with?

Bayesian Learning. You hear a which which could equally be Thanks or Tanks, which would you go with? Bayesian Learning A powerfu and growing approach in machine earning We use it in our own decision making a the time You hear a which which coud equay be Thanks or Tanks, which woud you go with? Combine

More information

17 Lecture 17: Recombination and Dark Matter Production

17 Lecture 17: Recombination and Dark Matter Production PYS 652: Astrophysics 88 17 Lecture 17: Recombination and Dark Matter Production New ideas pass through three periods: It can t be done. It probaby can be done, but it s not worth doing. I knew it was

More information

Power Control and Transmission Scheduling for Network Utility Maximization in Wireless Networks

Power Control and Transmission Scheduling for Network Utility Maximization in Wireless Networks ower Contro and Transmission Scheduing for Network Utiity Maximization in Wireess Networks Min Cao, Vivek Raghunathan, Stephen Hany, Vinod Sharma and. R. Kumar Abstract We consider a joint power contro

More information

FRIEZE GROUPS IN R 2

FRIEZE GROUPS IN R 2 FRIEZE GROUPS IN R 2 MAXWELL STOLARSKI Abstract. Focusing on the Eucidean pane under the Pythagorean Metric, our goa is to cassify the frieze groups, discrete subgroups of the set of isometries of the

More information

Approximated MLC shape matrix decomposition with interleaf collision constraint

Approximated MLC shape matrix decomposition with interleaf collision constraint Agorithmic Operations Research Vo.4 (29) 49 57 Approximated MLC shape matrix decomposition with intereaf coision constraint Antje Kiese and Thomas Kainowski Institut für Mathematik, Universität Rostock,

More information

Sequential Decoding of Polar Codes with Arbitrary Binary Kernel

Sequential Decoding of Polar Codes with Arbitrary Binary Kernel Sequentia Decoding of Poar Codes with Arbitrary Binary Kerne Vera Miosavskaya, Peter Trifonov Saint-Petersburg State Poytechnic University Emai: veram,petert}@dcn.icc.spbstu.ru Abstract The probem of efficient

More information

An Algorithm for Pruning Redundant Modules in Min-Max Modular Network

An Algorithm for Pruning Redundant Modules in Min-Max Modular Network An Agorithm for Pruning Redundant Modues in Min-Max Moduar Network Hui-Cheng Lian and Bao-Liang Lu Department of Computer Science and Engineering, Shanghai Jiao Tong University 1954 Hua Shan Rd., Shanghai

More information

A globally and R-linearly convergent hybrid HS and PRP method and its inexact version with applications

A globally and R-linearly convergent hybrid HS and PRP method and its inexact version with applications A globally and R-linearly convergent hybrid HS and PRP method and its inexact version with applications Weijun Zhou 28 October 20 Abstract A hybrid HS and PRP type conjugate gradient method for smooth

More information

A. Distribution of the test statistic

A. Distribution of the test statistic A. Distribution of the test statistic In the sequentia test, we first compute the test statistic from a mini-batch of size m. If a decision cannot be made with this statistic, we keep increasing the mini-batch

More information

$, (2.1) n="# #. (2.2)

$, (2.1) n=# #. (2.2) Chapter. Eectrostatic II Notes: Most of the materia presented in this chapter is taken from Jackson, Chap.,, and 4, and Di Bartoo, Chap... Mathematica Considerations.. The Fourier series and the Fourier

More information

STA 216 Project: Spline Approach to Discrete Survival Analysis

STA 216 Project: Spline Approach to Discrete Survival Analysis : Spine Approach to Discrete Surviva Anaysis November 4, 005 1 Introduction Athough continuous surviva anaysis differs much from the discrete surviva anaysis, there is certain ink between the two modeing

More information

Math 124B January 17, 2012

Math 124B January 17, 2012 Math 124B January 17, 212 Viktor Grigoryan 3 Fu Fourier series We saw in previous ectures how the Dirichet and Neumann boundary conditions ead to respectivey sine and cosine Fourier series of the initia

More information

On a geometrical approach in contact mechanics

On a geometrical approach in contact mechanics Institut für Mechanik On a geometrica approach in contact mechanics Aexander Konyukhov, Kar Schweizerhof Universität Karsruhe, Institut für Mechanik Institut für Mechanik Kaiserstr. 12, Geb. 20.30 76128

More information

Unit 48: Structural Behaviour and Detailing for Construction. Deflection of Beams

Unit 48: Structural Behaviour and Detailing for Construction. Deflection of Beams Unit 48: Structura Behaviour and Detaiing for Construction 4.1 Introduction Defection of Beams This topic investigates the deformation of beams as the direct effect of that bending tendency, which affects

More information

c 2007 Society for Industrial and Applied Mathematics

c 2007 Society for Industrial and Applied Mathematics SIAM REVIEW Vo. 49,No. 1,pp. 111 1 c 7 Society for Industria and Appied Mathematics Domino Waves C. J. Efthimiou M. D. Johnson Abstract. Motivated by a proposa of Daykin [Probem 71-19*, SIAM Rev., 13 (1971),

More information

6 Wave Equation on an Interval: Separation of Variables

6 Wave Equation on an Interval: Separation of Variables 6 Wave Equation on an Interva: Separation of Variabes 6.1 Dirichet Boundary Conditions Ref: Strauss, Chapter 4 We now use the separation of variabes technique to study the wave equation on a finite interva.

More information

Unconditional security of differential phase shift quantum key distribution

Unconditional security of differential phase shift quantum key distribution Unconditiona security of differentia phase shift quantum key distribution Kai Wen, Yoshihisa Yamamoto Ginzton Lab and Dept of Eectrica Engineering Stanford University Basic idea of DPS-QKD Protoco. Aice

More information

FRST Multivariate Statistics. Multivariate Discriminant Analysis (MDA)

FRST Multivariate Statistics. Multivariate Discriminant Analysis (MDA) 1 FRST 531 -- Mutivariate Statistics Mutivariate Discriminant Anaysis (MDA) Purpose: 1. To predict which group (Y) an observation beongs to based on the characteristics of p predictor (X) variabes, using

More information

u(x) s.t. px w x 0 Denote the solution to this problem by ˆx(p, x). In order to obtain ˆx we may simply solve the standard problem max x 0

u(x) s.t. px w x 0 Denote the solution to this problem by ˆx(p, x). In order to obtain ˆx we may simply solve the standard problem max x 0 Bocconi University PhD in Economics - Microeconomics I Prof M Messner Probem Set 4 - Soution Probem : If an individua has an endowment instead of a monetary income his weath depends on price eves In particuar,

More information

Regularization for Nonlinear Complementarity Problems

Regularization for Nonlinear Complementarity Problems Int. Journa of Math. Anaysis, Vo. 3, 2009, no. 34, 1683-1691 Reguarization for Noninear Compementarity Probems Nguyen Buong a and Nguyen Thi Thuy Hoa b a Vietnamse Academy of Science and Technoogy Institute

More information

Lecture 6 Povh Krane Enge Williams Properties of 2-nucleon potential

Lecture 6 Povh Krane Enge Williams Properties of 2-nucleon potential Lecture 6 Povh Krane Enge Wiiams Properties of -nuceon potentia 16.1 4.4 3.6 9.9 Meson Theory of Nucear potentia 4.5 3.11 9.10 I recommend Eisberg and Resnik notes as distributed Probems, Lecture 6 1 Consider

More information

THE REACHABILITY CONES OF ESSENTIALLY NONNEGATIVE MATRICES

THE REACHABILITY CONES OF ESSENTIALLY NONNEGATIVE MATRICES THE REACHABILITY CONES OF ESSENTIALLY NONNEGATIVE MATRICES by Michae Neumann Department of Mathematics, University of Connecticut, Storrs, CT 06269 3009 and Ronad J. Stern Department of Mathematics, Concordia

More information

Do Schools Matter for High Math Achievement? Evidence from the American Mathematics Competitions Glenn Ellison and Ashley Swanson Online Appendix

Do Schools Matter for High Math Achievement? Evidence from the American Mathematics Competitions Glenn Ellison and Ashley Swanson Online Appendix VOL. NO. DO SCHOOLS MATTER FOR HIGH MATH ACHIEVEMENT? 43 Do Schoos Matter for High Math Achievement? Evidence from the American Mathematics Competitions Genn Eison and Ashey Swanson Onine Appendix Appendix

More information

Turbo Codes. Coding and Communication Laboratory. Dept. of Electrical Engineering, National Chung Hsing University

Turbo Codes. Coding and Communication Laboratory. Dept. of Electrical Engineering, National Chung Hsing University Turbo Codes Coding and Communication Laboratory Dept. of Eectrica Engineering, Nationa Chung Hsing University Turbo codes 1 Chapter 12: Turbo Codes 1. Introduction 2. Turbo code encoder 3. Design of intereaver

More information

Pattern Frequency Sequences and Internal Zeros

Pattern Frequency Sequences and Internal Zeros Advances in Appied Mathematics 28, 395 420 (2002 doi:10.1006/aama.2001.0789, avaiabe onine at http://www.ideaibrary.com on Pattern Frequency Sequences and Interna Zeros Mikós Bóna Department of Mathematics,

More information

Moreau-Yosida Regularization for Grouped Tree Structure Learning

Moreau-Yosida Regularization for Grouped Tree Structure Learning Moreau-Yosida Reguarization for Grouped Tree Structure Learning Jun Liu Computer Science and Engineering Arizona State University J.Liu@asu.edu Jieping Ye Computer Science and Engineering Arizona State

More information

Semidefinite relaxation and Branch-and-Bound Algorithm for LPECs

Semidefinite relaxation and Branch-and-Bound Algorithm for LPECs Semidefinite reaxation and Branch-and-Bound Agorithm for LPECs Marcia H. C. Fampa Universidade Federa do Rio de Janeiro Instituto de Matemática e COPPE. Caixa Posta 68530 Rio de Janeiro RJ 21941-590 Brasi

More information

Asynchronous Control for Coupled Markov Decision Systems

Asynchronous Control for Coupled Markov Decision Systems INFORMATION THEORY WORKSHOP (ITW) 22 Asynchronous Contro for Couped Marov Decision Systems Michae J. Neey University of Southern Caifornia Abstract This paper considers optima contro for a coection of

More information

Week 6 Lectures, Math 6451, Tanveer

Week 6 Lectures, Math 6451, Tanveer Fourier Series Week 6 Lectures, Math 645, Tanveer In the context of separation of variabe to find soutions of PDEs, we encountered or and in other cases f(x = f(x = a 0 + f(x = a 0 + b n sin nπx { a n

More information

<C 2 2. λ 2 l. λ 1 l 1 < C 1

<C 2 2. λ 2 l. λ 1 l 1 < C 1 Teecommunication Network Contro and Management (EE E694) Prof. A. A. Lazar Notes for the ecture of 7/Feb/95 by Huayan Wang (this document was ast LaT E X-ed on May 9,995) Queueing Primer for Muticass Optima

More information

Tracking Control of Multiple Mobile Robots

Tracking Control of Multiple Mobile Robots Proceedings of the 2001 IEEE Internationa Conference on Robotics & Automation Seou, Korea May 21-26, 2001 Tracking Contro of Mutipe Mobie Robots A Case Study of Inter-Robot Coision-Free Probem Jurachart

More information

MATH 172: MOTIVATION FOR FOURIER SERIES: SEPARATION OF VARIABLES

MATH 172: MOTIVATION FOR FOURIER SERIES: SEPARATION OF VARIABLES MATH 172: MOTIVATION FOR FOURIER SERIES: SEPARATION OF VARIABLES Separation of variabes is a method to sove certain PDEs which have a warped product structure. First, on R n, a inear PDE of order m is

More information

Stochastic Automata Networks (SAN) - Modelling. and Evaluation. Paulo Fernandes 1. Brigitte Plateau 2. May 29, 1997

Stochastic Automata Networks (SAN) - Modelling. and Evaluation. Paulo Fernandes 1. Brigitte Plateau 2. May 29, 1997 Stochastic utomata etworks (S) - Modeing and Evauation Pauo Fernandes rigitte Pateau 2 May 29, 997 Institut ationa Poytechnique de Grenobe { IPG Ecoe ationae Superieure d'informatique et de Mathematiques

More information

BP neural network-based sports performance prediction model applied research

BP neural network-based sports performance prediction model applied research Avaiabe onine www.jocpr.com Journa of Chemica and Pharmaceutica Research, 204, 6(7:93-936 Research Artice ISSN : 0975-7384 CODEN(USA : JCPRC5 BP neura networ-based sports performance prediction mode appied

More information

LECTURE NOTES 9 TRACELESS SYMMETRIC TENSOR APPROACH TO LEGENDRE POLYNOMIALS AND SPHERICAL HARMONICS

LECTURE NOTES 9 TRACELESS SYMMETRIC TENSOR APPROACH TO LEGENDRE POLYNOMIALS AND SPHERICAL HARMONICS MASSACHUSETTS INSTITUTE OF TECHNOLOGY Physics Department Physics 8.07: Eectromagnetism II October 7, 202 Prof. Aan Guth LECTURE NOTES 9 TRACELESS SYMMETRIC TENSOR APPROACH TO LEGENDRE POLYNOMIALS AND SPHERICAL

More information

arxiv: v2 [cond-mat.stat-mech] 14 Nov 2008

arxiv: v2 [cond-mat.stat-mech] 14 Nov 2008 Random Booean Networks Barbara Drosse Institute of Condensed Matter Physics, Darmstadt University of Technoogy, Hochschustraße 6, 64289 Darmstadt, Germany (Dated: June 27) arxiv:76.335v2 [cond-mat.stat-mech]

More information

Theory and implementation behind: Universal surface creation - smallest unitcell

Theory and implementation behind: Universal surface creation - smallest unitcell Teory and impementation beind: Universa surface creation - smaest unitce Bjare Brin Buus, Jaob Howat & Tomas Bigaard September 15, 218 1 Construction of surface sabs Te aim for tis part of te project is

More information

Distributed average consensus: Beyond the realm of linearity

Distributed average consensus: Beyond the realm of linearity Distributed average consensus: Beyond the ream of inearity Usman A. Khan, Soummya Kar, and José M. F. Moura Department of Eectrica and Computer Engineering Carnegie Meon University 5 Forbes Ave, Pittsburgh,

More information

Melodic contour estimation with B-spline models using a MDL criterion

Melodic contour estimation with B-spline models using a MDL criterion Meodic contour estimation with B-spine modes using a MDL criterion Damien Loive, Ney Barbot, Oivier Boeffard IRISA / University of Rennes 1 - ENSSAT 6 rue de Kerampont, B.P. 80518, F-305 Lannion Cedex

More information

Integrating Factor Methods as Exponential Integrators

Integrating Factor Methods as Exponential Integrators Integrating Factor Methods as Exponentia Integrators Borisav V. Minchev Department of Mathematica Science, NTNU, 7491 Trondheim, Norway Borko.Minchev@ii.uib.no Abstract. Recenty a ot of effort has been

More information

8 APPENDIX. E[m M] = (n S )(1 exp( exp(s min + c M))) (19) E[m M] n exp(s min + c M) (20) 8.1 EMPIRICAL EVALUATION OF SAMPLING

8 APPENDIX. E[m M] = (n S )(1 exp( exp(s min + c M))) (19) E[m M] n exp(s min + c M) (20) 8.1 EMPIRICAL EVALUATION OF SAMPLING 8 APPENDIX 8.1 EMPIRICAL EVALUATION OF SAMPLING We wish to evauate the empirica accuracy of our samping technique on concrete exampes. We do this in two ways. First, we can sort the eements by probabiity

More information

Bourgain s Theorem. Computational and Metric Geometry. Instructor: Yury Makarychev. d(s 1, s 2 ).

Bourgain s Theorem. Computational and Metric Geometry. Instructor: Yury Makarychev. d(s 1, s 2 ). Bourgain s Theorem Computationa and Metric Geometry Instructor: Yury Makarychev 1 Notation Given a metric space (X, d) and S X, the distance from x X to S equas d(x, S) = inf d(x, s). s S The distance

More information

On Non-Optimally Expanding Sets in Grassmann Graphs

On Non-Optimally Expanding Sets in Grassmann Graphs ectronic Cooquium on Computationa Compexity, Report No. 94 (07) On Non-Optimay xpanding Sets in Grassmann Graphs Irit Dinur Subhash Khot Guy Kinder Dor Minzer Mui Safra Abstract The paper investigates

More information

NEW DEVELOPMENT OF OPTIMAL COMPUTING BUDGET ALLOCATION FOR DISCRETE EVENT SIMULATION

NEW DEVELOPMENT OF OPTIMAL COMPUTING BUDGET ALLOCATION FOR DISCRETE EVENT SIMULATION NEW DEVELOPMENT OF OPTIMAL COMPUTING BUDGET ALLOCATION FOR DISCRETE EVENT SIMULATION Hsiao-Chang Chen Dept. of Systems Engineering University of Pennsyvania Phiadephia, PA 904-635, U.S.A. Chun-Hung Chen

More information

arxiv: v1 [cs.db] 1 Aug 2012

arxiv: v1 [cs.db] 1 Aug 2012 Functiona Mechanism: Regression Anaysis under Differentia Privacy arxiv:208.029v [cs.db] Aug 202 Jun Zhang Zhenjie Zhang 2 Xiaokui Xiao Yin Yang 2 Marianne Winsett 2,3 ABSTRACT Schoo of Computer Engineering

More information

Approximation and Fast Calculation of Non-local Boundary Conditions for the Time-dependent Schrödinger Equation

Approximation and Fast Calculation of Non-local Boundary Conditions for the Time-dependent Schrödinger Equation Approximation and Fast Cacuation of Non-oca Boundary Conditions for the Time-dependent Schrödinger Equation Anton Arnod, Matthias Ehrhardt 2, and Ivan Sofronov 3 Universität Münster, Institut für Numerische

More information

MONOCHROMATIC LOOSE PATHS IN MULTICOLORED k-uniform CLIQUES

MONOCHROMATIC LOOSE PATHS IN MULTICOLORED k-uniform CLIQUES MONOCHROMATIC LOOSE PATHS IN MULTICOLORED k-uniform CLIQUES ANDRZEJ DUDEK AND ANDRZEJ RUCIŃSKI Abstract. For positive integers k and, a k-uniform hypergraph is caed a oose path of ength, and denoted by

More information

arxiv: v1 [math.fa] 23 Aug 2018

arxiv: v1 [math.fa] 23 Aug 2018 An Exact Upper Bound on the L p Lebesgue Constant and The -Rényi Entropy Power Inequaity for Integer Vaued Random Variabes arxiv:808.0773v [math.fa] 3 Aug 08 Peng Xu, Mokshay Madiman, James Mebourne Abstract

More information

Statistical Learning Theory: a Primer

Statistical Learning Theory: a Primer ??,??, 1 6 (??) c?? Kuwer Academic Pubishers, Boston. Manufactured in The Netherands. Statistica Learning Theory: a Primer THEODOROS EVGENIOU AND MASSIMILIANO PONTIL Center for Bioogica and Computationa

More information

Paper presented at the Workshop on Space Charge Physics in High Intensity Hadron Rings, sponsored by Brookhaven National Laboratory, May 4-7,1998

Paper presented at the Workshop on Space Charge Physics in High Intensity Hadron Rings, sponsored by Brookhaven National Laboratory, May 4-7,1998 Paper presented at the Workshop on Space Charge Physics in High ntensity Hadron Rings, sponsored by Brookhaven Nationa Laboratory, May 4-7,998 Noninear Sef Consistent High Resoution Beam Hao Agorithm in

More information

Math 124B January 31, 2012

Math 124B January 31, 2012 Math 124B January 31, 212 Viktor Grigoryan 7 Inhomogeneous boundary vaue probems Having studied the theory of Fourier series, with which we successfuy soved boundary vaue probems for the homogeneous heat

More information

Nonlinear Analysis of Spatial Trusses

Nonlinear Analysis of Spatial Trusses Noninear Anaysis of Spatia Trusses João Barrigó October 14 Abstract The present work addresses the noninear behavior of space trusses A formuation for geometrica noninear anaysis is presented, which incudes

More information

On the Goal Value of a Boolean Function

On the Goal Value of a Boolean Function On the Goa Vaue of a Booean Function Eric Bach Dept. of CS University of Wisconsin 1210 W. Dayton St. Madison, WI 53706 Lisa Heerstein Dept of CSE NYU Schoo of Engineering 2 Metrotech Center, 10th Foor

More information

An approximate method for solving the inverse scattering problem with fixed-energy data

An approximate method for solving the inverse scattering problem with fixed-energy data J. Inv. I-Posed Probems, Vo. 7, No. 6, pp. 561 571 (1999) c VSP 1999 An approximate method for soving the inverse scattering probem with fixed-energy data A. G. Ramm and W. Scheid Received May 12, 1999

More information

STABILITY OF A PARAMETRICALLY EXCITED DAMPED INVERTED PENDULUM 1. INTRODUCTION

STABILITY OF A PARAMETRICALLY EXCITED DAMPED INVERTED PENDULUM 1. INTRODUCTION Journa of Sound and Vibration (996) 98(5), 643 65 STABILITY OF A PARAMETRICALLY EXCITED DAMPED INVERTED PENDULUM G. ERDOS AND T. SINGH Department of Mechanica and Aerospace Engineering, SUNY at Buffao,

More information

Wave Equation Dirichlet Boundary Conditions

Wave Equation Dirichlet Boundary Conditions Wave Equation Dirichet Boundary Conditions u tt x, t = c u xx x, t, < x 1 u, t =, u, t = ux, = fx u t x, = gx Look for simpe soutions in the form ux, t = XxT t Substituting into 13 and dividing

More information

Noname manuscript No. (will be inserted by the editor) Can Li Ignacio E. Grossmann

Noname manuscript No. (will be inserted by the editor) Can Li Ignacio E. Grossmann Noname manuscript No. (wi be inserted by the editor) A finite ɛ-convergence agorithm for two-stage convex 0-1 mixed-integer noninear stochastic programs with mixed-integer first and second stage variabes

More information

King Fahd University of Petroleum & Minerals

King Fahd University of Petroleum & Minerals King Fahd University of Petroeum & Mineras DEPARTMENT OF MATHEMATICAL SCIENCES Technica Report Series TR 369 December 6 Genera decay of soutions of a viscoeastic equation Saim A. Messaoudi DHAHRAN 3161

More information

Effect of transport ratio on source term in determination of surface emission coefficient

Effect of transport ratio on source term in determination of surface emission coefficient Internationa Journa of heoretica & Appied Sciences, (): 74-78(9) ISSN : 975-78 Effect of transport ratio on source term in determination of surface emission coefficient Sanjeev Kumar and Apna Mishra epartment

More information